source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(d: dict) ->dict:
d_out = {}
for k, v in d.items():
if v in d_out:
d_out[v].append(k)
else:
d_out[v] = [k]
return d_out
| group_dict_by_value | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(d: dict) ->dict:\n d_out = {}\n for k, v in d.items():\n if v in d_out:\n d_out[v].append(k)\n else:\n d_out[v] = [k]\n return d_out\n",
"role": "user"
},
{
"content": "group_dict_by_value",
"role": "assistant"
}
] |
def x(code: str, exclude_variable: Union[set, None]=None,
jsonable_parameter: bool=True) ->tuple:
if exclude_variable is None:
exclude_variable = set()
else:
exclude_variable = copy.deepcopy(exclude_variable)
root = ast.parse(code)
store_variable_name = set()
assign_only = True
dict_parameter = {}
for node in ast.iter_child_nodes(root):
if isinstance(node, ast.Assign):
for assign_node in ast.walk(node):
if isinstance(assign_node, ast.Name):
if isinstance(assign_node.ctx, ast.Store):
if jsonable_parameter is False:
store_variable_name |= {assign_node.id}
else:
exclude_variable |= {assign_node.id}
_is_literal_eval, _value = is_literal_eval(node.value)
if jsonable_parameter is True:
for assign_node in ast.iter_child_nodes(node):
if isinstance(assign_node, ast.Tuple):
i = 0
for assign_tuple_node in ast.iter_child_nodes(
assign_node):
if isinstance(assign_tuple_node, ast.Name):
if isinstance(_value, collections.Iterable
) and is_jsonable(_value[i]
) and _is_literal_eval:
dict_parameter[assign_tuple_node.id
] = _value[i]
store_variable_name |= {assign_tuple_node
.id}
else:
exclude_variable |= {assign_tuple_node.id}
i += 1
elif isinstance(assign_node, ast.Name):
if is_jsonable(_value) and _is_literal_eval:
dict_parameter[assign_node.id] = _value
store_variable_name |= {assign_node.id}
else:
exclude_variable |= {assign_node.id}
elif isinstance(node, ast.AugAssign):
for assign_node in ast.walk(node):
if isinstance(assign_node, ast.Name):
exclude_variable |= {assign_node.id}
elif isinstance(node, (ast.FunctionDef, ast.ClassDef)):
assign_only = False
exclude_variable |= {node.name}
elif isinstance(node, ast.Import):
assign_only = False
for node1 in ast.iter_child_nodes(node):
if node1.asname is not None:
exclude_variable |= {node1.asname}
else:
exclude_variable |= {node1.name}
elif isinstance(node, ast.ImportFrom):
assign_only = False
for node1 in ast.iter_child_nodes(node):
if node1.asname is not None:
exclude_variable |= {node1.asname}
else:
exclude_variable |= {node1.name}
else:
assign_only = False
if assign_only is True:
possible_parameter = store_variable_name - exclude_variable
if jsonable_parameter is True:
dict_parameter = {k: dict_parameter[k] for k in possible_parameter}
return (possible_parameter, store_variable_name | exclude_variable,
dict_parameter)
return set(), store_variable_name | exclude_variable, {}
| variable_status | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(code: str, exclude_variable: Union[set, None]=None,\n jsonable_parameter: bool=True) ->tuple:\n if exclude_variable is None:\n exclude_variable = set()\n else:\n exclude_variable = copy.deepcopy(exclude_variable)\n root = ast.parse(code)\n store_variable_name = set()\n assign_only = True\n dict_parameter = {}\n for node in ast.iter_child_nodes(root):\n if isinstance(node, ast.Assign):\n for assign_node in ast.walk(node):\n if isinstance(assign_node, ast.Name):\n if isinstance(assign_node.ctx, ast.Store):\n if jsonable_parameter is False:\n store_variable_name |= {assign_node.id}\n else:\n exclude_variable |= {assign_node.id}\n _is_literal_eval, _value = is_literal_eval(node.value)\n if jsonable_parameter is True:\n for assign_node in ast.iter_child_nodes(node):\n if isinstance(assign_node, ast.Tuple):\n i = 0\n for assign_tuple_node in ast.iter_child_nodes(\n assign_node):\n if isinstance(assign_tuple_node, ast.Name):\n if isinstance(_value, collections.Iterable\n ) and is_jsonable(_value[i]\n ) and _is_literal_eval:\n dict_parameter[assign_tuple_node.id\n ] = _value[i]\n store_variable_name |= {assign_tuple_node\n .id}\n else:\n exclude_variable |= {assign_tuple_node.id}\n i += 1\n elif isinstance(assign_node, ast.Name):\n if is_jsonable(_value) and _is_literal_eval:\n dict_parameter[assign_node.id] = _value\n store_variable_name |= {assign_node.id}\n else:\n exclude_variable |= {assign_node.id}\n elif isinstance(node, ast.AugAssign):\n for assign_node in ast.walk(node):\n if isinstance(assign_node, ast.Name):\n exclude_variable |= {assign_node.id}\n elif isinstance(node, (ast.FunctionDef, ast.ClassDef)):\n assign_only = False\n exclude_variable |= {node.name}\n elif isinstance(node, ast.Import):\n assign_only = False\n for node1 in ast.iter_child_nodes(node):\n if node1.asname is not None:\n exclude_variable |= {node1.asname}\n else:\n exclude_variable |= {node1.name}\n elif isinstance(node, ast.ImportFrom):\n assign_only = False\n for node1 in ast.iter_child_nodes(node):\n if node1.asname is not None:\n exclude_variable |= {node1.asname}\n else:\n exclude_variable |= {node1.name}\n else:\n assign_only = False\n if assign_only is True:\n possible_parameter = store_variable_name - exclude_variable\n if jsonable_parameter is True:\n dict_parameter = {k: dict_parameter[k] for k in possible_parameter}\n return (possible_parameter, store_variable_name | exclude_variable,\n dict_parameter)\n return set(), store_variable_name | exclude_variable, {}\n",
"role": "user"
},
{
"content": "variable_status",
"role": "assistant"
}
] |
def x(name: str, start_marker: str=' (', end_marker: str=')') ->str:
if start_marker == '':
raise ValueError('start_marker can not be the empty string.')
a = name
start = len(a) - a[::-1].find(start_marker[::-1])
if a[len(a) - len(end_marker):len(a)] == end_marker and start < len(a
) - len(end_marker) and a[start - len(start_marker):start
] == start_marker and a[start:len(a) - len(end_marker)].isdigit():
old_int = int(a[start:len(a) - len(end_marker)])
new_int = old_int + 1
new_name = a[:start] + str(new_int) + end_marker
else:
new_name = a + start_marker + '1' + end_marker
return new_name
| increment_name | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(name: str, start_marker: str=' (', end_marker: str=')') ->str:\n if start_marker == '':\n raise ValueError('start_marker can not be the empty string.')\n a = name\n start = len(a) - a[::-1].find(start_marker[::-1])\n if a[len(a) - len(end_marker):len(a)] == end_marker and start < len(a\n ) - len(end_marker) and a[start - len(start_marker):start\n ] == start_marker and a[start:len(a) - len(end_marker)].isdigit():\n old_int = int(a[start:len(a) - len(end_marker)])\n new_int = old_int + 1\n new_name = a[:start] + str(new_int) + end_marker\n else:\n new_name = a + start_marker + '1' + end_marker\n return new_name\n",
"role": "user"
},
{
"content": "increment_name",
"role": "assistant"
}
] |
def x(identifier, configuration=None):
resourceview = ResourceView(configuration=configuration)
result = resourceview._load_from_hdx('resource view', identifier)
if result:
return resourceview
return None
| read_from_hdx | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(identifier, configuration=None):\n resourceview = ResourceView(configuration=configuration)\n result = resourceview._load_from_hdx('resource view', identifier)\n if result:\n return resourceview\n return None\n",
"role": "user"
},
{
"content": "read_from_hdx",
"role": "assistant"
}
] |
def x(identifier, configuration=None):
resourceview = ResourceView(configuration=configuration)
success, result = resourceview._read_from_hdx('resource view',
identifier, 'id', ResourceView.actions()['list'])
resourceviews = list()
if success:
for resourceviewdict in result:
resourceview = ResourceView(resourceviewdict, configuration=
configuration)
resourceviews.append(resourceview)
return resourceviews
| get_all_for_resource | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(identifier, configuration=None):\n resourceview = ResourceView(configuration=configuration)\n success, result = resourceview._read_from_hdx('resource view',\n identifier, 'id', ResourceView.actions()['list'])\n resourceviews = list()\n if success:\n for resourceviewdict in result:\n resourceview = ResourceView(resourceviewdict, configuration=\n configuration)\n resourceviews.append(resourceview)\n return resourceviews\n",
"role": "user"
},
{
"content": "get_all_for_resource",
"role": "assistant"
}
] |
def x(self, log=False):
update = False
if 'id' in self.data and self._load_from_hdx('resource view', self.data
['id']):
update = True
elif 'resource_id' in self.data:
resource_views = self.get_all_for_resource(self.data['resource_id'])
for resource_view in resource_views:
if self.data['title'] == resource_view['title']:
self.old_data = self.data
self.data = resource_view.data
update = True
break
if update:
if log:
logger.warning('resource view exists. Updating %s' % self.data[
'id'])
self._merge_hdx_update('resource view', 'id')
return update
| update_resource_view | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, log=False):\n update = False\n if 'id' in self.data and self._load_from_hdx('resource view', self.data\n ['id']):\n update = True\n elif 'resource_id' in self.data:\n resource_views = self.get_all_for_resource(self.data['resource_id'])\n for resource_view in resource_views:\n if self.data['title'] == resource_view['title']:\n self.old_data = self.data\n self.data = resource_view.data\n update = True\n break\n if update:\n if log:\n logger.warning('resource view exists. Updating %s' % self.data[\n 'id'])\n self._merge_hdx_update('resource view', 'id')\n return update\n",
"role": "user"
},
{
"content": "update_resource_view",
"role": "assistant"
}
] |
def x(self):
self.check_required_fields()
if not self._update_resource_view(log=True):
self._save_to_hdx('create', 'title')
| create_in_hdx | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.check_required_fields()\n if not self._update_resource_view(log=True):\n self._save_to_hdx('create', 'title')\n",
"role": "user"
},
{
"content": "create_in_hdx",
"role": "assistant"
}
] |
def x(self, resource_view):
if isinstance(resource_view, str):
if is_valid_uuid(resource_view) is False:
raise HDXError('%s is not a valid resource view id!' %
resource_view)
resource_view = ResourceView.read_from_hdx(resource_view)
if not isinstance(resource_view, dict) and not isinstance(resource_view,
ResourceView):
raise HDXError('%s is not a valid resource view!' % resource_view)
for key in resource_view:
if key not in ('id', 'resource_id', 'package_id'):
self.data[key] = resource_view[key]
| copy | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, resource_view):\n if isinstance(resource_view, str):\n if is_valid_uuid(resource_view) is False:\n raise HDXError('%s is not a valid resource view id!' %\n resource_view)\n resource_view = ResourceView.read_from_hdx(resource_view)\n if not isinstance(resource_view, dict) and not isinstance(resource_view,\n ResourceView):\n raise HDXError('%s is not a valid resource view!' % resource_view)\n for key in resource_view:\n if key not in ('id', 'resource_id', 'package_id'):\n self.data[key] = resource_view[key]\n",
"role": "user"
},
{
"content": "copy",
"role": "assistant"
}
] |
def x(configuration=None, url=None, keycolumn=5, failchained=True):
if not Tags._tags_dict:
if configuration is None:
configuration = Configuration.read()
with Download(full_agent=configuration.get_user_agent()) as downloader:
if url is None:
url = configuration['tags_cleanup_url']
Tags._tags_dict = downloader.download_tabular_rows_as_dicts(url,
keycolumn=keycolumn)
keys = Tags._tags_dict.keys()
chainerror = False
for i, tag in enumerate(keys):
whattodo = Tags._tags_dict[tag]
action = whattodo[u'action']
final_tags = whattodo[u'final tags (semicolon separated)']
for final_tag in final_tags.split(';'):
if final_tag in keys:
index = list(keys).index(final_tag)
if index != i:
whattodo2 = Tags._tags_dict[final_tag]
action2 = whattodo2[u'action']
if action2 != 'OK' and action2 != 'Other':
final_tags2 = whattodo2[
u'final tags (semicolon separated)']
if final_tag not in final_tags2.split(';'):
chainerror = True
if failchained:
logger.error(
'Chained rules: %s (%s -> %s) | %s (%s -> %s)'
% (action, tag, final_tags,
action2, final_tag, final_tags2))
if failchained and chainerror:
raise ChainRuleError('Chained rules for tags detected!')
Tags._wildcard_tags = list()
for tag in Tags._tags_dict:
if '*' in tag:
Tags._wildcard_tags.append(tag)
return Tags._tags_dict, Tags._wildcard_tags
| tagscleanupdicts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(configuration=None, url=None, keycolumn=5, failchained=True):\n if not Tags._tags_dict:\n if configuration is None:\n configuration = Configuration.read()\n with Download(full_agent=configuration.get_user_agent()) as downloader:\n if url is None:\n url = configuration['tags_cleanup_url']\n Tags._tags_dict = downloader.download_tabular_rows_as_dicts(url,\n keycolumn=keycolumn)\n keys = Tags._tags_dict.keys()\n chainerror = False\n for i, tag in enumerate(keys):\n whattodo = Tags._tags_dict[tag]\n action = whattodo[u'action']\n final_tags = whattodo[u'final tags (semicolon separated)']\n for final_tag in final_tags.split(';'):\n if final_tag in keys:\n index = list(keys).index(final_tag)\n if index != i:\n whattodo2 = Tags._tags_dict[final_tag]\n action2 = whattodo2[u'action']\n if action2 != 'OK' and action2 != 'Other':\n final_tags2 = whattodo2[\n u'final tags (semicolon separated)']\n if final_tag not in final_tags2.split(';'):\n chainerror = True\n if failchained:\n logger.error(\n 'Chained rules: %s (%s -> %s) | %s (%s -> %s)'\n % (action, tag, final_tags,\n action2, final_tag, final_tags2))\n if failchained and chainerror:\n raise ChainRuleError('Chained rules for tags detected!')\n Tags._wildcard_tags = list()\n for tag in Tags._tags_dict:\n if '*' in tag:\n Tags._wildcard_tags.append(tag)\n return Tags._tags_dict, Tags._wildcard_tags\n",
"role": "user"
},
{
"content": "tagscleanupdicts",
"role": "assistant"
}
] |
def x(identifier, configuration=None):
user = User(configuration=configuration)
result = user._load_from_hdx('user', identifier)
if result:
return user
return None
| read_from_hdx | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(identifier, configuration=None):\n user = User(configuration=configuration)\n result = user._load_from_hdx('user', identifier)\n if result:\n return user\n return None\n",
"role": "user"
},
{
"content": "read_from_hdx",
"role": "assistant"
}
] |
def x(self):
capacity = self.data.get('capacity')
if capacity is not None:
del self.data['capacity']
self._update_in_hdx('user', 'id')
if capacity is not None:
self.data['capacity'] = capacity
| update_in_hdx | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n capacity = self.data.get('capacity')\n if capacity is not None:\n del self.data['capacity']\n self._update_in_hdx('user', 'id')\n if capacity is not None:\n self.data['capacity'] = capacity\n",
"role": "user"
},
{
"content": "update_in_hdx",
"role": "assistant"
}
] |
def x(self):
capacity = self.data.get('capacity')
if capacity is not None:
del self.data['capacity']
self._create_in_hdx('user', 'id', 'name')
if capacity is not None:
self.data['capacity'] = capacity
| create_in_hdx | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n capacity = self.data.get('capacity')\n if capacity is not None:\n del self.data['capacity']\n self._create_in_hdx('user', 'id', 'name')\n if capacity is not None:\n self.data['capacity'] = capacity\n",
"role": "user"
},
{
"content": "create_in_hdx",
"role": "assistant"
}
] |
def x(self, subject, text_body, html_body=None, sender=None, **kwargs):
self.configuration.emailer().send([self.data['email']], subject,
text_body, html_body=html_body, sender=sender, **kwargs)
| email | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, subject, text_body, html_body=None, sender=None, **kwargs):\n self.configuration.emailer().send([self.data['email']], subject,\n text_body, html_body=html_body, sender=sender, **kwargs)\n",
"role": "user"
},
{
"content": "email",
"role": "assistant"
}
] |
def x(configuration=None, **kwargs):
user = User(configuration=configuration)
user['id'] = 'all users'
result = user._write_to_hdx('list', kwargs, 'id')
users = list()
if result:
for userdict in result:
user = User(userdict, configuration=configuration)
users.append(user)
else:
logger.debug(result)
return users
| get_all_users | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(configuration=None, **kwargs):\n user = User(configuration=configuration)\n user['id'] = 'all users'\n result = user._write_to_hdx('list', kwargs, 'id')\n users = list()\n if result:\n for userdict in result:\n user = User(userdict, configuration=configuration)\n users.append(user)\n else:\n logger.debug(result)\n return users\n",
"role": "user"
},
{
"content": "get_all_users",
"role": "assistant"
}
] |
def x(users, subject, text_body, html_body=None, sender=None, configuration
=None, **kwargs):
if not users:
raise ValueError('No users supplied')
recipients = list()
for user in users:
recipients.append(user.data['email'])
if configuration is None:
configuration = users[0].configuration
configuration.emailer().send(recipients, subject, text_body, html_body=
html_body, sender=sender, **kwargs)
| email_users | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(users, subject, text_body, html_body=None, sender=None, configuration\n =None, **kwargs):\n if not users:\n raise ValueError('No users supplied')\n recipients = list()\n for user in users:\n recipients.append(user.data['email'])\n if configuration is None:\n configuration = users[0].configuration\n configuration.emailer().send(recipients, subject, text_body, html_body=\n html_body, sender=sender, **kwargs)\n",
"role": "user"
},
{
"content": "email_users",
"role": "assistant"
}
] |
def x(self, permission='read'):
success, result = self._read_from_hdx('user', self.data['name'], 'id',
self.actions()['listorgs'], permission=permission)
organizations = list()
if success:
for organizationdict in result:
organization = hdx.data.organization.Organization.read_from_hdx(
organizationdict['id'])
organizations.append(organization)
return organizations
| get_organizations | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, permission='read'):\n success, result = self._read_from_hdx('user', self.data['name'], 'id',\n self.actions()['listorgs'], permission=permission)\n organizations = list()\n if success:\n for organizationdict in result:\n organization = hdx.data.organization.Organization.read_from_hdx(\n organizationdict['id'])\n organizations.append(organization)\n return organizations\n",
"role": "user"
},
{
"content": "get_organizations",
"role": "assistant"
}
] |
def x(projectmainfn, **kwargs):
site_url = Configuration._create(**kwargs)
logger.info('--------------------------------------------------')
logger.info('> Using HDX Python API Library %s' % Configuration.apiversion)
logger.info('> HDX Site: %s' % site_url)
UserAgent.user_agent = Configuration.read().user_agent
projectmainfn()
| facade | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(projectmainfn, **kwargs):\n site_url = Configuration._create(**kwargs)\n logger.info('--------------------------------------------------')\n logger.info('> Using HDX Python API Library %s' % Configuration.apiversion)\n logger.info('> HDX Site: %s' % site_url)\n UserAgent.user_agent = Configuration.read().user_agent\n projectmainfn()\n",
"role": "user"
},
{
"content": "facade",
"role": "assistant"
}
] |
def x(config_path=None):
if config_path:
config = LintConfig.load_from_file(config_path)
click.echo('Using config from {0}'.format(config_path))
elif os.path.exists(DEFAULT_CONFIG_FILE):
config = LintConfig.load_from_file(DEFAULT_CONFIG_FILE)
click.echo('Using config from {0}'.format(DEFAULT_CONFIG_FILE))
else:
config = LintConfig()
return config
| get_lint_config | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(config_path=None):\n if config_path:\n config = LintConfig.load_from_file(config_path)\n click.echo('Using config from {0}'.format(config_path))\n elif os.path.exists(DEFAULT_CONFIG_FILE):\n config = LintConfig.load_from_file(DEFAULT_CONFIG_FILE)\n click.echo('Using config from {0}'.format(DEFAULT_CONFIG_FILE))\n else:\n config = LintConfig()\n return config\n",
"role": "user"
},
{
"content": "get_lint_config",
"role": "assistant"
}
] |
def x(list_files, config, ignore, path):
files = MarkdownFileFinder.find_files(path)
if list_files:
echo_files(files)
lint_config = get_lint_config(config)
lint_config.apply_on_csv_string(ignore, lint_config.disable_rule)
linter = MarkdownLinter(lint_config)
error_count = linter.lint_files(files)
exit(error_count)
| cli | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(list_files, config, ignore, path):\n files = MarkdownFileFinder.find_files(path)\n if list_files:\n echo_files(files)\n lint_config = get_lint_config(config)\n lint_config.apply_on_csv_string(ignore, lint_config.disable_rule)\n linter = MarkdownLinter(lint_config)\n error_count = linter.lint_files(files)\n exit(error_count)\n",
"role": "user"
},
{
"content": "cli",
"role": "assistant"
}
] |
def x(self, check_interval=300):
while True:
if args.config:
config = config_file_parser.get_configuration(args.config)
access_key_id = config['access-key-id']
secret_access_key = config['secret-access-key']
region = config['region']
else:
access_key_id = args.access_key_id
secret_access_key = args.secret_access_key
region = args.region
connection = connection_manager.connect_to_ec2(region,
access_key_id, secret_access_key)
snapshot_manager.run(connection)
logger.info('Sleeping {} seconds until next check'.format(
check_interval))
time.sleep(check_interval)
| run | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, check_interval=300):\n while True:\n if args.config:\n config = config_file_parser.get_configuration(args.config)\n access_key_id = config['access-key-id']\n secret_access_key = config['secret-access-key']\n region = config['region']\n else:\n access_key_id = args.access_key_id\n secret_access_key = args.secret_access_key\n region = args.region\n connection = connection_manager.connect_to_ec2(region,\n access_key_id, secret_access_key)\n snapshot_manager.run(connection)\n logger.info('Sleeping {} seconds until next check'.format(\n check_interval))\n time.sleep(check_interval)\n",
"role": "user"
},
{
"content": "run",
"role": "assistant"
}
] |
def x(self, markdown_string):
all_violations = []
lines = markdown_string.split('\n')
line_rules = self.line_rules
line_nr = 1
ignoring = False
for line in lines:
if ignoring:
if line.strip() == '<!-- markdownlint:enable -->':
ignoring = False
else:
if line.strip() == '<!-- markdownlint:disable -->':
ignoring = True
continue
for rule in line_rules:
violation = rule.validate(line)
if violation:
violation.line_nr = line_nr
all_violations.append(violation)
line_nr += 1
return all_violations
| apply_line_rules | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, markdown_string):\n all_violations = []\n lines = markdown_string.split('\\n')\n line_rules = self.line_rules\n line_nr = 1\n ignoring = False\n for line in lines:\n if ignoring:\n if line.strip() == '<!-- markdownlint:enable -->':\n ignoring = False\n else:\n if line.strip() == '<!-- markdownlint:disable -->':\n ignoring = True\n continue\n for rule in line_rules:\n violation = rule.validate(line)\n if violation:\n violation.line_nr = line_nr\n all_violations.append(violation)\n line_nr += 1\n return all_violations\n",
"role": "user"
},
{
"content": "apply_line_rules",
"role": "assistant"
}
] |
def x(self, files):
all_violations = []
for filename in files:
with open(filename, 'r') as f:
content = f.read()
violations = self.lint(content)
all_violations.extend(violations)
for e in violations:
print('{0}:{1}: {2} {3}'.format(filename, e.line_nr, e.
rule_id, e.message))
return len(all_violations)
| lint_files | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, files):\n all_violations = []\n for filename in files:\n with open(filename, 'r') as f:\n content = f.read()\n violations = self.lint(content)\n all_violations.extend(violations)\n for e in violations:\n print('{0}:{1}: {2} {3}'.format(filename, e.line_nr, e.\n rule_id, e.message))\n return len(all_violations)\n",
"role": "user"
},
{
"content": "lint_files",
"role": "assistant"
}
] |
def x(self, byte_stream):
try:
return self._struct.unpack_from(byte_stream)
except (TypeError, struct.error) as exception:
raise IOError('Unable to read byte stream with error: {0!s}'.format
(exception))
| ReadFrom | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, byte_stream):\n try:\n return self._struct.unpack_from(byte_stream)\n except (TypeError, struct.error) as exception:\n raise IOError('Unable to read byte stream with error: {0!s}'.format\n (exception))\n",
"role": "user"
},
{
"content": "ReadFrom",
"role": "assistant"
}
] |
def x(self, values):
try:
return self._struct.pack(*values)
except (TypeError, struct.error) as exception:
raise IOError('Unable to write stream with error: {0!s}'.format(
exception))
| WriteTo | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, values):\n try:\n return self._struct.pack(*values)\n except (TypeError, struct.error) as exception:\n raise IOError('Unable to write stream with error: {0!s}'.format(\n exception))\n",
"role": "user"
},
{
"content": "WriteTo",
"role": "assistant"
}
] |
def x(connection):
volumes = volume_manager.get_watched_volumes(connection)
for volume in volumes:
_ensure_snapshot(connection, volume)
_remove_old_snapshots(connection, volume)
| run | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(connection):\n volumes = volume_manager.get_watched_volumes(connection)\n for volume in volumes:\n _ensure_snapshot(connection, volume)\n _remove_old_snapshots(connection, volume)\n",
"role": "user"
},
{
"content": "run",
"role": "assistant"
}
] |
def x(volume):
logger.info('Creating new snapshot for {}'.format(volume.id))
snapshot = volume.create_snapshot(description=
'Automatic snapshot by Automated EBS Snapshots')
logger.info('Created snapshot {} for volume {}'.format(snapshot.id,
volume.id))
return snapshot
| create_snapshot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(volume):\n logger.info('Creating new snapshot for {}'.format(volume.id))\n snapshot = volume.create_snapshot(description=\n 'Automatic snapshot by Automated EBS Snapshots')\n logger.info('Created snapshot {} for volume {}'.format(snapshot.id,\n volume.id))\n return snapshot\n",
"role": "user"
},
{
"content": "create_snapshot",
"role": "assistant"
}
] |
def x(connection, volume):
if 'AutomatedEBSSnapshots' not in volume.tags:
logger.warning('Missing tag AutomatedEBSSnapshots for volume {}'.
format(volume.id))
return
interval = volume.tags['AutomatedEBSSnapshots']
if volume.tags['AutomatedEBSSnapshots'] not in VALID_INTERVALS:
logger.warning(
'"{}" is not a valid snapshotting interval for volume {}'.
format(interval, volume.id))
return
snapshots = connection.get_all_snapshots(filters={'volume-id': volume.id})
if not snapshots:
_create_snapshot(volume)
return
min_delta = 3600 * 24 * 365 * 10
for snapshot in snapshots:
timestamp = datetime.datetime.strptime(snapshot.start_time,
'%Y-%m-%dT%H:%M:%S.000Z')
delta_seconds = int((datetime.datetime.utcnow() - timestamp).
total_seconds())
if delta_seconds < min_delta:
min_delta = delta_seconds
logger.info('The newest snapshot for {} is {} seconds old'.format(
volume.id, min_delta))
if interval == 'hourly' and min_delta > 3600:
_create_snapshot(volume)
elif interval == 'daily' and min_delta > 3600 * 24:
_create_snapshot(volume)
elif interval == 'weekly' and min_delta > 3600 * 24 * 7:
_create_snapshot(volume)
elif interval == 'monthly' and min_delta > 3600 * 24 * 30:
_create_snapshot(volume)
elif interval == 'yearly' and min_delta > 3600 * 24 * 365:
_create_snapshot(volume)
else:
logger.info('No need for a new snapshot of {}'.format(volume.id))
| ensure_snapshot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(connection, volume):\n if 'AutomatedEBSSnapshots' not in volume.tags:\n logger.warning('Missing tag AutomatedEBSSnapshots for volume {}'.\n format(volume.id))\n return\n interval = volume.tags['AutomatedEBSSnapshots']\n if volume.tags['AutomatedEBSSnapshots'] not in VALID_INTERVALS:\n logger.warning(\n '\"{}\" is not a valid snapshotting interval for volume {}'.\n format(interval, volume.id))\n return\n snapshots = connection.get_all_snapshots(filters={'volume-id': volume.id})\n if not snapshots:\n _create_snapshot(volume)\n return\n min_delta = 3600 * 24 * 365 * 10\n for snapshot in snapshots:\n timestamp = datetime.datetime.strptime(snapshot.start_time,\n '%Y-%m-%dT%H:%M:%S.000Z')\n delta_seconds = int((datetime.datetime.utcnow() - timestamp).\n total_seconds())\n if delta_seconds < min_delta:\n min_delta = delta_seconds\n logger.info('The newest snapshot for {} is {} seconds old'.format(\n volume.id, min_delta))\n if interval == 'hourly' and min_delta > 3600:\n _create_snapshot(volume)\n elif interval == 'daily' and min_delta > 3600 * 24:\n _create_snapshot(volume)\n elif interval == 'weekly' and min_delta > 3600 * 24 * 7:\n _create_snapshot(volume)\n elif interval == 'monthly' and min_delta > 3600 * 24 * 30:\n _create_snapshot(volume)\n elif interval == 'yearly' and min_delta > 3600 * 24 * 365:\n _create_snapshot(volume)\n else:\n logger.info('No need for a new snapshot of {}'.format(volume.id))\n",
"role": "user"
},
{
"content": "ensure_snapshot",
"role": "assistant"
}
] |
def x(connection, volume):
if 'AutomatedEBSSnapshotsRetention' not in volume.tags:
logger.warning(
'Missing tag AutomatedEBSSnapshotsRetention for volume {}'.
format(volume.id))
return
retention = int(volume.tags['AutomatedEBSSnapshotsRetention'])
snapshots = connection.get_all_snapshots(filters={'volume-id': volume.id})
snapshots.sort(key=lambda x: x.start_time)
snapshots = snapshots[:-int(retention)]
if not snapshots:
logger.info('No old snapshots to remove')
return
for snapshot in snapshots:
logger.info('Deleting snapshot {}'.format(snapshot.id))
try:
snapshot.delete()
except EC2ResponseError as error:
logger.warning('Could not remove snapshot: {}'.format(error.
message))
logger.info('Done deleting snapshots')
| remove_old_snapshots | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(connection, volume):\n if 'AutomatedEBSSnapshotsRetention' not in volume.tags:\n logger.warning(\n 'Missing tag AutomatedEBSSnapshotsRetention for volume {}'.\n format(volume.id))\n return\n retention = int(volume.tags['AutomatedEBSSnapshotsRetention'])\n snapshots = connection.get_all_snapshots(filters={'volume-id': volume.id})\n snapshots.sort(key=lambda x: x.start_time)\n snapshots = snapshots[:-int(retention)]\n if not snapshots:\n logger.info('No old snapshots to remove')\n return\n for snapshot in snapshots:\n logger.info('Deleting snapshot {}'.format(snapshot.id))\n try:\n snapshot.delete()\n except EC2ResponseError as error:\n logger.warning('Could not remove snapshot: {}'.format(error.\n message))\n logger.info('Done deleting snapshots')\n",
"role": "user"
},
{
"content": "remove_old_snapshots",
"role": "assistant"
}
] |
def x(connection):
volumes = get_watched_volumes(connection)
if not volumes:
logger.info('No watched volumes found')
return
logger.info(
'+-----------------------+----------------------+--------------+------------+'
)
logger.info(
'| {volume:<21} | {volume_name:<20.20} | {interval:<12} | {retention:<10} |'
.format(volume='Volume ID', volume_name='Volume name', interval=
'Interval', retention='Retention'))
logger.info(
'+-----------------------+----------------------+--------------+------------+'
)
for volume in volumes:
if 'AutomatedEBSSnapshots' not in volume.tags:
interval = 'Interval tag not found'
elif volume.tags['AutomatedEBSSnapshots'] not in VALID_INTERVALS:
interval = 'Invalid interval'
else:
interval = volume.tags['AutomatedEBSSnapshots']
if 'AutomatedEBSSnapshotsRetention' not in volume.tags:
retention = 0
else:
retention = volume.tags['AutomatedEBSSnapshotsRetention']
try:
volume_name = volume.tags['Name']
except KeyError:
volume_name = ''
logger.info(
'|\xa0{volume_id:<14} | {volume_name:<20.20} | {interval:<12} | {retention:<10} |'
.format(volume_id=volume.id, volume_name=volume_name, interval=
interval, retention=retention))
logger.info(
'+-----------------------+----------------------+--------------+------------+'
)
| list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(connection):\n volumes = get_watched_volumes(connection)\n if not volumes:\n logger.info('No watched volumes found')\n return\n logger.info(\n '+-----------------------+----------------------+--------------+------------+'\n )\n logger.info(\n '| {volume:<21} | {volume_name:<20.20} | {interval:<12} | {retention:<10} |'\n .format(volume='Volume ID', volume_name='Volume name', interval=\n 'Interval', retention='Retention'))\n logger.info(\n '+-----------------------+----------------------+--------------+------------+'\n )\n for volume in volumes:\n if 'AutomatedEBSSnapshots' not in volume.tags:\n interval = 'Interval tag not found'\n elif volume.tags['AutomatedEBSSnapshots'] not in VALID_INTERVALS:\n interval = 'Invalid interval'\n else:\n interval = volume.tags['AutomatedEBSSnapshots']\n if 'AutomatedEBSSnapshotsRetention' not in volume.tags:\n retention = 0\n else:\n retention = volume.tags['AutomatedEBSSnapshotsRetention']\n try:\n volume_name = volume.tags['Name']\n except KeyError:\n volume_name = ''\n logger.info(\n '|\\xa0{volume_id:<14} | {volume_name:<20.20} | {interval:<12} | {retention:<10} |'\n .format(volume_id=volume.id, volume_name=volume_name, interval=\n interval, retention=retention))\n logger.info(\n '+-----------------------+----------------------+--------------+------------+'\n )\n",
"role": "user"
},
{
"content": "list",
"role": "assistant"
}
] |
def x(connection, volume_id):
try:
volume = connection.get_all_volumes(volume_ids=[volume_id])[0]
volume.remove_tag('AutomatedEBSSnapshots')
except EC2ResponseError:
pass
logger.info('Removed {} from the watchlist'.format(volume_id))
return True
| unwatch | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(connection, volume_id):\n try:\n volume = connection.get_all_volumes(volume_ids=[volume_id])[0]\n volume.remove_tag('AutomatedEBSSnapshots')\n except EC2ResponseError:\n pass\n logger.info('Removed {} from the watchlist'.format(volume_id))\n return True\n",
"role": "user"
},
{
"content": "unwatch",
"role": "assistant"
}
] |
def x(connection, volume_id, interval='daily', retention=0):
try:
volume = connection.get_all_volumes(volume_ids=[volume_id])[0]
except EC2ResponseError:
logger.warning('Volume {} not found'.format(volume_id))
return False
if interval not in VALID_INTERVALS:
logger.warning('{} is not a valid interval. Valid intervals are {}'
.format(interval, ', '.join(VALID_INTERVALS)))
volume.remove_tag('AutomatedEBSSnapshots')
volume.add_tag('AutomatedEBSSnapshots', value=interval)
volume.remove_tag('AutomatedEBSSnapshotsRetention')
volume.add_tag('AutomatedEBSSnapshotsRetention', value=int(retention))
logger.info('Updated the rotation interval to {} for {}'.format(
interval, volume_id))
return True
| watch | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(connection, volume_id, interval='daily', retention=0):\n try:\n volume = connection.get_all_volumes(volume_ids=[volume_id])[0]\n except EC2ResponseError:\n logger.warning('Volume {} not found'.format(volume_id))\n return False\n if interval not in VALID_INTERVALS:\n logger.warning('{} is not a valid interval. Valid intervals are {}'\n .format(interval, ', '.join(VALID_INTERVALS)))\n volume.remove_tag('AutomatedEBSSnapshots')\n volume.add_tag('AutomatedEBSSnapshots', value=interval)\n volume.remove_tag('AutomatedEBSSnapshotsRetention')\n volume.add_tag('AutomatedEBSSnapshotsRetention', value=int(retention))\n logger.info('Updated the rotation interval to {} for {}'.format(\n interval, volume_id))\n return True\n",
"role": "user"
},
{
"content": "watch",
"role": "assistant"
}
] |
def x(connection, volume):
volume_id_pattern = re.compile('vol-\\w{8}')
if volume_id_pattern.match(volume):
try:
connection.get_all_volumes(volume_ids=[volume])
volume_id = volume
except EC2ResponseError:
logger.warning('Volume {} not found'.format(volume))
return None
else:
name_filter = {'tag-key': 'Name', 'tag-value': volume}
volumes = connection.get_all_volumes(filters=name_filter)
if not volumes:
logger.warning('Volume {} not found'.format(volume))
return None
if len(volumes) > 1:
logger.warning('Volume {} not unique'.format(volume))
volume_id = volumes[0].id
return volume_id
| get_volume_id | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(connection, volume):\n volume_id_pattern = re.compile('vol-\\\\w{8}')\n if volume_id_pattern.match(volume):\n try:\n connection.get_all_volumes(volume_ids=[volume])\n volume_id = volume\n except EC2ResponseError:\n logger.warning('Volume {} not found'.format(volume))\n return None\n else:\n name_filter = {'tag-key': 'Name', 'tag-value': volume}\n volumes = connection.get_all_volumes(filters=name_filter)\n if not volumes:\n logger.warning('Volume {} not found'.format(volume))\n return None\n if len(volumes) > 1:\n logger.warning('Volume {} not unique'.format(volume))\n volume_id = volumes[0].id\n return volume_id\n",
"role": "user"
},
{
"content": "get_volume_id",
"role": "assistant"
}
] |
def x(connection, file_name):
with open(file_name, 'r') as filehandle:
for line in filehandle.xreadlines():
volume, interval, retention = line.rstrip().split(',')
watch(connection, get_volume_id(connection, volume), interval,
retention)
| watch_from_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(connection, file_name):\n with open(file_name, 'r') as filehandle:\n for line in filehandle.xreadlines():\n volume, interval, retention = line.rstrip().split(',')\n watch(connection, get_volume_id(connection, volume), interval,\n retention)\n",
"role": "user"
},
{
"content": "watch_from_file",
"role": "assistant"
}
] |
def x(connection, file_name):
with open(file_name, 'r') as filehandle:
for line in filehandle.xreadlines():
volume, interval, retention = line.rstrip().split(',')
unwatch(connection, get_volume_id(connection, volume))
| unwatch_from_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(connection, file_name):\n with open(file_name, 'r') as filehandle:\n for line in filehandle.xreadlines():\n volume, interval, retention = line.rstrip().split(',')\n unwatch(connection, get_volume_id(connection, volume))\n",
"role": "user"
},
{
"content": "unwatch_from_file",
"role": "assistant"
}
] |
def x(connection, volume):
logger.info(
'+----------------+----------------------+---------------------------+'
)
logger.info('| {snapshot:<14} | {snapshot_name:<20.20} | {created:<25} |'
.format(snapshot='Snapshot ID', snapshot_name='Snapshot name',
created='Created'))
logger.info(
'+----------------+----------------------+---------------------------+'
)
vid = get_volume_id(connection, volume)
if vid:
vol = connection.get_all_volumes(volume_ids=[vid])[0]
for snap in vol.snapshots():
logger.info(
'| {snapshot:<14} | {snapshot_name:<20.20} | {created:<25} |'
.format(snapshot=snap.id, snapshot_name=snap.tags.get(
'Name', ''), created=snap.start_time))
logger.info(
'+----------------+----------------------+---------------------------+'
)
| list_snapshots | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(connection, volume):\n logger.info(\n '+----------------+----------------------+---------------------------+'\n )\n logger.info('| {snapshot:<14} | {snapshot_name:<20.20} | {created:<25} |'\n .format(snapshot='Snapshot ID', snapshot_name='Snapshot name',\n created='Created'))\n logger.info(\n '+----------------+----------------------+---------------------------+'\n )\n vid = get_volume_id(connection, volume)\n if vid:\n vol = connection.get_all_volumes(volume_ids=[vid])[0]\n for snap in vol.snapshots():\n logger.info(\n '| {snapshot:<14} | {snapshot_name:<20.20} | {created:<25} |'\n .format(snapshot=snap.id, snapshot_name=snap.tags.get(\n 'Name', ''), created=snap.start_time))\n logger.info(\n '+----------------+----------------------+---------------------------+'\n )\n",
"role": "user"
},
{
"content": "list_snapshots",
"role": "assistant"
}
] |
def x(self, words, parser, **kwargs):
output = self._run_morfologik(words)
return parser.parse(output, **kwargs)
| stem | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, words, parser, **kwargs):\n output = self._run_morfologik(words)\n return parser.parse(output, **kwargs)\n",
"role": "user"
},
{
"content": "stem",
"role": "assistant"
}
] |
def x(self, words):
p = subprocess.Popen(['java', '-jar', self.jar_path, 'plstem', '-ie',
'UTF-8', '-oe', 'UTF-8'], bufsize=-1, stdin=subprocess.PIPE, stdout
=subprocess.PIPE, stderr=subprocess.STDOUT)
out, _ = p.communicate(input=bytes('\n'.join(words), 'utf-8'))
return decode(out, 'utf-8')
| run_morfologik | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, words):\n p = subprocess.Popen(['java', '-jar', self.jar_path, 'plstem', '-ie',\n 'UTF-8', '-oe', 'UTF-8'], bufsize=-1, stdin=subprocess.PIPE, stdout\n =subprocess.PIPE, stderr=subprocess.STDOUT)\n out, _ = p.communicate(input=bytes('\\n'.join(words), 'utf-8'))\n return decode(out, 'utf-8')\n",
"role": "user"
},
{
"content": "run_morfologik",
"role": "assistant"
}
] |
def x(identifier, configuration=None):
showcase = Showcase(configuration=configuration)
result = showcase._load_from_hdx('showcase', identifier)
if result:
return showcase
return None
| read_from_hdx | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(identifier, configuration=None):\n showcase = Showcase(configuration=configuration)\n result = showcase._load_from_hdx('showcase', identifier)\n if result:\n return showcase\n return None\n",
"role": "user"
},
{
"content": "read_from_hdx",
"role": "assistant"
}
] |
def x(self):
assoc_result, datasets_dicts = self._read_from_hdx('showcase', self.
data['id'], fieldname='showcase_id', action=self.actions()[
'list_datasets'])
datasets = list()
if assoc_result:
for dataset_dict in datasets_dicts:
dataset = hdx.data.dataset.Dataset(dataset_dict, configuration=
self.configuration)
datasets.append(dataset)
return datasets
| get_datasets | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n assoc_result, datasets_dicts = self._read_from_hdx('showcase', self.\n data['id'], fieldname='showcase_id', action=self.actions()[\n 'list_datasets'])\n datasets = list()\n if assoc_result:\n for dataset_dict in datasets_dicts:\n dataset = hdx.data.dataset.Dataset(dataset_dict, configuration=\n self.configuration)\n datasets.append(dataset)\n return datasets\n",
"role": "user"
},
{
"content": "get_datasets",
"role": "assistant"
}
] |
def x(self, dataset):
if isinstance(dataset, hdx.data.dataset.Dataset) or isinstance(dataset,
dict):
if 'id' not in dataset:
dataset = hdx.data.dataset.Dataset.read_from_hdx(dataset['name'])
dataset = dataset['id']
elif not isinstance(dataset, str):
raise hdx.data.hdxobject.HDXError(
'Type %s cannot be added as a dataset!' % type(dataset).__name__)
if is_valid_uuid(dataset) is False:
raise hdx.data.hdxobject.HDXError('%s is not a valid dataset id!' %
dataset)
return {'showcase_id': self.data['id'], 'package_id': dataset}
| get_showcase_dataset_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dataset):\n if isinstance(dataset, hdx.data.dataset.Dataset) or isinstance(dataset,\n dict):\n if 'id' not in dataset:\n dataset = hdx.data.dataset.Dataset.read_from_hdx(dataset['name'])\n dataset = dataset['id']\n elif not isinstance(dataset, str):\n raise hdx.data.hdxobject.HDXError(\n 'Type %s cannot be added as a dataset!' % type(dataset).__name__)\n if is_valid_uuid(dataset) is False:\n raise hdx.data.hdxobject.HDXError('%s is not a valid dataset id!' %\n dataset)\n return {'showcase_id': self.data['id'], 'package_id': dataset}\n",
"role": "user"
},
{
"content": "get_showcase_dataset_dict",
"role": "assistant"
}
] |
def x(self, dataset, datasets_to_check=None):
showcase_dataset = self._get_showcase_dataset_dict(dataset)
if datasets_to_check is None:
datasets_to_check = self.get_datasets()
for dataset in datasets_to_check:
if showcase_dataset['package_id'] == dataset['id']:
return False
self._write_to_hdx('associate', showcase_dataset, 'package_id')
return True
| add_dataset | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dataset, datasets_to_check=None):\n showcase_dataset = self._get_showcase_dataset_dict(dataset)\n if datasets_to_check is None:\n datasets_to_check = self.get_datasets()\n for dataset in datasets_to_check:\n if showcase_dataset['package_id'] == dataset['id']:\n return False\n self._write_to_hdx('associate', showcase_dataset, 'package_id')\n return True\n",
"role": "user"
},
{
"content": "add_dataset",
"role": "assistant"
}
] |
def x(self, datasets, datasets_to_check=None):
if datasets_to_check is None:
datasets_to_check = self.get_datasets()
alldatasetsadded = True
for dataset in datasets:
if not self.add_dataset(dataset, datasets_to_check=datasets_to_check):
alldatasetsadded = False
return alldatasetsadded
| add_datasets | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, datasets, datasets_to_check=None):\n if datasets_to_check is None:\n datasets_to_check = self.get_datasets()\n alldatasetsadded = True\n for dataset in datasets:\n if not self.add_dataset(dataset, datasets_to_check=datasets_to_check):\n alldatasetsadded = False\n return alldatasetsadded\n",
"role": "user"
},
{
"content": "add_datasets",
"role": "assistant"
}
] |
def x(nb, jsonable_parameter=True, end_cell_index=None):
jh = _JupyterNotebookHelper(nb, jsonable_parameter, end_cell_index)
if jsonable_parameter is True:
PossibleParameter = collections.namedtuple('PossibleParameter', [
'name', 'value', 'cell_index'])
else:
PossibleParameter = collections.namedtuple('PossibleParameter', [
'name', 'cell_index'])
res = []
for name, cell_index in jh.param_cell_index.items():
if jsonable_parameter is True:
res.append(PossibleParameter(name=name, value=jh.param_value[
name], cell_index=cell_index))
else:
res.append(PossibleParameter(name=name, cell_index=cell_index))
return sorted(res, key=lambda x: x.name)
| possible_parameter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(nb, jsonable_parameter=True, end_cell_index=None):\n jh = _JupyterNotebookHelper(nb, jsonable_parameter, end_cell_index)\n if jsonable_parameter is True:\n PossibleParameter = collections.namedtuple('PossibleParameter', [\n 'name', 'value', 'cell_index'])\n else:\n PossibleParameter = collections.namedtuple('PossibleParameter', [\n 'name', 'cell_index'])\n res = []\n for name, cell_index in jh.param_cell_index.items():\n if jsonable_parameter is True:\n res.append(PossibleParameter(name=name, value=jh.param_value[\n name], cell_index=cell_index))\n else:\n res.append(PossibleParameter(name=name, cell_index=cell_index))\n return sorted(res, key=lambda x: x.name)\n",
"role": "user"
},
{
"content": "possible_parameter",
"role": "assistant"
}
] |
def x(self, joiner, formatter=lambda s, t: t.format(s), template='{}'):
return ww.s(joiner).join(self, formatter, template)
| join | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, joiner, formatter=lambda s, t: t.format(s), template='{}'):\n return ww.s(joiner).join(self, formatter, template)\n",
"role": "user"
},
{
"content": "join",
"role": "assistant"
}
] |
def x(self, *values):
for value in values:
list.append(self, value)
return self
| append | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *values):\n for value in values:\n list.append(self, value)\n return self\n",
"role": "user"
},
{
"content": "append",
"role": "assistant"
}
] |
def x(self, *iterables):
for value in iterables:
list.extend(self, value)
return self
| extend | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *iterables):\n for value in iterables:\n list.extend(self, value)\n return self\n",
"role": "user"
},
{
"content": "extend",
"role": "assistant"
}
] |
def x(value):
if isinstance(value, dict) or isinstance(value, list):
return json.dumps(value)
return value
| normalize_cell_value | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(value):\n if isinstance(value, dict) or isinstance(value, list):\n return json.dumps(value)\n return value\n",
"role": "user"
},
{
"content": "normalize_cell_value",
"role": "assistant"
}
] |
def x(input_file_name):
mode = 'r'
if sys.version_info[0] < 3:
mode = 'rb'
with io.open(input_file_name, mode) as input_file:
reader = csv.reader(input_file, delimiter=',', quotechar='"')
addresses = list(map(tuple, reader))
if len(addresses) == 0:
raise Exception('No addresses found in input file')
header_columns = list(column.lower() for column in addresses.pop(0))
try:
address_index = header_columns.index('address')
zipcode_index = header_columns.index('zipcode')
except ValueError:
raise Exception(
"The first row of the input CSV must be a header that contains a column labeled 'address' and a column labeled 'zipcode'."
)
return list((row[address_index], row[zipcode_index]) for row in
addresses)
| get_addresses_from_input_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(input_file_name):\n mode = 'r'\n if sys.version_info[0] < 3:\n mode = 'rb'\n with io.open(input_file_name, mode) as input_file:\n reader = csv.reader(input_file, delimiter=',', quotechar='\"')\n addresses = list(map(tuple, reader))\n if len(addresses) == 0:\n raise Exception('No addresses found in input file')\n header_columns = list(column.lower() for column in addresses.pop(0))\n try:\n address_index = header_columns.index('address')\n zipcode_index = header_columns.index('zipcode')\n except ValueError:\n raise Exception(\n \"The first row of the input CSV must be a header that contains a column labeled 'address' and a column labeled 'zipcode'.\"\n )\n return list((row[address_index], row[zipcode_index]) for row in\n addresses)\n",
"role": "user"
},
{
"content": "get_addresses_from_input_file",
"role": "assistant"
}
] |
def x(input_file_name):
valid_identifiers = ['address', 'zipcode', 'unit', 'city', 'state',
'slug', 'block_id', 'msa', 'num_bins', 'property_type',
'client_value', 'client_value_sqft', 'meta']
mode = 'r'
if sys.version_info[0] < 3:
mode = 'rb'
with io.open(input_file_name, mode) as input_file:
result = [{identifier: val for identifier, val in list(row.items()) if
identifier in valid_identifiers} for row in csv.DictReader(
input_file, skipinitialspace=True)]
return result
| get_identifiers_from_input_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(input_file_name):\n valid_identifiers = ['address', 'zipcode', 'unit', 'city', 'state',\n 'slug', 'block_id', 'msa', 'num_bins', 'property_type',\n 'client_value', 'client_value_sqft', 'meta']\n mode = 'r'\n if sys.version_info[0] < 3:\n mode = 'rb'\n with io.open(input_file_name, mode) as input_file:\n result = [{identifier: val for identifier, val in list(row.items()) if\n identifier in valid_identifiers} for row in csv.DictReader(\n input_file, skipinitialspace=True)]\n return result\n",
"role": "user"
},
{
"content": "get_identifiers_from_input_file",
"role": "assistant"
}
] |
def x(ipAddress, auth, url):
real_time_locate_url = (
'/imcrs/res/access/realtimeLocate?type=2&value=' + str(ipAddress) +
'&total=false')
f_url = url + real_time_locate_url
r = requests.get(f_url, auth=auth, headers=HEADERS)
try:
if r.status_code == 200:
response = json.loads(r.text)
if 'realtimeLocation' in response:
real_time_locate = json.loads(r.text)['realtimeLocation']
if type(real_time_locate) is dict:
real_time_locate = [real_time_locate]
return real_time_locate
else:
return json.loads(r.text)['realtimeLocation']
else:
return json.loads(r.text)
except requests.exceptions.RequestException as e:
return 'Error:\n' + str(e
) + ' get_real_time_locate: An Error has occured'
| get_real_time_locate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ipAddress, auth, url):\n real_time_locate_url = (\n '/imcrs/res/access/realtimeLocate?type=2&value=' + str(ipAddress) +\n '&total=false')\n f_url = url + real_time_locate_url\n r = requests.get(f_url, auth=auth, headers=HEADERS)\n try:\n if r.status_code == 200:\n response = json.loads(r.text)\n if 'realtimeLocation' in response:\n real_time_locate = json.loads(r.text)['realtimeLocation']\n if type(real_time_locate) is dict:\n real_time_locate = [real_time_locate]\n return real_time_locate\n else:\n return json.loads(r.text)['realtimeLocation']\n else:\n return json.loads(r.text)\n except requests.exceptions.RequestException as e:\n return 'Error:\\n' + str(e\n ) + ' get_real_time_locate: An Error has occured'\n",
"role": "user"
},
{
"content": "get_real_time_locate",
"role": "assistant"
}
] |
def x(devId, auth, url):
if auth is None or url is None:
set_imc_creds()
ip_mac_arp_list_url = '/imcrs/res/access/ipMacArp/' + str(devId)
f_url = url + ip_mac_arp_list_url
r = requests.get(f_url, auth=auth, headers=HEADERS)
try:
if r.status_code == 200:
macarplist = json.loads(r.text)
if len(macarplist) > 1:
return macarplist['ipMacArp']
else:
return ['this function is unsupported']
except requests.exceptions.RequestException as e:
return 'Error:\n' + str(e
) + ' get_ip_mac_arp_list: An Error has occured'
| get_ip_mac_arp_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(devId, auth, url):\n if auth is None or url is None:\n set_imc_creds()\n ip_mac_arp_list_url = '/imcrs/res/access/ipMacArp/' + str(devId)\n f_url = url + ip_mac_arp_list_url\n r = requests.get(f_url, auth=auth, headers=HEADERS)\n try:\n if r.status_code == 200:\n macarplist = json.loads(r.text)\n if len(macarplist) > 1:\n return macarplist['ipMacArp']\n else:\n return ['this function is unsupported']\n except requests.exceptions.RequestException as e:\n return 'Error:\\n' + str(e\n ) + ' get_ip_mac_arp_list: An Error has occured'\n",
"role": "user"
},
{
"content": "get_ip_mac_arp_list",
"role": "assistant"
}
] |
def x(scopeId, auth, url):
get_ip_scope_url = '/imcrs/res/access/assignedIpScope/' + str(scopeId)
f_url = url + get_ip_scope_url
r = requests.get(f_url, auth=auth, headers=HEADERS)
try:
if r.status_code == 200:
ipscopelist = json.loads(r.text)
return ipscopelist
except requests.exceptions.RequestException as e:
return 'Error:\n' + str(e) + ' get_ip_scope: An Error has occured'
| get_ip_scope_detail | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(scopeId, auth, url):\n get_ip_scope_url = '/imcrs/res/access/assignedIpScope/' + str(scopeId)\n f_url = url + get_ip_scope_url\n r = requests.get(f_url, auth=auth, headers=HEADERS)\n try:\n if r.status_code == 200:\n ipscopelist = json.loads(r.text)\n return ipscopelist\n except requests.exceptions.RequestException as e:\n return 'Error:\\n' + str(e) + ' get_ip_scope: An Error has occured'\n",
"role": "user"
},
{
"content": "get_ip_scope_detail",
"role": "assistant"
}
] |
def x(network_address, auth, url):
scope_id = get_scope_id(network_address, auth, url)
delete_ip_address_url = '/imcrs/res/access/assignedIpScope/' + str(scope_id
)
f_url = url + delete_ip_address_url
r = requests.delete(f_url, auth=auth, headers=HEADERS)
try:
return r
if r.status_code == 204:
return r.status_code
except requests.exceptions.RequestException as e:
return 'Error:\n' + str(e) + ' delete_ip_scope: An Error has occured'
| delete_ip_scope | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(network_address, auth, url):\n scope_id = get_scope_id(network_address, auth, url)\n delete_ip_address_url = '/imcrs/res/access/assignedIpScope/' + str(scope_id\n )\n f_url = url + delete_ip_address_url\n r = requests.delete(f_url, auth=auth, headers=HEADERS)\n try:\n return r\n if r.status_code == 204:\n return r.status_code\n except requests.exceptions.RequestException as e:\n return 'Error:\\n' + str(e) + ' delete_ip_scope: An Error has occured'\n",
"role": "user"
},
{
"content": "delete_ip_scope",
"role": "assistant"
}
] |
def x(ipaddress, name, description, scopeid, auth, url):
new_ip = {'ip': ipaddress, 'name': name, 'description': description}
add_scope_ip_url = '/imcrs/res/access/assignedIpScope/ip?ipScopeId=' + str(
scopeid)
f_url = url + add_scope_ip_url
payload = json.dumps(new_ip)
r = requests.post(f_url, auth=auth, headers=HEADERS, data=payload)
try:
if r.status_code == 200:
return r.status_code
elif r.status_code == 409:
return r.status_code
except requests.exceptions.RequestException as e:
return 'Error:\n' + str(e) + ' add_ip_scope: An Error has occured'
| add_scope_ip | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ipaddress, name, description, scopeid, auth, url):\n new_ip = {'ip': ipaddress, 'name': name, 'description': description}\n add_scope_ip_url = '/imcrs/res/access/assignedIpScope/ip?ipScopeId=' + str(\n scopeid)\n f_url = url + add_scope_ip_url\n payload = json.dumps(new_ip)\n r = requests.post(f_url, auth=auth, headers=HEADERS, data=payload)\n try:\n if r.status_code == 200:\n return r.status_code\n elif r.status_code == 409:\n return r.status_code\n except requests.exceptions.RequestException as e:\n return 'Error:\\n' + str(e) + ' add_ip_scope: An Error has occured'\n",
"role": "user"
},
{
"content": "add_scope_ip",
"role": "assistant"
}
] |
def x(hostid, auth, url):
add_scope_ip_url = '/imcrs/res/access/assignedIpScope/ip/' + str(hostid)
f_url = url + add_scope_ip_url
r = requests.delete(f_url, auth=auth, headers=HEADERS)
try:
if r.status_code == 204:
return r.status_code
elif r.status_code == 409:
return r.status_code
except requests.exceptions.RequestException as e:
return 'Error:\n' + str(e) + ' add_ip_scope: An Error has occured'
| remove_scope_ip | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(hostid, auth, url):\n add_scope_ip_url = '/imcrs/res/access/assignedIpScope/ip/' + str(hostid)\n f_url = url + add_scope_ip_url\n r = requests.delete(f_url, auth=auth, headers=HEADERS)\n try:\n if r.status_code == 204:\n return r.status_code\n elif r.status_code == 409:\n return r.status_code\n except requests.exceptions.RequestException as e:\n return 'Error:\\n' + str(e) + ' add_ip_scope: An Error has occured'\n",
"role": "user"
},
{
"content": "remove_scope_ip",
"role": "assistant"
}
] |
def x(scopeId, auth, url):
get_ip_scope_url = (
'/imcrs/res/access/assignedIpScope/ip?size=10000&ipScopeId=' + str(
scopeId))
f_url = url + get_ip_scope_url
r = requests.get(f_url, auth=auth, headers=HEADERS)
try:
if r.status_code == 200:
ipscopelist = json.loads(r.text)
if ipscopelist == {}:
return ipscopelist
else:
ipscopelist = ipscopelist['assignedIpInfo']
if type(ipscopelist) is dict:
ipscope = []
ipscope.append(ipscopelist)
return ipscope
return ipscopelist
except requests.exceptions.RequestException as e:
return 'Error:\n' + str(e) + ' get_ip_scope: An Error has occured'
| get_ip_scope_hosts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(scopeId, auth, url):\n get_ip_scope_url = (\n '/imcrs/res/access/assignedIpScope/ip?size=10000&ipScopeId=' + str(\n scopeId))\n f_url = url + get_ip_scope_url\n r = requests.get(f_url, auth=auth, headers=HEADERS)\n try:\n if r.status_code == 200:\n ipscopelist = json.loads(r.text)\n if ipscopelist == {}:\n return ipscopelist\n else:\n ipscopelist = ipscopelist['assignedIpInfo']\n if type(ipscopelist) is dict:\n ipscope = []\n ipscope.append(ipscopelist)\n return ipscope\n return ipscopelist\n except requests.exceptions.RequestException as e:\n return 'Error:\\n' + str(e) + ' get_ip_scope: An Error has occured'\n",
"role": "user"
},
{
"content": "get_ip_scope_hosts",
"role": "assistant"
}
] |
def x(ipaddress, name, description, network_address, auth, url):
scope_id = get_scope_id(network_address, auth, url)
add_scope_ip(ipaddress, name, description, scope_id, auth, url)
| add_host_to_segment | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ipaddress, name, description, network_address, auth, url):\n scope_id = get_scope_id(network_address, auth, url)\n add_scope_ip(ipaddress, name, description, scope_id, auth, url)\n",
"role": "user"
},
{
"content": "add_host_to_segment",
"role": "assistant"
}
] |
def x(ipaddress, networkaddress, auth, url):
host_id = get_host_id(ipaddress, networkaddress, auth, url)
remove_scope_ip(host_id, auth.creds, auth.url)
| delete_host_from_segment | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ipaddress, networkaddress, auth, url):\n host_id = get_host_id(ipaddress, networkaddress, auth, url)\n remove_scope_ip(host_id, auth.creds, auth.url)\n",
"role": "user"
},
{
"content": "delete_host_from_segment",
"role": "assistant"
}
] |
def x(method, version, endpoint, date, rel_url, content_type, content,
access_key, secret_key, hash_type):
hash_type = hash_type
hostname = endpoint._val.netloc
if version >= 'v4.20181215':
content = b''
elif content_type.startswith('multipart/'):
content = b''
body_hash = hashlib.new(hash_type, content).hexdigest()
sign_str = (
'{}\n{}\n{}\nhost:{}\ncontent-type:{}\nx-backendai-version:{}\n{}'.
format(method.upper(), rel_url, date.isoformat(), hostname,
content_type.lower(), version, body_hash))
sign_bytes = sign_str.encode()
sign_key = hmac.new(secret_key.encode(), date.strftime('%Y%m%d').encode
(), hash_type).digest()
sign_key = hmac.new(sign_key, hostname.encode(), hash_type).digest()
signature = hmac.new(sign_key, sign_bytes, hash_type).hexdigest()
headers = {'Authorization':
'BackendAI signMethod=HMAC-{}, credential={}:{}'.format(hash_type.
upper(), access_key, signature)}
return headers, signature
| generate_signature | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(method, version, endpoint, date, rel_url, content_type, content,\n access_key, secret_key, hash_type):\n hash_type = hash_type\n hostname = endpoint._val.netloc\n if version >= 'v4.20181215':\n content = b''\n elif content_type.startswith('multipart/'):\n content = b''\n body_hash = hashlib.new(hash_type, content).hexdigest()\n sign_str = (\n '{}\\n{}\\n{}\\nhost:{}\\ncontent-type:{}\\nx-backendai-version:{}\\n{}'.\n format(method.upper(), rel_url, date.isoformat(), hostname,\n content_type.lower(), version, body_hash))\n sign_bytes = sign_str.encode()\n sign_key = hmac.new(secret_key.encode(), date.strftime('%Y%m%d').encode\n (), hash_type).digest()\n sign_key = hmac.new(sign_key, hostname.encode(), hash_type).digest()\n signature = hmac.new(sign_key, sign_bytes, hash_type).hexdigest()\n headers = {'Authorization':\n 'BackendAI signMethod=HMAC-{}, credential={}:{}'.format(hash_type.\n upper(), access_key, signature)}\n return headers, signature\n",
"role": "user"
},
{
"content": "generate_signature",
"role": "assistant"
}
] |
async def list_with_limit(cls, limit, offset, status: str='ALIVE', fields:
Iterable[str]=None) ->Sequence[dict]:
"""
Fetches the list of agents with the given status with limit and offset for
pagination.
:param limit: number of agents to get
:param offset: offset index of agents to get
:param status: An upper-cased string constant representing agent
status (one of ``'ALIVE'``, ``'TERMINATED'``, ``'LOST'``,
etc.)
:param fields: Additional per-agent query fields to fetch.
"""
if fields is None:
fields = ('id', 'addr', 'status', 'first_contact', 'mem_slots',
'cpu_slots', 'gpu_slots')
q = (
'query($limit: Int!, $offset: Int!, $status: String) { agent_list(limit: $limit, offset: $offset, status: $status) { items { $fields } total_count }}'
)
q = q.replace('$fields', ' '.join(fields))
variables = {'limit': limit, 'offset': offset, 'status': status}
rqst = Request(cls.session, 'POST', '/admin/graphql')
rqst.set_json({'query': q, 'variables': variables})
async with rqst.fetch() as resp:
data = await resp.json()
return data['agent_list']
| list_with_limit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def list_with_limit(cls, limit, offset, status: str='ALIVE', fields:\n Iterable[str]=None) ->Sequence[dict]:\n \"\"\"\n Fetches the list of agents with the given status with limit and offset for\n pagination.\n\n :param limit: number of agents to get\n :param offset: offset index of agents to get\n :param status: An upper-cased string constant representing agent\n status (one of ``'ALIVE'``, ``'TERMINATED'``, ``'LOST'``,\n etc.)\n :param fields: Additional per-agent query fields to fetch.\n \"\"\"\n if fields is None:\n fields = ('id', 'addr', 'status', 'first_contact', 'mem_slots',\n 'cpu_slots', 'gpu_slots')\n q = (\n 'query($limit: Int!, $offset: Int!, $status: String) { agent_list(limit: $limit, offset: $offset, status: $status) { items { $fields } total_count }}'\n )\n q = q.replace('$fields', ' '.join(fields))\n variables = {'limit': limit, 'offset': offset, 'status': status}\n rqst = Request(cls.session, 'POST', '/admin/graphql')\n rqst.set_json({'query': q, 'variables': variables})\n async with rqst.fetch() as resp:\n data = await resp.json()\n return data['agent_list']\n",
"role": "user"
},
{
"content": "list_with_limit",
"role": "assistant"
}
] |
def x(self, value: RequestContent, *, content_type: str=None):
assert self._attached_files is None, 'cannot set content because you already attached files.'
guessed_content_type = 'application/octet-stream'
if value is None:
guessed_content_type = 'text/plain'
self._content = b''
elif isinstance(value, str):
guessed_content_type = 'text/plain'
self._content = value.encode('utf-8')
else:
guessed_content_type = 'application/octet-stream'
self._content = value
self.content_type = (content_type if content_type is not None else
guessed_content_type)
| set_content | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, value: RequestContent, *, content_type: str=None):\n assert self._attached_files is None, 'cannot set content because you already attached files.'\n guessed_content_type = 'application/octet-stream'\n if value is None:\n guessed_content_type = 'text/plain'\n self._content = b''\n elif isinstance(value, str):\n guessed_content_type = 'text/plain'\n self._content = value.encode('utf-8')\n else:\n guessed_content_type = 'application/octet-stream'\n self._content = value\n self.content_type = (content_type if content_type is not None else\n guessed_content_type)\n",
"role": "user"
},
{
"content": "set_content",
"role": "assistant"
}
] |
def x(self, value: object):
self.set_content(modjson.dumps(value, cls=ExtendedJSONEncoder),
content_type='application/json')
| set_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, value: object):\n self.set_content(modjson.dumps(value, cls=ExtendedJSONEncoder),\n content_type='application/json')\n",
"role": "user"
},
{
"content": "set_json",
"role": "assistant"
}
] |
def x(self, files: Sequence[AttachedFile]):
assert not self._content, 'content must be empty to attach files.'
self.content_type = 'multipart/form-data'
self._attached_files = files
| attach_files | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, files: Sequence[AttachedFile]):\n assert not self._content, 'content must be empty to attach files.'\n self.content_type = 'multipart/form-data'\n self._attached_files = files\n",
"role": "user"
},
{
"content": "attach_files",
"role": "assistant"
}
] |
def x(self, rel_url, access_key=None, secret_key=None, hash_type=None):
if access_key is None:
access_key = self.config.access_key
if secret_key is None:
secret_key = self.config.secret_key
if hash_type is None:
hash_type = self.config.hash_type
hdrs, _ = generate_signature(self.method, self.config.version, self.
config.endpoint, self.date, str(rel_url), self.content_type, self.
_content, access_key, secret_key, hash_type)
self.headers.update(hdrs)
| sign | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, rel_url, access_key=None, secret_key=None, hash_type=None):\n if access_key is None:\n access_key = self.config.access_key\n if secret_key is None:\n secret_key = self.config.secret_key\n if hash_type is None:\n hash_type = self.config.hash_type\n hdrs, _ = generate_signature(self.method, self.config.version, self.\n config.endpoint, self.date, str(rel_url), self.content_type, self.\n _content, access_key, secret_key, hash_type)\n self.headers.update(hdrs)\n",
"role": "user"
},
{
"content": "sign",
"role": "assistant"
}
] |
def x(self, **kwargs) ->'FetchContextManager':
assert self.method in self._allowed_methods, 'Disallowed HTTP method: {}'.format(
self.method)
self.date = datetime.now(tzutc())
self.headers['Date'] = self.date.isoformat()
if self.content_type is not None:
self.headers['Content-Type'] = self.content_type
full_url = self._build_url()
self._sign(full_url.relative())
rqst_ctx = self.session.aiohttp_session.request(self.method, str(
full_url), data=self._pack_content(), timeout=
_default_request_timeout, headers=self.headers)
return FetchContextManager(self.session, rqst_ctx, **kwargs)
| fetch | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs) ->'FetchContextManager':\n assert self.method in self._allowed_methods, 'Disallowed HTTP method: {}'.format(\n self.method)\n self.date = datetime.now(tzutc())\n self.headers['Date'] = self.date.isoformat()\n if self.content_type is not None:\n self.headers['Content-Type'] = self.content_type\n full_url = self._build_url()\n self._sign(full_url.relative())\n rqst_ctx = self.session.aiohttp_session.request(self.method, str(\n full_url), data=self._pack_content(), timeout=\n _default_request_timeout, headers=self.headers)\n return FetchContextManager(self.session, rqst_ctx, **kwargs)\n",
"role": "user"
},
{
"content": "fetch",
"role": "assistant"
}
] |
def x(self, **kwargs) ->'WebSocketContextManager':
assert isinstance(self.session, AsyncSession
), 'Cannot use websockets with sessions in the synchronous mode'
assert self.method == 'GET', 'Invalid websocket method'
self.date = datetime.now(tzutc())
self.headers['Date'] = self.date.isoformat()
self.content_type = 'application/octet-stream'
full_url = self._build_url()
self._sign(full_url.relative())
ws_ctx = self.session.aiohttp_session.ws_connect(str(full_url),
autoping=True, heartbeat=30.0, headers=self.headers)
return WebSocketContextManager(self.session, ws_ctx, **kwargs)
| connect_websocket | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs) ->'WebSocketContextManager':\n assert isinstance(self.session, AsyncSession\n ), 'Cannot use websockets with sessions in the synchronous mode'\n assert self.method == 'GET', 'Invalid websocket method'\n self.date = datetime.now(tzutc())\n self.headers['Date'] = self.date.isoformat()\n self.content_type = 'application/octet-stream'\n full_url = self._build_url()\n self._sign(full_url.relative())\n ws_ctx = self.session.aiohttp_session.ws_connect(str(full_url),\n autoping=True, heartbeat=30.0, headers=self.headers)\n return WebSocketContextManager(self.session, ws_ctx, **kwargs)\n",
"role": "user"
},
{
"content": "connect_websocket",
"role": "assistant"
}
] |
def x():
try:
src_path = os.path.dirname(os.path.dirname(os.path.dirname(os.path.
abspath(__file__))))
if src_path not in sys.path:
sys.path.append(src_path)
from aoiklivereload import LiveReloader
reloader = LiveReloader(reload_mode='spawn_exit' if sys.platform ==
'win32' else 'exec', force_exit=True)
reloader.start_watcher_thread()
server_host = '0.0.0.0'
server_port = 8000
msg = '# ----- Run server -----\nHost: {}\nPort: {}'.format(server_host
, server_port)
print(msg)
class HelloHandler(tornado.web.RequestHandler):
"""
Request handler class.
"""
def x(self):
self.write('hello')
handler_tuples = [('/', HelloHandler)]
tornado_app = tornado.web.Application(handler_tuples, debug=False)
tornado_app.listen(server_port, address=server_host)
io_loop = tornado.ioloop.IOLoop.current()
io_loop.start()
except KeyboardInterrupt:
pass
| main | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n try:\n src_path = os.path.dirname(os.path.dirname(os.path.dirname(os.path.\n abspath(__file__))))\n if src_path not in sys.path:\n sys.path.append(src_path)\n from aoiklivereload import LiveReloader\n reloader = LiveReloader(reload_mode='spawn_exit' if sys.platform ==\n 'win32' else 'exec', force_exit=True)\n reloader.start_watcher_thread()\n server_host = '0.0.0.0'\n server_port = 8000\n msg = '# ----- Run server -----\\nHost: {}\\nPort: {}'.format(server_host\n , server_port)\n print(msg)\n\n\n class HelloHandler(tornado.web.RequestHandler):\n \"\"\"\n Request handler class.\n \"\"\"\n\n def x(self):\n self.write('hello')\n handler_tuples = [('/', HelloHandler)]\n tornado_app = tornado.web.Application(handler_tuples, debug=False)\n tornado_app.listen(server_port, address=server_host)\n io_loop = tornado.ioloop.IOLoop.current()\n io_loop.start()\n except KeyboardInterrupt:\n pass\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x(task, auth, url):
add_perf_task_url = '/imcrs/perf/task'
f_url = url + add_perf_task_url
payload = json.dumps(task)
response = requests.post(f_url, data=payload, auth=auth, headers=HEADERS)
try:
return response.status_code
except requests.exceptions.RequestException as error:
return 'Error:\n' + str(error) + ' add_perf_task: An Error has occured'
| add_perf_task | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(task, auth, url):\n add_perf_task_url = '/imcrs/perf/task'\n f_url = url + add_perf_task_url\n payload = json.dumps(task)\n response = requests.post(f_url, data=payload, auth=auth, headers=HEADERS)\n try:\n return response.status_code\n except requests.exceptions.RequestException as error:\n return 'Error:\\n' + str(error) + ' add_perf_task: An Error has occured'\n",
"role": "user"
},
{
"content": "add_perf_task",
"role": "assistant"
}
] |
def x(task_name, auth, url):
get_perf_task_url = ('/imcrs/perf/task?name=' + task_name +
'&orderBy=taskId&desc=false')
f_url = url + get_perf_task_url
response = requests.get(f_url, auth=auth, headers=HEADERS)
try:
if response.status_code == 200:
perf_task_info = json.loads(response.text)
if 'task' in perf_task_info:
perf_task_info = json.loads(response.text)['task']
else:
perf_task_info = "Task Doesn't Exist"
return perf_task_info
except requests.exceptions.RequestException as error:
return 'Error:\n' + str(error) + ' get_perf_task: An Error has occured'
| get_perf_task | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(task_name, auth, url):\n get_perf_task_url = ('/imcrs/perf/task?name=' + task_name +\n '&orderBy=taskId&desc=false')\n f_url = url + get_perf_task_url\n response = requests.get(f_url, auth=auth, headers=HEADERS)\n try:\n if response.status_code == 200:\n perf_task_info = json.loads(response.text)\n if 'task' in perf_task_info:\n perf_task_info = json.loads(response.text)['task']\n else:\n perf_task_info = \"Task Doesn't Exist\"\n return perf_task_info\n except requests.exceptions.RequestException as error:\n return 'Error:\\n' + str(error) + ' get_perf_task: An Error has occured'\n",
"role": "user"
},
{
"content": "get_perf_task",
"role": "assistant"
}
] |
def x(task_name, auth, url):
task_id = get_perf_task(task_name, auth, url)
if isinstance(task_id, str):
print("Perf task doesn't exist")
return 403
task_id = task_id['taskId']
get_perf_task_url = '/imcrs/perf/task/delete/' + str(task_id)
f_url = url + get_perf_task_url
response = requests.delete(f_url, auth=auth, headers=HEADERS)
try:
if response.status_code == 204:
print('Perf Task successfully delete')
return response.status_code
except requests.exceptions.RequestException as error:
return 'Error:\n' + str(error
) + ' delete_perf_task: An Error has occured'
| delete_perf_task | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(task_name, auth, url):\n task_id = get_perf_task(task_name, auth, url)\n if isinstance(task_id, str):\n print(\"Perf task doesn't exist\")\n return 403\n task_id = task_id['taskId']\n get_perf_task_url = '/imcrs/perf/task/delete/' + str(task_id)\n f_url = url + get_perf_task_url\n response = requests.delete(f_url, auth=auth, headers=HEADERS)\n try:\n if response.status_code == 204:\n print('Perf Task successfully delete')\n return response.status_code\n except requests.exceptions.RequestException as error:\n return 'Error:\\n' + str(error\n ) + ' delete_perf_task: An Error has occured'\n",
"role": "user"
},
{
"content": "delete_perf_task",
"role": "assistant"
}
] |
def x(self, response):
response_json = response.json()
code_key = 'code'
if code_key in response_json and response_json[code_key
] != constants.HTTP_CODE_OK:
code = response_json[code_key]
message = response_json
if 'message' in response_json:
message = response_json['message']
elif 'code_description' in response_json:
message = response_json['code_description']
if code == constants.HTTP_FORBIDDEN:
raise housecanary.exceptions.UnauthorizedException(code, message)
if code == constants.HTTP_TOO_MANY_REQUESTS:
raise housecanary.exceptions.RateLimitException(code, message,
response)
else:
raise housecanary.exceptions.RequestException(code, message)
request_url = response.request.url
endpoint_name = self._parse_endpoint_name_from_url(request_url)
return Response.create(endpoint_name, response_json, response)
| process_json_response | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, response):\n response_json = response.json()\n code_key = 'code'\n if code_key in response_json and response_json[code_key\n ] != constants.HTTP_CODE_OK:\n code = response_json[code_key]\n message = response_json\n if 'message' in response_json:\n message = response_json['message']\n elif 'code_description' in response_json:\n message = response_json['code_description']\n if code == constants.HTTP_FORBIDDEN:\n raise housecanary.exceptions.UnauthorizedException(code, message)\n if code == constants.HTTP_TOO_MANY_REQUESTS:\n raise housecanary.exceptions.RateLimitException(code, message,\n response)\n else:\n raise housecanary.exceptions.RequestException(code, message)\n request_url = response.request.url\n endpoint_name = self._parse_endpoint_name_from_url(request_url)\n return Response.create(endpoint_name, response_json, response)\n",
"role": "user"
},
{
"content": "process_json_response",
"role": "assistant"
}
] |
def x(self):
watcher_thread = threading.Thread(target=self.run_watcher)
if self._reload_mode == self.RELOAD_MODE_V_SPAWN_WAIT:
daemon = False
else:
daemon = True
watcher_thread.setDaemon(daemon)
watcher_thread.start()
return watcher_thread
| start_watcher_thread | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n watcher_thread = threading.Thread(target=self.run_watcher)\n if self._reload_mode == self.RELOAD_MODE_V_SPAWN_WAIT:\n daemon = False\n else:\n daemon = True\n watcher_thread.setDaemon(daemon)\n watcher_thread.start()\n return watcher_thread\n",
"role": "user"
},
{
"content": "start_watcher_thread",
"role": "assistant"
}
] |
def x(self):
observer = Observer()
observer.start()
watche_obj_map = {}
while not self._watcher_to_stop:
old_watch_path_s = set(watche_obj_map)
new_watch_path_s = self._find_watch_paths()
for new_watch_path in new_watch_path_s:
old_watch_path_s.discard(new_watch_path)
if new_watch_path not in watche_obj_map:
try:
watch_obj = observer.schedule(self, new_watch_path,
recursive=True)
watche_obj_map[new_watch_path] = watch_obj
except OSError:
watche_obj_map[new_watch_path] = None
for old_watch_path in old_watch_path_s:
watch_obj = watche_obj_map.pop(old_watch_path, None)
if watch_obj is not None:
observer.unschedule(watch_obj)
self._watch_paths = new_watch_path_s
time.sleep(self._interval)
| run_watcher | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n observer = Observer()\n observer.start()\n watche_obj_map = {}\n while not self._watcher_to_stop:\n old_watch_path_s = set(watche_obj_map)\n new_watch_path_s = self._find_watch_paths()\n for new_watch_path in new_watch_path_s:\n old_watch_path_s.discard(new_watch_path)\n if new_watch_path not in watche_obj_map:\n try:\n watch_obj = observer.schedule(self, new_watch_path,\n recursive=True)\n watche_obj_map[new_watch_path] = watch_obj\n except OSError:\n watche_obj_map[new_watch_path] = None\n for old_watch_path in old_watch_path_s:\n watch_obj = watche_obj_map.pop(old_watch_path, None)\n if watch_obj is not None:\n observer.unschedule(watch_obj)\n self._watch_paths = new_watch_path_s\n time.sleep(self._interval)\n",
"role": "user"
},
{
"content": "run_watcher",
"role": "assistant"
}
] |
def x(self):
watch_path_s = set(os.path.abspath(x) for x in sys.path)
for extra_path in (self._extra_paths or ()):
extra_dir_path = os.path.dirname(os.path.abspath(extra_path))
watch_path_s.add(extra_dir_path)
for module in list(sys.modules.values()):
module_path = getattr(module, '__file__', None)
if module_path is not None:
module_dir_path = os.path.dirname(os.path.abspath(module_path))
watch_path_s.add(module_dir_path)
watch_path_s = self._find_short_paths(watch_path_s)
return watch_path_s
| find_watch_paths | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n watch_path_s = set(os.path.abspath(x) for x in sys.path)\n for extra_path in (self._extra_paths or ()):\n extra_dir_path = os.path.dirname(os.path.abspath(extra_path))\n watch_path_s.add(extra_dir_path)\n for module in list(sys.modules.values()):\n module_path = getattr(module, '__file__', None)\n if module_path is not None:\n module_dir_path = os.path.dirname(os.path.abspath(module_path))\n watch_path_s.add(module_dir_path)\n watch_path_s = self._find_short_paths(watch_path_s)\n return watch_path_s\n",
"role": "user"
},
{
"content": "find_watch_paths",
"role": "assistant"
}
] |
def x(self, paths):
path_parts_s = [path.split(os.path.sep) for path in paths]
root_node = {}
for parts in sorted(path_parts_s, key=len, reverse=True):
node = root_node
for part in parts:
node = node.setdefault(part, {})
node.clear()
short_path_s = set()
self._collect_leaf_paths(node=root_node, path_parts=(), leaf_paths=
short_path_s)
return short_path_s
| find_short_paths | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, paths):\n path_parts_s = [path.split(os.path.sep) for path in paths]\n root_node = {}\n for parts in sorted(path_parts_s, key=len, reverse=True):\n node = root_node\n for part in parts:\n node = node.setdefault(part, {})\n node.clear()\n short_path_s = set()\n self._collect_leaf_paths(node=root_node, path_parts=(), leaf_paths=\n short_path_s)\n return short_path_s\n",
"role": "user"
},
{
"content": "find_short_paths",
"role": "assistant"
}
] |
def x(self, node, path_parts, leaf_paths):
if not node:
node_path = '/'.join(path_parts)
leaf_paths.add(node_path)
else:
for child_path_part, child_node in node.items():
child_path_part_s = path_parts + (child_path_part,)
self._collect_leaf_paths(node=child_node, path_parts=
child_path_part_s, leaf_paths=leaf_paths)
| collect_leaf_paths | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, node, path_parts, leaf_paths):\n if not node:\n node_path = '/'.join(path_parts)\n leaf_paths.add(node_path)\n else:\n for child_path_part, child_node in node.items():\n child_path_part_s = path_parts + (child_path_part,)\n self._collect_leaf_paths(node=child_node, path_parts=\n child_path_part_s, leaf_paths=leaf_paths)\n",
"role": "user"
},
{
"content": "collect_leaf_paths",
"role": "assistant"
}
] |
def x(self, event):
file_path = event.src_path
if file_path in self._extra_paths:
self.reload()
if file_path.endswith(('.pyc', '.pyo')):
file_path = file_path[:-1]
if file_path.endswith('.py'):
file_dir = os.path.dirname(file_path)
if file_dir.startswith(tuple(self._watch_paths)):
self.reload()
| dispatch | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, event):\n file_path = event.src_path\n if file_path in self._extra_paths:\n self.reload()\n if file_path.endswith(('.pyc', '.pyo')):\n file_path = file_path[:-1]\n if file_path.endswith('.py'):\n file_dir = os.path.dirname(file_path)\n if file_dir.startswith(tuple(self._watch_paths)):\n self.reload()\n",
"role": "user"
},
{
"content": "dispatch",
"role": "assistant"
}
] |
def x(self):
reload_mode = self._reload_mode
if self._reload_mode == self.RELOAD_MODE_V_EXEC:
self.reload_using_exec()
elif self._reload_mode == self.RELOAD_MODE_V_SPAWN_EXIT:
self.reload_using_spawn_exit()
elif self._reload_mode == self.RELOAD_MODE_V_SPAWN_WAIT:
self.reload_using_spawn_wait()
else:
error_msg = 'Invalid reload mode: {}.'.format(repr(reload_mode))
raise ValueError(error_msg)
| reload | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n reload_mode = self._reload_mode\n if self._reload_mode == self.RELOAD_MODE_V_EXEC:\n self.reload_using_exec()\n elif self._reload_mode == self.RELOAD_MODE_V_SPAWN_EXIT:\n self.reload_using_spawn_exit()\n elif self._reload_mode == self.RELOAD_MODE_V_SPAWN_WAIT:\n self.reload_using_spawn_wait()\n else:\n error_msg = 'Invalid reload mode: {}.'.format(repr(reload_mode))\n raise ValueError(error_msg)\n",
"role": "user"
},
{
"content": "reload",
"role": "assistant"
}
] |
def x(self):
cmd_parts = [sys.executable] + sys.argv
env_copy = os.environ.copy()
os.execvpe(sys.executable, cmd_parts, env_copy)
| reload_using_exec | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n cmd_parts = [sys.executable] + sys.argv\n env_copy = os.environ.copy()\n os.execvpe(sys.executable, cmd_parts, env_copy)\n",
"role": "user"
},
{
"content": "reload_using_exec",
"role": "assistant"
}
] |
def x(self):
cmd_parts = [sys.executable] + sys.argv
env_copy = os.environ.copy()
subprocess.Popen(cmd_parts, env=env_copy, close_fds=True)
if self._force_exit:
os._exit(0)
else:
interrupt_main()
self._watcher_to_stop = True
sys.exit(0)
| reload_using_spawn_exit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n cmd_parts = [sys.executable] + sys.argv\n env_copy = os.environ.copy()\n subprocess.Popen(cmd_parts, env=env_copy, close_fds=True)\n if self._force_exit:\n os._exit(0)\n else:\n interrupt_main()\n self._watcher_to_stop = True\n sys.exit(0)\n",
"role": "user"
},
{
"content": "reload_using_spawn_exit",
"role": "assistant"
}
] |
def x(self):
cmd_parts = [sys.executable] + sys.argv
env_copy = os.environ.copy()
interrupt_main()
subprocess.call(cmd_parts, env=env_copy, close_fds=True)
sys.exit(0)
| reload_using_spawn_wait | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n cmd_parts = [sys.executable] + sys.argv\n env_copy = os.environ.copy()\n interrupt_main()\n subprocess.call(cmd_parts, env=env_copy, close_fds=True)\n sys.exit(0)\n",
"role": "user"
},
{
"content": "reload_using_spawn_wait",
"role": "assistant"
}
] |
def x(agent_id):
fields = [('ID', 'id'), ('Status', 'status'), ('Region', 'region'), (
'First Contact', 'first_contact'), ('CPU Usage (%)', 'cpu_cur_pct'),
('Used Memory (MiB)', 'mem_cur_bytes'), ('Total slots',
'available_slots'), ('Occupied slots', 'occupied_slots')]
if is_legacy_server():
del fields[9]
del fields[6]
q = 'query($agent_id:String!) { agent(agent_id:$agent_id) { $fields }}'
q = q.replace('$fields', ' '.join(item[1] for item in fields))
v = {'agent_id': agent_id}
with Session() as session:
try:
resp = session.Admin.query(q, v)
except Exception as e:
print_error(e)
sys.exit(1)
info = resp['agent']
rows = []
for name, key in fields:
if key == 'mem_cur_bytes' and info[key] is not None:
info[key] = round(info[key] / 2 ** 20, 1)
if key in info:
rows.append((name, info[key]))
print(tabulate(rows, headers=('Field', 'Value')))
| agent | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(agent_id):\n fields = [('ID', 'id'), ('Status', 'status'), ('Region', 'region'), (\n 'First Contact', 'first_contact'), ('CPU Usage (%)', 'cpu_cur_pct'),\n ('Used Memory (MiB)', 'mem_cur_bytes'), ('Total slots',\n 'available_slots'), ('Occupied slots', 'occupied_slots')]\n if is_legacy_server():\n del fields[9]\n del fields[6]\n q = 'query($agent_id:String!) { agent(agent_id:$agent_id) { $fields }}'\n q = q.replace('$fields', ' '.join(item[1] for item in fields))\n v = {'agent_id': agent_id}\n with Session() as session:\n try:\n resp = session.Admin.query(q, v)\n except Exception as e:\n print_error(e)\n sys.exit(1)\n info = resp['agent']\n rows = []\n for name, key in fields:\n if key == 'mem_cur_bytes' and info[key] is not None:\n info[key] = round(info[key] / 2 ** 20, 1)\n if key in info:\n rows.append((name, info[key]))\n print(tabulate(rows, headers=('Field', 'Value')))\n",
"role": "user"
},
{
"content": "agent",
"role": "assistant"
}
] |
def x(status, all):
fields = [('ID', 'id'), ('Status', 'status'), ('Region', 'region'), (
'First Contact', 'first_contact'), ('CPU Usage (%)', 'cpu_cur_pct'),
('Used Memory (MiB)', 'mem_cur_bytes'), ('Total slots',
'available_slots'), ('Occupied slots', 'occupied_slots')]
if is_legacy_server():
del fields[9]
del fields[6]
def x(limit, offset):
try:
resp_agents = session.Agent.list_with_limit(limit, offset,
status, fields=(item[1] for item in fields))
except Exception as e:
print_error(e)
sys.exit(1)
return resp_agents
def x(results):
for item in results:
if 'mem_cur_bytes' in item and item['mem_cur_bytes'] is not None:
item['mem_cur_bytes'] = round(item['mem_cur_bytes'] / 2 **
20, 1)
return results
def x(interval):
offset = 0
is_first = True
total_count = -1
while True:
limit = interval if is_first else min(interval, total_count -
offset)
try:
result = execute_paginated_query(limit, offset)
except Exception as e:
print_error(e)
sys.exit(1)
offset += interval
total_count = result['total_count']
items = result['items']
items = round_mem(items)
table = tabulate((item.values() for item in items), headers=(
item[0] for item in fields))
if is_first:
is_first = False
else:
table_rows = table.split('\n')
table = '\n'.join(table_rows[2:])
yield table + '\n'
if not offset < total_count:
break
with Session() as session:
paginating_interval = 10
if all:
click.echo_via_pager(_generate_paginated_results(
paginating_interval))
else:
result = execute_paginated_query(paginating_interval, offset=0)
total_count = result['total_count']
if total_count == 0:
print('There are no matching agents.')
return
items = result['items']
items = round_mem(items)
fields = [field for field in fields if field[1] in items[0]]
print(tabulate((item.values() for item in items), headers=(item
[0] for item in fields)))
if total_count > paginating_interval:
print('More agents can be displayed by using --all option.')
| agents | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(status, all):\n fields = [('ID', 'id'), ('Status', 'status'), ('Region', 'region'), (\n 'First Contact', 'first_contact'), ('CPU Usage (%)', 'cpu_cur_pct'),\n ('Used Memory (MiB)', 'mem_cur_bytes'), ('Total slots',\n 'available_slots'), ('Occupied slots', 'occupied_slots')]\n if is_legacy_server():\n del fields[9]\n del fields[6]\n\n def x(limit, offset):\n try:\n resp_agents = session.Agent.list_with_limit(limit, offset,\n status, fields=(item[1] for item in fields))\n except Exception as e:\n print_error(e)\n sys.exit(1)\n return resp_agents\n\n def x(results):\n for item in results:\n if 'mem_cur_bytes' in item and item['mem_cur_bytes'] is not None:\n item['mem_cur_bytes'] = round(item['mem_cur_bytes'] / 2 ** \n 20, 1)\n return results\n\n def x(interval):\n offset = 0\n is_first = True\n total_count = -1\n while True:\n limit = interval if is_first else min(interval, total_count -\n offset)\n try:\n result = execute_paginated_query(limit, offset)\n except Exception as e:\n print_error(e)\n sys.exit(1)\n offset += interval\n total_count = result['total_count']\n items = result['items']\n items = round_mem(items)\n table = tabulate((item.values() for item in items), headers=(\n item[0] for item in fields))\n if is_first:\n is_first = False\n else:\n table_rows = table.split('\\n')\n table = '\\n'.join(table_rows[2:])\n yield table + '\\n'\n if not offset < total_count:\n break\n with Session() as session:\n paginating_interval = 10\n if all:\n click.echo_via_pager(_generate_paginated_results(\n paginating_interval))\n else:\n result = execute_paginated_query(paginating_interval, offset=0)\n total_count = result['total_count']\n if total_count == 0:\n print('There are no matching agents.')\n return\n items = result['items']\n items = round_mem(items)\n fields = [field for field in fields if field[1] in items[0]]\n print(tabulate((item.values() for item in items), headers=(item\n [0] for item in fields)))\n if total_count > paginating_interval:\n print('More agents can be displayed by using --all option.')\n",
"role": "user"
},
{
"content": "agents",
"role": "assistant"
}
] |
def x(name):
fields = [('Name', 'name'), ('Created At', 'created_at'), (
'Default for Unspecified', 'default_for_unspecified'), (
'Total Resource Slot', 'total_resource_slots'), (
'Max Concurrent Sessions', 'max_concurrent_sessions'), (
'Max Containers per Session', 'max_containers_per_session'), (
'Max vFolder Count', 'max_vfolder_count'), ('Max vFolder Size',
'max_vfolder_size'), ('Idle Timeeout', 'idle_timeout'), (
'Allowed vFolder Hosts', 'allowed_vfolder_hosts')]
with Session() as session:
try:
rp = session.ResourcePolicy(session.config.access_key)
info = rp.info(name, fields=(item[1] for item in fields))
except Exception as e:
print_error(e)
sys.exit(1)
rows = []
if info is None:
print('No such resource policy.')
sys.exit(1)
for name, key in fields:
rows.append((name, info[key]))
print(tabulate(rows, headers=('Field', 'Value')))
| resource_policy | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(name):\n fields = [('Name', 'name'), ('Created At', 'created_at'), (\n 'Default for Unspecified', 'default_for_unspecified'), (\n 'Total Resource Slot', 'total_resource_slots'), (\n 'Max Concurrent Sessions', 'max_concurrent_sessions'), (\n 'Max Containers per Session', 'max_containers_per_session'), (\n 'Max vFolder Count', 'max_vfolder_count'), ('Max vFolder Size',\n 'max_vfolder_size'), ('Idle Timeeout', 'idle_timeout'), (\n 'Allowed vFolder Hosts', 'allowed_vfolder_hosts')]\n with Session() as session:\n try:\n rp = session.ResourcePolicy(session.config.access_key)\n info = rp.info(name, fields=(item[1] for item in fields))\n except Exception as e:\n print_error(e)\n sys.exit(1)\n rows = []\n if info is None:\n print('No such resource policy.')\n sys.exit(1)\n for name, key in fields:\n rows.append((name, info[key]))\n print(tabulate(rows, headers=('Field', 'Value')))\n",
"role": "user"
},
{
"content": "resource_policy",
"role": "assistant"
}
] |
def x(ctx):
if ctx.invoked_subcommand is not None:
return
fields = [('Name', 'name'), ('Created At', 'created_at'), (
'Default for Unspecified', 'default_for_unspecified'), (
'Total Resource Slot', 'total_resource_slots'), (
'Max Concurrent Sessions', 'max_concurrent_sessions'), (
'Max Containers per Session', 'max_containers_per_session'), (
'Max vFolder Count', 'max_vfolder_count'), ('Max vFolder Size',
'max_vfolder_size'), ('Idle Timeeout', 'idle_timeout'), (
'Allowed vFolder Hosts', 'allowed_vfolder_hosts')]
with Session() as session:
try:
items = session.ResourcePolicy.list(fields=(item[1] for item in
fields))
except Exception as e:
print_error(e)
sys.exit(1)
if len(items) == 0:
print('There are no keypair resource policies.')
return
print(tabulate((item.values() for item in items), headers=(item[0] for
item in fields)))
| resource_policies | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ctx):\n if ctx.invoked_subcommand is not None:\n return\n fields = [('Name', 'name'), ('Created At', 'created_at'), (\n 'Default for Unspecified', 'default_for_unspecified'), (\n 'Total Resource Slot', 'total_resource_slots'), (\n 'Max Concurrent Sessions', 'max_concurrent_sessions'), (\n 'Max Containers per Session', 'max_containers_per_session'), (\n 'Max vFolder Count', 'max_vfolder_count'), ('Max vFolder Size',\n 'max_vfolder_size'), ('Idle Timeeout', 'idle_timeout'), (\n 'Allowed vFolder Hosts', 'allowed_vfolder_hosts')]\n with Session() as session:\n try:\n items = session.ResourcePolicy.list(fields=(item[1] for item in\n fields))\n except Exception as e:\n print_error(e)\n sys.exit(1)\n if len(items) == 0:\n print('There are no keypair resource policies.')\n return\n print(tabulate((item.values() for item in items), headers=(item[0] for\n item in fields)))\n",
"role": "user"
},
{
"content": "resource_policies",
"role": "assistant"
}
] |
def x(name, default_for_unspecified, total_resource_slots,
max_concurrent_sessions, max_containers_per_session, max_vfolder_count,
max_vfolder_size, idle_timeout, allowed_vfolder_hosts):
with Session() as session:
try:
data = session.ResourcePolicy.create(name,
default_for_unspecified=default_for_unspecified,
total_resource_slots=total_resource_slots,
max_concurrent_sessions=max_concurrent_sessions,
max_containers_per_session=max_containers_per_session,
max_vfolder_count=max_vfolder_count, max_vfolder_size=
max_vfolder_size, idle_timeout=idle_timeout,
allowed_vfolder_hosts=allowed_vfolder_hosts)
except Exception as e:
print_error(e)
sys.exit(1)
if not data['ok']:
print_fail('KeyPair Resource Policy creation has failed: {0}'.
format(data['msg']))
sys.exit(1)
item = data['resource_policy']
print('Keypair resource policy ' + item['name'] + ' is created.')
| add | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(name, default_for_unspecified, total_resource_slots,\n max_concurrent_sessions, max_containers_per_session, max_vfolder_count,\n max_vfolder_size, idle_timeout, allowed_vfolder_hosts):\n with Session() as session:\n try:\n data = session.ResourcePolicy.create(name,\n default_for_unspecified=default_for_unspecified,\n total_resource_slots=total_resource_slots,\n max_concurrent_sessions=max_concurrent_sessions,\n max_containers_per_session=max_containers_per_session,\n max_vfolder_count=max_vfolder_count, max_vfolder_size=\n max_vfolder_size, idle_timeout=idle_timeout,\n allowed_vfolder_hosts=allowed_vfolder_hosts)\n except Exception as e:\n print_error(e)\n sys.exit(1)\n if not data['ok']:\n print_fail('KeyPair Resource Policy creation has failed: {0}'.\n format(data['msg']))\n sys.exit(1)\n item = data['resource_policy']\n print('Keypair resource policy ' + item['name'] + ' is created.')\n",
"role": "user"
},
{
"content": "add",
"role": "assistant"
}
] |
def x(name):
with Session() as session:
if input('Are you sure? (y/n): ').lower().strip()[:1] != 'y':
print('Canceled.')
sys.exit(1)
try:
data = session.ResourcePolicy.delete(name)
except Exception as e:
print_error(e)
sys.exit(1)
if not data['ok']:
print_fail('KeyPair Resource Policy deletion has failed: {0}'.
format(data['msg']))
sys.exit(1)
print('Resource policy ' + name + ' is deleted.')
| delete | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(name):\n with Session() as session:\n if input('Are you sure? (y/n): ').lower().strip()[:1] != 'y':\n print('Canceled.')\n sys.exit(1)\n try:\n data = session.ResourcePolicy.delete(name)\n except Exception as e:\n print_error(e)\n sys.exit(1)\n if not data['ok']:\n print_fail('KeyPair Resource Policy deletion has failed: {0}'.\n format(data['msg']))\n sys.exit(1)\n print('Resource policy ' + name + ' is deleted.')\n",
"role": "user"
},
{
"content": "delete",
"role": "assistant"
}
] |
def x(session_id, app, bind, port):
api_session = None
runner = None
async def app_setup():
nonlocal api_session, runner
loop = current_loop()
api_session = AsyncSession()
protocol = 'http'
runner = ProxyRunner(api_session, session_id, app, protocol, bind,
port, loop=loop)
await runner.ready()
print_info('A local proxy to the application "{0}" '.format(app) +
'provided by the session "{0}" '.format(session_id) +
'is available at: {0}://{1}:{2}'.format(protocol, bind, port))
async def app_shutdown():
nonlocal api_session, runner
print_info('Shutting down....')
await runner.close()
await api_session.close()
print_info('The local proxy to "{}" has terminated.'.format(app))
asyncio_run_forever(app_setup(), app_shutdown(), stop_signals={signal.
SIGINT, signal.SIGTERM})
| app | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(session_id, app, bind, port):\n api_session = None\n runner = None\n\n async def app_setup():\n nonlocal api_session, runner\n loop = current_loop()\n api_session = AsyncSession()\n protocol = 'http'\n runner = ProxyRunner(api_session, session_id, app, protocol, bind,\n port, loop=loop)\n await runner.ready()\n print_info('A local proxy to the application \"{0}\" '.format(app) +\n 'provided by the session \"{0}\" '.format(session_id) +\n 'is available at: {0}://{1}:{2}'.format(protocol, bind, port))\n\n async def app_shutdown():\n nonlocal api_session, runner\n print_info('Shutting down....')\n await runner.close()\n await api_session.close()\n print_info('The local proxy to \"{}\" has terminated.'.format(app))\n asyncio_run_forever(app_setup(), app_shutdown(), stop_signals={signal.\n SIGINT, signal.SIGTERM})\n",
"role": "user"
},
{
"content": "app",
"role": "assistant"
}
] |
def x(key: str, default: Any=None, clean: Callable[[str], Any]=lambda v: v):
key = key.upper()
v = os.environ.get('BACKEND_' + key)
if v is None:
v = os.environ.get('SORNA_' + key)
if v is None:
if default is None:
raise KeyError(key)
v = default
return clean(v)
| get_env | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(key: str, default: Any=None, clean: Callable[[str], Any]=lambda v: v):\n key = key.upper()\n v = os.environ.get('BACKEND_' + key)\n if v is None:\n v = os.environ.get('SORNA_' + key)\n if v is None:\n if default is None:\n raise KeyError(key)\n v = default\n return clean(v)\n",
"role": "user"
},
{
"content": "get_env",
"role": "assistant"
}
] |
def x(sess_id_or_alias):
with Session() as session:
try:
print_wait('Retrieving container logs...')
kernel = session.Kernel(sess_id_or_alias)
result = kernel.get_logs().get('result')
logs = result.get('logs') if 'logs' in result else ''
print(logs)
print_done('End of logs.')
except Exception as e:
print_error(e)
sys.exit(1)
| logs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sess_id_or_alias):\n with Session() as session:\n try:\n print_wait('Retrieving container logs...')\n kernel = session.Kernel(sess_id_or_alias)\n result = kernel.get_logs().get('result')\n logs = result.get('logs') if 'logs' in result else ''\n print(logs)\n print_done('End of logs.')\n except Exception as e:\n print_error(e)\n sys.exit(1)\n",
"role": "user"
},
{
"content": "logs",
"role": "assistant"
}
] |
def x(function, args, kws):
return hashlib.md5(pickle.dumps((_from_file(function) + function.
__name__, args, kws))).hexdigest()
| wrap_key | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(function, args, kws):\n return hashlib.md5(pickle.dumps((_from_file(function) + function.\n __name__, args, kws))).hexdigest()\n",
"role": "user"
},
{
"content": "wrap_key",
"role": "assistant"
}
] |
def x(key, adapter=MemoryAdapter):
try:
return pickle.loads(adapter().get(key))
except CacheExpiredException:
return None
| get | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(key, adapter=MemoryAdapter):\n try:\n return pickle.loads(adapter().get(key))\n except CacheExpiredException:\n return None\n",
"role": "user"
},
{
"content": "get",
"role": "assistant"
}
] |
def x(key, value, timeout=-1, adapter=MemoryAdapter):
if adapter(timeout=timeout).set(key, pickle.dumps(value)):
return value
else:
return None
| set | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(key, value, timeout=-1, adapter=MemoryAdapter):\n if adapter(timeout=timeout).set(key, pickle.dumps(value)):\n return value\n else:\n return None\n",
"role": "user"
},
{
"content": "set",
"role": "assistant"
}
] |
def x(timeout=-1, adapter=MemoryAdapter):
def x(function):
@wraps(function)
def x(*args, **kws):
hash_key = _wrap_key(function, args, kws)
try:
adapter_instance = adapter()
return pickle.loads(adapter_instance.get(hash_key))
except CacheExpiredException:
value = function(*args, **kws)
set(hash_key, value, timeout, adapter)
return value
return __wrapcache
return _wrapcache
| wrapcache | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(timeout=-1, adapter=MemoryAdapter):\n\n def x(function):\n\n @wraps(function)\n def x(*args, **kws):\n hash_key = _wrap_key(function, args, kws)\n try:\n adapter_instance = adapter()\n return pickle.loads(adapter_instance.get(hash_key))\n except CacheExpiredException:\n value = function(*args, **kws)\n set(hash_key, value, timeout, adapter)\n return value\n return __wrapcache\n return _wrapcache\n",
"role": "user"
},
{
"content": "wrapcache",
"role": "assistant"
}
] |
def x(data, output_file_name, result_info_key, identifier_keys):
workbook = create_excel_workbook(data, result_info_key, identifier_keys)
workbook.save(output_file_name)
print('Saved Excel file to {}'.format(output_file_name))
| export_analytics_data_to_excel | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(data, output_file_name, result_info_key, identifier_keys):\n workbook = create_excel_workbook(data, result_info_key, identifier_keys)\n workbook.save(output_file_name)\n print('Saved Excel file to {}'.format(output_file_name))\n",
"role": "user"
},
{
"content": "export_analytics_data_to_excel",
"role": "assistant"
}
] |
def x(data, output_folder, result_info_key, identifier_keys):
workbook = create_excel_workbook(data, result_info_key, identifier_keys)
suffix = '.csv'
if not os.path.exists(output_folder):
os.makedirs(output_folder)
for worksheet in workbook.worksheets:
file_name = utilities.convert_title_to_snake_case(worksheet.title)
file_path = os.path.join(output_folder, file_name + suffix)
mode = 'w'
if sys.version_info[0] < 3:
mode = 'wb'
with io.open(file_path, mode) as output_file:
csv_writer = csv.writer(output_file)
for row in worksheet.rows:
csv_writer.writerow([cell.value for cell in row])
print('Saved CSV files to {}'.format(output_folder))
| export_analytics_data_to_csv | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(data, output_folder, result_info_key, identifier_keys):\n workbook = create_excel_workbook(data, result_info_key, identifier_keys)\n suffix = '.csv'\n if not os.path.exists(output_folder):\n os.makedirs(output_folder)\n for worksheet in workbook.worksheets:\n file_name = utilities.convert_title_to_snake_case(worksheet.title)\n file_path = os.path.join(output_folder, file_name + suffix)\n mode = 'w'\n if sys.version_info[0] < 3:\n mode = 'wb'\n with io.open(file_path, mode) as output_file:\n csv_writer = csv.writer(output_file)\n for row in worksheet.rows:\n csv_writer.writerow([cell.value for cell in row])\n print('Saved CSV files to {}'.format(output_folder))\n",
"role": "user"
},
{
"content": "export_analytics_data_to_csv",
"role": "assistant"
}
] |
def x(addresses, output_file_name, endpoint, report_type, retry, api_key,
api_secret, files_path):
master_workbook = openpyxl.Workbook()
if api_key is not None and api_secret is not None:
client = ApiClient(api_key, api_secret)
else:
client = ApiClient()
errors = []
for index, addr in enumerate(addresses):
print('Processing {}'.format(addr[0]))
result = _get_excel_report(client, endpoint, addr[0], addr[1],
report_type, retry)
if not result['success']:
print('Error retrieving report for {}'.format(addr[0]))
print(result['content'])
errors.append({'address': addr[0], 'message': result['content']})
continue
orig_wb = openpyxl.load_workbook(filename=io.BytesIO(result['content'])
)
_save_individual_file(orig_wb, files_path, addr[0])
for sheet_name in orig_wb.get_sheet_names():
if sheet_name in master_workbook.get_sheet_names():
master_ws = master_workbook.get_sheet_by_name(sheet_name)
else:
master_ws = master_workbook.create_sheet(sheet_name)
orig_rows = orig_wb.get_sheet_by_name(sheet_name).rows
if sheet_name == 'Summary' or sheet_name == 'Chart Data':
_process_non_standard_sheet(master_ws, orig_rows, addr, index)
continue
_process_standard_sheet(master_ws, orig_rows, addr, index)
master_workbook.remove(master_workbook.worksheets[0])
if len(errors) > 0:
errors_sheet = master_workbook.create_sheet('Errors')
for error_idx, error in enumerate(errors):
errors_sheet.cell(row=error_idx + 1, column=1, value=error[
'address'])
errors_sheet.cell(row=error_idx + 1, column=2, value=error[
'message'])
adjust_column_width_workbook(master_workbook)
output_file_path = os.path.join(files_path, output_file_name)
master_workbook.save(output_file_path)
print('Saved output to {}'.format(output_file_path))
| concat_excel_reports | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(addresses, output_file_name, endpoint, report_type, retry, api_key,\n api_secret, files_path):\n master_workbook = openpyxl.Workbook()\n if api_key is not None and api_secret is not None:\n client = ApiClient(api_key, api_secret)\n else:\n client = ApiClient()\n errors = []\n for index, addr in enumerate(addresses):\n print('Processing {}'.format(addr[0]))\n result = _get_excel_report(client, endpoint, addr[0], addr[1],\n report_type, retry)\n if not result['success']:\n print('Error retrieving report for {}'.format(addr[0]))\n print(result['content'])\n errors.append({'address': addr[0], 'message': result['content']})\n continue\n orig_wb = openpyxl.load_workbook(filename=io.BytesIO(result['content'])\n )\n _save_individual_file(orig_wb, files_path, addr[0])\n for sheet_name in orig_wb.get_sheet_names():\n if sheet_name in master_workbook.get_sheet_names():\n master_ws = master_workbook.get_sheet_by_name(sheet_name)\n else:\n master_ws = master_workbook.create_sheet(sheet_name)\n orig_rows = orig_wb.get_sheet_by_name(sheet_name).rows\n if sheet_name == 'Summary' or sheet_name == 'Chart Data':\n _process_non_standard_sheet(master_ws, orig_rows, addr, index)\n continue\n _process_standard_sheet(master_ws, orig_rows, addr, index)\n master_workbook.remove(master_workbook.worksheets[0])\n if len(errors) > 0:\n errors_sheet = master_workbook.create_sheet('Errors')\n for error_idx, error in enumerate(errors):\n errors_sheet.cell(row=error_idx + 1, column=1, value=error[\n 'address'])\n errors_sheet.cell(row=error_idx + 1, column=2, value=error[\n 'message'])\n adjust_column_width_workbook(master_workbook)\n output_file_path = os.path.join(files_path, output_file_name)\n master_workbook.save(output_file_path)\n print('Saved output to {}'.format(output_file_path))\n",
"role": "user"
},
{
"content": "concat_excel_reports",
"role": "assistant"
}
] |
def x(data, result_info_key, identifier_keys):
workbook = analytics_data_excel.get_excel_workbook(data,
result_info_key, identifier_keys)
adjust_column_width_workbook(workbook)
return workbook
| create_excel_workbook | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(data, result_info_key, identifier_keys):\n workbook = analytics_data_excel.get_excel_workbook(data,\n result_info_key, identifier_keys)\n adjust_column_width_workbook(workbook)\n return workbook\n",
"role": "user"
},
{
"content": "create_excel_workbook",
"role": "assistant"
}
] |
def x(worksheet):
dims = {}
padding = 1
for row in worksheet.rows:
for cell in row:
if not cell.value:
continue
dims[cell.column] = max(dims.get(cell.column, 0), len(str(cell.
value)))
for col, value in list(dims.items()):
worksheet.column_dimensions[col].width = value + padding
| adjust_column_width | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(worksheet):\n dims = {}\n padding = 1\n for row in worksheet.rows:\n for cell in row:\n if not cell.value:\n continue\n dims[cell.column] = max(dims.get(cell.column, 0), len(str(cell.\n value)))\n for col, value in list(dims.items()):\n worksheet.column_dimensions[col].width = value + padding\n",
"role": "user"
},
{
"content": "adjust_column_width",
"role": "assistant"
}
] |
def x(ipaddress, auth, url):
get_ap_info_url = (
'/imcrs/wlan/apInfo/queryApBasicInfoByCondition?ipAddress=' + str(
ipaddress))
f_url = url + get_ap_info_url
payload = None
r = requests.get(f_url, auth=auth, headers=HEADERS)
try:
if r.status_code == 200:
if len(r.text) > 0:
return json.loads(r.text)['apBasicInfo']
except requests.exceptions.RequestException as e:
return 'Error:\n' + str(e) + ' get_ap_info_all: An Error has occured'
| get_ap_info | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ipaddress, auth, url):\n get_ap_info_url = (\n '/imcrs/wlan/apInfo/queryApBasicInfoByCondition?ipAddress=' + str(\n ipaddress))\n f_url = url + get_ap_info_url\n payload = None\n r = requests.get(f_url, auth=auth, headers=HEADERS)\n try:\n if r.status_code == 200:\n if len(r.text) > 0:\n return json.loads(r.text)['apBasicInfo']\n except requests.exceptions.RequestException as e:\n return 'Error:\\n' + str(e) + ' get_ap_info_all: An Error has occured'\n",
"role": "user"
},
{
"content": "get_ap_info",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.