source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(self, model_id, filename, token=None, url=API_GET_PREDICTION_IMAGE_URL):
auth = 'Bearer ' + self.check_for_token(token)
h = {'Authorization': auth, 'Cache-Control': 'no-cache'}
the_url = url
with open(filename, 'rb') as image_file:
encoded_string = base64.b64encode(image_file.read())
m = MultipartEncoder(fields={'sampleBase64Content': encoded_string,
'modelId': model_id})
h = {'Authorization': auth, 'Cache-Control': 'no-cache', 'Content-Type':
m.content_type}
r = requests.post(the_url, headers=h, data=m)
return r
| get_fileb64_image_prediction | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, model_id, filename, token=None, url=API_GET_PREDICTION_IMAGE_URL):\n auth = 'Bearer ' + self.check_for_token(token)\n h = {'Authorization': auth, 'Cache-Control': 'no-cache'}\n the_url = url\n with open(filename, 'rb') as image_file:\n encoded_string = base64.b64encode(image_file.read())\n m = MultipartEncoder(fields={'sampleBase64Content': encoded_string,\n 'modelId': model_id})\n h = {'Authorization': auth, 'Cache-Control': 'no-cache', 'Content-Type':\n m.content_type}\n r = requests.post(the_url, headers=h, data=m)\n return r\n",
"role": "user"
},
{
"content": "get_fileb64_image_prediction",
"role": "assistant"
}
] |
def x(self, model_id, b64_encoded_string, token=None, url=
API_GET_PREDICTION_IMAGE_URL):
auth = 'Bearer ' + self.check_for_token(token)
h = {'Authorization': auth, 'Cache-Control': 'no-cache'}
the_url = url
encoded_string = b64_encoded_string
m = MultipartEncoder(fields={'sampleBase64Content': encoded_string,
'modelId': model_id})
h = {'Authorization': auth, 'Cache-Control': 'no-cache', 'Content-Type':
m.content_type}
r = requests.post(the_url, headers=h, data=m)
return r
| get_b64_image_prediction | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, model_id, b64_encoded_string, token=None, url=\n API_GET_PREDICTION_IMAGE_URL):\n auth = 'Bearer ' + self.check_for_token(token)\n h = {'Authorization': auth, 'Cache-Control': 'no-cache'}\n the_url = url\n encoded_string = b64_encoded_string\n m = MultipartEncoder(fields={'sampleBase64Content': encoded_string,\n 'modelId': model_id})\n h = {'Authorization': auth, 'Cache-Control': 'no-cache', 'Content-Type':\n m.content_type}\n r = requests.post(the_url, headers=h, data=m)\n return r\n",
"role": "user"
},
{
"content": "get_b64_image_prediction",
"role": "assistant"
}
] |
def x(self, file_url, dataset_type='image', token=None, url=API_CREATE_DATASET
):
auth = 'Bearer ' + self.check_for_token(token)
m = MultipartEncoder(fields={'type': dataset_type, 'path': file_url})
h = {'Authorization': auth, 'Cache-Control': 'no-cache', 'Content-Type':
m.content_type}
the_url = url
r = requests.post(the_url, headers=h, data=m)
return r
| create_dataset_synchronous | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, file_url, dataset_type='image', token=None, url=API_CREATE_DATASET\n ):\n auth = 'Bearer ' + self.check_for_token(token)\n m = MultipartEncoder(fields={'type': dataset_type, 'path': file_url})\n h = {'Authorization': auth, 'Cache-Control': 'no-cache', 'Content-Type':\n m.content_type}\n the_url = url\n r = requests.post(the_url, headers=h, data=m)\n return r\n",
"role": "user"
},
{
"content": "create_dataset_synchronous",
"role": "assistant"
}
] |
def x(self, dataset_id, model_name, token=None, url=API_TRAIN_MODEL):
auth = 'Bearer ' + self.check_for_token(token)
m = MultipartEncoder(fields={'name': model_name, 'datasetId': dataset_id})
h = {'Authorization': auth, 'Cache-Control': 'no-cache', 'Content-Type':
m.content_type}
the_url = url
r = requests.post(the_url, headers=h, data=m)
return r
| train_model | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dataset_id, model_name, token=None, url=API_TRAIN_MODEL):\n auth = 'Bearer ' + self.check_for_token(token)\n m = MultipartEncoder(fields={'name': model_name, 'datasetId': dataset_id})\n h = {'Authorization': auth, 'Cache-Control': 'no-cache', 'Content-Type':\n m.content_type}\n the_url = url\n r = requests.post(the_url, headers=h, data=m)\n return r\n",
"role": "user"
},
{
"content": "train_model",
"role": "assistant"
}
] |
def x(self, model_id, token=None, url=API_TRAIN_MODEL):
auth = 'Bearer ' + self.check_for_token(token)
h = {'Authorization': auth, 'Cache-Control': 'no-cache'}
the_url = url + '/' + model_id
r = requests.get(the_url, headers=h)
return r
| get_training_status | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, model_id, token=None, url=API_TRAIN_MODEL):\n auth = 'Bearer ' + self.check_for_token(token)\n h = {'Authorization': auth, 'Cache-Control': 'no-cache'}\n the_url = url + '/' + model_id\n r = requests.get(the_url, headers=h)\n return r\n",
"role": "user"
},
{
"content": "get_training_status",
"role": "assistant"
}
] |
def x(self, dataset_id, token=None, url=API_GET_MODELS):
auth = 'Bearer ' + self.check_for_token(token)
h = {'Authorization': auth, 'Cache-Control': 'no-cache'}
if url != API_GET_MODELS:
r = requests.get(the_url, headers=h)
return r
the_url = url.replace('<dataset_id>', dataset_id)
r = requests.get(the_url, headers=h)
return r
| get_models_info_for_dataset | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dataset_id, token=None, url=API_GET_MODELS):\n auth = 'Bearer ' + self.check_for_token(token)\n h = {'Authorization': auth, 'Cache-Control': 'no-cache'}\n if url != API_GET_MODELS:\n r = requests.get(the_url, headers=h)\n return r\n the_url = url.replace('<dataset_id>', dataset_id)\n r = requests.get(the_url, headers=h)\n return r\n",
"role": "user"
},
{
"content": "get_models_info_for_dataset",
"role": "assistant"
}
] |
def x(self, file_url, token=None, url=API_CREATE_LANGUAGE_DATASET):
auth = 'Bearer ' + self.check_for_token(token)
dummy_files = {'type': (None, 'text-intent'), 'path': (None, file_url)}
h = {'Authorization': auth, 'Cache-Control': 'no-cache'}
the_url = url
r = requests.post(the_url, headers=h, files=dummy_files)
return r
| create_language_dataset_from_url | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, file_url, token=None, url=API_CREATE_LANGUAGE_DATASET):\n auth = 'Bearer ' + self.check_for_token(token)\n dummy_files = {'type': (None, 'text-intent'), 'path': (None, file_url)}\n h = {'Authorization': auth, 'Cache-Control': 'no-cache'}\n the_url = url\n r = requests.post(the_url, headers=h, files=dummy_files)\n return r\n",
"role": "user"
},
{
"content": "create_language_dataset_from_url",
"role": "assistant"
}
] |
def x(self, dataset_id, name, token=None, url=API_TRAIN_LANGUAGE_MODEL):
auth = 'Bearer ' + self.check_for_token(token)
dummy_files = {'name': (None, name), 'datasetId': (None, dataset_id)}
h = {'Authorization': auth, 'Cache-Control': 'no-cache'}
the_url = url
r = requests.post(the_url, headers=h, files=dummy_files)
return r
| train_language_model_from_dataset | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dataset_id, name, token=None, url=API_TRAIN_LANGUAGE_MODEL):\n auth = 'Bearer ' + self.check_for_token(token)\n dummy_files = {'name': (None, name), 'datasetId': (None, dataset_id)}\n h = {'Authorization': auth, 'Cache-Control': 'no-cache'}\n the_url = url\n r = requests.post(the_url, headers=h, files=dummy_files)\n return r\n",
"role": "user"
},
{
"content": "train_language_model_from_dataset",
"role": "assistant"
}
] |
def x(self, model_id, token=None, url=API_TRAIN_LANGUAGE_MODEL):
auth = 'Bearer ' + self.check_for_token(token)
h = {'Authorization': auth, 'Cache-Control': 'no-cache'}
the_url = url + '/' + model_id
r = requests.get(the_url, headers=h)
return r
| get_language_model_status | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, model_id, token=None, url=API_TRAIN_LANGUAGE_MODEL):\n auth = 'Bearer ' + self.check_for_token(token)\n h = {'Authorization': auth, 'Cache-Control': 'no-cache'}\n the_url = url + '/' + model_id\n r = requests.get(the_url, headers=h)\n return r\n",
"role": "user"
},
{
"content": "get_language_model_status",
"role": "assistant"
}
] |
def x(self, model_id, document, token=None, url=API_GET_LANGUAGE_PREDICTION):
auth = 'Bearer ' + self.check_for_token(token)
dummy_files = {'modelId': (None, model_id), 'document': (None, document)}
h = {'Authorization': auth, 'Cache-Control': 'no-cache'}
the_url = url
r = requests.post(the_url, headers=h, files=dummy_files)
return r
| get_language_prediction_from_model | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, model_id, document, token=None, url=API_GET_LANGUAGE_PREDICTION):\n auth = 'Bearer ' + self.check_for_token(token)\n dummy_files = {'modelId': (None, model_id), 'document': (None, document)}\n h = {'Authorization': auth, 'Cache-Control': 'no-cache'}\n the_url = url\n r = requests.post(the_url, headers=h, files=dummy_files)\n return r\n",
"role": "user"
},
{
"content": "get_language_prediction_from_model",
"role": "assistant"
}
] |
def x(self):
files = []
files = [f for f in os.listdir() if f[-5:] == '.json']
if len(files) == 0:
print('No json files found in this directory')
return None
max_boxes = 0
rows = []
for each_file in files:
f = open(each_file, 'r')
j = f.read()
j = json.loads(j)
f.close()
if len(j['objects']) > max_boxes:
max_boxes = len(j['objects'])
row = []
for o in j['objects']:
labels = {}
labels['label'] = o['label']
labels['x'] = o['x_y_w_h'][0]
labels['y'] = o['x_y_w_h'][1]
labels['width'] = o['x_y_w_h'][2]
labels['height'] = o['x_y_w_h'][3]
labels_right_format = '"' + json.dumps(labels).replace('"', '""'
) + '"'
row.append(labels_right_format)
row.insert(0, '"' + j['filename'] + '"')
rows.append(row)
rows = [','.join(i) for i in rows]
header = '"image"'
for box_num in range(0, max_boxes):
header += ', "box"' + str(box_num)
rows.insert(0, header)
return rows
| parse_rectlabel_app_output | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n files = []\n files = [f for f in os.listdir() if f[-5:] == '.json']\n if len(files) == 0:\n print('No json files found in this directory')\n return None\n max_boxes = 0\n rows = []\n for each_file in files:\n f = open(each_file, 'r')\n j = f.read()\n j = json.loads(j)\n f.close()\n if len(j['objects']) > max_boxes:\n max_boxes = len(j['objects'])\n row = []\n for o in j['objects']:\n labels = {}\n labels['label'] = o['label']\n labels['x'] = o['x_y_w_h'][0]\n labels['y'] = o['x_y_w_h'][1]\n labels['width'] = o['x_y_w_h'][2]\n labels['height'] = o['x_y_w_h'][3]\n labels_right_format = '\"' + json.dumps(labels).replace('\"', '\"\"'\n ) + '\"'\n row.append(labels_right_format)\n row.insert(0, '\"' + j['filename'] + '\"')\n rows.append(row)\n rows = [','.join(i) for i in rows]\n header = '\"image\"'\n for box_num in range(0, max_boxes):\n header += ', \"box\"' + str(box_num)\n rows.insert(0, header)\n return rows\n",
"role": "user"
},
{
"content": "parse_rectlabel_app_output",
"role": "assistant"
}
] |
def x(self, output_filename='output.csv'):
result = self.parse_rectlabel_app_output()
ff = open(output_filename, 'w', encoding='utf8')
for line in result:
ff.write(line + '\n')
ff.close()
| save_parsed_data_to_csv | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, output_filename='output.csv'):\n result = self.parse_rectlabel_app_output()\n ff = open(output_filename, 'w', encoding='utf8')\n for line in result:\n ff.write(line + '\\n')\n ff.close()\n",
"role": "user"
},
{
"content": "save_parsed_data_to_csv",
"role": "assistant"
}
] |
def x(self, output_filename='output.csv'):
result = self.XML_parse_rectlabel_app_output()
ff = open(output_filename, 'w', encoding='utf8')
for line in result:
ff.write(line + '\n')
ff.close()
| XML_save_parsed_data_to_csv | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, output_filename='output.csv'):\n result = self.XML_parse_rectlabel_app_output()\n ff = open(output_filename, 'w', encoding='utf8')\n for line in result:\n ff.write(line + '\\n')\n ff.close()\n",
"role": "user"
},
{
"content": "XML_save_parsed_data_to_csv",
"role": "assistant"
}
] |
def x(istr):
toks = re.findall('[^\\*\\\\\\+\\-\\^\\(\\)]+\\(?', istr)
return [t for t in toks if not t.endswith('(')]
| tokens | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(istr):\n toks = re.findall('[^\\\\*\\\\\\\\\\\\+\\\\-\\\\^\\\\(\\\\)]+\\\\(?', istr)\n return [t for t in toks if not t.endswith('(')]\n",
"role": "user"
},
{
"content": "tokens",
"role": "assistant"
}
] |
def x(istr):
data = set(i.rstrip('0123456789') for i in tokens(istr))
has_events = not data.isdisjoint(istr_type_evts)
has_2d = not data.isdisjoint(istr_type_2d)
has_1d = data.difference(istr_type_evts).difference(istr_type_2d) != set()
if has_events and not (has_1d or has_2d):
return 'events'
elif has_1d and not has_events:
return '1d'
elif has_2d and not (has_events or has_1d):
return '2d'
else:
return None
| istr_type | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(istr):\n data = set(i.rstrip('0123456789') for i in tokens(istr))\n has_events = not data.isdisjoint(istr_type_evts)\n has_2d = not data.isdisjoint(istr_type_2d)\n has_1d = data.difference(istr_type_evts).difference(istr_type_2d) != set()\n if has_events and not (has_1d or has_2d):\n return 'events'\n elif has_1d and not has_events:\n return '1d'\n elif has_2d and not (has_events or has_1d):\n return '2d'\n else:\n return None\n",
"role": "user"
},
{
"content": "istr_type",
"role": "assistant"
}
] |
def x(istr):
fxn = istr.split('(')[0]
if not fxn.isalnum() and fxn != '(' or istr[-1] != ')':
return False
plevel = 1
for c in '('.join(istr[:-1].split('(')[1:]):
if c == '(':
plevel += 1
elif c == ')':
plevel -= 1
if plevel == 0:
return False
return True
| is_parans_exp | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(istr):\n fxn = istr.split('(')[0]\n if not fxn.isalnum() and fxn != '(' or istr[-1] != ')':\n return False\n plevel = 1\n for c in '('.join(istr[:-1].split('(')[1:]):\n if c == '(':\n plevel += 1\n elif c == ')':\n plevel -= 1\n if plevel == 0:\n return False\n return True\n",
"role": "user"
},
{
"content": "is_parans_exp",
"role": "assistant"
}
] |
def x(istr, analyses, twin=None):
if istr.strip() == '':
return Trace()
istr = istr.lstrip('+')
if istr[0] == '-':
return -parse_ion_string(istr[1:], analyses, twin)
if is_parans_exp(istr):
if ')' not in istr:
pass
fxn = istr.split('(')[0]
args = istr[istr.find('(') + 1:istr.find(')')].split(',')
if fxn == '':
istr = args[0]
else:
ts = parse_ion_string(args[0], analyses, twin)
return ts
if set(istr).intersection(set('+-/*()')) == set():
if istr in SHORTCUTS:
return parse_ion_string(SHORTCUTS[istr], analyses, twin)
elif istr[0] == '!' and all(i in '0123456789.' for i in istr[1:]):
return float(istr[1:])
elif istr == '!pi':
return np.pi
elif istr == '!e':
return np.e
else:
return trace_resolver(istr, analyses, twin)
for token in '/*+-^':
if len(tokenize(istr, token)) != 1:
ts = tokenize(istr, token)
s = parse_ion_string(ts[0], analyses, twin)
for t in ts[1:]:
if token == '/':
s /= parse_ion_string(t, analyses, twin)
elif token == '*':
s *= parse_ion_string(t, analyses, twin)
elif token == '+':
s += parse_ion_string(t, analyses, twin)
elif token == '-':
s -= parse_ion_string(t, analyses, twin)
elif token == '^':
s **= parse_ion_string(t, analyses, twin)
return s
raise Exception("Parser hit a point it shouldn't have!")
| parse_ion_string | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(istr, analyses, twin=None):\n if istr.strip() == '':\n return Trace()\n istr = istr.lstrip('+')\n if istr[0] == '-':\n return -parse_ion_string(istr[1:], analyses, twin)\n if is_parans_exp(istr):\n if ')' not in istr:\n pass\n fxn = istr.split('(')[0]\n args = istr[istr.find('(') + 1:istr.find(')')].split(',')\n if fxn == '':\n istr = args[0]\n else:\n ts = parse_ion_string(args[0], analyses, twin)\n return ts\n if set(istr).intersection(set('+-/*()')) == set():\n if istr in SHORTCUTS:\n return parse_ion_string(SHORTCUTS[istr], analyses, twin)\n elif istr[0] == '!' and all(i in '0123456789.' for i in istr[1:]):\n return float(istr[1:])\n elif istr == '!pi':\n return np.pi\n elif istr == '!e':\n return np.e\n else:\n return trace_resolver(istr, analyses, twin)\n for token in '/*+-^':\n if len(tokenize(istr, token)) != 1:\n ts = tokenize(istr, token)\n s = parse_ion_string(ts[0], analyses, twin)\n for t in ts[1:]:\n if token == '/':\n s /= parse_ion_string(t, analyses, twin)\n elif token == '*':\n s *= parse_ion_string(t, analyses, twin)\n elif token == '+':\n s += parse_ion_string(t, analyses, twin)\n elif token == '-':\n s -= parse_ion_string(t, analyses, twin)\n elif token == '^':\n s **= parse_ion_string(t, analyses, twin)\n return s\n raise Exception(\"Parser hit a point it shouldn't have!\")\n",
"role": "user"
},
{
"content": "parse_ion_string",
"role": "assistant"
}
] |
def x(self, pedalboard, notify=True, force=False):
if pedalboard is not None and pedalboard.bank is None:
raise CurrentPedalboardError('Pedalboard {} has not added in any bank'
.format(pedalboard))
if pedalboard == self.pedalboard and not force:
return
self._pedalboard = pedalboard
self._device_controller.pedalboard = pedalboard
self._save_current_pedalboard()
if notify:
self.app.components_observer.on_current_pedalboard_changed(self.
pedalboard)
| set_pedalboard | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, pedalboard, notify=True, force=False):\n if pedalboard is not None and pedalboard.bank is None:\n raise CurrentPedalboardError('Pedalboard {} has not added in any bank'\n .format(pedalboard))\n if pedalboard == self.pedalboard and not force:\n return\n self._pedalboard = pedalboard\n self._device_controller.pedalboard = pedalboard\n self._save_current_pedalboard()\n if notify:\n self.app.components_observer.on_current_pedalboard_changed(self.\n pedalboard)\n",
"role": "user"
},
{
"content": "set_pedalboard",
"role": "assistant"
}
] |
def x(self):
if self.pedalboard is None:
raise CurrentPedalboardError('The current pedalboard is None')
before_index = self.pedalboard.index - 1
if before_index == -1:
before_index = len(self.bank.pedalboards) - 1
self.set_pedalboard(self.bank.pedalboards[before_index])
| to_before_pedalboard | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.pedalboard is None:\n raise CurrentPedalboardError('The current pedalboard is None')\n before_index = self.pedalboard.index - 1\n if before_index == -1:\n before_index = len(self.bank.pedalboards) - 1\n self.set_pedalboard(self.bank.pedalboards[before_index])\n",
"role": "user"
},
{
"content": "to_before_pedalboard",
"role": "assistant"
}
] |
def x(self):
if self.pedalboard is None:
raise CurrentPedalboardError('The current pedalboard is None')
next_index = self.pedalboard.index + 1
if next_index == len(self.bank.pedalboards):
next_index = 0
self.set_pedalboard(self.bank.pedalboards[next_index])
| to_next_pedalboard | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.pedalboard is None:\n raise CurrentPedalboardError('The current pedalboard is None')\n next_index = self.pedalboard.index + 1\n if next_index == len(self.bank.pedalboards):\n next_index = 0\n self.set_pedalboard(self.bank.pedalboards[next_index])\n",
"role": "user"
},
{
"content": "to_next_pedalboard",
"role": "assistant"
}
] |
def x(self):
if self.pedalboard is None:
raise CurrentPedalboardError('The current pedalboard is None')
before_index = self.bank.index - 1
if before_index == -1:
before_index = len(self._manager.banks) - 1
self.set_bank(self._manager.banks[before_index])
| to_before_bank | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.pedalboard is None:\n raise CurrentPedalboardError('The current pedalboard is None')\n before_index = self.bank.index - 1\n if before_index == -1:\n before_index = len(self._manager.banks) - 1\n self.set_bank(self._manager.banks[before_index])\n",
"role": "user"
},
{
"content": "to_before_bank",
"role": "assistant"
}
] |
def x(self):
if self.pedalboard is None:
raise CurrentPedalboardError('The current pedalboard is None')
next_index = self.next_bank_index(self.bank.index)
self.set_bank(self._manager.banks[next_index])
| to_next_bank | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.pedalboard is None:\n raise CurrentPedalboardError('The current pedalboard is None')\n next_index = self.next_bank_index(self.bank.index)\n self.set_bank(self._manager.banks[next_index])\n",
"role": "user"
},
{
"content": "to_next_bank",
"role": "assistant"
}
] |
def x(self, bank, try_preserve_index=False):
if bank not in self._manager:
raise CurrentPedalboardError('Bank {} has not added in banks manager'
.format(bank))
if self.bank == bank:
return
if bank.pedalboards:
pedalboard = self._equivalent_pedalboard(bank
) if try_preserve_index else bank.pedalboards[0]
self.set_pedalboard(pedalboard)
else:
self.set_pedalboard(None)
| set_bank | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, bank, try_preserve_index=False):\n if bank not in self._manager:\n raise CurrentPedalboardError('Bank {} has not added in banks manager'\n .format(bank))\n if self.bank == bank:\n return\n if bank.pedalboards:\n pedalboard = self._equivalent_pedalboard(bank\n ) if try_preserve_index else bank.pedalboards[0]\n self.set_pedalboard(pedalboard)\n else:\n self.set_pedalboard(None)\n",
"role": "user"
},
{
"content": "set_bank",
"role": "assistant"
}
] |
def x(options, options_name, key, _type):
if isinstance(options, dict):
has = lambda k: k in options
get = lambda k: options[k]
elif isinstance(options, object):
has = lambda k: hasattr(options, k)
get = lambda k: getattr(options, k)
else:
raise ImproperlyConfigured('`{}` must be a dictionary-like object.'
.format(options_name))
if not has(key):
raise ImproperlyConfigured('`{}` must be specified in `{}`'.format(
key, options_name))
value = get(key)
if not isinstance(value, _type):
raise ImproperlyConfigured('`{}` in `{}` must be a {}'.format(key,
options_name, repr(_type)))
return value
| validate_and_get_value | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(options, options_name, key, _type):\n if isinstance(options, dict):\n has = lambda k: k in options\n get = lambda k: options[k]\n elif isinstance(options, object):\n has = lambda k: hasattr(options, k)\n get = lambda k: getattr(options, k)\n else:\n raise ImproperlyConfigured('`{}` must be a dictionary-like object.'\n .format(options_name))\n if not has(key):\n raise ImproperlyConfigured('`{}` must be specified in `{}`'.format(\n key, options_name))\n value = get(key)\n if not isinstance(value, _type):\n raise ImproperlyConfigured('`{}` in `{}` must be a {}'.format(key,\n options_name, repr(_type)))\n return value\n",
"role": "user"
},
{
"content": "validate_and_get_value",
"role": "assistant"
}
] |
def x(event):
event_time = event.get(TIMESTAMP_FIELD)
if event_time is None:
event[TIMESTAMP_FIELD] = event_time = epoch_time_to_kronos_time(time
.time())
elif type(event_time) not in (int, long):
raise InvalidEventTime(event_time)
_id = uuid_from_kronos_time(event_time)
event[ID_FIELD] = str(_id)
return _id, event
| validate_event_and_assign_id | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(event):\n event_time = event.get(TIMESTAMP_FIELD)\n if event_time is None:\n event[TIMESTAMP_FIELD] = event_time = epoch_time_to_kronos_time(time\n .time())\n elif type(event_time) not in (int, long):\n raise InvalidEventTime(event_time)\n _id = uuid_from_kronos_time(event_time)\n event[ID_FIELD] = str(_id)\n return _id, event\n",
"role": "user"
},
{
"content": "validate_event_and_assign_id",
"role": "assistant"
}
] |
def x(stream):
if not STREAM_REGEX.match(stream) or len(stream) > MAX_STREAM_LENGTH:
raise InvalidStreamName(stream)
| validate_stream | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(stream):\n if not STREAM_REGEX.match(stream) or len(stream) > MAX_STREAM_LENGTH:\n raise InvalidStreamName(stream)\n",
"role": "user"
},
{
"content": "validate_stream",
"role": "assistant"
}
] |
def x(storage_class, settings):
if not isinstance(settings, dict):
raise ImproperlyConfigured('{}: storage class settings must be a dict'
.format(storage_class))
if not hasattr(storage_class, 'SETTINGS_VALIDATORS'):
raise NotImplementedError(
'{}: storage class must define `SETTINGS_VALIDATORS`'.format(
storage_class))
settings_validators = getattr(storage_class, 'SETTINGS_VALIDATORS')
settings = settings.copy()
settings.pop('backend', None)
invalid_settings = set(settings.keys()) - set(settings_validators.keys())
if invalid_settings:
raise ImproperlyConfigured('{}: invalid settings: {}'.format(
storage_class, invalid_settings))
for setting, value in settings.iteritems():
if not settings_validators[setting](value):
raise ImproperlyConfigured('{}: invalid value for {}'.format(
storage_class, setting))
| validate_storage_settings | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(storage_class, settings):\n if not isinstance(settings, dict):\n raise ImproperlyConfigured('{}: storage class settings must be a dict'\n .format(storage_class))\n if not hasattr(storage_class, 'SETTINGS_VALIDATORS'):\n raise NotImplementedError(\n '{}: storage class must define `SETTINGS_VALIDATORS`'.format(\n storage_class))\n settings_validators = getattr(storage_class, 'SETTINGS_VALIDATORS')\n settings = settings.copy()\n settings.pop('backend', None)\n invalid_settings = set(settings.keys()) - set(settings_validators.keys())\n if invalid_settings:\n raise ImproperlyConfigured('{}: invalid settings: {}'.format(\n storage_class, invalid_settings))\n for setting, value in settings.iteritems():\n if not settings_validators[setting](value):\n raise ImproperlyConfigured('{}: invalid value for {}'.format(\n storage_class, setting))\n",
"role": "user"
},
{
"content": "validate_storage_settings",
"role": "assistant"
}
] |
def x(settings):
storage = _validate_and_get_value(settings, 'settings', 'storage', dict)
for name, options in storage.iteritems():
if 'backend' not in options:
raise ImproperlyConfigured(
"`storage['{}'] must contain a `backend` key".format(name))
path = options['backend']
module, cls = path.rsplit('.', 1)
module = import_module(module)
if not hasattr(module, cls):
raise NotImplementedError('`{}` not implemented.'.format(cls))
validate_storage_settings(getattr(module, cls), options)
namespace_to_streams_configuration = _validate_and_get_value(settings,
'settings', 'namespace_to_streams_configuration', dict)
for namespace, prefix_confs in namespace_to_streams_configuration.iteritems(
):
if '' not in prefix_confs:
raise ImproperlyConfigured(
'Must specify backends for the null prefix')
for prefix, options in prefix_confs.iteritems():
if prefix != '':
validate_stream(prefix)
backends = _validate_and_get_value(options,
"namespace_to_streams_configuration['{}']['{}']".format(
namespace, prefix), 'backends', dict)
for backend in backends.keys():
if backend not in storage:
raise ImproperlyConfigured(
"`{}` backend for `namespace_to_streams_configuration['{}']['{}']` is not configured in `storage`"
.format(backend, namespace, prefix))
read_backend = _validate_and_get_value(options,
"namespace_to_streams_configuration['{}']['{}']".format(
namespace, prefix), 'read_backend', str)
if read_backend not in storage:
raise ImproperlyConfigured(
"`{}` backend for `namespace_to_streams_configuration['{}']['{}']` is not configured in `storage`"
.format(read_backend, namespace, prefix))
stream = getattr(settings, 'stream', dict)
_validate_and_get_value(stream, 'stream', 'format', re._pattern_type)
node = getattr(settings, 'node', dict)
_validate_and_get_value(node, 'node', 'greenlet_pool_size', int)
_validate_and_get_value(node, 'node', 'id', str)
| validate_settings | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(settings):\n storage = _validate_and_get_value(settings, 'settings', 'storage', dict)\n for name, options in storage.iteritems():\n if 'backend' not in options:\n raise ImproperlyConfigured(\n \"`storage['{}'] must contain a `backend` key\".format(name))\n path = options['backend']\n module, cls = path.rsplit('.', 1)\n module = import_module(module)\n if not hasattr(module, cls):\n raise NotImplementedError('`{}` not implemented.'.format(cls))\n validate_storage_settings(getattr(module, cls), options)\n namespace_to_streams_configuration = _validate_and_get_value(settings,\n 'settings', 'namespace_to_streams_configuration', dict)\n for namespace, prefix_confs in namespace_to_streams_configuration.iteritems(\n ):\n if '' not in prefix_confs:\n raise ImproperlyConfigured(\n 'Must specify backends for the null prefix')\n for prefix, options in prefix_confs.iteritems():\n if prefix != '':\n validate_stream(prefix)\n backends = _validate_and_get_value(options,\n \"namespace_to_streams_configuration['{}']['{}']\".format(\n namespace, prefix), 'backends', dict)\n for backend in backends.keys():\n if backend not in storage:\n raise ImproperlyConfigured(\n \"`{}` backend for `namespace_to_streams_configuration['{}']['{}']` is not configured in `storage`\"\n .format(backend, namespace, prefix))\n read_backend = _validate_and_get_value(options,\n \"namespace_to_streams_configuration['{}']['{}']\".format(\n namespace, prefix), 'read_backend', str)\n if read_backend not in storage:\n raise ImproperlyConfigured(\n \"`{}` backend for `namespace_to_streams_configuration['{}']['{}']` is not configured in `storage`\"\n .format(read_backend, namespace, prefix))\n stream = getattr(settings, 'stream', dict)\n _validate_and_get_value(stream, 'stream', 'format', re._pattern_type)\n node = getattr(settings, 'node', dict)\n _validate_and_get_value(node, 'node', 'greenlet_pool_size', int)\n _validate_and_get_value(node, 'node', 'id', str)\n",
"role": "user"
},
{
"content": "validate_settings",
"role": "assistant"
}
] |
def x(cls, **kwargs):
vals = cls.get_non_empty_vals({cls._to_snake_case(k): v for k, v in
kwargs.items()})
remove = []
for attr, val in vals.items():
try:
vals[attr] = cls._parse_property(attr, val)
except HelpScoutValidationException:
remove.append(attr)
logger.info('Unexpected property received in API response',
exc_info=True)
for attr in remove:
del vals[attr]
return cls(**cls.get_non_empty_vals(vals))
| from_api | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, **kwargs):\n vals = cls.get_non_empty_vals({cls._to_snake_case(k): v for k, v in\n kwargs.items()})\n remove = []\n for attr, val in vals.items():\n try:\n vals[attr] = cls._parse_property(attr, val)\n except HelpScoutValidationException:\n remove.append(attr)\n logger.info('Unexpected property received in API response',\n exc_info=True)\n for attr in remove:\n del vals[attr]\n return cls(**cls.get_non_empty_vals(vals))\n",
"role": "user"
},
{
"content": "from_api",
"role": "assistant"
}
] |
def x(self):
vals = {}
for attribute, attribute_type in self._props.items():
prop = getattr(self, attribute)
vals[self._to_camel_case(attribute)] = self._to_api_value(
attribute_type, prop)
return vals
| to_api | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n vals = {}\n for attribute, attribute_type in self._props.items():\n prop = getattr(self, attribute)\n vals[self._to_camel_case(attribute)] = self._to_api_value(\n attribute_type, prop)\n return vals\n",
"role": "user"
},
{
"content": "to_api",
"role": "assistant"
}
] |
def x(self, attribute_type, value):
if not value:
return None
if isinstance(attribute_type, properties.Instance):
return value.to_api()
if isinstance(attribute_type, properties.List):
return self._parse_api_value_list(value)
return attribute_type.serialize(value)
| to_api_value | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, attribute_type, value):\n if not value:\n return None\n if isinstance(attribute_type, properties.Instance):\n return value.to_api()\n if isinstance(attribute_type, properties.List):\n return self._parse_api_value_list(value)\n return attribute_type.serialize(value)\n",
"role": "user"
},
{
"content": "to_api_value",
"role": "assistant"
}
] |
def x(self, values):
try:
return [v.to_api() for v in values]
except AttributeError:
return list(values)
| parse_api_value_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, values):\n try:\n return [v.to_api() for v in values]\n except AttributeError:\n return list(values)\n",
"role": "user"
},
{
"content": "parse_api_value_list",
"role": "assistant"
}
] |
def x(cls, name, value):
prop = cls._props.get(name)
return_value = value
if not prop:
logger.debug(
'"%s" with value "%s" is not a valid property for "%s".' % (
name, value, cls))
return_value = None
elif isinstance(prop, properties.Instance):
return_value = prop.instance_class.from_api(**value)
elif isinstance(prop, properties.List):
return_value = cls._parse_property_list(prop, value)
elif isinstance(prop, properties.Color):
return_value = cls._parse_property_color(value)
return return_value
| parse_property | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, name, value):\n prop = cls._props.get(name)\n return_value = value\n if not prop:\n logger.debug(\n '\"%s\" with value \"%s\" is not a valid property for \"%s\".' % (\n name, value, cls))\n return_value = None\n elif isinstance(prop, properties.Instance):\n return_value = prop.instance_class.from_api(**value)\n elif isinstance(prop, properties.List):\n return_value = cls._parse_property_list(prop, value)\n elif isinstance(prop, properties.Color):\n return_value = cls._parse_property_color(value)\n return return_value\n",
"role": "user"
},
{
"content": "parse_property",
"role": "assistant"
}
] |
def x(prop, value):
attributes = []
for v in value:
try:
attributes.append(prop.prop.instance_class.from_api(**v))
except AttributeError:
attributes.append(v)
return attributes
| parse_property_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(prop, value):\n attributes = []\n for v in value:\n try:\n attributes.append(prop.prop.instance_class.from_api(**v))\n except AttributeError:\n attributes.append(v)\n return attributes\n",
"role": "user"
},
{
"content": "parse_property_list",
"role": "assistant"
}
] |
def x(string):
sub_string = '\\1_\\2'
string = REGEX_CAMEL_FIRST.sub(sub_string, string)
return REGEX_CAMEL_SECOND.sub(sub_string, string).lower()
| to_snake_case | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(string):\n sub_string = '\\\\1_\\\\2'\n string = REGEX_CAMEL_FIRST.sub(sub_string, string)\n return REGEX_CAMEL_SECOND.sub(sub_string, string).lower()\n",
"role": "user"
},
{
"content": "to_snake_case",
"role": "assistant"
}
] |
def x(string):
components = string.split('_')
return '%s%s' % (components[0], ''.join(c.title() for c in components[1:]))
| to_camel_case | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(string):\n components = string.split('_')\n return '%s%s' % (components[0], ''.join(c.title() for c in components[1:]))\n",
"role": "user"
},
{
"content": "to_camel_case",
"role": "assistant"
}
] |
def x(self, output=None):
if output is None:
if self.dbfile is None:
return
output = self.dbfile
with open(output, 'w') as f:
f.write(self._dump())
| save | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, output=None):\n if output is None:\n if self.dbfile is None:\n return\n output = self.dbfile\n with open(output, 'w') as f:\n f.write(self._dump())\n",
"role": "user"
},
{
"content": "save",
"role": "assistant"
}
] |
def x(self, text=None, source=None):
if text is not None:
words = re.split('\\s+', text)
wlen = len(words)
for i in range(wlen - 2):
self._insert(words[i:i + 3])
if source is not None:
with open(source, 'r') as f:
self.feed(f.read())
| feed | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, text=None, source=None):\n if text is not None:\n words = re.split('\\\\s+', text)\n wlen = len(words)\n for i in range(wlen - 2):\n self._insert(words[i:i + 3])\n if source is not None:\n with open(source, 'r') as f:\n self.feed(f.read())\n",
"role": "user"
},
{
"content": "feed",
"role": "assistant"
}
] |
def x(self, **kwargs):
words = list(map(self._sanitize, kwargs.get('words', [])))
max_wlen = kwargs.get('wlen', 70)
wlen = len(words)
if wlen < 2:
if not self._db:
return ''
if wlen == 0:
words = sample(self._db.keys(), 1)[0].split(self._WSEP)
elif wlen == 1:
spl = [k for k in self._db.keys() if k.startswith(words[0] +
self._WSEP)]
words.append(sample(spl, 1)[0].split(self._WSEP)[1])
wlen = 2
while wlen < max_wlen:
next_word = self._get(words[-2], words[-1])
if next_word is None:
break
words.append(next_word)
wlen += 1
return ' '.join(words)
| generate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n words = list(map(self._sanitize, kwargs.get('words', [])))\n max_wlen = kwargs.get('wlen', 70)\n wlen = len(words)\n if wlen < 2:\n if not self._db:\n return ''\n if wlen == 0:\n words = sample(self._db.keys(), 1)[0].split(self._WSEP)\n elif wlen == 1:\n spl = [k for k in self._db.keys() if k.startswith(words[0] +\n self._WSEP)]\n words.append(sample(spl, 1)[0].split(self._WSEP)[1])\n wlen = 2\n while wlen < max_wlen:\n next_word = self._get(words[-2], words[-1])\n if next_word is None:\n break\n words.append(next_word)\n wlen += 1\n return ' '.join(words)\n",
"role": "user"
},
{
"content": "generate",
"role": "assistant"
}
] |
def x(self):
if self.dbfile is not None:
with open(self.dbfile, 'r') as f:
self._db = json.loads(f.read())
else:
self._db = {}
| load | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.dbfile is not None:\n with open(self.dbfile, 'r') as f:\n self._db = json.loads(f.read())\n else:\n self._db = {}\n",
"role": "user"
},
{
"content": "load",
"role": "assistant"
}
] |
def x(self, word1, word2):
key = self._WSEP.join([self._sanitize(word1), self._sanitize(word2)])
key = key.lower()
if key not in self._db:
return
return sample(self._db[key], 1)[0]
| get | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, word1, word2):\n key = self._WSEP.join([self._sanitize(word1), self._sanitize(word2)])\n key = key.lower()\n if key not in self._db:\n return\n return sample(self._db[key], 1)[0]\n",
"role": "user"
},
{
"content": "get",
"role": "assistant"
}
] |
def x(self, trigram):
words = list(map(self._sanitize, trigram))
key = self._WSEP.join(words[:2]).lower()
next_word = words[2]
self._db.setdefault(key, [])
if next_word not in self._db[key]:
self._db[key].append(next_word)
| insert | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, trigram):\n words = list(map(self._sanitize, trigram))\n key = self._WSEP.join(words[:2]).lower()\n next_word = words[2]\n self._db.setdefault(key, [])\n if next_word not in self._db[key]:\n self._db[key].append(next_word)\n",
"role": "user"
},
{
"content": "insert",
"role": "assistant"
}
] |
def x(plan):
cohort = plan['cohort']
action = plan['action']
source = plan['source']
cohort_start = datetime_to_kronos_time(_date_to_datetime(cohort['start']))
cohort_span = timedelta(**{cohort['unit']: cohort['cohorts']})
cohort_end = cohort['start'] + cohort_span
action_span = timedelta(**{action['unit']: action['repetitions']})
action_end = cohort_end + action_span
cohort_end = datetime_to_kronos_time(_date_to_datetime(cohort_end)) + 1
action_end = datetime_to_kronos_time(_date_to_datetime(action_end)) + 1
left = _cohort_stream_transform(source, cohort['stream'], cohort_start,
cohort_end, cohort.get('transform'), cohort['grouping_key'], cohort
['unit'])
right = _cohort_stream_transform(source, action['stream'], cohort_start,
action_end, action.get('transform'), action['grouping_key'], action
['unit'])
additional_action_time = DateUnit.unit_to_kronos_time(action['unit']
) * action['repetitions']
left.alias = 'cohort'
right.alias = 'action'
joined = Join(left, right, Condition(Condition.Op.EQ, Property(
'cohort.%s' % cohort['grouping_key']), Property('action.%s' %
action['grouping_key'])) & Condition(Condition.Op.GTE, Property(
'action.%s' % TIMESTAMP_FIELD), Property('cohort.%s' %
TIMESTAMP_FIELD)) & Condition(Condition.Op.LT, Property('action.%s' %
TIMESTAMP_FIELD), Add([Property('cohort.%s' % TIMESTAMP_FIELD),
Constant(additional_action_time)])))
user_aggregated = Aggregate(joined, GroupBy([Property('cohort.date',
alias=TIMESTAMP_FIELD), Property('cohort.%s' % cohort[
'grouping_key'], alias='group'), Floor([Subtract([Property(
'action.%s' % TIMESTAMP_FIELD), Property('cohort.%s' %
TIMESTAMP_FIELD)]), Constant(DateUnit.unit_to_kronos_time(action[
'unit']))], alias='action_step')]), [Count([], alias='count')])
aggregated = Aggregate(user_aggregated, GroupBy([Property(
TIMESTAMP_FIELD, alias=TIMESTAMP_FIELD), Property('action_step',
alias='action_step')]), [Count([], alias='cohort_actions')])
return aggregated.to_dict()
| cohort_queryplan | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(plan):\n cohort = plan['cohort']\n action = plan['action']\n source = plan['source']\n cohort_start = datetime_to_kronos_time(_date_to_datetime(cohort['start']))\n cohort_span = timedelta(**{cohort['unit']: cohort['cohorts']})\n cohort_end = cohort['start'] + cohort_span\n action_span = timedelta(**{action['unit']: action['repetitions']})\n action_end = cohort_end + action_span\n cohort_end = datetime_to_kronos_time(_date_to_datetime(cohort_end)) + 1\n action_end = datetime_to_kronos_time(_date_to_datetime(action_end)) + 1\n left = _cohort_stream_transform(source, cohort['stream'], cohort_start,\n cohort_end, cohort.get('transform'), cohort['grouping_key'], cohort\n ['unit'])\n right = _cohort_stream_transform(source, action['stream'], cohort_start,\n action_end, action.get('transform'), action['grouping_key'], action\n ['unit'])\n additional_action_time = DateUnit.unit_to_kronos_time(action['unit']\n ) * action['repetitions']\n left.alias = 'cohort'\n right.alias = 'action'\n joined = Join(left, right, Condition(Condition.Op.EQ, Property(\n 'cohort.%s' % cohort['grouping_key']), Property('action.%s' %\n action['grouping_key'])) & Condition(Condition.Op.GTE, Property(\n 'action.%s' % TIMESTAMP_FIELD), Property('cohort.%s' %\n TIMESTAMP_FIELD)) & Condition(Condition.Op.LT, Property('action.%s' %\n TIMESTAMP_FIELD), Add([Property('cohort.%s' % TIMESTAMP_FIELD),\n Constant(additional_action_time)])))\n user_aggregated = Aggregate(joined, GroupBy([Property('cohort.date',\n alias=TIMESTAMP_FIELD), Property('cohort.%s' % cohort[\n 'grouping_key'], alias='group'), Floor([Subtract([Property(\n 'action.%s' % TIMESTAMP_FIELD), Property('cohort.%s' %\n TIMESTAMP_FIELD)]), Constant(DateUnit.unit_to_kronos_time(action[\n 'unit']))], alias='action_step')]), [Count([], alias='count')])\n aggregated = Aggregate(user_aggregated, GroupBy([Property(\n TIMESTAMP_FIELD, alias=TIMESTAMP_FIELD), Property('action_step',\n alias='action_step')]), [Count([], alias='cohort_actions')])\n return aggregated.to_dict()\n",
"role": "user"
},
{
"content": "cohort_queryplan",
"role": "assistant"
}
] |
def x(mass):
if mass is np.nan:
return None
for massLimit, massType in planetAssumptions['massType']:
if mass < massLimit:
return massType
| planetMassType | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(mass):\n if mass is np.nan:\n return None\n for massLimit, massType in planetAssumptions['massType']:\n if mass < massLimit:\n return massType\n",
"role": "user"
},
{
"content": "planetMassType",
"role": "assistant"
}
] |
def x(radius):
if radius is np.nan:
return None
for radiusLimit, radiusType in planetAssumptions['radiusType']:
if radius < radiusLimit:
return radiusType
| planetRadiusType | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(radius):\n if radius is np.nan:\n return None\n for radiusLimit, radiusType in planetAssumptions['radiusType']:\n if radius < radiusLimit:\n return radiusType\n",
"role": "user"
},
{
"content": "planetRadiusType",
"role": "assistant"
}
] |
def x(temperature, mass, radius):
if mass is not np.nan:
sizeType = planetMassType(mass)
elif radius is not np.nan:
sizeType = planetRadiusType(radius)
else:
return None
return '{0} {1}'.format(planetTempType(temperature), sizeType)
| planetType | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(temperature, mass, radius):\n if mass is not np.nan:\n sizeType = planetMassType(mass)\n elif radius is not np.nan:\n sizeType = planetRadiusType(radius)\n else:\n return None\n return '{0} {1}'.format(planetTempType(temperature), sizeType)\n",
"role": "user"
},
{
"content": "planetType",
"role": "assistant"
}
] |
def x(df, columns=None):
dtypes = df.dtypes
if columns is None:
columns = df.columns
elif isinstance(columns, str):
columns = [columns]
for column in columns:
expanded = np.repeat(df.values, df[column].apply(len).values, axis=0)
expanded[:, df.columns.get_loc(column)] = np.concatenate(df[column]
.tolist())
df = pd.DataFrame(expanded, columns=df.columns)
for i, dtype in enumerate(dtypes):
df.iloc[:, i] = df.iloc[:, i].astype(dtype)
return df
| split_array_like | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(df, columns=None):\n dtypes = df.dtypes\n if columns is None:\n columns = df.columns\n elif isinstance(columns, str):\n columns = [columns]\n for column in columns:\n expanded = np.repeat(df.values, df[column].apply(len).values, axis=0)\n expanded[:, df.columns.get_loc(column)] = np.concatenate(df[column]\n .tolist())\n df = pd.DataFrame(expanded, columns=df.columns)\n for i, dtype in enumerate(dtypes):\n df.iloc[:, i] = df.iloc[:, i].astype(dtype)\n return df\n",
"role": "user"
},
{
"content": "split_array_like",
"role": "assistant"
}
] |
def x(df1, df2, ignore_order=set(), ignore_indices=set(), all_close=False,
_return_reason=False):
result = _equals(df1, df2, ignore_order, ignore_indices, all_close)
if _return_reason:
return result
else:
return result[0]
| equals | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(df1, df2, ignore_order=set(), ignore_indices=set(), all_close=False,\n _return_reason=False):\n result = _equals(df1, df2, ignore_order, ignore_indices, all_close)\n if _return_reason:\n return result\n else:\n return result[0]\n",
"role": "user"
},
{
"content": "equals",
"role": "assistant"
}
] |
def x(row, values, all_close, ignore_order):
for row2 in values:
mask = ma.getmaskarray(row2)
assert mask.sum() in (0, len(mask))
if mask[0]:
continue
if _try_mask_row(row, row2, all_close, ignore_order):
return True
return False
| try_mask_first_row | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(row, values, all_close, ignore_order):\n for row2 in values:\n mask = ma.getmaskarray(row2)\n assert mask.sum() in (0, len(mask))\n if mask[0]:\n continue\n if _try_mask_row(row, row2, all_close, ignore_order):\n return True\n return False\n",
"role": "user"
},
{
"content": "try_mask_first_row",
"role": "assistant"
}
] |
def x(row1, row2, all_close, ignore_order):
if ignore_order:
for value1 in row1:
if not _try_mask_first_value(value1, row2, all_close):
row2.mask = ma.nomask
return False
else:
for value1, value2 in zip(row1, row2):
if not _value_equals(value1, value2, all_close):
return False
row2[:] = ma.masked
assert row2.mask.all()
return True
| try_mask_row | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(row1, row2, all_close, ignore_order):\n if ignore_order:\n for value1 in row1:\n if not _try_mask_first_value(value1, row2, all_close):\n row2.mask = ma.nomask\n return False\n else:\n for value1, value2 in zip(row1, row2):\n if not _value_equals(value1, value2, all_close):\n return False\n row2[:] = ma.masked\n assert row2.mask.all()\n return True\n",
"role": "user"
},
{
"content": "try_mask_row",
"role": "assistant"
}
] |
def x(value, row, all_close):
for i, value2 in enumerate(row):
if _value_equals(value, value2, all_close):
row[i] = ma.masked
return True
return False
| try_mask_first_value | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(value, row, all_close):\n for i, value2 in enumerate(row):\n if _value_equals(value, value2, all_close):\n row[i] = ma.masked\n return True\n return False\n",
"role": "user"
},
{
"content": "try_mask_first_value",
"role": "assistant"
}
] |
def x(value1, value2, all_close):
if value1 is None:
value1 = np.nan
if value2 is None:
value2 = np.nan
are_floats = np.can_cast(type(value1), float) and np.can_cast(type(
value2), float)
if all_close and are_floats:
return np.isclose(value1, value2, equal_nan=True)
elif are_floats:
return value1 == value2 or value1 != value1 and value2 != value2
else:
return value1 == value2
| value_equals | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(value1, value2, all_close):\n if value1 is None:\n value1 = np.nan\n if value2 is None:\n value2 = np.nan\n are_floats = np.can_cast(type(value1), float) and np.can_cast(type(\n value2), float)\n if all_close and are_floats:\n return np.isclose(value1, value2, equal_nan=True)\n elif are_floats:\n return value1 == value2 or value1 != value1 and value2 != value2\n else:\n return value1 == value2\n",
"role": "user"
},
{
"content": "value_equals",
"role": "assistant"
}
] |
def x(df1, df2, ignore_order=set(), ignore_indices=set(), all_close=False,
_return_reason=False):
equals_, reason = equals(df1, df2, ignore_order, ignore_indices,
all_close, _return_reason=True)
assert equals_, '{}\n\n{}\n\n{}'.format(reason, df1.to_string(), df2.
to_string())
| assert_equals | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(df1, df2, ignore_order=set(), ignore_indices=set(), all_close=False,\n _return_reason=False):\n equals_, reason = equals(df1, df2, ignore_order, ignore_indices,\n all_close, _return_reason=True)\n assert equals_, '{}\\n\\n{}\\n\\n{}'.format(reason, df1.to_string(), df2.\n to_string())\n",
"role": "user"
},
{
"content": "assert_equals",
"role": "assistant"
}
] |
def x(first, second):
if first and second:
return first.upper() == second.upper()
else:
return not (first or second)
| is_equal_strings_ignore_case | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(first, second):\n if first and second:\n return first.upper() == second.upper()\n else:\n return not (first or second)\n",
"role": "user"
},
{
"content": "is_equal_strings_ignore_case",
"role": "assistant"
}
] |
def x(self, dim_name):
for dim in self.dimensions:
if is_equal_strings_ignore_case(dim.name, dim_name):
return dim
return None
| find_dimension_by_name | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dim_name):\n for dim in self.dimensions:\n if is_equal_strings_ignore_case(dim.name, dim_name):\n return dim\n return None\n",
"role": "user"
},
{
"content": "find_dimension_by_name",
"role": "assistant"
}
] |
def x(self, dim_id):
for dim in self.dimensions:
if is_equal_strings_ignore_case(dim.id, dim_id):
return dim
return None
| find_dimension_by_id | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dim_id):\n for dim in self.dimensions:\n if is_equal_strings_ignore_case(dim.id, dim_id):\n return dim\n return None\n",
"role": "user"
},
{
"content": "find_dimension_by_id",
"role": "assistant"
}
] |
def x(self):
requestvalues = {'Dataset': self.dataset, 'Header': self.
_get_item_array(self.header), 'Filter': self._get_item_array(self.
filter), 'Stub': self._get_item_array(self.stub), 'Frequencies':
self.frequencies}
return json.dumps(requestvalues)
| save_to_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n requestvalues = {'Dataset': self.dataset, 'Header': self.\n _get_item_array(self.header), 'Filter': self._get_item_array(self.\n filter), 'Stub': self._get_item_array(self.stub), 'Frequencies':\n self.frequencies}\n return json.dumps(requestvalues)\n",
"role": "user"
},
{
"content": "save_to_json",
"role": "assistant"
}
] |
def x(self):
requestvalues = {'DatasetId': self.dataset, 'Name': self.name,
'Description': self.description, 'Source': self.source, 'PubDate':
self.publication_date, 'AccessedOn': self.accessed_on, 'Url': self.
dataset_ref, 'UploadFormatType': self.upload_format_type, 'Columns':
self.columns, 'FileProperty': self.file_property.__dict__,
'FlatDSUpdateOptions': self.flat_ds_update_options, 'Public': self.
public}
return json.dumps(requestvalues)
| save_to_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n requestvalues = {'DatasetId': self.dataset, 'Name': self.name,\n 'Description': self.description, 'Source': self.source, 'PubDate':\n self.publication_date, 'AccessedOn': self.accessed_on, 'Url': self.\n dataset_ref, 'UploadFormatType': self.upload_format_type, 'Columns':\n self.columns, 'FileProperty': self.file_property.__dict__,\n 'FlatDSUpdateOptions': self.flat_ds_update_options, 'Public': self.\n public}\n return json.dumps(requestvalues)\n",
"role": "user"
},
{
"content": "save_to_json",
"role": "assistant"
}
] |
def x(self):
requestvalues = {'id': self.dataset, 'publicationDate': self.
publication_date.strftime('%Y-%m-%d'), 'source': self.source,
'refUrl': self.refernce_url}
return json.dumps(requestvalues)
| save_to_json | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n requestvalues = {'id': self.dataset, 'publicationDate': self.\n publication_date.strftime('%Y-%m-%d'), 'source': self.source,\n 'refUrl': self.refernce_url}\n return json.dumps(requestvalues)\n",
"role": "user"
},
{
"content": "save_to_json",
"role": "assistant"
}
] |
def x(self, keyspace):
if FORMAT_SPEC.search(keyspace):
return KeyspacedProxy(self, keyspace)
else:
return KeyspacedProxy(self, self._keyspaces[keyspace])
| keyspace | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, keyspace):\n if FORMAT_SPEC.search(keyspace):\n return KeyspacedProxy(self, keyspace)\n else:\n return KeyspacedProxy(self, self._keyspaces[keyspace])\n",
"role": "user"
},
{
"content": "keyspace",
"role": "assistant"
}
] |
def x(dir1, dir2):
compared = dircmp(dir1, dir2, IGNORED_FILES)
if (compared.left_only or compared.right_only or compared.diff_files or
compared.funny_files):
return False
for subdir in compared.common_dirs:
if not is_same(os.path.join(dir1, subdir), os.path.join(dir2, subdir)):
return False
return True
| is_same | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dir1, dir2):\n compared = dircmp(dir1, dir2, IGNORED_FILES)\n if (compared.left_only or compared.right_only or compared.diff_files or\n compared.funny_files):\n return False\n for subdir in compared.common_dirs:\n if not is_same(os.path.join(dir1, subdir), os.path.join(dir2, subdir)):\n return False\n return True\n",
"role": "user"
},
{
"content": "is_same",
"role": "assistant"
}
] |
def x(self):
fcomp = filecmp.cmpfiles(self.left, self.right, self.common_files,
shallow=False)
self.same_files, self.diff_files, self.funny_files = fcomp
| phase3 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n fcomp = filecmp.cmpfiles(self.left, self.right, self.common_files,\n shallow=False)\n self.same_files, self.diff_files, self.funny_files = fcomp\n",
"role": "user"
},
{
"content": "phase3",
"role": "assistant"
}
] |
def x(self, namespace, stream, events, configuration):
index = self.index_manager.get_index(namespace)
start_dts_to_add = set()
def x():
for _id, event in events:
dt = kronos_time_to_datetime(uuid_to_kronos_time(_id))
start_dts_to_add.add(_round_datetime_down(dt))
event['_index'] = index
event['_type'] = stream
event[LOGSTASH_TIMESTAMP_FIELD] = dt.isoformat()
yield event
list(es_helpers.streaming_bulk(self.es, actions(), chunk_size=1000,
refresh=self.force_refresh))
self.index_manager.add_aliases(namespace, index, start_dts_to_add)
| insert | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, namespace, stream, events, configuration):\n index = self.index_manager.get_index(namespace)\n start_dts_to_add = set()\n\n def x():\n for _id, event in events:\n dt = kronos_time_to_datetime(uuid_to_kronos_time(_id))\n start_dts_to_add.add(_round_datetime_down(dt))\n event['_index'] = index\n event['_type'] = stream\n event[LOGSTASH_TIMESTAMP_FIELD] = dt.isoformat()\n yield event\n list(es_helpers.streaming_bulk(self.es, actions(), chunk_size=1000,\n refresh=self.force_refresh))\n self.index_manager.add_aliases(namespace, index, start_dts_to_add)\n",
"role": "user"
},
{
"content": "insert",
"role": "assistant"
}
] |
def x(self, namespace, stream, start_id, end_time, order, limit, configuration
):
indices = self.index_manager.get_aliases(namespace, uuid_to_kronos_time
(start_id), end_time)
if not indices:
return
end_id = uuid_from_kronos_time(end_time, _type=UUIDType.HIGHEST)
end_id.descending = start_id.descending = descending = (order ==
ResultOrder.DESCENDING)
start_time = uuid_to_kronos_time(start_id)
body_query = {'query': {'filtered': {'query': {'match_all': {}},
'filter': {'range': {TIMESTAMP_FIELD: {'gte': start_time, 'lte':
end_time}}}}}}
order = 'desc' if descending else 'asc'
sort_query = ['%s:%s' % (TIMESTAMP_FIELD, order), '%s:%s' % (ID_FIELD,
order)]
last_id = end_id if descending else start_id
scroll_id = None
while True:
size = max(min(limit, configuration['read_size']) / self.shards, 10)
if scroll_id is None:
res = self.es.search(index=indices, doc_type=stream, size=size,
body=body_query, sort=sort_query, _source=True, scroll='1m',
ignore=[400, 404], allow_no_indices=True,
ignore_unavailable=True)
else:
res = self.es.scroll(scroll_id, scroll='1m')
if '_scroll_id' not in res:
break
scroll_id = res['_scroll_id']
hits = res.get('hits', {}).get('hits')
if not hits:
break
for hit in hits:
_id = TimeUUID(hit['_id'], descending=descending)
if _id <= last_id:
continue
last_id = _id
event = hit['_source']
del event[LOGSTASH_TIMESTAMP_FIELD]
yield json.dumps(event)
limit -= 1
if limit == 0:
break
if scroll_id is not None:
self.es.clear_scroll(scroll_id)
| retrieve | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, namespace, stream, start_id, end_time, order, limit, configuration\n ):\n indices = self.index_manager.get_aliases(namespace, uuid_to_kronos_time\n (start_id), end_time)\n if not indices:\n return\n end_id = uuid_from_kronos_time(end_time, _type=UUIDType.HIGHEST)\n end_id.descending = start_id.descending = descending = (order ==\n ResultOrder.DESCENDING)\n start_time = uuid_to_kronos_time(start_id)\n body_query = {'query': {'filtered': {'query': {'match_all': {}},\n 'filter': {'range': {TIMESTAMP_FIELD: {'gte': start_time, 'lte':\n end_time}}}}}}\n order = 'desc' if descending else 'asc'\n sort_query = ['%s:%s' % (TIMESTAMP_FIELD, order), '%s:%s' % (ID_FIELD,\n order)]\n last_id = end_id if descending else start_id\n scroll_id = None\n while True:\n size = max(min(limit, configuration['read_size']) / self.shards, 10)\n if scroll_id is None:\n res = self.es.search(index=indices, doc_type=stream, size=size,\n body=body_query, sort=sort_query, _source=True, scroll='1m',\n ignore=[400, 404], allow_no_indices=True,\n ignore_unavailable=True)\n else:\n res = self.es.scroll(scroll_id, scroll='1m')\n if '_scroll_id' not in res:\n break\n scroll_id = res['_scroll_id']\n hits = res.get('hits', {}).get('hits')\n if not hits:\n break\n for hit in hits:\n _id = TimeUUID(hit['_id'], descending=descending)\n if _id <= last_id:\n continue\n last_id = _id\n event = hit['_source']\n del event[LOGSTASH_TIMESTAMP_FIELD]\n yield json.dumps(event)\n limit -= 1\n if limit == 0:\n break\n if scroll_id is not None:\n self.es.clear_scroll(scroll_id)\n",
"role": "user"
},
{
"content": "retrieve",
"role": "assistant"
}
] |
def x(self):
cdir = os.path.dirname(os.path.realpath(__file__))
schema = os.path.join(cdir, 'data', 'partsmaster.sql')
if self.debug:
print(self.hdr, 'parts master schema is ', schema)
self.populate(schema)
schema = os.path.join(cdir, 'data', 'approvedmfg.sql')
if self.debug:
print(self.hdr, 'approved mfg list schema is ', schema)
self.populate(schema)
schema = os.path.join(cdir, 'data', 'attachment.sql')
if self.debug:
print(self.hdr, 'attachment schema is ', schema)
self.populate(schema)
schema = os.path.join(cdir, 'data', 'bom.sql')
if self.debug:
print(self.hdr, 'bill of materials schema is ', schema)
self.populate(schema)
return
| create_tables | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n cdir = os.path.dirname(os.path.realpath(__file__))\n schema = os.path.join(cdir, 'data', 'partsmaster.sql')\n if self.debug:\n print(self.hdr, 'parts master schema is ', schema)\n self.populate(schema)\n schema = os.path.join(cdir, 'data', 'approvedmfg.sql')\n if self.debug:\n print(self.hdr, 'approved mfg list schema is ', schema)\n self.populate(schema)\n schema = os.path.join(cdir, 'data', 'attachment.sql')\n if self.debug:\n print(self.hdr, 'attachment schema is ', schema)\n self.populate(schema)\n schema = os.path.join(cdir, 'data', 'bom.sql')\n if self.debug:\n print(self.hdr, 'bill of materials schema is ', schema)\n self.populate(schema)\n return\n",
"role": "user"
},
{
"content": "create_tables",
"role": "assistant"
}
] |
def x(f):
f.seek(0)
try:
p_rec_type = None
while True:
rec_off = f.tell()
while True:
if f.read(2) == b'\xff\xff':
h = struct.unpack('<HH', f.read(4))
if h[1] < 64 and h[1] != 0:
rec_type = f.read(h[1])
if rec_type[0] == 67:
break
if f.read(1) == b'':
raise EOFError
f.seek(f.tell() - 2)
if p_rec_type is not None:
rec_len = f.tell() - 6 - len(rec_type) - rec_off
f.seek(rec_off)
yield p_rec_type, f.read(rec_len)
f.seek(f.tell() + 6 + len(rec_type))
p_rec_type = rec_type
except EOFError:
rec_len = f.tell() - 6 - len(rec_type) - rec_off
f.seek(rec_off)
yield p_rec_type, f.read(rec_len)
| parse_c_serialized | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(f):\n f.seek(0)\n try:\n p_rec_type = None\n while True:\n rec_off = f.tell()\n while True:\n if f.read(2) == b'\\xff\\xff':\n h = struct.unpack('<HH', f.read(4))\n if h[1] < 64 and h[1] != 0:\n rec_type = f.read(h[1])\n if rec_type[0] == 67:\n break\n if f.read(1) == b'':\n raise EOFError\n f.seek(f.tell() - 2)\n if p_rec_type is not None:\n rec_len = f.tell() - 6 - len(rec_type) - rec_off\n f.seek(rec_off)\n yield p_rec_type, f.read(rec_len)\n f.seek(f.tell() + 6 + len(rec_type))\n p_rec_type = rec_type\n except EOFError:\n rec_len = f.tell() - 6 - len(rec_type) - rec_off\n f.seek(rec_off)\n yield p_rec_type, f.read(rec_len)\n",
"role": "user"
},
{
"content": "parse_c_serialized",
"role": "assistant"
}
] |
def x(self, t, dt=None, aggfunc=None):
return self.data.scan(t, dt, aggfunc)
| scan | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, t, dt=None, aggfunc=None):\n return self.data.scan(t, dt, aggfunc)\n",
"role": "user"
},
{
"content": "scan",
"role": "assistant"
}
] |
def x():
storage_broker_lookup = dict()
for entrypoint in iter_entry_points('dtool.storage_brokers'):
StorageBroker = entrypoint.load()
storage_broker_lookup[StorageBroker.key] = StorageBroker
return storage_broker_lookup
| generate_storage_broker_lookup | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n storage_broker_lookup = dict()\n for entrypoint in iter_entry_points('dtool.storage_brokers'):\n StorageBroker = entrypoint.load()\n storage_broker_lookup[StorageBroker.key] = StorageBroker\n return storage_broker_lookup\n",
"role": "user"
},
{
"content": "generate_storage_broker_lookup",
"role": "assistant"
}
] |
def x(uri, config_path):
uri = dtoolcore.utils.sanitise_uri(uri)
storage_broker_lookup = _generate_storage_broker_lookup()
parsed_uri = dtoolcore.utils.generous_parse_uri(uri)
StorageBroker = storage_broker_lookup[parsed_uri.scheme]
return StorageBroker(uri, config_path)
| get_storage_broker | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(uri, config_path):\n uri = dtoolcore.utils.sanitise_uri(uri)\n storage_broker_lookup = _generate_storage_broker_lookup()\n parsed_uri = dtoolcore.utils.generous_parse_uri(uri)\n StorageBroker = storage_broker_lookup[parsed_uri.scheme]\n return StorageBroker(uri, config_path)\n",
"role": "user"
},
{
"content": "get_storage_broker",
"role": "assistant"
}
] |
def x(uri, config_path):
uri = dtoolcore.utils.sanitise_uri(uri)
storage_broker = _get_storage_broker(uri, config_path)
admin_metadata = storage_broker.get_admin_metadata()
return admin_metadata
| admin_metadata_from_uri | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(uri, config_path):\n uri = dtoolcore.utils.sanitise_uri(uri)\n storage_broker = _get_storage_broker(uri, config_path)\n admin_metadata = storage_broker.get_admin_metadata()\n return admin_metadata\n",
"role": "user"
},
{
"content": "admin_metadata_from_uri",
"role": "assistant"
}
] |
def x(uri, config_path):
uri = dtoolcore.utils.sanitise_uri(uri)
storage_broker = _get_storage_broker(uri, config_path)
return storage_broker.has_admin_metadata()
| is_dataset | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(uri, config_path):\n uri = dtoolcore.utils.sanitise_uri(uri)\n storage_broker = _get_storage_broker(uri, config_path)\n return storage_broker.has_admin_metadata()\n",
"role": "user"
},
{
"content": "is_dataset",
"role": "assistant"
}
] |
def x(name, creator_username=None):
if not dtoolcore.utils.name_is_valid(name):
raise DtoolCoreInvalidNameError()
if creator_username is None:
creator_username = dtoolcore.utils.getuser()
datetime_obj = datetime.datetime.utcnow()
admin_metadata = {'uuid': str(uuid.uuid4()), 'dtoolcore_version':
__version__, 'name': name, 'type': 'protodataset',
'creator_username': creator_username, 'created_at': dtoolcore.utils
.timestamp(datetime_obj)}
return admin_metadata
| generate_admin_metadata | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(name, creator_username=None):\n if not dtoolcore.utils.name_is_valid(name):\n raise DtoolCoreInvalidNameError()\n if creator_username is None:\n creator_username = dtoolcore.utils.getuser()\n datetime_obj = datetime.datetime.utcnow()\n admin_metadata = {'uuid': str(uuid.uuid4()), 'dtoolcore_version':\n __version__, 'name': name, 'type': 'protodataset',\n 'creator_username': creator_username, 'created_at': dtoolcore.utils\n .timestamp(datetime_obj)}\n return admin_metadata\n",
"role": "user"
},
{
"content": "generate_admin_metadata",
"role": "assistant"
}
] |
def x(admin_metadata, base_uri):
name = admin_metadata['name']
uuid = admin_metadata['uuid']
StorageBroker = _get_storage_broker(base_uri, config_path=None)
return StorageBroker.generate_uri(name, uuid, base_uri)
| generate_uri | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(admin_metadata, base_uri):\n name = admin_metadata['name']\n uuid = admin_metadata['uuid']\n StorageBroker = _get_storage_broker(base_uri, config_path=None)\n return StorageBroker.generate_uri(name, uuid, base_uri)\n",
"role": "user"
},
{
"content": "generate_uri",
"role": "assistant"
}
] |
def x(admin_metadata, base_uri, config_path=None):
uri = _generate_uri(admin_metadata, base_uri)
return ProtoDataSet(uri, admin_metadata, config_path)
| generate_proto_dataset | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(admin_metadata, base_uri, config_path=None):\n uri = _generate_uri(admin_metadata, base_uri)\n return ProtoDataSet(uri, admin_metadata, config_path)\n",
"role": "user"
},
{
"content": "generate_proto_dataset",
"role": "assistant"
}
] |
def x(src_uri, dest_base_uri, config_path=None, progressbar=None):
dataset = DataSet.from_uri(src_uri)
proto_dataset = _copy_create_proto_dataset(dataset, dest_base_uri,
config_path, progressbar)
_copy_content(dataset, proto_dataset, progressbar)
proto_dataset.freeze(progressbar=progressbar)
return proto_dataset.uri
| copy | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(src_uri, dest_base_uri, config_path=None, progressbar=None):\n dataset = DataSet.from_uri(src_uri)\n proto_dataset = _copy_create_proto_dataset(dataset, dest_base_uri,\n config_path, progressbar)\n _copy_content(dataset, proto_dataset, progressbar)\n proto_dataset.freeze(progressbar=progressbar)\n return proto_dataset.uri\n",
"role": "user"
},
{
"content": "copy",
"role": "assistant"
}
] |
def x(src_uri, dest_base_uri, config_path=None, progressbar=None):
dataset = DataSet.from_uri(src_uri)
dest_uri = _generate_uri(dataset._admin_metadata, dest_base_uri)
proto_dataset = ProtoDataSet.from_uri(dest_uri)
_copy_content(dataset, proto_dataset, progressbar)
proto_dataset.freeze(progressbar=progressbar)
return proto_dataset.uri
| copy_resume | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(src_uri, dest_base_uri, config_path=None, progressbar=None):\n dataset = DataSet.from_uri(src_uri)\n dest_uri = _generate_uri(dataset._admin_metadata, dest_base_uri)\n proto_dataset = ProtoDataSet.from_uri(dest_uri)\n _copy_content(dataset, proto_dataset, progressbar)\n proto_dataset.freeze(progressbar=progressbar)\n return proto_dataset.uri\n",
"role": "user"
},
{
"content": "copy_resume",
"role": "assistant"
}
] |
def x(self, new_name):
if not dtoolcore.utils.name_is_valid(new_name):
raise DtoolCoreInvalidNameError()
self._admin_metadata['name'] = new_name
if self._storage_broker.has_admin_metadata():
self._storage_broker.put_admin_metadata(self._admin_metadata)
| update_name | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, new_name):\n if not dtoolcore.utils.name_is_valid(new_name):\n raise DtoolCoreInvalidNameError()\n self._admin_metadata['name'] = new_name\n if self._storage_broker.has_admin_metadata():\n self._storage_broker.put_admin_metadata(self._admin_metadata)\n",
"role": "user"
},
{
"content": "update_name",
"role": "assistant"
}
] |
def x(self, overlay_name, overlay):
if not isinstance(overlay, dict):
raise TypeError('Overlay must be dict')
if set(self._identifiers()) != set(overlay.keys()):
raise ValueError('Overlay keys must be dataset identifiers')
self._storage_broker.put_overlay(overlay_name, overlay)
| put_overlay | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, overlay_name, overlay):\n if not isinstance(overlay, dict):\n raise TypeError('Overlay must be dict')\n if set(self._identifiers()) != set(overlay.keys()):\n raise ValueError('Overlay keys must be dataset identifiers')\n self._storage_broker.put_overlay(overlay_name, overlay)\n",
"role": "user"
},
{
"content": "put_overlay",
"role": "assistant"
}
] |
def x(self, progressbar=None):
items = dict()
if progressbar:
progressbar.label = 'Generating manifest'
for handle in self._storage_broker.iter_item_handles():
key = dtoolcore.utils.generate_identifier(handle)
value = self._storage_broker.item_properties(handle)
items[key] = value
if progressbar:
progressbar.item_show_func = lambda x: handle
progressbar.update(1)
manifest = {'items': items, 'dtoolcore_version': __version__,
'hash_function': self._storage_broker.hasher.name}
return manifest
| generate_manifest | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, progressbar=None):\n items = dict()\n if progressbar:\n progressbar.label = 'Generating manifest'\n for handle in self._storage_broker.iter_item_handles():\n key = dtoolcore.utils.generate_identifier(handle)\n value = self._storage_broker.item_properties(handle)\n items[key] = value\n if progressbar:\n progressbar.item_show_func = lambda x: handle\n progressbar.update(1)\n manifest = {'items': items, 'dtoolcore_version': __version__,\n 'hash_function': self._storage_broker.hasher.name}\n return manifest\n",
"role": "user"
},
{
"content": "generate_manifest",
"role": "assistant"
}
] |
def x(self):
if self._manifest_cache is None:
self._manifest_cache = self._storage_broker.get_manifest()
return self._manifest_cache
| manifest | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._manifest_cache is None:\n self._manifest_cache = self._storage_broker.get_manifest()\n return self._manifest_cache\n",
"role": "user"
},
{
"content": "manifest",
"role": "assistant"
}
] |
def x(self):
for handle in self._storage_broker.iter_item_handles():
yield dtoolcore.utils.generate_identifier(handle)
| identifiers | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n for handle in self._storage_broker.iter_item_handles():\n yield dtoolcore.utils.generate_identifier(handle)\n",
"role": "user"
},
{
"content": "identifiers",
"role": "assistant"
}
] |
def x(self):
self._storage_broker.create_structure()
self._storage_broker.put_admin_metadata(self._admin_metadata)
| create | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self._storage_broker.create_structure()\n self._storage_broker.put_admin_metadata(self._admin_metadata)\n",
"role": "user"
},
{
"content": "create",
"role": "assistant"
}
] |
def x(self, handle, key, value):
self._storage_broker.add_item_metadata(handle, key, value)
| add_item_metadata | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, handle, key, value):\n self._storage_broker.add_item_metadata(handle, key, value)\n",
"role": "user"
},
{
"content": "add_item_metadata",
"role": "assistant"
}
] |
def x(self):
overlays = defaultdict(dict)
for handle in self._storage_broker.iter_item_handles():
identifier = dtoolcore.utils.generate_identifier(handle)
item_metadata = self._storage_broker.get_item_metadata(handle)
for k, v in item_metadata.items():
overlays[k][identifier] = v
return overlays
| generate_overlays | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n overlays = defaultdict(dict)\n for handle in self._storage_broker.iter_item_handles():\n identifier = dtoolcore.utils.generate_identifier(handle)\n item_metadata = self._storage_broker.get_item_metadata(handle)\n for k, v in item_metadata.items():\n overlays[k][identifier] = v\n return overlays\n",
"role": "user"
},
{
"content": "generate_overlays",
"role": "assistant"
}
] |
def x(self, progressbar=None):
self._storage_broker.pre_freeze_hook()
if progressbar:
progressbar.label = 'Freezing dataset'
manifest = self.generate_manifest(progressbar=progressbar)
self._storage_broker.put_manifest(manifest)
overlays = self._generate_overlays()
for overlay_name, overlay in overlays.items():
self._put_overlay(overlay_name, overlay)
datetime_obj = datetime.datetime.utcnow()
metadata_update = {'type': 'dataset', 'frozen_at': dtoolcore.utils.
timestamp(datetime_obj)}
self._admin_metadata.update(metadata_update)
self._storage_broker.put_admin_metadata(self._admin_metadata)
self._storage_broker.post_freeze_hook()
| freeze | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, progressbar=None):\n self._storage_broker.pre_freeze_hook()\n if progressbar:\n progressbar.label = 'Freezing dataset'\n manifest = self.generate_manifest(progressbar=progressbar)\n self._storage_broker.put_manifest(manifest)\n overlays = self._generate_overlays()\n for overlay_name, overlay in overlays.items():\n self._put_overlay(overlay_name, overlay)\n datetime_obj = datetime.datetime.utcnow()\n metadata_update = {'type': 'dataset', 'frozen_at': dtoolcore.utils.\n timestamp(datetime_obj)}\n self._admin_metadata.update(metadata_update)\n self._storage_broker.put_admin_metadata(self._admin_metadata)\n self._storage_broker.post_freeze_hook()\n",
"role": "user"
},
{
"content": "freeze",
"role": "assistant"
}
] |
def x(self, route_name_prefix, base_url, handler):
route_name = route_name_prefix + '_read_many'
self.add_handler(route_name, base_url, handler, action='read_many',
request_method='GET')
route_name = route_name_prefix + '_read_one'
self.add_handler(route_name, base_url + '/{id}', handler, action=
'read_one', request_method='GET')
route_name = route_name_prefix + '_count'
self.add_handler(route_name, base_url + '/count', handler, action=
'count', request_method='GET')
route_name = route_name_prefix + '_create'
self.add_handler(route_name, base_url, handler, action='create',
request_method='POST')
route_name = route_name_prefix + '_update'
self.add_handler(route_name, base_url + '/{id}', handler, action=
'update', request_method='PUT')
route_name = route_name_prefix + '_delete'
self.add_handler(route_name, base_url + '/{id}', handler, action=
'delete', request_method='DELETE')
| add_papyrus_handler | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, route_name_prefix, base_url, handler):\n route_name = route_name_prefix + '_read_many'\n self.add_handler(route_name, base_url, handler, action='read_many',\n request_method='GET')\n route_name = route_name_prefix + '_read_one'\n self.add_handler(route_name, base_url + '/{id}', handler, action=\n 'read_one', request_method='GET')\n route_name = route_name_prefix + '_count'\n self.add_handler(route_name, base_url + '/count', handler, action=\n 'count', request_method='GET')\n route_name = route_name_prefix + '_create'\n self.add_handler(route_name, base_url, handler, action='create',\n request_method='POST')\n route_name = route_name_prefix + '_update'\n self.add_handler(route_name, base_url + '/{id}', handler, action=\n 'update', request_method='PUT')\n route_name = route_name_prefix + '_delete'\n self.add_handler(route_name, base_url + '/{id}', handler, action=\n 'delete', request_method='DELETE')\n",
"role": "user"
},
{
"content": "add_papyrus_handler",
"role": "assistant"
}
] |
def x(self, route_name_prefix, base_url):
route_name = route_name_prefix + '_read_many'
self.add_route(route_name, base_url, request_method='GET')
route_name = route_name_prefix + '_read_one'
self.add_route(route_name, base_url + '/{id}', request_method='GET')
route_name = route_name_prefix + '_count'
self.add_route(route_name, base_url + '/count', request_method='GET')
route_name = route_name_prefix + '_create'
self.add_route(route_name, base_url, request_method='POST')
route_name = route_name_prefix + '_update'
self.add_route(route_name, base_url + '/{id}', request_method='PUT')
route_name = route_name_prefix + '_delete'
self.add_route(route_name, base_url + '/{id}', request_method='DELETE')
| add_papyrus_routes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, route_name_prefix, base_url):\n route_name = route_name_prefix + '_read_many'\n self.add_route(route_name, base_url, request_method='GET')\n route_name = route_name_prefix + '_read_one'\n self.add_route(route_name, base_url + '/{id}', request_method='GET')\n route_name = route_name_prefix + '_count'\n self.add_route(route_name, base_url + '/count', request_method='GET')\n route_name = route_name_prefix + '_create'\n self.add_route(route_name, base_url, request_method='POST')\n route_name = route_name_prefix + '_update'\n self.add_route(route_name, base_url + '/{id}', request_method='PUT')\n route_name = route_name_prefix + '_delete'\n self.add_route(route_name, base_url + '/{id}', request_method='DELETE')\n",
"role": "user"
},
{
"content": "add_papyrus_routes",
"role": "assistant"
}
] |
def x(f):
@wraps(f)
def x(t, **kw):
def_vals = {'v': 0.0, 'h': 1.0, 'x': 0.0, 'w': 1.0, 's': 1.1, 'e': 1.0}
for v in def_vals:
if v not in kw:
kw[v] = def_vals[v]
anames, _, _, _ = inspect.getargspec(f)
fkw = dict([(arg, kw[arg]) for arg in anames if arg in kw])
ta = t
if 'x' not in anames:
ta = ta - kw['x']
if 'w' not in anames:
ta = ta / kw['w']
mod = f(ta, **fkw)
mod = f(ta + ta[mod.argmax()], **fkw)
return kw['v'] + kw['h'] / max(mod) * mod
args = set(['v', 'h', 'x', 'w'])
anames, _, _, _ = inspect.getargspec(f)
wrapped_f._peakargs = list(args.union([a for a in anames if a not in (
't', 'r')]))
return wrapped_f
| peak_model | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(f):\n\n @wraps(f)\n def x(t, **kw):\n def_vals = {'v': 0.0, 'h': 1.0, 'x': 0.0, 'w': 1.0, 's': 1.1, 'e': 1.0}\n for v in def_vals:\n if v not in kw:\n kw[v] = def_vals[v]\n anames, _, _, _ = inspect.getargspec(f)\n fkw = dict([(arg, kw[arg]) for arg in anames if arg in kw])\n ta = t\n if 'x' not in anames:\n ta = ta - kw['x']\n if 'w' not in anames:\n ta = ta / kw['w']\n mod = f(ta, **fkw)\n mod = f(ta + ta[mod.argmax()], **fkw)\n return kw['v'] + kw['h'] / max(mod) * mod\n args = set(['v', 'h', 'x', 'w'])\n anames, _, _, _ = inspect.getargspec(f)\n wrapped_f._peakargs = list(args.union([a for a in anames if a not in (\n 't', 'r')]))\n return wrapped_f\n",
"role": "user"
},
{
"content": "peak_model",
"role": "assistant"
}
] |
def x(match_downloaded_callback, on_exit_callback, conf,
synchronize_callback=True):
logger = logging.getLogger(__name__)
if conf['logging_level'] != logging.NOTSET:
logger.setLevel(conf['logging_level'])
else:
pass
def x(players_to_analyze, analyzed_players, matches_to_download,
downloaded_matches):
logger.info('Reached the checkpoint.'.format(datetime.datetime.now(
).strftime('%m-%d %H:%M:%S'), len(downloaded_matches)))
if on_exit_callback:
on_exit_callback(players_to_analyze, analyzed_players,
matches_to_download, downloaded_matches)
players_to_analyze = set(conf['seed_players_id'])
downloaded_matches = set(conf['downloaded_matches'])
logger.info('{} previously downloaded matches'.format(len(
downloaded_matches)))
matches_to_download = set(conf['matches_to_download'])
logger.info('{} matches to download'.format(len(matches_to_download)))
analyzed_players = set()
pta_lock = threading.Lock()
players_available_condition = threading.Condition(pta_lock)
mtd_lock = threading.Lock()
matches_Available_condition = threading.Condition(mtd_lock)
user_function_lock = threading.Lock(
) if synchronize_callback else NoOpContextManager()
logger_lock = threading.Lock()
player_downloader_threads = []
match_downloader_threads = []
try:
def x():
if len(player_downloader_threads) < max_players_download_threads:
player_downloader = PlayerDownloader(conf,
players_to_analyze, analyzed_players, pta_lock,
players_available_condition, matches_to_download,
mtd_lock, matches_Available_condition, logger, logger_lock)
player_downloader.start()
player_downloader_threads.append(player_downloader)
with logger_lock:
logger.info(
'Adding a player download thread. Threads: ' + str(
len(player_downloader_threads)))
else:
with logger_lock:
logger.debug(
'Tried adding a player download thread, but there are already the maximum number: '
+ str(max_players_download_threads))
def x():
if len(player_downloader_threads) > 1:
player_downloader_threads.pop().shutdown()
with logger_lock:
logger.info(
'Removing a player downloader thread. Threads: ' +
str(len(player_downloader_threads)))
else:
with logger_lock:
logger.debug(
'Tried removing a player download thread, but there is only one left'
)
logger.info('Starting fetching..')
create_thread()
for _ in range(matches_download_threads):
match_downloader = MatchDownloader(conf, players_to_analyze,
pta_lock, players_available_condition, matches_to_download,
downloaded_matches, mtd_lock, matches_Available_condition,
match_downloaded_callback, user_function_lock, logger,
logger_lock)
match_downloader.start()
match_downloader_threads.append(match_downloader)
auto_tuner = ThreadAutoTuner(create_thread, shutdown_thread)
for i, _ in enumerate(do_every(1)):
if conf.get('exit', False):
break
if i % 5 == 0:
with mtd_lock:
matches_in_queue = len(matches_to_download)
total_players = sum(th.total_downloads for th in
player_downloader_threads)
auto_tuner.update_thread_number(total_players, matches_in_queue
)
if i % logging_interval == 0:
with mtd_lock:
matches_in_queue = len(matches_to_download)
total_matches = sum(th.total_downloads for th in
match_downloader_threads)
with pta_lock:
players_in_queue = len(players_to_analyze)
total_players = sum(th.total_downloads for th in
player_downloader_threads)
with logger_lock:
logger.info(
'Players in queue: {}. Downloaded players: {}. Matches in queue: {}. Downloaded matches: {}'
.format(players_in_queue, total_players,
matches_in_queue, total_matches))
with pta_lock:
players_available_condition.notify_all()
with mtd_lock:
matches_Available_condition.notify_all()
logger.info('Terminating fetching')
finally:
conf['exit'] = True
for thread in (player_downloader_threads + match_downloader_threads):
thread.join()
logger.info('Calling checkpoint callback')
checkpoint(players_to_analyze, analyzed_players,
matches_to_download, downloaded_matches)
| download_matches | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(match_downloaded_callback, on_exit_callback, conf,\n synchronize_callback=True):\n logger = logging.getLogger(__name__)\n if conf['logging_level'] != logging.NOTSET:\n logger.setLevel(conf['logging_level'])\n else:\n pass\n\n def x(players_to_analyze, analyzed_players, matches_to_download,\n downloaded_matches):\n logger.info('Reached the checkpoint.'.format(datetime.datetime.now(\n ).strftime('%m-%d %H:%M:%S'), len(downloaded_matches)))\n if on_exit_callback:\n on_exit_callback(players_to_analyze, analyzed_players,\n matches_to_download, downloaded_matches)\n players_to_analyze = set(conf['seed_players_id'])\n downloaded_matches = set(conf['downloaded_matches'])\n logger.info('{} previously downloaded matches'.format(len(\n downloaded_matches)))\n matches_to_download = set(conf['matches_to_download'])\n logger.info('{} matches to download'.format(len(matches_to_download)))\n analyzed_players = set()\n pta_lock = threading.Lock()\n players_available_condition = threading.Condition(pta_lock)\n mtd_lock = threading.Lock()\n matches_Available_condition = threading.Condition(mtd_lock)\n user_function_lock = threading.Lock(\n ) if synchronize_callback else NoOpContextManager()\n logger_lock = threading.Lock()\n player_downloader_threads = []\n match_downloader_threads = []\n try:\n\n def x():\n if len(player_downloader_threads) < max_players_download_threads:\n player_downloader = PlayerDownloader(conf,\n players_to_analyze, analyzed_players, pta_lock,\n players_available_condition, matches_to_download,\n mtd_lock, matches_Available_condition, logger, logger_lock)\n player_downloader.start()\n player_downloader_threads.append(player_downloader)\n with logger_lock:\n logger.info(\n 'Adding a player download thread. Threads: ' + str(\n len(player_downloader_threads)))\n else:\n with logger_lock:\n logger.debug(\n 'Tried adding a player download thread, but there are already the maximum number: '\n + str(max_players_download_threads))\n\n def x():\n if len(player_downloader_threads) > 1:\n player_downloader_threads.pop().shutdown()\n with logger_lock:\n logger.info(\n 'Removing a player downloader thread. Threads: ' +\n str(len(player_downloader_threads)))\n else:\n with logger_lock:\n logger.debug(\n 'Tried removing a player download thread, but there is only one left'\n )\n logger.info('Starting fetching..')\n create_thread()\n for _ in range(matches_download_threads):\n match_downloader = MatchDownloader(conf, players_to_analyze,\n pta_lock, players_available_condition, matches_to_download,\n downloaded_matches, mtd_lock, matches_Available_condition,\n match_downloaded_callback, user_function_lock, logger,\n logger_lock)\n match_downloader.start()\n match_downloader_threads.append(match_downloader)\n auto_tuner = ThreadAutoTuner(create_thread, shutdown_thread)\n for i, _ in enumerate(do_every(1)):\n if conf.get('exit', False):\n break\n if i % 5 == 0:\n with mtd_lock:\n matches_in_queue = len(matches_to_download)\n total_players = sum(th.total_downloads for th in\n player_downloader_threads)\n auto_tuner.update_thread_number(total_players, matches_in_queue\n )\n if i % logging_interval == 0:\n with mtd_lock:\n matches_in_queue = len(matches_to_download)\n total_matches = sum(th.total_downloads for th in\n match_downloader_threads)\n with pta_lock:\n players_in_queue = len(players_to_analyze)\n total_players = sum(th.total_downloads for th in\n player_downloader_threads)\n with logger_lock:\n logger.info(\n 'Players in queue: {}. Downloaded players: {}. Matches in queue: {}. Downloaded matches: {}'\n .format(players_in_queue, total_players,\n matches_in_queue, total_matches))\n with pta_lock:\n players_available_condition.notify_all()\n with mtd_lock:\n matches_Available_condition.notify_all()\n logger.info('Terminating fetching')\n finally:\n conf['exit'] = True\n for thread in (player_downloader_threads + match_downloader_threads):\n thread.join()\n logger.info('Calling checkpoint callback')\n checkpoint(players_to_analyze, analyzed_players,\n matches_to_download, downloaded_matches)\n",
"role": "user"
},
{
"content": "download_matches",
"role": "assistant"
}
] |
def x(arr, value):
arr = np.array(arr)
idx = abs(arr - value).argmin()
return arr[idx]
| findNearest | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(arr, value):\n arr = np.array(arr)\n idx = abs(arr - value).argmin()\n return arr[idx]\n",
"role": "user"
},
{
"content": "findNearest",
"role": "assistant"
}
] |
def x():
magnitude_conversion_filepath = resource_stream(__name__,
'data/magnitude_conversion.dat')
raw_table = np.loadtxt(magnitude_conversion_filepath, '|S5')
magDict = {}
for row in raw_table:
if sys.hexversion >= 50331648:
starClass = row[1].decode('utf-8')
tableData = [x.decode('utf-8') for x in row[3:]]
else:
starClass = row[1]
tableData = row[3:]
magDict[starClass] = tableData
return magDict
| createMagConversionDict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n magnitude_conversion_filepath = resource_stream(__name__,\n 'data/magnitude_conversion.dat')\n raw_table = np.loadtxt(magnitude_conversion_filepath, '|S5')\n magDict = {}\n for row in raw_table:\n if sys.hexversion >= 50331648:\n starClass = row[1].decode('utf-8')\n tableData = [x.decode('utf-8') for x in row[3:]]\n else:\n starClass = row[1]\n tableData = row[3:]\n magDict[starClass] = tableData\n return magDict\n",
"role": "user"
},
{
"content": "createMagConversionDict",
"role": "assistant"
}
] |
def x(self, startClass, parentClass):
try:
if not startClass:
raise AttributeError
except AttributeError:
raise HierarchyError('This object ({0}) has no {1} as a parent object'
.format(self.name, parentClass))
if startClass.classType == parentClass:
return startClass
else:
return self._getParentClass(startClass.parent, parentClass)
| getParentClass | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, startClass, parentClass):\n try:\n if not startClass:\n raise AttributeError\n except AttributeError:\n raise HierarchyError('This object ({0}) has no {1} as a parent object'\n .format(self.name, parentClass))\n if startClass.classType == parentClass:\n return startClass\n else:\n return self._getParentClass(startClass.parent, parentClass)\n",
"role": "user"
},
{
"content": "getParentClass",
"role": "assistant"
}
] |
def x(self):
paramTemp = self.getParam('temperature')
if not paramTemp is np.nan:
return paramTemp
elif ed_params.estimateMissingValues:
self.flags.addFlag('Calculated Temperature')
return self.calcTemperature()
else:
return np.nan
| T | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n paramTemp = self.getParam('temperature')\n if not paramTemp is np.nan:\n return paramTemp\n elif ed_params.estimateMissingValues:\n self.flags.addFlag('Calculated Temperature')\n return self.calcTemperature()\n else:\n return np.nan\n",
"role": "user"
},
{
"content": "T",
"role": "assistant"
}
] |
def x(self):
d = self.parent.d
if ed_params.estimateMissingValues:
if d is np.nan:
d = self.estimateDistance()
if d is not np.nan:
self.flags.addFlag('Estimated Distance')
return d
else:
return np.nan
| d | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n d = self.parent.d\n if ed_params.estimateMissingValues:\n if d is np.nan:\n d = self.estimateDistance()\n if d is not np.nan:\n self.flags.addFlag('Estimated Distance')\n return d\n else:\n return np.nan\n",
"role": "user"
},
{
"content": "d",
"role": "assistant"
}
] |
def x(self, mag_letter):
allowed_mags = 'UBVJIHKLMN'
catalogue_mags = 'BVIJHK'
if mag_letter not in allowed_mags or not len(mag_letter) == 1:
raise ValueError('Magnitude letter must be a single letter in {0}'.
format(allowed_mags))
mag_str = 'mag' + mag_letter
mag_val = self.getParam(mag_str)
if isNanOrNone(mag_val) and ed_params.estimateMissingValues:
mag_dict = dict(('mag' + letter, self.getParam('mag' + letter)) for
letter in catalogue_mags)
mag_class = Magnitude(self.spectralType, **mag_dict)
try:
mag_conversion = mag_class.convert(mag_letter)
self.flags.addFlag('Estimated mag{0}'.format(mag_letter))
return mag_conversion
except ValueError as e:
logger.exception(e)
return np.nan
else:
return mag_val
| get_or_convert_magnitude | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, mag_letter):\n allowed_mags = 'UBVJIHKLMN'\n catalogue_mags = 'BVIJHK'\n if mag_letter not in allowed_mags or not len(mag_letter) == 1:\n raise ValueError('Magnitude letter must be a single letter in {0}'.\n format(allowed_mags))\n mag_str = 'mag' + mag_letter\n mag_val = self.getParam(mag_str)\n if isNanOrNone(mag_val) and ed_params.estimateMissingValues:\n mag_dict = dict(('mag' + letter, self.getParam('mag' + letter)) for\n letter in catalogue_mags)\n mag_class = Magnitude(self.spectralType, **mag_dict)\n try:\n mag_conversion = mag_class.convert(mag_letter)\n self.flags.addFlag('Estimated mag{0}'.format(mag_letter))\n return mag_conversion\n except ValueError as e:\n logger.exception(e)\n return np.nan\n else:\n return mag_val\n",
"role": "user"
},
{
"content": "get_or_convert_magnitude",
"role": "assistant"
}
] |
def x(self, wavelength=1.22):
tempind = [3500.0, 3750.0, 4000.0, 4250.0, 4500.0, 4750.0, 5000.0,
5250.0, 5500.0, 5750.0, 6000.0, 6250.0, 6500.0, 6750.0, 7000.0,
7250.0, 7500.0, 7750.0, 8000.0, 8250.0, 8500.0, 8750.0, 9000.0,
9250.0, 9500.0, 9750.0, 10000.0, 10250.0, 10500.0, 10750.0, 11000.0,
11250.0, 11500.0, 11750.0, 12000.0, 12250.0, 12500.0, 12750.0,
13000.0, 14000.0, 15000.0, 16000.0, 17000.0, 19000.0, 20000.0,
21000.0, 22000.0, 23000.0, 24000.0, 25000.0, 26000.0, 27000.0,
28000.0, 29000.0, 30000.0, 31000.0, 32000.0, 33000.0, 34000.0,
35000.0, 36000.0, 37000.0, 38000.0, 39000.0, 40000.0, 41000.0,
42000.0, 43000.0, 44000.0, 45000.0, 46000.0, 47000.0, 48000.0,
49000.0, 50000.0]
lggind = [0.0, 0.5, 1.0, 1.5, 2.0, 2.5, 3.0, 3.5, 4.0, 4.5, 5.0]
mhind = [-5.0, -4.5, -4.0, -3.5, -3.0, -2.5, -2.0, -1.5, -1.0, -0.5, -
0.3, -0.2, -0.1, 0.0, 0.1, 0.2, 0.3, 0.5, 1.0]
tempselect = _findNearest(tempind, float(self.T))
lgselect = _findNearest(lggind, float(self.calcLogg()))
mhselect = _findNearest(mhind, float(self.Z))
quadratic_filepath = resource_stream(__name__, 'data/quadratic.dat')
coeffTable = np.loadtxt(quadratic_filepath)
foundValues = False
for i in range(len(coeffTable)):
if coeffTable[i, 2] == lgselect and coeffTable[i, 3
] == tempselect and coeffTable[i, 4] == mhselect:
if coeffTable[i, 0] == 1:
u1array = coeffTable[i, 8:]
u2array = coeffTable[i + 1, 8:]
foundValues = True
break
if not foundValues:
raise ValueError('No limb darkening values could be found')
waveind = [0.365, 0.445, 0.551, 0.658, 0.806, 1.22, 1.63, 2.19, 3.45]
u1AtWavelength = np.interp(wavelength, waveind, u1array, left=0, right=0)
u2AtWavelength = np.interp(wavelength, waveind, u2array, left=0, right=0)
return u1AtWavelength, u2AtWavelength
| getLimbdarkeningCoeff | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, wavelength=1.22):\n tempind = [3500.0, 3750.0, 4000.0, 4250.0, 4500.0, 4750.0, 5000.0, \n 5250.0, 5500.0, 5750.0, 6000.0, 6250.0, 6500.0, 6750.0, 7000.0, \n 7250.0, 7500.0, 7750.0, 8000.0, 8250.0, 8500.0, 8750.0, 9000.0, \n 9250.0, 9500.0, 9750.0, 10000.0, 10250.0, 10500.0, 10750.0, 11000.0,\n 11250.0, 11500.0, 11750.0, 12000.0, 12250.0, 12500.0, 12750.0, \n 13000.0, 14000.0, 15000.0, 16000.0, 17000.0, 19000.0, 20000.0, \n 21000.0, 22000.0, 23000.0, 24000.0, 25000.0, 26000.0, 27000.0, \n 28000.0, 29000.0, 30000.0, 31000.0, 32000.0, 33000.0, 34000.0, \n 35000.0, 36000.0, 37000.0, 38000.0, 39000.0, 40000.0, 41000.0, \n 42000.0, 43000.0, 44000.0, 45000.0, 46000.0, 47000.0, 48000.0, \n 49000.0, 50000.0]\n lggind = [0.0, 0.5, 1.0, 1.5, 2.0, 2.5, 3.0, 3.5, 4.0, 4.5, 5.0]\n mhind = [-5.0, -4.5, -4.0, -3.5, -3.0, -2.5, -2.0, -1.5, -1.0, -0.5, -\n 0.3, -0.2, -0.1, 0.0, 0.1, 0.2, 0.3, 0.5, 1.0]\n tempselect = _findNearest(tempind, float(self.T))\n lgselect = _findNearest(lggind, float(self.calcLogg()))\n mhselect = _findNearest(mhind, float(self.Z))\n quadratic_filepath = resource_stream(__name__, 'data/quadratic.dat')\n coeffTable = np.loadtxt(quadratic_filepath)\n foundValues = False\n for i in range(len(coeffTable)):\n if coeffTable[i, 2] == lgselect and coeffTable[i, 3\n ] == tempselect and coeffTable[i, 4] == mhselect:\n if coeffTable[i, 0] == 1:\n u1array = coeffTable[i, 8:]\n u2array = coeffTable[i + 1, 8:]\n foundValues = True\n break\n if not foundValues:\n raise ValueError('No limb darkening values could be found')\n waveind = [0.365, 0.445, 0.551, 0.658, 0.806, 1.22, 1.63, 2.19, 3.45]\n u1AtWavelength = np.interp(wavelength, waveind, u1array, left=0, right=0)\n u2AtWavelength = np.interp(wavelength, waveind, u2array, left=0, right=0)\n return u1AtWavelength, u2AtWavelength\n",
"role": "user"
},
{
"content": "getLimbdarkeningCoeff",
"role": "assistant"
}
] |
def x(self):
try:
isTransiting = self.params['istransiting']
except KeyError:
return False
if isTransiting == '1':
return True
else:
return False
| isTransiting | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n try:\n isTransiting = self.params['istransiting']\n except KeyError:\n return False\n if isTransiting == '1':\n return True\n else:\n return False\n",
"role": "user"
},
{
"content": "isTransiting",
"role": "assistant"
}
] |
def x(self, circular=False):
try:
if circular:
return eq.transitDurationCircular(self.P, self.star.R, self.R,
self.a, self.i)
else:
return eq.TransitDuration(self.P, self.a, self.R, self.star.R,
self.i, self.e, self.periastron).Td
except (ValueError, AttributeError, HierarchyError):
return np.nan
| calcTransitDuration | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, circular=False):\n try:\n if circular:\n return eq.transitDurationCircular(self.P, self.star.R, self.R,\n self.a, self.i)\n else:\n return eq.TransitDuration(self.P, self.a, self.R, self.star.R,\n self.i, self.e, self.periastron).Td\n except (ValueError, AttributeError, HierarchyError):\n return np.nan\n",
"role": "user"
},
{
"content": "calcTransitDuration",
"role": "assistant"
}
] |
def x(self):
try:
return eq.MeanPlanetTemp(self.albedo, self.star.T, self.star.R, self.a
).T_p
except (ValueError, HierarchyError):
return np.nan
| calcTemperature | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n try:\n return eq.MeanPlanetTemp(self.albedo, self.star.T, self.star.R, self.a\n ).T_p\n except (ValueError, HierarchyError):\n return np.nan\n",
"role": "user"
},
{
"content": "calcTemperature",
"role": "assistant"
}
] |
def x(self):
try:
return eq.KeplersThirdLaw(None, self.star.M, self.P).a
except HierarchyError:
return np.nan
| calcSMA | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n try:\n return eq.KeplersThirdLaw(None, self.star.M, self.P).a\n except HierarchyError:\n return np.nan\n",
"role": "user"
},
{
"content": "calcSMA",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.