source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(self, model_id, filename, token=None, url=API_GET_PREDICTION_IMAGE_URL): auth = 'Bearer ' + self.check_for_token(token) h = {'Authorization': auth, 'Cache-Control': 'no-cache'} the_url = url with open(filename, 'rb') as image_file: encoded_string = base64.b64encode(image_file.read()) m = MultipartEncoder(fields={'sampleBase64Content': encoded_string, 'modelId': model_id}) h = {'Authorization': auth, 'Cache-Control': 'no-cache', 'Content-Type': m.content_type} r = requests.post(the_url, headers=h, data=m) return r
get_fileb64_image_prediction
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, model_id, filename, token=None, url=API_GET_PREDICTION_IMAGE_URL):\n auth = 'Bearer ' + self.check_for_token(token)\n h = {'Authorization': auth, 'Cache-Control': 'no-cache'}\n the_url = url\n with open(filename, 'rb') as image_file:\n encoded_string = base64.b64encode(image_file.read())\n m = MultipartEncoder(fields={'sampleBase64Content': encoded_string,\n 'modelId': model_id})\n h = {'Authorization': auth, 'Cache-Control': 'no-cache', 'Content-Type':\n m.content_type}\n r = requests.post(the_url, headers=h, data=m)\n return r\n", "role": "user" }, { "content": "get_fileb64_image_prediction", "role": "assistant" } ]
def x(self, model_id, b64_encoded_string, token=None, url= API_GET_PREDICTION_IMAGE_URL): auth = 'Bearer ' + self.check_for_token(token) h = {'Authorization': auth, 'Cache-Control': 'no-cache'} the_url = url encoded_string = b64_encoded_string m = MultipartEncoder(fields={'sampleBase64Content': encoded_string, 'modelId': model_id}) h = {'Authorization': auth, 'Cache-Control': 'no-cache', 'Content-Type': m.content_type} r = requests.post(the_url, headers=h, data=m) return r
get_b64_image_prediction
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, model_id, b64_encoded_string, token=None, url=\n API_GET_PREDICTION_IMAGE_URL):\n auth = 'Bearer ' + self.check_for_token(token)\n h = {'Authorization': auth, 'Cache-Control': 'no-cache'}\n the_url = url\n encoded_string = b64_encoded_string\n m = MultipartEncoder(fields={'sampleBase64Content': encoded_string,\n 'modelId': model_id})\n h = {'Authorization': auth, 'Cache-Control': 'no-cache', 'Content-Type':\n m.content_type}\n r = requests.post(the_url, headers=h, data=m)\n return r\n", "role": "user" }, { "content": "get_b64_image_prediction", "role": "assistant" } ]
def x(self, file_url, dataset_type='image', token=None, url=API_CREATE_DATASET ): auth = 'Bearer ' + self.check_for_token(token) m = MultipartEncoder(fields={'type': dataset_type, 'path': file_url}) h = {'Authorization': auth, 'Cache-Control': 'no-cache', 'Content-Type': m.content_type} the_url = url r = requests.post(the_url, headers=h, data=m) return r
create_dataset_synchronous
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, file_url, dataset_type='image', token=None, url=API_CREATE_DATASET\n ):\n auth = 'Bearer ' + self.check_for_token(token)\n m = MultipartEncoder(fields={'type': dataset_type, 'path': file_url})\n h = {'Authorization': auth, 'Cache-Control': 'no-cache', 'Content-Type':\n m.content_type}\n the_url = url\n r = requests.post(the_url, headers=h, data=m)\n return r\n", "role": "user" }, { "content": "create_dataset_synchronous", "role": "assistant" } ]
def x(self, dataset_id, model_name, token=None, url=API_TRAIN_MODEL): auth = 'Bearer ' + self.check_for_token(token) m = MultipartEncoder(fields={'name': model_name, 'datasetId': dataset_id}) h = {'Authorization': auth, 'Cache-Control': 'no-cache', 'Content-Type': m.content_type} the_url = url r = requests.post(the_url, headers=h, data=m) return r
train_model
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dataset_id, model_name, token=None, url=API_TRAIN_MODEL):\n auth = 'Bearer ' + self.check_for_token(token)\n m = MultipartEncoder(fields={'name': model_name, 'datasetId': dataset_id})\n h = {'Authorization': auth, 'Cache-Control': 'no-cache', 'Content-Type':\n m.content_type}\n the_url = url\n r = requests.post(the_url, headers=h, data=m)\n return r\n", "role": "user" }, { "content": "train_model", "role": "assistant" } ]
def x(self, model_id, token=None, url=API_TRAIN_MODEL): auth = 'Bearer ' + self.check_for_token(token) h = {'Authorization': auth, 'Cache-Control': 'no-cache'} the_url = url + '/' + model_id r = requests.get(the_url, headers=h) return r
get_training_status
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, model_id, token=None, url=API_TRAIN_MODEL):\n auth = 'Bearer ' + self.check_for_token(token)\n h = {'Authorization': auth, 'Cache-Control': 'no-cache'}\n the_url = url + '/' + model_id\n r = requests.get(the_url, headers=h)\n return r\n", "role": "user" }, { "content": "get_training_status", "role": "assistant" } ]
def x(self, dataset_id, token=None, url=API_GET_MODELS): auth = 'Bearer ' + self.check_for_token(token) h = {'Authorization': auth, 'Cache-Control': 'no-cache'} if url != API_GET_MODELS: r = requests.get(the_url, headers=h) return r the_url = url.replace('<dataset_id>', dataset_id) r = requests.get(the_url, headers=h) return r
get_models_info_for_dataset
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dataset_id, token=None, url=API_GET_MODELS):\n auth = 'Bearer ' + self.check_for_token(token)\n h = {'Authorization': auth, 'Cache-Control': 'no-cache'}\n if url != API_GET_MODELS:\n r = requests.get(the_url, headers=h)\n return r\n the_url = url.replace('<dataset_id>', dataset_id)\n r = requests.get(the_url, headers=h)\n return r\n", "role": "user" }, { "content": "get_models_info_for_dataset", "role": "assistant" } ]
def x(self, file_url, token=None, url=API_CREATE_LANGUAGE_DATASET): auth = 'Bearer ' + self.check_for_token(token) dummy_files = {'type': (None, 'text-intent'), 'path': (None, file_url)} h = {'Authorization': auth, 'Cache-Control': 'no-cache'} the_url = url r = requests.post(the_url, headers=h, files=dummy_files) return r
create_language_dataset_from_url
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, file_url, token=None, url=API_CREATE_LANGUAGE_DATASET):\n auth = 'Bearer ' + self.check_for_token(token)\n dummy_files = {'type': (None, 'text-intent'), 'path': (None, file_url)}\n h = {'Authorization': auth, 'Cache-Control': 'no-cache'}\n the_url = url\n r = requests.post(the_url, headers=h, files=dummy_files)\n return r\n", "role": "user" }, { "content": "create_language_dataset_from_url", "role": "assistant" } ]
def x(self, dataset_id, name, token=None, url=API_TRAIN_LANGUAGE_MODEL): auth = 'Bearer ' + self.check_for_token(token) dummy_files = {'name': (None, name), 'datasetId': (None, dataset_id)} h = {'Authorization': auth, 'Cache-Control': 'no-cache'} the_url = url r = requests.post(the_url, headers=h, files=dummy_files) return r
train_language_model_from_dataset
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dataset_id, name, token=None, url=API_TRAIN_LANGUAGE_MODEL):\n auth = 'Bearer ' + self.check_for_token(token)\n dummy_files = {'name': (None, name), 'datasetId': (None, dataset_id)}\n h = {'Authorization': auth, 'Cache-Control': 'no-cache'}\n the_url = url\n r = requests.post(the_url, headers=h, files=dummy_files)\n return r\n", "role": "user" }, { "content": "train_language_model_from_dataset", "role": "assistant" } ]
def x(self, model_id, token=None, url=API_TRAIN_LANGUAGE_MODEL): auth = 'Bearer ' + self.check_for_token(token) h = {'Authorization': auth, 'Cache-Control': 'no-cache'} the_url = url + '/' + model_id r = requests.get(the_url, headers=h) return r
get_language_model_status
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, model_id, token=None, url=API_TRAIN_LANGUAGE_MODEL):\n auth = 'Bearer ' + self.check_for_token(token)\n h = {'Authorization': auth, 'Cache-Control': 'no-cache'}\n the_url = url + '/' + model_id\n r = requests.get(the_url, headers=h)\n return r\n", "role": "user" }, { "content": "get_language_model_status", "role": "assistant" } ]
def x(self, model_id, document, token=None, url=API_GET_LANGUAGE_PREDICTION): auth = 'Bearer ' + self.check_for_token(token) dummy_files = {'modelId': (None, model_id), 'document': (None, document)} h = {'Authorization': auth, 'Cache-Control': 'no-cache'} the_url = url r = requests.post(the_url, headers=h, files=dummy_files) return r
get_language_prediction_from_model
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, model_id, document, token=None, url=API_GET_LANGUAGE_PREDICTION):\n auth = 'Bearer ' + self.check_for_token(token)\n dummy_files = {'modelId': (None, model_id), 'document': (None, document)}\n h = {'Authorization': auth, 'Cache-Control': 'no-cache'}\n the_url = url\n r = requests.post(the_url, headers=h, files=dummy_files)\n return r\n", "role": "user" }, { "content": "get_language_prediction_from_model", "role": "assistant" } ]
def x(self): files = [] files = [f for f in os.listdir() if f[-5:] == '.json'] if len(files) == 0: print('No json files found in this directory') return None max_boxes = 0 rows = [] for each_file in files: f = open(each_file, 'r') j = f.read() j = json.loads(j) f.close() if len(j['objects']) > max_boxes: max_boxes = len(j['objects']) row = [] for o in j['objects']: labels = {} labels['label'] = o['label'] labels['x'] = o['x_y_w_h'][0] labels['y'] = o['x_y_w_h'][1] labels['width'] = o['x_y_w_h'][2] labels['height'] = o['x_y_w_h'][3] labels_right_format = '"' + json.dumps(labels).replace('"', '""' ) + '"' row.append(labels_right_format) row.insert(0, '"' + j['filename'] + '"') rows.append(row) rows = [','.join(i) for i in rows] header = '"image"' for box_num in range(0, max_boxes): header += ', "box"' + str(box_num) rows.insert(0, header) return rows
parse_rectlabel_app_output
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n files = []\n files = [f for f in os.listdir() if f[-5:] == '.json']\n if len(files) == 0:\n print('No json files found in this directory')\n return None\n max_boxes = 0\n rows = []\n for each_file in files:\n f = open(each_file, 'r')\n j = f.read()\n j = json.loads(j)\n f.close()\n if len(j['objects']) > max_boxes:\n max_boxes = len(j['objects'])\n row = []\n for o in j['objects']:\n labels = {}\n labels['label'] = o['label']\n labels['x'] = o['x_y_w_h'][0]\n labels['y'] = o['x_y_w_h'][1]\n labels['width'] = o['x_y_w_h'][2]\n labels['height'] = o['x_y_w_h'][3]\n labels_right_format = '\"' + json.dumps(labels).replace('\"', '\"\"'\n ) + '\"'\n row.append(labels_right_format)\n row.insert(0, '\"' + j['filename'] + '\"')\n rows.append(row)\n rows = [','.join(i) for i in rows]\n header = '\"image\"'\n for box_num in range(0, max_boxes):\n header += ', \"box\"' + str(box_num)\n rows.insert(0, header)\n return rows\n", "role": "user" }, { "content": "parse_rectlabel_app_output", "role": "assistant" } ]
def x(self, output_filename='output.csv'): result = self.parse_rectlabel_app_output() ff = open(output_filename, 'w', encoding='utf8') for line in result: ff.write(line + '\n') ff.close()
save_parsed_data_to_csv
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, output_filename='output.csv'):\n result = self.parse_rectlabel_app_output()\n ff = open(output_filename, 'w', encoding='utf8')\n for line in result:\n ff.write(line + '\\n')\n ff.close()\n", "role": "user" }, { "content": "save_parsed_data_to_csv", "role": "assistant" } ]
def x(self, output_filename='output.csv'): result = self.XML_parse_rectlabel_app_output() ff = open(output_filename, 'w', encoding='utf8') for line in result: ff.write(line + '\n') ff.close()
XML_save_parsed_data_to_csv
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, output_filename='output.csv'):\n result = self.XML_parse_rectlabel_app_output()\n ff = open(output_filename, 'w', encoding='utf8')\n for line in result:\n ff.write(line + '\\n')\n ff.close()\n", "role": "user" }, { "content": "XML_save_parsed_data_to_csv", "role": "assistant" } ]
def x(istr): toks = re.findall('[^\\*\\\\\\+\\-\\^\\(\\)]+\\(?', istr) return [t for t in toks if not t.endswith('(')]
tokens
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(istr):\n toks = re.findall('[^\\\\*\\\\\\\\\\\\+\\\\-\\\\^\\\\(\\\\)]+\\\\(?', istr)\n return [t for t in toks if not t.endswith('(')]\n", "role": "user" }, { "content": "tokens", "role": "assistant" } ]
def x(istr): data = set(i.rstrip('0123456789') for i in tokens(istr)) has_events = not data.isdisjoint(istr_type_evts) has_2d = not data.isdisjoint(istr_type_2d) has_1d = data.difference(istr_type_evts).difference(istr_type_2d) != set() if has_events and not (has_1d or has_2d): return 'events' elif has_1d and not has_events: return '1d' elif has_2d and not (has_events or has_1d): return '2d' else: return None
istr_type
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(istr):\n data = set(i.rstrip('0123456789') for i in tokens(istr))\n has_events = not data.isdisjoint(istr_type_evts)\n has_2d = not data.isdisjoint(istr_type_2d)\n has_1d = data.difference(istr_type_evts).difference(istr_type_2d) != set()\n if has_events and not (has_1d or has_2d):\n return 'events'\n elif has_1d and not has_events:\n return '1d'\n elif has_2d and not (has_events or has_1d):\n return '2d'\n else:\n return None\n", "role": "user" }, { "content": "istr_type", "role": "assistant" } ]
def x(istr): fxn = istr.split('(')[0] if not fxn.isalnum() and fxn != '(' or istr[-1] != ')': return False plevel = 1 for c in '('.join(istr[:-1].split('(')[1:]): if c == '(': plevel += 1 elif c == ')': plevel -= 1 if plevel == 0: return False return True
is_parans_exp
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(istr):\n fxn = istr.split('(')[0]\n if not fxn.isalnum() and fxn != '(' or istr[-1] != ')':\n return False\n plevel = 1\n for c in '('.join(istr[:-1].split('(')[1:]):\n if c == '(':\n plevel += 1\n elif c == ')':\n plevel -= 1\n if plevel == 0:\n return False\n return True\n", "role": "user" }, { "content": "is_parans_exp", "role": "assistant" } ]
def x(istr, analyses, twin=None): if istr.strip() == '': return Trace() istr = istr.lstrip('+') if istr[0] == '-': return -parse_ion_string(istr[1:], analyses, twin) if is_parans_exp(istr): if ')' not in istr: pass fxn = istr.split('(')[0] args = istr[istr.find('(') + 1:istr.find(')')].split(',') if fxn == '': istr = args[0] else: ts = parse_ion_string(args[0], analyses, twin) return ts if set(istr).intersection(set('+-/*()')) == set(): if istr in SHORTCUTS: return parse_ion_string(SHORTCUTS[istr], analyses, twin) elif istr[0] == '!' and all(i in '0123456789.' for i in istr[1:]): return float(istr[1:]) elif istr == '!pi': return np.pi elif istr == '!e': return np.e else: return trace_resolver(istr, analyses, twin) for token in '/*+-^': if len(tokenize(istr, token)) != 1: ts = tokenize(istr, token) s = parse_ion_string(ts[0], analyses, twin) for t in ts[1:]: if token == '/': s /= parse_ion_string(t, analyses, twin) elif token == '*': s *= parse_ion_string(t, analyses, twin) elif token == '+': s += parse_ion_string(t, analyses, twin) elif token == '-': s -= parse_ion_string(t, analyses, twin) elif token == '^': s **= parse_ion_string(t, analyses, twin) return s raise Exception("Parser hit a point it shouldn't have!")
parse_ion_string
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(istr, analyses, twin=None):\n if istr.strip() == '':\n return Trace()\n istr = istr.lstrip('+')\n if istr[0] == '-':\n return -parse_ion_string(istr[1:], analyses, twin)\n if is_parans_exp(istr):\n if ')' not in istr:\n pass\n fxn = istr.split('(')[0]\n args = istr[istr.find('(') + 1:istr.find(')')].split(',')\n if fxn == '':\n istr = args[0]\n else:\n ts = parse_ion_string(args[0], analyses, twin)\n return ts\n if set(istr).intersection(set('+-/*()')) == set():\n if istr in SHORTCUTS:\n return parse_ion_string(SHORTCUTS[istr], analyses, twin)\n elif istr[0] == '!' and all(i in '0123456789.' for i in istr[1:]):\n return float(istr[1:])\n elif istr == '!pi':\n return np.pi\n elif istr == '!e':\n return np.e\n else:\n return trace_resolver(istr, analyses, twin)\n for token in '/*+-^':\n if len(tokenize(istr, token)) != 1:\n ts = tokenize(istr, token)\n s = parse_ion_string(ts[0], analyses, twin)\n for t in ts[1:]:\n if token == '/':\n s /= parse_ion_string(t, analyses, twin)\n elif token == '*':\n s *= parse_ion_string(t, analyses, twin)\n elif token == '+':\n s += parse_ion_string(t, analyses, twin)\n elif token == '-':\n s -= parse_ion_string(t, analyses, twin)\n elif token == '^':\n s **= parse_ion_string(t, analyses, twin)\n return s\n raise Exception(\"Parser hit a point it shouldn't have!\")\n", "role": "user" }, { "content": "parse_ion_string", "role": "assistant" } ]
def x(self, pedalboard, notify=True, force=False): if pedalboard is not None and pedalboard.bank is None: raise CurrentPedalboardError('Pedalboard {} has not added in any bank' .format(pedalboard)) if pedalboard == self.pedalboard and not force: return self._pedalboard = pedalboard self._device_controller.pedalboard = pedalboard self._save_current_pedalboard() if notify: self.app.components_observer.on_current_pedalboard_changed(self. pedalboard)
set_pedalboard
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, pedalboard, notify=True, force=False):\n if pedalboard is not None and pedalboard.bank is None:\n raise CurrentPedalboardError('Pedalboard {} has not added in any bank'\n .format(pedalboard))\n if pedalboard == self.pedalboard and not force:\n return\n self._pedalboard = pedalboard\n self._device_controller.pedalboard = pedalboard\n self._save_current_pedalboard()\n if notify:\n self.app.components_observer.on_current_pedalboard_changed(self.\n pedalboard)\n", "role": "user" }, { "content": "set_pedalboard", "role": "assistant" } ]
def x(self): if self.pedalboard is None: raise CurrentPedalboardError('The current pedalboard is None') before_index = self.pedalboard.index - 1 if before_index == -1: before_index = len(self.bank.pedalboards) - 1 self.set_pedalboard(self.bank.pedalboards[before_index])
to_before_pedalboard
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.pedalboard is None:\n raise CurrentPedalboardError('The current pedalboard is None')\n before_index = self.pedalboard.index - 1\n if before_index == -1:\n before_index = len(self.bank.pedalboards) - 1\n self.set_pedalboard(self.bank.pedalboards[before_index])\n", "role": "user" }, { "content": "to_before_pedalboard", "role": "assistant" } ]
def x(self): if self.pedalboard is None: raise CurrentPedalboardError('The current pedalboard is None') next_index = self.pedalboard.index + 1 if next_index == len(self.bank.pedalboards): next_index = 0 self.set_pedalboard(self.bank.pedalboards[next_index])
to_next_pedalboard
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.pedalboard is None:\n raise CurrentPedalboardError('The current pedalboard is None')\n next_index = self.pedalboard.index + 1\n if next_index == len(self.bank.pedalboards):\n next_index = 0\n self.set_pedalboard(self.bank.pedalboards[next_index])\n", "role": "user" }, { "content": "to_next_pedalboard", "role": "assistant" } ]
def x(self): if self.pedalboard is None: raise CurrentPedalboardError('The current pedalboard is None') before_index = self.bank.index - 1 if before_index == -1: before_index = len(self._manager.banks) - 1 self.set_bank(self._manager.banks[before_index])
to_before_bank
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.pedalboard is None:\n raise CurrentPedalboardError('The current pedalboard is None')\n before_index = self.bank.index - 1\n if before_index == -1:\n before_index = len(self._manager.banks) - 1\n self.set_bank(self._manager.banks[before_index])\n", "role": "user" }, { "content": "to_before_bank", "role": "assistant" } ]
def x(self): if self.pedalboard is None: raise CurrentPedalboardError('The current pedalboard is None') next_index = self.next_bank_index(self.bank.index) self.set_bank(self._manager.banks[next_index])
to_next_bank
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.pedalboard is None:\n raise CurrentPedalboardError('The current pedalboard is None')\n next_index = self.next_bank_index(self.bank.index)\n self.set_bank(self._manager.banks[next_index])\n", "role": "user" }, { "content": "to_next_bank", "role": "assistant" } ]
def x(self, bank, try_preserve_index=False): if bank not in self._manager: raise CurrentPedalboardError('Bank {} has not added in banks manager' .format(bank)) if self.bank == bank: return if bank.pedalboards: pedalboard = self._equivalent_pedalboard(bank ) if try_preserve_index else bank.pedalboards[0] self.set_pedalboard(pedalboard) else: self.set_pedalboard(None)
set_bank
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, bank, try_preserve_index=False):\n if bank not in self._manager:\n raise CurrentPedalboardError('Bank {} has not added in banks manager'\n .format(bank))\n if self.bank == bank:\n return\n if bank.pedalboards:\n pedalboard = self._equivalent_pedalboard(bank\n ) if try_preserve_index else bank.pedalboards[0]\n self.set_pedalboard(pedalboard)\n else:\n self.set_pedalboard(None)\n", "role": "user" }, { "content": "set_bank", "role": "assistant" } ]
def x(options, options_name, key, _type): if isinstance(options, dict): has = lambda k: k in options get = lambda k: options[k] elif isinstance(options, object): has = lambda k: hasattr(options, k) get = lambda k: getattr(options, k) else: raise ImproperlyConfigured('`{}` must be a dictionary-like object.' .format(options_name)) if not has(key): raise ImproperlyConfigured('`{}` must be specified in `{}`'.format( key, options_name)) value = get(key) if not isinstance(value, _type): raise ImproperlyConfigured('`{}` in `{}` must be a {}'.format(key, options_name, repr(_type))) return value
validate_and_get_value
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(options, options_name, key, _type):\n if isinstance(options, dict):\n has = lambda k: k in options\n get = lambda k: options[k]\n elif isinstance(options, object):\n has = lambda k: hasattr(options, k)\n get = lambda k: getattr(options, k)\n else:\n raise ImproperlyConfigured('`{}` must be a dictionary-like object.'\n .format(options_name))\n if not has(key):\n raise ImproperlyConfigured('`{}` must be specified in `{}`'.format(\n key, options_name))\n value = get(key)\n if not isinstance(value, _type):\n raise ImproperlyConfigured('`{}` in `{}` must be a {}'.format(key,\n options_name, repr(_type)))\n return value\n", "role": "user" }, { "content": "validate_and_get_value", "role": "assistant" } ]
def x(event): event_time = event.get(TIMESTAMP_FIELD) if event_time is None: event[TIMESTAMP_FIELD] = event_time = epoch_time_to_kronos_time(time .time()) elif type(event_time) not in (int, long): raise InvalidEventTime(event_time) _id = uuid_from_kronos_time(event_time) event[ID_FIELD] = str(_id) return _id, event
validate_event_and_assign_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(event):\n event_time = event.get(TIMESTAMP_FIELD)\n if event_time is None:\n event[TIMESTAMP_FIELD] = event_time = epoch_time_to_kronos_time(time\n .time())\n elif type(event_time) not in (int, long):\n raise InvalidEventTime(event_time)\n _id = uuid_from_kronos_time(event_time)\n event[ID_FIELD] = str(_id)\n return _id, event\n", "role": "user" }, { "content": "validate_event_and_assign_id", "role": "assistant" } ]
def x(stream): if not STREAM_REGEX.match(stream) or len(stream) > MAX_STREAM_LENGTH: raise InvalidStreamName(stream)
validate_stream
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(stream):\n if not STREAM_REGEX.match(stream) or len(stream) > MAX_STREAM_LENGTH:\n raise InvalidStreamName(stream)\n", "role": "user" }, { "content": "validate_stream", "role": "assistant" } ]
def x(storage_class, settings): if not isinstance(settings, dict): raise ImproperlyConfigured('{}: storage class settings must be a dict' .format(storage_class)) if not hasattr(storage_class, 'SETTINGS_VALIDATORS'): raise NotImplementedError( '{}: storage class must define `SETTINGS_VALIDATORS`'.format( storage_class)) settings_validators = getattr(storage_class, 'SETTINGS_VALIDATORS') settings = settings.copy() settings.pop('backend', None) invalid_settings = set(settings.keys()) - set(settings_validators.keys()) if invalid_settings: raise ImproperlyConfigured('{}: invalid settings: {}'.format( storage_class, invalid_settings)) for setting, value in settings.iteritems(): if not settings_validators[setting](value): raise ImproperlyConfigured('{}: invalid value for {}'.format( storage_class, setting))
validate_storage_settings
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(storage_class, settings):\n if not isinstance(settings, dict):\n raise ImproperlyConfigured('{}: storage class settings must be a dict'\n .format(storage_class))\n if not hasattr(storage_class, 'SETTINGS_VALIDATORS'):\n raise NotImplementedError(\n '{}: storage class must define `SETTINGS_VALIDATORS`'.format(\n storage_class))\n settings_validators = getattr(storage_class, 'SETTINGS_VALIDATORS')\n settings = settings.copy()\n settings.pop('backend', None)\n invalid_settings = set(settings.keys()) - set(settings_validators.keys())\n if invalid_settings:\n raise ImproperlyConfigured('{}: invalid settings: {}'.format(\n storage_class, invalid_settings))\n for setting, value in settings.iteritems():\n if not settings_validators[setting](value):\n raise ImproperlyConfigured('{}: invalid value for {}'.format(\n storage_class, setting))\n", "role": "user" }, { "content": "validate_storage_settings", "role": "assistant" } ]
def x(settings): storage = _validate_and_get_value(settings, 'settings', 'storage', dict) for name, options in storage.iteritems(): if 'backend' not in options: raise ImproperlyConfigured( "`storage['{}'] must contain a `backend` key".format(name)) path = options['backend'] module, cls = path.rsplit('.', 1) module = import_module(module) if not hasattr(module, cls): raise NotImplementedError('`{}` not implemented.'.format(cls)) validate_storage_settings(getattr(module, cls), options) namespace_to_streams_configuration = _validate_and_get_value(settings, 'settings', 'namespace_to_streams_configuration', dict) for namespace, prefix_confs in namespace_to_streams_configuration.iteritems( ): if '' not in prefix_confs: raise ImproperlyConfigured( 'Must specify backends for the null prefix') for prefix, options in prefix_confs.iteritems(): if prefix != '': validate_stream(prefix) backends = _validate_and_get_value(options, "namespace_to_streams_configuration['{}']['{}']".format( namespace, prefix), 'backends', dict) for backend in backends.keys(): if backend not in storage: raise ImproperlyConfigured( "`{}` backend for `namespace_to_streams_configuration['{}']['{}']` is not configured in `storage`" .format(backend, namespace, prefix)) read_backend = _validate_and_get_value(options, "namespace_to_streams_configuration['{}']['{}']".format( namespace, prefix), 'read_backend', str) if read_backend not in storage: raise ImproperlyConfigured( "`{}` backend for `namespace_to_streams_configuration['{}']['{}']` is not configured in `storage`" .format(read_backend, namespace, prefix)) stream = getattr(settings, 'stream', dict) _validate_and_get_value(stream, 'stream', 'format', re._pattern_type) node = getattr(settings, 'node', dict) _validate_and_get_value(node, 'node', 'greenlet_pool_size', int) _validate_and_get_value(node, 'node', 'id', str)
validate_settings
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(settings):\n storage = _validate_and_get_value(settings, 'settings', 'storage', dict)\n for name, options in storage.iteritems():\n if 'backend' not in options:\n raise ImproperlyConfigured(\n \"`storage['{}'] must contain a `backend` key\".format(name))\n path = options['backend']\n module, cls = path.rsplit('.', 1)\n module = import_module(module)\n if not hasattr(module, cls):\n raise NotImplementedError('`{}` not implemented.'.format(cls))\n validate_storage_settings(getattr(module, cls), options)\n namespace_to_streams_configuration = _validate_and_get_value(settings,\n 'settings', 'namespace_to_streams_configuration', dict)\n for namespace, prefix_confs in namespace_to_streams_configuration.iteritems(\n ):\n if '' not in prefix_confs:\n raise ImproperlyConfigured(\n 'Must specify backends for the null prefix')\n for prefix, options in prefix_confs.iteritems():\n if prefix != '':\n validate_stream(prefix)\n backends = _validate_and_get_value(options,\n \"namespace_to_streams_configuration['{}']['{}']\".format(\n namespace, prefix), 'backends', dict)\n for backend in backends.keys():\n if backend not in storage:\n raise ImproperlyConfigured(\n \"`{}` backend for `namespace_to_streams_configuration['{}']['{}']` is not configured in `storage`\"\n .format(backend, namespace, prefix))\n read_backend = _validate_and_get_value(options,\n \"namespace_to_streams_configuration['{}']['{}']\".format(\n namespace, prefix), 'read_backend', str)\n if read_backend not in storage:\n raise ImproperlyConfigured(\n \"`{}` backend for `namespace_to_streams_configuration['{}']['{}']` is not configured in `storage`\"\n .format(read_backend, namespace, prefix))\n stream = getattr(settings, 'stream', dict)\n _validate_and_get_value(stream, 'stream', 'format', re._pattern_type)\n node = getattr(settings, 'node', dict)\n _validate_and_get_value(node, 'node', 'greenlet_pool_size', int)\n _validate_and_get_value(node, 'node', 'id', str)\n", "role": "user" }, { "content": "validate_settings", "role": "assistant" } ]
def x(cls, **kwargs): vals = cls.get_non_empty_vals({cls._to_snake_case(k): v for k, v in kwargs.items()}) remove = [] for attr, val in vals.items(): try: vals[attr] = cls._parse_property(attr, val) except HelpScoutValidationException: remove.append(attr) logger.info('Unexpected property received in API response', exc_info=True) for attr in remove: del vals[attr] return cls(**cls.get_non_empty_vals(vals))
from_api
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, **kwargs):\n vals = cls.get_non_empty_vals({cls._to_snake_case(k): v for k, v in\n kwargs.items()})\n remove = []\n for attr, val in vals.items():\n try:\n vals[attr] = cls._parse_property(attr, val)\n except HelpScoutValidationException:\n remove.append(attr)\n logger.info('Unexpected property received in API response',\n exc_info=True)\n for attr in remove:\n del vals[attr]\n return cls(**cls.get_non_empty_vals(vals))\n", "role": "user" }, { "content": "from_api", "role": "assistant" } ]
def x(self): vals = {} for attribute, attribute_type in self._props.items(): prop = getattr(self, attribute) vals[self._to_camel_case(attribute)] = self._to_api_value( attribute_type, prop) return vals
to_api
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n vals = {}\n for attribute, attribute_type in self._props.items():\n prop = getattr(self, attribute)\n vals[self._to_camel_case(attribute)] = self._to_api_value(\n attribute_type, prop)\n return vals\n", "role": "user" }, { "content": "to_api", "role": "assistant" } ]
def x(self, attribute_type, value): if not value: return None if isinstance(attribute_type, properties.Instance): return value.to_api() if isinstance(attribute_type, properties.List): return self._parse_api_value_list(value) return attribute_type.serialize(value)
to_api_value
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, attribute_type, value):\n if not value:\n return None\n if isinstance(attribute_type, properties.Instance):\n return value.to_api()\n if isinstance(attribute_type, properties.List):\n return self._parse_api_value_list(value)\n return attribute_type.serialize(value)\n", "role": "user" }, { "content": "to_api_value", "role": "assistant" } ]
def x(self, values): try: return [v.to_api() for v in values] except AttributeError: return list(values)
parse_api_value_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, values):\n try:\n return [v.to_api() for v in values]\n except AttributeError:\n return list(values)\n", "role": "user" }, { "content": "parse_api_value_list", "role": "assistant" } ]
def x(cls, name, value): prop = cls._props.get(name) return_value = value if not prop: logger.debug( '"%s" with value "%s" is not a valid property for "%s".' % ( name, value, cls)) return_value = None elif isinstance(prop, properties.Instance): return_value = prop.instance_class.from_api(**value) elif isinstance(prop, properties.List): return_value = cls._parse_property_list(prop, value) elif isinstance(prop, properties.Color): return_value = cls._parse_property_color(value) return return_value
parse_property
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, name, value):\n prop = cls._props.get(name)\n return_value = value\n if not prop:\n logger.debug(\n '\"%s\" with value \"%s\" is not a valid property for \"%s\".' % (\n name, value, cls))\n return_value = None\n elif isinstance(prop, properties.Instance):\n return_value = prop.instance_class.from_api(**value)\n elif isinstance(prop, properties.List):\n return_value = cls._parse_property_list(prop, value)\n elif isinstance(prop, properties.Color):\n return_value = cls._parse_property_color(value)\n return return_value\n", "role": "user" }, { "content": "parse_property", "role": "assistant" } ]
def x(prop, value): attributes = [] for v in value: try: attributes.append(prop.prop.instance_class.from_api(**v)) except AttributeError: attributes.append(v) return attributes
parse_property_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(prop, value):\n attributes = []\n for v in value:\n try:\n attributes.append(prop.prop.instance_class.from_api(**v))\n except AttributeError:\n attributes.append(v)\n return attributes\n", "role": "user" }, { "content": "parse_property_list", "role": "assistant" } ]
def x(string): sub_string = '\\1_\\2' string = REGEX_CAMEL_FIRST.sub(sub_string, string) return REGEX_CAMEL_SECOND.sub(sub_string, string).lower()
to_snake_case
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(string):\n sub_string = '\\\\1_\\\\2'\n string = REGEX_CAMEL_FIRST.sub(sub_string, string)\n return REGEX_CAMEL_SECOND.sub(sub_string, string).lower()\n", "role": "user" }, { "content": "to_snake_case", "role": "assistant" } ]
def x(string): components = string.split('_') return '%s%s' % (components[0], ''.join(c.title() for c in components[1:]))
to_camel_case
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(string):\n components = string.split('_')\n return '%s%s' % (components[0], ''.join(c.title() for c in components[1:]))\n", "role": "user" }, { "content": "to_camel_case", "role": "assistant" } ]
def x(self, output=None): if output is None: if self.dbfile is None: return output = self.dbfile with open(output, 'w') as f: f.write(self._dump())
save
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, output=None):\n if output is None:\n if self.dbfile is None:\n return\n output = self.dbfile\n with open(output, 'w') as f:\n f.write(self._dump())\n", "role": "user" }, { "content": "save", "role": "assistant" } ]
def x(self, text=None, source=None): if text is not None: words = re.split('\\s+', text) wlen = len(words) for i in range(wlen - 2): self._insert(words[i:i + 3]) if source is not None: with open(source, 'r') as f: self.feed(f.read())
feed
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, text=None, source=None):\n if text is not None:\n words = re.split('\\\\s+', text)\n wlen = len(words)\n for i in range(wlen - 2):\n self._insert(words[i:i + 3])\n if source is not None:\n with open(source, 'r') as f:\n self.feed(f.read())\n", "role": "user" }, { "content": "feed", "role": "assistant" } ]
def x(self, **kwargs): words = list(map(self._sanitize, kwargs.get('words', []))) max_wlen = kwargs.get('wlen', 70) wlen = len(words) if wlen < 2: if not self._db: return '' if wlen == 0: words = sample(self._db.keys(), 1)[0].split(self._WSEP) elif wlen == 1: spl = [k for k in self._db.keys() if k.startswith(words[0] + self._WSEP)] words.append(sample(spl, 1)[0].split(self._WSEP)[1]) wlen = 2 while wlen < max_wlen: next_word = self._get(words[-2], words[-1]) if next_word is None: break words.append(next_word) wlen += 1 return ' '.join(words)
generate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n words = list(map(self._sanitize, kwargs.get('words', [])))\n max_wlen = kwargs.get('wlen', 70)\n wlen = len(words)\n if wlen < 2:\n if not self._db:\n return ''\n if wlen == 0:\n words = sample(self._db.keys(), 1)[0].split(self._WSEP)\n elif wlen == 1:\n spl = [k for k in self._db.keys() if k.startswith(words[0] +\n self._WSEP)]\n words.append(sample(spl, 1)[0].split(self._WSEP)[1])\n wlen = 2\n while wlen < max_wlen:\n next_word = self._get(words[-2], words[-1])\n if next_word is None:\n break\n words.append(next_word)\n wlen += 1\n return ' '.join(words)\n", "role": "user" }, { "content": "generate", "role": "assistant" } ]
def x(self): if self.dbfile is not None: with open(self.dbfile, 'r') as f: self._db = json.loads(f.read()) else: self._db = {}
load
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.dbfile is not None:\n with open(self.dbfile, 'r') as f:\n self._db = json.loads(f.read())\n else:\n self._db = {}\n", "role": "user" }, { "content": "load", "role": "assistant" } ]
def x(self, word1, word2): key = self._WSEP.join([self._sanitize(word1), self._sanitize(word2)]) key = key.lower() if key not in self._db: return return sample(self._db[key], 1)[0]
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, word1, word2):\n key = self._WSEP.join([self._sanitize(word1), self._sanitize(word2)])\n key = key.lower()\n if key not in self._db:\n return\n return sample(self._db[key], 1)[0]\n", "role": "user" }, { "content": "get", "role": "assistant" } ]
def x(self, trigram): words = list(map(self._sanitize, trigram)) key = self._WSEP.join(words[:2]).lower() next_word = words[2] self._db.setdefault(key, []) if next_word not in self._db[key]: self._db[key].append(next_word)
insert
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, trigram):\n words = list(map(self._sanitize, trigram))\n key = self._WSEP.join(words[:2]).lower()\n next_word = words[2]\n self._db.setdefault(key, [])\n if next_word not in self._db[key]:\n self._db[key].append(next_word)\n", "role": "user" }, { "content": "insert", "role": "assistant" } ]
def x(plan): cohort = plan['cohort'] action = plan['action'] source = plan['source'] cohort_start = datetime_to_kronos_time(_date_to_datetime(cohort['start'])) cohort_span = timedelta(**{cohort['unit']: cohort['cohorts']}) cohort_end = cohort['start'] + cohort_span action_span = timedelta(**{action['unit']: action['repetitions']}) action_end = cohort_end + action_span cohort_end = datetime_to_kronos_time(_date_to_datetime(cohort_end)) + 1 action_end = datetime_to_kronos_time(_date_to_datetime(action_end)) + 1 left = _cohort_stream_transform(source, cohort['stream'], cohort_start, cohort_end, cohort.get('transform'), cohort['grouping_key'], cohort ['unit']) right = _cohort_stream_transform(source, action['stream'], cohort_start, action_end, action.get('transform'), action['grouping_key'], action ['unit']) additional_action_time = DateUnit.unit_to_kronos_time(action['unit'] ) * action['repetitions'] left.alias = 'cohort' right.alias = 'action' joined = Join(left, right, Condition(Condition.Op.EQ, Property( 'cohort.%s' % cohort['grouping_key']), Property('action.%s' % action['grouping_key'])) & Condition(Condition.Op.GTE, Property( 'action.%s' % TIMESTAMP_FIELD), Property('cohort.%s' % TIMESTAMP_FIELD)) & Condition(Condition.Op.LT, Property('action.%s' % TIMESTAMP_FIELD), Add([Property('cohort.%s' % TIMESTAMP_FIELD), Constant(additional_action_time)]))) user_aggregated = Aggregate(joined, GroupBy([Property('cohort.date', alias=TIMESTAMP_FIELD), Property('cohort.%s' % cohort[ 'grouping_key'], alias='group'), Floor([Subtract([Property( 'action.%s' % TIMESTAMP_FIELD), Property('cohort.%s' % TIMESTAMP_FIELD)]), Constant(DateUnit.unit_to_kronos_time(action[ 'unit']))], alias='action_step')]), [Count([], alias='count')]) aggregated = Aggregate(user_aggregated, GroupBy([Property( TIMESTAMP_FIELD, alias=TIMESTAMP_FIELD), Property('action_step', alias='action_step')]), [Count([], alias='cohort_actions')]) return aggregated.to_dict()
cohort_queryplan
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(plan):\n cohort = plan['cohort']\n action = plan['action']\n source = plan['source']\n cohort_start = datetime_to_kronos_time(_date_to_datetime(cohort['start']))\n cohort_span = timedelta(**{cohort['unit']: cohort['cohorts']})\n cohort_end = cohort['start'] + cohort_span\n action_span = timedelta(**{action['unit']: action['repetitions']})\n action_end = cohort_end + action_span\n cohort_end = datetime_to_kronos_time(_date_to_datetime(cohort_end)) + 1\n action_end = datetime_to_kronos_time(_date_to_datetime(action_end)) + 1\n left = _cohort_stream_transform(source, cohort['stream'], cohort_start,\n cohort_end, cohort.get('transform'), cohort['grouping_key'], cohort\n ['unit'])\n right = _cohort_stream_transform(source, action['stream'], cohort_start,\n action_end, action.get('transform'), action['grouping_key'], action\n ['unit'])\n additional_action_time = DateUnit.unit_to_kronos_time(action['unit']\n ) * action['repetitions']\n left.alias = 'cohort'\n right.alias = 'action'\n joined = Join(left, right, Condition(Condition.Op.EQ, Property(\n 'cohort.%s' % cohort['grouping_key']), Property('action.%s' %\n action['grouping_key'])) & Condition(Condition.Op.GTE, Property(\n 'action.%s' % TIMESTAMP_FIELD), Property('cohort.%s' %\n TIMESTAMP_FIELD)) & Condition(Condition.Op.LT, Property('action.%s' %\n TIMESTAMP_FIELD), Add([Property('cohort.%s' % TIMESTAMP_FIELD),\n Constant(additional_action_time)])))\n user_aggregated = Aggregate(joined, GroupBy([Property('cohort.date',\n alias=TIMESTAMP_FIELD), Property('cohort.%s' % cohort[\n 'grouping_key'], alias='group'), Floor([Subtract([Property(\n 'action.%s' % TIMESTAMP_FIELD), Property('cohort.%s' %\n TIMESTAMP_FIELD)]), Constant(DateUnit.unit_to_kronos_time(action[\n 'unit']))], alias='action_step')]), [Count([], alias='count')])\n aggregated = Aggregate(user_aggregated, GroupBy([Property(\n TIMESTAMP_FIELD, alias=TIMESTAMP_FIELD), Property('action_step',\n alias='action_step')]), [Count([], alias='cohort_actions')])\n return aggregated.to_dict()\n", "role": "user" }, { "content": "cohort_queryplan", "role": "assistant" } ]
def x(mass): if mass is np.nan: return None for massLimit, massType in planetAssumptions['massType']: if mass < massLimit: return massType
planetMassType
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(mass):\n if mass is np.nan:\n return None\n for massLimit, massType in planetAssumptions['massType']:\n if mass < massLimit:\n return massType\n", "role": "user" }, { "content": "planetMassType", "role": "assistant" } ]
def x(radius): if radius is np.nan: return None for radiusLimit, radiusType in planetAssumptions['radiusType']: if radius < radiusLimit: return radiusType
planetRadiusType
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(radius):\n if radius is np.nan:\n return None\n for radiusLimit, radiusType in planetAssumptions['radiusType']:\n if radius < radiusLimit:\n return radiusType\n", "role": "user" }, { "content": "planetRadiusType", "role": "assistant" } ]
def x(temperature, mass, radius): if mass is not np.nan: sizeType = planetMassType(mass) elif radius is not np.nan: sizeType = planetRadiusType(radius) else: return None return '{0} {1}'.format(planetTempType(temperature), sizeType)
planetType
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(temperature, mass, radius):\n if mass is not np.nan:\n sizeType = planetMassType(mass)\n elif radius is not np.nan:\n sizeType = planetRadiusType(radius)\n else:\n return None\n return '{0} {1}'.format(planetTempType(temperature), sizeType)\n", "role": "user" }, { "content": "planetType", "role": "assistant" } ]
def x(df, columns=None): dtypes = df.dtypes if columns is None: columns = df.columns elif isinstance(columns, str): columns = [columns] for column in columns: expanded = np.repeat(df.values, df[column].apply(len).values, axis=0) expanded[:, df.columns.get_loc(column)] = np.concatenate(df[column] .tolist()) df = pd.DataFrame(expanded, columns=df.columns) for i, dtype in enumerate(dtypes): df.iloc[:, i] = df.iloc[:, i].astype(dtype) return df
split_array_like
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(df, columns=None):\n dtypes = df.dtypes\n if columns is None:\n columns = df.columns\n elif isinstance(columns, str):\n columns = [columns]\n for column in columns:\n expanded = np.repeat(df.values, df[column].apply(len).values, axis=0)\n expanded[:, df.columns.get_loc(column)] = np.concatenate(df[column]\n .tolist())\n df = pd.DataFrame(expanded, columns=df.columns)\n for i, dtype in enumerate(dtypes):\n df.iloc[:, i] = df.iloc[:, i].astype(dtype)\n return df\n", "role": "user" }, { "content": "split_array_like", "role": "assistant" } ]
def x(df1, df2, ignore_order=set(), ignore_indices=set(), all_close=False, _return_reason=False): result = _equals(df1, df2, ignore_order, ignore_indices, all_close) if _return_reason: return result else: return result[0]
equals
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(df1, df2, ignore_order=set(), ignore_indices=set(), all_close=False,\n _return_reason=False):\n result = _equals(df1, df2, ignore_order, ignore_indices, all_close)\n if _return_reason:\n return result\n else:\n return result[0]\n", "role": "user" }, { "content": "equals", "role": "assistant" } ]
def x(row, values, all_close, ignore_order): for row2 in values: mask = ma.getmaskarray(row2) assert mask.sum() in (0, len(mask)) if mask[0]: continue if _try_mask_row(row, row2, all_close, ignore_order): return True return False
try_mask_first_row
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(row, values, all_close, ignore_order):\n for row2 in values:\n mask = ma.getmaskarray(row2)\n assert mask.sum() in (0, len(mask))\n if mask[0]:\n continue\n if _try_mask_row(row, row2, all_close, ignore_order):\n return True\n return False\n", "role": "user" }, { "content": "try_mask_first_row", "role": "assistant" } ]
def x(row1, row2, all_close, ignore_order): if ignore_order: for value1 in row1: if not _try_mask_first_value(value1, row2, all_close): row2.mask = ma.nomask return False else: for value1, value2 in zip(row1, row2): if not _value_equals(value1, value2, all_close): return False row2[:] = ma.masked assert row2.mask.all() return True
try_mask_row
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(row1, row2, all_close, ignore_order):\n if ignore_order:\n for value1 in row1:\n if not _try_mask_first_value(value1, row2, all_close):\n row2.mask = ma.nomask\n return False\n else:\n for value1, value2 in zip(row1, row2):\n if not _value_equals(value1, value2, all_close):\n return False\n row2[:] = ma.masked\n assert row2.mask.all()\n return True\n", "role": "user" }, { "content": "try_mask_row", "role": "assistant" } ]
def x(value, row, all_close): for i, value2 in enumerate(row): if _value_equals(value, value2, all_close): row[i] = ma.masked return True return False
try_mask_first_value
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(value, row, all_close):\n for i, value2 in enumerate(row):\n if _value_equals(value, value2, all_close):\n row[i] = ma.masked\n return True\n return False\n", "role": "user" }, { "content": "try_mask_first_value", "role": "assistant" } ]
def x(value1, value2, all_close): if value1 is None: value1 = np.nan if value2 is None: value2 = np.nan are_floats = np.can_cast(type(value1), float) and np.can_cast(type( value2), float) if all_close and are_floats: return np.isclose(value1, value2, equal_nan=True) elif are_floats: return value1 == value2 or value1 != value1 and value2 != value2 else: return value1 == value2
value_equals
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(value1, value2, all_close):\n if value1 is None:\n value1 = np.nan\n if value2 is None:\n value2 = np.nan\n are_floats = np.can_cast(type(value1), float) and np.can_cast(type(\n value2), float)\n if all_close and are_floats:\n return np.isclose(value1, value2, equal_nan=True)\n elif are_floats:\n return value1 == value2 or value1 != value1 and value2 != value2\n else:\n return value1 == value2\n", "role": "user" }, { "content": "value_equals", "role": "assistant" } ]
def x(df1, df2, ignore_order=set(), ignore_indices=set(), all_close=False, _return_reason=False): equals_, reason = equals(df1, df2, ignore_order, ignore_indices, all_close, _return_reason=True) assert equals_, '{}\n\n{}\n\n{}'.format(reason, df1.to_string(), df2. to_string())
assert_equals
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(df1, df2, ignore_order=set(), ignore_indices=set(), all_close=False,\n _return_reason=False):\n equals_, reason = equals(df1, df2, ignore_order, ignore_indices,\n all_close, _return_reason=True)\n assert equals_, '{}\\n\\n{}\\n\\n{}'.format(reason, df1.to_string(), df2.\n to_string())\n", "role": "user" }, { "content": "assert_equals", "role": "assistant" } ]
def x(first, second): if first and second: return first.upper() == second.upper() else: return not (first or second)
is_equal_strings_ignore_case
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(first, second):\n if first and second:\n return first.upper() == second.upper()\n else:\n return not (first or second)\n", "role": "user" }, { "content": "is_equal_strings_ignore_case", "role": "assistant" } ]
def x(self, dim_name): for dim in self.dimensions: if is_equal_strings_ignore_case(dim.name, dim_name): return dim return None
find_dimension_by_name
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dim_name):\n for dim in self.dimensions:\n if is_equal_strings_ignore_case(dim.name, dim_name):\n return dim\n return None\n", "role": "user" }, { "content": "find_dimension_by_name", "role": "assistant" } ]
def x(self, dim_id): for dim in self.dimensions: if is_equal_strings_ignore_case(dim.id, dim_id): return dim return None
find_dimension_by_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dim_id):\n for dim in self.dimensions:\n if is_equal_strings_ignore_case(dim.id, dim_id):\n return dim\n return None\n", "role": "user" }, { "content": "find_dimension_by_id", "role": "assistant" } ]
def x(self): requestvalues = {'Dataset': self.dataset, 'Header': self. _get_item_array(self.header), 'Filter': self._get_item_array(self. filter), 'Stub': self._get_item_array(self.stub), 'Frequencies': self.frequencies} return json.dumps(requestvalues)
save_to_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n requestvalues = {'Dataset': self.dataset, 'Header': self.\n _get_item_array(self.header), 'Filter': self._get_item_array(self.\n filter), 'Stub': self._get_item_array(self.stub), 'Frequencies':\n self.frequencies}\n return json.dumps(requestvalues)\n", "role": "user" }, { "content": "save_to_json", "role": "assistant" } ]
def x(self): requestvalues = {'DatasetId': self.dataset, 'Name': self.name, 'Description': self.description, 'Source': self.source, 'PubDate': self.publication_date, 'AccessedOn': self.accessed_on, 'Url': self. dataset_ref, 'UploadFormatType': self.upload_format_type, 'Columns': self.columns, 'FileProperty': self.file_property.__dict__, 'FlatDSUpdateOptions': self.flat_ds_update_options, 'Public': self. public} return json.dumps(requestvalues)
save_to_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n requestvalues = {'DatasetId': self.dataset, 'Name': self.name,\n 'Description': self.description, 'Source': self.source, 'PubDate':\n self.publication_date, 'AccessedOn': self.accessed_on, 'Url': self.\n dataset_ref, 'UploadFormatType': self.upload_format_type, 'Columns':\n self.columns, 'FileProperty': self.file_property.__dict__,\n 'FlatDSUpdateOptions': self.flat_ds_update_options, 'Public': self.\n public}\n return json.dumps(requestvalues)\n", "role": "user" }, { "content": "save_to_json", "role": "assistant" } ]
def x(self): requestvalues = {'id': self.dataset, 'publicationDate': self. publication_date.strftime('%Y-%m-%d'), 'source': self.source, 'refUrl': self.refernce_url} return json.dumps(requestvalues)
save_to_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n requestvalues = {'id': self.dataset, 'publicationDate': self.\n publication_date.strftime('%Y-%m-%d'), 'source': self.source,\n 'refUrl': self.refernce_url}\n return json.dumps(requestvalues)\n", "role": "user" }, { "content": "save_to_json", "role": "assistant" } ]
def x(self, keyspace): if FORMAT_SPEC.search(keyspace): return KeyspacedProxy(self, keyspace) else: return KeyspacedProxy(self, self._keyspaces[keyspace])
keyspace
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, keyspace):\n if FORMAT_SPEC.search(keyspace):\n return KeyspacedProxy(self, keyspace)\n else:\n return KeyspacedProxy(self, self._keyspaces[keyspace])\n", "role": "user" }, { "content": "keyspace", "role": "assistant" } ]
def x(dir1, dir2): compared = dircmp(dir1, dir2, IGNORED_FILES) if (compared.left_only or compared.right_only or compared.diff_files or compared.funny_files): return False for subdir in compared.common_dirs: if not is_same(os.path.join(dir1, subdir), os.path.join(dir2, subdir)): return False return True
is_same
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(dir1, dir2):\n compared = dircmp(dir1, dir2, IGNORED_FILES)\n if (compared.left_only or compared.right_only or compared.diff_files or\n compared.funny_files):\n return False\n for subdir in compared.common_dirs:\n if not is_same(os.path.join(dir1, subdir), os.path.join(dir2, subdir)):\n return False\n return True\n", "role": "user" }, { "content": "is_same", "role": "assistant" } ]
def x(self): fcomp = filecmp.cmpfiles(self.left, self.right, self.common_files, shallow=False) self.same_files, self.diff_files, self.funny_files = fcomp
phase3
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n fcomp = filecmp.cmpfiles(self.left, self.right, self.common_files,\n shallow=False)\n self.same_files, self.diff_files, self.funny_files = fcomp\n", "role": "user" }, { "content": "phase3", "role": "assistant" } ]
def x(self, namespace, stream, events, configuration): index = self.index_manager.get_index(namespace) start_dts_to_add = set() def x(): for _id, event in events: dt = kronos_time_to_datetime(uuid_to_kronos_time(_id)) start_dts_to_add.add(_round_datetime_down(dt)) event['_index'] = index event['_type'] = stream event[LOGSTASH_TIMESTAMP_FIELD] = dt.isoformat() yield event list(es_helpers.streaming_bulk(self.es, actions(), chunk_size=1000, refresh=self.force_refresh)) self.index_manager.add_aliases(namespace, index, start_dts_to_add)
insert
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, namespace, stream, events, configuration):\n index = self.index_manager.get_index(namespace)\n start_dts_to_add = set()\n\n def x():\n for _id, event in events:\n dt = kronos_time_to_datetime(uuid_to_kronos_time(_id))\n start_dts_to_add.add(_round_datetime_down(dt))\n event['_index'] = index\n event['_type'] = stream\n event[LOGSTASH_TIMESTAMP_FIELD] = dt.isoformat()\n yield event\n list(es_helpers.streaming_bulk(self.es, actions(), chunk_size=1000,\n refresh=self.force_refresh))\n self.index_manager.add_aliases(namespace, index, start_dts_to_add)\n", "role": "user" }, { "content": "insert", "role": "assistant" } ]
def x(self, namespace, stream, start_id, end_time, order, limit, configuration ): indices = self.index_manager.get_aliases(namespace, uuid_to_kronos_time (start_id), end_time) if not indices: return end_id = uuid_from_kronos_time(end_time, _type=UUIDType.HIGHEST) end_id.descending = start_id.descending = descending = (order == ResultOrder.DESCENDING) start_time = uuid_to_kronos_time(start_id) body_query = {'query': {'filtered': {'query': {'match_all': {}}, 'filter': {'range': {TIMESTAMP_FIELD: {'gte': start_time, 'lte': end_time}}}}}} order = 'desc' if descending else 'asc' sort_query = ['%s:%s' % (TIMESTAMP_FIELD, order), '%s:%s' % (ID_FIELD, order)] last_id = end_id if descending else start_id scroll_id = None while True: size = max(min(limit, configuration['read_size']) / self.shards, 10) if scroll_id is None: res = self.es.search(index=indices, doc_type=stream, size=size, body=body_query, sort=sort_query, _source=True, scroll='1m', ignore=[400, 404], allow_no_indices=True, ignore_unavailable=True) else: res = self.es.scroll(scroll_id, scroll='1m') if '_scroll_id' not in res: break scroll_id = res['_scroll_id'] hits = res.get('hits', {}).get('hits') if not hits: break for hit in hits: _id = TimeUUID(hit['_id'], descending=descending) if _id <= last_id: continue last_id = _id event = hit['_source'] del event[LOGSTASH_TIMESTAMP_FIELD] yield json.dumps(event) limit -= 1 if limit == 0: break if scroll_id is not None: self.es.clear_scroll(scroll_id)
retrieve
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, namespace, stream, start_id, end_time, order, limit, configuration\n ):\n indices = self.index_manager.get_aliases(namespace, uuid_to_kronos_time\n (start_id), end_time)\n if not indices:\n return\n end_id = uuid_from_kronos_time(end_time, _type=UUIDType.HIGHEST)\n end_id.descending = start_id.descending = descending = (order ==\n ResultOrder.DESCENDING)\n start_time = uuid_to_kronos_time(start_id)\n body_query = {'query': {'filtered': {'query': {'match_all': {}},\n 'filter': {'range': {TIMESTAMP_FIELD: {'gte': start_time, 'lte':\n end_time}}}}}}\n order = 'desc' if descending else 'asc'\n sort_query = ['%s:%s' % (TIMESTAMP_FIELD, order), '%s:%s' % (ID_FIELD,\n order)]\n last_id = end_id if descending else start_id\n scroll_id = None\n while True:\n size = max(min(limit, configuration['read_size']) / self.shards, 10)\n if scroll_id is None:\n res = self.es.search(index=indices, doc_type=stream, size=size,\n body=body_query, sort=sort_query, _source=True, scroll='1m',\n ignore=[400, 404], allow_no_indices=True,\n ignore_unavailable=True)\n else:\n res = self.es.scroll(scroll_id, scroll='1m')\n if '_scroll_id' not in res:\n break\n scroll_id = res['_scroll_id']\n hits = res.get('hits', {}).get('hits')\n if not hits:\n break\n for hit in hits:\n _id = TimeUUID(hit['_id'], descending=descending)\n if _id <= last_id:\n continue\n last_id = _id\n event = hit['_source']\n del event[LOGSTASH_TIMESTAMP_FIELD]\n yield json.dumps(event)\n limit -= 1\n if limit == 0:\n break\n if scroll_id is not None:\n self.es.clear_scroll(scroll_id)\n", "role": "user" }, { "content": "retrieve", "role": "assistant" } ]
def x(self): cdir = os.path.dirname(os.path.realpath(__file__)) schema = os.path.join(cdir, 'data', 'partsmaster.sql') if self.debug: print(self.hdr, 'parts master schema is ', schema) self.populate(schema) schema = os.path.join(cdir, 'data', 'approvedmfg.sql') if self.debug: print(self.hdr, 'approved mfg list schema is ', schema) self.populate(schema) schema = os.path.join(cdir, 'data', 'attachment.sql') if self.debug: print(self.hdr, 'attachment schema is ', schema) self.populate(schema) schema = os.path.join(cdir, 'data', 'bom.sql') if self.debug: print(self.hdr, 'bill of materials schema is ', schema) self.populate(schema) return
create_tables
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n cdir = os.path.dirname(os.path.realpath(__file__))\n schema = os.path.join(cdir, 'data', 'partsmaster.sql')\n if self.debug:\n print(self.hdr, 'parts master schema is ', schema)\n self.populate(schema)\n schema = os.path.join(cdir, 'data', 'approvedmfg.sql')\n if self.debug:\n print(self.hdr, 'approved mfg list schema is ', schema)\n self.populate(schema)\n schema = os.path.join(cdir, 'data', 'attachment.sql')\n if self.debug:\n print(self.hdr, 'attachment schema is ', schema)\n self.populate(schema)\n schema = os.path.join(cdir, 'data', 'bom.sql')\n if self.debug:\n print(self.hdr, 'bill of materials schema is ', schema)\n self.populate(schema)\n return\n", "role": "user" }, { "content": "create_tables", "role": "assistant" } ]
def x(f): f.seek(0) try: p_rec_type = None while True: rec_off = f.tell() while True: if f.read(2) == b'\xff\xff': h = struct.unpack('<HH', f.read(4)) if h[1] < 64 and h[1] != 0: rec_type = f.read(h[1]) if rec_type[0] == 67: break if f.read(1) == b'': raise EOFError f.seek(f.tell() - 2) if p_rec_type is not None: rec_len = f.tell() - 6 - len(rec_type) - rec_off f.seek(rec_off) yield p_rec_type, f.read(rec_len) f.seek(f.tell() + 6 + len(rec_type)) p_rec_type = rec_type except EOFError: rec_len = f.tell() - 6 - len(rec_type) - rec_off f.seek(rec_off) yield p_rec_type, f.read(rec_len)
parse_c_serialized
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(f):\n f.seek(0)\n try:\n p_rec_type = None\n while True:\n rec_off = f.tell()\n while True:\n if f.read(2) == b'\\xff\\xff':\n h = struct.unpack('<HH', f.read(4))\n if h[1] < 64 and h[1] != 0:\n rec_type = f.read(h[1])\n if rec_type[0] == 67:\n break\n if f.read(1) == b'':\n raise EOFError\n f.seek(f.tell() - 2)\n if p_rec_type is not None:\n rec_len = f.tell() - 6 - len(rec_type) - rec_off\n f.seek(rec_off)\n yield p_rec_type, f.read(rec_len)\n f.seek(f.tell() + 6 + len(rec_type))\n p_rec_type = rec_type\n except EOFError:\n rec_len = f.tell() - 6 - len(rec_type) - rec_off\n f.seek(rec_off)\n yield p_rec_type, f.read(rec_len)\n", "role": "user" }, { "content": "parse_c_serialized", "role": "assistant" } ]
def x(self, t, dt=None, aggfunc=None): return self.data.scan(t, dt, aggfunc)
scan
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, t, dt=None, aggfunc=None):\n return self.data.scan(t, dt, aggfunc)\n", "role": "user" }, { "content": "scan", "role": "assistant" } ]
def x(): storage_broker_lookup = dict() for entrypoint in iter_entry_points('dtool.storage_brokers'): StorageBroker = entrypoint.load() storage_broker_lookup[StorageBroker.key] = StorageBroker return storage_broker_lookup
generate_storage_broker_lookup
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n storage_broker_lookup = dict()\n for entrypoint in iter_entry_points('dtool.storage_brokers'):\n StorageBroker = entrypoint.load()\n storage_broker_lookup[StorageBroker.key] = StorageBroker\n return storage_broker_lookup\n", "role": "user" }, { "content": "generate_storage_broker_lookup", "role": "assistant" } ]
def x(uri, config_path): uri = dtoolcore.utils.sanitise_uri(uri) storage_broker_lookup = _generate_storage_broker_lookup() parsed_uri = dtoolcore.utils.generous_parse_uri(uri) StorageBroker = storage_broker_lookup[parsed_uri.scheme] return StorageBroker(uri, config_path)
get_storage_broker
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(uri, config_path):\n uri = dtoolcore.utils.sanitise_uri(uri)\n storage_broker_lookup = _generate_storage_broker_lookup()\n parsed_uri = dtoolcore.utils.generous_parse_uri(uri)\n StorageBroker = storage_broker_lookup[parsed_uri.scheme]\n return StorageBroker(uri, config_path)\n", "role": "user" }, { "content": "get_storage_broker", "role": "assistant" } ]
def x(uri, config_path): uri = dtoolcore.utils.sanitise_uri(uri) storage_broker = _get_storage_broker(uri, config_path) admin_metadata = storage_broker.get_admin_metadata() return admin_metadata
admin_metadata_from_uri
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(uri, config_path):\n uri = dtoolcore.utils.sanitise_uri(uri)\n storage_broker = _get_storage_broker(uri, config_path)\n admin_metadata = storage_broker.get_admin_metadata()\n return admin_metadata\n", "role": "user" }, { "content": "admin_metadata_from_uri", "role": "assistant" } ]
def x(uri, config_path): uri = dtoolcore.utils.sanitise_uri(uri) storage_broker = _get_storage_broker(uri, config_path) return storage_broker.has_admin_metadata()
is_dataset
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(uri, config_path):\n uri = dtoolcore.utils.sanitise_uri(uri)\n storage_broker = _get_storage_broker(uri, config_path)\n return storage_broker.has_admin_metadata()\n", "role": "user" }, { "content": "is_dataset", "role": "assistant" } ]
def x(name, creator_username=None): if not dtoolcore.utils.name_is_valid(name): raise DtoolCoreInvalidNameError() if creator_username is None: creator_username = dtoolcore.utils.getuser() datetime_obj = datetime.datetime.utcnow() admin_metadata = {'uuid': str(uuid.uuid4()), 'dtoolcore_version': __version__, 'name': name, 'type': 'protodataset', 'creator_username': creator_username, 'created_at': dtoolcore.utils .timestamp(datetime_obj)} return admin_metadata
generate_admin_metadata
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(name, creator_username=None):\n if not dtoolcore.utils.name_is_valid(name):\n raise DtoolCoreInvalidNameError()\n if creator_username is None:\n creator_username = dtoolcore.utils.getuser()\n datetime_obj = datetime.datetime.utcnow()\n admin_metadata = {'uuid': str(uuid.uuid4()), 'dtoolcore_version':\n __version__, 'name': name, 'type': 'protodataset',\n 'creator_username': creator_username, 'created_at': dtoolcore.utils\n .timestamp(datetime_obj)}\n return admin_metadata\n", "role": "user" }, { "content": "generate_admin_metadata", "role": "assistant" } ]
def x(admin_metadata, base_uri): name = admin_metadata['name'] uuid = admin_metadata['uuid'] StorageBroker = _get_storage_broker(base_uri, config_path=None) return StorageBroker.generate_uri(name, uuid, base_uri)
generate_uri
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(admin_metadata, base_uri):\n name = admin_metadata['name']\n uuid = admin_metadata['uuid']\n StorageBroker = _get_storage_broker(base_uri, config_path=None)\n return StorageBroker.generate_uri(name, uuid, base_uri)\n", "role": "user" }, { "content": "generate_uri", "role": "assistant" } ]
def x(admin_metadata, base_uri, config_path=None): uri = _generate_uri(admin_metadata, base_uri) return ProtoDataSet(uri, admin_metadata, config_path)
generate_proto_dataset
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(admin_metadata, base_uri, config_path=None):\n uri = _generate_uri(admin_metadata, base_uri)\n return ProtoDataSet(uri, admin_metadata, config_path)\n", "role": "user" }, { "content": "generate_proto_dataset", "role": "assistant" } ]
def x(src_uri, dest_base_uri, config_path=None, progressbar=None): dataset = DataSet.from_uri(src_uri) proto_dataset = _copy_create_proto_dataset(dataset, dest_base_uri, config_path, progressbar) _copy_content(dataset, proto_dataset, progressbar) proto_dataset.freeze(progressbar=progressbar) return proto_dataset.uri
copy
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(src_uri, dest_base_uri, config_path=None, progressbar=None):\n dataset = DataSet.from_uri(src_uri)\n proto_dataset = _copy_create_proto_dataset(dataset, dest_base_uri,\n config_path, progressbar)\n _copy_content(dataset, proto_dataset, progressbar)\n proto_dataset.freeze(progressbar=progressbar)\n return proto_dataset.uri\n", "role": "user" }, { "content": "copy", "role": "assistant" } ]
def x(src_uri, dest_base_uri, config_path=None, progressbar=None): dataset = DataSet.from_uri(src_uri) dest_uri = _generate_uri(dataset._admin_metadata, dest_base_uri) proto_dataset = ProtoDataSet.from_uri(dest_uri) _copy_content(dataset, proto_dataset, progressbar) proto_dataset.freeze(progressbar=progressbar) return proto_dataset.uri
copy_resume
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(src_uri, dest_base_uri, config_path=None, progressbar=None):\n dataset = DataSet.from_uri(src_uri)\n dest_uri = _generate_uri(dataset._admin_metadata, dest_base_uri)\n proto_dataset = ProtoDataSet.from_uri(dest_uri)\n _copy_content(dataset, proto_dataset, progressbar)\n proto_dataset.freeze(progressbar=progressbar)\n return proto_dataset.uri\n", "role": "user" }, { "content": "copy_resume", "role": "assistant" } ]
def x(self, new_name): if not dtoolcore.utils.name_is_valid(new_name): raise DtoolCoreInvalidNameError() self._admin_metadata['name'] = new_name if self._storage_broker.has_admin_metadata(): self._storage_broker.put_admin_metadata(self._admin_metadata)
update_name
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, new_name):\n if not dtoolcore.utils.name_is_valid(new_name):\n raise DtoolCoreInvalidNameError()\n self._admin_metadata['name'] = new_name\n if self._storage_broker.has_admin_metadata():\n self._storage_broker.put_admin_metadata(self._admin_metadata)\n", "role": "user" }, { "content": "update_name", "role": "assistant" } ]
def x(self, overlay_name, overlay): if not isinstance(overlay, dict): raise TypeError('Overlay must be dict') if set(self._identifiers()) != set(overlay.keys()): raise ValueError('Overlay keys must be dataset identifiers') self._storage_broker.put_overlay(overlay_name, overlay)
put_overlay
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, overlay_name, overlay):\n if not isinstance(overlay, dict):\n raise TypeError('Overlay must be dict')\n if set(self._identifiers()) != set(overlay.keys()):\n raise ValueError('Overlay keys must be dataset identifiers')\n self._storage_broker.put_overlay(overlay_name, overlay)\n", "role": "user" }, { "content": "put_overlay", "role": "assistant" } ]
def x(self, progressbar=None): items = dict() if progressbar: progressbar.label = 'Generating manifest' for handle in self._storage_broker.iter_item_handles(): key = dtoolcore.utils.generate_identifier(handle) value = self._storage_broker.item_properties(handle) items[key] = value if progressbar: progressbar.item_show_func = lambda x: handle progressbar.update(1) manifest = {'items': items, 'dtoolcore_version': __version__, 'hash_function': self._storage_broker.hasher.name} return manifest
generate_manifest
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, progressbar=None):\n items = dict()\n if progressbar:\n progressbar.label = 'Generating manifest'\n for handle in self._storage_broker.iter_item_handles():\n key = dtoolcore.utils.generate_identifier(handle)\n value = self._storage_broker.item_properties(handle)\n items[key] = value\n if progressbar:\n progressbar.item_show_func = lambda x: handle\n progressbar.update(1)\n manifest = {'items': items, 'dtoolcore_version': __version__,\n 'hash_function': self._storage_broker.hasher.name}\n return manifest\n", "role": "user" }, { "content": "generate_manifest", "role": "assistant" } ]
def x(self): if self._manifest_cache is None: self._manifest_cache = self._storage_broker.get_manifest() return self._manifest_cache
manifest
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self._manifest_cache is None:\n self._manifest_cache = self._storage_broker.get_manifest()\n return self._manifest_cache\n", "role": "user" }, { "content": "manifest", "role": "assistant" } ]
def x(self): for handle in self._storage_broker.iter_item_handles(): yield dtoolcore.utils.generate_identifier(handle)
identifiers
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for handle in self._storage_broker.iter_item_handles():\n yield dtoolcore.utils.generate_identifier(handle)\n", "role": "user" }, { "content": "identifiers", "role": "assistant" } ]
def x(self): self._storage_broker.create_structure() self._storage_broker.put_admin_metadata(self._admin_metadata)
create
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self._storage_broker.create_structure()\n self._storage_broker.put_admin_metadata(self._admin_metadata)\n", "role": "user" }, { "content": "create", "role": "assistant" } ]
def x(self, handle, key, value): self._storage_broker.add_item_metadata(handle, key, value)
add_item_metadata
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, handle, key, value):\n self._storage_broker.add_item_metadata(handle, key, value)\n", "role": "user" }, { "content": "add_item_metadata", "role": "assistant" } ]
def x(self): overlays = defaultdict(dict) for handle in self._storage_broker.iter_item_handles(): identifier = dtoolcore.utils.generate_identifier(handle) item_metadata = self._storage_broker.get_item_metadata(handle) for k, v in item_metadata.items(): overlays[k][identifier] = v return overlays
generate_overlays
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n overlays = defaultdict(dict)\n for handle in self._storage_broker.iter_item_handles():\n identifier = dtoolcore.utils.generate_identifier(handle)\n item_metadata = self._storage_broker.get_item_metadata(handle)\n for k, v in item_metadata.items():\n overlays[k][identifier] = v\n return overlays\n", "role": "user" }, { "content": "generate_overlays", "role": "assistant" } ]
def x(self, progressbar=None): self._storage_broker.pre_freeze_hook() if progressbar: progressbar.label = 'Freezing dataset' manifest = self.generate_manifest(progressbar=progressbar) self._storage_broker.put_manifest(manifest) overlays = self._generate_overlays() for overlay_name, overlay in overlays.items(): self._put_overlay(overlay_name, overlay) datetime_obj = datetime.datetime.utcnow() metadata_update = {'type': 'dataset', 'frozen_at': dtoolcore.utils. timestamp(datetime_obj)} self._admin_metadata.update(metadata_update) self._storage_broker.put_admin_metadata(self._admin_metadata) self._storage_broker.post_freeze_hook()
freeze
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, progressbar=None):\n self._storage_broker.pre_freeze_hook()\n if progressbar:\n progressbar.label = 'Freezing dataset'\n manifest = self.generate_manifest(progressbar=progressbar)\n self._storage_broker.put_manifest(manifest)\n overlays = self._generate_overlays()\n for overlay_name, overlay in overlays.items():\n self._put_overlay(overlay_name, overlay)\n datetime_obj = datetime.datetime.utcnow()\n metadata_update = {'type': 'dataset', 'frozen_at': dtoolcore.utils.\n timestamp(datetime_obj)}\n self._admin_metadata.update(metadata_update)\n self._storage_broker.put_admin_metadata(self._admin_metadata)\n self._storage_broker.post_freeze_hook()\n", "role": "user" }, { "content": "freeze", "role": "assistant" } ]
def x(self, route_name_prefix, base_url, handler): route_name = route_name_prefix + '_read_many' self.add_handler(route_name, base_url, handler, action='read_many', request_method='GET') route_name = route_name_prefix + '_read_one' self.add_handler(route_name, base_url + '/{id}', handler, action= 'read_one', request_method='GET') route_name = route_name_prefix + '_count' self.add_handler(route_name, base_url + '/count', handler, action= 'count', request_method='GET') route_name = route_name_prefix + '_create' self.add_handler(route_name, base_url, handler, action='create', request_method='POST') route_name = route_name_prefix + '_update' self.add_handler(route_name, base_url + '/{id}', handler, action= 'update', request_method='PUT') route_name = route_name_prefix + '_delete' self.add_handler(route_name, base_url + '/{id}', handler, action= 'delete', request_method='DELETE')
add_papyrus_handler
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, route_name_prefix, base_url, handler):\n route_name = route_name_prefix + '_read_many'\n self.add_handler(route_name, base_url, handler, action='read_many',\n request_method='GET')\n route_name = route_name_prefix + '_read_one'\n self.add_handler(route_name, base_url + '/{id}', handler, action=\n 'read_one', request_method='GET')\n route_name = route_name_prefix + '_count'\n self.add_handler(route_name, base_url + '/count', handler, action=\n 'count', request_method='GET')\n route_name = route_name_prefix + '_create'\n self.add_handler(route_name, base_url, handler, action='create',\n request_method='POST')\n route_name = route_name_prefix + '_update'\n self.add_handler(route_name, base_url + '/{id}', handler, action=\n 'update', request_method='PUT')\n route_name = route_name_prefix + '_delete'\n self.add_handler(route_name, base_url + '/{id}', handler, action=\n 'delete', request_method='DELETE')\n", "role": "user" }, { "content": "add_papyrus_handler", "role": "assistant" } ]
def x(self, route_name_prefix, base_url): route_name = route_name_prefix + '_read_many' self.add_route(route_name, base_url, request_method='GET') route_name = route_name_prefix + '_read_one' self.add_route(route_name, base_url + '/{id}', request_method='GET') route_name = route_name_prefix + '_count' self.add_route(route_name, base_url + '/count', request_method='GET') route_name = route_name_prefix + '_create' self.add_route(route_name, base_url, request_method='POST') route_name = route_name_prefix + '_update' self.add_route(route_name, base_url + '/{id}', request_method='PUT') route_name = route_name_prefix + '_delete' self.add_route(route_name, base_url + '/{id}', request_method='DELETE')
add_papyrus_routes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, route_name_prefix, base_url):\n route_name = route_name_prefix + '_read_many'\n self.add_route(route_name, base_url, request_method='GET')\n route_name = route_name_prefix + '_read_one'\n self.add_route(route_name, base_url + '/{id}', request_method='GET')\n route_name = route_name_prefix + '_count'\n self.add_route(route_name, base_url + '/count', request_method='GET')\n route_name = route_name_prefix + '_create'\n self.add_route(route_name, base_url, request_method='POST')\n route_name = route_name_prefix + '_update'\n self.add_route(route_name, base_url + '/{id}', request_method='PUT')\n route_name = route_name_prefix + '_delete'\n self.add_route(route_name, base_url + '/{id}', request_method='DELETE')\n", "role": "user" }, { "content": "add_papyrus_routes", "role": "assistant" } ]
def x(f): @wraps(f) def x(t, **kw): def_vals = {'v': 0.0, 'h': 1.0, 'x': 0.0, 'w': 1.0, 's': 1.1, 'e': 1.0} for v in def_vals: if v not in kw: kw[v] = def_vals[v] anames, _, _, _ = inspect.getargspec(f) fkw = dict([(arg, kw[arg]) for arg in anames if arg in kw]) ta = t if 'x' not in anames: ta = ta - kw['x'] if 'w' not in anames: ta = ta / kw['w'] mod = f(ta, **fkw) mod = f(ta + ta[mod.argmax()], **fkw) return kw['v'] + kw['h'] / max(mod) * mod args = set(['v', 'h', 'x', 'w']) anames, _, _, _ = inspect.getargspec(f) wrapped_f._peakargs = list(args.union([a for a in anames if a not in ( 't', 'r')])) return wrapped_f
peak_model
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(f):\n\n @wraps(f)\n def x(t, **kw):\n def_vals = {'v': 0.0, 'h': 1.0, 'x': 0.0, 'w': 1.0, 's': 1.1, 'e': 1.0}\n for v in def_vals:\n if v not in kw:\n kw[v] = def_vals[v]\n anames, _, _, _ = inspect.getargspec(f)\n fkw = dict([(arg, kw[arg]) for arg in anames if arg in kw])\n ta = t\n if 'x' not in anames:\n ta = ta - kw['x']\n if 'w' not in anames:\n ta = ta / kw['w']\n mod = f(ta, **fkw)\n mod = f(ta + ta[mod.argmax()], **fkw)\n return kw['v'] + kw['h'] / max(mod) * mod\n args = set(['v', 'h', 'x', 'w'])\n anames, _, _, _ = inspect.getargspec(f)\n wrapped_f._peakargs = list(args.union([a for a in anames if a not in (\n 't', 'r')]))\n return wrapped_f\n", "role": "user" }, { "content": "peak_model", "role": "assistant" } ]
def x(match_downloaded_callback, on_exit_callback, conf, synchronize_callback=True): logger = logging.getLogger(__name__) if conf['logging_level'] != logging.NOTSET: logger.setLevel(conf['logging_level']) else: pass def x(players_to_analyze, analyzed_players, matches_to_download, downloaded_matches): logger.info('Reached the checkpoint.'.format(datetime.datetime.now( ).strftime('%m-%d %H:%M:%S'), len(downloaded_matches))) if on_exit_callback: on_exit_callback(players_to_analyze, analyzed_players, matches_to_download, downloaded_matches) players_to_analyze = set(conf['seed_players_id']) downloaded_matches = set(conf['downloaded_matches']) logger.info('{} previously downloaded matches'.format(len( downloaded_matches))) matches_to_download = set(conf['matches_to_download']) logger.info('{} matches to download'.format(len(matches_to_download))) analyzed_players = set() pta_lock = threading.Lock() players_available_condition = threading.Condition(pta_lock) mtd_lock = threading.Lock() matches_Available_condition = threading.Condition(mtd_lock) user_function_lock = threading.Lock( ) if synchronize_callback else NoOpContextManager() logger_lock = threading.Lock() player_downloader_threads = [] match_downloader_threads = [] try: def x(): if len(player_downloader_threads) < max_players_download_threads: player_downloader = PlayerDownloader(conf, players_to_analyze, analyzed_players, pta_lock, players_available_condition, matches_to_download, mtd_lock, matches_Available_condition, logger, logger_lock) player_downloader.start() player_downloader_threads.append(player_downloader) with logger_lock: logger.info( 'Adding a player download thread. Threads: ' + str( len(player_downloader_threads))) else: with logger_lock: logger.debug( 'Tried adding a player download thread, but there are already the maximum number: ' + str(max_players_download_threads)) def x(): if len(player_downloader_threads) > 1: player_downloader_threads.pop().shutdown() with logger_lock: logger.info( 'Removing a player downloader thread. Threads: ' + str(len(player_downloader_threads))) else: with logger_lock: logger.debug( 'Tried removing a player download thread, but there is only one left' ) logger.info('Starting fetching..') create_thread() for _ in range(matches_download_threads): match_downloader = MatchDownloader(conf, players_to_analyze, pta_lock, players_available_condition, matches_to_download, downloaded_matches, mtd_lock, matches_Available_condition, match_downloaded_callback, user_function_lock, logger, logger_lock) match_downloader.start() match_downloader_threads.append(match_downloader) auto_tuner = ThreadAutoTuner(create_thread, shutdown_thread) for i, _ in enumerate(do_every(1)): if conf.get('exit', False): break if i % 5 == 0: with mtd_lock: matches_in_queue = len(matches_to_download) total_players = sum(th.total_downloads for th in player_downloader_threads) auto_tuner.update_thread_number(total_players, matches_in_queue ) if i % logging_interval == 0: with mtd_lock: matches_in_queue = len(matches_to_download) total_matches = sum(th.total_downloads for th in match_downloader_threads) with pta_lock: players_in_queue = len(players_to_analyze) total_players = sum(th.total_downloads for th in player_downloader_threads) with logger_lock: logger.info( 'Players in queue: {}. Downloaded players: {}. Matches in queue: {}. Downloaded matches: {}' .format(players_in_queue, total_players, matches_in_queue, total_matches)) with pta_lock: players_available_condition.notify_all() with mtd_lock: matches_Available_condition.notify_all() logger.info('Terminating fetching') finally: conf['exit'] = True for thread in (player_downloader_threads + match_downloader_threads): thread.join() logger.info('Calling checkpoint callback') checkpoint(players_to_analyze, analyzed_players, matches_to_download, downloaded_matches)
download_matches
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(match_downloaded_callback, on_exit_callback, conf,\n synchronize_callback=True):\n logger = logging.getLogger(__name__)\n if conf['logging_level'] != logging.NOTSET:\n logger.setLevel(conf['logging_level'])\n else:\n pass\n\n def x(players_to_analyze, analyzed_players, matches_to_download,\n downloaded_matches):\n logger.info('Reached the checkpoint.'.format(datetime.datetime.now(\n ).strftime('%m-%d %H:%M:%S'), len(downloaded_matches)))\n if on_exit_callback:\n on_exit_callback(players_to_analyze, analyzed_players,\n matches_to_download, downloaded_matches)\n players_to_analyze = set(conf['seed_players_id'])\n downloaded_matches = set(conf['downloaded_matches'])\n logger.info('{} previously downloaded matches'.format(len(\n downloaded_matches)))\n matches_to_download = set(conf['matches_to_download'])\n logger.info('{} matches to download'.format(len(matches_to_download)))\n analyzed_players = set()\n pta_lock = threading.Lock()\n players_available_condition = threading.Condition(pta_lock)\n mtd_lock = threading.Lock()\n matches_Available_condition = threading.Condition(mtd_lock)\n user_function_lock = threading.Lock(\n ) if synchronize_callback else NoOpContextManager()\n logger_lock = threading.Lock()\n player_downloader_threads = []\n match_downloader_threads = []\n try:\n\n def x():\n if len(player_downloader_threads) < max_players_download_threads:\n player_downloader = PlayerDownloader(conf,\n players_to_analyze, analyzed_players, pta_lock,\n players_available_condition, matches_to_download,\n mtd_lock, matches_Available_condition, logger, logger_lock)\n player_downloader.start()\n player_downloader_threads.append(player_downloader)\n with logger_lock:\n logger.info(\n 'Adding a player download thread. Threads: ' + str(\n len(player_downloader_threads)))\n else:\n with logger_lock:\n logger.debug(\n 'Tried adding a player download thread, but there are already the maximum number: '\n + str(max_players_download_threads))\n\n def x():\n if len(player_downloader_threads) > 1:\n player_downloader_threads.pop().shutdown()\n with logger_lock:\n logger.info(\n 'Removing a player downloader thread. Threads: ' +\n str(len(player_downloader_threads)))\n else:\n with logger_lock:\n logger.debug(\n 'Tried removing a player download thread, but there is only one left'\n )\n logger.info('Starting fetching..')\n create_thread()\n for _ in range(matches_download_threads):\n match_downloader = MatchDownloader(conf, players_to_analyze,\n pta_lock, players_available_condition, matches_to_download,\n downloaded_matches, mtd_lock, matches_Available_condition,\n match_downloaded_callback, user_function_lock, logger,\n logger_lock)\n match_downloader.start()\n match_downloader_threads.append(match_downloader)\n auto_tuner = ThreadAutoTuner(create_thread, shutdown_thread)\n for i, _ in enumerate(do_every(1)):\n if conf.get('exit', False):\n break\n if i % 5 == 0:\n with mtd_lock:\n matches_in_queue = len(matches_to_download)\n total_players = sum(th.total_downloads for th in\n player_downloader_threads)\n auto_tuner.update_thread_number(total_players, matches_in_queue\n )\n if i % logging_interval == 0:\n with mtd_lock:\n matches_in_queue = len(matches_to_download)\n total_matches = sum(th.total_downloads for th in\n match_downloader_threads)\n with pta_lock:\n players_in_queue = len(players_to_analyze)\n total_players = sum(th.total_downloads for th in\n player_downloader_threads)\n with logger_lock:\n logger.info(\n 'Players in queue: {}. Downloaded players: {}. Matches in queue: {}. Downloaded matches: {}'\n .format(players_in_queue, total_players,\n matches_in_queue, total_matches))\n with pta_lock:\n players_available_condition.notify_all()\n with mtd_lock:\n matches_Available_condition.notify_all()\n logger.info('Terminating fetching')\n finally:\n conf['exit'] = True\n for thread in (player_downloader_threads + match_downloader_threads):\n thread.join()\n logger.info('Calling checkpoint callback')\n checkpoint(players_to_analyze, analyzed_players,\n matches_to_download, downloaded_matches)\n", "role": "user" }, { "content": "download_matches", "role": "assistant" } ]
def x(arr, value): arr = np.array(arr) idx = abs(arr - value).argmin() return arr[idx]
findNearest
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(arr, value):\n arr = np.array(arr)\n idx = abs(arr - value).argmin()\n return arr[idx]\n", "role": "user" }, { "content": "findNearest", "role": "assistant" } ]
def x(): magnitude_conversion_filepath = resource_stream(__name__, 'data/magnitude_conversion.dat') raw_table = np.loadtxt(magnitude_conversion_filepath, '|S5') magDict = {} for row in raw_table: if sys.hexversion >= 50331648: starClass = row[1].decode('utf-8') tableData = [x.decode('utf-8') for x in row[3:]] else: starClass = row[1] tableData = row[3:] magDict[starClass] = tableData return magDict
createMagConversionDict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n magnitude_conversion_filepath = resource_stream(__name__,\n 'data/magnitude_conversion.dat')\n raw_table = np.loadtxt(magnitude_conversion_filepath, '|S5')\n magDict = {}\n for row in raw_table:\n if sys.hexversion >= 50331648:\n starClass = row[1].decode('utf-8')\n tableData = [x.decode('utf-8') for x in row[3:]]\n else:\n starClass = row[1]\n tableData = row[3:]\n magDict[starClass] = tableData\n return magDict\n", "role": "user" }, { "content": "createMagConversionDict", "role": "assistant" } ]
def x(self, startClass, parentClass): try: if not startClass: raise AttributeError except AttributeError: raise HierarchyError('This object ({0}) has no {1} as a parent object' .format(self.name, parentClass)) if startClass.classType == parentClass: return startClass else: return self._getParentClass(startClass.parent, parentClass)
getParentClass
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, startClass, parentClass):\n try:\n if not startClass:\n raise AttributeError\n except AttributeError:\n raise HierarchyError('This object ({0}) has no {1} as a parent object'\n .format(self.name, parentClass))\n if startClass.classType == parentClass:\n return startClass\n else:\n return self._getParentClass(startClass.parent, parentClass)\n", "role": "user" }, { "content": "getParentClass", "role": "assistant" } ]
def x(self): paramTemp = self.getParam('temperature') if not paramTemp is np.nan: return paramTemp elif ed_params.estimateMissingValues: self.flags.addFlag('Calculated Temperature') return self.calcTemperature() else: return np.nan
T
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n paramTemp = self.getParam('temperature')\n if not paramTemp is np.nan:\n return paramTemp\n elif ed_params.estimateMissingValues:\n self.flags.addFlag('Calculated Temperature')\n return self.calcTemperature()\n else:\n return np.nan\n", "role": "user" }, { "content": "T", "role": "assistant" } ]
def x(self): d = self.parent.d if ed_params.estimateMissingValues: if d is np.nan: d = self.estimateDistance() if d is not np.nan: self.flags.addFlag('Estimated Distance') return d else: return np.nan
d
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n d = self.parent.d\n if ed_params.estimateMissingValues:\n if d is np.nan:\n d = self.estimateDistance()\n if d is not np.nan:\n self.flags.addFlag('Estimated Distance')\n return d\n else:\n return np.nan\n", "role": "user" }, { "content": "d", "role": "assistant" } ]
def x(self, mag_letter): allowed_mags = 'UBVJIHKLMN' catalogue_mags = 'BVIJHK' if mag_letter not in allowed_mags or not len(mag_letter) == 1: raise ValueError('Magnitude letter must be a single letter in {0}'. format(allowed_mags)) mag_str = 'mag' + mag_letter mag_val = self.getParam(mag_str) if isNanOrNone(mag_val) and ed_params.estimateMissingValues: mag_dict = dict(('mag' + letter, self.getParam('mag' + letter)) for letter in catalogue_mags) mag_class = Magnitude(self.spectralType, **mag_dict) try: mag_conversion = mag_class.convert(mag_letter) self.flags.addFlag('Estimated mag{0}'.format(mag_letter)) return mag_conversion except ValueError as e: logger.exception(e) return np.nan else: return mag_val
get_or_convert_magnitude
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, mag_letter):\n allowed_mags = 'UBVJIHKLMN'\n catalogue_mags = 'BVIJHK'\n if mag_letter not in allowed_mags or not len(mag_letter) == 1:\n raise ValueError('Magnitude letter must be a single letter in {0}'.\n format(allowed_mags))\n mag_str = 'mag' + mag_letter\n mag_val = self.getParam(mag_str)\n if isNanOrNone(mag_val) and ed_params.estimateMissingValues:\n mag_dict = dict(('mag' + letter, self.getParam('mag' + letter)) for\n letter in catalogue_mags)\n mag_class = Magnitude(self.spectralType, **mag_dict)\n try:\n mag_conversion = mag_class.convert(mag_letter)\n self.flags.addFlag('Estimated mag{0}'.format(mag_letter))\n return mag_conversion\n except ValueError as e:\n logger.exception(e)\n return np.nan\n else:\n return mag_val\n", "role": "user" }, { "content": "get_or_convert_magnitude", "role": "assistant" } ]
def x(self, wavelength=1.22): tempind = [3500.0, 3750.0, 4000.0, 4250.0, 4500.0, 4750.0, 5000.0, 5250.0, 5500.0, 5750.0, 6000.0, 6250.0, 6500.0, 6750.0, 7000.0, 7250.0, 7500.0, 7750.0, 8000.0, 8250.0, 8500.0, 8750.0, 9000.0, 9250.0, 9500.0, 9750.0, 10000.0, 10250.0, 10500.0, 10750.0, 11000.0, 11250.0, 11500.0, 11750.0, 12000.0, 12250.0, 12500.0, 12750.0, 13000.0, 14000.0, 15000.0, 16000.0, 17000.0, 19000.0, 20000.0, 21000.0, 22000.0, 23000.0, 24000.0, 25000.0, 26000.0, 27000.0, 28000.0, 29000.0, 30000.0, 31000.0, 32000.0, 33000.0, 34000.0, 35000.0, 36000.0, 37000.0, 38000.0, 39000.0, 40000.0, 41000.0, 42000.0, 43000.0, 44000.0, 45000.0, 46000.0, 47000.0, 48000.0, 49000.0, 50000.0] lggind = [0.0, 0.5, 1.0, 1.5, 2.0, 2.5, 3.0, 3.5, 4.0, 4.5, 5.0] mhind = [-5.0, -4.5, -4.0, -3.5, -3.0, -2.5, -2.0, -1.5, -1.0, -0.5, - 0.3, -0.2, -0.1, 0.0, 0.1, 0.2, 0.3, 0.5, 1.0] tempselect = _findNearest(tempind, float(self.T)) lgselect = _findNearest(lggind, float(self.calcLogg())) mhselect = _findNearest(mhind, float(self.Z)) quadratic_filepath = resource_stream(__name__, 'data/quadratic.dat') coeffTable = np.loadtxt(quadratic_filepath) foundValues = False for i in range(len(coeffTable)): if coeffTable[i, 2] == lgselect and coeffTable[i, 3 ] == tempselect and coeffTable[i, 4] == mhselect: if coeffTable[i, 0] == 1: u1array = coeffTable[i, 8:] u2array = coeffTable[i + 1, 8:] foundValues = True break if not foundValues: raise ValueError('No limb darkening values could be found') waveind = [0.365, 0.445, 0.551, 0.658, 0.806, 1.22, 1.63, 2.19, 3.45] u1AtWavelength = np.interp(wavelength, waveind, u1array, left=0, right=0) u2AtWavelength = np.interp(wavelength, waveind, u2array, left=0, right=0) return u1AtWavelength, u2AtWavelength
getLimbdarkeningCoeff
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, wavelength=1.22):\n tempind = [3500.0, 3750.0, 4000.0, 4250.0, 4500.0, 4750.0, 5000.0, \n 5250.0, 5500.0, 5750.0, 6000.0, 6250.0, 6500.0, 6750.0, 7000.0, \n 7250.0, 7500.0, 7750.0, 8000.0, 8250.0, 8500.0, 8750.0, 9000.0, \n 9250.0, 9500.0, 9750.0, 10000.0, 10250.0, 10500.0, 10750.0, 11000.0,\n 11250.0, 11500.0, 11750.0, 12000.0, 12250.0, 12500.0, 12750.0, \n 13000.0, 14000.0, 15000.0, 16000.0, 17000.0, 19000.0, 20000.0, \n 21000.0, 22000.0, 23000.0, 24000.0, 25000.0, 26000.0, 27000.0, \n 28000.0, 29000.0, 30000.0, 31000.0, 32000.0, 33000.0, 34000.0, \n 35000.0, 36000.0, 37000.0, 38000.0, 39000.0, 40000.0, 41000.0, \n 42000.0, 43000.0, 44000.0, 45000.0, 46000.0, 47000.0, 48000.0, \n 49000.0, 50000.0]\n lggind = [0.0, 0.5, 1.0, 1.5, 2.0, 2.5, 3.0, 3.5, 4.0, 4.5, 5.0]\n mhind = [-5.0, -4.5, -4.0, -3.5, -3.0, -2.5, -2.0, -1.5, -1.0, -0.5, -\n 0.3, -0.2, -0.1, 0.0, 0.1, 0.2, 0.3, 0.5, 1.0]\n tempselect = _findNearest(tempind, float(self.T))\n lgselect = _findNearest(lggind, float(self.calcLogg()))\n mhselect = _findNearest(mhind, float(self.Z))\n quadratic_filepath = resource_stream(__name__, 'data/quadratic.dat')\n coeffTable = np.loadtxt(quadratic_filepath)\n foundValues = False\n for i in range(len(coeffTable)):\n if coeffTable[i, 2] == lgselect and coeffTable[i, 3\n ] == tempselect and coeffTable[i, 4] == mhselect:\n if coeffTable[i, 0] == 1:\n u1array = coeffTable[i, 8:]\n u2array = coeffTable[i + 1, 8:]\n foundValues = True\n break\n if not foundValues:\n raise ValueError('No limb darkening values could be found')\n waveind = [0.365, 0.445, 0.551, 0.658, 0.806, 1.22, 1.63, 2.19, 3.45]\n u1AtWavelength = np.interp(wavelength, waveind, u1array, left=0, right=0)\n u2AtWavelength = np.interp(wavelength, waveind, u2array, left=0, right=0)\n return u1AtWavelength, u2AtWavelength\n", "role": "user" }, { "content": "getLimbdarkeningCoeff", "role": "assistant" } ]
def x(self): try: isTransiting = self.params['istransiting'] except KeyError: return False if isTransiting == '1': return True else: return False
isTransiting
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n try:\n isTransiting = self.params['istransiting']\n except KeyError:\n return False\n if isTransiting == '1':\n return True\n else:\n return False\n", "role": "user" }, { "content": "isTransiting", "role": "assistant" } ]
def x(self, circular=False): try: if circular: return eq.transitDurationCircular(self.P, self.star.R, self.R, self.a, self.i) else: return eq.TransitDuration(self.P, self.a, self.R, self.star.R, self.i, self.e, self.periastron).Td except (ValueError, AttributeError, HierarchyError): return np.nan
calcTransitDuration
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, circular=False):\n try:\n if circular:\n return eq.transitDurationCircular(self.P, self.star.R, self.R,\n self.a, self.i)\n else:\n return eq.TransitDuration(self.P, self.a, self.R, self.star.R,\n self.i, self.e, self.periastron).Td\n except (ValueError, AttributeError, HierarchyError):\n return np.nan\n", "role": "user" }, { "content": "calcTransitDuration", "role": "assistant" } ]
def x(self): try: return eq.MeanPlanetTemp(self.albedo, self.star.T, self.star.R, self.a ).T_p except (ValueError, HierarchyError): return np.nan
calcTemperature
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n try:\n return eq.MeanPlanetTemp(self.albedo, self.star.T, self.star.R, self.a\n ).T_p\n except (ValueError, HierarchyError):\n return np.nan\n", "role": "user" }, { "content": "calcTemperature", "role": "assistant" } ]
def x(self): try: return eq.KeplersThirdLaw(None, self.star.M, self.P).a except HierarchyError: return np.nan
calcSMA
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n try:\n return eq.KeplersThirdLaw(None, self.star.M, self.P).a\n except HierarchyError:\n return np.nan\n", "role": "user" }, { "content": "calcSMA", "role": "assistant" } ]