source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(self): data = get_data(self.endpoint, self.id_, force_lookup=self.__force_lookup) for key, val in data.items(): if key == 'location_area_encounters' and self.endpoint == 'pokemon': params = val.split('/')[-3:] ep, id_, subr = params encounters = get_data(ep, int(id_), subr) data[key] = [_make_obj(enc) for enc in encounters] continue if isinstance(val, dict): data[key] = _make_obj(val) elif isinstance(val, list): data[key] = [_make_obj(i) for i in val] self.__dict__.update(data) return None
load
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n data = get_data(self.endpoint, self.id_, force_lookup=self.__force_lookup)\n for key, val in data.items():\n if key == 'location_area_encounters' and self.endpoint == 'pokemon':\n params = val.split('/')[-3:]\n ep, id_, subr = params\n encounters = get_data(ep, int(id_), subr)\n data[key] = [_make_obj(enc) for enc in encounters]\n continue\n if isinstance(val, dict):\n data[key] = _make_obj(val)\n elif isinstance(val, list):\n data[key] = [_make_obj(i) for i in val]\n self.__dict__.update(data)\n return None\n", "role": "user" }, { "content": "load", "role": "assistant" } ]
def x(path, mode=511): try: os.makedirs(path, mode) except OSError as error: if error.errno != 17: raise return path
safe_make_dirs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path, mode=511):\n try:\n os.makedirs(path, mode)\n except OSError as error:\n if error.errno != 17:\n raise\n return path\n", "role": "user" }, { "content": "safe_make_dirs", "role": "assistant" } ]
def x(): xdg_cache_home = os.environ.get('XDG_CACHE_HOME') or os.path.join(os. path.expanduser('~'), '.cache') return os.path.join(xdg_cache_home, 'pokebase')
get_default_cache
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n xdg_cache_home = os.environ.get('XDG_CACHE_HOME') or os.path.join(os.\n path.expanduser('~'), '.cache')\n return os.path.join(xdg_cache_home, 'pokebase')\n", "role": "user" }, { "content": "get_default_cache", "role": "assistant" } ]
def x(new_path=None): global CACHE_DIR, API_CACHE, SPRITE_CACHE if new_path is None: new_path = get_default_cache() CACHE_DIR = safe_make_dirs(os.path.abspath(new_path)) API_CACHE = os.path.join(CACHE_DIR, 'api.cache') SPRITE_CACHE = safe_make_dirs(os.path.join(CACHE_DIR, 'sprite')) return CACHE_DIR, API_CACHE, SPRITE_CACHE
set_cache
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(new_path=None):\n global CACHE_DIR, API_CACHE, SPRITE_CACHE\n if new_path is None:\n new_path = get_default_cache()\n CACHE_DIR = safe_make_dirs(os.path.abspath(new_path))\n API_CACHE = os.path.join(CACHE_DIR, 'api.cache')\n SPRITE_CACHE = safe_make_dirs(os.path.join(CACHE_DIR, 'sprite'))\n return CACHE_DIR, API_CACHE, SPRITE_CACHE\n", "role": "user" }, { "content": "set_cache", "role": "assistant" } ]
def x(self, lun_or_snap, skip_hlu_0=False): try: return self._attach_with_retry(lun_or_snap, skip_hlu_0) except ex.SystemAPINotSupported: raise except ex.UnityAttachExceedLimitError: raise except: self.detach(lun_or_snap) raise
attach
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, lun_or_snap, skip_hlu_0=False):\n try:\n return self._attach_with_retry(lun_or_snap, skip_hlu_0)\n except ex.SystemAPINotSupported:\n raise\n except ex.UnityAttachExceedLimitError:\n raise\n except:\n self.detach(lun_or_snap)\n raise\n", "role": "user" }, { "content": "attach", "role": "assistant" } ]
def x(self, lun_or_snap, cg_member=None): hlu = self.get_hlu(lun_or_snap, cg_member=cg_member) return hlu is not None
has_hlu
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, lun_or_snap, cg_member=None):\n hlu = self.get_hlu(lun_or_snap, cg_member=cg_member)\n return hlu is not None\n", "role": "user" }, { "content": "has_hlu", "role": "assistant" } ]
def x(self, lun_or_snap, cg_member=None): import storops.unity.resource.lun as lun_module import storops.unity.resource.snap as snap_module which = None if isinstance(lun_or_snap, lun_module.UnityLun): which = self._get_host_luns(lun=lun_or_snap) elif isinstance(lun_or_snap, snap_module.UnitySnap): if lun_or_snap.is_cg_snap(): if cg_member is None: log.debug( 'None host lun for CG snap {}. Use its member snap instead or pass in cg_member.' .format(lun_or_snap.id)) return None lun_or_snap = lun_or_snap.get_member_snap(cg_member) which = self._get_host_luns(lun=cg_member, snap=lun_or_snap) else: which = self._get_host_luns(snap=lun_or_snap) if not which: log.debug('Resource(LUN or Snap) {} is not attached to host {}'. format(lun_or_snap.name, self.name)) return None return which[0]
get_host_lun
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, lun_or_snap, cg_member=None):\n import storops.unity.resource.lun as lun_module\n import storops.unity.resource.snap as snap_module\n which = None\n if isinstance(lun_or_snap, lun_module.UnityLun):\n which = self._get_host_luns(lun=lun_or_snap)\n elif isinstance(lun_or_snap, snap_module.UnitySnap):\n if lun_or_snap.is_cg_snap():\n if cg_member is None:\n log.debug(\n 'None host lun for CG snap {}. Use its member snap instead or pass in cg_member.'\n .format(lun_or_snap.id))\n return None\n lun_or_snap = lun_or_snap.get_member_snap(cg_member)\n which = self._get_host_luns(lun=cg_member, snap=lun_or_snap)\n else:\n which = self._get_host_luns(snap=lun_or_snap)\n if not which:\n log.debug('Resource(LUN or Snap) {} is not attached to host {}'.\n format(lun_or_snap.name, self.name))\n return None\n return which[0]\n", "role": "user" }, { "content": "get_host_lun", "role": "assistant" } ]
def x(self, resource, cg_member=None): host_lun = self.get_host_lun(resource, cg_member=cg_member) return host_lun if host_lun is None else host_lun.hlu
get_hlu
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, resource, cg_member=None):\n host_lun = self.get_host_lun(resource, cg_member=cg_member)\n return host_lun if host_lun is None else host_lun.hlu\n", "role": "user" }, { "content": "get_hlu", "role": "assistant" } ]
def x(self, iqns=None, wwns=None): iqns = set(iqns) if iqns else set() current_iqns = set() if self.iscsi_host_initiators: current_iqns = {initiator.initiator_id for initiator in self. iscsi_host_initiators} wwns = set(wwns) if wwns else set() current_wwns = set() if self.fc_host_initiators: current_wwns = {initiator.initiator_id for initiator in self. fc_host_initiators} updater = UnityHostInitiatorUpdater(self, current_iqns | current_wwns, iqns | wwns) return updater.update()
update_initiators
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, iqns=None, wwns=None):\n iqns = set(iqns) if iqns else set()\n current_iqns = set()\n if self.iscsi_host_initiators:\n current_iqns = {initiator.initiator_id for initiator in self.\n iscsi_host_initiators}\n wwns = set(wwns) if wwns else set()\n current_wwns = set()\n if self.fc_host_initiators:\n current_wwns = {initiator.initiator_id for initiator in self.\n fc_host_initiators}\n updater = UnityHostInitiatorUpdater(self, current_iqns | current_wwns, \n iqns | wwns)\n return updater.update()\n", "role": "user" }, { "content": "update_initiators", "role": "assistant" } ]
def x(tag=None, sat_id=None, data_path=None, format_str=None, supported_tags=None, fake_daily_files_from_monthly=False, two_digit_year_break=None): if data_path is not None: if format_str is None: try: format_str = supported_tags[sat_id][tag] except KeyError: raise ValueError('Unknown tag') out = pysat.Files.from_os(data_path=data_path, format_str=format_str) if not out.empty and fake_daily_files_from_monthly: out.ix[out.index[-1] + pds.DateOffset(months=1) - pds. DateOffset(days=1)] = out.iloc[-1] out = out.asfreq('D', 'pad') out = out + '_' + out.index.strftime('%Y-%m-%d') return out return out else: estr = ( 'A directory must be passed to the loading routine for <Instrument Code>' ) raise ValueError(estr)
list_files
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(tag=None, sat_id=None, data_path=None, format_str=None,\n supported_tags=None, fake_daily_files_from_monthly=False,\n two_digit_year_break=None):\n if data_path is not None:\n if format_str is None:\n try:\n format_str = supported_tags[sat_id][tag]\n except KeyError:\n raise ValueError('Unknown tag')\n out = pysat.Files.from_os(data_path=data_path, format_str=format_str)\n if not out.empty and fake_daily_files_from_monthly:\n out.ix[out.index[-1] + pds.DateOffset(months=1) - pds.\n DateOffset(days=1)] = out.iloc[-1]\n out = out.asfreq('D', 'pad')\n out = out + '_' + out.index.strftime('%Y-%m-%d')\n return out\n return out\n else:\n estr = (\n 'A directory must be passed to the loading routine for <Instrument Code>'\n )\n raise ValueError(estr)\n", "role": "user" }, { "content": "list_files", "role": "assistant" } ]
def x(fnames, tag=None, sat_id=None, fake_daily_files_from_monthly=False, flatten_twod=True): import pysatCDF if len(fnames) <= 0: return pysat.DataFrame(None), None elif fake_daily_files_from_monthly: fname = fnames[0][0:-11] date = pysat.datetime.strptime(fnames[0][-10:], '%Y-%m-%d') with pysatCDF.CDF(fname) as cdf: data, meta = cdf.to_pysat(flatten_twod=flatten_twod) data = data.ix[date:date + pds.DateOffset(days=1) - pds. DateOffset(microseconds=1), :] return data, meta else: with pysatCDF.CDF(fnames[0]) as cdf: return cdf.to_pysat(flatten_twod=flatten_twod)
load
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fnames, tag=None, sat_id=None, fake_daily_files_from_monthly=False,\n flatten_twod=True):\n import pysatCDF\n if len(fnames) <= 0:\n return pysat.DataFrame(None), None\n elif fake_daily_files_from_monthly:\n fname = fnames[0][0:-11]\n date = pysat.datetime.strptime(fnames[0][-10:], '%Y-%m-%d')\n with pysatCDF.CDF(fname) as cdf:\n data, meta = cdf.to_pysat(flatten_twod=flatten_twod)\n data = data.ix[date:date + pds.DateOffset(days=1) - pds.\n DateOffset(microseconds=1), :]\n return data, meta\n else:\n with pysatCDF.CDF(fnames[0]) as cdf:\n return cdf.to_pysat(flatten_twod=flatten_twod)\n", "role": "user" }, { "content": "load", "role": "assistant" } ]
def x(supported_tags, date_array, tag, sat_id, ftp_site= 'cdaweb.gsfc.nasa.gov', data_path=None, user=None, password=None, fake_daily_files_from_monthly=False): import os import ftplib ftp = ftplib.FTP(ftp_site) ftp.login() try: ftp_dict = supported_tags[tag] except KeyError: raise ValueError('Tag name unknown.') ftp.cwd(ftp_dict['dir']) remote_fname = ftp_dict['remote_fname'] local_fname = ftp_dict['local_fname'] for date in date_array: formatted_remote_fname = remote_fname.format(year=date.year, month= date.month, day=date.day) formatted_local_fname = local_fname.format(year=date.year, month= date.month, day=date.day) saved_local_fname = os.path.join(data_path, formatted_local_fname) try: print('Attempting to download file for ' + date.strftime('%x')) sys.stdout.flush() ftp.retrbinary('RETR ' + formatted_remote_fname, open( saved_local_fname, 'wb').write) print('Finished.') except ftplib.error_perm as exception: if str(exception.args[0]).split(' ', 1)[0] != '550': raise else: os.remove(saved_local_fname) print('File not available for ' + date.strftime('%x')) ftp.close()
download
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(supported_tags, date_array, tag, sat_id, ftp_site=\n 'cdaweb.gsfc.nasa.gov', data_path=None, user=None, password=None,\n fake_daily_files_from_monthly=False):\n import os\n import ftplib\n ftp = ftplib.FTP(ftp_site)\n ftp.login()\n try:\n ftp_dict = supported_tags[tag]\n except KeyError:\n raise ValueError('Tag name unknown.')\n ftp.cwd(ftp_dict['dir'])\n remote_fname = ftp_dict['remote_fname']\n local_fname = ftp_dict['local_fname']\n for date in date_array:\n formatted_remote_fname = remote_fname.format(year=date.year, month=\n date.month, day=date.day)\n formatted_local_fname = local_fname.format(year=date.year, month=\n date.month, day=date.day)\n saved_local_fname = os.path.join(data_path, formatted_local_fname)\n try:\n print('Attempting to download file for ' + date.strftime('%x'))\n sys.stdout.flush()\n ftp.retrbinary('RETR ' + formatted_remote_fname, open(\n saved_local_fname, 'wb').write)\n print('Finished.')\n except ftplib.error_perm as exception:\n if str(exception.args[0]).split(' ', 1)[0] != '550':\n raise\n else:\n os.remove(saved_local_fname)\n print('File not available for ' + date.strftime('%x'))\n ftp.close()\n", "role": "user" }, { "content": "download", "role": "assistant" } ]
def x(response, method, url): req_id = response.headers.get('x-openstack-request-id') if not req_id: req_id = response.headers.get('x-compute-request-id') kwargs = {'http_status': response.status_code, 'response': response, 'method': method, 'url': url, 'request_id': req_id} if 'retry-after' in response.headers: kwargs['retry_after'] = response.headers['retry-after'] content_type = response.headers.get('Content-Type', '') if content_type.startswith('application/json'): try: body = response.json() except ValueError: pass else: if isinstance(body, dict): error = body.get(list(body)[0]) if isinstance(error, dict): kwargs['message'] = error.get('message') or error.get( 'faultstring') kwargs['details'] = error.get('details') or six.text_type( body) elif content_type.startswith('text/'): kwargs['details'] = response.text try: cls = _code_map[response.status_code] except KeyError: if 500 <= response.status_code < 600: cls = HttpServerError elif 400 <= response.status_code < 500: cls = HTTPClientError else: cls = HttpError return cls(**kwargs)
from_response
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(response, method, url):\n req_id = response.headers.get('x-openstack-request-id')\n if not req_id:\n req_id = response.headers.get('x-compute-request-id')\n kwargs = {'http_status': response.status_code, 'response': response,\n 'method': method, 'url': url, 'request_id': req_id}\n if 'retry-after' in response.headers:\n kwargs['retry_after'] = response.headers['retry-after']\n content_type = response.headers.get('Content-Type', '')\n if content_type.startswith('application/json'):\n try:\n body = response.json()\n except ValueError:\n pass\n else:\n if isinstance(body, dict):\n error = body.get(list(body)[0])\n if isinstance(error, dict):\n kwargs['message'] = error.get('message') or error.get(\n 'faultstring')\n kwargs['details'] = error.get('details') or six.text_type(\n body)\n elif content_type.startswith('text/'):\n kwargs['details'] = response.text\n try:\n cls = _code_map[response.status_code]\n except KeyError:\n if 500 <= response.status_code < 600:\n cls = HttpServerError\n elif 400 <= response.status_code < 500:\n cls = HTTPClientError\n else:\n cls = HttpError\n return cls(**kwargs)\n", "role": "user" }, { "content": "from_response", "role": "assistant" } ]
def x(self, name, raid_groups, description=None, **kwargs): return UnityPool.create(self._cli, name=name, description=description, raid_groups=raid_groups, **kwargs)
create_pool
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, name, raid_groups, description=None, **kwargs):\n return UnityPool.create(self._cli, name=name, description=description,\n raid_groups=raid_groups, **kwargs)\n", "role": "user" }, { "content": "create_pool", "role": "assistant" } ]
def x(self): eths = self.get_ethernet_port(bond=False) las = self.get_link_aggregation() return eths + las
get_file_port
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n eths = self.get_ethernet_port(bond=False)\n las = self.get_link_aggregation()\n return eths + las\n", "role": "user" }, { "content": "get_file_port", "role": "assistant" } ]
def x(self, management_address, local_username=None, local_password=None, remote_username=None, remote_password=None, connection_type=None): return UnityRemoteSystem.create(self._cli, management_address, local_username=local_username, local_password=local_password, remote_username=remote_username, remote_password=remote_password, connection_type=connection_type)
create_remote_system
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, management_address, local_username=None, local_password=None,\n remote_username=None, remote_password=None, connection_type=None):\n return UnityRemoteSystem.create(self._cli, management_address,\n local_username=local_username, local_password=local_password,\n remote_username=remote_username, remote_password=remote_password,\n connection_type=connection_type)\n", "role": "user" }, { "content": "create_remote_system", "role": "assistant" } ]
def x(self, sp, ip_port, ip_address, netmask=None, v6_prefix_length=None, gateway=None, vlan_id=None): return UnityReplicationInterface.create(self._cli, sp, ip_port, ip_address, netmask=netmask, v6_prefix_length=v6_prefix_length, gateway=gateway, vlan_id=vlan_id)
create_replication_interface
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, sp, ip_port, ip_address, netmask=None, v6_prefix_length=None,\n gateway=None, vlan_id=None):\n return UnityReplicationInterface.create(self._cli, sp, ip_port,\n ip_address, netmask=netmask, v6_prefix_length=v6_prefix_length,\n gateway=gateway, vlan_id=vlan_id)\n", "role": "user" }, { "content": "create_replication_interface", "role": "assistant" } ]
def x(self, res_id, backup_snap=None): name = self._get_name() out = self._cli.restore_snap(name, res_id, backup_snap) ex.raise_if_err(out, 'failed to restore snap {}.'.format(name), default =ex.VNXSnapError)
restore
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, res_id, backup_snap=None):\n name = self._get_name()\n out = self._cli.restore_snap(name, res_id, backup_snap)\n ex.raise_if_err(out, 'failed to restore snap {}.'.format(name), default\n =ex.VNXSnapError)\n", "role": "user" }, { "content": "restore", "role": "assistant" } ]
def x(tag='', sat_id=None, data_path=None, format_str=None): if format_str is None and data_path is not None: file_base = 'supermag_magnetometer' if tag == 'indices' or tag == 'all': file_base += '_all' if tag == 'indices': psplit = path.split(data_path[:-1]) data_path = path.join(psplit[0], 'all', '') if tag == 'stations': min_fmt = '_'.join([file_base, '{year:4d}.???']) doff = pds.DateOffset(years=1) else: min_fmt = '_'.join([file_base, '{year:4d}{month:02d}{day:02d}.???'] ) doff = pds.DateOffset(days=1) files = pysat.Files.from_os(data_path=data_path, format_str=min_fmt) if not files.empty: files = files.sort_index() if tag == 'stations': orig_files = files.copy() new_files = [] for orig in orig_files.iteritems(): files.ix[orig[0] + doff - pds.DateOffset(days=1)] = orig[1] files = files.sort_index() new_files.append(files.ix[orig[0]:orig[0] + doff - pds. DateOffset(days=1)].asfreq('D', method='pad')) files = pds.concat(new_files) files = files.dropna() files = files.sort_index() files = files + '_' + files.index.strftime('%Y-%m-%d') return files elif format_str is None: estr = 'A directory must be passed to the loading routine for SuperMAG' raise ValueError(estr) else: return pysat.Files.from_os(data_path=data_path, format_str=format_str)
list_files
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(tag='', sat_id=None, data_path=None, format_str=None):\n if format_str is None and data_path is not None:\n file_base = 'supermag_magnetometer'\n if tag == 'indices' or tag == 'all':\n file_base += '_all'\n if tag == 'indices':\n psplit = path.split(data_path[:-1])\n data_path = path.join(psplit[0], 'all', '')\n if tag == 'stations':\n min_fmt = '_'.join([file_base, '{year:4d}.???'])\n doff = pds.DateOffset(years=1)\n else:\n min_fmt = '_'.join([file_base, '{year:4d}{month:02d}{day:02d}.???']\n )\n doff = pds.DateOffset(days=1)\n files = pysat.Files.from_os(data_path=data_path, format_str=min_fmt)\n if not files.empty:\n files = files.sort_index()\n if tag == 'stations':\n orig_files = files.copy()\n new_files = []\n for orig in orig_files.iteritems():\n files.ix[orig[0] + doff - pds.DateOffset(days=1)] = orig[1]\n files = files.sort_index()\n new_files.append(files.ix[orig[0]:orig[0] + doff - pds.\n DateOffset(days=1)].asfreq('D', method='pad'))\n files = pds.concat(new_files)\n files = files.dropna()\n files = files.sort_index()\n files = files + '_' + files.index.strftime('%Y-%m-%d')\n return files\n elif format_str is None:\n estr = 'A directory must be passed to the loading routine for SuperMAG'\n raise ValueError(estr)\n else:\n return pysat.Files.from_os(data_path=data_path, format_str=format_str)\n", "role": "user" }, { "content": "list_files", "role": "assistant" } ]
def x(fnames, tag='', sat_id=None): if len(fnames) <= 0: return pysat.DataFrame(None), pysat.Meta(None) if isinstance(fnames, str): fnames = [fnames] data = pds.DataFrame() baseline = list() for fname in fnames: fname = fname[:-11] file_type = path.splitext(fname)[1].lower() if file_type == '.csv': if tag != 'indices': temp = load_csv_data(fname, tag) else: temp, bline = load_ascii_data(fname, tag) if bline is not None: baseline.append(bline) if len(temp.columns) > 0: data = pds.concat([data, temp], axis=0) del temp if len(data.columns) > 0: meta = pysat.Meta() for cc in data.columns: meta[cc] = update_smag_metadata(cc) meta.info = {'baseline': format_baseline_list(baseline)} else: meta = pysat.Meta(None) return data, meta
load
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fnames, tag='', sat_id=None):\n if len(fnames) <= 0:\n return pysat.DataFrame(None), pysat.Meta(None)\n if isinstance(fnames, str):\n fnames = [fnames]\n data = pds.DataFrame()\n baseline = list()\n for fname in fnames:\n fname = fname[:-11]\n file_type = path.splitext(fname)[1].lower()\n if file_type == '.csv':\n if tag != 'indices':\n temp = load_csv_data(fname, tag)\n else:\n temp, bline = load_ascii_data(fname, tag)\n if bline is not None:\n baseline.append(bline)\n if len(temp.columns) > 0:\n data = pds.concat([data, temp], axis=0)\n del temp\n if len(data.columns) > 0:\n meta = pysat.Meta()\n for cc in data.columns:\n meta[cc] = update_smag_metadata(cc)\n meta.info = {'baseline': format_baseline_list(baseline)}\n else:\n meta = pysat.Meta(None)\n return data, meta\n", "role": "user" }, { "content": "load", "role": "assistant" } ]
def x(fname, tag): import re if tag == 'stations': ddict = dict() dkeys = list() date_list = list() with open(fname, 'r') as fopen: dtime = pds.datetime.strptime(fname.split('_')[-1].split('.')[0 ], '%Y') for fline in fopen.readlines(): sline = [ll for ll in re.split('[,\\n]+', fline) if len(ll) > 0 ] if len(ddict.items()) == 0: for kk in sline: kk = re.sub('-', '_', kk) ddict[kk] = list() dkeys.append(kk) else: date_list.append(dtime) for i, ll in enumerate(sline): if i >= 1 and i <= 4: ddict[dkeys[i]].append(float(ll)) elif i == 6: ddict[dkeys[i]].append(int(ll)) elif i < len(dkeys): ddict[dkeys[i]].append(ll) else: ddict[dkeys[-1]][-1] += ' {:s}'.format(ll) data = pds.DataFrame(ddict, index=date_list, columns=ddict.keys()) else: def x(dd): return pysat.datetime.strptime(dd, '%Y-%m-%d %H:%M:%S') data = pds.read_csv(fname, parse_dates={'datetime': [0]}, date_parser=parse_smag_date, index_col='datetime') return data
load_csv_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fname, tag):\n import re\n if tag == 'stations':\n ddict = dict()\n dkeys = list()\n date_list = list()\n with open(fname, 'r') as fopen:\n dtime = pds.datetime.strptime(fname.split('_')[-1].split('.')[0\n ], '%Y')\n for fline in fopen.readlines():\n sline = [ll for ll in re.split('[,\\\\n]+', fline) if len(ll) > 0\n ]\n if len(ddict.items()) == 0:\n for kk in sline:\n kk = re.sub('-', '_', kk)\n ddict[kk] = list()\n dkeys.append(kk)\n else:\n date_list.append(dtime)\n for i, ll in enumerate(sline):\n if i >= 1 and i <= 4:\n ddict[dkeys[i]].append(float(ll))\n elif i == 6:\n ddict[dkeys[i]].append(int(ll))\n elif i < len(dkeys):\n ddict[dkeys[i]].append(ll)\n else:\n ddict[dkeys[-1]][-1] += ' {:s}'.format(ll)\n data = pds.DataFrame(ddict, index=date_list, columns=ddict.keys())\n else:\n\n def x(dd):\n return pysat.datetime.strptime(dd, '%Y-%m-%d %H:%M:%S')\n data = pds.read_csv(fname, parse_dates={'datetime': [0]},\n date_parser=parse_smag_date, index_col='datetime')\n return data\n", "role": "user" }, { "content": "load_csv_data", "role": "assistant" } ]
def x(fname, tag): import re ndata = {'indices': 2, '': 4, 'all': 4, 'stations': 8} dkeys = {'stations': list(), '': ['IAGA', 'N', 'E', 'Z']} data = pds.DataFrame(None) baseline = None if not tag in ndata.keys(): return data, baseline with open(fname, 'r') as fopen: hflag = True pflag = False dflag = False if tag == 'stations' else True snum = 0 ddict = dict() date_list = list() if tag == 'stations': dtime = pds.datetime.strptime(fname.split('_')[-1].split('.')[0 ], '%Y') for fline in fopen.readlines(): line_len = len(fline) if hflag: if pflag: pflag = False if fline.find('-mlt') > 0: ndata[''] += 2 dkeys[''].extend(['MLT', 'MLAT']) if fline.find('-sza') > 0: ndata[''] += 1 dkeys[''].append('SZA') if fline.find('-decl') > 0: ndata[''] += 1 dkeys[''].append('IGRF_DECL') if tag == 'indices' and fline.find('-envelope') < 0: break lsplit = fline.split() idelta = lsplit.index('-delta') + 1 ibase = lsplit.index('-baseline') + 1 isd = lsplit.index('-sd') + 1 ist = lsplit.index('-st') + 1 iex = lsplit.index('-ex') + 1 baseline = ' '.join([lsplit[ibase], lsplit[idelta], lsplit[isd], lsplit[ist], lsplit[iex]]) if fline.find('Selected parameters:') >= 0: pflag = True if fline.count('=') == line_len - 1 and line_len > 2: hflag = False else: lsplit = [ll for ll in re.split('[\\t\\n]+', fline) if len( ll) > 0] if dflag: dflag = False dstring = ' '.join(lsplit[:6]) dtime = pysat.datetime.strptime(dstring, '%Y %m %d %H %M %S') snum = int(lsplit[6]) if tag == 'indices': date_list.append(dtime) else: date_list.extend([dtime for i in range(snum)]) elif len(lsplit) == ndata['indices']: if tag is not '': if lsplit[0] not in ddict.keys(): ddict[lsplit[0]] = list() if tag == 'indices': ddict[lsplit[0]].append(int(lsplit[1])) else: ddict[lsplit[0]].extend([int(lsplit[1]) for i in range(snum)]) elif tag == 'stations' and len(lsplit) >= ndata[tag]: if len(dkeys[tag]) == 0: for ll in lsplit: ll = re.sub('-', '_', ll) dkeys[tag].append(ll) ddict[ll] = list() else: date_list.append(dtime) for i, ll in enumerate(lsplit): if i >= 1 and i <= 4: ddict[dkeys[tag][i]].append(float(ll)) elif i == 6: ddict[dkeys[tag][i]].append(int(ll)) elif i < len(dkeys[tag]): ddict[dkeys[tag][i]].append(ll) else: ddict[dkeys[tag][-1]][-1] += ' {:s}'.format(ll) elif len(lsplit) == ndata['']: snum -= 1 if tag != 'indices': if len(ddict.keys()) < ndata['']: for kk in dkeys['']: ddict[kk] = list() for i, kk in enumerate(dkeys['']): if i == 0: ddict[kk].append(lsplit[i]) else: ddict[kk].append(float(lsplit[i])) if tag != 'stations' and snum == 0 and len(ddict.items()) >= 2: dflag = True data = pds.DataFrame(ddict, index=date_list, columns=ddict.keys()) fopen.close() return data, baseline
load_ascii_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fname, tag):\n import re\n ndata = {'indices': 2, '': 4, 'all': 4, 'stations': 8}\n dkeys = {'stations': list(), '': ['IAGA', 'N', 'E', 'Z']}\n data = pds.DataFrame(None)\n baseline = None\n if not tag in ndata.keys():\n return data, baseline\n with open(fname, 'r') as fopen:\n hflag = True\n pflag = False\n dflag = False if tag == 'stations' else True\n snum = 0\n ddict = dict()\n date_list = list()\n if tag == 'stations':\n dtime = pds.datetime.strptime(fname.split('_')[-1].split('.')[0\n ], '%Y')\n for fline in fopen.readlines():\n line_len = len(fline)\n if hflag:\n if pflag:\n pflag = False\n if fline.find('-mlt') > 0:\n ndata[''] += 2\n dkeys[''].extend(['MLT', 'MLAT'])\n if fline.find('-sza') > 0:\n ndata[''] += 1\n dkeys[''].append('SZA')\n if fline.find('-decl') > 0:\n ndata[''] += 1\n dkeys[''].append('IGRF_DECL')\n if tag == 'indices' and fline.find('-envelope') < 0:\n break\n lsplit = fline.split()\n idelta = lsplit.index('-delta') + 1\n ibase = lsplit.index('-baseline') + 1\n isd = lsplit.index('-sd') + 1\n ist = lsplit.index('-st') + 1\n iex = lsplit.index('-ex') + 1\n baseline = ' '.join([lsplit[ibase], lsplit[idelta],\n lsplit[isd], lsplit[ist], lsplit[iex]])\n if fline.find('Selected parameters:') >= 0:\n pflag = True\n if fline.count('=') == line_len - 1 and line_len > 2:\n hflag = False\n else:\n lsplit = [ll for ll in re.split('[\\\\t\\\\n]+', fline) if len(\n ll) > 0]\n if dflag:\n dflag = False\n dstring = ' '.join(lsplit[:6])\n dtime = pysat.datetime.strptime(dstring,\n '%Y %m %d %H %M %S')\n snum = int(lsplit[6])\n if tag == 'indices':\n date_list.append(dtime)\n else:\n date_list.extend([dtime for i in range(snum)])\n elif len(lsplit) == ndata['indices']:\n if tag is not '':\n if lsplit[0] not in ddict.keys():\n ddict[lsplit[0]] = list()\n if tag == 'indices':\n ddict[lsplit[0]].append(int(lsplit[1]))\n else:\n ddict[lsplit[0]].extend([int(lsplit[1]) for i in\n range(snum)])\n elif tag == 'stations' and len(lsplit) >= ndata[tag]:\n if len(dkeys[tag]) == 0:\n for ll in lsplit:\n ll = re.sub('-', '_', ll)\n dkeys[tag].append(ll)\n ddict[ll] = list()\n else:\n date_list.append(dtime)\n for i, ll in enumerate(lsplit):\n if i >= 1 and i <= 4:\n ddict[dkeys[tag][i]].append(float(ll))\n elif i == 6:\n ddict[dkeys[tag][i]].append(int(ll))\n elif i < len(dkeys[tag]):\n ddict[dkeys[tag][i]].append(ll)\n else:\n ddict[dkeys[tag][-1]][-1] += ' {:s}'.format(ll)\n elif len(lsplit) == ndata['']:\n snum -= 1\n if tag != 'indices':\n if len(ddict.keys()) < ndata['']:\n for kk in dkeys['']:\n ddict[kk] = list()\n for i, kk in enumerate(dkeys['']):\n if i == 0:\n ddict[kk].append(lsplit[i])\n else:\n ddict[kk].append(float(lsplit[i]))\n if tag != 'stations' and snum == 0 and len(ddict.items()) >= 2:\n dflag = True\n data = pds.DataFrame(ddict, index=date_list, columns=ddict.keys())\n fopen.close()\n return data, baseline\n", "role": "user" }, { "content": "load_ascii_data", "role": "assistant" } ]
def x(col_name): smag_units = {'IAGA': 'none', 'N': 'nT', 'E': 'nT', 'Z': 'nT', 'MLT': 'hours', 'MLAT': 'degrees', 'SZA': 'degrees', 'IGRF_DECL': 'degrees', 'SMU': 'none', 'SML': 'none', 'datetime': 'YYYY-MM-DD HH:MM:SS', 'GEOLON': 'degrees', 'GEOLAT': 'degrees', 'AACGMLON': 'degrees', 'AACGMLAT': 'degrees', 'STATION_NAME': 'none', 'OPERATOR_NUM': 'none', 'OPERATORS': 'none'} smag_name = {'IAGA': 'Station Code', 'N': 'B along local magnetic North', 'E': 'B along local magnetic East', 'Z': 'B vertically downward', 'MLT': 'Magnetic Local Time', 'MLAT': 'Magnetic Latitude', 'SZA': 'Solar Zenith Angle', 'IGRF_DECL': 'IGRF magnetic declination', 'SMU': """Maximum eastward auroral electrojets strength. Upper envelope of N-component for stations between 40 and 80 degrees magnetic north.""" , 'SML': """Maximum westward auroral electrojets strength. Lower envelope of N-component for stations between 40 and 80 degrees magnetic north.""" , 'datetime': 'UT date and time', 'GEOLON': 'geographic longitude', 'GEOLAT': 'geographic latitude', 'AACGMLON': 'Altitude-Adjusted Corrected Geomagnetic longitude', 'AACGMLAT': 'Altitude-Adjusted Corrected Geomagnetic latitude', 'STATION_NAME': 'Long form station name', 'OPERATOR_NUM': 'Number of station operators', 'OPERATORS': 'Station operator name(s)'} ackn = 'When using this data please include the following reference:\n' ackn += 'Gjerloev, J. W., The SuperMAG data processing technique, ' ackn += 'Geophys. Res., 117, A09213, doi:10.1029/2012JA017683, 2012\n\n' ackn += 'For publications and presentations, please include the following' ackn += 'acknowledgement:\nFor the ground magnetometer data we gratefully ' ackn += 'acknowledge: Intermagnet; USGS, Jeffrey J. Love; CARISMA, PI Ian ' ackn += 'Mann; CANMOS; The S-RAMP Database, PI K. Yumoto and Dr. K. ' ackn += 'Shiokawa; The SPIDR database; AARI, PI Oleg Troshichev; The ' ackn += 'MACCS program, PI M. Engebretson, Geomagnetism Unit of the ' ackn += 'Geological Survey of Canada; GIMA; MEASURE, UCLA IGPP and Florida' ackn += ' Institute of Technology; SAMBA, PI Eftyhia Zesta; 210 Chain, PI ' ackn += 'K. Yumoto; SAMNET, PI Farideh Honary; The institutes who maintain' ackn += ( ' the IMAGE magnetometer array, PI Eija Tanskanen; PENGUIN; AUTUMN,') ackn += ( ' PI Martin Connors; DTU Space, PI Dr. Rico Behlke; South Pole and ') ackn += " McMurdo Magnetometer, PI's Louis J. Lanzarotti and Alan T. " ackn += ( 'Weatherwax; ICESTAR; RAPIDMAG; PENGUIn; British Artarctic Survey; ') ackn += 'McMac, PI Dr. Peter Chi; BGS, PI Dr. Susan Macmillan; Pushkov ' ackn += 'Institute of Terrestrial Magnetism, Ionosphere and Radio Wave ' ackn += 'Propagation (IZMIRAN); GFZ, PI Dr. Juergen Matzka; MFGI, PI B. ' ackn += 'Heilig; IGFPAS, PI J. Reda; University of L’Aquila, PI M. ' ackn += 'Vellante; BCMT, V. Lesur and A. Chambodut; Data obtained in ' ackn += 'cooperation with Geoscience Australia, PI Marina Costelloe; ' ackn += 'SuperMAG, PI Jesper W. Gjerloev.' col_dict = {'units': smag_units[col_name], 'long_name': smag_name[ col_name], 'acknowledgements': ackn} return col_dict
update_smag_metadata
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(col_name):\n smag_units = {'IAGA': 'none', 'N': 'nT', 'E': 'nT', 'Z': 'nT', 'MLT':\n 'hours', 'MLAT': 'degrees', 'SZA': 'degrees', 'IGRF_DECL':\n 'degrees', 'SMU': 'none', 'SML': 'none', 'datetime':\n 'YYYY-MM-DD HH:MM:SS', 'GEOLON': 'degrees', 'GEOLAT': 'degrees',\n 'AACGMLON': 'degrees', 'AACGMLAT': 'degrees', 'STATION_NAME':\n 'none', 'OPERATOR_NUM': 'none', 'OPERATORS': 'none'}\n smag_name = {'IAGA': 'Station Code', 'N':\n 'B along local magnetic North', 'E': 'B along local magnetic East',\n 'Z': 'B vertically downward', 'MLT': 'Magnetic Local Time', 'MLAT':\n 'Magnetic Latitude', 'SZA': 'Solar Zenith Angle', 'IGRF_DECL':\n 'IGRF magnetic declination', 'SMU':\n \"\"\"Maximum eastward auroral electrojets strength.\nUpper envelope of N-component for stations between 40 and 80 degrees magnetic north.\"\"\"\n , 'SML':\n \"\"\"Maximum westward auroral electrojets strength.\nLower envelope of N-component for stations between 40 and 80 degrees magnetic north.\"\"\"\n , 'datetime': 'UT date and time', 'GEOLON': 'geographic longitude',\n 'GEOLAT': 'geographic latitude', 'AACGMLON':\n 'Altitude-Adjusted Corrected Geomagnetic longitude', 'AACGMLAT':\n 'Altitude-Adjusted Corrected Geomagnetic latitude', 'STATION_NAME':\n 'Long form station name', 'OPERATOR_NUM':\n 'Number of station operators', 'OPERATORS': 'Station operator name(s)'}\n ackn = 'When using this data please include the following reference:\\n'\n ackn += 'Gjerloev, J. W., The SuperMAG data processing technique, '\n ackn += 'Geophys. Res., 117, A09213, doi:10.1029/2012JA017683, 2012\\n\\n'\n ackn += 'For publications and presentations, please include the following'\n ackn += 'acknowledgement:\\nFor the ground magnetometer data we gratefully '\n ackn += 'acknowledge: Intermagnet; USGS, Jeffrey J. Love; CARISMA, PI Ian '\n ackn += 'Mann; CANMOS; The S-RAMP Database, PI K. Yumoto and Dr. K. '\n ackn += 'Shiokawa; The SPIDR database; AARI, PI Oleg Troshichev; The '\n ackn += 'MACCS program, PI M. Engebretson, Geomagnetism Unit of the '\n ackn += 'Geological Survey of Canada; GIMA; MEASURE, UCLA IGPP and Florida'\n ackn += ' Institute of Technology; SAMBA, PI Eftyhia Zesta; 210 Chain, PI '\n ackn += 'K. Yumoto; SAMNET, PI Farideh Honary; The institutes who maintain'\n ackn += (\n ' the IMAGE magnetometer array, PI Eija Tanskanen; PENGUIN; AUTUMN,')\n ackn += (\n ' PI Martin Connors; DTU Space, PI Dr. Rico Behlke; South Pole and ')\n ackn += \" McMurdo Magnetometer, PI's Louis J. Lanzarotti and Alan T. \"\n ackn += (\n 'Weatherwax; ICESTAR; RAPIDMAG; PENGUIn; British Artarctic Survey; ')\n ackn += 'McMac, PI Dr. Peter Chi; BGS, PI Dr. Susan Macmillan; Pushkov '\n ackn += 'Institute of Terrestrial Magnetism, Ionosphere and Radio Wave '\n ackn += 'Propagation (IZMIRAN); GFZ, PI Dr. Juergen Matzka; MFGI, PI B. '\n ackn += 'Heilig; IGFPAS, PI J. Reda; University of L’Aquila, PI M. '\n ackn += 'Vellante; BCMT, V. Lesur and A. Chambodut; Data obtained in '\n ackn += 'cooperation with Geoscience Australia, PI Marina Costelloe; '\n ackn += 'SuperMAG, PI Jesper W. Gjerloev.'\n col_dict = {'units': smag_units[col_name], 'long_name': smag_name[\n col_name], 'acknowledgements': ackn}\n return col_dict\n", "role": "user" }, { "content": "update_smag_metadata", "role": "assistant" } ]
def x(baseline_list): uniq_base = dict() uniq_delta = dict() for bline in baseline_list: bsplit = bline.split() bdate = ' '.join(bsplit[2:]) if bsplit[0] not in uniq_base.keys(): uniq_base[bsplit[0]] = '' if bsplit[1] not in uniq_delta.keys(): uniq_delta[bsplit[1]] = '' uniq_base[bsplit[0]] += '{:s}, '.format(bdate) uniq_delta[bsplit[1]] += '{:s}, '.format(bdate) if len(uniq_base.items()) == 1: base_string = 'Baseline {:s}'.format(list(uniq_base.keys())[0]) else: base_string = 'Baseline ' for i, kk in enumerate(uniq_base.keys()): if i == 1: base_string += '{:s}: {:s}'.format(kk, uniq_base[kk][:-2]) else: base_string += ' {:s}: {:s}'.format(kk, uniq_base[ kk][:-2]) else: base_string += 'unknown' if len(uniq_delta.items()) == 1: base_string += '\nDelta {:s}'.format(list(uniq_delta.keys())[0]) else: base_string += '\nDelta ' for i, kk in enumerate(uniq_delta.keys()): if i == 1: base_string += '{:s}: {:s}'.format(kk, uniq_delta[kk][:-2]) else: base_string += ' {:s}: {:s}'.format(kk, uniq_delta[ kk][:-2]) else: base_string += 'unknown' return base_string
format_baseline_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(baseline_list):\n uniq_base = dict()\n uniq_delta = dict()\n for bline in baseline_list:\n bsplit = bline.split()\n bdate = ' '.join(bsplit[2:])\n if bsplit[0] not in uniq_base.keys():\n uniq_base[bsplit[0]] = ''\n if bsplit[1] not in uniq_delta.keys():\n uniq_delta[bsplit[1]] = ''\n uniq_base[bsplit[0]] += '{:s}, '.format(bdate)\n uniq_delta[bsplit[1]] += '{:s}, '.format(bdate)\n if len(uniq_base.items()) == 1:\n base_string = 'Baseline {:s}'.format(list(uniq_base.keys())[0])\n else:\n base_string = 'Baseline '\n for i, kk in enumerate(uniq_base.keys()):\n if i == 1:\n base_string += '{:s}: {:s}'.format(kk, uniq_base[kk][:-2])\n else:\n base_string += ' {:s}: {:s}'.format(kk, uniq_base[\n kk][:-2])\n else:\n base_string += 'unknown'\n if len(uniq_delta.items()) == 1:\n base_string += '\\nDelta {:s}'.format(list(uniq_delta.keys())[0])\n else:\n base_string += '\\nDelta '\n for i, kk in enumerate(uniq_delta.keys()):\n if i == 1:\n base_string += '{:s}: {:s}'.format(kk, uniq_delta[kk][:-2])\n else:\n base_string += ' {:s}: {:s}'.format(kk, uniq_delta[\n kk][:-2])\n else:\n base_string += 'unknown'\n return base_string\n", "role": "user" }, { "content": "format_baseline_list", "role": "assistant" } ]
def x(file_strings, file_fmt, tag): if file_fmt.lower() == 'csv': return append_csv_data(file_strings) else: return append_ascii_data(file_strings, tag)
append_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(file_strings, file_fmt, tag):\n if file_fmt.lower() == 'csv':\n return append_csv_data(file_strings)\n else:\n return append_ascii_data(file_strings, tag)\n", "role": "user" }, { "content": "append_data", "role": "assistant" } ]
def x(file_strings, tag): import re out_lines = file_strings[0].split('\n') iparam = -1 ihead = -1 idates = list() date_list = list() num_stations = list() ind_num = 2 if tag in ['all', 'indices', ''] else 0 for i, line in enumerate(out_lines): if line == 'Selected parameters:': iparam = i + 1 elif line.count('=') == len(line) and len(line) > 2: ihead = i break i = ihead + 1 while i < len(out_lines) - 1: idates.append(i) lsplit = re.split('\t+', out_lines[i]) dtime = pds.datetime.strptime(' '.join(lsplit[0:-1]), '%Y %m %d %H %M %S') date_list.append(dtime) num_stations.append(int(lsplit[-1])) i += num_stations[-1] + 1 + ind_num idates = np.array(idates) station_names = list() for ff in range(len(file_strings) - 1): file_lines = file_strings[ff + 1].split('\n') head = True snum = 0 for i, line in enumerate(file_lines): if head: if line.count('=') == len(line) and len(line) > 2: head = False elif len(line) > 0: lsplit = re.split('\t+', line) if snum == 0: dtime = pds.datetime.strptime(' '.join(lsplit[0:-1]), '%Y %m %d %H %M %S') try: idate = date_list.index(dtime) except: raise ValueError('Unexpected date ', dtime) snum = int(lsplit[-1]) onum = num_stations[idate] inum = ind_num idates[idate + 1:] += snum num_stations[idate] += snum oline = '{:s}\t{:d}'.format(dtime.strftime( '%Y\t%m\t%d\t%H\t%M\t%S'), num_stations[idate]) out_lines[idates[idate]] = oline elif inum > 0: inum -= 1 else: onum += 1 snum -= 1 out_lines.insert(idates[idate] + onum, line) if not lsplit[0] in station_names: station_names.append(lsplit[0]) out_lines[iparam] += ',' + ','.join(station_names) out_string = '\n'.join(out_lines) return out_string
append_ascii_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(file_strings, tag):\n import re\n out_lines = file_strings[0].split('\\n')\n iparam = -1\n ihead = -1\n idates = list()\n date_list = list()\n num_stations = list()\n ind_num = 2 if tag in ['all', 'indices', ''] else 0\n for i, line in enumerate(out_lines):\n if line == 'Selected parameters:':\n iparam = i + 1\n elif line.count('=') == len(line) and len(line) > 2:\n ihead = i\n break\n i = ihead + 1\n while i < len(out_lines) - 1:\n idates.append(i)\n lsplit = re.split('\\t+', out_lines[i])\n dtime = pds.datetime.strptime(' '.join(lsplit[0:-1]),\n '%Y %m %d %H %M %S')\n date_list.append(dtime)\n num_stations.append(int(lsplit[-1]))\n i += num_stations[-1] + 1 + ind_num\n idates = np.array(idates)\n station_names = list()\n for ff in range(len(file_strings) - 1):\n file_lines = file_strings[ff + 1].split('\\n')\n head = True\n snum = 0\n for i, line in enumerate(file_lines):\n if head:\n if line.count('=') == len(line) and len(line) > 2:\n head = False\n elif len(line) > 0:\n lsplit = re.split('\\t+', line)\n if snum == 0:\n dtime = pds.datetime.strptime(' '.join(lsplit[0:-1]),\n '%Y %m %d %H %M %S')\n try:\n idate = date_list.index(dtime)\n except:\n raise ValueError('Unexpected date ', dtime)\n snum = int(lsplit[-1])\n onum = num_stations[idate]\n inum = ind_num\n idates[idate + 1:] += snum\n num_stations[idate] += snum\n oline = '{:s}\\t{:d}'.format(dtime.strftime(\n '%Y\\t%m\\t%d\\t%H\\t%M\\t%S'), num_stations[idate])\n out_lines[idates[idate]] = oline\n elif inum > 0:\n inum -= 1\n else:\n onum += 1\n snum -= 1\n out_lines.insert(idates[idate] + onum, line)\n if not lsplit[0] in station_names:\n station_names.append(lsplit[0])\n out_lines[iparam] += ',' + ','.join(station_names)\n out_string = '\\n'.join(out_lines)\n return out_string\n", "role": "user" }, { "content": "append_ascii_data", "role": "assistant" } ]
def x(file_strings): out_lines = list() head_line = None for fstrings in file_strings: file_lines = fstrings.split('\n') head_line = file_lines.pop(0) out_lines.extend(file_lines) out_lines.sort() i = 0 while len(out_lines[i]) == 0: out_lines.pop(i) out_lines.insert(0, head_line) out_lines.append('') out_string = '\n'.join(out_lines) return out_string
append_csv_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(file_strings):\n out_lines = list()\n head_line = None\n for fstrings in file_strings:\n file_lines = fstrings.split('\\n')\n head_line = file_lines.pop(0)\n out_lines.extend(file_lines)\n out_lines.sort()\n i = 0\n while len(out_lines[i]) == 0:\n out_lines.pop(i)\n out_lines.insert(0, head_line)\n out_lines.append('')\n out_string = '\\n'.join(out_lines)\n return out_string\n", "role": "user" }, { "content": "append_csv_data", "role": "assistant" } ]
def x(self): self.custom.add(add_quasi_dipole_coordinates, 'modify') self.custom.add(add_aacgm_coordinates, 'modify') self.custom.add(calculate_ecef_velocity, 'modify') self.custom.add(add_sc_attitude_vectors, 'modify') self.custom.add(add_iri_thermal_plasma, 'modify') self.custom.add(add_hwm_winds_and_ecef_vectors, 'modify') self.custom.add(add_igrf, 'modify') in_meta = {'desc': 'IGRF geomagnetic field expressed in the s/c basis.', 'units': 'nT'} self.custom.add(project_ecef_vector_onto_sc, 'modify', 'end', 'B_ecef_x', 'B_ecef_y', 'B_ecef_z', 'B_sc_x', 'B_sc_y', 'B_sc_z', meta=[in_meta.copy(), in_meta.copy(), in_meta.copy()]) self.custom.add(project_hwm_onto_sc, 'modify') self.custom.add(add_msis, 'modify')
init
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.custom.add(add_quasi_dipole_coordinates, 'modify')\n self.custom.add(add_aacgm_coordinates, 'modify')\n self.custom.add(calculate_ecef_velocity, 'modify')\n self.custom.add(add_sc_attitude_vectors, 'modify')\n self.custom.add(add_iri_thermal_plasma, 'modify')\n self.custom.add(add_hwm_winds_and_ecef_vectors, 'modify')\n self.custom.add(add_igrf, 'modify')\n in_meta = {'desc': 'IGRF geomagnetic field expressed in the s/c basis.',\n 'units': 'nT'}\n self.custom.add(project_ecef_vector_onto_sc, 'modify', 'end',\n 'B_ecef_x', 'B_ecef_y', 'B_ecef_z', 'B_sc_x', 'B_sc_y', 'B_sc_z',\n meta=[in_meta.copy(), in_meta.copy(), in_meta.copy()])\n self.custom.add(project_hwm_onto_sc, 'modify')\n self.custom.add(add_msis, 'modify')\n", "role": "user" }, { "content": "init", "role": "assistant" } ]
def x(fnames, tag=None, sat_id=None, obs_long=0.0, obs_lat=0.0, obs_alt=0.0, TLE1=None, TLE2=None): import sgp4 from sgp4.earth_gravity import wgs72 from sgp4.io import twoline2rv import ephem import pysatMagVect line1 = ( '1 25544U 98067A 18135.61844383 .00002728 00000-0 48567-4 0 9998' ) line2 = ( '2 25544 51.6402 181.0633 0004018 88.8954 22.2246 15.54059185113452' ) if TLE1 is not None: line1 = TLE1 if TLE2 is not None: line2 = TLE2 satellite = twoline2rv(line1, line2, wgs72) parts = os.path.split(fnames[0])[-1].split('-') yr = int(parts[0]) month = int(parts[1]) day = int(parts[2][0:2]) date = pysat.datetime(yr, month, day) times = pds.date_range(start=date, end=date + pds.DateOffset(seconds= 86399), freq='1S') on_travis = os.environ.get('ONTRAVIS') == 'True' if on_travis: times = times[0:100] position = [] velocity = [] for time in times: pos, vel = satellite.propagate(time.year, time.month, time.day, time.hour, time.minute, time.second) position.extend(pos) velocity.extend(vel) data = pysat.DataFrame({'position_eci_x': position[::3], 'position_eci_y': position[1::3], 'position_eci_z': position[2::3], 'velocity_eci_x': velocity[::3], 'velocity_eci_y': velocity[1::3], 'velocity_eci_z': velocity[2::3]}, index=times) data.index.name = 'Epoch' site = ephem.Observer() site.lon = str(obs_long) site.lat = str(obs_lat) site.elevation = obs_alt sat = ephem.readtle('pysat', line1, line2) output_params = [] for time in times: lp = {} site.date = time sat.compute(site) lp['obs_sat_az_angle'] = ephem.degrees(sat.az) lp['obs_sat_el_angle'] = ephem.degrees(sat.alt) lp['obs_sat_slant_range'] = sat.range lp['glat'] = np.degrees(sat.sublat) lp['glong'] = np.degrees(sat.sublong) lp['alt'] = sat.elevation / 1000.0 lp['x'], lp['y'], lp['z'] = pysatMagVect.geodetic_to_ecef(lp['glat' ], lp['glong'], lp['alt']) output_params.append(lp) output = pds.DataFrame(output_params, index=times) data[['glong', 'glat', 'alt']] = output[['glong', 'glat', 'alt']] data[['position_ecef_x', 'position_ecef_y', 'position_ecef_z']] = output[[ 'x', 'y', 'z']] data['obs_sat_az_angle'] = output['obs_sat_az_angle'] data['obs_sat_el_angle'] = output['obs_sat_el_angle'] data['obs_sat_slant_range'] = output['obs_sat_slant_range'] return data, meta.copy()
load
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fnames, tag=None, sat_id=None, obs_long=0.0, obs_lat=0.0, obs_alt=0.0,\n TLE1=None, TLE2=None):\n import sgp4\n from sgp4.earth_gravity import wgs72\n from sgp4.io import twoline2rv\n import ephem\n import pysatMagVect\n line1 = (\n '1 25544U 98067A 18135.61844383 .00002728 00000-0 48567-4 0 9998'\n )\n line2 = (\n '2 25544 51.6402 181.0633 0004018 88.8954 22.2246 15.54059185113452'\n )\n if TLE1 is not None:\n line1 = TLE1\n if TLE2 is not None:\n line2 = TLE2\n satellite = twoline2rv(line1, line2, wgs72)\n parts = os.path.split(fnames[0])[-1].split('-')\n yr = int(parts[0])\n month = int(parts[1])\n day = int(parts[2][0:2])\n date = pysat.datetime(yr, month, day)\n times = pds.date_range(start=date, end=date + pds.DateOffset(seconds=\n 86399), freq='1S')\n on_travis = os.environ.get('ONTRAVIS') == 'True'\n if on_travis:\n times = times[0:100]\n position = []\n velocity = []\n for time in times:\n pos, vel = satellite.propagate(time.year, time.month, time.day,\n time.hour, time.minute, time.second)\n position.extend(pos)\n velocity.extend(vel)\n data = pysat.DataFrame({'position_eci_x': position[::3],\n 'position_eci_y': position[1::3], 'position_eci_z': position[2::3],\n 'velocity_eci_x': velocity[::3], 'velocity_eci_y': velocity[1::3],\n 'velocity_eci_z': velocity[2::3]}, index=times)\n data.index.name = 'Epoch'\n site = ephem.Observer()\n site.lon = str(obs_long)\n site.lat = str(obs_lat)\n site.elevation = obs_alt\n sat = ephem.readtle('pysat', line1, line2)\n output_params = []\n for time in times:\n lp = {}\n site.date = time\n sat.compute(site)\n lp['obs_sat_az_angle'] = ephem.degrees(sat.az)\n lp['obs_sat_el_angle'] = ephem.degrees(sat.alt)\n lp['obs_sat_slant_range'] = sat.range\n lp['glat'] = np.degrees(sat.sublat)\n lp['glong'] = np.degrees(sat.sublong)\n lp['alt'] = sat.elevation / 1000.0\n lp['x'], lp['y'], lp['z'] = pysatMagVect.geodetic_to_ecef(lp['glat'\n ], lp['glong'], lp['alt'])\n output_params.append(lp)\n output = pds.DataFrame(output_params, index=times)\n data[['glong', 'glat', 'alt']] = output[['glong', 'glat', 'alt']]\n data[['position_ecef_x', 'position_ecef_y', 'position_ecef_z']] = output[[\n 'x', 'y', 'z']]\n data['obs_sat_az_angle'] = output['obs_sat_az_angle']\n data['obs_sat_el_angle'] = output['obs_sat_el_angle']\n data['obs_sat_slant_range'] = output['obs_sat_slant_range']\n return data, meta.copy()\n", "role": "user" }, { "content": "load", "role": "assistant" } ]
def x(tag=None, sat_id=None, data_path=None, format_str=None): index = pds.date_range(pysat.datetime(2017, 12, 1), pysat.datetime(2018, 12, 1)) names = [(data_path + date.strftime('%Y-%m-%d') + '.nofile') for date in index] return pysat.Series(names, index=index)
list_files
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(tag=None, sat_id=None, data_path=None, format_str=None):\n index = pds.date_range(pysat.datetime(2017, 12, 1), pysat.datetime(2018,\n 12, 1))\n names = [(data_path + date.strftime('%Y-%m-%d') + '.nofile') for date in\n index]\n return pysat.Series(names, index=index)\n", "role": "user" }, { "content": "list_files", "role": "assistant" } ]
def x(inst): import pysatMagVect inst['sc_xhat_ecef_x'], inst['sc_xhat_ecef_y'], inst['sc_xhat_ecef_z' ] = pysatMagVect.normalize_vector(inst['velocity_ecef_x'], inst[ 'velocity_ecef_y'], inst['velocity_ecef_z']) inst['sc_zhat_ecef_x'], inst['sc_zhat_ecef_y'], inst['sc_zhat_ecef_z' ] = pysatMagVect.normalize_vector(-inst['position_ecef_x'], -inst[ 'position_ecef_y'], -inst['position_ecef_z']) inst['sc_yhat_ecef_x'], inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z' ] = pysatMagVect.cross_product(inst['sc_zhat_ecef_x'], inst[ 'sc_zhat_ecef_y'], inst['sc_zhat_ecef_z'], inst['sc_xhat_ecef_x'], inst['sc_xhat_ecef_y'], inst['sc_xhat_ecef_z']) inst['sc_yhat_ecef_x'], inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z' ] = pysatMagVect.normalize_vector(inst['sc_yhat_ecef_x'], inst[ 'sc_yhat_ecef_y'], inst['sc_yhat_ecef_z']) inst['sc_zhat_ecef_x'], inst['sc_zhat_ecef_y'], inst['sc_zhat_ecef_z' ] = pysatMagVect.cross_product(inst['sc_xhat_ecef_x'], inst[ 'sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'], inst['sc_yhat_ecef_x'], inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z']) inst.meta['sc_xhat_ecef_x'] = {'units': '', 'desc': 'S/C attitude (x-direction, ram) unit vector, expressed in ECEF basis, x-component' } inst.meta['sc_xhat_ecef_y'] = {'units': '', 'desc': 'S/C attitude (x-direction, ram) unit vector, expressed in ECEF basis, y-component' } inst.meta['sc_xhat_ecef_z'] = {'units': '', 'desc': 'S/C attitude (x-direction, ram) unit vector, expressed in ECEF basis, z-component' } inst.meta['sc_zhat_ecef_x'] = {'units': '', 'desc': 'S/C attitude (z-direction, generally nadir) unit vector, expressed in ECEF basis, x-component' } inst.meta['sc_zhat_ecef_y'] = {'units': '', 'desc': 'S/C attitude (z-direction, generally nadir) unit vector, expressed in ECEF basis, y-component' } inst.meta['sc_zhat_ecef_z'] = {'units': '', 'desc': 'S/C attitude (z-direction, generally nadir) unit vector, expressed in ECEF basis, z-component' } inst.meta['sc_yhat_ecef_x'] = {'units': '', 'desc': 'S/C attitude (y-direction, generally south) unit vector, expressed in ECEF basis, x-component' } inst.meta['sc_yhat_ecef_y'] = {'units': '', 'desc': 'S/C attitude (y-direction, generally south) unit vector, expressed in ECEF basis, y-component' } inst.meta['sc_yhat_ecef_z'] = {'units': '', 'desc': 'S/C attitude (y-direction, generally south) unit vector, expressed in ECEF basis, z-component' } mag = np.sqrt(inst['sc_zhat_ecef_x'] ** 2 + inst['sc_zhat_ecef_y'] ** 2 + inst['sc_zhat_ecef_z'] ** 2) idx, = np.where((mag < 0.999999999) | (mag > 1.000000001)) if len(idx) > 0: print(mag[idx]) raise RuntimeError( 'Unit vector generation failure. Not sufficently orthogonal.') return
add_sc_attitude_vectors
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(inst):\n import pysatMagVect\n inst['sc_xhat_ecef_x'], inst['sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'\n ] = pysatMagVect.normalize_vector(inst['velocity_ecef_x'], inst[\n 'velocity_ecef_y'], inst['velocity_ecef_z'])\n inst['sc_zhat_ecef_x'], inst['sc_zhat_ecef_y'], inst['sc_zhat_ecef_z'\n ] = pysatMagVect.normalize_vector(-inst['position_ecef_x'], -inst[\n 'position_ecef_y'], -inst['position_ecef_z'])\n inst['sc_yhat_ecef_x'], inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'\n ] = pysatMagVect.cross_product(inst['sc_zhat_ecef_x'], inst[\n 'sc_zhat_ecef_y'], inst['sc_zhat_ecef_z'], inst['sc_xhat_ecef_x'],\n inst['sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'])\n inst['sc_yhat_ecef_x'], inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'\n ] = pysatMagVect.normalize_vector(inst['sc_yhat_ecef_x'], inst[\n 'sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'])\n inst['sc_zhat_ecef_x'], inst['sc_zhat_ecef_y'], inst['sc_zhat_ecef_z'\n ] = pysatMagVect.cross_product(inst['sc_xhat_ecef_x'], inst[\n 'sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'], inst['sc_yhat_ecef_x'],\n inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'])\n inst.meta['sc_xhat_ecef_x'] = {'units': '', 'desc':\n 'S/C attitude (x-direction, ram) unit vector, expressed in ECEF basis, x-component'\n }\n inst.meta['sc_xhat_ecef_y'] = {'units': '', 'desc':\n 'S/C attitude (x-direction, ram) unit vector, expressed in ECEF basis, y-component'\n }\n inst.meta['sc_xhat_ecef_z'] = {'units': '', 'desc':\n 'S/C attitude (x-direction, ram) unit vector, expressed in ECEF basis, z-component'\n }\n inst.meta['sc_zhat_ecef_x'] = {'units': '', 'desc':\n 'S/C attitude (z-direction, generally nadir) unit vector, expressed in ECEF basis, x-component'\n }\n inst.meta['sc_zhat_ecef_y'] = {'units': '', 'desc':\n 'S/C attitude (z-direction, generally nadir) unit vector, expressed in ECEF basis, y-component'\n }\n inst.meta['sc_zhat_ecef_z'] = {'units': '', 'desc':\n 'S/C attitude (z-direction, generally nadir) unit vector, expressed in ECEF basis, z-component'\n }\n inst.meta['sc_yhat_ecef_x'] = {'units': '', 'desc':\n 'S/C attitude (y-direction, generally south) unit vector, expressed in ECEF basis, x-component'\n }\n inst.meta['sc_yhat_ecef_y'] = {'units': '', 'desc':\n 'S/C attitude (y-direction, generally south) unit vector, expressed in ECEF basis, y-component'\n }\n inst.meta['sc_yhat_ecef_z'] = {'units': '', 'desc':\n 'S/C attitude (y-direction, generally south) unit vector, expressed in ECEF basis, z-component'\n }\n mag = np.sqrt(inst['sc_zhat_ecef_x'] ** 2 + inst['sc_zhat_ecef_y'] ** 2 +\n inst['sc_zhat_ecef_z'] ** 2)\n idx, = np.where((mag < 0.999999999) | (mag > 1.000000001))\n if len(idx) > 0:\n print(mag[idx])\n raise RuntimeError(\n 'Unit vector generation failure. Not sufficently orthogonal.')\n return\n", "role": "user" }, { "content": "add_sc_attitude_vectors", "role": "assistant" } ]
def x(inst): x = inst['position_ecef_x'] vel_x = (x.values[2:] - x.values[0:-2]) / 2.0 y = inst['position_ecef_y'] vel_y = (y.values[2:] - y.values[0:-2]) / 2.0 z = inst['position_ecef_z'] vel_z = (z.values[2:] - z.values[0:-2]) / 2.0 inst[1:-1, 'velocity_ecef_x'] = vel_x inst[1:-1, 'velocity_ecef_y'] = vel_y inst[1:-1, 'velocity_ecef_z'] = vel_z inst.meta['velocity_ecef_x'] = {'units': 'km/s', 'desc': 'Velocity of satellite calculated with respect to ECEF frame.'} inst.meta['velocity_ecef_y'] = {'units': 'km/s', 'desc': 'Velocity of satellite calculated with respect to ECEF frame.'} inst.meta['velocity_ecef_z'] = {'units': 'km/s', 'desc': 'Velocity of satellite calculated with respect to ECEF frame.'} return
calculate_ecef_velocity
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(inst):\n x = inst['position_ecef_x']\n vel_x = (x.values[2:] - x.values[0:-2]) / 2.0\n y = inst['position_ecef_y']\n vel_y = (y.values[2:] - y.values[0:-2]) / 2.0\n z = inst['position_ecef_z']\n vel_z = (z.values[2:] - z.values[0:-2]) / 2.0\n inst[1:-1, 'velocity_ecef_x'] = vel_x\n inst[1:-1, 'velocity_ecef_y'] = vel_y\n inst[1:-1, 'velocity_ecef_z'] = vel_z\n inst.meta['velocity_ecef_x'] = {'units': 'km/s', 'desc':\n 'Velocity of satellite calculated with respect to ECEF frame.'}\n inst.meta['velocity_ecef_y'] = {'units': 'km/s', 'desc':\n 'Velocity of satellite calculated with respect to ECEF frame.'}\n inst.meta['velocity_ecef_z'] = {'units': 'km/s', 'desc':\n 'Velocity of satellite calculated with respect to ECEF frame.'}\n return\n", "role": "user" }, { "content": "calculate_ecef_velocity", "role": "assistant" } ]
def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'): import apexpy ap = apexpy.Apex(date=inst.date) qd_lat = [] qd_lon = [] mlt = [] for lat, lon, alt, time in zip(inst[glat_label], inst[glong_label], inst[alt_label], inst.data.index): tlat, tlon = ap.geo2qd(lat, lon, alt) qd_lat.append(tlat) qd_lon.append(tlon) mlt.append(ap.mlon2mlt(tlon, time)) inst['qd_lat'] = qd_lat inst['qd_long'] = qd_lon inst['mlt'] = mlt inst.meta['qd_lat'] = {'units': 'degrees', 'long_name': 'Quasi dipole latitude'} inst.meta['qd_long'] = {'units': 'degrees', 'long_name': 'Quasi dipole longitude'} inst.meta['qd_mlt'] = {'units': 'hrs', 'long_name': 'Magnetic local time'} return
add_quasi_dipole_coordinates
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):\n import apexpy\n ap = apexpy.Apex(date=inst.date)\n qd_lat = []\n qd_lon = []\n mlt = []\n for lat, lon, alt, time in zip(inst[glat_label], inst[glong_label],\n inst[alt_label], inst.data.index):\n tlat, tlon = ap.geo2qd(lat, lon, alt)\n qd_lat.append(tlat)\n qd_lon.append(tlon)\n mlt.append(ap.mlon2mlt(tlon, time))\n inst['qd_lat'] = qd_lat\n inst['qd_long'] = qd_lon\n inst['mlt'] = mlt\n inst.meta['qd_lat'] = {'units': 'degrees', 'long_name':\n 'Quasi dipole latitude'}\n inst.meta['qd_long'] = {'units': 'degrees', 'long_name':\n 'Quasi dipole longitude'}\n inst.meta['qd_mlt'] = {'units': 'hrs', 'long_name': 'Magnetic local time'}\n return\n", "role": "user" }, { "content": "add_quasi_dipole_coordinates", "role": "assistant" } ]
def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'): import aacgmv2 aalat = [] aalon = [] mlt = [] for lat, lon, alt, time in zip(inst[glat_label], inst[glong_label], inst[alt_label], inst.data.index): tlat, tlon, tmlt = aacgmv2.get_aacgm_coord(lat, lon, alt, time) aalat.append(tlat) aalon.append(tlon) mlt.append(tmlt) inst['aacgm_lat'] = aalat inst['aacgm_long'] = aalon inst['aacgm_mlt'] = mlt inst.meta['aacgm_lat'] = {'units': 'degrees', 'long_name': 'AACGM latitude' } inst.meta['aacgm_long'] = {'units': 'degrees', 'long_name': 'AACGM longitude'} inst.meta['aacgm_mlt'] = {'units': 'hrs', 'long_name': 'AACGM Magnetic local time'} return
add_aacgm_coordinates
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):\n import aacgmv2\n aalat = []\n aalon = []\n mlt = []\n for lat, lon, alt, time in zip(inst[glat_label], inst[glong_label],\n inst[alt_label], inst.data.index):\n tlat, tlon, tmlt = aacgmv2.get_aacgm_coord(lat, lon, alt, time)\n aalat.append(tlat)\n aalon.append(tlon)\n mlt.append(tmlt)\n inst['aacgm_lat'] = aalat\n inst['aacgm_long'] = aalon\n inst['aacgm_mlt'] = mlt\n inst.meta['aacgm_lat'] = {'units': 'degrees', 'long_name': 'AACGM latitude'\n }\n inst.meta['aacgm_long'] = {'units': 'degrees', 'long_name':\n 'AACGM longitude'}\n inst.meta['aacgm_mlt'] = {'units': 'hrs', 'long_name':\n 'AACGM Magnetic local time'}\n return\n", "role": "user" }, { "content": "add_aacgm_coordinates", "role": "assistant" } ]
def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'): import pyglow from pyglow.pyglow import Point iri_params = [] for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[ glong_label], inst[alt_label]): pt = Point(time, lat, lon, alt) pt.run_iri() iri = {} iri['ion_temp'] = pt.Ti iri['e_temp'] = pt.Te iri['ion_dens'] = pt.ni['O+'] + pt.ni['H+'] + pt.ni['HE+'] iri['frac_dens_o'] = pt.ni['O+'] / iri['ion_dens'] iri['frac_dens_h'] = pt.ni['H+'] / iri['ion_dens'] iri['frac_dens_he'] = pt.ni['HE+'] / iri['ion_dens'] iri_params.append(iri) iri = pds.DataFrame(iri_params) iri.index = inst.data.index inst[iri.keys()] = iri inst.meta['ion_temp'] = {'units': 'Kelvin', 'long_name': 'Ion Temperature'} inst.meta['ion_dens'] = {'units': 'N/cc', 'long_name': 'Ion Density', 'desc': 'Total ion density including O+ and H+ from IRI model run.'} inst.meta['frac_dens_o'] = {'units': '', 'long_name': 'Fractional O+ Density'} inst.meta['frac_dens_h'] = {'units': '', 'long_name': 'Fractional H+ Density'}
add_iri_thermal_plasma
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):\n import pyglow\n from pyglow.pyglow import Point\n iri_params = []\n for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[\n glong_label], inst[alt_label]):\n pt = Point(time, lat, lon, alt)\n pt.run_iri()\n iri = {}\n iri['ion_temp'] = pt.Ti\n iri['e_temp'] = pt.Te\n iri['ion_dens'] = pt.ni['O+'] + pt.ni['H+'] + pt.ni['HE+']\n iri['frac_dens_o'] = pt.ni['O+'] / iri['ion_dens']\n iri['frac_dens_h'] = pt.ni['H+'] / iri['ion_dens']\n iri['frac_dens_he'] = pt.ni['HE+'] / iri['ion_dens']\n iri_params.append(iri)\n iri = pds.DataFrame(iri_params)\n iri.index = inst.data.index\n inst[iri.keys()] = iri\n inst.meta['ion_temp'] = {'units': 'Kelvin', 'long_name': 'Ion Temperature'}\n inst.meta['ion_dens'] = {'units': 'N/cc', 'long_name': 'Ion Density',\n 'desc': 'Total ion density including O+ and H+ from IRI model run.'}\n inst.meta['frac_dens_o'] = {'units': '', 'long_name':\n 'Fractional O+ Density'}\n inst.meta['frac_dens_h'] = {'units': '', 'long_name':\n 'Fractional H+ Density'}\n", "role": "user" }, { "content": "add_iri_thermal_plasma", "role": "assistant" } ]
def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'): import pyglow import pysatMagVect hwm_params = [] for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[ glong_label], inst[alt_label]): pt = pyglow.Point(time, lat, lon, alt) pt.run_hwm() hwm = {} hwm['zonal_wind'] = pt.u hwm['meridional_wind'] = pt.v hwm_params.append(hwm) hwm = pds.DataFrame(hwm_params) hwm.index = inst.data.index inst[['zonal_wind', 'meridional_wind']] = hwm[['zonal_wind', 'meridional_wind']] mag = np.sqrt(inst['position_ecef_x'] ** 2 + inst['position_ecef_y'] ** 2) inst['unit_zonal_wind_ecef_x'] = -inst['position_ecef_y'] / mag inst['unit_zonal_wind_ecef_y'] = inst['position_ecef_x'] / mag inst['unit_zonal_wind_ecef_z'] = 0 * inst['position_ecef_x'] unit_pos_x, unit_pos_y, unit_pos_z = pysatMagVect.normalize_vector(- inst['position_ecef_x'], -inst['position_ecef_y'], -inst[ 'position_ecef_z']) inst['unit_mer_wind_ecef_x'], inst['unit_mer_wind_ecef_y'], inst[ 'unit_mer_wind_ecef_z'] = pysatMagVect.cross_product(unit_pos_x, unit_pos_y, unit_pos_z, inst['unit_zonal_wind_ecef_x'], inst[ 'unit_zonal_wind_ecef_y'], inst['unit_zonal_wind_ecef_z']) inst.meta['zonal_wind'] = {'units': 'm/s', 'long_name': 'Zonal Wind', 'desc': 'HWM model zonal wind'} inst.meta['meridional_wind'] = {'units': 'm/s', 'long_name': 'Meridional Wind', 'desc': 'HWM model meridional wind'} inst.meta['unit_zonal_wind_ecef_x'] = {'units': '', 'long_name': 'Zonal Wind Unit ECEF x-vector', 'desc': 'x-value of zonal wind unit vector in ECEF co ordinates'} inst.meta['unit_zonal_wind_ecef_y'] = {'units': '', 'long_name': 'Zonal Wind Unit ECEF y-vector', 'desc': 'y-value of zonal wind unit vector in ECEF co ordinates'} inst.meta['unit_zonal_wind_ecef_z'] = {'units': '', 'long_name': 'Zonal Wind Unit ECEF z-vector', 'desc': 'z-value of zonal wind unit vector in ECEF co ordinates'} inst.meta['unit_mer_wind_ecef_x'] = {'units': '', 'long_name': 'Meridional Wind Unit ECEF x-vector', 'desc': 'x-value of meridional wind unit vector in ECEF co ordinates'} inst.meta['unit_mer_wind_ecef_y'] = {'units': '', 'long_name': 'Meridional Wind Unit ECEF y-vector', 'desc': 'y-value of meridional wind unit vector in ECEF co ordinates'} inst.meta['unit_mer_wind_ecef_z'] = {'units': '', 'long_name': 'Meridional Wind Unit ECEF z-vector', 'desc': 'z-value of meridional wind unit vector in ECEF co ordinates'} return
add_hwm_winds_and_ecef_vectors
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):\n import pyglow\n import pysatMagVect\n hwm_params = []\n for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[\n glong_label], inst[alt_label]):\n pt = pyglow.Point(time, lat, lon, alt)\n pt.run_hwm()\n hwm = {}\n hwm['zonal_wind'] = pt.u\n hwm['meridional_wind'] = pt.v\n hwm_params.append(hwm)\n hwm = pds.DataFrame(hwm_params)\n hwm.index = inst.data.index\n inst[['zonal_wind', 'meridional_wind']] = hwm[['zonal_wind',\n 'meridional_wind']]\n mag = np.sqrt(inst['position_ecef_x'] ** 2 + inst['position_ecef_y'] ** 2)\n inst['unit_zonal_wind_ecef_x'] = -inst['position_ecef_y'] / mag\n inst['unit_zonal_wind_ecef_y'] = inst['position_ecef_x'] / mag\n inst['unit_zonal_wind_ecef_z'] = 0 * inst['position_ecef_x']\n unit_pos_x, unit_pos_y, unit_pos_z = pysatMagVect.normalize_vector(-\n inst['position_ecef_x'], -inst['position_ecef_y'], -inst[\n 'position_ecef_z'])\n inst['unit_mer_wind_ecef_x'], inst['unit_mer_wind_ecef_y'], inst[\n 'unit_mer_wind_ecef_z'] = pysatMagVect.cross_product(unit_pos_x,\n unit_pos_y, unit_pos_z, inst['unit_zonal_wind_ecef_x'], inst[\n 'unit_zonal_wind_ecef_y'], inst['unit_zonal_wind_ecef_z'])\n inst.meta['zonal_wind'] = {'units': 'm/s', 'long_name': 'Zonal Wind',\n 'desc': 'HWM model zonal wind'}\n inst.meta['meridional_wind'] = {'units': 'm/s', 'long_name':\n 'Meridional Wind', 'desc': 'HWM model meridional wind'}\n inst.meta['unit_zonal_wind_ecef_x'] = {'units': '', 'long_name':\n 'Zonal Wind Unit ECEF x-vector', 'desc':\n 'x-value of zonal wind unit vector in ECEF co ordinates'}\n inst.meta['unit_zonal_wind_ecef_y'] = {'units': '', 'long_name':\n 'Zonal Wind Unit ECEF y-vector', 'desc':\n 'y-value of zonal wind unit vector in ECEF co ordinates'}\n inst.meta['unit_zonal_wind_ecef_z'] = {'units': '', 'long_name':\n 'Zonal Wind Unit ECEF z-vector', 'desc':\n 'z-value of zonal wind unit vector in ECEF co ordinates'}\n inst.meta['unit_mer_wind_ecef_x'] = {'units': '', 'long_name':\n 'Meridional Wind Unit ECEF x-vector', 'desc':\n 'x-value of meridional wind unit vector in ECEF co ordinates'}\n inst.meta['unit_mer_wind_ecef_y'] = {'units': '', 'long_name':\n 'Meridional Wind Unit ECEF y-vector', 'desc':\n 'y-value of meridional wind unit vector in ECEF co ordinates'}\n inst.meta['unit_mer_wind_ecef_z'] = {'units': '', 'long_name':\n 'Meridional Wind Unit ECEF z-vector', 'desc':\n 'z-value of meridional wind unit vector in ECEF co ordinates'}\n return\n", "role": "user" }, { "content": "add_hwm_winds_and_ecef_vectors", "role": "assistant" } ]
def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'): import pyglow from pyglow.pyglow import Point import pysatMagVect igrf_params = [] for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[ glong_label], inst[alt_label]): pt = Point(time, lat, lon, alt) pt.run_igrf() igrf = {} igrf['B'] = pt.B igrf['B_east'] = pt.Bx igrf['B_north'] = pt.By igrf['B_up'] = pt.Bz igrf_params.append(igrf) igrf = pds.DataFrame(igrf_params) igrf.index = inst.data.index inst[igrf.keys()] = igrf x, y, z = pysatMagVect.enu_to_ecef_vector(inst['B_east'], inst[ 'B_north'], inst['B_up'], inst[glat_label], inst[glong_label]) inst['B_ecef_x'] = x inst['B_ecef_y'] = y inst['B_ecef_z'] = z inst.meta['B'] = {'units': 'nT', 'desc': 'Total geomagnetic field from IGRF.'} inst.meta['B_east'] = {'units': 'nT', 'desc': 'Geomagnetic field from IGRF expressed using the East/North/Up (ENU) basis.' } inst.meta['B_north'] = {'units': 'nT', 'desc': 'Geomagnetic field from IGRF expressed using the East/North/Up (ENU) basis.' } inst.meta['B_up'] = {'units': 'nT', 'desc': 'Geomagnetic field from IGRF expressed using the East/North/Up (ENU) basis.' } inst.meta['B_ecef_x'] = {'units': 'nT', 'desc': 'Geomagnetic field from IGRF expressed using the Earth Centered Earth Fixed (ECEF) basis.' } inst.meta['B_ecef_y'] = {'units': 'nT', 'desc': 'Geomagnetic field from IGRF expressed using the Earth Centered Earth Fixed (ECEF) basis.' } inst.meta['B_ecef_z'] = {'units': 'nT', 'desc': 'Geomagnetic field from IGRF expressed using the Earth Centered Earth Fixed (ECEF) basis.' } return
add_igrf
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):\n import pyglow\n from pyglow.pyglow import Point\n import pysatMagVect\n igrf_params = []\n for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[\n glong_label], inst[alt_label]):\n pt = Point(time, lat, lon, alt)\n pt.run_igrf()\n igrf = {}\n igrf['B'] = pt.B\n igrf['B_east'] = pt.Bx\n igrf['B_north'] = pt.By\n igrf['B_up'] = pt.Bz\n igrf_params.append(igrf)\n igrf = pds.DataFrame(igrf_params)\n igrf.index = inst.data.index\n inst[igrf.keys()] = igrf\n x, y, z = pysatMagVect.enu_to_ecef_vector(inst['B_east'], inst[\n 'B_north'], inst['B_up'], inst[glat_label], inst[glong_label])\n inst['B_ecef_x'] = x\n inst['B_ecef_y'] = y\n inst['B_ecef_z'] = z\n inst.meta['B'] = {'units': 'nT', 'desc':\n 'Total geomagnetic field from IGRF.'}\n inst.meta['B_east'] = {'units': 'nT', 'desc':\n 'Geomagnetic field from IGRF expressed using the East/North/Up (ENU) basis.'\n }\n inst.meta['B_north'] = {'units': 'nT', 'desc':\n 'Geomagnetic field from IGRF expressed using the East/North/Up (ENU) basis.'\n }\n inst.meta['B_up'] = {'units': 'nT', 'desc':\n 'Geomagnetic field from IGRF expressed using the East/North/Up (ENU) basis.'\n }\n inst.meta['B_ecef_x'] = {'units': 'nT', 'desc':\n 'Geomagnetic field from IGRF expressed using the Earth Centered Earth Fixed (ECEF) basis.'\n }\n inst.meta['B_ecef_y'] = {'units': 'nT', 'desc':\n 'Geomagnetic field from IGRF expressed using the Earth Centered Earth Fixed (ECEF) basis.'\n }\n inst.meta['B_ecef_z'] = {'units': 'nT', 'desc':\n 'Geomagnetic field from IGRF expressed using the Earth Centered Earth Fixed (ECEF) basis.'\n }\n return\n", "role": "user" }, { "content": "add_igrf", "role": "assistant" } ]
def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'): import pyglow from pyglow.pyglow import Point msis_params = [] for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[ glong_label], inst[alt_label]): pt = Point(time, lat, lon, alt) pt.run_msis() msis = {} total = 0 for key in pt.nn.keys(): total += pt.nn[key] msis['Nn'] = total msis['Nn_N'] = pt.nn['N'] msis['Nn_N2'] = pt.nn['N2'] msis['Nn_O'] = pt.nn['O'] msis['Nn_O2'] = pt.nn['O2'] msis['Tn_msis'] = pt.Tn_msis msis_params.append(msis) msis = pds.DataFrame(msis_params) msis.index = inst.data.index inst[msis.keys()] = msis inst.meta['Nn'] = {'units': 'cm^-3', 'desc': 'Total neutral number particle density from MSIS.'} inst.meta['Nn_N'] = {'units': 'cm^-3', 'desc': 'Total nitrogen number particle density from MSIS.'} inst.meta['Nn_N2'] = {'units': 'cm^-3', 'desc': 'Total N2 number particle density from MSIS.'} inst.meta['Nn_O'] = {'units': 'cm^-3', 'desc': 'Total oxygen number particle density from MSIS.'} inst.meta['Nn_O2'] = {'units': 'cm^-3', 'desc': 'Total O2 number particle density from MSIS.'} inst.meta['Tn_msis'] = {'units': 'K', 'desc': 'Neutral temperature from MSIS.'} return
add_msis
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):\n import pyglow\n from pyglow.pyglow import Point\n msis_params = []\n for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[\n glong_label], inst[alt_label]):\n pt = Point(time, lat, lon, alt)\n pt.run_msis()\n msis = {}\n total = 0\n for key in pt.nn.keys():\n total += pt.nn[key]\n msis['Nn'] = total\n msis['Nn_N'] = pt.nn['N']\n msis['Nn_N2'] = pt.nn['N2']\n msis['Nn_O'] = pt.nn['O']\n msis['Nn_O2'] = pt.nn['O2']\n msis['Tn_msis'] = pt.Tn_msis\n msis_params.append(msis)\n msis = pds.DataFrame(msis_params)\n msis.index = inst.data.index\n inst[msis.keys()] = msis\n inst.meta['Nn'] = {'units': 'cm^-3', 'desc':\n 'Total neutral number particle density from MSIS.'}\n inst.meta['Nn_N'] = {'units': 'cm^-3', 'desc':\n 'Total nitrogen number particle density from MSIS.'}\n inst.meta['Nn_N2'] = {'units': 'cm^-3', 'desc':\n 'Total N2 number particle density from MSIS.'}\n inst.meta['Nn_O'] = {'units': 'cm^-3', 'desc':\n 'Total oxygen number particle density from MSIS.'}\n inst.meta['Nn_O2'] = {'units': 'cm^-3', 'desc':\n 'Total O2 number particle density from MSIS.'}\n inst.meta['Tn_msis'] = {'units': 'K', 'desc':\n 'Neutral temperature from MSIS.'}\n return\n", "role": "user" }, { "content": "add_msis", "role": "assistant" } ]
def x(inst, x_label, y_label, z_label, new_x_label, new_y_label, new_z_label, meta=None): import pysatMagVect x, y, z = pysatMagVect.project_ecef_vector_onto_basis(inst[x_label], inst[y_label], inst[z_label], inst['sc_xhat_ecef_x'], inst[ 'sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'], inst['sc_yhat_ecef_x'], inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'], inst[ 'sc_zhat_ecef_x'], inst['sc_zhat_ecef_y'], inst['sc_zhat_ecef_z']) inst[new_x_label] = x inst[new_y_label] = y inst[new_z_label] = z if meta is not None: inst.meta[new_x_label] = meta[0] inst.meta[new_y_label] = meta[1] inst.meta[new_z_label] = meta[2] return
project_ecef_vector_onto_sc
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(inst, x_label, y_label, z_label, new_x_label, new_y_label,\n new_z_label, meta=None):\n import pysatMagVect\n x, y, z = pysatMagVect.project_ecef_vector_onto_basis(inst[x_label],\n inst[y_label], inst[z_label], inst['sc_xhat_ecef_x'], inst[\n 'sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'], inst['sc_yhat_ecef_x'],\n inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'], inst[\n 'sc_zhat_ecef_x'], inst['sc_zhat_ecef_y'], inst['sc_zhat_ecef_z'])\n inst[new_x_label] = x\n inst[new_y_label] = y\n inst[new_z_label] = z\n if meta is not None:\n inst.meta[new_x_label] = meta[0]\n inst.meta[new_y_label] = meta[1]\n inst.meta[new_z_label] = meta[2]\n return\n", "role": "user" }, { "content": "project_ecef_vector_onto_sc", "role": "assistant" } ]
def x(inst, labelx, labely, data_label, datalim, xlim=None, ylim=None): plt.ioff() figs = [] axs = [] if type(data_label) is str: figs.append(plt.figure()) ax1 = figs[0].add_subplot(211, projection='3d') ax2 = figs[0].add_subplot(212) axs.append((ax1, ax2)) plt.suptitle(data_label) if xlim is not None: ax1.set_xlim(xlim) ax2.set_xlim(xlim) if ylim is not None: ax1.set_ylim(ylim) ax2.set_ylim(ylim) data_label = [data_label] plt.hold(True) else: for i in np.arange(len(data_label)): figs.append(plt.figure()) ax1 = figs[i].add_subplot(211, projection='3d') ax2 = figs[i].add_subplot(212) axs.append((ax1, ax2)) plt.suptitle(data_label[i]) if xlim is not None: ax1.set_xlim(xlim) ax2.set_xlim(xlim) if ylim is not None: ax1.set_ylim(ylim) ax2.set_ylim(ylim) plt.hold(True) norm = colors.Normalize(vmin=datalim[0], vmax=datalim[1]) p = [i for i in np.arange(len(figs))] for i, inst in enumerate(inst): for j, (fig, ax) in enumerate(zip(figs, axs)): if len(inst.data) > 0: if (len(inst.data[labelx]) > 0) & (len(inst.data[labely]) > 0 ) & (len(inst.data[data_label[j]]) > 0): p[j] = ax[0].scatter(inst.data[labelx], inst.data[ labely], inst.data[data_label[j]], zdir='z', c=inst .data[data_label[j]], cmap=cm.jet, norm=norm, linewidth='0', edgecolors=None) ax[1].scatter(inst.data[labelx], inst.data[labely], c= inst.data[data_label[j]], cmap=cm.jet, norm=norm, linewidth=1e-11, alpha=0.5, edgecolor=None) for j, (fig, ax) in enumerate(zip(figs, axs)): try: cbar = plt.colorbar(p[j], ax=ax[0], label='Amplitude (m/s)') except: print('Tried colorbar but failed, thus no colorbar.') ax[0].elev = 30.0 plt.ion() return figs
scatterplot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(inst, labelx, labely, data_label, datalim, xlim=None, ylim=None):\n plt.ioff()\n figs = []\n axs = []\n if type(data_label) is str:\n figs.append(plt.figure())\n ax1 = figs[0].add_subplot(211, projection='3d')\n ax2 = figs[0].add_subplot(212)\n axs.append((ax1, ax2))\n plt.suptitle(data_label)\n if xlim is not None:\n ax1.set_xlim(xlim)\n ax2.set_xlim(xlim)\n if ylim is not None:\n ax1.set_ylim(ylim)\n ax2.set_ylim(ylim)\n data_label = [data_label]\n plt.hold(True)\n else:\n for i in np.arange(len(data_label)):\n figs.append(plt.figure())\n ax1 = figs[i].add_subplot(211, projection='3d')\n ax2 = figs[i].add_subplot(212)\n axs.append((ax1, ax2))\n plt.suptitle(data_label[i])\n if xlim is not None:\n ax1.set_xlim(xlim)\n ax2.set_xlim(xlim)\n if ylim is not None:\n ax1.set_ylim(ylim)\n ax2.set_ylim(ylim)\n plt.hold(True)\n norm = colors.Normalize(vmin=datalim[0], vmax=datalim[1])\n p = [i for i in np.arange(len(figs))]\n for i, inst in enumerate(inst):\n for j, (fig, ax) in enumerate(zip(figs, axs)):\n if len(inst.data) > 0:\n if (len(inst.data[labelx]) > 0) & (len(inst.data[labely]) > 0\n ) & (len(inst.data[data_label[j]]) > 0):\n p[j] = ax[0].scatter(inst.data[labelx], inst.data[\n labely], inst.data[data_label[j]], zdir='z', c=inst\n .data[data_label[j]], cmap=cm.jet, norm=norm,\n linewidth='0', edgecolors=None)\n ax[1].scatter(inst.data[labelx], inst.data[labely], c=\n inst.data[data_label[j]], cmap=cm.jet, norm=norm,\n linewidth=1e-11, alpha=0.5, edgecolor=None)\n for j, (fig, ax) in enumerate(zip(figs, axs)):\n try:\n cbar = plt.colorbar(p[j], ax=ax[0], label='Amplitude (m/s)')\n except:\n print('Tried colorbar but failed, thus no colorbar.')\n ax[0].elev = 30.0\n plt.ion()\n return figs\n", "role": "user" }, { "content": "scatterplot", "role": "assistant" } ]
def x(level=logging.DEBUG): logger = logging.getLogger(__name__) logger.setLevel(level) if not logger.handlers: logger.info('enabling logging to console.') logger.addHandler(logging.StreamHandler(sys.stdout))
enable_log
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(level=logging.DEBUG):\n logger = logging.getLogger(__name__)\n logger.setLevel(level)\n if not logger.handlers:\n logger.info('enabling logging to console.')\n logger.addHandler(logging.StreamHandler(sys.stdout))\n", "role": "user" }, { "content": "enable_log", "role": "assistant" } ]
def x(tag=None, sat_id=None, data_path=None, format_str=None): import sys estr = 'Building a list of COSMIC files, which can possibly take time. ' estr = '{:s}~1s per 100K files'.format(estr) print(estr) sys.stdout.flush() cosmicFiles = glob.glob(os.path.join(data_path, '*/*_nc')) num = len(cosmicFiles) if num != 0: print('Estimated time:', num * 1e-05, 'seconds') sys.stdout.flush() year = [None] * num days = [None] * num hours = [None] * num minutes = [None] * num microseconds = [None] * num for i, f in enumerate(cosmicFiles): f2 = f.split('.') year[i] = f2[-6] days[i] = f2[-5] hours[i] = f2[-4] minutes[i] = f2[-3] microseconds[i] = i year = np.array(year).astype(int) days = np.array(days).astype(int) uts = np.array(hours).astype(int) * 3600.0 + np.array(minutes).astype( int) * 60.0 uts += np.mod(np.array(microseconds).astype(int) * 4, 8000) * 1e-05 index = pysat.utils.create_datetime_index(year=year, day=days, uts=uts) file_list = pysat.Series(cosmicFiles, index=index) return file_list else: print('Found no files, check your path or download them.') return pysat.Series(None)
list_files
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(tag=None, sat_id=None, data_path=None, format_str=None):\n import sys\n estr = 'Building a list of COSMIC files, which can possibly take time. '\n estr = '{:s}~1s per 100K files'.format(estr)\n print(estr)\n sys.stdout.flush()\n cosmicFiles = glob.glob(os.path.join(data_path, '*/*_nc'))\n num = len(cosmicFiles)\n if num != 0:\n print('Estimated time:', num * 1e-05, 'seconds')\n sys.stdout.flush()\n year = [None] * num\n days = [None] * num\n hours = [None] * num\n minutes = [None] * num\n microseconds = [None] * num\n for i, f in enumerate(cosmicFiles):\n f2 = f.split('.')\n year[i] = f2[-6]\n days[i] = f2[-5]\n hours[i] = f2[-4]\n minutes[i] = f2[-3]\n microseconds[i] = i\n year = np.array(year).astype(int)\n days = np.array(days).astype(int)\n uts = np.array(hours).astype(int) * 3600.0 + np.array(minutes).astype(\n int) * 60.0\n uts += np.mod(np.array(microseconds).astype(int) * 4, 8000) * 1e-05\n index = pysat.utils.create_datetime_index(year=year, day=days, uts=uts)\n file_list = pysat.Series(cosmicFiles, index=index)\n return file_list\n else:\n print('Found no files, check your path or download them.')\n return pysat.Series(None)\n", "role": "user" }, { "content": "list_files", "role": "assistant" } ]
def x(files, tag=None, sat_id=None, altitude_bin=None): output = [None] * len(files) drop_idx = [] for i, file in enumerate(files): try: data = netcdf_file(file, mode='r', mmap=False) new = {} ncattrsList = data._attributes.keys() for d in ncattrsList: new[d] = data._attributes[d] loadedVars = {} keys = data.variables.keys() for key in keys: if data.variables[key][:].dtype.byteorder != '=': loadedVars[key] = data.variables[key][:].byteswap( ).newbyteorder() else: loadedVars[key] = data.variables[key][:] new['profiles'] = pysat.DataFrame(loadedVars) output[i] = new data.close() except RuntimeError: drop_idx.append(i) drop_idx.reverse() for i in drop_idx: del output[i] if tag == 'ionprf': if altitude_bin is not None: for out in output: out['profiles'].index = (out['profiles']['MSL_alt'] / altitude_bin).round().values * altitude_bin out['profiles'] = out['profiles'].groupby(out['profiles']. index.values).mean() else: for out in output: out['profiles'].index = out['profiles']['MSL_alt'] return output
load_files
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(files, tag=None, sat_id=None, altitude_bin=None):\n output = [None] * len(files)\n drop_idx = []\n for i, file in enumerate(files):\n try:\n data = netcdf_file(file, mode='r', mmap=False)\n new = {}\n ncattrsList = data._attributes.keys()\n for d in ncattrsList:\n new[d] = data._attributes[d]\n loadedVars = {}\n keys = data.variables.keys()\n for key in keys:\n if data.variables[key][:].dtype.byteorder != '=':\n loadedVars[key] = data.variables[key][:].byteswap(\n ).newbyteorder()\n else:\n loadedVars[key] = data.variables[key][:]\n new['profiles'] = pysat.DataFrame(loadedVars)\n output[i] = new\n data.close()\n except RuntimeError:\n drop_idx.append(i)\n drop_idx.reverse()\n for i in drop_idx:\n del output[i]\n if tag == 'ionprf':\n if altitude_bin is not None:\n for out in output:\n out['profiles'].index = (out['profiles']['MSL_alt'] /\n altitude_bin).round().values * altitude_bin\n out['profiles'] = out['profiles'].groupby(out['profiles'].\n index.values).mean()\n else:\n for out in output:\n out['profiles'].index = out['profiles']['MSL_alt']\n return output\n", "role": "user" }, { "content": "load_files", "role": "assistant" } ]
def x(value): t = 60 if value is not None: r = value % t if r > t / 2: ret = value + (t - r) else: ret = value - r else: ret = NaN return ret
round_60
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(value):\n t = 60\n if value is not None:\n r = value % t\n if r > t / 2:\n ret = value + (t - r)\n else:\n ret = value - r\n else:\n ret = NaN\n return ret\n", "role": "user" }, { "content": "round_60", "role": "assistant" } ]
def x(prev, curr, counters): busy_prop, idle_prop = counters pb = getattr(prev, busy_prop) pi = getattr(prev, idle_prop) cb = getattr(curr, busy_prop) ci = getattr(curr, idle_prop) db = minus(cb, pb) di = minus(ci, pi) return mul(div(db, add(db, di)), 100)
utilization
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(prev, curr, counters):\n busy_prop, idle_prop = counters\n pb = getattr(prev, busy_prop)\n pi = getattr(prev, idle_prop)\n cb = getattr(curr, busy_prop)\n ci = getattr(curr, idle_prop)\n db = minus(cb, pb)\n di = minus(ci, pi)\n return mul(div(db, add(db, di)), 100)\n", "role": "user" }, { "content": "utilization", "role": "assistant" } ]
def x(prev, curr, counters): counter = get_counter(counters) pv = getattr(prev, counter) cv = getattr(curr, counter) return minus(cv, pv)
delta_ps
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(prev, curr, counters):\n counter = get_counter(counters)\n pv = getattr(prev, counter)\n cv = getattr(curr, counter)\n return minus(cv, pv)\n", "role": "user" }, { "content": "delta_ps", "role": "assistant" } ]
def x(prev, curr, counters): bw_stats, io_stats = counters size_mb = div(getattr(curr, bw_stats), getattr(curr, io_stats)) return mul(size_mb, 1024)
io_size_kb
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(prev, curr, counters):\n bw_stats, io_stats = counters\n size_mb = div(getattr(curr, bw_stats), getattr(curr, io_stats))\n return mul(size_mb, 1024)\n", "role": "user" }, { "content": "io_size_kb", "role": "assistant" } ]
def x(tag='', sat_id=None, data_path=None, format_str=None): if format_str is None and tag is not None: if tag == '' or tag == 'ascii': ascii_fmt = 'Density_3deg_{year:02d}_{doy:03d}.ascii' return pysat.Files.from_os(data_path=data_path, format_str= ascii_fmt) else: raise ValueError('Unrecognized tag name for CHAMP STAR') elif format_str is None: estr = 'A tag name must be passed to the loading routine for CHAMP' raise ValueError(estr) else: return pysat.Files.from_os(data_path=data_path, format_str=format_str)
list_files
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(tag='', sat_id=None, data_path=None, format_str=None):\n if format_str is None and tag is not None:\n if tag == '' or tag == 'ascii':\n ascii_fmt = 'Density_3deg_{year:02d}_{doy:03d}.ascii'\n return pysat.Files.from_os(data_path=data_path, format_str=\n ascii_fmt)\n else:\n raise ValueError('Unrecognized tag name for CHAMP STAR')\n elif format_str is None:\n estr = 'A tag name must be passed to the loading routine for CHAMP'\n raise ValueError(estr)\n else:\n return pysat.Files.from_os(data_path=data_path, format_str=format_str)\n", "role": "user" }, { "content": "list_files", "role": "assistant" } ]
def x(fnames, tag=None, sat_id=None): import re if len(fnames) <= 0: return pysat.DataFrame(None), pysat.Meta(None) if isinstance(fnames, str): fnames = [fnames] champ_labels = {'Two-digit Year (years)': 'year', 'Day of the Year (days)': 'doy', 'Second of the Day (GPS time,sec)': 'sod', 'Center Latitude of 3-degree Bin (deg)': 'bin_lat', 'Satellite Geodetic Latitude (deg)': 'sat_glat', 'Satellite Longitude (deg)': 'sat_lon', 'Satellite Height (km)': 'sat_h', 'Satellite Local Time (hours)': 'sat_lt', 'Satellite Quasi-Dipole Latitude (deg)': 'sat_qdlat', 'Satellite Magnetic Longitude (deg)': 'sat_mlon', 'Satellite Magnetic Local Time (hours)': 'sat_mlt', 'Neutral Density (kg/m^3)': 'ndens', 'Neutral Density Normalized to 400km using NRLMSISe00': 'ndens400', 'Neutral Density Normalized to 410km using NRLMSISe00': 'ndens410', 'NRLMSISe00 Neutral Density at Satellite Height': 'nrlmsis_ndens', 'Uncertainty in Neutral Density (kg/m^3)': 'ndens_err', 'Number of Data Points in Current Averaging Bin': 'npnts', 'Number of Points in Current Averaging Bin that Required Interpolation' : 'npnts_interp', 'Average Coefficient of Drag Used in Current Averaging Bin': 'avg_drag_coeff'} champ_dtypes = {'year': np.int32, 'doy': np.int32, 'sod': float, 'bin_lat': float, 'sat_glat': float, 'sat_lon': float, 'sat_h': float, 'sat_lt': float, 'sat_qdlat': float, 'sat_mlon': float, 'sat_mlt': float, 'ndens': float, 'ndens400': float, 'ndens410': float, 'nrlmsis_ndens': float, 'ndens_err': float, 'npnts': int, 'npnts_interp': float, 'avg_drag_coeff': float} champ_units = {'year': '2-digit years', 'doy': 'day of year', 'sod': 'seconds of day', 'bin_lat': 'degrees', 'sat_glat': 'degrees', 'sat_lon': 'degrees', 'sat_h': 'km', 'sat_lt': 'hours', 'sat_qdlat': 'degrees', 'sat_mlon': 'degrees', 'sat_mlt': 'hours', 'ndens': 'km m^{-3}', 'ndens400': 'km m^{-3}', 'ndens410': 'km m^{-3}', 'nrlmsis_ndens': 'km m^{-3}', 'ndens_err': 'km m^{-3}', 'npnts': 'number', 'npnts_interp': 'number', 'avg_drag_coeff': 'unitless'} def x(y, d, s): import datetime as dt t = dt.datetime.strptime('{:02d} {:03d}'.format(int(y), int(d)), '%y %j') fsec = float(s) isec = np.floor(fsec) microsec = (fsec - isec) * 1000000.0 t += dt.timedelta(seconds=isec, microseconds=microsec) return t f = open(fnames[0], 'r') vdata = re.split(';|\n', f.readline()) hdata = re.split(';|\n', f.readline()) f.close() try: hdata.pop(hdata.index('')) except: pass data = pds.read_csv(fnames[0], delim_whitespace=True, skiprows=2, header=None, names=[champ_labels[h] for h in hdata], keep_date_col= True, index_col='datetime', parse_dates={'datetime': [0, 1, 2]}, date_parser=parse_champdate) meta = pysat.Meta() for h in hdata: col = champ_labels[h] data[col].astype(champ_dtypes[col]) meta[col] = {'units': champ_units[col], 'long_name': h} return data, meta
load
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fnames, tag=None, sat_id=None):\n import re\n if len(fnames) <= 0:\n return pysat.DataFrame(None), pysat.Meta(None)\n if isinstance(fnames, str):\n fnames = [fnames]\n champ_labels = {'Two-digit Year (years)': 'year',\n 'Day of the Year (days)': 'doy', 'Second of the Day (GPS time,sec)':\n 'sod', 'Center Latitude of 3-degree Bin (deg)': 'bin_lat',\n 'Satellite Geodetic Latitude (deg)': 'sat_glat',\n 'Satellite Longitude (deg)': 'sat_lon', 'Satellite Height (km)':\n 'sat_h', 'Satellite Local Time (hours)': 'sat_lt',\n 'Satellite Quasi-Dipole Latitude (deg)': 'sat_qdlat',\n 'Satellite Magnetic Longitude (deg)': 'sat_mlon',\n 'Satellite Magnetic Local Time (hours)': 'sat_mlt',\n 'Neutral Density (kg/m^3)': 'ndens',\n 'Neutral Density Normalized to 400km using NRLMSISe00': 'ndens400',\n 'Neutral Density Normalized to 410km using NRLMSISe00': 'ndens410',\n 'NRLMSISe00 Neutral Density at Satellite Height': 'nrlmsis_ndens',\n 'Uncertainty in Neutral Density (kg/m^3)': 'ndens_err',\n 'Number of Data Points in Current Averaging Bin': 'npnts',\n 'Number of Points in Current Averaging Bin that Required Interpolation'\n : 'npnts_interp',\n 'Average Coefficient of Drag Used in Current Averaging Bin':\n 'avg_drag_coeff'}\n champ_dtypes = {'year': np.int32, 'doy': np.int32, 'sod': float,\n 'bin_lat': float, 'sat_glat': float, 'sat_lon': float, 'sat_h':\n float, 'sat_lt': float, 'sat_qdlat': float, 'sat_mlon': float,\n 'sat_mlt': float, 'ndens': float, 'ndens400': float, 'ndens410':\n float, 'nrlmsis_ndens': float, 'ndens_err': float, 'npnts': int,\n 'npnts_interp': float, 'avg_drag_coeff': float}\n champ_units = {'year': '2-digit years', 'doy': 'day of year', 'sod':\n 'seconds of day', 'bin_lat': 'degrees', 'sat_glat': 'degrees',\n 'sat_lon': 'degrees', 'sat_h': 'km', 'sat_lt': 'hours', 'sat_qdlat':\n 'degrees', 'sat_mlon': 'degrees', 'sat_mlt': 'hours', 'ndens':\n 'km m^{-3}', 'ndens400': 'km m^{-3}', 'ndens410': 'km m^{-3}',\n 'nrlmsis_ndens': 'km m^{-3}', 'ndens_err': 'km m^{-3}', 'npnts':\n 'number', 'npnts_interp': 'number', 'avg_drag_coeff': 'unitless'}\n\n def x(y, d, s):\n import datetime as dt\n t = dt.datetime.strptime('{:02d} {:03d}'.format(int(y), int(d)),\n '%y %j')\n fsec = float(s)\n isec = np.floor(fsec)\n microsec = (fsec - isec) * 1000000.0\n t += dt.timedelta(seconds=isec, microseconds=microsec)\n return t\n f = open(fnames[0], 'r')\n vdata = re.split(';|\\n', f.readline())\n hdata = re.split(';|\\n', f.readline())\n f.close()\n try:\n hdata.pop(hdata.index(''))\n except:\n pass\n data = pds.read_csv(fnames[0], delim_whitespace=True, skiprows=2,\n header=None, names=[champ_labels[h] for h in hdata], keep_date_col=\n True, index_col='datetime', parse_dates={'datetime': [0, 1, 2]},\n date_parser=parse_champdate)\n meta = pysat.Meta()\n for h in hdata:\n col = champ_labels[h]\n data[col].astype(champ_dtypes[col])\n meta[col] = {'units': champ_units[col], 'long_name': h}\n return data, meta\n", "role": "user" }, { "content": "load", "role": "assistant" } ]
def x(self, by_name=False, inst_module=None): import importlib self._list_rtn = self._pass_func self._load_rtn = self._pass_func self._default_rtn = self._pass_func self._clean_rtn = self._pass_func self._init_rtn = self._pass_func self._download_rtn = self._pass_func self.directory_format = None self.file_format = None self.multi_file_day = False self.orbit_info = None if by_name: inst = importlib.import_module(''.join(('.', self.platform, '_', self.name)), package='pysat.instruments') elif inst_module is not None: inst = inst_module else: return try: self._load_rtn = inst.load self._list_rtn = inst.list_files self._download_rtn = inst.download except AttributeError: estr = 'A load, file_list, and download routine are required for ' raise AttributeError('{:s}every instrument.'.format(estr)) try: self._default_rtn = inst.default except AttributeError: pass try: self._init_rtn = inst.init except AttributeError: pass try: self._clean_rtn = inst.clean except AttributeError: pass try: self.directory_format = inst.directory_format except AttributeError: pass try: self.multi_file_day = inst.multi_file_day except AttributeError: pass try: self.orbit_info = inst.orbit_info except AttributeError: pass return
assign_funcs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, by_name=False, inst_module=None):\n import importlib\n self._list_rtn = self._pass_func\n self._load_rtn = self._pass_func\n self._default_rtn = self._pass_func\n self._clean_rtn = self._pass_func\n self._init_rtn = self._pass_func\n self._download_rtn = self._pass_func\n self.directory_format = None\n self.file_format = None\n self.multi_file_day = False\n self.orbit_info = None\n if by_name:\n inst = importlib.import_module(''.join(('.', self.platform, '_',\n self.name)), package='pysat.instruments')\n elif inst_module is not None:\n inst = inst_module\n else:\n return\n try:\n self._load_rtn = inst.load\n self._list_rtn = inst.list_files\n self._download_rtn = inst.download\n except AttributeError:\n estr = 'A load, file_list, and download routine are required for '\n raise AttributeError('{:s}every instrument.'.format(estr))\n try:\n self._default_rtn = inst.default\n except AttributeError:\n pass\n try:\n self._init_rtn = inst.init\n except AttributeError:\n pass\n try:\n self._clean_rtn = inst.clean\n except AttributeError:\n pass\n try:\n self.directory_format = inst.directory_format\n except AttributeError:\n pass\n try:\n self.multi_file_day = inst.multi_file_day\n except AttributeError:\n pass\n try:\n self.orbit_info = inst.orbit_info\n except AttributeError:\n pass\n return\n", "role": "user" }, { "content": "assign_funcs", "role": "assistant" } ]
def x(self, date=None, fid=None): if fid is not None: fname = self.files[fid:fid + 1] elif date is not None: fname = self.files[date:date + pds.DateOffset(days=1)] else: raise ValueError('Must supply either a date or file id number.') if len(fname) > 0: load_fname = [os.path.join(self.files.data_path, f) for f in fname] data, mdata = self._load_rtn(load_fname, tag=self.tag, sat_id=self. sat_id, **self.kwargs) mdata.accept_default_labels(self) else: data = DataFrame(None) mdata = _meta.Meta(units_label=self.units_label, name_label=self. name_label, notes_label=self.notes_label, desc_label=self. desc_label, plot_label=self.plot_label, axis_label=self. axis_label, scale_label=self.scale_label, min_label=self. min_label, max_label=self.max_label, fill_label=self.fill_label) output_str = '{platform} {name} {tag} {sat_id}' output_str = output_str.format(platform=self.platform, name=self.name, tag=self.tag, sat_id=self.sat_id) if not data.empty: if not isinstance(data, DataFrame): raise TypeError(' '.join(('Data returned by instrument load', 'routine must be a pandas.DataFrame'))) if not isinstance(mdata, _meta.Meta): raise TypeError('Metadata returned must be a pysat.Meta object') if date is not None: output_str = ' '.join(('Returning', output_str, 'data for', date.strftime('%x'))) elif len(fname) == 1: output_str = ' '.join(('Returning', output_str, 'data from', fname[0])) else: output_str = ' '.join(('Returning', output_str, 'data from', fname[0], '::', fname[-1])) else: output_str = ' '.join(('No', output_str, 'data for', date.strftime( '%m/%d/%y'))) output_str = ' '.join(output_str.split()) print(output_str) return data, mdata
load_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, date=None, fid=None):\n if fid is not None:\n fname = self.files[fid:fid + 1]\n elif date is not None:\n fname = self.files[date:date + pds.DateOffset(days=1)]\n else:\n raise ValueError('Must supply either a date or file id number.')\n if len(fname) > 0:\n load_fname = [os.path.join(self.files.data_path, f) for f in fname]\n data, mdata = self._load_rtn(load_fname, tag=self.tag, sat_id=self.\n sat_id, **self.kwargs)\n mdata.accept_default_labels(self)\n else:\n data = DataFrame(None)\n mdata = _meta.Meta(units_label=self.units_label, name_label=self.\n name_label, notes_label=self.notes_label, desc_label=self.\n desc_label, plot_label=self.plot_label, axis_label=self.\n axis_label, scale_label=self.scale_label, min_label=self.\n min_label, max_label=self.max_label, fill_label=self.fill_label)\n output_str = '{platform} {name} {tag} {sat_id}'\n output_str = output_str.format(platform=self.platform, name=self.name,\n tag=self.tag, sat_id=self.sat_id)\n if not data.empty:\n if not isinstance(data, DataFrame):\n raise TypeError(' '.join(('Data returned by instrument load',\n 'routine must be a pandas.DataFrame')))\n if not isinstance(mdata, _meta.Meta):\n raise TypeError('Metadata returned must be a pysat.Meta object')\n if date is not None:\n output_str = ' '.join(('Returning', output_str, 'data for',\n date.strftime('%x')))\n elif len(fname) == 1:\n output_str = ' '.join(('Returning', output_str, 'data from',\n fname[0]))\n else:\n output_str = ' '.join(('Returning', output_str, 'data from',\n fname[0], '::', fname[-1]))\n else:\n output_str = ' '.join(('No', output_str, 'data for', date.strftime(\n '%m/%d/%y')))\n output_str = ' '.join(output_str.split())\n print(output_str)\n return data, mdata\n", "role": "user" }, { "content": "load_data", "role": "assistant" } ]
def x(self): if self._load_by_date: next_date = self.date + pds.DateOffset(days=1) return self._load_data(date=next_date) else: return self._load_data(fid=self._fid + 1)
load_next
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self._load_by_date:\n next_date = self.date + pds.DateOffset(days=1)\n return self._load_data(date=next_date)\n else:\n return self._load_data(fid=self._fid + 1)\n", "role": "user" }, { "content": "load_next", "role": "assistant" } ]
def x(self): if self._load_by_date: prev_date = self.date - pds.DateOffset(days=1) return self._load_data(date=prev_date) else: return self._load_data(fid=self._fid - 1)
load_prev
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self._load_by_date:\n prev_date = self.date - pds.DateOffset(days=1)\n return self._load_data(date=prev_date)\n else:\n return self._load_data(fid=self._fid - 1)\n", "role": "user" }, { "content": "load_prev", "role": "assistant" } ]
def x(self, start, stop, freq='D', user=None, password=None, **kwargs): import errno try: os.makedirs(self.files.data_path) except OSError as e: if e.errno != errno.EEXIST: raise print('Downloading data to: ', self.files.data_path) date_array = utils.season_date_range(start, stop, freq=freq) if user is None: self._download_rtn(date_array, tag=self.tag, sat_id=self.sat_id, data_path=self.files.data_path, **kwargs) else: self._download_rtn(date_array, tag=self.tag, sat_id=self.sat_id, data_path=self.files.data_path, user=user, password=password, **kwargs) first_date = self.files.start_date last_date = self.files.stop_date print('Updating pysat file list') self.files.refresh() if len(self.bounds[0]) == 1: if self.bounds[0][0] == first_date and self.bounds[1][0] == last_date: print('Updating instrument object bounds.') self.bounds = None
download
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, start, stop, freq='D', user=None, password=None, **kwargs):\n import errno\n try:\n os.makedirs(self.files.data_path)\n except OSError as e:\n if e.errno != errno.EEXIST:\n raise\n print('Downloading data to: ', self.files.data_path)\n date_array = utils.season_date_range(start, stop, freq=freq)\n if user is None:\n self._download_rtn(date_array, tag=self.tag, sat_id=self.sat_id,\n data_path=self.files.data_path, **kwargs)\n else:\n self._download_rtn(date_array, tag=self.tag, sat_id=self.sat_id,\n data_path=self.files.data_path, user=user, password=password,\n **kwargs)\n first_date = self.files.start_date\n last_date = self.files.stop_date\n print('Updating pysat file list')\n self.files.refresh()\n if len(self.bounds[0]) == 1:\n if self.bounds[0][0] == first_date and self.bounds[1][0] == last_date:\n print('Updating instrument object bounds.')\n self.bounds = None\n", "role": "user" }, { "content": "download", "role": "assistant" } ]
def x(self, verifyPad=False): if self._iter_type == 'date': if self.date is not None: idx, = np.where(self._iter_list == self.date) if len(idx) == 0: raise StopIteration('File list is empty. Nothing to be done.') elif idx[-1] + 1 >= len(self._iter_list): raise StopIteration('Outside the set date boundaries.') else: idx += 1 self.load(date=self._iter_list[idx[0]], verifyPad=verifyPad) else: self.load(date=self._iter_list[0], verifyPad=verifyPad) elif self._iter_type == 'file': if self._fid is not None: first = self.files.get_index(self._iter_list[0]) last = self.files.get_index(self._iter_list[-1]) if (self._fid < first) | (self._fid + 1 > last): raise StopIteration('Outside the set file boundaries.') else: self.load(fname=self._iter_list[self._fid + 1 - first], verifyPad=verifyPad) else: self.load(fname=self._iter_list[0], verifyPad=verifyPad)
next
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, verifyPad=False):\n if self._iter_type == 'date':\n if self.date is not None:\n idx, = np.where(self._iter_list == self.date)\n if len(idx) == 0:\n raise StopIteration('File list is empty. Nothing to be done.')\n elif idx[-1] + 1 >= len(self._iter_list):\n raise StopIteration('Outside the set date boundaries.')\n else:\n idx += 1\n self.load(date=self._iter_list[idx[0]], verifyPad=verifyPad)\n else:\n self.load(date=self._iter_list[0], verifyPad=verifyPad)\n elif self._iter_type == 'file':\n if self._fid is not None:\n first = self.files.get_index(self._iter_list[0])\n last = self.files.get_index(self._iter_list[-1])\n if (self._fid < first) | (self._fid + 1 > last):\n raise StopIteration('Outside the set file boundaries.')\n else:\n self.load(fname=self._iter_list[self._fid + 1 - first],\n verifyPad=verifyPad)\n else:\n self.load(fname=self._iter_list[0], verifyPad=verifyPad)\n", "role": "user" }, { "content": "next", "role": "assistant" } ]
def x(self, coltype): if type(coltype) is np.dtype: var_type = coltype.kind + str(coltype.itemsize) return var_type elif coltype is np.int64: return 'i8' elif coltype is np.int32: return 'i4' elif coltype is np.int16: return 'i2' elif coltype is np.int8: return 'i1' elif coltype is np.uint64: return 'u8' elif coltype is np.uint32: return 'u4' elif coltype is np.uint16: return 'u2' elif coltype is np.uint8: return 'u1' elif coltype is np.float64: return 'f8' elif coltype is np.float32: return 'f4' elif issubclass(coltype, basestring): return 'S1' else: raise TypeError('Unknown Variable Type' + str(coltype))
get_var_type_code
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, coltype):\n if type(coltype) is np.dtype:\n var_type = coltype.kind + str(coltype.itemsize)\n return var_type\n elif coltype is np.int64:\n return 'i8'\n elif coltype is np.int32:\n return 'i4'\n elif coltype is np.int16:\n return 'i2'\n elif coltype is np.int8:\n return 'i1'\n elif coltype is np.uint64:\n return 'u8'\n elif coltype is np.uint32:\n return 'u4'\n elif coltype is np.uint16:\n return 'u2'\n elif coltype is np.uint8:\n return 'u1'\n elif coltype is np.float64:\n return 'f8'\n elif coltype is np.float32:\n return 'f4'\n elif issubclass(coltype, basestring):\n return 'S1'\n else:\n raise TypeError('Unknown Variable Type' + str(coltype))\n", "role": "user" }, { "content": "get_var_type_code", "role": "assistant" } ]
def x(self, data, file_format): data_type = data.dtype if file_format != 'NETCDF4': old_format = True else: old_format = False if data_type != np.dtype('O'): if (data_type == np.int64) & old_format: data = data.astype(np.int32) data_type = np.int32 if data_type == np.dtype('<M8[ns]'): if not old_format: data_type = np.int64 else: data_type = np.float datetime_flag = True else: datetime_flag = False else: data_type = type(data.iloc[0]) for i in np.arange(len(data)): if len(data.iloc[i]) > 0: data_type = type(data.iloc[i]) if not isinstance(data_type, np.float): break datetime_flag = False return data, data_type, datetime_flag
get_data_info
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data, file_format):\n data_type = data.dtype\n if file_format != 'NETCDF4':\n old_format = True\n else:\n old_format = False\n if data_type != np.dtype('O'):\n if (data_type == np.int64) & old_format:\n data = data.astype(np.int32)\n data_type = np.int32\n if data_type == np.dtype('<M8[ns]'):\n if not old_format:\n data_type = np.int64\n else:\n data_type = np.float\n datetime_flag = True\n else:\n datetime_flag = False\n else:\n data_type = type(data.iloc[0])\n for i in np.arange(len(data)):\n if len(data.iloc[i]) > 0:\n data_type = type(data.iloc[i])\n if not isinstance(data_type, np.float):\n break\n datetime_flag = False\n return data, data_type, datetime_flag\n", "role": "user" }, { "content": "get_data_info", "role": "assistant" } ]
def x(self, mdata_dict, coltype, remove=False): for key in mdata_dict: if type(mdata_dict[key]) == bool: mdata_dict[key] = int(mdata_dict[key]) if coltype == type(' ') or coltype == type(u' '): remove = True if u'_FillValue' in mdata_dict.keys(): if remove: mdata_dict.pop('_FillValue') else: mdata_dict['_FillValue'] = np.array(mdata_dict['_FillValue'] ).astype(coltype) if u'FillVal' in mdata_dict.keys(): if remove: mdata_dict.pop('FillVal') else: mdata_dict['FillVal'] = np.array(mdata_dict['FillVal']).astype( coltype) return mdata_dict
filter_netcdf4_metadata
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, mdata_dict, coltype, remove=False):\n for key in mdata_dict:\n if type(mdata_dict[key]) == bool:\n mdata_dict[key] = int(mdata_dict[key])\n if coltype == type(' ') or coltype == type(u' '):\n remove = True\n if u'_FillValue' in mdata_dict.keys():\n if remove:\n mdata_dict.pop('_FillValue')\n else:\n mdata_dict['_FillValue'] = np.array(mdata_dict['_FillValue']\n ).astype(coltype)\n if u'FillVal' in mdata_dict.keys():\n if remove:\n mdata_dict.pop('FillVal')\n else:\n mdata_dict['FillVal'] = np.array(mdata_dict['FillVal']).astype(\n coltype)\n return mdata_dict\n", "role": "user" }, { "content": "filter_netcdf4_metadata", "role": "assistant" } ]
def x(self, meta_to_translate): export_dict = {} if self._meta_translation_table is not None: translation_table = {} for key in self._meta_translation_table: translation_table[getattr(self, key) ] = self._meta_translation_table[key] else: translation_table = None for key in meta_to_translate.data.index: if translation_table is None: export_dict[key] = meta_to_translate.data.loc[key].to_dict() else: export_dict[key] = {} meta_dict = meta_to_translate.data.loc[key].to_dict() for original_key in meta_dict: if original_key in translation_table: for translated_key in translation_table[original_key]: export_dict[key][translated_key] = meta_dict[ original_key] else: export_dict[key][original_key] = meta_dict[original_key] for key in meta_to_translate.ho_data: if key not in export_dict: export_dict[key] = {} for ho_key in meta_to_translate.ho_data[key].data.index: if translation_table is None: export_dict[key + '_' + ho_key] = meta_to_translate.ho_data[key ].data.loc[ho_key].to_dict() else: export_dict[key + '_' + ho_key] = {} meta_dict = meta_to_translate.ho_data[key].data.loc[ho_key ].to_dict() for original_key in meta_dict: if original_key in translation_table: for translated_key in translation_table[original_key]: export_dict[key + '_' + ho_key][translated_key ] = meta_dict[original_key] else: export_dict[key + '_' + ho_key][original_key ] = meta_dict[original_key] return export_dict
generic_meta_translator
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, meta_to_translate):\n export_dict = {}\n if self._meta_translation_table is not None:\n translation_table = {}\n for key in self._meta_translation_table:\n translation_table[getattr(self, key)\n ] = self._meta_translation_table[key]\n else:\n translation_table = None\n for key in meta_to_translate.data.index:\n if translation_table is None:\n export_dict[key] = meta_to_translate.data.loc[key].to_dict()\n else:\n export_dict[key] = {}\n meta_dict = meta_to_translate.data.loc[key].to_dict()\n for original_key in meta_dict:\n if original_key in translation_table:\n for translated_key in translation_table[original_key]:\n export_dict[key][translated_key] = meta_dict[\n original_key]\n else:\n export_dict[key][original_key] = meta_dict[original_key]\n for key in meta_to_translate.ho_data:\n if key not in export_dict:\n export_dict[key] = {}\n for ho_key in meta_to_translate.ho_data[key].data.index:\n if translation_table is None:\n export_dict[key + '_' + ho_key] = meta_to_translate.ho_data[key\n ].data.loc[ho_key].to_dict()\n else:\n export_dict[key + '_' + ho_key] = {}\n meta_dict = meta_to_translate.ho_data[key].data.loc[ho_key\n ].to_dict()\n for original_key in meta_dict:\n if original_key in translation_table:\n for translated_key in translation_table[original_key]:\n export_dict[key + '_' + ho_key][translated_key\n ] = meta_dict[original_key]\n else:\n export_dict[key + '_' + ho_key][original_key\n ] = meta_dict[original_key]\n return export_dict\n", "role": "user" }, { "content": "generic_meta_translator", "role": "assistant" } ]
def x(input_str): ret = {} if input_str is not None: pattern = re.compile('(\\d+)\\s*(\\d+)') for line in input_str.split('\n'): line = line.strip() if len(line) == 0: continue matched = re.search(pattern, line) if matched is None or len(matched.groups()) < 2: continue else: hlu = matched.group(1) alu = matched.group(2) ret[int(alu)] = int(hlu) return ret
to_alu_hlu_map
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(input_str):\n ret = {}\n if input_str is not None:\n pattern = re.compile('(\\\\d+)\\\\s*(\\\\d+)')\n for line in input_str.split('\\n'):\n line = line.strip()\n if len(line) == 0:\n continue\n matched = re.search(pattern, line)\n if matched is None or len(matched.groups()) < 2:\n continue\n else:\n hlu = matched.group(1)\n alu = matched.group(2)\n ret[int(alu)] = int(hlu)\n return ret\n", "role": "user" }, { "content": "to_alu_hlu_map", "role": "assistant" } ]
def x(value): ret = [] p = re.compile('Bus\\s+(\\w+)\\s+Enclosure\\s+(\\w+)\\s+Disk\\s+(\\w+)') if value is not None: for line in value.split('\n'): line = line.strip() if len(line) == 0: continue matched = re.search(p, line) if matched is None or len(matched.groups()) < 3: continue else: ret.append('{}_{}_{}'.format(*matched.groups())) return ret
to_disk_indices
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(value):\n ret = []\n p = re.compile('Bus\\\\s+(\\\\w+)\\\\s+Enclosure\\\\s+(\\\\w+)\\\\s+Disk\\\\s+(\\\\w+)')\n if value is not None:\n for line in value.split('\\n'):\n line = line.strip()\n if len(line) == 0:\n continue\n matched = re.search(p, line)\n if matched is None or len(matched.groups()) < 3:\n continue\n else:\n ret.append('{}_{}_{}'.format(*matched.groups()))\n return ret\n", "role": "user" }, { "content": "to_disk_indices", "role": "assistant" } ]
def x(url): regex_url = ('([a-z][a-z0-9+\\-.]*://)?' + "([a-z0-9\\-._~%!$&'()*+,;=]+@)?" + '([a-z0-9\\-._~%]+' + "|\\[[a-z0-9\\-._~%!$&'()*+,;=:]+\\])?" + '(:(?P<port>[0-9]+))?') m = re.match(regex_url, url, re.IGNORECASE) if m and m.group(3): return url[m.start(3):m.end(3)] else: raise ValueError('URL without a valid host or ip')
url_to_host
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(url):\n regex_url = ('([a-z][a-z0-9+\\\\-.]*://)?' +\n \"([a-z0-9\\\\-._~%!$&'()*+,;=]+@)?\" + '([a-z0-9\\\\-._~%]+' +\n \"|\\\\[[a-z0-9\\\\-._~%!$&'()*+,;=:]+\\\\])?\" + '(:(?P<port>[0-9]+))?')\n m = re.match(regex_url, url, re.IGNORECASE)\n if m and m.group(3):\n return url[m.start(3):m.end(3)]\n else:\n raise ValueError('URL without a valid host or ip')\n", "role": "user" }, { "content": "url_to_host", "role": "assistant" } ]
def x(addr): if addr.startswith('[') and addr.endswith(']'): addr = addr[1:-1] parts = addr.split('/') if len(parts) == 1: return parts[0], None, None if len(parts) > 2: raise ValueError('Illegal host address') else: domain_or_ip, prefix = parts prefix = int(prefix) if re.match('^(?:[0-9]{1,3}\\.){3}[0-9]{1,3}$', domain_or_ip): return domain_or_ip, prefix, ipv4_prefix_to_mask(prefix) elif ':' in domain_or_ip: return domain_or_ip, prefix, ipv6_prefix_to_mask(prefix) else: return domain_or_ip, None, None
parse_host_address
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(addr):\n if addr.startswith('[') and addr.endswith(']'):\n addr = addr[1:-1]\n parts = addr.split('/')\n if len(parts) == 1:\n return parts[0], None, None\n if len(parts) > 2:\n raise ValueError('Illegal host address')\n else:\n domain_or_ip, prefix = parts\n prefix = int(prefix)\n if re.match('^(?:[0-9]{1,3}\\\\.){3}[0-9]{1,3}$', domain_or_ip):\n return domain_or_ip, prefix, ipv4_prefix_to_mask(prefix)\n elif ':' in domain_or_ip:\n return domain_or_ip, prefix, ipv6_prefix_to_mask(prefix)\n else:\n return domain_or_ip, None, None\n", "role": "user" }, { "content": "parse_host_address", "role": "assistant" } ]
def x(prefix): if prefix > 32 or prefix < 0: raise ValueError('invalid cidr prefix for ipv4') else: mask = (1 << 32) - 1 ^ (1 << 32 - prefix) - 1 eight_ones = 255 mask_str = '' for i in range(0, 4): mask_str = str(mask & eight_ones) + mask_str mask = mask >> 8 if i != 3: mask_str = '.' + mask_str return mask_str
ipv4_prefix_to_mask
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(prefix):\n if prefix > 32 or prefix < 0:\n raise ValueError('invalid cidr prefix for ipv4')\n else:\n mask = (1 << 32) - 1 ^ (1 << 32 - prefix) - 1\n eight_ones = 255\n mask_str = ''\n for i in range(0, 4):\n mask_str = str(mask & eight_ones) + mask_str\n mask = mask >> 8\n if i != 3:\n mask_str = '.' + mask_str\n return mask_str\n", "role": "user" }, { "content": "ipv4_prefix_to_mask", "role": "assistant" } ]
def x(prefix): if prefix > 128 or prefix < 0: raise ValueError('invalid cidr prefix for ipv6') else: mask = (1 << 128) - 1 ^ (1 << 128 - prefix) - 1 f = 15 hex_mask_str = '' for i in range(0, 32): hex_mask_str = format(mask & f, 'x') + hex_mask_str mask = mask >> 4 if i != 31 and i & 3 == 3: hex_mask_str = ':' + hex_mask_str return hex_mask_str
ipv6_prefix_to_mask
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(prefix):\n if prefix > 128 or prefix < 0:\n raise ValueError('invalid cidr prefix for ipv6')\n else:\n mask = (1 << 128) - 1 ^ (1 << 128 - prefix) - 1\n f = 15\n hex_mask_str = ''\n for i in range(0, 32):\n hex_mask_str = format(mask & f, 'x') + hex_mask_str\n mask = mask >> 4\n if i != 31 and i & 3 == 3:\n hex_mask_str = ':' + hex_mask_str\n return hex_mask_str\n", "role": "user" }, { "content": "ipv6_prefix_to_mask", "role": "assistant" } ]
def x(self, new_size): ret = self.size_total resp = self.modify(size=new_size) resp.raise_if_err() return ret
expand
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, new_size):\n ret = self.size_total\n resp = self.modify(size=new_size)\n resp.raise_if_err()\n return ret\n", "role": "user" }, { "content": "expand", "role": "assistant" } ]
def x(self, host_names): if self.host_access: curr_hosts = [access.host.name for access in self.host_access] else: curr_hosts = [] if set(curr_hosts) == set(host_names): log.info( 'Hosts for updating is equal to current hosts, skip modification.') return None new_hosts = [UnityHostList.get(cli=self._cli, name=host_name)[0] for host_name in host_names] new_access = [{'host': item, 'accessMask': HostLUNAccessEnum.PRODUCTION } for item in new_hosts] resp = self.modify(host_access=new_access) resp.raise_if_err() return resp
update_hosts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, host_names):\n if self.host_access:\n curr_hosts = [access.host.name for access in self.host_access]\n else:\n curr_hosts = []\n if set(curr_hosts) == set(host_names):\n log.info(\n 'Hosts for updating is equal to current hosts, skip modification.')\n return None\n new_hosts = [UnityHostList.get(cli=self._cli, name=host_name)[0] for\n host_name in host_names]\n new_access = [{'host': item, 'accessMask': HostLUNAccessEnum.PRODUCTION\n } for item in new_hosts]\n resp = self.modify(host_access=new_access)\n resp.raise_if_err()\n return resp\n", "role": "user" }, { "content": "update_hosts", "role": "assistant" } ]
def x(self, dst_lun_id, max_time_out_of_sync, replication_name=None, replicate_existing_snaps=None, remote_system=None): return UnityReplicationSession.create(self._cli, self.get_id(), dst_lun_id, max_time_out_of_sync, name=replication_name, replicate_existing_snaps=replicate_existing_snaps, remote_system= remote_system)
replicate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dst_lun_id, max_time_out_of_sync, replication_name=None,\n replicate_existing_snaps=None, remote_system=None):\n return UnityReplicationSession.create(self._cli, self.get_id(),\n dst_lun_id, max_time_out_of_sync, name=replication_name,\n replicate_existing_snaps=replicate_existing_snaps, remote_system=\n remote_system)\n", "role": "user" }, { "content": "replicate", "role": "assistant" } ]
def x(self, max_time_out_of_sync, dst_pool_id, dst_lun_name=None, remote_system=None, replication_name=None, dst_size=None, dst_sp=None, is_dst_thin=None, dst_tiering_policy=None, is_dst_compression=None): dst_size = self.size_total if dst_size is None else dst_size dst_resource = UnityResourceConfig.to_embedded(name=dst_lun_name, pool_id=dst_pool_id, size=dst_size, default_sp=dst_sp, tiering_policy=dst_tiering_policy, is_thin_enabled=is_dst_thin, is_compression_enabled=is_dst_compression) return UnityReplicationSession.create_with_dst_resource_provisioning(self ._cli, self.get_id(), dst_resource, max_time_out_of_sync, remote_system=remote_system, name=replication_name)
replicate_with_dst_resource_provisioning
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, max_time_out_of_sync, dst_pool_id, dst_lun_name=None,\n remote_system=None, replication_name=None, dst_size=None, dst_sp=None,\n is_dst_thin=None, dst_tiering_policy=None, is_dst_compression=None):\n dst_size = self.size_total if dst_size is None else dst_size\n dst_resource = UnityResourceConfig.to_embedded(name=dst_lun_name,\n pool_id=dst_pool_id, size=dst_size, default_sp=dst_sp,\n tiering_policy=dst_tiering_policy, is_thin_enabled=is_dst_thin,\n is_compression_enabled=is_dst_compression)\n return UnityReplicationSession.create_with_dst_resource_provisioning(self\n ._cli, self.get_id(), dst_resource, max_time_out_of_sync,\n remote_system=remote_system, name=replication_name)\n", "role": "user" }, { "content": "replicate_with_dst_resource_provisioning", "role": "assistant" } ]
def x(self): obj = None if self.is_link_aggregation(): obj = UnityLinkAggregation.get(self._cli, self.get_id()) else: obj = UnityEthernetPort.get(self._cli, self.get_id()) return obj
get_physical_port
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n obj = None\n if self.is_link_aggregation():\n obj = UnityLinkAggregation.get(self._cli, self.get_id())\n else:\n obj = UnityEthernetPort.get(self._cli, self.get_id())\n return obj\n", "role": "user" }, { "content": "get_physical_port", "role": "assistant" } ]
def x(pool_id=None, is_thin_enabled=None, is_deduplication_enabled=None, is_compression_enabled=None, is_backup_only=None, size=None, tiering_policy=None, request_id=None, src_id=None, name=None, default_sp=None, replication_resource_type=None): return {'poolId': pool_id, 'isThinEnabled': is_thin_enabled, 'isDeduplicationEnabled': is_deduplication_enabled, 'isCompressionEnabled': is_compression_enabled, 'isBackupOnly': is_backup_only, 'size': size, 'tieringPolicy': tiering_policy, 'requestId': request_id, 'srcId': src_id, 'name': name, 'defaultSP': default_sp, 'replicationResourceType': replication_resource_type}
to_embedded
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(pool_id=None, is_thin_enabled=None, is_deduplication_enabled=None,\n is_compression_enabled=None, is_backup_only=None, size=None,\n tiering_policy=None, request_id=None, src_id=None, name=None,\n default_sp=None, replication_resource_type=None):\n return {'poolId': pool_id, 'isThinEnabled': is_thin_enabled,\n 'isDeduplicationEnabled': is_deduplication_enabled,\n 'isCompressionEnabled': is_compression_enabled, 'isBackupOnly':\n is_backup_only, 'size': size, 'tieringPolicy': tiering_policy,\n 'requestId': request_id, 'srcId': src_id, 'name': name, 'defaultSP':\n default_sp, 'replicationResourceType': replication_resource_type}\n", "role": "user" }, { "content": "to_embedded", "role": "assistant" } ]
def x(cls, cli, src_resource_id, dst_resource_id, max_time_out_of_sync, name=None, members=None, auto_initiate=None, hourly_snap_replication_policy=None, daily_snap_replication_policy=None, replicate_existing_snaps=None, remote_system=None, src_spa_interface= None, src_spb_interface=None, dst_spa_interface=None, dst_spb_interface =None): req_body = cli.make_body(srcResourceId=src_resource_id, dstResourceId= dst_resource_id, maxTimeOutOfSync=max_time_out_of_sync, members= members, autoInitiate=auto_initiate, name=name, hourlySnapReplicationPolicy=hourly_snap_replication_policy, dailySnapReplicationPolicy=daily_snap_replication_policy, replicateExistingSnaps=replicate_existing_snaps, remoteSystem= remote_system, srcSPAInterface=src_spa_interface, srcSPBInterface= src_spb_interface, dstSPAInterface=dst_spa_interface, dstSPBInterface=dst_spb_interface) resp = cli.post(cls().resource_class, **req_body) resp.raise_if_err() return cls.get(cli, resp.resource_id)
create
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, cli, src_resource_id, dst_resource_id, max_time_out_of_sync,\n name=None, members=None, auto_initiate=None,\n hourly_snap_replication_policy=None, daily_snap_replication_policy=None,\n replicate_existing_snaps=None, remote_system=None, src_spa_interface=\n None, src_spb_interface=None, dst_spa_interface=None, dst_spb_interface\n =None):\n req_body = cli.make_body(srcResourceId=src_resource_id, dstResourceId=\n dst_resource_id, maxTimeOutOfSync=max_time_out_of_sync, members=\n members, autoInitiate=auto_initiate, name=name,\n hourlySnapReplicationPolicy=hourly_snap_replication_policy,\n dailySnapReplicationPolicy=daily_snap_replication_policy,\n replicateExistingSnaps=replicate_existing_snaps, remoteSystem=\n remote_system, srcSPAInterface=src_spa_interface, srcSPBInterface=\n src_spb_interface, dstSPAInterface=dst_spa_interface,\n dstSPBInterface=dst_spb_interface)\n resp = cli.post(cls().resource_class, **req_body)\n resp.raise_if_err()\n return cls.get(cli, resp.resource_id)\n", "role": "user" }, { "content": "create", "role": "assistant" } ]
def x(cls, cli, src_resource_id, dst_resource_config, max_time_out_of_sync, name=None, remote_system=None, src_spa_interface=None, src_spb_interface=None, dst_spa_interface=None, dst_spb_interface=None, dst_resource_element_configs=None, auto_initiate=None, hourly_snap_replication_policy=None, daily_snap_replication_policy=None, replicate_existing_snaps=None): req_body = cli.make_body(srcResourceId=src_resource_id, dstResourceConfig=dst_resource_config, maxTimeOutOfSync= max_time_out_of_sync, name=name, remoteSystem=remote_system, srcSPAInterface=src_spa_interface, srcSPBInterface= src_spb_interface, dstSPAInterface=dst_spa_interface, dstSPBInterface=dst_spb_interface, dstResourceElementConfigs= dst_resource_element_configs, autoInitiate=auto_initiate, hourlySnapReplicationPolicy=hourly_snap_replication_policy, dailySnapReplicationPolicy=daily_snap_replication_policy, replicateExistingSnaps=replicate_existing_snaps) resp = cli.type_action(cls().resource_class, 'createReplicationSessionWDestResProvisioning', **req_body) resp.raise_if_err() session_resp = resp.first_content['id'] return cls.get(cli, _id=session_resp['id'])
create_with_dst_resource_provisioning
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, cli, src_resource_id, dst_resource_config, max_time_out_of_sync,\n name=None, remote_system=None, src_spa_interface=None,\n src_spb_interface=None, dst_spa_interface=None, dst_spb_interface=None,\n dst_resource_element_configs=None, auto_initiate=None,\n hourly_snap_replication_policy=None, daily_snap_replication_policy=None,\n replicate_existing_snaps=None):\n req_body = cli.make_body(srcResourceId=src_resource_id,\n dstResourceConfig=dst_resource_config, maxTimeOutOfSync=\n max_time_out_of_sync, name=name, remoteSystem=remote_system,\n srcSPAInterface=src_spa_interface, srcSPBInterface=\n src_spb_interface, dstSPAInterface=dst_spa_interface,\n dstSPBInterface=dst_spb_interface, dstResourceElementConfigs=\n dst_resource_element_configs, autoInitiate=auto_initiate,\n hourlySnapReplicationPolicy=hourly_snap_replication_policy,\n dailySnapReplicationPolicy=daily_snap_replication_policy,\n replicateExistingSnaps=replicate_existing_snaps)\n resp = cli.type_action(cls().resource_class,\n 'createReplicationSessionWDestResProvisioning', **req_body)\n resp.raise_if_err()\n session_resp = resp.first_content['id']\n return cls.get(cli, _id=session_resp['id'])\n", "role": "user" }, { "content": "create_with_dst_resource_provisioning", "role": "assistant" } ]
def x(self, max_time_out_of_sync=None, name=None, hourly_snap_replication_policy=None, daily_snap_replication_policy=None, src_spa_interface=None, src_spb_interface=None, dst_spa_interface=None, dst_spb_interface=None): req_body = self._cli.make_body(maxTimeOutOfSync=max_time_out_of_sync, name=name, hourlySnapReplicationPolicy= hourly_snap_replication_policy, dailySnapReplicationPolicy= daily_snap_replication_policy, srcSPAInterface=src_spa_interface, srcSPBInterface=src_spb_interface, dstSPAInterface= dst_spa_interface, dstSPBInterface=dst_spb_interface) resp = self.action('modify', **req_body) resp.raise_if_err() return resp
modify
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, max_time_out_of_sync=None, name=None,\n hourly_snap_replication_policy=None, daily_snap_replication_policy=None,\n src_spa_interface=None, src_spb_interface=None, dst_spa_interface=None,\n dst_spb_interface=None):\n req_body = self._cli.make_body(maxTimeOutOfSync=max_time_out_of_sync,\n name=name, hourlySnapReplicationPolicy=\n hourly_snap_replication_policy, dailySnapReplicationPolicy=\n daily_snap_replication_policy, srcSPAInterface=src_spa_interface,\n srcSPBInterface=src_spb_interface, dstSPAInterface=\n dst_spa_interface, dstSPBInterface=dst_spb_interface)\n resp = self.action('modify', **req_body)\n resp.raise_if_err()\n return resp\n", "role": "user" }, { "content": "modify", "role": "assistant" } ]
def x(self, force_full_copy=None, src_spa_interface=None, src_spb_interface =None, dst_spa_interface=None, dst_spb_interface=None): req_body = self._cli.make_body(forceFullCopy=force_full_copy, srcSPAInterface=src_spa_interface, srcSPBInterface= src_spb_interface, dstSPAInterface=dst_spa_interface, dstSPBInterface=dst_spb_interface) resp = self.action('resume', **req_body) resp.raise_if_err() return resp
resume
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, force_full_copy=None, src_spa_interface=None, src_spb_interface\n =None, dst_spa_interface=None, dst_spb_interface=None):\n req_body = self._cli.make_body(forceFullCopy=force_full_copy,\n srcSPAInterface=src_spa_interface, srcSPBInterface=\n src_spb_interface, dstSPAInterface=dst_spa_interface,\n dstSPBInterface=dst_spb_interface)\n resp = self.action('resume', **req_body)\n resp.raise_if_err()\n return resp\n", "role": "user" }, { "content": "resume", "role": "assistant" } ]
def x(self, sync=None, force=None): req_body = self._cli.make_body(sync=sync, force=force) resp = self.action('failover', **req_body) resp.raise_if_err() return resp
failover
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, sync=None, force=None):\n req_body = self._cli.make_body(sync=sync, force=force)\n resp = self.action('failover', **req_body)\n resp.raise_if_err()\n return resp\n", "role": "user" }, { "content": "failover", "role": "assistant" } ]
def x(self, force_full_copy=None): req_body = self._cli.make_body(forceFullCopy=force_full_copy) resp = self.action('failback', **req_body) resp.raise_if_err() return resp
failback
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, force_full_copy=None):\n req_body = self._cli.make_body(forceFullCopy=force_full_copy)\n resp = self.action('failback', **req_body)\n resp.raise_if_err()\n return resp\n", "role": "user" }, { "content": "failback", "role": "assistant" } ]
def x(self): if len(self._orbit_breaks) == 0: self._detBreaks() self._fullDayData = self.sat.data.copy() self._current = 0
calcOrbits
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if len(self._orbit_breaks) == 0:\n self._detBreaks()\n self._fullDayData = self.sat.data.copy()\n self._current = 0\n", "role": "user" }, { "content": "calcOrbits", "role": "assistant" } ]
def x(self, orbit_index_period=24.0): if self.orbit_index is None: raise ValueError('Orbit properties must be defined at ' + 'pysat.Instrument object instantiation.' + 'See Instrument docs.') else: try: self.sat[self.orbit_index] except ValueError: raise ValueError('Provided orbit index does not exist in ' + 'loaded data') lt_diff = self.sat[self.orbit_index].diff() ut_vals = Series(self.sat.data.index) ut_diff = ut_vals.diff() ind, = np.where(lt_diff < -0.1) if len(ind) > 0: ind = np.hstack((ind, np.array([len(self.sat[self.orbit_index])]))) dist = ind[1:] - ind[0:-1] if len(ind) > 1: if min(dist) == 1: print('There are orbit breaks right next to each other') ind = ind[:-1][dist > 1] new_ind = [] for idx in ind: tidx, = np.where(lt_diff[idx - 5:idx + 6] > 0.1) if len(tidx) != 0: for tidx in tidx: if ut_diff[idx - 5:idx + 6].iloc[tidx] < lt_diff[idx - 5:idx + 6].iloc[tidx ] / orbit_index_period * self.orbit_period: pass else: new_ind.append(idx) break else: new_ind.append(idx) ind = np.array(new_ind) ut_change_vs_period = ut_diff > self.orbit_period norm_ut = ut_diff / self.orbit_period norm_ut_vs_norm_lt = norm_ut.gt(np.abs(lt_diff.values / orbit_index_period) ) ut_ind, = np.where(ut_change_vs_period | norm_ut_vs_norm_lt & (norm_ut > 0.95)) if len(ut_ind) > 0: ind = np.hstack((ind, ut_ind)) ind = np.sort(ind) ind = np.unique(ind) print('Time Gap') orbit_ut_diff = ut_vals[ind].diff() orbit_lt_diff = self.sat[self.orbit_index][ind].diff() idx, = np.where(orbit_ut_diff / self.orbit_period - orbit_lt_diff. values / orbit_index_period > 0.97) if len(idx) > 0: if idx[0] != 0: idx = np.hstack((0, idx)) else: idx = np.array([0]) if len(ind) > 0: ind = ind[idx] if ind[0] != 0: ind = np.hstack((np.array([0]), ind)) else: ind = np.array([0]) num_orbits = len(ind) self._orbit_breaks = ind self.num = num_orbits
equaBreaks
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, orbit_index_period=24.0):\n if self.orbit_index is None:\n raise ValueError('Orbit properties must be defined at ' +\n 'pysat.Instrument object instantiation.' + 'See Instrument docs.')\n else:\n try:\n self.sat[self.orbit_index]\n except ValueError:\n raise ValueError('Provided orbit index does not exist in ' +\n 'loaded data')\n lt_diff = self.sat[self.orbit_index].diff()\n ut_vals = Series(self.sat.data.index)\n ut_diff = ut_vals.diff()\n ind, = np.where(lt_diff < -0.1)\n if len(ind) > 0:\n ind = np.hstack((ind, np.array([len(self.sat[self.orbit_index])])))\n dist = ind[1:] - ind[0:-1]\n if len(ind) > 1:\n if min(dist) == 1:\n print('There are orbit breaks right next to each other')\n ind = ind[:-1][dist > 1]\n new_ind = []\n for idx in ind:\n tidx, = np.where(lt_diff[idx - 5:idx + 6] > 0.1)\n if len(tidx) != 0:\n for tidx in tidx:\n if ut_diff[idx - 5:idx + 6].iloc[tidx] < lt_diff[idx - \n 5:idx + 6].iloc[tidx\n ] / orbit_index_period * self.orbit_period:\n pass\n else:\n new_ind.append(idx)\n break\n else:\n new_ind.append(idx)\n ind = np.array(new_ind)\n ut_change_vs_period = ut_diff > self.orbit_period\n norm_ut = ut_diff / self.orbit_period\n norm_ut_vs_norm_lt = norm_ut.gt(np.abs(lt_diff.values / orbit_index_period)\n )\n ut_ind, = np.where(ut_change_vs_period | norm_ut_vs_norm_lt & (norm_ut >\n 0.95))\n if len(ut_ind) > 0:\n ind = np.hstack((ind, ut_ind))\n ind = np.sort(ind)\n ind = np.unique(ind)\n print('Time Gap')\n orbit_ut_diff = ut_vals[ind].diff()\n orbit_lt_diff = self.sat[self.orbit_index][ind].diff()\n idx, = np.where(orbit_ut_diff / self.orbit_period - orbit_lt_diff.\n values / orbit_index_period > 0.97)\n if len(idx) > 0:\n if idx[0] != 0:\n idx = np.hstack((0, idx))\n else:\n idx = np.array([0])\n if len(ind) > 0:\n ind = ind[idx]\n if ind[0] != 0:\n ind = np.hstack((np.array([0]), ind))\n else:\n ind = np.array([0])\n num_orbits = len(ind)\n self._orbit_breaks = ind\n self.num = num_orbits\n", "role": "user" }, { "content": "equaBreaks", "role": "assistant" } ]
def x(self): if self.orbit_index is None: raise ValueError('Orbit properties must be defined at ' + 'pysat.Instrument object instantiation.' + 'See Instrument docs.') else: try: self.sat[self.orbit_index] except ValueError: raise ValueError('Provided orbit index does not appear to ' + 'exist in loaded data') pos = self.sat[self.orbit_index] >= 0 npos = -pos change = pos.values[:-1] & npos.values[1:] | npos.values[:-1] & pos.values[ 1:] ind, = np.where(change) ind += 1 ut_diff = Series(self.sat.data.index).diff() ut_ind, = np.where(ut_diff / self.orbit_period > 0.95) if len(ut_ind) > 0: ind = np.hstack((ind, ut_ind)) ind = np.sort(ind) ind = np.unique(ind) if ind[0] != 0: ind = np.hstack((np.array([0]), ind)) num_orbits = len(ind) self._orbit_breaks = ind self.num = num_orbits
polarBreaks
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.orbit_index is None:\n raise ValueError('Orbit properties must be defined at ' +\n 'pysat.Instrument object instantiation.' + 'See Instrument docs.')\n else:\n try:\n self.sat[self.orbit_index]\n except ValueError:\n raise ValueError('Provided orbit index does not appear to ' +\n 'exist in loaded data')\n pos = self.sat[self.orbit_index] >= 0\n npos = -pos\n change = pos.values[:-1] & npos.values[1:] | npos.values[:-1] & pos.values[\n 1:]\n ind, = np.where(change)\n ind += 1\n ut_diff = Series(self.sat.data.index).diff()\n ut_ind, = np.where(ut_diff / self.orbit_period > 0.95)\n if len(ut_ind) > 0:\n ind = np.hstack((ind, ut_ind))\n ind = np.sort(ind)\n ind = np.unique(ind)\n if ind[0] != 0:\n ind = np.hstack((np.array([0]), ind))\n num_orbits = len(ind)\n self._orbit_breaks = ind\n self.num = num_orbits\n", "role": "user" }, { "content": "polarBreaks", "role": "assistant" } ]
def x(self): if self.orbit_index is None: raise ValueError('Orbit properties must be defined at ' + 'pysat.Instrument object instantiation.' + 'See Instrument docs.') else: try: self.sat[self.orbit_index] except ValueError: raise ValueError('Provided orbit index does not appear to ' + 'exist in loaded data') uniq_vals = self.sat[self.orbit_index].unique() orbit_index = [] for val in uniq_vals: idx, = np.where(val == self.sat[self.orbit_index].values) orbit_index.append(idx[0]) if orbit_index[0] != 0: ind = np.hstack((np.array([0]), orbit_index)) else: ind = orbit_index num_orbits = len(ind) self._orbit_breaks = ind self.num = num_orbits
orbitNumberBreaks
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.orbit_index is None:\n raise ValueError('Orbit properties must be defined at ' +\n 'pysat.Instrument object instantiation.' + 'See Instrument docs.')\n else:\n try:\n self.sat[self.orbit_index]\n except ValueError:\n raise ValueError('Provided orbit index does not appear to ' +\n 'exist in loaded data')\n uniq_vals = self.sat[self.orbit_index].unique()\n orbit_index = []\n for val in uniq_vals:\n idx, = np.where(val == self.sat[self.orbit_index].values)\n orbit_index.append(idx[0])\n if orbit_index[0] != 0:\n ind = np.hstack((np.array([0]), orbit_index))\n else:\n ind = orbit_index\n num_orbits = len(ind)\n self._orbit_breaks = ind\n self.num = num_orbits\n", "role": "user" }, { "content": "orbitNumberBreaks", "role": "assistant" } ]
def x(self, orbit=None): if not self.sat.empty: self._calcOrbits() if orbit is not None: if orbit == -1: self.sat.data = self._fullDayData[self._orbit_breaks[self. num + orbit]:] self._current = self.num + orbit + 1 elif (orbit < 0) & (orbit >= -self.num): self.sat.data = self._fullDayData[self._orbit_breaks[self. num + orbit]:self._orbit_breaks[self.num + orbit + 1]] self._current = self.num + orbit + 1 elif (orbit < self.num) & (orbit != 0): self.sat.data = self._fullDayData[self._orbit_breaks[orbit - 1]:self._orbit_breaks[orbit]] self._current = orbit elif orbit == self.num: self.sat.data = self._fullDayData[self._orbit_breaks[orbit - 1]:] self._current = orbit elif orbit == 0: raise ValueError('Orbits internally indexed by 1, 0 not ' + 'allowed') else: self.sat.data = [] raise ValueError('Requested an orbit past total orbits ' + 'for day') else: raise ValueError('Must set an orbit')
getBasicOrbit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, orbit=None):\n if not self.sat.empty:\n self._calcOrbits()\n if orbit is not None:\n if orbit == -1:\n self.sat.data = self._fullDayData[self._orbit_breaks[self.\n num + orbit]:]\n self._current = self.num + orbit + 1\n elif (orbit < 0) & (orbit >= -self.num):\n self.sat.data = self._fullDayData[self._orbit_breaks[self.\n num + orbit]:self._orbit_breaks[self.num + orbit + 1]]\n self._current = self.num + orbit + 1\n elif (orbit < self.num) & (orbit != 0):\n self.sat.data = self._fullDayData[self._orbit_breaks[orbit -\n 1]:self._orbit_breaks[orbit]]\n self._current = orbit\n elif orbit == self.num:\n self.sat.data = self._fullDayData[self._orbit_breaks[orbit -\n 1]:]\n self._current = orbit\n elif orbit == 0:\n raise ValueError('Orbits internally indexed by 1, 0 not ' +\n 'allowed')\n else:\n self.sat.data = []\n raise ValueError('Requested an orbit past total orbits ' +\n 'for day')\n else:\n raise ValueError('Must set an orbit')\n", "role": "user" }, { "content": "getBasicOrbit", "role": "assistant" } ]
def x(self, orbit=None): if not self.sat.empty: self._calcOrbits() if orbit is not None: if orbit < 0: orbit = self.num + 1 + orbit if orbit == 1: try: true_date = self.sat.date self.sat.prev() if not self.sat.empty: self.load(orbit=-1) else: self.sat.next() self._getBasicOrbit(orbit=1) delta = true_date - self.sat.data.index[0] if delta >= self.orbit_period: self.next() except StopIteration: self._getBasicOrbit(orbit=1) print('Loaded Orbit:%i' % (self._current - 1)) elif orbit == self.num: if self.num != 1: self._getBasicOrbit(self.num - 1) self.next() else: self._getBasicOrbit(orbit=-1) elif orbit < self.num: self._getBasicOrbit(orbit) print('Loaded Orbit:%i' % (self._current - 1)) else: self.sat.data = DataFrame() raise Exception('Requested an orbit past total orbits for day') else: raise Exception('Must set an orbit') else: print('No data loaded in instrument object to determine orbits.')
load
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, orbit=None):\n if not self.sat.empty:\n self._calcOrbits()\n if orbit is not None:\n if orbit < 0:\n orbit = self.num + 1 + orbit\n if orbit == 1:\n try:\n true_date = self.sat.date\n self.sat.prev()\n if not self.sat.empty:\n self.load(orbit=-1)\n else:\n self.sat.next()\n self._getBasicOrbit(orbit=1)\n delta = true_date - self.sat.data.index[0]\n if delta >= self.orbit_period:\n self.next()\n except StopIteration:\n self._getBasicOrbit(orbit=1)\n print('Loaded Orbit:%i' % (self._current - 1))\n elif orbit == self.num:\n if self.num != 1:\n self._getBasicOrbit(self.num - 1)\n self.next()\n else:\n self._getBasicOrbit(orbit=-1)\n elif orbit < self.num:\n self._getBasicOrbit(orbit)\n print('Loaded Orbit:%i' % (self._current - 1))\n else:\n self.sat.data = DataFrame()\n raise Exception('Requested an orbit past total orbits for day')\n else:\n raise Exception('Must set an orbit')\n else:\n print('No data loaded in instrument object to determine orbits.')\n", "role": "user" }, { "content": "load", "role": "assistant" } ]
def x(self, *arg, **kwarg): if not self.sat.empty: self._calcOrbits() if self._current == self.num - 1: self._getBasicOrbit(orbit=-1) load_next = True if self.sat._iter_type == 'date': delta = self.sat.date - self.sat.data.index[-1 ] + pds.Timedelta('1 day') if delta >= self.orbit_period: load_next = False if load_next: temp_orbit_data = self.sat.data.copy() try: self.sat.next() if not self.sat.empty: self.sat.data = pds.concat([temp_orbit_data[:self. sat.data.index[0] - pds.DateOffset(microseconds =1)], self.sat.data]) self._getBasicOrbit(orbit=1) else: self.sat.prev() self._getBasicOrbit(orbit=-1) except StopIteration: pass del temp_orbit_data print('Loaded Orbit:%i' % (self._current - 1)) elif self._current == self.num: temp_orbit_data = self.sat.data.copy() self.sat.next() if not self.sat.empty: pad_next = True if self.sat._iter_type == 'date': delta = self.sat.date - temp_orbit_data.index[-1] if delta >= self.orbit_period: pad_next = False if pad_next: self.sat.data = pds.concat([temp_orbit_data[:self.sat. data.index[0] - pds.DateOffset(microseconds=1)], self.sat.data]) self._getBasicOrbit(orbit=2) else: self._getBasicOrbit(orbit=1) if self.sat._iter_type == 'date': delta = self.sat.date + pds.DateOffset(days=1 ) - self.sat.data.index[0] if delta < self.orbit_period: self._current = self.num - 1 self.next() else: while self.sat.empty: self.sat.next() self._getBasicOrbit(orbit=1) del temp_orbit_data print('Loaded Orbit:%i' % (self._current - 1)) elif self._current == 0: self.load(orbit=1) elif self._current < self.num - 1: self._getBasicOrbit(orbit=self._current + 1) print('Loaded Orbit:%i' % (self._current - 1)) else: raise Exception('You ended up where nobody should ever be. ' + 'Talk to someone about this fundamental ' + 'failure.') else: while self.sat.empty: self.sat.next() self.next()
next
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *arg, **kwarg):\n if not self.sat.empty:\n self._calcOrbits()\n if self._current == self.num - 1:\n self._getBasicOrbit(orbit=-1)\n load_next = True\n if self.sat._iter_type == 'date':\n delta = self.sat.date - self.sat.data.index[-1\n ] + pds.Timedelta('1 day')\n if delta >= self.orbit_period:\n load_next = False\n if load_next:\n temp_orbit_data = self.sat.data.copy()\n try:\n self.sat.next()\n if not self.sat.empty:\n self.sat.data = pds.concat([temp_orbit_data[:self.\n sat.data.index[0] - pds.DateOffset(microseconds\n =1)], self.sat.data])\n self._getBasicOrbit(orbit=1)\n else:\n self.sat.prev()\n self._getBasicOrbit(orbit=-1)\n except StopIteration:\n pass\n del temp_orbit_data\n print('Loaded Orbit:%i' % (self._current - 1))\n elif self._current == self.num:\n temp_orbit_data = self.sat.data.copy()\n self.sat.next()\n if not self.sat.empty:\n pad_next = True\n if self.sat._iter_type == 'date':\n delta = self.sat.date - temp_orbit_data.index[-1]\n if delta >= self.orbit_period:\n pad_next = False\n if pad_next:\n self.sat.data = pds.concat([temp_orbit_data[:self.sat.\n data.index[0] - pds.DateOffset(microseconds=1)],\n self.sat.data])\n self._getBasicOrbit(orbit=2)\n else:\n self._getBasicOrbit(orbit=1)\n if self.sat._iter_type == 'date':\n delta = self.sat.date + pds.DateOffset(days=1\n ) - self.sat.data.index[0]\n if delta < self.orbit_period:\n self._current = self.num - 1\n self.next()\n else:\n while self.sat.empty:\n self.sat.next()\n self._getBasicOrbit(orbit=1)\n del temp_orbit_data\n print('Loaded Orbit:%i' % (self._current - 1))\n elif self._current == 0:\n self.load(orbit=1)\n elif self._current < self.num - 1:\n self._getBasicOrbit(orbit=self._current + 1)\n print('Loaded Orbit:%i' % (self._current - 1))\n else:\n raise Exception('You ended up where nobody should ever be. ' +\n 'Talk to someone about this fundamental ' + 'failure.')\n else:\n while self.sat.empty:\n self.sat.next()\n self.next()\n", "role": "user" }, { "content": "next", "role": "assistant" } ]
def x(fnames, tag=None, sat_id=None): data = pds.DataFrame() for filename in fnames: fname = filename[0:-11] with open(fname) as f: lines = f.readlines() idx = 0 max_lines = 0 for line in lines: if len(line) > 1: max_lines += 1 yr = np.zeros(max_lines * 24, dtype=int) mo = np.zeros(max_lines * 24, dtype=int) day = np.zeros(max_lines * 24, dtype=int) ut = np.zeros(max_lines * 24, dtype=int) dst = np.zeros(max_lines * 24, dtype=int) for line in lines: if len(line) > 1: temp_year = int(line[14:16] + line[3:5]) if temp_year > 57: temp_year += 1900 else: temp_year += 2000 yr[idx:idx + 24] = temp_year mo[idx:idx + 24] = int(line[5:7]) day[idx:idx + 24] = int(line[8:10]) ut[idx:idx + 24] = np.arange(24) temp = line.strip()[20:-4] temp2 = [temp[4 * i:4 * (i + 1)] for i in np.arange(24)] dst[idx:idx + 24] = temp2 idx += 24 start = pds.datetime(yr[0], mo[0], day[0], ut[0]) stop = pds.datetime(yr[-1], mo[-1], day[-1], ut[-1]) dates = pds.date_range(start, stop, freq='H') new_data = pds.DataFrame(dst, index=dates, columns=['dst']) new_date = pysat.datetime.strptime(filename[-10:], '%Y-%m-%d') idx, = np.where((new_data.index >= new_date) & (new_data.index < new_date + pds.DateOffset(days=1))) new_data = new_data.iloc[idx, :] data = pds.concat([data, new_data], axis=0) return data, pysat.Meta()
load
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fnames, tag=None, sat_id=None):\n data = pds.DataFrame()\n for filename in fnames:\n fname = filename[0:-11]\n with open(fname) as f:\n lines = f.readlines()\n idx = 0\n max_lines = 0\n for line in lines:\n if len(line) > 1:\n max_lines += 1\n yr = np.zeros(max_lines * 24, dtype=int)\n mo = np.zeros(max_lines * 24, dtype=int)\n day = np.zeros(max_lines * 24, dtype=int)\n ut = np.zeros(max_lines * 24, dtype=int)\n dst = np.zeros(max_lines * 24, dtype=int)\n for line in lines:\n if len(line) > 1:\n temp_year = int(line[14:16] + line[3:5])\n if temp_year > 57:\n temp_year += 1900\n else:\n temp_year += 2000\n yr[idx:idx + 24] = temp_year\n mo[idx:idx + 24] = int(line[5:7])\n day[idx:idx + 24] = int(line[8:10])\n ut[idx:idx + 24] = np.arange(24)\n temp = line.strip()[20:-4]\n temp2 = [temp[4 * i:4 * (i + 1)] for i in np.arange(24)]\n dst[idx:idx + 24] = temp2\n idx += 24\n start = pds.datetime(yr[0], mo[0], day[0], ut[0])\n stop = pds.datetime(yr[-1], mo[-1], day[-1], ut[-1])\n dates = pds.date_range(start, stop, freq='H')\n new_data = pds.DataFrame(dst, index=dates, columns=['dst'])\n new_date = pysat.datetime.strptime(filename[-10:], '%Y-%m-%d')\n idx, = np.where((new_data.index >= new_date) & (new_data.index <\n new_date + pds.DateOffset(days=1)))\n new_data = new_data.iloc[idx, :]\n data = pds.concat([data, new_data], axis=0)\n return data, pysat.Meta()\n", "role": "user" }, { "content": "load", "role": "assistant" } ]
def x(yr, mo, day): yr = '20' + yr yr = int(yr) mo = int(mo) day = int(day) return pds.datetime(yr, mo, day)
parse
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(yr, mo, day):\n yr = '20' + yr\n yr = int(yr)\n mo = int(mo)\n day = int(day)\n return pds.datetime(yr, mo, day)\n", "role": "user" }, { "content": "parse", "role": "assistant" } ]
def x(fnames, tag=None, sat_id=None): data = pds.DataFrame() colspec = [(0, 2), (2, 4), (4, 6), (7, 10), (10, 13), (13, 16), (16, 19 ), (19, 23), (23, 26), (26, 29), (29, 32), (32, 50)] for filename in fnames: fname = filename[0:-11] date = pysat.datetime.strptime(filename[-10:], '%Y-%m-%d') temp = pds.read_fwf(fname, colspecs=colspec, skipfooter=4, header= None, parse_dates=[[0, 1, 2]], date_parser=_parse, index_col= '0_1_2') idx, = np.where((temp.index >= date) & (temp.index < date + pds. DateOffset(days=1))) temp = temp.iloc[idx, :] data = pds.concat([data, temp], axis=0) data = data.iloc[:, 0:-1] s = pds.Series() for i in np.arange(8): temp = pds.Series(data.iloc[:, i].values, index=data.index + pds. DateOffset(hours=int(3 * i))) s = s.append(temp) s = s.sort_index() s.index.name = 'time' first = np.array([float(x[0]) for x in s]) flag = np.array([x[1] for x in s]) ind, = np.where(flag == '+') first[ind] += 1.0 / 3.0 ind, = np.where(flag == '-') first[ind] -= 1.0 / 3.0 result = pds.DataFrame(first, columns=['kp'], index=s.index) return result, pysat.Meta()
load
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fnames, tag=None, sat_id=None):\n data = pds.DataFrame()\n colspec = [(0, 2), (2, 4), (4, 6), (7, 10), (10, 13), (13, 16), (16, 19\n ), (19, 23), (23, 26), (26, 29), (29, 32), (32, 50)]\n for filename in fnames:\n fname = filename[0:-11]\n date = pysat.datetime.strptime(filename[-10:], '%Y-%m-%d')\n temp = pds.read_fwf(fname, colspecs=colspec, skipfooter=4, header=\n None, parse_dates=[[0, 1, 2]], date_parser=_parse, index_col=\n '0_1_2')\n idx, = np.where((temp.index >= date) & (temp.index < date + pds.\n DateOffset(days=1)))\n temp = temp.iloc[idx, :]\n data = pds.concat([data, temp], axis=0)\n data = data.iloc[:, 0:-1]\n s = pds.Series()\n for i in np.arange(8):\n temp = pds.Series(data.iloc[:, i].values, index=data.index + pds.\n DateOffset(hours=int(3 * i)))\n s = s.append(temp)\n s = s.sort_index()\n s.index.name = 'time'\n first = np.array([float(x[0]) for x in s])\n flag = np.array([x[1] for x in s])\n ind, = np.where(flag == '+')\n first[ind] += 1.0 / 3.0\n ind, = np.where(flag == '-')\n first[ind] -= 1.0 / 3.0\n result = pds.DataFrame(first, columns=['kp'], index=s.index)\n return result, pysat.Meta()\n", "role": "user" }, { "content": "load", "role": "assistant" } ]
def x(date_array, tag, sat_id, data_path, user=None, password=None): import ftplib from ftplib import FTP import sys ftp = FTP('ftp.gfz-potsdam.de') ftp.login() ftp.cwd('/pub/home/obs/kp-ap/tab') for date in date_array: fname = 'kp{year:02d}{month:02d}.tab' fname = fname.format(year=date.year - date.year // 100 * 100, month =date.month) local_fname = fname saved_fname = os.path.join(data_path, local_fname) try: print('Downloading file for ' + date.strftime('%D')) sys.stdout.flush() ftp.retrbinary('RETR ' + fname, open(saved_fname, 'wb').write) except ftplib.error_perm as exception: if str(exception.args[0]).split(' ', 1)[0] != '550': raise else: os.remove(saved_fname) print('File not available for ' + date.strftime('%D')) ftp.close() return
download
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(date_array, tag, sat_id, data_path, user=None, password=None):\n import ftplib\n from ftplib import FTP\n import sys\n ftp = FTP('ftp.gfz-potsdam.de')\n ftp.login()\n ftp.cwd('/pub/home/obs/kp-ap/tab')\n for date in date_array:\n fname = 'kp{year:02d}{month:02d}.tab'\n fname = fname.format(year=date.year - date.year // 100 * 100, month\n =date.month)\n local_fname = fname\n saved_fname = os.path.join(data_path, local_fname)\n try:\n print('Downloading file for ' + date.strftime('%D'))\n sys.stdout.flush()\n ftp.retrbinary('RETR ' + fname, open(saved_fname, 'wb').write)\n except ftplib.error_perm as exception:\n if str(exception.args[0]).split(' ', 1)[0] != '550':\n raise\n else:\n os.remove(saved_fname)\n print('File not available for ' + date.strftime('%D'))\n ftp.close()\n return\n", "role": "user" }, { "content": "download", "role": "assistant" } ]
def x(sat, maxKp=None, filterTime=None, kpData=None, kp_inst=None): if kp_inst is not None: kp_inst.load(date=sat.date, verifyPad=True) kpData = kp_inst elif kpData is None: kp = pysat.Instrument('sw', 'kp', pad=pds.DateOffset(days=1)) kp.load(date=sat.date, verifyPad=True) kpData = kp if maxKp is None: maxKp = 3 + 1.0 / 3.0 if filterTime is None: filterTime = 24 date = sat.date selData = kpData[date - pds.DateOffset(days=1):date + pds.DateOffset( days=1)] ind, = np.where(selData['kp'] >= maxKp) for lind in ind: sat.data[selData.index[lind]:selData.index[lind] + pds.DateOffset( hours=filterTime)] = np.NaN sat.data = sat.data.dropna(axis=0, how='all') return
filter_geoquiet
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sat, maxKp=None, filterTime=None, kpData=None, kp_inst=None):\n if kp_inst is not None:\n kp_inst.load(date=sat.date, verifyPad=True)\n kpData = kp_inst\n elif kpData is None:\n kp = pysat.Instrument('sw', 'kp', pad=pds.DateOffset(days=1))\n kp.load(date=sat.date, verifyPad=True)\n kpData = kp\n if maxKp is None:\n maxKp = 3 + 1.0 / 3.0\n if filterTime is None:\n filterTime = 24\n date = sat.date\n selData = kpData[date - pds.DateOffset(days=1):date + pds.DateOffset(\n days=1)]\n ind, = np.where(selData['kp'] >= maxKp)\n for lind in ind:\n sat.data[selData.index[lind]:selData.index[lind] + pds.DateOffset(\n hours=filterTime)] = np.NaN\n sat.data = sat.data.dropna(axis=0, how='all')\n return\n", "role": "user" }, { "content": "filter_geoquiet", "role": "assistant" } ]
def x(self, converter_str): ret = None if converter_str is not None: converter_desc_list = converter_str.split('.') if len(converter_desc_list) == 1: converter = converter_desc_list[0] ret = getattr(cvt, converter, None) if ret is None: ret = self.get_converter(converter) if ret is None: ret = self.get_resource_clz_by_name(converter) if ret is None: ret = self.get_enum_by_name(converter) if ret is None: ret = self.get(converter) if ret is None and converter_str is not None: raise ValueError('Specified converter not supported: {}'.format (converter_str)) return ret
get_converter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, converter_str):\n ret = None\n if converter_str is not None:\n converter_desc_list = converter_str.split('.')\n if len(converter_desc_list) == 1:\n converter = converter_desc_list[0]\n ret = getattr(cvt, converter, None)\n if ret is None:\n ret = self.get_converter(converter)\n if ret is None:\n ret = self.get_resource_clz_by_name(converter)\n if ret is None:\n ret = self.get_enum_by_name(converter)\n if ret is None:\n ret = self.get(converter)\n if ret is None and converter_str is not None:\n raise ValueError('Specified converter not supported: {}'.format\n (converter_str))\n return ret\n", "role": "user" }, { "content": "get_converter", "role": "assistant" } ]
def x(self, local_path, remote_path): sftp_client = self.transport.open_sftp_client() LOG.debug( 'Copy the local file to remote. Source=%(src)s. Target=%(target)s.' % {'src': local_path, 'target': remote_path}) try: sftp_client.put(local_path, remote_path) except Exception as ex: LOG.error('Failed to copy the local file to remote. Reason: %s.' % six.text_type(ex)) raise SFtpExecutionError(err=ex)
copy_file_to_remote
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, local_path, remote_path):\n sftp_client = self.transport.open_sftp_client()\n LOG.debug(\n 'Copy the local file to remote. Source=%(src)s. Target=%(target)s.' %\n {'src': local_path, 'target': remote_path})\n try:\n sftp_client.put(local_path, remote_path)\n except Exception as ex:\n LOG.error('Failed to copy the local file to remote. Reason: %s.' %\n six.text_type(ex))\n raise SFtpExecutionError(err=ex)\n", "role": "user" }, { "content": "copy_file_to_remote", "role": "assistant" } ]
def x(self, remote_path, local_path): sftp_client = self.transport.open_sftp_client() LOG.debug('Get the remote file. Source=%(src)s. Target=%(target)s.' % { 'src': remote_path, 'target': local_path}) try: sftp_client.get(remote_path, local_path) except Exception as ex: LOG.error('Failed to secure copy. Reason: %s.' % six.text_type(ex)) raise SFtpExecutionError(err=ex)
get_remote_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, remote_path, local_path):\n sftp_client = self.transport.open_sftp_client()\n LOG.debug('Get the remote file. Source=%(src)s. Target=%(target)s.' % {\n 'src': remote_path, 'target': local_path})\n try:\n sftp_client.get(remote_path, local_path)\n except Exception as ex:\n LOG.error('Failed to secure copy. Reason: %s.' % six.text_type(ex))\n raise SFtpExecutionError(err=ex)\n", "role": "user" }, { "content": "get_remote_file", "role": "assistant" } ]
def x(self): if 'isLive' in self.__dict__ and self.isLive: self.transport.close() self.isLive = False
close
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if 'isLive' in self.__dict__ and self.isLive:\n self.transport.close()\n self.isLive = False\n", "role": "user" }, { "content": "close", "role": "assistant" } ]
def x(check_object=False, check_invalid_data_mover=False): def x(f): @functools.wraps(f) def x(self, *argv, **kwargs): request = f(self, *argv, **kwargs) return self.request(request, check_object=check_object, check_invalid_data_mover=check_invalid_data_mover) return func_wrapper return decorator
xml_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(check_object=False, check_invalid_data_mover=False):\n\n def x(f):\n\n @functools.wraps(f)\n def x(self, *argv, **kwargs):\n request = f(self, *argv, **kwargs)\n return self.request(request, check_object=check_object,\n check_invalid_data_mover=check_invalid_data_mover)\n return func_wrapper\n return decorator\n", "role": "user" }, { "content": "xml_request", "role": "assistant" } ]
def x(f): @functools.wraps(f) def x(self, *argv, **kwargs): commands = f(self, *argv, **kwargs) return self.ssh_execute(['env', 'NAS_DB=/nas'] + commands) return func_wrapper
nas_command
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(f):\n\n @functools.wraps(f)\n def x(self, *argv, **kwargs):\n commands = f(self, *argv, **kwargs)\n return self.ssh_execute(['env', 'NAS_DB=/nas'] + commands)\n return func_wrapper\n", "role": "user" }, { "content": "nas_command", "role": "assistant" } ]
def x(self, backup=None, delete_backup=False): resp = self._cli.action(self.resource_class, self.get_id(), 'restore', copyName=backup) resp.raise_if_err() backup = resp.first_content['backup'] backup_snap = UnitySnap(_id=backup['id'], cli=self._cli) if delete_backup: log.info('Deleting the backup snap {} as the restoration succeeded.' .format(backup['id'])) backup_snap.delete() return backup_snap
restore
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, backup=None, delete_backup=False):\n resp = self._cli.action(self.resource_class, self.get_id(), 'restore',\n copyName=backup)\n resp.raise_if_err()\n backup = resp.first_content['backup']\n backup_snap = UnitySnap(_id=backup['id'], cli=self._cli)\n if delete_backup:\n log.info('Deleting the backup snap {} as the restoration succeeded.'\n .format(backup['id']))\n backup_snap.delete()\n return backup_snap\n", "role": "user" }, { "content": "restore", "role": "assistant" } ]
def x(self, name, io_limit_policy=None, description=None): if self.is_member_snap(): raise UnityCGMemberActionNotSupportError() if self.lun and not self.lun.is_thin_enabled: raise UnityThinCloneNotAllowedError() return TCHelper.thin_clone(self._cli, self, name, io_limit_policy, description)
thin_clone
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, name, io_limit_policy=None, description=None):\n if self.is_member_snap():\n raise UnityCGMemberActionNotSupportError()\n if self.lun and not self.lun.is_thin_enabled:\n raise UnityThinCloneNotAllowedError()\n return TCHelper.thin_clone(self._cli, self, name, io_limit_policy,\n description)\n", "role": "user" }, { "content": "thin_clone", "role": "assistant" } ]
def x(self, async_mode=False, even_attached=False): try: return super(UnitySnap, self).delete(async_mode=async_mode) except UnityDeleteAttachedSnapError: if even_attached: log.debug( 'Force delete the snapshot even if it is attached. First detach the snapshot from hosts, then delete again.' ) self.detach_from(None) return super(UnitySnap, self).delete(async_mode=async_mode) else: raise
delete
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, async_mode=False, even_attached=False):\n try:\n return super(UnitySnap, self).delete(async_mode=async_mode)\n except UnityDeleteAttachedSnapError:\n if even_attached:\n log.debug(\n 'Force delete the snapshot even if it is attached. First detach the snapshot from hosts, then delete again.'\n )\n self.detach_from(None)\n return super(UnitySnap, self).delete(async_mode=async_mode)\n else:\n raise\n", "role": "user" }, { "content": "delete", "role": "assistant" } ]
def x(self, type_name, base_fields=None, the_filter=None, nested_fields=None): fields = self.get_fields(type_name, base_fields, nested_fields) the_filter = self.dict_to_filter_string(the_filter) url = '/api/types/{}/instances'.format(type_name) resp = self.rest_get(url, fields=fields, filter=the_filter) ret = resp while resp.has_next_page: resp = self.rest_get(url, fields=fields, filter=the_filter, page= resp.next_page) ret.entries.extend(resp.entries) return ret
get_all
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, type_name, base_fields=None, the_filter=None, nested_fields=None):\n fields = self.get_fields(type_name, base_fields, nested_fields)\n the_filter = self.dict_to_filter_string(the_filter)\n url = '/api/types/{}/instances'.format(type_name)\n resp = self.rest_get(url, fields=fields, filter=the_filter)\n ret = resp\n while resp.has_next_page:\n resp = self.rest_get(url, fields=fields, filter=the_filter, page=\n resp.next_page)\n ret.entries.extend(resp.entries)\n return ret\n", "role": "user" }, { "content": "get_all", "role": "assistant" } ]
def x(self, type_name, obj_id, base_fields=None, nested_fields=None): base_fields = self.get_fields(type_name, base_fields, nested_fields) url = '/api/instances/{}/{}'.format(type_name, obj_id) return self.rest_get(url, fields=base_fields)
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, type_name, obj_id, base_fields=None, nested_fields=None):\n base_fields = self.get_fields(type_name, base_fields, nested_fields)\n url = '/api/instances/{}/{}'.format(type_name, obj_id)\n return self.rest_get(url, fields=base_fields)\n", "role": "user" }, { "content": "get", "role": "assistant" } ]