source_code
stringlengths 45
7.24k
| function_name
stringlengths 1
111
| messages
list |
|---|---|---|
def x(self):
data = get_data(self.endpoint, self.id_, force_lookup=self.__force_lookup)
for key, val in data.items():
if key == 'location_area_encounters' and self.endpoint == 'pokemon':
params = val.split('/')[-3:]
ep, id_, subr = params
encounters = get_data(ep, int(id_), subr)
data[key] = [_make_obj(enc) for enc in encounters]
continue
if isinstance(val, dict):
data[key] = _make_obj(val)
elif isinstance(val, list):
data[key] = [_make_obj(i) for i in val]
self.__dict__.update(data)
return None
|
load
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n data = get_data(self.endpoint, self.id_, force_lookup=self.__force_lookup)\n for key, val in data.items():\n if key == 'location_area_encounters' and self.endpoint == 'pokemon':\n params = val.split('/')[-3:]\n ep, id_, subr = params\n encounters = get_data(ep, int(id_), subr)\n data[key] = [_make_obj(enc) for enc in encounters]\n continue\n if isinstance(val, dict):\n data[key] = _make_obj(val)\n elif isinstance(val, list):\n data[key] = [_make_obj(i) for i in val]\n self.__dict__.update(data)\n return None\n",
"role": "user"
},
{
"content": "load",
"role": "assistant"
}
] |
def x(path, mode=511):
try:
os.makedirs(path, mode)
except OSError as error:
if error.errno != 17:
raise
return path
|
safe_make_dirs
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path, mode=511):\n try:\n os.makedirs(path, mode)\n except OSError as error:\n if error.errno != 17:\n raise\n return path\n",
"role": "user"
},
{
"content": "safe_make_dirs",
"role": "assistant"
}
] |
def x():
xdg_cache_home = os.environ.get('XDG_CACHE_HOME') or os.path.join(os.
path.expanduser('~'), '.cache')
return os.path.join(xdg_cache_home, 'pokebase')
|
get_default_cache
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n xdg_cache_home = os.environ.get('XDG_CACHE_HOME') or os.path.join(os.\n path.expanduser('~'), '.cache')\n return os.path.join(xdg_cache_home, 'pokebase')\n",
"role": "user"
},
{
"content": "get_default_cache",
"role": "assistant"
}
] |
def x(new_path=None):
global CACHE_DIR, API_CACHE, SPRITE_CACHE
if new_path is None:
new_path = get_default_cache()
CACHE_DIR = safe_make_dirs(os.path.abspath(new_path))
API_CACHE = os.path.join(CACHE_DIR, 'api.cache')
SPRITE_CACHE = safe_make_dirs(os.path.join(CACHE_DIR, 'sprite'))
return CACHE_DIR, API_CACHE, SPRITE_CACHE
|
set_cache
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(new_path=None):\n global CACHE_DIR, API_CACHE, SPRITE_CACHE\n if new_path is None:\n new_path = get_default_cache()\n CACHE_DIR = safe_make_dirs(os.path.abspath(new_path))\n API_CACHE = os.path.join(CACHE_DIR, 'api.cache')\n SPRITE_CACHE = safe_make_dirs(os.path.join(CACHE_DIR, 'sprite'))\n return CACHE_DIR, API_CACHE, SPRITE_CACHE\n",
"role": "user"
},
{
"content": "set_cache",
"role": "assistant"
}
] |
def x(self, lun_or_snap, skip_hlu_0=False):
try:
return self._attach_with_retry(lun_or_snap, skip_hlu_0)
except ex.SystemAPINotSupported:
raise
except ex.UnityAttachExceedLimitError:
raise
except:
self.detach(lun_or_snap)
raise
|
attach
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, lun_or_snap, skip_hlu_0=False):\n try:\n return self._attach_with_retry(lun_or_snap, skip_hlu_0)\n except ex.SystemAPINotSupported:\n raise\n except ex.UnityAttachExceedLimitError:\n raise\n except:\n self.detach(lun_or_snap)\n raise\n",
"role": "user"
},
{
"content": "attach",
"role": "assistant"
}
] |
def x(self, lun_or_snap, cg_member=None):
hlu = self.get_hlu(lun_or_snap, cg_member=cg_member)
return hlu is not None
|
has_hlu
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, lun_or_snap, cg_member=None):\n hlu = self.get_hlu(lun_or_snap, cg_member=cg_member)\n return hlu is not None\n",
"role": "user"
},
{
"content": "has_hlu",
"role": "assistant"
}
] |
def x(self, lun_or_snap, cg_member=None):
import storops.unity.resource.lun as lun_module
import storops.unity.resource.snap as snap_module
which = None
if isinstance(lun_or_snap, lun_module.UnityLun):
which = self._get_host_luns(lun=lun_or_snap)
elif isinstance(lun_or_snap, snap_module.UnitySnap):
if lun_or_snap.is_cg_snap():
if cg_member is None:
log.debug(
'None host lun for CG snap {}. Use its member snap instead or pass in cg_member.'
.format(lun_or_snap.id))
return None
lun_or_snap = lun_or_snap.get_member_snap(cg_member)
which = self._get_host_luns(lun=cg_member, snap=lun_or_snap)
else:
which = self._get_host_luns(snap=lun_or_snap)
if not which:
log.debug('Resource(LUN or Snap) {} is not attached to host {}'.
format(lun_or_snap.name, self.name))
return None
return which[0]
|
get_host_lun
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, lun_or_snap, cg_member=None):\n import storops.unity.resource.lun as lun_module\n import storops.unity.resource.snap as snap_module\n which = None\n if isinstance(lun_or_snap, lun_module.UnityLun):\n which = self._get_host_luns(lun=lun_or_snap)\n elif isinstance(lun_or_snap, snap_module.UnitySnap):\n if lun_or_snap.is_cg_snap():\n if cg_member is None:\n log.debug(\n 'None host lun for CG snap {}. Use its member snap instead or pass in cg_member.'\n .format(lun_or_snap.id))\n return None\n lun_or_snap = lun_or_snap.get_member_snap(cg_member)\n which = self._get_host_luns(lun=cg_member, snap=lun_or_snap)\n else:\n which = self._get_host_luns(snap=lun_or_snap)\n if not which:\n log.debug('Resource(LUN or Snap) {} is not attached to host {}'.\n format(lun_or_snap.name, self.name))\n return None\n return which[0]\n",
"role": "user"
},
{
"content": "get_host_lun",
"role": "assistant"
}
] |
def x(self, resource, cg_member=None):
host_lun = self.get_host_lun(resource, cg_member=cg_member)
return host_lun if host_lun is None else host_lun.hlu
|
get_hlu
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, resource, cg_member=None):\n host_lun = self.get_host_lun(resource, cg_member=cg_member)\n return host_lun if host_lun is None else host_lun.hlu\n",
"role": "user"
},
{
"content": "get_hlu",
"role": "assistant"
}
] |
def x(self, iqns=None, wwns=None):
iqns = set(iqns) if iqns else set()
current_iqns = set()
if self.iscsi_host_initiators:
current_iqns = {initiator.initiator_id for initiator in self.
iscsi_host_initiators}
wwns = set(wwns) if wwns else set()
current_wwns = set()
if self.fc_host_initiators:
current_wwns = {initiator.initiator_id for initiator in self.
fc_host_initiators}
updater = UnityHostInitiatorUpdater(self, current_iqns | current_wwns,
iqns | wwns)
return updater.update()
|
update_initiators
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, iqns=None, wwns=None):\n iqns = set(iqns) if iqns else set()\n current_iqns = set()\n if self.iscsi_host_initiators:\n current_iqns = {initiator.initiator_id for initiator in self.\n iscsi_host_initiators}\n wwns = set(wwns) if wwns else set()\n current_wwns = set()\n if self.fc_host_initiators:\n current_wwns = {initiator.initiator_id for initiator in self.\n fc_host_initiators}\n updater = UnityHostInitiatorUpdater(self, current_iqns | current_wwns, \n iqns | wwns)\n return updater.update()\n",
"role": "user"
},
{
"content": "update_initiators",
"role": "assistant"
}
] |
def x(tag=None, sat_id=None, data_path=None, format_str=None,
supported_tags=None, fake_daily_files_from_monthly=False,
two_digit_year_break=None):
if data_path is not None:
if format_str is None:
try:
format_str = supported_tags[sat_id][tag]
except KeyError:
raise ValueError('Unknown tag')
out = pysat.Files.from_os(data_path=data_path, format_str=format_str)
if not out.empty and fake_daily_files_from_monthly:
out.ix[out.index[-1] + pds.DateOffset(months=1) - pds.
DateOffset(days=1)] = out.iloc[-1]
out = out.asfreq('D', 'pad')
out = out + '_' + out.index.strftime('%Y-%m-%d')
return out
return out
else:
estr = (
'A directory must be passed to the loading routine for <Instrument Code>'
)
raise ValueError(estr)
|
list_files
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tag=None, sat_id=None, data_path=None, format_str=None,\n supported_tags=None, fake_daily_files_from_monthly=False,\n two_digit_year_break=None):\n if data_path is not None:\n if format_str is None:\n try:\n format_str = supported_tags[sat_id][tag]\n except KeyError:\n raise ValueError('Unknown tag')\n out = pysat.Files.from_os(data_path=data_path, format_str=format_str)\n if not out.empty and fake_daily_files_from_monthly:\n out.ix[out.index[-1] + pds.DateOffset(months=1) - pds.\n DateOffset(days=1)] = out.iloc[-1]\n out = out.asfreq('D', 'pad')\n out = out + '_' + out.index.strftime('%Y-%m-%d')\n return out\n return out\n else:\n estr = (\n 'A directory must be passed to the loading routine for <Instrument Code>'\n )\n raise ValueError(estr)\n",
"role": "user"
},
{
"content": "list_files",
"role": "assistant"
}
] |
def x(fnames, tag=None, sat_id=None, fake_daily_files_from_monthly=False,
flatten_twod=True):
import pysatCDF
if len(fnames) <= 0:
return pysat.DataFrame(None), None
elif fake_daily_files_from_monthly:
fname = fnames[0][0:-11]
date = pysat.datetime.strptime(fnames[0][-10:], '%Y-%m-%d')
with pysatCDF.CDF(fname) as cdf:
data, meta = cdf.to_pysat(flatten_twod=flatten_twod)
data = data.ix[date:date + pds.DateOffset(days=1) - pds.
DateOffset(microseconds=1), :]
return data, meta
else:
with pysatCDF.CDF(fnames[0]) as cdf:
return cdf.to_pysat(flatten_twod=flatten_twod)
|
load
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fnames, tag=None, sat_id=None, fake_daily_files_from_monthly=False,\n flatten_twod=True):\n import pysatCDF\n if len(fnames) <= 0:\n return pysat.DataFrame(None), None\n elif fake_daily_files_from_monthly:\n fname = fnames[0][0:-11]\n date = pysat.datetime.strptime(fnames[0][-10:], '%Y-%m-%d')\n with pysatCDF.CDF(fname) as cdf:\n data, meta = cdf.to_pysat(flatten_twod=flatten_twod)\n data = data.ix[date:date + pds.DateOffset(days=1) - pds.\n DateOffset(microseconds=1), :]\n return data, meta\n else:\n with pysatCDF.CDF(fnames[0]) as cdf:\n return cdf.to_pysat(flatten_twod=flatten_twod)\n",
"role": "user"
},
{
"content": "load",
"role": "assistant"
}
] |
def x(supported_tags, date_array, tag, sat_id, ftp_site=
'cdaweb.gsfc.nasa.gov', data_path=None, user=None, password=None,
fake_daily_files_from_monthly=False):
import os
import ftplib
ftp = ftplib.FTP(ftp_site)
ftp.login()
try:
ftp_dict = supported_tags[tag]
except KeyError:
raise ValueError('Tag name unknown.')
ftp.cwd(ftp_dict['dir'])
remote_fname = ftp_dict['remote_fname']
local_fname = ftp_dict['local_fname']
for date in date_array:
formatted_remote_fname = remote_fname.format(year=date.year, month=
date.month, day=date.day)
formatted_local_fname = local_fname.format(year=date.year, month=
date.month, day=date.day)
saved_local_fname = os.path.join(data_path, formatted_local_fname)
try:
print('Attempting to download file for ' + date.strftime('%x'))
sys.stdout.flush()
ftp.retrbinary('RETR ' + formatted_remote_fname, open(
saved_local_fname, 'wb').write)
print('Finished.')
except ftplib.error_perm as exception:
if str(exception.args[0]).split(' ', 1)[0] != '550':
raise
else:
os.remove(saved_local_fname)
print('File not available for ' + date.strftime('%x'))
ftp.close()
|
download
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(supported_tags, date_array, tag, sat_id, ftp_site=\n 'cdaweb.gsfc.nasa.gov', data_path=None, user=None, password=None,\n fake_daily_files_from_monthly=False):\n import os\n import ftplib\n ftp = ftplib.FTP(ftp_site)\n ftp.login()\n try:\n ftp_dict = supported_tags[tag]\n except KeyError:\n raise ValueError('Tag name unknown.')\n ftp.cwd(ftp_dict['dir'])\n remote_fname = ftp_dict['remote_fname']\n local_fname = ftp_dict['local_fname']\n for date in date_array:\n formatted_remote_fname = remote_fname.format(year=date.year, month=\n date.month, day=date.day)\n formatted_local_fname = local_fname.format(year=date.year, month=\n date.month, day=date.day)\n saved_local_fname = os.path.join(data_path, formatted_local_fname)\n try:\n print('Attempting to download file for ' + date.strftime('%x'))\n sys.stdout.flush()\n ftp.retrbinary('RETR ' + formatted_remote_fname, open(\n saved_local_fname, 'wb').write)\n print('Finished.')\n except ftplib.error_perm as exception:\n if str(exception.args[0]).split(' ', 1)[0] != '550':\n raise\n else:\n os.remove(saved_local_fname)\n print('File not available for ' + date.strftime('%x'))\n ftp.close()\n",
"role": "user"
},
{
"content": "download",
"role": "assistant"
}
] |
def x(response, method, url):
req_id = response.headers.get('x-openstack-request-id')
if not req_id:
req_id = response.headers.get('x-compute-request-id')
kwargs = {'http_status': response.status_code, 'response': response,
'method': method, 'url': url, 'request_id': req_id}
if 'retry-after' in response.headers:
kwargs['retry_after'] = response.headers['retry-after']
content_type = response.headers.get('Content-Type', '')
if content_type.startswith('application/json'):
try:
body = response.json()
except ValueError:
pass
else:
if isinstance(body, dict):
error = body.get(list(body)[0])
if isinstance(error, dict):
kwargs['message'] = error.get('message') or error.get(
'faultstring')
kwargs['details'] = error.get('details') or six.text_type(
body)
elif content_type.startswith('text/'):
kwargs['details'] = response.text
try:
cls = _code_map[response.status_code]
except KeyError:
if 500 <= response.status_code < 600:
cls = HttpServerError
elif 400 <= response.status_code < 500:
cls = HTTPClientError
else:
cls = HttpError
return cls(**kwargs)
|
from_response
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(response, method, url):\n req_id = response.headers.get('x-openstack-request-id')\n if not req_id:\n req_id = response.headers.get('x-compute-request-id')\n kwargs = {'http_status': response.status_code, 'response': response,\n 'method': method, 'url': url, 'request_id': req_id}\n if 'retry-after' in response.headers:\n kwargs['retry_after'] = response.headers['retry-after']\n content_type = response.headers.get('Content-Type', '')\n if content_type.startswith('application/json'):\n try:\n body = response.json()\n except ValueError:\n pass\n else:\n if isinstance(body, dict):\n error = body.get(list(body)[0])\n if isinstance(error, dict):\n kwargs['message'] = error.get('message') or error.get(\n 'faultstring')\n kwargs['details'] = error.get('details') or six.text_type(\n body)\n elif content_type.startswith('text/'):\n kwargs['details'] = response.text\n try:\n cls = _code_map[response.status_code]\n except KeyError:\n if 500 <= response.status_code < 600:\n cls = HttpServerError\n elif 400 <= response.status_code < 500:\n cls = HTTPClientError\n else:\n cls = HttpError\n return cls(**kwargs)\n",
"role": "user"
},
{
"content": "from_response",
"role": "assistant"
}
] |
def x(self, name, raid_groups, description=None, **kwargs):
return UnityPool.create(self._cli, name=name, description=description,
raid_groups=raid_groups, **kwargs)
|
create_pool
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name, raid_groups, description=None, **kwargs):\n return UnityPool.create(self._cli, name=name, description=description,\n raid_groups=raid_groups, **kwargs)\n",
"role": "user"
},
{
"content": "create_pool",
"role": "assistant"
}
] |
def x(self):
eths = self.get_ethernet_port(bond=False)
las = self.get_link_aggregation()
return eths + las
|
get_file_port
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n eths = self.get_ethernet_port(bond=False)\n las = self.get_link_aggregation()\n return eths + las\n",
"role": "user"
},
{
"content": "get_file_port",
"role": "assistant"
}
] |
def x(self, management_address, local_username=None, local_password=None,
remote_username=None, remote_password=None, connection_type=None):
return UnityRemoteSystem.create(self._cli, management_address,
local_username=local_username, local_password=local_password,
remote_username=remote_username, remote_password=remote_password,
connection_type=connection_type)
|
create_remote_system
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, management_address, local_username=None, local_password=None,\n remote_username=None, remote_password=None, connection_type=None):\n return UnityRemoteSystem.create(self._cli, management_address,\n local_username=local_username, local_password=local_password,\n remote_username=remote_username, remote_password=remote_password,\n connection_type=connection_type)\n",
"role": "user"
},
{
"content": "create_remote_system",
"role": "assistant"
}
] |
def x(self, sp, ip_port, ip_address, netmask=None, v6_prefix_length=None,
gateway=None, vlan_id=None):
return UnityReplicationInterface.create(self._cli, sp, ip_port,
ip_address, netmask=netmask, v6_prefix_length=v6_prefix_length,
gateway=gateway, vlan_id=vlan_id)
|
create_replication_interface
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, sp, ip_port, ip_address, netmask=None, v6_prefix_length=None,\n gateway=None, vlan_id=None):\n return UnityReplicationInterface.create(self._cli, sp, ip_port,\n ip_address, netmask=netmask, v6_prefix_length=v6_prefix_length,\n gateway=gateway, vlan_id=vlan_id)\n",
"role": "user"
},
{
"content": "create_replication_interface",
"role": "assistant"
}
] |
def x(self, res_id, backup_snap=None):
name = self._get_name()
out = self._cli.restore_snap(name, res_id, backup_snap)
ex.raise_if_err(out, 'failed to restore snap {}.'.format(name), default
=ex.VNXSnapError)
|
restore
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, res_id, backup_snap=None):\n name = self._get_name()\n out = self._cli.restore_snap(name, res_id, backup_snap)\n ex.raise_if_err(out, 'failed to restore snap {}.'.format(name), default\n =ex.VNXSnapError)\n",
"role": "user"
},
{
"content": "restore",
"role": "assistant"
}
] |
def x(tag='', sat_id=None, data_path=None, format_str=None):
if format_str is None and data_path is not None:
file_base = 'supermag_magnetometer'
if tag == 'indices' or tag == 'all':
file_base += '_all'
if tag == 'indices':
psplit = path.split(data_path[:-1])
data_path = path.join(psplit[0], 'all', '')
if tag == 'stations':
min_fmt = '_'.join([file_base, '{year:4d}.???'])
doff = pds.DateOffset(years=1)
else:
min_fmt = '_'.join([file_base, '{year:4d}{month:02d}{day:02d}.???']
)
doff = pds.DateOffset(days=1)
files = pysat.Files.from_os(data_path=data_path, format_str=min_fmt)
if not files.empty:
files = files.sort_index()
if tag == 'stations':
orig_files = files.copy()
new_files = []
for orig in orig_files.iteritems():
files.ix[orig[0] + doff - pds.DateOffset(days=1)] = orig[1]
files = files.sort_index()
new_files.append(files.ix[orig[0]:orig[0] + doff - pds.
DateOffset(days=1)].asfreq('D', method='pad'))
files = pds.concat(new_files)
files = files.dropna()
files = files.sort_index()
files = files + '_' + files.index.strftime('%Y-%m-%d')
return files
elif format_str is None:
estr = 'A directory must be passed to the loading routine for SuperMAG'
raise ValueError(estr)
else:
return pysat.Files.from_os(data_path=data_path, format_str=format_str)
|
list_files
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tag='', sat_id=None, data_path=None, format_str=None):\n if format_str is None and data_path is not None:\n file_base = 'supermag_magnetometer'\n if tag == 'indices' or tag == 'all':\n file_base += '_all'\n if tag == 'indices':\n psplit = path.split(data_path[:-1])\n data_path = path.join(psplit[0], 'all', '')\n if tag == 'stations':\n min_fmt = '_'.join([file_base, '{year:4d}.???'])\n doff = pds.DateOffset(years=1)\n else:\n min_fmt = '_'.join([file_base, '{year:4d}{month:02d}{day:02d}.???']\n )\n doff = pds.DateOffset(days=1)\n files = pysat.Files.from_os(data_path=data_path, format_str=min_fmt)\n if not files.empty:\n files = files.sort_index()\n if tag == 'stations':\n orig_files = files.copy()\n new_files = []\n for orig in orig_files.iteritems():\n files.ix[orig[0] + doff - pds.DateOffset(days=1)] = orig[1]\n files = files.sort_index()\n new_files.append(files.ix[orig[0]:orig[0] + doff - pds.\n DateOffset(days=1)].asfreq('D', method='pad'))\n files = pds.concat(new_files)\n files = files.dropna()\n files = files.sort_index()\n files = files + '_' + files.index.strftime('%Y-%m-%d')\n return files\n elif format_str is None:\n estr = 'A directory must be passed to the loading routine for SuperMAG'\n raise ValueError(estr)\n else:\n return pysat.Files.from_os(data_path=data_path, format_str=format_str)\n",
"role": "user"
},
{
"content": "list_files",
"role": "assistant"
}
] |
def x(fnames, tag='', sat_id=None):
if len(fnames) <= 0:
return pysat.DataFrame(None), pysat.Meta(None)
if isinstance(fnames, str):
fnames = [fnames]
data = pds.DataFrame()
baseline = list()
for fname in fnames:
fname = fname[:-11]
file_type = path.splitext(fname)[1].lower()
if file_type == '.csv':
if tag != 'indices':
temp = load_csv_data(fname, tag)
else:
temp, bline = load_ascii_data(fname, tag)
if bline is not None:
baseline.append(bline)
if len(temp.columns) > 0:
data = pds.concat([data, temp], axis=0)
del temp
if len(data.columns) > 0:
meta = pysat.Meta()
for cc in data.columns:
meta[cc] = update_smag_metadata(cc)
meta.info = {'baseline': format_baseline_list(baseline)}
else:
meta = pysat.Meta(None)
return data, meta
|
load
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fnames, tag='', sat_id=None):\n if len(fnames) <= 0:\n return pysat.DataFrame(None), pysat.Meta(None)\n if isinstance(fnames, str):\n fnames = [fnames]\n data = pds.DataFrame()\n baseline = list()\n for fname in fnames:\n fname = fname[:-11]\n file_type = path.splitext(fname)[1].lower()\n if file_type == '.csv':\n if tag != 'indices':\n temp = load_csv_data(fname, tag)\n else:\n temp, bline = load_ascii_data(fname, tag)\n if bline is not None:\n baseline.append(bline)\n if len(temp.columns) > 0:\n data = pds.concat([data, temp], axis=0)\n del temp\n if len(data.columns) > 0:\n meta = pysat.Meta()\n for cc in data.columns:\n meta[cc] = update_smag_metadata(cc)\n meta.info = {'baseline': format_baseline_list(baseline)}\n else:\n meta = pysat.Meta(None)\n return data, meta\n",
"role": "user"
},
{
"content": "load",
"role": "assistant"
}
] |
def x(fname, tag):
import re
if tag == 'stations':
ddict = dict()
dkeys = list()
date_list = list()
with open(fname, 'r') as fopen:
dtime = pds.datetime.strptime(fname.split('_')[-1].split('.')[0
], '%Y')
for fline in fopen.readlines():
sline = [ll for ll in re.split('[,\\n]+', fline) if len(ll) > 0
]
if len(ddict.items()) == 0:
for kk in sline:
kk = re.sub('-', '_', kk)
ddict[kk] = list()
dkeys.append(kk)
else:
date_list.append(dtime)
for i, ll in enumerate(sline):
if i >= 1 and i <= 4:
ddict[dkeys[i]].append(float(ll))
elif i == 6:
ddict[dkeys[i]].append(int(ll))
elif i < len(dkeys):
ddict[dkeys[i]].append(ll)
else:
ddict[dkeys[-1]][-1] += ' {:s}'.format(ll)
data = pds.DataFrame(ddict, index=date_list, columns=ddict.keys())
else:
def x(dd):
return pysat.datetime.strptime(dd, '%Y-%m-%d %H:%M:%S')
data = pds.read_csv(fname, parse_dates={'datetime': [0]},
date_parser=parse_smag_date, index_col='datetime')
return data
|
load_csv_data
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fname, tag):\n import re\n if tag == 'stations':\n ddict = dict()\n dkeys = list()\n date_list = list()\n with open(fname, 'r') as fopen:\n dtime = pds.datetime.strptime(fname.split('_')[-1].split('.')[0\n ], '%Y')\n for fline in fopen.readlines():\n sline = [ll for ll in re.split('[,\\\\n]+', fline) if len(ll) > 0\n ]\n if len(ddict.items()) == 0:\n for kk in sline:\n kk = re.sub('-', '_', kk)\n ddict[kk] = list()\n dkeys.append(kk)\n else:\n date_list.append(dtime)\n for i, ll in enumerate(sline):\n if i >= 1 and i <= 4:\n ddict[dkeys[i]].append(float(ll))\n elif i == 6:\n ddict[dkeys[i]].append(int(ll))\n elif i < len(dkeys):\n ddict[dkeys[i]].append(ll)\n else:\n ddict[dkeys[-1]][-1] += ' {:s}'.format(ll)\n data = pds.DataFrame(ddict, index=date_list, columns=ddict.keys())\n else:\n\n def x(dd):\n return pysat.datetime.strptime(dd, '%Y-%m-%d %H:%M:%S')\n data = pds.read_csv(fname, parse_dates={'datetime': [0]},\n date_parser=parse_smag_date, index_col='datetime')\n return data\n",
"role": "user"
},
{
"content": "load_csv_data",
"role": "assistant"
}
] |
def x(fname, tag):
import re
ndata = {'indices': 2, '': 4, 'all': 4, 'stations': 8}
dkeys = {'stations': list(), '': ['IAGA', 'N', 'E', 'Z']}
data = pds.DataFrame(None)
baseline = None
if not tag in ndata.keys():
return data, baseline
with open(fname, 'r') as fopen:
hflag = True
pflag = False
dflag = False if tag == 'stations' else True
snum = 0
ddict = dict()
date_list = list()
if tag == 'stations':
dtime = pds.datetime.strptime(fname.split('_')[-1].split('.')[0
], '%Y')
for fline in fopen.readlines():
line_len = len(fline)
if hflag:
if pflag:
pflag = False
if fline.find('-mlt') > 0:
ndata[''] += 2
dkeys[''].extend(['MLT', 'MLAT'])
if fline.find('-sza') > 0:
ndata[''] += 1
dkeys[''].append('SZA')
if fline.find('-decl') > 0:
ndata[''] += 1
dkeys[''].append('IGRF_DECL')
if tag == 'indices' and fline.find('-envelope') < 0:
break
lsplit = fline.split()
idelta = lsplit.index('-delta') + 1
ibase = lsplit.index('-baseline') + 1
isd = lsplit.index('-sd') + 1
ist = lsplit.index('-st') + 1
iex = lsplit.index('-ex') + 1
baseline = ' '.join([lsplit[ibase], lsplit[idelta],
lsplit[isd], lsplit[ist], lsplit[iex]])
if fline.find('Selected parameters:') >= 0:
pflag = True
if fline.count('=') == line_len - 1 and line_len > 2:
hflag = False
else:
lsplit = [ll for ll in re.split('[\\t\\n]+', fline) if len(
ll) > 0]
if dflag:
dflag = False
dstring = ' '.join(lsplit[:6])
dtime = pysat.datetime.strptime(dstring,
'%Y %m %d %H %M %S')
snum = int(lsplit[6])
if tag == 'indices':
date_list.append(dtime)
else:
date_list.extend([dtime for i in range(snum)])
elif len(lsplit) == ndata['indices']:
if tag is not '':
if lsplit[0] not in ddict.keys():
ddict[lsplit[0]] = list()
if tag == 'indices':
ddict[lsplit[0]].append(int(lsplit[1]))
else:
ddict[lsplit[0]].extend([int(lsplit[1]) for i in
range(snum)])
elif tag == 'stations' and len(lsplit) >= ndata[tag]:
if len(dkeys[tag]) == 0:
for ll in lsplit:
ll = re.sub('-', '_', ll)
dkeys[tag].append(ll)
ddict[ll] = list()
else:
date_list.append(dtime)
for i, ll in enumerate(lsplit):
if i >= 1 and i <= 4:
ddict[dkeys[tag][i]].append(float(ll))
elif i == 6:
ddict[dkeys[tag][i]].append(int(ll))
elif i < len(dkeys[tag]):
ddict[dkeys[tag][i]].append(ll)
else:
ddict[dkeys[tag][-1]][-1] += ' {:s}'.format(ll)
elif len(lsplit) == ndata['']:
snum -= 1
if tag != 'indices':
if len(ddict.keys()) < ndata['']:
for kk in dkeys['']:
ddict[kk] = list()
for i, kk in enumerate(dkeys['']):
if i == 0:
ddict[kk].append(lsplit[i])
else:
ddict[kk].append(float(lsplit[i]))
if tag != 'stations' and snum == 0 and len(ddict.items()) >= 2:
dflag = True
data = pds.DataFrame(ddict, index=date_list, columns=ddict.keys())
fopen.close()
return data, baseline
|
load_ascii_data
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fname, tag):\n import re\n ndata = {'indices': 2, '': 4, 'all': 4, 'stations': 8}\n dkeys = {'stations': list(), '': ['IAGA', 'N', 'E', 'Z']}\n data = pds.DataFrame(None)\n baseline = None\n if not tag in ndata.keys():\n return data, baseline\n with open(fname, 'r') as fopen:\n hflag = True\n pflag = False\n dflag = False if tag == 'stations' else True\n snum = 0\n ddict = dict()\n date_list = list()\n if tag == 'stations':\n dtime = pds.datetime.strptime(fname.split('_')[-1].split('.')[0\n ], '%Y')\n for fline in fopen.readlines():\n line_len = len(fline)\n if hflag:\n if pflag:\n pflag = False\n if fline.find('-mlt') > 0:\n ndata[''] += 2\n dkeys[''].extend(['MLT', 'MLAT'])\n if fline.find('-sza') > 0:\n ndata[''] += 1\n dkeys[''].append('SZA')\n if fline.find('-decl') > 0:\n ndata[''] += 1\n dkeys[''].append('IGRF_DECL')\n if tag == 'indices' and fline.find('-envelope') < 0:\n break\n lsplit = fline.split()\n idelta = lsplit.index('-delta') + 1\n ibase = lsplit.index('-baseline') + 1\n isd = lsplit.index('-sd') + 1\n ist = lsplit.index('-st') + 1\n iex = lsplit.index('-ex') + 1\n baseline = ' '.join([lsplit[ibase], lsplit[idelta],\n lsplit[isd], lsplit[ist], lsplit[iex]])\n if fline.find('Selected parameters:') >= 0:\n pflag = True\n if fline.count('=') == line_len - 1 and line_len > 2:\n hflag = False\n else:\n lsplit = [ll for ll in re.split('[\\\\t\\\\n]+', fline) if len(\n ll) > 0]\n if dflag:\n dflag = False\n dstring = ' '.join(lsplit[:6])\n dtime = pysat.datetime.strptime(dstring,\n '%Y %m %d %H %M %S')\n snum = int(lsplit[6])\n if tag == 'indices':\n date_list.append(dtime)\n else:\n date_list.extend([dtime for i in range(snum)])\n elif len(lsplit) == ndata['indices']:\n if tag is not '':\n if lsplit[0] not in ddict.keys():\n ddict[lsplit[0]] = list()\n if tag == 'indices':\n ddict[lsplit[0]].append(int(lsplit[1]))\n else:\n ddict[lsplit[0]].extend([int(lsplit[1]) for i in\n range(snum)])\n elif tag == 'stations' and len(lsplit) >= ndata[tag]:\n if len(dkeys[tag]) == 0:\n for ll in lsplit:\n ll = re.sub('-', '_', ll)\n dkeys[tag].append(ll)\n ddict[ll] = list()\n else:\n date_list.append(dtime)\n for i, ll in enumerate(lsplit):\n if i >= 1 and i <= 4:\n ddict[dkeys[tag][i]].append(float(ll))\n elif i == 6:\n ddict[dkeys[tag][i]].append(int(ll))\n elif i < len(dkeys[tag]):\n ddict[dkeys[tag][i]].append(ll)\n else:\n ddict[dkeys[tag][-1]][-1] += ' {:s}'.format(ll)\n elif len(lsplit) == ndata['']:\n snum -= 1\n if tag != 'indices':\n if len(ddict.keys()) < ndata['']:\n for kk in dkeys['']:\n ddict[kk] = list()\n for i, kk in enumerate(dkeys['']):\n if i == 0:\n ddict[kk].append(lsplit[i])\n else:\n ddict[kk].append(float(lsplit[i]))\n if tag != 'stations' and snum == 0 and len(ddict.items()) >= 2:\n dflag = True\n data = pds.DataFrame(ddict, index=date_list, columns=ddict.keys())\n fopen.close()\n return data, baseline\n",
"role": "user"
},
{
"content": "load_ascii_data",
"role": "assistant"
}
] |
def x(col_name):
smag_units = {'IAGA': 'none', 'N': 'nT', 'E': 'nT', 'Z': 'nT', 'MLT':
'hours', 'MLAT': 'degrees', 'SZA': 'degrees', 'IGRF_DECL':
'degrees', 'SMU': 'none', 'SML': 'none', 'datetime':
'YYYY-MM-DD HH:MM:SS', 'GEOLON': 'degrees', 'GEOLAT': 'degrees',
'AACGMLON': 'degrees', 'AACGMLAT': 'degrees', 'STATION_NAME':
'none', 'OPERATOR_NUM': 'none', 'OPERATORS': 'none'}
smag_name = {'IAGA': 'Station Code', 'N':
'B along local magnetic North', 'E': 'B along local magnetic East',
'Z': 'B vertically downward', 'MLT': 'Magnetic Local Time', 'MLAT':
'Magnetic Latitude', 'SZA': 'Solar Zenith Angle', 'IGRF_DECL':
'IGRF magnetic declination', 'SMU':
"""Maximum eastward auroral electrojets strength.
Upper envelope of N-component for stations between 40 and 80 degrees magnetic north."""
, 'SML':
"""Maximum westward auroral electrojets strength.
Lower envelope of N-component for stations between 40 and 80 degrees magnetic north."""
, 'datetime': 'UT date and time', 'GEOLON': 'geographic longitude',
'GEOLAT': 'geographic latitude', 'AACGMLON':
'Altitude-Adjusted Corrected Geomagnetic longitude', 'AACGMLAT':
'Altitude-Adjusted Corrected Geomagnetic latitude', 'STATION_NAME':
'Long form station name', 'OPERATOR_NUM':
'Number of station operators', 'OPERATORS': 'Station operator name(s)'}
ackn = 'When using this data please include the following reference:\n'
ackn += 'Gjerloev, J. W., The SuperMAG data processing technique, '
ackn += 'Geophys. Res., 117, A09213, doi:10.1029/2012JA017683, 2012\n\n'
ackn += 'For publications and presentations, please include the following'
ackn += 'acknowledgement:\nFor the ground magnetometer data we gratefully '
ackn += 'acknowledge: Intermagnet; USGS, Jeffrey J. Love; CARISMA, PI Ian '
ackn += 'Mann; CANMOS; The S-RAMP Database, PI K. Yumoto and Dr. K. '
ackn += 'Shiokawa; The SPIDR database; AARI, PI Oleg Troshichev; The '
ackn += 'MACCS program, PI M. Engebretson, Geomagnetism Unit of the '
ackn += 'Geological Survey of Canada; GIMA; MEASURE, UCLA IGPP and Florida'
ackn += ' Institute of Technology; SAMBA, PI Eftyhia Zesta; 210 Chain, PI '
ackn += 'K. Yumoto; SAMNET, PI Farideh Honary; The institutes who maintain'
ackn += (
' the IMAGE magnetometer array, PI Eija Tanskanen; PENGUIN; AUTUMN,')
ackn += (
' PI Martin Connors; DTU Space, PI Dr. Rico Behlke; South Pole and ')
ackn += " McMurdo Magnetometer, PI's Louis J. Lanzarotti and Alan T. "
ackn += (
'Weatherwax; ICESTAR; RAPIDMAG; PENGUIn; British Artarctic Survey; ')
ackn += 'McMac, PI Dr. Peter Chi; BGS, PI Dr. Susan Macmillan; Pushkov '
ackn += 'Institute of Terrestrial Magnetism, Ionosphere and Radio Wave '
ackn += 'Propagation (IZMIRAN); GFZ, PI Dr. Juergen Matzka; MFGI, PI B. '
ackn += 'Heilig; IGFPAS, PI J. Reda; University of L’Aquila, PI M. '
ackn += 'Vellante; BCMT, V. Lesur and A. Chambodut; Data obtained in '
ackn += 'cooperation with Geoscience Australia, PI Marina Costelloe; '
ackn += 'SuperMAG, PI Jesper W. Gjerloev.'
col_dict = {'units': smag_units[col_name], 'long_name': smag_name[
col_name], 'acknowledgements': ackn}
return col_dict
|
update_smag_metadata
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(col_name):\n smag_units = {'IAGA': 'none', 'N': 'nT', 'E': 'nT', 'Z': 'nT', 'MLT':\n 'hours', 'MLAT': 'degrees', 'SZA': 'degrees', 'IGRF_DECL':\n 'degrees', 'SMU': 'none', 'SML': 'none', 'datetime':\n 'YYYY-MM-DD HH:MM:SS', 'GEOLON': 'degrees', 'GEOLAT': 'degrees',\n 'AACGMLON': 'degrees', 'AACGMLAT': 'degrees', 'STATION_NAME':\n 'none', 'OPERATOR_NUM': 'none', 'OPERATORS': 'none'}\n smag_name = {'IAGA': 'Station Code', 'N':\n 'B along local magnetic North', 'E': 'B along local magnetic East',\n 'Z': 'B vertically downward', 'MLT': 'Magnetic Local Time', 'MLAT':\n 'Magnetic Latitude', 'SZA': 'Solar Zenith Angle', 'IGRF_DECL':\n 'IGRF magnetic declination', 'SMU':\n \"\"\"Maximum eastward auroral electrojets strength.\nUpper envelope of N-component for stations between 40 and 80 degrees magnetic north.\"\"\"\n , 'SML':\n \"\"\"Maximum westward auroral electrojets strength.\nLower envelope of N-component for stations between 40 and 80 degrees magnetic north.\"\"\"\n , 'datetime': 'UT date and time', 'GEOLON': 'geographic longitude',\n 'GEOLAT': 'geographic latitude', 'AACGMLON':\n 'Altitude-Adjusted Corrected Geomagnetic longitude', 'AACGMLAT':\n 'Altitude-Adjusted Corrected Geomagnetic latitude', 'STATION_NAME':\n 'Long form station name', 'OPERATOR_NUM':\n 'Number of station operators', 'OPERATORS': 'Station operator name(s)'}\n ackn = 'When using this data please include the following reference:\\n'\n ackn += 'Gjerloev, J. W., The SuperMAG data processing technique, '\n ackn += 'Geophys. Res., 117, A09213, doi:10.1029/2012JA017683, 2012\\n\\n'\n ackn += 'For publications and presentations, please include the following'\n ackn += 'acknowledgement:\\nFor the ground magnetometer data we gratefully '\n ackn += 'acknowledge: Intermagnet; USGS, Jeffrey J. Love; CARISMA, PI Ian '\n ackn += 'Mann; CANMOS; The S-RAMP Database, PI K. Yumoto and Dr. K. '\n ackn += 'Shiokawa; The SPIDR database; AARI, PI Oleg Troshichev; The '\n ackn += 'MACCS program, PI M. Engebretson, Geomagnetism Unit of the '\n ackn += 'Geological Survey of Canada; GIMA; MEASURE, UCLA IGPP and Florida'\n ackn += ' Institute of Technology; SAMBA, PI Eftyhia Zesta; 210 Chain, PI '\n ackn += 'K. Yumoto; SAMNET, PI Farideh Honary; The institutes who maintain'\n ackn += (\n ' the IMAGE magnetometer array, PI Eija Tanskanen; PENGUIN; AUTUMN,')\n ackn += (\n ' PI Martin Connors; DTU Space, PI Dr. Rico Behlke; South Pole and ')\n ackn += \" McMurdo Magnetometer, PI's Louis J. Lanzarotti and Alan T. \"\n ackn += (\n 'Weatherwax; ICESTAR; RAPIDMAG; PENGUIn; British Artarctic Survey; ')\n ackn += 'McMac, PI Dr. Peter Chi; BGS, PI Dr. Susan Macmillan; Pushkov '\n ackn += 'Institute of Terrestrial Magnetism, Ionosphere and Radio Wave '\n ackn += 'Propagation (IZMIRAN); GFZ, PI Dr. Juergen Matzka; MFGI, PI B. '\n ackn += 'Heilig; IGFPAS, PI J. Reda; University of L’Aquila, PI M. '\n ackn += 'Vellante; BCMT, V. Lesur and A. Chambodut; Data obtained in '\n ackn += 'cooperation with Geoscience Australia, PI Marina Costelloe; '\n ackn += 'SuperMAG, PI Jesper W. Gjerloev.'\n col_dict = {'units': smag_units[col_name], 'long_name': smag_name[\n col_name], 'acknowledgements': ackn}\n return col_dict\n",
"role": "user"
},
{
"content": "update_smag_metadata",
"role": "assistant"
}
] |
def x(baseline_list):
uniq_base = dict()
uniq_delta = dict()
for bline in baseline_list:
bsplit = bline.split()
bdate = ' '.join(bsplit[2:])
if bsplit[0] not in uniq_base.keys():
uniq_base[bsplit[0]] = ''
if bsplit[1] not in uniq_delta.keys():
uniq_delta[bsplit[1]] = ''
uniq_base[bsplit[0]] += '{:s}, '.format(bdate)
uniq_delta[bsplit[1]] += '{:s}, '.format(bdate)
if len(uniq_base.items()) == 1:
base_string = 'Baseline {:s}'.format(list(uniq_base.keys())[0])
else:
base_string = 'Baseline '
for i, kk in enumerate(uniq_base.keys()):
if i == 1:
base_string += '{:s}: {:s}'.format(kk, uniq_base[kk][:-2])
else:
base_string += ' {:s}: {:s}'.format(kk, uniq_base[
kk][:-2])
else:
base_string += 'unknown'
if len(uniq_delta.items()) == 1:
base_string += '\nDelta {:s}'.format(list(uniq_delta.keys())[0])
else:
base_string += '\nDelta '
for i, kk in enumerate(uniq_delta.keys()):
if i == 1:
base_string += '{:s}: {:s}'.format(kk, uniq_delta[kk][:-2])
else:
base_string += ' {:s}: {:s}'.format(kk, uniq_delta[
kk][:-2])
else:
base_string += 'unknown'
return base_string
|
format_baseline_list
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(baseline_list):\n uniq_base = dict()\n uniq_delta = dict()\n for bline in baseline_list:\n bsplit = bline.split()\n bdate = ' '.join(bsplit[2:])\n if bsplit[0] not in uniq_base.keys():\n uniq_base[bsplit[0]] = ''\n if bsplit[1] not in uniq_delta.keys():\n uniq_delta[bsplit[1]] = ''\n uniq_base[bsplit[0]] += '{:s}, '.format(bdate)\n uniq_delta[bsplit[1]] += '{:s}, '.format(bdate)\n if len(uniq_base.items()) == 1:\n base_string = 'Baseline {:s}'.format(list(uniq_base.keys())[0])\n else:\n base_string = 'Baseline '\n for i, kk in enumerate(uniq_base.keys()):\n if i == 1:\n base_string += '{:s}: {:s}'.format(kk, uniq_base[kk][:-2])\n else:\n base_string += ' {:s}: {:s}'.format(kk, uniq_base[\n kk][:-2])\n else:\n base_string += 'unknown'\n if len(uniq_delta.items()) == 1:\n base_string += '\\nDelta {:s}'.format(list(uniq_delta.keys())[0])\n else:\n base_string += '\\nDelta '\n for i, kk in enumerate(uniq_delta.keys()):\n if i == 1:\n base_string += '{:s}: {:s}'.format(kk, uniq_delta[kk][:-2])\n else:\n base_string += ' {:s}: {:s}'.format(kk, uniq_delta[\n kk][:-2])\n else:\n base_string += 'unknown'\n return base_string\n",
"role": "user"
},
{
"content": "format_baseline_list",
"role": "assistant"
}
] |
def x(file_strings, file_fmt, tag):
if file_fmt.lower() == 'csv':
return append_csv_data(file_strings)
else:
return append_ascii_data(file_strings, tag)
|
append_data
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(file_strings, file_fmt, tag):\n if file_fmt.lower() == 'csv':\n return append_csv_data(file_strings)\n else:\n return append_ascii_data(file_strings, tag)\n",
"role": "user"
},
{
"content": "append_data",
"role": "assistant"
}
] |
def x(file_strings, tag):
import re
out_lines = file_strings[0].split('\n')
iparam = -1
ihead = -1
idates = list()
date_list = list()
num_stations = list()
ind_num = 2 if tag in ['all', 'indices', ''] else 0
for i, line in enumerate(out_lines):
if line == 'Selected parameters:':
iparam = i + 1
elif line.count('=') == len(line) and len(line) > 2:
ihead = i
break
i = ihead + 1
while i < len(out_lines) - 1:
idates.append(i)
lsplit = re.split('\t+', out_lines[i])
dtime = pds.datetime.strptime(' '.join(lsplit[0:-1]),
'%Y %m %d %H %M %S')
date_list.append(dtime)
num_stations.append(int(lsplit[-1]))
i += num_stations[-1] + 1 + ind_num
idates = np.array(idates)
station_names = list()
for ff in range(len(file_strings) - 1):
file_lines = file_strings[ff + 1].split('\n')
head = True
snum = 0
for i, line in enumerate(file_lines):
if head:
if line.count('=') == len(line) and len(line) > 2:
head = False
elif len(line) > 0:
lsplit = re.split('\t+', line)
if snum == 0:
dtime = pds.datetime.strptime(' '.join(lsplit[0:-1]),
'%Y %m %d %H %M %S')
try:
idate = date_list.index(dtime)
except:
raise ValueError('Unexpected date ', dtime)
snum = int(lsplit[-1])
onum = num_stations[idate]
inum = ind_num
idates[idate + 1:] += snum
num_stations[idate] += snum
oline = '{:s}\t{:d}'.format(dtime.strftime(
'%Y\t%m\t%d\t%H\t%M\t%S'), num_stations[idate])
out_lines[idates[idate]] = oline
elif inum > 0:
inum -= 1
else:
onum += 1
snum -= 1
out_lines.insert(idates[idate] + onum, line)
if not lsplit[0] in station_names:
station_names.append(lsplit[0])
out_lines[iparam] += ',' + ','.join(station_names)
out_string = '\n'.join(out_lines)
return out_string
|
append_ascii_data
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(file_strings, tag):\n import re\n out_lines = file_strings[0].split('\\n')\n iparam = -1\n ihead = -1\n idates = list()\n date_list = list()\n num_stations = list()\n ind_num = 2 if tag in ['all', 'indices', ''] else 0\n for i, line in enumerate(out_lines):\n if line == 'Selected parameters:':\n iparam = i + 1\n elif line.count('=') == len(line) and len(line) > 2:\n ihead = i\n break\n i = ihead + 1\n while i < len(out_lines) - 1:\n idates.append(i)\n lsplit = re.split('\\t+', out_lines[i])\n dtime = pds.datetime.strptime(' '.join(lsplit[0:-1]),\n '%Y %m %d %H %M %S')\n date_list.append(dtime)\n num_stations.append(int(lsplit[-1]))\n i += num_stations[-1] + 1 + ind_num\n idates = np.array(idates)\n station_names = list()\n for ff in range(len(file_strings) - 1):\n file_lines = file_strings[ff + 1].split('\\n')\n head = True\n snum = 0\n for i, line in enumerate(file_lines):\n if head:\n if line.count('=') == len(line) and len(line) > 2:\n head = False\n elif len(line) > 0:\n lsplit = re.split('\\t+', line)\n if snum == 0:\n dtime = pds.datetime.strptime(' '.join(lsplit[0:-1]),\n '%Y %m %d %H %M %S')\n try:\n idate = date_list.index(dtime)\n except:\n raise ValueError('Unexpected date ', dtime)\n snum = int(lsplit[-1])\n onum = num_stations[idate]\n inum = ind_num\n idates[idate + 1:] += snum\n num_stations[idate] += snum\n oline = '{:s}\\t{:d}'.format(dtime.strftime(\n '%Y\\t%m\\t%d\\t%H\\t%M\\t%S'), num_stations[idate])\n out_lines[idates[idate]] = oline\n elif inum > 0:\n inum -= 1\n else:\n onum += 1\n snum -= 1\n out_lines.insert(idates[idate] + onum, line)\n if not lsplit[0] in station_names:\n station_names.append(lsplit[0])\n out_lines[iparam] += ',' + ','.join(station_names)\n out_string = '\\n'.join(out_lines)\n return out_string\n",
"role": "user"
},
{
"content": "append_ascii_data",
"role": "assistant"
}
] |
def x(file_strings):
out_lines = list()
head_line = None
for fstrings in file_strings:
file_lines = fstrings.split('\n')
head_line = file_lines.pop(0)
out_lines.extend(file_lines)
out_lines.sort()
i = 0
while len(out_lines[i]) == 0:
out_lines.pop(i)
out_lines.insert(0, head_line)
out_lines.append('')
out_string = '\n'.join(out_lines)
return out_string
|
append_csv_data
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(file_strings):\n out_lines = list()\n head_line = None\n for fstrings in file_strings:\n file_lines = fstrings.split('\\n')\n head_line = file_lines.pop(0)\n out_lines.extend(file_lines)\n out_lines.sort()\n i = 0\n while len(out_lines[i]) == 0:\n out_lines.pop(i)\n out_lines.insert(0, head_line)\n out_lines.append('')\n out_string = '\\n'.join(out_lines)\n return out_string\n",
"role": "user"
},
{
"content": "append_csv_data",
"role": "assistant"
}
] |
def x(self):
self.custom.add(add_quasi_dipole_coordinates, 'modify')
self.custom.add(add_aacgm_coordinates, 'modify')
self.custom.add(calculate_ecef_velocity, 'modify')
self.custom.add(add_sc_attitude_vectors, 'modify')
self.custom.add(add_iri_thermal_plasma, 'modify')
self.custom.add(add_hwm_winds_and_ecef_vectors, 'modify')
self.custom.add(add_igrf, 'modify')
in_meta = {'desc': 'IGRF geomagnetic field expressed in the s/c basis.',
'units': 'nT'}
self.custom.add(project_ecef_vector_onto_sc, 'modify', 'end',
'B_ecef_x', 'B_ecef_y', 'B_ecef_z', 'B_sc_x', 'B_sc_y', 'B_sc_z',
meta=[in_meta.copy(), in_meta.copy(), in_meta.copy()])
self.custom.add(project_hwm_onto_sc, 'modify')
self.custom.add(add_msis, 'modify')
|
init
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.custom.add(add_quasi_dipole_coordinates, 'modify')\n self.custom.add(add_aacgm_coordinates, 'modify')\n self.custom.add(calculate_ecef_velocity, 'modify')\n self.custom.add(add_sc_attitude_vectors, 'modify')\n self.custom.add(add_iri_thermal_plasma, 'modify')\n self.custom.add(add_hwm_winds_and_ecef_vectors, 'modify')\n self.custom.add(add_igrf, 'modify')\n in_meta = {'desc': 'IGRF geomagnetic field expressed in the s/c basis.',\n 'units': 'nT'}\n self.custom.add(project_ecef_vector_onto_sc, 'modify', 'end',\n 'B_ecef_x', 'B_ecef_y', 'B_ecef_z', 'B_sc_x', 'B_sc_y', 'B_sc_z',\n meta=[in_meta.copy(), in_meta.copy(), in_meta.copy()])\n self.custom.add(project_hwm_onto_sc, 'modify')\n self.custom.add(add_msis, 'modify')\n",
"role": "user"
},
{
"content": "init",
"role": "assistant"
}
] |
def x(fnames, tag=None, sat_id=None, obs_long=0.0, obs_lat=0.0, obs_alt=0.0,
TLE1=None, TLE2=None):
import sgp4
from sgp4.earth_gravity import wgs72
from sgp4.io import twoline2rv
import ephem
import pysatMagVect
line1 = (
'1 25544U 98067A 18135.61844383 .00002728 00000-0 48567-4 0 9998'
)
line2 = (
'2 25544 51.6402 181.0633 0004018 88.8954 22.2246 15.54059185113452'
)
if TLE1 is not None:
line1 = TLE1
if TLE2 is not None:
line2 = TLE2
satellite = twoline2rv(line1, line2, wgs72)
parts = os.path.split(fnames[0])[-1].split('-')
yr = int(parts[0])
month = int(parts[1])
day = int(parts[2][0:2])
date = pysat.datetime(yr, month, day)
times = pds.date_range(start=date, end=date + pds.DateOffset(seconds=
86399), freq='1S')
on_travis = os.environ.get('ONTRAVIS') == 'True'
if on_travis:
times = times[0:100]
position = []
velocity = []
for time in times:
pos, vel = satellite.propagate(time.year, time.month, time.day,
time.hour, time.minute, time.second)
position.extend(pos)
velocity.extend(vel)
data = pysat.DataFrame({'position_eci_x': position[::3],
'position_eci_y': position[1::3], 'position_eci_z': position[2::3],
'velocity_eci_x': velocity[::3], 'velocity_eci_y': velocity[1::3],
'velocity_eci_z': velocity[2::3]}, index=times)
data.index.name = 'Epoch'
site = ephem.Observer()
site.lon = str(obs_long)
site.lat = str(obs_lat)
site.elevation = obs_alt
sat = ephem.readtle('pysat', line1, line2)
output_params = []
for time in times:
lp = {}
site.date = time
sat.compute(site)
lp['obs_sat_az_angle'] = ephem.degrees(sat.az)
lp['obs_sat_el_angle'] = ephem.degrees(sat.alt)
lp['obs_sat_slant_range'] = sat.range
lp['glat'] = np.degrees(sat.sublat)
lp['glong'] = np.degrees(sat.sublong)
lp['alt'] = sat.elevation / 1000.0
lp['x'], lp['y'], lp['z'] = pysatMagVect.geodetic_to_ecef(lp['glat'
], lp['glong'], lp['alt'])
output_params.append(lp)
output = pds.DataFrame(output_params, index=times)
data[['glong', 'glat', 'alt']] = output[['glong', 'glat', 'alt']]
data[['position_ecef_x', 'position_ecef_y', 'position_ecef_z']] = output[[
'x', 'y', 'z']]
data['obs_sat_az_angle'] = output['obs_sat_az_angle']
data['obs_sat_el_angle'] = output['obs_sat_el_angle']
data['obs_sat_slant_range'] = output['obs_sat_slant_range']
return data, meta.copy()
|
load
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fnames, tag=None, sat_id=None, obs_long=0.0, obs_lat=0.0, obs_alt=0.0,\n TLE1=None, TLE2=None):\n import sgp4\n from sgp4.earth_gravity import wgs72\n from sgp4.io import twoline2rv\n import ephem\n import pysatMagVect\n line1 = (\n '1 25544U 98067A 18135.61844383 .00002728 00000-0 48567-4 0 9998'\n )\n line2 = (\n '2 25544 51.6402 181.0633 0004018 88.8954 22.2246 15.54059185113452'\n )\n if TLE1 is not None:\n line1 = TLE1\n if TLE2 is not None:\n line2 = TLE2\n satellite = twoline2rv(line1, line2, wgs72)\n parts = os.path.split(fnames[0])[-1].split('-')\n yr = int(parts[0])\n month = int(parts[1])\n day = int(parts[2][0:2])\n date = pysat.datetime(yr, month, day)\n times = pds.date_range(start=date, end=date + pds.DateOffset(seconds=\n 86399), freq='1S')\n on_travis = os.environ.get('ONTRAVIS') == 'True'\n if on_travis:\n times = times[0:100]\n position = []\n velocity = []\n for time in times:\n pos, vel = satellite.propagate(time.year, time.month, time.day,\n time.hour, time.minute, time.second)\n position.extend(pos)\n velocity.extend(vel)\n data = pysat.DataFrame({'position_eci_x': position[::3],\n 'position_eci_y': position[1::3], 'position_eci_z': position[2::3],\n 'velocity_eci_x': velocity[::3], 'velocity_eci_y': velocity[1::3],\n 'velocity_eci_z': velocity[2::3]}, index=times)\n data.index.name = 'Epoch'\n site = ephem.Observer()\n site.lon = str(obs_long)\n site.lat = str(obs_lat)\n site.elevation = obs_alt\n sat = ephem.readtle('pysat', line1, line2)\n output_params = []\n for time in times:\n lp = {}\n site.date = time\n sat.compute(site)\n lp['obs_sat_az_angle'] = ephem.degrees(sat.az)\n lp['obs_sat_el_angle'] = ephem.degrees(sat.alt)\n lp['obs_sat_slant_range'] = sat.range\n lp['glat'] = np.degrees(sat.sublat)\n lp['glong'] = np.degrees(sat.sublong)\n lp['alt'] = sat.elevation / 1000.0\n lp['x'], lp['y'], lp['z'] = pysatMagVect.geodetic_to_ecef(lp['glat'\n ], lp['glong'], lp['alt'])\n output_params.append(lp)\n output = pds.DataFrame(output_params, index=times)\n data[['glong', 'glat', 'alt']] = output[['glong', 'glat', 'alt']]\n data[['position_ecef_x', 'position_ecef_y', 'position_ecef_z']] = output[[\n 'x', 'y', 'z']]\n data['obs_sat_az_angle'] = output['obs_sat_az_angle']\n data['obs_sat_el_angle'] = output['obs_sat_el_angle']\n data['obs_sat_slant_range'] = output['obs_sat_slant_range']\n return data, meta.copy()\n",
"role": "user"
},
{
"content": "load",
"role": "assistant"
}
] |
def x(tag=None, sat_id=None, data_path=None, format_str=None):
index = pds.date_range(pysat.datetime(2017, 12, 1), pysat.datetime(2018,
12, 1))
names = [(data_path + date.strftime('%Y-%m-%d') + '.nofile') for date in
index]
return pysat.Series(names, index=index)
|
list_files
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tag=None, sat_id=None, data_path=None, format_str=None):\n index = pds.date_range(pysat.datetime(2017, 12, 1), pysat.datetime(2018,\n 12, 1))\n names = [(data_path + date.strftime('%Y-%m-%d') + '.nofile') for date in\n index]\n return pysat.Series(names, index=index)\n",
"role": "user"
},
{
"content": "list_files",
"role": "assistant"
}
] |
def x(inst):
import pysatMagVect
inst['sc_xhat_ecef_x'], inst['sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'
] = pysatMagVect.normalize_vector(inst['velocity_ecef_x'], inst[
'velocity_ecef_y'], inst['velocity_ecef_z'])
inst['sc_zhat_ecef_x'], inst['sc_zhat_ecef_y'], inst['sc_zhat_ecef_z'
] = pysatMagVect.normalize_vector(-inst['position_ecef_x'], -inst[
'position_ecef_y'], -inst['position_ecef_z'])
inst['sc_yhat_ecef_x'], inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'
] = pysatMagVect.cross_product(inst['sc_zhat_ecef_x'], inst[
'sc_zhat_ecef_y'], inst['sc_zhat_ecef_z'], inst['sc_xhat_ecef_x'],
inst['sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'])
inst['sc_yhat_ecef_x'], inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'
] = pysatMagVect.normalize_vector(inst['sc_yhat_ecef_x'], inst[
'sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'])
inst['sc_zhat_ecef_x'], inst['sc_zhat_ecef_y'], inst['sc_zhat_ecef_z'
] = pysatMagVect.cross_product(inst['sc_xhat_ecef_x'], inst[
'sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'], inst['sc_yhat_ecef_x'],
inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'])
inst.meta['sc_xhat_ecef_x'] = {'units': '', 'desc':
'S/C attitude (x-direction, ram) unit vector, expressed in ECEF basis, x-component'
}
inst.meta['sc_xhat_ecef_y'] = {'units': '', 'desc':
'S/C attitude (x-direction, ram) unit vector, expressed in ECEF basis, y-component'
}
inst.meta['sc_xhat_ecef_z'] = {'units': '', 'desc':
'S/C attitude (x-direction, ram) unit vector, expressed in ECEF basis, z-component'
}
inst.meta['sc_zhat_ecef_x'] = {'units': '', 'desc':
'S/C attitude (z-direction, generally nadir) unit vector, expressed in ECEF basis, x-component'
}
inst.meta['sc_zhat_ecef_y'] = {'units': '', 'desc':
'S/C attitude (z-direction, generally nadir) unit vector, expressed in ECEF basis, y-component'
}
inst.meta['sc_zhat_ecef_z'] = {'units': '', 'desc':
'S/C attitude (z-direction, generally nadir) unit vector, expressed in ECEF basis, z-component'
}
inst.meta['sc_yhat_ecef_x'] = {'units': '', 'desc':
'S/C attitude (y-direction, generally south) unit vector, expressed in ECEF basis, x-component'
}
inst.meta['sc_yhat_ecef_y'] = {'units': '', 'desc':
'S/C attitude (y-direction, generally south) unit vector, expressed in ECEF basis, y-component'
}
inst.meta['sc_yhat_ecef_z'] = {'units': '', 'desc':
'S/C attitude (y-direction, generally south) unit vector, expressed in ECEF basis, z-component'
}
mag = np.sqrt(inst['sc_zhat_ecef_x'] ** 2 + inst['sc_zhat_ecef_y'] ** 2 +
inst['sc_zhat_ecef_z'] ** 2)
idx, = np.where((mag < 0.999999999) | (mag > 1.000000001))
if len(idx) > 0:
print(mag[idx])
raise RuntimeError(
'Unit vector generation failure. Not sufficently orthogonal.')
return
|
add_sc_attitude_vectors
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst):\n import pysatMagVect\n inst['sc_xhat_ecef_x'], inst['sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'\n ] = pysatMagVect.normalize_vector(inst['velocity_ecef_x'], inst[\n 'velocity_ecef_y'], inst['velocity_ecef_z'])\n inst['sc_zhat_ecef_x'], inst['sc_zhat_ecef_y'], inst['sc_zhat_ecef_z'\n ] = pysatMagVect.normalize_vector(-inst['position_ecef_x'], -inst[\n 'position_ecef_y'], -inst['position_ecef_z'])\n inst['sc_yhat_ecef_x'], inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'\n ] = pysatMagVect.cross_product(inst['sc_zhat_ecef_x'], inst[\n 'sc_zhat_ecef_y'], inst['sc_zhat_ecef_z'], inst['sc_xhat_ecef_x'],\n inst['sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'])\n inst['sc_yhat_ecef_x'], inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'\n ] = pysatMagVect.normalize_vector(inst['sc_yhat_ecef_x'], inst[\n 'sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'])\n inst['sc_zhat_ecef_x'], inst['sc_zhat_ecef_y'], inst['sc_zhat_ecef_z'\n ] = pysatMagVect.cross_product(inst['sc_xhat_ecef_x'], inst[\n 'sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'], inst['sc_yhat_ecef_x'],\n inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'])\n inst.meta['sc_xhat_ecef_x'] = {'units': '', 'desc':\n 'S/C attitude (x-direction, ram) unit vector, expressed in ECEF basis, x-component'\n }\n inst.meta['sc_xhat_ecef_y'] = {'units': '', 'desc':\n 'S/C attitude (x-direction, ram) unit vector, expressed in ECEF basis, y-component'\n }\n inst.meta['sc_xhat_ecef_z'] = {'units': '', 'desc':\n 'S/C attitude (x-direction, ram) unit vector, expressed in ECEF basis, z-component'\n }\n inst.meta['sc_zhat_ecef_x'] = {'units': '', 'desc':\n 'S/C attitude (z-direction, generally nadir) unit vector, expressed in ECEF basis, x-component'\n }\n inst.meta['sc_zhat_ecef_y'] = {'units': '', 'desc':\n 'S/C attitude (z-direction, generally nadir) unit vector, expressed in ECEF basis, y-component'\n }\n inst.meta['sc_zhat_ecef_z'] = {'units': '', 'desc':\n 'S/C attitude (z-direction, generally nadir) unit vector, expressed in ECEF basis, z-component'\n }\n inst.meta['sc_yhat_ecef_x'] = {'units': '', 'desc':\n 'S/C attitude (y-direction, generally south) unit vector, expressed in ECEF basis, x-component'\n }\n inst.meta['sc_yhat_ecef_y'] = {'units': '', 'desc':\n 'S/C attitude (y-direction, generally south) unit vector, expressed in ECEF basis, y-component'\n }\n inst.meta['sc_yhat_ecef_z'] = {'units': '', 'desc':\n 'S/C attitude (y-direction, generally south) unit vector, expressed in ECEF basis, z-component'\n }\n mag = np.sqrt(inst['sc_zhat_ecef_x'] ** 2 + inst['sc_zhat_ecef_y'] ** 2 +\n inst['sc_zhat_ecef_z'] ** 2)\n idx, = np.where((mag < 0.999999999) | (mag > 1.000000001))\n if len(idx) > 0:\n print(mag[idx])\n raise RuntimeError(\n 'Unit vector generation failure. Not sufficently orthogonal.')\n return\n",
"role": "user"
},
{
"content": "add_sc_attitude_vectors",
"role": "assistant"
}
] |
def x(inst):
x = inst['position_ecef_x']
vel_x = (x.values[2:] - x.values[0:-2]) / 2.0
y = inst['position_ecef_y']
vel_y = (y.values[2:] - y.values[0:-2]) / 2.0
z = inst['position_ecef_z']
vel_z = (z.values[2:] - z.values[0:-2]) / 2.0
inst[1:-1, 'velocity_ecef_x'] = vel_x
inst[1:-1, 'velocity_ecef_y'] = vel_y
inst[1:-1, 'velocity_ecef_z'] = vel_z
inst.meta['velocity_ecef_x'] = {'units': 'km/s', 'desc':
'Velocity of satellite calculated with respect to ECEF frame.'}
inst.meta['velocity_ecef_y'] = {'units': 'km/s', 'desc':
'Velocity of satellite calculated with respect to ECEF frame.'}
inst.meta['velocity_ecef_z'] = {'units': 'km/s', 'desc':
'Velocity of satellite calculated with respect to ECEF frame.'}
return
|
calculate_ecef_velocity
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst):\n x = inst['position_ecef_x']\n vel_x = (x.values[2:] - x.values[0:-2]) / 2.0\n y = inst['position_ecef_y']\n vel_y = (y.values[2:] - y.values[0:-2]) / 2.0\n z = inst['position_ecef_z']\n vel_z = (z.values[2:] - z.values[0:-2]) / 2.0\n inst[1:-1, 'velocity_ecef_x'] = vel_x\n inst[1:-1, 'velocity_ecef_y'] = vel_y\n inst[1:-1, 'velocity_ecef_z'] = vel_z\n inst.meta['velocity_ecef_x'] = {'units': 'km/s', 'desc':\n 'Velocity of satellite calculated with respect to ECEF frame.'}\n inst.meta['velocity_ecef_y'] = {'units': 'km/s', 'desc':\n 'Velocity of satellite calculated with respect to ECEF frame.'}\n inst.meta['velocity_ecef_z'] = {'units': 'km/s', 'desc':\n 'Velocity of satellite calculated with respect to ECEF frame.'}\n return\n",
"role": "user"
},
{
"content": "calculate_ecef_velocity",
"role": "assistant"
}
] |
def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):
import apexpy
ap = apexpy.Apex(date=inst.date)
qd_lat = []
qd_lon = []
mlt = []
for lat, lon, alt, time in zip(inst[glat_label], inst[glong_label],
inst[alt_label], inst.data.index):
tlat, tlon = ap.geo2qd(lat, lon, alt)
qd_lat.append(tlat)
qd_lon.append(tlon)
mlt.append(ap.mlon2mlt(tlon, time))
inst['qd_lat'] = qd_lat
inst['qd_long'] = qd_lon
inst['mlt'] = mlt
inst.meta['qd_lat'] = {'units': 'degrees', 'long_name':
'Quasi dipole latitude'}
inst.meta['qd_long'] = {'units': 'degrees', 'long_name':
'Quasi dipole longitude'}
inst.meta['qd_mlt'] = {'units': 'hrs', 'long_name': 'Magnetic local time'}
return
|
add_quasi_dipole_coordinates
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):\n import apexpy\n ap = apexpy.Apex(date=inst.date)\n qd_lat = []\n qd_lon = []\n mlt = []\n for lat, lon, alt, time in zip(inst[glat_label], inst[glong_label],\n inst[alt_label], inst.data.index):\n tlat, tlon = ap.geo2qd(lat, lon, alt)\n qd_lat.append(tlat)\n qd_lon.append(tlon)\n mlt.append(ap.mlon2mlt(tlon, time))\n inst['qd_lat'] = qd_lat\n inst['qd_long'] = qd_lon\n inst['mlt'] = mlt\n inst.meta['qd_lat'] = {'units': 'degrees', 'long_name':\n 'Quasi dipole latitude'}\n inst.meta['qd_long'] = {'units': 'degrees', 'long_name':\n 'Quasi dipole longitude'}\n inst.meta['qd_mlt'] = {'units': 'hrs', 'long_name': 'Magnetic local time'}\n return\n",
"role": "user"
},
{
"content": "add_quasi_dipole_coordinates",
"role": "assistant"
}
] |
def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):
import aacgmv2
aalat = []
aalon = []
mlt = []
for lat, lon, alt, time in zip(inst[glat_label], inst[glong_label],
inst[alt_label], inst.data.index):
tlat, tlon, tmlt = aacgmv2.get_aacgm_coord(lat, lon, alt, time)
aalat.append(tlat)
aalon.append(tlon)
mlt.append(tmlt)
inst['aacgm_lat'] = aalat
inst['aacgm_long'] = aalon
inst['aacgm_mlt'] = mlt
inst.meta['aacgm_lat'] = {'units': 'degrees', 'long_name': 'AACGM latitude'
}
inst.meta['aacgm_long'] = {'units': 'degrees', 'long_name':
'AACGM longitude'}
inst.meta['aacgm_mlt'] = {'units': 'hrs', 'long_name':
'AACGM Magnetic local time'}
return
|
add_aacgm_coordinates
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):\n import aacgmv2\n aalat = []\n aalon = []\n mlt = []\n for lat, lon, alt, time in zip(inst[glat_label], inst[glong_label],\n inst[alt_label], inst.data.index):\n tlat, tlon, tmlt = aacgmv2.get_aacgm_coord(lat, lon, alt, time)\n aalat.append(tlat)\n aalon.append(tlon)\n mlt.append(tmlt)\n inst['aacgm_lat'] = aalat\n inst['aacgm_long'] = aalon\n inst['aacgm_mlt'] = mlt\n inst.meta['aacgm_lat'] = {'units': 'degrees', 'long_name': 'AACGM latitude'\n }\n inst.meta['aacgm_long'] = {'units': 'degrees', 'long_name':\n 'AACGM longitude'}\n inst.meta['aacgm_mlt'] = {'units': 'hrs', 'long_name':\n 'AACGM Magnetic local time'}\n return\n",
"role": "user"
},
{
"content": "add_aacgm_coordinates",
"role": "assistant"
}
] |
def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):
import pyglow
from pyglow.pyglow import Point
iri_params = []
for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[
glong_label], inst[alt_label]):
pt = Point(time, lat, lon, alt)
pt.run_iri()
iri = {}
iri['ion_temp'] = pt.Ti
iri['e_temp'] = pt.Te
iri['ion_dens'] = pt.ni['O+'] + pt.ni['H+'] + pt.ni['HE+']
iri['frac_dens_o'] = pt.ni['O+'] / iri['ion_dens']
iri['frac_dens_h'] = pt.ni['H+'] / iri['ion_dens']
iri['frac_dens_he'] = pt.ni['HE+'] / iri['ion_dens']
iri_params.append(iri)
iri = pds.DataFrame(iri_params)
iri.index = inst.data.index
inst[iri.keys()] = iri
inst.meta['ion_temp'] = {'units': 'Kelvin', 'long_name': 'Ion Temperature'}
inst.meta['ion_dens'] = {'units': 'N/cc', 'long_name': 'Ion Density',
'desc': 'Total ion density including O+ and H+ from IRI model run.'}
inst.meta['frac_dens_o'] = {'units': '', 'long_name':
'Fractional O+ Density'}
inst.meta['frac_dens_h'] = {'units': '', 'long_name':
'Fractional H+ Density'}
|
add_iri_thermal_plasma
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):\n import pyglow\n from pyglow.pyglow import Point\n iri_params = []\n for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[\n glong_label], inst[alt_label]):\n pt = Point(time, lat, lon, alt)\n pt.run_iri()\n iri = {}\n iri['ion_temp'] = pt.Ti\n iri['e_temp'] = pt.Te\n iri['ion_dens'] = pt.ni['O+'] + pt.ni['H+'] + pt.ni['HE+']\n iri['frac_dens_o'] = pt.ni['O+'] / iri['ion_dens']\n iri['frac_dens_h'] = pt.ni['H+'] / iri['ion_dens']\n iri['frac_dens_he'] = pt.ni['HE+'] / iri['ion_dens']\n iri_params.append(iri)\n iri = pds.DataFrame(iri_params)\n iri.index = inst.data.index\n inst[iri.keys()] = iri\n inst.meta['ion_temp'] = {'units': 'Kelvin', 'long_name': 'Ion Temperature'}\n inst.meta['ion_dens'] = {'units': 'N/cc', 'long_name': 'Ion Density',\n 'desc': 'Total ion density including O+ and H+ from IRI model run.'}\n inst.meta['frac_dens_o'] = {'units': '', 'long_name':\n 'Fractional O+ Density'}\n inst.meta['frac_dens_h'] = {'units': '', 'long_name':\n 'Fractional H+ Density'}\n",
"role": "user"
},
{
"content": "add_iri_thermal_plasma",
"role": "assistant"
}
] |
def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):
import pyglow
import pysatMagVect
hwm_params = []
for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[
glong_label], inst[alt_label]):
pt = pyglow.Point(time, lat, lon, alt)
pt.run_hwm()
hwm = {}
hwm['zonal_wind'] = pt.u
hwm['meridional_wind'] = pt.v
hwm_params.append(hwm)
hwm = pds.DataFrame(hwm_params)
hwm.index = inst.data.index
inst[['zonal_wind', 'meridional_wind']] = hwm[['zonal_wind',
'meridional_wind']]
mag = np.sqrt(inst['position_ecef_x'] ** 2 + inst['position_ecef_y'] ** 2)
inst['unit_zonal_wind_ecef_x'] = -inst['position_ecef_y'] / mag
inst['unit_zonal_wind_ecef_y'] = inst['position_ecef_x'] / mag
inst['unit_zonal_wind_ecef_z'] = 0 * inst['position_ecef_x']
unit_pos_x, unit_pos_y, unit_pos_z = pysatMagVect.normalize_vector(-
inst['position_ecef_x'], -inst['position_ecef_y'], -inst[
'position_ecef_z'])
inst['unit_mer_wind_ecef_x'], inst['unit_mer_wind_ecef_y'], inst[
'unit_mer_wind_ecef_z'] = pysatMagVect.cross_product(unit_pos_x,
unit_pos_y, unit_pos_z, inst['unit_zonal_wind_ecef_x'], inst[
'unit_zonal_wind_ecef_y'], inst['unit_zonal_wind_ecef_z'])
inst.meta['zonal_wind'] = {'units': 'm/s', 'long_name': 'Zonal Wind',
'desc': 'HWM model zonal wind'}
inst.meta['meridional_wind'] = {'units': 'm/s', 'long_name':
'Meridional Wind', 'desc': 'HWM model meridional wind'}
inst.meta['unit_zonal_wind_ecef_x'] = {'units': '', 'long_name':
'Zonal Wind Unit ECEF x-vector', 'desc':
'x-value of zonal wind unit vector in ECEF co ordinates'}
inst.meta['unit_zonal_wind_ecef_y'] = {'units': '', 'long_name':
'Zonal Wind Unit ECEF y-vector', 'desc':
'y-value of zonal wind unit vector in ECEF co ordinates'}
inst.meta['unit_zonal_wind_ecef_z'] = {'units': '', 'long_name':
'Zonal Wind Unit ECEF z-vector', 'desc':
'z-value of zonal wind unit vector in ECEF co ordinates'}
inst.meta['unit_mer_wind_ecef_x'] = {'units': '', 'long_name':
'Meridional Wind Unit ECEF x-vector', 'desc':
'x-value of meridional wind unit vector in ECEF co ordinates'}
inst.meta['unit_mer_wind_ecef_y'] = {'units': '', 'long_name':
'Meridional Wind Unit ECEF y-vector', 'desc':
'y-value of meridional wind unit vector in ECEF co ordinates'}
inst.meta['unit_mer_wind_ecef_z'] = {'units': '', 'long_name':
'Meridional Wind Unit ECEF z-vector', 'desc':
'z-value of meridional wind unit vector in ECEF co ordinates'}
return
|
add_hwm_winds_and_ecef_vectors
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):\n import pyglow\n import pysatMagVect\n hwm_params = []\n for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[\n glong_label], inst[alt_label]):\n pt = pyglow.Point(time, lat, lon, alt)\n pt.run_hwm()\n hwm = {}\n hwm['zonal_wind'] = pt.u\n hwm['meridional_wind'] = pt.v\n hwm_params.append(hwm)\n hwm = pds.DataFrame(hwm_params)\n hwm.index = inst.data.index\n inst[['zonal_wind', 'meridional_wind']] = hwm[['zonal_wind',\n 'meridional_wind']]\n mag = np.sqrt(inst['position_ecef_x'] ** 2 + inst['position_ecef_y'] ** 2)\n inst['unit_zonal_wind_ecef_x'] = -inst['position_ecef_y'] / mag\n inst['unit_zonal_wind_ecef_y'] = inst['position_ecef_x'] / mag\n inst['unit_zonal_wind_ecef_z'] = 0 * inst['position_ecef_x']\n unit_pos_x, unit_pos_y, unit_pos_z = pysatMagVect.normalize_vector(-\n inst['position_ecef_x'], -inst['position_ecef_y'], -inst[\n 'position_ecef_z'])\n inst['unit_mer_wind_ecef_x'], inst['unit_mer_wind_ecef_y'], inst[\n 'unit_mer_wind_ecef_z'] = pysatMagVect.cross_product(unit_pos_x,\n unit_pos_y, unit_pos_z, inst['unit_zonal_wind_ecef_x'], inst[\n 'unit_zonal_wind_ecef_y'], inst['unit_zonal_wind_ecef_z'])\n inst.meta['zonal_wind'] = {'units': 'm/s', 'long_name': 'Zonal Wind',\n 'desc': 'HWM model zonal wind'}\n inst.meta['meridional_wind'] = {'units': 'm/s', 'long_name':\n 'Meridional Wind', 'desc': 'HWM model meridional wind'}\n inst.meta['unit_zonal_wind_ecef_x'] = {'units': '', 'long_name':\n 'Zonal Wind Unit ECEF x-vector', 'desc':\n 'x-value of zonal wind unit vector in ECEF co ordinates'}\n inst.meta['unit_zonal_wind_ecef_y'] = {'units': '', 'long_name':\n 'Zonal Wind Unit ECEF y-vector', 'desc':\n 'y-value of zonal wind unit vector in ECEF co ordinates'}\n inst.meta['unit_zonal_wind_ecef_z'] = {'units': '', 'long_name':\n 'Zonal Wind Unit ECEF z-vector', 'desc':\n 'z-value of zonal wind unit vector in ECEF co ordinates'}\n inst.meta['unit_mer_wind_ecef_x'] = {'units': '', 'long_name':\n 'Meridional Wind Unit ECEF x-vector', 'desc':\n 'x-value of meridional wind unit vector in ECEF co ordinates'}\n inst.meta['unit_mer_wind_ecef_y'] = {'units': '', 'long_name':\n 'Meridional Wind Unit ECEF y-vector', 'desc':\n 'y-value of meridional wind unit vector in ECEF co ordinates'}\n inst.meta['unit_mer_wind_ecef_z'] = {'units': '', 'long_name':\n 'Meridional Wind Unit ECEF z-vector', 'desc':\n 'z-value of meridional wind unit vector in ECEF co ordinates'}\n return\n",
"role": "user"
},
{
"content": "add_hwm_winds_and_ecef_vectors",
"role": "assistant"
}
] |
def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):
import pyglow
from pyglow.pyglow import Point
import pysatMagVect
igrf_params = []
for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[
glong_label], inst[alt_label]):
pt = Point(time, lat, lon, alt)
pt.run_igrf()
igrf = {}
igrf['B'] = pt.B
igrf['B_east'] = pt.Bx
igrf['B_north'] = pt.By
igrf['B_up'] = pt.Bz
igrf_params.append(igrf)
igrf = pds.DataFrame(igrf_params)
igrf.index = inst.data.index
inst[igrf.keys()] = igrf
x, y, z = pysatMagVect.enu_to_ecef_vector(inst['B_east'], inst[
'B_north'], inst['B_up'], inst[glat_label], inst[glong_label])
inst['B_ecef_x'] = x
inst['B_ecef_y'] = y
inst['B_ecef_z'] = z
inst.meta['B'] = {'units': 'nT', 'desc':
'Total geomagnetic field from IGRF.'}
inst.meta['B_east'] = {'units': 'nT', 'desc':
'Geomagnetic field from IGRF expressed using the East/North/Up (ENU) basis.'
}
inst.meta['B_north'] = {'units': 'nT', 'desc':
'Geomagnetic field from IGRF expressed using the East/North/Up (ENU) basis.'
}
inst.meta['B_up'] = {'units': 'nT', 'desc':
'Geomagnetic field from IGRF expressed using the East/North/Up (ENU) basis.'
}
inst.meta['B_ecef_x'] = {'units': 'nT', 'desc':
'Geomagnetic field from IGRF expressed using the Earth Centered Earth Fixed (ECEF) basis.'
}
inst.meta['B_ecef_y'] = {'units': 'nT', 'desc':
'Geomagnetic field from IGRF expressed using the Earth Centered Earth Fixed (ECEF) basis.'
}
inst.meta['B_ecef_z'] = {'units': 'nT', 'desc':
'Geomagnetic field from IGRF expressed using the Earth Centered Earth Fixed (ECEF) basis.'
}
return
|
add_igrf
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):\n import pyglow\n from pyglow.pyglow import Point\n import pysatMagVect\n igrf_params = []\n for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[\n glong_label], inst[alt_label]):\n pt = Point(time, lat, lon, alt)\n pt.run_igrf()\n igrf = {}\n igrf['B'] = pt.B\n igrf['B_east'] = pt.Bx\n igrf['B_north'] = pt.By\n igrf['B_up'] = pt.Bz\n igrf_params.append(igrf)\n igrf = pds.DataFrame(igrf_params)\n igrf.index = inst.data.index\n inst[igrf.keys()] = igrf\n x, y, z = pysatMagVect.enu_to_ecef_vector(inst['B_east'], inst[\n 'B_north'], inst['B_up'], inst[glat_label], inst[glong_label])\n inst['B_ecef_x'] = x\n inst['B_ecef_y'] = y\n inst['B_ecef_z'] = z\n inst.meta['B'] = {'units': 'nT', 'desc':\n 'Total geomagnetic field from IGRF.'}\n inst.meta['B_east'] = {'units': 'nT', 'desc':\n 'Geomagnetic field from IGRF expressed using the East/North/Up (ENU) basis.'\n }\n inst.meta['B_north'] = {'units': 'nT', 'desc':\n 'Geomagnetic field from IGRF expressed using the East/North/Up (ENU) basis.'\n }\n inst.meta['B_up'] = {'units': 'nT', 'desc':\n 'Geomagnetic field from IGRF expressed using the East/North/Up (ENU) basis.'\n }\n inst.meta['B_ecef_x'] = {'units': 'nT', 'desc':\n 'Geomagnetic field from IGRF expressed using the Earth Centered Earth Fixed (ECEF) basis.'\n }\n inst.meta['B_ecef_y'] = {'units': 'nT', 'desc':\n 'Geomagnetic field from IGRF expressed using the Earth Centered Earth Fixed (ECEF) basis.'\n }\n inst.meta['B_ecef_z'] = {'units': 'nT', 'desc':\n 'Geomagnetic field from IGRF expressed using the Earth Centered Earth Fixed (ECEF) basis.'\n }\n return\n",
"role": "user"
},
{
"content": "add_igrf",
"role": "assistant"
}
] |
def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):
import pyglow
from pyglow.pyglow import Point
msis_params = []
for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[
glong_label], inst[alt_label]):
pt = Point(time, lat, lon, alt)
pt.run_msis()
msis = {}
total = 0
for key in pt.nn.keys():
total += pt.nn[key]
msis['Nn'] = total
msis['Nn_N'] = pt.nn['N']
msis['Nn_N2'] = pt.nn['N2']
msis['Nn_O'] = pt.nn['O']
msis['Nn_O2'] = pt.nn['O2']
msis['Tn_msis'] = pt.Tn_msis
msis_params.append(msis)
msis = pds.DataFrame(msis_params)
msis.index = inst.data.index
inst[msis.keys()] = msis
inst.meta['Nn'] = {'units': 'cm^-3', 'desc':
'Total neutral number particle density from MSIS.'}
inst.meta['Nn_N'] = {'units': 'cm^-3', 'desc':
'Total nitrogen number particle density from MSIS.'}
inst.meta['Nn_N2'] = {'units': 'cm^-3', 'desc':
'Total N2 number particle density from MSIS.'}
inst.meta['Nn_O'] = {'units': 'cm^-3', 'desc':
'Total oxygen number particle density from MSIS.'}
inst.meta['Nn_O2'] = {'units': 'cm^-3', 'desc':
'Total O2 number particle density from MSIS.'}
inst.meta['Tn_msis'] = {'units': 'K', 'desc':
'Neutral temperature from MSIS.'}
return
|
add_msis
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst, glat_label='glat', glong_label='glong', alt_label='alt'):\n import pyglow\n from pyglow.pyglow import Point\n msis_params = []\n for time, lat, lon, alt in zip(inst.data.index, inst[glat_label], inst[\n glong_label], inst[alt_label]):\n pt = Point(time, lat, lon, alt)\n pt.run_msis()\n msis = {}\n total = 0\n for key in pt.nn.keys():\n total += pt.nn[key]\n msis['Nn'] = total\n msis['Nn_N'] = pt.nn['N']\n msis['Nn_N2'] = pt.nn['N2']\n msis['Nn_O'] = pt.nn['O']\n msis['Nn_O2'] = pt.nn['O2']\n msis['Tn_msis'] = pt.Tn_msis\n msis_params.append(msis)\n msis = pds.DataFrame(msis_params)\n msis.index = inst.data.index\n inst[msis.keys()] = msis\n inst.meta['Nn'] = {'units': 'cm^-3', 'desc':\n 'Total neutral number particle density from MSIS.'}\n inst.meta['Nn_N'] = {'units': 'cm^-3', 'desc':\n 'Total nitrogen number particle density from MSIS.'}\n inst.meta['Nn_N2'] = {'units': 'cm^-3', 'desc':\n 'Total N2 number particle density from MSIS.'}\n inst.meta['Nn_O'] = {'units': 'cm^-3', 'desc':\n 'Total oxygen number particle density from MSIS.'}\n inst.meta['Nn_O2'] = {'units': 'cm^-3', 'desc':\n 'Total O2 number particle density from MSIS.'}\n inst.meta['Tn_msis'] = {'units': 'K', 'desc':\n 'Neutral temperature from MSIS.'}\n return\n",
"role": "user"
},
{
"content": "add_msis",
"role": "assistant"
}
] |
def x(inst, x_label, y_label, z_label, new_x_label, new_y_label,
new_z_label, meta=None):
import pysatMagVect
x, y, z = pysatMagVect.project_ecef_vector_onto_basis(inst[x_label],
inst[y_label], inst[z_label], inst['sc_xhat_ecef_x'], inst[
'sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'], inst['sc_yhat_ecef_x'],
inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'], inst[
'sc_zhat_ecef_x'], inst['sc_zhat_ecef_y'], inst['sc_zhat_ecef_z'])
inst[new_x_label] = x
inst[new_y_label] = y
inst[new_z_label] = z
if meta is not None:
inst.meta[new_x_label] = meta[0]
inst.meta[new_y_label] = meta[1]
inst.meta[new_z_label] = meta[2]
return
|
project_ecef_vector_onto_sc
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst, x_label, y_label, z_label, new_x_label, new_y_label,\n new_z_label, meta=None):\n import pysatMagVect\n x, y, z = pysatMagVect.project_ecef_vector_onto_basis(inst[x_label],\n inst[y_label], inst[z_label], inst['sc_xhat_ecef_x'], inst[\n 'sc_xhat_ecef_y'], inst['sc_xhat_ecef_z'], inst['sc_yhat_ecef_x'],\n inst['sc_yhat_ecef_y'], inst['sc_yhat_ecef_z'], inst[\n 'sc_zhat_ecef_x'], inst['sc_zhat_ecef_y'], inst['sc_zhat_ecef_z'])\n inst[new_x_label] = x\n inst[new_y_label] = y\n inst[new_z_label] = z\n if meta is not None:\n inst.meta[new_x_label] = meta[0]\n inst.meta[new_y_label] = meta[1]\n inst.meta[new_z_label] = meta[2]\n return\n",
"role": "user"
},
{
"content": "project_ecef_vector_onto_sc",
"role": "assistant"
}
] |
def x(inst, labelx, labely, data_label, datalim, xlim=None, ylim=None):
plt.ioff()
figs = []
axs = []
if type(data_label) is str:
figs.append(plt.figure())
ax1 = figs[0].add_subplot(211, projection='3d')
ax2 = figs[0].add_subplot(212)
axs.append((ax1, ax2))
plt.suptitle(data_label)
if xlim is not None:
ax1.set_xlim(xlim)
ax2.set_xlim(xlim)
if ylim is not None:
ax1.set_ylim(ylim)
ax2.set_ylim(ylim)
data_label = [data_label]
plt.hold(True)
else:
for i in np.arange(len(data_label)):
figs.append(plt.figure())
ax1 = figs[i].add_subplot(211, projection='3d')
ax2 = figs[i].add_subplot(212)
axs.append((ax1, ax2))
plt.suptitle(data_label[i])
if xlim is not None:
ax1.set_xlim(xlim)
ax2.set_xlim(xlim)
if ylim is not None:
ax1.set_ylim(ylim)
ax2.set_ylim(ylim)
plt.hold(True)
norm = colors.Normalize(vmin=datalim[0], vmax=datalim[1])
p = [i for i in np.arange(len(figs))]
for i, inst in enumerate(inst):
for j, (fig, ax) in enumerate(zip(figs, axs)):
if len(inst.data) > 0:
if (len(inst.data[labelx]) > 0) & (len(inst.data[labely]) > 0
) & (len(inst.data[data_label[j]]) > 0):
p[j] = ax[0].scatter(inst.data[labelx], inst.data[
labely], inst.data[data_label[j]], zdir='z', c=inst
.data[data_label[j]], cmap=cm.jet, norm=norm,
linewidth='0', edgecolors=None)
ax[1].scatter(inst.data[labelx], inst.data[labely], c=
inst.data[data_label[j]], cmap=cm.jet, norm=norm,
linewidth=1e-11, alpha=0.5, edgecolor=None)
for j, (fig, ax) in enumerate(zip(figs, axs)):
try:
cbar = plt.colorbar(p[j], ax=ax[0], label='Amplitude (m/s)')
except:
print('Tried colorbar but failed, thus no colorbar.')
ax[0].elev = 30.0
plt.ion()
return figs
|
scatterplot
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst, labelx, labely, data_label, datalim, xlim=None, ylim=None):\n plt.ioff()\n figs = []\n axs = []\n if type(data_label) is str:\n figs.append(plt.figure())\n ax1 = figs[0].add_subplot(211, projection='3d')\n ax2 = figs[0].add_subplot(212)\n axs.append((ax1, ax2))\n plt.suptitle(data_label)\n if xlim is not None:\n ax1.set_xlim(xlim)\n ax2.set_xlim(xlim)\n if ylim is not None:\n ax1.set_ylim(ylim)\n ax2.set_ylim(ylim)\n data_label = [data_label]\n plt.hold(True)\n else:\n for i in np.arange(len(data_label)):\n figs.append(plt.figure())\n ax1 = figs[i].add_subplot(211, projection='3d')\n ax2 = figs[i].add_subplot(212)\n axs.append((ax1, ax2))\n plt.suptitle(data_label[i])\n if xlim is not None:\n ax1.set_xlim(xlim)\n ax2.set_xlim(xlim)\n if ylim is not None:\n ax1.set_ylim(ylim)\n ax2.set_ylim(ylim)\n plt.hold(True)\n norm = colors.Normalize(vmin=datalim[0], vmax=datalim[1])\n p = [i for i in np.arange(len(figs))]\n for i, inst in enumerate(inst):\n for j, (fig, ax) in enumerate(zip(figs, axs)):\n if len(inst.data) > 0:\n if (len(inst.data[labelx]) > 0) & (len(inst.data[labely]) > 0\n ) & (len(inst.data[data_label[j]]) > 0):\n p[j] = ax[0].scatter(inst.data[labelx], inst.data[\n labely], inst.data[data_label[j]], zdir='z', c=inst\n .data[data_label[j]], cmap=cm.jet, norm=norm,\n linewidth='0', edgecolors=None)\n ax[1].scatter(inst.data[labelx], inst.data[labely], c=\n inst.data[data_label[j]], cmap=cm.jet, norm=norm,\n linewidth=1e-11, alpha=0.5, edgecolor=None)\n for j, (fig, ax) in enumerate(zip(figs, axs)):\n try:\n cbar = plt.colorbar(p[j], ax=ax[0], label='Amplitude (m/s)')\n except:\n print('Tried colorbar but failed, thus no colorbar.')\n ax[0].elev = 30.0\n plt.ion()\n return figs\n",
"role": "user"
},
{
"content": "scatterplot",
"role": "assistant"
}
] |
def x(level=logging.DEBUG):
logger = logging.getLogger(__name__)
logger.setLevel(level)
if not logger.handlers:
logger.info('enabling logging to console.')
logger.addHandler(logging.StreamHandler(sys.stdout))
|
enable_log
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(level=logging.DEBUG):\n logger = logging.getLogger(__name__)\n logger.setLevel(level)\n if not logger.handlers:\n logger.info('enabling logging to console.')\n logger.addHandler(logging.StreamHandler(sys.stdout))\n",
"role": "user"
},
{
"content": "enable_log",
"role": "assistant"
}
] |
def x(tag=None, sat_id=None, data_path=None, format_str=None):
import sys
estr = 'Building a list of COSMIC files, which can possibly take time. '
estr = '{:s}~1s per 100K files'.format(estr)
print(estr)
sys.stdout.flush()
cosmicFiles = glob.glob(os.path.join(data_path, '*/*_nc'))
num = len(cosmicFiles)
if num != 0:
print('Estimated time:', num * 1e-05, 'seconds')
sys.stdout.flush()
year = [None] * num
days = [None] * num
hours = [None] * num
minutes = [None] * num
microseconds = [None] * num
for i, f in enumerate(cosmicFiles):
f2 = f.split('.')
year[i] = f2[-6]
days[i] = f2[-5]
hours[i] = f2[-4]
minutes[i] = f2[-3]
microseconds[i] = i
year = np.array(year).astype(int)
days = np.array(days).astype(int)
uts = np.array(hours).astype(int) * 3600.0 + np.array(minutes).astype(
int) * 60.0
uts += np.mod(np.array(microseconds).astype(int) * 4, 8000) * 1e-05
index = pysat.utils.create_datetime_index(year=year, day=days, uts=uts)
file_list = pysat.Series(cosmicFiles, index=index)
return file_list
else:
print('Found no files, check your path or download them.')
return pysat.Series(None)
|
list_files
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tag=None, sat_id=None, data_path=None, format_str=None):\n import sys\n estr = 'Building a list of COSMIC files, which can possibly take time. '\n estr = '{:s}~1s per 100K files'.format(estr)\n print(estr)\n sys.stdout.flush()\n cosmicFiles = glob.glob(os.path.join(data_path, '*/*_nc'))\n num = len(cosmicFiles)\n if num != 0:\n print('Estimated time:', num * 1e-05, 'seconds')\n sys.stdout.flush()\n year = [None] * num\n days = [None] * num\n hours = [None] * num\n minutes = [None] * num\n microseconds = [None] * num\n for i, f in enumerate(cosmicFiles):\n f2 = f.split('.')\n year[i] = f2[-6]\n days[i] = f2[-5]\n hours[i] = f2[-4]\n minutes[i] = f2[-3]\n microseconds[i] = i\n year = np.array(year).astype(int)\n days = np.array(days).astype(int)\n uts = np.array(hours).astype(int) * 3600.0 + np.array(minutes).astype(\n int) * 60.0\n uts += np.mod(np.array(microseconds).astype(int) * 4, 8000) * 1e-05\n index = pysat.utils.create_datetime_index(year=year, day=days, uts=uts)\n file_list = pysat.Series(cosmicFiles, index=index)\n return file_list\n else:\n print('Found no files, check your path or download them.')\n return pysat.Series(None)\n",
"role": "user"
},
{
"content": "list_files",
"role": "assistant"
}
] |
def x(files, tag=None, sat_id=None, altitude_bin=None):
output = [None] * len(files)
drop_idx = []
for i, file in enumerate(files):
try:
data = netcdf_file(file, mode='r', mmap=False)
new = {}
ncattrsList = data._attributes.keys()
for d in ncattrsList:
new[d] = data._attributes[d]
loadedVars = {}
keys = data.variables.keys()
for key in keys:
if data.variables[key][:].dtype.byteorder != '=':
loadedVars[key] = data.variables[key][:].byteswap(
).newbyteorder()
else:
loadedVars[key] = data.variables[key][:]
new['profiles'] = pysat.DataFrame(loadedVars)
output[i] = new
data.close()
except RuntimeError:
drop_idx.append(i)
drop_idx.reverse()
for i in drop_idx:
del output[i]
if tag == 'ionprf':
if altitude_bin is not None:
for out in output:
out['profiles'].index = (out['profiles']['MSL_alt'] /
altitude_bin).round().values * altitude_bin
out['profiles'] = out['profiles'].groupby(out['profiles'].
index.values).mean()
else:
for out in output:
out['profiles'].index = out['profiles']['MSL_alt']
return output
|
load_files
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(files, tag=None, sat_id=None, altitude_bin=None):\n output = [None] * len(files)\n drop_idx = []\n for i, file in enumerate(files):\n try:\n data = netcdf_file(file, mode='r', mmap=False)\n new = {}\n ncattrsList = data._attributes.keys()\n for d in ncattrsList:\n new[d] = data._attributes[d]\n loadedVars = {}\n keys = data.variables.keys()\n for key in keys:\n if data.variables[key][:].dtype.byteorder != '=':\n loadedVars[key] = data.variables[key][:].byteswap(\n ).newbyteorder()\n else:\n loadedVars[key] = data.variables[key][:]\n new['profiles'] = pysat.DataFrame(loadedVars)\n output[i] = new\n data.close()\n except RuntimeError:\n drop_idx.append(i)\n drop_idx.reverse()\n for i in drop_idx:\n del output[i]\n if tag == 'ionprf':\n if altitude_bin is not None:\n for out in output:\n out['profiles'].index = (out['profiles']['MSL_alt'] /\n altitude_bin).round().values * altitude_bin\n out['profiles'] = out['profiles'].groupby(out['profiles'].\n index.values).mean()\n else:\n for out in output:\n out['profiles'].index = out['profiles']['MSL_alt']\n return output\n",
"role": "user"
},
{
"content": "load_files",
"role": "assistant"
}
] |
def x(value):
t = 60
if value is not None:
r = value % t
if r > t / 2:
ret = value + (t - r)
else:
ret = value - r
else:
ret = NaN
return ret
|
round_60
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(value):\n t = 60\n if value is not None:\n r = value % t\n if r > t / 2:\n ret = value + (t - r)\n else:\n ret = value - r\n else:\n ret = NaN\n return ret\n",
"role": "user"
},
{
"content": "round_60",
"role": "assistant"
}
] |
def x(prev, curr, counters):
busy_prop, idle_prop = counters
pb = getattr(prev, busy_prop)
pi = getattr(prev, idle_prop)
cb = getattr(curr, busy_prop)
ci = getattr(curr, idle_prop)
db = minus(cb, pb)
di = minus(ci, pi)
return mul(div(db, add(db, di)), 100)
|
utilization
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(prev, curr, counters):\n busy_prop, idle_prop = counters\n pb = getattr(prev, busy_prop)\n pi = getattr(prev, idle_prop)\n cb = getattr(curr, busy_prop)\n ci = getattr(curr, idle_prop)\n db = minus(cb, pb)\n di = minus(ci, pi)\n return mul(div(db, add(db, di)), 100)\n",
"role": "user"
},
{
"content": "utilization",
"role": "assistant"
}
] |
def x(prev, curr, counters):
counter = get_counter(counters)
pv = getattr(prev, counter)
cv = getattr(curr, counter)
return minus(cv, pv)
|
delta_ps
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(prev, curr, counters):\n counter = get_counter(counters)\n pv = getattr(prev, counter)\n cv = getattr(curr, counter)\n return minus(cv, pv)\n",
"role": "user"
},
{
"content": "delta_ps",
"role": "assistant"
}
] |
def x(prev, curr, counters):
bw_stats, io_stats = counters
size_mb = div(getattr(curr, bw_stats), getattr(curr, io_stats))
return mul(size_mb, 1024)
|
io_size_kb
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(prev, curr, counters):\n bw_stats, io_stats = counters\n size_mb = div(getattr(curr, bw_stats), getattr(curr, io_stats))\n return mul(size_mb, 1024)\n",
"role": "user"
},
{
"content": "io_size_kb",
"role": "assistant"
}
] |
def x(tag='', sat_id=None, data_path=None, format_str=None):
if format_str is None and tag is not None:
if tag == '' or tag == 'ascii':
ascii_fmt = 'Density_3deg_{year:02d}_{doy:03d}.ascii'
return pysat.Files.from_os(data_path=data_path, format_str=
ascii_fmt)
else:
raise ValueError('Unrecognized tag name for CHAMP STAR')
elif format_str is None:
estr = 'A tag name must be passed to the loading routine for CHAMP'
raise ValueError(estr)
else:
return pysat.Files.from_os(data_path=data_path, format_str=format_str)
|
list_files
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tag='', sat_id=None, data_path=None, format_str=None):\n if format_str is None and tag is not None:\n if tag == '' or tag == 'ascii':\n ascii_fmt = 'Density_3deg_{year:02d}_{doy:03d}.ascii'\n return pysat.Files.from_os(data_path=data_path, format_str=\n ascii_fmt)\n else:\n raise ValueError('Unrecognized tag name for CHAMP STAR')\n elif format_str is None:\n estr = 'A tag name must be passed to the loading routine for CHAMP'\n raise ValueError(estr)\n else:\n return pysat.Files.from_os(data_path=data_path, format_str=format_str)\n",
"role": "user"
},
{
"content": "list_files",
"role": "assistant"
}
] |
def x(fnames, tag=None, sat_id=None):
import re
if len(fnames) <= 0:
return pysat.DataFrame(None), pysat.Meta(None)
if isinstance(fnames, str):
fnames = [fnames]
champ_labels = {'Two-digit Year (years)': 'year',
'Day of the Year (days)': 'doy', 'Second of the Day (GPS time,sec)':
'sod', 'Center Latitude of 3-degree Bin (deg)': 'bin_lat',
'Satellite Geodetic Latitude (deg)': 'sat_glat',
'Satellite Longitude (deg)': 'sat_lon', 'Satellite Height (km)':
'sat_h', 'Satellite Local Time (hours)': 'sat_lt',
'Satellite Quasi-Dipole Latitude (deg)': 'sat_qdlat',
'Satellite Magnetic Longitude (deg)': 'sat_mlon',
'Satellite Magnetic Local Time (hours)': 'sat_mlt',
'Neutral Density (kg/m^3)': 'ndens',
'Neutral Density Normalized to 400km using NRLMSISe00': 'ndens400',
'Neutral Density Normalized to 410km using NRLMSISe00': 'ndens410',
'NRLMSISe00 Neutral Density at Satellite Height': 'nrlmsis_ndens',
'Uncertainty in Neutral Density (kg/m^3)': 'ndens_err',
'Number of Data Points in Current Averaging Bin': 'npnts',
'Number of Points in Current Averaging Bin that Required Interpolation'
: 'npnts_interp',
'Average Coefficient of Drag Used in Current Averaging Bin':
'avg_drag_coeff'}
champ_dtypes = {'year': np.int32, 'doy': np.int32, 'sod': float,
'bin_lat': float, 'sat_glat': float, 'sat_lon': float, 'sat_h':
float, 'sat_lt': float, 'sat_qdlat': float, 'sat_mlon': float,
'sat_mlt': float, 'ndens': float, 'ndens400': float, 'ndens410':
float, 'nrlmsis_ndens': float, 'ndens_err': float, 'npnts': int,
'npnts_interp': float, 'avg_drag_coeff': float}
champ_units = {'year': '2-digit years', 'doy': 'day of year', 'sod':
'seconds of day', 'bin_lat': 'degrees', 'sat_glat': 'degrees',
'sat_lon': 'degrees', 'sat_h': 'km', 'sat_lt': 'hours', 'sat_qdlat':
'degrees', 'sat_mlon': 'degrees', 'sat_mlt': 'hours', 'ndens':
'km m^{-3}', 'ndens400': 'km m^{-3}', 'ndens410': 'km m^{-3}',
'nrlmsis_ndens': 'km m^{-3}', 'ndens_err': 'km m^{-3}', 'npnts':
'number', 'npnts_interp': 'number', 'avg_drag_coeff': 'unitless'}
def x(y, d, s):
import datetime as dt
t = dt.datetime.strptime('{:02d} {:03d}'.format(int(y), int(d)),
'%y %j')
fsec = float(s)
isec = np.floor(fsec)
microsec = (fsec - isec) * 1000000.0
t += dt.timedelta(seconds=isec, microseconds=microsec)
return t
f = open(fnames[0], 'r')
vdata = re.split(';|\n', f.readline())
hdata = re.split(';|\n', f.readline())
f.close()
try:
hdata.pop(hdata.index(''))
except:
pass
data = pds.read_csv(fnames[0], delim_whitespace=True, skiprows=2,
header=None, names=[champ_labels[h] for h in hdata], keep_date_col=
True, index_col='datetime', parse_dates={'datetime': [0, 1, 2]},
date_parser=parse_champdate)
meta = pysat.Meta()
for h in hdata:
col = champ_labels[h]
data[col].astype(champ_dtypes[col])
meta[col] = {'units': champ_units[col], 'long_name': h}
return data, meta
|
load
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fnames, tag=None, sat_id=None):\n import re\n if len(fnames) <= 0:\n return pysat.DataFrame(None), pysat.Meta(None)\n if isinstance(fnames, str):\n fnames = [fnames]\n champ_labels = {'Two-digit Year (years)': 'year',\n 'Day of the Year (days)': 'doy', 'Second of the Day (GPS time,sec)':\n 'sod', 'Center Latitude of 3-degree Bin (deg)': 'bin_lat',\n 'Satellite Geodetic Latitude (deg)': 'sat_glat',\n 'Satellite Longitude (deg)': 'sat_lon', 'Satellite Height (km)':\n 'sat_h', 'Satellite Local Time (hours)': 'sat_lt',\n 'Satellite Quasi-Dipole Latitude (deg)': 'sat_qdlat',\n 'Satellite Magnetic Longitude (deg)': 'sat_mlon',\n 'Satellite Magnetic Local Time (hours)': 'sat_mlt',\n 'Neutral Density (kg/m^3)': 'ndens',\n 'Neutral Density Normalized to 400km using NRLMSISe00': 'ndens400',\n 'Neutral Density Normalized to 410km using NRLMSISe00': 'ndens410',\n 'NRLMSISe00 Neutral Density at Satellite Height': 'nrlmsis_ndens',\n 'Uncertainty in Neutral Density (kg/m^3)': 'ndens_err',\n 'Number of Data Points in Current Averaging Bin': 'npnts',\n 'Number of Points in Current Averaging Bin that Required Interpolation'\n : 'npnts_interp',\n 'Average Coefficient of Drag Used in Current Averaging Bin':\n 'avg_drag_coeff'}\n champ_dtypes = {'year': np.int32, 'doy': np.int32, 'sod': float,\n 'bin_lat': float, 'sat_glat': float, 'sat_lon': float, 'sat_h':\n float, 'sat_lt': float, 'sat_qdlat': float, 'sat_mlon': float,\n 'sat_mlt': float, 'ndens': float, 'ndens400': float, 'ndens410':\n float, 'nrlmsis_ndens': float, 'ndens_err': float, 'npnts': int,\n 'npnts_interp': float, 'avg_drag_coeff': float}\n champ_units = {'year': '2-digit years', 'doy': 'day of year', 'sod':\n 'seconds of day', 'bin_lat': 'degrees', 'sat_glat': 'degrees',\n 'sat_lon': 'degrees', 'sat_h': 'km', 'sat_lt': 'hours', 'sat_qdlat':\n 'degrees', 'sat_mlon': 'degrees', 'sat_mlt': 'hours', 'ndens':\n 'km m^{-3}', 'ndens400': 'km m^{-3}', 'ndens410': 'km m^{-3}',\n 'nrlmsis_ndens': 'km m^{-3}', 'ndens_err': 'km m^{-3}', 'npnts':\n 'number', 'npnts_interp': 'number', 'avg_drag_coeff': 'unitless'}\n\n def x(y, d, s):\n import datetime as dt\n t = dt.datetime.strptime('{:02d} {:03d}'.format(int(y), int(d)),\n '%y %j')\n fsec = float(s)\n isec = np.floor(fsec)\n microsec = (fsec - isec) * 1000000.0\n t += dt.timedelta(seconds=isec, microseconds=microsec)\n return t\n f = open(fnames[0], 'r')\n vdata = re.split(';|\\n', f.readline())\n hdata = re.split(';|\\n', f.readline())\n f.close()\n try:\n hdata.pop(hdata.index(''))\n except:\n pass\n data = pds.read_csv(fnames[0], delim_whitespace=True, skiprows=2,\n header=None, names=[champ_labels[h] for h in hdata], keep_date_col=\n True, index_col='datetime', parse_dates={'datetime': [0, 1, 2]},\n date_parser=parse_champdate)\n meta = pysat.Meta()\n for h in hdata:\n col = champ_labels[h]\n data[col].astype(champ_dtypes[col])\n meta[col] = {'units': champ_units[col], 'long_name': h}\n return data, meta\n",
"role": "user"
},
{
"content": "load",
"role": "assistant"
}
] |
def x(self, by_name=False, inst_module=None):
import importlib
self._list_rtn = self._pass_func
self._load_rtn = self._pass_func
self._default_rtn = self._pass_func
self._clean_rtn = self._pass_func
self._init_rtn = self._pass_func
self._download_rtn = self._pass_func
self.directory_format = None
self.file_format = None
self.multi_file_day = False
self.orbit_info = None
if by_name:
inst = importlib.import_module(''.join(('.', self.platform, '_',
self.name)), package='pysat.instruments')
elif inst_module is not None:
inst = inst_module
else:
return
try:
self._load_rtn = inst.load
self._list_rtn = inst.list_files
self._download_rtn = inst.download
except AttributeError:
estr = 'A load, file_list, and download routine are required for '
raise AttributeError('{:s}every instrument.'.format(estr))
try:
self._default_rtn = inst.default
except AttributeError:
pass
try:
self._init_rtn = inst.init
except AttributeError:
pass
try:
self._clean_rtn = inst.clean
except AttributeError:
pass
try:
self.directory_format = inst.directory_format
except AttributeError:
pass
try:
self.multi_file_day = inst.multi_file_day
except AttributeError:
pass
try:
self.orbit_info = inst.orbit_info
except AttributeError:
pass
return
|
assign_funcs
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, by_name=False, inst_module=None):\n import importlib\n self._list_rtn = self._pass_func\n self._load_rtn = self._pass_func\n self._default_rtn = self._pass_func\n self._clean_rtn = self._pass_func\n self._init_rtn = self._pass_func\n self._download_rtn = self._pass_func\n self.directory_format = None\n self.file_format = None\n self.multi_file_day = False\n self.orbit_info = None\n if by_name:\n inst = importlib.import_module(''.join(('.', self.platform, '_',\n self.name)), package='pysat.instruments')\n elif inst_module is not None:\n inst = inst_module\n else:\n return\n try:\n self._load_rtn = inst.load\n self._list_rtn = inst.list_files\n self._download_rtn = inst.download\n except AttributeError:\n estr = 'A load, file_list, and download routine are required for '\n raise AttributeError('{:s}every instrument.'.format(estr))\n try:\n self._default_rtn = inst.default\n except AttributeError:\n pass\n try:\n self._init_rtn = inst.init\n except AttributeError:\n pass\n try:\n self._clean_rtn = inst.clean\n except AttributeError:\n pass\n try:\n self.directory_format = inst.directory_format\n except AttributeError:\n pass\n try:\n self.multi_file_day = inst.multi_file_day\n except AttributeError:\n pass\n try:\n self.orbit_info = inst.orbit_info\n except AttributeError:\n pass\n return\n",
"role": "user"
},
{
"content": "assign_funcs",
"role": "assistant"
}
] |
def x(self, date=None, fid=None):
if fid is not None:
fname = self.files[fid:fid + 1]
elif date is not None:
fname = self.files[date:date + pds.DateOffset(days=1)]
else:
raise ValueError('Must supply either a date or file id number.')
if len(fname) > 0:
load_fname = [os.path.join(self.files.data_path, f) for f in fname]
data, mdata = self._load_rtn(load_fname, tag=self.tag, sat_id=self.
sat_id, **self.kwargs)
mdata.accept_default_labels(self)
else:
data = DataFrame(None)
mdata = _meta.Meta(units_label=self.units_label, name_label=self.
name_label, notes_label=self.notes_label, desc_label=self.
desc_label, plot_label=self.plot_label, axis_label=self.
axis_label, scale_label=self.scale_label, min_label=self.
min_label, max_label=self.max_label, fill_label=self.fill_label)
output_str = '{platform} {name} {tag} {sat_id}'
output_str = output_str.format(platform=self.platform, name=self.name,
tag=self.tag, sat_id=self.sat_id)
if not data.empty:
if not isinstance(data, DataFrame):
raise TypeError(' '.join(('Data returned by instrument load',
'routine must be a pandas.DataFrame')))
if not isinstance(mdata, _meta.Meta):
raise TypeError('Metadata returned must be a pysat.Meta object')
if date is not None:
output_str = ' '.join(('Returning', output_str, 'data for',
date.strftime('%x')))
elif len(fname) == 1:
output_str = ' '.join(('Returning', output_str, 'data from',
fname[0]))
else:
output_str = ' '.join(('Returning', output_str, 'data from',
fname[0], '::', fname[-1]))
else:
output_str = ' '.join(('No', output_str, 'data for', date.strftime(
'%m/%d/%y')))
output_str = ' '.join(output_str.split())
print(output_str)
return data, mdata
|
load_data
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, date=None, fid=None):\n if fid is not None:\n fname = self.files[fid:fid + 1]\n elif date is not None:\n fname = self.files[date:date + pds.DateOffset(days=1)]\n else:\n raise ValueError('Must supply either a date or file id number.')\n if len(fname) > 0:\n load_fname = [os.path.join(self.files.data_path, f) for f in fname]\n data, mdata = self._load_rtn(load_fname, tag=self.tag, sat_id=self.\n sat_id, **self.kwargs)\n mdata.accept_default_labels(self)\n else:\n data = DataFrame(None)\n mdata = _meta.Meta(units_label=self.units_label, name_label=self.\n name_label, notes_label=self.notes_label, desc_label=self.\n desc_label, plot_label=self.plot_label, axis_label=self.\n axis_label, scale_label=self.scale_label, min_label=self.\n min_label, max_label=self.max_label, fill_label=self.fill_label)\n output_str = '{platform} {name} {tag} {sat_id}'\n output_str = output_str.format(platform=self.platform, name=self.name,\n tag=self.tag, sat_id=self.sat_id)\n if not data.empty:\n if not isinstance(data, DataFrame):\n raise TypeError(' '.join(('Data returned by instrument load',\n 'routine must be a pandas.DataFrame')))\n if not isinstance(mdata, _meta.Meta):\n raise TypeError('Metadata returned must be a pysat.Meta object')\n if date is not None:\n output_str = ' '.join(('Returning', output_str, 'data for',\n date.strftime('%x')))\n elif len(fname) == 1:\n output_str = ' '.join(('Returning', output_str, 'data from',\n fname[0]))\n else:\n output_str = ' '.join(('Returning', output_str, 'data from',\n fname[0], '::', fname[-1]))\n else:\n output_str = ' '.join(('No', output_str, 'data for', date.strftime(\n '%m/%d/%y')))\n output_str = ' '.join(output_str.split())\n print(output_str)\n return data, mdata\n",
"role": "user"
},
{
"content": "load_data",
"role": "assistant"
}
] |
def x(self):
if self._load_by_date:
next_date = self.date + pds.DateOffset(days=1)
return self._load_data(date=next_date)
else:
return self._load_data(fid=self._fid + 1)
|
load_next
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._load_by_date:\n next_date = self.date + pds.DateOffset(days=1)\n return self._load_data(date=next_date)\n else:\n return self._load_data(fid=self._fid + 1)\n",
"role": "user"
},
{
"content": "load_next",
"role": "assistant"
}
] |
def x(self):
if self._load_by_date:
prev_date = self.date - pds.DateOffset(days=1)
return self._load_data(date=prev_date)
else:
return self._load_data(fid=self._fid - 1)
|
load_prev
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._load_by_date:\n prev_date = self.date - pds.DateOffset(days=1)\n return self._load_data(date=prev_date)\n else:\n return self._load_data(fid=self._fid - 1)\n",
"role": "user"
},
{
"content": "load_prev",
"role": "assistant"
}
] |
def x(self, start, stop, freq='D', user=None, password=None, **kwargs):
import errno
try:
os.makedirs(self.files.data_path)
except OSError as e:
if e.errno != errno.EEXIST:
raise
print('Downloading data to: ', self.files.data_path)
date_array = utils.season_date_range(start, stop, freq=freq)
if user is None:
self._download_rtn(date_array, tag=self.tag, sat_id=self.sat_id,
data_path=self.files.data_path, **kwargs)
else:
self._download_rtn(date_array, tag=self.tag, sat_id=self.sat_id,
data_path=self.files.data_path, user=user, password=password,
**kwargs)
first_date = self.files.start_date
last_date = self.files.stop_date
print('Updating pysat file list')
self.files.refresh()
if len(self.bounds[0]) == 1:
if self.bounds[0][0] == first_date and self.bounds[1][0] == last_date:
print('Updating instrument object bounds.')
self.bounds = None
|
download
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, start, stop, freq='D', user=None, password=None, **kwargs):\n import errno\n try:\n os.makedirs(self.files.data_path)\n except OSError as e:\n if e.errno != errno.EEXIST:\n raise\n print('Downloading data to: ', self.files.data_path)\n date_array = utils.season_date_range(start, stop, freq=freq)\n if user is None:\n self._download_rtn(date_array, tag=self.tag, sat_id=self.sat_id,\n data_path=self.files.data_path, **kwargs)\n else:\n self._download_rtn(date_array, tag=self.tag, sat_id=self.sat_id,\n data_path=self.files.data_path, user=user, password=password,\n **kwargs)\n first_date = self.files.start_date\n last_date = self.files.stop_date\n print('Updating pysat file list')\n self.files.refresh()\n if len(self.bounds[0]) == 1:\n if self.bounds[0][0] == first_date and self.bounds[1][0] == last_date:\n print('Updating instrument object bounds.')\n self.bounds = None\n",
"role": "user"
},
{
"content": "download",
"role": "assistant"
}
] |
def x(self, verifyPad=False):
if self._iter_type == 'date':
if self.date is not None:
idx, = np.where(self._iter_list == self.date)
if len(idx) == 0:
raise StopIteration('File list is empty. Nothing to be done.')
elif idx[-1] + 1 >= len(self._iter_list):
raise StopIteration('Outside the set date boundaries.')
else:
idx += 1
self.load(date=self._iter_list[idx[0]], verifyPad=verifyPad)
else:
self.load(date=self._iter_list[0], verifyPad=verifyPad)
elif self._iter_type == 'file':
if self._fid is not None:
first = self.files.get_index(self._iter_list[0])
last = self.files.get_index(self._iter_list[-1])
if (self._fid < first) | (self._fid + 1 > last):
raise StopIteration('Outside the set file boundaries.')
else:
self.load(fname=self._iter_list[self._fid + 1 - first],
verifyPad=verifyPad)
else:
self.load(fname=self._iter_list[0], verifyPad=verifyPad)
|
next
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, verifyPad=False):\n if self._iter_type == 'date':\n if self.date is not None:\n idx, = np.where(self._iter_list == self.date)\n if len(idx) == 0:\n raise StopIteration('File list is empty. Nothing to be done.')\n elif idx[-1] + 1 >= len(self._iter_list):\n raise StopIteration('Outside the set date boundaries.')\n else:\n idx += 1\n self.load(date=self._iter_list[idx[0]], verifyPad=verifyPad)\n else:\n self.load(date=self._iter_list[0], verifyPad=verifyPad)\n elif self._iter_type == 'file':\n if self._fid is not None:\n first = self.files.get_index(self._iter_list[0])\n last = self.files.get_index(self._iter_list[-1])\n if (self._fid < first) | (self._fid + 1 > last):\n raise StopIteration('Outside the set file boundaries.')\n else:\n self.load(fname=self._iter_list[self._fid + 1 - first],\n verifyPad=verifyPad)\n else:\n self.load(fname=self._iter_list[0], verifyPad=verifyPad)\n",
"role": "user"
},
{
"content": "next",
"role": "assistant"
}
] |
def x(self, coltype):
if type(coltype) is np.dtype:
var_type = coltype.kind + str(coltype.itemsize)
return var_type
elif coltype is np.int64:
return 'i8'
elif coltype is np.int32:
return 'i4'
elif coltype is np.int16:
return 'i2'
elif coltype is np.int8:
return 'i1'
elif coltype is np.uint64:
return 'u8'
elif coltype is np.uint32:
return 'u4'
elif coltype is np.uint16:
return 'u2'
elif coltype is np.uint8:
return 'u1'
elif coltype is np.float64:
return 'f8'
elif coltype is np.float32:
return 'f4'
elif issubclass(coltype, basestring):
return 'S1'
else:
raise TypeError('Unknown Variable Type' + str(coltype))
|
get_var_type_code
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, coltype):\n if type(coltype) is np.dtype:\n var_type = coltype.kind + str(coltype.itemsize)\n return var_type\n elif coltype is np.int64:\n return 'i8'\n elif coltype is np.int32:\n return 'i4'\n elif coltype is np.int16:\n return 'i2'\n elif coltype is np.int8:\n return 'i1'\n elif coltype is np.uint64:\n return 'u8'\n elif coltype is np.uint32:\n return 'u4'\n elif coltype is np.uint16:\n return 'u2'\n elif coltype is np.uint8:\n return 'u1'\n elif coltype is np.float64:\n return 'f8'\n elif coltype is np.float32:\n return 'f4'\n elif issubclass(coltype, basestring):\n return 'S1'\n else:\n raise TypeError('Unknown Variable Type' + str(coltype))\n",
"role": "user"
},
{
"content": "get_var_type_code",
"role": "assistant"
}
] |
def x(self, data, file_format):
data_type = data.dtype
if file_format != 'NETCDF4':
old_format = True
else:
old_format = False
if data_type != np.dtype('O'):
if (data_type == np.int64) & old_format:
data = data.astype(np.int32)
data_type = np.int32
if data_type == np.dtype('<M8[ns]'):
if not old_format:
data_type = np.int64
else:
data_type = np.float
datetime_flag = True
else:
datetime_flag = False
else:
data_type = type(data.iloc[0])
for i in np.arange(len(data)):
if len(data.iloc[i]) > 0:
data_type = type(data.iloc[i])
if not isinstance(data_type, np.float):
break
datetime_flag = False
return data, data_type, datetime_flag
|
get_data_info
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data, file_format):\n data_type = data.dtype\n if file_format != 'NETCDF4':\n old_format = True\n else:\n old_format = False\n if data_type != np.dtype('O'):\n if (data_type == np.int64) & old_format:\n data = data.astype(np.int32)\n data_type = np.int32\n if data_type == np.dtype('<M8[ns]'):\n if not old_format:\n data_type = np.int64\n else:\n data_type = np.float\n datetime_flag = True\n else:\n datetime_flag = False\n else:\n data_type = type(data.iloc[0])\n for i in np.arange(len(data)):\n if len(data.iloc[i]) > 0:\n data_type = type(data.iloc[i])\n if not isinstance(data_type, np.float):\n break\n datetime_flag = False\n return data, data_type, datetime_flag\n",
"role": "user"
},
{
"content": "get_data_info",
"role": "assistant"
}
] |
def x(self, mdata_dict, coltype, remove=False):
for key in mdata_dict:
if type(mdata_dict[key]) == bool:
mdata_dict[key] = int(mdata_dict[key])
if coltype == type(' ') or coltype == type(u' '):
remove = True
if u'_FillValue' in mdata_dict.keys():
if remove:
mdata_dict.pop('_FillValue')
else:
mdata_dict['_FillValue'] = np.array(mdata_dict['_FillValue']
).astype(coltype)
if u'FillVal' in mdata_dict.keys():
if remove:
mdata_dict.pop('FillVal')
else:
mdata_dict['FillVal'] = np.array(mdata_dict['FillVal']).astype(
coltype)
return mdata_dict
|
filter_netcdf4_metadata
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, mdata_dict, coltype, remove=False):\n for key in mdata_dict:\n if type(mdata_dict[key]) == bool:\n mdata_dict[key] = int(mdata_dict[key])\n if coltype == type(' ') or coltype == type(u' '):\n remove = True\n if u'_FillValue' in mdata_dict.keys():\n if remove:\n mdata_dict.pop('_FillValue')\n else:\n mdata_dict['_FillValue'] = np.array(mdata_dict['_FillValue']\n ).astype(coltype)\n if u'FillVal' in mdata_dict.keys():\n if remove:\n mdata_dict.pop('FillVal')\n else:\n mdata_dict['FillVal'] = np.array(mdata_dict['FillVal']).astype(\n coltype)\n return mdata_dict\n",
"role": "user"
},
{
"content": "filter_netcdf4_metadata",
"role": "assistant"
}
] |
def x(self, meta_to_translate):
export_dict = {}
if self._meta_translation_table is not None:
translation_table = {}
for key in self._meta_translation_table:
translation_table[getattr(self, key)
] = self._meta_translation_table[key]
else:
translation_table = None
for key in meta_to_translate.data.index:
if translation_table is None:
export_dict[key] = meta_to_translate.data.loc[key].to_dict()
else:
export_dict[key] = {}
meta_dict = meta_to_translate.data.loc[key].to_dict()
for original_key in meta_dict:
if original_key in translation_table:
for translated_key in translation_table[original_key]:
export_dict[key][translated_key] = meta_dict[
original_key]
else:
export_dict[key][original_key] = meta_dict[original_key]
for key in meta_to_translate.ho_data:
if key not in export_dict:
export_dict[key] = {}
for ho_key in meta_to_translate.ho_data[key].data.index:
if translation_table is None:
export_dict[key + '_' + ho_key] = meta_to_translate.ho_data[key
].data.loc[ho_key].to_dict()
else:
export_dict[key + '_' + ho_key] = {}
meta_dict = meta_to_translate.ho_data[key].data.loc[ho_key
].to_dict()
for original_key in meta_dict:
if original_key in translation_table:
for translated_key in translation_table[original_key]:
export_dict[key + '_' + ho_key][translated_key
] = meta_dict[original_key]
else:
export_dict[key + '_' + ho_key][original_key
] = meta_dict[original_key]
return export_dict
|
generic_meta_translator
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, meta_to_translate):\n export_dict = {}\n if self._meta_translation_table is not None:\n translation_table = {}\n for key in self._meta_translation_table:\n translation_table[getattr(self, key)\n ] = self._meta_translation_table[key]\n else:\n translation_table = None\n for key in meta_to_translate.data.index:\n if translation_table is None:\n export_dict[key] = meta_to_translate.data.loc[key].to_dict()\n else:\n export_dict[key] = {}\n meta_dict = meta_to_translate.data.loc[key].to_dict()\n for original_key in meta_dict:\n if original_key in translation_table:\n for translated_key in translation_table[original_key]:\n export_dict[key][translated_key] = meta_dict[\n original_key]\n else:\n export_dict[key][original_key] = meta_dict[original_key]\n for key in meta_to_translate.ho_data:\n if key not in export_dict:\n export_dict[key] = {}\n for ho_key in meta_to_translate.ho_data[key].data.index:\n if translation_table is None:\n export_dict[key + '_' + ho_key] = meta_to_translate.ho_data[key\n ].data.loc[ho_key].to_dict()\n else:\n export_dict[key + '_' + ho_key] = {}\n meta_dict = meta_to_translate.ho_data[key].data.loc[ho_key\n ].to_dict()\n for original_key in meta_dict:\n if original_key in translation_table:\n for translated_key in translation_table[original_key]:\n export_dict[key + '_' + ho_key][translated_key\n ] = meta_dict[original_key]\n else:\n export_dict[key + '_' + ho_key][original_key\n ] = meta_dict[original_key]\n return export_dict\n",
"role": "user"
},
{
"content": "generic_meta_translator",
"role": "assistant"
}
] |
def x(input_str):
ret = {}
if input_str is not None:
pattern = re.compile('(\\d+)\\s*(\\d+)')
for line in input_str.split('\n'):
line = line.strip()
if len(line) == 0:
continue
matched = re.search(pattern, line)
if matched is None or len(matched.groups()) < 2:
continue
else:
hlu = matched.group(1)
alu = matched.group(2)
ret[int(alu)] = int(hlu)
return ret
|
to_alu_hlu_map
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(input_str):\n ret = {}\n if input_str is not None:\n pattern = re.compile('(\\\\d+)\\\\s*(\\\\d+)')\n for line in input_str.split('\\n'):\n line = line.strip()\n if len(line) == 0:\n continue\n matched = re.search(pattern, line)\n if matched is None or len(matched.groups()) < 2:\n continue\n else:\n hlu = matched.group(1)\n alu = matched.group(2)\n ret[int(alu)] = int(hlu)\n return ret\n",
"role": "user"
},
{
"content": "to_alu_hlu_map",
"role": "assistant"
}
] |
def x(value):
ret = []
p = re.compile('Bus\\s+(\\w+)\\s+Enclosure\\s+(\\w+)\\s+Disk\\s+(\\w+)')
if value is not None:
for line in value.split('\n'):
line = line.strip()
if len(line) == 0:
continue
matched = re.search(p, line)
if matched is None or len(matched.groups()) < 3:
continue
else:
ret.append('{}_{}_{}'.format(*matched.groups()))
return ret
|
to_disk_indices
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(value):\n ret = []\n p = re.compile('Bus\\\\s+(\\\\w+)\\\\s+Enclosure\\\\s+(\\\\w+)\\\\s+Disk\\\\s+(\\\\w+)')\n if value is not None:\n for line in value.split('\\n'):\n line = line.strip()\n if len(line) == 0:\n continue\n matched = re.search(p, line)\n if matched is None or len(matched.groups()) < 3:\n continue\n else:\n ret.append('{}_{}_{}'.format(*matched.groups()))\n return ret\n",
"role": "user"
},
{
"content": "to_disk_indices",
"role": "assistant"
}
] |
def x(url):
regex_url = ('([a-z][a-z0-9+\\-.]*://)?' +
"([a-z0-9\\-._~%!$&'()*+,;=]+@)?" + '([a-z0-9\\-._~%]+' +
"|\\[[a-z0-9\\-._~%!$&'()*+,;=:]+\\])?" + '(:(?P<port>[0-9]+))?')
m = re.match(regex_url, url, re.IGNORECASE)
if m and m.group(3):
return url[m.start(3):m.end(3)]
else:
raise ValueError('URL without a valid host or ip')
|
url_to_host
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(url):\n regex_url = ('([a-z][a-z0-9+\\\\-.]*://)?' +\n \"([a-z0-9\\\\-._~%!$&'()*+,;=]+@)?\" + '([a-z0-9\\\\-._~%]+' +\n \"|\\\\[[a-z0-9\\\\-._~%!$&'()*+,;=:]+\\\\])?\" + '(:(?P<port>[0-9]+))?')\n m = re.match(regex_url, url, re.IGNORECASE)\n if m and m.group(3):\n return url[m.start(3):m.end(3)]\n else:\n raise ValueError('URL without a valid host or ip')\n",
"role": "user"
},
{
"content": "url_to_host",
"role": "assistant"
}
] |
def x(addr):
if addr.startswith('[') and addr.endswith(']'):
addr = addr[1:-1]
parts = addr.split('/')
if len(parts) == 1:
return parts[0], None, None
if len(parts) > 2:
raise ValueError('Illegal host address')
else:
domain_or_ip, prefix = parts
prefix = int(prefix)
if re.match('^(?:[0-9]{1,3}\\.){3}[0-9]{1,3}$', domain_or_ip):
return domain_or_ip, prefix, ipv4_prefix_to_mask(prefix)
elif ':' in domain_or_ip:
return domain_or_ip, prefix, ipv6_prefix_to_mask(prefix)
else:
return domain_or_ip, None, None
|
parse_host_address
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(addr):\n if addr.startswith('[') and addr.endswith(']'):\n addr = addr[1:-1]\n parts = addr.split('/')\n if len(parts) == 1:\n return parts[0], None, None\n if len(parts) > 2:\n raise ValueError('Illegal host address')\n else:\n domain_or_ip, prefix = parts\n prefix = int(prefix)\n if re.match('^(?:[0-9]{1,3}\\\\.){3}[0-9]{1,3}$', domain_or_ip):\n return domain_or_ip, prefix, ipv4_prefix_to_mask(prefix)\n elif ':' in domain_or_ip:\n return domain_or_ip, prefix, ipv6_prefix_to_mask(prefix)\n else:\n return domain_or_ip, None, None\n",
"role": "user"
},
{
"content": "parse_host_address",
"role": "assistant"
}
] |
def x(prefix):
if prefix > 32 or prefix < 0:
raise ValueError('invalid cidr prefix for ipv4')
else:
mask = (1 << 32) - 1 ^ (1 << 32 - prefix) - 1
eight_ones = 255
mask_str = ''
for i in range(0, 4):
mask_str = str(mask & eight_ones) + mask_str
mask = mask >> 8
if i != 3:
mask_str = '.' + mask_str
return mask_str
|
ipv4_prefix_to_mask
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(prefix):\n if prefix > 32 or prefix < 0:\n raise ValueError('invalid cidr prefix for ipv4')\n else:\n mask = (1 << 32) - 1 ^ (1 << 32 - prefix) - 1\n eight_ones = 255\n mask_str = ''\n for i in range(0, 4):\n mask_str = str(mask & eight_ones) + mask_str\n mask = mask >> 8\n if i != 3:\n mask_str = '.' + mask_str\n return mask_str\n",
"role": "user"
},
{
"content": "ipv4_prefix_to_mask",
"role": "assistant"
}
] |
def x(prefix):
if prefix > 128 or prefix < 0:
raise ValueError('invalid cidr prefix for ipv6')
else:
mask = (1 << 128) - 1 ^ (1 << 128 - prefix) - 1
f = 15
hex_mask_str = ''
for i in range(0, 32):
hex_mask_str = format(mask & f, 'x') + hex_mask_str
mask = mask >> 4
if i != 31 and i & 3 == 3:
hex_mask_str = ':' + hex_mask_str
return hex_mask_str
|
ipv6_prefix_to_mask
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(prefix):\n if prefix > 128 or prefix < 0:\n raise ValueError('invalid cidr prefix for ipv6')\n else:\n mask = (1 << 128) - 1 ^ (1 << 128 - prefix) - 1\n f = 15\n hex_mask_str = ''\n for i in range(0, 32):\n hex_mask_str = format(mask & f, 'x') + hex_mask_str\n mask = mask >> 4\n if i != 31 and i & 3 == 3:\n hex_mask_str = ':' + hex_mask_str\n return hex_mask_str\n",
"role": "user"
},
{
"content": "ipv6_prefix_to_mask",
"role": "assistant"
}
] |
def x(self, new_size):
ret = self.size_total
resp = self.modify(size=new_size)
resp.raise_if_err()
return ret
|
expand
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, new_size):\n ret = self.size_total\n resp = self.modify(size=new_size)\n resp.raise_if_err()\n return ret\n",
"role": "user"
},
{
"content": "expand",
"role": "assistant"
}
] |
def x(self, host_names):
if self.host_access:
curr_hosts = [access.host.name for access in self.host_access]
else:
curr_hosts = []
if set(curr_hosts) == set(host_names):
log.info(
'Hosts for updating is equal to current hosts, skip modification.')
return None
new_hosts = [UnityHostList.get(cli=self._cli, name=host_name)[0] for
host_name in host_names]
new_access = [{'host': item, 'accessMask': HostLUNAccessEnum.PRODUCTION
} for item in new_hosts]
resp = self.modify(host_access=new_access)
resp.raise_if_err()
return resp
|
update_hosts
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, host_names):\n if self.host_access:\n curr_hosts = [access.host.name for access in self.host_access]\n else:\n curr_hosts = []\n if set(curr_hosts) == set(host_names):\n log.info(\n 'Hosts for updating is equal to current hosts, skip modification.')\n return None\n new_hosts = [UnityHostList.get(cli=self._cli, name=host_name)[0] for\n host_name in host_names]\n new_access = [{'host': item, 'accessMask': HostLUNAccessEnum.PRODUCTION\n } for item in new_hosts]\n resp = self.modify(host_access=new_access)\n resp.raise_if_err()\n return resp\n",
"role": "user"
},
{
"content": "update_hosts",
"role": "assistant"
}
] |
def x(self, dst_lun_id, max_time_out_of_sync, replication_name=None,
replicate_existing_snaps=None, remote_system=None):
return UnityReplicationSession.create(self._cli, self.get_id(),
dst_lun_id, max_time_out_of_sync, name=replication_name,
replicate_existing_snaps=replicate_existing_snaps, remote_system=
remote_system)
|
replicate
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dst_lun_id, max_time_out_of_sync, replication_name=None,\n replicate_existing_snaps=None, remote_system=None):\n return UnityReplicationSession.create(self._cli, self.get_id(),\n dst_lun_id, max_time_out_of_sync, name=replication_name,\n replicate_existing_snaps=replicate_existing_snaps, remote_system=\n remote_system)\n",
"role": "user"
},
{
"content": "replicate",
"role": "assistant"
}
] |
def x(self, max_time_out_of_sync, dst_pool_id, dst_lun_name=None,
remote_system=None, replication_name=None, dst_size=None, dst_sp=None,
is_dst_thin=None, dst_tiering_policy=None, is_dst_compression=None):
dst_size = self.size_total if dst_size is None else dst_size
dst_resource = UnityResourceConfig.to_embedded(name=dst_lun_name,
pool_id=dst_pool_id, size=dst_size, default_sp=dst_sp,
tiering_policy=dst_tiering_policy, is_thin_enabled=is_dst_thin,
is_compression_enabled=is_dst_compression)
return UnityReplicationSession.create_with_dst_resource_provisioning(self
._cli, self.get_id(), dst_resource, max_time_out_of_sync,
remote_system=remote_system, name=replication_name)
|
replicate_with_dst_resource_provisioning
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, max_time_out_of_sync, dst_pool_id, dst_lun_name=None,\n remote_system=None, replication_name=None, dst_size=None, dst_sp=None,\n is_dst_thin=None, dst_tiering_policy=None, is_dst_compression=None):\n dst_size = self.size_total if dst_size is None else dst_size\n dst_resource = UnityResourceConfig.to_embedded(name=dst_lun_name,\n pool_id=dst_pool_id, size=dst_size, default_sp=dst_sp,\n tiering_policy=dst_tiering_policy, is_thin_enabled=is_dst_thin,\n is_compression_enabled=is_dst_compression)\n return UnityReplicationSession.create_with_dst_resource_provisioning(self\n ._cli, self.get_id(), dst_resource, max_time_out_of_sync,\n remote_system=remote_system, name=replication_name)\n",
"role": "user"
},
{
"content": "replicate_with_dst_resource_provisioning",
"role": "assistant"
}
] |
def x(self):
obj = None
if self.is_link_aggregation():
obj = UnityLinkAggregation.get(self._cli, self.get_id())
else:
obj = UnityEthernetPort.get(self._cli, self.get_id())
return obj
|
get_physical_port
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n obj = None\n if self.is_link_aggregation():\n obj = UnityLinkAggregation.get(self._cli, self.get_id())\n else:\n obj = UnityEthernetPort.get(self._cli, self.get_id())\n return obj\n",
"role": "user"
},
{
"content": "get_physical_port",
"role": "assistant"
}
] |
def x(pool_id=None, is_thin_enabled=None, is_deduplication_enabled=None,
is_compression_enabled=None, is_backup_only=None, size=None,
tiering_policy=None, request_id=None, src_id=None, name=None,
default_sp=None, replication_resource_type=None):
return {'poolId': pool_id, 'isThinEnabled': is_thin_enabled,
'isDeduplicationEnabled': is_deduplication_enabled,
'isCompressionEnabled': is_compression_enabled, 'isBackupOnly':
is_backup_only, 'size': size, 'tieringPolicy': tiering_policy,
'requestId': request_id, 'srcId': src_id, 'name': name, 'defaultSP':
default_sp, 'replicationResourceType': replication_resource_type}
|
to_embedded
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(pool_id=None, is_thin_enabled=None, is_deduplication_enabled=None,\n is_compression_enabled=None, is_backup_only=None, size=None,\n tiering_policy=None, request_id=None, src_id=None, name=None,\n default_sp=None, replication_resource_type=None):\n return {'poolId': pool_id, 'isThinEnabled': is_thin_enabled,\n 'isDeduplicationEnabled': is_deduplication_enabled,\n 'isCompressionEnabled': is_compression_enabled, 'isBackupOnly':\n is_backup_only, 'size': size, 'tieringPolicy': tiering_policy,\n 'requestId': request_id, 'srcId': src_id, 'name': name, 'defaultSP':\n default_sp, 'replicationResourceType': replication_resource_type}\n",
"role": "user"
},
{
"content": "to_embedded",
"role": "assistant"
}
] |
def x(cls, cli, src_resource_id, dst_resource_id, max_time_out_of_sync,
name=None, members=None, auto_initiate=None,
hourly_snap_replication_policy=None, daily_snap_replication_policy=None,
replicate_existing_snaps=None, remote_system=None, src_spa_interface=
None, src_spb_interface=None, dst_spa_interface=None, dst_spb_interface
=None):
req_body = cli.make_body(srcResourceId=src_resource_id, dstResourceId=
dst_resource_id, maxTimeOutOfSync=max_time_out_of_sync, members=
members, autoInitiate=auto_initiate, name=name,
hourlySnapReplicationPolicy=hourly_snap_replication_policy,
dailySnapReplicationPolicy=daily_snap_replication_policy,
replicateExistingSnaps=replicate_existing_snaps, remoteSystem=
remote_system, srcSPAInterface=src_spa_interface, srcSPBInterface=
src_spb_interface, dstSPAInterface=dst_spa_interface,
dstSPBInterface=dst_spb_interface)
resp = cli.post(cls().resource_class, **req_body)
resp.raise_if_err()
return cls.get(cli, resp.resource_id)
|
create
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, cli, src_resource_id, dst_resource_id, max_time_out_of_sync,\n name=None, members=None, auto_initiate=None,\n hourly_snap_replication_policy=None, daily_snap_replication_policy=None,\n replicate_existing_snaps=None, remote_system=None, src_spa_interface=\n None, src_spb_interface=None, dst_spa_interface=None, dst_spb_interface\n =None):\n req_body = cli.make_body(srcResourceId=src_resource_id, dstResourceId=\n dst_resource_id, maxTimeOutOfSync=max_time_out_of_sync, members=\n members, autoInitiate=auto_initiate, name=name,\n hourlySnapReplicationPolicy=hourly_snap_replication_policy,\n dailySnapReplicationPolicy=daily_snap_replication_policy,\n replicateExistingSnaps=replicate_existing_snaps, remoteSystem=\n remote_system, srcSPAInterface=src_spa_interface, srcSPBInterface=\n src_spb_interface, dstSPAInterface=dst_spa_interface,\n dstSPBInterface=dst_spb_interface)\n resp = cli.post(cls().resource_class, **req_body)\n resp.raise_if_err()\n return cls.get(cli, resp.resource_id)\n",
"role": "user"
},
{
"content": "create",
"role": "assistant"
}
] |
def x(cls, cli, src_resource_id, dst_resource_config, max_time_out_of_sync,
name=None, remote_system=None, src_spa_interface=None,
src_spb_interface=None, dst_spa_interface=None, dst_spb_interface=None,
dst_resource_element_configs=None, auto_initiate=None,
hourly_snap_replication_policy=None, daily_snap_replication_policy=None,
replicate_existing_snaps=None):
req_body = cli.make_body(srcResourceId=src_resource_id,
dstResourceConfig=dst_resource_config, maxTimeOutOfSync=
max_time_out_of_sync, name=name, remoteSystem=remote_system,
srcSPAInterface=src_spa_interface, srcSPBInterface=
src_spb_interface, dstSPAInterface=dst_spa_interface,
dstSPBInterface=dst_spb_interface, dstResourceElementConfigs=
dst_resource_element_configs, autoInitiate=auto_initiate,
hourlySnapReplicationPolicy=hourly_snap_replication_policy,
dailySnapReplicationPolicy=daily_snap_replication_policy,
replicateExistingSnaps=replicate_existing_snaps)
resp = cli.type_action(cls().resource_class,
'createReplicationSessionWDestResProvisioning', **req_body)
resp.raise_if_err()
session_resp = resp.first_content['id']
return cls.get(cli, _id=session_resp['id'])
|
create_with_dst_resource_provisioning
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, cli, src_resource_id, dst_resource_config, max_time_out_of_sync,\n name=None, remote_system=None, src_spa_interface=None,\n src_spb_interface=None, dst_spa_interface=None, dst_spb_interface=None,\n dst_resource_element_configs=None, auto_initiate=None,\n hourly_snap_replication_policy=None, daily_snap_replication_policy=None,\n replicate_existing_snaps=None):\n req_body = cli.make_body(srcResourceId=src_resource_id,\n dstResourceConfig=dst_resource_config, maxTimeOutOfSync=\n max_time_out_of_sync, name=name, remoteSystem=remote_system,\n srcSPAInterface=src_spa_interface, srcSPBInterface=\n src_spb_interface, dstSPAInterface=dst_spa_interface,\n dstSPBInterface=dst_spb_interface, dstResourceElementConfigs=\n dst_resource_element_configs, autoInitiate=auto_initiate,\n hourlySnapReplicationPolicy=hourly_snap_replication_policy,\n dailySnapReplicationPolicy=daily_snap_replication_policy,\n replicateExistingSnaps=replicate_existing_snaps)\n resp = cli.type_action(cls().resource_class,\n 'createReplicationSessionWDestResProvisioning', **req_body)\n resp.raise_if_err()\n session_resp = resp.first_content['id']\n return cls.get(cli, _id=session_resp['id'])\n",
"role": "user"
},
{
"content": "create_with_dst_resource_provisioning",
"role": "assistant"
}
] |
def x(self, max_time_out_of_sync=None, name=None,
hourly_snap_replication_policy=None, daily_snap_replication_policy=None,
src_spa_interface=None, src_spb_interface=None, dst_spa_interface=None,
dst_spb_interface=None):
req_body = self._cli.make_body(maxTimeOutOfSync=max_time_out_of_sync,
name=name, hourlySnapReplicationPolicy=
hourly_snap_replication_policy, dailySnapReplicationPolicy=
daily_snap_replication_policy, srcSPAInterface=src_spa_interface,
srcSPBInterface=src_spb_interface, dstSPAInterface=
dst_spa_interface, dstSPBInterface=dst_spb_interface)
resp = self.action('modify', **req_body)
resp.raise_if_err()
return resp
|
modify
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, max_time_out_of_sync=None, name=None,\n hourly_snap_replication_policy=None, daily_snap_replication_policy=None,\n src_spa_interface=None, src_spb_interface=None, dst_spa_interface=None,\n dst_spb_interface=None):\n req_body = self._cli.make_body(maxTimeOutOfSync=max_time_out_of_sync,\n name=name, hourlySnapReplicationPolicy=\n hourly_snap_replication_policy, dailySnapReplicationPolicy=\n daily_snap_replication_policy, srcSPAInterface=src_spa_interface,\n srcSPBInterface=src_spb_interface, dstSPAInterface=\n dst_spa_interface, dstSPBInterface=dst_spb_interface)\n resp = self.action('modify', **req_body)\n resp.raise_if_err()\n return resp\n",
"role": "user"
},
{
"content": "modify",
"role": "assistant"
}
] |
def x(self, force_full_copy=None, src_spa_interface=None, src_spb_interface
=None, dst_spa_interface=None, dst_spb_interface=None):
req_body = self._cli.make_body(forceFullCopy=force_full_copy,
srcSPAInterface=src_spa_interface, srcSPBInterface=
src_spb_interface, dstSPAInterface=dst_spa_interface,
dstSPBInterface=dst_spb_interface)
resp = self.action('resume', **req_body)
resp.raise_if_err()
return resp
|
resume
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, force_full_copy=None, src_spa_interface=None, src_spb_interface\n =None, dst_spa_interface=None, dst_spb_interface=None):\n req_body = self._cli.make_body(forceFullCopy=force_full_copy,\n srcSPAInterface=src_spa_interface, srcSPBInterface=\n src_spb_interface, dstSPAInterface=dst_spa_interface,\n dstSPBInterface=dst_spb_interface)\n resp = self.action('resume', **req_body)\n resp.raise_if_err()\n return resp\n",
"role": "user"
},
{
"content": "resume",
"role": "assistant"
}
] |
def x(self, sync=None, force=None):
req_body = self._cli.make_body(sync=sync, force=force)
resp = self.action('failover', **req_body)
resp.raise_if_err()
return resp
|
failover
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, sync=None, force=None):\n req_body = self._cli.make_body(sync=sync, force=force)\n resp = self.action('failover', **req_body)\n resp.raise_if_err()\n return resp\n",
"role": "user"
},
{
"content": "failover",
"role": "assistant"
}
] |
def x(self, force_full_copy=None):
req_body = self._cli.make_body(forceFullCopy=force_full_copy)
resp = self.action('failback', **req_body)
resp.raise_if_err()
return resp
|
failback
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, force_full_copy=None):\n req_body = self._cli.make_body(forceFullCopy=force_full_copy)\n resp = self.action('failback', **req_body)\n resp.raise_if_err()\n return resp\n",
"role": "user"
},
{
"content": "failback",
"role": "assistant"
}
] |
def x(self):
if len(self._orbit_breaks) == 0:
self._detBreaks()
self._fullDayData = self.sat.data.copy()
self._current = 0
|
calcOrbits
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if len(self._orbit_breaks) == 0:\n self._detBreaks()\n self._fullDayData = self.sat.data.copy()\n self._current = 0\n",
"role": "user"
},
{
"content": "calcOrbits",
"role": "assistant"
}
] |
def x(self, orbit_index_period=24.0):
if self.orbit_index is None:
raise ValueError('Orbit properties must be defined at ' +
'pysat.Instrument object instantiation.' + 'See Instrument docs.')
else:
try:
self.sat[self.orbit_index]
except ValueError:
raise ValueError('Provided orbit index does not exist in ' +
'loaded data')
lt_diff = self.sat[self.orbit_index].diff()
ut_vals = Series(self.sat.data.index)
ut_diff = ut_vals.diff()
ind, = np.where(lt_diff < -0.1)
if len(ind) > 0:
ind = np.hstack((ind, np.array([len(self.sat[self.orbit_index])])))
dist = ind[1:] - ind[0:-1]
if len(ind) > 1:
if min(dist) == 1:
print('There are orbit breaks right next to each other')
ind = ind[:-1][dist > 1]
new_ind = []
for idx in ind:
tidx, = np.where(lt_diff[idx - 5:idx + 6] > 0.1)
if len(tidx) != 0:
for tidx in tidx:
if ut_diff[idx - 5:idx + 6].iloc[tidx] < lt_diff[idx -
5:idx + 6].iloc[tidx
] / orbit_index_period * self.orbit_period:
pass
else:
new_ind.append(idx)
break
else:
new_ind.append(idx)
ind = np.array(new_ind)
ut_change_vs_period = ut_diff > self.orbit_period
norm_ut = ut_diff / self.orbit_period
norm_ut_vs_norm_lt = norm_ut.gt(np.abs(lt_diff.values / orbit_index_period)
)
ut_ind, = np.where(ut_change_vs_period | norm_ut_vs_norm_lt & (norm_ut >
0.95))
if len(ut_ind) > 0:
ind = np.hstack((ind, ut_ind))
ind = np.sort(ind)
ind = np.unique(ind)
print('Time Gap')
orbit_ut_diff = ut_vals[ind].diff()
orbit_lt_diff = self.sat[self.orbit_index][ind].diff()
idx, = np.where(orbit_ut_diff / self.orbit_period - orbit_lt_diff.
values / orbit_index_period > 0.97)
if len(idx) > 0:
if idx[0] != 0:
idx = np.hstack((0, idx))
else:
idx = np.array([0])
if len(ind) > 0:
ind = ind[idx]
if ind[0] != 0:
ind = np.hstack((np.array([0]), ind))
else:
ind = np.array([0])
num_orbits = len(ind)
self._orbit_breaks = ind
self.num = num_orbits
|
equaBreaks
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, orbit_index_period=24.0):\n if self.orbit_index is None:\n raise ValueError('Orbit properties must be defined at ' +\n 'pysat.Instrument object instantiation.' + 'See Instrument docs.')\n else:\n try:\n self.sat[self.orbit_index]\n except ValueError:\n raise ValueError('Provided orbit index does not exist in ' +\n 'loaded data')\n lt_diff = self.sat[self.orbit_index].diff()\n ut_vals = Series(self.sat.data.index)\n ut_diff = ut_vals.diff()\n ind, = np.where(lt_diff < -0.1)\n if len(ind) > 0:\n ind = np.hstack((ind, np.array([len(self.sat[self.orbit_index])])))\n dist = ind[1:] - ind[0:-1]\n if len(ind) > 1:\n if min(dist) == 1:\n print('There are orbit breaks right next to each other')\n ind = ind[:-1][dist > 1]\n new_ind = []\n for idx in ind:\n tidx, = np.where(lt_diff[idx - 5:idx + 6] > 0.1)\n if len(tidx) != 0:\n for tidx in tidx:\n if ut_diff[idx - 5:idx + 6].iloc[tidx] < lt_diff[idx - \n 5:idx + 6].iloc[tidx\n ] / orbit_index_period * self.orbit_period:\n pass\n else:\n new_ind.append(idx)\n break\n else:\n new_ind.append(idx)\n ind = np.array(new_ind)\n ut_change_vs_period = ut_diff > self.orbit_period\n norm_ut = ut_diff / self.orbit_period\n norm_ut_vs_norm_lt = norm_ut.gt(np.abs(lt_diff.values / orbit_index_period)\n )\n ut_ind, = np.where(ut_change_vs_period | norm_ut_vs_norm_lt & (norm_ut >\n 0.95))\n if len(ut_ind) > 0:\n ind = np.hstack((ind, ut_ind))\n ind = np.sort(ind)\n ind = np.unique(ind)\n print('Time Gap')\n orbit_ut_diff = ut_vals[ind].diff()\n orbit_lt_diff = self.sat[self.orbit_index][ind].diff()\n idx, = np.where(orbit_ut_diff / self.orbit_period - orbit_lt_diff.\n values / orbit_index_period > 0.97)\n if len(idx) > 0:\n if idx[0] != 0:\n idx = np.hstack((0, idx))\n else:\n idx = np.array([0])\n if len(ind) > 0:\n ind = ind[idx]\n if ind[0] != 0:\n ind = np.hstack((np.array([0]), ind))\n else:\n ind = np.array([0])\n num_orbits = len(ind)\n self._orbit_breaks = ind\n self.num = num_orbits\n",
"role": "user"
},
{
"content": "equaBreaks",
"role": "assistant"
}
] |
def x(self):
if self.orbit_index is None:
raise ValueError('Orbit properties must be defined at ' +
'pysat.Instrument object instantiation.' + 'See Instrument docs.')
else:
try:
self.sat[self.orbit_index]
except ValueError:
raise ValueError('Provided orbit index does not appear to ' +
'exist in loaded data')
pos = self.sat[self.orbit_index] >= 0
npos = -pos
change = pos.values[:-1] & npos.values[1:] | npos.values[:-1] & pos.values[
1:]
ind, = np.where(change)
ind += 1
ut_diff = Series(self.sat.data.index).diff()
ut_ind, = np.where(ut_diff / self.orbit_period > 0.95)
if len(ut_ind) > 0:
ind = np.hstack((ind, ut_ind))
ind = np.sort(ind)
ind = np.unique(ind)
if ind[0] != 0:
ind = np.hstack((np.array([0]), ind))
num_orbits = len(ind)
self._orbit_breaks = ind
self.num = num_orbits
|
polarBreaks
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.orbit_index is None:\n raise ValueError('Orbit properties must be defined at ' +\n 'pysat.Instrument object instantiation.' + 'See Instrument docs.')\n else:\n try:\n self.sat[self.orbit_index]\n except ValueError:\n raise ValueError('Provided orbit index does not appear to ' +\n 'exist in loaded data')\n pos = self.sat[self.orbit_index] >= 0\n npos = -pos\n change = pos.values[:-1] & npos.values[1:] | npos.values[:-1] & pos.values[\n 1:]\n ind, = np.where(change)\n ind += 1\n ut_diff = Series(self.sat.data.index).diff()\n ut_ind, = np.where(ut_diff / self.orbit_period > 0.95)\n if len(ut_ind) > 0:\n ind = np.hstack((ind, ut_ind))\n ind = np.sort(ind)\n ind = np.unique(ind)\n if ind[0] != 0:\n ind = np.hstack((np.array([0]), ind))\n num_orbits = len(ind)\n self._orbit_breaks = ind\n self.num = num_orbits\n",
"role": "user"
},
{
"content": "polarBreaks",
"role": "assistant"
}
] |
def x(self):
if self.orbit_index is None:
raise ValueError('Orbit properties must be defined at ' +
'pysat.Instrument object instantiation.' + 'See Instrument docs.')
else:
try:
self.sat[self.orbit_index]
except ValueError:
raise ValueError('Provided orbit index does not appear to ' +
'exist in loaded data')
uniq_vals = self.sat[self.orbit_index].unique()
orbit_index = []
for val in uniq_vals:
idx, = np.where(val == self.sat[self.orbit_index].values)
orbit_index.append(idx[0])
if orbit_index[0] != 0:
ind = np.hstack((np.array([0]), orbit_index))
else:
ind = orbit_index
num_orbits = len(ind)
self._orbit_breaks = ind
self.num = num_orbits
|
orbitNumberBreaks
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.orbit_index is None:\n raise ValueError('Orbit properties must be defined at ' +\n 'pysat.Instrument object instantiation.' + 'See Instrument docs.')\n else:\n try:\n self.sat[self.orbit_index]\n except ValueError:\n raise ValueError('Provided orbit index does not appear to ' +\n 'exist in loaded data')\n uniq_vals = self.sat[self.orbit_index].unique()\n orbit_index = []\n for val in uniq_vals:\n idx, = np.where(val == self.sat[self.orbit_index].values)\n orbit_index.append(idx[0])\n if orbit_index[0] != 0:\n ind = np.hstack((np.array([0]), orbit_index))\n else:\n ind = orbit_index\n num_orbits = len(ind)\n self._orbit_breaks = ind\n self.num = num_orbits\n",
"role": "user"
},
{
"content": "orbitNumberBreaks",
"role": "assistant"
}
] |
def x(self, orbit=None):
if not self.sat.empty:
self._calcOrbits()
if orbit is not None:
if orbit == -1:
self.sat.data = self._fullDayData[self._orbit_breaks[self.
num + orbit]:]
self._current = self.num + orbit + 1
elif (orbit < 0) & (orbit >= -self.num):
self.sat.data = self._fullDayData[self._orbit_breaks[self.
num + orbit]:self._orbit_breaks[self.num + orbit + 1]]
self._current = self.num + orbit + 1
elif (orbit < self.num) & (orbit != 0):
self.sat.data = self._fullDayData[self._orbit_breaks[orbit -
1]:self._orbit_breaks[orbit]]
self._current = orbit
elif orbit == self.num:
self.sat.data = self._fullDayData[self._orbit_breaks[orbit -
1]:]
self._current = orbit
elif orbit == 0:
raise ValueError('Orbits internally indexed by 1, 0 not ' +
'allowed')
else:
self.sat.data = []
raise ValueError('Requested an orbit past total orbits ' +
'for day')
else:
raise ValueError('Must set an orbit')
|
getBasicOrbit
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, orbit=None):\n if not self.sat.empty:\n self._calcOrbits()\n if orbit is not None:\n if orbit == -1:\n self.sat.data = self._fullDayData[self._orbit_breaks[self.\n num + orbit]:]\n self._current = self.num + orbit + 1\n elif (orbit < 0) & (orbit >= -self.num):\n self.sat.data = self._fullDayData[self._orbit_breaks[self.\n num + orbit]:self._orbit_breaks[self.num + orbit + 1]]\n self._current = self.num + orbit + 1\n elif (orbit < self.num) & (orbit != 0):\n self.sat.data = self._fullDayData[self._orbit_breaks[orbit -\n 1]:self._orbit_breaks[orbit]]\n self._current = orbit\n elif orbit == self.num:\n self.sat.data = self._fullDayData[self._orbit_breaks[orbit -\n 1]:]\n self._current = orbit\n elif orbit == 0:\n raise ValueError('Orbits internally indexed by 1, 0 not ' +\n 'allowed')\n else:\n self.sat.data = []\n raise ValueError('Requested an orbit past total orbits ' +\n 'for day')\n else:\n raise ValueError('Must set an orbit')\n",
"role": "user"
},
{
"content": "getBasicOrbit",
"role": "assistant"
}
] |
def x(self, orbit=None):
if not self.sat.empty:
self._calcOrbits()
if orbit is not None:
if orbit < 0:
orbit = self.num + 1 + orbit
if orbit == 1:
try:
true_date = self.sat.date
self.sat.prev()
if not self.sat.empty:
self.load(orbit=-1)
else:
self.sat.next()
self._getBasicOrbit(orbit=1)
delta = true_date - self.sat.data.index[0]
if delta >= self.orbit_period:
self.next()
except StopIteration:
self._getBasicOrbit(orbit=1)
print('Loaded Orbit:%i' % (self._current - 1))
elif orbit == self.num:
if self.num != 1:
self._getBasicOrbit(self.num - 1)
self.next()
else:
self._getBasicOrbit(orbit=-1)
elif orbit < self.num:
self._getBasicOrbit(orbit)
print('Loaded Orbit:%i' % (self._current - 1))
else:
self.sat.data = DataFrame()
raise Exception('Requested an orbit past total orbits for day')
else:
raise Exception('Must set an orbit')
else:
print('No data loaded in instrument object to determine orbits.')
|
load
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, orbit=None):\n if not self.sat.empty:\n self._calcOrbits()\n if orbit is not None:\n if orbit < 0:\n orbit = self.num + 1 + orbit\n if orbit == 1:\n try:\n true_date = self.sat.date\n self.sat.prev()\n if not self.sat.empty:\n self.load(orbit=-1)\n else:\n self.sat.next()\n self._getBasicOrbit(orbit=1)\n delta = true_date - self.sat.data.index[0]\n if delta >= self.orbit_period:\n self.next()\n except StopIteration:\n self._getBasicOrbit(orbit=1)\n print('Loaded Orbit:%i' % (self._current - 1))\n elif orbit == self.num:\n if self.num != 1:\n self._getBasicOrbit(self.num - 1)\n self.next()\n else:\n self._getBasicOrbit(orbit=-1)\n elif orbit < self.num:\n self._getBasicOrbit(orbit)\n print('Loaded Orbit:%i' % (self._current - 1))\n else:\n self.sat.data = DataFrame()\n raise Exception('Requested an orbit past total orbits for day')\n else:\n raise Exception('Must set an orbit')\n else:\n print('No data loaded in instrument object to determine orbits.')\n",
"role": "user"
},
{
"content": "load",
"role": "assistant"
}
] |
def x(self, *arg, **kwarg):
if not self.sat.empty:
self._calcOrbits()
if self._current == self.num - 1:
self._getBasicOrbit(orbit=-1)
load_next = True
if self.sat._iter_type == 'date':
delta = self.sat.date - self.sat.data.index[-1
] + pds.Timedelta('1 day')
if delta >= self.orbit_period:
load_next = False
if load_next:
temp_orbit_data = self.sat.data.copy()
try:
self.sat.next()
if not self.sat.empty:
self.sat.data = pds.concat([temp_orbit_data[:self.
sat.data.index[0] - pds.DateOffset(microseconds
=1)], self.sat.data])
self._getBasicOrbit(orbit=1)
else:
self.sat.prev()
self._getBasicOrbit(orbit=-1)
except StopIteration:
pass
del temp_orbit_data
print('Loaded Orbit:%i' % (self._current - 1))
elif self._current == self.num:
temp_orbit_data = self.sat.data.copy()
self.sat.next()
if not self.sat.empty:
pad_next = True
if self.sat._iter_type == 'date':
delta = self.sat.date - temp_orbit_data.index[-1]
if delta >= self.orbit_period:
pad_next = False
if pad_next:
self.sat.data = pds.concat([temp_orbit_data[:self.sat.
data.index[0] - pds.DateOffset(microseconds=1)],
self.sat.data])
self._getBasicOrbit(orbit=2)
else:
self._getBasicOrbit(orbit=1)
if self.sat._iter_type == 'date':
delta = self.sat.date + pds.DateOffset(days=1
) - self.sat.data.index[0]
if delta < self.orbit_period:
self._current = self.num - 1
self.next()
else:
while self.sat.empty:
self.sat.next()
self._getBasicOrbit(orbit=1)
del temp_orbit_data
print('Loaded Orbit:%i' % (self._current - 1))
elif self._current == 0:
self.load(orbit=1)
elif self._current < self.num - 1:
self._getBasicOrbit(orbit=self._current + 1)
print('Loaded Orbit:%i' % (self._current - 1))
else:
raise Exception('You ended up where nobody should ever be. ' +
'Talk to someone about this fundamental ' + 'failure.')
else:
while self.sat.empty:
self.sat.next()
self.next()
|
next
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *arg, **kwarg):\n if not self.sat.empty:\n self._calcOrbits()\n if self._current == self.num - 1:\n self._getBasicOrbit(orbit=-1)\n load_next = True\n if self.sat._iter_type == 'date':\n delta = self.sat.date - self.sat.data.index[-1\n ] + pds.Timedelta('1 day')\n if delta >= self.orbit_period:\n load_next = False\n if load_next:\n temp_orbit_data = self.sat.data.copy()\n try:\n self.sat.next()\n if not self.sat.empty:\n self.sat.data = pds.concat([temp_orbit_data[:self.\n sat.data.index[0] - pds.DateOffset(microseconds\n =1)], self.sat.data])\n self._getBasicOrbit(orbit=1)\n else:\n self.sat.prev()\n self._getBasicOrbit(orbit=-1)\n except StopIteration:\n pass\n del temp_orbit_data\n print('Loaded Orbit:%i' % (self._current - 1))\n elif self._current == self.num:\n temp_orbit_data = self.sat.data.copy()\n self.sat.next()\n if not self.sat.empty:\n pad_next = True\n if self.sat._iter_type == 'date':\n delta = self.sat.date - temp_orbit_data.index[-1]\n if delta >= self.orbit_period:\n pad_next = False\n if pad_next:\n self.sat.data = pds.concat([temp_orbit_data[:self.sat.\n data.index[0] - pds.DateOffset(microseconds=1)],\n self.sat.data])\n self._getBasicOrbit(orbit=2)\n else:\n self._getBasicOrbit(orbit=1)\n if self.sat._iter_type == 'date':\n delta = self.sat.date + pds.DateOffset(days=1\n ) - self.sat.data.index[0]\n if delta < self.orbit_period:\n self._current = self.num - 1\n self.next()\n else:\n while self.sat.empty:\n self.sat.next()\n self._getBasicOrbit(orbit=1)\n del temp_orbit_data\n print('Loaded Orbit:%i' % (self._current - 1))\n elif self._current == 0:\n self.load(orbit=1)\n elif self._current < self.num - 1:\n self._getBasicOrbit(orbit=self._current + 1)\n print('Loaded Orbit:%i' % (self._current - 1))\n else:\n raise Exception('You ended up where nobody should ever be. ' +\n 'Talk to someone about this fundamental ' + 'failure.')\n else:\n while self.sat.empty:\n self.sat.next()\n self.next()\n",
"role": "user"
},
{
"content": "next",
"role": "assistant"
}
] |
def x(fnames, tag=None, sat_id=None):
data = pds.DataFrame()
for filename in fnames:
fname = filename[0:-11]
with open(fname) as f:
lines = f.readlines()
idx = 0
max_lines = 0
for line in lines:
if len(line) > 1:
max_lines += 1
yr = np.zeros(max_lines * 24, dtype=int)
mo = np.zeros(max_lines * 24, dtype=int)
day = np.zeros(max_lines * 24, dtype=int)
ut = np.zeros(max_lines * 24, dtype=int)
dst = np.zeros(max_lines * 24, dtype=int)
for line in lines:
if len(line) > 1:
temp_year = int(line[14:16] + line[3:5])
if temp_year > 57:
temp_year += 1900
else:
temp_year += 2000
yr[idx:idx + 24] = temp_year
mo[idx:idx + 24] = int(line[5:7])
day[idx:idx + 24] = int(line[8:10])
ut[idx:idx + 24] = np.arange(24)
temp = line.strip()[20:-4]
temp2 = [temp[4 * i:4 * (i + 1)] for i in np.arange(24)]
dst[idx:idx + 24] = temp2
idx += 24
start = pds.datetime(yr[0], mo[0], day[0], ut[0])
stop = pds.datetime(yr[-1], mo[-1], day[-1], ut[-1])
dates = pds.date_range(start, stop, freq='H')
new_data = pds.DataFrame(dst, index=dates, columns=['dst'])
new_date = pysat.datetime.strptime(filename[-10:], '%Y-%m-%d')
idx, = np.where((new_data.index >= new_date) & (new_data.index <
new_date + pds.DateOffset(days=1)))
new_data = new_data.iloc[idx, :]
data = pds.concat([data, new_data], axis=0)
return data, pysat.Meta()
|
load
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fnames, tag=None, sat_id=None):\n data = pds.DataFrame()\n for filename in fnames:\n fname = filename[0:-11]\n with open(fname) as f:\n lines = f.readlines()\n idx = 0\n max_lines = 0\n for line in lines:\n if len(line) > 1:\n max_lines += 1\n yr = np.zeros(max_lines * 24, dtype=int)\n mo = np.zeros(max_lines * 24, dtype=int)\n day = np.zeros(max_lines * 24, dtype=int)\n ut = np.zeros(max_lines * 24, dtype=int)\n dst = np.zeros(max_lines * 24, dtype=int)\n for line in lines:\n if len(line) > 1:\n temp_year = int(line[14:16] + line[3:5])\n if temp_year > 57:\n temp_year += 1900\n else:\n temp_year += 2000\n yr[idx:idx + 24] = temp_year\n mo[idx:idx + 24] = int(line[5:7])\n day[idx:idx + 24] = int(line[8:10])\n ut[idx:idx + 24] = np.arange(24)\n temp = line.strip()[20:-4]\n temp2 = [temp[4 * i:4 * (i + 1)] for i in np.arange(24)]\n dst[idx:idx + 24] = temp2\n idx += 24\n start = pds.datetime(yr[0], mo[0], day[0], ut[0])\n stop = pds.datetime(yr[-1], mo[-1], day[-1], ut[-1])\n dates = pds.date_range(start, stop, freq='H')\n new_data = pds.DataFrame(dst, index=dates, columns=['dst'])\n new_date = pysat.datetime.strptime(filename[-10:], '%Y-%m-%d')\n idx, = np.where((new_data.index >= new_date) & (new_data.index <\n new_date + pds.DateOffset(days=1)))\n new_data = new_data.iloc[idx, :]\n data = pds.concat([data, new_data], axis=0)\n return data, pysat.Meta()\n",
"role": "user"
},
{
"content": "load",
"role": "assistant"
}
] |
def x(yr, mo, day):
yr = '20' + yr
yr = int(yr)
mo = int(mo)
day = int(day)
return pds.datetime(yr, mo, day)
|
parse
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(yr, mo, day):\n yr = '20' + yr\n yr = int(yr)\n mo = int(mo)\n day = int(day)\n return pds.datetime(yr, mo, day)\n",
"role": "user"
},
{
"content": "parse",
"role": "assistant"
}
] |
def x(fnames, tag=None, sat_id=None):
data = pds.DataFrame()
colspec = [(0, 2), (2, 4), (4, 6), (7, 10), (10, 13), (13, 16), (16, 19
), (19, 23), (23, 26), (26, 29), (29, 32), (32, 50)]
for filename in fnames:
fname = filename[0:-11]
date = pysat.datetime.strptime(filename[-10:], '%Y-%m-%d')
temp = pds.read_fwf(fname, colspecs=colspec, skipfooter=4, header=
None, parse_dates=[[0, 1, 2]], date_parser=_parse, index_col=
'0_1_2')
idx, = np.where((temp.index >= date) & (temp.index < date + pds.
DateOffset(days=1)))
temp = temp.iloc[idx, :]
data = pds.concat([data, temp], axis=0)
data = data.iloc[:, 0:-1]
s = pds.Series()
for i in np.arange(8):
temp = pds.Series(data.iloc[:, i].values, index=data.index + pds.
DateOffset(hours=int(3 * i)))
s = s.append(temp)
s = s.sort_index()
s.index.name = 'time'
first = np.array([float(x[0]) for x in s])
flag = np.array([x[1] for x in s])
ind, = np.where(flag == '+')
first[ind] += 1.0 / 3.0
ind, = np.where(flag == '-')
first[ind] -= 1.0 / 3.0
result = pds.DataFrame(first, columns=['kp'], index=s.index)
return result, pysat.Meta()
|
load
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fnames, tag=None, sat_id=None):\n data = pds.DataFrame()\n colspec = [(0, 2), (2, 4), (4, 6), (7, 10), (10, 13), (13, 16), (16, 19\n ), (19, 23), (23, 26), (26, 29), (29, 32), (32, 50)]\n for filename in fnames:\n fname = filename[0:-11]\n date = pysat.datetime.strptime(filename[-10:], '%Y-%m-%d')\n temp = pds.read_fwf(fname, colspecs=colspec, skipfooter=4, header=\n None, parse_dates=[[0, 1, 2]], date_parser=_parse, index_col=\n '0_1_2')\n idx, = np.where((temp.index >= date) & (temp.index < date + pds.\n DateOffset(days=1)))\n temp = temp.iloc[idx, :]\n data = pds.concat([data, temp], axis=0)\n data = data.iloc[:, 0:-1]\n s = pds.Series()\n for i in np.arange(8):\n temp = pds.Series(data.iloc[:, i].values, index=data.index + pds.\n DateOffset(hours=int(3 * i)))\n s = s.append(temp)\n s = s.sort_index()\n s.index.name = 'time'\n first = np.array([float(x[0]) for x in s])\n flag = np.array([x[1] for x in s])\n ind, = np.where(flag == '+')\n first[ind] += 1.0 / 3.0\n ind, = np.where(flag == '-')\n first[ind] -= 1.0 / 3.0\n result = pds.DataFrame(first, columns=['kp'], index=s.index)\n return result, pysat.Meta()\n",
"role": "user"
},
{
"content": "load",
"role": "assistant"
}
] |
def x(date_array, tag, sat_id, data_path, user=None, password=None):
import ftplib
from ftplib import FTP
import sys
ftp = FTP('ftp.gfz-potsdam.de')
ftp.login()
ftp.cwd('/pub/home/obs/kp-ap/tab')
for date in date_array:
fname = 'kp{year:02d}{month:02d}.tab'
fname = fname.format(year=date.year - date.year // 100 * 100, month
=date.month)
local_fname = fname
saved_fname = os.path.join(data_path, local_fname)
try:
print('Downloading file for ' + date.strftime('%D'))
sys.stdout.flush()
ftp.retrbinary('RETR ' + fname, open(saved_fname, 'wb').write)
except ftplib.error_perm as exception:
if str(exception.args[0]).split(' ', 1)[0] != '550':
raise
else:
os.remove(saved_fname)
print('File not available for ' + date.strftime('%D'))
ftp.close()
return
|
download
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(date_array, tag, sat_id, data_path, user=None, password=None):\n import ftplib\n from ftplib import FTP\n import sys\n ftp = FTP('ftp.gfz-potsdam.de')\n ftp.login()\n ftp.cwd('/pub/home/obs/kp-ap/tab')\n for date in date_array:\n fname = 'kp{year:02d}{month:02d}.tab'\n fname = fname.format(year=date.year - date.year // 100 * 100, month\n =date.month)\n local_fname = fname\n saved_fname = os.path.join(data_path, local_fname)\n try:\n print('Downloading file for ' + date.strftime('%D'))\n sys.stdout.flush()\n ftp.retrbinary('RETR ' + fname, open(saved_fname, 'wb').write)\n except ftplib.error_perm as exception:\n if str(exception.args[0]).split(' ', 1)[0] != '550':\n raise\n else:\n os.remove(saved_fname)\n print('File not available for ' + date.strftime('%D'))\n ftp.close()\n return\n",
"role": "user"
},
{
"content": "download",
"role": "assistant"
}
] |
def x(sat, maxKp=None, filterTime=None, kpData=None, kp_inst=None):
if kp_inst is not None:
kp_inst.load(date=sat.date, verifyPad=True)
kpData = kp_inst
elif kpData is None:
kp = pysat.Instrument('sw', 'kp', pad=pds.DateOffset(days=1))
kp.load(date=sat.date, verifyPad=True)
kpData = kp
if maxKp is None:
maxKp = 3 + 1.0 / 3.0
if filterTime is None:
filterTime = 24
date = sat.date
selData = kpData[date - pds.DateOffset(days=1):date + pds.DateOffset(
days=1)]
ind, = np.where(selData['kp'] >= maxKp)
for lind in ind:
sat.data[selData.index[lind]:selData.index[lind] + pds.DateOffset(
hours=filterTime)] = np.NaN
sat.data = sat.data.dropna(axis=0, how='all')
return
|
filter_geoquiet
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sat, maxKp=None, filterTime=None, kpData=None, kp_inst=None):\n if kp_inst is not None:\n kp_inst.load(date=sat.date, verifyPad=True)\n kpData = kp_inst\n elif kpData is None:\n kp = pysat.Instrument('sw', 'kp', pad=pds.DateOffset(days=1))\n kp.load(date=sat.date, verifyPad=True)\n kpData = kp\n if maxKp is None:\n maxKp = 3 + 1.0 / 3.0\n if filterTime is None:\n filterTime = 24\n date = sat.date\n selData = kpData[date - pds.DateOffset(days=1):date + pds.DateOffset(\n days=1)]\n ind, = np.where(selData['kp'] >= maxKp)\n for lind in ind:\n sat.data[selData.index[lind]:selData.index[lind] + pds.DateOffset(\n hours=filterTime)] = np.NaN\n sat.data = sat.data.dropna(axis=0, how='all')\n return\n",
"role": "user"
},
{
"content": "filter_geoquiet",
"role": "assistant"
}
] |
def x(self, converter_str):
ret = None
if converter_str is not None:
converter_desc_list = converter_str.split('.')
if len(converter_desc_list) == 1:
converter = converter_desc_list[0]
ret = getattr(cvt, converter, None)
if ret is None:
ret = self.get_converter(converter)
if ret is None:
ret = self.get_resource_clz_by_name(converter)
if ret is None:
ret = self.get_enum_by_name(converter)
if ret is None:
ret = self.get(converter)
if ret is None and converter_str is not None:
raise ValueError('Specified converter not supported: {}'.format
(converter_str))
return ret
|
get_converter
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, converter_str):\n ret = None\n if converter_str is not None:\n converter_desc_list = converter_str.split('.')\n if len(converter_desc_list) == 1:\n converter = converter_desc_list[0]\n ret = getattr(cvt, converter, None)\n if ret is None:\n ret = self.get_converter(converter)\n if ret is None:\n ret = self.get_resource_clz_by_name(converter)\n if ret is None:\n ret = self.get_enum_by_name(converter)\n if ret is None:\n ret = self.get(converter)\n if ret is None and converter_str is not None:\n raise ValueError('Specified converter not supported: {}'.format\n (converter_str))\n return ret\n",
"role": "user"
},
{
"content": "get_converter",
"role": "assistant"
}
] |
def x(self, local_path, remote_path):
sftp_client = self.transport.open_sftp_client()
LOG.debug(
'Copy the local file to remote. Source=%(src)s. Target=%(target)s.' %
{'src': local_path, 'target': remote_path})
try:
sftp_client.put(local_path, remote_path)
except Exception as ex:
LOG.error('Failed to copy the local file to remote. Reason: %s.' %
six.text_type(ex))
raise SFtpExecutionError(err=ex)
|
copy_file_to_remote
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, local_path, remote_path):\n sftp_client = self.transport.open_sftp_client()\n LOG.debug(\n 'Copy the local file to remote. Source=%(src)s. Target=%(target)s.' %\n {'src': local_path, 'target': remote_path})\n try:\n sftp_client.put(local_path, remote_path)\n except Exception as ex:\n LOG.error('Failed to copy the local file to remote. Reason: %s.' %\n six.text_type(ex))\n raise SFtpExecutionError(err=ex)\n",
"role": "user"
},
{
"content": "copy_file_to_remote",
"role": "assistant"
}
] |
def x(self, remote_path, local_path):
sftp_client = self.transport.open_sftp_client()
LOG.debug('Get the remote file. Source=%(src)s. Target=%(target)s.' % {
'src': remote_path, 'target': local_path})
try:
sftp_client.get(remote_path, local_path)
except Exception as ex:
LOG.error('Failed to secure copy. Reason: %s.' % six.text_type(ex))
raise SFtpExecutionError(err=ex)
|
get_remote_file
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, remote_path, local_path):\n sftp_client = self.transport.open_sftp_client()\n LOG.debug('Get the remote file. Source=%(src)s. Target=%(target)s.' % {\n 'src': remote_path, 'target': local_path})\n try:\n sftp_client.get(remote_path, local_path)\n except Exception as ex:\n LOG.error('Failed to secure copy. Reason: %s.' % six.text_type(ex))\n raise SFtpExecutionError(err=ex)\n",
"role": "user"
},
{
"content": "get_remote_file",
"role": "assistant"
}
] |
def x(self):
if 'isLive' in self.__dict__ and self.isLive:
self.transport.close()
self.isLive = False
|
close
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if 'isLive' in self.__dict__ and self.isLive:\n self.transport.close()\n self.isLive = False\n",
"role": "user"
},
{
"content": "close",
"role": "assistant"
}
] |
def x(check_object=False, check_invalid_data_mover=False):
def x(f):
@functools.wraps(f)
def x(self, *argv, **kwargs):
request = f(self, *argv, **kwargs)
return self.request(request, check_object=check_object,
check_invalid_data_mover=check_invalid_data_mover)
return func_wrapper
return decorator
|
xml_request
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(check_object=False, check_invalid_data_mover=False):\n\n def x(f):\n\n @functools.wraps(f)\n def x(self, *argv, **kwargs):\n request = f(self, *argv, **kwargs)\n return self.request(request, check_object=check_object,\n check_invalid_data_mover=check_invalid_data_mover)\n return func_wrapper\n return decorator\n",
"role": "user"
},
{
"content": "xml_request",
"role": "assistant"
}
] |
def x(f):
@functools.wraps(f)
def x(self, *argv, **kwargs):
commands = f(self, *argv, **kwargs)
return self.ssh_execute(['env', 'NAS_DB=/nas'] + commands)
return func_wrapper
|
nas_command
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(f):\n\n @functools.wraps(f)\n def x(self, *argv, **kwargs):\n commands = f(self, *argv, **kwargs)\n return self.ssh_execute(['env', 'NAS_DB=/nas'] + commands)\n return func_wrapper\n",
"role": "user"
},
{
"content": "nas_command",
"role": "assistant"
}
] |
def x(self, backup=None, delete_backup=False):
resp = self._cli.action(self.resource_class, self.get_id(), 'restore',
copyName=backup)
resp.raise_if_err()
backup = resp.first_content['backup']
backup_snap = UnitySnap(_id=backup['id'], cli=self._cli)
if delete_backup:
log.info('Deleting the backup snap {} as the restoration succeeded.'
.format(backup['id']))
backup_snap.delete()
return backup_snap
|
restore
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, backup=None, delete_backup=False):\n resp = self._cli.action(self.resource_class, self.get_id(), 'restore',\n copyName=backup)\n resp.raise_if_err()\n backup = resp.first_content['backup']\n backup_snap = UnitySnap(_id=backup['id'], cli=self._cli)\n if delete_backup:\n log.info('Deleting the backup snap {} as the restoration succeeded.'\n .format(backup['id']))\n backup_snap.delete()\n return backup_snap\n",
"role": "user"
},
{
"content": "restore",
"role": "assistant"
}
] |
def x(self, name, io_limit_policy=None, description=None):
if self.is_member_snap():
raise UnityCGMemberActionNotSupportError()
if self.lun and not self.lun.is_thin_enabled:
raise UnityThinCloneNotAllowedError()
return TCHelper.thin_clone(self._cli, self, name, io_limit_policy,
description)
|
thin_clone
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name, io_limit_policy=None, description=None):\n if self.is_member_snap():\n raise UnityCGMemberActionNotSupportError()\n if self.lun and not self.lun.is_thin_enabled:\n raise UnityThinCloneNotAllowedError()\n return TCHelper.thin_clone(self._cli, self, name, io_limit_policy,\n description)\n",
"role": "user"
},
{
"content": "thin_clone",
"role": "assistant"
}
] |
def x(self, async_mode=False, even_attached=False):
try:
return super(UnitySnap, self).delete(async_mode=async_mode)
except UnityDeleteAttachedSnapError:
if even_attached:
log.debug(
'Force delete the snapshot even if it is attached. First detach the snapshot from hosts, then delete again.'
)
self.detach_from(None)
return super(UnitySnap, self).delete(async_mode=async_mode)
else:
raise
|
delete
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, async_mode=False, even_attached=False):\n try:\n return super(UnitySnap, self).delete(async_mode=async_mode)\n except UnityDeleteAttachedSnapError:\n if even_attached:\n log.debug(\n 'Force delete the snapshot even if it is attached. First detach the snapshot from hosts, then delete again.'\n )\n self.detach_from(None)\n return super(UnitySnap, self).delete(async_mode=async_mode)\n else:\n raise\n",
"role": "user"
},
{
"content": "delete",
"role": "assistant"
}
] |
def x(self, type_name, base_fields=None, the_filter=None, nested_fields=None):
fields = self.get_fields(type_name, base_fields, nested_fields)
the_filter = self.dict_to_filter_string(the_filter)
url = '/api/types/{}/instances'.format(type_name)
resp = self.rest_get(url, fields=fields, filter=the_filter)
ret = resp
while resp.has_next_page:
resp = self.rest_get(url, fields=fields, filter=the_filter, page=
resp.next_page)
ret.entries.extend(resp.entries)
return ret
|
get_all
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, type_name, base_fields=None, the_filter=None, nested_fields=None):\n fields = self.get_fields(type_name, base_fields, nested_fields)\n the_filter = self.dict_to_filter_string(the_filter)\n url = '/api/types/{}/instances'.format(type_name)\n resp = self.rest_get(url, fields=fields, filter=the_filter)\n ret = resp\n while resp.has_next_page:\n resp = self.rest_get(url, fields=fields, filter=the_filter, page=\n resp.next_page)\n ret.entries.extend(resp.entries)\n return ret\n",
"role": "user"
},
{
"content": "get_all",
"role": "assistant"
}
] |
def x(self, type_name, obj_id, base_fields=None, nested_fields=None):
base_fields = self.get_fields(type_name, base_fields, nested_fields)
url = '/api/instances/{}/{}'.format(type_name, obj_id)
return self.rest_get(url, fields=base_fields)
|
get
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, type_name, obj_id, base_fields=None, nested_fields=None):\n base_fields = self.get_fields(type_name, base_fields, nested_fields)\n url = '/api/instances/{}/{}'.format(type_name, obj_id)\n return self.rest_get(url, fields=base_fields)\n",
"role": "user"
},
{
"content": "get",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.