source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(self, where=None):
table = self.lconfig.get('table')
db = self.lconfig.get('db_schema_name') or self.lconfig.get('db')
_oid = self.lconfig.get('_oid')
if is_array(_oid):
_oid = _oid[0]
sql = 'SELECT DISTINCT %s.%s FROM %s.%s' % (table, _oid, db, table)
if where:
where = [where] if isinstance(where, basestring) else list(where)
sql += ' WHERE %s' % ' OR '.join(where)
result = sorted([r[_oid] for r in self._load_sql(sql)])
return result
| sql_get_oids | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, where=None):\n table = self.lconfig.get('table')\n db = self.lconfig.get('db_schema_name') or self.lconfig.get('db')\n _oid = self.lconfig.get('_oid')\n if is_array(_oid):\n _oid = _oid[0]\n sql = 'SELECT DISTINCT %s.%s FROM %s.%s' % (table, _oid, db, table)\n if where:\n where = [where] if isinstance(where, basestring) else list(where)\n sql += ' WHERE %s' % ' OR '.join(where)\n result = sorted([r[_oid] for r in self._load_sql(sql)])\n return result\n",
"role": "user"
},
{
"content": "sql_get_oids",
"role": "assistant"
}
] |
def x(self, flush=False, autosnap=True, **kwargs):
logger.debug('Running get_objects(flush=%s, autosnap=%s, %s)' % (flush,
autosnap, kwargs))
if flush:
s = time()
result = self.flush(autosnap=autosnap, **kwargs)
diff = time() - s
logger.debug('Flush complete (%ss)' % int(diff))
return result
else:
return self
| get_objects | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, flush=False, autosnap=True, **kwargs):\n logger.debug('Running get_objects(flush=%s, autosnap=%s, %s)' % (flush,\n autosnap, kwargs))\n if flush:\n s = time()\n result = self.flush(autosnap=autosnap, **kwargs)\n diff = time() - s\n logger.debug('Flush complete (%ss)' % int(diff))\n return result\n else:\n return self\n",
"role": "user"
},
{
"content": "get_objects",
"role": "assistant"
}
] |
def x(self, cube, init=True, name=None, copy_config=True, **kwargs):
name = name or cube
config = copy(self.config) if copy_config else {}
config_file = self.config_file
container = type(self.container)
container_config = copy(self.container_config)
proxy = str(type(self.proxy))
return get_cube(cube=cube, init=init, name=name, config=config,
config_file=config_file, container=container, container_config=
container_config, proxy=proxy, proxy_config=self.proxy_config, **kwargs
)
| get_cube | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, cube, init=True, name=None, copy_config=True, **kwargs):\n name = name or cube\n config = copy(self.config) if copy_config else {}\n config_file = self.config_file\n container = type(self.container)\n container_config = copy(self.container_config)\n proxy = str(type(self.proxy))\n return get_cube(cube=cube, init=init, name=name, config=config,\n config_file=config_file, container=container, container_config=\n container_config, proxy=proxy, proxy_config=self.proxy_config, **kwargs\n )\n",
"role": "user"
},
{
"content": "get_cube",
"role": "assistant"
}
] |
def x(aln, moltype, params={}, write_log=True):
new_aln = get_align_for_phylip(StringIO(aln))
aln2 = Alignment(new_aln)
seqs = aln2.toFasta()
ih = '_input_as_multiline_string'
pplacer_app = Pplacer(params=params, InputHandler=ih, WorkingDir=None,
SuppressStderr=False, SuppressStdout=False)
pplacer_result = pplacer_app(seqs)
if write_log:
log_fp = join(params['--out-dir'], 'log_pplacer_' + split(
get_tmp_filename())[-1])
log_file = open(log_fp, 'w')
log_file.write(pplacer_result['StdOut'].read())
log_file.close()
guppy_params = {'tog': None}
new_tree = build_tree_from_json_using_params(pplacer_result['json'].
name, output_dir=params['--out-dir'], params=guppy_params)
pplacer_result.cleanUp()
return new_tree
| insert_sequences_into_tree | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(aln, moltype, params={}, write_log=True):\n new_aln = get_align_for_phylip(StringIO(aln))\n aln2 = Alignment(new_aln)\n seqs = aln2.toFasta()\n ih = '_input_as_multiline_string'\n pplacer_app = Pplacer(params=params, InputHandler=ih, WorkingDir=None,\n SuppressStderr=False, SuppressStdout=False)\n pplacer_result = pplacer_app(seqs)\n if write_log:\n log_fp = join(params['--out-dir'], 'log_pplacer_' + split(\n get_tmp_filename())[-1])\n log_file = open(log_fp, 'w')\n log_file.write(pplacer_result['StdOut'].read())\n log_file.close()\n guppy_params = {'tog': None}\n new_tree = build_tree_from_json_using_params(pplacer_result['json'].\n name, output_dir=params['--out-dir'], params=guppy_params)\n pplacer_result.cleanUp()\n return new_tree\n",
"role": "user"
},
{
"content": "insert_sequences_into_tree",
"role": "assistant"
}
] |
def x(self, tmp_dir='/tmp', prefix='tmp', suffix='.fasta', include_class_id
=False, result_constructor=FilePath):
return super(Pplacer, self).getTmpFilename(tmp_dir=tmp_dir, prefix=
prefix, suffix=suffix, include_class_id=include_class_id,
result_constructor=result_constructor)
| getTmpFilename | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tmp_dir='/tmp', prefix='tmp', suffix='.fasta', include_class_id\n =False, result_constructor=FilePath):\n return super(Pplacer, self).getTmpFilename(tmp_dir=tmp_dir, prefix=\n prefix, suffix=suffix, include_class_id=include_class_id,\n result_constructor=result_constructor)\n",
"role": "user"
},
{
"content": "getTmpFilename",
"role": "assistant"
}
] |
def x(self, data):
output_dir = self.Parameters['--out-dir'].Value
result = {}
result['json'] = ResultPath(Path=join(output_dir, splitext(split(self.
_input_filename)[-1])[0] + '.jplace'))
return result
| get_result_paths | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n output_dir = self.Parameters['--out-dir'].Value\n result = {}\n result['json'] = ResultPath(Path=join(output_dir, splitext(split(self.\n _input_filename)[-1])[0] + '.jplace'))\n return result\n",
"role": "user"
},
{
"content": "get_result_paths",
"role": "assistant"
}
] |
def x(inst):
inst['iv_zon'] = {'data': inst['unit_zon_x'] * inst['iv_x'] + inst[
'unit_zon_y'] * inst['iv_y'] + inst['unit_zon_z'] * inst['iv_z'],
'units': 'm/s', 'long_name': 'Zonal ion velocity', 'notes':
'Ion velocity relative to co-rotation along zonal direction, normal to meridional plane. Positive east. Velocity obtained using ion velocities relative to co-rotation in the instrument frame along with the corresponding unit vectors expressed in the instrument frame. '
, 'label': 'Zonal Ion Velocity', 'axis': 'Zonal Ion Velocity',
'desc': 'Zonal ion velocity', 'scale': 'Linear', 'value_min': -
500.0, 'value_max': 500.0}
inst['iv_fa'] = {'data': inst['unit_fa_x'] * inst['iv_x'] + inst[
'unit_fa_y'] * inst['iv_y'] + inst['unit_fa_z'] * inst['iv_z'],
'units': 'm/s', 'long_name': 'Field-Aligned ion velocity', 'notes':
(
'Ion velocity relative to co-rotation along magnetic field line. Positive along the field. '
,
'Velocity obtained using ion velocities relative to co-rotation in the instrument frame along with the corresponding unit vectors expressed in the instrument frame. '
), 'label': 'Field-Aligned Ion Velocity', 'axis':
'Field-Aligned Ion Velocity', 'desc': 'Field-Aligned Ion Velocity',
'scale': 'Linear', 'value_min': -500.0, 'value_max': 500.0}
inst['iv_mer'] = {'data': inst['unit_mer_x'] * inst['iv_x'] + inst[
'unit_mer_y'] * inst['iv_y'] + inst['unit_mer_z'] * inst['iv_z'],
'units': 'm/s', 'long_name': 'Meridional ion velocity', 'notes': (
'Velocity along meridional direction, perpendicular to field and within meridional plane. Positive is up at magnetic equator. '
,
'Velocity obtained using ion velocities relative to co-rotation in the instrument frame along with the corresponding unit vectors expressed in the instrument frame. '
), 'label': 'Meridional Ion Velocity', 'axis':
'Meridional Ion Velocity', 'desc': 'Meridional Ion Velocity',
'scale': 'Linear', 'value_min': -500.0, 'value_max': 500.0}
return
| add_mag_drifts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inst):\n inst['iv_zon'] = {'data': inst['unit_zon_x'] * inst['iv_x'] + inst[\n 'unit_zon_y'] * inst['iv_y'] + inst['unit_zon_z'] * inst['iv_z'],\n 'units': 'm/s', 'long_name': 'Zonal ion velocity', 'notes':\n 'Ion velocity relative to co-rotation along zonal direction, normal to meridional plane. Positive east. Velocity obtained using ion velocities relative to co-rotation in the instrument frame along with the corresponding unit vectors expressed in the instrument frame. '\n , 'label': 'Zonal Ion Velocity', 'axis': 'Zonal Ion Velocity',\n 'desc': 'Zonal ion velocity', 'scale': 'Linear', 'value_min': -\n 500.0, 'value_max': 500.0}\n inst['iv_fa'] = {'data': inst['unit_fa_x'] * inst['iv_x'] + inst[\n 'unit_fa_y'] * inst['iv_y'] + inst['unit_fa_z'] * inst['iv_z'],\n 'units': 'm/s', 'long_name': 'Field-Aligned ion velocity', 'notes':\n (\n 'Ion velocity relative to co-rotation along magnetic field line. Positive along the field. '\n ,\n 'Velocity obtained using ion velocities relative to co-rotation in the instrument frame along with the corresponding unit vectors expressed in the instrument frame. '\n ), 'label': 'Field-Aligned Ion Velocity', 'axis':\n 'Field-Aligned Ion Velocity', 'desc': 'Field-Aligned Ion Velocity',\n 'scale': 'Linear', 'value_min': -500.0, 'value_max': 500.0}\n inst['iv_mer'] = {'data': inst['unit_mer_x'] * inst['iv_x'] + inst[\n 'unit_mer_y'] * inst['iv_y'] + inst['unit_mer_z'] * inst['iv_z'],\n 'units': 'm/s', 'long_name': 'Meridional ion velocity', 'notes': (\n 'Velocity along meridional direction, perpendicular to field and within meridional plane. Positive is up at magnetic equator. '\n ,\n 'Velocity obtained using ion velocities relative to co-rotation in the instrument frame along with the corresponding unit vectors expressed in the instrument frame. '\n ), 'label': 'Meridional Ion Velocity', 'axis':\n 'Meridional Ion Velocity', 'desc': 'Meridional Ion Velocity',\n 'scale': 'Linear', 'value_min': -500.0, 'value_max': 500.0}\n return\n",
"role": "user"
},
{
"content": "add_mag_drifts",
"role": "assistant"
}
] |
def x(*choices):
term = lambda expected, observed: float((expected - observed) ** 2) / max(
expected, 1)
mean_success_rate = float(sum([c.rewards for c in choices])) / max(sum(
[c.plays for c in choices]), 1)
mean_failure_rate = 1 - mean_success_rate
return sum([(term(mean_success_rate * c.plays, c.rewards) + term(
mean_failure_rate * c.plays, c.plays - c.rewards)) for c in choices])
| chi_squared | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(*choices):\n term = lambda expected, observed: float((expected - observed) ** 2) / max(\n expected, 1)\n mean_success_rate = float(sum([c.rewards for c in choices])) / max(sum(\n [c.plays for c in choices]), 1)\n mean_failure_rate = 1 - mean_success_rate\n return sum([(term(mean_success_rate * c.plays, c.rewards) + term(\n mean_failure_rate * c.plays, c.plays - c.rewards)) for c in choices])\n",
"role": "user"
},
{
"content": "chi_squared",
"role": "assistant"
}
] |
def x(redis, schema):
def x(active=True):
return dict((experiment.name, experiment) for experiment in
get_experiments(redis, active=active))
active_experiments = get_experiments_dict()
archived_experiments = get_experiments_dict(active=False)
new_experiment_names = set(schema.keys())
active_experiment_names = set(active_experiments.keys())
unarchivable_experiment_names = (new_experiment_names -
active_experiment_names & set(archived_experiments.keys()))
for unarchivable_experiment_name in unarchivable_experiment_names:
print('- De-archiving %s' % unarchivable_experiment_name)
pipe = redis.pipeline(transaction=True)
pipe.sadd(ACTIVE_EXPERIMENTS_REDIS_KEY, unarchivable_experiment_name)
pipe.srem(ARCHIVED_EXPERIMENTS_REDIS_KEY, unarchivable_experiment_name)
pipe.execute()
if unarchivable_experiment_names:
active_experiments = get_experiments_dict()
active_experiment_names = set(active_experiments.keys())
for new_experiment_name in (new_experiment_names - active_experiment_names
):
print('- Creating experiment %s' % new_experiment_name)
experiment = add_experiment(redis, new_experiment_name)
for choice in schema[new_experiment_name]:
print(' - Adding choice %s' % choice)
experiment.add_choice(choice)
for archivable_experiment_name in (active_experiment_names -
new_experiment_names):
print('- Archiving %s' % archivable_experiment_name)
active_experiments[archivable_experiment_name].archive()
for experiment_name in (new_experiment_names & active_experiment_names):
experiment = active_experiments[experiment_name]
new_choice_names = set(schema[experiment_name])
old_choice_names = set(experiment.choice_names)
for new_choice_name in (new_choice_names - old_choice_names):
print('- Adding choice %s to existing experiment %s' % (
new_choice_name, experiment_name))
experiment.add_choice(new_choice_name)
for removable_choice_name in (old_choice_names - new_choice_names):
print('- Removing choice %s from existing experiment %s' % (
removable_choice_name, experiment_name))
experiment.remove_choice(removable_choice_name)
| sync_from_spec | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(redis, schema):\n\n def x(active=True):\n return dict((experiment.name, experiment) for experiment in\n get_experiments(redis, active=active))\n active_experiments = get_experiments_dict()\n archived_experiments = get_experiments_dict(active=False)\n new_experiment_names = set(schema.keys())\n active_experiment_names = set(active_experiments.keys())\n unarchivable_experiment_names = (new_experiment_names -\n active_experiment_names & set(archived_experiments.keys()))\n for unarchivable_experiment_name in unarchivable_experiment_names:\n print('- De-archiving %s' % unarchivable_experiment_name)\n pipe = redis.pipeline(transaction=True)\n pipe.sadd(ACTIVE_EXPERIMENTS_REDIS_KEY, unarchivable_experiment_name)\n pipe.srem(ARCHIVED_EXPERIMENTS_REDIS_KEY, unarchivable_experiment_name)\n pipe.execute()\n if unarchivable_experiment_names:\n active_experiments = get_experiments_dict()\n active_experiment_names = set(active_experiments.keys())\n for new_experiment_name in (new_experiment_names - active_experiment_names\n ):\n print('- Creating experiment %s' % new_experiment_name)\n experiment = add_experiment(redis, new_experiment_name)\n for choice in schema[new_experiment_name]:\n print(' - Adding choice %s' % choice)\n experiment.add_choice(choice)\n for archivable_experiment_name in (active_experiment_names -\n new_experiment_names):\n print('- Archiving %s' % archivable_experiment_name)\n active_experiments[archivable_experiment_name].archive()\n for experiment_name in (new_experiment_names & active_experiment_names):\n experiment = active_experiments[experiment_name]\n new_choice_names = set(schema[experiment_name])\n old_choice_names = set(experiment.choice_names)\n for new_choice_name in (new_choice_names - old_choice_names):\n print('- Adding choice %s to existing experiment %s' % (\n new_choice_name, experiment_name))\n experiment.add_choice(new_choice_name)\n for removable_choice_name in (old_choice_names - new_choice_names):\n print('- Removing choice %s from existing experiment %s' % (\n removable_choice_name, experiment_name))\n experiment.remove_choice(removable_choice_name)\n",
"role": "user"
},
{
"content": "sync_from_spec",
"role": "assistant"
}
] |
def x(redis, name):
if not ALLOWED_NAMES.match(name):
raise ExperimentException(name, 'Illegal name')
if redis.exists(EXPERIMENT_REDIS_KEY_TEMPLATE % name):
raise ExperimentException(name, 'Already exists')
json = dict(creation_date=util.unicode_type(datetime.datetime.now()))
pipe = redis.pipeline(transaction=True)
pipe.sadd(ACTIVE_EXPERIMENTS_REDIS_KEY, name)
pipe.hset(EXPERIMENT_REDIS_KEY_TEMPLATE % name, 'metadata', escape.
json_encode(json))
pipe.execute()
return Experiment(redis, name)
| add_experiment | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(redis, name):\n if not ALLOWED_NAMES.match(name):\n raise ExperimentException(name, 'Illegal name')\n if redis.exists(EXPERIMENT_REDIS_KEY_TEMPLATE % name):\n raise ExperimentException(name, 'Already exists')\n json = dict(creation_date=util.unicode_type(datetime.datetime.now()))\n pipe = redis.pipeline(transaction=True)\n pipe.sadd(ACTIVE_EXPERIMENTS_REDIS_KEY, name)\n pipe.hset(EXPERIMENT_REDIS_KEY_TEMPLATE % name, 'metadata', escape.\n json_encode(json))\n pipe.execute()\n return Experiment(redis, name)\n",
"role": "user"
},
{
"content": "add_experiment",
"role": "assistant"
}
] |
def x(redis, active=True):
key = (ACTIVE_EXPERIMENTS_REDIS_KEY if active else
ARCHIVED_EXPERIMENTS_REDIS_KEY)
return [Experiment(redis, escape.to_unicode(name)) for name in redis.
smembers(key)]
| get_experiments | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(redis, active=True):\n key = (ACTIVE_EXPERIMENTS_REDIS_KEY if active else\n ARCHIVED_EXPERIMENTS_REDIS_KEY)\n return [Experiment(redis, escape.to_unicode(name)) for name in redis.\n smembers(key)]\n",
"role": "user"
},
{
"content": "get_experiments",
"role": "assistant"
}
] |
def x(self):
pipe = self.redis.pipeline()
pipe.hget(EXPERIMENT_REDIS_KEY_TEMPLATE % self.name, 'metadata')
pipe.hget(EXPERIMENT_REDIS_KEY_TEMPLATE % self.name, 'choices')
pipe.hget(EXPERIMENT_REDIS_KEY_TEMPLATE % self.name, 'default-choice')
results = pipe.execute()
if results[0] == None:
raise ExperimentException(self.name, 'Does not exist')
self.metadata = parse_json(results[0])
self.choice_names = parse_json(results[1]) if results[1] != None else []
self.default_choice = escape.to_unicode(results[2])
self._choices = None
| refresh | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n pipe = self.redis.pipeline()\n pipe.hget(EXPERIMENT_REDIS_KEY_TEMPLATE % self.name, 'metadata')\n pipe.hget(EXPERIMENT_REDIS_KEY_TEMPLATE % self.name, 'choices')\n pipe.hget(EXPERIMENT_REDIS_KEY_TEMPLATE % self.name, 'default-choice')\n results = pipe.execute()\n if results[0] == None:\n raise ExperimentException(self.name, 'Does not exist')\n self.metadata = parse_json(results[0])\n self.choice_names = parse_json(results[1]) if results[1] != None else []\n self.default_choice = escape.to_unicode(results[2])\n self._choices = None\n",
"role": "user"
},
{
"content": "refresh",
"role": "assistant"
}
] |
def x(self):
if self._choices == None:
self._choices = [ExperimentChoice(self, choice_name) for
choice_name in self.choice_names]
return self._choices
| choices | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._choices == None:\n self._choices = [ExperimentChoice(self, choice_name) for\n choice_name in self.choice_names]\n return self._choices\n",
"role": "user"
},
{
"content": "choices",
"role": "assistant"
}
] |
def x(self):
choices = self.choices
if len(choices) >= 2:
csq = chi_squared(*choices)
confident = is_confident(csq, len(choices)) if len(choices
) <= 10 else None
else:
csq = None
confident = False
return csq, confident
| confidence | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n choices = self.choices\n if len(choices) >= 2:\n csq = chi_squared(*choices)\n confident = is_confident(csq, len(choices)) if len(choices\n ) <= 10 else None\n else:\n csq = None\n confident = False\n return csq, confident\n",
"role": "user"
},
{
"content": "confidence",
"role": "assistant"
}
] |
def x(self):
pipe = self.redis.pipeline(transaction=True)
pipe.srem(ACTIVE_EXPERIMENTS_REDIS_KEY, self.name)
pipe.sadd(ARCHIVED_EXPERIMENTS_REDIS_KEY, self.name)
pipe.execute()
| archive | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n pipe = self.redis.pipeline(transaction=True)\n pipe.srem(ACTIVE_EXPERIMENTS_REDIS_KEY, self.name)\n pipe.sadd(ARCHIVED_EXPERIMENTS_REDIS_KEY, self.name)\n pipe.execute()\n",
"role": "user"
},
{
"content": "archive",
"role": "assistant"
}
] |
def x(self, choice_name):
if not ALLOWED_NAMES.match(choice_name):
raise ExperimentException(self.name, 'Illegal choice name: %s' %
choice_name)
if choice_name in self.choice_names:
raise ExperimentException(self.name, 'Choice already exists: %s' %
choice_name)
self.choice_names.append(choice_name)
self.redis.hset(EXPERIMENT_REDIS_KEY_TEMPLATE % self.name, 'choices',
escape.json_encode(self.choice_names))
self.refresh()
| add_choice | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, choice_name):\n if not ALLOWED_NAMES.match(choice_name):\n raise ExperimentException(self.name, 'Illegal choice name: %s' %\n choice_name)\n if choice_name in self.choice_names:\n raise ExperimentException(self.name, 'Choice already exists: %s' %\n choice_name)\n self.choice_names.append(choice_name)\n self.redis.hset(EXPERIMENT_REDIS_KEY_TEMPLATE % self.name, 'choices',\n escape.json_encode(self.choice_names))\n self.refresh()\n",
"role": "user"
},
{
"content": "add_choice",
"role": "assistant"
}
] |
def x(self, choice_name):
self.choice_names.remove(choice_name)
self.redis.hset(EXPERIMENT_REDIS_KEY_TEMPLATE % self.name, 'choices',
escape.json_encode(self.choice_names))
self.refresh()
| remove_choice | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, choice_name):\n self.choice_names.remove(choice_name)\n self.redis.hset(EXPERIMENT_REDIS_KEY_TEMPLATE % self.name, 'choices',\n escape.json_encode(self.choice_names))\n self.refresh()\n",
"role": "user"
},
{
"content": "remove_choice",
"role": "assistant"
}
] |
def x(self, choice, count=1):
self.redis.hincrby(EXPERIMENT_REDIS_KEY_TEMPLATE % self.name,
'%s:plays' % choice, count)
self._choices = None
| add_play | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, choice, count=1):\n self.redis.hincrby(EXPERIMENT_REDIS_KEY_TEMPLATE % self.name, \n '%s:plays' % choice, count)\n self._choices = None\n",
"role": "user"
},
{
"content": "add_play",
"role": "assistant"
}
] |
def x(self):
choices = self.choices
if len(choices) == 0:
return None
high_choice = max(choices, key=lambda choice: choice.performance)
self.redis.hset(EXPERIMENT_REDIS_KEY_TEMPLATE % self.name,
'default-choice', high_choice.name)
self.refresh()
return high_choice
| compute_default_choice | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n choices = self.choices\n if len(choices) == 0:\n return None\n high_choice = max(choices, key=lambda choice: choice.performance)\n self.redis.hset(EXPERIMENT_REDIS_KEY_TEMPLATE % self.name,\n 'default-choice', high_choice.name)\n self.refresh()\n return high_choice\n",
"role": "user"
},
{
"content": "compute_default_choice",
"role": "assistant"
}
] |
def x(self):
redis_key = EXPERIMENT_REDIS_KEY_TEMPLATE % self.experiment.name
self.plays = int(self.experiment.redis.hget(redis_key, '%s:plays' %
self.name) or 0)
self.rewards = int(self.experiment.redis.hget(redis_key, '%s:rewards' %
self.name) or 0)
self.performance = float(self.rewards) / max(self.plays, 1)
| refresh | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n redis_key = EXPERIMENT_REDIS_KEY_TEMPLATE % self.experiment.name\n self.plays = int(self.experiment.redis.hget(redis_key, '%s:plays' %\n self.name) or 0)\n self.rewards = int(self.experiment.redis.hget(redis_key, '%s:rewards' %\n self.name) or 0)\n self.performance = float(self.rewards) / max(self.plays, 1)\n",
"role": "user"
},
{
"content": "refresh",
"role": "assistant"
}
] |
def x(self, pageNum, itemsPerPage):
return self.get_all_alerts(self.status, pageNum, itemsPerPage)
| fetch | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, pageNum, itemsPerPage):\n return self.get_all_alerts(self.status, pageNum, itemsPerPage)\n",
"role": "user"
},
{
"content": "fetch",
"role": "assistant"
}
] |
async def await_all():
"""
Simple utility function that drains all pending tasks
"""
tasks = asyncio.Task.all_tasks()
for task in tasks:
try:
await task
except RuntimeError as e:
if 'Task cannot await on itself' not in str(e):
raise e
except AssertionError as e:
if "yield from wasn't used with future" not in str(e):
raise e
| await_all | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def await_all():\n \"\"\"\n Simple utility function that drains all pending tasks\n \"\"\"\n tasks = asyncio.Task.all_tasks()\n for task in tasks:\n try:\n await task\n except RuntimeError as e:\n if 'Task cannot await on itself' not in str(e):\n raise e\n except AssertionError as e:\n if \"yield from wasn't used with future\" not in str(e):\n raise e\n",
"role": "user"
},
{
"content": "await_all",
"role": "assistant"
}
] |
def x(func):
if inspect.iscoroutinefunction(func):
@functools.wraps(func)
def x(*args, **kwargs):
loop = asyncio.get_event_loop()
try:
loop.run_until_complete(func(*args, **kwargs))
finally:
loop.close()
return sync_wrapper
return func
| coerce_to_synchronous | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(func):\n if inspect.iscoroutinefunction(func):\n\n @functools.wraps(func)\n def x(*args, **kwargs):\n loop = asyncio.get_event_loop()\n try:\n loop.run_until_complete(func(*args, **kwargs))\n finally:\n loop.close()\n return sync_wrapper\n return func\n",
"role": "user"
},
{
"content": "coerce_to_synchronous",
"role": "assistant"
}
] |
def x(a, b, segmenter=None):
a, b = list(a), list(b)
segmenter = segmenter or SEGMENTER
a_segments = segmenter.segment(a)
b_segments = segmenter.segment(b)
return diff_segments(a_segments, b_segments)
| diff | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(a, b, segmenter=None):\n a, b = list(a), list(b)\n segmenter = segmenter or SEGMENTER\n a_segments = segmenter.segment(a)\n b_segments = segmenter.segment(b)\n return diff_segments(a_segments, b_segments)\n",
"role": "user"
},
{
"content": "diff",
"role": "assistant"
}
] |
def x(a_segments, b_segments):
a_segment_tokens, b_segment_tokens = _cluster_matching_segments(a_segments,
b_segments)
clustered_ops = sequence_matcher.diff(a_segment_tokens, b_segment_tokens)
return (op for op in SegmentOperationsExpander(clustered_ops,
a_segment_tokens, b_segment_tokens).expand())
| diff_segments | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(a_segments, b_segments):\n a_segment_tokens, b_segment_tokens = _cluster_matching_segments(a_segments,\n b_segments)\n clustered_ops = sequence_matcher.diff(a_segment_tokens, b_segment_tokens)\n return (op for op in SegmentOperationsExpander(clustered_ops,\n a_segment_tokens, b_segment_tokens).expand())\n",
"role": "user"
},
{
"content": "diff_segments",
"role": "assistant"
}
] |
def x(texts, *args, **kwargs):
processor = SegmentMatcher.Processor(*args, **kwargs)
for text in texts:
yield processor.process(text)
| process | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(texts, *args, **kwargs):\n processor = SegmentMatcher.Processor(*args, **kwargs)\n for text in texts:\n yield processor.process(text)\n",
"role": "user"
},
{
"content": "process",
"role": "assistant"
}
] |
def x(segments):
for subsegment in segments:
if isinstance(subsegment, Token):
break
if isinstance(subsegment, Segment):
if isinstance(subsegment, MatchableSegment):
yield subsegment
for matchable_subsegment in _get_matchable_segments(subsegment):
yield matchable_subsegment
| get_matchable_segments | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(segments):\n for subsegment in segments:\n if isinstance(subsegment, Token):\n break\n if isinstance(subsegment, Segment):\n if isinstance(subsegment, MatchableSegment):\n yield subsegment\n for matchable_subsegment in _get_matchable_segments(subsegment):\n yield matchable_subsegment\n",
"role": "user"
},
{
"content": "get_matchable_segments",
"role": "assistant"
}
] |
def x(self, orm):
for document in orm['document_library.Document'].objects.all():
self.migrate_placeholder(orm, document, 'content',
'document_library_content', 'content')
| forwards | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, orm):\n for document in orm['document_library.Document'].objects.all():\n self.migrate_placeholder(orm, document, 'content',\n 'document_library_content', 'content')\n",
"role": "user"
},
{
"content": "forwards",
"role": "assistant"
}
] |
def x(predictions='/dev/stdout', quiet=True, save_resume=True, q_colon=None,
**kwargs):
args = ['vw']
if q_colon:
kwargs['q:'] = q_colon
kwargs['predictions'] = predictions
kwargs['quiet'] = quiet
kwargs['save_resume'] = save_resume
for key, value in kwargs.items():
if len(key) == 1:
option = '-{}'.format(key)
else:
option = '--{}'.format(key)
if value is True:
arg_list = [option]
elif isinstance(value, basestring):
arg_list = ['{} {}'.format(option, value)]
elif hasattr(value, '__getitem__'):
arg_list = ['{} {}'.format(option, subvalue) for subvalue in value]
else:
arg_list = ['{} {}'.format(option, value)]
args.extend(arg_list)
command = ' '.join(args)
return command
| make_command_line | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(predictions='/dev/stdout', quiet=True, save_resume=True, q_colon=None,\n **kwargs):\n args = ['vw']\n if q_colon:\n kwargs['q:'] = q_colon\n kwargs['predictions'] = predictions\n kwargs['quiet'] = quiet\n kwargs['save_resume'] = save_resume\n for key, value in kwargs.items():\n if len(key) == 1:\n option = '-{}'.format(key)\n else:\n option = '--{}'.format(key)\n if value is True:\n arg_list = [option]\n elif isinstance(value, basestring):\n arg_list = ['{} {}'.format(option, value)]\n elif hasattr(value, '__getitem__'):\n arg_list = ['{} {}'.format(option, subvalue) for subvalue in value]\n else:\n arg_list = ['{} {}'.format(option, value)]\n args.extend(arg_list)\n command = ' '.join(args)\n return command\n",
"role": "user"
},
{
"content": "make_command_line",
"role": "assistant"
}
] |
def x(self, features):
for feature in features:
if isinstance(feature, basestring):
label = feature
value = None
else:
label, value = feature
self.add_feature(label, value)
| add_features | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, features):\n for feature in features:\n if isinstance(feature, basestring):\n label = feature\n value = None\n else:\n label, value = feature\n self.add_feature(label, value)\n",
"role": "user"
},
{
"content": "add_features",
"role": "assistant"
}
] |
def x(self, label, value=None):
if self.escape:
label = escape_vw_string(label)
elif self.validate:
validate_vw_string(label)
feature = label, value
self.features.append(feature)
| add_feature | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, label, value=None):\n if self.escape:\n label = escape_vw_string(label)\n elif self.validate:\n validate_vw_string(label)\n feature = label, value\n self.features.append(feature)\n",
"role": "user"
},
{
"content": "add_feature",
"role": "assistant"
}
] |
def x(self):
if self._string is None:
tokens = []
if self.name:
if self.scale:
token = self.name + ':' + str(self.scale)
else:
token = self.name
else:
token = ''
tokens.append(token)
for label, value in self.features:
if value is None:
token = label
else:
token = label + ':' + str(value)
tokens.append(token)
tokens.append('')
output = ' '.join(tokens)
if self.cache_string:
self._string = output
else:
output = self._string
return output
| to_string | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._string is None:\n tokens = []\n if self.name:\n if self.scale:\n token = self.name + ':' + str(self.scale)\n else:\n token = self.name\n else:\n token = ''\n tokens.append(token)\n for label, value in self.features:\n if value is None:\n token = label\n else:\n token = label + ':' + str(value)\n tokens.append(token)\n tokens.append('')\n output = ' '.join(tokens)\n if self.cache_string:\n self._string = output\n else:\n output = self._string\n return output\n",
"role": "user"
},
{
"content": "to_string",
"role": "assistant"
}
] |
def x(self, delimiter='\\'):
result_list = []
for feature in self.features:
result = '{}{}{}'.format(self.name, delimiter, feature)
result_list.append(result)
return result_list
| export_features | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, delimiter='\\\\'):\n result_list = []\n for feature in self.features:\n result = '{}{}{}'.format(self.name, delimiter, feature)\n result_list.append(result)\n return result_list\n",
"role": "user"
},
{
"content": "export_features",
"role": "assistant"
}
] |
def x(self, line, parse_result=True):
self.vw_process.sendline(line)
result = self._get_response(parse_result=parse_result)
return result
| send_line | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, line, parse_result=True):\n self.vw_process.sendline(line)\n result = self._get_response(parse_result=parse_result)\n return result\n",
"role": "user"
},
{
"content": "send_line",
"role": "assistant"
}
] |
def x(self, parse_result=True):
self.vw_process.expect_exact('\r\n', searchwindowsize=-1)
if parse_result:
output = self.vw_process.before
result_struct = VWResult(output, active_mode=self.active_mode)
else:
result_struct = None
return result_struct
| get_response | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, parse_result=True):\n self.vw_process.expect_exact('\\r\\n', searchwindowsize=-1)\n if parse_result:\n output = self.vw_process.before\n result_struct = VWResult(output, active_mode=self.active_mode)\n else:\n result_struct = None\n return result_struct\n",
"role": "user"
},
{
"content": "get_response",
"role": "assistant"
}
] |
def x(self, *args, **kwargs):
parse_result = kwargs.pop('parse_result', True)
line = self.make_line(*args, **kwargs)
result = self.send_line(line, parse_result=parse_result)
return result
| send_example | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *args, **kwargs):\n parse_result = kwargs.pop('parse_result', True)\n line = self.make_line(*args, **kwargs)\n result = self.send_line(line, parse_result=parse_result)\n return result\n",
"role": "user"
},
{
"content": "send_example",
"role": "assistant"
}
] |
def x(self, response=None, importance=None, base=None, tag=None, features=
None, namespaces=None):
if namespaces is not None:
self.add_namespaces(namespaces)
if features is not None:
namespace = Namespace(features=features)
self.add_namespace(namespace)
substrings = []
tokens = []
if response is not None:
token = str(response)
tokens.append(token)
if importance is not None:
token = str(importance)
tokens.append(token)
if base is not None:
token = str(base)
tokens.append(token)
if tag is not None:
token = "'" + str(tag)
tokens.append(token)
else:
token = ''
tokens.append(token)
substring = ' '.join(tokens)
substrings.append(substring)
if self.namespaces:
for namespace in self.namespaces:
substring = namespace.to_string()
substrings.append(substring)
else:
substrings.append('')
line = '|'.join(substrings)
self._line = line
self.namespaces = []
return line
| make_line | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, response=None, importance=None, base=None, tag=None, features=\n None, namespaces=None):\n if namespaces is not None:\n self.add_namespaces(namespaces)\n if features is not None:\n namespace = Namespace(features=features)\n self.add_namespace(namespace)\n substrings = []\n tokens = []\n if response is not None:\n token = str(response)\n tokens.append(token)\n if importance is not None:\n token = str(importance)\n tokens.append(token)\n if base is not None:\n token = str(base)\n tokens.append(token)\n if tag is not None:\n token = \"'\" + str(tag)\n tokens.append(token)\n else:\n token = ''\n tokens.append(token)\n substring = ' '.join(tokens)\n substrings.append(substring)\n if self.namespaces:\n for namespace in self.namespaces:\n substring = namespace.to_string()\n substrings.append(substring)\n else:\n substrings.append('')\n line = '|'.join(substrings)\n self._line = line\n self.namespaces = []\n return line\n",
"role": "user"
},
{
"content": "make_line",
"role": "assistant"
}
] |
def x(self, *args, **kwargs):
if args and isinstance(args[0], Namespace):
namespace = args[0]
elif isinstance(kwargs.get('namespace'), Namespace):
namespace = kwargs.get('namespace')
else:
namespace = Namespace(*args, **kwargs)
self.namespaces.append(namespace)
return self
| add_namespace | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *args, **kwargs):\n if args and isinstance(args[0], Namespace):\n namespace = args[0]\n elif isinstance(kwargs.get('namespace'), Namespace):\n namespace = kwargs.get('namespace')\n else:\n namespace = Namespace(*args, **kwargs)\n self.namespaces.append(namespace)\n return self\n",
"role": "user"
},
{
"content": "add_namespace",
"role": "assistant"
}
] |
def x(self, features=None, tag=None, namespaces=None):
if features is not None:
namespace = Namespace(features=features)
self.add_namespace(namespace)
result = self.send_example(tag=tag, namespaces=namespaces)
return result
| get_prediction | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, features=None, tag=None, namespaces=None):\n if features is not None:\n namespace = Namespace(features=features)\n self.add_namespace(namespace)\n result = self.send_example(tag=tag, namespaces=namespaces)\n return result\n",
"role": "user"
},
{
"content": "get_prediction",
"role": "assistant"
}
] |
def x(self, model_filename):
line = 'save_{}|'.format(model_filename)
self.vw_process.sendline(line)
| save_model | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, model_filename):\n line = 'save_{}|'.format(model_filename)\n self.vw_process.sendline(line)\n",
"role": "user"
},
{
"content": "save_model",
"role": "assistant"
}
] |
def x(self, path):
path = path.strip('/')
name, _, subpath = path.partition('/')
for service in singletons.settings.load_all('SERVICES'):
if service.SERVICE_NAME == name:
break
else:
return [], None
for partial_url, view in service.urls.items():
partial_url = partial_url.strip('/')
if isinstance(view, str):
view = getattr(service, view)
regexp = re.sub('<[^>]+>', '([^/]+)', partial_url)
matches = re.findall('^%s$' % regexp, subpath)
if matches:
if '(' not in regexp:
matches = []
return matches, view
return [], None
| route_path | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, path):\n path = path.strip('/')\n name, _, subpath = path.partition('/')\n for service in singletons.settings.load_all('SERVICES'):\n if service.SERVICE_NAME == name:\n break\n else:\n return [], None\n for partial_url, view in service.urls.items():\n partial_url = partial_url.strip('/')\n if isinstance(view, str):\n view = getattr(service, view)\n regexp = re.sub('<[^>]+>', '([^/]+)', partial_url)\n matches = re.findall('^%s$' % regexp, subpath)\n if matches:\n if '(' not in regexp:\n matches = []\n return matches, view\n return [], None\n",
"role": "user"
},
{
"content": "route_path",
"role": "assistant"
}
] |
def x(iterator):
left = START
for item in iterator:
if left is not START:
yield left, item
left = item
| pair_looper | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(iterator):\n left = START\n for item in iterator:\n if left is not START:\n yield left, item\n left = item\n",
"role": "user"
},
{
"content": "pair_looper",
"role": "assistant"
}
] |
def x(data, batch_size):
data = data or []
for i in range(0, len(data), batch_size):
yield data[i:i + batch_size]
| batch_gen | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(data, batch_size):\n data = data or []\n for i in range(0, len(data), batch_size):\n yield data[i:i + batch_size]\n",
"role": "user"
},
{
"content": "batch_gen",
"role": "assistant"
}
] |
def x(pids, pid_dir='/tmp', prefix='', multi=False):
if isinstance(pids, (int, float, long)):
pids = [pids]
pids = str2list(pids, map_=unicode)
procs = map(unicode, os.listdir('/proc'))
running = [pid for pid in pids if pid in procs]
logger.warn('Found %s pids running: %s' % (len(running), running))
prefix = prefix.rstrip('.') if prefix else None
for pid in pids:
if prefix:
_prefix = prefix
else:
_prefix = unicode(pid)
if pid in running:
continue
if multi:
pid_file = '%s%s.pid' % (_prefix, pid)
else:
pid_file = '%s.pid' % _prefix
path = os.path.join(pid_dir, pid_file)
if os.path.exists(path):
logger.debug('Removing pidfile: %s' % path)
try:
remove_file(path)
except OSError as e:
logger.warn(e)
return running
| clear_stale_pids | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(pids, pid_dir='/tmp', prefix='', multi=False):\n if isinstance(pids, (int, float, long)):\n pids = [pids]\n pids = str2list(pids, map_=unicode)\n procs = map(unicode, os.listdir('/proc'))\n running = [pid for pid in pids if pid in procs]\n logger.warn('Found %s pids running: %s' % (len(running), running))\n prefix = prefix.rstrip('.') if prefix else None\n for pid in pids:\n if prefix:\n _prefix = prefix\n else:\n _prefix = unicode(pid)\n if pid in running:\n continue\n if multi:\n pid_file = '%s%s.pid' % (_prefix, pid)\n else:\n pid_file = '%s.pid' % _prefix\n path = os.path.join(pid_dir, pid_file)\n if os.path.exists(path):\n logger.debug('Removing pidfile: %s' % path)\n try:\n remove_file(path)\n except OSError as e:\n logger.warn(e)\n return running\n",
"role": "user"
},
{
"content": "clear_stale_pids",
"role": "assistant"
}
] |
def x(pid_file=None, cwd=None):
cwd = cwd or '/'
try:
pid = os.fork()
except OSError as e:
raise Exception('%s [%d]' % (e.strerror, e.errno))
if pid == 0:
os.setsid()
try:
pid = os.fork()
except OSError as e:
raise Exception('%s [%d]' % (e.strerror, e.errno))
if pid == 0:
os.chdir(cwd)
os.umask(0)
else:
os._exit(0)
else:
os._exit(0)
maxfd = resource.getrlimit(resource.RLIMIT_NOFILE)[1]
if maxfd == resource.RLIM_INFINITY:
maxfd = 1024
for fd in range(0, maxfd):
try:
os.close(fd)
except OSError:
pass
os.open('/dev/null', os.O_RDWR)
os.dup2(0, 1)
os.dup2(0, 2)
pid_file = pid_file or '%s.pid' % os.getpid()
write_file(pid_file, os.getpid())
return 0
| daemonize | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(pid_file=None, cwd=None):\n cwd = cwd or '/'\n try:\n pid = os.fork()\n except OSError as e:\n raise Exception('%s [%d]' % (e.strerror, e.errno))\n if pid == 0:\n os.setsid()\n try:\n pid = os.fork()\n except OSError as e:\n raise Exception('%s [%d]' % (e.strerror, e.errno))\n if pid == 0:\n os.chdir(cwd)\n os.umask(0)\n else:\n os._exit(0)\n else:\n os._exit(0)\n maxfd = resource.getrlimit(resource.RLIMIT_NOFILE)[1]\n if maxfd == resource.RLIM_INFINITY:\n maxfd = 1024\n for fd in range(0, maxfd):\n try:\n os.close(fd)\n except OSError:\n pass\n os.open('/dev/null', os.O_RDWR)\n os.dup2(0, 1)\n os.dup2(0, 2)\n pid_file = pid_file or '%s.pid' % os.getpid()\n write_file(pid_file, os.getpid())\n return 0\n",
"role": "user"
},
{
"content": "daemonize",
"role": "assistant"
}
] |
def x(cube):
_cube = cube.split('_')
pkg = _cube[0]
mod = '_'.join(_cube[1:])
_cls = ''.join([(s[0].upper() + s[1:]) for s in _cube[1:]])
return pkg, mod, _cls
| cube_pkg_mod_cls | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cube):\n _cube = cube.split('_')\n pkg = _cube[0]\n mod = '_'.join(_cube[1:])\n _cls = ''.join([(s[0].upper() + s[1:]) for s in _cube[1:]])\n return pkg, mod, _cls\n",
"role": "user"
},
{
"content": "cube_pkg_mod_cls",
"role": "assistant"
}
] |
def x(logger=None, level=None, log2file=None, log_file=None, log_format=
None, log_dir=None, log2stdout=None, truncate=False):
log2stdout = False if log2stdout is None else log2stdout
_log_format = '%(levelname)s.%(name)s.%(process)s:%(asctime)s:%(message)s'
log_format = log_format or _log_format
if isinstance(log_format, basestring):
log_format = logging.Formatter(log_format, '%Y%m%dT%H%M%S')
log2file = True if log2file is None else log2file
logger = logger or 'metrique'
if isinstance(logger, basestring):
logger = logging.getLogger(logger)
else:
logger = logger or logging.getLogger(logger)
logger.propagate = 0
logger.handlers = []
if log2file:
log_dir = log_dir or LOGS_DIR
log_file = log_file or 'metrique'
log_file = os.path.join(log_dir, log_file)
if truncate:
open(log_file, 'w+').close()
hdlr = logging.FileHandler(log_file)
hdlr.setFormatter(log_format)
logger.addHandler(hdlr)
else:
log2stdout = True
if log2stdout:
hdlr = logging.StreamHandler()
hdlr.setFormatter(log_format)
logger.addHandler(hdlr)
logger = _debug_set_level(logger, level)
return logger
| debug_setup | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(logger=None, level=None, log2file=None, log_file=None, log_format=\n None, log_dir=None, log2stdout=None, truncate=False):\n log2stdout = False if log2stdout is None else log2stdout\n _log_format = '%(levelname)s.%(name)s.%(process)s:%(asctime)s:%(message)s'\n log_format = log_format or _log_format\n if isinstance(log_format, basestring):\n log_format = logging.Formatter(log_format, '%Y%m%dT%H%M%S')\n log2file = True if log2file is None else log2file\n logger = logger or 'metrique'\n if isinstance(logger, basestring):\n logger = logging.getLogger(logger)\n else:\n logger = logger or logging.getLogger(logger)\n logger.propagate = 0\n logger.handlers = []\n if log2file:\n log_dir = log_dir or LOGS_DIR\n log_file = log_file or 'metrique'\n log_file = os.path.join(log_dir, log_file)\n if truncate:\n open(log_file, 'w+').close()\n hdlr = logging.FileHandler(log_file)\n hdlr.setFormatter(log_format)\n logger.addHandler(hdlr)\n else:\n log2stdout = True\n if log2stdout:\n hdlr = logging.StreamHandler()\n hdlr.setFormatter(log_format)\n logger.addHandler(hdlr)\n logger = _debug_set_level(logger, level)\n return logger\n",
"role": "user"
},
{
"content": "debug_setup",
"role": "assistant"
}
] |
def x(dt, drop_micro=False):
is_true(HAS_DATEUTIL, '`pip install python_dateutil` required')
if is_empty(dt, except_=False):
ts = None
elif isinstance(dt, (int, long, float)):
ts = float(dt)
elif isinstance(dt, basestring):
try:
parsed_dt = float(dt)
except (TypeError, ValueError):
parsed_dt = dt_parse(dt)
ts = dt2ts(parsed_dt)
else:
assert isinstance(dt, (datetime, date))
ts = (timegm(dt.timetuple()) * 1000.0 + dt.microsecond / 1000.0
) / 1000.0
if ts is None:
pass
elif drop_micro:
ts = float(int(ts))
else:
ts = float(ts)
return ts
| dt2ts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dt, drop_micro=False):\n is_true(HAS_DATEUTIL, '`pip install python_dateutil` required')\n if is_empty(dt, except_=False):\n ts = None\n elif isinstance(dt, (int, long, float)):\n ts = float(dt)\n elif isinstance(dt, basestring):\n try:\n parsed_dt = float(dt)\n except (TypeError, ValueError):\n parsed_dt = dt_parse(dt)\n ts = dt2ts(parsed_dt)\n else:\n assert isinstance(dt, (datetime, date))\n ts = (timegm(dt.timetuple()) * 1000.0 + dt.microsecond / 1000.0\n ) / 1000.0\n if ts is None:\n pass\n elif drop_micro:\n ts = float(int(ts))\n else:\n ts = float(ts)\n return ts\n",
"role": "user"
},
{
"content": "dt2ts",
"role": "assistant"
}
] |
def x(cube, init=False, pkgs=None, cube_paths=None, config=None, backends=
None, **kwargs):
pkgs = pkgs or ['cubes']
pkgs = [pkgs] if isinstance(pkgs, basestring) else pkgs
cube_paths = cube_paths or []
cube_paths_is_basestring = isinstance(cube_paths, basestring)
cube_paths = [cube_paths] if cube_paths_is_basestring else cube_paths
cube_paths = [os.path.expanduser(path) for path in cube_paths]
[sys.path.append(path) for path in cube_paths if path not in sys.path]
pkgs = pkgs + DEFAULT_PKGS
err = False
for pkg in pkgs:
try:
_cube = _load_cube_pkg(pkg, cube)
except ImportError as err:
_cube = None
if _cube:
break
else:
logger.error(err)
raise RuntimeError('"%s" not found! %s; %s \n%s)' % (cube, pkgs,
cube_paths, sys.path))
if init:
_cube = _cube(config=config, **kwargs)
return _cube
| get_cube | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cube, init=False, pkgs=None, cube_paths=None, config=None, backends=\n None, **kwargs):\n pkgs = pkgs or ['cubes']\n pkgs = [pkgs] if isinstance(pkgs, basestring) else pkgs\n cube_paths = cube_paths or []\n cube_paths_is_basestring = isinstance(cube_paths, basestring)\n cube_paths = [cube_paths] if cube_paths_is_basestring else cube_paths\n cube_paths = [os.path.expanduser(path) for path in cube_paths]\n [sys.path.append(path) for path in cube_paths if path not in sys.path]\n pkgs = pkgs + DEFAULT_PKGS\n err = False\n for pkg in pkgs:\n try:\n _cube = _load_cube_pkg(pkg, cube)\n except ImportError as err:\n _cube = None\n if _cube:\n break\n else:\n logger.error(err)\n raise RuntimeError('\"%s\" not found! %s; %s \\n%s)' % (cube, pkgs,\n cube_paths, sys.path))\n if init:\n _cube = _cube(config=config, **kwargs)\n return _cube\n",
"role": "user"
},
{
"content": "get_cube",
"role": "assistant"
}
] |
def x(from_timezone, to_tz=None, tz_aware=False):
if not from_timezone:
return None
is_true(HAS_DATEUTIL, '`pip install python_dateutil` required')
is_true(HAS_PYTZ, '`pip install pytz` required')
from_tz = pytz.timezone(from_timezone)
return partial(_get_timezone_converter, from_tz=from_tz, to_tz=to_tz,
tz_aware=tz_aware)
| get_timezone_converter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(from_timezone, to_tz=None, tz_aware=False):\n if not from_timezone:\n return None\n is_true(HAS_DATEUTIL, '`pip install python_dateutil` required')\n is_true(HAS_PYTZ, '`pip install pytz` required')\n from_tz = pytz.timezone(from_timezone)\n return partial(_get_timezone_converter, from_tz=from_tz, to_tz=to_tz,\n tz_aware=tz_aware)\n",
"role": "user"
},
{
"content": "get_timezone_converter",
"role": "assistant"
}
] |
def x(uri, pull=True, reflect=False, cache_dir=None, chdir=True):
cache_dir = cache_dir or CACHE_DIR
repo_path = os.path.expanduser(os.path.join(cache_dir, safestr(uri)))
if not os.path.exists(repo_path):
from_cache = False
logger.info('Locally caching git repo [%s] to [%s]' % (uri, repo_path))
cmd = 'git clone %s %s' % (uri, repo_path)
sys_call(cmd)
else:
from_cache = True
logger.info('GIT repo loaded from local cache [%s])' % repo_path)
if pull and not from_cache:
os.chdir(repo_path)
cmd = 'git pull'
sys_call(cmd, cwd=repo_path)
if chdir:
os.chdir(repo_path)
if reflect:
if not HAS_DULWICH:
raise RuntimeError('`pip install dulwich` required!')
return Repo(repo_path)
else:
return repo_path
| git_clone | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(uri, pull=True, reflect=False, cache_dir=None, chdir=True):\n cache_dir = cache_dir or CACHE_DIR\n repo_path = os.path.expanduser(os.path.join(cache_dir, safestr(uri)))\n if not os.path.exists(repo_path):\n from_cache = False\n logger.info('Locally caching git repo [%s] to [%s]' % (uri, repo_path))\n cmd = 'git clone %s %s' % (uri, repo_path)\n sys_call(cmd)\n else:\n from_cache = True\n logger.info('GIT repo loaded from local cache [%s])' % repo_path)\n if pull and not from_cache:\n os.chdir(repo_path)\n cmd = 'git pull'\n sys_call(cmd, cwd=repo_path)\n if chdir:\n os.chdir(repo_path)\n if reflect:\n if not HAS_DULWICH:\n raise RuntimeError('`pip install dulwich` required!')\n return Repo(repo_path)\n else:\n return repo_path\n",
"role": "user"
},
{
"content": "git_clone",
"role": "assistant"
}
] |
def x(value, msg=None, except_=None, inc_zeros=True):
if hasattr(value, 'empty'):
value = not bool(value.empty)
elif inc_zeros and value in ZEROS:
value = True
else:
pass
_is_null = is_null(value, except_=False)
result = bool(_is_null or not value)
if except_:
return is_true(result, msg=msg, except_=except_)
else:
return bool(result)
| is_empty | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(value, msg=None, except_=None, inc_zeros=True):\n if hasattr(value, 'empty'):\n value = not bool(value.empty)\n elif inc_zeros and value in ZEROS:\n value = True\n else:\n pass\n _is_null = is_null(value, except_=False)\n result = bool(_is_null or not value)\n if except_:\n return is_true(result, msg=msg, except_=except_)\n else:\n return bool(result)\n",
"role": "user"
},
{
"content": "is_empty",
"role": "assistant"
}
] |
def x(value, msg=None, except_=None):
value = False if hasattr(value, 'empty') else value
result = bool(value is None or value != value or repr(value) == 'NaT')
if except_:
return is_true(result, msg=msg, except_=except_)
else:
return bool(result)
| is_null | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(value, msg=None, except_=None):\n value = False if hasattr(value, 'empty') else value\n result = bool(value is None or value != value or repr(value) == 'NaT')\n if except_:\n return is_true(result, msg=msg, except_=except_)\n else:\n return bool(result)\n",
"role": "user"
},
{
"content": "is_null",
"role": "assistant"
}
] |
def x(obj):
if isinstance(obj, (datetime, date)):
result = dt2ts(obj)
else:
result = json_encoder.default(obj)
return to_encoding(result)
| json_encode_default | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(obj):\n if isinstance(obj, (datetime, date)):\n result = dt2ts(obj)\n else:\n result = json_encoder.default(obj)\n return to_encoding(result)\n",
"role": "user"
},
{
"content": "json_encode_default",
"role": "assistant"
}
] |
def x(obj, root=True, exclude=None, hash_func=_jsonhash_sha1):
if isinstance(obj, Mapping):
if root and exclude:
obj = {k: v for k, v in obj.iteritems() if k not in exclude}
result = sorted((k, jsonhash(v, False)) for k, v in obj.iteritems())
elif isinstance(obj, list):
result = tuple(jsonhash(e, False) for e in obj)
else:
result = obj
if root:
result = unicode(hash_func(result))
return result
| jsonhash | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(obj, root=True, exclude=None, hash_func=_jsonhash_sha1):\n if isinstance(obj, Mapping):\n if root and exclude:\n obj = {k: v for k, v in obj.iteritems() if k not in exclude}\n result = sorted((k, jsonhash(v, False)) for k, v in obj.iteritems())\n elif isinstance(obj, list):\n result = tuple(jsonhash(e, False) for e in obj)\n else:\n result = obj\n if root:\n result = unicode(hash_func(result))\n return result\n",
"role": "user"
},
{
"content": "jsonhash",
"role": "assistant"
}
] |
def x(path, filetype=None, as_df=False, retries=None, _oid=None, quiet=
False, **kwargs):
is_true(HAS_PANDAS, '`pip install pandas` required')
set_oid = set_oid_func(_oid)
if not isinstance(path, basestring):
objects = path
if not isinstance(objects, pd.DataFrame):
raise ValueError('loading raw values must be DataFrames')
elif re.match('https?://', path):
logger.debug('Saving %s to tmp file' % path)
_path = urlretrieve(path, retries)
logger.debug('%s saved to tmp file: %s' % (path, _path))
try:
objects = load_file(_path, filetype, **kwargs)
finally:
remove_file(_path)
else:
path = re.sub('^file://', '', path)
path = os.path.expanduser(path)
path = path if os.path.isabs(path) else pjoin(os.getcwd(), path)
files = sorted(glob.glob(os.path.expanduser(path)))
if not files:
raise IOError('failed to load: %s' % path)
objects = []
[objects.extend(load_file(ds, filetype, **kwargs)) for ds in files]
if is_empty(objects, except_=False) and not quiet:
raise RuntimeError('no objects extracted!')
else:
logger.debug('Data loaded successfully from %s' % path)
if set_oid:
objects = [set_oid(o) for o in objects]
if as_df:
return pd.DataFrame(objects)
else:
return objects
| load | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path, filetype=None, as_df=False, retries=None, _oid=None, quiet=\n False, **kwargs):\n is_true(HAS_PANDAS, '`pip install pandas` required')\n set_oid = set_oid_func(_oid)\n if not isinstance(path, basestring):\n objects = path\n if not isinstance(objects, pd.DataFrame):\n raise ValueError('loading raw values must be DataFrames')\n elif re.match('https?://', path):\n logger.debug('Saving %s to tmp file' % path)\n _path = urlretrieve(path, retries)\n logger.debug('%s saved to tmp file: %s' % (path, _path))\n try:\n objects = load_file(_path, filetype, **kwargs)\n finally:\n remove_file(_path)\n else:\n path = re.sub('^file://', '', path)\n path = os.path.expanduser(path)\n path = path if os.path.isabs(path) else pjoin(os.getcwd(), path)\n files = sorted(glob.glob(os.path.expanduser(path)))\n if not files:\n raise IOError('failed to load: %s' % path)\n objects = []\n [objects.extend(load_file(ds, filetype, **kwargs)) for ds in files]\n if is_empty(objects, except_=False) and not quiet:\n raise RuntimeError('no objects extracted!')\n else:\n logger.debug('Data loaded successfully from %s' % path)\n if set_oid:\n objects = [set_oid(o) for o in objects]\n if as_df:\n return pd.DataFrame(objects)\n else:\n return objects\n",
"role": "user"
},
{
"content": "load",
"role": "assistant"
}
] |
def x(pkg, cube):
try:
fromlist = map(str, [cube])
mcubes = __import__(pkg, fromlist=fromlist)
return getattr(mcubes, cube)
except AttributeError:
_pkg, _mod, _cls = cube_pkg_mod_cls(cube)
fromlist = map(str, [_cls])
mcubes = __import__('%s.%s.%s' % (pkg, _pkg, _mod), fromlist=fromlist)
return getattr(mcubes, _cls)
| load_cube_pkg | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(pkg, cube):\n try:\n fromlist = map(str, [cube])\n mcubes = __import__(pkg, fromlist=fromlist)\n return getattr(mcubes, cube)\n except AttributeError:\n _pkg, _mod, _cls = cube_pkg_mod_cls(cube)\n fromlist = map(str, [_cls])\n mcubes = __import__('%s.%s.%s' % (pkg, _pkg, _mod), fromlist=fromlist)\n return getattr(mcubes, _cls)\n",
"role": "user"
},
{
"content": "load_cube_pkg",
"role": "assistant"
}
] |
def x(rel_path, paths=None, raw=False, as_list=False, as_iter=False, *args,
**kwargs):
if not rel_path:
raise ValueError('rel_path can not be null!')
paths = str2list(paths)
paths.extend([STATIC_DIR, os.path.join(SRC_DIR, 'static')])
paths = [os.path.expanduser(p) for p in set(paths)]
for path in paths:
path = os.path.join(path, rel_path)
logger.debug('trying to read: %s ' % path)
if os.path.exists(path):
break
else:
raise IOError('path %s does not exist!' % rel_path)
args = args if args else ['rU']
fd = open(path, *args, **kwargs)
if raw:
return fd
if as_iter:
return read_in_chunks(fd)
else:
fd_lines = fd.readlines()
if as_list:
return fd_lines
else:
return ''.join(fd_lines)
| read_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rel_path, paths=None, raw=False, as_list=False, as_iter=False, *args,\n **kwargs):\n if not rel_path:\n raise ValueError('rel_path can not be null!')\n paths = str2list(paths)\n paths.extend([STATIC_DIR, os.path.join(SRC_DIR, 'static')])\n paths = [os.path.expanduser(p) for p in set(paths)]\n for path in paths:\n path = os.path.join(path, rel_path)\n logger.debug('trying to read: %s ' % path)\n if os.path.exists(path):\n break\n else:\n raise IOError('path %s does not exist!' % rel_path)\n args = args if args else ['rU']\n fd = open(path, *args, **kwargs)\n if raw:\n return fd\n if as_iter:\n return read_in_chunks(fd)\n else:\n fd_lines = fd.readlines()\n if as_list:\n return fd_lines\n else:\n return ''.join(fd_lines)\n",
"role": "user"
},
{
"content": "read_file",
"role": "assistant"
}
] |
def x(source, target):
for k, v in target.iteritems():
if isinstance(v, Mapping):
r = rupdate(source.get(k, {}), v)
source[k] = r
else:
source[k] = target[k]
return source
| rupdate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(source, target):\n for k, v in target.iteritems():\n if isinstance(v, Mapping):\n r = rupdate(source.get(k, {}), v)\n source[k] = r\n else:\n source[k] = target[k]\n return source\n",
"role": "user"
},
{
"content": "rupdate",
"role": "assistant"
}
] |
def x(str_):
str_ = str_ or ''
return ''.join(x for x in str_ if x.isalnum())
| safestr | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(str_):\n str_ = str_ or ''\n return ''.join(x for x in str_ if x.isalnum())\n",
"role": "user"
},
{
"content": "safestr",
"role": "assistant"
}
] |
def x(ts, milli=False, tz_aware=False):
is_true(HAS_DATEUTIL, '`pip install python_dateutil` required')
if isinstance(ts, (datetime, date)):
pass
elif is_empty(ts, except_=False):
return None
elif isinstance(ts, (int, float, long)) and ts < 0:
return None
elif isinstance(ts, basestring):
try:
ts = float(ts)
except (TypeError, ValueError):
try:
ts = dt_parse(ts)
except Exception:
raise TypeError(
'unable to derive datetime from timestamp string: %s' % ts)
elif milli:
ts = float(ts) / 1000.0
else:
ts = float(ts)
return _get_datetime(ts, tz_aware)
| ts2dt | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ts, milli=False, tz_aware=False):\n is_true(HAS_DATEUTIL, '`pip install python_dateutil` required')\n if isinstance(ts, (datetime, date)):\n pass\n elif is_empty(ts, except_=False):\n return None\n elif isinstance(ts, (int, float, long)) and ts < 0:\n return None\n elif isinstance(ts, basestring):\n try:\n ts = float(ts)\n except (TypeError, ValueError):\n try:\n ts = dt_parse(ts)\n except Exception:\n raise TypeError(\n 'unable to derive datetime from timestamp string: %s' % ts)\n elif milli:\n ts = float(ts) / 1000.0\n else:\n ts = float(ts)\n return _get_datetime(ts, tz_aware)\n",
"role": "user"
},
{
"content": "ts2dt",
"role": "assistant"
}
] |
def x(uri, saveas=None, retries=3, cache_dir=None):
retries = int(retries) if retries else 3
while retries:
try:
_path, headers = urllib.urlretrieve(uri, saveas)
except Exception as e:
retries -= 1
logger.warn('Failed getting uri "%s": %s (retry:%s in 1s)' % (
uri, e, retries))
time.sleep(0.2)
continue
else:
break
else:
raise RuntimeError('Failed to retrieve uri: %s' % uri)
return _path
| urlretrieve | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(uri, saveas=None, retries=3, cache_dir=None):\n retries = int(retries) if retries else 3\n while retries:\n try:\n _path, headers = urllib.urlretrieve(uri, saveas)\n except Exception as e:\n retries -= 1\n logger.warn('Failed getting uri \"%s\": %s (retry:%s in 1s)' % (\n uri, e, retries))\n time.sleep(0.2)\n continue\n else:\n break\n else:\n raise RuntimeError('Failed to retrieve uri: %s' % uri)\n return _path\n",
"role": "user"
},
{
"content": "urlretrieve",
"role": "assistant"
}
] |
def x(self, orm):
for category in orm['document_library.DocumentCategory'].objects.all():
category.is_published = True
category.save()
| forwards | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, orm):\n for category in orm['document_library.DocumentCategory'].objects.all():\n category.is_published = True\n category.save()\n",
"role": "user"
},
{
"content": "forwards",
"role": "assistant"
}
] |
def x(target_type, url_string, *args, **kwargs):
args_str = '<%s>' % '><'.join(args)
kwargs_str = '<%s>' % '><'.join('%s:%s' % pair for pair in kwargs.items())
url_str = ''.join([url_string, args_str, kwargs_str])
normalized_url = str(ResourceURL(url_str))
query_tuples = []
if singletons.settings.SECURITY and 'Sha1' in singletons.settings.SECURITY:
secret = singletons.settings.HMAC_SECRET
digest = get_hmac_sha1_digest(secret, normalized_url, target_type)
query_tuples.append(('digest', digest))
query_tuples.append(('url', normalized_url))
querystring = urlencode(query_tuples)
scheme = singletons.settings.EXTERNAL_SCHEME
host = singletons.settings.EXTERNAL_HOST
port = singletons.settings.EXTERNAL_PORT
if not host:
host = singletons.settings.HOST
if not port:
port = singletons.settings.PORT
port_suffix = ':%s' % port if port != 80 else ''
typestring_normalized = str(TypeString(target_type))
return '%s://%s%s/media/%s/?%s' % (scheme, host, port_suffix,
typestring_normalized, querystring)
| reverse_media_url | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(target_type, url_string, *args, **kwargs):\n args_str = '<%s>' % '><'.join(args)\n kwargs_str = '<%s>' % '><'.join('%s:%s' % pair for pair in kwargs.items())\n url_str = ''.join([url_string, args_str, kwargs_str])\n normalized_url = str(ResourceURL(url_str))\n query_tuples = []\n if singletons.settings.SECURITY and 'Sha1' in singletons.settings.SECURITY:\n secret = singletons.settings.HMAC_SECRET\n digest = get_hmac_sha1_digest(secret, normalized_url, target_type)\n query_tuples.append(('digest', digest))\n query_tuples.append(('url', normalized_url))\n querystring = urlencode(query_tuples)\n scheme = singletons.settings.EXTERNAL_SCHEME\n host = singletons.settings.EXTERNAL_HOST\n port = singletons.settings.EXTERNAL_PORT\n if not host:\n host = singletons.settings.HOST\n if not port:\n port = singletons.settings.PORT\n port_suffix = ':%s' % port if port != 80 else ''\n typestring_normalized = str(TypeString(target_type))\n return '%s://%s%s/media/%s/?%s' % (scheme, host, port_suffix,\n typestring_normalized, querystring)\n",
"role": "user"
},
{
"content": "reverse_media_url",
"role": "assistant"
}
] |
def x(isbn, checksum=True):
if not isinstance(isbn, string_types):
raise TypeError('ISBN must be a string, received %r' % isbn)
if PY2 and isinstance(isbn, str):
isbn = unicode(isbn)
uni_input = False
else:
uni_input = True
for dash in DASHES:
isbn = isbn.replace(dash, unicode())
if checksum:
if not isbn[:-1].isdigit():
raise IsbnError('non-digit parts')
if len(isbn) == 9:
isbn = '0' + isbn
if len(isbn) == 10:
if not (isbn[-1].isdigit() or isbn[-1] in 'Xx'):
raise IsbnError('non-digit or X checksum')
elif len(isbn) == 13:
if not isbn[-1].isdigit():
raise IsbnError('non-digit checksum')
if not isbn.startswith(('978', '979')):
raise IsbnError('invalid Bookland region')
else:
raise IsbnError('ISBN must be either 10 or 13 characters long')
else:
if len(isbn) == 8:
isbn = '0' + isbn
elif len(isbn) == 12 and not isbn[:3].startswith(('978', '979')):
raise IsbnError('invalid Bookland region')
if not isbn.isdigit():
raise IsbnError('non-digit parts')
if not len(isbn) in (9, 12):
raise IsbnError(
'ISBN must be either 9 or 12 characters long without checksum')
if PY2 and not uni_input:
return str(isbn)
else:
return isbn
| isbn_cleanse | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(isbn, checksum=True):\n if not isinstance(isbn, string_types):\n raise TypeError('ISBN must be a string, received %r' % isbn)\n if PY2 and isinstance(isbn, str):\n isbn = unicode(isbn)\n uni_input = False\n else:\n uni_input = True\n for dash in DASHES:\n isbn = isbn.replace(dash, unicode())\n if checksum:\n if not isbn[:-1].isdigit():\n raise IsbnError('non-digit parts')\n if len(isbn) == 9:\n isbn = '0' + isbn\n if len(isbn) == 10:\n if not (isbn[-1].isdigit() or isbn[-1] in 'Xx'):\n raise IsbnError('non-digit or X checksum')\n elif len(isbn) == 13:\n if not isbn[-1].isdigit():\n raise IsbnError('non-digit checksum')\n if not isbn.startswith(('978', '979')):\n raise IsbnError('invalid Bookland region')\n else:\n raise IsbnError('ISBN must be either 10 or 13 characters long')\n else:\n if len(isbn) == 8:\n isbn = '0' + isbn\n elif len(isbn) == 12 and not isbn[:3].startswith(('978', '979')):\n raise IsbnError('invalid Bookland region')\n if not isbn.isdigit():\n raise IsbnError('non-digit parts')\n if not len(isbn) in (9, 12):\n raise IsbnError(\n 'ISBN must be either 9 or 12 characters long without checksum')\n if PY2 and not uni_input:\n return str(isbn)\n else:\n return isbn\n",
"role": "user"
},
{
"content": "isbn_cleanse",
"role": "assistant"
}
] |
def x(isbn):
isbn = [int(i) for i in _isbn_cleanse(isbn, checksum=False)]
if len(isbn) == 9:
products = [(x * y) for x, y in zip(isbn, range(1, 10))]
check = sum(products) % 11
if check == 10:
check = 'X'
else:
products = []
for i in range(12):
if i % 2 == 0:
products.append(isbn[i])
else:
products.append(isbn[i] * 3)
check = 10 - sum(products) % 10
if check == 10:
check = 0
return str(check)
| calculate_checksum | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(isbn):\n isbn = [int(i) for i in _isbn_cleanse(isbn, checksum=False)]\n if len(isbn) == 9:\n products = [(x * y) for x, y in zip(isbn, range(1, 10))]\n check = sum(products) % 11\n if check == 10:\n check = 'X'\n else:\n products = []\n for i in range(12):\n if i % 2 == 0:\n products.append(isbn[i])\n else:\n products.append(isbn[i] * 3)\n check = 10 - sum(products) % 10\n if check == 10:\n check = 0\n return str(check)\n",
"role": "user"
},
{
"content": "calculate_checksum",
"role": "assistant"
}
] |
def x(isbn, code='978'):
isbn = _isbn_cleanse(isbn)
if len(isbn) == 10:
isbn = code + isbn[:-1]
return isbn + calculate_checksum(isbn)
elif isbn.startswith('978'):
return isbn[3:-1] + calculate_checksum(isbn[3:-1])
else:
raise IsbnError(
'Only ISBN-13s with 978 Bookland code can be converted to ISBN-10.'
)
| convert | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(isbn, code='978'):\n isbn = _isbn_cleanse(isbn)\n if len(isbn) == 10:\n isbn = code + isbn[:-1]\n return isbn + calculate_checksum(isbn)\n elif isbn.startswith('978'):\n return isbn[3:-1] + calculate_checksum(isbn[3:-1])\n else:\n raise IsbnError(\n 'Only ISBN-13s with 978 Bookland code can be converted to ISBN-10.'\n )\n",
"role": "user"
},
{
"content": "convert",
"role": "assistant"
}
] |
def x(self):
if len(self.isbn) in (9, 12):
return calculate_checksum(self.isbn)
else:
return calculate_checksum(self.isbn[:-1])
| calculate_checksum | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if len(self.isbn) in (9, 12):\n return calculate_checksum(self.isbn)\n else:\n return calculate_checksum(self.isbn[:-1])\n",
"role": "user"
},
{
"content": "calculate_checksum",
"role": "assistant"
}
] |
def x(self, site='amazon', country='us'):
try:
try:
url, tlds = URL_MAP[site]
except ValueError:
tlds = None
url = URL_MAP[site]
except KeyError:
raise SiteError(site)
inject = {'isbn': self._isbn}
if tlds:
if country not in tlds:
raise CountryError(country)
tld = tlds[country]
if not tld:
tld = country
inject['tld'] = tld
return url % inject
| to_url | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, site='amazon', country='us'):\n try:\n try:\n url, tlds = URL_MAP[site]\n except ValueError:\n tlds = None\n url = URL_MAP[site]\n except KeyError:\n raise SiteError(site)\n inject = {'isbn': self._isbn}\n if tlds:\n if country not in tlds:\n raise CountryError(country)\n tld = tlds[country]\n if not tld:\n tld = country\n inject['tld'] = tld\n return url % inject\n",
"role": "user"
},
{
"content": "to_url",
"role": "assistant"
}
] |
def x(self, text, token_class=None):
token_class = token_class or Token
tokens = {}
for i, match in enumerate(self.regex.finditer(text)):
value = match.group(0)
try:
token = tokens[value]
except KeyError:
type = match.lastgroup
token = token_class(value, type=type)
tokens[value] = token
yield token
| tokenize | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, text, token_class=None):\n token_class = token_class or Token\n tokens = {}\n for i, match in enumerate(self.regex.finditer(text)):\n value = match.group(0)\n try:\n token = tokens[value]\n except KeyError:\n type = match.lastgroup\n token = token_class(value, type=type)\n tokens[value] = token\n yield token\n",
"role": "user"
},
{
"content": "tokenize",
"role": "assistant"
}
] |
def x(facter_results):
last_key, last_value = None, []
for line in filter(None, facter_results.splitlines()):
res = line.split(six.u(' => '), 1)
if len(res) == 1:
if not last_key:
raise ValueError('parse error')
else:
last_value.append(res[0])
else:
if last_key:
yield last_key, os.linesep.join(last_value)
last_key, last_value = res[0], [res[1]]
else:
if last_key:
yield last_key, os.linesep.join(last_value)
| parse_cli_facter_results | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(facter_results):\n last_key, last_value = None, []\n for line in filter(None, facter_results.splitlines()):\n res = line.split(six.u(' => '), 1)\n if len(res) == 1:\n if not last_key:\n raise ValueError('parse error')\n else:\n last_value.append(res[0])\n else:\n if last_key:\n yield last_key, os.linesep.join(last_value)\n last_key, last_value = res[0], [res[1]]\n else:\n if last_key:\n yield last_key, os.linesep.join(last_value)\n",
"role": "user"
},
{
"content": "parse_cli_facter_results",
"role": "assistant"
}
] |
def x(self, key=None):
args = [self.facter_path]
args.append('--puppet')
if self.external_dir is not None:
args.append('--external-dir')
args.append(self.external_dir)
if self.uses_yaml:
args.append('--yaml')
if key is not None:
args.append(key)
proc = subprocess.Popen(args, stdout=subprocess.PIPE)
results = proc.stdout.read()
if self.uses_yaml:
parsed_results = yaml.load(results)
if key is not None:
return parsed_results[key]
else:
return parsed_results
results = results.decode()
if key is not None:
return results.strip()
else:
return dict(_parse_cli_facter_results(results))
| run_facter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key=None):\n args = [self.facter_path]\n args.append('--puppet')\n if self.external_dir is not None:\n args.append('--external-dir')\n args.append(self.external_dir)\n if self.uses_yaml:\n args.append('--yaml')\n if key is not None:\n args.append(key)\n proc = subprocess.Popen(args, stdout=subprocess.PIPE)\n results = proc.stdout.read()\n if self.uses_yaml:\n parsed_results = yaml.load(results)\n if key is not None:\n return parsed_results[key]\n else:\n return parsed_results\n results = results.decode()\n if key is not None:\n return results.strip()\n else:\n return dict(_parse_cli_facter_results(results))\n",
"role": "user"
},
{
"content": "run_facter",
"role": "assistant"
}
] |
def x(self):
if not self.cache_enabled:
return False
if self._cache is None:
self.build_cache()
return True
| has_cache | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if not self.cache_enabled:\n return False\n if self._cache is None:\n self.build_cache()\n return True\n",
"role": "user"
},
{
"content": "has_cache",
"role": "assistant"
}
] |
def x(self, fact, cache=True):
if not cache or not self.has_cache():
val = self.run_facter(fact)
if val is None or val == '':
raise KeyError(fact)
return val
return self._cache[fact]
| lookup | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, fact, cache=True):\n if not cache or not self.has_cache():\n val = self.run_facter(fact)\n if val is None or val == '':\n raise KeyError(fact)\n return val\n return self._cache[fact]\n",
"role": "user"
},
{
"content": "lookup",
"role": "assistant"
}
] |
def x(self) ->None:
self.project: str = namesgenerator.get_random_name()
self._processed: List = list()
self.results: List = list()
| reset | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->None:\n self.project: str = namesgenerator.get_random_name()\n self._processed: List = list()\n self.results: List = list()\n",
"role": "user"
},
{
"content": "reset",
"role": "assistant"
}
] |
def x(self) ->None:
if self.output:
self.folder: str = os.getcwd() + '/' + self.project
os.mkdir(self.folder)
| config_bootstrap | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->None:\n if self.output:\n self.folder: str = os.getcwd() + '/' + self.project\n os.mkdir(self.folder)\n",
"role": "user"
},
{
"content": "config_bootstrap",
"role": "assistant"
}
] |
def x(self, module: str, kwargs: str):
package_directory: str = os.path.dirname(os.path.abspath(__file__))
modules: str = package_directory + '/modules'
module = module + '.py'
if module not in os.listdir(modules):
raise Exception('Module %s is not valid' % module)
module_name: str = module[:-3]
import_path: str = '%s.%s' % (self.MODULE_PATH, module_name)
imported = import_module(import_path)
obj = getattr(imported, 'Module')
return obj(**kwargs)
| dyn_loader | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, module: str, kwargs: str):\n package_directory: str = os.path.dirname(os.path.abspath(__file__))\n modules: str = package_directory + '/modules'\n module = module + '.py'\n if module not in os.listdir(modules):\n raise Exception('Module %s is not valid' % module)\n module_name: str = module[:-3]\n import_path: str = '%s.%s' % (self.MODULE_PATH, module_name)\n imported = import_module(import_path)\n obj = getattr(imported, 'Module')\n return obj(**kwargs)\n",
"role": "user"
},
{
"content": "dyn_loader",
"role": "assistant"
}
] |
def x(self) ->bool:
while True:
try:
task = self._unfullfilled.get_nowait()
except queue.Empty:
break
else:
self._log.debug('Job: %s' % str(task))
engine = self._dyn_loader(task['engine'], task)
task['start_time'] = now_time()
results = engine.search()
task['end_time'] = now_time()
duration: str = str((task['end_time'] - task['start_time']).seconds
)
task['duration'] = duration
task.update({'results': results})
self._fulfilled.put(task)
return True
| job_handler | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->bool:\n while True:\n try:\n task = self._unfullfilled.get_nowait()\n except queue.Empty:\n break\n else:\n self._log.debug('Job: %s' % str(task))\n engine = self._dyn_loader(task['engine'], task)\n task['start_time'] = now_time()\n results = engine.search()\n task['end_time'] = now_time()\n duration: str = str((task['end_time'] - task['start_time']).seconds\n )\n task['duration'] = duration\n task.update({'results': results})\n self._fulfilled.put(task)\n return True\n",
"role": "user"
},
{
"content": "job_handler",
"role": "assistant"
}
] |
def x(self) ->None:
self._log.info("Saving results to '%s'" % self.folder)
path: str = self.folder + '/'
for job in self.results:
if job['domain'] in self.saved:
continue
job['start_time'] = str_datetime(job['start_time'])
job['end_time'] = str_datetime(job['end_time'])
jid: int = random.randint(100000, 999999)
filename: str = '%s_%s_%d_job.json' % (self.project, job['domain'], jid
)
handle = open(path + filename, 'w')
handle.write(json.dumps(job, indent=4))
handle.close()
filename = '%s_%s_%d_emails.txt' % (self.project, job['domain'], jid)
handle = open(path + filename, 'w')
for email in job['results']['emails']:
handle.write(email + '\n')
handle.close()
self.saved.append(job['domain'])
| save | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->None:\n self._log.info(\"Saving results to '%s'\" % self.folder)\n path: str = self.folder + '/'\n for job in self.results:\n if job['domain'] in self.saved:\n continue\n job['start_time'] = str_datetime(job['start_time'])\n job['end_time'] = str_datetime(job['end_time'])\n jid: int = random.randint(100000, 999999)\n filename: str = '%s_%s_%d_job.json' % (self.project, job['domain'], jid\n )\n handle = open(path + filename, 'w')\n handle.write(json.dumps(job, indent=4))\n handle.close()\n filename = '%s_%s_%d_emails.txt' % (self.project, job['domain'], jid)\n handle = open(path + filename, 'w')\n for email in job['results']['emails']:\n handle.write(email + '\\n')\n handle.close()\n self.saved.append(job['domain'])\n",
"role": "user"
},
{
"content": "save",
"role": "assistant"
}
] |
def x(self, jobs: List[Dict[str, str]]) ->None:
if not isinstance(jobs, list):
raise Exception('Jobs must be of type list.')
self._log.info('Project: %s' % self.project)
self._log.info('Processing jobs: %d', len(jobs))
for _, job in enumerate(jobs):
self._unfullfilled.put(job)
for _ in range(self.PROCESSES):
proc: Process = Process(target=self._job_handler)
self._processes.append(proc)
proc.start()
for proc in self._processes:
proc.join()
while not self._fulfilled.empty():
output: Dict = self._fulfilled.get()
output.update({'project': self.project})
self._processed.append(output['domain'])
self.results.append(output)
if output['greedy']:
bonus_jobs: List = list()
observed: List = list()
for item in output['results']['emails']:
found: str = item.split('@')[1]
if found in self._processed or found in observed:
continue
observed.append(found)
base: Dict = dict()
base['limit'] = output['limit']
base['modifier'] = output['modifier']
base['engine'] = output['engine']
base['greedy'] = False
base['domain'] = found
bonus_jobs.append(base)
if len(bonus_jobs) > 0:
self.search(bonus_jobs)
self._log.info('All jobs processed')
if self.output:
self._save()
| search | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, jobs: List[Dict[str, str]]) ->None:\n if not isinstance(jobs, list):\n raise Exception('Jobs must be of type list.')\n self._log.info('Project: %s' % self.project)\n self._log.info('Processing jobs: %d', len(jobs))\n for _, job in enumerate(jobs):\n self._unfullfilled.put(job)\n for _ in range(self.PROCESSES):\n proc: Process = Process(target=self._job_handler)\n self._processes.append(proc)\n proc.start()\n for proc in self._processes:\n proc.join()\n while not self._fulfilled.empty():\n output: Dict = self._fulfilled.get()\n output.update({'project': self.project})\n self._processed.append(output['domain'])\n self.results.append(output)\n if output['greedy']:\n bonus_jobs: List = list()\n observed: List = list()\n for item in output['results']['emails']:\n found: str = item.split('@')[1]\n if found in self._processed or found in observed:\n continue\n observed.append(found)\n base: Dict = dict()\n base['limit'] = output['limit']\n base['modifier'] = output['modifier']\n base['engine'] = output['engine']\n base['greedy'] = False\n base['domain'] = found\n bonus_jobs.append(base)\n if len(bonus_jobs) > 0:\n self.search(bonus_jobs)\n self._log.info('All jobs processed')\n if self.output:\n self._save()\n",
"role": "user"
},
{
"content": "search",
"role": "assistant"
}
] |
def x(latitude, longitude, altitude):
r = earth_geo_radius + altitude
x = r * np.cos(np.deg2rad(latitude)) * np.cos(np.deg2rad(longitude))
y = r * np.cos(np.deg2rad(latitude)) * np.sin(np.deg2rad(longitude))
z = r * np.sin(np.deg2rad(latitude))
return x, y, z
| geocentric_to_ecef | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(latitude, longitude, altitude):\n r = earth_geo_radius + altitude\n x = r * np.cos(np.deg2rad(latitude)) * np.cos(np.deg2rad(longitude))\n y = r * np.cos(np.deg2rad(latitude)) * np.sin(np.deg2rad(longitude))\n z = r * np.sin(np.deg2rad(latitude))\n return x, y, z\n",
"role": "user"
},
{
"content": "geocentric_to_ecef",
"role": "assistant"
}
] |
def x(x, y, z, ref_height=None):
if ref_height is None:
ref_height = earth_geo_radius
r = np.sqrt(x ** 2 + y ** 2 + z ** 2)
colatitude = np.rad2deg(np.arccos(z / r))
longitude = np.rad2deg(np.arctan2(y, x))
latitude = 90.0 - colatitude
return latitude, longitude, r - ref_height
| ecef_to_geocentric | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(x, y, z, ref_height=None):\n if ref_height is None:\n ref_height = earth_geo_radius\n r = np.sqrt(x ** 2 + y ** 2 + z ** 2)\n colatitude = np.rad2deg(np.arccos(z / r))\n longitude = np.rad2deg(np.arctan2(y, x))\n latitude = 90.0 - colatitude\n return latitude, longitude, r - ref_height\n",
"role": "user"
},
{
"content": "ecef_to_geocentric",
"role": "assistant"
}
] |
def x(latitude, longitude, altitude):
ellip = np.sqrt(1.0 - earth_b ** 2 / earth_a ** 2)
r_n = earth_a / np.sqrt(1.0 - ellip ** 2 * np.sin(np.deg2rad(latitude)) **
2)
x = (r_n + altitude) * np.cos(np.deg2rad(latitude)) * np.cos(np.deg2rad
(longitude))
y = (r_n + altitude) * np.cos(np.deg2rad(latitude)) * np.sin(np.deg2rad
(longitude))
z = (r_n * (1.0 - ellip ** 2) + altitude) * np.sin(np.deg2rad(latitude))
return x, y, z
| geodetic_to_ecef | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(latitude, longitude, altitude):\n ellip = np.sqrt(1.0 - earth_b ** 2 / earth_a ** 2)\n r_n = earth_a / np.sqrt(1.0 - ellip ** 2 * np.sin(np.deg2rad(latitude)) **\n 2)\n x = (r_n + altitude) * np.cos(np.deg2rad(latitude)) * np.cos(np.deg2rad\n (longitude))\n y = (r_n + altitude) * np.cos(np.deg2rad(latitude)) * np.sin(np.deg2rad\n (longitude))\n z = (r_n * (1.0 - ellip ** 2) + altitude) * np.sin(np.deg2rad(latitude))\n return x, y, z\n",
"role": "user"
},
{
"content": "geodetic_to_ecef",
"role": "assistant"
}
] |
def x(x, y, z, method=None):
method = method or 'closed'
ellip = np.sqrt(1.0 - earth_b ** 2 / earth_a ** 2)
e2 = ellip ** 2
longitude = np.arctan2(y, x)
p = np.sqrt(x ** 2 + y ** 2)
if method == 'closed':
e_prime = np.sqrt((earth_a ** 2 - earth_b ** 2) / earth_b ** 2)
theta = np.arctan2(z * earth_a, p * earth_b)
latitude = np.arctan2(z + e_prime ** 2 * earth_b * np.sin(theta) **
3, p - e2 * earth_a * np.cos(theta) ** 3)
r_n = earth_a / np.sqrt(1.0 - e2 * np.sin(latitude) ** 2)
h = p / np.cos(latitude) - r_n
if method == 'iterative':
latitude = np.arctan2(p, z)
r_n = earth_a / np.sqrt(1.0 - e2 * np.sin(latitude) ** 2)
for i in np.arange(6):
r_n = earth_a / np.sqrt(1.0 - e2 * np.sin(latitude) ** 2)
h = p / np.cos(latitude) - r_n
latitude = np.arctan(z / p / (1.0 - e2 * (r_n / (r_n + h))))
h = p / np.cos(latitude) - r_n
return np.rad2deg(latitude), np.rad2deg(longitude), h
| ecef_to_geodetic | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(x, y, z, method=None):\n method = method or 'closed'\n ellip = np.sqrt(1.0 - earth_b ** 2 / earth_a ** 2)\n e2 = ellip ** 2\n longitude = np.arctan2(y, x)\n p = np.sqrt(x ** 2 + y ** 2)\n if method == 'closed':\n e_prime = np.sqrt((earth_a ** 2 - earth_b ** 2) / earth_b ** 2)\n theta = np.arctan2(z * earth_a, p * earth_b)\n latitude = np.arctan2(z + e_prime ** 2 * earth_b * np.sin(theta) **\n 3, p - e2 * earth_a * np.cos(theta) ** 3)\n r_n = earth_a / np.sqrt(1.0 - e2 * np.sin(latitude) ** 2)\n h = p / np.cos(latitude) - r_n\n if method == 'iterative':\n latitude = np.arctan2(p, z)\n r_n = earth_a / np.sqrt(1.0 - e2 * np.sin(latitude) ** 2)\n for i in np.arange(6):\n r_n = earth_a / np.sqrt(1.0 - e2 * np.sin(latitude) ** 2)\n h = p / np.cos(latitude) - r_n\n latitude = np.arctan(z / p / (1.0 - e2 * (r_n / (r_n + h))))\n h = p / np.cos(latitude) - r_n\n return np.rad2deg(latitude), np.rad2deg(longitude), h\n",
"role": "user"
},
{
"content": "ecef_to_geodetic",
"role": "assistant"
}
] |
def x(east, north, up, glat, glong):
rlat = np.radians(glat)
rlon = np.radians(glong)
x = -east * np.sin(rlon) - north * np.cos(rlon) * np.sin(rlat
) + up * np.cos(rlon) * np.cos(rlat)
y = east * np.cos(rlon) - north * np.sin(rlon) * np.sin(rlat
) + up * np.sin(rlon) * np.cos(rlat)
z = north * np.cos(rlat) + up * np.sin(rlat)
return x, y, z
| enu_to_ecef_vector | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(east, north, up, glat, glong):\n rlat = np.radians(glat)\n rlon = np.radians(glong)\n x = -east * np.sin(rlon) - north * np.cos(rlon) * np.sin(rlat\n ) + up * np.cos(rlon) * np.cos(rlat)\n y = east * np.cos(rlon) - north * np.sin(rlon) * np.sin(rlat\n ) + up * np.sin(rlon) * np.cos(rlat)\n z = north * np.cos(rlat) + up * np.sin(rlat)\n return x, y, z\n",
"role": "user"
},
{
"content": "enu_to_ecef_vector",
"role": "assistant"
}
] |
def x(x, y, z, xx, xy, xz, yx, yy, yz, zx, zy, zz):
out_x = x * xx + y * xy + z * xz
out_y = x * yx + y * yy + z * yz
out_z = x * zx + y * zy + z * zz
return out_x, out_y, out_z
| project_ecef_vector_onto_basis | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(x, y, z, xx, xy, xz, yx, yy, yz, zx, zy, zz):\n out_x = x * xx + y * xy + z * xz\n out_y = x * yx + y * yy + z * yz\n out_z = x * zx + y * zy + z * zz\n return out_x, out_y, out_z\n",
"role": "user"
},
{
"content": "project_ecef_vector_onto_basis",
"role": "assistant"
}
] |
def x(x, y, z):
mag = np.sqrt(x ** 2 + y ** 2 + z ** 2)
x = x / mag
y = y / mag
z = z / mag
return x, y, z
| normalize_vector | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(x, y, z):\n mag = np.sqrt(x ** 2 + y ** 2 + z ** 2)\n x = x / mag\n y = y / mag\n z = z / mag\n return x, y, z\n",
"role": "user"
},
{
"content": "normalize_vector",
"role": "assistant"
}
] |
def x(x1, y1, z1, x2, y2, z2):
x = y1 * z2 - y2 * z1
y = z1 * x2 - x1 * z2
z = x1 * y2 - y1 * x2
return x, y, z
| cross_product | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(x1, y1, z1, x2, y2, z2):\n x = y1 * z2 - y2 * z1\n y = z1 * x2 - x1 * z2\n z = x1 * y2 - y1 * x2\n return x, y, z\n",
"role": "user"
},
{
"content": "cross_product",
"role": "assistant"
}
] |
def x(init, date, direction, height, steps=None, max_steps=10000.0,
step_size=10.0, recursive_loop_count=None, recurse=True):
if recursive_loop_count is None:
recursive_loop_count = 0
if steps is None:
steps = np.arange(max_steps)
if not isinstance(date, float):
doy = (date - datetime.datetime(date.year, 1, 1)).days
num_doy_year = (datetime.datetime(date.year + 1, 1, 1) - datetime.
datetime(date.year, 1, 1)).days
date = float(date.year) + float(doy) / float(num_doy_year) + float(
date.hour + date.minute / 60.0 + date.second / 3600.0) / 24.0
trace_north = scipy.integrate.odeint(igrf.igrf_step, init.copy(), steps,
args=(date, step_size, direction, height), full_output=False,
printmessg=False, ixpr=False)
check = trace_north[-1, :]
x, y, z = ecef_to_geodetic(*check)
if height == 0:
check_height = 1.0
else:
check_height = height
if recurse & (z > check_height * 1.000001):
if recursive_loop_count < 1000:
recursive_loop_count = recursive_loop_count + 1
trace_north1 = field_line_trace(check, date, direction, height,
step_size=step_size, max_steps=max_steps,
recursive_loop_count=recursive_loop_count, steps=steps)
else:
raise RuntimeError(
"After 1000 iterations couldn't reach target altitude")
return np.vstack((trace_north, trace_north1))
else:
return trace_north
| field_line_trace | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(init, date, direction, height, steps=None, max_steps=10000.0,\n step_size=10.0, recursive_loop_count=None, recurse=True):\n if recursive_loop_count is None:\n recursive_loop_count = 0\n if steps is None:\n steps = np.arange(max_steps)\n if not isinstance(date, float):\n doy = (date - datetime.datetime(date.year, 1, 1)).days\n num_doy_year = (datetime.datetime(date.year + 1, 1, 1) - datetime.\n datetime(date.year, 1, 1)).days\n date = float(date.year) + float(doy) / float(num_doy_year) + float(\n date.hour + date.minute / 60.0 + date.second / 3600.0) / 24.0\n trace_north = scipy.integrate.odeint(igrf.igrf_step, init.copy(), steps,\n args=(date, step_size, direction, height), full_output=False,\n printmessg=False, ixpr=False)\n check = trace_north[-1, :]\n x, y, z = ecef_to_geodetic(*check)\n if height == 0:\n check_height = 1.0\n else:\n check_height = height\n if recurse & (z > check_height * 1.000001):\n if recursive_loop_count < 1000:\n recursive_loop_count = recursive_loop_count + 1\n trace_north1 = field_line_trace(check, date, direction, height,\n step_size=step_size, max_steps=max_steps,\n recursive_loop_count=recursive_loop_count, steps=steps)\n else:\n raise RuntimeError(\n \"After 1000 iterations couldn't reach target altitude\")\n return np.vstack((trace_north, trace_north1))\n else:\n return trace_north\n",
"role": "user"
},
{
"content": "field_line_trace",
"role": "assistant"
}
] |
def x(init, date, height, step_size=100.0, max_steps=1000, steps=None, **kwargs
):
if steps is None:
steps = np.arange(max_steps)
trace_south = field_line_trace(init, date, -1.0, height, steps=steps,
step_size=step_size, max_steps=max_steps, **kwargs)
trace_north = field_line_trace(init, date, 1.0, height, steps=steps,
step_size=step_size, max_steps=max_steps, **kwargs)
trace = np.vstack((trace_south[::-1][:-1, :], trace_north))
return trace
| full_field_line | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(init, date, height, step_size=100.0, max_steps=1000, steps=None, **kwargs\n ):\n if steps is None:\n steps = np.arange(max_steps)\n trace_south = field_line_trace(init, date, -1.0, height, steps=steps,\n step_size=step_size, max_steps=max_steps, **kwargs)\n trace_north = field_line_trace(init, date, 1.0, height, steps=steps,\n step_size=step_size, max_steps=max_steps, **kwargs)\n trace = np.vstack((trace_south[::-1][:-1, :], trace_north))\n return trace\n",
"role": "user"
},
{
"content": "full_field_line",
"role": "assistant"
}
] |
def x(latitude, longitude, altitude, datetimes, steps=None, max_steps=1000,
step_size=100.0, ref_height=120.0, filter_zonal=True):
if steps is None:
steps = np.arange(max_steps)
ecef_x, ecef_y, ecef_z = geodetic_to_ecef(latitude, longitude, altitude)
geo_lat, geo_long, geo_alt = ecef_to_geocentric(ecef_x, ecef_y, ecef_z,
ref_height=0.0)
idx, = np.where(geo_long < 0)
geo_long[idx] = geo_long[idx] + 360.0
north_x = []
north_y = []
north_z = []
south_x = []
south_y = []
south_z = []
bn = []
be = []
bd = []
for x, y, z, alt, colat, elong, time in zip(ecef_x, ecef_y, ecef_z,
geo_alt, np.deg2rad(90.0 - geo_lat), np.deg2rad(geo_long), datetimes):
init = np.array([x, y, z])
trace_north = field_line_trace(init, time, 1.0, ref_height, steps=
steps, step_size=step_size, max_steps=max_steps)
trace_south = field_line_trace(init, time, -1.0, ref_height, steps=
steps, step_size=step_size, max_steps=max_steps)
trace_north = trace_north[-1, :]
trace_south = trace_south[-1, :]
doy = (time - datetime.datetime(time.year, 1, 1)).days
num_doy_year = (datetime.datetime(time.year + 1, 1, 1) - datetime.
datetime(time.year, 1, 1)).days
date = time.year + float(doy) / float(num_doy_year) + (time.hour +
time.minute / 60.0 + time.second / 3600.0) / 24.0
tbn, tbe, tbd, tbmag = igrf.igrf12syn(0, date, 1, alt, colat, elong)
south_x.append(trace_south[0])
south_y.append(trace_south[1])
south_z.append(trace_south[2])
north_x.append(trace_north[0])
north_y.append(trace_north[1])
north_z.append(trace_north[2])
bn.append(tbn)
be.append(tbe)
bd.append(tbd)
north_x = np.array(north_x)
north_y = np.array(north_y)
north_z = np.array(north_z)
south_x = np.array(south_x)
south_y = np.array(south_y)
south_z = np.array(south_z)
bn = np.array(bn)
be = np.array(be)
bd = np.array(bd)
north_x = north_x - ecef_x
north_y = north_y - ecef_y
north_z = north_z - ecef_z
north_x, north_y, north_z = normalize_vector(north_x, north_y, north_z)
south_x = south_x - ecef_x
south_y = south_y - ecef_y
south_z = south_z - ecef_z
south_x, south_y, south_z = normalize_vector(south_x, south_y, south_z)
bx, by, bz = enu_to_ecef_vector(be, bn, -bd, geo_lat, geo_long)
bx, by, bz = normalize_vector(bx, by, bz)
zvx_foot, zvy_foot, zvz_foot = cross_product(south_x, south_y, south_z,
north_x, north_y, north_z)
zvx_north, zvy_north, zvz_north = cross_product(north_x, north_y,
north_z, bx, by, bz)
zvx_south, zvy_south, zvz_south = cross_product(south_x, south_y,
south_z, bx, by, bz)
norm_foot = np.sqrt(zvx_foot ** 2 + zvy_foot ** 2 + zvz_foot ** 2)
zvx = zvx_foot / norm_foot
zvy = zvy_foot / norm_foot
zvz = zvz_foot / norm_foot
dot_fa = zvx * bx + zvy * by + zvz * bz
zvx -= dot_fa * bx
zvy -= dot_fa * by
zvz -= dot_fa * bz
zvx, zvy, zvz = normalize_vector(zvx, zvy, zvz)
mx, my, mz = cross_product(zvx, zvy, zvz, bx, by, bz)
return zvx, zvy, zvz, bx, by, bz, mx, my, mz
| calculate_mag_drift_unit_vectors_ecef | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(latitude, longitude, altitude, datetimes, steps=None, max_steps=1000,\n step_size=100.0, ref_height=120.0, filter_zonal=True):\n if steps is None:\n steps = np.arange(max_steps)\n ecef_x, ecef_y, ecef_z = geodetic_to_ecef(latitude, longitude, altitude)\n geo_lat, geo_long, geo_alt = ecef_to_geocentric(ecef_x, ecef_y, ecef_z,\n ref_height=0.0)\n idx, = np.where(geo_long < 0)\n geo_long[idx] = geo_long[idx] + 360.0\n north_x = []\n north_y = []\n north_z = []\n south_x = []\n south_y = []\n south_z = []\n bn = []\n be = []\n bd = []\n for x, y, z, alt, colat, elong, time in zip(ecef_x, ecef_y, ecef_z,\n geo_alt, np.deg2rad(90.0 - geo_lat), np.deg2rad(geo_long), datetimes):\n init = np.array([x, y, z])\n trace_north = field_line_trace(init, time, 1.0, ref_height, steps=\n steps, step_size=step_size, max_steps=max_steps)\n trace_south = field_line_trace(init, time, -1.0, ref_height, steps=\n steps, step_size=step_size, max_steps=max_steps)\n trace_north = trace_north[-1, :]\n trace_south = trace_south[-1, :]\n doy = (time - datetime.datetime(time.year, 1, 1)).days\n num_doy_year = (datetime.datetime(time.year + 1, 1, 1) - datetime.\n datetime(time.year, 1, 1)).days\n date = time.year + float(doy) / float(num_doy_year) + (time.hour + \n time.minute / 60.0 + time.second / 3600.0) / 24.0\n tbn, tbe, tbd, tbmag = igrf.igrf12syn(0, date, 1, alt, colat, elong)\n south_x.append(trace_south[0])\n south_y.append(trace_south[1])\n south_z.append(trace_south[2])\n north_x.append(trace_north[0])\n north_y.append(trace_north[1])\n north_z.append(trace_north[2])\n bn.append(tbn)\n be.append(tbe)\n bd.append(tbd)\n north_x = np.array(north_x)\n north_y = np.array(north_y)\n north_z = np.array(north_z)\n south_x = np.array(south_x)\n south_y = np.array(south_y)\n south_z = np.array(south_z)\n bn = np.array(bn)\n be = np.array(be)\n bd = np.array(bd)\n north_x = north_x - ecef_x\n north_y = north_y - ecef_y\n north_z = north_z - ecef_z\n north_x, north_y, north_z = normalize_vector(north_x, north_y, north_z)\n south_x = south_x - ecef_x\n south_y = south_y - ecef_y\n south_z = south_z - ecef_z\n south_x, south_y, south_z = normalize_vector(south_x, south_y, south_z)\n bx, by, bz = enu_to_ecef_vector(be, bn, -bd, geo_lat, geo_long)\n bx, by, bz = normalize_vector(bx, by, bz)\n zvx_foot, zvy_foot, zvz_foot = cross_product(south_x, south_y, south_z,\n north_x, north_y, north_z)\n zvx_north, zvy_north, zvz_north = cross_product(north_x, north_y,\n north_z, bx, by, bz)\n zvx_south, zvy_south, zvz_south = cross_product(south_x, south_y,\n south_z, bx, by, bz)\n norm_foot = np.sqrt(zvx_foot ** 2 + zvy_foot ** 2 + zvz_foot ** 2)\n zvx = zvx_foot / norm_foot\n zvy = zvy_foot / norm_foot\n zvz = zvz_foot / norm_foot\n dot_fa = zvx * bx + zvy * by + zvz * bz\n zvx -= dot_fa * bx\n zvy -= dot_fa * by\n zvz -= dot_fa * bz\n zvx, zvy, zvz = normalize_vector(zvx, zvy, zvz)\n mx, my, mz = cross_product(zvx, zvy, zvz, bx, by, bz)\n return zvx, zvy, zvz, bx, by, bz, mx, my, mz\n",
"role": "user"
},
{
"content": "calculate_mag_drift_unit_vectors_ecef",
"role": "assistant"
}
] |
def x(pos, field_line, sign, time, direction=None, step_size_goal=5.0,
field_step_size=None):
field_copy = field_line
last_min_dist = 2500000.0
scalar = 0.0
repeat = True
first = True
factor = 1
while repeat:
unit_steps = np.abs(scalar // step_size_goal)
if unit_steps == 0:
unit_steps = 1
pos_step = step_along_mag_unit_vector(pos[0], pos[1], pos[2], time,
direction=direction, num_steps=unit_steps, step_size=np.abs(
scalar) / unit_steps, scalar=sign)
diff = field_copy - pos_step
diff_mag = np.sqrt((diff ** 2).sum(axis=1))
min_idx = np.argmin(diff_mag)
if first:
init = field_copy[min_idx, :]
field_copy = full_field_line(init, time, 0.0, step_size=0.01,
max_steps=int(field_step_size / 0.01), recurse=False)
diff = field_copy - pos_step
diff_mag = np.sqrt((diff ** 2).sum(axis=1))
min_idx = np.argmin(diff_mag)
first = False
min_dist = diff_mag[min_idx]
if min_dist > last_min_dist:
if factor > 4:
repeat = False
scalar = scalar - last_min_dist / (2 * factor)
pos_step = step_along_mag_unit_vector(pos[0], pos[1], pos[2
], time, direction=direction, num_steps=unit_steps,
step_size=np.abs(scalar) / unit_steps, scalar=sign)
else:
scalar = scalar - last_min_dist / (2 * factor)
factor = factor + 1.0
scalar = scalar + last_min_dist / (2 * factor)
else:
scalar = scalar + min_dist / (2 * factor)
last_min_dist = min_dist.copy()
return scalar, pos_step, min_dist
| step_until_intersect | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(pos, field_line, sign, time, direction=None, step_size_goal=5.0,\n field_step_size=None):\n field_copy = field_line\n last_min_dist = 2500000.0\n scalar = 0.0\n repeat = True\n first = True\n factor = 1\n while repeat:\n unit_steps = np.abs(scalar // step_size_goal)\n if unit_steps == 0:\n unit_steps = 1\n pos_step = step_along_mag_unit_vector(pos[0], pos[1], pos[2], time,\n direction=direction, num_steps=unit_steps, step_size=np.abs(\n scalar) / unit_steps, scalar=sign)\n diff = field_copy - pos_step\n diff_mag = np.sqrt((diff ** 2).sum(axis=1))\n min_idx = np.argmin(diff_mag)\n if first:\n init = field_copy[min_idx, :]\n field_copy = full_field_line(init, time, 0.0, step_size=0.01,\n max_steps=int(field_step_size / 0.01), recurse=False)\n diff = field_copy - pos_step\n diff_mag = np.sqrt((diff ** 2).sum(axis=1))\n min_idx = np.argmin(diff_mag)\n first = False\n min_dist = diff_mag[min_idx]\n if min_dist > last_min_dist:\n if factor > 4:\n repeat = False\n scalar = scalar - last_min_dist / (2 * factor)\n pos_step = step_along_mag_unit_vector(pos[0], pos[1], pos[2\n ], time, direction=direction, num_steps=unit_steps,\n step_size=np.abs(scalar) / unit_steps, scalar=sign)\n else:\n scalar = scalar - last_min_dist / (2 * factor)\n factor = factor + 1.0\n scalar = scalar + last_min_dist / (2 * factor)\n else:\n scalar = scalar + min_dist / (2 * factor)\n last_min_dist = min_dist.copy()\n return scalar, pos_step, min_dist\n",
"role": "user"
},
{
"content": "step_until_intersect",
"role": "assistant"
}
] |
def x(x, y, z, date, direction=None, num_steps=5.0, step_size=5.0, scalar=1):
field_step_size = 100.0
field_max_steps = 1000
field_steps = np.arange(field_max_steps)
for i in np.arange(num_steps):
lat, lon, alt = ecef_to_geodetic(x, y, z)
zvx, zvy, zvz, bx, by, bz, mx, my, mz = (
calculate_mag_drift_unit_vectors_ecef([lat], [lon], [alt], [
date], steps=field_steps, max_steps=field_max_steps, step_size=
field_step_size, ref_height=0.0))
if direction == 'meridional':
ux, uy, uz = mx, my, mz
elif direction == 'zonal':
ux, uy, uz = zvx, zvy, zvz
elif direction == 'aligned':
ux, uy, uz = bx, by, bz
x = x + step_size * ux[0] * scalar
y = y + step_size * uy[0] * scalar
z = z + step_size * uz[0] * scalar
return np.array([x, y, z])
| step_along_mag_unit_vector | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(x, y, z, date, direction=None, num_steps=5.0, step_size=5.0, scalar=1):\n field_step_size = 100.0\n field_max_steps = 1000\n field_steps = np.arange(field_max_steps)\n for i in np.arange(num_steps):\n lat, lon, alt = ecef_to_geodetic(x, y, z)\n zvx, zvy, zvz, bx, by, bz, mx, my, mz = (\n calculate_mag_drift_unit_vectors_ecef([lat], [lon], [alt], [\n date], steps=field_steps, max_steps=field_max_steps, step_size=\n field_step_size, ref_height=0.0))\n if direction == 'meridional':\n ux, uy, uz = mx, my, mz\n elif direction == 'zonal':\n ux, uy, uz = zvx, zvy, zvz\n elif direction == 'aligned':\n ux, uy, uz = bx, by, bz\n x = x + step_size * ux[0] * scalar\n y = y + step_size * uy[0] * scalar\n z = z + step_size * uz[0] * scalar\n return np.array([x, y, z])\n",
"role": "user"
},
{
"content": "step_along_mag_unit_vector",
"role": "assistant"
}
] |
def x(glats, glons, alts, dates):
ecef_xs, ecef_ys, ecef_zs = geodetic_to_ecef(glats, glons, alts)
step_size = 100.0
max_steps = 1000
steps = np.arange(max_steps)
fine_step_size = 0.01
fine_max_steps = int(step_size / fine_step_size) + 10
fine_steps = np.arange(fine_max_steps)
out_x = []
out_y = []
out_z = []
for ecef_x, ecef_y, ecef_z, glat, glon, alt, date in zip(ecef_xs,
ecef_ys, ecef_zs, glats, glons, alts, dates):
trace = full_field_line(np.array([ecef_x, ecef_y, ecef_z]), date,
0.0, steps=steps, step_size=step_size, max_steps=max_steps)
tlat, tlon, talt = ecef_to_geodetic(trace[:, 0], trace[:, 1], trace
[:, 2])
max_idx = np.argmax(talt)
trace = full_field_line(trace[max_idx, :], date, 0.0, steps=
fine_steps, step_size=fine_step_size, max_steps=fine_max_steps,
recurse=False)
tlat, tlon, talt = ecef_to_geodetic(trace[:, 0], trace[:, 1], trace
[:, 2])
max_idx = np.argmax(talt)
out_x.append(trace[max_idx, 0])
out_y.append(trace[max_idx, 1])
out_z.append(trace[max_idx, 2])
out_x = np.array(out_x)
out_y = np.array(out_y)
out_z = np.array(out_z)
glat, glon, alt = ecef_to_geodetic(out_x, out_y, out_z)
return out_x, out_y, out_z, glat, glon, alt
| apex_location_info | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(glats, glons, alts, dates):\n ecef_xs, ecef_ys, ecef_zs = geodetic_to_ecef(glats, glons, alts)\n step_size = 100.0\n max_steps = 1000\n steps = np.arange(max_steps)\n fine_step_size = 0.01\n fine_max_steps = int(step_size / fine_step_size) + 10\n fine_steps = np.arange(fine_max_steps)\n out_x = []\n out_y = []\n out_z = []\n for ecef_x, ecef_y, ecef_z, glat, glon, alt, date in zip(ecef_xs,\n ecef_ys, ecef_zs, glats, glons, alts, dates):\n trace = full_field_line(np.array([ecef_x, ecef_y, ecef_z]), date, \n 0.0, steps=steps, step_size=step_size, max_steps=max_steps)\n tlat, tlon, talt = ecef_to_geodetic(trace[:, 0], trace[:, 1], trace\n [:, 2])\n max_idx = np.argmax(talt)\n trace = full_field_line(trace[max_idx, :], date, 0.0, steps=\n fine_steps, step_size=fine_step_size, max_steps=fine_max_steps,\n recurse=False)\n tlat, tlon, talt = ecef_to_geodetic(trace[:, 0], trace[:, 1], trace\n [:, 2])\n max_idx = np.argmax(talt)\n out_x.append(trace[max_idx, 0])\n out_y.append(trace[max_idx, 1])\n out_z.append(trace[max_idx, 2])\n out_x = np.array(out_x)\n out_y = np.array(out_y)\n out_z = np.array(out_z)\n glat, glon, alt = ecef_to_geodetic(out_x, out_y, out_z)\n return out_x, out_y, out_z, glat, glon, alt\n",
"role": "user"
},
{
"content": "apex_location_info",
"role": "assistant"
}
] |
def x(glats, glons, alts, dates, direction, vector_direction, step_size=
None, max_steps=None, edge_length=25.0, edge_steps=5):
if step_size is None:
step_size = 100.0
if max_steps is None:
max_steps = 1000
steps = np.arange(max_steps)
if direction == 'south':
direct = -1
elif direction == 'north':
direct = 1
ecef_xs, ecef_ys, ecef_zs = geodetic_to_ecef(glats, glons, alts)
full_local_step = []
min_distance_plus = []
min_distance_minus = []
for ecef_x, ecef_y, ecef_z, glat, glon, alt, date in zip(ecef_xs,
ecef_ys, ecef_zs, glats, glons, alts, dates):
yr, doy = pysat.utils.getyrdoy(date)
double_date = float(yr) + float(doy) / 366.0
sc_root = np.array([ecef_x, ecef_y, ecef_z])
trace = field_line_trace(sc_root, double_date, direct, 120.0, steps
=steps, step_size=step_size, max_steps=max_steps)
ftpnt = trace[-1, :]
ft_glat, ft_glon, ft_alt = ecef_to_geodetic(*ftpnt)
plus_step = step_along_mag_unit_vector(ftpnt[0], ftpnt[1], ftpnt[2],
date, direction=vector_direction, num_steps=edge_steps,
step_size=edge_length / edge_steps)
other_plus = field_line_trace(plus_step, double_date, -direct, 0.0,
steps=steps, step_size=step_size, max_steps=max_steps)
minus_step = step_along_mag_unit_vector(ftpnt[0], ftpnt[1], ftpnt[2
], date, direction=vector_direction, scalar=-1, num_steps=
edge_steps, step_size=edge_length / edge_steps)
other_minus = field_line_trace(minus_step, double_date, -direct,
0.0, steps=steps, step_size=step_size, max_steps=max_steps)
pos_edge_length, _, mind_pos = step_until_intersect(sc_root,
other_plus, 1, date, direction=vector_direction,
field_step_size=step_size, step_size_goal=edge_length / edge_steps)
minus_edge_length, _, mind_minus = step_until_intersect(sc_root,
other_minus, -1, date, direction=vector_direction,
field_step_size=step_size, step_size_goal=edge_length / edge_steps)
full_local_step.append(pos_edge_length + minus_edge_length)
min_distance_plus.append(mind_pos)
min_distance_minus.append(mind_minus)
return np.array(full_local_step), np.array(min_distance_plus), np.array(
min_distance_minus)
| closed_loop_edge_lengths_via_footpoint | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(glats, glons, alts, dates, direction, vector_direction, step_size=\n None, max_steps=None, edge_length=25.0, edge_steps=5):\n if step_size is None:\n step_size = 100.0\n if max_steps is None:\n max_steps = 1000\n steps = np.arange(max_steps)\n if direction == 'south':\n direct = -1\n elif direction == 'north':\n direct = 1\n ecef_xs, ecef_ys, ecef_zs = geodetic_to_ecef(glats, glons, alts)\n full_local_step = []\n min_distance_plus = []\n min_distance_minus = []\n for ecef_x, ecef_y, ecef_z, glat, glon, alt, date in zip(ecef_xs,\n ecef_ys, ecef_zs, glats, glons, alts, dates):\n yr, doy = pysat.utils.getyrdoy(date)\n double_date = float(yr) + float(doy) / 366.0\n sc_root = np.array([ecef_x, ecef_y, ecef_z])\n trace = field_line_trace(sc_root, double_date, direct, 120.0, steps\n =steps, step_size=step_size, max_steps=max_steps)\n ftpnt = trace[-1, :]\n ft_glat, ft_glon, ft_alt = ecef_to_geodetic(*ftpnt)\n plus_step = step_along_mag_unit_vector(ftpnt[0], ftpnt[1], ftpnt[2],\n date, direction=vector_direction, num_steps=edge_steps,\n step_size=edge_length / edge_steps)\n other_plus = field_line_trace(plus_step, double_date, -direct, 0.0,\n steps=steps, step_size=step_size, max_steps=max_steps)\n minus_step = step_along_mag_unit_vector(ftpnt[0], ftpnt[1], ftpnt[2\n ], date, direction=vector_direction, scalar=-1, num_steps=\n edge_steps, step_size=edge_length / edge_steps)\n other_minus = field_line_trace(minus_step, double_date, -direct, \n 0.0, steps=steps, step_size=step_size, max_steps=max_steps)\n pos_edge_length, _, mind_pos = step_until_intersect(sc_root,\n other_plus, 1, date, direction=vector_direction,\n field_step_size=step_size, step_size_goal=edge_length / edge_steps)\n minus_edge_length, _, mind_minus = step_until_intersect(sc_root,\n other_minus, -1, date, direction=vector_direction,\n field_step_size=step_size, step_size_goal=edge_length / edge_steps)\n full_local_step.append(pos_edge_length + minus_edge_length)\n min_distance_plus.append(mind_pos)\n min_distance_minus.append(mind_minus)\n return np.array(full_local_step), np.array(min_distance_plus), np.array(\n min_distance_minus)\n",
"role": "user"
},
{
"content": "closed_loop_edge_lengths_via_footpoint",
"role": "assistant"
}
] |
def x(glats, glons, alts, dates, vector_direction, edge_length=25.0,
edge_steps=5):
ecef_xs, ecef_ys, ecef_zs = geodetic_to_ecef(glats, glons, alts)
apex_edge_length = []
full_local_step = []
min_distance_plus = []
min_distance_minus = []
for ecef_x, ecef_y, ecef_z, glat, glon, alt, date in zip(ecef_xs,
ecef_ys, ecef_zs, glats, glons, alts, dates):
yr, doy = pysat.utils.getyrdoy(date)
double_date = float(yr) + float(doy) / 366.0
apex_x, apex_y, apex_z, apex_lat, apex_lon, apex_alt = (
apex_location_info([glat], [glon], [alt], [date]))
apex_root = np.array([apex_x[0], apex_y[0], apex_z[0]])
plus = step_along_mag_unit_vector(ecef_x, ecef_y, ecef_z, date,
direction=vector_direction, num_steps=edge_steps, step_size=
edge_length / edge_steps)
plus_lat, plus_lon, plus_alt = ecef_to_geodetic(*plus)
(plus_apex_x, plus_apex_y, plus_apex_z, plus_apex_lat,
plus_apex_lon, plus_apex_alt) = (apex_location_info([plus_lat],
[plus_lon], [plus_alt], [date]))
plus_apex_root = np.array([plus_apex_x[0], plus_apex_y[0],
plus_apex_z[0]])
minus = step_along_mag_unit_vector(ecef_x, ecef_y, ecef_z, date,
direction=vector_direction, scalar=-1, num_steps=edge_steps,
step_size=edge_length / edge_steps)
minus_lat, minus_lon, minus_alt = ecef_to_geodetic(*minus)
(minus_apex_x, minus_apex_y, minus_apex_z, minus_apex_lat,
minus_apex_lon, minus_apex_alt) = (apex_location_info([
minus_lat], [minus_lon], [minus_alt], [date]))
minus_apex_root = np.array([minus_apex_x[0], minus_apex_y[0],
minus_apex_z[0]])
apex_edge_length.append(np.sqrt((plus_apex_x[0] - minus_apex_x[0]) **
2 + (plus_apex_y[0] - minus_apex_y[0]) ** 2 + (plus_apex_z[0] -
minus_apex_z[0]) ** 2))
return np.array(apex_edge_length)
| closed_loop_edge_lengths_via_equator | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(glats, glons, alts, dates, vector_direction, edge_length=25.0,\n edge_steps=5):\n ecef_xs, ecef_ys, ecef_zs = geodetic_to_ecef(glats, glons, alts)\n apex_edge_length = []\n full_local_step = []\n min_distance_plus = []\n min_distance_minus = []\n for ecef_x, ecef_y, ecef_z, glat, glon, alt, date in zip(ecef_xs,\n ecef_ys, ecef_zs, glats, glons, alts, dates):\n yr, doy = pysat.utils.getyrdoy(date)\n double_date = float(yr) + float(doy) / 366.0\n apex_x, apex_y, apex_z, apex_lat, apex_lon, apex_alt = (\n apex_location_info([glat], [glon], [alt], [date]))\n apex_root = np.array([apex_x[0], apex_y[0], apex_z[0]])\n plus = step_along_mag_unit_vector(ecef_x, ecef_y, ecef_z, date,\n direction=vector_direction, num_steps=edge_steps, step_size=\n edge_length / edge_steps)\n plus_lat, plus_lon, plus_alt = ecef_to_geodetic(*plus)\n (plus_apex_x, plus_apex_y, plus_apex_z, plus_apex_lat,\n plus_apex_lon, plus_apex_alt) = (apex_location_info([plus_lat],\n [plus_lon], [plus_alt], [date]))\n plus_apex_root = np.array([plus_apex_x[0], plus_apex_y[0],\n plus_apex_z[0]])\n minus = step_along_mag_unit_vector(ecef_x, ecef_y, ecef_z, date,\n direction=vector_direction, scalar=-1, num_steps=edge_steps,\n step_size=edge_length / edge_steps)\n minus_lat, minus_lon, minus_alt = ecef_to_geodetic(*minus)\n (minus_apex_x, minus_apex_y, minus_apex_z, minus_apex_lat,\n minus_apex_lon, minus_apex_alt) = (apex_location_info([\n minus_lat], [minus_lon], [minus_alt], [date]))\n minus_apex_root = np.array([minus_apex_x[0], minus_apex_y[0],\n minus_apex_z[0]])\n apex_edge_length.append(np.sqrt((plus_apex_x[0] - minus_apex_x[0]) **\n 2 + (plus_apex_y[0] - minus_apex_y[0]) ** 2 + (plus_apex_z[0] -\n minus_apex_z[0]) ** 2))\n return np.array(apex_edge_length)\n",
"role": "user"
},
{
"content": "closed_loop_edge_lengths_via_equator",
"role": "assistant"
}
] |
def x(self):
if not self._ref:
raise ValueError('Object has no reference id for deletion')
if 'save' not in self._supports:
raise AssertionError('Can not save this object type')
response = self._session.delete(self._path)
if response.status_code == 200:
self._ref = None
self.clear()
return True
try:
error = response.json()
raise exceptions.ProtocolError(error['text'])
except ValueError:
raise exceptions.ProtocolError(response.content)
| delete | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if not self._ref:\n raise ValueError('Object has no reference id for deletion')\n if 'save' not in self._supports:\n raise AssertionError('Can not save this object type')\n response = self._session.delete(self._path)\n if response.status_code == 200:\n self._ref = None\n self.clear()\n return True\n try:\n error = response.json()\n raise exceptions.ProtocolError(error['text'])\n except ValueError:\n raise exceptions.ProtocolError(response.content)\n",
"role": "user"
},
{
"content": "delete",
"role": "assistant"
}
] |
def x(self):
LOGGER.debug('Fetching %s, %s', self._path, self._search_values)
response = self._session.get(self._path, self._search_values, {
'_return_fields': self._return_fields})
if response.status_code == 200:
values = response.json()
self._assign(values)
return bool(values)
elif response.status_code >= 400:
try:
error = response.json()
raise exceptions.ProtocolError(error['text'])
except ValueError:
raise exceptions.ProtocolError(response.content)
return False
| fetch | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n LOGGER.debug('Fetching %s, %s', self._path, self._search_values)\n response = self._session.get(self._path, self._search_values, {\n '_return_fields': self._return_fields})\n if response.status_code == 200:\n values = response.json()\n self._assign(values)\n return bool(values)\n elif response.status_code >= 400:\n try:\n error = response.json()\n raise exceptions.ProtocolError(error['text'])\n except ValueError:\n raise exceptions.ProtocolError(response.content)\n return False\n",
"role": "user"
},
{
"content": "fetch",
"role": "assistant"
}
] |
def x(self):
if 'save' not in self._supports:
raise AssertionError('Can not save this object type')
values = {}
for key in [key for key in self.keys() if key not in self._save_ignore]:
if not getattr(self, key) and getattr(self, key) != False:
continue
if isinstance(getattr(self, key, None), list):
value = list()
for item in getattr(self, key):
if isinstance(item, dict):
value.append(item)
elif hasattr(item, '_save_as'):
value.append(item._save_as())
elif hasattr(item, '_ref') and getattr(item, '_ref'):
value.append(getattr(item, '_ref'))
else:
LOGGER.warning('Cant assign %r', item)
values[key] = value
elif getattr(self, key, None):
values[key] = getattr(self, key)
if not self._ref:
response = self._session.post(self._path, values)
else:
values['_ref'] = self._ref
response = self._session.put(self._path, values)
LOGGER.debug('Response: %r, %r', response.status_code, response.content)
if 200 <= response.status_code <= 201:
self.fetch()
return True
else:
try:
error = response.json()
raise exceptions.ProtocolError(error['text'])
except ValueError:
raise exceptions.ProtocolError(response.content)
| save | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if 'save' not in self._supports:\n raise AssertionError('Can not save this object type')\n values = {}\n for key in [key for key in self.keys() if key not in self._save_ignore]:\n if not getattr(self, key) and getattr(self, key) != False:\n continue\n if isinstance(getattr(self, key, None), list):\n value = list()\n for item in getattr(self, key):\n if isinstance(item, dict):\n value.append(item)\n elif hasattr(item, '_save_as'):\n value.append(item._save_as())\n elif hasattr(item, '_ref') and getattr(item, '_ref'):\n value.append(getattr(item, '_ref'))\n else:\n LOGGER.warning('Cant assign %r', item)\n values[key] = value\n elif getattr(self, key, None):\n values[key] = getattr(self, key)\n if not self._ref:\n response = self._session.post(self._path, values)\n else:\n values['_ref'] = self._ref\n response = self._session.put(self._path, values)\n LOGGER.debug('Response: %r, %r', response.status_code, response.content)\n if 200 <= response.status_code <= 201:\n self.fetch()\n return True\n else:\n try:\n error = response.json()\n raise exceptions.ProtocolError(error['text'])\n except ValueError:\n raise exceptions.ProtocolError(response.content)\n",
"role": "user"
},
{
"content": "save",
"role": "assistant"
}
] |
def x(self, values):
LOGGER.debug('Assigning values: %r', values)
if not values:
return
keys = self.keys()
if not self._ref:
keys.append('_ref')
if isinstance(values, dict):
for key in keys:
if values.get(key):
if isinstance(values.get(key), list):
items = list()
for item in values[key]:
if isinstance(item, dict):
if '_ref' in item:
obj_class = get_class(item['_ref'])
if obj_class:
items.append(obj_class(self._session,
**item))
else:
items.append(item)
setattr(self, key, items)
else:
setattr(self, key, values[key])
elif isinstance(values, list):
self._assign(values[0])
else:
LOGGER.critical('Unhandled return type: %r', values)
| assign | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, values):\n LOGGER.debug('Assigning values: %r', values)\n if not values:\n return\n keys = self.keys()\n if not self._ref:\n keys.append('_ref')\n if isinstance(values, dict):\n for key in keys:\n if values.get(key):\n if isinstance(values.get(key), list):\n items = list()\n for item in values[key]:\n if isinstance(item, dict):\n if '_ref' in item:\n obj_class = get_class(item['_ref'])\n if obj_class:\n items.append(obj_class(self._session,\n **item))\n else:\n items.append(item)\n setattr(self, key, items)\n else:\n setattr(self, key, values[key])\n elif isinstance(values, list):\n self._assign(values[0])\n else:\n LOGGER.critical('Unhandled return type: %r', values)\n",
"role": "user"
},
{
"content": "assign",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.