source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(time):
if time.tzinfo is None:
time_utc = pytz.utc.localize(time)
else:
time_utc = time.astimezone(pytz.utc)
djd_start = pytz.utc.localize(dt.datetime(1899, 12, 31, 12))
djd = (time_utc - djd_start).total_seconds() * 1.0 / (60 * 60 * 24)
return djd
| datetime_to_djd | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(time):\n if time.tzinfo is None:\n time_utc = pytz.utc.localize(time)\n else:\n time_utc = time.astimezone(pytz.utc)\n djd_start = pytz.utc.localize(dt.datetime(1899, 12, 31, 12))\n djd = (time_utc - djd_start).total_seconds() * 1.0 / (60 * 60 * 24)\n return djd\n",
"role": "user"
},
{
"content": "datetime_to_djd",
"role": "assistant"
}
] |
def x(djd, tz='UTC'):
djd_start = pytz.utc.localize(dt.datetime(1899, 12, 31, 12))
utc_time = djd_start + dt.timedelta(days=djd)
return utc_time.astimezone(pytz.timezone(tz))
| djd_to_datetime | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(djd, tz='UTC'):\n djd_start = pytz.utc.localize(dt.datetime(1899, 12, 31, 12))\n utc_time = djd_start + dt.timedelta(days=djd)\n return utc_time.astimezone(pytz.timezone(tz))\n",
"role": "user"
},
{
"content": "djd_to_datetime",
"role": "assistant"
}
] |
def x(doy, epoch_year=2014):
doy = np.atleast_1d(doy).astype('float')
epoch = pd.Timestamp('{}-12-31'.format(epoch_year - 1))
timestamps = [(epoch + dt.timedelta(days=adoy)) for adoy in doy]
return pd.DatetimeIndex(timestamps)
| doy_to_datetimeindex | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(doy, epoch_year=2014):\n doy = np.atleast_1d(doy).astype('float')\n epoch = pd.Timestamp('{}-12-31'.format(epoch_year - 1))\n timestamps = [(epoch + dt.timedelta(days=adoy)) for adoy in doy]\n return pd.DatetimeIndex(timestamps)\n",
"role": "user"
},
{
"content": "doy_to_datetimeindex",
"role": "assistant"
}
] |
def x(keys, input_dict):
kwargs = {}
for key in keys:
try:
kwargs[key] = input_dict[key]
except KeyError:
pass
return kwargs
| build_kwargs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(keys, input_dict):\n kwargs = {}\n for key in keys:\n try:\n kwargs[key] = input_dict[key]\n except KeyError:\n pass\n return kwargs\n",
"role": "user"
},
{
"content": "build_kwargs",
"role": "assistant"
}
] |
def x(func, x0, fprime, args, tol, maxiter, fprime2, converged=False):
try:
p = np.asarray(x0, dtype=float)
except TypeError:
p = np.asarray(x0)
failures = np.ones_like(p, dtype=bool)
nz_der = np.copy(failures)
if fprime is not None:
for iteration in range(maxiter):
fval = np.asarray(func(p, *args))
if not fval.any():
failures = fval.astype(bool)
break
fder = np.asarray(fprime(p, *args))
nz_der = fder != 0
if not nz_der.any():
break
dp = fval[nz_der] / fder[nz_der]
if fprime2 is not None:
fder2 = np.asarray(fprime2(p, *args))
dp = dp / (1.0 - 0.5 * dp * fder2[nz_der] / fder[nz_der])
p[nz_der] -= dp
failures[nz_der] = np.abs(dp) >= tol
if not failures[nz_der].any():
break
else:
dx = np.finfo(float).eps ** 0.33
p1 = p * (1 + dx) + np.where(p >= 0, dx, -dx)
q0 = np.asarray(func(p, *args))
q1 = np.asarray(func(p1, *args))
active = np.ones_like(p, dtype=bool)
for iteration in range(maxiter):
nz_der = q1 != q0
if not nz_der.any():
p = (p1 + p) / 2.0
break
dp = (q1 * (p1 - p))[nz_der] / (q1 - q0)[nz_der]
p[nz_der] = p1[nz_der] - dp
active_zero_der = ~nz_der & active
p[active_zero_der] = (p1 + p)[active_zero_der] / 2.0
active &= nz_der
failures[nz_der] = np.abs(dp) >= tol
if not failures[nz_der].any():
break
p1, p = p, p1
q0 = q1
q1 = np.asarray(func(p1, *args))
zero_der = ~nz_der & failures
if zero_der.any():
if fprime is None:
nonzero_dp = p1 != p
zero_der_nz_dp = zero_der & nonzero_dp
if zero_der_nz_dp.any():
rms = np.sqrt(sum((p1[zero_der_nz_dp] - p[zero_der_nz_dp]) **
2))
warnings.warn('RMS of {:g} reached'.format(rms), RuntimeWarning
)
else:
all_or_some = 'all' if zero_der.all() else 'some'
msg = '{:s} derivatives were zero'.format(all_or_some)
warnings.warn(msg, RuntimeWarning)
elif failures.any():
all_or_some = 'all' if failures.all() else 'some'
msg = '{0:s} failed to converge after {1:d} iterations'.format(
all_or_some, maxiter)
if failures.all():
raise RuntimeError(msg)
warnings.warn(msg, RuntimeWarning)
if converged:
result = namedtuple('result', ('root', 'converged', 'zero_der'))
p = result(p, ~failures, zero_der)
return p
| array_newton | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(func, x0, fprime, args, tol, maxiter, fprime2, converged=False):\n try:\n p = np.asarray(x0, dtype=float)\n except TypeError:\n p = np.asarray(x0)\n failures = np.ones_like(p, dtype=bool)\n nz_der = np.copy(failures)\n if fprime is not None:\n for iteration in range(maxiter):\n fval = np.asarray(func(p, *args))\n if not fval.any():\n failures = fval.astype(bool)\n break\n fder = np.asarray(fprime(p, *args))\n nz_der = fder != 0\n if not nz_der.any():\n break\n dp = fval[nz_der] / fder[nz_der]\n if fprime2 is not None:\n fder2 = np.asarray(fprime2(p, *args))\n dp = dp / (1.0 - 0.5 * dp * fder2[nz_der] / fder[nz_der])\n p[nz_der] -= dp\n failures[nz_der] = np.abs(dp) >= tol\n if not failures[nz_der].any():\n break\n else:\n dx = np.finfo(float).eps ** 0.33\n p1 = p * (1 + dx) + np.where(p >= 0, dx, -dx)\n q0 = np.asarray(func(p, *args))\n q1 = np.asarray(func(p1, *args))\n active = np.ones_like(p, dtype=bool)\n for iteration in range(maxiter):\n nz_der = q1 != q0\n if not nz_der.any():\n p = (p1 + p) / 2.0\n break\n dp = (q1 * (p1 - p))[nz_der] / (q1 - q0)[nz_der]\n p[nz_der] = p1[nz_der] - dp\n active_zero_der = ~nz_der & active\n p[active_zero_der] = (p1 + p)[active_zero_der] / 2.0\n active &= nz_der\n failures[nz_der] = np.abs(dp) >= tol\n if not failures[nz_der].any():\n break\n p1, p = p, p1\n q0 = q1\n q1 = np.asarray(func(p1, *args))\n zero_der = ~nz_der & failures\n if zero_der.any():\n if fprime is None:\n nonzero_dp = p1 != p\n zero_der_nz_dp = zero_der & nonzero_dp\n if zero_der_nz_dp.any():\n rms = np.sqrt(sum((p1[zero_der_nz_dp] - p[zero_der_nz_dp]) **\n 2))\n warnings.warn('RMS of {:g} reached'.format(rms), RuntimeWarning\n )\n else:\n all_or_some = 'all' if zero_der.all() else 'some'\n msg = '{:s} derivatives were zero'.format(all_or_some)\n warnings.warn(msg, RuntimeWarning)\n elif failures.any():\n all_or_some = 'all' if failures.all() else 'some'\n msg = '{0:s} failed to converge after {1:d} iterations'.format(\n all_or_some, maxiter)\n if failures.all():\n raise RuntimeError(msg)\n warnings.warn(msg, RuntimeWarning)\n if converged:\n result = namedtuple('result', ('root', 'converged', 'zero_der'))\n p = result(p, ~failures, zero_der)\n return p\n",
"role": "user"
},
{
"content": "array_newton",
"role": "assistant"
}
] |
def x(params, VL, VH, func):
df = params
df['VH'] = VH
df['VL'] = VL
errflag = True
iterations = 0
while errflag:
phi = (np.sqrt(5) - 1) / 2 * (df['VH'] - df['VL'])
df['V1'] = df['VL'] + phi
df['V2'] = df['VH'] - phi
df['f1'] = func(df, 'V1')
df['f2'] = func(df, 'V2')
df['SW_Flag'] = df['f1'] > df['f2']
df['VL'] = df['V2'] * df['SW_Flag'] + df['VL'] * ~df['SW_Flag']
df['VH'] = df['V1'] * ~df['SW_Flag'] + df['VH'] * df['SW_Flag']
err = df['V1'] - df['V2']
try:
errflag = (abs(err) > 0.01).any()
except ValueError:
errflag = abs(err) > 0.01
iterations += 1
if iterations > 50:
raise Exception('EXCEPTION:iterations exceeded maximum (50)')
return func(df, 'V1'), df['V1']
| golden_sect_DataFrame | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(params, VL, VH, func):\n df = params\n df['VH'] = VH\n df['VL'] = VL\n errflag = True\n iterations = 0\n while errflag:\n phi = (np.sqrt(5) - 1) / 2 * (df['VH'] - df['VL'])\n df['V1'] = df['VL'] + phi\n df['V2'] = df['VH'] - phi\n df['f1'] = func(df, 'V1')\n df['f2'] = func(df, 'V2')\n df['SW_Flag'] = df['f1'] > df['f2']\n df['VL'] = df['V2'] * df['SW_Flag'] + df['VL'] * ~df['SW_Flag']\n df['VH'] = df['V1'] * ~df['SW_Flag'] + df['VH'] * df['SW_Flag']\n err = df['V1'] - df['V2']\n try:\n errflag = (abs(err) > 0.01).any()\n except ValueError:\n errflag = abs(err) > 0.01\n iterations += 1\n if iterations > 50:\n raise Exception('EXCEPTION:iterations exceeded maximum (50)')\n return func(df, 'V1'), df['V1']\n",
"role": "user"
},
{
"content": "golden_sect_DataFrame",
"role": "assistant"
}
] |
def x(apparent_zenith, apparent_azimuth, axis_tilt=0, axis_azimuth=0,
max_angle=90, backtrack=True, gcr=2.0 / 7.0):
if isinstance(apparent_zenith, pd.Series):
index = apparent_zenith.index
else:
index = None
apparent_azimuth = np.atleast_1d(apparent_azimuth)
apparent_zenith = np.atleast_1d(apparent_zenith)
if apparent_azimuth.ndim > 1 or apparent_zenith.ndim > 1:
raise ValueError('Input dimensions must not exceed 1')
az = apparent_azimuth - 180
apparent_elevation = 90 - apparent_zenith
x = cosd(apparent_elevation) * sind(az)
y = cosd(apparent_elevation) * cosd(az)
z = sind(apparent_elevation)
axis_azimuth_south = axis_azimuth - 180
xp = x * cosd(axis_azimuth_south) - y * sind(axis_azimuth_south)
yp = x * cosd(axis_tilt) * sind(axis_azimuth_south) + y * cosd(axis_tilt
) * cosd(axis_azimuth_south) - z * sind(axis_tilt)
zp = x * sind(axis_tilt) * sind(axis_azimuth_south) + y * sind(axis_tilt
) * cosd(axis_azimuth_south) + z * cosd(axis_tilt)
wid = 90 - np.degrees(np.arctan2(zp, xp))
zen_gt_90 = apparent_zenith > 90
wid[zen_gt_90] = np.nan
if backtrack:
axes_distance = 1 / gcr
temp = np.minimum(axes_distance * cosd(wid), 1)
wc = np.degrees(np.arccos(temp))
tracker_theta = np.where(wid < 0, wid + wc, wid - wc)
else:
tracker_theta = wid
tracker_theta[tracker_theta > max_angle] = max_angle
tracker_theta[tracker_theta < -max_angle] = -max_angle
panel_norm = np.array([sind(tracker_theta), tracker_theta * 0, cosd(
tracker_theta)])
sun_vec = np.array([xp, yp, zp])
aoi = np.degrees(np.arccos(np.abs(np.sum(sun_vec * panel_norm, axis=0))))
rot_x = np.array([[1, 0, 0], [0, cosd(-axis_tilt), -sind(-axis_tilt)],
[0, sind(-axis_tilt), cosd(-axis_tilt)]])
panel_norm_earth = np.dot(rot_x, panel_norm).T
projected_normal = np.array([panel_norm_earth[:, 0], panel_norm_earth[:,
1], panel_norm_earth[:, 2] * 0]).T
projected_normal_mag = np.sqrt(np.nansum(projected_normal ** 2, axis=1))
non_zeros = projected_normal_mag != 0
projected_normal[non_zeros] = (projected_normal[non_zeros].T /
projected_normal_mag[non_zeros]).T
surface_azimuth = np.degrees(np.arctan2(projected_normal[:, 1],
projected_normal[:, 0]))
surface_azimuth = 90 - surface_azimuth + axis_azimuth
surface_azimuth[surface_azimuth < 0] += 360
surface_azimuth[surface_azimuth >= 360] -= 360
dotproduct = (panel_norm_earth * projected_normal).sum(axis=1)
surface_tilt = 90 - np.degrees(np.arccos(dotproduct))
out = {'tracker_theta': tracker_theta, 'aoi': aoi, 'surface_azimuth':
surface_azimuth, 'surface_tilt': surface_tilt}
if index is not None:
out = pd.DataFrame(out, index=index)
out = out[['tracker_theta', 'aoi', 'surface_azimuth', 'surface_tilt']]
out[zen_gt_90] = np.nan
else:
out = {k: np.where(zen_gt_90, np.nan, v) for k, v in out.items()}
return out
| singleaxis | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(apparent_zenith, apparent_azimuth, axis_tilt=0, axis_azimuth=0,\n max_angle=90, backtrack=True, gcr=2.0 / 7.0):\n if isinstance(apparent_zenith, pd.Series):\n index = apparent_zenith.index\n else:\n index = None\n apparent_azimuth = np.atleast_1d(apparent_azimuth)\n apparent_zenith = np.atleast_1d(apparent_zenith)\n if apparent_azimuth.ndim > 1 or apparent_zenith.ndim > 1:\n raise ValueError('Input dimensions must not exceed 1')\n az = apparent_azimuth - 180\n apparent_elevation = 90 - apparent_zenith\n x = cosd(apparent_elevation) * sind(az)\n y = cosd(apparent_elevation) * cosd(az)\n z = sind(apparent_elevation)\n axis_azimuth_south = axis_azimuth - 180\n xp = x * cosd(axis_azimuth_south) - y * sind(axis_azimuth_south)\n yp = x * cosd(axis_tilt) * sind(axis_azimuth_south) + y * cosd(axis_tilt\n ) * cosd(axis_azimuth_south) - z * sind(axis_tilt)\n zp = x * sind(axis_tilt) * sind(axis_azimuth_south) + y * sind(axis_tilt\n ) * cosd(axis_azimuth_south) + z * cosd(axis_tilt)\n wid = 90 - np.degrees(np.arctan2(zp, xp))\n zen_gt_90 = apparent_zenith > 90\n wid[zen_gt_90] = np.nan\n if backtrack:\n axes_distance = 1 / gcr\n temp = np.minimum(axes_distance * cosd(wid), 1)\n wc = np.degrees(np.arccos(temp))\n tracker_theta = np.where(wid < 0, wid + wc, wid - wc)\n else:\n tracker_theta = wid\n tracker_theta[tracker_theta > max_angle] = max_angle\n tracker_theta[tracker_theta < -max_angle] = -max_angle\n panel_norm = np.array([sind(tracker_theta), tracker_theta * 0, cosd(\n tracker_theta)])\n sun_vec = np.array([xp, yp, zp])\n aoi = np.degrees(np.arccos(np.abs(np.sum(sun_vec * panel_norm, axis=0))))\n rot_x = np.array([[1, 0, 0], [0, cosd(-axis_tilt), -sind(-axis_tilt)],\n [0, sind(-axis_tilt), cosd(-axis_tilt)]])\n panel_norm_earth = np.dot(rot_x, panel_norm).T\n projected_normal = np.array([panel_norm_earth[:, 0], panel_norm_earth[:,\n 1], panel_norm_earth[:, 2] * 0]).T\n projected_normal_mag = np.sqrt(np.nansum(projected_normal ** 2, axis=1))\n non_zeros = projected_normal_mag != 0\n projected_normal[non_zeros] = (projected_normal[non_zeros].T /\n projected_normal_mag[non_zeros]).T\n surface_azimuth = np.degrees(np.arctan2(projected_normal[:, 1],\n projected_normal[:, 0]))\n surface_azimuth = 90 - surface_azimuth + axis_azimuth\n surface_azimuth[surface_azimuth < 0] += 360\n surface_azimuth[surface_azimuth >= 360] -= 360\n dotproduct = (panel_norm_earth * projected_normal).sum(axis=1)\n surface_tilt = 90 - np.degrees(np.arccos(dotproduct))\n out = {'tracker_theta': tracker_theta, 'aoi': aoi, 'surface_azimuth':\n surface_azimuth, 'surface_tilt': surface_tilt}\n if index is not None:\n out = pd.DataFrame(out, index=index)\n out = out[['tracker_theta', 'aoi', 'surface_azimuth', 'surface_tilt']]\n out[zen_gt_90] = np.nan\n else:\n out = {k: np.where(zen_gt_90, np.nan, v) for k, v in out.items()}\n return out\n",
"role": "user"
},
{
"content": "singleaxis",
"role": "assistant"
}
] |
def x(self, apparent_zenith, apparent_azimuth):
tracking_data = singleaxis(apparent_zenith, apparent_azimuth, self.
axis_tilt, self.axis_azimuth, self.max_angle, self.backtrack, self.gcr)
return tracking_data
| singleaxis | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, apparent_zenith, apparent_azimuth):\n tracking_data = singleaxis(apparent_zenith, apparent_azimuth, self.\n axis_tilt, self.axis_azimuth, self.max_angle, self.backtrack, self.gcr)\n return tracking_data\n",
"role": "user"
},
{
"content": "singleaxis",
"role": "assistant"
}
] |
def x(self, surface_tilt, surface_azimuth, solar_zenith, solar_azimuth, dni,
ghi, dhi, dni_extra=None, airmass=None, model='haydavies', **kwargs):
if dni_extra is None:
dni_extra = irradiance.get_extra_radiation(solar_zenith.index)
if airmass is None:
airmass = atmosphere.get_relative_airmass(solar_zenith)
return irradiance.get_total_irradiance(surface_tilt, surface_azimuth,
solar_zenith, solar_azimuth, dni, ghi, dhi, dni_extra=dni_extra,
airmass=airmass, model=model, albedo=self.albedo, **kwargs)
| get_irradiance | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, surface_tilt, surface_azimuth, solar_zenith, solar_azimuth, dni,\n ghi, dhi, dni_extra=None, airmass=None, model='haydavies', **kwargs):\n if dni_extra is None:\n dni_extra = irradiance.get_extra_radiation(solar_zenith.index)\n if airmass is None:\n airmass = atmosphere.get_relative_airmass(solar_zenith)\n return irradiance.get_total_irradiance(surface_tilt, surface_azimuth,\n solar_zenith, solar_azimuth, dni, ghi, dhi, dni_extra=dni_extra,\n airmass=airmass, model=model, albedo=self.albedo, **kwargs)\n",
"role": "user"
},
{
"content": "get_irradiance",
"role": "assistant"
}
] |
def x(filename):
tsv_data = pd.read_csv(filename, delimiter='\t')
data = format_index(tsv_data)
data = data[data.columns[2:]]
data = data.rename(columns=map_columns)
columns = data.columns
flag_label_map = {flag: (columns[columns.get_loc(flag) - 1] + '_flag') for
flag in columns[1::2]}
data = data.rename(columns=flag_label_map)
for col in columns[::2]:
missing = data[col + '_flag'] == 99
data[col] = data[col].where(~missing, np.NaN)
return data
| read_srml | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename):\n tsv_data = pd.read_csv(filename, delimiter='\\t')\n data = format_index(tsv_data)\n data = data[data.columns[2:]]\n data = data.rename(columns=map_columns)\n columns = data.columns\n flag_label_map = {flag: (columns[columns.get_loc(flag) - 1] + '_flag') for\n flag in columns[1::2]}\n data = data.rename(columns=flag_label_map)\n for col in columns[::2]:\n missing = data[col + '_flag'] == 99\n data[col] = data[col].where(~missing, np.NaN)\n return data\n",
"role": "user"
},
{
"content": "read_srml",
"role": "assistant"
}
] |
def x(col):
if col.startswith('7'):
try:
return VARIABLE_MAP[col]
except KeyError:
return col
try:
variable_name = VARIABLE_MAP[col[:3]]
variable_number = col[3:]
return variable_name + '_' + variable_number
except KeyError:
return col
| map_columns | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(col):\n if col.startswith('7'):\n try:\n return VARIABLE_MAP[col]\n except KeyError:\n return col\n try:\n variable_name = VARIABLE_MAP[col[:3]]\n variable_number = col[3:]\n return variable_name + '_' + variable_number\n except KeyError:\n return col\n",
"role": "user"
},
{
"content": "map_columns",
"role": "assistant"
}
] |
def x(df):
year = int(df.columns[1])
df_doy = df[df.columns[0]]
df_time = df[df.columns[1]] - 1
fifty_nines = df_time % 100 == 99
times = df_time.where(~fifty_nines, df_time - 40)
times = times.apply(lambda x: '{:04.0f}'.format(x))
doy = df_doy.apply(lambda x: '{:03.0f}'.format(x))
dts = pd.to_datetime(str(year) + '-' + doy + '-' + times, format=
'%Y-%j-%H%M')
df.index = dts
df = df.tz_localize('Etc/GMT+8')
return df
| format_index | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(df):\n year = int(df.columns[1])\n df_doy = df[df.columns[0]]\n df_time = df[df.columns[1]] - 1\n fifty_nines = df_time % 100 == 99\n times = df_time.where(~fifty_nines, df_time - 40)\n times = times.apply(lambda x: '{:04.0f}'.format(x))\n doy = df_doy.apply(lambda x: '{:03.0f}'.format(x))\n dts = pd.to_datetime(str(year) + '-' + doy + '-' + times, format=\n '%Y-%j-%H%M')\n df.index = dts\n df = df.tz_localize('Etc/GMT+8')\n return df\n",
"role": "user"
},
{
"content": "format_index",
"role": "assistant"
}
] |
def x(station, year, month, filetype='PO'):
file_name = '{station}{filetype}{year:02d}{month:02d}.txt'.format(station
=station, filetype=filetype, year=year % 100, month=month)
url = 'http://solardat.uoregon.edu/download/Archive/'
data = read_srml(url + file_name)
return data
| read_srml_month_from_solardat | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(station, year, month, filetype='PO'):\n file_name = '{station}{filetype}{year:02d}{month:02d}.txt'.format(station\n =station, filetype=filetype, year=year % 100, month=month)\n url = 'http://solardat.uoregon.edu/download/Archive/'\n data = read_srml(url + file_name)\n return data\n",
"role": "user"
},
{
"content": "read_srml_month_from_solardat",
"role": "assistant"
}
] |
def x(filename, params, startdate, stopdate, lookup_params=True, server=
None, target=_ecmwf):
if not filename.endswith('nc'):
filename += '.nc'
if lookup_params:
try:
params = '/'.join(PARAMS.get(p) for p in params)
except TypeError:
params = PARAMS.get(params)
startdate = startdate.strftime('%Y-%m-%d')
stopdate = stopdate.strftime('%Y-%m-%d')
if not server:
server = ECMWFDataServer()
t = threading.Thread(target=target, daemon=True, args=(server,
startdate, stopdate, params, filename))
t.start()
return t
| get_ecmwf_macc | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename, params, startdate, stopdate, lookup_params=True, server=\n None, target=_ecmwf):\n if not filename.endswith('nc'):\n filename += '.nc'\n if lookup_params:\n try:\n params = '/'.join(PARAMS.get(p) for p in params)\n except TypeError:\n params = PARAMS.get(params)\n startdate = startdate.strftime('%Y-%m-%d')\n stopdate = stopdate.strftime('%Y-%m-%d')\n if not server:\n server = ECMWFDataServer()\n t = threading.Thread(target=target, daemon=True, args=(server,\n startdate, stopdate, params, filename))\n t.start()\n return t\n",
"role": "user"
},
{
"content": "get_ecmwf_macc",
"role": "assistant"
}
] |
def x(filename, latitude, longitude, utc_time_range=None):
ecmwf_macc = ECMWF_MACC(filename)
try:
ilat, ilon = ecmwf_macc.get_nearest_indices(latitude, longitude)
nctime = ecmwf_macc.data['time']
if utc_time_range:
start_idx = netCDF4.date2index(utc_time_range[0], nctime,
select='before')
stop_idx = netCDF4.date2index(utc_time_range[-1], nctime,
select='after')
time_slice = slice(start_idx, stop_idx + 1)
else:
time_slice = slice(0, ecmwf_macc.time_size)
times = netCDF4.num2date(nctime[time_slice], nctime.units)
df = {k: ecmwf_macc.data[k][time_slice, ilat, ilon] for k in
ecmwf_macc.keys}
if ECMWF_MACC.TCWV in df:
df['precipitable_water'] = df[ECMWF_MACC.TCWV] / 10.0
finally:
ecmwf_macc.data.close()
return pd.DataFrame(df, index=times.astype('datetime64[s]'))
| read_ecmwf_macc | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename, latitude, longitude, utc_time_range=None):\n ecmwf_macc = ECMWF_MACC(filename)\n try:\n ilat, ilon = ecmwf_macc.get_nearest_indices(latitude, longitude)\n nctime = ecmwf_macc.data['time']\n if utc_time_range:\n start_idx = netCDF4.date2index(utc_time_range[0], nctime,\n select='before')\n stop_idx = netCDF4.date2index(utc_time_range[-1], nctime,\n select='after')\n time_slice = slice(start_idx, stop_idx + 1)\n else:\n time_slice = slice(0, ecmwf_macc.time_size)\n times = netCDF4.num2date(nctime[time_slice], nctime.units)\n df = {k: ecmwf_macc.data[k][time_slice, ilat, ilon] for k in\n ecmwf_macc.keys}\n if ECMWF_MACC.TCWV in df:\n df['precipitable_water'] = df[ECMWF_MACC.TCWV] / 10.0\n finally:\n ecmwf_macc.data.close()\n return pd.DataFrame(df, index=times.astype('datetime64[s]'))\n",
"role": "user"
},
{
"content": "read_ecmwf_macc",
"role": "assistant"
}
] |
def x(self, latitude, longitude):
idx_lat = int(round((latitude - 90.0) / self.delta_lat))
if idx_lat < 0:
idx_lat = 0
elif idx_lat > self.lat_size:
idx_lat = self.lat_size
longitude = longitude % 360.0
idx_lon = int(round(longitude / self.delta_lon)) % self.lon_size
return idx_lat, idx_lon
| get_nearest_indices | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, latitude, longitude):\n idx_lat = int(round((latitude - 90.0) / self.delta_lat))\n if idx_lat < 0:\n idx_lat = 0\n elif idx_lat > self.lat_size:\n idx_lat = self.lat_size\n longitude = longitude % 360.0\n idx_lon = int(round(longitude / self.delta_lon)) % self.lon_size\n return idx_lat, idx_lon\n",
"role": "user"
},
{
"content": "get_nearest_indices",
"role": "assistant"
}
] |
def x(self, latitude, longitude, utc_time, param):
nctime = self.data['time']
ilat, ilon = self.get_nearest_indices(latitude, longitude)
before = netCDF4.date2index(utc_time, nctime, select='before')
fbefore = self.data[param][before, ilat, ilon]
fafter = self.data[param][before + 1, ilat, ilon]
dt_num = netCDF4.date2num(utc_time, nctime.units)
time_ratio = (dt_num - nctime[before]) / self.delta_time
return fbefore + (fafter - fbefore) * time_ratio
| interp_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, latitude, longitude, utc_time, param):\n nctime = self.data['time']\n ilat, ilon = self.get_nearest_indices(latitude, longitude)\n before = netCDF4.date2index(utc_time, nctime, select='before')\n fbefore = self.data[param][before, ilat, ilon]\n fafter = self.data[param][before + 1, ilat, ilon]\n dt_num = netCDF4.date2num(utc_time, nctime.units)\n time_ratio = (dt_num - nctime[before]) / self.delta_time\n return fbefore + (fafter - fbefore) * time_ratio\n",
"role": "user"
},
{
"content": "interp_data",
"role": "assistant"
}
] |
def x(filename):
if 'msn' in filename:
names = MADISON_HEADERS
widths = MADISON_WIDTHS
dtypes = MADISON_DTYPES
else:
names = HEADERS
widths = WIDTHS
dtypes = DTYPES
data = pd.read_fwf(filename, header=None, skiprows=2, names=names,
widths=widths, na_values=-9999.9)
for col, _dtype in zip(data.columns, dtypes):
ser = data[col].astype(_dtype)
if _dtype == 'float64':
ser = ser.where(ser > -9999, other=np.nan)
data[col] = ser
dts = data[['month', 'day', 'hour', 'minute']].astype(str).apply(lambda
x: x.str.zfill(2))
dtindex = pd.to_datetime(data['year'].astype(str) + dts['month'] + dts[
'day'] + dts['hour'] + dts['minute'], format='%Y%m%d%H%M', utc=True)
data = data.set_index(dtindex)
try:
data = data.tz_localize('UTC')
except TypeError:
pass
return data
| read_solrad | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename):\n if 'msn' in filename:\n names = MADISON_HEADERS\n widths = MADISON_WIDTHS\n dtypes = MADISON_DTYPES\n else:\n names = HEADERS\n widths = WIDTHS\n dtypes = DTYPES\n data = pd.read_fwf(filename, header=None, skiprows=2, names=names,\n widths=widths, na_values=-9999.9)\n for col, _dtype in zip(data.columns, dtypes):\n ser = data[col].astype(_dtype)\n if _dtype == 'float64':\n ser = ser.where(ser > -9999, other=np.nan)\n data[col] = ser\n dts = data[['month', 'day', 'hour', 'minute']].astype(str).apply(lambda\n x: x.str.zfill(2))\n dtindex = pd.to_datetime(data['year'].astype(str) + dts['month'] + dts[\n 'day'] + dts['hour'] + dts['minute'], format='%Y%m%d%H%M', utc=True)\n data = data.set_index(dtindex)\n try:\n data = data.tz_localize('UTC')\n except TypeError:\n pass\n return data\n",
"role": "user"
},
{
"content": "read_solrad",
"role": "assistant"
}
] |
def x(filename):
data = pd.read_fwf(filename, header=None, names=HEADERS.split(' '),
widths=WIDTHS)
for col, _dtype in zip(data.columns, DTYPES):
data[col] = data[col].astype(_dtype)
dts = data[['UTC_DATE', 'UTC_TIME']].astype(str)
dtindex = pd.to_datetime(dts['UTC_DATE'] + dts['UTC_TIME'].str.zfill(4),
format='%Y%m%d%H%M', utc=True)
data = data.set_index(dtindex)
try:
data = data.tz_localize('UTC')
except TypeError:
pass
for val in [-99, -999, -9999]:
data = data.where(data != val, np.nan)
data = data.rename(columns=VARIABLE_MAP)
return data
| read_crn | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename):\n data = pd.read_fwf(filename, header=None, names=HEADERS.split(' '),\n widths=WIDTHS)\n for col, _dtype in zip(data.columns, DTYPES):\n data[col] = data[col].astype(_dtype)\n dts = data[['UTC_DATE', 'UTC_TIME']].astype(str)\n dtindex = pd.to_datetime(dts['UTC_DATE'] + dts['UTC_TIME'].str.zfill(4),\n format='%Y%m%d%H%M', utc=True)\n data = data.set_index(dtindex)\n try:\n data = data.tz_localize('UTC')\n except TypeError:\n pass\n for val in [-99, -999, -9999]:\n data = data.where(data != val, np.nan)\n data = data.rename(columns=VARIABLE_MAP)\n return data\n",
"role": "user"
},
{
"content": "read_crn",
"role": "assistant"
}
] |
def x(filename, coerce_year=None):
if filename.startswith('http'):
request = Request(filename, headers={'User-Agent':
'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_13_5) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/67.0.3396.87 Safari/537.36'
})
response = urlopen(request)
csvdata = io.StringIO(response.read().decode(errors='ignore'))
else:
csvdata = open(filename, 'r')
firstline = csvdata.readline()
head = ['loc', 'city', 'state-prov', 'country', 'data_type', 'WMO_code',
'latitude', 'longitude', 'TZ', 'altitude']
meta = dict(zip(head, firstline.rstrip('\n').split(',')))
meta['altitude'] = float(meta['altitude'])
meta['latitude'] = float(meta['latitude'])
meta['longitude'] = float(meta['longitude'])
meta['TZ'] = float(meta['TZ'])
colnames = ['year', 'month', 'day', 'hour', 'minute',
'data_source_unct', 'temp_air', 'temp_dew', 'relative_humidity',
'atmospheric_pressure', 'etr', 'etrn', 'ghi_infrared', 'ghi', 'dni',
'dhi', 'global_hor_illum', 'direct_normal_illum',
'diffuse_horizontal_illum', 'zenith_luminance', 'wind_direction',
'wind_speed', 'total_sky_cover', 'opaque_sky_cover', 'visibility',
'ceiling_height', 'present_weather_observation',
'present_weather_codes', 'precipitable_water',
'aerosol_optical_depth', 'snow_depth', 'days_since_last_snowfall',
'albedo', 'liquid_precipitation_depth', 'liquid_precipitation_quantity'
]
data = pd.read_csv(csvdata, skiprows=6, header=0, names=colnames)
if coerce_year is not None:
data['year'] = coerce_year
dts = data[['month', 'day']].astype(str).apply(lambda x: x.str.zfill(2))
hrs = (data['hour'] - 1).astype(str).str.zfill(2)
dtscat = data['year'].astype(str) + dts['month'] + dts['day'] + hrs
idx = pd.to_datetime(dtscat, format='%Y%m%d%H')
idx = idx.dt.tz_localize(int(meta['TZ'] * 3600))
data.index = idx
return data, meta
| read_epw | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename, coerce_year=None):\n if filename.startswith('http'):\n request = Request(filename, headers={'User-Agent':\n 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_13_5) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/67.0.3396.87 Safari/537.36'\n })\n response = urlopen(request)\n csvdata = io.StringIO(response.read().decode(errors='ignore'))\n else:\n csvdata = open(filename, 'r')\n firstline = csvdata.readline()\n head = ['loc', 'city', 'state-prov', 'country', 'data_type', 'WMO_code',\n 'latitude', 'longitude', 'TZ', 'altitude']\n meta = dict(zip(head, firstline.rstrip('\\n').split(',')))\n meta['altitude'] = float(meta['altitude'])\n meta['latitude'] = float(meta['latitude'])\n meta['longitude'] = float(meta['longitude'])\n meta['TZ'] = float(meta['TZ'])\n colnames = ['year', 'month', 'day', 'hour', 'minute',\n 'data_source_unct', 'temp_air', 'temp_dew', 'relative_humidity',\n 'atmospheric_pressure', 'etr', 'etrn', 'ghi_infrared', 'ghi', 'dni',\n 'dhi', 'global_hor_illum', 'direct_normal_illum',\n 'diffuse_horizontal_illum', 'zenith_luminance', 'wind_direction',\n 'wind_speed', 'total_sky_cover', 'opaque_sky_cover', 'visibility',\n 'ceiling_height', 'present_weather_observation',\n 'present_weather_codes', 'precipitable_water',\n 'aerosol_optical_depth', 'snow_depth', 'days_since_last_snowfall',\n 'albedo', 'liquid_precipitation_depth', 'liquid_precipitation_quantity'\n ]\n data = pd.read_csv(csvdata, skiprows=6, header=0, names=colnames)\n if coerce_year is not None:\n data['year'] = coerce_year\n dts = data[['month', 'day']].astype(str).apply(lambda x: x.str.zfill(2))\n hrs = (data['hour'] - 1).astype(str).str.zfill(2)\n dtscat = data['year'].astype(str) + dts['month'] + dts['day'] + hrs\n idx = pd.to_datetime(dtscat, format='%Y%m%d%H')\n idx = idx.dt.tz_localize(int(meta['TZ'] * 3600))\n data.index = idx\n return data, meta\n",
"role": "user"
},
{
"content": "read_epw",
"role": "assistant"
}
] |
def x(cls, tmy_metadata, tmy_data=None, **kwargs):
tmy2 = tmy_metadata.get('City', False)
latitude = tmy_metadata['latitude']
longitude = tmy_metadata['longitude']
if tmy2:
name = tmy_metadata['City']
else:
name = tmy_metadata['Name']
tz = tmy_metadata['TZ']
altitude = tmy_metadata['altitude']
new_object = cls(latitude, longitude, tz=tz, altitude=altitude, name=
name, **kwargs)
if tmy_data is not None:
new_object.tmy_data = tmy_data
return new_object
| from_tmy | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, tmy_metadata, tmy_data=None, **kwargs):\n tmy2 = tmy_metadata.get('City', False)\n latitude = tmy_metadata['latitude']\n longitude = tmy_metadata['longitude']\n if tmy2:\n name = tmy_metadata['City']\n else:\n name = tmy_metadata['Name']\n tz = tmy_metadata['TZ']\n altitude = tmy_metadata['altitude']\n new_object = cls(latitude, longitude, tz=tz, altitude=altitude, name=\n name, **kwargs)\n if tmy_data is not None:\n new_object.tmy_data = tmy_data\n return new_object\n",
"role": "user"
},
{
"content": "from_tmy",
"role": "assistant"
}
] |
def x(self, times, pressure=None, temperature=12, **kwargs):
if pressure is None:
pressure = atmosphere.alt2pres(self.altitude)
return solarposition.get_solarposition(times, latitude=self.latitude,
longitude=self.longitude, altitude=self.altitude, pressure=pressure,
temperature=temperature, **kwargs)
| get_solarposition | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, times, pressure=None, temperature=12, **kwargs):\n if pressure is None:\n pressure = atmosphere.alt2pres(self.altitude)\n return solarposition.get_solarposition(times, latitude=self.latitude,\n longitude=self.longitude, altitude=self.altitude, pressure=pressure,\n temperature=temperature, **kwargs)\n",
"role": "user"
},
{
"content": "get_solarposition",
"role": "assistant"
}
] |
def x(self, times, model='ineichen', solar_position=None, dni_extra=None,
**kwargs):
if dni_extra is None:
dni_extra = irradiance.get_extra_radiation(times)
try:
pressure = kwargs.pop('pressure')
except KeyError:
pressure = atmosphere.alt2pres(self.altitude)
if solar_position is None:
solar_position = self.get_solarposition(times, pressure=pressure,
**kwargs)
apparent_zenith = solar_position['apparent_zenith']
apparent_elevation = solar_position['apparent_elevation']
if model == 'ineichen':
try:
linke_turbidity = kwargs.pop('linke_turbidity')
except KeyError:
interp_turbidity = kwargs.pop('interp_turbidity', True)
linke_turbidity = clearsky.lookup_linke_turbidity(times, self.
latitude, self.longitude, interp_turbidity=interp_turbidity)
try:
airmass_absolute = kwargs.pop('airmass_absolute')
except KeyError:
airmass_absolute = self.get_airmass(times, solar_position=
solar_position)['airmass_absolute']
cs = clearsky.ineichen(apparent_zenith, airmass_absolute,
linke_turbidity, altitude=self.altitude, dni_extra=dni_extra,
**kwargs)
elif model == 'haurwitz':
cs = clearsky.haurwitz(apparent_zenith)
elif model == 'simplified_solis':
cs = clearsky.simplified_solis(apparent_elevation, pressure=
pressure, dni_extra=dni_extra, **kwargs)
else:
raise ValueError(
'{} is not a valid clear sky model. Must be one of ineichen, simplified_solis, haurwitz'
.format(model))
return cs
| get_clearsky | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, times, model='ineichen', solar_position=None, dni_extra=None,\n **kwargs):\n if dni_extra is None:\n dni_extra = irradiance.get_extra_radiation(times)\n try:\n pressure = kwargs.pop('pressure')\n except KeyError:\n pressure = atmosphere.alt2pres(self.altitude)\n if solar_position is None:\n solar_position = self.get_solarposition(times, pressure=pressure,\n **kwargs)\n apparent_zenith = solar_position['apparent_zenith']\n apparent_elevation = solar_position['apparent_elevation']\n if model == 'ineichen':\n try:\n linke_turbidity = kwargs.pop('linke_turbidity')\n except KeyError:\n interp_turbidity = kwargs.pop('interp_turbidity', True)\n linke_turbidity = clearsky.lookup_linke_turbidity(times, self.\n latitude, self.longitude, interp_turbidity=interp_turbidity)\n try:\n airmass_absolute = kwargs.pop('airmass_absolute')\n except KeyError:\n airmass_absolute = self.get_airmass(times, solar_position=\n solar_position)['airmass_absolute']\n cs = clearsky.ineichen(apparent_zenith, airmass_absolute,\n linke_turbidity, altitude=self.altitude, dni_extra=dni_extra,\n **kwargs)\n elif model == 'haurwitz':\n cs = clearsky.haurwitz(apparent_zenith)\n elif model == 'simplified_solis':\n cs = clearsky.simplified_solis(apparent_elevation, pressure=\n pressure, dni_extra=dni_extra, **kwargs)\n else:\n raise ValueError(\n '{} is not a valid clear sky model. Must be one of ineichen, simplified_solis, haurwitz'\n .format(model))\n return cs\n",
"role": "user"
},
{
"content": "get_clearsky",
"role": "assistant"
}
] |
def x(self, times=None, solar_position=None, model='kastenyoung1989'):
if solar_position is None:
solar_position = self.get_solarposition(times)
if model in atmosphere.APPARENT_ZENITH_MODELS:
zenith = solar_position['apparent_zenith']
elif model in atmosphere.TRUE_ZENITH_MODELS:
zenith = solar_position['zenith']
else:
raise ValueError('{} is not a valid airmass model'.format(model))
airmass_relative = atmosphere.get_relative_airmass(zenith, model)
pressure = atmosphere.alt2pres(self.altitude)
airmass_absolute = atmosphere.get_absolute_airmass(airmass_relative,
pressure)
airmass = pd.DataFrame(index=solar_position.index)
airmass['airmass_relative'] = airmass_relative
airmass['airmass_absolute'] = airmass_absolute
return airmass
| get_airmass | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, times=None, solar_position=None, model='kastenyoung1989'):\n if solar_position is None:\n solar_position = self.get_solarposition(times)\n if model in atmosphere.APPARENT_ZENITH_MODELS:\n zenith = solar_position['apparent_zenith']\n elif model in atmosphere.TRUE_ZENITH_MODELS:\n zenith = solar_position['zenith']\n else:\n raise ValueError('{} is not a valid airmass model'.format(model))\n airmass_relative = atmosphere.get_relative_airmass(zenith, model)\n pressure = atmosphere.alt2pres(self.altitude)\n airmass_absolute = atmosphere.get_absolute_airmass(airmass_relative,\n pressure)\n airmass = pd.DataFrame(index=solar_position.index)\n airmass['airmass_relative'] = airmass_relative\n airmass['airmass_absolute'] = airmass_absolute\n return airmass\n",
"role": "user"
},
{
"content": "get_airmass",
"role": "assistant"
}
] |
def x(self, times, method='pyephem', **kwargs):
if method == 'pyephem':
result = solarposition.sun_rise_set_transit_ephem(times, self.
latitude, self.longitude, **kwargs)
elif method == 'spa':
result = solarposition.sun_rise_set_transit_spa(times, self.
latitude, self.longitude, **kwargs)
elif method == 'geometric':
sr, ss, tr = solarposition.sun_rise_set_transit_geometric(times,
self.latitude, self.longitude, **kwargs)
result = pd.DataFrame(index=times, data={'sunrise': sr, 'sunset':
ss, 'transit': tr})
else:
raise ValueError(
'{} is not a valid method. Must be one of pyephem, spa, geometric'
.format(method))
return result
| get_sun_rise_set_transit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, times, method='pyephem', **kwargs):\n if method == 'pyephem':\n result = solarposition.sun_rise_set_transit_ephem(times, self.\n latitude, self.longitude, **kwargs)\n elif method == 'spa':\n result = solarposition.sun_rise_set_transit_spa(times, self.\n latitude, self.longitude, **kwargs)\n elif method == 'geometric':\n sr, ss, tr = solarposition.sun_rise_set_transit_geometric(times,\n self.latitude, self.longitude, **kwargs)\n result = pd.DataFrame(index=times, data={'sunrise': sr, 'sunset':\n ss, 'transit': tr})\n else:\n raise ValueError(\n '{} is not a valid method. Must be one of pyephem, spa, geometric'\n .format(method))\n return result\n",
"role": "user"
},
{
"content": "get_sun_rise_set_transit",
"role": "assistant"
}
] |
def x(year, month, day, hour, minute, second, microsecond):
if month <= 2:
year = year - 1
month = month + 12
a = int(year / 100)
b = 2 - a + int(a * 0.25)
frac_of_day = (microsecond + (second + minute * 60 + hour * 3600)
) * 1.0 / (3600 * 24)
d = day + frac_of_day
jd = int(365.25 * (year + 4716)) + int(30.6001 * (month + 1)
) + d + b - 1524.5
return jd
| julian_day_dt | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(year, month, day, hour, minute, second, microsecond):\n if month <= 2:\n year = year - 1\n month = month + 12\n a = int(year / 100)\n b = 2 - a + int(a * 0.25)\n frac_of_day = (microsecond + (second + minute * 60 + hour * 3600)\n ) * 1.0 / (3600 * 24)\n d = day + frac_of_day\n jd = int(365.25 * (year + 4716)) + int(30.6001 * (month + 1)\n ) + d + b - 1524.5\n return jd\n",
"role": "user"
},
{
"content": "julian_day_dt",
"role": "assistant"
}
] |
def x(unixtime, loc_args, out):
lat = loc_args[0]
lon = loc_args[1]
elev = loc_args[2]
pressure = loc_args[3]
temp = loc_args[4]
delta_t = loc_args[5]
atmos_refract = loc_args[6]
sst = loc_args[7]
esd = loc_args[8]
for i in range(unixtime.shape[0]):
utime = unixtime[i]
jd = julian_day(utime)
jde = julian_ephemeris_day(jd, delta_t)
jc = julian_century(jd)
jce = julian_ephemeris_century(jde)
jme = julian_ephemeris_millennium(jce)
R = heliocentric_radius_vector(jme)
if esd:
out[0, i] = R
continue
L = heliocentric_longitude(jme)
B = heliocentric_latitude(jme)
Theta = geocentric_longitude(L)
beta = geocentric_latitude(B)
x0 = mean_elongation(jce)
x1 = mean_anomaly_sun(jce)
x2 = mean_anomaly_moon(jce)
x3 = moon_argument_latitude(jce)
x4 = moon_ascending_longitude(jce)
delta_psi = longitude_nutation(jce, x0, x1, x2, x3, x4)
delta_epsilon = obliquity_nutation(jce, x0, x1, x2, x3, x4)
epsilon0 = mean_ecliptic_obliquity(jme)
epsilon = true_ecliptic_obliquity(epsilon0, delta_epsilon)
delta_tau = aberration_correction(R)
lamd = apparent_sun_longitude(Theta, delta_psi, delta_tau)
v0 = mean_sidereal_time(jd, jc)
v = apparent_sidereal_time(v0, delta_psi, epsilon)
alpha = geocentric_sun_right_ascension(lamd, epsilon, beta)
delta = geocentric_sun_declination(lamd, epsilon, beta)
if sst:
out[0, i] = v
out[1, i] = alpha
out[2, i] = delta
continue
m = sun_mean_longitude(jme)
eot = equation_of_time(m, alpha, delta_psi, epsilon)
H = local_hour_angle(v, lon, alpha)
xi = equatorial_horizontal_parallax(R)
u = uterm(lat)
x = xterm(u, lat, elev)
y = yterm(u, lat, elev)
delta_alpha = parallax_sun_right_ascension(x, xi, H, delta)
delta_prime = topocentric_sun_declination(delta, x, y, xi,
delta_alpha, H)
H_prime = topocentric_local_hour_angle(H, delta_alpha)
e0 = topocentric_elevation_angle_without_atmosphere(lat,
delta_prime, H_prime)
delta_e = atmospheric_refraction_correction(pressure, temp, e0,
atmos_refract)
e = topocentric_elevation_angle(e0, delta_e)
theta = topocentric_zenith_angle(e)
theta0 = topocentric_zenith_angle(e0)
gamma = topocentric_astronomers_azimuth(H_prime, delta_prime, lat)
phi = topocentric_azimuth_angle(gamma)
out[0, i] = theta
out[1, i] = theta0
out[2, i] = e
out[3, i] = e0
out[4, i] = phi
out[5, i] = eot
| solar_position_loop | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(unixtime, loc_args, out):\n lat = loc_args[0]\n lon = loc_args[1]\n elev = loc_args[2]\n pressure = loc_args[3]\n temp = loc_args[4]\n delta_t = loc_args[5]\n atmos_refract = loc_args[6]\n sst = loc_args[7]\n esd = loc_args[8]\n for i in range(unixtime.shape[0]):\n utime = unixtime[i]\n jd = julian_day(utime)\n jde = julian_ephemeris_day(jd, delta_t)\n jc = julian_century(jd)\n jce = julian_ephemeris_century(jde)\n jme = julian_ephemeris_millennium(jce)\n R = heliocentric_radius_vector(jme)\n if esd:\n out[0, i] = R\n continue\n L = heliocentric_longitude(jme)\n B = heliocentric_latitude(jme)\n Theta = geocentric_longitude(L)\n beta = geocentric_latitude(B)\n x0 = mean_elongation(jce)\n x1 = mean_anomaly_sun(jce)\n x2 = mean_anomaly_moon(jce)\n x3 = moon_argument_latitude(jce)\n x4 = moon_ascending_longitude(jce)\n delta_psi = longitude_nutation(jce, x0, x1, x2, x3, x4)\n delta_epsilon = obliquity_nutation(jce, x0, x1, x2, x3, x4)\n epsilon0 = mean_ecliptic_obliquity(jme)\n epsilon = true_ecliptic_obliquity(epsilon0, delta_epsilon)\n delta_tau = aberration_correction(R)\n lamd = apparent_sun_longitude(Theta, delta_psi, delta_tau)\n v0 = mean_sidereal_time(jd, jc)\n v = apparent_sidereal_time(v0, delta_psi, epsilon)\n alpha = geocentric_sun_right_ascension(lamd, epsilon, beta)\n delta = geocentric_sun_declination(lamd, epsilon, beta)\n if sst:\n out[0, i] = v\n out[1, i] = alpha\n out[2, i] = delta\n continue\n m = sun_mean_longitude(jme)\n eot = equation_of_time(m, alpha, delta_psi, epsilon)\n H = local_hour_angle(v, lon, alpha)\n xi = equatorial_horizontal_parallax(R)\n u = uterm(lat)\n x = xterm(u, lat, elev)\n y = yterm(u, lat, elev)\n delta_alpha = parallax_sun_right_ascension(x, xi, H, delta)\n delta_prime = topocentric_sun_declination(delta, x, y, xi,\n delta_alpha, H)\n H_prime = topocentric_local_hour_angle(H, delta_alpha)\n e0 = topocentric_elevation_angle_without_atmosphere(lat,\n delta_prime, H_prime)\n delta_e = atmospheric_refraction_correction(pressure, temp, e0,\n atmos_refract)\n e = topocentric_elevation_angle(e0, delta_e)\n theta = topocentric_zenith_angle(e)\n theta0 = topocentric_zenith_angle(e0)\n gamma = topocentric_astronomers_azimuth(H_prime, delta_prime, lat)\n phi = topocentric_azimuth_angle(gamma)\n out[0, i] = theta\n out[1, i] = theta0\n out[2, i] = e\n out[3, i] = e0\n out[4, i] = phi\n out[5, i] = eot\n",
"role": "user"
},
{
"content": "solar_position_loop",
"role": "assistant"
}
] |
def x(unixtime, lat, lon, elev, pressure, temp, delta_t, atmos_refract,
numthreads, sst=False, esd=False):
loc_args = np.array([lat, lon, elev, pressure, temp, delta_t,
atmos_refract, sst, esd])
ulength = unixtime.shape[0]
if sst:
dims = 3
elif esd:
dims = 1
else:
dims = 6
result = np.empty((dims, ulength), dtype=np.float64)
if unixtime.dtype != np.float64:
unixtime = unixtime.astype(np.float64)
if ulength < numthreads:
warnings.warn(
'The number of threads is more than the length of the time array. Only using %s threads.'
.format(ulength))
numthreads = ulength
if numthreads <= 1:
solar_position_loop(unixtime, loc_args, result)
return result
split0 = np.array_split(unixtime, numthreads)
split2 = np.array_split(result, numthreads, axis=1)
chunks = [[a0, loc_args, split2[i]] for i, a0 in enumerate(split0)]
threads = [threading.Thread(target=solar_position_loop, args=chunk) for
chunk in chunks]
for thread in threads:
thread.start()
for thread in threads:
thread.join()
return result
| solar_position_numba | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(unixtime, lat, lon, elev, pressure, temp, delta_t, atmos_refract,\n numthreads, sst=False, esd=False):\n loc_args = np.array([lat, lon, elev, pressure, temp, delta_t,\n atmos_refract, sst, esd])\n ulength = unixtime.shape[0]\n if sst:\n dims = 3\n elif esd:\n dims = 1\n else:\n dims = 6\n result = np.empty((dims, ulength), dtype=np.float64)\n if unixtime.dtype != np.float64:\n unixtime = unixtime.astype(np.float64)\n if ulength < numthreads:\n warnings.warn(\n 'The number of threads is more than the length of the time array. Only using %s threads.'\n .format(ulength))\n numthreads = ulength\n if numthreads <= 1:\n solar_position_loop(unixtime, loc_args, result)\n return result\n split0 = np.array_split(unixtime, numthreads)\n split2 = np.array_split(result, numthreads, axis=1)\n chunks = [[a0, loc_args, split2[i]] for i, a0 in enumerate(split0)]\n threads = [threading.Thread(target=solar_position_loop, args=chunk) for\n chunk in chunks]\n for thread in threads:\n thread.start()\n for thread in threads:\n thread.join()\n return result\n",
"role": "user"
},
{
"content": "solar_position_numba",
"role": "assistant"
}
] |
def x(unixtime, delta_t, numthreads):
R = solar_position(unixtime, 0, 0, 0, 0, 0, delta_t, 0, numthreads, esd
=True)[0]
return R
| earthsun_distance | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(unixtime, delta_t, numthreads):\n R = solar_position(unixtime, 0, 0, 0, 0, 0, delta_t, 0, numthreads, esd\n =True)[0]\n return R\n",
"role": "user"
},
{
"content": "earthsun_distance",
"role": "assistant"
}
] |
def x(solar_azimuth, solar_zenith, surface_azimuth, surface_tilt,
timestamps, dni, dhi, gcr, pvrow_height, pvrow_width, albedo, n_pvrows=
3, index_observed_pvrow=1, rho_front_pvrow=0.03, rho_back_pvrow=0.05,
horizon_band_angle=15.0, run_parallel_calculations=True,
n_workers_for_parallel_calcs=None):
if isinstance(solar_azimuth, pd.Series):
solar_azimuth = solar_azimuth.values
if isinstance(solar_zenith, pd.Series):
solar_zenith = solar_zenith.values
if isinstance(surface_azimuth, pd.Series):
surface_azimuth = surface_azimuth.values
if isinstance(surface_tilt, pd.Series):
surface_tilt = surface_tilt.values
if isinstance(dni, pd.Series):
dni = dni.values
if isinstance(dhi, pd.Series):
dhi = dhi.values
from pvfactors.timeseries import calculate_radiosities_parallel_perez, calculate_radiosities_serially_perez, get_average_pvrow_outputs
idx_slice = pd.IndexSlice
pvarray_parameters = {'n_pvrows': n_pvrows, 'pvrow_height':
pvrow_height, 'pvrow_width': pvrow_width, 'gcr': gcr, 'rho_ground':
albedo, 'rho_front_pvrow': rho_front_pvrow, 'rho_back_pvrow':
rho_back_pvrow, 'horizon_band_angle': horizon_band_angle}
if run_parallel_calculations:
df_registries, df_custom_perez = calculate_radiosities_parallel_perez(
pvarray_parameters, timestamps, solar_zenith, solar_azimuth,
surface_tilt, surface_azimuth, dni, dhi, n_processes=
n_workers_for_parallel_calcs)
else:
inputs = (pvarray_parameters, timestamps, solar_zenith,
solar_azimuth, surface_tilt, surface_azimuth, dni, dhi)
df_registries, df_custom_perez = calculate_radiosities_serially_perez(
inputs)
df_outputs = get_average_pvrow_outputs(df_registries, values=['qinc'],
include_shading=True)
ipoa_front = df_outputs.loc[:, idx_slice[index_observed_pvrow, 'front',
'qinc']]
ipoa_back = df_outputs.loc[:, idx_slice[index_observed_pvrow, 'back',
'qinc']]
df_registries = df_registries.set_index('timestamps')
return ipoa_front, ipoa_back, df_registries
| pvfactors_timeseries | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(solar_azimuth, solar_zenith, surface_azimuth, surface_tilt,\n timestamps, dni, dhi, gcr, pvrow_height, pvrow_width, albedo, n_pvrows=\n 3, index_observed_pvrow=1, rho_front_pvrow=0.03, rho_back_pvrow=0.05,\n horizon_band_angle=15.0, run_parallel_calculations=True,\n n_workers_for_parallel_calcs=None):\n if isinstance(solar_azimuth, pd.Series):\n solar_azimuth = solar_azimuth.values\n if isinstance(solar_zenith, pd.Series):\n solar_zenith = solar_zenith.values\n if isinstance(surface_azimuth, pd.Series):\n surface_azimuth = surface_azimuth.values\n if isinstance(surface_tilt, pd.Series):\n surface_tilt = surface_tilt.values\n if isinstance(dni, pd.Series):\n dni = dni.values\n if isinstance(dhi, pd.Series):\n dhi = dhi.values\n from pvfactors.timeseries import calculate_radiosities_parallel_perez, calculate_radiosities_serially_perez, get_average_pvrow_outputs\n idx_slice = pd.IndexSlice\n pvarray_parameters = {'n_pvrows': n_pvrows, 'pvrow_height':\n pvrow_height, 'pvrow_width': pvrow_width, 'gcr': gcr, 'rho_ground':\n albedo, 'rho_front_pvrow': rho_front_pvrow, 'rho_back_pvrow':\n rho_back_pvrow, 'horizon_band_angle': horizon_band_angle}\n if run_parallel_calculations:\n df_registries, df_custom_perez = calculate_radiosities_parallel_perez(\n pvarray_parameters, timestamps, solar_zenith, solar_azimuth,\n surface_tilt, surface_azimuth, dni, dhi, n_processes=\n n_workers_for_parallel_calcs)\n else:\n inputs = (pvarray_parameters, timestamps, solar_zenith,\n solar_azimuth, surface_tilt, surface_azimuth, dni, dhi)\n df_registries, df_custom_perez = calculate_radiosities_serially_perez(\n inputs)\n df_outputs = get_average_pvrow_outputs(df_registries, values=['qinc'],\n include_shading=True)\n ipoa_front = df_outputs.loc[:, idx_slice[index_observed_pvrow, 'front',\n 'qinc']]\n ipoa_back = df_outputs.loc[:, idx_slice[index_observed_pvrow, 'back',\n 'qinc']]\n df_registries = df_registries.set_index('timestamps')\n return ipoa_front, ipoa_back, df_registries\n",
"role": "user"
},
{
"content": "pvfactors_timeseries",
"role": "assistant"
}
] |
def x(api_endpoint, credentials, device_model_id, device_id, lang, verbose,
input_audio_file, output_audio_file, block_size, grpc_deadline, *args,
**kwargs):
logging.basicConfig(level=logging.DEBUG if verbose else logging.INFO)
try:
with open(credentials, 'r') as f:
credentials = google.oauth2.credentials.Credentials(token=None,
**json.load(f))
http_request = google.auth.transport.requests.Request()
credentials.refresh(http_request)
except Exception as e:
logging.error('Error loading credentials: %s', e)
logging.error(
'Run google-oauthlib-tool to initialize new OAuth 2.0 credentials.'
)
sys.exit(-1)
grpc_channel = google.auth.transport.grpc.secure_authorized_channel(
credentials, http_request, api_endpoint)
logging.info('Connecting to %s', api_endpoint)
assistant = embedded_assistant_pb2_grpc.EmbeddedAssistantStub(grpc_channel)
def x(input_stream):
dialog_state_in = embedded_assistant_pb2.DialogStateIn(language_code
=lang, conversation_state=b'')
config = embedded_assistant_pb2.AssistConfig(audio_in_config=
embedded_assistant_pb2.AudioInConfig(encoding='LINEAR16',
sample_rate_hertz=16000), audio_out_config=
embedded_assistant_pb2.AudioOutConfig(encoding='LINEAR16',
sample_rate_hertz=16000, volume_percentage=100),
dialog_state_in=dialog_state_in, device_config=
embedded_assistant_pb2.DeviceConfig(device_id=device_id,
device_model_id=device_model_id))
yield embedded_assistant_pb2.AssistRequest(config=config)
while True:
data = input_stream.read(block_size)
if not data:
break
yield embedded_assistant_pb2.AssistRequest(audio_in=data)
for resp in assistant.Assist(gen_assist_requests(input_audio_file),
grpc_deadline):
if resp.event_type == END_OF_UTTERANCE:
logging.info('End of audio request detected')
if resp.speech_results:
logging.info('Transcript of user request: "%s".', ' '.join(r.
transcript for r in resp.speech_results))
if len(resp.audio_out.audio_data) > 0:
output_audio_file.write(resp.audio_out.audio_data)
if resp.dialog_state_out.supplemental_display_text:
logging.info('Assistant display text: "%s"', resp.
dialog_state_out.supplemental_display_text)
if resp.device_action.device_request_json:
device_request = json.loads(resp.device_action.device_request_json)
logging.info('Device request: %s', device_request)
| main | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(api_endpoint, credentials, device_model_id, device_id, lang, verbose,\n input_audio_file, output_audio_file, block_size, grpc_deadline, *args,\n **kwargs):\n logging.basicConfig(level=logging.DEBUG if verbose else logging.INFO)\n try:\n with open(credentials, 'r') as f:\n credentials = google.oauth2.credentials.Credentials(token=None,\n **json.load(f))\n http_request = google.auth.transport.requests.Request()\n credentials.refresh(http_request)\n except Exception as e:\n logging.error('Error loading credentials: %s', e)\n logging.error(\n 'Run google-oauthlib-tool to initialize new OAuth 2.0 credentials.'\n )\n sys.exit(-1)\n grpc_channel = google.auth.transport.grpc.secure_authorized_channel(\n credentials, http_request, api_endpoint)\n logging.info('Connecting to %s', api_endpoint)\n assistant = embedded_assistant_pb2_grpc.EmbeddedAssistantStub(grpc_channel)\n\n def x(input_stream):\n dialog_state_in = embedded_assistant_pb2.DialogStateIn(language_code\n =lang, conversation_state=b'')\n config = embedded_assistant_pb2.AssistConfig(audio_in_config=\n embedded_assistant_pb2.AudioInConfig(encoding='LINEAR16',\n sample_rate_hertz=16000), audio_out_config=\n embedded_assistant_pb2.AudioOutConfig(encoding='LINEAR16',\n sample_rate_hertz=16000, volume_percentage=100),\n dialog_state_in=dialog_state_in, device_config=\n embedded_assistant_pb2.DeviceConfig(device_id=device_id,\n device_model_id=device_model_id))\n yield embedded_assistant_pb2.AssistRequest(config=config)\n while True:\n data = input_stream.read(block_size)\n if not data:\n break\n yield embedded_assistant_pb2.AssistRequest(audio_in=data)\n for resp in assistant.Assist(gen_assist_requests(input_audio_file),\n grpc_deadline):\n if resp.event_type == END_OF_UTTERANCE:\n logging.info('End of audio request detected')\n if resp.speech_results:\n logging.info('Transcript of user request: \"%s\".', ' '.join(r.\n transcript for r in resp.speech_results))\n if len(resp.audio_out.audio_data) > 0:\n output_audio_file.write(resp.audio_out.audio_data)\n if resp.dialog_state_out.supplemental_display_text:\n logging.info('Assistant display text: \"%s\"', resp.\n dialog_state_out.supplemental_display_text)\n if resp.device_action.device_request_json:\n device_request = json.loads(resp.device_action.device_request_json)\n logging.info('Device request: %s', device_request)\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x(event):
if event.type == EventType.ON_CONVERSATION_TURN_STARTED:
print()
print(event)
if (event.type == EventType.ON_CONVERSATION_TURN_FINISHED and event.
args and not event.args['with_follow_on_turn']):
print()
if event.type == EventType.ON_DEVICE_ACTION:
for command, params in event.actions:
print('Do command', command, 'with params', str(params))
| process_event | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(event):\n if event.type == EventType.ON_CONVERSATION_TURN_STARTED:\n print()\n print(event)\n if (event.type == EventType.ON_CONVERSATION_TURN_FINISHED and event.\n args and not event.args['with_follow_on_turn']):\n print()\n if event.type == EventType.ON_DEVICE_ACTION:\n for command, params in event.actions:\n print('Do command', command, 'with params', str(params))\n",
"role": "user"
},
{
"content": "process_event",
"role": "assistant"
}
] |
def x(message, r):
try:
resp = json.loads(r.text)
message = '%s: %d\n%s' % (message, resp['error']['code'], resp[
'error']['message'])
return click.ClickException(message)
except ValueError:
return click.ClickException('%s: %d\n%s' % (message, r.status_code,
r.text))
| failed_request_exception | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(message, r):\n try:\n resp = json.loads(r.text)\n message = '%s: %d\\n%s' % (message, resp['error']['code'], resp[\n 'error']['message'])\n return click.ClickException(message)\n except ValueError:\n return click.ClickException('%s: %d\\n%s' % (message, r.status_code,\n r.text))\n",
"role": "user"
},
{
"content": "failed_request_exception",
"role": "assistant"
}
] |
def x(devicemodel):
PRETTY_PRINT_MODEL = """Device Model ID: %(deviceModelId)s
Project ID: %(projectId)s
Device Type: %(deviceType)s"""
logging.info(PRETTY_PRINT_MODEL % devicemodel)
if 'traits' in devicemodel:
for trait in devicemodel['traits']:
logging.info(' Trait %s' % trait)
else:
logging.info('No traits')
logging.info('')
| pretty_print_model | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(devicemodel):\n PRETTY_PRINT_MODEL = \"\"\"Device Model ID: %(deviceModelId)s\n Project ID: %(projectId)s\n Device Type: %(deviceType)s\"\"\"\n logging.info(PRETTY_PRINT_MODEL % devicemodel)\n if 'traits' in devicemodel:\n for trait in devicemodel['traits']:\n logging.info(' Trait %s' % trait)\n else:\n logging.info('No traits')\n logging.info('')\n",
"role": "user"
},
{
"content": "pretty_print_model",
"role": "assistant"
}
] |
def x(device):
logging.info('Device Instance ID: %s' % device['id'])
if 'nickname' in device:
logging.info(' Nickname: %s' % device['nickname'])
if 'modelId' in device:
logging.info(' Model: %s' % device['modelId'])
logging.info('')
| pretty_print_device | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(device):\n logging.info('Device Instance ID: %s' % device['id'])\n if 'nickname' in device:\n logging.info(' Nickname: %s' % device['nickname'])\n if 'modelId' in device:\n logging.info(' Model: %s' % device['modelId'])\n logging.info('')\n",
"role": "user"
},
{
"content": "pretty_print_device",
"role": "assistant"
}
] |
def x(ctx, model, type, trait, manufacturer, product_name, description,
device, nickname, client_type):
ctx.obj['SESSION'] = google.auth.transport.requests.AuthorizedSession(ctx
.obj['CREDENTIALS'])
ctx.invoke(register_model, model=model, type=type, trait=trait,
manufacturer=manufacturer, product_name=product_name, description=
description)
ctx.invoke(register_device, device=device, model=model, nickname=
nickname, client_type=client_type)
| register | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ctx, model, type, trait, manufacturer, product_name, description,\n device, nickname, client_type):\n ctx.obj['SESSION'] = google.auth.transport.requests.AuthorizedSession(ctx\n .obj['CREDENTIALS'])\n ctx.invoke(register_model, model=model, type=type, trait=trait,\n manufacturer=manufacturer, product_name=product_name, description=\n description)\n ctx.invoke(register_device, device=device, model=model, nickname=\n nickname, client_type=client_type)\n",
"role": "user"
},
{
"content": "register",
"role": "assistant"
}
] |
def x(ctx, model, type, trait, manufacturer, product_name, description):
session, api_url, project_id = build_client_from_context(ctx)
model_base_url = '/'.join([api_url, 'deviceModels'])
model_url = '/'.join([model_base_url, model])
payload = {'device_model_id': model, 'project_id': project_id,
'device_type': 'action.devices.types.' + type}
if trait:
payload['traits'] = trait
if manufacturer:
payload.setdefault('manifest', {})['manufacturer'] = manufacturer
if product_name:
payload.setdefault('manifest', {})['productName'] = product_name
if description:
payload.setdefault('manifest', {})['deviceDescription'] = description
logging.debug(json.dumps(payload))
r = session.get(model_url)
logging.debug(r.text)
if r.status_code == 200:
click.echo('Updating existing device model: %s' % model)
r = session.put(model_url, data=json.dumps(payload))
elif r.status_code in (400, 403, 404):
click.echo('Creating new device model')
r = session.post(model_base_url, data=json.dumps(payload))
else:
raise failed_request_exception('Failed to check existing device model',
r)
if r.status_code != 200:
raise failed_request_exception('Failed to register model', r)
click.echo('Model %s successfully registered' % model)
| register_model | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ctx, model, type, trait, manufacturer, product_name, description):\n session, api_url, project_id = build_client_from_context(ctx)\n model_base_url = '/'.join([api_url, 'deviceModels'])\n model_url = '/'.join([model_base_url, model])\n payload = {'device_model_id': model, 'project_id': project_id,\n 'device_type': 'action.devices.types.' + type}\n if trait:\n payload['traits'] = trait\n if manufacturer:\n payload.setdefault('manifest', {})['manufacturer'] = manufacturer\n if product_name:\n payload.setdefault('manifest', {})['productName'] = product_name\n if description:\n payload.setdefault('manifest', {})['deviceDescription'] = description\n logging.debug(json.dumps(payload))\n r = session.get(model_url)\n logging.debug(r.text)\n if r.status_code == 200:\n click.echo('Updating existing device model: %s' % model)\n r = session.put(model_url, data=json.dumps(payload))\n elif r.status_code in (400, 403, 404):\n click.echo('Creating new device model')\n r = session.post(model_base_url, data=json.dumps(payload))\n else:\n raise failed_request_exception('Failed to check existing device model',\n r)\n if r.status_code != 200:\n raise failed_request_exception('Failed to register model', r)\n click.echo('Model %s successfully registered' % model)\n",
"role": "user"
},
{
"content": "register_model",
"role": "assistant"
}
] |
def x(ctx, device, model, nickname, client_type):
session, api_url, project_id = build_client_from_context(ctx)
device_base_url = '/'.join([api_url, 'devices'])
device_url = '/'.join([device_base_url, device])
payload = {'id': device, 'model_id': model}
if client_type:
payload['client_type'] = 'SDK_' + client_type
if nickname:
payload['nickname'] = nickname
logging.debug(json.dumps(payload))
r = session.get(device_url)
if r.status_code == 200:
click.echo('Updating existing device: %s' % device)
session.delete(device_url)
r = session.post(device_base_url, data=json.dumps(payload))
elif r.status_code in (400, 403, 404):
click.echo('Creating new device')
r = session.post(device_base_url, data=json.dumps(payload))
else:
raise failed_request_exception('Failed to check existing device', r)
if r.status_code != 200:
raise failed_request_exception('Failed to register device', r)
click.echo('Device instance %s successfully registered' % device)
logging.debug(r.text)
| register_device | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ctx, device, model, nickname, client_type):\n session, api_url, project_id = build_client_from_context(ctx)\n device_base_url = '/'.join([api_url, 'devices'])\n device_url = '/'.join([device_base_url, device])\n payload = {'id': device, 'model_id': model}\n if client_type:\n payload['client_type'] = 'SDK_' + client_type\n if nickname:\n payload['nickname'] = nickname\n logging.debug(json.dumps(payload))\n r = session.get(device_url)\n if r.status_code == 200:\n click.echo('Updating existing device: %s' % device)\n session.delete(device_url)\n r = session.post(device_base_url, data=json.dumps(payload))\n elif r.status_code in (400, 403, 404):\n click.echo('Creating new device')\n r = session.post(device_base_url, data=json.dumps(payload))\n else:\n raise failed_request_exception('Failed to check existing device', r)\n if r.status_code != 200:\n raise failed_request_exception('Failed to register device', r)\n click.echo('Device instance %s successfully registered' % device)\n logging.debug(r.text)\n",
"role": "user"
},
{
"content": "register_device",
"role": "assistant"
}
] |
def x(ctx, resource, id):
session, api_url, project_id = build_client_from_context(ctx)
url = '/'.join([api_url, resource, id])
r = session.get(url)
if r.status_code != 200:
raise failed_request_exception('Failed to get resource', r)
response = json.loads(r.text)
if resource == 'deviceModels':
pretty_print_model(response)
elif resource == 'devices':
pretty_print_device(response)
logging.debug(r.text)
| get | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ctx, resource, id):\n session, api_url, project_id = build_client_from_context(ctx)\n url = '/'.join([api_url, resource, id])\n r = session.get(url)\n if r.status_code != 200:\n raise failed_request_exception('Failed to get resource', r)\n response = json.loads(r.text)\n if resource == 'deviceModels':\n pretty_print_model(response)\n elif resource == 'devices':\n pretty_print_device(response)\n logging.debug(r.text)\n",
"role": "user"
},
{
"content": "get",
"role": "assistant"
}
] |
def x(ctx, resource, id):
session, api_url, project_id = build_client_from_context(ctx)
url = '/'.join([api_url, resource, id])
r = session.delete(url)
if r.status_code != 200:
raise failed_request_exception('failed to delete resource', r)
click.echo(r.text)
| delete | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ctx, resource, id):\n session, api_url, project_id = build_client_from_context(ctx)\n url = '/'.join([api_url, resource, id])\n r = session.delete(url)\n if r.status_code != 200:\n raise failed_request_exception('failed to delete resource', r)\n click.echo(r.text)\n",
"role": "user"
},
{
"content": "delete",
"role": "assistant"
}
] |
def x(self, request_iterator, context):
context.set_code(grpc.StatusCode.UNIMPLEMENTED)
context.set_details('Method not implemented!')
raise NotImplementedError('Method not implemented!')
| Converse | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, request_iterator, context):\n context.set_code(grpc.StatusCode.UNIMPLEMENTED)\n context.set_details('Method not implemented!')\n raise NotImplementedError('Method not implemented!')\n",
"role": "user"
},
{
"content": "Converse",
"role": "assistant"
}
] |
def x(assist_request):
if logging.getLogger().isEnabledFor(logging.DEBUG):
resp_copy = embedded_assistant_pb2.AssistRequest()
resp_copy.CopyFrom(assist_request)
if len(resp_copy.audio_in) > 0:
size = len(resp_copy.audio_in)
resp_copy.ClearField('audio_in')
logging.debug('AssistRequest: audio_in (%d bytes)', size)
return
logging.debug('AssistRequest: %s', resp_copy)
| log_assist_request_without_audio | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(assist_request):\n if logging.getLogger().isEnabledFor(logging.DEBUG):\n resp_copy = embedded_assistant_pb2.AssistRequest()\n resp_copy.CopyFrom(assist_request)\n if len(resp_copy.audio_in) > 0:\n size = len(resp_copy.audio_in)\n resp_copy.ClearField('audio_in')\n logging.debug('AssistRequest: audio_in (%d bytes)', size)\n return\n logging.debug('AssistRequest: %s', resp_copy)\n",
"role": "user"
},
{
"content": "log_assist_request_without_audio",
"role": "assistant"
}
] |
def x(assist_response):
if logging.getLogger().isEnabledFor(logging.DEBUG):
resp_copy = embedded_assistant_pb2.AssistResponse()
resp_copy.CopyFrom(assist_response)
has_audio_data = resp_copy.HasField('audio_out') and len(resp_copy.
audio_out.audio_data) > 0
if has_audio_data:
size = len(resp_copy.audio_out.audio_data)
resp_copy.audio_out.ClearField('audio_data')
if resp_copy.audio_out.ListFields():
logging.debug('AssistResponse: %s audio_data (%d bytes)',
resp_copy, size)
else:
logging.debug('AssistResponse: audio_data (%d bytes)', size)
return
logging.debug('AssistResponse: %s', resp_copy)
| log_assist_response_without_audio | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(assist_response):\n if logging.getLogger().isEnabledFor(logging.DEBUG):\n resp_copy = embedded_assistant_pb2.AssistResponse()\n resp_copy.CopyFrom(assist_response)\n has_audio_data = resp_copy.HasField('audio_out') and len(resp_copy.\n audio_out.audio_data) > 0\n if has_audio_data:\n size = len(resp_copy.audio_out.audio_data)\n resp_copy.audio_out.ClearField('audio_data')\n if resp_copy.audio_out.ListFields():\n logging.debug('AssistResponse: %s audio_data (%d bytes)',\n resp_copy, size)\n else:\n logging.debug('AssistResponse: audio_data (%d bytes)', size)\n return\n logging.debug('AssistResponse: %s', resp_copy)\n",
"role": "user"
},
{
"content": "log_assist_response_without_audio",
"role": "assistant"
}
] |
def x(self, intent):
def x(fn):
self.handlers[intent] = fn
return decorator
| command | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, intent):\n\n def x(fn):\n self.handlers[intent] = fn\n return decorator\n",
"role": "user"
},
{
"content": "command",
"role": "assistant"
}
] |
def x(self, devices, execution):
fs = []
for device in devices:
if device[key_id_] != self.device_id:
logging.warning('Ignoring command for unknown device: %s' %
device[key_id_])
continue
if not execution:
logging.warning('Ignoring noop execution')
continue
for command in execution:
f = self.executor.submit(self.dispatch_command, **command)
fs.append(f)
return fs
| submit_commands | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, devices, execution):\n fs = []\n for device in devices:\n if device[key_id_] != self.device_id:\n logging.warning('Ignoring command for unknown device: %s' %\n device[key_id_])\n continue\n if not execution:\n logging.warning('Ignoring noop execution')\n continue\n for command in execution:\n f = self.executor.submit(self.dispatch_command, **command)\n fs.append(f)\n return fs\n",
"role": "user"
},
{
"content": "submit_commands",
"role": "assistant"
}
] |
def x(self, command, params=None):
try:
if command in self.handlers:
self.handlers[command](**params)
else:
logging.warning('Unsupported command: %s: %s', command, params)
except Exception as e:
logging.warning('Error during command execution', exc_info=sys.
exc_info())
raise e
| dispatch_command | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, command, params=None):\n try:\n if command in self.handlers:\n self.handlers[command](**params)\n else:\n logging.warning('Unsupported command: %s: %s', command, params)\n except Exception as e:\n logging.warning('Error during command execution', exc_info=sys.\n exc_info())\n raise e\n",
"role": "user"
},
{
"content": "dispatch_command",
"role": "assistant"
}
] |
def x(buf, volume_percentage, sample_width=2):
if sample_width != 2:
raise Exception('unsupported sample width:', sample_width)
scale = math.pow(2, 1.0 * volume_percentage / 100) - 1
arr = array.array('h', buf)
for idx in range(0, len(arr)):
arr[idx] = int(arr[idx] * scale)
buf = arr.tostring()
return buf
| normalize_audio_buffer | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(buf, volume_percentage, sample_width=2):\n if sample_width != 2:\n raise Exception('unsupported sample width:', sample_width)\n scale = math.pow(2, 1.0 * volume_percentage / 100) - 1\n arr = array.array('h', buf)\n for idx in range(0, len(arr)):\n arr[idx] = int(arr[idx] * scale)\n buf = arr.tostring()\n return buf\n",
"role": "user"
},
{
"content": "normalize_audio_buffer",
"role": "assistant"
}
] |
def x(buf, sample_width):
remainder = len(buf) % sample_width
if remainder != 0:
buf += b'\x00' * (sample_width - remainder)
return buf
| align_buf | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(buf, sample_width):\n remainder = len(buf) % sample_width\n if remainder != 0:\n buf += b'\\x00' * (sample_width - remainder)\n return buf\n",
"role": "user"
},
{
"content": "align_buf",
"role": "assistant"
}
] |
def x(record_time, audio_sample_rate, audio_sample_width, audio_iter_size,
audio_block_size, audio_flush_size):
end_time = time.time() + record_time
audio_device = SoundDeviceStream(sample_rate=audio_sample_rate,
sample_width=audio_sample_width, block_size=audio_block_size,
flush_size=audio_flush_size)
stream = ConversationStream(source=audio_device, sink=audio_device,
iter_size=audio_iter_size, sample_width=audio_sample_width)
samples = []
logging.basicConfig(level=logging.INFO)
logging.info('Starting audio test.')
stream.start_recording()
logging.info('Recording samples.')
while time.time() < end_time:
samples.append(stream.read(audio_block_size))
logging.info('Finished recording.')
stream.stop_recording()
stream.start_playback()
logging.info('Playing back samples.')
while len(samples):
stream.write(samples.pop(0))
logging.info('Finished playback.')
stream.stop_playback()
logging.info('audio test completed.')
stream.close()
| main | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(record_time, audio_sample_rate, audio_sample_width, audio_iter_size,\n audio_block_size, audio_flush_size):\n end_time = time.time() + record_time\n audio_device = SoundDeviceStream(sample_rate=audio_sample_rate,\n sample_width=audio_sample_width, block_size=audio_block_size,\n flush_size=audio_flush_size)\n stream = ConversationStream(source=audio_device, sink=audio_device,\n iter_size=audio_iter_size, sample_width=audio_sample_width)\n samples = []\n logging.basicConfig(level=logging.INFO)\n logging.info('Starting audio test.')\n stream.start_recording()\n logging.info('Recording samples.')\n while time.time() < end_time:\n samples.append(stream.read(audio_block_size))\n logging.info('Finished recording.')\n stream.stop_recording()\n stream.start_playback()\n logging.info('Playing back samples.')\n while len(samples):\n stream.write(samples.pop(0))\n logging.info('Finished playback.')\n stream.stop_playback()\n logging.info('audio test completed.')\n stream.close()\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x(self, size):
now = time.time()
missing_dt = self._sleep_until - now
if missing_dt > 0:
time.sleep(missing_dt)
self._sleep_until = time.time() + self._sleep_time(size)
data = self._wavep.readframes(size) if self._wavep else self._fp.read(size)
if not data:
return b'\x00' * size
return data
| read | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, size):\n now = time.time()\n missing_dt = self._sleep_until - now\n if missing_dt > 0:\n time.sleep(missing_dt)\n self._sleep_until = time.time() + self._sleep_time(size)\n data = self._wavep.readframes(size) if self._wavep else self._fp.read(size)\n if not data:\n return b'\\x00' * size\n return data\n",
"role": "user"
},
{
"content": "read",
"role": "assistant"
}
] |
def x(self):
if self._wavep:
self._wavep.close()
self._fp.close()
| close | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._wavep:\n self._wavep.close()\n self._fp.close()\n",
"role": "user"
},
{
"content": "close",
"role": "assistant"
}
] |
def x(self, size):
buf, overflow = self._audio_stream.read(size)
if overflow:
logging.warning('SoundDeviceStream read overflow (%d, %d)', size,
len(buf))
return bytes(buf)
| read | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, size):\n buf, overflow = self._audio_stream.read(size)\n if overflow:\n logging.warning('SoundDeviceStream read overflow (%d, %d)', size,\n len(buf))\n return bytes(buf)\n",
"role": "user"
},
{
"content": "read",
"role": "assistant"
}
] |
def x(self, buf):
underflow = self._audio_stream.write(buf)
if underflow:
logging.warning('SoundDeviceStream write underflow (size: %d)', len
(buf))
return len(buf)
| write | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, buf):\n underflow = self._audio_stream.write(buf)\n if underflow:\n logging.warning('SoundDeviceStream write underflow (size: %d)', len\n (buf))\n return len(buf)\n",
"role": "user"
},
{
"content": "write",
"role": "assistant"
}
] |
def x(self):
if self._audio_stream:
self.stop()
self._audio_stream.close()
self._audio_stream = None
| close | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._audio_stream:\n self.stop()\n self._audio_stream.close()\n self._audio_stream = None\n",
"role": "user"
},
{
"content": "close",
"role": "assistant"
}
] |
def x(self):
self._recording = True
self._stop_recording.clear()
self._source.start()
| start_recording | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self._recording = True\n self._stop_recording.clear()\n self._source.start()\n",
"role": "user"
},
{
"content": "start_recording",
"role": "assistant"
}
] |
def x(self):
self._stop_recording.set()
with self._source_lock:
self._source.stop()
self._recording = False
| stop_recording | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self._stop_recording.set()\n with self._source_lock:\n self._source.stop()\n self._recording = False\n",
"role": "user"
},
{
"content": "stop_recording",
"role": "assistant"
}
] |
def x(self):
self._sink.flush()
self._sink.stop()
self._playing = False
| stop_playback | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self._sink.flush()\n self._sink.stop()\n self._playing = False\n",
"role": "user"
},
{
"content": "stop_playback",
"role": "assistant"
}
] |
def x(self, buf):
buf = align_buf(buf, self._sample_width)
buf = normalize_audio_buffer(buf, self.volume_percentage)
return self._sink.write(buf)
| write | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, buf):\n buf = align_buf(buf, self._sample_width)\n buf = normalize_audio_buffer(buf, self.volume_percentage)\n return self._sink.write(buf)\n",
"role": "user"
},
{
"content": "write",
"role": "assistant"
}
] |
def x(self, text_query):
def x():
config = embedded_assistant_pb2.AssistConfig(audio_out_config=
embedded_assistant_pb2.AudioOutConfig(encoding='LINEAR16',
sample_rate_hertz=16000, volume_percentage=0), dialog_state_in=
embedded_assistant_pb2.DialogStateIn(language_code=self.
language_code, conversation_state=self.conversation_state,
is_new_conversation=self.is_new_conversation), device_config=
embedded_assistant_pb2.DeviceConfig(device_id=self.device_id,
device_model_id=self.device_model_id), text_query=text_query)
self.is_new_conversation = False
if self.display:
config.screen_out_config.screen_mode = PLAYING
req = embedded_assistant_pb2.AssistRequest(config=config)
assistant_helpers.log_assist_request_without_audio(req)
yield req
text_response = None
html_response = None
for resp in self.assistant.Assist(iter_assist_requests(), self.deadline):
assistant_helpers.log_assist_response_without_audio(resp)
if resp.screen_out.data:
html_response = resp.screen_out.data
if resp.dialog_state_out.conversation_state:
conversation_state = resp.dialog_state_out.conversation_state
self.conversation_state = conversation_state
if resp.dialog_state_out.supplemental_display_text:
text_response = resp.dialog_state_out.supplemental_display_text
return text_response, html_response
| assist | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, text_query):\n\n def x():\n config = embedded_assistant_pb2.AssistConfig(audio_out_config=\n embedded_assistant_pb2.AudioOutConfig(encoding='LINEAR16',\n sample_rate_hertz=16000, volume_percentage=0), dialog_state_in=\n embedded_assistant_pb2.DialogStateIn(language_code=self.\n language_code, conversation_state=self.conversation_state,\n is_new_conversation=self.is_new_conversation), device_config=\n embedded_assistant_pb2.DeviceConfig(device_id=self.device_id,\n device_model_id=self.device_model_id), text_query=text_query)\n self.is_new_conversation = False\n if self.display:\n config.screen_out_config.screen_mode = PLAYING\n req = embedded_assistant_pb2.AssistRequest(config=config)\n assistant_helpers.log_assist_request_without_audio(req)\n yield req\n text_response = None\n html_response = None\n for resp in self.assistant.Assist(iter_assist_requests(), self.deadline):\n assistant_helpers.log_assist_response_without_audio(resp)\n if resp.screen_out.data:\n html_response = resp.screen_out.data\n if resp.dialog_state_out.conversation_state:\n conversation_state = resp.dialog_state_out.conversation_state\n self.conversation_state = conversation_state\n if resp.dialog_state_out.supplemental_display_text:\n text_response = resp.dialog_state_out.supplemental_display_text\n return text_response, html_response\n",
"role": "user"
},
{
"content": "assist",
"role": "assistant"
}
] |
def x(api_endpoint, credentials, project_id, device_model_id, device_id,
device_config, lang, display, verbose, input_audio_file,
output_audio_file, audio_sample_rate, audio_sample_width,
audio_iter_size, audio_block_size, audio_flush_size, grpc_deadline,
once, *args, **kwargs):
logging.basicConfig(level=logging.DEBUG if verbose else logging.INFO)
try:
with open(credentials, 'r') as f:
credentials = google.oauth2.credentials.Credentials(token=None,
**json.load(f))
http_request = google.auth.transport.requests.Request()
credentials.refresh(http_request)
except Exception as e:
logging.error('Error loading credentials: %s', e)
logging.error(
'Run google-oauthlib-tool to initialize new OAuth 2.0 credentials.'
)
sys.exit(-1)
grpc_channel = google.auth.transport.grpc.secure_authorized_channel(
credentials, http_request, api_endpoint)
logging.info('Connecting to %s', api_endpoint)
audio_device = None
if input_audio_file:
audio_source = audio_helpers.WaveSource(open(input_audio_file, 'rb'
), sample_rate=audio_sample_rate, sample_width=audio_sample_width)
else:
audio_source = audio_device = (audio_device or audio_helpers.
SoundDeviceStream(sample_rate=audio_sample_rate, sample_width=
audio_sample_width, block_size=audio_block_size, flush_size=
audio_flush_size))
if output_audio_file:
audio_sink = audio_helpers.WaveSink(open(output_audio_file, 'wb'),
sample_rate=audio_sample_rate, sample_width=audio_sample_width)
else:
audio_sink = audio_device = (audio_device or audio_helpers.
SoundDeviceStream(sample_rate=audio_sample_rate, sample_width=
audio_sample_width, block_size=audio_block_size, flush_size=
audio_flush_size))
conversation_stream = audio_helpers.ConversationStream(source=
audio_source, sink=audio_sink, iter_size=audio_iter_size,
sample_width=audio_sample_width)
if not device_id or not device_model_id:
try:
with open(device_config) as f:
device = json.load(f)
device_id = device['id']
device_model_id = device['model_id']
logging.info('Using device model %s and device id %s',
device_model_id, device_id)
except Exception as e:
logging.warning('Device config not found: %s' % e)
logging.info('Registering device')
if not device_model_id:
logging.error(
'Option --device-model-id required when registering a device instance.'
)
sys.exit(-1)
if not project_id:
logging.error(
'Option --project-id required when registering a device instance.'
)
sys.exit(-1)
device_base_url = 'https://%s/v1alpha2/projects/%s/devices' % (
api_endpoint, project_id)
device_id = str(uuid.uuid1())
payload = {'id': device_id, 'model_id': device_model_id,
'client_type': 'SDK_SERVICE'}
session = google.auth.transport.requests.AuthorizedSession(
credentials)
r = session.post(device_base_url, data=json.dumps(payload))
if r.status_code != 200:
logging.error('Failed to register device: %s', r.text)
sys.exit(-1)
logging.info('Device registered: %s', device_id)
pathlib.Path(os.path.dirname(device_config)).mkdir(exist_ok=True)
with open(device_config, 'w') as f:
json.dump(payload, f)
device_handler = device_helpers.DeviceRequestHandler(device_id)
@device_handler.command('action.devices.commands.OnOff')
def x(on):
if on:
logging.info('Turning device on')
else:
logging.info('Turning device off')
@device_handler.command('com.example.commands.BlinkLight')
def x(speed, number):
logging.info('Blinking device %s times.' % number)
delay = 1
if speed == 'SLOWLY':
delay = 2
elif speed == 'QUICKLY':
delay = 0.5
for i in range(int(number)):
logging.info('Device is blinking.')
time.sleep(delay)
with SampleAssistant(lang, device_model_id, device_id,
conversation_stream, display, grpc_channel, grpc_deadline,
device_handler) as assistant:
if input_audio_file or output_audio_file:
assistant.assist()
return
wait_for_user_trigger = not once
while True:
if wait_for_user_trigger:
click.pause(info='Press Enter to send a new request...')
continue_conversation = assistant.assist()
wait_for_user_trigger = not continue_conversation
if once and not continue_conversation:
break
| main | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(api_endpoint, credentials, project_id, device_model_id, device_id,\n device_config, lang, display, verbose, input_audio_file,\n output_audio_file, audio_sample_rate, audio_sample_width,\n audio_iter_size, audio_block_size, audio_flush_size, grpc_deadline,\n once, *args, **kwargs):\n logging.basicConfig(level=logging.DEBUG if verbose else logging.INFO)\n try:\n with open(credentials, 'r') as f:\n credentials = google.oauth2.credentials.Credentials(token=None,\n **json.load(f))\n http_request = google.auth.transport.requests.Request()\n credentials.refresh(http_request)\n except Exception as e:\n logging.error('Error loading credentials: %s', e)\n logging.error(\n 'Run google-oauthlib-tool to initialize new OAuth 2.0 credentials.'\n )\n sys.exit(-1)\n grpc_channel = google.auth.transport.grpc.secure_authorized_channel(\n credentials, http_request, api_endpoint)\n logging.info('Connecting to %s', api_endpoint)\n audio_device = None\n if input_audio_file:\n audio_source = audio_helpers.WaveSource(open(input_audio_file, 'rb'\n ), sample_rate=audio_sample_rate, sample_width=audio_sample_width)\n else:\n audio_source = audio_device = (audio_device or audio_helpers.\n SoundDeviceStream(sample_rate=audio_sample_rate, sample_width=\n audio_sample_width, block_size=audio_block_size, flush_size=\n audio_flush_size))\n if output_audio_file:\n audio_sink = audio_helpers.WaveSink(open(output_audio_file, 'wb'),\n sample_rate=audio_sample_rate, sample_width=audio_sample_width)\n else:\n audio_sink = audio_device = (audio_device or audio_helpers.\n SoundDeviceStream(sample_rate=audio_sample_rate, sample_width=\n audio_sample_width, block_size=audio_block_size, flush_size=\n audio_flush_size))\n conversation_stream = audio_helpers.ConversationStream(source=\n audio_source, sink=audio_sink, iter_size=audio_iter_size,\n sample_width=audio_sample_width)\n if not device_id or not device_model_id:\n try:\n with open(device_config) as f:\n device = json.load(f)\n device_id = device['id']\n device_model_id = device['model_id']\n logging.info('Using device model %s and device id %s',\n device_model_id, device_id)\n except Exception as e:\n logging.warning('Device config not found: %s' % e)\n logging.info('Registering device')\n if not device_model_id:\n logging.error(\n 'Option --device-model-id required when registering a device instance.'\n )\n sys.exit(-1)\n if not project_id:\n logging.error(\n 'Option --project-id required when registering a device instance.'\n )\n sys.exit(-1)\n device_base_url = 'https://%s/v1alpha2/projects/%s/devices' % (\n api_endpoint, project_id)\n device_id = str(uuid.uuid1())\n payload = {'id': device_id, 'model_id': device_model_id,\n 'client_type': 'SDK_SERVICE'}\n session = google.auth.transport.requests.AuthorizedSession(\n credentials)\n r = session.post(device_base_url, data=json.dumps(payload))\n if r.status_code != 200:\n logging.error('Failed to register device: %s', r.text)\n sys.exit(-1)\n logging.info('Device registered: %s', device_id)\n pathlib.Path(os.path.dirname(device_config)).mkdir(exist_ok=True)\n with open(device_config, 'w') as f:\n json.dump(payload, f)\n device_handler = device_helpers.DeviceRequestHandler(device_id)\n\n @device_handler.command('action.devices.commands.OnOff')\n def x(on):\n if on:\n logging.info('Turning device on')\n else:\n logging.info('Turning device off')\n\n @device_handler.command('com.example.commands.BlinkLight')\n def x(speed, number):\n logging.info('Blinking device %s times.' % number)\n delay = 1\n if speed == 'SLOWLY':\n delay = 2\n elif speed == 'QUICKLY':\n delay = 0.5\n for i in range(int(number)):\n logging.info('Device is blinking.')\n time.sleep(delay)\n with SampleAssistant(lang, device_model_id, device_id,\n conversation_stream, display, grpc_channel, grpc_deadline,\n device_handler) as assistant:\n if input_audio_file or output_audio_file:\n assistant.assist()\n return\n wait_for_user_trigger = not once\n while True:\n if wait_for_user_trigger:\n click.pause(info='Press Enter to send a new request...')\n continue_conversation = assistant.assist()\n wait_for_user_trigger = not continue_conversation\n if once and not continue_conversation:\n break\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x(self):
continue_conversation = False
device_actions_futures = []
self.conversation_stream.start_recording()
logging.info('Recording audio request.')
def x():
for c in self.gen_assist_requests():
assistant_helpers.log_assist_request_without_audio(c)
yield c
logging.debug('Reached end of AssistRequest iteration.')
for resp in self.assistant.Assist(iter_log_assist_requests(), self.deadline
):
assistant_helpers.log_assist_response_without_audio(resp)
if resp.event_type == END_OF_UTTERANCE:
logging.info('End of audio request detected.')
logging.info('Stopping recording.')
self.conversation_stream.stop_recording()
if resp.speech_results:
logging.info('Transcript of user request: "%s".', ' '.join(r.
transcript for r in resp.speech_results))
if len(resp.audio_out.audio_data) > 0:
if not self.conversation_stream.playing:
self.conversation_stream.stop_recording()
self.conversation_stream.start_playback()
logging.info('Playing assistant response.')
self.conversation_stream.write(resp.audio_out.audio_data)
if resp.dialog_state_out.conversation_state:
conversation_state = resp.dialog_state_out.conversation_state
logging.debug('Updating conversation state.')
self.conversation_state = conversation_state
if resp.dialog_state_out.volume_percentage != 0:
volume_percentage = resp.dialog_state_out.volume_percentage
logging.info('Setting volume to %s%%', volume_percentage)
self.conversation_stream.volume_percentage = volume_percentage
if resp.dialog_state_out.microphone_mode == DIALOG_FOLLOW_ON:
continue_conversation = True
logging.info('Expecting follow-on query from user.')
elif resp.dialog_state_out.microphone_mode == CLOSE_MICROPHONE:
continue_conversation = False
if resp.device_action.device_request_json:
device_request = json.loads(resp.device_action.device_request_json)
fs = self.device_handler(device_request)
if fs:
device_actions_futures.extend(fs)
if self.display and resp.screen_out.data:
system_browser = browser_helpers.system_browser
system_browser.display(resp.screen_out.data)
if len(device_actions_futures):
logging.info('Waiting for device executions to complete.')
concurrent.futures.wait(device_actions_futures)
logging.info('Finished playing assistant response.')
self.conversation_stream.stop_playback()
return continue_conversation
| assist | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n continue_conversation = False\n device_actions_futures = []\n self.conversation_stream.start_recording()\n logging.info('Recording audio request.')\n\n def x():\n for c in self.gen_assist_requests():\n assistant_helpers.log_assist_request_without_audio(c)\n yield c\n logging.debug('Reached end of AssistRequest iteration.')\n for resp in self.assistant.Assist(iter_log_assist_requests(), self.deadline\n ):\n assistant_helpers.log_assist_response_without_audio(resp)\n if resp.event_type == END_OF_UTTERANCE:\n logging.info('End of audio request detected.')\n logging.info('Stopping recording.')\n self.conversation_stream.stop_recording()\n if resp.speech_results:\n logging.info('Transcript of user request: \"%s\".', ' '.join(r.\n transcript for r in resp.speech_results))\n if len(resp.audio_out.audio_data) > 0:\n if not self.conversation_stream.playing:\n self.conversation_stream.stop_recording()\n self.conversation_stream.start_playback()\n logging.info('Playing assistant response.')\n self.conversation_stream.write(resp.audio_out.audio_data)\n if resp.dialog_state_out.conversation_state:\n conversation_state = resp.dialog_state_out.conversation_state\n logging.debug('Updating conversation state.')\n self.conversation_state = conversation_state\n if resp.dialog_state_out.volume_percentage != 0:\n volume_percentage = resp.dialog_state_out.volume_percentage\n logging.info('Setting volume to %s%%', volume_percentage)\n self.conversation_stream.volume_percentage = volume_percentage\n if resp.dialog_state_out.microphone_mode == DIALOG_FOLLOW_ON:\n continue_conversation = True\n logging.info('Expecting follow-on query from user.')\n elif resp.dialog_state_out.microphone_mode == CLOSE_MICROPHONE:\n continue_conversation = False\n if resp.device_action.device_request_json:\n device_request = json.loads(resp.device_action.device_request_json)\n fs = self.device_handler(device_request)\n if fs:\n device_actions_futures.extend(fs)\n if self.display and resp.screen_out.data:\n system_browser = browser_helpers.system_browser\n system_browser.display(resp.screen_out.data)\n if len(device_actions_futures):\n logging.info('Waiting for device executions to complete.')\n concurrent.futures.wait(device_actions_futures)\n logging.info('Finished playing assistant response.')\n self.conversation_stream.stop_playback()\n return continue_conversation\n",
"role": "user"
},
{
"content": "assist",
"role": "assistant"
}
] |
def x(self):
config = embedded_assistant_pb2.AssistConfig(audio_in_config=
embedded_assistant_pb2.AudioInConfig(encoding='LINEAR16',
sample_rate_hertz=self.conversation_stream.sample_rate),
audio_out_config=embedded_assistant_pb2.AudioOutConfig(encoding=
'LINEAR16', sample_rate_hertz=self.conversation_stream.sample_rate,
volume_percentage=self.conversation_stream.volume_percentage),
dialog_state_in=embedded_assistant_pb2.DialogStateIn(language_code=
self.language_code, conversation_state=self.conversation_state,
is_new_conversation=self.is_new_conversation), device_config=
embedded_assistant_pb2.DeviceConfig(device_id=self.device_id,
device_model_id=self.device_model_id))
if self.display:
config.screen_out_config.screen_mode = PLAYING
self.is_new_conversation = False
yield embedded_assistant_pb2.AssistRequest(config=config)
for data in self.conversation_stream:
yield embedded_assistant_pb2.AssistRequest(audio_in=data)
| gen_assist_requests | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n config = embedded_assistant_pb2.AssistConfig(audio_in_config=\n embedded_assistant_pb2.AudioInConfig(encoding='LINEAR16',\n sample_rate_hertz=self.conversation_stream.sample_rate),\n audio_out_config=embedded_assistant_pb2.AudioOutConfig(encoding=\n 'LINEAR16', sample_rate_hertz=self.conversation_stream.sample_rate,\n volume_percentage=self.conversation_stream.volume_percentage),\n dialog_state_in=embedded_assistant_pb2.DialogStateIn(language_code=\n self.language_code, conversation_state=self.conversation_state,\n is_new_conversation=self.is_new_conversation), device_config=\n embedded_assistant_pb2.DeviceConfig(device_id=self.device_id,\n device_model_id=self.device_model_id))\n if self.display:\n config.screen_out_config.screen_mode = PLAYING\n self.is_new_conversation = False\n yield embedded_assistant_pb2.AssistRequest(config=config)\n for data in self.conversation_stream:\n yield embedded_assistant_pb2.AssistRequest(audio_in=data)\n",
"role": "user"
},
{
"content": "gen_assist_requests",
"role": "assistant"
}
] |
def x(self, X, shrinkage_value, max_rank=None):
if max_rank:
U, s, V = randomized_svd(X, max_rank, n_iter=self.n_power_iterations)
else:
U, s, V = np.linalg.svd(X, full_matrices=False, compute_uv=True)
s_thresh = np.maximum(s - shrinkage_value, 0)
rank = (s_thresh > 0).sum()
s_thresh = s_thresh[:rank]
U_thresh = U[:, :rank]
V_thresh = V[:rank, :]
S_thresh = np.diag(s_thresh)
X_reconstruction = np.dot(U_thresh, np.dot(S_thresh, V_thresh))
return X_reconstruction, rank
| svd_step | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, shrinkage_value, max_rank=None):\n if max_rank:\n U, s, V = randomized_svd(X, max_rank, n_iter=self.n_power_iterations)\n else:\n U, s, V = np.linalg.svd(X, full_matrices=False, compute_uv=True)\n s_thresh = np.maximum(s - shrinkage_value, 0)\n rank = (s_thresh > 0).sum()\n s_thresh = s_thresh[:rank]\n U_thresh = U[:, :rank]\n V_thresh = V[:rank, :]\n S_thresh = np.diag(s_thresh)\n X_reconstruction = np.dot(U_thresh, np.dot(S_thresh, V_thresh))\n return X_reconstruction, rank\n",
"role": "user"
},
{
"content": "svd_step",
"role": "assistant"
}
] |
def x(self, X, missing_mask, fill_method=None, inplace=False):
X = check_array(X, force_all_finite=False)
if not inplace:
X = X.copy()
if not fill_method:
fill_method = self.fill_method
if fill_method not in ('zero', 'mean', 'median', 'min', 'random'):
raise ValueError("Invalid fill method: '%s'" % fill_method)
elif fill_method == 'zero':
X[missing_mask] = 0
elif fill_method == 'mean':
self._fill_columns_with_fn(X, missing_mask, np.nanmean)
elif fill_method == 'median':
self._fill_columns_with_fn(X, missing_mask, np.nanmedian)
elif fill_method == 'min':
self._fill_columns_with_fn(X, missing_mask, np.nanmin)
elif fill_method == 'random':
self._fill_columns_with_fn(X, missing_mask, col_fn=
generate_random_column_samples)
return X
| fill | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, missing_mask, fill_method=None, inplace=False):\n X = check_array(X, force_all_finite=False)\n if not inplace:\n X = X.copy()\n if not fill_method:\n fill_method = self.fill_method\n if fill_method not in ('zero', 'mean', 'median', 'min', 'random'):\n raise ValueError(\"Invalid fill method: '%s'\" % fill_method)\n elif fill_method == 'zero':\n X[missing_mask] = 0\n elif fill_method == 'mean':\n self._fill_columns_with_fn(X, missing_mask, np.nanmean)\n elif fill_method == 'median':\n self._fill_columns_with_fn(X, missing_mask, np.nanmedian)\n elif fill_method == 'min':\n self._fill_columns_with_fn(X, missing_mask, np.nanmin)\n elif fill_method == 'random':\n self._fill_columns_with_fn(X, missing_mask, col_fn=\n generate_random_column_samples)\n return X\n",
"role": "user"
},
{
"content": "fill",
"role": "assistant"
}
] |
def x(self, X):
X = check_array(X, force_all_finite=False)
if X.dtype != 'f' and X.dtype != 'd':
X = X.astype(float)
self._check_input(X)
missing_mask = np.isnan(X)
self._check_missing_value_mask(missing_mask)
return X, missing_mask
| prepare_input_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n X = check_array(X, force_all_finite=False)\n if X.dtype != 'f' and X.dtype != 'd':\n X = X.astype(float)\n self._check_input(X)\n missing_mask = np.isnan(X)\n self._check_missing_value_mask(missing_mask)\n return X, missing_mask\n",
"role": "user"
},
{
"content": "prepare_input_data",
"role": "assistant"
}
] |
def x(self, X):
X = np.asarray(X)
if self.min_value is not None:
X[X < self.min_value] = self.min_value
if self.max_value is not None:
X[X > self.max_value] = self.max_value
return X
| clip | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n X = np.asarray(X)\n if self.min_value is not None:\n X[X < self.min_value] = self.min_value\n if self.max_value is not None:\n X[X > self.max_value] = self.max_value\n return X\n",
"role": "user"
},
{
"content": "clip",
"role": "assistant"
}
] |
def x(self, X):
X = np.asarray(X)
if self.normalizer is not None:
X = self.normalizer.inverse_transform(X)
return self.clip(X)
| project_result | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n X = np.asarray(X)\n if self.normalizer is not None:\n X = self.normalizer.inverse_transform(X)\n return self.clip(X)\n",
"role": "user"
},
{
"content": "project_result",
"role": "assistant"
}
] |
def x(self, X, y=None):
X_original, missing_mask = self.prepare_input_data(X)
observed_mask = ~missing_mask
X = X_original.copy()
if self.normalizer is not None:
X = self.normalizer.fit_transform(X)
X_filled = self.fill(X, missing_mask, inplace=True)
if not isinstance(X_filled, np.ndarray):
raise TypeError(
'Expected %s.fill() to return NumPy array but got %s' % (self.
__class__.__name__, type(X_filled)))
X_result = self.solve(X_filled, missing_mask)
if not isinstance(X_result, np.ndarray):
raise TypeError(
'Expected %s.solve() to return NumPy array but got %s' % (self.
__class__.__name__, type(X_result)))
X_result = self.project_result(X=X_result)
X_result[observed_mask] = X_original[observed_mask]
return X_result
| fit_transform | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, y=None):\n X_original, missing_mask = self.prepare_input_data(X)\n observed_mask = ~missing_mask\n X = X_original.copy()\n if self.normalizer is not None:\n X = self.normalizer.fit_transform(X)\n X_filled = self.fill(X, missing_mask, inplace=True)\n if not isinstance(X_filled, np.ndarray):\n raise TypeError(\n 'Expected %s.fill() to return NumPy array but got %s' % (self.\n __class__.__name__, type(X_filled)))\n X_result = self.solve(X_filled, missing_mask)\n if not isinstance(X_result, np.ndarray):\n raise TypeError(\n 'Expected %s.solve() to return NumPy array but got %s' % (self.\n __class__.__name__, type(X_result)))\n X_result = self.project_result(X=X_result)\n X_result[observed_mask] = X_original[observed_mask]\n return X_result\n",
"role": "user"
},
{
"content": "fit_transform",
"role": "assistant"
}
] |
def x(self, X, y=None):
raise ValueError(
"%s.transform not implemented! This imputation algorithm likely doesn't support inductive mode. Only %s.fit_transform is supported at this time."
% (self.__class__.__name__, self.__class__.__name__))
| transform | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, y=None):\n raise ValueError(\n \"%s.transform not implemented! This imputation algorithm likely doesn't support inductive mode. Only %s.fit_transform is supported at this time.\"\n % (self.__class__.__name__, self.__class__.__name__))\n",
"role": "user"
},
{
"content": "transform",
"role": "assistant"
}
] |
def x(images, order='C'):
images = np.asarray(images, order=order).astype('float32')
images -= images.min()
images /= images.max()
return images
| rescale_pixel_values | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(images, order='C'):\n images = np.asarray(images, order=order).astype('float32')\n images -= images.min()\n images /= images.max()\n return images\n",
"role": "user"
},
{
"content": "rescale_pixel_values",
"role": "assistant"
}
] |
def x(self):
for i, (name, mae) in enumerate(sorted(self.mae_dict.items(), key=lambda
x: x[1])):
yield i + 1, name, self.mse_dict[name], self.mae_dict[name]
| sorted_errors | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n for i, (name, mae) in enumerate(sorted(self.mae_dict.items(), key=lambda\n x: x[1])):\n yield i + 1, name, self.mse_dict[name], self.mae_dict[name]\n",
"role": "user"
},
{
"content": "sorted_errors",
"role": "assistant"
}
] |
def x(x):
return bool(isinstance(x, (numbers.Real, np.floating)) and np.isnan(x))
| is_scalar_nan | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(x):\n return bool(isinstance(x, (numbers.Real, np.floating)) and np.isnan(x))\n",
"role": "user"
},
{
"content": "is_scalar_nan",
"role": "assistant"
}
] |
def x(X, value_to_mask):
if is_scalar_nan(value_to_mask):
if X.dtype.kind == 'f':
return np.isnan(X)
elif X.dtype.kind in ('i', 'u'):
return np.zeros(X.shape, dtype=bool)
else:
return _object_dtype_isnan(X)
else:
return np.equal(X, value_to_mask)
| get_mask | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(X, value_to_mask):\n if is_scalar_nan(value_to_mask):\n if X.dtype.kind == 'f':\n return np.isnan(X)\n elif X.dtype.kind in ('i', 'u'):\n return np.zeros(X.shape, dtype=bool)\n else:\n return _object_dtype_isnan(X)\n else:\n return np.equal(X, value_to_mask)\n",
"role": "user"
},
{
"content": "get_mask",
"role": "assistant"
}
] |
def x(array, extra_value, n_repeat):
if array.size > 0:
with warnings.catch_warnings():
warnings.simplefilter('ignore', RuntimeWarning)
mode = stats.mode(array)
most_frequent_value = mode[0][0]
most_frequent_count = mode[1][0]
else:
most_frequent_value = 0
most_frequent_count = 0
if most_frequent_count == 0 and n_repeat == 0:
return np.nan
elif most_frequent_count < n_repeat:
return extra_value
elif most_frequent_count > n_repeat:
return most_frequent_value
elif most_frequent_count == n_repeat:
if most_frequent_value < extra_value:
return most_frequent_value
else:
return extra_value
| most_frequent | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(array, extra_value, n_repeat):\n if array.size > 0:\n with warnings.catch_warnings():\n warnings.simplefilter('ignore', RuntimeWarning)\n mode = stats.mode(array)\n most_frequent_value = mode[0][0]\n most_frequent_count = mode[1][0]\n else:\n most_frequent_value = 0\n most_frequent_count = 0\n if most_frequent_count == 0 and n_repeat == 0:\n return np.nan\n elif most_frequent_count < n_repeat:\n return extra_value\n elif most_frequent_count > n_repeat:\n return most_frequent_value\n elif most_frequent_count == n_repeat:\n if most_frequent_value < extra_value:\n return most_frequent_value\n else:\n return extra_value\n",
"role": "user"
},
{
"content": "most_frequent",
"role": "assistant"
}
] |
def x(self, X, y=None):
X = self._validate_input(X)
if self.fill_value is None:
if X.dtype.kind in ('i', 'u', 'f'):
fill_value = 0
else:
fill_value = 'missing_value'
else:
fill_value = self.fill_value
if self.strategy == 'constant' and X.dtype.kind in ('i', 'u', 'f'
) and not isinstance(fill_value, numbers.Real):
raise ValueError(
"'fill_value'={0} is invalid. Expected a numerical value when imputing numerical data"
.format(fill_value))
if sparse.issparse(X):
if self.missing_values == 0:
raise ValueError(
'Imputation not possible when missing_values == 0 and input is sparse. Provide a dense array instead.'
)
else:
self.statistics_ = self._sparse_fit(X, self.strategy, self.
missing_values, fill_value)
else:
self.statistics_ = self._dense_fit(X, self.strategy, self.
missing_values, fill_value)
return self
| fit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, y=None):\n X = self._validate_input(X)\n if self.fill_value is None:\n if X.dtype.kind in ('i', 'u', 'f'):\n fill_value = 0\n else:\n fill_value = 'missing_value'\n else:\n fill_value = self.fill_value\n if self.strategy == 'constant' and X.dtype.kind in ('i', 'u', 'f'\n ) and not isinstance(fill_value, numbers.Real):\n raise ValueError(\n \"'fill_value'={0} is invalid. Expected a numerical value when imputing numerical data\"\n .format(fill_value))\n if sparse.issparse(X):\n if self.missing_values == 0:\n raise ValueError(\n 'Imputation not possible when missing_values == 0 and input is sparse. Provide a dense array instead.'\n )\n else:\n self.statistics_ = self._sparse_fit(X, self.strategy, self.\n missing_values, fill_value)\n else:\n self.statistics_ = self._dense_fit(X, self.strategy, self.\n missing_values, fill_value)\n return self\n",
"role": "user"
},
{
"content": "fit",
"role": "assistant"
}
] |
def x(self, X, strategy, missing_values, fill_value):
mask_data = _get_mask(X.data, missing_values)
n_implicit_zeros = X.shape[0] - np.diff(X.indptr)
statistics = np.empty(X.shape[1])
if strategy == 'constant':
statistics.fill(fill_value)
else:
for i in range(X.shape[1]):
column = X.data[X.indptr[i]:X.indptr[i + 1]]
mask_column = mask_data[X.indptr[i]:X.indptr[i + 1]]
column = column[~mask_column]
mask_zeros = _get_mask(column, 0)
column = column[~mask_zeros]
n_explicit_zeros = mask_zeros.sum()
n_zeros = n_implicit_zeros[i] + n_explicit_zeros
if strategy == 'mean':
s = column.size + n_zeros
statistics[i] = np.nan if s == 0 else column.sum() / s
elif strategy == 'median':
statistics[i] = _get_median(column, n_zeros)
elif strategy == 'most_frequent':
statistics[i] = _most_frequent(column, 0, n_zeros)
return statistics
| sparse_fit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, strategy, missing_values, fill_value):\n mask_data = _get_mask(X.data, missing_values)\n n_implicit_zeros = X.shape[0] - np.diff(X.indptr)\n statistics = np.empty(X.shape[1])\n if strategy == 'constant':\n statistics.fill(fill_value)\n else:\n for i in range(X.shape[1]):\n column = X.data[X.indptr[i]:X.indptr[i + 1]]\n mask_column = mask_data[X.indptr[i]:X.indptr[i + 1]]\n column = column[~mask_column]\n mask_zeros = _get_mask(column, 0)\n column = column[~mask_zeros]\n n_explicit_zeros = mask_zeros.sum()\n n_zeros = n_implicit_zeros[i] + n_explicit_zeros\n if strategy == 'mean':\n s = column.size + n_zeros\n statistics[i] = np.nan if s == 0 else column.sum() / s\n elif strategy == 'median':\n statistics[i] = _get_median(column, n_zeros)\n elif strategy == 'most_frequent':\n statistics[i] = _most_frequent(column, 0, n_zeros)\n return statistics\n",
"role": "user"
},
{
"content": "sparse_fit",
"role": "assistant"
}
] |
def x(self, X, strategy, missing_values, fill_value):
mask = _get_mask(X, missing_values)
masked_X = ma.masked_array(X, mask=mask)
if strategy == 'mean':
mean_masked = np.ma.mean(masked_X, axis=0)
mean = np.ma.getdata(mean_masked)
mean[np.ma.getmask(mean_masked)] = np.nan
return mean
elif strategy == 'median':
median_masked = np.ma.median(masked_X, axis=0)
median = np.ma.getdata(median_masked)
median[np.ma.getmaskarray(median_masked)] = np.nan
return median
elif strategy == 'most_frequent':
X = X.transpose()
mask = mask.transpose()
if X.dtype.kind == 'O':
most_frequent = np.empty(X.shape[0], dtype=object)
else:
most_frequent = np.empty(X.shape[0])
for i, (row, row_mask) in enumerate(zip(X[:], mask[:])):
row_mask = np.logical_not(row_mask).astype(np.bool)
row = row[row_mask]
most_frequent[i] = _most_frequent(row, np.nan, 0)
return most_frequent
elif strategy == 'constant':
return np.full(X.shape[1], fill_value, dtype=X.dtype)
| dense_fit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, strategy, missing_values, fill_value):\n mask = _get_mask(X, missing_values)\n masked_X = ma.masked_array(X, mask=mask)\n if strategy == 'mean':\n mean_masked = np.ma.mean(masked_X, axis=0)\n mean = np.ma.getdata(mean_masked)\n mean[np.ma.getmask(mean_masked)] = np.nan\n return mean\n elif strategy == 'median':\n median_masked = np.ma.median(masked_X, axis=0)\n median = np.ma.getdata(median_masked)\n median[np.ma.getmaskarray(median_masked)] = np.nan\n return median\n elif strategy == 'most_frequent':\n X = X.transpose()\n mask = mask.transpose()\n if X.dtype.kind == 'O':\n most_frequent = np.empty(X.shape[0], dtype=object)\n else:\n most_frequent = np.empty(X.shape[0])\n for i, (row, row_mask) in enumerate(zip(X[:], mask[:])):\n row_mask = np.logical_not(row_mask).astype(np.bool)\n row = row[row_mask]\n most_frequent[i] = _most_frequent(row, np.nan, 0)\n return most_frequent\n elif strategy == 'constant':\n return np.full(X.shape[1], fill_value, dtype=X.dtype)\n",
"role": "user"
},
{
"content": "dense_fit",
"role": "assistant"
}
] |
def x(self, X):
check_is_fitted(self, 'statistics_')
X = self._validate_input(X)
statistics = self.statistics_
if X.shape[1] != statistics.shape[0]:
raise ValueError('X has %d features per sample, expected %d' % (X.
shape[1], self.statistics_.shape[0]))
if self.strategy == 'constant':
valid_statistics = statistics
else:
invalid_mask = _get_mask(statistics, np.nan)
valid_mask = np.logical_not(invalid_mask)
valid_statistics = statistics[valid_mask]
valid_statistics_indexes = np.flatnonzero(valid_mask)
if invalid_mask.any():
missing = np.arange(X.shape[1])[invalid_mask]
if self.verbose:
warnings.warn(
'Deleting features without observed values: %s' % missing)
X = X[:, valid_statistics_indexes]
if sparse.issparse(X):
if self.missing_values == 0:
raise ValueError(
'Imputation not possible when missing_values == 0 and input is sparse. Provide a dense array instead.'
)
else:
mask = _get_mask(X.data, self.missing_values)
indexes = np.repeat(np.arange(len(X.indptr) - 1, dtype=np.int),
np.diff(X.indptr))[mask]
X.data[mask] = valid_statistics[indexes].astype(X.dtype, copy=False
)
else:
mask = _get_mask(X, self.missing_values)
n_missing = np.sum(mask, axis=0)
values = np.repeat(valid_statistics, n_missing)
coordinates = np.where(mask.transpose())[::-1]
X[coordinates] = values
return X
| transform | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n check_is_fitted(self, 'statistics_')\n X = self._validate_input(X)\n statistics = self.statistics_\n if X.shape[1] != statistics.shape[0]:\n raise ValueError('X has %d features per sample, expected %d' % (X.\n shape[1], self.statistics_.shape[0]))\n if self.strategy == 'constant':\n valid_statistics = statistics\n else:\n invalid_mask = _get_mask(statistics, np.nan)\n valid_mask = np.logical_not(invalid_mask)\n valid_statistics = statistics[valid_mask]\n valid_statistics_indexes = np.flatnonzero(valid_mask)\n if invalid_mask.any():\n missing = np.arange(X.shape[1])[invalid_mask]\n if self.verbose:\n warnings.warn(\n 'Deleting features without observed values: %s' % missing)\n X = X[:, valid_statistics_indexes]\n if sparse.issparse(X):\n if self.missing_values == 0:\n raise ValueError(\n 'Imputation not possible when missing_values == 0 and input is sparse. Provide a dense array instead.'\n )\n else:\n mask = _get_mask(X.data, self.missing_values)\n indexes = np.repeat(np.arange(len(X.indptr) - 1, dtype=np.int),\n np.diff(X.indptr))[mask]\n X.data[mask] = valid_statistics[indexes].astype(X.dtype, copy=False\n )\n else:\n mask = _get_mask(X, self.missing_values)\n n_missing = np.sum(mask, axis=0)\n values = np.repeat(valid_statistics, n_missing)\n coordinates = np.where(mask.transpose())[::-1]\n X[coordinates] = values\n return X\n",
"role": "user"
},
{
"content": "transform",
"role": "assistant"
}
] |
def x(self, X_filled, mask_missing_values, feat_idx, neighbor_feat_idx,
predictor=None, fit_mode=True):
missing_row_mask = mask_missing_values[:, feat_idx]
if not np.any(missing_row_mask):
return X_filled, predictor
if predictor is None and fit_mode is False:
raise ValueError(
'If fit_mode is False, then an already-fitted predictor should be passed in.'
)
if predictor is None:
predictor = clone(self._predictor)
if fit_mode:
X_train = safe_indexing(X_filled[:, neighbor_feat_idx], ~
missing_row_mask)
y_train = safe_indexing(X_filled[:, feat_idx], ~missing_row_mask)
predictor.fit(X_train, y_train)
X_test = safe_indexing(X_filled[:, neighbor_feat_idx], missing_row_mask)
if self.sample_posterior:
mus, sigmas = predictor.predict(X_test, return_std=True)
good_sigmas = sigmas > 0
imputed_values = np.zeros(mus.shape, dtype=X_filled.dtype)
imputed_values[~good_sigmas] = mus[~good_sigmas]
imputed_values[good_sigmas] = self.random_state_.normal(loc=mus[
good_sigmas], scale=sigmas[good_sigmas])
else:
imputed_values = predictor.predict(X_test)
imputed_values = self.clip(imputed_values)
X_filled[missing_row_mask, feat_idx] = imputed_values
return X_filled, predictor
| impute_one_feature | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X_filled, mask_missing_values, feat_idx, neighbor_feat_idx,\n predictor=None, fit_mode=True):\n missing_row_mask = mask_missing_values[:, feat_idx]\n if not np.any(missing_row_mask):\n return X_filled, predictor\n if predictor is None and fit_mode is False:\n raise ValueError(\n 'If fit_mode is False, then an already-fitted predictor should be passed in.'\n )\n if predictor is None:\n predictor = clone(self._predictor)\n if fit_mode:\n X_train = safe_indexing(X_filled[:, neighbor_feat_idx], ~\n missing_row_mask)\n y_train = safe_indexing(X_filled[:, feat_idx], ~missing_row_mask)\n predictor.fit(X_train, y_train)\n X_test = safe_indexing(X_filled[:, neighbor_feat_idx], missing_row_mask)\n if self.sample_posterior:\n mus, sigmas = predictor.predict(X_test, return_std=True)\n good_sigmas = sigmas > 0\n imputed_values = np.zeros(mus.shape, dtype=X_filled.dtype)\n imputed_values[~good_sigmas] = mus[~good_sigmas]\n imputed_values[good_sigmas] = self.random_state_.normal(loc=mus[\n good_sigmas], scale=sigmas[good_sigmas])\n else:\n imputed_values = predictor.predict(X_test)\n imputed_values = self.clip(imputed_values)\n X_filled[missing_row_mask, feat_idx] = imputed_values\n return X_filled, predictor\n",
"role": "user"
},
{
"content": "impute_one_feature",
"role": "assistant"
}
] |
def x(self, n_features, feat_idx, abs_corr_mat):
if (self.n_nearest_features is not None and self.n_nearest_features <
n_features):
p = abs_corr_mat[:, feat_idx]
neighbor_feat_idx = self.random_state_.choice(np.arange(n_features),
self.n_nearest_features, replace=False, p=p)
else:
inds_left = np.arange(feat_idx)
inds_right = np.arange(feat_idx + 1, n_features)
neighbor_feat_idx = np.concatenate((inds_left, inds_right))
return neighbor_feat_idx
| get_neighbor_feat_idx | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, n_features, feat_idx, abs_corr_mat):\n if (self.n_nearest_features is not None and self.n_nearest_features <\n n_features):\n p = abs_corr_mat[:, feat_idx]\n neighbor_feat_idx = self.random_state_.choice(np.arange(n_features),\n self.n_nearest_features, replace=False, p=p)\n else:\n inds_left = np.arange(feat_idx)\n inds_right = np.arange(feat_idx + 1, n_features)\n neighbor_feat_idx = np.concatenate((inds_left, inds_right))\n return neighbor_feat_idx\n",
"role": "user"
},
{
"content": "get_neighbor_feat_idx",
"role": "assistant"
}
] |
def x(self, mask_missing_values):
frac_of_missing_values = mask_missing_values.mean(axis=0)
missing_values_idx = np.nonzero(frac_of_missing_values)[0]
if self.imputation_order == 'roman':
ordered_idx = missing_values_idx
elif self.imputation_order == 'arabic':
ordered_idx = missing_values_idx[::-1]
elif self.imputation_order == 'ascending':
n = len(frac_of_missing_values) - len(missing_values_idx)
ordered_idx = np.argsort(frac_of_missing_values, kind='mergesort')[n:][
::-1]
elif self.imputation_order == 'descending':
n = len(frac_of_missing_values) - len(missing_values_idx)
ordered_idx = np.argsort(frac_of_missing_values, kind='mergesort')[n:]
elif self.imputation_order == 'random':
ordered_idx = missing_values_idx
self.random_state_.shuffle(ordered_idx)
else:
raise ValueError(
"Got an invalid imputation order: '{0}'. It must be one of the following: 'roman', 'arabic', 'ascending', 'descending', or 'random'."
.format(self.imputation_order))
return ordered_idx
| get_ordered_idx | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, mask_missing_values):\n frac_of_missing_values = mask_missing_values.mean(axis=0)\n missing_values_idx = np.nonzero(frac_of_missing_values)[0]\n if self.imputation_order == 'roman':\n ordered_idx = missing_values_idx\n elif self.imputation_order == 'arabic':\n ordered_idx = missing_values_idx[::-1]\n elif self.imputation_order == 'ascending':\n n = len(frac_of_missing_values) - len(missing_values_idx)\n ordered_idx = np.argsort(frac_of_missing_values, kind='mergesort')[n:][\n ::-1]\n elif self.imputation_order == 'descending':\n n = len(frac_of_missing_values) - len(missing_values_idx)\n ordered_idx = np.argsort(frac_of_missing_values, kind='mergesort')[n:]\n elif self.imputation_order == 'random':\n ordered_idx = missing_values_idx\n self.random_state_.shuffle(ordered_idx)\n else:\n raise ValueError(\n \"Got an invalid imputation order: '{0}'. It must be one of the following: 'roman', 'arabic', 'ascending', 'descending', or 'random'.\"\n .format(self.imputation_order))\n return ordered_idx\n",
"role": "user"
},
{
"content": "get_ordered_idx",
"role": "assistant"
}
] |
def x(self, X_filled, tolerance=1e-06):
n_features = X_filled.shape[1]
if (self.n_nearest_features is None or self.n_nearest_features >=
n_features):
return None
abs_corr_mat = np.abs(np.corrcoef(X_filled.T))
abs_corr_mat[np.isnan(abs_corr_mat)] = tolerance
np.clip(abs_corr_mat, tolerance, None, out=abs_corr_mat)
np.fill_diagonal(abs_corr_mat, 0)
abs_corr_mat = normalize(abs_corr_mat, norm='l1', axis=0, copy=False)
return abs_corr_mat
| get_abs_corr_mat | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X_filled, tolerance=1e-06):\n n_features = X_filled.shape[1]\n if (self.n_nearest_features is None or self.n_nearest_features >=\n n_features):\n return None\n abs_corr_mat = np.abs(np.corrcoef(X_filled.T))\n abs_corr_mat[np.isnan(abs_corr_mat)] = tolerance\n np.clip(abs_corr_mat, tolerance, None, out=abs_corr_mat)\n np.fill_diagonal(abs_corr_mat, 0)\n abs_corr_mat = normalize(abs_corr_mat, norm='l1', axis=0, copy=False)\n return abs_corr_mat\n",
"role": "user"
},
{
"content": "get_abs_corr_mat",
"role": "assistant"
}
] |
def x(self, X):
if is_scalar_nan(self.missing_values):
force_all_finite = False
else:
force_all_finite = True
X = check_array(X, dtype=FLOAT_DTYPES, order='F', force_all_finite=
force_all_finite)
_check_inputs_dtype(X, self.missing_values)
mask_missing_values = _get_mask(X, self.missing_values)
if self.initial_imputer_ is None:
self.initial_imputer_ = _SimpleImputer(missing_values=self.
missing_values, strategy=self.initial_strategy)
X_filled = self.initial_imputer_.fit_transform(X)
else:
X_filled = self.initial_imputer_.transform(X)
valid_mask = np.flatnonzero(np.logical_not(np.isnan(self.
initial_imputer_.statistics_)))
Xt = X[:, valid_mask]
mask_missing_values = mask_missing_values[:, valid_mask]
return Xt, X_filled, mask_missing_values
| initial_imputation | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n if is_scalar_nan(self.missing_values):\n force_all_finite = False\n else:\n force_all_finite = True\n X = check_array(X, dtype=FLOAT_DTYPES, order='F', force_all_finite=\n force_all_finite)\n _check_inputs_dtype(X, self.missing_values)\n mask_missing_values = _get_mask(X, self.missing_values)\n if self.initial_imputer_ is None:\n self.initial_imputer_ = _SimpleImputer(missing_values=self.\n missing_values, strategy=self.initial_strategy)\n X_filled = self.initial_imputer_.fit_transform(X)\n else:\n X_filled = self.initial_imputer_.transform(X)\n valid_mask = np.flatnonzero(np.logical_not(np.isnan(self.\n initial_imputer_.statistics_)))\n Xt = X[:, valid_mask]\n mask_missing_values = mask_missing_values[:, valid_mask]\n return Xt, X_filled, mask_missing_values\n",
"role": "user"
},
{
"content": "initial_imputation",
"role": "assistant"
}
] |
def x(self, X, y=None):
self.random_state_ = getattr(self, 'random_state_', check_random_state(
self.random_state))
if self.n_iter < 0:
raise ValueError(
"'n_iter' should be a positive integer. Got {} instead.".format
(self.n_iter))
if self.predictor is None:
if self.sample_posterior:
from sklearn.linear_model import BayesianRidge
self._predictor = BayesianRidge()
else:
from sklearn.linear_model import RidgeCV
self._predictor = RidgeCV(alphas=np.array([1e-05, 0.1, 1, 10]))
else:
self._predictor = clone(self.predictor)
if hasattr(self._predictor, 'random_state'):
self._predictor.random_state = self.random_state_
self._min_value = np.nan if self.min_value is None else self.min_value
self._max_value = np.nan if self.max_value is None else self.max_value
self.initial_imputer_ = None
X, Xt, mask_missing_values = self._initial_imputation(X)
if self.n_iter == 0:
return Xt
ordered_idx = self._get_ordered_idx(mask_missing_values)
self.n_features_with_missing_ = len(ordered_idx)
abs_corr_mat = self._get_abs_corr_mat(Xt)
n_samples, n_features = Xt.shape
self.imputation_sequence_ = []
if self.verbose > 0:
print('[IterativeImputer] Completing matrix with shape %s' % (X.shape,)
)
start_t = time()
for i_rnd in range(self.n_iter):
if self.imputation_order == 'random':
ordered_idx = self._get_ordered_idx(mask_missing_values)
for feat_idx in ordered_idx:
neighbor_feat_idx = self._get_neighbor_feat_idx(n_features,
feat_idx, abs_corr_mat)
Xt, predictor = self._impute_one_feature(Xt,
mask_missing_values, feat_idx, neighbor_feat_idx, predictor
=None, fit_mode=True)
predictor_triplet = ImputerTriplet(feat_idx, neighbor_feat_idx,
predictor)
self.imputation_sequence_.append(predictor_triplet)
if self.verbose > 0:
print(
'[IterativeImputer] Ending imputation round %d/%d, elapsed time %0.2f'
% (i_rnd + 1, self.n_iter, time() - start_t))
Xt[~mask_missing_values] = X[~mask_missing_values]
return Xt
| fit_transform | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, y=None):\n self.random_state_ = getattr(self, 'random_state_', check_random_state(\n self.random_state))\n if self.n_iter < 0:\n raise ValueError(\n \"'n_iter' should be a positive integer. Got {} instead.\".format\n (self.n_iter))\n if self.predictor is None:\n if self.sample_posterior:\n from sklearn.linear_model import BayesianRidge\n self._predictor = BayesianRidge()\n else:\n from sklearn.linear_model import RidgeCV\n self._predictor = RidgeCV(alphas=np.array([1e-05, 0.1, 1, 10]))\n else:\n self._predictor = clone(self.predictor)\n if hasattr(self._predictor, 'random_state'):\n self._predictor.random_state = self.random_state_\n self._min_value = np.nan if self.min_value is None else self.min_value\n self._max_value = np.nan if self.max_value is None else self.max_value\n self.initial_imputer_ = None\n X, Xt, mask_missing_values = self._initial_imputation(X)\n if self.n_iter == 0:\n return Xt\n ordered_idx = self._get_ordered_idx(mask_missing_values)\n self.n_features_with_missing_ = len(ordered_idx)\n abs_corr_mat = self._get_abs_corr_mat(Xt)\n n_samples, n_features = Xt.shape\n self.imputation_sequence_ = []\n if self.verbose > 0:\n print('[IterativeImputer] Completing matrix with shape %s' % (X.shape,)\n )\n start_t = time()\n for i_rnd in range(self.n_iter):\n if self.imputation_order == 'random':\n ordered_idx = self._get_ordered_idx(mask_missing_values)\n for feat_idx in ordered_idx:\n neighbor_feat_idx = self._get_neighbor_feat_idx(n_features,\n feat_idx, abs_corr_mat)\n Xt, predictor = self._impute_one_feature(Xt,\n mask_missing_values, feat_idx, neighbor_feat_idx, predictor\n =None, fit_mode=True)\n predictor_triplet = ImputerTriplet(feat_idx, neighbor_feat_idx,\n predictor)\n self.imputation_sequence_.append(predictor_triplet)\n if self.verbose > 0:\n print(\n '[IterativeImputer] Ending imputation round %d/%d, elapsed time %0.2f'\n % (i_rnd + 1, self.n_iter, time() - start_t))\n Xt[~mask_missing_values] = X[~mask_missing_values]\n return Xt\n",
"role": "user"
},
{
"content": "fit_transform",
"role": "assistant"
}
] |
def x(self, X):
check_is_fitted(self, 'initial_imputer_')
X, Xt, mask_missing_values = self._initial_imputation(X)
if self.n_iter == 0:
return Xt
imputations_per_round = len(self.imputation_sequence_) // self.n_iter
i_rnd = 0
if self.verbose > 0:
print('[IterativeImputer] Completing matrix with shape %s' % (X.shape,)
)
start_t = time()
for it, predictor_triplet in enumerate(self.imputation_sequence_):
Xt, _ = self._impute_one_feature(Xt, mask_missing_values,
predictor_triplet.feat_idx, predictor_triplet.neighbor_feat_idx,
predictor=predictor_triplet.predictor, fit_mode=False)
if not (it + 1) % imputations_per_round:
if self.verbose > 1:
print(
'[IterativeImputer] Ending imputation round %d/%d, elapsed time %0.2f'
% (i_rnd + 1, self.n_iter, time() - start_t))
i_rnd += 1
Xt[~mask_missing_values] = X[~mask_missing_values]
return Xt
| transform | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n check_is_fitted(self, 'initial_imputer_')\n X, Xt, mask_missing_values = self._initial_imputation(X)\n if self.n_iter == 0:\n return Xt\n imputations_per_round = len(self.imputation_sequence_) // self.n_iter\n i_rnd = 0\n if self.verbose > 0:\n print('[IterativeImputer] Completing matrix with shape %s' % (X.shape,)\n )\n start_t = time()\n for it, predictor_triplet in enumerate(self.imputation_sequence_):\n Xt, _ = self._impute_one_feature(Xt, mask_missing_values,\n predictor_triplet.feat_idx, predictor_triplet.neighbor_feat_idx,\n predictor=predictor_triplet.predictor, fit_mode=False)\n if not (it + 1) % imputations_per_round:\n if self.verbose > 1:\n print(\n '[IterativeImputer] Ending imputation round %d/%d, elapsed time %0.2f'\n % (i_rnd + 1, self.n_iter, time() - start_t))\n i_rnd += 1\n Xt[~mask_missing_values] = X[~mask_missing_values]\n return Xt\n",
"role": "user"
},
{
"content": "transform",
"role": "assistant"
}
] |
def x(self, X, missing_mask, S, error_tolerance):
ok_mask = ~missing_mask
masked_X = cvxpy.multiply(ok_mask, X)
masked_S = cvxpy.multiply(ok_mask, S)
abs_diff = cvxpy.abs(masked_S - masked_X)
close_to_data = abs_diff <= error_tolerance
constraints = [close_to_data]
if self.require_symmetric_solution:
constraints.append(S == S.T)
if self.min_value is not None:
constraints.append(S >= self.min_value)
if self.max_value is not None:
constraints.append(S <= self.max_value)
return constraints
| constraints | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, missing_mask, S, error_tolerance):\n ok_mask = ~missing_mask\n masked_X = cvxpy.multiply(ok_mask, X)\n masked_S = cvxpy.multiply(ok_mask, S)\n abs_diff = cvxpy.abs(masked_S - masked_X)\n close_to_data = abs_diff <= error_tolerance\n constraints = [close_to_data]\n if self.require_symmetric_solution:\n constraints.append(S == S.T)\n if self.min_value is not None:\n constraints.append(S >= self.min_value)\n if self.max_value is not None:\n constraints.append(S <= self.max_value)\n return constraints\n",
"role": "user"
},
{
"content": "constraints",
"role": "assistant"
}
] |
def x(self, m, n):
shape = m, n
S = cvxpy.Variable(shape, name='S')
norm = cvxpy.norm(S, 'nuc')
objective = cvxpy.Minimize(norm)
return S, objective
| create_objective | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, m, n):\n shape = m, n\n S = cvxpy.Variable(shape, name='S')\n norm = cvxpy.norm(S, 'nuc')\n objective = cvxpy.Minimize(norm)\n return S, objective\n",
"role": "user"
},
{
"content": "create_objective",
"role": "assistant"
}
] |
def x(module, name):
module = importlib.import_module(module)
return getattr(module, name)
| import_from | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(module, name):\n module = importlib.import_module(module)\n return getattr(module, name)\n",
"role": "user"
},
{
"content": "import_from",
"role": "assistant"
}
] |
def x(X_original, solutions, parameters=None, verbose=False, percentiles=
list(range(10, 100, 10))):
missing_mask = np.isnan(X_original)
min_mse = np.inf
best_solution = None
for i, candidate in enumerate(solutions):
for col_idx in range(X_original.shape[1]):
col_data = candidate[:, col_idx]
col_missing = missing_mask[:, col_idx]
col_observed = ~col_missing
if col_missing.sum() < 2:
continue
elif col_observed.sum() < 2:
continue
missing_data = col_data[col_missing]
observed_data = col_data[col_observed]
missing_percentiles = np.array([np.percentile(missing_data, p) for
p in percentiles])
observed_percentiles = np.array([np.percentile(observed_data, p
) for p in percentiles])
mse = np.mean((missing_percentiles - observed_percentiles) ** 2)
if mse < min_mse:
min_mse = mse
best_solution = candidate
if verbose:
print('Candidate #%d/%d%s: %f' % (i + 1, len(solutions),
' (parameter=%s) ' % parameters[i] if parameters is not
None else '', mse))
return best_solution
| choose_solution_using_percentiles | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(X_original, solutions, parameters=None, verbose=False, percentiles=\n list(range(10, 100, 10))):\n missing_mask = np.isnan(X_original)\n min_mse = np.inf\n best_solution = None\n for i, candidate in enumerate(solutions):\n for col_idx in range(X_original.shape[1]):\n col_data = candidate[:, col_idx]\n col_missing = missing_mask[:, col_idx]\n col_observed = ~col_missing\n if col_missing.sum() < 2:\n continue\n elif col_observed.sum() < 2:\n continue\n missing_data = col_data[col_missing]\n observed_data = col_data[col_observed]\n missing_percentiles = np.array([np.percentile(missing_data, p) for\n p in percentiles])\n observed_percentiles = np.array([np.percentile(observed_data, p\n ) for p in percentiles])\n mse = np.mean((missing_percentiles - observed_percentiles) ** 2)\n if mse < min_mse:\n min_mse = mse\n best_solution = candidate\n if verbose:\n print('Candidate #%d/%d%s: %f' % (i + 1, len(solutions), \n ' (parameter=%s) ' % parameters[i] if parameters is not\n None else '', mse))\n return best_solution\n",
"role": "user"
},
{
"content": "choose_solution_using_percentiles",
"role": "assistant"
}
] |
def x(self, X, observed, column_means, column_scales):
n_rows, n_cols = X.shape
column_means = np.asarray(column_means)
if len(column_means) != n_cols:
raise ValueError('Expected length %d but got shape %s' % (n_cols,
column_means.shape))
X = X - column_means.reshape((1, n_cols))
column_weights = 1.0 / column_scales
X *= column_weights.reshape((1, n_cols))
row_means = np.zeros(n_rows, dtype=X.dtype)
row_residual_sums = np.nansum(X, axis=1)
for i in range(n_rows):
row_mask = observed[i, :]
sum_weights = column_weights[row_mask].sum()
row_means[i] = row_residual_sums[i] / sum_weights
return row_means
| estimate_row_means | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, observed, column_means, column_scales):\n n_rows, n_cols = X.shape\n column_means = np.asarray(column_means)\n if len(column_means) != n_cols:\n raise ValueError('Expected length %d but got shape %s' % (n_cols,\n column_means.shape))\n X = X - column_means.reshape((1, n_cols))\n column_weights = 1.0 / column_scales\n X *= column_weights.reshape((1, n_cols))\n row_means = np.zeros(n_rows, dtype=X.dtype)\n row_residual_sums = np.nansum(X, axis=1)\n for i in range(n_rows):\n row_mask = observed[i, :]\n sum_weights = column_weights[row_mask].sum()\n row_means[i] = row_residual_sums[i] / sum_weights\n return row_means\n",
"role": "user"
},
{
"content": "estimate_row_means",
"role": "assistant"
}
] |
def x(self, X, observed, row_means, row_scales):
n_rows, n_cols = X.shape
row_means = np.asarray(row_means)
if len(row_means) != n_rows:
raise ValueError('Expected length %d but got shape %s' % (n_rows,
row_means.shape))
column_means = np.zeros(n_cols, dtype=X.dtype)
X = X - row_means.reshape((n_rows, 1))
row_weights = 1.0 / row_scales
X *= row_weights.reshape((n_rows, 1))
col_residual_sums = np.nansum(X, axis=0)
for j in range(n_cols):
col_mask = observed[:, j]
sum_weights = row_weights[col_mask].sum()
column_means[j] = col_residual_sums[j] / sum_weights
return column_means
| estimate_column_means | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, observed, row_means, row_scales):\n n_rows, n_cols = X.shape\n row_means = np.asarray(row_means)\n if len(row_means) != n_rows:\n raise ValueError('Expected length %d but got shape %s' % (n_rows,\n row_means.shape))\n column_means = np.zeros(n_cols, dtype=X.dtype)\n X = X - row_means.reshape((n_rows, 1))\n row_weights = 1.0 / row_scales\n X *= row_weights.reshape((n_rows, 1))\n col_residual_sums = np.nansum(X, axis=0)\n for j in range(n_cols):\n col_mask = observed[:, j]\n sum_weights = row_weights[col_mask].sum()\n column_means[j] = col_residual_sums[j] / sum_weights\n return column_means\n",
"role": "user"
},
{
"content": "estimate_column_means",
"role": "assistant"
}
] |
def x(self, X_centered, column_scales):
n_rows, n_cols = X_centered.shape
column_scales = np.asarray(column_scales)
if len(column_scales) != n_cols:
raise ValueError('Expected length %d but got shape %s' % (n_cols,
column_scales))
row_variances = np.nanmean(X_centered ** 2 / (column_scales ** 2).
reshape((1, n_cols)), axis=1)
row_variances[row_variances == 0] = 1.0
assert len(row_variances) == n_rows, '%d != %d' % (len(row_variances),
n_rows)
return np.sqrt(row_variances)
| estimate_row_scales | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X_centered, column_scales):\n n_rows, n_cols = X_centered.shape\n column_scales = np.asarray(column_scales)\n if len(column_scales) != n_cols:\n raise ValueError('Expected length %d but got shape %s' % (n_cols,\n column_scales))\n row_variances = np.nanmean(X_centered ** 2 / (column_scales ** 2).\n reshape((1, n_cols)), axis=1)\n row_variances[row_variances == 0] = 1.0\n assert len(row_variances) == n_rows, '%d != %d' % (len(row_variances),\n n_rows)\n return np.sqrt(row_variances)\n",
"role": "user"
},
{
"content": "estimate_row_scales",
"role": "assistant"
}
] |
def x(self, X_centered, row_scales):
n_rows, n_cols = X_centered.shape
row_scales = np.asarray(row_scales)
if len(row_scales) != n_rows:
raise ValueError('Expected length %s, got shape %s' % (n_rows,
row_scales.shape))
column_variances = np.nanmean(X_centered ** 2 / (row_scales ** 2).
reshape((n_rows, 1)), axis=0)
column_variances[column_variances == 0] = 1.0
assert len(column_variances) == n_cols, '%d != %d' % (len(
column_variances), n_cols)
return np.sqrt(column_variances)
| estimate_column_scales | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X_centered, row_scales):\n n_rows, n_cols = X_centered.shape\n row_scales = np.asarray(row_scales)\n if len(row_scales) != n_rows:\n raise ValueError('Expected length %s, got shape %s' % (n_rows,\n row_scales.shape))\n column_variances = np.nanmean(X_centered ** 2 / (row_scales ** 2).\n reshape((n_rows, 1)), axis=0)\n column_variances[column_variances == 0] = 1.0\n assert len(column_variances) == n_cols, '%d != %d' % (len(\n column_variances), n_cols)\n return np.sqrt(column_variances)\n",
"role": "user"
},
{
"content": "estimate_column_scales",
"role": "assistant"
}
] |
def x(self, nested_dictionaries):
sims = {}
overlaps = {}
weights = {}
for a, column_dict_a in nested_dictionaries.items():
row_set_a = set(column_dict_a.keys())
for b, column_dict_b in nested_dictionaries.items():
row_set_b = set(column_dict_b.keys())
common_rows = row_set_a.intersection(row_set_b)
n_overlap = len(common_rows)
overlaps[a, b] = n_overlap
total = 0.0
weight = 0.0
for row_name in common_rows:
value_a = column_dict_a[row_name]
value_b = column_dict_b[row_name]
minval = min(value_a, value_b)
maxval = max(value_a, value_b)
total += minval
weight += maxval
weights[a, b] = weight
if weight < self.min_weight_for_similarity:
continue
if n_overlap < self.min_count_for_similarity:
continue
sims[a, b] = total / weight
return sims, overlaps, weights
| jacard_similarity_from_nested_dicts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, nested_dictionaries):\n sims = {}\n overlaps = {}\n weights = {}\n for a, column_dict_a in nested_dictionaries.items():\n row_set_a = set(column_dict_a.keys())\n for b, column_dict_b in nested_dictionaries.items():\n row_set_b = set(column_dict_b.keys())\n common_rows = row_set_a.intersection(row_set_b)\n n_overlap = len(common_rows)\n overlaps[a, b] = n_overlap\n total = 0.0\n weight = 0.0\n for row_name in common_rows:\n value_a = column_dict_a[row_name]\n value_b = column_dict_b[row_name]\n minval = min(value_a, value_b)\n maxval = max(value_a, value_b)\n total += minval\n weight += maxval\n weights[a, b] = weight\n if weight < self.min_weight_for_similarity:\n continue\n if n_overlap < self.min_count_for_similarity:\n continue\n sims[a, b] = total / weight\n return sims, overlaps, weights\n",
"role": "user"
},
{
"content": "jacard_similarity_from_nested_dicts",
"role": "assistant"
}
] |
def x(self, values_dict):
if self.orientation != 'rows':
values_dict = transpose_nested_dictionary(values_dict)
row_keys, column_keys = collect_nested_keys(values_dict)
if self.verbose:
print('[SimilarityWeightedAveraging] # rows = %d' % len(row_keys))
print('[SimilarityWeightedAveraging] # columns = %d' % len(column_keys)
)
similarities, overlaps, weights = self.jacard_similarity_from_nested_dicts(
values_dict)
if self.verbose:
print(
'[SimilarityWeightedAveraging] Computed %d similarities between rows'
% (len(similarities),))
column_to_row_values = reverse_lookup_from_nested_dict(values_dict)
result = defaultdict(dict)
exponent = self.similarity_exponent
shrinkage_coef = self.shrinkage_coef
for i, row_key in enumerate(row_keys):
for column_key, value_triplets in column_to_row_values.items():
total = 0
denom = shrinkage_coef
for other_row_key, y in value_triplets:
sample_weight = 1.0
sim = similarities.get((row_key, other_row_key), 0)
combined_weight = sim ** exponent
combined_weight *= sample_weight
total += combined_weight * y
denom += combined_weight
if denom > shrinkage_coef:
result[row_key][column_key] = total / denom
if self.orientation != 'rows':
result = transpose_nested_dictionary(result)
return result
| complete_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, values_dict):\n if self.orientation != 'rows':\n values_dict = transpose_nested_dictionary(values_dict)\n row_keys, column_keys = collect_nested_keys(values_dict)\n if self.verbose:\n print('[SimilarityWeightedAveraging] # rows = %d' % len(row_keys))\n print('[SimilarityWeightedAveraging] # columns = %d' % len(column_keys)\n )\n similarities, overlaps, weights = self.jacard_similarity_from_nested_dicts(\n values_dict)\n if self.verbose:\n print(\n '[SimilarityWeightedAveraging] Computed %d similarities between rows'\n % (len(similarities),))\n column_to_row_values = reverse_lookup_from_nested_dict(values_dict)\n result = defaultdict(dict)\n exponent = self.similarity_exponent\n shrinkage_coef = self.shrinkage_coef\n for i, row_key in enumerate(row_keys):\n for column_key, value_triplets in column_to_row_values.items():\n total = 0\n denom = shrinkage_coef\n for other_row_key, y in value_triplets:\n sample_weight = 1.0\n sim = similarities.get((row_key, other_row_key), 0)\n combined_weight = sim ** exponent\n combined_weight *= sample_weight\n total += combined_weight * y\n denom += combined_weight\n if denom > shrinkage_coef:\n result[row_key][column_key] = total / denom\n if self.orientation != 'rows':\n result = transpose_nested_dictionary(result)\n return result\n",
"role": "user"
},
{
"content": "complete_dict",
"role": "assistant"
}
] |
def x(nested_dict):
outer_keys, inner_keys = collect_nested_keys(nested_dict)
outer_key_indices = {k: i for i, k in enumerate(outer_keys)}
inner_key_indices = {k: i for i, k in enumerate(inner_keys)}
return outer_key_indices, inner_key_indices
| nested_key_indices | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(nested_dict):\n outer_keys, inner_keys = collect_nested_keys(nested_dict)\n outer_key_indices = {k: i for i, k in enumerate(outer_keys)}\n inner_key_indices = {k: i for i, k in enumerate(inner_keys)}\n return outer_key_indices, inner_key_indices\n",
"role": "user"
},
{
"content": "nested_key_indices",
"role": "assistant"
}
] |
def x(nested_dict):
outer_keys, inner_keys = collect_nested_keys(nested_dict)
combined_keys = list(sorted(set(outer_keys + inner_keys)))
return {k: i for i, k in enumerate(combined_keys)}
| flattened_nested_key_indices | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(nested_dict):\n outer_keys, inner_keys = collect_nested_keys(nested_dict)\n combined_keys = list(sorted(set(outer_keys + inner_keys)))\n return {k: i for i, k in enumerate(combined_keys)}\n",
"role": "user"
},
{
"content": "flattened_nested_key_indices",
"role": "assistant"
}
] |
def x(nested_dict, array_fn, dtype='float32', square_result=False):
if square_result:
outer_key_indices = inner_key_indices = flattened_nested_key_indices(
nested_dict)
else:
outer_key_indices, inner_key_indices = nested_key_indices(nested_dict)
n_rows = len(outer_key_indices)
n_cols = len(inner_key_indices)
shape = n_rows, n_cols
result = array_fn(shape, dtype)
for outer_key, sub_dictionary in nested_dict.items():
i = outer_key_indices[outer_key]
for inner_key, value in sub_dictionary.items():
j = inner_key_indices[inner_key]
result[i, j] = value
outer_key_list = index_dict_to_sorted_list(outer_key_indices)
inner_key_list = index_dict_to_sorted_list(inner_key_indices)
return result, outer_key_list, inner_key_list
| array_from_nested_dictionary | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(nested_dict, array_fn, dtype='float32', square_result=False):\n if square_result:\n outer_key_indices = inner_key_indices = flattened_nested_key_indices(\n nested_dict)\n else:\n outer_key_indices, inner_key_indices = nested_key_indices(nested_dict)\n n_rows = len(outer_key_indices)\n n_cols = len(inner_key_indices)\n shape = n_rows, n_cols\n result = array_fn(shape, dtype)\n for outer_key, sub_dictionary in nested_dict.items():\n i = outer_key_indices[outer_key]\n for inner_key, value in sub_dictionary.items():\n j = inner_key_indices[inner_key]\n result[i, j] = value\n outer_key_list = index_dict_to_sorted_list(outer_key_indices)\n inner_key_list = index_dict_to_sorted_list(inner_key_indices)\n return result, outer_key_list, inner_key_list\n",
"role": "user"
},
{
"content": "array_from_nested_dictionary",
"role": "assistant"
}
] |
def x(X, row_keys=None, column_keys=None, filter_fn=None):
n_rows, n_cols = X.shape
if row_keys is None:
row_keys = {i: i for i in range(n_rows)}
if column_keys is None:
if n_rows == n_cols:
column_keys = row_keys
else:
column_keys = {j: j for j in range(n_cols)}
if len(row_keys) != n_rows:
raise ValueError('Need %d row keys but got list of length %d' % (
n_rows, len(row_keys)))
if len(column_keys) != n_cols:
raise ValueError('Need %d column keys but got list of length %d' %
(n_cols, len(column_keys)))
result_dict = {}
for i, X_i in enumerate(X):
row_key = row_keys[i]
for j, X_ij in enumerate(X_i):
if filter_fn and not filter_fn(X_ij):
continue
column_key = column_keys[j]
key_pair = row_key, column_key
result_dict[key_pair] = X_ij
return result_dict
| matrix_to_pair_dictionary | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(X, row_keys=None, column_keys=None, filter_fn=None):\n n_rows, n_cols = X.shape\n if row_keys is None:\n row_keys = {i: i for i in range(n_rows)}\n if column_keys is None:\n if n_rows == n_cols:\n column_keys = row_keys\n else:\n column_keys = {j: j for j in range(n_cols)}\n if len(row_keys) != n_rows:\n raise ValueError('Need %d row keys but got list of length %d' % (\n n_rows, len(row_keys)))\n if len(column_keys) != n_cols:\n raise ValueError('Need %d column keys but got list of length %d' %\n (n_cols, len(column_keys)))\n result_dict = {}\n for i, X_i in enumerate(X):\n row_key = row_keys[i]\n for j, X_ij in enumerate(X_i):\n if filter_fn and not filter_fn(X_ij):\n continue\n column_key = column_keys[j]\n key_pair = row_key, column_key\n result_dict[key_pair] = X_ij\n return result_dict\n",
"role": "user"
},
{
"content": "matrix_to_pair_dictionary",
"role": "assistant"
}
] |
def x(key_pair_dict, default_value=0.0):
result = defaultdict(dict)
for (a, b), value in key_pair_dict.items():
result[a][b] = value
return result
| curry_pair_dictionary | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(key_pair_dict, default_value=0.0):\n result = defaultdict(dict)\n for (a, b), value in key_pair_dict.items():\n result[a][b] = value\n return result\n",
"role": "user"
},
{
"content": "curry_pair_dictionary",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.