source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(input_dict):
import copy
input_dict = copy.deepcopy(input_dict)
likelihood_class = input_dict.pop('class')
input_dict['name'] = str(input_dict['name'])
name = input_dict.pop('name')
import GPy
likelihood_class = eval(likelihood_class)
return likelihood_class._build_from_input_dict(likelihood_class, input_dict
)
| from_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(input_dict):\n import copy\n input_dict = copy.deepcopy(input_dict)\n likelihood_class = input_dict.pop('class')\n input_dict['name'] = str(input_d... |
def x(self, y_test, mu_star, var_star, Y_metadata=None):
assert y_test.shape == mu_star.shape
assert y_test.shape == var_star.shape
assert y_test.shape[1] == 1
flat_y_test = y_test.flatten()
flat_mu_star = mu_star.flatten()
flat_var_star = var_star.flatten()
if Y_metadata is not None:
Y_metadata_flat = {}
if Y_metadata is not None:
for key, val in Y_metadata.items():
Y_metadata_flat[key] = np.atleast_1d(val).reshape(-1, 1)
zipped_values = []
for i in range(y_test.shape[0]):
y_m = {}
for key, val in Y_metadata_flat.items():
if np.isscalar(val) or val.shape[0] == 1:
y_m[key] = val
else:
y_m[key] = val[i]
zipped_values.append((flat_y_test[i], flat_mu_star[i],
flat_var_star[i], y_m))
else:
zipped_values = zip(flat_y_test, flat_mu_star, flat_var_star, [None
] * y_test.shape[0])
def x(yi, mi, vi, yi_m):
def x(fi_star):
res = np.exp(self.logpdf(fi_star, yi, yi_m) - 0.5 * np.log(2 *
np.pi * vi) - 0.5 * np.square(fi_star - mi) / vi)
if not np.isfinite(res):
import ipdb
ipdb.set_trace()
return res
return f
p_ystar, _ = zip(*[quad(integral_generator(yi, mi, vi, yi_m), -np.inf,
np.inf) for yi, mi, vi, yi_m in zipped_values])
p_ystar = np.array(p_ystar).reshape(*y_test.shape)
return np.log(p_ystar)
| log_predictive_density | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, y_test, mu_star, var_star, Y_metadata=None):\n assert y_test.shape == mu_star.shape\n assert y_test.shape == var_star.shape\n assert y_test.shap... |
def x(self, y_test, mu_star, var_star, Y_metadata=None, num_samples=1000):
assert y_test.shape == mu_star.shape
assert y_test.shape == var_star.shape
assert y_test.shape[1] == 1
fi_samples = np.random.normal(mu_star, np.sqrt(var_star), size=(mu_star
.shape[0], num_samples))
from scipy.misc import logsumexp
log_p_ystar = -np.log(num_samples) + logsumexp(self.logpdf(fi_samples,
y_test, Y_metadata=Y_metadata), axis=1)
log_p_ystar = np.array(log_p_ystar).reshape(*y_test.shape)
return log_p_ystar
| log_predictive_density_sampling | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, y_test, mu_star, var_star, Y_metadata=None, num_samples=1000):\n assert y_test.shape == mu_star.shape\n assert y_test.shape == var_star.shape\n ... |
def x(self, obs, tau, v, Y_metadata_i=None):
mu = v / tau
sigma2 = 1.0 / tau
f_min = mu - 20 * np.sqrt(sigma2)
f_max = mu + 20 * np.sqrt(sigma2)
def x(f):
return self.pdf(f, obs, Y_metadata=Y_metadata_i) * np.exp(-0.5 *
tau * np.square(mu - f))
z_scaled, accuracy = quad(int_1, f_min, f_max)
def x(f):
return f * self.pdf(f, obs, Y_metadata=Y_metadata_i) * np.exp(-0.5 *
tau * np.square(mu - f))
mean, accuracy = quad(int_2, f_min, f_max)
mean /= z_scaled
def x(f):
return f ** 2 * self.pdf(f, obs, Y_metadata=Y_metadata_i) * np.exp(
-0.5 * tau * np.square(mu - f))
Ef2, accuracy = quad(int_3, f_min, f_max)
Ef2 /= z_scaled
variance = Ef2 - mean ** 2
z = z_scaled / np.sqrt(2 * np.pi / tau)
return z, mean, variance
| moments_match_ep | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, obs, tau, v, Y_metadata_i=None):\n mu = v / tau\n sigma2 = 1.0 / tau\n f_min = mu - 20 * np.sqrt(sigma2)\n f_max = mu + 20 * np.sqrt(sigma2)\... |
def x(self, Y, m, v, gh_points=None, Y_metadata=None):
if gh_points is None:
gh_x, gh_w = self._gh_points()
else:
gh_x, gh_w = gh_points
shape = m.shape
m, v, Y = m.flatten(), v.flatten(), Y.flatten()
X = gh_x[None, :] * np.sqrt(2.0 * v[:, None]) + m[:, None]
logp = self.logpdf(X, Y[:, None], Y_metadata=Y_metadata)
dlogp_dx = self.dlogpdf_df(X, Y[:, None], Y_metadata=Y_metadata)
d2logp_dx2 = self.d2logpdf_df2(X, Y[:, None], Y_metadata=Y_metadata)
F = np.dot(logp, gh_w) / np.sqrt(np.pi)
dF_dm = np.dot(dlogp_dx, gh_w) / np.sqrt(np.pi)
dF_dv = np.dot(d2logp_dx2, gh_w) / np.sqrt(np.pi)
dF_dv /= 2.0
if np.any(np.isnan(dF_dv)) or np.any(np.isinf(dF_dv)):
stop
if np.any(np.isnan(dF_dm)) or np.any(np.isinf(dF_dm)):
stop
if self.size:
dF_dtheta = self.dlogpdf_dtheta(X, Y[:, None], Y_metadata=Y_metadata)
dF_dtheta = np.dot(dF_dtheta, gh_w) / np.sqrt(np.pi)
dF_dtheta = dF_dtheta.reshape(self.size, shape[0], shape[1])
else:
dF_dtheta = None
return F.reshape(*shape), dF_dm.reshape(*shape), dF_dv.reshape(*shape
), dF_dtheta
| variational_expectations | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Y, m, v, gh_points=None, Y_metadata=None):\n if gh_points is None:\n gh_x, gh_w = self._gh_points()\n else:\n gh_x, gh_w = gh_points\... |
def x(self, mu, variance, Y_metadata=None):
fmin = -np.inf
fmax = np.inf
def x(f, m, v):
exponent = -(0.5 / v) * np.square(f - m)
assert v.all() > 0
p = safe_exp(exponent)
if p < 1e-10:
return 0.0
else:
return self.conditional_mean(f) * p
scaled_mean = [quad(int_mean, fmin, fmax, args=(mj, s2j))[0] for mj,
s2j in zip(mu, variance)]
mean = np.array(scaled_mean)[:, None] / np.sqrt(2 * np.pi * variance)
return mean
| predictive_mean | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, mu, variance, Y_metadata=None):\n fmin = -np.inf\n fmax = np.inf\n\n def x(f, m, v):\n exponent = -(0.5 / v) * np.square(f - m)\n ... |
def x(self, mu, variance, predictive_mean=None, Y_metadata=None):
normalizer = np.sqrt(2 * np.pi * variance)
fmin_v = -np.inf
fmin_m = np.inf
fmin = -np.inf
fmax = np.inf
from ..util.misc import safe_exp
def x(f, m, v):
exponent = -(0.5 / v) * np.square(f - m)
p = safe_exp(exponent)
if p < 1e-10:
return 0.0
else:
return self.conditional_variance(f) * p
scaled_exp_variance = [quad(int_var, fmin_v, fmax, args=(mj, s2j))[0] for
mj, s2j in zip(mu, variance)]
exp_var = np.array(scaled_exp_variance)[:, None] / normalizer
if predictive_mean is None:
predictive_mean = self.predictive_mean(mu, variance)
predictive_mean_sq = predictive_mean ** 2
def x(f, m, v, predictive_mean_sq):
exponent = -(0.5 / v) * np.square(f - m)
p = np.exp(exponent)
if p < 1e-10:
return 0.0
else:
return self.conditional_mean(f) ** 2 * p
scaled_exp_exp2 = [quad(int_pred_mean_sq, fmin_m, fmax, args=(mj, s2j,
pm2j))[0] for mj, s2j, pm2j in zip(mu, variance, predictive_mean_sq)]
exp_exp2 = np.array(scaled_exp_exp2)[:, None] / normalizer
var_exp = exp_exp2 - predictive_mean_sq
return exp_var + var_exp
| predictive_variance | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, mu, variance, predictive_mean=None, Y_metadata=None):\n normalizer = np.sqrt(2 * np.pi * variance)\n fmin_v = -np.inf\n fmin_m = np.inf\n fmi... |
def x(self, f, y, Y_metadata=None):
if isinstance(self.gp_link, link_functions.Identity):
return self.pdf_link(f, y, Y_metadata=Y_metadata)
else:
inv_link_f = self.gp_link.transf(f)
return self.pdf_link(inv_link_f, y, Y_metadata=Y_metadata)
| pdf | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, f, y, Y_metadata=None):\n if isinstance(self.gp_link, link_functions.Identity):\n return self.pdf_link(f, y, Y_metadata=Y_metadata)\n else:\... |
def x(self, f, y, Y_metadata=None):
return np.sum(self.logpdf(f, y, Y_metadata=Y_metadata))
| logpdf_sum | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, f, y, Y_metadata=None):\n return np.sum(self.logpdf(f, y, Y_metadata=Y_metadata))\n",
"role": "user"
},
{
"content": "logpdf_sum",
"ro... |
def x(self, f, y, Y_metadata=None):
if isinstance(self.gp_link, link_functions.Identity):
return self.logpdf_link(f, y, Y_metadata=Y_metadata)
else:
inv_link_f = self.gp_link.transf(f)
return self.logpdf_link(inv_link_f, y, Y_metadata=Y_metadata)
| logpdf | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, f, y, Y_metadata=None):\n if isinstance(self.gp_link, link_functions.Identity):\n return self.logpdf_link(f, y, Y_metadata=Y_metadata)\n els... |
def x(self, f, y, Y_metadata=None):
if isinstance(self.gp_link, link_functions.Identity):
return self.dlogpdf_dlink(f, y, Y_metadata=Y_metadata)
else:
inv_link_f = self.gp_link.transf(f)
dlogpdf_dlink = self.dlogpdf_dlink(inv_link_f, y, Y_metadata=Y_metadata
)
dlink_df = self.gp_link.dtransf_df(f)
return chain_1(dlogpdf_dlink, dlink_df)
| dlogpdf_df | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, f, y, Y_metadata=None):\n if isinstance(self.gp_link, link_functions.Identity):\n return self.dlogpdf_dlink(f, y, Y_metadata=Y_metadata)\n e... |
def x(self, f, y, Y_metadata=None):
if isinstance(self.gp_link, link_functions.Identity):
d2logpdf_df2 = self.d2logpdf_dlink2(f, y, Y_metadata=Y_metadata)
else:
inv_link_f = self.gp_link.transf(f)
d2logpdf_dlink2 = self.d2logpdf_dlink2(inv_link_f, y, Y_metadata=
Y_metadata)
dlink_df = self.gp_link.dtransf_df(f)
dlogpdf_dlink = self.dlogpdf_dlink(inv_link_f, y, Y_metadata=Y_metadata
)
d2link_df2 = self.gp_link.d2transf_df2(f)
d2logpdf_df2 = chain_2(d2logpdf_dlink2, dlink_df, dlogpdf_dlink,
d2link_df2)
return d2logpdf_df2
| d2logpdf_df2 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, f, y, Y_metadata=None):\n if isinstance(self.gp_link, link_functions.Identity):\n d2logpdf_df2 = self.d2logpdf_dlink2(f, y, Y_metadata=Y_metada... |
def x(self, f, y, Y_metadata=None):
if isinstance(self.gp_link, link_functions.Identity):
d3logpdf_df3 = self.d3logpdf_dlink3(f, y, Y_metadata=Y_metadata)
else:
inv_link_f = self.gp_link.transf(f)
d3logpdf_dlink3 = self.d3logpdf_dlink3(inv_link_f, y, Y_metadata=
Y_metadata)
dlink_df = self.gp_link.dtransf_df(f)
d2logpdf_dlink2 = self.d2logpdf_dlink2(inv_link_f, y, Y_metadata=
Y_metadata)
d2link_df2 = self.gp_link.d2transf_df2(f)
dlogpdf_dlink = self.dlogpdf_dlink(inv_link_f, y, Y_metadata=Y_metadata
)
d3link_df3 = self.gp_link.d3transf_df3(f)
d3logpdf_df3 = chain_3(d3logpdf_dlink3, dlink_df, d2logpdf_dlink2,
d2link_df2, dlogpdf_dlink, d3link_df3)
return d3logpdf_df3
| d3logpdf_df3 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, f, y, Y_metadata=None):\n if isinstance(self.gp_link, link_functions.Identity):\n d3logpdf_df3 = self.d3logpdf_dlink3(f, y, Y_metadata=Y_metada... |
def x(self, f, y, Y_metadata=None):
if self.size > 0:
if self.not_block_really:
raise NotImplementedError('Need to make a decorator for this!')
if isinstance(self.gp_link, link_functions.Identity):
return self.dlogpdf_link_dtheta(f, y, Y_metadata=Y_metadata)
else:
inv_link_f = self.gp_link.transf(f)
return self.dlogpdf_link_dtheta(inv_link_f, y, Y_metadata=
Y_metadata)
else:
return np.zeros((0, f.shape[0], f.shape[1]))
| dlogpdf_dtheta | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, f, y, Y_metadata=None):\n if self.size > 0:\n if self.not_block_really:\n raise NotImplementedError('Need to make a decorator for th... |
def x(self, f, y, Y_metadata=None):
if self.size > 0:
if self.not_block_really:
raise NotImplementedError('Need to make a decorator for this!')
if isinstance(self.gp_link, link_functions.Identity):
return self.dlogpdf_dlink_dtheta(f, y, Y_metadata=Y_metadata)
else:
inv_link_f = self.gp_link.transf(f)
dlink_df = self.gp_link.dtransf_df(f)
dlogpdf_dlink_dtheta = self.dlogpdf_dlink_dtheta(inv_link_f, y,
Y_metadata=Y_metadata)
dlogpdf_df_dtheta = np.zeros((self.size, f.shape[0], f.shape[1]))
for p in range(self.size):
dlogpdf_df_dtheta[p, :, :] = chain_1(dlogpdf_dlink_dtheta[p,
:, :], dlink_df)
return dlogpdf_df_dtheta
else:
return np.zeros((0, f.shape[0], f.shape[1]))
| dlogpdf_df_dtheta | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, f, y, Y_metadata=None):\n if self.size > 0:\n if self.not_block_really:\n raise NotImplementedError('Need to make a decorator for th... |
def x(self, f, y, Y_metadata=None):
if self.size > 0:
if self.not_block_really:
raise NotImplementedError('Need to make a decorator for this!')
if isinstance(self.gp_link, link_functions.Identity):
return self.d2logpdf_dlink2_dtheta(f, y, Y_metadata=Y_metadata)
else:
inv_link_f = self.gp_link.transf(f)
dlink_df = self.gp_link.dtransf_df(f)
d2link_df2 = self.gp_link.d2transf_df2(f)
d2logpdf_dlink2_dtheta = self.d2logpdf_dlink2_dtheta(inv_link_f,
y, Y_metadata=Y_metadata)
dlogpdf_dlink_dtheta = self.dlogpdf_dlink_dtheta(inv_link_f, y,
Y_metadata=Y_metadata)
d2logpdf_df2_dtheta = np.zeros((self.size, f.shape[0], f.shape[1]))
for p in range(self.size):
d2logpdf_df2_dtheta[p, :, :] = chain_2(d2logpdf_dlink2_dtheta
[p, :, :], dlink_df, dlogpdf_dlink_dtheta[p, :, :],
d2link_df2)
return d2logpdf_df2_dtheta
else:
return np.zeros((0, f.shape[0], f.shape[1]))
| d2logpdf_df2_dtheta | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, f, y, Y_metadata=None):\n if self.size > 0:\n if self.not_block_really:\n raise NotImplementedError('Need to make a decorator for th... |
def x(self, mu, var, full_cov=False, Y_metadata=None):
try:
pred_mean = self.predictive_mean(mu, var, Y_metadata=Y_metadata)
pred_var = self.predictive_variance(mu, var, pred_mean, Y_metadata=
Y_metadata)
except NotImplementedError:
print(
'Finding predictive mean and variance via sampling rather than quadrature'
)
Nf_samp = 300
Ny_samp = 1
s = np.random.randn(mu.shape[0], Nf_samp) * np.sqrt(var) + mu
ss_y = self.samples(s, Y_metadata, samples=Ny_samp)
pred_mean = np.mean(ss_y, axis=1)[:, None]
pred_var = np.var(ss_y, axis=1)[:, None]
return pred_mean, pred_var
| predictive_values | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, mu, var, full_cov=False, Y_metadata=None):\n try:\n pred_mean = self.predictive_mean(mu, var, Y_metadata=Y_metadata)\n pred_var = self.p... |
def x(self, fNew, num_samples=1000, starting_loc=None, stepsize=0.1,
burn_in=1000, Y_metadata=None):
print('Warning, using MCMC for sampling y*, needs to be tuned!')
if starting_loc is None:
starting_loc = fNew
from functools import partial
logpdf = partial(self.logpdf, f=fNew, Y_metadata=Y_metadata)
pdf = lambda y_star: np.exp(logpdf(y=y_star[:, None]))
par_chains = starting_loc.shape[0]
chain_values = np.zeros((par_chains, num_samples))
chain_values[:, 0][:, None] = starting_loc
stepsize = np.ones(par_chains) * stepsize
accepted = np.zeros((par_chains, num_samples + burn_in))
accept_ratio = np.zeros(num_samples + burn_in)
burnin_cache = np.zeros(par_chains)
burnin_cache[:] = starting_loc.flatten()
burning_in = True
for i in range(burn_in + num_samples):
next_ind = i - burn_in
if burning_in:
old_y = burnin_cache
else:
old_y = chain_values[:, next_ind - 1]
old_lik = pdf(old_y)
new_y = np.random.normal(loc=old_y, scale=stepsize)
new_lik = pdf(new_y)
accept_probability = np.minimum(1, new_lik / old_lik)
u = np.random.uniform(0, 1, par_chains)
accepts = u < accept_probability
if burning_in:
burnin_cache[accepts] = new_y[accepts]
burnin_cache[~accepts] = old_y[~accepts]
if i == burn_in:
burning_in = False
chain_values[:, 0] = burnin_cache
else:
chain_values[accepts, next_ind] = new_y[accepts]
chain_values[~accepts, next_ind] = old_y[~accepts]
accepted[~accepts, i] = 0
accepted[accepts, i] = 1
accept_ratio[i] = np.sum(accepted[:, i]) / float(par_chains)
if i % int((burn_in + num_samples) * 0.1) == 0:
print('{}% of samples taken ({})'.format(i / int((burn_in +
num_samples) * 0.1) * 10, i))
print('Last run accept ratio: ', accept_ratio[i])
print('Average accept ratio: ', np.mean(accept_ratio))
return chain_values
| MCMC_pdf_samples | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, fNew, num_samples=1000, starting_loc=None, stepsize=0.1,\n burn_in=1000, Y_metadata=None):\n print('Warning, using MCMC for sampling y*, needs to b... |
def x(variance, Z, variational_posterior, return_psi2_n=False):
mu = variational_posterior.mean
S = variational_posterior.variance
gamma = variational_posterior.binary_prob
psi0 = (gamma * (np.square(mu) + S) * variance).sum(axis=-1)
psi1 = np.inner(variance * gamma * mu, Z)
psi2 = np.inner(np.square(variance) * (gamma * ((1 - gamma) * np.square
(mu) + S)).sum(axis=0) * Z, Z) + tdot(psi1.T)
return psi0, psi1, psi2
| psicomputations | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(variance, Z, variational_posterior, return_psi2_n=False):\n mu = variational_posterior.mean\n S = variational_posterior.variance\n gamma = variational... |
def x(dL_dpsi2, variance, Z, mu, S, gamma):
mu2 = np.square(mu)
gamma2 = np.square(gamma)
variance2 = np.square(variance)
mu2S = mu2 + S
gvm = gamma * mu * variance
common_sum = gvm.dot(Z.T)
Z_expect = (np.dot(dL_dpsi2, Z) * Z).sum(axis=0)
Z_expect_var2 = Z_expect * variance2
dL_dpsi2T = dL_dpsi2 + dL_dpsi2.T
common_expect = common_sum.dot(dL_dpsi2T).dot(Z)
Z2_expect = common_sum.dot(dL_dpsi2T)
Z1_expect = dL_dpsi2T.dot(Z)
dL_dvar = variance * Z_expect * 2.0 * (gamma * mu2S - gamma2 * mu2).sum(
axis=0) + (common_expect * gamma * mu).sum(axis=0)
dL_dgamma = Z_expect_var2 * (mu2S - 2.0 * gamma * mu2
) + common_expect * mu * variance
dL_dmu = Z_expect_var2 * mu * 2.0 * (gamma - gamma2
) + common_expect * gamma * variance
dL_dS = gamma * Z_expect_var2
dL_dZ = (gamma * (mu2S - gamma * mu2)).sum(axis=0
) * variance2 * Z1_expect + Z2_expect.T.dot(gamma * mu) * variance
return dL_dvar, dL_dgamma, dL_dmu, dL_dS, dL_dZ
| psi2computations | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dL_dpsi2, variance, Z, mu, S, gamma):\n mu2 = np.square(mu)\n gamma2 = np.square(gamma)\n variance2 = np.square(variance)\n mu2S = mu2 + S\n gvm... |
def x(self):
if self._mean is None:
self._mean = np.dot(self._K, self.woodbury_vector)
return self._mean
| mean | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._mean is None:\n self._mean = np.dot(self._K, self.woodbury_vector)\n return self._mean\n",
"role": "user"
},
{
"cont... |
def x(self):
if self._covariance is None:
self._covariance = (np.atleast_3d(self._K) - np.tensordot(np.dot(np
.atleast_3d(self.woodbury_inv).T, self._K), self._K, [1, 0]).T
).squeeze()
return self._covariance
| covariance | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._covariance is None:\n self._covariance = (np.atleast_3d(self._K) - np.tensordot(np.dot(np\n .atleast_3d(self.woodbury_in... |
def x(self, kern, X, X1, X2):
if self.woodbury_chol.ndim != 2:
raise RuntimeError(
'This method does not support posterior for missing data models')
Kx1 = kern.K(X, X1)
Kx2 = kern.K(X, X2)
K12 = kern.K(X1, X2)
tmp1 = dtrtrs(self.woodbury_chol, Kx1)[0]
tmp2 = dtrtrs(self.woodbury_chol, Kx2)[0]
var = K12 - tmp1.T.dot(tmp2)
return var
| covariance_between_points | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, kern, X, X1, X2):\n if self.woodbury_chol.ndim != 2:\n raise RuntimeError(\n 'This method does not support posterior for missing dat... |
def x(self):
if self._precision is None:
cov = np.atleast_3d(self.covariance)
self._precision = np.zeros(cov.shape)
for p in range(cov.shape[-1]):
self._precision[:, :, p] = pdinv(cov[:, :, p])[0]
return self._precision
| precision | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._precision is None:\n cov = np.atleast_3d(self.covariance)\n self._precision = np.zeros(cov.shape)\n for p in range(co... |
def x(self):
if self._woodbury_chol is None:
if self._woodbury_inv is not None:
winv = np.atleast_3d(self._woodbury_inv)
self._woodbury_chol = np.zeros(winv.shape)
for p in range(winv.shape[-1]):
self._woodbury_chol[:, :, p] = pdinv(winv[:, :, p])[2]
elif self._covariance is not None:
raise NotImplementedError('TODO: check code here')
B = self._K - self._covariance
tmp, _ = dpotrs(self.K_chol, B)
self._woodbury_inv, _ = dpotrs(self.K_chol, tmp.T)
_, _, self._woodbury_chol, _ = pdinv(self._woodbury_inv)
else:
raise ValueError('insufficient information to compute posterior')
return self._woodbury_chol
| woodbury_chol | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._woodbury_chol is None:\n if self._woodbury_inv is not None:\n winv = np.atleast_3d(self._woodbury_inv)\n self... |
def x(self):
if self._woodbury_inv is None:
if self._woodbury_chol is not None:
self._woodbury_inv, _ = dpotri(self._woodbury_chol, lower=1)
symmetrify(self._woodbury_inv)
elif self._covariance is not None:
B = np.atleast_3d(self._K) - np.atleast_3d(self._covariance)
self._woodbury_inv = np.empty_like(B)
for i in range(B.shape[-1]):
tmp, _ = dpotrs(self.K_chol, B[:, :, i])
self._woodbury_inv[:, :, i], _ = dpotrs(self.K_chol, tmp.T)
return self._woodbury_inv
| woodbury_inv | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._woodbury_inv is None:\n if self._woodbury_chol is not None:\n self._woodbury_inv, _ = dpotri(self._woodbury_chol, lower=... |
def x(self):
if self._woodbury_vector is None:
self._woodbury_vector, _ = dpotrs(self.K_chol, self.mean - self.
_prior_mean)
return self._woodbury_vector
| woodbury_vector | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._woodbury_vector is None:\n self._woodbury_vector, _ = dpotrs(self.K_chol, self.mean - self.\n _prior_mean)\n return s... |
def x(self):
if self._K_chol is None:
self._K_chol = jitchol(self._K)
return self._K_chol
| K_chol | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._K_chol is None:\n self._K_chol = jitchol(self._K)\n return self._K_chol\n",
"role": "user"
},
{
"content": "K_chol",... |
def x(self, X, target):
np.add(target, self.variance, target)
| Kdiag | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, target):\n np.add(target, self.variance, target)\n",
"role": "user"
},
{
"content": "Kdiag",
"role": "assistant"
}
] |
def x(self, dL_dK, X, X2, target):
self._K_computations(X, X2)
self._dK_computations(dL_dK)
if X2 == None:
gmapping = self.mapping.df_dtheta(2 * self._dL_dl[:, None], X)
else:
gmapping = self.mapping.df_dtheta(self._dL_dl[:, None], X)
gmapping += self.mapping.df_dtheta(self._dL_dl_two[:, None], X2)
target += np.hstack([(dL_dK * self._K_dvar).sum(), gmapping])
| param_grad_helper | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dL_dK, X, X2, target):\n self._K_computations(X, X2)\n self._dK_computations(dL_dK)\n if X2 == None:\n gmapping = self.mapping.df_dtheta(... |
def x(self, dL_dK, X, X2, target):
self._K_computations(X, X2)
if X2 is None:
_K_dist = 2 * (X[:, None, :] - X[None, :, :])
else:
_K_dist = X[:, None, :] - X2[None, :, :]
gradients_X = -2.0 * self.variance * np.transpose((self._K_dvar / self.
_w2)[:, :, None] * _K_dist, (1, 0, 2))
target += np.sum(gradients_X * dL_dK.T[:, :, None], 0)
self._dK_computations(dL_dK)
if X2 is None:
target += 2.0 * self.mapping.df_dX(self._dL_dl[:, None], X)
else:
target += self.mapping.df_dX(self._dL_dl[:, None], X)
| gradients_X | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dL_dK, X, X2, target):\n self._K_computations(X, X2)\n if X2 is None:\n _K_dist = 2 * (X[:, None, :] - X[None, :, :])\n else:\n _K... |
def x(self, X, X2=None):
self._lengthscales = self.mapping.f(X)
self._lengthscales2 = np.square(self._lengthscales)
if X2 == None:
self._lengthscales_two = self._lengthscales
self._lengthscales_two2 = self._lengthscales2
Xsquare = np.square(X).sum(1)
self._K_dist2 = -2.0 * tdot(X) + Xsquare[:, None] + Xsquare[None, :]
else:
self._lengthscales_two = self.mapping.f(X2)
self._lengthscales_two2 = np.square(self._lengthscales_two)
self._K_dist2 = -2.0 * np.dot(X, X2.T) + np.square(X).sum(1)[:, None
] + np.square(X2).sum(1)[None, :]
self._w2 = self._lengthscales2 + self._lengthscales_two2.T
prod_length = self._lengthscales * self._lengthscales_two.T
self._K_exponential = np.exp(-self._K_dist2 / self._w2)
self._K_dvar = np.sign(prod_length) * (2 * np.abs(prod_length) / self._w2
) ** (self.input_dim / 2.0) * np.exp(-self._K_dist2 / self._w2)
| K_computations | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, X2=None):\n self._lengthscales = self.mapping.f(X)\n self._lengthscales2 = np.square(self._lengthscales)\n if X2 == None:\n self._leng... |
def x(self, dL_dK):
self._dL_dl = (dL_dK * self.variance * self._K_dvar * (self.input_dim /
2.0 * (self._lengthscales_two.T ** 4 - self._lengthscales ** 4) + 2 *
self._lengthscales2 * self._K_dist2) / (self._w2 * self._w2 * self.
_lengthscales)).sum(1)
if self._lengthscales_two is self._lengthscales:
self._dL_dl_two = None
else:
self._dL_dl_two = (dL_dK * self.variance * self._K_dvar * (self.
input_dim / 2.0 * (self._lengthscales ** 4 - self.
_lengthscales_two.T ** 4) + 2 * self._lengthscales_two2.T *
self._K_dist2) / (self._w2 * self._w2 * self._lengthscales_two.T)
).sum(0)
| dK_computations | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dL_dK):\n self._dL_dl = (dL_dK * self.variance * self._K_dvar * (self.input_dim /\n 2.0 * (self._lengthscales_two.T ** 4 - self._lengthscales *... |
def x(self, summarize=True):
if not summarize:
num_params = [0]
parts = []
def x(x):
if not isinstance(x, CombinationKernel) and isinstance(x, Kern):
num_params[0] += 1
parts.append(x)
self.traverse(sum_params)
i_s = np.zeros((num_params[0], self.input_dim))
from operator import setitem
[setitem(i_s, (i, k._all_dims_active), k.input_sensitivity(
summarize)) for i, k in enumerate(parts)]
return i_s
else:
raise NotImplementedError(
'Choose the kernel you want to get the sensitivity for. You need to override the default behaviour for getting the input sensitivity to be able to get the input sensitivity. For sum kernel it is the sum of all sensitivities, TODO: product kernel? Other kernels?, also TODO: shall we return all the sensitivities here in the combination kernel? So we can combine them however we want? This could lead to just plot all the sensitivities here...'
)
| input_sensitivity | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, summarize=True):\n if not summarize:\n num_params = [0]\n parts = []\n\n def x(x):\n if not isinstance(x, CombinationK... |
def x(D1, D2, D3, N, num_inducing, plot_sim=False):
Q_signal = 4
import GPy
import numpy as np
np.random.seed(3000)
k = GPy.kern.Matern32(Q_signal, 1.0, lengthscale=np.random.uniform(1, 6,
Q_signal), ARD=1)
for i in range(Q_signal):
k += GPy.kern.PeriodicExponential(1, variance=1.0, active_dims=[i],
period=3.0, lower=-2, upper=6)
t = np.c_[[np.linspace(-1, 5, N) for _ in range(Q_signal)]].T
K = k.K(t)
s2, s1, s3, sS = np.random.multivariate_normal(np.zeros(K.shape[0]), K,
size=4)[:, :, None]
Y1, Y2, Y3, S1, S2, S3 = _generate_high_dimensional_output(D1, D2, D3,
s1, s2, s3, sS)
slist = [sS, s1, s2, s3]
slist_names = ['sS', 's1', 's2', 's3']
Ylist = [Y1, Y2, Y3]
if plot_sim:
from matplotlib import pyplot as plt
import matplotlib.cm as cm
import itertools
fig = plt.figure('MRD Simulation Data', figsize=(8, 6))
fig.clf()
ax = fig.add_subplot(2, 1, 1)
labls = slist_names
for S, lab in zip(slist, labls):
ax.plot(S, label=lab)
ax.legend()
for i, Y in enumerate(Ylist):
ax = fig.add_subplot(2, len(Ylist), len(Ylist) + 1 + i)
ax.imshow(Y, aspect='auto', cmap=cm.gray)
ax.set_title('Y{}'.format(i + 1))
plt.draw()
plt.tight_layout()
return slist, [S1, S2, S3], Ylist
| simulate_matern | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(D1, D2, D3, N, num_inducing, plot_sim=False):\n Q_signal = 4\n import GPy\n import numpy as np\n np.random.seed(3000)\n k = GPy.kern.Matern32(Q_... |
def x(D1, D2, D3, N, num_inducing, plot_sim=False):
_np.random.seed(1234)
x = _np.linspace(0, 4 * _np.pi, N)[:, None]
s1 = _np.vectorize(lambda x: _np.sin(x))
s2 = _np.vectorize(lambda x: _np.cos(x))
s3 = _np.vectorize(lambda x: -_np.exp(-_np.cos(2 * x)))
sS = _np.vectorize(lambda x: _np.cos(x))
s1 = s1(x)
s2 = s2(x)
s3 = s3(x)
sS = sS(x)
s1 -= s1.mean()
s1 /= s1.std(0)
s2 -= s2.mean()
s2 /= s2.std(0)
s3 -= s3.mean()
s3 /= s3.std(0)
sS -= sS.mean()
sS /= sS.std(0)
Y1, Y2, Y3, S1, S2, S3 = _generate_high_dimensional_output(D1, D2, D3,
s1, s2, s3, sS)
slist = [sS, s1, s2, s3]
slist_names = ['sS', 's1', 's2', 's3']
Ylist = [Y1, Y2, Y3]
if plot_sim:
from matplotlib import pyplot as plt
import matplotlib.cm as cm
import itertools
fig = plt.figure('MRD Simulation Data', figsize=(8, 6))
fig.clf()
ax = fig.add_subplot(2, 1, 1)
labls = slist_names
for S, lab in zip(slist, labls):
ax.plot(S, label=lab)
ax.legend()
for i, Y in enumerate(Ylist):
ax = fig.add_subplot(2, len(Ylist), len(Ylist) + 1 + i)
ax.imshow(Y, aspect='auto', cmap=cm.gray)
ax.set_title('Y{}'.format(i + 1))
plt.draw()
plt.tight_layout()
return slist, [S1, S2, S3], Ylist
| simulate_sincos | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(D1, D2, D3, N, num_inducing, plot_sim=False):\n _np.random.seed(1234)\n x = _np.linspace(0, 4 * _np.pi, N)[:, None]\n s1 = _np.vectorize(lambda x: _np... |
def x(model=None, optimize=True, verbose=True, plot=True):
from GPy.models import BayesianGPLVM
from matplotlib import pyplot as plt
import numpy as np
import GPy
import pods
data = pods.datasets.osu_run1()
Q = 6
kernel = GPy.kern.RBF(Q, lengthscale=np.repeat(0.5, Q), ARD=True)
m = BayesianGPLVM(data['Y'], Q, init='PCA', num_inducing=20, kernel=kernel)
m.data = data
m.likelihood.variance = 0.001
try:
if optimize:
m.optimize('bfgs', messages=verbose, max_iters=5000.0,
bfgs_factor=10)
except KeyboardInterrupt:
print('Keyboard interrupt, continuing to plot and return')
if plot:
fig, (latent_axes, sense_axes) = plt.subplots(1, 2)
plt.sca(latent_axes)
m.plot_latent(ax=latent_axes)
y = m.Y[:1, :].copy()
data_show = GPy.plotting.matplot_dep.visualize.stick_show(y,
connect=data['connect'])
dim_select = GPy.plotting.matplot_dep.visualize.lvm_dimselect(m.X.
mean[:1, :].copy(), m, data_show, latent_axes=latent_axes,
sense_axes=sense_axes)
fig.canvas.draw()
input('Press enter to finish')
return m
| stick_bgplvm | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model=None, optimize=True, verbose=True, plot=True):\n from GPy.models import BayesianGPLVM\n from matplotlib import pyplot as plt\n import numpy as n... |
def x(inputs, data, clust):
S = data[0].shape[0]
X = np.zeros([0, 1])
Y = np.zeros([0, S])
for p in clust:
X = np.vstack([X, inputs[p]])
Y = np.vstack([Y, data[p].T])
m = GPy.models.GPRegression(X, Y)
m.optimize()
ll = m.log_likelihood()
return ll, 0
| get_log_likelihood | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inputs, data, clust):\n S = data[0].shape[0]\n X = np.zeros([0, 1])\n Y = np.zeros([0, S])\n for p in clust:\n X = np.vstack([X, inputs[p]])... |
def x(inputs, data, clust):
if len(clust) == 1:
return get_log_likelihood(inputs, data, clust)
S = data[0].shape[0]
X = np.zeros([0, 2])
Y = np.zeros([0, S])
for i, p in enumerate(clust):
idx = i * np.ones([inputs[p].shape[0], 1])
X = np.vstack([X, np.hstack([inputs[p], idx])])
Y = np.vstack([Y, data[p].T])
m = GPy.models.GPOffsetRegression(X, Y)
m.offset.set_prior(GPy.priors.Gaussian(0, 20))
m.optimize()
ll = m.log_likelihood()
offset = m.offset.values[0]
return ll, offset
| get_log_likelihood_offset | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(inputs, data, clust):\n if len(clust) == 1:\n return get_log_likelihood(inputs, data, clust)\n S = data[0].shape[0]\n X = np.zeros([0, 2])\n ... |
def x(data, inputs, verbose=False):
N = len(data)
active = []
for p in range(0, N):
active.append([p])
loglikes = np.zeros(len(active))
loglikes[:] = None
pairloglikes = np.zeros([len(active), len(active)])
pairloglikes[:] = None
pairoffset = np.zeros([len(active), len(active)])
it = 0
while True:
if verbose:
it += 1
print('Iteration %d' % it)
for clusti in range(len(active)):
if verbose:
sys.stdout.write('.')
sys.stdout.flush()
if np.isnan(loglikes[clusti]):
loglikes[clusti], unused_offset = get_log_likelihood_offset(
inputs, data, [clusti])
for clustj in range(clusti):
temp = [clusti, clustj]
if np.isnan(pairloglikes[clusti, clustj]):
pairloglikes[clusti, clustj], pairoffset[clusti, clustj
] = get_log_likelihood_offset(inputs, data, temp)
seploglikes = np.repeat(loglikes[:, None].T, len(loglikes), 0
) + np.repeat(loglikes[:, None], len(loglikes), 1)
loglikeimprovement = pairloglikes - seploglikes
top = np.unravel_index(np.nanargmax(pairloglikes - seploglikes),
pairloglikes.shape)
if loglikeimprovement[top[0], top[1]] > 0:
active[top[0]].extend(active[top[1]])
offset = pairoffset[top[0], top[1]]
inputs[top[0]] = np.vstack([inputs[top[0]], inputs[top[1]] -
offset])
data[top[0]] = np.hstack([data[top[0]], data[top[1]]])
del inputs[top[1]]
del data[top[1]]
del active[top[1]]
pairloglikes[:, top[0]] = None
pairloglikes[top[0], :] = None
pairloglikes = np.delete(pairloglikes, top[1], 0)
pairloglikes = np.delete(pairloglikes, top[1], 1)
loglikes[top[0]] = None
loglikes = np.delete(loglikes, top[1])
else:
break
return active
| cluster | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(data, inputs, verbose=False):\n N = len(data)\n active = []\n for p in range(0, N):\n active.append([p])\n loglikes = np.zeros(len(active))\... |
def x(self, X, X2, target):
if X2 is None or X2 is X:
target[np.diag_indices_from(target)] += self._Kdiag(X)
| K | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, X2, target):\n if X2 is None or X2 is X:\n target[np.diag_indices_from(target)] += self._Kdiag(X)\n",
"role": "user"
},
{
"conte... |
def x(self, dL_dK, X, X2, target):
if X2 is None or X2 is X:
dL_dKdiag = dL_dK.flat[::dL_dK.shape[0] + 1]
self.dKdiag_dtheta(dL_dKdiag, X, target)
| param_grad_helper | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dL_dK, X, X2, target):\n if X2 is None or X2 is X:\n dL_dKdiag = dL_dK.flat[::dL_dK.shape[0] + 1]\n self.dKdiag_dtheta(dL_dKdiag, X, tar... |
def x(self, dL_dKdiag, X, target):
target += 2.0 * self.mapping.df_dtheta(dL_dKdiag[:, None] * self.
mapping.f(X), X)
| dKdiag_dtheta | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dL_dKdiag, X, target):\n target += 2.0 * self.mapping.df_dtheta(dL_dKdiag[:, None] * self.\n mapping.f(X), X)\n",
"role": "user"
},
{
... |
def x(self, dL_dK, X, X2, target):
if X2 == None or X2 is X:
dL_dKdiag = dL_dK.flat[::dL_dK.shape[0] + 1]
self.dKdiag_dX(dL_dKdiag, X, target)
| gradients_X | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dL_dK, X, X2, target):\n if X2 == None or X2 is X:\n dL_dKdiag = dL_dK.flat[::dL_dK.shape[0] + 1]\n self.dKdiag_dX(dL_dKdiag, X, target)... |
def x(self, dL_dKdiag, X, target):
target += 2.0 * self.mapping.df_dX(dL_dKdiag[:, None], X) * self.mapping.f(
X)
| dKdiag_dX | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dL_dKdiag, X, target):\n target += 2.0 * self.mapping.df_dX(dL_dKdiag[:, None], X) * self.mapping.f(\n X)\n",
"role": "user"
},
{
"... |
def x(self, link_f, y, Y_metadata=None):
c = np.zeros_like(link_f)
if Y_metadata is not None and 'censored' in Y_metadata.keys():
c = Y_metadata['censored']
link_f = np.clip(link_f, 1e-150, 1e+100)
y_r = np.clip(y ** self.r, 1e-150, 1e+200)
link_f_r = np.clip(link_f ** self.r, 1e-150, 1e+200)
y_link_f_r = np.clip(y_r / link_f_r, 1e-150, 1e+200)
uncensored = (1 - c) * (np.log(self.r) + (self.r - 1) * np.log(y) -
self.r * np.log(link_f) - 2 * np.log1p(y_link_f_r))
censored = c * -np.log1p(y_link_f_r)
return uncensored + censored
| logpdf_link | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, link_f, y, Y_metadata=None):\n c = np.zeros_like(link_f)\n if Y_metadata is not None and 'censored' in Y_metadata.keys():\n c = Y_metadata['... |
def x(self, link_f, y, Y_metadata=None):
c = np.zeros_like(link_f)
if Y_metadata is not None and 'censored' in Y_metadata.keys():
c = Y_metadata['censored']
y_link_f_r = np.clip(y ** self.r, 1e-150, 1e+200) / np.clip(link_f **
self.r, 1e-150, 1e+200)
uncensored = (1 - c) * self.r * (y_link_f_r - 1) / (link_f * (1 +
y_link_f_r))
censored = c * (self.r * y_link_f_r / (link_f * y_link_f_r + link_f))
return uncensored + censored
| dlogpdf_dlink | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, link_f, y, Y_metadata=None):\n c = np.zeros_like(link_f)\n if Y_metadata is not None and 'censored' in Y_metadata.keys():\n c = Y_metadata['... |
def x(self, link_f, y, Y_metadata=None):
c = np.zeros_like(link_f)
if Y_metadata is not None and 'censored' in Y_metadata.keys():
c = Y_metadata['censored']
y_link_f = y / link_f
y_link_f_r = y_link_f ** self.r
censored = c * (-self.r * y_link_f_r * (y_link_f_r + self.r + 1) / (
link_f ** 2 * (y_link_f_r + 1) ** 2))
uncensored = (1 - c) * (-self.r * (2 * self.r * y_link_f_r + y_link_f **
(2 * self.r) - 1) / (link_f ** 2 * (1 + y_link_f_r) ** 2))
hess = censored + uncensored
return hess
| d2logpdf_dlink2 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, link_f, y, Y_metadata=None):\n c = np.zeros_like(link_f)\n if Y_metadata is not None and 'censored' in Y_metadata.keys():\n c = Y_metadata['... |
def x(self, link_f, y, Y_metadata=None):
c = np.zeros_like(link_f)
if Y_metadata is not None and 'censored' in Y_metadata.keys():
c = Y_metadata['censored']
y_link_f = y / link_f
y_link_f_r = y_link_f ** self.r
censored = c * (self.r * y_link_f_r * (self.r ** 2 * -(y_link_f_r - 1) +
3 * self.r * (y_link_f_r + 1) + 2 * (y_link_f_r + 1) ** 2) / (
link_f ** 3 * (y_link_f_r + 1) ** 3))
uncensored = (1 - c) * (2 * self.r * (-self.r ** 2 * (y_link_f_r - 1) *
y_link_f_r + 3 * self.r * (y_link_f_r + 1) * y_link_f_r + (
y_link_f_r - 1) * (y_link_f_r + 1) ** 2) / (link_f ** 3 * (
y_link_f_r + 1) ** 3))
d3lik_dlink3 = censored + uncensored
return d3lik_dlink3
| d3logpdf_dlink3 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, link_f, y, Y_metadata=None):\n c = np.zeros_like(link_f)\n if Y_metadata is not None and 'censored' in Y_metadata.keys():\n c = Y_metadata['... |
def x(self, inv_link_f, y, Y_metadata=None):
c = np.zeros_like(y)
if Y_metadata is not None and 'censored' in Y_metadata.keys():
c = Y_metadata['censored']
link_f = inv_link_f
y_link_f = y / link_f
log_y_link_f = np.log(y) - np.log(link_f)
y_link_f_r = y_link_f ** self.r
censored = c * (-y_link_f_r * log_y_link_f / (1 + y_link_f_r))
uncensored = (1 - c) * (1.0 / self.r + np.log(y) - np.log(link_f) - 2 *
y_link_f_r * log_y_link_f / (1 + y_link_f_r))
dlogpdf_dr = censored + uncensored
return dlogpdf_dr
| dlogpdf_link_dr | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, inv_link_f, y, Y_metadata=None):\n c = np.zeros_like(y)\n if Y_metadata is not None and 'censored' in Y_metadata.keys():\n c = Y_metadata['c... |
def x(self, inv_link_f, y, Y_metadata=None):
c = np.zeros_like(y)
if Y_metadata is not None and 'censored' in Y_metadata.keys():
c = Y_metadata['censored']
link_f = inv_link_f
y_link_f = y / link_f
y_link_f_r = y_link_f ** self.r
log_y_link_f = np.log(y) - np.log(link_f)
y_link_f_2r = y_link_f ** (2 * self.r)
denom2 = link_f ** 2 * (1 + y_link_f_r) ** 2
denom3 = link_f ** 2 * (1 + y_link_f_r) ** 3
censored = c * (-((y_link_f_r + self.r + 1) * y_link_f_r) / denom2 -
self.r * (y_link_f_r + self.r + 1) * y_link_f_r * log_y_link_f /
denom2 - self.r * y_link_f_r * (y_link_f_r * log_y_link_f + 1) /
denom2 + 2 * self.r * (y_link_f_r + self.r + 1) * y_link_f_2r *
log_y_link_f / denom3)
uncensored = (1 - c) * (-(2 * self.r * y_link_f_r + y_link_f_2r - 1) /
denom2 - self.r * (2 * y_link_f_r + 2 * self.r * y_link_f_r *
log_y_link_f + 2 * y_link_f_2r * log_y_link_f) / denom2 + 2 * self.
r * (2 * self.r * y_link_f_r + y_link_f_2r - 1) * y_link_f_r *
log_y_link_f / denom3)
d2logpdf_dlink2_dr = censored + uncensored
return d2logpdf_dlink2_dr
| d2logpdf_dlink2_dr | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, inv_link_f, y, Y_metadata=None):\n c = np.zeros_like(y)\n if Y_metadata is not None and 'censored' in Y_metadata.keys():\n c = Y_metadata['c... |
def x(self, max_iters=10000, verbose=True):
self.Z.fix(warning=False)
self.kern.fix(warning=False)
self.kern_row.fix(warning=False)
self.Zr.fix(warning=False)
self.Xr.fix(warning=False)
self.optimize(max_iters=int(0.1 * max_iters), messages=verbose)
self.unfix()
self.optimize(max_iters=max_iters, messages=verbose)
| optimize_auto | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, max_iters=10000, verbose=True):\n self.Z.fix(warning=False)\n self.kern.fix(warning=False)\n self.kern_row.fix(warning=False)\n self.Zr.fix(w... |
def x(x1, x2, return_sign=False):
x1 = np.require(x1).real
x2 = np.require(x2).real
if x1.size == 1:
x1 = np.reshape(x1, (1, 1))
if x2.size == 1:
x2 = np.reshape(x2, (1, 1))
if x1.shape == x2.shape:
v = np.zeros_like(x1)
elif x1.size == 1:
v = np.zeros(x2.shape)
elif x2.size == 1:
v = np.zeros(x1.shape)
else:
raise ValueError(
'This function does not broadcast unless provided with a scalar.')
if x1.size == 1:
x1 = np.tile(x1, x2.shape)
if x2.size == 1:
x2 = np.tile(x2, x1.shape)
sign = np.sign(x1 - x2)
if x1.size == 1:
if sign == -1:
swap = x1
x1 = x2
x2 = swap
else:
I = sign == -1
swap = x1[I]
x1[I] = x2[I]
x2[I] = swap
with np.errstate(divide='ignore'):
I0 = np.logical_or(np.logical_and(x1 > 0, x2 < 0), np.logical_and(
x2 > 0, x1 < 0))
I1 = x1 == x2
I2 = np.logical_and(x1 > 0, np.logical_and(np.logical_not(I0), np.
logical_not(I1)))
I3 = np.logical_and(np.logical_and(np.logical_not(I0), np.
logical_not(I1)), np.logical_not(I2))
_x2 = x2.flatten()
_x1 = x1.flatten()
for group, flags in zip((0, 1, 2, 3), (I0, I1, I2, I3)):
if np.any(flags):
if not x1.size == 1:
_x1 = x1[flags]
if not x2.size == 1:
_x2 = x2[flags]
if group == 0:
v[flags] = np.log(erf(_x1) - erf(_x2))
elif group == 1:
v[flags] = -np.inf
elif group == 2:
v[flags] = np.log(erfcx(_x2) - erfcx(_x1) * np.exp(_x2 **
2 - _x1 ** 2)) - _x2 ** 2
elif group == 3:
v[flags] = np.log(erfcx(-_x1) - erfcx(-_x2) * np.exp(
_x1 ** 2 - _x2 ** 2)) - _x1 ** 2
if return_sign:
return v, sign
elif v.size == 1:
if sign == -1:
v = v.view('complex64')
v += np.pi * 1.0j
else:
v = v.view('complex64')
v[I] += np.pi * 1.0j
return v
| ln_diff_erfs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(x1, x2, return_sign=False):\n x1 = np.require(x1).real\n x2 = np.require(x2).real\n if x1.size == 1:\n x1 = np.reshape(x1, (1, 1))\n if x2.s... |
def x(self, kern, X, likelihood, Y, mean_function=None, Y_metadata=None, K=
None, variance=None, Z_tilde=None):
if mean_function is None:
m = 0
else:
m = mean_function.f(X)
if variance is None:
variance = likelihood.gaussian_variance(Y_metadata)
YYT_factor = Y - m
if K is None:
K = kern.K(X)
Ky = K.copy()
diag.add(Ky, variance + 1e-08)
Wi, LW, LWi, W_logdet = pdinv(Ky)
alpha, _ = dpotrs(LW, YYT_factor, lower=1)
log_marginal = 0.5 * (-Y.size * log_2_pi - Y.shape[1] * W_logdet - np.
sum(alpha * YYT_factor))
if Z_tilde is not None:
log_marginal += Z_tilde
dL_dK = 0.5 * (tdot(alpha) - Y.shape[1] * Wi)
dL_dthetaL = likelihood.exact_inference_gradients(np.diag(dL_dK),
Y_metadata)
return Posterior(woodbury_chol=LW, woodbury_vector=alpha, K=K
), log_marginal, {'dL_dK': dL_dK, 'dL_dthetaL': dL_dthetaL, 'dL_dm':
alpha}
| inference | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, kern, X, likelihood, Y, mean_function=None, Y_metadata=None, K=\n None, variance=None, Z_tilde=None):\n if mean_function is None:\n m = 0\n ... |
def x(self, kern, X, Y, likelihood, posterior, Y_metadata=None, K=None):
g = posterior.woodbury_vector
c = posterior.woodbury_inv
c_diag = np.diag(c)[:, None]
neg_log_marginal_LOO = 0.5 * np.log(2 * np.pi) - 0.5 * np.log(c_diag
) + 0.5 * g ** 2 / c_diag
return -neg_log_marginal_LOO
| LOO | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, kern, X, Y, likelihood, posterior, Y_metadata=None, K=None):\n g = posterior.woodbury_vector\n c = posterior.woodbury_inv\n c_diag = np.diag(c)[... |
def x(self, kern, X, Y, likelihood, posterior, Y_metadata=None, K=None,
f_hat=None, W=None, Ki_W_i=None):
Ki_f_init = np.zeros_like(Y)
if K is None:
K = kern.K(X)
if f_hat is None:
f_hat, _ = self.rasm_mode(K, Y, likelihood, Ki_f_init, Y_metadata=
Y_metadata)
if W is None:
W = -likelihood.d2logpdf_df2(f_hat, Y, Y_metadata=Y_metadata)
if Ki_W_i is None:
_, _, _, Ki_W_i = self._compute_B_statistics(K, W, likelihood.
log_concave)
logpdf_dfhat = likelihood.dlogpdf_df(f_hat, Y, Y_metadata=Y_metadata)
if W.shape[1] == 1:
W = np.diagflat(W)
var_site = 1.0 / np.diag(W)[:, None]
mu_site = f_hat + var_site * logpdf_dfhat
prec_site = 1.0 / var_site
marginal_cov = Ki_W_i
marginal_mu = marginal_cov.dot(np.diagflat(prec_site)).dot(mu_site)
marginal_var = np.diag(marginal_cov)[:, None]
posterior_cav_var = 1.0 / (1.0 / marginal_var - 1.0 / var_site)
posterior_cav_mean = posterior_cav_var * (1.0 / marginal_var *
marginal_mu - 1.0 / var_site * Y)
flat_y = Y.flatten()
flat_mu = posterior_cav_mean.flatten()
flat_var = posterior_cav_var.flatten()
if Y_metadata is not None:
Y_metadata_flat = {}
if Y_metadata is not None:
for key, val in Y_metadata.items():
Y_metadata_flat[key] = np.atleast_1d(val).reshape(-1, 1)
zipped_values = []
for i in range(Y.shape[0]):
y_m = {}
for key, val in Y_metadata_flat.items():
if np.isscalar(val) or val.shape[0] == 1:
y_m[key] = val
else:
y_m[key] = val[i]
zipped_values.append((flat_y[i], flat_mu[i], flat_var[i], y_m))
else:
zipped_values = zip(flat_y, flat_mu, flat_var, [None] * Y.shape[0])
def x(yi, mi, vi, yi_m):
def x(fi_star):
p_fi = np.exp(likelihood.logpdf(fi_star, yi, yi_m) - 0.5 * np.
log(2 * np.pi * vi) - 0.5 * np.square(mi - fi_star) / vi)
return p_fi
return f
p_ystar, _ = zip(*[quad(integral_generator(y, m, v, yi_m), -np.inf, np.
inf) for y, m, v, yi_m in zipped_values])
p_ystar = np.array(p_ystar).reshape(-1, 1)
return np.log(p_ystar)
| LOO | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, kern, X, Y, likelihood, posterior, Y_metadata=None, K=None,\n f_hat=None, W=None, Ki_W_i=None):\n Ki_f_init = np.zeros_like(Y)\n if K is None:\n... |
def x(self, kern, X, likelihood, Y, mean_function=None, Y_metadata=None):
assert mean_function is None, 'inference with a mean function not implemented'
K = kern.K(X)
if self.bad_fhat or self.first_run:
Ki_f_init = np.zeros_like(Y)
self.first_run = False
else:
Ki_f_init = self._previous_Ki_fhat
Ki_f_init = np.zeros_like(Y)
f_hat, Ki_fhat = self.rasm_mode(K, Y, likelihood, Ki_f_init, Y_metadata
=Y_metadata)
log_marginal, woodbury_inv, dL_dK, dL_dthetaL = self.mode_computations(
f_hat, Ki_fhat, K, Y, likelihood, kern, Y_metadata)
self._previous_Ki_fhat = Ki_fhat.copy()
return Posterior(woodbury_vector=Ki_fhat, woodbury_inv=woodbury_inv, K=K
), log_marginal, {'dL_dK': dL_dK, 'dL_dthetaL': dL_dthetaL}
| inference | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, kern, X, likelihood, Y, mean_function=None, Y_metadata=None):\n assert mean_function is None, 'inference with a mean function not implemented'\n K ... |
def x(self, K, Y, likelihood, Ki_f_init, Y_metadata=None, *args, **kwargs):
Ki_f = Ki_f_init.copy()
f = np.dot(K, Ki_f)
def x(Ki_f, f):
ll = -0.5 * np.sum(np.dot(Ki_f.T, f)) + np.sum(likelihood.logpdf(f,
Y, Y_metadata=Y_metadata))
if np.isnan(ll):
import ipdb
ipdb.set_trace()
return -np.inf
else:
return ll
difference = np.inf
iteration = 0
while (difference > self._mode_finding_tolerance and iteration < self.
_mode_finding_max_iter):
W = -likelihood.d2logpdf_df2(f, Y, Y_metadata=Y_metadata)
if np.any(np.isnan(W)):
raise ValueError('One or more element(s) of W is NaN')
grad = likelihood.dlogpdf_df(f, Y, Y_metadata=Y_metadata)
if np.any(np.isnan(grad)):
raise ValueError('One or more element(s) of grad is NaN')
W_f = W * f
b = W_f + grad
W12BiW12, _, _, _ = self._compute_B_statistics(K, W, likelihood.
log_concave, *args, **kwargs)
W12BiW12Kb = np.dot(W12BiW12, np.dot(K, b))
full_step_Ki_f = b - W12BiW12Kb
dKi_f = full_step_Ki_f - Ki_f
def x(step_size):
Ki_f_trial = Ki_f + step_size * dKi_f
f_trial = np.dot(K, Ki_f_trial)
return -obj(Ki_f_trial, f_trial)
step = optimize.brent(inner_obj, tol=0.0001, maxiter=12)
Ki_f_new = Ki_f + step * dKi_f
f_new = np.dot(K, Ki_f_new)
old_obj = obj(Ki_f, f)
new_obj = obj(Ki_f_new, f_new)
if new_obj < old_obj:
raise ValueError("Shouldn't happen, brent optimization failing")
difference = np.abs(new_obj - old_obj)
Ki_f = Ki_f_new
f = f_new
iteration += 1
if difference > self._mode_finding_tolerance:
if not self.bad_fhat:
warnings.warn(
'Not perfect mode found (f_hat). difference: {}, iteration: {} out of max {}'
.format(difference, iteration, self._mode_finding_max_iter))
self.bad_fhat = True
elif self.bad_fhat:
self.bad_fhat = False
warnings.warn(
'f_hat now fine again. difference: {}, iteration: {} out of max {}'
.format(difference, iteration, self._mode_finding_max_iter))
return f, Ki_f
| rasm_mode | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, K, Y, likelihood, Ki_f_init, Y_metadata=None, *args, **kwargs):\n Ki_f = Ki_f_init.copy()\n f = np.dot(K, Ki_f)\n\n def x(Ki_f, f):\n ll ... |
def x(self, f_hat, Ki_f, K, Y, likelihood, kern, Y_metadata):
W = -likelihood.d2logpdf_df2(f_hat, Y, Y_metadata=Y_metadata)
if np.any(np.isnan(W)):
raise ValueError('One or more element(s) of W is NaN')
K_Wi_i, logdet_I_KW, I_KW_i, Ki_W_i = self._compute_B_statistics(K, W,
likelihood.log_concave)
log_marginal = -0.5 * np.sum(np.dot(Ki_f.T, f_hat)) + np.sum(likelihood
.logpdf(f_hat, Y, Y_metadata=Y_metadata)) - 0.5 * logdet_I_KW
dW_df = -likelihood.d3logpdf_df3(f_hat, Y, Y_metadata=Y_metadata)
if np.any(np.isnan(dW_df)):
raise ValueError('One or more element(s) of dW_df is NaN')
dL_dfhat = -0.5 * (np.diag(Ki_W_i)[:, None] * dW_df)
I_KW_i = np.eye(Y.shape[0]) - np.dot(K, K_Wi_i)
if kern.size > 0 and not kern.is_fixed:
explicit_part = 0.5 * (np.dot(Ki_f, Ki_f.T) - K_Wi_i)
implicit_part = np.dot(Ki_f, dL_dfhat.T).dot(I_KW_i)
dL_dK = explicit_part + implicit_part
else:
dL_dK = np.zeros(likelihood.size)
if likelihood.size > 0 and not likelihood.is_fixed:
dlik_dthetaL, dlik_grad_dthetaL, dlik_hess_dthetaL = (likelihood.
_laplace_gradients(f_hat, Y, Y_metadata=Y_metadata))
num_params = likelihood.size
dL_dthetaL = np.zeros(num_params)
for thetaL_i in range(num_params):
dL_dthetaL_exp = np.sum(dlik_dthetaL[thetaL_i, :, :]
) + 0.5 * np.sum(np.diag(Ki_W_i) * np.squeeze(
dlik_hess_dthetaL[thetaL_i, :, :]))
dfhat_dthetaL = mdot(I_KW_i, K, dlik_grad_dthetaL[thetaL_i, :, :])
dL_dthetaL_imp = np.dot(dL_dfhat.T, dfhat_dthetaL)
dL_dthetaL[thetaL_i] = np.sum(dL_dthetaL_exp + dL_dthetaL_imp)
else:
dL_dthetaL = np.zeros(likelihood.size)
self.Ki_W_i = Ki_W_i
self.K = K
self.W = W
self.f_hat = f_hat
return log_marginal, K_Wi_i, dL_dK, dL_dthetaL
| mode_computations | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, f_hat, Ki_f, K, Y, likelihood, kern, Y_metadata):\n W = -likelihood.d2logpdf_df2(f_hat, Y, Y_metadata=Y_metadata)\n if np.any(np.isnan(W)):\n ... |
def x(self, K, W, log_concave, *args, **kwargs):
if not log_concave:
W = np.clip(W, 1e-06, 1e+30)
if np.any(np.isnan(W)):
raise ValueError('One or more element(s) of W is NaN')
W_12 = np.sqrt(W)
B = np.eye(K.shape[0]) + W_12 * K * W_12.T
L = jitchol(B)
LiW12, _ = dtrtrs(L, np.diagflat(W_12), lower=1, trans=0)
K_Wi_i = np.dot(LiW12.T, LiW12)
C = np.dot(LiW12, K)
Ki_W_i = K - C.T.dot(C)
I_KW_i = np.eye(K.shape[0]) - np.dot(K, K_Wi_i)
logdet_I_KW = 2 * np.sum(np.log(np.diag(L)))
return K_Wi_i, logdet_I_KW, I_KW_i, Ki_W_i
| compute_B_statistics | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, K, W, log_concave, *args, **kwargs):\n if not log_concave:\n W = np.clip(W, 1e-06, 1e+30)\n if np.any(np.isnan(W)):\n raise ValueErro... |
def x(self, K, W, log_concave, *args, **kwargs):
B = np.eye(K.shape[0]) + np.dot(K, W)
Bi = np.linalg.inv(B)
K_Wi_i = np.dot(W, Bi)
Ki_W_i = np.dot(Bi, K)
sign, logdetB = np.linalg.slogdet(B)
return K_Wi_i, sign * logdetB, Bi, Ki_W_i
| compute_B_statistics | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, K, W, log_concave, *args, **kwargs):\n B = np.eye(K.shape[0]) + np.dot(K, W)\n Bi = np.linalg.inv(B)\n K_Wi_i = np.dot(W, Bi)\n Ki_W_i = np.d... |
def x(dL, L):
dL_dK = np.tril(dL).copy()
N = L.shape[0]
for k in range(N - 1, -1, -1):
for j in range(k + 1, N):
for i in range(j, N):
dL_dK[i, k] -= dL_dK[i, j] * L[j, k]
dL_dK[j, k] -= dL_dK[i, j] * L[i, k]
for j in range(k + 1, N):
dL_dK[j, k] /= L[k, k]
dL_dK[k, k] -= L[j, k] * dL_dK[j, k]
dL_dK[k, k] /= 2 * L[k, k]
return dL_dK
| backprop_gradient_pure | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dL, L):\n dL_dK = np.tril(dL).copy()\n N = L.shape[0]\n for k in range(N - 1, -1, -1):\n for j in range(k + 1, N):\n for i in range(... |
def x(rank, size):
n = np.arange(size)
triangulars = (n ** 2 + n) / 2
keep = []
for i in range(rank):
keep.append(triangulars[i:] + i)
keep.append(triangulars[1:] - 1)
keep.append((size ** 2 + size) / 2 - 1)
keep = np.hstack(keep)
return np.setdiff1d(np.arange((size ** 2 + size) / 2), keep)
| indexes_to_fix_for_low_rank | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rank, size):\n n = np.arange(size)\n triangulars = (n ** 2 + n) / 2\n keep = []\n for i in range(rank):\n keep.append(triangulars[i:] + i)\n... |
def x(self, inv_link_f, y, Y_metadata):
return np.exp(self.logpdf_link(inv_link_f, y, Y_metadata))
| pdf_link | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, inv_link_f, y, Y_metadata):\n return np.exp(self.logpdf_link(inv_link_f, y, Y_metadata))\n",
"role": "user"
},
{
"content": "pdf_link",
... |
def x(self, inv_link_f, y, Y_metadata=None):
N = Y_metadata['trials']
np.testing.assert_array_equal(N.shape, y.shape)
nchoosey = special.gammaln(N + 1) - special.gammaln(y + 1
) - special.gammaln(N - y + 1)
Ny = N - y
t1 = np.zeros(y.shape)
t2 = np.zeros(y.shape)
t1[y > 0] = y[y > 0] * np.log(inv_link_f[y > 0])
t2[Ny > 0] = Ny[Ny > 0] * np.log(1.0 - inv_link_f[Ny > 0])
return nchoosey + t1 + t2
| logpdf_link | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, inv_link_f, y, Y_metadata=None):\n N = Y_metadata['trials']\n np.testing.assert_array_equal(N.shape, y.shape)\n nchoosey = special.gammaln(N + 1... |
def x(self, inv_link_f, y, Y_metadata=None):
N = Y_metadata['trials']
np.testing.assert_array_equal(N.shape, y.shape)
Ny = N - y
t1 = np.zeros(y.shape)
t2 = np.zeros(y.shape)
t1[y > 0] = y[y > 0] / inv_link_f[y > 0]
t2[Ny > 0] = Ny[Ny > 0] / (1.0 - inv_link_f[Ny > 0])
return t1 - t2
| dlogpdf_dlink | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, inv_link_f, y, Y_metadata=None):\n N = Y_metadata['trials']\n np.testing.assert_array_equal(N.shape, y.shape)\n Ny = N - y\n t1 = np.zeros(y.... |
def x(self, gp, Y_metadata=None, **kw):
orig_shape = gp.shape
gp = gp.flatten()
N = Y_metadata['trials']
Ysim = np.random.binomial(N, self.gp_link.transf(gp))
return Ysim.reshape(orig_shape)
| samples | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, gp, Y_metadata=None, **kw):\n orig_shape = gp.shape\n gp = gp.flatten()\n N = Y_metadata['trials']\n Ysim = np.random.binomial(N, self.gp_lin... |
def x(filename, *args, **kwargs):
try:
netpbm = NetpbmFile(filename)
image = netpbm.asarray()
finally:
netpbm.close()
return image
| imread | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename, *args, **kwargs):\n try:\n netpbm = NetpbmFile(filename)\n image = netpbm.asarray()\n finally:\n netpbm.close()\n retur... |
def x(filename, data, maxval=None, pam=False):
try:
netpbm = NetpbmFile(data, maxval=maxval)
netpbm.write(filename, pam=pam)
finally:
netpbm.close()
| imsave | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename, data, maxval=None, pam=False):\n try:\n netpbm = NetpbmFile(data, maxval=maxval)\n netpbm.write(filename, pam=pam)\n finally:\n ... |
def x(self, copy=True, cache=False, **kwargs):
data = self._data
if data is None:
data = self._read_data(self._fh, **kwargs)
if cache:
self._data = data
else:
return data
return deepcopy(data) if copy else data
| asarray | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, copy=True, cache=False, **kwargs):\n data = self._data\n if data is None:\n data = self._read_data(self._fh, **kwargs)\n if cache:\n ... |
def x(self, arg, **kwargs):
if hasattr(arg, 'seek'):
self._tofile(arg, **kwargs)
else:
with open(arg, 'wb') as fid:
self._tofile(fid, **kwargs)
| write | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, arg, **kwargs):\n if hasattr(arg, 'seek'):\n self._tofile(arg, **kwargs)\n else:\n with open(arg, 'wb') as fid:\n self._to... |
def x(self):
if self._filename and self._fh:
self._fh.close()
self._fh = None
| close | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._filename and self._fh:\n self._fh.close()\n self._fh = None\n",
"role": "user"
},
{
"content": "close",
"rol... |
def x(self, fh):
fh.seek(0)
data = fh.read(4096)
if len(data) < 7 or not b'0' < data[1:2] < b'8':
raise ValueError('Not a Netpbm file:\n%s' % data[:32])
try:
self._read_pam_header(data)
except Exception:
try:
self._read_pnm_header(data)
except Exception:
raise ValueError('Not a Netpbm file:\n%s' % data[:32])
| fromfile | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, fh):\n fh.seek(0)\n data = fh.read(4096)\n if len(data) < 7 or not b'0' < data[1:2] < b'8':\n raise ValueError('Not a Netpbm file:\\n%s' ... |
def x(self, data):
regroups = re.search(
b'(^P7[\n\r]+(?:(?:[\n\r]+)|(?:#.*)|(HEIGHT\\s+\\d+)|(WIDTH\\s+\\d+)|(DEPTH\\s+\\d+)|(MAXVAL\\s+\\d+)|(?:TUPLTYPE\\s+\\w+))*ENDHDR\n)'
, data).groups()
self.header = regroups[0]
self.magicnum = b'P7'
for group in regroups[1:]:
key, value = group.split()
setattr(self, unicode(key).lower(), int(value))
matches = re.findall(b'(TUPLTYPE\\s+\\w+)', self.header)
self.tupltypes = [s.split(None, 1)[1] for s in matches]
| read_pam_header | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n regroups = re.search(\n b'(^P7[\\n\\r]+(?:(?:[\\n\\r]+)|(?:#.*)|(HEIGHT\\\\s+\\\\d+)|(WIDTH\\\\s+\\\\d+)|(DEPTH\\\\s+\\\\d+)|(MAXVAL\\... |
def x(self, data):
bpm = data[1:2] in b'14'
regroups = re.search(b''.join((
b'(^(P[123456]|P7 332)\\s+(?:#.*[\r\n])*',
b'\\s*(\\d+)\\s+(?:#.*[\r\n])*', b'\\s*(\\d+)\\s+(?:#.*[\r\n])*' *
(not bpm), b'\\s*(\\d+)\\s(?:\\s*#.*[\r\n]\\s)*)')), data).groups() + (
1,) * bpm
self.header = regroups[0]
self.magicnum = regroups[1]
self.width = int(regroups[2])
self.height = int(regroups[3])
self.maxval = int(regroups[4])
self.depth = 3 if self.magicnum in b'P3P6P7 332' else 1
self.tupltypes = [self._types[self.magicnum]]
| read_pnm_header | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n bpm = data[1:2] in b'14'\n regroups = re.search(b''.join((\n b'(^(P[123456]|P7 332)\\\\s+(?:#.*[\\r\\n])*',\n b'\\\\s*(\\\\d... |
def x(self, fh, byteorder='>'):
fh.seek(len(self.header))
data = fh.read()
dtype = 'u1' if self.maxval < 256 else byteorder + 'u2'
depth = 1 if self.magicnum == b'P7 332' else self.depth
shape = [-1, self.height, self.width, depth]
size = numpy.prod(shape[1:])
if self.magicnum in b'P1P2P3':
data = numpy.array(data.split(None, size)[:size], dtype)
data = data.reshape(shape)
elif self.maxval == 1:
shape[2] = int(math.ceil(self.width / 8))
data = numpy.frombuffer(data, dtype).reshape(shape)
data = numpy.unpackbits(data, axis=-2)[:, :, :self.width, :]
else:
data = numpy.frombuffer(data, dtype)
data = data[:size * (data.size // size)].reshape(shape)
if data.shape[0] < 2:
data = data.reshape(data.shape[1:])
if data.shape[-1] < 2:
data = data.reshape(data.shape[:-1])
if self.magicnum == b'P7 332':
rgb332 = numpy.array(list(numpy.ndindex(8, 8, 4)), numpy.uint8)
rgb332 *= [36, 36, 85]
data = numpy.take(rgb332, data, axis=0)
return data
| read_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, fh, byteorder='>'):\n fh.seek(len(self.header))\n data = fh.read()\n dtype = 'u1' if self.maxval < 256 else byteorder + 'u2'\n depth = 1 if s... |
def x(self, data, maxval=None):
data = numpy.array(data, ndmin=2, copy=True)
if data.dtype.kind not in 'uib':
raise ValueError('not an integer type: %s' % data.dtype)
if data.dtype.kind == 'i' and numpy.min(data) < 0:
raise ValueError('data out of range: %i' % numpy.min(data))
if maxval is None:
maxval = numpy.max(data)
maxval = 255 if maxval < 256 else 65535
if maxval < 0 or maxval > 65535:
raise ValueError('data out of range: %i' % maxval)
data = data.astype('u1' if maxval < 256 else '>u2')
self._data = data
if data.ndim > 2 and data.shape[-1] in (3, 4):
self.depth = data.shape[-1]
self.width = data.shape[-2]
self.height = data.shape[-3]
self.magicnum = b'P7' if self.depth == 4 else b'P6'
else:
self.depth = 1
self.width = data.shape[-1]
self.height = data.shape[-2]
self.magicnum = b'P5' if maxval > 1 else b'P4'
self.maxval = maxval
self.tupltypes = [self._types[self.magicnum]]
self.header = self._header()
| fromdata | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data, maxval=None):\n data = numpy.array(data, ndmin=2, copy=True)\n if data.dtype.kind not in 'uib':\n raise ValueError('not an integer typ... |
def x(self, fh, pam=False):
fh.seek(0)
fh.write(self._header(pam))
data = self.asarray(copy=False)
if self.maxval == 1:
data = numpy.packbits(data, axis=-1)
data.tofile(fh)
| tofile | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, fh, pam=False):\n fh.seek(0)\n fh.write(self._header(pam))\n data = self.asarray(copy=False)\n if self.maxval == 1:\n data = numpy.pac... |
def x(self, pam=False):
if pam or self.magicnum == b'P7':
header = '\n'.join(('P7', 'HEIGHT %i' % self.height, 'WIDTH %i' %
self.width, 'DEPTH %i' % self.depth, 'MAXVAL %i' % self.maxval,
'\n'.join('TUPLTYPE %s' % unicode(i) for i in self.tupltypes),
'ENDHDR\n'))
elif self.maxval == 1:
header = 'P4 %i %i\n' % (self.width, self.height)
elif self.depth == 1:
header = 'P5 %i %i %i\n' % (self.width, self.height, self.maxval)
else:
header = 'P6 %i %i %i\n' % (self.width, self.height, self.maxval)
if sys.version_info[0] > 2:
header = bytes(header, 'ascii')
return header
| header | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, pam=False):\n if pam or self.magicnum == b'P7':\n header = '\\n'.join(('P7', 'HEIGHT %i' % self.height, 'WIDTH %i' %\n self.width, '... |
def x(f):
@wraps(f)
def x(*args, **kwds):
status = np.seterr(all='ignore')
result = f(*args, **kwds)
np.seterr(**status)
return result
return wrapper
| silence_errors | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(f):\n\n @wraps(f)\n def x(*args, **kwds):\n status = np.seterr(all='ignore')\n result = f(*args, **kwds)\n np.seterr(**status)\n ... |
def x(self):
if self.approx_order is not None:
N = int(self.approx_order)
else:
N = 7
p_period = float(self.period)
p_lengthscale = 2 * float(self.lengthscale)
p_variance = float(self.variance)
w0 = 2 * np.pi / p_period
[q2, dq2l] = seriescoeff(N, p_lengthscale, p_variance)
dq2l = 2 * dq2l
eps = 1e-12
if np.any(np.isfinite(q2) == False) or np.any(np.abs(q2) > 1.0 / eps
) or np.any(np.abs(q2) < eps):
warnings.warn(
'sde_Periodic: Infinite, too small, or too large (eps={0:e}) values in q2 :'
.format(eps) + q2.__format__(''))
if np.any(np.isfinite(dq2l) == False) or np.any(np.abs(dq2l) > 1.0 / eps
) or np.any(np.abs(dq2l) < eps):
warnings.warn(
'sde_Periodic: Infinite, too small, or too large (eps={0:e}) values in dq2l :'
.format(eps) + q2.__format__(''))
F = np.kron(np.diag(range(0, N + 1)), np.array(((0, -w0), (w0, 0))))
L = np.eye(2 * (N + 1))
Qc = np.zeros((2 * (N + 1), 2 * (N + 1)))
P_inf = np.kron(np.diag(q2), np.eye(2))
H = np.kron(np.ones((1, N + 1)), np.array((1, 0)))
P0 = P_inf.copy()
dF = np.empty((F.shape[0], F.shape[1], 3))
dQc = np.empty((Qc.shape[0], Qc.shape[1], 3))
dP_inf = np.empty((P_inf.shape[0], P_inf.shape[1], 3))
dF[:, :, 0] = np.zeros(F.shape)
dQc[:, :, 0] = np.zeros(Qc.shape)
dP_inf[:, :, 0] = P_inf / p_variance
dF[:, :, 1] = np.kron(np.diag(range(0, N + 1)), np.array(((0, w0), (-w0,
0))) / p_period)
dQc[:, :, 1] = np.zeros(Qc.shape)
dP_inf[:, :, 1] = np.zeros(P_inf.shape)
dF[:, :, 2] = np.zeros(F.shape)
dQc[:, :, 2] = np.zeros(Qc.shape)
dP_inf[:, :, 2] = np.kron(np.diag(dq2l), np.eye(2))
dP0 = dP_inf.copy()
if self.balance:
import GPy.models.state_space_main as ssm
F, L, Qc, H, P_inf, P0, dF, dQc, dP_inf, dP0 = ssm.balance_ss_model(F,
L, Qc, H, P_inf, P0, dF, dQc, dP_inf, dP0)
return F, L, Qc, H, P_inf, P0, dF, dQc, dP_inf, dP0
| sde | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.approx_order is not None:\n N = int(self.approx_order)\n else:\n N = 7\n p_period = float(self.period)\n p_lengthsca... |
def x(self, link_f, y, Y_metadata=None):
assert np.atleast_1d(link_f).shape == np.atleast_1d(y).shape
log_objective = link_f * np.exp(-y * link_f)
return np.exp(np.sum(np.log(log_objective)))
| pdf_link | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, link_f, y, Y_metadata=None):\n assert np.atleast_1d(link_f).shape == np.atleast_1d(y).shape\n log_objective = link_f * np.exp(-y * link_f)\n ret... |
def x(self, link_f, y, Y_metadata=None):
log_objective = np.log(link_f) - y * link_f
return log_objective
| logpdf_link | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, link_f, y, Y_metadata=None):\n log_objective = np.log(link_f) - y * link_f\n return log_objective\n",
"role": "user"
},
{
"content": "l... |
def x(self, link_f, y, Y_metadata=None):
assert np.atleast_1d(link_f).shape == np.atleast_1d(y).shape
return np.exp(self.logpdf_link(link_f, y, Y_metadata))
| pdf_link | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, link_f, y, Y_metadata=None):\n assert np.atleast_1d(link_f).shape == np.atleast_1d(y).shape\n return np.exp(self.logpdf_link(link_f, y, Y_metadata)... |
def x(self, link_f, y, Y_metadata=None):
return -link_f + y * np.log(link_f) - special.gammaln(y + 1)
| logpdf_link | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, link_f, y, Y_metadata=None):\n return -link_f + y * np.log(link_f) - special.gammaln(y + 1)\n",
"role": "user"
},
{
"content": "logpdf_lin... |
def x(self, gp, Y_metadata=None):
orig_shape = gp.shape
gp = gp.flatten()
Ysim = np.random.poisson(self.gp_link.transf(gp))
return Ysim.reshape(orig_shape)
| samples | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, gp, Y_metadata=None):\n orig_shape = gp.shape\n gp = gp.flatten()\n Ysim = np.random.poisson(self.gp_link.transf(gp))\n return Ysim.reshape(o... |
def x(self, z, max_iterations=250, y=None):
z = z.copy()
y = np.ones_like(z)
it = 0
update = np.inf
while np.abs(update).sum() > 1e-10 and it < max_iterations:
fy = self.f(y)
fgrady = self.fgrad_y(y)
update = (fy - z) / fgrady
y -= self.rate * update
it += 1
return y
| f_inv | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, z, max_iterations=250, y=None):\n z = z.copy()\n y = np.ones_like(z)\n it = 0\n update = np.inf\n while np.abs(update).sum() > 1e-10 and i... |
def x(self, y):
d = self.d
mpsi = self.psi
z = d * y.copy()
for i in range(len(mpsi)):
a, b, c = mpsi[i]
z += a * np.tanh(b * (y + c))
return z
| f | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, y):\n d = self.d\n mpsi = self.psi\n z = d * y.copy()\n for i in range(len(mpsi)):\n a, b, c = mpsi[i]\n z += a * np.tanh(b * (... |
def x(self, y, return_precalc=False):
d = self.d
mpsi = self.psi
S = (mpsi[:, 1] * (y[:, :, None] + mpsi[:, 2])).T
R = np.tanh(S)
D = 1 - R ** 2
GRAD = (d + (mpsi[:, 0:1][:, :, None] * mpsi[:, 1:2][:, :, None] * D).
sum(axis=0)).T
if return_precalc:
return GRAD, S, R, D
return GRAD
| fgrad_y | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, y, return_precalc=False):\n d = self.d\n mpsi = self.psi\n S = (mpsi[:, 1] * (y[:, :, None] + mpsi[:, 2])).T\n R = np.tanh(S)\n D = 1 - R ... |
def x(self, y, return_covar_chain=False):
mpsi = self.psi
w, s, r, d = self.fgrad_y(y, return_precalc=True)
gradients = np.zeros((y.shape[0], y.shape[1], len(mpsi), 4))
for i in range(len(mpsi)):
a, b, c = mpsi[i]
gradients[:, :, i, 0] = (b * (1.0 / np.cosh(s[i])) ** 2).T
gradients[:, :, i, 1] = a * (d[i] - 2.0 * s[i] * r[i] * (1.0 / np.
cosh(s[i])) ** 2).T
gradients[:, :, i, 2] = (-2.0 * a * b ** 2 * r[i] * (1.0 / np.cosh(
s[i])) ** 2).T
gradients[:, :, 0, 3] = 1.0
if return_covar_chain:
covar_grad_chain = np.zeros((y.shape[0], y.shape[1], len(mpsi), 4))
for i in range(len(mpsi)):
a, b, c = mpsi[i]
covar_grad_chain[:, :, i, 0] = r[i].T
covar_grad_chain[:, :, i, 1] = a * (y + c) * ((1.0 / np.cosh(s[
i])) ** 2).T
covar_grad_chain[:, :, i, 2] = a * b * ((1.0 / np.cosh(s[i])) ** 2
).T
covar_grad_chain[:, :, 0, 3] = y
return gradients, covar_grad_chain
return gradients
| fgrad_y_psi | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, y, return_covar_chain=False):\n mpsi = self.psi\n w, s, r, d = self.fgrad_y(y, return_precalc=True)\n gradients = np.zeros((y.shape[0], y.shape[... |
def x(x):
z = 2 * x - 1
ln1z2 = np.log(1 - z ** 2)
a = 8 * (np.pi - 3) / (3 * np.pi * (4 - np.pi))
b = 2 / (np.pi * a) + ln1z2 / 2
inv_erf = np.sign(z) * np.sqrt(np.sqrt(b ** 2 - ln1z2 / a) - b)
return np.sqrt(2) * inv_erf
| inv_std_norm_cdf | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(x):\n z = 2 * x - 1\n ln1z2 = np.log(1 - z ** 2)\n a = 8 * (np.pi - 3) / (3 * np.pi * (4 - np.pi))\n b = 2 / (np.pi * a) + ln1z2 / 2\n inv_erf =... |
def x(z):
if abs(z) < ERF_CODY_LIMIT1:
return 0.5 * (1.0 + z / M_SQRT2 * _erfRationalHelperR3(0.5 * z * z))
elif z < 0.0:
return np.exp(logPdfNormal(z)) * _erfRationalHelper(-z) / -z
else:
return 1.0 - np.exp(logPdfNormal(z)) * _erfRationalHelper(z) / z
| cdfNormal | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(z):\n if abs(z) < ERF_CODY_LIMIT1:\n return 0.5 * (1.0 + z / M_SQRT2 * _erfRationalHelperR3(0.5 * z * z))\n elif z < 0.0:\n return np.exp(l... |
def x(z):
if abs(z) < ERF_CODY_LIMIT1:
return np.log1p(z / M_SQRT2 * _erfRationalHelperR3(0.5 * z * z)
) - M_LN2
elif z < 0.0:
return logPdfNormal(z) - np.log(-z) + np.log(_erfRationalHelper(-z))
else:
return np.log1p(-np.exp(logPdfNormal(z)) * _erfRationalHelper(z) / z)
| logCdfNormal | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(z):\n if abs(z) < ERF_CODY_LIMIT1:\n return np.log1p(z / M_SQRT2 * _erfRationalHelperR3(0.5 * z * z)\n ) - M_LN2\n elif z < 0.0:\n ... |
def x(z):
if abs(z) < ERF_CODY_LIMIT1:
return 2.0 * np.exp(logPdfNormal(z)) / (1.0 + z / M_SQRT2 *
_erfRationalHelperR3(0.5 * z * z))
elif z < 0.0:
return -z / _erfRationalHelper(-z)
else:
t = np.exp(logPdfNormal(z))
return t / (1.0 - t * _erfRationalHelper(z) / z)
| derivLogCdfNormal | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(z):\n if abs(z) < ERF_CODY_LIMIT1:\n return 2.0 * np.exp(logPdfNormal(z)) / (1.0 + z / M_SQRT2 *\n _erfRationalHelperR3(0.5 * z * z))\n ... |
def x(index):
if len(index) == 0:
return []
ind = np.asarray(index, dtype=np.int)
ret = [[] for i in range(ind.max() + 1)]
ind_ = np.hstack((ind, ind[0] + ind[-1] + 1))
switchpoints = np.nonzero(ind_ - np.roll(ind_, +1))[0]
[ret[ind_i].append(slice(*indexes_i)) for ind_i, indexes_i in zip(ind[
switchpoints[:-1]], zip(switchpoints, switchpoints[1:]))]
return ret
| index_to_slices | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(index):\n if len(index) == 0:\n return []\n ind = np.asarray(index, dtype=np.int)\n ret = [[] for i in range(ind.max() + 1)]\n ind_ = np.hst... |
def x(prompt):
yes = set(['yes', 'y'])
no = set(['no', 'n'])
try:
print(prompt)
choice = input().lower()
except:
print('Stdin is not implemented.')
print('You need to set')
print('overide_manual_authorize=True')
print(
'to proceed with the download. Please set that variable and continue.'
)
raise
if choice in yes:
return True
elif choice in no:
return False
else:
print('Your response was a ' + choice)
print("Please respond with 'yes', 'y' or 'no', 'n'")
| prompt_user | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(prompt):\n yes = set(['yes', 'y'])\n no = set(['no', 'n'])\n try:\n print(prompt)\n choice = input().lower()\n except:\n print... |
def x(dataset_name=None):
try:
from itertools import zip_longest
except ImportError:
from itertools import izip_longest as zip_longest
dr = data_resources[dataset_name]
zip_urls = dr['files'],
if 'save_names' in dr:
zip_urls += dr['save_names'],
else:
zip_urls += [],
for file_list, save_list in zip_longest(*zip_urls, fillvalue=[]):
for f, s in zip_longest(file_list, save_list, fillvalue=None):
if s is not None:
f = s
if not os.path.exists(os.path.join(data_path, dataset_name, f)):
return False
return True
| data_available | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dataset_name=None):\n try:\n from itertools import zip_longest\n except ImportError:\n from itertools import izip_longest as zip_longest\n ... |
def x(url, store_directory, save_name=None, messages=True, suffix=''):
i = url.rfind('/')
file = url[i + 1:]
print(file)
dir_name = os.path.join(data_path, store_directory)
if save_name is None:
save_name = os.path.join(dir_name, file)
else:
save_name = os.path.join(dir_name, save_name)
if suffix is None:
suffix = ''
print('Downloading ', url, '->', save_name)
if not os.path.exists(dir_name):
os.makedirs(dir_name)
try:
response = urlopen(url + suffix)
except URLError as e:
if not hasattr(e, 'code'):
raise
response = e
if response.code > 399 and response.code < 500:
raise ValueError('Tried url ' + url + suffix +
' and received client error ' + str(response.code))
elif response.code > 499:
raise ValueError('Tried url ' + url + suffix +
' and received server error ' + str(response.code))
with open(save_name, 'wb') as f:
meta = response.info()
content_length_str = meta.get('Content-Length')
if content_length_str:
file_size = int(content_length_str)
else:
file_size = None
status = ''
file_size_dl = 0
block_sz = 8192
line_length = 30
while True:
buff = response.read(block_sz)
if not buff:
break
file_size_dl += len(buff)
f.write(buff)
sys.stdout.write(' ' * len(status) + '\r')
if file_size:
status = '[{perc: <{ll}}] {dl:7.3f}/{full:.3f}MB'.format(dl
=file_size_dl / 1048576.0, full=file_size / 1048576.0,
ll=line_length, perc='=' * int(line_length * float(
file_size_dl) / file_size))
else:
status = '[{perc: <{ll}}] {dl:7.3f}MB'.format(dl=
file_size_dl / 1048576.0, ll=line_length, perc='.' *
int(line_length * float(file_size_dl / (10 * 1048576.0))))
sys.stdout.write(status)
sys.stdout.flush()
sys.stdout.write(' ' * len(status) + '\r')
print(status)
| download_url | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(url, store_directory, save_name=None, messages=True, suffix=''):\n i = url.rfind('/')\n file = url[i + 1:]\n print(file)\n dir_name = os.path.join(... |
def x(dataset_name=None):
print('Acquiring resource: ' + dataset_name)
print('')
dr = data_resources[dataset_name]
print('Details of data: ')
print(dr['details'])
print('')
if dr['citation']:
print('Please cite:')
print(dr['citation'])
print('')
if dr['size']:
print('After downloading the data will take up ' + str(dr['size']) +
' bytes of space.')
print('')
print('Data will be stored in ' + os.path.join(data_path, dataset_name) +
'.')
print('')
if overide_manual_authorize:
if dr['license']:
print('You have agreed to the following license:')
print(dr['license'])
print('')
return True
else:
if dr['license']:
print('You must also agree to the following license:')
print(dr['license'])
print('')
return prompt_user('Do you wish to proceed with the download? [yes/no]'
)
| authorize_download | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dataset_name=None):\n print('Acquiring resource: ' + dataset_name)\n print('')\n dr = data_resources[dataset_name]\n print('Details of data: ')\n ... |
def x(dataset_name=None):
try:
from itertools import zip_longest
except ImportError:
from itertools import izip_longest as zip_longest
dr = data_resources[dataset_name]
if not authorize_download(dataset_name):
raise Exception('Permission to download data set denied.')
zip_urls = dr['urls'], dr['files']
if 'save_names' in dr:
zip_urls += dr['save_names'],
else:
zip_urls += [],
if 'suffices' in dr:
zip_urls += dr['suffices'],
else:
zip_urls += [],
for url, files, save_names, suffices in zip_longest(*zip_urls, fillvalue=[]
):
for f, save_name, suffix in zip_longest(files, save_names, suffices,
fillvalue=None):
download_url(os.path.join(url, f), dataset_name, save_name,
suffix=suffix)
return True
| download_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dataset_name=None):\n try:\n from itertools import zip_longest\n except ImportError:\n from itertools import izip_longest as zip_longest\n ... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.