source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(self, fid):
lin = self.read_line(fid)
while lin[0] != ':':
parts = lin.split()
if parts[0] == 'order':
order = []
for i in range(1, len(parts)):
if parts[i].lower() == 'rx':
chan = 'Xrotation'
order.append('x')
elif parts[i].lower() == 'ry':
chan = 'Yrotation'
order.append('y')
elif parts[i].lower() == 'rz':
chan = 'Zrotation'
order.append('z')
elif parts[i].lower() == 'tx':
chan = 'Xposition'
elif parts[i].lower() == 'ty':
chan = 'Yposition'
elif parts[i].lower() == 'tz':
chan = 'Zposition'
elif parts[i].lower() == 'l':
chan = 'length'
self.vertices[0].meta['channels'].append(chan)
self.vertices[0].meta['order'] = order[::-1]
elif parts[0] == 'axis':
self.vertices[0].meta['axis_order'] = parts[1][::-1].lower()
elif parts[0] == 'position':
self.vertices[0].meta['offset'] = [float(parts[1]), float(parts
[2]), float(parts[3])]
elif parts[0] == 'orientation':
self.vertices[0].meta['orientation'] = [float(parts[1]), float(
parts[2]), float(parts[3])]
lin = self.read_line(fid)
return lin
| read_root | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, fid):\n lin = self.read_line(fid)\n while lin[0] != ':':\n parts = lin.split()\n if parts[0] == 'order':\n order = []\n ... |
def x(self, fid):
lin = self.read_line(fid)
while lin:
if lin[0] == ':':
if lin[1:] == 'name':
lin = self.read_line(fid)
self.name = lin
elif lin[1:] == 'units':
lin = self.read_units(fid)
elif lin[1:] == 'documentation':
lin = self.read_documentation(fid)
elif lin[1:] == 'root':
lin = self.read_root(fid)
elif lin[1:] == 'bonedata':
lin = self.read_bonedata(fid)
elif lin[1:] == 'hierarchy':
lin = self.read_hierarchy(fid)
elif lin[1:8] == 'version':
lin = self.read_line(fid)
continue
else:
if not lin:
self.finalize()
return
lin = self.read_line(fid)
else:
raise ValueError('Unrecognised file format')
self.finalize()
| read_skel | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, fid):\n lin = self.read_line(fid)\n while lin:\n if lin[0] == ':':\n if lin[1:] == 'name':\n lin = self.read_line(... |
def x(self, fid):
lin = self.read_line(fid)
while lin[0] != ':':
parts = lin.split()
if parts[0] == 'mass':
self.mass = float(parts[1])
elif parts[0] == 'length':
self.length = float(parts[1])
elif parts[0] == 'angle':
self.angle = parts[1]
lin = self.read_line(fid)
return lin
| read_units | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, fid):\n lin = self.read_line(fid)\n while lin[0] != ':':\n parts = lin.split()\n if parts[0] == 'mass':\n self.mass = floa... |
def x(self, index, start_val):
channels = self.vertices[index].meta['channels']
base_channel = start_val
rot_ind = -np.ones(3, dtype=int)
pos_ind = -np.ones(3, dtype=int)
for i in range(len(channels)):
if channels[i] == 'Xrotation':
rot_ind[0] = base_channel + i
elif channels[i] == 'Yrotation':
rot_ind[1] = base_channel + i
elif channels[i] == 'Zrotation':
rot_ind[2] = base_channel + i
elif channels[i] == 'Xposition':
pos_ind[0] = base_channel + i
elif channels[i] == 'Yposition':
pos_ind[1] = base_channel + i
elif channels[i] == 'Zposition':
pos_ind[2] = base_channel + i
self.vertices[index].meta['rot_ind'] = list(rot_ind)
self.vertices[index].meta['pos_ind'] = list(pos_ind)
| resolve_indices | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, index, start_val):\n channels = self.vertices[index].meta['channels']\n base_channel = start_val\n rot_ind = -np.ones(3, dtype=int)\n pos_ind... |
def x(self):
for i in range(len(self.vertices)):
self.vertices[i].meta['C'] = rotation_matrix(self.vertices[i].meta[
'axis'][0], self.vertices[i].meta['axis'][1], self.vertices[i].
meta['axis'][2], self.vertices[i].meta['axis_order'], degrees=True)
self.vertices[i].meta['Cinv'] = np.linalg.inv(self.vertices[i].meta
['C'])
| set_rotation_matrices | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n for i in range(len(self.vertices)):\n self.vertices[i].meta['C'] = rotation_matrix(self.vertices[i].meta[\n 'axis'][0], self.vert... |
def x(self, X):
X_prod = self._comp_prod(X)
return self.variance * four_over_tau * np.arcsin(X_prod / (X_prod + 1.0))
| Kdiag | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n X_prod = self._comp_prod(X)\n return self.variance * four_over_tau * np.arcsin(X_prod / (X_prod + 1.0))\n",
"role": "user"
},
{
"... |
def x(self, dL_dK, X, X2=None):
dvar, dw, db = self._comp_grads(dL_dK, X, X2)[:3]
self.variance.gradient = dvar
self.weight_variance.gradient = dw
self.bias_variance.gradient = db
| update_gradients_full | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dL_dK, X, X2=None):\n dvar, dw, db = self._comp_grads(dL_dK, X, X2)[:3]\n self.variance.gradient = dvar\n self.weight_variance.gradient = dw\n ... |
def x(self, dL_dK, X, X2):
return self._comp_grads(dL_dK, X, X2)[3]
| gradients_X | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dL_dK, X, X2):\n return self._comp_grads(dL_dK, X, X2)[3]\n",
"role": "user"
},
{
"content": "gradients_X",
"role": "assistant"
}
] |
def x(self, dL_dK, X, X2):
return self._comp_grads(dL_dK, X, X2)[3:]
| gradients_X_X2 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dL_dK, X, X2):\n return self._comp_grads(dL_dK, X, X2)[3:]\n",
"role": "user"
},
{
"content": "gradients_X_X2",
"role": "assistant"
... |
def x(self):
return np.hstack((self.varianceU, self.varianceY, self.lengthscaleU,
self.lengthscaleY))
| get_params | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return np.hstack((self.varianceU, self.varianceY, self.lengthscaleU,\n self.lengthscaleY))\n",
"role": "user"
},
{
"content": "g... |
def x(self, x):
assert x.size == self.num_params
self.varianceU = x[0]
self.varianceY = x[1]
self.lengthscaleU = x[2]
self.lengthscaleY = x[3]
| set_params | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, x):\n assert x.size == self.num_params\n self.varianceU = x[0]\n self.varianceY = x[1]\n self.lengthscaleU = x[2]\n self.lengthscaleY = x[... |
def x(self, X, X2, target):
if X2 is None:
X2 = X
dist = np.abs(X - X2.T)
ly = 1 / self.lengthscaleY
lu = np.sqrt(3) / self.lengthscaleU
k1 = np.exp(-ly * dist) * (2 * lu + ly) / (lu + ly) ** 2
k2 = (np.exp(-lu * dist) * (ly - 2 * lu + lu * ly * dist - lu ** 2 *
dist) + np.exp(-ly * dist) * (2 * lu - ly)) / (ly - lu) ** 2
k3 = np.exp(-lu * dist) * ((1 + lu * dist) / (lu + ly) + lu / (lu + ly) **
2)
np.add(self.varianceU * self.varianceY * (k1 + k2 + k3), target, target)
| K | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, X2, target):\n if X2 is None:\n X2 = X\n dist = np.abs(X - X2.T)\n ly = 1 / self.lengthscaleY\n lu = np.sqrt(3) / self.lengthscaleU... |
def x(self, X, target):
ly = 1 / self.lengthscaleY
lu = np.sqrt(3) / self.lengthscaleU
k1 = (2 * lu + ly) / (lu + ly) ** 2
k2 = (ly - 2 * lu + 2 * lu - ly) / (ly - lu) ** 2
k3 = 1 / (lu + ly) + lu / (lu + ly) ** 2
np.add(self.varianceU * self.varianceY * (k1 + k2 + k3), target, target)
| Kdiag | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, target):\n ly = 1 / self.lengthscaleY\n lu = np.sqrt(3) / self.lengthscaleU\n k1 = (2 * lu + ly) / (lu + ly) ** 2\n k2 = (ly - 2 * lu + 2 ... |
def x(self, dL_dK, X, X2, target):
if X2 is None:
X2 = X
dist = np.abs(X - X2.T)
ly = 1 / self.lengthscaleY
lu = np.sqrt(3) / self.lengthscaleU
dk1theta1 = np.exp(-ly * dist) * 2 * -lu / (lu + ly) ** 3
dk2theta1 = 1 * (np.exp(-lu * dist) * dist * (-ly + 2 * lu - lu * ly *
dist + dist * lu ** 2) * (ly - lu) ** -2 + np.exp(-lu * dist) * (-2 +
ly * dist - 2 * dist * lu) * (ly - lu) ** -2 + np.exp(-dist * lu) *
(ly - 2 * lu + ly * lu * dist - dist * lu ** 2) * 2 * (ly - lu) **
-3 + np.exp(-dist * ly) * 2 * (ly - lu) ** -2 + np.exp(-dist * ly) *
2 * (2 * lu - ly) * (ly - lu) ** -3)
dk3theta1 = np.exp(-dist * lu) * (lu + ly) ** -2 * ((2 * lu + ly + dist *
lu ** 2 + lu * ly * dist) * (-dist - 2 / (lu + ly)) + 2 + 2 * lu *
dist + ly * dist)
dktheta1 = self.varianceU * self.varianceY * (dk1theta1 + dk2theta1 +
dk3theta1)
dk1theta2 = np.exp(-ly * dist) * (lu + ly) ** -2 * (-dist * (2 * lu +
ly) + 1 + -2 * (2 * lu + ly) / (lu + ly))
dk2theta2 = 1 * (np.exp(-dist * lu) * (ly - lu) ** -2 * (1 + lu * dist +
-2 * (ly - 2 * lu + lu * ly * dist - dist * lu ** 2) * (ly - lu) **
-1) + np.exp(-dist * ly) * (ly - lu) ** -2 * (-dist * (2 * lu - ly) -
1 + (2 * lu - ly) * -2 * (ly - lu) ** -1))
dk3theta2 = np.exp(-dist * lu) * (-3 * lu - ly - dist * lu ** 2 - lu *
ly * dist) / (lu + ly) ** 3
dktheta2 = self.varianceU * self.varianceY * (dk1theta2 + dk2theta2 +
dk3theta2)
k1 = np.exp(-ly * dist) * (2 * lu + ly) / (lu + ly) ** 2
k2 = (np.exp(-lu * dist) * (ly - 2 * lu + lu * ly * dist - lu ** 2 *
dist) + np.exp(-ly * dist) * (2 * lu - ly)) / (ly - lu) ** 2
k3 = np.exp(-lu * dist) * ((1 + lu * dist) / (lu + ly) + lu / (lu + ly) **
2)
dkdvar = k1 + k2 + k3
target[0] += np.sum(self.varianceY * dkdvar * dL_dK)
target[1] += np.sum(self.varianceU * dkdvar * dL_dK)
target[2] += np.sum(dktheta1 * (-np.sqrt(3) * self.lengthscaleU ** -2) *
dL_dK)
target[3] += np.sum(dktheta2 * -self.lengthscaleY ** -2 * dL_dK)
| param_grad_helper | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dL_dK, X, X2, target):\n if X2 is None:\n X2 = X\n dist = np.abs(X - X2.T)\n ly = 1 / self.lengthscaleY\n lu = np.sqrt(3) / self.lengt... |
def x(self, kern, X, likelihood, Y, Y_metadata=None):
N = X.shape[0]
D = X.shape[1]
Kds = np.zeros(D, dtype=object)
Qs = np.zeros(D, dtype=object)
QTs = np.zeros(D, dtype=object)
V_kron = 1
oneDkernel = kern.get_one_dimensional_kernel(D)
for d in range(D):
xg = list(set(X[:, d]))
xg = np.reshape(xg, (len(xg), 1))
oneDkernel.lengthscale = kern.lengthscale[d]
Kds[d] = oneDkernel.K(xg)
[V, Q] = np.linalg.eig(Kds[d])
V_kron = np.kron(V_kron, V)
Qs[d] = Q
QTs[d] = Q.T
noise = likelihood.variance + 1e-08
alpha_kron = self.kron_mvprod(QTs, Y)
V_kron = V_kron.reshape(-1, 1)
alpha_kron = alpha_kron / (V_kron + noise)
alpha_kron = self.kron_mvprod(Qs, alpha_kron)
log_likelihood = -0.5 * (np.dot(Y.T, alpha_kron) + np.sum(np.log(V_kron +
noise)) + N * log_2_pi)
derivs = np.zeros(D + 2, dtype='object')
for t in range(len(derivs)):
dKd_dTheta = np.zeros(D, dtype='object')
gamma = np.zeros(D, dtype='object')
gam = 1
for d in range(D):
xg = list(set(X[:, d]))
xg = np.reshape(xg, (len(xg), 1))
oneDkernel.lengthscale = kern.lengthscale[d]
if t < D:
dKd_dTheta[d] = oneDkernel.dKd_dLen(xg, t == d, lengthscale
=kern.lengthscale[t])
elif t == D:
dKd_dTheta[d] = oneDkernel.dKd_dVar(xg)
else:
dKd_dTheta[d] = np.identity(len(xg))
gamma[d] = np.diag(np.dot(np.dot(QTs[d], dKd_dTheta[d].T), Qs[d]))
gam = np.kron(gam, gamma[d])
gam = gam.reshape(-1, 1)
kappa = self.kron_mvprod(dKd_dTheta, alpha_kron)
derivs[t] = 0.5 * np.dot(alpha_kron.T, kappa) - 0.5 * np.sum(gam /
(V_kron + noise))
dL_dLen = derivs[:D]
dL_dVar = derivs[D]
dL_dThetaL = derivs[D + 1]
return GridPosterior(alpha_kron=alpha_kron, QTs=QTs, Qs=Qs, V_kron=V_kron
), log_likelihood, {'dL_dLen': dL_dLen, 'dL_dVar': dL_dVar,
'dL_dthetaL': dL_dThetaL}
| inference | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, kern, X, likelihood, Y, Y_metadata=None):\n N = X.shape[0]\n D = X.shape[1]\n Kds = np.zeros(D, dtype=object)\n Qs = np.zeros(D, dtype=object... |
def x(self):
input_dict = super(Constant, self)._save_to_input_dict()
input_dict['class'] = 'GPy.mappings.Constant'
input_dict['value'] = self.C.values[0]
return input_dict
| to_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n input_dict = super(Constant, self)._save_to_input_dict()\n input_dict['class'] = 'GPy.mappings.Constant'\n input_dict['value'] = self.C.value... |
def x(self, grads):
self.sigma2.gradient = grads[0]
self.v.gradient = grads[1]
| update_gradients | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, grads):\n self.sigma2.gradient = grads[0]\n self.v.gradient = grads[1]\n",
"role": "user"
},
{
"content": "update_gradients",
"role... |
def x(self, inv_link_f, y, Y_metadata=None):
assert np.atleast_1d(inv_link_f).shape == np.atleast_1d(y).shape
e = y - inv_link_f
objective = np.exp(gammaln((self.v + 1) * 0.5) - gammaln(self.v * 0.5)
) / np.sqrt(self.v * np.pi * self.sigma2) * (1 + 1.0 / float(self.v
) * (e ** 2 / float(self.sigma2))) ** (-0.5 * (self.v + 1))
return np.prod(objective)
| pdf_link | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, inv_link_f, y, Y_metadata=None):\n assert np.atleast_1d(inv_link_f).shape == np.atleast_1d(y).shape\n e = y - inv_link_f\n objective = np.exp(ga... |
def x(self, inv_link_f, y, Y_metadata=None):
e = y - inv_link_f
objective = +gammaln((self.v + 1) * 0.5) - gammaln(self.v * 0.5
) - 0.5 * np.log(self.sigma2 * self.v * np.pi) - 0.5 * (self.v + 1
) * np.log(1 + 1 / np.float(self.v) * (e ** 2 / self.sigma2))
return objective
| logpdf_link | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, inv_link_f, y, Y_metadata=None):\n e = y - inv_link_f\n objective = +gammaln((self.v + 1) * 0.5) - gammaln(self.v * 0.5\n ) - 0.5 * np.log(s... |
def x(self, inv_link_f, y, Y_metadata=None):
e = y - inv_link_f
grad = (self.v + 1) * e / (self.v * self.sigma2 + e ** 2)
return grad
| dlogpdf_dlink | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, inv_link_f, y, Y_metadata=None):\n e = y - inv_link_f\n grad = (self.v + 1) * e / (self.v * self.sigma2 + e ** 2)\n return grad\n",
"role": ... |
def x(self, inv_link_f, y, Y_metadata=None):
e = y - inv_link_f
hess = (self.v + 1) * (e ** 2 - self.v * self.sigma2) / (self.sigma2 *
self.v + e ** 2) ** 2
return hess
| d2logpdf_dlink2 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, inv_link_f, y, Y_metadata=None):\n e = y - inv_link_f\n hess = (self.v + 1) * (e ** 2 - self.v * self.sigma2) / (self.sigma2 *\n self.v + e ... |
def x(self, inv_link_f, y, Y_metadata=None):
e = y - inv_link_f
d3lik_dlink3 = -(2 * (self.v + 1) * -e * (e ** 2 - 3 * self.v * self.
sigma2)) / (e ** 2 + self.sigma2 * self.v) ** 3
return d3lik_dlink3
| d3logpdf_dlink3 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, inv_link_f, y, Y_metadata=None):\n e = y - inv_link_f\n d3lik_dlink3 = -(2 * (self.v + 1) * -e * (e ** 2 - 3 * self.v * self.\n sigma2)) / (... |
def x(self, inv_link_f, y, Y_metadata=None):
e = y - inv_link_f
e2 = np.square(e)
dlogpdf_dvar = self.v * (e2 - self.sigma2) / (2 * self.sigma2 * (self.
sigma2 * self.v + e2))
return dlogpdf_dvar
| dlogpdf_link_dvar | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, inv_link_f, y, Y_metadata=None):\n e = y - inv_link_f\n e2 = np.square(e)\n dlogpdf_dvar = self.v * (e2 - self.sigma2) / (2 * self.sigma2 * (sel... |
def x(self, inv_link_f, y, Y_metadata=None):
e = y - inv_link_f
dlogpdf_dlink_dvar = self.v * (self.v + 1) * -e / (self.sigma2 * self.v +
e ** 2) ** 2
return dlogpdf_dlink_dvar
| dlogpdf_dlink_dvar | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, inv_link_f, y, Y_metadata=None):\n e = y - inv_link_f\n dlogpdf_dlink_dvar = self.v * (self.v + 1) * -e / (self.sigma2 * self.v +\n e ** 2) ... |
def x(self, inv_link_f, y, Y_metadata=None):
e = y - inv_link_f
d2logpdf_dlink2_dvar = self.v * (self.v + 1) * (self.sigma2 * self.v -
3 * e ** 2) / (self.sigma2 * self.v + e ** 2) ** 3
return d2logpdf_dlink2_dvar
| d2logpdf_dlink2_dvar | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, inv_link_f, y, Y_metadata=None):\n e = y - inv_link_f\n d2logpdf_dlink2_dvar = self.v * (self.v + 1) * (self.sigma2 * self.v - \n 3 * e ** 2... |
def x(self, gp, Y_metadata=None):
orig_shape = gp.shape
gp = gp.flatten()
dfs = np.ones_like(gp) * self.v
scales = np.ones_like(gp) * np.sqrt(self.sigma2)
student_t_samples = stats.t.rvs(dfs, loc=self.gp_link.transf(gp), scale
=scales)
return student_t_samples.reshape(orig_shape)
| samples | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, gp, Y_metadata=None):\n orig_shape = gp.shape\n gp = gp.flatten()\n dfs = np.ones_like(gp) * self.v\n scales = np.ones_like(gp) * np.sqrt(sel... |
def x(self, kern, X, Z, likelihood, Y, Y_metadata, Lm=None, dL_dKmm=None,
psi0=None, psi1=None, psi2=None, **kwargs):
return self.inference_method.inference(kern, X, Z, likelihood, Y,
Y_metadata, Lm=Lm, dL_dKmm=dL_dKmm, psi0=psi0, psi1=psi1, psi2=psi2,
**kwargs)
| inner_parameters_changed | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, kern, X, Z, likelihood, Y, Y_metadata, Lm=None, dL_dKmm=None,\n psi0=None, psi1=None, psi2=None, **kwargs):\n return self.inference_method.inferenc... |
def x(self, full_values=None, current_values=None, value_indices=None):
for key in current_values.keys():
if value_indices is not None and key in value_indices:
index = value_indices[key]
else:
index = slice(None)
if key in full_values:
try:
full_values[key][index] += current_values[key]
except:
full_values[key] += current_values[key]
else:
full_values[key] = current_values[key]
| inner_take_over_or_update | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, full_values=None, current_values=None, value_indices=None):\n for key in current_values.keys():\n if value_indices is not None and key in value... |
def x(self, full_values):
if self.has_uncertain_inputs():
dL_dKmm = full_values['dL_dKmm']
self.kern.update_gradients_full(dL_dKmm, self.Z, None)
kgrad = self.kern.gradient.copy()
self.kern.update_gradients_expectations(variational_posterior=self.
X, Z=self.Z, dL_dpsi0=full_values['dL_dpsi0'], dL_dpsi1=
full_values['dL_dpsi1'], dL_dpsi2=full_values['dL_dpsi2'])
self.kern.gradient += kgrad
self.Z.gradient = self.kern.gradients_X(dL_dKmm, self.Z)
self.Z.gradient += self.kern.gradients_Z_expectations(
variational_posterior=self.X, Z=self.Z, dL_dpsi0=full_values[
'dL_dpsi0'], dL_dpsi1=full_values['dL_dpsi1'], dL_dpsi2=
full_values['dL_dpsi2'])
else:
self.kern.update_gradients_diag(full_values['dL_dKdiag'], self.X)
kgrad = self.kern.gradient.copy()
self.kern.update_gradients_full(full_values['dL_dKnm'], self.X, self.Z)
kgrad += self.kern.gradient
self.kern.update_gradients_full(full_values['dL_dKmm'], self.Z, None)
self.kern.gradient += kgrad
self.Z.gradient = self.kern.gradients_X(full_values['dL_dKmm'], self.Z)
self.Z.gradient += self.kern.gradients_X(full_values['dL_dKnm'].T,
self.Z, self.X)
self.likelihood.update_gradients(full_values['dL_dthetaL'])
| outer_values_update | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, full_values):\n if self.has_uncertain_inputs():\n dL_dKmm = full_values['dL_dKmm']\n self.kern.update_gradients_full(dL_dKmm, self.Z, No... |
def x(self):
retd = dict(dL_dKmm=np.zeros((self.Z.shape[0], self.Z.shape[0])))
if self.has_uncertain_inputs():
retd.update(dict(dL_dpsi0=np.zeros(self.X.shape[0]), dL_dpsi1=np.
zeros((self.X.shape[0], self.Z.shape[0])), dL_dpsi2=np.zeros((
self.X.shape[0], self.Z.shape[0], self.Z.shape[0]))))
else:
retd.update({'dL_dKdiag': np.zeros(self.X.shape[0]), 'dL_dKnm': np.
zeros((self.X.shape[0], self.Z.shape[0]))})
return retd
| outer_init_full_values | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n retd = dict(dL_dKmm=np.zeros((self.Z.shape[0], self.Z.shape[0])))\n if self.has_uncertain_inputs():\n retd.update(dict(dL_dpsi0=np.zeros(... |
def x(self, save_data=True):
model_dict = super(SparseGPClassification, self).to_dict(save_data)
model_dict['class'] = 'GPy.models.SparseGPClassification'
return model_dict
| to_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, save_data=True):\n model_dict = super(SparseGPClassification, self).to_dict(save_data)\n model_dict['class'] = 'GPy.models.SparseGPClassification'\... |
def x(input_dict, data=None):
import GPy
m = GPy.core.model.Model.from_dict(input_dict, data)
from copy import deepcopy
sparse_gp = deepcopy(m)
return SparseGPClassification(sparse_gp.X, sparse_gp.Y, sparse_gp.Z,
sparse_gp.kern, sparse_gp.likelihood, sparse_gp.inference_method,
sparse_gp.mean_function, name='sparse_gp_classification')
| from_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(input_dict, data=None):\n import GPy\n m = GPy.core.model.Model.from_dict(input_dict, data)\n from copy import deepcopy\n sparse_gp = deepcopy(m)\n... |
def x(self):
input_dict = super(Gaussian, self)._save_to_input_dict()
input_dict['class'] = 'GPy.likelihoods.Gaussian'
input_dict['variance'] = self.variance.values.tolist()
return input_dict
| to_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n input_dict = super(Gaussian, self)._save_to_input_dict()\n input_dict['class'] = 'GPy.likelihoods.Gaussian'\n input_dict['variance'] = self.v... |
def x(self, data_i, tau_i, v_i, Y_metadata_i=None):
sigma2_hat = 1.0 / (1.0 / self.variance + tau_i)
mu_hat = sigma2_hat * (data_i / self.variance + v_i)
sum_var = self.variance + 1.0 / tau_i
Z_hat = 1.0 / np.sqrt(2.0 * np.pi * sum_var) * np.exp(-0.5 * (data_i -
v_i / tau_i) ** 2.0 / sum_var)
return Z_hat, mu_hat, sigma2_hat
| moments_match_ep | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data_i, tau_i, v_i, Y_metadata_i=None):\n sigma2_hat = 1.0 / (1.0 / self.variance + tau_i)\n mu_hat = sigma2_hat * (data_i / self.variance + v_i)\n... |
def x(self, link_f, y, Y_metadata=None):
return np.exp(np.sum(np.log(stats.norm.pdf(y, link_f, np.sqrt(self.
variance)))))
| pdf_link | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, link_f, y, Y_metadata=None):\n return np.exp(np.sum(np.log(stats.norm.pdf(y, link_f, np.sqrt(self.\n variance)))))\n",
"role": "user"
},
... |
def x(self, link_f, y, Y_metadata=None):
ln_det_cov = np.log(self.variance)
return -(1.0 / (2 * self.variance)) * (y - link_f
) ** 2 - 0.5 * ln_det_cov - 0.5 * np.log(2.0 * np.pi)
| logpdf_link | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, link_f, y, Y_metadata=None):\n ln_det_cov = np.log(self.variance)\n return -(1.0 / (2 * self.variance)) * (y - link_f\n ) ** 2 - 0.5 * ln_de... |
def x(self, link_f, y, Y_metadata=None):
s2_i = 1.0 / self.variance
grad = s2_i * y - s2_i * link_f
return grad
| dlogpdf_dlink | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, link_f, y, Y_metadata=None):\n s2_i = 1.0 / self.variance\n grad = s2_i * y - s2_i * link_f\n return grad\n",
"role": "user"
},
{
"c... |
def x(self, link_f, y, Y_metadata=None):
N = y.shape[0]
D = link_f.shape[1]
hess = -(1.0 / self.variance) * np.ones((N, D))
return hess
| d2logpdf_dlink2 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, link_f, y, Y_metadata=None):\n N = y.shape[0]\n D = link_f.shape[1]\n hess = -(1.0 / self.variance) * np.ones((N, D))\n return hess\n",
"... |
def x(self, link_f, y, Y_metadata=None):
N = y.shape[0]
D = link_f.shape[1]
d3logpdf_dlink3 = np.zeros((N, D))
return d3logpdf_dlink3
| d3logpdf_dlink3 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, link_f, y, Y_metadata=None):\n N = y.shape[0]\n D = link_f.shape[1]\n d3logpdf_dlink3 = np.zeros((N, D))\n return d3logpdf_dlink3\n",
"ro... |
def x(self, link_f, y, Y_metadata=None):
e = y - link_f
s_4 = 1.0 / self.variance ** 2
dlik_dsigma = -0.5 / self.variance + 0.5 * s_4 * np.square(e)
return dlik_dsigma
| dlogpdf_link_dvar | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, link_f, y, Y_metadata=None):\n e = y - link_f\n s_4 = 1.0 / self.variance ** 2\n dlik_dsigma = -0.5 / self.variance + 0.5 * s_4 * np.square(e)\n... |
def x(self, link_f, y, Y_metadata=None):
s_4 = 1.0 / self.variance ** 2
dlik_grad_dsigma = -s_4 * y + s_4 * link_f
return dlik_grad_dsigma
| dlogpdf_dlink_dvar | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, link_f, y, Y_metadata=None):\n s_4 = 1.0 / self.variance ** 2\n dlik_grad_dsigma = -s_4 * y + s_4 * link_f\n return dlik_grad_dsigma\n",
"ro... |
def x(self, link_f, y, Y_metadata=None):
s_4 = 1.0 / self.variance ** 2
N = y.shape[0]
D = link_f.shape[1]
d2logpdf_dlink2_dvar = np.ones((N, D)) * s_4
return d2logpdf_dlink2_dvar
| d2logpdf_dlink2_dvar | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, link_f, y, Y_metadata=None):\n s_4 = 1.0 / self.variance ** 2\n N = y.shape[0]\n D = link_f.shape[1]\n d2logpdf_dlink2_dvar = np.ones((N, D))... |
def x(self, gp, Y_metadata=None):
orig_shape = gp.shape
gp = gp.flatten()
gp = gp.flatten()
Ysim = np.array([np.random.normal(self.gp_link.transf(gpj), scale=np.
sqrt(self.variance), size=1) for gpj in gp])
return Ysim.reshape(orig_shape)
| samples | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, gp, Y_metadata=None):\n orig_shape = gp.shape\n gp = gp.flatten()\n gp = gp.flatten()\n Ysim = np.array([np.random.normal(self.gp_link.transf... |
def x(self, y_test, mu_star, var_star, Y_metadata=None):
v = var_star + self.variance
return -0.5 * np.log(2 * np.pi) - 0.5 * np.log(v) - 0.5 * np.square(
y_test - mu_star) / v
| log_predictive_density | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, y_test, mu_star, var_star, Y_metadata=None):\n v = var_star + self.variance\n return -0.5 * np.log(2 * np.pi) - 0.5 * np.log(v) - 0.5 * np.square(\... |
def x(self, save_data=True):
input_dict = super(GP, self)._save_to_input_dict()
input_dict['class'] = 'GPy.core.GP'
if not save_data:
input_dict['X'] = None
input_dict['Y'] = None
else:
try:
input_dict['X'] = self.X.values.tolist()
except:
input_dict['X'] = self.X.tolist()
try:
input_dict['Y'] = self.Y.values.tolist()
except:
input_dict['Y'] = self.Y.tolist()
input_dict['kernel'] = self.kern.to_dict()
input_dict['likelihood'] = self.likelihood.to_dict()
if self.mean_function is not None:
input_dict['mean_function'] = self.mean_function.to_dict()
input_dict['inference_method'] = self.inference_method.to_dict()
if self.Y_metadata is not None:
input_dict['Y_metadata'] = self.Y_metadata
if self.normalizer is not None:
input_dict['normalizer'] = self.normalizer.to_dict()
return input_dict
| to_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, save_data=True):\n input_dict = super(GP, self)._save_to_input_dict()\n input_dict['class'] = 'GPy.core.GP'\n if not save_data:\n input_d... |
def x(self, X=None, Y=None):
self.update_model(False)
if Y is not None:
if self.normalizer is not None:
self.normalizer.scale_by(Y)
self.Y_normalized = ObsAr(self.normalizer.normalize(Y))
self.Y = Y
else:
self.Y = ObsAr(Y)
self.Y_normalized = self.Y
if X is not None:
if self.X in self.parameters:
if isinstance(self.X, VariationalPosterior):
assert isinstance(X, type(self.X)
), 'The given X must have the same type as the X in the model!'
index = self.X._parent_index_
self.unlink_parameter(self.X)
self.X = X
self.link_parameter(self.X, index=index)
else:
index = self.X._parent_index_
self.unlink_parameter(self.X)
from ..core import Param
self.X = Param('latent mean', X)
self.link_parameter(self.X, index=index)
else:
self.X = ObsAr(X)
self.update_model(True)
| set_XY | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X=None, Y=None):\n self.update_model(False)\n if Y is not None:\n if self.normalizer is not None:\n self.normalizer.scale_by(Y)\n... |
def x(self, Xnew, full_cov=False, Y_metadata=None, kern=None, likelihood=
None, include_likelihood=True):
mean, var = self._raw_predict(Xnew, full_cov=full_cov, kern=kern)
if include_likelihood:
if likelihood is None:
likelihood = self.likelihood
mean, var = likelihood.predictive_values(mean, var, full_cov,
Y_metadata=Y_metadata)
if self.normalizer is not None:
mean = self.normalizer.inverse_mean(mean)
if full_cov & (mean.shape[1] > 1):
var = self.normalizer.inverse_covariance(var)
else:
var = self.normalizer.inverse_variance(var)
return mean, var
| predict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Xnew, full_cov=False, Y_metadata=None, kern=None, likelihood=\n None, include_likelihood=True):\n mean, var = self._raw_predict(Xnew, full_cov=full... |
def x(self, Xnew, full_cov=False, Y_metadata=None, kern=None):
return self.predict(Xnew, full_cov, Y_metadata, kern, None, False)
| predict_noiseless | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Xnew, full_cov=False, Y_metadata=None, kern=None):\n return self.predict(Xnew, full_cov, Y_metadata, kern, None, False)\n",
"role": "user"
},
... |
def x(self, X, quantiles=(2.5, 97.5), Y_metadata=None, kern=None,
likelihood=None):
m, v = self._raw_predict(X, full_cov=False, kern=kern)
if likelihood is None:
likelihood = self.likelihood
quantiles = likelihood.predictive_quantiles(m, v, quantiles, Y_metadata
=Y_metadata)
if self.normalizer is not None:
quantiles = [self.normalizer.inverse_mean(q) for q in quantiles]
return quantiles
| predict_quantiles | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, quantiles=(2.5, 97.5), Y_metadata=None, kern=None,\n likelihood=None):\n m, v = self._raw_predict(X, full_cov=False, kern=kern)\n if likeliho... |
def x(self, Xnew, kern=None):
if kern is None:
kern = self.kern
mean_jac = np.empty((Xnew.shape[0], Xnew.shape[1], self.output_dim))
for i in range(self.output_dim):
mean_jac[:, :, i] = kern.gradients_X(self.posterior.woodbury_vector
[:, i:i + 1].T, Xnew, self._predictive_variable)
dv_dX = kern.gradients_X_diag(np.ones(Xnew.shape[0]), Xnew)
if self.posterior.woodbury_inv.ndim == 3:
var_jac = np.empty(dv_dX.shape + (self.posterior.woodbury_inv.shape
[2],))
var_jac[:] = dv_dX[:, :, None]
for i in range(self.posterior.woodbury_inv.shape[2]):
alpha = -2.0 * np.dot(kern.K(Xnew, self._predictive_variable),
self.posterior.woodbury_inv[:, :, i])
var_jac[:, :, i] += kern.gradients_X(alpha, Xnew, self.
_predictive_variable)
else:
var_jac = dv_dX
alpha = -2.0 * np.dot(kern.K(Xnew, self._predictive_variable), self
.posterior.woodbury_inv)
var_jac += kern.gradients_X(alpha, Xnew, self._predictive_variable)
return mean_jac, var_jac
| predictive_gradients | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Xnew, kern=None):\n if kern is None:\n kern = self.kern\n mean_jac = np.empty((Xnew.shape[0], Xnew.shape[1], self.output_dim))\n for i in... |
def x(self, Xnew, kern=None, full_cov=False):
if kern is None:
kern = self.kern
mean_jac = np.empty((Xnew.shape[0], Xnew.shape[1], self.output_dim))
for i in range(self.output_dim):
mean_jac[:, :, i] = kern.gradients_X(self.posterior.woodbury_vector
[:, i:i + 1].T, Xnew, self._predictive_variable)
dK_dXnew_full = np.empty((self._predictive_variable.shape[0], Xnew.
shape[0], Xnew.shape[1]))
one = np.ones((1, 1))
for i in range(self._predictive_variable.shape[0]):
dK_dXnew_full[i] = kern.gradients_X(one, Xnew, self.
_predictive_variable[[i]])
if full_cov:
dK2_dXdX = kern.gradients_XX(one, Xnew)
else:
dK2_dXdX = kern.gradients_XX_diag(one, Xnew)
def x(wi):
if full_cov:
var_jac = dK2_dXdX - np.einsum('qnm,msr->nsqr', dK_dXnew_full.T
.dot(wi), dK_dXnew_full)
else:
var_jac = dK2_dXdX - np.einsum('qnm,mnr->nqr', dK_dXnew_full.T.
dot(wi), dK_dXnew_full)
return var_jac
if self.posterior.woodbury_inv.ndim == 3:
if full_cov:
var_jac = np.empty((Xnew.shape[0], Xnew.shape[0], Xnew.shape[1],
Xnew.shape[1], self.output_dim))
for d in range(self.posterior.woodbury_inv.shape[2]):
var_jac[:, :, :, :, d] = compute_cov_inner(self.posterior.
woodbury_inv[:, :, d])
else:
var_jac = np.empty((Xnew.shape[0], Xnew.shape[1], Xnew.shape[1],
self.output_dim))
for d in range(self.posterior.woodbury_inv.shape[2]):
var_jac[:, :, :, d] = compute_cov_inner(self.posterior.
woodbury_inv[:, :, d])
else:
var_jac = compute_cov_inner(self.posterior.woodbury_inv)
return mean_jac, var_jac
| predict_jacobian | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Xnew, kern=None, full_cov=False):\n if kern is None:\n kern = self.kern\n mean_jac = np.empty((Xnew.shape[0], Xnew.shape[1], self.output_dim... |
def x(self, Xnew, kern=None, mean=True, covariance=True):
if kern is None:
kern = self.kern
mu_jac, var_jac = self.predict_jacobian(Xnew, kern, full_cov=False)
mumuT = np.einsum('iqd,ipd->iqp', mu_jac, mu_jac)
Sigma = np.zeros(mumuT.shape)
if var_jac.ndim == 4:
Sigma = var_jac.sum(-1)
else:
Sigma = self.output_dim * var_jac
G = 0.0
if mean:
G += mumuT
if covariance:
G += Sigma
return G
| predict_wishart_embedding | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Xnew, kern=None, mean=True, covariance=True):\n if kern is None:\n kern = self.kern\n mu_jac, var_jac = self.predict_jacobian(Xnew, kern, fu... |
def x(self, Xnew, kern=None, mean=True, covariance=True, dimensions=None):
G = self.predict_wishart_embedding(Xnew, kern, mean, covariance)
if dimensions is None:
dimensions = self.get_most_significant_input_dimensions()[:2]
G = G[:, dimensions][:, :, dimensions]
from ..util.linalg import jitchol
mag = np.empty(Xnew.shape[0])
for n in range(Xnew.shape[0]):
try:
mag[n] = np.sqrt(np.exp(2 * np.sum(np.log(np.diag(jitchol(G[n,
:, :]))))))
except:
mag[n] = np.sqrt(np.linalg.det(G[n, :, :]))
return mag
| predict_magnification | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Xnew, kern=None, mean=True, covariance=True, dimensions=None):\n G = self.predict_wishart_embedding(Xnew, kern, mean, covariance)\n if dimensions i... |
def x(self, X, size=10, **predict_kwargs):
predict_kwargs['full_cov'] = True
m, v = self._raw_predict(X, **predict_kwargs)
if self.normalizer is not None:
m, v = self.normalizer.inverse_mean(m
), self.normalizer.inverse_variance(v)
def x(m, v):
return np.random.multivariate_normal(m, v, size).T
if self.output_dim == 1:
return sim_one_dim(m.flatten(), v)[:, np.newaxis, :]
else:
fsim = np.empty((X.shape[0], self.output_dim, size))
for d in range(self.output_dim):
if v.ndim == 3:
fsim[:, d, :] = sim_one_dim(m[:, d], v[:, :, d])
else:
fsim[:, d, :] = sim_one_dim(m[:, d], v)
return fsim
| posterior_samples_f | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, size=10, **predict_kwargs):\n predict_kwargs['full_cov'] = True\n m, v = self._raw_predict(X, **predict_kwargs)\n if self.normalizer is not N... |
def x(self, X, size=10, Y_metadata=None, likelihood=None, **predict_kwargs):
fsim = self.posterior_samples_f(X, size, **predict_kwargs)
if likelihood is None:
likelihood = self.likelihood
if fsim.ndim == 3:
for d in range(fsim.shape[1]):
fsim[:, d] = likelihood.samples(fsim[:, d], Y_metadata=Y_metadata)
else:
fsim = likelihood.samples(fsim, Y_metadata=Y_metadata)
return fsim
| posterior_samples | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, size=10, Y_metadata=None, likelihood=None, **predict_kwargs):\n fsim = self.posterior_samples_f(X, size, **predict_kwargs)\n if likelihood is No... |
def x(self, optimizer=None, start=None, messages=False, max_iters=1000,
ipython_notebook=True, clear_after_finish=False, **kwargs):
self.inference_method.on_optimization_start()
try:
ret = super(GP, self).optimize(optimizer, start, messages,
max_iters, ipython_notebook, clear_after_finish, **kwargs)
except KeyboardInterrupt:
print(
'KeyboardInterrupt caught, calling on_optimization_end() to round things up'
)
self.inference_method.on_optimization_end()
raise
return ret
| optimize | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, optimizer=None, start=None, messages=False, max_iters=1000,\n ipython_notebook=True, clear_after_finish=False, **kwargs):\n self.inference_method.o... |
def x(self, Y_new, optimize=True):
from ..inference.latent_function_inference.inferenceX import infer_newX
return infer_newX(self, Y_new, optimize=optimize)
| infer_newX | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Y_new, optimize=True):\n from ..inference.latent_function_inference.inferenceX import infer_newX\n return infer_newX(self, Y_new, optimize=optimize... |
def x(self, x_test, y_test, Y_metadata=None):
mu_star, var_star = self._raw_predict(x_test)
return self.likelihood.log_predictive_density(y_test, mu_star, var_star,
Y_metadata=Y_metadata)
| log_predictive_density | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, x_test, y_test, Y_metadata=None):\n mu_star, var_star = self._raw_predict(x_test)\n return self.likelihood.log_predictive_density(y_test, mu_star, ... |
def x(self, x_test, y_test, Y_metadata=None, num_samples=1000):
mu_star, var_star = self._raw_predict(x_test)
return self.likelihood.log_predictive_density_sampling(y_test, mu_star,
var_star, Y_metadata=Y_metadata, num_samples=num_samples)
| log_predictive_density_sampling | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, x_test, y_test, Y_metadata=None, num_samples=1000):\n mu_star, var_star = self._raw_predict(x_test)\n return self.likelihood.log_predictive_density... |
def x(self, X1, X2):
return self.posterior.covariance_between_points(self.kern, self.X, X1, X2)
| posterior_covariance_between_points | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X1, X2):\n return self.posterior.covariance_between_points(self.kern, self.X, X1, X2)\n",
"role": "user"
},
{
"content": "posterior_covari... |
def x(shape, var_name, dims=None, coords=None, default_dims=None):
if default_dims is None:
default_dims = []
if dims is None:
dims = []
if len([dim for dim in dims if dim not in default_dims]) > len(shape):
warnings.warn(
'More dims ({dims_len}) given than exists ({shape_len}). Passed array should have shape (chains, draws, *shape)'
.format(dims_len=len(dims), shape_len=len(shape)), SyntaxWarning)
if coords is None:
coords = {}
coords = deepcopy(coords)
dims = deepcopy(dims)
for idx, dim_len in enumerate(shape):
if len(dims) < idx + 1 or dims[idx] is None:
dim_name = '{var_name}_dim_{idx}'.format(var_name=var_name, idx=idx
)
if len(dims) < idx + 1:
dims.append(dim_name)
else:
dims[idx] = dim_name
dim_name = dims[idx]
if dim_name not in coords:
coords[dim_name] = np.arange(dim_len)
coords = {key: coord for key, coord in coords.items() if any(key == dim for
dim in dims)}
return dims, coords
| generate_dims_coords | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(shape, var_name, dims=None, coords=None, default_dims=None):\n if default_dims is None:\n default_dims = []\n if dims is None:\n dims = []\... |
def x(ary, *, var_name='data', coords=None, dims=None):
default_dims = ['chain', 'draw']
ary = np.atleast_2d(ary)
n_chains, n_samples, *shape = ary.shape
if n_chains > n_samples:
warnings.warn(
'More chains ({n_chains}) than draws ({n_samples}). Passed array should have shape (chains, draws, *shape)'
.format(n_chains=n_chains, n_samples=n_samples), SyntaxWarning)
dims, coords = generate_dims_coords(shape, var_name, dims=dims, coords=
coords, default_dims=default_dims)
if 'draw' not in dims:
dims = ['draw'] + dims
if 'chain' not in dims:
dims = ['chain'] + dims
if 'chain' not in coords:
coords['chain'] = np.arange(n_chains)
if 'draw' not in coords:
coords['draw'] = np.arange(n_samples)
coords = {key: xr.IndexVariable((key,), data=coords[key]) for key in dims}
return xr.DataArray(ary, coords=coords, dims=dims)
| numpy_to_data_array | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ary, *, var_name='data', coords=None, dims=None):\n default_dims = ['chain', 'draw']\n ary = np.atleast_2d(ary)\n n_chains, n_samples, *shape = ary.sh... |
def x(data, *, attrs=None, library=None, coords=None, dims=None):
if dims is None:
dims = {}
data_vars = {}
for key, values in data.items():
data_vars[key] = numpy_to_data_array(values, var_name=key, coords=
coords, dims=dims.get(key))
return xr.Dataset(data_vars=data_vars, attrs=make_attrs(attrs=attrs,
library=library))
| dict_to_dataset | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(data, *, attrs=None, library=None, coords=None, dims=None):\n if dims is None:\n dims = {}\n data_vars = {}\n for key, values in data.items():\... |
def x(attrs=None, library=None):
default_attrs = {'created_at': datetime.datetime.utcnow().isoformat()}
if library is not None:
library_name = library.__name__
default_attrs['inference_library'] = library_name
try:
version = pkg_resources.get_distribution(library_name).version
default_attrs['inference_library_version'] = version
except pkg_resources.DistributionNotFound:
if hasattr(library, '__version__'):
version = library.__version__
default_attrs['inference_library_version'] = version
if attrs is not None:
default_attrs.update(attrs)
return default_attrs
| make_attrs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(attrs=None, library=None):\n default_attrs = {'created_at': datetime.datetime.utcnow().isoformat()}\n if library is not None:\n library_name = lib... |
def x(data_home=None):
if data_home is None:
data_home = os.environ.get('ARVIZ_DATA', os.path.join('~',
'arviz_data'))
data_home = os.path.expanduser(data_home)
if not os.path.exists(data_home):
os.makedirs(data_home)
return data_home
| get_data_home | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(data_home=None):\n if data_home is None:\n data_home = os.environ.get('ARVIZ_DATA', os.path.join('~',\n 'arviz_data'))\n data_home = os... |
def x(path):
sha256hash = hashlib.sha256()
chunk_size = 8192
with open(path, 'rb') as buff:
while True:
buffer = buff.read(chunk_size)
if not buffer:
break
sha256hash.update(buffer)
return sha256hash.hexdigest()
| sha256 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path):\n sha256hash = hashlib.sha256()\n chunk_size = 8192\n with open(path, 'rb') as buff:\n while True:\n buffer = buff.read(chunk... |
def x(dataset=None, data_home=None):
if dataset in LOCAL_DATASETS:
resource = LOCAL_DATASETS[dataset]
return from_netcdf(resource.filename)
elif dataset in REMOTE_DATASETS:
remote = REMOTE_DATASETS[dataset]
home_dir = get_data_home(data_home=data_home)
file_path = os.path.join(home_dir, remote.filename)
if not os.path.exists(file_path):
urlretrieve(remote.url, file_path)
checksum = _sha256(file_path)
if remote.checksum != checksum:
raise IOError(
'{} has an SHA256 checksum ({}) differing from expected ({}), file may be corrupted. Run `arviz.clear_data_home()` and try again, or please open an issue.'
.format(file_path, checksum, remote.checksum))
return from_netcdf(file_path)
else:
raise ValueError(
'Dataset {} not found! The following are available:\n{}'.format
(dataset, list_datasets()))
| load_arviz_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dataset=None, data_home=None):\n if dataset in LOCAL_DATASETS:\n resource = LOCAL_DATASETS[dataset]\n return from_netcdf(resource.filename)\n ... |
def x():
lines = []
for name, resource in itertools.chain(LOCAL_DATASETS.items(),
REMOTE_DATASETS.items()):
if isinstance(resource, LocalFileMetadata):
location = 'local: {}'.format(resource.filename)
elif isinstance(resource, RemoteFileMetadata):
location = 'remote: {}'.format(resource.url)
else:
location = 'unknown'
lines.append('{}\n{}\n{}\n{}'.format(name, '=' * len(name),
resource.description, location))
return '\n\n{}\n\n'.format(10 * '-').join(lines)
| list_datasets | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n lines = []\n for name, resource in itertools.chain(LOCAL_DATASETS.items(),\n REMOTE_DATASETS.items()):\n if isinstance(resource, Local... |
def x(sampler, var_names, arg_names):
if hasattr(sampler, 'args'):
num_vars = sampler.chain.shape[-1]
num_args = len(sampler.args)
elif hasattr(sampler, 'log_prob_fn'):
num_vars = sampler.get_chain().shape[-1]
num_args = len(sampler.log_prob_fn.args)
else:
num_vars = sampler.get_chain().shape[-1]
num_args = 0
if var_names is None:
var_names = ['var_{}'.format(idx) for idx in range(num_vars)]
if arg_names is None:
arg_names = ['arg_{}'.format(idx) for idx in range(num_args)]
if len(var_names) != num_vars:
raise ValueError(
'The sampler has {} variables, but only {} var_names were provided!'
.format(num_vars, len(var_names)))
if len(arg_names) != num_args:
raise ValueError(
'The sampler has {} args, but only {} arg_names were provided!'
.format(num_args, len(arg_names)))
return var_names, arg_names
| verify_names | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sampler, var_names, arg_names):\n if hasattr(sampler, 'args'):\n num_vars = sampler.chain.shape[-1]\n num_args = len(sampler.args)\n elif h... |
def x(sampler=None, *, var_names=None, arg_names=None, coords=None, dims=None):
return EmceeConverter(sampler=sampler, var_names=var_names, arg_names=
arg_names, coords=coords, dims=dims).to_inference_data()
| from_emcee | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sampler=None, *, var_names=None, arg_names=None, coords=None, dims=None):\n return EmceeConverter(sampler=sampler, var_names=var_names, arg_names=\n ... |
def x(self):
data = {}
for idx, var_name in enumerate(self.var_names):
data[var_name] = self.sampler.get_chain()[..., idx].T if hasattr(self
.sampler, 'get_chain') else self.sampler.chain[..., idx]
return dict_to_dataset(data, library=self.emcee, coords=self.coords,
dims=self.dims)
| posterior_to_xarray | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n data = {}\n for idx, var_name in enumerate(self.var_names):\n data[var_name] = self.sampler.get_chain()[..., idx].T if hasattr(self\n ... |
def x(self):
data = {}
for idx, var_name in enumerate(self.arg_names):
data[var_name] = self.sampler.log_prob_fn.args[idx] if hasattr(self
.sampler, 'log_prob_fn') else self.sampler.args[idx]
return dict_to_dataset(data, library=self.emcee, coords=self.coords,
dims=self.dims)
| observed_data_to_xarray | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n data = {}\n for idx, var_name in enumerate(self.arg_names):\n data[var_name] = self.sampler.log_prob_fn.args[idx] if hasattr(self\n ... |
def x(posterior=None, *, var_names=None, model_fn=None, feed_dict=None,
posterior_predictive_samples=100, posterior_predictive_size=1, observed
=None, coords=None, dims=None):
return TfpConverter(posterior=posterior, var_names=var_names, model_fn=
model_fn, feed_dict=feed_dict, posterior_predictive_samples=
posterior_predictive_samples, posterior_predictive_size=
posterior_predictive_size, observed=observed, coords=coords, dims=dims
).to_inference_data()
| from_tfp | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(posterior=None, *, var_names=None, model_fn=None, feed_dict=None,\n posterior_predictive_samples=100, posterior_predictive_size=1, observed\n =None, coor... |
def x(self):
data = {}
for i, var_name in enumerate(self.var_names):
data[var_name] = np.expand_dims(self.posterior[i], axis=0)
return dict_to_dataset(data, library=self.tfp, coords=self.coords, dims
=self.dims)
| posterior_to_xarray | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n data = {}\n for i, var_name in enumerate(self.var_names):\n data[var_name] = np.expand_dims(self.posterior[i], axis=0)\n return dict_t... |
def x(self):
if self.observed is None:
return None
observed_data = {}
if isinstance(self.observed, self.tf.Tensor):
with self.tf.Session() as sess:
vals = sess.run(self.observed, feed_dict=self.feed_dict)
else:
vals = self.observed
if self.dims is None:
dims = {}
else:
dims = self.dims
name = 'obs'
val_dims = dims.get(name)
vals = np.atleast_1d(vals)
val_dims, coords = generate_dims_coords(vals.shape, name, dims=val_dims,
coords=self.coords)
observed_data[name] = xr.DataArray(vals, dims=val_dims, coords=coords)
return xr.Dataset(data_vars=observed_data, attrs=make_attrs(library=
self.tfp))
| observed_data_to_xarray | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.observed is None:\n return None\n observed_data = {}\n if isinstance(self.observed, self.tf.Tensor):\n with self.tf.Ses... |
def x(self):
if self.model_fn is None:
return None
posterior_preds = []
sample_size = self.posterior[0].shape[0]
for i in np.arange(0, sample_size, int(sample_size / self.
posterior_predictive_samples)):
variables = {}
for var_i, var_name in enumerate(self.var_names):
variables[var_name] = self.posterior[var_i][i]
with self.ed.interception(self._value_setter(variables)):
if self.posterior_predictive_size > 1:
posterior_preds.append([self.model_fn() for _ in range(self
.posterior_predictive_size)])
else:
posterior_preds.append(self.model_fn())
data = {}
with self.tf.Session() as sess:
data['obs'] = np.expand_dims(sess.run(posterior_preds, feed_dict=
self.feed_dict), axis=0)
return dict_to_dataset(data, library=self.tfp, coords=self.coords, dims
=self.dims)
| posterior_predictive_to_xarray | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.model_fn is None:\n return None\n posterior_preds = []\n sample_size = self.posterior[0].shape[0]\n for i in np.arange(0, s... |
def x(self):
if self.model_fn is None or self.observed is None:
return None
log_likelihood = []
sample_size = self.posterior[0].shape[0]
for i in range(sample_size):
variables = {}
for var_i, var_name in enumerate(self.var_names):
variables[var_name] = self.posterior[var_i][i]
with self.ed.interception(self._value_setter(variables)):
log_likelihood.append(self.model_fn().distribution.log_prob(
self.observed))
data = {}
if self.dims is not None:
coord_name = self.dims.get('obs')
else:
coord_name = None
dims = {'log_likelihood': coord_name}
with self.tf.Session() as sess:
data['log_likelihood'] = np.expand_dims(sess.run(log_likelihood,
feed_dict=self.feed_dict), axis=0)
return dict_to_dataset(data, library=self.tfp, coords=self.coords, dims
=dims)
| sample_stats_to_xarray | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.model_fn is None or self.observed is None:\n return None\n log_likelihood = []\n sample_size = self.posterior[0].shape[0]\n ... |
def x(self):
return InferenceData(**{'posterior': self.posterior_to_xarray(),
'sample_stats': self.sample_stats_to_xarray(),
'posterior_predictive': self.posterior_predictive_to_xarray(),
'observed_data': self.observed_data_to_xarray()})
| to_inference_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return InferenceData(**{'posterior': self.posterior_to_xarray(),\n 'sample_stats': self.sample_stats_to_xarray(),\n 'posterior_predic... |
def x(trace=None, *, prior=None, posterior_predictive=None, coords=None,
dims=None):
return PyMC3Converter(trace=trace, prior=prior, posterior_predictive=
posterior_predictive, coords=coords, dims=dims).to_inference_data()
| from_pymc3 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(trace=None, *, prior=None, posterior_predictive=None, coords=None,\n dims=None):\n return PyMC3Converter(trace=trace, prior=prior, posterior_predictive=\... |
def x(self):
model = self.trace._straces[0].model
if len(model.observed_RVs) != 1:
return None, None
elif self.dims is not None:
coord_name = self.dims.get(model.observed_RVs[0].name)
else:
coord_name = None
cached = [(var, var.logp_elemwise) for var in model.observed_RVs]
def x(point):
log_like_vals = []
for var, log_like in cached:
log_like_val = log_like(point)
if var.missing_values:
log_like_val = log_like_val[~var.observations.mask]
log_like_vals.append(log_like_val.ravel())
return np.concatenate(log_like_vals)
chain_likelihoods = []
for chain in self.trace.chains:
log_like = (log_likelihood_vals_point(point) for point in self.
trace.points([chain]))
chain_likelihoods.append(np.stack(log_like))
return np.stack(chain_likelihoods), coord_name
| extract_log_likelihood | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n model = self.trace._straces[0].model\n if len(model.observed_RVs) != 1:\n return None, None\n elif self.dims is not None:\n coo... |
def x(self):
var_names = self.pymc3.util.get_default_varnames(self.trace.varnames,
include_transformed=False)
data = {}
for var_name in var_names:
data[var_name] = np.array(self.trace.get_values(var_name, combine=
False, squeeze=False))
return dict_to_dataset(data, library=self.pymc3, coords=self.coords,
dims=self.dims)
| posterior_to_xarray | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n var_names = self.pymc3.util.get_default_varnames(self.trace.varnames,\n include_transformed=False)\n data = {}\n for var_name in var_n... |
def x(self):
rename_key = {'model_logp': 'lp'}
data = {}
for stat in self.trace.stat_names:
name = rename_key.get(stat, stat)
data[name] = np.array(self.trace.get_sampler_stats(stat, combine=False)
)
log_likelihood, dims = self._extract_log_likelihood()
if log_likelihood is not None:
data['log_likelihood'] = log_likelihood
dims = {'log_likelihood': dims}
else:
dims = None
return dict_to_dataset(data, library=self.pymc3, dims=dims, coords=self
.coords)
| sample_stats_to_xarray | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n rename_key = {'model_logp': 'lp'}\n data = {}\n for stat in self.trace.stat_names:\n name = rename_key.get(stat, stat)\n data[n... |
def x(self):
data = {k: np.expand_dims(v, 0) for k, v in self.posterior_predictive.
items()}
return dict_to_dataset(data, library=self.pymc3, coords=self.coords,
dims=self.dims)
| posterior_predictive_to_xarray | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n data = {k: np.expand_dims(v, 0) for k, v in self.posterior_predictive.\n items()}\n return dict_to_dataset(data, library=self.pymc3, coor... |
def x(self):
return dict_to_dataset({k: np.expand_dims(v, 0) for k, v in self.prior.
items()}, library=self.pymc3, coords=self.coords, dims=self.dims)
| prior_to_xarray | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return dict_to_dataset({k: np.expand_dims(v, 0) for k, v in self.prior.\n items()}, library=self.pymc3, coords=self.coords, dims=self.dims)\... |
def x(self):
model = self.trace._straces[0].model
observations = {obs.name: obs.observations for obs in model.observed_RVs}
if self.dims is None:
dims = {}
else:
dims = self.dims
observed_data = {}
for name, vals in observations.items():
vals = np.atleast_1d(vals)
val_dims = dims.get(name)
val_dims, coords = generate_dims_coords(vals.shape, name, dims=
val_dims, coords=self.coords)
coords = {key: xr.IndexVariable((key,), data=coords[key]) for key in
val_dims}
observed_data[name] = xr.DataArray(vals, dims=val_dims, coords=coords)
return xr.Dataset(data_vars=observed_data, attrs=make_attrs(library=
self.pymc3))
| observed_data_to_xarray | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n model = self.trace._straces[0].model\n observations = {obs.name: obs.observations for obs in model.observed_RVs}\n if self.dims is None:\n ... |
def x(data, var_names=None, coords=None, figsize=None, textsize=None,
credible_interval=0.94, round_to=1, point_estimate='mean', rope=None,
ref_val=None, kind='kde', bw=4.5, bins=None, ax=None, **kwargs):
data = convert_to_dataset(data, group='posterior')
var_names = _var_names(var_names, data)
if coords is None:
coords = {}
plotters = list(xarray_var_iter(get_coords(data, coords), var_names=
var_names, combined=True))
length_plotters = len(plotters)
rows, cols = default_grid(length_plotters)
figsize, ax_labelsize, titlesize, xt_labelsize, _linewidth, _ = (
_scale_fig_size(figsize, textsize, rows, cols))
kwargs.setdefault('linewidth', _linewidth)
if ax is None:
_, ax = _create_axes_grid(length_plotters, rows, cols, figsize=
figsize, squeeze=False, constrained_layout=True)
for (var_name, selection, x), ax_ in zip(plotters, np.ravel(ax)):
_plot_posterior_op(x.flatten(), var_name, selection, ax=ax_, bw=bw,
bins=bins, kind=kind, point_estimate=point_estimate, round_to=
round_to, credible_interval=credible_interval, ref_val=ref_val,
rope=rope, ax_labelsize=ax_labelsize, xt_labelsize=xt_labelsize,
**kwargs)
ax_.set_title(make_label(var_name, selection), fontsize=titlesize,
wrap=True)
return ax
| plot_posterior | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(data, var_names=None, coords=None, figsize=None, textsize=None,\n credible_interval=0.94, round_to=1, point_estimate='mean', rope=None,\n ref_val=None, k... |
def x(energy):
energy_mat = np.atleast_2d(energy)
num = np.square(np.diff(energy_mat, axis=1)).mean(axis=1)
den = np.var(energy_mat, axis=1)
return num / den
| bfmi | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(energy):\n energy_mat = np.atleast_2d(energy)\n num = np.square(np.diff(energy_mat, axis=1)).mean(axis=1)\n den = np.var(energy_mat, axis=1)\n retu... |
def x(dataset_dict, ic='waic', method='BB-pseudo-BMA', b_samples=1000,
alpha=1, seed=None, scale='deviance'):
names = list(dataset_dict.keys())
scale = scale.lower()
if scale == 'log':
scale_value = 1
ascending = False
else:
if scale == 'negative_log':
scale_value = -1
else:
scale_value = -2
ascending = True
if ic == 'waic':
ic_func = waic
df_comp = pd.DataFrame(index=names, columns=['waic', 'p_waic',
'd_waic', 'weight', 'se', 'dse', 'warning', 'waic_scale'])
scale_col = 'waic_scale'
elif ic == 'loo':
ic_func = loo
df_comp = pd.DataFrame(index=names, columns=['loo', 'p_loo',
'd_loo', 'weight', 'se', 'dse', 'warning', 'loo_scale'])
scale_col = 'loo_scale'
else:
raise NotImplementedError(
'The information criterion {} is not supported.'.format(ic))
if method.lower() not in ['stacking', 'bb-pseudo-bma', 'pseudo-bma']:
raise ValueError('The method {}, to compute weights, is not supported.'
.format(method))
ic_se = '{}_se'.format(ic)
p_ic = 'p_{}'.format(ic)
ic_i = '{}_i'.format(ic)
ics = pd.DataFrame()
names = []
for name, dataset in dataset_dict.items():
names.append(name)
ics = ics.append([ic_func(dataset, pointwise=True, scale=scale)])
ics.index = names
ics.sort_values(by=ic, inplace=True, ascending=ascending)
if method.lower() == 'stacking':
rows, cols, ic_i_val = _ic_matrix(ics, ic_i)
exp_ic_i = np.exp(ic_i_val / scale_value)
last_col = cols - 1
def x(weights):
return np.concatenate((weights, [max(1.0 - np.sum(weights), 0.0)]))
def x(weights):
w_full = w_fuller(weights)
score = 0.0
for i in range(rows):
score += np.log(np.dot(exp_ic_i[i], w_full))
return -score
def x(weights):
w_full = w_fuller(weights)
grad = np.zeros(last_col)
for k in range(last_col - 1):
for i in range(rows):
grad[k] += (exp_ic_i[i, k] - exp_ic_i[i, last_col]
) / np.dot(exp_ic_i[i], w_full)
return -grad
theta = np.full(last_col, 1.0 / cols)
bounds = [(0.0, 1.0) for _ in range(last_col)]
constraints = [{'type': 'ineq', 'fun': lambda x: 1.0 - np.sum(x)},
{'type': 'ineq', 'fun': np.sum}]
weights = minimize(fun=log_score, x0=theta, jac=gradient, bounds=
bounds, constraints=constraints)
weights = w_fuller(weights['x'])
ses = ics[ic_se]
elif method.lower() == 'bb-pseudo-bma':
rows, cols, ic_i_val = _ic_matrix(ics, ic_i)
ic_i_val = ic_i_val * rows
b_weighting = st.dirichlet.rvs(alpha=[alpha] * rows, size=b_samples,
random_state=seed)
weights = np.zeros((b_samples, cols))
z_bs = np.zeros_like(weights)
for i in range(b_samples):
z_b = np.dot(b_weighting[i], ic_i_val)
u_weights = np.exp((z_b - np.min(z_b)) / scale_value)
z_bs[i] = z_b
weights[i] = u_weights / np.sum(u_weights)
weights = weights.mean(axis=0)
ses = pd.Series(z_bs.std(axis=0), index=names)
elif method.lower() == 'pseudo-bma':
min_ic = ics.iloc[0][ic]
z_rv = np.exp((ics[ic] - min_ic) / scale_value)
weights = z_rv / np.sum(z_rv)
ses = ics[ic_se]
if np.any(weights):
min_ic_i_val = ics[ic_i].iloc[0]
for idx, val in enumerate(ics.index):
res = ics.loc[val]
if scale_value < 0:
diff = res[ic_i] - min_ic_i_val
else:
diff = min_ic_i_val - res[ic_i]
d_ic = np.sum(diff)
d_std_err = np.sqrt(len(diff) * np.var(diff))
std_err = ses.loc[val]
weight = weights[idx]
df_comp.at[val] = res[ic], res[p_ic
], d_ic, weight, std_err, d_std_err, res['warning'], res[
scale_col]
return df_comp.sort_values(by=ic, ascending=ascending)
| compare | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dataset_dict, ic='waic', method='BB-pseudo-BMA', b_samples=1000,\n alpha=1, seed=None, scale='deviance'):\n names = list(dataset_dict.keys())\n scale ... |
def x(ics, ic_i):
cols, _ = ics.shape
rows = len(ics[ic_i].iloc[0])
ic_i_val = np.zeros((rows, cols))
for idx, val in enumerate(ics.index):
ic = ics.loc[val][ic_i]
if len(ic) != rows:
raise ValueError(
'The number of observations should be the same across all models'
)
ic_i_val[:, idx] = ic
return rows, cols, ic_i_val
| ic_matrix | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ics, ic_i):\n cols, _ = ics.shape\n rows = len(ics[ic_i].iloc[0])\n ic_i_val = np.zeros((rows, cols))\n for idx, val in enumerate(ics.index):\n ... |
def x(x, credible_interval=0.94, circular=False):
if x.ndim > 1:
hpd_array = np.array([hpd(row, credible_interval=credible_interval,
circular=circular) for row in x.T])
return hpd_array
x = x.copy()
len_x = len(x)
if circular:
mean = st.circmean(x, high=np.pi, low=-np.pi)
x = x - mean
x = np.arctan2(np.sin(x), np.cos(x))
x = np.sort(x)
interval_idx_inc = int(np.floor(credible_interval * len_x))
n_intervals = len_x - interval_idx_inc
interval_width = x[interval_idx_inc:] - x[:n_intervals]
if len(interval_width) == 0:
raise ValueError(
'Too few elements for interval calculation. Check that credible_interval meets condition 0 =< credible_interval < 1'
)
min_idx = np.argmin(interval_width)
hdi_min = x[min_idx]
hdi_max = x[min_idx + interval_idx_inc]
if circular:
hdi_min = hdi_min + mean
hdi_max = hdi_max + mean
hdi_min = np.arctan2(np.sin(hdi_min), np.cos(hdi_min))
hdi_max = np.arctan2(np.sin(hdi_max), np.cos(hdi_max))
return np.array([hdi_min, hdi_max])
| hpd | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(x, credible_interval=0.94, circular=False):\n if x.ndim > 1:\n hpd_array = np.array([hpd(row, credible_interval=credible_interval,\n circu... |
def x(ary, *, b=None, b_inv=None, axis=None, keepdims=False, out=None, copy
=True):
ary = np.asarray(ary)
if ary.dtype.kind == 'i':
ary = ary.astype(np.float64)
dtype = ary.dtype.type
shape = ary.shape
shape_len = len(shape)
if isinstance(axis, Sequence):
axis = tuple(axis_i if axis_i >= 0 else shape_len + axis_i for
axis_i in axis)
agroup = axis
else:
axis = axis if axis is None or axis >= 0 else shape_len + axis
agroup = axis,
shape_max = tuple(1 for _ in shape) if axis is None else tuple(1 if i in
agroup else d for i, d in enumerate(shape))
if out is None:
if not keepdims:
out_shape = tuple() if axis is None else tuple(d for i, d in
enumerate(shape) if i not in agroup)
else:
out_shape = shape_max
out = np.empty(out_shape, dtype=dtype)
if b_inv == 0:
return np.full_like(out, np.inf, dtype=dtype) if out.shape else np.inf
if b_inv is None and b == 0:
return np.full_like(out, -np.inf) if out.shape else -np.inf
ary_max = np.empty(shape_max, dtype=dtype)
ary.max(axis=axis, keepdims=True, out=ary_max)
if copy:
ary = ary.copy()
ary -= ary_max
np.exp(ary, out=ary)
ary.sum(axis=axis, keepdims=keepdims, out=out)
np.log(out, out=out)
if b_inv is not None:
ary_max -= np.log(b_inv)
elif b:
ary_max += np.log(b)
out += ary_max.squeeze() if not keepdims else ary_max
return out if out.shape else dtype(out)
| logsumexp | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ary, *, b=None, b_inv=None, axis=None, keepdims=False, out=None, copy\n =True):\n ary = np.asarray(ary)\n if ary.dtype.kind == 'i':\n ary = ary... |
def x(data, pointwise=False, reff=None, scale='deviance'):
inference_data = convert_to_inference_data(data)
for group in ('posterior', 'sample_stats'):
if not hasattr(inference_data, group):
raise TypeError('Must be able to extract a {group}group from data!'
.format(group=group))
if 'log_likelihood' not in inference_data.sample_stats:
raise TypeError('Data must include log_likelihood in sample_stats')
posterior = inference_data.posterior
log_likelihood = inference_data.sample_stats.log_likelihood
n_samples = log_likelihood.chain.size * log_likelihood.draw.size
new_shape = (n_samples,) + log_likelihood.shape[2:]
log_likelihood = log_likelihood.values.reshape(*new_shape)
if scale.lower() == 'deviance':
scale_value = -2
elif scale.lower() == 'log':
scale_value = 1
elif scale.lower() == 'negative_log':
scale_value = -1
else:
raise TypeError(
'Valid scale values are "deviance", "log", "negative_log"')
if reff is None:
n_chains = len(posterior.chain)
if n_chains == 1:
reff = 1.0
else:
ess = effective_sample_size(posterior)
reff = np.hstack([ess[v].values.flatten() for v in ess.data_vars]
).mean() / n_samples
log_weights, pareto_shape = psislw(-log_likelihood, reff)
log_weights += log_likelihood
warn_mg = 0
if np.any(pareto_shape > 0.7):
warnings.warn(
"""Estimated shape parameter of Pareto distribution is greater than 0.7 for
one or more samples. You should consider using a more robust model, this is because
importance sampling is less likely to work well if the marginal posterior and LOO posterior
are very different. This is more likely to happen with a non-robust model and highly
influential observations."""
)
warn_mg = 1
loo_lppd_i = scale_value * _logsumexp(log_weights, axis=0)
loo_lppd = loo_lppd_i.sum()
loo_lppd_se = (len(loo_lppd_i) * np.var(loo_lppd_i)) ** 0.5
lppd = np.sum(_logsumexp(log_likelihood, axis=0, b_inv=log_likelihood.
shape[0]))
p_loo = lppd - loo_lppd / scale_value
if pointwise:
if np.equal(loo_lppd, loo_lppd_i).all():
warnings.warn(
"""The point-wise LOO is the same with the sum LOO, please double check
the Observed RV in your model to make sure it returns element-wise logp.
"""
)
return pd.Series(data=[loo_lppd, loo_lppd_se, p_loo, warn_mg,
loo_lppd_i, pareto_shape, scale], index=['loo', 'loo_se',
'p_loo', 'warning', 'loo_i', 'pareto_k', 'loo_scale'])
else:
return pd.Series(data=[loo_lppd, loo_lppd_se, p_loo, warn_mg, scale
], index=['loo', 'loo_se', 'p_loo', 'warning', 'loo_scale'])
| loo | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(data, pointwise=False, reff=None, scale='deviance'):\n inference_data = convert_to_inference_data(data)\n for group in ('posterior', 'sample_stats'):\n ... |
def x(log_weights, reff=1.0):
rows, cols = log_weights.shape
log_weights_out = np.copy(log_weights, order='F')
kss = np.empty(cols)
cutoff_ind = -int(np.ceil(min(rows / 5.0, 3 * (rows / reff) ** 0.5))) - 1
cutoffmin = np.log(np.finfo(float).tiny)
k_min = 1.0 / 3
for i, x in enumerate(log_weights_out.T):
x -= np.max(x)
x_sort_ind = np.argsort(x)
xcutoff = max(x[x_sort_ind[cutoff_ind]], cutoffmin)
expxcutoff = np.exp(xcutoff)
tailinds, = np.where(x > xcutoff)
x_tail = x[tailinds]
tail_len = len(x_tail)
if tail_len <= 4:
k = np.inf
else:
x_tail_si = np.argsort(x_tail)
x_tail = np.exp(x_tail) - expxcutoff
k, sigma = _gpdfit(x_tail[x_tail_si])
if k >= k_min:
sti = np.arange(0.5, tail_len) / tail_len
smoothed_tail = _gpinv(sti, k, sigma)
smoothed_tail = np.log(smoothed_tail + expxcutoff)
x[tailinds[x_tail_si]] = smoothed_tail
x[x > 0] = 0
x -= _logsumexp(x)
kss[i] = k
return log_weights_out, kss
| psislw | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(log_weights, reff=1.0):\n rows, cols = log_weights.shape\n log_weights_out = np.copy(log_weights, order='F')\n kss = np.empty(cols)\n cutoff_ind = ... |
def x(x):
prior_bs = 3
prior_k = 10
len_x = len(x)
m_est = 30 + int(len_x ** 0.5)
b_ary = 1 - np.sqrt(m_est / (np.arange(1, m_est + 1, dtype=float) - 0.5))
b_ary /= prior_bs * x[int(len_x / 4 + 0.5) - 1]
b_ary += 1 / x[-1]
k_ary = np.log1p(-b_ary[:, None] * x).mean(axis=1)
len_scale = len_x * (np.log(-(b_ary / k_ary)) - k_ary - 1)
weights = 1 / np.exp(len_scale - len_scale[:, None]).sum(axis=1)
real_idxs = weights >= 10 * np.finfo(float).eps
if not np.all(real_idxs):
weights = weights[real_idxs]
b_ary = b_ary[real_idxs]
weights /= weights.sum()
b_post = np.sum(b_ary * weights)
k_post = np.log1p(-b_post * x).mean()
k_post = (len_x * k_post + prior_k * 0.5) / (len_x + prior_k)
sigma = -k_post / b_post
return k_post, sigma
| gpdfit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(x):\n prior_bs = 3\n prior_k = 10\n len_x = len(x)\n m_est = 30 + int(len_x ** 0.5)\n b_ary = 1 - np.sqrt(m_est / (np.arange(1, m_est + 1, dtype... |
def x(probs, kappa, sigma):
x = np.full_like(probs, np.nan)
if sigma <= 0:
return x
ok = (probs > 0) & (probs < 1)
if np.all(ok):
if np.abs(kappa) < np.finfo(float).eps:
x = -np.log1p(-probs)
else:
x = np.expm1(-kappa * np.log1p(-probs)) / kappa
x *= sigma
else:
if np.abs(kappa) < np.finfo(float).eps:
x[ok] = -np.log1p(-probs[ok])
else:
x[ok] = np.expm1(-kappa * np.log1p(-probs[ok])) / kappa
x *= sigma
x[probs == 0] = 0
if kappa >= 0:
x[probs == 1] = np.inf
else:
x[probs == 1] = -sigma / kappa
return x
| gpinv | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(probs, kappa, sigma):\n x = np.full_like(probs, np.nan)\n if sigma <= 0:\n return x\n ok = (probs > 0) & (probs < 1)\n if np.all(ok):\n ... |
def x(y_true, y_pred):
if y_pred.ndim == 1:
var_y_est = np.var(y_pred)
var_e = np.var(y_true - y_pred)
else:
var_y_est = np.var(y_pred.mean(0))
var_e = np.var(y_true - y_pred, 0)
r_squared = var_y_est / (var_y_est + var_e)
return pd.Series([np.mean(r_squared), np.std(r_squared)], index=['r2',
'r2_std'])
| r2_score | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(y_true, y_pred):\n if y_pred.ndim == 1:\n var_y_est = np.var(y_pred)\n var_e = np.var(y_true - y_pred)\n else:\n var_y_est = np.var(... |
def x(data, var_names=None, fmt='wide', round_to=2, include_circ=None,
stat_funcs=None, extend=True, credible_interval=0.94, order='C'):
posterior = convert_to_dataset(data, group='posterior')
var_names = _var_names(var_names, posterior)
posterior = posterior if var_names is None else posterior[var_names]
fmt_group = 'wide', 'long', 'xarray'
if not isinstance(fmt, str) or fmt.lower() not in fmt_group:
raise TypeError("Invalid format: '{}'! Formatting options are: {}".
format(fmt, fmt_group))
unpack_order_group = 'C', 'F'
if not isinstance(order, str) or order.upper() not in unpack_order_group:
raise TypeError("Invalid order: '{}'! Unpacking options are: {}".
format(order, unpack_order_group))
alpha = 1 - credible_interval
metrics = []
metric_names = []
if stat_funcs is not None:
if isinstance(stat_funcs, dict):
for stat_func_name, stat_func in stat_funcs.items():
metrics.append(xr.apply_ufunc(_make_ufunc(stat_func),
posterior, input_core_dims=(('chain', 'draw'),)))
metric_names.append(stat_func_name)
else:
for stat_func in stat_funcs:
metrics.append(xr.apply_ufunc(_make_ufunc(stat_func),
posterior, input_core_dims=(('chain', 'draw'),)))
metric_names.append(stat_func.__name__)
if extend:
metrics.append(posterior.mean(dim=('chain', 'draw')))
metric_names.append('mean')
metrics.append(posterior.std(dim=('chain', 'draw')))
metric_names.append('sd')
metrics.append(xr.apply_ufunc(_make_ufunc(_mc_error), posterior,
input_core_dims=(('chain', 'draw'),)))
metric_names.append('mc error')
metrics.append(xr.apply_ufunc(_make_ufunc(hpd, index=0,
credible_interval=credible_interval), posterior,
input_core_dims=(('chain', 'draw'),)))
metric_names.append('hpd {:g}%'.format(100 * alpha / 2))
metrics.append(xr.apply_ufunc(_make_ufunc(hpd, index=1,
credible_interval=credible_interval), posterior,
input_core_dims=(('chain', 'draw'),)))
metric_names.append('hpd {:g}%'.format(100 * (1 - alpha / 2)))
if include_circ:
metrics.append(xr.apply_ufunc(_make_ufunc(st.circmean, high=np.pi,
low=-np.pi), posterior, input_core_dims=(('chain', 'draw'),)))
metric_names.append('circular mean')
metrics.append(xr.apply_ufunc(_make_ufunc(st.circstd, high=np.pi,
low=-np.pi), posterior, input_core_dims=(('chain', 'draw'),)))
metric_names.append('circular standard deviation')
metrics.append(xr.apply_ufunc(_make_ufunc(_mc_error, circular=True),
posterior, input_core_dims=(('chain', 'draw'),)))
metric_names.append('circular mc error')
metrics.append(xr.apply_ufunc(_make_ufunc(hpd, index=0,
credible_interval=credible_interval, circular=True), posterior,
input_core_dims=(('chain', 'draw'),)))
metric_names.append('circular hpd {:.2%}'.format(alpha / 2))
metrics.append(xr.apply_ufunc(_make_ufunc(hpd, index=1,
credible_interval=credible_interval, circular=True), posterior,
input_core_dims=(('chain', 'draw'),)))
metric_names.append('circular hpd {:.2%}'.format(1 - alpha / 2))
if len(posterior.chain) > 1:
metrics.append(effective_sample_size(posterior, var_names=var_names))
metric_names.append('ess')
metrics.append(rhat(posterior, var_names=var_names))
metric_names.append('r_hat')
joined = xr.concat(metrics, dim='metric').assign_coords(metric=metric_names
)
if fmt.lower() == 'wide':
dfs = []
for var_name, values in joined.data_vars.items():
if len(values.shape[1:]):
metric = list(values.metric.values)
data_dict = OrderedDict()
for idx in np.ndindex(values.shape[1:] if order == 'C' else
values.shape[1:][::-1]):
if order == 'F':
idx = tuple(idx[::-1])
ser = pd.Series(values[Ellipsis, *idx].values, index=metric
)
key = '{}[{}]'.format(var_name, ','.join(map(str, idx)))
data_dict[key] = ser
df = pd.DataFrame.from_dict(data_dict, orient='index')
df = df.loc[list(data_dict.keys())]
else:
df = values.to_dataframe()
df.index = list(df.index)
df = df.T
dfs.append(df)
summary_df = pd.concat(dfs, sort=False)
elif fmt.lower() == 'long':
df = joined.to_dataframe().reset_index().set_index('metric')
df.index = list(df.index)
summary_df = df
else:
summary_df = joined
return summary_df.round(round_to)
| summary | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(data, var_names=None, fmt='wide', round_to=2, include_circ=None,\n stat_funcs=None, extend=True, credible_interval=0.94, order='C'):\n posterior = conver... |
def x(func, index=Ellipsis, **kwargs):
def x(ary):
target = np.empty(ary.shape[:-2])
for idx in np.ndindex(target.shape):
target[idx] = np.asarray(func(ary[idx].ravel(), **kwargs))[index]
return target
return _ufunc
| make_ufunc | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(func, index=Ellipsis, **kwargs):\n\n def x(ary):\n target = np.empty(ary.shape[:-2])\n for idx in np.ndindex(target.shape):\n targe... |
def x(x, batches=5, circular=False):
if x.ndim > 1:
dims = np.shape(x)
trace = np.transpose([t.ravel() for t in x])
return np.reshape([_mc_error(t, batches) for t in trace], dims[1:])
else:
if batches == 1:
if circular:
std = st.circstd(x, high=np.pi, low=-np.pi)
else:
std = np.std(x)
return std / np.sqrt(len(x))
batched_traces = np.resize(x, (batches, int(len(x) / batches)))
if circular:
means = st.circmean(batched_traces, high=np.pi, low=-np.pi, axis=1)
std = st.circstd(means, high=np.pi, low=-np.pi)
else:
means = np.mean(batched_traces, 1)
std = np.std(means)
return std / np.sqrt(batches)
| mc_error | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(x, batches=5, circular=False):\n if x.ndim > 1:\n dims = np.shape(x)\n trace = np.transpose([t.ravel() for t in x])\n return np.reshape... |
def x(data, pointwise=False, scale='deviance'):
inference_data = convert_to_inference_data(data)
for group in ('sample_stats',):
if not hasattr(inference_data, group):
raise TypeError(
'Must be able to extract a {group} group from data!'.format
(group=group))
if 'log_likelihood' not in inference_data.sample_stats:
raise TypeError('Data must include log_likelihood in sample_stats')
log_likelihood = inference_data.sample_stats.log_likelihood
if scale.lower() == 'deviance':
scale_value = -2
elif scale.lower() == 'log':
scale_value = 1
elif scale.lower() == 'negative_log':
scale_value = -1
else:
raise TypeError(
'Valid scale values are "deviance", "log", "negative_log"')
n_samples = log_likelihood.chain.size * log_likelihood.draw.size
new_shape = (n_samples,) + log_likelihood.shape[2:]
log_likelihood = log_likelihood.values.reshape(*new_shape)
lppd_i = _logsumexp(log_likelihood, axis=0, b_inv=log_likelihood.shape[0])
vars_lpd = np.var(log_likelihood, axis=0)
warn_mg = 0
if np.any(vars_lpd > 0.4):
warnings.warn(
"""For one or more samples the posterior variance of the log predictive
densities exceeds 0.4. This could be indication of WAIC starting to fail see
http://arxiv.org/abs/1507.04544 for details
"""
)
warn_mg = 1
waic_i = scale_value * (lppd_i - vars_lpd)
waic_se = (len(waic_i) * np.var(waic_i)) ** 0.5
waic_sum = np.sum(waic_i)
p_waic = np.sum(vars_lpd)
if pointwise:
if np.equal(waic_sum, waic_i).all():
warnings.warn(
"""The point-wise WAIC is the same with the sum WAIC, please double check
the Observed RV in your model to make sure it returns element-wise logp.
"""
)
return pd.Series(data=[waic_sum, waic_se, p_waic, warn_mg, waic_i,
scale], index=['waic', 'waic_se', 'p_waic', 'warning', 'waic_i',
'waic_scale'])
else:
return pd.Series(data=[waic_sum, waic_se, p_waic, warn_mg, scale],
index=['waic', 'waic_se', 'p_waic', 'warning', 'waic_scale'])
| waic | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(data, pointwise=False, scale='deviance'):\n inference_data = convert_to_inference_data(data)\n for group in ('sample_stats',):\n if not hasattr(in... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.