source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(self, fid): lin = self.read_line(fid) while lin[0] != ':': parts = lin.split() if parts[0] == 'order': order = [] for i in range(1, len(parts)): if parts[i].lower() == 'rx': chan = 'Xrotation' order.append('x') elif parts[i].lower() == 'ry': chan = 'Yrotation' order.append('y') elif parts[i].lower() == 'rz': chan = 'Zrotation' order.append('z') elif parts[i].lower() == 'tx': chan = 'Xposition' elif parts[i].lower() == 'ty': chan = 'Yposition' elif parts[i].lower() == 'tz': chan = 'Zposition' elif parts[i].lower() == 'l': chan = 'length' self.vertices[0].meta['channels'].append(chan) self.vertices[0].meta['order'] = order[::-1] elif parts[0] == 'axis': self.vertices[0].meta['axis_order'] = parts[1][::-1].lower() elif parts[0] == 'position': self.vertices[0].meta['offset'] = [float(parts[1]), float(parts [2]), float(parts[3])] elif parts[0] == 'orientation': self.vertices[0].meta['orientation'] = [float(parts[1]), float( parts[2]), float(parts[3])] lin = self.read_line(fid) return lin
read_root
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, fid):\n lin = self.read_line(fid)\n while lin[0] != ':':\n parts = lin.split()\n if parts[0] == 'order':\n order = []\n ...
def x(self, fid): lin = self.read_line(fid) while lin: if lin[0] == ':': if lin[1:] == 'name': lin = self.read_line(fid) self.name = lin elif lin[1:] == 'units': lin = self.read_units(fid) elif lin[1:] == 'documentation': lin = self.read_documentation(fid) elif lin[1:] == 'root': lin = self.read_root(fid) elif lin[1:] == 'bonedata': lin = self.read_bonedata(fid) elif lin[1:] == 'hierarchy': lin = self.read_hierarchy(fid) elif lin[1:8] == 'version': lin = self.read_line(fid) continue else: if not lin: self.finalize() return lin = self.read_line(fid) else: raise ValueError('Unrecognised file format') self.finalize()
read_skel
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, fid):\n lin = self.read_line(fid)\n while lin:\n if lin[0] == ':':\n if lin[1:] == 'name':\n lin = self.read_line(...
def x(self, fid): lin = self.read_line(fid) while lin[0] != ':': parts = lin.split() if parts[0] == 'mass': self.mass = float(parts[1]) elif parts[0] == 'length': self.length = float(parts[1]) elif parts[0] == 'angle': self.angle = parts[1] lin = self.read_line(fid) return lin
read_units
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, fid):\n lin = self.read_line(fid)\n while lin[0] != ':':\n parts = lin.split()\n if parts[0] == 'mass':\n self.mass = floa...
def x(self, index, start_val): channels = self.vertices[index].meta['channels'] base_channel = start_val rot_ind = -np.ones(3, dtype=int) pos_ind = -np.ones(3, dtype=int) for i in range(len(channels)): if channels[i] == 'Xrotation': rot_ind[0] = base_channel + i elif channels[i] == 'Yrotation': rot_ind[1] = base_channel + i elif channels[i] == 'Zrotation': rot_ind[2] = base_channel + i elif channels[i] == 'Xposition': pos_ind[0] = base_channel + i elif channels[i] == 'Yposition': pos_ind[1] = base_channel + i elif channels[i] == 'Zposition': pos_ind[2] = base_channel + i self.vertices[index].meta['rot_ind'] = list(rot_ind) self.vertices[index].meta['pos_ind'] = list(pos_ind)
resolve_indices
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, index, start_val):\n channels = self.vertices[index].meta['channels']\n base_channel = start_val\n rot_ind = -np.ones(3, dtype=int)\n pos_ind...
def x(self): for i in range(len(self.vertices)): self.vertices[i].meta['C'] = rotation_matrix(self.vertices[i].meta[ 'axis'][0], self.vertices[i].meta['axis'][1], self.vertices[i]. meta['axis'][2], self.vertices[i].meta['axis_order'], degrees=True) self.vertices[i].meta['Cinv'] = np.linalg.inv(self.vertices[i].meta ['C'])
set_rotation_matrices
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for i in range(len(self.vertices)):\n self.vertices[i].meta['C'] = rotation_matrix(self.vertices[i].meta[\n 'axis'][0], self.vert...
def x(self, X): X_prod = self._comp_prod(X) return self.variance * four_over_tau * np.arcsin(X_prod / (X_prod + 1.0))
Kdiag
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X):\n X_prod = self._comp_prod(X)\n return self.variance * four_over_tau * np.arcsin(X_prod / (X_prod + 1.0))\n", "role": "user" }, { "...
def x(self, dL_dK, X, X2=None): dvar, dw, db = self._comp_grads(dL_dK, X, X2)[:3] self.variance.gradient = dvar self.weight_variance.gradient = dw self.bias_variance.gradient = db
update_gradients_full
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dL_dK, X, X2=None):\n dvar, dw, db = self._comp_grads(dL_dK, X, X2)[:3]\n self.variance.gradient = dvar\n self.weight_variance.gradient = dw\n ...
def x(self, dL_dK, X, X2): return self._comp_grads(dL_dK, X, X2)[3]
gradients_X
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dL_dK, X, X2):\n return self._comp_grads(dL_dK, X, X2)[3]\n", "role": "user" }, { "content": "gradients_X", "role": "assistant" } ]
def x(self, dL_dK, X, X2): return self._comp_grads(dL_dK, X, X2)[3:]
gradients_X_X2
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dL_dK, X, X2):\n return self._comp_grads(dL_dK, X, X2)[3:]\n", "role": "user" }, { "content": "gradients_X_X2", "role": "assistant" ...
def x(self): return np.hstack((self.varianceU, self.varianceY, self.lengthscaleU, self.lengthscaleY))
get_params
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return np.hstack((self.varianceU, self.varianceY, self.lengthscaleU,\n self.lengthscaleY))\n", "role": "user" }, { "content": "g...
def x(self, x): assert x.size == self.num_params self.varianceU = x[0] self.varianceY = x[1] self.lengthscaleU = x[2] self.lengthscaleY = x[3]
set_params
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, x):\n assert x.size == self.num_params\n self.varianceU = x[0]\n self.varianceY = x[1]\n self.lengthscaleU = x[2]\n self.lengthscaleY = x[...
def x(self, X, X2, target): if X2 is None: X2 = X dist = np.abs(X - X2.T) ly = 1 / self.lengthscaleY lu = np.sqrt(3) / self.lengthscaleU k1 = np.exp(-ly * dist) * (2 * lu + ly) / (lu + ly) ** 2 k2 = (np.exp(-lu * dist) * (ly - 2 * lu + lu * ly * dist - lu ** 2 * dist) + np.exp(-ly * dist) * (2 * lu - ly)) / (ly - lu) ** 2 k3 = np.exp(-lu * dist) * ((1 + lu * dist) / (lu + ly) + lu / (lu + ly) ** 2) np.add(self.varianceU * self.varianceY * (k1 + k2 + k3), target, target)
K
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, X2, target):\n if X2 is None:\n X2 = X\n dist = np.abs(X - X2.T)\n ly = 1 / self.lengthscaleY\n lu = np.sqrt(3) / self.lengthscaleU...
def x(self, X, target): ly = 1 / self.lengthscaleY lu = np.sqrt(3) / self.lengthscaleU k1 = (2 * lu + ly) / (lu + ly) ** 2 k2 = (ly - 2 * lu + 2 * lu - ly) / (ly - lu) ** 2 k3 = 1 / (lu + ly) + lu / (lu + ly) ** 2 np.add(self.varianceU * self.varianceY * (k1 + k2 + k3), target, target)
Kdiag
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, target):\n ly = 1 / self.lengthscaleY\n lu = np.sqrt(3) / self.lengthscaleU\n k1 = (2 * lu + ly) / (lu + ly) ** 2\n k2 = (ly - 2 * lu + 2 ...
def x(self, dL_dK, X, X2, target): if X2 is None: X2 = X dist = np.abs(X - X2.T) ly = 1 / self.lengthscaleY lu = np.sqrt(3) / self.lengthscaleU dk1theta1 = np.exp(-ly * dist) * 2 * -lu / (lu + ly) ** 3 dk2theta1 = 1 * (np.exp(-lu * dist) * dist * (-ly + 2 * lu - lu * ly * dist + dist * lu ** 2) * (ly - lu) ** -2 + np.exp(-lu * dist) * (-2 + ly * dist - 2 * dist * lu) * (ly - lu) ** -2 + np.exp(-dist * lu) * (ly - 2 * lu + ly * lu * dist - dist * lu ** 2) * 2 * (ly - lu) ** -3 + np.exp(-dist * ly) * 2 * (ly - lu) ** -2 + np.exp(-dist * ly) * 2 * (2 * lu - ly) * (ly - lu) ** -3) dk3theta1 = np.exp(-dist * lu) * (lu + ly) ** -2 * ((2 * lu + ly + dist * lu ** 2 + lu * ly * dist) * (-dist - 2 / (lu + ly)) + 2 + 2 * lu * dist + ly * dist) dktheta1 = self.varianceU * self.varianceY * (dk1theta1 + dk2theta1 + dk3theta1) dk1theta2 = np.exp(-ly * dist) * (lu + ly) ** -2 * (-dist * (2 * lu + ly) + 1 + -2 * (2 * lu + ly) / (lu + ly)) dk2theta2 = 1 * (np.exp(-dist * lu) * (ly - lu) ** -2 * (1 + lu * dist + -2 * (ly - 2 * lu + lu * ly * dist - dist * lu ** 2) * (ly - lu) ** -1) + np.exp(-dist * ly) * (ly - lu) ** -2 * (-dist * (2 * lu - ly) - 1 + (2 * lu - ly) * -2 * (ly - lu) ** -1)) dk3theta2 = np.exp(-dist * lu) * (-3 * lu - ly - dist * lu ** 2 - lu * ly * dist) / (lu + ly) ** 3 dktheta2 = self.varianceU * self.varianceY * (dk1theta2 + dk2theta2 + dk3theta2) k1 = np.exp(-ly * dist) * (2 * lu + ly) / (lu + ly) ** 2 k2 = (np.exp(-lu * dist) * (ly - 2 * lu + lu * ly * dist - lu ** 2 * dist) + np.exp(-ly * dist) * (2 * lu - ly)) / (ly - lu) ** 2 k3 = np.exp(-lu * dist) * ((1 + lu * dist) / (lu + ly) + lu / (lu + ly) ** 2) dkdvar = k1 + k2 + k3 target[0] += np.sum(self.varianceY * dkdvar * dL_dK) target[1] += np.sum(self.varianceU * dkdvar * dL_dK) target[2] += np.sum(dktheta1 * (-np.sqrt(3) * self.lengthscaleU ** -2) * dL_dK) target[3] += np.sum(dktheta2 * -self.lengthscaleY ** -2 * dL_dK)
param_grad_helper
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dL_dK, X, X2, target):\n if X2 is None:\n X2 = X\n dist = np.abs(X - X2.T)\n ly = 1 / self.lengthscaleY\n lu = np.sqrt(3) / self.lengt...
def x(self, kern, X, likelihood, Y, Y_metadata=None): N = X.shape[0] D = X.shape[1] Kds = np.zeros(D, dtype=object) Qs = np.zeros(D, dtype=object) QTs = np.zeros(D, dtype=object) V_kron = 1 oneDkernel = kern.get_one_dimensional_kernel(D) for d in range(D): xg = list(set(X[:, d])) xg = np.reshape(xg, (len(xg), 1)) oneDkernel.lengthscale = kern.lengthscale[d] Kds[d] = oneDkernel.K(xg) [V, Q] = np.linalg.eig(Kds[d]) V_kron = np.kron(V_kron, V) Qs[d] = Q QTs[d] = Q.T noise = likelihood.variance + 1e-08 alpha_kron = self.kron_mvprod(QTs, Y) V_kron = V_kron.reshape(-1, 1) alpha_kron = alpha_kron / (V_kron + noise) alpha_kron = self.kron_mvprod(Qs, alpha_kron) log_likelihood = -0.5 * (np.dot(Y.T, alpha_kron) + np.sum(np.log(V_kron + noise)) + N * log_2_pi) derivs = np.zeros(D + 2, dtype='object') for t in range(len(derivs)): dKd_dTheta = np.zeros(D, dtype='object') gamma = np.zeros(D, dtype='object') gam = 1 for d in range(D): xg = list(set(X[:, d])) xg = np.reshape(xg, (len(xg), 1)) oneDkernel.lengthscale = kern.lengthscale[d] if t < D: dKd_dTheta[d] = oneDkernel.dKd_dLen(xg, t == d, lengthscale =kern.lengthscale[t]) elif t == D: dKd_dTheta[d] = oneDkernel.dKd_dVar(xg) else: dKd_dTheta[d] = np.identity(len(xg)) gamma[d] = np.diag(np.dot(np.dot(QTs[d], dKd_dTheta[d].T), Qs[d])) gam = np.kron(gam, gamma[d]) gam = gam.reshape(-1, 1) kappa = self.kron_mvprod(dKd_dTheta, alpha_kron) derivs[t] = 0.5 * np.dot(alpha_kron.T, kappa) - 0.5 * np.sum(gam / (V_kron + noise)) dL_dLen = derivs[:D] dL_dVar = derivs[D] dL_dThetaL = derivs[D + 1] return GridPosterior(alpha_kron=alpha_kron, QTs=QTs, Qs=Qs, V_kron=V_kron ), log_likelihood, {'dL_dLen': dL_dLen, 'dL_dVar': dL_dVar, 'dL_dthetaL': dL_dThetaL}
inference
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, kern, X, likelihood, Y, Y_metadata=None):\n N = X.shape[0]\n D = X.shape[1]\n Kds = np.zeros(D, dtype=object)\n Qs = np.zeros(D, dtype=object...
def x(self): input_dict = super(Constant, self)._save_to_input_dict() input_dict['class'] = 'GPy.mappings.Constant' input_dict['value'] = self.C.values[0] return input_dict
to_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n input_dict = super(Constant, self)._save_to_input_dict()\n input_dict['class'] = 'GPy.mappings.Constant'\n input_dict['value'] = self.C.value...
def x(self, grads): self.sigma2.gradient = grads[0] self.v.gradient = grads[1]
update_gradients
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, grads):\n self.sigma2.gradient = grads[0]\n self.v.gradient = grads[1]\n", "role": "user" }, { "content": "update_gradients", "role...
def x(self, inv_link_f, y, Y_metadata=None): assert np.atleast_1d(inv_link_f).shape == np.atleast_1d(y).shape e = y - inv_link_f objective = np.exp(gammaln((self.v + 1) * 0.5) - gammaln(self.v * 0.5) ) / np.sqrt(self.v * np.pi * self.sigma2) * (1 + 1.0 / float(self.v ) * (e ** 2 / float(self.sigma2))) ** (-0.5 * (self.v + 1)) return np.prod(objective)
pdf_link
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, inv_link_f, y, Y_metadata=None):\n assert np.atleast_1d(inv_link_f).shape == np.atleast_1d(y).shape\n e = y - inv_link_f\n objective = np.exp(ga...
def x(self, inv_link_f, y, Y_metadata=None): e = y - inv_link_f objective = +gammaln((self.v + 1) * 0.5) - gammaln(self.v * 0.5 ) - 0.5 * np.log(self.sigma2 * self.v * np.pi) - 0.5 * (self.v + 1 ) * np.log(1 + 1 / np.float(self.v) * (e ** 2 / self.sigma2)) return objective
logpdf_link
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, inv_link_f, y, Y_metadata=None):\n e = y - inv_link_f\n objective = +gammaln((self.v + 1) * 0.5) - gammaln(self.v * 0.5\n ) - 0.5 * np.log(s...
def x(self, inv_link_f, y, Y_metadata=None): e = y - inv_link_f grad = (self.v + 1) * e / (self.v * self.sigma2 + e ** 2) return grad
dlogpdf_dlink
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, inv_link_f, y, Y_metadata=None):\n e = y - inv_link_f\n grad = (self.v + 1) * e / (self.v * self.sigma2 + e ** 2)\n return grad\n", "role": ...
def x(self, inv_link_f, y, Y_metadata=None): e = y - inv_link_f hess = (self.v + 1) * (e ** 2 - self.v * self.sigma2) / (self.sigma2 * self.v + e ** 2) ** 2 return hess
d2logpdf_dlink2
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, inv_link_f, y, Y_metadata=None):\n e = y - inv_link_f\n hess = (self.v + 1) * (e ** 2 - self.v * self.sigma2) / (self.sigma2 *\n self.v + e ...
def x(self, inv_link_f, y, Y_metadata=None): e = y - inv_link_f d3lik_dlink3 = -(2 * (self.v + 1) * -e * (e ** 2 - 3 * self.v * self. sigma2)) / (e ** 2 + self.sigma2 * self.v) ** 3 return d3lik_dlink3
d3logpdf_dlink3
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, inv_link_f, y, Y_metadata=None):\n e = y - inv_link_f\n d3lik_dlink3 = -(2 * (self.v + 1) * -e * (e ** 2 - 3 * self.v * self.\n sigma2)) / (...
def x(self, inv_link_f, y, Y_metadata=None): e = y - inv_link_f e2 = np.square(e) dlogpdf_dvar = self.v * (e2 - self.sigma2) / (2 * self.sigma2 * (self. sigma2 * self.v + e2)) return dlogpdf_dvar
dlogpdf_link_dvar
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, inv_link_f, y, Y_metadata=None):\n e = y - inv_link_f\n e2 = np.square(e)\n dlogpdf_dvar = self.v * (e2 - self.sigma2) / (2 * self.sigma2 * (sel...
def x(self, inv_link_f, y, Y_metadata=None): e = y - inv_link_f dlogpdf_dlink_dvar = self.v * (self.v + 1) * -e / (self.sigma2 * self.v + e ** 2) ** 2 return dlogpdf_dlink_dvar
dlogpdf_dlink_dvar
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, inv_link_f, y, Y_metadata=None):\n e = y - inv_link_f\n dlogpdf_dlink_dvar = self.v * (self.v + 1) * -e / (self.sigma2 * self.v +\n e ** 2) ...
def x(self, inv_link_f, y, Y_metadata=None): e = y - inv_link_f d2logpdf_dlink2_dvar = self.v * (self.v + 1) * (self.sigma2 * self.v - 3 * e ** 2) / (self.sigma2 * self.v + e ** 2) ** 3 return d2logpdf_dlink2_dvar
d2logpdf_dlink2_dvar
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, inv_link_f, y, Y_metadata=None):\n e = y - inv_link_f\n d2logpdf_dlink2_dvar = self.v * (self.v + 1) * (self.sigma2 * self.v - \n 3 * e ** 2...
def x(self, gp, Y_metadata=None): orig_shape = gp.shape gp = gp.flatten() dfs = np.ones_like(gp) * self.v scales = np.ones_like(gp) * np.sqrt(self.sigma2) student_t_samples = stats.t.rvs(dfs, loc=self.gp_link.transf(gp), scale =scales) return student_t_samples.reshape(orig_shape)
samples
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, gp, Y_metadata=None):\n orig_shape = gp.shape\n gp = gp.flatten()\n dfs = np.ones_like(gp) * self.v\n scales = np.ones_like(gp) * np.sqrt(sel...
def x(self, kern, X, Z, likelihood, Y, Y_metadata, Lm=None, dL_dKmm=None, psi0=None, psi1=None, psi2=None, **kwargs): return self.inference_method.inference(kern, X, Z, likelihood, Y, Y_metadata, Lm=Lm, dL_dKmm=dL_dKmm, psi0=psi0, psi1=psi1, psi2=psi2, **kwargs)
inner_parameters_changed
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, kern, X, Z, likelihood, Y, Y_metadata, Lm=None, dL_dKmm=None,\n psi0=None, psi1=None, psi2=None, **kwargs):\n return self.inference_method.inferenc...
def x(self, full_values=None, current_values=None, value_indices=None): for key in current_values.keys(): if value_indices is not None and key in value_indices: index = value_indices[key] else: index = slice(None) if key in full_values: try: full_values[key][index] += current_values[key] except: full_values[key] += current_values[key] else: full_values[key] = current_values[key]
inner_take_over_or_update
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, full_values=None, current_values=None, value_indices=None):\n for key in current_values.keys():\n if value_indices is not None and key in value...
def x(self, full_values): if self.has_uncertain_inputs(): dL_dKmm = full_values['dL_dKmm'] self.kern.update_gradients_full(dL_dKmm, self.Z, None) kgrad = self.kern.gradient.copy() self.kern.update_gradients_expectations(variational_posterior=self. X, Z=self.Z, dL_dpsi0=full_values['dL_dpsi0'], dL_dpsi1= full_values['dL_dpsi1'], dL_dpsi2=full_values['dL_dpsi2']) self.kern.gradient += kgrad self.Z.gradient = self.kern.gradients_X(dL_dKmm, self.Z) self.Z.gradient += self.kern.gradients_Z_expectations( variational_posterior=self.X, Z=self.Z, dL_dpsi0=full_values[ 'dL_dpsi0'], dL_dpsi1=full_values['dL_dpsi1'], dL_dpsi2= full_values['dL_dpsi2']) else: self.kern.update_gradients_diag(full_values['dL_dKdiag'], self.X) kgrad = self.kern.gradient.copy() self.kern.update_gradients_full(full_values['dL_dKnm'], self.X, self.Z) kgrad += self.kern.gradient self.kern.update_gradients_full(full_values['dL_dKmm'], self.Z, None) self.kern.gradient += kgrad self.Z.gradient = self.kern.gradients_X(full_values['dL_dKmm'], self.Z) self.Z.gradient += self.kern.gradients_X(full_values['dL_dKnm'].T, self.Z, self.X) self.likelihood.update_gradients(full_values['dL_dthetaL'])
outer_values_update
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, full_values):\n if self.has_uncertain_inputs():\n dL_dKmm = full_values['dL_dKmm']\n self.kern.update_gradients_full(dL_dKmm, self.Z, No...
def x(self): retd = dict(dL_dKmm=np.zeros((self.Z.shape[0], self.Z.shape[0]))) if self.has_uncertain_inputs(): retd.update(dict(dL_dpsi0=np.zeros(self.X.shape[0]), dL_dpsi1=np. zeros((self.X.shape[0], self.Z.shape[0])), dL_dpsi2=np.zeros(( self.X.shape[0], self.Z.shape[0], self.Z.shape[0])))) else: retd.update({'dL_dKdiag': np.zeros(self.X.shape[0]), 'dL_dKnm': np. zeros((self.X.shape[0], self.Z.shape[0]))}) return retd
outer_init_full_values
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n retd = dict(dL_dKmm=np.zeros((self.Z.shape[0], self.Z.shape[0])))\n if self.has_uncertain_inputs():\n retd.update(dict(dL_dpsi0=np.zeros(...
def x(self, save_data=True): model_dict = super(SparseGPClassification, self).to_dict(save_data) model_dict['class'] = 'GPy.models.SparseGPClassification' return model_dict
to_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, save_data=True):\n model_dict = super(SparseGPClassification, self).to_dict(save_data)\n model_dict['class'] = 'GPy.models.SparseGPClassification'\...
def x(input_dict, data=None): import GPy m = GPy.core.model.Model.from_dict(input_dict, data) from copy import deepcopy sparse_gp = deepcopy(m) return SparseGPClassification(sparse_gp.X, sparse_gp.Y, sparse_gp.Z, sparse_gp.kern, sparse_gp.likelihood, sparse_gp.inference_method, sparse_gp.mean_function, name='sparse_gp_classification')
from_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(input_dict, data=None):\n import GPy\n m = GPy.core.model.Model.from_dict(input_dict, data)\n from copy import deepcopy\n sparse_gp = deepcopy(m)\n...
def x(self): input_dict = super(Gaussian, self)._save_to_input_dict() input_dict['class'] = 'GPy.likelihoods.Gaussian' input_dict['variance'] = self.variance.values.tolist() return input_dict
to_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n input_dict = super(Gaussian, self)._save_to_input_dict()\n input_dict['class'] = 'GPy.likelihoods.Gaussian'\n input_dict['variance'] = self.v...
def x(self, data_i, tau_i, v_i, Y_metadata_i=None): sigma2_hat = 1.0 / (1.0 / self.variance + tau_i) mu_hat = sigma2_hat * (data_i / self.variance + v_i) sum_var = self.variance + 1.0 / tau_i Z_hat = 1.0 / np.sqrt(2.0 * np.pi * sum_var) * np.exp(-0.5 * (data_i - v_i / tau_i) ** 2.0 / sum_var) return Z_hat, mu_hat, sigma2_hat
moments_match_ep
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data_i, tau_i, v_i, Y_metadata_i=None):\n sigma2_hat = 1.0 / (1.0 / self.variance + tau_i)\n mu_hat = sigma2_hat * (data_i / self.variance + v_i)\n...
def x(self, link_f, y, Y_metadata=None): return np.exp(np.sum(np.log(stats.norm.pdf(y, link_f, np.sqrt(self. variance)))))
pdf_link
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, link_f, y, Y_metadata=None):\n return np.exp(np.sum(np.log(stats.norm.pdf(y, link_f, np.sqrt(self.\n variance)))))\n", "role": "user" }, ...
def x(self, link_f, y, Y_metadata=None): ln_det_cov = np.log(self.variance) return -(1.0 / (2 * self.variance)) * (y - link_f ) ** 2 - 0.5 * ln_det_cov - 0.5 * np.log(2.0 * np.pi)
logpdf_link
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, link_f, y, Y_metadata=None):\n ln_det_cov = np.log(self.variance)\n return -(1.0 / (2 * self.variance)) * (y - link_f\n ) ** 2 - 0.5 * ln_de...
def x(self, link_f, y, Y_metadata=None): s2_i = 1.0 / self.variance grad = s2_i * y - s2_i * link_f return grad
dlogpdf_dlink
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, link_f, y, Y_metadata=None):\n s2_i = 1.0 / self.variance\n grad = s2_i * y - s2_i * link_f\n return grad\n", "role": "user" }, { "c...
def x(self, link_f, y, Y_metadata=None): N = y.shape[0] D = link_f.shape[1] hess = -(1.0 / self.variance) * np.ones((N, D)) return hess
d2logpdf_dlink2
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, link_f, y, Y_metadata=None):\n N = y.shape[0]\n D = link_f.shape[1]\n hess = -(1.0 / self.variance) * np.ones((N, D))\n return hess\n", "...
def x(self, link_f, y, Y_metadata=None): N = y.shape[0] D = link_f.shape[1] d3logpdf_dlink3 = np.zeros((N, D)) return d3logpdf_dlink3
d3logpdf_dlink3
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, link_f, y, Y_metadata=None):\n N = y.shape[0]\n D = link_f.shape[1]\n d3logpdf_dlink3 = np.zeros((N, D))\n return d3logpdf_dlink3\n", "ro...
def x(self, link_f, y, Y_metadata=None): e = y - link_f s_4 = 1.0 / self.variance ** 2 dlik_dsigma = -0.5 / self.variance + 0.5 * s_4 * np.square(e) return dlik_dsigma
dlogpdf_link_dvar
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, link_f, y, Y_metadata=None):\n e = y - link_f\n s_4 = 1.0 / self.variance ** 2\n dlik_dsigma = -0.5 / self.variance + 0.5 * s_4 * np.square(e)\n...
def x(self, link_f, y, Y_metadata=None): s_4 = 1.0 / self.variance ** 2 dlik_grad_dsigma = -s_4 * y + s_4 * link_f return dlik_grad_dsigma
dlogpdf_dlink_dvar
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, link_f, y, Y_metadata=None):\n s_4 = 1.0 / self.variance ** 2\n dlik_grad_dsigma = -s_4 * y + s_4 * link_f\n return dlik_grad_dsigma\n", "ro...
def x(self, link_f, y, Y_metadata=None): s_4 = 1.0 / self.variance ** 2 N = y.shape[0] D = link_f.shape[1] d2logpdf_dlink2_dvar = np.ones((N, D)) * s_4 return d2logpdf_dlink2_dvar
d2logpdf_dlink2_dvar
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, link_f, y, Y_metadata=None):\n s_4 = 1.0 / self.variance ** 2\n N = y.shape[0]\n D = link_f.shape[1]\n d2logpdf_dlink2_dvar = np.ones((N, D))...
def x(self, gp, Y_metadata=None): orig_shape = gp.shape gp = gp.flatten() gp = gp.flatten() Ysim = np.array([np.random.normal(self.gp_link.transf(gpj), scale=np. sqrt(self.variance), size=1) for gpj in gp]) return Ysim.reshape(orig_shape)
samples
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, gp, Y_metadata=None):\n orig_shape = gp.shape\n gp = gp.flatten()\n gp = gp.flatten()\n Ysim = np.array([np.random.normal(self.gp_link.transf...
def x(self, y_test, mu_star, var_star, Y_metadata=None): v = var_star + self.variance return -0.5 * np.log(2 * np.pi) - 0.5 * np.log(v) - 0.5 * np.square( y_test - mu_star) / v
log_predictive_density
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, y_test, mu_star, var_star, Y_metadata=None):\n v = var_star + self.variance\n return -0.5 * np.log(2 * np.pi) - 0.5 * np.log(v) - 0.5 * np.square(\...
def x(self, save_data=True): input_dict = super(GP, self)._save_to_input_dict() input_dict['class'] = 'GPy.core.GP' if not save_data: input_dict['X'] = None input_dict['Y'] = None else: try: input_dict['X'] = self.X.values.tolist() except: input_dict['X'] = self.X.tolist() try: input_dict['Y'] = self.Y.values.tolist() except: input_dict['Y'] = self.Y.tolist() input_dict['kernel'] = self.kern.to_dict() input_dict['likelihood'] = self.likelihood.to_dict() if self.mean_function is not None: input_dict['mean_function'] = self.mean_function.to_dict() input_dict['inference_method'] = self.inference_method.to_dict() if self.Y_metadata is not None: input_dict['Y_metadata'] = self.Y_metadata if self.normalizer is not None: input_dict['normalizer'] = self.normalizer.to_dict() return input_dict
to_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, save_data=True):\n input_dict = super(GP, self)._save_to_input_dict()\n input_dict['class'] = 'GPy.core.GP'\n if not save_data:\n input_d...
def x(self, X=None, Y=None): self.update_model(False) if Y is not None: if self.normalizer is not None: self.normalizer.scale_by(Y) self.Y_normalized = ObsAr(self.normalizer.normalize(Y)) self.Y = Y else: self.Y = ObsAr(Y) self.Y_normalized = self.Y if X is not None: if self.X in self.parameters: if isinstance(self.X, VariationalPosterior): assert isinstance(X, type(self.X) ), 'The given X must have the same type as the X in the model!' index = self.X._parent_index_ self.unlink_parameter(self.X) self.X = X self.link_parameter(self.X, index=index) else: index = self.X._parent_index_ self.unlink_parameter(self.X) from ..core import Param self.X = Param('latent mean', X) self.link_parameter(self.X, index=index) else: self.X = ObsAr(X) self.update_model(True)
set_XY
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X=None, Y=None):\n self.update_model(False)\n if Y is not None:\n if self.normalizer is not None:\n self.normalizer.scale_by(Y)\n...
def x(self, Xnew, full_cov=False, Y_metadata=None, kern=None, likelihood= None, include_likelihood=True): mean, var = self._raw_predict(Xnew, full_cov=full_cov, kern=kern) if include_likelihood: if likelihood is None: likelihood = self.likelihood mean, var = likelihood.predictive_values(mean, var, full_cov, Y_metadata=Y_metadata) if self.normalizer is not None: mean = self.normalizer.inverse_mean(mean) if full_cov & (mean.shape[1] > 1): var = self.normalizer.inverse_covariance(var) else: var = self.normalizer.inverse_variance(var) return mean, var
predict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Xnew, full_cov=False, Y_metadata=None, kern=None, likelihood=\n None, include_likelihood=True):\n mean, var = self._raw_predict(Xnew, full_cov=full...
def x(self, Xnew, full_cov=False, Y_metadata=None, kern=None): return self.predict(Xnew, full_cov, Y_metadata, kern, None, False)
predict_noiseless
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Xnew, full_cov=False, Y_metadata=None, kern=None):\n return self.predict(Xnew, full_cov, Y_metadata, kern, None, False)\n", "role": "user" }, ...
def x(self, X, quantiles=(2.5, 97.5), Y_metadata=None, kern=None, likelihood=None): m, v = self._raw_predict(X, full_cov=False, kern=kern) if likelihood is None: likelihood = self.likelihood quantiles = likelihood.predictive_quantiles(m, v, quantiles, Y_metadata =Y_metadata) if self.normalizer is not None: quantiles = [self.normalizer.inverse_mean(q) for q in quantiles] return quantiles
predict_quantiles
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, quantiles=(2.5, 97.5), Y_metadata=None, kern=None,\n likelihood=None):\n m, v = self._raw_predict(X, full_cov=False, kern=kern)\n if likeliho...
def x(self, Xnew, kern=None): if kern is None: kern = self.kern mean_jac = np.empty((Xnew.shape[0], Xnew.shape[1], self.output_dim)) for i in range(self.output_dim): mean_jac[:, :, i] = kern.gradients_X(self.posterior.woodbury_vector [:, i:i + 1].T, Xnew, self._predictive_variable) dv_dX = kern.gradients_X_diag(np.ones(Xnew.shape[0]), Xnew) if self.posterior.woodbury_inv.ndim == 3: var_jac = np.empty(dv_dX.shape + (self.posterior.woodbury_inv.shape [2],)) var_jac[:] = dv_dX[:, :, None] for i in range(self.posterior.woodbury_inv.shape[2]): alpha = -2.0 * np.dot(kern.K(Xnew, self._predictive_variable), self.posterior.woodbury_inv[:, :, i]) var_jac[:, :, i] += kern.gradients_X(alpha, Xnew, self. _predictive_variable) else: var_jac = dv_dX alpha = -2.0 * np.dot(kern.K(Xnew, self._predictive_variable), self .posterior.woodbury_inv) var_jac += kern.gradients_X(alpha, Xnew, self._predictive_variable) return mean_jac, var_jac
predictive_gradients
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Xnew, kern=None):\n if kern is None:\n kern = self.kern\n mean_jac = np.empty((Xnew.shape[0], Xnew.shape[1], self.output_dim))\n for i in...
def x(self, Xnew, kern=None, full_cov=False): if kern is None: kern = self.kern mean_jac = np.empty((Xnew.shape[0], Xnew.shape[1], self.output_dim)) for i in range(self.output_dim): mean_jac[:, :, i] = kern.gradients_X(self.posterior.woodbury_vector [:, i:i + 1].T, Xnew, self._predictive_variable) dK_dXnew_full = np.empty((self._predictive_variable.shape[0], Xnew. shape[0], Xnew.shape[1])) one = np.ones((1, 1)) for i in range(self._predictive_variable.shape[0]): dK_dXnew_full[i] = kern.gradients_X(one, Xnew, self. _predictive_variable[[i]]) if full_cov: dK2_dXdX = kern.gradients_XX(one, Xnew) else: dK2_dXdX = kern.gradients_XX_diag(one, Xnew) def x(wi): if full_cov: var_jac = dK2_dXdX - np.einsum('qnm,msr->nsqr', dK_dXnew_full.T .dot(wi), dK_dXnew_full) else: var_jac = dK2_dXdX - np.einsum('qnm,mnr->nqr', dK_dXnew_full.T. dot(wi), dK_dXnew_full) return var_jac if self.posterior.woodbury_inv.ndim == 3: if full_cov: var_jac = np.empty((Xnew.shape[0], Xnew.shape[0], Xnew.shape[1], Xnew.shape[1], self.output_dim)) for d in range(self.posterior.woodbury_inv.shape[2]): var_jac[:, :, :, :, d] = compute_cov_inner(self.posterior. woodbury_inv[:, :, d]) else: var_jac = np.empty((Xnew.shape[0], Xnew.shape[1], Xnew.shape[1], self.output_dim)) for d in range(self.posterior.woodbury_inv.shape[2]): var_jac[:, :, :, d] = compute_cov_inner(self.posterior. woodbury_inv[:, :, d]) else: var_jac = compute_cov_inner(self.posterior.woodbury_inv) return mean_jac, var_jac
predict_jacobian
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Xnew, kern=None, full_cov=False):\n if kern is None:\n kern = self.kern\n mean_jac = np.empty((Xnew.shape[0], Xnew.shape[1], self.output_dim...
def x(self, Xnew, kern=None, mean=True, covariance=True): if kern is None: kern = self.kern mu_jac, var_jac = self.predict_jacobian(Xnew, kern, full_cov=False) mumuT = np.einsum('iqd,ipd->iqp', mu_jac, mu_jac) Sigma = np.zeros(mumuT.shape) if var_jac.ndim == 4: Sigma = var_jac.sum(-1) else: Sigma = self.output_dim * var_jac G = 0.0 if mean: G += mumuT if covariance: G += Sigma return G
predict_wishart_embedding
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Xnew, kern=None, mean=True, covariance=True):\n if kern is None:\n kern = self.kern\n mu_jac, var_jac = self.predict_jacobian(Xnew, kern, fu...
def x(self, Xnew, kern=None, mean=True, covariance=True, dimensions=None): G = self.predict_wishart_embedding(Xnew, kern, mean, covariance) if dimensions is None: dimensions = self.get_most_significant_input_dimensions()[:2] G = G[:, dimensions][:, :, dimensions] from ..util.linalg import jitchol mag = np.empty(Xnew.shape[0]) for n in range(Xnew.shape[0]): try: mag[n] = np.sqrt(np.exp(2 * np.sum(np.log(np.diag(jitchol(G[n, :, :])))))) except: mag[n] = np.sqrt(np.linalg.det(G[n, :, :])) return mag
predict_magnification
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Xnew, kern=None, mean=True, covariance=True, dimensions=None):\n G = self.predict_wishart_embedding(Xnew, kern, mean, covariance)\n if dimensions i...
def x(self, X, size=10, **predict_kwargs): predict_kwargs['full_cov'] = True m, v = self._raw_predict(X, **predict_kwargs) if self.normalizer is not None: m, v = self.normalizer.inverse_mean(m ), self.normalizer.inverse_variance(v) def x(m, v): return np.random.multivariate_normal(m, v, size).T if self.output_dim == 1: return sim_one_dim(m.flatten(), v)[:, np.newaxis, :] else: fsim = np.empty((X.shape[0], self.output_dim, size)) for d in range(self.output_dim): if v.ndim == 3: fsim[:, d, :] = sim_one_dim(m[:, d], v[:, :, d]) else: fsim[:, d, :] = sim_one_dim(m[:, d], v) return fsim
posterior_samples_f
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, size=10, **predict_kwargs):\n predict_kwargs['full_cov'] = True\n m, v = self._raw_predict(X, **predict_kwargs)\n if self.normalizer is not N...
def x(self, X, size=10, Y_metadata=None, likelihood=None, **predict_kwargs): fsim = self.posterior_samples_f(X, size, **predict_kwargs) if likelihood is None: likelihood = self.likelihood if fsim.ndim == 3: for d in range(fsim.shape[1]): fsim[:, d] = likelihood.samples(fsim[:, d], Y_metadata=Y_metadata) else: fsim = likelihood.samples(fsim, Y_metadata=Y_metadata) return fsim
posterior_samples
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, size=10, Y_metadata=None, likelihood=None, **predict_kwargs):\n fsim = self.posterior_samples_f(X, size, **predict_kwargs)\n if likelihood is No...
def x(self, optimizer=None, start=None, messages=False, max_iters=1000, ipython_notebook=True, clear_after_finish=False, **kwargs): self.inference_method.on_optimization_start() try: ret = super(GP, self).optimize(optimizer, start, messages, max_iters, ipython_notebook, clear_after_finish, **kwargs) except KeyboardInterrupt: print( 'KeyboardInterrupt caught, calling on_optimization_end() to round things up' ) self.inference_method.on_optimization_end() raise return ret
optimize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, optimizer=None, start=None, messages=False, max_iters=1000,\n ipython_notebook=True, clear_after_finish=False, **kwargs):\n self.inference_method.o...
def x(self, Y_new, optimize=True): from ..inference.latent_function_inference.inferenceX import infer_newX return infer_newX(self, Y_new, optimize=optimize)
infer_newX
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Y_new, optimize=True):\n from ..inference.latent_function_inference.inferenceX import infer_newX\n return infer_newX(self, Y_new, optimize=optimize...
def x(self, x_test, y_test, Y_metadata=None): mu_star, var_star = self._raw_predict(x_test) return self.likelihood.log_predictive_density(y_test, mu_star, var_star, Y_metadata=Y_metadata)
log_predictive_density
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, x_test, y_test, Y_metadata=None):\n mu_star, var_star = self._raw_predict(x_test)\n return self.likelihood.log_predictive_density(y_test, mu_star, ...
def x(self, x_test, y_test, Y_metadata=None, num_samples=1000): mu_star, var_star = self._raw_predict(x_test) return self.likelihood.log_predictive_density_sampling(y_test, mu_star, var_star, Y_metadata=Y_metadata, num_samples=num_samples)
log_predictive_density_sampling
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, x_test, y_test, Y_metadata=None, num_samples=1000):\n mu_star, var_star = self._raw_predict(x_test)\n return self.likelihood.log_predictive_density...
def x(self, X1, X2): return self.posterior.covariance_between_points(self.kern, self.X, X1, X2)
posterior_covariance_between_points
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X1, X2):\n return self.posterior.covariance_between_points(self.kern, self.X, X1, X2)\n", "role": "user" }, { "content": "posterior_covari...
def x(shape, var_name, dims=None, coords=None, default_dims=None): if default_dims is None: default_dims = [] if dims is None: dims = [] if len([dim for dim in dims if dim not in default_dims]) > len(shape): warnings.warn( 'More dims ({dims_len}) given than exists ({shape_len}). Passed array should have shape (chains, draws, *shape)' .format(dims_len=len(dims), shape_len=len(shape)), SyntaxWarning) if coords is None: coords = {} coords = deepcopy(coords) dims = deepcopy(dims) for idx, dim_len in enumerate(shape): if len(dims) < idx + 1 or dims[idx] is None: dim_name = '{var_name}_dim_{idx}'.format(var_name=var_name, idx=idx ) if len(dims) < idx + 1: dims.append(dim_name) else: dims[idx] = dim_name dim_name = dims[idx] if dim_name not in coords: coords[dim_name] = np.arange(dim_len) coords = {key: coord for key, coord in coords.items() if any(key == dim for dim in dims)} return dims, coords
generate_dims_coords
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(shape, var_name, dims=None, coords=None, default_dims=None):\n if default_dims is None:\n default_dims = []\n if dims is None:\n dims = []\...
def x(ary, *, var_name='data', coords=None, dims=None): default_dims = ['chain', 'draw'] ary = np.atleast_2d(ary) n_chains, n_samples, *shape = ary.shape if n_chains > n_samples: warnings.warn( 'More chains ({n_chains}) than draws ({n_samples}). Passed array should have shape (chains, draws, *shape)' .format(n_chains=n_chains, n_samples=n_samples), SyntaxWarning) dims, coords = generate_dims_coords(shape, var_name, dims=dims, coords= coords, default_dims=default_dims) if 'draw' not in dims: dims = ['draw'] + dims if 'chain' not in dims: dims = ['chain'] + dims if 'chain' not in coords: coords['chain'] = np.arange(n_chains) if 'draw' not in coords: coords['draw'] = np.arange(n_samples) coords = {key: xr.IndexVariable((key,), data=coords[key]) for key in dims} return xr.DataArray(ary, coords=coords, dims=dims)
numpy_to_data_array
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ary, *, var_name='data', coords=None, dims=None):\n default_dims = ['chain', 'draw']\n ary = np.atleast_2d(ary)\n n_chains, n_samples, *shape = ary.sh...
def x(data, *, attrs=None, library=None, coords=None, dims=None): if dims is None: dims = {} data_vars = {} for key, values in data.items(): data_vars[key] = numpy_to_data_array(values, var_name=key, coords= coords, dims=dims.get(key)) return xr.Dataset(data_vars=data_vars, attrs=make_attrs(attrs=attrs, library=library))
dict_to_dataset
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(data, *, attrs=None, library=None, coords=None, dims=None):\n if dims is None:\n dims = {}\n data_vars = {}\n for key, values in data.items():\...
def x(attrs=None, library=None): default_attrs = {'created_at': datetime.datetime.utcnow().isoformat()} if library is not None: library_name = library.__name__ default_attrs['inference_library'] = library_name try: version = pkg_resources.get_distribution(library_name).version default_attrs['inference_library_version'] = version except pkg_resources.DistributionNotFound: if hasattr(library, '__version__'): version = library.__version__ default_attrs['inference_library_version'] = version if attrs is not None: default_attrs.update(attrs) return default_attrs
make_attrs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(attrs=None, library=None):\n default_attrs = {'created_at': datetime.datetime.utcnow().isoformat()}\n if library is not None:\n library_name = lib...
def x(data_home=None): if data_home is None: data_home = os.environ.get('ARVIZ_DATA', os.path.join('~', 'arviz_data')) data_home = os.path.expanduser(data_home) if not os.path.exists(data_home): os.makedirs(data_home) return data_home
get_data_home
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(data_home=None):\n if data_home is None:\n data_home = os.environ.get('ARVIZ_DATA', os.path.join('~',\n 'arviz_data'))\n data_home = os...
def x(path): sha256hash = hashlib.sha256() chunk_size = 8192 with open(path, 'rb') as buff: while True: buffer = buff.read(chunk_size) if not buffer: break sha256hash.update(buffer) return sha256hash.hexdigest()
sha256
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path):\n sha256hash = hashlib.sha256()\n chunk_size = 8192\n with open(path, 'rb') as buff:\n while True:\n buffer = buff.read(chunk...
def x(dataset=None, data_home=None): if dataset in LOCAL_DATASETS: resource = LOCAL_DATASETS[dataset] return from_netcdf(resource.filename) elif dataset in REMOTE_DATASETS: remote = REMOTE_DATASETS[dataset] home_dir = get_data_home(data_home=data_home) file_path = os.path.join(home_dir, remote.filename) if not os.path.exists(file_path): urlretrieve(remote.url, file_path) checksum = _sha256(file_path) if remote.checksum != checksum: raise IOError( '{} has an SHA256 checksum ({}) differing from expected ({}), file may be corrupted. Run `arviz.clear_data_home()` and try again, or please open an issue.' .format(file_path, checksum, remote.checksum)) return from_netcdf(file_path) else: raise ValueError( 'Dataset {} not found! The following are available:\n{}'.format (dataset, list_datasets()))
load_arviz_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(dataset=None, data_home=None):\n if dataset in LOCAL_DATASETS:\n resource = LOCAL_DATASETS[dataset]\n return from_netcdf(resource.filename)\n ...
def x(): lines = [] for name, resource in itertools.chain(LOCAL_DATASETS.items(), REMOTE_DATASETS.items()): if isinstance(resource, LocalFileMetadata): location = 'local: {}'.format(resource.filename) elif isinstance(resource, RemoteFileMetadata): location = 'remote: {}'.format(resource.url) else: location = 'unknown' lines.append('{}\n{}\n{}\n{}'.format(name, '=' * len(name), resource.description, location)) return '\n\n{}\n\n'.format(10 * '-').join(lines)
list_datasets
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n lines = []\n for name, resource in itertools.chain(LOCAL_DATASETS.items(),\n REMOTE_DATASETS.items()):\n if isinstance(resource, Local...
def x(sampler, var_names, arg_names): if hasattr(sampler, 'args'): num_vars = sampler.chain.shape[-1] num_args = len(sampler.args) elif hasattr(sampler, 'log_prob_fn'): num_vars = sampler.get_chain().shape[-1] num_args = len(sampler.log_prob_fn.args) else: num_vars = sampler.get_chain().shape[-1] num_args = 0 if var_names is None: var_names = ['var_{}'.format(idx) for idx in range(num_vars)] if arg_names is None: arg_names = ['arg_{}'.format(idx) for idx in range(num_args)] if len(var_names) != num_vars: raise ValueError( 'The sampler has {} variables, but only {} var_names were provided!' .format(num_vars, len(var_names))) if len(arg_names) != num_args: raise ValueError( 'The sampler has {} args, but only {} arg_names were provided!' .format(num_args, len(arg_names))) return var_names, arg_names
verify_names
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sampler, var_names, arg_names):\n if hasattr(sampler, 'args'):\n num_vars = sampler.chain.shape[-1]\n num_args = len(sampler.args)\n elif h...
def x(sampler=None, *, var_names=None, arg_names=None, coords=None, dims=None): return EmceeConverter(sampler=sampler, var_names=var_names, arg_names= arg_names, coords=coords, dims=dims).to_inference_data()
from_emcee
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sampler=None, *, var_names=None, arg_names=None, coords=None, dims=None):\n return EmceeConverter(sampler=sampler, var_names=var_names, arg_names=\n ...
def x(self): data = {} for idx, var_name in enumerate(self.var_names): data[var_name] = self.sampler.get_chain()[..., idx].T if hasattr(self .sampler, 'get_chain') else self.sampler.chain[..., idx] return dict_to_dataset(data, library=self.emcee, coords=self.coords, dims=self.dims)
posterior_to_xarray
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n data = {}\n for idx, var_name in enumerate(self.var_names):\n data[var_name] = self.sampler.get_chain()[..., idx].T if hasattr(self\n ...
def x(self): data = {} for idx, var_name in enumerate(self.arg_names): data[var_name] = self.sampler.log_prob_fn.args[idx] if hasattr(self .sampler, 'log_prob_fn') else self.sampler.args[idx] return dict_to_dataset(data, library=self.emcee, coords=self.coords, dims=self.dims)
observed_data_to_xarray
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n data = {}\n for idx, var_name in enumerate(self.arg_names):\n data[var_name] = self.sampler.log_prob_fn.args[idx] if hasattr(self\n ...
def x(posterior=None, *, var_names=None, model_fn=None, feed_dict=None, posterior_predictive_samples=100, posterior_predictive_size=1, observed =None, coords=None, dims=None): return TfpConverter(posterior=posterior, var_names=var_names, model_fn= model_fn, feed_dict=feed_dict, posterior_predictive_samples= posterior_predictive_samples, posterior_predictive_size= posterior_predictive_size, observed=observed, coords=coords, dims=dims ).to_inference_data()
from_tfp
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(posterior=None, *, var_names=None, model_fn=None, feed_dict=None,\n posterior_predictive_samples=100, posterior_predictive_size=1, observed\n =None, coor...
def x(self): data = {} for i, var_name in enumerate(self.var_names): data[var_name] = np.expand_dims(self.posterior[i], axis=0) return dict_to_dataset(data, library=self.tfp, coords=self.coords, dims =self.dims)
posterior_to_xarray
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n data = {}\n for i, var_name in enumerate(self.var_names):\n data[var_name] = np.expand_dims(self.posterior[i], axis=0)\n return dict_t...
def x(self): if self.observed is None: return None observed_data = {} if isinstance(self.observed, self.tf.Tensor): with self.tf.Session() as sess: vals = sess.run(self.observed, feed_dict=self.feed_dict) else: vals = self.observed if self.dims is None: dims = {} else: dims = self.dims name = 'obs' val_dims = dims.get(name) vals = np.atleast_1d(vals) val_dims, coords = generate_dims_coords(vals.shape, name, dims=val_dims, coords=self.coords) observed_data[name] = xr.DataArray(vals, dims=val_dims, coords=coords) return xr.Dataset(data_vars=observed_data, attrs=make_attrs(library= self.tfp))
observed_data_to_xarray
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.observed is None:\n return None\n observed_data = {}\n if isinstance(self.observed, self.tf.Tensor):\n with self.tf.Ses...
def x(self): if self.model_fn is None: return None posterior_preds = [] sample_size = self.posterior[0].shape[0] for i in np.arange(0, sample_size, int(sample_size / self. posterior_predictive_samples)): variables = {} for var_i, var_name in enumerate(self.var_names): variables[var_name] = self.posterior[var_i][i] with self.ed.interception(self._value_setter(variables)): if self.posterior_predictive_size > 1: posterior_preds.append([self.model_fn() for _ in range(self .posterior_predictive_size)]) else: posterior_preds.append(self.model_fn()) data = {} with self.tf.Session() as sess: data['obs'] = np.expand_dims(sess.run(posterior_preds, feed_dict= self.feed_dict), axis=0) return dict_to_dataset(data, library=self.tfp, coords=self.coords, dims =self.dims)
posterior_predictive_to_xarray
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.model_fn is None:\n return None\n posterior_preds = []\n sample_size = self.posterior[0].shape[0]\n for i in np.arange(0, s...
def x(self): if self.model_fn is None or self.observed is None: return None log_likelihood = [] sample_size = self.posterior[0].shape[0] for i in range(sample_size): variables = {} for var_i, var_name in enumerate(self.var_names): variables[var_name] = self.posterior[var_i][i] with self.ed.interception(self._value_setter(variables)): log_likelihood.append(self.model_fn().distribution.log_prob( self.observed)) data = {} if self.dims is not None: coord_name = self.dims.get('obs') else: coord_name = None dims = {'log_likelihood': coord_name} with self.tf.Session() as sess: data['log_likelihood'] = np.expand_dims(sess.run(log_likelihood, feed_dict=self.feed_dict), axis=0) return dict_to_dataset(data, library=self.tfp, coords=self.coords, dims =dims)
sample_stats_to_xarray
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.model_fn is None or self.observed is None:\n return None\n log_likelihood = []\n sample_size = self.posterior[0].shape[0]\n ...
def x(self): return InferenceData(**{'posterior': self.posterior_to_xarray(), 'sample_stats': self.sample_stats_to_xarray(), 'posterior_predictive': self.posterior_predictive_to_xarray(), 'observed_data': self.observed_data_to_xarray()})
to_inference_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return InferenceData(**{'posterior': self.posterior_to_xarray(),\n 'sample_stats': self.sample_stats_to_xarray(),\n 'posterior_predic...
def x(trace=None, *, prior=None, posterior_predictive=None, coords=None, dims=None): return PyMC3Converter(trace=trace, prior=prior, posterior_predictive= posterior_predictive, coords=coords, dims=dims).to_inference_data()
from_pymc3
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(trace=None, *, prior=None, posterior_predictive=None, coords=None,\n dims=None):\n return PyMC3Converter(trace=trace, prior=prior, posterior_predictive=\...
def x(self): model = self.trace._straces[0].model if len(model.observed_RVs) != 1: return None, None elif self.dims is not None: coord_name = self.dims.get(model.observed_RVs[0].name) else: coord_name = None cached = [(var, var.logp_elemwise) for var in model.observed_RVs] def x(point): log_like_vals = [] for var, log_like in cached: log_like_val = log_like(point) if var.missing_values: log_like_val = log_like_val[~var.observations.mask] log_like_vals.append(log_like_val.ravel()) return np.concatenate(log_like_vals) chain_likelihoods = [] for chain in self.trace.chains: log_like = (log_likelihood_vals_point(point) for point in self. trace.points([chain])) chain_likelihoods.append(np.stack(log_like)) return np.stack(chain_likelihoods), coord_name
extract_log_likelihood
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n model = self.trace._straces[0].model\n if len(model.observed_RVs) != 1:\n return None, None\n elif self.dims is not None:\n coo...
def x(self): var_names = self.pymc3.util.get_default_varnames(self.trace.varnames, include_transformed=False) data = {} for var_name in var_names: data[var_name] = np.array(self.trace.get_values(var_name, combine= False, squeeze=False)) return dict_to_dataset(data, library=self.pymc3, coords=self.coords, dims=self.dims)
posterior_to_xarray
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n var_names = self.pymc3.util.get_default_varnames(self.trace.varnames,\n include_transformed=False)\n data = {}\n for var_name in var_n...
def x(self): rename_key = {'model_logp': 'lp'} data = {} for stat in self.trace.stat_names: name = rename_key.get(stat, stat) data[name] = np.array(self.trace.get_sampler_stats(stat, combine=False) ) log_likelihood, dims = self._extract_log_likelihood() if log_likelihood is not None: data['log_likelihood'] = log_likelihood dims = {'log_likelihood': dims} else: dims = None return dict_to_dataset(data, library=self.pymc3, dims=dims, coords=self .coords)
sample_stats_to_xarray
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n rename_key = {'model_logp': 'lp'}\n data = {}\n for stat in self.trace.stat_names:\n name = rename_key.get(stat, stat)\n data[n...
def x(self): data = {k: np.expand_dims(v, 0) for k, v in self.posterior_predictive. items()} return dict_to_dataset(data, library=self.pymc3, coords=self.coords, dims=self.dims)
posterior_predictive_to_xarray
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n data = {k: np.expand_dims(v, 0) for k, v in self.posterior_predictive.\n items()}\n return dict_to_dataset(data, library=self.pymc3, coor...
def x(self): return dict_to_dataset({k: np.expand_dims(v, 0) for k, v in self.prior. items()}, library=self.pymc3, coords=self.coords, dims=self.dims)
prior_to_xarray
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return dict_to_dataset({k: np.expand_dims(v, 0) for k, v in self.prior.\n items()}, library=self.pymc3, coords=self.coords, dims=self.dims)\...
def x(self): model = self.trace._straces[0].model observations = {obs.name: obs.observations for obs in model.observed_RVs} if self.dims is None: dims = {} else: dims = self.dims observed_data = {} for name, vals in observations.items(): vals = np.atleast_1d(vals) val_dims = dims.get(name) val_dims, coords = generate_dims_coords(vals.shape, name, dims= val_dims, coords=self.coords) coords = {key: xr.IndexVariable((key,), data=coords[key]) for key in val_dims} observed_data[name] = xr.DataArray(vals, dims=val_dims, coords=coords) return xr.Dataset(data_vars=observed_data, attrs=make_attrs(library= self.pymc3))
observed_data_to_xarray
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n model = self.trace._straces[0].model\n observations = {obs.name: obs.observations for obs in model.observed_RVs}\n if self.dims is None:\n ...
def x(data, var_names=None, coords=None, figsize=None, textsize=None, credible_interval=0.94, round_to=1, point_estimate='mean', rope=None, ref_val=None, kind='kde', bw=4.5, bins=None, ax=None, **kwargs): data = convert_to_dataset(data, group='posterior') var_names = _var_names(var_names, data) if coords is None: coords = {} plotters = list(xarray_var_iter(get_coords(data, coords), var_names= var_names, combined=True)) length_plotters = len(plotters) rows, cols = default_grid(length_plotters) figsize, ax_labelsize, titlesize, xt_labelsize, _linewidth, _ = ( _scale_fig_size(figsize, textsize, rows, cols)) kwargs.setdefault('linewidth', _linewidth) if ax is None: _, ax = _create_axes_grid(length_plotters, rows, cols, figsize= figsize, squeeze=False, constrained_layout=True) for (var_name, selection, x), ax_ in zip(plotters, np.ravel(ax)): _plot_posterior_op(x.flatten(), var_name, selection, ax=ax_, bw=bw, bins=bins, kind=kind, point_estimate=point_estimate, round_to= round_to, credible_interval=credible_interval, ref_val=ref_val, rope=rope, ax_labelsize=ax_labelsize, xt_labelsize=xt_labelsize, **kwargs) ax_.set_title(make_label(var_name, selection), fontsize=titlesize, wrap=True) return ax
plot_posterior
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(data, var_names=None, coords=None, figsize=None, textsize=None,\n credible_interval=0.94, round_to=1, point_estimate='mean', rope=None,\n ref_val=None, k...
def x(energy): energy_mat = np.atleast_2d(energy) num = np.square(np.diff(energy_mat, axis=1)).mean(axis=1) den = np.var(energy_mat, axis=1) return num / den
bfmi
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(energy):\n energy_mat = np.atleast_2d(energy)\n num = np.square(np.diff(energy_mat, axis=1)).mean(axis=1)\n den = np.var(energy_mat, axis=1)\n retu...
def x(dataset_dict, ic='waic', method='BB-pseudo-BMA', b_samples=1000, alpha=1, seed=None, scale='deviance'): names = list(dataset_dict.keys()) scale = scale.lower() if scale == 'log': scale_value = 1 ascending = False else: if scale == 'negative_log': scale_value = -1 else: scale_value = -2 ascending = True if ic == 'waic': ic_func = waic df_comp = pd.DataFrame(index=names, columns=['waic', 'p_waic', 'd_waic', 'weight', 'se', 'dse', 'warning', 'waic_scale']) scale_col = 'waic_scale' elif ic == 'loo': ic_func = loo df_comp = pd.DataFrame(index=names, columns=['loo', 'p_loo', 'd_loo', 'weight', 'se', 'dse', 'warning', 'loo_scale']) scale_col = 'loo_scale' else: raise NotImplementedError( 'The information criterion {} is not supported.'.format(ic)) if method.lower() not in ['stacking', 'bb-pseudo-bma', 'pseudo-bma']: raise ValueError('The method {}, to compute weights, is not supported.' .format(method)) ic_se = '{}_se'.format(ic) p_ic = 'p_{}'.format(ic) ic_i = '{}_i'.format(ic) ics = pd.DataFrame() names = [] for name, dataset in dataset_dict.items(): names.append(name) ics = ics.append([ic_func(dataset, pointwise=True, scale=scale)]) ics.index = names ics.sort_values(by=ic, inplace=True, ascending=ascending) if method.lower() == 'stacking': rows, cols, ic_i_val = _ic_matrix(ics, ic_i) exp_ic_i = np.exp(ic_i_val / scale_value) last_col = cols - 1 def x(weights): return np.concatenate((weights, [max(1.0 - np.sum(weights), 0.0)])) def x(weights): w_full = w_fuller(weights) score = 0.0 for i in range(rows): score += np.log(np.dot(exp_ic_i[i], w_full)) return -score def x(weights): w_full = w_fuller(weights) grad = np.zeros(last_col) for k in range(last_col - 1): for i in range(rows): grad[k] += (exp_ic_i[i, k] - exp_ic_i[i, last_col] ) / np.dot(exp_ic_i[i], w_full) return -grad theta = np.full(last_col, 1.0 / cols) bounds = [(0.0, 1.0) for _ in range(last_col)] constraints = [{'type': 'ineq', 'fun': lambda x: 1.0 - np.sum(x)}, {'type': 'ineq', 'fun': np.sum}] weights = minimize(fun=log_score, x0=theta, jac=gradient, bounds= bounds, constraints=constraints) weights = w_fuller(weights['x']) ses = ics[ic_se] elif method.lower() == 'bb-pseudo-bma': rows, cols, ic_i_val = _ic_matrix(ics, ic_i) ic_i_val = ic_i_val * rows b_weighting = st.dirichlet.rvs(alpha=[alpha] * rows, size=b_samples, random_state=seed) weights = np.zeros((b_samples, cols)) z_bs = np.zeros_like(weights) for i in range(b_samples): z_b = np.dot(b_weighting[i], ic_i_val) u_weights = np.exp((z_b - np.min(z_b)) / scale_value) z_bs[i] = z_b weights[i] = u_weights / np.sum(u_weights) weights = weights.mean(axis=0) ses = pd.Series(z_bs.std(axis=0), index=names) elif method.lower() == 'pseudo-bma': min_ic = ics.iloc[0][ic] z_rv = np.exp((ics[ic] - min_ic) / scale_value) weights = z_rv / np.sum(z_rv) ses = ics[ic_se] if np.any(weights): min_ic_i_val = ics[ic_i].iloc[0] for idx, val in enumerate(ics.index): res = ics.loc[val] if scale_value < 0: diff = res[ic_i] - min_ic_i_val else: diff = min_ic_i_val - res[ic_i] d_ic = np.sum(diff) d_std_err = np.sqrt(len(diff) * np.var(diff)) std_err = ses.loc[val] weight = weights[idx] df_comp.at[val] = res[ic], res[p_ic ], d_ic, weight, std_err, d_std_err, res['warning'], res[ scale_col] return df_comp.sort_values(by=ic, ascending=ascending)
compare
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(dataset_dict, ic='waic', method='BB-pseudo-BMA', b_samples=1000,\n alpha=1, seed=None, scale='deviance'):\n names = list(dataset_dict.keys())\n scale ...
def x(ics, ic_i): cols, _ = ics.shape rows = len(ics[ic_i].iloc[0]) ic_i_val = np.zeros((rows, cols)) for idx, val in enumerate(ics.index): ic = ics.loc[val][ic_i] if len(ic) != rows: raise ValueError( 'The number of observations should be the same across all models' ) ic_i_val[:, idx] = ic return rows, cols, ic_i_val
ic_matrix
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ics, ic_i):\n cols, _ = ics.shape\n rows = len(ics[ic_i].iloc[0])\n ic_i_val = np.zeros((rows, cols))\n for idx, val in enumerate(ics.index):\n ...
def x(x, credible_interval=0.94, circular=False): if x.ndim > 1: hpd_array = np.array([hpd(row, credible_interval=credible_interval, circular=circular) for row in x.T]) return hpd_array x = x.copy() len_x = len(x) if circular: mean = st.circmean(x, high=np.pi, low=-np.pi) x = x - mean x = np.arctan2(np.sin(x), np.cos(x)) x = np.sort(x) interval_idx_inc = int(np.floor(credible_interval * len_x)) n_intervals = len_x - interval_idx_inc interval_width = x[interval_idx_inc:] - x[:n_intervals] if len(interval_width) == 0: raise ValueError( 'Too few elements for interval calculation. Check that credible_interval meets condition 0 =< credible_interval < 1' ) min_idx = np.argmin(interval_width) hdi_min = x[min_idx] hdi_max = x[min_idx + interval_idx_inc] if circular: hdi_min = hdi_min + mean hdi_max = hdi_max + mean hdi_min = np.arctan2(np.sin(hdi_min), np.cos(hdi_min)) hdi_max = np.arctan2(np.sin(hdi_max), np.cos(hdi_max)) return np.array([hdi_min, hdi_max])
hpd
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(x, credible_interval=0.94, circular=False):\n if x.ndim > 1:\n hpd_array = np.array([hpd(row, credible_interval=credible_interval,\n circu...
def x(ary, *, b=None, b_inv=None, axis=None, keepdims=False, out=None, copy =True): ary = np.asarray(ary) if ary.dtype.kind == 'i': ary = ary.astype(np.float64) dtype = ary.dtype.type shape = ary.shape shape_len = len(shape) if isinstance(axis, Sequence): axis = tuple(axis_i if axis_i >= 0 else shape_len + axis_i for axis_i in axis) agroup = axis else: axis = axis if axis is None or axis >= 0 else shape_len + axis agroup = axis, shape_max = tuple(1 for _ in shape) if axis is None else tuple(1 if i in agroup else d for i, d in enumerate(shape)) if out is None: if not keepdims: out_shape = tuple() if axis is None else tuple(d for i, d in enumerate(shape) if i not in agroup) else: out_shape = shape_max out = np.empty(out_shape, dtype=dtype) if b_inv == 0: return np.full_like(out, np.inf, dtype=dtype) if out.shape else np.inf if b_inv is None and b == 0: return np.full_like(out, -np.inf) if out.shape else -np.inf ary_max = np.empty(shape_max, dtype=dtype) ary.max(axis=axis, keepdims=True, out=ary_max) if copy: ary = ary.copy() ary -= ary_max np.exp(ary, out=ary) ary.sum(axis=axis, keepdims=keepdims, out=out) np.log(out, out=out) if b_inv is not None: ary_max -= np.log(b_inv) elif b: ary_max += np.log(b) out += ary_max.squeeze() if not keepdims else ary_max return out if out.shape else dtype(out)
logsumexp
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ary, *, b=None, b_inv=None, axis=None, keepdims=False, out=None, copy\n =True):\n ary = np.asarray(ary)\n if ary.dtype.kind == 'i':\n ary = ary...
def x(data, pointwise=False, reff=None, scale='deviance'): inference_data = convert_to_inference_data(data) for group in ('posterior', 'sample_stats'): if not hasattr(inference_data, group): raise TypeError('Must be able to extract a {group}group from data!' .format(group=group)) if 'log_likelihood' not in inference_data.sample_stats: raise TypeError('Data must include log_likelihood in sample_stats') posterior = inference_data.posterior log_likelihood = inference_data.sample_stats.log_likelihood n_samples = log_likelihood.chain.size * log_likelihood.draw.size new_shape = (n_samples,) + log_likelihood.shape[2:] log_likelihood = log_likelihood.values.reshape(*new_shape) if scale.lower() == 'deviance': scale_value = -2 elif scale.lower() == 'log': scale_value = 1 elif scale.lower() == 'negative_log': scale_value = -1 else: raise TypeError( 'Valid scale values are "deviance", "log", "negative_log"') if reff is None: n_chains = len(posterior.chain) if n_chains == 1: reff = 1.0 else: ess = effective_sample_size(posterior) reff = np.hstack([ess[v].values.flatten() for v in ess.data_vars] ).mean() / n_samples log_weights, pareto_shape = psislw(-log_likelihood, reff) log_weights += log_likelihood warn_mg = 0 if np.any(pareto_shape > 0.7): warnings.warn( """Estimated shape parameter of Pareto distribution is greater than 0.7 for one or more samples. You should consider using a more robust model, this is because importance sampling is less likely to work well if the marginal posterior and LOO posterior are very different. This is more likely to happen with a non-robust model and highly influential observations.""" ) warn_mg = 1 loo_lppd_i = scale_value * _logsumexp(log_weights, axis=0) loo_lppd = loo_lppd_i.sum() loo_lppd_se = (len(loo_lppd_i) * np.var(loo_lppd_i)) ** 0.5 lppd = np.sum(_logsumexp(log_likelihood, axis=0, b_inv=log_likelihood. shape[0])) p_loo = lppd - loo_lppd / scale_value if pointwise: if np.equal(loo_lppd, loo_lppd_i).all(): warnings.warn( """The point-wise LOO is the same with the sum LOO, please double check the Observed RV in your model to make sure it returns element-wise logp. """ ) return pd.Series(data=[loo_lppd, loo_lppd_se, p_loo, warn_mg, loo_lppd_i, pareto_shape, scale], index=['loo', 'loo_se', 'p_loo', 'warning', 'loo_i', 'pareto_k', 'loo_scale']) else: return pd.Series(data=[loo_lppd, loo_lppd_se, p_loo, warn_mg, scale ], index=['loo', 'loo_se', 'p_loo', 'warning', 'loo_scale'])
loo
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(data, pointwise=False, reff=None, scale='deviance'):\n inference_data = convert_to_inference_data(data)\n for group in ('posterior', 'sample_stats'):\n ...
def x(log_weights, reff=1.0): rows, cols = log_weights.shape log_weights_out = np.copy(log_weights, order='F') kss = np.empty(cols) cutoff_ind = -int(np.ceil(min(rows / 5.0, 3 * (rows / reff) ** 0.5))) - 1 cutoffmin = np.log(np.finfo(float).tiny) k_min = 1.0 / 3 for i, x in enumerate(log_weights_out.T): x -= np.max(x) x_sort_ind = np.argsort(x) xcutoff = max(x[x_sort_ind[cutoff_ind]], cutoffmin) expxcutoff = np.exp(xcutoff) tailinds, = np.where(x > xcutoff) x_tail = x[tailinds] tail_len = len(x_tail) if tail_len <= 4: k = np.inf else: x_tail_si = np.argsort(x_tail) x_tail = np.exp(x_tail) - expxcutoff k, sigma = _gpdfit(x_tail[x_tail_si]) if k >= k_min: sti = np.arange(0.5, tail_len) / tail_len smoothed_tail = _gpinv(sti, k, sigma) smoothed_tail = np.log(smoothed_tail + expxcutoff) x[tailinds[x_tail_si]] = smoothed_tail x[x > 0] = 0 x -= _logsumexp(x) kss[i] = k return log_weights_out, kss
psislw
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(log_weights, reff=1.0):\n rows, cols = log_weights.shape\n log_weights_out = np.copy(log_weights, order='F')\n kss = np.empty(cols)\n cutoff_ind = ...
def x(x): prior_bs = 3 prior_k = 10 len_x = len(x) m_est = 30 + int(len_x ** 0.5) b_ary = 1 - np.sqrt(m_est / (np.arange(1, m_est + 1, dtype=float) - 0.5)) b_ary /= prior_bs * x[int(len_x / 4 + 0.5) - 1] b_ary += 1 / x[-1] k_ary = np.log1p(-b_ary[:, None] * x).mean(axis=1) len_scale = len_x * (np.log(-(b_ary / k_ary)) - k_ary - 1) weights = 1 / np.exp(len_scale - len_scale[:, None]).sum(axis=1) real_idxs = weights >= 10 * np.finfo(float).eps if not np.all(real_idxs): weights = weights[real_idxs] b_ary = b_ary[real_idxs] weights /= weights.sum() b_post = np.sum(b_ary * weights) k_post = np.log1p(-b_post * x).mean() k_post = (len_x * k_post + prior_k * 0.5) / (len_x + prior_k) sigma = -k_post / b_post return k_post, sigma
gpdfit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(x):\n prior_bs = 3\n prior_k = 10\n len_x = len(x)\n m_est = 30 + int(len_x ** 0.5)\n b_ary = 1 - np.sqrt(m_est / (np.arange(1, m_est + 1, dtype...
def x(probs, kappa, sigma): x = np.full_like(probs, np.nan) if sigma <= 0: return x ok = (probs > 0) & (probs < 1) if np.all(ok): if np.abs(kappa) < np.finfo(float).eps: x = -np.log1p(-probs) else: x = np.expm1(-kappa * np.log1p(-probs)) / kappa x *= sigma else: if np.abs(kappa) < np.finfo(float).eps: x[ok] = -np.log1p(-probs[ok]) else: x[ok] = np.expm1(-kappa * np.log1p(-probs[ok])) / kappa x *= sigma x[probs == 0] = 0 if kappa >= 0: x[probs == 1] = np.inf else: x[probs == 1] = -sigma / kappa return x
gpinv
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(probs, kappa, sigma):\n x = np.full_like(probs, np.nan)\n if sigma <= 0:\n return x\n ok = (probs > 0) & (probs < 1)\n if np.all(ok):\n ...
def x(y_true, y_pred): if y_pred.ndim == 1: var_y_est = np.var(y_pred) var_e = np.var(y_true - y_pred) else: var_y_est = np.var(y_pred.mean(0)) var_e = np.var(y_true - y_pred, 0) r_squared = var_y_est / (var_y_est + var_e) return pd.Series([np.mean(r_squared), np.std(r_squared)], index=['r2', 'r2_std'])
r2_score
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(y_true, y_pred):\n if y_pred.ndim == 1:\n var_y_est = np.var(y_pred)\n var_e = np.var(y_true - y_pred)\n else:\n var_y_est = np.var(...
def x(data, var_names=None, fmt='wide', round_to=2, include_circ=None, stat_funcs=None, extend=True, credible_interval=0.94, order='C'): posterior = convert_to_dataset(data, group='posterior') var_names = _var_names(var_names, posterior) posterior = posterior if var_names is None else posterior[var_names] fmt_group = 'wide', 'long', 'xarray' if not isinstance(fmt, str) or fmt.lower() not in fmt_group: raise TypeError("Invalid format: '{}'! Formatting options are: {}". format(fmt, fmt_group)) unpack_order_group = 'C', 'F' if not isinstance(order, str) or order.upper() not in unpack_order_group: raise TypeError("Invalid order: '{}'! Unpacking options are: {}". format(order, unpack_order_group)) alpha = 1 - credible_interval metrics = [] metric_names = [] if stat_funcs is not None: if isinstance(stat_funcs, dict): for stat_func_name, stat_func in stat_funcs.items(): metrics.append(xr.apply_ufunc(_make_ufunc(stat_func), posterior, input_core_dims=(('chain', 'draw'),))) metric_names.append(stat_func_name) else: for stat_func in stat_funcs: metrics.append(xr.apply_ufunc(_make_ufunc(stat_func), posterior, input_core_dims=(('chain', 'draw'),))) metric_names.append(stat_func.__name__) if extend: metrics.append(posterior.mean(dim=('chain', 'draw'))) metric_names.append('mean') metrics.append(posterior.std(dim=('chain', 'draw'))) metric_names.append('sd') metrics.append(xr.apply_ufunc(_make_ufunc(_mc_error), posterior, input_core_dims=(('chain', 'draw'),))) metric_names.append('mc error') metrics.append(xr.apply_ufunc(_make_ufunc(hpd, index=0, credible_interval=credible_interval), posterior, input_core_dims=(('chain', 'draw'),))) metric_names.append('hpd {:g}%'.format(100 * alpha / 2)) metrics.append(xr.apply_ufunc(_make_ufunc(hpd, index=1, credible_interval=credible_interval), posterior, input_core_dims=(('chain', 'draw'),))) metric_names.append('hpd {:g}%'.format(100 * (1 - alpha / 2))) if include_circ: metrics.append(xr.apply_ufunc(_make_ufunc(st.circmean, high=np.pi, low=-np.pi), posterior, input_core_dims=(('chain', 'draw'),))) metric_names.append('circular mean') metrics.append(xr.apply_ufunc(_make_ufunc(st.circstd, high=np.pi, low=-np.pi), posterior, input_core_dims=(('chain', 'draw'),))) metric_names.append('circular standard deviation') metrics.append(xr.apply_ufunc(_make_ufunc(_mc_error, circular=True), posterior, input_core_dims=(('chain', 'draw'),))) metric_names.append('circular mc error') metrics.append(xr.apply_ufunc(_make_ufunc(hpd, index=0, credible_interval=credible_interval, circular=True), posterior, input_core_dims=(('chain', 'draw'),))) metric_names.append('circular hpd {:.2%}'.format(alpha / 2)) metrics.append(xr.apply_ufunc(_make_ufunc(hpd, index=1, credible_interval=credible_interval, circular=True), posterior, input_core_dims=(('chain', 'draw'),))) metric_names.append('circular hpd {:.2%}'.format(1 - alpha / 2)) if len(posterior.chain) > 1: metrics.append(effective_sample_size(posterior, var_names=var_names)) metric_names.append('ess') metrics.append(rhat(posterior, var_names=var_names)) metric_names.append('r_hat') joined = xr.concat(metrics, dim='metric').assign_coords(metric=metric_names ) if fmt.lower() == 'wide': dfs = [] for var_name, values in joined.data_vars.items(): if len(values.shape[1:]): metric = list(values.metric.values) data_dict = OrderedDict() for idx in np.ndindex(values.shape[1:] if order == 'C' else values.shape[1:][::-1]): if order == 'F': idx = tuple(idx[::-1]) ser = pd.Series(values[Ellipsis, *idx].values, index=metric ) key = '{}[{}]'.format(var_name, ','.join(map(str, idx))) data_dict[key] = ser df = pd.DataFrame.from_dict(data_dict, orient='index') df = df.loc[list(data_dict.keys())] else: df = values.to_dataframe() df.index = list(df.index) df = df.T dfs.append(df) summary_df = pd.concat(dfs, sort=False) elif fmt.lower() == 'long': df = joined.to_dataframe().reset_index().set_index('metric') df.index = list(df.index) summary_df = df else: summary_df = joined return summary_df.round(round_to)
summary
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(data, var_names=None, fmt='wide', round_to=2, include_circ=None,\n stat_funcs=None, extend=True, credible_interval=0.94, order='C'):\n posterior = conver...
def x(func, index=Ellipsis, **kwargs): def x(ary): target = np.empty(ary.shape[:-2]) for idx in np.ndindex(target.shape): target[idx] = np.asarray(func(ary[idx].ravel(), **kwargs))[index] return target return _ufunc
make_ufunc
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(func, index=Ellipsis, **kwargs):\n\n def x(ary):\n target = np.empty(ary.shape[:-2])\n for idx in np.ndindex(target.shape):\n targe...
def x(x, batches=5, circular=False): if x.ndim > 1: dims = np.shape(x) trace = np.transpose([t.ravel() for t in x]) return np.reshape([_mc_error(t, batches) for t in trace], dims[1:]) else: if batches == 1: if circular: std = st.circstd(x, high=np.pi, low=-np.pi) else: std = np.std(x) return std / np.sqrt(len(x)) batched_traces = np.resize(x, (batches, int(len(x) / batches))) if circular: means = st.circmean(batched_traces, high=np.pi, low=-np.pi, axis=1) std = st.circstd(means, high=np.pi, low=-np.pi) else: means = np.mean(batched_traces, 1) std = np.std(means) return std / np.sqrt(batches)
mc_error
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(x, batches=5, circular=False):\n if x.ndim > 1:\n dims = np.shape(x)\n trace = np.transpose([t.ravel() for t in x])\n return np.reshape...
def x(data, pointwise=False, scale='deviance'): inference_data = convert_to_inference_data(data) for group in ('sample_stats',): if not hasattr(inference_data, group): raise TypeError( 'Must be able to extract a {group} group from data!'.format (group=group)) if 'log_likelihood' not in inference_data.sample_stats: raise TypeError('Data must include log_likelihood in sample_stats') log_likelihood = inference_data.sample_stats.log_likelihood if scale.lower() == 'deviance': scale_value = -2 elif scale.lower() == 'log': scale_value = 1 elif scale.lower() == 'negative_log': scale_value = -1 else: raise TypeError( 'Valid scale values are "deviance", "log", "negative_log"') n_samples = log_likelihood.chain.size * log_likelihood.draw.size new_shape = (n_samples,) + log_likelihood.shape[2:] log_likelihood = log_likelihood.values.reshape(*new_shape) lppd_i = _logsumexp(log_likelihood, axis=0, b_inv=log_likelihood.shape[0]) vars_lpd = np.var(log_likelihood, axis=0) warn_mg = 0 if np.any(vars_lpd > 0.4): warnings.warn( """For one or more samples the posterior variance of the log predictive densities exceeds 0.4. This could be indication of WAIC starting to fail see http://arxiv.org/abs/1507.04544 for details """ ) warn_mg = 1 waic_i = scale_value * (lppd_i - vars_lpd) waic_se = (len(waic_i) * np.var(waic_i)) ** 0.5 waic_sum = np.sum(waic_i) p_waic = np.sum(vars_lpd) if pointwise: if np.equal(waic_sum, waic_i).all(): warnings.warn( """The point-wise WAIC is the same with the sum WAIC, please double check the Observed RV in your model to make sure it returns element-wise logp. """ ) return pd.Series(data=[waic_sum, waic_se, p_waic, warn_mg, waic_i, scale], index=['waic', 'waic_se', 'p_waic', 'warning', 'waic_i', 'waic_scale']) else: return pd.Series(data=[waic_sum, waic_se, p_waic, warn_mg, scale], index=['waic', 'waic_se', 'p_waic', 'warning', 'waic_scale'])
waic
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(data, pointwise=False, scale='deviance'):\n inference_data = convert_to_inference_data(data)\n for group in ('sample_stats',):\n if not hasattr(in...