source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(self, function, args): param = self.model.param_array fs = [] for p in self.chain: self.model.param_array = p fs.append(function(*args)) self.model.param_array = param return fs
predict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, function, args):\n param = self.model.param_array\n fs = []\n for p in self.chain:\n self.model.param_array = p\n fs.append(functi...
def x(self, X, X2, target): if X2 is None: X2 = X FX = np.column_stack([f(X) for f in self.F]) FX2 = np.column_stack([f(X2) for f in self.F]) DER = np.zeros((self.n, self.n, self.n)) for i in range(self.n): DER[i, i, i] = np.sqrt(self.weights[i]) dw = self.variance * mdot(FX, DER, self.G_1, np.diag(np.sqrt(self. weights)), FX2.T) dv = mdot(FX, np.diag(np.sqrt(self.weights)), self.G_1, np.diag(np.sqrt (self.weights)), FX2.T) np.add(target[:, :, 0], np.transpose(dv, (0, 2, 1)), target[:, :, 0]) np.add(target[:, :, 1:], np.transpose(dw, (0, 2, 1)), target[:, :, 1:])
param_grad_helper
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, X2, target):\n if X2 is None:\n X2 = X\n FX = np.column_stack([f(X) for f in self.F])\n FX2 = np.column_stack([f(X2) for f in self.F])...
def x(self, X, Y): self.update_model(False) self.set_Y(Y) self.set_X(X) self.update_model(True)
set_XY
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, Y):\n self.update_model(False)\n self.set_Y(Y)\n self.set_X(X)\n self.update_model(True)\n", "role": "user" }, { "content": "s...
def x(self, X): assert isinstance(X, np.ndarray) state = self.update_model() self.update_model(False) self.X = ObsAr(X) self.update_model(state)
set_X
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X):\n assert isinstance(X, np.ndarray)\n state = self.update_model()\n self.update_model(False)\n self.X = ObsAr(X)\n self.update_model(st...
def x(self, Y): assert isinstance(Y, (np.ndarray, ObsAr)) state = self.update_model() self.update_model(False) if self.normalizer is not None: self.normalizer.scale_by(Y) self.Y_normalized = ObsAr(self.normalizer.normalize(Y)) self.Y = Y else: self.Y = ObsAr(Y) if isinstance(Y, np.ndarray) else Y self.Y_normalized = self.Y self.update_model(state)
set_Y
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Y):\n assert isinstance(Y, (np.ndarray, ObsAr))\n state = self.update_model()\n self.update_model(False)\n if self.normalizer is not None:\n ...
def x(self): self.posterior, self._log_marginal_likelihood, grad_dict = (self. inference_method.inference(self.kern, self.X, self.Y_normalized, self.nu + 2 + np.finfo(float).eps, self.mean_function)) self.kern.update_gradients_full(grad_dict['dL_dK'], self.X) if self.mean_function is not None: self.mean_function.update_gradients(grad_dict['dL_dm'], self.X) self.nu.gradient = grad_dict['dL_dnu']
parameters_changed
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.posterior, self._log_marginal_likelihood, grad_dict = (self.\n inference_method.inference(self.kern, self.X, self.Y_normalized, \n ...
def x(self, Xnew, full_cov=False, kern=None): mu, var = self.posterior._raw_predict(kern=self.kern if kern is None else kern, Xnew=Xnew, pred_var=self._predictive_variable, full_cov=full_cov) if self.mean_function is not None: mu += self.mean_function.f(Xnew) return mu, var
raw_predict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Xnew, full_cov=False, kern=None):\n mu, var = self.posterior._raw_predict(kern=self.kern if kern is None else\n kern, Xnew=Xnew, pred_var=self....
def x(self, Xnew, full_cov=False, kern=None, **kwargs): return self.predict_noiseless(Xnew, full_cov=full_cov, kern=kern)
predict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Xnew, full_cov=False, kern=None, **kwargs):\n return self.predict_noiseless(Xnew, full_cov=full_cov, kern=kern)\n", "role": "user" }, { "c...
def x(self, Xnew, full_cov=False, kern=None): mu, var = self._raw_predict(Xnew, full_cov=full_cov, kern=kern) if self.normalizer is not None: mu, var = self.normalizer.inverse_mean(mu ), self.normalizer.inverse_variance(var) return mu, var
predict_noiseless
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Xnew, full_cov=False, kern=None):\n mu, var = self._raw_predict(Xnew, full_cov=full_cov, kern=kern)\n if self.normalizer is not None:\n mu, ...
def x(self, X, quantiles=(2.5, 97.5), kern=None, **kwargs): mu, var = self._raw_predict(X, full_cov=False, kern=kern) quantiles = [(stats.t.ppf(q / 100.0, self.nu + 2 + self.num_data) * np. sqrt(var) + mu) for q in quantiles] if self.normalizer is not None: quantiles = [self.normalizer.inverse_mean(q) for q in quantiles] return quantiles
predict_quantiles
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, quantiles=(2.5, 97.5), kern=None, **kwargs):\n mu, var = self._raw_predict(X, full_cov=False, kern=kern)\n quantiles = [(stats.t.ppf(q / 100.0, ...
def x(self, X, size=10, full_cov=False, Y_metadata=None, likelihood=None, **predict_kwargs): return self.posterior_samples_f(X, size, full_cov=full_cov, ** predict_kwargs)
posterior_samples
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, size=10, full_cov=False, Y_metadata=None, likelihood=None,\n **predict_kwargs):\n return self.posterior_samples_f(X, size, full_cov=full_cov, **...
def x(self, X, size=10, full_cov=True, **predict_kwargs): mu, var = self._raw_predict(X, full_cov=full_cov, **predict_kwargs) if self.normalizer is not None: mu, var = self.normalizer.inverse_mean(mu ), self.normalizer.inverse_variance(var) def x(m, v): nu = self.nu + 2 + self.num_data v = np.diag(v.flatten()) if not full_cov else v Z = np.random.multivariate_normal(np.zeros(X.shape[0]), v, size).T g = np.tile(np.random.gamma(nu / 2.0, 2.0 / nu, size), (X.shape[0], 1)) return m + Z / np.sqrt(g) if self.output_dim == 1: return sim_one_dim(mu, var) else: fsim = np.empty((self.output_dim, self.num_data, size)) for d in range(self.output_dim): if full_cov and var.ndim == 3: fsim[d] = sim_one_dim(mu[:, d], var[:, :, d]) elif not full_cov and var.ndim == 2: fsim[d] = sim_one_dim(mu[:, d], var[:, d]) else: fsim[d] = sim_one_dim(mu[:, d], var) return fsim
posterior_samples_f
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, size=10, full_cov=True, **predict_kwargs):\n mu, var = self._raw_predict(X, full_cov=full_cov, **predict_kwargs)\n if self.normalizer is not Non...
def x(A, offset=0): from numpy.lib.stride_tricks import as_strided assert A.ndim == 2, 'only implemented for 2 dimensions' assert A.shape[0] == A.shape[1 ], 'attempting to get the view of non-square matrix?!' if offset > 0: return as_strided(A[0, offset:], shape=(A.shape[0] - offset,), strides=((A.shape[0] + 1) * A.itemsize,)) elif offset < 0: return as_strided(A[-offset:, 0], shape=(A.shape[0] + offset,), strides=((A.shape[0] + 1) * A.itemsize,)) else: return as_strided(A, shape=(A.shape[0],), strides=((A.shape[0] + 1) * A.itemsize,))
view
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, offset=0):\n from numpy.lib.stride_tricks import as_strided\n assert A.ndim == 2, 'only implemented for 2 dimensions'\n assert A.shape[0] == A.shap...
def x(A, b, offset=0): return _diag_ufunc(A, b, offset, np.multiply)
times
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, b, offset=0):\n return _diag_ufunc(A, b, offset, np.multiply)\n", "role": "user" }, { "content": "times", "role": "assistant" } ]
def x(A, b, offset=0): return _diag_ufunc(A, b, offset, np.divide)
divide
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, b, offset=0):\n return _diag_ufunc(A, b, offset, np.divide)\n", "role": "user" }, { "content": "divide", "role": "assistant" } ]
def x(A, b, offset=0): return _diag_ufunc(A, b, offset, np.add)
add
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, b, offset=0):\n return _diag_ufunc(A, b, offset, np.add)\n", "role": "user" }, { "content": "add", "role": "assistant" } ]
def x(A, b, offset=0): return _diag_ufunc(A, b, offset, np.subtract)
subtract
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, b, offset=0):\n return _diag_ufunc(A, b, offset, np.subtract)\n", "role": "user" }, { "content": "subtract", "role": "assistant" } ]
def x(input_dict): import copy input_dict = copy.deepcopy(input_dict) inference_class = input_dict.pop('class') import GPy inference_class = eval(inference_class) return inference_class._build_from_input_dict(inference_class, input_dict)
from_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(input_dict):\n import copy\n input_dict = copy.deepcopy(input_dict)\n inference_class = input_dict.pop('class')\n import GPy\n inference_class =...
def x(self): input_dict = super(Add, self)._save_to_input_dict() input_dict['class'] = str('GPy.kern.Add') return input_dict
to_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n input_dict = super(Add, self)._save_to_input_dict()\n input_dict['class'] = str('GPy.kern.Add')\n return input_dict\n", "role": "user" ...
def x(self, dL_dK, X, X2=None): target = np.zeros(X.shape) [target.__iadd__(p.gradients_X(dL_dK, X, X2)) for p in self.parts] return target
gradients_X
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dL_dK, X, X2=None):\n target = np.zeros(X.shape)\n [target.__iadd__(p.gradients_X(dL_dK, X, X2)) for p in self.parts]\n return target\n", "r...
def x(self): import scipy.linalg as la F = None L = None Qc = None H = None Pinf = None P0 = None dF = None dQc = None dPinf = None dP0 = None n = 0 nq = 0 nd = 0 for p in self.parts: Ft, Lt, Qct, Ht, Pinft, P0t, dFt, dQct, dPinft, dP0t = p.sde() F = la.block_diag(F, Ft) if F is not None else Ft L = la.block_diag(L, Lt) if L is not None else Lt Qc = la.block_diag(Qc, Qct) if Qc is not None else Qct H = np.hstack((H, Ht)) if H is not None else Ht Pinf = la.block_diag(Pinf, Pinft) if Pinf is not None else Pinft P0 = la.block_diag(P0, P0t) if P0 is not None else P0t if dF is not None: dF = np.pad(dF, ((0, dFt.shape[0]), (0, dFt.shape[1]), (0, dFt. shape[2])), 'constant', constant_values=0) dF[-dFt.shape[0]:, -dFt.shape[1]:, -dFt.shape[2]:] = dFt else: dF = dFt if dQc is not None: dQc = np.pad(dQc, ((0, dQct.shape[0]), (0, dQct.shape[1]), (0, dQct.shape[2])), 'constant', constant_values=0) dQc[-dQct.shape[0]:, -dQct.shape[1]:, -dQct.shape[2]:] = dQct else: dQc = dQct if dPinf is not None: dPinf = np.pad(dPinf, ((0, dPinft.shape[0]), (0, dPinft.shape[1 ]), (0, dPinft.shape[2])), 'constant', constant_values=0) dPinf[-dPinft.shape[0]:, -dPinft.shape[1]:, -dPinft.shape[2]: ] = dPinft else: dPinf = dPinft if dP0 is not None: dP0 = np.pad(dP0, ((0, dP0t.shape[0]), (0, dP0t.shape[1]), (0, dP0t.shape[2])), 'constant', constant_values=0) dP0[-dP0t.shape[0]:, -dP0t.shape[1]:, -dP0t.shape[2]:] = dP0t else: dP0 = dP0t n += Ft.shape[0] nq += Qct.shape[0] nd += dFt.shape[2] assert F.shape[0] == n and F.shape[1 ] == n, 'SDE add: Check of F Dimensions failed' assert L.shape[0] == n and L.shape[1 ] == nq, 'SDE add: Check of L Dimensions failed' assert Qc.shape[0] == nq and Qc.shape[1 ] == nq, 'SDE add: Check of Qc Dimensions failed' assert H.shape[0] == 1 and H.shape[1 ] == n, 'SDE add: Check of H Dimensions failed' assert Pinf.shape[0] == n and Pinf.shape[1 ] == n, 'SDE add: Check of Pinf Dimensions failed' assert P0.shape[0] == n and P0.shape[1 ] == n, 'SDE add: Check of P0 Dimensions failed' assert dF.shape[0] == n and dF.shape[1] == n and dF.shape[2 ] == nd, 'SDE add: Check of dF Dimensions failed' assert dQc.shape[0] == nq and dQc.shape[1] == nq and dQc.shape[2 ] == nd, 'SDE add: Check of dQc Dimensions failed' assert dPinf.shape[0] == n and dPinf.shape[1] == n and dPinf.shape[2 ] == nd, 'SDE add: Check of dPinf Dimensions failed' assert dP0.shape[0] == n and dP0.shape[1] == n and dP0.shape[2 ] == nd, 'SDE add: Check of dP0 Dimensions failed' return F, L, Qc, H, Pinf, P0, dF, dQc, dPinf, dP0
sde
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n import scipy.linalg as la\n F = None\n L = None\n Qc = None\n H = None\n Pinf = None\n P0 = None\n dF = None\n dQc = None\n...
def x(p, labels, names=['1', '0'], threshold=0.5, show=True): assert p.size == labels.size, 'Arrays p and labels have different dimensions.' decision = np.ones((labels.size, 1)) decision[p < threshold] = 0 diff = decision - labels false_0 = diff[diff == -1].size false_1 = diff[diff == 1].size true_1 = np.sum(decision[diff == 0]) true_0 = labels.size - true_1 - false_0 - false_1 error = (false_1 + false_0) / np.float(labels.size) if show: print(100.0 - error * 100, '% instances correctly classified') print('%-10s| %-10s| %-10s| ' % ('', names[0], names[1])) print('----------|------------|------------|') print('%-10s| %-10s| %-10s| ' % (names[0], true_1, false_0)) print('%-10s| %-10s| %-10s| ' % (names[1], false_1, true_0)) return error, true_1, false_1, true_0, false_0
conf_matrix
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(p, labels, names=['1', '0'], threshold=0.5, show=True):\n assert p.size == labels.size, 'Arrays p and labels have different dimensions.'\n decision = np....
def x(self, save_data=True): input_dict = super(SparseGP, self).to_dict(save_data) input_dict['class'] = 'GPy.core.SparseGP' input_dict['Z'] = self.Z.tolist() return input_dict
to_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, save_data=True):\n input_dict = super(SparseGP, self).to_dict(save_data)\n input_dict['class'] = 'GPy.core.SparseGP'\n input_dict['Z'] = self.Z....
def x(A): if A.flags['F_CONTIGUOUS']: return A if A.flags['C_CONTIGUOUS']: return A.T else: return np.asfortranarray(A)
force_F_ordered_symmetric
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A):\n if A.flags['F_CONTIGUOUS']:\n return A\n if A.flags['C_CONTIGUOUS']:\n return A.T\n else:\n return np.asfortranarray(A)\n",...
def x(A, B, lower=1, trans=0, unitdiag=0): A = np.asfortranarray(A) return lapack.dtrtrs(A, B, lower=lower, trans=trans, unitdiag=unitdiag)
dtrtrs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, B, lower=1, trans=0, unitdiag=0):\n A = np.asfortranarray(A)\n return lapack.dtrtrs(A, B, lower=lower, trans=trans, unitdiag=unitdiag)\n", "role":...
def x(A, B, lower=1): A = force_F_ordered(A) return lapack.dpotrs(A, B, lower=lower)
dpotrs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, B, lower=1):\n A = force_F_ordered(A)\n return lapack.dpotrs(A, B, lower=lower)\n", "role": "user" }, { "content": "dpotrs", "role": "...
def x(A, lower=1): A = force_F_ordered(A) R, info = lapack.dpotri(A, lower=lower) symmetrify(R) return R, info
dpotri
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, lower=1):\n A = force_F_ordered(A)\n R, info = lapack.dpotri(A, lower=lower)\n symmetrify(R)\n return R, info\n", "role": "user" }, { ...
def x(A): L = jitchol(A) logdetA = 2 * sum(np.log(np.diag(L))) return logdetA
pddet
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A):\n L = jitchol(A)\n logdetA = 2 * sum(np.log(np.diag(L)))\n return logdetA\n", "role": "user" }, { "content": "pddet", "role": "ass...
def x(*args): if len(args) == 1: return args[0] elif len(args) == 2: return _mdot_r(args[0], args[1]) else: return _mdot_r(args[:-1], args[-1])
mdot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*args):\n if len(args) == 1:\n return args[0]\n elif len(args) == 2:\n return _mdot_r(args[0], args[1])\n else:\n return _mdot_r(...
def x(a, b): if type(a) == tuple: if len(a) > 1: a = mdot(*a) else: a = a[0] if type(b) == tuple: if len(b) > 1: b = mdot(*b) else: b = b[0] return np.dot(a, b)
mdot_r
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(a, b):\n if type(a) == tuple:\n if len(a) > 1:\n a = mdot(*a)\n else:\n a = a[0]\n if type(b) == tuple:\n if l...
def x(A, *args): L = jitchol(A, *args) logdet = 2.0 * np.sum(np.log(np.diag(L))) Li = dtrtri(L) Ai, _ = dpotri(L, lower=1) symmetrify(Ai) return Ai, L, Li, logdet
pdinv
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, *args):\n L = jitchol(A, *args)\n logdet = 2.0 * np.sum(np.log(np.diag(L)))\n Li = dtrtri(L)\n Ai, _ = dpotri(L, lower=1)\n symmetrify(Ai)\n ...
def x(A): N = A.shape[-1] chols = [jitchol(A[:, :, i]) for i in range(N)] halflogdets = [np.sum(np.log(np.diag(L[0]))) for L in chols] invs = [dpotri(L[0], True)[0] for L in chols] invs = [(np.triu(I) + np.triu(I, 1).T) for I in invs] return np.dstack(invs), np.array(halflogdets)
multiple_pdinv
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A):\n N = A.shape[-1]\n chols = [jitchol(A[:, :, i]) for i in range(N)]\n halflogdets = [np.sum(np.log(np.diag(L[0]))) for L in chols]\n invs = [dp...
def x(Y, input_dim): if not np.allclose(Y.mean(axis=0), 0.0): print('Y is not zero mean, centering it locally (GPy.util.linalg.pca)') Z = linalg.svd(Y - Y.mean(axis=0), full_matrices=False) [X, W] = [Z[0][:, 0:input_dim], np.dot(np.diag(Z[1]), Z[2]).T[:, 0: input_dim]] v = X.std(axis=0) X /= v W *= v return X, W.T
pca
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(Y, input_dim):\n if not np.allclose(Y.mean(axis=0), 0.0):\n print('Y is not zero mean, centering it locally (GPy.util.linalg.pca)')\n Z = linalg.s...
def x(Y, Q, iterations=100): from numpy.ma import dot as madot N, D = Y.shape W = np.random.randn(D, Q) * 0.001 Y = np.ma.masked_invalid(Y, copy=0) mu = Y.mean(0) Ycentered = Y - mu try: for _ in range(iterations): exp_x = np.asarray_chkfinite(np.linalg.solve(W.T.dot(W), madot( W.T, Ycentered.T))).T W = np.asarray_chkfinite(np.linalg.solve(exp_x.T.dot(exp_x), madot(exp_x.T, Ycentered))).T except np.linalg.linalg.LinAlgError: pass return np.asarray_chkfinite(exp_x), np.asarray_chkfinite(W)
ppca
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(Y, Q, iterations=100):\n from numpy.ma import dot as madot\n N, D = Y.shape\n W = np.random.randn(D, Q) * 0.001\n Y = np.ma.masked_invalid(Y, copy=...
def x(mat, out=None): if mat.dtype != 'float64' or len(mat.shape) != 2: return np.dot(mat, mat.T) nn = mat.shape[0] if out is None: out = np.zeros((nn, nn)) else: assert out.dtype == 'float64' assert out.shape == (nn, nn) assert 8 in out.strides out[:] = 0.0 mat = np.asfortranarray(mat) out = blas.dsyrk(alpha=1.0, a=mat, beta=0.0, c=out, overwrite_c=1, trans=0, lower=0) symmetrify(out, upper=True) return np.ascontiguousarray(out)
tdot_blas
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(mat, out=None):\n if mat.dtype != 'float64' or len(mat.shape) != 2:\n return np.dot(mat, mat.T)\n nn = mat.shape[0]\n if out is None:\n ...
def x(A, x, alpha=1.0): At = blas.dsyr(lower=0, x=x, a=A, alpha=alpha, overwrite_a=False) A[:] = At symmetrify(A, upper=True)
DSYR_blas
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, x, alpha=1.0):\n At = blas.dsyr(lower=0, x=x, a=A, alpha=alpha, overwrite_a=False)\n A[:] = At\n symmetrify(A, upper=True)\n", "role": "user" ...
def x(A, x, alpha=1.0): A += alpha * np.dot(x[:, None], x[None, :])
DSYR_numpy
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, x, alpha=1.0):\n A += alpha * np.dot(x[:, None], x[None, :])\n", "role": "user" }, { "content": "DSYR_numpy", "role": "assistant" } ]
def x(A, upper=False): if use_linalg_cython: _symmetrify_cython(A, upper) else: _symmetrify_numpy(A, upper)
symmetrify
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, upper=False):\n if use_linalg_cython:\n _symmetrify_cython(A, upper)\n else:\n _symmetrify_numpy(A, upper)\n", "role": "user" }, ...
def x(L, X, transpose='left'): if transpose == 'left': tmp, _ = dtrtrs(L, X, lower=1, trans=1) return dtrtrs(L, tmp.T, lower=1, trans=1)[0].T else: tmp, _ = dtrtrs(L, X, lower=1, trans=0) return dtrtrs(L, tmp.T, lower=1, trans=0)[0].T
backsub_both_sides
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(L, X, transpose='left'):\n if transpose == 'left':\n tmp, _ = dtrtrs(L, X, lower=1, trans=1)\n return dtrtrs(L, tmp.T, lower=1, trans=1)[0].T\...
def x(A, B): return A.dot(B.reshape(B.shape[0], -1)).reshape(A.shape[0], B.shape[1], B.shape[2])
ij_jlk_to_ilk
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, B):\n return A.dot(B.reshape(B.shape[0], -1)).reshape(A.shape[0], B.shape[1],\n B.shape[2])\n", "role": "user" }, { "content": "ij_jlk...
def x(A, B): res = np.zeros((A.shape[0], B.shape[1])) [np.add(np.dot(A[:, :, k], B[:, :, k]), res, out=res) for k in range(B. shape[-1])] return res
ijk_jlk_to_il
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, B):\n res = np.zeros((A.shape[0], B.shape[1]))\n [np.add(np.dot(A[:, :, k], B[:, :, k]), res, out=res) for k in range(B.\n shape[-1])]\n ret...
def x(A, B): res = np.zeros((A.shape[-1], A.shape[0], B.shape[0])) [np.dot(A[:, :, i], B[:, :, i].T, out=res[i, :, :]) for i in range(A. shape[-1])] res = res.swapaxes(0, 2).swapaxes(0, 1) return res
ijk_ljk_to_ilk
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, B):\n res = np.zeros((A.shape[-1], A.shape[0], B.shape[0]))\n [np.dot(A[:, :, i], B[:, :, i].T, out=res[i, :, :]) for i in range(A.\n shape[-1]...
def x(file_or_path): import sys import inspect sys.modules['GPy.kern._src'] = kern.src for name, module in inspect.getmembers(kern.src): if not name.startswith('_'): sys.modules['GPy.kern._src.{}'.format(name)] = module sys.modules['GPy.inference.optimization'] = inference.optimization import paramz return paramz.load(file_or_path)
load
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(file_or_path):\n import sys\n import inspect\n sys.modules['GPy.kern._src'] = kern.src\n for name, module in inspect.getmembers(kern.src):\n ...
def x(feval, fmin, fmax, tol1=1e-05, tol2=1e-05): XK = np.array([-0.991455371120813, -0.949107912342759, - 0.864864423359769, -0.741531185599394, -0.586087235467691, - 0.405845151377397, -0.207784955007898, 0.0, 0.207784955007898, 0.405845151377397, 0.586087235467691, 0.741531185599394, 0.864864423359769, 0.949107912342759, 0.991455371120813]) WK = np.array([0.022935322010529, 0.063092092629979, 0.10479001032225, 0.140653259715525, 0.169004726639267, 0.190350578064785, 0.204432940075298, 0.209482141084728, 0.204432940075298, 0.190350578064785, 0.169004726639267, 0.140653259715525, 0.10479001032225, 0.063092092629979, 0.022935322010529]) WG = np.array([0.12948496616887, 0.279705391489277, 0.381830050505119, 0.417959183673469, 0.381830050505119, 0.279705391489277, 0.12948496616887]) NK = WK.size G = np.arange(2, NK, 2) tol1 = 0.0001 tol2 = 0.0001 Subs = np.array([[fmin], [fmax]]) NF = feval(np.zeros(1)).size Q = np.zeros(NF) neval = 0 while Subs.size > 0: Subs = getSubs(Subs, XK) M = (Subs[1, :] - Subs[0, :]) / 2 C = (Subs[1, :] + Subs[0, :]) / 2 NM = M.size x = XK[:, None] * M + C x = x.flatten() FV = feval(x) Q1 = np.zeros((NF, NM)) Q2 = np.zeros((NF, NM)) Q1 = np.dot(FV.reshape(NF, NK, NM).swapaxes(2, 1), WK) * M Q2 = np.dot(FV.reshape(NF, NK, NM).swapaxes(2, 1)[:, :, 1::2], WG) * M ind = np.nonzero(np.logical_or(np.max(np.abs(Q1 - Q2), 0) < tol1, Subs[1, :] - Subs[0, :] < tol2))[0] Q = Q + np.sum(Q1[:, ind], axis=1) Subs = np.delete(Subs, ind, axis=1) return Q
quadvgk
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(feval, fmin, fmax, tol1=1e-05, tol2=1e-05):\n XK = np.array([-0.991455371120813, -0.949107912342759, -\n 0.864864423359769, -0.741531185599394, -0.58...
def x(f, fmin=-np.inf, fmax=np.inf, difftol=0.1): difftol = 0.0001 def x(r): r2 = np.square(r) x = r / (1 - r2) dx_dr = (1 + r2) / (1 - r2) ** 2 return f(x) * dx_dr integrand = quadvgk(trans_func, -1.0, 1.0, difftol, difftol) return integrand
quadgk_int
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(f, fmin=-np.inf, fmax=np.inf, difftol=0.1):\n difftol = 0.0001\n\n def x(r):\n r2 = np.square(r)\n x = r / (1 - r2)\n dx_dr = (1 + r...
def x(self, varp_list): if self.group_spike: self._b_prob_all = 1.0 - param_to_array(varp_list[0].gamma_group) [np.multiply(self._b_prob_all, 1.0 - vp.gamma_group, self. _b_prob_all) for vp in varp_list[1:]] else: self._b_prob_all = 1.0 - param_to_array(varp_list[0].binary_prob) [np.multiply(self._b_prob_all, 1.0 - vp.binary_prob, self. _b_prob_all) for vp in varp_list[1:]]
update_inernal
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, varp_list):\n if self.group_spike:\n self._b_prob_all = 1.0 - param_to_array(varp_list[0].gamma_group)\n [np.multiply(self._b_prob_all, ...
def x(self, varp_list): self._b_prob_all = 1.0 - param_to_array(varp_list[0].gamma_group) [np.multiply(self._b_prob_all, 1.0 - vp.gamma_group, self._b_prob_all) for vp in varp_list[1:]]
update_inernal
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, varp_list):\n self._b_prob_all = 1.0 - param_to_array(varp_list[0].gamma_group)\n [np.multiply(self._b_prob_all, 1.0 - vp.gamma_group, self._b_prob...
def x(self, X, X2, target): AX = np.dot(X, self.transform) if X2 is None: X2 = X AX2 = AX else: AX2 = np.dot(X2, self.transform) self.k.K(X, X2, target) self.k.K(AX, X2, target) self.k.K(X, AX2, target) self.k.K(AX, AX2, target)
K
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, X2, target):\n AX = np.dot(X, self.transform)\n if X2 is None:\n X2 = X\n AX2 = AX\n else:\n AX2 = np.dot(X2, self.trans...
def x(self, dL_dK, X, X2, target): AX = np.dot(X, self.transform) if X2 is None: X2 = X ZX2 = AX else: AX2 = np.dot(X2, self.transform) self.k._param_grad_helper(dL_dK, X, X2, target) self.k._param_grad_helper(dL_dK, AX, X2, target) self.k._param_grad_helper(dL_dK, X, AX2, target) self.k._param_grad_helper(dL_dK, AX, AX2, target)
param_grad_helper
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dL_dK, X, X2, target):\n AX = np.dot(X, self.transform)\n if X2 is None:\n X2 = X\n ZX2 = AX\n else:\n AX2 = np.dot(X2, sel...
def x(self, X, target): foo = np.zeros((X.shape[0], X.shape[0])) self.K(X, X, foo) target += np.diag(foo)
Kdiag
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, target):\n foo = np.zeros((X.shape[0], X.shape[0]))\n self.K(X, X, foo)\n target += np.diag(foo)\n", "role": "user" }, { "content...
def x(self, dL_dpsi2, Z, mu, S, target): self._psi_computations(Z, mu, S) d_var = 2.0 * self._psi2 / self.variance d_length = -2.0 * self._psi2[:, :, :, None] * (self._psi2_Zdist_sq * self._psi2_denom + self._psi2_mudist_sq + S[:, None, None, :] * self.inv_lengthscale2) / (self.inv_lengthscale * self._psi2_denom) target[0] += np.sum(dL_dpsi2 * d_var) dpsi2_dlength = d_length * dL_dpsi2[:, :, :, None] if not self.ARD: target[1] += dpsi2_dlength.sum() else: target[1:] += dpsi2_dlength.sum(0).sum(0).sum(0)
dpsi2_dtheta
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dL_dpsi2, Z, mu, S, target):\n self._psi_computations(Z, mu, S)\n d_var = 2.0 * self._psi2 / self.variance\n d_length = -2.0 * self._psi2[:, :, ...
def x(self, dL_dpsi2, Z, mu, S, target_mu, target_S): self._psi_computations(Z, mu, S) tmp = self.inv_lengthscale2 * self._psi2[:, :, :, None] / self._psi2_denom target_mu += -2.0 * (dL_dpsi2[:, :, :, None] * tmp * self._psi2_mudist ).sum(1).sum(1) target_S += (dL_dpsi2[:, :, :, None] * tmp * (2.0 * self. _psi2_mudist_sq - 1)).sum(1).sum(1)
dpsi2_dmuS
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dL_dpsi2, Z, mu, S, target_mu, target_S):\n self._psi_computations(Z, mu, S)\n tmp = self.inv_lengthscale2 * self._psi2[:, :, :, None] / self._psi2...
def x(self, kern, Z, variational_posterior, return_psi2_n=False): variance, lengthscale = kern.variance, kern.lengthscale N, M, Q = self.get_dimensions(Z, variational_posterior) self._initGPUCache(N, M, Q) self.sync_params(lengthscale, Z, variational_posterior.mean, variational_posterior.variance) psi1_gpu = self.gpuCache['psi1_gpu'] psi2_gpu = self.gpuCache['psi2_gpu'] psi2n_gpu = self.gpuCache['psi2n_gpu'] l_gpu = self.gpuCache['l_gpu'] Z_gpu = self.gpuCache['Z_gpu'] mu_gpu = self.gpuCache['mu_gpu'] S_gpu = self.gpuCache['S_gpu'] log_denom1_gpu = self.gpuCache['log_denom1_gpu'] log_denom2_gpu = self.gpuCache['log_denom2_gpu'] psi0 = np.empty((N,)) psi0[:] = variance self.g_psi1computations.prepared_call((self.blocknum, 1), (self. threadnum, 1, 1), psi1_gpu.gpudata, log_denom1_gpu.gpudata, np. float64(variance), l_gpu.gpudata, Z_gpu.gpudata, mu_gpu.gpudata, S_gpu.gpudata, np.int32(N), np.int32(M), np.int32(Q)) self.g_psi2computations.prepared_call((self.blocknum, 1), (self. threadnum, 1, 1), psi2_gpu.gpudata, psi2n_gpu.gpudata, log_denom2_gpu.gpudata, np.float64(variance), l_gpu.gpudata, Z_gpu. gpudata, mu_gpu.gpudata, S_gpu.gpudata, np.int32(N), np.int32(M), np.int32(Q)) if self.GPU_direct: return psi0, psi1_gpu, psi2_gpu elif return_psi2_n: return psi0, psi1_gpu.get(), psi2n_gpu.get() else: return psi0, psi1_gpu.get(), psi2_gpu.get()
psicomputations
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, kern, Z, variational_posterior, return_psi2_n=False):\n variance, lengthscale = kern.variance, kern.lengthscale\n N, M, Q = self.get_dimensions(Z, ...
def x(self, prior, warning=True): repriorized = self.unset_priors() self._add_to_index_operations(self.priors, repriorized, prior, warning) from paramz.domains import _REAL, _POSITIVE, _NEGATIVE if prior.domain is _POSITIVE: self.constrain_positive(warning) elif prior.domain is _NEGATIVE: self.constrain_negative(warning) elif prior.domain is _REAL: rav_i = self._raveled_index() assert all(all(False if c is __fixed__ else c.domain is _REAL for c in con) for con in self.constraints.properties_for(rav_i) ), 'Domain of prior and constraint have to match, please unconstrain if you REALLY wish to use this prior'
set_prior
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, prior, warning=True):\n repriorized = self.unset_priors()\n self._add_to_index_operations(self.priors, repriorized, prior, warning)\n from param...
def x(self): if self.priors.size == 0: return 0.0 x = self.param_array log_p = reduce(lambda a, b: a + b, (p.lnpdf(x[ind]).sum() for p, ind in self.priors.items()), 0) log_j = 0.0 priored_indexes = np.hstack([i for p, i in self.priors.items()]) for c, j in self.constraints.items(): if not isinstance(c, Transformation): continue for jj in j: if jj in priored_indexes: log_j += c.log_jacobian(x[jj]) return log_p + log_j
log_prior
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.priors.size == 0:\n return 0.0\n x = self.param_array\n log_p = reduce(lambda a, b: a + b, (p.lnpdf(x[ind]).sum() for p, ind i...
def x(self): if self.priors.size == 0: return 0.0 x = self.param_array ret = np.zeros(x.size) [np.put(ret, ind, p.lnpdf_grad(x[ind])) for p, ind in self.priors.items()] priored_indexes = np.hstack([i for p, i in self.priors.items()]) for c, j in self.constraints.items(): if not isinstance(c, Transformation): continue for jj in j: if jj in priored_indexes: ret[jj] += c.log_jacobian_grad(x[jj]) return ret
log_prior_gradients
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.priors.size == 0:\n return 0.0\n x = self.param_array\n ret = np.zeros(x.size)\n [np.put(ret, ind, p.lnpdf_grad(x[ind])) fo...
def x(self, data, train): if train: inputs, labels = data self.data_mean = inputs.mean(axis=0) self.data_std = inputs.std(axis=0) self.labels_mean = labels.mean(axis=0) self.labels_std = labels.std(axis=0) return (inputs - self.data_mean) / self.data_std, (labels - self. labels_mean) / self.labels_std else: return (data - self.data_mean) / self.data_std
preprocess
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data, train):\n if train:\n inputs, labels = data\n self.data_mean = inputs.mean(axis=0)\n self.data_std = inputs.std(axis=0)\n ...
def x(A, blocksizes, pagesizes=None): assert A.shape[0] == A.shape[1] and len(A.shape ) == 3, "can't blockify this non-square matrix, may need to use 2d version" N = np.sum(blocksizes) assert A.shape[0] == N, 'bad blocksizes' num_blocks = len(blocksizes) if pagesizes == None: pagesizes = range(A.shape[2]) num_pages = len(pagesizes) B = np.empty(shape=(num_blocks, num_blocks, num_pages), dtype=np.object) count_k = 0 for Bk in pagesizes: count_i = 0 for Bi, i in enumerate(blocksizes): count_j = 0 for Bj, j in enumerate(blocksizes): B[Bi, Bj, Bk] = A[count_i:count_i + i, count_j:count_j + j, Bk] count_j += j count_i += i return B
get_blocks_3d
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, blocksizes, pagesizes=None):\n assert A.shape[0] == A.shape[1] and len(A.shape\n ) == 3, \"can't blockify this non-square matrix, may need to use ...
def x(A, B, diagonal=False): assert A.dtype is np.dtype('object'), 'Must be a block matrix' assert B.dtype is np.dtype('object'), 'Must be a block matrix' assert A.shape == B.shape def x(C, D): Cshape = C.shape Dshape = D.shape if diagonal and (len(Cshape) == 1 or len(Dshape) == 1 or C.shape[0] != C.shape[1] or D.shape[0] != D.shape[1]): print('Broadcasting, C: {} D:{}'.format(C.shape, D.shape)) return C * D else: print('Dotting, C: {} C:{}'.format(C.shape, D.shape)) return np.dot(C, D) dot = np.vectorize(f, otypes=[np.object]) return dot(A, B)
block_dot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, B, diagonal=False):\n assert A.dtype is np.dtype('object'), 'Must be a block matrix'\n assert B.dtype is np.dtype('object'), 'Must be a block matrix'\...
def x(num_inducing=50, max_iters=100, kernel=None, optimize=True, plot=True): try: import pods except ImportError: raise ImportWarning( 'Need pods for example datasets. See https://github.com/sods/ods, or pip install pods.' ) data = pods.datasets.oil() X = data['X'] Xtest = data['Xtest'] Y = data['Y'][:, 0:1] Ytest = data['Ytest'][:, 0:1] Y[Y.flatten() == -1] = 0 Ytest[Ytest.flatten() == -1] = 0 m = GPy.models.SparseGPClassification(X, Y, kernel=kernel, num_inducing =num_inducing) m.Ytest = Ytest m['.*len'] = 10.0 if optimize: m.optimize(messages=1) print(m) probs = m.predict(Xtest)[0] GPy.util.classification.conf_matrix(probs, Ytest) return m
oil
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(num_inducing=50, max_iters=100, kernel=None, optimize=True, plot=True):\n try:\n import pods\n except ImportError:\n raise ImportWarning(\n...
def x(seed=default_seed, optimize=True, plot=True): try: import pods except ImportError: raise ImportWarning( 'Need pods for example datasets. See https://github.com/sods/ods, or pip install pods.' ) data = pods.datasets.toy_linear_1d_classification(seed=seed) Y = data['Y'][:, 0:1] Y[Y.flatten() == -1] = 0 m = GPy.models.GPClassification(data['X'], Y) if optimize: m.optimize() if plot: from matplotlib import pyplot as plt fig, axes = plt.subplots(2, 1) m.plot_f(ax=axes[0]) m.plot(ax=axes[1]) print(m) return m
toy_linear_1d_classification
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seed=default_seed, optimize=True, plot=True):\n try:\n import pods\n except ImportError:\n raise ImportWarning(\n 'Need pods for...
def x(seed=default_seed, optimize=True, plot=True): try: import pods except ImportError: print( 'pods unavailable, see https://github.com/sods/ods for example datasets' ) data = pods.datasets.toy_linear_1d_classification(seed=seed) Y = data['Y'][:, 0:1] Y[Y.flatten() == -1] = 0 likelihood = GPy.likelihoods.Bernoulli() laplace_inf = GPy.inference.latent_function_inference.Laplace() kernel = GPy.kern.RBF(1) m = GPy.core.GP(data['X'], Y, kernel=kernel, likelihood=likelihood, inference_method=laplace_inf) if optimize: try: m.optimize('scg', messages=1) except Exception as e: return m if plot: from matplotlib import pyplot as plt fig, axes = plt.subplots(2, 1) m.plot_f(ax=axes[0]) m.plot(ax=axes[1]) print(m) return m
toy_linear_1d_classification_laplace
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seed=default_seed, optimize=True, plot=True):\n try:\n import pods\n except ImportError:\n print(\n 'pods unavailable, see https...
def x(num_inducing=10, seed=default_seed, optimize=True, plot=True): try: import pods except ImportError: print( 'pods unavailable, see https://github.com/sods/ods for example datasets' ) import numpy as np data = pods.datasets.toy_linear_1d_classification(seed=seed) Y = data['Y'][:, 0:1] Y[Y.flatten() == -1] = 0 X = data['X'] X_var = np.random.uniform(0.3, 0.5, X.shape) m = GPy.models.SparseGPClassificationUncertainInput(X, X_var, Y, num_inducing=num_inducing) m['.*len'] = 4.0 if optimize: m.optimize() if plot: from matplotlib import pyplot as plt fig, axes = plt.subplots(2, 1) m.plot_f(ax=axes[0]) m.plot(ax=axes[1]) print(m) return m
sparse_toy_linear_1d_classification_uncertain_input
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(num_inducing=10, seed=default_seed, optimize=True, plot=True):\n try:\n import pods\n except ImportError:\n print(\n 'pods unava...
def x(seed=default_seed, max_iters=100, optimize=True, plot=True): try: import pods except ImportError: print( 'pods unavailable, see https://github.com/sods/ods for example datasets' ) data = pods.datasets.toy_linear_1d_classification(seed=seed) Y = data['Y'][:, 0:1] Y[Y.flatten() == -1] = 0 kernel = GPy.kern.RBF(1) likelihood = GPy.likelihoods.Bernoulli(gp_link=GPy.likelihoods. link_functions.Heaviside()) ep = GPy.inference.latent_function_inference.expectation_propagation.EP() m = GPy.core.GP(X=data['X'], Y=Y, kernel=kernel, likelihood=likelihood, inference_method=ep, name='gp_classification_heaviside') if optimize: for _ in range(5): m.optimize(max_iters=int(max_iters / 5)) print(m) if plot: from matplotlib import pyplot as plt fig, axes = plt.subplots(2, 1) m.plot_f(ax=axes[0]) m.plot(ax=axes[1]) print(m) return m
toy_heaviside
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seed=default_seed, max_iters=100, optimize=True, plot=True):\n try:\n import pods\n except ImportError:\n print(\n 'pods unavail...
def x(model_type='Full', num_inducing=10, seed=default_seed, kernel=None, optimize=True, plot=True): try: import pods except ImportError: print( 'pods unavailable, see https://github.com/sods/ods for example datasets' ) data = pods.datasets.crescent_data(seed=seed) Y = data['Y'] Y[Y.flatten() == -1] = 0 if model_type == 'Full': m = GPy.models.GPClassification(data['X'], Y, kernel=kernel) elif model_type == 'DTC': m = GPy.models.SparseGPClassification(data['X'], Y, kernel=kernel, num_inducing=num_inducing) m['.*len'] = 10.0 elif model_type == 'FITC': m = GPy.models.FITCClassification(data['X'], Y, kernel=kernel, num_inducing=num_inducing) m['.*len'] = 3.0 if optimize: m.optimize(messages=1) if plot: m.plot() print(m) return m
crescent_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model_type='Full', num_inducing=10, seed=default_seed, kernel=None,\n optimize=True, plot=True):\n try:\n import pods\n except ImportError:\n ...
def x(self, X, test_data=False): X_warped = X.copy() if test_data: X_normalized = (X - self.Xmin) / (self.Xmax - self.Xmin) else: X_normalized = self.X_normalized for i_seq, i_fea in enumerate(self.warping_indices): a, b = self.params[i_seq][0], self.params[i_seq][1] X_warped[:, i_fea] = 1 - np.power(1 - np.power(X_normalized[:, i_fea], a), b) return X_warped
f
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, test_data=False):\n X_warped = X.copy()\n if test_data:\n X_normalized = (X - self.Xmin) / (self.Xmax - self.Xmin)\n else:\n X_...
def x(self, X): grad = np.zeros(X.shape) for i_seq, i_fea in enumerate(self.warping_indices): a, b = self.params[i_seq][0], self.params[i_seq][1] grad[:, i_fea] = a * b * np.power(self.X_normalized[:, i_fea], a - 1 ) * np.power(1 - np.power(self.X_normalized[:, i_fea], a), b - 1 ) * self.scaling[i_fea] return grad
fgrad_X
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X):\n grad = np.zeros(X.shape)\n for i_seq, i_fea in enumerate(self.warping_indices):\n a, b = self.params[i_seq][0], self.params[i_seq][1]\...
def x(self, X, dL_dW): for i_seq, i_fea in enumerate(self.warping_indices): ai, bi = self.params[i_seq][0], self.params[i_seq][1] x_pow_a = np.power(self.X_normalized[:, i_fea], ai) dz_dai = bi * np.power(1 - x_pow_a, bi - 1) * x_pow_a * np.log(self .X_normalized[:, i_fea]) dz_dbi = -np.power(1 - x_pow_a, bi) * np.log(1 - x_pow_a) dL_dai = np.sum(dL_dW[:, i_fea] * dz_dai) dL_dbi = np.sum(dL_dW[:, i_fea] * dz_dbi) self.params[i_seq][0].gradient[:] = dL_dai self.params[i_seq][1].gradient[:] = dL_dbi
update_grads
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, dL_dW):\n for i_seq, i_fea in enumerate(self.warping_indices):\n ai, bi = self.params[i_seq][0], self.params[i_seq][1]\n x_pow_a = np...
def x(self, kern, X, Z, likelihood, Y, indexD, output_dim, Y_metadata=None, Lm=None, dL_dKmm=None, Kuu_sigma=None): input_dim = Z.shape[0] uncertain_inputs = isinstance(X, VariationalPosterior) beta = 1.0 / likelihood.variance if len(beta) == 1: beta = np.zeros(output_dim) + beta beta_exp = np.zeros(indexD.shape[0]) for d in range(output_dim): beta_exp[indexD == d] = beta[d] psi0, psi1, psi2 = self.gatherPsiStat(kern, X, Z, Y, beta, uncertain_inputs ) psi2_sum = (beta_exp[:, None, None] * psi2).sum(0) / output_dim Kmm = kern.K(Z).copy() if Kuu_sigma is not None: diag.add(Kmm, Kuu_sigma) else: diag.add(Kmm, self.const_jitter) Lm = jitchol(Kmm) logL = 0.0 dL_dthetaL = np.zeros(output_dim) dL_dKmm = np.zeros_like(Kmm) dL_dpsi0 = np.zeros_like(psi0) dL_dpsi1 = np.zeros_like(psi1) dL_dpsi2 = np.zeros_like(psi2) wv = np.empty((Kmm.shape[0], output_dim)) for d in range(output_dim): idx_d = indexD == d Y_d = Y[idx_d] N_d = Y_d.shape[0] beta_d = beta[d] psi2_d = psi2[idx_d].sum(0) * beta_d psi1Y = Y_d.T.dot(psi1[idx_d]) * beta_d psi0_d = psi0[idx_d].sum() * beta_d YRY_d = np.square(Y_d).sum() * beta_d LmInvPsi2LmInvT = backsub_both_sides(Lm, psi2_d, 'right') Lambda = np.eye(Kmm.shape[0]) + LmInvPsi2LmInvT LL = jitchol(Lambda) LmLL = Lm.dot(LL) b = dtrtrs(LmLL, psi1Y.T)[0].T bbt = np.square(b).sum() v = dtrtrs(LmLL, b.T, trans=1)[0].T LLinvPsi1TYYTPsi1LLinvT = tdot(b.T) tmp = -backsub_both_sides(LL, LLinvPsi1TYYTPsi1LLinvT) dL_dpsi2R = backsub_both_sides(Lm, tmp + np.eye(input_dim)) / 2 logL_R = -N_d * np.log(beta_d) logL += -(N_d * log_2_pi + logL_R + psi0_d - np.trace( LmInvPsi2LmInvT) + YRY_d - bbt) / 2.0 dL_dKmm += dL_dpsi2R - backsub_both_sides(Lm, LmInvPsi2LmInvT) / 2 dL_dthetaL[d:d + 1] = (YRY_d * beta_d + beta_d * psi0_d - N_d * beta_d ) / 2.0 - beta_d * (dL_dpsi2R * psi2_d).sum() - beta_d * np.trace( LLinvPsi1TYYTPsi1LLinvT) dL_dpsi0[idx_d] = -beta_d / 2.0 dL_dpsi1[idx_d] = beta_d * np.dot(Y_d, v) dL_dpsi2[idx_d] = beta_d * dL_dpsi2R wv[:, d] = v LmInvPsi2LmInvT = backsub_both_sides(Lm, psi2_sum, 'right') Lambda = np.eye(Kmm.shape[0]) + LmInvPsi2LmInvT LL = jitchol(Lambda) LmLL = Lm.dot(LL) logdet_L = 2.0 * np.sum(np.log(np.diag(LL))) dL_dpsi2R_common = dpotri(LmLL)[0] / -2.0 dL_dpsi2 += dL_dpsi2R_common[None, :, :] * beta_exp[:, None, None] for d in range(output_dim): dL_dthetaL[d] += (dL_dpsi2R_common * psi2[indexD == d].sum(0)).sum( ) * -beta[d] * beta[d] dL_dKmm += dL_dpsi2R_common * output_dim logL += -output_dim * logdet_L / 2.0 LLInvLmT = dtrtrs(LL, Lm.T)[0] cov = tdot(LLInvLmT.T) wd_inv = backsub_both_sides(Lm, np.eye(input_dim) - backsub_both_sides( LL, np.identity(input_dim), transpose='left'), transpose='left') post = Posterior(woodbury_inv=wd_inv, woodbury_vector=wv, K=Kmm, mean= None, cov=cov, K_chol=Lm) if not uncertain_inputs: dL_dpsi1 += (psi1[:, None, :] * dL_dpsi2).sum(2) * 2.0 if uncertain_inputs: grad_dict = {'dL_dKmm': dL_dKmm, 'dL_dpsi0': dL_dpsi0, 'dL_dpsi1': dL_dpsi1, 'dL_dpsi2': dL_dpsi2, 'dL_dthetaL': dL_dthetaL} else: grad_dict = {'dL_dKmm': dL_dKmm, 'dL_dKdiag': dL_dpsi0, 'dL_dKnm': dL_dpsi1, 'dL_dthetaL': dL_dthetaL} return post, logL, grad_dict
inference
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, kern, X, Z, likelihood, Y, indexD, output_dim, Y_metadata=None,\n Lm=None, dL_dKmm=None, Kuu_sigma=None):\n input_dim = Z.shape[0]\n uncertain_i...
def x(parameterized, fignum=None, ax=None, colors=None, figsize=(12, 6)): if ax is None: fig = pb.figure(num=fignum, figsize=figsize) if colors is None: from ..Tango import mediumList from itertools import cycle colors = cycle(mediumList) pb.clf() else: colors = iter(colors) lines = [] fills = [] bg_lines = [] means, variances = parameterized.mean.values, parameterized.variance.values x = np.arange(means.shape[0]) for i in range(means.shape[1]): if ax is None: a = fig.add_subplot(means.shape[1], 1, i + 1) elif isinstance(ax, (tuple, list)): a = ax[i] else: raise ValueError('Need one ax per latent dimension input_dim') bg_lines.append(a.plot(means, c='k', alpha=0.3)) lines.extend(a.plot(x, means.T[i], c=next(colors), label= '$\\mathbf{{X_{{{}}}}}$'.format(i))) fills.append(a.fill_between(x, means.T[i] - 2 * np.sqrt(variances.T [i]), means.T[i] + 2 * np.sqrt(variances.T[i]), facecolor=lines [-1].get_color(), alpha=0.3)) a.legend(borderaxespad=0.0) a.set_xlim(x.min(), x.max()) if i < means.shape[1] - 1: a.set_xticklabels('') pb.draw() a.figure.tight_layout(h_pad=0.01) return dict(lines=lines, fills=fills, bg_lines=bg_lines)
plot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(parameterized, fignum=None, ax=None, colors=None, figsize=(12, 6)):\n if ax is None:\n fig = pb.figure(num=fignum, figsize=figsize)\n if colors is...
def x(parameterized, fignum=None, ax=None, colors=None, side_by_side=True): if ax is None: if side_by_side: fig = pb.figure(num=fignum, figsize=(16, min(12, 2 * parameterized.mean.shape[1]))) else: fig = pb.figure(num=fignum, figsize=(8, min(12, 2 * parameterized.mean.shape[1]))) if colors is None: from ..Tango import mediumList from itertools import cycle colors = cycle(mediumList) pb.clf() else: colors = iter(colors) plots = [] means, variances, gamma = (parameterized.mean, parameterized.variance, parameterized.binary_prob) x = np.arange(means.shape[0]) for i in range(means.shape[1]): if side_by_side: sub1 = means.shape[1], 2, 2 * i + 1 sub2 = means.shape[1], 2, 2 * i + 2 else: sub1 = means.shape[1] * 2, 1, 2 * i + 1 sub2 = means.shape[1] * 2, 1, 2 * i + 2 a = fig.add_subplot(*sub1) a.plot(means, c='k', alpha=0.3) plots.extend(a.plot(x, means.T[i], c=next(colors), label= '$\\mathbf{{X_{{{}}}}}$'.format(i))) a.fill_between(x, means.T[i] - 2 * np.sqrt(variances.T[i]), means.T [i] + 2 * np.sqrt(variances.T[i]), facecolor=plots[-1]. get_color(), alpha=0.3) a.legend(borderaxespad=0.0) a.set_xlim(x.min(), x.max()) if i < means.shape[1] - 1: a.set_xticklabels('') a = fig.add_subplot(*sub2) a.bar(x, gamma[:, i], bottom=0.0, linewidth=1.0, width=1.0, align= 'center') a.set_xlim(x.min(), x.max()) a.set_ylim([0.0, 1.0]) pb.draw() fig.tight_layout(h_pad=0.01) return fig
plot_SpikeSlab
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(parameterized, fignum=None, ax=None, colors=None, side_by_side=True):\n if ax is None:\n if side_by_side:\n fig = pb.figure(num=fignum, fi...
def x(self, plot_limits=None, which_data='all', which_parts='all', resolution=None, levels=20, samples=0, fignum=None, ax=None, fixed_inputs=[], linecol=Tango.colorsHex['darkBlue']): if which_data == 'all': which_data = slice(None) if ax is None: fig = pb.figure(num=fignum) ax = fig.add_subplot(111) plotdims = self.input_dim - len(fixed_inputs) from ..gpy_plot.plot_util import x_frame1D, x_frame2D if plotdims == 1: Xu = self.X * self._Xscale + self._Xoffset fixed_dims = np.array([i for i, v in fixed_inputs]) freedim = np.setdiff1d(np.arange(self.input_dim), fixed_dims) Xnew, xmin, xmax = x_frame1D(Xu[:, freedim], plot_limits=plot_limits) Xgrid = np.empty((Xnew.shape[0], self.input_dim)) Xgrid[:, freedim] = Xnew for i, v in fixed_inputs: Xgrid[:, i] = v f = self.predict(Xgrid, which_parts=which_parts) for d in range(y.shape[1]): ax.plot(Xnew, f[:, d], edgecol=linecol) elif self.X.shape[1] == 2: resolution = resolution or 50 Xnew, _, _, xmin, xmax = x_frame2D(self.X, plot_limits, resolution) x, y = np.linspace(xmin[0], xmax[0], resolution), np.linspace(xmin[ 1], xmax[1], resolution) f = self.predict(Xnew, which_parts=which_parts) m = m.reshape(resolution, resolution).T ax.contour(x, y, f, levels, vmin=m.min(), vmax=m.max(), cmap=pb.cm.jet) ax.set_xlim(xmin[0], xmax[0]) ax.set_ylim(xmin[1], xmax[1]) else: raise NotImplementedError( 'Cannot define a frame with more than two input dimensions')
plot_mapping
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, plot_limits=None, which_data='all', which_parts='all',\n resolution=None, levels=20, samples=0, fignum=None, ax=None,\n fixed_inputs=[], linecol=Ta...
def x(self): F, L, Qc, H, P_inf, P0, dFt, dQct, dP_inft, dP0t = self.kern.sde() measurement_dim = self.output_dim grad_params_no = dFt.shape[2] + 1 dF = np.zeros([dFt.shape[0], dFt.shape[1], grad_params_no]) dQc = np.zeros([dQct.shape[0], dQct.shape[1], grad_params_no]) dP_inf = np.zeros([dP_inft.shape[0], dP_inft.shape[1], grad_params_no]) dP0 = np.zeros([dP0t.shape[0], dP0t.shape[1], grad_params_no]) dF[:, :, :-1] = dFt dQc[:, :, :-1] = dQct dP_inf[:, :, :-1] = dP_inft dP0[:, :, :-1] = dP0t dR = np.zeros([measurement_dim, measurement_dim, grad_params_no]) dR[:, :, -1] = np.eye(measurement_dim) if self.balance: F, L, Qc, H, P_inf, P0, dF, dQc, dP_inf, dP0 = ssm.balance_ss_model(F, L, Qc, H, P_inf, P0, dF, dQc, dP_inf, dP0) print('SSM parameters_changed balancing!') grad_calc_params = {} grad_calc_params['dP_inf'] = dP_inf grad_calc_params['dF'] = dF grad_calc_params['dQc'] = dQc grad_calc_params['dR'] = dR grad_calc_params['dP_init'] = dP0 kalman_filter_type = self.kalman_filter_type Y = self.Y if self.ts_number is None: Y.shape = self.num_data, 1 else: Y.shape = self.num_data, 1, self.ts_number (filter_means, filter_covs, log_likelihood, grad_log_likelihood, SmootherMatrObject) = (ssm.ContDescrStateSpace. cont_discr_kalman_filter(F, L, Qc, H, float(self.Gaussian_noise. variance), P_inf, self.X, Y, m_init=None, P_init=P0, p_kalman_filter_type=kalman_filter_type, calc_log_likelihood=True, calc_grad_log_likelihood=True, grad_params_no=grad_params_no, grad_calc_params=grad_calc_params)) if np.any(np.isfinite(log_likelihood) == False): print('State-Space: NaN valkues in the log_likelihood') if np.any(np.isfinite(grad_log_likelihood) == False): print('State-Space: NaN values in the grad_log_likelihood') grad_log_likelihood_sum = np.sum(grad_log_likelihood, axis=1) grad_log_likelihood_sum.shape = grad_log_likelihood_sum.shape[0], 1 self._log_marginal_likelihood = np.sum(log_likelihood, axis=1) self.likelihood.update_gradients(grad_log_likelihood_sum[-1, 0]) self.kern.sde_update_gradient_full(grad_log_likelihood_sum[:-1, 0])
parameters_changed
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n F, L, Qc, H, P_inf, P0, dFt, dQct, dP_inft, dP0t = self.kern.sde()\n measurement_dim = self.output_dim\n grad_params_no = dFt.shape[2] + 1\n ...
def x(self, Xnew=None, Ynew=None, filteronly=False, p_balance=False, **kw): if Ynew is None: Ynew = self.Y if Xnew is not None: X = np.vstack((self.X, Xnew)) Y = np.vstack((Ynew, np.nan * np.zeros(Xnew.shape))) predict_only_training = False else: X = self.X Y = Ynew predict_only_training = True _, return_index, return_inverse = np.unique(X, True, True) X = X[return_index] Y = Y[return_index] F, L, Qc, H, P_inf, P0, dF, dQc, dP_inf, dP0 = self.kern.sde() state_dim = F.shape[0] if p_balance == True: F, L, Qc, H, P_inf, P0, dF, dQc, dP_inf, dP0 = ssm.balance_ss_model(F, L, Qc, H, P_inf, P0, dF, dQc, dP_inf, dP0) print('SSM _raw_predict balancing!') kalman_filter_type = self.kalman_filter_type M, P, log_likelihood, grad_log_likelihood, SmootherMatrObject = (ssm. ContDescrStateSpace.cont_discr_kalman_filter(F, L, Qc, H, float( self.Gaussian_noise.variance), P_inf, X, Y, m_init=None, P_init=P0, p_kalman_filter_type=kalman_filter_type, calc_log_likelihood=False, calc_grad_log_likelihood=False)) if not filteronly: M, P = ssm.ContDescrStateSpace.cont_discr_rts_smoother(state_dim, M, P, p_dynamic_callables=SmootherMatrObject, X=X, F=F, L=L, Qc=Qc) M = M[1:, :, :] P = P[1:, :, :] M = M[return_inverse, :, :] P = P[return_inverse, :, :] if not predict_only_training: M = M[self.num_data:, :, :] P = P[self.num_data:, :, :] m = np.einsum('ijl,kj', M, H) m.shape = m.shape[0], m.shape[1] V = np.einsum('ij,ajk,kl', H, P, H.T) V.shape = V.shape[0], V.shape[1] return m, V
raw_predict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Xnew=None, Ynew=None, filteronly=False, p_balance=False, **kw):\n if Ynew is None:\n Ynew = self.Y\n if Xnew is not None:\n X = np.vs...
def x(self, Xnew=None, filteronly=False, include_likelihood=True, balance= None, **kw): if balance is None: p_balance = self.balance else: p_balance = balance m, V = self._raw_predict(Xnew, filteronly=filteronly, p_balance=p_balance) if include_likelihood: V += float(self.likelihood.variance) return m, V
predict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Xnew=None, filteronly=False, include_likelihood=True, balance=\n None, **kw):\n if balance is None:\n p_balance = self.balance\n else:\n ...
def x(self, Xnew=None, quantiles=(2.5, 97.5), balance=None, **kw): if balance is None: p_balance = self.balance else: p_balance = balance mu, var = self._raw_predict(Xnew, p_balance=p_balance) return [(stats.norm.ppf(q / 100.0) * np.sqrt(var + float(self. Gaussian_noise.variance)) + mu) for q in quantiles]
predict_quantiles
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Xnew=None, quantiles=(2.5, 97.5), balance=None, **kw):\n if balance is None:\n p_balance = self.balance\n else:\n p_balance = balance...
def x(Y, visualizer, frame_rate=30): for y in Y: visualizer.modify(y[None, :]) time.sleep(1.0 / float(frame_rate))
data_play
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(Y, visualizer, frame_rate=30):\n for y in Y:\n visualizer.modify(y[None, :])\n time.sleep(1.0 / float(frame_rate))\n", "role": "user" },...
def x(self, vals): self.vals = vals.view(np.ndarray).copy() y = self.model.predict(self.vals)[0] self.data_visualize.modify(y) self.latent_handle.set_data(self.vals[0, self.latent_index[0]], self. vals[0, self.latent_index[1]]) self.axes.figure.canvas.draw()
modify
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, vals):\n self.vals = vals.view(np.ndarray).copy()\n y = self.model.predict(self.vals)[0]\n self.data_visualize.modify(y)\n self.latent_handle...
def x(self, boundary=0.05): bs = [((self.vals[:, i].max() - self.vals[:, i].min()) * boundary) for i in range(3)] self.x_lim = np.array([self.vals[:, 0].min() - bs[0], self.vals[:, 0]. max() + bs[0]]) self.y_lim = np.array([self.vals[:, 1].min() - bs[1], self.vals[:, 1]. max() + bs[1]]) self.z_lim = np.array([self.vals[:, 2].min() - bs[2], self.vals[:, 2]. max() + bs[2]])
initialize_axes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, boundary=0.05):\n bs = [((self.vals[:, i].max() - self.vals[:, i].min()) * boundary) for\n i in range(3)]\n self.x_lim = np.array([self.vals...
def x(self): if self.padding > 0: channels = np.zeros((self.vals.shape[0], self.vals.shape[1] + self. padding)) channels[:, 0:self.vals.shape[0]] = self.vals else: channels = self.vals vals_mat = self.skel.to_xyz(channels.flatten()) self.vals = np.zeros_like(vals_mat) self.vals[:, 0] = vals_mat[:, 0].copy() self.vals[:, 1] = vals_mat[:, 2].copy() self.vals[:, 2] = vals_mat[:, 1].copy()
process_values
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.padding > 0:\n channels = np.zeros((self.vals.shape[0], self.vals.shape[1] + self.\n padding))\n channels[:, 0:sel...
def x(self, figure=None, col=1, row=1, projection='2d', xlabel=None, ylabel =None, zlabel=None, title=None, xlim=None, ylim=None, zlim=None, **kwargs): raise NotImplementedError( 'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library' )
new_canvas
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, figure=None, col=1, row=1, projection='2d', xlabel=None, ylabel\n =None, zlabel=None, title=None, xlim=None, ylim=None, zlim=None, **kwargs):\n rai...
def x(self, cavas, X, Y, Z=None, color=None, label=None, **kwargs): raise NotImplementedError( 'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library' )
plot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, cavas, X, Y, Z=None, color=None, label=None, **kwargs):\n raise NotImplementedError(\n 'Implement all plot functions in AbstractPlottingLibrary...
def x(self, canvas, X, Y, Z, color=None, label=None, **kwargs): raise NotImplementedError( 'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library' )
surface
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, canvas, X, Y, Z, color=None, label=None, **kwargs):\n raise NotImplementedError(\n 'Implement all plot functions in AbstractPlottingLibrary in ...
def x(self, canvas, X, Y, Z=None, color=None, vmin=None, vmax=None, label= None, **kwargs): raise NotImplementedError( 'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library' )
scatter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, canvas, X, Y, Z=None, color=None, vmin=None, vmax=None, label=\n None, **kwargs):\n raise NotImplementedError(\n 'Implement all plot functio...
def x(self, canvas, x, height, width=0.8, bottom=0, color=None, label=None, **kwargs): raise NotImplementedError( 'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library' )
barplot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, canvas, x, height, width=0.8, bottom=0, color=None, label=None,\n **kwargs):\n raise NotImplementedError(\n 'Implement all plot functions in...
def x(self, canvas, X, Y, error, color=None, label=None, **kwargs): raise NotImplementedError( 'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library' )
xerrorbar
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, canvas, X, Y, error, color=None, label=None, **kwargs):\n raise NotImplementedError(\n 'Implement all plot functions in AbstractPlottingLibrary...
def x(self, canvas, X, extent=None, label=None, vmin=None, vmax=None, **kwargs ): raise NotImplementedError( 'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library' )
imshow
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, canvas, X, extent=None, label=None, vmin=None, vmax=None, **kwargs\n ):\n raise NotImplementedError(\n 'Implement all plot functions in Abst...
def x(self, canvas, plot_function, extent=None, label=None, vmin=None, vmax =None, **kwargs): raise NotImplementedError( 'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library' )
imshow_interact
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, canvas, plot_function, extent=None, label=None, vmin=None, vmax\n =None, **kwargs):\n raise NotImplementedError(\n 'Implement all plot funct...
def x(self, canvas, X, Y, C, Z=None, color=None, label=None, **kwargs): raise NotImplementedError( 'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library' )
contour
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, canvas, X, Y, C, Z=None, color=None, label=None, **kwargs):\n raise NotImplementedError(\n 'Implement all plot functions in AbstractPlottingLib...
def x(self, canvas, X, lower, upper, color=None, label=None, **kwargs): raise NotImplementedError( 'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library' )
fill_between
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, canvas, X, lower, upper, color=None, label=None, **kwargs):\n raise NotImplementedError(\n 'Implement all plot functions in AbstractPlottingLib...
def x(self, rand_gen=None, *args, **kwargs): if rand_gen is None: rand_gen = np.random.normal x = rand_gen(*args, size=self._size_transformed(), **kwargs) updates = self.update_model() self.update_model(False) self.optimizer_array = x x = self.param_array.copy() [np.put(x, ind, p.rvs(ind.size)) for p, ind in self.priors.items() if not p is None] unfixlist = np.ones((self.size,), dtype=np.bool) from paramz.transformations import __fixed__ unfixlist[self.constraints[__fixed__]] = False self.param_array.flat[unfixlist] = x.view(np.ndarray).ravel()[unfixlist] self.update_model(updates)
randomize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, rand_gen=None, *args, **kwargs):\n if rand_gen is None:\n rand_gen = np.random.normal\n x = rand_gen(*args, size=self._size_transformed(), *...
def x(self, t, tprime, s, sprime, l): return 0.5 * l ** 2 * (self.g((t - sprime) / l) + self.g((tprime - s) / l) - self.g((t - tprime) / l) - self.g((s - sprime) / l))
k_xx
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, t, tprime, s, sprime, l):\n return 0.5 * l ** 2 * (self.g((t - sprime) / l) + self.g((tprime - s) /\n l) - self.g((t - tprime) / l) - self.g((s...
def x(self, X): K_xx = np.ones([X.shape[0], X.shape[0]]) for i, x in enumerate(X): for j, x2 in enumerate(X): for il, l in enumerate(self.lengthscale): idx = il * 2 K_xx[i, j] *= self.k_xx(x[idx], x2[idx], x[idx + 1], x2[idx + 1], l) return K_xx
calc_K_xx_wo_variance
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X):\n K_xx = np.ones([X.shape[0], X.shape[0]])\n for i, x in enumerate(X):\n for j, x2 in enumerate(X):\n for il, l in enumerate(...
def x(self, X): K_ff = np.ones(X.shape[0]) for i, x in enumerate(X): for il, l in enumerate(self.lengthscale): idx = il * 2 K_ff[i] *= self.k_ff(x[idx], x[idx], l) return K_ff * self.variances[0]
Kdiag
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X):\n K_ff = np.ones(X.shape[0])\n for i, x in enumerate(X):\n for il, l in enumerate(self.lengthscale):\n idx = il * 2\n ...
def x(self, target_param=None, verbose=False, step=1e-06, tolerance=0.001, block_indices=None, plot=False): try: import numdifftools as nd except: raise ImportError( "Don't have numdifftools package installed, it is not a GPy dependency as of yet, it is only used for hessian tests" ) if target_param: raise NotImplementedError( 'Only basic functionality is provided with this gradchecker') current_index = 0 for name, shape in zip(self.names, self.shapes): current_size = numpy.prod(shape) x = self.optimizer_array.copy() x = x[current_index:current_index + current_size].reshape(shape) analytic_hess = self._ddf(x) if analytic_hess.shape[1] == 1: analytic_hess = numpy.diagflat(analytic_hess) numeric_hess_partial = nd.Jacobian(self._df, vectorized=False) numeric_hess = numeric_hess_partial(x) check_passed = self.checkgrad_block(analytic_hess, numeric_hess, verbose=verbose, step=step, tolerance=tolerance, block_indices= block_indices, plot=plot) current_index += current_size return check_passed
checkgrad
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, target_param=None, verbose=False, step=1e-06, tolerance=0.001,\n block_indices=None, plot=False):\n try:\n import numdifftools as nd\n ex...
def x(self, analytic_hess, numeric_hess, verbose=False, step=1e-06, tolerance=0.001, block_indices=None, plot=False): if analytic_hess.dtype is np.dtype('object'): real_size = get_block_shapes(analytic_hess) num_elements = np.sum(real_size) if (num_elements, num_elements) == numeric_hess.shape: numeric_hess = get_blocks(numeric_hess, real_size) else: tmp_numeric_hess = get_blocks(np.zeros((num_elements, num_elements)), real_size) tmp_numeric_hess[block_indices] = numeric_hess.copy() numeric_hess = tmp_numeric_hess if block_indices is not None: analytic_hess = analytic_hess[block_indices] numeric_hess = numeric_hess[block_indices] elif analytic_hess.dtype is np.dtype('object'): analytic_hess = unblock(analytic_hess) numeric_hess = unblock(numeric_hess) ratio = numeric_hess / numpy.where(analytic_hess == 0, 1e-10, analytic_hess ) difference = numpy.abs(analytic_hess - numeric_hess) check_passed = numpy.all(numpy.abs(1 - ratio) < tolerance ) or numpy.allclose(numeric_hess, analytic_hess, atol=tolerance) if verbose: if block_indices: print('\nBlock {}'.format(block_indices)) else: print('\nAll blocks') header = ['Checked', 'Max-Ratio', 'Min-Ratio', 'Min-Difference', 'Max-Difference'] header_string = map(lambda x: ' | '.join(header), [header]) separator = '-' * len(header_string[0]) print('\n'.join([header_string[0], separator])) min_r = '%.6f' % float(numpy.min(ratio)) max_r = '%.6f' % float(numpy.max(ratio)) max_d = '%.6f' % float(numpy.max(difference)) min_d = '%.6f' % float(numpy.min(difference)) cols = [max_r, min_r, min_d, max_d] if check_passed: checked = '\x1b[92m True \x1b[0m' else: checked = '\x1b[91m False \x1b[0m' grad_string = '{} | {} | {} | {} | {} '.format(checked, cols[0], cols[1], cols[2], cols[3]) print(grad_string) if plot: from matplotlib import pyplot as pb fig, axes = pb.subplots(2, 2) max_lim = numpy.max(numpy.vstack((analytic_hess, numeric_hess))) min_lim = numpy.min(numpy.vstack((analytic_hess, numeric_hess))) msa = axes[0, 0].matshow(analytic_hess, vmin=min_lim, vmax=max_lim) axes[0, 0].set_title('Analytic hessian') axes[0, 0].xaxis.set_ticklabels([None]) axes[0, 0].yaxis.set_ticklabels([None]) axes[0, 0].xaxis.set_ticks([None]) axes[0, 0].yaxis.set_ticks([None]) msn = axes[0, 1].matshow(numeric_hess, vmin=min_lim, vmax=max_lim) pb.colorbar(msn, ax=axes[0, 1]) axes[0, 1].set_title('Numeric hessian') axes[0, 1].xaxis.set_ticklabels([None]) axes[0, 1].yaxis.set_ticklabels([None]) axes[0, 1].xaxis.set_ticks([None]) axes[0, 1].yaxis.set_ticks([None]) msr = axes[1, 0].matshow(ratio) pb.colorbar(msr, ax=axes[1, 0]) axes[1, 0].set_title('Ratio') axes[1, 0].xaxis.set_ticklabels([None]) axes[1, 0].yaxis.set_ticklabels([None]) axes[1, 0].xaxis.set_ticks([None]) axes[1, 0].yaxis.set_ticks([None]) msd = axes[1, 1].matshow(difference) pb.colorbar(msd, ax=axes[1, 1]) axes[1, 1].set_title('difference') axes[1, 1].xaxis.set_ticklabels([None]) axes[1, 1].yaxis.set_ticklabels([None]) axes[1, 1].xaxis.set_ticks([None]) axes[1, 1].yaxis.set_ticks([None]) if block_indices: fig.suptitle('Block: {}'.format(block_indices)) pb.show() return check_passed
checkgrad_block
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, analytic_hess, numeric_hess, verbose=False, step=1e-06,\n tolerance=0.001, block_indices=None, plot=False):\n if analytic_hess.dtype is np.dtype('o...
def x(self, target_param=None, verbose=False, step=1e-06, tolerance=0.001, block_indices=None, plot=False, super_plot=False): try: import numdifftools as nd except: raise ImportError( "Don't have numdifftools package installed, it is not a GPy dependency as of yet, it is only used for hessian tests" ) if target_param: raise NotImplementedError( 'Only basic functionality is provided with this gradchecker') current_index = 0 for name, n_shape in zip(self.names, self.shapes): current_size = numpy.prod(n_shape) x = self.optimizer_array.copy() x = x[current_index:current_index + current_size].reshape(n_shape) analytic_hess = self._ddf(x) numeric_hess_partial = nd.Jacobian(self._df, vectorized=True) numeric_hess = numeric_hess_partial(x) print('Done making numerical hessian') if analytic_hess.dtype is np.dtype('object'): blocksizes, pagesizes = get_block_shapes_3d(analytic_hess) real_block_size = np.sum(blocksizes) numeric_hess = numeric_hess.reshape(real_block_size, real_block_size, pagesizes) else: numeric_hess = numeric_hess.reshape(*analytic_hess.shape) check_passed = [False] * numeric_hess.shape[2] for block_ind in range(numeric_hess.shape[2]): p = True if plot and block_ind == numeric_hess.shape[2 ] - 1 or super_plot else False if verbose: print('Checking derivative of hessian wrt parameter number {}' .format(block_ind)) check_passed[block_ind] = self.checkgrad_block(analytic_hess[:, :, block_ind], numeric_hess[:, :, block_ind], verbose= verbose, step=step, tolerance=tolerance, block_indices= block_indices, plot=p) current_index += current_size return np.all(check_passed)
checkgrad
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, target_param=None, verbose=False, step=1e-06, tolerance=0.001,\n block_indices=None, plot=False, super_plot=False):\n try:\n import numdifft...
def x(self, X, X2=None): r = self._scaled_dist(X, X2) return self.dKdVar_of_r(r)
dKd_dVar
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, X2=None):\n r = self._scaled_dist(X, X2)\n return self.dKdVar_of_r(r)\n", "role": "user" }, { "content": "dKd_dVar", "role": "as...
def x(self, X, dimension, lengthscale, X2=None): r = self._scaled_dist(X, X2) return self.dKdLen_of_r(r, dimension, lengthscale)
dKd_dLen
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, dimension, lengthscale, X2=None):\n r = self._scaled_dist(X, X2)\n return self.dKdLen_of_r(r, dimension, lengthscale)\n", "role": "user" }...
def x(self, r, dimCheck, lengthscale): if dimCheck == True: return self.variance ** (float(1) / self.originalDimensions) * np.exp( -0.5 * r ** 2) * r ** 2 / lengthscale ** (float(1) / self. originalDimensions) else: return self.variance ** (float(1) / self.originalDimensions) * np.exp( -0.5 * r ** 2) / lengthscale ** (float(1) / self.originalDimensions )
dKdLen_of_r
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, r, dimCheck, lengthscale):\n if dimCheck == True:\n return self.variance ** (float(1) / self.originalDimensions) * np.exp(\n -0.5 * ...