source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(self, function, args):
param = self.model.param_array
fs = []
for p in self.chain:
self.model.param_array = p
fs.append(function(*args))
self.model.param_array = param
return fs
| predict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, function, args):\n param = self.model.param_array\n fs = []\n for p in self.chain:\n self.model.param_array = p\n fs.append(functi... |
def x(self, X, X2, target):
if X2 is None:
X2 = X
FX = np.column_stack([f(X) for f in self.F])
FX2 = np.column_stack([f(X2) for f in self.F])
DER = np.zeros((self.n, self.n, self.n))
for i in range(self.n):
DER[i, i, i] = np.sqrt(self.weights[i])
dw = self.variance * mdot(FX, DER, self.G_1, np.diag(np.sqrt(self.
weights)), FX2.T)
dv = mdot(FX, np.diag(np.sqrt(self.weights)), self.G_1, np.diag(np.sqrt
(self.weights)), FX2.T)
np.add(target[:, :, 0], np.transpose(dv, (0, 2, 1)), target[:, :, 0])
np.add(target[:, :, 1:], np.transpose(dw, (0, 2, 1)), target[:, :, 1:])
| param_grad_helper | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, X2, target):\n if X2 is None:\n X2 = X\n FX = np.column_stack([f(X) for f in self.F])\n FX2 = np.column_stack([f(X2) for f in self.F])... |
def x(self, X, Y):
self.update_model(False)
self.set_Y(Y)
self.set_X(X)
self.update_model(True)
| set_XY | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, Y):\n self.update_model(False)\n self.set_Y(Y)\n self.set_X(X)\n self.update_model(True)\n",
"role": "user"
},
{
"content": "s... |
def x(self, X):
assert isinstance(X, np.ndarray)
state = self.update_model()
self.update_model(False)
self.X = ObsAr(X)
self.update_model(state)
| set_X | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n assert isinstance(X, np.ndarray)\n state = self.update_model()\n self.update_model(False)\n self.X = ObsAr(X)\n self.update_model(st... |
def x(self, Y):
assert isinstance(Y, (np.ndarray, ObsAr))
state = self.update_model()
self.update_model(False)
if self.normalizer is not None:
self.normalizer.scale_by(Y)
self.Y_normalized = ObsAr(self.normalizer.normalize(Y))
self.Y = Y
else:
self.Y = ObsAr(Y) if isinstance(Y, np.ndarray) else Y
self.Y_normalized = self.Y
self.update_model(state)
| set_Y | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Y):\n assert isinstance(Y, (np.ndarray, ObsAr))\n state = self.update_model()\n self.update_model(False)\n if self.normalizer is not None:\n ... |
def x(self):
self.posterior, self._log_marginal_likelihood, grad_dict = (self.
inference_method.inference(self.kern, self.X, self.Y_normalized,
self.nu + 2 + np.finfo(float).eps, self.mean_function))
self.kern.update_gradients_full(grad_dict['dL_dK'], self.X)
if self.mean_function is not None:
self.mean_function.update_gradients(grad_dict['dL_dm'], self.X)
self.nu.gradient = grad_dict['dL_dnu']
| parameters_changed | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.posterior, self._log_marginal_likelihood, grad_dict = (self.\n inference_method.inference(self.kern, self.X, self.Y_normalized, \n ... |
def x(self, Xnew, full_cov=False, kern=None):
mu, var = self.posterior._raw_predict(kern=self.kern if kern is None else
kern, Xnew=Xnew, pred_var=self._predictive_variable, full_cov=full_cov)
if self.mean_function is not None:
mu += self.mean_function.f(Xnew)
return mu, var
| raw_predict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Xnew, full_cov=False, kern=None):\n mu, var = self.posterior._raw_predict(kern=self.kern if kern is None else\n kern, Xnew=Xnew, pred_var=self.... |
def x(self, Xnew, full_cov=False, kern=None, **kwargs):
return self.predict_noiseless(Xnew, full_cov=full_cov, kern=kern)
| predict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Xnew, full_cov=False, kern=None, **kwargs):\n return self.predict_noiseless(Xnew, full_cov=full_cov, kern=kern)\n",
"role": "user"
},
{
"c... |
def x(self, Xnew, full_cov=False, kern=None):
mu, var = self._raw_predict(Xnew, full_cov=full_cov, kern=kern)
if self.normalizer is not None:
mu, var = self.normalizer.inverse_mean(mu
), self.normalizer.inverse_variance(var)
return mu, var
| predict_noiseless | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Xnew, full_cov=False, kern=None):\n mu, var = self._raw_predict(Xnew, full_cov=full_cov, kern=kern)\n if self.normalizer is not None:\n mu, ... |
def x(self, X, quantiles=(2.5, 97.5), kern=None, **kwargs):
mu, var = self._raw_predict(X, full_cov=False, kern=kern)
quantiles = [(stats.t.ppf(q / 100.0, self.nu + 2 + self.num_data) * np.
sqrt(var) + mu) for q in quantiles]
if self.normalizer is not None:
quantiles = [self.normalizer.inverse_mean(q) for q in quantiles]
return quantiles
| predict_quantiles | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, quantiles=(2.5, 97.5), kern=None, **kwargs):\n mu, var = self._raw_predict(X, full_cov=False, kern=kern)\n quantiles = [(stats.t.ppf(q / 100.0, ... |
def x(self, X, size=10, full_cov=False, Y_metadata=None, likelihood=None,
**predict_kwargs):
return self.posterior_samples_f(X, size, full_cov=full_cov, **
predict_kwargs)
| posterior_samples | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, size=10, full_cov=False, Y_metadata=None, likelihood=None,\n **predict_kwargs):\n return self.posterior_samples_f(X, size, full_cov=full_cov, **... |
def x(self, X, size=10, full_cov=True, **predict_kwargs):
mu, var = self._raw_predict(X, full_cov=full_cov, **predict_kwargs)
if self.normalizer is not None:
mu, var = self.normalizer.inverse_mean(mu
), self.normalizer.inverse_variance(var)
def x(m, v):
nu = self.nu + 2 + self.num_data
v = np.diag(v.flatten()) if not full_cov else v
Z = np.random.multivariate_normal(np.zeros(X.shape[0]), v, size).T
g = np.tile(np.random.gamma(nu / 2.0, 2.0 / nu, size), (X.shape[0], 1))
return m + Z / np.sqrt(g)
if self.output_dim == 1:
return sim_one_dim(mu, var)
else:
fsim = np.empty((self.output_dim, self.num_data, size))
for d in range(self.output_dim):
if full_cov and var.ndim == 3:
fsim[d] = sim_one_dim(mu[:, d], var[:, :, d])
elif not full_cov and var.ndim == 2:
fsim[d] = sim_one_dim(mu[:, d], var[:, d])
else:
fsim[d] = sim_one_dim(mu[:, d], var)
return fsim
| posterior_samples_f | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, size=10, full_cov=True, **predict_kwargs):\n mu, var = self._raw_predict(X, full_cov=full_cov, **predict_kwargs)\n if self.normalizer is not Non... |
def x(A, offset=0):
from numpy.lib.stride_tricks import as_strided
assert A.ndim == 2, 'only implemented for 2 dimensions'
assert A.shape[0] == A.shape[1
], 'attempting to get the view of non-square matrix?!'
if offset > 0:
return as_strided(A[0, offset:], shape=(A.shape[0] - offset,),
strides=((A.shape[0] + 1) * A.itemsize,))
elif offset < 0:
return as_strided(A[-offset:, 0], shape=(A.shape[0] + offset,),
strides=((A.shape[0] + 1) * A.itemsize,))
else:
return as_strided(A, shape=(A.shape[0],), strides=((A.shape[0] + 1) *
A.itemsize,))
| view | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, offset=0):\n from numpy.lib.stride_tricks import as_strided\n assert A.ndim == 2, 'only implemented for 2 dimensions'\n assert A.shape[0] == A.shap... |
def x(A, b, offset=0):
return _diag_ufunc(A, b, offset, np.multiply)
| times | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, b, offset=0):\n return _diag_ufunc(A, b, offset, np.multiply)\n",
"role": "user"
},
{
"content": "times",
"role": "assistant"
}
] |
def x(A, b, offset=0):
return _diag_ufunc(A, b, offset, np.divide)
| divide | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, b, offset=0):\n return _diag_ufunc(A, b, offset, np.divide)\n",
"role": "user"
},
{
"content": "divide",
"role": "assistant"
}
] |
def x(A, b, offset=0):
return _diag_ufunc(A, b, offset, np.add)
| add | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, b, offset=0):\n return _diag_ufunc(A, b, offset, np.add)\n",
"role": "user"
},
{
"content": "add",
"role": "assistant"
}
] |
def x(A, b, offset=0):
return _diag_ufunc(A, b, offset, np.subtract)
| subtract | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, b, offset=0):\n return _diag_ufunc(A, b, offset, np.subtract)\n",
"role": "user"
},
{
"content": "subtract",
"role": "assistant"
}
] |
def x(input_dict):
import copy
input_dict = copy.deepcopy(input_dict)
inference_class = input_dict.pop('class')
import GPy
inference_class = eval(inference_class)
return inference_class._build_from_input_dict(inference_class, input_dict)
| from_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(input_dict):\n import copy\n input_dict = copy.deepcopy(input_dict)\n inference_class = input_dict.pop('class')\n import GPy\n inference_class =... |
def x(self):
input_dict = super(Add, self)._save_to_input_dict()
input_dict['class'] = str('GPy.kern.Add')
return input_dict
| to_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n input_dict = super(Add, self)._save_to_input_dict()\n input_dict['class'] = str('GPy.kern.Add')\n return input_dict\n",
"role": "user"
... |
def x(self, dL_dK, X, X2=None):
target = np.zeros(X.shape)
[target.__iadd__(p.gradients_X(dL_dK, X, X2)) for p in self.parts]
return target
| gradients_X | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dL_dK, X, X2=None):\n target = np.zeros(X.shape)\n [target.__iadd__(p.gradients_X(dL_dK, X, X2)) for p in self.parts]\n return target\n",
"r... |
def x(self):
import scipy.linalg as la
F = None
L = None
Qc = None
H = None
Pinf = None
P0 = None
dF = None
dQc = None
dPinf = None
dP0 = None
n = 0
nq = 0
nd = 0
for p in self.parts:
Ft, Lt, Qct, Ht, Pinft, P0t, dFt, dQct, dPinft, dP0t = p.sde()
F = la.block_diag(F, Ft) if F is not None else Ft
L = la.block_diag(L, Lt) if L is not None else Lt
Qc = la.block_diag(Qc, Qct) if Qc is not None else Qct
H = np.hstack((H, Ht)) if H is not None else Ht
Pinf = la.block_diag(Pinf, Pinft) if Pinf is not None else Pinft
P0 = la.block_diag(P0, P0t) if P0 is not None else P0t
if dF is not None:
dF = np.pad(dF, ((0, dFt.shape[0]), (0, dFt.shape[1]), (0, dFt.
shape[2])), 'constant', constant_values=0)
dF[-dFt.shape[0]:, -dFt.shape[1]:, -dFt.shape[2]:] = dFt
else:
dF = dFt
if dQc is not None:
dQc = np.pad(dQc, ((0, dQct.shape[0]), (0, dQct.shape[1]), (0,
dQct.shape[2])), 'constant', constant_values=0)
dQc[-dQct.shape[0]:, -dQct.shape[1]:, -dQct.shape[2]:] = dQct
else:
dQc = dQct
if dPinf is not None:
dPinf = np.pad(dPinf, ((0, dPinft.shape[0]), (0, dPinft.shape[1
]), (0, dPinft.shape[2])), 'constant', constant_values=0)
dPinf[-dPinft.shape[0]:, -dPinft.shape[1]:, -dPinft.shape[2]:
] = dPinft
else:
dPinf = dPinft
if dP0 is not None:
dP0 = np.pad(dP0, ((0, dP0t.shape[0]), (0, dP0t.shape[1]), (0,
dP0t.shape[2])), 'constant', constant_values=0)
dP0[-dP0t.shape[0]:, -dP0t.shape[1]:, -dP0t.shape[2]:] = dP0t
else:
dP0 = dP0t
n += Ft.shape[0]
nq += Qct.shape[0]
nd += dFt.shape[2]
assert F.shape[0] == n and F.shape[1
] == n, 'SDE add: Check of F Dimensions failed'
assert L.shape[0] == n and L.shape[1
] == nq, 'SDE add: Check of L Dimensions failed'
assert Qc.shape[0] == nq and Qc.shape[1
] == nq, 'SDE add: Check of Qc Dimensions failed'
assert H.shape[0] == 1 and H.shape[1
] == n, 'SDE add: Check of H Dimensions failed'
assert Pinf.shape[0] == n and Pinf.shape[1
] == n, 'SDE add: Check of Pinf Dimensions failed'
assert P0.shape[0] == n and P0.shape[1
] == n, 'SDE add: Check of P0 Dimensions failed'
assert dF.shape[0] == n and dF.shape[1] == n and dF.shape[2
] == nd, 'SDE add: Check of dF Dimensions failed'
assert dQc.shape[0] == nq and dQc.shape[1] == nq and dQc.shape[2
] == nd, 'SDE add: Check of dQc Dimensions failed'
assert dPinf.shape[0] == n and dPinf.shape[1] == n and dPinf.shape[2
] == nd, 'SDE add: Check of dPinf Dimensions failed'
assert dP0.shape[0] == n and dP0.shape[1] == n and dP0.shape[2
] == nd, 'SDE add: Check of dP0 Dimensions failed'
return F, L, Qc, H, Pinf, P0, dF, dQc, dPinf, dP0
| sde | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n import scipy.linalg as la\n F = None\n L = None\n Qc = None\n H = None\n Pinf = None\n P0 = None\n dF = None\n dQc = None\n... |
def x(p, labels, names=['1', '0'], threshold=0.5, show=True):
assert p.size == labels.size, 'Arrays p and labels have different dimensions.'
decision = np.ones((labels.size, 1))
decision[p < threshold] = 0
diff = decision - labels
false_0 = diff[diff == -1].size
false_1 = diff[diff == 1].size
true_1 = np.sum(decision[diff == 0])
true_0 = labels.size - true_1 - false_0 - false_1
error = (false_1 + false_0) / np.float(labels.size)
if show:
print(100.0 - error * 100, '% instances correctly classified')
print('%-10s| %-10s| %-10s| ' % ('', names[0], names[1]))
print('----------|------------|------------|')
print('%-10s| %-10s| %-10s| ' % (names[0], true_1, false_0))
print('%-10s| %-10s| %-10s| ' % (names[1], false_1, true_0))
return error, true_1, false_1, true_0, false_0
| conf_matrix | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(p, labels, names=['1', '0'], threshold=0.5, show=True):\n assert p.size == labels.size, 'Arrays p and labels have different dimensions.'\n decision = np.... |
def x(self, save_data=True):
input_dict = super(SparseGP, self).to_dict(save_data)
input_dict['class'] = 'GPy.core.SparseGP'
input_dict['Z'] = self.Z.tolist()
return input_dict
| to_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, save_data=True):\n input_dict = super(SparseGP, self).to_dict(save_data)\n input_dict['class'] = 'GPy.core.SparseGP'\n input_dict['Z'] = self.Z.... |
def x(A):
if A.flags['F_CONTIGUOUS']:
return A
if A.flags['C_CONTIGUOUS']:
return A.T
else:
return np.asfortranarray(A)
| force_F_ordered_symmetric | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A):\n if A.flags['F_CONTIGUOUS']:\n return A\n if A.flags['C_CONTIGUOUS']:\n return A.T\n else:\n return np.asfortranarray(A)\n",... |
def x(A, B, lower=1, trans=0, unitdiag=0):
A = np.asfortranarray(A)
return lapack.dtrtrs(A, B, lower=lower, trans=trans, unitdiag=unitdiag)
| dtrtrs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, B, lower=1, trans=0, unitdiag=0):\n A = np.asfortranarray(A)\n return lapack.dtrtrs(A, B, lower=lower, trans=trans, unitdiag=unitdiag)\n",
"role":... |
def x(A, B, lower=1):
A = force_F_ordered(A)
return lapack.dpotrs(A, B, lower=lower)
| dpotrs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, B, lower=1):\n A = force_F_ordered(A)\n return lapack.dpotrs(A, B, lower=lower)\n",
"role": "user"
},
{
"content": "dpotrs",
"role": "... |
def x(A, lower=1):
A = force_F_ordered(A)
R, info = lapack.dpotri(A, lower=lower)
symmetrify(R)
return R, info
| dpotri | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, lower=1):\n A = force_F_ordered(A)\n R, info = lapack.dpotri(A, lower=lower)\n symmetrify(R)\n return R, info\n",
"role": "user"
},
{
... |
def x(A):
L = jitchol(A)
logdetA = 2 * sum(np.log(np.diag(L)))
return logdetA
| pddet | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A):\n L = jitchol(A)\n logdetA = 2 * sum(np.log(np.diag(L)))\n return logdetA\n",
"role": "user"
},
{
"content": "pddet",
"role": "ass... |
def x(*args):
if len(args) == 1:
return args[0]
elif len(args) == 2:
return _mdot_r(args[0], args[1])
else:
return _mdot_r(args[:-1], args[-1])
| mdot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(*args):\n if len(args) == 1:\n return args[0]\n elif len(args) == 2:\n return _mdot_r(args[0], args[1])\n else:\n return _mdot_r(... |
def x(a, b):
if type(a) == tuple:
if len(a) > 1:
a = mdot(*a)
else:
a = a[0]
if type(b) == tuple:
if len(b) > 1:
b = mdot(*b)
else:
b = b[0]
return np.dot(a, b)
| mdot_r | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(a, b):\n if type(a) == tuple:\n if len(a) > 1:\n a = mdot(*a)\n else:\n a = a[0]\n if type(b) == tuple:\n if l... |
def x(A, *args):
L = jitchol(A, *args)
logdet = 2.0 * np.sum(np.log(np.diag(L)))
Li = dtrtri(L)
Ai, _ = dpotri(L, lower=1)
symmetrify(Ai)
return Ai, L, Li, logdet
| pdinv | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, *args):\n L = jitchol(A, *args)\n logdet = 2.0 * np.sum(np.log(np.diag(L)))\n Li = dtrtri(L)\n Ai, _ = dpotri(L, lower=1)\n symmetrify(Ai)\n ... |
def x(A):
N = A.shape[-1]
chols = [jitchol(A[:, :, i]) for i in range(N)]
halflogdets = [np.sum(np.log(np.diag(L[0]))) for L in chols]
invs = [dpotri(L[0], True)[0] for L in chols]
invs = [(np.triu(I) + np.triu(I, 1).T) for I in invs]
return np.dstack(invs), np.array(halflogdets)
| multiple_pdinv | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A):\n N = A.shape[-1]\n chols = [jitchol(A[:, :, i]) for i in range(N)]\n halflogdets = [np.sum(np.log(np.diag(L[0]))) for L in chols]\n invs = [dp... |
def x(Y, input_dim):
if not np.allclose(Y.mean(axis=0), 0.0):
print('Y is not zero mean, centering it locally (GPy.util.linalg.pca)')
Z = linalg.svd(Y - Y.mean(axis=0), full_matrices=False)
[X, W] = [Z[0][:, 0:input_dim], np.dot(np.diag(Z[1]), Z[2]).T[:, 0:
input_dim]]
v = X.std(axis=0)
X /= v
W *= v
return X, W.T
| pca | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(Y, input_dim):\n if not np.allclose(Y.mean(axis=0), 0.0):\n print('Y is not zero mean, centering it locally (GPy.util.linalg.pca)')\n Z = linalg.s... |
def x(Y, Q, iterations=100):
from numpy.ma import dot as madot
N, D = Y.shape
W = np.random.randn(D, Q) * 0.001
Y = np.ma.masked_invalid(Y, copy=0)
mu = Y.mean(0)
Ycentered = Y - mu
try:
for _ in range(iterations):
exp_x = np.asarray_chkfinite(np.linalg.solve(W.T.dot(W), madot(
W.T, Ycentered.T))).T
W = np.asarray_chkfinite(np.linalg.solve(exp_x.T.dot(exp_x),
madot(exp_x.T, Ycentered))).T
except np.linalg.linalg.LinAlgError:
pass
return np.asarray_chkfinite(exp_x), np.asarray_chkfinite(W)
| ppca | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(Y, Q, iterations=100):\n from numpy.ma import dot as madot\n N, D = Y.shape\n W = np.random.randn(D, Q) * 0.001\n Y = np.ma.masked_invalid(Y, copy=... |
def x(mat, out=None):
if mat.dtype != 'float64' or len(mat.shape) != 2:
return np.dot(mat, mat.T)
nn = mat.shape[0]
if out is None:
out = np.zeros((nn, nn))
else:
assert out.dtype == 'float64'
assert out.shape == (nn, nn)
assert 8 in out.strides
out[:] = 0.0
mat = np.asfortranarray(mat)
out = blas.dsyrk(alpha=1.0, a=mat, beta=0.0, c=out, overwrite_c=1,
trans=0, lower=0)
symmetrify(out, upper=True)
return np.ascontiguousarray(out)
| tdot_blas | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(mat, out=None):\n if mat.dtype != 'float64' or len(mat.shape) != 2:\n return np.dot(mat, mat.T)\n nn = mat.shape[0]\n if out is None:\n ... |
def x(A, x, alpha=1.0):
At = blas.dsyr(lower=0, x=x, a=A, alpha=alpha, overwrite_a=False)
A[:] = At
symmetrify(A, upper=True)
| DSYR_blas | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, x, alpha=1.0):\n At = blas.dsyr(lower=0, x=x, a=A, alpha=alpha, overwrite_a=False)\n A[:] = At\n symmetrify(A, upper=True)\n",
"role": "user"
... |
def x(A, x, alpha=1.0):
A += alpha * np.dot(x[:, None], x[None, :])
| DSYR_numpy | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, x, alpha=1.0):\n A += alpha * np.dot(x[:, None], x[None, :])\n",
"role": "user"
},
{
"content": "DSYR_numpy",
"role": "assistant"
}
] |
def x(A, upper=False):
if use_linalg_cython:
_symmetrify_cython(A, upper)
else:
_symmetrify_numpy(A, upper)
| symmetrify | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, upper=False):\n if use_linalg_cython:\n _symmetrify_cython(A, upper)\n else:\n _symmetrify_numpy(A, upper)\n",
"role": "user"
},
... |
def x(L, X, transpose='left'):
if transpose == 'left':
tmp, _ = dtrtrs(L, X, lower=1, trans=1)
return dtrtrs(L, tmp.T, lower=1, trans=1)[0].T
else:
tmp, _ = dtrtrs(L, X, lower=1, trans=0)
return dtrtrs(L, tmp.T, lower=1, trans=0)[0].T
| backsub_both_sides | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(L, X, transpose='left'):\n if transpose == 'left':\n tmp, _ = dtrtrs(L, X, lower=1, trans=1)\n return dtrtrs(L, tmp.T, lower=1, trans=1)[0].T\... |
def x(A, B):
return A.dot(B.reshape(B.shape[0], -1)).reshape(A.shape[0], B.shape[1],
B.shape[2])
| ij_jlk_to_ilk | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, B):\n return A.dot(B.reshape(B.shape[0], -1)).reshape(A.shape[0], B.shape[1],\n B.shape[2])\n",
"role": "user"
},
{
"content": "ij_jlk... |
def x(A, B):
res = np.zeros((A.shape[0], B.shape[1]))
[np.add(np.dot(A[:, :, k], B[:, :, k]), res, out=res) for k in range(B.
shape[-1])]
return res
| ijk_jlk_to_il | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, B):\n res = np.zeros((A.shape[0], B.shape[1]))\n [np.add(np.dot(A[:, :, k], B[:, :, k]), res, out=res) for k in range(B.\n shape[-1])]\n ret... |
def x(A, B):
res = np.zeros((A.shape[-1], A.shape[0], B.shape[0]))
[np.dot(A[:, :, i], B[:, :, i].T, out=res[i, :, :]) for i in range(A.
shape[-1])]
res = res.swapaxes(0, 2).swapaxes(0, 1)
return res
| ijk_ljk_to_ilk | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, B):\n res = np.zeros((A.shape[-1], A.shape[0], B.shape[0]))\n [np.dot(A[:, :, i], B[:, :, i].T, out=res[i, :, :]) for i in range(A.\n shape[-1]... |
def x(file_or_path):
import sys
import inspect
sys.modules['GPy.kern._src'] = kern.src
for name, module in inspect.getmembers(kern.src):
if not name.startswith('_'):
sys.modules['GPy.kern._src.{}'.format(name)] = module
sys.modules['GPy.inference.optimization'] = inference.optimization
import paramz
return paramz.load(file_or_path)
| load | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(file_or_path):\n import sys\n import inspect\n sys.modules['GPy.kern._src'] = kern.src\n for name, module in inspect.getmembers(kern.src):\n ... |
def x(feval, fmin, fmax, tol1=1e-05, tol2=1e-05):
XK = np.array([-0.991455371120813, -0.949107912342759, -
0.864864423359769, -0.741531185599394, -0.586087235467691, -
0.405845151377397, -0.207784955007898, 0.0, 0.207784955007898,
0.405845151377397, 0.586087235467691, 0.741531185599394,
0.864864423359769, 0.949107912342759, 0.991455371120813])
WK = np.array([0.022935322010529, 0.063092092629979, 0.10479001032225,
0.140653259715525, 0.169004726639267, 0.190350578064785,
0.204432940075298, 0.209482141084728, 0.204432940075298,
0.190350578064785, 0.169004726639267, 0.140653259715525,
0.10479001032225, 0.063092092629979, 0.022935322010529])
WG = np.array([0.12948496616887, 0.279705391489277, 0.381830050505119,
0.417959183673469, 0.381830050505119, 0.279705391489277,
0.12948496616887])
NK = WK.size
G = np.arange(2, NK, 2)
tol1 = 0.0001
tol2 = 0.0001
Subs = np.array([[fmin], [fmax]])
NF = feval(np.zeros(1)).size
Q = np.zeros(NF)
neval = 0
while Subs.size > 0:
Subs = getSubs(Subs, XK)
M = (Subs[1, :] - Subs[0, :]) / 2
C = (Subs[1, :] + Subs[0, :]) / 2
NM = M.size
x = XK[:, None] * M + C
x = x.flatten()
FV = feval(x)
Q1 = np.zeros((NF, NM))
Q2 = np.zeros((NF, NM))
Q1 = np.dot(FV.reshape(NF, NK, NM).swapaxes(2, 1), WK) * M
Q2 = np.dot(FV.reshape(NF, NK, NM).swapaxes(2, 1)[:, :, 1::2], WG) * M
ind = np.nonzero(np.logical_or(np.max(np.abs(Q1 - Q2), 0) < tol1,
Subs[1, :] - Subs[0, :] < tol2))[0]
Q = Q + np.sum(Q1[:, ind], axis=1)
Subs = np.delete(Subs, ind, axis=1)
return Q
| quadvgk | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(feval, fmin, fmax, tol1=1e-05, tol2=1e-05):\n XK = np.array([-0.991455371120813, -0.949107912342759, -\n 0.864864423359769, -0.741531185599394, -0.58... |
def x(f, fmin=-np.inf, fmax=np.inf, difftol=0.1):
difftol = 0.0001
def x(r):
r2 = np.square(r)
x = r / (1 - r2)
dx_dr = (1 + r2) / (1 - r2) ** 2
return f(x) * dx_dr
integrand = quadvgk(trans_func, -1.0, 1.0, difftol, difftol)
return integrand
| quadgk_int | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(f, fmin=-np.inf, fmax=np.inf, difftol=0.1):\n difftol = 0.0001\n\n def x(r):\n r2 = np.square(r)\n x = r / (1 - r2)\n dx_dr = (1 + r... |
def x(self, varp_list):
if self.group_spike:
self._b_prob_all = 1.0 - param_to_array(varp_list[0].gamma_group)
[np.multiply(self._b_prob_all, 1.0 - vp.gamma_group, self.
_b_prob_all) for vp in varp_list[1:]]
else:
self._b_prob_all = 1.0 - param_to_array(varp_list[0].binary_prob)
[np.multiply(self._b_prob_all, 1.0 - vp.binary_prob, self.
_b_prob_all) for vp in varp_list[1:]]
| update_inernal | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, varp_list):\n if self.group_spike:\n self._b_prob_all = 1.0 - param_to_array(varp_list[0].gamma_group)\n [np.multiply(self._b_prob_all, ... |
def x(self, varp_list):
self._b_prob_all = 1.0 - param_to_array(varp_list[0].gamma_group)
[np.multiply(self._b_prob_all, 1.0 - vp.gamma_group, self._b_prob_all) for
vp in varp_list[1:]]
| update_inernal | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, varp_list):\n self._b_prob_all = 1.0 - param_to_array(varp_list[0].gamma_group)\n [np.multiply(self._b_prob_all, 1.0 - vp.gamma_group, self._b_prob... |
def x(self, X, X2, target):
AX = np.dot(X, self.transform)
if X2 is None:
X2 = X
AX2 = AX
else:
AX2 = np.dot(X2, self.transform)
self.k.K(X, X2, target)
self.k.K(AX, X2, target)
self.k.K(X, AX2, target)
self.k.K(AX, AX2, target)
| K | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, X2, target):\n AX = np.dot(X, self.transform)\n if X2 is None:\n X2 = X\n AX2 = AX\n else:\n AX2 = np.dot(X2, self.trans... |
def x(self, dL_dK, X, X2, target):
AX = np.dot(X, self.transform)
if X2 is None:
X2 = X
ZX2 = AX
else:
AX2 = np.dot(X2, self.transform)
self.k._param_grad_helper(dL_dK, X, X2, target)
self.k._param_grad_helper(dL_dK, AX, X2, target)
self.k._param_grad_helper(dL_dK, X, AX2, target)
self.k._param_grad_helper(dL_dK, AX, AX2, target)
| param_grad_helper | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dL_dK, X, X2, target):\n AX = np.dot(X, self.transform)\n if X2 is None:\n X2 = X\n ZX2 = AX\n else:\n AX2 = np.dot(X2, sel... |
def x(self, X, target):
foo = np.zeros((X.shape[0], X.shape[0]))
self.K(X, X, foo)
target += np.diag(foo)
| Kdiag | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, target):\n foo = np.zeros((X.shape[0], X.shape[0]))\n self.K(X, X, foo)\n target += np.diag(foo)\n",
"role": "user"
},
{
"content... |
def x(self, dL_dpsi2, Z, mu, S, target):
self._psi_computations(Z, mu, S)
d_var = 2.0 * self._psi2 / self.variance
d_length = -2.0 * self._psi2[:, :, :, None] * (self._psi2_Zdist_sq *
self._psi2_denom + self._psi2_mudist_sq + S[:, None, None, :] *
self.inv_lengthscale2) / (self.inv_lengthscale * self._psi2_denom)
target[0] += np.sum(dL_dpsi2 * d_var)
dpsi2_dlength = d_length * dL_dpsi2[:, :, :, None]
if not self.ARD:
target[1] += dpsi2_dlength.sum()
else:
target[1:] += dpsi2_dlength.sum(0).sum(0).sum(0)
| dpsi2_dtheta | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dL_dpsi2, Z, mu, S, target):\n self._psi_computations(Z, mu, S)\n d_var = 2.0 * self._psi2 / self.variance\n d_length = -2.0 * self._psi2[:, :, ... |
def x(self, dL_dpsi2, Z, mu, S, target_mu, target_S):
self._psi_computations(Z, mu, S)
tmp = self.inv_lengthscale2 * self._psi2[:, :, :, None] / self._psi2_denom
target_mu += -2.0 * (dL_dpsi2[:, :, :, None] * tmp * self._psi2_mudist
).sum(1).sum(1)
target_S += (dL_dpsi2[:, :, :, None] * tmp * (2.0 * self.
_psi2_mudist_sq - 1)).sum(1).sum(1)
| dpsi2_dmuS | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dL_dpsi2, Z, mu, S, target_mu, target_S):\n self._psi_computations(Z, mu, S)\n tmp = self.inv_lengthscale2 * self._psi2[:, :, :, None] / self._psi2... |
def x(self, kern, Z, variational_posterior, return_psi2_n=False):
variance, lengthscale = kern.variance, kern.lengthscale
N, M, Q = self.get_dimensions(Z, variational_posterior)
self._initGPUCache(N, M, Q)
self.sync_params(lengthscale, Z, variational_posterior.mean,
variational_posterior.variance)
psi1_gpu = self.gpuCache['psi1_gpu']
psi2_gpu = self.gpuCache['psi2_gpu']
psi2n_gpu = self.gpuCache['psi2n_gpu']
l_gpu = self.gpuCache['l_gpu']
Z_gpu = self.gpuCache['Z_gpu']
mu_gpu = self.gpuCache['mu_gpu']
S_gpu = self.gpuCache['S_gpu']
log_denom1_gpu = self.gpuCache['log_denom1_gpu']
log_denom2_gpu = self.gpuCache['log_denom2_gpu']
psi0 = np.empty((N,))
psi0[:] = variance
self.g_psi1computations.prepared_call((self.blocknum, 1), (self.
threadnum, 1, 1), psi1_gpu.gpudata, log_denom1_gpu.gpudata, np.
float64(variance), l_gpu.gpudata, Z_gpu.gpudata, mu_gpu.gpudata,
S_gpu.gpudata, np.int32(N), np.int32(M), np.int32(Q))
self.g_psi2computations.prepared_call((self.blocknum, 1), (self.
threadnum, 1, 1), psi2_gpu.gpudata, psi2n_gpu.gpudata,
log_denom2_gpu.gpudata, np.float64(variance), l_gpu.gpudata, Z_gpu.
gpudata, mu_gpu.gpudata, S_gpu.gpudata, np.int32(N), np.int32(M),
np.int32(Q))
if self.GPU_direct:
return psi0, psi1_gpu, psi2_gpu
elif return_psi2_n:
return psi0, psi1_gpu.get(), psi2n_gpu.get()
else:
return psi0, psi1_gpu.get(), psi2_gpu.get()
| psicomputations | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, kern, Z, variational_posterior, return_psi2_n=False):\n variance, lengthscale = kern.variance, kern.lengthscale\n N, M, Q = self.get_dimensions(Z, ... |
def x(self, prior, warning=True):
repriorized = self.unset_priors()
self._add_to_index_operations(self.priors, repriorized, prior, warning)
from paramz.domains import _REAL, _POSITIVE, _NEGATIVE
if prior.domain is _POSITIVE:
self.constrain_positive(warning)
elif prior.domain is _NEGATIVE:
self.constrain_negative(warning)
elif prior.domain is _REAL:
rav_i = self._raveled_index()
assert all(all(False if c is __fixed__ else c.domain is _REAL for c in
con) for con in self.constraints.properties_for(rav_i)
), 'Domain of prior and constraint have to match, please unconstrain if you REALLY wish to use this prior'
| set_prior | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, prior, warning=True):\n repriorized = self.unset_priors()\n self._add_to_index_operations(self.priors, repriorized, prior, warning)\n from param... |
def x(self):
if self.priors.size == 0:
return 0.0
x = self.param_array
log_p = reduce(lambda a, b: a + b, (p.lnpdf(x[ind]).sum() for p, ind in
self.priors.items()), 0)
log_j = 0.0
priored_indexes = np.hstack([i for p, i in self.priors.items()])
for c, j in self.constraints.items():
if not isinstance(c, Transformation):
continue
for jj in j:
if jj in priored_indexes:
log_j += c.log_jacobian(x[jj])
return log_p + log_j
| log_prior | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.priors.size == 0:\n return 0.0\n x = self.param_array\n log_p = reduce(lambda a, b: a + b, (p.lnpdf(x[ind]).sum() for p, ind i... |
def x(self):
if self.priors.size == 0:
return 0.0
x = self.param_array
ret = np.zeros(x.size)
[np.put(ret, ind, p.lnpdf_grad(x[ind])) for p, ind in self.priors.items()]
priored_indexes = np.hstack([i for p, i in self.priors.items()])
for c, j in self.constraints.items():
if not isinstance(c, Transformation):
continue
for jj in j:
if jj in priored_indexes:
ret[jj] += c.log_jacobian_grad(x[jj])
return ret
| log_prior_gradients | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.priors.size == 0:\n return 0.0\n x = self.param_array\n ret = np.zeros(x.size)\n [np.put(ret, ind, p.lnpdf_grad(x[ind])) fo... |
def x(self, data, train):
if train:
inputs, labels = data
self.data_mean = inputs.mean(axis=0)
self.data_std = inputs.std(axis=0)
self.labels_mean = labels.mean(axis=0)
self.labels_std = labels.std(axis=0)
return (inputs - self.data_mean) / self.data_std, (labels - self.
labels_mean) / self.labels_std
else:
return (data - self.data_mean) / self.data_std
| preprocess | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data, train):\n if train:\n inputs, labels = data\n self.data_mean = inputs.mean(axis=0)\n self.data_std = inputs.std(axis=0)\n ... |
def x(A, blocksizes, pagesizes=None):
assert A.shape[0] == A.shape[1] and len(A.shape
) == 3, "can't blockify this non-square matrix, may need to use 2d version"
N = np.sum(blocksizes)
assert A.shape[0] == N, 'bad blocksizes'
num_blocks = len(blocksizes)
if pagesizes == None:
pagesizes = range(A.shape[2])
num_pages = len(pagesizes)
B = np.empty(shape=(num_blocks, num_blocks, num_pages), dtype=np.object)
count_k = 0
for Bk in pagesizes:
count_i = 0
for Bi, i in enumerate(blocksizes):
count_j = 0
for Bj, j in enumerate(blocksizes):
B[Bi, Bj, Bk] = A[count_i:count_i + i, count_j:count_j + j, Bk]
count_j += j
count_i += i
return B
| get_blocks_3d | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, blocksizes, pagesizes=None):\n assert A.shape[0] == A.shape[1] and len(A.shape\n ) == 3, \"can't blockify this non-square matrix, may need to use ... |
def x(A, B, diagonal=False):
assert A.dtype is np.dtype('object'), 'Must be a block matrix'
assert B.dtype is np.dtype('object'), 'Must be a block matrix'
assert A.shape == B.shape
def x(C, D):
Cshape = C.shape
Dshape = D.shape
if diagonal and (len(Cshape) == 1 or len(Dshape) == 1 or C.shape[0] !=
C.shape[1] or D.shape[0] != D.shape[1]):
print('Broadcasting, C: {} D:{}'.format(C.shape, D.shape))
return C * D
else:
print('Dotting, C: {} C:{}'.format(C.shape, D.shape))
return np.dot(C, D)
dot = np.vectorize(f, otypes=[np.object])
return dot(A, B)
| block_dot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, B, diagonal=False):\n assert A.dtype is np.dtype('object'), 'Must be a block matrix'\n assert B.dtype is np.dtype('object'), 'Must be a block matrix'\... |
def x(num_inducing=50, max_iters=100, kernel=None, optimize=True, plot=True):
try:
import pods
except ImportError:
raise ImportWarning(
'Need pods for example datasets. See https://github.com/sods/ods, or pip install pods.'
)
data = pods.datasets.oil()
X = data['X']
Xtest = data['Xtest']
Y = data['Y'][:, 0:1]
Ytest = data['Ytest'][:, 0:1]
Y[Y.flatten() == -1] = 0
Ytest[Ytest.flatten() == -1] = 0
m = GPy.models.SparseGPClassification(X, Y, kernel=kernel, num_inducing
=num_inducing)
m.Ytest = Ytest
m['.*len'] = 10.0
if optimize:
m.optimize(messages=1)
print(m)
probs = m.predict(Xtest)[0]
GPy.util.classification.conf_matrix(probs, Ytest)
return m
| oil | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(num_inducing=50, max_iters=100, kernel=None, optimize=True, plot=True):\n try:\n import pods\n except ImportError:\n raise ImportWarning(\n... |
def x(seed=default_seed, optimize=True, plot=True):
try:
import pods
except ImportError:
raise ImportWarning(
'Need pods for example datasets. See https://github.com/sods/ods, or pip install pods.'
)
data = pods.datasets.toy_linear_1d_classification(seed=seed)
Y = data['Y'][:, 0:1]
Y[Y.flatten() == -1] = 0
m = GPy.models.GPClassification(data['X'], Y)
if optimize:
m.optimize()
if plot:
from matplotlib import pyplot as plt
fig, axes = plt.subplots(2, 1)
m.plot_f(ax=axes[0])
m.plot(ax=axes[1])
print(m)
return m
| toy_linear_1d_classification | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seed=default_seed, optimize=True, plot=True):\n try:\n import pods\n except ImportError:\n raise ImportWarning(\n 'Need pods for... |
def x(seed=default_seed, optimize=True, plot=True):
try:
import pods
except ImportError:
print(
'pods unavailable, see https://github.com/sods/ods for example datasets'
)
data = pods.datasets.toy_linear_1d_classification(seed=seed)
Y = data['Y'][:, 0:1]
Y[Y.flatten() == -1] = 0
likelihood = GPy.likelihoods.Bernoulli()
laplace_inf = GPy.inference.latent_function_inference.Laplace()
kernel = GPy.kern.RBF(1)
m = GPy.core.GP(data['X'], Y, kernel=kernel, likelihood=likelihood,
inference_method=laplace_inf)
if optimize:
try:
m.optimize('scg', messages=1)
except Exception as e:
return m
if plot:
from matplotlib import pyplot as plt
fig, axes = plt.subplots(2, 1)
m.plot_f(ax=axes[0])
m.plot(ax=axes[1])
print(m)
return m
| toy_linear_1d_classification_laplace | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seed=default_seed, optimize=True, plot=True):\n try:\n import pods\n except ImportError:\n print(\n 'pods unavailable, see https... |
def x(num_inducing=10, seed=default_seed, optimize=True, plot=True):
try:
import pods
except ImportError:
print(
'pods unavailable, see https://github.com/sods/ods for example datasets'
)
import numpy as np
data = pods.datasets.toy_linear_1d_classification(seed=seed)
Y = data['Y'][:, 0:1]
Y[Y.flatten() == -1] = 0
X = data['X']
X_var = np.random.uniform(0.3, 0.5, X.shape)
m = GPy.models.SparseGPClassificationUncertainInput(X, X_var, Y,
num_inducing=num_inducing)
m['.*len'] = 4.0
if optimize:
m.optimize()
if plot:
from matplotlib import pyplot as plt
fig, axes = plt.subplots(2, 1)
m.plot_f(ax=axes[0])
m.plot(ax=axes[1])
print(m)
return m
| sparse_toy_linear_1d_classification_uncertain_input | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(num_inducing=10, seed=default_seed, optimize=True, plot=True):\n try:\n import pods\n except ImportError:\n print(\n 'pods unava... |
def x(seed=default_seed, max_iters=100, optimize=True, plot=True):
try:
import pods
except ImportError:
print(
'pods unavailable, see https://github.com/sods/ods for example datasets'
)
data = pods.datasets.toy_linear_1d_classification(seed=seed)
Y = data['Y'][:, 0:1]
Y[Y.flatten() == -1] = 0
kernel = GPy.kern.RBF(1)
likelihood = GPy.likelihoods.Bernoulli(gp_link=GPy.likelihoods.
link_functions.Heaviside())
ep = GPy.inference.latent_function_inference.expectation_propagation.EP()
m = GPy.core.GP(X=data['X'], Y=Y, kernel=kernel, likelihood=likelihood,
inference_method=ep, name='gp_classification_heaviside')
if optimize:
for _ in range(5):
m.optimize(max_iters=int(max_iters / 5))
print(m)
if plot:
from matplotlib import pyplot as plt
fig, axes = plt.subplots(2, 1)
m.plot_f(ax=axes[0])
m.plot(ax=axes[1])
print(m)
return m
| toy_heaviside | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seed=default_seed, max_iters=100, optimize=True, plot=True):\n try:\n import pods\n except ImportError:\n print(\n 'pods unavail... |
def x(model_type='Full', num_inducing=10, seed=default_seed, kernel=None,
optimize=True, plot=True):
try:
import pods
except ImportError:
print(
'pods unavailable, see https://github.com/sods/ods for example datasets'
)
data = pods.datasets.crescent_data(seed=seed)
Y = data['Y']
Y[Y.flatten() == -1] = 0
if model_type == 'Full':
m = GPy.models.GPClassification(data['X'], Y, kernel=kernel)
elif model_type == 'DTC':
m = GPy.models.SparseGPClassification(data['X'], Y, kernel=kernel,
num_inducing=num_inducing)
m['.*len'] = 10.0
elif model_type == 'FITC':
m = GPy.models.FITCClassification(data['X'], Y, kernel=kernel,
num_inducing=num_inducing)
m['.*len'] = 3.0
if optimize:
m.optimize(messages=1)
if plot:
m.plot()
print(m)
return m
| crescent_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model_type='Full', num_inducing=10, seed=default_seed, kernel=None,\n optimize=True, plot=True):\n try:\n import pods\n except ImportError:\n ... |
def x(self, X, test_data=False):
X_warped = X.copy()
if test_data:
X_normalized = (X - self.Xmin) / (self.Xmax - self.Xmin)
else:
X_normalized = self.X_normalized
for i_seq, i_fea in enumerate(self.warping_indices):
a, b = self.params[i_seq][0], self.params[i_seq][1]
X_warped[:, i_fea] = 1 - np.power(1 - np.power(X_normalized[:,
i_fea], a), b)
return X_warped
| f | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, test_data=False):\n X_warped = X.copy()\n if test_data:\n X_normalized = (X - self.Xmin) / (self.Xmax - self.Xmin)\n else:\n X_... |
def x(self, X):
grad = np.zeros(X.shape)
for i_seq, i_fea in enumerate(self.warping_indices):
a, b = self.params[i_seq][0], self.params[i_seq][1]
grad[:, i_fea] = a * b * np.power(self.X_normalized[:, i_fea], a - 1
) * np.power(1 - np.power(self.X_normalized[:, i_fea], a), b - 1
) * self.scaling[i_fea]
return grad
| fgrad_X | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n grad = np.zeros(X.shape)\n for i_seq, i_fea in enumerate(self.warping_indices):\n a, b = self.params[i_seq][0], self.params[i_seq][1]\... |
def x(self, X, dL_dW):
for i_seq, i_fea in enumerate(self.warping_indices):
ai, bi = self.params[i_seq][0], self.params[i_seq][1]
x_pow_a = np.power(self.X_normalized[:, i_fea], ai)
dz_dai = bi * np.power(1 - x_pow_a, bi - 1) * x_pow_a * np.log(self
.X_normalized[:, i_fea])
dz_dbi = -np.power(1 - x_pow_a, bi) * np.log(1 - x_pow_a)
dL_dai = np.sum(dL_dW[:, i_fea] * dz_dai)
dL_dbi = np.sum(dL_dW[:, i_fea] * dz_dbi)
self.params[i_seq][0].gradient[:] = dL_dai
self.params[i_seq][1].gradient[:] = dL_dbi
| update_grads | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, dL_dW):\n for i_seq, i_fea in enumerate(self.warping_indices):\n ai, bi = self.params[i_seq][0], self.params[i_seq][1]\n x_pow_a = np... |
def x(self, kern, X, Z, likelihood, Y, indexD, output_dim, Y_metadata=None,
Lm=None, dL_dKmm=None, Kuu_sigma=None):
input_dim = Z.shape[0]
uncertain_inputs = isinstance(X, VariationalPosterior)
beta = 1.0 / likelihood.variance
if len(beta) == 1:
beta = np.zeros(output_dim) + beta
beta_exp = np.zeros(indexD.shape[0])
for d in range(output_dim):
beta_exp[indexD == d] = beta[d]
psi0, psi1, psi2 = self.gatherPsiStat(kern, X, Z, Y, beta, uncertain_inputs
)
psi2_sum = (beta_exp[:, None, None] * psi2).sum(0) / output_dim
Kmm = kern.K(Z).copy()
if Kuu_sigma is not None:
diag.add(Kmm, Kuu_sigma)
else:
diag.add(Kmm, self.const_jitter)
Lm = jitchol(Kmm)
logL = 0.0
dL_dthetaL = np.zeros(output_dim)
dL_dKmm = np.zeros_like(Kmm)
dL_dpsi0 = np.zeros_like(psi0)
dL_dpsi1 = np.zeros_like(psi1)
dL_dpsi2 = np.zeros_like(psi2)
wv = np.empty((Kmm.shape[0], output_dim))
for d in range(output_dim):
idx_d = indexD == d
Y_d = Y[idx_d]
N_d = Y_d.shape[0]
beta_d = beta[d]
psi2_d = psi2[idx_d].sum(0) * beta_d
psi1Y = Y_d.T.dot(psi1[idx_d]) * beta_d
psi0_d = psi0[idx_d].sum() * beta_d
YRY_d = np.square(Y_d).sum() * beta_d
LmInvPsi2LmInvT = backsub_both_sides(Lm, psi2_d, 'right')
Lambda = np.eye(Kmm.shape[0]) + LmInvPsi2LmInvT
LL = jitchol(Lambda)
LmLL = Lm.dot(LL)
b = dtrtrs(LmLL, psi1Y.T)[0].T
bbt = np.square(b).sum()
v = dtrtrs(LmLL, b.T, trans=1)[0].T
LLinvPsi1TYYTPsi1LLinvT = tdot(b.T)
tmp = -backsub_both_sides(LL, LLinvPsi1TYYTPsi1LLinvT)
dL_dpsi2R = backsub_both_sides(Lm, tmp + np.eye(input_dim)) / 2
logL_R = -N_d * np.log(beta_d)
logL += -(N_d * log_2_pi + logL_R + psi0_d - np.trace(
LmInvPsi2LmInvT) + YRY_d - bbt) / 2.0
dL_dKmm += dL_dpsi2R - backsub_both_sides(Lm, LmInvPsi2LmInvT) / 2
dL_dthetaL[d:d + 1] = (YRY_d * beta_d + beta_d * psi0_d - N_d * beta_d
) / 2.0 - beta_d * (dL_dpsi2R * psi2_d).sum() - beta_d * np.trace(
LLinvPsi1TYYTPsi1LLinvT)
dL_dpsi0[idx_d] = -beta_d / 2.0
dL_dpsi1[idx_d] = beta_d * np.dot(Y_d, v)
dL_dpsi2[idx_d] = beta_d * dL_dpsi2R
wv[:, d] = v
LmInvPsi2LmInvT = backsub_both_sides(Lm, psi2_sum, 'right')
Lambda = np.eye(Kmm.shape[0]) + LmInvPsi2LmInvT
LL = jitchol(Lambda)
LmLL = Lm.dot(LL)
logdet_L = 2.0 * np.sum(np.log(np.diag(LL)))
dL_dpsi2R_common = dpotri(LmLL)[0] / -2.0
dL_dpsi2 += dL_dpsi2R_common[None, :, :] * beta_exp[:, None, None]
for d in range(output_dim):
dL_dthetaL[d] += (dL_dpsi2R_common * psi2[indexD == d].sum(0)).sum(
) * -beta[d] * beta[d]
dL_dKmm += dL_dpsi2R_common * output_dim
logL += -output_dim * logdet_L / 2.0
LLInvLmT = dtrtrs(LL, Lm.T)[0]
cov = tdot(LLInvLmT.T)
wd_inv = backsub_both_sides(Lm, np.eye(input_dim) - backsub_both_sides(
LL, np.identity(input_dim), transpose='left'), transpose='left')
post = Posterior(woodbury_inv=wd_inv, woodbury_vector=wv, K=Kmm, mean=
None, cov=cov, K_chol=Lm)
if not uncertain_inputs:
dL_dpsi1 += (psi1[:, None, :] * dL_dpsi2).sum(2) * 2.0
if uncertain_inputs:
grad_dict = {'dL_dKmm': dL_dKmm, 'dL_dpsi0': dL_dpsi0, 'dL_dpsi1':
dL_dpsi1, 'dL_dpsi2': dL_dpsi2, 'dL_dthetaL': dL_dthetaL}
else:
grad_dict = {'dL_dKmm': dL_dKmm, 'dL_dKdiag': dL_dpsi0, 'dL_dKnm':
dL_dpsi1, 'dL_dthetaL': dL_dthetaL}
return post, logL, grad_dict
| inference | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, kern, X, Z, likelihood, Y, indexD, output_dim, Y_metadata=None,\n Lm=None, dL_dKmm=None, Kuu_sigma=None):\n input_dim = Z.shape[0]\n uncertain_i... |
def x(parameterized, fignum=None, ax=None, colors=None, figsize=(12, 6)):
if ax is None:
fig = pb.figure(num=fignum, figsize=figsize)
if colors is None:
from ..Tango import mediumList
from itertools import cycle
colors = cycle(mediumList)
pb.clf()
else:
colors = iter(colors)
lines = []
fills = []
bg_lines = []
means, variances = parameterized.mean.values, parameterized.variance.values
x = np.arange(means.shape[0])
for i in range(means.shape[1]):
if ax is None:
a = fig.add_subplot(means.shape[1], 1, i + 1)
elif isinstance(ax, (tuple, list)):
a = ax[i]
else:
raise ValueError('Need one ax per latent dimension input_dim')
bg_lines.append(a.plot(means, c='k', alpha=0.3))
lines.extend(a.plot(x, means.T[i], c=next(colors), label=
'$\\mathbf{{X_{{{}}}}}$'.format(i)))
fills.append(a.fill_between(x, means.T[i] - 2 * np.sqrt(variances.T
[i]), means.T[i] + 2 * np.sqrt(variances.T[i]), facecolor=lines
[-1].get_color(), alpha=0.3))
a.legend(borderaxespad=0.0)
a.set_xlim(x.min(), x.max())
if i < means.shape[1] - 1:
a.set_xticklabels('')
pb.draw()
a.figure.tight_layout(h_pad=0.01)
return dict(lines=lines, fills=fills, bg_lines=bg_lines)
| plot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(parameterized, fignum=None, ax=None, colors=None, figsize=(12, 6)):\n if ax is None:\n fig = pb.figure(num=fignum, figsize=figsize)\n if colors is... |
def x(parameterized, fignum=None, ax=None, colors=None, side_by_side=True):
if ax is None:
if side_by_side:
fig = pb.figure(num=fignum, figsize=(16, min(12, 2 *
parameterized.mean.shape[1])))
else:
fig = pb.figure(num=fignum, figsize=(8, min(12, 2 *
parameterized.mean.shape[1])))
if colors is None:
from ..Tango import mediumList
from itertools import cycle
colors = cycle(mediumList)
pb.clf()
else:
colors = iter(colors)
plots = []
means, variances, gamma = (parameterized.mean, parameterized.variance,
parameterized.binary_prob)
x = np.arange(means.shape[0])
for i in range(means.shape[1]):
if side_by_side:
sub1 = means.shape[1], 2, 2 * i + 1
sub2 = means.shape[1], 2, 2 * i + 2
else:
sub1 = means.shape[1] * 2, 1, 2 * i + 1
sub2 = means.shape[1] * 2, 1, 2 * i + 2
a = fig.add_subplot(*sub1)
a.plot(means, c='k', alpha=0.3)
plots.extend(a.plot(x, means.T[i], c=next(colors), label=
'$\\mathbf{{X_{{{}}}}}$'.format(i)))
a.fill_between(x, means.T[i] - 2 * np.sqrt(variances.T[i]), means.T
[i] + 2 * np.sqrt(variances.T[i]), facecolor=plots[-1].
get_color(), alpha=0.3)
a.legend(borderaxespad=0.0)
a.set_xlim(x.min(), x.max())
if i < means.shape[1] - 1:
a.set_xticklabels('')
a = fig.add_subplot(*sub2)
a.bar(x, gamma[:, i], bottom=0.0, linewidth=1.0, width=1.0, align=
'center')
a.set_xlim(x.min(), x.max())
a.set_ylim([0.0, 1.0])
pb.draw()
fig.tight_layout(h_pad=0.01)
return fig
| plot_SpikeSlab | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(parameterized, fignum=None, ax=None, colors=None, side_by_side=True):\n if ax is None:\n if side_by_side:\n fig = pb.figure(num=fignum, fi... |
def x(self, plot_limits=None, which_data='all', which_parts='all',
resolution=None, levels=20, samples=0, fignum=None, ax=None,
fixed_inputs=[], linecol=Tango.colorsHex['darkBlue']):
if which_data == 'all':
which_data = slice(None)
if ax is None:
fig = pb.figure(num=fignum)
ax = fig.add_subplot(111)
plotdims = self.input_dim - len(fixed_inputs)
from ..gpy_plot.plot_util import x_frame1D, x_frame2D
if plotdims == 1:
Xu = self.X * self._Xscale + self._Xoffset
fixed_dims = np.array([i for i, v in fixed_inputs])
freedim = np.setdiff1d(np.arange(self.input_dim), fixed_dims)
Xnew, xmin, xmax = x_frame1D(Xu[:, freedim], plot_limits=plot_limits)
Xgrid = np.empty((Xnew.shape[0], self.input_dim))
Xgrid[:, freedim] = Xnew
for i, v in fixed_inputs:
Xgrid[:, i] = v
f = self.predict(Xgrid, which_parts=which_parts)
for d in range(y.shape[1]):
ax.plot(Xnew, f[:, d], edgecol=linecol)
elif self.X.shape[1] == 2:
resolution = resolution or 50
Xnew, _, _, xmin, xmax = x_frame2D(self.X, plot_limits, resolution)
x, y = np.linspace(xmin[0], xmax[0], resolution), np.linspace(xmin[
1], xmax[1], resolution)
f = self.predict(Xnew, which_parts=which_parts)
m = m.reshape(resolution, resolution).T
ax.contour(x, y, f, levels, vmin=m.min(), vmax=m.max(), cmap=pb.cm.jet)
ax.set_xlim(xmin[0], xmax[0])
ax.set_ylim(xmin[1], xmax[1])
else:
raise NotImplementedError(
'Cannot define a frame with more than two input dimensions')
| plot_mapping | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, plot_limits=None, which_data='all', which_parts='all',\n resolution=None, levels=20, samples=0, fignum=None, ax=None,\n fixed_inputs=[], linecol=Ta... |
def x(self):
F, L, Qc, H, P_inf, P0, dFt, dQct, dP_inft, dP0t = self.kern.sde()
measurement_dim = self.output_dim
grad_params_no = dFt.shape[2] + 1
dF = np.zeros([dFt.shape[0], dFt.shape[1], grad_params_no])
dQc = np.zeros([dQct.shape[0], dQct.shape[1], grad_params_no])
dP_inf = np.zeros([dP_inft.shape[0], dP_inft.shape[1], grad_params_no])
dP0 = np.zeros([dP0t.shape[0], dP0t.shape[1], grad_params_no])
dF[:, :, :-1] = dFt
dQc[:, :, :-1] = dQct
dP_inf[:, :, :-1] = dP_inft
dP0[:, :, :-1] = dP0t
dR = np.zeros([measurement_dim, measurement_dim, grad_params_no])
dR[:, :, -1] = np.eye(measurement_dim)
if self.balance:
F, L, Qc, H, P_inf, P0, dF, dQc, dP_inf, dP0 = ssm.balance_ss_model(F,
L, Qc, H, P_inf, P0, dF, dQc, dP_inf, dP0)
print('SSM parameters_changed balancing!')
grad_calc_params = {}
grad_calc_params['dP_inf'] = dP_inf
grad_calc_params['dF'] = dF
grad_calc_params['dQc'] = dQc
grad_calc_params['dR'] = dR
grad_calc_params['dP_init'] = dP0
kalman_filter_type = self.kalman_filter_type
Y = self.Y
if self.ts_number is None:
Y.shape = self.num_data, 1
else:
Y.shape = self.num_data, 1, self.ts_number
(filter_means, filter_covs, log_likelihood, grad_log_likelihood,
SmootherMatrObject) = (ssm.ContDescrStateSpace.
cont_discr_kalman_filter(F, L, Qc, H, float(self.Gaussian_noise.
variance), P_inf, self.X, Y, m_init=None, P_init=P0,
p_kalman_filter_type=kalman_filter_type, calc_log_likelihood=True,
calc_grad_log_likelihood=True, grad_params_no=grad_params_no,
grad_calc_params=grad_calc_params))
if np.any(np.isfinite(log_likelihood) == False):
print('State-Space: NaN valkues in the log_likelihood')
if np.any(np.isfinite(grad_log_likelihood) == False):
print('State-Space: NaN values in the grad_log_likelihood')
grad_log_likelihood_sum = np.sum(grad_log_likelihood, axis=1)
grad_log_likelihood_sum.shape = grad_log_likelihood_sum.shape[0], 1
self._log_marginal_likelihood = np.sum(log_likelihood, axis=1)
self.likelihood.update_gradients(grad_log_likelihood_sum[-1, 0])
self.kern.sde_update_gradient_full(grad_log_likelihood_sum[:-1, 0])
| parameters_changed | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n F, L, Qc, H, P_inf, P0, dFt, dQct, dP_inft, dP0t = self.kern.sde()\n measurement_dim = self.output_dim\n grad_params_no = dFt.shape[2] + 1\n ... |
def x(self, Xnew=None, Ynew=None, filteronly=False, p_balance=False, **kw):
if Ynew is None:
Ynew = self.Y
if Xnew is not None:
X = np.vstack((self.X, Xnew))
Y = np.vstack((Ynew, np.nan * np.zeros(Xnew.shape)))
predict_only_training = False
else:
X = self.X
Y = Ynew
predict_only_training = True
_, return_index, return_inverse = np.unique(X, True, True)
X = X[return_index]
Y = Y[return_index]
F, L, Qc, H, P_inf, P0, dF, dQc, dP_inf, dP0 = self.kern.sde()
state_dim = F.shape[0]
if p_balance == True:
F, L, Qc, H, P_inf, P0, dF, dQc, dP_inf, dP0 = ssm.balance_ss_model(F,
L, Qc, H, P_inf, P0, dF, dQc, dP_inf, dP0)
print('SSM _raw_predict balancing!')
kalman_filter_type = self.kalman_filter_type
M, P, log_likelihood, grad_log_likelihood, SmootherMatrObject = (ssm.
ContDescrStateSpace.cont_discr_kalman_filter(F, L, Qc, H, float(
self.Gaussian_noise.variance), P_inf, X, Y, m_init=None, P_init=P0,
p_kalman_filter_type=kalman_filter_type, calc_log_likelihood=False,
calc_grad_log_likelihood=False))
if not filteronly:
M, P = ssm.ContDescrStateSpace.cont_discr_rts_smoother(state_dim, M,
P, p_dynamic_callables=SmootherMatrObject, X=X, F=F, L=L, Qc=Qc)
M = M[1:, :, :]
P = P[1:, :, :]
M = M[return_inverse, :, :]
P = P[return_inverse, :, :]
if not predict_only_training:
M = M[self.num_data:, :, :]
P = P[self.num_data:, :, :]
m = np.einsum('ijl,kj', M, H)
m.shape = m.shape[0], m.shape[1]
V = np.einsum('ij,ajk,kl', H, P, H.T)
V.shape = V.shape[0], V.shape[1]
return m, V
| raw_predict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Xnew=None, Ynew=None, filteronly=False, p_balance=False, **kw):\n if Ynew is None:\n Ynew = self.Y\n if Xnew is not None:\n X = np.vs... |
def x(self, Xnew=None, filteronly=False, include_likelihood=True, balance=
None, **kw):
if balance is None:
p_balance = self.balance
else:
p_balance = balance
m, V = self._raw_predict(Xnew, filteronly=filteronly, p_balance=p_balance)
if include_likelihood:
V += float(self.likelihood.variance)
return m, V
| predict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Xnew=None, filteronly=False, include_likelihood=True, balance=\n None, **kw):\n if balance is None:\n p_balance = self.balance\n else:\n ... |
def x(self, Xnew=None, quantiles=(2.5, 97.5), balance=None, **kw):
if balance is None:
p_balance = self.balance
else:
p_balance = balance
mu, var = self._raw_predict(Xnew, p_balance=p_balance)
return [(stats.norm.ppf(q / 100.0) * np.sqrt(var + float(self.
Gaussian_noise.variance)) + mu) for q in quantiles]
| predict_quantiles | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Xnew=None, quantiles=(2.5, 97.5), balance=None, **kw):\n if balance is None:\n p_balance = self.balance\n else:\n p_balance = balance... |
def x(Y, visualizer, frame_rate=30):
for y in Y:
visualizer.modify(y[None, :])
time.sleep(1.0 / float(frame_rate))
| data_play | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(Y, visualizer, frame_rate=30):\n for y in Y:\n visualizer.modify(y[None, :])\n time.sleep(1.0 / float(frame_rate))\n",
"role": "user"
},... |
def x(self, vals):
self.vals = vals.view(np.ndarray).copy()
y = self.model.predict(self.vals)[0]
self.data_visualize.modify(y)
self.latent_handle.set_data(self.vals[0, self.latent_index[0]], self.
vals[0, self.latent_index[1]])
self.axes.figure.canvas.draw()
| modify | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, vals):\n self.vals = vals.view(np.ndarray).copy()\n y = self.model.predict(self.vals)[0]\n self.data_visualize.modify(y)\n self.latent_handle... |
def x(self, boundary=0.05):
bs = [((self.vals[:, i].max() - self.vals[:, i].min()) * boundary) for
i in range(3)]
self.x_lim = np.array([self.vals[:, 0].min() - bs[0], self.vals[:, 0].
max() + bs[0]])
self.y_lim = np.array([self.vals[:, 1].min() - bs[1], self.vals[:, 1].
max() + bs[1]])
self.z_lim = np.array([self.vals[:, 2].min() - bs[2], self.vals[:, 2].
max() + bs[2]])
| initialize_axes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, boundary=0.05):\n bs = [((self.vals[:, i].max() - self.vals[:, i].min()) * boundary) for\n i in range(3)]\n self.x_lim = np.array([self.vals... |
def x(self):
if self.padding > 0:
channels = np.zeros((self.vals.shape[0], self.vals.shape[1] + self.
padding))
channels[:, 0:self.vals.shape[0]] = self.vals
else:
channels = self.vals
vals_mat = self.skel.to_xyz(channels.flatten())
self.vals = np.zeros_like(vals_mat)
self.vals[:, 0] = vals_mat[:, 0].copy()
self.vals[:, 1] = vals_mat[:, 2].copy()
self.vals[:, 2] = vals_mat[:, 1].copy()
| process_values | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.padding > 0:\n channels = np.zeros((self.vals.shape[0], self.vals.shape[1] + self.\n padding))\n channels[:, 0:sel... |
def x(self, figure=None, col=1, row=1, projection='2d', xlabel=None, ylabel
=None, zlabel=None, title=None, xlim=None, ylim=None, zlim=None, **kwargs):
raise NotImplementedError(
'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library'
)
| new_canvas | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, figure=None, col=1, row=1, projection='2d', xlabel=None, ylabel\n =None, zlabel=None, title=None, xlim=None, ylim=None, zlim=None, **kwargs):\n rai... |
def x(self, cavas, X, Y, Z=None, color=None, label=None, **kwargs):
raise NotImplementedError(
'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library'
)
| plot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, cavas, X, Y, Z=None, color=None, label=None, **kwargs):\n raise NotImplementedError(\n 'Implement all plot functions in AbstractPlottingLibrary... |
def x(self, canvas, X, Y, Z, color=None, label=None, **kwargs):
raise NotImplementedError(
'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library'
)
| surface | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, canvas, X, Y, Z, color=None, label=None, **kwargs):\n raise NotImplementedError(\n 'Implement all plot functions in AbstractPlottingLibrary in ... |
def x(self, canvas, X, Y, Z=None, color=None, vmin=None, vmax=None, label=
None, **kwargs):
raise NotImplementedError(
'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library'
)
| scatter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, canvas, X, Y, Z=None, color=None, vmin=None, vmax=None, label=\n None, **kwargs):\n raise NotImplementedError(\n 'Implement all plot functio... |
def x(self, canvas, x, height, width=0.8, bottom=0, color=None, label=None,
**kwargs):
raise NotImplementedError(
'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library'
)
| barplot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, canvas, x, height, width=0.8, bottom=0, color=None, label=None,\n **kwargs):\n raise NotImplementedError(\n 'Implement all plot functions in... |
def x(self, canvas, X, Y, error, color=None, label=None, **kwargs):
raise NotImplementedError(
'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library'
)
| xerrorbar | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, canvas, X, Y, error, color=None, label=None, **kwargs):\n raise NotImplementedError(\n 'Implement all plot functions in AbstractPlottingLibrary... |
def x(self, canvas, X, extent=None, label=None, vmin=None, vmax=None, **kwargs
):
raise NotImplementedError(
'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library'
)
| imshow | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, canvas, X, extent=None, label=None, vmin=None, vmax=None, **kwargs\n ):\n raise NotImplementedError(\n 'Implement all plot functions in Abst... |
def x(self, canvas, plot_function, extent=None, label=None, vmin=None, vmax
=None, **kwargs):
raise NotImplementedError(
'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library'
)
| imshow_interact | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, canvas, plot_function, extent=None, label=None, vmin=None, vmax\n =None, **kwargs):\n raise NotImplementedError(\n 'Implement all plot funct... |
def x(self, canvas, X, Y, C, Z=None, color=None, label=None, **kwargs):
raise NotImplementedError(
'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library'
)
| contour | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, canvas, X, Y, C, Z=None, color=None, label=None, **kwargs):\n raise NotImplementedError(\n 'Implement all plot functions in AbstractPlottingLib... |
def x(self, canvas, X, lower, upper, color=None, label=None, **kwargs):
raise NotImplementedError(
'Implement all plot functions in AbstractPlottingLibrary in order to use your own plotting library'
)
| fill_between | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, canvas, X, lower, upper, color=None, label=None, **kwargs):\n raise NotImplementedError(\n 'Implement all plot functions in AbstractPlottingLib... |
def x(self, rand_gen=None, *args, **kwargs):
if rand_gen is None:
rand_gen = np.random.normal
x = rand_gen(*args, size=self._size_transformed(), **kwargs)
updates = self.update_model()
self.update_model(False)
self.optimizer_array = x
x = self.param_array.copy()
[np.put(x, ind, p.rvs(ind.size)) for p, ind in self.priors.items() if
not p is None]
unfixlist = np.ones((self.size,), dtype=np.bool)
from paramz.transformations import __fixed__
unfixlist[self.constraints[__fixed__]] = False
self.param_array.flat[unfixlist] = x.view(np.ndarray).ravel()[unfixlist]
self.update_model(updates)
| randomize | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, rand_gen=None, *args, **kwargs):\n if rand_gen is None:\n rand_gen = np.random.normal\n x = rand_gen(*args, size=self._size_transformed(), *... |
def x(self, t, tprime, s, sprime, l):
return 0.5 * l ** 2 * (self.g((t - sprime) / l) + self.g((tprime - s) /
l) - self.g((t - tprime) / l) - self.g((s - sprime) / l))
| k_xx | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, t, tprime, s, sprime, l):\n return 0.5 * l ** 2 * (self.g((t - sprime) / l) + self.g((tprime - s) /\n l) - self.g((t - tprime) / l) - self.g((s... |
def x(self, X):
K_xx = np.ones([X.shape[0], X.shape[0]])
for i, x in enumerate(X):
for j, x2 in enumerate(X):
for il, l in enumerate(self.lengthscale):
idx = il * 2
K_xx[i, j] *= self.k_xx(x[idx], x2[idx], x[idx + 1], x2[idx +
1], l)
return K_xx
| calc_K_xx_wo_variance | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n K_xx = np.ones([X.shape[0], X.shape[0]])\n for i, x in enumerate(X):\n for j, x2 in enumerate(X):\n for il, l in enumerate(... |
def x(self, X):
K_ff = np.ones(X.shape[0])
for i, x in enumerate(X):
for il, l in enumerate(self.lengthscale):
idx = il * 2
K_ff[i] *= self.k_ff(x[idx], x[idx], l)
return K_ff * self.variances[0]
| Kdiag | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n K_ff = np.ones(X.shape[0])\n for i, x in enumerate(X):\n for il, l in enumerate(self.lengthscale):\n idx = il * 2\n ... |
def x(self, target_param=None, verbose=False, step=1e-06, tolerance=0.001,
block_indices=None, plot=False):
try:
import numdifftools as nd
except:
raise ImportError(
"Don't have numdifftools package installed, it is not a GPy dependency as of yet, it is only used for hessian tests"
)
if target_param:
raise NotImplementedError(
'Only basic functionality is provided with this gradchecker')
current_index = 0
for name, shape in zip(self.names, self.shapes):
current_size = numpy.prod(shape)
x = self.optimizer_array.copy()
x = x[current_index:current_index + current_size].reshape(shape)
analytic_hess = self._ddf(x)
if analytic_hess.shape[1] == 1:
analytic_hess = numpy.diagflat(analytic_hess)
numeric_hess_partial = nd.Jacobian(self._df, vectorized=False)
numeric_hess = numeric_hess_partial(x)
check_passed = self.checkgrad_block(analytic_hess, numeric_hess,
verbose=verbose, step=step, tolerance=tolerance, block_indices=
block_indices, plot=plot)
current_index += current_size
return check_passed
| checkgrad | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, target_param=None, verbose=False, step=1e-06, tolerance=0.001,\n block_indices=None, plot=False):\n try:\n import numdifftools as nd\n ex... |
def x(self, analytic_hess, numeric_hess, verbose=False, step=1e-06,
tolerance=0.001, block_indices=None, plot=False):
if analytic_hess.dtype is np.dtype('object'):
real_size = get_block_shapes(analytic_hess)
num_elements = np.sum(real_size)
if (num_elements, num_elements) == numeric_hess.shape:
numeric_hess = get_blocks(numeric_hess, real_size)
else:
tmp_numeric_hess = get_blocks(np.zeros((num_elements,
num_elements)), real_size)
tmp_numeric_hess[block_indices] = numeric_hess.copy()
numeric_hess = tmp_numeric_hess
if block_indices is not None:
analytic_hess = analytic_hess[block_indices]
numeric_hess = numeric_hess[block_indices]
elif analytic_hess.dtype is np.dtype('object'):
analytic_hess = unblock(analytic_hess)
numeric_hess = unblock(numeric_hess)
ratio = numeric_hess / numpy.where(analytic_hess == 0, 1e-10, analytic_hess
)
difference = numpy.abs(analytic_hess - numeric_hess)
check_passed = numpy.all(numpy.abs(1 - ratio) < tolerance
) or numpy.allclose(numeric_hess, analytic_hess, atol=tolerance)
if verbose:
if block_indices:
print('\nBlock {}'.format(block_indices))
else:
print('\nAll blocks')
header = ['Checked', 'Max-Ratio', 'Min-Ratio', 'Min-Difference',
'Max-Difference']
header_string = map(lambda x: ' | '.join(header), [header])
separator = '-' * len(header_string[0])
print('\n'.join([header_string[0], separator]))
min_r = '%.6f' % float(numpy.min(ratio))
max_r = '%.6f' % float(numpy.max(ratio))
max_d = '%.6f' % float(numpy.max(difference))
min_d = '%.6f' % float(numpy.min(difference))
cols = [max_r, min_r, min_d, max_d]
if check_passed:
checked = '\x1b[92m True \x1b[0m'
else:
checked = '\x1b[91m False \x1b[0m'
grad_string = '{} | {} | {} | {} | {} '.format(checked,
cols[0], cols[1], cols[2], cols[3])
print(grad_string)
if plot:
from matplotlib import pyplot as pb
fig, axes = pb.subplots(2, 2)
max_lim = numpy.max(numpy.vstack((analytic_hess, numeric_hess)))
min_lim = numpy.min(numpy.vstack((analytic_hess, numeric_hess)))
msa = axes[0, 0].matshow(analytic_hess, vmin=min_lim, vmax=max_lim)
axes[0, 0].set_title('Analytic hessian')
axes[0, 0].xaxis.set_ticklabels([None])
axes[0, 0].yaxis.set_ticklabels([None])
axes[0, 0].xaxis.set_ticks([None])
axes[0, 0].yaxis.set_ticks([None])
msn = axes[0, 1].matshow(numeric_hess, vmin=min_lim, vmax=max_lim)
pb.colorbar(msn, ax=axes[0, 1])
axes[0, 1].set_title('Numeric hessian')
axes[0, 1].xaxis.set_ticklabels([None])
axes[0, 1].yaxis.set_ticklabels([None])
axes[0, 1].xaxis.set_ticks([None])
axes[0, 1].yaxis.set_ticks([None])
msr = axes[1, 0].matshow(ratio)
pb.colorbar(msr, ax=axes[1, 0])
axes[1, 0].set_title('Ratio')
axes[1, 0].xaxis.set_ticklabels([None])
axes[1, 0].yaxis.set_ticklabels([None])
axes[1, 0].xaxis.set_ticks([None])
axes[1, 0].yaxis.set_ticks([None])
msd = axes[1, 1].matshow(difference)
pb.colorbar(msd, ax=axes[1, 1])
axes[1, 1].set_title('difference')
axes[1, 1].xaxis.set_ticklabels([None])
axes[1, 1].yaxis.set_ticklabels([None])
axes[1, 1].xaxis.set_ticks([None])
axes[1, 1].yaxis.set_ticks([None])
if block_indices:
fig.suptitle('Block: {}'.format(block_indices))
pb.show()
return check_passed
| checkgrad_block | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, analytic_hess, numeric_hess, verbose=False, step=1e-06,\n tolerance=0.001, block_indices=None, plot=False):\n if analytic_hess.dtype is np.dtype('o... |
def x(self, target_param=None, verbose=False, step=1e-06, tolerance=0.001,
block_indices=None, plot=False, super_plot=False):
try:
import numdifftools as nd
except:
raise ImportError(
"Don't have numdifftools package installed, it is not a GPy dependency as of yet, it is only used for hessian tests"
)
if target_param:
raise NotImplementedError(
'Only basic functionality is provided with this gradchecker')
current_index = 0
for name, n_shape in zip(self.names, self.shapes):
current_size = numpy.prod(n_shape)
x = self.optimizer_array.copy()
x = x[current_index:current_index + current_size].reshape(n_shape)
analytic_hess = self._ddf(x)
numeric_hess_partial = nd.Jacobian(self._df, vectorized=True)
numeric_hess = numeric_hess_partial(x)
print('Done making numerical hessian')
if analytic_hess.dtype is np.dtype('object'):
blocksizes, pagesizes = get_block_shapes_3d(analytic_hess)
real_block_size = np.sum(blocksizes)
numeric_hess = numeric_hess.reshape(real_block_size,
real_block_size, pagesizes)
else:
numeric_hess = numeric_hess.reshape(*analytic_hess.shape)
check_passed = [False] * numeric_hess.shape[2]
for block_ind in range(numeric_hess.shape[2]):
p = True if plot and block_ind == numeric_hess.shape[2
] - 1 or super_plot else False
if verbose:
print('Checking derivative of hessian wrt parameter number {}'
.format(block_ind))
check_passed[block_ind] = self.checkgrad_block(analytic_hess[:,
:, block_ind], numeric_hess[:, :, block_ind], verbose=
verbose, step=step, tolerance=tolerance, block_indices=
block_indices, plot=p)
current_index += current_size
return np.all(check_passed)
| checkgrad | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, target_param=None, verbose=False, step=1e-06, tolerance=0.001,\n block_indices=None, plot=False, super_plot=False):\n try:\n import numdifft... |
def x(self, X, X2=None):
r = self._scaled_dist(X, X2)
return self.dKdVar_of_r(r)
| dKd_dVar | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, X2=None):\n r = self._scaled_dist(X, X2)\n return self.dKdVar_of_r(r)\n",
"role": "user"
},
{
"content": "dKd_dVar",
"role": "as... |
def x(self, X, dimension, lengthscale, X2=None):
r = self._scaled_dist(X, X2)
return self.dKdLen_of_r(r, dimension, lengthscale)
| dKd_dLen | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, dimension, lengthscale, X2=None):\n r = self._scaled_dist(X, X2)\n return self.dKdLen_of_r(r, dimension, lengthscale)\n",
"role": "user"
}... |
def x(self, r, dimCheck, lengthscale):
if dimCheck == True:
return self.variance ** (float(1) / self.originalDimensions) * np.exp(
-0.5 * r ** 2) * r ** 2 / lengthscale ** (float(1) / self.
originalDimensions)
else:
return self.variance ** (float(1) / self.originalDimensions) * np.exp(
-0.5 * r ** 2) / lengthscale ** (float(1) / self.originalDimensions
)
| dKdLen_of_r | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, r, dimCheck, lengthscale):\n if dimCheck == True:\n return self.variance ** (float(1) / self.originalDimensions) * np.exp(\n -0.5 * ... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.