source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(num_samples=400, num_inducing=5, max_iters=100, optimize=True, plot=
True, checkgrad=False):
X = np.random.uniform(-3.0, 3.0, (num_samples, 1))
Y = np.sin(X) + np.random.randn(num_samples, 1) * 0.05
rbf = GPy.kern.RBF(1)
m = GPy.models.SparseGPRegression(X, Y, kernel=rbf, num_inducing=
num_inducing)
if checkgrad:
m.checkgrad()
if optimize:
m.optimize('tnc', max_iters=max_iters)
if plot:
m.plot()
return m
| sparse_GP_regression_1D | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(num_samples=400, num_inducing=5, max_iters=100, optimize=True, plot=\n True, checkgrad=False):\n X = np.random.uniform(-3.0, 3.0, (num_samples, 1))\n ... |
def x(num_samples=400, num_inducing=50, max_iters=100, optimize=True, plot=
True, nan=False):
np.random.seed(1234)
X = np.random.uniform(-3.0, 3.0, (num_samples, 2))
Y = np.sin(X[:, 0:1]) * np.sin(X[:, 1:2]) + np.random.randn(num_samples, 1
) * 0.05
if nan:
inan = np.random.binomial(1, 0.2, size=Y.shape)
Y[inan] = np.nan
rbf = GPy.kern.RBF(2)
m = GPy.models.SparseGPRegression(X, Y, kernel=rbf, num_inducing=
num_inducing)
m['.*len'] = 2.0
m.checkgrad()
if optimize:
m.optimize('tnc', messages=1, max_iters=max_iters)
if plot:
m.plot()
print(m)
return m
| sparse_GP_regression_2D | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(num_samples=400, num_inducing=50, max_iters=100, optimize=True, plot=\n True, nan=False):\n np.random.seed(1234)\n X = np.random.uniform(-3.0, 3.0, (n... |
def x(max_iters=200, optimize=True, plot=True):
fig, axes = pb.subplots(1, 2, figsize=(12, 5), sharex=True, sharey=True)
S = np.ones((20, 1))
X = np.random.uniform(-3.0, 3.0, (20, 1))
Y = np.sin(X) + np.random.randn(20, 1) * 0.05
Z = np.random.uniform(-3.0, 3.0, (7, 1))
k = GPy.kern.RBF(1)
m = GPy.models.SparseGPRegression(X, Y, kernel=k, Z=Z)
if optimize:
m.optimize('scg', messages=1, max_iters=max_iters)
if plot:
m.plot(ax=axes[0])
axes[0].set_title('no input uncertainty')
print(m)
m = GPy.models.SparseGPRegression(X, Y, kernel=GPy.kern.RBF(1), Z=Z,
X_variance=S)
if optimize:
m.optimize('scg', messages=1, max_iters=max_iters)
if plot:
m.plot(ax=axes[1])
axes[1].set_title('with input uncertainty')
fig.canvas.draw()
print(m)
return m
| uncertain_inputs_sparse_regression | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(max_iters=200, optimize=True, plot=True):\n fig, axes = pb.subplots(1, 2, figsize=(12, 5), sharex=True, sharey=True)\n S = np.ones((20, 1))\n X = np.r... |
def x(max_iters=100, optimize=True, plot=True):
mf = GPy.core.Mapping(1, 1)
mf.f = np.sin
mf.update_gradients = lambda a, b: None
X = np.linspace(0, 10, 50).reshape(-1, 1)
Y = np.sin(X) + 0.5 * np.cos(3 * X) + 0.1 * np.random.randn(*X.shape)
k = GPy.kern.RBF(1)
lik = GPy.likelihoods.Gaussian()
m = GPy.core.GP(X, Y, kernel=k, likelihood=lik, mean_function=mf)
if optimize:
m.optimize(max_iters=max_iters)
if plot:
m.plot(plot_limits=(-10, 15))
return m
| simple_mean_function | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(max_iters=100, optimize=True, plot=True):\n mf = GPy.core.Mapping(1, 1)\n mf.f = np.sin\n mf.update_gradients = lambda a, b: None\n X = np.linspace... |
def x(max_iters=100, optimize=True, plot=True):
mf = GPy.core.Mapping(1, 1)
mf.f = np.sin
X = np.linspace(0, 10, 50).reshape(-1, 1)
Y = np.sin(X) + 0.5 * np.cos(3 * X) + 0.1 * np.random.randn(*X.shape
) + 3 * X
mf = GPy.mappings.Linear(1, 1)
k = GPy.kern.RBF(1)
lik = GPy.likelihoods.Gaussian()
m = GPy.core.GP(X, Y, kernel=k, likelihood=lik, mean_function=mf)
if optimize:
m.optimize(max_iters=max_iters)
if plot:
m.plot()
return m
| parametric_mean_function | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(max_iters=100, optimize=True, plot=True):\n mf = GPy.core.Mapping(1, 1)\n mf.f = np.sin\n X = np.linspace(0, 10, 50).reshape(-1, 1)\n Y = np.sin(X)... |
def x(max_iters=100):
X = 2 * np.pi * np.random.random(151) - np.pi
Y = np.sin(X) + np.random.normal(0, 0.2, 151)
Y = np.array([(np.power(abs(y), float(1) / 3) * (1, -1)[y < 0]) for y in Y]
)
X = X[:, None]
Y = Y[:, None]
warp_k = GPy.kern.RBF(1)
warp_f = GPy.util.warping_functions.TanhFunction(n_terms=2)
warp_m = GPy.models.WarpedGP(X, Y, kernel=warp_k, warping_function=warp_f)
warp_m['.*\\.d'].constrain_fixed(1.0)
m = GPy.models.GPRegression(X, Y)
m.optimize_restarts(parallel=False, robust=True, num_restarts=5,
max_iters=max_iters)
warp_m.optimize_restarts(parallel=False, robust=True, num_restarts=5,
max_iters=max_iters)
print(warp_m)
print(warp_m['.*warp.*'])
warp_m.predict_in_warped_space = False
warp_m.plot(title='Warped GP - Latent space')
warp_m.predict_in_warped_space = True
warp_m.plot(title='Warped GP - Warped space')
m.plot(title='Standard GP')
warp_m.plot_warping()
pb.show()
| warped_gp_cubic_sine | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(max_iters=100):\n X = 2 * np.pi * np.random.random(151) - np.pi\n Y = np.sin(X) + np.random.normal(0, 0.2, 151)\n Y = np.array([(np.power(abs(y), floa... |
def x(self, X1new, X2new):
k1xf = self.kern1.K(X1new, self.X1)
k2xf = self.kern2.K(X2new, self.X2)
A = k1xf.dot(self.U1)
B = k2xf.dot(self.U2)
mu = A.dot(self.Ytilde.reshape(self.num_data1, self.num_data2, order='F')
).dot(B.T).flatten(order='F')
k1xx = self.kern1.Kdiag(X1new)
k2xx = self.kern2.Kdiag(X2new)
BA = np.kron(B, A)
var = np.kron(k2xx, k1xx) - np.sum(BA ** 2 * self.Wi, 1
) + self.likelihood.variance
return mu[:, None], var[:, None]
| predict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X1new, X2new):\n k1xf = self.kern1.K(X1new, self.X1)\n k2xf = self.kern2.K(X2new, self.X2)\n A = k1xf.dot(self.U1)\n B = k2xf.dot(self.U2)\n ... |
def x(self, Xgrid, plot_raw, apply_link, percentiles, which_data_ycols,
predict_kw, samples=0):
if predict_kw is None:
predict_kw = {}
if 'likelihood' not in predict_kw:
if plot_raw:
from ...likelihoods import Gaussian
from ...likelihoods.link_functions import Identity
lik = Gaussian(Identity(), 1e-09)
else:
lik = None
predict_kw['likelihood'] = lik
if 'Y_metadata' not in predict_kw:
predict_kw['Y_metadata'] = {}
if 'output_index' not in predict_kw['Y_metadata']:
predict_kw['Y_metadata']['output_index'] = Xgrid[:, -1:].astype(np.int)
mu, _ = self.predict(Xgrid, **predict_kw)
if percentiles is not None:
percentiles = self.predict_quantiles(Xgrid, quantiles=percentiles,
**predict_kw)
else:
percentiles = []
if samples > 0:
fsamples = self.posterior_samples(Xgrid, size=samples, **predict_kw)
fsamples = fsamples[:, which_data_ycols, :]
else:
fsamples = None
retmu = mu[:, which_data_ycols]
percs = [p[:, which_data_ycols] for p in percentiles]
if plot_raw and apply_link:
for i in range(len(which_data_ycols)):
retmu[:, [i]] = self.likelihood.gp_link.transf(mu[:, [i]])
for perc in percs:
perc[:, [i]] = self.likelihood.gp_link.transf(perc[:, [i]])
if fsamples is not None:
for s in range(fsamples.shape[-1]):
fsamples[:, i, s] = self.likelihood.gp_link.transf(fsamples
[:, i, s])
return retmu, percs, fsamples
| helper_predict_with_model | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Xgrid, plot_raw, apply_link, percentiles, which_data_ycols,\n predict_kw, samples=0):\n if predict_kw is None:\n predict_kw = {}\n if 'li... |
def x(self, X, plot_limits, visible_dims, fixed_inputs, resolution):
if fixed_inputs is None:
fixed_inputs = []
fixed_dims = get_fixed_dims(fixed_inputs)
free_dims = get_free_dims(self, visible_dims, fixed_dims)
if len(free_dims) == 1:
resolution = resolution or 200
Xnew, xmin, xmax = x_frame1D(X[:, free_dims], plot_limits=
plot_limits, resolution=resolution)
Xgrid = np.zeros((Xnew.shape[0], self.input_dim))
Xgrid[:, free_dims] = Xnew
for i, v in fixed_inputs:
Xgrid[:, i] = v
x = Xgrid
y = None
elif len(free_dims) == 2:
resolution = resolution or 35
Xnew, x, y, xmin, xmax = x_frame2D(X[:, free_dims], plot_limits,
resolution)
Xgrid = np.zeros((Xnew.shape[0], self.input_dim))
Xgrid[:, free_dims] = Xnew
for i, v in fixed_inputs:
Xgrid[:, i] = v
else:
raise TypeError(
'calculated free_dims {} from visible_dims {} and fixed_dims {} is neither 1D nor 2D'
.format(free_dims, visible_dims, fixed_dims))
return fixed_dims, free_dims, Xgrid, x, y, xmin, xmax, resolution
| helper_for_plot_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, plot_limits, visible_dims, fixed_inputs, resolution):\n if fixed_inputs is None:\n fixed_inputs = []\n fixed_dims = get_fixed_dims(fixed_... |
def x(X, labels, num_samples=1000):
if X.shape[0] > num_samples:
print(
'Warning: subsampling X, as it has more samples then {}. X.shape={!s}'
.format(int(num_samples), X.shape))
if labels is not None:
subsample = []
for _, _, _, _, index, _ in scatter_label_generator(labels, X,
(0, None, None)):
subsample.append(np.random.choice(index, size=max(2, int(
index.size * (float(num_samples) / X.shape[0]))),
replace=False))
subsample = np.hstack(subsample)
else:
subsample = np.random.choice(X.shape[0], size=1000, replace=False)
X = X[subsample]
labels = labels[subsample]
return X, labels
| subsample_X | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(X, labels, num_samples=1000):\n if X.shape[0] > num_samples:\n print(\n 'Warning: subsampling X, as it has more samples then {}. X.shape={... |
def x(to_update, update_from):
if to_update is None:
to_update = {}
to_update.update({k: v for k, v in update_from.items() if k not in
to_update})
return to_update
| update_not_existing_kwargs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(to_update, update_from):\n if to_update is None:\n to_update = {}\n to_update.update({k: v for k, v in update_from.items() if k not in\n to... |
def x(model):
if hasattr(model, 'has_uncertain_inputs') and model.has_uncertain_inputs():
X = model.X.mean.values
X_variance = model.X.variance.values
else:
try:
X = model.X.values
except AttributeError:
X = model.X
X_variance = None
try:
Y = model.Y.values
except AttributeError:
Y = model.Y
if isinstance(model, WarpedGP) and not model.predict_in_warped_space:
Y = model.Y_normalized
if sparse.issparse(Y):
Y = Y.todense().view(np.ndarray)
return X, X_variance, Y
| get_x_y_var | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n if hasattr(model, 'has_uncertain_inputs') and model.has_uncertain_inputs():\n X = model.X.mean.values\n X_variance = model.X.varianc... |
def x(model, visible_dims, fixed_dims):
if visible_dims is None:
visible_dims = np.arange(model.input_dim)
dims = np.asanyarray(visible_dims)
if fixed_dims is not None:
dims = [dim for dim in dims if dim not in fixed_dims]
return np.asanyarray([dim for dim in dims if dim is not None])
| get_free_dims | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model, visible_dims, fixed_dims):\n if visible_dims is None:\n visible_dims = np.arange(model.input_dim)\n dims = np.asanyarray(visible_dims)\n ... |
def x(model, which_data_ycols):
if which_data_ycols == 'all' or which_data_ycols is None:
return np.arange(model.output_dim)
return which_data_ycols
| get_which_data_ycols | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model, which_data_ycols):\n if which_data_ycols == 'all' or which_data_ycols is None:\n return np.arange(model.output_dim)\n return which_data_yco... |
def x(X, plot_limits=None, resolution=None):
assert X.shape[1] == 1, 'x_frame1D is defined for one-dimensional inputs'
if plot_limits is None:
from GPy.core.parameterization.variational import VariationalPosterior
if isinstance(X, VariationalPosterior):
xmin, xmax = X.mean.min(0), X.mean.max(0)
else:
xmin, xmax = X.min(0), X.max(0)
xmin, xmax = xmin - 0.25 * (xmax - xmin), xmax + 0.25 * (xmax - xmin)
elif len(plot_limits) == 2:
xmin, xmax = map(np.atleast_1d, plot_limits)
else:
raise ValueError('Bad limits for plotting')
Xnew = np.linspace(float(xmin), float(xmax), int(resolution) or 200)[:,
None]
return Xnew, xmin, xmax
| x_frame1D | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(X, plot_limits=None, resolution=None):\n assert X.shape[1] == 1, 'x_frame1D is defined for one-dimensional inputs'\n if plot_limits is None:\n fro... |
def x(X, plot_limits=None, resolution=None):
assert X.shape[1] == 2, 'x_frame2D is defined for two-dimensional inputs'
if plot_limits is None:
xmin, xmax = X.min(0), X.max(0)
xmin, xmax = xmin - 0.075 * (xmax - xmin), xmax + 0.075 * (xmax - xmin)
elif len(plot_limits) == 2:
xmin, xmax = plot_limits
try:
xmin = xmin[0], xmin[1]
except:
xmin = [plot_limits[0], plot_limits[0]]
xmax = [plot_limits[1], plot_limits[1]]
elif len(plot_limits) == 4:
xmin, xmax = (plot_limits[0], plot_limits[2]), (plot_limits[1],
plot_limits[3])
else:
raise ValueError('Bad limits for plotting')
resolution = resolution or 50
xx, yy = np.mgrid[xmin[0]:xmax[0]:1.0j * resolution, xmin[1]:xmax[1]:
1.0j * resolution]
Xnew = np.c_[xx.flat, yy.flat]
return Xnew, xx, yy, xmin, xmax
| x_frame2D | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(X, plot_limits=None, resolution=None):\n assert X.shape[1] == 2, 'x_frame2D is defined for two-dimensional inputs'\n if plot_limits is None:\n xmi... |
def x(self, X):
X = X.copy()
inan = numpy.isnan(X)
if self.mu is None:
X_ = numpy.ma.masked_array(X, inan)
self.mu = X_.mean(0).base
self.sigma = X_.std(0).base
reduce(lambda y, x: setitem(x[0], x[1], x[2]), zip(X.T, inan.T, self.mu
), None)
X = X - self.mu
X = X / numpy.where(self.sigma == 0, 1e-30, self.sigma)
return X
| center | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n X = X.copy()\n inan = numpy.isnan(X)\n if self.mu is None:\n X_ = numpy.ma.masked_array(X, inan)\n self.mu = X_.mean(0).base... |
def x(self, X, Q=None):
if Q is None:
Q = self.Q
if Q > X.shape[1]:
raise IndexError('requested dimension larger then input dimension')
X = self.center(X)
return X.dot(self.eigvectors[:, :Q])
| project | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, Q=None):\n if Q is None:\n Q = self.Q\n if Q > X.shape[1]:\n raise IndexError('requested dimension larger then input dimension')\n... |
def x(self, Q=None, ax=None, fignum=None):
from ..plotting import Tango
Tango.reset()
col = Tango.nextMedium()
if ax is None:
fig = pylab.figure(fignum)
ax = fig.add_subplot(111)
if Q is None:
Q = self.Q
ticks = numpy.arange(Q)
bar = ax.bar(ticks - 0.4, self.fracs[:Q], color=col)
ax.set_xticks(ticks, map(lambda x: '${}$'.format(x), ticks + 1))
ax.set_ylabel('Eigenvalue fraction')
ax.set_xlabel('PC')
ax.set_ylim(0, ax.get_ylim()[1])
ax.set_xlim(ticks.min() - 0.5, ticks.max() + 0.5)
try:
pylab.tight_layout()
except:
pass
return bar
| plot_fracs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Q=None, ax=None, fignum=None):\n from ..plotting import Tango\n Tango.reset()\n col = Tango.nextMedium()\n if ax is None:\n fig = pyla... |
def x(self, X, labels=None, s=20, marker='o', dimensions=(0, 1), ax=None,
colors=None, fignum=None, cmap=None, **kwargs):
if cmap is None:
cmap = matplotlib.cm.jet
if ax is None:
fig = pylab.figure(fignum)
ax = fig.add_subplot(111)
if labels is None:
labels = numpy.zeros(X.shape[0])
ulabels = []
for lab in labels:
if not lab in ulabels:
ulabels.append(lab)
nlabels = len(ulabels)
if colors is None:
colors = iter([cmap(float(i) / nlabels) for i in range(nlabels)])
else:
colors = iter(colors)
X_ = self.project(X, self.Q)[:, dimensions]
kwargs.update(dict(s=s))
plots = list()
for i, l in enumerate(ulabels):
kwargs.update(dict(color=next(colors), marker=marker[i % len(marker)]))
plots.append(ax.scatter(*X_[labels == l, :].T, label=str(l), **kwargs))
ax.set_xlabel('PC$_1$')
ax.set_ylabel('PC$_2$')
try:
pylab.tight_layout()
except:
pass
return plots
| plot_2d | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, labels=None, s=20, marker='o', dimensions=(0, 1), ax=None,\n colors=None, fignum=None, cmap=None, **kwargs):\n if cmap is None:\n cmap = ... |
def x(self, labels=None, which_indices=None, legend=True, plot_limits=None,
marker='<>^vsd', num_samples=1000, projection='2d', **kwargs):
canvas, projection, kwargs, sig_dims = _new_canvas(self, projection,
kwargs, which_indices)
X, _, _ = get_x_y_var(self)
if labels is None:
labels = np.ones(self.num_data)
legend = False
else:
legend = find_best_layout_for_subplots(len(np.unique(labels)))[1]
scatters = _plot_latent_scatter(canvas, X, sig_dims, labels, marker,
num_samples, projection=projection, **kwargs)
return pl().add_to_canvas(canvas, dict(scatter=scatters), legend=legend)
| plot_latent_scatter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, labels=None, which_indices=None, legend=True, plot_limits=None,\n marker='<>^vsd', num_samples=1000, projection='2d', **kwargs):\n canvas, projecti... |
def x(self, which_indices=None, legend=False, plot_limits=None, marker=None,
projection='2d', **kwargs):
canvas, projection, kwargs, sig_dims = _new_canvas(self, projection,
kwargs, which_indices)
if legend:
label = 'inducing'
else:
label = None
if marker is not None:
kwargs['marker'] = marker
update_not_existing_kwargs(kwargs, pl().defaults.inducing_2d)
from .data_plots import _plot_inducing
scatters = _plot_inducing(self, canvas, sig_dims[:2], projection, label,
**kwargs)
return pl().add_to_canvas(canvas, dict(scatter=scatters), legend=legend)
| plot_latent_inducing | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, which_indices=None, legend=False, plot_limits=None, marker=None,\n projection='2d', **kwargs):\n canvas, projection, kwargs, sig_dims = _new_canvas... |
def x(self, labels=None, which_indices=None, resolution=60, marker='<>^vsd',
legend=True, plot_limits=None, updates=False, mean=True, covariance=
True, kern=None, num_samples=1000, scatter_kwargs=None, plot_scatter=
True, **imshow_kwargs):
input_1, input_2 = which_indices = (self.
get_most_significant_input_dimensions(which_indices)[:2])
X = get_x_y_var(self)[0]
_, _, Xgrid, _, _, xmin, xmax, resolution = helper_for_plot_data(self,
X, plot_limits, which_indices, None, resolution)
canvas, imshow_kwargs = pl().new_canvas(xlim=(xmin[0], xmax[0]), ylim=(
xmin[1], xmax[1]), xlabel='latent dimension %i' % input_1, ylabel=
'latent dimension %i' % input_2, **imshow_kwargs)
plots = {}
if legend and plot_scatter:
if labels is not None:
legend = find_best_layout_for_subplots(len(np.unique(labels)))[1]
else:
labels = np.ones(self.num_data)
legend = False
if plot_scatter:
plots['scatters'] = _plot_latent_scatter(canvas, X, which_indices,
labels, marker, num_samples, projection='2d', **scatter_kwargs or
{})
plots['view'] = _plot_magnification(self, canvas, which_indices, Xgrid,
xmin, xmax, resolution, updates, mean, covariance, kern, **
imshow_kwargs)
retval = pl().add_to_canvas(canvas, plots, legend=legend)
_wait_for_updates(plots['view'], updates)
return retval
| plot_magnification | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, labels=None, which_indices=None, resolution=60, marker='<>^vsd',\n legend=True, plot_limits=None, updates=False, mean=True, covariance=\n True, ker... |
def x(self, labels=None, which_indices=None, resolution=60, legend=True,
plot_limits=None, updates=False, kern=None, marker='<>^vsd',
num_samples=1000, projection='2d', scatter_kwargs=None, **imshow_kwargs):
if projection != '2d':
raise ValueError(
'Cannot plot latent in other then 2 dimensions, consider plot_scatter'
)
input_1, input_2 = which_indices = (self.
get_most_significant_input_dimensions(which_indices)[:2])
X = get_x_y_var(self)[0]
_, _, Xgrid, _, _, xmin, xmax, resolution = helper_for_plot_data(self,
X, plot_limits, which_indices, None, resolution)
canvas, imshow_kwargs = pl().new_canvas(xlim=(xmin[0], xmax[0]), ylim=(
xmin[1], xmax[1]), xlabel='latent dimension %i' % input_1, ylabel=
'latent dimension %i' % input_2, **imshow_kwargs)
if legend:
if labels is not None:
legend = find_best_layout_for_subplots(len(np.unique(labels)))[1]
else:
labels = np.ones(self.num_data)
legend = False
scatters = _plot_latent_scatter(canvas, X, which_indices, labels,
marker, num_samples, projection='2d', **scatter_kwargs or {})
view = _plot_latent(self, canvas, which_indices, Xgrid, xmin, xmax,
resolution, updates, kern, **imshow_kwargs)
retval = pl().add_to_canvas(canvas, dict(scatter=scatters, imshow=view),
legend=legend)
_wait_for_updates(view, updates)
return retval
| plot_latent | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, labels=None, which_indices=None, resolution=60, legend=True,\n plot_limits=None, updates=False, kern=None, marker='<>^vsd',\n num_samples=1000, pro... |
def x(self, output_labels=None, data_labels=None, which_indices=None,
resolution=15, legend=True, plot_limits=None, updates=False, kern=None,
marker='<>^vsd', num_samples=1000, annotation_kwargs=None,
scatter_kwargs=None, **imshow_kwargs):
input_1, input_2 = which_indices = (self.
get_most_significant_input_dimensions(which_indices)[:2])
X = get_x_y_var(self)[0]
_, _, Xgrid, _, _, xmin, xmax, resolution = helper_for_plot_data(self,
X, plot_limits, which_indices, None, resolution)
canvas, imshow_kwargs = pl().new_canvas(xlim=(xmin[0], xmax[0]), ylim=(
xmin[1], xmax[1]), xlabel='latent dimension %i' % input_1, ylabel=
'latent dimension %i' % input_2, **imshow_kwargs)
if data_labels is not None:
legend = find_best_layout_for_subplots(len(np.unique(data_labels)))[1]
else:
data_labels = np.ones(self.num_data)
legend = False
plots = dict(scatter=_plot_latent_scatter(canvas, X, which_indices,
data_labels, marker, num_samples, **scatter_kwargs or {}))
plots.update(_plot_steepest_gradient_map(self, canvas, which_indices,
Xgrid, xmin, xmax, resolution, output_labels, updates, kern,
annotation_kwargs=annotation_kwargs, **imshow_kwargs))
retval = pl().add_to_canvas(canvas, plots, legend=legend)
_wait_for_updates(plots['annotation'], updates)
return retval
| plot_steepest_gradient_map | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, output_labels=None, data_labels=None, which_indices=None,\n resolution=15, legend=True, plot_limits=None, updates=False, kern=None,\n marker='<>^vs... |
def x(self, expressions):
Symbolic_core._set_expressions(self, expressions)
Symbolic_core._set_variables(self, self.cacheable)
for x, z in zip(self.variables['X'], self.variables['Z']):
expressions['kdiag'] = expressions['kdiag'].subs(z, x)
Symbolic_core._set_expressions(self, expressions)
| set_expressions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, expressions):\n Symbolic_core._set_expressions(self, expressions)\n Symbolic_core._set_variables(self, self.cacheable)\n for x, z in zip(self.va... |
def x(M):
if M.shape[0] == 1:
Mexp = np.array(((np.exp(M[0, 0]),),))
else:
method = None
try:
Mexp = linalg.expm(M)
method = 1
except (Exception,) as e:
Mexp = linalg.expm3(M)
method = 2
finally:
if np.any(np.isnan(Mexp)):
if method == 2:
raise ValueError('Matrix Exponent is not computed 1')
else:
Mexp = linalg.expm3(M)
method = 2
if np.any(np.isnan(Mexp)):
raise ValueError('Matrix Exponent is not computed 2')
return Mexp
| matrix_exponent | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(M):\n if M.shape[0] == 1:\n Mexp = np.array(((np.exp(M[0, 0]),),))\n else:\n method = None\n try:\n Mexp = linalg.expm(M)... |
def x(A):
if len(A.shape) != 2 or A.shape[0] != A.shape[1]:
raise ValueError('balance_matrix: Expecting square matrix')
N = A.shape[0]
gebal = sp.linalg.lapack.get_lapack_funcs('gebal', (A,))
bA, lo, hi, pivscale, info = gebal(A, permute=True, scale=True,
overwrite_a=False)
if info < 0:
raise ValueError(
'balance_matrix: Illegal value in %d-th argument of internal gebal '
% -info)
def x(D, c1, c2):
i1 = c1 - 1
i2 = c2 - 1
P = np.eye(D)
P[i1, i1] = 0.0
P[i2, i2] = 0.0
P[i1, i2] = 1.0
P[i2, i1] = 1.0
return P
P = np.eye(N)
if hi != N - 1:
for k in range(N - 1, hi, -1):
new_perm = perm_matr(N, k + 1, pivscale[k])
P = np.dot(P, new_perm)
if lo != 0:
for k in range(0, lo, 1):
new_perm = perm_matr(N, k + 1, pivscale[k])
P = np.dot(P, new_perm)
D = pivscale.copy()
D[0:lo] = 1.0
D[hi + 1:N] = 1.0
T = np.dot(P, np.diag(D))
T_inv = np.dot(np.diag(D ** -1), P.T)
return bA.copy(), T, T_inv
| balance_matrix | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A):\n if len(A.shape) != 2 or A.shape[0] != A.shape[1]:\n raise ValueError('balance_matrix: Expecting square matrix')\n N = A.shape[0]\n gebal ... |
def x(F, L, Qc, H, Pinf, P0, dF=None, dQc=None, dPinf=None, dP0=None):
bF, T, T_inv = balance_matrix(F)
bL = np.dot(T_inv, L)
bQc = Qc
bH = np.dot(H, T)
bPinf = np.dot(T_inv, np.dot(Pinf, T_inv.T))
bP0 = np.dot(T_inv, np.dot(P0, T_inv.T))
if dF is not None:
bdF = np.zeros(dF.shape)
for i in range(dF.shape[2]):
bdF[:, :, i] = np.dot(T_inv, np.dot(dF[:, :, i], T))
else:
bdF = None
if dPinf is not None:
bdPinf = np.zeros(dPinf.shape)
for i in range(dPinf.shape[2]):
bdPinf[:, :, i] = np.dot(T_inv, np.dot(dPinf[:, :, i], T_inv.T))
else:
bdPinf = None
if dP0 is not None:
bdP0 = np.zeros(dP0.shape)
for i in range(dP0.shape[2]):
bdP0[:, :, i] = np.dot(T_inv, np.dot(dP0[:, :, i], T_inv.T))
else:
bdP0 = None
bdQc = dQc
return bF, bL, bQc, bH, bPinf, bP0, bdF, bdQc, bdPinf, bdP0
| balance_ss_model | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(F, L, Qc, H, Pinf, P0, dF=None, dQc=None, dPinf=None, dP0=None):\n bF, T, T_inv = balance_matrix(F)\n bL = np.dot(T_inv, L)\n bQc = Qc\n bH = np.do... |
def x(self, k):
ind = int(self.index[self.R_time_var_index, k])
R = self.R[:, :, ind]
if R.shape[0] == 1:
inv_square_root = np.sqrt(1.0 / R)
elif self.svd_each_time:
U, S, Vh = sp.linalg.svd(R, full_matrices=False, compute_uv=True,
overwrite_a=False, check_finite=True)
inv_square_root = U * 1.0 / np.sqrt(S)
elif ind in self.R_square_root:
inv_square_root = self.R_square_root[ind]
else:
U, S, Vh = sp.linalg.svd(R, full_matrices=False, compute_uv=True,
overwrite_a=False, check_finite=True)
inv_square_root = U * 1.0 / np.sqrt(S)
self.R_square_root[ind] = inv_square_root
return inv_square_root
| R_isrk | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, k):\n ind = int(self.index[self.R_time_var_index, k])\n R = self.R[:, :, ind]\n if R.shape[0] == 1:\n inv_square_root = np.sqrt(1.0 / R)\... |
def x(self, k, m_pred, P_pred):
return self.H[:, :, int(self.index[self.H_time_var_index, k])]
| Hk | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, k, m_pred, P_pred):\n return self.H[:, :, int(self.index[self.H_time_var_index, k])]\n",
"role": "user"
},
{
"content": "Hk",
"role": ... |
def x(self, k):
return self.Q[:, :, self.index[self.Q_time_var_index, k]]
| Qk | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, k):\n return self.Q[:, :, self.index[self.Q_time_var_index, k]]\n",
"role": "user"
},
{
"content": "Qk",
"role": "assistant"
}
] |
def x(self, k):
ind = self.index[self.Q_time_var_index, k]
Q = self.Q[:, :, ind]
if Q.shape[0] == 1:
square_root = np.sqrt(Q)
elif self.svd_each_time:
U, S, Vh = sp.linalg.svd(Q, full_matrices=False, compute_uv=True,
overwrite_a=False, check_finite=True)
square_root = U * np.sqrt(S)
elif ind in self.Q_square_root:
square_root = self.Q_square_root[ind]
else:
U, S, Vh = sp.linalg.svd(Q, full_matrices=False, compute_uv=True,
overwrite_a=False, check_finite=True)
square_root = U * np.sqrt(S)
self.Q_square_root[ind] = square_root
return square_root
| Q_srk | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, k):\n ind = self.index[self.Q_time_var_index, k]\n Q = self.Q[:, :, ind]\n if Q.shape[0] == 1:\n square_root = np.sqrt(Q)\n elif self.... |
def x(self, k, m_pred, P_pred):
return self.A[:, :, self.index[self.A_time_var_index, k]]
| Ak | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, k, m_pred, P_pred):\n return self.A[:, :, self.index[self.A_time_var_index, k]]\n",
"role": "user"
},
{
"content": "Ak",
"role": "assi... |
def x(shape, desired_dim=3):
if len(shape) > 3:
raise ValueError(
"""Input array is not supposed to be more
than 3 dimensional."""
)
if len(shape) > desired_dim:
raise ValueError('Input array shape is more than desired shape.')
elif len(shape) == 1:
if desired_dim == 3:
return (shape[0], 1, 1), shape
elif desired_dim == 2:
return (shape[0], 1), shape
elif len(shape) == 2:
if desired_dim == 3:
return ((shape[1], 1, 1), shape) if shape[0] == 1 else ((shape[
0], shape[1], 1), shape)
elif desired_dim == 2:
return ((shape[1], 1), shape) if shape[0] == 1 else ((shape[0],
shape[1]), None)
else:
return shape, None
| reshape_input_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(shape, desired_dim=3):\n if len(shape) > 3:\n raise ValueError(\n \"\"\"Input array is not supposed to be more\n ... |
def x(cls, p_state_dim, p_a, p_f_A, p_f_Q, p_h, p_f_H, p_f_R, Y, m_init=
None, P_init=None, calc_log_likelihood=False):
Y.shape, old_Y_shape = cls._reshape_input_data(Y.shape)
if m_init is None:
m_init = np.zeros((p_state_dim, 1))
else:
m_init = np.atleast_2d(m_init).T
if P_init is None:
P_init = np.eye(p_state_dim)
elif not isinstance(P_init, collections.Iterable):
P_init = P_init * np.eye(p_state_dim)
if p_a is None:
p_a = lambda k, m, A: np.dot(A, m)
old_A_shape = None
if not isinstance(p_f_A, types.FunctionType):
p_f_A = np.atleast_1d(p_f_A)
p_A, old_A_shape = cls._check_A_matrix(p_f_A)
p_f_A = lambda k, m, P: p_A[:, :, 0]
elif p_f_A(1, m_init, P_init).shape[0] != m_init.shape[0]:
raise ValueError('p_f_A function returns matrix of wrong size')
old_Q_shape = None
if not isinstance(p_f_Q, types.FunctionType):
p_f_Q = np.atleast_1d(p_f_Q)
p_Q, old_Q_shape = cls._check_Q_matrix(p_f_Q)
p_f_Q = lambda k: p_Q[:, :, 0]
elif p_f_Q(1).shape[0] != m_init.shape[0]:
raise ValueError('p_f_Q function returns matrix of wrong size')
if p_h is None:
lambda k, m, H: np.dot(H, m)
old_H_shape = None
if not isinstance(p_f_H, types.FunctionType):
p_f_H = np.atleast_1d(p_f_H)
p_H, old_H_shape = cls._check_H_matrix(p_f_H)
p_f_H = lambda k, m, P: p_H
elif p_f_H(1, m_init, P_init).shape[0] != Y.shape[1]:
raise ValueError('p_f_H function returns matrix of wrong size')
old_R_shape = None
if not isinstance(p_f_R, types.FunctionType):
p_f_R = np.atleast_1d(p_f_R)
p_R, old_R_shape = cls._check_H_matrix(p_f_R)
p_f_R = lambda k: p_R
elif p_f_R(1).shape[0] != m_init.shape[0]:
raise ValueError('p_f_R function returns matrix of wrong size')
class measurement_callables_class(R_handling_Class):
def x(self, R, index, R_time_var_index, unique_R_number):
super(measurement_callables_class, self).__init__(R, index,
R_time_var_index, unique_R_number)
Hk = AddMethodToClass(f_H)
f_h = AddMethodToClass(f_hl)
M, P, log_likelihood, grad_log_likelihood = cls._kalman_algorithm_raw(
p_state_dim, p_a, p_f_A, p_f_Q, p_h, p_f_H, p_f_R, Y, m_init,
P_init, calc_log_likelihood, calc_grad_log_likelihood=False,
grad_calc_params=None)
if old_Y_shape is not None:
Y.shape = old_Y_shape
if old_A_shape is not None:
p_A.shape = old_A_shape
if old_Q_shape is not None:
p_Q.shape = old_Q_shape
if old_H_shape is not None:
p_H.shape = old_H_shape
if old_R_shape is not None:
p_R.shape = old_R_shape
return M, P
| extended_kalman_filter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, p_state_dim, p_a, p_f_A, p_f_Q, p_h, p_f_H, p_f_R, Y, m_init=\n None, P_init=None, calc_log_likelihood=False):\n Y.shape, old_Y_shape = cls._reshape... |
def x(cls, state_dim, p_dynamic_callables, p_measurement_callables, Y,
m_init, P_init, p_kalman_filter_type='regular', calc_log_likelihood=
False, calc_grad_log_likelihood=False, grad_params_no=None, dm_init=
None, dP_init=None):
steps_no = Y.shape[0]
time_series_no = Y.shape[2]
M = np.empty((steps_no + 1, state_dim, time_series_no))
M[0, :, :] = m_init
P = np.empty((steps_no + 1, state_dim, state_dim))
P_init = 0.5 * (P_init + P_init.T)
P[0, :, :] = P_init
if p_kalman_filter_type == 'svd':
U, S, Vh = sp.linalg.svd(P_init, full_matrices=False, compute_uv=
True, overwrite_a=False, check_finite=True)
S[S == 0] = 1e-17
P_upd = P_init, S, U
log_likelihood = 0 if calc_log_likelihood else None
grad_log_likelihood = 0 if calc_grad_log_likelihood else None
dm_upd = dm_init
dP_upd = dP_init
for k in range(0, steps_no):
prev_mean = M[k, :, :]
if p_kalman_filter_type == 'svd':
m_pred, P_pred, dm_pred, dP_pred = cls._kalman_prediction_step_SVD(
k, prev_mean, P_upd, p_dynamic_callables,
calc_grad_log_likelihood=calc_grad_log_likelihood, p_dm=
dm_upd, p_dP=dP_upd)
else:
m_pred, P_pred, dm_pred, dP_pred = cls._kalman_prediction_step(k,
prev_mean, P[k, :, :], p_dynamic_callables,
calc_grad_log_likelihood=calc_grad_log_likelihood, p_dm=
dm_upd, p_dP=dP_upd)
k_measurment = Y[k, :, :]
if np.any(np.isnan(k_measurment)) == False:
if p_kalman_filter_type == 'svd':
(m_upd, P_upd, log_likelihood_update, dm_upd, dP_upd,
d_log_likelihood_update) = (cls._kalman_update_step_SVD
(k, m_pred, P_pred, p_measurement_callables,
k_measurment, calc_log_likelihood=calc_log_likelihood,
calc_grad_log_likelihood=calc_grad_log_likelihood, p_dm
=dm_pred, p_dP=dP_pred))
else:
(m_upd, P_upd, log_likelihood_update, dm_upd, dP_upd,
d_log_likelihood_update) = (cls._kalman_update_step(k,
m_pred, P_pred, p_measurement_callables, k_measurment,
calc_log_likelihood=calc_log_likelihood,
calc_grad_log_likelihood=calc_grad_log_likelihood, p_dm
=dm_pred, p_dP=dP_pred))
elif not np.all(np.isnan(k_measurment)):
raise ValueError(
"""Nan measurements are currently not supported if
they are intermixed with not NaN measurements"""
)
else:
m_upd = m_pred
P_upd = P_pred
dm_upd = dm_pred
dP_upd = dP_pred
if calc_log_likelihood:
log_likelihood_update = np.zeros((time_series_no,))
if calc_grad_log_likelihood:
d_log_likelihood_update = np.zeros((grad_params_no,
time_series_no))
if calc_log_likelihood:
log_likelihood += log_likelihood_update
if calc_grad_log_likelihood:
grad_log_likelihood += d_log_likelihood_update
M[k + 1, :, :] = m_upd
if p_kalman_filter_type == 'svd':
P[k + 1, :, :] = P_upd[0]
else:
P[k + 1, :, :] = P_upd
return (M, P, log_likelihood, grad_log_likelihood, p_dynamic_callables.
reset(False))
| kalman_algorithm_raw | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, state_dim, p_dynamic_callables, p_measurement_callables, Y,\n m_init, P_init, p_kalman_filter_type='regular', calc_log_likelihood=\n False, calc_gra... |
def x(k, p_m, p_P, p_dyn_model_callable, calc_grad_log_likelihood=False,
p_dm=None, p_dP=None):
A = p_dyn_model_callable.Ak(k, p_m, p_P)
Q = p_dyn_model_callable.Qk(k)
m_pred = p_dyn_model_callable.f_a(k, p_m, A)
P_pred = A.dot(p_P).dot(A.T) + Q
if calc_grad_log_likelihood:
dA_all_params = p_dyn_model_callable.dAk(k)
dQ_all_params = p_dyn_model_callable.dQk(k)
param_number = p_dP.shape[2]
dm_pred = np.empty(p_dm.shape)
dP_pred = np.empty(p_dP.shape)
for j in range(param_number):
dA = dA_all_params[:, :, j]
dQ = dQ_all_params[:, :, j]
dP = p_dP[:, :, j]
dm = p_dm[:, :, j]
dm_pred[:, :, j] = np.dot(dA, p_m) + np.dot(A, dm)
dP_pred[:, :, j] = np.dot(dA, np.dot(p_P, A.T))
dP_pred[:, :, j] += dP_pred[:, :, j].T
dP_pred[:, :, j] += np.dot(A, np.dot(dP, A.T)) + dQ
dP_pred[:, :, j] = 0.5 * (dP_pred[:, :, j] + dP_pred[:, :, j].T)
else:
dm_pred = None
dP_pred = None
return m_pred, P_pred, dm_pred, dP_pred
| kalman_prediction_step | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(k, p_m, p_P, p_dyn_model_callable, calc_grad_log_likelihood=False,\n p_dm=None, p_dP=None):\n A = p_dyn_model_callable.Ak(k, p_m, p_P)\n Q = p_dyn_mod... |
def x(k, p_m, p_P, p_dyn_model_callable, calc_grad_log_likelihood=False,
p_dm=None, p_dP=None):
Prev_cov, S_old, V_old = p_P
A = p_dyn_model_callable.Ak(k, p_m, Prev_cov)
Q = p_dyn_model_callable.Qk(k)
Q_sr = p_dyn_model_callable.Q_srk(k)
m_pred = p_dyn_model_callable.f_a(k, p_m, A)
svd_1_matr = np.vstack(((np.sqrt(S_old) * np.dot(A, V_old)).T, Q_sr.T))
U, S, Vh = sp.linalg.svd(svd_1_matr, full_matrices=False, compute_uv=
True, overwrite_a=False, check_finite=True)
V_new = Vh.T
S_new = S ** 2
P_pred = np.dot(V_new * S_new, V_new.T)
P_pred = P_pred, S_new, Vh.T
if calc_grad_log_likelihood:
dA_all_params = p_dyn_model_callable.dAk(k)
dQ_all_params = p_dyn_model_callable.dQk(k)
param_number = p_dP.shape[2]
dm_pred = np.empty(p_dm.shape)
dP_pred = np.empty(p_dP.shape)
for j in range(param_number):
dA = dA_all_params[:, :, j]
dQ = dQ_all_params[:, :, j]
dm_pred[:, :, j] = np.dot(dA, p_m) + np.dot(A, p_dm[:, :, j])
dP_pred[:, :, j] = np.dot(dA, np.dot(Prev_cov, A.T))
dP_pred[:, :, j] += dP_pred[:, :, j].T
dP_pred[:, :, j] += np.dot(A, np.dot(p_dP[:, :, j], A.T)) + dQ
dP_pred[:, :, j] = 0.5 * (dP_pred[:, :, j] + dP_pred[:, :, j].T)
else:
dm_pred = None
dP_pred = None
return m_pred, P_pred, dm_pred, dP_pred
| kalman_prediction_step_SVD | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(k, p_m, p_P, p_dyn_model_callable, calc_grad_log_likelihood=False,\n p_dm=None, p_dP=None):\n Prev_cov, S_old, V_old = p_P\n A = p_dyn_model_callable.... |
def x(k, p_m, p_P, p_meas_model_callable, measurement, calc_log_likelihood=
False, calc_grad_log_likelihood=False, p_dm=None, p_dP=None):
m_pred = p_m
P_pred = p_P
H = p_meas_model_callable.Hk(k, m_pred, P_pred)
R = p_meas_model_callable.Rk(k)
time_series_no = p_m.shape[1]
log_likelihood_update = None
dm_upd = None
dP_upd = None
d_log_likelihood_update = None
v = measurement - p_meas_model_callable.f_h(k, m_pred, H)
S = H.dot(P_pred).dot(H.T) + R
if measurement.shape[0] == 1:
if S < 0:
raise ValueError('Kalman Filter Update: S is negative step %i' % k)
K = P_pred.dot(H.T) / S
if calc_log_likelihood:
log_likelihood_update = -0.5 * (np.log(2 * np.pi) + np.log(S) +
v * v / S)
if np.any(np.isnan(log_likelihood_update)):
raise ValueError('Nan values in likelihood update!')
LL = None
islower = None
else:
LL, islower = linalg.cho_factor(S)
K = linalg.cho_solve((LL, islower), H.dot(P_pred.T)).T
if calc_log_likelihood:
log_likelihood_update = -0.5 * (v.shape[0] * np.log(2 * np.pi) +
2 * np.sum(np.log(np.diag(LL))) + np.sum(linalg.cho_solve((
LL, islower), v) * v, axis=0))
if calc_grad_log_likelihood:
dm_pred_all_params = p_dm
dP_pred_all_params = p_dP
param_number = p_dP.shape[2]
dH_all_params = p_meas_model_callable.dHk(k)
dR_all_params = p_meas_model_callable.dRk(k)
dm_upd = np.empty(dm_pred_all_params.shape)
dP_upd = np.empty(dP_pred_all_params.shape)
d_log_likelihood_update = np.empty((param_number, time_series_no))
for param in range(param_number):
dH = dH_all_params[:, :, param]
dR = dR_all_params[:, :, param]
dm_pred = dm_pred_all_params[:, :, param]
dP_pred = dP_pred_all_params[:, :, param]
dv = -np.dot(dH, m_pred) - np.dot(H, dm_pred)
dS = np.dot(dH, np.dot(P_pred, H.T))
dS += dS.T
dS += np.dot(H, np.dot(dP_pred, H.T)) + dR
if LL is not None:
tmp1 = linalg.cho_solve((LL, islower), H).T
tmp2 = linalg.cho_solve((LL, islower), dH).T
tmp3 = linalg.cho_solve((LL, islower), dS).T
else:
tmp1 = H.T / S
tmp2 = dH.T / S
tmp3 = dS.T / S
dK = np.dot(dP_pred, tmp1) + np.dot(P_pred, tmp2) - np.dot(P_pred,
np.dot(tmp1, tmp3))
dm_upd[:, :, param] = dm_pred + np.dot(dK, v) + np.dot(K, dv)
dP_upd[:, :, param] = -np.dot(dK, np.dot(S, K.T))
dP_upd[:, :, param] += dP_upd[:, :, param].T
dP_upd[:, :, param] += dP_pred - np.dot(K, np.dot(dS, K.T))
dP_upd[:, :, param] = 0.5 * (dP_upd[:, :, param] + dP_upd[:, :,
param].T)
if LL is not None:
tmp5 = linalg.cho_solve((LL, islower), v)
else:
tmp5 = v / S
d_log_likelihood_update[param, :] = -(0.5 * np.sum(np.diag(tmp3
)) + np.sum(tmp5 * dv, axis=0) - 0.5 * np.sum(tmp5 * np.dot
(dS, tmp5), axis=0))
m_upd = m_pred + K.dot(v)
P_upd = K.dot(S).dot(K.T)
P_upd = 0.5 * (P_upd + P_upd.T)
P_upd = P_pred - P_upd
return (m_upd, P_upd, log_likelihood_update, dm_upd, dP_upd,
d_log_likelihood_update)
| kalman_update_step | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(k, p_m, p_P, p_meas_model_callable, measurement, calc_log_likelihood=\n False, calc_grad_log_likelihood=False, p_dm=None, p_dP=None):\n m_pred = p_m\n ... |
def x(k, p_m, p_P, p_meas_model_callable, measurement, calc_log_likelihood=
False, calc_grad_log_likelihood=False, p_dm=None, p_dP=None):
m_pred = p_m
P_pred, S_pred, V_pred = p_P
H = p_meas_model_callable.Hk(k, m_pred, P_pred)
R = p_meas_model_callable.Rk(k)
R_isr = p_meas_model_callable.R_isrk(k)
time_series_no = p_m.shape[1]
log_likelihood_update = None
dm_upd = None
dP_upd = None
d_log_likelihood_update = None
v = measurement - p_meas_model_callable.f_h(k, m_pred, H)
svd_2_matr = np.vstack((np.dot(R_isr.T, np.dot(H, V_pred)), np.diag(1.0 /
np.sqrt(S_pred))))
U, S, Vh = sp.linalg.svd(svd_2_matr, full_matrices=False, compute_uv=
True, overwrite_a=False, check_finite=True)
U_upd = np.dot(V_pred, Vh.T)
S_upd = (1.0 / S) ** 2
P_upd = np.dot(U_upd * S_upd, U_upd.T)
P_upd = P_upd, S_upd, U_upd
S = H.dot(P_pred).dot(H.T) + R
if measurement.shape[0] == 1:
if S < 0:
raise ValueError('Kalman Filter Update: S is negative step %i' % k)
K = P_pred.dot(H.T) / S
if calc_log_likelihood:
log_likelihood_update = -0.5 * (np.log(2 * np.pi) + np.log(S) +
v * v / S)
if np.any(np.isnan(log_likelihood_update)):
raise ValueError('Nan values in likelihood update!')
LL = None
islower = None
else:
LL, islower = linalg.cho_factor(S)
K = linalg.cho_solve((LL, islower), H.dot(P_pred.T)).T
if calc_log_likelihood:
log_likelihood_update = -0.5 * (v.shape[0] * np.log(2 * np.pi) +
2 * np.sum(np.log(np.diag(LL))) + np.sum(linalg.cho_solve((
LL, islower), v) * v, axis=0))
if calc_grad_log_likelihood:
dm_pred_all_params = p_dm
dP_pred_all_params = p_dP
param_number = p_dP.shape[2]
dH_all_params = p_meas_model_callable.dHk(k)
dR_all_params = p_meas_model_callable.dRk(k)
dm_upd = np.empty(dm_pred_all_params.shape)
dP_upd = np.empty(dP_pred_all_params.shape)
d_log_likelihood_update = np.empty((param_number, time_series_no))
for param in range(param_number):
dH = dH_all_params[:, :, param]
dR = dR_all_params[:, :, param]
dm_pred = dm_pred_all_params[:, :, param]
dP_pred = dP_pred_all_params[:, :, param]
dv = -np.dot(dH, m_pred) - np.dot(H, dm_pred)
dS = np.dot(dH, np.dot(P_pred, H.T))
dS += dS.T
dS += np.dot(H, np.dot(dP_pred, H.T)) + dR
if LL is not None:
tmp1 = linalg.cho_solve((LL, islower), H).T
tmp2 = linalg.cho_solve((LL, islower), dH).T
tmp3 = linalg.cho_solve((LL, islower), dS).T
else:
tmp1 = H.T / S
tmp2 = dH.T / S
tmp3 = dS.T / S
dK = np.dot(dP_pred, tmp1) + np.dot(P_pred, tmp2) - np.dot(P_pred,
np.dot(tmp1, tmp3))
dm_upd[:, :, param] = dm_pred + np.dot(dK, v) + np.dot(K, dv)
dP_upd[:, :, param] = -np.dot(dK, np.dot(S, K.T))
dP_upd[:, :, param] += dP_upd[:, :, param].T
dP_upd[:, :, param] += dP_pred - np.dot(K, np.dot(dS, K.T))
dP_upd[:, :, param] = 0.5 * (dP_upd[:, :, param] + dP_upd[:, :,
param].T)
if LL is not None:
tmp5 = linalg.cho_solve((LL, islower), v)
else:
tmp5 = v / S
d_log_likelihood_update[param, :] = -(0.5 * np.sum(np.diag(tmp3
)) + np.sum(tmp5 * dv, axis=0) - 0.5 * np.sum(tmp5 * np.dot
(dS, tmp5), axis=0))
m_upd = m_pred + K.dot(v)
return (m_upd, P_upd, log_likelihood_update, dm_upd, dP_upd,
d_log_likelihood_update)
| kalman_update_step_SVD | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(k, p_m, p_P, p_meas_model_callable, measurement, calc_log_likelihood=\n False, calc_grad_log_likelihood=False, p_dm=None, p_dP=None):\n m_pred = p_m\n ... |
def x(k, p_m, p_P, p_m_pred, p_P_pred, p_m_prev_step, p_P_prev_step,
p_dynamic_callables):
A = p_dynamic_callables.Ak(k, p_m, p_P)
tmp = np.dot(A, p_P.T)
if A.shape[0] == 1:
G = tmp.T / p_P_pred
else:
try:
LL, islower = linalg.cho_factor(p_P_pred)
G = linalg.cho_solve((LL, islower), tmp).T
except:
res = sp.linalg.lstsq(p_P_pred, tmp)
G = res[0].T
m_upd = p_m + G.dot(p_m_prev_step - p_m_pred)
P_upd = p_P + G.dot(p_P_prev_step - p_P_pred).dot(G.T)
P_upd = 0.5 * (P_upd + P_upd.T)
return m_upd, P_upd, G
| rts_smoother_update_step | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(k, p_m, p_P, p_m_pred, p_P_pred, p_m_prev_step, p_P_prev_step,\n p_dynamic_callables):\n A = p_dynamic_callables.Ak(k, p_m, p_P)\n tmp = np.dot(A, p_P... |
def x(cls, state_dim, p_dynamic_callables, filter_means, filter_covars):
no_steps = filter_covars.shape[0] - 1
M = np.empty(filter_means.shape)
P = np.empty(filter_covars.shape)
M[-1, :] = filter_means[-1, :]
P[-1, :, :] = filter_covars[-1, :, :]
for k in range(no_steps - 1, -1, -1):
m_pred, P_pred, tmp1, tmp2 = cls._kalman_prediction_step(k,
filter_means[k, :], filter_covars[k, :, :], p_dynamic_callables,
calc_grad_log_likelihood=False)
p_m = filter_means[k, :]
if len(p_m.shape) < 2:
p_m.shape = p_m.shape[0], 1
p_m_prev_step = M[k + 1, :]
if len(p_m_prev_step.shape) < 2:
p_m_prev_step.shape = p_m_prev_step.shape[0], 1
m_upd, P_upd, G_tmp = cls._rts_smoother_update_step(k, p_m,
filter_covars[k, :, :], m_pred, P_pred, p_m_prev_step, P[k + 1,
:, :], p_dynamic_callables)
M[k, :] = m_upd
P[k, :, :] = P_upd
return M, P
| rts_smoother | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, state_dim, p_dynamic_callables, filter_means, filter_covars):\n no_steps = filter_covars.shape[0] - 1\n M = np.empty(filter_means.shape)\n P = np... |
def x(A, Q, H, R, m_init, P_init, measurements, M, P, G, dA, dQ, dH, dR,
dm_init, dP_init):
import pdb
pdb.set_trace()
param_number = dA.shape[-1]
d_log_likelihood_update = np.empty((param_number, 1))
sample_no = measurements.shape[0]
P_1 = P[1:, :, :]
P_2 = P[0:-1, :, :]
M_1 = M[1:, :]
M_2 = M[0:-1, :]
Sigma = np.mean(P_1, axis=0) + np.dot(M_1.T, M_1) / sample_no
Phi = np.mean(P_2, axis=0) + np.dot(M_2.T, M_2) / sample_no
B = np.dot(measurements.T, M_1) / sample_no
C = (sp.einsum('ijk,ikl', P_1, G) + np.dot(M_1.T, M_2)) / sample_no
D = np.dot(measurements.T, measurements) / sample_no
try:
P_init_inv = sp.linalg.inv(P_init)
if np.max(np.abs(P_init_inv)) > 100000000000000.0:
compute_P_init_terms = False
else:
compute_P_init_terms = True
except np.linalg.LinAlgError:
compute_P_init_terms = False
try:
Q_inv = sp.linalg.inv(Q)
if np.max(np.abs(Q_inv)) > 100000000000000.0:
compute_Q_terms = False
else:
compute_Q_terms = True
except np.linalg.LinAlgError:
compute_Q_terms = False
try:
R_inv = sp.linalg.inv(R)
if np.max(np.abs(R_inv)) > 100000000000000.0:
compute_R_terms = False
else:
compute_R_terms = True
except np.linalg.LinAlgError:
compute_R_terms = False
d_log_likelihood_update = np.zeros((param_number, 1))
for j in range(param_number):
if compute_P_init_terms:
d_log_likelihood_update[j, :] -= 0.5 * np.sum(P_init_inv *
dP_init[:, :, j].T)
M0_smoothed = M[0]
M0_smoothed.shape = M0_smoothed.shape[0], 1
tmp1 = np.dot(dP_init[:, :, j], np.dot(P_init_inv, P[0, :, :] +
sp.outer(M0_smoothed - m_init, M0_smoothed - m_init)))
d_log_likelihood_update[j, :] += 0.5 * np.sum(P_init_inv * tmp1.T)
tmp2 = sp.outer(dm_init[:, j], M0_smoothed)
tmp2 += tmp2.T
d_log_likelihood_update[j, :] += 0.5 * np.sum(P_init_inv * tmp2.T)
if compute_Q_terms:
d_log_likelihood_update[j, :] -= sample_no / 2.0 * np.sum(Q_inv *
dQ[:, :, j].T)
tmp1 = np.dot(C, A.T)
tmp1 += tmp1.T
tmp1 = Sigma - tmp1 + np.dot(A, np.dot(Phi, A.T))
tmp1 = np.dot(dQ[:, :, j], np.dot(Q_inv, tmp1))
d_log_likelihood_update[j, :] += sample_no / 2.0 * np.sum(Q_inv *
tmp1.T)
tmp2 = np.dot(dA[:, :, j], C.T)
tmp2 += tmp2.T
tmp3 = np.dot(dA[:, :, j], np.dot(Phi, A.T))
tmp3 += tmp3.T
d_log_likelihood_update[j, :] -= sample_no / 2.0 * np.sum(Q_inv
.T * (tmp3 - tmp2))
if compute_R_terms:
d_log_likelihood_update[j, :] -= sample_no / 2.0 * np.sum(R_inv *
dR[:, :, j].T)
tmp1 = np.dot(B, H.T)
tmp1 += tmp1.T
tmp1 = D - tmp1 + np.dot(H, np.dot(Sigma, H.T))
tmp1 = np.dot(dR[:, :, j], np.dot(R_inv, tmp1))
d_log_likelihood_update[j, :] += sample_no / 2.0 * np.sum(R_inv *
tmp1.T)
tmp2 = np.dot(dH[:, :, j], B.T)
tmp2 += tmp2.T
tmp3 = np.dot(dH[:, :, j], np.dot(Sigma, H.T))
tmp3 += tmp3.T
d_log_likelihood_update[j, :] -= sample_no / 2.0 * np.sum(R_inv
.T * (tmp3 - tmp2))
return d_log_likelihood_update
| EM_gradient | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, Q, H, R, m_init, P_init, measurements, M, P, G, dA, dQ, dH, dR,\n dm_init, dP_init):\n import pdb\n pdb.set_trace()\n param_number = dA.shape[-1... |
def x(p_M, state_dim, measurement_dim, which='A'):
old_M_shape = None
if len(p_M.shape) < 3:
old_M_shape = p_M.shape
if len(p_M.shape) == 2:
p_M.shape = p_M.shape[0], p_M.shape[1], 1
elif len(p_M.shape) == 1:
if p_M.shape[0] != 1:
raise ValueError(
'Matrix %s is an 1D array, while it must be a matrix or scalar'
, which)
else:
p_M.shape = 1, 1, 1
if which == 'A' or which == 'Q':
if p_M.shape[0] != state_dim or p_M.shape[1] != state_dim:
raise ValueError('%s must be a square matrix of size (%i,%i)' %
(which, state_dim, state_dim))
if which == 'H':
if p_M.shape[0] != measurement_dim or p_M.shape[1] != state_dim:
raise ValueError(
'H must be of shape (measurement_dim, state_dim) (%i,%i)' %
(measurement_dim, state_dim))
if which == 'R':
if p_M.shape[0] != measurement_dim or p_M.shape[1] != measurement_dim:
raise ValueError(
'R must be of shape (measurement_dim, measurement_dim) (%i,%i)'
% (measurement_dim, measurement_dim))
return p_M, old_M_shape
| check_SS_matrix | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(p_M, state_dim, measurement_dim, which='A'):\n old_M_shape = None\n if len(p_M.shape) < 3:\n old_M_shape = p_M.shape\n if len(p_M.shape) ==... |
def x(dM, state_dim, grad_params_no, which='dA'):
if dM is None:
dM = np.zeros((state_dim, state_dim, grad_params_no))
elif isinstance(dM, np.ndarray):
if state_dim == 1:
if len(dM.shape) < 3:
dM.shape = 1, 1, 1
elif len(dM.shape) < 3:
dM.shape = state_dim, state_dim, 1
elif isinstance(dM, np.int):
if state_dim > 1:
raise ValueError(
'When computing likelihood gradient wrong %s dimension.' %
which)
else:
dM = np.ones((1, 1, 1)) * dM
return dM
| check_grad_state_matrices | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dM, state_dim, grad_params_no, which='dA'):\n if dM is None:\n dM = np.zeros((state_dim, state_dim, grad_params_no))\n elif isinstance(dM, np.ndar... |
def x(dM, state_dim, grad_params_no, measurement_dim, which='dH'):
if dM is None:
if which == 'dH':
dM = np.zeros((measurement_dim, state_dim, grad_params_no))
elif which == 'dR':
dM = np.zeros((measurement_dim, measurement_dim, grad_params_no))
elif isinstance(dM, np.ndarray):
if state_dim == 1:
if len(dM.shape) < 3:
dM.shape = 1, 1, 1
elif len(dM.shape) < 3:
if which == 'dH':
dM.shape = measurement_dim, state_dim, 1
elif which == 'dR':
dM.shape = measurement_dim, measurement_dim, 1
elif isinstance(dM, np.int):
if state_dim > 1:
raise ValueError(
'When computing likelihood gradient wrong dH dimension.')
else:
dM = np.ones((1, 1, 1)) * dM
return dM
| check_grad_measurement_matrices | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dM, state_dim, grad_params_no, measurement_dim, which='dH'):\n if dM is None:\n if which == 'dH':\n dM = np.zeros((measurement_dim, state_... |
def x(cls, F, L, Qc, p_H, p_R, P_inf, X, Y, index=None, m_init=None, P_init
=None, p_kalman_filter_type='regular', calc_log_likelihood=False,
calc_grad_log_likelihood=False, grad_params_no=0, grad_calc_params=None):
p_H = np.atleast_1d(p_H)
p_R = np.atleast_1d(p_R)
X.shape, old_X_shape = cls._reshape_input_data(X.shape, 2)
if X.shape[1] != 1:
raise ValueError('Only one dimensional X data is supported.')
Y.shape, old_Y_shape = cls._reshape_input_data(Y.shape)
state_dim = F.shape[0]
measurement_dim = Y.shape[1]
time_series_no = Y.shape[2]
if len(p_H.shape) == 3 and len(p_H.shape[2]) != 1 or len(p_R.shape
) == 3 and len(p_R.shape[2]) != 1:
model_matrices_chage_with_time = True
else:
model_matrices_chage_with_time = False
old_index_shape = None
if index is None:
if len(p_H.shape) == 3 or len(p_R.shape) == 3:
raise ValueError(
'Parameter index can not be None for time varying matrices (third dimension is present)'
)
else:
index = np.zeros((1, Y.shape[0]))
else:
if len(index.shape) == 1:
index.shape = 1, index.shape[0]
old_index_shape = index.shape[0],
if index.shape[1] != Y.shape[0]:
raise ValueError(
'Number of measurements must be equal the number of H_{k}, R_{k}'
)
if index.shape[0] == 1:
H_time_var_index = 0
R_time_var_index = 0
elif index.shape[0] == 4:
H_time_var_index = 0
R_time_var_index = 1
else:
raise ValueError('First Dimension of index must be either 1 or 2.')
p_H, old_H_shape = cls._check_SS_matrix(p_H, state_dim, measurement_dim,
which='H')
p_R, old_R_shape = cls._check_SS_matrix(p_R, state_dim, measurement_dim,
which='R')
if m_init is None:
m_init = np.zeros((state_dim, time_series_no))
else:
m_init = np.atleast_2d(m_init).T
if P_init is None:
P_init = P_inf.copy()
if p_kalman_filter_type not in ('regular', 'svd'):
raise ValueError("Kalman filer type neither 'regular nor 'svd'.")
if calc_grad_log_likelihood:
dF = cls._check_grad_state_matrices(grad_calc_params.get('dF'),
state_dim, grad_params_no, which='dA')
dQc = cls._check_grad_state_matrices(grad_calc_params.get('dQc'),
state_dim, grad_params_no, which='dQ')
dP_inf = cls._check_grad_state_matrices(grad_calc_params.get(
'dP_inf'), state_dim, grad_params_no, which='dA')
dH = cls._check_grad_measurement_matrices(grad_calc_params.get('dH'
), state_dim, grad_params_no, measurement_dim, which='dH')
dR = cls._check_grad_measurement_matrices(grad_calc_params.get('dR'
), state_dim, grad_params_no, measurement_dim, which='dR')
dm_init = grad_calc_params.get('dm_init')
if dm_init is None:
dm_init = np.zeros((state_dim, time_series_no, grad_params_no))
dP_init = grad_calc_params.get('dP_init')
if dP_init is None:
dP_init = dP_inf(0).copy()
else:
dP_inf = None
dF = None
dQc = None
dH = None
dR = None
dm_init = None
dP_init = None
measurement_callables = Std_Measurement_Callables_Class(p_H,
H_time_var_index, p_R, index, R_time_var_index, 20, dH, dR)
dynamic_callables = cls._cont_to_discrete_object(X, F, L, Qc,
compute_derivatives=calc_grad_log_likelihood, grad_params_no=
grad_params_no, P_inf=P_inf, dP_inf=dP_inf, dF=dF, dQc=dQc)
if print_verbose:
print('General: run Continuos-Discrete Kalman Filter')
M, P, log_likelihood, grad_log_likelihood, AQcomp = (cls.
_cont_discr_kalman_filter_raw(state_dim, dynamic_callables,
measurement_callables, X, Y, m_init=m_init, P_init=P_init,
p_kalman_filter_type=p_kalman_filter_type, calc_log_likelihood=
calc_log_likelihood, calc_grad_log_likelihood=
calc_grad_log_likelihood, grad_params_no=grad_params_no, dm_init=
dm_init, dP_init=dP_init))
if old_index_shape is not None:
index.shape = old_index_shape
if old_X_shape is not None:
X.shape = old_X_shape
if old_Y_shape is not None:
Y.shape = old_Y_shape
if old_H_shape is not None:
p_H.shape = old_H_shape
if old_R_shape is not None:
p_R.shape = old_R_shape
return M, P, log_likelihood, grad_log_likelihood, AQcomp
| cont_discr_kalman_filter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, F, L, Qc, p_H, p_R, P_inf, X, Y, index=None, m_init=None, P_init\n =None, p_kalman_filter_type='regular', calc_log_likelihood=False,\n calc_grad_log... |
def x(cls, X, F, L, Qc, compute_derivatives=False, grad_params_no=None,
P_inf=None, dP_inf=None, dF=None, dQc=None, dt0=None):
unique_round_decimals = 10
threshold_number_of_unique_time_steps = 20
dt = np.empty((X.shape[0],))
dt[1:] = np.diff(X[:, 0], axis=0)
if dt0 is None:
dt[0] = 0
elif isinstance(dt0, str):
dt = dt[1:]
else:
dt[0] = dt0
unique_indices = np.unique(np.round(dt, decimals=unique_round_decimals))
number_unique_indices = len(unique_indices)
if use_cython:
class AQcompute_batch(state_space_cython.AQcompute_batch_Cython):
def x(self, F, L, Qc, dt, compute_derivatives=False,
grad_params_no=None, P_inf=None, dP_inf=None, dF=None, dQc=None
):
As, Qs, reconstruct_indices, dAs, dQs = (ContDescrStateSpace
.lti_sde_to_descrete(F, L, Qc, dt, compute_derivatives,
grad_params_no=grad_params_no, P_inf=P_inf, dP_inf=
dP_inf, dF=dF, dQc=dQc))
super(AQcompute_batch, self).__init__(As, Qs,
reconstruct_indices, dAs, dQs)
else:
AQcompute_batch = cls.AQcompute_batch_Python
if number_unique_indices > threshold_number_of_unique_time_steps:
AQcomp = cls.AQcompute_once(F, L, Qc, dt, compute_derivatives=
compute_derivatives, grad_params_no=grad_params_no, P_inf=P_inf,
dP_inf=dP_inf, dF=dF, dQc=dQc)
if print_verbose:
print('CDO: Continue-to-discrete INSTANTANEOUS object is created.'
)
print('CDO: Number of different time steps: %i' % (
number_unique_indices,))
else:
AQcomp = AQcompute_batch(F, L, Qc, dt, compute_derivatives=
compute_derivatives, grad_params_no=grad_params_no, P_inf=P_inf,
dP_inf=dP_inf, dF=dF, dQc=dQc)
if print_verbose:
print('CDO: Continue-to-discrete BATCH object is created.')
print('CDO: Number of different time steps: %i' % (
number_unique_indices,))
print('CDO: Total size if its data: %i' % (AQcomp.
total_size_of_data,))
return AQcomp
| cont_to_discrete_object | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, X, F, L, Qc, compute_derivatives=False, grad_params_no=None,\n P_inf=None, dP_inf=None, dF=None, dQc=None, dt0=None):\n unique_round_decimals = 10\n... |
def x(F, L, Qc, dt, compute_derivatives=False, grad_params_no=None, P_inf=
None, dP_inf=None, dF=None, dQc=None):
n = F.shape[0]
if not isinstance(dt, collections.Iterable):
A = matrix_exponent(F * dt)
Phi = np.zeros((2 * n, 2 * n))
Phi[:n, :n] = F
Phi[:n, n:] = L.dot(Qc).dot(L.T)
Phi[n:, n:] = -F.T
AB = matrix_exponent(Phi * dt)
AB = np.dot(AB, np.vstack((np.zeros((n, n)), np.eye(n))))
Q_noise_1 = linalg.solve(AB[n:, :].T, AB[:n, :].T)
Q_noise_2 = P_inf - A.dot(P_inf).dot(A.T)
if compute_derivatives:
dA = np.zeros([n, n, grad_params_no])
dQ = np.zeros([n, n, grad_params_no])
FF = np.zeros([2 * n, 2 * n])
AA = np.zeros([2 * n, 2 * n, grad_params_no])
for p in range(0, grad_params_no):
FF[:n, :n] = F
FF[n:, :n] = dF[:, :, p]
FF[n:, n:] = F
AA[:, :, p] = matrix_exponent(FF * dt)
if p == 0:
A = AA[:n, :n, p]
Q_noise_3 = P_inf - A.dot(P_inf).dot(A.T)
Q_noise = Q_noise_3
dA[:, :, p] = AA[n:, :n, p]
tmp = dA[:, :, p].dot(P_inf).dot(A.T)
dQ[:, :, p] = dP_inf[:, :, p] - tmp - A.dot(dP_inf[:, :, p]
).dot(A.T) - tmp.T
dQ[:, :, p] = 0.5 * (dQ[:, :, p] + dQ[:, :, p].T)
else:
dA = None
dQ = None
Q_noise = Q_noise_2
Q_noise = 0.5 * (Q_noise + Q_noise.T)
return A, Q_noise, None, dA, dQ
else:
dt_unique, tmp, reconstruct_index = np.unique(np.round(dt, 8),
return_index=True, return_inverse=True)
del tmp
A = np.empty((n, n, dt_unique.shape[0]))
Q_noise = np.empty((n, n, dt_unique.shape[0]))
if compute_derivatives:
dA = np.empty((n, n, grad_params_no, dt_unique.shape[0]))
dQ = np.empty((n, n, grad_params_no, dt_unique.shape[0]))
else:
dA = None
dQ = None
for j in range(0, dt_unique.shape[0]):
A[:, :, j], Q_noise[:, :, j
], tmp1, dA_t, dQ_t = ContDescrStateSpace.lti_sde_to_descrete(F
, L, Qc, dt_unique[j], compute_derivatives=
compute_derivatives, grad_params_no=grad_params_no, P_inf=
P_inf, dP_inf=dP_inf, dF=dF, dQc=dQc)
if compute_derivatives:
dA[:, :, :, j] = dA_t
dQ[:, :, :, j] = dQ_t
return A, Q_noise, reconstruct_index, dA, dQ
| lti_sde_to_descrete | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(F, L, Qc, dt, compute_derivatives=False, grad_params_no=None, P_inf=\n None, dP_inf=None, dF=None, dQc=None):\n n = F.shape[0]\n if not isinstance(dt,... |
def x(self, X):
preactivations_list = list()
net = X
preactivations_list.append(X)
for W, b, i in zip(self.W_list, self.b_list, np.arange(len(self.W_list))):
net = np.dot(net, W)
net = net + b
if i < len(self.W_list) - 1:
preactivations_list.append(net)
net = self.act(net)
return preactivations_list
| f_preactivations | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n preactivations_list = list()\n net = X\n preactivations_list.append(X)\n for W, b, i in zip(self.W_list, self.b_list, np.arange(len(sel... |
def x(self):
for W, b in zip(self.W_list, self.b_list):
W.fix()
b.fix()
| fix_parameters | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n for W, b in zip(self.W_list, self.b_list):\n W.fix()\n b.fix()\n",
"role": "user"
},
{
"content": "fix_parameters",
"... |
def x(self):
for W, b in zip(self.W_list, self.b_list):
W.unfix()
b.unfix()
| unfix_parameters | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n for W, b in zip(self.W_list, self.b_list):\n W.unfix()\n b.unfix()\n",
"role": "user"
},
{
"content": "unfix_parameters",... |
def x(func):
def x(*args, **kwargs):
np.seterr(invalid='raise')
result = func(*args, **kwargs)
np.seterr(invalid='warn')
return result
return func_wrapper
| numpy_invalid_op_as_exception | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(func):\n\n def x(*args, **kwargs):\n np.seterr(invalid='raise')\n result = func(*args, **kwargs)\n np.seterr(invalid='warn')\n r... |
def x(A, dA, B, dB, operation='prod'):
if dA is None:
dA_param_num = 0
dA = np.zeros((A.shape[0], A.shape[1], 1))
else:
dA_param_num = dA.shape[2]
if dB is None:
dB_param_num = 0
dB = np.zeros((B.shape[0], B.shape[1], 1))
else:
dB_param_num = dB.shape[2]
dC = np.zeros((A.shape[0] * B.shape[0], A.shape[1] * B.shape[1],
dA_param_num + dB_param_num))
for k in range(dA_param_num):
if operation == 'prod':
dC[:, :, k] = np.kron(dA[:, :, k], B)
else:
dC[:, :, k] = np.kron(dA[:, :, k], np.eye(B.shape[0]))
for k in range(dB_param_num):
if operation == 'prod':
dC[:, :, dA_param_num + k] = np.kron(A, dB[:, :, k])
else:
dC[:, :, dA_param_num + k] = np.kron(np.eye(A.shape[0]), dB[:,
:, k])
return dC
| dkron | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(A, dA, B, dB, operation='prod'):\n if dA is None:\n dA_param_num = 0\n dA = np.zeros((A.shape[0], A.shape[1], 1))\n else:\n dA_param... |
def x(self):
input_dict = super(Prod, self)._save_to_input_dict()
input_dict['class'] = str('GPy.kern.Prod')
return input_dict
| to_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n input_dict = super(Prod, self)._save_to_input_dict()\n input_dict['class'] = str('GPy.kern.Prod')\n return input_dict\n",
"role": "user"
... |
def x(self, gradients):
part_start_param_index = 0
for p in self.parts:
if not p.is_fixed:
part_param_num = len(p.param_array)
p.sde_update_gradient_full(gradients[part_start_param_index:
part_start_param_index + part_param_num])
part_start_param_index += part_param_num
| sde_update_gradient_full | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, gradients):\n part_start_param_index = 0\n for p in self.parts:\n if not p.is_fixed:\n part_param_num = len(p.param_array)\n ... |
def x(self):
input_dict = super(StdPeriodic, self)._save_to_input_dict()
input_dict['class'] = 'GPy.kern.StdPeriodic'
input_dict['variance'] = self.variance.values.tolist()
input_dict['period'] = self.period.values.tolist()
input_dict['lengthscale'] = self.lengthscale.values.tolist()
input_dict['ARD1'] = self.ARD1
input_dict['ARD2'] = self.ARD2
return input_dict
| to_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n input_dict = super(StdPeriodic, self)._save_to_input_dict()\n input_dict['class'] = 'GPy.kern.StdPeriodic'\n input_dict['variance'] = self.va... |
def x(self, X, X2=None):
if X2 is None:
X2 = X
base = np.pi * (X[:, None, :] - X2[None, :, :]) / self.period
exp_dist = np.exp(-0.5 * np.sum(np.square(np.sin(base) / self.
lengthscale), axis=-1))
return self.variance * exp_dist
| K | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, X2=None):\n if X2 is None:\n X2 = X\n base = np.pi * (X[:, None, :] - X2[None, :, :]) / self.period\n exp_dist = np.exp(-0.5 * np.sum(... |
def x(self, X):
ret = np.empty(X.shape[0])
ret[:] = self.variance
return ret
| Kdiag | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n ret = np.empty(X.shape[0])\n ret[:] = self.variance\n return ret\n",
"role": "user"
},
{
"content": "Kdiag",
"role": "assi... |
def x(self, dL_dK, X, X2=None):
if X2 is None:
X2 = X
base = np.pi * (X[:, None, :] - X2[None, :, :]) / self.period
sin_base = np.sin(base)
exp_dist = np.exp(-0.5 * np.sum(np.square(sin_base / self.lengthscale),
axis=-1))
dwl = self.variance * (1.0 / np.square(self.lengthscale)
) * sin_base * np.cos(base) * (base / self.period)
dl = self.variance * np.square(sin_base) / np.power(self.lengthscale, 3)
self.variance.gradient = np.sum(exp_dist * dL_dK)
if self.ARD1:
self.period.gradient = (dwl * exp_dist[:, :, None] * dL_dK[:, :, None]
).sum(0).sum(0)
else:
self.period.gradient = np.sum(dwl.sum(-1) * exp_dist * dL_dK)
if self.ARD2:
self.lengthscale.gradient = (dl * exp_dist[:, :, None] * dL_dK[:, :,
None]).sum(0).sum(0)
else:
self.lengthscale.gradient = np.sum(dl.sum(-1) * exp_dist * dL_dK)
| update_gradients_full | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dL_dK, X, X2=None):\n if X2 is None:\n X2 = X\n base = np.pi * (X[:, None, :] - X2[None, :, :]) / self.period\n sin_base = np.sin(base)\n... |
def x(self, dL_dKdiag, X):
self.variance.gradient = np.sum(dL_dKdiag)
self.period.gradient = 0
self.lengthscale.gradient = 0
| update_gradients_diag | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dL_dKdiag, X):\n self.variance.gradient = np.sum(dL_dKdiag)\n self.period.gradient = 0\n self.lengthscale.gradient = 0\n",
"role": "user"
... |
def x(shape_records, facecolor='w', edgecolor='k', linewidths=0.5, ax=None,
xlims=None, ylims=None):
if ax is None:
fig = pb.figure()
ax = fig.add_subplot(111)
for srec in shape_records:
points = np.vstack(srec.shape.points)
sparts = srec.shape.parts
par = list(sparts) + [points.shape[0]]
polygs = []
for pj in range(len(sparts)):
polygs.append(Polygon(points[par[pj]:par[pj + 1]]))
ax.add_collection(PatchCollection(polygs, facecolor=facecolor,
edgecolor=edgecolor, linewidths=linewidths))
_box = np.vstack([srec.shape.bbox for srec in shape_records])
minx, miny = np.min(_box[:, :2], 0)
maxx, maxy = np.max(_box[:, 2:], 0)
if xlims is not None:
minx, maxx = xlims
if ylims is not None:
miny, maxy = ylims
ax.set_xlim(minx, maxx)
ax.set_ylim(miny, maxy)
| plot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(shape_records, facecolor='w', edgecolor='k', linewidths=0.5, ax=None,\n xlims=None, ylims=None):\n if ax is None:\n fig = pb.figure()\n ax ... |
def x(sf, regex, field=2):
index = []
shape_records = []
for rec in enumerate(sf.shapeRecords()):
m = re.search(regex, rec[1].record[field])
if m is not None:
index.append(rec[0])
shape_records.append(rec[1])
return index, shape_records
| string_match | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sf, regex, field=2):\n index = []\n shape_records = []\n for rec in enumerate(sf.shapeRecords()):\n m = re.search(regex, rec[1].record[field])\... |
def x(sf, bbox, inside_only=True):
A, B, C, D = bbox
index = []
shape_records = []
for rec in enumerate(sf.shapeRecords()):
a, b, c, d = rec[1].shape.bbox
if inside_only:
if A <= a and B <= b and C >= c and D >= d:
index.append(rec[0])
shape_records.append(rec[1])
else:
cond1 = A <= a and B <= b and C >= a and D >= b
cond2 = A <= c and B <= d and C >= c and D >= d
cond3 = A <= a and D >= d and C >= a and B <= d
cond4 = A <= c and D >= b and C >= c and B <= b
cond5 = a <= C and b <= B and d >= D
cond6 = c <= A and b <= B and d >= D
cond7 = d <= B and a <= A and c >= C
cond8 = b <= D and a <= A and c >= C
if (cond1 or cond2 or cond3 or cond4 or cond5 or cond6 or cond7 or
cond8):
index.append(rec[0])
shape_records.append(rec[1])
return index, shape_records
| bbox_match | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sf, bbox, inside_only=True):\n A, B, C, D = bbox\n index = []\n shape_records = []\n for rec in enumerate(sf.shapeRecords()):\n a, b, c, d =... |
def x(sf, bbox, inside_only=True):
index, shape_records = bbox_match(sf, bbox, inside_only)
A, B, C, D = bbox
plot(shape_records, xlims=[bbox[0], bbox[2]], ylims=[bbox[1], bbox[3]])
| plot_bbox | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sf, bbox, inside_only=True):\n index, shape_records = bbox_match(sf, bbox, inside_only)\n A, B, C, D = bbox\n plot(shape_records, xlims=[bbox[0], bbox... |
def x(sf, regex, field, **kwargs):
index, shape_records = string_match(sf, regex, field)
plot(shape_records, **kwargs)
| plot_string_match | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sf, regex, field, **kwargs):\n index, shape_records = string_match(sf, regex, field)\n plot(shape_records, **kwargs)\n",
"role": "user"
},
{
... |
def x(sf, ax):
limits = sf.bbox
xlim = limits[0], limits[2]
ylim = limits[1], limits[3]
ax.set_xlim(xlim)
ax.set_ylim(ylim)
| apply_bbox | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sf, ax):\n limits = sf.bbox\n xlim = limits[0], limits[2]\n ylim = limits[1], limits[3]\n ax.set_xlim(xlim)\n ax.set_ylim(ylim)\n",
"role": ... |
def x(self):
self.Y_normalized[:] = self.transform_data()
super(WarpedGP, self).parameters_changed()
Kiy = self.posterior.woodbury_vector.flatten()
self.warping_function.update_grads(self.Y_untransformed, Kiy)
| parameters_changed | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.Y_normalized[:] = self.transform_data()\n super(WarpedGP, self).parameters_changed()\n Kiy = self.posterior.woodbury_vector.flatten()\n ... |
def x(self):
ll = GP.log_likelihood(self)
jacobian = self.warping_function.fgrad_y(self.Y_untransformed)
return ll + np.log(jacobian).sum()
| log_likelihood | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n ll = GP.log_likelihood(self)\n jacobian = self.warping_function.fgrad_y(self.Y_untransformed)\n return ll + np.log(jacobian).sum()\n",
"r... |
def x(self, mean, std, pred_init=None, deg_gauss_hermite=20):
gh_samples, gh_weights = np.polynomial.hermite.hermgauss(deg_gauss_hermite)
gh_samples = gh_samples[:, None]
gh_weights = gh_weights[None, :]
return gh_weights.dot(self._get_warped_term(mean, std, gh_samples)
) / np.sqrt(np.pi)
| get_warped_mean | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, mean, std, pred_init=None, deg_gauss_hermite=20):\n gh_samples, gh_weights = np.polynomial.hermite.hermgauss(deg_gauss_hermite)\n gh_samples = gh_s... |
def x(self, mean, std, pred_init=None, deg_gauss_hermite=20):
gh_samples, gh_weights = np.polynomial.hermite.hermgauss(deg_gauss_hermite)
gh_samples = gh_samples[:, None]
gh_weights = gh_weights[None, :]
arg1 = gh_weights.dot(self._get_warped_term(mean, std, gh_samples,
pred_init=pred_init) ** 2) / np.sqrt(np.pi)
arg2 = self._get_warped_mean(mean, std, pred_init=pred_init,
deg_gauss_hermite=deg_gauss_hermite)
return arg1 - arg2 ** 2
| get_warped_variance | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, mean, std, pred_init=None, deg_gauss_hermite=20):\n gh_samples, gh_weights = np.polynomial.hermite.hermgauss(deg_gauss_hermite)\n gh_samples = gh_s... |
def x(self, Xnew, kern=None, pred_init=None, Y_metadata=None, median=False,
deg_gauss_hermite=20, likelihood=None):
mean, var = super(WarpedGP, self).predict(Xnew, kern=kern, full_cov=
False, likelihood=likelihood)
if self.predict_in_warped_space:
std = np.sqrt(var)
if median:
wmean = self.warping_function.f_inv(mean, y=pred_init)
else:
wmean = self._get_warped_mean(mean, std, pred_init=pred_init,
deg_gauss_hermite=deg_gauss_hermite).T
wvar = self._get_warped_variance(mean, std, pred_init=pred_init,
deg_gauss_hermite=deg_gauss_hermite).T
else:
wmean = mean
wvar = var
return wmean, wvar
| predict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Xnew, kern=None, pred_init=None, Y_metadata=None, median=False,\n deg_gauss_hermite=20, likelihood=None):\n mean, var = super(WarpedGP, self).predi... |
def x(self, X, quantiles=(2.5, 97.5), Y_metadata=None, likelihood=None,
kern=None):
qs = super(WarpedGP, self).predict_quantiles(X, quantiles, Y_metadata=
Y_metadata, likelihood=likelihood, kern=kern)
if self.predict_in_warped_space:
return [self.warping_function.f_inv(q) for q in qs]
return qs
| predict_quantiles | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, quantiles=(2.5, 97.5), Y_metadata=None, likelihood=None,\n kern=None):\n qs = super(WarpedGP, self).predict_quantiles(X, quantiles, Y_metadata=\... |
def x(self, x_test, y_test, Y_metadata=None):
mu_star, var_star = self._raw_predict(x_test)
fy = self.warping_function.f(y_test)
ll_lpd = self.likelihood.log_predictive_density(fy, mu_star, var_star,
Y_metadata=Y_metadata)
return ll_lpd + np.log(self.warping_function.fgrad_y(y_test))
| log_predictive_density | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, x_test, y_test, Y_metadata=None):\n mu_star, var_star = self._raw_predict(x_test)\n fy = self.warping_function.f(y_test)\n ll_lpd = self.likelih... |
def x(self, gradients):
self.variance.gradient = gradients[0]
self.lengthscale.gradient = gradients[1]
| sde_update_gradient_full | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, gradients):\n self.variance.gradient = gradients[0]\n self.lengthscale.gradient = gradients[1]\n",
"role": "user"
},
{
"content": "sde_... |
def x(self):
variance = float(self.variance.values)
lengthscale = float(self.lengthscale.values)
lamda = np.sqrt(5.0) / lengthscale
kappa = 5.0 / 3.0 * variance / lengthscale ** 2
F = np.array(((0, 1, 0), (0, 0, 1), (-lamda ** 3, -3.0 * lamda ** 2, -3 *
lamda)))
L = np.array(((0,), (0,), (1,)))
Qc = np.array(((variance * 400.0 * np.sqrt(5.0) / 3.0 / lengthscale **
5,),))
H = np.array(((1, 0, 0),))
Pinf = np.array(((variance, 0, -kappa), (0, kappa, 0), (-kappa, 0, 25.0 *
variance / lengthscale ** 4)))
P0 = Pinf.copy()
dF = np.empty((3, 3, 2))
dQc = np.empty((1, 1, 2))
dPinf = np.empty((3, 3, 2))
dFvariance = np.zeros((3, 3))
dFlengthscale = np.array(((0, 0, 0), (0, 0, 0), (15.0 * np.sqrt(5.0) /
lengthscale ** 4, 30.0 / lengthscale ** 3, 3 * np.sqrt(5.0) /
lengthscale ** 2)))
dQcvariance = np.array(((400 * np.sqrt(5) / 3 / lengthscale ** 5,),))
dQclengthscale = np.array(((-variance * 2000 * np.sqrt(5) / 3 /
lengthscale ** 6,),))
dPinf_variance = Pinf / variance
kappa2 = -2.0 * kappa / lengthscale
dPinf_lengthscale = np.array(((0, 0, -kappa2), (0, kappa2, 0), (-kappa2,
0, -100 * variance / lengthscale ** 5)))
dF[:, :, 0] = dFvariance
dF[:, :, 1] = dFlengthscale
dQc[:, :, 0] = dQcvariance
dQc[:, :, 1] = dQclengthscale
dPinf[:, :, 0] = dPinf_variance
dPinf[:, :, 1] = dPinf_lengthscale
dP0 = dPinf.copy()
return F, L, Qc, H, Pinf, P0, dF, dQc, dPinf, dP0
| sde | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n variance = float(self.variance.values)\n lengthscale = float(self.lengthscale.values)\n lamda = np.sqrt(5.0) / lengthscale\n kappa = 5.0 /... |
def x(self, num_samples=1000, hmc_iters=20):
params = np.empty((num_samples, self.p.size))
for i in range(num_samples):
self.p[:] = np.random.multivariate_normal(np.zeros(self.p.size), self.M
)
H_old = self._computeH()
theta_old = self.model.optimizer_array.copy()
params[i] = self.model.unfixed_param_array
self._update(hmc_iters)
H_new = self._computeH()
if H_old > H_new:
k = 1.0
else:
k = np.exp(H_old - H_new)
if np.random.rand() < k:
params[i] = self.model.unfixed_param_array
else:
self.model.optimizer_array = theta_old
return params
| sample | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, num_samples=1000, hmc_iters=20):\n params = np.empty((num_samples, self.p.size))\n for i in range(num_samples):\n self.p[:] = np.random.mult... |
def x(self):
self.posterior, self._log_marginal_likelihood, self.grad_dict = (self.
inference_method.inference(self.kern, self.X, self.likelihood, self
.Y_normalized, self.Y_metadata))
self.likelihood.update_gradients(self.grad_dict['dL_dthetaL'])
self.kern.update_gradients_direct(self.grad_dict['dL_dVar'], self.
grad_dict['dL_dLen'])
| parameters_changed | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.posterior, self._log_marginal_likelihood, self.grad_dict = (self.\n inference_method.inference(self.kern, self.X, self.likelihood, self... |
def x(self, Xnew, full_cov=False, kern=None):
if kern is None:
kern = self.kern
Kmn = kern.K(Xnew, self.X)
alpha_kron = self.posterior.alpha
mu = np.dot(Kmn, alpha_kron)
mu = mu.reshape(-1, 1)
Knm = Kmn.T
noise = self.likelihood.variance
V_kron = self.posterior.V_kron
Qs = self.posterior.Qs
QTs = self.posterior.QTs
A = self.kron_mmprod(QTs, Knm)
V_kron = V_kron.reshape(-1, 1)
A = A / (V_kron + noise)
A = self.kron_mmprod(Qs, A)
Kmm = kern.K(Xnew)
var = np.diag(Kmm - np.dot(Kmn, A)).copy()
var = var.reshape(-1, 1)
return mu, var
| raw_predict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Xnew, full_cov=False, kern=None):\n if kern is None:\n kern = self.kern\n Kmn = kern.K(Xnew, self.X)\n alpha_kron = self.posterior.alpha\... |
def x(self):
variance = float(self.variances.values)
t0 = float(self.t0)
F = np.array(((0, 1.0), (0, 0)))
L = np.array(((0,), (1.0,)))
Qc = np.zeros((1, 1))
H = np.array(((1.0, 0),))
Pinf = np.zeros((2, 2))
P0 = np.array(((t0 ** 2, t0), (t0, 1))) * variance
dF = np.zeros((2, 2, 1))
dQc = np.zeros((1, 1, 1))
dPinf = np.zeros((2, 2, 1))
dP0 = np.zeros((2, 2, 1))
dP0[:, :, 0] = P0 / variance
return F, L, Qc, H, Pinf, P0, dF, dQc, dPinf, dP0
| sde | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n variance = float(self.variances.values)\n t0 = float(self.t0)\n F = np.array(((0, 1.0), (0, 0)))\n L = np.array(((0,), (1.0,)))\n Qc = ... |
def x(self, X, X2=None):
X, slices = X[:, :-1], index_to_slices(X[:, -1])
if X2 is None:
X2, slices2 = X, slices
K = np.zeros((X.shape[0], X.shape[0]))
else:
X2, slices2 = X2[:, :-1], index_to_slices(X2[:, -1])
K = np.zeros((X.shape[0], X2.shape[0]))
tdist = (X[:, 0][:, None] - X2[:, 0][None, :]) ** 2
xdist = (X[:, 1][:, None] - X2[:, 1][None, :]) ** 2
ttdist = X[:, 0][:, None] - X2[:, 0][None, :]
vyt = self.variance_Yt
vyx = self.variance_Yx
lyt = 1 / (2 * self.lengthscale_Yt)
lyx = 1 / (2 * self.lengthscale_Yx)
a = self.a
b = self.b
c = self.c
kyy = lambda tdist, xdist: np.exp(-lyt * tdist - lyx * xdist)
k1 = lambda tdist: 2 * lyt - 4 * lyt ** 2 * tdist
k2 = lambda xdist: 4 * lyx ** 2 * xdist - 2 * lyx
k3 = (lambda xdist: 3 * 4 * lyx ** 2 - 6 * 8 * xdist * lyx ** 3 + 16 *
xdist ** 2 * lyx ** 4)
k4 = lambda ttdist: 2 * lyt * ttdist
for i, s1 in enumerate(slices):
for j, s2 in enumerate(slices2):
for ss1 in s1:
for ss2 in s2:
if i == 0 and j == 0:
K[ss1, ss2] = vyt * vyx * kyy(tdist[ss1, ss2],
xdist[ss1, ss2])
elif i == 0 and j == 1:
K[ss1, ss2] = (-a * k2(xdist[ss1, ss2]) + b * k4(
ttdist[ss1, ss2]) + c) * vyt * vyx * kyy(tdist[
ss1, ss2], xdist[ss1, ss2])
elif i == 1 and j == 1:
K[ss1, ss2] = (b ** 2 * k1(tdist[ss1, ss2]) - 2 * a *
c * k2(xdist[ss1, ss2]) + a ** 2 * k3(xdist[ss1,
ss2]) + c ** 2) * vyt * vyx * kyy(tdist[ss1,
ss2], xdist[ss1, ss2])
else:
K[ss1, ss2] = (-a * k2(xdist[ss1, ss2]) - b * k4(
ttdist[ss1, ss2]) + c) * vyt * vyx * kyy(tdist[
ss1, ss2], xdist[ss1, ss2])
return K
| K | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, X2=None):\n X, slices = X[:, :-1], index_to_slices(X[:, -1])\n if X2 is None:\n X2, slices2 = X, slices\n K = np.zeros((X.shape[0]... |
def x(self, X):
vyt = self.variance_Yt
vyx = self.variance_Yx
lyt = 1.0 / (2 * self.lengthscale_Yt)
lyx = 1.0 / (2 * self.lengthscale_Yx)
a = self.a
b = self.b
c = self.c
k1 = 2 * lyt * vyt * vyx
k2 = -2 * lyx * vyt * vyx
k3 = 4 * 3 * lyx ** 2 * vyt * vyx
Kdiag = np.zeros(X.shape[0])
slices = index_to_slices(X[:, -1])
for i, ss1 in enumerate(slices):
for s1 in ss1:
if i == 0:
Kdiag[s1] += vyt * vyx
elif i == 1:
Kdiag[s1] += (b ** 2 * k1 - 2 * a * c * k2 + a ** 2 * k3 +
c ** 2 * vyt * vyx)
else:
raise ValueError('invalid input/output index')
return Kdiag
| Kdiag | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n vyt = self.variance_Yt\n vyx = self.variance_Yx\n lyt = 1.0 / (2 * self.lengthscale_Yt)\n lyx = 1.0 / (2 * self.lengthscale_Yx)\n a ... |
def x(self, X):
return X.mean.gradient, X.variance.gradient, X.binary_prob.gradient
| get_X_gradients | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n return X.mean.gradient, X.variance.gradient, X.binary_prob.gradient\n",
"role": "user"
},
{
"content": "get_X_gradients",
"role"... |
def x(self, nSamples, raw_samples=False):
assert isinstance(self.kern, kern.Linear)
from ..util.linalg import pdinv
N, D = self.Y.shape
Q = self.X.shape[1]
noise_var = self.likelihood.variance.values
Xs = np.random.randn(*((nSamples,) + self.X.shape)) * np.sqrt(self.X.
variance.values) + self.X.mean.values
b = np.random.rand(*((nSamples,) + self.X.shape))
Xs[b > self.X.gamma.values] = 0
invcov = (Xs[:, :, :, None] * Xs[:, :, None, :]).sum(1
) / noise_var + np.eye(Q)
cov = np.array([pdinv(invcov[s_idx])[0] for s_idx in range(invcov.shape
[0])])
Ws = np.empty((nSamples, Q, D))
tmp = np.transpose(Xs, (0, 2, 1)).reshape(nSamples * Q, N).dot(self.Y
).reshape(nSamples, Q, D)
mean = (cov[:, :, :, None] * tmp[:, None, :, :]).sum(2) / noise_var
zeros = np.zeros((Q,))
for s_idx in range(Xs.shape[0]):
Ws[s_idx] = np.random.multivariate_normal(mean=zeros, cov=cov[s_idx
], size=(D,)).T + mean[s_idx]
if raw_samples:
return Ws
else:
return Ws.mean(0), Ws.std(0)
| sample_W | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, nSamples, raw_samples=False):\n assert isinstance(self.kern, kern.Linear)\n from ..util.linalg import pdinv\n N, D = self.Y.shape\n Q = self.... |
def x(self):
if self.L is None:
return {'mu': self.mu.tolist(), 'Sigma': self.Sigma.tolist()}
else:
return {'mu': self.mu.tolist(), 'Sigma': self.Sigma.tolist(), 'L':
self.L.tolist()}
| to_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.L is None:\n return {'mu': self.mu.tolist(), 'Sigma': self.Sigma.tolist()}\n else:\n return {'mu': self.mu.tolist(), 'Sigm... |
def x(self):
input_dict = super(EP, self)._save_to_input_dict()
input_dict['class'
] = 'GPy.inference.latent_function_inference.expectation_propagation.EP'
if self.ga_approx_old is not None:
input_dict['ga_approx_old'] = self.ga_approx_old.to_dict()
if self._ep_approximation is not None:
input_dict['_ep_approximation'] = {}
input_dict['_ep_approximation']['post_params'
] = self._ep_approximation[0].to_dict()
input_dict['_ep_approximation']['ga_approx'] = self._ep_approximation[1
].to_dict()
input_dict['_ep_approximation']['cav_params'] = self._ep_approximation[
2].to_dict()
input_dict['_ep_approximation']['log_Z_tilde'
] = self._ep_approximation[3].tolist()
return input_dict
| to_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n input_dict = super(EP, self)._save_to_input_dict()\n input_dict['class'\n ] = 'GPy.inference.latent_function_inference.expectation_propag... |
def x(self):
input_dict = super(EPDTC, self)._save_to_input_dict()
input_dict['class'
] = 'GPy.inference.latent_function_inference.expectation_propagation.EPDTC'
if self.ga_approx_old is not None:
input_dict['ga_approx_old'] = self.ga_approx_old.to_dict()
if self._ep_approximation is not None:
input_dict['_ep_approximation'] = {}
input_dict['_ep_approximation']['post_params'
] = self._ep_approximation[0].to_dict()
input_dict['_ep_approximation']['ga_approx'] = self._ep_approximation[1
].to_dict()
input_dict['_ep_approximation']['log_Z_tilde'
] = self._ep_approximation[2]
return input_dict
| to_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n input_dict = super(EPDTC, self)._save_to_input_dict()\n input_dict['class'\n ] = 'GPy.inference.latent_function_inference.expectation_pro... |
def x(self, other):
return 0.5 * (np.sum(self.variance / other.variance) + ((other.mean -
self.mean) ** 2 / other.variance).sum() - self.num_data * self.
input_dim + np.sum(np.log(other.variance)) - np.sum(np.log(self.
variance)))
| KL | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, other):\n return 0.5 * (np.sum(self.variance / other.variance) + ((other.mean -\n self.mean) ** 2 / other.variance).sum() - self.num_data * sel... |
def x(self, *args, **kwargs):
import sys
assert 'matplotlib' in sys.modules, 'matplotlib package has not been imported.'
from ...plotting.matplot_dep import variational_plots
return variational_plots.plot_SpikeSlab(self, *args, **kwargs)
| plot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *args, **kwargs):\n import sys\n assert 'matplotlib' in sys.modules, 'matplotlib package has not been imported.'\n from ...plotting.matplot_dep ... |
def x(d2f_dg2, dg_dx, df_dg, d2g_dx2):
if np.all(dg_dx == 1.0) and np.all(d2g_dx2 == 0):
return d2f_dg2
dg_dx_2 = np.clip(dg_dx, -np.inf, _lim_val_square) ** 2
return d2f_dg2 * dg_dx_2 + df_dg * d2g_dx2
| chain_2 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(d2f_dg2, dg_dx, df_dg, d2g_dx2):\n if np.all(dg_dx == 1.0) and np.all(d2g_dx2 == 0):\n return d2f_dg2\n dg_dx_2 = np.clip(dg_dx, -np.inf, _lim_val... |
def x(d3f_dg3, dg_dx, d2f_dg2, d2g_dx2, df_dg, d3g_dx3):
if np.all(dg_dx == 1.0) and np.all(d2g_dx2 == 0) and np.all(d3g_dx3 == 0):
return d3f_dg3
dg_dx_3 = np.clip(dg_dx, -np.inf, _lim_val_cube) ** 3
return d3f_dg3 * dg_dx_3 + 3 * d2f_dg2 * dg_dx * d2g_dx2 + df_dg * d3g_dx3
| chain_3 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(d3f_dg3, dg_dx, d2f_dg2, d2g_dx2, df_dg, d3g_dx3):\n if np.all(dg_dx == 1.0) and np.all(d2g_dx2 == 0) and np.all(d3g_dx3 == 0):\n return d3f_dg3\n ... |
def x(D, n=100, min_max=(-100, 100)):
g = np.linspace(min_max[0], min_max[1], n)
G = np.ones((n, D))
return G * g[:, None]
| linear_grid | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(D, n=100, min_max=(-100, 100)):\n g = np.linspace(min_max[0], min_max[1], n)\n G = np.ones((n, D))\n return G * g[:, None]\n",
"role": "user"
},... |
def x(X, m=10):
XXT = np.dot(X, X.T)
D = -2.0 * XXT + np.diag(XXT)[:, np.newaxis] + np.diag(XXT)[np.newaxis, :]
s = np.random.permutation(X.shape[0])[0]
inducing = [s]
prob = D[s] / D[s].sum()
for z in range(m - 1):
s = np.random.multinomial(1, prob.flatten()).argmax()
inducing.append(s)
prob = D[s] / D[s].sum()
inducing = np.array(inducing)
return X[inducing]
| kmm_init | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(X, m=10):\n XXT = np.dot(X, X.T)\n D = -2.0 * XXT + np.diag(XXT)[:, np.newaxis] + np.diag(XXT)[np.newaxis, :]\n s = np.random.permutation(X.shape[0])[... |
def x(*param):
import warnings
warnings.warn(
'Please use param.values, as this function will be deprecated in the next release.'
, DeprecationWarning)
assert len(param) > 0, 'At least one parameter needed'
if len(param) == 1:
return param[0].view(np.ndarray)
return [x.view(np.ndarray) for x in param]
| param_to_array | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(*param):\n import warnings\n warnings.warn(\n 'Please use param.values, as this function will be deprecated in the next release.'\n , Depre... |
def x(self, X):
Kdiag = np.zeros(X.shape[0])
ly = 1 / self.lengthscale_Y
lu = np.sqrt(3) / self.lengthscale_U
Vu = self.variance_U
Vy = self.variance_Y
k1 = (2 * lu + ly) / (lu + ly) ** 2
k2 = (ly - 2 * lu + 2 * lu - ly) / (ly - lu) ** 2
k3 = 1 / (lu + ly) + lu / (lu + ly) ** 2
slices = index_to_slices(X[:, -1])
for i, ss1 in enumerate(slices):
for s1 in ss1:
if i == 0:
Kdiag[s1] += self.variance_U + self.ubias
elif i == 1:
Kdiag[s1] += Vu * Vy * (k1 + k2 + k3)
else:
raise ValueError('invalid input/output index')
return Kdiag
| Kdiag | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n Kdiag = np.zeros(X.shape[0])\n ly = 1 / self.lengthscale_Y\n lu = np.sqrt(3) / self.lengthscale_U\n Vu = self.variance_U\n Vy = self... |
def x(self, X, Y):
assert X.shape[1] == self.Z.shape[1]
self.X, self.Y = X, Y
| set_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, Y):\n assert X.shape[1] == self.Z.shape[1]\n self.X, self.Y = X, Y\n",
"role": "user"
},
{
"content": "set_data",
"role": "assis... |
def x(self):
i = next(self.slicer)
return self.X_all[i], self.Y_all[i]
| new_batch | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n i = next(self.slicer)\n return self.X_all[i], self.Y_all[i]\n",
"role": "user"
},
{
"content": "new_batch",
"role": "assistant"
... |
def x(input_dict):
import copy
input_dict = copy.deepcopy(input_dict)
normalizer_class = input_dict.pop('class')
import GPy
normalizer_class = eval(normalizer_class)
return normalizer_class._build_from_input_dict(normalizer_class, input_dict
)
| from_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(input_dict):\n import copy\n input_dict = copy.deepcopy(input_dict)\n normalizer_class = input_dict.pop('class')\n import GPy\n normalizer_class... |
def x(self):
input_dict = super(Standardize, self)._save_to_input_dict()
input_dict['class'] = 'GPy.util.normalizer.Standardize'
if self.mean is not None:
input_dict['mean'] = self.mean.tolist()
input_dict['std'] = self.std.tolist()
return input_dict
| to_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n input_dict = super(Standardize, self)._save_to_input_dict()\n input_dict['class'] = 'GPy.util.normalizer.Standardize'\n if self.mean is not N... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.