source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(num_samples=400, num_inducing=5, max_iters=100, optimize=True, plot= True, checkgrad=False): X = np.random.uniform(-3.0, 3.0, (num_samples, 1)) Y = np.sin(X) + np.random.randn(num_samples, 1) * 0.05 rbf = GPy.kern.RBF(1) m = GPy.models.SparseGPRegression(X, Y, kernel=rbf, num_inducing= num_inducing) if checkgrad: m.checkgrad() if optimize: m.optimize('tnc', max_iters=max_iters) if plot: m.plot() return m
sparse_GP_regression_1D
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(num_samples=400, num_inducing=5, max_iters=100, optimize=True, plot=\n True, checkgrad=False):\n X = np.random.uniform(-3.0, 3.0, (num_samples, 1))\n ...
def x(num_samples=400, num_inducing=50, max_iters=100, optimize=True, plot= True, nan=False): np.random.seed(1234) X = np.random.uniform(-3.0, 3.0, (num_samples, 2)) Y = np.sin(X[:, 0:1]) * np.sin(X[:, 1:2]) + np.random.randn(num_samples, 1 ) * 0.05 if nan: inan = np.random.binomial(1, 0.2, size=Y.shape) Y[inan] = np.nan rbf = GPy.kern.RBF(2) m = GPy.models.SparseGPRegression(X, Y, kernel=rbf, num_inducing= num_inducing) m['.*len'] = 2.0 m.checkgrad() if optimize: m.optimize('tnc', messages=1, max_iters=max_iters) if plot: m.plot() print(m) return m
sparse_GP_regression_2D
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(num_samples=400, num_inducing=50, max_iters=100, optimize=True, plot=\n True, nan=False):\n np.random.seed(1234)\n X = np.random.uniform(-3.0, 3.0, (n...
def x(max_iters=200, optimize=True, plot=True): fig, axes = pb.subplots(1, 2, figsize=(12, 5), sharex=True, sharey=True) S = np.ones((20, 1)) X = np.random.uniform(-3.0, 3.0, (20, 1)) Y = np.sin(X) + np.random.randn(20, 1) * 0.05 Z = np.random.uniform(-3.0, 3.0, (7, 1)) k = GPy.kern.RBF(1) m = GPy.models.SparseGPRegression(X, Y, kernel=k, Z=Z) if optimize: m.optimize('scg', messages=1, max_iters=max_iters) if plot: m.plot(ax=axes[0]) axes[0].set_title('no input uncertainty') print(m) m = GPy.models.SparseGPRegression(X, Y, kernel=GPy.kern.RBF(1), Z=Z, X_variance=S) if optimize: m.optimize('scg', messages=1, max_iters=max_iters) if plot: m.plot(ax=axes[1]) axes[1].set_title('with input uncertainty') fig.canvas.draw() print(m) return m
uncertain_inputs_sparse_regression
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(max_iters=200, optimize=True, plot=True):\n fig, axes = pb.subplots(1, 2, figsize=(12, 5), sharex=True, sharey=True)\n S = np.ones((20, 1))\n X = np.r...
def x(max_iters=100, optimize=True, plot=True): mf = GPy.core.Mapping(1, 1) mf.f = np.sin mf.update_gradients = lambda a, b: None X = np.linspace(0, 10, 50).reshape(-1, 1) Y = np.sin(X) + 0.5 * np.cos(3 * X) + 0.1 * np.random.randn(*X.shape) k = GPy.kern.RBF(1) lik = GPy.likelihoods.Gaussian() m = GPy.core.GP(X, Y, kernel=k, likelihood=lik, mean_function=mf) if optimize: m.optimize(max_iters=max_iters) if plot: m.plot(plot_limits=(-10, 15)) return m
simple_mean_function
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(max_iters=100, optimize=True, plot=True):\n mf = GPy.core.Mapping(1, 1)\n mf.f = np.sin\n mf.update_gradients = lambda a, b: None\n X = np.linspace...
def x(max_iters=100, optimize=True, plot=True): mf = GPy.core.Mapping(1, 1) mf.f = np.sin X = np.linspace(0, 10, 50).reshape(-1, 1) Y = np.sin(X) + 0.5 * np.cos(3 * X) + 0.1 * np.random.randn(*X.shape ) + 3 * X mf = GPy.mappings.Linear(1, 1) k = GPy.kern.RBF(1) lik = GPy.likelihoods.Gaussian() m = GPy.core.GP(X, Y, kernel=k, likelihood=lik, mean_function=mf) if optimize: m.optimize(max_iters=max_iters) if plot: m.plot() return m
parametric_mean_function
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(max_iters=100, optimize=True, plot=True):\n mf = GPy.core.Mapping(1, 1)\n mf.f = np.sin\n X = np.linspace(0, 10, 50).reshape(-1, 1)\n Y = np.sin(X)...
def x(max_iters=100): X = 2 * np.pi * np.random.random(151) - np.pi Y = np.sin(X) + np.random.normal(0, 0.2, 151) Y = np.array([(np.power(abs(y), float(1) / 3) * (1, -1)[y < 0]) for y in Y] ) X = X[:, None] Y = Y[:, None] warp_k = GPy.kern.RBF(1) warp_f = GPy.util.warping_functions.TanhFunction(n_terms=2) warp_m = GPy.models.WarpedGP(X, Y, kernel=warp_k, warping_function=warp_f) warp_m['.*\\.d'].constrain_fixed(1.0) m = GPy.models.GPRegression(X, Y) m.optimize_restarts(parallel=False, robust=True, num_restarts=5, max_iters=max_iters) warp_m.optimize_restarts(parallel=False, robust=True, num_restarts=5, max_iters=max_iters) print(warp_m) print(warp_m['.*warp.*']) warp_m.predict_in_warped_space = False warp_m.plot(title='Warped GP - Latent space') warp_m.predict_in_warped_space = True warp_m.plot(title='Warped GP - Warped space') m.plot(title='Standard GP') warp_m.plot_warping() pb.show()
warped_gp_cubic_sine
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(max_iters=100):\n X = 2 * np.pi * np.random.random(151) - np.pi\n Y = np.sin(X) + np.random.normal(0, 0.2, 151)\n Y = np.array([(np.power(abs(y), floa...
def x(self, X1new, X2new): k1xf = self.kern1.K(X1new, self.X1) k2xf = self.kern2.K(X2new, self.X2) A = k1xf.dot(self.U1) B = k2xf.dot(self.U2) mu = A.dot(self.Ytilde.reshape(self.num_data1, self.num_data2, order='F') ).dot(B.T).flatten(order='F') k1xx = self.kern1.Kdiag(X1new) k2xx = self.kern2.Kdiag(X2new) BA = np.kron(B, A) var = np.kron(k2xx, k1xx) - np.sum(BA ** 2 * self.Wi, 1 ) + self.likelihood.variance return mu[:, None], var[:, None]
predict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X1new, X2new):\n k1xf = self.kern1.K(X1new, self.X1)\n k2xf = self.kern2.K(X2new, self.X2)\n A = k1xf.dot(self.U1)\n B = k2xf.dot(self.U2)\n ...
def x(self, Xgrid, plot_raw, apply_link, percentiles, which_data_ycols, predict_kw, samples=0): if predict_kw is None: predict_kw = {} if 'likelihood' not in predict_kw: if plot_raw: from ...likelihoods import Gaussian from ...likelihoods.link_functions import Identity lik = Gaussian(Identity(), 1e-09) else: lik = None predict_kw['likelihood'] = lik if 'Y_metadata' not in predict_kw: predict_kw['Y_metadata'] = {} if 'output_index' not in predict_kw['Y_metadata']: predict_kw['Y_metadata']['output_index'] = Xgrid[:, -1:].astype(np.int) mu, _ = self.predict(Xgrid, **predict_kw) if percentiles is not None: percentiles = self.predict_quantiles(Xgrid, quantiles=percentiles, **predict_kw) else: percentiles = [] if samples > 0: fsamples = self.posterior_samples(Xgrid, size=samples, **predict_kw) fsamples = fsamples[:, which_data_ycols, :] else: fsamples = None retmu = mu[:, which_data_ycols] percs = [p[:, which_data_ycols] for p in percentiles] if plot_raw and apply_link: for i in range(len(which_data_ycols)): retmu[:, [i]] = self.likelihood.gp_link.transf(mu[:, [i]]) for perc in percs: perc[:, [i]] = self.likelihood.gp_link.transf(perc[:, [i]]) if fsamples is not None: for s in range(fsamples.shape[-1]): fsamples[:, i, s] = self.likelihood.gp_link.transf(fsamples [:, i, s]) return retmu, percs, fsamples
helper_predict_with_model
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Xgrid, plot_raw, apply_link, percentiles, which_data_ycols,\n predict_kw, samples=0):\n if predict_kw is None:\n predict_kw = {}\n if 'li...
def x(self, X, plot_limits, visible_dims, fixed_inputs, resolution): if fixed_inputs is None: fixed_inputs = [] fixed_dims = get_fixed_dims(fixed_inputs) free_dims = get_free_dims(self, visible_dims, fixed_dims) if len(free_dims) == 1: resolution = resolution or 200 Xnew, xmin, xmax = x_frame1D(X[:, free_dims], plot_limits= plot_limits, resolution=resolution) Xgrid = np.zeros((Xnew.shape[0], self.input_dim)) Xgrid[:, free_dims] = Xnew for i, v in fixed_inputs: Xgrid[:, i] = v x = Xgrid y = None elif len(free_dims) == 2: resolution = resolution or 35 Xnew, x, y, xmin, xmax = x_frame2D(X[:, free_dims], plot_limits, resolution) Xgrid = np.zeros((Xnew.shape[0], self.input_dim)) Xgrid[:, free_dims] = Xnew for i, v in fixed_inputs: Xgrid[:, i] = v else: raise TypeError( 'calculated free_dims {} from visible_dims {} and fixed_dims {} is neither 1D nor 2D' .format(free_dims, visible_dims, fixed_dims)) return fixed_dims, free_dims, Xgrid, x, y, xmin, xmax, resolution
helper_for_plot_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, plot_limits, visible_dims, fixed_inputs, resolution):\n if fixed_inputs is None:\n fixed_inputs = []\n fixed_dims = get_fixed_dims(fixed_...
def x(X, labels, num_samples=1000): if X.shape[0] > num_samples: print( 'Warning: subsampling X, as it has more samples then {}. X.shape={!s}' .format(int(num_samples), X.shape)) if labels is not None: subsample = [] for _, _, _, _, index, _ in scatter_label_generator(labels, X, (0, None, None)): subsample.append(np.random.choice(index, size=max(2, int( index.size * (float(num_samples) / X.shape[0]))), replace=False)) subsample = np.hstack(subsample) else: subsample = np.random.choice(X.shape[0], size=1000, replace=False) X = X[subsample] labels = labels[subsample] return X, labels
subsample_X
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(X, labels, num_samples=1000):\n if X.shape[0] > num_samples:\n print(\n 'Warning: subsampling X, as it has more samples then {}. X.shape={...
def x(to_update, update_from): if to_update is None: to_update = {} to_update.update({k: v for k, v in update_from.items() if k not in to_update}) return to_update
update_not_existing_kwargs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(to_update, update_from):\n if to_update is None:\n to_update = {}\n to_update.update({k: v for k, v in update_from.items() if k not in\n to...
def x(model): if hasattr(model, 'has_uncertain_inputs') and model.has_uncertain_inputs(): X = model.X.mean.values X_variance = model.X.variance.values else: try: X = model.X.values except AttributeError: X = model.X X_variance = None try: Y = model.Y.values except AttributeError: Y = model.Y if isinstance(model, WarpedGP) and not model.predict_in_warped_space: Y = model.Y_normalized if sparse.issparse(Y): Y = Y.todense().view(np.ndarray) return X, X_variance, Y
get_x_y_var
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n if hasattr(model, 'has_uncertain_inputs') and model.has_uncertain_inputs():\n X = model.X.mean.values\n X_variance = model.X.varianc...
def x(model, visible_dims, fixed_dims): if visible_dims is None: visible_dims = np.arange(model.input_dim) dims = np.asanyarray(visible_dims) if fixed_dims is not None: dims = [dim for dim in dims if dim not in fixed_dims] return np.asanyarray([dim for dim in dims if dim is not None])
get_free_dims
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model, visible_dims, fixed_dims):\n if visible_dims is None:\n visible_dims = np.arange(model.input_dim)\n dims = np.asanyarray(visible_dims)\n ...
def x(model, which_data_ycols): if which_data_ycols == 'all' or which_data_ycols is None: return np.arange(model.output_dim) return which_data_ycols
get_which_data_ycols
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model, which_data_ycols):\n if which_data_ycols == 'all' or which_data_ycols is None:\n return np.arange(model.output_dim)\n return which_data_yco...
def x(X, plot_limits=None, resolution=None): assert X.shape[1] == 1, 'x_frame1D is defined for one-dimensional inputs' if plot_limits is None: from GPy.core.parameterization.variational import VariationalPosterior if isinstance(X, VariationalPosterior): xmin, xmax = X.mean.min(0), X.mean.max(0) else: xmin, xmax = X.min(0), X.max(0) xmin, xmax = xmin - 0.25 * (xmax - xmin), xmax + 0.25 * (xmax - xmin) elif len(plot_limits) == 2: xmin, xmax = map(np.atleast_1d, plot_limits) else: raise ValueError('Bad limits for plotting') Xnew = np.linspace(float(xmin), float(xmax), int(resolution) or 200)[:, None] return Xnew, xmin, xmax
x_frame1D
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(X, plot_limits=None, resolution=None):\n assert X.shape[1] == 1, 'x_frame1D is defined for one-dimensional inputs'\n if plot_limits is None:\n fro...
def x(X, plot_limits=None, resolution=None): assert X.shape[1] == 2, 'x_frame2D is defined for two-dimensional inputs' if plot_limits is None: xmin, xmax = X.min(0), X.max(0) xmin, xmax = xmin - 0.075 * (xmax - xmin), xmax + 0.075 * (xmax - xmin) elif len(plot_limits) == 2: xmin, xmax = plot_limits try: xmin = xmin[0], xmin[1] except: xmin = [plot_limits[0], plot_limits[0]] xmax = [plot_limits[1], plot_limits[1]] elif len(plot_limits) == 4: xmin, xmax = (plot_limits[0], plot_limits[2]), (plot_limits[1], plot_limits[3]) else: raise ValueError('Bad limits for plotting') resolution = resolution or 50 xx, yy = np.mgrid[xmin[0]:xmax[0]:1.0j * resolution, xmin[1]:xmax[1]: 1.0j * resolution] Xnew = np.c_[xx.flat, yy.flat] return Xnew, xx, yy, xmin, xmax
x_frame2D
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(X, plot_limits=None, resolution=None):\n assert X.shape[1] == 2, 'x_frame2D is defined for two-dimensional inputs'\n if plot_limits is None:\n xmi...
def x(self, X): X = X.copy() inan = numpy.isnan(X) if self.mu is None: X_ = numpy.ma.masked_array(X, inan) self.mu = X_.mean(0).base self.sigma = X_.std(0).base reduce(lambda y, x: setitem(x[0], x[1], x[2]), zip(X.T, inan.T, self.mu ), None) X = X - self.mu X = X / numpy.where(self.sigma == 0, 1e-30, self.sigma) return X
center
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X):\n X = X.copy()\n inan = numpy.isnan(X)\n if self.mu is None:\n X_ = numpy.ma.masked_array(X, inan)\n self.mu = X_.mean(0).base...
def x(self, X, Q=None): if Q is None: Q = self.Q if Q > X.shape[1]: raise IndexError('requested dimension larger then input dimension') X = self.center(X) return X.dot(self.eigvectors[:, :Q])
project
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, Q=None):\n if Q is None:\n Q = self.Q\n if Q > X.shape[1]:\n raise IndexError('requested dimension larger then input dimension')\n...
def x(self, Q=None, ax=None, fignum=None): from ..plotting import Tango Tango.reset() col = Tango.nextMedium() if ax is None: fig = pylab.figure(fignum) ax = fig.add_subplot(111) if Q is None: Q = self.Q ticks = numpy.arange(Q) bar = ax.bar(ticks - 0.4, self.fracs[:Q], color=col) ax.set_xticks(ticks, map(lambda x: '${}$'.format(x), ticks + 1)) ax.set_ylabel('Eigenvalue fraction') ax.set_xlabel('PC') ax.set_ylim(0, ax.get_ylim()[1]) ax.set_xlim(ticks.min() - 0.5, ticks.max() + 0.5) try: pylab.tight_layout() except: pass return bar
plot_fracs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Q=None, ax=None, fignum=None):\n from ..plotting import Tango\n Tango.reset()\n col = Tango.nextMedium()\n if ax is None:\n fig = pyla...
def x(self, X, labels=None, s=20, marker='o', dimensions=(0, 1), ax=None, colors=None, fignum=None, cmap=None, **kwargs): if cmap is None: cmap = matplotlib.cm.jet if ax is None: fig = pylab.figure(fignum) ax = fig.add_subplot(111) if labels is None: labels = numpy.zeros(X.shape[0]) ulabels = [] for lab in labels: if not lab in ulabels: ulabels.append(lab) nlabels = len(ulabels) if colors is None: colors = iter([cmap(float(i) / nlabels) for i in range(nlabels)]) else: colors = iter(colors) X_ = self.project(X, self.Q)[:, dimensions] kwargs.update(dict(s=s)) plots = list() for i, l in enumerate(ulabels): kwargs.update(dict(color=next(colors), marker=marker[i % len(marker)])) plots.append(ax.scatter(*X_[labels == l, :].T, label=str(l), **kwargs)) ax.set_xlabel('PC$_1$') ax.set_ylabel('PC$_2$') try: pylab.tight_layout() except: pass return plots
plot_2d
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, labels=None, s=20, marker='o', dimensions=(0, 1), ax=None,\n colors=None, fignum=None, cmap=None, **kwargs):\n if cmap is None:\n cmap = ...
def x(self, labels=None, which_indices=None, legend=True, plot_limits=None, marker='<>^vsd', num_samples=1000, projection='2d', **kwargs): canvas, projection, kwargs, sig_dims = _new_canvas(self, projection, kwargs, which_indices) X, _, _ = get_x_y_var(self) if labels is None: labels = np.ones(self.num_data) legend = False else: legend = find_best_layout_for_subplots(len(np.unique(labels)))[1] scatters = _plot_latent_scatter(canvas, X, sig_dims, labels, marker, num_samples, projection=projection, **kwargs) return pl().add_to_canvas(canvas, dict(scatter=scatters), legend=legend)
plot_latent_scatter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, labels=None, which_indices=None, legend=True, plot_limits=None,\n marker='<>^vsd', num_samples=1000, projection='2d', **kwargs):\n canvas, projecti...
def x(self, which_indices=None, legend=False, plot_limits=None, marker=None, projection='2d', **kwargs): canvas, projection, kwargs, sig_dims = _new_canvas(self, projection, kwargs, which_indices) if legend: label = 'inducing' else: label = None if marker is not None: kwargs['marker'] = marker update_not_existing_kwargs(kwargs, pl().defaults.inducing_2d) from .data_plots import _plot_inducing scatters = _plot_inducing(self, canvas, sig_dims[:2], projection, label, **kwargs) return pl().add_to_canvas(canvas, dict(scatter=scatters), legend=legend)
plot_latent_inducing
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, which_indices=None, legend=False, plot_limits=None, marker=None,\n projection='2d', **kwargs):\n canvas, projection, kwargs, sig_dims = _new_canvas...
def x(self, labels=None, which_indices=None, resolution=60, marker='<>^vsd', legend=True, plot_limits=None, updates=False, mean=True, covariance= True, kern=None, num_samples=1000, scatter_kwargs=None, plot_scatter= True, **imshow_kwargs): input_1, input_2 = which_indices = (self. get_most_significant_input_dimensions(which_indices)[:2]) X = get_x_y_var(self)[0] _, _, Xgrid, _, _, xmin, xmax, resolution = helper_for_plot_data(self, X, plot_limits, which_indices, None, resolution) canvas, imshow_kwargs = pl().new_canvas(xlim=(xmin[0], xmax[0]), ylim=( xmin[1], xmax[1]), xlabel='latent dimension %i' % input_1, ylabel= 'latent dimension %i' % input_2, **imshow_kwargs) plots = {} if legend and plot_scatter: if labels is not None: legend = find_best_layout_for_subplots(len(np.unique(labels)))[1] else: labels = np.ones(self.num_data) legend = False if plot_scatter: plots['scatters'] = _plot_latent_scatter(canvas, X, which_indices, labels, marker, num_samples, projection='2d', **scatter_kwargs or {}) plots['view'] = _plot_magnification(self, canvas, which_indices, Xgrid, xmin, xmax, resolution, updates, mean, covariance, kern, ** imshow_kwargs) retval = pl().add_to_canvas(canvas, plots, legend=legend) _wait_for_updates(plots['view'], updates) return retval
plot_magnification
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, labels=None, which_indices=None, resolution=60, marker='<>^vsd',\n legend=True, plot_limits=None, updates=False, mean=True, covariance=\n True, ker...
def x(self, labels=None, which_indices=None, resolution=60, legend=True, plot_limits=None, updates=False, kern=None, marker='<>^vsd', num_samples=1000, projection='2d', scatter_kwargs=None, **imshow_kwargs): if projection != '2d': raise ValueError( 'Cannot plot latent in other then 2 dimensions, consider plot_scatter' ) input_1, input_2 = which_indices = (self. get_most_significant_input_dimensions(which_indices)[:2]) X = get_x_y_var(self)[0] _, _, Xgrid, _, _, xmin, xmax, resolution = helper_for_plot_data(self, X, plot_limits, which_indices, None, resolution) canvas, imshow_kwargs = pl().new_canvas(xlim=(xmin[0], xmax[0]), ylim=( xmin[1], xmax[1]), xlabel='latent dimension %i' % input_1, ylabel= 'latent dimension %i' % input_2, **imshow_kwargs) if legend: if labels is not None: legend = find_best_layout_for_subplots(len(np.unique(labels)))[1] else: labels = np.ones(self.num_data) legend = False scatters = _plot_latent_scatter(canvas, X, which_indices, labels, marker, num_samples, projection='2d', **scatter_kwargs or {}) view = _plot_latent(self, canvas, which_indices, Xgrid, xmin, xmax, resolution, updates, kern, **imshow_kwargs) retval = pl().add_to_canvas(canvas, dict(scatter=scatters, imshow=view), legend=legend) _wait_for_updates(view, updates) return retval
plot_latent
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, labels=None, which_indices=None, resolution=60, legend=True,\n plot_limits=None, updates=False, kern=None, marker='<>^vsd',\n num_samples=1000, pro...
def x(self, output_labels=None, data_labels=None, which_indices=None, resolution=15, legend=True, plot_limits=None, updates=False, kern=None, marker='<>^vsd', num_samples=1000, annotation_kwargs=None, scatter_kwargs=None, **imshow_kwargs): input_1, input_2 = which_indices = (self. get_most_significant_input_dimensions(which_indices)[:2]) X = get_x_y_var(self)[0] _, _, Xgrid, _, _, xmin, xmax, resolution = helper_for_plot_data(self, X, plot_limits, which_indices, None, resolution) canvas, imshow_kwargs = pl().new_canvas(xlim=(xmin[0], xmax[0]), ylim=( xmin[1], xmax[1]), xlabel='latent dimension %i' % input_1, ylabel= 'latent dimension %i' % input_2, **imshow_kwargs) if data_labels is not None: legend = find_best_layout_for_subplots(len(np.unique(data_labels)))[1] else: data_labels = np.ones(self.num_data) legend = False plots = dict(scatter=_plot_latent_scatter(canvas, X, which_indices, data_labels, marker, num_samples, **scatter_kwargs or {})) plots.update(_plot_steepest_gradient_map(self, canvas, which_indices, Xgrid, xmin, xmax, resolution, output_labels, updates, kern, annotation_kwargs=annotation_kwargs, **imshow_kwargs)) retval = pl().add_to_canvas(canvas, plots, legend=legend) _wait_for_updates(plots['annotation'], updates) return retval
plot_steepest_gradient_map
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, output_labels=None, data_labels=None, which_indices=None,\n resolution=15, legend=True, plot_limits=None, updates=False, kern=None,\n marker='<>^vs...
def x(self, expressions): Symbolic_core._set_expressions(self, expressions) Symbolic_core._set_variables(self, self.cacheable) for x, z in zip(self.variables['X'], self.variables['Z']): expressions['kdiag'] = expressions['kdiag'].subs(z, x) Symbolic_core._set_expressions(self, expressions)
set_expressions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, expressions):\n Symbolic_core._set_expressions(self, expressions)\n Symbolic_core._set_variables(self, self.cacheable)\n for x, z in zip(self.va...
def x(M): if M.shape[0] == 1: Mexp = np.array(((np.exp(M[0, 0]),),)) else: method = None try: Mexp = linalg.expm(M) method = 1 except (Exception,) as e: Mexp = linalg.expm3(M) method = 2 finally: if np.any(np.isnan(Mexp)): if method == 2: raise ValueError('Matrix Exponent is not computed 1') else: Mexp = linalg.expm3(M) method = 2 if np.any(np.isnan(Mexp)): raise ValueError('Matrix Exponent is not computed 2') return Mexp
matrix_exponent
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(M):\n if M.shape[0] == 1:\n Mexp = np.array(((np.exp(M[0, 0]),),))\n else:\n method = None\n try:\n Mexp = linalg.expm(M)...
def x(A): if len(A.shape) != 2 or A.shape[0] != A.shape[1]: raise ValueError('balance_matrix: Expecting square matrix') N = A.shape[0] gebal = sp.linalg.lapack.get_lapack_funcs('gebal', (A,)) bA, lo, hi, pivscale, info = gebal(A, permute=True, scale=True, overwrite_a=False) if info < 0: raise ValueError( 'balance_matrix: Illegal value in %d-th argument of internal gebal ' % -info) def x(D, c1, c2): i1 = c1 - 1 i2 = c2 - 1 P = np.eye(D) P[i1, i1] = 0.0 P[i2, i2] = 0.0 P[i1, i2] = 1.0 P[i2, i1] = 1.0 return P P = np.eye(N) if hi != N - 1: for k in range(N - 1, hi, -1): new_perm = perm_matr(N, k + 1, pivscale[k]) P = np.dot(P, new_perm) if lo != 0: for k in range(0, lo, 1): new_perm = perm_matr(N, k + 1, pivscale[k]) P = np.dot(P, new_perm) D = pivscale.copy() D[0:lo] = 1.0 D[hi + 1:N] = 1.0 T = np.dot(P, np.diag(D)) T_inv = np.dot(np.diag(D ** -1), P.T) return bA.copy(), T, T_inv
balance_matrix
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A):\n if len(A.shape) != 2 or A.shape[0] != A.shape[1]:\n raise ValueError('balance_matrix: Expecting square matrix')\n N = A.shape[0]\n gebal ...
def x(F, L, Qc, H, Pinf, P0, dF=None, dQc=None, dPinf=None, dP0=None): bF, T, T_inv = balance_matrix(F) bL = np.dot(T_inv, L) bQc = Qc bH = np.dot(H, T) bPinf = np.dot(T_inv, np.dot(Pinf, T_inv.T)) bP0 = np.dot(T_inv, np.dot(P0, T_inv.T)) if dF is not None: bdF = np.zeros(dF.shape) for i in range(dF.shape[2]): bdF[:, :, i] = np.dot(T_inv, np.dot(dF[:, :, i], T)) else: bdF = None if dPinf is not None: bdPinf = np.zeros(dPinf.shape) for i in range(dPinf.shape[2]): bdPinf[:, :, i] = np.dot(T_inv, np.dot(dPinf[:, :, i], T_inv.T)) else: bdPinf = None if dP0 is not None: bdP0 = np.zeros(dP0.shape) for i in range(dP0.shape[2]): bdP0[:, :, i] = np.dot(T_inv, np.dot(dP0[:, :, i], T_inv.T)) else: bdP0 = None bdQc = dQc return bF, bL, bQc, bH, bPinf, bP0, bdF, bdQc, bdPinf, bdP0
balance_ss_model
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(F, L, Qc, H, Pinf, P0, dF=None, dQc=None, dPinf=None, dP0=None):\n bF, T, T_inv = balance_matrix(F)\n bL = np.dot(T_inv, L)\n bQc = Qc\n bH = np.do...
def x(self, k): ind = int(self.index[self.R_time_var_index, k]) R = self.R[:, :, ind] if R.shape[0] == 1: inv_square_root = np.sqrt(1.0 / R) elif self.svd_each_time: U, S, Vh = sp.linalg.svd(R, full_matrices=False, compute_uv=True, overwrite_a=False, check_finite=True) inv_square_root = U * 1.0 / np.sqrt(S) elif ind in self.R_square_root: inv_square_root = self.R_square_root[ind] else: U, S, Vh = sp.linalg.svd(R, full_matrices=False, compute_uv=True, overwrite_a=False, check_finite=True) inv_square_root = U * 1.0 / np.sqrt(S) self.R_square_root[ind] = inv_square_root return inv_square_root
R_isrk
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, k):\n ind = int(self.index[self.R_time_var_index, k])\n R = self.R[:, :, ind]\n if R.shape[0] == 1:\n inv_square_root = np.sqrt(1.0 / R)\...
def x(self, k, m_pred, P_pred): return self.H[:, :, int(self.index[self.H_time_var_index, k])]
Hk
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, k, m_pred, P_pred):\n return self.H[:, :, int(self.index[self.H_time_var_index, k])]\n", "role": "user" }, { "content": "Hk", "role": ...
def x(self, k): return self.Q[:, :, self.index[self.Q_time_var_index, k]]
Qk
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, k):\n return self.Q[:, :, self.index[self.Q_time_var_index, k]]\n", "role": "user" }, { "content": "Qk", "role": "assistant" } ]
def x(self, k): ind = self.index[self.Q_time_var_index, k] Q = self.Q[:, :, ind] if Q.shape[0] == 1: square_root = np.sqrt(Q) elif self.svd_each_time: U, S, Vh = sp.linalg.svd(Q, full_matrices=False, compute_uv=True, overwrite_a=False, check_finite=True) square_root = U * np.sqrt(S) elif ind in self.Q_square_root: square_root = self.Q_square_root[ind] else: U, S, Vh = sp.linalg.svd(Q, full_matrices=False, compute_uv=True, overwrite_a=False, check_finite=True) square_root = U * np.sqrt(S) self.Q_square_root[ind] = square_root return square_root
Q_srk
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, k):\n ind = self.index[self.Q_time_var_index, k]\n Q = self.Q[:, :, ind]\n if Q.shape[0] == 1:\n square_root = np.sqrt(Q)\n elif self....
def x(self, k, m_pred, P_pred): return self.A[:, :, self.index[self.A_time_var_index, k]]
Ak
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, k, m_pred, P_pred):\n return self.A[:, :, self.index[self.A_time_var_index, k]]\n", "role": "user" }, { "content": "Ak", "role": "assi...
def x(shape, desired_dim=3): if len(shape) > 3: raise ValueError( """Input array is not supposed to be more than 3 dimensional.""" ) if len(shape) > desired_dim: raise ValueError('Input array shape is more than desired shape.') elif len(shape) == 1: if desired_dim == 3: return (shape[0], 1, 1), shape elif desired_dim == 2: return (shape[0], 1), shape elif len(shape) == 2: if desired_dim == 3: return ((shape[1], 1, 1), shape) if shape[0] == 1 else ((shape[ 0], shape[1], 1), shape) elif desired_dim == 2: return ((shape[1], 1), shape) if shape[0] == 1 else ((shape[0], shape[1]), None) else: return shape, None
reshape_input_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(shape, desired_dim=3):\n if len(shape) > 3:\n raise ValueError(\n \"\"\"Input array is not supposed to be more\n ...
def x(cls, p_state_dim, p_a, p_f_A, p_f_Q, p_h, p_f_H, p_f_R, Y, m_init= None, P_init=None, calc_log_likelihood=False): Y.shape, old_Y_shape = cls._reshape_input_data(Y.shape) if m_init is None: m_init = np.zeros((p_state_dim, 1)) else: m_init = np.atleast_2d(m_init).T if P_init is None: P_init = np.eye(p_state_dim) elif not isinstance(P_init, collections.Iterable): P_init = P_init * np.eye(p_state_dim) if p_a is None: p_a = lambda k, m, A: np.dot(A, m) old_A_shape = None if not isinstance(p_f_A, types.FunctionType): p_f_A = np.atleast_1d(p_f_A) p_A, old_A_shape = cls._check_A_matrix(p_f_A) p_f_A = lambda k, m, P: p_A[:, :, 0] elif p_f_A(1, m_init, P_init).shape[0] != m_init.shape[0]: raise ValueError('p_f_A function returns matrix of wrong size') old_Q_shape = None if not isinstance(p_f_Q, types.FunctionType): p_f_Q = np.atleast_1d(p_f_Q) p_Q, old_Q_shape = cls._check_Q_matrix(p_f_Q) p_f_Q = lambda k: p_Q[:, :, 0] elif p_f_Q(1).shape[0] != m_init.shape[0]: raise ValueError('p_f_Q function returns matrix of wrong size') if p_h is None: lambda k, m, H: np.dot(H, m) old_H_shape = None if not isinstance(p_f_H, types.FunctionType): p_f_H = np.atleast_1d(p_f_H) p_H, old_H_shape = cls._check_H_matrix(p_f_H) p_f_H = lambda k, m, P: p_H elif p_f_H(1, m_init, P_init).shape[0] != Y.shape[1]: raise ValueError('p_f_H function returns matrix of wrong size') old_R_shape = None if not isinstance(p_f_R, types.FunctionType): p_f_R = np.atleast_1d(p_f_R) p_R, old_R_shape = cls._check_H_matrix(p_f_R) p_f_R = lambda k: p_R elif p_f_R(1).shape[0] != m_init.shape[0]: raise ValueError('p_f_R function returns matrix of wrong size') class measurement_callables_class(R_handling_Class): def x(self, R, index, R_time_var_index, unique_R_number): super(measurement_callables_class, self).__init__(R, index, R_time_var_index, unique_R_number) Hk = AddMethodToClass(f_H) f_h = AddMethodToClass(f_hl) M, P, log_likelihood, grad_log_likelihood = cls._kalman_algorithm_raw( p_state_dim, p_a, p_f_A, p_f_Q, p_h, p_f_H, p_f_R, Y, m_init, P_init, calc_log_likelihood, calc_grad_log_likelihood=False, grad_calc_params=None) if old_Y_shape is not None: Y.shape = old_Y_shape if old_A_shape is not None: p_A.shape = old_A_shape if old_Q_shape is not None: p_Q.shape = old_Q_shape if old_H_shape is not None: p_H.shape = old_H_shape if old_R_shape is not None: p_R.shape = old_R_shape return M, P
extended_kalman_filter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, p_state_dim, p_a, p_f_A, p_f_Q, p_h, p_f_H, p_f_R, Y, m_init=\n None, P_init=None, calc_log_likelihood=False):\n Y.shape, old_Y_shape = cls._reshape...
def x(cls, state_dim, p_dynamic_callables, p_measurement_callables, Y, m_init, P_init, p_kalman_filter_type='regular', calc_log_likelihood= False, calc_grad_log_likelihood=False, grad_params_no=None, dm_init= None, dP_init=None): steps_no = Y.shape[0] time_series_no = Y.shape[2] M = np.empty((steps_no + 1, state_dim, time_series_no)) M[0, :, :] = m_init P = np.empty((steps_no + 1, state_dim, state_dim)) P_init = 0.5 * (P_init + P_init.T) P[0, :, :] = P_init if p_kalman_filter_type == 'svd': U, S, Vh = sp.linalg.svd(P_init, full_matrices=False, compute_uv= True, overwrite_a=False, check_finite=True) S[S == 0] = 1e-17 P_upd = P_init, S, U log_likelihood = 0 if calc_log_likelihood else None grad_log_likelihood = 0 if calc_grad_log_likelihood else None dm_upd = dm_init dP_upd = dP_init for k in range(0, steps_no): prev_mean = M[k, :, :] if p_kalman_filter_type == 'svd': m_pred, P_pred, dm_pred, dP_pred = cls._kalman_prediction_step_SVD( k, prev_mean, P_upd, p_dynamic_callables, calc_grad_log_likelihood=calc_grad_log_likelihood, p_dm= dm_upd, p_dP=dP_upd) else: m_pred, P_pred, dm_pred, dP_pred = cls._kalman_prediction_step(k, prev_mean, P[k, :, :], p_dynamic_callables, calc_grad_log_likelihood=calc_grad_log_likelihood, p_dm= dm_upd, p_dP=dP_upd) k_measurment = Y[k, :, :] if np.any(np.isnan(k_measurment)) == False: if p_kalman_filter_type == 'svd': (m_upd, P_upd, log_likelihood_update, dm_upd, dP_upd, d_log_likelihood_update) = (cls._kalman_update_step_SVD (k, m_pred, P_pred, p_measurement_callables, k_measurment, calc_log_likelihood=calc_log_likelihood, calc_grad_log_likelihood=calc_grad_log_likelihood, p_dm =dm_pred, p_dP=dP_pred)) else: (m_upd, P_upd, log_likelihood_update, dm_upd, dP_upd, d_log_likelihood_update) = (cls._kalman_update_step(k, m_pred, P_pred, p_measurement_callables, k_measurment, calc_log_likelihood=calc_log_likelihood, calc_grad_log_likelihood=calc_grad_log_likelihood, p_dm =dm_pred, p_dP=dP_pred)) elif not np.all(np.isnan(k_measurment)): raise ValueError( """Nan measurements are currently not supported if they are intermixed with not NaN measurements""" ) else: m_upd = m_pred P_upd = P_pred dm_upd = dm_pred dP_upd = dP_pred if calc_log_likelihood: log_likelihood_update = np.zeros((time_series_no,)) if calc_grad_log_likelihood: d_log_likelihood_update = np.zeros((grad_params_no, time_series_no)) if calc_log_likelihood: log_likelihood += log_likelihood_update if calc_grad_log_likelihood: grad_log_likelihood += d_log_likelihood_update M[k + 1, :, :] = m_upd if p_kalman_filter_type == 'svd': P[k + 1, :, :] = P_upd[0] else: P[k + 1, :, :] = P_upd return (M, P, log_likelihood, grad_log_likelihood, p_dynamic_callables. reset(False))
kalman_algorithm_raw
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, state_dim, p_dynamic_callables, p_measurement_callables, Y,\n m_init, P_init, p_kalman_filter_type='regular', calc_log_likelihood=\n False, calc_gra...
def x(k, p_m, p_P, p_dyn_model_callable, calc_grad_log_likelihood=False, p_dm=None, p_dP=None): A = p_dyn_model_callable.Ak(k, p_m, p_P) Q = p_dyn_model_callable.Qk(k) m_pred = p_dyn_model_callable.f_a(k, p_m, A) P_pred = A.dot(p_P).dot(A.T) + Q if calc_grad_log_likelihood: dA_all_params = p_dyn_model_callable.dAk(k) dQ_all_params = p_dyn_model_callable.dQk(k) param_number = p_dP.shape[2] dm_pred = np.empty(p_dm.shape) dP_pred = np.empty(p_dP.shape) for j in range(param_number): dA = dA_all_params[:, :, j] dQ = dQ_all_params[:, :, j] dP = p_dP[:, :, j] dm = p_dm[:, :, j] dm_pred[:, :, j] = np.dot(dA, p_m) + np.dot(A, dm) dP_pred[:, :, j] = np.dot(dA, np.dot(p_P, A.T)) dP_pred[:, :, j] += dP_pred[:, :, j].T dP_pred[:, :, j] += np.dot(A, np.dot(dP, A.T)) + dQ dP_pred[:, :, j] = 0.5 * (dP_pred[:, :, j] + dP_pred[:, :, j].T) else: dm_pred = None dP_pred = None return m_pred, P_pred, dm_pred, dP_pred
kalman_prediction_step
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(k, p_m, p_P, p_dyn_model_callable, calc_grad_log_likelihood=False,\n p_dm=None, p_dP=None):\n A = p_dyn_model_callable.Ak(k, p_m, p_P)\n Q = p_dyn_mod...
def x(k, p_m, p_P, p_dyn_model_callable, calc_grad_log_likelihood=False, p_dm=None, p_dP=None): Prev_cov, S_old, V_old = p_P A = p_dyn_model_callable.Ak(k, p_m, Prev_cov) Q = p_dyn_model_callable.Qk(k) Q_sr = p_dyn_model_callable.Q_srk(k) m_pred = p_dyn_model_callable.f_a(k, p_m, A) svd_1_matr = np.vstack(((np.sqrt(S_old) * np.dot(A, V_old)).T, Q_sr.T)) U, S, Vh = sp.linalg.svd(svd_1_matr, full_matrices=False, compute_uv= True, overwrite_a=False, check_finite=True) V_new = Vh.T S_new = S ** 2 P_pred = np.dot(V_new * S_new, V_new.T) P_pred = P_pred, S_new, Vh.T if calc_grad_log_likelihood: dA_all_params = p_dyn_model_callable.dAk(k) dQ_all_params = p_dyn_model_callable.dQk(k) param_number = p_dP.shape[2] dm_pred = np.empty(p_dm.shape) dP_pred = np.empty(p_dP.shape) for j in range(param_number): dA = dA_all_params[:, :, j] dQ = dQ_all_params[:, :, j] dm_pred[:, :, j] = np.dot(dA, p_m) + np.dot(A, p_dm[:, :, j]) dP_pred[:, :, j] = np.dot(dA, np.dot(Prev_cov, A.T)) dP_pred[:, :, j] += dP_pred[:, :, j].T dP_pred[:, :, j] += np.dot(A, np.dot(p_dP[:, :, j], A.T)) + dQ dP_pred[:, :, j] = 0.5 * (dP_pred[:, :, j] + dP_pred[:, :, j].T) else: dm_pred = None dP_pred = None return m_pred, P_pred, dm_pred, dP_pred
kalman_prediction_step_SVD
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(k, p_m, p_P, p_dyn_model_callable, calc_grad_log_likelihood=False,\n p_dm=None, p_dP=None):\n Prev_cov, S_old, V_old = p_P\n A = p_dyn_model_callable....
def x(k, p_m, p_P, p_meas_model_callable, measurement, calc_log_likelihood= False, calc_grad_log_likelihood=False, p_dm=None, p_dP=None): m_pred = p_m P_pred = p_P H = p_meas_model_callable.Hk(k, m_pred, P_pred) R = p_meas_model_callable.Rk(k) time_series_no = p_m.shape[1] log_likelihood_update = None dm_upd = None dP_upd = None d_log_likelihood_update = None v = measurement - p_meas_model_callable.f_h(k, m_pred, H) S = H.dot(P_pred).dot(H.T) + R if measurement.shape[0] == 1: if S < 0: raise ValueError('Kalman Filter Update: S is negative step %i' % k) K = P_pred.dot(H.T) / S if calc_log_likelihood: log_likelihood_update = -0.5 * (np.log(2 * np.pi) + np.log(S) + v * v / S) if np.any(np.isnan(log_likelihood_update)): raise ValueError('Nan values in likelihood update!') LL = None islower = None else: LL, islower = linalg.cho_factor(S) K = linalg.cho_solve((LL, islower), H.dot(P_pred.T)).T if calc_log_likelihood: log_likelihood_update = -0.5 * (v.shape[0] * np.log(2 * np.pi) + 2 * np.sum(np.log(np.diag(LL))) + np.sum(linalg.cho_solve(( LL, islower), v) * v, axis=0)) if calc_grad_log_likelihood: dm_pred_all_params = p_dm dP_pred_all_params = p_dP param_number = p_dP.shape[2] dH_all_params = p_meas_model_callable.dHk(k) dR_all_params = p_meas_model_callable.dRk(k) dm_upd = np.empty(dm_pred_all_params.shape) dP_upd = np.empty(dP_pred_all_params.shape) d_log_likelihood_update = np.empty((param_number, time_series_no)) for param in range(param_number): dH = dH_all_params[:, :, param] dR = dR_all_params[:, :, param] dm_pred = dm_pred_all_params[:, :, param] dP_pred = dP_pred_all_params[:, :, param] dv = -np.dot(dH, m_pred) - np.dot(H, dm_pred) dS = np.dot(dH, np.dot(P_pred, H.T)) dS += dS.T dS += np.dot(H, np.dot(dP_pred, H.T)) + dR if LL is not None: tmp1 = linalg.cho_solve((LL, islower), H).T tmp2 = linalg.cho_solve((LL, islower), dH).T tmp3 = linalg.cho_solve((LL, islower), dS).T else: tmp1 = H.T / S tmp2 = dH.T / S tmp3 = dS.T / S dK = np.dot(dP_pred, tmp1) + np.dot(P_pred, tmp2) - np.dot(P_pred, np.dot(tmp1, tmp3)) dm_upd[:, :, param] = dm_pred + np.dot(dK, v) + np.dot(K, dv) dP_upd[:, :, param] = -np.dot(dK, np.dot(S, K.T)) dP_upd[:, :, param] += dP_upd[:, :, param].T dP_upd[:, :, param] += dP_pred - np.dot(K, np.dot(dS, K.T)) dP_upd[:, :, param] = 0.5 * (dP_upd[:, :, param] + dP_upd[:, :, param].T) if LL is not None: tmp5 = linalg.cho_solve((LL, islower), v) else: tmp5 = v / S d_log_likelihood_update[param, :] = -(0.5 * np.sum(np.diag(tmp3 )) + np.sum(tmp5 * dv, axis=0) - 0.5 * np.sum(tmp5 * np.dot (dS, tmp5), axis=0)) m_upd = m_pred + K.dot(v) P_upd = K.dot(S).dot(K.T) P_upd = 0.5 * (P_upd + P_upd.T) P_upd = P_pred - P_upd return (m_upd, P_upd, log_likelihood_update, dm_upd, dP_upd, d_log_likelihood_update)
kalman_update_step
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(k, p_m, p_P, p_meas_model_callable, measurement, calc_log_likelihood=\n False, calc_grad_log_likelihood=False, p_dm=None, p_dP=None):\n m_pred = p_m\n ...
def x(k, p_m, p_P, p_meas_model_callable, measurement, calc_log_likelihood= False, calc_grad_log_likelihood=False, p_dm=None, p_dP=None): m_pred = p_m P_pred, S_pred, V_pred = p_P H = p_meas_model_callable.Hk(k, m_pred, P_pred) R = p_meas_model_callable.Rk(k) R_isr = p_meas_model_callable.R_isrk(k) time_series_no = p_m.shape[1] log_likelihood_update = None dm_upd = None dP_upd = None d_log_likelihood_update = None v = measurement - p_meas_model_callable.f_h(k, m_pred, H) svd_2_matr = np.vstack((np.dot(R_isr.T, np.dot(H, V_pred)), np.diag(1.0 / np.sqrt(S_pred)))) U, S, Vh = sp.linalg.svd(svd_2_matr, full_matrices=False, compute_uv= True, overwrite_a=False, check_finite=True) U_upd = np.dot(V_pred, Vh.T) S_upd = (1.0 / S) ** 2 P_upd = np.dot(U_upd * S_upd, U_upd.T) P_upd = P_upd, S_upd, U_upd S = H.dot(P_pred).dot(H.T) + R if measurement.shape[0] == 1: if S < 0: raise ValueError('Kalman Filter Update: S is negative step %i' % k) K = P_pred.dot(H.T) / S if calc_log_likelihood: log_likelihood_update = -0.5 * (np.log(2 * np.pi) + np.log(S) + v * v / S) if np.any(np.isnan(log_likelihood_update)): raise ValueError('Nan values in likelihood update!') LL = None islower = None else: LL, islower = linalg.cho_factor(S) K = linalg.cho_solve((LL, islower), H.dot(P_pred.T)).T if calc_log_likelihood: log_likelihood_update = -0.5 * (v.shape[0] * np.log(2 * np.pi) + 2 * np.sum(np.log(np.diag(LL))) + np.sum(linalg.cho_solve(( LL, islower), v) * v, axis=0)) if calc_grad_log_likelihood: dm_pred_all_params = p_dm dP_pred_all_params = p_dP param_number = p_dP.shape[2] dH_all_params = p_meas_model_callable.dHk(k) dR_all_params = p_meas_model_callable.dRk(k) dm_upd = np.empty(dm_pred_all_params.shape) dP_upd = np.empty(dP_pred_all_params.shape) d_log_likelihood_update = np.empty((param_number, time_series_no)) for param in range(param_number): dH = dH_all_params[:, :, param] dR = dR_all_params[:, :, param] dm_pred = dm_pred_all_params[:, :, param] dP_pred = dP_pred_all_params[:, :, param] dv = -np.dot(dH, m_pred) - np.dot(H, dm_pred) dS = np.dot(dH, np.dot(P_pred, H.T)) dS += dS.T dS += np.dot(H, np.dot(dP_pred, H.T)) + dR if LL is not None: tmp1 = linalg.cho_solve((LL, islower), H).T tmp2 = linalg.cho_solve((LL, islower), dH).T tmp3 = linalg.cho_solve((LL, islower), dS).T else: tmp1 = H.T / S tmp2 = dH.T / S tmp3 = dS.T / S dK = np.dot(dP_pred, tmp1) + np.dot(P_pred, tmp2) - np.dot(P_pred, np.dot(tmp1, tmp3)) dm_upd[:, :, param] = dm_pred + np.dot(dK, v) + np.dot(K, dv) dP_upd[:, :, param] = -np.dot(dK, np.dot(S, K.T)) dP_upd[:, :, param] += dP_upd[:, :, param].T dP_upd[:, :, param] += dP_pred - np.dot(K, np.dot(dS, K.T)) dP_upd[:, :, param] = 0.5 * (dP_upd[:, :, param] + dP_upd[:, :, param].T) if LL is not None: tmp5 = linalg.cho_solve((LL, islower), v) else: tmp5 = v / S d_log_likelihood_update[param, :] = -(0.5 * np.sum(np.diag(tmp3 )) + np.sum(tmp5 * dv, axis=0) - 0.5 * np.sum(tmp5 * np.dot (dS, tmp5), axis=0)) m_upd = m_pred + K.dot(v) return (m_upd, P_upd, log_likelihood_update, dm_upd, dP_upd, d_log_likelihood_update)
kalman_update_step_SVD
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(k, p_m, p_P, p_meas_model_callable, measurement, calc_log_likelihood=\n False, calc_grad_log_likelihood=False, p_dm=None, p_dP=None):\n m_pred = p_m\n ...
def x(k, p_m, p_P, p_m_pred, p_P_pred, p_m_prev_step, p_P_prev_step, p_dynamic_callables): A = p_dynamic_callables.Ak(k, p_m, p_P) tmp = np.dot(A, p_P.T) if A.shape[0] == 1: G = tmp.T / p_P_pred else: try: LL, islower = linalg.cho_factor(p_P_pred) G = linalg.cho_solve((LL, islower), tmp).T except: res = sp.linalg.lstsq(p_P_pred, tmp) G = res[0].T m_upd = p_m + G.dot(p_m_prev_step - p_m_pred) P_upd = p_P + G.dot(p_P_prev_step - p_P_pred).dot(G.T) P_upd = 0.5 * (P_upd + P_upd.T) return m_upd, P_upd, G
rts_smoother_update_step
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(k, p_m, p_P, p_m_pred, p_P_pred, p_m_prev_step, p_P_prev_step,\n p_dynamic_callables):\n A = p_dynamic_callables.Ak(k, p_m, p_P)\n tmp = np.dot(A, p_P...
def x(cls, state_dim, p_dynamic_callables, filter_means, filter_covars): no_steps = filter_covars.shape[0] - 1 M = np.empty(filter_means.shape) P = np.empty(filter_covars.shape) M[-1, :] = filter_means[-1, :] P[-1, :, :] = filter_covars[-1, :, :] for k in range(no_steps - 1, -1, -1): m_pred, P_pred, tmp1, tmp2 = cls._kalman_prediction_step(k, filter_means[k, :], filter_covars[k, :, :], p_dynamic_callables, calc_grad_log_likelihood=False) p_m = filter_means[k, :] if len(p_m.shape) < 2: p_m.shape = p_m.shape[0], 1 p_m_prev_step = M[k + 1, :] if len(p_m_prev_step.shape) < 2: p_m_prev_step.shape = p_m_prev_step.shape[0], 1 m_upd, P_upd, G_tmp = cls._rts_smoother_update_step(k, p_m, filter_covars[k, :, :], m_pred, P_pred, p_m_prev_step, P[k + 1, :, :], p_dynamic_callables) M[k, :] = m_upd P[k, :, :] = P_upd return M, P
rts_smoother
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, state_dim, p_dynamic_callables, filter_means, filter_covars):\n no_steps = filter_covars.shape[0] - 1\n M = np.empty(filter_means.shape)\n P = np...
def x(A, Q, H, R, m_init, P_init, measurements, M, P, G, dA, dQ, dH, dR, dm_init, dP_init): import pdb pdb.set_trace() param_number = dA.shape[-1] d_log_likelihood_update = np.empty((param_number, 1)) sample_no = measurements.shape[0] P_1 = P[1:, :, :] P_2 = P[0:-1, :, :] M_1 = M[1:, :] M_2 = M[0:-1, :] Sigma = np.mean(P_1, axis=0) + np.dot(M_1.T, M_1) / sample_no Phi = np.mean(P_2, axis=0) + np.dot(M_2.T, M_2) / sample_no B = np.dot(measurements.T, M_1) / sample_no C = (sp.einsum('ijk,ikl', P_1, G) + np.dot(M_1.T, M_2)) / sample_no D = np.dot(measurements.T, measurements) / sample_no try: P_init_inv = sp.linalg.inv(P_init) if np.max(np.abs(P_init_inv)) > 100000000000000.0: compute_P_init_terms = False else: compute_P_init_terms = True except np.linalg.LinAlgError: compute_P_init_terms = False try: Q_inv = sp.linalg.inv(Q) if np.max(np.abs(Q_inv)) > 100000000000000.0: compute_Q_terms = False else: compute_Q_terms = True except np.linalg.LinAlgError: compute_Q_terms = False try: R_inv = sp.linalg.inv(R) if np.max(np.abs(R_inv)) > 100000000000000.0: compute_R_terms = False else: compute_R_terms = True except np.linalg.LinAlgError: compute_R_terms = False d_log_likelihood_update = np.zeros((param_number, 1)) for j in range(param_number): if compute_P_init_terms: d_log_likelihood_update[j, :] -= 0.5 * np.sum(P_init_inv * dP_init[:, :, j].T) M0_smoothed = M[0] M0_smoothed.shape = M0_smoothed.shape[0], 1 tmp1 = np.dot(dP_init[:, :, j], np.dot(P_init_inv, P[0, :, :] + sp.outer(M0_smoothed - m_init, M0_smoothed - m_init))) d_log_likelihood_update[j, :] += 0.5 * np.sum(P_init_inv * tmp1.T) tmp2 = sp.outer(dm_init[:, j], M0_smoothed) tmp2 += tmp2.T d_log_likelihood_update[j, :] += 0.5 * np.sum(P_init_inv * tmp2.T) if compute_Q_terms: d_log_likelihood_update[j, :] -= sample_no / 2.0 * np.sum(Q_inv * dQ[:, :, j].T) tmp1 = np.dot(C, A.T) tmp1 += tmp1.T tmp1 = Sigma - tmp1 + np.dot(A, np.dot(Phi, A.T)) tmp1 = np.dot(dQ[:, :, j], np.dot(Q_inv, tmp1)) d_log_likelihood_update[j, :] += sample_no / 2.0 * np.sum(Q_inv * tmp1.T) tmp2 = np.dot(dA[:, :, j], C.T) tmp2 += tmp2.T tmp3 = np.dot(dA[:, :, j], np.dot(Phi, A.T)) tmp3 += tmp3.T d_log_likelihood_update[j, :] -= sample_no / 2.0 * np.sum(Q_inv .T * (tmp3 - tmp2)) if compute_R_terms: d_log_likelihood_update[j, :] -= sample_no / 2.0 * np.sum(R_inv * dR[:, :, j].T) tmp1 = np.dot(B, H.T) tmp1 += tmp1.T tmp1 = D - tmp1 + np.dot(H, np.dot(Sigma, H.T)) tmp1 = np.dot(dR[:, :, j], np.dot(R_inv, tmp1)) d_log_likelihood_update[j, :] += sample_no / 2.0 * np.sum(R_inv * tmp1.T) tmp2 = np.dot(dH[:, :, j], B.T) tmp2 += tmp2.T tmp3 = np.dot(dH[:, :, j], np.dot(Sigma, H.T)) tmp3 += tmp3.T d_log_likelihood_update[j, :] -= sample_no / 2.0 * np.sum(R_inv .T * (tmp3 - tmp2)) return d_log_likelihood_update
EM_gradient
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, Q, H, R, m_init, P_init, measurements, M, P, G, dA, dQ, dH, dR,\n dm_init, dP_init):\n import pdb\n pdb.set_trace()\n param_number = dA.shape[-1...
def x(p_M, state_dim, measurement_dim, which='A'): old_M_shape = None if len(p_M.shape) < 3: old_M_shape = p_M.shape if len(p_M.shape) == 2: p_M.shape = p_M.shape[0], p_M.shape[1], 1 elif len(p_M.shape) == 1: if p_M.shape[0] != 1: raise ValueError( 'Matrix %s is an 1D array, while it must be a matrix or scalar' , which) else: p_M.shape = 1, 1, 1 if which == 'A' or which == 'Q': if p_M.shape[0] != state_dim or p_M.shape[1] != state_dim: raise ValueError('%s must be a square matrix of size (%i,%i)' % (which, state_dim, state_dim)) if which == 'H': if p_M.shape[0] != measurement_dim or p_M.shape[1] != state_dim: raise ValueError( 'H must be of shape (measurement_dim, state_dim) (%i,%i)' % (measurement_dim, state_dim)) if which == 'R': if p_M.shape[0] != measurement_dim or p_M.shape[1] != measurement_dim: raise ValueError( 'R must be of shape (measurement_dim, measurement_dim) (%i,%i)' % (measurement_dim, measurement_dim)) return p_M, old_M_shape
check_SS_matrix
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(p_M, state_dim, measurement_dim, which='A'):\n old_M_shape = None\n if len(p_M.shape) < 3:\n old_M_shape = p_M.shape\n if len(p_M.shape) ==...
def x(dM, state_dim, grad_params_no, which='dA'): if dM is None: dM = np.zeros((state_dim, state_dim, grad_params_no)) elif isinstance(dM, np.ndarray): if state_dim == 1: if len(dM.shape) < 3: dM.shape = 1, 1, 1 elif len(dM.shape) < 3: dM.shape = state_dim, state_dim, 1 elif isinstance(dM, np.int): if state_dim > 1: raise ValueError( 'When computing likelihood gradient wrong %s dimension.' % which) else: dM = np.ones((1, 1, 1)) * dM return dM
check_grad_state_matrices
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(dM, state_dim, grad_params_no, which='dA'):\n if dM is None:\n dM = np.zeros((state_dim, state_dim, grad_params_no))\n elif isinstance(dM, np.ndar...
def x(dM, state_dim, grad_params_no, measurement_dim, which='dH'): if dM is None: if which == 'dH': dM = np.zeros((measurement_dim, state_dim, grad_params_no)) elif which == 'dR': dM = np.zeros((measurement_dim, measurement_dim, grad_params_no)) elif isinstance(dM, np.ndarray): if state_dim == 1: if len(dM.shape) < 3: dM.shape = 1, 1, 1 elif len(dM.shape) < 3: if which == 'dH': dM.shape = measurement_dim, state_dim, 1 elif which == 'dR': dM.shape = measurement_dim, measurement_dim, 1 elif isinstance(dM, np.int): if state_dim > 1: raise ValueError( 'When computing likelihood gradient wrong dH dimension.') else: dM = np.ones((1, 1, 1)) * dM return dM
check_grad_measurement_matrices
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(dM, state_dim, grad_params_no, measurement_dim, which='dH'):\n if dM is None:\n if which == 'dH':\n dM = np.zeros((measurement_dim, state_...
def x(cls, F, L, Qc, p_H, p_R, P_inf, X, Y, index=None, m_init=None, P_init =None, p_kalman_filter_type='regular', calc_log_likelihood=False, calc_grad_log_likelihood=False, grad_params_no=0, grad_calc_params=None): p_H = np.atleast_1d(p_H) p_R = np.atleast_1d(p_R) X.shape, old_X_shape = cls._reshape_input_data(X.shape, 2) if X.shape[1] != 1: raise ValueError('Only one dimensional X data is supported.') Y.shape, old_Y_shape = cls._reshape_input_data(Y.shape) state_dim = F.shape[0] measurement_dim = Y.shape[1] time_series_no = Y.shape[2] if len(p_H.shape) == 3 and len(p_H.shape[2]) != 1 or len(p_R.shape ) == 3 and len(p_R.shape[2]) != 1: model_matrices_chage_with_time = True else: model_matrices_chage_with_time = False old_index_shape = None if index is None: if len(p_H.shape) == 3 or len(p_R.shape) == 3: raise ValueError( 'Parameter index can not be None for time varying matrices (third dimension is present)' ) else: index = np.zeros((1, Y.shape[0])) else: if len(index.shape) == 1: index.shape = 1, index.shape[0] old_index_shape = index.shape[0], if index.shape[1] != Y.shape[0]: raise ValueError( 'Number of measurements must be equal the number of H_{k}, R_{k}' ) if index.shape[0] == 1: H_time_var_index = 0 R_time_var_index = 0 elif index.shape[0] == 4: H_time_var_index = 0 R_time_var_index = 1 else: raise ValueError('First Dimension of index must be either 1 or 2.') p_H, old_H_shape = cls._check_SS_matrix(p_H, state_dim, measurement_dim, which='H') p_R, old_R_shape = cls._check_SS_matrix(p_R, state_dim, measurement_dim, which='R') if m_init is None: m_init = np.zeros((state_dim, time_series_no)) else: m_init = np.atleast_2d(m_init).T if P_init is None: P_init = P_inf.copy() if p_kalman_filter_type not in ('regular', 'svd'): raise ValueError("Kalman filer type neither 'regular nor 'svd'.") if calc_grad_log_likelihood: dF = cls._check_grad_state_matrices(grad_calc_params.get('dF'), state_dim, grad_params_no, which='dA') dQc = cls._check_grad_state_matrices(grad_calc_params.get('dQc'), state_dim, grad_params_no, which='dQ') dP_inf = cls._check_grad_state_matrices(grad_calc_params.get( 'dP_inf'), state_dim, grad_params_no, which='dA') dH = cls._check_grad_measurement_matrices(grad_calc_params.get('dH' ), state_dim, grad_params_no, measurement_dim, which='dH') dR = cls._check_grad_measurement_matrices(grad_calc_params.get('dR' ), state_dim, grad_params_no, measurement_dim, which='dR') dm_init = grad_calc_params.get('dm_init') if dm_init is None: dm_init = np.zeros((state_dim, time_series_no, grad_params_no)) dP_init = grad_calc_params.get('dP_init') if dP_init is None: dP_init = dP_inf(0).copy() else: dP_inf = None dF = None dQc = None dH = None dR = None dm_init = None dP_init = None measurement_callables = Std_Measurement_Callables_Class(p_H, H_time_var_index, p_R, index, R_time_var_index, 20, dH, dR) dynamic_callables = cls._cont_to_discrete_object(X, F, L, Qc, compute_derivatives=calc_grad_log_likelihood, grad_params_no= grad_params_no, P_inf=P_inf, dP_inf=dP_inf, dF=dF, dQc=dQc) if print_verbose: print('General: run Continuos-Discrete Kalman Filter') M, P, log_likelihood, grad_log_likelihood, AQcomp = (cls. _cont_discr_kalman_filter_raw(state_dim, dynamic_callables, measurement_callables, X, Y, m_init=m_init, P_init=P_init, p_kalman_filter_type=p_kalman_filter_type, calc_log_likelihood= calc_log_likelihood, calc_grad_log_likelihood= calc_grad_log_likelihood, grad_params_no=grad_params_no, dm_init= dm_init, dP_init=dP_init)) if old_index_shape is not None: index.shape = old_index_shape if old_X_shape is not None: X.shape = old_X_shape if old_Y_shape is not None: Y.shape = old_Y_shape if old_H_shape is not None: p_H.shape = old_H_shape if old_R_shape is not None: p_R.shape = old_R_shape return M, P, log_likelihood, grad_log_likelihood, AQcomp
cont_discr_kalman_filter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, F, L, Qc, p_H, p_R, P_inf, X, Y, index=None, m_init=None, P_init\n =None, p_kalman_filter_type='regular', calc_log_likelihood=False,\n calc_grad_log...
def x(cls, X, F, L, Qc, compute_derivatives=False, grad_params_no=None, P_inf=None, dP_inf=None, dF=None, dQc=None, dt0=None): unique_round_decimals = 10 threshold_number_of_unique_time_steps = 20 dt = np.empty((X.shape[0],)) dt[1:] = np.diff(X[:, 0], axis=0) if dt0 is None: dt[0] = 0 elif isinstance(dt0, str): dt = dt[1:] else: dt[0] = dt0 unique_indices = np.unique(np.round(dt, decimals=unique_round_decimals)) number_unique_indices = len(unique_indices) if use_cython: class AQcompute_batch(state_space_cython.AQcompute_batch_Cython): def x(self, F, L, Qc, dt, compute_derivatives=False, grad_params_no=None, P_inf=None, dP_inf=None, dF=None, dQc=None ): As, Qs, reconstruct_indices, dAs, dQs = (ContDescrStateSpace .lti_sde_to_descrete(F, L, Qc, dt, compute_derivatives, grad_params_no=grad_params_no, P_inf=P_inf, dP_inf= dP_inf, dF=dF, dQc=dQc)) super(AQcompute_batch, self).__init__(As, Qs, reconstruct_indices, dAs, dQs) else: AQcompute_batch = cls.AQcompute_batch_Python if number_unique_indices > threshold_number_of_unique_time_steps: AQcomp = cls.AQcompute_once(F, L, Qc, dt, compute_derivatives= compute_derivatives, grad_params_no=grad_params_no, P_inf=P_inf, dP_inf=dP_inf, dF=dF, dQc=dQc) if print_verbose: print('CDO: Continue-to-discrete INSTANTANEOUS object is created.' ) print('CDO: Number of different time steps: %i' % ( number_unique_indices,)) else: AQcomp = AQcompute_batch(F, L, Qc, dt, compute_derivatives= compute_derivatives, grad_params_no=grad_params_no, P_inf=P_inf, dP_inf=dP_inf, dF=dF, dQc=dQc) if print_verbose: print('CDO: Continue-to-discrete BATCH object is created.') print('CDO: Number of different time steps: %i' % ( number_unique_indices,)) print('CDO: Total size if its data: %i' % (AQcomp. total_size_of_data,)) return AQcomp
cont_to_discrete_object
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, X, F, L, Qc, compute_derivatives=False, grad_params_no=None,\n P_inf=None, dP_inf=None, dF=None, dQc=None, dt0=None):\n unique_round_decimals = 10\n...
def x(F, L, Qc, dt, compute_derivatives=False, grad_params_no=None, P_inf= None, dP_inf=None, dF=None, dQc=None): n = F.shape[0] if not isinstance(dt, collections.Iterable): A = matrix_exponent(F * dt) Phi = np.zeros((2 * n, 2 * n)) Phi[:n, :n] = F Phi[:n, n:] = L.dot(Qc).dot(L.T) Phi[n:, n:] = -F.T AB = matrix_exponent(Phi * dt) AB = np.dot(AB, np.vstack((np.zeros((n, n)), np.eye(n)))) Q_noise_1 = linalg.solve(AB[n:, :].T, AB[:n, :].T) Q_noise_2 = P_inf - A.dot(P_inf).dot(A.T) if compute_derivatives: dA = np.zeros([n, n, grad_params_no]) dQ = np.zeros([n, n, grad_params_no]) FF = np.zeros([2 * n, 2 * n]) AA = np.zeros([2 * n, 2 * n, grad_params_no]) for p in range(0, grad_params_no): FF[:n, :n] = F FF[n:, :n] = dF[:, :, p] FF[n:, n:] = F AA[:, :, p] = matrix_exponent(FF * dt) if p == 0: A = AA[:n, :n, p] Q_noise_3 = P_inf - A.dot(P_inf).dot(A.T) Q_noise = Q_noise_3 dA[:, :, p] = AA[n:, :n, p] tmp = dA[:, :, p].dot(P_inf).dot(A.T) dQ[:, :, p] = dP_inf[:, :, p] - tmp - A.dot(dP_inf[:, :, p] ).dot(A.T) - tmp.T dQ[:, :, p] = 0.5 * (dQ[:, :, p] + dQ[:, :, p].T) else: dA = None dQ = None Q_noise = Q_noise_2 Q_noise = 0.5 * (Q_noise + Q_noise.T) return A, Q_noise, None, dA, dQ else: dt_unique, tmp, reconstruct_index = np.unique(np.round(dt, 8), return_index=True, return_inverse=True) del tmp A = np.empty((n, n, dt_unique.shape[0])) Q_noise = np.empty((n, n, dt_unique.shape[0])) if compute_derivatives: dA = np.empty((n, n, grad_params_no, dt_unique.shape[0])) dQ = np.empty((n, n, grad_params_no, dt_unique.shape[0])) else: dA = None dQ = None for j in range(0, dt_unique.shape[0]): A[:, :, j], Q_noise[:, :, j ], tmp1, dA_t, dQ_t = ContDescrStateSpace.lti_sde_to_descrete(F , L, Qc, dt_unique[j], compute_derivatives= compute_derivatives, grad_params_no=grad_params_no, P_inf= P_inf, dP_inf=dP_inf, dF=dF, dQc=dQc) if compute_derivatives: dA[:, :, :, j] = dA_t dQ[:, :, :, j] = dQ_t return A, Q_noise, reconstruct_index, dA, dQ
lti_sde_to_descrete
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(F, L, Qc, dt, compute_derivatives=False, grad_params_no=None, P_inf=\n None, dP_inf=None, dF=None, dQc=None):\n n = F.shape[0]\n if not isinstance(dt,...
def x(self, X): preactivations_list = list() net = X preactivations_list.append(X) for W, b, i in zip(self.W_list, self.b_list, np.arange(len(self.W_list))): net = np.dot(net, W) net = net + b if i < len(self.W_list) - 1: preactivations_list.append(net) net = self.act(net) return preactivations_list
f_preactivations
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X):\n preactivations_list = list()\n net = X\n preactivations_list.append(X)\n for W, b, i in zip(self.W_list, self.b_list, np.arange(len(sel...
def x(self): for W, b in zip(self.W_list, self.b_list): W.fix() b.fix()
fix_parameters
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for W, b in zip(self.W_list, self.b_list):\n W.fix()\n b.fix()\n", "role": "user" }, { "content": "fix_parameters", "...
def x(self): for W, b in zip(self.W_list, self.b_list): W.unfix() b.unfix()
unfix_parameters
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for W, b in zip(self.W_list, self.b_list):\n W.unfix()\n b.unfix()\n", "role": "user" }, { "content": "unfix_parameters",...
def x(func): def x(*args, **kwargs): np.seterr(invalid='raise') result = func(*args, **kwargs) np.seterr(invalid='warn') return result return func_wrapper
numpy_invalid_op_as_exception
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(func):\n\n def x(*args, **kwargs):\n np.seterr(invalid='raise')\n result = func(*args, **kwargs)\n np.seterr(invalid='warn')\n r...
def x(A, dA, B, dB, operation='prod'): if dA is None: dA_param_num = 0 dA = np.zeros((A.shape[0], A.shape[1], 1)) else: dA_param_num = dA.shape[2] if dB is None: dB_param_num = 0 dB = np.zeros((B.shape[0], B.shape[1], 1)) else: dB_param_num = dB.shape[2] dC = np.zeros((A.shape[0] * B.shape[0], A.shape[1] * B.shape[1], dA_param_num + dB_param_num)) for k in range(dA_param_num): if operation == 'prod': dC[:, :, k] = np.kron(dA[:, :, k], B) else: dC[:, :, k] = np.kron(dA[:, :, k], np.eye(B.shape[0])) for k in range(dB_param_num): if operation == 'prod': dC[:, :, dA_param_num + k] = np.kron(A, dB[:, :, k]) else: dC[:, :, dA_param_num + k] = np.kron(np.eye(A.shape[0]), dB[:, :, k]) return dC
dkron
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(A, dA, B, dB, operation='prod'):\n if dA is None:\n dA_param_num = 0\n dA = np.zeros((A.shape[0], A.shape[1], 1))\n else:\n dA_param...
def x(self): input_dict = super(Prod, self)._save_to_input_dict() input_dict['class'] = str('GPy.kern.Prod') return input_dict
to_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n input_dict = super(Prod, self)._save_to_input_dict()\n input_dict['class'] = str('GPy.kern.Prod')\n return input_dict\n", "role": "user" ...
def x(self, gradients): part_start_param_index = 0 for p in self.parts: if not p.is_fixed: part_param_num = len(p.param_array) p.sde_update_gradient_full(gradients[part_start_param_index: part_start_param_index + part_param_num]) part_start_param_index += part_param_num
sde_update_gradient_full
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, gradients):\n part_start_param_index = 0\n for p in self.parts:\n if not p.is_fixed:\n part_param_num = len(p.param_array)\n ...
def x(self): input_dict = super(StdPeriodic, self)._save_to_input_dict() input_dict['class'] = 'GPy.kern.StdPeriodic' input_dict['variance'] = self.variance.values.tolist() input_dict['period'] = self.period.values.tolist() input_dict['lengthscale'] = self.lengthscale.values.tolist() input_dict['ARD1'] = self.ARD1 input_dict['ARD2'] = self.ARD2 return input_dict
to_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n input_dict = super(StdPeriodic, self)._save_to_input_dict()\n input_dict['class'] = 'GPy.kern.StdPeriodic'\n input_dict['variance'] = self.va...
def x(self, X, X2=None): if X2 is None: X2 = X base = np.pi * (X[:, None, :] - X2[None, :, :]) / self.period exp_dist = np.exp(-0.5 * np.sum(np.square(np.sin(base) / self. lengthscale), axis=-1)) return self.variance * exp_dist
K
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, X2=None):\n if X2 is None:\n X2 = X\n base = np.pi * (X[:, None, :] - X2[None, :, :]) / self.period\n exp_dist = np.exp(-0.5 * np.sum(...
def x(self, X): ret = np.empty(X.shape[0]) ret[:] = self.variance return ret
Kdiag
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X):\n ret = np.empty(X.shape[0])\n ret[:] = self.variance\n return ret\n", "role": "user" }, { "content": "Kdiag", "role": "assi...
def x(self, dL_dK, X, X2=None): if X2 is None: X2 = X base = np.pi * (X[:, None, :] - X2[None, :, :]) / self.period sin_base = np.sin(base) exp_dist = np.exp(-0.5 * np.sum(np.square(sin_base / self.lengthscale), axis=-1)) dwl = self.variance * (1.0 / np.square(self.lengthscale) ) * sin_base * np.cos(base) * (base / self.period) dl = self.variance * np.square(sin_base) / np.power(self.lengthscale, 3) self.variance.gradient = np.sum(exp_dist * dL_dK) if self.ARD1: self.period.gradient = (dwl * exp_dist[:, :, None] * dL_dK[:, :, None] ).sum(0).sum(0) else: self.period.gradient = np.sum(dwl.sum(-1) * exp_dist * dL_dK) if self.ARD2: self.lengthscale.gradient = (dl * exp_dist[:, :, None] * dL_dK[:, :, None]).sum(0).sum(0) else: self.lengthscale.gradient = np.sum(dl.sum(-1) * exp_dist * dL_dK)
update_gradients_full
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dL_dK, X, X2=None):\n if X2 is None:\n X2 = X\n base = np.pi * (X[:, None, :] - X2[None, :, :]) / self.period\n sin_base = np.sin(base)\n...
def x(self, dL_dKdiag, X): self.variance.gradient = np.sum(dL_dKdiag) self.period.gradient = 0 self.lengthscale.gradient = 0
update_gradients_diag
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dL_dKdiag, X):\n self.variance.gradient = np.sum(dL_dKdiag)\n self.period.gradient = 0\n self.lengthscale.gradient = 0\n", "role": "user" ...
def x(shape_records, facecolor='w', edgecolor='k', linewidths=0.5, ax=None, xlims=None, ylims=None): if ax is None: fig = pb.figure() ax = fig.add_subplot(111) for srec in shape_records: points = np.vstack(srec.shape.points) sparts = srec.shape.parts par = list(sparts) + [points.shape[0]] polygs = [] for pj in range(len(sparts)): polygs.append(Polygon(points[par[pj]:par[pj + 1]])) ax.add_collection(PatchCollection(polygs, facecolor=facecolor, edgecolor=edgecolor, linewidths=linewidths)) _box = np.vstack([srec.shape.bbox for srec in shape_records]) minx, miny = np.min(_box[:, :2], 0) maxx, maxy = np.max(_box[:, 2:], 0) if xlims is not None: minx, maxx = xlims if ylims is not None: miny, maxy = ylims ax.set_xlim(minx, maxx) ax.set_ylim(miny, maxy)
plot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(shape_records, facecolor='w', edgecolor='k', linewidths=0.5, ax=None,\n xlims=None, ylims=None):\n if ax is None:\n fig = pb.figure()\n ax ...
def x(sf, regex, field=2): index = [] shape_records = [] for rec in enumerate(sf.shapeRecords()): m = re.search(regex, rec[1].record[field]) if m is not None: index.append(rec[0]) shape_records.append(rec[1]) return index, shape_records
string_match
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sf, regex, field=2):\n index = []\n shape_records = []\n for rec in enumerate(sf.shapeRecords()):\n m = re.search(regex, rec[1].record[field])\...
def x(sf, bbox, inside_only=True): A, B, C, D = bbox index = [] shape_records = [] for rec in enumerate(sf.shapeRecords()): a, b, c, d = rec[1].shape.bbox if inside_only: if A <= a and B <= b and C >= c and D >= d: index.append(rec[0]) shape_records.append(rec[1]) else: cond1 = A <= a and B <= b and C >= a and D >= b cond2 = A <= c and B <= d and C >= c and D >= d cond3 = A <= a and D >= d and C >= a and B <= d cond4 = A <= c and D >= b and C >= c and B <= b cond5 = a <= C and b <= B and d >= D cond6 = c <= A and b <= B and d >= D cond7 = d <= B and a <= A and c >= C cond8 = b <= D and a <= A and c >= C if (cond1 or cond2 or cond3 or cond4 or cond5 or cond6 or cond7 or cond8): index.append(rec[0]) shape_records.append(rec[1]) return index, shape_records
bbox_match
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sf, bbox, inside_only=True):\n A, B, C, D = bbox\n index = []\n shape_records = []\n for rec in enumerate(sf.shapeRecords()):\n a, b, c, d =...
def x(sf, bbox, inside_only=True): index, shape_records = bbox_match(sf, bbox, inside_only) A, B, C, D = bbox plot(shape_records, xlims=[bbox[0], bbox[2]], ylims=[bbox[1], bbox[3]])
plot_bbox
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sf, bbox, inside_only=True):\n index, shape_records = bbox_match(sf, bbox, inside_only)\n A, B, C, D = bbox\n plot(shape_records, xlims=[bbox[0], bbox...
def x(sf, regex, field, **kwargs): index, shape_records = string_match(sf, regex, field) plot(shape_records, **kwargs)
plot_string_match
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sf, regex, field, **kwargs):\n index, shape_records = string_match(sf, regex, field)\n plot(shape_records, **kwargs)\n", "role": "user" }, { ...
def x(sf, ax): limits = sf.bbox xlim = limits[0], limits[2] ylim = limits[1], limits[3] ax.set_xlim(xlim) ax.set_ylim(ylim)
apply_bbox
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sf, ax):\n limits = sf.bbox\n xlim = limits[0], limits[2]\n ylim = limits[1], limits[3]\n ax.set_xlim(xlim)\n ax.set_ylim(ylim)\n", "role": ...
def x(self): self.Y_normalized[:] = self.transform_data() super(WarpedGP, self).parameters_changed() Kiy = self.posterior.woodbury_vector.flatten() self.warping_function.update_grads(self.Y_untransformed, Kiy)
parameters_changed
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.Y_normalized[:] = self.transform_data()\n super(WarpedGP, self).parameters_changed()\n Kiy = self.posterior.woodbury_vector.flatten()\n ...
def x(self): ll = GP.log_likelihood(self) jacobian = self.warping_function.fgrad_y(self.Y_untransformed) return ll + np.log(jacobian).sum()
log_likelihood
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n ll = GP.log_likelihood(self)\n jacobian = self.warping_function.fgrad_y(self.Y_untransformed)\n return ll + np.log(jacobian).sum()\n", "r...
def x(self, mean, std, pred_init=None, deg_gauss_hermite=20): gh_samples, gh_weights = np.polynomial.hermite.hermgauss(deg_gauss_hermite) gh_samples = gh_samples[:, None] gh_weights = gh_weights[None, :] return gh_weights.dot(self._get_warped_term(mean, std, gh_samples) ) / np.sqrt(np.pi)
get_warped_mean
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, mean, std, pred_init=None, deg_gauss_hermite=20):\n gh_samples, gh_weights = np.polynomial.hermite.hermgauss(deg_gauss_hermite)\n gh_samples = gh_s...
def x(self, mean, std, pred_init=None, deg_gauss_hermite=20): gh_samples, gh_weights = np.polynomial.hermite.hermgauss(deg_gauss_hermite) gh_samples = gh_samples[:, None] gh_weights = gh_weights[None, :] arg1 = gh_weights.dot(self._get_warped_term(mean, std, gh_samples, pred_init=pred_init) ** 2) / np.sqrt(np.pi) arg2 = self._get_warped_mean(mean, std, pred_init=pred_init, deg_gauss_hermite=deg_gauss_hermite) return arg1 - arg2 ** 2
get_warped_variance
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, mean, std, pred_init=None, deg_gauss_hermite=20):\n gh_samples, gh_weights = np.polynomial.hermite.hermgauss(deg_gauss_hermite)\n gh_samples = gh_s...
def x(self, Xnew, kern=None, pred_init=None, Y_metadata=None, median=False, deg_gauss_hermite=20, likelihood=None): mean, var = super(WarpedGP, self).predict(Xnew, kern=kern, full_cov= False, likelihood=likelihood) if self.predict_in_warped_space: std = np.sqrt(var) if median: wmean = self.warping_function.f_inv(mean, y=pred_init) else: wmean = self._get_warped_mean(mean, std, pred_init=pred_init, deg_gauss_hermite=deg_gauss_hermite).T wvar = self._get_warped_variance(mean, std, pred_init=pred_init, deg_gauss_hermite=deg_gauss_hermite).T else: wmean = mean wvar = var return wmean, wvar
predict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Xnew, kern=None, pred_init=None, Y_metadata=None, median=False,\n deg_gauss_hermite=20, likelihood=None):\n mean, var = super(WarpedGP, self).predi...
def x(self, X, quantiles=(2.5, 97.5), Y_metadata=None, likelihood=None, kern=None): qs = super(WarpedGP, self).predict_quantiles(X, quantiles, Y_metadata= Y_metadata, likelihood=likelihood, kern=kern) if self.predict_in_warped_space: return [self.warping_function.f_inv(q) for q in qs] return qs
predict_quantiles
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, quantiles=(2.5, 97.5), Y_metadata=None, likelihood=None,\n kern=None):\n qs = super(WarpedGP, self).predict_quantiles(X, quantiles, Y_metadata=\...
def x(self, x_test, y_test, Y_metadata=None): mu_star, var_star = self._raw_predict(x_test) fy = self.warping_function.f(y_test) ll_lpd = self.likelihood.log_predictive_density(fy, mu_star, var_star, Y_metadata=Y_metadata) return ll_lpd + np.log(self.warping_function.fgrad_y(y_test))
log_predictive_density
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, x_test, y_test, Y_metadata=None):\n mu_star, var_star = self._raw_predict(x_test)\n fy = self.warping_function.f(y_test)\n ll_lpd = self.likelih...
def x(self, gradients): self.variance.gradient = gradients[0] self.lengthscale.gradient = gradients[1]
sde_update_gradient_full
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, gradients):\n self.variance.gradient = gradients[0]\n self.lengthscale.gradient = gradients[1]\n", "role": "user" }, { "content": "sde_...
def x(self): variance = float(self.variance.values) lengthscale = float(self.lengthscale.values) lamda = np.sqrt(5.0) / lengthscale kappa = 5.0 / 3.0 * variance / lengthscale ** 2 F = np.array(((0, 1, 0), (0, 0, 1), (-lamda ** 3, -3.0 * lamda ** 2, -3 * lamda))) L = np.array(((0,), (0,), (1,))) Qc = np.array(((variance * 400.0 * np.sqrt(5.0) / 3.0 / lengthscale ** 5,),)) H = np.array(((1, 0, 0),)) Pinf = np.array(((variance, 0, -kappa), (0, kappa, 0), (-kappa, 0, 25.0 * variance / lengthscale ** 4))) P0 = Pinf.copy() dF = np.empty((3, 3, 2)) dQc = np.empty((1, 1, 2)) dPinf = np.empty((3, 3, 2)) dFvariance = np.zeros((3, 3)) dFlengthscale = np.array(((0, 0, 0), (0, 0, 0), (15.0 * np.sqrt(5.0) / lengthscale ** 4, 30.0 / lengthscale ** 3, 3 * np.sqrt(5.0) / lengthscale ** 2))) dQcvariance = np.array(((400 * np.sqrt(5) / 3 / lengthscale ** 5,),)) dQclengthscale = np.array(((-variance * 2000 * np.sqrt(5) / 3 / lengthscale ** 6,),)) dPinf_variance = Pinf / variance kappa2 = -2.0 * kappa / lengthscale dPinf_lengthscale = np.array(((0, 0, -kappa2), (0, kappa2, 0), (-kappa2, 0, -100 * variance / lengthscale ** 5))) dF[:, :, 0] = dFvariance dF[:, :, 1] = dFlengthscale dQc[:, :, 0] = dQcvariance dQc[:, :, 1] = dQclengthscale dPinf[:, :, 0] = dPinf_variance dPinf[:, :, 1] = dPinf_lengthscale dP0 = dPinf.copy() return F, L, Qc, H, Pinf, P0, dF, dQc, dPinf, dP0
sde
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n variance = float(self.variance.values)\n lengthscale = float(self.lengthscale.values)\n lamda = np.sqrt(5.0) / lengthscale\n kappa = 5.0 /...
def x(self, num_samples=1000, hmc_iters=20): params = np.empty((num_samples, self.p.size)) for i in range(num_samples): self.p[:] = np.random.multivariate_normal(np.zeros(self.p.size), self.M ) H_old = self._computeH() theta_old = self.model.optimizer_array.copy() params[i] = self.model.unfixed_param_array self._update(hmc_iters) H_new = self._computeH() if H_old > H_new: k = 1.0 else: k = np.exp(H_old - H_new) if np.random.rand() < k: params[i] = self.model.unfixed_param_array else: self.model.optimizer_array = theta_old return params
sample
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, num_samples=1000, hmc_iters=20):\n params = np.empty((num_samples, self.p.size))\n for i in range(num_samples):\n self.p[:] = np.random.mult...
def x(self): self.posterior, self._log_marginal_likelihood, self.grad_dict = (self. inference_method.inference(self.kern, self.X, self.likelihood, self .Y_normalized, self.Y_metadata)) self.likelihood.update_gradients(self.grad_dict['dL_dthetaL']) self.kern.update_gradients_direct(self.grad_dict['dL_dVar'], self. grad_dict['dL_dLen'])
parameters_changed
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.posterior, self._log_marginal_likelihood, self.grad_dict = (self.\n inference_method.inference(self.kern, self.X, self.likelihood, self...
def x(self, Xnew, full_cov=False, kern=None): if kern is None: kern = self.kern Kmn = kern.K(Xnew, self.X) alpha_kron = self.posterior.alpha mu = np.dot(Kmn, alpha_kron) mu = mu.reshape(-1, 1) Knm = Kmn.T noise = self.likelihood.variance V_kron = self.posterior.V_kron Qs = self.posterior.Qs QTs = self.posterior.QTs A = self.kron_mmprod(QTs, Knm) V_kron = V_kron.reshape(-1, 1) A = A / (V_kron + noise) A = self.kron_mmprod(Qs, A) Kmm = kern.K(Xnew) var = np.diag(Kmm - np.dot(Kmn, A)).copy() var = var.reshape(-1, 1) return mu, var
raw_predict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Xnew, full_cov=False, kern=None):\n if kern is None:\n kern = self.kern\n Kmn = kern.K(Xnew, self.X)\n alpha_kron = self.posterior.alpha\...
def x(self): variance = float(self.variances.values) t0 = float(self.t0) F = np.array(((0, 1.0), (0, 0))) L = np.array(((0,), (1.0,))) Qc = np.zeros((1, 1)) H = np.array(((1.0, 0),)) Pinf = np.zeros((2, 2)) P0 = np.array(((t0 ** 2, t0), (t0, 1))) * variance dF = np.zeros((2, 2, 1)) dQc = np.zeros((1, 1, 1)) dPinf = np.zeros((2, 2, 1)) dP0 = np.zeros((2, 2, 1)) dP0[:, :, 0] = P0 / variance return F, L, Qc, H, Pinf, P0, dF, dQc, dPinf, dP0
sde
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n variance = float(self.variances.values)\n t0 = float(self.t0)\n F = np.array(((0, 1.0), (0, 0)))\n L = np.array(((0,), (1.0,)))\n Qc = ...
def x(self, X, X2=None): X, slices = X[:, :-1], index_to_slices(X[:, -1]) if X2 is None: X2, slices2 = X, slices K = np.zeros((X.shape[0], X.shape[0])) else: X2, slices2 = X2[:, :-1], index_to_slices(X2[:, -1]) K = np.zeros((X.shape[0], X2.shape[0])) tdist = (X[:, 0][:, None] - X2[:, 0][None, :]) ** 2 xdist = (X[:, 1][:, None] - X2[:, 1][None, :]) ** 2 ttdist = X[:, 0][:, None] - X2[:, 0][None, :] vyt = self.variance_Yt vyx = self.variance_Yx lyt = 1 / (2 * self.lengthscale_Yt) lyx = 1 / (2 * self.lengthscale_Yx) a = self.a b = self.b c = self.c kyy = lambda tdist, xdist: np.exp(-lyt * tdist - lyx * xdist) k1 = lambda tdist: 2 * lyt - 4 * lyt ** 2 * tdist k2 = lambda xdist: 4 * lyx ** 2 * xdist - 2 * lyx k3 = (lambda xdist: 3 * 4 * lyx ** 2 - 6 * 8 * xdist * lyx ** 3 + 16 * xdist ** 2 * lyx ** 4) k4 = lambda ttdist: 2 * lyt * ttdist for i, s1 in enumerate(slices): for j, s2 in enumerate(slices2): for ss1 in s1: for ss2 in s2: if i == 0 and j == 0: K[ss1, ss2] = vyt * vyx * kyy(tdist[ss1, ss2], xdist[ss1, ss2]) elif i == 0 and j == 1: K[ss1, ss2] = (-a * k2(xdist[ss1, ss2]) + b * k4( ttdist[ss1, ss2]) + c) * vyt * vyx * kyy(tdist[ ss1, ss2], xdist[ss1, ss2]) elif i == 1 and j == 1: K[ss1, ss2] = (b ** 2 * k1(tdist[ss1, ss2]) - 2 * a * c * k2(xdist[ss1, ss2]) + a ** 2 * k3(xdist[ss1, ss2]) + c ** 2) * vyt * vyx * kyy(tdist[ss1, ss2], xdist[ss1, ss2]) else: K[ss1, ss2] = (-a * k2(xdist[ss1, ss2]) - b * k4( ttdist[ss1, ss2]) + c) * vyt * vyx * kyy(tdist[ ss1, ss2], xdist[ss1, ss2]) return K
K
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, X2=None):\n X, slices = X[:, :-1], index_to_slices(X[:, -1])\n if X2 is None:\n X2, slices2 = X, slices\n K = np.zeros((X.shape[0]...
def x(self, X): vyt = self.variance_Yt vyx = self.variance_Yx lyt = 1.0 / (2 * self.lengthscale_Yt) lyx = 1.0 / (2 * self.lengthscale_Yx) a = self.a b = self.b c = self.c k1 = 2 * lyt * vyt * vyx k2 = -2 * lyx * vyt * vyx k3 = 4 * 3 * lyx ** 2 * vyt * vyx Kdiag = np.zeros(X.shape[0]) slices = index_to_slices(X[:, -1]) for i, ss1 in enumerate(slices): for s1 in ss1: if i == 0: Kdiag[s1] += vyt * vyx elif i == 1: Kdiag[s1] += (b ** 2 * k1 - 2 * a * c * k2 + a ** 2 * k3 + c ** 2 * vyt * vyx) else: raise ValueError('invalid input/output index') return Kdiag
Kdiag
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X):\n vyt = self.variance_Yt\n vyx = self.variance_Yx\n lyt = 1.0 / (2 * self.lengthscale_Yt)\n lyx = 1.0 / (2 * self.lengthscale_Yx)\n a ...
def x(self, X): return X.mean.gradient, X.variance.gradient, X.binary_prob.gradient
get_X_gradients
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X):\n return X.mean.gradient, X.variance.gradient, X.binary_prob.gradient\n", "role": "user" }, { "content": "get_X_gradients", "role"...
def x(self, nSamples, raw_samples=False): assert isinstance(self.kern, kern.Linear) from ..util.linalg import pdinv N, D = self.Y.shape Q = self.X.shape[1] noise_var = self.likelihood.variance.values Xs = np.random.randn(*((nSamples,) + self.X.shape)) * np.sqrt(self.X. variance.values) + self.X.mean.values b = np.random.rand(*((nSamples,) + self.X.shape)) Xs[b > self.X.gamma.values] = 0 invcov = (Xs[:, :, :, None] * Xs[:, :, None, :]).sum(1 ) / noise_var + np.eye(Q) cov = np.array([pdinv(invcov[s_idx])[0] for s_idx in range(invcov.shape [0])]) Ws = np.empty((nSamples, Q, D)) tmp = np.transpose(Xs, (0, 2, 1)).reshape(nSamples * Q, N).dot(self.Y ).reshape(nSamples, Q, D) mean = (cov[:, :, :, None] * tmp[:, None, :, :]).sum(2) / noise_var zeros = np.zeros((Q,)) for s_idx in range(Xs.shape[0]): Ws[s_idx] = np.random.multivariate_normal(mean=zeros, cov=cov[s_idx ], size=(D,)).T + mean[s_idx] if raw_samples: return Ws else: return Ws.mean(0), Ws.std(0)
sample_W
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, nSamples, raw_samples=False):\n assert isinstance(self.kern, kern.Linear)\n from ..util.linalg import pdinv\n N, D = self.Y.shape\n Q = self....
def x(self): if self.L is None: return {'mu': self.mu.tolist(), 'Sigma': self.Sigma.tolist()} else: return {'mu': self.mu.tolist(), 'Sigma': self.Sigma.tolist(), 'L': self.L.tolist()}
to_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.L is None:\n return {'mu': self.mu.tolist(), 'Sigma': self.Sigma.tolist()}\n else:\n return {'mu': self.mu.tolist(), 'Sigm...
def x(self): input_dict = super(EP, self)._save_to_input_dict() input_dict['class' ] = 'GPy.inference.latent_function_inference.expectation_propagation.EP' if self.ga_approx_old is not None: input_dict['ga_approx_old'] = self.ga_approx_old.to_dict() if self._ep_approximation is not None: input_dict['_ep_approximation'] = {} input_dict['_ep_approximation']['post_params' ] = self._ep_approximation[0].to_dict() input_dict['_ep_approximation']['ga_approx'] = self._ep_approximation[1 ].to_dict() input_dict['_ep_approximation']['cav_params'] = self._ep_approximation[ 2].to_dict() input_dict['_ep_approximation']['log_Z_tilde' ] = self._ep_approximation[3].tolist() return input_dict
to_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n input_dict = super(EP, self)._save_to_input_dict()\n input_dict['class'\n ] = 'GPy.inference.latent_function_inference.expectation_propag...
def x(self): input_dict = super(EPDTC, self)._save_to_input_dict() input_dict['class' ] = 'GPy.inference.latent_function_inference.expectation_propagation.EPDTC' if self.ga_approx_old is not None: input_dict['ga_approx_old'] = self.ga_approx_old.to_dict() if self._ep_approximation is not None: input_dict['_ep_approximation'] = {} input_dict['_ep_approximation']['post_params' ] = self._ep_approximation[0].to_dict() input_dict['_ep_approximation']['ga_approx'] = self._ep_approximation[1 ].to_dict() input_dict['_ep_approximation']['log_Z_tilde' ] = self._ep_approximation[2] return input_dict
to_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n input_dict = super(EPDTC, self)._save_to_input_dict()\n input_dict['class'\n ] = 'GPy.inference.latent_function_inference.expectation_pro...
def x(self, other): return 0.5 * (np.sum(self.variance / other.variance) + ((other.mean - self.mean) ** 2 / other.variance).sum() - self.num_data * self. input_dim + np.sum(np.log(other.variance)) - np.sum(np.log(self. variance)))
KL
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, other):\n return 0.5 * (np.sum(self.variance / other.variance) + ((other.mean -\n self.mean) ** 2 / other.variance).sum() - self.num_data * sel...
def x(self, *args, **kwargs): import sys assert 'matplotlib' in sys.modules, 'matplotlib package has not been imported.' from ...plotting.matplot_dep import variational_plots return variational_plots.plot_SpikeSlab(self, *args, **kwargs)
plot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *args, **kwargs):\n import sys\n assert 'matplotlib' in sys.modules, 'matplotlib package has not been imported.'\n from ...plotting.matplot_dep ...
def x(d2f_dg2, dg_dx, df_dg, d2g_dx2): if np.all(dg_dx == 1.0) and np.all(d2g_dx2 == 0): return d2f_dg2 dg_dx_2 = np.clip(dg_dx, -np.inf, _lim_val_square) ** 2 return d2f_dg2 * dg_dx_2 + df_dg * d2g_dx2
chain_2
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(d2f_dg2, dg_dx, df_dg, d2g_dx2):\n if np.all(dg_dx == 1.0) and np.all(d2g_dx2 == 0):\n return d2f_dg2\n dg_dx_2 = np.clip(dg_dx, -np.inf, _lim_val...
def x(d3f_dg3, dg_dx, d2f_dg2, d2g_dx2, df_dg, d3g_dx3): if np.all(dg_dx == 1.0) and np.all(d2g_dx2 == 0) and np.all(d3g_dx3 == 0): return d3f_dg3 dg_dx_3 = np.clip(dg_dx, -np.inf, _lim_val_cube) ** 3 return d3f_dg3 * dg_dx_3 + 3 * d2f_dg2 * dg_dx * d2g_dx2 + df_dg * d3g_dx3
chain_3
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(d3f_dg3, dg_dx, d2f_dg2, d2g_dx2, df_dg, d3g_dx3):\n if np.all(dg_dx == 1.0) and np.all(d2g_dx2 == 0) and np.all(d3g_dx3 == 0):\n return d3f_dg3\n ...
def x(D, n=100, min_max=(-100, 100)): g = np.linspace(min_max[0], min_max[1], n) G = np.ones((n, D)) return G * g[:, None]
linear_grid
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(D, n=100, min_max=(-100, 100)):\n g = np.linspace(min_max[0], min_max[1], n)\n G = np.ones((n, D))\n return G * g[:, None]\n", "role": "user" },...
def x(X, m=10): XXT = np.dot(X, X.T) D = -2.0 * XXT + np.diag(XXT)[:, np.newaxis] + np.diag(XXT)[np.newaxis, :] s = np.random.permutation(X.shape[0])[0] inducing = [s] prob = D[s] / D[s].sum() for z in range(m - 1): s = np.random.multinomial(1, prob.flatten()).argmax() inducing.append(s) prob = D[s] / D[s].sum() inducing = np.array(inducing) return X[inducing]
kmm_init
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(X, m=10):\n XXT = np.dot(X, X.T)\n D = -2.0 * XXT + np.diag(XXT)[:, np.newaxis] + np.diag(XXT)[np.newaxis, :]\n s = np.random.permutation(X.shape[0])[...
def x(*param): import warnings warnings.warn( 'Please use param.values, as this function will be deprecated in the next release.' , DeprecationWarning) assert len(param) > 0, 'At least one parameter needed' if len(param) == 1: return param[0].view(np.ndarray) return [x.view(np.ndarray) for x in param]
param_to_array
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*param):\n import warnings\n warnings.warn(\n 'Please use param.values, as this function will be deprecated in the next release.'\n , Depre...
def x(self, X): Kdiag = np.zeros(X.shape[0]) ly = 1 / self.lengthscale_Y lu = np.sqrt(3) / self.lengthscale_U Vu = self.variance_U Vy = self.variance_Y k1 = (2 * lu + ly) / (lu + ly) ** 2 k2 = (ly - 2 * lu + 2 * lu - ly) / (ly - lu) ** 2 k3 = 1 / (lu + ly) + lu / (lu + ly) ** 2 slices = index_to_slices(X[:, -1]) for i, ss1 in enumerate(slices): for s1 in ss1: if i == 0: Kdiag[s1] += self.variance_U + self.ubias elif i == 1: Kdiag[s1] += Vu * Vy * (k1 + k2 + k3) else: raise ValueError('invalid input/output index') return Kdiag
Kdiag
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X):\n Kdiag = np.zeros(X.shape[0])\n ly = 1 / self.lengthscale_Y\n lu = np.sqrt(3) / self.lengthscale_U\n Vu = self.variance_U\n Vy = self...
def x(self, X, Y): assert X.shape[1] == self.Z.shape[1] self.X, self.Y = X, Y
set_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, Y):\n assert X.shape[1] == self.Z.shape[1]\n self.X, self.Y = X, Y\n", "role": "user" }, { "content": "set_data", "role": "assis...
def x(self): i = next(self.slicer) return self.X_all[i], self.Y_all[i]
new_batch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n i = next(self.slicer)\n return self.X_all[i], self.Y_all[i]\n", "role": "user" }, { "content": "new_batch", "role": "assistant" ...
def x(input_dict): import copy input_dict = copy.deepcopy(input_dict) normalizer_class = input_dict.pop('class') import GPy normalizer_class = eval(normalizer_class) return normalizer_class._build_from_input_dict(normalizer_class, input_dict )
from_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(input_dict):\n import copy\n input_dict = copy.deepcopy(input_dict)\n normalizer_class = input_dict.pop('class')\n import GPy\n normalizer_class...
def x(self): input_dict = super(Standardize, self)._save_to_input_dict() input_dict['class'] = 'GPy.util.normalizer.Standardize' if self.mean is not None: input_dict['mean'] = self.mean.tolist() input_dict['std'] = self.std.tolist() return input_dict
to_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n input_dict = super(Standardize, self)._save_to_input_dict()\n input_dict['class'] = 'GPy.util.normalizer.Standardize'\n if self.mean is not N...