source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(survival_table, observed_deaths_col='observed', censored_col='censored'): T_ = [] E_ = [] W_ = [] for t, row in survival_table.iterrows(): if row[observed_deaths_col] > 0: T_.append(t) E_.append(1) W_.append(row[observed_deaths_col]) if row[censored_col] > 0: T_.append(t) E_.append(0) W_.append(row[censored_col]) return np.asarray(T_), np.asarray(E_), np.asarray(W_)
survival_events_from_table
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(survival_table, observed_deaths_col='observed', censored_col='censored'):\n T_ = []\n E_ = []\n W_ = []\n for t, row in survival_table.iterrows():\n if row[observed_deaths_col] > 0:\n T_.append(t)\n E_.append(1)\n W_.append(row[observed_deaths_col])\n if row[censored_col] > 0:\n T_.append(t)\n E_.append(0)\n W_.append(row[censored_col])\n return np.asarray(T_), np.asarray(E_), np.asarray(W_)\n", "role": "user" }, { "content": "survival_events_from_table", "role": "assistant" } ]
def x(start_times, end_times, fill_date=datetime.today(), freq='D', dayfirst=False, na_values=None): fill_date = pd.to_datetime(fill_date) freq_string = 'timedelta64[%s]' % freq start_times = pd.Series(start_times).copy() end_times = pd.Series(end_times).copy() C = ~(pd.isnull(end_times).values | end_times.isin(na_values or [''])) end_times[~C] = fill_date start_times_ = pd.to_datetime(start_times, dayfirst=dayfirst) end_times_ = pd.to_datetime(end_times, dayfirst=dayfirst, errors='coerce') deaths_after_cutoff = end_times_ > fill_date C[deaths_after_cutoff] = False T = (end_times_ - start_times_).values.astype(freq_string).astype(float) if (T < 0).sum(): warnings.warn('Warning: some values of start_times are after end_times' ) return T, C.values
datetimes_to_durations
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(start_times, end_times, fill_date=datetime.today(), freq='D',\n dayfirst=False, na_values=None):\n fill_date = pd.to_datetime(fill_date)\n freq_string = 'timedelta64[%s]' % freq\n start_times = pd.Series(start_times).copy()\n end_times = pd.Series(end_times).copy()\n C = ~(pd.isnull(end_times).values | end_times.isin(na_values or ['']))\n end_times[~C] = fill_date\n start_times_ = pd.to_datetime(start_times, dayfirst=dayfirst)\n end_times_ = pd.to_datetime(end_times, dayfirst=dayfirst, errors='coerce')\n deaths_after_cutoff = end_times_ > fill_date\n C[deaths_after_cutoff] = False\n T = (end_times_ - start_times_).values.astype(freq_string).astype(float)\n if (T < 0).sum():\n warnings.warn('Warning: some values of start_times are after end_times'\n )\n return T, C.values\n", "role": "user" }, { "content": "datetimes_to_durations", "role": "assistant" } ]
def x(event_times, predicted_event_times, event_observed=None): if event_observed is None: event_observed = np.ones_like(event_times) ix = event_observed.astype(bool) return np.abs(np.log(event_times[ix]) - np.log(predicted_event_times[ix]) ).mean()
l1_log_loss
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(event_times, predicted_event_times, event_observed=None):\n if event_observed is None:\n event_observed = np.ones_like(event_times)\n ix = event_observed.astype(bool)\n return np.abs(np.log(event_times[ix]) - np.log(predicted_event_times[ix])\n ).mean()\n", "role": "user" }, { "content": "l1_log_loss", "role": "assistant" } ]
def x(event_times, predicted_event_times, event_observed=None): if event_observed is None: event_observed = np.ones_like(event_times) ix = event_observed.astype(bool) return np.power(np.log(event_times[ix]) - np.log(predicted_event_times[ ix]), 2).mean()
l2_log_loss
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(event_times, predicted_event_times, event_observed=None):\n if event_observed is None:\n event_observed = np.ones_like(event_times)\n ix = event_observed.astype(bool)\n return np.power(np.log(event_times[ix]) - np.log(predicted_event_times[\n ix]), 2).mean()\n", "role": "user" }, { "content": "l2_log_loss", "role": "assistant" } ]
def x(fitters, df, duration_col, event_col=None, k=5, evaluation_measure= concordance_index, predictor='predict_expectation', predictor_kwargs={}, fitter_kwargs={}): try: fitters = list(fitters) except TypeError: fitters = [fitters] fitterscores = [[] for _ in fitters] n, _ = df.shape df = df.copy() if event_col is None: event_col = 'E' df[event_col] = 1.0 df = df.reindex(np.random.permutation(df.index)).sort_values(event_col) assignments = np.array((n // k + 1) * list(range(1, k + 1))) assignments = assignments[:n] testing_columns = df.columns.drop([duration_col, event_col]) for i in range(1, k + 1): ix = assignments == i training_data = df.loc[~ix] testing_data = df.loc[ix] T_actual = testing_data[duration_col].values E_actual = testing_data[event_col].values X_testing = testing_data[testing_columns] for fitter, scores in zip(fitters, fitterscores): fitter.fit(training_data, duration_col=duration_col, event_col= event_col, **fitter_kwargs) T_pred = getattr(fitter, predictor)(X_testing, **predictor_kwargs ).values try: scores.append(evaluation_measure(T_actual, T_pred, E_actual)) except TypeError: scores.append(evaluation_measure(T_actual, T_pred)) if len(fitters) == 1: return fitterscores[0] return fitterscores
k_fold_cross_validation
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fitters, df, duration_col, event_col=None, k=5, evaluation_measure=\n concordance_index, predictor='predict_expectation', predictor_kwargs={},\n fitter_kwargs={}):\n try:\n fitters = list(fitters)\n except TypeError:\n fitters = [fitters]\n fitterscores = [[] for _ in fitters]\n n, _ = df.shape\n df = df.copy()\n if event_col is None:\n event_col = 'E'\n df[event_col] = 1.0\n df = df.reindex(np.random.permutation(df.index)).sort_values(event_col)\n assignments = np.array((n // k + 1) * list(range(1, k + 1)))\n assignments = assignments[:n]\n testing_columns = df.columns.drop([duration_col, event_col])\n for i in range(1, k + 1):\n ix = assignments == i\n training_data = df.loc[~ix]\n testing_data = df.loc[ix]\n T_actual = testing_data[duration_col].values\n E_actual = testing_data[event_col].values\n X_testing = testing_data[testing_columns]\n for fitter, scores in zip(fitters, fitterscores):\n fitter.fit(training_data, duration_col=duration_col, event_col=\n event_col, **fitter_kwargs)\n T_pred = getattr(fitter, predictor)(X_testing, **predictor_kwargs\n ).values\n try:\n scores.append(evaluation_measure(T_actual, T_pred, E_actual))\n except TypeError:\n scores.append(evaluation_measure(T_actual, T_pred))\n if len(fitters) == 1:\n return fitterscores[0]\n return fitterscores\n", "role": "user" }, { "content": "k_fold_cross_validation", "role": "assistant" } ]
def x(X, mean=None, std=None): if mean is None or std is None: mean = X.mean(0) std = X.std(0) return (X - mean) / std
normalize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(X, mean=None, std=None):\n if mean is None or std is None:\n mean = X.mean(0)\n std = X.std(0)\n return (X - mean) / std\n", "role": "user" }, { "content": "normalize", "role": "assistant" } ]
def x(X, Y, c1=0.0, c2=0.0, offset=None, ix=None): _, d = X.shape if c1 > 0 or c2 > 0: penalizer_matrix = (c1 + c2) * np.eye(d) A = np.dot(X.T, X) + penalizer_matrix else: A = np.dot(X.T, X) if offset is None or c2 == 0: b = np.dot(X.T, Y) else: b = np.dot(X.T, Y) + c2 * offset if ix is not None: M = np.c_[X.T[:, ix], b] else: M = np.c_[X.T, b] R = solve(A, M, assume_a='pos', check_finite=False) return R[:, -1], R[:, :-1]
ridge_regression
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(X, Y, c1=0.0, c2=0.0, offset=None, ix=None):\n _, d = X.shape\n if c1 > 0 or c2 > 0:\n penalizer_matrix = (c1 + c2) * np.eye(d)\n A = np.dot(X.T, X) + penalizer_matrix\n else:\n A = np.dot(X.T, X)\n if offset is None or c2 == 0:\n b = np.dot(X.T, Y)\n else:\n b = np.dot(X.T, Y) + c2 * offset\n if ix is not None:\n M = np.c_[X.T[:, ix], b]\n else:\n M = np.c_[X.T, b]\n R = solve(A, M, assume_a='pos', check_finite=False)\n return R[:, -1], R[:, :-1]\n", "role": "user" }, { "content": "ridge_regression", "role": "assistant" } ]
def x(events, timeline, _additive_f, _additive_var, reverse): if reverse: events = events.sort_index(ascending=False) at_risk = events['entrance'].sum() - events['removed'].cumsum().shift(1 ).fillna(0) deaths = events['observed'] estimate_ = np.cumsum(_additive_f(at_risk, deaths)).sort_index().shift( -1).fillna(0) var_ = np.cumsum(_additive_var(at_risk, deaths)).sort_index().shift(-1 ).fillna(0) else: deaths = events['observed'] entrances = events['entrance'].copy() entrances.iloc[0] = 0 population = events['at_risk'] - entrances estimate_ = np.cumsum(_additive_f(population, deaths)) var_ = np.cumsum(_additive_var(population, deaths)) timeline = sorted(timeline) estimate_ = estimate_.reindex(timeline, method='pad').fillna(0) var_ = var_.reindex(timeline, method='pad') var_.index.name = 'timeline' estimate_.index.name = 'timeline' return estimate_, var_
additive_estimate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(events, timeline, _additive_f, _additive_var, reverse):\n if reverse:\n events = events.sort_index(ascending=False)\n at_risk = events['entrance'].sum() - events['removed'].cumsum().shift(1\n ).fillna(0)\n deaths = events['observed']\n estimate_ = np.cumsum(_additive_f(at_risk, deaths)).sort_index().shift(\n -1).fillna(0)\n var_ = np.cumsum(_additive_var(at_risk, deaths)).sort_index().shift(-1\n ).fillna(0)\n else:\n deaths = events['observed']\n entrances = events['entrance'].copy()\n entrances.iloc[0] = 0\n population = events['at_risk'] - entrances\n estimate_ = np.cumsum(_additive_f(population, deaths))\n var_ = np.cumsum(_additive_var(population, deaths))\n timeline = sorted(timeline)\n estimate_ = estimate_.reindex(timeline, method='pad').fillna(0)\n var_ = var_.reindex(timeline, method='pad')\n var_.index.name = 'timeline'\n estimate_.index.name = 'timeline'\n return estimate_, var_\n", "role": "user" }, { "content": "additive_estimate", "role": "assistant" } ]
def x(durations, event_observed, timeline, entry, weights): n = len(durations) durations = np.asarray(pass_for_numeric_dtypes_or_raise_array(durations) ).reshape((n,)) if event_observed is None: event_observed = np.ones(n, dtype=int) else: event_observed = np.asarray(event_observed).reshape((n,)).copy( ).astype(int) if entry is not None: entry = np.asarray(entry).reshape((n,)) event_table = survival_table_from_events(durations, event_observed, entry, weights=weights) if timeline is None: timeline = event_table.index.values else: timeline = np.asarray(timeline) return durations, event_observed, timeline.astype(float ), entry, event_table
preprocess_inputs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(durations, event_observed, timeline, entry, weights):\n n = len(durations)\n durations = np.asarray(pass_for_numeric_dtypes_or_raise_array(durations)\n ).reshape((n,))\n if event_observed is None:\n event_observed = np.ones(n, dtype=int)\n else:\n event_observed = np.asarray(event_observed).reshape((n,)).copy(\n ).astype(int)\n if entry is not None:\n entry = np.asarray(entry).reshape((n,))\n event_table = survival_table_from_events(durations, event_observed,\n entry, weights=weights)\n if timeline is None:\n timeline = event_table.index.values\n else:\n timeline = np.asarray(timeline)\n return durations, event_observed, timeline.astype(float\n ), entry, event_table\n", "role": "user" }, { "content": "preprocess_inputs", "role": "assistant" } ]
def x(x): try: if isinstance(x, (pd.Series, pd.DataFrame)): v = pd.to_numeric(x.squeeze()) else: v = pd.to_numeric(np.asarray(x).squeeze()) if v.size == 0: raise ValueError('Empty array/Series passed in.') return v except: raise ValueError( 'Values must be numeric: no strings, datetimes, objects, etc.')
pass_for_numeric_dtypes_or_raise_array
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(x):\n try:\n if isinstance(x, (pd.Series, pd.DataFrame)):\n v = pd.to_numeric(x.squeeze())\n else:\n v = pd.to_numeric(np.asarray(x).squeeze())\n if v.size == 0:\n raise ValueError('Empty array/Series passed in.')\n return v\n except:\n raise ValueError(\n 'Values must be numeric: no strings, datetimes, objects, etc.')\n", "role": "user" }, { "content": "pass_for_numeric_dtypes_or_raise_array", "role": "assistant" } ]
def x(df, duration_col, event_col, id_col=None, time_gaps=1): df = df.copy() df[duration_col] /= time_gaps df = to_long_format(df, duration_col) stop_col = 'stop' start_col = 'start' _, d = df.shape if id_col is None: id_col = 'id' df.index.rename(id_col, inplace=True) df = df.reset_index() d_dftv = d + 1 else: d_dftv = d dtype_dftv = object if (df.dtypes == object).any() else float n_dftv = int(np.ceil(df[stop_col]).sum()) tv_array = np.empty((n_dftv, d_dftv), dtype=dtype_dftv) special_columns = [stop_col, start_col, event_col] non_special_columns = df.columns.difference(special_columns).tolist() order_I_want = special_columns + non_special_columns df = df[order_I_want] position_counter = 0 for _, row in df.iterrows(): T, E = row[stop_col], row[event_col] T_int = int(np.ceil(T)) values = np.tile(row.values, (T_int, 1)) values[:, 0] = np.arange(1, T + 1, dtype=float) values[-1, 0] = T values[:, 1] = np.arange(0, T, dtype=float) values[:, 2] = 0.0 values[-1, 2] = float(E) tv_array[position_counter:position_counter + T_int, :] = values position_counter += T_int dftv = pd.DataFrame(tv_array, columns=df.columns) dftv = dftv.astype(dtype=df.dtypes[non_special_columns + [event_col]]. to_dict()) dftv[start_col] *= time_gaps dftv[stop_col] *= time_gaps return dftv
to_episodic_format
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(df, duration_col, event_col, id_col=None, time_gaps=1):\n df = df.copy()\n df[duration_col] /= time_gaps\n df = to_long_format(df, duration_col)\n stop_col = 'stop'\n start_col = 'start'\n _, d = df.shape\n if id_col is None:\n id_col = 'id'\n df.index.rename(id_col, inplace=True)\n df = df.reset_index()\n d_dftv = d + 1\n else:\n d_dftv = d\n dtype_dftv = object if (df.dtypes == object).any() else float\n n_dftv = int(np.ceil(df[stop_col]).sum())\n tv_array = np.empty((n_dftv, d_dftv), dtype=dtype_dftv)\n special_columns = [stop_col, start_col, event_col]\n non_special_columns = df.columns.difference(special_columns).tolist()\n order_I_want = special_columns + non_special_columns\n df = df[order_I_want]\n position_counter = 0\n for _, row in df.iterrows():\n T, E = row[stop_col], row[event_col]\n T_int = int(np.ceil(T))\n values = np.tile(row.values, (T_int, 1))\n values[:, 0] = np.arange(1, T + 1, dtype=float)\n values[-1, 0] = T\n values[:, 1] = np.arange(0, T, dtype=float)\n values[:, 2] = 0.0\n values[-1, 2] = float(E)\n tv_array[position_counter:position_counter + T_int, :] = values\n position_counter += T_int\n dftv = pd.DataFrame(tv_array, columns=df.columns)\n dftv = dftv.astype(dtype=df.dtypes[non_special_columns + [event_col]].\n to_dict())\n dftv[start_col] *= time_gaps\n dftv[stop_col] *= time_gaps\n return dftv\n", "role": "user" }, { "content": "to_episodic_format", "role": "assistant" } ]
def x(df, duration_col): return df.assign(start=0, stop=lambda s: s[duration_col]).drop(duration_col , axis=1)
to_long_format
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(df, duration_col):\n return df.assign(start=0, stop=lambda s: s[duration_col]).drop(duration_col\n , axis=1)\n", "role": "user" }, { "content": "to_long_format", "role": "assistant" } ]
def x(long_form_df, cv, id_col, duration_col, event_col, start_col='start', stop_col='stop', add_enum=False, overwrite=True, cumulative_sum=False, cumulative_sum_prefix='cumsum_', delay=0): def x(cv): if cumulative_sum: return cv cols = cv.columns.difference([duration_col]) cv = cv.loc[(cv[cols].shift() != cv[cols]).any(axis=1)] return cv def x(cv): return cv.rename(columns={duration_col: start_col}) def x(df, cvs): id_ = df.name try: cv = cvs.get_group(id_) except KeyError: return df final_state = bool(df[event_col].iloc[-1]) final_stop_time = df[stop_col].iloc[-1] df = df.drop([id_col, event_col, stop_col], axis=1).set_index(start_col ) cv = cv.drop([id_col], axis=1).set_index(start_col).loc[: final_stop_time] if cumulative_sum: cv = cv.cumsum() cv = cv.add_prefix(cumulative_sum_prefix) if not overwrite: expanded_df = cv.combine(df, lambda s1, s2: s1 + s2, fill_value =0, overwrite=False) elif overwrite: expanded_df = cv.combine_first(df) n = expanded_df.shape[0] expanded_df = expanded_df.reset_index() expanded_df[stop_col] = expanded_df[start_col].shift(-1) expanded_df[id_col] = id_ expanded_df[event_col] = False expanded_df.at[n - 1, event_col] = final_state expanded_df.at[n - 1, stop_col] = final_stop_time if add_enum: expanded_df['enum'] = np.arange(1, n + 1) if cumulative_sum: expanded_df[cv.columns] = expanded_df[cv.columns].ffill().fillna(0) return expanded_df.ffill() if delay < 0: raise ValueError('delay parameter must be equal to or greater than 0') if any(col not in long_form_df for col in (id_col, event_col, start_col, stop_col)): raise IndexError('Missing column in long_form_df') cv[duration_col] += delay cv = cv.dropna() cv = cv.sort_values([id_col, duration_col]) cvs = cv.pipe(remove_redundant_rows).pipe(transform_cv_to_long_format ).groupby(id_col) long_form_df = long_form_df.groupby(id_col, group_keys=False).apply(expand, cvs=cvs) return long_form_df.reset_index(drop=True)
add_covariate_to_timeline
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(long_form_df, cv, id_col, duration_col, event_col, start_col='start',\n stop_col='stop', add_enum=False, overwrite=True, cumulative_sum=False,\n cumulative_sum_prefix='cumsum_', delay=0):\n\n def x(cv):\n if cumulative_sum:\n return cv\n cols = cv.columns.difference([duration_col])\n cv = cv.loc[(cv[cols].shift() != cv[cols]).any(axis=1)]\n return cv\n\n def x(cv):\n return cv.rename(columns={duration_col: start_col})\n\n def x(df, cvs):\n id_ = df.name\n try:\n cv = cvs.get_group(id_)\n except KeyError:\n return df\n final_state = bool(df[event_col].iloc[-1])\n final_stop_time = df[stop_col].iloc[-1]\n df = df.drop([id_col, event_col, stop_col], axis=1).set_index(start_col\n )\n cv = cv.drop([id_col], axis=1).set_index(start_col).loc[:\n final_stop_time]\n if cumulative_sum:\n cv = cv.cumsum()\n cv = cv.add_prefix(cumulative_sum_prefix)\n if not overwrite:\n expanded_df = cv.combine(df, lambda s1, s2: s1 + s2, fill_value\n =0, overwrite=False)\n elif overwrite:\n expanded_df = cv.combine_first(df)\n n = expanded_df.shape[0]\n expanded_df = expanded_df.reset_index()\n expanded_df[stop_col] = expanded_df[start_col].shift(-1)\n expanded_df[id_col] = id_\n expanded_df[event_col] = False\n expanded_df.at[n - 1, event_col] = final_state\n expanded_df.at[n - 1, stop_col] = final_stop_time\n if add_enum:\n expanded_df['enum'] = np.arange(1, n + 1)\n if cumulative_sum:\n expanded_df[cv.columns] = expanded_df[cv.columns].ffill().fillna(0)\n return expanded_df.ffill()\n if delay < 0:\n raise ValueError('delay parameter must be equal to or greater than 0')\n if any(col not in long_form_df for col in (id_col, event_col, start_col,\n stop_col)):\n raise IndexError('Missing column in long_form_df')\n cv[duration_col] += delay\n cv = cv.dropna()\n cv = cv.sort_values([id_col, duration_col])\n cvs = cv.pipe(remove_redundant_rows).pipe(transform_cv_to_long_format\n ).groupby(id_col)\n long_form_df = long_form_df.groupby(id_col, group_keys=False).apply(expand,\n cvs=cvs)\n return long_form_df.reset_index(drop=True)\n", "role": "user" }, { "content": "add_covariate_to_timeline", "role": "assistant" } ]
def x(df, id_col): df = df.set_index(id_col) df = df.stack().reset_index() df.columns = [id_col, 'event', 'duration'] df['_counter'] = 1 return df.pivot_table(index=[id_col, 'duration'], columns='event', fill_value=0)['_counter'].reset_index()
covariates_from_event_matrix
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(df, id_col):\n df = df.set_index(id_col)\n df = df.stack().reset_index()\n df.columns = [id_col, 'event', 'duration']\n df['_counter'] = 1\n return df.pivot_table(index=[id_col, 'duration'], columns='event',\n fill_value=0)['_counter'].reset_index()\n", "role": "user" }, { "content": "covariates_from_event_matrix", "role": "assistant" } ]
def x(decimals): threshold = 10 ** 5 return lambda n: '{:.{prec}e}'.format(n, prec=decimals ) if n > threshold else '{:4.{prec}f}'.format(n, prec=decimals)
format_exp_floats
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(decimals):\n threshold = 10 ** 5\n return lambda n: '{:.{prec}e}'.format(n, prec=decimals\n ) if n > threshold else '{:4.{prec}f}'.format(n, prec=decimals)\n", "role": "user" }, { "content": "format_exp_floats", "role": "assistant" } ]
def x(self, durations, event_observed=None, timeline=None, entry=None, label='BFH_estimate', alpha=None, ci_labels=None): self._censoring_type = CensoringType.RIGHT self._label = label alpha = coalesce(alpha, self.alpha) naf = NelsonAalenFitter(alpha=alpha) naf.fit(durations, event_observed=event_observed, timeline=timeline, label=label, entry=entry, ci_labels=ci_labels) (self.durations, self.event_observed, self.timeline, self.entry, self. event_table) = (naf.durations, naf.event_observed, naf.timeline, naf.entry, naf.event_table) self.survival_function_ = np.exp(-naf.cumulative_hazard_) self.confidence_interval_ = np.exp(-naf.confidence_interval_) self._estimation_method = 'survival_function_' self._estimate_name = 'survival_function_' self._predict_label = label self._update_docstrings() self.plot_survival_function = self.plot return self
fit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, durations, event_observed=None, timeline=None, entry=None,\n label='BFH_estimate', alpha=None, ci_labels=None):\n self._censoring_type = CensoringType.RIGHT\n self._label = label\n alpha = coalesce(alpha, self.alpha)\n naf = NelsonAalenFitter(alpha=alpha)\n naf.fit(durations, event_observed=event_observed, timeline=timeline,\n label=label, entry=entry, ci_labels=ci_labels)\n (self.durations, self.event_observed, self.timeline, self.entry, self.\n event_table) = (naf.durations, naf.event_observed, naf.timeline,\n naf.entry, naf.event_table)\n self.survival_function_ = np.exp(-naf.cumulative_hazard_)\n self.confidence_interval_ = np.exp(-naf.confidence_interval_)\n self._estimation_method = 'survival_function_'\n self._estimate_name = 'survival_function_'\n self._predict_label = label\n self._update_docstrings()\n self.plot_survival_function = self.plot\n return self\n", "role": "user" }, { "content": "fit", "role": "assistant" } ]
def x(self, durations, event_observed=None, timeline=None, entry=None, label='KM_estimate', left_censorship=False, alpha=None, ci_labels=None, weights=None): if left_censorship: warnings.warn( 'kwarg left_censorship is deprecated and will be removed in a future release. Please use ``.fit_left_censoring`` instead.' , DeprecationWarning) self._censoring_type = CensoringType.RIGHT return self._fit(durations, event_observed, timeline, entry, label, alpha, ci_labels, weights)
fit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, durations, event_observed=None, timeline=None, entry=None,\n label='KM_estimate', left_censorship=False, alpha=None, ci_labels=None,\n weights=None):\n if left_censorship:\n warnings.warn(\n 'kwarg left_censorship is deprecated and will be removed in a future release. Please use ``.fit_left_censoring`` instead.'\n , DeprecationWarning)\n self._censoring_type = CensoringType.RIGHT\n return self._fit(durations, event_observed, timeline, entry, label,\n alpha, ci_labels, weights)\n", "role": "user" }, { "content": "fit", "role": "assistant" } ]
def x(self, durations, event_observed=None, timeline=None, entry=None, label='KM_estimate', alpha=None, ci_labels=None, weights=None): self._censoring_type = CensoringType.LEFT return self._fit(durations, event_observed, timeline, entry, label, alpha, ci_labels, weights)
fit_left_censoring
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, durations, event_observed=None, timeline=None, entry=None,\n label='KM_estimate', alpha=None, ci_labels=None, weights=None):\n self._censoring_type = CensoringType.LEFT\n return self._fit(durations, event_observed, timeline, entry, label,\n alpha, ci_labels, weights)\n", "role": "user" }, { "content": "fit_left_censoring", "role": "assistant" } ]
def x(self, durations, event_observed=None, timeline=None, entry=None, label='KM_estimate', alpha=None, ci_labels=None, weights=None): self._check_values(durations) if event_observed is not None: self._check_values(event_observed) self._label = label if weights is not None: weights = np.asarray(weights) if (weights.astype(int) != weights).any(): warnings.warn( """It looks like your weights are not integers, possibly propensity scores then? It's important to know that the naive variance estimates of the coefficients are biased. Instead use Monte Carlo to estimate the variances. See paper "Variance estimation when using inverse probability of treatment weighting (IPTW) with survival analysis" or "Adjusted Kaplan-Meier estimator and log-rank test with inverse probability of treatment weighting for survival data." """ , StatisticalWarning) is_left_censoring = self._censoring_type == CensoringType.LEFT primary_estimate_name = ('survival_function_' if not is_left_censoring else 'cumulative_density_') secondary_estimate_name = ('cumulative_density_' if not is_left_censoring else 'survival_function_') (self.durations, self.event_observed, self.timeline, self.entry, self. event_table) = (_preprocess_inputs(durations, event_observed, timeline, entry, weights)) alpha = alpha if alpha else self.alpha log_estimate, cumulative_sq_ = _additive_estimate(self.event_table, self.timeline, self._additive_f, self._additive_var, is_left_censoring) if entry is not None: n = self.event_table.shape[0] net_population = (self.event_table['entrance'] - self.event_table[ 'removed']).cumsum() if net_population.iloc[:int(n / 2)].min() == 0: ix = net_population.iloc[:int(n / 2)].idxmin() raise StatError( 'There are too few early truncation times and too many events. S(t)==0 for all t>%g. Recommend BreslowFlemingHarringtonFitter.' % ix) setattr(self, primary_estimate_name, pd.DataFrame(np.exp(log_estimate), columns=[self._label])) setattr(self, secondary_estimate_name, pd.DataFrame(1 - np.exp( log_estimate), columns=[self._label])) self.__estimate = getattr(self, primary_estimate_name) self.confidence_interval_ = self._bounds(cumulative_sq_[:, None], alpha, ci_labels) self.median_ = median_survival_times(self.__estimate, left_censorship= is_left_censoring) self._cumulative_sq_ = cumulative_sq_ setattr(self, 'confidence_interval_' + primary_estimate_name, self. confidence_interval_) setattr(self, 'confidence_interval_' + secondary_estimate_name, 1 - self.confidence_interval_) self._estimation_method = primary_estimate_name self._estimate_name = primary_estimate_name self._predict_label = label self._update_docstrings() return self
fit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, durations, event_observed=None, timeline=None, entry=None,\n label='KM_estimate', alpha=None, ci_labels=None, weights=None):\n self._check_values(durations)\n if event_observed is not None:\n self._check_values(event_observed)\n self._label = label\n if weights is not None:\n weights = np.asarray(weights)\n if (weights.astype(int) != weights).any():\n warnings.warn(\n \"\"\"It looks like your weights are not integers, possibly propensity scores then?\n It's important to know that the naive variance estimates of the coefficients are biased. Instead use Monte Carlo to\n estimate the variances. See paper \"Variance estimation when using inverse probability of treatment weighting (IPTW) with survival analysis\"\n or \"Adjusted Kaplan-Meier estimator and log-rank test with inverse probability of treatment weighting for survival data.\"\n \"\"\"\n , StatisticalWarning)\n is_left_censoring = self._censoring_type == CensoringType.LEFT\n primary_estimate_name = ('survival_function_' if not is_left_censoring else\n 'cumulative_density_')\n secondary_estimate_name = ('cumulative_density_' if not\n is_left_censoring else 'survival_function_')\n (self.durations, self.event_observed, self.timeline, self.entry, self.\n event_table) = (_preprocess_inputs(durations, event_observed,\n timeline, entry, weights))\n alpha = alpha if alpha else self.alpha\n log_estimate, cumulative_sq_ = _additive_estimate(self.event_table,\n self.timeline, self._additive_f, self._additive_var, is_left_censoring)\n if entry is not None:\n n = self.event_table.shape[0]\n net_population = (self.event_table['entrance'] - self.event_table[\n 'removed']).cumsum()\n if net_population.iloc[:int(n / 2)].min() == 0:\n ix = net_population.iloc[:int(n / 2)].idxmin()\n raise StatError(\n 'There are too few early truncation times and too many events. S(t)==0 for all t>%g. Recommend BreslowFlemingHarringtonFitter.'\n % ix)\n setattr(self, primary_estimate_name, pd.DataFrame(np.exp(log_estimate),\n columns=[self._label]))\n setattr(self, secondary_estimate_name, pd.DataFrame(1 - np.exp(\n log_estimate), columns=[self._label]))\n self.__estimate = getattr(self, primary_estimate_name)\n self.confidence_interval_ = self._bounds(cumulative_sq_[:, None], alpha,\n ci_labels)\n self.median_ = median_survival_times(self.__estimate, left_censorship=\n is_left_censoring)\n self._cumulative_sq_ = cumulative_sq_\n setattr(self, 'confidence_interval_' + primary_estimate_name, self.\n confidence_interval_)\n setattr(self, 'confidence_interval_' + secondary_estimate_name, 1 -\n self.confidence_interval_)\n self._estimation_method = primary_estimate_name\n self._estimate_name = primary_estimate_name\n self._predict_label = label\n self._update_docstrings()\n return self\n", "role": "user" }, { "content": "fit", "role": "assistant" } ]
def x(self, times, label=None): label = coalesce(label, self._label) return pd.Series(1 - self.predict(times), index=_to_array(times), name= label)
cumulative_density_at_times
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, times, label=None):\n label = coalesce(label, self._label)\n return pd.Series(1 - self.predict(times), index=_to_array(times), name=\n label)\n", "role": "user" }, { "content": "cumulative_density_at_times", "role": "assistant" } ]
def x(self, **kwargs): return _plot_estimate(self, estimate=self.survival_function_, confidence_intervals=self.confidence_interval_survival_function_, **kwargs)
plot_survival_function
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n return _plot_estimate(self, estimate=self.survival_function_,\n confidence_intervals=self.confidence_interval_survival_function_,\n **kwargs)\n", "role": "user" }, { "content": "plot_survival_function", "role": "assistant" } ]
def x(self, **kwargs): return _plot_estimate(self, estimate=self.cumulative_density_, confidence_intervals=self.confidence_interval_cumulative_density_, **kwargs)
plot_cumulative_density
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n return _plot_estimate(self, estimate=self.cumulative_density_,\n confidence_intervals=self.confidence_interval_cumulative_density_,\n **kwargs)\n", "role": "user" }, { "content": "plot_cumulative_density", "role": "assistant" } ]
def x(self, df, id_col, event_col, start_col='start', stop_col='stop', weights_col=None, show_progress=False, step_size=None, robust=False, strata=None, initial_point=None): self.strata = coalesce(strata, self.strata) self.robust = robust if self.robust: raise NotImplementedError('Not available yet.') self.event_col = event_col self.id_col = id_col self.stop_col = stop_col self.start_col = start_col self._time_fit_was_called = datetime.utcnow().strftime('%Y-%m-%d %H:%M:%S') df = df.copy() if not (id_col in df and event_col in df and start_col in df and stop_col in df): raise KeyError( 'A column specified in the call to `fit` does not exist in the DataFrame provided.' ) if weights_col is None: self.weights_col = None assert '__weights' not in df.columns, '__weights is an internal lifelines column, please rename your column first.' df['__weights'] = 1.0 else: self.weights_col = weights_col if (df[weights_col] <= 0).any(): raise ValueError('values in weights_col must be positive.') df = df.rename(columns={id_col: 'id', event_col: 'event', start_col: 'start', stop_col: 'stop', weights_col: '__weights'}) if self.strata is None: df = df.set_index('id') else: df = df.set_index(_to_list(self.strata) + ['id']) df = df.sort_index() events, start, stop = pass_for_numeric_dtypes_or_raise_array(df.pop( 'event')).astype(bool), df.pop('start'), df.pop('stop') weights = df.pop('__weights').astype(float) df = df.astype(float) self._check_values(df, events, start, stop) self._norm_mean = df.mean(0) self._norm_std = df.std(0) hazards_ = self._newton_rhaphson(normalize(df, self._norm_mean, self. _norm_std), events, start, stop, weights, initial_point= initial_point, show_progress=show_progress, step_size=step_size) self.hazards_ = pd.Series(hazards_, index=df.columns, name='coef' ) / self._norm_std self.variance_matrix_ = -inv(self._hessian_) / np.outer(self._norm_std, self._norm_std) self.standard_errors_ = self._compute_standard_errors(normalize(df, self._norm_mean, self._norm_std), events, start, stop, weights) self.confidence_intervals_ = self._compute_confidence_intervals() self.baseline_cumulative_hazard_ = (self. _compute_cumulative_baseline_hazard(df, events, start, stop, weights)) self.baseline_survival_ = self._compute_baseline_survival() self.event_observed = events self.start_stop_and_events = pd.DataFrame({'event': events, 'start': start, 'stop': stop}) self.weights = weights self._n_examples = df.shape[0] self._n_unique = df.index.unique().shape[0] return self
fit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, df, id_col, event_col, start_col='start', stop_col='stop',\n weights_col=None, show_progress=False, step_size=None, robust=False,\n strata=None, initial_point=None):\n self.strata = coalesce(strata, self.strata)\n self.robust = robust\n if self.robust:\n raise NotImplementedError('Not available yet.')\n self.event_col = event_col\n self.id_col = id_col\n self.stop_col = stop_col\n self.start_col = start_col\n self._time_fit_was_called = datetime.utcnow().strftime('%Y-%m-%d %H:%M:%S')\n df = df.copy()\n if not (id_col in df and event_col in df and start_col in df and \n stop_col in df):\n raise KeyError(\n 'A column specified in the call to `fit` does not exist in the DataFrame provided.'\n )\n if weights_col is None:\n self.weights_col = None\n assert '__weights' not in df.columns, '__weights is an internal lifelines column, please rename your column first.'\n df['__weights'] = 1.0\n else:\n self.weights_col = weights_col\n if (df[weights_col] <= 0).any():\n raise ValueError('values in weights_col must be positive.')\n df = df.rename(columns={id_col: 'id', event_col: 'event', start_col:\n 'start', stop_col: 'stop', weights_col: '__weights'})\n if self.strata is None:\n df = df.set_index('id')\n else:\n df = df.set_index(_to_list(self.strata) + ['id'])\n df = df.sort_index()\n events, start, stop = pass_for_numeric_dtypes_or_raise_array(df.pop(\n 'event')).astype(bool), df.pop('start'), df.pop('stop')\n weights = df.pop('__weights').astype(float)\n df = df.astype(float)\n self._check_values(df, events, start, stop)\n self._norm_mean = df.mean(0)\n self._norm_std = df.std(0)\n hazards_ = self._newton_rhaphson(normalize(df, self._norm_mean, self.\n _norm_std), events, start, stop, weights, initial_point=\n initial_point, show_progress=show_progress, step_size=step_size)\n self.hazards_ = pd.Series(hazards_, index=df.columns, name='coef'\n ) / self._norm_std\n self.variance_matrix_ = -inv(self._hessian_) / np.outer(self._norm_std,\n self._norm_std)\n self.standard_errors_ = self._compute_standard_errors(normalize(df,\n self._norm_mean, self._norm_std), events, start, stop, weights)\n self.confidence_intervals_ = self._compute_confidence_intervals()\n self.baseline_cumulative_hazard_ = (self.\n _compute_cumulative_baseline_hazard(df, events, start, stop, weights))\n self.baseline_survival_ = self._compute_baseline_survival()\n self.event_observed = events\n self.start_stop_and_events = pd.DataFrame({'event': events, 'start':\n start, 'stop': stop})\n self.weights = weights\n self._n_examples = df.shape[0]\n self._n_unique = df.index.unique().shape[0]\n return self\n", "role": "user" }, { "content": "fit", "role": "assistant" } ]
def x(self, df, events, start, stop, weights, show_progress=False, step_size=None, precision=1e-05, max_steps=50, initial_point=None): assert precision <= 1.0, 'precision must be less than or equal to 1.' _, d = df.shape if initial_point is not None: beta = initial_point else: beta = np.zeros((d,)) i = 0 converging = True ll, previous_ll = 0, 0 start_time = time.time() step_sizer = StepSizer(step_size) step_size = step_sizer.next() while converging: i += 1 if self.strata is None: h, g, ll = self._get_gradients(df.values, events.values, start. values, stop.values, weights.values, beta) else: g = np.zeros_like(beta) h = np.zeros((d, d)) ll = 0 for _h, _g, _ll in self._partition_by_strata_and_apply(df, events, start, stop, weights, self._get_gradients, beta): g += _g h += _h ll += _ll if i == 1 and np.all(beta == 0): self._log_likelihood_null = ll if self.penalizer > 0: g -= self.penalizer * beta h.flat[::d + 1] -= self.penalizer try: inv_h_dot_g_T = spsolve(-h, g, sym_pos=True) except ValueError as e: if 'infs or NaNs' in str(e): raise ConvergenceError( """hessian or gradient contains nan or inf value(s). Convergence halted. Please see the following tips in the lifelines documentation: https://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model """ , e) else: raise e except LinAlgError as e: raise ConvergenceError( """Convergence halted due to matrix inversion problems. Suspicion is high colinearity. Please see the following tips in the lifelines documentation: https://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model """ , e) delta = step_size * inv_h_dot_g_T if np.any(np.isnan(delta)): raise ConvergenceError( """delta contains nan value(s). Convergence halted. Please see the following tips in the lifelines documentation: https://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model """ ) hessian, gradient = h, g norm_delta = norm(delta) newton_decrement = g.dot(inv_h_dot_g_T) / 2 if show_progress: print( 'Iteration %d: norm_delta = %.5f, step_size = %.5f, ll = %.5f, newton_decrement = %.5f, seconds_since_start = %.1f' % (i, norm_delta, step_size, ll, newton_decrement, time. time() - start_time)) if norm_delta < precision: converging, completed = False, True elif previous_ll > 0 and abs(ll - previous_ll) / -previous_ll < 1e-09: converging, completed = False, True elif newton_decrement < 1e-07: converging, completed = False, True elif i >= max_steps: converging, completed = False, False elif step_size <= 0.0001: converging, completed = False, False elif abs(ll) < 0.0001 and norm_delta > 1.0: warnings.warn( 'The log-likelihood is getting suspiciously close to 0 and the delta is still large. There may be complete separation in the dataset. This may result in incorrect inference of coefficients. See https://stats.stackexchange.com/questions/11109/how-to-deal-with-perfect-separation-in-logistic-regression' , ConvergenceWarning) converging, completed = False, False step_size = step_sizer.update(norm_delta).next() beta += delta self._hessian_ = hessian self._score_ = gradient self._log_likelihood = ll if show_progress and completed: print('Convergence completed after %d iterations.' % i) elif show_progress and not completed: print('Convergence failed. See any warning messages.') if completed and norm_delta > 0.1: warnings.warn( 'Newton-Rhapson convergence completed but norm(delta) is still high, %.3f. This may imply non-unique solutions to the maximum likelihood. Perhaps there is colinearity or complete separation in the dataset?' % norm_delta, ConvergenceWarning) elif not completed: warnings.warn( 'Newton-Rhapson failed to converge sufficiently in %d steps.' % max_steps, ConvergenceWarning) return beta
newton_rhaphson
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, df, events, start, stop, weights, show_progress=False,\n step_size=None, precision=1e-05, max_steps=50, initial_point=None):\n assert precision <= 1.0, 'precision must be less than or equal to 1.'\n _, d = df.shape\n if initial_point is not None:\n beta = initial_point\n else:\n beta = np.zeros((d,))\n i = 0\n converging = True\n ll, previous_ll = 0, 0\n start_time = time.time()\n step_sizer = StepSizer(step_size)\n step_size = step_sizer.next()\n while converging:\n i += 1\n if self.strata is None:\n h, g, ll = self._get_gradients(df.values, events.values, start.\n values, stop.values, weights.values, beta)\n else:\n g = np.zeros_like(beta)\n h = np.zeros((d, d))\n ll = 0\n for _h, _g, _ll in self._partition_by_strata_and_apply(df,\n events, start, stop, weights, self._get_gradients, beta):\n g += _g\n h += _h\n ll += _ll\n if i == 1 and np.all(beta == 0):\n self._log_likelihood_null = ll\n if self.penalizer > 0:\n g -= self.penalizer * beta\n h.flat[::d + 1] -= self.penalizer\n try:\n inv_h_dot_g_T = spsolve(-h, g, sym_pos=True)\n except ValueError as e:\n if 'infs or NaNs' in str(e):\n raise ConvergenceError(\n \"\"\"hessian or gradient contains nan or inf value(s). Convergence halted. Please see the following tips in the lifelines documentation:\nhttps://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model\n\"\"\"\n , e)\n else:\n raise e\n except LinAlgError as e:\n raise ConvergenceError(\n \"\"\"Convergence halted due to matrix inversion problems. Suspicion is high colinearity. Please see the following tips in the lifelines documentation:\nhttps://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model\n\"\"\"\n , e)\n delta = step_size * inv_h_dot_g_T\n if np.any(np.isnan(delta)):\n raise ConvergenceError(\n \"\"\"delta contains nan value(s). Convergence halted. Please see the following tips in the lifelines documentation:\nhttps://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model\n\"\"\"\n )\n hessian, gradient = h, g\n norm_delta = norm(delta)\n newton_decrement = g.dot(inv_h_dot_g_T) / 2\n if show_progress:\n print(\n 'Iteration %d: norm_delta = %.5f, step_size = %.5f, ll = %.5f, newton_decrement = %.5f, seconds_since_start = %.1f'\n % (i, norm_delta, step_size, ll, newton_decrement, time.\n time() - start_time))\n if norm_delta < precision:\n converging, completed = False, True\n elif previous_ll > 0 and abs(ll - previous_ll) / -previous_ll < 1e-09:\n converging, completed = False, True\n elif newton_decrement < 1e-07:\n converging, completed = False, True\n elif i >= max_steps:\n converging, completed = False, False\n elif step_size <= 0.0001:\n converging, completed = False, False\n elif abs(ll) < 0.0001 and norm_delta > 1.0:\n warnings.warn(\n 'The log-likelihood is getting suspiciously close to 0 and the delta is still large. There may be complete separation in the dataset. This may result in incorrect inference of coefficients. See https://stats.stackexchange.com/questions/11109/how-to-deal-with-perfect-separation-in-logistic-regression'\n , ConvergenceWarning)\n converging, completed = False, False\n step_size = step_sizer.update(norm_delta).next()\n beta += delta\n self._hessian_ = hessian\n self._score_ = gradient\n self._log_likelihood = ll\n if show_progress and completed:\n print('Convergence completed after %d iterations.' % i)\n elif show_progress and not completed:\n print('Convergence failed. See any warning messages.')\n if completed and norm_delta > 0.1:\n warnings.warn(\n 'Newton-Rhapson convergence completed but norm(delta) is still high, %.3f. This may imply non-unique solutions to the maximum likelihood. Perhaps there is colinearity or complete separation in the dataset?'\n % norm_delta, ConvergenceWarning)\n elif not completed:\n warnings.warn(\n 'Newton-Rhapson failed to converge sufficiently in %d steps.' %\n max_steps, ConvergenceWarning)\n return beta\n", "role": "user" }, { "content": "newton_rhaphson", "role": "assistant" } ]
def x(self, X, events, start, stop, weights, beta): _, d = X.shape hessian = np.zeros((d, d)) gradient = np.zeros(d) log_lik = 0 unique_death_times = np.unique(stop[events]) for t in unique_death_times: ix = (start < t) & (t <= stop) X_at_t = X[ix] weights_at_t = weights[ix] stops_events_at_t = stop[ix] events_at_t = events[ix] phi_i = weights_at_t * np.exp(np.dot(X_at_t, beta)) phi_x_i = phi_i[:, None] * X_at_t phi_x_x_i = np.dot(X_at_t.T, phi_x_i) risk_phi = array_sum_to_scalar(phi_i) risk_phi_x = matrix_axis_0_sum_to_array(phi_x_i) risk_phi_x_x = phi_x_x_i deaths = events_at_t & (stops_events_at_t == t) tied_death_counts = array_sum_to_scalar(deaths.astype(int)) xi_deaths = X_at_t[deaths] x_death_sum = matrix_axis_0_sum_to_array(weights_at_t[deaths, None] * xi_deaths) weight_count = array_sum_to_scalar(weights_at_t[deaths]) weighted_average = weight_count / tied_death_counts if tied_death_counts > 1: tie_phi = array_sum_to_scalar(phi_i[deaths]) tie_phi_x = matrix_axis_0_sum_to_array(phi_x_i[deaths]) tie_phi_x_x = np.dot(xi_deaths.T, phi_i[deaths, None] * xi_deaths) increasing_proportion = np.arange(tied_death_counts ) / tied_death_counts denom = 1.0 / (risk_phi - increasing_proportion * tie_phi) numer = risk_phi_x - np.outer(increasing_proportion, tie_phi_x) a1 = np.einsum('ab, i->ab', risk_phi_x_x, denom) - np.einsum( 'ab, i->ab', tie_phi_x_x, increasing_proportion * denom) else: denom = 1.0 / np.array([risk_phi]) numer = risk_phi_x a1 = risk_phi_x_x * denom summand = numer * denom[:, None] a2 = summand.T.dot(summand) gradient = gradient + x_death_sum - weighted_average * summand.sum(0) log_lik = log_lik + np.dot(x_death_sum, beta ) + weighted_average * np.log(denom).sum() hessian = hessian + weighted_average * (a2 - a1) return hessian, gradient, log_lik
get_gradients
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, events, start, stop, weights, beta):\n _, d = X.shape\n hessian = np.zeros((d, d))\n gradient = np.zeros(d)\n log_lik = 0\n unique_death_times = np.unique(stop[events])\n for t in unique_death_times:\n ix = (start < t) & (t <= stop)\n X_at_t = X[ix]\n weights_at_t = weights[ix]\n stops_events_at_t = stop[ix]\n events_at_t = events[ix]\n phi_i = weights_at_t * np.exp(np.dot(X_at_t, beta))\n phi_x_i = phi_i[:, None] * X_at_t\n phi_x_x_i = np.dot(X_at_t.T, phi_x_i)\n risk_phi = array_sum_to_scalar(phi_i)\n risk_phi_x = matrix_axis_0_sum_to_array(phi_x_i)\n risk_phi_x_x = phi_x_x_i\n deaths = events_at_t & (stops_events_at_t == t)\n tied_death_counts = array_sum_to_scalar(deaths.astype(int))\n xi_deaths = X_at_t[deaths]\n x_death_sum = matrix_axis_0_sum_to_array(weights_at_t[deaths, None] *\n xi_deaths)\n weight_count = array_sum_to_scalar(weights_at_t[deaths])\n weighted_average = weight_count / tied_death_counts\n if tied_death_counts > 1:\n tie_phi = array_sum_to_scalar(phi_i[deaths])\n tie_phi_x = matrix_axis_0_sum_to_array(phi_x_i[deaths])\n tie_phi_x_x = np.dot(xi_deaths.T, phi_i[deaths, None] * xi_deaths)\n increasing_proportion = np.arange(tied_death_counts\n ) / tied_death_counts\n denom = 1.0 / (risk_phi - increasing_proportion * tie_phi)\n numer = risk_phi_x - np.outer(increasing_proportion, tie_phi_x)\n a1 = np.einsum('ab, i->ab', risk_phi_x_x, denom) - np.einsum(\n 'ab, i->ab', tie_phi_x_x, increasing_proportion * denom)\n else:\n denom = 1.0 / np.array([risk_phi])\n numer = risk_phi_x\n a1 = risk_phi_x_x * denom\n summand = numer * denom[:, None]\n a2 = summand.T.dot(summand)\n gradient = gradient + x_death_sum - weighted_average * summand.sum(0)\n log_lik = log_lik + np.dot(x_death_sum, beta\n ) + weighted_average * np.log(denom).sum()\n hessian = hessian + weighted_average * (a2 - a1)\n return hessian, gradient, log_lik\n", "role": "user" }, { "content": "get_gradients", "role": "assistant" } ]
def x(self, X): if isinstance(X, pd.DataFrame): order = self.hazards_.index X = X[order] check_for_numeric_dtypes_or_raise(X) X = X.astype(float) index = _get_index(X) X = normalize(X, self._norm_mean.values, 1) return pd.DataFrame(np.dot(X, self.hazards_), index=index)
predict_log_partial_hazard
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X):\n if isinstance(X, pd.DataFrame):\n order = self.hazards_.index\n X = X[order]\n check_for_numeric_dtypes_or_raise(X)\n X = X.astype(float)\n index = _get_index(X)\n X = normalize(X, self._norm_mean.values, 1)\n return pd.DataFrame(np.dot(X, self.hazards_), index=index)\n", "role": "user" }, { "content": "predict_log_partial_hazard", "role": "assistant" } ]
def x(self, decimals=2, **kwargs): justify = string_justify(18) print(self) print("{} = '{}'".format(justify('event col'), self.event_col)) if self.weights_col: print("{} = '{}'".format(justify('weights col'), self.weights_col)) if self.strata: print('{} = {}'.format(justify('strata'), self.strata)) if self.penalizer > 0: print('{} = {}'.format(justify('penalizer'), self.penalizer)) print('{} = {}'.format(justify('number of subjects'), self._n_unique)) print('{} = {}'.format(justify('number of periods'), self._n_examples)) print('{} = {}'.format(justify('number of events'), self.event_observed .sum())) print('{} = {:.{prec}f}'.format(justify('log-likelihood'), self. _log_likelihood, prec=decimals)) print('{} = {} UTC'.format(justify('time fit was run'), self. _time_fit_was_called)) for k, v in kwargs.items(): print('{} = {}\n'.format(justify(k), v)) print(end='\n') print('---') df = self.summary print(df.to_string(float_format=format_floats(decimals), formatters={ 'p': format_p_value(decimals), 'exp(coef)': format_exp_floats( decimals)})) print('---') print( 'Log-likelihood ratio test = {:.{prec}f} on {} df, -log2(p)={:.{prec}f}' .format(*self._compute_likelihood_ratio_test(), prec=decimals))
print_summary
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, decimals=2, **kwargs):\n justify = string_justify(18)\n print(self)\n print(\"{} = '{}'\".format(justify('event col'), self.event_col))\n if self.weights_col:\n print(\"{} = '{}'\".format(justify('weights col'), self.weights_col))\n if self.strata:\n print('{} = {}'.format(justify('strata'), self.strata))\n if self.penalizer > 0:\n print('{} = {}'.format(justify('penalizer'), self.penalizer))\n print('{} = {}'.format(justify('number of subjects'), self._n_unique))\n print('{} = {}'.format(justify('number of periods'), self._n_examples))\n print('{} = {}'.format(justify('number of events'), self.event_observed\n .sum()))\n print('{} = {:.{prec}f}'.format(justify('log-likelihood'), self.\n _log_likelihood, prec=decimals))\n print('{} = {} UTC'.format(justify('time fit was run'), self.\n _time_fit_was_called))\n for k, v in kwargs.items():\n print('{} = {}\\n'.format(justify(k), v))\n print(end='\\n')\n print('---')\n df = self.summary\n print(df.to_string(float_format=format_floats(decimals), formatters={\n 'p': format_p_value(decimals), 'exp(coef)': format_exp_floats(\n decimals)}))\n print('---')\n print(\n 'Log-likelihood ratio test = {:.{prec}f} on {} df, -log2(p)={:.{prec}f}'\n .format(*self._compute_likelihood_ratio_test(), prec=decimals))\n", "role": "user" }, { "content": "print_summary", "role": "assistant" } ]
def x(self, columns=None, **errorbar_kwargs): from matplotlib import pyplot as plt ax = errorbar_kwargs.pop('ax', None) or plt.figure().add_subplot(111) errorbar_kwargs.setdefault('c', 'k') errorbar_kwargs.setdefault('fmt', 's') errorbar_kwargs.setdefault('markerfacecolor', 'white') errorbar_kwargs.setdefault('markeredgewidth', 1.25) errorbar_kwargs.setdefault('elinewidth', 1.25) errorbar_kwargs.setdefault('capsize', 3) z = inv_normal_cdf(1 - self.alpha / 2) if columns is None: columns = self.hazards_.index yaxis_locations = list(range(len(columns))) symmetric_errors = z * self.standard_errors_[columns].to_frame().squeeze( axis=1).values.copy() hazards = self.hazards_[columns].values.copy() order = np.argsort(hazards) ax.errorbar(hazards[order], yaxis_locations, xerr=symmetric_errors[ order], **errorbar_kwargs) best_ylim = ax.get_ylim() ax.vlines(0, -2, len(columns) + 1, linestyles='dashed', linewidths=1, alpha=0.65) ax.set_ylim(best_ylim) tick_labels = [columns[i] for i in order] ax.set_yticks(yaxis_locations) ax.set_yticklabels(tick_labels) ax.set_xlabel('log(HR) (%g%% CI)' % ((1 - self.alpha) * 100)) return ax
plot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, columns=None, **errorbar_kwargs):\n from matplotlib import pyplot as plt\n ax = errorbar_kwargs.pop('ax', None) or plt.figure().add_subplot(111)\n errorbar_kwargs.setdefault('c', 'k')\n errorbar_kwargs.setdefault('fmt', 's')\n errorbar_kwargs.setdefault('markerfacecolor', 'white')\n errorbar_kwargs.setdefault('markeredgewidth', 1.25)\n errorbar_kwargs.setdefault('elinewidth', 1.25)\n errorbar_kwargs.setdefault('capsize', 3)\n z = inv_normal_cdf(1 - self.alpha / 2)\n if columns is None:\n columns = self.hazards_.index\n yaxis_locations = list(range(len(columns)))\n symmetric_errors = z * self.standard_errors_[columns].to_frame().squeeze(\n axis=1).values.copy()\n hazards = self.hazards_[columns].values.copy()\n order = np.argsort(hazards)\n ax.errorbar(hazards[order], yaxis_locations, xerr=symmetric_errors[\n order], **errorbar_kwargs)\n best_ylim = ax.get_ylim()\n ax.vlines(0, -2, len(columns) + 1, linestyles='dashed', linewidths=1,\n alpha=0.65)\n ax.set_ylim(best_ylim)\n tick_labels = [columns[i] for i in order]\n ax.set_yticks(yaxis_locations)\n ax.set_yticklabels(tick_labels)\n ax.set_xlabel('log(HR) (%g%% CI)' % ((1 - self.alpha) * 100))\n return ax\n", "role": "user" }, { "content": "plot", "role": "assistant" } ]
def x(self, df, events, start, stop, weights): score_residuals = self._compute_residuals(df, events, start, stop, weights ) * weights[:, None] naive_var = inv(self._hessian_) delta_betas = -score_residuals.dot(naive_var) / self._norm_std.values return delta_betas
compute_delta_beta
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, df, events, start, stop, weights):\n score_residuals = self._compute_residuals(df, events, start, stop, weights\n ) * weights[:, None]\n naive_var = inv(self._hessian_)\n delta_betas = -score_residuals.dot(naive_var) / self._norm_std.values\n return delta_betas\n", "role": "user" }, { "content": "compute_delta_beta", "role": "assistant" } ]
def x(self, df, duration_col=None, event_col=None, show_progress=False, initial_point=None, strata=None, step_size=None, weights_col=None, cluster_col=None, robust=False, batch_mode=None): if duration_col is None: raise TypeError('duration_col cannot be None.') self._censoring_type = CensoringType.RIGHT self._time_fit_was_called = datetime.utcnow().strftime('%Y-%m-%d %H:%M:%S' ) + ' UTC' self.duration_col = duration_col self.event_col = event_col self.robust = robust self.cluster_col = cluster_col self.weights_col = weights_col self._n_examples = df.shape[0] self._batch_mode = batch_mode self.strata = coalesce(strata, self.strata) X, T, E, weights, original_index, self._clusters = (self. _preprocess_dataframe(df)) self.durations = T.copy() self.event_observed = E.copy() self.weights = weights.copy() if self.strata is not None: self.durations.index = original_index self.event_observed.index = original_index self.weights.index = original_index self._norm_mean = X.mean(0) self._norm_std = X.std(0) X_norm = normalize(X, self._norm_mean, self._norm_std) hazards_ = self._fit_model(X_norm, T, E, weights=weights, initial_point =initial_point, show_progress=show_progress, step_size=step_size) self.hazards_ = pd.Series(hazards_, index=X.columns, name='coef' ) / self._norm_std self.variance_matrix_ = -inv(self._hessian_) / np.outer(self._norm_std, self._norm_std) self.standard_errors_ = self._compute_standard_errors(X_norm, T, E, weights ) self.confidence_intervals_ = self._compute_confidence_intervals() self._predicted_partial_hazards_ = self.predict_partial_hazard(X).rename( columns={(0): 'P'}).assign(T=self.durations.values, E=self. event_observed.values, W=self.weights.values).set_index(X.index) self.baseline_hazard_ = self._compute_baseline_hazards() self.baseline_cumulative_hazard_ = (self. _compute_baseline_cumulative_hazard()) self.baseline_survival_ = self._compute_baseline_survival() if hasattr(self, '_concordance_score_'): del self._concordance_score_ return self
fit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, df, duration_col=None, event_col=None, show_progress=False,\n initial_point=None, strata=None, step_size=None, weights_col=None,\n cluster_col=None, robust=False, batch_mode=None):\n if duration_col is None:\n raise TypeError('duration_col cannot be None.')\n self._censoring_type = CensoringType.RIGHT\n self._time_fit_was_called = datetime.utcnow().strftime('%Y-%m-%d %H:%M:%S'\n ) + ' UTC'\n self.duration_col = duration_col\n self.event_col = event_col\n self.robust = robust\n self.cluster_col = cluster_col\n self.weights_col = weights_col\n self._n_examples = df.shape[0]\n self._batch_mode = batch_mode\n self.strata = coalesce(strata, self.strata)\n X, T, E, weights, original_index, self._clusters = (self.\n _preprocess_dataframe(df))\n self.durations = T.copy()\n self.event_observed = E.copy()\n self.weights = weights.copy()\n if self.strata is not None:\n self.durations.index = original_index\n self.event_observed.index = original_index\n self.weights.index = original_index\n self._norm_mean = X.mean(0)\n self._norm_std = X.std(0)\n X_norm = normalize(X, self._norm_mean, self._norm_std)\n hazards_ = self._fit_model(X_norm, T, E, weights=weights, initial_point\n =initial_point, show_progress=show_progress, step_size=step_size)\n self.hazards_ = pd.Series(hazards_, index=X.columns, name='coef'\n ) / self._norm_std\n self.variance_matrix_ = -inv(self._hessian_) / np.outer(self._norm_std,\n self._norm_std)\n self.standard_errors_ = self._compute_standard_errors(X_norm, T, E, weights\n )\n self.confidence_intervals_ = self._compute_confidence_intervals()\n self._predicted_partial_hazards_ = self.predict_partial_hazard(X).rename(\n columns={(0): 'P'}).assign(T=self.durations.values, E=self.\n event_observed.values, W=self.weights.values).set_index(X.index)\n self.baseline_hazard_ = self._compute_baseline_hazards()\n self.baseline_cumulative_hazard_ = (self.\n _compute_baseline_cumulative_hazard())\n self.baseline_survival_ = self._compute_baseline_survival()\n if hasattr(self, '_concordance_score_'):\n del self._concordance_score_\n return self\n", "role": "user" }, { "content": "fit", "role": "assistant" } ]
def x(self, X, T, E, weights=None, initial_point=None, step_size=None, precision=1e-07, show_progress=True, max_steps=50): self.path = [] assert precision <= 1.0, 'precision must be less than or equal to 1.' _, d = X.shape if initial_point is not None: assert initial_point.shape == (d,) beta = initial_point else: beta = np.zeros((d,)) step_sizer = StepSizer(step_size) step_size = step_sizer.next() if self.tie_method == 'Efron': decision = BatchVsSingle.decide(self._batch_mode, T) get_gradients = getattr(self, '_get_efron_values_%s' % decision) self._batch_mode = decision == 'batch' else: raise NotImplementedError('Only Efron is available.') i = 0 converging = True ll, previous_ll = 0, 0 start = time.time() while converging: self.path.append(beta.copy()) i += 1 if self.strata is None: h, g, ll = get_gradients(X.values, T.values, E.values, weights. values, beta) else: g = np.zeros_like(beta) h = np.zeros((beta.shape[0], beta.shape[0])) ll = 0 for _h, _g, _ll in self._partition_by_strata_and_apply(X, T, E, weights, get_gradients, beta): g += _g h += _h ll += _ll if i == 1 and np.all(beta == 0): self._log_likelihood_null = ll if self.penalizer > 0: g -= self.penalizer * beta h.flat[::d + 1] -= self.penalizer try: inv_h_dot_g_T = spsolve(-h, g, assume_a='pos', check_finite=False) except ValueError as e: if 'infs or NaNs' in str(e): raise ConvergenceError( """Hessian or gradient contains nan or inf value(s). Convergence halted. Please see the following tips in the lifelines documentation: https://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model """ , e) else: raise e except LinAlgError as e: raise ConvergenceError( """Convergence halted due to matrix inversion problems. Suspicion is high collinearity. Please see the following tips in the lifelines documentation: https://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model """ , e) delta = inv_h_dot_g_T if np.any(np.isnan(delta)): raise ConvergenceError( """delta contains nan value(s). Convergence halted. Please see the following tips in the lifelines documentation: https://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model """ ) hessian, gradient = h, g norm_delta = norm(delta) newton_decrement = g.dot(inv_h_dot_g_T) / 2 if show_progress: print( 'Iteration %d: norm_delta = %.5f, step_size = %.4f, ll = %.5f, newton_decrement = %.5f, seconds_since_start = %.1f' % (i, norm_delta, step_size, ll, newton_decrement, time. time() - start)) if norm_delta < precision: converging, completed = False, True elif previous_ll != 0 and abs(ll - previous_ll) / -previous_ll < 1e-09: converging, completed = False, True elif newton_decrement < precision: converging, completed = False, True elif i >= max_steps: converging, completed = False, False elif step_size <= 1e-05: converging, completed = False, False elif abs(ll) < 0.0001 and norm_delta > 1.0: warnings.warn( 'The log-likelihood is getting suspiciously close to 0 and the delta is still large. There may be complete separation in the dataset. This may result in incorrect inference of coefficients. See https://stats.stackexchange.com/questions/11109/how-to-deal-with-perfect-separation-in-logistic-regression' , ConvergenceWarning) converging, completed = False, False beta += step_size * delta previous_ll = ll step_size = step_sizer.update(norm_delta).next() self._hessian_ = hessian self._score_ = gradient self._log_likelihood = ll if show_progress and completed: print('Convergence completed after %d iterations.' % i) elif show_progress and not completed: print('Convergence failed. See any warning messages.') if completed and norm_delta > 0.1: warnings.warn( 'Newton-Rhapson convergence completed but norm(delta) is still high, %.3f. This may imply non-unique solutions to the maximum likelihood. Perhaps there is collinearity or complete separation in the dataset?' % norm_delta, ConvergenceWarning) elif not completed: warnings.warn( 'Newton-Rhapson failed to converge sufficiently in %d steps.' % max_steps, ConvergenceWarning) return beta
fit_model
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, T, E, weights=None, initial_point=None, step_size=None,\n precision=1e-07, show_progress=True, max_steps=50):\n self.path = []\n assert precision <= 1.0, 'precision must be less than or equal to 1.'\n _, d = X.shape\n if initial_point is not None:\n assert initial_point.shape == (d,)\n beta = initial_point\n else:\n beta = np.zeros((d,))\n step_sizer = StepSizer(step_size)\n step_size = step_sizer.next()\n if self.tie_method == 'Efron':\n decision = BatchVsSingle.decide(self._batch_mode, T)\n get_gradients = getattr(self, '_get_efron_values_%s' % decision)\n self._batch_mode = decision == 'batch'\n else:\n raise NotImplementedError('Only Efron is available.')\n i = 0\n converging = True\n ll, previous_ll = 0, 0\n start = time.time()\n while converging:\n self.path.append(beta.copy())\n i += 1\n if self.strata is None:\n h, g, ll = get_gradients(X.values, T.values, E.values, weights.\n values, beta)\n else:\n g = np.zeros_like(beta)\n h = np.zeros((beta.shape[0], beta.shape[0]))\n ll = 0\n for _h, _g, _ll in self._partition_by_strata_and_apply(X, T, E,\n weights, get_gradients, beta):\n g += _g\n h += _h\n ll += _ll\n if i == 1 and np.all(beta == 0):\n self._log_likelihood_null = ll\n if self.penalizer > 0:\n g -= self.penalizer * beta\n h.flat[::d + 1] -= self.penalizer\n try:\n inv_h_dot_g_T = spsolve(-h, g, assume_a='pos', check_finite=False)\n except ValueError as e:\n if 'infs or NaNs' in str(e):\n raise ConvergenceError(\n \"\"\"Hessian or gradient contains nan or inf value(s). Convergence halted. Please see the following tips in the lifelines documentation:\nhttps://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model\n\"\"\"\n , e)\n else:\n raise e\n except LinAlgError as e:\n raise ConvergenceError(\n \"\"\"Convergence halted due to matrix inversion problems. Suspicion is high collinearity. Please see the following tips in the lifelines documentation:\nhttps://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model\n\"\"\"\n , e)\n delta = inv_h_dot_g_T\n if np.any(np.isnan(delta)):\n raise ConvergenceError(\n \"\"\"delta contains nan value(s). Convergence halted. Please see the following tips in the lifelines documentation:\nhttps://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model\n\"\"\"\n )\n hessian, gradient = h, g\n norm_delta = norm(delta)\n newton_decrement = g.dot(inv_h_dot_g_T) / 2\n if show_progress:\n print(\n 'Iteration %d: norm_delta = %.5f, step_size = %.4f, ll = %.5f, newton_decrement = %.5f, seconds_since_start = %.1f'\n % (i, norm_delta, step_size, ll, newton_decrement, time.\n time() - start))\n if norm_delta < precision:\n converging, completed = False, True\n elif previous_ll != 0 and abs(ll - previous_ll) / -previous_ll < 1e-09:\n converging, completed = False, True\n elif newton_decrement < precision:\n converging, completed = False, True\n elif i >= max_steps:\n converging, completed = False, False\n elif step_size <= 1e-05:\n converging, completed = False, False\n elif abs(ll) < 0.0001 and norm_delta > 1.0:\n warnings.warn(\n 'The log-likelihood is getting suspiciously close to 0 and the delta is still large. There may be complete separation in the dataset. This may result in incorrect inference of coefficients. See https://stats.stackexchange.com/questions/11109/how-to-deal-with-perfect-separation-in-logistic-regression'\n , ConvergenceWarning)\n converging, completed = False, False\n beta += step_size * delta\n previous_ll = ll\n step_size = step_sizer.update(norm_delta).next()\n self._hessian_ = hessian\n self._score_ = gradient\n self._log_likelihood = ll\n if show_progress and completed:\n print('Convergence completed after %d iterations.' % i)\n elif show_progress and not completed:\n print('Convergence failed. See any warning messages.')\n if completed and norm_delta > 0.1:\n warnings.warn(\n 'Newton-Rhapson convergence completed but norm(delta) is still high, %.3f. This may imply non-unique solutions to the maximum likelihood. Perhaps there is collinearity or complete separation in the dataset?'\n % norm_delta, ConvergenceWarning)\n elif not completed:\n warnings.warn(\n 'Newton-Rhapson failed to converge sufficiently in %d steps.' %\n max_steps, ConvergenceWarning)\n return beta\n", "role": "user" }, { "content": "fit_model", "role": "assistant" } ]
def x(self, X, T, E, weights, beta): n, d = X.shape hessian = np.zeros((d, d)) gradient = np.zeros((d,)) log_lik = 0 x_death_sum = np.zeros((d,)) risk_phi, tie_phi = 0, 0 risk_phi_x, tie_phi_x = np.zeros((d,)), np.zeros((d,)) risk_phi_x_x, tie_phi_x_x = np.zeros((d, d)), np.zeros((d, d)) weight_count = 0.0 tied_death_counts = 0 scores = weights * np.exp(np.dot(X, beta)) for i in range(n - 1, -1, -1): ti = T[i] ei = E[i] xi = X[i] score = scores[i] w = weights[i] phi_i = score phi_x_i = phi_i * xi phi_x_x_i = np.outer(xi, phi_x_i) risk_phi = risk_phi + phi_i risk_phi_x = risk_phi_x + phi_x_i risk_phi_x_x = risk_phi_x_x + phi_x_x_i if ei: x_death_sum = x_death_sum + w * xi tie_phi = tie_phi + phi_i tie_phi_x = tie_phi_x + phi_x_i tie_phi_x_x = tie_phi_x_x + phi_x_x_i tied_death_counts += 1 weight_count += w if i > 0 and T[i - 1] == ti: continue elif tied_death_counts == 0: continue weighted_average = weight_count / tied_death_counts if tied_death_counts > 1: increasing_proportion = np.arange(tied_death_counts ) / tied_death_counts denom = 1.0 / (risk_phi - increasing_proportion * tie_phi) numer = risk_phi_x - np.outer(increasing_proportion, tie_phi_x) a1 = np.einsum('ab,i->ab', risk_phi_x_x, denom) - np.einsum( 'ab,i->ab', tie_phi_x_x, increasing_proportion * denom) else: denom = 1.0 / np.array([risk_phi]) numer = risk_phi_x a1 = risk_phi_x_x * denom summand = numer * denom[:, None] a2 = summand.T.dot(summand) gradient = gradient + x_death_sum - weighted_average * summand.sum(0) log_lik = log_lik + np.dot(x_death_sum, beta ) + weighted_average * np.log(denom).sum() hessian = hessian + weighted_average * (a2 - a1) tied_death_counts = 0 weight_count = 0.0 x_death_sum = np.zeros((d,)) tie_phi = 0 tie_phi_x = np.zeros((d,)) tie_phi_x_x = np.zeros((d, d)) return hessian, gradient, log_lik
get_efron_values_single
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, T, E, weights, beta):\n n, d = X.shape\n hessian = np.zeros((d, d))\n gradient = np.zeros((d,))\n log_lik = 0\n x_death_sum = np.zeros((d,))\n risk_phi, tie_phi = 0, 0\n risk_phi_x, tie_phi_x = np.zeros((d,)), np.zeros((d,))\n risk_phi_x_x, tie_phi_x_x = np.zeros((d, d)), np.zeros((d, d))\n weight_count = 0.0\n tied_death_counts = 0\n scores = weights * np.exp(np.dot(X, beta))\n for i in range(n - 1, -1, -1):\n ti = T[i]\n ei = E[i]\n xi = X[i]\n score = scores[i]\n w = weights[i]\n phi_i = score\n phi_x_i = phi_i * xi\n phi_x_x_i = np.outer(xi, phi_x_i)\n risk_phi = risk_phi + phi_i\n risk_phi_x = risk_phi_x + phi_x_i\n risk_phi_x_x = risk_phi_x_x + phi_x_x_i\n if ei:\n x_death_sum = x_death_sum + w * xi\n tie_phi = tie_phi + phi_i\n tie_phi_x = tie_phi_x + phi_x_i\n tie_phi_x_x = tie_phi_x_x + phi_x_x_i\n tied_death_counts += 1\n weight_count += w\n if i > 0 and T[i - 1] == ti:\n continue\n elif tied_death_counts == 0:\n continue\n weighted_average = weight_count / tied_death_counts\n if tied_death_counts > 1:\n increasing_proportion = np.arange(tied_death_counts\n ) / tied_death_counts\n denom = 1.0 / (risk_phi - increasing_proportion * tie_phi)\n numer = risk_phi_x - np.outer(increasing_proportion, tie_phi_x)\n a1 = np.einsum('ab,i->ab', risk_phi_x_x, denom) - np.einsum(\n 'ab,i->ab', tie_phi_x_x, increasing_proportion * denom)\n else:\n denom = 1.0 / np.array([risk_phi])\n numer = risk_phi_x\n a1 = risk_phi_x_x * denom\n summand = numer * denom[:, None]\n a2 = summand.T.dot(summand)\n gradient = gradient + x_death_sum - weighted_average * summand.sum(0)\n log_lik = log_lik + np.dot(x_death_sum, beta\n ) + weighted_average * np.log(denom).sum()\n hessian = hessian + weighted_average * (a2 - a1)\n tied_death_counts = 0\n weight_count = 0.0\n x_death_sum = np.zeros((d,))\n tie_phi = 0\n tie_phi_x = np.zeros((d,))\n tie_phi_x_x = np.zeros((d, d))\n return hessian, gradient, log_lik\n", "role": "user" }, { "content": "get_efron_values_single", "role": "assistant" } ]
def x(self, X, T, E, weights, beta): n, d = X.shape hessian = np.zeros((d, d)) gradient = np.zeros((d,)) log_lik = 0 risk_phi, tie_phi = 0, 0 risk_phi_x, tie_phi_x = np.zeros((d,)), np.zeros((d,)) risk_phi_x_x, tie_phi_x_x = np.zeros((d, d)), np.zeros((d, d)) _, counts = np.unique(-T, return_counts=True) scores = weights * np.exp(np.dot(X, beta)) pos = n for count_of_removals in counts: slice_ = slice(pos - count_of_removals, pos) X_at_t = X[slice_] weights_at_t = weights[slice_] phi_i = scores[slice_, None] phi_x_i = phi_i * X_at_t phi_x_x_i = np.dot(X_at_t.T, phi_x_i) risk_phi = risk_phi + array_sum_to_scalar(phi_i) risk_phi_x = risk_phi_x + matrix_axis_0_sum_to_array(phi_x_i) risk_phi_x_x = risk_phi_x_x + phi_x_x_i deaths = E[slice_] tied_death_counts = array_sum_to_scalar(deaths.astype(int)) if tied_death_counts == 0: pos -= count_of_removals continue xi_deaths = X_at_t[deaths] weights_deaths = weights_at_t[deaths] x_death_sum = matrix_axis_0_sum_to_array(weights_deaths[:, None] * xi_deaths) weight_count = array_sum_to_scalar(weights_deaths) weighted_average = weight_count / tied_death_counts if tied_death_counts > 1: tie_phi = array_sum_to_scalar(phi_i[deaths]) tie_phi_x = matrix_axis_0_sum_to_array(phi_x_i[deaths]) tie_phi_x_x = np.dot(xi_deaths.T, phi_i[deaths] * xi_deaths) increasing_proportion = np.arange(tied_death_counts ) / tied_death_counts denom = 1.0 / (risk_phi - increasing_proportion * tie_phi) numer = risk_phi_x - np.outer(increasing_proportion, tie_phi_x) a1 = np.einsum('ab,i->ab', risk_phi_x_x, denom) - np.einsum( 'ab,i->ab', tie_phi_x_x, increasing_proportion * denom) else: denom = 1.0 / np.array([risk_phi]) numer = risk_phi_x a1 = risk_phi_x_x * denom summand = numer * denom[:, None] a2 = summand.T.dot(summand) gradient = gradient + x_death_sum - weighted_average * summand.sum(0) log_lik = log_lik + np.dot(x_death_sum, beta ) + weighted_average * np.log(denom).sum() hessian = hessian + weighted_average * (a2 - a1) pos -= count_of_removals return hessian, gradient, log_lik
get_efron_values_batch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, T, E, weights, beta):\n n, d = X.shape\n hessian = np.zeros((d, d))\n gradient = np.zeros((d,))\n log_lik = 0\n risk_phi, tie_phi = 0, 0\n risk_phi_x, tie_phi_x = np.zeros((d,)), np.zeros((d,))\n risk_phi_x_x, tie_phi_x_x = np.zeros((d, d)), np.zeros((d, d))\n _, counts = np.unique(-T, return_counts=True)\n scores = weights * np.exp(np.dot(X, beta))\n pos = n\n for count_of_removals in counts:\n slice_ = slice(pos - count_of_removals, pos)\n X_at_t = X[slice_]\n weights_at_t = weights[slice_]\n phi_i = scores[slice_, None]\n phi_x_i = phi_i * X_at_t\n phi_x_x_i = np.dot(X_at_t.T, phi_x_i)\n risk_phi = risk_phi + array_sum_to_scalar(phi_i)\n risk_phi_x = risk_phi_x + matrix_axis_0_sum_to_array(phi_x_i)\n risk_phi_x_x = risk_phi_x_x + phi_x_x_i\n deaths = E[slice_]\n tied_death_counts = array_sum_to_scalar(deaths.astype(int))\n if tied_death_counts == 0:\n pos -= count_of_removals\n continue\n xi_deaths = X_at_t[deaths]\n weights_deaths = weights_at_t[deaths]\n x_death_sum = matrix_axis_0_sum_to_array(weights_deaths[:, None] *\n xi_deaths)\n weight_count = array_sum_to_scalar(weights_deaths)\n weighted_average = weight_count / tied_death_counts\n if tied_death_counts > 1:\n tie_phi = array_sum_to_scalar(phi_i[deaths])\n tie_phi_x = matrix_axis_0_sum_to_array(phi_x_i[deaths])\n tie_phi_x_x = np.dot(xi_deaths.T, phi_i[deaths] * xi_deaths)\n increasing_proportion = np.arange(tied_death_counts\n ) / tied_death_counts\n denom = 1.0 / (risk_phi - increasing_proportion * tie_phi)\n numer = risk_phi_x - np.outer(increasing_proportion, tie_phi_x)\n a1 = np.einsum('ab,i->ab', risk_phi_x_x, denom) - np.einsum(\n 'ab,i->ab', tie_phi_x_x, increasing_proportion * denom)\n else:\n denom = 1.0 / np.array([risk_phi])\n numer = risk_phi_x\n a1 = risk_phi_x_x * denom\n summand = numer * denom[:, None]\n a2 = summand.T.dot(summand)\n gradient = gradient + x_death_sum - weighted_average * summand.sum(0)\n log_lik = log_lik + np.dot(x_death_sum, beta\n ) + weighted_average * np.log(denom).sum()\n hessian = hessian + weighted_average * (a2 - a1)\n pos -= count_of_removals\n return hessian, gradient, log_lik\n", "role": "user" }, { "content": "get_efron_values_batch", "role": "assistant" } ]
def x(self, X, T, E, weights, index=None): n_deaths = self.event_observed.sum() scaled_schoenfeld_resids = n_deaths * self._compute_schoenfeld(X, T, E, weights, index).dot(self.variance_matrix_) scaled_schoenfeld_resids.columns = self.hazards_.index return scaled_schoenfeld_resids
compute_scaled_schoenfeld
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, T, E, weights, index=None):\n n_deaths = self.event_observed.sum()\n scaled_schoenfeld_resids = n_deaths * self._compute_schoenfeld(X, T, E,\n weights, index).dot(self.variance_matrix_)\n scaled_schoenfeld_resids.columns = self.hazards_.index\n return scaled_schoenfeld_resids\n", "role": "user" }, { "content": "compute_scaled_schoenfeld", "role": "assistant" } ]
def x(self, X, T, E, weights): n, d = X.shape if not np.any(E): return np.zeros((n, d)) risk_phi, tie_phi = 0, 0 risk_phi_x, tie_phi_x = np.zeros((1, d)), np.zeros((1, d)) weight_count = 0.0 tie_count = 0 scores = weights * np.exp(np.dot(X, self.hazards_)) diff_against = [] schoenfeld_residuals = np.empty((0, d)) for i in range(n - 1, -1, -1): ti = T[i] ei = E[i] xi = X[i:i + 1] score = scores[i:i + 1] w = weights[i] phi_i = score phi_x_i = phi_i * xi risk_phi = risk_phi + phi_i risk_phi_x = risk_phi_x + phi_x_i diff_against.append((xi, ei)) if ei: tie_phi = tie_phi + phi_i tie_phi_x = tie_phi_x + phi_x_i tie_count += 1 weight_count += w if i > 0 and T[i - 1] == ti: continue elif tie_count == 0: for _ in diff_against: schoenfeld_residuals = np.append(schoenfeld_residuals, np. zeros((1, d)), axis=0) diff_against = [] continue weighted_mean = np.zeros((1, d)) for l in range(tie_count): numer = risk_phi_x - l * tie_phi_x / tie_count denom = risk_phi - l * tie_phi / tie_count weighted_mean += numer / (denom * tie_count) for xi, ei in diff_against: schoenfeld_residuals = np.append(schoenfeld_residuals, ei * (xi - weighted_mean), axis=0) tie_count = 0 weight_count = 0.0 tie_phi = 0 tie_phi_x = np.zeros((1, d)) diff_against = [] return schoenfeld_residuals[::-1]
compute_schoenfeld_within_strata
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, T, E, weights):\n n, d = X.shape\n if not np.any(E):\n return np.zeros((n, d))\n risk_phi, tie_phi = 0, 0\n risk_phi_x, tie_phi_x = np.zeros((1, d)), np.zeros((1, d))\n weight_count = 0.0\n tie_count = 0\n scores = weights * np.exp(np.dot(X, self.hazards_))\n diff_against = []\n schoenfeld_residuals = np.empty((0, d))\n for i in range(n - 1, -1, -1):\n ti = T[i]\n ei = E[i]\n xi = X[i:i + 1]\n score = scores[i:i + 1]\n w = weights[i]\n phi_i = score\n phi_x_i = phi_i * xi\n risk_phi = risk_phi + phi_i\n risk_phi_x = risk_phi_x + phi_x_i\n diff_against.append((xi, ei))\n if ei:\n tie_phi = tie_phi + phi_i\n tie_phi_x = tie_phi_x + phi_x_i\n tie_count += 1\n weight_count += w\n if i > 0 and T[i - 1] == ti:\n continue\n elif tie_count == 0:\n for _ in diff_against:\n schoenfeld_residuals = np.append(schoenfeld_residuals, np.\n zeros((1, d)), axis=0)\n diff_against = []\n continue\n weighted_mean = np.zeros((1, d))\n for l in range(tie_count):\n numer = risk_phi_x - l * tie_phi_x / tie_count\n denom = risk_phi - l * tie_phi / tie_count\n weighted_mean += numer / (denom * tie_count)\n for xi, ei in diff_against:\n schoenfeld_residuals = np.append(schoenfeld_residuals, ei * (xi -\n weighted_mean), axis=0)\n tie_count = 0\n weight_count = 0.0\n tie_phi = 0\n tie_phi_x = np.zeros((1, d))\n diff_against = []\n return schoenfeld_residuals[::-1]\n", "role": "user" }, { "content": "compute_schoenfeld_within_strata", "role": "assistant" } ]
def x(self, X, T, E, weights, index=None): score_residuals = self._compute_score(X, T, E, weights, index=index) d = X.shape[1] scaled_variance_matrix = self.variance_matrix_ * np.tile(self._norm_std .values, (d, 1)).T delta_betas = score_residuals.dot(scaled_variance_matrix) delta_betas.columns = self.hazards_.index return delta_betas
compute_delta_beta
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, T, E, weights, index=None):\n score_residuals = self._compute_score(X, T, E, weights, index=index)\n d = X.shape[1]\n scaled_variance_matrix = self.variance_matrix_ * np.tile(self._norm_std\n .values, (d, 1)).T\n delta_betas = score_residuals.dot(scaled_variance_matrix)\n delta_betas.columns = self.hazards_.index\n return delta_betas\n", "role": "user" }, { "content": "compute_delta_beta", "role": "assistant" } ]
def x(self): ci = 1 - self.alpha with np.errstate(invalid='ignore', divide='ignore'): df = pd.DataFrame(index=self.hazards_.index) df['coef'] = self.hazards_ df['exp(coef)'] = np.exp(self.hazards_) df['se(coef)'] = self.standard_errors_ df['z'] = self._compute_z_values() df['p'] = self._compute_p_values() df['-log2(p)'] = -np.log2(df['p']) df['lower %g' % ci] = self.confidence_intervals_['lower-bound'] df['upper %g' % ci] = self.confidence_intervals_['upper-bound'] return df
summary
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n ci = 1 - self.alpha\n with np.errstate(invalid='ignore', divide='ignore'):\n df = pd.DataFrame(index=self.hazards_.index)\n df['coef'] = self.hazards_\n df['exp(coef)'] = np.exp(self.hazards_)\n df['se(coef)'] = self.standard_errors_\n df['z'] = self._compute_z_values()\n df['p'] = self._compute_p_values()\n df['-log2(p)'] = -np.log2(df['p'])\n df['lower %g' % ci] = self.confidence_intervals_['lower-bound']\n df['upper %g' % ci] = self.confidence_intervals_['upper-bound']\n return df\n", "role": "user" }, { "content": "summary", "role": "assistant" } ]
def x(self, X): hazard_names = self.hazards_.index if isinstance(X, pd.Series) and (X.shape[0] == len(hazard_names) + 2 or X.shape[0] == len(hazard_names)): X = X.to_frame().T return self.predict_log_partial_hazard(X) elif isinstance(X, pd.Series): assert len(hazard_names) == 1, 'Series not the correct argument' X = X.to_frame().T return self.predict_log_partial_hazard(X) index = _get_index(X) if isinstance(X, pd.DataFrame): order = hazard_names X = X.reindex(order, axis='columns') check_for_numeric_dtypes_or_raise(X) X = X.values X = X.astype(float) X = normalize(X, self._norm_mean.values, 1) return pd.DataFrame(np.dot(X, self.hazards_), index=index)
predict_log_partial_hazard
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X):\n hazard_names = self.hazards_.index\n if isinstance(X, pd.Series) and (X.shape[0] == len(hazard_names) + 2 or\n X.shape[0] == len(hazard_names)):\n X = X.to_frame().T\n return self.predict_log_partial_hazard(X)\n elif isinstance(X, pd.Series):\n assert len(hazard_names) == 1, 'Series not the correct argument'\n X = X.to_frame().T\n return self.predict_log_partial_hazard(X)\n index = _get_index(X)\n if isinstance(X, pd.DataFrame):\n order = hazard_names\n X = X.reindex(order, axis='columns')\n check_for_numeric_dtypes_or_raise(X)\n X = X.values\n X = X.astype(float)\n X = normalize(X, self._norm_mean.values, 1)\n return pd.DataFrame(np.dot(X, self.hazards_), index=index)\n", "role": "user" }, { "content": "predict_log_partial_hazard", "role": "assistant" } ]
def x(self, X, times=None): if self.strata: cumulative_hazard_ = pd.DataFrame() for stratum, stratified_X in X.groupby(self.strata): try: c_0 = self.baseline_cumulative_hazard_[[stratum]] except KeyError: raise StatError( """The stratum %s was not found in the original training data. For example, try the following on the original dataset, df: `df.groupby(%s).size()`. Expected is that %s is not present in the output. """ % (stratum, self.strata, stratum)) col = _get_index(stratified_X) v = self.predict_partial_hazard(stratified_X) cumulative_hazard_ = cumulative_hazard_.merge(pd.DataFrame(np. dot(c_0, v.T), index=c_0.index, columns=col), how='outer', right_index=True, left_index=True) else: c_0 = self.baseline_cumulative_hazard_ v = self.predict_partial_hazard(X) col = _get_index(v) cumulative_hazard_ = pd.DataFrame(np.dot(c_0, v.T), columns=col, index=c_0.index) if times is not None: return dataframe_interpolate_at_times(cumulative_hazard_, times) return cumulative_hazard_
predict_cumulative_hazard
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, times=None):\n if self.strata:\n cumulative_hazard_ = pd.DataFrame()\n for stratum, stratified_X in X.groupby(self.strata):\n try:\n c_0 = self.baseline_cumulative_hazard_[[stratum]]\n except KeyError:\n raise StatError(\n \"\"\"The stratum %s was not found in the original training data. For example, try\nthe following on the original dataset, df: `df.groupby(%s).size()`. Expected is that %s is not present in the output.\n\"\"\"\n % (stratum, self.strata, stratum))\n col = _get_index(stratified_X)\n v = self.predict_partial_hazard(stratified_X)\n cumulative_hazard_ = cumulative_hazard_.merge(pd.DataFrame(np.\n dot(c_0, v.T), index=c_0.index, columns=col), how='outer',\n right_index=True, left_index=True)\n else:\n c_0 = self.baseline_cumulative_hazard_\n v = self.predict_partial_hazard(X)\n col = _get_index(v)\n cumulative_hazard_ = pd.DataFrame(np.dot(c_0, v.T), columns=col,\n index=c_0.index)\n if times is not None:\n return dataframe_interpolate_at_times(cumulative_hazard_, times)\n return cumulative_hazard_\n", "role": "user" }, { "content": "predict_cumulative_hazard", "role": "assistant" } ]
def x(self, X, times=None): return np.exp(-self.predict_cumulative_hazard(X, times=times))
predict_survival_function
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, times=None):\n return np.exp(-self.predict_cumulative_hazard(X, times=times))\n", "role": "user" }, { "content": "predict_survival_function", "role": "assistant" } ]
def x(self, X, p=0.5): subjects = _get_index(X) return qth_survival_times(p, self.predict_survival_function(X)[subjects]).T
predict_percentile
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, p=0.5):\n subjects = _get_index(X)\n return qth_survival_times(p, self.predict_survival_function(X)[subjects]).T\n", "role": "user" }, { "content": "predict_percentile", "role": "assistant" } ]
def x(self, X): subjects = _get_index(X) v = self.predict_survival_function(X)[subjects] return pd.DataFrame(trapz(v.values.T, v.index), index=subjects)
predict_expectation
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X):\n subjects = _get_index(X)\n v = self.predict_survival_function(X)[subjects]\n return pd.DataFrame(trapz(v.values.T, v.index), index=subjects)\n", "role": "user" }, { "content": "predict_expectation", "role": "assistant" } ]
def x(self): survival_df = np.exp(-self.baseline_cumulative_hazard_) if self.strata is None: survival_df.columns = ['baseline survival'] return survival_df
compute_baseline_survival
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n survival_df = np.exp(-self.baseline_cumulative_hazard_)\n if self.strata is None:\n survival_df.columns = ['baseline survival']\n return survival_df\n", "role": "user" }, { "content": "compute_baseline_survival", "role": "assistant" } ]
def x(self, covariates, values, plot_baseline=True, **kwargs): from matplotlib import pyplot as plt covariates = _to_list(covariates) n_covariates = len(covariates) values = _to_array(values) if len(values.shape) == 1: values = values[None, :].T if n_covariates != values.shape[1]: raise ValueError( 'The number of covariates must equal to second dimension of the values array.' ) for covariate in covariates: if covariate not in self.hazards_.index: raise KeyError( 'covariate `%s` is not present in the original dataset' % covariate) set_kwargs_drawstyle(kwargs, 'steps-post') if self.strata is None: axes = kwargs.pop('ax', None) or plt.figure().add_subplot(111) x_bar = self._norm_mean.to_frame().T X = pd.concat([x_bar] * values.shape[0]) if np.array_equal(np.eye(n_covariates), values): X.index = [('%s=1' % c) for c in covariates] else: X.index = [', '.join('%s=%g' % (c, v) for c, v in zip( covariates, row)) for row in values] for covariate, value in zip(covariates, values.T): X[covariate] = value self.predict_survival_function(X).plot(ax=axes, **kwargs) if plot_baseline: self.baseline_survival_.plot(ax=axes, ls=':', color='k', drawstyle='steps-post') else: axes = [] for stratum, baseline_survival_ in self.baseline_survival_.iteritems(): ax = plt.figure().add_subplot(1, 1, 1) x_bar = self._norm_mean.to_frame().T for name, value in zip(_to_list(self.strata), _to_tuple(stratum)): x_bar[name] = value X = pd.concat([x_bar] * values.shape[0]) if np.array_equal(np.eye(len(covariates)), values): X.index = [('%s=1' % c) for c in covariates] else: X.index = [', '.join('%s=%g' % (c, v) for c, v in zip( covariates, row)) for row in values] for covariate, value in zip(covariates, values.T): X[covariate] = value self.predict_survival_function(X).plot(ax=ax, **kwargs) if plot_baseline: baseline_survival_.plot(ax=ax, ls=':', label= 'stratum %s baseline survival' % str(stratum), drawstyle='steps-post') plt.legend() axes.append(ax) return axes
plot_covariate_groups
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, covariates, values, plot_baseline=True, **kwargs):\n from matplotlib import pyplot as plt\n covariates = _to_list(covariates)\n n_covariates = len(covariates)\n values = _to_array(values)\n if len(values.shape) == 1:\n values = values[None, :].T\n if n_covariates != values.shape[1]:\n raise ValueError(\n 'The number of covariates must equal to second dimension of the values array.'\n )\n for covariate in covariates:\n if covariate not in self.hazards_.index:\n raise KeyError(\n 'covariate `%s` is not present in the original dataset' %\n covariate)\n set_kwargs_drawstyle(kwargs, 'steps-post')\n if self.strata is None:\n axes = kwargs.pop('ax', None) or plt.figure().add_subplot(111)\n x_bar = self._norm_mean.to_frame().T\n X = pd.concat([x_bar] * values.shape[0])\n if np.array_equal(np.eye(n_covariates), values):\n X.index = [('%s=1' % c) for c in covariates]\n else:\n X.index = [', '.join('%s=%g' % (c, v) for c, v in zip(\n covariates, row)) for row in values]\n for covariate, value in zip(covariates, values.T):\n X[covariate] = value\n self.predict_survival_function(X).plot(ax=axes, **kwargs)\n if plot_baseline:\n self.baseline_survival_.plot(ax=axes, ls=':', color='k',\n drawstyle='steps-post')\n else:\n axes = []\n for stratum, baseline_survival_ in self.baseline_survival_.iteritems():\n ax = plt.figure().add_subplot(1, 1, 1)\n x_bar = self._norm_mean.to_frame().T\n for name, value in zip(_to_list(self.strata), _to_tuple(stratum)):\n x_bar[name] = value\n X = pd.concat([x_bar] * values.shape[0])\n if np.array_equal(np.eye(len(covariates)), values):\n X.index = [('%s=1' % c) for c in covariates]\n else:\n X.index = [', '.join('%s=%g' % (c, v) for c, v in zip(\n covariates, row)) for row in values]\n for covariate, value in zip(covariates, values.T):\n X[covariate] = value\n self.predict_survival_function(X).plot(ax=ax, **kwargs)\n if plot_baseline:\n baseline_survival_.plot(ax=ax, ls=':', label=\n 'stratum %s baseline survival' % str(stratum),\n drawstyle='steps-post')\n plt.legend()\n axes.append(ax)\n return axes\n", "role": "user" }, { "content": "plot_covariate_groups", "role": "assistant" } ]
def x(self, training_df, advice=True, show_plots=False, p_value_threshold= 0.01, plot_n_bootstraps=10): if not training_df.index.is_unique: raise IndexError( '`training_df` index should be unique for this exercise. Please make it unique or use `.reset_index(drop=True)` to force a unique index' ) residuals = self.compute_residuals(training_df, kind='scaled_schoenfeld') test_results = proportional_hazard_test(self, training_df, time_transform=['rank', 'km'], precomputed_residuals=residuals) residuals_and_duration = residuals.join(training_df[self.duration_col]) counter = 0 n = residuals_and_duration.shape[0] for variable in self.hazards_.index: minumum_observed_p_value = test_results.summary.loc[variable, 'p'].min( ) if np.round(minumum_observed_p_value, 2) > p_value_threshold: continue counter += 1 if counter == 1: if advice: print(fill( 'The ``p_value_threshold`` is set at %g. Even under the null hypothesis of no violations, some covariates will be below the threshold by chance. This is compounded when there are many covariates. Similarly, when there are lots of observations, even minor deviances from the proportional hazard assumption will be flagged.' % p_value_threshold, width=100)) print() print(fill( "With that in mind, it's best to use a combination of statistical tests and visual tests to determine the most serious violations. Produce visual plots using ``check_assumptions(..., show_plots=True)`` and looking for non-constant lines. See link [A] below for a full example." , width=100)) print() test_results.print_summary() print() print() print( "%d. Variable '%s' failed the non-proportional test: p-value is %s." % (counter, variable, format_p_value(4)( minumum_observed_p_value)), end='\n\n') if advice: values = training_df[variable] value_counts = values.value_counts() n_uniques = value_counts.shape[0] if n_uniques <= 10 and value_counts.min() >= 5: print(fill( " Advice: with so few unique values (only {0}), you can include `strata=['{1}', ...]` in the call in `.fit`. See documentation in link [E] below." .format(n_uniques, variable), width=100)) else: print(fill( " Advice 1: the functional form of the variable '{var}' might be incorrect. That is, there may be non-linear terms missing. The proportional hazard test used is very sensitive to incorrect functional forms. See documentation in link [D] below on how to specify a functional form." .format(var=variable), width=100), end='\n\n') print(fill( " Advice 2: try binning the variable '{var}' using pd.cut, and then specify it in `strata=['{var}', ...]` in the call in `.fit`. See documentation in link [B] below." .format(var=variable), width=100), end='\n\n') print(fill( ' Advice 3: try adding an interaction term with your time variable. See documentation in link [C] below.' .format(var=variable), width=100), end='\n\n') if show_plots: from matplotlib import pyplot as plt fig = plt.figure() for i, (transform_name, transformer) in enumerate(TimeTransformers ().iter(['rank', 'km']), start=1): p_value = test_results.summary.loc[(variable, transform_name), 'p'] ax = fig.add_subplot(1, 2, i) y = residuals_and_duration[variable] tt = transformer(self.durations, self.event_observed, self. weights)[self.event_observed.values] ax.scatter(tt, y, alpha=0.75) y_lowess = lowess(tt.values, y.values) ax.plot(tt, y_lowess, color='k', alpha=1.0, linewidth=2) for _ in range(plot_n_bootstraps): ix = sorted(np.random.choice(n, n)) tt_ = tt.values[ix] y_lowess = lowess(tt_, y.values[ix]) ax.plot(tt_, y_lowess, color='k', alpha=0.3) best_xlim = ax.get_xlim() ax.hlines(0, 0, tt.max(), linestyles='dashed', linewidths=1) ax.set_xlim(best_xlim) ax.set_xlabel('%s-transformed time\n(p=%.4f)' % ( transform_name, p_value), fontsize=10) fig.suptitle("Scaled Schoenfeld residuals of '%s'" % variable, fontsize=14) plt.tight_layout() plt.subplots_adjust(top=0.9) if advice and counter > 0: print(dedent( """ --- [A] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html [B] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html#Bin-variable-and-stratify-on-it [C] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html#Introduce-time-varying-covariates [D] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html#Modify-the-functional-form [E] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html#Stratification """ )) if counter == 0: print('Proportional hazard assumption looks okay.')
check_assumptions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, training_df, advice=True, show_plots=False, p_value_threshold=\n 0.01, plot_n_bootstraps=10):\n if not training_df.index.is_unique:\n raise IndexError(\n '`training_df` index should be unique for this exercise. Please make it unique or use `.reset_index(drop=True)` to force a unique index'\n )\n residuals = self.compute_residuals(training_df, kind='scaled_schoenfeld')\n test_results = proportional_hazard_test(self, training_df,\n time_transform=['rank', 'km'], precomputed_residuals=residuals)\n residuals_and_duration = residuals.join(training_df[self.duration_col])\n counter = 0\n n = residuals_and_duration.shape[0]\n for variable in self.hazards_.index:\n minumum_observed_p_value = test_results.summary.loc[variable, 'p'].min(\n )\n if np.round(minumum_observed_p_value, 2) > p_value_threshold:\n continue\n counter += 1\n if counter == 1:\n if advice:\n print(fill(\n 'The ``p_value_threshold`` is set at %g. Even under the null hypothesis of no violations, some covariates will be below the threshold by chance. This is compounded when there are many covariates. Similarly, when there are lots of observations, even minor deviances from the proportional hazard assumption will be flagged.'\n % p_value_threshold, width=100))\n print()\n print(fill(\n \"With that in mind, it's best to use a combination of statistical tests and visual tests to determine the most serious violations. Produce visual plots using ``check_assumptions(..., show_plots=True)`` and looking for non-constant lines. See link [A] below for a full example.\"\n , width=100))\n print()\n test_results.print_summary()\n print()\n print()\n print(\n \"%d. Variable '%s' failed the non-proportional test: p-value is %s.\"\n % (counter, variable, format_p_value(4)(\n minumum_observed_p_value)), end='\\n\\n')\n if advice:\n values = training_df[variable]\n value_counts = values.value_counts()\n n_uniques = value_counts.shape[0]\n if n_uniques <= 10 and value_counts.min() >= 5:\n print(fill(\n \" Advice: with so few unique values (only {0}), you can include `strata=['{1}', ...]` in the call in `.fit`. See documentation in link [E] below.\"\n .format(n_uniques, variable), width=100))\n else:\n print(fill(\n \" Advice 1: the functional form of the variable '{var}' might be incorrect. That is, there may be non-linear terms missing. The proportional hazard test used is very sensitive to incorrect functional forms. See documentation in link [D] below on how to specify a functional form.\"\n .format(var=variable), width=100), end='\\n\\n')\n print(fill(\n \" Advice 2: try binning the variable '{var}' using pd.cut, and then specify it in `strata=['{var}', ...]` in the call in `.fit`. See documentation in link [B] below.\"\n .format(var=variable), width=100), end='\\n\\n')\n print(fill(\n ' Advice 3: try adding an interaction term with your time variable. See documentation in link [C] below.'\n .format(var=variable), width=100), end='\\n\\n')\n if show_plots:\n from matplotlib import pyplot as plt\n fig = plt.figure()\n for i, (transform_name, transformer) in enumerate(TimeTransformers\n ().iter(['rank', 'km']), start=1):\n p_value = test_results.summary.loc[(variable,\n transform_name), 'p']\n ax = fig.add_subplot(1, 2, i)\n y = residuals_and_duration[variable]\n tt = transformer(self.durations, self.event_observed, self.\n weights)[self.event_observed.values]\n ax.scatter(tt, y, alpha=0.75)\n y_lowess = lowess(tt.values, y.values)\n ax.plot(tt, y_lowess, color='k', alpha=1.0, linewidth=2)\n for _ in range(plot_n_bootstraps):\n ix = sorted(np.random.choice(n, n))\n tt_ = tt.values[ix]\n y_lowess = lowess(tt_, y.values[ix])\n ax.plot(tt_, y_lowess, color='k', alpha=0.3)\n best_xlim = ax.get_xlim()\n ax.hlines(0, 0, tt.max(), linestyles='dashed', linewidths=1)\n ax.set_xlim(best_xlim)\n ax.set_xlabel('%s-transformed time\\n(p=%.4f)' % (\n transform_name, p_value), fontsize=10)\n fig.suptitle(\"Scaled Schoenfeld residuals of '%s'\" % variable,\n fontsize=14)\n plt.tight_layout()\n plt.subplots_adjust(top=0.9)\n if advice and counter > 0:\n print(dedent(\n \"\"\"\n ---\n [A] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html\n [B] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html#Bin-variable-and-stratify-on-it\n [C] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html#Introduce-time-varying-covariates\n [D] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html#Modify-the-functional-form\n [E] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html#Stratification\n \"\"\"\n ))\n if counter == 0:\n print('Proportional hazard assumption looks okay.')\n", "role": "user" }, { "content": "check_assumptions", "role": "assistant" } ]
def x(self): if not hasattr(self, '_concordance_score_'): if self.strata: num_correct, num_tied, num_pairs = 0, 0, 0 for _, _df in self._predicted_partial_hazards_.groupby(self.strata ): if _df.shape[0] == 1: continue _num_correct, _num_tied, _num_pairs = ( _concordance_summary_statistics(_df['T'].values, -_df[ 'P'].values, _df['E'].values)) num_correct += _num_correct num_tied += _num_tied num_pairs += _num_pairs else: df = self._predicted_partial_hazards_ num_correct, num_tied, num_pairs = _concordance_summary_statistics( df['T'].values, -df['P'].values, df['E'].values) self._concordance_score_ = _concordance_ratio(num_correct, num_tied, num_pairs) return self._concordance_score_ return self._concordance_score_
score_
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if not hasattr(self, '_concordance_score_'):\n if self.strata:\n num_correct, num_tied, num_pairs = 0, 0, 0\n for _, _df in self._predicted_partial_hazards_.groupby(self.strata\n ):\n if _df.shape[0] == 1:\n continue\n _num_correct, _num_tied, _num_pairs = (\n _concordance_summary_statistics(_df['T'].values, -_df[\n 'P'].values, _df['E'].values))\n num_correct += _num_correct\n num_tied += _num_tied\n num_pairs += _num_pairs\n else:\n df = self._predicted_partial_hazards_\n num_correct, num_tied, num_pairs = _concordance_summary_statistics(\n df['T'].values, -df['P'].values, df['E'].values)\n self._concordance_score_ = _concordance_ratio(num_correct, num_tied,\n num_pairs)\n return self._concordance_score_\n return self._concordance_score_\n", "role": "user" }, { "content": "score_", "role": "assistant" } ]
def x(event_times, predicted_scores, event_observed=None): event_times = np.asarray(event_times, dtype=float) predicted_scores = np.asarray(predicted_scores, dtype=float) if event_times.ndim == 2 and (event_times.shape[0] == 1 or event_times. shape[1] == 1): event_times = event_times.ravel() if predicted_scores.ndim == 2 and (predicted_scores.shape[0] == 1 or predicted_scores.shape[1] == 1): predicted_scores = predicted_scores.ravel() if event_times.shape != predicted_scores.shape: raise ValueError('Event times and predictions must have the same shape' ) if event_times.ndim != 1: raise ValueError('Event times can only be 1-dimensional: (n,)') if event_observed is None: event_observed = np.ones(event_times.shape[0], dtype=float) else: event_observed = np.asarray(event_observed, dtype=float).ravel() if event_observed.shape != event_times.shape: raise ValueError( 'Observed events must be 1-dimensional of same length as event times' ) num_correct, num_tied, num_pairs = _concordance_summary_statistics( event_times, predicted_scores, event_observed) return _concordance_ratio(num_correct, num_tied, num_pairs)
concordance_index
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(event_times, predicted_scores, event_observed=None):\n event_times = np.asarray(event_times, dtype=float)\n predicted_scores = np.asarray(predicted_scores, dtype=float)\n if event_times.ndim == 2 and (event_times.shape[0] == 1 or event_times.\n shape[1] == 1):\n event_times = event_times.ravel()\n if predicted_scores.ndim == 2 and (predicted_scores.shape[0] == 1 or \n predicted_scores.shape[1] == 1):\n predicted_scores = predicted_scores.ravel()\n if event_times.shape != predicted_scores.shape:\n raise ValueError('Event times and predictions must have the same shape'\n )\n if event_times.ndim != 1:\n raise ValueError('Event times can only be 1-dimensional: (n,)')\n if event_observed is None:\n event_observed = np.ones(event_times.shape[0], dtype=float)\n else:\n event_observed = np.asarray(event_observed, dtype=float).ravel()\n if event_observed.shape != event_times.shape:\n raise ValueError(\n 'Observed events must be 1-dimensional of same length as event times'\n )\n num_correct, num_tied, num_pairs = _concordance_summary_statistics(\n event_times, predicted_scores, event_observed)\n return _concordance_ratio(num_correct, num_tied, num_pairs)\n", "role": "user" }, { "content": "concordance_index", "role": "assistant" } ]
def x(event_times, predicted_event_times, event_observed): if np.logical_not(event_observed).all(): return 0, 0, 0 died_mask = event_observed.astype(bool) died_truth = event_times[died_mask] ix = np.argsort(died_truth) died_truth = died_truth[ix] died_pred = predicted_event_times[died_mask][ix] censored_truth = event_times[~died_mask] ix = np.argsort(censored_truth) censored_truth = censored_truth[ix] censored_pred = predicted_event_times[~died_mask][ix] censored_ix = 0 died_ix = 0 times_to_compare = _BTree(np.unique(died_pred)) num_pairs = np.int64(0) num_correct = np.int64(0) num_tied = np.int64(0) while True: has_more_censored = censored_ix < len(censored_truth) has_more_died = died_ix < len(died_truth) if has_more_censored and (not has_more_died or died_truth[died_ix] > censored_truth[censored_ix]): pairs, correct, tied, next_ix = _handle_pairs(censored_truth, censored_pred, censored_ix, times_to_compare) censored_ix = next_ix elif has_more_died and (not has_more_censored or died_truth[died_ix ] <= censored_truth[censored_ix]): pairs, correct, tied, next_ix = _handle_pairs(died_truth, died_pred, died_ix, times_to_compare) for pred in died_pred[died_ix:next_ix]: times_to_compare.insert(pred) died_ix = next_ix else: assert not (has_more_died or has_more_censored) break num_pairs += pairs num_correct += correct num_tied += tied return num_correct, num_tied, num_pairs
concordance_summary_statistics
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(event_times, predicted_event_times, event_observed):\n if np.logical_not(event_observed).all():\n return 0, 0, 0\n died_mask = event_observed.astype(bool)\n died_truth = event_times[died_mask]\n ix = np.argsort(died_truth)\n died_truth = died_truth[ix]\n died_pred = predicted_event_times[died_mask][ix]\n censored_truth = event_times[~died_mask]\n ix = np.argsort(censored_truth)\n censored_truth = censored_truth[ix]\n censored_pred = predicted_event_times[~died_mask][ix]\n censored_ix = 0\n died_ix = 0\n times_to_compare = _BTree(np.unique(died_pred))\n num_pairs = np.int64(0)\n num_correct = np.int64(0)\n num_tied = np.int64(0)\n while True:\n has_more_censored = censored_ix < len(censored_truth)\n has_more_died = died_ix < len(died_truth)\n if has_more_censored and (not has_more_died or died_truth[died_ix] >\n censored_truth[censored_ix]):\n pairs, correct, tied, next_ix = _handle_pairs(censored_truth,\n censored_pred, censored_ix, times_to_compare)\n censored_ix = next_ix\n elif has_more_died and (not has_more_censored or died_truth[died_ix\n ] <= censored_truth[censored_ix]):\n pairs, correct, tied, next_ix = _handle_pairs(died_truth,\n died_pred, died_ix, times_to_compare)\n for pred in died_pred[died_ix:next_ix]:\n times_to_compare.insert(pred)\n died_ix = next_ix\n else:\n assert not (has_more_died or has_more_censored)\n break\n num_pairs += pairs\n num_correct += correct\n num_tied += tied\n return num_correct, num_tied, num_pairs\n", "role": "user" }, { "content": "concordance_summary_statistics", "role": "assistant" } ]
def x(truth, pred, first_ix, times_to_compare): next_ix = first_ix while next_ix < len(truth) and truth[next_ix] == truth[first_ix]: next_ix += 1 pairs = len(times_to_compare) * (next_ix - first_ix) correct = np.int64(0) tied = np.int64(0) for i in range(first_ix, next_ix): rank, count = times_to_compare.rank(pred[i]) correct += rank tied += count return pairs, correct, tied, next_ix
handle_pairs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(truth, pred, first_ix, times_to_compare):\n next_ix = first_ix\n while next_ix < len(truth) and truth[next_ix] == truth[first_ix]:\n next_ix += 1\n pairs = len(times_to_compare) * (next_ix - first_ix)\n correct = np.int64(0)\n tied = np.int64(0)\n for i in range(first_ix, next_ix):\n rank, count = times_to_compare.rank(pred[i])\n correct += rank\n tied += count\n return pairs, correct, tied, next_ix\n", "role": "user" }, { "content": "handle_pairs", "role": "assistant" } ]
def x(event_times, predicted_event_times, event_observed): num_pairs = 0.0 num_correct = 0.0 num_tied = 0.0 for a, time_a in enumerate(event_times): pred_a = predicted_event_times[a] event_a = event_observed[a] for b in range(a + 1, len(event_times)): time_b = event_times[b] pred_b = predicted_event_times[b] event_b = event_observed[b] if _valid_comparison(time_a, time_b, event_a, event_b): num_pairs += 1.0 crct, ties = _concordance_value(time_a, time_b, pred_a, pred_b, event_a, event_b) num_correct += crct num_tied += ties return num_correct, num_tied, num_pairs
naive_concordance_summary_statistics
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(event_times, predicted_event_times, event_observed):\n num_pairs = 0.0\n num_correct = 0.0\n num_tied = 0.0\n for a, time_a in enumerate(event_times):\n pred_a = predicted_event_times[a]\n event_a = event_observed[a]\n for b in range(a + 1, len(event_times)):\n time_b = event_times[b]\n pred_b = predicted_event_times[b]\n event_b = event_observed[b]\n if _valid_comparison(time_a, time_b, event_a, event_b):\n num_pairs += 1.0\n crct, ties = _concordance_value(time_a, time_b, pred_a,\n pred_b, event_a, event_b)\n num_correct += crct\n num_tied += ties\n return num_correct, num_tied, num_pairs\n", "role": "user" }, { "content": "naive_concordance_summary_statistics", "role": "assistant" } ]
def x(time_a, time_b, event_a, event_b): if time_a == time_b: return event_a != event_b if event_a and event_b: return True if event_a and time_a < time_b: return True if event_b and time_b < time_a: return True return False
valid_comparison
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(time_a, time_b, event_a, event_b):\n if time_a == time_b:\n return event_a != event_b\n if event_a and event_b:\n return True\n if event_a and time_a < time_b:\n return True\n if event_b and time_b < time_a:\n return True\n return False\n", "role": "user" }, { "content": "valid_comparison", "role": "assistant" } ]
def x(self, X, ancillary_X=None, p=0.5): exp_mu_, sigma_ = self._prep_inputs_for_prediction_and_return_scores(X, ancillary_X) return pd.DataFrame(exp_mu_ * np.exp(np.sqrt(2) * sigma_ * erfinv(2 * p - 1)), index=_get_index(X))
predict_percentile
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, ancillary_X=None, p=0.5):\n exp_mu_, sigma_ = self._prep_inputs_for_prediction_and_return_scores(X,\n ancillary_X)\n return pd.DataFrame(exp_mu_ * np.exp(np.sqrt(2) * sigma_ * erfinv(2 * p -\n 1)), index=_get_index(X))\n", "role": "user" }, { "content": "predict_percentile", "role": "assistant" } ]
def x(self, X, ancillary_X=None): exp_mu_, _ = self._prep_inputs_for_prediction_and_return_scores(X, ancillary_X) return pd.DataFrame(exp_mu_, index=_get_index(X))
predict_median
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, ancillary_X=None):\n exp_mu_, _ = self._prep_inputs_for_prediction_and_return_scores(X,\n ancillary_X)\n return pd.DataFrame(exp_mu_, index=_get_index(X))\n", "role": "user" }, { "content": "predict_median", "role": "assistant" } ]
def x(self, X, ancillary_X=None): exp_mu_, sigma_ = self._prep_inputs_for_prediction_and_return_scores(X, ancillary_X) return pd.DataFrame(exp_mu_ * np.exp(sigma_ ** 2 / 2), index=_get_index(X))
predict_expectation
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, ancillary_X=None):\n exp_mu_, sigma_ = self._prep_inputs_for_prediction_and_return_scores(X,\n ancillary_X)\n return pd.DataFrame(exp_mu_ * np.exp(sigma_ ** 2 / 2), index=_get_index(X))\n", "role": "user" }, { "content": "predict_expectation", "role": "assistant" } ]
def x(self, X, times=None, ancillary_X=None): import numpy as np times = coalesce(times, self.timeline, np.unique(self.durations)) exp_mu_, sigma_ = self._prep_inputs_for_prediction_and_return_scores(X, ancillary_X) mu_ = np.log(exp_mu_) Z = np.subtract.outer(np.log(times), mu_) / sigma_ return pd.DataFrame(-logsf(Z), columns=_get_index(X), index=times)
predict_cumulative_hazard
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, times=None, ancillary_X=None):\n import numpy as np\n times = coalesce(times, self.timeline, np.unique(self.durations))\n exp_mu_, sigma_ = self._prep_inputs_for_prediction_and_return_scores(X,\n ancillary_X)\n mu_ = np.log(exp_mu_)\n Z = np.subtract.outer(np.log(times), mu_) / sigma_\n return pd.DataFrame(-logsf(Z), columns=_get_index(X), index=times)\n", "role": "user" }, { "content": "predict_cumulative_hazard", "role": "assistant" } ]
def x(self, other): self_estimate = getattr(self, self._estimate_name) other_estimate = getattr(other, other._estimate_name) new_index = np.concatenate((other_estimate.index, self_estimate.index)) new_index = np.unique(new_index) return pd.DataFrame(self_estimate.reindex(new_index, method='ffill'). values - other_estimate.reindex(new_index, method='ffill').values, index=new_index, columns=['diff'])
subtract
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, other):\n self_estimate = getattr(self, self._estimate_name)\n other_estimate = getattr(other, other._estimate_name)\n new_index = np.concatenate((other_estimate.index, self_estimate.index))\n new_index = np.unique(new_index)\n return pd.DataFrame(self_estimate.reindex(new_index, method='ffill').\n values - other_estimate.reindex(new_index, method='ffill').values,\n index=new_index, columns=['diff'])\n", "role": "user" }, { "content": "subtract", "role": "assistant" } ]
def x(self, times): if callable(self._estimation_method): return pd.DataFrame(self._estimation_method(_to_array(times)), index=_to_array(times)).loc[times].squeeze() estimate = getattr(self, self._estimation_method) return dataframe_interpolate_at_times(estimate, times)
predict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, times):\n if callable(self._estimation_method):\n return pd.DataFrame(self._estimation_method(_to_array(times)),\n index=_to_array(times)).loc[times].squeeze()\n estimate = getattr(self, self._estimation_method)\n return dataframe_interpolate_at_times(estimate, times)\n", "role": "user" }, { "content": "predict", "role": "assistant" } ]
def x(self): age = self.survival_function_.index.values[:, None] columns = ['%s - Conditional time remaining to event' % self._label] return pd.DataFrame(qth_survival_times(self.survival_function_[self. _label] * 0.5, self.survival_function_).sort_index(ascending=False) .values, index=self.survival_function_.index, columns=columns) - age
conditional_time_to_event_
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n age = self.survival_function_.index.values[:, None]\n columns = ['%s - Conditional time remaining to event' % self._label]\n return pd.DataFrame(qth_survival_times(self.survival_function_[self.\n _label] * 0.5, self.survival_function_).sort_index(ascending=False)\n .values, index=self.survival_function_.index, columns=columns) - age\n", "role": "user" }, { "content": "conditional_time_to_event_", "role": "assistant" } ]
def x(self, transform, alpha, ci_labels): alpha2 = 1 - alpha / 2.0 z = inv_normal_cdf(alpha2) df = pd.DataFrame(index=self.timeline) gradient_of_cum_hazard_at_mle = make_jvp_reversemode(transform)(self. _fitted_parameters_, self.timeline.astype(float)) gradient_at_times = np.vstack([gradient_of_cum_hazard_at_mle(basis) for basis in np.eye(len(self._fitted_parameters_), dtype=float)]) std_cumulative_hazard = np.sqrt(np.einsum('nj,jk,nk->n', gradient_at_times.T, self.variance_matrix_, gradient_at_times.T)) if ci_labels is None: ci_labels = ['%s_upper_%g' % (self._label, 1 - alpha), '%s_lower_%g' % (self._label, 1 - alpha)] assert len(ci_labels) == 2, 'ci_labels should be a length 2 array.' df[ci_labels[0]] = transform(self._fitted_parameters_, self.timeline ) + z * std_cumulative_hazard df[ci_labels[1]] = transform(self._fitted_parameters_, self.timeline ) - z * std_cumulative_hazard return df
compute_confidence_bounds_of_transform
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, transform, alpha, ci_labels):\n alpha2 = 1 - alpha / 2.0\n z = inv_normal_cdf(alpha2)\n df = pd.DataFrame(index=self.timeline)\n gradient_of_cum_hazard_at_mle = make_jvp_reversemode(transform)(self.\n _fitted_parameters_, self.timeline.astype(float))\n gradient_at_times = np.vstack([gradient_of_cum_hazard_at_mle(basis) for\n basis in np.eye(len(self._fitted_parameters_), dtype=float)])\n std_cumulative_hazard = np.sqrt(np.einsum('nj,jk,nk->n',\n gradient_at_times.T, self.variance_matrix_, gradient_at_times.T))\n if ci_labels is None:\n ci_labels = ['%s_upper_%g' % (self._label, 1 - alpha), \n '%s_lower_%g' % (self._label, 1 - alpha)]\n assert len(ci_labels) == 2, 'ci_labels should be a length 2 array.'\n df[ci_labels[0]] = transform(self._fitted_parameters_, self.timeline\n ) + z * std_cumulative_hazard\n df[ci_labels[1]] = transform(self._fitted_parameters_, self.timeline\n ) - z * std_cumulative_hazard\n return df\n", "role": "user" }, { "content": "compute_confidence_bounds_of_transform", "role": "assistant" } ]
def x(self): ci = 1 - self.alpha lower_upper_bounds = self._compute_confidence_bounds_of_parameters() df = pd.DataFrame(index=self._fitted_parameter_names) df['coef'] = self._fitted_parameters_ df['se(coef)'] = self._compute_standard_errors().loc['se'] df['lower %g' % ci] = lower_upper_bounds['lower-bound'] df['upper %g' % ci] = lower_upper_bounds['upper-bound'] df['p'] = self._compute_p_values() with np.errstate(invalid='ignore', divide='ignore'): df['-log2(p)'] = -np.log2(df['p']) return df
summary
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n ci = 1 - self.alpha\n lower_upper_bounds = self._compute_confidence_bounds_of_parameters()\n df = pd.DataFrame(index=self._fitted_parameter_names)\n df['coef'] = self._fitted_parameters_\n df['se(coef)'] = self._compute_standard_errors().loc['se']\n df['lower %g' % ci] = lower_upper_bounds['lower-bound']\n df['upper %g' % ci] = lower_upper_bounds['upper-bound']\n df['p'] = self._compute_p_values()\n with np.errstate(invalid='ignore', divide='ignore'):\n df['-log2(p)'] = -np.log2(df['p'])\n return df\n", "role": "user" }, { "content": "summary", "role": "assistant" } ]
def x(self, decimals=2, **kwargs): justify = string_justify(18) print(self) print('{} = {}'.format(justify('number of subjects'), self. event_observed.shape[0])) print('{} = {}'.format(justify('number of events'), np.where(self. event_observed)[0].shape[0])) print('{} = {:.{prec}f}'.format(justify('log-likelihood'), self. _log_likelihood, prec=decimals)) print('{} = {}'.format(justify('hypothesis'), ', '.join('%s != %d' % ( name, iv) for name, iv in zip(self._fitted_parameter_names, self. _initial_values)))) for k, v in kwargs.items(): print('{} = {}\n'.format(justify(k), v)) print(end='\n') print('---') df = self.summary print(df.to_string(float_format=format_floats(decimals), formatters={ 'p': format_p_value(decimals), 'exp(coef)': format_exp_floats( decimals)}))
print_summary
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, decimals=2, **kwargs):\n justify = string_justify(18)\n print(self)\n print('{} = {}'.format(justify('number of subjects'), self.\n event_observed.shape[0]))\n print('{} = {}'.format(justify('number of events'), np.where(self.\n event_observed)[0].shape[0]))\n print('{} = {:.{prec}f}'.format(justify('log-likelihood'), self.\n _log_likelihood, prec=decimals))\n print('{} = {}'.format(justify('hypothesis'), ', '.join('%s != %d' % (\n name, iv) for name, iv in zip(self._fitted_parameter_names, self.\n _initial_values))))\n for k, v in kwargs.items():\n print('{} = {}\\n'.format(justify(k), v))\n print(end='\\n')\n print('---')\n df = self.summary\n print(df.to_string(float_format=format_floats(decimals), formatters={\n 'p': format_p_value(decimals), 'exp(coef)': format_exp_floats(\n decimals)}))\n", "role": "user" }, { "content": "print_summary", "role": "assistant" } ]
def x(self, durations, event_observed=None, timeline=None, label=None, alpha=None, ci_labels=None, show_progress=False, entry=None, weights= None, left_censorship=False): if left_censorship: warnings.warn( 'kwarg left_censorship is deprecated and will be removed in a future release. Please use ``.fit_left_censoring`` instead.' , DeprecationWarning) return self.fit_left_censoring(durations, event_observed, timeline, label, alpha, ci_labels, show_progress, entry, weights) self.durations = np.asarray(pass_for_numeric_dtypes_or_raise_array( durations)) check_nans_or_infs(self.durations) check_positivity(self.durations) self._censoring_type = CensoringType.RIGHT return self._fit((self.durations, None), event_observed=event_observed, timeline=timeline, label=label, alpha=alpha, ci_labels=ci_labels, show_progress=show_progress, entry=entry, weights=weights)
fit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, durations, event_observed=None, timeline=None, label=None,\n alpha=None, ci_labels=None, show_progress=False, entry=None, weights=\n None, left_censorship=False):\n if left_censorship:\n warnings.warn(\n 'kwarg left_censorship is deprecated and will be removed in a future release. Please use ``.fit_left_censoring`` instead.'\n , DeprecationWarning)\n return self.fit_left_censoring(durations, event_observed, timeline,\n label, alpha, ci_labels, show_progress, entry, weights)\n self.durations = np.asarray(pass_for_numeric_dtypes_or_raise_array(\n durations))\n check_nans_or_infs(self.durations)\n check_positivity(self.durations)\n self._censoring_type = CensoringType.RIGHT\n return self._fit((self.durations, None), event_observed=event_observed,\n timeline=timeline, label=label, alpha=alpha, ci_labels=ci_labels,\n show_progress=show_progress, entry=entry, weights=weights)\n", "role": "user" }, { "content": "fit", "role": "assistant" } ]
def x(self, durations, event_observed=None, timeline=None, label=None, alpha=None, ci_labels=None, show_progress=False, entry=None, weights=None): self.durations = np.asarray(pass_for_numeric_dtypes_or_raise_array( durations)) check_nans_or_infs(self.durations) check_positivity(self.durations) self._censoring_type = CensoringType.LEFT return self._fit((None, self.durations), event_observed=event_observed, timeline=timeline, label=label, alpha=alpha, ci_labels=ci_labels, show_progress=show_progress, entry=entry, weights=weights)
fit_left_censoring
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, durations, event_observed=None, timeline=None, label=None,\n alpha=None, ci_labels=None, show_progress=False, entry=None, weights=None):\n self.durations = np.asarray(pass_for_numeric_dtypes_or_raise_array(\n durations))\n check_nans_or_infs(self.durations)\n check_positivity(self.durations)\n self._censoring_type = CensoringType.LEFT\n return self._fit((None, self.durations), event_observed=event_observed,\n timeline=timeline, label=label, alpha=alpha, ci_labels=ci_labels,\n show_progress=show_progress, entry=entry, weights=weights)\n", "role": "user" }, { "content": "fit_left_censoring", "role": "assistant" } ]
def x(self, lower_bound, upper_bound, event_observed=None, timeline=None, label=None, alpha=None, ci_labels=None, show_progress=False, entry=None, weights=None): check_nans_or_infs(lower_bound) check_positivity(upper_bound) self.upper_bound = np.asarray(pass_for_numeric_dtypes_or_raise_array( upper_bound)) self.lower_bound = np.asarray(pass_for_numeric_dtypes_or_raise_array( lower_bound)) if (self.upper_bound < self.lower_bound).any(): raise ValueError( 'All upper_bound times must be greater than or equal to lower_bound times.' ) if event_observed is None: event_observed = self.upper_bound == self.lower_bound if ((self.lower_bound == self.upper_bound) != event_observed).any(): raise ValueError( 'For all rows, lower_bound == upper_bound if and only if event observed = 1 (uncensored). Likewise, lower_bound < upper_bound if and only if event observed = 0 (censored)' ) self._censoring_type = CensoringType.INTERVAL return self._fit((np.clip(self.lower_bound, 1e-20, 1e+25), np.clip(self .upper_bound, 1e-20, 1e+25)), event_observed=event_observed, timeline=timeline, label=label, alpha=alpha, ci_labels=ci_labels, show_progress=show_progress, entry=entry, weights=weights)
fit_interval_censoring
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, lower_bound, upper_bound, event_observed=None, timeline=None,\n label=None, alpha=None, ci_labels=None, show_progress=False, entry=None,\n weights=None):\n check_nans_or_infs(lower_bound)\n check_positivity(upper_bound)\n self.upper_bound = np.asarray(pass_for_numeric_dtypes_or_raise_array(\n upper_bound))\n self.lower_bound = np.asarray(pass_for_numeric_dtypes_or_raise_array(\n lower_bound))\n if (self.upper_bound < self.lower_bound).any():\n raise ValueError(\n 'All upper_bound times must be greater than or equal to lower_bound times.'\n )\n if event_observed is None:\n event_observed = self.upper_bound == self.lower_bound\n if ((self.lower_bound == self.upper_bound) != event_observed).any():\n raise ValueError(\n 'For all rows, lower_bound == upper_bound if and only if event observed = 1 (uncensored). Likewise, lower_bound < upper_bound if and only if event observed = 0 (censored)'\n )\n self._censoring_type = CensoringType.INTERVAL\n return self._fit((np.clip(self.lower_bound, 1e-20, 1e+25), np.clip(self\n .upper_bound, 1e-20, 1e+25)), event_observed=event_observed,\n timeline=timeline, label=label, alpha=alpha, ci_labels=ci_labels,\n show_progress=show_progress, entry=entry, weights=weights)\n", "role": "user" }, { "content": "fit_interval_censoring", "role": "assistant" } ]
def x(self, times, label=None): label = coalesce(label, self._label) return pd.Series(self._survival_function(self._fitted_parameters_, times), index=_to_array(times), name=label)
survival_function_at_times
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, times, label=None):\n label = coalesce(label, self._label)\n return pd.Series(self._survival_function(self._fitted_parameters_,\n times), index=_to_array(times), name=label)\n", "role": "user" }, { "content": "survival_function_at_times", "role": "assistant" } ]
def x(self, times, label=None): label = coalesce(label, self._label) return pd.Series(self._cumulative_density(self._fitted_parameters_, times), index=_to_array(times), name=label)
cumulative_density_at_times
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, times, label=None):\n label = coalesce(label, self._label)\n return pd.Series(self._cumulative_density(self._fitted_parameters_,\n times), index=_to_array(times), name=label)\n", "role": "user" }, { "content": "cumulative_density_at_times", "role": "assistant" } ]
def x(self, times, label=None): label = coalesce(label, self._label) return pd.Series(self._cumulative_hazard(self._fitted_parameters_, times), index=_to_array(times), name=label)
cumulative_hazard_at_times
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, times, label=None):\n label = coalesce(label, self._label)\n return pd.Series(self._cumulative_hazard(self._fitted_parameters_,\n times), index=_to_array(times), name=label)\n", "role": "user" }, { "content": "cumulative_hazard_at_times", "role": "assistant" } ]
def x(self, times, label=None): label = coalesce(label, self._label) return pd.Series(self._hazard(self._fitted_parameters_, times), index= _to_array(times), name=label)
hazard_at_times
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, times, label=None):\n label = coalesce(label, self._label)\n return pd.Series(self._hazard(self._fitted_parameters_, times), index=\n _to_array(times), name=label)\n", "role": "user" }, { "content": "hazard_at_times", "role": "assistant" } ]
def x(self): return self._compute_confidence_bounds_of_transform(self._hazard, self. alpha, self._ci_labels)
confidence_interval_hazard_
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return self._compute_confidence_bounds_of_transform(self._hazard, self.\n alpha, self._ci_labels)\n", "role": "user" }, { "content": "confidence_interval_hazard_", "role": "assistant" } ]
def x(self): return self._compute_confidence_bounds_of_transform(self. _survival_function, self.alpha, self._ci_labels)
confidence_interval_survival_function_
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return self._compute_confidence_bounds_of_transform(self.\n _survival_function, self.alpha, self._ci_labels)\n", "role": "user" }, { "content": "confidence_interval_survival_function_", "role": "assistant" } ]
def x(self): return self._compute_confidence_bounds_of_transform(self. _cumulative_density, self.alpha, self._ci_labels)
confidence_interval_cumulative_density_
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return self._compute_confidence_bounds_of_transform(self.\n _cumulative_density, self.alpha, self._ci_labels)\n", "role": "user" }, { "content": "confidence_interval_cumulative_density_", "role": "assistant" } ]
def x(self, **kwargs): set_kwargs_drawstyle(kwargs, 'default') return _plot_estimate(self, estimate=getattr(self, self._estimate_name), confidence_intervals=self.confidence_interval_, **kwargs)
plot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n set_kwargs_drawstyle(kwargs, 'default')\n return _plot_estimate(self, estimate=getattr(self, self._estimate_name),\n confidence_intervals=self.confidence_interval_, **kwargs)\n", "role": "user" }, { "content": "plot", "role": "assistant" } ]
def x(self, df, duration_col, event_col=None, ancillary_df=None, show_progress=False, timeline=None, weights_col=None, robust=False, initial_point=None, entry_col=None): self.duration_col = duration_col self._time_cols = [duration_col] self._censoring_type = CensoringType.RIGHT df = df.copy() T = pass_for_numeric_dtypes_or_raise_array(df.pop(duration_col)).astype( float) self.durations = T.copy() self._fit(self._log_likelihood_right_censoring, df, (T.values, None), event_col=event_col, ancillary_df=ancillary_df, show_progress= show_progress, timeline=timeline, weights_col=weights_col, robust= robust, initial_point=initial_point, entry_col=entry_col) return self
fit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, df, duration_col, event_col=None, ancillary_df=None,\n show_progress=False, timeline=None, weights_col=None, robust=False,\n initial_point=None, entry_col=None):\n self.duration_col = duration_col\n self._time_cols = [duration_col]\n self._censoring_type = CensoringType.RIGHT\n df = df.copy()\n T = pass_for_numeric_dtypes_or_raise_array(df.pop(duration_col)).astype(\n float)\n self.durations = T.copy()\n self._fit(self._log_likelihood_right_censoring, df, (T.values, None),\n event_col=event_col, ancillary_df=ancillary_df, show_progress=\n show_progress, timeline=timeline, weights_col=weights_col, robust=\n robust, initial_point=initial_point, entry_col=entry_col)\n return self\n", "role": "user" }, { "content": "fit", "role": "assistant" } ]
def x(self, df, lower_bound_col, upper_bound_col, event_col=None, ancillary_df=None, show_progress=False, timeline=None, weights_col=None, robust=False, initial_point=None, entry_col=None): self.lower_bound_col = lower_bound_col self.upper_bound_col = upper_bound_col self._time_cols = [lower_bound_col, upper_bound_col] self._censoring_type = CensoringType.INTERVAL df = df.copy() lower_bound = pass_for_numeric_dtypes_or_raise_array(df.pop( lower_bound_col)).astype(float) upper_bound = pass_for_numeric_dtypes_or_raise_array(df.pop( upper_bound_col)).astype(float) if event_col is None: event_col = 'E' df['E'] = lower_bound == upper_bound if ((lower_bound == upper_bound) != df[event_col]).any(): raise ValueError( 'For all rows, lower_bound == upper_bound if and only if event observed = 1 (uncensored). Likewise, lower_bound < upper_bound if and only if event observed = 0 (censored)' ) if (lower_bound > upper_bound).any(): raise ValueError( 'All upper bound measurements must be greater than or equal to lower bound measurements.' ) self.lower_bound = lower_bound self.upper_bound = upper_bound self._fit(self._log_likelihood_interval_censoring, df, (lower_bound. values, np.clip(upper_bound.values, 0, 1e+25)), event_col=event_col, ancillary_df=ancillary_df, show_progress=show_progress, timeline= timeline, weights_col=weights_col, robust=robust, initial_point= initial_point, entry_col=entry_col) return self
fit_interval_censoring
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, df, lower_bound_col, upper_bound_col, event_col=None,\n ancillary_df=None, show_progress=False, timeline=None, weights_col=None,\n robust=False, initial_point=None, entry_col=None):\n self.lower_bound_col = lower_bound_col\n self.upper_bound_col = upper_bound_col\n self._time_cols = [lower_bound_col, upper_bound_col]\n self._censoring_type = CensoringType.INTERVAL\n df = df.copy()\n lower_bound = pass_for_numeric_dtypes_or_raise_array(df.pop(\n lower_bound_col)).astype(float)\n upper_bound = pass_for_numeric_dtypes_or_raise_array(df.pop(\n upper_bound_col)).astype(float)\n if event_col is None:\n event_col = 'E'\n df['E'] = lower_bound == upper_bound\n if ((lower_bound == upper_bound) != df[event_col]).any():\n raise ValueError(\n 'For all rows, lower_bound == upper_bound if and only if event observed = 1 (uncensored). Likewise, lower_bound < upper_bound if and only if event observed = 0 (censored)'\n )\n if (lower_bound > upper_bound).any():\n raise ValueError(\n 'All upper bound measurements must be greater than or equal to lower bound measurements.'\n )\n self.lower_bound = lower_bound\n self.upper_bound = upper_bound\n self._fit(self._log_likelihood_interval_censoring, df, (lower_bound.\n values, np.clip(upper_bound.values, 0, 1e+25)), event_col=event_col,\n ancillary_df=ancillary_df, show_progress=show_progress, timeline=\n timeline, weights_col=weights_col, robust=robust, initial_point=\n initial_point, entry_col=entry_col)\n return self\n", "role": "user" }, { "content": "fit_interval_censoring", "role": "assistant" } ]
def x(self, df, duration_col=None, event_col=None, ancillary_df=None, show_progress=False, timeline=None, weights_col=None, robust=False, initial_point=None, entry_col=None): self._censoring_type = CensoringType.LEFT df = df.copy() T = pass_for_numeric_dtypes_or_raise_array(df.pop(duration_col)).astype( float) self.durations = T.copy() self._fit(self._log_likelihood_left_censoring, df, (None, T.values), event_col=event_col, ancillary_df=ancillary_df, show_progress= show_progress, timeline=timeline, weights_col=weights_col, robust= robust, initial_point=initial_point, entry_col=entry_col) return self
fit_left_censoring
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, df, duration_col=None, event_col=None, ancillary_df=None,\n show_progress=False, timeline=None, weights_col=None, robust=False,\n initial_point=None, entry_col=None):\n self._censoring_type = CensoringType.LEFT\n df = df.copy()\n T = pass_for_numeric_dtypes_or_raise_array(df.pop(duration_col)).astype(\n float)\n self.durations = T.copy()\n self._fit(self._log_likelihood_left_censoring, df, (None, T.values),\n event_col=event_col, ancillary_df=ancillary_df, show_progress=\n show_progress, timeline=timeline, weights_col=weights_col, robust=\n robust, initial_point=initial_point, entry_col=entry_col)\n return self\n", "role": "user" }, { "content": "fit_left_censoring", "role": "assistant" } ]
def x(self, Ts, E, entries, weights, Xs): import lifelines def x(model, i): param = model._fitted_parameters_[i] if param <= 0: return param return np.log(param) name = self.__class__.__name__.replace('AFT', '') uni_model = getattr(lifelines, name)() if self._censoring_type == CensoringType.RIGHT: uni_model.fit_right_censoring(Ts[0], event_observed=E, entry= entries, weights=weights) elif self._censoring_type == CensoringType.INTERVAL: uni_model.fit_interval_censoring(Ts[0], Ts[1], event_observed=E, entry=entries, weights=weights) elif self._censoring_type == CensoringType.LEFT: uni_model.fit_left_censoring(Ts[1], event_observed=E, entry=entries, weights=weights) self._ll_null_ = uni_model._log_likelihood return np.concatenate([([0] * (_X.shape[1] - 1) + [transform_ith_param( uni_model, i)]) for i, _X in enumerate(Xs)])
create_initial_point
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, Ts, E, entries, weights, Xs):\n import lifelines\n\n def x(model, i):\n param = model._fitted_parameters_[i]\n if param <= 0:\n return param\n return np.log(param)\n name = self.__class__.__name__.replace('AFT', '')\n uni_model = getattr(lifelines, name)()\n if self._censoring_type == CensoringType.RIGHT:\n uni_model.fit_right_censoring(Ts[0], event_observed=E, entry=\n entries, weights=weights)\n elif self._censoring_type == CensoringType.INTERVAL:\n uni_model.fit_interval_censoring(Ts[0], Ts[1], event_observed=E,\n entry=entries, weights=weights)\n elif self._censoring_type == CensoringType.LEFT:\n uni_model.fit_left_censoring(Ts[1], event_observed=E, entry=entries,\n weights=weights)\n self._ll_null_ = uni_model._log_likelihood\n return np.concatenate([([0] * (_X.shape[1] - 1) + [transform_ith_param(\n uni_model, i)]) for i, _X in enumerate(Xs)])\n", "role": "user" }, { "content": "create_initial_point", "role": "assistant" } ]
def x(self, X, ancillary_X=None): return self.predict_percentile(X, p=0.5, ancillary_X=ancillary_X)
predict_median
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, ancillary_X=None):\n return self.predict_percentile(X, p=0.5, ancillary_X=ancillary_X)\n", "role": "user" }, { "content": "predict_median", "role": "assistant" } ]
def x(self): if hasattr(self, '_predicted_median'): self._concordance_score_ = concordance_index(self.durations, self. _predicted_median, self.event_observed) del self._predicted_median return self._concordance_score_ return self._concordance_score_
score_
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if hasattr(self, '_predicted_median'):\n self._concordance_score_ = concordance_index(self.durations, self.\n _predicted_median, self.event_observed)\n del self._predicted_median\n return self._concordance_score_\n return self._concordance_score_\n", "role": "user" }, { "content": "score_", "role": "assistant" } ]
def x(self, columns=None, parameter=None, **errorbar_kwargs): from matplotlib import pyplot as plt set_kwargs_ax(errorbar_kwargs) ax = errorbar_kwargs.pop('ax') errorbar_kwargs.setdefault('c', 'k') errorbar_kwargs.setdefault('fmt', 's') errorbar_kwargs.setdefault('markerfacecolor', 'white') errorbar_kwargs.setdefault('markeredgewidth', 1.25) errorbar_kwargs.setdefault('elinewidth', 1.25) errorbar_kwargs.setdefault('capsize', 3) z = inv_normal_cdf(1 - self.alpha / 2) params_ = self.params_.copy() standard_errors_ = self.standard_errors_.copy() if columns is not None: params_ = params_.loc[:, columns] standard_errors_ = standard_errors_.loc[:, columns] if parameter is not None: params_ = params_.loc[parameter] standard_errors_ = standard_errors_.loc[parameter] columns = params_.index hazards = params_.loc[columns].to_frame(name='coefs') hazards['se'] = z * standard_errors_.loc[columns] if isinstance(hazards.index, pd.MultiIndex): hazards = hazards.groupby(level=0, group_keys=False).apply(lambda x: x.sort_values(by='coefs', ascending=True)) else: hazards = hazards.sort_values(by='coefs', ascending=True) yaxis_locations = list(range(len(columns))) ax.errorbar(hazards['coefs'], yaxis_locations, xerr=hazards['se'], ** errorbar_kwargs) best_ylim = ax.get_ylim() ax.vlines(0, -2, len(columns) + 1, linestyles='dashed', linewidths=1, alpha=0.65) ax.set_ylim(best_ylim) if isinstance(columns[0], tuple): tick_labels = [('%s: %s' % (c, p)) for p, c in hazards.index] else: tick_labels = [i for i in hazards.index] plt.yticks(yaxis_locations, tick_labels) plt.xlabel('log(accelerated failure rate) (%g%% CI)' % ((1 - self.alpha ) * 100)) return ax
plot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, columns=None, parameter=None, **errorbar_kwargs):\n from matplotlib import pyplot as plt\n set_kwargs_ax(errorbar_kwargs)\n ax = errorbar_kwargs.pop('ax')\n errorbar_kwargs.setdefault('c', 'k')\n errorbar_kwargs.setdefault('fmt', 's')\n errorbar_kwargs.setdefault('markerfacecolor', 'white')\n errorbar_kwargs.setdefault('markeredgewidth', 1.25)\n errorbar_kwargs.setdefault('elinewidth', 1.25)\n errorbar_kwargs.setdefault('capsize', 3)\n z = inv_normal_cdf(1 - self.alpha / 2)\n params_ = self.params_.copy()\n standard_errors_ = self.standard_errors_.copy()\n if columns is not None:\n params_ = params_.loc[:, columns]\n standard_errors_ = standard_errors_.loc[:, columns]\n if parameter is not None:\n params_ = params_.loc[parameter]\n standard_errors_ = standard_errors_.loc[parameter]\n columns = params_.index\n hazards = params_.loc[columns].to_frame(name='coefs')\n hazards['se'] = z * standard_errors_.loc[columns]\n if isinstance(hazards.index, pd.MultiIndex):\n hazards = hazards.groupby(level=0, group_keys=False).apply(lambda x:\n x.sort_values(by='coefs', ascending=True))\n else:\n hazards = hazards.sort_values(by='coefs', ascending=True)\n yaxis_locations = list(range(len(columns)))\n ax.errorbar(hazards['coefs'], yaxis_locations, xerr=hazards['se'], **\n errorbar_kwargs)\n best_ylim = ax.get_ylim()\n ax.vlines(0, -2, len(columns) + 1, linestyles='dashed', linewidths=1,\n alpha=0.65)\n ax.set_ylim(best_ylim)\n if isinstance(columns[0], tuple):\n tick_labels = [('%s: %s' % (c, p)) for p, c in hazards.index]\n else:\n tick_labels = [i for i in hazards.index]\n plt.yticks(yaxis_locations, tick_labels)\n plt.xlabel('log(accelerated failure rate) (%g%% CI)' % ((1 - self.alpha\n ) * 100))\n return ax\n", "role": "user" }, { "content": "plot", "role": "assistant" } ]
def x(self, covariates, values, plot_baseline=True, **kwargs): from matplotlib import pyplot as plt covariates = _to_list(covariates) values = _to_array(values) if len(values.shape) == 1: values = values[None, :].T if len(covariates) != values.shape[1]: raise ValueError( 'The number of covariates must equal to second dimension of the values array.' ) original_columns = self.params_.index.get_level_values(1) for covariate in covariates: if covariate not in original_columns: raise KeyError( 'covariate `%s` is not present in the original dataset' % covariate) ax = kwargs.pop('ax', None) or plt.figure().add_subplot(111) x_bar = self._norm_mean.to_frame().T X = pd.concat([x_bar] * values.shape[0]) if np.array_equal(np.eye(len(covariates)), values): X.index = [('%s=1' % c) for c in covariates] else: X.index = [', '.join('%s=%g' % (c, v) for c, v in zip(covariates, row)) for row in values] for covariate, value in zip(covariates, values.T): X[covariate] = value x_bar_anc = self._norm_mean_ancillary.to_frame().T ancillary_X = pd.concat([x_bar_anc] * values.shape[0]) for covariate, value in zip(covariates, values.T): ancillary_X[covariate] = value if self.fit_intercept: X['_intercept'] = 1.0 ancillary_X['_intercept'] = 1.0 self.predict_survival_function(X, ancillary_X=ancillary_X).plot(ax=ax, **kwargs) if plot_baseline: self.predict_survival_function(x_bar, ancillary_X=x_bar_anc).rename( columns={(0): 'baseline survival'}).plot(ax=ax, ls=':', color='k') return ax
plot_covariate_groups
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, covariates, values, plot_baseline=True, **kwargs):\n from matplotlib import pyplot as plt\n covariates = _to_list(covariates)\n values = _to_array(values)\n if len(values.shape) == 1:\n values = values[None, :].T\n if len(covariates) != values.shape[1]:\n raise ValueError(\n 'The number of covariates must equal to second dimension of the values array.'\n )\n original_columns = self.params_.index.get_level_values(1)\n for covariate in covariates:\n if covariate not in original_columns:\n raise KeyError(\n 'covariate `%s` is not present in the original dataset' %\n covariate)\n ax = kwargs.pop('ax', None) or plt.figure().add_subplot(111)\n x_bar = self._norm_mean.to_frame().T\n X = pd.concat([x_bar] * values.shape[0])\n if np.array_equal(np.eye(len(covariates)), values):\n X.index = [('%s=1' % c) for c in covariates]\n else:\n X.index = [', '.join('%s=%g' % (c, v) for c, v in zip(covariates,\n row)) for row in values]\n for covariate, value in zip(covariates, values.T):\n X[covariate] = value\n x_bar_anc = self._norm_mean_ancillary.to_frame().T\n ancillary_X = pd.concat([x_bar_anc] * values.shape[0])\n for covariate, value in zip(covariates, values.T):\n ancillary_X[covariate] = value\n if self.fit_intercept:\n X['_intercept'] = 1.0\n ancillary_X['_intercept'] = 1.0\n self.predict_survival_function(X, ancillary_X=ancillary_X).plot(ax=ax,\n **kwargs)\n if plot_baseline:\n self.predict_survival_function(x_bar, ancillary_X=x_bar_anc).rename(\n columns={(0): 'baseline survival'}).plot(ax=ax, ls=':', color='k')\n return ax\n", "role": "user" }, { "content": "plot_covariate_groups", "role": "assistant" } ]
def x(self, X, ancillary_X=None, p=0.5): alpha_, beta_ = self._prep_inputs_for_prediction_and_return_scores(X, ancillary_X) return pd.DataFrame(alpha_ * (1 / p - 1) ** beta_, index=_get_index(X))
predict_percentile
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, ancillary_X=None, p=0.5):\n alpha_, beta_ = self._prep_inputs_for_prediction_and_return_scores(X,\n ancillary_X)\n return pd.DataFrame(alpha_ * (1 / p - 1) ** beta_, index=_get_index(X))\n", "role": "user" }, { "content": "predict_percentile", "role": "assistant" } ]
def x(self, X, ancillary_X=None): alpha_, beta_ = self._prep_inputs_for_prediction_and_return_scores(X, ancillary_X) v = alpha_ * np.pi / beta_ / np.sin(np.pi / beta_) v = np.where(beta_ > 1, v, np.nan) return pd.DataFrame(v, index=_get_index(X))
predict_expectation
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, ancillary_X=None):\n alpha_, beta_ = self._prep_inputs_for_prediction_and_return_scores(X,\n ancillary_X)\n v = alpha_ * np.pi / beta_ / np.sin(np.pi / beta_)\n v = np.where(beta_ > 1, v, np.nan)\n return pd.DataFrame(v, index=_get_index(X))\n", "role": "user" }, { "content": "predict_expectation", "role": "assistant" } ]
def x(self, X, times=None, ancillary_X=None): times = coalesce(times, self.timeline, np.unique(self.durations)) alpha_, beta_ = self._prep_inputs_for_prediction_and_return_scores(X, ancillary_X) return pd.DataFrame(np.log1p(np.outer(times, 1 / alpha_) ** beta_), columns=_get_index(X), index=times)
predict_cumulative_hazard
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, X, times=None, ancillary_X=None):\n times = coalesce(times, self.timeline, np.unique(self.durations))\n alpha_, beta_ = self._prep_inputs_for_prediction_and_return_scores(X,\n ancillary_X)\n return pd.DataFrame(np.log1p(np.outer(times, 1 / alpha_) ** beta_),\n columns=_get_index(X), index=times)\n", "role": "user" }, { "content": "predict_cumulative_hazard", "role": "assistant" } ]
def x(self, durations, event_observed=None, timeline=None, entry=None, label='NA_estimate', alpha=None, ci_labels=None, weights=None): self._censoring_type = CensoringType.RIGHT check_nans_or_infs(durations) if event_observed is not None: check_nans_or_infs(event_observed) if weights is not None: if (weights.astype(int) != weights).any(): warnings.warn( """It looks like your weights are not integers, possibly prospenity scores then? It's important to know that the naive variance estimates of the coefficients are biased. Instead use Monte Carlo to estimate the variances. See paper "Variance estimation when using inverse probability of treatment weighting (IPTW) with survival analysis" or "Adjusted Kaplan-Meier estimator and log-rank test with inverse probability of treatment weighting for survival data." """ , StatisticalWarning) v = _preprocess_inputs(durations, event_observed, timeline, entry, weights) (self.durations, self.event_observed, self.timeline, self.entry, self. event_table) = v cumulative_hazard_, cumulative_sq_ = _additive_estimate(self. event_table, self.timeline, self._additive_f, self._variance_f, False) self._label = label self.cumulative_hazard_ = pd.DataFrame(cumulative_hazard_, columns=[ self._label]) self.confidence_interval_ = self._bounds(cumulative_sq_[:, None], alpha if alpha else self.alpha, ci_labels) self._cumulative_sq = cumulative_sq_ self._estimation_method = 'cumulative_hazard_' self._estimate_name = 'cumulative_hazard_' self._predict_label = label self._update_docstrings() self.plot_cumulative_hazard = self.plot return self
fit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, durations, event_observed=None, timeline=None, entry=None,\n label='NA_estimate', alpha=None, ci_labels=None, weights=None):\n self._censoring_type = CensoringType.RIGHT\n check_nans_or_infs(durations)\n if event_observed is not None:\n check_nans_or_infs(event_observed)\n if weights is not None:\n if (weights.astype(int) != weights).any():\n warnings.warn(\n \"\"\"It looks like your weights are not integers, possibly prospenity scores then?\n It's important to know that the naive variance estimates of the coefficients are biased. Instead use Monte Carlo to\n estimate the variances. See paper \"Variance estimation when using inverse probability of treatment weighting (IPTW) with survival analysis\"\n or \"Adjusted Kaplan-Meier estimator and log-rank test with inverse probability of treatment weighting for survival data.\"\n \"\"\"\n , StatisticalWarning)\n v = _preprocess_inputs(durations, event_observed, timeline, entry, weights)\n (self.durations, self.event_observed, self.timeline, self.entry, self.\n event_table) = v\n cumulative_hazard_, cumulative_sq_ = _additive_estimate(self.\n event_table, self.timeline, self._additive_f, self._variance_f, False)\n self._label = label\n self.cumulative_hazard_ = pd.DataFrame(cumulative_hazard_, columns=[\n self._label])\n self.confidence_interval_ = self._bounds(cumulative_sq_[:, None], alpha if\n alpha else self.alpha, ci_labels)\n self._cumulative_sq = cumulative_sq_\n self._estimation_method = 'cumulative_hazard_'\n self._estimate_name = 'cumulative_hazard_'\n self._predict_label = label\n self._update_docstrings()\n self.plot_cumulative_hazard = self.plot\n return self\n", "role": "user" }, { "content": "fit", "role": "assistant" } ]
def x(self, bandwidth): timeline = self.timeline cumulative_hazard_name = self.cumulative_hazard_.columns[0] hazard_name = 'differenced-' + cumulative_hazard_name hazard_ = self.cumulative_hazard_.diff().fillna(self.cumulative_hazard_ .iloc[0]) C = (hazard_[cumulative_hazard_name] != 0.0).values return pd.DataFrame(1.0 / bandwidth * np.dot(epanechnikov_kernel( timeline[:, None], timeline[C][None, :], bandwidth), hazard_.values [C, :]), columns=[hazard_name], index=timeline)
smoothed_hazard_
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, bandwidth):\n timeline = self.timeline\n cumulative_hazard_name = self.cumulative_hazard_.columns[0]\n hazard_name = 'differenced-' + cumulative_hazard_name\n hazard_ = self.cumulative_hazard_.diff().fillna(self.cumulative_hazard_\n .iloc[0])\n C = (hazard_[cumulative_hazard_name] != 0.0).values\n return pd.DataFrame(1.0 / bandwidth * np.dot(epanechnikov_kernel(\n timeline[:, None], timeline[C][None, :], bandwidth), hazard_.values\n [C, :]), columns=[hazard_name], index=timeline)\n", "role": "user" }, { "content": "smoothed_hazard_", "role": "assistant" } ]
def x(self, bandwidth, hazard_=None): if hazard_ is None: hazard_ = self.smoothed_hazard_(bandwidth).values[:, 0] timeline = self.timeline z = inv_normal_cdf(1 - self.alpha / 2) self._cumulative_sq.iloc[0] = 0 var_hazard_ = self._cumulative_sq.diff().fillna(self._cumulative_sq.iloc[0] ) C = var_hazard_.values != 0.0 std_hazard_ = np.sqrt(1.0 / bandwidth ** 2 * np.dot(epanechnikov_kernel (timeline[:, None], timeline[C][None, :], bandwidth) ** 2, var_hazard_.values[C])) values = {self.ci_labels[0]: hazard_ * np.exp(z * std_hazard_ / hazard_ ), self.ci_labels[1]: hazard_ * np.exp(-z * std_hazard_ / hazard_)} return pd.DataFrame(values, index=timeline)
smoothed_hazard_confidence_intervals_
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, bandwidth, hazard_=None):\n if hazard_ is None:\n hazard_ = self.smoothed_hazard_(bandwidth).values[:, 0]\n timeline = self.timeline\n z = inv_normal_cdf(1 - self.alpha / 2)\n self._cumulative_sq.iloc[0] = 0\n var_hazard_ = self._cumulative_sq.diff().fillna(self._cumulative_sq.iloc[0]\n )\n C = var_hazard_.values != 0.0\n std_hazard_ = np.sqrt(1.0 / bandwidth ** 2 * np.dot(epanechnikov_kernel\n (timeline[:, None], timeline[C][None, :], bandwidth) ** 2,\n var_hazard_.values[C]))\n values = {self.ci_labels[0]: hazard_ * np.exp(z * std_hazard_ / hazard_\n ), self.ci_labels[1]: hazard_ * np.exp(-z * std_hazard_ / hazard_)}\n return pd.DataFrame(values, index=timeline)\n", "role": "user" }, { "content": "smoothed_hazard_confidence_intervals_", "role": "assistant" } ]
def x(values): if len(values) == 1: return values tree = np.empty_like(values) last_full_row = int(np.log2(len(values) + 1) - 1) len_ragged_row = len(values) - (2 ** (last_full_row + 1) - 1) if len_ragged_row > 0: bottom_row_ix = np.s_[:2 * len_ragged_row:2] tree[-len_ragged_row:] = values[bottom_row_ix] values = np.delete(values, bottom_row_ix) values_start = 0 values_space = 2 values_len = 2 ** last_full_row while values_start < len(values): tree[values_len - 1:2 * values_len - 1] = values[values_start:: values_space] values_start += int(values_space / 2) values_space *= 2 values_len = int(values_len / 2) return tree
treeify
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(values):\n if len(values) == 1:\n return values\n tree = np.empty_like(values)\n last_full_row = int(np.log2(len(values) + 1) - 1)\n len_ragged_row = len(values) - (2 ** (last_full_row + 1) - 1)\n if len_ragged_row > 0:\n bottom_row_ix = np.s_[:2 * len_ragged_row:2]\n tree[-len_ragged_row:] = values[bottom_row_ix]\n values = np.delete(values, bottom_row_ix)\n values_start = 0\n values_space = 2\n values_len = 2 ** last_full_row\n while values_start < len(values):\n tree[values_len - 1:2 * values_len - 1] = values[values_start::\n values_space]\n values_start += int(values_space / 2)\n values_space *= 2\n values_len = int(values_len / 2)\n return tree\n", "role": "user" }, { "content": "treeify", "role": "assistant" } ]
def x(self, value): i = 0 n = len(self._tree) while i < n: cur = self._tree[i] self._counts[i] += 1 if value < cur: i = 2 * i + 1 elif value > cur: i = 2 * i + 2 else: return raise ValueError( 'Value %s not contained in tree.Also, the counts are now messed up.' % value)
insert
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, value):\n i = 0\n n = len(self._tree)\n while i < n:\n cur = self._tree[i]\n self._counts[i] += 1\n if value < cur:\n i = 2 * i + 1\n elif value > cur:\n i = 2 * i + 2\n else:\n return\n raise ValueError(\n 'Value %s not contained in tree.Also, the counts are now messed up.' %\n value)\n", "role": "user" }, { "content": "insert", "role": "assistant" } ]
def x(self, value): i = 0 n = len(self._tree) rank = 0 count = 0 while i < n: cur = self._tree[i] if value < cur: i = 2 * i + 1 continue elif value > cur: rank += self._counts[i] nexti = 2 * i + 2 if nexti < n: rank -= self._counts[nexti] i = nexti continue else: return rank, count else: count = self._counts[i] lefti = 2 * i + 1 if lefti < n: nleft = self._counts[lefti] count -= nleft rank += nleft righti = lefti + 1 if righti < n: count -= self._counts[righti] return rank, count return rank, count
rank
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, value):\n i = 0\n n = len(self._tree)\n rank = 0\n count = 0\n while i < n:\n cur = self._tree[i]\n if value < cur:\n i = 2 * i + 1\n continue\n elif value > cur:\n rank += self._counts[i]\n nexti = 2 * i + 2\n if nexti < n:\n rank -= self._counts[nexti]\n i = nexti\n continue\n else:\n return rank, count\n else:\n count = self._counts[i]\n lefti = 2 * i + 1\n if lefti < n:\n nleft = self._counts[lefti]\n count -= nleft\n rank += nleft\n righti = lefti + 1\n if righti < n:\n count -= self._counts[righti]\n return rank, count\n return rank, count\n", "role": "user" }, { "content": "rank", "role": "assistant" } ]
def x(model, **plot_kwargs): from lifelines.utils import qth_survival_times from lifelines import KaplanMeierFitter set_kwargs_ax(plot_kwargs) ax = plot_kwargs.pop('ax') dist = get_distribution_name_of_lifelines_model(model) dist_object = create_scipy_stats_model_from_lifelines_model(model) COL_EMP = 'empirical quantiles' COL_THEO = 'fitted %s quantiles' % dist if model._censoring_type == CensoringType.LEFT: kmf = KaplanMeierFitter().fit_left_censoring(model.durations, model .event_observed, label=COL_EMP) elif model._censoring_type == CensoringType.RIGHT: kmf = KaplanMeierFitter().fit_right_censoring(model.durations, model.event_observed, label=COL_EMP) elif model._censoring_type == CensoringType.INTERVAL: raise NotImplementedError() q = np.unique(kmf.cumulative_density_.values[:, 0]) quantiles = qth_survival_times(q, kmf.cumulative_density_, cdf=True) quantiles[COL_THEO] = dist_object.ppf(q) quantiles = quantiles.replace([-np.inf, 0, np.inf], np.nan).dropna() max_, min_ = quantiles[COL_EMP].max(), quantiles[COL_EMP].min() quantiles.plot.scatter(COL_THEO, COL_EMP, c='none', edgecolor='k', lw= 0.5, ax=ax) ax.plot([min_, max_], [min_, max_], c='k', ls=':', lw=1.0) ax.set_ylim(min_, max_) ax.set_xlim(min_, max_) return ax
qq_plot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model, **plot_kwargs):\n from lifelines.utils import qth_survival_times\n from lifelines import KaplanMeierFitter\n set_kwargs_ax(plot_kwargs)\n ax = plot_kwargs.pop('ax')\n dist = get_distribution_name_of_lifelines_model(model)\n dist_object = create_scipy_stats_model_from_lifelines_model(model)\n COL_EMP = 'empirical quantiles'\n COL_THEO = 'fitted %s quantiles' % dist\n if model._censoring_type == CensoringType.LEFT:\n kmf = KaplanMeierFitter().fit_left_censoring(model.durations, model\n .event_observed, label=COL_EMP)\n elif model._censoring_type == CensoringType.RIGHT:\n kmf = KaplanMeierFitter().fit_right_censoring(model.durations,\n model.event_observed, label=COL_EMP)\n elif model._censoring_type == CensoringType.INTERVAL:\n raise NotImplementedError()\n q = np.unique(kmf.cumulative_density_.values[:, 0])\n quantiles = qth_survival_times(q, kmf.cumulative_density_, cdf=True)\n quantiles[COL_THEO] = dist_object.ppf(q)\n quantiles = quantiles.replace([-np.inf, 0, np.inf], np.nan).dropna()\n max_, min_ = quantiles[COL_EMP].max(), quantiles[COL_EMP].min()\n quantiles.plot.scatter(COL_THEO, COL_EMP, c='none', edgecolor='k', lw=\n 0.5, ax=ax)\n ax.plot([min_, max_], [min_, max_], c='k', ls=':', lw=1.0)\n ax.set_ylim(min_, max_)\n ax.set_xlim(min_, max_)\n return ax\n", "role": "user" }, { "content": "qq_plot", "role": "assistant" } ]
def x(ax, sides): for side in sides: ax.spines[side].set_visible(False) return ax
remove_spines
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ax, sides):\n for side in sides:\n ax.spines[side].set_visible(False)\n return ax\n", "role": "user" }, { "content": "remove_spines", "role": "assistant" } ]
def x(ax, sides, dists): for side, dist in zip(sides, dists): ax.spines[side].set_position(('axes', dist)) return ax
move_spines
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ax, sides, dists):\n for side, dist in zip(sides, dists):\n ax.spines[side].set_position(('axes', dist))\n return ax\n", "role": "user" }, { "content": "move_spines", "role": "assistant" } ]
def x(ax, x=False, y=False): if x: ax.xaxis.set_ticks_position('none') if y: ax.yaxis.set_ticks_position('none') return ax
remove_ticks
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ax, x=False, y=False):\n if x:\n ax.xaxis.set_ticks_position('none')\n if y:\n ax.yaxis.set_ticks_position('none')\n return ax\n", "role": "user" }, { "content": "remove_ticks", "role": "assistant" } ]
def x(*fitters, **kwargs): from matplotlib import pyplot as plt ax = kwargs.get('ax', None) if ax is None: ax = plt.gca() fig = kwargs.get('fig', None) if fig is None: fig = plt.gcf() if 'labels' not in kwargs: labels = [f._label for f in fitters] else: labels = kwargs['labels'] if labels is None: labels = [None] * len(fitters) ax2 = plt.twiny(ax=ax) ax2_ypos = -0.15 * 6.0 / fig.get_figheight() move_spines(ax2, ['bottom'], [ax2_ypos]) remove_spines(ax2, ['top', 'right', 'bottom', 'left']) ax2.xaxis.tick_bottom() ax2.set_xlim(ax.get_xlim()) ax2.set_xticks(ax.get_xticks()) remove_ticks(ax2, x=True, y=True) ticklabels = [] for tick in ax2.get_xticks(): lbl = '' for f, l in zip(fitters, labels): if tick == ax2.get_xticks()[0] and l is not None: if is_latex_enabled(): s = '\n{}\\quad'.format(l) + '{}' else: s = '\n{} '.format(l) + '{}' else: s = '\n{}' lbl += s.format(f.durations[f.durations >= tick].shape[0]) ticklabels.append(lbl.strip()) ax2.set_xticklabels(ticklabels, ha='right') ax2.xaxis.set_label_coords(0, ax2_ypos) ax2.set_xlabel('At risk') plt.tight_layout() return ax
add_at_risk_counts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*fitters, **kwargs):\n from matplotlib import pyplot as plt\n ax = kwargs.get('ax', None)\n if ax is None:\n ax = plt.gca()\n fig = kwargs.get('fig', None)\n if fig is None:\n fig = plt.gcf()\n if 'labels' not in kwargs:\n labels = [f._label for f in fitters]\n else:\n labels = kwargs['labels']\n if labels is None:\n labels = [None] * len(fitters)\n ax2 = plt.twiny(ax=ax)\n ax2_ypos = -0.15 * 6.0 / fig.get_figheight()\n move_spines(ax2, ['bottom'], [ax2_ypos])\n remove_spines(ax2, ['top', 'right', 'bottom', 'left'])\n ax2.xaxis.tick_bottom()\n ax2.set_xlim(ax.get_xlim())\n ax2.set_xticks(ax.get_xticks())\n remove_ticks(ax2, x=True, y=True)\n ticklabels = []\n for tick in ax2.get_xticks():\n lbl = ''\n for f, l in zip(fitters, labels):\n if tick == ax2.get_xticks()[0] and l is not None:\n if is_latex_enabled():\n s = '\\n{}\\\\quad'.format(l) + '{}'\n else:\n s = '\\n{} '.format(l) + '{}'\n else:\n s = '\\n{}'\n lbl += s.format(f.durations[f.durations >= tick].shape[0])\n ticklabels.append(lbl.strip())\n ax2.set_xticklabels(ticklabels, ha='right')\n ax2.xaxis.set_label_coords(0, ax2_ypos)\n ax2.set_xlabel('At risk')\n plt.tight_layout()\n return ax\n", "role": "user" }, { "content": "add_at_risk_counts", "role": "assistant" } ]
def x(durations, event_observed=None, entry=None, left_truncated=False, sort_by_duration=True, event_observed_color='#A60628', event_censored_color='#348ABD', **kwargs): set_kwargs_ax(kwargs) ax = kwargs.pop('ax') N = durations.shape[0] if N > 80: warnings.warn( 'For less visual clutter, you may want to subsample to less than 80 individuals.' ) if event_observed is None: event_observed = np.ones(N, dtype=bool) if entry is None: entry = np.zeros(N) assert durations.shape[0] == N assert event_observed.shape[0] == N if sort_by_duration: ix = np.argsort(entry + durations, 0) durations = durations[ix] event_observed = event_observed[ix] entry = entry[ix] for i in range(N): c = event_observed_color if event_observed[i] else event_censored_color ax.hlines(i, entry[i], entry[i] + durations[i], color=c, lw=1.5) if left_truncated: ax.hlines(i, 0, entry[i], color=c, lw=1.0, linestyle='--') m = '' if not event_observed[i] else 'o' ax.scatter(entry[i] + durations[i], i, color=c, marker=m, s=10) ax.set_ylim(-0.5, N) return ax
plot_lifetimes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(durations, event_observed=None, entry=None, left_truncated=False,\n sort_by_duration=True, event_observed_color='#A60628',\n event_censored_color='#348ABD', **kwargs):\n set_kwargs_ax(kwargs)\n ax = kwargs.pop('ax')\n N = durations.shape[0]\n if N > 80:\n warnings.warn(\n 'For less visual clutter, you may want to subsample to less than 80 individuals.'\n )\n if event_observed is None:\n event_observed = np.ones(N, dtype=bool)\n if entry is None:\n entry = np.zeros(N)\n assert durations.shape[0] == N\n assert event_observed.shape[0] == N\n if sort_by_duration:\n ix = np.argsort(entry + durations, 0)\n durations = durations[ix]\n event_observed = event_observed[ix]\n entry = entry[ix]\n for i in range(N):\n c = event_observed_color if event_observed[i] else event_censored_color\n ax.hlines(i, entry[i], entry[i] + durations[i], color=c, lw=1.5)\n if left_truncated:\n ax.hlines(i, 0, entry[i], color=c, lw=1.0, linestyle='--')\n m = '' if not event_observed[i] else 'o'\n ax.scatter(entry[i] + durations[i], i, color=c, marker=m, s=10)\n ax.set_ylim(-0.5, N)\n return ax\n", "role": "user" }, { "content": "plot_lifetimes", "role": "assistant" } ]
def x(cls, loc=None, iloc=None, show_censors=False, censor_styles=None, ** kwargs): def x(s): return np.log(-np.log(s)) if loc is not None and iloc is not None: raise ValueError('Cannot set both loc and iloc in call to .plot().') if censor_styles is None: censor_styles = {} set_kwargs_ax(kwargs) set_kwargs_color(kwargs) set_kwargs_drawstyle(kwargs) kwargs['logx'] = True dataframe_slicer = create_dataframe_slicer(iloc, loc) ax = kwargs['ax'] colour = kwargs['c'] if show_censors and cls.event_table['censored'].sum() > 0: cs = {'marker': '+', 'ms': 12, 'mew': 1} cs.update(censor_styles) times = dataframe_slicer(cls.event_table.loc[cls.event_table[ 'censored'] > 0]).index.values.astype(float) v = cls.predict(times) ax.plot(times, loglog(v), linestyle='None', color=colour, **cs) dataframe_slicer(loglog(cls.survival_function_)).plot(**kwargs) ax.set_xlabel('log(timeline)') ax.set_ylabel('log(-log(survival_function_))') return ax
plot_loglogs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, loc=None, iloc=None, show_censors=False, censor_styles=None, **\n kwargs):\n\n def x(s):\n return np.log(-np.log(s))\n if loc is not None and iloc is not None:\n raise ValueError('Cannot set both loc and iloc in call to .plot().')\n if censor_styles is None:\n censor_styles = {}\n set_kwargs_ax(kwargs)\n set_kwargs_color(kwargs)\n set_kwargs_drawstyle(kwargs)\n kwargs['logx'] = True\n dataframe_slicer = create_dataframe_slicer(iloc, loc)\n ax = kwargs['ax']\n colour = kwargs['c']\n if show_censors and cls.event_table['censored'].sum() > 0:\n cs = {'marker': '+', 'ms': 12, 'mew': 1}\n cs.update(censor_styles)\n times = dataframe_slicer(cls.event_table.loc[cls.event_table[\n 'censored'] > 0]).index.values.astype(float)\n v = cls.predict(times)\n ax.plot(times, loglog(v), linestyle='None', color=colour, **cs)\n dataframe_slicer(loglog(cls.survival_function_)).plot(**kwargs)\n ax.set_xlabel('log(timeline)')\n ax.set_ylabel('log(-log(survival_function_))')\n return ax\n", "role": "user" }, { "content": "plot_loglogs", "role": "assistant" } ]
def x(cls, estimate=None, confidence_intervals=None, loc=None, iloc=None, show_censors=False, censor_styles=None, ci_legend=False, ci_force_lines =False, ci_alpha=0.25, ci_show=True, at_risk_counts=False, **kwargs): plot_estimate_config = PlotEstimateConfig(cls, estimate, confidence_intervals, loc, iloc, show_censors, censor_styles, **kwargs) dataframe_slicer = create_dataframe_slicer(iloc, loc) if show_censors and cls.event_table['censored'].sum() > 0: cs = {'marker': '+', 'ms': 12, 'mew': 1} cs.update(plot_estimate_config.censor_styles) times = dataframe_slicer(cls.event_table.loc[cls.event_table[ 'censored'] > 0]).index.values.astype(float) v = cls.predict(times) plot_estimate_config.ax.plot(times, v, linestyle='None', color= plot_estimate_config.colour, **cs) dataframe_slicer(plot_estimate_config.estimate_).rename(columns=lambda _: plot_estimate_config.kwargs.pop('label')).plot(** plot_estimate_config.kwargs) if ci_show: if ci_force_lines: dataframe_slicer(plot_estimate_config.confidence_interval_).plot( linestyle='-', linewidth=1, color=[plot_estimate_config. colour], legend=ci_legend, drawstyle=plot_estimate_config. kwargs['drawstyle'], ax=plot_estimate_config.ax, alpha=0.6) else: x = dataframe_slicer(plot_estimate_config.confidence_interval_ ).index.values.astype(float) lower = dataframe_slicer(plot_estimate_config. confidence_interval_.filter(like='lower')).values[:, 0] upper = dataframe_slicer(plot_estimate_config. confidence_interval_.filter(like='upper')).values[:, 0] if plot_estimate_config.kwargs['drawstyle'] == 'default': step = None elif plot_estimate_config.kwargs['drawstyle'].startswith('step'): step = plot_estimate_config.kwargs['drawstyle'].replace( 'steps-', '') plot_estimate_config.ax.fill_between(x, lower, upper, alpha= ci_alpha, color=plot_estimate_config.colour, linewidth=1.0, step=step) if at_risk_counts: add_at_risk_counts(cls, ax=plot_estimate_config.ax) return plot_estimate_config.ax
plot_estimate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, estimate=None, confidence_intervals=None, loc=None, iloc=None,\n show_censors=False, censor_styles=None, ci_legend=False, ci_force_lines\n =False, ci_alpha=0.25, ci_show=True, at_risk_counts=False, **kwargs):\n plot_estimate_config = PlotEstimateConfig(cls, estimate,\n confidence_intervals, loc, iloc, show_censors, censor_styles, **kwargs)\n dataframe_slicer = create_dataframe_slicer(iloc, loc)\n if show_censors and cls.event_table['censored'].sum() > 0:\n cs = {'marker': '+', 'ms': 12, 'mew': 1}\n cs.update(plot_estimate_config.censor_styles)\n times = dataframe_slicer(cls.event_table.loc[cls.event_table[\n 'censored'] > 0]).index.values.astype(float)\n v = cls.predict(times)\n plot_estimate_config.ax.plot(times, v, linestyle='None', color=\n plot_estimate_config.colour, **cs)\n dataframe_slicer(plot_estimate_config.estimate_).rename(columns=lambda\n _: plot_estimate_config.kwargs.pop('label')).plot(**\n plot_estimate_config.kwargs)\n if ci_show:\n if ci_force_lines:\n dataframe_slicer(plot_estimate_config.confidence_interval_).plot(\n linestyle='-', linewidth=1, color=[plot_estimate_config.\n colour], legend=ci_legend, drawstyle=plot_estimate_config.\n kwargs['drawstyle'], ax=plot_estimate_config.ax, alpha=0.6)\n else:\n x = dataframe_slicer(plot_estimate_config.confidence_interval_\n ).index.values.astype(float)\n lower = dataframe_slicer(plot_estimate_config.\n confidence_interval_.filter(like='lower')).values[:, 0]\n upper = dataframe_slicer(plot_estimate_config.\n confidence_interval_.filter(like='upper')).values[:, 0]\n if plot_estimate_config.kwargs['drawstyle'] == 'default':\n step = None\n elif plot_estimate_config.kwargs['drawstyle'].startswith('step'):\n step = plot_estimate_config.kwargs['drawstyle'].replace(\n 'steps-', '')\n plot_estimate_config.ax.fill_between(x, lower, upper, alpha=\n ci_alpha, color=plot_estimate_config.colour, linewidth=1.0,\n step=step)\n if at_risk_counts:\n add_at_risk_counts(cls, ax=plot_estimate_config.ax)\n return plot_estimate_config.ax\n", "role": "user" }, { "content": "plot_estimate", "role": "assistant" } ]
def x(filename, destination='zh-CN', mix=True): def x(t): return Translator().translate(t, dest=destination).text doc = Document(filename) for p in doc.paragraphs: txd = tx(p.text) p.text = p.text + ('\n' + txd if mix else '') for table in doc.tables: for row in table.rows: for cell in row.cells: txd = tx(cell.text) p.text = cell.text + ('\n' + txd if mix else '') f = filename.replace('.doc', destination.lower() + '.doc') doc.save(f)
translate_doc
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(filename, destination='zh-CN', mix=True):\n\n def x(t):\n return Translator().translate(t, dest=destination).text\n doc = Document(filename)\n for p in doc.paragraphs:\n txd = tx(p.text)\n p.text = p.text + ('\\n' + txd if mix else '')\n for table in doc.tables:\n for row in table.rows:\n for cell in row.cells:\n txd = tx(cell.text)\n p.text = cell.text + ('\\n' + txd if mix else '')\n f = filename.replace('.doc', destination.lower() + '.doc')\n doc.save(f)\n", "role": "user" }, { "content": "translate_doc", "role": "assistant" } ]
def x(self, text, dest='en', src='auto'): dest = dest.lower().split('_', 1)[0] src = src.lower().split('_', 1)[0] if src != 'auto' and src not in LANGUAGES: if src in SPECIAL_CASES: src = SPECIAL_CASES[src] elif src in LANGCODES: src = LANGCODES[src] else: raise ValueError('invalid source language') if dest not in LANGUAGES: if dest in SPECIAL_CASES: dest = SPECIAL_CASES[dest] elif dest in LANGCODES: dest = LANGCODES[dest] else: raise ValueError('invalid destination language') if isinstance(text, list): result = [] for item in text: translated = self.translate(item, dest=dest, src=src) result.append(translated) return result origin = text data = self._translate(text, dest, src) translated = ''.join([(d[0] if d[0] else '') for d in data[0]]) extra_data = self._parse_extra_data(data) try: src = data[2] except Exception: pass pron = origin try: pron = data[0][1][-2] except Exception: pass if not PY3 and isinstance(pron, unicode) and isinstance(origin, str): origin = origin.decode('utf-8') if dest in EXCLUDES and pron == origin: pron = translated if not PY3: if isinstance(src, str): src = src.decode('utf-8') if isinstance(dest, str): dest = dest.decode('utf-8') if isinstance(translated, str): translated = translated.decode('utf-8') result = Translated(src=src, dest=dest, origin=origin, text=translated, pronunciation=pron, extra_data=extra_data) return result
translate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, text, dest='en', src='auto'):\n dest = dest.lower().split('_', 1)[0]\n src = src.lower().split('_', 1)[0]\n if src != 'auto' and src not in LANGUAGES:\n if src in SPECIAL_CASES:\n src = SPECIAL_CASES[src]\n elif src in LANGCODES:\n src = LANGCODES[src]\n else:\n raise ValueError('invalid source language')\n if dest not in LANGUAGES:\n if dest in SPECIAL_CASES:\n dest = SPECIAL_CASES[dest]\n elif dest in LANGCODES:\n dest = LANGCODES[dest]\n else:\n raise ValueError('invalid destination language')\n if isinstance(text, list):\n result = []\n for item in text:\n translated = self.translate(item, dest=dest, src=src)\n result.append(translated)\n return result\n origin = text\n data = self._translate(text, dest, src)\n translated = ''.join([(d[0] if d[0] else '') for d in data[0]])\n extra_data = self._parse_extra_data(data)\n try:\n src = data[2]\n except Exception:\n pass\n pron = origin\n try:\n pron = data[0][1][-2]\n except Exception:\n pass\n if not PY3 and isinstance(pron, unicode) and isinstance(origin, str):\n origin = origin.decode('utf-8')\n if dest in EXCLUDES and pron == origin:\n pron = translated\n if not PY3:\n if isinstance(src, str):\n src = src.decode('utf-8')\n if isinstance(dest, str):\n dest = dest.decode('utf-8')\n if isinstance(translated, str):\n translated = translated.decode('utf-8')\n result = Translated(src=src, dest=dest, origin=origin, text=translated,\n pronunciation=pron, extra_data=extra_data)\n return result\n", "role": "user" }, { "content": "translate", "role": "assistant" } ]
def x(self, text): if isinstance(text, list): result = [] for item in text: lang = self.detect(item) result.append(lang) return result data = self._translate(text, dest='en', src='auto') src = '' confidence = 0.0 try: src = ''.join(data[8][0]) confidence = data[8][-2][0] except Exception: pass result = Detected(lang=src, confidence=confidence) return result
detect
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, text):\n if isinstance(text, list):\n result = []\n for item in text:\n lang = self.detect(item)\n result.append(lang)\n return result\n data = self._translate(text, dest='en', src='auto')\n src = ''\n confidence = 0.0\n try:\n src = ''.join(data[8][0])\n confidence = data[8][-2][0]\n except Exception:\n pass\n result = Detected(lang=src, confidence=confidence)\n return result\n", "role": "user" }, { "content": "detect", "role": "assistant" } ]