source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(survival_table, observed_deaths_col='observed', censored_col='censored'):
T_ = []
E_ = []
W_ = []
for t, row in survival_table.iterrows():
if row[observed_deaths_col] > 0:
T_.append(t)
E_.append(1)
W_.append(row[observed_deaths_col])
if row[censored_col] > 0:
T_.append(t)
E_.append(0)
W_.append(row[censored_col])
return np.asarray(T_), np.asarray(E_), np.asarray(W_)
| survival_events_from_table | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(survival_table, observed_deaths_col='observed', censored_col='censored'):\n T_ = []\n E_ = []\n W_ = []\n for t, row in survival_table.iterrows():\n if row[observed_deaths_col] > 0:\n T_.append(t)\n E_.append(1)\n W_.append(row[observed_deaths_col])\n if row[censored_col] > 0:\n T_.append(t)\n E_.append(0)\n W_.append(row[censored_col])\n return np.asarray(T_), np.asarray(E_), np.asarray(W_)\n",
"role": "user"
},
{
"content": "survival_events_from_table",
"role": "assistant"
}
] |
def x(start_times, end_times, fill_date=datetime.today(), freq='D',
dayfirst=False, na_values=None):
fill_date = pd.to_datetime(fill_date)
freq_string = 'timedelta64[%s]' % freq
start_times = pd.Series(start_times).copy()
end_times = pd.Series(end_times).copy()
C = ~(pd.isnull(end_times).values | end_times.isin(na_values or ['']))
end_times[~C] = fill_date
start_times_ = pd.to_datetime(start_times, dayfirst=dayfirst)
end_times_ = pd.to_datetime(end_times, dayfirst=dayfirst, errors='coerce')
deaths_after_cutoff = end_times_ > fill_date
C[deaths_after_cutoff] = False
T = (end_times_ - start_times_).values.astype(freq_string).astype(float)
if (T < 0).sum():
warnings.warn('Warning: some values of start_times are after end_times'
)
return T, C.values
| datetimes_to_durations | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(start_times, end_times, fill_date=datetime.today(), freq='D',\n dayfirst=False, na_values=None):\n fill_date = pd.to_datetime(fill_date)\n freq_string = 'timedelta64[%s]' % freq\n start_times = pd.Series(start_times).copy()\n end_times = pd.Series(end_times).copy()\n C = ~(pd.isnull(end_times).values | end_times.isin(na_values or ['']))\n end_times[~C] = fill_date\n start_times_ = pd.to_datetime(start_times, dayfirst=dayfirst)\n end_times_ = pd.to_datetime(end_times, dayfirst=dayfirst, errors='coerce')\n deaths_after_cutoff = end_times_ > fill_date\n C[deaths_after_cutoff] = False\n T = (end_times_ - start_times_).values.astype(freq_string).astype(float)\n if (T < 0).sum():\n warnings.warn('Warning: some values of start_times are after end_times'\n )\n return T, C.values\n",
"role": "user"
},
{
"content": "datetimes_to_durations",
"role": "assistant"
}
] |
def x(event_times, predicted_event_times, event_observed=None):
if event_observed is None:
event_observed = np.ones_like(event_times)
ix = event_observed.astype(bool)
return np.abs(np.log(event_times[ix]) - np.log(predicted_event_times[ix])
).mean()
| l1_log_loss | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(event_times, predicted_event_times, event_observed=None):\n if event_observed is None:\n event_observed = np.ones_like(event_times)\n ix = event_observed.astype(bool)\n return np.abs(np.log(event_times[ix]) - np.log(predicted_event_times[ix])\n ).mean()\n",
"role": "user"
},
{
"content": "l1_log_loss",
"role": "assistant"
}
] |
def x(event_times, predicted_event_times, event_observed=None):
if event_observed is None:
event_observed = np.ones_like(event_times)
ix = event_observed.astype(bool)
return np.power(np.log(event_times[ix]) - np.log(predicted_event_times[
ix]), 2).mean()
| l2_log_loss | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(event_times, predicted_event_times, event_observed=None):\n if event_observed is None:\n event_observed = np.ones_like(event_times)\n ix = event_observed.astype(bool)\n return np.power(np.log(event_times[ix]) - np.log(predicted_event_times[\n ix]), 2).mean()\n",
"role": "user"
},
{
"content": "l2_log_loss",
"role": "assistant"
}
] |
def x(fitters, df, duration_col, event_col=None, k=5, evaluation_measure=
concordance_index, predictor='predict_expectation', predictor_kwargs={},
fitter_kwargs={}):
try:
fitters = list(fitters)
except TypeError:
fitters = [fitters]
fitterscores = [[] for _ in fitters]
n, _ = df.shape
df = df.copy()
if event_col is None:
event_col = 'E'
df[event_col] = 1.0
df = df.reindex(np.random.permutation(df.index)).sort_values(event_col)
assignments = np.array((n // k + 1) * list(range(1, k + 1)))
assignments = assignments[:n]
testing_columns = df.columns.drop([duration_col, event_col])
for i in range(1, k + 1):
ix = assignments == i
training_data = df.loc[~ix]
testing_data = df.loc[ix]
T_actual = testing_data[duration_col].values
E_actual = testing_data[event_col].values
X_testing = testing_data[testing_columns]
for fitter, scores in zip(fitters, fitterscores):
fitter.fit(training_data, duration_col=duration_col, event_col=
event_col, **fitter_kwargs)
T_pred = getattr(fitter, predictor)(X_testing, **predictor_kwargs
).values
try:
scores.append(evaluation_measure(T_actual, T_pred, E_actual))
except TypeError:
scores.append(evaluation_measure(T_actual, T_pred))
if len(fitters) == 1:
return fitterscores[0]
return fitterscores
| k_fold_cross_validation | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fitters, df, duration_col, event_col=None, k=5, evaluation_measure=\n concordance_index, predictor='predict_expectation', predictor_kwargs={},\n fitter_kwargs={}):\n try:\n fitters = list(fitters)\n except TypeError:\n fitters = [fitters]\n fitterscores = [[] for _ in fitters]\n n, _ = df.shape\n df = df.copy()\n if event_col is None:\n event_col = 'E'\n df[event_col] = 1.0\n df = df.reindex(np.random.permutation(df.index)).sort_values(event_col)\n assignments = np.array((n // k + 1) * list(range(1, k + 1)))\n assignments = assignments[:n]\n testing_columns = df.columns.drop([duration_col, event_col])\n for i in range(1, k + 1):\n ix = assignments == i\n training_data = df.loc[~ix]\n testing_data = df.loc[ix]\n T_actual = testing_data[duration_col].values\n E_actual = testing_data[event_col].values\n X_testing = testing_data[testing_columns]\n for fitter, scores in zip(fitters, fitterscores):\n fitter.fit(training_data, duration_col=duration_col, event_col=\n event_col, **fitter_kwargs)\n T_pred = getattr(fitter, predictor)(X_testing, **predictor_kwargs\n ).values\n try:\n scores.append(evaluation_measure(T_actual, T_pred, E_actual))\n except TypeError:\n scores.append(evaluation_measure(T_actual, T_pred))\n if len(fitters) == 1:\n return fitterscores[0]\n return fitterscores\n",
"role": "user"
},
{
"content": "k_fold_cross_validation",
"role": "assistant"
}
] |
def x(X, mean=None, std=None):
if mean is None or std is None:
mean = X.mean(0)
std = X.std(0)
return (X - mean) / std
| normalize | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(X, mean=None, std=None):\n if mean is None or std is None:\n mean = X.mean(0)\n std = X.std(0)\n return (X - mean) / std\n",
"role": "user"
},
{
"content": "normalize",
"role": "assistant"
}
] |
def x(X, Y, c1=0.0, c2=0.0, offset=None, ix=None):
_, d = X.shape
if c1 > 0 or c2 > 0:
penalizer_matrix = (c1 + c2) * np.eye(d)
A = np.dot(X.T, X) + penalizer_matrix
else:
A = np.dot(X.T, X)
if offset is None or c2 == 0:
b = np.dot(X.T, Y)
else:
b = np.dot(X.T, Y) + c2 * offset
if ix is not None:
M = np.c_[X.T[:, ix], b]
else:
M = np.c_[X.T, b]
R = solve(A, M, assume_a='pos', check_finite=False)
return R[:, -1], R[:, :-1]
| ridge_regression | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(X, Y, c1=0.0, c2=0.0, offset=None, ix=None):\n _, d = X.shape\n if c1 > 0 or c2 > 0:\n penalizer_matrix = (c1 + c2) * np.eye(d)\n A = np.dot(X.T, X) + penalizer_matrix\n else:\n A = np.dot(X.T, X)\n if offset is None or c2 == 0:\n b = np.dot(X.T, Y)\n else:\n b = np.dot(X.T, Y) + c2 * offset\n if ix is not None:\n M = np.c_[X.T[:, ix], b]\n else:\n M = np.c_[X.T, b]\n R = solve(A, M, assume_a='pos', check_finite=False)\n return R[:, -1], R[:, :-1]\n",
"role": "user"
},
{
"content": "ridge_regression",
"role": "assistant"
}
] |
def x(events, timeline, _additive_f, _additive_var, reverse):
if reverse:
events = events.sort_index(ascending=False)
at_risk = events['entrance'].sum() - events['removed'].cumsum().shift(1
).fillna(0)
deaths = events['observed']
estimate_ = np.cumsum(_additive_f(at_risk, deaths)).sort_index().shift(
-1).fillna(0)
var_ = np.cumsum(_additive_var(at_risk, deaths)).sort_index().shift(-1
).fillna(0)
else:
deaths = events['observed']
entrances = events['entrance'].copy()
entrances.iloc[0] = 0
population = events['at_risk'] - entrances
estimate_ = np.cumsum(_additive_f(population, deaths))
var_ = np.cumsum(_additive_var(population, deaths))
timeline = sorted(timeline)
estimate_ = estimate_.reindex(timeline, method='pad').fillna(0)
var_ = var_.reindex(timeline, method='pad')
var_.index.name = 'timeline'
estimate_.index.name = 'timeline'
return estimate_, var_
| additive_estimate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(events, timeline, _additive_f, _additive_var, reverse):\n if reverse:\n events = events.sort_index(ascending=False)\n at_risk = events['entrance'].sum() - events['removed'].cumsum().shift(1\n ).fillna(0)\n deaths = events['observed']\n estimate_ = np.cumsum(_additive_f(at_risk, deaths)).sort_index().shift(\n -1).fillna(0)\n var_ = np.cumsum(_additive_var(at_risk, deaths)).sort_index().shift(-1\n ).fillna(0)\n else:\n deaths = events['observed']\n entrances = events['entrance'].copy()\n entrances.iloc[0] = 0\n population = events['at_risk'] - entrances\n estimate_ = np.cumsum(_additive_f(population, deaths))\n var_ = np.cumsum(_additive_var(population, deaths))\n timeline = sorted(timeline)\n estimate_ = estimate_.reindex(timeline, method='pad').fillna(0)\n var_ = var_.reindex(timeline, method='pad')\n var_.index.name = 'timeline'\n estimate_.index.name = 'timeline'\n return estimate_, var_\n",
"role": "user"
},
{
"content": "additive_estimate",
"role": "assistant"
}
] |
def x(durations, event_observed, timeline, entry, weights):
n = len(durations)
durations = np.asarray(pass_for_numeric_dtypes_or_raise_array(durations)
).reshape((n,))
if event_observed is None:
event_observed = np.ones(n, dtype=int)
else:
event_observed = np.asarray(event_observed).reshape((n,)).copy(
).astype(int)
if entry is not None:
entry = np.asarray(entry).reshape((n,))
event_table = survival_table_from_events(durations, event_observed,
entry, weights=weights)
if timeline is None:
timeline = event_table.index.values
else:
timeline = np.asarray(timeline)
return durations, event_observed, timeline.astype(float
), entry, event_table
| preprocess_inputs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(durations, event_observed, timeline, entry, weights):\n n = len(durations)\n durations = np.asarray(pass_for_numeric_dtypes_or_raise_array(durations)\n ).reshape((n,))\n if event_observed is None:\n event_observed = np.ones(n, dtype=int)\n else:\n event_observed = np.asarray(event_observed).reshape((n,)).copy(\n ).astype(int)\n if entry is not None:\n entry = np.asarray(entry).reshape((n,))\n event_table = survival_table_from_events(durations, event_observed,\n entry, weights=weights)\n if timeline is None:\n timeline = event_table.index.values\n else:\n timeline = np.asarray(timeline)\n return durations, event_observed, timeline.astype(float\n ), entry, event_table\n",
"role": "user"
},
{
"content": "preprocess_inputs",
"role": "assistant"
}
] |
def x(x):
try:
if isinstance(x, (pd.Series, pd.DataFrame)):
v = pd.to_numeric(x.squeeze())
else:
v = pd.to_numeric(np.asarray(x).squeeze())
if v.size == 0:
raise ValueError('Empty array/Series passed in.')
return v
except:
raise ValueError(
'Values must be numeric: no strings, datetimes, objects, etc.')
| pass_for_numeric_dtypes_or_raise_array | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(x):\n try:\n if isinstance(x, (pd.Series, pd.DataFrame)):\n v = pd.to_numeric(x.squeeze())\n else:\n v = pd.to_numeric(np.asarray(x).squeeze())\n if v.size == 0:\n raise ValueError('Empty array/Series passed in.')\n return v\n except:\n raise ValueError(\n 'Values must be numeric: no strings, datetimes, objects, etc.')\n",
"role": "user"
},
{
"content": "pass_for_numeric_dtypes_or_raise_array",
"role": "assistant"
}
] |
def x(df, duration_col, event_col, id_col=None, time_gaps=1):
df = df.copy()
df[duration_col] /= time_gaps
df = to_long_format(df, duration_col)
stop_col = 'stop'
start_col = 'start'
_, d = df.shape
if id_col is None:
id_col = 'id'
df.index.rename(id_col, inplace=True)
df = df.reset_index()
d_dftv = d + 1
else:
d_dftv = d
dtype_dftv = object if (df.dtypes == object).any() else float
n_dftv = int(np.ceil(df[stop_col]).sum())
tv_array = np.empty((n_dftv, d_dftv), dtype=dtype_dftv)
special_columns = [stop_col, start_col, event_col]
non_special_columns = df.columns.difference(special_columns).tolist()
order_I_want = special_columns + non_special_columns
df = df[order_I_want]
position_counter = 0
for _, row in df.iterrows():
T, E = row[stop_col], row[event_col]
T_int = int(np.ceil(T))
values = np.tile(row.values, (T_int, 1))
values[:, 0] = np.arange(1, T + 1, dtype=float)
values[-1, 0] = T
values[:, 1] = np.arange(0, T, dtype=float)
values[:, 2] = 0.0
values[-1, 2] = float(E)
tv_array[position_counter:position_counter + T_int, :] = values
position_counter += T_int
dftv = pd.DataFrame(tv_array, columns=df.columns)
dftv = dftv.astype(dtype=df.dtypes[non_special_columns + [event_col]].
to_dict())
dftv[start_col] *= time_gaps
dftv[stop_col] *= time_gaps
return dftv
| to_episodic_format | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(df, duration_col, event_col, id_col=None, time_gaps=1):\n df = df.copy()\n df[duration_col] /= time_gaps\n df = to_long_format(df, duration_col)\n stop_col = 'stop'\n start_col = 'start'\n _, d = df.shape\n if id_col is None:\n id_col = 'id'\n df.index.rename(id_col, inplace=True)\n df = df.reset_index()\n d_dftv = d + 1\n else:\n d_dftv = d\n dtype_dftv = object if (df.dtypes == object).any() else float\n n_dftv = int(np.ceil(df[stop_col]).sum())\n tv_array = np.empty((n_dftv, d_dftv), dtype=dtype_dftv)\n special_columns = [stop_col, start_col, event_col]\n non_special_columns = df.columns.difference(special_columns).tolist()\n order_I_want = special_columns + non_special_columns\n df = df[order_I_want]\n position_counter = 0\n for _, row in df.iterrows():\n T, E = row[stop_col], row[event_col]\n T_int = int(np.ceil(T))\n values = np.tile(row.values, (T_int, 1))\n values[:, 0] = np.arange(1, T + 1, dtype=float)\n values[-1, 0] = T\n values[:, 1] = np.arange(0, T, dtype=float)\n values[:, 2] = 0.0\n values[-1, 2] = float(E)\n tv_array[position_counter:position_counter + T_int, :] = values\n position_counter += T_int\n dftv = pd.DataFrame(tv_array, columns=df.columns)\n dftv = dftv.astype(dtype=df.dtypes[non_special_columns + [event_col]].\n to_dict())\n dftv[start_col] *= time_gaps\n dftv[stop_col] *= time_gaps\n return dftv\n",
"role": "user"
},
{
"content": "to_episodic_format",
"role": "assistant"
}
] |
def x(df, duration_col):
return df.assign(start=0, stop=lambda s: s[duration_col]).drop(duration_col
, axis=1)
| to_long_format | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(df, duration_col):\n return df.assign(start=0, stop=lambda s: s[duration_col]).drop(duration_col\n , axis=1)\n",
"role": "user"
},
{
"content": "to_long_format",
"role": "assistant"
}
] |
def x(long_form_df, cv, id_col, duration_col, event_col, start_col='start',
stop_col='stop', add_enum=False, overwrite=True, cumulative_sum=False,
cumulative_sum_prefix='cumsum_', delay=0):
def x(cv):
if cumulative_sum:
return cv
cols = cv.columns.difference([duration_col])
cv = cv.loc[(cv[cols].shift() != cv[cols]).any(axis=1)]
return cv
def x(cv):
return cv.rename(columns={duration_col: start_col})
def x(df, cvs):
id_ = df.name
try:
cv = cvs.get_group(id_)
except KeyError:
return df
final_state = bool(df[event_col].iloc[-1])
final_stop_time = df[stop_col].iloc[-1]
df = df.drop([id_col, event_col, stop_col], axis=1).set_index(start_col
)
cv = cv.drop([id_col], axis=1).set_index(start_col).loc[:
final_stop_time]
if cumulative_sum:
cv = cv.cumsum()
cv = cv.add_prefix(cumulative_sum_prefix)
if not overwrite:
expanded_df = cv.combine(df, lambda s1, s2: s1 + s2, fill_value
=0, overwrite=False)
elif overwrite:
expanded_df = cv.combine_first(df)
n = expanded_df.shape[0]
expanded_df = expanded_df.reset_index()
expanded_df[stop_col] = expanded_df[start_col].shift(-1)
expanded_df[id_col] = id_
expanded_df[event_col] = False
expanded_df.at[n - 1, event_col] = final_state
expanded_df.at[n - 1, stop_col] = final_stop_time
if add_enum:
expanded_df['enum'] = np.arange(1, n + 1)
if cumulative_sum:
expanded_df[cv.columns] = expanded_df[cv.columns].ffill().fillna(0)
return expanded_df.ffill()
if delay < 0:
raise ValueError('delay parameter must be equal to or greater than 0')
if any(col not in long_form_df for col in (id_col, event_col, start_col,
stop_col)):
raise IndexError('Missing column in long_form_df')
cv[duration_col] += delay
cv = cv.dropna()
cv = cv.sort_values([id_col, duration_col])
cvs = cv.pipe(remove_redundant_rows).pipe(transform_cv_to_long_format
).groupby(id_col)
long_form_df = long_form_df.groupby(id_col, group_keys=False).apply(expand,
cvs=cvs)
return long_form_df.reset_index(drop=True)
| add_covariate_to_timeline | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(long_form_df, cv, id_col, duration_col, event_col, start_col='start',\n stop_col='stop', add_enum=False, overwrite=True, cumulative_sum=False,\n cumulative_sum_prefix='cumsum_', delay=0):\n\n def x(cv):\n if cumulative_sum:\n return cv\n cols = cv.columns.difference([duration_col])\n cv = cv.loc[(cv[cols].shift() != cv[cols]).any(axis=1)]\n return cv\n\n def x(cv):\n return cv.rename(columns={duration_col: start_col})\n\n def x(df, cvs):\n id_ = df.name\n try:\n cv = cvs.get_group(id_)\n except KeyError:\n return df\n final_state = bool(df[event_col].iloc[-1])\n final_stop_time = df[stop_col].iloc[-1]\n df = df.drop([id_col, event_col, stop_col], axis=1).set_index(start_col\n )\n cv = cv.drop([id_col], axis=1).set_index(start_col).loc[:\n final_stop_time]\n if cumulative_sum:\n cv = cv.cumsum()\n cv = cv.add_prefix(cumulative_sum_prefix)\n if not overwrite:\n expanded_df = cv.combine(df, lambda s1, s2: s1 + s2, fill_value\n =0, overwrite=False)\n elif overwrite:\n expanded_df = cv.combine_first(df)\n n = expanded_df.shape[0]\n expanded_df = expanded_df.reset_index()\n expanded_df[stop_col] = expanded_df[start_col].shift(-1)\n expanded_df[id_col] = id_\n expanded_df[event_col] = False\n expanded_df.at[n - 1, event_col] = final_state\n expanded_df.at[n - 1, stop_col] = final_stop_time\n if add_enum:\n expanded_df['enum'] = np.arange(1, n + 1)\n if cumulative_sum:\n expanded_df[cv.columns] = expanded_df[cv.columns].ffill().fillna(0)\n return expanded_df.ffill()\n if delay < 0:\n raise ValueError('delay parameter must be equal to or greater than 0')\n if any(col not in long_form_df for col in (id_col, event_col, start_col,\n stop_col)):\n raise IndexError('Missing column in long_form_df')\n cv[duration_col] += delay\n cv = cv.dropna()\n cv = cv.sort_values([id_col, duration_col])\n cvs = cv.pipe(remove_redundant_rows).pipe(transform_cv_to_long_format\n ).groupby(id_col)\n long_form_df = long_form_df.groupby(id_col, group_keys=False).apply(expand,\n cvs=cvs)\n return long_form_df.reset_index(drop=True)\n",
"role": "user"
},
{
"content": "add_covariate_to_timeline",
"role": "assistant"
}
] |
def x(df, id_col):
df = df.set_index(id_col)
df = df.stack().reset_index()
df.columns = [id_col, 'event', 'duration']
df['_counter'] = 1
return df.pivot_table(index=[id_col, 'duration'], columns='event',
fill_value=0)['_counter'].reset_index()
| covariates_from_event_matrix | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(df, id_col):\n df = df.set_index(id_col)\n df = df.stack().reset_index()\n df.columns = [id_col, 'event', 'duration']\n df['_counter'] = 1\n return df.pivot_table(index=[id_col, 'duration'], columns='event',\n fill_value=0)['_counter'].reset_index()\n",
"role": "user"
},
{
"content": "covariates_from_event_matrix",
"role": "assistant"
}
] |
def x(decimals):
threshold = 10 ** 5
return lambda n: '{:.{prec}e}'.format(n, prec=decimals
) if n > threshold else '{:4.{prec}f}'.format(n, prec=decimals)
| format_exp_floats | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(decimals):\n threshold = 10 ** 5\n return lambda n: '{:.{prec}e}'.format(n, prec=decimals\n ) if n > threshold else '{:4.{prec}f}'.format(n, prec=decimals)\n",
"role": "user"
},
{
"content": "format_exp_floats",
"role": "assistant"
}
] |
def x(self, durations, event_observed=None, timeline=None, entry=None,
label='BFH_estimate', alpha=None, ci_labels=None):
self._censoring_type = CensoringType.RIGHT
self._label = label
alpha = coalesce(alpha, self.alpha)
naf = NelsonAalenFitter(alpha=alpha)
naf.fit(durations, event_observed=event_observed, timeline=timeline,
label=label, entry=entry, ci_labels=ci_labels)
(self.durations, self.event_observed, self.timeline, self.entry, self.
event_table) = (naf.durations, naf.event_observed, naf.timeline,
naf.entry, naf.event_table)
self.survival_function_ = np.exp(-naf.cumulative_hazard_)
self.confidence_interval_ = np.exp(-naf.confidence_interval_)
self._estimation_method = 'survival_function_'
self._estimate_name = 'survival_function_'
self._predict_label = label
self._update_docstrings()
self.plot_survival_function = self.plot
return self
| fit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, durations, event_observed=None, timeline=None, entry=None,\n label='BFH_estimate', alpha=None, ci_labels=None):\n self._censoring_type = CensoringType.RIGHT\n self._label = label\n alpha = coalesce(alpha, self.alpha)\n naf = NelsonAalenFitter(alpha=alpha)\n naf.fit(durations, event_observed=event_observed, timeline=timeline,\n label=label, entry=entry, ci_labels=ci_labels)\n (self.durations, self.event_observed, self.timeline, self.entry, self.\n event_table) = (naf.durations, naf.event_observed, naf.timeline,\n naf.entry, naf.event_table)\n self.survival_function_ = np.exp(-naf.cumulative_hazard_)\n self.confidence_interval_ = np.exp(-naf.confidence_interval_)\n self._estimation_method = 'survival_function_'\n self._estimate_name = 'survival_function_'\n self._predict_label = label\n self._update_docstrings()\n self.plot_survival_function = self.plot\n return self\n",
"role": "user"
},
{
"content": "fit",
"role": "assistant"
}
] |
def x(self, durations, event_observed=None, timeline=None, entry=None,
label='KM_estimate', left_censorship=False, alpha=None, ci_labels=None,
weights=None):
if left_censorship:
warnings.warn(
'kwarg left_censorship is deprecated and will be removed in a future release. Please use ``.fit_left_censoring`` instead.'
, DeprecationWarning)
self._censoring_type = CensoringType.RIGHT
return self._fit(durations, event_observed, timeline, entry, label,
alpha, ci_labels, weights)
| fit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, durations, event_observed=None, timeline=None, entry=None,\n label='KM_estimate', left_censorship=False, alpha=None, ci_labels=None,\n weights=None):\n if left_censorship:\n warnings.warn(\n 'kwarg left_censorship is deprecated and will be removed in a future release. Please use ``.fit_left_censoring`` instead.'\n , DeprecationWarning)\n self._censoring_type = CensoringType.RIGHT\n return self._fit(durations, event_observed, timeline, entry, label,\n alpha, ci_labels, weights)\n",
"role": "user"
},
{
"content": "fit",
"role": "assistant"
}
] |
def x(self, durations, event_observed=None, timeline=None, entry=None,
label='KM_estimate', alpha=None, ci_labels=None, weights=None):
self._censoring_type = CensoringType.LEFT
return self._fit(durations, event_observed, timeline, entry, label,
alpha, ci_labels, weights)
| fit_left_censoring | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, durations, event_observed=None, timeline=None, entry=None,\n label='KM_estimate', alpha=None, ci_labels=None, weights=None):\n self._censoring_type = CensoringType.LEFT\n return self._fit(durations, event_observed, timeline, entry, label,\n alpha, ci_labels, weights)\n",
"role": "user"
},
{
"content": "fit_left_censoring",
"role": "assistant"
}
] |
def x(self, durations, event_observed=None, timeline=None, entry=None,
label='KM_estimate', alpha=None, ci_labels=None, weights=None):
self._check_values(durations)
if event_observed is not None:
self._check_values(event_observed)
self._label = label
if weights is not None:
weights = np.asarray(weights)
if (weights.astype(int) != weights).any():
warnings.warn(
"""It looks like your weights are not integers, possibly propensity scores then?
It's important to know that the naive variance estimates of the coefficients are biased. Instead use Monte Carlo to
estimate the variances. See paper "Variance estimation when using inverse probability of treatment weighting (IPTW) with survival analysis"
or "Adjusted Kaplan-Meier estimator and log-rank test with inverse probability of treatment weighting for survival data."
"""
, StatisticalWarning)
is_left_censoring = self._censoring_type == CensoringType.LEFT
primary_estimate_name = ('survival_function_' if not is_left_censoring else
'cumulative_density_')
secondary_estimate_name = ('cumulative_density_' if not
is_left_censoring else 'survival_function_')
(self.durations, self.event_observed, self.timeline, self.entry, self.
event_table) = (_preprocess_inputs(durations, event_observed,
timeline, entry, weights))
alpha = alpha if alpha else self.alpha
log_estimate, cumulative_sq_ = _additive_estimate(self.event_table,
self.timeline, self._additive_f, self._additive_var, is_left_censoring)
if entry is not None:
n = self.event_table.shape[0]
net_population = (self.event_table['entrance'] - self.event_table[
'removed']).cumsum()
if net_population.iloc[:int(n / 2)].min() == 0:
ix = net_population.iloc[:int(n / 2)].idxmin()
raise StatError(
'There are too few early truncation times and too many events. S(t)==0 for all t>%g. Recommend BreslowFlemingHarringtonFitter.'
% ix)
setattr(self, primary_estimate_name, pd.DataFrame(np.exp(log_estimate),
columns=[self._label]))
setattr(self, secondary_estimate_name, pd.DataFrame(1 - np.exp(
log_estimate), columns=[self._label]))
self.__estimate = getattr(self, primary_estimate_name)
self.confidence_interval_ = self._bounds(cumulative_sq_[:, None], alpha,
ci_labels)
self.median_ = median_survival_times(self.__estimate, left_censorship=
is_left_censoring)
self._cumulative_sq_ = cumulative_sq_
setattr(self, 'confidence_interval_' + primary_estimate_name, self.
confidence_interval_)
setattr(self, 'confidence_interval_' + secondary_estimate_name, 1 -
self.confidence_interval_)
self._estimation_method = primary_estimate_name
self._estimate_name = primary_estimate_name
self._predict_label = label
self._update_docstrings()
return self
| fit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, durations, event_observed=None, timeline=None, entry=None,\n label='KM_estimate', alpha=None, ci_labels=None, weights=None):\n self._check_values(durations)\n if event_observed is not None:\n self._check_values(event_observed)\n self._label = label\n if weights is not None:\n weights = np.asarray(weights)\n if (weights.astype(int) != weights).any():\n warnings.warn(\n \"\"\"It looks like your weights are not integers, possibly propensity scores then?\n It's important to know that the naive variance estimates of the coefficients are biased. Instead use Monte Carlo to\n estimate the variances. See paper \"Variance estimation when using inverse probability of treatment weighting (IPTW) with survival analysis\"\n or \"Adjusted Kaplan-Meier estimator and log-rank test with inverse probability of treatment weighting for survival data.\"\n \"\"\"\n , StatisticalWarning)\n is_left_censoring = self._censoring_type == CensoringType.LEFT\n primary_estimate_name = ('survival_function_' if not is_left_censoring else\n 'cumulative_density_')\n secondary_estimate_name = ('cumulative_density_' if not\n is_left_censoring else 'survival_function_')\n (self.durations, self.event_observed, self.timeline, self.entry, self.\n event_table) = (_preprocess_inputs(durations, event_observed,\n timeline, entry, weights))\n alpha = alpha if alpha else self.alpha\n log_estimate, cumulative_sq_ = _additive_estimate(self.event_table,\n self.timeline, self._additive_f, self._additive_var, is_left_censoring)\n if entry is not None:\n n = self.event_table.shape[0]\n net_population = (self.event_table['entrance'] - self.event_table[\n 'removed']).cumsum()\n if net_population.iloc[:int(n / 2)].min() == 0:\n ix = net_population.iloc[:int(n / 2)].idxmin()\n raise StatError(\n 'There are too few early truncation times and too many events. S(t)==0 for all t>%g. Recommend BreslowFlemingHarringtonFitter.'\n % ix)\n setattr(self, primary_estimate_name, pd.DataFrame(np.exp(log_estimate),\n columns=[self._label]))\n setattr(self, secondary_estimate_name, pd.DataFrame(1 - np.exp(\n log_estimate), columns=[self._label]))\n self.__estimate = getattr(self, primary_estimate_name)\n self.confidence_interval_ = self._bounds(cumulative_sq_[:, None], alpha,\n ci_labels)\n self.median_ = median_survival_times(self.__estimate, left_censorship=\n is_left_censoring)\n self._cumulative_sq_ = cumulative_sq_\n setattr(self, 'confidence_interval_' + primary_estimate_name, self.\n confidence_interval_)\n setattr(self, 'confidence_interval_' + secondary_estimate_name, 1 -\n self.confidence_interval_)\n self._estimation_method = primary_estimate_name\n self._estimate_name = primary_estimate_name\n self._predict_label = label\n self._update_docstrings()\n return self\n",
"role": "user"
},
{
"content": "fit",
"role": "assistant"
}
] |
def x(self, times, label=None):
label = coalesce(label, self._label)
return pd.Series(1 - self.predict(times), index=_to_array(times), name=
label)
| cumulative_density_at_times | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, times, label=None):\n label = coalesce(label, self._label)\n return pd.Series(1 - self.predict(times), index=_to_array(times), name=\n label)\n",
"role": "user"
},
{
"content": "cumulative_density_at_times",
"role": "assistant"
}
] |
def x(self, **kwargs):
return _plot_estimate(self, estimate=self.survival_function_,
confidence_intervals=self.confidence_interval_survival_function_,
**kwargs)
| plot_survival_function | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n return _plot_estimate(self, estimate=self.survival_function_,\n confidence_intervals=self.confidence_interval_survival_function_,\n **kwargs)\n",
"role": "user"
},
{
"content": "plot_survival_function",
"role": "assistant"
}
] |
def x(self, **kwargs):
return _plot_estimate(self, estimate=self.cumulative_density_,
confidence_intervals=self.confidence_interval_cumulative_density_,
**kwargs)
| plot_cumulative_density | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n return _plot_estimate(self, estimate=self.cumulative_density_,\n confidence_intervals=self.confidence_interval_cumulative_density_,\n **kwargs)\n",
"role": "user"
},
{
"content": "plot_cumulative_density",
"role": "assistant"
}
] |
def x(self, df, id_col, event_col, start_col='start', stop_col='stop',
weights_col=None, show_progress=False, step_size=None, robust=False,
strata=None, initial_point=None):
self.strata = coalesce(strata, self.strata)
self.robust = robust
if self.robust:
raise NotImplementedError('Not available yet.')
self.event_col = event_col
self.id_col = id_col
self.stop_col = stop_col
self.start_col = start_col
self._time_fit_was_called = datetime.utcnow().strftime('%Y-%m-%d %H:%M:%S')
df = df.copy()
if not (id_col in df and event_col in df and start_col in df and
stop_col in df):
raise KeyError(
'A column specified in the call to `fit` does not exist in the DataFrame provided.'
)
if weights_col is None:
self.weights_col = None
assert '__weights' not in df.columns, '__weights is an internal lifelines column, please rename your column first.'
df['__weights'] = 1.0
else:
self.weights_col = weights_col
if (df[weights_col] <= 0).any():
raise ValueError('values in weights_col must be positive.')
df = df.rename(columns={id_col: 'id', event_col: 'event', start_col:
'start', stop_col: 'stop', weights_col: '__weights'})
if self.strata is None:
df = df.set_index('id')
else:
df = df.set_index(_to_list(self.strata) + ['id'])
df = df.sort_index()
events, start, stop = pass_for_numeric_dtypes_or_raise_array(df.pop(
'event')).astype(bool), df.pop('start'), df.pop('stop')
weights = df.pop('__weights').astype(float)
df = df.astype(float)
self._check_values(df, events, start, stop)
self._norm_mean = df.mean(0)
self._norm_std = df.std(0)
hazards_ = self._newton_rhaphson(normalize(df, self._norm_mean, self.
_norm_std), events, start, stop, weights, initial_point=
initial_point, show_progress=show_progress, step_size=step_size)
self.hazards_ = pd.Series(hazards_, index=df.columns, name='coef'
) / self._norm_std
self.variance_matrix_ = -inv(self._hessian_) / np.outer(self._norm_std,
self._norm_std)
self.standard_errors_ = self._compute_standard_errors(normalize(df,
self._norm_mean, self._norm_std), events, start, stop, weights)
self.confidence_intervals_ = self._compute_confidence_intervals()
self.baseline_cumulative_hazard_ = (self.
_compute_cumulative_baseline_hazard(df, events, start, stop, weights))
self.baseline_survival_ = self._compute_baseline_survival()
self.event_observed = events
self.start_stop_and_events = pd.DataFrame({'event': events, 'start':
start, 'stop': stop})
self.weights = weights
self._n_examples = df.shape[0]
self._n_unique = df.index.unique().shape[0]
return self
| fit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, df, id_col, event_col, start_col='start', stop_col='stop',\n weights_col=None, show_progress=False, step_size=None, robust=False,\n strata=None, initial_point=None):\n self.strata = coalesce(strata, self.strata)\n self.robust = robust\n if self.robust:\n raise NotImplementedError('Not available yet.')\n self.event_col = event_col\n self.id_col = id_col\n self.stop_col = stop_col\n self.start_col = start_col\n self._time_fit_was_called = datetime.utcnow().strftime('%Y-%m-%d %H:%M:%S')\n df = df.copy()\n if not (id_col in df and event_col in df and start_col in df and \n stop_col in df):\n raise KeyError(\n 'A column specified in the call to `fit` does not exist in the DataFrame provided.'\n )\n if weights_col is None:\n self.weights_col = None\n assert '__weights' not in df.columns, '__weights is an internal lifelines column, please rename your column first.'\n df['__weights'] = 1.0\n else:\n self.weights_col = weights_col\n if (df[weights_col] <= 0).any():\n raise ValueError('values in weights_col must be positive.')\n df = df.rename(columns={id_col: 'id', event_col: 'event', start_col:\n 'start', stop_col: 'stop', weights_col: '__weights'})\n if self.strata is None:\n df = df.set_index('id')\n else:\n df = df.set_index(_to_list(self.strata) + ['id'])\n df = df.sort_index()\n events, start, stop = pass_for_numeric_dtypes_or_raise_array(df.pop(\n 'event')).astype(bool), df.pop('start'), df.pop('stop')\n weights = df.pop('__weights').astype(float)\n df = df.astype(float)\n self._check_values(df, events, start, stop)\n self._norm_mean = df.mean(0)\n self._norm_std = df.std(0)\n hazards_ = self._newton_rhaphson(normalize(df, self._norm_mean, self.\n _norm_std), events, start, stop, weights, initial_point=\n initial_point, show_progress=show_progress, step_size=step_size)\n self.hazards_ = pd.Series(hazards_, index=df.columns, name='coef'\n ) / self._norm_std\n self.variance_matrix_ = -inv(self._hessian_) / np.outer(self._norm_std,\n self._norm_std)\n self.standard_errors_ = self._compute_standard_errors(normalize(df,\n self._norm_mean, self._norm_std), events, start, stop, weights)\n self.confidence_intervals_ = self._compute_confidence_intervals()\n self.baseline_cumulative_hazard_ = (self.\n _compute_cumulative_baseline_hazard(df, events, start, stop, weights))\n self.baseline_survival_ = self._compute_baseline_survival()\n self.event_observed = events\n self.start_stop_and_events = pd.DataFrame({'event': events, 'start':\n start, 'stop': stop})\n self.weights = weights\n self._n_examples = df.shape[0]\n self._n_unique = df.index.unique().shape[0]\n return self\n",
"role": "user"
},
{
"content": "fit",
"role": "assistant"
}
] |
def x(self, df, events, start, stop, weights, show_progress=False,
step_size=None, precision=1e-05, max_steps=50, initial_point=None):
assert precision <= 1.0, 'precision must be less than or equal to 1.'
_, d = df.shape
if initial_point is not None:
beta = initial_point
else:
beta = np.zeros((d,))
i = 0
converging = True
ll, previous_ll = 0, 0
start_time = time.time()
step_sizer = StepSizer(step_size)
step_size = step_sizer.next()
while converging:
i += 1
if self.strata is None:
h, g, ll = self._get_gradients(df.values, events.values, start.
values, stop.values, weights.values, beta)
else:
g = np.zeros_like(beta)
h = np.zeros((d, d))
ll = 0
for _h, _g, _ll in self._partition_by_strata_and_apply(df,
events, start, stop, weights, self._get_gradients, beta):
g += _g
h += _h
ll += _ll
if i == 1 and np.all(beta == 0):
self._log_likelihood_null = ll
if self.penalizer > 0:
g -= self.penalizer * beta
h.flat[::d + 1] -= self.penalizer
try:
inv_h_dot_g_T = spsolve(-h, g, sym_pos=True)
except ValueError as e:
if 'infs or NaNs' in str(e):
raise ConvergenceError(
"""hessian or gradient contains nan or inf value(s). Convergence halted. Please see the following tips in the lifelines documentation:
https://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model
"""
, e)
else:
raise e
except LinAlgError as e:
raise ConvergenceError(
"""Convergence halted due to matrix inversion problems. Suspicion is high colinearity. Please see the following tips in the lifelines documentation:
https://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model
"""
, e)
delta = step_size * inv_h_dot_g_T
if np.any(np.isnan(delta)):
raise ConvergenceError(
"""delta contains nan value(s). Convergence halted. Please see the following tips in the lifelines documentation:
https://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model
"""
)
hessian, gradient = h, g
norm_delta = norm(delta)
newton_decrement = g.dot(inv_h_dot_g_T) / 2
if show_progress:
print(
'Iteration %d: norm_delta = %.5f, step_size = %.5f, ll = %.5f, newton_decrement = %.5f, seconds_since_start = %.1f'
% (i, norm_delta, step_size, ll, newton_decrement, time.
time() - start_time))
if norm_delta < precision:
converging, completed = False, True
elif previous_ll > 0 and abs(ll - previous_ll) / -previous_ll < 1e-09:
converging, completed = False, True
elif newton_decrement < 1e-07:
converging, completed = False, True
elif i >= max_steps:
converging, completed = False, False
elif step_size <= 0.0001:
converging, completed = False, False
elif abs(ll) < 0.0001 and norm_delta > 1.0:
warnings.warn(
'The log-likelihood is getting suspiciously close to 0 and the delta is still large. There may be complete separation in the dataset. This may result in incorrect inference of coefficients. See https://stats.stackexchange.com/questions/11109/how-to-deal-with-perfect-separation-in-logistic-regression'
, ConvergenceWarning)
converging, completed = False, False
step_size = step_sizer.update(norm_delta).next()
beta += delta
self._hessian_ = hessian
self._score_ = gradient
self._log_likelihood = ll
if show_progress and completed:
print('Convergence completed after %d iterations.' % i)
elif show_progress and not completed:
print('Convergence failed. See any warning messages.')
if completed and norm_delta > 0.1:
warnings.warn(
'Newton-Rhapson convergence completed but norm(delta) is still high, %.3f. This may imply non-unique solutions to the maximum likelihood. Perhaps there is colinearity or complete separation in the dataset?'
% norm_delta, ConvergenceWarning)
elif not completed:
warnings.warn(
'Newton-Rhapson failed to converge sufficiently in %d steps.' %
max_steps, ConvergenceWarning)
return beta
| newton_rhaphson | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, df, events, start, stop, weights, show_progress=False,\n step_size=None, precision=1e-05, max_steps=50, initial_point=None):\n assert precision <= 1.0, 'precision must be less than or equal to 1.'\n _, d = df.shape\n if initial_point is not None:\n beta = initial_point\n else:\n beta = np.zeros((d,))\n i = 0\n converging = True\n ll, previous_ll = 0, 0\n start_time = time.time()\n step_sizer = StepSizer(step_size)\n step_size = step_sizer.next()\n while converging:\n i += 1\n if self.strata is None:\n h, g, ll = self._get_gradients(df.values, events.values, start.\n values, stop.values, weights.values, beta)\n else:\n g = np.zeros_like(beta)\n h = np.zeros((d, d))\n ll = 0\n for _h, _g, _ll in self._partition_by_strata_and_apply(df,\n events, start, stop, weights, self._get_gradients, beta):\n g += _g\n h += _h\n ll += _ll\n if i == 1 and np.all(beta == 0):\n self._log_likelihood_null = ll\n if self.penalizer > 0:\n g -= self.penalizer * beta\n h.flat[::d + 1] -= self.penalizer\n try:\n inv_h_dot_g_T = spsolve(-h, g, sym_pos=True)\n except ValueError as e:\n if 'infs or NaNs' in str(e):\n raise ConvergenceError(\n \"\"\"hessian or gradient contains nan or inf value(s). Convergence halted. Please see the following tips in the lifelines documentation:\nhttps://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model\n\"\"\"\n , e)\n else:\n raise e\n except LinAlgError as e:\n raise ConvergenceError(\n \"\"\"Convergence halted due to matrix inversion problems. Suspicion is high colinearity. Please see the following tips in the lifelines documentation:\nhttps://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model\n\"\"\"\n , e)\n delta = step_size * inv_h_dot_g_T\n if np.any(np.isnan(delta)):\n raise ConvergenceError(\n \"\"\"delta contains nan value(s). Convergence halted. Please see the following tips in the lifelines documentation:\nhttps://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model\n\"\"\"\n )\n hessian, gradient = h, g\n norm_delta = norm(delta)\n newton_decrement = g.dot(inv_h_dot_g_T) / 2\n if show_progress:\n print(\n 'Iteration %d: norm_delta = %.5f, step_size = %.5f, ll = %.5f, newton_decrement = %.5f, seconds_since_start = %.1f'\n % (i, norm_delta, step_size, ll, newton_decrement, time.\n time() - start_time))\n if norm_delta < precision:\n converging, completed = False, True\n elif previous_ll > 0 and abs(ll - previous_ll) / -previous_ll < 1e-09:\n converging, completed = False, True\n elif newton_decrement < 1e-07:\n converging, completed = False, True\n elif i >= max_steps:\n converging, completed = False, False\n elif step_size <= 0.0001:\n converging, completed = False, False\n elif abs(ll) < 0.0001 and norm_delta > 1.0:\n warnings.warn(\n 'The log-likelihood is getting suspiciously close to 0 and the delta is still large. There may be complete separation in the dataset. This may result in incorrect inference of coefficients. See https://stats.stackexchange.com/questions/11109/how-to-deal-with-perfect-separation-in-logistic-regression'\n , ConvergenceWarning)\n converging, completed = False, False\n step_size = step_sizer.update(norm_delta).next()\n beta += delta\n self._hessian_ = hessian\n self._score_ = gradient\n self._log_likelihood = ll\n if show_progress and completed:\n print('Convergence completed after %d iterations.' % i)\n elif show_progress and not completed:\n print('Convergence failed. See any warning messages.')\n if completed and norm_delta > 0.1:\n warnings.warn(\n 'Newton-Rhapson convergence completed but norm(delta) is still high, %.3f. This may imply non-unique solutions to the maximum likelihood. Perhaps there is colinearity or complete separation in the dataset?'\n % norm_delta, ConvergenceWarning)\n elif not completed:\n warnings.warn(\n 'Newton-Rhapson failed to converge sufficiently in %d steps.' %\n max_steps, ConvergenceWarning)\n return beta\n",
"role": "user"
},
{
"content": "newton_rhaphson",
"role": "assistant"
}
] |
def x(self, X, events, start, stop, weights, beta):
_, d = X.shape
hessian = np.zeros((d, d))
gradient = np.zeros(d)
log_lik = 0
unique_death_times = np.unique(stop[events])
for t in unique_death_times:
ix = (start < t) & (t <= stop)
X_at_t = X[ix]
weights_at_t = weights[ix]
stops_events_at_t = stop[ix]
events_at_t = events[ix]
phi_i = weights_at_t * np.exp(np.dot(X_at_t, beta))
phi_x_i = phi_i[:, None] * X_at_t
phi_x_x_i = np.dot(X_at_t.T, phi_x_i)
risk_phi = array_sum_to_scalar(phi_i)
risk_phi_x = matrix_axis_0_sum_to_array(phi_x_i)
risk_phi_x_x = phi_x_x_i
deaths = events_at_t & (stops_events_at_t == t)
tied_death_counts = array_sum_to_scalar(deaths.astype(int))
xi_deaths = X_at_t[deaths]
x_death_sum = matrix_axis_0_sum_to_array(weights_at_t[deaths, None] *
xi_deaths)
weight_count = array_sum_to_scalar(weights_at_t[deaths])
weighted_average = weight_count / tied_death_counts
if tied_death_counts > 1:
tie_phi = array_sum_to_scalar(phi_i[deaths])
tie_phi_x = matrix_axis_0_sum_to_array(phi_x_i[deaths])
tie_phi_x_x = np.dot(xi_deaths.T, phi_i[deaths, None] * xi_deaths)
increasing_proportion = np.arange(tied_death_counts
) / tied_death_counts
denom = 1.0 / (risk_phi - increasing_proportion * tie_phi)
numer = risk_phi_x - np.outer(increasing_proportion, tie_phi_x)
a1 = np.einsum('ab, i->ab', risk_phi_x_x, denom) - np.einsum(
'ab, i->ab', tie_phi_x_x, increasing_proportion * denom)
else:
denom = 1.0 / np.array([risk_phi])
numer = risk_phi_x
a1 = risk_phi_x_x * denom
summand = numer * denom[:, None]
a2 = summand.T.dot(summand)
gradient = gradient + x_death_sum - weighted_average * summand.sum(0)
log_lik = log_lik + np.dot(x_death_sum, beta
) + weighted_average * np.log(denom).sum()
hessian = hessian + weighted_average * (a2 - a1)
return hessian, gradient, log_lik
| get_gradients | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, events, start, stop, weights, beta):\n _, d = X.shape\n hessian = np.zeros((d, d))\n gradient = np.zeros(d)\n log_lik = 0\n unique_death_times = np.unique(stop[events])\n for t in unique_death_times:\n ix = (start < t) & (t <= stop)\n X_at_t = X[ix]\n weights_at_t = weights[ix]\n stops_events_at_t = stop[ix]\n events_at_t = events[ix]\n phi_i = weights_at_t * np.exp(np.dot(X_at_t, beta))\n phi_x_i = phi_i[:, None] * X_at_t\n phi_x_x_i = np.dot(X_at_t.T, phi_x_i)\n risk_phi = array_sum_to_scalar(phi_i)\n risk_phi_x = matrix_axis_0_sum_to_array(phi_x_i)\n risk_phi_x_x = phi_x_x_i\n deaths = events_at_t & (stops_events_at_t == t)\n tied_death_counts = array_sum_to_scalar(deaths.astype(int))\n xi_deaths = X_at_t[deaths]\n x_death_sum = matrix_axis_0_sum_to_array(weights_at_t[deaths, None] *\n xi_deaths)\n weight_count = array_sum_to_scalar(weights_at_t[deaths])\n weighted_average = weight_count / tied_death_counts\n if tied_death_counts > 1:\n tie_phi = array_sum_to_scalar(phi_i[deaths])\n tie_phi_x = matrix_axis_0_sum_to_array(phi_x_i[deaths])\n tie_phi_x_x = np.dot(xi_deaths.T, phi_i[deaths, None] * xi_deaths)\n increasing_proportion = np.arange(tied_death_counts\n ) / tied_death_counts\n denom = 1.0 / (risk_phi - increasing_proportion * tie_phi)\n numer = risk_phi_x - np.outer(increasing_proportion, tie_phi_x)\n a1 = np.einsum('ab, i->ab', risk_phi_x_x, denom) - np.einsum(\n 'ab, i->ab', tie_phi_x_x, increasing_proportion * denom)\n else:\n denom = 1.0 / np.array([risk_phi])\n numer = risk_phi_x\n a1 = risk_phi_x_x * denom\n summand = numer * denom[:, None]\n a2 = summand.T.dot(summand)\n gradient = gradient + x_death_sum - weighted_average * summand.sum(0)\n log_lik = log_lik + np.dot(x_death_sum, beta\n ) + weighted_average * np.log(denom).sum()\n hessian = hessian + weighted_average * (a2 - a1)\n return hessian, gradient, log_lik\n",
"role": "user"
},
{
"content": "get_gradients",
"role": "assistant"
}
] |
def x(self, X):
if isinstance(X, pd.DataFrame):
order = self.hazards_.index
X = X[order]
check_for_numeric_dtypes_or_raise(X)
X = X.astype(float)
index = _get_index(X)
X = normalize(X, self._norm_mean.values, 1)
return pd.DataFrame(np.dot(X, self.hazards_), index=index)
| predict_log_partial_hazard | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n if isinstance(X, pd.DataFrame):\n order = self.hazards_.index\n X = X[order]\n check_for_numeric_dtypes_or_raise(X)\n X = X.astype(float)\n index = _get_index(X)\n X = normalize(X, self._norm_mean.values, 1)\n return pd.DataFrame(np.dot(X, self.hazards_), index=index)\n",
"role": "user"
},
{
"content": "predict_log_partial_hazard",
"role": "assistant"
}
] |
def x(self, decimals=2, **kwargs):
justify = string_justify(18)
print(self)
print("{} = '{}'".format(justify('event col'), self.event_col))
if self.weights_col:
print("{} = '{}'".format(justify('weights col'), self.weights_col))
if self.strata:
print('{} = {}'.format(justify('strata'), self.strata))
if self.penalizer > 0:
print('{} = {}'.format(justify('penalizer'), self.penalizer))
print('{} = {}'.format(justify('number of subjects'), self._n_unique))
print('{} = {}'.format(justify('number of periods'), self._n_examples))
print('{} = {}'.format(justify('number of events'), self.event_observed
.sum()))
print('{} = {:.{prec}f}'.format(justify('log-likelihood'), self.
_log_likelihood, prec=decimals))
print('{} = {} UTC'.format(justify('time fit was run'), self.
_time_fit_was_called))
for k, v in kwargs.items():
print('{} = {}\n'.format(justify(k), v))
print(end='\n')
print('---')
df = self.summary
print(df.to_string(float_format=format_floats(decimals), formatters={
'p': format_p_value(decimals), 'exp(coef)': format_exp_floats(
decimals)}))
print('---')
print(
'Log-likelihood ratio test = {:.{prec}f} on {} df, -log2(p)={:.{prec}f}'
.format(*self._compute_likelihood_ratio_test(), prec=decimals))
| print_summary | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, decimals=2, **kwargs):\n justify = string_justify(18)\n print(self)\n print(\"{} = '{}'\".format(justify('event col'), self.event_col))\n if self.weights_col:\n print(\"{} = '{}'\".format(justify('weights col'), self.weights_col))\n if self.strata:\n print('{} = {}'.format(justify('strata'), self.strata))\n if self.penalizer > 0:\n print('{} = {}'.format(justify('penalizer'), self.penalizer))\n print('{} = {}'.format(justify('number of subjects'), self._n_unique))\n print('{} = {}'.format(justify('number of periods'), self._n_examples))\n print('{} = {}'.format(justify('number of events'), self.event_observed\n .sum()))\n print('{} = {:.{prec}f}'.format(justify('log-likelihood'), self.\n _log_likelihood, prec=decimals))\n print('{} = {} UTC'.format(justify('time fit was run'), self.\n _time_fit_was_called))\n for k, v in kwargs.items():\n print('{} = {}\\n'.format(justify(k), v))\n print(end='\\n')\n print('---')\n df = self.summary\n print(df.to_string(float_format=format_floats(decimals), formatters={\n 'p': format_p_value(decimals), 'exp(coef)': format_exp_floats(\n decimals)}))\n print('---')\n print(\n 'Log-likelihood ratio test = {:.{prec}f} on {} df, -log2(p)={:.{prec}f}'\n .format(*self._compute_likelihood_ratio_test(), prec=decimals))\n",
"role": "user"
},
{
"content": "print_summary",
"role": "assistant"
}
] |
def x(self, columns=None, **errorbar_kwargs):
from matplotlib import pyplot as plt
ax = errorbar_kwargs.pop('ax', None) or plt.figure().add_subplot(111)
errorbar_kwargs.setdefault('c', 'k')
errorbar_kwargs.setdefault('fmt', 's')
errorbar_kwargs.setdefault('markerfacecolor', 'white')
errorbar_kwargs.setdefault('markeredgewidth', 1.25)
errorbar_kwargs.setdefault('elinewidth', 1.25)
errorbar_kwargs.setdefault('capsize', 3)
z = inv_normal_cdf(1 - self.alpha / 2)
if columns is None:
columns = self.hazards_.index
yaxis_locations = list(range(len(columns)))
symmetric_errors = z * self.standard_errors_[columns].to_frame().squeeze(
axis=1).values.copy()
hazards = self.hazards_[columns].values.copy()
order = np.argsort(hazards)
ax.errorbar(hazards[order], yaxis_locations, xerr=symmetric_errors[
order], **errorbar_kwargs)
best_ylim = ax.get_ylim()
ax.vlines(0, -2, len(columns) + 1, linestyles='dashed', linewidths=1,
alpha=0.65)
ax.set_ylim(best_ylim)
tick_labels = [columns[i] for i in order]
ax.set_yticks(yaxis_locations)
ax.set_yticklabels(tick_labels)
ax.set_xlabel('log(HR) (%g%% CI)' % ((1 - self.alpha) * 100))
return ax
| plot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, columns=None, **errorbar_kwargs):\n from matplotlib import pyplot as plt\n ax = errorbar_kwargs.pop('ax', None) or plt.figure().add_subplot(111)\n errorbar_kwargs.setdefault('c', 'k')\n errorbar_kwargs.setdefault('fmt', 's')\n errorbar_kwargs.setdefault('markerfacecolor', 'white')\n errorbar_kwargs.setdefault('markeredgewidth', 1.25)\n errorbar_kwargs.setdefault('elinewidth', 1.25)\n errorbar_kwargs.setdefault('capsize', 3)\n z = inv_normal_cdf(1 - self.alpha / 2)\n if columns is None:\n columns = self.hazards_.index\n yaxis_locations = list(range(len(columns)))\n symmetric_errors = z * self.standard_errors_[columns].to_frame().squeeze(\n axis=1).values.copy()\n hazards = self.hazards_[columns].values.copy()\n order = np.argsort(hazards)\n ax.errorbar(hazards[order], yaxis_locations, xerr=symmetric_errors[\n order], **errorbar_kwargs)\n best_ylim = ax.get_ylim()\n ax.vlines(0, -2, len(columns) + 1, linestyles='dashed', linewidths=1,\n alpha=0.65)\n ax.set_ylim(best_ylim)\n tick_labels = [columns[i] for i in order]\n ax.set_yticks(yaxis_locations)\n ax.set_yticklabels(tick_labels)\n ax.set_xlabel('log(HR) (%g%% CI)' % ((1 - self.alpha) * 100))\n return ax\n",
"role": "user"
},
{
"content": "plot",
"role": "assistant"
}
] |
def x(self, df, events, start, stop, weights):
score_residuals = self._compute_residuals(df, events, start, stop, weights
) * weights[:, None]
naive_var = inv(self._hessian_)
delta_betas = -score_residuals.dot(naive_var) / self._norm_std.values
return delta_betas
| compute_delta_beta | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, df, events, start, stop, weights):\n score_residuals = self._compute_residuals(df, events, start, stop, weights\n ) * weights[:, None]\n naive_var = inv(self._hessian_)\n delta_betas = -score_residuals.dot(naive_var) / self._norm_std.values\n return delta_betas\n",
"role": "user"
},
{
"content": "compute_delta_beta",
"role": "assistant"
}
] |
def x(self, df, duration_col=None, event_col=None, show_progress=False,
initial_point=None, strata=None, step_size=None, weights_col=None,
cluster_col=None, robust=False, batch_mode=None):
if duration_col is None:
raise TypeError('duration_col cannot be None.')
self._censoring_type = CensoringType.RIGHT
self._time_fit_was_called = datetime.utcnow().strftime('%Y-%m-%d %H:%M:%S'
) + ' UTC'
self.duration_col = duration_col
self.event_col = event_col
self.robust = robust
self.cluster_col = cluster_col
self.weights_col = weights_col
self._n_examples = df.shape[0]
self._batch_mode = batch_mode
self.strata = coalesce(strata, self.strata)
X, T, E, weights, original_index, self._clusters = (self.
_preprocess_dataframe(df))
self.durations = T.copy()
self.event_observed = E.copy()
self.weights = weights.copy()
if self.strata is not None:
self.durations.index = original_index
self.event_observed.index = original_index
self.weights.index = original_index
self._norm_mean = X.mean(0)
self._norm_std = X.std(0)
X_norm = normalize(X, self._norm_mean, self._norm_std)
hazards_ = self._fit_model(X_norm, T, E, weights=weights, initial_point
=initial_point, show_progress=show_progress, step_size=step_size)
self.hazards_ = pd.Series(hazards_, index=X.columns, name='coef'
) / self._norm_std
self.variance_matrix_ = -inv(self._hessian_) / np.outer(self._norm_std,
self._norm_std)
self.standard_errors_ = self._compute_standard_errors(X_norm, T, E, weights
)
self.confidence_intervals_ = self._compute_confidence_intervals()
self._predicted_partial_hazards_ = self.predict_partial_hazard(X).rename(
columns={(0): 'P'}).assign(T=self.durations.values, E=self.
event_observed.values, W=self.weights.values).set_index(X.index)
self.baseline_hazard_ = self._compute_baseline_hazards()
self.baseline_cumulative_hazard_ = (self.
_compute_baseline_cumulative_hazard())
self.baseline_survival_ = self._compute_baseline_survival()
if hasattr(self, '_concordance_score_'):
del self._concordance_score_
return self
| fit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, df, duration_col=None, event_col=None, show_progress=False,\n initial_point=None, strata=None, step_size=None, weights_col=None,\n cluster_col=None, robust=False, batch_mode=None):\n if duration_col is None:\n raise TypeError('duration_col cannot be None.')\n self._censoring_type = CensoringType.RIGHT\n self._time_fit_was_called = datetime.utcnow().strftime('%Y-%m-%d %H:%M:%S'\n ) + ' UTC'\n self.duration_col = duration_col\n self.event_col = event_col\n self.robust = robust\n self.cluster_col = cluster_col\n self.weights_col = weights_col\n self._n_examples = df.shape[0]\n self._batch_mode = batch_mode\n self.strata = coalesce(strata, self.strata)\n X, T, E, weights, original_index, self._clusters = (self.\n _preprocess_dataframe(df))\n self.durations = T.copy()\n self.event_observed = E.copy()\n self.weights = weights.copy()\n if self.strata is not None:\n self.durations.index = original_index\n self.event_observed.index = original_index\n self.weights.index = original_index\n self._norm_mean = X.mean(0)\n self._norm_std = X.std(0)\n X_norm = normalize(X, self._norm_mean, self._norm_std)\n hazards_ = self._fit_model(X_norm, T, E, weights=weights, initial_point\n =initial_point, show_progress=show_progress, step_size=step_size)\n self.hazards_ = pd.Series(hazards_, index=X.columns, name='coef'\n ) / self._norm_std\n self.variance_matrix_ = -inv(self._hessian_) / np.outer(self._norm_std,\n self._norm_std)\n self.standard_errors_ = self._compute_standard_errors(X_norm, T, E, weights\n )\n self.confidence_intervals_ = self._compute_confidence_intervals()\n self._predicted_partial_hazards_ = self.predict_partial_hazard(X).rename(\n columns={(0): 'P'}).assign(T=self.durations.values, E=self.\n event_observed.values, W=self.weights.values).set_index(X.index)\n self.baseline_hazard_ = self._compute_baseline_hazards()\n self.baseline_cumulative_hazard_ = (self.\n _compute_baseline_cumulative_hazard())\n self.baseline_survival_ = self._compute_baseline_survival()\n if hasattr(self, '_concordance_score_'):\n del self._concordance_score_\n return self\n",
"role": "user"
},
{
"content": "fit",
"role": "assistant"
}
] |
def x(self, X, T, E, weights=None, initial_point=None, step_size=None,
precision=1e-07, show_progress=True, max_steps=50):
self.path = []
assert precision <= 1.0, 'precision must be less than or equal to 1.'
_, d = X.shape
if initial_point is not None:
assert initial_point.shape == (d,)
beta = initial_point
else:
beta = np.zeros((d,))
step_sizer = StepSizer(step_size)
step_size = step_sizer.next()
if self.tie_method == 'Efron':
decision = BatchVsSingle.decide(self._batch_mode, T)
get_gradients = getattr(self, '_get_efron_values_%s' % decision)
self._batch_mode = decision == 'batch'
else:
raise NotImplementedError('Only Efron is available.')
i = 0
converging = True
ll, previous_ll = 0, 0
start = time.time()
while converging:
self.path.append(beta.copy())
i += 1
if self.strata is None:
h, g, ll = get_gradients(X.values, T.values, E.values, weights.
values, beta)
else:
g = np.zeros_like(beta)
h = np.zeros((beta.shape[0], beta.shape[0]))
ll = 0
for _h, _g, _ll in self._partition_by_strata_and_apply(X, T, E,
weights, get_gradients, beta):
g += _g
h += _h
ll += _ll
if i == 1 and np.all(beta == 0):
self._log_likelihood_null = ll
if self.penalizer > 0:
g -= self.penalizer * beta
h.flat[::d + 1] -= self.penalizer
try:
inv_h_dot_g_T = spsolve(-h, g, assume_a='pos', check_finite=False)
except ValueError as e:
if 'infs or NaNs' in str(e):
raise ConvergenceError(
"""Hessian or gradient contains nan or inf value(s). Convergence halted. Please see the following tips in the lifelines documentation:
https://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model
"""
, e)
else:
raise e
except LinAlgError as e:
raise ConvergenceError(
"""Convergence halted due to matrix inversion problems. Suspicion is high collinearity. Please see the following tips in the lifelines documentation:
https://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model
"""
, e)
delta = inv_h_dot_g_T
if np.any(np.isnan(delta)):
raise ConvergenceError(
"""delta contains nan value(s). Convergence halted. Please see the following tips in the lifelines documentation:
https://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model
"""
)
hessian, gradient = h, g
norm_delta = norm(delta)
newton_decrement = g.dot(inv_h_dot_g_T) / 2
if show_progress:
print(
'Iteration %d: norm_delta = %.5f, step_size = %.4f, ll = %.5f, newton_decrement = %.5f, seconds_since_start = %.1f'
% (i, norm_delta, step_size, ll, newton_decrement, time.
time() - start))
if norm_delta < precision:
converging, completed = False, True
elif previous_ll != 0 and abs(ll - previous_ll) / -previous_ll < 1e-09:
converging, completed = False, True
elif newton_decrement < precision:
converging, completed = False, True
elif i >= max_steps:
converging, completed = False, False
elif step_size <= 1e-05:
converging, completed = False, False
elif abs(ll) < 0.0001 and norm_delta > 1.0:
warnings.warn(
'The log-likelihood is getting suspiciously close to 0 and the delta is still large. There may be complete separation in the dataset. This may result in incorrect inference of coefficients. See https://stats.stackexchange.com/questions/11109/how-to-deal-with-perfect-separation-in-logistic-regression'
, ConvergenceWarning)
converging, completed = False, False
beta += step_size * delta
previous_ll = ll
step_size = step_sizer.update(norm_delta).next()
self._hessian_ = hessian
self._score_ = gradient
self._log_likelihood = ll
if show_progress and completed:
print('Convergence completed after %d iterations.' % i)
elif show_progress and not completed:
print('Convergence failed. See any warning messages.')
if completed and norm_delta > 0.1:
warnings.warn(
'Newton-Rhapson convergence completed but norm(delta) is still high, %.3f. This may imply non-unique solutions to the maximum likelihood. Perhaps there is collinearity or complete separation in the dataset?'
% norm_delta, ConvergenceWarning)
elif not completed:
warnings.warn(
'Newton-Rhapson failed to converge sufficiently in %d steps.' %
max_steps, ConvergenceWarning)
return beta
| fit_model | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, T, E, weights=None, initial_point=None, step_size=None,\n precision=1e-07, show_progress=True, max_steps=50):\n self.path = []\n assert precision <= 1.0, 'precision must be less than or equal to 1.'\n _, d = X.shape\n if initial_point is not None:\n assert initial_point.shape == (d,)\n beta = initial_point\n else:\n beta = np.zeros((d,))\n step_sizer = StepSizer(step_size)\n step_size = step_sizer.next()\n if self.tie_method == 'Efron':\n decision = BatchVsSingle.decide(self._batch_mode, T)\n get_gradients = getattr(self, '_get_efron_values_%s' % decision)\n self._batch_mode = decision == 'batch'\n else:\n raise NotImplementedError('Only Efron is available.')\n i = 0\n converging = True\n ll, previous_ll = 0, 0\n start = time.time()\n while converging:\n self.path.append(beta.copy())\n i += 1\n if self.strata is None:\n h, g, ll = get_gradients(X.values, T.values, E.values, weights.\n values, beta)\n else:\n g = np.zeros_like(beta)\n h = np.zeros((beta.shape[0], beta.shape[0]))\n ll = 0\n for _h, _g, _ll in self._partition_by_strata_and_apply(X, T, E,\n weights, get_gradients, beta):\n g += _g\n h += _h\n ll += _ll\n if i == 1 and np.all(beta == 0):\n self._log_likelihood_null = ll\n if self.penalizer > 0:\n g -= self.penalizer * beta\n h.flat[::d + 1] -= self.penalizer\n try:\n inv_h_dot_g_T = spsolve(-h, g, assume_a='pos', check_finite=False)\n except ValueError as e:\n if 'infs or NaNs' in str(e):\n raise ConvergenceError(\n \"\"\"Hessian or gradient contains nan or inf value(s). Convergence halted. Please see the following tips in the lifelines documentation:\nhttps://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model\n\"\"\"\n , e)\n else:\n raise e\n except LinAlgError as e:\n raise ConvergenceError(\n \"\"\"Convergence halted due to matrix inversion problems. Suspicion is high collinearity. Please see the following tips in the lifelines documentation:\nhttps://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model\n\"\"\"\n , e)\n delta = inv_h_dot_g_T\n if np.any(np.isnan(delta)):\n raise ConvergenceError(\n \"\"\"delta contains nan value(s). Convergence halted. Please see the following tips in the lifelines documentation:\nhttps://lifelines.readthedocs.io/en/latest/Examples.html#problems-with-convergence-in-the-cox-proportional-hazard-model\n\"\"\"\n )\n hessian, gradient = h, g\n norm_delta = norm(delta)\n newton_decrement = g.dot(inv_h_dot_g_T) / 2\n if show_progress:\n print(\n 'Iteration %d: norm_delta = %.5f, step_size = %.4f, ll = %.5f, newton_decrement = %.5f, seconds_since_start = %.1f'\n % (i, norm_delta, step_size, ll, newton_decrement, time.\n time() - start))\n if norm_delta < precision:\n converging, completed = False, True\n elif previous_ll != 0 and abs(ll - previous_ll) / -previous_ll < 1e-09:\n converging, completed = False, True\n elif newton_decrement < precision:\n converging, completed = False, True\n elif i >= max_steps:\n converging, completed = False, False\n elif step_size <= 1e-05:\n converging, completed = False, False\n elif abs(ll) < 0.0001 and norm_delta > 1.0:\n warnings.warn(\n 'The log-likelihood is getting suspiciously close to 0 and the delta is still large. There may be complete separation in the dataset. This may result in incorrect inference of coefficients. See https://stats.stackexchange.com/questions/11109/how-to-deal-with-perfect-separation-in-logistic-regression'\n , ConvergenceWarning)\n converging, completed = False, False\n beta += step_size * delta\n previous_ll = ll\n step_size = step_sizer.update(norm_delta).next()\n self._hessian_ = hessian\n self._score_ = gradient\n self._log_likelihood = ll\n if show_progress and completed:\n print('Convergence completed after %d iterations.' % i)\n elif show_progress and not completed:\n print('Convergence failed. See any warning messages.')\n if completed and norm_delta > 0.1:\n warnings.warn(\n 'Newton-Rhapson convergence completed but norm(delta) is still high, %.3f. This may imply non-unique solutions to the maximum likelihood. Perhaps there is collinearity or complete separation in the dataset?'\n % norm_delta, ConvergenceWarning)\n elif not completed:\n warnings.warn(\n 'Newton-Rhapson failed to converge sufficiently in %d steps.' %\n max_steps, ConvergenceWarning)\n return beta\n",
"role": "user"
},
{
"content": "fit_model",
"role": "assistant"
}
] |
def x(self, X, T, E, weights, beta):
n, d = X.shape
hessian = np.zeros((d, d))
gradient = np.zeros((d,))
log_lik = 0
x_death_sum = np.zeros((d,))
risk_phi, tie_phi = 0, 0
risk_phi_x, tie_phi_x = np.zeros((d,)), np.zeros((d,))
risk_phi_x_x, tie_phi_x_x = np.zeros((d, d)), np.zeros((d, d))
weight_count = 0.0
tied_death_counts = 0
scores = weights * np.exp(np.dot(X, beta))
for i in range(n - 1, -1, -1):
ti = T[i]
ei = E[i]
xi = X[i]
score = scores[i]
w = weights[i]
phi_i = score
phi_x_i = phi_i * xi
phi_x_x_i = np.outer(xi, phi_x_i)
risk_phi = risk_phi + phi_i
risk_phi_x = risk_phi_x + phi_x_i
risk_phi_x_x = risk_phi_x_x + phi_x_x_i
if ei:
x_death_sum = x_death_sum + w * xi
tie_phi = tie_phi + phi_i
tie_phi_x = tie_phi_x + phi_x_i
tie_phi_x_x = tie_phi_x_x + phi_x_x_i
tied_death_counts += 1
weight_count += w
if i > 0 and T[i - 1] == ti:
continue
elif tied_death_counts == 0:
continue
weighted_average = weight_count / tied_death_counts
if tied_death_counts > 1:
increasing_proportion = np.arange(tied_death_counts
) / tied_death_counts
denom = 1.0 / (risk_phi - increasing_proportion * tie_phi)
numer = risk_phi_x - np.outer(increasing_proportion, tie_phi_x)
a1 = np.einsum('ab,i->ab', risk_phi_x_x, denom) - np.einsum(
'ab,i->ab', tie_phi_x_x, increasing_proportion * denom)
else:
denom = 1.0 / np.array([risk_phi])
numer = risk_phi_x
a1 = risk_phi_x_x * denom
summand = numer * denom[:, None]
a2 = summand.T.dot(summand)
gradient = gradient + x_death_sum - weighted_average * summand.sum(0)
log_lik = log_lik + np.dot(x_death_sum, beta
) + weighted_average * np.log(denom).sum()
hessian = hessian + weighted_average * (a2 - a1)
tied_death_counts = 0
weight_count = 0.0
x_death_sum = np.zeros((d,))
tie_phi = 0
tie_phi_x = np.zeros((d,))
tie_phi_x_x = np.zeros((d, d))
return hessian, gradient, log_lik
| get_efron_values_single | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, T, E, weights, beta):\n n, d = X.shape\n hessian = np.zeros((d, d))\n gradient = np.zeros((d,))\n log_lik = 0\n x_death_sum = np.zeros((d,))\n risk_phi, tie_phi = 0, 0\n risk_phi_x, tie_phi_x = np.zeros((d,)), np.zeros((d,))\n risk_phi_x_x, tie_phi_x_x = np.zeros((d, d)), np.zeros((d, d))\n weight_count = 0.0\n tied_death_counts = 0\n scores = weights * np.exp(np.dot(X, beta))\n for i in range(n - 1, -1, -1):\n ti = T[i]\n ei = E[i]\n xi = X[i]\n score = scores[i]\n w = weights[i]\n phi_i = score\n phi_x_i = phi_i * xi\n phi_x_x_i = np.outer(xi, phi_x_i)\n risk_phi = risk_phi + phi_i\n risk_phi_x = risk_phi_x + phi_x_i\n risk_phi_x_x = risk_phi_x_x + phi_x_x_i\n if ei:\n x_death_sum = x_death_sum + w * xi\n tie_phi = tie_phi + phi_i\n tie_phi_x = tie_phi_x + phi_x_i\n tie_phi_x_x = tie_phi_x_x + phi_x_x_i\n tied_death_counts += 1\n weight_count += w\n if i > 0 and T[i - 1] == ti:\n continue\n elif tied_death_counts == 0:\n continue\n weighted_average = weight_count / tied_death_counts\n if tied_death_counts > 1:\n increasing_proportion = np.arange(tied_death_counts\n ) / tied_death_counts\n denom = 1.0 / (risk_phi - increasing_proportion * tie_phi)\n numer = risk_phi_x - np.outer(increasing_proportion, tie_phi_x)\n a1 = np.einsum('ab,i->ab', risk_phi_x_x, denom) - np.einsum(\n 'ab,i->ab', tie_phi_x_x, increasing_proportion * denom)\n else:\n denom = 1.0 / np.array([risk_phi])\n numer = risk_phi_x\n a1 = risk_phi_x_x * denom\n summand = numer * denom[:, None]\n a2 = summand.T.dot(summand)\n gradient = gradient + x_death_sum - weighted_average * summand.sum(0)\n log_lik = log_lik + np.dot(x_death_sum, beta\n ) + weighted_average * np.log(denom).sum()\n hessian = hessian + weighted_average * (a2 - a1)\n tied_death_counts = 0\n weight_count = 0.0\n x_death_sum = np.zeros((d,))\n tie_phi = 0\n tie_phi_x = np.zeros((d,))\n tie_phi_x_x = np.zeros((d, d))\n return hessian, gradient, log_lik\n",
"role": "user"
},
{
"content": "get_efron_values_single",
"role": "assistant"
}
] |
def x(self, X, T, E, weights, beta):
n, d = X.shape
hessian = np.zeros((d, d))
gradient = np.zeros((d,))
log_lik = 0
risk_phi, tie_phi = 0, 0
risk_phi_x, tie_phi_x = np.zeros((d,)), np.zeros((d,))
risk_phi_x_x, tie_phi_x_x = np.zeros((d, d)), np.zeros((d, d))
_, counts = np.unique(-T, return_counts=True)
scores = weights * np.exp(np.dot(X, beta))
pos = n
for count_of_removals in counts:
slice_ = slice(pos - count_of_removals, pos)
X_at_t = X[slice_]
weights_at_t = weights[slice_]
phi_i = scores[slice_, None]
phi_x_i = phi_i * X_at_t
phi_x_x_i = np.dot(X_at_t.T, phi_x_i)
risk_phi = risk_phi + array_sum_to_scalar(phi_i)
risk_phi_x = risk_phi_x + matrix_axis_0_sum_to_array(phi_x_i)
risk_phi_x_x = risk_phi_x_x + phi_x_x_i
deaths = E[slice_]
tied_death_counts = array_sum_to_scalar(deaths.astype(int))
if tied_death_counts == 0:
pos -= count_of_removals
continue
xi_deaths = X_at_t[deaths]
weights_deaths = weights_at_t[deaths]
x_death_sum = matrix_axis_0_sum_to_array(weights_deaths[:, None] *
xi_deaths)
weight_count = array_sum_to_scalar(weights_deaths)
weighted_average = weight_count / tied_death_counts
if tied_death_counts > 1:
tie_phi = array_sum_to_scalar(phi_i[deaths])
tie_phi_x = matrix_axis_0_sum_to_array(phi_x_i[deaths])
tie_phi_x_x = np.dot(xi_deaths.T, phi_i[deaths] * xi_deaths)
increasing_proportion = np.arange(tied_death_counts
) / tied_death_counts
denom = 1.0 / (risk_phi - increasing_proportion * tie_phi)
numer = risk_phi_x - np.outer(increasing_proportion, tie_phi_x)
a1 = np.einsum('ab,i->ab', risk_phi_x_x, denom) - np.einsum(
'ab,i->ab', tie_phi_x_x, increasing_proportion * denom)
else:
denom = 1.0 / np.array([risk_phi])
numer = risk_phi_x
a1 = risk_phi_x_x * denom
summand = numer * denom[:, None]
a2 = summand.T.dot(summand)
gradient = gradient + x_death_sum - weighted_average * summand.sum(0)
log_lik = log_lik + np.dot(x_death_sum, beta
) + weighted_average * np.log(denom).sum()
hessian = hessian + weighted_average * (a2 - a1)
pos -= count_of_removals
return hessian, gradient, log_lik
| get_efron_values_batch | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, T, E, weights, beta):\n n, d = X.shape\n hessian = np.zeros((d, d))\n gradient = np.zeros((d,))\n log_lik = 0\n risk_phi, tie_phi = 0, 0\n risk_phi_x, tie_phi_x = np.zeros((d,)), np.zeros((d,))\n risk_phi_x_x, tie_phi_x_x = np.zeros((d, d)), np.zeros((d, d))\n _, counts = np.unique(-T, return_counts=True)\n scores = weights * np.exp(np.dot(X, beta))\n pos = n\n for count_of_removals in counts:\n slice_ = slice(pos - count_of_removals, pos)\n X_at_t = X[slice_]\n weights_at_t = weights[slice_]\n phi_i = scores[slice_, None]\n phi_x_i = phi_i * X_at_t\n phi_x_x_i = np.dot(X_at_t.T, phi_x_i)\n risk_phi = risk_phi + array_sum_to_scalar(phi_i)\n risk_phi_x = risk_phi_x + matrix_axis_0_sum_to_array(phi_x_i)\n risk_phi_x_x = risk_phi_x_x + phi_x_x_i\n deaths = E[slice_]\n tied_death_counts = array_sum_to_scalar(deaths.astype(int))\n if tied_death_counts == 0:\n pos -= count_of_removals\n continue\n xi_deaths = X_at_t[deaths]\n weights_deaths = weights_at_t[deaths]\n x_death_sum = matrix_axis_0_sum_to_array(weights_deaths[:, None] *\n xi_deaths)\n weight_count = array_sum_to_scalar(weights_deaths)\n weighted_average = weight_count / tied_death_counts\n if tied_death_counts > 1:\n tie_phi = array_sum_to_scalar(phi_i[deaths])\n tie_phi_x = matrix_axis_0_sum_to_array(phi_x_i[deaths])\n tie_phi_x_x = np.dot(xi_deaths.T, phi_i[deaths] * xi_deaths)\n increasing_proportion = np.arange(tied_death_counts\n ) / tied_death_counts\n denom = 1.0 / (risk_phi - increasing_proportion * tie_phi)\n numer = risk_phi_x - np.outer(increasing_proportion, tie_phi_x)\n a1 = np.einsum('ab,i->ab', risk_phi_x_x, denom) - np.einsum(\n 'ab,i->ab', tie_phi_x_x, increasing_proportion * denom)\n else:\n denom = 1.0 / np.array([risk_phi])\n numer = risk_phi_x\n a1 = risk_phi_x_x * denom\n summand = numer * denom[:, None]\n a2 = summand.T.dot(summand)\n gradient = gradient + x_death_sum - weighted_average * summand.sum(0)\n log_lik = log_lik + np.dot(x_death_sum, beta\n ) + weighted_average * np.log(denom).sum()\n hessian = hessian + weighted_average * (a2 - a1)\n pos -= count_of_removals\n return hessian, gradient, log_lik\n",
"role": "user"
},
{
"content": "get_efron_values_batch",
"role": "assistant"
}
] |
def x(self, X, T, E, weights, index=None):
n_deaths = self.event_observed.sum()
scaled_schoenfeld_resids = n_deaths * self._compute_schoenfeld(X, T, E,
weights, index).dot(self.variance_matrix_)
scaled_schoenfeld_resids.columns = self.hazards_.index
return scaled_schoenfeld_resids
| compute_scaled_schoenfeld | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, T, E, weights, index=None):\n n_deaths = self.event_observed.sum()\n scaled_schoenfeld_resids = n_deaths * self._compute_schoenfeld(X, T, E,\n weights, index).dot(self.variance_matrix_)\n scaled_schoenfeld_resids.columns = self.hazards_.index\n return scaled_schoenfeld_resids\n",
"role": "user"
},
{
"content": "compute_scaled_schoenfeld",
"role": "assistant"
}
] |
def x(self, X, T, E, weights):
n, d = X.shape
if not np.any(E):
return np.zeros((n, d))
risk_phi, tie_phi = 0, 0
risk_phi_x, tie_phi_x = np.zeros((1, d)), np.zeros((1, d))
weight_count = 0.0
tie_count = 0
scores = weights * np.exp(np.dot(X, self.hazards_))
diff_against = []
schoenfeld_residuals = np.empty((0, d))
for i in range(n - 1, -1, -1):
ti = T[i]
ei = E[i]
xi = X[i:i + 1]
score = scores[i:i + 1]
w = weights[i]
phi_i = score
phi_x_i = phi_i * xi
risk_phi = risk_phi + phi_i
risk_phi_x = risk_phi_x + phi_x_i
diff_against.append((xi, ei))
if ei:
tie_phi = tie_phi + phi_i
tie_phi_x = tie_phi_x + phi_x_i
tie_count += 1
weight_count += w
if i > 0 and T[i - 1] == ti:
continue
elif tie_count == 0:
for _ in diff_against:
schoenfeld_residuals = np.append(schoenfeld_residuals, np.
zeros((1, d)), axis=0)
diff_against = []
continue
weighted_mean = np.zeros((1, d))
for l in range(tie_count):
numer = risk_phi_x - l * tie_phi_x / tie_count
denom = risk_phi - l * tie_phi / tie_count
weighted_mean += numer / (denom * tie_count)
for xi, ei in diff_against:
schoenfeld_residuals = np.append(schoenfeld_residuals, ei * (xi -
weighted_mean), axis=0)
tie_count = 0
weight_count = 0.0
tie_phi = 0
tie_phi_x = np.zeros((1, d))
diff_against = []
return schoenfeld_residuals[::-1]
| compute_schoenfeld_within_strata | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, T, E, weights):\n n, d = X.shape\n if not np.any(E):\n return np.zeros((n, d))\n risk_phi, tie_phi = 0, 0\n risk_phi_x, tie_phi_x = np.zeros((1, d)), np.zeros((1, d))\n weight_count = 0.0\n tie_count = 0\n scores = weights * np.exp(np.dot(X, self.hazards_))\n diff_against = []\n schoenfeld_residuals = np.empty((0, d))\n for i in range(n - 1, -1, -1):\n ti = T[i]\n ei = E[i]\n xi = X[i:i + 1]\n score = scores[i:i + 1]\n w = weights[i]\n phi_i = score\n phi_x_i = phi_i * xi\n risk_phi = risk_phi + phi_i\n risk_phi_x = risk_phi_x + phi_x_i\n diff_against.append((xi, ei))\n if ei:\n tie_phi = tie_phi + phi_i\n tie_phi_x = tie_phi_x + phi_x_i\n tie_count += 1\n weight_count += w\n if i > 0 and T[i - 1] == ti:\n continue\n elif tie_count == 0:\n for _ in diff_against:\n schoenfeld_residuals = np.append(schoenfeld_residuals, np.\n zeros((1, d)), axis=0)\n diff_against = []\n continue\n weighted_mean = np.zeros((1, d))\n for l in range(tie_count):\n numer = risk_phi_x - l * tie_phi_x / tie_count\n denom = risk_phi - l * tie_phi / tie_count\n weighted_mean += numer / (denom * tie_count)\n for xi, ei in diff_against:\n schoenfeld_residuals = np.append(schoenfeld_residuals, ei * (xi -\n weighted_mean), axis=0)\n tie_count = 0\n weight_count = 0.0\n tie_phi = 0\n tie_phi_x = np.zeros((1, d))\n diff_against = []\n return schoenfeld_residuals[::-1]\n",
"role": "user"
},
{
"content": "compute_schoenfeld_within_strata",
"role": "assistant"
}
] |
def x(self, X, T, E, weights, index=None):
score_residuals = self._compute_score(X, T, E, weights, index=index)
d = X.shape[1]
scaled_variance_matrix = self.variance_matrix_ * np.tile(self._norm_std
.values, (d, 1)).T
delta_betas = score_residuals.dot(scaled_variance_matrix)
delta_betas.columns = self.hazards_.index
return delta_betas
| compute_delta_beta | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, T, E, weights, index=None):\n score_residuals = self._compute_score(X, T, E, weights, index=index)\n d = X.shape[1]\n scaled_variance_matrix = self.variance_matrix_ * np.tile(self._norm_std\n .values, (d, 1)).T\n delta_betas = score_residuals.dot(scaled_variance_matrix)\n delta_betas.columns = self.hazards_.index\n return delta_betas\n",
"role": "user"
},
{
"content": "compute_delta_beta",
"role": "assistant"
}
] |
def x(self):
ci = 1 - self.alpha
with np.errstate(invalid='ignore', divide='ignore'):
df = pd.DataFrame(index=self.hazards_.index)
df['coef'] = self.hazards_
df['exp(coef)'] = np.exp(self.hazards_)
df['se(coef)'] = self.standard_errors_
df['z'] = self._compute_z_values()
df['p'] = self._compute_p_values()
df['-log2(p)'] = -np.log2(df['p'])
df['lower %g' % ci] = self.confidence_intervals_['lower-bound']
df['upper %g' % ci] = self.confidence_intervals_['upper-bound']
return df
| summary | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n ci = 1 - self.alpha\n with np.errstate(invalid='ignore', divide='ignore'):\n df = pd.DataFrame(index=self.hazards_.index)\n df['coef'] = self.hazards_\n df['exp(coef)'] = np.exp(self.hazards_)\n df['se(coef)'] = self.standard_errors_\n df['z'] = self._compute_z_values()\n df['p'] = self._compute_p_values()\n df['-log2(p)'] = -np.log2(df['p'])\n df['lower %g' % ci] = self.confidence_intervals_['lower-bound']\n df['upper %g' % ci] = self.confidence_intervals_['upper-bound']\n return df\n",
"role": "user"
},
{
"content": "summary",
"role": "assistant"
}
] |
def x(self, X):
hazard_names = self.hazards_.index
if isinstance(X, pd.Series) and (X.shape[0] == len(hazard_names) + 2 or
X.shape[0] == len(hazard_names)):
X = X.to_frame().T
return self.predict_log_partial_hazard(X)
elif isinstance(X, pd.Series):
assert len(hazard_names) == 1, 'Series not the correct argument'
X = X.to_frame().T
return self.predict_log_partial_hazard(X)
index = _get_index(X)
if isinstance(X, pd.DataFrame):
order = hazard_names
X = X.reindex(order, axis='columns')
check_for_numeric_dtypes_or_raise(X)
X = X.values
X = X.astype(float)
X = normalize(X, self._norm_mean.values, 1)
return pd.DataFrame(np.dot(X, self.hazards_), index=index)
| predict_log_partial_hazard | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n hazard_names = self.hazards_.index\n if isinstance(X, pd.Series) and (X.shape[0] == len(hazard_names) + 2 or\n X.shape[0] == len(hazard_names)):\n X = X.to_frame().T\n return self.predict_log_partial_hazard(X)\n elif isinstance(X, pd.Series):\n assert len(hazard_names) == 1, 'Series not the correct argument'\n X = X.to_frame().T\n return self.predict_log_partial_hazard(X)\n index = _get_index(X)\n if isinstance(X, pd.DataFrame):\n order = hazard_names\n X = X.reindex(order, axis='columns')\n check_for_numeric_dtypes_or_raise(X)\n X = X.values\n X = X.astype(float)\n X = normalize(X, self._norm_mean.values, 1)\n return pd.DataFrame(np.dot(X, self.hazards_), index=index)\n",
"role": "user"
},
{
"content": "predict_log_partial_hazard",
"role": "assistant"
}
] |
def x(self, X, times=None):
if self.strata:
cumulative_hazard_ = pd.DataFrame()
for stratum, stratified_X in X.groupby(self.strata):
try:
c_0 = self.baseline_cumulative_hazard_[[stratum]]
except KeyError:
raise StatError(
"""The stratum %s was not found in the original training data. For example, try
the following on the original dataset, df: `df.groupby(%s).size()`. Expected is that %s is not present in the output.
"""
% (stratum, self.strata, stratum))
col = _get_index(stratified_X)
v = self.predict_partial_hazard(stratified_X)
cumulative_hazard_ = cumulative_hazard_.merge(pd.DataFrame(np.
dot(c_0, v.T), index=c_0.index, columns=col), how='outer',
right_index=True, left_index=True)
else:
c_0 = self.baseline_cumulative_hazard_
v = self.predict_partial_hazard(X)
col = _get_index(v)
cumulative_hazard_ = pd.DataFrame(np.dot(c_0, v.T), columns=col,
index=c_0.index)
if times is not None:
return dataframe_interpolate_at_times(cumulative_hazard_, times)
return cumulative_hazard_
| predict_cumulative_hazard | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, times=None):\n if self.strata:\n cumulative_hazard_ = pd.DataFrame()\n for stratum, stratified_X in X.groupby(self.strata):\n try:\n c_0 = self.baseline_cumulative_hazard_[[stratum]]\n except KeyError:\n raise StatError(\n \"\"\"The stratum %s was not found in the original training data. For example, try\nthe following on the original dataset, df: `df.groupby(%s).size()`. Expected is that %s is not present in the output.\n\"\"\"\n % (stratum, self.strata, stratum))\n col = _get_index(stratified_X)\n v = self.predict_partial_hazard(stratified_X)\n cumulative_hazard_ = cumulative_hazard_.merge(pd.DataFrame(np.\n dot(c_0, v.T), index=c_0.index, columns=col), how='outer',\n right_index=True, left_index=True)\n else:\n c_0 = self.baseline_cumulative_hazard_\n v = self.predict_partial_hazard(X)\n col = _get_index(v)\n cumulative_hazard_ = pd.DataFrame(np.dot(c_0, v.T), columns=col,\n index=c_0.index)\n if times is not None:\n return dataframe_interpolate_at_times(cumulative_hazard_, times)\n return cumulative_hazard_\n",
"role": "user"
},
{
"content": "predict_cumulative_hazard",
"role": "assistant"
}
] |
def x(self, X, times=None):
return np.exp(-self.predict_cumulative_hazard(X, times=times))
| predict_survival_function | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, times=None):\n return np.exp(-self.predict_cumulative_hazard(X, times=times))\n",
"role": "user"
},
{
"content": "predict_survival_function",
"role": "assistant"
}
] |
def x(self, X, p=0.5):
subjects = _get_index(X)
return qth_survival_times(p, self.predict_survival_function(X)[subjects]).T
| predict_percentile | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, p=0.5):\n subjects = _get_index(X)\n return qth_survival_times(p, self.predict_survival_function(X)[subjects]).T\n",
"role": "user"
},
{
"content": "predict_percentile",
"role": "assistant"
}
] |
def x(self, X):
subjects = _get_index(X)
v = self.predict_survival_function(X)[subjects]
return pd.DataFrame(trapz(v.values.T, v.index), index=subjects)
| predict_expectation | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X):\n subjects = _get_index(X)\n v = self.predict_survival_function(X)[subjects]\n return pd.DataFrame(trapz(v.values.T, v.index), index=subjects)\n",
"role": "user"
},
{
"content": "predict_expectation",
"role": "assistant"
}
] |
def x(self):
survival_df = np.exp(-self.baseline_cumulative_hazard_)
if self.strata is None:
survival_df.columns = ['baseline survival']
return survival_df
| compute_baseline_survival | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n survival_df = np.exp(-self.baseline_cumulative_hazard_)\n if self.strata is None:\n survival_df.columns = ['baseline survival']\n return survival_df\n",
"role": "user"
},
{
"content": "compute_baseline_survival",
"role": "assistant"
}
] |
def x(self, covariates, values, plot_baseline=True, **kwargs):
from matplotlib import pyplot as plt
covariates = _to_list(covariates)
n_covariates = len(covariates)
values = _to_array(values)
if len(values.shape) == 1:
values = values[None, :].T
if n_covariates != values.shape[1]:
raise ValueError(
'The number of covariates must equal to second dimension of the values array.'
)
for covariate in covariates:
if covariate not in self.hazards_.index:
raise KeyError(
'covariate `%s` is not present in the original dataset' %
covariate)
set_kwargs_drawstyle(kwargs, 'steps-post')
if self.strata is None:
axes = kwargs.pop('ax', None) or plt.figure().add_subplot(111)
x_bar = self._norm_mean.to_frame().T
X = pd.concat([x_bar] * values.shape[0])
if np.array_equal(np.eye(n_covariates), values):
X.index = [('%s=1' % c) for c in covariates]
else:
X.index = [', '.join('%s=%g' % (c, v) for c, v in zip(
covariates, row)) for row in values]
for covariate, value in zip(covariates, values.T):
X[covariate] = value
self.predict_survival_function(X).plot(ax=axes, **kwargs)
if plot_baseline:
self.baseline_survival_.plot(ax=axes, ls=':', color='k',
drawstyle='steps-post')
else:
axes = []
for stratum, baseline_survival_ in self.baseline_survival_.iteritems():
ax = plt.figure().add_subplot(1, 1, 1)
x_bar = self._norm_mean.to_frame().T
for name, value in zip(_to_list(self.strata), _to_tuple(stratum)):
x_bar[name] = value
X = pd.concat([x_bar] * values.shape[0])
if np.array_equal(np.eye(len(covariates)), values):
X.index = [('%s=1' % c) for c in covariates]
else:
X.index = [', '.join('%s=%g' % (c, v) for c, v in zip(
covariates, row)) for row in values]
for covariate, value in zip(covariates, values.T):
X[covariate] = value
self.predict_survival_function(X).plot(ax=ax, **kwargs)
if plot_baseline:
baseline_survival_.plot(ax=ax, ls=':', label=
'stratum %s baseline survival' % str(stratum),
drawstyle='steps-post')
plt.legend()
axes.append(ax)
return axes
| plot_covariate_groups | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, covariates, values, plot_baseline=True, **kwargs):\n from matplotlib import pyplot as plt\n covariates = _to_list(covariates)\n n_covariates = len(covariates)\n values = _to_array(values)\n if len(values.shape) == 1:\n values = values[None, :].T\n if n_covariates != values.shape[1]:\n raise ValueError(\n 'The number of covariates must equal to second dimension of the values array.'\n )\n for covariate in covariates:\n if covariate not in self.hazards_.index:\n raise KeyError(\n 'covariate `%s` is not present in the original dataset' %\n covariate)\n set_kwargs_drawstyle(kwargs, 'steps-post')\n if self.strata is None:\n axes = kwargs.pop('ax', None) or plt.figure().add_subplot(111)\n x_bar = self._norm_mean.to_frame().T\n X = pd.concat([x_bar] * values.shape[0])\n if np.array_equal(np.eye(n_covariates), values):\n X.index = [('%s=1' % c) for c in covariates]\n else:\n X.index = [', '.join('%s=%g' % (c, v) for c, v in zip(\n covariates, row)) for row in values]\n for covariate, value in zip(covariates, values.T):\n X[covariate] = value\n self.predict_survival_function(X).plot(ax=axes, **kwargs)\n if plot_baseline:\n self.baseline_survival_.plot(ax=axes, ls=':', color='k',\n drawstyle='steps-post')\n else:\n axes = []\n for stratum, baseline_survival_ in self.baseline_survival_.iteritems():\n ax = plt.figure().add_subplot(1, 1, 1)\n x_bar = self._norm_mean.to_frame().T\n for name, value in zip(_to_list(self.strata), _to_tuple(stratum)):\n x_bar[name] = value\n X = pd.concat([x_bar] * values.shape[0])\n if np.array_equal(np.eye(len(covariates)), values):\n X.index = [('%s=1' % c) for c in covariates]\n else:\n X.index = [', '.join('%s=%g' % (c, v) for c, v in zip(\n covariates, row)) for row in values]\n for covariate, value in zip(covariates, values.T):\n X[covariate] = value\n self.predict_survival_function(X).plot(ax=ax, **kwargs)\n if plot_baseline:\n baseline_survival_.plot(ax=ax, ls=':', label=\n 'stratum %s baseline survival' % str(stratum),\n drawstyle='steps-post')\n plt.legend()\n axes.append(ax)\n return axes\n",
"role": "user"
},
{
"content": "plot_covariate_groups",
"role": "assistant"
}
] |
def x(self, training_df, advice=True, show_plots=False, p_value_threshold=
0.01, plot_n_bootstraps=10):
if not training_df.index.is_unique:
raise IndexError(
'`training_df` index should be unique for this exercise. Please make it unique or use `.reset_index(drop=True)` to force a unique index'
)
residuals = self.compute_residuals(training_df, kind='scaled_schoenfeld')
test_results = proportional_hazard_test(self, training_df,
time_transform=['rank', 'km'], precomputed_residuals=residuals)
residuals_and_duration = residuals.join(training_df[self.duration_col])
counter = 0
n = residuals_and_duration.shape[0]
for variable in self.hazards_.index:
minumum_observed_p_value = test_results.summary.loc[variable, 'p'].min(
)
if np.round(minumum_observed_p_value, 2) > p_value_threshold:
continue
counter += 1
if counter == 1:
if advice:
print(fill(
'The ``p_value_threshold`` is set at %g. Even under the null hypothesis of no violations, some covariates will be below the threshold by chance. This is compounded when there are many covariates. Similarly, when there are lots of observations, even minor deviances from the proportional hazard assumption will be flagged.'
% p_value_threshold, width=100))
print()
print(fill(
"With that in mind, it's best to use a combination of statistical tests and visual tests to determine the most serious violations. Produce visual plots using ``check_assumptions(..., show_plots=True)`` and looking for non-constant lines. See link [A] below for a full example."
, width=100))
print()
test_results.print_summary()
print()
print()
print(
"%d. Variable '%s' failed the non-proportional test: p-value is %s."
% (counter, variable, format_p_value(4)(
minumum_observed_p_value)), end='\n\n')
if advice:
values = training_df[variable]
value_counts = values.value_counts()
n_uniques = value_counts.shape[0]
if n_uniques <= 10 and value_counts.min() >= 5:
print(fill(
" Advice: with so few unique values (only {0}), you can include `strata=['{1}', ...]` in the call in `.fit`. See documentation in link [E] below."
.format(n_uniques, variable), width=100))
else:
print(fill(
" Advice 1: the functional form of the variable '{var}' might be incorrect. That is, there may be non-linear terms missing. The proportional hazard test used is very sensitive to incorrect functional forms. See documentation in link [D] below on how to specify a functional form."
.format(var=variable), width=100), end='\n\n')
print(fill(
" Advice 2: try binning the variable '{var}' using pd.cut, and then specify it in `strata=['{var}', ...]` in the call in `.fit`. See documentation in link [B] below."
.format(var=variable), width=100), end='\n\n')
print(fill(
' Advice 3: try adding an interaction term with your time variable. See documentation in link [C] below.'
.format(var=variable), width=100), end='\n\n')
if show_plots:
from matplotlib import pyplot as plt
fig = plt.figure()
for i, (transform_name, transformer) in enumerate(TimeTransformers
().iter(['rank', 'km']), start=1):
p_value = test_results.summary.loc[(variable,
transform_name), 'p']
ax = fig.add_subplot(1, 2, i)
y = residuals_and_duration[variable]
tt = transformer(self.durations, self.event_observed, self.
weights)[self.event_observed.values]
ax.scatter(tt, y, alpha=0.75)
y_lowess = lowess(tt.values, y.values)
ax.plot(tt, y_lowess, color='k', alpha=1.0, linewidth=2)
for _ in range(plot_n_bootstraps):
ix = sorted(np.random.choice(n, n))
tt_ = tt.values[ix]
y_lowess = lowess(tt_, y.values[ix])
ax.plot(tt_, y_lowess, color='k', alpha=0.3)
best_xlim = ax.get_xlim()
ax.hlines(0, 0, tt.max(), linestyles='dashed', linewidths=1)
ax.set_xlim(best_xlim)
ax.set_xlabel('%s-transformed time\n(p=%.4f)' % (
transform_name, p_value), fontsize=10)
fig.suptitle("Scaled Schoenfeld residuals of '%s'" % variable,
fontsize=14)
plt.tight_layout()
plt.subplots_adjust(top=0.9)
if advice and counter > 0:
print(dedent(
"""
---
[A] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html
[B] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html#Bin-variable-and-stratify-on-it
[C] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html#Introduce-time-varying-covariates
[D] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html#Modify-the-functional-form
[E] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html#Stratification
"""
))
if counter == 0:
print('Proportional hazard assumption looks okay.')
| check_assumptions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, training_df, advice=True, show_plots=False, p_value_threshold=\n 0.01, plot_n_bootstraps=10):\n if not training_df.index.is_unique:\n raise IndexError(\n '`training_df` index should be unique for this exercise. Please make it unique or use `.reset_index(drop=True)` to force a unique index'\n )\n residuals = self.compute_residuals(training_df, kind='scaled_schoenfeld')\n test_results = proportional_hazard_test(self, training_df,\n time_transform=['rank', 'km'], precomputed_residuals=residuals)\n residuals_and_duration = residuals.join(training_df[self.duration_col])\n counter = 0\n n = residuals_and_duration.shape[0]\n for variable in self.hazards_.index:\n minumum_observed_p_value = test_results.summary.loc[variable, 'p'].min(\n )\n if np.round(minumum_observed_p_value, 2) > p_value_threshold:\n continue\n counter += 1\n if counter == 1:\n if advice:\n print(fill(\n 'The ``p_value_threshold`` is set at %g. Even under the null hypothesis of no violations, some covariates will be below the threshold by chance. This is compounded when there are many covariates. Similarly, when there are lots of observations, even minor deviances from the proportional hazard assumption will be flagged.'\n % p_value_threshold, width=100))\n print()\n print(fill(\n \"With that in mind, it's best to use a combination of statistical tests and visual tests to determine the most serious violations. Produce visual plots using ``check_assumptions(..., show_plots=True)`` and looking for non-constant lines. See link [A] below for a full example.\"\n , width=100))\n print()\n test_results.print_summary()\n print()\n print()\n print(\n \"%d. Variable '%s' failed the non-proportional test: p-value is %s.\"\n % (counter, variable, format_p_value(4)(\n minumum_observed_p_value)), end='\\n\\n')\n if advice:\n values = training_df[variable]\n value_counts = values.value_counts()\n n_uniques = value_counts.shape[0]\n if n_uniques <= 10 and value_counts.min() >= 5:\n print(fill(\n \" Advice: with so few unique values (only {0}), you can include `strata=['{1}', ...]` in the call in `.fit`. See documentation in link [E] below.\"\n .format(n_uniques, variable), width=100))\n else:\n print(fill(\n \" Advice 1: the functional form of the variable '{var}' might be incorrect. That is, there may be non-linear terms missing. The proportional hazard test used is very sensitive to incorrect functional forms. See documentation in link [D] below on how to specify a functional form.\"\n .format(var=variable), width=100), end='\\n\\n')\n print(fill(\n \" Advice 2: try binning the variable '{var}' using pd.cut, and then specify it in `strata=['{var}', ...]` in the call in `.fit`. See documentation in link [B] below.\"\n .format(var=variable), width=100), end='\\n\\n')\n print(fill(\n ' Advice 3: try adding an interaction term with your time variable. See documentation in link [C] below.'\n .format(var=variable), width=100), end='\\n\\n')\n if show_plots:\n from matplotlib import pyplot as plt\n fig = plt.figure()\n for i, (transform_name, transformer) in enumerate(TimeTransformers\n ().iter(['rank', 'km']), start=1):\n p_value = test_results.summary.loc[(variable,\n transform_name), 'p']\n ax = fig.add_subplot(1, 2, i)\n y = residuals_and_duration[variable]\n tt = transformer(self.durations, self.event_observed, self.\n weights)[self.event_observed.values]\n ax.scatter(tt, y, alpha=0.75)\n y_lowess = lowess(tt.values, y.values)\n ax.plot(tt, y_lowess, color='k', alpha=1.0, linewidth=2)\n for _ in range(plot_n_bootstraps):\n ix = sorted(np.random.choice(n, n))\n tt_ = tt.values[ix]\n y_lowess = lowess(tt_, y.values[ix])\n ax.plot(tt_, y_lowess, color='k', alpha=0.3)\n best_xlim = ax.get_xlim()\n ax.hlines(0, 0, tt.max(), linestyles='dashed', linewidths=1)\n ax.set_xlim(best_xlim)\n ax.set_xlabel('%s-transformed time\\n(p=%.4f)' % (\n transform_name, p_value), fontsize=10)\n fig.suptitle(\"Scaled Schoenfeld residuals of '%s'\" % variable,\n fontsize=14)\n plt.tight_layout()\n plt.subplots_adjust(top=0.9)\n if advice and counter > 0:\n print(dedent(\n \"\"\"\n ---\n [A] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html\n [B] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html#Bin-variable-and-stratify-on-it\n [C] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html#Introduce-time-varying-covariates\n [D] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html#Modify-the-functional-form\n [E] https://lifelines.readthedocs.io/en/latest/jupyter_notebooks/Proportional%20hazard%20assumption.html#Stratification\n \"\"\"\n ))\n if counter == 0:\n print('Proportional hazard assumption looks okay.')\n",
"role": "user"
},
{
"content": "check_assumptions",
"role": "assistant"
}
] |
def x(self):
if not hasattr(self, '_concordance_score_'):
if self.strata:
num_correct, num_tied, num_pairs = 0, 0, 0
for _, _df in self._predicted_partial_hazards_.groupby(self.strata
):
if _df.shape[0] == 1:
continue
_num_correct, _num_tied, _num_pairs = (
_concordance_summary_statistics(_df['T'].values, -_df[
'P'].values, _df['E'].values))
num_correct += _num_correct
num_tied += _num_tied
num_pairs += _num_pairs
else:
df = self._predicted_partial_hazards_
num_correct, num_tied, num_pairs = _concordance_summary_statistics(
df['T'].values, -df['P'].values, df['E'].values)
self._concordance_score_ = _concordance_ratio(num_correct, num_tied,
num_pairs)
return self._concordance_score_
return self._concordance_score_
| score_ | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if not hasattr(self, '_concordance_score_'):\n if self.strata:\n num_correct, num_tied, num_pairs = 0, 0, 0\n for _, _df in self._predicted_partial_hazards_.groupby(self.strata\n ):\n if _df.shape[0] == 1:\n continue\n _num_correct, _num_tied, _num_pairs = (\n _concordance_summary_statistics(_df['T'].values, -_df[\n 'P'].values, _df['E'].values))\n num_correct += _num_correct\n num_tied += _num_tied\n num_pairs += _num_pairs\n else:\n df = self._predicted_partial_hazards_\n num_correct, num_tied, num_pairs = _concordance_summary_statistics(\n df['T'].values, -df['P'].values, df['E'].values)\n self._concordance_score_ = _concordance_ratio(num_correct, num_tied,\n num_pairs)\n return self._concordance_score_\n return self._concordance_score_\n",
"role": "user"
},
{
"content": "score_",
"role": "assistant"
}
] |
def x(event_times, predicted_scores, event_observed=None):
event_times = np.asarray(event_times, dtype=float)
predicted_scores = np.asarray(predicted_scores, dtype=float)
if event_times.ndim == 2 and (event_times.shape[0] == 1 or event_times.
shape[1] == 1):
event_times = event_times.ravel()
if predicted_scores.ndim == 2 and (predicted_scores.shape[0] == 1 or
predicted_scores.shape[1] == 1):
predicted_scores = predicted_scores.ravel()
if event_times.shape != predicted_scores.shape:
raise ValueError('Event times and predictions must have the same shape'
)
if event_times.ndim != 1:
raise ValueError('Event times can only be 1-dimensional: (n,)')
if event_observed is None:
event_observed = np.ones(event_times.shape[0], dtype=float)
else:
event_observed = np.asarray(event_observed, dtype=float).ravel()
if event_observed.shape != event_times.shape:
raise ValueError(
'Observed events must be 1-dimensional of same length as event times'
)
num_correct, num_tied, num_pairs = _concordance_summary_statistics(
event_times, predicted_scores, event_observed)
return _concordance_ratio(num_correct, num_tied, num_pairs)
| concordance_index | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(event_times, predicted_scores, event_observed=None):\n event_times = np.asarray(event_times, dtype=float)\n predicted_scores = np.asarray(predicted_scores, dtype=float)\n if event_times.ndim == 2 and (event_times.shape[0] == 1 or event_times.\n shape[1] == 1):\n event_times = event_times.ravel()\n if predicted_scores.ndim == 2 and (predicted_scores.shape[0] == 1 or \n predicted_scores.shape[1] == 1):\n predicted_scores = predicted_scores.ravel()\n if event_times.shape != predicted_scores.shape:\n raise ValueError('Event times and predictions must have the same shape'\n )\n if event_times.ndim != 1:\n raise ValueError('Event times can only be 1-dimensional: (n,)')\n if event_observed is None:\n event_observed = np.ones(event_times.shape[0], dtype=float)\n else:\n event_observed = np.asarray(event_observed, dtype=float).ravel()\n if event_observed.shape != event_times.shape:\n raise ValueError(\n 'Observed events must be 1-dimensional of same length as event times'\n )\n num_correct, num_tied, num_pairs = _concordance_summary_statistics(\n event_times, predicted_scores, event_observed)\n return _concordance_ratio(num_correct, num_tied, num_pairs)\n",
"role": "user"
},
{
"content": "concordance_index",
"role": "assistant"
}
] |
def x(event_times, predicted_event_times, event_observed):
if np.logical_not(event_observed).all():
return 0, 0, 0
died_mask = event_observed.astype(bool)
died_truth = event_times[died_mask]
ix = np.argsort(died_truth)
died_truth = died_truth[ix]
died_pred = predicted_event_times[died_mask][ix]
censored_truth = event_times[~died_mask]
ix = np.argsort(censored_truth)
censored_truth = censored_truth[ix]
censored_pred = predicted_event_times[~died_mask][ix]
censored_ix = 0
died_ix = 0
times_to_compare = _BTree(np.unique(died_pred))
num_pairs = np.int64(0)
num_correct = np.int64(0)
num_tied = np.int64(0)
while True:
has_more_censored = censored_ix < len(censored_truth)
has_more_died = died_ix < len(died_truth)
if has_more_censored and (not has_more_died or died_truth[died_ix] >
censored_truth[censored_ix]):
pairs, correct, tied, next_ix = _handle_pairs(censored_truth,
censored_pred, censored_ix, times_to_compare)
censored_ix = next_ix
elif has_more_died and (not has_more_censored or died_truth[died_ix
] <= censored_truth[censored_ix]):
pairs, correct, tied, next_ix = _handle_pairs(died_truth,
died_pred, died_ix, times_to_compare)
for pred in died_pred[died_ix:next_ix]:
times_to_compare.insert(pred)
died_ix = next_ix
else:
assert not (has_more_died or has_more_censored)
break
num_pairs += pairs
num_correct += correct
num_tied += tied
return num_correct, num_tied, num_pairs
| concordance_summary_statistics | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(event_times, predicted_event_times, event_observed):\n if np.logical_not(event_observed).all():\n return 0, 0, 0\n died_mask = event_observed.astype(bool)\n died_truth = event_times[died_mask]\n ix = np.argsort(died_truth)\n died_truth = died_truth[ix]\n died_pred = predicted_event_times[died_mask][ix]\n censored_truth = event_times[~died_mask]\n ix = np.argsort(censored_truth)\n censored_truth = censored_truth[ix]\n censored_pred = predicted_event_times[~died_mask][ix]\n censored_ix = 0\n died_ix = 0\n times_to_compare = _BTree(np.unique(died_pred))\n num_pairs = np.int64(0)\n num_correct = np.int64(0)\n num_tied = np.int64(0)\n while True:\n has_more_censored = censored_ix < len(censored_truth)\n has_more_died = died_ix < len(died_truth)\n if has_more_censored and (not has_more_died or died_truth[died_ix] >\n censored_truth[censored_ix]):\n pairs, correct, tied, next_ix = _handle_pairs(censored_truth,\n censored_pred, censored_ix, times_to_compare)\n censored_ix = next_ix\n elif has_more_died and (not has_more_censored or died_truth[died_ix\n ] <= censored_truth[censored_ix]):\n pairs, correct, tied, next_ix = _handle_pairs(died_truth,\n died_pred, died_ix, times_to_compare)\n for pred in died_pred[died_ix:next_ix]:\n times_to_compare.insert(pred)\n died_ix = next_ix\n else:\n assert not (has_more_died or has_more_censored)\n break\n num_pairs += pairs\n num_correct += correct\n num_tied += tied\n return num_correct, num_tied, num_pairs\n",
"role": "user"
},
{
"content": "concordance_summary_statistics",
"role": "assistant"
}
] |
def x(truth, pred, first_ix, times_to_compare):
next_ix = first_ix
while next_ix < len(truth) and truth[next_ix] == truth[first_ix]:
next_ix += 1
pairs = len(times_to_compare) * (next_ix - first_ix)
correct = np.int64(0)
tied = np.int64(0)
for i in range(first_ix, next_ix):
rank, count = times_to_compare.rank(pred[i])
correct += rank
tied += count
return pairs, correct, tied, next_ix
| handle_pairs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(truth, pred, first_ix, times_to_compare):\n next_ix = first_ix\n while next_ix < len(truth) and truth[next_ix] == truth[first_ix]:\n next_ix += 1\n pairs = len(times_to_compare) * (next_ix - first_ix)\n correct = np.int64(0)\n tied = np.int64(0)\n for i in range(first_ix, next_ix):\n rank, count = times_to_compare.rank(pred[i])\n correct += rank\n tied += count\n return pairs, correct, tied, next_ix\n",
"role": "user"
},
{
"content": "handle_pairs",
"role": "assistant"
}
] |
def x(event_times, predicted_event_times, event_observed):
num_pairs = 0.0
num_correct = 0.0
num_tied = 0.0
for a, time_a in enumerate(event_times):
pred_a = predicted_event_times[a]
event_a = event_observed[a]
for b in range(a + 1, len(event_times)):
time_b = event_times[b]
pred_b = predicted_event_times[b]
event_b = event_observed[b]
if _valid_comparison(time_a, time_b, event_a, event_b):
num_pairs += 1.0
crct, ties = _concordance_value(time_a, time_b, pred_a,
pred_b, event_a, event_b)
num_correct += crct
num_tied += ties
return num_correct, num_tied, num_pairs
| naive_concordance_summary_statistics | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(event_times, predicted_event_times, event_observed):\n num_pairs = 0.0\n num_correct = 0.0\n num_tied = 0.0\n for a, time_a in enumerate(event_times):\n pred_a = predicted_event_times[a]\n event_a = event_observed[a]\n for b in range(a + 1, len(event_times)):\n time_b = event_times[b]\n pred_b = predicted_event_times[b]\n event_b = event_observed[b]\n if _valid_comparison(time_a, time_b, event_a, event_b):\n num_pairs += 1.0\n crct, ties = _concordance_value(time_a, time_b, pred_a,\n pred_b, event_a, event_b)\n num_correct += crct\n num_tied += ties\n return num_correct, num_tied, num_pairs\n",
"role": "user"
},
{
"content": "naive_concordance_summary_statistics",
"role": "assistant"
}
] |
def x(time_a, time_b, event_a, event_b):
if time_a == time_b:
return event_a != event_b
if event_a and event_b:
return True
if event_a and time_a < time_b:
return True
if event_b and time_b < time_a:
return True
return False
| valid_comparison | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(time_a, time_b, event_a, event_b):\n if time_a == time_b:\n return event_a != event_b\n if event_a and event_b:\n return True\n if event_a and time_a < time_b:\n return True\n if event_b and time_b < time_a:\n return True\n return False\n",
"role": "user"
},
{
"content": "valid_comparison",
"role": "assistant"
}
] |
def x(self, X, ancillary_X=None, p=0.5):
exp_mu_, sigma_ = self._prep_inputs_for_prediction_and_return_scores(X,
ancillary_X)
return pd.DataFrame(exp_mu_ * np.exp(np.sqrt(2) * sigma_ * erfinv(2 * p -
1)), index=_get_index(X))
| predict_percentile | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, ancillary_X=None, p=0.5):\n exp_mu_, sigma_ = self._prep_inputs_for_prediction_and_return_scores(X,\n ancillary_X)\n return pd.DataFrame(exp_mu_ * np.exp(np.sqrt(2) * sigma_ * erfinv(2 * p -\n 1)), index=_get_index(X))\n",
"role": "user"
},
{
"content": "predict_percentile",
"role": "assistant"
}
] |
def x(self, X, ancillary_X=None):
exp_mu_, _ = self._prep_inputs_for_prediction_and_return_scores(X,
ancillary_X)
return pd.DataFrame(exp_mu_, index=_get_index(X))
| predict_median | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, ancillary_X=None):\n exp_mu_, _ = self._prep_inputs_for_prediction_and_return_scores(X,\n ancillary_X)\n return pd.DataFrame(exp_mu_, index=_get_index(X))\n",
"role": "user"
},
{
"content": "predict_median",
"role": "assistant"
}
] |
def x(self, X, ancillary_X=None):
exp_mu_, sigma_ = self._prep_inputs_for_prediction_and_return_scores(X,
ancillary_X)
return pd.DataFrame(exp_mu_ * np.exp(sigma_ ** 2 / 2), index=_get_index(X))
| predict_expectation | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, ancillary_X=None):\n exp_mu_, sigma_ = self._prep_inputs_for_prediction_and_return_scores(X,\n ancillary_X)\n return pd.DataFrame(exp_mu_ * np.exp(sigma_ ** 2 / 2), index=_get_index(X))\n",
"role": "user"
},
{
"content": "predict_expectation",
"role": "assistant"
}
] |
def x(self, X, times=None, ancillary_X=None):
import numpy as np
times = coalesce(times, self.timeline, np.unique(self.durations))
exp_mu_, sigma_ = self._prep_inputs_for_prediction_and_return_scores(X,
ancillary_X)
mu_ = np.log(exp_mu_)
Z = np.subtract.outer(np.log(times), mu_) / sigma_
return pd.DataFrame(-logsf(Z), columns=_get_index(X), index=times)
| predict_cumulative_hazard | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, times=None, ancillary_X=None):\n import numpy as np\n times = coalesce(times, self.timeline, np.unique(self.durations))\n exp_mu_, sigma_ = self._prep_inputs_for_prediction_and_return_scores(X,\n ancillary_X)\n mu_ = np.log(exp_mu_)\n Z = np.subtract.outer(np.log(times), mu_) / sigma_\n return pd.DataFrame(-logsf(Z), columns=_get_index(X), index=times)\n",
"role": "user"
},
{
"content": "predict_cumulative_hazard",
"role": "assistant"
}
] |
def x(self, other):
self_estimate = getattr(self, self._estimate_name)
other_estimate = getattr(other, other._estimate_name)
new_index = np.concatenate((other_estimate.index, self_estimate.index))
new_index = np.unique(new_index)
return pd.DataFrame(self_estimate.reindex(new_index, method='ffill').
values - other_estimate.reindex(new_index, method='ffill').values,
index=new_index, columns=['diff'])
| subtract | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, other):\n self_estimate = getattr(self, self._estimate_name)\n other_estimate = getattr(other, other._estimate_name)\n new_index = np.concatenate((other_estimate.index, self_estimate.index))\n new_index = np.unique(new_index)\n return pd.DataFrame(self_estimate.reindex(new_index, method='ffill').\n values - other_estimate.reindex(new_index, method='ffill').values,\n index=new_index, columns=['diff'])\n",
"role": "user"
},
{
"content": "subtract",
"role": "assistant"
}
] |
def x(self, times):
if callable(self._estimation_method):
return pd.DataFrame(self._estimation_method(_to_array(times)),
index=_to_array(times)).loc[times].squeeze()
estimate = getattr(self, self._estimation_method)
return dataframe_interpolate_at_times(estimate, times)
| predict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, times):\n if callable(self._estimation_method):\n return pd.DataFrame(self._estimation_method(_to_array(times)),\n index=_to_array(times)).loc[times].squeeze()\n estimate = getattr(self, self._estimation_method)\n return dataframe_interpolate_at_times(estimate, times)\n",
"role": "user"
},
{
"content": "predict",
"role": "assistant"
}
] |
def x(self):
age = self.survival_function_.index.values[:, None]
columns = ['%s - Conditional time remaining to event' % self._label]
return pd.DataFrame(qth_survival_times(self.survival_function_[self.
_label] * 0.5, self.survival_function_).sort_index(ascending=False)
.values, index=self.survival_function_.index, columns=columns) - age
| conditional_time_to_event_ | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n age = self.survival_function_.index.values[:, None]\n columns = ['%s - Conditional time remaining to event' % self._label]\n return pd.DataFrame(qth_survival_times(self.survival_function_[self.\n _label] * 0.5, self.survival_function_).sort_index(ascending=False)\n .values, index=self.survival_function_.index, columns=columns) - age\n",
"role": "user"
},
{
"content": "conditional_time_to_event_",
"role": "assistant"
}
] |
def x(self, transform, alpha, ci_labels):
alpha2 = 1 - alpha / 2.0
z = inv_normal_cdf(alpha2)
df = pd.DataFrame(index=self.timeline)
gradient_of_cum_hazard_at_mle = make_jvp_reversemode(transform)(self.
_fitted_parameters_, self.timeline.astype(float))
gradient_at_times = np.vstack([gradient_of_cum_hazard_at_mle(basis) for
basis in np.eye(len(self._fitted_parameters_), dtype=float)])
std_cumulative_hazard = np.sqrt(np.einsum('nj,jk,nk->n',
gradient_at_times.T, self.variance_matrix_, gradient_at_times.T))
if ci_labels is None:
ci_labels = ['%s_upper_%g' % (self._label, 1 - alpha),
'%s_lower_%g' % (self._label, 1 - alpha)]
assert len(ci_labels) == 2, 'ci_labels should be a length 2 array.'
df[ci_labels[0]] = transform(self._fitted_parameters_, self.timeline
) + z * std_cumulative_hazard
df[ci_labels[1]] = transform(self._fitted_parameters_, self.timeline
) - z * std_cumulative_hazard
return df
| compute_confidence_bounds_of_transform | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, transform, alpha, ci_labels):\n alpha2 = 1 - alpha / 2.0\n z = inv_normal_cdf(alpha2)\n df = pd.DataFrame(index=self.timeline)\n gradient_of_cum_hazard_at_mle = make_jvp_reversemode(transform)(self.\n _fitted_parameters_, self.timeline.astype(float))\n gradient_at_times = np.vstack([gradient_of_cum_hazard_at_mle(basis) for\n basis in np.eye(len(self._fitted_parameters_), dtype=float)])\n std_cumulative_hazard = np.sqrt(np.einsum('nj,jk,nk->n',\n gradient_at_times.T, self.variance_matrix_, gradient_at_times.T))\n if ci_labels is None:\n ci_labels = ['%s_upper_%g' % (self._label, 1 - alpha), \n '%s_lower_%g' % (self._label, 1 - alpha)]\n assert len(ci_labels) == 2, 'ci_labels should be a length 2 array.'\n df[ci_labels[0]] = transform(self._fitted_parameters_, self.timeline\n ) + z * std_cumulative_hazard\n df[ci_labels[1]] = transform(self._fitted_parameters_, self.timeline\n ) - z * std_cumulative_hazard\n return df\n",
"role": "user"
},
{
"content": "compute_confidence_bounds_of_transform",
"role": "assistant"
}
] |
def x(self):
ci = 1 - self.alpha
lower_upper_bounds = self._compute_confidence_bounds_of_parameters()
df = pd.DataFrame(index=self._fitted_parameter_names)
df['coef'] = self._fitted_parameters_
df['se(coef)'] = self._compute_standard_errors().loc['se']
df['lower %g' % ci] = lower_upper_bounds['lower-bound']
df['upper %g' % ci] = lower_upper_bounds['upper-bound']
df['p'] = self._compute_p_values()
with np.errstate(invalid='ignore', divide='ignore'):
df['-log2(p)'] = -np.log2(df['p'])
return df
| summary | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n ci = 1 - self.alpha\n lower_upper_bounds = self._compute_confidence_bounds_of_parameters()\n df = pd.DataFrame(index=self._fitted_parameter_names)\n df['coef'] = self._fitted_parameters_\n df['se(coef)'] = self._compute_standard_errors().loc['se']\n df['lower %g' % ci] = lower_upper_bounds['lower-bound']\n df['upper %g' % ci] = lower_upper_bounds['upper-bound']\n df['p'] = self._compute_p_values()\n with np.errstate(invalid='ignore', divide='ignore'):\n df['-log2(p)'] = -np.log2(df['p'])\n return df\n",
"role": "user"
},
{
"content": "summary",
"role": "assistant"
}
] |
def x(self, decimals=2, **kwargs):
justify = string_justify(18)
print(self)
print('{} = {}'.format(justify('number of subjects'), self.
event_observed.shape[0]))
print('{} = {}'.format(justify('number of events'), np.where(self.
event_observed)[0].shape[0]))
print('{} = {:.{prec}f}'.format(justify('log-likelihood'), self.
_log_likelihood, prec=decimals))
print('{} = {}'.format(justify('hypothesis'), ', '.join('%s != %d' % (
name, iv) for name, iv in zip(self._fitted_parameter_names, self.
_initial_values))))
for k, v in kwargs.items():
print('{} = {}\n'.format(justify(k), v))
print(end='\n')
print('---')
df = self.summary
print(df.to_string(float_format=format_floats(decimals), formatters={
'p': format_p_value(decimals), 'exp(coef)': format_exp_floats(
decimals)}))
| print_summary | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, decimals=2, **kwargs):\n justify = string_justify(18)\n print(self)\n print('{} = {}'.format(justify('number of subjects'), self.\n event_observed.shape[0]))\n print('{} = {}'.format(justify('number of events'), np.where(self.\n event_observed)[0].shape[0]))\n print('{} = {:.{prec}f}'.format(justify('log-likelihood'), self.\n _log_likelihood, prec=decimals))\n print('{} = {}'.format(justify('hypothesis'), ', '.join('%s != %d' % (\n name, iv) for name, iv in zip(self._fitted_parameter_names, self.\n _initial_values))))\n for k, v in kwargs.items():\n print('{} = {}\\n'.format(justify(k), v))\n print(end='\\n')\n print('---')\n df = self.summary\n print(df.to_string(float_format=format_floats(decimals), formatters={\n 'p': format_p_value(decimals), 'exp(coef)': format_exp_floats(\n decimals)}))\n",
"role": "user"
},
{
"content": "print_summary",
"role": "assistant"
}
] |
def x(self, durations, event_observed=None, timeline=None, label=None,
alpha=None, ci_labels=None, show_progress=False, entry=None, weights=
None, left_censorship=False):
if left_censorship:
warnings.warn(
'kwarg left_censorship is deprecated and will be removed in a future release. Please use ``.fit_left_censoring`` instead.'
, DeprecationWarning)
return self.fit_left_censoring(durations, event_observed, timeline,
label, alpha, ci_labels, show_progress, entry, weights)
self.durations = np.asarray(pass_for_numeric_dtypes_or_raise_array(
durations))
check_nans_or_infs(self.durations)
check_positivity(self.durations)
self._censoring_type = CensoringType.RIGHT
return self._fit((self.durations, None), event_observed=event_observed,
timeline=timeline, label=label, alpha=alpha, ci_labels=ci_labels,
show_progress=show_progress, entry=entry, weights=weights)
| fit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, durations, event_observed=None, timeline=None, label=None,\n alpha=None, ci_labels=None, show_progress=False, entry=None, weights=\n None, left_censorship=False):\n if left_censorship:\n warnings.warn(\n 'kwarg left_censorship is deprecated and will be removed in a future release. Please use ``.fit_left_censoring`` instead.'\n , DeprecationWarning)\n return self.fit_left_censoring(durations, event_observed, timeline,\n label, alpha, ci_labels, show_progress, entry, weights)\n self.durations = np.asarray(pass_for_numeric_dtypes_or_raise_array(\n durations))\n check_nans_or_infs(self.durations)\n check_positivity(self.durations)\n self._censoring_type = CensoringType.RIGHT\n return self._fit((self.durations, None), event_observed=event_observed,\n timeline=timeline, label=label, alpha=alpha, ci_labels=ci_labels,\n show_progress=show_progress, entry=entry, weights=weights)\n",
"role": "user"
},
{
"content": "fit",
"role": "assistant"
}
] |
def x(self, durations, event_observed=None, timeline=None, label=None,
alpha=None, ci_labels=None, show_progress=False, entry=None, weights=None):
self.durations = np.asarray(pass_for_numeric_dtypes_or_raise_array(
durations))
check_nans_or_infs(self.durations)
check_positivity(self.durations)
self._censoring_type = CensoringType.LEFT
return self._fit((None, self.durations), event_observed=event_observed,
timeline=timeline, label=label, alpha=alpha, ci_labels=ci_labels,
show_progress=show_progress, entry=entry, weights=weights)
| fit_left_censoring | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, durations, event_observed=None, timeline=None, label=None,\n alpha=None, ci_labels=None, show_progress=False, entry=None, weights=None):\n self.durations = np.asarray(pass_for_numeric_dtypes_or_raise_array(\n durations))\n check_nans_or_infs(self.durations)\n check_positivity(self.durations)\n self._censoring_type = CensoringType.LEFT\n return self._fit((None, self.durations), event_observed=event_observed,\n timeline=timeline, label=label, alpha=alpha, ci_labels=ci_labels,\n show_progress=show_progress, entry=entry, weights=weights)\n",
"role": "user"
},
{
"content": "fit_left_censoring",
"role": "assistant"
}
] |
def x(self, lower_bound, upper_bound, event_observed=None, timeline=None,
label=None, alpha=None, ci_labels=None, show_progress=False, entry=None,
weights=None):
check_nans_or_infs(lower_bound)
check_positivity(upper_bound)
self.upper_bound = np.asarray(pass_for_numeric_dtypes_or_raise_array(
upper_bound))
self.lower_bound = np.asarray(pass_for_numeric_dtypes_or_raise_array(
lower_bound))
if (self.upper_bound < self.lower_bound).any():
raise ValueError(
'All upper_bound times must be greater than or equal to lower_bound times.'
)
if event_observed is None:
event_observed = self.upper_bound == self.lower_bound
if ((self.lower_bound == self.upper_bound) != event_observed).any():
raise ValueError(
'For all rows, lower_bound == upper_bound if and only if event observed = 1 (uncensored). Likewise, lower_bound < upper_bound if and only if event observed = 0 (censored)'
)
self._censoring_type = CensoringType.INTERVAL
return self._fit((np.clip(self.lower_bound, 1e-20, 1e+25), np.clip(self
.upper_bound, 1e-20, 1e+25)), event_observed=event_observed,
timeline=timeline, label=label, alpha=alpha, ci_labels=ci_labels,
show_progress=show_progress, entry=entry, weights=weights)
| fit_interval_censoring | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, lower_bound, upper_bound, event_observed=None, timeline=None,\n label=None, alpha=None, ci_labels=None, show_progress=False, entry=None,\n weights=None):\n check_nans_or_infs(lower_bound)\n check_positivity(upper_bound)\n self.upper_bound = np.asarray(pass_for_numeric_dtypes_or_raise_array(\n upper_bound))\n self.lower_bound = np.asarray(pass_for_numeric_dtypes_or_raise_array(\n lower_bound))\n if (self.upper_bound < self.lower_bound).any():\n raise ValueError(\n 'All upper_bound times must be greater than or equal to lower_bound times.'\n )\n if event_observed is None:\n event_observed = self.upper_bound == self.lower_bound\n if ((self.lower_bound == self.upper_bound) != event_observed).any():\n raise ValueError(\n 'For all rows, lower_bound == upper_bound if and only if event observed = 1 (uncensored). Likewise, lower_bound < upper_bound if and only if event observed = 0 (censored)'\n )\n self._censoring_type = CensoringType.INTERVAL\n return self._fit((np.clip(self.lower_bound, 1e-20, 1e+25), np.clip(self\n .upper_bound, 1e-20, 1e+25)), event_observed=event_observed,\n timeline=timeline, label=label, alpha=alpha, ci_labels=ci_labels,\n show_progress=show_progress, entry=entry, weights=weights)\n",
"role": "user"
},
{
"content": "fit_interval_censoring",
"role": "assistant"
}
] |
def x(self, times, label=None):
label = coalesce(label, self._label)
return pd.Series(self._survival_function(self._fitted_parameters_,
times), index=_to_array(times), name=label)
| survival_function_at_times | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, times, label=None):\n label = coalesce(label, self._label)\n return pd.Series(self._survival_function(self._fitted_parameters_,\n times), index=_to_array(times), name=label)\n",
"role": "user"
},
{
"content": "survival_function_at_times",
"role": "assistant"
}
] |
def x(self, times, label=None):
label = coalesce(label, self._label)
return pd.Series(self._cumulative_density(self._fitted_parameters_,
times), index=_to_array(times), name=label)
| cumulative_density_at_times | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, times, label=None):\n label = coalesce(label, self._label)\n return pd.Series(self._cumulative_density(self._fitted_parameters_,\n times), index=_to_array(times), name=label)\n",
"role": "user"
},
{
"content": "cumulative_density_at_times",
"role": "assistant"
}
] |
def x(self, times, label=None):
label = coalesce(label, self._label)
return pd.Series(self._cumulative_hazard(self._fitted_parameters_,
times), index=_to_array(times), name=label)
| cumulative_hazard_at_times | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, times, label=None):\n label = coalesce(label, self._label)\n return pd.Series(self._cumulative_hazard(self._fitted_parameters_,\n times), index=_to_array(times), name=label)\n",
"role": "user"
},
{
"content": "cumulative_hazard_at_times",
"role": "assistant"
}
] |
def x(self, times, label=None):
label = coalesce(label, self._label)
return pd.Series(self._hazard(self._fitted_parameters_, times), index=
_to_array(times), name=label)
| hazard_at_times | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, times, label=None):\n label = coalesce(label, self._label)\n return pd.Series(self._hazard(self._fitted_parameters_, times), index=\n _to_array(times), name=label)\n",
"role": "user"
},
{
"content": "hazard_at_times",
"role": "assistant"
}
] |
def x(self):
return self._compute_confidence_bounds_of_transform(self._hazard, self.
alpha, self._ci_labels)
| confidence_interval_hazard_ | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return self._compute_confidence_bounds_of_transform(self._hazard, self.\n alpha, self._ci_labels)\n",
"role": "user"
},
{
"content": "confidence_interval_hazard_",
"role": "assistant"
}
] |
def x(self):
return self._compute_confidence_bounds_of_transform(self.
_survival_function, self.alpha, self._ci_labels)
| confidence_interval_survival_function_ | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return self._compute_confidence_bounds_of_transform(self.\n _survival_function, self.alpha, self._ci_labels)\n",
"role": "user"
},
{
"content": "confidence_interval_survival_function_",
"role": "assistant"
}
] |
def x(self):
return self._compute_confidence_bounds_of_transform(self.
_cumulative_density, self.alpha, self._ci_labels)
| confidence_interval_cumulative_density_ | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return self._compute_confidence_bounds_of_transform(self.\n _cumulative_density, self.alpha, self._ci_labels)\n",
"role": "user"
},
{
"content": "confidence_interval_cumulative_density_",
"role": "assistant"
}
] |
def x(self, **kwargs):
set_kwargs_drawstyle(kwargs, 'default')
return _plot_estimate(self, estimate=getattr(self, self._estimate_name),
confidence_intervals=self.confidence_interval_, **kwargs)
| plot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n set_kwargs_drawstyle(kwargs, 'default')\n return _plot_estimate(self, estimate=getattr(self, self._estimate_name),\n confidence_intervals=self.confidence_interval_, **kwargs)\n",
"role": "user"
},
{
"content": "plot",
"role": "assistant"
}
] |
def x(self, df, duration_col, event_col=None, ancillary_df=None,
show_progress=False, timeline=None, weights_col=None, robust=False,
initial_point=None, entry_col=None):
self.duration_col = duration_col
self._time_cols = [duration_col]
self._censoring_type = CensoringType.RIGHT
df = df.copy()
T = pass_for_numeric_dtypes_or_raise_array(df.pop(duration_col)).astype(
float)
self.durations = T.copy()
self._fit(self._log_likelihood_right_censoring, df, (T.values, None),
event_col=event_col, ancillary_df=ancillary_df, show_progress=
show_progress, timeline=timeline, weights_col=weights_col, robust=
robust, initial_point=initial_point, entry_col=entry_col)
return self
| fit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, df, duration_col, event_col=None, ancillary_df=None,\n show_progress=False, timeline=None, weights_col=None, robust=False,\n initial_point=None, entry_col=None):\n self.duration_col = duration_col\n self._time_cols = [duration_col]\n self._censoring_type = CensoringType.RIGHT\n df = df.copy()\n T = pass_for_numeric_dtypes_or_raise_array(df.pop(duration_col)).astype(\n float)\n self.durations = T.copy()\n self._fit(self._log_likelihood_right_censoring, df, (T.values, None),\n event_col=event_col, ancillary_df=ancillary_df, show_progress=\n show_progress, timeline=timeline, weights_col=weights_col, robust=\n robust, initial_point=initial_point, entry_col=entry_col)\n return self\n",
"role": "user"
},
{
"content": "fit",
"role": "assistant"
}
] |
def x(self, df, lower_bound_col, upper_bound_col, event_col=None,
ancillary_df=None, show_progress=False, timeline=None, weights_col=None,
robust=False, initial_point=None, entry_col=None):
self.lower_bound_col = lower_bound_col
self.upper_bound_col = upper_bound_col
self._time_cols = [lower_bound_col, upper_bound_col]
self._censoring_type = CensoringType.INTERVAL
df = df.copy()
lower_bound = pass_for_numeric_dtypes_or_raise_array(df.pop(
lower_bound_col)).astype(float)
upper_bound = pass_for_numeric_dtypes_or_raise_array(df.pop(
upper_bound_col)).astype(float)
if event_col is None:
event_col = 'E'
df['E'] = lower_bound == upper_bound
if ((lower_bound == upper_bound) != df[event_col]).any():
raise ValueError(
'For all rows, lower_bound == upper_bound if and only if event observed = 1 (uncensored). Likewise, lower_bound < upper_bound if and only if event observed = 0 (censored)'
)
if (lower_bound > upper_bound).any():
raise ValueError(
'All upper bound measurements must be greater than or equal to lower bound measurements.'
)
self.lower_bound = lower_bound
self.upper_bound = upper_bound
self._fit(self._log_likelihood_interval_censoring, df, (lower_bound.
values, np.clip(upper_bound.values, 0, 1e+25)), event_col=event_col,
ancillary_df=ancillary_df, show_progress=show_progress, timeline=
timeline, weights_col=weights_col, robust=robust, initial_point=
initial_point, entry_col=entry_col)
return self
| fit_interval_censoring | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, df, lower_bound_col, upper_bound_col, event_col=None,\n ancillary_df=None, show_progress=False, timeline=None, weights_col=None,\n robust=False, initial_point=None, entry_col=None):\n self.lower_bound_col = lower_bound_col\n self.upper_bound_col = upper_bound_col\n self._time_cols = [lower_bound_col, upper_bound_col]\n self._censoring_type = CensoringType.INTERVAL\n df = df.copy()\n lower_bound = pass_for_numeric_dtypes_or_raise_array(df.pop(\n lower_bound_col)).astype(float)\n upper_bound = pass_for_numeric_dtypes_or_raise_array(df.pop(\n upper_bound_col)).astype(float)\n if event_col is None:\n event_col = 'E'\n df['E'] = lower_bound == upper_bound\n if ((lower_bound == upper_bound) != df[event_col]).any():\n raise ValueError(\n 'For all rows, lower_bound == upper_bound if and only if event observed = 1 (uncensored). Likewise, lower_bound < upper_bound if and only if event observed = 0 (censored)'\n )\n if (lower_bound > upper_bound).any():\n raise ValueError(\n 'All upper bound measurements must be greater than or equal to lower bound measurements.'\n )\n self.lower_bound = lower_bound\n self.upper_bound = upper_bound\n self._fit(self._log_likelihood_interval_censoring, df, (lower_bound.\n values, np.clip(upper_bound.values, 0, 1e+25)), event_col=event_col,\n ancillary_df=ancillary_df, show_progress=show_progress, timeline=\n timeline, weights_col=weights_col, robust=robust, initial_point=\n initial_point, entry_col=entry_col)\n return self\n",
"role": "user"
},
{
"content": "fit_interval_censoring",
"role": "assistant"
}
] |
def x(self, df, duration_col=None, event_col=None, ancillary_df=None,
show_progress=False, timeline=None, weights_col=None, robust=False,
initial_point=None, entry_col=None):
self._censoring_type = CensoringType.LEFT
df = df.copy()
T = pass_for_numeric_dtypes_or_raise_array(df.pop(duration_col)).astype(
float)
self.durations = T.copy()
self._fit(self._log_likelihood_left_censoring, df, (None, T.values),
event_col=event_col, ancillary_df=ancillary_df, show_progress=
show_progress, timeline=timeline, weights_col=weights_col, robust=
robust, initial_point=initial_point, entry_col=entry_col)
return self
| fit_left_censoring | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, df, duration_col=None, event_col=None, ancillary_df=None,\n show_progress=False, timeline=None, weights_col=None, robust=False,\n initial_point=None, entry_col=None):\n self._censoring_type = CensoringType.LEFT\n df = df.copy()\n T = pass_for_numeric_dtypes_or_raise_array(df.pop(duration_col)).astype(\n float)\n self.durations = T.copy()\n self._fit(self._log_likelihood_left_censoring, df, (None, T.values),\n event_col=event_col, ancillary_df=ancillary_df, show_progress=\n show_progress, timeline=timeline, weights_col=weights_col, robust=\n robust, initial_point=initial_point, entry_col=entry_col)\n return self\n",
"role": "user"
},
{
"content": "fit_left_censoring",
"role": "assistant"
}
] |
def x(self, Ts, E, entries, weights, Xs):
import lifelines
def x(model, i):
param = model._fitted_parameters_[i]
if param <= 0:
return param
return np.log(param)
name = self.__class__.__name__.replace('AFT', '')
uni_model = getattr(lifelines, name)()
if self._censoring_type == CensoringType.RIGHT:
uni_model.fit_right_censoring(Ts[0], event_observed=E, entry=
entries, weights=weights)
elif self._censoring_type == CensoringType.INTERVAL:
uni_model.fit_interval_censoring(Ts[0], Ts[1], event_observed=E,
entry=entries, weights=weights)
elif self._censoring_type == CensoringType.LEFT:
uni_model.fit_left_censoring(Ts[1], event_observed=E, entry=entries,
weights=weights)
self._ll_null_ = uni_model._log_likelihood
return np.concatenate([([0] * (_X.shape[1] - 1) + [transform_ith_param(
uni_model, i)]) for i, _X in enumerate(Xs)])
| create_initial_point | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, Ts, E, entries, weights, Xs):\n import lifelines\n\n def x(model, i):\n param = model._fitted_parameters_[i]\n if param <= 0:\n return param\n return np.log(param)\n name = self.__class__.__name__.replace('AFT', '')\n uni_model = getattr(lifelines, name)()\n if self._censoring_type == CensoringType.RIGHT:\n uni_model.fit_right_censoring(Ts[0], event_observed=E, entry=\n entries, weights=weights)\n elif self._censoring_type == CensoringType.INTERVAL:\n uni_model.fit_interval_censoring(Ts[0], Ts[1], event_observed=E,\n entry=entries, weights=weights)\n elif self._censoring_type == CensoringType.LEFT:\n uni_model.fit_left_censoring(Ts[1], event_observed=E, entry=entries,\n weights=weights)\n self._ll_null_ = uni_model._log_likelihood\n return np.concatenate([([0] * (_X.shape[1] - 1) + [transform_ith_param(\n uni_model, i)]) for i, _X in enumerate(Xs)])\n",
"role": "user"
},
{
"content": "create_initial_point",
"role": "assistant"
}
] |
def x(self, X, ancillary_X=None):
return self.predict_percentile(X, p=0.5, ancillary_X=ancillary_X)
| predict_median | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, ancillary_X=None):\n return self.predict_percentile(X, p=0.5, ancillary_X=ancillary_X)\n",
"role": "user"
},
{
"content": "predict_median",
"role": "assistant"
}
] |
def x(self):
if hasattr(self, '_predicted_median'):
self._concordance_score_ = concordance_index(self.durations, self.
_predicted_median, self.event_observed)
del self._predicted_median
return self._concordance_score_
return self._concordance_score_
| score_ | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if hasattr(self, '_predicted_median'):\n self._concordance_score_ = concordance_index(self.durations, self.\n _predicted_median, self.event_observed)\n del self._predicted_median\n return self._concordance_score_\n return self._concordance_score_\n",
"role": "user"
},
{
"content": "score_",
"role": "assistant"
}
] |
def x(self, columns=None, parameter=None, **errorbar_kwargs):
from matplotlib import pyplot as plt
set_kwargs_ax(errorbar_kwargs)
ax = errorbar_kwargs.pop('ax')
errorbar_kwargs.setdefault('c', 'k')
errorbar_kwargs.setdefault('fmt', 's')
errorbar_kwargs.setdefault('markerfacecolor', 'white')
errorbar_kwargs.setdefault('markeredgewidth', 1.25)
errorbar_kwargs.setdefault('elinewidth', 1.25)
errorbar_kwargs.setdefault('capsize', 3)
z = inv_normal_cdf(1 - self.alpha / 2)
params_ = self.params_.copy()
standard_errors_ = self.standard_errors_.copy()
if columns is not None:
params_ = params_.loc[:, columns]
standard_errors_ = standard_errors_.loc[:, columns]
if parameter is not None:
params_ = params_.loc[parameter]
standard_errors_ = standard_errors_.loc[parameter]
columns = params_.index
hazards = params_.loc[columns].to_frame(name='coefs')
hazards['se'] = z * standard_errors_.loc[columns]
if isinstance(hazards.index, pd.MultiIndex):
hazards = hazards.groupby(level=0, group_keys=False).apply(lambda x:
x.sort_values(by='coefs', ascending=True))
else:
hazards = hazards.sort_values(by='coefs', ascending=True)
yaxis_locations = list(range(len(columns)))
ax.errorbar(hazards['coefs'], yaxis_locations, xerr=hazards['se'], **
errorbar_kwargs)
best_ylim = ax.get_ylim()
ax.vlines(0, -2, len(columns) + 1, linestyles='dashed', linewidths=1,
alpha=0.65)
ax.set_ylim(best_ylim)
if isinstance(columns[0], tuple):
tick_labels = [('%s: %s' % (c, p)) for p, c in hazards.index]
else:
tick_labels = [i for i in hazards.index]
plt.yticks(yaxis_locations, tick_labels)
plt.xlabel('log(accelerated failure rate) (%g%% CI)' % ((1 - self.alpha
) * 100))
return ax
| plot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, columns=None, parameter=None, **errorbar_kwargs):\n from matplotlib import pyplot as plt\n set_kwargs_ax(errorbar_kwargs)\n ax = errorbar_kwargs.pop('ax')\n errorbar_kwargs.setdefault('c', 'k')\n errorbar_kwargs.setdefault('fmt', 's')\n errorbar_kwargs.setdefault('markerfacecolor', 'white')\n errorbar_kwargs.setdefault('markeredgewidth', 1.25)\n errorbar_kwargs.setdefault('elinewidth', 1.25)\n errorbar_kwargs.setdefault('capsize', 3)\n z = inv_normal_cdf(1 - self.alpha / 2)\n params_ = self.params_.copy()\n standard_errors_ = self.standard_errors_.copy()\n if columns is not None:\n params_ = params_.loc[:, columns]\n standard_errors_ = standard_errors_.loc[:, columns]\n if parameter is not None:\n params_ = params_.loc[parameter]\n standard_errors_ = standard_errors_.loc[parameter]\n columns = params_.index\n hazards = params_.loc[columns].to_frame(name='coefs')\n hazards['se'] = z * standard_errors_.loc[columns]\n if isinstance(hazards.index, pd.MultiIndex):\n hazards = hazards.groupby(level=0, group_keys=False).apply(lambda x:\n x.sort_values(by='coefs', ascending=True))\n else:\n hazards = hazards.sort_values(by='coefs', ascending=True)\n yaxis_locations = list(range(len(columns)))\n ax.errorbar(hazards['coefs'], yaxis_locations, xerr=hazards['se'], **\n errorbar_kwargs)\n best_ylim = ax.get_ylim()\n ax.vlines(0, -2, len(columns) + 1, linestyles='dashed', linewidths=1,\n alpha=0.65)\n ax.set_ylim(best_ylim)\n if isinstance(columns[0], tuple):\n tick_labels = [('%s: %s' % (c, p)) for p, c in hazards.index]\n else:\n tick_labels = [i for i in hazards.index]\n plt.yticks(yaxis_locations, tick_labels)\n plt.xlabel('log(accelerated failure rate) (%g%% CI)' % ((1 - self.alpha\n ) * 100))\n return ax\n",
"role": "user"
},
{
"content": "plot",
"role": "assistant"
}
] |
def x(self, covariates, values, plot_baseline=True, **kwargs):
from matplotlib import pyplot as plt
covariates = _to_list(covariates)
values = _to_array(values)
if len(values.shape) == 1:
values = values[None, :].T
if len(covariates) != values.shape[1]:
raise ValueError(
'The number of covariates must equal to second dimension of the values array.'
)
original_columns = self.params_.index.get_level_values(1)
for covariate in covariates:
if covariate not in original_columns:
raise KeyError(
'covariate `%s` is not present in the original dataset' %
covariate)
ax = kwargs.pop('ax', None) or plt.figure().add_subplot(111)
x_bar = self._norm_mean.to_frame().T
X = pd.concat([x_bar] * values.shape[0])
if np.array_equal(np.eye(len(covariates)), values):
X.index = [('%s=1' % c) for c in covariates]
else:
X.index = [', '.join('%s=%g' % (c, v) for c, v in zip(covariates,
row)) for row in values]
for covariate, value in zip(covariates, values.T):
X[covariate] = value
x_bar_anc = self._norm_mean_ancillary.to_frame().T
ancillary_X = pd.concat([x_bar_anc] * values.shape[0])
for covariate, value in zip(covariates, values.T):
ancillary_X[covariate] = value
if self.fit_intercept:
X['_intercept'] = 1.0
ancillary_X['_intercept'] = 1.0
self.predict_survival_function(X, ancillary_X=ancillary_X).plot(ax=ax,
**kwargs)
if plot_baseline:
self.predict_survival_function(x_bar, ancillary_X=x_bar_anc).rename(
columns={(0): 'baseline survival'}).plot(ax=ax, ls=':', color='k')
return ax
| plot_covariate_groups | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, covariates, values, plot_baseline=True, **kwargs):\n from matplotlib import pyplot as plt\n covariates = _to_list(covariates)\n values = _to_array(values)\n if len(values.shape) == 1:\n values = values[None, :].T\n if len(covariates) != values.shape[1]:\n raise ValueError(\n 'The number of covariates must equal to second dimension of the values array.'\n )\n original_columns = self.params_.index.get_level_values(1)\n for covariate in covariates:\n if covariate not in original_columns:\n raise KeyError(\n 'covariate `%s` is not present in the original dataset' %\n covariate)\n ax = kwargs.pop('ax', None) or plt.figure().add_subplot(111)\n x_bar = self._norm_mean.to_frame().T\n X = pd.concat([x_bar] * values.shape[0])\n if np.array_equal(np.eye(len(covariates)), values):\n X.index = [('%s=1' % c) for c in covariates]\n else:\n X.index = [', '.join('%s=%g' % (c, v) for c, v in zip(covariates,\n row)) for row in values]\n for covariate, value in zip(covariates, values.T):\n X[covariate] = value\n x_bar_anc = self._norm_mean_ancillary.to_frame().T\n ancillary_X = pd.concat([x_bar_anc] * values.shape[0])\n for covariate, value in zip(covariates, values.T):\n ancillary_X[covariate] = value\n if self.fit_intercept:\n X['_intercept'] = 1.0\n ancillary_X['_intercept'] = 1.0\n self.predict_survival_function(X, ancillary_X=ancillary_X).plot(ax=ax,\n **kwargs)\n if plot_baseline:\n self.predict_survival_function(x_bar, ancillary_X=x_bar_anc).rename(\n columns={(0): 'baseline survival'}).plot(ax=ax, ls=':', color='k')\n return ax\n",
"role": "user"
},
{
"content": "plot_covariate_groups",
"role": "assistant"
}
] |
def x(self, X, ancillary_X=None, p=0.5):
alpha_, beta_ = self._prep_inputs_for_prediction_and_return_scores(X,
ancillary_X)
return pd.DataFrame(alpha_ * (1 / p - 1) ** beta_, index=_get_index(X))
| predict_percentile | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, ancillary_X=None, p=0.5):\n alpha_, beta_ = self._prep_inputs_for_prediction_and_return_scores(X,\n ancillary_X)\n return pd.DataFrame(alpha_ * (1 / p - 1) ** beta_, index=_get_index(X))\n",
"role": "user"
},
{
"content": "predict_percentile",
"role": "assistant"
}
] |
def x(self, X, ancillary_X=None):
alpha_, beta_ = self._prep_inputs_for_prediction_and_return_scores(X,
ancillary_X)
v = alpha_ * np.pi / beta_ / np.sin(np.pi / beta_)
v = np.where(beta_ > 1, v, np.nan)
return pd.DataFrame(v, index=_get_index(X))
| predict_expectation | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, ancillary_X=None):\n alpha_, beta_ = self._prep_inputs_for_prediction_and_return_scores(X,\n ancillary_X)\n v = alpha_ * np.pi / beta_ / np.sin(np.pi / beta_)\n v = np.where(beta_ > 1, v, np.nan)\n return pd.DataFrame(v, index=_get_index(X))\n",
"role": "user"
},
{
"content": "predict_expectation",
"role": "assistant"
}
] |
def x(self, X, times=None, ancillary_X=None):
times = coalesce(times, self.timeline, np.unique(self.durations))
alpha_, beta_ = self._prep_inputs_for_prediction_and_return_scores(X,
ancillary_X)
return pd.DataFrame(np.log1p(np.outer(times, 1 / alpha_) ** beta_),
columns=_get_index(X), index=times)
| predict_cumulative_hazard | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, X, times=None, ancillary_X=None):\n times = coalesce(times, self.timeline, np.unique(self.durations))\n alpha_, beta_ = self._prep_inputs_for_prediction_and_return_scores(X,\n ancillary_X)\n return pd.DataFrame(np.log1p(np.outer(times, 1 / alpha_) ** beta_),\n columns=_get_index(X), index=times)\n",
"role": "user"
},
{
"content": "predict_cumulative_hazard",
"role": "assistant"
}
] |
def x(self, durations, event_observed=None, timeline=None, entry=None,
label='NA_estimate', alpha=None, ci_labels=None, weights=None):
self._censoring_type = CensoringType.RIGHT
check_nans_or_infs(durations)
if event_observed is not None:
check_nans_or_infs(event_observed)
if weights is not None:
if (weights.astype(int) != weights).any():
warnings.warn(
"""It looks like your weights are not integers, possibly prospenity scores then?
It's important to know that the naive variance estimates of the coefficients are biased. Instead use Monte Carlo to
estimate the variances. See paper "Variance estimation when using inverse probability of treatment weighting (IPTW) with survival analysis"
or "Adjusted Kaplan-Meier estimator and log-rank test with inverse probability of treatment weighting for survival data."
"""
, StatisticalWarning)
v = _preprocess_inputs(durations, event_observed, timeline, entry, weights)
(self.durations, self.event_observed, self.timeline, self.entry, self.
event_table) = v
cumulative_hazard_, cumulative_sq_ = _additive_estimate(self.
event_table, self.timeline, self._additive_f, self._variance_f, False)
self._label = label
self.cumulative_hazard_ = pd.DataFrame(cumulative_hazard_, columns=[
self._label])
self.confidence_interval_ = self._bounds(cumulative_sq_[:, None], alpha if
alpha else self.alpha, ci_labels)
self._cumulative_sq = cumulative_sq_
self._estimation_method = 'cumulative_hazard_'
self._estimate_name = 'cumulative_hazard_'
self._predict_label = label
self._update_docstrings()
self.plot_cumulative_hazard = self.plot
return self
| fit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, durations, event_observed=None, timeline=None, entry=None,\n label='NA_estimate', alpha=None, ci_labels=None, weights=None):\n self._censoring_type = CensoringType.RIGHT\n check_nans_or_infs(durations)\n if event_observed is not None:\n check_nans_or_infs(event_observed)\n if weights is not None:\n if (weights.astype(int) != weights).any():\n warnings.warn(\n \"\"\"It looks like your weights are not integers, possibly prospenity scores then?\n It's important to know that the naive variance estimates of the coefficients are biased. Instead use Monte Carlo to\n estimate the variances. See paper \"Variance estimation when using inverse probability of treatment weighting (IPTW) with survival analysis\"\n or \"Adjusted Kaplan-Meier estimator and log-rank test with inverse probability of treatment weighting for survival data.\"\n \"\"\"\n , StatisticalWarning)\n v = _preprocess_inputs(durations, event_observed, timeline, entry, weights)\n (self.durations, self.event_observed, self.timeline, self.entry, self.\n event_table) = v\n cumulative_hazard_, cumulative_sq_ = _additive_estimate(self.\n event_table, self.timeline, self._additive_f, self._variance_f, False)\n self._label = label\n self.cumulative_hazard_ = pd.DataFrame(cumulative_hazard_, columns=[\n self._label])\n self.confidence_interval_ = self._bounds(cumulative_sq_[:, None], alpha if\n alpha else self.alpha, ci_labels)\n self._cumulative_sq = cumulative_sq_\n self._estimation_method = 'cumulative_hazard_'\n self._estimate_name = 'cumulative_hazard_'\n self._predict_label = label\n self._update_docstrings()\n self.plot_cumulative_hazard = self.plot\n return self\n",
"role": "user"
},
{
"content": "fit",
"role": "assistant"
}
] |
def x(self, bandwidth):
timeline = self.timeline
cumulative_hazard_name = self.cumulative_hazard_.columns[0]
hazard_name = 'differenced-' + cumulative_hazard_name
hazard_ = self.cumulative_hazard_.diff().fillna(self.cumulative_hazard_
.iloc[0])
C = (hazard_[cumulative_hazard_name] != 0.0).values
return pd.DataFrame(1.0 / bandwidth * np.dot(epanechnikov_kernel(
timeline[:, None], timeline[C][None, :], bandwidth), hazard_.values
[C, :]), columns=[hazard_name], index=timeline)
| smoothed_hazard_ | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, bandwidth):\n timeline = self.timeline\n cumulative_hazard_name = self.cumulative_hazard_.columns[0]\n hazard_name = 'differenced-' + cumulative_hazard_name\n hazard_ = self.cumulative_hazard_.diff().fillna(self.cumulative_hazard_\n .iloc[0])\n C = (hazard_[cumulative_hazard_name] != 0.0).values\n return pd.DataFrame(1.0 / bandwidth * np.dot(epanechnikov_kernel(\n timeline[:, None], timeline[C][None, :], bandwidth), hazard_.values\n [C, :]), columns=[hazard_name], index=timeline)\n",
"role": "user"
},
{
"content": "smoothed_hazard_",
"role": "assistant"
}
] |
def x(self, bandwidth, hazard_=None):
if hazard_ is None:
hazard_ = self.smoothed_hazard_(bandwidth).values[:, 0]
timeline = self.timeline
z = inv_normal_cdf(1 - self.alpha / 2)
self._cumulative_sq.iloc[0] = 0
var_hazard_ = self._cumulative_sq.diff().fillna(self._cumulative_sq.iloc[0]
)
C = var_hazard_.values != 0.0
std_hazard_ = np.sqrt(1.0 / bandwidth ** 2 * np.dot(epanechnikov_kernel
(timeline[:, None], timeline[C][None, :], bandwidth) ** 2,
var_hazard_.values[C]))
values = {self.ci_labels[0]: hazard_ * np.exp(z * std_hazard_ / hazard_
), self.ci_labels[1]: hazard_ * np.exp(-z * std_hazard_ / hazard_)}
return pd.DataFrame(values, index=timeline)
| smoothed_hazard_confidence_intervals_ | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, bandwidth, hazard_=None):\n if hazard_ is None:\n hazard_ = self.smoothed_hazard_(bandwidth).values[:, 0]\n timeline = self.timeline\n z = inv_normal_cdf(1 - self.alpha / 2)\n self._cumulative_sq.iloc[0] = 0\n var_hazard_ = self._cumulative_sq.diff().fillna(self._cumulative_sq.iloc[0]\n )\n C = var_hazard_.values != 0.0\n std_hazard_ = np.sqrt(1.0 / bandwidth ** 2 * np.dot(epanechnikov_kernel\n (timeline[:, None], timeline[C][None, :], bandwidth) ** 2,\n var_hazard_.values[C]))\n values = {self.ci_labels[0]: hazard_ * np.exp(z * std_hazard_ / hazard_\n ), self.ci_labels[1]: hazard_ * np.exp(-z * std_hazard_ / hazard_)}\n return pd.DataFrame(values, index=timeline)\n",
"role": "user"
},
{
"content": "smoothed_hazard_confidence_intervals_",
"role": "assistant"
}
] |
def x(values):
if len(values) == 1:
return values
tree = np.empty_like(values)
last_full_row = int(np.log2(len(values) + 1) - 1)
len_ragged_row = len(values) - (2 ** (last_full_row + 1) - 1)
if len_ragged_row > 0:
bottom_row_ix = np.s_[:2 * len_ragged_row:2]
tree[-len_ragged_row:] = values[bottom_row_ix]
values = np.delete(values, bottom_row_ix)
values_start = 0
values_space = 2
values_len = 2 ** last_full_row
while values_start < len(values):
tree[values_len - 1:2 * values_len - 1] = values[values_start::
values_space]
values_start += int(values_space / 2)
values_space *= 2
values_len = int(values_len / 2)
return tree
| treeify | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(values):\n if len(values) == 1:\n return values\n tree = np.empty_like(values)\n last_full_row = int(np.log2(len(values) + 1) - 1)\n len_ragged_row = len(values) - (2 ** (last_full_row + 1) - 1)\n if len_ragged_row > 0:\n bottom_row_ix = np.s_[:2 * len_ragged_row:2]\n tree[-len_ragged_row:] = values[bottom_row_ix]\n values = np.delete(values, bottom_row_ix)\n values_start = 0\n values_space = 2\n values_len = 2 ** last_full_row\n while values_start < len(values):\n tree[values_len - 1:2 * values_len - 1] = values[values_start::\n values_space]\n values_start += int(values_space / 2)\n values_space *= 2\n values_len = int(values_len / 2)\n return tree\n",
"role": "user"
},
{
"content": "treeify",
"role": "assistant"
}
] |
def x(self, value):
i = 0
n = len(self._tree)
while i < n:
cur = self._tree[i]
self._counts[i] += 1
if value < cur:
i = 2 * i + 1
elif value > cur:
i = 2 * i + 2
else:
return
raise ValueError(
'Value %s not contained in tree.Also, the counts are now messed up.' %
value)
| insert | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, value):\n i = 0\n n = len(self._tree)\n while i < n:\n cur = self._tree[i]\n self._counts[i] += 1\n if value < cur:\n i = 2 * i + 1\n elif value > cur:\n i = 2 * i + 2\n else:\n return\n raise ValueError(\n 'Value %s not contained in tree.Also, the counts are now messed up.' %\n value)\n",
"role": "user"
},
{
"content": "insert",
"role": "assistant"
}
] |
def x(self, value):
i = 0
n = len(self._tree)
rank = 0
count = 0
while i < n:
cur = self._tree[i]
if value < cur:
i = 2 * i + 1
continue
elif value > cur:
rank += self._counts[i]
nexti = 2 * i + 2
if nexti < n:
rank -= self._counts[nexti]
i = nexti
continue
else:
return rank, count
else:
count = self._counts[i]
lefti = 2 * i + 1
if lefti < n:
nleft = self._counts[lefti]
count -= nleft
rank += nleft
righti = lefti + 1
if righti < n:
count -= self._counts[righti]
return rank, count
return rank, count
| rank | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, value):\n i = 0\n n = len(self._tree)\n rank = 0\n count = 0\n while i < n:\n cur = self._tree[i]\n if value < cur:\n i = 2 * i + 1\n continue\n elif value > cur:\n rank += self._counts[i]\n nexti = 2 * i + 2\n if nexti < n:\n rank -= self._counts[nexti]\n i = nexti\n continue\n else:\n return rank, count\n else:\n count = self._counts[i]\n lefti = 2 * i + 1\n if lefti < n:\n nleft = self._counts[lefti]\n count -= nleft\n rank += nleft\n righti = lefti + 1\n if righti < n:\n count -= self._counts[righti]\n return rank, count\n return rank, count\n",
"role": "user"
},
{
"content": "rank",
"role": "assistant"
}
] |
def x(model, **plot_kwargs):
from lifelines.utils import qth_survival_times
from lifelines import KaplanMeierFitter
set_kwargs_ax(plot_kwargs)
ax = plot_kwargs.pop('ax')
dist = get_distribution_name_of_lifelines_model(model)
dist_object = create_scipy_stats_model_from_lifelines_model(model)
COL_EMP = 'empirical quantiles'
COL_THEO = 'fitted %s quantiles' % dist
if model._censoring_type == CensoringType.LEFT:
kmf = KaplanMeierFitter().fit_left_censoring(model.durations, model
.event_observed, label=COL_EMP)
elif model._censoring_type == CensoringType.RIGHT:
kmf = KaplanMeierFitter().fit_right_censoring(model.durations,
model.event_observed, label=COL_EMP)
elif model._censoring_type == CensoringType.INTERVAL:
raise NotImplementedError()
q = np.unique(kmf.cumulative_density_.values[:, 0])
quantiles = qth_survival_times(q, kmf.cumulative_density_, cdf=True)
quantiles[COL_THEO] = dist_object.ppf(q)
quantiles = quantiles.replace([-np.inf, 0, np.inf], np.nan).dropna()
max_, min_ = quantiles[COL_EMP].max(), quantiles[COL_EMP].min()
quantiles.plot.scatter(COL_THEO, COL_EMP, c='none', edgecolor='k', lw=
0.5, ax=ax)
ax.plot([min_, max_], [min_, max_], c='k', ls=':', lw=1.0)
ax.set_ylim(min_, max_)
ax.set_xlim(min_, max_)
return ax
| qq_plot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model, **plot_kwargs):\n from lifelines.utils import qth_survival_times\n from lifelines import KaplanMeierFitter\n set_kwargs_ax(plot_kwargs)\n ax = plot_kwargs.pop('ax')\n dist = get_distribution_name_of_lifelines_model(model)\n dist_object = create_scipy_stats_model_from_lifelines_model(model)\n COL_EMP = 'empirical quantiles'\n COL_THEO = 'fitted %s quantiles' % dist\n if model._censoring_type == CensoringType.LEFT:\n kmf = KaplanMeierFitter().fit_left_censoring(model.durations, model\n .event_observed, label=COL_EMP)\n elif model._censoring_type == CensoringType.RIGHT:\n kmf = KaplanMeierFitter().fit_right_censoring(model.durations,\n model.event_observed, label=COL_EMP)\n elif model._censoring_type == CensoringType.INTERVAL:\n raise NotImplementedError()\n q = np.unique(kmf.cumulative_density_.values[:, 0])\n quantiles = qth_survival_times(q, kmf.cumulative_density_, cdf=True)\n quantiles[COL_THEO] = dist_object.ppf(q)\n quantiles = quantiles.replace([-np.inf, 0, np.inf], np.nan).dropna()\n max_, min_ = quantiles[COL_EMP].max(), quantiles[COL_EMP].min()\n quantiles.plot.scatter(COL_THEO, COL_EMP, c='none', edgecolor='k', lw=\n 0.5, ax=ax)\n ax.plot([min_, max_], [min_, max_], c='k', ls=':', lw=1.0)\n ax.set_ylim(min_, max_)\n ax.set_xlim(min_, max_)\n return ax\n",
"role": "user"
},
{
"content": "qq_plot",
"role": "assistant"
}
] |
def x(ax, sides):
for side in sides:
ax.spines[side].set_visible(False)
return ax
| remove_spines | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ax, sides):\n for side in sides:\n ax.spines[side].set_visible(False)\n return ax\n",
"role": "user"
},
{
"content": "remove_spines",
"role": "assistant"
}
] |
def x(ax, sides, dists):
for side, dist in zip(sides, dists):
ax.spines[side].set_position(('axes', dist))
return ax
| move_spines | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ax, sides, dists):\n for side, dist in zip(sides, dists):\n ax.spines[side].set_position(('axes', dist))\n return ax\n",
"role": "user"
},
{
"content": "move_spines",
"role": "assistant"
}
] |
def x(ax, x=False, y=False):
if x:
ax.xaxis.set_ticks_position('none')
if y:
ax.yaxis.set_ticks_position('none')
return ax
| remove_ticks | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ax, x=False, y=False):\n if x:\n ax.xaxis.set_ticks_position('none')\n if y:\n ax.yaxis.set_ticks_position('none')\n return ax\n",
"role": "user"
},
{
"content": "remove_ticks",
"role": "assistant"
}
] |
def x(*fitters, **kwargs):
from matplotlib import pyplot as plt
ax = kwargs.get('ax', None)
if ax is None:
ax = plt.gca()
fig = kwargs.get('fig', None)
if fig is None:
fig = plt.gcf()
if 'labels' not in kwargs:
labels = [f._label for f in fitters]
else:
labels = kwargs['labels']
if labels is None:
labels = [None] * len(fitters)
ax2 = plt.twiny(ax=ax)
ax2_ypos = -0.15 * 6.0 / fig.get_figheight()
move_spines(ax2, ['bottom'], [ax2_ypos])
remove_spines(ax2, ['top', 'right', 'bottom', 'left'])
ax2.xaxis.tick_bottom()
ax2.set_xlim(ax.get_xlim())
ax2.set_xticks(ax.get_xticks())
remove_ticks(ax2, x=True, y=True)
ticklabels = []
for tick in ax2.get_xticks():
lbl = ''
for f, l in zip(fitters, labels):
if tick == ax2.get_xticks()[0] and l is not None:
if is_latex_enabled():
s = '\n{}\\quad'.format(l) + '{}'
else:
s = '\n{} '.format(l) + '{}'
else:
s = '\n{}'
lbl += s.format(f.durations[f.durations >= tick].shape[0])
ticklabels.append(lbl.strip())
ax2.set_xticklabels(ticklabels, ha='right')
ax2.xaxis.set_label_coords(0, ax2_ypos)
ax2.set_xlabel('At risk')
plt.tight_layout()
return ax
| add_at_risk_counts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(*fitters, **kwargs):\n from matplotlib import pyplot as plt\n ax = kwargs.get('ax', None)\n if ax is None:\n ax = plt.gca()\n fig = kwargs.get('fig', None)\n if fig is None:\n fig = plt.gcf()\n if 'labels' not in kwargs:\n labels = [f._label for f in fitters]\n else:\n labels = kwargs['labels']\n if labels is None:\n labels = [None] * len(fitters)\n ax2 = plt.twiny(ax=ax)\n ax2_ypos = -0.15 * 6.0 / fig.get_figheight()\n move_spines(ax2, ['bottom'], [ax2_ypos])\n remove_spines(ax2, ['top', 'right', 'bottom', 'left'])\n ax2.xaxis.tick_bottom()\n ax2.set_xlim(ax.get_xlim())\n ax2.set_xticks(ax.get_xticks())\n remove_ticks(ax2, x=True, y=True)\n ticklabels = []\n for tick in ax2.get_xticks():\n lbl = ''\n for f, l in zip(fitters, labels):\n if tick == ax2.get_xticks()[0] and l is not None:\n if is_latex_enabled():\n s = '\\n{}\\\\quad'.format(l) + '{}'\n else:\n s = '\\n{} '.format(l) + '{}'\n else:\n s = '\\n{}'\n lbl += s.format(f.durations[f.durations >= tick].shape[0])\n ticklabels.append(lbl.strip())\n ax2.set_xticklabels(ticklabels, ha='right')\n ax2.xaxis.set_label_coords(0, ax2_ypos)\n ax2.set_xlabel('At risk')\n plt.tight_layout()\n return ax\n",
"role": "user"
},
{
"content": "add_at_risk_counts",
"role": "assistant"
}
] |
def x(durations, event_observed=None, entry=None, left_truncated=False,
sort_by_duration=True, event_observed_color='#A60628',
event_censored_color='#348ABD', **kwargs):
set_kwargs_ax(kwargs)
ax = kwargs.pop('ax')
N = durations.shape[0]
if N > 80:
warnings.warn(
'For less visual clutter, you may want to subsample to less than 80 individuals.'
)
if event_observed is None:
event_observed = np.ones(N, dtype=bool)
if entry is None:
entry = np.zeros(N)
assert durations.shape[0] == N
assert event_observed.shape[0] == N
if sort_by_duration:
ix = np.argsort(entry + durations, 0)
durations = durations[ix]
event_observed = event_observed[ix]
entry = entry[ix]
for i in range(N):
c = event_observed_color if event_observed[i] else event_censored_color
ax.hlines(i, entry[i], entry[i] + durations[i], color=c, lw=1.5)
if left_truncated:
ax.hlines(i, 0, entry[i], color=c, lw=1.0, linestyle='--')
m = '' if not event_observed[i] else 'o'
ax.scatter(entry[i] + durations[i], i, color=c, marker=m, s=10)
ax.set_ylim(-0.5, N)
return ax
| plot_lifetimes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(durations, event_observed=None, entry=None, left_truncated=False,\n sort_by_duration=True, event_observed_color='#A60628',\n event_censored_color='#348ABD', **kwargs):\n set_kwargs_ax(kwargs)\n ax = kwargs.pop('ax')\n N = durations.shape[0]\n if N > 80:\n warnings.warn(\n 'For less visual clutter, you may want to subsample to less than 80 individuals.'\n )\n if event_observed is None:\n event_observed = np.ones(N, dtype=bool)\n if entry is None:\n entry = np.zeros(N)\n assert durations.shape[0] == N\n assert event_observed.shape[0] == N\n if sort_by_duration:\n ix = np.argsort(entry + durations, 0)\n durations = durations[ix]\n event_observed = event_observed[ix]\n entry = entry[ix]\n for i in range(N):\n c = event_observed_color if event_observed[i] else event_censored_color\n ax.hlines(i, entry[i], entry[i] + durations[i], color=c, lw=1.5)\n if left_truncated:\n ax.hlines(i, 0, entry[i], color=c, lw=1.0, linestyle='--')\n m = '' if not event_observed[i] else 'o'\n ax.scatter(entry[i] + durations[i], i, color=c, marker=m, s=10)\n ax.set_ylim(-0.5, N)\n return ax\n",
"role": "user"
},
{
"content": "plot_lifetimes",
"role": "assistant"
}
] |
def x(cls, loc=None, iloc=None, show_censors=False, censor_styles=None, **
kwargs):
def x(s):
return np.log(-np.log(s))
if loc is not None and iloc is not None:
raise ValueError('Cannot set both loc and iloc in call to .plot().')
if censor_styles is None:
censor_styles = {}
set_kwargs_ax(kwargs)
set_kwargs_color(kwargs)
set_kwargs_drawstyle(kwargs)
kwargs['logx'] = True
dataframe_slicer = create_dataframe_slicer(iloc, loc)
ax = kwargs['ax']
colour = kwargs['c']
if show_censors and cls.event_table['censored'].sum() > 0:
cs = {'marker': '+', 'ms': 12, 'mew': 1}
cs.update(censor_styles)
times = dataframe_slicer(cls.event_table.loc[cls.event_table[
'censored'] > 0]).index.values.astype(float)
v = cls.predict(times)
ax.plot(times, loglog(v), linestyle='None', color=colour, **cs)
dataframe_slicer(loglog(cls.survival_function_)).plot(**kwargs)
ax.set_xlabel('log(timeline)')
ax.set_ylabel('log(-log(survival_function_))')
return ax
| plot_loglogs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, loc=None, iloc=None, show_censors=False, censor_styles=None, **\n kwargs):\n\n def x(s):\n return np.log(-np.log(s))\n if loc is not None and iloc is not None:\n raise ValueError('Cannot set both loc and iloc in call to .plot().')\n if censor_styles is None:\n censor_styles = {}\n set_kwargs_ax(kwargs)\n set_kwargs_color(kwargs)\n set_kwargs_drawstyle(kwargs)\n kwargs['logx'] = True\n dataframe_slicer = create_dataframe_slicer(iloc, loc)\n ax = kwargs['ax']\n colour = kwargs['c']\n if show_censors and cls.event_table['censored'].sum() > 0:\n cs = {'marker': '+', 'ms': 12, 'mew': 1}\n cs.update(censor_styles)\n times = dataframe_slicer(cls.event_table.loc[cls.event_table[\n 'censored'] > 0]).index.values.astype(float)\n v = cls.predict(times)\n ax.plot(times, loglog(v), linestyle='None', color=colour, **cs)\n dataframe_slicer(loglog(cls.survival_function_)).plot(**kwargs)\n ax.set_xlabel('log(timeline)')\n ax.set_ylabel('log(-log(survival_function_))')\n return ax\n",
"role": "user"
},
{
"content": "plot_loglogs",
"role": "assistant"
}
] |
def x(cls, estimate=None, confidence_intervals=None, loc=None, iloc=None,
show_censors=False, censor_styles=None, ci_legend=False, ci_force_lines
=False, ci_alpha=0.25, ci_show=True, at_risk_counts=False, **kwargs):
plot_estimate_config = PlotEstimateConfig(cls, estimate,
confidence_intervals, loc, iloc, show_censors, censor_styles, **kwargs)
dataframe_slicer = create_dataframe_slicer(iloc, loc)
if show_censors and cls.event_table['censored'].sum() > 0:
cs = {'marker': '+', 'ms': 12, 'mew': 1}
cs.update(plot_estimate_config.censor_styles)
times = dataframe_slicer(cls.event_table.loc[cls.event_table[
'censored'] > 0]).index.values.astype(float)
v = cls.predict(times)
plot_estimate_config.ax.plot(times, v, linestyle='None', color=
plot_estimate_config.colour, **cs)
dataframe_slicer(plot_estimate_config.estimate_).rename(columns=lambda
_: plot_estimate_config.kwargs.pop('label')).plot(**
plot_estimate_config.kwargs)
if ci_show:
if ci_force_lines:
dataframe_slicer(plot_estimate_config.confidence_interval_).plot(
linestyle='-', linewidth=1, color=[plot_estimate_config.
colour], legend=ci_legend, drawstyle=plot_estimate_config.
kwargs['drawstyle'], ax=plot_estimate_config.ax, alpha=0.6)
else:
x = dataframe_slicer(plot_estimate_config.confidence_interval_
).index.values.astype(float)
lower = dataframe_slicer(plot_estimate_config.
confidence_interval_.filter(like='lower')).values[:, 0]
upper = dataframe_slicer(plot_estimate_config.
confidence_interval_.filter(like='upper')).values[:, 0]
if plot_estimate_config.kwargs['drawstyle'] == 'default':
step = None
elif plot_estimate_config.kwargs['drawstyle'].startswith('step'):
step = plot_estimate_config.kwargs['drawstyle'].replace(
'steps-', '')
plot_estimate_config.ax.fill_between(x, lower, upper, alpha=
ci_alpha, color=plot_estimate_config.colour, linewidth=1.0,
step=step)
if at_risk_counts:
add_at_risk_counts(cls, ax=plot_estimate_config.ax)
return plot_estimate_config.ax
| plot_estimate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, estimate=None, confidence_intervals=None, loc=None, iloc=None,\n show_censors=False, censor_styles=None, ci_legend=False, ci_force_lines\n =False, ci_alpha=0.25, ci_show=True, at_risk_counts=False, **kwargs):\n plot_estimate_config = PlotEstimateConfig(cls, estimate,\n confidence_intervals, loc, iloc, show_censors, censor_styles, **kwargs)\n dataframe_slicer = create_dataframe_slicer(iloc, loc)\n if show_censors and cls.event_table['censored'].sum() > 0:\n cs = {'marker': '+', 'ms': 12, 'mew': 1}\n cs.update(plot_estimate_config.censor_styles)\n times = dataframe_slicer(cls.event_table.loc[cls.event_table[\n 'censored'] > 0]).index.values.astype(float)\n v = cls.predict(times)\n plot_estimate_config.ax.plot(times, v, linestyle='None', color=\n plot_estimate_config.colour, **cs)\n dataframe_slicer(plot_estimate_config.estimate_).rename(columns=lambda\n _: plot_estimate_config.kwargs.pop('label')).plot(**\n plot_estimate_config.kwargs)\n if ci_show:\n if ci_force_lines:\n dataframe_slicer(plot_estimate_config.confidence_interval_).plot(\n linestyle='-', linewidth=1, color=[plot_estimate_config.\n colour], legend=ci_legend, drawstyle=plot_estimate_config.\n kwargs['drawstyle'], ax=plot_estimate_config.ax, alpha=0.6)\n else:\n x = dataframe_slicer(plot_estimate_config.confidence_interval_\n ).index.values.astype(float)\n lower = dataframe_slicer(plot_estimate_config.\n confidence_interval_.filter(like='lower')).values[:, 0]\n upper = dataframe_slicer(plot_estimate_config.\n confidence_interval_.filter(like='upper')).values[:, 0]\n if plot_estimate_config.kwargs['drawstyle'] == 'default':\n step = None\n elif plot_estimate_config.kwargs['drawstyle'].startswith('step'):\n step = plot_estimate_config.kwargs['drawstyle'].replace(\n 'steps-', '')\n plot_estimate_config.ax.fill_between(x, lower, upper, alpha=\n ci_alpha, color=plot_estimate_config.colour, linewidth=1.0,\n step=step)\n if at_risk_counts:\n add_at_risk_counts(cls, ax=plot_estimate_config.ax)\n return plot_estimate_config.ax\n",
"role": "user"
},
{
"content": "plot_estimate",
"role": "assistant"
}
] |
def x(filename, destination='zh-CN', mix=True):
def x(t):
return Translator().translate(t, dest=destination).text
doc = Document(filename)
for p in doc.paragraphs:
txd = tx(p.text)
p.text = p.text + ('\n' + txd if mix else '')
for table in doc.tables:
for row in table.rows:
for cell in row.cells:
txd = tx(cell.text)
p.text = cell.text + ('\n' + txd if mix else '')
f = filename.replace('.doc', destination.lower() + '.doc')
doc.save(f)
| translate_doc | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename, destination='zh-CN', mix=True):\n\n def x(t):\n return Translator().translate(t, dest=destination).text\n doc = Document(filename)\n for p in doc.paragraphs:\n txd = tx(p.text)\n p.text = p.text + ('\\n' + txd if mix else '')\n for table in doc.tables:\n for row in table.rows:\n for cell in row.cells:\n txd = tx(cell.text)\n p.text = cell.text + ('\\n' + txd if mix else '')\n f = filename.replace('.doc', destination.lower() + '.doc')\n doc.save(f)\n",
"role": "user"
},
{
"content": "translate_doc",
"role": "assistant"
}
] |
def x(self, text, dest='en', src='auto'):
dest = dest.lower().split('_', 1)[0]
src = src.lower().split('_', 1)[0]
if src != 'auto' and src not in LANGUAGES:
if src in SPECIAL_CASES:
src = SPECIAL_CASES[src]
elif src in LANGCODES:
src = LANGCODES[src]
else:
raise ValueError('invalid source language')
if dest not in LANGUAGES:
if dest in SPECIAL_CASES:
dest = SPECIAL_CASES[dest]
elif dest in LANGCODES:
dest = LANGCODES[dest]
else:
raise ValueError('invalid destination language')
if isinstance(text, list):
result = []
for item in text:
translated = self.translate(item, dest=dest, src=src)
result.append(translated)
return result
origin = text
data = self._translate(text, dest, src)
translated = ''.join([(d[0] if d[0] else '') for d in data[0]])
extra_data = self._parse_extra_data(data)
try:
src = data[2]
except Exception:
pass
pron = origin
try:
pron = data[0][1][-2]
except Exception:
pass
if not PY3 and isinstance(pron, unicode) and isinstance(origin, str):
origin = origin.decode('utf-8')
if dest in EXCLUDES and pron == origin:
pron = translated
if not PY3:
if isinstance(src, str):
src = src.decode('utf-8')
if isinstance(dest, str):
dest = dest.decode('utf-8')
if isinstance(translated, str):
translated = translated.decode('utf-8')
result = Translated(src=src, dest=dest, origin=origin, text=translated,
pronunciation=pron, extra_data=extra_data)
return result
| translate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, text, dest='en', src='auto'):\n dest = dest.lower().split('_', 1)[0]\n src = src.lower().split('_', 1)[0]\n if src != 'auto' and src not in LANGUAGES:\n if src in SPECIAL_CASES:\n src = SPECIAL_CASES[src]\n elif src in LANGCODES:\n src = LANGCODES[src]\n else:\n raise ValueError('invalid source language')\n if dest not in LANGUAGES:\n if dest in SPECIAL_CASES:\n dest = SPECIAL_CASES[dest]\n elif dest in LANGCODES:\n dest = LANGCODES[dest]\n else:\n raise ValueError('invalid destination language')\n if isinstance(text, list):\n result = []\n for item in text:\n translated = self.translate(item, dest=dest, src=src)\n result.append(translated)\n return result\n origin = text\n data = self._translate(text, dest, src)\n translated = ''.join([(d[0] if d[0] else '') for d in data[0]])\n extra_data = self._parse_extra_data(data)\n try:\n src = data[2]\n except Exception:\n pass\n pron = origin\n try:\n pron = data[0][1][-2]\n except Exception:\n pass\n if not PY3 and isinstance(pron, unicode) and isinstance(origin, str):\n origin = origin.decode('utf-8')\n if dest in EXCLUDES and pron == origin:\n pron = translated\n if not PY3:\n if isinstance(src, str):\n src = src.decode('utf-8')\n if isinstance(dest, str):\n dest = dest.decode('utf-8')\n if isinstance(translated, str):\n translated = translated.decode('utf-8')\n result = Translated(src=src, dest=dest, origin=origin, text=translated,\n pronunciation=pron, extra_data=extra_data)\n return result\n",
"role": "user"
},
{
"content": "translate",
"role": "assistant"
}
] |
def x(self, text):
if isinstance(text, list):
result = []
for item in text:
lang = self.detect(item)
result.append(lang)
return result
data = self._translate(text, dest='en', src='auto')
src = ''
confidence = 0.0
try:
src = ''.join(data[8][0])
confidence = data[8][-2][0]
except Exception:
pass
result = Detected(lang=src, confidence=confidence)
return result
| detect | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, text):\n if isinstance(text, list):\n result = []\n for item in text:\n lang = self.detect(item)\n result.append(lang)\n return result\n data = self._translate(text, dest='en', src='auto')\n src = ''\n confidence = 0.0\n try:\n src = ''.join(data[8][0])\n confidence = data[8][-2][0]\n except Exception:\n pass\n result = Detected(lang=src, confidence=confidence)\n return result\n",
"role": "user"
},
{
"content": "detect",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.