source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(requestContext, *seriesLists):
if not seriesLists or not any(seriesLists):
series = constantLine(requestContext, 0).pop()
series.pathExpression = 'countSeries()'
else:
seriesList, start, end, step = normalize(seriesLists)
name = 'countSeries(%s)' % formatPathExpressions(seriesList)
values = (int(len(row)) for row in zip_longest(*seriesList))
series = TimeSeries(name, start, end, step, values)
series.pathExpression = name
return [series]
| countSeries | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(requestContext, *seriesLists):\n if not seriesLists or not any(seriesLists):\n series = constantLine(requestContext, 0).pop()\n series.pathExpression = 'countSeries()'\n else:\n seriesList, start, end, step = normalize(seriesLists)\n name = 'countSeries(%s)' % formatPathExpressions(seriesList)\n values = (int(len(row)) for row in zip_longest(*seriesList))\n series = TimeSeries(name, start, end, step, values)\n series.pathExpression = name\n return [series]\n",
"role": "user"
},
{
"content": "countSeries",
"role": "assistant"
}
] |
def x(requestContext, *seriesLists):
seriesGroup = []
for s in seriesLists:
seriesGroup.extend(s)
return seriesGroup
| group | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(requestContext, *seriesLists):\n seriesGroup = []\n for s in seriesLists:\n seriesGroup.extend(s)\n return seriesGroup\n",
"role": "user"
},
{
"content": "group",
"role": "assistant"
}
] |
def x(requestContext, seriesList, mapNode):
metaSeries = {}
keys = []
for series in seriesList:
key = series.name.split('.')[mapNode]
if key not in metaSeries:
metaSeries[key] = [series]
keys.append(key)
else:
metaSeries[key].append(series)
return [metaSeries[k] for k in keys]
| mapSeries | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(requestContext, seriesList, mapNode):\n metaSeries = {}\n keys = []\n for series in seriesList:\n key = series.name.split('.')[mapNode]\n if key not in metaSeries:\n metaSeries[key] = [series]\n keys.append(key)\n else:\n metaSeries[key].append(series)\n return [metaSeries[k] for k in keys]\n",
"role": "user"
},
{
"content": "mapSeries",
"role": "assistant"
}
] |
def x(requestContext, seriesLists, reduceFunction, reduceNode, *reduceMatchers
):
metaSeries = {}
keys = []
for seriesList in seriesLists:
for series in seriesList:
nodes = series.name.split('.')
node = nodes[reduceNode]
reduceSeriesName = '.'.join(nodes[0:reduceNode]
) + '.reduce.' + reduceFunction
if node in reduceMatchers:
if reduceSeriesName not in metaSeries:
metaSeries[reduceSeriesName] = [None] * len(reduceMatchers)
keys.append(reduceSeriesName)
i = reduceMatchers.index(node)
metaSeries[reduceSeriesName][i] = series
for key in keys:
metaSeries[key] = app.functions[reduceFunction](requestContext, *[[
s] for s in metaSeries[key]])[0]
metaSeries[key].name = key
return [metaSeries[key] for key in keys]
| reduceSeries | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(requestContext, seriesLists, reduceFunction, reduceNode, *reduceMatchers\n ):\n metaSeries = {}\n keys = []\n for seriesList in seriesLists:\n for series in seriesList:\n nodes = series.name.split('.')\n node = nodes[reduceNode]\n reduceSeriesName = '.'.join(nodes[0:reduceNode]\n ) + '.reduce.' + reduceFunction\n if node in reduceMatchers:\n if reduceSeriesName not in metaSeries:\n metaSeries[reduceSeriesName] = [None] * len(reduceMatchers)\n keys.append(reduceSeriesName)\n i = reduceMatchers.index(node)\n metaSeries[reduceSeriesName][i] = series\n for key in keys:\n metaSeries[key] = app.functions[reduceFunction](requestContext, *[[\n s] for s in metaSeries[key]])[0]\n metaSeries[key].name = key\n return [metaSeries[key] for key in keys]\n",
"role": "user"
},
{
"content": "reduceSeries",
"role": "assistant"
}
] |
def x(requestContext, seriesList, nodeNum, templateFunction, newName=None):
from .app import evaluateTarget
prefixes = set()
for series in seriesList:
prefix = '.'.join(series.name.split('.')[:nodeNum + 1])
prefixes.add(prefix)
results = []
for prefix in sorted(prefixes):
target = templateFunction.replace('%', prefix)
for resultSeries in evaluateTarget(requestContext, target):
if newName:
resultSeries.name = newName.replace('%', prefix)
resultSeries.pathExpression = prefix
resultSeries.start = series.start
resultSeries.end = series.end
results.append(resultSeries)
return results
| applyByNode | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(requestContext, seriesList, nodeNum, templateFunction, newName=None):\n from .app import evaluateTarget\n prefixes = set()\n for series in seriesList:\n prefix = '.'.join(series.name.split('.')[:nodeNum + 1])\n prefixes.add(prefix)\n results = []\n for prefix in sorted(prefixes):\n target = templateFunction.replace('%', prefix)\n for resultSeries in evaluateTarget(requestContext, target):\n if newName:\n resultSeries.name = newName.replace('%', prefix)\n resultSeries.pathExpression = prefix\n resultSeries.start = series.start\n resultSeries.end = series.end\n results.append(resultSeries)\n return results\n",
"role": "user"
},
{
"content": "applyByNode",
"role": "assistant"
}
] |
def x(requestContext, seriesList, nodeNum, callback):
return groupByNodes(requestContext, seriesList, callback, nodeNum)
| groupByNode | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(requestContext, seriesList, nodeNum, callback):\n return groupByNodes(requestContext, seriesList, callback, nodeNum)\n",
"role": "user"
},
{
"content": "groupByNode",
"role": "assistant"
}
] |
def x(requestContext, seriesList, callback, *nodes):
from .app import app
metaSeries = {}
keys = []
if isinstance(nodes, int):
nodes = [nodes]
for series in seriesList:
key = '.'.join(series.name.split('.')[n] for n in nodes)
if key not in metaSeries:
metaSeries[key] = [series]
keys.append(key)
else:
metaSeries[key].append(series)
for key in metaSeries:
metaSeries[key] = app.functions[callback](requestContext,
metaSeries[key])[0]
metaSeries[key].name = key
return [metaSeries[key] for key in keys]
| groupByNodes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(requestContext, seriesList, callback, *nodes):\n from .app import app\n metaSeries = {}\n keys = []\n if isinstance(nodes, int):\n nodes = [nodes]\n for series in seriesList:\n key = '.'.join(series.name.split('.')[n] for n in nodes)\n if key not in metaSeries:\n metaSeries[key] = [series]\n keys.append(key)\n else:\n metaSeries[key].append(series)\n for key in metaSeries:\n metaSeries[key] = app.functions[callback](requestContext,\n metaSeries[key])[0]\n metaSeries[key].name = key\n return [metaSeries[key] for key in keys]\n",
"role": "user"
},
{
"content": "groupByNodes",
"role": "assistant"
}
] |
def x(requestContext, seriesList, pattern):
regex = re.compile(pattern)
return [s for s in seriesList if not regex.search(s.name)]
| exclude | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(requestContext, seriesList, pattern):\n regex = re.compile(pattern)\n return [s for s in seriesList if not regex.search(s.name)]\n",
"role": "user"
},
{
"content": "exclude",
"role": "assistant"
}
] |
def x(requestContext, seriesList, intervalString, func='sum'):
results = []
delta = parseTimeOffset(intervalString)
interval = to_seconds(delta)
requestContext = requestContext.copy()
tzinfo = requestContext['tzinfo']
s = requestContext['startTime']
if interval >= DAY:
requestContext['startTime'] = datetime(s.year, s.month, s.day,
tzinfo=tzinfo)
elif interval >= HOUR:
requestContext['startTime'] = datetime(s.year, s.month, s.day, s.
hour, tzinfo=tzinfo)
elif interval >= MINUTE:
requestContext['startTime'] = datetime(s.year, s.month, s.day, s.
hour, s.minute, tzinfo=tzinfo)
paths = []
for series in seriesList:
paths.extend(pathsFromTarget(requestContext, series.pathExpression))
data_store = fetchData(requestContext, paths)
for series in seriesList:
newSeries = evaluateTarget(requestContext, series.pathExpression,
data_store)[0]
series[0:len(series)] = newSeries
series.start = newSeries.start
series.end = newSeries.end
series.step = newSeries.step
for series in seriesList:
buckets = {}
timestamps = range(int(series.start), int(series.end), int(series.step)
)
datapoints = zip_longest(timestamps, series)
for timestamp, value in datapoints:
if not timestamp:
continue
bucketInterval = int((timestamp - series.start) / interval)
if bucketInterval not in buckets:
buckets[bucketInterval] = []
if value is not None:
buckets[bucketInterval].append(value)
newValues = []
for timestamp in range(series.start, series.end, interval):
bucketInterval = int((timestamp - series.start) / interval)
bucket = buckets.get(bucketInterval, [])
if bucket:
if func == 'avg':
newValues.append(float(sum(bucket)) / float(len(bucket)))
elif func == 'last':
newValues.append(bucket[len(bucket) - 1])
elif func == 'max':
newValues.append(max(bucket))
elif func == 'min':
newValues.append(min(bucket))
else:
newValues.append(sum(bucket))
else:
newValues.append(None)
newName = 'smartSummarize(%s, "%s", "%s")' % (series.name,
intervalString, func)
alignedEnd = series.start + bucketInterval * interval + interval
newSeries = TimeSeries(newName, series.start, alignedEnd, interval,
newValues)
newSeries.pathExpression = newName
results.append(newSeries)
return results
| smartSummarize | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(requestContext, seriesList, intervalString, func='sum'):\n results = []\n delta = parseTimeOffset(intervalString)\n interval = to_seconds(delta)\n requestContext = requestContext.copy()\n tzinfo = requestContext['tzinfo']\n s = requestContext['startTime']\n if interval >= DAY:\n requestContext['startTime'] = datetime(s.year, s.month, s.day,\n tzinfo=tzinfo)\n elif interval >= HOUR:\n requestContext['startTime'] = datetime(s.year, s.month, s.day, s.\n hour, tzinfo=tzinfo)\n elif interval >= MINUTE:\n requestContext['startTime'] = datetime(s.year, s.month, s.day, s.\n hour, s.minute, tzinfo=tzinfo)\n paths = []\n for series in seriesList:\n paths.extend(pathsFromTarget(requestContext, series.pathExpression))\n data_store = fetchData(requestContext, paths)\n for series in seriesList:\n newSeries = evaluateTarget(requestContext, series.pathExpression,\n data_store)[0]\n series[0:len(series)] = newSeries\n series.start = newSeries.start\n series.end = newSeries.end\n series.step = newSeries.step\n for series in seriesList:\n buckets = {}\n timestamps = range(int(series.start), int(series.end), int(series.step)\n )\n datapoints = zip_longest(timestamps, series)\n for timestamp, value in datapoints:\n if not timestamp:\n continue\n bucketInterval = int((timestamp - series.start) / interval)\n if bucketInterval not in buckets:\n buckets[bucketInterval] = []\n if value is not None:\n buckets[bucketInterval].append(value)\n newValues = []\n for timestamp in range(series.start, series.end, interval):\n bucketInterval = int((timestamp - series.start) / interval)\n bucket = buckets.get(bucketInterval, [])\n if bucket:\n if func == 'avg':\n newValues.append(float(sum(bucket)) / float(len(bucket)))\n elif func == 'last':\n newValues.append(bucket[len(bucket) - 1])\n elif func == 'max':\n newValues.append(max(bucket))\n elif func == 'min':\n newValues.append(min(bucket))\n else:\n newValues.append(sum(bucket))\n else:\n newValues.append(None)\n newName = 'smartSummarize(%s, \"%s\", \"%s\")' % (series.name,\n intervalString, func)\n alignedEnd = series.start + bucketInterval * interval + interval\n newSeries = TimeSeries(newName, series.start, alignedEnd, interval,\n newValues)\n newSeries.pathExpression = newName\n results.append(newSeries)\n return results\n",
"role": "user"
},
{
"content": "smartSummarize",
"role": "assistant"
}
] |
def x(requestContext, seriesList, intervalString, func='sum', alignToFrom=False
):
results = []
delta = parseTimeOffset(intervalString)
interval = to_seconds(delta)
for series in seriesList:
buckets = {}
timestamps = range(int(series.start), int(series.end) + 1, int(
series.step))
datapoints = zip_longest(timestamps, series)
for timestamp, value in datapoints:
if timestamp is None:
continue
if alignToFrom:
bucketInterval = int((timestamp - series.start) / interval)
else:
bucketInterval = timestamp - timestamp % interval
if bucketInterval not in buckets:
buckets[bucketInterval] = []
if value is not None:
buckets[bucketInterval].append(value)
if alignToFrom:
newStart = series.start
newEnd = series.end
else:
newStart = series.start - series.start % interval
newEnd = series.end - series.end % interval + interval
newValues = []
for timestamp in range(newStart, newEnd, interval):
if alignToFrom:
newEnd = timestamp
bucketInterval = int((timestamp - series.start) / interval)
else:
bucketInterval = timestamp - timestamp % interval
bucket = buckets.get(bucketInterval, [])
if bucket:
if func == 'avg':
newValues.append(float(sum(bucket)) / float(len(bucket)))
elif func == 'last':
newValues.append(bucket[len(bucket) - 1])
elif func == 'max':
newValues.append(max(bucket))
elif func == 'min':
newValues.append(min(bucket))
else:
newValues.append(sum(bucket))
else:
newValues.append(None)
if alignToFrom:
newEnd += interval
newName = 'summarize(%s, "%s", "%s"%s)' % (series.name,
intervalString, func, alignToFrom and ', true' or '')
newSeries = TimeSeries(newName, newStart, newEnd, interval, newValues)
newSeries.pathExpression = newName
results.append(newSeries)
return results
| summarize | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(requestContext, seriesList, intervalString, func='sum', alignToFrom=False\n ):\n results = []\n delta = parseTimeOffset(intervalString)\n interval = to_seconds(delta)\n for series in seriesList:\n buckets = {}\n timestamps = range(int(series.start), int(series.end) + 1, int(\n series.step))\n datapoints = zip_longest(timestamps, series)\n for timestamp, value in datapoints:\n if timestamp is None:\n continue\n if alignToFrom:\n bucketInterval = int((timestamp - series.start) / interval)\n else:\n bucketInterval = timestamp - timestamp % interval\n if bucketInterval not in buckets:\n buckets[bucketInterval] = []\n if value is not None:\n buckets[bucketInterval].append(value)\n if alignToFrom:\n newStart = series.start\n newEnd = series.end\n else:\n newStart = series.start - series.start % interval\n newEnd = series.end - series.end % interval + interval\n newValues = []\n for timestamp in range(newStart, newEnd, interval):\n if alignToFrom:\n newEnd = timestamp\n bucketInterval = int((timestamp - series.start) / interval)\n else:\n bucketInterval = timestamp - timestamp % interval\n bucket = buckets.get(bucketInterval, [])\n if bucket:\n if func == 'avg':\n newValues.append(float(sum(bucket)) / float(len(bucket)))\n elif func == 'last':\n newValues.append(bucket[len(bucket) - 1])\n elif func == 'max':\n newValues.append(max(bucket))\n elif func == 'min':\n newValues.append(min(bucket))\n else:\n newValues.append(sum(bucket))\n else:\n newValues.append(None)\n if alignToFrom:\n newEnd += interval\n newName = 'summarize(%s, \"%s\", \"%s\"%s)' % (series.name,\n intervalString, func, alignToFrom and ', true' or '')\n newSeries = TimeSeries(newName, newStart, newEnd, interval, newValues)\n newSeries.pathExpression = newName\n results.append(newSeries)\n return results\n",
"role": "user"
},
{
"content": "summarize",
"role": "assistant"
}
] |
def x(requestContext, seriesList, intervalString, alignToInterval=False):
results = []
delta = parseTimeOffset(intervalString)
interval = to_seconds(delta)
if alignToInterval:
requestContext = requestContext.copy()
tzinfo = requestContext['tzinfo']
s = requestContext['startTime']
if interval >= DAY:
requestContext['startTime'] = datetime(s.year, s.month, s.day,
tzinfo=tzinfo)
elif interval >= HOUR:
requestContext['startTime'] = datetime(s.year, s.month, s.day,
s.hour, tzinfo=tzinfo)
elif interval >= MINUTE:
requestContext['startTime'] = datetime(s.year, s.month, s.day,
s.hour, s.minute, tzinfo=tzinfo)
paths = []
for series in seriesList:
paths.extend(pathsFromTarget(requestContext, series.pathExpression)
)
data_store = fetchData(requestContext, paths)
for series in seriesList:
newSeries = evaluateTarget(requestContext, series.
pathExpression, data_store)[0]
intervalCount = int((series.end - series.start) / interval)
series[0:len(series)] = newSeries
series.start = newSeries.start
series.end = newSeries.start + intervalCount * interval + interval
series.step = newSeries.step
for series in seriesList:
step = int(series.step)
bucket_count = int(math.ceil(float(series.end - series.start) /
interval))
buckets = [[] for _ in range(bucket_count)]
newStart = int(series.end - bucket_count * interval)
for i, value in enumerate(series):
if value is None:
continue
start_time = int(series.start + i * step)
start_bucket, start_mod = divmod(start_time - newStart, interval)
end_time = start_time + step
end_bucket, end_mod = divmod(end_time - newStart, interval)
if end_bucket >= bucket_count:
end_bucket = bucket_count - 1
end_mod = interval
if start_bucket == end_bucket:
if start_bucket >= 0:
buckets[start_bucket].append(value * (end_mod - start_mod))
else:
if start_bucket >= 0:
buckets[start_bucket].append(value * (interval - start_mod)
)
hits_per_bucket = value * interval
for j in range(start_bucket + 1, end_bucket):
buckets[j].append(hits_per_bucket)
if end_mod > 0:
buckets[end_bucket].append(value * end_mod)
newValues = []
for bucket in buckets:
if bucket:
newValues.append(sum(bucket))
else:
newValues.append(None)
newName = 'hitcount(%s, "%s"%s)' % (series.name, intervalString,
alignToInterval and ', true' or '')
newSeries = TimeSeries(newName, newStart, series.end, interval,
newValues)
newSeries.pathExpression = newName
results.append(newSeries)
return results
| hitcount | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(requestContext, seriesList, intervalString, alignToInterval=False):\n results = []\n delta = parseTimeOffset(intervalString)\n interval = to_seconds(delta)\n if alignToInterval:\n requestContext = requestContext.copy()\n tzinfo = requestContext['tzinfo']\n s = requestContext['startTime']\n if interval >= DAY:\n requestContext['startTime'] = datetime(s.year, s.month, s.day,\n tzinfo=tzinfo)\n elif interval >= HOUR:\n requestContext['startTime'] = datetime(s.year, s.month, s.day,\n s.hour, tzinfo=tzinfo)\n elif interval >= MINUTE:\n requestContext['startTime'] = datetime(s.year, s.month, s.day,\n s.hour, s.minute, tzinfo=tzinfo)\n paths = []\n for series in seriesList:\n paths.extend(pathsFromTarget(requestContext, series.pathExpression)\n )\n data_store = fetchData(requestContext, paths)\n for series in seriesList:\n newSeries = evaluateTarget(requestContext, series.\n pathExpression, data_store)[0]\n intervalCount = int((series.end - series.start) / interval)\n series[0:len(series)] = newSeries\n series.start = newSeries.start\n series.end = newSeries.start + intervalCount * interval + interval\n series.step = newSeries.step\n for series in seriesList:\n step = int(series.step)\n bucket_count = int(math.ceil(float(series.end - series.start) /\n interval))\n buckets = [[] for _ in range(bucket_count)]\n newStart = int(series.end - bucket_count * interval)\n for i, value in enumerate(series):\n if value is None:\n continue\n start_time = int(series.start + i * step)\n start_bucket, start_mod = divmod(start_time - newStart, interval)\n end_time = start_time + step\n end_bucket, end_mod = divmod(end_time - newStart, interval)\n if end_bucket >= bucket_count:\n end_bucket = bucket_count - 1\n end_mod = interval\n if start_bucket == end_bucket:\n if start_bucket >= 0:\n buckets[start_bucket].append(value * (end_mod - start_mod))\n else:\n if start_bucket >= 0:\n buckets[start_bucket].append(value * (interval - start_mod)\n )\n hits_per_bucket = value * interval\n for j in range(start_bucket + 1, end_bucket):\n buckets[j].append(hits_per_bucket)\n if end_mod > 0:\n buckets[end_bucket].append(value * end_mod)\n newValues = []\n for bucket in buckets:\n if bucket:\n newValues.append(sum(bucket))\n else:\n newValues.append(None)\n newName = 'hitcount(%s, \"%s\"%s)' % (series.name, intervalString, \n alignToInterval and ', true' or '')\n newSeries = TimeSeries(newName, newStart, series.end, interval,\n newValues)\n newSeries.pathExpression = newName\n results.append(newSeries)\n return results\n",
"role": "user"
},
{
"content": "hitcount",
"role": "assistant"
}
] |
def x(requestContext, name, step=60):
start = int(epoch(requestContext['startTime']))
end = int(epoch(requestContext['endTime']))
delta = timedelta(seconds=step)
when = requestContext['startTime']
values = []
while when < requestContext['endTime']:
values.append(epoch(when))
when += delta
series = TimeSeries(name, start, end, step, values)
series.pathExpression = name
return [series]
| timeFunction | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(requestContext, name, step=60):\n start = int(epoch(requestContext['startTime']))\n end = int(epoch(requestContext['endTime']))\n delta = timedelta(seconds=step)\n when = requestContext['startTime']\n values = []\n while when < requestContext['endTime']:\n values.append(epoch(when))\n when += delta\n series = TimeSeries(name, start, end, step, values)\n series.pathExpression = name\n return [series]\n",
"role": "user"
},
{
"content": "timeFunction",
"role": "assistant"
}
] |
def x(requestContext, name, amplitude=1, step=60):
delta = timedelta(seconds=step)
when = requestContext['startTime']
values = []
while when < requestContext['endTime']:
values.append(math.sin(epoch(when)) * amplitude)
when += delta
series = TimeSeries(name, int(epoch(requestContext['startTime'])), int(
epoch(requestContext['endTime'])), step, values)
series.pathExpression = 'sin({0})'.format(name)
return [series]
| sinFunction | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(requestContext, name, amplitude=1, step=60):\n delta = timedelta(seconds=step)\n when = requestContext['startTime']\n values = []\n while when < requestContext['endTime']:\n values.append(math.sin(epoch(when)) * amplitude)\n when += delta\n series = TimeSeries(name, int(epoch(requestContext['startTime'])), int(\n epoch(requestContext['endTime'])), step, values)\n series.pathExpression = 'sin({0})'.format(name)\n return [series]\n",
"role": "user"
},
{
"content": "sinFunction",
"role": "assistant"
}
] |
def x(requestContext, name, step=60):
delta = timedelta(seconds=step)
when = requestContext['startTime']
values = []
current = 0
while when < requestContext['endTime']:
values.append(current)
current += random.random() - 0.5
when += delta
return [TimeSeries(name, int(epoch(requestContext['startTime'])), int(
epoch(requestContext['endTime'])), step, values)]
| randomWalkFunction | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(requestContext, name, step=60):\n delta = timedelta(seconds=step)\n when = requestContext['startTime']\n values = []\n current = 0\n while when < requestContext['endTime']:\n values.append(current)\n current += random.random() - 0.5\n when += delta\n return [TimeSeries(name, int(epoch(requestContext['startTime'])), int(\n epoch(requestContext['endTime'])), step, values)]\n",
"role": "user"
},
{
"content": "randomWalkFunction",
"role": "assistant"
}
] |
def x(self, model, checks=[]):
custom = [check_partial(reaction_id_check, frozenset(r.id for r in
model.reactions))]
super(Medium, self).validate(model=model, checks=checks + custom)
| validate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, model, checks=[]):\n custom = [check_partial(reaction_id_check, frozenset(r.id for r in\n model.reactions))]\n super(Medium, self).validate(model=model, checks=checks + custom)\n",
"role": "user"
},
{
"content": "validate",
"role": "assistant"
}
] |
def x(self, model):
model.medium = {row.exchange: row.uptake for row in self.data.
itertuples(index=False)}
| apply | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, model):\n model.medium = {row.exchange: row.uptake for row in self.data.\n itertuples(index=False)}\n",
"role": "user"
},
{
"content": "apply",
"role": "assistant"
}
] |
def x(meta):
meta['timestamp'] = datetime.utcnow().isoformat(' ')
meta['platform'] = platform.system()
meta['release'] = platform.release()
meta['python'] = platform.python_version()
meta['packages'] = get_pkg_info('memote')
| add_environment_information | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(meta):\n meta['timestamp'] = datetime.utcnow().isoformat(' ')\n meta['platform'] = platform.system()\n meta['release'] = platform.release()\n meta['python'] = platform.python_version()\n meta['packages'] = get_pkg_info('memote')\n",
"role": "user"
},
{
"content": "add_environment_information",
"role": "assistant"
}
] |
def x(rxn):
element_dist = defaultdict()
for met in rxn.metabolites:
if met.compartment not in element_dist:
element_dist[met.compartment] = {k: (v * rxn.metabolites[met]) for
k, v in iteritems(met.elements)}
else:
x = {k: (v * rxn.metabolites[met]) for k, v in iteritems(met.
elements)}
y = element_dist[met.compartment]
element_dist[met.compartment] = {k: (x.get(k, 0) + y.get(k, 0)) for
k in set(x) | set(y)}
delta_dict = defaultdict()
for elements in itervalues(element_dist):
delta_dict.update(elements)
delta_dict = {k: abs(v) for k, v in iteritems(delta_dict) if v != 0}
return delta_dict
| find_transported_elements | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rxn):\n element_dist = defaultdict()\n for met in rxn.metabolites:\n if met.compartment not in element_dist:\n element_dist[met.compartment] = {k: (v * rxn.metabolites[met]) for\n k, v in iteritems(met.elements)}\n else:\n x = {k: (v * rxn.metabolites[met]) for k, v in iteritems(met.\n elements)}\n y = element_dist[met.compartment]\n element_dist[met.compartment] = {k: (x.get(k, 0) + y.get(k, 0)) for\n k in set(x) | set(y)}\n delta_dict = defaultdict()\n for elements in itervalues(element_dist):\n delta_dict.update(elements)\n delta_dict = {k: abs(v) for k, v in iteritems(delta_dict) if v != 0}\n return delta_dict\n",
"role": "user"
},
{
"content": "find_transported_elements",
"role": "assistant"
}
] |
def x(model):
transport_reactions = []
transport_rxn_candidates = set(model.reactions) - set(model.boundary
) - set(find_biomass_reaction(model))
transport_rxn_candidates = set([rxn for rxn in transport_rxn_candidates if
len(rxn.compartments) >= 2])
sbo_matches = set([rxn for rxn in transport_rxn_candidates if rxn.
annotation is not None and 'sbo' in rxn.annotation and rxn.
annotation['sbo'] in TRANSPORT_RXN_SBO_TERMS])
if len(sbo_matches) > 0:
transport_reactions += list(sbo_matches)
for rxn in transport_rxn_candidates:
rxn_mets = set([met.formula for met in rxn.metabolites])
if None not in rxn_mets and len(rxn_mets) != 0:
if is_transport_reaction_formulae(rxn):
transport_reactions.append(rxn)
elif is_transport_reaction_annotations(rxn):
transport_reactions.append(rxn)
return set(transport_reactions)
| find_transport_reactions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n transport_reactions = []\n transport_rxn_candidates = set(model.reactions) - set(model.boundary\n ) - set(find_biomass_reaction(model))\n transport_rxn_candidates = set([rxn for rxn in transport_rxn_candidates if\n len(rxn.compartments) >= 2])\n sbo_matches = set([rxn for rxn in transport_rxn_candidates if rxn.\n annotation is not None and 'sbo' in rxn.annotation and rxn.\n annotation['sbo'] in TRANSPORT_RXN_SBO_TERMS])\n if len(sbo_matches) > 0:\n transport_reactions += list(sbo_matches)\n for rxn in transport_rxn_candidates:\n rxn_mets = set([met.formula for met in rxn.metabolites])\n if None not in rxn_mets and len(rxn_mets) != 0:\n if is_transport_reaction_formulae(rxn):\n transport_reactions.append(rxn)\n elif is_transport_reaction_annotations(rxn):\n transport_reactions.append(rxn)\n return set(transport_reactions)\n",
"role": "user"
},
{
"content": "find_transport_reactions",
"role": "assistant"
}
] |
def x(rxn):
rxn_reactants = set([met.formula for met in rxn.reactants])
rxn_products = set([met.formula for met in rxn.products])
transported_mets = [formula for formula in rxn_reactants if formula in
rxn_products]
delta_dicts = find_transported_elements(rxn)
non_zero_array = [v for k, v in iteritems(delta_dicts) if v != 0]
if set(transported_mets) != set('H') and list(delta_dicts.keys()) == ['H']:
pass
elif sum(non_zero_array):
return True
| is_transport_reaction_formulae | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rxn):\n rxn_reactants = set([met.formula for met in rxn.reactants])\n rxn_products = set([met.formula for met in rxn.products])\n transported_mets = [formula for formula in rxn_reactants if formula in\n rxn_products]\n delta_dicts = find_transported_elements(rxn)\n non_zero_array = [v for k, v in iteritems(delta_dicts) if v != 0]\n if set(transported_mets) != set('H') and list(delta_dicts.keys()) == ['H']:\n pass\n elif sum(non_zero_array):\n return True\n",
"role": "user"
},
{
"content": "is_transport_reaction_formulae",
"role": "assistant"
}
] |
def x(rxn):
reactants = set([(k, tuple(v)) for met in rxn.reactants for k, v in
iteritems(met.annotation) if met.id != 'H' and k is not None and k !=
'sbo' and v is not None])
products = set([(k, tuple(v)) for met in rxn.products for k, v in
iteritems(met.annotation) if met.id != 'H' and k is not None and k !=
'sbo' and v is not None])
transported_mets = reactants & products
if len(transported_mets) > 0:
return True
| is_transport_reaction_annotations | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rxn):\n reactants = set([(k, tuple(v)) for met in rxn.reactants for k, v in\n iteritems(met.annotation) if met.id != 'H' and k is not None and k !=\n 'sbo' and v is not None])\n products = set([(k, tuple(v)) for met in rxn.products for k, v in\n iteritems(met.annotation) if met.id != 'H' and k is not None and k !=\n 'sbo' and v is not None])\n transported_mets = reactants & products\n if len(transported_mets) > 0:\n return True\n",
"role": "user"
},
{
"content": "is_transport_reaction_annotations",
"role": "assistant"
}
] |
def x(model, pair):
first = set(find_met_in_model(model, pair[0]))
second = set(find_met_in_model(model, pair[1]))
hits = list()
for rxn in model.reactions:
if len(first & set(rxn.reactants)) > 0 and len(second & set(rxn.
products)) > 0:
hits.append(rxn)
elif len(first & set(rxn.products)) > 0 and len(second & set(rxn.
reactants)) > 0:
hits.append(rxn)
return frozenset(hits)
| find_converting_reactions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model, pair):\n first = set(find_met_in_model(model, pair[0]))\n second = set(find_met_in_model(model, pair[1]))\n hits = list()\n for rxn in model.reactions:\n if len(first & set(rxn.reactants)) > 0 and len(second & set(rxn.\n products)) > 0:\n hits.append(rxn)\n elif len(first & set(rxn.products)) > 0 and len(second & set(rxn.\n reactants)) > 0:\n hits.append(rxn)\n return frozenset(hits)\n",
"role": "user"
},
{
"content": "find_converting_reactions",
"role": "assistant"
}
] |
def x(model):
sbo_matches = set([rxn for rxn in model.reactions if rxn.annotation is not
None and 'sbo' in rxn.annotation and rxn.annotation['sbo'] ==
'SBO:0000629'])
if len(sbo_matches) > 0:
return list(sbo_matches)
buzzwords = ['biomass', 'growth', 'bof']
buzzword_matches = set([rxn for rxn in model.reactions if any(string in
rxn.id.lower() for string in buzzwords)])
biomass_met = []
for met in model.metabolites:
if 'biomass' in met.id.lower(
) or met.name is not None and 'biomass' in met.name.lower():
biomass_met.append(met)
if biomass_met == 1:
biomass_met_matches = set(biomass_met.reactions) - set(model.boundary)
else:
biomass_met_matches = set()
return list(buzzword_matches | biomass_met_matches)
| find_biomass_reaction | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n sbo_matches = set([rxn for rxn in model.reactions if rxn.annotation is not\n None and 'sbo' in rxn.annotation and rxn.annotation['sbo'] ==\n 'SBO:0000629'])\n if len(sbo_matches) > 0:\n return list(sbo_matches)\n buzzwords = ['biomass', 'growth', 'bof']\n buzzword_matches = set([rxn for rxn in model.reactions if any(string in\n rxn.id.lower() for string in buzzwords)])\n biomass_met = []\n for met in model.metabolites:\n if 'biomass' in met.id.lower(\n ) or met.name is not None and 'biomass' in met.name.lower():\n biomass_met.append(met)\n if biomass_met == 1:\n biomass_met_matches = set(biomass_met.reactions) - set(model.boundary)\n else:\n biomass_met_matches = set()\n return list(buzzword_matches | biomass_met_matches)\n",
"role": "user"
},
{
"content": "find_biomass_reaction",
"role": "assistant"
}
] |
def x(model):
try:
extracellular = find_compartment_id_in_model(model, 'e')
except KeyError:
extracellular = None
return find_boundary_types(model, 'demand', extracellular)
| find_demand_reactions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n try:\n extracellular = find_compartment_id_in_model(model, 'e')\n except KeyError:\n extracellular = None\n return find_boundary_types(model, 'demand', extracellular)\n",
"role": "user"
},
{
"content": "find_demand_reactions",
"role": "assistant"
}
] |
def x(model):
try:
extracellular = find_compartment_id_in_model(model, 'e')
except KeyError:
extracellular = None
return find_boundary_types(model, 'sink', extracellular)
| find_sink_reactions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n try:\n extracellular = find_compartment_id_in_model(model, 'e')\n except KeyError:\n extracellular = None\n return find_boundary_types(model, 'sink', extracellular)\n",
"role": "user"
},
{
"content": "find_sink_reactions",
"role": "assistant"
}
] |
def x(model):
try:
extracellular = find_compartment_id_in_model(model, 'e')
except KeyError:
extracellular = None
return find_boundary_types(model, 'exchange', extracellular)
| find_exchange_rxns | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n try:\n extracellular = find_compartment_id_in_model(model, 'e')\n except KeyError:\n extracellular = None\n return find_boundary_types(model, 'exchange', extracellular)\n",
"role": "user"
},
{
"content": "find_exchange_rxns",
"role": "assistant"
}
] |
def x(model, biomass=None):
boundary = set(model.boundary)
transporters = find_transport_reactions(model)
if biomass is None:
biomass = set(find_biomass_reaction(model))
return boundary | transporters | biomass
| find_interchange_biomass_reactions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model, biomass=None):\n boundary = set(model.boundary)\n transporters = find_transport_reactions(model)\n if biomass is None:\n biomass = set(find_biomass_reaction(model))\n return boundary | transporters | biomass\n",
"role": "user"
},
{
"content": "find_interchange_biomass_reactions",
"role": "assistant"
}
] |
def x(model, rxn_id, direction='max', single_value=True):
model.objective = model.reactions.get_by_id(rxn_id)
model.objective_direction = direction
if single_value:
try:
return model.slim_optimize()
except Infeasible:
return np.nan
else:
try:
solution = model.optimize()
return solution
except Infeasible:
return np.nan
| run_fba | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model, rxn_id, direction='max', single_value=True):\n model.objective = model.reactions.get_by_id(rxn_id)\n model.objective_direction = direction\n if single_value:\n try:\n return model.slim_optimize()\n except Infeasible:\n return np.nan\n else:\n try:\n solution = model.optimize()\n return solution\n except Infeasible:\n return np.nan\n",
"role": "user"
},
{
"content": "run_fba",
"role": "assistant"
}
] |
def x(model):
for rxn in model.reactions:
if rxn.reversibility:
rxn.bounds = -1, 1
else:
rxn.bounds = 0, 1
for boundary in model.boundary:
boundary.bounds = 0, 0
| close_boundaries_sensibly | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n for rxn in model.reactions:\n if rxn.reversibility:\n rxn.bounds = -1, 1\n else:\n rxn.bounds = 0, 1\n for boundary in model.boundary:\n boundary.bounds = 0, 0\n",
"role": "user"
},
{
"content": "close_boundaries_sensibly",
"role": "assistant"
}
] |
def x(model, compartment_id):
return [met for met in model.metabolites if met.compartment ==
compartment_id]
| metabolites_per_compartment | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model, compartment_id):\n return [met for met in model.metabolites if met.compartment ==\n compartment_id]\n",
"role": "user"
},
{
"content": "metabolites_per_compartment",
"role": "assistant"
}
] |
def x(model):
candidate, second = sorted(((c, len(metabolites_per_compartment(model,
c))) for c in model.compartments), reverse=True, key=itemgetter(1))[:2]
if candidate[1] == second[1]:
raise RuntimeError(
'There is a tie for the largest compartment. Compartment {} and {} have equal amounts of metabolites.'
.format(candidate[0], second[0]))
else:
return candidate[0]
| largest_compartment_id_met | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n candidate, second = sorted(((c, len(metabolites_per_compartment(model,\n c))) for c in model.compartments), reverse=True, key=itemgetter(1))[:2]\n if candidate[1] == second[1]:\n raise RuntimeError(\n 'There is a tie for the largest compartment. Compartment {} and {} have equal amounts of metabolites.'\n .format(candidate[0], second[0]))\n else:\n return candidate[0]\n",
"role": "user"
},
{
"content": "largest_compartment_id_met",
"role": "assistant"
}
] |
def x(model, compartment_id):
if compartment_id not in COMPARTMENT_SHORTLIST.keys():
raise KeyError(
'{} is not in the COMPARTMENT_SHORTLIST! Make sure you typed the ID correctly, if yes, update the shortlist manually.'
.format(compartment_id))
if len(model.compartments) == 0:
raise KeyError(
'It was not possible to identify the compartment {}, since the model has no compartments at all.'
.format(COMPARTMENT_SHORTLIST[compartment_id][0]))
if compartment_id in model.compartments.keys():
return compartment_id
for name in COMPARTMENT_SHORTLIST[compartment_id]:
for c_id, c_name in model.compartments.items():
if c_name.lower() == name:
return c_id
if compartment_id == 'c':
return largest_compartment_id_met(model)
| find_compartment_id_in_model | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model, compartment_id):\n if compartment_id not in COMPARTMENT_SHORTLIST.keys():\n raise KeyError(\n '{} is not in the COMPARTMENT_SHORTLIST! Make sure you typed the ID correctly, if yes, update the shortlist manually.'\n .format(compartment_id))\n if len(model.compartments) == 0:\n raise KeyError(\n 'It was not possible to identify the compartment {}, since the model has no compartments at all.'\n .format(COMPARTMENT_SHORTLIST[compartment_id][0]))\n if compartment_id in model.compartments.keys():\n return compartment_id\n for name in COMPARTMENT_SHORTLIST[compartment_id]:\n for c_id, c_name in model.compartments.items():\n if c_name.lower() == name:\n return c_id\n if compartment_id == 'c':\n return largest_compartment_id_met(model)\n",
"role": "user"
},
{
"content": "find_compartment_id_in_model",
"role": "assistant"
}
] |
def x(model, mnx_id, compartment_id=None):
def x(annotation):
query_values = set(utils.flatten(annotation.values()))
ref_values = set(utils.flatten(METANETX_SHORTLIST[mnx_id]))
return query_values & ref_values
if mnx_id not in METANETX_SHORTLIST.columns:
raise ValueError(
'{} is not in the MetaNetX Shortlist! Make sure you typed the ID correctly, if yes, update the shortlist by updating and re-running the script generate_mnx_shortlists.py.'
.format(mnx_id))
candidates = []
regex = re.compile('^{}(_[a-zA-Z0-9]+)?$'.format(mnx_id))
if model.metabolites.query(regex):
candidates = model.metabolites.query(regex)
elif model.metabolites.query(compare_annotation, attribute='annotation'):
candidates = model.metabolites.query(compare_annotation, attribute=
'annotation')
else:
for value in METANETX_SHORTLIST[mnx_id]:
if value:
for ident in value:
regex = re.compile('^{}(_[a-zA-Z0-9]+)?$'.format(ident))
if model.metabolites.query(regex, attribute='id'):
candidates.extend(model.metabolites.query(regex,
attribute='id'))
if compartment_id is None:
print('compartment_id = None?')
return candidates
else:
candidates_in_compartment = [cand for cand in candidates if cand.
compartment == compartment_id]
if len(candidates_in_compartment) == 0:
raise RuntimeError(
'It was not possible to identify any metabolite in compartment {} corresponding to the following MetaNetX identifier: {}.Make sure that a cross-reference to this ID in the MetaNetX Database exists for your identifier namespace.'
.format(compartment_id, mnx_id))
elif len(candidates_in_compartment) > 1:
raise RuntimeError(
'It was not possible to uniquely identify a single metabolite in compartment {} that corresponds to the following MetaNetX identifier: {}.Instead these candidates were found: {}.Check that metabolite compartment tags are correct. Consider switching to a namespace scheme where identifiers are truly unique.'
.format(compartment_id, mnx_id, utils.get_ids(
candidates_in_compartment)))
else:
return candidates_in_compartment
| find_met_in_model | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model, mnx_id, compartment_id=None):\n\n def x(annotation):\n query_values = set(utils.flatten(annotation.values()))\n ref_values = set(utils.flatten(METANETX_SHORTLIST[mnx_id]))\n return query_values & ref_values\n if mnx_id not in METANETX_SHORTLIST.columns:\n raise ValueError(\n '{} is not in the MetaNetX Shortlist! Make sure you typed the ID correctly, if yes, update the shortlist by updating and re-running the script generate_mnx_shortlists.py.'\n .format(mnx_id))\n candidates = []\n regex = re.compile('^{}(_[a-zA-Z0-9]+)?$'.format(mnx_id))\n if model.metabolites.query(regex):\n candidates = model.metabolites.query(regex)\n elif model.metabolites.query(compare_annotation, attribute='annotation'):\n candidates = model.metabolites.query(compare_annotation, attribute=\n 'annotation')\n else:\n for value in METANETX_SHORTLIST[mnx_id]:\n if value:\n for ident in value:\n regex = re.compile('^{}(_[a-zA-Z0-9]+)?$'.format(ident))\n if model.metabolites.query(regex, attribute='id'):\n candidates.extend(model.metabolites.query(regex,\n attribute='id'))\n if compartment_id is None:\n print('compartment_id = None?')\n return candidates\n else:\n candidates_in_compartment = [cand for cand in candidates if cand.\n compartment == compartment_id]\n if len(candidates_in_compartment) == 0:\n raise RuntimeError(\n 'It was not possible to identify any metabolite in compartment {} corresponding to the following MetaNetX identifier: {}.Make sure that a cross-reference to this ID in the MetaNetX Database exists for your identifier namespace.'\n .format(compartment_id, mnx_id))\n elif len(candidates_in_compartment) > 1:\n raise RuntimeError(\n 'It was not possible to uniquely identify a single metabolite in compartment {} that corresponds to the following MetaNetX identifier: {}.Instead these candidates were found: {}.Check that metabolite compartment tags are correct. Consider switching to a namespace scheme where identifiers are truly unique.'\n .format(compartment_id, mnx_id, utils.get_ids(\n candidates_in_compartment)))\n else:\n return candidates_in_compartment\n",
"role": "user"
},
{
"content": "find_met_in_model",
"role": "assistant"
}
] |
def x(model):
lower_bounds = np.asarray([rxn.lower_bound for rxn in model.reactions],
dtype=float)
upper_bounds = np.asarray([rxn.upper_bound for rxn in model.reactions],
dtype=float)
lower_bound = np.nanmedian(lower_bounds[lower_bounds != 0.0])
upper_bound = np.nanmedian(upper_bounds[upper_bounds != 0.0])
if np.isnan(lower_bound):
LOGGER.warning('Could not identify a median lower bound.')
lower_bound = -1000.0
if np.isnan(upper_bound):
LOGGER.warning('Could not identify a median upper bound.')
upper_bound = 1000.0
return lower_bound, upper_bound
| find_bounds | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n lower_bounds = np.asarray([rxn.lower_bound for rxn in model.reactions],\n dtype=float)\n upper_bounds = np.asarray([rxn.upper_bound for rxn in model.reactions],\n dtype=float)\n lower_bound = np.nanmedian(lower_bounds[lower_bounds != 0.0])\n upper_bound = np.nanmedian(upper_bounds[upper_bounds != 0.0])\n if np.isnan(lower_bound):\n LOGGER.warning('Could not identify a median lower bound.')\n lower_bound = -1000.0\n if np.isnan(upper_bound):\n LOGGER.warning('Could not identify a median upper bound.')\n upper_bound = 1000.0\n return lower_bound, upper_bound\n",
"role": "user"
},
{
"content": "find_bounds",
"role": "assistant"
}
] |
def x(self):
return self._template.safe_substitute(report_type=self._report_type,
results=self.render_json())
| render_html | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return self._template.safe_substitute(report_type=self._report_type,\n results=self.render_json())\n",
"role": "user"
},
{
"content": "render_html",
"role": "assistant"
}
] |
def x(self):
cases = self.get_configured_tests() | set(self.result.cases)
scores = DataFrame({'score': 0.0, 'max': 1.0}, index=sorted(cases))
self.result.setdefault('score', dict())
self.result['score']['sections'] = list()
for test, result in iteritems(self.result.cases):
metric = result['metric']
if hasattr(metric, 'items'):
result['score'] = test_score = dict()
total = 0.0
for key, value in iteritems(metric):
value = 1.0 - value
total += value
test_score[key] = value
if len(metric) == 0:
metric = 0.0
else:
metric = total / len(metric)
else:
metric = 1.0 - metric
scores.at[test, 'score'] = metric
scores.loc[test, :] *= self.config['weights'].get(test, 1.0)
score = 0.0
maximum = 0.0
for section_id, card in iteritems(self.config['cards']['scored'][
'sections']):
cases = card.get('cases', None)
if cases is None:
continue
card_score = scores.loc[cases, 'score'].sum()
card_total = scores.loc[cases, 'max'].sum()
section_score = {'section': section_id, 'score': card_score /
card_total}
self.result['score']['sections'].append(section_score)
weight = card.get('weight', 1.0)
score += card_score * weight
maximum += card_total * weight
self.result['score']['total_score'] = score / maximum
| compute_score | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n cases = self.get_configured_tests() | set(self.result.cases)\n scores = DataFrame({'score': 0.0, 'max': 1.0}, index=sorted(cases))\n self.result.setdefault('score', dict())\n self.result['score']['sections'] = list()\n for test, result in iteritems(self.result.cases):\n metric = result['metric']\n if hasattr(metric, 'items'):\n result['score'] = test_score = dict()\n total = 0.0\n for key, value in iteritems(metric):\n value = 1.0 - value\n total += value\n test_score[key] = value\n if len(metric) == 0:\n metric = 0.0\n else:\n metric = total / len(metric)\n else:\n metric = 1.0 - metric\n scores.at[test, 'score'] = metric\n scores.loc[test, :] *= self.config['weights'].get(test, 1.0)\n score = 0.0\n maximum = 0.0\n for section_id, card in iteritems(self.config['cards']['scored'][\n 'sections']):\n cases = card.get('cases', None)\n if cases is None:\n continue\n card_score = scores.loc[cases, 'score'].sum()\n card_total = scores.loc[cases, 'max'].sum()\n section_score = {'section': section_id, 'score': card_score /\n card_total}\n self.result['score']['sections'].append(section_score)\n weight = card.get('weight', 1.0)\n score += card_score * weight\n maximum += card_total * weight\n self.result['score']['total_score'] = score / maximum\n",
"role": "user"
},
{
"content": "compute_score",
"role": "assistant"
}
] |
def x(model, components):
return [elem for elem in getattr(model, components) if elem.annotation is
None or 'sbo' not in elem.annotation]
| find_components_without_sbo_terms | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model, components):\n return [elem for elem in getattr(model, components) if elem.annotation is\n None or 'sbo' not in elem.annotation]\n",
"role": "user"
},
{
"content": "find_components_without_sbo_terms",
"role": "assistant"
}
] |
def x(items, term):
if isinstance(term, list):
return [elem for elem in items if elem.annotation is None or 'sbo'
not in elem.annotation or not any(i in elem.annotation['sbo'] for
i in term)]
else:
return [elem for elem in items if elem.annotation is None or 'sbo'
not in elem.annotation or term not in elem.annotation['sbo']]
| check_component_for_specific_sbo_term | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(items, term):\n if isinstance(term, list):\n return [elem for elem in items if elem.annotation is None or 'sbo'\n not in elem.annotation or not any(i in elem.annotation['sbo'] for\n i in term)]\n else:\n return [elem for elem in items if elem.annotation is None or 'sbo'\n not in elem.annotation or term not in elem.annotation['sbo']]\n",
"role": "user"
},
{
"content": "check_component_for_specific_sbo_term",
"role": "assistant"
}
] |
def x(compounds_identifiers):
return min((c for c in compounds_identifiers if c.startswith('C')), key
=lambda c: int(c[1:]))
| get_smallest_compound_id | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(compounds_identifiers):\n return min((c for c in compounds_identifiers if c.startswith('C')), key\n =lambda c: int(c[1:]))\n",
"role": "user"
},
{
"content": "get_smallest_compound_id",
"role": "assistant"
}
] |
def x(metabolite):
logger.debug('Looking for KEGG compound identifier for %s.', metabolite.id)
kegg_annotation = metabolite.annotation.get('kegg.compound')
if kegg_annotation is None:
logger.warning('No kegg.compound annotation for metabolite %s.',
metabolite.id)
return
if isinstance(kegg_annotation, string_types
) and kegg_annotation.startswith('C'):
return kegg_annotation
elif isinstance(kegg_annotation, Iterable):
try:
return get_smallest_compound_id(kegg_annotation)
except ValueError:
return
logger.warning('No matching kegg.compound annotation for metabolite %s.',
metabolite.id)
return
| map_metabolite2kegg | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(metabolite):\n logger.debug('Looking for KEGG compound identifier for %s.', metabolite.id)\n kegg_annotation = metabolite.annotation.get('kegg.compound')\n if kegg_annotation is None:\n logger.warning('No kegg.compound annotation for metabolite %s.',\n metabolite.id)\n return\n if isinstance(kegg_annotation, string_types\n ) and kegg_annotation.startswith('C'):\n return kegg_annotation\n elif isinstance(kegg_annotation, Iterable):\n try:\n return get_smallest_compound_id(kegg_annotation)\n except ValueError:\n return\n logger.warning('No matching kegg.compound annotation for metabolite %s.',\n metabolite.id)\n return\n",
"role": "user"
},
{
"content": "map_metabolite2kegg",
"role": "assistant"
}
] |
def x(reaction, metabolite_mapping):
stoichiometry = defaultdict(float)
for met, coef in iteritems(reaction.metabolites):
kegg_id = metabolite_mapping.setdefault(met, map_metabolite2kegg(met))
if kegg_id is None:
continue
stoichiometry[kegg_id] += coef
return dict(stoichiometry)
| translate_reaction | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(reaction, metabolite_mapping):\n stoichiometry = defaultdict(float)\n for met, coef in iteritems(reaction.metabolites):\n kegg_id = metabolite_mapping.setdefault(met, map_metabolite2kegg(met))\n if kegg_id is None:\n continue\n stoichiometry[kegg_id] += coef\n return dict(stoichiometry)\n",
"role": "user"
},
{
"content": "translate_reaction",
"role": "assistant"
}
] |
def x(reactions):
incomplete_mapping = []
problematic_calculation = []
reversibility_indexes = []
unbalanced = []
metabolite_mapping = {}
for rxn in reactions:
stoich = translate_reaction(rxn, metabolite_mapping)
if len(stoich) < len(rxn.metabolites):
incomplete_mapping.append(rxn)
continue
try:
if 'C00080' in stoich:
del stoich['C00080']
eq_rxn = Reaction(stoich, rxn.id)
except KeyError:
incomplete_mapping.append(rxn)
continue
if eq_rxn.check_full_reaction_balancing():
try:
ln_rev_index = eq_rxn.reversibility_index()
except Exception:
problematic_calculation.append(rxn)
continue
reversibility_indexes.append((rxn, ln_rev_index))
else:
unbalanced.append(rxn)
reversibility_indexes.sort(key=lambda p: abs(p[1]), reverse=True)
return (reversibility_indexes, incomplete_mapping,
problematic_calculation, unbalanced)
| find_thermodynamic_reversibility_index | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(reactions):\n incomplete_mapping = []\n problematic_calculation = []\n reversibility_indexes = []\n unbalanced = []\n metabolite_mapping = {}\n for rxn in reactions:\n stoich = translate_reaction(rxn, metabolite_mapping)\n if len(stoich) < len(rxn.metabolites):\n incomplete_mapping.append(rxn)\n continue\n try:\n if 'C00080' in stoich:\n del stoich['C00080']\n eq_rxn = Reaction(stoich, rxn.id)\n except KeyError:\n incomplete_mapping.append(rxn)\n continue\n if eq_rxn.check_full_reaction_balancing():\n try:\n ln_rev_index = eq_rxn.reversibility_index()\n except Exception:\n problematic_calculation.append(rxn)\n continue\n reversibility_indexes.append((rxn, ln_rev_index))\n else:\n unbalanced.append(rxn)\n reversibility_indexes.sort(key=lambda p: abs(p[1]), reverse=True)\n return (reversibility_indexes, incomplete_mapping,\n problematic_calculation, unbalanced)\n",
"role": "user"
},
{
"content": "find_thermodynamic_reversibility_index",
"role": "assistant"
}
] |
def x(model):
problem = model.problem
stoich_trans = problem.Model()
internal_rxns = con_helpers.get_internals(model)
metabolites = set(met for rxn in internal_rxns for met in rxn.metabolites)
LOGGER.info("model '%s' has %d internal reactions", model.id, len(
internal_rxns))
LOGGER.info("model '%s' has %d internal metabolites", model.id, len(
metabolites))
stoich_trans.add([problem.Variable(m.id, lb=1) for m in metabolites])
stoich_trans.update()
con_helpers.add_reaction_constraints(stoich_trans, internal_rxns,
problem.Constraint)
stoich_trans.objective = problem.Objective(Zero, direction='min',
sloppy=True)
stoich_trans.objective.set_linear_coefficients({var: (1.0) for var in
stoich_trans.variables})
status = stoich_trans.optimize()
if status == OPTIMAL:
return True
elif status == INFEASIBLE:
return False
else:
raise RuntimeError(
"Could not determine stoichiometric consistencty. Solver status is '{}' (only optimal or infeasible expected)."
.format(status))
| check_stoichiometric_consistency | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n problem = model.problem\n stoich_trans = problem.Model()\n internal_rxns = con_helpers.get_internals(model)\n metabolites = set(met for rxn in internal_rxns for met in rxn.metabolites)\n LOGGER.info(\"model '%s' has %d internal reactions\", model.id, len(\n internal_rxns))\n LOGGER.info(\"model '%s' has %d internal metabolites\", model.id, len(\n metabolites))\n stoich_trans.add([problem.Variable(m.id, lb=1) for m in metabolites])\n stoich_trans.update()\n con_helpers.add_reaction_constraints(stoich_trans, internal_rxns,\n problem.Constraint)\n stoich_trans.objective = problem.Objective(Zero, direction='min',\n sloppy=True)\n stoich_trans.objective.set_linear_coefficients({var: (1.0) for var in\n stoich_trans.variables})\n status = stoich_trans.optimize()\n if status == OPTIMAL:\n return True\n elif status == INFEASIBLE:\n return False\n else:\n raise RuntimeError(\n \"Could not determine stoichiometric consistencty. Solver status is '{}' (only optimal or infeasible expected).\"\n .format(status))\n",
"role": "user"
},
{
"content": "check_stoichiometric_consistency",
"role": "assistant"
}
] |
def x(model):
problem = model.problem
stoich_trans = problem.Model()
internal_rxns = con_helpers.get_internals(model)
metabolites = set(met for rxn in internal_rxns for met in rxn.metabolites)
k_vars = list()
for met in metabolites:
m_var = problem.Variable(met.id)
k_var = problem.Variable('k_{}'.format(met.id), type='binary')
k_vars.append(k_var)
stoich_trans.add([m_var, k_var])
stoich_trans.add(problem.Constraint(k_var - m_var, ub=0, name=
'switch_{}'.format(met.id)))
stoich_trans.update()
con_helpers.add_reaction_constraints(stoich_trans, internal_rxns,
problem.Constraint)
stoich_trans.objective = problem.Objective(Zero, sloppy=True, direction
='max')
stoich_trans.objective.set_linear_coefficients({var: (1.0) for var in
k_vars})
status = stoich_trans.optimize()
if status == OPTIMAL:
return set([model.metabolites.get_by_id(var.name[2:]) for var in
k_vars if var.primal < 0.8])
else:
raise RuntimeError(
"Could not compute list of unconserved metabolites. Solver status is '{}' (only optimal expected)."
.format(status))
| find_unconserved_metabolites | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n problem = model.problem\n stoich_trans = problem.Model()\n internal_rxns = con_helpers.get_internals(model)\n metabolites = set(met for rxn in internal_rxns for met in rxn.metabolites)\n k_vars = list()\n for met in metabolites:\n m_var = problem.Variable(met.id)\n k_var = problem.Variable('k_{}'.format(met.id), type='binary')\n k_vars.append(k_var)\n stoich_trans.add([m_var, k_var])\n stoich_trans.add(problem.Constraint(k_var - m_var, ub=0, name=\n 'switch_{}'.format(met.id)))\n stoich_trans.update()\n con_helpers.add_reaction_constraints(stoich_trans, internal_rxns,\n problem.Constraint)\n stoich_trans.objective = problem.Objective(Zero, sloppy=True, direction\n ='max')\n stoich_trans.objective.set_linear_coefficients({var: (1.0) for var in\n k_vars})\n status = stoich_trans.optimize()\n if status == OPTIMAL:\n return set([model.metabolites.get_by_id(var.name[2:]) for var in\n k_vars if var.primal < 0.8])\n else:\n raise RuntimeError(\n \"Could not compute list of unconserved metabolites. Solver status is '{}' (only optimal expected).\"\n .format(status))\n",
"role": "user"
},
{
"content": "find_unconserved_metabolites",
"role": "assistant"
}
] |
def x(model, atol=1e-13):
if check_stoichiometric_consistency(model):
return set()
Model, Constraint, Variable, Objective = con_helpers.get_interface(model)
unconserved_mets = find_unconserved_metabolites(model)
LOGGER.info('model has %d unconserved metabolites', len(unconserved_mets))
internal_rxns = con_helpers.get_internals(model)
internal_mets = set(met for rxn in internal_rxns for met in rxn.metabolites
)
get_id = attrgetter('id')
reactions = sorted(internal_rxns, key=get_id)
metabolites = sorted(internal_mets, key=get_id)
stoich, met_index, rxn_index = con_helpers.stoichiometry_matrix(metabolites
, reactions)
left_ns = con_helpers.nullspace(stoich.T)
left_ns[np.abs(left_ns) < atol] = 0.0
LOGGER.info('nullspace has dimension %d', left_ns.shape[1])
inc_minimal = set()
problem, indicators = con_helpers.create_milp_problem(left_ns,
metabolites, Model, Variable, Constraint, Objective)
LOGGER.debug(str(problem))
cuts = list()
for met in unconserved_mets:
row = met_index[met]
if (left_ns[row] == 0.0).all():
LOGGER.debug('%s: singleton minimal unconservable set', met.id)
inc_minimal.add((met,))
continue
problem.variables[met.id].lb = 0.001
status = problem.optimize()
while status == 'optimal':
LOGGER.debug('%s: status %s', met.id, status)
LOGGER.debug('sum of all primal values: %f', sum(problem.
primal_values.values()))
LOGGER.debug('sum of binary indicators: %f', sum(var.primal for
var in indicators))
solution = [model.metabolites.get_by_id(var.name[2:]) for var in
indicators if var.primal > 0.2]
LOGGER.debug('%s: set size %d', met.id, len(solution))
inc_minimal.add(tuple(solution))
if len(solution) == 1:
break
cuts.append(con_helpers.add_cut(problem, indicators, len(
solution) - 1, Constraint))
status = problem.optimize()
LOGGER.debug('%s: last status %s', met.id, status)
problem.variables[met.id].lb = 0.0
problem.remove(cuts)
cuts.clear()
return inc_minimal
| find_inconsistent_min_stoichiometry | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model, atol=1e-13):\n if check_stoichiometric_consistency(model):\n return set()\n Model, Constraint, Variable, Objective = con_helpers.get_interface(model)\n unconserved_mets = find_unconserved_metabolites(model)\n LOGGER.info('model has %d unconserved metabolites', len(unconserved_mets))\n internal_rxns = con_helpers.get_internals(model)\n internal_mets = set(met for rxn in internal_rxns for met in rxn.metabolites\n )\n get_id = attrgetter('id')\n reactions = sorted(internal_rxns, key=get_id)\n metabolites = sorted(internal_mets, key=get_id)\n stoich, met_index, rxn_index = con_helpers.stoichiometry_matrix(metabolites\n , reactions)\n left_ns = con_helpers.nullspace(stoich.T)\n left_ns[np.abs(left_ns) < atol] = 0.0\n LOGGER.info('nullspace has dimension %d', left_ns.shape[1])\n inc_minimal = set()\n problem, indicators = con_helpers.create_milp_problem(left_ns,\n metabolites, Model, Variable, Constraint, Objective)\n LOGGER.debug(str(problem))\n cuts = list()\n for met in unconserved_mets:\n row = met_index[met]\n if (left_ns[row] == 0.0).all():\n LOGGER.debug('%s: singleton minimal unconservable set', met.id)\n inc_minimal.add((met,))\n continue\n problem.variables[met.id].lb = 0.001\n status = problem.optimize()\n while status == 'optimal':\n LOGGER.debug('%s: status %s', met.id, status)\n LOGGER.debug('sum of all primal values: %f', sum(problem.\n primal_values.values()))\n LOGGER.debug('sum of binary indicators: %f', sum(var.primal for\n var in indicators))\n solution = [model.metabolites.get_by_id(var.name[2:]) for var in\n indicators if var.primal > 0.2]\n LOGGER.debug('%s: set size %d', met.id, len(solution))\n inc_minimal.add(tuple(solution))\n if len(solution) == 1:\n break\n cuts.append(con_helpers.add_cut(problem, indicators, len(\n solution) - 1, Constraint))\n status = problem.optimize()\n LOGGER.debug('%s: last status %s', met.id, status)\n problem.variables[met.id].lb = 0.0\n problem.remove(cuts)\n cuts.clear()\n return inc_minimal\n",
"role": "user"
},
{
"content": "find_inconsistent_min_stoichiometry",
"role": "assistant"
}
] |
def x(model, metabolite_id):
main_comp = helpers.find_compartment_id_in_model(model, 'c')
met = helpers.find_met_in_model(model, metabolite_id, main_comp)[0]
dissipation_rxn = Reaction('Dissipation')
if metabolite_id in ['MNXM3', 'MNXM63', 'MNXM51', 'MNXM121', 'MNXM423']:
dissipation_rxn.add_metabolites({helpers.find_met_in_model(model,
'MNXM2', main_comp)[0]: -1, helpers.find_met_in_model(model,
'MNXM1', main_comp)[0]: 1, helpers.find_met_in_model(model,
'MNXM9', main_comp)[0]: 1})
elif metabolite_id in ['MNXM6', 'MNXM10']:
dissipation_rxn.add_metabolites({helpers.find_met_in_model(model,
'MNXM1', main_comp)[0]: 1})
elif metabolite_id in ['MNXM38', 'MNXM208', 'MNXM191', 'MNXM223',
'MNXM7517', 'MNXM12233', 'MNXM558']:
dissipation_rxn.add_metabolites({helpers.find_met_in_model(model,
'MNXM1', main_comp)[0]: 2})
elif metabolite_id == 'MNXM21':
dissipation_rxn.add_metabolites({helpers.find_met_in_model(model,
'MNXM2', main_comp)[0]: -1, helpers.find_met_in_model(model,
'MNXM1', main_comp)[0]: 1, helpers.find_met_in_model(model,
'MNXM26', main_comp)[0]: 1})
elif metabolite_id == 'MNXM89557':
dissipation_rxn.add_metabolites({helpers.find_met_in_model(model,
'MNXM2', main_comp)[0]: -1, helpers.find_met_in_model(model,
'MNXM1', main_comp)[0]: 2, helpers.find_met_in_model(model,
'MNXM15', main_comp)[0]: 1})
dissipation_product = helpers.find_met_in_model(model, ENERGY_COUPLES[
metabolite_id], main_comp)[0]
dissipation_rxn.add_metabolites({met: -1, dissipation_product: 1})
helpers.close_boundaries_sensibly(model)
model.add_reactions([dissipation_rxn])
model.objective = dissipation_rxn
solution = model.optimize(raise_error=True)
if solution.objective_value > 0.0:
return solution.fluxes[solution.fluxes.abs() > 0.0].index.drop([
'Dissipation']).tolist()
else:
return []
| detect_energy_generating_cycles | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model, metabolite_id):\n main_comp = helpers.find_compartment_id_in_model(model, 'c')\n met = helpers.find_met_in_model(model, metabolite_id, main_comp)[0]\n dissipation_rxn = Reaction('Dissipation')\n if metabolite_id in ['MNXM3', 'MNXM63', 'MNXM51', 'MNXM121', 'MNXM423']:\n dissipation_rxn.add_metabolites({helpers.find_met_in_model(model,\n 'MNXM2', main_comp)[0]: -1, helpers.find_met_in_model(model,\n 'MNXM1', main_comp)[0]: 1, helpers.find_met_in_model(model,\n 'MNXM9', main_comp)[0]: 1})\n elif metabolite_id in ['MNXM6', 'MNXM10']:\n dissipation_rxn.add_metabolites({helpers.find_met_in_model(model,\n 'MNXM1', main_comp)[0]: 1})\n elif metabolite_id in ['MNXM38', 'MNXM208', 'MNXM191', 'MNXM223',\n 'MNXM7517', 'MNXM12233', 'MNXM558']:\n dissipation_rxn.add_metabolites({helpers.find_met_in_model(model,\n 'MNXM1', main_comp)[0]: 2})\n elif metabolite_id == 'MNXM21':\n dissipation_rxn.add_metabolites({helpers.find_met_in_model(model,\n 'MNXM2', main_comp)[0]: -1, helpers.find_met_in_model(model,\n 'MNXM1', main_comp)[0]: 1, helpers.find_met_in_model(model,\n 'MNXM26', main_comp)[0]: 1})\n elif metabolite_id == 'MNXM89557':\n dissipation_rxn.add_metabolites({helpers.find_met_in_model(model,\n 'MNXM2', main_comp)[0]: -1, helpers.find_met_in_model(model,\n 'MNXM1', main_comp)[0]: 2, helpers.find_met_in_model(model,\n 'MNXM15', main_comp)[0]: 1})\n dissipation_product = helpers.find_met_in_model(model, ENERGY_COUPLES[\n metabolite_id], main_comp)[0]\n dissipation_rxn.add_metabolites({met: -1, dissipation_product: 1})\n helpers.close_boundaries_sensibly(model)\n model.add_reactions([dissipation_rxn])\n model.objective = dissipation_rxn\n solution = model.optimize(raise_error=True)\n if solution.objective_value > 0.0:\n return solution.fluxes[solution.fluxes.abs() > 0.0].index.drop([\n 'Dissipation']).tolist()\n else:\n return []\n",
"role": "user"
},
{
"content": "detect_energy_generating_cycles",
"role": "assistant"
}
] |
def x(model):
helpers.close_boundaries_sensibly(model)
fva_result = flux_variability_analysis(model, loopless=False)
return fva_result.index[(fva_result['minimum'] <= -1 +
TOLERANCE_THRESHOLD) | (fva_result['maximum'] >= 1 -
TOLERANCE_THRESHOLD)].tolist()
| find_stoichiometrically_balanced_cycles | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n helpers.close_boundaries_sensibly(model)\n fva_result = flux_variability_analysis(model, loopless=False)\n return fva_result.index[(fva_result['minimum'] <= -1 +\n TOLERANCE_THRESHOLD) | (fva_result['maximum'] >= 1 -\n TOLERANCE_THRESHOLD)].tolist()\n",
"role": "user"
},
{
"content": "find_stoichiometrically_balanced_cycles",
"role": "assistant"
}
] |
def x(model):
exchange = frozenset(model.exchanges)
return [met for met in model.metabolites if len(met.reactions) > 0 and
all(not rxn.reversibility and rxn not in exchange and rxn.
metabolites[met] < 0 for rxn in met.reactions)]
| find_orphans | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n exchange = frozenset(model.exchanges)\n return [met for met in model.metabolites if len(met.reactions) > 0 and\n all(not rxn.reversibility and rxn not in exchange and rxn.\n metabolites[met] < 0 for rxn in met.reactions)]\n",
"role": "user"
},
{
"content": "find_orphans",
"role": "assistant"
}
] |
def x(model):
mets_not_produced = list()
helpers.open_exchanges(model)
for met in model.metabolites:
with model:
exch = model.add_boundary(met, type='irrex', reaction_id=
'IRREX', lb=0, ub=1000)
solution = helpers.run_fba(model, exch.id)
if np.isnan(solution) or solution < TOLERANCE_THRESHOLD:
mets_not_produced.append(met)
return mets_not_produced
| find_metabolites_not_produced_with_open_bounds | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n mets_not_produced = list()\n helpers.open_exchanges(model)\n for met in model.metabolites:\n with model:\n exch = model.add_boundary(met, type='irrex', reaction_id=\n 'IRREX', lb=0, ub=1000)\n solution = helpers.run_fba(model, exch.id)\n if np.isnan(solution) or solution < TOLERANCE_THRESHOLD:\n mets_not_produced.append(met)\n return mets_not_produced\n",
"role": "user"
},
{
"content": "find_metabolites_not_produced_with_open_bounds",
"role": "assistant"
}
] |
def x(model):
mets_not_consumed = list()
helpers.open_exchanges(model)
for met in model.metabolites:
with model:
exch = model.add_boundary(met, type='irrex', reaction_id=
'IRREX', lb=-1000, ub=0)
solution = helpers.run_fba(model, exch.id, direction='min')
if np.isnan(solution) or abs(solution) < TOLERANCE_THRESHOLD:
mets_not_consumed.append(met)
return mets_not_consumed
| find_metabolites_not_consumed_with_open_bounds | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n mets_not_consumed = list()\n helpers.open_exchanges(model)\n for met in model.metabolites:\n with model:\n exch = model.add_boundary(met, type='irrex', reaction_id=\n 'IRREX', lb=-1000, ub=0)\n solution = helpers.run_fba(model, exch.id, direction='min')\n if np.isnan(solution) or abs(solution) < TOLERANCE_THRESHOLD:\n mets_not_consumed.append(met)\n return mets_not_consumed\n",
"role": "user"
},
{
"content": "find_metabolites_not_consumed_with_open_bounds",
"role": "assistant"
}
] |
def x(model):
try:
fva_result = flux_variability_analysis(model, fraction_of_optimum=1.0)
except Infeasible as err:
LOGGER.error(
"Failed to find reactions with unbounded flux because '{}'. This may be a bug."
.format(err))
raise Infeasible(
'It was not possible to run flux variability analysis on the model. Make sure that the model can be solved! Check if the constraints are not too strict.'
)
conditionally_blocked = fva_result.loc[fva_result.abs().max(axis=1) <
TOLERANCE_THRESHOLD].index.tolist()
small, large = helpers.find_bounds(model)
unlimited_flux = fva_result.loc[np.isclose(fva_result['maximum'], large,
atol=TOLERANCE_THRESHOLD) | (fva_result['maximum'] > large) | np.
isclose(fva_result['minimum'], small, atol=TOLERANCE_THRESHOLD) | (
fva_result['minimum'] < small)].index.tolist()
try:
fraction = len(unlimited_flux) / (len(model.reactions) - len(
conditionally_blocked))
except ZeroDivisionError:
LOGGER.error(
'Division by Zero! Failed to calculate the fraction of unbounded reactions. Does this model have any reactions at all?'
)
raise ZeroDivisionError(
"It was not possible to calculate the fraction of unbounded reactions to un-blocked reactions. This may be becausethe model doesn't have any reactions at all or that none of the reactions can carry a flux larger than zero!"
)
return unlimited_flux, fraction, conditionally_blocked
| find_reactions_with_unbounded_flux_default_condition | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n try:\n fva_result = flux_variability_analysis(model, fraction_of_optimum=1.0)\n except Infeasible as err:\n LOGGER.error(\n \"Failed to find reactions with unbounded flux because '{}'. This may be a bug.\"\n .format(err))\n raise Infeasible(\n 'It was not possible to run flux variability analysis on the model. Make sure that the model can be solved! Check if the constraints are not too strict.'\n )\n conditionally_blocked = fva_result.loc[fva_result.abs().max(axis=1) <\n TOLERANCE_THRESHOLD].index.tolist()\n small, large = helpers.find_bounds(model)\n unlimited_flux = fva_result.loc[np.isclose(fva_result['maximum'], large,\n atol=TOLERANCE_THRESHOLD) | (fva_result['maximum'] > large) | np.\n isclose(fva_result['minimum'], small, atol=TOLERANCE_THRESHOLD) | (\n fva_result['minimum'] < small)].index.tolist()\n try:\n fraction = len(unlimited_flux) / (len(model.reactions) - len(\n conditionally_blocked))\n except ZeroDivisionError:\n LOGGER.error(\n 'Division by Zero! Failed to calculate the fraction of unbounded reactions. Does this model have any reactions at all?'\n )\n raise ZeroDivisionError(\n \"It was not possible to calculate the fraction of unbounded reactions to un-blocked reactions. This may be becausethe model doesn't have any reactions at all or that none of the reactions can carry a flux larger than zero!\"\n )\n return unlimited_flux, fraction, conditionally_blocked\n",
"role": "user"
},
{
"content": "find_reactions_with_unbounded_flux_default_condition",
"role": "assistant"
}
] |
def x(filename, dtype_conversion=None):
if dtype_conversion is None:
dtype_conversion = {}
name, ext = filename.split('.', 1)
ext = ext.lower()
dtype_conversion['comment'] = str
if 'csv' in ext:
df = pd.read_csv(filename, dtype=dtype_conversion, encoding='utf-8')
elif 'tsv' in ext:
df = pd.read_table(filename, dtype=dtype_conversion, encoding='utf-8')
elif 'xls' in ext or 'xlsx' in ext:
df = pd.read_excel(filename, dtype=dtype_conversion, encoding='utf-8')
else:
raise ValueError("Unknown file format '{}'.".format(ext))
return df
| read_tabular | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename, dtype_conversion=None):\n if dtype_conversion is None:\n dtype_conversion = {}\n name, ext = filename.split('.', 1)\n ext = ext.lower()\n dtype_conversion['comment'] = str\n if 'csv' in ext:\n df = pd.read_csv(filename, dtype=dtype_conversion, encoding='utf-8')\n elif 'tsv' in ext:\n df = pd.read_table(filename, dtype=dtype_conversion, encoding='utf-8')\n elif 'xls' in ext or 'xlsx' in ext:\n df = pd.read_excel(filename, dtype=dtype_conversion, encoding='utf-8')\n else:\n raise ValueError(\"Unknown file format '{}'.\".format(ext))\n return df\n",
"role": "user"
},
{
"content": "read_tabular",
"role": "assistant"
}
] |
def x(model, filename, pytest_args, exclusive, skip, solver, experimental,
custom_tests, custom_config):
model_obj, sbml_ver, notifications = api.validate_model(model)
if model_obj is None:
LOGGER.critical(
'The model could not be loaded due to the following SBML errors.')
utils.stdout_notifications(notifications)
api.validation_report(model, notifications, filename)
sys.exit(1)
if not any(a.startswith('--tb') for a in pytest_args):
pytest_args = ['--tb', 'no'] + pytest_args
pytest_args.extend(custom_tests)
config = ReportConfiguration.load()
for custom in custom_config:
config.merge(ReportConfiguration.load(custom))
model_obj.solver = solver
_, results = api.test_model(model_obj, sbml_version=sbml_ver, results=
True, pytest_args=pytest_args, skip=skip, exclusive=exclusive,
experimental=experimental)
with open(filename, 'w', encoding='utf-8') as file_handle:
LOGGER.info("Writing snapshot report to '%s'.", filename)
file_handle.write(api.snapshot_report(results, config))
| snapshot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model, filename, pytest_args, exclusive, skip, solver, experimental,\n custom_tests, custom_config):\n model_obj, sbml_ver, notifications = api.validate_model(model)\n if model_obj is None:\n LOGGER.critical(\n 'The model could not be loaded due to the following SBML errors.')\n utils.stdout_notifications(notifications)\n api.validation_report(model, notifications, filename)\n sys.exit(1)\n if not any(a.startswith('--tb') for a in pytest_args):\n pytest_args = ['--tb', 'no'] + pytest_args\n pytest_args.extend(custom_tests)\n config = ReportConfiguration.load()\n for custom in custom_config:\n config.merge(ReportConfiguration.load(custom))\n model_obj.solver = solver\n _, results = api.test_model(model_obj, sbml_version=sbml_ver, results=\n True, pytest_args=pytest_args, skip=skip, exclusive=exclusive,\n experimental=experimental)\n with open(filename, 'w', encoding='utf-8') as file_handle:\n LOGGER.info(\"Writing snapshot report to '%s'.\", filename)\n file_handle.write(api.snapshot_report(results, config))\n",
"role": "user"
},
{
"content": "snapshot",
"role": "assistant"
}
] |
def x(location, model, filename, deployment, custom_config):
callbacks.git_installed()
LOGGER.info('Initialising history report generation.')
if location is None:
raise click.BadParameter("No 'location' given or configured.")
try:
repo = git.Repo()
except git.InvalidGitRepositoryError:
LOGGER.critical(
"The history report requires a git repository in order to check the model's commit history."
)
sys.exit(1)
LOGGER.info('Obtaining history of results from the deployment branch {}.'
.format(deployment))
repo.git.checkout(deployment)
try:
manager = managers.SQLResultManager(repository=repo, location=location)
except (AttributeError, ArgumentError):
manager = managers.RepoResultManager(repository=repo, location=location
)
config = ReportConfiguration.load()
for custom in custom_config:
config.merge(ReportConfiguration.load(custom))
LOGGER.info('Tracing the commit history.')
history = managers.HistoryManager(repository=repo, manager=manager)
history.load_history(model, skip={deployment})
LOGGER.info('Composing the history report.')
report = api.history_report(history, config=config)
with open(filename, 'w', encoding='utf-8') as file_handle:
file_handle.write(report)
| history | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(location, model, filename, deployment, custom_config):\n callbacks.git_installed()\n LOGGER.info('Initialising history report generation.')\n if location is None:\n raise click.BadParameter(\"No 'location' given or configured.\")\n try:\n repo = git.Repo()\n except git.InvalidGitRepositoryError:\n LOGGER.critical(\n \"The history report requires a git repository in order to check the model's commit history.\"\n )\n sys.exit(1)\n LOGGER.info('Obtaining history of results from the deployment branch {}.'\n .format(deployment))\n repo.git.checkout(deployment)\n try:\n manager = managers.SQLResultManager(repository=repo, location=location)\n except (AttributeError, ArgumentError):\n manager = managers.RepoResultManager(repository=repo, location=location\n )\n config = ReportConfiguration.load()\n for custom in custom_config:\n config.merge(ReportConfiguration.load(custom))\n LOGGER.info('Tracing the commit history.')\n history = managers.HistoryManager(repository=repo, manager=manager)\n history.load_history(model, skip={deployment})\n LOGGER.info('Composing the history report.')\n report = api.history_report(history, config=config)\n with open(filename, 'w', encoding='utf-8') as file_handle:\n file_handle.write(report)\n",
"role": "user"
},
{
"content": "history",
"role": "assistant"
}
] |
def x(models, filename, pytest_args, exclusive, skip, solver, experimental,
custom_tests, custom_config):
if not any(a.startswith('--tb') for a in pytest_args):
pytest_args = ['--tb', 'no'] + pytest_args
pytest_args.extend(custom_tests)
config = ReportConfiguration.load()
for custom in custom_config:
config.merge(ReportConfiguration.load(custom))
diff_results = dict()
model_and_model_ver_tuple = list()
for model_path in models:
try:
model_filename = os.path.basename(model_path)
diff_results.setdefault(model_filename, dict())
model, model_ver, notifications = api.validate_model(model_path)
if model is None:
head, tail = os.path.split(filename)
report_path = os.path.join(head,
'{}_structural_report.html'.format(model_filename))
api.validation_report(model_path, notifications, report_path)
LOGGER.critical(
'The model {} could not be loaded due to SBML errors reported in {}.'
.format(model_filename, report_path))
continue
model.solver = solver
model_and_model_ver_tuple.append((model, model_ver))
except (IOError, SBMLError):
LOGGER.debug(exc_info=True)
LOGGER.warning(
"An error occurred while loading the model '%s'. Skipping.",
model_filename)
if len(model_and_model_ver_tuple) < 2:
LOGGER.critical(
'Out of the %d provided models only %d could be loaded. Please, check if the models that could not be loaded are valid SBML. Aborting.'
, len(models), len(model_and_model_ver_tuple))
sys.exit(1)
partial_test_diff = partial(_test_diff, pytest_args=pytest_args, skip=
skip, exclusive=exclusive, experimental=experimental)
pool = Pool(min(len(models), cpu_count()))
results = pool.map(partial_test_diff, model_and_model_ver_tuple)
for model_path, result in zip(models, results):
model_filename = os.path.basename(model_path)
diff_results[model_filename] = result
with open(filename, 'w', encoding='utf-8') as file_handle:
LOGGER.info("Writing diff report to '%s'.", filename)
file_handle.write(api.diff_report(diff_results, config))
| diff | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(models, filename, pytest_args, exclusive, skip, solver, experimental,\n custom_tests, custom_config):\n if not any(a.startswith('--tb') for a in pytest_args):\n pytest_args = ['--tb', 'no'] + pytest_args\n pytest_args.extend(custom_tests)\n config = ReportConfiguration.load()\n for custom in custom_config:\n config.merge(ReportConfiguration.load(custom))\n diff_results = dict()\n model_and_model_ver_tuple = list()\n for model_path in models:\n try:\n model_filename = os.path.basename(model_path)\n diff_results.setdefault(model_filename, dict())\n model, model_ver, notifications = api.validate_model(model_path)\n if model is None:\n head, tail = os.path.split(filename)\n report_path = os.path.join(head,\n '{}_structural_report.html'.format(model_filename))\n api.validation_report(model_path, notifications, report_path)\n LOGGER.critical(\n 'The model {} could not be loaded due to SBML errors reported in {}.'\n .format(model_filename, report_path))\n continue\n model.solver = solver\n model_and_model_ver_tuple.append((model, model_ver))\n except (IOError, SBMLError):\n LOGGER.debug(exc_info=True)\n LOGGER.warning(\n \"An error occurred while loading the model '%s'. Skipping.\",\n model_filename)\n if len(model_and_model_ver_tuple) < 2:\n LOGGER.critical(\n 'Out of the %d provided models only %d could be loaded. Please, check if the models that could not be loaded are valid SBML. Aborting.'\n , len(models), len(model_and_model_ver_tuple))\n sys.exit(1)\n partial_test_diff = partial(_test_diff, pytest_args=pytest_args, skip=\n skip, exclusive=exclusive, experimental=experimental)\n pool = Pool(min(len(models), cpu_count()))\n results = pool.map(partial_test_diff, model_and_model_ver_tuple)\n for model_path, result in zip(models, results):\n model_filename = os.path.basename(model_path)\n diff_results[model_filename] = result\n with open(filename, 'w', encoding='utf-8') as file_handle:\n LOGGER.info(\"Writing diff report to '%s'.\", filename)\n file_handle.write(api.diff_report(diff_results, config))\n",
"role": "user"
},
{
"content": "diff",
"role": "assistant"
}
] |
def x(self, model, skip):
self._history = dict()
self._history['commits'] = commits = dict()
self._history['branches'] = branches = dict()
for branch in self._repo.refs:
LOGGER.debug(branch.name)
if branch.name in skip:
continue
branches[branch.name] = branch_history = list()
latest = branch.commit
history = [latest] + list(latest.iter_parents())
for commit in history:
if not is_modified(model, commit):
LOGGER.info(
"The model was not modified in commit '{}'. Skipping.".
format(commit))
continue
branch_history.append(commit.hexsha)
if commit.hexsha not in commits:
commits[commit.hexsha] = sub = dict()
sub['timestamp'] = commit.authored_datetime.isoformat(' ')
sub['author'] = commit.author.name
sub['email'] = commit.author.email
LOGGER.debug('%s', json.dumps(self._history, indent=2))
| build_branch_structure | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, model, skip):\n self._history = dict()\n self._history['commits'] = commits = dict()\n self._history['branches'] = branches = dict()\n for branch in self._repo.refs:\n LOGGER.debug(branch.name)\n if branch.name in skip:\n continue\n branches[branch.name] = branch_history = list()\n latest = branch.commit\n history = [latest] + list(latest.iter_parents())\n for commit in history:\n if not is_modified(model, commit):\n LOGGER.info(\n \"The model was not modified in commit '{}'. Skipping.\".\n format(commit))\n continue\n branch_history.append(commit.hexsha)\n if commit.hexsha not in commits:\n commits[commit.hexsha] = sub = dict()\n sub['timestamp'] = commit.authored_datetime.isoformat(' ')\n sub['author'] = commit.author.name\n sub['email'] = commit.author.email\n LOGGER.debug('%s', json.dumps(self._history, indent=2))\n",
"role": "user"
},
{
"content": "build_branch_structure",
"role": "assistant"
}
] |
def x(self, model, skip={'gh-pages'}):
if self._history is None:
self.build_branch_structure(model, skip)
self._results = dict()
all_commits = list(self._history['commits'])
for commit in all_commits:
try:
self._results[commit] = self.manager.load(commit)
except (IOError, NoResultFound) as err:
LOGGER.error("Could not load result '%s'.", commit)
LOGGER.debug('%s', str(err))
| load_history | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, model, skip={'gh-pages'}):\n if self._history is None:\n self.build_branch_structure(model, skip)\n self._results = dict()\n all_commits = list(self._history['commits'])\n for commit in all_commits:\n try:\n self._results[commit] = self.manager.load(commit)\n except (IOError, NoResultFound) as err:\n LOGGER.error(\"Could not load result '%s'.\", commit)\n LOGGER.debug('%s', str(err))\n",
"role": "user"
},
{
"content": "load_history",
"role": "assistant"
}
] |
def x(self, commit, default=MemoteResult()):
assert self._results is not None, 'Please call the method `load_history` first.'
return self._results.get(commit, default)
| get_result | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, commit, default=MemoteResult()):\n assert self._results is not None, 'Please call the method `load_history` first.'\n return self._results.get(commit, default)\n",
"role": "user"
},
{
"content": "get_result",
"role": "assistant"
}
] |
def x(model):
s_matrix, _, _ = con_helpers.stoichiometry_matrix(model.metabolites,
model.reactions)
abs_matrix = np.abs(s_matrix)
return abs_matrix.max(), abs_matrix[abs_matrix > 0].min()
| absolute_extreme_coefficient_ratio | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n s_matrix, _, _ = con_helpers.stoichiometry_matrix(model.metabolites,\n model.reactions)\n abs_matrix = np.abs(s_matrix)\n return abs_matrix.max(), abs_matrix[abs_matrix > 0].min()\n",
"role": "user"
},
{
"content": "absolute_extreme_coefficient_ratio",
"role": "assistant"
}
] |
def x(model):
s_matrix, _, _ = con_helpers.stoichiometry_matrix(model.metabolites,
model.reactions)
ln_matrix = con_helpers.nullspace(s_matrix.T)
return ln_matrix.shape[1]
| number_independent_conservation_relations | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n s_matrix, _, _ = con_helpers.stoichiometry_matrix(model.metabolites,\n model.reactions)\n ln_matrix = con_helpers.nullspace(s_matrix.T)\n return ln_matrix.shape[1]\n",
"role": "user"
},
{
"content": "number_independent_conservation_relations",
"role": "assistant"
}
] |
def x(model):
s_matrix, _, _ = con_helpers.stoichiometry_matrix(model.metabolites,
model.reactions)
return con_helpers.rank(s_matrix)
| matrix_rank | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n s_matrix, _, _ = con_helpers.stoichiometry_matrix(model.metabolites,\n model.reactions)\n return con_helpers.rank(s_matrix)\n",
"role": "user"
},
{
"content": "matrix_rank",
"role": "assistant"
}
] |
def x(model):
s_matrix, _, _ = con_helpers.stoichiometry_matrix(model.metabolites,
model.reactions)
return s_matrix.shape[1] - matrix_rank(model)
| degrees_of_freedom | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n s_matrix, _, _ = con_helpers.stoichiometry_matrix(model.metabolites,\n model.reactions)\n return s_matrix.shape[1] - matrix_rank(model)\n",
"role": "user"
},
{
"content": "degrees_of_freedom",
"role": "assistant"
}
] |
def x(self, model):
self.load_medium(model)
self.load_essentiality(model)
self.load_growth(model)
return self
| load | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, model):\n self.load_medium(model)\n self.load_essentiality(model)\n self.load_growth(model)\n return self\n",
"role": "user"
},
{
"content": "load",
"role": "assistant"
}
] |
def x(self):
validator = Draft4Validator(self.SCHEMA)
if not validator.is_valid(self.config):
for err in validator.iter_errors(self.config):
LOGGER.error(str(err.message))
validator.validate(self.config)
| validate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n validator = Draft4Validator(self.SCHEMA)\n if not validator.is_valid(self.config):\n for err in validator.iter_errors(self.config):\n LOGGER.error(str(err.message))\n validator.validate(self.config)\n",
"role": "user"
},
{
"content": "validate",
"role": "assistant"
}
] |
def x(self, model):
media = self.config.get('medium')
if media is None:
return
definitions = media.get('definitions')
if definitions is None or len(definitions) == 0:
return
path = self.get_path(media, join('data', 'experimental', 'media'))
for medium_id, medium in iteritems(definitions):
if medium is None:
medium = dict()
filename = medium.get('filename')
if filename is None:
filename = join(path, '{}.csv'.format(medium_id))
elif not isabs(filename):
filename = join(path, filename)
tmp = Medium(identifier=medium_id, obj=medium, filename=filename)
tmp.load()
tmp.validate(model)
self.media[medium_id] = tmp
| load_medium | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, model):\n media = self.config.get('medium')\n if media is None:\n return\n definitions = media.get('definitions')\n if definitions is None or len(definitions) == 0:\n return\n path = self.get_path(media, join('data', 'experimental', 'media'))\n for medium_id, medium in iteritems(definitions):\n if medium is None:\n medium = dict()\n filename = medium.get('filename')\n if filename is None:\n filename = join(path, '{}.csv'.format(medium_id))\n elif not isabs(filename):\n filename = join(path, filename)\n tmp = Medium(identifier=medium_id, obj=medium, filename=filename)\n tmp.load()\n tmp.validate(model)\n self.media[medium_id] = tmp\n",
"role": "user"
},
{
"content": "load_medium",
"role": "assistant"
}
] |
def x(self, model):
data = self.config.get('essentiality')
if data is None:
return
experiments = data.get('experiments')
if experiments is None or len(experiments) == 0:
return
path = self.get_path(data, join('data', 'experimental', 'essentiality'))
for exp_id, exp in iteritems(experiments):
if exp is None:
exp = dict()
filename = exp.get('filename')
if filename is None:
filename = join(path, '{}.csv'.format(exp_id))
elif not isabs(filename):
filename = join(path, filename)
experiment = EssentialityExperiment(identifier=exp_id, obj=exp,
filename=filename)
if experiment.medium is not None:
assert experiment.medium in self.media, "Experiment '{}' has an undefined medium '{}'.".format(
exp_id, experiment.medium)
experiment.medium = self.media[experiment.medium]
experiment.load()
experiment.validate(model)
self.essentiality[exp_id] = experiment
| load_essentiality | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, model):\n data = self.config.get('essentiality')\n if data is None:\n return\n experiments = data.get('experiments')\n if experiments is None or len(experiments) == 0:\n return\n path = self.get_path(data, join('data', 'experimental', 'essentiality'))\n for exp_id, exp in iteritems(experiments):\n if exp is None:\n exp = dict()\n filename = exp.get('filename')\n if filename is None:\n filename = join(path, '{}.csv'.format(exp_id))\n elif not isabs(filename):\n filename = join(path, filename)\n experiment = EssentialityExperiment(identifier=exp_id, obj=exp,\n filename=filename)\n if experiment.medium is not None:\n assert experiment.medium in self.media, \"Experiment '{}' has an undefined medium '{}'.\".format(\n exp_id, experiment.medium)\n experiment.medium = self.media[experiment.medium]\n experiment.load()\n experiment.validate(model)\n self.essentiality[exp_id] = experiment\n",
"role": "user"
},
{
"content": "load_essentiality",
"role": "assistant"
}
] |
def x(self, model):
data = self.config.get('growth')
if data is None:
return
experiments = data.get('experiments')
if experiments is None or len(experiments) == 0:
return
path = self.get_path(data, join('data', 'experimental', 'growth'))
for exp_id, exp in iteritems(experiments):
if exp is None:
exp = dict()
filename = exp.get('filename')
if filename is None:
filename = join(path, '{}.csv'.format(exp_id))
elif not isabs(filename):
filename = join(path, filename)
growth = GrowthExperiment(identifier=exp_id, obj=exp, filename=filename
)
if growth.medium is not None:
assert growth.medium in self.media, "Growth-experiment '{}' has an undefined medium '{}'.".format(
exp_id, growth.medium)
growth.medium = self.media[growth.medium]
growth.load()
growth.validate(model)
self.growth[exp_id] = growth
| load_growth | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, model):\n data = self.config.get('growth')\n if data is None:\n return\n experiments = data.get('experiments')\n if experiments is None or len(experiments) == 0:\n return\n path = self.get_path(data, join('data', 'experimental', 'growth'))\n for exp_id, exp in iteritems(experiments):\n if exp is None:\n exp = dict()\n filename = exp.get('filename')\n if filename is None:\n filename = join(path, '{}.csv'.format(exp_id))\n elif not isabs(filename):\n filename = join(path, filename)\n growth = GrowthExperiment(identifier=exp_id, obj=exp, filename=filename\n )\n if growth.medium is not None:\n assert growth.medium in self.media, \"Growth-experiment '{}' has an undefined medium '{}'.\".format(\n exp_id, growth.medium)\n growth.medium = self.media[growth.medium]\n growth.load()\n growth.validate(model)\n self.growth[exp_id] = growth\n",
"role": "user"
},
{
"content": "load_growth",
"role": "assistant"
}
] |
def x(self, obj, default):
path = obj.get('path')
if path is None:
path = join(self._base, default)
if not isabs(path):
path = join(self._base, path)
return path
| get_path | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, obj, default):\n path = obj.get('path')\n if path is None:\n path = join(self._base, default)\n if not isabs(path):\n path = join(self._base, path)\n return path\n",
"role": "user"
},
{
"content": "get_path",
"role": "assistant"
}
] |
def x(model, components):
return [elem for elem in getattr(model, components) if elem.annotation is
None or len(elem.annotation) == 0]
| find_components_without_annotation | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model, components):\n return [elem for elem in getattr(model, components) if elem.annotation is\n None or len(elem.annotation) == 0]\n",
"role": "user"
},
{
"content": "find_components_without_annotation",
"role": "assistant"
}
] |
def x(elements, component, db):
def x(annotation, key, pattern):
if key not in annotation:
return False
test = annotation[key]
if isinstance(test, native_str):
return pattern.match(test) is None
else:
return any(pattern.match(elem) is None for elem in test)
pattern = {'metabolites': METABOLITE_ANNOTATIONS, 'reactions':
REACTION_ANNOTATIONS, 'genes': GENE_PRODUCT_ANNOTATIONS}[component][db]
return [elem for elem in elements if is_faulty(elem.annotation, db,
pattern)]
| generate_component_annotation_miriam_match | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(elements, component, db):\n\n def x(annotation, key, pattern):\n if key not in annotation:\n return False\n test = annotation[key]\n if isinstance(test, native_str):\n return pattern.match(test) is None\n else:\n return any(pattern.match(elem) is None for elem in test)\n pattern = {'metabolites': METABOLITE_ANNOTATIONS, 'reactions':\n REACTION_ANNOTATIONS, 'genes': GENE_PRODUCT_ANNOTATIONS}[component][db]\n return [elem for elem in elements if is_faulty(elem.annotation, db,\n pattern)]\n",
"role": "user"
},
{
"content": "generate_component_annotation_miriam_match",
"role": "assistant"
}
] |
def x(model, components):
patterns = {'metabolites': METABOLITE_ANNOTATIONS, 'reactions':
REACTION_ANNOTATIONS, 'genes': GENE_PRODUCT_ANNOTATIONS}[components]
databases = list(patterns)
data = list()
index = list()
for elem in getattr(model, components):
index.append(elem.id)
data.append(tuple(patterns[db].match(elem.id) is not None for db in
databases))
df = pd.DataFrame(data, index=index, columns=databases)
if components != 'genes':
df['duplicate'] = df[df['biocyc']].sum(axis=1) >= 2
df['duplicate'].fillna(False, inplace=True)
df.loc[df['duplicate'], 'biocyc'] = False
del df['duplicate']
return df
| generate_component_id_namespace_overview | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model, components):\n patterns = {'metabolites': METABOLITE_ANNOTATIONS, 'reactions':\n REACTION_ANNOTATIONS, 'genes': GENE_PRODUCT_ANNOTATIONS}[components]\n databases = list(patterns)\n data = list()\n index = list()\n for elem in getattr(model, components):\n index.append(elem.id)\n data.append(tuple(patterns[db].match(elem.id) is not None for db in\n databases))\n df = pd.DataFrame(data, index=index, columns=databases)\n if components != 'genes':\n df['duplicate'] = df[df['biocyc']].sum(axis=1) >= 2\n df['duplicate'].fillna(False, inplace=True)\n df.loc[df['duplicate'], 'biocyc'] = False\n del df['duplicate']\n return df\n",
"role": "user"
},
{
"content": "generate_component_id_namespace_overview",
"role": "assistant"
}
] |
def x(predicted_essential, expected_essential, predicted_nonessential,
expected_nonessential):
true_positive = predicted_essential & expected_essential
tp = len(true_positive)
true_negative = predicted_nonessential & expected_nonessential
tn = len(true_negative)
false_positive = predicted_essential - expected_essential
fp = len(false_positive)
false_negative = predicted_nonessential - expected_nonessential
fn = len(false_negative)
try:
tpr = tp / (tp + fn)
except ZeroDivisionError:
tpr = None
try:
tnr = tn / (tn + fp)
except ZeroDivisionError:
tnr = None
try:
ppv = tp / (tp + fp)
except ZeroDivisionError:
ppv = None
fdr = 1 - ppv
try:
acc = (tp + tn) / (tp + tn + fp + fn)
except ZeroDivisionError:
acc = None
try:
mcc = (tp * tn - fp * fn) / sqrt((tp + fp) * (tp + fn) * (tn + fp) *
(tn + fn))
except ZeroDivisionError:
mcc = None
return {'TP': list(true_positive), 'TN': list(true_negative), 'FP':
list(false_positive), 'FN': list(false_negative), 'TPR': tpr, 'TNR':
tnr, 'PPV': ppv, 'FDR': fdr, 'ACC': acc, 'MCC': mcc}
| confusion_matrix | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(predicted_essential, expected_essential, predicted_nonessential,\n expected_nonessential):\n true_positive = predicted_essential & expected_essential\n tp = len(true_positive)\n true_negative = predicted_nonessential & expected_nonessential\n tn = len(true_negative)\n false_positive = predicted_essential - expected_essential\n fp = len(false_positive)\n false_negative = predicted_nonessential - expected_nonessential\n fn = len(false_negative)\n try:\n tpr = tp / (tp + fn)\n except ZeroDivisionError:\n tpr = None\n try:\n tnr = tn / (tn + fp)\n except ZeroDivisionError:\n tnr = None\n try:\n ppv = tp / (tp + fp)\n except ZeroDivisionError:\n ppv = None\n fdr = 1 - ppv\n try:\n acc = (tp + tn) / (tp + tn + fp + fn)\n except ZeroDivisionError:\n acc = None\n try:\n mcc = (tp * tn - fp * fn) / sqrt((tp + fp) * (tp + fn) * (tn + fp) *\n (tn + fn))\n except ZeroDivisionError:\n mcc = None\n return {'TP': list(true_positive), 'TN': list(true_negative), 'FP':\n list(false_positive), 'FN': list(false_negative), 'TPR': tpr, 'TNR':\n tnr, 'PPV': ppv, 'FDR': fdr, 'ACC': acc, 'MCC': mcc}\n",
"role": "user"
},
{
"content": "confusion_matrix",
"role": "assistant"
}
] |
def x(path):
notifications = {'warnings': [], 'errors': []}
model, sbml_ver = val.load_cobra_model(path, notifications)
return model, sbml_ver, notifications
| validate_model | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path):\n notifications = {'warnings': [], 'errors': []}\n model, sbml_ver = val.load_cobra_model(path, notifications)\n return model, sbml_ver, notifications\n",
"role": "user"
},
{
"content": "validate_model",
"role": "assistant"
}
] |
def x(result, config=None, html=True):
if config is None:
config = ReportConfiguration.load()
report = SnapshotReport(result=result, configuration=config)
if html:
return report.render_html()
else:
return report.render_json()
| snapshot_report | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(result, config=None, html=True):\n if config is None:\n config = ReportConfiguration.load()\n report = SnapshotReport(result=result, configuration=config)\n if html:\n return report.render_html()\n else:\n return report.render_json()\n",
"role": "user"
},
{
"content": "snapshot_report",
"role": "assistant"
}
] |
def x(history, config=None, html=True):
if config is None:
config = ReportConfiguration.load()
report = HistoryReport(history=history, configuration=config)
if html:
return report.render_html()
else:
return report.render_json()
| history_report | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(history, config=None, html=True):\n if config is None:\n config = ReportConfiguration.load()\n report = HistoryReport(history=history, configuration=config)\n if html:\n return report.render_html()\n else:\n return report.render_json()\n",
"role": "user"
},
{
"content": "history_report",
"role": "assistant"
}
] |
def x(diff_results, config=None, html=True):
if config is None:
config = ReportConfiguration.load()
report = DiffReport(diff_results=diff_results, configuration=config)
if html:
return report.render_html()
else:
return report.render_json()
| diff_report | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(diff_results, config=None, html=True):\n if config is None:\n config = ReportConfiguration.load()\n report = DiffReport(diff_results=diff_results, configuration=config)\n if html:\n return report.render_html()\n else:\n return report.render_json()\n",
"role": "user"
},
{
"content": "diff_report",
"role": "assistant"
}
] |
def x(path, notifications, filename):
env = Environment(loader=PackageLoader('memote.suite', 'templates'),
autoescape=select_autoescape(['html', 'xml']))
template = env.get_template('validation_template.html')
model = os.path.basename(path)
with open(filename, 'w') as file_h:
file_h.write(template.render(model=model, notifications=notifications))
| validation_report | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path, notifications, filename):\n env = Environment(loader=PackageLoader('memote.suite', 'templates'),\n autoescape=select_autoescape(['html', 'xml']))\n template = env.get_template('validation_template.html')\n model = os.path.basename(path)\n with open(filename, 'w') as file_h:\n file_h.write(template.render(model=model, notifications=notifications))\n",
"role": "user"
},
{
"content": "validation_report",
"role": "assistant"
}
] |
def x(cls, filename=None):
if filename is None:
LOGGER.debug('Loading default configuration.')
with open_text(templates, 'test_config.yml', encoding='utf-8'
) as file_handle:
content = yaml.load(file_handle)
else:
LOGGER.debug("Loading custom configuration '%s'.", filename)
try:
with open(filename, encoding='utf-8') as file_handle:
content = yaml.load(file_handle)
except IOError as err:
LOGGER.error(
"Failed to load the custom configuration '%s'. Skipping.",
filename)
LOGGER.debug(str(err))
content = dict()
return cls(content)
| load | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, filename=None):\n if filename is None:\n LOGGER.debug('Loading default configuration.')\n with open_text(templates, 'test_config.yml', encoding='utf-8'\n ) as file_handle:\n content = yaml.load(file_handle)\n else:\n LOGGER.debug(\"Loading custom configuration '%s'.\", filename)\n try:\n with open(filename, encoding='utf-8') as file_handle:\n content = yaml.load(file_handle)\n except IOError as err:\n LOGGER.error(\n \"Failed to load the custom configuration '%s'. Skipping.\",\n filename)\n LOGGER.debug(str(err))\n content = dict()\n return cls(content)\n",
"role": "user"
},
{
"content": "load",
"role": "assistant"
}
] |
def x(gpr):
logger.debug('%r', gpr)
conform = logical_and.sub('and', gpr)
conform = logical_or.sub('or', conform)
conform = escape_chars.sub('_', conform)
expression = ast.parse(conform)
walker = GPRVisitor()
walker.visit(expression)
return len(walker.left ^ walker.right)
| find_top_level_complex | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(gpr):\n logger.debug('%r', gpr)\n conform = logical_and.sub('and', gpr)\n conform = logical_or.sub('or', conform)\n conform = escape_chars.sub('_', conform)\n expression = ast.parse(conform)\n walker = GPRVisitor()\n walker.visit(expression)\n return len(walker.left ^ walker.right)\n",
"role": "user"
},
{
"content": "find_top_level_complex",
"role": "assistant"
}
] |
def x(self, node):
if self._is_top and isinstance(node.op, ast.And):
self._is_top = False
self._current = self.left
self.visit(node.values[0])
self._current = self.right
for successor in node.values[1:]:
self.visit(successor)
else:
self.generic_visit(node)
| visit_BoolOp | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, node):\n if self._is_top and isinstance(node.op, ast.And):\n self._is_top = False\n self._current = self.left\n self.visit(node.values[0])\n self._current = self.right\n for successor in node.values[1:]:\n self.visit(successor)\n else:\n self.generic_visit(node)\n",
"role": "user"
},
{
"content": "visit_BoolOp",
"role": "assistant"
}
] |
def x(model):
lower_bound, upper_bound = helpers.find_bounds(model)
return [rxn for rxn in model.reactions if 0 > rxn.lower_bound >
lower_bound or 0 < rxn.upper_bound < upper_bound]
| find_nonzero_constrained_reactions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n lower_bound, upper_bound = helpers.find_bounds(model)\n return [rxn for rxn in model.reactions if 0 > rxn.lower_bound >\n lower_bound or 0 < rxn.upper_bound < upper_bound]\n",
"role": "user"
},
{
"content": "find_nonzero_constrained_reactions",
"role": "assistant"
}
] |
def x(model):
return [rxn for rxn in model.reactions if rxn.lower_bound == 0 and rxn.
upper_bound == 0]
| find_zero_constrained_reactions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n return [rxn for rxn in model.reactions if rxn.lower_bound == 0 and rxn.\n upper_bound == 0]\n",
"role": "user"
},
{
"content": "find_zero_constrained_reactions",
"role": "assistant"
}
] |
def x(model):
lower_bound, upper_bound = helpers.find_bounds(model)
return [rxn for rxn in model.reactions if rxn.lower_bound <=
lower_bound and rxn.upper_bound >= upper_bound]
| find_unconstrained_reactions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n lower_bound, upper_bound = helpers.find_bounds(model)\n return [rxn for rxn in model.reactions if rxn.lower_bound <=\n lower_bound and rxn.upper_bound >= upper_bound]\n",
"role": "user"
},
{
"content": "find_unconstrained_reactions",
"role": "assistant"
}
] |
def x(model):
atp_adp_conv_rxns = helpers.find_converting_reactions(model, ('MNXM3',
'MNXM7'))
id_of_main_compartment = helpers.find_compartment_id_in_model(model, 'c')
reactants = {helpers.find_met_in_model(model, 'MNXM3',
id_of_main_compartment)[0], helpers.find_met_in_model(model,
'MNXM2', id_of_main_compartment)[0]}
products = {helpers.find_met_in_model(model, 'MNXM7',
id_of_main_compartment)[0], helpers.find_met_in_model(model,
'MNXM1', id_of_main_compartment)[0], helpers.find_met_in_model(
model, 'MNXM9', id_of_main_compartment)[0]}
candidates = [rxn for rxn in atp_adp_conv_rxns if rxn.reversibility is
False and set(rxn.reactants) == reactants and set(rxn.products) ==
products]
buzzwords = ['maintenance', 'atpm', 'requirement', 'ngam', 'non-growth',
'associated']
refined_candidates = [rxn for rxn in candidates if any(string in
filter_none(rxn.name, '').lower() for string in buzzwords)]
if refined_candidates:
return refined_candidates
else:
return candidates
| find_ngam | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n atp_adp_conv_rxns = helpers.find_converting_reactions(model, ('MNXM3',\n 'MNXM7'))\n id_of_main_compartment = helpers.find_compartment_id_in_model(model, 'c')\n reactants = {helpers.find_met_in_model(model, 'MNXM3',\n id_of_main_compartment)[0], helpers.find_met_in_model(model,\n 'MNXM2', id_of_main_compartment)[0]}\n products = {helpers.find_met_in_model(model, 'MNXM7',\n id_of_main_compartment)[0], helpers.find_met_in_model(model,\n 'MNXM1', id_of_main_compartment)[0], helpers.find_met_in_model(\n model, 'MNXM9', id_of_main_compartment)[0]}\n candidates = [rxn for rxn in atp_adp_conv_rxns if rxn.reversibility is \n False and set(rxn.reactants) == reactants and set(rxn.products) ==\n products]\n buzzwords = ['maintenance', 'atpm', 'requirement', 'ngam', 'non-growth',\n 'associated']\n refined_candidates = [rxn for rxn in candidates if any(string in\n filter_none(rxn.name, '').lower() for string in buzzwords)]\n if refined_candidates:\n return refined_candidates\n else:\n return candidates\n",
"role": "user"
},
{
"content": "find_ngam",
"role": "assistant"
}
] |
def x(model):
if len(model.reactions) == 0 or len(model.genes) == 0:
raise ValueError('The model contains no reactions or genes.')
return float(len(model.reactions)) / float(len(model.genes))
| calculate_metabolic_coverage | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n if len(model.reactions) == 0 or len(model.genes) == 0:\n raise ValueError('The model contains no reactions or genes.')\n return float(len(model.reactions)) / float(len(model.genes))\n",
"role": "user"
},
{
"content": "calculate_metabolic_coverage",
"role": "assistant"
}
] |
def x(model):
complexes = []
for rxn in model.reactions:
if not rxn.gene_reaction_rule:
continue
size = find_top_level_complex(rxn.gene_reaction_rule)
if size >= 2:
complexes.append(rxn)
return complexes
| find_protein_complexes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n complexes = []\n for rxn in model.reactions:\n if not rxn.gene_reaction_rule:\n continue\n size = find_top_level_complex(rxn.gene_reaction_rule)\n if size >= 2:\n complexes.append(rxn)\n return complexes\n",
"role": "user"
},
{
"content": "find_protein_complexes",
"role": "assistant"
}
] |
def x(model, rxn):
lower_bound, upper_bound = helpers.find_bounds(model)
if rxn.reversibility:
return rxn.lower_bound > lower_bound or rxn.upper_bound < upper_bound
else:
return rxn.lower_bound > 0 or rxn.upper_bound < upper_bound
| is_constrained_reaction | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model, rxn):\n lower_bound, upper_bound = helpers.find_bounds(model)\n if rxn.reversibility:\n return rxn.lower_bound > lower_bound or rxn.upper_bound < upper_bound\n else:\n return rxn.lower_bound > 0 or rxn.upper_bound < upper_bound\n",
"role": "user"
},
{
"content": "is_constrained_reaction",
"role": "assistant"
}
] |
def x(model):
o2_in_model = helpers.find_met_in_model(model, 'MNXM4')
return set([rxn for met in model.metabolites for rxn in met.reactions if
met.formula == 'O2' or met in o2_in_model])
| find_oxygen_reactions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n o2_in_model = helpers.find_met_in_model(model, 'MNXM4')\n return set([rxn for met in model.metabolites for rxn in met.reactions if\n met.formula == 'O2' or met in o2_in_model])\n",
"role": "user"
},
{
"content": "find_oxygen_reactions",
"role": "assistant"
}
] |
def x(model):
unique = set()
for met in model.metabolites:
is_missing = True
for comp in model.compartments:
if met.id.endswith('_{}'.format(comp)):
unique.add(met.id[:-(len(comp) + 1)])
is_missing = False
break
if is_missing:
unique.add(met.id)
return unique
| find_unique_metabolites | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n unique = set()\n for met in model.metabolites:\n is_missing = True\n for comp in model.compartments:\n if met.id.endswith('_{}'.format(comp)):\n unique.add(met.id[:-(len(comp) + 1)])\n is_missing = False\n break\n if is_missing:\n unique.add(met.id)\n return unique\n",
"role": "user"
},
{
"content": "find_unique_metabolites",
"role": "assistant"
}
] |
def x(model):
unique_identifiers = ['inchikey', 'inchi']
duplicates = []
for met_1, met_2 in combinations(model.metabolites, 2):
if met_1.compartment == met_2.compartment:
for key in unique_identifiers:
if key in met_1.annotation and key in met_2.annotation:
if met_1.annotation[key] == met_2.annotation[key]:
duplicates.append((met_1.id, met_2.id))
break
return duplicates
| find_duplicate_metabolites_in_compartments | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n unique_identifiers = ['inchikey', 'inchi']\n duplicates = []\n for met_1, met_2 in combinations(model.metabolites, 2):\n if met_1.compartment == met_2.compartment:\n for key in unique_identifiers:\n if key in met_1.annotation and key in met_2.annotation:\n if met_1.annotation[key] == met_2.annotation[key]:\n duplicates.append((met_1.id, met_2.id))\n break\n return duplicates\n",
"role": "user"
},
{
"content": "find_duplicate_metabolites_in_compartments",
"role": "assistant"
}
] |
def x(model):
duplicates = {}
rxn_db_identifiers = ['metanetx.reaction', 'kegg.reaction', 'brenda',
'rhea', 'biocyc', 'bigg.reaction']
ann_rxns = []
for rxn in model.reactions:
ann = []
for key in rxn_db_identifiers:
if key in rxn.annotation:
if isinstance(rxn.annotation[key], list):
ann.extend([(key, elem) for elem in rxn.annotation[key]])
else:
ann.append((key, rxn.annotation[key]))
ann_rxns.append((rxn, frozenset(ann)))
for (rxn_a, ann_a), (rxn_b, ann_b) in combinations(ann_rxns, 2):
mutual_pair = tuple(ann_a & ann_b)
if len(mutual_pair) > 0:
duplicates.setdefault(mutual_pair, set()).update([rxn_a.id,
rxn_b.id])
num_duplicated = set()
duplicated = {}
for key in duplicates:
new_key = ','.join(sorted('{}:{}'.format(ns, term) for ns, term in key)
)
duplicated[new_key] = rxns = list(duplicates[key])
num_duplicated.update(rxns)
return duplicated, len(num_duplicated)
| find_reactions_with_partially_identical_annotations | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n duplicates = {}\n rxn_db_identifiers = ['metanetx.reaction', 'kegg.reaction', 'brenda',\n 'rhea', 'biocyc', 'bigg.reaction']\n ann_rxns = []\n for rxn in model.reactions:\n ann = []\n for key in rxn_db_identifiers:\n if key in rxn.annotation:\n if isinstance(rxn.annotation[key], list):\n ann.extend([(key, elem) for elem in rxn.annotation[key]])\n else:\n ann.append((key, rxn.annotation[key]))\n ann_rxns.append((rxn, frozenset(ann)))\n for (rxn_a, ann_a), (rxn_b, ann_b) in combinations(ann_rxns, 2):\n mutual_pair = tuple(ann_a & ann_b)\n if len(mutual_pair) > 0:\n duplicates.setdefault(mutual_pair, set()).update([rxn_a.id,\n rxn_b.id])\n num_duplicated = set()\n duplicated = {}\n for key in duplicates:\n new_key = ','.join(sorted('{}:{}'.format(ns, term) for ns, term in key)\n )\n duplicated[new_key] = rxns = list(duplicates[key])\n num_duplicated.update(rxns)\n return duplicated, len(num_duplicated)\n",
"role": "user"
},
{
"content": "find_reactions_with_partially_identical_annotations",
"role": "assistant"
}
] |
def x(metabolites, compartments):
unique_identifiers = ['inchikey', 'inchi']
met2mol = {}
molecules = {c: [] for c in compartments}
for met in metabolites:
ann = []
for key in unique_identifiers:
mol = met.annotation.get(key)
if mol is not None:
ann.append(mol)
if len(ann) == 0:
continue
ann = set(ann)
mols = molecules[met.compartment]
for i, mol_group in enumerate(mols):
if len(ann & mol_group) > 0:
mol_group.update(ann)
met2mol[met] = '{}-{}'.format(met.compartment, i)
break
if met not in met2mol:
met2mol[met] = '{}-{}'.format(met.compartment, len(mols))
mols.append(ann)
return met2mol
| map_metabolites_to_structures | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(metabolites, compartments):\n unique_identifiers = ['inchikey', 'inchi']\n met2mol = {}\n molecules = {c: [] for c in compartments}\n for met in metabolites:\n ann = []\n for key in unique_identifiers:\n mol = met.annotation.get(key)\n if mol is not None:\n ann.append(mol)\n if len(ann) == 0:\n continue\n ann = set(ann)\n mols = molecules[met.compartment]\n for i, mol_group in enumerate(mols):\n if len(ann & mol_group) > 0:\n mol_group.update(ann)\n met2mol[met] = '{}-{}'.format(met.compartment, i)\n break\n if met not in met2mol:\n met2mol[met] = '{}-{}'.format(met.compartment, len(mols))\n mols.append(ann)\n return met2mol\n",
"role": "user"
},
{
"content": "map_metabolites_to_structures",
"role": "assistant"
}
] |
def x(model):
met2mol = map_metabolites_to_structures(model.metabolites, model.
compartments)
structural = []
for rxn in model.reactions:
if not all(met in met2mol for met in rxn.metabolites):
continue
substrates = {met2mol[met]: rxn.get_coefficient(met) for met in rxn
.reactants}
products = {met2mol[met]: rxn.get_coefficient(met) for met in rxn.
products}
structural.append((rxn, substrates, products))
num_duplicated = set()
duplicates = []
for (rxn_a, sub_a, prod_a), (rxn_b, sub_b, prod_b) in combinations(
structural, 2):
if sub_a != sub_b:
continue
if prod_a != prod_b:
continue
if rxn_a.reversibility != rxn_b.reversibility:
continue
duplicates.append((rxn_a.id, rxn_b.id))
num_duplicated.add(rxn_a.id)
num_duplicated.add(rxn_b.id)
return duplicates, len(num_duplicated)
| find_duplicate_reactions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n met2mol = map_metabolites_to_structures(model.metabolites, model.\n compartments)\n structural = []\n for rxn in model.reactions:\n if not all(met in met2mol for met in rxn.metabolites):\n continue\n substrates = {met2mol[met]: rxn.get_coefficient(met) for met in rxn\n .reactants}\n products = {met2mol[met]: rxn.get_coefficient(met) for met in rxn.\n products}\n structural.append((rxn, substrates, products))\n num_duplicated = set()\n duplicates = []\n for (rxn_a, sub_a, prod_a), (rxn_b, sub_b, prod_b) in combinations(\n structural, 2):\n if sub_a != sub_b:\n continue\n if prod_a != prod_b:\n continue\n if rxn_a.reversibility != rxn_b.reversibility:\n continue\n duplicates.append((rxn_a.id, rxn_b.id))\n num_duplicated.add(rxn_a.id)\n num_duplicated.add(rxn_b.id)\n return duplicates, len(num_duplicated)\n",
"role": "user"
},
{
"content": "find_duplicate_reactions",
"role": "assistant"
}
] |
def x(model):
duplicates = dict()
for rxn_a, rxn_b in combinations(model.reactions, 2):
if not (rxn_a.genes and rxn_b.genes):
continue
if rxn_a.genes == rxn_b.genes:
identifiers = rxn_a.genes
duplicates.setdefault(identifiers, set()).update([rxn_a.id,
rxn_b.id])
num_duplicated = set()
duplicated = {}
for key in duplicates:
new_key = ','.join(sorted(g.id for g in key))
duplicated[new_key] = rxns = list(duplicates[key])
num_duplicated.update(rxns)
return duplicated, len(num_duplicated)
| find_reactions_with_identical_genes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n duplicates = dict()\n for rxn_a, rxn_b in combinations(model.reactions, 2):\n if not (rxn_a.genes and rxn_b.genes):\n continue\n if rxn_a.genes == rxn_b.genes:\n identifiers = rxn_a.genes\n duplicates.setdefault(identifiers, set()).update([rxn_a.id,\n rxn_b.id])\n num_duplicated = set()\n duplicated = {}\n for key in duplicates:\n new_key = ','.join(sorted(g.id for g in key))\n duplicated[new_key] = rxns = list(duplicates[key])\n num_duplicated.update(rxns)\n return duplicated, len(num_duplicated)\n",
"role": "user"
},
{
"content": "find_reactions_with_identical_genes",
"role": "assistant"
}
] |
def x(model):
return [met.id for rxn in model.medium for met in model.reactions.
get_by_id(rxn).metabolites]
| find_medium_metabolites | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n return [met.id for rxn in model.medium for met in model.reactions.\n get_by_id(rxn).metabolites]\n",
"role": "user"
},
{
"content": "find_medium_metabolites",
"role": "assistant"
}
] |
def x(model):
ex_comp = find_external_compartment(model)
return [met for met in model.metabolites if met.compartment == ex_comp]
| find_external_metabolites | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model):\n ex_comp = find_external_compartment(model)\n return [met for met in model.metabolites if met.compartment == ex_comp]\n",
"role": "user"
},
{
"content": "find_external_metabolites",
"role": "assistant"
}
] |
def x(self, result, filename, pretty=True):
LOGGER.info("Storing result in '%s'.", filename)
if filename.endswith('.gz'):
with gzip.open(filename, 'wb') as file_handle:
file_handle.write(jsonify(result, pretty=pretty).encode('utf-8'))
else:
with open(filename, 'w', encoding='utf-8') as file_handle:
file_handle.write(jsonify(result, pretty=pretty))
| store | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, result, filename, pretty=True):\n LOGGER.info(\"Storing result in '%s'.\", filename)\n if filename.endswith('.gz'):\n with gzip.open(filename, 'wb') as file_handle:\n file_handle.write(jsonify(result, pretty=pretty).encode('utf-8'))\n else:\n with open(filename, 'w', encoding='utf-8') as file_handle:\n file_handle.write(jsonify(result, pretty=pretty))\n",
"role": "user"
},
{
"content": "store",
"role": "assistant"
}
] |
def x(self, filename):
LOGGER.info("Loading result from '%s'.", filename)
if filename.endswith('.gz'):
with gzip.open(filename, 'rb') as file_handle:
result = MemoteResult(json.loads(file_handle.read().decode(
'utf-8')))
else:
with open(filename, 'r', encoding='utf-8') as file_handle:
result = MemoteResult(json.load(file_handle))
return result
| load | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, filename):\n LOGGER.info(\"Loading result from '%s'.\", filename)\n if filename.endswith('.gz'):\n with gzip.open(filename, 'rb') as file_handle:\n result = MemoteResult(json.loads(file_handle.read().decode(\n 'utf-8')))\n else:\n with open(filename, 'r', encoding='utf-8') as file_handle:\n result = MemoteResult(json.load(file_handle))\n return result\n",
"role": "user"
},
{
"content": "load",
"role": "assistant"
}
] |
def x(path, notifications):
doc = libsbml.readSBML(path)
fbc = doc.getPlugin('fbc')
sbml_ver = doc.getLevel(), doc.getVersion(
), fbc if fbc is None else fbc.getVersion()
with catch_warnings(record=True) as warnings:
simplefilter('always')
try:
model = read_sbml_model(path)
except Exception as err:
notifications['errors'].append(str(err))
model = None
validate = True
else:
validate = False
notifications['warnings'].extend([str(w.message) for w in warnings])
if validate:
run_sbml_validation(doc, notifications)
return model, sbml_ver
| load_cobra_model | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path, notifications):\n doc = libsbml.readSBML(path)\n fbc = doc.getPlugin('fbc')\n sbml_ver = doc.getLevel(), doc.getVersion(\n ), fbc if fbc is None else fbc.getVersion()\n with catch_warnings(record=True) as warnings:\n simplefilter('always')\n try:\n model = read_sbml_model(path)\n except Exception as err:\n notifications['errors'].append(str(err))\n model = None\n validate = True\n else:\n validate = False\n notifications['warnings'].extend([str(w.message) for w in warnings])\n if validate:\n run_sbml_validation(doc, notifications)\n return model, sbml_ver\n",
"role": "user"
},
{
"content": "load_cobra_model",
"role": "assistant"
}
] |
def x(failure):
return 'Line {}, Column {} - #{}: {} - Category: {}, Severity: {}'.format(
failure.getLine(), failure.getColumn(), failure.getErrorId(),
failure.getMessage(), failure.getCategoryAsString(), failure.
getSeverity())
| format_failure | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(failure):\n return 'Line {}, Column {} - #{}: {} - Category: {}, Severity: {}'.format(\n failure.getLine(), failure.getColumn(), failure.getErrorId(),\n failure.getMessage(), failure.getCategoryAsString(), failure.\n getSeverity())\n",
"role": "user"
},
{
"content": "format_failure",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.