source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(requestContext, *seriesLists): if not seriesLists or not any(seriesLists): series = constantLine(requestContext, 0).pop() series.pathExpression = 'countSeries()' else: seriesList, start, end, step = normalize(seriesLists) name = 'countSeries(%s)' % formatPathExpressions(seriesList) values = (int(len(row)) for row in zip_longest(*seriesList)) series = TimeSeries(name, start, end, step, values) series.pathExpression = name return [series]
countSeries
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(requestContext, *seriesLists):\n if not seriesLists or not any(seriesLists):\n series = constantLine(requestContext, 0).pop()\n series.pathExpression = 'countSeries()'\n else:\n seriesList, start, end, step = normalize(seriesLists)\n name = 'countSeries(%s)' % formatPathExpressions(seriesList)\n values = (int(len(row)) for row in zip_longest(*seriesList))\n series = TimeSeries(name, start, end, step, values)\n series.pathExpression = name\n return [series]\n", "role": "user" }, { "content": "countSeries", "role": "assistant" } ]
def x(requestContext, *seriesLists): seriesGroup = [] for s in seriesLists: seriesGroup.extend(s) return seriesGroup
group
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(requestContext, *seriesLists):\n seriesGroup = []\n for s in seriesLists:\n seriesGroup.extend(s)\n return seriesGroup\n", "role": "user" }, { "content": "group", "role": "assistant" } ]
def x(requestContext, seriesList, mapNode): metaSeries = {} keys = [] for series in seriesList: key = series.name.split('.')[mapNode] if key not in metaSeries: metaSeries[key] = [series] keys.append(key) else: metaSeries[key].append(series) return [metaSeries[k] for k in keys]
mapSeries
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(requestContext, seriesList, mapNode):\n metaSeries = {}\n keys = []\n for series in seriesList:\n key = series.name.split('.')[mapNode]\n if key not in metaSeries:\n metaSeries[key] = [series]\n keys.append(key)\n else:\n metaSeries[key].append(series)\n return [metaSeries[k] for k in keys]\n", "role": "user" }, { "content": "mapSeries", "role": "assistant" } ]
def x(requestContext, seriesLists, reduceFunction, reduceNode, *reduceMatchers ): metaSeries = {} keys = [] for seriesList in seriesLists: for series in seriesList: nodes = series.name.split('.') node = nodes[reduceNode] reduceSeriesName = '.'.join(nodes[0:reduceNode] ) + '.reduce.' + reduceFunction if node in reduceMatchers: if reduceSeriesName not in metaSeries: metaSeries[reduceSeriesName] = [None] * len(reduceMatchers) keys.append(reduceSeriesName) i = reduceMatchers.index(node) metaSeries[reduceSeriesName][i] = series for key in keys: metaSeries[key] = app.functions[reduceFunction](requestContext, *[[ s] for s in metaSeries[key]])[0] metaSeries[key].name = key return [metaSeries[key] for key in keys]
reduceSeries
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(requestContext, seriesLists, reduceFunction, reduceNode, *reduceMatchers\n ):\n metaSeries = {}\n keys = []\n for seriesList in seriesLists:\n for series in seriesList:\n nodes = series.name.split('.')\n node = nodes[reduceNode]\n reduceSeriesName = '.'.join(nodes[0:reduceNode]\n ) + '.reduce.' + reduceFunction\n if node in reduceMatchers:\n if reduceSeriesName not in metaSeries:\n metaSeries[reduceSeriesName] = [None] * len(reduceMatchers)\n keys.append(reduceSeriesName)\n i = reduceMatchers.index(node)\n metaSeries[reduceSeriesName][i] = series\n for key in keys:\n metaSeries[key] = app.functions[reduceFunction](requestContext, *[[\n s] for s in metaSeries[key]])[0]\n metaSeries[key].name = key\n return [metaSeries[key] for key in keys]\n", "role": "user" }, { "content": "reduceSeries", "role": "assistant" } ]
def x(requestContext, seriesList, nodeNum, templateFunction, newName=None): from .app import evaluateTarget prefixes = set() for series in seriesList: prefix = '.'.join(series.name.split('.')[:nodeNum + 1]) prefixes.add(prefix) results = [] for prefix in sorted(prefixes): target = templateFunction.replace('%', prefix) for resultSeries in evaluateTarget(requestContext, target): if newName: resultSeries.name = newName.replace('%', prefix) resultSeries.pathExpression = prefix resultSeries.start = series.start resultSeries.end = series.end results.append(resultSeries) return results
applyByNode
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(requestContext, seriesList, nodeNum, templateFunction, newName=None):\n from .app import evaluateTarget\n prefixes = set()\n for series in seriesList:\n prefix = '.'.join(series.name.split('.')[:nodeNum + 1])\n prefixes.add(prefix)\n results = []\n for prefix in sorted(prefixes):\n target = templateFunction.replace('%', prefix)\n for resultSeries in evaluateTarget(requestContext, target):\n if newName:\n resultSeries.name = newName.replace('%', prefix)\n resultSeries.pathExpression = prefix\n resultSeries.start = series.start\n resultSeries.end = series.end\n results.append(resultSeries)\n return results\n", "role": "user" }, { "content": "applyByNode", "role": "assistant" } ]
def x(requestContext, seriesList, nodeNum, callback): return groupByNodes(requestContext, seriesList, callback, nodeNum)
groupByNode
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(requestContext, seriesList, nodeNum, callback):\n return groupByNodes(requestContext, seriesList, callback, nodeNum)\n", "role": "user" }, { "content": "groupByNode", "role": "assistant" } ]
def x(requestContext, seriesList, callback, *nodes): from .app import app metaSeries = {} keys = [] if isinstance(nodes, int): nodes = [nodes] for series in seriesList: key = '.'.join(series.name.split('.')[n] for n in nodes) if key not in metaSeries: metaSeries[key] = [series] keys.append(key) else: metaSeries[key].append(series) for key in metaSeries: metaSeries[key] = app.functions[callback](requestContext, metaSeries[key])[0] metaSeries[key].name = key return [metaSeries[key] for key in keys]
groupByNodes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(requestContext, seriesList, callback, *nodes):\n from .app import app\n metaSeries = {}\n keys = []\n if isinstance(nodes, int):\n nodes = [nodes]\n for series in seriesList:\n key = '.'.join(series.name.split('.')[n] for n in nodes)\n if key not in metaSeries:\n metaSeries[key] = [series]\n keys.append(key)\n else:\n metaSeries[key].append(series)\n for key in metaSeries:\n metaSeries[key] = app.functions[callback](requestContext,\n metaSeries[key])[0]\n metaSeries[key].name = key\n return [metaSeries[key] for key in keys]\n", "role": "user" }, { "content": "groupByNodes", "role": "assistant" } ]
def x(requestContext, seriesList, pattern): regex = re.compile(pattern) return [s for s in seriesList if not regex.search(s.name)]
exclude
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(requestContext, seriesList, pattern):\n regex = re.compile(pattern)\n return [s for s in seriesList if not regex.search(s.name)]\n", "role": "user" }, { "content": "exclude", "role": "assistant" } ]
def x(requestContext, seriesList, intervalString, func='sum'): results = [] delta = parseTimeOffset(intervalString) interval = to_seconds(delta) requestContext = requestContext.copy() tzinfo = requestContext['tzinfo'] s = requestContext['startTime'] if interval >= DAY: requestContext['startTime'] = datetime(s.year, s.month, s.day, tzinfo=tzinfo) elif interval >= HOUR: requestContext['startTime'] = datetime(s.year, s.month, s.day, s. hour, tzinfo=tzinfo) elif interval >= MINUTE: requestContext['startTime'] = datetime(s.year, s.month, s.day, s. hour, s.minute, tzinfo=tzinfo) paths = [] for series in seriesList: paths.extend(pathsFromTarget(requestContext, series.pathExpression)) data_store = fetchData(requestContext, paths) for series in seriesList: newSeries = evaluateTarget(requestContext, series.pathExpression, data_store)[0] series[0:len(series)] = newSeries series.start = newSeries.start series.end = newSeries.end series.step = newSeries.step for series in seriesList: buckets = {} timestamps = range(int(series.start), int(series.end), int(series.step) ) datapoints = zip_longest(timestamps, series) for timestamp, value in datapoints: if not timestamp: continue bucketInterval = int((timestamp - series.start) / interval) if bucketInterval not in buckets: buckets[bucketInterval] = [] if value is not None: buckets[bucketInterval].append(value) newValues = [] for timestamp in range(series.start, series.end, interval): bucketInterval = int((timestamp - series.start) / interval) bucket = buckets.get(bucketInterval, []) if bucket: if func == 'avg': newValues.append(float(sum(bucket)) / float(len(bucket))) elif func == 'last': newValues.append(bucket[len(bucket) - 1]) elif func == 'max': newValues.append(max(bucket)) elif func == 'min': newValues.append(min(bucket)) else: newValues.append(sum(bucket)) else: newValues.append(None) newName = 'smartSummarize(%s, "%s", "%s")' % (series.name, intervalString, func) alignedEnd = series.start + bucketInterval * interval + interval newSeries = TimeSeries(newName, series.start, alignedEnd, interval, newValues) newSeries.pathExpression = newName results.append(newSeries) return results
smartSummarize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(requestContext, seriesList, intervalString, func='sum'):\n results = []\n delta = parseTimeOffset(intervalString)\n interval = to_seconds(delta)\n requestContext = requestContext.copy()\n tzinfo = requestContext['tzinfo']\n s = requestContext['startTime']\n if interval >= DAY:\n requestContext['startTime'] = datetime(s.year, s.month, s.day,\n tzinfo=tzinfo)\n elif interval >= HOUR:\n requestContext['startTime'] = datetime(s.year, s.month, s.day, s.\n hour, tzinfo=tzinfo)\n elif interval >= MINUTE:\n requestContext['startTime'] = datetime(s.year, s.month, s.day, s.\n hour, s.minute, tzinfo=tzinfo)\n paths = []\n for series in seriesList:\n paths.extend(pathsFromTarget(requestContext, series.pathExpression))\n data_store = fetchData(requestContext, paths)\n for series in seriesList:\n newSeries = evaluateTarget(requestContext, series.pathExpression,\n data_store)[0]\n series[0:len(series)] = newSeries\n series.start = newSeries.start\n series.end = newSeries.end\n series.step = newSeries.step\n for series in seriesList:\n buckets = {}\n timestamps = range(int(series.start), int(series.end), int(series.step)\n )\n datapoints = zip_longest(timestamps, series)\n for timestamp, value in datapoints:\n if not timestamp:\n continue\n bucketInterval = int((timestamp - series.start) / interval)\n if bucketInterval not in buckets:\n buckets[bucketInterval] = []\n if value is not None:\n buckets[bucketInterval].append(value)\n newValues = []\n for timestamp in range(series.start, series.end, interval):\n bucketInterval = int((timestamp - series.start) / interval)\n bucket = buckets.get(bucketInterval, [])\n if bucket:\n if func == 'avg':\n newValues.append(float(sum(bucket)) / float(len(bucket)))\n elif func == 'last':\n newValues.append(bucket[len(bucket) - 1])\n elif func == 'max':\n newValues.append(max(bucket))\n elif func == 'min':\n newValues.append(min(bucket))\n else:\n newValues.append(sum(bucket))\n else:\n newValues.append(None)\n newName = 'smartSummarize(%s, \"%s\", \"%s\")' % (series.name,\n intervalString, func)\n alignedEnd = series.start + bucketInterval * interval + interval\n newSeries = TimeSeries(newName, series.start, alignedEnd, interval,\n newValues)\n newSeries.pathExpression = newName\n results.append(newSeries)\n return results\n", "role": "user" }, { "content": "smartSummarize", "role": "assistant" } ]
def x(requestContext, seriesList, intervalString, func='sum', alignToFrom=False ): results = [] delta = parseTimeOffset(intervalString) interval = to_seconds(delta) for series in seriesList: buckets = {} timestamps = range(int(series.start), int(series.end) + 1, int( series.step)) datapoints = zip_longest(timestamps, series) for timestamp, value in datapoints: if timestamp is None: continue if alignToFrom: bucketInterval = int((timestamp - series.start) / interval) else: bucketInterval = timestamp - timestamp % interval if bucketInterval not in buckets: buckets[bucketInterval] = [] if value is not None: buckets[bucketInterval].append(value) if alignToFrom: newStart = series.start newEnd = series.end else: newStart = series.start - series.start % interval newEnd = series.end - series.end % interval + interval newValues = [] for timestamp in range(newStart, newEnd, interval): if alignToFrom: newEnd = timestamp bucketInterval = int((timestamp - series.start) / interval) else: bucketInterval = timestamp - timestamp % interval bucket = buckets.get(bucketInterval, []) if bucket: if func == 'avg': newValues.append(float(sum(bucket)) / float(len(bucket))) elif func == 'last': newValues.append(bucket[len(bucket) - 1]) elif func == 'max': newValues.append(max(bucket)) elif func == 'min': newValues.append(min(bucket)) else: newValues.append(sum(bucket)) else: newValues.append(None) if alignToFrom: newEnd += interval newName = 'summarize(%s, "%s", "%s"%s)' % (series.name, intervalString, func, alignToFrom and ', true' or '') newSeries = TimeSeries(newName, newStart, newEnd, interval, newValues) newSeries.pathExpression = newName results.append(newSeries) return results
summarize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(requestContext, seriesList, intervalString, func='sum', alignToFrom=False\n ):\n results = []\n delta = parseTimeOffset(intervalString)\n interval = to_seconds(delta)\n for series in seriesList:\n buckets = {}\n timestamps = range(int(series.start), int(series.end) + 1, int(\n series.step))\n datapoints = zip_longest(timestamps, series)\n for timestamp, value in datapoints:\n if timestamp is None:\n continue\n if alignToFrom:\n bucketInterval = int((timestamp - series.start) / interval)\n else:\n bucketInterval = timestamp - timestamp % interval\n if bucketInterval not in buckets:\n buckets[bucketInterval] = []\n if value is not None:\n buckets[bucketInterval].append(value)\n if alignToFrom:\n newStart = series.start\n newEnd = series.end\n else:\n newStart = series.start - series.start % interval\n newEnd = series.end - series.end % interval + interval\n newValues = []\n for timestamp in range(newStart, newEnd, interval):\n if alignToFrom:\n newEnd = timestamp\n bucketInterval = int((timestamp - series.start) / interval)\n else:\n bucketInterval = timestamp - timestamp % interval\n bucket = buckets.get(bucketInterval, [])\n if bucket:\n if func == 'avg':\n newValues.append(float(sum(bucket)) / float(len(bucket)))\n elif func == 'last':\n newValues.append(bucket[len(bucket) - 1])\n elif func == 'max':\n newValues.append(max(bucket))\n elif func == 'min':\n newValues.append(min(bucket))\n else:\n newValues.append(sum(bucket))\n else:\n newValues.append(None)\n if alignToFrom:\n newEnd += interval\n newName = 'summarize(%s, \"%s\", \"%s\"%s)' % (series.name,\n intervalString, func, alignToFrom and ', true' or '')\n newSeries = TimeSeries(newName, newStart, newEnd, interval, newValues)\n newSeries.pathExpression = newName\n results.append(newSeries)\n return results\n", "role": "user" }, { "content": "summarize", "role": "assistant" } ]
def x(requestContext, seriesList, intervalString, alignToInterval=False): results = [] delta = parseTimeOffset(intervalString) interval = to_seconds(delta) if alignToInterval: requestContext = requestContext.copy() tzinfo = requestContext['tzinfo'] s = requestContext['startTime'] if interval >= DAY: requestContext['startTime'] = datetime(s.year, s.month, s.day, tzinfo=tzinfo) elif interval >= HOUR: requestContext['startTime'] = datetime(s.year, s.month, s.day, s.hour, tzinfo=tzinfo) elif interval >= MINUTE: requestContext['startTime'] = datetime(s.year, s.month, s.day, s.hour, s.minute, tzinfo=tzinfo) paths = [] for series in seriesList: paths.extend(pathsFromTarget(requestContext, series.pathExpression) ) data_store = fetchData(requestContext, paths) for series in seriesList: newSeries = evaluateTarget(requestContext, series. pathExpression, data_store)[0] intervalCount = int((series.end - series.start) / interval) series[0:len(series)] = newSeries series.start = newSeries.start series.end = newSeries.start + intervalCount * interval + interval series.step = newSeries.step for series in seriesList: step = int(series.step) bucket_count = int(math.ceil(float(series.end - series.start) / interval)) buckets = [[] for _ in range(bucket_count)] newStart = int(series.end - bucket_count * interval) for i, value in enumerate(series): if value is None: continue start_time = int(series.start + i * step) start_bucket, start_mod = divmod(start_time - newStart, interval) end_time = start_time + step end_bucket, end_mod = divmod(end_time - newStart, interval) if end_bucket >= bucket_count: end_bucket = bucket_count - 1 end_mod = interval if start_bucket == end_bucket: if start_bucket >= 0: buckets[start_bucket].append(value * (end_mod - start_mod)) else: if start_bucket >= 0: buckets[start_bucket].append(value * (interval - start_mod) ) hits_per_bucket = value * interval for j in range(start_bucket + 1, end_bucket): buckets[j].append(hits_per_bucket) if end_mod > 0: buckets[end_bucket].append(value * end_mod) newValues = [] for bucket in buckets: if bucket: newValues.append(sum(bucket)) else: newValues.append(None) newName = 'hitcount(%s, "%s"%s)' % (series.name, intervalString, alignToInterval and ', true' or '') newSeries = TimeSeries(newName, newStart, series.end, interval, newValues) newSeries.pathExpression = newName results.append(newSeries) return results
hitcount
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(requestContext, seriesList, intervalString, alignToInterval=False):\n results = []\n delta = parseTimeOffset(intervalString)\n interval = to_seconds(delta)\n if alignToInterval:\n requestContext = requestContext.copy()\n tzinfo = requestContext['tzinfo']\n s = requestContext['startTime']\n if interval >= DAY:\n requestContext['startTime'] = datetime(s.year, s.month, s.day,\n tzinfo=tzinfo)\n elif interval >= HOUR:\n requestContext['startTime'] = datetime(s.year, s.month, s.day,\n s.hour, tzinfo=tzinfo)\n elif interval >= MINUTE:\n requestContext['startTime'] = datetime(s.year, s.month, s.day,\n s.hour, s.minute, tzinfo=tzinfo)\n paths = []\n for series in seriesList:\n paths.extend(pathsFromTarget(requestContext, series.pathExpression)\n )\n data_store = fetchData(requestContext, paths)\n for series in seriesList:\n newSeries = evaluateTarget(requestContext, series.\n pathExpression, data_store)[0]\n intervalCount = int((series.end - series.start) / interval)\n series[0:len(series)] = newSeries\n series.start = newSeries.start\n series.end = newSeries.start + intervalCount * interval + interval\n series.step = newSeries.step\n for series in seriesList:\n step = int(series.step)\n bucket_count = int(math.ceil(float(series.end - series.start) /\n interval))\n buckets = [[] for _ in range(bucket_count)]\n newStart = int(series.end - bucket_count * interval)\n for i, value in enumerate(series):\n if value is None:\n continue\n start_time = int(series.start + i * step)\n start_bucket, start_mod = divmod(start_time - newStart, interval)\n end_time = start_time + step\n end_bucket, end_mod = divmod(end_time - newStart, interval)\n if end_bucket >= bucket_count:\n end_bucket = bucket_count - 1\n end_mod = interval\n if start_bucket == end_bucket:\n if start_bucket >= 0:\n buckets[start_bucket].append(value * (end_mod - start_mod))\n else:\n if start_bucket >= 0:\n buckets[start_bucket].append(value * (interval - start_mod)\n )\n hits_per_bucket = value * interval\n for j in range(start_bucket + 1, end_bucket):\n buckets[j].append(hits_per_bucket)\n if end_mod > 0:\n buckets[end_bucket].append(value * end_mod)\n newValues = []\n for bucket in buckets:\n if bucket:\n newValues.append(sum(bucket))\n else:\n newValues.append(None)\n newName = 'hitcount(%s, \"%s\"%s)' % (series.name, intervalString, \n alignToInterval and ', true' or '')\n newSeries = TimeSeries(newName, newStart, series.end, interval,\n newValues)\n newSeries.pathExpression = newName\n results.append(newSeries)\n return results\n", "role": "user" }, { "content": "hitcount", "role": "assistant" } ]
def x(requestContext, name, step=60): start = int(epoch(requestContext['startTime'])) end = int(epoch(requestContext['endTime'])) delta = timedelta(seconds=step) when = requestContext['startTime'] values = [] while when < requestContext['endTime']: values.append(epoch(when)) when += delta series = TimeSeries(name, start, end, step, values) series.pathExpression = name return [series]
timeFunction
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(requestContext, name, step=60):\n start = int(epoch(requestContext['startTime']))\n end = int(epoch(requestContext['endTime']))\n delta = timedelta(seconds=step)\n when = requestContext['startTime']\n values = []\n while when < requestContext['endTime']:\n values.append(epoch(when))\n when += delta\n series = TimeSeries(name, start, end, step, values)\n series.pathExpression = name\n return [series]\n", "role": "user" }, { "content": "timeFunction", "role": "assistant" } ]
def x(requestContext, name, amplitude=1, step=60): delta = timedelta(seconds=step) when = requestContext['startTime'] values = [] while when < requestContext['endTime']: values.append(math.sin(epoch(when)) * amplitude) when += delta series = TimeSeries(name, int(epoch(requestContext['startTime'])), int( epoch(requestContext['endTime'])), step, values) series.pathExpression = 'sin({0})'.format(name) return [series]
sinFunction
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(requestContext, name, amplitude=1, step=60):\n delta = timedelta(seconds=step)\n when = requestContext['startTime']\n values = []\n while when < requestContext['endTime']:\n values.append(math.sin(epoch(when)) * amplitude)\n when += delta\n series = TimeSeries(name, int(epoch(requestContext['startTime'])), int(\n epoch(requestContext['endTime'])), step, values)\n series.pathExpression = 'sin({0})'.format(name)\n return [series]\n", "role": "user" }, { "content": "sinFunction", "role": "assistant" } ]
def x(requestContext, name, step=60): delta = timedelta(seconds=step) when = requestContext['startTime'] values = [] current = 0 while when < requestContext['endTime']: values.append(current) current += random.random() - 0.5 when += delta return [TimeSeries(name, int(epoch(requestContext['startTime'])), int( epoch(requestContext['endTime'])), step, values)]
randomWalkFunction
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(requestContext, name, step=60):\n delta = timedelta(seconds=step)\n when = requestContext['startTime']\n values = []\n current = 0\n while when < requestContext['endTime']:\n values.append(current)\n current += random.random() - 0.5\n when += delta\n return [TimeSeries(name, int(epoch(requestContext['startTime'])), int(\n epoch(requestContext['endTime'])), step, values)]\n", "role": "user" }, { "content": "randomWalkFunction", "role": "assistant" } ]
def x(self, model, checks=[]): custom = [check_partial(reaction_id_check, frozenset(r.id for r in model.reactions))] super(Medium, self).validate(model=model, checks=checks + custom)
validate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, model, checks=[]):\n custom = [check_partial(reaction_id_check, frozenset(r.id for r in\n model.reactions))]\n super(Medium, self).validate(model=model, checks=checks + custom)\n", "role": "user" }, { "content": "validate", "role": "assistant" } ]
def x(self, model): model.medium = {row.exchange: row.uptake for row in self.data. itertuples(index=False)}
apply
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, model):\n model.medium = {row.exchange: row.uptake for row in self.data.\n itertuples(index=False)}\n", "role": "user" }, { "content": "apply", "role": "assistant" } ]
def x(meta): meta['timestamp'] = datetime.utcnow().isoformat(' ') meta['platform'] = platform.system() meta['release'] = platform.release() meta['python'] = platform.python_version() meta['packages'] = get_pkg_info('memote')
add_environment_information
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(meta):\n meta['timestamp'] = datetime.utcnow().isoformat(' ')\n meta['platform'] = platform.system()\n meta['release'] = platform.release()\n meta['python'] = platform.python_version()\n meta['packages'] = get_pkg_info('memote')\n", "role": "user" }, { "content": "add_environment_information", "role": "assistant" } ]
def x(rxn): element_dist = defaultdict() for met in rxn.metabolites: if met.compartment not in element_dist: element_dist[met.compartment] = {k: (v * rxn.metabolites[met]) for k, v in iteritems(met.elements)} else: x = {k: (v * rxn.metabolites[met]) for k, v in iteritems(met. elements)} y = element_dist[met.compartment] element_dist[met.compartment] = {k: (x.get(k, 0) + y.get(k, 0)) for k in set(x) | set(y)} delta_dict = defaultdict() for elements in itervalues(element_dist): delta_dict.update(elements) delta_dict = {k: abs(v) for k, v in iteritems(delta_dict) if v != 0} return delta_dict
find_transported_elements
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(rxn):\n element_dist = defaultdict()\n for met in rxn.metabolites:\n if met.compartment not in element_dist:\n element_dist[met.compartment] = {k: (v * rxn.metabolites[met]) for\n k, v in iteritems(met.elements)}\n else:\n x = {k: (v * rxn.metabolites[met]) for k, v in iteritems(met.\n elements)}\n y = element_dist[met.compartment]\n element_dist[met.compartment] = {k: (x.get(k, 0) + y.get(k, 0)) for\n k in set(x) | set(y)}\n delta_dict = defaultdict()\n for elements in itervalues(element_dist):\n delta_dict.update(elements)\n delta_dict = {k: abs(v) for k, v in iteritems(delta_dict) if v != 0}\n return delta_dict\n", "role": "user" }, { "content": "find_transported_elements", "role": "assistant" } ]
def x(model): transport_reactions = [] transport_rxn_candidates = set(model.reactions) - set(model.boundary ) - set(find_biomass_reaction(model)) transport_rxn_candidates = set([rxn for rxn in transport_rxn_candidates if len(rxn.compartments) >= 2]) sbo_matches = set([rxn for rxn in transport_rxn_candidates if rxn. annotation is not None and 'sbo' in rxn.annotation and rxn. annotation['sbo'] in TRANSPORT_RXN_SBO_TERMS]) if len(sbo_matches) > 0: transport_reactions += list(sbo_matches) for rxn in transport_rxn_candidates: rxn_mets = set([met.formula for met in rxn.metabolites]) if None not in rxn_mets and len(rxn_mets) != 0: if is_transport_reaction_formulae(rxn): transport_reactions.append(rxn) elif is_transport_reaction_annotations(rxn): transport_reactions.append(rxn) return set(transport_reactions)
find_transport_reactions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n transport_reactions = []\n transport_rxn_candidates = set(model.reactions) - set(model.boundary\n ) - set(find_biomass_reaction(model))\n transport_rxn_candidates = set([rxn for rxn in transport_rxn_candidates if\n len(rxn.compartments) >= 2])\n sbo_matches = set([rxn for rxn in transport_rxn_candidates if rxn.\n annotation is not None and 'sbo' in rxn.annotation and rxn.\n annotation['sbo'] in TRANSPORT_RXN_SBO_TERMS])\n if len(sbo_matches) > 0:\n transport_reactions += list(sbo_matches)\n for rxn in transport_rxn_candidates:\n rxn_mets = set([met.formula for met in rxn.metabolites])\n if None not in rxn_mets and len(rxn_mets) != 0:\n if is_transport_reaction_formulae(rxn):\n transport_reactions.append(rxn)\n elif is_transport_reaction_annotations(rxn):\n transport_reactions.append(rxn)\n return set(transport_reactions)\n", "role": "user" }, { "content": "find_transport_reactions", "role": "assistant" } ]
def x(rxn): rxn_reactants = set([met.formula for met in rxn.reactants]) rxn_products = set([met.formula for met in rxn.products]) transported_mets = [formula for formula in rxn_reactants if formula in rxn_products] delta_dicts = find_transported_elements(rxn) non_zero_array = [v for k, v in iteritems(delta_dicts) if v != 0] if set(transported_mets) != set('H') and list(delta_dicts.keys()) == ['H']: pass elif sum(non_zero_array): return True
is_transport_reaction_formulae
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(rxn):\n rxn_reactants = set([met.formula for met in rxn.reactants])\n rxn_products = set([met.formula for met in rxn.products])\n transported_mets = [formula for formula in rxn_reactants if formula in\n rxn_products]\n delta_dicts = find_transported_elements(rxn)\n non_zero_array = [v for k, v in iteritems(delta_dicts) if v != 0]\n if set(transported_mets) != set('H') and list(delta_dicts.keys()) == ['H']:\n pass\n elif sum(non_zero_array):\n return True\n", "role": "user" }, { "content": "is_transport_reaction_formulae", "role": "assistant" } ]
def x(rxn): reactants = set([(k, tuple(v)) for met in rxn.reactants for k, v in iteritems(met.annotation) if met.id != 'H' and k is not None and k != 'sbo' and v is not None]) products = set([(k, tuple(v)) for met in rxn.products for k, v in iteritems(met.annotation) if met.id != 'H' and k is not None and k != 'sbo' and v is not None]) transported_mets = reactants & products if len(transported_mets) > 0: return True
is_transport_reaction_annotations
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(rxn):\n reactants = set([(k, tuple(v)) for met in rxn.reactants for k, v in\n iteritems(met.annotation) if met.id != 'H' and k is not None and k !=\n 'sbo' and v is not None])\n products = set([(k, tuple(v)) for met in rxn.products for k, v in\n iteritems(met.annotation) if met.id != 'H' and k is not None and k !=\n 'sbo' and v is not None])\n transported_mets = reactants & products\n if len(transported_mets) > 0:\n return True\n", "role": "user" }, { "content": "is_transport_reaction_annotations", "role": "assistant" } ]
def x(model, pair): first = set(find_met_in_model(model, pair[0])) second = set(find_met_in_model(model, pair[1])) hits = list() for rxn in model.reactions: if len(first & set(rxn.reactants)) > 0 and len(second & set(rxn. products)) > 0: hits.append(rxn) elif len(first & set(rxn.products)) > 0 and len(second & set(rxn. reactants)) > 0: hits.append(rxn) return frozenset(hits)
find_converting_reactions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model, pair):\n first = set(find_met_in_model(model, pair[0]))\n second = set(find_met_in_model(model, pair[1]))\n hits = list()\n for rxn in model.reactions:\n if len(first & set(rxn.reactants)) > 0 and len(second & set(rxn.\n products)) > 0:\n hits.append(rxn)\n elif len(first & set(rxn.products)) > 0 and len(second & set(rxn.\n reactants)) > 0:\n hits.append(rxn)\n return frozenset(hits)\n", "role": "user" }, { "content": "find_converting_reactions", "role": "assistant" } ]
def x(model): sbo_matches = set([rxn for rxn in model.reactions if rxn.annotation is not None and 'sbo' in rxn.annotation and rxn.annotation['sbo'] == 'SBO:0000629']) if len(sbo_matches) > 0: return list(sbo_matches) buzzwords = ['biomass', 'growth', 'bof'] buzzword_matches = set([rxn for rxn in model.reactions if any(string in rxn.id.lower() for string in buzzwords)]) biomass_met = [] for met in model.metabolites: if 'biomass' in met.id.lower( ) or met.name is not None and 'biomass' in met.name.lower(): biomass_met.append(met) if biomass_met == 1: biomass_met_matches = set(biomass_met.reactions) - set(model.boundary) else: biomass_met_matches = set() return list(buzzword_matches | biomass_met_matches)
find_biomass_reaction
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n sbo_matches = set([rxn for rxn in model.reactions if rxn.annotation is not\n None and 'sbo' in rxn.annotation and rxn.annotation['sbo'] ==\n 'SBO:0000629'])\n if len(sbo_matches) > 0:\n return list(sbo_matches)\n buzzwords = ['biomass', 'growth', 'bof']\n buzzword_matches = set([rxn for rxn in model.reactions if any(string in\n rxn.id.lower() for string in buzzwords)])\n biomass_met = []\n for met in model.metabolites:\n if 'biomass' in met.id.lower(\n ) or met.name is not None and 'biomass' in met.name.lower():\n biomass_met.append(met)\n if biomass_met == 1:\n biomass_met_matches = set(biomass_met.reactions) - set(model.boundary)\n else:\n biomass_met_matches = set()\n return list(buzzword_matches | biomass_met_matches)\n", "role": "user" }, { "content": "find_biomass_reaction", "role": "assistant" } ]
def x(model): try: extracellular = find_compartment_id_in_model(model, 'e') except KeyError: extracellular = None return find_boundary_types(model, 'demand', extracellular)
find_demand_reactions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n try:\n extracellular = find_compartment_id_in_model(model, 'e')\n except KeyError:\n extracellular = None\n return find_boundary_types(model, 'demand', extracellular)\n", "role": "user" }, { "content": "find_demand_reactions", "role": "assistant" } ]
def x(model): try: extracellular = find_compartment_id_in_model(model, 'e') except KeyError: extracellular = None return find_boundary_types(model, 'sink', extracellular)
find_sink_reactions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n try:\n extracellular = find_compartment_id_in_model(model, 'e')\n except KeyError:\n extracellular = None\n return find_boundary_types(model, 'sink', extracellular)\n", "role": "user" }, { "content": "find_sink_reactions", "role": "assistant" } ]
def x(model): try: extracellular = find_compartment_id_in_model(model, 'e') except KeyError: extracellular = None return find_boundary_types(model, 'exchange', extracellular)
find_exchange_rxns
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n try:\n extracellular = find_compartment_id_in_model(model, 'e')\n except KeyError:\n extracellular = None\n return find_boundary_types(model, 'exchange', extracellular)\n", "role": "user" }, { "content": "find_exchange_rxns", "role": "assistant" } ]
def x(model, biomass=None): boundary = set(model.boundary) transporters = find_transport_reactions(model) if biomass is None: biomass = set(find_biomass_reaction(model)) return boundary | transporters | biomass
find_interchange_biomass_reactions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model, biomass=None):\n boundary = set(model.boundary)\n transporters = find_transport_reactions(model)\n if biomass is None:\n biomass = set(find_biomass_reaction(model))\n return boundary | transporters | biomass\n", "role": "user" }, { "content": "find_interchange_biomass_reactions", "role": "assistant" } ]
def x(model, rxn_id, direction='max', single_value=True): model.objective = model.reactions.get_by_id(rxn_id) model.objective_direction = direction if single_value: try: return model.slim_optimize() except Infeasible: return np.nan else: try: solution = model.optimize() return solution except Infeasible: return np.nan
run_fba
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model, rxn_id, direction='max', single_value=True):\n model.objective = model.reactions.get_by_id(rxn_id)\n model.objective_direction = direction\n if single_value:\n try:\n return model.slim_optimize()\n except Infeasible:\n return np.nan\n else:\n try:\n solution = model.optimize()\n return solution\n except Infeasible:\n return np.nan\n", "role": "user" }, { "content": "run_fba", "role": "assistant" } ]
def x(model): for rxn in model.reactions: if rxn.reversibility: rxn.bounds = -1, 1 else: rxn.bounds = 0, 1 for boundary in model.boundary: boundary.bounds = 0, 0
close_boundaries_sensibly
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n for rxn in model.reactions:\n if rxn.reversibility:\n rxn.bounds = -1, 1\n else:\n rxn.bounds = 0, 1\n for boundary in model.boundary:\n boundary.bounds = 0, 0\n", "role": "user" }, { "content": "close_boundaries_sensibly", "role": "assistant" } ]
def x(model, compartment_id): return [met for met in model.metabolites if met.compartment == compartment_id]
metabolites_per_compartment
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model, compartment_id):\n return [met for met in model.metabolites if met.compartment ==\n compartment_id]\n", "role": "user" }, { "content": "metabolites_per_compartment", "role": "assistant" } ]
def x(model): candidate, second = sorted(((c, len(metabolites_per_compartment(model, c))) for c in model.compartments), reverse=True, key=itemgetter(1))[:2] if candidate[1] == second[1]: raise RuntimeError( 'There is a tie for the largest compartment. Compartment {} and {} have equal amounts of metabolites.' .format(candidate[0], second[0])) else: return candidate[0]
largest_compartment_id_met
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n candidate, second = sorted(((c, len(metabolites_per_compartment(model,\n c))) for c in model.compartments), reverse=True, key=itemgetter(1))[:2]\n if candidate[1] == second[1]:\n raise RuntimeError(\n 'There is a tie for the largest compartment. Compartment {} and {} have equal amounts of metabolites.'\n .format(candidate[0], second[0]))\n else:\n return candidate[0]\n", "role": "user" }, { "content": "largest_compartment_id_met", "role": "assistant" } ]
def x(model, compartment_id): if compartment_id not in COMPARTMENT_SHORTLIST.keys(): raise KeyError( '{} is not in the COMPARTMENT_SHORTLIST! Make sure you typed the ID correctly, if yes, update the shortlist manually.' .format(compartment_id)) if len(model.compartments) == 0: raise KeyError( 'It was not possible to identify the compartment {}, since the model has no compartments at all.' .format(COMPARTMENT_SHORTLIST[compartment_id][0])) if compartment_id in model.compartments.keys(): return compartment_id for name in COMPARTMENT_SHORTLIST[compartment_id]: for c_id, c_name in model.compartments.items(): if c_name.lower() == name: return c_id if compartment_id == 'c': return largest_compartment_id_met(model)
find_compartment_id_in_model
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model, compartment_id):\n if compartment_id not in COMPARTMENT_SHORTLIST.keys():\n raise KeyError(\n '{} is not in the COMPARTMENT_SHORTLIST! Make sure you typed the ID correctly, if yes, update the shortlist manually.'\n .format(compartment_id))\n if len(model.compartments) == 0:\n raise KeyError(\n 'It was not possible to identify the compartment {}, since the model has no compartments at all.'\n .format(COMPARTMENT_SHORTLIST[compartment_id][0]))\n if compartment_id in model.compartments.keys():\n return compartment_id\n for name in COMPARTMENT_SHORTLIST[compartment_id]:\n for c_id, c_name in model.compartments.items():\n if c_name.lower() == name:\n return c_id\n if compartment_id == 'c':\n return largest_compartment_id_met(model)\n", "role": "user" }, { "content": "find_compartment_id_in_model", "role": "assistant" } ]
def x(model, mnx_id, compartment_id=None): def x(annotation): query_values = set(utils.flatten(annotation.values())) ref_values = set(utils.flatten(METANETX_SHORTLIST[mnx_id])) return query_values & ref_values if mnx_id not in METANETX_SHORTLIST.columns: raise ValueError( '{} is not in the MetaNetX Shortlist! Make sure you typed the ID correctly, if yes, update the shortlist by updating and re-running the script generate_mnx_shortlists.py.' .format(mnx_id)) candidates = [] regex = re.compile('^{}(_[a-zA-Z0-9]+)?$'.format(mnx_id)) if model.metabolites.query(regex): candidates = model.metabolites.query(regex) elif model.metabolites.query(compare_annotation, attribute='annotation'): candidates = model.metabolites.query(compare_annotation, attribute= 'annotation') else: for value in METANETX_SHORTLIST[mnx_id]: if value: for ident in value: regex = re.compile('^{}(_[a-zA-Z0-9]+)?$'.format(ident)) if model.metabolites.query(regex, attribute='id'): candidates.extend(model.metabolites.query(regex, attribute='id')) if compartment_id is None: print('compartment_id = None?') return candidates else: candidates_in_compartment = [cand for cand in candidates if cand. compartment == compartment_id] if len(candidates_in_compartment) == 0: raise RuntimeError( 'It was not possible to identify any metabolite in compartment {} corresponding to the following MetaNetX identifier: {}.Make sure that a cross-reference to this ID in the MetaNetX Database exists for your identifier namespace.' .format(compartment_id, mnx_id)) elif len(candidates_in_compartment) > 1: raise RuntimeError( 'It was not possible to uniquely identify a single metabolite in compartment {} that corresponds to the following MetaNetX identifier: {}.Instead these candidates were found: {}.Check that metabolite compartment tags are correct. Consider switching to a namespace scheme where identifiers are truly unique.' .format(compartment_id, mnx_id, utils.get_ids( candidates_in_compartment))) else: return candidates_in_compartment
find_met_in_model
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model, mnx_id, compartment_id=None):\n\n def x(annotation):\n query_values = set(utils.flatten(annotation.values()))\n ref_values = set(utils.flatten(METANETX_SHORTLIST[mnx_id]))\n return query_values & ref_values\n if mnx_id not in METANETX_SHORTLIST.columns:\n raise ValueError(\n '{} is not in the MetaNetX Shortlist! Make sure you typed the ID correctly, if yes, update the shortlist by updating and re-running the script generate_mnx_shortlists.py.'\n .format(mnx_id))\n candidates = []\n regex = re.compile('^{}(_[a-zA-Z0-9]+)?$'.format(mnx_id))\n if model.metabolites.query(regex):\n candidates = model.metabolites.query(regex)\n elif model.metabolites.query(compare_annotation, attribute='annotation'):\n candidates = model.metabolites.query(compare_annotation, attribute=\n 'annotation')\n else:\n for value in METANETX_SHORTLIST[mnx_id]:\n if value:\n for ident in value:\n regex = re.compile('^{}(_[a-zA-Z0-9]+)?$'.format(ident))\n if model.metabolites.query(regex, attribute='id'):\n candidates.extend(model.metabolites.query(regex,\n attribute='id'))\n if compartment_id is None:\n print('compartment_id = None?')\n return candidates\n else:\n candidates_in_compartment = [cand for cand in candidates if cand.\n compartment == compartment_id]\n if len(candidates_in_compartment) == 0:\n raise RuntimeError(\n 'It was not possible to identify any metabolite in compartment {} corresponding to the following MetaNetX identifier: {}.Make sure that a cross-reference to this ID in the MetaNetX Database exists for your identifier namespace.'\n .format(compartment_id, mnx_id))\n elif len(candidates_in_compartment) > 1:\n raise RuntimeError(\n 'It was not possible to uniquely identify a single metabolite in compartment {} that corresponds to the following MetaNetX identifier: {}.Instead these candidates were found: {}.Check that metabolite compartment tags are correct. Consider switching to a namespace scheme where identifiers are truly unique.'\n .format(compartment_id, mnx_id, utils.get_ids(\n candidates_in_compartment)))\n else:\n return candidates_in_compartment\n", "role": "user" }, { "content": "find_met_in_model", "role": "assistant" } ]
def x(model): lower_bounds = np.asarray([rxn.lower_bound for rxn in model.reactions], dtype=float) upper_bounds = np.asarray([rxn.upper_bound for rxn in model.reactions], dtype=float) lower_bound = np.nanmedian(lower_bounds[lower_bounds != 0.0]) upper_bound = np.nanmedian(upper_bounds[upper_bounds != 0.0]) if np.isnan(lower_bound): LOGGER.warning('Could not identify a median lower bound.') lower_bound = -1000.0 if np.isnan(upper_bound): LOGGER.warning('Could not identify a median upper bound.') upper_bound = 1000.0 return lower_bound, upper_bound
find_bounds
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n lower_bounds = np.asarray([rxn.lower_bound for rxn in model.reactions],\n dtype=float)\n upper_bounds = np.asarray([rxn.upper_bound for rxn in model.reactions],\n dtype=float)\n lower_bound = np.nanmedian(lower_bounds[lower_bounds != 0.0])\n upper_bound = np.nanmedian(upper_bounds[upper_bounds != 0.0])\n if np.isnan(lower_bound):\n LOGGER.warning('Could not identify a median lower bound.')\n lower_bound = -1000.0\n if np.isnan(upper_bound):\n LOGGER.warning('Could not identify a median upper bound.')\n upper_bound = 1000.0\n return lower_bound, upper_bound\n", "role": "user" }, { "content": "find_bounds", "role": "assistant" } ]
def x(self): return self._template.safe_substitute(report_type=self._report_type, results=self.render_json())
render_html
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return self._template.safe_substitute(report_type=self._report_type,\n results=self.render_json())\n", "role": "user" }, { "content": "render_html", "role": "assistant" } ]
def x(self): cases = self.get_configured_tests() | set(self.result.cases) scores = DataFrame({'score': 0.0, 'max': 1.0}, index=sorted(cases)) self.result.setdefault('score', dict()) self.result['score']['sections'] = list() for test, result in iteritems(self.result.cases): metric = result['metric'] if hasattr(metric, 'items'): result['score'] = test_score = dict() total = 0.0 for key, value in iteritems(metric): value = 1.0 - value total += value test_score[key] = value if len(metric) == 0: metric = 0.0 else: metric = total / len(metric) else: metric = 1.0 - metric scores.at[test, 'score'] = metric scores.loc[test, :] *= self.config['weights'].get(test, 1.0) score = 0.0 maximum = 0.0 for section_id, card in iteritems(self.config['cards']['scored'][ 'sections']): cases = card.get('cases', None) if cases is None: continue card_score = scores.loc[cases, 'score'].sum() card_total = scores.loc[cases, 'max'].sum() section_score = {'section': section_id, 'score': card_score / card_total} self.result['score']['sections'].append(section_score) weight = card.get('weight', 1.0) score += card_score * weight maximum += card_total * weight self.result['score']['total_score'] = score / maximum
compute_score
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n cases = self.get_configured_tests() | set(self.result.cases)\n scores = DataFrame({'score': 0.0, 'max': 1.0}, index=sorted(cases))\n self.result.setdefault('score', dict())\n self.result['score']['sections'] = list()\n for test, result in iteritems(self.result.cases):\n metric = result['metric']\n if hasattr(metric, 'items'):\n result['score'] = test_score = dict()\n total = 0.0\n for key, value in iteritems(metric):\n value = 1.0 - value\n total += value\n test_score[key] = value\n if len(metric) == 0:\n metric = 0.0\n else:\n metric = total / len(metric)\n else:\n metric = 1.0 - metric\n scores.at[test, 'score'] = metric\n scores.loc[test, :] *= self.config['weights'].get(test, 1.0)\n score = 0.0\n maximum = 0.0\n for section_id, card in iteritems(self.config['cards']['scored'][\n 'sections']):\n cases = card.get('cases', None)\n if cases is None:\n continue\n card_score = scores.loc[cases, 'score'].sum()\n card_total = scores.loc[cases, 'max'].sum()\n section_score = {'section': section_id, 'score': card_score /\n card_total}\n self.result['score']['sections'].append(section_score)\n weight = card.get('weight', 1.0)\n score += card_score * weight\n maximum += card_total * weight\n self.result['score']['total_score'] = score / maximum\n", "role": "user" }, { "content": "compute_score", "role": "assistant" } ]
def x(model, components): return [elem for elem in getattr(model, components) if elem.annotation is None or 'sbo' not in elem.annotation]
find_components_without_sbo_terms
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model, components):\n return [elem for elem in getattr(model, components) if elem.annotation is\n None or 'sbo' not in elem.annotation]\n", "role": "user" }, { "content": "find_components_without_sbo_terms", "role": "assistant" } ]
def x(items, term): if isinstance(term, list): return [elem for elem in items if elem.annotation is None or 'sbo' not in elem.annotation or not any(i in elem.annotation['sbo'] for i in term)] else: return [elem for elem in items if elem.annotation is None or 'sbo' not in elem.annotation or term not in elem.annotation['sbo']]
check_component_for_specific_sbo_term
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(items, term):\n if isinstance(term, list):\n return [elem for elem in items if elem.annotation is None or 'sbo'\n not in elem.annotation or not any(i in elem.annotation['sbo'] for\n i in term)]\n else:\n return [elem for elem in items if elem.annotation is None or 'sbo'\n not in elem.annotation or term not in elem.annotation['sbo']]\n", "role": "user" }, { "content": "check_component_for_specific_sbo_term", "role": "assistant" } ]
def x(compounds_identifiers): return min((c for c in compounds_identifiers if c.startswith('C')), key =lambda c: int(c[1:]))
get_smallest_compound_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(compounds_identifiers):\n return min((c for c in compounds_identifiers if c.startswith('C')), key\n =lambda c: int(c[1:]))\n", "role": "user" }, { "content": "get_smallest_compound_id", "role": "assistant" } ]
def x(metabolite): logger.debug('Looking for KEGG compound identifier for %s.', metabolite.id) kegg_annotation = metabolite.annotation.get('kegg.compound') if kegg_annotation is None: logger.warning('No kegg.compound annotation for metabolite %s.', metabolite.id) return if isinstance(kegg_annotation, string_types ) and kegg_annotation.startswith('C'): return kegg_annotation elif isinstance(kegg_annotation, Iterable): try: return get_smallest_compound_id(kegg_annotation) except ValueError: return logger.warning('No matching kegg.compound annotation for metabolite %s.', metabolite.id) return
map_metabolite2kegg
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(metabolite):\n logger.debug('Looking for KEGG compound identifier for %s.', metabolite.id)\n kegg_annotation = metabolite.annotation.get('kegg.compound')\n if kegg_annotation is None:\n logger.warning('No kegg.compound annotation for metabolite %s.',\n metabolite.id)\n return\n if isinstance(kegg_annotation, string_types\n ) and kegg_annotation.startswith('C'):\n return kegg_annotation\n elif isinstance(kegg_annotation, Iterable):\n try:\n return get_smallest_compound_id(kegg_annotation)\n except ValueError:\n return\n logger.warning('No matching kegg.compound annotation for metabolite %s.',\n metabolite.id)\n return\n", "role": "user" }, { "content": "map_metabolite2kegg", "role": "assistant" } ]
def x(reaction, metabolite_mapping): stoichiometry = defaultdict(float) for met, coef in iteritems(reaction.metabolites): kegg_id = metabolite_mapping.setdefault(met, map_metabolite2kegg(met)) if kegg_id is None: continue stoichiometry[kegg_id] += coef return dict(stoichiometry)
translate_reaction
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(reaction, metabolite_mapping):\n stoichiometry = defaultdict(float)\n for met, coef in iteritems(reaction.metabolites):\n kegg_id = metabolite_mapping.setdefault(met, map_metabolite2kegg(met))\n if kegg_id is None:\n continue\n stoichiometry[kegg_id] += coef\n return dict(stoichiometry)\n", "role": "user" }, { "content": "translate_reaction", "role": "assistant" } ]
def x(reactions): incomplete_mapping = [] problematic_calculation = [] reversibility_indexes = [] unbalanced = [] metabolite_mapping = {} for rxn in reactions: stoich = translate_reaction(rxn, metabolite_mapping) if len(stoich) < len(rxn.metabolites): incomplete_mapping.append(rxn) continue try: if 'C00080' in stoich: del stoich['C00080'] eq_rxn = Reaction(stoich, rxn.id) except KeyError: incomplete_mapping.append(rxn) continue if eq_rxn.check_full_reaction_balancing(): try: ln_rev_index = eq_rxn.reversibility_index() except Exception: problematic_calculation.append(rxn) continue reversibility_indexes.append((rxn, ln_rev_index)) else: unbalanced.append(rxn) reversibility_indexes.sort(key=lambda p: abs(p[1]), reverse=True) return (reversibility_indexes, incomplete_mapping, problematic_calculation, unbalanced)
find_thermodynamic_reversibility_index
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(reactions):\n incomplete_mapping = []\n problematic_calculation = []\n reversibility_indexes = []\n unbalanced = []\n metabolite_mapping = {}\n for rxn in reactions:\n stoich = translate_reaction(rxn, metabolite_mapping)\n if len(stoich) < len(rxn.metabolites):\n incomplete_mapping.append(rxn)\n continue\n try:\n if 'C00080' in stoich:\n del stoich['C00080']\n eq_rxn = Reaction(stoich, rxn.id)\n except KeyError:\n incomplete_mapping.append(rxn)\n continue\n if eq_rxn.check_full_reaction_balancing():\n try:\n ln_rev_index = eq_rxn.reversibility_index()\n except Exception:\n problematic_calculation.append(rxn)\n continue\n reversibility_indexes.append((rxn, ln_rev_index))\n else:\n unbalanced.append(rxn)\n reversibility_indexes.sort(key=lambda p: abs(p[1]), reverse=True)\n return (reversibility_indexes, incomplete_mapping,\n problematic_calculation, unbalanced)\n", "role": "user" }, { "content": "find_thermodynamic_reversibility_index", "role": "assistant" } ]
def x(model): problem = model.problem stoich_trans = problem.Model() internal_rxns = con_helpers.get_internals(model) metabolites = set(met for rxn in internal_rxns for met in rxn.metabolites) LOGGER.info("model '%s' has %d internal reactions", model.id, len( internal_rxns)) LOGGER.info("model '%s' has %d internal metabolites", model.id, len( metabolites)) stoich_trans.add([problem.Variable(m.id, lb=1) for m in metabolites]) stoich_trans.update() con_helpers.add_reaction_constraints(stoich_trans, internal_rxns, problem.Constraint) stoich_trans.objective = problem.Objective(Zero, direction='min', sloppy=True) stoich_trans.objective.set_linear_coefficients({var: (1.0) for var in stoich_trans.variables}) status = stoich_trans.optimize() if status == OPTIMAL: return True elif status == INFEASIBLE: return False else: raise RuntimeError( "Could not determine stoichiometric consistencty. Solver status is '{}' (only optimal or infeasible expected)." .format(status))
check_stoichiometric_consistency
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n problem = model.problem\n stoich_trans = problem.Model()\n internal_rxns = con_helpers.get_internals(model)\n metabolites = set(met for rxn in internal_rxns for met in rxn.metabolites)\n LOGGER.info(\"model '%s' has %d internal reactions\", model.id, len(\n internal_rxns))\n LOGGER.info(\"model '%s' has %d internal metabolites\", model.id, len(\n metabolites))\n stoich_trans.add([problem.Variable(m.id, lb=1) for m in metabolites])\n stoich_trans.update()\n con_helpers.add_reaction_constraints(stoich_trans, internal_rxns,\n problem.Constraint)\n stoich_trans.objective = problem.Objective(Zero, direction='min',\n sloppy=True)\n stoich_trans.objective.set_linear_coefficients({var: (1.0) for var in\n stoich_trans.variables})\n status = stoich_trans.optimize()\n if status == OPTIMAL:\n return True\n elif status == INFEASIBLE:\n return False\n else:\n raise RuntimeError(\n \"Could not determine stoichiometric consistencty. Solver status is '{}' (only optimal or infeasible expected).\"\n .format(status))\n", "role": "user" }, { "content": "check_stoichiometric_consistency", "role": "assistant" } ]
def x(model): problem = model.problem stoich_trans = problem.Model() internal_rxns = con_helpers.get_internals(model) metabolites = set(met for rxn in internal_rxns for met in rxn.metabolites) k_vars = list() for met in metabolites: m_var = problem.Variable(met.id) k_var = problem.Variable('k_{}'.format(met.id), type='binary') k_vars.append(k_var) stoich_trans.add([m_var, k_var]) stoich_trans.add(problem.Constraint(k_var - m_var, ub=0, name= 'switch_{}'.format(met.id))) stoich_trans.update() con_helpers.add_reaction_constraints(stoich_trans, internal_rxns, problem.Constraint) stoich_trans.objective = problem.Objective(Zero, sloppy=True, direction ='max') stoich_trans.objective.set_linear_coefficients({var: (1.0) for var in k_vars}) status = stoich_trans.optimize() if status == OPTIMAL: return set([model.metabolites.get_by_id(var.name[2:]) for var in k_vars if var.primal < 0.8]) else: raise RuntimeError( "Could not compute list of unconserved metabolites. Solver status is '{}' (only optimal expected)." .format(status))
find_unconserved_metabolites
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n problem = model.problem\n stoich_trans = problem.Model()\n internal_rxns = con_helpers.get_internals(model)\n metabolites = set(met for rxn in internal_rxns for met in rxn.metabolites)\n k_vars = list()\n for met in metabolites:\n m_var = problem.Variable(met.id)\n k_var = problem.Variable('k_{}'.format(met.id), type='binary')\n k_vars.append(k_var)\n stoich_trans.add([m_var, k_var])\n stoich_trans.add(problem.Constraint(k_var - m_var, ub=0, name=\n 'switch_{}'.format(met.id)))\n stoich_trans.update()\n con_helpers.add_reaction_constraints(stoich_trans, internal_rxns,\n problem.Constraint)\n stoich_trans.objective = problem.Objective(Zero, sloppy=True, direction\n ='max')\n stoich_trans.objective.set_linear_coefficients({var: (1.0) for var in\n k_vars})\n status = stoich_trans.optimize()\n if status == OPTIMAL:\n return set([model.metabolites.get_by_id(var.name[2:]) for var in\n k_vars if var.primal < 0.8])\n else:\n raise RuntimeError(\n \"Could not compute list of unconserved metabolites. Solver status is '{}' (only optimal expected).\"\n .format(status))\n", "role": "user" }, { "content": "find_unconserved_metabolites", "role": "assistant" } ]
def x(model, atol=1e-13): if check_stoichiometric_consistency(model): return set() Model, Constraint, Variable, Objective = con_helpers.get_interface(model) unconserved_mets = find_unconserved_metabolites(model) LOGGER.info('model has %d unconserved metabolites', len(unconserved_mets)) internal_rxns = con_helpers.get_internals(model) internal_mets = set(met for rxn in internal_rxns for met in rxn.metabolites ) get_id = attrgetter('id') reactions = sorted(internal_rxns, key=get_id) metabolites = sorted(internal_mets, key=get_id) stoich, met_index, rxn_index = con_helpers.stoichiometry_matrix(metabolites , reactions) left_ns = con_helpers.nullspace(stoich.T) left_ns[np.abs(left_ns) < atol] = 0.0 LOGGER.info('nullspace has dimension %d', left_ns.shape[1]) inc_minimal = set() problem, indicators = con_helpers.create_milp_problem(left_ns, metabolites, Model, Variable, Constraint, Objective) LOGGER.debug(str(problem)) cuts = list() for met in unconserved_mets: row = met_index[met] if (left_ns[row] == 0.0).all(): LOGGER.debug('%s: singleton minimal unconservable set', met.id) inc_minimal.add((met,)) continue problem.variables[met.id].lb = 0.001 status = problem.optimize() while status == 'optimal': LOGGER.debug('%s: status %s', met.id, status) LOGGER.debug('sum of all primal values: %f', sum(problem. primal_values.values())) LOGGER.debug('sum of binary indicators: %f', sum(var.primal for var in indicators)) solution = [model.metabolites.get_by_id(var.name[2:]) for var in indicators if var.primal > 0.2] LOGGER.debug('%s: set size %d', met.id, len(solution)) inc_minimal.add(tuple(solution)) if len(solution) == 1: break cuts.append(con_helpers.add_cut(problem, indicators, len( solution) - 1, Constraint)) status = problem.optimize() LOGGER.debug('%s: last status %s', met.id, status) problem.variables[met.id].lb = 0.0 problem.remove(cuts) cuts.clear() return inc_minimal
find_inconsistent_min_stoichiometry
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model, atol=1e-13):\n if check_stoichiometric_consistency(model):\n return set()\n Model, Constraint, Variable, Objective = con_helpers.get_interface(model)\n unconserved_mets = find_unconserved_metabolites(model)\n LOGGER.info('model has %d unconserved metabolites', len(unconserved_mets))\n internal_rxns = con_helpers.get_internals(model)\n internal_mets = set(met for rxn in internal_rxns for met in rxn.metabolites\n )\n get_id = attrgetter('id')\n reactions = sorted(internal_rxns, key=get_id)\n metabolites = sorted(internal_mets, key=get_id)\n stoich, met_index, rxn_index = con_helpers.stoichiometry_matrix(metabolites\n , reactions)\n left_ns = con_helpers.nullspace(stoich.T)\n left_ns[np.abs(left_ns) < atol] = 0.0\n LOGGER.info('nullspace has dimension %d', left_ns.shape[1])\n inc_minimal = set()\n problem, indicators = con_helpers.create_milp_problem(left_ns,\n metabolites, Model, Variable, Constraint, Objective)\n LOGGER.debug(str(problem))\n cuts = list()\n for met in unconserved_mets:\n row = met_index[met]\n if (left_ns[row] == 0.0).all():\n LOGGER.debug('%s: singleton minimal unconservable set', met.id)\n inc_minimal.add((met,))\n continue\n problem.variables[met.id].lb = 0.001\n status = problem.optimize()\n while status == 'optimal':\n LOGGER.debug('%s: status %s', met.id, status)\n LOGGER.debug('sum of all primal values: %f', sum(problem.\n primal_values.values()))\n LOGGER.debug('sum of binary indicators: %f', sum(var.primal for\n var in indicators))\n solution = [model.metabolites.get_by_id(var.name[2:]) for var in\n indicators if var.primal > 0.2]\n LOGGER.debug('%s: set size %d', met.id, len(solution))\n inc_minimal.add(tuple(solution))\n if len(solution) == 1:\n break\n cuts.append(con_helpers.add_cut(problem, indicators, len(\n solution) - 1, Constraint))\n status = problem.optimize()\n LOGGER.debug('%s: last status %s', met.id, status)\n problem.variables[met.id].lb = 0.0\n problem.remove(cuts)\n cuts.clear()\n return inc_minimal\n", "role": "user" }, { "content": "find_inconsistent_min_stoichiometry", "role": "assistant" } ]
def x(model, metabolite_id): main_comp = helpers.find_compartment_id_in_model(model, 'c') met = helpers.find_met_in_model(model, metabolite_id, main_comp)[0] dissipation_rxn = Reaction('Dissipation') if metabolite_id in ['MNXM3', 'MNXM63', 'MNXM51', 'MNXM121', 'MNXM423']: dissipation_rxn.add_metabolites({helpers.find_met_in_model(model, 'MNXM2', main_comp)[0]: -1, helpers.find_met_in_model(model, 'MNXM1', main_comp)[0]: 1, helpers.find_met_in_model(model, 'MNXM9', main_comp)[0]: 1}) elif metabolite_id in ['MNXM6', 'MNXM10']: dissipation_rxn.add_metabolites({helpers.find_met_in_model(model, 'MNXM1', main_comp)[0]: 1}) elif metabolite_id in ['MNXM38', 'MNXM208', 'MNXM191', 'MNXM223', 'MNXM7517', 'MNXM12233', 'MNXM558']: dissipation_rxn.add_metabolites({helpers.find_met_in_model(model, 'MNXM1', main_comp)[0]: 2}) elif metabolite_id == 'MNXM21': dissipation_rxn.add_metabolites({helpers.find_met_in_model(model, 'MNXM2', main_comp)[0]: -1, helpers.find_met_in_model(model, 'MNXM1', main_comp)[0]: 1, helpers.find_met_in_model(model, 'MNXM26', main_comp)[0]: 1}) elif metabolite_id == 'MNXM89557': dissipation_rxn.add_metabolites({helpers.find_met_in_model(model, 'MNXM2', main_comp)[0]: -1, helpers.find_met_in_model(model, 'MNXM1', main_comp)[0]: 2, helpers.find_met_in_model(model, 'MNXM15', main_comp)[0]: 1}) dissipation_product = helpers.find_met_in_model(model, ENERGY_COUPLES[ metabolite_id], main_comp)[0] dissipation_rxn.add_metabolites({met: -1, dissipation_product: 1}) helpers.close_boundaries_sensibly(model) model.add_reactions([dissipation_rxn]) model.objective = dissipation_rxn solution = model.optimize(raise_error=True) if solution.objective_value > 0.0: return solution.fluxes[solution.fluxes.abs() > 0.0].index.drop([ 'Dissipation']).tolist() else: return []
detect_energy_generating_cycles
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model, metabolite_id):\n main_comp = helpers.find_compartment_id_in_model(model, 'c')\n met = helpers.find_met_in_model(model, metabolite_id, main_comp)[0]\n dissipation_rxn = Reaction('Dissipation')\n if metabolite_id in ['MNXM3', 'MNXM63', 'MNXM51', 'MNXM121', 'MNXM423']:\n dissipation_rxn.add_metabolites({helpers.find_met_in_model(model,\n 'MNXM2', main_comp)[0]: -1, helpers.find_met_in_model(model,\n 'MNXM1', main_comp)[0]: 1, helpers.find_met_in_model(model,\n 'MNXM9', main_comp)[0]: 1})\n elif metabolite_id in ['MNXM6', 'MNXM10']:\n dissipation_rxn.add_metabolites({helpers.find_met_in_model(model,\n 'MNXM1', main_comp)[0]: 1})\n elif metabolite_id in ['MNXM38', 'MNXM208', 'MNXM191', 'MNXM223',\n 'MNXM7517', 'MNXM12233', 'MNXM558']:\n dissipation_rxn.add_metabolites({helpers.find_met_in_model(model,\n 'MNXM1', main_comp)[0]: 2})\n elif metabolite_id == 'MNXM21':\n dissipation_rxn.add_metabolites({helpers.find_met_in_model(model,\n 'MNXM2', main_comp)[0]: -1, helpers.find_met_in_model(model,\n 'MNXM1', main_comp)[0]: 1, helpers.find_met_in_model(model,\n 'MNXM26', main_comp)[0]: 1})\n elif metabolite_id == 'MNXM89557':\n dissipation_rxn.add_metabolites({helpers.find_met_in_model(model,\n 'MNXM2', main_comp)[0]: -1, helpers.find_met_in_model(model,\n 'MNXM1', main_comp)[0]: 2, helpers.find_met_in_model(model,\n 'MNXM15', main_comp)[0]: 1})\n dissipation_product = helpers.find_met_in_model(model, ENERGY_COUPLES[\n metabolite_id], main_comp)[0]\n dissipation_rxn.add_metabolites({met: -1, dissipation_product: 1})\n helpers.close_boundaries_sensibly(model)\n model.add_reactions([dissipation_rxn])\n model.objective = dissipation_rxn\n solution = model.optimize(raise_error=True)\n if solution.objective_value > 0.0:\n return solution.fluxes[solution.fluxes.abs() > 0.0].index.drop([\n 'Dissipation']).tolist()\n else:\n return []\n", "role": "user" }, { "content": "detect_energy_generating_cycles", "role": "assistant" } ]
def x(model): helpers.close_boundaries_sensibly(model) fva_result = flux_variability_analysis(model, loopless=False) return fva_result.index[(fva_result['minimum'] <= -1 + TOLERANCE_THRESHOLD) | (fva_result['maximum'] >= 1 - TOLERANCE_THRESHOLD)].tolist()
find_stoichiometrically_balanced_cycles
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n helpers.close_boundaries_sensibly(model)\n fva_result = flux_variability_analysis(model, loopless=False)\n return fva_result.index[(fva_result['minimum'] <= -1 +\n TOLERANCE_THRESHOLD) | (fva_result['maximum'] >= 1 -\n TOLERANCE_THRESHOLD)].tolist()\n", "role": "user" }, { "content": "find_stoichiometrically_balanced_cycles", "role": "assistant" } ]
def x(model): exchange = frozenset(model.exchanges) return [met for met in model.metabolites if len(met.reactions) > 0 and all(not rxn.reversibility and rxn not in exchange and rxn. metabolites[met] < 0 for rxn in met.reactions)]
find_orphans
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n exchange = frozenset(model.exchanges)\n return [met for met in model.metabolites if len(met.reactions) > 0 and\n all(not rxn.reversibility and rxn not in exchange and rxn.\n metabolites[met] < 0 for rxn in met.reactions)]\n", "role": "user" }, { "content": "find_orphans", "role": "assistant" } ]
def x(model): mets_not_produced = list() helpers.open_exchanges(model) for met in model.metabolites: with model: exch = model.add_boundary(met, type='irrex', reaction_id= 'IRREX', lb=0, ub=1000) solution = helpers.run_fba(model, exch.id) if np.isnan(solution) or solution < TOLERANCE_THRESHOLD: mets_not_produced.append(met) return mets_not_produced
find_metabolites_not_produced_with_open_bounds
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n mets_not_produced = list()\n helpers.open_exchanges(model)\n for met in model.metabolites:\n with model:\n exch = model.add_boundary(met, type='irrex', reaction_id=\n 'IRREX', lb=0, ub=1000)\n solution = helpers.run_fba(model, exch.id)\n if np.isnan(solution) or solution < TOLERANCE_THRESHOLD:\n mets_not_produced.append(met)\n return mets_not_produced\n", "role": "user" }, { "content": "find_metabolites_not_produced_with_open_bounds", "role": "assistant" } ]
def x(model): mets_not_consumed = list() helpers.open_exchanges(model) for met in model.metabolites: with model: exch = model.add_boundary(met, type='irrex', reaction_id= 'IRREX', lb=-1000, ub=0) solution = helpers.run_fba(model, exch.id, direction='min') if np.isnan(solution) or abs(solution) < TOLERANCE_THRESHOLD: mets_not_consumed.append(met) return mets_not_consumed
find_metabolites_not_consumed_with_open_bounds
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n mets_not_consumed = list()\n helpers.open_exchanges(model)\n for met in model.metabolites:\n with model:\n exch = model.add_boundary(met, type='irrex', reaction_id=\n 'IRREX', lb=-1000, ub=0)\n solution = helpers.run_fba(model, exch.id, direction='min')\n if np.isnan(solution) or abs(solution) < TOLERANCE_THRESHOLD:\n mets_not_consumed.append(met)\n return mets_not_consumed\n", "role": "user" }, { "content": "find_metabolites_not_consumed_with_open_bounds", "role": "assistant" } ]
def x(model): try: fva_result = flux_variability_analysis(model, fraction_of_optimum=1.0) except Infeasible as err: LOGGER.error( "Failed to find reactions with unbounded flux because '{}'. This may be a bug." .format(err)) raise Infeasible( 'It was not possible to run flux variability analysis on the model. Make sure that the model can be solved! Check if the constraints are not too strict.' ) conditionally_blocked = fva_result.loc[fva_result.abs().max(axis=1) < TOLERANCE_THRESHOLD].index.tolist() small, large = helpers.find_bounds(model) unlimited_flux = fva_result.loc[np.isclose(fva_result['maximum'], large, atol=TOLERANCE_THRESHOLD) | (fva_result['maximum'] > large) | np. isclose(fva_result['minimum'], small, atol=TOLERANCE_THRESHOLD) | ( fva_result['minimum'] < small)].index.tolist() try: fraction = len(unlimited_flux) / (len(model.reactions) - len( conditionally_blocked)) except ZeroDivisionError: LOGGER.error( 'Division by Zero! Failed to calculate the fraction of unbounded reactions. Does this model have any reactions at all?' ) raise ZeroDivisionError( "It was not possible to calculate the fraction of unbounded reactions to un-blocked reactions. This may be becausethe model doesn't have any reactions at all or that none of the reactions can carry a flux larger than zero!" ) return unlimited_flux, fraction, conditionally_blocked
find_reactions_with_unbounded_flux_default_condition
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n try:\n fva_result = flux_variability_analysis(model, fraction_of_optimum=1.0)\n except Infeasible as err:\n LOGGER.error(\n \"Failed to find reactions with unbounded flux because '{}'. This may be a bug.\"\n .format(err))\n raise Infeasible(\n 'It was not possible to run flux variability analysis on the model. Make sure that the model can be solved! Check if the constraints are not too strict.'\n )\n conditionally_blocked = fva_result.loc[fva_result.abs().max(axis=1) <\n TOLERANCE_THRESHOLD].index.tolist()\n small, large = helpers.find_bounds(model)\n unlimited_flux = fva_result.loc[np.isclose(fva_result['maximum'], large,\n atol=TOLERANCE_THRESHOLD) | (fva_result['maximum'] > large) | np.\n isclose(fva_result['minimum'], small, atol=TOLERANCE_THRESHOLD) | (\n fva_result['minimum'] < small)].index.tolist()\n try:\n fraction = len(unlimited_flux) / (len(model.reactions) - len(\n conditionally_blocked))\n except ZeroDivisionError:\n LOGGER.error(\n 'Division by Zero! Failed to calculate the fraction of unbounded reactions. Does this model have any reactions at all?'\n )\n raise ZeroDivisionError(\n \"It was not possible to calculate the fraction of unbounded reactions to un-blocked reactions. This may be becausethe model doesn't have any reactions at all or that none of the reactions can carry a flux larger than zero!\"\n )\n return unlimited_flux, fraction, conditionally_blocked\n", "role": "user" }, { "content": "find_reactions_with_unbounded_flux_default_condition", "role": "assistant" } ]
def x(filename, dtype_conversion=None): if dtype_conversion is None: dtype_conversion = {} name, ext = filename.split('.', 1) ext = ext.lower() dtype_conversion['comment'] = str if 'csv' in ext: df = pd.read_csv(filename, dtype=dtype_conversion, encoding='utf-8') elif 'tsv' in ext: df = pd.read_table(filename, dtype=dtype_conversion, encoding='utf-8') elif 'xls' in ext or 'xlsx' in ext: df = pd.read_excel(filename, dtype=dtype_conversion, encoding='utf-8') else: raise ValueError("Unknown file format '{}'.".format(ext)) return df
read_tabular
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(filename, dtype_conversion=None):\n if dtype_conversion is None:\n dtype_conversion = {}\n name, ext = filename.split('.', 1)\n ext = ext.lower()\n dtype_conversion['comment'] = str\n if 'csv' in ext:\n df = pd.read_csv(filename, dtype=dtype_conversion, encoding='utf-8')\n elif 'tsv' in ext:\n df = pd.read_table(filename, dtype=dtype_conversion, encoding='utf-8')\n elif 'xls' in ext or 'xlsx' in ext:\n df = pd.read_excel(filename, dtype=dtype_conversion, encoding='utf-8')\n else:\n raise ValueError(\"Unknown file format '{}'.\".format(ext))\n return df\n", "role": "user" }, { "content": "read_tabular", "role": "assistant" } ]
def x(model, filename, pytest_args, exclusive, skip, solver, experimental, custom_tests, custom_config): model_obj, sbml_ver, notifications = api.validate_model(model) if model_obj is None: LOGGER.critical( 'The model could not be loaded due to the following SBML errors.') utils.stdout_notifications(notifications) api.validation_report(model, notifications, filename) sys.exit(1) if not any(a.startswith('--tb') for a in pytest_args): pytest_args = ['--tb', 'no'] + pytest_args pytest_args.extend(custom_tests) config = ReportConfiguration.load() for custom in custom_config: config.merge(ReportConfiguration.load(custom)) model_obj.solver = solver _, results = api.test_model(model_obj, sbml_version=sbml_ver, results= True, pytest_args=pytest_args, skip=skip, exclusive=exclusive, experimental=experimental) with open(filename, 'w', encoding='utf-8') as file_handle: LOGGER.info("Writing snapshot report to '%s'.", filename) file_handle.write(api.snapshot_report(results, config))
snapshot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model, filename, pytest_args, exclusive, skip, solver, experimental,\n custom_tests, custom_config):\n model_obj, sbml_ver, notifications = api.validate_model(model)\n if model_obj is None:\n LOGGER.critical(\n 'The model could not be loaded due to the following SBML errors.')\n utils.stdout_notifications(notifications)\n api.validation_report(model, notifications, filename)\n sys.exit(1)\n if not any(a.startswith('--tb') for a in pytest_args):\n pytest_args = ['--tb', 'no'] + pytest_args\n pytest_args.extend(custom_tests)\n config = ReportConfiguration.load()\n for custom in custom_config:\n config.merge(ReportConfiguration.load(custom))\n model_obj.solver = solver\n _, results = api.test_model(model_obj, sbml_version=sbml_ver, results=\n True, pytest_args=pytest_args, skip=skip, exclusive=exclusive,\n experimental=experimental)\n with open(filename, 'w', encoding='utf-8') as file_handle:\n LOGGER.info(\"Writing snapshot report to '%s'.\", filename)\n file_handle.write(api.snapshot_report(results, config))\n", "role": "user" }, { "content": "snapshot", "role": "assistant" } ]
def x(location, model, filename, deployment, custom_config): callbacks.git_installed() LOGGER.info('Initialising history report generation.') if location is None: raise click.BadParameter("No 'location' given or configured.") try: repo = git.Repo() except git.InvalidGitRepositoryError: LOGGER.critical( "The history report requires a git repository in order to check the model's commit history." ) sys.exit(1) LOGGER.info('Obtaining history of results from the deployment branch {}.' .format(deployment)) repo.git.checkout(deployment) try: manager = managers.SQLResultManager(repository=repo, location=location) except (AttributeError, ArgumentError): manager = managers.RepoResultManager(repository=repo, location=location ) config = ReportConfiguration.load() for custom in custom_config: config.merge(ReportConfiguration.load(custom)) LOGGER.info('Tracing the commit history.') history = managers.HistoryManager(repository=repo, manager=manager) history.load_history(model, skip={deployment}) LOGGER.info('Composing the history report.') report = api.history_report(history, config=config) with open(filename, 'w', encoding='utf-8') as file_handle: file_handle.write(report)
history
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(location, model, filename, deployment, custom_config):\n callbacks.git_installed()\n LOGGER.info('Initialising history report generation.')\n if location is None:\n raise click.BadParameter(\"No 'location' given or configured.\")\n try:\n repo = git.Repo()\n except git.InvalidGitRepositoryError:\n LOGGER.critical(\n \"The history report requires a git repository in order to check the model's commit history.\"\n )\n sys.exit(1)\n LOGGER.info('Obtaining history of results from the deployment branch {}.'\n .format(deployment))\n repo.git.checkout(deployment)\n try:\n manager = managers.SQLResultManager(repository=repo, location=location)\n except (AttributeError, ArgumentError):\n manager = managers.RepoResultManager(repository=repo, location=location\n )\n config = ReportConfiguration.load()\n for custom in custom_config:\n config.merge(ReportConfiguration.load(custom))\n LOGGER.info('Tracing the commit history.')\n history = managers.HistoryManager(repository=repo, manager=manager)\n history.load_history(model, skip={deployment})\n LOGGER.info('Composing the history report.')\n report = api.history_report(history, config=config)\n with open(filename, 'w', encoding='utf-8') as file_handle:\n file_handle.write(report)\n", "role": "user" }, { "content": "history", "role": "assistant" } ]
def x(models, filename, pytest_args, exclusive, skip, solver, experimental, custom_tests, custom_config): if not any(a.startswith('--tb') for a in pytest_args): pytest_args = ['--tb', 'no'] + pytest_args pytest_args.extend(custom_tests) config = ReportConfiguration.load() for custom in custom_config: config.merge(ReportConfiguration.load(custom)) diff_results = dict() model_and_model_ver_tuple = list() for model_path in models: try: model_filename = os.path.basename(model_path) diff_results.setdefault(model_filename, dict()) model, model_ver, notifications = api.validate_model(model_path) if model is None: head, tail = os.path.split(filename) report_path = os.path.join(head, '{}_structural_report.html'.format(model_filename)) api.validation_report(model_path, notifications, report_path) LOGGER.critical( 'The model {} could not be loaded due to SBML errors reported in {}.' .format(model_filename, report_path)) continue model.solver = solver model_and_model_ver_tuple.append((model, model_ver)) except (IOError, SBMLError): LOGGER.debug(exc_info=True) LOGGER.warning( "An error occurred while loading the model '%s'. Skipping.", model_filename) if len(model_and_model_ver_tuple) < 2: LOGGER.critical( 'Out of the %d provided models only %d could be loaded. Please, check if the models that could not be loaded are valid SBML. Aborting.' , len(models), len(model_and_model_ver_tuple)) sys.exit(1) partial_test_diff = partial(_test_diff, pytest_args=pytest_args, skip= skip, exclusive=exclusive, experimental=experimental) pool = Pool(min(len(models), cpu_count())) results = pool.map(partial_test_diff, model_and_model_ver_tuple) for model_path, result in zip(models, results): model_filename = os.path.basename(model_path) diff_results[model_filename] = result with open(filename, 'w', encoding='utf-8') as file_handle: LOGGER.info("Writing diff report to '%s'.", filename) file_handle.write(api.diff_report(diff_results, config))
diff
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(models, filename, pytest_args, exclusive, skip, solver, experimental,\n custom_tests, custom_config):\n if not any(a.startswith('--tb') for a in pytest_args):\n pytest_args = ['--tb', 'no'] + pytest_args\n pytest_args.extend(custom_tests)\n config = ReportConfiguration.load()\n for custom in custom_config:\n config.merge(ReportConfiguration.load(custom))\n diff_results = dict()\n model_and_model_ver_tuple = list()\n for model_path in models:\n try:\n model_filename = os.path.basename(model_path)\n diff_results.setdefault(model_filename, dict())\n model, model_ver, notifications = api.validate_model(model_path)\n if model is None:\n head, tail = os.path.split(filename)\n report_path = os.path.join(head,\n '{}_structural_report.html'.format(model_filename))\n api.validation_report(model_path, notifications, report_path)\n LOGGER.critical(\n 'The model {} could not be loaded due to SBML errors reported in {}.'\n .format(model_filename, report_path))\n continue\n model.solver = solver\n model_and_model_ver_tuple.append((model, model_ver))\n except (IOError, SBMLError):\n LOGGER.debug(exc_info=True)\n LOGGER.warning(\n \"An error occurred while loading the model '%s'. Skipping.\",\n model_filename)\n if len(model_and_model_ver_tuple) < 2:\n LOGGER.critical(\n 'Out of the %d provided models only %d could be loaded. Please, check if the models that could not be loaded are valid SBML. Aborting.'\n , len(models), len(model_and_model_ver_tuple))\n sys.exit(1)\n partial_test_diff = partial(_test_diff, pytest_args=pytest_args, skip=\n skip, exclusive=exclusive, experimental=experimental)\n pool = Pool(min(len(models), cpu_count()))\n results = pool.map(partial_test_diff, model_and_model_ver_tuple)\n for model_path, result in zip(models, results):\n model_filename = os.path.basename(model_path)\n diff_results[model_filename] = result\n with open(filename, 'w', encoding='utf-8') as file_handle:\n LOGGER.info(\"Writing diff report to '%s'.\", filename)\n file_handle.write(api.diff_report(diff_results, config))\n", "role": "user" }, { "content": "diff", "role": "assistant" } ]
def x(self, model, skip): self._history = dict() self._history['commits'] = commits = dict() self._history['branches'] = branches = dict() for branch in self._repo.refs: LOGGER.debug(branch.name) if branch.name in skip: continue branches[branch.name] = branch_history = list() latest = branch.commit history = [latest] + list(latest.iter_parents()) for commit in history: if not is_modified(model, commit): LOGGER.info( "The model was not modified in commit '{}'. Skipping.". format(commit)) continue branch_history.append(commit.hexsha) if commit.hexsha not in commits: commits[commit.hexsha] = sub = dict() sub['timestamp'] = commit.authored_datetime.isoformat(' ') sub['author'] = commit.author.name sub['email'] = commit.author.email LOGGER.debug('%s', json.dumps(self._history, indent=2))
build_branch_structure
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, model, skip):\n self._history = dict()\n self._history['commits'] = commits = dict()\n self._history['branches'] = branches = dict()\n for branch in self._repo.refs:\n LOGGER.debug(branch.name)\n if branch.name in skip:\n continue\n branches[branch.name] = branch_history = list()\n latest = branch.commit\n history = [latest] + list(latest.iter_parents())\n for commit in history:\n if not is_modified(model, commit):\n LOGGER.info(\n \"The model was not modified in commit '{}'. Skipping.\".\n format(commit))\n continue\n branch_history.append(commit.hexsha)\n if commit.hexsha not in commits:\n commits[commit.hexsha] = sub = dict()\n sub['timestamp'] = commit.authored_datetime.isoformat(' ')\n sub['author'] = commit.author.name\n sub['email'] = commit.author.email\n LOGGER.debug('%s', json.dumps(self._history, indent=2))\n", "role": "user" }, { "content": "build_branch_structure", "role": "assistant" } ]
def x(self, model, skip={'gh-pages'}): if self._history is None: self.build_branch_structure(model, skip) self._results = dict() all_commits = list(self._history['commits']) for commit in all_commits: try: self._results[commit] = self.manager.load(commit) except (IOError, NoResultFound) as err: LOGGER.error("Could not load result '%s'.", commit) LOGGER.debug('%s', str(err))
load_history
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, model, skip={'gh-pages'}):\n if self._history is None:\n self.build_branch_structure(model, skip)\n self._results = dict()\n all_commits = list(self._history['commits'])\n for commit in all_commits:\n try:\n self._results[commit] = self.manager.load(commit)\n except (IOError, NoResultFound) as err:\n LOGGER.error(\"Could not load result '%s'.\", commit)\n LOGGER.debug('%s', str(err))\n", "role": "user" }, { "content": "load_history", "role": "assistant" } ]
def x(self, commit, default=MemoteResult()): assert self._results is not None, 'Please call the method `load_history` first.' return self._results.get(commit, default)
get_result
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, commit, default=MemoteResult()):\n assert self._results is not None, 'Please call the method `load_history` first.'\n return self._results.get(commit, default)\n", "role": "user" }, { "content": "get_result", "role": "assistant" } ]
def x(model): s_matrix, _, _ = con_helpers.stoichiometry_matrix(model.metabolites, model.reactions) abs_matrix = np.abs(s_matrix) return abs_matrix.max(), abs_matrix[abs_matrix > 0].min()
absolute_extreme_coefficient_ratio
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n s_matrix, _, _ = con_helpers.stoichiometry_matrix(model.metabolites,\n model.reactions)\n abs_matrix = np.abs(s_matrix)\n return abs_matrix.max(), abs_matrix[abs_matrix > 0].min()\n", "role": "user" }, { "content": "absolute_extreme_coefficient_ratio", "role": "assistant" } ]
def x(model): s_matrix, _, _ = con_helpers.stoichiometry_matrix(model.metabolites, model.reactions) ln_matrix = con_helpers.nullspace(s_matrix.T) return ln_matrix.shape[1]
number_independent_conservation_relations
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n s_matrix, _, _ = con_helpers.stoichiometry_matrix(model.metabolites,\n model.reactions)\n ln_matrix = con_helpers.nullspace(s_matrix.T)\n return ln_matrix.shape[1]\n", "role": "user" }, { "content": "number_independent_conservation_relations", "role": "assistant" } ]
def x(model): s_matrix, _, _ = con_helpers.stoichiometry_matrix(model.metabolites, model.reactions) return con_helpers.rank(s_matrix)
matrix_rank
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n s_matrix, _, _ = con_helpers.stoichiometry_matrix(model.metabolites,\n model.reactions)\n return con_helpers.rank(s_matrix)\n", "role": "user" }, { "content": "matrix_rank", "role": "assistant" } ]
def x(model): s_matrix, _, _ = con_helpers.stoichiometry_matrix(model.metabolites, model.reactions) return s_matrix.shape[1] - matrix_rank(model)
degrees_of_freedom
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n s_matrix, _, _ = con_helpers.stoichiometry_matrix(model.metabolites,\n model.reactions)\n return s_matrix.shape[1] - matrix_rank(model)\n", "role": "user" }, { "content": "degrees_of_freedom", "role": "assistant" } ]
def x(self, model): self.load_medium(model) self.load_essentiality(model) self.load_growth(model) return self
load
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, model):\n self.load_medium(model)\n self.load_essentiality(model)\n self.load_growth(model)\n return self\n", "role": "user" }, { "content": "load", "role": "assistant" } ]
def x(self): validator = Draft4Validator(self.SCHEMA) if not validator.is_valid(self.config): for err in validator.iter_errors(self.config): LOGGER.error(str(err.message)) validator.validate(self.config)
validate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n validator = Draft4Validator(self.SCHEMA)\n if not validator.is_valid(self.config):\n for err in validator.iter_errors(self.config):\n LOGGER.error(str(err.message))\n validator.validate(self.config)\n", "role": "user" }, { "content": "validate", "role": "assistant" } ]
def x(self, model): media = self.config.get('medium') if media is None: return definitions = media.get('definitions') if definitions is None or len(definitions) == 0: return path = self.get_path(media, join('data', 'experimental', 'media')) for medium_id, medium in iteritems(definitions): if medium is None: medium = dict() filename = medium.get('filename') if filename is None: filename = join(path, '{}.csv'.format(medium_id)) elif not isabs(filename): filename = join(path, filename) tmp = Medium(identifier=medium_id, obj=medium, filename=filename) tmp.load() tmp.validate(model) self.media[medium_id] = tmp
load_medium
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, model):\n media = self.config.get('medium')\n if media is None:\n return\n definitions = media.get('definitions')\n if definitions is None or len(definitions) == 0:\n return\n path = self.get_path(media, join('data', 'experimental', 'media'))\n for medium_id, medium in iteritems(definitions):\n if medium is None:\n medium = dict()\n filename = medium.get('filename')\n if filename is None:\n filename = join(path, '{}.csv'.format(medium_id))\n elif not isabs(filename):\n filename = join(path, filename)\n tmp = Medium(identifier=medium_id, obj=medium, filename=filename)\n tmp.load()\n tmp.validate(model)\n self.media[medium_id] = tmp\n", "role": "user" }, { "content": "load_medium", "role": "assistant" } ]
def x(self, model): data = self.config.get('essentiality') if data is None: return experiments = data.get('experiments') if experiments is None or len(experiments) == 0: return path = self.get_path(data, join('data', 'experimental', 'essentiality')) for exp_id, exp in iteritems(experiments): if exp is None: exp = dict() filename = exp.get('filename') if filename is None: filename = join(path, '{}.csv'.format(exp_id)) elif not isabs(filename): filename = join(path, filename) experiment = EssentialityExperiment(identifier=exp_id, obj=exp, filename=filename) if experiment.medium is not None: assert experiment.medium in self.media, "Experiment '{}' has an undefined medium '{}'.".format( exp_id, experiment.medium) experiment.medium = self.media[experiment.medium] experiment.load() experiment.validate(model) self.essentiality[exp_id] = experiment
load_essentiality
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, model):\n data = self.config.get('essentiality')\n if data is None:\n return\n experiments = data.get('experiments')\n if experiments is None or len(experiments) == 0:\n return\n path = self.get_path(data, join('data', 'experimental', 'essentiality'))\n for exp_id, exp in iteritems(experiments):\n if exp is None:\n exp = dict()\n filename = exp.get('filename')\n if filename is None:\n filename = join(path, '{}.csv'.format(exp_id))\n elif not isabs(filename):\n filename = join(path, filename)\n experiment = EssentialityExperiment(identifier=exp_id, obj=exp,\n filename=filename)\n if experiment.medium is not None:\n assert experiment.medium in self.media, \"Experiment '{}' has an undefined medium '{}'.\".format(\n exp_id, experiment.medium)\n experiment.medium = self.media[experiment.medium]\n experiment.load()\n experiment.validate(model)\n self.essentiality[exp_id] = experiment\n", "role": "user" }, { "content": "load_essentiality", "role": "assistant" } ]
def x(self, model): data = self.config.get('growth') if data is None: return experiments = data.get('experiments') if experiments is None or len(experiments) == 0: return path = self.get_path(data, join('data', 'experimental', 'growth')) for exp_id, exp in iteritems(experiments): if exp is None: exp = dict() filename = exp.get('filename') if filename is None: filename = join(path, '{}.csv'.format(exp_id)) elif not isabs(filename): filename = join(path, filename) growth = GrowthExperiment(identifier=exp_id, obj=exp, filename=filename ) if growth.medium is not None: assert growth.medium in self.media, "Growth-experiment '{}' has an undefined medium '{}'.".format( exp_id, growth.medium) growth.medium = self.media[growth.medium] growth.load() growth.validate(model) self.growth[exp_id] = growth
load_growth
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, model):\n data = self.config.get('growth')\n if data is None:\n return\n experiments = data.get('experiments')\n if experiments is None or len(experiments) == 0:\n return\n path = self.get_path(data, join('data', 'experimental', 'growth'))\n for exp_id, exp in iteritems(experiments):\n if exp is None:\n exp = dict()\n filename = exp.get('filename')\n if filename is None:\n filename = join(path, '{}.csv'.format(exp_id))\n elif not isabs(filename):\n filename = join(path, filename)\n growth = GrowthExperiment(identifier=exp_id, obj=exp, filename=filename\n )\n if growth.medium is not None:\n assert growth.medium in self.media, \"Growth-experiment '{}' has an undefined medium '{}'.\".format(\n exp_id, growth.medium)\n growth.medium = self.media[growth.medium]\n growth.load()\n growth.validate(model)\n self.growth[exp_id] = growth\n", "role": "user" }, { "content": "load_growth", "role": "assistant" } ]
def x(self, obj, default): path = obj.get('path') if path is None: path = join(self._base, default) if not isabs(path): path = join(self._base, path) return path
get_path
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, obj, default):\n path = obj.get('path')\n if path is None:\n path = join(self._base, default)\n if not isabs(path):\n path = join(self._base, path)\n return path\n", "role": "user" }, { "content": "get_path", "role": "assistant" } ]
def x(model, components): return [elem for elem in getattr(model, components) if elem.annotation is None or len(elem.annotation) == 0]
find_components_without_annotation
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model, components):\n return [elem for elem in getattr(model, components) if elem.annotation is\n None or len(elem.annotation) == 0]\n", "role": "user" }, { "content": "find_components_without_annotation", "role": "assistant" } ]
def x(elements, component, db): def x(annotation, key, pattern): if key not in annotation: return False test = annotation[key] if isinstance(test, native_str): return pattern.match(test) is None else: return any(pattern.match(elem) is None for elem in test) pattern = {'metabolites': METABOLITE_ANNOTATIONS, 'reactions': REACTION_ANNOTATIONS, 'genes': GENE_PRODUCT_ANNOTATIONS}[component][db] return [elem for elem in elements if is_faulty(elem.annotation, db, pattern)]
generate_component_annotation_miriam_match
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(elements, component, db):\n\n def x(annotation, key, pattern):\n if key not in annotation:\n return False\n test = annotation[key]\n if isinstance(test, native_str):\n return pattern.match(test) is None\n else:\n return any(pattern.match(elem) is None for elem in test)\n pattern = {'metabolites': METABOLITE_ANNOTATIONS, 'reactions':\n REACTION_ANNOTATIONS, 'genes': GENE_PRODUCT_ANNOTATIONS}[component][db]\n return [elem for elem in elements if is_faulty(elem.annotation, db,\n pattern)]\n", "role": "user" }, { "content": "generate_component_annotation_miriam_match", "role": "assistant" } ]
def x(model, components): patterns = {'metabolites': METABOLITE_ANNOTATIONS, 'reactions': REACTION_ANNOTATIONS, 'genes': GENE_PRODUCT_ANNOTATIONS}[components] databases = list(patterns) data = list() index = list() for elem in getattr(model, components): index.append(elem.id) data.append(tuple(patterns[db].match(elem.id) is not None for db in databases)) df = pd.DataFrame(data, index=index, columns=databases) if components != 'genes': df['duplicate'] = df[df['biocyc']].sum(axis=1) >= 2 df['duplicate'].fillna(False, inplace=True) df.loc[df['duplicate'], 'biocyc'] = False del df['duplicate'] return df
generate_component_id_namespace_overview
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model, components):\n patterns = {'metabolites': METABOLITE_ANNOTATIONS, 'reactions':\n REACTION_ANNOTATIONS, 'genes': GENE_PRODUCT_ANNOTATIONS}[components]\n databases = list(patterns)\n data = list()\n index = list()\n for elem in getattr(model, components):\n index.append(elem.id)\n data.append(tuple(patterns[db].match(elem.id) is not None for db in\n databases))\n df = pd.DataFrame(data, index=index, columns=databases)\n if components != 'genes':\n df['duplicate'] = df[df['biocyc']].sum(axis=1) >= 2\n df['duplicate'].fillna(False, inplace=True)\n df.loc[df['duplicate'], 'biocyc'] = False\n del df['duplicate']\n return df\n", "role": "user" }, { "content": "generate_component_id_namespace_overview", "role": "assistant" } ]
def x(predicted_essential, expected_essential, predicted_nonessential, expected_nonessential): true_positive = predicted_essential & expected_essential tp = len(true_positive) true_negative = predicted_nonessential & expected_nonessential tn = len(true_negative) false_positive = predicted_essential - expected_essential fp = len(false_positive) false_negative = predicted_nonessential - expected_nonessential fn = len(false_negative) try: tpr = tp / (tp + fn) except ZeroDivisionError: tpr = None try: tnr = tn / (tn + fp) except ZeroDivisionError: tnr = None try: ppv = tp / (tp + fp) except ZeroDivisionError: ppv = None fdr = 1 - ppv try: acc = (tp + tn) / (tp + tn + fp + fn) except ZeroDivisionError: acc = None try: mcc = (tp * tn - fp * fn) / sqrt((tp + fp) * (tp + fn) * (tn + fp) * (tn + fn)) except ZeroDivisionError: mcc = None return {'TP': list(true_positive), 'TN': list(true_negative), 'FP': list(false_positive), 'FN': list(false_negative), 'TPR': tpr, 'TNR': tnr, 'PPV': ppv, 'FDR': fdr, 'ACC': acc, 'MCC': mcc}
confusion_matrix
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(predicted_essential, expected_essential, predicted_nonessential,\n expected_nonessential):\n true_positive = predicted_essential & expected_essential\n tp = len(true_positive)\n true_negative = predicted_nonessential & expected_nonessential\n tn = len(true_negative)\n false_positive = predicted_essential - expected_essential\n fp = len(false_positive)\n false_negative = predicted_nonessential - expected_nonessential\n fn = len(false_negative)\n try:\n tpr = tp / (tp + fn)\n except ZeroDivisionError:\n tpr = None\n try:\n tnr = tn / (tn + fp)\n except ZeroDivisionError:\n tnr = None\n try:\n ppv = tp / (tp + fp)\n except ZeroDivisionError:\n ppv = None\n fdr = 1 - ppv\n try:\n acc = (tp + tn) / (tp + tn + fp + fn)\n except ZeroDivisionError:\n acc = None\n try:\n mcc = (tp * tn - fp * fn) / sqrt((tp + fp) * (tp + fn) * (tn + fp) *\n (tn + fn))\n except ZeroDivisionError:\n mcc = None\n return {'TP': list(true_positive), 'TN': list(true_negative), 'FP':\n list(false_positive), 'FN': list(false_negative), 'TPR': tpr, 'TNR':\n tnr, 'PPV': ppv, 'FDR': fdr, 'ACC': acc, 'MCC': mcc}\n", "role": "user" }, { "content": "confusion_matrix", "role": "assistant" } ]
def x(path): notifications = {'warnings': [], 'errors': []} model, sbml_ver = val.load_cobra_model(path, notifications) return model, sbml_ver, notifications
validate_model
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path):\n notifications = {'warnings': [], 'errors': []}\n model, sbml_ver = val.load_cobra_model(path, notifications)\n return model, sbml_ver, notifications\n", "role": "user" }, { "content": "validate_model", "role": "assistant" } ]
def x(result, config=None, html=True): if config is None: config = ReportConfiguration.load() report = SnapshotReport(result=result, configuration=config) if html: return report.render_html() else: return report.render_json()
snapshot_report
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(result, config=None, html=True):\n if config is None:\n config = ReportConfiguration.load()\n report = SnapshotReport(result=result, configuration=config)\n if html:\n return report.render_html()\n else:\n return report.render_json()\n", "role": "user" }, { "content": "snapshot_report", "role": "assistant" } ]
def x(history, config=None, html=True): if config is None: config = ReportConfiguration.load() report = HistoryReport(history=history, configuration=config) if html: return report.render_html() else: return report.render_json()
history_report
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(history, config=None, html=True):\n if config is None:\n config = ReportConfiguration.load()\n report = HistoryReport(history=history, configuration=config)\n if html:\n return report.render_html()\n else:\n return report.render_json()\n", "role": "user" }, { "content": "history_report", "role": "assistant" } ]
def x(diff_results, config=None, html=True): if config is None: config = ReportConfiguration.load() report = DiffReport(diff_results=diff_results, configuration=config) if html: return report.render_html() else: return report.render_json()
diff_report
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(diff_results, config=None, html=True):\n if config is None:\n config = ReportConfiguration.load()\n report = DiffReport(diff_results=diff_results, configuration=config)\n if html:\n return report.render_html()\n else:\n return report.render_json()\n", "role": "user" }, { "content": "diff_report", "role": "assistant" } ]
def x(path, notifications, filename): env = Environment(loader=PackageLoader('memote.suite', 'templates'), autoescape=select_autoescape(['html', 'xml'])) template = env.get_template('validation_template.html') model = os.path.basename(path) with open(filename, 'w') as file_h: file_h.write(template.render(model=model, notifications=notifications))
validation_report
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path, notifications, filename):\n env = Environment(loader=PackageLoader('memote.suite', 'templates'),\n autoescape=select_autoescape(['html', 'xml']))\n template = env.get_template('validation_template.html')\n model = os.path.basename(path)\n with open(filename, 'w') as file_h:\n file_h.write(template.render(model=model, notifications=notifications))\n", "role": "user" }, { "content": "validation_report", "role": "assistant" } ]
def x(cls, filename=None): if filename is None: LOGGER.debug('Loading default configuration.') with open_text(templates, 'test_config.yml', encoding='utf-8' ) as file_handle: content = yaml.load(file_handle) else: LOGGER.debug("Loading custom configuration '%s'.", filename) try: with open(filename, encoding='utf-8') as file_handle: content = yaml.load(file_handle) except IOError as err: LOGGER.error( "Failed to load the custom configuration '%s'. Skipping.", filename) LOGGER.debug(str(err)) content = dict() return cls(content)
load
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, filename=None):\n if filename is None:\n LOGGER.debug('Loading default configuration.')\n with open_text(templates, 'test_config.yml', encoding='utf-8'\n ) as file_handle:\n content = yaml.load(file_handle)\n else:\n LOGGER.debug(\"Loading custom configuration '%s'.\", filename)\n try:\n with open(filename, encoding='utf-8') as file_handle:\n content = yaml.load(file_handle)\n except IOError as err:\n LOGGER.error(\n \"Failed to load the custom configuration '%s'. Skipping.\",\n filename)\n LOGGER.debug(str(err))\n content = dict()\n return cls(content)\n", "role": "user" }, { "content": "load", "role": "assistant" } ]
def x(gpr): logger.debug('%r', gpr) conform = logical_and.sub('and', gpr) conform = logical_or.sub('or', conform) conform = escape_chars.sub('_', conform) expression = ast.parse(conform) walker = GPRVisitor() walker.visit(expression) return len(walker.left ^ walker.right)
find_top_level_complex
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(gpr):\n logger.debug('%r', gpr)\n conform = logical_and.sub('and', gpr)\n conform = logical_or.sub('or', conform)\n conform = escape_chars.sub('_', conform)\n expression = ast.parse(conform)\n walker = GPRVisitor()\n walker.visit(expression)\n return len(walker.left ^ walker.right)\n", "role": "user" }, { "content": "find_top_level_complex", "role": "assistant" } ]
def x(self, node): if self._is_top and isinstance(node.op, ast.And): self._is_top = False self._current = self.left self.visit(node.values[0]) self._current = self.right for successor in node.values[1:]: self.visit(successor) else: self.generic_visit(node)
visit_BoolOp
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, node):\n if self._is_top and isinstance(node.op, ast.And):\n self._is_top = False\n self._current = self.left\n self.visit(node.values[0])\n self._current = self.right\n for successor in node.values[1:]:\n self.visit(successor)\n else:\n self.generic_visit(node)\n", "role": "user" }, { "content": "visit_BoolOp", "role": "assistant" } ]
def x(model): lower_bound, upper_bound = helpers.find_bounds(model) return [rxn for rxn in model.reactions if 0 > rxn.lower_bound > lower_bound or 0 < rxn.upper_bound < upper_bound]
find_nonzero_constrained_reactions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n lower_bound, upper_bound = helpers.find_bounds(model)\n return [rxn for rxn in model.reactions if 0 > rxn.lower_bound >\n lower_bound or 0 < rxn.upper_bound < upper_bound]\n", "role": "user" }, { "content": "find_nonzero_constrained_reactions", "role": "assistant" } ]
def x(model): return [rxn for rxn in model.reactions if rxn.lower_bound == 0 and rxn. upper_bound == 0]
find_zero_constrained_reactions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n return [rxn for rxn in model.reactions if rxn.lower_bound == 0 and rxn.\n upper_bound == 0]\n", "role": "user" }, { "content": "find_zero_constrained_reactions", "role": "assistant" } ]
def x(model): lower_bound, upper_bound = helpers.find_bounds(model) return [rxn for rxn in model.reactions if rxn.lower_bound <= lower_bound and rxn.upper_bound >= upper_bound]
find_unconstrained_reactions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n lower_bound, upper_bound = helpers.find_bounds(model)\n return [rxn for rxn in model.reactions if rxn.lower_bound <=\n lower_bound and rxn.upper_bound >= upper_bound]\n", "role": "user" }, { "content": "find_unconstrained_reactions", "role": "assistant" } ]
def x(model): atp_adp_conv_rxns = helpers.find_converting_reactions(model, ('MNXM3', 'MNXM7')) id_of_main_compartment = helpers.find_compartment_id_in_model(model, 'c') reactants = {helpers.find_met_in_model(model, 'MNXM3', id_of_main_compartment)[0], helpers.find_met_in_model(model, 'MNXM2', id_of_main_compartment)[0]} products = {helpers.find_met_in_model(model, 'MNXM7', id_of_main_compartment)[0], helpers.find_met_in_model(model, 'MNXM1', id_of_main_compartment)[0], helpers.find_met_in_model( model, 'MNXM9', id_of_main_compartment)[0]} candidates = [rxn for rxn in atp_adp_conv_rxns if rxn.reversibility is False and set(rxn.reactants) == reactants and set(rxn.products) == products] buzzwords = ['maintenance', 'atpm', 'requirement', 'ngam', 'non-growth', 'associated'] refined_candidates = [rxn for rxn in candidates if any(string in filter_none(rxn.name, '').lower() for string in buzzwords)] if refined_candidates: return refined_candidates else: return candidates
find_ngam
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n atp_adp_conv_rxns = helpers.find_converting_reactions(model, ('MNXM3',\n 'MNXM7'))\n id_of_main_compartment = helpers.find_compartment_id_in_model(model, 'c')\n reactants = {helpers.find_met_in_model(model, 'MNXM3',\n id_of_main_compartment)[0], helpers.find_met_in_model(model,\n 'MNXM2', id_of_main_compartment)[0]}\n products = {helpers.find_met_in_model(model, 'MNXM7',\n id_of_main_compartment)[0], helpers.find_met_in_model(model,\n 'MNXM1', id_of_main_compartment)[0], helpers.find_met_in_model(\n model, 'MNXM9', id_of_main_compartment)[0]}\n candidates = [rxn for rxn in atp_adp_conv_rxns if rxn.reversibility is \n False and set(rxn.reactants) == reactants and set(rxn.products) ==\n products]\n buzzwords = ['maintenance', 'atpm', 'requirement', 'ngam', 'non-growth',\n 'associated']\n refined_candidates = [rxn for rxn in candidates if any(string in\n filter_none(rxn.name, '').lower() for string in buzzwords)]\n if refined_candidates:\n return refined_candidates\n else:\n return candidates\n", "role": "user" }, { "content": "find_ngam", "role": "assistant" } ]
def x(model): if len(model.reactions) == 0 or len(model.genes) == 0: raise ValueError('The model contains no reactions or genes.') return float(len(model.reactions)) / float(len(model.genes))
calculate_metabolic_coverage
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n if len(model.reactions) == 0 or len(model.genes) == 0:\n raise ValueError('The model contains no reactions or genes.')\n return float(len(model.reactions)) / float(len(model.genes))\n", "role": "user" }, { "content": "calculate_metabolic_coverage", "role": "assistant" } ]
def x(model): complexes = [] for rxn in model.reactions: if not rxn.gene_reaction_rule: continue size = find_top_level_complex(rxn.gene_reaction_rule) if size >= 2: complexes.append(rxn) return complexes
find_protein_complexes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n complexes = []\n for rxn in model.reactions:\n if not rxn.gene_reaction_rule:\n continue\n size = find_top_level_complex(rxn.gene_reaction_rule)\n if size >= 2:\n complexes.append(rxn)\n return complexes\n", "role": "user" }, { "content": "find_protein_complexes", "role": "assistant" } ]
def x(model, rxn): lower_bound, upper_bound = helpers.find_bounds(model) if rxn.reversibility: return rxn.lower_bound > lower_bound or rxn.upper_bound < upper_bound else: return rxn.lower_bound > 0 or rxn.upper_bound < upper_bound
is_constrained_reaction
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model, rxn):\n lower_bound, upper_bound = helpers.find_bounds(model)\n if rxn.reversibility:\n return rxn.lower_bound > lower_bound or rxn.upper_bound < upper_bound\n else:\n return rxn.lower_bound > 0 or rxn.upper_bound < upper_bound\n", "role": "user" }, { "content": "is_constrained_reaction", "role": "assistant" } ]
def x(model): o2_in_model = helpers.find_met_in_model(model, 'MNXM4') return set([rxn for met in model.metabolites for rxn in met.reactions if met.formula == 'O2' or met in o2_in_model])
find_oxygen_reactions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n o2_in_model = helpers.find_met_in_model(model, 'MNXM4')\n return set([rxn for met in model.metabolites for rxn in met.reactions if\n met.formula == 'O2' or met in o2_in_model])\n", "role": "user" }, { "content": "find_oxygen_reactions", "role": "assistant" } ]
def x(model): unique = set() for met in model.metabolites: is_missing = True for comp in model.compartments: if met.id.endswith('_{}'.format(comp)): unique.add(met.id[:-(len(comp) + 1)]) is_missing = False break if is_missing: unique.add(met.id) return unique
find_unique_metabolites
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n unique = set()\n for met in model.metabolites:\n is_missing = True\n for comp in model.compartments:\n if met.id.endswith('_{}'.format(comp)):\n unique.add(met.id[:-(len(comp) + 1)])\n is_missing = False\n break\n if is_missing:\n unique.add(met.id)\n return unique\n", "role": "user" }, { "content": "find_unique_metabolites", "role": "assistant" } ]
def x(model): unique_identifiers = ['inchikey', 'inchi'] duplicates = [] for met_1, met_2 in combinations(model.metabolites, 2): if met_1.compartment == met_2.compartment: for key in unique_identifiers: if key in met_1.annotation and key in met_2.annotation: if met_1.annotation[key] == met_2.annotation[key]: duplicates.append((met_1.id, met_2.id)) break return duplicates
find_duplicate_metabolites_in_compartments
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n unique_identifiers = ['inchikey', 'inchi']\n duplicates = []\n for met_1, met_2 in combinations(model.metabolites, 2):\n if met_1.compartment == met_2.compartment:\n for key in unique_identifiers:\n if key in met_1.annotation and key in met_2.annotation:\n if met_1.annotation[key] == met_2.annotation[key]:\n duplicates.append((met_1.id, met_2.id))\n break\n return duplicates\n", "role": "user" }, { "content": "find_duplicate_metabolites_in_compartments", "role": "assistant" } ]
def x(model): duplicates = {} rxn_db_identifiers = ['metanetx.reaction', 'kegg.reaction', 'brenda', 'rhea', 'biocyc', 'bigg.reaction'] ann_rxns = [] for rxn in model.reactions: ann = [] for key in rxn_db_identifiers: if key in rxn.annotation: if isinstance(rxn.annotation[key], list): ann.extend([(key, elem) for elem in rxn.annotation[key]]) else: ann.append((key, rxn.annotation[key])) ann_rxns.append((rxn, frozenset(ann))) for (rxn_a, ann_a), (rxn_b, ann_b) in combinations(ann_rxns, 2): mutual_pair = tuple(ann_a & ann_b) if len(mutual_pair) > 0: duplicates.setdefault(mutual_pair, set()).update([rxn_a.id, rxn_b.id]) num_duplicated = set() duplicated = {} for key in duplicates: new_key = ','.join(sorted('{}:{}'.format(ns, term) for ns, term in key) ) duplicated[new_key] = rxns = list(duplicates[key]) num_duplicated.update(rxns) return duplicated, len(num_duplicated)
find_reactions_with_partially_identical_annotations
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n duplicates = {}\n rxn_db_identifiers = ['metanetx.reaction', 'kegg.reaction', 'brenda',\n 'rhea', 'biocyc', 'bigg.reaction']\n ann_rxns = []\n for rxn in model.reactions:\n ann = []\n for key in rxn_db_identifiers:\n if key in rxn.annotation:\n if isinstance(rxn.annotation[key], list):\n ann.extend([(key, elem) for elem in rxn.annotation[key]])\n else:\n ann.append((key, rxn.annotation[key]))\n ann_rxns.append((rxn, frozenset(ann)))\n for (rxn_a, ann_a), (rxn_b, ann_b) in combinations(ann_rxns, 2):\n mutual_pair = tuple(ann_a & ann_b)\n if len(mutual_pair) > 0:\n duplicates.setdefault(mutual_pair, set()).update([rxn_a.id,\n rxn_b.id])\n num_duplicated = set()\n duplicated = {}\n for key in duplicates:\n new_key = ','.join(sorted('{}:{}'.format(ns, term) for ns, term in key)\n )\n duplicated[new_key] = rxns = list(duplicates[key])\n num_duplicated.update(rxns)\n return duplicated, len(num_duplicated)\n", "role": "user" }, { "content": "find_reactions_with_partially_identical_annotations", "role": "assistant" } ]
def x(metabolites, compartments): unique_identifiers = ['inchikey', 'inchi'] met2mol = {} molecules = {c: [] for c in compartments} for met in metabolites: ann = [] for key in unique_identifiers: mol = met.annotation.get(key) if mol is not None: ann.append(mol) if len(ann) == 0: continue ann = set(ann) mols = molecules[met.compartment] for i, mol_group in enumerate(mols): if len(ann & mol_group) > 0: mol_group.update(ann) met2mol[met] = '{}-{}'.format(met.compartment, i) break if met not in met2mol: met2mol[met] = '{}-{}'.format(met.compartment, len(mols)) mols.append(ann) return met2mol
map_metabolites_to_structures
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(metabolites, compartments):\n unique_identifiers = ['inchikey', 'inchi']\n met2mol = {}\n molecules = {c: [] for c in compartments}\n for met in metabolites:\n ann = []\n for key in unique_identifiers:\n mol = met.annotation.get(key)\n if mol is not None:\n ann.append(mol)\n if len(ann) == 0:\n continue\n ann = set(ann)\n mols = molecules[met.compartment]\n for i, mol_group in enumerate(mols):\n if len(ann & mol_group) > 0:\n mol_group.update(ann)\n met2mol[met] = '{}-{}'.format(met.compartment, i)\n break\n if met not in met2mol:\n met2mol[met] = '{}-{}'.format(met.compartment, len(mols))\n mols.append(ann)\n return met2mol\n", "role": "user" }, { "content": "map_metabolites_to_structures", "role": "assistant" } ]
def x(model): met2mol = map_metabolites_to_structures(model.metabolites, model. compartments) structural = [] for rxn in model.reactions: if not all(met in met2mol for met in rxn.metabolites): continue substrates = {met2mol[met]: rxn.get_coefficient(met) for met in rxn .reactants} products = {met2mol[met]: rxn.get_coefficient(met) for met in rxn. products} structural.append((rxn, substrates, products)) num_duplicated = set() duplicates = [] for (rxn_a, sub_a, prod_a), (rxn_b, sub_b, prod_b) in combinations( structural, 2): if sub_a != sub_b: continue if prod_a != prod_b: continue if rxn_a.reversibility != rxn_b.reversibility: continue duplicates.append((rxn_a.id, rxn_b.id)) num_duplicated.add(rxn_a.id) num_duplicated.add(rxn_b.id) return duplicates, len(num_duplicated)
find_duplicate_reactions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n met2mol = map_metabolites_to_structures(model.metabolites, model.\n compartments)\n structural = []\n for rxn in model.reactions:\n if not all(met in met2mol for met in rxn.metabolites):\n continue\n substrates = {met2mol[met]: rxn.get_coefficient(met) for met in rxn\n .reactants}\n products = {met2mol[met]: rxn.get_coefficient(met) for met in rxn.\n products}\n structural.append((rxn, substrates, products))\n num_duplicated = set()\n duplicates = []\n for (rxn_a, sub_a, prod_a), (rxn_b, sub_b, prod_b) in combinations(\n structural, 2):\n if sub_a != sub_b:\n continue\n if prod_a != prod_b:\n continue\n if rxn_a.reversibility != rxn_b.reversibility:\n continue\n duplicates.append((rxn_a.id, rxn_b.id))\n num_duplicated.add(rxn_a.id)\n num_duplicated.add(rxn_b.id)\n return duplicates, len(num_duplicated)\n", "role": "user" }, { "content": "find_duplicate_reactions", "role": "assistant" } ]
def x(model): duplicates = dict() for rxn_a, rxn_b in combinations(model.reactions, 2): if not (rxn_a.genes and rxn_b.genes): continue if rxn_a.genes == rxn_b.genes: identifiers = rxn_a.genes duplicates.setdefault(identifiers, set()).update([rxn_a.id, rxn_b.id]) num_duplicated = set() duplicated = {} for key in duplicates: new_key = ','.join(sorted(g.id for g in key)) duplicated[new_key] = rxns = list(duplicates[key]) num_duplicated.update(rxns) return duplicated, len(num_duplicated)
find_reactions_with_identical_genes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n duplicates = dict()\n for rxn_a, rxn_b in combinations(model.reactions, 2):\n if not (rxn_a.genes and rxn_b.genes):\n continue\n if rxn_a.genes == rxn_b.genes:\n identifiers = rxn_a.genes\n duplicates.setdefault(identifiers, set()).update([rxn_a.id,\n rxn_b.id])\n num_duplicated = set()\n duplicated = {}\n for key in duplicates:\n new_key = ','.join(sorted(g.id for g in key))\n duplicated[new_key] = rxns = list(duplicates[key])\n num_duplicated.update(rxns)\n return duplicated, len(num_duplicated)\n", "role": "user" }, { "content": "find_reactions_with_identical_genes", "role": "assistant" } ]
def x(model): return [met.id for rxn in model.medium for met in model.reactions. get_by_id(rxn).metabolites]
find_medium_metabolites
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n return [met.id for rxn in model.medium for met in model.reactions.\n get_by_id(rxn).metabolites]\n", "role": "user" }, { "content": "find_medium_metabolites", "role": "assistant" } ]
def x(model): ex_comp = find_external_compartment(model) return [met for met in model.metabolites if met.compartment == ex_comp]
find_external_metabolites
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model):\n ex_comp = find_external_compartment(model)\n return [met for met in model.metabolites if met.compartment == ex_comp]\n", "role": "user" }, { "content": "find_external_metabolites", "role": "assistant" } ]
def x(self, result, filename, pretty=True): LOGGER.info("Storing result in '%s'.", filename) if filename.endswith('.gz'): with gzip.open(filename, 'wb') as file_handle: file_handle.write(jsonify(result, pretty=pretty).encode('utf-8')) else: with open(filename, 'w', encoding='utf-8') as file_handle: file_handle.write(jsonify(result, pretty=pretty))
store
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, result, filename, pretty=True):\n LOGGER.info(\"Storing result in '%s'.\", filename)\n if filename.endswith('.gz'):\n with gzip.open(filename, 'wb') as file_handle:\n file_handle.write(jsonify(result, pretty=pretty).encode('utf-8'))\n else:\n with open(filename, 'w', encoding='utf-8') as file_handle:\n file_handle.write(jsonify(result, pretty=pretty))\n", "role": "user" }, { "content": "store", "role": "assistant" } ]
def x(self, filename): LOGGER.info("Loading result from '%s'.", filename) if filename.endswith('.gz'): with gzip.open(filename, 'rb') as file_handle: result = MemoteResult(json.loads(file_handle.read().decode( 'utf-8'))) else: with open(filename, 'r', encoding='utf-8') as file_handle: result = MemoteResult(json.load(file_handle)) return result
load
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, filename):\n LOGGER.info(\"Loading result from '%s'.\", filename)\n if filename.endswith('.gz'):\n with gzip.open(filename, 'rb') as file_handle:\n result = MemoteResult(json.loads(file_handle.read().decode(\n 'utf-8')))\n else:\n with open(filename, 'r', encoding='utf-8') as file_handle:\n result = MemoteResult(json.load(file_handle))\n return result\n", "role": "user" }, { "content": "load", "role": "assistant" } ]
def x(path, notifications): doc = libsbml.readSBML(path) fbc = doc.getPlugin('fbc') sbml_ver = doc.getLevel(), doc.getVersion( ), fbc if fbc is None else fbc.getVersion() with catch_warnings(record=True) as warnings: simplefilter('always') try: model = read_sbml_model(path) except Exception as err: notifications['errors'].append(str(err)) model = None validate = True else: validate = False notifications['warnings'].extend([str(w.message) for w in warnings]) if validate: run_sbml_validation(doc, notifications) return model, sbml_ver
load_cobra_model
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path, notifications):\n doc = libsbml.readSBML(path)\n fbc = doc.getPlugin('fbc')\n sbml_ver = doc.getLevel(), doc.getVersion(\n ), fbc if fbc is None else fbc.getVersion()\n with catch_warnings(record=True) as warnings:\n simplefilter('always')\n try:\n model = read_sbml_model(path)\n except Exception as err:\n notifications['errors'].append(str(err))\n model = None\n validate = True\n else:\n validate = False\n notifications['warnings'].extend([str(w.message) for w in warnings])\n if validate:\n run_sbml_validation(doc, notifications)\n return model, sbml_ver\n", "role": "user" }, { "content": "load_cobra_model", "role": "assistant" } ]
def x(failure): return 'Line {}, Column {} - #{}: {} - Category: {}, Severity: {}'.format( failure.getLine(), failure.getColumn(), failure.getErrorId(), failure.getMessage(), failure.getCategoryAsString(), failure. getSeverity())
format_failure
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(failure):\n return 'Line {}, Column {} - #{}: {} - Category: {}, Severity: {}'.format(\n failure.getLine(), failure.getColumn(), failure.getErrorId(),\n failure.getMessage(), failure.getCategoryAsString(), failure.\n getSeverity())\n", "role": "user" }, { "content": "format_failure", "role": "assistant" } ]