source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(time_filter): if not time_filter: time_filter = '[* TO *]' start, end = parse_solr_time_range_as_pair(time_filter) start, end = parse_datetime(start), parse_datetime(end) return start, end
parse_datetime_range
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(time_filter):\n if not time_filter:\n time_filter = '[* TO *]'\n start, end = parse_solr_time_range_as_pair(time_filter)\n start, end = parse_datetime(start), parse_datetime(end)\n return start, end\n", "role": "user" }, { "content": "parse_datetime_range", "role": "assistant" } ]
def x(time_gap): matcher = None if time_gap.count('T'): units = {'H': ('HOURS', isodate.Duration(hours=1)), 'M': ('MINUTES', isodate.Duration(minutes=1)), 'S': ('SECONDS', isodate.Duration (seconds=1))} matcher = re.search('PT(\\d+)([HMS])', time_gap) if matcher: quantity = int(matcher.group(1)) unit = matcher.group(2) return quantity, units.get(unit) else: raise Exception('Does not match the pattern: {}'.format(time_gap)) else: units = {'Y': ('YEARS', isodate.Duration(years=1)), 'M': ('MONTHS', isodate.Duration(months=1)), 'W': ('WEEKS', isodate.Duration( weeks=1)), 'D': ('DAYS', isodate.Duration(days=1))} matcher = re.search('P(\\d+)([YMWD])', time_gap) if matcher: quantity = int(matcher.group(1)) unit = matcher.group(2) else: raise Exception('Does not match the pattern: {}'.format(time_gap)) return quantity, units.get(unit)
parse_ISO8601
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(time_gap):\n matcher = None\n if time_gap.count('T'):\n units = {'H': ('HOURS', isodate.Duration(hours=1)), 'M': ('MINUTES',\n isodate.Duration(minutes=1)), 'S': ('SECONDS', isodate.Duration\n (seconds=1))}\n matcher = re.search('PT(\\\\d+)([HMS])', time_gap)\n if matcher:\n quantity = int(matcher.group(1))\n unit = matcher.group(2)\n return quantity, units.get(unit)\n else:\n raise Exception('Does not match the pattern: {}'.format(time_gap))\n else:\n units = {'Y': ('YEARS', isodate.Duration(years=1)), 'M': ('MONTHS',\n isodate.Duration(months=1)), 'W': ('WEEKS', isodate.Duration(\n weeks=1)), 'D': ('DAYS', isodate.Duration(days=1))}\n matcher = re.search('P(\\\\d+)([YMWD])', time_gap)\n if matcher:\n quantity = int(matcher.group(1))\n unit = matcher.group(2)\n else:\n raise Exception('Does not match the pattern: {}'.format(time_gap))\n return quantity, units.get(unit)\n", "role": "user" }, { "content": "parse_ISO8601", "role": "assistant" } ]
def x(start, end, time_limit): if is_range_common_era(start, end): duration = end.get('parsed_datetime') - start.get('parsed_datetime') unit = int(math.ceil(duration.days / float(time_limit))) return '+{0}DAYS'.format(unit) else: return '+100YEARS'
compute_gap
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(start, end, time_limit):\n if is_range_common_era(start, end):\n duration = end.get('parsed_datetime') - start.get('parsed_datetime')\n unit = int(math.ceil(duration.days / float(time_limit)))\n return '+{0}DAYS'.format(unit)\n else:\n return '+100YEARS'\n", "role": "user" }, { "content": "compute_gap", "role": "assistant" } ]
def x(time_gap): quantity, unit = parse_ISO8601(time_gap) if unit[0] == 'WEEKS': return '+{0}DAYS'.format(quantity * 7) else: return '+{0}{1}'.format(quantity, unit[0])
gap_to_sorl
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(time_gap):\n quantity, unit = parse_ISO8601(time_gap)\n if unit[0] == 'WEEKS':\n return '+{0}DAYS'.format(quantity * 7)\n else:\n return '+{0}{1}'.format(quantity, unit[0])\n", "role": "user" }, { "content": "gap_to_sorl", "role": "assistant" } ]
def x(field, time_filter, time_gap, time_limit=100): start, end = parse_datetime_range(time_filter) key_range_start = 'f.{0}.facet.range.start'.format(field) key_range_end = 'f.{0}.facet.range.end'.format(field) key_range_gap = 'f.{0}.facet.range.gap'.format(field) key_range_mincount = 'f.{0}.facet.mincount'.format(field) if time_gap: gap = gap_to_sorl(time_gap) else: gap = compute_gap(start, end, time_limit) value_range_start = start.get('parsed_datetime') if start.get('is_common_era'): value_range_start = start.get('parsed_datetime').isoformat().replace( '+00:00', '') + 'Z' value_range_end = start.get('parsed_datetime') if end.get('is_common_era'): value_range_end = end.get('parsed_datetime').isoformat().replace( '+00:00', '') + 'Z' value_range_gap = gap params = {'facet.range': field, key_range_start: value_range_start, key_range_end: value_range_end, key_range_gap: value_range_gap, key_range_mincount: 1} return params
request_time_facet
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(field, time_filter, time_gap, time_limit=100):\n start, end = parse_datetime_range(time_filter)\n key_range_start = 'f.{0}.facet.range.start'.format(field)\n key_range_end = 'f.{0}.facet.range.end'.format(field)\n key_range_gap = 'f.{0}.facet.range.gap'.format(field)\n key_range_mincount = 'f.{0}.facet.mincount'.format(field)\n if time_gap:\n gap = gap_to_sorl(time_gap)\n else:\n gap = compute_gap(start, end, time_limit)\n value_range_start = start.get('parsed_datetime')\n if start.get('is_common_era'):\n value_range_start = start.get('parsed_datetime').isoformat().replace(\n '+00:00', '') + 'Z'\n value_range_end = start.get('parsed_datetime')\n if end.get('is_common_era'):\n value_range_end = end.get('parsed_datetime').isoformat().replace(\n '+00:00', '') + 'Z'\n value_range_gap = gap\n params = {'facet.range': field, key_range_start: value_range_start,\n key_range_end: value_range_end, key_range_gap: value_range_gap,\n key_range_mincount: 1}\n return params\n", "role": "user" }, { "content": "request_time_facet", "role": "assistant" } ]
def x(geo_box_str): pattern = '\\[(.*) TO (.*)\\]' matcher = re.search(pattern, geo_box_str) if matcher: return matcher.group(1), matcher.group(2) else: raise Exception('Regex {0} could not parse {1}'.format(pattern, geo_box_str))
parse_solr_geo_range_as_pair
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(geo_box_str):\n pattern = '\\\\[(.*) TO (.*)\\\\]'\n matcher = re.search(pattern, geo_box_str)\n if matcher:\n return matcher.group(1), matcher.group(2)\n else:\n raise Exception('Regex {0} could not parse {1}'.format(pattern,\n geo_box_str))\n", "role": "user" }, { "content": "parse_solr_geo_range_as_pair", "role": "assistant" } ]
def x(geo_box_str): from_point_str, to_point_str = parse_solr_geo_range_as_pair(geo_box_str) from_point = parse_lat_lon(from_point_str) to_point = parse_lat_lon(to_point_str) rectangle = box(from_point[0], from_point[1], to_point[0], to_point[1]) return rectangle
parse_geo_box
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(geo_box_str):\n from_point_str, to_point_str = parse_solr_geo_range_as_pair(geo_box_str)\n from_point = parse_lat_lon(from_point_str)\n to_point = parse_lat_lon(to_point_str)\n rectangle = box(from_point[0], from_point[1], to_point[0], to_point[1])\n return rectangle\n", "role": "user" }, { "content": "parse_geo_box", "role": "assistant" } ]
def x(field, hm_filter, hm_grid_level, hm_limit): if not hm_filter: hm_filter = '[-90,-180 TO 90,180]' params = {'facet': 'on', 'facet.heatmap': field, 'facet.heatmap.geom': hm_filter} if hm_grid_level: params['facet.heatmap.gridLevel'] = hm_grid_level else: rectangle = parse_geo_box(hm_filter) degrees_side_length = rectangle.length / 2 cell_side_length = math.sqrt(float(hm_limit)) cell_side_length_degrees = degrees_side_length / cell_side_length * 2 params['facet.heatmap.distErr'] = str(float(cell_side_length_degrees)) return params
request_heatmap_facet
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(field, hm_filter, hm_grid_level, hm_limit):\n if not hm_filter:\n hm_filter = '[-90,-180 TO 90,180]'\n params = {'facet': 'on', 'facet.heatmap': field, 'facet.heatmap.geom':\n hm_filter}\n if hm_grid_level:\n params['facet.heatmap.gridLevel'] = hm_grid_level\n else:\n rectangle = parse_geo_box(hm_filter)\n degrees_side_length = rectangle.length / 2\n cell_side_length = math.sqrt(float(hm_limit))\n cell_side_length_degrees = degrees_side_length / cell_side_length * 2\n params['facet.heatmap.distErr'] = str(float(cell_side_length_degrees))\n return params\n", "role": "user" }, { "content": "request_heatmap_facet", "role": "assistant" } ]
def x(field, time_filter, search_engine_endpoint, actual_params=None): if actual_params: raise NotImplemented('actual_params') start, end = parse_solr_time_range_as_pair(time_filter) if start == '*' or end == '*': params_stats = {'q': '*:*', 'rows': 0, 'stats.field': field, 'stats': 'true', 'wt': 'json'} res_stats = requests.get(search_engine_endpoint, params=params_stats) if res_stats.ok: stats_date_field = res_stats.json()['stats']['stats_fields'][field] date_min = stats_date_field['min'] date_max = stats_date_field['max'] if start != '*': date_min = start if end != '*': date_max = end time_filter = '[{0} TO {1}]'.format(date_min, date_max) return time_filter
asterisk_to_min_max
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(field, time_filter, search_engine_endpoint, actual_params=None):\n if actual_params:\n raise NotImplemented('actual_params')\n start, end = parse_solr_time_range_as_pair(time_filter)\n if start == '*' or end == '*':\n params_stats = {'q': '*:*', 'rows': 0, 'stats.field': field,\n 'stats': 'true', 'wt': 'json'}\n res_stats = requests.get(search_engine_endpoint, params=params_stats)\n if res_stats.ok:\n stats_date_field = res_stats.json()['stats']['stats_fields'][field]\n date_min = stats_date_field['min']\n date_max = stats_date_field['max']\n if start != '*':\n date_min = start\n if end != '*':\n date_max = end\n time_filter = '[{0} TO {1}]'.format(date_min, date_max)\n return time_filter\n", "role": "user" }, { "content": "asterisk_to_min_max", "role": "assistant" } ]
def x(raw_xml): from pycsw.core.etree import etree parsed = etree.fromstring(raw_xml, etree.XMLParser(resolve_entities=False)) source_tag = parsed.find('{http://purl.org/dc/elements/1.1/}source') format_tag = parsed.find('{http://purl.org/dc/elements/1.1/}format') service_url = None service_type = None if hasattr(source_tag, 'text'): service_url = source_tag.text if hasattr(format_tag, 'text'): service_type = format_tag.text if hasattr(format_tag, 'text'): service_type = format_tag.text service, created = Service.objects.get_or_create(url=service_url, is_monitored=False, type=service_type) srs, created = SpatialReferenceSystem.objects.get_or_create(code= 'EPSG:4326') service.srs.add(srs) return service
get_service
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(raw_xml):\n from pycsw.core.etree import etree\n parsed = etree.fromstring(raw_xml, etree.XMLParser(resolve_entities=False))\n source_tag = parsed.find('{http://purl.org/dc/elements/1.1/}source')\n format_tag = parsed.find('{http://purl.org/dc/elements/1.1/}format')\n service_url = None\n service_type = None\n if hasattr(source_tag, 'text'):\n service_url = source_tag.text\n if hasattr(format_tag, 'text'):\n service_type = format_tag.text\n if hasattr(format_tag, 'text'):\n service_type = format_tag.text\n service, created = Service.objects.get_or_create(url=service_url,\n is_monitored=False, type=service_type)\n srs, created = SpatialReferenceSystem.objects.get_or_create(code=\n 'EPSG:4326')\n service.srs.add(srs)\n return service\n", "role": "user" }, { "content": "get_service", "role": "assistant" } ]
def x(self, ids): results = self._get_repo_filter(Layer.objects).filter(uuid__in=ids).all() if len(results) == 0: results = self._get_repo_filter(Service.objects).filter(uuid__in=ids ).all() return results
query_ids
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ids):\n results = self._get_repo_filter(Layer.objects).filter(uuid__in=ids).all()\n if len(results) == 0:\n results = self._get_repo_filter(Service.objects).filter(uuid__in=ids\n ).all()\n return results\n", "role": "user" }, { "content": "query_ids", "role": "assistant" } ]
def x(self, domain, typenames, domainquerytype='list', count=False): objects = self._get_repo_filter(Layer.objects) if domainquerytype == 'range': return [tuple(objects.aggregate(Min(domain), Max(domain)).values())] elif count: return [(d[domain], d['%s__count' % domain]) for d in objects. values(domain).annotate(Count(domain))] else: return objects.values_list(domain).distinct()
query_domain
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, domain, typenames, domainquerytype='list', count=False):\n objects = self._get_repo_filter(Layer.objects)\n if domainquerytype == 'range':\n return [tuple(objects.aggregate(Min(domain), Max(domain)).values())]\n elif count:\n return [(d[domain], d['%s__count' % domain]) for d in objects.\n values(domain).annotate(Count(domain))]\n else:\n return objects.values_list(domain).distinct()\n", "role": "user" }, { "content": "query_domain", "role": "assistant" } ]
def x(self, direction='max'): if direction == 'min': return Layer.objects.aggregate(Min('last_updated'))['last_updated__min' ].strftime('%Y-%m-%dT%H:%M:%SZ') return self._get_repo_filter(Layer.objects).aggregate(Max('last_updated'))[ 'last_updated__max'].strftime('%Y-%m-%dT%H:%M:%SZ')
query_insert
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, direction='max'):\n if direction == 'min':\n return Layer.objects.aggregate(Min('last_updated'))['last_updated__min'\n ].strftime('%Y-%m-%dT%H:%M:%SZ')\n return self._get_repo_filter(Layer.objects).aggregate(Max('last_updated'))[\n 'last_updated__max'].strftime('%Y-%m-%dT%H:%M:%SZ')\n", "role": "user" }, { "content": "query_insert", "role": "assistant" } ]
def x(self, source): return self._get_repo_filter(Layer.objects).filter(url=source)
query_source
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, source):\n return self._get_repo_filter(Layer.objects).filter(url=source)\n", "role": "user" }, { "content": "query_source", "role": "assistant" } ]
def x(self, constraint, sortby=None, typenames=None, maxrecords=10, startposition=0): if 'where' in constraint: query = self._get_repo_filter(Layer.objects).filter(is_valid=True ).extra(where=[constraint['where']], params=constraint['values']) else: query = self._get_repo_filter(Layer.objects).filter(is_valid=True) total = query.count() if sortby is not None: if 'spatial' in sortby and sortby['spatial']: desc = False if sortby['order'] == 'DESC': desc = True query = query.all() return [str(total), sorted(query, key=lambda x: float(util. get_geometry_area(getattr(x, sortby['propertyname']))), reverse=desc)[startposition:startposition + int(maxrecords)]] else: if sortby['order'] == 'DESC': pname = '-%s' % sortby['propertyname'] else: pname = sortby['propertyname'] return [str(total), query.order_by(pname)[startposition: startposition + int(maxrecords)]] else: return [str(total), query.all()[startposition:startposition + int( maxrecords)]]
query
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, constraint, sortby=None, typenames=None, maxrecords=10,\n startposition=0):\n if 'where' in constraint:\n query = self._get_repo_filter(Layer.objects).filter(is_valid=True\n ).extra(where=[constraint['where']], params=constraint['values'])\n else:\n query = self._get_repo_filter(Layer.objects).filter(is_valid=True)\n total = query.count()\n if sortby is not None:\n if 'spatial' in sortby and sortby['spatial']:\n desc = False\n if sortby['order'] == 'DESC':\n desc = True\n query = query.all()\n return [str(total), sorted(query, key=lambda x: float(util.\n get_geometry_area(getattr(x, sortby['propertyname']))),\n reverse=desc)[startposition:startposition + int(maxrecords)]]\n else:\n if sortby['order'] == 'DESC':\n pname = '-%s' % sortby['propertyname']\n else:\n pname = sortby['propertyname']\n return [str(total), query.order_by(pname)[startposition:\n startposition + int(maxrecords)]]\n else:\n return [str(total), query.all()[startposition:startposition + int(\n maxrecords)]]\n", "role": "user" }, { "content": "query", "role": "assistant" } ]
def x(self, resourcetype, source, insert_date=None): caller = inspect.stack()[1][3] if caller == 'transaction': hhclass = 'Layer' source = resourcetype resourcetype = resourcetype.csw_schema else: hhclass = 'Service' if resourcetype not in HYPERMAP_SERVICE_TYPES.keys(): raise RuntimeError('Unsupported Service Type') return self._insert_or_update(resourcetype, source, mode='insert', hhclass=hhclass)
insert
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, resourcetype, source, insert_date=None):\n caller = inspect.stack()[1][3]\n if caller == 'transaction':\n hhclass = 'Layer'\n source = resourcetype\n resourcetype = resourcetype.csw_schema\n else:\n hhclass = 'Service'\n if resourcetype not in HYPERMAP_SERVICE_TYPES.keys():\n raise RuntimeError('Unsupported Service Type')\n return self._insert_or_update(resourcetype, source, mode='insert',\n hhclass=hhclass)\n", "role": "user" }, { "content": "insert", "role": "assistant" } ]
def x(self, resourcetype, source, mode='insert', hhclass='Service'): keywords = [] if self.filter is not None: catalog = Catalog.objects.get(id=int(self.filter.split()[-1])) try: if hhclass == 'Layer': match = Layer.objects.filter(name=source.name, title=source. title, abstract=source.abstract, is_monitored=False) matches = match.all() if matches: if mode == 'insert': raise RuntimeError( "HHypermap error: Layer %d '%s' already exists" % ( matches[0].id, source.title)) elif mode == 'update': match.update(name=source.name, title=source.title, abstract=source.abstract, is_monitored=False, xml= source.xml, wkt_geometry=source.wkt_geometry, anytext=util.get_anytext([source.title, source. abstract, source.keywords_csv])) service = get_service(source.xml) res, keywords = create_layer_from_metadata_xml(resourcetype, source.xml, monitor=False, service=service, catalog=catalog) res.save() LOGGER.debug('Indexing layer with id %s on search engine' % res .uuid) index_layer(res.id, use_cache=True) else: if resourcetype == 'http://www.opengis.net/cat/csw/2.0.2': res = Endpoint(url=source, catalog=catalog) else: res = Service(type=HYPERMAP_SERVICE_TYPES[resourcetype], url=source, catalog=catalog) res.save() if keywords: for kw in keywords: res.keywords.add(kw) except Exception as err: raise RuntimeError('HHypermap error: %s' % err) ids = [] if hhclass == 'Layer': ids.append({'identifier': res.uuid, 'title': res.title}) elif resourcetype == 'http://www.opengis.net/cat/csw/2.0.2': for res in Endpoint.objects.filter(url=source).all(): ids.append({'identifier': res.uuid, 'title': res.url}) else: for res in Service.objects.filter(url=source).all(): ids.append({'identifier': res.uuid, 'title': res.title}) return ids
insert_or_update
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, resourcetype, source, mode='insert', hhclass='Service'):\n keywords = []\n if self.filter is not None:\n catalog = Catalog.objects.get(id=int(self.filter.split()[-1]))\n try:\n if hhclass == 'Layer':\n match = Layer.objects.filter(name=source.name, title=source.\n title, abstract=source.abstract, is_monitored=False)\n matches = match.all()\n if matches:\n if mode == 'insert':\n raise RuntimeError(\n \"HHypermap error: Layer %d '%s' already exists\" % (\n matches[0].id, source.title))\n elif mode == 'update':\n match.update(name=source.name, title=source.title,\n abstract=source.abstract, is_monitored=False, xml=\n source.xml, wkt_geometry=source.wkt_geometry,\n anytext=util.get_anytext([source.title, source.\n abstract, source.keywords_csv]))\n service = get_service(source.xml)\n res, keywords = create_layer_from_metadata_xml(resourcetype,\n source.xml, monitor=False, service=service, catalog=catalog)\n res.save()\n LOGGER.debug('Indexing layer with id %s on search engine' % res\n .uuid)\n index_layer(res.id, use_cache=True)\n else:\n if resourcetype == 'http://www.opengis.net/cat/csw/2.0.2':\n res = Endpoint(url=source, catalog=catalog)\n else:\n res = Service(type=HYPERMAP_SERVICE_TYPES[resourcetype],\n url=source, catalog=catalog)\n res.save()\n if keywords:\n for kw in keywords:\n res.keywords.add(kw)\n except Exception as err:\n raise RuntimeError('HHypermap error: %s' % err)\n ids = []\n if hhclass == 'Layer':\n ids.append({'identifier': res.uuid, 'title': res.title})\n elif resourcetype == 'http://www.opengis.net/cat/csw/2.0.2':\n for res in Endpoint.objects.filter(url=source).all():\n ids.append({'identifier': res.uuid, 'title': res.url})\n else:\n for res in Service.objects.filter(url=source).all():\n ids.append({'identifier': res.uuid, 'title': res.title})\n return ids\n", "role": "user" }, { "content": "insert_or_update", "role": "assistant" } ]
def x(self, constraint): results = self._get_repo_filter(Service.objects).extra(where=[ constraint['where']], params=constraint['values']).all() deleted = len(results) results.delete() return deleted
delete
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, constraint):\n results = self._get_repo_filter(Service.objects).extra(where=[\n constraint['where']], params=constraint['values']).all()\n deleted = len(results)\n results.delete()\n return deleted\n", "role": "user" }, { "content": "delete", "role": "assistant" } ]
def x(self, query): if self.filter is not None: return query.extra(where=[self.filter]) return query
get_repo_filter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, query):\n if self.filter is not None:\n return query.extra(where=[self.filter])\n return query\n", "role": "user" }, { "content": "get_repo_filter", "role": "assistant" } ]
def x(meshcode): length = len(str(meshcode)) if length == 4: return 1 if length == 5: return 40000 if length == 6: return 2 if length == 7: if meshcode[6:7] in ['1', '2', '3', '4']: return 5000 if meshcode[6:7] == '6': return 8000 if meshcode[6:7] == '5': return 20000 if meshcode[6:7] == '7': return 16000 if length == 8: return 3 if length == 9: if meshcode[8:9] in ['1', '2', '3', '4']: return 4 if meshcode[8:9] == '5': return 2000 if meshcode[8:9] == '6': return 2500 if meshcode[8:9] == '7': return 4000 if length == 10: if meshcode[9:10] in ['1', '2', '3', '4']: return 5 if length == 11: if meshcode[10:11] in ['1', '2', '3', '4']: return 6 raise ValueError('the meshcode is unsupported.')
to_meshlevel
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(meshcode):\n length = len(str(meshcode))\n if length == 4:\n return 1\n if length == 5:\n return 40000\n if length == 6:\n return 2\n if length == 7:\n if meshcode[6:7] in ['1', '2', '3', '4']:\n return 5000\n if meshcode[6:7] == '6':\n return 8000\n if meshcode[6:7] == '5':\n return 20000\n if meshcode[6:7] == '7':\n return 16000\n if length == 8:\n return 3\n if length == 9:\n if meshcode[8:9] in ['1', '2', '3', '4']:\n return 4\n if meshcode[8:9] == '5':\n return 2000\n if meshcode[8:9] == '6':\n return 2500\n if meshcode[8:9] == '7':\n return 4000\n if length == 10:\n if meshcode[9:10] in ['1', '2', '3', '4']:\n return 5\n if length == 11:\n if meshcode[10:11] in ['1', '2', '3', '4']:\n return 6\n raise ValueError('the meshcode is unsupported.')\n", "role": "user" }, { "content": "to_meshlevel", "role": "assistant" } ]
def x(func): def x(request, *args, **kwargs): if not request.method == 'POST': return HttpResponseBadRequest('Must be POST request.') follow = func(request, *args, **kwargs) if request.is_ajax(): return HttpResponse('ok') try: if 'next' in request.GET: return HttpResponseRedirect(request.GET.get('next')) if 'next' in request.POST: return HttpResponseRedirect(request.POST.get('next')) return HttpResponseRedirect(follow.target.get_absolute_url()) except (AttributeError, TypeError): if 'HTTP_REFERER' in request.META: return HttpResponseRedirect(request.META.get('HTTP_REFERER', '/')) if follow: return HttpResponseServerError( '"%s" object of type ``%s`` has no method ``get_absolute_url()``.' % (unicode(follow.target), follow.target.__class__)) return HttpResponseServerError( 'No follow object and `next` parameter found.') return iCheck
check
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(func):\n\n def x(request, *args, **kwargs):\n if not request.method == 'POST':\n return HttpResponseBadRequest('Must be POST request.')\n follow = func(request, *args, **kwargs)\n if request.is_ajax():\n return HttpResponse('ok')\n try:\n if 'next' in request.GET:\n return HttpResponseRedirect(request.GET.get('next'))\n if 'next' in request.POST:\n return HttpResponseRedirect(request.POST.get('next'))\n return HttpResponseRedirect(follow.target.get_absolute_url())\n except (AttributeError, TypeError):\n if 'HTTP_REFERER' in request.META:\n return HttpResponseRedirect(request.META.get('HTTP_REFERER',\n '/'))\n if follow:\n return HttpResponseServerError(\n '\"%s\" object of type ``%s`` has no method ``get_absolute_url()``.'\n % (unicode(follow.target), follow.target.__class__))\n return HttpResponseServerError(\n 'No follow object and `next` parameter found.')\n return iCheck\n", "role": "user" }, { "content": "check", "role": "assistant" } ]
def x(model, field_name=None, related_name=None, lookup_method_name= 'get_follows'): if model in registry: return registry.append(model) if not field_name: field_name = 'target_%s' % model._meta.module_name if not related_name: related_name = 'follow_%s' % model._meta.module_name field = ForeignKey(model, related_name=related_name, null=True, blank= True, db_index=True) field.contribute_to_class(Follow, field_name) setattr(model, lookup_method_name, get_followers_for_object) model_map[model] = [related_name, field_name]
register
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(model, field_name=None, related_name=None, lookup_method_name=\n 'get_follows'):\n if model in registry:\n return\n registry.append(model)\n if not field_name:\n field_name = 'target_%s' % model._meta.module_name\n if not related_name:\n related_name = 'follow_%s' % model._meta.module_name\n field = ForeignKey(model, related_name=related_name, null=True, blank=\n True, db_index=True)\n field.contribute_to_class(Follow, field_name)\n setattr(model, lookup_method_name, get_followers_for_object)\n model_map[model] = [related_name, field_name]\n", "role": "user" }, { "content": "register", "role": "assistant" } ]
def x(user, obj): follow, created = Follow.objects.get_or_create(user, obj) return follow
follow
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(user, obj):\n follow, created = Follow.objects.get_or_create(user, obj)\n return follow\n", "role": "user" }, { "content": "follow", "role": "assistant" } ]
def x(user, obj): try: follow = Follow.objects.get_follows(obj).get(user=user) follow.delete() return follow except Follow.DoesNotExist: pass
unfollow
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(user, obj):\n try:\n follow = Follow.objects.get_follows(obj).get(user=user)\n follow.delete()\n return follow\n except Follow.DoesNotExist:\n pass\n", "role": "user" }, { "content": "unfollow", "role": "assistant" } ]
def x(user, obj): if Follow.objects.is_following(user, obj): return unfollow(user, obj) return follow(user, obj)
toggle
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(user, obj):\n if Follow.objects.is_following(user, obj):\n return unfollow(user, obj)\n return follow(user, obj)\n", "role": "user" }, { "content": "toggle", "role": "assistant" } ]
def x(self, value): if value: try: range = utils.parse_datetime_range_to_solr(value) return range except Exception as e: raise serializers.ValidationError(e.message) return value
validate_q_time
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, value):\n if value:\n try:\n range = utils.parse_datetime_range_to_solr(value)\n return range\n except Exception as e:\n raise serializers.ValidationError(e.message)\n return value\n", "role": "user" }, { "content": "validate_q_time", "role": "assistant" } ]
def x(self, value): if value: try: rectangle = utils.parse_geo_box(value) return '[{0},{1} TO {2},{3}]'.format(rectangle.bounds[0], rectangle.bounds[1], rectangle.bounds[2], rectangle.bounds[3]) except Exception as e: raise serializers.ValidationError(e.message) return value
validate_q_geo
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, value):\n if value:\n try:\n rectangle = utils.parse_geo_box(value)\n return '[{0},{1} TO {2},{3}]'.format(rectangle.bounds[0],\n rectangle.bounds[1], rectangle.bounds[2], rectangle.bounds[3])\n except Exception as e:\n raise serializers.ValidationError(e.message)\n return value\n", "role": "user" }, { "content": "validate_q_geo", "role": "assistant" } ]
def x(self, value): if value: try: utils.parse_datetime_range(value) except Exception as e: raise serializers.ValidationError(e.message) return value
validate_a_time_filter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, value):\n if value:\n try:\n utils.parse_datetime_range(value)\n except Exception as e:\n raise serializers.ValidationError(e.message)\n return value\n", "role": "user" }, { "content": "validate_a_time_filter", "role": "assistant" } ]
def x(self, model_or_obj_or_qs): if isinstance(model_or_obj_or_qs, QuerySet): _, fname = model_map[model_or_obj_or_qs.model] else: cls = model_or_obj_or_qs if inspect.isclass(model_or_obj_or_qs ) else model_or_obj_or_qs.__class__ _, fname = model_map[cls] return fname
fname
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, model_or_obj_or_qs):\n if isinstance(model_or_obj_or_qs, QuerySet):\n _, fname = model_map[model_or_obj_or_qs.model]\n else:\n cls = model_or_obj_or_qs if inspect.isclass(model_or_obj_or_qs\n ) else model_or_obj_or_qs.__class__\n _, fname = model_map[cls]\n return fname\n", "role": "user" }, { "content": "fname", "role": "assistant" } ]
def x(self, user, obj, **kwargs): follow = Follow(user=user) follow.target = obj follow.save() return follow
create
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, user, obj, **kwargs):\n follow = Follow(user=user)\n follow.target = obj\n follow.save()\n return follow\n", "role": "user" }, { "content": "create", "role": "assistant" } ]
def x(self, user, obj, **kwargs): if not self.is_following(user, obj): return self.create(user, obj, **kwargs), True return self.get_follows(obj).get(user=user), False
get_or_create
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, user, obj, **kwargs):\n if not self.is_following(user, obj):\n return self.create(user, obj, **kwargs), True\n return self.get_follows(obj).get(user=user), False\n", "role": "user" }, { "content": "get_or_create", "role": "assistant" } ]
def x(self, user, obj): if isinstance(user, AnonymousUser): return False return 0 < self.get_follows(obj).filter(user=user).count()
is_following
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, user, obj):\n if isinstance(user, AnonymousUser):\n return False\n return 0 < self.get_follows(obj).filter(user=user).count()\n", "role": "user" }, { "content": "is_following", "role": "assistant" } ]
def x(self, model_or_obj_or_qs): fname = self.fname(model_or_obj_or_qs) if isinstance(model_or_obj_or_qs, QuerySet): return self.filter(**{('%s__in' % fname): model_or_obj_or_qs}) if inspect.isclass(model_or_obj_or_qs): return self.exclude(**{fname: None}) return self.filter(**{fname: model_or_obj_or_qs})
get_follows
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, model_or_obj_or_qs):\n fname = self.fname(model_or_obj_or_qs)\n if isinstance(model_or_obj_or_qs, QuerySet):\n return self.filter(**{('%s__in' % fname): model_or_obj_or_qs})\n if inspect.isclass(model_or_obj_or_qs):\n return self.exclude(**{fname: None})\n return self.filter(**{fname: model_or_obj_or_qs})\n", "role": "user" }, { "content": "get_follows", "role": "assistant" } ]
def x(self, event_times_indices, covariates=None, durations=None): if covariates is None: covariates = np.ones(self.event_times_indices.shape) if durations is None: durations = np.ones(self.event_times_indices.shape) else: durations = np.round(durations * self.deconvolution_frequency).astype( int) mean_duration = np.mean(durations) regressors_for_event = np.zeros((self.deconvolution_interval_size, self .resampled_signal_size)) for cov, eti, dur in zip(covariates, event_times_indices, durations): valid = True if eti < 0: self.logger.debug( 'deconv samples are starting before the data starts.') valid = False if eti + self.deconvolution_interval_size > self.resampled_signal_size: self.logger.debug( 'deconv samples are continuing after the data stops.') valid = False if eti > self.resampled_signal_size: self.logger.debug('event falls outside of the scope of the data.') valid = False if valid: this_event_design_matrix = np.diag(np.ones(self. deconvolution_interval_size)) * cov over_durations_dm = np.copy(this_event_design_matrix) if dur > 1: for d in np.arange(1, dur): over_durations_dm[d:] += this_event_design_matrix[:-d] over_durations_dm /= mean_duration regressors_for_event[:, eti:int(eti + self. deconvolution_interval_size)] += over_durations_dm return regressors_for_event
create_event_regressors
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, event_times_indices, covariates=None, durations=None):\n if covariates is None:\n covariates = np.ones(self.event_times_indices.shape)\n if durations is None:\n durations = np.ones(self.event_times_indices.shape)\n else:\n durations = np.round(durations * self.deconvolution_frequency).astype(\n int)\n mean_duration = np.mean(durations)\n regressors_for_event = np.zeros((self.deconvolution_interval_size, self\n .resampled_signal_size))\n for cov, eti, dur in zip(covariates, event_times_indices, durations):\n valid = True\n if eti < 0:\n self.logger.debug(\n 'deconv samples are starting before the data starts.')\n valid = False\n if eti + self.deconvolution_interval_size > self.resampled_signal_size:\n self.logger.debug(\n 'deconv samples are continuing after the data stops.')\n valid = False\n if eti > self.resampled_signal_size:\n self.logger.debug('event falls outside of the scope of the data.')\n valid = False\n if valid:\n this_event_design_matrix = np.diag(np.ones(self.\n deconvolution_interval_size)) * cov\n over_durations_dm = np.copy(this_event_design_matrix)\n if dur > 1:\n for d in np.arange(1, dur):\n over_durations_dm[d:] += this_event_design_matrix[:-d]\n over_durations_dm /= mean_duration\n regressors_for_event[:, eti:int(eti + self.\n deconvolution_interval_size)] += over_durations_dm\n return regressors_for_event\n", "role": "user" }, { "content": "create_event_regressors", "role": "assistant" } ]
def x(self, demean=False, intercept=True): self.design_matrix = np.zeros((int(self.number_of_event_types * self. deconvolution_interval_size), self.resampled_signal_size)) for i, covariate in enumerate(self.covariates.keys()): self.logger.debug('creating regressor for ' + covariate) indices = np.arange(i * self.deconvolution_interval_size, (i + 1) * self.deconvolution_interval_size, dtype=int) if len(covariate.split('.')) > 0: which_event_time_indices = covariate.split('.')[0] else: which_event_time_indices = covariate self.design_matrix[indices] = self.create_event_regressors(self. event_times_indices[which_event_time_indices], self.covariates[ covariate], self.durations[which_event_time_indices]) if demean: self.design_matrix = (self.design_matrix.T - self.design_matrix. mean(axis=-1)).T if intercept: self.design_matrix = np.vstack((self.design_matrix, np.ones((1, self.design_matrix.shape[-1])))) self.logger.debug('created %s design_matrix' % str(self.design_matrix. shape))
create_design_matrix
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, demean=False, intercept=True):\n self.design_matrix = np.zeros((int(self.number_of_event_types * self.\n deconvolution_interval_size), self.resampled_signal_size))\n for i, covariate in enumerate(self.covariates.keys()):\n self.logger.debug('creating regressor for ' + covariate)\n indices = np.arange(i * self.deconvolution_interval_size, (i + 1) *\n self.deconvolution_interval_size, dtype=int)\n if len(covariate.split('.')) > 0:\n which_event_time_indices = covariate.split('.')[0]\n else:\n which_event_time_indices = covariate\n self.design_matrix[indices] = self.create_event_regressors(self.\n event_times_indices[which_event_time_indices], self.covariates[\n covariate], self.durations[which_event_time_indices])\n if demean:\n self.design_matrix = (self.design_matrix.T - self.design_matrix.\n mean(axis=-1)).T\n if intercept:\n self.design_matrix = np.vstack((self.design_matrix, np.ones((1,\n self.design_matrix.shape[-1]))))\n self.logger.debug('created %s design_matrix' % str(self.design_matrix.\n shape))\n", "role": "user" }, { "content": "create_design_matrix", "role": "assistant" } ]
def x(self, regressors): previous_design_matrix_shape = self.design_matrix.shape if len(regressors.shape) == 1: regressors = regressors[np.newaxis, :] if regressors.shape[1] != self.resampled_signal.shape[1]: self.logger.warning( 'additional regressor shape %s does not conform to designmatrix shape %s' % (regressors.shape, self.resampled_signal.shape)) self.design_matrix = np.vstack((self.design_matrix, regressors)) self.logger.debug( 'added %s continuous regressors to %s design_matrix, shape now %s' % (str(regressors.shape), str(previous_design_matrix_shape), str(self .design_matrix.shape)))
add_continuous_regressors_to_design_matrix
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, regressors):\n previous_design_matrix_shape = self.design_matrix.shape\n if len(regressors.shape) == 1:\n regressors = regressors[np.newaxis, :]\n if regressors.shape[1] != self.resampled_signal.shape[1]:\n self.logger.warning(\n 'additional regressor shape %s does not conform to designmatrix shape %s'\n % (regressors.shape, self.resampled_signal.shape))\n self.design_matrix = np.vstack((self.design_matrix, regressors))\n self.logger.debug(\n 'added %s continuous regressors to %s design_matrix, shape now %s' %\n (str(regressors.shape), str(previous_design_matrix_shape), str(self\n .design_matrix.shape)))\n", "role": "user" }, { "content": "add_continuous_regressors_to_design_matrix", "role": "assistant" } ]
def x(self, method='lstsq'): if method is 'lstsq': self.betas, residuals_sum, rank, s = LA.lstsq(self.design_matrix.T, self.resampled_signal.T) self.residuals = (self.resampled_signal - self. predict_from_design_matrix(self.design_matrix)) elif method is 'sm_ols': import statsmodels.api as sm assert self.resampled_signal.shape[0 ] == 1, 'signal input into statsmodels OLS cannot contain multiple signals at once, present shape %s' % str( self.resampled_signal.shape) model = sm.OLS(np.squeeze(self.resampled_signal), self.design_matrix.T) results = model.fit() self.betas = np.array(results.params).reshape((self.design_matrix. shape[0], self.resampled_signal.shape[0])) self.residuals = np.array(results.resid).reshape(self. resampled_signal.shape) self.logger.debug( 'performed %s regression on %s design_matrix and %s signal' % ( method, str(self.design_matrix.shape), str(self.resampled_signal. shape)))
regress
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, method='lstsq'):\n if method is 'lstsq':\n self.betas, residuals_sum, rank, s = LA.lstsq(self.design_matrix.T,\n self.resampled_signal.T)\n self.residuals = (self.resampled_signal - self.\n predict_from_design_matrix(self.design_matrix))\n elif method is 'sm_ols':\n import statsmodels.api as sm\n assert self.resampled_signal.shape[0\n ] == 1, 'signal input into statsmodels OLS cannot contain multiple signals at once, present shape %s' % str(\n self.resampled_signal.shape)\n model = sm.OLS(np.squeeze(self.resampled_signal), self.design_matrix.T)\n results = model.fit()\n self.betas = np.array(results.params).reshape((self.design_matrix.\n shape[0], self.resampled_signal.shape[0]))\n self.residuals = np.array(results.resid).reshape(self.\n resampled_signal.shape)\n self.logger.debug(\n 'performed %s regression on %s design_matrix and %s signal' % (\n method, str(self.design_matrix.shape), str(self.resampled_signal.\n shape)))\n", "role": "user" }, { "content": "regress", "role": "assistant" } ]
def x(self, cv=20, alphas=None): if alphas is None: alphas = np.logspace(7, 0, 20) self.rcv = linear_model.RidgeCV(alphas=alphas, fit_intercept=False, cv=cv) self.rcv.fit(self.design_matrix.T, self.resampled_signal.T) self.betas = self.rcv.coef_.T self.residuals = self.resampled_signal - self.rcv.predict(self. design_matrix.T) self.logger.debug( 'performed ridge regression on %s design_matrix and %s signal, resulting alpha value is %f' % (str(self.design_matrix.shape), str(self.resampled_signal.shape), self.rcv.alpha_))
ridge_regress
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, cv=20, alphas=None):\n if alphas is None:\n alphas = np.logspace(7, 0, 20)\n self.rcv = linear_model.RidgeCV(alphas=alphas, fit_intercept=False, cv=cv)\n self.rcv.fit(self.design_matrix.T, self.resampled_signal.T)\n self.betas = self.rcv.coef_.T\n self.residuals = self.resampled_signal - self.rcv.predict(self.\n design_matrix.T)\n self.logger.debug(\n 'performed ridge regression on %s design_matrix and %s signal, resulting alpha value is %f'\n % (str(self.design_matrix.shape), str(self.resampled_signal.shape),\n self.rcv.alpha_))\n", "role": "user" }, { "content": "ridge_regress", "role": "assistant" } ]
def x(self, covariate='0'): this_covariate_index = list(self.covariates.keys()).index(covariate) return self.betas[int(this_covariate_index * self. deconvolution_interval_size):int((this_covariate_index + 1) * self. deconvolution_interval_size)]
betas_for_cov
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, covariate='0'):\n this_covariate_index = list(self.covariates.keys()).index(covariate)\n return self.betas[int(this_covariate_index * self.\n deconvolution_interval_size):int((this_covariate_index + 1) * self.\n deconvolution_interval_size)]\n", "role": "user" }, { "content": "betas_for_cov", "role": "assistant" } ]
def x(self): self.betas_per_event_type = np.zeros((len(self.covariates), self. deconvolution_interval_size, self.resampled_signal.shape[0])) for i, covariate in enumerate(self.covariates.keys()): self.betas_per_event_type[i] = self.betas_for_cov(covariate)
betas_for_events
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.betas_per_event_type = np.zeros((len(self.covariates), self.\n deconvolution_interval_size, self.resampled_signal.shape[0]))\n for i, covariate in enumerate(self.covariates.keys()):\n self.betas_per_event_type[i] = self.betas_for_cov(covariate)\n", "role": "user" }, { "content": "betas_for_events", "role": "assistant" } ]
def x(self, design_matrix): assert hasattr(self, 'betas' ), 'no betas found, please run regression before prediction' assert design_matrix.shape[0] == self.betas.shape[0 ], 'designmatrix needs to have the same number of regressors as the betas already calculated' prediction = np.dot(self.betas.astype(np.float32).T, design_matrix. astype(np.float32)) return prediction
predict_from_design_matrix
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, design_matrix):\n assert hasattr(self, 'betas'\n ), 'no betas found, please run regression before prediction'\n assert design_matrix.shape[0] == self.betas.shape[0\n ], 'designmatrix needs to have the same number of regressors as the betas already calculated'\n prediction = np.dot(self.betas.astype(np.float32).T, design_matrix.\n astype(np.float32))\n return prediction\n", "role": "user" }, { "content": "predict_from_design_matrix", "role": "assistant" } ]
def x(self): assert hasattr(self, 'betas' ), 'no betas found, please run regression before rsq' explained_times = self.design_matrix.sum(axis=0) != 0 explained_signal = self.predict_from_design_matrix(self.design_matrix) self.rsq = 1.0 - np.sum((explained_signal[:, explained_times] - self. resampled_signal[:, explained_times]) ** 2, axis=-1) / np.sum(self. resampled_signal[:, explained_times].squeeze() ** 2, axis=-1) self.ssr = np.sum((explained_signal[:, explained_times] - self. resampled_signal[:, explained_times]) ** 2, axis=-1) return np.squeeze(self.rsq)
calculate_rsq
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n assert hasattr(self, 'betas'\n ), 'no betas found, please run regression before rsq'\n explained_times = self.design_matrix.sum(axis=0) != 0\n explained_signal = self.predict_from_design_matrix(self.design_matrix)\n self.rsq = 1.0 - np.sum((explained_signal[:, explained_times] - self.\n resampled_signal[:, explained_times]) ** 2, axis=-1) / np.sum(self.\n resampled_signal[:, explained_times].squeeze() ** 2, axis=-1)\n self.ssr = np.sum((explained_signal[:, explained_times] - self.\n resampled_signal[:, explained_times]) ** 2, axis=-1)\n return np.squeeze(self.rsq)\n", "role": "user" }, { "content": "calculate_rsq", "role": "assistant" } ]
def x(self, nr_repetitions=1000): assert self.resampled_signal.shape[0 ] == 1, 'signal input into bootstrap_on_residuals cannot contain signals from multiple channels at once, present shape %s' % str( self.resampled_signal.shape) assert hasattr(self, 'betas' ), 'no betas found, please run regression before bootstrapping' bootstrap_data = np.zeros((self.resampled_signal_size, nr_repetitions)) explained_signal = self.predict_from_design_matrix(self.design_matrix).T for x in range(bootstrap_data.shape[-1]): bootstrap_data[:, x] = (self.residuals.T[np.random.permutation(self .resampled_signal_size)] + explained_signal).squeeze() self.bootstrap_betas, bs_residuals, rank, s = LA.lstsq(self. design_matrix.T, bootstrap_data) self.bootstrap_betas_per_event_type = np.zeros((len(self.covariates), self.deconvolution_interval_size, nr_repetitions)) for i, covariate in enumerate(list(self.covariates.keys())): this_covariate_index = list(self.covariates.keys()).index(covariate) self.bootstrap_betas_per_event_type[i] = self.bootstrap_betas[ this_covariate_index * self.deconvolution_interval_size:( this_covariate_index + 1) * self.deconvolution_interval_size]
bootstrap_on_residuals
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, nr_repetitions=1000):\n assert self.resampled_signal.shape[0\n ] == 1, 'signal input into bootstrap_on_residuals cannot contain signals from multiple channels at once, present shape %s' % str(\n self.resampled_signal.shape)\n assert hasattr(self, 'betas'\n ), 'no betas found, please run regression before bootstrapping'\n bootstrap_data = np.zeros((self.resampled_signal_size, nr_repetitions))\n explained_signal = self.predict_from_design_matrix(self.design_matrix).T\n for x in range(bootstrap_data.shape[-1]):\n bootstrap_data[:, x] = (self.residuals.T[np.random.permutation(self\n .resampled_signal_size)] + explained_signal).squeeze()\n self.bootstrap_betas, bs_residuals, rank, s = LA.lstsq(self.\n design_matrix.T, bootstrap_data)\n self.bootstrap_betas_per_event_type = np.zeros((len(self.covariates),\n self.deconvolution_interval_size, nr_repetitions))\n for i, covariate in enumerate(list(self.covariates.keys())):\n this_covariate_index = list(self.covariates.keys()).index(covariate)\n self.bootstrap_betas_per_event_type[i] = self.bootstrap_betas[\n this_covariate_index * self.deconvolution_interval_size:(\n this_covariate_index + 1) * self.deconvolution_interval_size]\n", "role": "user" }, { "content": "bootstrap_on_residuals", "role": "assistant" } ]
def x(request): url_parsed = urlparse(settings.SEARCH_URL) defaults = dict(APP_NAME=__description__, APP_VERSION=__version__, SITE_URL=settings.SITE_URL.rstrip('/'), SEARCH_TYPE=settings. SEARCH_TYPE, SEARCH_URL=settings.SEARCH_URL, SEARCH_IP='%s://%s:%s' % (url_parsed.scheme, url_parsed.hostname, url_parsed.port)) return defaults
resource_urls
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(request):\n url_parsed = urlparse(settings.SEARCH_URL)\n defaults = dict(APP_NAME=__description__, APP_VERSION=__version__,\n SITE_URL=settings.SITE_URL.rstrip('/'), SEARCH_TYPE=settings.\n SEARCH_TYPE, SEARCH_URL=settings.SEARCH_URL, SEARCH_IP='%s://%s:%s' %\n (url_parsed.scheme, url_parsed.hostname, url_parsed.port))\n return defaults\n", "role": "user" }, { "content": "resource_urls", "role": "assistant" } ]
def x(self): from hypermap.aggregator.models import Layer if SEARCH_TYPE == 'solr': from hypermap.aggregator.solr import SolrHypermap solrobject = SolrHypermap() else: from hypermap.aggregator.elasticsearch_client import ESHypermap from elasticsearch import helpers es_client = ESHypermap() layers_cache = cache.get('layers') deleted_layers_cache = cache.get('deleted_layers') if layers_cache: layers_list = list(layers_cache) LOGGER.debug('There are %s layers in cache: %s' % (len(layers_list), layers_list)) batch_size = settings.REGISTRY_SEARCH_BATCH_SIZE batch_lists = [layers_list[i:i + batch_size] for i in range(0, len( layers_list), batch_size)] for batch_list_ids in batch_lists: layers = Layer.objects.filter(id__in=batch_list_ids) if batch_size > len(layers): batch_size = len(layers) LOGGER.debug('Syncing %s/%s layers to %s: %s' % (batch_size, len(layers_cache), layers, SEARCH_TYPE)) try: if SEARCH_TYPE == 'solr': success, layers_errors_ids = solrobject.layers_to_solr( layers) if success: layers_cache = layers_cache.difference(set( batch_list_ids)) LOGGER.debug( 'Removing layers with id %s from cache' % batch_list_ids) cache.set('layers', layers_cache) elif SEARCH_TYPE == 'elasticsearch': with_bulk, success = True, False layers_to_index = [es_client.layer_to_es(layer, with_bulk) for layer in layers] message = helpers.bulk(es_client.es, layers_to_index) len_indexed_layers = message[0] if len_indexed_layers == len(layers): LOGGER.debug('%d layers indexed successfully' % len_indexed_layers) success = True if success: layers_cache = layers_cache.difference(set( batch_list_ids)) cache.set('layers', layers_cache) else: raise Exception('Incorrect SEARCH_TYPE=%s' % SEARCH_TYPE) except Exception as e: LOGGER.error('Layers were NOT indexed correctly') LOGGER.error(e, exc_info=True) else: LOGGER.debug('No cached layers to add in search engine.') if deleted_layers_cache: layers_list = list(deleted_layers_cache) LOGGER.debug('There are %s layers in cache for deleting: %s' % (len (layers_list), layers_list)) for layer_id in layers_list: if SEARCH_TYPE == 'solr': if Layer.objects.filter(pk=layer_id).exists(): layer = Layer.objects.get(id=layer_id) unindex_layer(layer.id, use_cache=False) deleted_layers_cache = deleted_layers_cache.difference(set ([layer_id])) cache.set('deleted_layers', deleted_layers_cache) else: raise NotImplementedError else: LOGGER.debug('No cached layers to remove in search engine.')
index_cached_layers
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from hypermap.aggregator.models import Layer\n if SEARCH_TYPE == 'solr':\n from hypermap.aggregator.solr import SolrHypermap\n solrobject = SolrHypermap()\n else:\n from hypermap.aggregator.elasticsearch_client import ESHypermap\n from elasticsearch import helpers\n es_client = ESHypermap()\n layers_cache = cache.get('layers')\n deleted_layers_cache = cache.get('deleted_layers')\n if layers_cache:\n layers_list = list(layers_cache)\n LOGGER.debug('There are %s layers in cache: %s' % (len(layers_list),\n layers_list))\n batch_size = settings.REGISTRY_SEARCH_BATCH_SIZE\n batch_lists = [layers_list[i:i + batch_size] for i in range(0, len(\n layers_list), batch_size)]\n for batch_list_ids in batch_lists:\n layers = Layer.objects.filter(id__in=batch_list_ids)\n if batch_size > len(layers):\n batch_size = len(layers)\n LOGGER.debug('Syncing %s/%s layers to %s: %s' % (batch_size,\n len(layers_cache), layers, SEARCH_TYPE))\n try:\n if SEARCH_TYPE == 'solr':\n success, layers_errors_ids = solrobject.layers_to_solr(\n layers)\n if success:\n layers_cache = layers_cache.difference(set(\n batch_list_ids))\n LOGGER.debug(\n 'Removing layers with id %s from cache' %\n batch_list_ids)\n cache.set('layers', layers_cache)\n elif SEARCH_TYPE == 'elasticsearch':\n with_bulk, success = True, False\n layers_to_index = [es_client.layer_to_es(layer,\n with_bulk) for layer in layers]\n message = helpers.bulk(es_client.es, layers_to_index)\n len_indexed_layers = message[0]\n if len_indexed_layers == len(layers):\n LOGGER.debug('%d layers indexed successfully' %\n len_indexed_layers)\n success = True\n if success:\n layers_cache = layers_cache.difference(set(\n batch_list_ids))\n cache.set('layers', layers_cache)\n else:\n raise Exception('Incorrect SEARCH_TYPE=%s' % SEARCH_TYPE)\n except Exception as e:\n LOGGER.error('Layers were NOT indexed correctly')\n LOGGER.error(e, exc_info=True)\n else:\n LOGGER.debug('No cached layers to add in search engine.')\n if deleted_layers_cache:\n layers_list = list(deleted_layers_cache)\n LOGGER.debug('There are %s layers in cache for deleting: %s' % (len\n (layers_list), layers_list))\n for layer_id in layers_list:\n if SEARCH_TYPE == 'solr':\n if Layer.objects.filter(pk=layer_id).exists():\n layer = Layer.objects.get(id=layer_id)\n unindex_layer(layer.id, use_cache=False)\n deleted_layers_cache = deleted_layers_cache.difference(set\n ([layer_id]))\n cache.set('deleted_layers', deleted_layers_cache)\n else:\n raise NotImplementedError\n else:\n LOGGER.debug('No cached layers to remove in search engine.')\n", "role": "user" }, { "content": "index_cached_layers", "role": "assistant" } ]
def x(self, service_id): from hypermap.aggregator.models import Service service = Service.objects.get(id=service_id) service.check_set.all().delete() layer_to_process = service.layer_set.all() for layer in layer_to_process: layer.check_set.all().delete()
remove_service_checks
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, service_id):\n from hypermap.aggregator.models import Service\n service = Service.objects.get(id=service_id)\n service.check_set.all().delete()\n layer_to_process = service.layer_set.all()\n for layer in layer_to_process:\n layer.check_set.all().delete()\n", "role": "user" }, { "content": "remove_service_checks", "role": "assistant" } ]
def x(self, service_id): from hypermap.aggregator.models import Service service = Service.objects.get(id=service_id) if not service.is_valid: LOGGER.debug( 'Not indexing service with id %s in search engine as it is not valid' % service.id) return LOGGER.debug('Indexing service %s' % service.id) layer_to_process = service.layer_set.all() for layer in layer_to_process: if not settings.REGISTRY_SKIP_CELERY: index_layer(layer.id, use_cache=True) else: index_layer(layer.id)
index_service
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, service_id):\n from hypermap.aggregator.models import Service\n service = Service.objects.get(id=service_id)\n if not service.is_valid:\n LOGGER.debug(\n 'Not indexing service with id %s in search engine as it is not valid'\n % service.id)\n return\n LOGGER.debug('Indexing service %s' % service.id)\n layer_to_process = service.layer_set.all()\n for layer in layer_to_process:\n if not settings.REGISTRY_SKIP_CELERY:\n index_layer(layer.id, use_cache=True)\n else:\n index_layer(layer.id)\n", "role": "user" }, { "content": "index_service", "role": "assistant" } ]
def x(self, layer_id, use_cache=False): from hypermap.aggregator.models import Layer layer = Layer.objects.get(id=layer_id) if not layer.is_valid: LOGGER.debug( 'Not indexing or removing layer with id %s in search engine as it is not valid' % layer.id) unindex_layer(layer.id, use_cache) return if layer.was_deleted: LOGGER.debug( 'Not indexing or removing layer with id %s in search engine as was_deleted is true' % layer.id) unindex_layer(layer.id, use_cache) return if use_cache: LOGGER.debug( 'Caching layer with id %s for syncing with search engine' % layer.id) layers = cache.get('layers') if layers is None: layers = set([layer.id]) else: layers.add(layer.id) cache.set('layers', layers) return if SEARCH_TYPE == 'solr': from hypermap.aggregator.solr import SolrHypermap LOGGER.debug('Syncing layer %s to solr' % layer.name) solrobject = SolrHypermap() success, message = solrobject.layer_to_solr(layer) if not settings.REGISTRY_SKIP_CELERY: if not success: self.update_state(state=states.FAILURE, meta=message) raise Ignore() elif SEARCH_TYPE == 'elasticsearch': from hypermap.aggregator.elasticsearch_client import ESHypermap LOGGER.debug('Syncing layer %s to es' % layer.name) esobject = ESHypermap() success, message = esobject.layer_to_es(layer) if not settings.REGISTRY_SKIP_CELERY: if not success: self.update_state(state=states.FAILURE, meta=message) raise Ignore()
index_layer
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, layer_id, use_cache=False):\n from hypermap.aggregator.models import Layer\n layer = Layer.objects.get(id=layer_id)\n if not layer.is_valid:\n LOGGER.debug(\n 'Not indexing or removing layer with id %s in search engine as it is not valid'\n % layer.id)\n unindex_layer(layer.id, use_cache)\n return\n if layer.was_deleted:\n LOGGER.debug(\n 'Not indexing or removing layer with id %s in search engine as was_deleted is true'\n % layer.id)\n unindex_layer(layer.id, use_cache)\n return\n if use_cache:\n LOGGER.debug(\n 'Caching layer with id %s for syncing with search engine' %\n layer.id)\n layers = cache.get('layers')\n if layers is None:\n layers = set([layer.id])\n else:\n layers.add(layer.id)\n cache.set('layers', layers)\n return\n if SEARCH_TYPE == 'solr':\n from hypermap.aggregator.solr import SolrHypermap\n LOGGER.debug('Syncing layer %s to solr' % layer.name)\n solrobject = SolrHypermap()\n success, message = solrobject.layer_to_solr(layer)\n if not settings.REGISTRY_SKIP_CELERY:\n if not success:\n self.update_state(state=states.FAILURE, meta=message)\n raise Ignore()\n elif SEARCH_TYPE == 'elasticsearch':\n from hypermap.aggregator.elasticsearch_client import ESHypermap\n LOGGER.debug('Syncing layer %s to es' % layer.name)\n esobject = ESHypermap()\n success, message = esobject.layer_to_es(layer)\n if not settings.REGISTRY_SKIP_CELERY:\n if not success:\n self.update_state(state=states.FAILURE, meta=message)\n raise Ignore()\n", "role": "user" }, { "content": "index_layer", "role": "assistant" } ]
def x(self, use_cache=False): from hypermap.aggregator.models import Issue, Layer, Service from django.contrib.contenttypes.models import ContentType layer_type = ContentType.objects.get_for_model(Layer) service_type = ContentType.objects.get_for_model(Service) for issue in Issue.objects.filter(content_type__pk=layer_type.id): unindex_layer(issue.content_object.id, use_cache) for issue in Issue.objects.filter(content_type__pk=service_type.id): for layer in issue.content_object.layer_set.all(): unindex_layer(layer.id, use_cache)
unindex_layers_with_issues
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, use_cache=False):\n from hypermap.aggregator.models import Issue, Layer, Service\n from django.contrib.contenttypes.models import ContentType\n layer_type = ContentType.objects.get_for_model(Layer)\n service_type = ContentType.objects.get_for_model(Service)\n for issue in Issue.objects.filter(content_type__pk=layer_type.id):\n unindex_layer(issue.content_object.id, use_cache)\n for issue in Issue.objects.filter(content_type__pk=service_type.id):\n for layer in issue.content_object.layer_set.all():\n unindex_layer(layer.id, use_cache)\n", "role": "user" }, { "content": "unindex_layers_with_issues", "role": "assistant" } ]
def x(self, layer_id, use_cache=False): from hypermap.aggregator.models import Layer layer = Layer.objects.get(id=layer_id) if use_cache: LOGGER.debug( 'Caching layer with id %s for being removed from search engine' % layer.id) deleted_layers = cache.get('deleted_layers') if deleted_layers is None: deleted_layers = set([layer.id]) else: deleted_layers.add(layer.id) cache.set('deleted_layers', deleted_layers) return if SEARCH_TYPE == 'solr': from hypermap.aggregator.solr import SolrHypermap LOGGER.debug('Removing layer %s from solr' % layer.id) try: solrobject = SolrHypermap() solrobject.remove_layer(layer.uuid) except Exception: LOGGER.error('Layer NOT correctly removed from Solr') elif SEARCH_TYPE == 'elasticsearch': pass
unindex_layer
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, layer_id, use_cache=False):\n from hypermap.aggregator.models import Layer\n layer = Layer.objects.get(id=layer_id)\n if use_cache:\n LOGGER.debug(\n 'Caching layer with id %s for being removed from search engine' %\n layer.id)\n deleted_layers = cache.get('deleted_layers')\n if deleted_layers is None:\n deleted_layers = set([layer.id])\n else:\n deleted_layers.add(layer.id)\n cache.set('deleted_layers', deleted_layers)\n return\n if SEARCH_TYPE == 'solr':\n from hypermap.aggregator.solr import SolrHypermap\n LOGGER.debug('Removing layer %s from solr' % layer.id)\n try:\n solrobject = SolrHypermap()\n solrobject.remove_layer(layer.uuid)\n except Exception:\n LOGGER.error('Layer NOT correctly removed from Solr')\n elif SEARCH_TYPE == 'elasticsearch':\n pass\n", "role": "user" }, { "content": "unindex_layer", "role": "assistant" } ]
def x(self): from hypermap.aggregator.models import Layer if not settings.REGISTRY_SKIP_CELERY: layers_cache = set(Layer.objects.filter(is_valid=True).values_list( 'id', flat=True)) deleted_layers_cache = set(Layer.objects.filter(is_valid=False). values_list('id', flat=True)) cache.set('layers', layers_cache) cache.set('deleted_layers', deleted_layers_cache) else: for layer in Layer.objects.all(): index_layer(layer.id)
index_all_layers
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from hypermap.aggregator.models import Layer\n if not settings.REGISTRY_SKIP_CELERY:\n layers_cache = set(Layer.objects.filter(is_valid=True).values_list(\n 'id', flat=True))\n deleted_layers_cache = set(Layer.objects.filter(is_valid=False).\n values_list('id', flat=True))\n cache.set('layers', layers_cache)\n cache.set('deleted_layers', deleted_layers_cache)\n else:\n for layer in Layer.objects.all():\n index_layer(layer.id)\n", "role": "user" }, { "content": "index_all_layers", "role": "assistant" } ]
def x(self, service_id, num_layers=10): from hypermap.aggregator.models import Service LOGGER.debug( 'Updating the index the last %s added and %s deleted layers in WorldMap service' % (num_layers, num_layers)) service = Service.objects.get(id=service_id) if service.type == 'Hypermap:WorldMapLegacy': from hypermap.aggregator.models import update_layers_wm_legacy as update_layers_wm if service.type == 'Hypermap:WorldMap': from hypermap.aggregator.models import update_layers_geonode_wm as update_layers_wm update_layers_wm(service, num_layers) LOGGER.debug('Removing the index for the last %s deleted layers' % num_layers) layer_to_unindex = service.layer_set.filter(was_deleted=True).order_by( '-last_updated')[0:num_layers] for layer in layer_to_unindex: if not settings.REGISTRY_SKIP_CELERY: unindex_layer(layer.id, use_cache=True) else: unindex_layer(layer.id) LOGGER.debug('Adding/Updating the index for the last %s added layers' % num_layers) layer_to_index = service.layer_set.filter(was_deleted=False).order_by( '-last_updated')[0:num_layers] for layer in layer_to_index: if not settings.REGISTRY_SKIP_CELERY: index_layer(layer.id, use_cache=True) else: index_layer(layer.id)
update_last_wm_layers
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, service_id, num_layers=10):\n from hypermap.aggregator.models import Service\n LOGGER.debug(\n 'Updating the index the last %s added and %s deleted layers in WorldMap service'\n % (num_layers, num_layers))\n service = Service.objects.get(id=service_id)\n if service.type == 'Hypermap:WorldMapLegacy':\n from hypermap.aggregator.models import update_layers_wm_legacy as update_layers_wm\n if service.type == 'Hypermap:WorldMap':\n from hypermap.aggregator.models import update_layers_geonode_wm as update_layers_wm\n update_layers_wm(service, num_layers)\n LOGGER.debug('Removing the index for the last %s deleted layers' %\n num_layers)\n layer_to_unindex = service.layer_set.filter(was_deleted=True).order_by(\n '-last_updated')[0:num_layers]\n for layer in layer_to_unindex:\n if not settings.REGISTRY_SKIP_CELERY:\n unindex_layer(layer.id, use_cache=True)\n else:\n unindex_layer(layer.id)\n LOGGER.debug('Adding/Updating the index for the last %s added layers' %\n num_layers)\n layer_to_index = service.layer_set.filter(was_deleted=False).order_by(\n '-last_updated')[0:num_layers]\n for layer in layer_to_index:\n if not settings.REGISTRY_SKIP_CELERY:\n index_layer(layer.id, use_cache=True)\n else:\n index_layer(layer.id)\n", "role": "user" }, { "content": "update_last_wm_layers", "role": "assistant" } ]
def x(bbox): try: minx = float(bbox[0]) miny = float(bbox[1]) maxx = float(bbox[2]) maxy = float(bbox[3]) except: LOGGER.debug('Invalid bbox, setting it to a zero POLYGON') minx = 0 miny = 0 maxx = 0 maxy = 0 return ( 'POLYGON((%.2f %.2f, %.2f %.2f, %.2f %.2f, %.2f %.2f, %.2f %.2f))' % (minx, miny, minx, maxy, maxx, maxy, maxx, miny, minx, miny))
bbox2wktpolygon
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bbox):\n try:\n minx = float(bbox[0])\n miny = float(bbox[1])\n maxx = float(bbox[2])\n maxy = float(bbox[3])\n except:\n LOGGER.debug('Invalid bbox, setting it to a zero POLYGON')\n minx = 0\n miny = 0\n maxx = 0\n maxy = 0\n return (\n 'POLYGON((%.2f %.2f, %.2f %.2f, %.2f %.2f, %.2f %.2f, %.2f %.2f))' %\n (minx, miny, minx, maxy, maxx, maxy, maxx, miny, minx, miny))\n", "role": "user" }, { "content": "bbox2wktpolygon", "role": "assistant" } ]
def x(**kwargs): if 'srs' in kwargs: srs = kwargs['srs'] else: srs = '4326' modified = '%sZ' % datetime.datetime.utcnow().isoformat().split('.')[0] nsmap = Namespaces().get_namespaces(['csw', 'dc', 'dct', 'ows']) e = etree.Element(nspath_eval('csw:Record', nsmap), nsmap=nsmap) etree.SubElement(e, nspath_eval('dc:identifier', nsmap)).text = kwargs[ 'identifier'] etree.SubElement(e, nspath_eval('dc:title', nsmap)).text = kwargs['title'] if 'alternative' in kwargs: etree.SubElement(e, nspath_eval('dct:alternative', nsmap) ).text = kwargs['alternative'] etree.SubElement(e, nspath_eval('dct:modified', nsmap)).text = modified etree.SubElement(e, nspath_eval('dct:abstract', nsmap)).text = kwargs[ 'abstract'] etree.SubElement(e, nspath_eval('dc:type', nsmap)).text = kwargs['type'] etree.SubElement(e, nspath_eval('dc:format', nsmap)).text = kwargs['format' ] etree.SubElement(e, nspath_eval('dc:source', nsmap)).text = kwargs['source' ] if 'relation' in kwargs: etree.SubElement(e, nspath_eval('dc:relation', nsmap)).text = kwargs[ 'relation'] if 'keywords' in kwargs: if kwargs['keywords'] is not None: for keyword in kwargs['keywords']: etree.SubElement(e, nspath_eval('dc:subject', nsmap) ).text = keyword for link in kwargs['links']: etree.SubElement(e, nspath_eval('dct:references', nsmap), scheme= link[0]).text = link[1] bbox2 = loads(kwargs['wkt_geometry']).bounds bbox = etree.SubElement(e, nspath_eval('ows:BoundingBox', nsmap), crs= 'http://www.opengis.net/def/crs/EPSG/0/%s' % srs, dimensions='2') etree.SubElement(bbox, nspath_eval('ows:LowerCorner', nsmap) ).text = '%s %s' % (bbox2[1], bbox2[0]) etree.SubElement(bbox, nspath_eval('ows:UpperCorner', nsmap) ).text = '%s %s' % (bbox2[3], bbox2[2]) return etree.tostring(e, pretty_print=True)
create_metadata_record
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(**kwargs):\n if 'srs' in kwargs:\n srs = kwargs['srs']\n else:\n srs = '4326'\n modified = '%sZ' % datetime.datetime.utcnow().isoformat().split('.')[0]\n nsmap = Namespaces().get_namespaces(['csw', 'dc', 'dct', 'ows'])\n e = etree.Element(nspath_eval('csw:Record', nsmap), nsmap=nsmap)\n etree.SubElement(e, nspath_eval('dc:identifier', nsmap)).text = kwargs[\n 'identifier']\n etree.SubElement(e, nspath_eval('dc:title', nsmap)).text = kwargs['title']\n if 'alternative' in kwargs:\n etree.SubElement(e, nspath_eval('dct:alternative', nsmap)\n ).text = kwargs['alternative']\n etree.SubElement(e, nspath_eval('dct:modified', nsmap)).text = modified\n etree.SubElement(e, nspath_eval('dct:abstract', nsmap)).text = kwargs[\n 'abstract']\n etree.SubElement(e, nspath_eval('dc:type', nsmap)).text = kwargs['type']\n etree.SubElement(e, nspath_eval('dc:format', nsmap)).text = kwargs['format'\n ]\n etree.SubElement(e, nspath_eval('dc:source', nsmap)).text = kwargs['source'\n ]\n if 'relation' in kwargs:\n etree.SubElement(e, nspath_eval('dc:relation', nsmap)).text = kwargs[\n 'relation']\n if 'keywords' in kwargs:\n if kwargs['keywords'] is not None:\n for keyword in kwargs['keywords']:\n etree.SubElement(e, nspath_eval('dc:subject', nsmap)\n ).text = keyword\n for link in kwargs['links']:\n etree.SubElement(e, nspath_eval('dct:references', nsmap), scheme=\n link[0]).text = link[1]\n bbox2 = loads(kwargs['wkt_geometry']).bounds\n bbox = etree.SubElement(e, nspath_eval('ows:BoundingBox', nsmap), crs=\n 'http://www.opengis.net/def/crs/EPSG/0/%s' % srs, dimensions='2')\n etree.SubElement(bbox, nspath_eval('ows:LowerCorner', nsmap)\n ).text = '%s %s' % (bbox2[1], bbox2[0])\n etree.SubElement(bbox, nspath_eval('ows:UpperCorner', nsmap)\n ).text = '%s %s' % (bbox2[3], bbox2[2])\n return etree.tostring(e, pretty_print=True)\n", "role": "user" }, { "content": "create_metadata_record", "role": "assistant" } ]
def x(*args): bag = [] for term in args: if term is not None: if isinstance(term, list): for term2 in term: if term2 is not None: bag.append(term2) else: bag.append(term) return ' '.join(bag)
gen_anytext
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*args):\n bag = []\n for term in args:\n if term is not None:\n if isinstance(term, list):\n for term2 in term:\n if term2 is not None:\n bag.append(term2)\n else:\n bag.append(term)\n return ' '.join(bag)\n", "role": "user" }, { "content": "gen_anytext", "role": "assistant" } ]
def x(service): try: wmts = WebMapTileService(service.url) srs, created = SpatialReferenceSystem.objects.get_or_create(code= 'EPSG:4326') service.srs.add(srs) service.update_validity() layer_names = list(wmts.contents) layer_n = 0 total = len(layer_names) for layer_name in layer_names: ows_layer = wmts.contents[layer_name] LOGGER.debug('Updating layer %s' % ows_layer.name) layer, created = Layer.objects.get_or_create(name=ows_layer. name, service=service, catalog=service.catalog) if layer.active: links = [['OGC:WMTS', service.url], ['OGC:WMTS', settings. SITE_URL.rstrip('/') + '/' + layer.get_url_endpoint()]] layer.type = 'OGC:WMTS' layer.title = ows_layer.title layer.abstract = ows_layer.abstract keywords = None if hasattr(ows_layer, 'keywords'): keywords = ows_layer.keywords for keyword in keywords: layer.keywords.add(keyword) layer.url = service.url layer.page_url = layer.get_absolute_url links.append(['WWW:LINK', settings.SITE_URL.rstrip('/') + layer.page_url]) bbox = list(ows_layer.boundingBoxWGS84 or (-179.0, -89.0, 179.0, 89.0)) layer.bbox_x0 = bbox[0] layer.bbox_y0 = bbox[1] layer.bbox_x1 = bbox[2] layer.bbox_y1 = bbox[3] layer.wkt_geometry = bbox2wktpolygon(bbox) layer.xml = create_metadata_record(identifier=str(layer. uuid), source=service.url, links=links, format= 'OGC:WMS', type=layer.csw_type, relation=service. id_string, title=ows_layer.title, alternative=ows_layer .name, abstract=layer.abstract, keywords=keywords, wkt_geometry=layer.wkt_geometry) layer.anytext = gen_anytext(layer.title, layer.abstract, keywords) layer.save() add_mined_dates(layer) layer_n = layer_n + 1 LOGGER.debug('Updating layer n. %s/%s' % (layer_n, total)) if DEBUG_SERVICES and layer_n == DEBUG_LAYER_NUMBER: return except Exception as err: message = 'update_layers_wmts: {0}'.format(err) check = Check(content_object=service, success=False, response_time= 0, message=message) check.save()
update_layers_wmts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(service):\n try:\n wmts = WebMapTileService(service.url)\n srs, created = SpatialReferenceSystem.objects.get_or_create(code=\n 'EPSG:4326')\n service.srs.add(srs)\n service.update_validity()\n layer_names = list(wmts.contents)\n layer_n = 0\n total = len(layer_names)\n for layer_name in layer_names:\n ows_layer = wmts.contents[layer_name]\n LOGGER.debug('Updating layer %s' % ows_layer.name)\n layer, created = Layer.objects.get_or_create(name=ows_layer.\n name, service=service, catalog=service.catalog)\n if layer.active:\n links = [['OGC:WMTS', service.url], ['OGC:WMTS', settings.\n SITE_URL.rstrip('/') + '/' + layer.get_url_endpoint()]]\n layer.type = 'OGC:WMTS'\n layer.title = ows_layer.title\n layer.abstract = ows_layer.abstract\n keywords = None\n if hasattr(ows_layer, 'keywords'):\n keywords = ows_layer.keywords\n for keyword in keywords:\n layer.keywords.add(keyword)\n layer.url = service.url\n layer.page_url = layer.get_absolute_url\n links.append(['WWW:LINK', settings.SITE_URL.rstrip('/') +\n layer.page_url])\n bbox = list(ows_layer.boundingBoxWGS84 or (-179.0, -89.0, \n 179.0, 89.0))\n layer.bbox_x0 = bbox[0]\n layer.bbox_y0 = bbox[1]\n layer.bbox_x1 = bbox[2]\n layer.bbox_y1 = bbox[3]\n layer.wkt_geometry = bbox2wktpolygon(bbox)\n layer.xml = create_metadata_record(identifier=str(layer.\n uuid), source=service.url, links=links, format=\n 'OGC:WMS', type=layer.csw_type, relation=service.\n id_string, title=ows_layer.title, alternative=ows_layer\n .name, abstract=layer.abstract, keywords=keywords,\n wkt_geometry=layer.wkt_geometry)\n layer.anytext = gen_anytext(layer.title, layer.abstract,\n keywords)\n layer.save()\n add_mined_dates(layer)\n layer_n = layer_n + 1\n LOGGER.debug('Updating layer n. %s/%s' % (layer_n, total))\n if DEBUG_SERVICES and layer_n == DEBUG_LAYER_NUMBER:\n return\n except Exception as err:\n message = 'update_layers_wmts: {0}'.format(err)\n check = Check(content_object=service, success=False, response_time=\n 0, message=message)\n check.save()\n", "role": "user" }, { "content": "update_layers_wmts", "role": "assistant" } ]
def x(service): params = {'field': 'title', 'query': '', 'show_warped': '1', 'format': 'json'} headers = {'Content-Type': 'application/json', 'Accept': 'application/json' } request = requests.get(service.url, headers=headers, params=params) try: records = json.loads(request.content) total_pages = int(records['total_pages']) for crs_code in ['EPSG:4326', 'EPSG:900913', 'EPSG:3857']: srs, created = SpatialReferenceSystem.objects.get_or_create(code =crs_code) service.srs.add(srs) service.update_validity() for i in range(1, total_pages + 1): params = {'field': 'title', 'query': '', 'show_warped': '1', 'format': 'json', 'page': i} request = requests.get(service.url, headers=headers, params=params) records = json.loads(request.content) LOGGER.debug('Fetched %s' % request.url) layers = records['items'] layer_n = 0 total = len(layers) for layer in layers: name = layer['id'] title = layer['title'] abstract = layer['description'] bbox = layer['bbox'] dates = [] if 'published_date' in layer: dates.append(layer['published_date']) if 'date_depicted' in layer: dates.append(layer['date_depicted']) if 'depicts_year' in layer: dates.append(layer['depicts_year']) if 'issue_year' in layer: dates.append(layer['issue_year']) layer, created = Layer.objects.get_or_create(name=name, service=service, catalog=service.catalog) if layer.active: layer.type = 'Hypermap:WARPER' layer.title = title layer.abstract = abstract layer.is_public = True layer.url = '%s/wms/%s?' % (service.url, name) layer.page_url = '%s/%s' % (service.url, name) x0 = None y0 = None x1 = None y1 = None if bbox: bbox_list = bbox.split(',') x0 = format_float(bbox_list[0]) y0 = format_float(bbox_list[1]) x1 = format_float(bbox_list[2]) y1 = format_float(bbox_list[3]) layer.bbox_x0 = x0 layer.bbox_y0 = y0 layer.bbox_x1 = x1 layer.bbox_y1 = y1 layer.save() add_mined_dates(layer) add_metadata_dates_to_layer(dates, layer) layer_n = layer_n + 1 LOGGER.debug('Updating layer n. %s/%s' % (layer_n, total)) if DEBUG_SERVICES and layer_n == DEBUG_LAYER_NUMBER: return except Exception as err: message = 'update_layers_warper: {0}. request={1} response={2}'.format( err, service.url, request.text) check = Check(content_object=service, success=False, response_time= 0, message=message) check.save()
update_layers_warper
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(service):\n params = {'field': 'title', 'query': '', 'show_warped': '1', 'format':\n 'json'}\n headers = {'Content-Type': 'application/json', 'Accept': 'application/json'\n }\n request = requests.get(service.url, headers=headers, params=params)\n try:\n records = json.loads(request.content)\n total_pages = int(records['total_pages'])\n for crs_code in ['EPSG:4326', 'EPSG:900913', 'EPSG:3857']:\n srs, created = SpatialReferenceSystem.objects.get_or_create(code\n =crs_code)\n service.srs.add(srs)\n service.update_validity()\n for i in range(1, total_pages + 1):\n params = {'field': 'title', 'query': '', 'show_warped': '1',\n 'format': 'json', 'page': i}\n request = requests.get(service.url, headers=headers, params=params)\n records = json.loads(request.content)\n LOGGER.debug('Fetched %s' % request.url)\n layers = records['items']\n layer_n = 0\n total = len(layers)\n for layer in layers:\n name = layer['id']\n title = layer['title']\n abstract = layer['description']\n bbox = layer['bbox']\n dates = []\n if 'published_date' in layer:\n dates.append(layer['published_date'])\n if 'date_depicted' in layer:\n dates.append(layer['date_depicted'])\n if 'depicts_year' in layer:\n dates.append(layer['depicts_year'])\n if 'issue_year' in layer:\n dates.append(layer['issue_year'])\n layer, created = Layer.objects.get_or_create(name=name,\n service=service, catalog=service.catalog)\n if layer.active:\n layer.type = 'Hypermap:WARPER'\n layer.title = title\n layer.abstract = abstract\n layer.is_public = True\n layer.url = '%s/wms/%s?' % (service.url, name)\n layer.page_url = '%s/%s' % (service.url, name)\n x0 = None\n y0 = None\n x1 = None\n y1 = None\n if bbox:\n bbox_list = bbox.split(',')\n x0 = format_float(bbox_list[0])\n y0 = format_float(bbox_list[1])\n x1 = format_float(bbox_list[2])\n y1 = format_float(bbox_list[3])\n layer.bbox_x0 = x0\n layer.bbox_y0 = y0\n layer.bbox_x1 = x1\n layer.bbox_y1 = y1\n layer.save()\n add_mined_dates(layer)\n add_metadata_dates_to_layer(dates, layer)\n layer_n = layer_n + 1\n LOGGER.debug('Updating layer n. %s/%s' % (layer_n, total))\n if DEBUG_SERVICES and layer_n == DEBUG_LAYER_NUMBER:\n return\n except Exception as err:\n message = 'update_layers_warper: {0}. request={1} response={2}'.format(\n err, service.url, request.text)\n check = Check(content_object=service, success=False, response_time=\n 0, message=message)\n check.save()\n", "role": "user" }, { "content": "update_layers_warper", "role": "assistant" } ]
def x(service, greedy_opt=False): try: esri_service = ArcMapService(service.url) srs_code = esri_service.spatialReference.wkid srs, created = SpatialReferenceSystem.objects.get_or_create(code= srs_code) service.srs.add(srs) service.update_validity() if 'supportedExtensions' in esri_service._json_struct and greedy_opt: if 'WMSServer' in esri_service._json_struct['supportedExtensions']: wms_url = service.url.replace('/rest/services/', '/services/') if '?f=pjson' in wms_url: wms_url = wms_url.replace('?f=pjson', 'WMSServer?') if '?f=json' in wms_url: wms_url = wms_url.replace('?f=json', 'WMSServer?') LOGGER.debug( 'This ESRI REST endpoint has an WMS interface to process: %s' % wms_url) from utils import create_service_from_endpoint create_service_from_endpoint(wms_url, 'OGC:WMS', catalog= service.catalog) layer_n = 0 total = len(esri_service.layers) for esri_layer in esri_service.layers: if 'error' not in esri_layer._json_struct: LOGGER.debug('Updating layer %s' % esri_layer.name) layer, created = Layer.objects.get_or_create(name= esri_layer.id, service=service, catalog=service.catalog) if layer.active: layer.type = 'ESRI:ArcGIS:MapServer' links = [[layer.type, service.url], ['OGC:WMTS', settings.SITE_URL.rstrip('/') + '/' + layer. get_url_endpoint()]] layer.title = esri_layer.name layer.abstract = esri_service.serviceDescription layer.url = service.url layer.page_url = layer.get_absolute_url links.append(['WWW:LINK', settings.SITE_URL.rstrip('/') + layer.page_url]) try: layer.bbox_x0 = esri_layer.extent.xmin layer.bbox_y0 = esri_layer.extent.ymin layer.bbox_x1 = esri_layer.extent.xmax layer.bbox_y1 = esri_layer.extent.ymax except KeyError: pass try: layer.bbox_x0 = esri_layer._json_struct['extent'][ 'xmin'] layer.bbox_y0 = esri_layer._json_struct['extent'][ 'ymin'] layer.bbox_x1 = esri_layer._json_struct['extent'][ 'xmax'] layer.bbox_y1 = esri_layer._json_struct['extent'][ 'ymax'] except Exception: pass layer.wkt_geometry = bbox2wktpolygon([layer.bbox_x0, layer.bbox_y0, layer.bbox_x1, layer.bbox_y1]) layer.xml = create_metadata_record(identifier=str(layer .uuid), source=service.url, links=links, format= 'ESRI:ArcGIS:MapServer', type=layer.csw_type, relation=service.id_string, title=layer.title, alternative=layer.title, abstract=layer.abstract, wkt_geometry=layer.wkt_geometry) layer.anytext = gen_anytext(layer.title, layer.abstract) layer.save() add_mined_dates(layer) layer_n = layer_n + 1 LOGGER.debug('Updating layer n. %s/%s' % (layer_n, total)) if DEBUG_SERVICES and layer_n == DEBUG_LAYER_NUMBER: return except Exception as err: message = 'update_layers_esri_mapserver: {0}'.format(err) check = Check(content_object=service, success=False, response_time= 0, message=message) check.save()
update_layers_esri_mapserver
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(service, greedy_opt=False):\n try:\n esri_service = ArcMapService(service.url)\n srs_code = esri_service.spatialReference.wkid\n srs, created = SpatialReferenceSystem.objects.get_or_create(code=\n srs_code)\n service.srs.add(srs)\n service.update_validity()\n if 'supportedExtensions' in esri_service._json_struct and greedy_opt:\n if 'WMSServer' in esri_service._json_struct['supportedExtensions']:\n wms_url = service.url.replace('/rest/services/', '/services/')\n if '?f=pjson' in wms_url:\n wms_url = wms_url.replace('?f=pjson', 'WMSServer?')\n if '?f=json' in wms_url:\n wms_url = wms_url.replace('?f=json', 'WMSServer?')\n LOGGER.debug(\n 'This ESRI REST endpoint has an WMS interface to process: %s'\n % wms_url)\n from utils import create_service_from_endpoint\n create_service_from_endpoint(wms_url, 'OGC:WMS', catalog=\n service.catalog)\n layer_n = 0\n total = len(esri_service.layers)\n for esri_layer in esri_service.layers:\n if 'error' not in esri_layer._json_struct:\n LOGGER.debug('Updating layer %s' % esri_layer.name)\n layer, created = Layer.objects.get_or_create(name=\n esri_layer.id, service=service, catalog=service.catalog)\n if layer.active:\n layer.type = 'ESRI:ArcGIS:MapServer'\n links = [[layer.type, service.url], ['OGC:WMTS', \n settings.SITE_URL.rstrip('/') + '/' + layer.\n get_url_endpoint()]]\n layer.title = esri_layer.name\n layer.abstract = esri_service.serviceDescription\n layer.url = service.url\n layer.page_url = layer.get_absolute_url\n links.append(['WWW:LINK', settings.SITE_URL.rstrip('/') +\n layer.page_url])\n try:\n layer.bbox_x0 = esri_layer.extent.xmin\n layer.bbox_y0 = esri_layer.extent.ymin\n layer.bbox_x1 = esri_layer.extent.xmax\n layer.bbox_y1 = esri_layer.extent.ymax\n except KeyError:\n pass\n try:\n layer.bbox_x0 = esri_layer._json_struct['extent'][\n 'xmin']\n layer.bbox_y0 = esri_layer._json_struct['extent'][\n 'ymin']\n layer.bbox_x1 = esri_layer._json_struct['extent'][\n 'xmax']\n layer.bbox_y1 = esri_layer._json_struct['extent'][\n 'ymax']\n except Exception:\n pass\n layer.wkt_geometry = bbox2wktpolygon([layer.bbox_x0,\n layer.bbox_y0, layer.bbox_x1, layer.bbox_y1])\n layer.xml = create_metadata_record(identifier=str(layer\n .uuid), source=service.url, links=links, format=\n 'ESRI:ArcGIS:MapServer', type=layer.csw_type,\n relation=service.id_string, title=layer.title,\n alternative=layer.title, abstract=layer.abstract,\n wkt_geometry=layer.wkt_geometry)\n layer.anytext = gen_anytext(layer.title, layer.abstract)\n layer.save()\n add_mined_dates(layer)\n layer_n = layer_n + 1\n LOGGER.debug('Updating layer n. %s/%s' % (layer_n, total))\n if DEBUG_SERVICES and layer_n == DEBUG_LAYER_NUMBER:\n return\n except Exception as err:\n message = 'update_layers_esri_mapserver: {0}'.format(err)\n check = Check(content_object=service, success=False, response_time=\n 0, message=message)\n check.save()\n", "role": "user" }, { "content": "update_layers_esri_mapserver", "role": "assistant" } ]
def x(service): try: esri_service = ArcImageService(service.url) obj = json.loads(esri_service._contents) srs_code = obj['spatialReference']['wkid'] srs, created = SpatialReferenceSystem.objects.get_or_create(code= srs_code) service.srs.add(srs) service.update_validity() layer, created = Layer.objects.get_or_create(name=obj['name'], service=service, catalog=service.catalog) if layer.active: layer.type = 'ESRI:ArcGIS:ImageServer' links = [[layer.type, service.url], ['OGC:WMTS', settings. SITE_URL.rstrip('/') + '/' + layer.get_url_endpoint()]] layer.title = obj['name'] layer.abstract = esri_service.serviceDescription layer.url = service.url layer.bbox_x0 = str(obj['extent']['xmin']) layer.bbox_y0 = str(obj['extent']['ymin']) layer.bbox_x1 = str(obj['extent']['xmax']) layer.bbox_y1 = str(obj['extent']['ymax']) layer.page_url = layer.get_absolute_url links.append(['WWW:LINK', settings.SITE_URL.rstrip('/') + layer .page_url]) layer.wkt_geometry = bbox2wktpolygon([layer.bbox_x0, layer. bbox_y0, layer.bbox_x1, layer.bbox_y1]) layer.xml = create_metadata_record(identifier=str(layer.uuid), source=service.url, links=links, format= 'ESRI:ArcGIS:ImageServer', type=layer.csw_type, relation= service.id_string, title=layer.title, alternative=layer. title, abstract=layer.abstract, wkt_geometry=layer.wkt_geometry ) layer.anytext = gen_anytext(layer.title, layer.abstract) layer.save() add_mined_dates(layer) except Exception as err: message = 'update_layers_esri_imageserver: {0}'.format(err) check = Check(content_object=service, success=False, response_time= 0, message=message) check.save()
update_layers_esri_imageserver
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(service):\n try:\n esri_service = ArcImageService(service.url)\n obj = json.loads(esri_service._contents)\n srs_code = obj['spatialReference']['wkid']\n srs, created = SpatialReferenceSystem.objects.get_or_create(code=\n srs_code)\n service.srs.add(srs)\n service.update_validity()\n layer, created = Layer.objects.get_or_create(name=obj['name'],\n service=service, catalog=service.catalog)\n if layer.active:\n layer.type = 'ESRI:ArcGIS:ImageServer'\n links = [[layer.type, service.url], ['OGC:WMTS', settings.\n SITE_URL.rstrip('/') + '/' + layer.get_url_endpoint()]]\n layer.title = obj['name']\n layer.abstract = esri_service.serviceDescription\n layer.url = service.url\n layer.bbox_x0 = str(obj['extent']['xmin'])\n layer.bbox_y0 = str(obj['extent']['ymin'])\n layer.bbox_x1 = str(obj['extent']['xmax'])\n layer.bbox_y1 = str(obj['extent']['ymax'])\n layer.page_url = layer.get_absolute_url\n links.append(['WWW:LINK', settings.SITE_URL.rstrip('/') + layer\n .page_url])\n layer.wkt_geometry = bbox2wktpolygon([layer.bbox_x0, layer.\n bbox_y0, layer.bbox_x1, layer.bbox_y1])\n layer.xml = create_metadata_record(identifier=str(layer.uuid),\n source=service.url, links=links, format=\n 'ESRI:ArcGIS:ImageServer', type=layer.csw_type, relation=\n service.id_string, title=layer.title, alternative=layer.\n title, abstract=layer.abstract, wkt_geometry=layer.wkt_geometry\n )\n layer.anytext = gen_anytext(layer.title, layer.abstract)\n layer.save()\n add_mined_dates(layer)\n except Exception as err:\n message = 'update_layers_esri_imageserver: {0}'.format(err)\n check = Check(content_object=service, success=False, response_time=\n 0, message=message)\n check.save()\n", "role": "user" }, { "content": "update_layers_esri_imageserver", "role": "assistant" } ]
def x(instance, *args, **kwargs): with open(instance.upload.file.name, mode='rb') as f: lines = f.readlines() for url in lines: if len(url) > 255: LOGGER.debug( 'Skipping this endpoint, as it is more than 255 characters: %s' % url) elif Endpoint.objects.filter(url=url, catalog=instance.catalog).count( ) == 0: endpoint = Endpoint(url=url, endpoint_list=instance) endpoint.catalog = instance.catalog endpoint.save() if not settings.REGISTRY_SKIP_CELERY: update_endpoints.delay(instance.id) else: update_endpoints(instance.id)
endpointlist_post_save
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(instance, *args, **kwargs):\n with open(instance.upload.file.name, mode='rb') as f:\n lines = f.readlines()\n for url in lines:\n if len(url) > 255:\n LOGGER.debug(\n 'Skipping this endpoint, as it is more than 255 characters: %s'\n % url)\n elif Endpoint.objects.filter(url=url, catalog=instance.catalog).count(\n ) == 0:\n endpoint = Endpoint(url=url, endpoint_list=instance)\n endpoint.catalog = instance.catalog\n endpoint.save()\n if not settings.REGISTRY_SKIP_CELERY:\n update_endpoints.delay(instance.id)\n else:\n update_endpoints(instance.id)\n", "role": "user" }, { "content": "endpointlist_post_save", "role": "assistant" } ]
def x(instance, *args, **kwargs): exists = Service.objects.filter(url=instance.url, type=instance.type, catalog=instance.catalog).count() > 0 if instance._state.adding and exists: raise Exception('There is already such a service. url={0} catalog={1}' .format(instance.url, instance.catalog))
service_pre_save
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(instance, *args, **kwargs):\n exists = Service.objects.filter(url=instance.url, type=instance.type,\n catalog=instance.catalog).count() > 0\n if instance._state.adding and exists:\n raise Exception('There is already such a service. url={0} catalog={1}'\n .format(instance.url, instance.catalog))\n", "role": "user" }, { "content": "service_pre_save", "role": "assistant" } ]
def x(instance, *args, **kwargs): if instance.is_monitored and settings.REGISTRY_SKIP_CELERY: check_service(instance.id) elif instance.is_monitored: check_service.delay(instance.id)
service_post_save
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(instance, *args, **kwargs):\n if instance.is_monitored and settings.REGISTRY_SKIP_CELERY:\n check_service(instance.id)\n elif instance.is_monitored:\n check_service.delay(instance.id)\n", "role": "user" }, { "content": "service_post_save", "role": "assistant" } ]
def x(instance, *args, **kwargs): is_valid = True if not instance.service.type == 'Hypermap:WorldMap': if not instance.service.is_valid: is_valid = False LOGGER.debug( 'Layer with id %s is marked invalid because its service is invalid' % instance.id) if (instance.bbox_x0 > -2 and instance.bbox_x1 < 2 and instance. bbox_y0 > -2 and instance.bbox_y1 < 2): is_valid = False LOGGER.debug( 'Layer with id %s is marked invalid because its extent is within (-2, -2, +2, +2)' % instance.id) instance.is_valid = is_valid
layer_pre_save
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(instance, *args, **kwargs):\n is_valid = True\n if not instance.service.type == 'Hypermap:WorldMap':\n if not instance.service.is_valid:\n is_valid = False\n LOGGER.debug(\n 'Layer with id %s is marked invalid because its service is invalid'\n % instance.id)\n if (instance.bbox_x0 > -2 and instance.bbox_x1 < 2 and instance.\n bbox_y0 > -2 and instance.bbox_y1 < 2):\n is_valid = False\n LOGGER.debug(\n 'Layer with id %s is marked invalid because its extent is within (-2, -2, +2, +2)'\n % instance.id)\n instance.is_valid = is_valid\n", "role": "user" }, { "content": "layer_pre_save", "role": "assistant" } ]
def x(instance, *args, **kwargs): if instance.is_monitored and instance.service.is_monitored: if not settings.REGISTRY_SKIP_CELERY: check_layer.delay(instance.id) else: check_layer(instance.id) else: index_layer(instance.id)
layer_post_save
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(instance, *args, **kwargs):\n if instance.is_monitored and instance.service.is_monitored:\n if not settings.REGISTRY_SKIP_CELERY:\n check_layer.delay(instance.id)\n else:\n check_layer(instance.id)\n else:\n index_layer(instance.id)\n", "role": "user" }, { "content": "layer_post_save", "role": "assistant" } ]
def x(instance, *args, **kwargs): LOGGER.debug('Re-adding layer/service to search engine index') if isinstance(instance.content_object, Service): if not settings.REGISTRY_SKIP_CELERY: index_service.delay(instance.content_object.id) else: index_service(instance.content_object.id) elif not settings.REGISTRY_SKIP_CELERY: index_layer.delay(instance.content_object.id) else: index_layer(instance.content_object.id)
issue_post_delete
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(instance, *args, **kwargs):\n LOGGER.debug('Re-adding layer/service to search engine index')\n if isinstance(instance.content_object, Service):\n if not settings.REGISTRY_SKIP_CELERY:\n index_service.delay(instance.content_object.id)\n else:\n index_service(instance.content_object.id)\n elif not settings.REGISTRY_SKIP_CELERY:\n index_layer.delay(instance.content_object.id)\n else:\n index_layer(instance.content_object.id)\n", "role": "user" }, { "content": "issue_post_delete", "role": "assistant" } ]
def x(self): path = self.get_checks_admin_url() content_type = ContentType.objects.get_for_model(self) params = '?content_type__id__exact={0}&q={1}&success__exact=0'.format( content_type.id, self.id) url = path + params return url
get_checks_admin_reliability_warning_url
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n path = self.get_checks_admin_url()\n content_type = ContentType.objects.get_for_model(self)\n params = '?content_type__id__exact={0}&q={1}&success__exact=0'.format(\n content_type.id, self.id)\n url = path + params\n return url\n", "role": "user" }, { "content": "get_checks_admin_reliability_warning_url", "role": "assistant" } ]
def x(self): signals.post_save.disconnect(layer_post_save, sender=Layer) try: LOGGER.debug('Updating layers for service id %s' % self.id) if self.type == 'OGC:WMS': update_layers_wms(self) elif self.type == 'OGC:WMTS': update_layers_wmts(self) elif self.type == 'ESRI:ArcGIS:MapServer': update_layers_esri_mapserver(self) elif self.type == 'ESRI:ArcGIS:ImageServer': update_layers_esri_imageserver(self) elif self.type == 'Hypermap:WorldMapLegacy': update_layers_wm_legacy(self) elif self.type == 'Hypermap:WorldMap': update_layers_geonode_wm(self) elif self.type == 'Hypermap:WARPER': update_layers_warper(self) except: LOGGER.error('Error updating layers for service %s' % self.uuid) signals.post_save.connect(layer_post_save, sender=Layer)
update_layers
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n signals.post_save.disconnect(layer_post_save, sender=Layer)\n try:\n LOGGER.debug('Updating layers for service id %s' % self.id)\n if self.type == 'OGC:WMS':\n update_layers_wms(self)\n elif self.type == 'OGC:WMTS':\n update_layers_wmts(self)\n elif self.type == 'ESRI:ArcGIS:MapServer':\n update_layers_esri_mapserver(self)\n elif self.type == 'ESRI:ArcGIS:ImageServer':\n update_layers_esri_imageserver(self)\n elif self.type == 'Hypermap:WorldMapLegacy':\n update_layers_wm_legacy(self)\n elif self.type == 'Hypermap:WorldMap':\n update_layers_geonode_wm(self)\n elif self.type == 'Hypermap:WARPER':\n update_layers_warper(self)\n except:\n LOGGER.error('Error updating layers for service %s' % self.uuid)\n signals.post_save.connect(layer_post_save, sender=Layer)\n", "role": "user" }, { "content": "update_layers", "role": "assistant" } ]
def x(self): if self.type == 'Hypermap:WorldMap': return signals.post_save.disconnect(service_post_save, sender=Service) try: is_valid = True if self.srs.filter(code__in=SUPPORTED_SRS).count() == 0: LOGGER.debug( 'Service with id %s is marked invalid because in not exposed in SUPPORTED_SRS' % self.id) is_valid = False if self.type == 'OGC:WMTS': LOGGER.debug( 'Service with id %s is marked invalid because it is of type OGC:WMTS' % self.id) is_valid = False if 'noaa' in self.url.lower(): LOGGER.debug( 'Service with id %s is marked invalid because it is from NOAA' % self.id) is_valid = False self.is_valid = is_valid self.save() except: LOGGER.error('Error updating validity of the service!') signals.post_save.connect(service_post_save, sender=Service)
update_validity
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.type == 'Hypermap:WorldMap':\n return\n signals.post_save.disconnect(service_post_save, sender=Service)\n try:\n is_valid = True\n if self.srs.filter(code__in=SUPPORTED_SRS).count() == 0:\n LOGGER.debug(\n 'Service with id %s is marked invalid because in not exposed in SUPPORTED_SRS'\n % self.id)\n is_valid = False\n if self.type == 'OGC:WMTS':\n LOGGER.debug(\n 'Service with id %s is marked invalid because it is of type OGC:WMTS'\n % self.id)\n is_valid = False\n if 'noaa' in self.url.lower():\n LOGGER.debug(\n 'Service with id %s is marked invalid because it is from NOAA'\n % self.id)\n is_valid = False\n self.is_valid = is_valid\n self.save()\n except:\n LOGGER.error('Error updating validity of the service!')\n signals.post_save.connect(service_post_save, sender=Service)\n", "role": "user" }, { "content": "update_validity", "role": "assistant" } ]
def x(self): if self.is_remote: return self.url return reverse('search_api', args=[self.slug])
get_search_url
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.is_remote:\n return self.url\n return reverse('search_api', args=[self.slug])\n", "role": "user" }, { "content": "get_search_url", "role": "assistant" } ]
def x(self): endpoint = self.url if self.type not in ('Hypermap:WorldMap',): endpoint = ( 'registry/%s/layer/%s/map/wmts/1.0.0/WMTSCapabilities.xml' % ( self.catalog.slug, self.id)) return endpoint
get_url_endpoint
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n endpoint = self.url\n if self.type not in ('Hypermap:WorldMap',):\n endpoint = (\n 'registry/%s/layer/%s/map/wmts/1.0.0/WMTSCapabilities.xml' % (\n self.catalog.slug, self.id))\n return endpoint\n", "role": "user" }, { "content": "get_url_endpoint", "role": "assistant" } ]
def x(url=None): if url is None: url = user_input( 'Input the URL of the GitHub repository to use as a `trytravis` repository: ' ) url = url.strip() http_match = _HTTPS_REGEX.match(url) ssh_match = _SSH_REGEX.match(url) if not http_match and not ssh_match: raise RuntimeError( "That URL doesn't look like a valid GitHub URL. We expect something of the form: `https://github.com/[USERNAME]/[REPOSITORY]` or `ssh://git@github.com/[USERNAME]/[REPOSITORY]" ) if http_match: _, name = http_match.groups() else: _, name = ssh_match.groups() if 'trytravis' not in name: raise RuntimeError( "You must have `trytravis` in the name of your repository. This is a security feature to reduce chances of running git push -f on a repository you don't mean to." ) accept = user_input( 'Remember that `trytravis` will make commits on your behalf to `%s`. Are you sure you wish to use this repository? Type `y` or `yes` to accept: ' % url) if accept.lower() not in ['y', 'yes']: raise RuntimeError('Operation aborted by user.') if not os.path.isdir(config_dir): os.makedirs(config_dir) with open(os.path.join(config_dir, 'repo'), 'w+') as f: f.truncate() f.write(url) print('Repository saved successfully.')
input_github_repo
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(url=None):\n if url is None:\n url = user_input(\n 'Input the URL of the GitHub repository to use as a `trytravis` repository: '\n )\n url = url.strip()\n http_match = _HTTPS_REGEX.match(url)\n ssh_match = _SSH_REGEX.match(url)\n if not http_match and not ssh_match:\n raise RuntimeError(\n \"That URL doesn't look like a valid GitHub URL. We expect something of the form: `https://github.com/[USERNAME]/[REPOSITORY]` or `ssh://git@github.com/[USERNAME]/[REPOSITORY]\"\n )\n if http_match:\n _, name = http_match.groups()\n else:\n _, name = ssh_match.groups()\n if 'trytravis' not in name:\n raise RuntimeError(\n \"You must have `trytravis` in the name of your repository. This is a security feature to reduce chances of running git push -f on a repository you don't mean to.\"\n )\n accept = user_input(\n 'Remember that `trytravis` will make commits on your behalf to `%s`. Are you sure you wish to use this repository? Type `y` or `yes` to accept: '\n % url)\n if accept.lower() not in ['y', 'yes']:\n raise RuntimeError('Operation aborted by user.')\n if not os.path.isdir(config_dir):\n os.makedirs(config_dir)\n with open(os.path.join(config_dir, 'repo'), 'w+') as f:\n f.truncate()\n f.write(url)\n print('Repository saved successfully.')\n", "role": "user" }, { "content": "input_github_repo", "role": "assistant" } ]
def x(): if 'TRAVIS' in os.environ: raise RuntimeError( 'Detected that we are running in Travis. Stopping to prevent infinite loops.' ) try: with open(os.path.join(config_dir, 'repo'), 'r') as f: return f.read() except (OSError, IOError): raise RuntimeError( 'Could not find your repository. Have you ran `trytravis --repo`?')
load_github_repo
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n if 'TRAVIS' in os.environ:\n raise RuntimeError(\n 'Detected that we are running in Travis. Stopping to prevent infinite loops.'\n )\n try:\n with open(os.path.join(config_dir, 'repo'), 'r') as f:\n return f.read()\n except (OSError, IOError):\n raise RuntimeError(\n 'Could not find your repository. Have you ran `trytravis --repo`?')\n", "role": "user" }, { "content": "load_github_repo", "role": "assistant" } ]
def x(path, url): try: repo = git.Repo(path) except Exception: raise RuntimeError("Couldn't locate a repository at `%s`." % path) commited = False try: try: repo.delete_remote('trytravis') except Exception: pass print('Adding a temporary remote to `%s`...' % url) remote = repo.create_remote('trytravis', url) print('Adding all local changes...') repo.git.add('--all') try: print('Committing local changes...') timestamp = datetime.datetime.now().isoformat() repo.git.commit(m='trytravis-' + timestamp) commited = True except git.exc.GitCommandError as e: if 'nothing to commit' in str(e): commited = False else: raise commit = repo.head.commit.hexsha committed_at = repo.head.commit.committed_datetime print('Pushing to `trytravis` remote...') remote.push(force=True) finally: if commited: print('Reverting to old state...') repo.git.reset('HEAD^') try: repo.delete_remote('trytravis') except Exception: pass return commit, committed_at
submit_changes_to_github_repo
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path, url):\n try:\n repo = git.Repo(path)\n except Exception:\n raise RuntimeError(\"Couldn't locate a repository at `%s`.\" % path)\n commited = False\n try:\n try:\n repo.delete_remote('trytravis')\n except Exception:\n pass\n print('Adding a temporary remote to `%s`...' % url)\n remote = repo.create_remote('trytravis', url)\n print('Adding all local changes...')\n repo.git.add('--all')\n try:\n print('Committing local changes...')\n timestamp = datetime.datetime.now().isoformat()\n repo.git.commit(m='trytravis-' + timestamp)\n commited = True\n except git.exc.GitCommandError as e:\n if 'nothing to commit' in str(e):\n commited = False\n else:\n raise\n commit = repo.head.commit.hexsha\n committed_at = repo.head.commit.committed_datetime\n print('Pushing to `trytravis` remote...')\n remote.push(force=True)\n finally:\n if commited:\n print('Reverting to old state...')\n repo.git.reset('HEAD^')\n try:\n repo.delete_remote('trytravis')\n except Exception:\n pass\n return commit, committed_at\n", "role": "user" }, { "content": "submit_changes_to_github_repo", "role": "assistant" } ]
def x(url, commit, committed_at): print('Waiting for a Travis build to appear for `%s` after `%s`...' % ( commit, committed_at)) import requests slug = _slug_from_url(url) start_time = time.time() build_id = None while time.time() - start_time < 60: with requests.get('https://api.travis-ci.org/repos/%s/builds' % slug, headers=_travis_headers()) as r: if not r.ok: raise RuntimeError( 'Could not reach the Travis API endpoint. Additional information: %s' % str(r.content)) commit_to_sha = {} json = r.json() for travis_commit in sorted(json['commits'], key=lambda x: x[ 'committed_at']): travis_committed_at = datetime.datetime.strptime(travis_commit ['committed_at'], '%Y-%m-%dT%H:%M:%SZ').replace(tzinfo=utc) if travis_committed_at < committed_at: continue commit_to_sha[travis_commit['id']] = travis_commit['sha'] for build in json['builds']: if build['commit_id'] in commit_to_sha and commit_to_sha[build ['commit_id']] == commit: build_id = build['id'] print('Travis build id: `%d`' % build_id) print( 'Travis build URL: `https://travis-ci.org/%s/builds/%d`' % (slug, build_id)) if build_id is not None: break time.sleep(3.0) else: raise RuntimeError( 'Timed out while waiting for a Travis build to start. Is Travis configured for `%s`?' % url) return build_id
wait_for_travis_build
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(url, commit, committed_at):\n print('Waiting for a Travis build to appear for `%s` after `%s`...' % (\n commit, committed_at))\n import requests\n slug = _slug_from_url(url)\n start_time = time.time()\n build_id = None\n while time.time() - start_time < 60:\n with requests.get('https://api.travis-ci.org/repos/%s/builds' %\n slug, headers=_travis_headers()) as r:\n if not r.ok:\n raise RuntimeError(\n 'Could not reach the Travis API endpoint. Additional information: %s'\n % str(r.content))\n commit_to_sha = {}\n json = r.json()\n for travis_commit in sorted(json['commits'], key=lambda x: x[\n 'committed_at']):\n travis_committed_at = datetime.datetime.strptime(travis_commit\n ['committed_at'], '%Y-%m-%dT%H:%M:%SZ').replace(tzinfo=utc)\n if travis_committed_at < committed_at:\n continue\n commit_to_sha[travis_commit['id']] = travis_commit['sha']\n for build in json['builds']:\n if build['commit_id'] in commit_to_sha and commit_to_sha[build\n ['commit_id']] == commit:\n build_id = build['id']\n print('Travis build id: `%d`' % build_id)\n print(\n 'Travis build URL: `https://travis-ci.org/%s/builds/%d`'\n % (slug, build_id))\n if build_id is not None:\n break\n time.sleep(3.0)\n else:\n raise RuntimeError(\n 'Timed out while waiting for a Travis build to start. Is Travis configured for `%s`?'\n % url)\n return build_id\n", "role": "user" }, { "content": "wait_for_travis_build", "role": "assistant" } ]
def x(build_id): import requests try: build_size = None running = True while running: with requests.get('https://api.travis-ci.org/builds/%d' % build_id, headers=_travis_headers()) as r: json = r.json() if build_size is not None: if build_size > 1: sys.stdout.write('\r\x1b[%dA' % build_size) else: sys.stdout.write('\r') build_size = len(json['jobs']) running = False current_number = 1 for job in json['jobs']: color, state, is_running = _travis_job_state(job['state']) if is_running: running = True platform = job['config']['os'] if platform == 'osx': platform = ' osx ' env = job['config'].get('env', '') sudo = 's' if job['config'].get('sudo', True) else 'c' lang = job['config'].get('language', 'generic') padding = ' ' * (len(str(build_size)) - len(str( current_number))) number = str(current_number) + padding current_number += 1 job_display = '#' + ' '.join([number, state, platform, sudo, lang, env]) print(color + job_display + colorama.Style.RESET_ALL) time.sleep(3.0) except KeyboardInterrupt: pass
watch_travis_build
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(build_id):\n import requests\n try:\n build_size = None\n running = True\n while running:\n with requests.get('https://api.travis-ci.org/builds/%d' %\n build_id, headers=_travis_headers()) as r:\n json = r.json()\n if build_size is not None:\n if build_size > 1:\n sys.stdout.write('\\r\\x1b[%dA' % build_size)\n else:\n sys.stdout.write('\\r')\n build_size = len(json['jobs'])\n running = False\n current_number = 1\n for job in json['jobs']:\n color, state, is_running = _travis_job_state(job['state'])\n if is_running:\n running = True\n platform = job['config']['os']\n if platform == 'osx':\n platform = ' osx '\n env = job['config'].get('env', '')\n sudo = 's' if job['config'].get('sudo', True) else 'c'\n lang = job['config'].get('language', 'generic')\n padding = ' ' * (len(str(build_size)) - len(str(\n current_number)))\n number = str(current_number) + padding\n current_number += 1\n job_display = '#' + ' '.join([number, state, platform,\n sudo, lang, env])\n print(color + job_display + colorama.Style.RESET_ALL)\n time.sleep(3.0)\n except KeyboardInterrupt:\n pass\n", "role": "user" }, { "content": "watch_travis_build", "role": "assistant" } ]
def x(state): if state in [None, 'queued', 'created', 'received']: return colorama.Fore.YELLOW, '*', True elif state in ['started', 'running']: return colorama.Fore.LIGHTYELLOW_EX, '*', True elif state == 'passed': return colorama.Fore.LIGHTGREEN_EX, 'P', False elif state == 'failed': return colorama.Fore.LIGHTRED_EX, 'X', False elif state == 'errored': return colorama.Fore.LIGHTRED_EX, '!', False elif state == 'canceled': return colorama.Fore.LIGHTBLACK_EX, 'X', False else: raise RuntimeError('unknown state: %s' % str(state))
travis_job_state
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(state):\n if state in [None, 'queued', 'created', 'received']:\n return colorama.Fore.YELLOW, '*', True\n elif state in ['started', 'running']:\n return colorama.Fore.LIGHTYELLOW_EX, '*', True\n elif state == 'passed':\n return colorama.Fore.LIGHTGREEN_EX, 'P', False\n elif state == 'failed':\n return colorama.Fore.LIGHTRED_EX, 'X', False\n elif state == 'errored':\n return colorama.Fore.LIGHTRED_EX, '!', False\n elif state == 'canceled':\n return colorama.Fore.LIGHTBLACK_EX, 'X', False\n else:\n raise RuntimeError('unknown state: %s' % str(state))\n", "role": "user" }, { "content": "travis_job_state", "role": "assistant" } ]
def x(url): http_match = _HTTPS_REGEX.match(url) ssh_match = _SSH_REGEX.match(url) if not http_match and not ssh_match: raise RuntimeError( 'Could not parse the URL (`%s`) for your repository.' % url) if http_match: return '/'.join(http_match.groups()) else: return '/'.join(ssh_match.groups())
slug_from_url
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(url):\n http_match = _HTTPS_REGEX.match(url)\n ssh_match = _SSH_REGEX.match(url)\n if not http_match and not ssh_match:\n raise RuntimeError(\n 'Could not parse the URL (`%s`) for your repository.' % url)\n if http_match:\n return '/'.join(http_match.groups())\n else:\n return '/'.join(ssh_match.groups())\n", "role": "user" }, { "content": "slug_from_url", "role": "assistant" } ]
def x(): platform_system = platform.system() if platform_system == 'Linux': os_name, os_version, _ = platform.dist() else: os_name = platform_system os_version = platform.version() python_version = platform.python_version() return 'trytravis %s (%s %s, python %s)' % (__version__, os_name.lower( ), os_version, python_version)
version_string
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n platform_system = platform.system()\n if platform_system == 'Linux':\n os_name, os_version, _ = platform.dist()\n else:\n os_name = platform_system\n os_version = platform.version()\n python_version = platform.python_version()\n return 'trytravis %s (%s %s, python %s)' % (__version__, os_name.lower(\n ), os_version, python_version)\n", "role": "user" }, { "content": "version_string", "role": "assistant" } ]
def x(argv): repo_input_argv = len(argv) == 2 and argv[0] in ['--repo', '-r', '-R'] if len(argv) > 1 and not repo_input_argv: _main(['--help']) if len(argv) == 1 or repo_input_argv: arg = argv[0] if arg in ['-h', '--help', '-H']: print(_USAGE) elif arg in ['-v', '--version', '-V']: print(_version_string()) elif arg in ['-r', '--repo', '-R']: if len(argv) == 2: url = argv[1] else: url = None _input_github_repo(url) elif arg in ['--no-wait', '-nw']: url = _load_github_repo() commit, committed = _submit_changes_to_github_repo(os.getcwd(), url ) build_id = _wait_for_travis_build(url, commit, committed) else: _main(['--help']) elif len(argv) == 0: url = _load_github_repo() commit, committed = _submit_changes_to_github_repo(os.getcwd(), url) build_id = _wait_for_travis_build(url, commit, committed) _watch_travis_build(build_id)
main
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(argv):\n repo_input_argv = len(argv) == 2 and argv[0] in ['--repo', '-r', '-R']\n if len(argv) > 1 and not repo_input_argv:\n _main(['--help'])\n if len(argv) == 1 or repo_input_argv:\n arg = argv[0]\n if arg in ['-h', '--help', '-H']:\n print(_USAGE)\n elif arg in ['-v', '--version', '-V']:\n print(_version_string())\n elif arg in ['-r', '--repo', '-R']:\n if len(argv) == 2:\n url = argv[1]\n else:\n url = None\n _input_github_repo(url)\n elif arg in ['--no-wait', '-nw']:\n url = _load_github_repo()\n commit, committed = _submit_changes_to_github_repo(os.getcwd(), url\n )\n build_id = _wait_for_travis_build(url, commit, committed)\n else:\n _main(['--help'])\n elif len(argv) == 0:\n url = _load_github_repo()\n commit, committed = _submit_changes_to_github_repo(os.getcwd(), url)\n build_id = _wait_for_travis_build(url, commit, committed)\n _watch_travis_build(build_id)\n", "role": "user" }, { "content": "main", "role": "assistant" } ]
def x(argv=None): try: colorama.init() if argv is None: argv = sys.argv[1:] _main(argv) except RuntimeError as e: print(colorama.Fore.RED + 'ERROR: ' + str(e) + colorama.Style.RESET_ALL ) sys.exit(1) else: sys.exit(0)
main
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(argv=None):\n try:\n colorama.init()\n if argv is None:\n argv = sys.argv[1:]\n _main(argv)\n except RuntimeError as e:\n print(colorama.Fore.RED + 'ERROR: ' + str(e) + colorama.Style.RESET_ALL\n )\n sys.exit(1)\n else:\n sys.exit(0)\n", "role": "user" }, { "content": "main", "role": "assistant" } ]
def x(request, url=None, catalog_id=None): if request.user.is_authenticated(): settings.REGISTRY_PYCSW['manager']['transactions'] = 'true' env = request.META.copy() if request.method == 'POST': from StringIO import StringIO env['wsgi.input'] = StringIO(request.body) env.update({'local.app_root': os.path.dirname(__file__), 'REQUEST_URI': request.build_absolute_uri()}) if url is not None: settings.REGISTRY_PYCSW['server']['url'] = url if catalog_id is not None: settings.REGISTRY_PYCSW['repository']['filter' ] = 'catalog_id = %d' % catalog_id csw = server.Csw(settings.REGISTRY_PYCSW, env) content = csw.dispatch_wsgi() if isinstance(content, list): content = content[1] response = HttpResponse(content, content_type=csw.contenttype) response['Access-Control-Allow-Origin'] = '*' return response
csw_global_dispatch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(request, url=None, catalog_id=None):\n if request.user.is_authenticated():\n settings.REGISTRY_PYCSW['manager']['transactions'] = 'true'\n env = request.META.copy()\n if request.method == 'POST':\n from StringIO import StringIO\n env['wsgi.input'] = StringIO(request.body)\n env.update({'local.app_root': os.path.dirname(__file__), 'REQUEST_URI':\n request.build_absolute_uri()})\n if url is not None:\n settings.REGISTRY_PYCSW['server']['url'] = url\n if catalog_id is not None:\n settings.REGISTRY_PYCSW['repository']['filter'\n ] = 'catalog_id = %d' % catalog_id\n csw = server.Csw(settings.REGISTRY_PYCSW, env)\n content = csw.dispatch_wsgi()\n if isinstance(content, list):\n content = content[1]\n response = HttpResponse(content, content_type=csw.contenttype)\n response['Access-Control-Allow-Origin'] = '*'\n return response\n", "role": "user" }, { "content": "csw_global_dispatch", "role": "assistant" } ]
def x(request, catalog_slug): catalog = get_object_or_404(Catalog, slug=catalog_slug) if catalog: url = settings.SITE_URL.rstrip('/') + request.path.rstrip('/') return csw_global_dispatch(request, url=url, catalog_id=catalog.id)
csw_global_dispatch_by_catalog
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(request, catalog_slug):\n catalog = get_object_or_404(Catalog, slug=catalog_slug)\n if catalog:\n url = settings.SITE_URL.rstrip('/') + request.path.rstrip('/')\n return csw_global_dispatch(request, url=url, catalog_id=catalog.id)\n", "role": "user" }, { "content": "csw_global_dispatch_by_catalog", "role": "assistant" } ]
def x(request): ctx = {'shortname': settings.REGISTRY_PYCSW['metadata:main'][ 'identification_title'], 'description': settings.REGISTRY_PYCSW[ 'metadata:main']['identification_abstract'], 'developer': settings. REGISTRY_PYCSW['metadata:main']['contact_name'], 'contact': settings.REGISTRY_PYCSW['metadata:main']['contact_email'], 'attribution': settings.REGISTRY_PYCSW['metadata:main'][ 'provider_name'], 'tags': settings.REGISTRY_PYCSW['metadata:main'][ 'identification_keywords'].replace(',', ' '), 'url': settings. SITE_URL.rstrip('/')} return render_to_response('search/opensearch_description.xml', ctx, content_type='application/opensearchdescription+xml')
opensearch_dispatch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(request):\n ctx = {'shortname': settings.REGISTRY_PYCSW['metadata:main'][\n 'identification_title'], 'description': settings.REGISTRY_PYCSW[\n 'metadata:main']['identification_abstract'], 'developer': settings.\n REGISTRY_PYCSW['metadata:main']['contact_name'], 'contact':\n settings.REGISTRY_PYCSW['metadata:main']['contact_email'],\n 'attribution': settings.REGISTRY_PYCSW['metadata:main'][\n 'provider_name'], 'tags': settings.REGISTRY_PYCSW['metadata:main'][\n 'identification_keywords'].replace(',', ' '), 'url': settings.\n SITE_URL.rstrip('/')}\n return render_to_response('search/opensearch_description.xml', ctx,\n content_type='application/opensearchdescription+xml')\n", "role": "user" }, { "content": "opensearch_dispatch", "role": "assistant" } ]
def x(coords): if len(coords) != 4: return False for coord in coords[0:3]: try: num = float(coord) if math.isnan(num): return False if math.isinf(num): return False except ValueError: return False return True
good_coords
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(coords):\n if len(coords) != 4:\n return False\n for coord in coords[0:3]:\n try:\n num = float(coord)\n if math.isnan(num):\n return False\n if math.isinf(num):\n return False\n except ValueError:\n return False\n return True\n", "role": "user" }, { "content": "good_coords", "role": "assistant" } ]
def x(): ESHypermap.es.indices.delete(ESHypermap.index_name, ignore=[400, 404]) LOGGER.debug('Elasticsearch: Index cleared')
clear_es
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n ESHypermap.es.indices.delete(ESHypermap.index_name, ignore=[400, 404])\n LOGGER.debug('Elasticsearch: Index cleared')\n", "role": "user" }, { "content": "clear_es", "role": "assistant" } ]
def x(catalog_slug): mapping = {'mappings': {'layer': {'properties': {'layer_geoshape': { 'type': 'geo_shape', 'tree': 'quadtree', 'precision': REGISTRY_MAPPING_PRECISION}}}}} ESHypermap.es.indices.create(catalog_slug, ignore=[400, 404], body=mapping)
create_indices
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(catalog_slug):\n mapping = {'mappings': {'layer': {'properties': {'layer_geoshape': {\n 'type': 'geo_shape', 'tree': 'quadtree', 'precision':\n REGISTRY_MAPPING_PRECISION}}}}}\n ESHypermap.es.indices.create(catalog_slug, ignore=[400, 404], body=mapping)\n", "role": "user" }, { "content": "create_indices", "role": "assistant" } ]
def x(procname, scriptname): import signal import subprocess p = subprocess.Popen(['ps', 'aux'], stdout=subprocess.PIPE) out, err = p.communicate() for line in out.decode().splitlines(): if procname in line and scriptname in line: pid = int(line.split()[1]) info('Stopping %s %s %d' % (procname, scriptname, pid)) os.kill(pid, signal.SIGKILL)
kill_process
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(procname, scriptname):\n import signal\n import subprocess\n p = subprocess.Popen(['ps', 'aux'], stdout=subprocess.PIPE)\n out, err = p.communicate()\n for line in out.decode().splitlines():\n if procname in line and scriptname in line:\n pid = int(line.split()[1])\n info('Stopping %s %s %d' % (procname, scriptname, pid))\n os.kill(pid, signal.SIGKILL)\n", "role": "user" }, { "content": "kill_process", "role": "assistant" } ]
def x(): services_list = ('Harvard WorldMap', 'Harvard WorldMap open source web geospatial platform', 'Hypermap:WorldMap', 'http://worldmap.harvard.edu'), ('NYPL MapWarper', 'The New York Public Library (NYPL) MapWarper web site', 'Hypermap:WARPER', 'http://maps.nypl.org/warper/maps'), ('Map Warper', 'The MapWarper web site developed, hosted and maintained by Tim Waters' , 'Hypermap:WARPER', 'http://mapwarper.net/maps'), ('WorldMap Warp', 'The MapWarper instance part of the Harvard WorldMap project', 'Hypermap:WARPER', 'http://warp.worldmap.harvard.edu/maps'), ( 'WFP GeoNode', 'World Food Programme GeoNode', 'OGC:WMS', 'http://geonode.wfp.org/geoserver/ows?'), ('NASA EARTHDATA', 'NASA EARTHDATA, powered by EOSDIS', 'OGC:WMTS', 'http://map1.vis.earthdata.nasa.gov/wmts-geo/1.0.0/WMTSCapabilities.xml' ) esri_endpoint = 'https://gis.ngdc.noaa.gov/arcgis/rest/services' LOGGER.debug('*** Importing esri endpoint: %s' % esri_endpoint) create_services_from_endpoint(esri_endpoint) for service in services_list: LOGGER.debug('*** Importing %s' % service[0]) service = Service(title=service[0], abstract=service[1], type= service[2], url=service[3]) service.save()
populate_initial_services
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n services_list = ('Harvard WorldMap',\n 'Harvard WorldMap open source web geospatial platform',\n 'Hypermap:WorldMap', 'http://worldmap.harvard.edu'), ('NYPL MapWarper',\n 'The New York Public Library (NYPL) MapWarper web site',\n 'Hypermap:WARPER', 'http://maps.nypl.org/warper/maps'), ('Map Warper',\n 'The MapWarper web site developed, hosted and maintained by Tim Waters'\n , 'Hypermap:WARPER', 'http://mapwarper.net/maps'), ('WorldMap Warp',\n 'The MapWarper instance part of the Harvard WorldMap project',\n 'Hypermap:WARPER', 'http://warp.worldmap.harvard.edu/maps'), (\n 'WFP GeoNode', 'World Food Programme GeoNode', 'OGC:WMS',\n 'http://geonode.wfp.org/geoserver/ows?'), ('NASA EARTHDATA',\n 'NASA EARTHDATA, powered by EOSDIS', 'OGC:WMTS',\n 'http://map1.vis.earthdata.nasa.gov/wmts-geo/1.0.0/WMTSCapabilities.xml'\n )\n esri_endpoint = 'https://gis.ngdc.noaa.gov/arcgis/rest/services'\n LOGGER.debug('*** Importing esri endpoint: %s' % esri_endpoint)\n create_services_from_endpoint(esri_endpoint)\n for service in services_list:\n LOGGER.debug('*** Importing %s' % service[0])\n service = Service(title=service[0], abstract=service[1], type=\n service[2], url=service[3])\n service.save()\n", "role": "user" }, { "content": "populate_initial_services", "role": "assistant" } ]
def x(request): get = request.GET.copy() new_get = request.GET.copy() for key in get.iterkeys(): if key.count('.') > 0: new_key = key.replace('.', '_') new_get[new_key] = get.get(key) del new_get[key] return new_get
parse_get_params
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(request):\n get = request.GET.copy()\n new_get = request.GET.copy()\n for key in get.iterkeys():\n if key.count('.') > 0:\n new_key = key.replace('.', '_')\n new_get[new_key] = get.get(key)\n del new_get[key]\n return new_get\n", "role": "user" }, { "content": "parse_get_params", "role": "assistant" } ]
def x(): tcp_adapter = TcpAdapter('192.168.1.3', name='HASS', activate_source=False) hdmi_network = HDMINetwork(tcp_adapter) hdmi_network.start() while True: for d in hdmi_network.devices: _LOGGER.info('Device: %s', d) time.sleep(7)
main
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n tcp_adapter = TcpAdapter('192.168.1.3', name='HASS', activate_source=False)\n hdmi_network = HDMINetwork(tcp_adapter)\n hdmi_network.start()\n while True:\n for d in hdmi_network.devices:\n _LOGGER.info('Device: %s', d)\n time.sleep(7)\n", "role": "user" }, { "content": "main", "role": "assistant" } ]
def x(digest1, digest2): digest1 = tuple([int(digest1[i:i + 2], 16) for i in range(0, 63, 2)]) digest2 = tuple([int(digest2[i:i + 2], 16) for i in range(0, 63, 2)]) bits = 0 for i in range(32): bits += POPC[255 & digest1[i] ^ digest2[i]] return 128 - bits
compare_hexdigests
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(digest1, digest2):\n digest1 = tuple([int(digest1[i:i + 2], 16) for i in range(0, 63, 2)])\n digest2 = tuple([int(digest2[i:i + 2], 16) for i in range(0, 63, 2)])\n bits = 0\n for i in range(32):\n bits += POPC[255 & digest1[i] ^ digest2[i]]\n return 128 - bits\n", "role": "user" }, { "content": "compare_hexdigests", "role": "assistant" } ]
def x(self, data): for character in data: if PY3: ch = character else: ch = ord(character) self.count += 1 if self.lastch[1] > -1: self.acc[self.tran3(ch, self.lastch[0], self.lastch[1], 0)] += 1 if self.lastch[2] > -1: self.acc[self.tran3(ch, self.lastch[0], self.lastch[2], 1)] += 1 self.acc[self.tran3(ch, self.lastch[1], self.lastch[2], 2)] += 1 if self.lastch[3] > -1: self.acc[self.tran3(ch, self.lastch[0], self.lastch[3], 3)] += 1 self.acc[self.tran3(ch, self.lastch[1], self.lastch[3], 4)] += 1 self.acc[self.tran3(ch, self.lastch[2], self.lastch[3], 5)] += 1 self.acc[self.tran3(self.lastch[3], self.lastch[0], ch, 6)] += 1 self.acc[self.tran3(self.lastch[3], self.lastch[2], ch, 7)] += 1 self.lastch = [ch] + self.lastch[:3]
update
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data):\n for character in data:\n if PY3:\n ch = character\n else:\n ch = ord(character)\n self.count += 1\n if self.lastch[1] > -1:\n self.acc[self.tran3(ch, self.lastch[0], self.lastch[1], 0)] += 1\n if self.lastch[2] > -1:\n self.acc[self.tran3(ch, self.lastch[0], self.lastch[2], 1)] += 1\n self.acc[self.tran3(ch, self.lastch[1], self.lastch[2], 2)] += 1\n if self.lastch[3] > -1:\n self.acc[self.tran3(ch, self.lastch[0], self.lastch[3], 3)] += 1\n self.acc[self.tran3(ch, self.lastch[1], self.lastch[3], 4)] += 1\n self.acc[self.tran3(ch, self.lastch[2], self.lastch[3], 5)] += 1\n self.acc[self.tran3(self.lastch[3], self.lastch[0], ch, 6)] += 1\n self.acc[self.tran3(self.lastch[3], self.lastch[2], ch, 7)] += 1\n self.lastch = [ch] + self.lastch[:3]\n", "role": "user" }, { "content": "update", "role": "assistant" } ]
def x(self): total = 0 if self.count == 3: total = 1 elif self.count == 4: total = 4 elif self.count > 4: total = 8 * self.count - 28 threshold = total / 256 code = [0] * 32 for i in range(256): if self.acc[i] > threshold: code[i >> 3] += 1 << (i & 7) return code[::-1]
digest
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n total = 0\n if self.count == 3:\n total = 1\n elif self.count == 4:\n total = 4\n elif self.count > 4:\n total = 8 * self.count - 28\n threshold = total / 256\n code = [0] * 32\n for i in range(256):\n if self.acc[i] > threshold:\n code[i >> 3] += 1 << (i & 7)\n return code[::-1]\n", "role": "user" }, { "content": "digest", "role": "assistant" } ]
def x(self, filename): f = open(filename, 'rb') while True: data = f.read(10480) if not data: break self.update(data) f.close()
from_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, filename):\n f = open(filename, 'rb')\n while True:\n data = f.read(10480)\n if not data:\n break\n self.update(data)\n f.close()\n", "role": "user" }, { "content": "from_file", "role": "assistant" } ]
def x(self, otherdigest, ishex=False): bits = 0 myd = self.digest() if ishex: otherdigest = tuple([int(otherdigest[i:i + 2], 16) for i in range(0, 63, 2)]) for i in range(32): bits += POPC[255 & myd[i] ^ otherdigest[i]] return 128 - bits
compare
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, otherdigest, ishex=False):\n bits = 0\n myd = self.digest()\n if ishex:\n otherdigest = tuple([int(otherdigest[i:i + 2], 16) for i in range(0,\n 63, 2)])\n for i in range(32):\n bits += POPC[255 & myd[i] ^ otherdigest[i]]\n return 128 - bits\n", "role": "user" }, { "content": "compare", "role": "assistant" } ]
def x(api_response): try: output = json.dumps(api_response.cgx_content, indent=4) except (TypeError, ValueError, AttributeError): try: output = json.dumps(api_response, indent=4) except (TypeError, ValueError, AttributeError): output = api_response return output
jdout
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(api_response):\n try:\n output = json.dumps(api_response.cgx_content, indent=4)\n except (TypeError, ValueError, AttributeError):\n try:\n output = json.dumps(api_response, indent=4)\n except (TypeError, ValueError, AttributeError):\n output = api_response\n return output\n", "role": "user" }, { "content": "jdout", "role": "assistant" } ]
def x(api_response, sensitive=False): try: output = 'REQUEST: {0} {1}\n'.format(api_response.request.method, api_response.request.path_url) output += 'REQUEST HEADERS:\n' for key, value in api_response.request.headers.items(): if key.lower() in ['cookie'] and not sensitive: cookie_list = value.split('; ') muted_cookie_list = [] for cookie in cookie_list: if cookie.lower().strip().startswith('auth_token='): newcookie = cookie.strip()[:11 ] + '"<SENSITIVE - NOT SHOWN BY DEFAULT>"' muted_cookie_list.append(newcookie) else: muted_cookie_list.append(cookie) muted_value = '; '.join(muted_cookie_list) output += '\t{0}: {1}\n'.format(key, muted_value) elif key.lower() in ['x-auth-token'] and not sensitive: output += '\t{0}: {1}\n'.format(key, '<SENSITIVE - NOT SHOWN BY DEFAULT>') else: output += '\t{0}: {1}\n'.format(key, value) if not api_response.request.body: output += 'REQUEST BODY:\n{0}\n\n'.format({}) else: try: output += 'REQUEST BODY:\n{0}\n\n'.format(json.dumps(json. loads(api_response.request.body), indent=4)) except (TypeError, ValueError, AttributeError): output += 'REQUEST BODY:\n{0}\n\n'.format(jdout( api_response.request.body)) output += 'RESPONSE: {0} {1}\n'.format(api_response.status_code, api_response.reason) output += 'RESPONSE HEADERS:\n' for key, value in api_response.headers.items(): output += '\t{0}: {1}\n'.format(key, value) try: output += 'RESPONSE DATA:\n{0}'.format(json.dumps(api_response. cgx_content, indent=4)) except (TypeError, ValueError, AttributeError): output += 'RESPONSE DATA:\n{0}'.format(json.dumps(json.loads( api_response.content), indent=4)) except (TypeError, ValueError, AttributeError, UnicodeDecodeError): try: output = json.dumps(api_response, indent=4) except (TypeError, ValueError, AttributeError): output = api_response return output
jdout_detailed
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(api_response, sensitive=False):\n try:\n output = 'REQUEST: {0} {1}\\n'.format(api_response.request.method,\n api_response.request.path_url)\n output += 'REQUEST HEADERS:\\n'\n for key, value in api_response.request.headers.items():\n if key.lower() in ['cookie'] and not sensitive:\n cookie_list = value.split('; ')\n muted_cookie_list = []\n for cookie in cookie_list:\n if cookie.lower().strip().startswith('auth_token='):\n newcookie = cookie.strip()[:11\n ] + '\"<SENSITIVE - NOT SHOWN BY DEFAULT>\"'\n muted_cookie_list.append(newcookie)\n else:\n muted_cookie_list.append(cookie)\n muted_value = '; '.join(muted_cookie_list)\n output += '\\t{0}: {1}\\n'.format(key, muted_value)\n elif key.lower() in ['x-auth-token'] and not sensitive:\n output += '\\t{0}: {1}\\n'.format(key,\n '<SENSITIVE - NOT SHOWN BY DEFAULT>')\n else:\n output += '\\t{0}: {1}\\n'.format(key, value)\n if not api_response.request.body:\n output += 'REQUEST BODY:\\n{0}\\n\\n'.format({})\n else:\n try:\n output += 'REQUEST BODY:\\n{0}\\n\\n'.format(json.dumps(json.\n loads(api_response.request.body), indent=4))\n except (TypeError, ValueError, AttributeError):\n output += 'REQUEST BODY:\\n{0}\\n\\n'.format(jdout(\n api_response.request.body))\n output += 'RESPONSE: {0} {1}\\n'.format(api_response.status_code,\n api_response.reason)\n output += 'RESPONSE HEADERS:\\n'\n for key, value in api_response.headers.items():\n output += '\\t{0}: {1}\\n'.format(key, value)\n try:\n output += 'RESPONSE DATA:\\n{0}'.format(json.dumps(api_response.\n cgx_content, indent=4))\n except (TypeError, ValueError, AttributeError):\n output += 'RESPONSE DATA:\\n{0}'.format(json.dumps(json.loads(\n api_response.content), indent=4))\n except (TypeError, ValueError, AttributeError, UnicodeDecodeError):\n try:\n output = json.dumps(api_response, indent=4)\n except (TypeError, ValueError, AttributeError):\n output = api_response\n return output\n", "role": "user" }, { "content": "jdout_detailed", "role": "assistant" } ]
def x(self): try: recommend_update = False update_check_resp = requests.get(self.update_info_url, timeout=3) web_version = update_check_resp.json()['info']['version'] api_logger.debug('RETRIEVED_VERSION: %s', web_version) available_version = SDK_BUILD_REGEX.search(web_version).groupdict() current_version = SDK_BUILD_REGEX.search(self.version).groupdict() available_major = available_version.get('major') available_minor = available_version.get('minor') available_patch = available_version.get('patch') available_build = available_version.get('build') current_major = current_version.get('major') current_minor = current_version.get('minor') current_patch = current_version.get('patch') current_build = current_version.get('build') api_logger.debug('AVAILABLE_VERSION: %s', available_version) api_logger.debug('CURRENT_VERSION: %s', current_version) if available_major > current_major: recommend_update = True elif available_major >= current_major and available_minor > current_minor: recommend_update = True elif available_major >= current_major and available_minor >= current_minor and available_patch > current_patch: recommend_update = True api_logger.debug('NEED_UPDATE: %s', recommend_update) if recommend_update: sys.stderr.write( """WARNING: CloudGenix Python SDK upgrade available. SDKs are typically deprecated 6 months after release of a new version. Latest Version: {0} Current Version: {1} For more info, see 'https://github.com/cloudgenix/sdk-python'. Additionally, this message can be suppressed by instantiating the API with API(update_check=False). """ .format(web_version, self.version)) return except Exception: return
notify_for_new_version
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n try:\n recommend_update = False\n update_check_resp = requests.get(self.update_info_url, timeout=3)\n web_version = update_check_resp.json()['info']['version']\n api_logger.debug('RETRIEVED_VERSION: %s', web_version)\n available_version = SDK_BUILD_REGEX.search(web_version).groupdict()\n current_version = SDK_BUILD_REGEX.search(self.version).groupdict()\n available_major = available_version.get('major')\n available_minor = available_version.get('minor')\n available_patch = available_version.get('patch')\n available_build = available_version.get('build')\n current_major = current_version.get('major')\n current_minor = current_version.get('minor')\n current_patch = current_version.get('patch')\n current_build = current_version.get('build')\n api_logger.debug('AVAILABLE_VERSION: %s', available_version)\n api_logger.debug('CURRENT_VERSION: %s', current_version)\n if available_major > current_major:\n recommend_update = True\n elif available_major >= current_major and available_minor > current_minor:\n recommend_update = True\n elif available_major >= current_major and available_minor >= current_minor and available_patch > current_patch:\n recommend_update = True\n api_logger.debug('NEED_UPDATE: %s', recommend_update)\n if recommend_update:\n sys.stderr.write(\n \"\"\"WARNING: CloudGenix Python SDK upgrade available. SDKs are typically deprecated 6 months after release of a new version.\n\tLatest Version: {0}\n\tCurrent Version: {1}\n\tFor more info, see 'https://github.com/cloudgenix/sdk-python'. Additionally, this message can be suppressed by instantiating the API with API(update_check=False).\n\n\"\"\"\n .format(web_version, self.version))\n return\n except Exception:\n return\n", "role": "user" }, { "content": "notify_for_new_version", "role": "assistant" } ]
def x(self, ssl_verify): self.verify = ssl_verify if isinstance(self.verify, bool): if self.verify: if os.name == 'nt': self._ca_verify_file_handle = temp_ca_bundle(delete=False) self._ca_verify_file_handle.write(BYTE_CA_BUNDLE) self._ca_verify_file_handle.flush() self.ca_verify_filename = self._ca_verify_file_handle.name self._ca_verify_file_handle.close() else: self._ca_verify_file_handle = temp_ca_bundle() self._ca_verify_file_handle.write(BYTE_CA_BUNDLE) self._ca_verify_file_handle.flush() self.ca_verify_filename = self._ca_verify_file_handle.name atexit.register(self._cleanup_ca_temp_file) else: urllib3.disable_warnings() self.ca_verify_filename = False else: self.ca_verify_filename = self.verify return
ssl_verify
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ssl_verify):\n self.verify = ssl_verify\n if isinstance(self.verify, bool):\n if self.verify:\n if os.name == 'nt':\n self._ca_verify_file_handle = temp_ca_bundle(delete=False)\n self._ca_verify_file_handle.write(BYTE_CA_BUNDLE)\n self._ca_verify_file_handle.flush()\n self.ca_verify_filename = self._ca_verify_file_handle.name\n self._ca_verify_file_handle.close()\n else:\n self._ca_verify_file_handle = temp_ca_bundle()\n self._ca_verify_file_handle.write(BYTE_CA_BUNDLE)\n self._ca_verify_file_handle.flush()\n self.ca_verify_filename = self._ca_verify_file_handle.name\n atexit.register(self._cleanup_ca_temp_file)\n else:\n urllib3.disable_warnings()\n self.ca_verify_filename = False\n else:\n self.ca_verify_filename = self.verify\n return\n", "role": "user" }, { "content": "ssl_verify", "role": "assistant" } ]
def x(self, total=8, connect=None, read=None, redirect=None, status=None, method_whitelist=urllib3.util.retry.Retry.DEFAULT_METHOD_WHITELIST, status_forcelist=None, backoff_factor=0.705883, raise_on_redirect=True, raise_on_status=True, respect_retry_after_header=True, adapter_url= 'https://'): if status_forcelist is None: status_forcelist = 413, 429, 502, 503, 504 retry = urllib3.util.retry.Retry(total=total, connect=connect, read= read, redirect=redirect, status=status, method_whitelist= method_whitelist, status_forcelist=status_forcelist, backoff_factor =backoff_factor, raise_on_redirect=raise_on_redirect, raise_on_status=raise_on_status, respect_retry_after_header= respect_retry_after_header) adapter = requests.adapters.HTTPAdapter(max_retries=retry) self._session.mount(adapter_url, adapter) return
modify_rest_retry
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, total=8, connect=None, read=None, redirect=None, status=None,\n method_whitelist=urllib3.util.retry.Retry.DEFAULT_METHOD_WHITELIST,\n status_forcelist=None, backoff_factor=0.705883, raise_on_redirect=True,\n raise_on_status=True, respect_retry_after_header=True, adapter_url=\n 'https://'):\n if status_forcelist is None:\n status_forcelist = 413, 429, 502, 503, 504\n retry = urllib3.util.retry.Retry(total=total, connect=connect, read=\n read, redirect=redirect, status=status, method_whitelist=\n method_whitelist, status_forcelist=status_forcelist, backoff_factor\n =backoff_factor, raise_on_redirect=raise_on_redirect,\n raise_on_status=raise_on_status, respect_retry_after_header=\n respect_retry_after_header)\n adapter = requests.adapters.HTTPAdapter(max_retries=retry)\n self._session.mount(adapter_url, adapter)\n return\n", "role": "user" }, { "content": "modify_rest_retry", "role": "assistant" } ]