source_code
stringlengths 45
7.24k
| function_name
stringlengths 1
111
| messages
list |
|---|---|---|
def x(time_filter):
if not time_filter:
time_filter = '[* TO *]'
start, end = parse_solr_time_range_as_pair(time_filter)
start, end = parse_datetime(start), parse_datetime(end)
return start, end
|
parse_datetime_range
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(time_filter):\n if not time_filter:\n time_filter = '[* TO *]'\n start, end = parse_solr_time_range_as_pair(time_filter)\n start, end = parse_datetime(start), parse_datetime(end)\n return start, end\n",
"role": "user"
},
{
"content": "parse_datetime_range",
"role": "assistant"
}
] |
def x(time_gap):
matcher = None
if time_gap.count('T'):
units = {'H': ('HOURS', isodate.Duration(hours=1)), 'M': ('MINUTES',
isodate.Duration(minutes=1)), 'S': ('SECONDS', isodate.Duration
(seconds=1))}
matcher = re.search('PT(\\d+)([HMS])', time_gap)
if matcher:
quantity = int(matcher.group(1))
unit = matcher.group(2)
return quantity, units.get(unit)
else:
raise Exception('Does not match the pattern: {}'.format(time_gap))
else:
units = {'Y': ('YEARS', isodate.Duration(years=1)), 'M': ('MONTHS',
isodate.Duration(months=1)), 'W': ('WEEKS', isodate.Duration(
weeks=1)), 'D': ('DAYS', isodate.Duration(days=1))}
matcher = re.search('P(\\d+)([YMWD])', time_gap)
if matcher:
quantity = int(matcher.group(1))
unit = matcher.group(2)
else:
raise Exception('Does not match the pattern: {}'.format(time_gap))
return quantity, units.get(unit)
|
parse_ISO8601
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(time_gap):\n matcher = None\n if time_gap.count('T'):\n units = {'H': ('HOURS', isodate.Duration(hours=1)), 'M': ('MINUTES',\n isodate.Duration(minutes=1)), 'S': ('SECONDS', isodate.Duration\n (seconds=1))}\n matcher = re.search('PT(\\\\d+)([HMS])', time_gap)\n if matcher:\n quantity = int(matcher.group(1))\n unit = matcher.group(2)\n return quantity, units.get(unit)\n else:\n raise Exception('Does not match the pattern: {}'.format(time_gap))\n else:\n units = {'Y': ('YEARS', isodate.Duration(years=1)), 'M': ('MONTHS',\n isodate.Duration(months=1)), 'W': ('WEEKS', isodate.Duration(\n weeks=1)), 'D': ('DAYS', isodate.Duration(days=1))}\n matcher = re.search('P(\\\\d+)([YMWD])', time_gap)\n if matcher:\n quantity = int(matcher.group(1))\n unit = matcher.group(2)\n else:\n raise Exception('Does not match the pattern: {}'.format(time_gap))\n return quantity, units.get(unit)\n",
"role": "user"
},
{
"content": "parse_ISO8601",
"role": "assistant"
}
] |
def x(start, end, time_limit):
if is_range_common_era(start, end):
duration = end.get('parsed_datetime') - start.get('parsed_datetime')
unit = int(math.ceil(duration.days / float(time_limit)))
return '+{0}DAYS'.format(unit)
else:
return '+100YEARS'
|
compute_gap
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(start, end, time_limit):\n if is_range_common_era(start, end):\n duration = end.get('parsed_datetime') - start.get('parsed_datetime')\n unit = int(math.ceil(duration.days / float(time_limit)))\n return '+{0}DAYS'.format(unit)\n else:\n return '+100YEARS'\n",
"role": "user"
},
{
"content": "compute_gap",
"role": "assistant"
}
] |
def x(time_gap):
quantity, unit = parse_ISO8601(time_gap)
if unit[0] == 'WEEKS':
return '+{0}DAYS'.format(quantity * 7)
else:
return '+{0}{1}'.format(quantity, unit[0])
|
gap_to_sorl
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(time_gap):\n quantity, unit = parse_ISO8601(time_gap)\n if unit[0] == 'WEEKS':\n return '+{0}DAYS'.format(quantity * 7)\n else:\n return '+{0}{1}'.format(quantity, unit[0])\n",
"role": "user"
},
{
"content": "gap_to_sorl",
"role": "assistant"
}
] |
def x(field, time_filter, time_gap, time_limit=100):
start, end = parse_datetime_range(time_filter)
key_range_start = 'f.{0}.facet.range.start'.format(field)
key_range_end = 'f.{0}.facet.range.end'.format(field)
key_range_gap = 'f.{0}.facet.range.gap'.format(field)
key_range_mincount = 'f.{0}.facet.mincount'.format(field)
if time_gap:
gap = gap_to_sorl(time_gap)
else:
gap = compute_gap(start, end, time_limit)
value_range_start = start.get('parsed_datetime')
if start.get('is_common_era'):
value_range_start = start.get('parsed_datetime').isoformat().replace(
'+00:00', '') + 'Z'
value_range_end = start.get('parsed_datetime')
if end.get('is_common_era'):
value_range_end = end.get('parsed_datetime').isoformat().replace(
'+00:00', '') + 'Z'
value_range_gap = gap
params = {'facet.range': field, key_range_start: value_range_start,
key_range_end: value_range_end, key_range_gap: value_range_gap,
key_range_mincount: 1}
return params
|
request_time_facet
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(field, time_filter, time_gap, time_limit=100):\n start, end = parse_datetime_range(time_filter)\n key_range_start = 'f.{0}.facet.range.start'.format(field)\n key_range_end = 'f.{0}.facet.range.end'.format(field)\n key_range_gap = 'f.{0}.facet.range.gap'.format(field)\n key_range_mincount = 'f.{0}.facet.mincount'.format(field)\n if time_gap:\n gap = gap_to_sorl(time_gap)\n else:\n gap = compute_gap(start, end, time_limit)\n value_range_start = start.get('parsed_datetime')\n if start.get('is_common_era'):\n value_range_start = start.get('parsed_datetime').isoformat().replace(\n '+00:00', '') + 'Z'\n value_range_end = start.get('parsed_datetime')\n if end.get('is_common_era'):\n value_range_end = end.get('parsed_datetime').isoformat().replace(\n '+00:00', '') + 'Z'\n value_range_gap = gap\n params = {'facet.range': field, key_range_start: value_range_start,\n key_range_end: value_range_end, key_range_gap: value_range_gap,\n key_range_mincount: 1}\n return params\n",
"role": "user"
},
{
"content": "request_time_facet",
"role": "assistant"
}
] |
def x(geo_box_str):
pattern = '\\[(.*) TO (.*)\\]'
matcher = re.search(pattern, geo_box_str)
if matcher:
return matcher.group(1), matcher.group(2)
else:
raise Exception('Regex {0} could not parse {1}'.format(pattern,
geo_box_str))
|
parse_solr_geo_range_as_pair
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(geo_box_str):\n pattern = '\\\\[(.*) TO (.*)\\\\]'\n matcher = re.search(pattern, geo_box_str)\n if matcher:\n return matcher.group(1), matcher.group(2)\n else:\n raise Exception('Regex {0} could not parse {1}'.format(pattern,\n geo_box_str))\n",
"role": "user"
},
{
"content": "parse_solr_geo_range_as_pair",
"role": "assistant"
}
] |
def x(geo_box_str):
from_point_str, to_point_str = parse_solr_geo_range_as_pair(geo_box_str)
from_point = parse_lat_lon(from_point_str)
to_point = parse_lat_lon(to_point_str)
rectangle = box(from_point[0], from_point[1], to_point[0], to_point[1])
return rectangle
|
parse_geo_box
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(geo_box_str):\n from_point_str, to_point_str = parse_solr_geo_range_as_pair(geo_box_str)\n from_point = parse_lat_lon(from_point_str)\n to_point = parse_lat_lon(to_point_str)\n rectangle = box(from_point[0], from_point[1], to_point[0], to_point[1])\n return rectangle\n",
"role": "user"
},
{
"content": "parse_geo_box",
"role": "assistant"
}
] |
def x(field, hm_filter, hm_grid_level, hm_limit):
if not hm_filter:
hm_filter = '[-90,-180 TO 90,180]'
params = {'facet': 'on', 'facet.heatmap': field, 'facet.heatmap.geom':
hm_filter}
if hm_grid_level:
params['facet.heatmap.gridLevel'] = hm_grid_level
else:
rectangle = parse_geo_box(hm_filter)
degrees_side_length = rectangle.length / 2
cell_side_length = math.sqrt(float(hm_limit))
cell_side_length_degrees = degrees_side_length / cell_side_length * 2
params['facet.heatmap.distErr'] = str(float(cell_side_length_degrees))
return params
|
request_heatmap_facet
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(field, hm_filter, hm_grid_level, hm_limit):\n if not hm_filter:\n hm_filter = '[-90,-180 TO 90,180]'\n params = {'facet': 'on', 'facet.heatmap': field, 'facet.heatmap.geom':\n hm_filter}\n if hm_grid_level:\n params['facet.heatmap.gridLevel'] = hm_grid_level\n else:\n rectangle = parse_geo_box(hm_filter)\n degrees_side_length = rectangle.length / 2\n cell_side_length = math.sqrt(float(hm_limit))\n cell_side_length_degrees = degrees_side_length / cell_side_length * 2\n params['facet.heatmap.distErr'] = str(float(cell_side_length_degrees))\n return params\n",
"role": "user"
},
{
"content": "request_heatmap_facet",
"role": "assistant"
}
] |
def x(field, time_filter, search_engine_endpoint, actual_params=None):
if actual_params:
raise NotImplemented('actual_params')
start, end = parse_solr_time_range_as_pair(time_filter)
if start == '*' or end == '*':
params_stats = {'q': '*:*', 'rows': 0, 'stats.field': field,
'stats': 'true', 'wt': 'json'}
res_stats = requests.get(search_engine_endpoint, params=params_stats)
if res_stats.ok:
stats_date_field = res_stats.json()['stats']['stats_fields'][field]
date_min = stats_date_field['min']
date_max = stats_date_field['max']
if start != '*':
date_min = start
if end != '*':
date_max = end
time_filter = '[{0} TO {1}]'.format(date_min, date_max)
return time_filter
|
asterisk_to_min_max
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(field, time_filter, search_engine_endpoint, actual_params=None):\n if actual_params:\n raise NotImplemented('actual_params')\n start, end = parse_solr_time_range_as_pair(time_filter)\n if start == '*' or end == '*':\n params_stats = {'q': '*:*', 'rows': 0, 'stats.field': field,\n 'stats': 'true', 'wt': 'json'}\n res_stats = requests.get(search_engine_endpoint, params=params_stats)\n if res_stats.ok:\n stats_date_field = res_stats.json()['stats']['stats_fields'][field]\n date_min = stats_date_field['min']\n date_max = stats_date_field['max']\n if start != '*':\n date_min = start\n if end != '*':\n date_max = end\n time_filter = '[{0} TO {1}]'.format(date_min, date_max)\n return time_filter\n",
"role": "user"
},
{
"content": "asterisk_to_min_max",
"role": "assistant"
}
] |
def x(raw_xml):
from pycsw.core.etree import etree
parsed = etree.fromstring(raw_xml, etree.XMLParser(resolve_entities=False))
source_tag = parsed.find('{http://purl.org/dc/elements/1.1/}source')
format_tag = parsed.find('{http://purl.org/dc/elements/1.1/}format')
service_url = None
service_type = None
if hasattr(source_tag, 'text'):
service_url = source_tag.text
if hasattr(format_tag, 'text'):
service_type = format_tag.text
if hasattr(format_tag, 'text'):
service_type = format_tag.text
service, created = Service.objects.get_or_create(url=service_url,
is_monitored=False, type=service_type)
srs, created = SpatialReferenceSystem.objects.get_or_create(code=
'EPSG:4326')
service.srs.add(srs)
return service
|
get_service
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(raw_xml):\n from pycsw.core.etree import etree\n parsed = etree.fromstring(raw_xml, etree.XMLParser(resolve_entities=False))\n source_tag = parsed.find('{http://purl.org/dc/elements/1.1/}source')\n format_tag = parsed.find('{http://purl.org/dc/elements/1.1/}format')\n service_url = None\n service_type = None\n if hasattr(source_tag, 'text'):\n service_url = source_tag.text\n if hasattr(format_tag, 'text'):\n service_type = format_tag.text\n if hasattr(format_tag, 'text'):\n service_type = format_tag.text\n service, created = Service.objects.get_or_create(url=service_url,\n is_monitored=False, type=service_type)\n srs, created = SpatialReferenceSystem.objects.get_or_create(code=\n 'EPSG:4326')\n service.srs.add(srs)\n return service\n",
"role": "user"
},
{
"content": "get_service",
"role": "assistant"
}
] |
def x(self, ids):
results = self._get_repo_filter(Layer.objects).filter(uuid__in=ids).all()
if len(results) == 0:
results = self._get_repo_filter(Service.objects).filter(uuid__in=ids
).all()
return results
|
query_ids
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ids):\n results = self._get_repo_filter(Layer.objects).filter(uuid__in=ids).all()\n if len(results) == 0:\n results = self._get_repo_filter(Service.objects).filter(uuid__in=ids\n ).all()\n return results\n",
"role": "user"
},
{
"content": "query_ids",
"role": "assistant"
}
] |
def x(self, domain, typenames, domainquerytype='list', count=False):
objects = self._get_repo_filter(Layer.objects)
if domainquerytype == 'range':
return [tuple(objects.aggregate(Min(domain), Max(domain)).values())]
elif count:
return [(d[domain], d['%s__count' % domain]) for d in objects.
values(domain).annotate(Count(domain))]
else:
return objects.values_list(domain).distinct()
|
query_domain
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, domain, typenames, domainquerytype='list', count=False):\n objects = self._get_repo_filter(Layer.objects)\n if domainquerytype == 'range':\n return [tuple(objects.aggregate(Min(domain), Max(domain)).values())]\n elif count:\n return [(d[domain], d['%s__count' % domain]) for d in objects.\n values(domain).annotate(Count(domain))]\n else:\n return objects.values_list(domain).distinct()\n",
"role": "user"
},
{
"content": "query_domain",
"role": "assistant"
}
] |
def x(self, direction='max'):
if direction == 'min':
return Layer.objects.aggregate(Min('last_updated'))['last_updated__min'
].strftime('%Y-%m-%dT%H:%M:%SZ')
return self._get_repo_filter(Layer.objects).aggregate(Max('last_updated'))[
'last_updated__max'].strftime('%Y-%m-%dT%H:%M:%SZ')
|
query_insert
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, direction='max'):\n if direction == 'min':\n return Layer.objects.aggregate(Min('last_updated'))['last_updated__min'\n ].strftime('%Y-%m-%dT%H:%M:%SZ')\n return self._get_repo_filter(Layer.objects).aggregate(Max('last_updated'))[\n 'last_updated__max'].strftime('%Y-%m-%dT%H:%M:%SZ')\n",
"role": "user"
},
{
"content": "query_insert",
"role": "assistant"
}
] |
def x(self, source):
return self._get_repo_filter(Layer.objects).filter(url=source)
|
query_source
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, source):\n return self._get_repo_filter(Layer.objects).filter(url=source)\n",
"role": "user"
},
{
"content": "query_source",
"role": "assistant"
}
] |
def x(self, constraint, sortby=None, typenames=None, maxrecords=10,
startposition=0):
if 'where' in constraint:
query = self._get_repo_filter(Layer.objects).filter(is_valid=True
).extra(where=[constraint['where']], params=constraint['values'])
else:
query = self._get_repo_filter(Layer.objects).filter(is_valid=True)
total = query.count()
if sortby is not None:
if 'spatial' in sortby and sortby['spatial']:
desc = False
if sortby['order'] == 'DESC':
desc = True
query = query.all()
return [str(total), sorted(query, key=lambda x: float(util.
get_geometry_area(getattr(x, sortby['propertyname']))),
reverse=desc)[startposition:startposition + int(maxrecords)]]
else:
if sortby['order'] == 'DESC':
pname = '-%s' % sortby['propertyname']
else:
pname = sortby['propertyname']
return [str(total), query.order_by(pname)[startposition:
startposition + int(maxrecords)]]
else:
return [str(total), query.all()[startposition:startposition + int(
maxrecords)]]
|
query
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, constraint, sortby=None, typenames=None, maxrecords=10,\n startposition=0):\n if 'where' in constraint:\n query = self._get_repo_filter(Layer.objects).filter(is_valid=True\n ).extra(where=[constraint['where']], params=constraint['values'])\n else:\n query = self._get_repo_filter(Layer.objects).filter(is_valid=True)\n total = query.count()\n if sortby is not None:\n if 'spatial' in sortby and sortby['spatial']:\n desc = False\n if sortby['order'] == 'DESC':\n desc = True\n query = query.all()\n return [str(total), sorted(query, key=lambda x: float(util.\n get_geometry_area(getattr(x, sortby['propertyname']))),\n reverse=desc)[startposition:startposition + int(maxrecords)]]\n else:\n if sortby['order'] == 'DESC':\n pname = '-%s' % sortby['propertyname']\n else:\n pname = sortby['propertyname']\n return [str(total), query.order_by(pname)[startposition:\n startposition + int(maxrecords)]]\n else:\n return [str(total), query.all()[startposition:startposition + int(\n maxrecords)]]\n",
"role": "user"
},
{
"content": "query",
"role": "assistant"
}
] |
def x(self, resourcetype, source, insert_date=None):
caller = inspect.stack()[1][3]
if caller == 'transaction':
hhclass = 'Layer'
source = resourcetype
resourcetype = resourcetype.csw_schema
else:
hhclass = 'Service'
if resourcetype not in HYPERMAP_SERVICE_TYPES.keys():
raise RuntimeError('Unsupported Service Type')
return self._insert_or_update(resourcetype, source, mode='insert',
hhclass=hhclass)
|
insert
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, resourcetype, source, insert_date=None):\n caller = inspect.stack()[1][3]\n if caller == 'transaction':\n hhclass = 'Layer'\n source = resourcetype\n resourcetype = resourcetype.csw_schema\n else:\n hhclass = 'Service'\n if resourcetype not in HYPERMAP_SERVICE_TYPES.keys():\n raise RuntimeError('Unsupported Service Type')\n return self._insert_or_update(resourcetype, source, mode='insert',\n hhclass=hhclass)\n",
"role": "user"
},
{
"content": "insert",
"role": "assistant"
}
] |
def x(self, resourcetype, source, mode='insert', hhclass='Service'):
keywords = []
if self.filter is not None:
catalog = Catalog.objects.get(id=int(self.filter.split()[-1]))
try:
if hhclass == 'Layer':
match = Layer.objects.filter(name=source.name, title=source.
title, abstract=source.abstract, is_monitored=False)
matches = match.all()
if matches:
if mode == 'insert':
raise RuntimeError(
"HHypermap error: Layer %d '%s' already exists" % (
matches[0].id, source.title))
elif mode == 'update':
match.update(name=source.name, title=source.title,
abstract=source.abstract, is_monitored=False, xml=
source.xml, wkt_geometry=source.wkt_geometry,
anytext=util.get_anytext([source.title, source.
abstract, source.keywords_csv]))
service = get_service(source.xml)
res, keywords = create_layer_from_metadata_xml(resourcetype,
source.xml, monitor=False, service=service, catalog=catalog)
res.save()
LOGGER.debug('Indexing layer with id %s on search engine' % res
.uuid)
index_layer(res.id, use_cache=True)
else:
if resourcetype == 'http://www.opengis.net/cat/csw/2.0.2':
res = Endpoint(url=source, catalog=catalog)
else:
res = Service(type=HYPERMAP_SERVICE_TYPES[resourcetype],
url=source, catalog=catalog)
res.save()
if keywords:
for kw in keywords:
res.keywords.add(kw)
except Exception as err:
raise RuntimeError('HHypermap error: %s' % err)
ids = []
if hhclass == 'Layer':
ids.append({'identifier': res.uuid, 'title': res.title})
elif resourcetype == 'http://www.opengis.net/cat/csw/2.0.2':
for res in Endpoint.objects.filter(url=source).all():
ids.append({'identifier': res.uuid, 'title': res.url})
else:
for res in Service.objects.filter(url=source).all():
ids.append({'identifier': res.uuid, 'title': res.title})
return ids
|
insert_or_update
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, resourcetype, source, mode='insert', hhclass='Service'):\n keywords = []\n if self.filter is not None:\n catalog = Catalog.objects.get(id=int(self.filter.split()[-1]))\n try:\n if hhclass == 'Layer':\n match = Layer.objects.filter(name=source.name, title=source.\n title, abstract=source.abstract, is_monitored=False)\n matches = match.all()\n if matches:\n if mode == 'insert':\n raise RuntimeError(\n \"HHypermap error: Layer %d '%s' already exists\" % (\n matches[0].id, source.title))\n elif mode == 'update':\n match.update(name=source.name, title=source.title,\n abstract=source.abstract, is_monitored=False, xml=\n source.xml, wkt_geometry=source.wkt_geometry,\n anytext=util.get_anytext([source.title, source.\n abstract, source.keywords_csv]))\n service = get_service(source.xml)\n res, keywords = create_layer_from_metadata_xml(resourcetype,\n source.xml, monitor=False, service=service, catalog=catalog)\n res.save()\n LOGGER.debug('Indexing layer with id %s on search engine' % res\n .uuid)\n index_layer(res.id, use_cache=True)\n else:\n if resourcetype == 'http://www.opengis.net/cat/csw/2.0.2':\n res = Endpoint(url=source, catalog=catalog)\n else:\n res = Service(type=HYPERMAP_SERVICE_TYPES[resourcetype],\n url=source, catalog=catalog)\n res.save()\n if keywords:\n for kw in keywords:\n res.keywords.add(kw)\n except Exception as err:\n raise RuntimeError('HHypermap error: %s' % err)\n ids = []\n if hhclass == 'Layer':\n ids.append({'identifier': res.uuid, 'title': res.title})\n elif resourcetype == 'http://www.opengis.net/cat/csw/2.0.2':\n for res in Endpoint.objects.filter(url=source).all():\n ids.append({'identifier': res.uuid, 'title': res.url})\n else:\n for res in Service.objects.filter(url=source).all():\n ids.append({'identifier': res.uuid, 'title': res.title})\n return ids\n",
"role": "user"
},
{
"content": "insert_or_update",
"role": "assistant"
}
] |
def x(self, constraint):
results = self._get_repo_filter(Service.objects).extra(where=[
constraint['where']], params=constraint['values']).all()
deleted = len(results)
results.delete()
return deleted
|
delete
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, constraint):\n results = self._get_repo_filter(Service.objects).extra(where=[\n constraint['where']], params=constraint['values']).all()\n deleted = len(results)\n results.delete()\n return deleted\n",
"role": "user"
},
{
"content": "delete",
"role": "assistant"
}
] |
def x(self, query):
if self.filter is not None:
return query.extra(where=[self.filter])
return query
|
get_repo_filter
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, query):\n if self.filter is not None:\n return query.extra(where=[self.filter])\n return query\n",
"role": "user"
},
{
"content": "get_repo_filter",
"role": "assistant"
}
] |
def x(meshcode):
length = len(str(meshcode))
if length == 4:
return 1
if length == 5:
return 40000
if length == 6:
return 2
if length == 7:
if meshcode[6:7] in ['1', '2', '3', '4']:
return 5000
if meshcode[6:7] == '6':
return 8000
if meshcode[6:7] == '5':
return 20000
if meshcode[6:7] == '7':
return 16000
if length == 8:
return 3
if length == 9:
if meshcode[8:9] in ['1', '2', '3', '4']:
return 4
if meshcode[8:9] == '5':
return 2000
if meshcode[8:9] == '6':
return 2500
if meshcode[8:9] == '7':
return 4000
if length == 10:
if meshcode[9:10] in ['1', '2', '3', '4']:
return 5
if length == 11:
if meshcode[10:11] in ['1', '2', '3', '4']:
return 6
raise ValueError('the meshcode is unsupported.')
|
to_meshlevel
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(meshcode):\n length = len(str(meshcode))\n if length == 4:\n return 1\n if length == 5:\n return 40000\n if length == 6:\n return 2\n if length == 7:\n if meshcode[6:7] in ['1', '2', '3', '4']:\n return 5000\n if meshcode[6:7] == '6':\n return 8000\n if meshcode[6:7] == '5':\n return 20000\n if meshcode[6:7] == '7':\n return 16000\n if length == 8:\n return 3\n if length == 9:\n if meshcode[8:9] in ['1', '2', '3', '4']:\n return 4\n if meshcode[8:9] == '5':\n return 2000\n if meshcode[8:9] == '6':\n return 2500\n if meshcode[8:9] == '7':\n return 4000\n if length == 10:\n if meshcode[9:10] in ['1', '2', '3', '4']:\n return 5\n if length == 11:\n if meshcode[10:11] in ['1', '2', '3', '4']:\n return 6\n raise ValueError('the meshcode is unsupported.')\n",
"role": "user"
},
{
"content": "to_meshlevel",
"role": "assistant"
}
] |
def x(func):
def x(request, *args, **kwargs):
if not request.method == 'POST':
return HttpResponseBadRequest('Must be POST request.')
follow = func(request, *args, **kwargs)
if request.is_ajax():
return HttpResponse('ok')
try:
if 'next' in request.GET:
return HttpResponseRedirect(request.GET.get('next'))
if 'next' in request.POST:
return HttpResponseRedirect(request.POST.get('next'))
return HttpResponseRedirect(follow.target.get_absolute_url())
except (AttributeError, TypeError):
if 'HTTP_REFERER' in request.META:
return HttpResponseRedirect(request.META.get('HTTP_REFERER',
'/'))
if follow:
return HttpResponseServerError(
'"%s" object of type ``%s`` has no method ``get_absolute_url()``.'
% (unicode(follow.target), follow.target.__class__))
return HttpResponseServerError(
'No follow object and `next` parameter found.')
return iCheck
|
check
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(func):\n\n def x(request, *args, **kwargs):\n if not request.method == 'POST':\n return HttpResponseBadRequest('Must be POST request.')\n follow = func(request, *args, **kwargs)\n if request.is_ajax():\n return HttpResponse('ok')\n try:\n if 'next' in request.GET:\n return HttpResponseRedirect(request.GET.get('next'))\n if 'next' in request.POST:\n return HttpResponseRedirect(request.POST.get('next'))\n return HttpResponseRedirect(follow.target.get_absolute_url())\n except (AttributeError, TypeError):\n if 'HTTP_REFERER' in request.META:\n return HttpResponseRedirect(request.META.get('HTTP_REFERER',\n '/'))\n if follow:\n return HttpResponseServerError(\n '\"%s\" object of type ``%s`` has no method ``get_absolute_url()``.'\n % (unicode(follow.target), follow.target.__class__))\n return HttpResponseServerError(\n 'No follow object and `next` parameter found.')\n return iCheck\n",
"role": "user"
},
{
"content": "check",
"role": "assistant"
}
] |
def x(model, field_name=None, related_name=None, lookup_method_name=
'get_follows'):
if model in registry:
return
registry.append(model)
if not field_name:
field_name = 'target_%s' % model._meta.module_name
if not related_name:
related_name = 'follow_%s' % model._meta.module_name
field = ForeignKey(model, related_name=related_name, null=True, blank=
True, db_index=True)
field.contribute_to_class(Follow, field_name)
setattr(model, lookup_method_name, get_followers_for_object)
model_map[model] = [related_name, field_name]
|
register
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(model, field_name=None, related_name=None, lookup_method_name=\n 'get_follows'):\n if model in registry:\n return\n registry.append(model)\n if not field_name:\n field_name = 'target_%s' % model._meta.module_name\n if not related_name:\n related_name = 'follow_%s' % model._meta.module_name\n field = ForeignKey(model, related_name=related_name, null=True, blank=\n True, db_index=True)\n field.contribute_to_class(Follow, field_name)\n setattr(model, lookup_method_name, get_followers_for_object)\n model_map[model] = [related_name, field_name]\n",
"role": "user"
},
{
"content": "register",
"role": "assistant"
}
] |
def x(user, obj):
follow, created = Follow.objects.get_or_create(user, obj)
return follow
|
follow
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(user, obj):\n follow, created = Follow.objects.get_or_create(user, obj)\n return follow\n",
"role": "user"
},
{
"content": "follow",
"role": "assistant"
}
] |
def x(user, obj):
try:
follow = Follow.objects.get_follows(obj).get(user=user)
follow.delete()
return follow
except Follow.DoesNotExist:
pass
|
unfollow
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(user, obj):\n try:\n follow = Follow.objects.get_follows(obj).get(user=user)\n follow.delete()\n return follow\n except Follow.DoesNotExist:\n pass\n",
"role": "user"
},
{
"content": "unfollow",
"role": "assistant"
}
] |
def x(user, obj):
if Follow.objects.is_following(user, obj):
return unfollow(user, obj)
return follow(user, obj)
|
toggle
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(user, obj):\n if Follow.objects.is_following(user, obj):\n return unfollow(user, obj)\n return follow(user, obj)\n",
"role": "user"
},
{
"content": "toggle",
"role": "assistant"
}
] |
def x(self, value):
if value:
try:
range = utils.parse_datetime_range_to_solr(value)
return range
except Exception as e:
raise serializers.ValidationError(e.message)
return value
|
validate_q_time
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, value):\n if value:\n try:\n range = utils.parse_datetime_range_to_solr(value)\n return range\n except Exception as e:\n raise serializers.ValidationError(e.message)\n return value\n",
"role": "user"
},
{
"content": "validate_q_time",
"role": "assistant"
}
] |
def x(self, value):
if value:
try:
rectangle = utils.parse_geo_box(value)
return '[{0},{1} TO {2},{3}]'.format(rectangle.bounds[0],
rectangle.bounds[1], rectangle.bounds[2], rectangle.bounds[3])
except Exception as e:
raise serializers.ValidationError(e.message)
return value
|
validate_q_geo
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, value):\n if value:\n try:\n rectangle = utils.parse_geo_box(value)\n return '[{0},{1} TO {2},{3}]'.format(rectangle.bounds[0],\n rectangle.bounds[1], rectangle.bounds[2], rectangle.bounds[3])\n except Exception as e:\n raise serializers.ValidationError(e.message)\n return value\n",
"role": "user"
},
{
"content": "validate_q_geo",
"role": "assistant"
}
] |
def x(self, value):
if value:
try:
utils.parse_datetime_range(value)
except Exception as e:
raise serializers.ValidationError(e.message)
return value
|
validate_a_time_filter
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, value):\n if value:\n try:\n utils.parse_datetime_range(value)\n except Exception as e:\n raise serializers.ValidationError(e.message)\n return value\n",
"role": "user"
},
{
"content": "validate_a_time_filter",
"role": "assistant"
}
] |
def x(self, model_or_obj_or_qs):
if isinstance(model_or_obj_or_qs, QuerySet):
_, fname = model_map[model_or_obj_or_qs.model]
else:
cls = model_or_obj_or_qs if inspect.isclass(model_or_obj_or_qs
) else model_or_obj_or_qs.__class__
_, fname = model_map[cls]
return fname
|
fname
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, model_or_obj_or_qs):\n if isinstance(model_or_obj_or_qs, QuerySet):\n _, fname = model_map[model_or_obj_or_qs.model]\n else:\n cls = model_or_obj_or_qs if inspect.isclass(model_or_obj_or_qs\n ) else model_or_obj_or_qs.__class__\n _, fname = model_map[cls]\n return fname\n",
"role": "user"
},
{
"content": "fname",
"role": "assistant"
}
] |
def x(self, user, obj, **kwargs):
follow = Follow(user=user)
follow.target = obj
follow.save()
return follow
|
create
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, user, obj, **kwargs):\n follow = Follow(user=user)\n follow.target = obj\n follow.save()\n return follow\n",
"role": "user"
},
{
"content": "create",
"role": "assistant"
}
] |
def x(self, user, obj, **kwargs):
if not self.is_following(user, obj):
return self.create(user, obj, **kwargs), True
return self.get_follows(obj).get(user=user), False
|
get_or_create
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, user, obj, **kwargs):\n if not self.is_following(user, obj):\n return self.create(user, obj, **kwargs), True\n return self.get_follows(obj).get(user=user), False\n",
"role": "user"
},
{
"content": "get_or_create",
"role": "assistant"
}
] |
def x(self, user, obj):
if isinstance(user, AnonymousUser):
return False
return 0 < self.get_follows(obj).filter(user=user).count()
|
is_following
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, user, obj):\n if isinstance(user, AnonymousUser):\n return False\n return 0 < self.get_follows(obj).filter(user=user).count()\n",
"role": "user"
},
{
"content": "is_following",
"role": "assistant"
}
] |
def x(self, model_or_obj_or_qs):
fname = self.fname(model_or_obj_or_qs)
if isinstance(model_or_obj_or_qs, QuerySet):
return self.filter(**{('%s__in' % fname): model_or_obj_or_qs})
if inspect.isclass(model_or_obj_or_qs):
return self.exclude(**{fname: None})
return self.filter(**{fname: model_or_obj_or_qs})
|
get_follows
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, model_or_obj_or_qs):\n fname = self.fname(model_or_obj_or_qs)\n if isinstance(model_or_obj_or_qs, QuerySet):\n return self.filter(**{('%s__in' % fname): model_or_obj_or_qs})\n if inspect.isclass(model_or_obj_or_qs):\n return self.exclude(**{fname: None})\n return self.filter(**{fname: model_or_obj_or_qs})\n",
"role": "user"
},
{
"content": "get_follows",
"role": "assistant"
}
] |
def x(self, event_times_indices, covariates=None, durations=None):
if covariates is None:
covariates = np.ones(self.event_times_indices.shape)
if durations is None:
durations = np.ones(self.event_times_indices.shape)
else:
durations = np.round(durations * self.deconvolution_frequency).astype(
int)
mean_duration = np.mean(durations)
regressors_for_event = np.zeros((self.deconvolution_interval_size, self
.resampled_signal_size))
for cov, eti, dur in zip(covariates, event_times_indices, durations):
valid = True
if eti < 0:
self.logger.debug(
'deconv samples are starting before the data starts.')
valid = False
if eti + self.deconvolution_interval_size > self.resampled_signal_size:
self.logger.debug(
'deconv samples are continuing after the data stops.')
valid = False
if eti > self.resampled_signal_size:
self.logger.debug('event falls outside of the scope of the data.')
valid = False
if valid:
this_event_design_matrix = np.diag(np.ones(self.
deconvolution_interval_size)) * cov
over_durations_dm = np.copy(this_event_design_matrix)
if dur > 1:
for d in np.arange(1, dur):
over_durations_dm[d:] += this_event_design_matrix[:-d]
over_durations_dm /= mean_duration
regressors_for_event[:, eti:int(eti + self.
deconvolution_interval_size)] += over_durations_dm
return regressors_for_event
|
create_event_regressors
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, event_times_indices, covariates=None, durations=None):\n if covariates is None:\n covariates = np.ones(self.event_times_indices.shape)\n if durations is None:\n durations = np.ones(self.event_times_indices.shape)\n else:\n durations = np.round(durations * self.deconvolution_frequency).astype(\n int)\n mean_duration = np.mean(durations)\n regressors_for_event = np.zeros((self.deconvolution_interval_size, self\n .resampled_signal_size))\n for cov, eti, dur in zip(covariates, event_times_indices, durations):\n valid = True\n if eti < 0:\n self.logger.debug(\n 'deconv samples are starting before the data starts.')\n valid = False\n if eti + self.deconvolution_interval_size > self.resampled_signal_size:\n self.logger.debug(\n 'deconv samples are continuing after the data stops.')\n valid = False\n if eti > self.resampled_signal_size:\n self.logger.debug('event falls outside of the scope of the data.')\n valid = False\n if valid:\n this_event_design_matrix = np.diag(np.ones(self.\n deconvolution_interval_size)) * cov\n over_durations_dm = np.copy(this_event_design_matrix)\n if dur > 1:\n for d in np.arange(1, dur):\n over_durations_dm[d:] += this_event_design_matrix[:-d]\n over_durations_dm /= mean_duration\n regressors_for_event[:, eti:int(eti + self.\n deconvolution_interval_size)] += over_durations_dm\n return regressors_for_event\n",
"role": "user"
},
{
"content": "create_event_regressors",
"role": "assistant"
}
] |
def x(self, demean=False, intercept=True):
self.design_matrix = np.zeros((int(self.number_of_event_types * self.
deconvolution_interval_size), self.resampled_signal_size))
for i, covariate in enumerate(self.covariates.keys()):
self.logger.debug('creating regressor for ' + covariate)
indices = np.arange(i * self.deconvolution_interval_size, (i + 1) *
self.deconvolution_interval_size, dtype=int)
if len(covariate.split('.')) > 0:
which_event_time_indices = covariate.split('.')[0]
else:
which_event_time_indices = covariate
self.design_matrix[indices] = self.create_event_regressors(self.
event_times_indices[which_event_time_indices], self.covariates[
covariate], self.durations[which_event_time_indices])
if demean:
self.design_matrix = (self.design_matrix.T - self.design_matrix.
mean(axis=-1)).T
if intercept:
self.design_matrix = np.vstack((self.design_matrix, np.ones((1,
self.design_matrix.shape[-1]))))
self.logger.debug('created %s design_matrix' % str(self.design_matrix.
shape))
|
create_design_matrix
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, demean=False, intercept=True):\n self.design_matrix = np.zeros((int(self.number_of_event_types * self.\n deconvolution_interval_size), self.resampled_signal_size))\n for i, covariate in enumerate(self.covariates.keys()):\n self.logger.debug('creating regressor for ' + covariate)\n indices = np.arange(i * self.deconvolution_interval_size, (i + 1) *\n self.deconvolution_interval_size, dtype=int)\n if len(covariate.split('.')) > 0:\n which_event_time_indices = covariate.split('.')[0]\n else:\n which_event_time_indices = covariate\n self.design_matrix[indices] = self.create_event_regressors(self.\n event_times_indices[which_event_time_indices], self.covariates[\n covariate], self.durations[which_event_time_indices])\n if demean:\n self.design_matrix = (self.design_matrix.T - self.design_matrix.\n mean(axis=-1)).T\n if intercept:\n self.design_matrix = np.vstack((self.design_matrix, np.ones((1,\n self.design_matrix.shape[-1]))))\n self.logger.debug('created %s design_matrix' % str(self.design_matrix.\n shape))\n",
"role": "user"
},
{
"content": "create_design_matrix",
"role": "assistant"
}
] |
def x(self, regressors):
previous_design_matrix_shape = self.design_matrix.shape
if len(regressors.shape) == 1:
regressors = regressors[np.newaxis, :]
if regressors.shape[1] != self.resampled_signal.shape[1]:
self.logger.warning(
'additional regressor shape %s does not conform to designmatrix shape %s'
% (regressors.shape, self.resampled_signal.shape))
self.design_matrix = np.vstack((self.design_matrix, regressors))
self.logger.debug(
'added %s continuous regressors to %s design_matrix, shape now %s' %
(str(regressors.shape), str(previous_design_matrix_shape), str(self
.design_matrix.shape)))
|
add_continuous_regressors_to_design_matrix
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, regressors):\n previous_design_matrix_shape = self.design_matrix.shape\n if len(regressors.shape) == 1:\n regressors = regressors[np.newaxis, :]\n if regressors.shape[1] != self.resampled_signal.shape[1]:\n self.logger.warning(\n 'additional regressor shape %s does not conform to designmatrix shape %s'\n % (regressors.shape, self.resampled_signal.shape))\n self.design_matrix = np.vstack((self.design_matrix, regressors))\n self.logger.debug(\n 'added %s continuous regressors to %s design_matrix, shape now %s' %\n (str(regressors.shape), str(previous_design_matrix_shape), str(self\n .design_matrix.shape)))\n",
"role": "user"
},
{
"content": "add_continuous_regressors_to_design_matrix",
"role": "assistant"
}
] |
def x(self, method='lstsq'):
if method is 'lstsq':
self.betas, residuals_sum, rank, s = LA.lstsq(self.design_matrix.T,
self.resampled_signal.T)
self.residuals = (self.resampled_signal - self.
predict_from_design_matrix(self.design_matrix))
elif method is 'sm_ols':
import statsmodels.api as sm
assert self.resampled_signal.shape[0
] == 1, 'signal input into statsmodels OLS cannot contain multiple signals at once, present shape %s' % str(
self.resampled_signal.shape)
model = sm.OLS(np.squeeze(self.resampled_signal), self.design_matrix.T)
results = model.fit()
self.betas = np.array(results.params).reshape((self.design_matrix.
shape[0], self.resampled_signal.shape[0]))
self.residuals = np.array(results.resid).reshape(self.
resampled_signal.shape)
self.logger.debug(
'performed %s regression on %s design_matrix and %s signal' % (
method, str(self.design_matrix.shape), str(self.resampled_signal.
shape)))
|
regress
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, method='lstsq'):\n if method is 'lstsq':\n self.betas, residuals_sum, rank, s = LA.lstsq(self.design_matrix.T,\n self.resampled_signal.T)\n self.residuals = (self.resampled_signal - self.\n predict_from_design_matrix(self.design_matrix))\n elif method is 'sm_ols':\n import statsmodels.api as sm\n assert self.resampled_signal.shape[0\n ] == 1, 'signal input into statsmodels OLS cannot contain multiple signals at once, present shape %s' % str(\n self.resampled_signal.shape)\n model = sm.OLS(np.squeeze(self.resampled_signal), self.design_matrix.T)\n results = model.fit()\n self.betas = np.array(results.params).reshape((self.design_matrix.\n shape[0], self.resampled_signal.shape[0]))\n self.residuals = np.array(results.resid).reshape(self.\n resampled_signal.shape)\n self.logger.debug(\n 'performed %s regression on %s design_matrix and %s signal' % (\n method, str(self.design_matrix.shape), str(self.resampled_signal.\n shape)))\n",
"role": "user"
},
{
"content": "regress",
"role": "assistant"
}
] |
def x(self, cv=20, alphas=None):
if alphas is None:
alphas = np.logspace(7, 0, 20)
self.rcv = linear_model.RidgeCV(alphas=alphas, fit_intercept=False, cv=cv)
self.rcv.fit(self.design_matrix.T, self.resampled_signal.T)
self.betas = self.rcv.coef_.T
self.residuals = self.resampled_signal - self.rcv.predict(self.
design_matrix.T)
self.logger.debug(
'performed ridge regression on %s design_matrix and %s signal, resulting alpha value is %f'
% (str(self.design_matrix.shape), str(self.resampled_signal.shape),
self.rcv.alpha_))
|
ridge_regress
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, cv=20, alphas=None):\n if alphas is None:\n alphas = np.logspace(7, 0, 20)\n self.rcv = linear_model.RidgeCV(alphas=alphas, fit_intercept=False, cv=cv)\n self.rcv.fit(self.design_matrix.T, self.resampled_signal.T)\n self.betas = self.rcv.coef_.T\n self.residuals = self.resampled_signal - self.rcv.predict(self.\n design_matrix.T)\n self.logger.debug(\n 'performed ridge regression on %s design_matrix and %s signal, resulting alpha value is %f'\n % (str(self.design_matrix.shape), str(self.resampled_signal.shape),\n self.rcv.alpha_))\n",
"role": "user"
},
{
"content": "ridge_regress",
"role": "assistant"
}
] |
def x(self, covariate='0'):
this_covariate_index = list(self.covariates.keys()).index(covariate)
return self.betas[int(this_covariate_index * self.
deconvolution_interval_size):int((this_covariate_index + 1) * self.
deconvolution_interval_size)]
|
betas_for_cov
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, covariate='0'):\n this_covariate_index = list(self.covariates.keys()).index(covariate)\n return self.betas[int(this_covariate_index * self.\n deconvolution_interval_size):int((this_covariate_index + 1) * self.\n deconvolution_interval_size)]\n",
"role": "user"
},
{
"content": "betas_for_cov",
"role": "assistant"
}
] |
def x(self):
self.betas_per_event_type = np.zeros((len(self.covariates), self.
deconvolution_interval_size, self.resampled_signal.shape[0]))
for i, covariate in enumerate(self.covariates.keys()):
self.betas_per_event_type[i] = self.betas_for_cov(covariate)
|
betas_for_events
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.betas_per_event_type = np.zeros((len(self.covariates), self.\n deconvolution_interval_size, self.resampled_signal.shape[0]))\n for i, covariate in enumerate(self.covariates.keys()):\n self.betas_per_event_type[i] = self.betas_for_cov(covariate)\n",
"role": "user"
},
{
"content": "betas_for_events",
"role": "assistant"
}
] |
def x(self, design_matrix):
assert hasattr(self, 'betas'
), 'no betas found, please run regression before prediction'
assert design_matrix.shape[0] == self.betas.shape[0
], 'designmatrix needs to have the same number of regressors as the betas already calculated'
prediction = np.dot(self.betas.astype(np.float32).T, design_matrix.
astype(np.float32))
return prediction
|
predict_from_design_matrix
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, design_matrix):\n assert hasattr(self, 'betas'\n ), 'no betas found, please run regression before prediction'\n assert design_matrix.shape[0] == self.betas.shape[0\n ], 'designmatrix needs to have the same number of regressors as the betas already calculated'\n prediction = np.dot(self.betas.astype(np.float32).T, design_matrix.\n astype(np.float32))\n return prediction\n",
"role": "user"
},
{
"content": "predict_from_design_matrix",
"role": "assistant"
}
] |
def x(self):
assert hasattr(self, 'betas'
), 'no betas found, please run regression before rsq'
explained_times = self.design_matrix.sum(axis=0) != 0
explained_signal = self.predict_from_design_matrix(self.design_matrix)
self.rsq = 1.0 - np.sum((explained_signal[:, explained_times] - self.
resampled_signal[:, explained_times]) ** 2, axis=-1) / np.sum(self.
resampled_signal[:, explained_times].squeeze() ** 2, axis=-1)
self.ssr = np.sum((explained_signal[:, explained_times] - self.
resampled_signal[:, explained_times]) ** 2, axis=-1)
return np.squeeze(self.rsq)
|
calculate_rsq
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n assert hasattr(self, 'betas'\n ), 'no betas found, please run regression before rsq'\n explained_times = self.design_matrix.sum(axis=0) != 0\n explained_signal = self.predict_from_design_matrix(self.design_matrix)\n self.rsq = 1.0 - np.sum((explained_signal[:, explained_times] - self.\n resampled_signal[:, explained_times]) ** 2, axis=-1) / np.sum(self.\n resampled_signal[:, explained_times].squeeze() ** 2, axis=-1)\n self.ssr = np.sum((explained_signal[:, explained_times] - self.\n resampled_signal[:, explained_times]) ** 2, axis=-1)\n return np.squeeze(self.rsq)\n",
"role": "user"
},
{
"content": "calculate_rsq",
"role": "assistant"
}
] |
def x(self, nr_repetitions=1000):
assert self.resampled_signal.shape[0
] == 1, 'signal input into bootstrap_on_residuals cannot contain signals from multiple channels at once, present shape %s' % str(
self.resampled_signal.shape)
assert hasattr(self, 'betas'
), 'no betas found, please run regression before bootstrapping'
bootstrap_data = np.zeros((self.resampled_signal_size, nr_repetitions))
explained_signal = self.predict_from_design_matrix(self.design_matrix).T
for x in range(bootstrap_data.shape[-1]):
bootstrap_data[:, x] = (self.residuals.T[np.random.permutation(self
.resampled_signal_size)] + explained_signal).squeeze()
self.bootstrap_betas, bs_residuals, rank, s = LA.lstsq(self.
design_matrix.T, bootstrap_data)
self.bootstrap_betas_per_event_type = np.zeros((len(self.covariates),
self.deconvolution_interval_size, nr_repetitions))
for i, covariate in enumerate(list(self.covariates.keys())):
this_covariate_index = list(self.covariates.keys()).index(covariate)
self.bootstrap_betas_per_event_type[i] = self.bootstrap_betas[
this_covariate_index * self.deconvolution_interval_size:(
this_covariate_index + 1) * self.deconvolution_interval_size]
|
bootstrap_on_residuals
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, nr_repetitions=1000):\n assert self.resampled_signal.shape[0\n ] == 1, 'signal input into bootstrap_on_residuals cannot contain signals from multiple channels at once, present shape %s' % str(\n self.resampled_signal.shape)\n assert hasattr(self, 'betas'\n ), 'no betas found, please run regression before bootstrapping'\n bootstrap_data = np.zeros((self.resampled_signal_size, nr_repetitions))\n explained_signal = self.predict_from_design_matrix(self.design_matrix).T\n for x in range(bootstrap_data.shape[-1]):\n bootstrap_data[:, x] = (self.residuals.T[np.random.permutation(self\n .resampled_signal_size)] + explained_signal).squeeze()\n self.bootstrap_betas, bs_residuals, rank, s = LA.lstsq(self.\n design_matrix.T, bootstrap_data)\n self.bootstrap_betas_per_event_type = np.zeros((len(self.covariates),\n self.deconvolution_interval_size, nr_repetitions))\n for i, covariate in enumerate(list(self.covariates.keys())):\n this_covariate_index = list(self.covariates.keys()).index(covariate)\n self.bootstrap_betas_per_event_type[i] = self.bootstrap_betas[\n this_covariate_index * self.deconvolution_interval_size:(\n this_covariate_index + 1) * self.deconvolution_interval_size]\n",
"role": "user"
},
{
"content": "bootstrap_on_residuals",
"role": "assistant"
}
] |
def x(request):
url_parsed = urlparse(settings.SEARCH_URL)
defaults = dict(APP_NAME=__description__, APP_VERSION=__version__,
SITE_URL=settings.SITE_URL.rstrip('/'), SEARCH_TYPE=settings.
SEARCH_TYPE, SEARCH_URL=settings.SEARCH_URL, SEARCH_IP='%s://%s:%s' %
(url_parsed.scheme, url_parsed.hostname, url_parsed.port))
return defaults
|
resource_urls
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(request):\n url_parsed = urlparse(settings.SEARCH_URL)\n defaults = dict(APP_NAME=__description__, APP_VERSION=__version__,\n SITE_URL=settings.SITE_URL.rstrip('/'), SEARCH_TYPE=settings.\n SEARCH_TYPE, SEARCH_URL=settings.SEARCH_URL, SEARCH_IP='%s://%s:%s' %\n (url_parsed.scheme, url_parsed.hostname, url_parsed.port))\n return defaults\n",
"role": "user"
},
{
"content": "resource_urls",
"role": "assistant"
}
] |
def x(self):
from hypermap.aggregator.models import Layer
if SEARCH_TYPE == 'solr':
from hypermap.aggregator.solr import SolrHypermap
solrobject = SolrHypermap()
else:
from hypermap.aggregator.elasticsearch_client import ESHypermap
from elasticsearch import helpers
es_client = ESHypermap()
layers_cache = cache.get('layers')
deleted_layers_cache = cache.get('deleted_layers')
if layers_cache:
layers_list = list(layers_cache)
LOGGER.debug('There are %s layers in cache: %s' % (len(layers_list),
layers_list))
batch_size = settings.REGISTRY_SEARCH_BATCH_SIZE
batch_lists = [layers_list[i:i + batch_size] for i in range(0, len(
layers_list), batch_size)]
for batch_list_ids in batch_lists:
layers = Layer.objects.filter(id__in=batch_list_ids)
if batch_size > len(layers):
batch_size = len(layers)
LOGGER.debug('Syncing %s/%s layers to %s: %s' % (batch_size,
len(layers_cache), layers, SEARCH_TYPE))
try:
if SEARCH_TYPE == 'solr':
success, layers_errors_ids = solrobject.layers_to_solr(
layers)
if success:
layers_cache = layers_cache.difference(set(
batch_list_ids))
LOGGER.debug(
'Removing layers with id %s from cache' %
batch_list_ids)
cache.set('layers', layers_cache)
elif SEARCH_TYPE == 'elasticsearch':
with_bulk, success = True, False
layers_to_index = [es_client.layer_to_es(layer,
with_bulk) for layer in layers]
message = helpers.bulk(es_client.es, layers_to_index)
len_indexed_layers = message[0]
if len_indexed_layers == len(layers):
LOGGER.debug('%d layers indexed successfully' %
len_indexed_layers)
success = True
if success:
layers_cache = layers_cache.difference(set(
batch_list_ids))
cache.set('layers', layers_cache)
else:
raise Exception('Incorrect SEARCH_TYPE=%s' % SEARCH_TYPE)
except Exception as e:
LOGGER.error('Layers were NOT indexed correctly')
LOGGER.error(e, exc_info=True)
else:
LOGGER.debug('No cached layers to add in search engine.')
if deleted_layers_cache:
layers_list = list(deleted_layers_cache)
LOGGER.debug('There are %s layers in cache for deleting: %s' % (len
(layers_list), layers_list))
for layer_id in layers_list:
if SEARCH_TYPE == 'solr':
if Layer.objects.filter(pk=layer_id).exists():
layer = Layer.objects.get(id=layer_id)
unindex_layer(layer.id, use_cache=False)
deleted_layers_cache = deleted_layers_cache.difference(set
([layer_id]))
cache.set('deleted_layers', deleted_layers_cache)
else:
raise NotImplementedError
else:
LOGGER.debug('No cached layers to remove in search engine.')
|
index_cached_layers
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from hypermap.aggregator.models import Layer\n if SEARCH_TYPE == 'solr':\n from hypermap.aggregator.solr import SolrHypermap\n solrobject = SolrHypermap()\n else:\n from hypermap.aggregator.elasticsearch_client import ESHypermap\n from elasticsearch import helpers\n es_client = ESHypermap()\n layers_cache = cache.get('layers')\n deleted_layers_cache = cache.get('deleted_layers')\n if layers_cache:\n layers_list = list(layers_cache)\n LOGGER.debug('There are %s layers in cache: %s' % (len(layers_list),\n layers_list))\n batch_size = settings.REGISTRY_SEARCH_BATCH_SIZE\n batch_lists = [layers_list[i:i + batch_size] for i in range(0, len(\n layers_list), batch_size)]\n for batch_list_ids in batch_lists:\n layers = Layer.objects.filter(id__in=batch_list_ids)\n if batch_size > len(layers):\n batch_size = len(layers)\n LOGGER.debug('Syncing %s/%s layers to %s: %s' % (batch_size,\n len(layers_cache), layers, SEARCH_TYPE))\n try:\n if SEARCH_TYPE == 'solr':\n success, layers_errors_ids = solrobject.layers_to_solr(\n layers)\n if success:\n layers_cache = layers_cache.difference(set(\n batch_list_ids))\n LOGGER.debug(\n 'Removing layers with id %s from cache' %\n batch_list_ids)\n cache.set('layers', layers_cache)\n elif SEARCH_TYPE == 'elasticsearch':\n with_bulk, success = True, False\n layers_to_index = [es_client.layer_to_es(layer,\n with_bulk) for layer in layers]\n message = helpers.bulk(es_client.es, layers_to_index)\n len_indexed_layers = message[0]\n if len_indexed_layers == len(layers):\n LOGGER.debug('%d layers indexed successfully' %\n len_indexed_layers)\n success = True\n if success:\n layers_cache = layers_cache.difference(set(\n batch_list_ids))\n cache.set('layers', layers_cache)\n else:\n raise Exception('Incorrect SEARCH_TYPE=%s' % SEARCH_TYPE)\n except Exception as e:\n LOGGER.error('Layers were NOT indexed correctly')\n LOGGER.error(e, exc_info=True)\n else:\n LOGGER.debug('No cached layers to add in search engine.')\n if deleted_layers_cache:\n layers_list = list(deleted_layers_cache)\n LOGGER.debug('There are %s layers in cache for deleting: %s' % (len\n (layers_list), layers_list))\n for layer_id in layers_list:\n if SEARCH_TYPE == 'solr':\n if Layer.objects.filter(pk=layer_id).exists():\n layer = Layer.objects.get(id=layer_id)\n unindex_layer(layer.id, use_cache=False)\n deleted_layers_cache = deleted_layers_cache.difference(set\n ([layer_id]))\n cache.set('deleted_layers', deleted_layers_cache)\n else:\n raise NotImplementedError\n else:\n LOGGER.debug('No cached layers to remove in search engine.')\n",
"role": "user"
},
{
"content": "index_cached_layers",
"role": "assistant"
}
] |
def x(self, service_id):
from hypermap.aggregator.models import Service
service = Service.objects.get(id=service_id)
service.check_set.all().delete()
layer_to_process = service.layer_set.all()
for layer in layer_to_process:
layer.check_set.all().delete()
|
remove_service_checks
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, service_id):\n from hypermap.aggregator.models import Service\n service = Service.objects.get(id=service_id)\n service.check_set.all().delete()\n layer_to_process = service.layer_set.all()\n for layer in layer_to_process:\n layer.check_set.all().delete()\n",
"role": "user"
},
{
"content": "remove_service_checks",
"role": "assistant"
}
] |
def x(self, service_id):
from hypermap.aggregator.models import Service
service = Service.objects.get(id=service_id)
if not service.is_valid:
LOGGER.debug(
'Not indexing service with id %s in search engine as it is not valid'
% service.id)
return
LOGGER.debug('Indexing service %s' % service.id)
layer_to_process = service.layer_set.all()
for layer in layer_to_process:
if not settings.REGISTRY_SKIP_CELERY:
index_layer(layer.id, use_cache=True)
else:
index_layer(layer.id)
|
index_service
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, service_id):\n from hypermap.aggregator.models import Service\n service = Service.objects.get(id=service_id)\n if not service.is_valid:\n LOGGER.debug(\n 'Not indexing service with id %s in search engine as it is not valid'\n % service.id)\n return\n LOGGER.debug('Indexing service %s' % service.id)\n layer_to_process = service.layer_set.all()\n for layer in layer_to_process:\n if not settings.REGISTRY_SKIP_CELERY:\n index_layer(layer.id, use_cache=True)\n else:\n index_layer(layer.id)\n",
"role": "user"
},
{
"content": "index_service",
"role": "assistant"
}
] |
def x(self, layer_id, use_cache=False):
from hypermap.aggregator.models import Layer
layer = Layer.objects.get(id=layer_id)
if not layer.is_valid:
LOGGER.debug(
'Not indexing or removing layer with id %s in search engine as it is not valid'
% layer.id)
unindex_layer(layer.id, use_cache)
return
if layer.was_deleted:
LOGGER.debug(
'Not indexing or removing layer with id %s in search engine as was_deleted is true'
% layer.id)
unindex_layer(layer.id, use_cache)
return
if use_cache:
LOGGER.debug(
'Caching layer with id %s for syncing with search engine' %
layer.id)
layers = cache.get('layers')
if layers is None:
layers = set([layer.id])
else:
layers.add(layer.id)
cache.set('layers', layers)
return
if SEARCH_TYPE == 'solr':
from hypermap.aggregator.solr import SolrHypermap
LOGGER.debug('Syncing layer %s to solr' % layer.name)
solrobject = SolrHypermap()
success, message = solrobject.layer_to_solr(layer)
if not settings.REGISTRY_SKIP_CELERY:
if not success:
self.update_state(state=states.FAILURE, meta=message)
raise Ignore()
elif SEARCH_TYPE == 'elasticsearch':
from hypermap.aggregator.elasticsearch_client import ESHypermap
LOGGER.debug('Syncing layer %s to es' % layer.name)
esobject = ESHypermap()
success, message = esobject.layer_to_es(layer)
if not settings.REGISTRY_SKIP_CELERY:
if not success:
self.update_state(state=states.FAILURE, meta=message)
raise Ignore()
|
index_layer
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, layer_id, use_cache=False):\n from hypermap.aggregator.models import Layer\n layer = Layer.objects.get(id=layer_id)\n if not layer.is_valid:\n LOGGER.debug(\n 'Not indexing or removing layer with id %s in search engine as it is not valid'\n % layer.id)\n unindex_layer(layer.id, use_cache)\n return\n if layer.was_deleted:\n LOGGER.debug(\n 'Not indexing or removing layer with id %s in search engine as was_deleted is true'\n % layer.id)\n unindex_layer(layer.id, use_cache)\n return\n if use_cache:\n LOGGER.debug(\n 'Caching layer with id %s for syncing with search engine' %\n layer.id)\n layers = cache.get('layers')\n if layers is None:\n layers = set([layer.id])\n else:\n layers.add(layer.id)\n cache.set('layers', layers)\n return\n if SEARCH_TYPE == 'solr':\n from hypermap.aggregator.solr import SolrHypermap\n LOGGER.debug('Syncing layer %s to solr' % layer.name)\n solrobject = SolrHypermap()\n success, message = solrobject.layer_to_solr(layer)\n if not settings.REGISTRY_SKIP_CELERY:\n if not success:\n self.update_state(state=states.FAILURE, meta=message)\n raise Ignore()\n elif SEARCH_TYPE == 'elasticsearch':\n from hypermap.aggregator.elasticsearch_client import ESHypermap\n LOGGER.debug('Syncing layer %s to es' % layer.name)\n esobject = ESHypermap()\n success, message = esobject.layer_to_es(layer)\n if not settings.REGISTRY_SKIP_CELERY:\n if not success:\n self.update_state(state=states.FAILURE, meta=message)\n raise Ignore()\n",
"role": "user"
},
{
"content": "index_layer",
"role": "assistant"
}
] |
def x(self, use_cache=False):
from hypermap.aggregator.models import Issue, Layer, Service
from django.contrib.contenttypes.models import ContentType
layer_type = ContentType.objects.get_for_model(Layer)
service_type = ContentType.objects.get_for_model(Service)
for issue in Issue.objects.filter(content_type__pk=layer_type.id):
unindex_layer(issue.content_object.id, use_cache)
for issue in Issue.objects.filter(content_type__pk=service_type.id):
for layer in issue.content_object.layer_set.all():
unindex_layer(layer.id, use_cache)
|
unindex_layers_with_issues
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, use_cache=False):\n from hypermap.aggregator.models import Issue, Layer, Service\n from django.contrib.contenttypes.models import ContentType\n layer_type = ContentType.objects.get_for_model(Layer)\n service_type = ContentType.objects.get_for_model(Service)\n for issue in Issue.objects.filter(content_type__pk=layer_type.id):\n unindex_layer(issue.content_object.id, use_cache)\n for issue in Issue.objects.filter(content_type__pk=service_type.id):\n for layer in issue.content_object.layer_set.all():\n unindex_layer(layer.id, use_cache)\n",
"role": "user"
},
{
"content": "unindex_layers_with_issues",
"role": "assistant"
}
] |
def x(self, layer_id, use_cache=False):
from hypermap.aggregator.models import Layer
layer = Layer.objects.get(id=layer_id)
if use_cache:
LOGGER.debug(
'Caching layer with id %s for being removed from search engine' %
layer.id)
deleted_layers = cache.get('deleted_layers')
if deleted_layers is None:
deleted_layers = set([layer.id])
else:
deleted_layers.add(layer.id)
cache.set('deleted_layers', deleted_layers)
return
if SEARCH_TYPE == 'solr':
from hypermap.aggregator.solr import SolrHypermap
LOGGER.debug('Removing layer %s from solr' % layer.id)
try:
solrobject = SolrHypermap()
solrobject.remove_layer(layer.uuid)
except Exception:
LOGGER.error('Layer NOT correctly removed from Solr')
elif SEARCH_TYPE == 'elasticsearch':
pass
|
unindex_layer
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, layer_id, use_cache=False):\n from hypermap.aggregator.models import Layer\n layer = Layer.objects.get(id=layer_id)\n if use_cache:\n LOGGER.debug(\n 'Caching layer with id %s for being removed from search engine' %\n layer.id)\n deleted_layers = cache.get('deleted_layers')\n if deleted_layers is None:\n deleted_layers = set([layer.id])\n else:\n deleted_layers.add(layer.id)\n cache.set('deleted_layers', deleted_layers)\n return\n if SEARCH_TYPE == 'solr':\n from hypermap.aggregator.solr import SolrHypermap\n LOGGER.debug('Removing layer %s from solr' % layer.id)\n try:\n solrobject = SolrHypermap()\n solrobject.remove_layer(layer.uuid)\n except Exception:\n LOGGER.error('Layer NOT correctly removed from Solr')\n elif SEARCH_TYPE == 'elasticsearch':\n pass\n",
"role": "user"
},
{
"content": "unindex_layer",
"role": "assistant"
}
] |
def x(self):
from hypermap.aggregator.models import Layer
if not settings.REGISTRY_SKIP_CELERY:
layers_cache = set(Layer.objects.filter(is_valid=True).values_list(
'id', flat=True))
deleted_layers_cache = set(Layer.objects.filter(is_valid=False).
values_list('id', flat=True))
cache.set('layers', layers_cache)
cache.set('deleted_layers', deleted_layers_cache)
else:
for layer in Layer.objects.all():
index_layer(layer.id)
|
index_all_layers
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from hypermap.aggregator.models import Layer\n if not settings.REGISTRY_SKIP_CELERY:\n layers_cache = set(Layer.objects.filter(is_valid=True).values_list(\n 'id', flat=True))\n deleted_layers_cache = set(Layer.objects.filter(is_valid=False).\n values_list('id', flat=True))\n cache.set('layers', layers_cache)\n cache.set('deleted_layers', deleted_layers_cache)\n else:\n for layer in Layer.objects.all():\n index_layer(layer.id)\n",
"role": "user"
},
{
"content": "index_all_layers",
"role": "assistant"
}
] |
def x(self, service_id, num_layers=10):
from hypermap.aggregator.models import Service
LOGGER.debug(
'Updating the index the last %s added and %s deleted layers in WorldMap service'
% (num_layers, num_layers))
service = Service.objects.get(id=service_id)
if service.type == 'Hypermap:WorldMapLegacy':
from hypermap.aggregator.models import update_layers_wm_legacy as update_layers_wm
if service.type == 'Hypermap:WorldMap':
from hypermap.aggregator.models import update_layers_geonode_wm as update_layers_wm
update_layers_wm(service, num_layers)
LOGGER.debug('Removing the index for the last %s deleted layers' %
num_layers)
layer_to_unindex = service.layer_set.filter(was_deleted=True).order_by(
'-last_updated')[0:num_layers]
for layer in layer_to_unindex:
if not settings.REGISTRY_SKIP_CELERY:
unindex_layer(layer.id, use_cache=True)
else:
unindex_layer(layer.id)
LOGGER.debug('Adding/Updating the index for the last %s added layers' %
num_layers)
layer_to_index = service.layer_set.filter(was_deleted=False).order_by(
'-last_updated')[0:num_layers]
for layer in layer_to_index:
if not settings.REGISTRY_SKIP_CELERY:
index_layer(layer.id, use_cache=True)
else:
index_layer(layer.id)
|
update_last_wm_layers
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, service_id, num_layers=10):\n from hypermap.aggregator.models import Service\n LOGGER.debug(\n 'Updating the index the last %s added and %s deleted layers in WorldMap service'\n % (num_layers, num_layers))\n service = Service.objects.get(id=service_id)\n if service.type == 'Hypermap:WorldMapLegacy':\n from hypermap.aggregator.models import update_layers_wm_legacy as update_layers_wm\n if service.type == 'Hypermap:WorldMap':\n from hypermap.aggregator.models import update_layers_geonode_wm as update_layers_wm\n update_layers_wm(service, num_layers)\n LOGGER.debug('Removing the index for the last %s deleted layers' %\n num_layers)\n layer_to_unindex = service.layer_set.filter(was_deleted=True).order_by(\n '-last_updated')[0:num_layers]\n for layer in layer_to_unindex:\n if not settings.REGISTRY_SKIP_CELERY:\n unindex_layer(layer.id, use_cache=True)\n else:\n unindex_layer(layer.id)\n LOGGER.debug('Adding/Updating the index for the last %s added layers' %\n num_layers)\n layer_to_index = service.layer_set.filter(was_deleted=False).order_by(\n '-last_updated')[0:num_layers]\n for layer in layer_to_index:\n if not settings.REGISTRY_SKIP_CELERY:\n index_layer(layer.id, use_cache=True)\n else:\n index_layer(layer.id)\n",
"role": "user"
},
{
"content": "update_last_wm_layers",
"role": "assistant"
}
] |
def x(bbox):
try:
minx = float(bbox[0])
miny = float(bbox[1])
maxx = float(bbox[2])
maxy = float(bbox[3])
except:
LOGGER.debug('Invalid bbox, setting it to a zero POLYGON')
minx = 0
miny = 0
maxx = 0
maxy = 0
return (
'POLYGON((%.2f %.2f, %.2f %.2f, %.2f %.2f, %.2f %.2f, %.2f %.2f))' %
(minx, miny, minx, maxy, maxx, maxy, maxx, miny, minx, miny))
|
bbox2wktpolygon
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(bbox):\n try:\n minx = float(bbox[0])\n miny = float(bbox[1])\n maxx = float(bbox[2])\n maxy = float(bbox[3])\n except:\n LOGGER.debug('Invalid bbox, setting it to a zero POLYGON')\n minx = 0\n miny = 0\n maxx = 0\n maxy = 0\n return (\n 'POLYGON((%.2f %.2f, %.2f %.2f, %.2f %.2f, %.2f %.2f, %.2f %.2f))' %\n (minx, miny, minx, maxy, maxx, maxy, maxx, miny, minx, miny))\n",
"role": "user"
},
{
"content": "bbox2wktpolygon",
"role": "assistant"
}
] |
def x(**kwargs):
if 'srs' in kwargs:
srs = kwargs['srs']
else:
srs = '4326'
modified = '%sZ' % datetime.datetime.utcnow().isoformat().split('.')[0]
nsmap = Namespaces().get_namespaces(['csw', 'dc', 'dct', 'ows'])
e = etree.Element(nspath_eval('csw:Record', nsmap), nsmap=nsmap)
etree.SubElement(e, nspath_eval('dc:identifier', nsmap)).text = kwargs[
'identifier']
etree.SubElement(e, nspath_eval('dc:title', nsmap)).text = kwargs['title']
if 'alternative' in kwargs:
etree.SubElement(e, nspath_eval('dct:alternative', nsmap)
).text = kwargs['alternative']
etree.SubElement(e, nspath_eval('dct:modified', nsmap)).text = modified
etree.SubElement(e, nspath_eval('dct:abstract', nsmap)).text = kwargs[
'abstract']
etree.SubElement(e, nspath_eval('dc:type', nsmap)).text = kwargs['type']
etree.SubElement(e, nspath_eval('dc:format', nsmap)).text = kwargs['format'
]
etree.SubElement(e, nspath_eval('dc:source', nsmap)).text = kwargs['source'
]
if 'relation' in kwargs:
etree.SubElement(e, nspath_eval('dc:relation', nsmap)).text = kwargs[
'relation']
if 'keywords' in kwargs:
if kwargs['keywords'] is not None:
for keyword in kwargs['keywords']:
etree.SubElement(e, nspath_eval('dc:subject', nsmap)
).text = keyword
for link in kwargs['links']:
etree.SubElement(e, nspath_eval('dct:references', nsmap), scheme=
link[0]).text = link[1]
bbox2 = loads(kwargs['wkt_geometry']).bounds
bbox = etree.SubElement(e, nspath_eval('ows:BoundingBox', nsmap), crs=
'http://www.opengis.net/def/crs/EPSG/0/%s' % srs, dimensions='2')
etree.SubElement(bbox, nspath_eval('ows:LowerCorner', nsmap)
).text = '%s %s' % (bbox2[1], bbox2[0])
etree.SubElement(bbox, nspath_eval('ows:UpperCorner', nsmap)
).text = '%s %s' % (bbox2[3], bbox2[2])
return etree.tostring(e, pretty_print=True)
|
create_metadata_record
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(**kwargs):\n if 'srs' in kwargs:\n srs = kwargs['srs']\n else:\n srs = '4326'\n modified = '%sZ' % datetime.datetime.utcnow().isoformat().split('.')[0]\n nsmap = Namespaces().get_namespaces(['csw', 'dc', 'dct', 'ows'])\n e = etree.Element(nspath_eval('csw:Record', nsmap), nsmap=nsmap)\n etree.SubElement(e, nspath_eval('dc:identifier', nsmap)).text = kwargs[\n 'identifier']\n etree.SubElement(e, nspath_eval('dc:title', nsmap)).text = kwargs['title']\n if 'alternative' in kwargs:\n etree.SubElement(e, nspath_eval('dct:alternative', nsmap)\n ).text = kwargs['alternative']\n etree.SubElement(e, nspath_eval('dct:modified', nsmap)).text = modified\n etree.SubElement(e, nspath_eval('dct:abstract', nsmap)).text = kwargs[\n 'abstract']\n etree.SubElement(e, nspath_eval('dc:type', nsmap)).text = kwargs['type']\n etree.SubElement(e, nspath_eval('dc:format', nsmap)).text = kwargs['format'\n ]\n etree.SubElement(e, nspath_eval('dc:source', nsmap)).text = kwargs['source'\n ]\n if 'relation' in kwargs:\n etree.SubElement(e, nspath_eval('dc:relation', nsmap)).text = kwargs[\n 'relation']\n if 'keywords' in kwargs:\n if kwargs['keywords'] is not None:\n for keyword in kwargs['keywords']:\n etree.SubElement(e, nspath_eval('dc:subject', nsmap)\n ).text = keyword\n for link in kwargs['links']:\n etree.SubElement(e, nspath_eval('dct:references', nsmap), scheme=\n link[0]).text = link[1]\n bbox2 = loads(kwargs['wkt_geometry']).bounds\n bbox = etree.SubElement(e, nspath_eval('ows:BoundingBox', nsmap), crs=\n 'http://www.opengis.net/def/crs/EPSG/0/%s' % srs, dimensions='2')\n etree.SubElement(bbox, nspath_eval('ows:LowerCorner', nsmap)\n ).text = '%s %s' % (bbox2[1], bbox2[0])\n etree.SubElement(bbox, nspath_eval('ows:UpperCorner', nsmap)\n ).text = '%s %s' % (bbox2[3], bbox2[2])\n return etree.tostring(e, pretty_print=True)\n",
"role": "user"
},
{
"content": "create_metadata_record",
"role": "assistant"
}
] |
def x(*args):
bag = []
for term in args:
if term is not None:
if isinstance(term, list):
for term2 in term:
if term2 is not None:
bag.append(term2)
else:
bag.append(term)
return ' '.join(bag)
|
gen_anytext
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(*args):\n bag = []\n for term in args:\n if term is not None:\n if isinstance(term, list):\n for term2 in term:\n if term2 is not None:\n bag.append(term2)\n else:\n bag.append(term)\n return ' '.join(bag)\n",
"role": "user"
},
{
"content": "gen_anytext",
"role": "assistant"
}
] |
def x(service):
try:
wmts = WebMapTileService(service.url)
srs, created = SpatialReferenceSystem.objects.get_or_create(code=
'EPSG:4326')
service.srs.add(srs)
service.update_validity()
layer_names = list(wmts.contents)
layer_n = 0
total = len(layer_names)
for layer_name in layer_names:
ows_layer = wmts.contents[layer_name]
LOGGER.debug('Updating layer %s' % ows_layer.name)
layer, created = Layer.objects.get_or_create(name=ows_layer.
name, service=service, catalog=service.catalog)
if layer.active:
links = [['OGC:WMTS', service.url], ['OGC:WMTS', settings.
SITE_URL.rstrip('/') + '/' + layer.get_url_endpoint()]]
layer.type = 'OGC:WMTS'
layer.title = ows_layer.title
layer.abstract = ows_layer.abstract
keywords = None
if hasattr(ows_layer, 'keywords'):
keywords = ows_layer.keywords
for keyword in keywords:
layer.keywords.add(keyword)
layer.url = service.url
layer.page_url = layer.get_absolute_url
links.append(['WWW:LINK', settings.SITE_URL.rstrip('/') +
layer.page_url])
bbox = list(ows_layer.boundingBoxWGS84 or (-179.0, -89.0,
179.0, 89.0))
layer.bbox_x0 = bbox[0]
layer.bbox_y0 = bbox[1]
layer.bbox_x1 = bbox[2]
layer.bbox_y1 = bbox[3]
layer.wkt_geometry = bbox2wktpolygon(bbox)
layer.xml = create_metadata_record(identifier=str(layer.
uuid), source=service.url, links=links, format=
'OGC:WMS', type=layer.csw_type, relation=service.
id_string, title=ows_layer.title, alternative=ows_layer
.name, abstract=layer.abstract, keywords=keywords,
wkt_geometry=layer.wkt_geometry)
layer.anytext = gen_anytext(layer.title, layer.abstract,
keywords)
layer.save()
add_mined_dates(layer)
layer_n = layer_n + 1
LOGGER.debug('Updating layer n. %s/%s' % (layer_n, total))
if DEBUG_SERVICES and layer_n == DEBUG_LAYER_NUMBER:
return
except Exception as err:
message = 'update_layers_wmts: {0}'.format(err)
check = Check(content_object=service, success=False, response_time=
0, message=message)
check.save()
|
update_layers_wmts
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(service):\n try:\n wmts = WebMapTileService(service.url)\n srs, created = SpatialReferenceSystem.objects.get_or_create(code=\n 'EPSG:4326')\n service.srs.add(srs)\n service.update_validity()\n layer_names = list(wmts.contents)\n layer_n = 0\n total = len(layer_names)\n for layer_name in layer_names:\n ows_layer = wmts.contents[layer_name]\n LOGGER.debug('Updating layer %s' % ows_layer.name)\n layer, created = Layer.objects.get_or_create(name=ows_layer.\n name, service=service, catalog=service.catalog)\n if layer.active:\n links = [['OGC:WMTS', service.url], ['OGC:WMTS', settings.\n SITE_URL.rstrip('/') + '/' + layer.get_url_endpoint()]]\n layer.type = 'OGC:WMTS'\n layer.title = ows_layer.title\n layer.abstract = ows_layer.abstract\n keywords = None\n if hasattr(ows_layer, 'keywords'):\n keywords = ows_layer.keywords\n for keyword in keywords:\n layer.keywords.add(keyword)\n layer.url = service.url\n layer.page_url = layer.get_absolute_url\n links.append(['WWW:LINK', settings.SITE_URL.rstrip('/') +\n layer.page_url])\n bbox = list(ows_layer.boundingBoxWGS84 or (-179.0, -89.0, \n 179.0, 89.0))\n layer.bbox_x0 = bbox[0]\n layer.bbox_y0 = bbox[1]\n layer.bbox_x1 = bbox[2]\n layer.bbox_y1 = bbox[3]\n layer.wkt_geometry = bbox2wktpolygon(bbox)\n layer.xml = create_metadata_record(identifier=str(layer.\n uuid), source=service.url, links=links, format=\n 'OGC:WMS', type=layer.csw_type, relation=service.\n id_string, title=ows_layer.title, alternative=ows_layer\n .name, abstract=layer.abstract, keywords=keywords,\n wkt_geometry=layer.wkt_geometry)\n layer.anytext = gen_anytext(layer.title, layer.abstract,\n keywords)\n layer.save()\n add_mined_dates(layer)\n layer_n = layer_n + 1\n LOGGER.debug('Updating layer n. %s/%s' % (layer_n, total))\n if DEBUG_SERVICES and layer_n == DEBUG_LAYER_NUMBER:\n return\n except Exception as err:\n message = 'update_layers_wmts: {0}'.format(err)\n check = Check(content_object=service, success=False, response_time=\n 0, message=message)\n check.save()\n",
"role": "user"
},
{
"content": "update_layers_wmts",
"role": "assistant"
}
] |
def x(service):
params = {'field': 'title', 'query': '', 'show_warped': '1', 'format':
'json'}
headers = {'Content-Type': 'application/json', 'Accept': 'application/json'
}
request = requests.get(service.url, headers=headers, params=params)
try:
records = json.loads(request.content)
total_pages = int(records['total_pages'])
for crs_code in ['EPSG:4326', 'EPSG:900913', 'EPSG:3857']:
srs, created = SpatialReferenceSystem.objects.get_or_create(code
=crs_code)
service.srs.add(srs)
service.update_validity()
for i in range(1, total_pages + 1):
params = {'field': 'title', 'query': '', 'show_warped': '1',
'format': 'json', 'page': i}
request = requests.get(service.url, headers=headers, params=params)
records = json.loads(request.content)
LOGGER.debug('Fetched %s' % request.url)
layers = records['items']
layer_n = 0
total = len(layers)
for layer in layers:
name = layer['id']
title = layer['title']
abstract = layer['description']
bbox = layer['bbox']
dates = []
if 'published_date' in layer:
dates.append(layer['published_date'])
if 'date_depicted' in layer:
dates.append(layer['date_depicted'])
if 'depicts_year' in layer:
dates.append(layer['depicts_year'])
if 'issue_year' in layer:
dates.append(layer['issue_year'])
layer, created = Layer.objects.get_or_create(name=name,
service=service, catalog=service.catalog)
if layer.active:
layer.type = 'Hypermap:WARPER'
layer.title = title
layer.abstract = abstract
layer.is_public = True
layer.url = '%s/wms/%s?' % (service.url, name)
layer.page_url = '%s/%s' % (service.url, name)
x0 = None
y0 = None
x1 = None
y1 = None
if bbox:
bbox_list = bbox.split(',')
x0 = format_float(bbox_list[0])
y0 = format_float(bbox_list[1])
x1 = format_float(bbox_list[2])
y1 = format_float(bbox_list[3])
layer.bbox_x0 = x0
layer.bbox_y0 = y0
layer.bbox_x1 = x1
layer.bbox_y1 = y1
layer.save()
add_mined_dates(layer)
add_metadata_dates_to_layer(dates, layer)
layer_n = layer_n + 1
LOGGER.debug('Updating layer n. %s/%s' % (layer_n, total))
if DEBUG_SERVICES and layer_n == DEBUG_LAYER_NUMBER:
return
except Exception as err:
message = 'update_layers_warper: {0}. request={1} response={2}'.format(
err, service.url, request.text)
check = Check(content_object=service, success=False, response_time=
0, message=message)
check.save()
|
update_layers_warper
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(service):\n params = {'field': 'title', 'query': '', 'show_warped': '1', 'format':\n 'json'}\n headers = {'Content-Type': 'application/json', 'Accept': 'application/json'\n }\n request = requests.get(service.url, headers=headers, params=params)\n try:\n records = json.loads(request.content)\n total_pages = int(records['total_pages'])\n for crs_code in ['EPSG:4326', 'EPSG:900913', 'EPSG:3857']:\n srs, created = SpatialReferenceSystem.objects.get_or_create(code\n =crs_code)\n service.srs.add(srs)\n service.update_validity()\n for i in range(1, total_pages + 1):\n params = {'field': 'title', 'query': '', 'show_warped': '1',\n 'format': 'json', 'page': i}\n request = requests.get(service.url, headers=headers, params=params)\n records = json.loads(request.content)\n LOGGER.debug('Fetched %s' % request.url)\n layers = records['items']\n layer_n = 0\n total = len(layers)\n for layer in layers:\n name = layer['id']\n title = layer['title']\n abstract = layer['description']\n bbox = layer['bbox']\n dates = []\n if 'published_date' in layer:\n dates.append(layer['published_date'])\n if 'date_depicted' in layer:\n dates.append(layer['date_depicted'])\n if 'depicts_year' in layer:\n dates.append(layer['depicts_year'])\n if 'issue_year' in layer:\n dates.append(layer['issue_year'])\n layer, created = Layer.objects.get_or_create(name=name,\n service=service, catalog=service.catalog)\n if layer.active:\n layer.type = 'Hypermap:WARPER'\n layer.title = title\n layer.abstract = abstract\n layer.is_public = True\n layer.url = '%s/wms/%s?' % (service.url, name)\n layer.page_url = '%s/%s' % (service.url, name)\n x0 = None\n y0 = None\n x1 = None\n y1 = None\n if bbox:\n bbox_list = bbox.split(',')\n x0 = format_float(bbox_list[0])\n y0 = format_float(bbox_list[1])\n x1 = format_float(bbox_list[2])\n y1 = format_float(bbox_list[3])\n layer.bbox_x0 = x0\n layer.bbox_y0 = y0\n layer.bbox_x1 = x1\n layer.bbox_y1 = y1\n layer.save()\n add_mined_dates(layer)\n add_metadata_dates_to_layer(dates, layer)\n layer_n = layer_n + 1\n LOGGER.debug('Updating layer n. %s/%s' % (layer_n, total))\n if DEBUG_SERVICES and layer_n == DEBUG_LAYER_NUMBER:\n return\n except Exception as err:\n message = 'update_layers_warper: {0}. request={1} response={2}'.format(\n err, service.url, request.text)\n check = Check(content_object=service, success=False, response_time=\n 0, message=message)\n check.save()\n",
"role": "user"
},
{
"content": "update_layers_warper",
"role": "assistant"
}
] |
def x(service, greedy_opt=False):
try:
esri_service = ArcMapService(service.url)
srs_code = esri_service.spatialReference.wkid
srs, created = SpatialReferenceSystem.objects.get_or_create(code=
srs_code)
service.srs.add(srs)
service.update_validity()
if 'supportedExtensions' in esri_service._json_struct and greedy_opt:
if 'WMSServer' in esri_service._json_struct['supportedExtensions']:
wms_url = service.url.replace('/rest/services/', '/services/')
if '?f=pjson' in wms_url:
wms_url = wms_url.replace('?f=pjson', 'WMSServer?')
if '?f=json' in wms_url:
wms_url = wms_url.replace('?f=json', 'WMSServer?')
LOGGER.debug(
'This ESRI REST endpoint has an WMS interface to process: %s'
% wms_url)
from utils import create_service_from_endpoint
create_service_from_endpoint(wms_url, 'OGC:WMS', catalog=
service.catalog)
layer_n = 0
total = len(esri_service.layers)
for esri_layer in esri_service.layers:
if 'error' not in esri_layer._json_struct:
LOGGER.debug('Updating layer %s' % esri_layer.name)
layer, created = Layer.objects.get_or_create(name=
esri_layer.id, service=service, catalog=service.catalog)
if layer.active:
layer.type = 'ESRI:ArcGIS:MapServer'
links = [[layer.type, service.url], ['OGC:WMTS',
settings.SITE_URL.rstrip('/') + '/' + layer.
get_url_endpoint()]]
layer.title = esri_layer.name
layer.abstract = esri_service.serviceDescription
layer.url = service.url
layer.page_url = layer.get_absolute_url
links.append(['WWW:LINK', settings.SITE_URL.rstrip('/') +
layer.page_url])
try:
layer.bbox_x0 = esri_layer.extent.xmin
layer.bbox_y0 = esri_layer.extent.ymin
layer.bbox_x1 = esri_layer.extent.xmax
layer.bbox_y1 = esri_layer.extent.ymax
except KeyError:
pass
try:
layer.bbox_x0 = esri_layer._json_struct['extent'][
'xmin']
layer.bbox_y0 = esri_layer._json_struct['extent'][
'ymin']
layer.bbox_x1 = esri_layer._json_struct['extent'][
'xmax']
layer.bbox_y1 = esri_layer._json_struct['extent'][
'ymax']
except Exception:
pass
layer.wkt_geometry = bbox2wktpolygon([layer.bbox_x0,
layer.bbox_y0, layer.bbox_x1, layer.bbox_y1])
layer.xml = create_metadata_record(identifier=str(layer
.uuid), source=service.url, links=links, format=
'ESRI:ArcGIS:MapServer', type=layer.csw_type,
relation=service.id_string, title=layer.title,
alternative=layer.title, abstract=layer.abstract,
wkt_geometry=layer.wkt_geometry)
layer.anytext = gen_anytext(layer.title, layer.abstract)
layer.save()
add_mined_dates(layer)
layer_n = layer_n + 1
LOGGER.debug('Updating layer n. %s/%s' % (layer_n, total))
if DEBUG_SERVICES and layer_n == DEBUG_LAYER_NUMBER:
return
except Exception as err:
message = 'update_layers_esri_mapserver: {0}'.format(err)
check = Check(content_object=service, success=False, response_time=
0, message=message)
check.save()
|
update_layers_esri_mapserver
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(service, greedy_opt=False):\n try:\n esri_service = ArcMapService(service.url)\n srs_code = esri_service.spatialReference.wkid\n srs, created = SpatialReferenceSystem.objects.get_or_create(code=\n srs_code)\n service.srs.add(srs)\n service.update_validity()\n if 'supportedExtensions' in esri_service._json_struct and greedy_opt:\n if 'WMSServer' in esri_service._json_struct['supportedExtensions']:\n wms_url = service.url.replace('/rest/services/', '/services/')\n if '?f=pjson' in wms_url:\n wms_url = wms_url.replace('?f=pjson', 'WMSServer?')\n if '?f=json' in wms_url:\n wms_url = wms_url.replace('?f=json', 'WMSServer?')\n LOGGER.debug(\n 'This ESRI REST endpoint has an WMS interface to process: %s'\n % wms_url)\n from utils import create_service_from_endpoint\n create_service_from_endpoint(wms_url, 'OGC:WMS', catalog=\n service.catalog)\n layer_n = 0\n total = len(esri_service.layers)\n for esri_layer in esri_service.layers:\n if 'error' not in esri_layer._json_struct:\n LOGGER.debug('Updating layer %s' % esri_layer.name)\n layer, created = Layer.objects.get_or_create(name=\n esri_layer.id, service=service, catalog=service.catalog)\n if layer.active:\n layer.type = 'ESRI:ArcGIS:MapServer'\n links = [[layer.type, service.url], ['OGC:WMTS', \n settings.SITE_URL.rstrip('/') + '/' + layer.\n get_url_endpoint()]]\n layer.title = esri_layer.name\n layer.abstract = esri_service.serviceDescription\n layer.url = service.url\n layer.page_url = layer.get_absolute_url\n links.append(['WWW:LINK', settings.SITE_URL.rstrip('/') +\n layer.page_url])\n try:\n layer.bbox_x0 = esri_layer.extent.xmin\n layer.bbox_y0 = esri_layer.extent.ymin\n layer.bbox_x1 = esri_layer.extent.xmax\n layer.bbox_y1 = esri_layer.extent.ymax\n except KeyError:\n pass\n try:\n layer.bbox_x0 = esri_layer._json_struct['extent'][\n 'xmin']\n layer.bbox_y0 = esri_layer._json_struct['extent'][\n 'ymin']\n layer.bbox_x1 = esri_layer._json_struct['extent'][\n 'xmax']\n layer.bbox_y1 = esri_layer._json_struct['extent'][\n 'ymax']\n except Exception:\n pass\n layer.wkt_geometry = bbox2wktpolygon([layer.bbox_x0,\n layer.bbox_y0, layer.bbox_x1, layer.bbox_y1])\n layer.xml = create_metadata_record(identifier=str(layer\n .uuid), source=service.url, links=links, format=\n 'ESRI:ArcGIS:MapServer', type=layer.csw_type,\n relation=service.id_string, title=layer.title,\n alternative=layer.title, abstract=layer.abstract,\n wkt_geometry=layer.wkt_geometry)\n layer.anytext = gen_anytext(layer.title, layer.abstract)\n layer.save()\n add_mined_dates(layer)\n layer_n = layer_n + 1\n LOGGER.debug('Updating layer n. %s/%s' % (layer_n, total))\n if DEBUG_SERVICES and layer_n == DEBUG_LAYER_NUMBER:\n return\n except Exception as err:\n message = 'update_layers_esri_mapserver: {0}'.format(err)\n check = Check(content_object=service, success=False, response_time=\n 0, message=message)\n check.save()\n",
"role": "user"
},
{
"content": "update_layers_esri_mapserver",
"role": "assistant"
}
] |
def x(service):
try:
esri_service = ArcImageService(service.url)
obj = json.loads(esri_service._contents)
srs_code = obj['spatialReference']['wkid']
srs, created = SpatialReferenceSystem.objects.get_or_create(code=
srs_code)
service.srs.add(srs)
service.update_validity()
layer, created = Layer.objects.get_or_create(name=obj['name'],
service=service, catalog=service.catalog)
if layer.active:
layer.type = 'ESRI:ArcGIS:ImageServer'
links = [[layer.type, service.url], ['OGC:WMTS', settings.
SITE_URL.rstrip('/') + '/' + layer.get_url_endpoint()]]
layer.title = obj['name']
layer.abstract = esri_service.serviceDescription
layer.url = service.url
layer.bbox_x0 = str(obj['extent']['xmin'])
layer.bbox_y0 = str(obj['extent']['ymin'])
layer.bbox_x1 = str(obj['extent']['xmax'])
layer.bbox_y1 = str(obj['extent']['ymax'])
layer.page_url = layer.get_absolute_url
links.append(['WWW:LINK', settings.SITE_URL.rstrip('/') + layer
.page_url])
layer.wkt_geometry = bbox2wktpolygon([layer.bbox_x0, layer.
bbox_y0, layer.bbox_x1, layer.bbox_y1])
layer.xml = create_metadata_record(identifier=str(layer.uuid),
source=service.url, links=links, format=
'ESRI:ArcGIS:ImageServer', type=layer.csw_type, relation=
service.id_string, title=layer.title, alternative=layer.
title, abstract=layer.abstract, wkt_geometry=layer.wkt_geometry
)
layer.anytext = gen_anytext(layer.title, layer.abstract)
layer.save()
add_mined_dates(layer)
except Exception as err:
message = 'update_layers_esri_imageserver: {0}'.format(err)
check = Check(content_object=service, success=False, response_time=
0, message=message)
check.save()
|
update_layers_esri_imageserver
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(service):\n try:\n esri_service = ArcImageService(service.url)\n obj = json.loads(esri_service._contents)\n srs_code = obj['spatialReference']['wkid']\n srs, created = SpatialReferenceSystem.objects.get_or_create(code=\n srs_code)\n service.srs.add(srs)\n service.update_validity()\n layer, created = Layer.objects.get_or_create(name=obj['name'],\n service=service, catalog=service.catalog)\n if layer.active:\n layer.type = 'ESRI:ArcGIS:ImageServer'\n links = [[layer.type, service.url], ['OGC:WMTS', settings.\n SITE_URL.rstrip('/') + '/' + layer.get_url_endpoint()]]\n layer.title = obj['name']\n layer.abstract = esri_service.serviceDescription\n layer.url = service.url\n layer.bbox_x0 = str(obj['extent']['xmin'])\n layer.bbox_y0 = str(obj['extent']['ymin'])\n layer.bbox_x1 = str(obj['extent']['xmax'])\n layer.bbox_y1 = str(obj['extent']['ymax'])\n layer.page_url = layer.get_absolute_url\n links.append(['WWW:LINK', settings.SITE_URL.rstrip('/') + layer\n .page_url])\n layer.wkt_geometry = bbox2wktpolygon([layer.bbox_x0, layer.\n bbox_y0, layer.bbox_x1, layer.bbox_y1])\n layer.xml = create_metadata_record(identifier=str(layer.uuid),\n source=service.url, links=links, format=\n 'ESRI:ArcGIS:ImageServer', type=layer.csw_type, relation=\n service.id_string, title=layer.title, alternative=layer.\n title, abstract=layer.abstract, wkt_geometry=layer.wkt_geometry\n )\n layer.anytext = gen_anytext(layer.title, layer.abstract)\n layer.save()\n add_mined_dates(layer)\n except Exception as err:\n message = 'update_layers_esri_imageserver: {0}'.format(err)\n check = Check(content_object=service, success=False, response_time=\n 0, message=message)\n check.save()\n",
"role": "user"
},
{
"content": "update_layers_esri_imageserver",
"role": "assistant"
}
] |
def x(instance, *args, **kwargs):
with open(instance.upload.file.name, mode='rb') as f:
lines = f.readlines()
for url in lines:
if len(url) > 255:
LOGGER.debug(
'Skipping this endpoint, as it is more than 255 characters: %s'
% url)
elif Endpoint.objects.filter(url=url, catalog=instance.catalog).count(
) == 0:
endpoint = Endpoint(url=url, endpoint_list=instance)
endpoint.catalog = instance.catalog
endpoint.save()
if not settings.REGISTRY_SKIP_CELERY:
update_endpoints.delay(instance.id)
else:
update_endpoints(instance.id)
|
endpointlist_post_save
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(instance, *args, **kwargs):\n with open(instance.upload.file.name, mode='rb') as f:\n lines = f.readlines()\n for url in lines:\n if len(url) > 255:\n LOGGER.debug(\n 'Skipping this endpoint, as it is more than 255 characters: %s'\n % url)\n elif Endpoint.objects.filter(url=url, catalog=instance.catalog).count(\n ) == 0:\n endpoint = Endpoint(url=url, endpoint_list=instance)\n endpoint.catalog = instance.catalog\n endpoint.save()\n if not settings.REGISTRY_SKIP_CELERY:\n update_endpoints.delay(instance.id)\n else:\n update_endpoints(instance.id)\n",
"role": "user"
},
{
"content": "endpointlist_post_save",
"role": "assistant"
}
] |
def x(instance, *args, **kwargs):
exists = Service.objects.filter(url=instance.url, type=instance.type,
catalog=instance.catalog).count() > 0
if instance._state.adding and exists:
raise Exception('There is already such a service. url={0} catalog={1}'
.format(instance.url, instance.catalog))
|
service_pre_save
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(instance, *args, **kwargs):\n exists = Service.objects.filter(url=instance.url, type=instance.type,\n catalog=instance.catalog).count() > 0\n if instance._state.adding and exists:\n raise Exception('There is already such a service. url={0} catalog={1}'\n .format(instance.url, instance.catalog))\n",
"role": "user"
},
{
"content": "service_pre_save",
"role": "assistant"
}
] |
def x(instance, *args, **kwargs):
if instance.is_monitored and settings.REGISTRY_SKIP_CELERY:
check_service(instance.id)
elif instance.is_monitored:
check_service.delay(instance.id)
|
service_post_save
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(instance, *args, **kwargs):\n if instance.is_monitored and settings.REGISTRY_SKIP_CELERY:\n check_service(instance.id)\n elif instance.is_monitored:\n check_service.delay(instance.id)\n",
"role": "user"
},
{
"content": "service_post_save",
"role": "assistant"
}
] |
def x(instance, *args, **kwargs):
is_valid = True
if not instance.service.type == 'Hypermap:WorldMap':
if not instance.service.is_valid:
is_valid = False
LOGGER.debug(
'Layer with id %s is marked invalid because its service is invalid'
% instance.id)
if (instance.bbox_x0 > -2 and instance.bbox_x1 < 2 and instance.
bbox_y0 > -2 and instance.bbox_y1 < 2):
is_valid = False
LOGGER.debug(
'Layer with id %s is marked invalid because its extent is within (-2, -2, +2, +2)'
% instance.id)
instance.is_valid = is_valid
|
layer_pre_save
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(instance, *args, **kwargs):\n is_valid = True\n if not instance.service.type == 'Hypermap:WorldMap':\n if not instance.service.is_valid:\n is_valid = False\n LOGGER.debug(\n 'Layer with id %s is marked invalid because its service is invalid'\n % instance.id)\n if (instance.bbox_x0 > -2 and instance.bbox_x1 < 2 and instance.\n bbox_y0 > -2 and instance.bbox_y1 < 2):\n is_valid = False\n LOGGER.debug(\n 'Layer with id %s is marked invalid because its extent is within (-2, -2, +2, +2)'\n % instance.id)\n instance.is_valid = is_valid\n",
"role": "user"
},
{
"content": "layer_pre_save",
"role": "assistant"
}
] |
def x(instance, *args, **kwargs):
if instance.is_monitored and instance.service.is_monitored:
if not settings.REGISTRY_SKIP_CELERY:
check_layer.delay(instance.id)
else:
check_layer(instance.id)
else:
index_layer(instance.id)
|
layer_post_save
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(instance, *args, **kwargs):\n if instance.is_monitored and instance.service.is_monitored:\n if not settings.REGISTRY_SKIP_CELERY:\n check_layer.delay(instance.id)\n else:\n check_layer(instance.id)\n else:\n index_layer(instance.id)\n",
"role": "user"
},
{
"content": "layer_post_save",
"role": "assistant"
}
] |
def x(instance, *args, **kwargs):
LOGGER.debug('Re-adding layer/service to search engine index')
if isinstance(instance.content_object, Service):
if not settings.REGISTRY_SKIP_CELERY:
index_service.delay(instance.content_object.id)
else:
index_service(instance.content_object.id)
elif not settings.REGISTRY_SKIP_CELERY:
index_layer.delay(instance.content_object.id)
else:
index_layer(instance.content_object.id)
|
issue_post_delete
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(instance, *args, **kwargs):\n LOGGER.debug('Re-adding layer/service to search engine index')\n if isinstance(instance.content_object, Service):\n if not settings.REGISTRY_SKIP_CELERY:\n index_service.delay(instance.content_object.id)\n else:\n index_service(instance.content_object.id)\n elif not settings.REGISTRY_SKIP_CELERY:\n index_layer.delay(instance.content_object.id)\n else:\n index_layer(instance.content_object.id)\n",
"role": "user"
},
{
"content": "issue_post_delete",
"role": "assistant"
}
] |
def x(self):
path = self.get_checks_admin_url()
content_type = ContentType.objects.get_for_model(self)
params = '?content_type__id__exact={0}&q={1}&success__exact=0'.format(
content_type.id, self.id)
url = path + params
return url
|
get_checks_admin_reliability_warning_url
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n path = self.get_checks_admin_url()\n content_type = ContentType.objects.get_for_model(self)\n params = '?content_type__id__exact={0}&q={1}&success__exact=0'.format(\n content_type.id, self.id)\n url = path + params\n return url\n",
"role": "user"
},
{
"content": "get_checks_admin_reliability_warning_url",
"role": "assistant"
}
] |
def x(self):
signals.post_save.disconnect(layer_post_save, sender=Layer)
try:
LOGGER.debug('Updating layers for service id %s' % self.id)
if self.type == 'OGC:WMS':
update_layers_wms(self)
elif self.type == 'OGC:WMTS':
update_layers_wmts(self)
elif self.type == 'ESRI:ArcGIS:MapServer':
update_layers_esri_mapserver(self)
elif self.type == 'ESRI:ArcGIS:ImageServer':
update_layers_esri_imageserver(self)
elif self.type == 'Hypermap:WorldMapLegacy':
update_layers_wm_legacy(self)
elif self.type == 'Hypermap:WorldMap':
update_layers_geonode_wm(self)
elif self.type == 'Hypermap:WARPER':
update_layers_warper(self)
except:
LOGGER.error('Error updating layers for service %s' % self.uuid)
signals.post_save.connect(layer_post_save, sender=Layer)
|
update_layers
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n signals.post_save.disconnect(layer_post_save, sender=Layer)\n try:\n LOGGER.debug('Updating layers for service id %s' % self.id)\n if self.type == 'OGC:WMS':\n update_layers_wms(self)\n elif self.type == 'OGC:WMTS':\n update_layers_wmts(self)\n elif self.type == 'ESRI:ArcGIS:MapServer':\n update_layers_esri_mapserver(self)\n elif self.type == 'ESRI:ArcGIS:ImageServer':\n update_layers_esri_imageserver(self)\n elif self.type == 'Hypermap:WorldMapLegacy':\n update_layers_wm_legacy(self)\n elif self.type == 'Hypermap:WorldMap':\n update_layers_geonode_wm(self)\n elif self.type == 'Hypermap:WARPER':\n update_layers_warper(self)\n except:\n LOGGER.error('Error updating layers for service %s' % self.uuid)\n signals.post_save.connect(layer_post_save, sender=Layer)\n",
"role": "user"
},
{
"content": "update_layers",
"role": "assistant"
}
] |
def x(self):
if self.type == 'Hypermap:WorldMap':
return
signals.post_save.disconnect(service_post_save, sender=Service)
try:
is_valid = True
if self.srs.filter(code__in=SUPPORTED_SRS).count() == 0:
LOGGER.debug(
'Service with id %s is marked invalid because in not exposed in SUPPORTED_SRS'
% self.id)
is_valid = False
if self.type == 'OGC:WMTS':
LOGGER.debug(
'Service with id %s is marked invalid because it is of type OGC:WMTS'
% self.id)
is_valid = False
if 'noaa' in self.url.lower():
LOGGER.debug(
'Service with id %s is marked invalid because it is from NOAA'
% self.id)
is_valid = False
self.is_valid = is_valid
self.save()
except:
LOGGER.error('Error updating validity of the service!')
signals.post_save.connect(service_post_save, sender=Service)
|
update_validity
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.type == 'Hypermap:WorldMap':\n return\n signals.post_save.disconnect(service_post_save, sender=Service)\n try:\n is_valid = True\n if self.srs.filter(code__in=SUPPORTED_SRS).count() == 0:\n LOGGER.debug(\n 'Service with id %s is marked invalid because in not exposed in SUPPORTED_SRS'\n % self.id)\n is_valid = False\n if self.type == 'OGC:WMTS':\n LOGGER.debug(\n 'Service with id %s is marked invalid because it is of type OGC:WMTS'\n % self.id)\n is_valid = False\n if 'noaa' in self.url.lower():\n LOGGER.debug(\n 'Service with id %s is marked invalid because it is from NOAA'\n % self.id)\n is_valid = False\n self.is_valid = is_valid\n self.save()\n except:\n LOGGER.error('Error updating validity of the service!')\n signals.post_save.connect(service_post_save, sender=Service)\n",
"role": "user"
},
{
"content": "update_validity",
"role": "assistant"
}
] |
def x(self):
if self.is_remote:
return self.url
return reverse('search_api', args=[self.slug])
|
get_search_url
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.is_remote:\n return self.url\n return reverse('search_api', args=[self.slug])\n",
"role": "user"
},
{
"content": "get_search_url",
"role": "assistant"
}
] |
def x(self):
endpoint = self.url
if self.type not in ('Hypermap:WorldMap',):
endpoint = (
'registry/%s/layer/%s/map/wmts/1.0.0/WMTSCapabilities.xml' % (
self.catalog.slug, self.id))
return endpoint
|
get_url_endpoint
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n endpoint = self.url\n if self.type not in ('Hypermap:WorldMap',):\n endpoint = (\n 'registry/%s/layer/%s/map/wmts/1.0.0/WMTSCapabilities.xml' % (\n self.catalog.slug, self.id))\n return endpoint\n",
"role": "user"
},
{
"content": "get_url_endpoint",
"role": "assistant"
}
] |
def x(url=None):
if url is None:
url = user_input(
'Input the URL of the GitHub repository to use as a `trytravis` repository: '
)
url = url.strip()
http_match = _HTTPS_REGEX.match(url)
ssh_match = _SSH_REGEX.match(url)
if not http_match and not ssh_match:
raise RuntimeError(
"That URL doesn't look like a valid GitHub URL. We expect something of the form: `https://github.com/[USERNAME]/[REPOSITORY]` or `ssh://git@github.com/[USERNAME]/[REPOSITORY]"
)
if http_match:
_, name = http_match.groups()
else:
_, name = ssh_match.groups()
if 'trytravis' not in name:
raise RuntimeError(
"You must have `trytravis` in the name of your repository. This is a security feature to reduce chances of running git push -f on a repository you don't mean to."
)
accept = user_input(
'Remember that `trytravis` will make commits on your behalf to `%s`. Are you sure you wish to use this repository? Type `y` or `yes` to accept: '
% url)
if accept.lower() not in ['y', 'yes']:
raise RuntimeError('Operation aborted by user.')
if not os.path.isdir(config_dir):
os.makedirs(config_dir)
with open(os.path.join(config_dir, 'repo'), 'w+') as f:
f.truncate()
f.write(url)
print('Repository saved successfully.')
|
input_github_repo
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(url=None):\n if url is None:\n url = user_input(\n 'Input the URL of the GitHub repository to use as a `trytravis` repository: '\n )\n url = url.strip()\n http_match = _HTTPS_REGEX.match(url)\n ssh_match = _SSH_REGEX.match(url)\n if not http_match and not ssh_match:\n raise RuntimeError(\n \"That URL doesn't look like a valid GitHub URL. We expect something of the form: `https://github.com/[USERNAME]/[REPOSITORY]` or `ssh://git@github.com/[USERNAME]/[REPOSITORY]\"\n )\n if http_match:\n _, name = http_match.groups()\n else:\n _, name = ssh_match.groups()\n if 'trytravis' not in name:\n raise RuntimeError(\n \"You must have `trytravis` in the name of your repository. This is a security feature to reduce chances of running git push -f on a repository you don't mean to.\"\n )\n accept = user_input(\n 'Remember that `trytravis` will make commits on your behalf to `%s`. Are you sure you wish to use this repository? Type `y` or `yes` to accept: '\n % url)\n if accept.lower() not in ['y', 'yes']:\n raise RuntimeError('Operation aborted by user.')\n if not os.path.isdir(config_dir):\n os.makedirs(config_dir)\n with open(os.path.join(config_dir, 'repo'), 'w+') as f:\n f.truncate()\n f.write(url)\n print('Repository saved successfully.')\n",
"role": "user"
},
{
"content": "input_github_repo",
"role": "assistant"
}
] |
def x():
if 'TRAVIS' in os.environ:
raise RuntimeError(
'Detected that we are running in Travis. Stopping to prevent infinite loops.'
)
try:
with open(os.path.join(config_dir, 'repo'), 'r') as f:
return f.read()
except (OSError, IOError):
raise RuntimeError(
'Could not find your repository. Have you ran `trytravis --repo`?')
|
load_github_repo
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n if 'TRAVIS' in os.environ:\n raise RuntimeError(\n 'Detected that we are running in Travis. Stopping to prevent infinite loops.'\n )\n try:\n with open(os.path.join(config_dir, 'repo'), 'r') as f:\n return f.read()\n except (OSError, IOError):\n raise RuntimeError(\n 'Could not find your repository. Have you ran `trytravis --repo`?')\n",
"role": "user"
},
{
"content": "load_github_repo",
"role": "assistant"
}
] |
def x(path, url):
try:
repo = git.Repo(path)
except Exception:
raise RuntimeError("Couldn't locate a repository at `%s`." % path)
commited = False
try:
try:
repo.delete_remote('trytravis')
except Exception:
pass
print('Adding a temporary remote to `%s`...' % url)
remote = repo.create_remote('trytravis', url)
print('Adding all local changes...')
repo.git.add('--all')
try:
print('Committing local changes...')
timestamp = datetime.datetime.now().isoformat()
repo.git.commit(m='trytravis-' + timestamp)
commited = True
except git.exc.GitCommandError as e:
if 'nothing to commit' in str(e):
commited = False
else:
raise
commit = repo.head.commit.hexsha
committed_at = repo.head.commit.committed_datetime
print('Pushing to `trytravis` remote...')
remote.push(force=True)
finally:
if commited:
print('Reverting to old state...')
repo.git.reset('HEAD^')
try:
repo.delete_remote('trytravis')
except Exception:
pass
return commit, committed_at
|
submit_changes_to_github_repo
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path, url):\n try:\n repo = git.Repo(path)\n except Exception:\n raise RuntimeError(\"Couldn't locate a repository at `%s`.\" % path)\n commited = False\n try:\n try:\n repo.delete_remote('trytravis')\n except Exception:\n pass\n print('Adding a temporary remote to `%s`...' % url)\n remote = repo.create_remote('trytravis', url)\n print('Adding all local changes...')\n repo.git.add('--all')\n try:\n print('Committing local changes...')\n timestamp = datetime.datetime.now().isoformat()\n repo.git.commit(m='trytravis-' + timestamp)\n commited = True\n except git.exc.GitCommandError as e:\n if 'nothing to commit' in str(e):\n commited = False\n else:\n raise\n commit = repo.head.commit.hexsha\n committed_at = repo.head.commit.committed_datetime\n print('Pushing to `trytravis` remote...')\n remote.push(force=True)\n finally:\n if commited:\n print('Reverting to old state...')\n repo.git.reset('HEAD^')\n try:\n repo.delete_remote('trytravis')\n except Exception:\n pass\n return commit, committed_at\n",
"role": "user"
},
{
"content": "submit_changes_to_github_repo",
"role": "assistant"
}
] |
def x(url, commit, committed_at):
print('Waiting for a Travis build to appear for `%s` after `%s`...' % (
commit, committed_at))
import requests
slug = _slug_from_url(url)
start_time = time.time()
build_id = None
while time.time() - start_time < 60:
with requests.get('https://api.travis-ci.org/repos/%s/builds' %
slug, headers=_travis_headers()) as r:
if not r.ok:
raise RuntimeError(
'Could not reach the Travis API endpoint. Additional information: %s'
% str(r.content))
commit_to_sha = {}
json = r.json()
for travis_commit in sorted(json['commits'], key=lambda x: x[
'committed_at']):
travis_committed_at = datetime.datetime.strptime(travis_commit
['committed_at'], '%Y-%m-%dT%H:%M:%SZ').replace(tzinfo=utc)
if travis_committed_at < committed_at:
continue
commit_to_sha[travis_commit['id']] = travis_commit['sha']
for build in json['builds']:
if build['commit_id'] in commit_to_sha and commit_to_sha[build
['commit_id']] == commit:
build_id = build['id']
print('Travis build id: `%d`' % build_id)
print(
'Travis build URL: `https://travis-ci.org/%s/builds/%d`'
% (slug, build_id))
if build_id is not None:
break
time.sleep(3.0)
else:
raise RuntimeError(
'Timed out while waiting for a Travis build to start. Is Travis configured for `%s`?'
% url)
return build_id
|
wait_for_travis_build
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(url, commit, committed_at):\n print('Waiting for a Travis build to appear for `%s` after `%s`...' % (\n commit, committed_at))\n import requests\n slug = _slug_from_url(url)\n start_time = time.time()\n build_id = None\n while time.time() - start_time < 60:\n with requests.get('https://api.travis-ci.org/repos/%s/builds' %\n slug, headers=_travis_headers()) as r:\n if not r.ok:\n raise RuntimeError(\n 'Could not reach the Travis API endpoint. Additional information: %s'\n % str(r.content))\n commit_to_sha = {}\n json = r.json()\n for travis_commit in sorted(json['commits'], key=lambda x: x[\n 'committed_at']):\n travis_committed_at = datetime.datetime.strptime(travis_commit\n ['committed_at'], '%Y-%m-%dT%H:%M:%SZ').replace(tzinfo=utc)\n if travis_committed_at < committed_at:\n continue\n commit_to_sha[travis_commit['id']] = travis_commit['sha']\n for build in json['builds']:\n if build['commit_id'] in commit_to_sha and commit_to_sha[build\n ['commit_id']] == commit:\n build_id = build['id']\n print('Travis build id: `%d`' % build_id)\n print(\n 'Travis build URL: `https://travis-ci.org/%s/builds/%d`'\n % (slug, build_id))\n if build_id is not None:\n break\n time.sleep(3.0)\n else:\n raise RuntimeError(\n 'Timed out while waiting for a Travis build to start. Is Travis configured for `%s`?'\n % url)\n return build_id\n",
"role": "user"
},
{
"content": "wait_for_travis_build",
"role": "assistant"
}
] |
def x(build_id):
import requests
try:
build_size = None
running = True
while running:
with requests.get('https://api.travis-ci.org/builds/%d' %
build_id, headers=_travis_headers()) as r:
json = r.json()
if build_size is not None:
if build_size > 1:
sys.stdout.write('\r\x1b[%dA' % build_size)
else:
sys.stdout.write('\r')
build_size = len(json['jobs'])
running = False
current_number = 1
for job in json['jobs']:
color, state, is_running = _travis_job_state(job['state'])
if is_running:
running = True
platform = job['config']['os']
if platform == 'osx':
platform = ' osx '
env = job['config'].get('env', '')
sudo = 's' if job['config'].get('sudo', True) else 'c'
lang = job['config'].get('language', 'generic')
padding = ' ' * (len(str(build_size)) - len(str(
current_number)))
number = str(current_number) + padding
current_number += 1
job_display = '#' + ' '.join([number, state, platform,
sudo, lang, env])
print(color + job_display + colorama.Style.RESET_ALL)
time.sleep(3.0)
except KeyboardInterrupt:
pass
|
watch_travis_build
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(build_id):\n import requests\n try:\n build_size = None\n running = True\n while running:\n with requests.get('https://api.travis-ci.org/builds/%d' %\n build_id, headers=_travis_headers()) as r:\n json = r.json()\n if build_size is not None:\n if build_size > 1:\n sys.stdout.write('\\r\\x1b[%dA' % build_size)\n else:\n sys.stdout.write('\\r')\n build_size = len(json['jobs'])\n running = False\n current_number = 1\n for job in json['jobs']:\n color, state, is_running = _travis_job_state(job['state'])\n if is_running:\n running = True\n platform = job['config']['os']\n if platform == 'osx':\n platform = ' osx '\n env = job['config'].get('env', '')\n sudo = 's' if job['config'].get('sudo', True) else 'c'\n lang = job['config'].get('language', 'generic')\n padding = ' ' * (len(str(build_size)) - len(str(\n current_number)))\n number = str(current_number) + padding\n current_number += 1\n job_display = '#' + ' '.join([number, state, platform,\n sudo, lang, env])\n print(color + job_display + colorama.Style.RESET_ALL)\n time.sleep(3.0)\n except KeyboardInterrupt:\n pass\n",
"role": "user"
},
{
"content": "watch_travis_build",
"role": "assistant"
}
] |
def x(state):
if state in [None, 'queued', 'created', 'received']:
return colorama.Fore.YELLOW, '*', True
elif state in ['started', 'running']:
return colorama.Fore.LIGHTYELLOW_EX, '*', True
elif state == 'passed':
return colorama.Fore.LIGHTGREEN_EX, 'P', False
elif state == 'failed':
return colorama.Fore.LIGHTRED_EX, 'X', False
elif state == 'errored':
return colorama.Fore.LIGHTRED_EX, '!', False
elif state == 'canceled':
return colorama.Fore.LIGHTBLACK_EX, 'X', False
else:
raise RuntimeError('unknown state: %s' % str(state))
|
travis_job_state
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(state):\n if state in [None, 'queued', 'created', 'received']:\n return colorama.Fore.YELLOW, '*', True\n elif state in ['started', 'running']:\n return colorama.Fore.LIGHTYELLOW_EX, '*', True\n elif state == 'passed':\n return colorama.Fore.LIGHTGREEN_EX, 'P', False\n elif state == 'failed':\n return colorama.Fore.LIGHTRED_EX, 'X', False\n elif state == 'errored':\n return colorama.Fore.LIGHTRED_EX, '!', False\n elif state == 'canceled':\n return colorama.Fore.LIGHTBLACK_EX, 'X', False\n else:\n raise RuntimeError('unknown state: %s' % str(state))\n",
"role": "user"
},
{
"content": "travis_job_state",
"role": "assistant"
}
] |
def x(url):
http_match = _HTTPS_REGEX.match(url)
ssh_match = _SSH_REGEX.match(url)
if not http_match and not ssh_match:
raise RuntimeError(
'Could not parse the URL (`%s`) for your repository.' % url)
if http_match:
return '/'.join(http_match.groups())
else:
return '/'.join(ssh_match.groups())
|
slug_from_url
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(url):\n http_match = _HTTPS_REGEX.match(url)\n ssh_match = _SSH_REGEX.match(url)\n if not http_match and not ssh_match:\n raise RuntimeError(\n 'Could not parse the URL (`%s`) for your repository.' % url)\n if http_match:\n return '/'.join(http_match.groups())\n else:\n return '/'.join(ssh_match.groups())\n",
"role": "user"
},
{
"content": "slug_from_url",
"role": "assistant"
}
] |
def x():
platform_system = platform.system()
if platform_system == 'Linux':
os_name, os_version, _ = platform.dist()
else:
os_name = platform_system
os_version = platform.version()
python_version = platform.python_version()
return 'trytravis %s (%s %s, python %s)' % (__version__, os_name.lower(
), os_version, python_version)
|
version_string
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n platform_system = platform.system()\n if platform_system == 'Linux':\n os_name, os_version, _ = platform.dist()\n else:\n os_name = platform_system\n os_version = platform.version()\n python_version = platform.python_version()\n return 'trytravis %s (%s %s, python %s)' % (__version__, os_name.lower(\n ), os_version, python_version)\n",
"role": "user"
},
{
"content": "version_string",
"role": "assistant"
}
] |
def x(argv):
repo_input_argv = len(argv) == 2 and argv[0] in ['--repo', '-r', '-R']
if len(argv) > 1 and not repo_input_argv:
_main(['--help'])
if len(argv) == 1 or repo_input_argv:
arg = argv[0]
if arg in ['-h', '--help', '-H']:
print(_USAGE)
elif arg in ['-v', '--version', '-V']:
print(_version_string())
elif arg in ['-r', '--repo', '-R']:
if len(argv) == 2:
url = argv[1]
else:
url = None
_input_github_repo(url)
elif arg in ['--no-wait', '-nw']:
url = _load_github_repo()
commit, committed = _submit_changes_to_github_repo(os.getcwd(), url
)
build_id = _wait_for_travis_build(url, commit, committed)
else:
_main(['--help'])
elif len(argv) == 0:
url = _load_github_repo()
commit, committed = _submit_changes_to_github_repo(os.getcwd(), url)
build_id = _wait_for_travis_build(url, commit, committed)
_watch_travis_build(build_id)
|
main
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(argv):\n repo_input_argv = len(argv) == 2 and argv[0] in ['--repo', '-r', '-R']\n if len(argv) > 1 and not repo_input_argv:\n _main(['--help'])\n if len(argv) == 1 or repo_input_argv:\n arg = argv[0]\n if arg in ['-h', '--help', '-H']:\n print(_USAGE)\n elif arg in ['-v', '--version', '-V']:\n print(_version_string())\n elif arg in ['-r', '--repo', '-R']:\n if len(argv) == 2:\n url = argv[1]\n else:\n url = None\n _input_github_repo(url)\n elif arg in ['--no-wait', '-nw']:\n url = _load_github_repo()\n commit, committed = _submit_changes_to_github_repo(os.getcwd(), url\n )\n build_id = _wait_for_travis_build(url, commit, committed)\n else:\n _main(['--help'])\n elif len(argv) == 0:\n url = _load_github_repo()\n commit, committed = _submit_changes_to_github_repo(os.getcwd(), url)\n build_id = _wait_for_travis_build(url, commit, committed)\n _watch_travis_build(build_id)\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x(argv=None):
try:
colorama.init()
if argv is None:
argv = sys.argv[1:]
_main(argv)
except RuntimeError as e:
print(colorama.Fore.RED + 'ERROR: ' + str(e) + colorama.Style.RESET_ALL
)
sys.exit(1)
else:
sys.exit(0)
|
main
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(argv=None):\n try:\n colorama.init()\n if argv is None:\n argv = sys.argv[1:]\n _main(argv)\n except RuntimeError as e:\n print(colorama.Fore.RED + 'ERROR: ' + str(e) + colorama.Style.RESET_ALL\n )\n sys.exit(1)\n else:\n sys.exit(0)\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x(request, url=None, catalog_id=None):
if request.user.is_authenticated():
settings.REGISTRY_PYCSW['manager']['transactions'] = 'true'
env = request.META.copy()
if request.method == 'POST':
from StringIO import StringIO
env['wsgi.input'] = StringIO(request.body)
env.update({'local.app_root': os.path.dirname(__file__), 'REQUEST_URI':
request.build_absolute_uri()})
if url is not None:
settings.REGISTRY_PYCSW['server']['url'] = url
if catalog_id is not None:
settings.REGISTRY_PYCSW['repository']['filter'
] = 'catalog_id = %d' % catalog_id
csw = server.Csw(settings.REGISTRY_PYCSW, env)
content = csw.dispatch_wsgi()
if isinstance(content, list):
content = content[1]
response = HttpResponse(content, content_type=csw.contenttype)
response['Access-Control-Allow-Origin'] = '*'
return response
|
csw_global_dispatch
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(request, url=None, catalog_id=None):\n if request.user.is_authenticated():\n settings.REGISTRY_PYCSW['manager']['transactions'] = 'true'\n env = request.META.copy()\n if request.method == 'POST':\n from StringIO import StringIO\n env['wsgi.input'] = StringIO(request.body)\n env.update({'local.app_root': os.path.dirname(__file__), 'REQUEST_URI':\n request.build_absolute_uri()})\n if url is not None:\n settings.REGISTRY_PYCSW['server']['url'] = url\n if catalog_id is not None:\n settings.REGISTRY_PYCSW['repository']['filter'\n ] = 'catalog_id = %d' % catalog_id\n csw = server.Csw(settings.REGISTRY_PYCSW, env)\n content = csw.dispatch_wsgi()\n if isinstance(content, list):\n content = content[1]\n response = HttpResponse(content, content_type=csw.contenttype)\n response['Access-Control-Allow-Origin'] = '*'\n return response\n",
"role": "user"
},
{
"content": "csw_global_dispatch",
"role": "assistant"
}
] |
def x(request, catalog_slug):
catalog = get_object_or_404(Catalog, slug=catalog_slug)
if catalog:
url = settings.SITE_URL.rstrip('/') + request.path.rstrip('/')
return csw_global_dispatch(request, url=url, catalog_id=catalog.id)
|
csw_global_dispatch_by_catalog
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(request, catalog_slug):\n catalog = get_object_or_404(Catalog, slug=catalog_slug)\n if catalog:\n url = settings.SITE_URL.rstrip('/') + request.path.rstrip('/')\n return csw_global_dispatch(request, url=url, catalog_id=catalog.id)\n",
"role": "user"
},
{
"content": "csw_global_dispatch_by_catalog",
"role": "assistant"
}
] |
def x(request):
ctx = {'shortname': settings.REGISTRY_PYCSW['metadata:main'][
'identification_title'], 'description': settings.REGISTRY_PYCSW[
'metadata:main']['identification_abstract'], 'developer': settings.
REGISTRY_PYCSW['metadata:main']['contact_name'], 'contact':
settings.REGISTRY_PYCSW['metadata:main']['contact_email'],
'attribution': settings.REGISTRY_PYCSW['metadata:main'][
'provider_name'], 'tags': settings.REGISTRY_PYCSW['metadata:main'][
'identification_keywords'].replace(',', ' '), 'url': settings.
SITE_URL.rstrip('/')}
return render_to_response('search/opensearch_description.xml', ctx,
content_type='application/opensearchdescription+xml')
|
opensearch_dispatch
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(request):\n ctx = {'shortname': settings.REGISTRY_PYCSW['metadata:main'][\n 'identification_title'], 'description': settings.REGISTRY_PYCSW[\n 'metadata:main']['identification_abstract'], 'developer': settings.\n REGISTRY_PYCSW['metadata:main']['contact_name'], 'contact':\n settings.REGISTRY_PYCSW['metadata:main']['contact_email'],\n 'attribution': settings.REGISTRY_PYCSW['metadata:main'][\n 'provider_name'], 'tags': settings.REGISTRY_PYCSW['metadata:main'][\n 'identification_keywords'].replace(',', ' '), 'url': settings.\n SITE_URL.rstrip('/')}\n return render_to_response('search/opensearch_description.xml', ctx,\n content_type='application/opensearchdescription+xml')\n",
"role": "user"
},
{
"content": "opensearch_dispatch",
"role": "assistant"
}
] |
def x(coords):
if len(coords) != 4:
return False
for coord in coords[0:3]:
try:
num = float(coord)
if math.isnan(num):
return False
if math.isinf(num):
return False
except ValueError:
return False
return True
|
good_coords
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(coords):\n if len(coords) != 4:\n return False\n for coord in coords[0:3]:\n try:\n num = float(coord)\n if math.isnan(num):\n return False\n if math.isinf(num):\n return False\n except ValueError:\n return False\n return True\n",
"role": "user"
},
{
"content": "good_coords",
"role": "assistant"
}
] |
def x():
ESHypermap.es.indices.delete(ESHypermap.index_name, ignore=[400, 404])
LOGGER.debug('Elasticsearch: Index cleared')
|
clear_es
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n ESHypermap.es.indices.delete(ESHypermap.index_name, ignore=[400, 404])\n LOGGER.debug('Elasticsearch: Index cleared')\n",
"role": "user"
},
{
"content": "clear_es",
"role": "assistant"
}
] |
def x(catalog_slug):
mapping = {'mappings': {'layer': {'properties': {'layer_geoshape': {
'type': 'geo_shape', 'tree': 'quadtree', 'precision':
REGISTRY_MAPPING_PRECISION}}}}}
ESHypermap.es.indices.create(catalog_slug, ignore=[400, 404], body=mapping)
|
create_indices
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(catalog_slug):\n mapping = {'mappings': {'layer': {'properties': {'layer_geoshape': {\n 'type': 'geo_shape', 'tree': 'quadtree', 'precision':\n REGISTRY_MAPPING_PRECISION}}}}}\n ESHypermap.es.indices.create(catalog_slug, ignore=[400, 404], body=mapping)\n",
"role": "user"
},
{
"content": "create_indices",
"role": "assistant"
}
] |
def x(procname, scriptname):
import signal
import subprocess
p = subprocess.Popen(['ps', 'aux'], stdout=subprocess.PIPE)
out, err = p.communicate()
for line in out.decode().splitlines():
if procname in line and scriptname in line:
pid = int(line.split()[1])
info('Stopping %s %s %d' % (procname, scriptname, pid))
os.kill(pid, signal.SIGKILL)
|
kill_process
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(procname, scriptname):\n import signal\n import subprocess\n p = subprocess.Popen(['ps', 'aux'], stdout=subprocess.PIPE)\n out, err = p.communicate()\n for line in out.decode().splitlines():\n if procname in line and scriptname in line:\n pid = int(line.split()[1])\n info('Stopping %s %s %d' % (procname, scriptname, pid))\n os.kill(pid, signal.SIGKILL)\n",
"role": "user"
},
{
"content": "kill_process",
"role": "assistant"
}
] |
def x():
services_list = ('Harvard WorldMap',
'Harvard WorldMap open source web geospatial platform',
'Hypermap:WorldMap', 'http://worldmap.harvard.edu'), ('NYPL MapWarper',
'The New York Public Library (NYPL) MapWarper web site',
'Hypermap:WARPER', 'http://maps.nypl.org/warper/maps'), ('Map Warper',
'The MapWarper web site developed, hosted and maintained by Tim Waters'
, 'Hypermap:WARPER', 'http://mapwarper.net/maps'), ('WorldMap Warp',
'The MapWarper instance part of the Harvard WorldMap project',
'Hypermap:WARPER', 'http://warp.worldmap.harvard.edu/maps'), (
'WFP GeoNode', 'World Food Programme GeoNode', 'OGC:WMS',
'http://geonode.wfp.org/geoserver/ows?'), ('NASA EARTHDATA',
'NASA EARTHDATA, powered by EOSDIS', 'OGC:WMTS',
'http://map1.vis.earthdata.nasa.gov/wmts-geo/1.0.0/WMTSCapabilities.xml'
)
esri_endpoint = 'https://gis.ngdc.noaa.gov/arcgis/rest/services'
LOGGER.debug('*** Importing esri endpoint: %s' % esri_endpoint)
create_services_from_endpoint(esri_endpoint)
for service in services_list:
LOGGER.debug('*** Importing %s' % service[0])
service = Service(title=service[0], abstract=service[1], type=
service[2], url=service[3])
service.save()
|
populate_initial_services
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n services_list = ('Harvard WorldMap',\n 'Harvard WorldMap open source web geospatial platform',\n 'Hypermap:WorldMap', 'http://worldmap.harvard.edu'), ('NYPL MapWarper',\n 'The New York Public Library (NYPL) MapWarper web site',\n 'Hypermap:WARPER', 'http://maps.nypl.org/warper/maps'), ('Map Warper',\n 'The MapWarper web site developed, hosted and maintained by Tim Waters'\n , 'Hypermap:WARPER', 'http://mapwarper.net/maps'), ('WorldMap Warp',\n 'The MapWarper instance part of the Harvard WorldMap project',\n 'Hypermap:WARPER', 'http://warp.worldmap.harvard.edu/maps'), (\n 'WFP GeoNode', 'World Food Programme GeoNode', 'OGC:WMS',\n 'http://geonode.wfp.org/geoserver/ows?'), ('NASA EARTHDATA',\n 'NASA EARTHDATA, powered by EOSDIS', 'OGC:WMTS',\n 'http://map1.vis.earthdata.nasa.gov/wmts-geo/1.0.0/WMTSCapabilities.xml'\n )\n esri_endpoint = 'https://gis.ngdc.noaa.gov/arcgis/rest/services'\n LOGGER.debug('*** Importing esri endpoint: %s' % esri_endpoint)\n create_services_from_endpoint(esri_endpoint)\n for service in services_list:\n LOGGER.debug('*** Importing %s' % service[0])\n service = Service(title=service[0], abstract=service[1], type=\n service[2], url=service[3])\n service.save()\n",
"role": "user"
},
{
"content": "populate_initial_services",
"role": "assistant"
}
] |
def x(request):
get = request.GET.copy()
new_get = request.GET.copy()
for key in get.iterkeys():
if key.count('.') > 0:
new_key = key.replace('.', '_')
new_get[new_key] = get.get(key)
del new_get[key]
return new_get
|
parse_get_params
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(request):\n get = request.GET.copy()\n new_get = request.GET.copy()\n for key in get.iterkeys():\n if key.count('.') > 0:\n new_key = key.replace('.', '_')\n new_get[new_key] = get.get(key)\n del new_get[key]\n return new_get\n",
"role": "user"
},
{
"content": "parse_get_params",
"role": "assistant"
}
] |
def x():
tcp_adapter = TcpAdapter('192.168.1.3', name='HASS', activate_source=False)
hdmi_network = HDMINetwork(tcp_adapter)
hdmi_network.start()
while True:
for d in hdmi_network.devices:
_LOGGER.info('Device: %s', d)
time.sleep(7)
|
main
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n tcp_adapter = TcpAdapter('192.168.1.3', name='HASS', activate_source=False)\n hdmi_network = HDMINetwork(tcp_adapter)\n hdmi_network.start()\n while True:\n for d in hdmi_network.devices:\n _LOGGER.info('Device: %s', d)\n time.sleep(7)\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x(digest1, digest2):
digest1 = tuple([int(digest1[i:i + 2], 16) for i in range(0, 63, 2)])
digest2 = tuple([int(digest2[i:i + 2], 16) for i in range(0, 63, 2)])
bits = 0
for i in range(32):
bits += POPC[255 & digest1[i] ^ digest2[i]]
return 128 - bits
|
compare_hexdigests
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(digest1, digest2):\n digest1 = tuple([int(digest1[i:i + 2], 16) for i in range(0, 63, 2)])\n digest2 = tuple([int(digest2[i:i + 2], 16) for i in range(0, 63, 2)])\n bits = 0\n for i in range(32):\n bits += POPC[255 & digest1[i] ^ digest2[i]]\n return 128 - bits\n",
"role": "user"
},
{
"content": "compare_hexdigests",
"role": "assistant"
}
] |
def x(self, data):
for character in data:
if PY3:
ch = character
else:
ch = ord(character)
self.count += 1
if self.lastch[1] > -1:
self.acc[self.tran3(ch, self.lastch[0], self.lastch[1], 0)] += 1
if self.lastch[2] > -1:
self.acc[self.tran3(ch, self.lastch[0], self.lastch[2], 1)] += 1
self.acc[self.tran3(ch, self.lastch[1], self.lastch[2], 2)] += 1
if self.lastch[3] > -1:
self.acc[self.tran3(ch, self.lastch[0], self.lastch[3], 3)] += 1
self.acc[self.tran3(ch, self.lastch[1], self.lastch[3], 4)] += 1
self.acc[self.tran3(ch, self.lastch[2], self.lastch[3], 5)] += 1
self.acc[self.tran3(self.lastch[3], self.lastch[0], ch, 6)] += 1
self.acc[self.tran3(self.lastch[3], self.lastch[2], ch, 7)] += 1
self.lastch = [ch] + self.lastch[:3]
|
update
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n for character in data:\n if PY3:\n ch = character\n else:\n ch = ord(character)\n self.count += 1\n if self.lastch[1] > -1:\n self.acc[self.tran3(ch, self.lastch[0], self.lastch[1], 0)] += 1\n if self.lastch[2] > -1:\n self.acc[self.tran3(ch, self.lastch[0], self.lastch[2], 1)] += 1\n self.acc[self.tran3(ch, self.lastch[1], self.lastch[2], 2)] += 1\n if self.lastch[3] > -1:\n self.acc[self.tran3(ch, self.lastch[0], self.lastch[3], 3)] += 1\n self.acc[self.tran3(ch, self.lastch[1], self.lastch[3], 4)] += 1\n self.acc[self.tran3(ch, self.lastch[2], self.lastch[3], 5)] += 1\n self.acc[self.tran3(self.lastch[3], self.lastch[0], ch, 6)] += 1\n self.acc[self.tran3(self.lastch[3], self.lastch[2], ch, 7)] += 1\n self.lastch = [ch] + self.lastch[:3]\n",
"role": "user"
},
{
"content": "update",
"role": "assistant"
}
] |
def x(self):
total = 0
if self.count == 3:
total = 1
elif self.count == 4:
total = 4
elif self.count > 4:
total = 8 * self.count - 28
threshold = total / 256
code = [0] * 32
for i in range(256):
if self.acc[i] > threshold:
code[i >> 3] += 1 << (i & 7)
return code[::-1]
|
digest
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n total = 0\n if self.count == 3:\n total = 1\n elif self.count == 4:\n total = 4\n elif self.count > 4:\n total = 8 * self.count - 28\n threshold = total / 256\n code = [0] * 32\n for i in range(256):\n if self.acc[i] > threshold:\n code[i >> 3] += 1 << (i & 7)\n return code[::-1]\n",
"role": "user"
},
{
"content": "digest",
"role": "assistant"
}
] |
def x(self, filename):
f = open(filename, 'rb')
while True:
data = f.read(10480)
if not data:
break
self.update(data)
f.close()
|
from_file
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, filename):\n f = open(filename, 'rb')\n while True:\n data = f.read(10480)\n if not data:\n break\n self.update(data)\n f.close()\n",
"role": "user"
},
{
"content": "from_file",
"role": "assistant"
}
] |
def x(self, otherdigest, ishex=False):
bits = 0
myd = self.digest()
if ishex:
otherdigest = tuple([int(otherdigest[i:i + 2], 16) for i in range(0,
63, 2)])
for i in range(32):
bits += POPC[255 & myd[i] ^ otherdigest[i]]
return 128 - bits
|
compare
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, otherdigest, ishex=False):\n bits = 0\n myd = self.digest()\n if ishex:\n otherdigest = tuple([int(otherdigest[i:i + 2], 16) for i in range(0,\n 63, 2)])\n for i in range(32):\n bits += POPC[255 & myd[i] ^ otherdigest[i]]\n return 128 - bits\n",
"role": "user"
},
{
"content": "compare",
"role": "assistant"
}
] |
def x(api_response):
try:
output = json.dumps(api_response.cgx_content, indent=4)
except (TypeError, ValueError, AttributeError):
try:
output = json.dumps(api_response, indent=4)
except (TypeError, ValueError, AttributeError):
output = api_response
return output
|
jdout
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(api_response):\n try:\n output = json.dumps(api_response.cgx_content, indent=4)\n except (TypeError, ValueError, AttributeError):\n try:\n output = json.dumps(api_response, indent=4)\n except (TypeError, ValueError, AttributeError):\n output = api_response\n return output\n",
"role": "user"
},
{
"content": "jdout",
"role": "assistant"
}
] |
def x(api_response, sensitive=False):
try:
output = 'REQUEST: {0} {1}\n'.format(api_response.request.method,
api_response.request.path_url)
output += 'REQUEST HEADERS:\n'
for key, value in api_response.request.headers.items():
if key.lower() in ['cookie'] and not sensitive:
cookie_list = value.split('; ')
muted_cookie_list = []
for cookie in cookie_list:
if cookie.lower().strip().startswith('auth_token='):
newcookie = cookie.strip()[:11
] + '"<SENSITIVE - NOT SHOWN BY DEFAULT>"'
muted_cookie_list.append(newcookie)
else:
muted_cookie_list.append(cookie)
muted_value = '; '.join(muted_cookie_list)
output += '\t{0}: {1}\n'.format(key, muted_value)
elif key.lower() in ['x-auth-token'] and not sensitive:
output += '\t{0}: {1}\n'.format(key,
'<SENSITIVE - NOT SHOWN BY DEFAULT>')
else:
output += '\t{0}: {1}\n'.format(key, value)
if not api_response.request.body:
output += 'REQUEST BODY:\n{0}\n\n'.format({})
else:
try:
output += 'REQUEST BODY:\n{0}\n\n'.format(json.dumps(json.
loads(api_response.request.body), indent=4))
except (TypeError, ValueError, AttributeError):
output += 'REQUEST BODY:\n{0}\n\n'.format(jdout(
api_response.request.body))
output += 'RESPONSE: {0} {1}\n'.format(api_response.status_code,
api_response.reason)
output += 'RESPONSE HEADERS:\n'
for key, value in api_response.headers.items():
output += '\t{0}: {1}\n'.format(key, value)
try:
output += 'RESPONSE DATA:\n{0}'.format(json.dumps(api_response.
cgx_content, indent=4))
except (TypeError, ValueError, AttributeError):
output += 'RESPONSE DATA:\n{0}'.format(json.dumps(json.loads(
api_response.content), indent=4))
except (TypeError, ValueError, AttributeError, UnicodeDecodeError):
try:
output = json.dumps(api_response, indent=4)
except (TypeError, ValueError, AttributeError):
output = api_response
return output
|
jdout_detailed
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(api_response, sensitive=False):\n try:\n output = 'REQUEST: {0} {1}\\n'.format(api_response.request.method,\n api_response.request.path_url)\n output += 'REQUEST HEADERS:\\n'\n for key, value in api_response.request.headers.items():\n if key.lower() in ['cookie'] and not sensitive:\n cookie_list = value.split('; ')\n muted_cookie_list = []\n for cookie in cookie_list:\n if cookie.lower().strip().startswith('auth_token='):\n newcookie = cookie.strip()[:11\n ] + '\"<SENSITIVE - NOT SHOWN BY DEFAULT>\"'\n muted_cookie_list.append(newcookie)\n else:\n muted_cookie_list.append(cookie)\n muted_value = '; '.join(muted_cookie_list)\n output += '\\t{0}: {1}\\n'.format(key, muted_value)\n elif key.lower() in ['x-auth-token'] and not sensitive:\n output += '\\t{0}: {1}\\n'.format(key,\n '<SENSITIVE - NOT SHOWN BY DEFAULT>')\n else:\n output += '\\t{0}: {1}\\n'.format(key, value)\n if not api_response.request.body:\n output += 'REQUEST BODY:\\n{0}\\n\\n'.format({})\n else:\n try:\n output += 'REQUEST BODY:\\n{0}\\n\\n'.format(json.dumps(json.\n loads(api_response.request.body), indent=4))\n except (TypeError, ValueError, AttributeError):\n output += 'REQUEST BODY:\\n{0}\\n\\n'.format(jdout(\n api_response.request.body))\n output += 'RESPONSE: {0} {1}\\n'.format(api_response.status_code,\n api_response.reason)\n output += 'RESPONSE HEADERS:\\n'\n for key, value in api_response.headers.items():\n output += '\\t{0}: {1}\\n'.format(key, value)\n try:\n output += 'RESPONSE DATA:\\n{0}'.format(json.dumps(api_response.\n cgx_content, indent=4))\n except (TypeError, ValueError, AttributeError):\n output += 'RESPONSE DATA:\\n{0}'.format(json.dumps(json.loads(\n api_response.content), indent=4))\n except (TypeError, ValueError, AttributeError, UnicodeDecodeError):\n try:\n output = json.dumps(api_response, indent=4)\n except (TypeError, ValueError, AttributeError):\n output = api_response\n return output\n",
"role": "user"
},
{
"content": "jdout_detailed",
"role": "assistant"
}
] |
def x(self):
try:
recommend_update = False
update_check_resp = requests.get(self.update_info_url, timeout=3)
web_version = update_check_resp.json()['info']['version']
api_logger.debug('RETRIEVED_VERSION: %s', web_version)
available_version = SDK_BUILD_REGEX.search(web_version).groupdict()
current_version = SDK_BUILD_REGEX.search(self.version).groupdict()
available_major = available_version.get('major')
available_minor = available_version.get('minor')
available_patch = available_version.get('patch')
available_build = available_version.get('build')
current_major = current_version.get('major')
current_minor = current_version.get('minor')
current_patch = current_version.get('patch')
current_build = current_version.get('build')
api_logger.debug('AVAILABLE_VERSION: %s', available_version)
api_logger.debug('CURRENT_VERSION: %s', current_version)
if available_major > current_major:
recommend_update = True
elif available_major >= current_major and available_minor > current_minor:
recommend_update = True
elif available_major >= current_major and available_minor >= current_minor and available_patch > current_patch:
recommend_update = True
api_logger.debug('NEED_UPDATE: %s', recommend_update)
if recommend_update:
sys.stderr.write(
"""WARNING: CloudGenix Python SDK upgrade available. SDKs are typically deprecated 6 months after release of a new version.
Latest Version: {0}
Current Version: {1}
For more info, see 'https://github.com/cloudgenix/sdk-python'. Additionally, this message can be suppressed by instantiating the API with API(update_check=False).
"""
.format(web_version, self.version))
return
except Exception:
return
|
notify_for_new_version
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n try:\n recommend_update = False\n update_check_resp = requests.get(self.update_info_url, timeout=3)\n web_version = update_check_resp.json()['info']['version']\n api_logger.debug('RETRIEVED_VERSION: %s', web_version)\n available_version = SDK_BUILD_REGEX.search(web_version).groupdict()\n current_version = SDK_BUILD_REGEX.search(self.version).groupdict()\n available_major = available_version.get('major')\n available_minor = available_version.get('minor')\n available_patch = available_version.get('patch')\n available_build = available_version.get('build')\n current_major = current_version.get('major')\n current_minor = current_version.get('minor')\n current_patch = current_version.get('patch')\n current_build = current_version.get('build')\n api_logger.debug('AVAILABLE_VERSION: %s', available_version)\n api_logger.debug('CURRENT_VERSION: %s', current_version)\n if available_major > current_major:\n recommend_update = True\n elif available_major >= current_major and available_minor > current_minor:\n recommend_update = True\n elif available_major >= current_major and available_minor >= current_minor and available_patch > current_patch:\n recommend_update = True\n api_logger.debug('NEED_UPDATE: %s', recommend_update)\n if recommend_update:\n sys.stderr.write(\n \"\"\"WARNING: CloudGenix Python SDK upgrade available. SDKs are typically deprecated 6 months after release of a new version.\n\tLatest Version: {0}\n\tCurrent Version: {1}\n\tFor more info, see 'https://github.com/cloudgenix/sdk-python'. Additionally, this message can be suppressed by instantiating the API with API(update_check=False).\n\n\"\"\"\n .format(web_version, self.version))\n return\n except Exception:\n return\n",
"role": "user"
},
{
"content": "notify_for_new_version",
"role": "assistant"
}
] |
def x(self, ssl_verify):
self.verify = ssl_verify
if isinstance(self.verify, bool):
if self.verify:
if os.name == 'nt':
self._ca_verify_file_handle = temp_ca_bundle(delete=False)
self._ca_verify_file_handle.write(BYTE_CA_BUNDLE)
self._ca_verify_file_handle.flush()
self.ca_verify_filename = self._ca_verify_file_handle.name
self._ca_verify_file_handle.close()
else:
self._ca_verify_file_handle = temp_ca_bundle()
self._ca_verify_file_handle.write(BYTE_CA_BUNDLE)
self._ca_verify_file_handle.flush()
self.ca_verify_filename = self._ca_verify_file_handle.name
atexit.register(self._cleanup_ca_temp_file)
else:
urllib3.disable_warnings()
self.ca_verify_filename = False
else:
self.ca_verify_filename = self.verify
return
|
ssl_verify
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ssl_verify):\n self.verify = ssl_verify\n if isinstance(self.verify, bool):\n if self.verify:\n if os.name == 'nt':\n self._ca_verify_file_handle = temp_ca_bundle(delete=False)\n self._ca_verify_file_handle.write(BYTE_CA_BUNDLE)\n self._ca_verify_file_handle.flush()\n self.ca_verify_filename = self._ca_verify_file_handle.name\n self._ca_verify_file_handle.close()\n else:\n self._ca_verify_file_handle = temp_ca_bundle()\n self._ca_verify_file_handle.write(BYTE_CA_BUNDLE)\n self._ca_verify_file_handle.flush()\n self.ca_verify_filename = self._ca_verify_file_handle.name\n atexit.register(self._cleanup_ca_temp_file)\n else:\n urllib3.disable_warnings()\n self.ca_verify_filename = False\n else:\n self.ca_verify_filename = self.verify\n return\n",
"role": "user"
},
{
"content": "ssl_verify",
"role": "assistant"
}
] |
def x(self, total=8, connect=None, read=None, redirect=None, status=None,
method_whitelist=urllib3.util.retry.Retry.DEFAULT_METHOD_WHITELIST,
status_forcelist=None, backoff_factor=0.705883, raise_on_redirect=True,
raise_on_status=True, respect_retry_after_header=True, adapter_url=
'https://'):
if status_forcelist is None:
status_forcelist = 413, 429, 502, 503, 504
retry = urllib3.util.retry.Retry(total=total, connect=connect, read=
read, redirect=redirect, status=status, method_whitelist=
method_whitelist, status_forcelist=status_forcelist, backoff_factor
=backoff_factor, raise_on_redirect=raise_on_redirect,
raise_on_status=raise_on_status, respect_retry_after_header=
respect_retry_after_header)
adapter = requests.adapters.HTTPAdapter(max_retries=retry)
self._session.mount(adapter_url, adapter)
return
|
modify_rest_retry
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, total=8, connect=None, read=None, redirect=None, status=None,\n method_whitelist=urllib3.util.retry.Retry.DEFAULT_METHOD_WHITELIST,\n status_forcelist=None, backoff_factor=0.705883, raise_on_redirect=True,\n raise_on_status=True, respect_retry_after_header=True, adapter_url=\n 'https://'):\n if status_forcelist is None:\n status_forcelist = 413, 429, 502, 503, 504\n retry = urllib3.util.retry.Retry(total=total, connect=connect, read=\n read, redirect=redirect, status=status, method_whitelist=\n method_whitelist, status_forcelist=status_forcelist, backoff_factor\n =backoff_factor, raise_on_redirect=raise_on_redirect,\n raise_on_status=raise_on_status, respect_retry_after_header=\n respect_retry_after_header)\n adapter = requests.adapters.HTTPAdapter(max_retries=retry)\n self._session.mount(adapter_url, adapter)\n return\n",
"role": "user"
},
{
"content": "modify_rest_retry",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.