source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(self, album, artist):
params = collections.OrderedDict()
params['gbv'] = '2'
params['q'] = '"%s" "%s" front cover' % (artist, album)
if abs(self.target_size - 500) < 300:
params['tbs'] = 'isz:m'
elif self.target_size > 800:
params['tbs'] = 'isz:l'
return __class__.assembleUrl(__class__.BASE_URL, params)
| getSearchUrl | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, album, artist):\n params = collections.OrderedDict()\n params['gbv'] = '2'\n params['q'] = '\"%s\" \"%s\" front cover' % (artist, album)\n if abs(self.target_size - 500) < 300:\n params['tbs'] = 'isz:m'\n elif self.target_size > 800:\n params['tbs'] = 'isz:l'\n return __class__.assembleUrl(__class__.BASE_URL, params)\n",
"role": "user"
},
{
"content": "getSearchUrl",
"role": "assistant"
}
] |
async def parseResults(self, api_data):
""" See CoverSource.parseResults. """
results = []
parser = lxml.etree.HTMLParser()
html = lxml.etree.XML(api_data.decode('latin-1'), parser)
for rank, result in enumerate(__class__.RESULTS_SELECTOR(html), 1):
metadata_div = result.find('div')
metadata_json = lxml.etree.tostring(metadata_div, encoding=
'unicode', method='text')
metadata_json = json.loads(metadata_json)
google_url = result.find('a').get('href')
if google_url is not None:
query = urllib.parse.urlsplit(google_url).query
else:
query = None
if not query:
img_url = metadata_json['ou']
else:
query = urllib.parse.parse_qs(query)
img_url = query['imgurl'][0]
check_metadata = CoverImageMetadata.NONE
format = metadata_json['ity'].lower()
try:
format = SUPPORTED_IMG_FORMATS[format]
except KeyError:
format = None
check_metadata = CoverImageMetadata.FORMAT
if not query:
size = metadata_json['ow'], metadata_json['oh']
else:
size = tuple(map(int, (query['w'][0], query['h'][0])))
thumbnail_url = metadata_json['tu']
results.append(GoogleImagesCoverSourceResult(img_url, size, format,
thumbnail_url=thumbnail_url, source=self, rank=rank,
check_metadata=check_metadata))
return results
| parseResults | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def parseResults(self, api_data):\n \"\"\" See CoverSource.parseResults. \"\"\"\n results = []\n parser = lxml.etree.HTMLParser()\n html = lxml.etree.XML(api_data.decode('latin-1'), parser)\n for rank, result in enumerate(__class__.RESULTS_SELECTOR(html), 1):\n metadata_div = result.find('div')\n metadata_json = lxml.etree.tostring(metadata_div, encoding=\n 'unicode', method='text')\n metadata_json = json.loads(metadata_json)\n google_url = result.find('a').get('href')\n if google_url is not None:\n query = urllib.parse.urlsplit(google_url).query\n else:\n query = None\n if not query:\n img_url = metadata_json['ou']\n else:\n query = urllib.parse.parse_qs(query)\n img_url = query['imgurl'][0]\n check_metadata = CoverImageMetadata.NONE\n format = metadata_json['ity'].lower()\n try:\n format = SUPPORTED_IMG_FORMATS[format]\n except KeyError:\n format = None\n check_metadata = CoverImageMetadata.FORMAT\n if not query:\n size = metadata_json['ow'], metadata_json['oh']\n else:\n size = tuple(map(int, (query['w'][0], query['h'][0])))\n thumbnail_url = metadata_json['tu']\n results.append(GoogleImagesCoverSourceResult(img_url, size, format,\n thumbnail_url=thumbnail_url, source=self, rank=rank,\n check_metadata=check_metadata))\n return results\n",
"role": "user"
},
{
"content": "parseResults",
"role": "assistant"
}
] |
async def waitAccessAsync(self):
""" Wait the needed time before sending a request to honor rate limit. """
async with self.lock:
while True:
last_access_ts = self.__getLastAccess()
if last_access_ts is not None:
now = time.time()
last_access_ts = last_access_ts[0]
time_since_last_access = now - last_access_ts
if time_since_last_access < self.min_delay_between_accesses:
time_to_wait = (self.min_delay_between_accesses -
time_since_last_access)
if self.jitter_range_ms is not None:
time_to_wait += random.randint(*self.jitter_range_ms
) / 1000
self.logger.debug(
'Sleeping for %.2fms because of rate limit for domain %s'
% (time_to_wait * 1000, self.domain))
await asyncio.sleep(time_to_wait)
access_time = time.time()
self.__access(access_time)
last_access_ts = self.__getLastAccess()
if last_access_ts[0] == access_time:
break
| waitAccessAsync | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def waitAccessAsync(self):\n \"\"\" Wait the needed time before sending a request to honor rate limit. \"\"\"\n async with self.lock:\n while True:\n last_access_ts = self.__getLastAccess()\n if last_access_ts is not None:\n now = time.time()\n last_access_ts = last_access_ts[0]\n time_since_last_access = now - last_access_ts\n if time_since_last_access < self.min_delay_between_accesses:\n time_to_wait = (self.min_delay_between_accesses -\n time_since_last_access)\n if self.jitter_range_ms is not None:\n time_to_wait += random.randint(*self.jitter_range_ms\n ) / 1000\n self.logger.debug(\n 'Sleeping for %.2fms because of rate limit for domain %s'\n % (time_to_wait * 1000, self.domain))\n await asyncio.sleep(time_to_wait)\n access_time = time.time()\n self.__access(access_time)\n last_access_ts = self.__getLastAccess()\n if last_access_ts[0] == access_time:\n break\n",
"role": "user"
},
{
"content": "waitAccessAsync",
"role": "assistant"
}
] |
def x(self, ts):
with self.connection:
self.connection.execute(
'INSERT OR REPLACE INTO access_timestamp (timestamp, domain) VALUES (?, ?)'
, (ts, self.domain))
| access | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ts):\n with self.connection:\n self.connection.execute(\n 'INSERT OR REPLACE INTO access_timestamp (timestamp, domain) VALUES (?, ?)'\n , (ts, self.domain))\n",
"role": "user"
},
{
"content": "access",
"role": "assistant"
}
] |
def x(socket_timeout_s):
return aiohttp.ClientTimeout(total=None, connect=None, sock_connect=
socket_timeout_s, sock_read=socket_timeout_s)
| aiohttp_socket_timeout | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(socket_timeout_s):\n return aiohttp.ClientTimeout(total=None, connect=None, sock_connect=\n socket_timeout_s, sock_read=socket_timeout_s)\n",
"role": "user"
},
{
"content": "aiohttp_socket_timeout",
"role": "assistant"
}
] |
async def query(self, url, *, post_data=None, headers=None, verify=True,
cache=None, pre_cache_callback=None):
""" Send a GET/POST request or get data from cache, retry if it fails, and return a tuple of store in cache callback, response content. """
async def store_in_cache_callback():
pass
if cache is not None:
if post_data is not None:
if (url, post_data) in cache:
self.logger.debug("Got data for URL '%s' %s from cache" % (
url, dict(post_data)))
return store_in_cache_callback, cache[url, post_data]
elif url in cache:
self.logger.debug("Got data for URL '%s' from cache" % url)
return store_in_cache_callback, cache[url]
domain_rate_watcher = rate_watcher.AccessRateWatcher(self.
watcher_db_filepath, url, self.min_delay_between_accesses,
jitter_range_ms=self.jitter_range_ms, logger=self.logger)
for attempt, time_to_sleep in enumerate(redo.retrier(max_attempts=
HTTP_MAX_ATTEMPTS, sleeptime=1, max_sleeptime=
HTTP_MAX_RETRY_SLEEP_S, sleepscale=1.5), 1):
await domain_rate_watcher.waitAccessAsync()
try:
if post_data is not None:
async with self.session.post(url, data=post_data, headers=
self._buildHeaders(headers), timeout=
HTTP_NORMAL_TIMEOUT, ssl=verify) as response:
content = await response.read()
else:
async with self.session.get(url, headers=self._buildHeaders
(headers), timeout=HTTP_NORMAL_TIMEOUT, ssl=verify
) as response:
content = await response.read()
if cache is not None:
async def store_in_cache_callback():
if pre_cache_callback is not None:
try:
data = await pre_cache_callback(content)
except Exception:
data = content
else:
data = content
if post_data is not None:
cache[url, post_data] = data
else:
cache[url] = data
except (asyncio.TimeoutError, aiohttp.ClientError) as e:
self.logger.warning(
"Querying '%s' failed (attempt %u/%u): %s %s" % (url,
attempt, HTTP_MAX_ATTEMPTS, e.__class__.__qualname__, e))
if attempt == HTTP_MAX_ATTEMPTS:
raise
else:
self.logger.debug('Retrying in %.3fs' % time_to_sleep)
await asyncio.sleep(time_to_sleep)
else:
break
response.raise_for_status()
return store_in_cache_callback, content
| query | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def query(self, url, *, post_data=None, headers=None, verify=True,\n cache=None, pre_cache_callback=None):\n \"\"\" Send a GET/POST request or get data from cache, retry if it fails, and return a tuple of store in cache callback, response content. \"\"\"\n\n async def store_in_cache_callback():\n pass\n if cache is not None:\n if post_data is not None:\n if (url, post_data) in cache:\n self.logger.debug(\"Got data for URL '%s' %s from cache\" % (\n url, dict(post_data)))\n return store_in_cache_callback, cache[url, post_data]\n elif url in cache:\n self.logger.debug(\"Got data for URL '%s' from cache\" % url)\n return store_in_cache_callback, cache[url]\n domain_rate_watcher = rate_watcher.AccessRateWatcher(self.\n watcher_db_filepath, url, self.min_delay_between_accesses,\n jitter_range_ms=self.jitter_range_ms, logger=self.logger)\n for attempt, time_to_sleep in enumerate(redo.retrier(max_attempts=\n HTTP_MAX_ATTEMPTS, sleeptime=1, max_sleeptime=\n HTTP_MAX_RETRY_SLEEP_S, sleepscale=1.5), 1):\n await domain_rate_watcher.waitAccessAsync()\n try:\n if post_data is not None:\n async with self.session.post(url, data=post_data, headers=\n self._buildHeaders(headers), timeout=\n HTTP_NORMAL_TIMEOUT, ssl=verify) as response:\n content = await response.read()\n else:\n async with self.session.get(url, headers=self._buildHeaders\n (headers), timeout=HTTP_NORMAL_TIMEOUT, ssl=verify\n ) as response:\n content = await response.read()\n if cache is not None:\n\n async def store_in_cache_callback():\n if pre_cache_callback is not None:\n try:\n data = await pre_cache_callback(content)\n except Exception:\n data = content\n else:\n data = content\n if post_data is not None:\n cache[url, post_data] = data\n else:\n cache[url] = data\n except (asyncio.TimeoutError, aiohttp.ClientError) as e:\n self.logger.warning(\n \"Querying '%s' failed (attempt %u/%u): %s %s\" % (url,\n attempt, HTTP_MAX_ATTEMPTS, e.__class__.__qualname__, e))\n if attempt == HTTP_MAX_ATTEMPTS:\n raise\n else:\n self.logger.debug('Retrying in %.3fs' % time_to_sleep)\n await asyncio.sleep(time_to_sleep)\n else:\n break\n response.raise_for_status()\n return store_in_cache_callback, content\n",
"role": "user"
},
{
"content": "query",
"role": "assistant"
}
] |
async def isReachable(self, url, *, headers=None, verify=True,
response_headers=None, cache=None):
""" Send a HEAD request with short timeout or get data from cache, return True if ressource has 2xx status code, False instead. """
if cache is not None and url in cache:
self.logger.debug("Got headers for URL '%s' from cache" % url)
resp_ok, response_headers = pickle.loads(cache[url])
return resp_ok
domain_rate_watcher = rate_watcher.AccessRateWatcher(self.
watcher_db_filepath, url, self.min_delay_between_accesses,
jitter_range_ms=self.jitter_range_ms, logger=self.logger)
resp_ok = True
try:
for attempt, time_to_sleep in enumerate(redo.retrier(max_attempts=
HTTP_MAX_ATTEMPTS, sleeptime=0.5, max_sleeptime=
HTTP_MAX_RETRY_SLEEP_SHORT_S, sleepscale=1.5), 1):
await domain_rate_watcher.waitAccessAsync()
try:
async with self.session.head(url, headers=self.
_buildHeaders(headers), timeout=HTTP_SHORT_TIMEOUT, ssl
=verify) as response:
pass
except (asyncio.TimeoutError, aiohttp.ClientError) as e:
self.logger.warning(
"Probing '%s' failed (attempt %u/%u): %s %s" % (url,
attempt, HTTP_MAX_ATTEMPTS, e.__class__.__qualname__, e))
if attempt == HTTP_MAX_ATTEMPTS:
resp_ok = False
else:
self.logger.debug('Retrying in %.3fs' % time_to_sleep)
await asyncio.sleep(time_to_sleep)
else:
response.raise_for_status()
if response_headers is not None:
response_headers.update(response.headers)
break
except aiohttp.ClientResponseError as e:
self.logger.debug("Probing '%s' failed: %s %s" % (url, e.__class__.
__qualname__, e))
resp_ok = False
if cache is not None:
cache[url] = pickle.dumps((resp_ok, response_headers))
return resp_ok
| isReachable | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def isReachable(self, url, *, headers=None, verify=True,\n response_headers=None, cache=None):\n \"\"\" Send a HEAD request with short timeout or get data from cache, return True if ressource has 2xx status code, False instead. \"\"\"\n if cache is not None and url in cache:\n self.logger.debug(\"Got headers for URL '%s' from cache\" % url)\n resp_ok, response_headers = pickle.loads(cache[url])\n return resp_ok\n domain_rate_watcher = rate_watcher.AccessRateWatcher(self.\n watcher_db_filepath, url, self.min_delay_between_accesses,\n jitter_range_ms=self.jitter_range_ms, logger=self.logger)\n resp_ok = True\n try:\n for attempt, time_to_sleep in enumerate(redo.retrier(max_attempts=\n HTTP_MAX_ATTEMPTS, sleeptime=0.5, max_sleeptime=\n HTTP_MAX_RETRY_SLEEP_SHORT_S, sleepscale=1.5), 1):\n await domain_rate_watcher.waitAccessAsync()\n try:\n async with self.session.head(url, headers=self.\n _buildHeaders(headers), timeout=HTTP_SHORT_TIMEOUT, ssl\n =verify) as response:\n pass\n except (asyncio.TimeoutError, aiohttp.ClientError) as e:\n self.logger.warning(\n \"Probing '%s' failed (attempt %u/%u): %s %s\" % (url,\n attempt, HTTP_MAX_ATTEMPTS, e.__class__.__qualname__, e))\n if attempt == HTTP_MAX_ATTEMPTS:\n resp_ok = False\n else:\n self.logger.debug('Retrying in %.3fs' % time_to_sleep)\n await asyncio.sleep(time_to_sleep)\n else:\n response.raise_for_status()\n if response_headers is not None:\n response_headers.update(response.headers)\n break\n except aiohttp.ClientResponseError as e:\n self.logger.debug(\"Probing '%s' failed: %s %s\" % (url, e.__class__.\n __qualname__, e))\n resp_ok = False\n if cache is not None:\n cache[url] = pickle.dumps((resp_ok, response_headers))\n return resp_ok\n",
"role": "user"
},
{
"content": "isReachable",
"role": "assistant"
}
] |
async def fastStreamedQuery(self, url, *, headers=None, verify=True):
""" Send a GET request with short timeout, do not retry, and return streamed response. """
response = await self.session.get(url, headers=self._buildHeaders(
headers), timeout=HTTP_SHORT_TIMEOUT, ssl=verify)
response.raise_for_status()
return response
| fastStreamedQuery | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def fastStreamedQuery(self, url, *, headers=None, verify=True):\n \"\"\" Send a GET request with short timeout, do not retry, and return streamed response. \"\"\"\n response = await self.session.get(url, headers=self._buildHeaders(\n headers), timeout=HTTP_SHORT_TIMEOUT, ssl=verify)\n response.raise_for_status()\n return response\n",
"role": "user"
},
{
"content": "fastStreamedQuery",
"role": "assistant"
}
] |
def x(self, album, artist):
params = collections.OrderedDict()
params['method'] = 'album.getinfo'
params['api_key'] = __class__.API_KEY
params['album'] = album
params['artist'] = artist
return __class__.assembleUrl(__class__.BASE_URL, params)
| getSearchUrl | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, album, artist):\n params = collections.OrderedDict()\n params['method'] = 'album.getinfo'\n params['api_key'] = __class__.API_KEY\n params['album'] = album\n params['artist'] = artist\n return __class__.assembleUrl(__class__.BASE_URL, params)\n",
"role": "user"
},
{
"content": "getSearchUrl",
"role": "assistant"
}
] |
def x(self, s):
char_blacklist = set(string.punctuation)
char_blacklist.remove("'")
char_blacklist.remove('&')
char_blacklist = frozenset(char_blacklist)
return __class__.unpunctuate(s.lower(), char_blacklist=char_blacklist)
| processQueryString | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, s):\n char_blacklist = set(string.punctuation)\n char_blacklist.remove(\"'\")\n char_blacklist.remove('&')\n char_blacklist = frozenset(char_blacklist)\n return __class__.unpunctuate(s.lower(), char_blacklist=char_blacklist)\n",
"role": "user"
},
{
"content": "processQueryString",
"role": "assistant"
}
] |
async def parseResults(self, api_data):
""" See CoverSource.parseResults. """
results = []
xml_text = api_data.decode('utf-8')
xml_root = xml.etree.ElementTree.fromstring(xml_text)
status = xml_root.get('status')
if status != 'ok':
raise Exception('Unexpected Last.fm response status: %s' % status)
img_elements = xml_root.findall('album/image')
thumbnail_url = None
thumbnail_size = None
for img_element in img_elements:
img_url = img_element.text
if not img_url:
continue
lfm_size = img_element.get('size')
if lfm_size == 'mega':
check_metadata = CoverImageMetadata.SIZE
else:
check_metadata = CoverImageMetadata.NONE
try:
size = __class__.SIZES[lfm_size]
except KeyError:
continue
if size[0] <= MAX_THUMBNAIL_SIZE and (thumbnail_size is None or
size[0] < thumbnail_size):
thumbnail_url = img_url
thumbnail_size = size[0]
format = os.path.splitext(img_url)[1][1:].lower()
format = SUPPORTED_IMG_FORMATS[format]
results.append(LastFmCoverSourceResult(img_url, size, format,
thumbnail_url=thumbnail_url, source=self, check_metadata=
check_metadata))
return results
| parseResults | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def parseResults(self, api_data):\n \"\"\" See CoverSource.parseResults. \"\"\"\n results = []\n xml_text = api_data.decode('utf-8')\n xml_root = xml.etree.ElementTree.fromstring(xml_text)\n status = xml_root.get('status')\n if status != 'ok':\n raise Exception('Unexpected Last.fm response status: %s' % status)\n img_elements = xml_root.findall('album/image')\n thumbnail_url = None\n thumbnail_size = None\n for img_element in img_elements:\n img_url = img_element.text\n if not img_url:\n continue\n lfm_size = img_element.get('size')\n if lfm_size == 'mega':\n check_metadata = CoverImageMetadata.SIZE\n else:\n check_metadata = CoverImageMetadata.NONE\n try:\n size = __class__.SIZES[lfm_size]\n except KeyError:\n continue\n if size[0] <= MAX_THUMBNAIL_SIZE and (thumbnail_size is None or \n size[0] < thumbnail_size):\n thumbnail_url = img_url\n thumbnail_size = size[0]\n format = os.path.splitext(img_url)[1][1:].lower()\n format = SUPPORTED_IMG_FORMATS[format]\n results.append(LastFmCoverSourceResult(img_url, size, format,\n thumbnail_url=thumbnail_url, source=self, check_metadata=\n check_metadata))\n return results\n",
"role": "user"
},
{
"content": "parseResults",
"role": "assistant"
}
] |
async def search_and_download(album, artist, format, size, out_filepath, *,
size_tolerance_prct, amazon_tlds, no_lq_sources, async_loop):
""" Search and download a cover, return True if success, False instead. """
source_args = size, size_tolerance_prct
cover_sources = [sources.LastFmCoverSource(*source_args), sources.
AmazonCdCoverSource(*source_args), sources.AmazonDigitalCoverSource
(*source_args)]
for tld in amazon_tlds:
cover_sources.append(sources.AmazonCdCoverSource(*source_args, tld=tld)
)
if not no_lq_sources:
cover_sources.append(sources.GoogleImagesWebScrapeCoverSource(*
source_args))
search_futures = []
for cover_source in cover_sources:
coroutine = cover_source.search(album, artist)
future = asyncio.ensure_future(coroutine, loop=async_loop)
search_futures.append(future)
await asyncio.wait(search_futures, loop=async_loop)
results = []
for future in search_futures:
source_results = future.result()
results.extend(source_results)
results = await CoverSourceResult.preProcessForComparison(results, size,
size_tolerance_prct)
results.sort(reverse=True, key=functools.cmp_to_key(functools.partial(
CoverSourceResult.compare, target_size=size, size_tolerance_prct=
size_tolerance_prct)))
if not results:
logging.getLogger('Main').info('No results')
for result in results:
try:
await result.get(format, size, size_tolerance_prct, out_filepath)
except Exception as e:
logging.getLogger('Main').warning(
'Download of %s failed: %s %s' % (result, e.__class__.
__qualname__, e))
continue
else:
return True
return False
| search_and_download | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def search_and_download(album, artist, format, size, out_filepath, *,\n size_tolerance_prct, amazon_tlds, no_lq_sources, async_loop):\n \"\"\" Search and download a cover, return True if success, False instead. \"\"\"\n source_args = size, size_tolerance_prct\n cover_sources = [sources.LastFmCoverSource(*source_args), sources.\n AmazonCdCoverSource(*source_args), sources.AmazonDigitalCoverSource\n (*source_args)]\n for tld in amazon_tlds:\n cover_sources.append(sources.AmazonCdCoverSource(*source_args, tld=tld)\n )\n if not no_lq_sources:\n cover_sources.append(sources.GoogleImagesWebScrapeCoverSource(*\n source_args))\n search_futures = []\n for cover_source in cover_sources:\n coroutine = cover_source.search(album, artist)\n future = asyncio.ensure_future(coroutine, loop=async_loop)\n search_futures.append(future)\n await asyncio.wait(search_futures, loop=async_loop)\n results = []\n for future in search_futures:\n source_results = future.result()\n results.extend(source_results)\n results = await CoverSourceResult.preProcessForComparison(results, size,\n size_tolerance_prct)\n results.sort(reverse=True, key=functools.cmp_to_key(functools.partial(\n CoverSourceResult.compare, target_size=size, size_tolerance_prct=\n size_tolerance_prct)))\n if not results:\n logging.getLogger('Main').info('No results')\n for result in results:\n try:\n await result.get(format, size, size_tolerance_prct, out_filepath)\n except Exception as e:\n logging.getLogger('Main').warning(\n 'Download of %s failed: %s %s' % (result, e.__class__.\n __qualname__, e))\n continue\n else:\n return True\n return False\n",
"role": "user"
},
{
"content": "search_and_download",
"role": "assistant"
}
] |
def x(self, album, artist):
url = '%s/search' % __class__.BASE_URL
params = collections.OrderedDict()
params['search-alias'] = 'digital-music'
params['field-keywords'] = ' '.join((artist, album))
params['sort'] = 'relevancerank'
return __class__.assembleUrl(url, params)
| getSearchUrl | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, album, artist):\n url = '%s/search' % __class__.BASE_URL\n params = collections.OrderedDict()\n params['search-alias'] = 'digital-music'\n params['field-keywords'] = ' '.join((artist, album))\n params['sort'] = 'relevancerank'\n return __class__.assembleUrl(url, params)\n",
"role": "user"
},
{
"content": "getSearchUrl",
"role": "assistant"
}
] |
async def parseResults(self, api_data):
""" See CoverSource.parseResults. """
results = []
parser = lxml.etree.HTMLParser()
html = lxml.etree.XML(api_data.decode('utf-8'), parser)
for page_struct_version, result_selector in enumerate(__class__.
RESULTS_SELECTORS):
result_nodes = result_selector(html)
if result_nodes:
break
for rank, result_node in enumerate(result_nodes, 1):
img_node = __class__.IMG_SELECTORS[page_struct_version](result_node)[0]
thumbnail_url = img_node.get('src')
thumbnail_url = thumbnail_url.replace('Stripe-Prime-Only', '')
url_parts = thumbnail_url.rsplit('.', 2)
img_url = '.'.join((url_parts[0], url_parts[2]))
size = 500, 500
if self.target_size > size[0]:
self.logger.debug('Looking for optimal subimages configuration...')
product_url = __class__.LINK_SELECTOR(result_node)[0].get('href')
product_url = urllib.parse.urlsplit(product_url)
product_id = product_url.path.split('/')[3]
for amazon_img_format in AMAZON_DIGITAL_IMAGE_FORMATS:
self.logger.debug('Trying %u subimages...' %
amazon_img_format.slice_count ** 2)
urls = tuple(self.generateImgUrls(product_id, __class__.
DYNAPI_KEY, amazon_img_format.id, amazon_img_format.
slice_count))
url_ok = await self.probeUrl(urls[-1])
if not url_ok:
continue
img_url = urls
size = (amazon_img_format.total_res,) * 2
break
format = CoverImageFormat.JPEG
results.append(AmazonDigitalCoverSourceResult(img_url, size, format,
thumbnail_url=thumbnail_url, source=self, rank=rank,
check_metadata=CoverImageMetadata.SIZE))
return results
| parseResults | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def parseResults(self, api_data):\n \"\"\" See CoverSource.parseResults. \"\"\"\n results = []\n parser = lxml.etree.HTMLParser()\n html = lxml.etree.XML(api_data.decode('utf-8'), parser)\n for page_struct_version, result_selector in enumerate(__class__.\n RESULTS_SELECTORS):\n result_nodes = result_selector(html)\n if result_nodes:\n break\n for rank, result_node in enumerate(result_nodes, 1):\n img_node = __class__.IMG_SELECTORS[page_struct_version](result_node)[0]\n thumbnail_url = img_node.get('src')\n thumbnail_url = thumbnail_url.replace('Stripe-Prime-Only', '')\n url_parts = thumbnail_url.rsplit('.', 2)\n img_url = '.'.join((url_parts[0], url_parts[2]))\n size = 500, 500\n if self.target_size > size[0]:\n self.logger.debug('Looking for optimal subimages configuration...')\n product_url = __class__.LINK_SELECTOR(result_node)[0].get('href')\n product_url = urllib.parse.urlsplit(product_url)\n product_id = product_url.path.split('/')[3]\n for amazon_img_format in AMAZON_DIGITAL_IMAGE_FORMATS:\n self.logger.debug('Trying %u subimages...' % \n amazon_img_format.slice_count ** 2)\n urls = tuple(self.generateImgUrls(product_id, __class__.\n DYNAPI_KEY, amazon_img_format.id, amazon_img_format.\n slice_count))\n url_ok = await self.probeUrl(urls[-1])\n if not url_ok:\n continue\n img_url = urls\n size = (amazon_img_format.total_res,) * 2\n break\n format = CoverImageFormat.JPEG\n results.append(AmazonDigitalCoverSourceResult(img_url, size, format,\n thumbnail_url=thumbnail_url, source=self, rank=rank,\n check_metadata=CoverImageMetadata.SIZE))\n return results\n",
"role": "user"
},
{
"content": "parseResults",
"role": "assistant"
}
] |
def x(self, product_id, dynapi_key, format_id, slice_count):
for x in range(slice_count):
for y in range(slice_count):
yield 'http://z2-ec2.images-amazon.com/R/1/a=' + product_id + '+c=' + dynapi_key + '+d=_SCR%28' + str(
format_id) + ',' + str(x) + ',' + str(y) + '%29_=.jpg'
| generateImgUrls | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, product_id, dynapi_key, format_id, slice_count):\n for x in range(slice_count):\n for y in range(slice_count):\n yield 'http://z2-ec2.images-amazon.com/R/1/a=' + product_id + '+c=' + dynapi_key + '+d=_SCR%28' + str(\n format_id) + ',' + str(x) + ',' + str(y) + '%29_=.jpg'\n",
"role": "user"
},
{
"content": "generateImgUrls",
"role": "assistant"
}
] |
def x(*, max_attempts, sleeptime, max_sleeptime, sleepscale=1.5, jitter=0.2):
assert max_attempts > 1
assert sleeptime >= 0
assert 0 <= jitter <= sleeptime
assert sleepscale >= 1
cur_sleeptime = min(max_sleeptime, sleeptime)
for attempt in range(max_attempts):
cur_jitter = random.randint(int(-jitter * 1000), int(jitter * 1000)
) / 1000
yield max(0, cur_sleeptime + cur_jitter)
cur_sleeptime = min(max_sleeptime, cur_sleeptime * sleepscale)
| retrier | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(*, max_attempts, sleeptime, max_sleeptime, sleepscale=1.5, jitter=0.2):\n assert max_attempts > 1\n assert sleeptime >= 0\n assert 0 <= jitter <= sleeptime\n assert sleepscale >= 1\n cur_sleeptime = min(max_sleeptime, sleeptime)\n for attempt in range(max_attempts):\n cur_jitter = random.randint(int(-jitter * 1000), int(jitter * 1000)\n ) / 1000\n yield max(0, cur_sleeptime + cur_jitter)\n cur_sleeptime = min(max_sleeptime, cur_sleeptime * sleepscale)\n",
"role": "user"
},
{
"content": "retrier",
"role": "assistant"
}
] |
async def get(self, target_format, target_size, size_tolerance_prct,
out_filepath):
""" Download cover and process it. """
if self.source_quality.value <= CoverSourceQuality.LOW.value:
logging.getLogger('Cover').warning(
'Cover is from a potentially unreliable source and may be unrelated to the search'
)
images_data = []
for i, url in enumerate(self.urls):
logging.getLogger('Cover').info(
"Downloading cover '%s' (part %u/%u)..." % (url, i + 1, len(
self.urls)))
headers = {}
self.source.updateHttpHeaders(headers)
async def pre_cache_callback(img_data):
return await __class__.crunch(img_data, self.format)
store_in_cache_callback, image_data = await self.source.http.query(url,
headers=headers, verify=False, cache=__class__.image_cache,
pre_cache_callback=pre_cache_callback)
await store_in_cache_callback()
images_data.append(image_data)
need_format_change = self.format != target_format
need_size_change = max(self.size) > target_size and abs(max(self.size) -
target_size) > target_size * size_tolerance_prct / 100
need_join = len(images_data) > 1
if need_join or need_format_change or need_size_change:
image_data = self.postProcess(images_data, target_format if
need_format_change else None, target_size if need_size_change else
None)
image_data = await __class__.crunch(image_data, target_format)
with open(out_filepath, 'wb') as file:
file.write(image_data)
| get | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def get(self, target_format, target_size, size_tolerance_prct,\n out_filepath):\n \"\"\" Download cover and process it. \"\"\"\n if self.source_quality.value <= CoverSourceQuality.LOW.value:\n logging.getLogger('Cover').warning(\n 'Cover is from a potentially unreliable source and may be unrelated to the search'\n )\n images_data = []\n for i, url in enumerate(self.urls):\n logging.getLogger('Cover').info(\n \"Downloading cover '%s' (part %u/%u)...\" % (url, i + 1, len(\n self.urls)))\n headers = {}\n self.source.updateHttpHeaders(headers)\n\n async def pre_cache_callback(img_data):\n return await __class__.crunch(img_data, self.format)\n store_in_cache_callback, image_data = await self.source.http.query(url,\n headers=headers, verify=False, cache=__class__.image_cache,\n pre_cache_callback=pre_cache_callback)\n await store_in_cache_callback()\n images_data.append(image_data)\n need_format_change = self.format != target_format\n need_size_change = max(self.size) > target_size and abs(max(self.size) -\n target_size) > target_size * size_tolerance_prct / 100\n need_join = len(images_data) > 1\n if need_join or need_format_change or need_size_change:\n image_data = self.postProcess(images_data, target_format if\n need_format_change else None, target_size if need_size_change else\n None)\n image_data = await __class__.crunch(image_data, target_format)\n with open(out_filepath, 'wb') as file:\n file.write(image_data)\n",
"role": "user"
},
{
"content": "get",
"role": "assistant"
}
] |
def x(self, images_data, new_format, new_size):
if len(images_data) == 1:
in_bytes = io.BytesIO(images_data[0])
img = PIL.Image.open(in_bytes)
if img.mode != 'RGB':
img = img.convert('RGB')
else:
logging.getLogger('Cover').info('Joining %u images...' % len(
images_data))
new_img = PIL.Image.new('RGB', self.size)
assert is_square(len(images_data))
sq = int(math.sqrt(len(images_data)))
images_data_it = iter(images_data)
img_sizes = {}
for x in range(sq):
for y in range(sq):
current_image_data = next(images_data_it)
img_stream = io.BytesIO(current_image_data)
img = PIL.Image.open(img_stream)
img_sizes[x, y] = img.size
box = [0, 0]
if x > 0:
for px in range(x):
box[0] += img_sizes[px, y][0]
if y > 0:
for py in range(y):
box[1] += img_sizes[x, py][1]
box.extend((box[0] + img.size[0], box[1] + img.size[1]))
new_img.paste(img, box=tuple(box))
img = new_img
out_bytes = io.BytesIO()
if new_size is not None:
logging.getLogger('Cover').info('Resizing from %ux%u to %ux%u...' %
(self.size[0], self.size[1], new_size, new_size))
img = img.resize((new_size, new_size), PIL.Image.LANCZOS)
unsharper = PIL.ImageFilter.UnsharpMask(radius=1.5, percent=70,
threshold=5)
img = img.filter(unsharper)
if new_format is not None:
logging.getLogger('Cover').info('Converting to %s...' % new_format.
name.upper())
target_format = new_format
else:
target_format = self.format
img.save(out_bytes, format=target_format.name, quality=90, optimize=True)
return out_bytes.getvalue()
| postProcess | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, images_data, new_format, new_size):\n if len(images_data) == 1:\n in_bytes = io.BytesIO(images_data[0])\n img = PIL.Image.open(in_bytes)\n if img.mode != 'RGB':\n img = img.convert('RGB')\n else:\n logging.getLogger('Cover').info('Joining %u images...' % len(\n images_data))\n new_img = PIL.Image.new('RGB', self.size)\n assert is_square(len(images_data))\n sq = int(math.sqrt(len(images_data)))\n images_data_it = iter(images_data)\n img_sizes = {}\n for x in range(sq):\n for y in range(sq):\n current_image_data = next(images_data_it)\n img_stream = io.BytesIO(current_image_data)\n img = PIL.Image.open(img_stream)\n img_sizes[x, y] = img.size\n box = [0, 0]\n if x > 0:\n for px in range(x):\n box[0] += img_sizes[px, y][0]\n if y > 0:\n for py in range(y):\n box[1] += img_sizes[x, py][1]\n box.extend((box[0] + img.size[0], box[1] + img.size[1]))\n new_img.paste(img, box=tuple(box))\n img = new_img\n out_bytes = io.BytesIO()\n if new_size is not None:\n logging.getLogger('Cover').info('Resizing from %ux%u to %ux%u...' %\n (self.size[0], self.size[1], new_size, new_size))\n img = img.resize((new_size, new_size), PIL.Image.LANCZOS)\n unsharper = PIL.ImageFilter.UnsharpMask(radius=1.5, percent=70,\n threshold=5)\n img = img.filter(unsharper)\n if new_format is not None:\n logging.getLogger('Cover').info('Converting to %s...' % new_format.\n name.upper())\n target_format = new_format\n else:\n target_format = self.format\n img.save(out_bytes, format=target_format.name, quality=90, optimize=True)\n return out_bytes.getvalue()\n",
"role": "user"
},
{
"content": "postProcess",
"role": "assistant"
}
] |
async def updateImageMetadata(self):
""" Partially download image file(s) to get its real metadata, or get it from cache. """
assert self.needMetadataUpdate()
width_sum, height_sum = 0, 0
idxs = []
assert is_square(len(self.urls))
sq = int(math.sqrt(len(self.urls)))
for x in range(sq):
for y in range(sq):
if x == y:
idxs.append((x * sq + y, x, y))
for idx, x, y in idxs:
url = self.urls[idx]
format, width, height = None, None, None
try:
format, width, height = pickle.loads(__class__.metadata_cache[url])
except KeyError:
pass
except Exception as e:
logging.getLogger('Cover').warning(
"Unable to load metadata for URL '%s' from cache: %s %s" %
(url, e.__class__.__qualname__, e))
else:
logging.getLogger('Cover').debug(
"Got metadata for URL '%s' from cache" % url)
if format is not None:
self.setFormatMetadata(format)
if self.needMetadataUpdate(CoverImageMetadata.FORMAT
) or self.needMetadataUpdate(CoverImageMetadata.SIZE) and (
width is None or height is None):
logging.getLogger('Cover').debug(
"Downloading file header for URL '%s'..." % url)
try:
headers = {}
self.source.updateHttpHeaders(headers)
response = await self.source.http.fastStreamedQuery(url,
headers=headers, verify=False)
try:
if self.needMetadataUpdate(CoverImageMetadata.FORMAT):
format = __class__.guessImageFormatFromHttpResponse(
response)
if format is not None:
self.setFormatMetadata(format)
if self.needMetadataUpdate():
metadata = (await __class__.
guessImageMetadataFromHttpData(response))
if metadata is not None:
format, width, height = metadata
if format is not None:
self.setFormatMetadata(format)
finally:
response.release()
except Exception as e:
logging.getLogger('Cover').warning(
"Failed to get file metadata for URL '%s' (%s %s)" % (
url, e.__class__.__qualname__, e))
if self.needMetadataUpdate():
if self.format is None or self.size is None and (width is
None or height is None):
logging.getLogger('Cover').debug(
"Unable to get file metadata from file or HTTP headers for URL '%s', skipping this result"
% url)
return
if self.format is not None and (self.size is not None and
width is None and height is None):
logging.getLogger('Cover').debug(
"Unable to get file metadata from file or HTTP headers for URL '%s', falling back to API data"
% url)
self.check_metadata = CoverImageMetadata.NONE
self.reliable_metadata = False
return
__class__.metadata_cache[url] = pickle.dumps((format, width,
height))
if width is not None and height is not None:
width_sum += width
height_sum += height
if self.needMetadataUpdate(CoverImageMetadata.SIZE
) and width_sum > 0 and height_sum > 0:
self.setSizeMetadata((width_sum, height_sum))
| updateImageMetadata | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def updateImageMetadata(self):\n \"\"\" Partially download image file(s) to get its real metadata, or get it from cache. \"\"\"\n assert self.needMetadataUpdate()\n width_sum, height_sum = 0, 0\n idxs = []\n assert is_square(len(self.urls))\n sq = int(math.sqrt(len(self.urls)))\n for x in range(sq):\n for y in range(sq):\n if x == y:\n idxs.append((x * sq + y, x, y))\n for idx, x, y in idxs:\n url = self.urls[idx]\n format, width, height = None, None, None\n try:\n format, width, height = pickle.loads(__class__.metadata_cache[url])\n except KeyError:\n pass\n except Exception as e:\n logging.getLogger('Cover').warning(\n \"Unable to load metadata for URL '%s' from cache: %s %s\" %\n (url, e.__class__.__qualname__, e))\n else:\n logging.getLogger('Cover').debug(\n \"Got metadata for URL '%s' from cache\" % url)\n if format is not None:\n self.setFormatMetadata(format)\n if self.needMetadataUpdate(CoverImageMetadata.FORMAT\n ) or self.needMetadataUpdate(CoverImageMetadata.SIZE) and (\n width is None or height is None):\n logging.getLogger('Cover').debug(\n \"Downloading file header for URL '%s'...\" % url)\n try:\n headers = {}\n self.source.updateHttpHeaders(headers)\n response = await self.source.http.fastStreamedQuery(url,\n headers=headers, verify=False)\n try:\n if self.needMetadataUpdate(CoverImageMetadata.FORMAT):\n format = __class__.guessImageFormatFromHttpResponse(\n response)\n if format is not None:\n self.setFormatMetadata(format)\n if self.needMetadataUpdate():\n metadata = (await __class__.\n guessImageMetadataFromHttpData(response))\n if metadata is not None:\n format, width, height = metadata\n if format is not None:\n self.setFormatMetadata(format)\n finally:\n response.release()\n except Exception as e:\n logging.getLogger('Cover').warning(\n \"Failed to get file metadata for URL '%s' (%s %s)\" % (\n url, e.__class__.__qualname__, e))\n if self.needMetadataUpdate():\n if self.format is None or self.size is None and (width is\n None or height is None):\n logging.getLogger('Cover').debug(\n \"Unable to get file metadata from file or HTTP headers for URL '%s', skipping this result\"\n % url)\n return\n if self.format is not None and (self.size is not None and \n width is None and height is None):\n logging.getLogger('Cover').debug(\n \"Unable to get file metadata from file or HTTP headers for URL '%s', falling back to API data\"\n % url)\n self.check_metadata = CoverImageMetadata.NONE\n self.reliable_metadata = False\n return\n __class__.metadata_cache[url] = pickle.dumps((format, width,\n height))\n if width is not None and height is not None:\n width_sum += width\n height_sum += height\n if self.needMetadataUpdate(CoverImageMetadata.SIZE\n ) and width_sum > 0 and height_sum > 0:\n self.setSizeMetadata((width_sum, height_sum))\n",
"role": "user"
},
{
"content": "updateImageMetadata",
"role": "assistant"
}
] |
def x(self, format):
assert self.needMetadataUpdate(CoverImageMetadata.FORMAT
) or self.format is format
self.format = format
self.check_metadata &= ~CoverImageMetadata.FORMAT
| setFormatMetadata | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, format):\n assert self.needMetadataUpdate(CoverImageMetadata.FORMAT\n ) or self.format is format\n self.format = format\n self.check_metadata &= ~CoverImageMetadata.FORMAT\n",
"role": "user"
},
{
"content": "setFormatMetadata",
"role": "assistant"
}
] |
def x(self, size):
assert self.needMetadataUpdate(CoverImageMetadata.SIZE
) or self.size == size
self.size = size
self.check_metadata &= ~CoverImageMetadata.SIZE
| setSizeMetadata | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, size):\n assert self.needMetadataUpdate(CoverImageMetadata.SIZE\n ) or self.size == size\n self.size = size\n self.check_metadata &= ~CoverImageMetadata.SIZE\n",
"role": "user"
},
{
"content": "setSizeMetadata",
"role": "assistant"
}
] |
async def updateSignature(self):
""" Calculate a cover's "signature" using its thumbnail url. """
assert self.thumbnail_sig is None
if self.thumbnail_url is None:
logging.getLogger('Cover').warning('No thumbnail available for %s' %
self)
return
logging.getLogger('Cover').debug("Downloading cover thumbnail '%s'..." %
self.thumbnail_url)
headers = {}
self.source.updateHttpHeaders(headers)
async def pre_cache_callback(img_data):
return await __class__.crunch(img_data, CoverImageFormat.JPEG,
silent=True)
try:
store_in_cache_callback, image_data = await self.source.http.query(self
.thumbnail_url, cache=__class__.image_cache, headers=headers,
pre_cache_callback=pre_cache_callback)
except Exception as e:
logging.getLogger('Cover').warning("Download of '%s' failed: %s %s" %
(self.thumbnail_url, e.__class__.__qualname__, e))
return
logging.getLogger('Cover').debug('Computing signature of %s...' % self)
try:
self.thumbnail_sig = __class__.computeImgSignature(image_data)
except Exception as e:
logging.getLogger('Cover').warning(
"Failed to compute signature of '%s': %s %s" % (self, e.
__class__.__qualname__, e))
else:
await store_in_cache_callback()
| updateSignature | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def updateSignature(self):\n \"\"\" Calculate a cover's \"signature\" using its thumbnail url. \"\"\"\n assert self.thumbnail_sig is None\n if self.thumbnail_url is None:\n logging.getLogger('Cover').warning('No thumbnail available for %s' %\n self)\n return\n logging.getLogger('Cover').debug(\"Downloading cover thumbnail '%s'...\" %\n self.thumbnail_url)\n headers = {}\n self.source.updateHttpHeaders(headers)\n\n async def pre_cache_callback(img_data):\n return await __class__.crunch(img_data, CoverImageFormat.JPEG,\n silent=True)\n try:\n store_in_cache_callback, image_data = await self.source.http.query(self\n .thumbnail_url, cache=__class__.image_cache, headers=headers,\n pre_cache_callback=pre_cache_callback)\n except Exception as e:\n logging.getLogger('Cover').warning(\"Download of '%s' failed: %s %s\" %\n (self.thumbnail_url, e.__class__.__qualname__, e))\n return\n logging.getLogger('Cover').debug('Computing signature of %s...' % self)\n try:\n self.thumbnail_sig = __class__.computeImgSignature(image_data)\n except Exception as e:\n logging.getLogger('Cover').warning(\n \"Failed to compute signature of '%s': %s %s\" % (self, e.\n __class__.__qualname__, e))\n else:\n await store_in_cache_callback()\n",
"role": "user"
},
{
"content": "updateSignature",
"role": "assistant"
}
] |
def x(first, second, *, target_size, size_tolerance_prct):
for c in (first, second):
assert c.format is not None
assert isinstance(c.size[0], int) and isinstance(c.size[1], int)
delta_ratio1 = abs(first.size[0] / first.size[1] - 1)
delta_ratio2 = abs(second.size[0] / second.size[1] - 1)
if abs(delta_ratio1 - delta_ratio2) > 0.15:
return -1 if delta_ratio1 > delta_ratio2 else 1
sr1 = first.is_similar_to_reference
sr2 = second.is_similar_to_reference
if sr1 and not sr2:
return 1
if not sr1 and sr2:
return -1
delta_size1 = (first.size[0] + first.size[1]) / 2 - target_size
delta_size2 = (second.size[0] + second.size[1]) / 2 - target_size
if (delta_size1 < 0 and delta_size2 >= 0 or delta_size1 >= 0 and
delta_size2 < 0):
return -1 if delta_size1 < delta_size2 else 1
if delta_size1 < 0 and delta_size2 < 0 and delta_size1 != delta_size2:
return -1 if delta_size1 < delta_size2 else 1
qs1 = first.source_quality.value
qs2 = second.source_quality.value
if qs1 != qs2:
return -1 if qs1 < qs2 else 1
if (first.rank is not None and second.rank is not None and first.
__class__ is second.__class__ and first.rank != second.rank):
return -1 if first.rank > second.rank else 1
if first.reliable_metadata != second.reliable_metadata:
return 1 if first.reliable_metadata else -1
ic1 = len(first.urls)
ic2 = len(second.urls)
if ic1 != ic2:
return -1 if ic1 > ic2 else 1
if abs(delta_size1) != abs(delta_size2):
return -1 if abs(delta_size1) > abs(delta_size2) else 1
if first.format != second.format:
return -1 if second.format is CoverImageFormat.PNG else 1
if delta_ratio1 != delta_ratio2:
return -1 if delta_ratio1 > delta_ratio2 else 1
return 0
| compare | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(first, second, *, target_size, size_tolerance_prct):\n for c in (first, second):\n assert c.format is not None\n assert isinstance(c.size[0], int) and isinstance(c.size[1], int)\n delta_ratio1 = abs(first.size[0] / first.size[1] - 1)\n delta_ratio2 = abs(second.size[0] / second.size[1] - 1)\n if abs(delta_ratio1 - delta_ratio2) > 0.15:\n return -1 if delta_ratio1 > delta_ratio2 else 1\n sr1 = first.is_similar_to_reference\n sr2 = second.is_similar_to_reference\n if sr1 and not sr2:\n return 1\n if not sr1 and sr2:\n return -1\n delta_size1 = (first.size[0] + first.size[1]) / 2 - target_size\n delta_size2 = (second.size[0] + second.size[1]) / 2 - target_size\n if (delta_size1 < 0 and delta_size2 >= 0 or delta_size1 >= 0 and \n delta_size2 < 0):\n return -1 if delta_size1 < delta_size2 else 1\n if delta_size1 < 0 and delta_size2 < 0 and delta_size1 != delta_size2:\n return -1 if delta_size1 < delta_size2 else 1\n qs1 = first.source_quality.value\n qs2 = second.source_quality.value\n if qs1 != qs2:\n return -1 if qs1 < qs2 else 1\n if (first.rank is not None and second.rank is not None and first.\n __class__ is second.__class__ and first.rank != second.rank):\n return -1 if first.rank > second.rank else 1\n if first.reliable_metadata != second.reliable_metadata:\n return 1 if first.reliable_metadata else -1\n ic1 = len(first.urls)\n ic2 = len(second.urls)\n if ic1 != ic2:\n return -1 if ic1 > ic2 else 1\n if abs(delta_size1) != abs(delta_size2):\n return -1 if abs(delta_size1) > abs(delta_size2) else 1\n if first.format != second.format:\n return -1 if second.format is CoverImageFormat.PNG else 1\n if delta_ratio1 != delta_ratio2:\n return -1 if delta_ratio1 > delta_ratio2 else 1\n return 0\n",
"role": "user"
},
{
"content": "compare",
"role": "assistant"
}
] |
async def crunch(image_data, format, silent=False):
""" Crunch image data, and return the processed data, or orignal data if operation failed. """
if (format is CoverImageFormat.PNG and not HAS_OPTIPNG or format is
CoverImageFormat.JPEG and not HAS_JPEGOPTIM):
return image_data
with mkstemp_ctx.mkstemp(suffix='.%s' % format.name.lower()
) as tmp_out_filepath:
if not silent:
logging.getLogger('Cover').info('Crunching %s image...' %
format.name.upper())
with open(tmp_out_filepath, 'wb') as tmp_out_file:
tmp_out_file.write(image_data)
size_before = len(image_data)
if format is CoverImageFormat.PNG:
cmd = ['optipng', '-quiet', '-o1']
elif format is CoverImageFormat.JPEG:
cmd = ['jpegoptim', '-q', '--strip-all']
cmd.append(tmp_out_filepath)
p = await asyncio.create_subprocess_exec(*cmd, stdin=asyncio.
subprocess.DEVNULL, stdout=asyncio.subprocess.DEVNULL, stderr=
asyncio.subprocess.DEVNULL)
await p.wait()
if p.returncode != 0:
if not silent:
logging.getLogger('Cover').warning('Crunching image failed')
return image_data
with open(tmp_out_filepath, 'rb') as tmp_out_file:
crunched_image_data = tmp_out_file.read()
size_after = len(crunched_image_data)
pct_saved = 100 * (size_before - size_after) / size_before
if not silent:
logging.getLogger('Cover').debug(
'Crunching image saved %.2f%% filesize' % pct_saved)
return crunched_image_data
| crunch | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def crunch(image_data, format, silent=False):\n \"\"\" Crunch image data, and return the processed data, or orignal data if operation failed. \"\"\"\n if (format is CoverImageFormat.PNG and not HAS_OPTIPNG or format is\n CoverImageFormat.JPEG and not HAS_JPEGOPTIM):\n return image_data\n with mkstemp_ctx.mkstemp(suffix='.%s' % format.name.lower()\n ) as tmp_out_filepath:\n if not silent:\n logging.getLogger('Cover').info('Crunching %s image...' %\n format.name.upper())\n with open(tmp_out_filepath, 'wb') as tmp_out_file:\n tmp_out_file.write(image_data)\n size_before = len(image_data)\n if format is CoverImageFormat.PNG:\n cmd = ['optipng', '-quiet', '-o1']\n elif format is CoverImageFormat.JPEG:\n cmd = ['jpegoptim', '-q', '--strip-all']\n cmd.append(tmp_out_filepath)\n p = await asyncio.create_subprocess_exec(*cmd, stdin=asyncio.\n subprocess.DEVNULL, stdout=asyncio.subprocess.DEVNULL, stderr=\n asyncio.subprocess.DEVNULL)\n await p.wait()\n if p.returncode != 0:\n if not silent:\n logging.getLogger('Cover').warning('Crunching image failed')\n return image_data\n with open(tmp_out_filepath, 'rb') as tmp_out_file:\n crunched_image_data = tmp_out_file.read()\n size_after = len(crunched_image_data)\n pct_saved = 100 * (size_before - size_after) / size_before\n if not silent:\n logging.getLogger('Cover').debug(\n 'Crunching image saved %.2f%% filesize' % pct_saved)\n return crunched_image_data\n",
"role": "user"
},
{
"content": "crunch",
"role": "assistant"
}
] |
def x(img_data):
format, width, height = None, None, None
img_stream = io.BytesIO(img_data)
try:
img = PIL.Image.open(img_stream)
except IOError:
format = imghdr.what(None, h=img_data)
format = SUPPORTED_IMG_FORMATS.get(format, None)
else:
format = img.format.lower()
format = SUPPORTED_IMG_FORMATS.get(format, None)
width, height = img.size
return format, width, height
| guessImageMetadataFromData | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(img_data):\n format, width, height = None, None, None\n img_stream = io.BytesIO(img_data)\n try:\n img = PIL.Image.open(img_stream)\n except IOError:\n format = imghdr.what(None, h=img_data)\n format = SUPPORTED_IMG_FORMATS.get(format, None)\n else:\n format = img.format.lower()\n format = SUPPORTED_IMG_FORMATS.get(format, None)\n width, height = img.size\n return format, width, height\n",
"role": "user"
},
{
"content": "guessImageMetadataFromData",
"role": "assistant"
}
] |
async def guessImageMetadataFromHttpData(response):
""" Identify an image format and size from the beginning of its HTTP data. """
metadata = None
img_data = bytearray()
while len(img_data) < CoverSourceResult.MAX_FILE_METADATA_PEEK_SIZE:
new_img_data = await response.content.read(__class__.
METADATA_PEEK_SIZE_INCREMENT)
if not new_img_data:
break
img_data.extend(new_img_data)
metadata = __class__.guessImageMetadataFromData(img_data)
if metadata is not None and all(metadata):
return metadata
return metadata
| guessImageMetadataFromHttpData | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def guessImageMetadataFromHttpData(response):\n \"\"\" Identify an image format and size from the beginning of its HTTP data. \"\"\"\n metadata = None\n img_data = bytearray()\n while len(img_data) < CoverSourceResult.MAX_FILE_METADATA_PEEK_SIZE:\n new_img_data = await response.content.read(__class__.\n METADATA_PEEK_SIZE_INCREMENT)\n if not new_img_data:\n break\n img_data.extend(new_img_data)\n metadata = __class__.guessImageMetadataFromData(img_data)\n if metadata is not None and all(metadata):\n return metadata\n return metadata\n",
"role": "user"
},
{
"content": "guessImageMetadataFromHttpData",
"role": "assistant"
}
] |
def x(response):
extensions = []
try:
content_type = response.headers['Content-Type']
except KeyError:
pass
else:
ext = mimetypes.guess_extension(content_type, strict=False)
if ext is not None:
extensions.append(ext)
urls = list(response.history) + [response.url]
for url in map(str, urls):
ext = os.path.splitext(urllib.parse.urlsplit(url).path)[-1]
if ext is not None and ext not in extensions:
extensions.append(ext)
for ext in extensions:
try:
return SUPPORTED_IMG_FORMATS[ext[1:]]
except KeyError:
pass
| guessImageFormatFromHttpResponse | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(response):\n extensions = []\n try:\n content_type = response.headers['Content-Type']\n except KeyError:\n pass\n else:\n ext = mimetypes.guess_extension(content_type, strict=False)\n if ext is not None:\n extensions.append(ext)\n urls = list(response.history) + [response.url]\n for url in map(str, urls):\n ext = os.path.splitext(urllib.parse.urlsplit(url).path)[-1]\n if ext is not None and ext not in extensions:\n extensions.append(ext)\n for ext in extensions:\n try:\n return SUPPORTED_IMG_FORMATS[ext[1:]]\n except KeyError:\n pass\n",
"role": "user"
},
{
"content": "guessImageFormatFromHttpResponse",
"role": "assistant"
}
] |
async def preProcessForComparison(results, target_size, size_tolerance_prct):
""" Process results to prepare them for future comparison and sorting. """
reference = None
for result in results:
if result.source_quality is CoverSourceQuality.REFERENCE:
if reference is None or CoverSourceResult.compare(result,
reference, target_size=target_size, size_tolerance_prct=
size_tolerance_prct) > 0:
reference = result
results = list(itertools.filterfalse(operator.attrgetter(
'is_only_reference'), results))
no_dup_results = []
for result in results:
is_dup = False
for result_comp in results:
if (result_comp is not result and result_comp.urls == result.
urls and __class__.compare(result, result_comp, target_size
=target_size, size_tolerance_prct=size_tolerance_prct) < 0):
is_dup = True
break
if not is_dup:
no_dup_results.append(result)
dup_count = len(results) - len(no_dup_results)
if dup_count > 0:
logging.getLogger('Cover').info('Removed %u duplicate results' %
dup_count)
results = no_dup_results
if reference is not None:
logging.getLogger('Cover').info('Reference is: %s' % reference)
reference.is_similar_to_reference = True
futures = []
for result in results:
coroutine = result.updateSignature()
future = asyncio.ensure_future(coroutine)
futures.append(future)
if reference.is_only_reference:
assert reference not in results
coroutine = reference.updateSignature()
future = asyncio.ensure_future(coroutine)
futures.append(future)
if futures:
await asyncio.wait(futures)
for future in futures:
future.result()
for result in results:
if (result is not reference and result.thumbnail_sig is not
None and reference.thumbnail_sig is not None):
result.is_similar_to_reference = __class__.areImageSigsSimilar(
result.thumbnail_sig, reference.thumbnail_sig)
if result.is_similar_to_reference:
logging.getLogger('Cover').debug(
'%s is similar to reference' % result)
else:
logging.getLogger('Cover').debug(
'%s is NOT similar to reference' % result)
else:
logging.getLogger('Cover').warning('No reference result found')
return results
| preProcessForComparison | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def preProcessForComparison(results, target_size, size_tolerance_prct):\n \"\"\" Process results to prepare them for future comparison and sorting. \"\"\"\n reference = None\n for result in results:\n if result.source_quality is CoverSourceQuality.REFERENCE:\n if reference is None or CoverSourceResult.compare(result,\n reference, target_size=target_size, size_tolerance_prct=\n size_tolerance_prct) > 0:\n reference = result\n results = list(itertools.filterfalse(operator.attrgetter(\n 'is_only_reference'), results))\n no_dup_results = []\n for result in results:\n is_dup = False\n for result_comp in results:\n if (result_comp is not result and result_comp.urls == result.\n urls and __class__.compare(result, result_comp, target_size\n =target_size, size_tolerance_prct=size_tolerance_prct) < 0):\n is_dup = True\n break\n if not is_dup:\n no_dup_results.append(result)\n dup_count = len(results) - len(no_dup_results)\n if dup_count > 0:\n logging.getLogger('Cover').info('Removed %u duplicate results' %\n dup_count)\n results = no_dup_results\n if reference is not None:\n logging.getLogger('Cover').info('Reference is: %s' % reference)\n reference.is_similar_to_reference = True\n futures = []\n for result in results:\n coroutine = result.updateSignature()\n future = asyncio.ensure_future(coroutine)\n futures.append(future)\n if reference.is_only_reference:\n assert reference not in results\n coroutine = reference.updateSignature()\n future = asyncio.ensure_future(coroutine)\n futures.append(future)\n if futures:\n await asyncio.wait(futures)\n for future in futures:\n future.result()\n for result in results:\n if (result is not reference and result.thumbnail_sig is not\n None and reference.thumbnail_sig is not None):\n result.is_similar_to_reference = __class__.areImageSigsSimilar(\n result.thumbnail_sig, reference.thumbnail_sig)\n if result.is_similar_to_reference:\n logging.getLogger('Cover').debug(\n '%s is similar to reference' % result)\n else:\n logging.getLogger('Cover').debug(\n '%s is NOT similar to reference' % result)\n else:\n logging.getLogger('Cover').warning('No reference result found')\n return results\n",
"role": "user"
},
{
"content": "preProcessForComparison",
"role": "assistant"
}
] |
def x(image_data):
parser = PIL.ImageFile.Parser()
parser.feed(image_data)
img = parser.close()
target_size = __class__.IMG_SIG_SIZE, __class__.IMG_SIG_SIZE
img.thumbnail(target_size, PIL.Image.BICUBIC)
if img.size != target_size:
logging.getLogger('Cover').debug(
'Non square thumbnail after resize to %ux%u, unable to compute signature'
% target_size)
return None
img = img.convert(mode='RGB')
pixels = img.getdata()
pixel_count = target_size[0] * target_size[1]
color_count = 3
r = bitarray.bitarray(pixel_count * color_count)
r.setall(False)
for ic in range(color_count):
mean = sum(p[ic] for p in pixels) // pixel_count
for ip, p in enumerate(pixels):
if p[ic] > mean:
r[pixel_count * ic + ip] = True
return r
| computeImgSignature | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(image_data):\n parser = PIL.ImageFile.Parser()\n parser.feed(image_data)\n img = parser.close()\n target_size = __class__.IMG_SIG_SIZE, __class__.IMG_SIG_SIZE\n img.thumbnail(target_size, PIL.Image.BICUBIC)\n if img.size != target_size:\n logging.getLogger('Cover').debug(\n 'Non square thumbnail after resize to %ux%u, unable to compute signature'\n % target_size)\n return None\n img = img.convert(mode='RGB')\n pixels = img.getdata()\n pixel_count = target_size[0] * target_size[1]\n color_count = 3\n r = bitarray.bitarray(pixel_count * color_count)\n r.setall(False)\n for ic in range(color_count):\n mean = sum(p[ic] for p in pixels) // pixel_count\n for ip, p in enumerate(pixels):\n if p[ic] > mean:\n r[pixel_count * ic + ip] = True\n return r\n",
"role": "user"
},
{
"content": "computeImgSignature",
"role": "assistant"
}
] |
def x(lib_dir, cover_filename, *, ignore_existing=False):
work = {}
stats = collections.OrderedDict((k, 0) for k in ('files', 'albums',
'missing covers', 'errors'))
with tqdm.tqdm(desc='Analyzing library', unit='dir', postfix=stats
) as progress, tqdm_logging.redirect_logging(progress):
for rootpath, rel_dirpaths, rel_filepaths in os.walk(lib_dir):
metadata = analyze_dir(stats, rootpath, rel_filepaths,
cover_filename, ignore_existing=ignore_existing)
progress.set_postfix(stats, refresh=False)
progress.update(1)
if all(metadata[:-1]):
work[rootpath] = metadata[:-1]
return work
| analyze_lib | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(lib_dir, cover_filename, *, ignore_existing=False):\n work = {}\n stats = collections.OrderedDict((k, 0) for k in ('files', 'albums',\n 'missing covers', 'errors'))\n with tqdm.tqdm(desc='Analyzing library', unit='dir', postfix=stats\n ) as progress, tqdm_logging.redirect_logging(progress):\n for rootpath, rel_dirpaths, rel_filepaths in os.walk(lib_dir):\n metadata = analyze_dir(stats, rootpath, rel_filepaths,\n cover_filename, ignore_existing=ignore_existing)\n progress.set_postfix(stats, refresh=False)\n progress.update(1)\n if all(metadata[:-1]):\n work[rootpath] = metadata[:-1]\n return work\n",
"role": "user"
},
{
"content": "analyze_lib",
"role": "assistant"
}
] |
def x(audio_filepaths):
artist, album, has_embedded_album_art = None, None, None
for audio_filepath in audio_filepaths:
try:
mf = mutagen.File(audio_filepath)
except Exception:
continue
if mf is None:
continue
for key in ('albumartist', 'artist', 'TPE1', 'TPE2', 'aART', '©ART'):
try:
val = mf.get(key, None)
except ValueError:
val = None
if val is not None:
artist = val[-1]
break
for key in ('_album', 'album', 'TALB', '©alb'):
try:
val = mf.get(key, None)
except ValueError:
val = None
if val is not None:
album = val[-1]
break
if artist and album:
if isinstance(mf, mutagen.ogg.OggFileType):
has_embedded_album_art = 'metadata_block_picture' in mf
elif isinstance(mf, mutagen.mp3.MP3):
has_embedded_album_art = any(map(operator.methodcaller(
'startswith', 'APIC:'), mf.keys()))
elif isinstance(mf, mutagen.mp4.MP4):
has_embedded_album_art = 'covr' in mf
break
return artist, album, has_embedded_album_art
| get_metadata | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(audio_filepaths):\n artist, album, has_embedded_album_art = None, None, None\n for audio_filepath in audio_filepaths:\n try:\n mf = mutagen.File(audio_filepath)\n except Exception:\n continue\n if mf is None:\n continue\n for key in ('albumartist', 'artist', 'TPE1', 'TPE2', 'aART', '©ART'):\n try:\n val = mf.get(key, None)\n except ValueError:\n val = None\n if val is not None:\n artist = val[-1]\n break\n for key in ('_album', 'album', 'TALB', '©alb'):\n try:\n val = mf.get(key, None)\n except ValueError:\n val = None\n if val is not None:\n album = val[-1]\n break\n if artist and album:\n if isinstance(mf, mutagen.ogg.OggFileType):\n has_embedded_album_art = 'metadata_block_picture' in mf\n elif isinstance(mf, mutagen.mp3.MP3):\n has_embedded_album_art = any(map(operator.methodcaller(\n 'startswith', 'APIC:'), mf.keys()))\n elif isinstance(mf, mutagen.mp4.MP4):\n has_embedded_album_art = 'covr' in mf\n break\n return artist, album, has_embedded_album_art\n",
"role": "user"
},
{
"content": "get_metadata",
"role": "assistant"
}
] |
def x(stats, parent_dir, rel_filepaths, cover_filename, *, ignore_existing=
False):
no_metadata = None, None, None
metadata = no_metadata
audio_filepaths = []
for rel_filepath in rel_filepaths:
stats['files'] += 1
try:
ext = os.path.splitext(rel_filepath)[1][1:].lower()
except IndexError:
continue
if ext in AUDIO_EXTENSIONS:
audio_filepaths.append(os.path.join(parent_dir, rel_filepath))
if audio_filepaths:
stats['albums'] += 1
if cover_filename != EMBEDDED_ALBUM_ART_SYMBOL:
missing = not os.path.isfile(os.path.join(parent_dir,
cover_filename)) or ignore_existing
if missing:
metadata = get_metadata(audio_filepaths)
else:
metadata = get_metadata(audio_filepaths)
missing = not metadata[2] or ignore_existing
if missing:
stats['missing covers'] += 1
if not all(metadata[:-1]):
stats['errors'] += 1
logging.getLogger('sacad_r').error(
"Unable to read metadata for album directory '%s'" %
parent_dir)
else:
metadata = no_metadata
return metadata
| analyze_dir | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(stats, parent_dir, rel_filepaths, cover_filename, *, ignore_existing=\n False):\n no_metadata = None, None, None\n metadata = no_metadata\n audio_filepaths = []\n for rel_filepath in rel_filepaths:\n stats['files'] += 1\n try:\n ext = os.path.splitext(rel_filepath)[1][1:].lower()\n except IndexError:\n continue\n if ext in AUDIO_EXTENSIONS:\n audio_filepaths.append(os.path.join(parent_dir, rel_filepath))\n if audio_filepaths:\n stats['albums'] += 1\n if cover_filename != EMBEDDED_ALBUM_ART_SYMBOL:\n missing = not os.path.isfile(os.path.join(parent_dir,\n cover_filename)) or ignore_existing\n if missing:\n metadata = get_metadata(audio_filepaths)\n else:\n metadata = get_metadata(audio_filepaths)\n missing = not metadata[2] or ignore_existing\n if missing:\n stats['missing covers'] += 1\n if not all(metadata[:-1]):\n stats['errors'] += 1\n logging.getLogger('sacad_r').error(\n \"Unable to read metadata for album directory '%s'\" %\n parent_dir)\n else:\n metadata = no_metadata\n return metadata\n",
"role": "user"
},
{
"content": "analyze_dir",
"role": "assistant"
}
] |
def x(cover_filepath, path):
with open(cover_filepath, 'rb') as f:
cover_data = f.read()
for filename in os.listdir(path):
try:
ext = os.path.splitext(filename)[1][1:].lower()
except IndexError:
continue
if ext in AUDIO_EXTENSIONS:
filepath = os.path.join(path, filename)
mf = mutagen.File(filepath)
if isinstance(mf.tags, mutagen._vorbis.VComment) or isinstance(mf,
mutagen.ogg.OggFileType):
picture = mutagen.flac.Picture()
picture.data = cover_data
picture.type = mutagen.id3.PictureType.COVER_FRONT
picture.mime = 'image/jpeg'
encoded_data = base64.b64encode(picture.write())
mf['metadata_block_picture'] = encoded_data.decode('ascii')
elif isinstance(mf.tags, mutagen.id3.ID3) or isinstance(mf,
mutagen.id3.ID3FileType):
mf.tags.add(mutagen.id3.APIC(mime='image/jpeg', type=
mutagen.id3.PictureType.COVER_FRONT, data=cover_data))
elif isinstance(mf.tags, mutagen.mp4.MP4Tags) or isinstance(mf,
mutagen.mp4.MP4):
mf['covr'] = [mutagen.mp4.MP4Cover(cover_data, imageformat=
mutagen.mp4.AtomDataType.JPEG)]
mf.save()
| embed_album_art | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cover_filepath, path):\n with open(cover_filepath, 'rb') as f:\n cover_data = f.read()\n for filename in os.listdir(path):\n try:\n ext = os.path.splitext(filename)[1][1:].lower()\n except IndexError:\n continue\n if ext in AUDIO_EXTENSIONS:\n filepath = os.path.join(path, filename)\n mf = mutagen.File(filepath)\n if isinstance(mf.tags, mutagen._vorbis.VComment) or isinstance(mf,\n mutagen.ogg.OggFileType):\n picture = mutagen.flac.Picture()\n picture.data = cover_data\n picture.type = mutagen.id3.PictureType.COVER_FRONT\n picture.mime = 'image/jpeg'\n encoded_data = base64.b64encode(picture.write())\n mf['metadata_block_picture'] = encoded_data.decode('ascii')\n elif isinstance(mf.tags, mutagen.id3.ID3) or isinstance(mf,\n mutagen.id3.ID3FileType):\n mf.tags.add(mutagen.id3.APIC(mime='image/jpeg', type=\n mutagen.id3.PictureType.COVER_FRONT, data=cover_data))\n elif isinstance(mf.tags, mutagen.mp4.MP4Tags) or isinstance(mf,\n mutagen.mp4.MP4):\n mf['covr'] = [mutagen.mp4.MP4Cover(cover_data, imageformat=\n mutagen.mp4.AtomDataType.JPEG)]\n mf.save()\n",
"role": "user"
},
{
"content": "embed_album_art",
"role": "assistant"
}
] |
def x(iterable, n):
it = iter(iterable)
while True:
chunk = tuple(itertools.islice(it, n))
if not chunk:
return
yield chunk
| ichunk | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(iterable, n):\n it = iter(iterable)\n while True:\n chunk = tuple(itertools.islice(it, n))\n if not chunk:\n return\n yield chunk\n",
"role": "user"
},
{
"content": "ichunk",
"role": "assistant"
}
] |
def x(work, args):
with contextlib.ExitStack() as cm:
if args.filename == EMBEDDED_ALBUM_ART_SYMBOL:
tmp_prefix = '%s_' % os.path.splitext(os.path.basename(inspect.
getfile(inspect.currentframe())))[0]
tmp_dir = cm.enter_context(tempfile.TemporaryDirectory(prefix=
tmp_prefix))
stats = collections.OrderedDict((k, 0) for k in ('ok', 'errors',
'no result found'))
progress = cm.enter_context(tqdm.tqdm(total=len(work), miniters=1,
desc='Searching covers', unit='cover', postfix=stats))
cm.enter_context(tqdm_logging.redirect_logging(progress))
def x(future):
path, cover_filepath, artist, album = futures[future]
try:
status = future.result()
except Exception as exception:
stats['errors'] += 1
logging.getLogger('sacad_r').error(
"Error occured while searching cover for '%s' by '%s' from '%s': %s %s"
% (album, artist, path, exception.__class__.
__qualname__, exception))
else:
if status:
if args.filename == EMBEDDED_ALBUM_ART_SYMBOL:
try:
embed_album_art(cover_filepath, path)
except Exception as exception:
stats['errors'] += 1
logging.getLogger('sacad_r').error(
"Error occured while embedding cover for '%s' by '%s' from '%s': %s %s"
% (album, artist, path, exception.
__class__.__qualname__, exception))
else:
stats['ok'] += 1
finally:
os.remove(cover_filepath)
else:
stats['ok'] += 1
else:
stats['no result found'] += 1
logging.getLogger('sacad_r').warning(
"Unable to find cover for '%s' by '%s' from '%s'" %
(album, artist, path))
progress.set_postfix(stats, refresh=False)
progress.update(1)
async_loop = asyncio.get_event_loop()
i = 0
work_chunk_length = 16
for work_chunk in ichunk(work.items(), work_chunk_length):
futures = {}
for i, (path, (artist, album)) in enumerate(work_chunk, i):
if args.filename == EMBEDDED_ALBUM_ART_SYMBOL:
cover_filepath = os.path.join(tmp_dir, '%00u.%s' % (i,
args.format.name.lower()))
else:
cover_filepath = os.path.join(path, args.filename)
coroutine = sacad.search_and_download(album, artist, args.
format, args.size, cover_filepath, size_tolerance_prct=
args.size_tolerance_prct, amazon_tlds=args.amazon_tlds,
no_lq_sources=args.no_lq_sources, async_loop=async_loop)
future = asyncio.ensure_future(coroutine, loop=async_loop)
futures[future] = path, cover_filepath, artist, album
for future in futures:
future.add_done_callback(update_progress)
root_future = asyncio.gather(*futures.keys(), loop=async_loop)
async_loop.run_until_complete(root_future)
| get_covers | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(work, args):\n with contextlib.ExitStack() as cm:\n if args.filename == EMBEDDED_ALBUM_ART_SYMBOL:\n tmp_prefix = '%s_' % os.path.splitext(os.path.basename(inspect.\n getfile(inspect.currentframe())))[0]\n tmp_dir = cm.enter_context(tempfile.TemporaryDirectory(prefix=\n tmp_prefix))\n stats = collections.OrderedDict((k, 0) for k in ('ok', 'errors',\n 'no result found'))\n progress = cm.enter_context(tqdm.tqdm(total=len(work), miniters=1,\n desc='Searching covers', unit='cover', postfix=stats))\n cm.enter_context(tqdm_logging.redirect_logging(progress))\n\n def x(future):\n path, cover_filepath, artist, album = futures[future]\n try:\n status = future.result()\n except Exception as exception:\n stats['errors'] += 1\n logging.getLogger('sacad_r').error(\n \"Error occured while searching cover for '%s' by '%s' from '%s': %s %s\"\n % (album, artist, path, exception.__class__.\n __qualname__, exception))\n else:\n if status:\n if args.filename == EMBEDDED_ALBUM_ART_SYMBOL:\n try:\n embed_album_art(cover_filepath, path)\n except Exception as exception:\n stats['errors'] += 1\n logging.getLogger('sacad_r').error(\n \"Error occured while embedding cover for '%s' by '%s' from '%s': %s %s\"\n % (album, artist, path, exception.\n __class__.__qualname__, exception))\n else:\n stats['ok'] += 1\n finally:\n os.remove(cover_filepath)\n else:\n stats['ok'] += 1\n else:\n stats['no result found'] += 1\n logging.getLogger('sacad_r').warning(\n \"Unable to find cover for '%s' by '%s' from '%s'\" %\n (album, artist, path))\n progress.set_postfix(stats, refresh=False)\n progress.update(1)\n async_loop = asyncio.get_event_loop()\n i = 0\n work_chunk_length = 16\n for work_chunk in ichunk(work.items(), work_chunk_length):\n futures = {}\n for i, (path, (artist, album)) in enumerate(work_chunk, i):\n if args.filename == EMBEDDED_ALBUM_ART_SYMBOL:\n cover_filepath = os.path.join(tmp_dir, '%00u.%s' % (i,\n args.format.name.lower()))\n else:\n cover_filepath = os.path.join(path, args.filename)\n coroutine = sacad.search_and_download(album, artist, args.\n format, args.size, cover_filepath, size_tolerance_prct=\n args.size_tolerance_prct, amazon_tlds=args.amazon_tlds,\n no_lq_sources=args.no_lq_sources, async_loop=async_loop)\n future = asyncio.ensure_future(coroutine, loop=async_loop)\n futures[future] = path, cover_filepath, artist, album\n for future in futures:\n future.add_done_callback(update_progress)\n root_future = asyncio.gather(*futures.keys(), loop=async_loop)\n async_loop.run_until_complete(root_future)\n",
"role": "user"
},
{
"content": "get_covers",
"role": "assistant"
}
] |
def x(*args, **kwargs):
fd, filename = tempfile.mkstemp(*args, **kwargs)
os.close(fd)
try:
yield filename
finally:
os.remove(filename)
| mkstemp | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(*args, **kwargs):\n fd, filename = tempfile.mkstemp(*args, **kwargs)\n os.close(fd)\n try:\n yield filename\n finally:\n os.remove(filename)\n",
"role": "user"
},
{
"content": "mkstemp",
"role": "assistant"
}
] |
def x(tqdm_obj, logger=logging.getLogger()):
assert len(logger.handlers) == 1
prev_handler = logger.handlers[0]
logger.removeHandler(prev_handler)
tqdm_handler = TqdmLoggingHandler(tqdm_obj)
if prev_handler.formatter is not None:
tqdm_handler.setFormatter(prev_handler.formatter)
logger.addHandler(tqdm_handler)
try:
yield
finally:
logger.removeHandler(tqdm_handler)
logger.addHandler(prev_handler)
| redirect_logging | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tqdm_obj, logger=logging.getLogger()):\n assert len(logger.handlers) == 1\n prev_handler = logger.handlers[0]\n logger.removeHandler(prev_handler)\n tqdm_handler = TqdmLoggingHandler(tqdm_obj)\n if prev_handler.formatter is not None:\n tqdm_handler.setFormatter(prev_handler.formatter)\n logger.addHandler(tqdm_handler)\n try:\n yield\n finally:\n logger.removeHandler(tqdm_handler)\n logger.addHandler(prev_handler)\n",
"role": "user"
},
{
"content": "redirect_logging",
"role": "assistant"
}
] |
async def search(self, album, artist):
""" Search for a given album/artist and return an iterable of CoverSourceResult. """
self.logger.debug("Searching with source '%s'..." % self.__class__.__name__
)
album = self.processAlbumString(album)
artist = self.processArtistString(artist)
url_data = self.getSearchUrl(album, artist)
if isinstance(url_data, tuple):
url, post_data = url_data
else:
url = url_data
post_data = None
try:
store_in_cache_callback, api_data = await self.fetchResults(url,
post_data)
results = await self.parseResults(api_data)
except Exception as e:
self.logger.warning("Search with source '%s' failed: %s %s" % (self
.__class__.__name__, e.__class__.__qualname__, e))
return ()
else:
if results:
await store_in_cache_callback()
futures = []
for result in filter(operator.methodcaller('needMetadataUpdate'), results):
coroutine = result.updateImageMetadata()
future = asyncio.ensure_future(coroutine)
futures.append(future)
if futures:
await asyncio.wait(futures)
for future in futures:
future.result()
results_excluded_count = 0
reference_only_count = 0
results_kept = []
for result in results:
if (result.size[0] + self.size_tolerance_prct * self.target_size /
100 < self.target_size or result.size[1] + self.
size_tolerance_prct * self.target_size / 100 < self.target_size or
result.format is None or result.needMetadataUpdate()):
if result.source_quality is CoverSourceQuality.REFERENCE:
result.is_only_reference = True
results_kept.append(result)
reference_only_count += 1
else:
results_excluded_count += 1
else:
results_kept.append(result)
result_kept_count = len(results_kept) - reference_only_count
self.logger.info(
"Got %u relevant (%u excluded) results from source '%s'" % (
result_kept_count, results_excluded_count + reference_only_count,
self.__class__.__name__))
for result in itertools.filterfalse(operator.attrgetter(
'is_only_reference'), results_kept):
self.logger.debug('%s %s%s %4dx%4d %s%s' % (result.__class__.
__name__, '(%02d) ' % result.rank if result.rank is not None else
'', result.format.name, result.size[0], result.size[1], result.
urls[0], ' [x%u]' % len(result.urls) if len(result.urls) > 1 else
''))
return results_kept
| search | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def search(self, album, artist):\n \"\"\" Search for a given album/artist and return an iterable of CoverSourceResult. \"\"\"\n self.logger.debug(\"Searching with source '%s'...\" % self.__class__.__name__\n )\n album = self.processAlbumString(album)\n artist = self.processArtistString(artist)\n url_data = self.getSearchUrl(album, artist)\n if isinstance(url_data, tuple):\n url, post_data = url_data\n else:\n url = url_data\n post_data = None\n try:\n store_in_cache_callback, api_data = await self.fetchResults(url,\n post_data)\n results = await self.parseResults(api_data)\n except Exception as e:\n self.logger.warning(\"Search with source '%s' failed: %s %s\" % (self\n .__class__.__name__, e.__class__.__qualname__, e))\n return ()\n else:\n if results:\n await store_in_cache_callback()\n futures = []\n for result in filter(operator.methodcaller('needMetadataUpdate'), results):\n coroutine = result.updateImageMetadata()\n future = asyncio.ensure_future(coroutine)\n futures.append(future)\n if futures:\n await asyncio.wait(futures)\n for future in futures:\n future.result()\n results_excluded_count = 0\n reference_only_count = 0\n results_kept = []\n for result in results:\n if (result.size[0] + self.size_tolerance_prct * self.target_size / \n 100 < self.target_size or result.size[1] + self.\n size_tolerance_prct * self.target_size / 100 < self.target_size or\n result.format is None or result.needMetadataUpdate()):\n if result.source_quality is CoverSourceQuality.REFERENCE:\n result.is_only_reference = True\n results_kept.append(result)\n reference_only_count += 1\n else:\n results_excluded_count += 1\n else:\n results_kept.append(result)\n result_kept_count = len(results_kept) - reference_only_count\n self.logger.info(\n \"Got %u relevant (%u excluded) results from source '%s'\" % (\n result_kept_count, results_excluded_count + reference_only_count,\n self.__class__.__name__))\n for result in itertools.filterfalse(operator.attrgetter(\n 'is_only_reference'), results_kept):\n self.logger.debug('%s %s%s %4dx%4d %s%s' % (result.__class__.\n __name__, '(%02d) ' % result.rank if result.rank is not None else\n '', result.format.name, result.size[0], result.size[1], result.\n urls[0], ' [x%u]' % len(result.urls) if len(result.urls) > 1 else\n ''))\n return results_kept\n",
"role": "user"
},
{
"content": "search",
"role": "assistant"
}
] |
async def fetchResults(self, url, post_data=None):
""" Get a (store in cache callback, search results) tuple from an URL. """
if post_data is not None:
self.logger.debug("Querying URL '%s' %s..." % (url, dict(post_data)))
else:
self.logger.debug("Querying URL '%s'..." % url)
headers = {}
self.updateHttpHeaders(headers)
return await self.http.query(url, post_data=post_data, headers=headers,
cache=__class__.api_cache)
| fetchResults | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def fetchResults(self, url, post_data=None):\n \"\"\" Get a (store in cache callback, search results) tuple from an URL. \"\"\"\n if post_data is not None:\n self.logger.debug(\"Querying URL '%s' %s...\" % (url, dict(post_data)))\n else:\n self.logger.debug(\"Querying URL '%s'...\" % url)\n headers = {}\n self.updateHttpHeaders(headers)\n return await self.http.query(url, post_data=post_data, headers=headers,\n cache=__class__.api_cache)\n",
"role": "user"
},
{
"content": "fetchResults",
"role": "assistant"
}
] |
async def probeUrl(self, url, response_headers=None):
""" Probe URL reachability from cache or HEAD request. """
self.logger.debug("Probing URL '%s'..." % url)
headers = {}
self.updateHttpHeaders(headers)
resp_headers = {}
resp_ok = await self.http.isReachable(url, headers=headers,
response_headers=resp_headers, cache=__class__.probe_cache)
if response_headers is not None:
response_headers.update(resp_headers)
return resp_ok
| probeUrl | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def probeUrl(self, url, response_headers=None):\n \"\"\" Probe URL reachability from cache or HEAD request. \"\"\"\n self.logger.debug(\"Probing URL '%s'...\" % url)\n headers = {}\n self.updateHttpHeaders(headers)\n resp_headers = {}\n resp_ok = await self.http.isReachable(url, headers=headers,\n response_headers=resp_headers, cache=__class__.probe_cache)\n if response_headers is not None:\n response_headers.update(resp_headers)\n return resp_ok\n",
"role": "user"
},
{
"content": "probeUrl",
"role": "assistant"
}
] |
def x(s):
return ''.join(c for c in unicodedata.normalize('NFKD', s) if not
unicodedata.combining(c))
| unaccentuate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(s):\n return ''.join(c for c in unicodedata.normalize('NFKD', s) if not\n unicodedata.combining(c))\n",
"role": "user"
},
{
"content": "unaccentuate",
"role": "assistant"
}
] |
def x(s, *, char_blacklist=string.punctuation):
s = ''.join(c for c in s if c not in char_blacklist)
return ' '.join(filter(None, s.split(' ')))
| unpunctuate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(s, *, char_blacklist=string.punctuation):\n s = ''.join(c for c in s if c not in char_blacklist)\n return ' '.join(filter(None, s.split(' ')))\n",
"role": "user"
},
{
"content": "unpunctuate",
"role": "assistant"
}
] |
def x(self, album, artist):
params = collections.OrderedDict()
params['search-alias'] = 'popular'
params['field-artist'] = artist
params['field-title'] = album
params['sort'] = 'relevancerank'
return __class__.assembleUrl(self.base_url, params)
| getSearchUrl | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, album, artist):\n params = collections.OrderedDict()\n params['search-alias'] = 'popular'\n params['field-artist'] = artist\n params['field-title'] = album\n params['sort'] = 'relevancerank'\n return __class__.assembleUrl(self.base_url, params)\n",
"role": "user"
},
{
"content": "getSearchUrl",
"role": "assistant"
}
] |
async def parseResults(self, api_data):
""" See CoverSource.parseResults. """
results = []
parser = lxml.etree.HTMLParser()
html = lxml.etree.XML(api_data.decode('utf-8', 'ignore'), parser)
for page_struct_version, result_selector in enumerate(__class__.
RESULTS_SELECTORS):
result_nodes = result_selector(html)
if result_nodes:
break
for rank, result_node in enumerate(result_nodes, 1):
try:
img_node = __class__.IMG_SELECTORS[page_struct_version](result_node
)[0]
except IndexError:
continue
thumbnail_url = img_node.get('src')
url_parts = thumbnail_url.rsplit('.', 2)
img_url = '.'.join((url_parts[0], url_parts[2]))
size = 500, 500
check_metadata = CoverImageMetadata.SIZE
if self.target_size > size[0] and rank <= 3:
product_url = __class__.PRODUCT_LINK_SELECTORS[page_struct_version
](result_node)[0].get('href')
product_url_split = urllib.parse.urlsplit(product_url)
if not product_url_split.scheme:
product_url_query = urllib.parse.parse_qsl(product_url_split
.query)
product_url_query = collections.OrderedDict(product_url_query)
try:
product_url = product_url_query['url']
except KeyError:
product_url = urllib.parse.urljoin(self.base_url,
product_url)
product_url_split = urllib.parse.urlsplit(product_url)
product_url_query = urllib.parse.parse_qsl(product_url_split.query)
product_url_query = collections.OrderedDict(product_url_query)
try:
del product_url_query['qid']
except KeyError:
pass
product_url_query = urllib.parse.urlencode(product_url_query)
product_url_no_ts = urllib.parse.urlunsplit(product_url_split[:
3] + (product_url_query,) + product_url_split[4:])
store_in_cache_callback, product_page_data = (await self.
fetchResults(product_url_no_ts))
product_page_html = lxml.etree.XML(product_page_data.decode(
'latin-1'), parser)
try:
img_node = __class__.PRODUCT_PAGE_IMG_SELECTOR(
product_page_html)[0]
except IndexError:
pass
else:
better_img_url = img_node.get('data-old-hires')
if better_img_url:
img_url = better_img_url
size_url_hint = img_url.rsplit('.', 2)[1].strip('_')
assert size_url_hint.startswith('SL')
size_url_hint = int(size_url_hint[2:])
size = size_url_hint, size_url_hint
check_metadata = CoverImageMetadata.NONE
await store_in_cache_callback()
format = CoverImageFormat.JPEG
results.append(AmazonCdCoverSourceResult(img_url, size, format,
thumbnail_url=thumbnail_url, source=self, rank=rank,
check_metadata=check_metadata))
return results
| parseResults | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def parseResults(self, api_data):\n \"\"\" See CoverSource.parseResults. \"\"\"\n results = []\n parser = lxml.etree.HTMLParser()\n html = lxml.etree.XML(api_data.decode('utf-8', 'ignore'), parser)\n for page_struct_version, result_selector in enumerate(__class__.\n RESULTS_SELECTORS):\n result_nodes = result_selector(html)\n if result_nodes:\n break\n for rank, result_node in enumerate(result_nodes, 1):\n try:\n img_node = __class__.IMG_SELECTORS[page_struct_version](result_node\n )[0]\n except IndexError:\n continue\n thumbnail_url = img_node.get('src')\n url_parts = thumbnail_url.rsplit('.', 2)\n img_url = '.'.join((url_parts[0], url_parts[2]))\n size = 500, 500\n check_metadata = CoverImageMetadata.SIZE\n if self.target_size > size[0] and rank <= 3:\n product_url = __class__.PRODUCT_LINK_SELECTORS[page_struct_version\n ](result_node)[0].get('href')\n product_url_split = urllib.parse.urlsplit(product_url)\n if not product_url_split.scheme:\n product_url_query = urllib.parse.parse_qsl(product_url_split\n .query)\n product_url_query = collections.OrderedDict(product_url_query)\n try:\n product_url = product_url_query['url']\n except KeyError:\n product_url = urllib.parse.urljoin(self.base_url,\n product_url)\n product_url_split = urllib.parse.urlsplit(product_url)\n product_url_query = urllib.parse.parse_qsl(product_url_split.query)\n product_url_query = collections.OrderedDict(product_url_query)\n try:\n del product_url_query['qid']\n except KeyError:\n pass\n product_url_query = urllib.parse.urlencode(product_url_query)\n product_url_no_ts = urllib.parse.urlunsplit(product_url_split[:\n 3] + (product_url_query,) + product_url_split[4:])\n store_in_cache_callback, product_page_data = (await self.\n fetchResults(product_url_no_ts))\n product_page_html = lxml.etree.XML(product_page_data.decode(\n 'latin-1'), parser)\n try:\n img_node = __class__.PRODUCT_PAGE_IMG_SELECTOR(\n product_page_html)[0]\n except IndexError:\n pass\n else:\n better_img_url = img_node.get('data-old-hires')\n if better_img_url:\n img_url = better_img_url\n size_url_hint = img_url.rsplit('.', 2)[1].strip('_')\n assert size_url_hint.startswith('SL')\n size_url_hint = int(size_url_hint[2:])\n size = size_url_hint, size_url_hint\n check_metadata = CoverImageMetadata.NONE\n await store_in_cache_callback()\n format = CoverImageFormat.JPEG\n results.append(AmazonCdCoverSourceResult(img_url, size, format,\n thumbnail_url=thumbnail_url, source=self, rank=rank,\n check_metadata=check_metadata))\n return results\n",
"role": "user"
},
{
"content": "parseResults",
"role": "assistant"
}
] |
def x(obj):
types = tuple([Shader, Mesh, VertexBuffer, IndexBuffer, Texture,
Program, Context])
for name in dir(obj):
child = getattr(obj, name)
if isinstance(child, types):
child.delete()
| delete_all | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(obj):\n types = tuple([Shader, Mesh, VertexBuffer, IndexBuffer, Texture,\n Program, Context])\n for name in dir(obj):\n child = getattr(obj, name)\n if isinstance(child, types):\n child.delete()\n",
"role": "user"
},
{
"content": "delete_all",
"role": "assistant"
}
] |
def x(filename):
version_checker_source = """
import sys
import ctypes
def get_version(library_handle):
'''
Queries and returns the library version tuple or None.
'''
major_value = ctypes.c_int(0)
major = ctypes.pointer(major_value)
minor_value = ctypes.c_int(0)
minor = ctypes.pointer(minor_value)
rev_value = ctypes.c_int(0)
rev = ctypes.pointer(rev_value)
if hasattr(library_handle, 'glfwGetVersion'):
library_handle.glfwGetVersion(major, minor, rev)
version = (major_value.value,
minor_value.value,
rev_value.value)
return version
else:
return None
try:
input_func = raw_input
except NameError:
input_func = input
filename = input_func().strip()
try:
library_handle = ctypes.CDLL(filename)
except OSError:
pass
else:
version = get_version(library_handle)
print(version)
"""
args = [sys.executable, '-c', textwrap.dedent(version_checker_source)]
process = subprocess.Popen(args, universal_newlines=True, stdin=
subprocess.PIPE, stdout=subprocess.PIPE)
out = process.communicate(_to_char_p(filename))[0]
out = out.strip()
if out:
return eval(out)
else:
return None
| glfw_get_version | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename):\n version_checker_source = \"\"\"\n import sys\n import ctypes\n\n def get_version(library_handle):\n '''\n Queries and returns the library version tuple or None.\n '''\n major_value = ctypes.c_int(0)\n major = ctypes.pointer(major_value)\n minor_value = ctypes.c_int(0)\n minor = ctypes.pointer(minor_value)\n rev_value = ctypes.c_int(0)\n rev = ctypes.pointer(rev_value)\n if hasattr(library_handle, 'glfwGetVersion'):\n library_handle.glfwGetVersion(major, minor, rev)\n version = (major_value.value,\n minor_value.value,\n rev_value.value)\n return version\n else:\n return None\n\n try:\n input_func = raw_input\n except NameError:\n input_func = input\n filename = input_func().strip()\n\n try:\n library_handle = ctypes.CDLL(filename)\n except OSError:\n pass\n else:\n version = get_version(library_handle)\n print(version)\n \"\"\"\n args = [sys.executable, '-c', textwrap.dedent(version_checker_source)]\n process = subprocess.Popen(args, universal_newlines=True, stdin=\n subprocess.PIPE, stdout=subprocess.PIPE)\n out = process.communicate(_to_char_p(filename))[0]\n out = out.strip()\n if out:\n return eval(out)\n else:\n return None\n",
"role": "user"
},
{
"content": "glfw_get_version",
"role": "assistant"
}
] |
def x(cbfun):
global _error_callback
previous_callback = _error_callback
if cbfun is None:
cbfun = 0
c_cbfun = _GLFWerrorfun(cbfun)
_error_callback = cbfun, c_cbfun
cbfun = c_cbfun
_glfw.glfwSetErrorCallback(cbfun)
if previous_callback is not None and previous_callback[0] != 0:
return previous_callback[0]
| set_error_callback | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cbfun):\n global _error_callback\n previous_callback = _error_callback\n if cbfun is None:\n cbfun = 0\n c_cbfun = _GLFWerrorfun(cbfun)\n _error_callback = cbfun, c_cbfun\n cbfun = c_cbfun\n _glfw.glfwSetErrorCallback(cbfun)\n if previous_callback is not None and previous_callback[0] != 0:\n return previous_callback[0]\n",
"role": "user"
},
{
"content": "set_error_callback",
"role": "assistant"
}
] |
def x(window):
_glfw.glfwDestroyWindow(window)
window_addr = ctypes.cast(ctypes.pointer(window), ctypes.POINTER(ctypes
.c_ulong)).contents.value
for callback_repository in _callback_repositories:
if window_addr in callback_repository:
del callback_repository[window_addr]
| destroy_window | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(window):\n _glfw.glfwDestroyWindow(window)\n window_addr = ctypes.cast(ctypes.pointer(window), ctypes.POINTER(ctypes\n .c_ulong)).contents.value\n for callback_repository in _callback_repositories:\n if window_addr in callback_repository:\n del callback_repository[window_addr]\n",
"role": "user"
},
{
"content": "destroy_window",
"role": "assistant"
}
] |
def x(self):
size = self.width, self.height
bits = self.red_bits, self.green_bits, self.blue_bits
return size, bits, self.refresh_rate
| unwrap | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n size = self.width, self.height\n bits = self.red_bits, self.green_bits, self.blue_bits\n return size, bits, self.refresh_rate\n",
"role": "user"
},
{
"content": "unwrap",
"role": "assistant"
}
] |
def x(self, gammaramp):
red, green, blue = gammaramp
size = min(len(red), len(green), len(blue))
array_type = ctypes.c_ushort * size
self.size = ctypes.c_uint(size)
self.red_array = array_type()
self.green_array = array_type()
self.blue_array = array_type()
for i in range(self.size):
self.red_array[i] = int(red[i] * 65535)
self.green_array[i] = int(green[i] * 65535)
self.blue_array[i] = int(blue[i] * 65535)
pointer_type = ctypes.POINTER(ctypes.c_ushort)
self.red = ctypes.cast(self.red_array, pointer_type)
self.green = ctypes.cast(self.green_array, pointer_type)
self.blue = ctypes.cast(self.blue_array, pointer_type)
| wrap | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, gammaramp):\n red, green, blue = gammaramp\n size = min(len(red), len(green), len(blue))\n array_type = ctypes.c_ushort * size\n self.size = ctypes.c_uint(size)\n self.red_array = array_type()\n self.green_array = array_type()\n self.blue_array = array_type()\n for i in range(self.size):\n self.red_array[i] = int(red[i] * 65535)\n self.green_array[i] = int(green[i] * 65535)\n self.blue_array[i] = int(blue[i] * 65535)\n pointer_type = ctypes.POINTER(ctypes.c_ushort)\n self.red = ctypes.cast(self.red_array, pointer_type)\n self.green = ctypes.cast(self.green_array, pointer_type)\n self.blue = ctypes.cast(self.blue_array, pointer_type)\n",
"role": "user"
},
{
"content": "wrap",
"role": "assistant"
}
] |
def x(self):
red = [(self.red[i] / 65535.0) for i in range(self.size)]
green = [(self.green[i] / 65535.0) for i in range(self.size)]
blue = [(self.blue[i] / 65535.0) for i in range(self.size)]
return red, green, blue
| unwrap | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n red = [(self.red[i] / 65535.0) for i in range(self.size)]\n green = [(self.green[i] / 65535.0) for i in range(self.size)]\n blue = [(self.blue[i] / 65535.0) for i in range(self.size)]\n return red, green, blue\n",
"role": "user"
},
{
"content": "unwrap",
"role": "assistant"
}
] |
def x(value):
r = (value >> 8 * 2 & 255) / 255.0
g = (value >> 8 * 1 & 255) / 255.0
b = (value >> 8 * 0 & 255) / 255.0
return r, g, b
| hex_color | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(value):\n r = (value >> 8 * 2 & 255) / 255.0\n g = (value >> 8 * 1 & 255) / 255.0\n b = (value >> 8 * 0 & 255) / 255.0\n return r, g, b\n",
"role": "user"
},
{
"content": "hex_color",
"role": "assistant"
}
] |
def x(vector):
d = sum(x * x for x in vector) ** 0.5
return tuple(x / d for x in vector)
| normalize | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(vector):\n d = sum(x * x for x in vector) ** 0.5\n return tuple(x / d for x in vector)\n",
"role": "user"
},
{
"content": "normalize",
"role": "assistant"
}
] |
def x(p1, p2):
return sum((a - b) ** 2 for a, b in zip(p1, p2)) ** 0.5
| distance | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(p1, p2):\n return sum((a - b) ** 2 for a, b in zip(p1, p2)) ** 0.5\n",
"role": "user"
},
{
"content": "distance",
"role": "assistant"
}
] |
def x(v1, v2):
return v1[1] * v2[2] - v1[2] * v2[1], v1[2] * v2[0] - v1[0] * v2[2], v1[0
] * v2[1] - v1[1] * v2[0]
| cross | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(v1, v2):\n return v1[1] * v2[2] - v1[2] * v2[1], v1[2] * v2[0] - v1[0] * v2[2], v1[0\n ] * v2[1] - v1[1] * v2[0]\n",
"role": "user"
},
{
"content": "cross",
"role": "assistant"
}
] |
def x(v1, v2):
x1, y1, z1 = v1
x2, y2, z2 = v2
return x1 * x2 + y1 * y2 + z1 * z2
| dot | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(v1, v2):\n x1, y1, z1 = v1\n x2, y2, z2 = v2\n return x1 * x2 + y1 * y2 + z1 * z2\n",
"role": "user"
},
{
"content": "dot",
"role": "assistant"
}
] |
def x(v1, v2):
return tuple(a + b for a, b in zip(v1, v2))
| add | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(v1, v2):\n return tuple(a + b for a, b in zip(v1, v2))\n",
"role": "user"
},
{
"content": "add",
"role": "assistant"
}
] |
def x(v1, v2):
return tuple(a - b for a, b in zip(v1, v2))
| sub | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(v1, v2):\n return tuple(a - b for a, b in zip(v1, v2))\n",
"role": "user"
},
{
"content": "sub",
"role": "assistant"
}
] |
def x(v1, v2, t):
return add(v1, mul(sub(v2, v1), t))
| interpolate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(v1, v2, t):\n return add(v1, mul(sub(v2, v1), t))\n",
"role": "user"
},
{
"content": "interpolate",
"role": "assistant"
}
] |
def x(a, b, c):
x1, y1, z1 = a
x2, y2, z2 = b
x3, y3, z3 = c
ab = x2 - x1, y2 - y1, z2 - z1
ac = x3 - x1, y3 - y1, z3 - z1
x, y, z = cross(ab, ac)
d = (x * x + y * y + z * z) ** 0.5
return x / d, y / d, z / d
| normal_from_points | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(a, b, c):\n x1, y1, z1 = a\n x2, y2, z2 = b\n x3, y3, z3 = c\n ab = x2 - x1, y2 - y1, z2 - z1\n ac = x3 - x1, y3 - y1, z3 - z1\n x, y, z = cross(ab, ac)\n d = (x * x + y * y + z * z) ** 0.5\n return x / d, y / d, z / d\n",
"role": "user"
},
{
"content": "normal_from_points",
"role": "assistant"
}
] |
def x(positions, normals):
lookup = defaultdict(list)
for position, normal in zip(positions, normals):
lookup[position].append(normal)
result = []
for position in positions:
tx = ty = tz = 0
for x, y, z in lookup[position]:
tx += x
ty += y
tz += z
d = (tx * tx + ty * ty + tz * tz) ** 0.5
result.append((tx / d, ty / d, tz / d))
return result
| smooth_normals | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(positions, normals):\n lookup = defaultdict(list)\n for position, normal in zip(positions, normals):\n lookup[position].append(normal)\n result = []\n for position in positions:\n tx = ty = tz = 0\n for x, y, z in lookup[position]:\n tx += x\n ty += y\n tz += z\n d = (tx * tx + ty * ty + tz * tz) ** 0.5\n result.append((tx / d, ty / d, tz / d))\n return result\n",
"role": "user"
},
{
"content": "smooth_normals",
"role": "assistant"
}
] |
def x(positions):
x0, y0, z0 = x1, y1, z1 = positions[0]
for x, y, z in positions:
x0 = min(x0, x)
y0 = min(y0, y)
z0 = min(z0, z)
x1 = max(x1, x)
y1 = max(y1, y)
z1 = max(z1, z)
return (x0, y0, z0), (x1, y1, z1)
| bounding_box | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(positions):\n x0, y0, z0 = x1, y1, z1 = positions[0]\n for x, y, z in positions:\n x0 = min(x0, x)\n y0 = min(y0, y)\n z0 = min(z0, z)\n x1 = max(x1, x)\n y1 = max(y1, y)\n z1 = max(z1, z)\n return (x0, y0, z0), (x1, y1, z1)\n",
"role": "user"
},
{
"content": "bounding_box",
"role": "assistant"
}
] |
def x(positions):
(x0, y0, z0), (x1, y1, z1) = bounding_box(positions)
dx = x1 - (x1 - x0) / 2.0
dy = y1 - (y1 - y0) / 2.0
dz = z1 - (z1 - z0) / 2.0
result = []
for x, y, z in positions:
result.append((x - dx, y - dy, z - dz))
return result
| recenter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(positions):\n (x0, y0, z0), (x1, y1, z1) = bounding_box(positions)\n dx = x1 - (x1 - x0) / 2.0\n dy = y1 - (y1 - y0) / 2.0\n dz = z1 - (z1 - z0) / 2.0\n result = []\n for x, y, z in positions:\n result.append((x - dx, y - dy, z - dz))\n return result\n",
"role": "user"
},
{
"content": "recenter",
"role": "assistant"
}
] |
def x(*args):
result = []
for array in zip(*args):
result.append(tuple(flatten(array)))
return result
| interleave | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(*args):\n result = []\n for array in zip(*args):\n result.append(tuple(flatten(array)))\n return result\n",
"role": "user"
},
{
"content": "interleave",
"role": "assistant"
}
] |
def x(iterable, keyfunc=None):
seen = set()
for item in iterable:
key = item if keyfunc is None else keyfunc(item)
if key not in seen:
seen.add(key)
yield item
| distinct | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(iterable, keyfunc=None):\n seen = set()\n for item in iterable:\n key = item if keyfunc is None else keyfunc(item)\n if key not in seen:\n seen.add(key)\n yield item\n",
"role": "user"
},
{
"content": "distinct",
"role": "assistant"
}
] |
def x(v1, v2, v3, o, d):
eps = 1e-06
e1 = sub(v2, v1)
e2 = sub(v3, v1)
p = cross(d, e2)
det = dot(e1, p)
if abs(det) < eps:
return None
inv = 1.0 / det
t = sub(o, v1)
u = dot(t, p) * inv
if u < 0 or u > 1:
return None
q = cross(t, e1)
v = dot(d, q) * inv
if v < 0 or v > 1:
return None
t = dot(e2, q) * inv
if t > eps:
return t
return None
| ray_triangle_intersection | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(v1, v2, v3, o, d):\n eps = 1e-06\n e1 = sub(v2, v1)\n e2 = sub(v3, v1)\n p = cross(d, e2)\n det = dot(e1, p)\n if abs(det) < eps:\n return None\n inv = 1.0 / det\n t = sub(o, v1)\n u = dot(t, p) * inv\n if u < 0 or u > 1:\n return None\n q = cross(t, e1)\n v = dot(d, q) * inv\n if v < 0 or v > 1:\n return None\n t = dot(e2, q) * inv\n if t > eps:\n return t\n return None\n",
"role": "user"
},
{
"content": "ray_triangle_intersection",
"role": "assistant"
}
] |
def x(fmt, data):
func = struct.Struct(fmt).pack
return create_string_buffer(''.join([func(x) for x in data]))
| pack_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fmt, data):\n func = struct.Struct(fmt).pack\n return create_string_buffer(''.join([func(x) for x in data]))\n",
"role": "user"
},
{
"content": "pack_list",
"role": "assistant"
}
] |
def x(self, jquery=False):
if jquery:
e = JQuery(self)
e.click()
else:
super(Clickable, self).click()
| click | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, jquery=False):\n if jquery:\n e = JQuery(self)\n e.click()\n else:\n super(Clickable, self).click()\n",
"role": "user"
},
{
"content": "click",
"role": "assistant"
}
] |
def x(cookie, keys_map=WEB_DRIVER_COOKIE_KEYS_MAP):
cookie_dict = dict()
for k in keys_map.keys():
key = _to_unicode_if_str(keys_map[k])
value = _to_unicode_if_str(getattr(cookie, k))
cookie_dict[key] = value
return cookie_dict
| convert_cookie_to_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cookie, keys_map=WEB_DRIVER_COOKIE_KEYS_MAP):\n cookie_dict = dict()\n for k in keys_map.keys():\n key = _to_unicode_if_str(keys_map[k])\n value = _to_unicode_if_str(getattr(cookie, k))\n cookie_dict[key] = value\n return cookie_dict\n",
"role": "user"
},
{
"content": "convert_cookie_to_dict",
"role": "assistant"
}
] |
def x(driver, cookies):
for cookie in cookies:
driver.add_cookie(convert_cookie_to_dict(cookie))
return driver
| add_cookies_to_web_driver | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(driver, cookies):\n for cookie in cookies:\n driver.add_cookie(convert_cookie_to_dict(cookie))\n return driver\n",
"role": "user"
},
{
"content": "add_cookies_to_web_driver",
"role": "assistant"
}
] |
def x(self, options, conf):
self.conf = conf
self.when = options.browser_closer_when
| configure | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, options, conf):\n self.conf = conf\n self.when = options.browser_closer_when\n",
"role": "user"
},
{
"content": "configure",
"role": "assistant"
}
] |
def x(self, root):
basename = os.path.basename(root)
if os.path.splitext(basename)[0] != '__init__' and basename.startswith('_'
):
return
location = self._determine_location_for(root)
if os.path.isfile(root):
self._index_module(root, location)
elif os.path.isdir(root) and os.path.exists(os.path.join(root,
'__init__.py')):
self._index_package(root, location)
| index_path | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, root):\n basename = os.path.basename(root)\n if os.path.splitext(basename)[0] != '__init__' and basename.startswith('_'\n ):\n return\n location = self._determine_location_for(root)\n if os.path.isfile(root):\n self._index_module(root, location)\n elif os.path.isdir(root) and os.path.exists(os.path.join(root,\n '__init__.py')):\n self._index_package(root, location)\n",
"role": "user"
},
{
"content": "index_path",
"role": "assistant"
}
] |
def x(self, paths=None, name=None, refresh=False):
if not paths:
paths = sys.path
if not name:
name = 'default'
self._name = name
idx_dir = get_cache_dir()
idx_file = os.path.join(idx_dir, name + '.json')
if os.path.exists(idx_file) and not refresh:
with open(idx_file) as fd:
self.deserialize(fd)
else:
self.build_index(paths)
with open(idx_file, 'w') as fd:
self.serialize(fd)
return self
| get_or_create_index | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, paths=None, name=None, refresh=False):\n if not paths:\n paths = sys.path\n if not name:\n name = 'default'\n self._name = name\n idx_dir = get_cache_dir()\n idx_file = os.path.join(idx_dir, name + '.json')\n if os.path.exists(idx_file) and not refresh:\n with open(idx_file) as fd:\n self.deserialize(fd)\n else:\n self.build_index(paths)\n with open(idx_file, 'w') as fd:\n self.serialize(fd)\n return self\n",
"role": "user"
},
{
"content": "get_or_create_index",
"role": "assistant"
}
] |
def x(self, symbol):
scores = []
path = []
def x(module, variable):
prefix = module.split('.')
if variable is not None:
prefix.append(variable)
seeking = symbol.split('.')
new_module = []
while prefix and seeking[0] != prefix[0]:
new_module.append(prefix.pop(0))
if new_module:
module, variable = '.'.join(new_module), prefix[0]
else:
variable = None
return module, variable
def x(scope, scale):
sub_path, score = self._score_key(scope, full_key)
if score > 0.1:
try:
i = sub_path.index(None)
sub_path, from_symbol = sub_path[:i], '.'.join(sub_path[i + 1:]
)
except ValueError:
from_symbol = None
package_path = '.'.join(path + sub_path)
package_path, from_symbol = fixup(package_path, from_symbol)
scores.append((score * scale, package_path, from_symbol))
for key, subscope in scope._tree.items():
if type(subscope) is not float:
path.append(key)
score_walk(subscope, subscope.score * scale - 0.1)
path.pop()
full_key = symbol.split('.')
score_walk(self, 1.0)
scores.sort(reverse=True)
return scores
| symbol_scores | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, symbol):\n scores = []\n path = []\n\n def x(module, variable):\n prefix = module.split('.')\n if variable is not None:\n prefix.append(variable)\n seeking = symbol.split('.')\n new_module = []\n while prefix and seeking[0] != prefix[0]:\n new_module.append(prefix.pop(0))\n if new_module:\n module, variable = '.'.join(new_module), prefix[0]\n else:\n variable = None\n return module, variable\n\n def x(scope, scale):\n sub_path, score = self._score_key(scope, full_key)\n if score > 0.1:\n try:\n i = sub_path.index(None)\n sub_path, from_symbol = sub_path[:i], '.'.join(sub_path[i + 1:]\n )\n except ValueError:\n from_symbol = None\n package_path = '.'.join(path + sub_path)\n package_path, from_symbol = fixup(package_path, from_symbol)\n scores.append((score * scale, package_path, from_symbol))\n for key, subscope in scope._tree.items():\n if type(subscope) is not float:\n path.append(key)\n score_walk(subscope, subscope.score * scale - 0.1)\n path.pop()\n full_key = symbol.split('.')\n score_walk(self, 1.0)\n scores.sort(reverse=True)\n return scores\n",
"role": "user"
},
{
"content": "symbol_scores",
"role": "assistant"
}
] |
def x(self, path):
path = path.split('.')
node = self
while node._parent:
node = node._parent
for name in path:
node = node._tree.get(name, None)
if node is None or type(node) is float:
return None
return node
| find | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, path):\n path = path.split('.')\n node = self\n while node._parent:\n node = node._parent\n for name in path:\n node = node._tree.get(name, None)\n if node is None or type(node) is float:\n return None\n return node\n",
"role": "user"
},
{
"content": "find",
"role": "assistant"
}
] |
def x(self, path):
path = path.split('.')
node = self
while node._parent:
node = node._parent
location = node.location
for name in path:
tree = node._tree.get(name, None)
if tree is None or type(tree) is float:
return location
location = tree.location
return location
| location_for | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, path):\n path = path.split('.')\n node = self\n while node._parent:\n node = node._parent\n location = node.location\n for name in path:\n tree = node._tree.get(name, None)\n if tree is None or type(tree) is float:\n return location\n location = tree.location\n return location\n",
"role": "user"
},
{
"content": "location_for",
"role": "assistant"
}
] |
def x(self, option):
items_list = self.get_options()
for item in items_list:
if item.get_attribute('value') == option:
item.click()
break
| select_option | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, option):\n items_list = self.get_options()\n for item in items_list:\n if item.get_attribute('value') == option:\n item.click()\n break\n",
"role": "user"
},
{
"content": "select_option",
"role": "assistant"
}
] |
def x(self, attribute):
items_list = self.get_options()
return next(iter([item.get_attribute(attribute) for item in items_list if
item.is_selected()]), None)
| get_attribute_selected | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, attribute):\n items_list = self.get_options()\n return next(iter([item.get_attribute(attribute) for item in items_list if\n item.is_selected()]), None)\n",
"role": "user"
},
{
"content": "get_attribute_selected",
"role": "assistant"
}
] |
def x(self, text):
xpath = './/option[normalize-space(.) = {0}]'.format(self.
_escape_string(text))
opts = self.find_elements_by_xpath(xpath)
matched = False
for opt in opts:
self._set_selected(opt)
if not self.is_multiple:
return
matched = True
if len(opts) == 0 and ' ' in text:
sub_string_without_space = self._get_longest_token(text)
if sub_string_without_space == '':
candidates = self.get_options()
else:
xpath = './/option[contains(.,{0})]'.format(self._escape_string
(sub_string_without_space))
candidates = self.find_elements_by_xpath(xpath)
for candidate in candidates:
if text == candidate.text:
self._set_selected(candidate)
if not self.is_multiple:
return
matched = True
if not matched:
raise NoSuchElementException(
'Could not locate element with visible text: ' + str(text))
| select_by_visible_text | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, text):\n xpath = './/option[normalize-space(.) = {0}]'.format(self.\n _escape_string(text))\n opts = self.find_elements_by_xpath(xpath)\n matched = False\n for opt in opts:\n self._set_selected(opt)\n if not self.is_multiple:\n return\n matched = True\n if len(opts) == 0 and ' ' in text:\n sub_string_without_space = self._get_longest_token(text)\n if sub_string_without_space == '':\n candidates = self.get_options()\n else:\n xpath = './/option[contains(.,{0})]'.format(self._escape_string\n (sub_string_without_space))\n candidates = self.find_elements_by_xpath(xpath)\n for candidate in candidates:\n if text == candidate.text:\n self._set_selected(candidate)\n if not self.is_multiple:\n return\n matched = True\n if not matched:\n raise NoSuchElementException(\n 'Could not locate element with visible text: ' + str(text))\n",
"role": "user"
},
{
"content": "select_by_visible_text",
"role": "assistant"
}
] |
def x(*args, **kwargs):
kwargs.setdefault('sleep_seconds', (1, None))
kwargs.setdefault('expected_exceptions', WebDriverException)
kwargs.setdefault('timeout_seconds', webium.settings.wait_timeout)
return wait_lib(*args, **kwargs)
| wait | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(*args, **kwargs):\n kwargs.setdefault('sleep_seconds', (1, None))\n kwargs.setdefault('expected_exceptions', WebDriverException)\n kwargs.setdefault('timeout_seconds', webium.settings.wait_timeout)\n return wait_lib(*args, **kwargs)\n",
"role": "user"
},
{
"content": "wait",
"role": "assistant"
}
] |
def x(source, filename=None):
if isinstance(source, text_type) and sys.version_info[0] == 2:
source = CODING_COOKIE_RE.sub('\\1', source, 1)
return ast.parse(source, filename or '<unknown>')
| parse_ast | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(source, filename=None):\n if isinstance(source, text_type) and sys.version_info[0] == 2:\n source = CODING_COOKIE_RE.sub('\\\\1', source, 1)\n return ast.parse(source, filename or '<unknown>')\n",
"role": "user"
},
{
"content": "parse_ast",
"role": "assistant"
}
] |
def x(self):
unresolved = set()
unreferenced = self._definitions.copy()
self._collect_unresolved_and_unreferenced(set(), set(), unresolved,
unreferenced, frozenset(self._definitions), start=True)
return unresolved, unreferenced - Scope.ALL_BUILTINS
| find_unresolved_and_unreferenced_symbols | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n unresolved = set()\n unreferenced = self._definitions.copy()\n self._collect_unresolved_and_unreferenced(set(), set(), unresolved,\n unreferenced, frozenset(self._definitions), start=True)\n return unresolved, unreferenced - Scope.ALL_BUILTINS\n",
"role": "user"
},
{
"content": "find_unresolved_and_unreferenced_symbols",
"role": "assistant"
}
] |
def x(key):
CACHED_KEY_FILE = os.path.join(CURRENT_DIR, key)
try:
return json.loads(open(CACHED_KEY_FILE, 'rb').read().decode('UTF-8'))[
'_']
except (IOError, ValueError):
return None
| get_item | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(key):\n CACHED_KEY_FILE = os.path.join(CURRENT_DIR, key)\n try:\n return json.loads(open(CACHED_KEY_FILE, 'rb').read().decode('UTF-8'))[\n '_']\n except (IOError, ValueError):\n return None\n",
"role": "user"
},
{
"content": "get_item",
"role": "assistant"
}
] |
def x(key, value):
CACHED_KEY_FILE = os.path.join(CURRENT_DIR, key)
open(CACHED_KEY_FILE, 'wb').write(json.dumps({'_': value}).encode('UTF-8'))
return value
| set_item | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(key, value):\n CACHED_KEY_FILE = os.path.join(CURRENT_DIR, key)\n open(CACHED_KEY_FILE, 'wb').write(json.dumps({'_': value}).encode('UTF-8'))\n return value\n",
"role": "user"
},
{
"content": "set_item",
"role": "assistant"
}
] |
def x(key):
CACHED_KEY_FILE = os.path.join(CURRENT_DIR, key)
if os.path.isfile(CACHED_KEY_FILE):
os.remove(CACHED_KEY_FILE)
| delete_item | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(key):\n CACHED_KEY_FILE = os.path.join(CURRENT_DIR, key)\n if os.path.isfile(CACHED_KEY_FILE):\n os.remove(CACHED_KEY_FILE)\n",
"role": "user"
},
{
"content": "delete_item",
"role": "assistant"
}
] |
def x(self, data):
if isinstance(data, dict) or isinstance(data, list):
self._raw_data = data
self._json_data = copy.deepcopy(self._raw_data)
else:
raise TypeError('Provided Data is not json')
| parse_json_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n if isinstance(data, dict) or isinstance(data, list):\n self._raw_data = data\n self._json_data = copy.deepcopy(self._raw_data)\n else:\n raise TypeError('Provided Data is not json')\n",
"role": "user"
},
{
"content": "parse_json_data",
"role": "assistant"
}
] |
def x(self, file_path):
if file_path == '' or os.path.splitext(file_path)[1] != '.json':
raise IOError('Invalid Json file')
with open(file_path) as json_file:
self._raw_data = json.load(json_file)
self._json_data = copy.deepcopy(self._raw_data)
| parse_json_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, file_path):\n if file_path == '' or os.path.splitext(file_path)[1] != '.json':\n raise IOError('Invalid Json file')\n with open(file_path) as json_file:\n self._raw_data = json.load(json_file)\n self._json_data = copy.deepcopy(self._raw_data)\n",
"role": "user"
},
{
"content": "parse_json_file",
"role": "assistant"
}
] |
def x(self, key, data):
if key.isdigit():
return data[int(key)]
if key not in data:
raise KeyError('Key not exists')
return data.get(key)
| get_value_from_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, data):\n if key.isdigit():\n return data[int(key)]\n if key not in data:\n raise KeyError('Key not exists')\n return data.get(key)\n",
"role": "user"
},
{
"content": "get_value_from_data",
"role": "assistant"
}
] |
def x(self, root):
leafs = root.strip(' ').split('.')
for leaf in leafs:
if leaf:
self._json_data = self.__get_value_from_data(leaf, self._json_data)
return self
| at | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, root):\n leafs = root.strip(' ').split('.')\n for leaf in leafs:\n if leaf:\n self._json_data = self.__get_value_from_data(leaf, self._json_data)\n return self\n",
"role": "user"
},
{
"content": "at",
"role": "assistant"
}
] |
def x(self, data={}):
if data and (isinstance(data, dict) or isinstance(data, list)):
self._json_data = data
else:
self._json_data = copy.deepcopy(self._raw_data)
self.__reset_queries()
return self
| reset | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data={}):\n if data and (isinstance(data, dict) or isinstance(data, list)):\n self._json_data = data\n else:\n self._json_data = copy.deepcopy(self._raw_data)\n self.__reset_queries()\n return self\n",
"role": "user"
},
{
"content": "reset",
"role": "assistant"
}
] |
def x(self, query_items):
temp_index = self._current_query_index
if len(self._queries) - 1 < temp_index:
self._queries.append([])
self._queries[temp_index].append(query_items)
| store_query | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, query_items):\n temp_index = self._current_query_index\n if len(self._queries) - 1 < temp_index:\n self._queries.append([])\n self._queries[temp_index].append(query_items)\n",
"role": "user"
},
{
"content": "store_query",
"role": "assistant"
}
] |
def x(self):
def x(item):
or_check = False
for queries in self._queries:
and_check = True
for query in queries:
and_check &= self._matcher._match(item.get(query.get('key'),
None), query.get('operator'), query.get('value'))
or_check |= and_check
return or_check
self._json_data = list(filter(lambda item: func(item), self._json_data))
| execute_queries | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n\n def x(item):\n or_check = False\n for queries in self._queries:\n and_check = True\n for query in queries:\n and_check &= self._matcher._match(item.get(query.get('key'),\n None), query.get('operator'), query.get('value'))\n or_check |= and_check\n return or_check\n self._json_data = list(filter(lambda item: func(item), self._json_data))\n",
"role": "user"
},
{
"content": "execute_queries",
"role": "assistant"
}
] |
def x(self, key, operator, value):
self.__store_query({'key': key, 'operator': operator, 'value': value})
return self
| where | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, operator, value):\n self.__store_query({'key': key, 'operator': operator, 'value': value})\n return self\n",
"role": "user"
},
{
"content": "where",
"role": "assistant"
}
] |
def x(self, key, operator, value):
if len(self._queries) > 0:
self._current_query_index += 1
self.__store_query({'key': key, 'operator': operator, 'value': value})
return self
| or_where | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, operator, value):\n if len(self._queries) > 0:\n self._current_query_index += 1\n self.__store_query({'key': key, 'operator': operator, 'value': value})\n return self\n",
"role": "user"
},
{
"content": "or_where",
"role": "assistant"
}
] |
def x(self, index):
self.__prepare()
return None if self.count() < math.fabs(index) else self._json_data[index]
| nth | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, index):\n self.__prepare()\n return None if self.count() < math.fabs(index) else self._json_data[index]\n",
"role": "user"
},
{
"content": "nth",
"role": "assistant"
}
] |
def x(self, property):
self.__prepare()
total = 0
for i in self._json_data:
total += i.get(property)
return total
| sum | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, property):\n self.__prepare()\n total = 0\n for i in self._json_data:\n total += i.get(property)\n return total\n",
"role": "user"
},
{
"content": "sum",
"role": "assistant"
}
] |
def x(self, property):
self.__prepare()
try:
return max(self._json_data, key=lambda x: x[property]).get(property)
except KeyError:
raise KeyError('Key is not exists')
| max | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, property):\n self.__prepare()\n try:\n return max(self._json_data, key=lambda x: x[property]).get(property)\n except KeyError:\n raise KeyError('Key is not exists')\n",
"role": "user"
},
{
"content": "max",
"role": "assistant"
}
] |
def x(self, property):
self.__prepare()
return self.sum(property) / self.count()
| avg | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, property):\n self.__prepare()\n return self.sum(property) / self.count()\n",
"role": "user"
},
{
"content": "avg",
"role": "assistant"
}
] |
def x(self, size=0):
if size == 0:
raise ValueError('Invalid chunk size')
self.__prepare()
_new_content = []
while len(self._json_data) > 0:
_new_content.append(self._json_data[0:size])
self._json_data = self._json_data[size:]
self._json_data = _new_content
return self._json_data
| chunk | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, size=0):\n if size == 0:\n raise ValueError('Invalid chunk size')\n self.__prepare()\n _new_content = []\n while len(self._json_data) > 0:\n _new_content.append(self._json_data[0:size])\n self._json_data = self._json_data[size:]\n self._json_data = _new_content\n return self._json_data\n",
"role": "user"
},
{
"content": "chunk",
"role": "assistant"
}
] |
def x(self, property):
self.__prepare()
group_data = {}
for data in self._json_data:
if data[property] not in group_data:
group_data[data[property]] = []
group_data[data[property]].append(data)
self._json_data = group_data
return self
| group_by | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, property):\n self.__prepare()\n group_data = {}\n for data in self._json_data:\n if data[property] not in group_data:\n group_data[data[property]] = []\n group_data[data[property]].append(data)\n self._json_data = group_data\n return self\n",
"role": "user"
},
{
"content": "group_by",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.