source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(self, album, artist): params = collections.OrderedDict() params['gbv'] = '2' params['q'] = '"%s" "%s" front cover' % (artist, album) if abs(self.target_size - 500) < 300: params['tbs'] = 'isz:m' elif self.target_size > 800: params['tbs'] = 'isz:l' return __class__.assembleUrl(__class__.BASE_URL, params)
getSearchUrl
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, album, artist):\n params = collections.OrderedDict()\n params['gbv'] = '2'\n params['q'] = '\"%s\" \"%s\" front cover' % (artist, album)\n if abs(self.target_size - 500) < 300:\n params['tbs'] = 'isz:m'\n elif self.target_size > 800:\n params['tbs'] = 'isz:l'\n return __class__.assembleUrl(__class__.BASE_URL, params)\n", "role": "user" }, { "content": "getSearchUrl", "role": "assistant" } ]
async def parseResults(self, api_data): """ See CoverSource.parseResults. """ results = [] parser = lxml.etree.HTMLParser() html = lxml.etree.XML(api_data.decode('latin-1'), parser) for rank, result in enumerate(__class__.RESULTS_SELECTOR(html), 1): metadata_div = result.find('div') metadata_json = lxml.etree.tostring(metadata_div, encoding= 'unicode', method='text') metadata_json = json.loads(metadata_json) google_url = result.find('a').get('href') if google_url is not None: query = urllib.parse.urlsplit(google_url).query else: query = None if not query: img_url = metadata_json['ou'] else: query = urllib.parse.parse_qs(query) img_url = query['imgurl'][0] check_metadata = CoverImageMetadata.NONE format = metadata_json['ity'].lower() try: format = SUPPORTED_IMG_FORMATS[format] except KeyError: format = None check_metadata = CoverImageMetadata.FORMAT if not query: size = metadata_json['ow'], metadata_json['oh'] else: size = tuple(map(int, (query['w'][0], query['h'][0]))) thumbnail_url = metadata_json['tu'] results.append(GoogleImagesCoverSourceResult(img_url, size, format, thumbnail_url=thumbnail_url, source=self, rank=rank, check_metadata=check_metadata)) return results
parseResults
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def parseResults(self, api_data):\n \"\"\" See CoverSource.parseResults. \"\"\"\n results = []\n parser = lxml.etree.HTMLParser()\n html = lxml.etree.XML(api_data.decode('latin-1'), parser)\n for rank, result in enumerate(__class__.RESULTS_SELECTOR(html), 1):\n metadata_div = result.find('div')\n metadata_json = lxml.etree.tostring(metadata_div, encoding=\n 'unicode', method='text')\n metadata_json = json.loads(metadata_json)\n google_url = result.find('a').get('href')\n if google_url is not None:\n query = urllib.parse.urlsplit(google_url).query\n else:\n query = None\n if not query:\n img_url = metadata_json['ou']\n else:\n query = urllib.parse.parse_qs(query)\n img_url = query['imgurl'][0]\n check_metadata = CoverImageMetadata.NONE\n format = metadata_json['ity'].lower()\n try:\n format = SUPPORTED_IMG_FORMATS[format]\n except KeyError:\n format = None\n check_metadata = CoverImageMetadata.FORMAT\n if not query:\n size = metadata_json['ow'], metadata_json['oh']\n else:\n size = tuple(map(int, (query['w'][0], query['h'][0])))\n thumbnail_url = metadata_json['tu']\n results.append(GoogleImagesCoverSourceResult(img_url, size, format,\n thumbnail_url=thumbnail_url, source=self, rank=rank,\n check_metadata=check_metadata))\n return results\n", "role": "user" }, { "content": "parseResults", "role": "assistant" } ]
async def waitAccessAsync(self): """ Wait the needed time before sending a request to honor rate limit. """ async with self.lock: while True: last_access_ts = self.__getLastAccess() if last_access_ts is not None: now = time.time() last_access_ts = last_access_ts[0] time_since_last_access = now - last_access_ts if time_since_last_access < self.min_delay_between_accesses: time_to_wait = (self.min_delay_between_accesses - time_since_last_access) if self.jitter_range_ms is not None: time_to_wait += random.randint(*self.jitter_range_ms ) / 1000 self.logger.debug( 'Sleeping for %.2fms because of rate limit for domain %s' % (time_to_wait * 1000, self.domain)) await asyncio.sleep(time_to_wait) access_time = time.time() self.__access(access_time) last_access_ts = self.__getLastAccess() if last_access_ts[0] == access_time: break
waitAccessAsync
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def waitAccessAsync(self):\n \"\"\" Wait the needed time before sending a request to honor rate limit. \"\"\"\n async with self.lock:\n while True:\n last_access_ts = self.__getLastAccess()\n if last_access_ts is not None:\n now = time.time()\n last_access_ts = last_access_ts[0]\n time_since_last_access = now - last_access_ts\n if time_since_last_access < self.min_delay_between_accesses:\n time_to_wait = (self.min_delay_between_accesses -\n time_since_last_access)\n if self.jitter_range_ms is not None:\n time_to_wait += random.randint(*self.jitter_range_ms\n ) / 1000\n self.logger.debug(\n 'Sleeping for %.2fms because of rate limit for domain %s'\n % (time_to_wait * 1000, self.domain))\n await asyncio.sleep(time_to_wait)\n access_time = time.time()\n self.__access(access_time)\n last_access_ts = self.__getLastAccess()\n if last_access_ts[0] == access_time:\n break\n", "role": "user" }, { "content": "waitAccessAsync", "role": "assistant" } ]
def x(self, ts): with self.connection: self.connection.execute( 'INSERT OR REPLACE INTO access_timestamp (timestamp, domain) VALUES (?, ?)' , (ts, self.domain))
access
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ts):\n with self.connection:\n self.connection.execute(\n 'INSERT OR REPLACE INTO access_timestamp (timestamp, domain) VALUES (?, ?)'\n , (ts, self.domain))\n", "role": "user" }, { "content": "access", "role": "assistant" } ]
def x(socket_timeout_s): return aiohttp.ClientTimeout(total=None, connect=None, sock_connect= socket_timeout_s, sock_read=socket_timeout_s)
aiohttp_socket_timeout
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(socket_timeout_s):\n return aiohttp.ClientTimeout(total=None, connect=None, sock_connect=\n socket_timeout_s, sock_read=socket_timeout_s)\n", "role": "user" }, { "content": "aiohttp_socket_timeout", "role": "assistant" } ]
async def query(self, url, *, post_data=None, headers=None, verify=True, cache=None, pre_cache_callback=None): """ Send a GET/POST request or get data from cache, retry if it fails, and return a tuple of store in cache callback, response content. """ async def store_in_cache_callback(): pass if cache is not None: if post_data is not None: if (url, post_data) in cache: self.logger.debug("Got data for URL '%s' %s from cache" % ( url, dict(post_data))) return store_in_cache_callback, cache[url, post_data] elif url in cache: self.logger.debug("Got data for URL '%s' from cache" % url) return store_in_cache_callback, cache[url] domain_rate_watcher = rate_watcher.AccessRateWatcher(self. watcher_db_filepath, url, self.min_delay_between_accesses, jitter_range_ms=self.jitter_range_ms, logger=self.logger) for attempt, time_to_sleep in enumerate(redo.retrier(max_attempts= HTTP_MAX_ATTEMPTS, sleeptime=1, max_sleeptime= HTTP_MAX_RETRY_SLEEP_S, sleepscale=1.5), 1): await domain_rate_watcher.waitAccessAsync() try: if post_data is not None: async with self.session.post(url, data=post_data, headers= self._buildHeaders(headers), timeout= HTTP_NORMAL_TIMEOUT, ssl=verify) as response: content = await response.read() else: async with self.session.get(url, headers=self._buildHeaders (headers), timeout=HTTP_NORMAL_TIMEOUT, ssl=verify ) as response: content = await response.read() if cache is not None: async def store_in_cache_callback(): if pre_cache_callback is not None: try: data = await pre_cache_callback(content) except Exception: data = content else: data = content if post_data is not None: cache[url, post_data] = data else: cache[url] = data except (asyncio.TimeoutError, aiohttp.ClientError) as e: self.logger.warning( "Querying '%s' failed (attempt %u/%u): %s %s" % (url, attempt, HTTP_MAX_ATTEMPTS, e.__class__.__qualname__, e)) if attempt == HTTP_MAX_ATTEMPTS: raise else: self.logger.debug('Retrying in %.3fs' % time_to_sleep) await asyncio.sleep(time_to_sleep) else: break response.raise_for_status() return store_in_cache_callback, content
query
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def query(self, url, *, post_data=None, headers=None, verify=True,\n cache=None, pre_cache_callback=None):\n \"\"\" Send a GET/POST request or get data from cache, retry if it fails, and return a tuple of store in cache callback, response content. \"\"\"\n\n async def store_in_cache_callback():\n pass\n if cache is not None:\n if post_data is not None:\n if (url, post_data) in cache:\n self.logger.debug(\"Got data for URL '%s' %s from cache\" % (\n url, dict(post_data)))\n return store_in_cache_callback, cache[url, post_data]\n elif url in cache:\n self.logger.debug(\"Got data for URL '%s' from cache\" % url)\n return store_in_cache_callback, cache[url]\n domain_rate_watcher = rate_watcher.AccessRateWatcher(self.\n watcher_db_filepath, url, self.min_delay_between_accesses,\n jitter_range_ms=self.jitter_range_ms, logger=self.logger)\n for attempt, time_to_sleep in enumerate(redo.retrier(max_attempts=\n HTTP_MAX_ATTEMPTS, sleeptime=1, max_sleeptime=\n HTTP_MAX_RETRY_SLEEP_S, sleepscale=1.5), 1):\n await domain_rate_watcher.waitAccessAsync()\n try:\n if post_data is not None:\n async with self.session.post(url, data=post_data, headers=\n self._buildHeaders(headers), timeout=\n HTTP_NORMAL_TIMEOUT, ssl=verify) as response:\n content = await response.read()\n else:\n async with self.session.get(url, headers=self._buildHeaders\n (headers), timeout=HTTP_NORMAL_TIMEOUT, ssl=verify\n ) as response:\n content = await response.read()\n if cache is not None:\n\n async def store_in_cache_callback():\n if pre_cache_callback is not None:\n try:\n data = await pre_cache_callback(content)\n except Exception:\n data = content\n else:\n data = content\n if post_data is not None:\n cache[url, post_data] = data\n else:\n cache[url] = data\n except (asyncio.TimeoutError, aiohttp.ClientError) as e:\n self.logger.warning(\n \"Querying '%s' failed (attempt %u/%u): %s %s\" % (url,\n attempt, HTTP_MAX_ATTEMPTS, e.__class__.__qualname__, e))\n if attempt == HTTP_MAX_ATTEMPTS:\n raise\n else:\n self.logger.debug('Retrying in %.3fs' % time_to_sleep)\n await asyncio.sleep(time_to_sleep)\n else:\n break\n response.raise_for_status()\n return store_in_cache_callback, content\n", "role": "user" }, { "content": "query", "role": "assistant" } ]
async def isReachable(self, url, *, headers=None, verify=True, response_headers=None, cache=None): """ Send a HEAD request with short timeout or get data from cache, return True if ressource has 2xx status code, False instead. """ if cache is not None and url in cache: self.logger.debug("Got headers for URL '%s' from cache" % url) resp_ok, response_headers = pickle.loads(cache[url]) return resp_ok domain_rate_watcher = rate_watcher.AccessRateWatcher(self. watcher_db_filepath, url, self.min_delay_between_accesses, jitter_range_ms=self.jitter_range_ms, logger=self.logger) resp_ok = True try: for attempt, time_to_sleep in enumerate(redo.retrier(max_attempts= HTTP_MAX_ATTEMPTS, sleeptime=0.5, max_sleeptime= HTTP_MAX_RETRY_SLEEP_SHORT_S, sleepscale=1.5), 1): await domain_rate_watcher.waitAccessAsync() try: async with self.session.head(url, headers=self. _buildHeaders(headers), timeout=HTTP_SHORT_TIMEOUT, ssl =verify) as response: pass except (asyncio.TimeoutError, aiohttp.ClientError) as e: self.logger.warning( "Probing '%s' failed (attempt %u/%u): %s %s" % (url, attempt, HTTP_MAX_ATTEMPTS, e.__class__.__qualname__, e)) if attempt == HTTP_MAX_ATTEMPTS: resp_ok = False else: self.logger.debug('Retrying in %.3fs' % time_to_sleep) await asyncio.sleep(time_to_sleep) else: response.raise_for_status() if response_headers is not None: response_headers.update(response.headers) break except aiohttp.ClientResponseError as e: self.logger.debug("Probing '%s' failed: %s %s" % (url, e.__class__. __qualname__, e)) resp_ok = False if cache is not None: cache[url] = pickle.dumps((resp_ok, response_headers)) return resp_ok
isReachable
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def isReachable(self, url, *, headers=None, verify=True,\n response_headers=None, cache=None):\n \"\"\" Send a HEAD request with short timeout or get data from cache, return True if ressource has 2xx status code, False instead. \"\"\"\n if cache is not None and url in cache:\n self.logger.debug(\"Got headers for URL '%s' from cache\" % url)\n resp_ok, response_headers = pickle.loads(cache[url])\n return resp_ok\n domain_rate_watcher = rate_watcher.AccessRateWatcher(self.\n watcher_db_filepath, url, self.min_delay_between_accesses,\n jitter_range_ms=self.jitter_range_ms, logger=self.logger)\n resp_ok = True\n try:\n for attempt, time_to_sleep in enumerate(redo.retrier(max_attempts=\n HTTP_MAX_ATTEMPTS, sleeptime=0.5, max_sleeptime=\n HTTP_MAX_RETRY_SLEEP_SHORT_S, sleepscale=1.5), 1):\n await domain_rate_watcher.waitAccessAsync()\n try:\n async with self.session.head(url, headers=self.\n _buildHeaders(headers), timeout=HTTP_SHORT_TIMEOUT, ssl\n =verify) as response:\n pass\n except (asyncio.TimeoutError, aiohttp.ClientError) as e:\n self.logger.warning(\n \"Probing '%s' failed (attempt %u/%u): %s %s\" % (url,\n attempt, HTTP_MAX_ATTEMPTS, e.__class__.__qualname__, e))\n if attempt == HTTP_MAX_ATTEMPTS:\n resp_ok = False\n else:\n self.logger.debug('Retrying in %.3fs' % time_to_sleep)\n await asyncio.sleep(time_to_sleep)\n else:\n response.raise_for_status()\n if response_headers is not None:\n response_headers.update(response.headers)\n break\n except aiohttp.ClientResponseError as e:\n self.logger.debug(\"Probing '%s' failed: %s %s\" % (url, e.__class__.\n __qualname__, e))\n resp_ok = False\n if cache is not None:\n cache[url] = pickle.dumps((resp_ok, response_headers))\n return resp_ok\n", "role": "user" }, { "content": "isReachable", "role": "assistant" } ]
async def fastStreamedQuery(self, url, *, headers=None, verify=True): """ Send a GET request with short timeout, do not retry, and return streamed response. """ response = await self.session.get(url, headers=self._buildHeaders( headers), timeout=HTTP_SHORT_TIMEOUT, ssl=verify) response.raise_for_status() return response
fastStreamedQuery
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def fastStreamedQuery(self, url, *, headers=None, verify=True):\n \"\"\" Send a GET request with short timeout, do not retry, and return streamed response. \"\"\"\n response = await self.session.get(url, headers=self._buildHeaders(\n headers), timeout=HTTP_SHORT_TIMEOUT, ssl=verify)\n response.raise_for_status()\n return response\n", "role": "user" }, { "content": "fastStreamedQuery", "role": "assistant" } ]
def x(self, album, artist): params = collections.OrderedDict() params['method'] = 'album.getinfo' params['api_key'] = __class__.API_KEY params['album'] = album params['artist'] = artist return __class__.assembleUrl(__class__.BASE_URL, params)
getSearchUrl
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, album, artist):\n params = collections.OrderedDict()\n params['method'] = 'album.getinfo'\n params['api_key'] = __class__.API_KEY\n params['album'] = album\n params['artist'] = artist\n return __class__.assembleUrl(__class__.BASE_URL, params)\n", "role": "user" }, { "content": "getSearchUrl", "role": "assistant" } ]
def x(self, s): char_blacklist = set(string.punctuation) char_blacklist.remove("'") char_blacklist.remove('&') char_blacklist = frozenset(char_blacklist) return __class__.unpunctuate(s.lower(), char_blacklist=char_blacklist)
processQueryString
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, s):\n char_blacklist = set(string.punctuation)\n char_blacklist.remove(\"'\")\n char_blacklist.remove('&')\n char_blacklist = frozenset(char_blacklist)\n return __class__.unpunctuate(s.lower(), char_blacklist=char_blacklist)\n", "role": "user" }, { "content": "processQueryString", "role": "assistant" } ]
async def parseResults(self, api_data): """ See CoverSource.parseResults. """ results = [] xml_text = api_data.decode('utf-8') xml_root = xml.etree.ElementTree.fromstring(xml_text) status = xml_root.get('status') if status != 'ok': raise Exception('Unexpected Last.fm response status: %s' % status) img_elements = xml_root.findall('album/image') thumbnail_url = None thumbnail_size = None for img_element in img_elements: img_url = img_element.text if not img_url: continue lfm_size = img_element.get('size') if lfm_size == 'mega': check_metadata = CoverImageMetadata.SIZE else: check_metadata = CoverImageMetadata.NONE try: size = __class__.SIZES[lfm_size] except KeyError: continue if size[0] <= MAX_THUMBNAIL_SIZE and (thumbnail_size is None or size[0] < thumbnail_size): thumbnail_url = img_url thumbnail_size = size[0] format = os.path.splitext(img_url)[1][1:].lower() format = SUPPORTED_IMG_FORMATS[format] results.append(LastFmCoverSourceResult(img_url, size, format, thumbnail_url=thumbnail_url, source=self, check_metadata= check_metadata)) return results
parseResults
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def parseResults(self, api_data):\n \"\"\" See CoverSource.parseResults. \"\"\"\n results = []\n xml_text = api_data.decode('utf-8')\n xml_root = xml.etree.ElementTree.fromstring(xml_text)\n status = xml_root.get('status')\n if status != 'ok':\n raise Exception('Unexpected Last.fm response status: %s' % status)\n img_elements = xml_root.findall('album/image')\n thumbnail_url = None\n thumbnail_size = None\n for img_element in img_elements:\n img_url = img_element.text\n if not img_url:\n continue\n lfm_size = img_element.get('size')\n if lfm_size == 'mega':\n check_metadata = CoverImageMetadata.SIZE\n else:\n check_metadata = CoverImageMetadata.NONE\n try:\n size = __class__.SIZES[lfm_size]\n except KeyError:\n continue\n if size[0] <= MAX_THUMBNAIL_SIZE and (thumbnail_size is None or \n size[0] < thumbnail_size):\n thumbnail_url = img_url\n thumbnail_size = size[0]\n format = os.path.splitext(img_url)[1][1:].lower()\n format = SUPPORTED_IMG_FORMATS[format]\n results.append(LastFmCoverSourceResult(img_url, size, format,\n thumbnail_url=thumbnail_url, source=self, check_metadata=\n check_metadata))\n return results\n", "role": "user" }, { "content": "parseResults", "role": "assistant" } ]
async def search_and_download(album, artist, format, size, out_filepath, *, size_tolerance_prct, amazon_tlds, no_lq_sources, async_loop): """ Search and download a cover, return True if success, False instead. """ source_args = size, size_tolerance_prct cover_sources = [sources.LastFmCoverSource(*source_args), sources. AmazonCdCoverSource(*source_args), sources.AmazonDigitalCoverSource (*source_args)] for tld in amazon_tlds: cover_sources.append(sources.AmazonCdCoverSource(*source_args, tld=tld) ) if not no_lq_sources: cover_sources.append(sources.GoogleImagesWebScrapeCoverSource(* source_args)) search_futures = [] for cover_source in cover_sources: coroutine = cover_source.search(album, artist) future = asyncio.ensure_future(coroutine, loop=async_loop) search_futures.append(future) await asyncio.wait(search_futures, loop=async_loop) results = [] for future in search_futures: source_results = future.result() results.extend(source_results) results = await CoverSourceResult.preProcessForComparison(results, size, size_tolerance_prct) results.sort(reverse=True, key=functools.cmp_to_key(functools.partial( CoverSourceResult.compare, target_size=size, size_tolerance_prct= size_tolerance_prct))) if not results: logging.getLogger('Main').info('No results') for result in results: try: await result.get(format, size, size_tolerance_prct, out_filepath) except Exception as e: logging.getLogger('Main').warning( 'Download of %s failed: %s %s' % (result, e.__class__. __qualname__, e)) continue else: return True return False
search_and_download
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def search_and_download(album, artist, format, size, out_filepath, *,\n size_tolerance_prct, amazon_tlds, no_lq_sources, async_loop):\n \"\"\" Search and download a cover, return True if success, False instead. \"\"\"\n source_args = size, size_tolerance_prct\n cover_sources = [sources.LastFmCoverSource(*source_args), sources.\n AmazonCdCoverSource(*source_args), sources.AmazonDigitalCoverSource\n (*source_args)]\n for tld in amazon_tlds:\n cover_sources.append(sources.AmazonCdCoverSource(*source_args, tld=tld)\n )\n if not no_lq_sources:\n cover_sources.append(sources.GoogleImagesWebScrapeCoverSource(*\n source_args))\n search_futures = []\n for cover_source in cover_sources:\n coroutine = cover_source.search(album, artist)\n future = asyncio.ensure_future(coroutine, loop=async_loop)\n search_futures.append(future)\n await asyncio.wait(search_futures, loop=async_loop)\n results = []\n for future in search_futures:\n source_results = future.result()\n results.extend(source_results)\n results = await CoverSourceResult.preProcessForComparison(results, size,\n size_tolerance_prct)\n results.sort(reverse=True, key=functools.cmp_to_key(functools.partial(\n CoverSourceResult.compare, target_size=size, size_tolerance_prct=\n size_tolerance_prct)))\n if not results:\n logging.getLogger('Main').info('No results')\n for result in results:\n try:\n await result.get(format, size, size_tolerance_prct, out_filepath)\n except Exception as e:\n logging.getLogger('Main').warning(\n 'Download of %s failed: %s %s' % (result, e.__class__.\n __qualname__, e))\n continue\n else:\n return True\n return False\n", "role": "user" }, { "content": "search_and_download", "role": "assistant" } ]
def x(self, album, artist): url = '%s/search' % __class__.BASE_URL params = collections.OrderedDict() params['search-alias'] = 'digital-music' params['field-keywords'] = ' '.join((artist, album)) params['sort'] = 'relevancerank' return __class__.assembleUrl(url, params)
getSearchUrl
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, album, artist):\n url = '%s/search' % __class__.BASE_URL\n params = collections.OrderedDict()\n params['search-alias'] = 'digital-music'\n params['field-keywords'] = ' '.join((artist, album))\n params['sort'] = 'relevancerank'\n return __class__.assembleUrl(url, params)\n", "role": "user" }, { "content": "getSearchUrl", "role": "assistant" } ]
async def parseResults(self, api_data): """ See CoverSource.parseResults. """ results = [] parser = lxml.etree.HTMLParser() html = lxml.etree.XML(api_data.decode('utf-8'), parser) for page_struct_version, result_selector in enumerate(__class__. RESULTS_SELECTORS): result_nodes = result_selector(html) if result_nodes: break for rank, result_node in enumerate(result_nodes, 1): img_node = __class__.IMG_SELECTORS[page_struct_version](result_node)[0] thumbnail_url = img_node.get('src') thumbnail_url = thumbnail_url.replace('Stripe-Prime-Only', '') url_parts = thumbnail_url.rsplit('.', 2) img_url = '.'.join((url_parts[0], url_parts[2])) size = 500, 500 if self.target_size > size[0]: self.logger.debug('Looking for optimal subimages configuration...') product_url = __class__.LINK_SELECTOR(result_node)[0].get('href') product_url = urllib.parse.urlsplit(product_url) product_id = product_url.path.split('/')[3] for amazon_img_format in AMAZON_DIGITAL_IMAGE_FORMATS: self.logger.debug('Trying %u subimages...' % amazon_img_format.slice_count ** 2) urls = tuple(self.generateImgUrls(product_id, __class__. DYNAPI_KEY, amazon_img_format.id, amazon_img_format. slice_count)) url_ok = await self.probeUrl(urls[-1]) if not url_ok: continue img_url = urls size = (amazon_img_format.total_res,) * 2 break format = CoverImageFormat.JPEG results.append(AmazonDigitalCoverSourceResult(img_url, size, format, thumbnail_url=thumbnail_url, source=self, rank=rank, check_metadata=CoverImageMetadata.SIZE)) return results
parseResults
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def parseResults(self, api_data):\n \"\"\" See CoverSource.parseResults. \"\"\"\n results = []\n parser = lxml.etree.HTMLParser()\n html = lxml.etree.XML(api_data.decode('utf-8'), parser)\n for page_struct_version, result_selector in enumerate(__class__.\n RESULTS_SELECTORS):\n result_nodes = result_selector(html)\n if result_nodes:\n break\n for rank, result_node in enumerate(result_nodes, 1):\n img_node = __class__.IMG_SELECTORS[page_struct_version](result_node)[0]\n thumbnail_url = img_node.get('src')\n thumbnail_url = thumbnail_url.replace('Stripe-Prime-Only', '')\n url_parts = thumbnail_url.rsplit('.', 2)\n img_url = '.'.join((url_parts[0], url_parts[2]))\n size = 500, 500\n if self.target_size > size[0]:\n self.logger.debug('Looking for optimal subimages configuration...')\n product_url = __class__.LINK_SELECTOR(result_node)[0].get('href')\n product_url = urllib.parse.urlsplit(product_url)\n product_id = product_url.path.split('/')[3]\n for amazon_img_format in AMAZON_DIGITAL_IMAGE_FORMATS:\n self.logger.debug('Trying %u subimages...' % \n amazon_img_format.slice_count ** 2)\n urls = tuple(self.generateImgUrls(product_id, __class__.\n DYNAPI_KEY, amazon_img_format.id, amazon_img_format.\n slice_count))\n url_ok = await self.probeUrl(urls[-1])\n if not url_ok:\n continue\n img_url = urls\n size = (amazon_img_format.total_res,) * 2\n break\n format = CoverImageFormat.JPEG\n results.append(AmazonDigitalCoverSourceResult(img_url, size, format,\n thumbnail_url=thumbnail_url, source=self, rank=rank,\n check_metadata=CoverImageMetadata.SIZE))\n return results\n", "role": "user" }, { "content": "parseResults", "role": "assistant" } ]
def x(self, product_id, dynapi_key, format_id, slice_count): for x in range(slice_count): for y in range(slice_count): yield 'http://z2-ec2.images-amazon.com/R/1/a=' + product_id + '+c=' + dynapi_key + '+d=_SCR%28' + str( format_id) + ',' + str(x) + ',' + str(y) + '%29_=.jpg'
generateImgUrls
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, product_id, dynapi_key, format_id, slice_count):\n for x in range(slice_count):\n for y in range(slice_count):\n yield 'http://z2-ec2.images-amazon.com/R/1/a=' + product_id + '+c=' + dynapi_key + '+d=_SCR%28' + str(\n format_id) + ',' + str(x) + ',' + str(y) + '%29_=.jpg'\n", "role": "user" }, { "content": "generateImgUrls", "role": "assistant" } ]
def x(*, max_attempts, sleeptime, max_sleeptime, sleepscale=1.5, jitter=0.2): assert max_attempts > 1 assert sleeptime >= 0 assert 0 <= jitter <= sleeptime assert sleepscale >= 1 cur_sleeptime = min(max_sleeptime, sleeptime) for attempt in range(max_attempts): cur_jitter = random.randint(int(-jitter * 1000), int(jitter * 1000) ) / 1000 yield max(0, cur_sleeptime + cur_jitter) cur_sleeptime = min(max_sleeptime, cur_sleeptime * sleepscale)
retrier
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*, max_attempts, sleeptime, max_sleeptime, sleepscale=1.5, jitter=0.2):\n assert max_attempts > 1\n assert sleeptime >= 0\n assert 0 <= jitter <= sleeptime\n assert sleepscale >= 1\n cur_sleeptime = min(max_sleeptime, sleeptime)\n for attempt in range(max_attempts):\n cur_jitter = random.randint(int(-jitter * 1000), int(jitter * 1000)\n ) / 1000\n yield max(0, cur_sleeptime + cur_jitter)\n cur_sleeptime = min(max_sleeptime, cur_sleeptime * sleepscale)\n", "role": "user" }, { "content": "retrier", "role": "assistant" } ]
async def get(self, target_format, target_size, size_tolerance_prct, out_filepath): """ Download cover and process it. """ if self.source_quality.value <= CoverSourceQuality.LOW.value: logging.getLogger('Cover').warning( 'Cover is from a potentially unreliable source and may be unrelated to the search' ) images_data = [] for i, url in enumerate(self.urls): logging.getLogger('Cover').info( "Downloading cover '%s' (part %u/%u)..." % (url, i + 1, len( self.urls))) headers = {} self.source.updateHttpHeaders(headers) async def pre_cache_callback(img_data): return await __class__.crunch(img_data, self.format) store_in_cache_callback, image_data = await self.source.http.query(url, headers=headers, verify=False, cache=__class__.image_cache, pre_cache_callback=pre_cache_callback) await store_in_cache_callback() images_data.append(image_data) need_format_change = self.format != target_format need_size_change = max(self.size) > target_size and abs(max(self.size) - target_size) > target_size * size_tolerance_prct / 100 need_join = len(images_data) > 1 if need_join or need_format_change or need_size_change: image_data = self.postProcess(images_data, target_format if need_format_change else None, target_size if need_size_change else None) image_data = await __class__.crunch(image_data, target_format) with open(out_filepath, 'wb') as file: file.write(image_data)
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def get(self, target_format, target_size, size_tolerance_prct,\n out_filepath):\n \"\"\" Download cover and process it. \"\"\"\n if self.source_quality.value <= CoverSourceQuality.LOW.value:\n logging.getLogger('Cover').warning(\n 'Cover is from a potentially unreliable source and may be unrelated to the search'\n )\n images_data = []\n for i, url in enumerate(self.urls):\n logging.getLogger('Cover').info(\n \"Downloading cover '%s' (part %u/%u)...\" % (url, i + 1, len(\n self.urls)))\n headers = {}\n self.source.updateHttpHeaders(headers)\n\n async def pre_cache_callback(img_data):\n return await __class__.crunch(img_data, self.format)\n store_in_cache_callback, image_data = await self.source.http.query(url,\n headers=headers, verify=False, cache=__class__.image_cache,\n pre_cache_callback=pre_cache_callback)\n await store_in_cache_callback()\n images_data.append(image_data)\n need_format_change = self.format != target_format\n need_size_change = max(self.size) > target_size and abs(max(self.size) -\n target_size) > target_size * size_tolerance_prct / 100\n need_join = len(images_data) > 1\n if need_join or need_format_change or need_size_change:\n image_data = self.postProcess(images_data, target_format if\n need_format_change else None, target_size if need_size_change else\n None)\n image_data = await __class__.crunch(image_data, target_format)\n with open(out_filepath, 'wb') as file:\n file.write(image_data)\n", "role": "user" }, { "content": "get", "role": "assistant" } ]
def x(self, images_data, new_format, new_size): if len(images_data) == 1: in_bytes = io.BytesIO(images_data[0]) img = PIL.Image.open(in_bytes) if img.mode != 'RGB': img = img.convert('RGB') else: logging.getLogger('Cover').info('Joining %u images...' % len( images_data)) new_img = PIL.Image.new('RGB', self.size) assert is_square(len(images_data)) sq = int(math.sqrt(len(images_data))) images_data_it = iter(images_data) img_sizes = {} for x in range(sq): for y in range(sq): current_image_data = next(images_data_it) img_stream = io.BytesIO(current_image_data) img = PIL.Image.open(img_stream) img_sizes[x, y] = img.size box = [0, 0] if x > 0: for px in range(x): box[0] += img_sizes[px, y][0] if y > 0: for py in range(y): box[1] += img_sizes[x, py][1] box.extend((box[0] + img.size[0], box[1] + img.size[1])) new_img.paste(img, box=tuple(box)) img = new_img out_bytes = io.BytesIO() if new_size is not None: logging.getLogger('Cover').info('Resizing from %ux%u to %ux%u...' % (self.size[0], self.size[1], new_size, new_size)) img = img.resize((new_size, new_size), PIL.Image.LANCZOS) unsharper = PIL.ImageFilter.UnsharpMask(radius=1.5, percent=70, threshold=5) img = img.filter(unsharper) if new_format is not None: logging.getLogger('Cover').info('Converting to %s...' % new_format. name.upper()) target_format = new_format else: target_format = self.format img.save(out_bytes, format=target_format.name, quality=90, optimize=True) return out_bytes.getvalue()
postProcess
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, images_data, new_format, new_size):\n if len(images_data) == 1:\n in_bytes = io.BytesIO(images_data[0])\n img = PIL.Image.open(in_bytes)\n if img.mode != 'RGB':\n img = img.convert('RGB')\n else:\n logging.getLogger('Cover').info('Joining %u images...' % len(\n images_data))\n new_img = PIL.Image.new('RGB', self.size)\n assert is_square(len(images_data))\n sq = int(math.sqrt(len(images_data)))\n images_data_it = iter(images_data)\n img_sizes = {}\n for x in range(sq):\n for y in range(sq):\n current_image_data = next(images_data_it)\n img_stream = io.BytesIO(current_image_data)\n img = PIL.Image.open(img_stream)\n img_sizes[x, y] = img.size\n box = [0, 0]\n if x > 0:\n for px in range(x):\n box[0] += img_sizes[px, y][0]\n if y > 0:\n for py in range(y):\n box[1] += img_sizes[x, py][1]\n box.extend((box[0] + img.size[0], box[1] + img.size[1]))\n new_img.paste(img, box=tuple(box))\n img = new_img\n out_bytes = io.BytesIO()\n if new_size is not None:\n logging.getLogger('Cover').info('Resizing from %ux%u to %ux%u...' %\n (self.size[0], self.size[1], new_size, new_size))\n img = img.resize((new_size, new_size), PIL.Image.LANCZOS)\n unsharper = PIL.ImageFilter.UnsharpMask(radius=1.5, percent=70,\n threshold=5)\n img = img.filter(unsharper)\n if new_format is not None:\n logging.getLogger('Cover').info('Converting to %s...' % new_format.\n name.upper())\n target_format = new_format\n else:\n target_format = self.format\n img.save(out_bytes, format=target_format.name, quality=90, optimize=True)\n return out_bytes.getvalue()\n", "role": "user" }, { "content": "postProcess", "role": "assistant" } ]
async def updateImageMetadata(self): """ Partially download image file(s) to get its real metadata, or get it from cache. """ assert self.needMetadataUpdate() width_sum, height_sum = 0, 0 idxs = [] assert is_square(len(self.urls)) sq = int(math.sqrt(len(self.urls))) for x in range(sq): for y in range(sq): if x == y: idxs.append((x * sq + y, x, y)) for idx, x, y in idxs: url = self.urls[idx] format, width, height = None, None, None try: format, width, height = pickle.loads(__class__.metadata_cache[url]) except KeyError: pass except Exception as e: logging.getLogger('Cover').warning( "Unable to load metadata for URL '%s' from cache: %s %s" % (url, e.__class__.__qualname__, e)) else: logging.getLogger('Cover').debug( "Got metadata for URL '%s' from cache" % url) if format is not None: self.setFormatMetadata(format) if self.needMetadataUpdate(CoverImageMetadata.FORMAT ) or self.needMetadataUpdate(CoverImageMetadata.SIZE) and ( width is None or height is None): logging.getLogger('Cover').debug( "Downloading file header for URL '%s'..." % url) try: headers = {} self.source.updateHttpHeaders(headers) response = await self.source.http.fastStreamedQuery(url, headers=headers, verify=False) try: if self.needMetadataUpdate(CoverImageMetadata.FORMAT): format = __class__.guessImageFormatFromHttpResponse( response) if format is not None: self.setFormatMetadata(format) if self.needMetadataUpdate(): metadata = (await __class__. guessImageMetadataFromHttpData(response)) if metadata is not None: format, width, height = metadata if format is not None: self.setFormatMetadata(format) finally: response.release() except Exception as e: logging.getLogger('Cover').warning( "Failed to get file metadata for URL '%s' (%s %s)" % ( url, e.__class__.__qualname__, e)) if self.needMetadataUpdate(): if self.format is None or self.size is None and (width is None or height is None): logging.getLogger('Cover').debug( "Unable to get file metadata from file or HTTP headers for URL '%s', skipping this result" % url) return if self.format is not None and (self.size is not None and width is None and height is None): logging.getLogger('Cover').debug( "Unable to get file metadata from file or HTTP headers for URL '%s', falling back to API data" % url) self.check_metadata = CoverImageMetadata.NONE self.reliable_metadata = False return __class__.metadata_cache[url] = pickle.dumps((format, width, height)) if width is not None and height is not None: width_sum += width height_sum += height if self.needMetadataUpdate(CoverImageMetadata.SIZE ) and width_sum > 0 and height_sum > 0: self.setSizeMetadata((width_sum, height_sum))
updateImageMetadata
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def updateImageMetadata(self):\n \"\"\" Partially download image file(s) to get its real metadata, or get it from cache. \"\"\"\n assert self.needMetadataUpdate()\n width_sum, height_sum = 0, 0\n idxs = []\n assert is_square(len(self.urls))\n sq = int(math.sqrt(len(self.urls)))\n for x in range(sq):\n for y in range(sq):\n if x == y:\n idxs.append((x * sq + y, x, y))\n for idx, x, y in idxs:\n url = self.urls[idx]\n format, width, height = None, None, None\n try:\n format, width, height = pickle.loads(__class__.metadata_cache[url])\n except KeyError:\n pass\n except Exception as e:\n logging.getLogger('Cover').warning(\n \"Unable to load metadata for URL '%s' from cache: %s %s\" %\n (url, e.__class__.__qualname__, e))\n else:\n logging.getLogger('Cover').debug(\n \"Got metadata for URL '%s' from cache\" % url)\n if format is not None:\n self.setFormatMetadata(format)\n if self.needMetadataUpdate(CoverImageMetadata.FORMAT\n ) or self.needMetadataUpdate(CoverImageMetadata.SIZE) and (\n width is None or height is None):\n logging.getLogger('Cover').debug(\n \"Downloading file header for URL '%s'...\" % url)\n try:\n headers = {}\n self.source.updateHttpHeaders(headers)\n response = await self.source.http.fastStreamedQuery(url,\n headers=headers, verify=False)\n try:\n if self.needMetadataUpdate(CoverImageMetadata.FORMAT):\n format = __class__.guessImageFormatFromHttpResponse(\n response)\n if format is not None:\n self.setFormatMetadata(format)\n if self.needMetadataUpdate():\n metadata = (await __class__.\n guessImageMetadataFromHttpData(response))\n if metadata is not None:\n format, width, height = metadata\n if format is not None:\n self.setFormatMetadata(format)\n finally:\n response.release()\n except Exception as e:\n logging.getLogger('Cover').warning(\n \"Failed to get file metadata for URL '%s' (%s %s)\" % (\n url, e.__class__.__qualname__, e))\n if self.needMetadataUpdate():\n if self.format is None or self.size is None and (width is\n None or height is None):\n logging.getLogger('Cover').debug(\n \"Unable to get file metadata from file or HTTP headers for URL '%s', skipping this result\"\n % url)\n return\n if self.format is not None and (self.size is not None and \n width is None and height is None):\n logging.getLogger('Cover').debug(\n \"Unable to get file metadata from file or HTTP headers for URL '%s', falling back to API data\"\n % url)\n self.check_metadata = CoverImageMetadata.NONE\n self.reliable_metadata = False\n return\n __class__.metadata_cache[url] = pickle.dumps((format, width,\n height))\n if width is not None and height is not None:\n width_sum += width\n height_sum += height\n if self.needMetadataUpdate(CoverImageMetadata.SIZE\n ) and width_sum > 0 and height_sum > 0:\n self.setSizeMetadata((width_sum, height_sum))\n", "role": "user" }, { "content": "updateImageMetadata", "role": "assistant" } ]
def x(self, format): assert self.needMetadataUpdate(CoverImageMetadata.FORMAT ) or self.format is format self.format = format self.check_metadata &= ~CoverImageMetadata.FORMAT
setFormatMetadata
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, format):\n assert self.needMetadataUpdate(CoverImageMetadata.FORMAT\n ) or self.format is format\n self.format = format\n self.check_metadata &= ~CoverImageMetadata.FORMAT\n", "role": "user" }, { "content": "setFormatMetadata", "role": "assistant" } ]
def x(self, size): assert self.needMetadataUpdate(CoverImageMetadata.SIZE ) or self.size == size self.size = size self.check_metadata &= ~CoverImageMetadata.SIZE
setSizeMetadata
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, size):\n assert self.needMetadataUpdate(CoverImageMetadata.SIZE\n ) or self.size == size\n self.size = size\n self.check_metadata &= ~CoverImageMetadata.SIZE\n", "role": "user" }, { "content": "setSizeMetadata", "role": "assistant" } ]
async def updateSignature(self): """ Calculate a cover's "signature" using its thumbnail url. """ assert self.thumbnail_sig is None if self.thumbnail_url is None: logging.getLogger('Cover').warning('No thumbnail available for %s' % self) return logging.getLogger('Cover').debug("Downloading cover thumbnail '%s'..." % self.thumbnail_url) headers = {} self.source.updateHttpHeaders(headers) async def pre_cache_callback(img_data): return await __class__.crunch(img_data, CoverImageFormat.JPEG, silent=True) try: store_in_cache_callback, image_data = await self.source.http.query(self .thumbnail_url, cache=__class__.image_cache, headers=headers, pre_cache_callback=pre_cache_callback) except Exception as e: logging.getLogger('Cover').warning("Download of '%s' failed: %s %s" % (self.thumbnail_url, e.__class__.__qualname__, e)) return logging.getLogger('Cover').debug('Computing signature of %s...' % self) try: self.thumbnail_sig = __class__.computeImgSignature(image_data) except Exception as e: logging.getLogger('Cover').warning( "Failed to compute signature of '%s': %s %s" % (self, e. __class__.__qualname__, e)) else: await store_in_cache_callback()
updateSignature
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def updateSignature(self):\n \"\"\" Calculate a cover's \"signature\" using its thumbnail url. \"\"\"\n assert self.thumbnail_sig is None\n if self.thumbnail_url is None:\n logging.getLogger('Cover').warning('No thumbnail available for %s' %\n self)\n return\n logging.getLogger('Cover').debug(\"Downloading cover thumbnail '%s'...\" %\n self.thumbnail_url)\n headers = {}\n self.source.updateHttpHeaders(headers)\n\n async def pre_cache_callback(img_data):\n return await __class__.crunch(img_data, CoverImageFormat.JPEG,\n silent=True)\n try:\n store_in_cache_callback, image_data = await self.source.http.query(self\n .thumbnail_url, cache=__class__.image_cache, headers=headers,\n pre_cache_callback=pre_cache_callback)\n except Exception as e:\n logging.getLogger('Cover').warning(\"Download of '%s' failed: %s %s\" %\n (self.thumbnail_url, e.__class__.__qualname__, e))\n return\n logging.getLogger('Cover').debug('Computing signature of %s...' % self)\n try:\n self.thumbnail_sig = __class__.computeImgSignature(image_data)\n except Exception as e:\n logging.getLogger('Cover').warning(\n \"Failed to compute signature of '%s': %s %s\" % (self, e.\n __class__.__qualname__, e))\n else:\n await store_in_cache_callback()\n", "role": "user" }, { "content": "updateSignature", "role": "assistant" } ]
def x(first, second, *, target_size, size_tolerance_prct): for c in (first, second): assert c.format is not None assert isinstance(c.size[0], int) and isinstance(c.size[1], int) delta_ratio1 = abs(first.size[0] / first.size[1] - 1) delta_ratio2 = abs(second.size[0] / second.size[1] - 1) if abs(delta_ratio1 - delta_ratio2) > 0.15: return -1 if delta_ratio1 > delta_ratio2 else 1 sr1 = first.is_similar_to_reference sr2 = second.is_similar_to_reference if sr1 and not sr2: return 1 if not sr1 and sr2: return -1 delta_size1 = (first.size[0] + first.size[1]) / 2 - target_size delta_size2 = (second.size[0] + second.size[1]) / 2 - target_size if (delta_size1 < 0 and delta_size2 >= 0 or delta_size1 >= 0 and delta_size2 < 0): return -1 if delta_size1 < delta_size2 else 1 if delta_size1 < 0 and delta_size2 < 0 and delta_size1 != delta_size2: return -1 if delta_size1 < delta_size2 else 1 qs1 = first.source_quality.value qs2 = second.source_quality.value if qs1 != qs2: return -1 if qs1 < qs2 else 1 if (first.rank is not None and second.rank is not None and first. __class__ is second.__class__ and first.rank != second.rank): return -1 if first.rank > second.rank else 1 if first.reliable_metadata != second.reliable_metadata: return 1 if first.reliable_metadata else -1 ic1 = len(first.urls) ic2 = len(second.urls) if ic1 != ic2: return -1 if ic1 > ic2 else 1 if abs(delta_size1) != abs(delta_size2): return -1 if abs(delta_size1) > abs(delta_size2) else 1 if first.format != second.format: return -1 if second.format is CoverImageFormat.PNG else 1 if delta_ratio1 != delta_ratio2: return -1 if delta_ratio1 > delta_ratio2 else 1 return 0
compare
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(first, second, *, target_size, size_tolerance_prct):\n for c in (first, second):\n assert c.format is not None\n assert isinstance(c.size[0], int) and isinstance(c.size[1], int)\n delta_ratio1 = abs(first.size[0] / first.size[1] - 1)\n delta_ratio2 = abs(second.size[0] / second.size[1] - 1)\n if abs(delta_ratio1 - delta_ratio2) > 0.15:\n return -1 if delta_ratio1 > delta_ratio2 else 1\n sr1 = first.is_similar_to_reference\n sr2 = second.is_similar_to_reference\n if sr1 and not sr2:\n return 1\n if not sr1 and sr2:\n return -1\n delta_size1 = (first.size[0] + first.size[1]) / 2 - target_size\n delta_size2 = (second.size[0] + second.size[1]) / 2 - target_size\n if (delta_size1 < 0 and delta_size2 >= 0 or delta_size1 >= 0 and \n delta_size2 < 0):\n return -1 if delta_size1 < delta_size2 else 1\n if delta_size1 < 0 and delta_size2 < 0 and delta_size1 != delta_size2:\n return -1 if delta_size1 < delta_size2 else 1\n qs1 = first.source_quality.value\n qs2 = second.source_quality.value\n if qs1 != qs2:\n return -1 if qs1 < qs2 else 1\n if (first.rank is not None and second.rank is not None and first.\n __class__ is second.__class__ and first.rank != second.rank):\n return -1 if first.rank > second.rank else 1\n if first.reliable_metadata != second.reliable_metadata:\n return 1 if first.reliable_metadata else -1\n ic1 = len(first.urls)\n ic2 = len(second.urls)\n if ic1 != ic2:\n return -1 if ic1 > ic2 else 1\n if abs(delta_size1) != abs(delta_size2):\n return -1 if abs(delta_size1) > abs(delta_size2) else 1\n if first.format != second.format:\n return -1 if second.format is CoverImageFormat.PNG else 1\n if delta_ratio1 != delta_ratio2:\n return -1 if delta_ratio1 > delta_ratio2 else 1\n return 0\n", "role": "user" }, { "content": "compare", "role": "assistant" } ]
async def crunch(image_data, format, silent=False): """ Crunch image data, and return the processed data, or orignal data if operation failed. """ if (format is CoverImageFormat.PNG and not HAS_OPTIPNG or format is CoverImageFormat.JPEG and not HAS_JPEGOPTIM): return image_data with mkstemp_ctx.mkstemp(suffix='.%s' % format.name.lower() ) as tmp_out_filepath: if not silent: logging.getLogger('Cover').info('Crunching %s image...' % format.name.upper()) with open(tmp_out_filepath, 'wb') as tmp_out_file: tmp_out_file.write(image_data) size_before = len(image_data) if format is CoverImageFormat.PNG: cmd = ['optipng', '-quiet', '-o1'] elif format is CoverImageFormat.JPEG: cmd = ['jpegoptim', '-q', '--strip-all'] cmd.append(tmp_out_filepath) p = await asyncio.create_subprocess_exec(*cmd, stdin=asyncio. subprocess.DEVNULL, stdout=asyncio.subprocess.DEVNULL, stderr= asyncio.subprocess.DEVNULL) await p.wait() if p.returncode != 0: if not silent: logging.getLogger('Cover').warning('Crunching image failed') return image_data with open(tmp_out_filepath, 'rb') as tmp_out_file: crunched_image_data = tmp_out_file.read() size_after = len(crunched_image_data) pct_saved = 100 * (size_before - size_after) / size_before if not silent: logging.getLogger('Cover').debug( 'Crunching image saved %.2f%% filesize' % pct_saved) return crunched_image_data
crunch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def crunch(image_data, format, silent=False):\n \"\"\" Crunch image data, and return the processed data, or orignal data if operation failed. \"\"\"\n if (format is CoverImageFormat.PNG and not HAS_OPTIPNG or format is\n CoverImageFormat.JPEG and not HAS_JPEGOPTIM):\n return image_data\n with mkstemp_ctx.mkstemp(suffix='.%s' % format.name.lower()\n ) as tmp_out_filepath:\n if not silent:\n logging.getLogger('Cover').info('Crunching %s image...' %\n format.name.upper())\n with open(tmp_out_filepath, 'wb') as tmp_out_file:\n tmp_out_file.write(image_data)\n size_before = len(image_data)\n if format is CoverImageFormat.PNG:\n cmd = ['optipng', '-quiet', '-o1']\n elif format is CoverImageFormat.JPEG:\n cmd = ['jpegoptim', '-q', '--strip-all']\n cmd.append(tmp_out_filepath)\n p = await asyncio.create_subprocess_exec(*cmd, stdin=asyncio.\n subprocess.DEVNULL, stdout=asyncio.subprocess.DEVNULL, stderr=\n asyncio.subprocess.DEVNULL)\n await p.wait()\n if p.returncode != 0:\n if not silent:\n logging.getLogger('Cover').warning('Crunching image failed')\n return image_data\n with open(tmp_out_filepath, 'rb') as tmp_out_file:\n crunched_image_data = tmp_out_file.read()\n size_after = len(crunched_image_data)\n pct_saved = 100 * (size_before - size_after) / size_before\n if not silent:\n logging.getLogger('Cover').debug(\n 'Crunching image saved %.2f%% filesize' % pct_saved)\n return crunched_image_data\n", "role": "user" }, { "content": "crunch", "role": "assistant" } ]
def x(img_data): format, width, height = None, None, None img_stream = io.BytesIO(img_data) try: img = PIL.Image.open(img_stream) except IOError: format = imghdr.what(None, h=img_data) format = SUPPORTED_IMG_FORMATS.get(format, None) else: format = img.format.lower() format = SUPPORTED_IMG_FORMATS.get(format, None) width, height = img.size return format, width, height
guessImageMetadataFromData
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(img_data):\n format, width, height = None, None, None\n img_stream = io.BytesIO(img_data)\n try:\n img = PIL.Image.open(img_stream)\n except IOError:\n format = imghdr.what(None, h=img_data)\n format = SUPPORTED_IMG_FORMATS.get(format, None)\n else:\n format = img.format.lower()\n format = SUPPORTED_IMG_FORMATS.get(format, None)\n width, height = img.size\n return format, width, height\n", "role": "user" }, { "content": "guessImageMetadataFromData", "role": "assistant" } ]
async def guessImageMetadataFromHttpData(response): """ Identify an image format and size from the beginning of its HTTP data. """ metadata = None img_data = bytearray() while len(img_data) < CoverSourceResult.MAX_FILE_METADATA_PEEK_SIZE: new_img_data = await response.content.read(__class__. METADATA_PEEK_SIZE_INCREMENT) if not new_img_data: break img_data.extend(new_img_data) metadata = __class__.guessImageMetadataFromData(img_data) if metadata is not None and all(metadata): return metadata return metadata
guessImageMetadataFromHttpData
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def guessImageMetadataFromHttpData(response):\n \"\"\" Identify an image format and size from the beginning of its HTTP data. \"\"\"\n metadata = None\n img_data = bytearray()\n while len(img_data) < CoverSourceResult.MAX_FILE_METADATA_PEEK_SIZE:\n new_img_data = await response.content.read(__class__.\n METADATA_PEEK_SIZE_INCREMENT)\n if not new_img_data:\n break\n img_data.extend(new_img_data)\n metadata = __class__.guessImageMetadataFromData(img_data)\n if metadata is not None and all(metadata):\n return metadata\n return metadata\n", "role": "user" }, { "content": "guessImageMetadataFromHttpData", "role": "assistant" } ]
def x(response): extensions = [] try: content_type = response.headers['Content-Type'] except KeyError: pass else: ext = mimetypes.guess_extension(content_type, strict=False) if ext is not None: extensions.append(ext) urls = list(response.history) + [response.url] for url in map(str, urls): ext = os.path.splitext(urllib.parse.urlsplit(url).path)[-1] if ext is not None and ext not in extensions: extensions.append(ext) for ext in extensions: try: return SUPPORTED_IMG_FORMATS[ext[1:]] except KeyError: pass
guessImageFormatFromHttpResponse
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(response):\n extensions = []\n try:\n content_type = response.headers['Content-Type']\n except KeyError:\n pass\n else:\n ext = mimetypes.guess_extension(content_type, strict=False)\n if ext is not None:\n extensions.append(ext)\n urls = list(response.history) + [response.url]\n for url in map(str, urls):\n ext = os.path.splitext(urllib.parse.urlsplit(url).path)[-1]\n if ext is not None and ext not in extensions:\n extensions.append(ext)\n for ext in extensions:\n try:\n return SUPPORTED_IMG_FORMATS[ext[1:]]\n except KeyError:\n pass\n", "role": "user" }, { "content": "guessImageFormatFromHttpResponse", "role": "assistant" } ]
async def preProcessForComparison(results, target_size, size_tolerance_prct): """ Process results to prepare them for future comparison and sorting. """ reference = None for result in results: if result.source_quality is CoverSourceQuality.REFERENCE: if reference is None or CoverSourceResult.compare(result, reference, target_size=target_size, size_tolerance_prct= size_tolerance_prct) > 0: reference = result results = list(itertools.filterfalse(operator.attrgetter( 'is_only_reference'), results)) no_dup_results = [] for result in results: is_dup = False for result_comp in results: if (result_comp is not result and result_comp.urls == result. urls and __class__.compare(result, result_comp, target_size =target_size, size_tolerance_prct=size_tolerance_prct) < 0): is_dup = True break if not is_dup: no_dup_results.append(result) dup_count = len(results) - len(no_dup_results) if dup_count > 0: logging.getLogger('Cover').info('Removed %u duplicate results' % dup_count) results = no_dup_results if reference is not None: logging.getLogger('Cover').info('Reference is: %s' % reference) reference.is_similar_to_reference = True futures = [] for result in results: coroutine = result.updateSignature() future = asyncio.ensure_future(coroutine) futures.append(future) if reference.is_only_reference: assert reference not in results coroutine = reference.updateSignature() future = asyncio.ensure_future(coroutine) futures.append(future) if futures: await asyncio.wait(futures) for future in futures: future.result() for result in results: if (result is not reference and result.thumbnail_sig is not None and reference.thumbnail_sig is not None): result.is_similar_to_reference = __class__.areImageSigsSimilar( result.thumbnail_sig, reference.thumbnail_sig) if result.is_similar_to_reference: logging.getLogger('Cover').debug( '%s is similar to reference' % result) else: logging.getLogger('Cover').debug( '%s is NOT similar to reference' % result) else: logging.getLogger('Cover').warning('No reference result found') return results
preProcessForComparison
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def preProcessForComparison(results, target_size, size_tolerance_prct):\n \"\"\" Process results to prepare them for future comparison and sorting. \"\"\"\n reference = None\n for result in results:\n if result.source_quality is CoverSourceQuality.REFERENCE:\n if reference is None or CoverSourceResult.compare(result,\n reference, target_size=target_size, size_tolerance_prct=\n size_tolerance_prct) > 0:\n reference = result\n results = list(itertools.filterfalse(operator.attrgetter(\n 'is_only_reference'), results))\n no_dup_results = []\n for result in results:\n is_dup = False\n for result_comp in results:\n if (result_comp is not result and result_comp.urls == result.\n urls and __class__.compare(result, result_comp, target_size\n =target_size, size_tolerance_prct=size_tolerance_prct) < 0):\n is_dup = True\n break\n if not is_dup:\n no_dup_results.append(result)\n dup_count = len(results) - len(no_dup_results)\n if dup_count > 0:\n logging.getLogger('Cover').info('Removed %u duplicate results' %\n dup_count)\n results = no_dup_results\n if reference is not None:\n logging.getLogger('Cover').info('Reference is: %s' % reference)\n reference.is_similar_to_reference = True\n futures = []\n for result in results:\n coroutine = result.updateSignature()\n future = asyncio.ensure_future(coroutine)\n futures.append(future)\n if reference.is_only_reference:\n assert reference not in results\n coroutine = reference.updateSignature()\n future = asyncio.ensure_future(coroutine)\n futures.append(future)\n if futures:\n await asyncio.wait(futures)\n for future in futures:\n future.result()\n for result in results:\n if (result is not reference and result.thumbnail_sig is not\n None and reference.thumbnail_sig is not None):\n result.is_similar_to_reference = __class__.areImageSigsSimilar(\n result.thumbnail_sig, reference.thumbnail_sig)\n if result.is_similar_to_reference:\n logging.getLogger('Cover').debug(\n '%s is similar to reference' % result)\n else:\n logging.getLogger('Cover').debug(\n '%s is NOT similar to reference' % result)\n else:\n logging.getLogger('Cover').warning('No reference result found')\n return results\n", "role": "user" }, { "content": "preProcessForComparison", "role": "assistant" } ]
def x(image_data): parser = PIL.ImageFile.Parser() parser.feed(image_data) img = parser.close() target_size = __class__.IMG_SIG_SIZE, __class__.IMG_SIG_SIZE img.thumbnail(target_size, PIL.Image.BICUBIC) if img.size != target_size: logging.getLogger('Cover').debug( 'Non square thumbnail after resize to %ux%u, unable to compute signature' % target_size) return None img = img.convert(mode='RGB') pixels = img.getdata() pixel_count = target_size[0] * target_size[1] color_count = 3 r = bitarray.bitarray(pixel_count * color_count) r.setall(False) for ic in range(color_count): mean = sum(p[ic] for p in pixels) // pixel_count for ip, p in enumerate(pixels): if p[ic] > mean: r[pixel_count * ic + ip] = True return r
computeImgSignature
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(image_data):\n parser = PIL.ImageFile.Parser()\n parser.feed(image_data)\n img = parser.close()\n target_size = __class__.IMG_SIG_SIZE, __class__.IMG_SIG_SIZE\n img.thumbnail(target_size, PIL.Image.BICUBIC)\n if img.size != target_size:\n logging.getLogger('Cover').debug(\n 'Non square thumbnail after resize to %ux%u, unable to compute signature'\n % target_size)\n return None\n img = img.convert(mode='RGB')\n pixels = img.getdata()\n pixel_count = target_size[0] * target_size[1]\n color_count = 3\n r = bitarray.bitarray(pixel_count * color_count)\n r.setall(False)\n for ic in range(color_count):\n mean = sum(p[ic] for p in pixels) // pixel_count\n for ip, p in enumerate(pixels):\n if p[ic] > mean:\n r[pixel_count * ic + ip] = True\n return r\n", "role": "user" }, { "content": "computeImgSignature", "role": "assistant" } ]
def x(lib_dir, cover_filename, *, ignore_existing=False): work = {} stats = collections.OrderedDict((k, 0) for k in ('files', 'albums', 'missing covers', 'errors')) with tqdm.tqdm(desc='Analyzing library', unit='dir', postfix=stats ) as progress, tqdm_logging.redirect_logging(progress): for rootpath, rel_dirpaths, rel_filepaths in os.walk(lib_dir): metadata = analyze_dir(stats, rootpath, rel_filepaths, cover_filename, ignore_existing=ignore_existing) progress.set_postfix(stats, refresh=False) progress.update(1) if all(metadata[:-1]): work[rootpath] = metadata[:-1] return work
analyze_lib
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(lib_dir, cover_filename, *, ignore_existing=False):\n work = {}\n stats = collections.OrderedDict((k, 0) for k in ('files', 'albums',\n 'missing covers', 'errors'))\n with tqdm.tqdm(desc='Analyzing library', unit='dir', postfix=stats\n ) as progress, tqdm_logging.redirect_logging(progress):\n for rootpath, rel_dirpaths, rel_filepaths in os.walk(lib_dir):\n metadata = analyze_dir(stats, rootpath, rel_filepaths,\n cover_filename, ignore_existing=ignore_existing)\n progress.set_postfix(stats, refresh=False)\n progress.update(1)\n if all(metadata[:-1]):\n work[rootpath] = metadata[:-1]\n return work\n", "role": "user" }, { "content": "analyze_lib", "role": "assistant" } ]
def x(audio_filepaths): artist, album, has_embedded_album_art = None, None, None for audio_filepath in audio_filepaths: try: mf = mutagen.File(audio_filepath) except Exception: continue if mf is None: continue for key in ('albumartist', 'artist', 'TPE1', 'TPE2', 'aART', '©ART'): try: val = mf.get(key, None) except ValueError: val = None if val is not None: artist = val[-1] break for key in ('_album', 'album', 'TALB', '©alb'): try: val = mf.get(key, None) except ValueError: val = None if val is not None: album = val[-1] break if artist and album: if isinstance(mf, mutagen.ogg.OggFileType): has_embedded_album_art = 'metadata_block_picture' in mf elif isinstance(mf, mutagen.mp3.MP3): has_embedded_album_art = any(map(operator.methodcaller( 'startswith', 'APIC:'), mf.keys())) elif isinstance(mf, mutagen.mp4.MP4): has_embedded_album_art = 'covr' in mf break return artist, album, has_embedded_album_art
get_metadata
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(audio_filepaths):\n artist, album, has_embedded_album_art = None, None, None\n for audio_filepath in audio_filepaths:\n try:\n mf = mutagen.File(audio_filepath)\n except Exception:\n continue\n if mf is None:\n continue\n for key in ('albumartist', 'artist', 'TPE1', 'TPE2', 'aART', '©ART'):\n try:\n val = mf.get(key, None)\n except ValueError:\n val = None\n if val is not None:\n artist = val[-1]\n break\n for key in ('_album', 'album', 'TALB', '©alb'):\n try:\n val = mf.get(key, None)\n except ValueError:\n val = None\n if val is not None:\n album = val[-1]\n break\n if artist and album:\n if isinstance(mf, mutagen.ogg.OggFileType):\n has_embedded_album_art = 'metadata_block_picture' in mf\n elif isinstance(mf, mutagen.mp3.MP3):\n has_embedded_album_art = any(map(operator.methodcaller(\n 'startswith', 'APIC:'), mf.keys()))\n elif isinstance(mf, mutagen.mp4.MP4):\n has_embedded_album_art = 'covr' in mf\n break\n return artist, album, has_embedded_album_art\n", "role": "user" }, { "content": "get_metadata", "role": "assistant" } ]
def x(stats, parent_dir, rel_filepaths, cover_filename, *, ignore_existing= False): no_metadata = None, None, None metadata = no_metadata audio_filepaths = [] for rel_filepath in rel_filepaths: stats['files'] += 1 try: ext = os.path.splitext(rel_filepath)[1][1:].lower() except IndexError: continue if ext in AUDIO_EXTENSIONS: audio_filepaths.append(os.path.join(parent_dir, rel_filepath)) if audio_filepaths: stats['albums'] += 1 if cover_filename != EMBEDDED_ALBUM_ART_SYMBOL: missing = not os.path.isfile(os.path.join(parent_dir, cover_filename)) or ignore_existing if missing: metadata = get_metadata(audio_filepaths) else: metadata = get_metadata(audio_filepaths) missing = not metadata[2] or ignore_existing if missing: stats['missing covers'] += 1 if not all(metadata[:-1]): stats['errors'] += 1 logging.getLogger('sacad_r').error( "Unable to read metadata for album directory '%s'" % parent_dir) else: metadata = no_metadata return metadata
analyze_dir
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(stats, parent_dir, rel_filepaths, cover_filename, *, ignore_existing=\n False):\n no_metadata = None, None, None\n metadata = no_metadata\n audio_filepaths = []\n for rel_filepath in rel_filepaths:\n stats['files'] += 1\n try:\n ext = os.path.splitext(rel_filepath)[1][1:].lower()\n except IndexError:\n continue\n if ext in AUDIO_EXTENSIONS:\n audio_filepaths.append(os.path.join(parent_dir, rel_filepath))\n if audio_filepaths:\n stats['albums'] += 1\n if cover_filename != EMBEDDED_ALBUM_ART_SYMBOL:\n missing = not os.path.isfile(os.path.join(parent_dir,\n cover_filename)) or ignore_existing\n if missing:\n metadata = get_metadata(audio_filepaths)\n else:\n metadata = get_metadata(audio_filepaths)\n missing = not metadata[2] or ignore_existing\n if missing:\n stats['missing covers'] += 1\n if not all(metadata[:-1]):\n stats['errors'] += 1\n logging.getLogger('sacad_r').error(\n \"Unable to read metadata for album directory '%s'\" %\n parent_dir)\n else:\n metadata = no_metadata\n return metadata\n", "role": "user" }, { "content": "analyze_dir", "role": "assistant" } ]
def x(cover_filepath, path): with open(cover_filepath, 'rb') as f: cover_data = f.read() for filename in os.listdir(path): try: ext = os.path.splitext(filename)[1][1:].lower() except IndexError: continue if ext in AUDIO_EXTENSIONS: filepath = os.path.join(path, filename) mf = mutagen.File(filepath) if isinstance(mf.tags, mutagen._vorbis.VComment) or isinstance(mf, mutagen.ogg.OggFileType): picture = mutagen.flac.Picture() picture.data = cover_data picture.type = mutagen.id3.PictureType.COVER_FRONT picture.mime = 'image/jpeg' encoded_data = base64.b64encode(picture.write()) mf['metadata_block_picture'] = encoded_data.decode('ascii') elif isinstance(mf.tags, mutagen.id3.ID3) or isinstance(mf, mutagen.id3.ID3FileType): mf.tags.add(mutagen.id3.APIC(mime='image/jpeg', type= mutagen.id3.PictureType.COVER_FRONT, data=cover_data)) elif isinstance(mf.tags, mutagen.mp4.MP4Tags) or isinstance(mf, mutagen.mp4.MP4): mf['covr'] = [mutagen.mp4.MP4Cover(cover_data, imageformat= mutagen.mp4.AtomDataType.JPEG)] mf.save()
embed_album_art
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cover_filepath, path):\n with open(cover_filepath, 'rb') as f:\n cover_data = f.read()\n for filename in os.listdir(path):\n try:\n ext = os.path.splitext(filename)[1][1:].lower()\n except IndexError:\n continue\n if ext in AUDIO_EXTENSIONS:\n filepath = os.path.join(path, filename)\n mf = mutagen.File(filepath)\n if isinstance(mf.tags, mutagen._vorbis.VComment) or isinstance(mf,\n mutagen.ogg.OggFileType):\n picture = mutagen.flac.Picture()\n picture.data = cover_data\n picture.type = mutagen.id3.PictureType.COVER_FRONT\n picture.mime = 'image/jpeg'\n encoded_data = base64.b64encode(picture.write())\n mf['metadata_block_picture'] = encoded_data.decode('ascii')\n elif isinstance(mf.tags, mutagen.id3.ID3) or isinstance(mf,\n mutagen.id3.ID3FileType):\n mf.tags.add(mutagen.id3.APIC(mime='image/jpeg', type=\n mutagen.id3.PictureType.COVER_FRONT, data=cover_data))\n elif isinstance(mf.tags, mutagen.mp4.MP4Tags) or isinstance(mf,\n mutagen.mp4.MP4):\n mf['covr'] = [mutagen.mp4.MP4Cover(cover_data, imageformat=\n mutagen.mp4.AtomDataType.JPEG)]\n mf.save()\n", "role": "user" }, { "content": "embed_album_art", "role": "assistant" } ]
def x(iterable, n): it = iter(iterable) while True: chunk = tuple(itertools.islice(it, n)) if not chunk: return yield chunk
ichunk
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(iterable, n):\n it = iter(iterable)\n while True:\n chunk = tuple(itertools.islice(it, n))\n if not chunk:\n return\n yield chunk\n", "role": "user" }, { "content": "ichunk", "role": "assistant" } ]
def x(work, args): with contextlib.ExitStack() as cm: if args.filename == EMBEDDED_ALBUM_ART_SYMBOL: tmp_prefix = '%s_' % os.path.splitext(os.path.basename(inspect. getfile(inspect.currentframe())))[0] tmp_dir = cm.enter_context(tempfile.TemporaryDirectory(prefix= tmp_prefix)) stats = collections.OrderedDict((k, 0) for k in ('ok', 'errors', 'no result found')) progress = cm.enter_context(tqdm.tqdm(total=len(work), miniters=1, desc='Searching covers', unit='cover', postfix=stats)) cm.enter_context(tqdm_logging.redirect_logging(progress)) def x(future): path, cover_filepath, artist, album = futures[future] try: status = future.result() except Exception as exception: stats['errors'] += 1 logging.getLogger('sacad_r').error( "Error occured while searching cover for '%s' by '%s' from '%s': %s %s" % (album, artist, path, exception.__class__. __qualname__, exception)) else: if status: if args.filename == EMBEDDED_ALBUM_ART_SYMBOL: try: embed_album_art(cover_filepath, path) except Exception as exception: stats['errors'] += 1 logging.getLogger('sacad_r').error( "Error occured while embedding cover for '%s' by '%s' from '%s': %s %s" % (album, artist, path, exception. __class__.__qualname__, exception)) else: stats['ok'] += 1 finally: os.remove(cover_filepath) else: stats['ok'] += 1 else: stats['no result found'] += 1 logging.getLogger('sacad_r').warning( "Unable to find cover for '%s' by '%s' from '%s'" % (album, artist, path)) progress.set_postfix(stats, refresh=False) progress.update(1) async_loop = asyncio.get_event_loop() i = 0 work_chunk_length = 16 for work_chunk in ichunk(work.items(), work_chunk_length): futures = {} for i, (path, (artist, album)) in enumerate(work_chunk, i): if args.filename == EMBEDDED_ALBUM_ART_SYMBOL: cover_filepath = os.path.join(tmp_dir, '%00u.%s' % (i, args.format.name.lower())) else: cover_filepath = os.path.join(path, args.filename) coroutine = sacad.search_and_download(album, artist, args. format, args.size, cover_filepath, size_tolerance_prct= args.size_tolerance_prct, amazon_tlds=args.amazon_tlds, no_lq_sources=args.no_lq_sources, async_loop=async_loop) future = asyncio.ensure_future(coroutine, loop=async_loop) futures[future] = path, cover_filepath, artist, album for future in futures: future.add_done_callback(update_progress) root_future = asyncio.gather(*futures.keys(), loop=async_loop) async_loop.run_until_complete(root_future)
get_covers
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(work, args):\n with contextlib.ExitStack() as cm:\n if args.filename == EMBEDDED_ALBUM_ART_SYMBOL:\n tmp_prefix = '%s_' % os.path.splitext(os.path.basename(inspect.\n getfile(inspect.currentframe())))[0]\n tmp_dir = cm.enter_context(tempfile.TemporaryDirectory(prefix=\n tmp_prefix))\n stats = collections.OrderedDict((k, 0) for k in ('ok', 'errors',\n 'no result found'))\n progress = cm.enter_context(tqdm.tqdm(total=len(work), miniters=1,\n desc='Searching covers', unit='cover', postfix=stats))\n cm.enter_context(tqdm_logging.redirect_logging(progress))\n\n def x(future):\n path, cover_filepath, artist, album = futures[future]\n try:\n status = future.result()\n except Exception as exception:\n stats['errors'] += 1\n logging.getLogger('sacad_r').error(\n \"Error occured while searching cover for '%s' by '%s' from '%s': %s %s\"\n % (album, artist, path, exception.__class__.\n __qualname__, exception))\n else:\n if status:\n if args.filename == EMBEDDED_ALBUM_ART_SYMBOL:\n try:\n embed_album_art(cover_filepath, path)\n except Exception as exception:\n stats['errors'] += 1\n logging.getLogger('sacad_r').error(\n \"Error occured while embedding cover for '%s' by '%s' from '%s': %s %s\"\n % (album, artist, path, exception.\n __class__.__qualname__, exception))\n else:\n stats['ok'] += 1\n finally:\n os.remove(cover_filepath)\n else:\n stats['ok'] += 1\n else:\n stats['no result found'] += 1\n logging.getLogger('sacad_r').warning(\n \"Unable to find cover for '%s' by '%s' from '%s'\" %\n (album, artist, path))\n progress.set_postfix(stats, refresh=False)\n progress.update(1)\n async_loop = asyncio.get_event_loop()\n i = 0\n work_chunk_length = 16\n for work_chunk in ichunk(work.items(), work_chunk_length):\n futures = {}\n for i, (path, (artist, album)) in enumerate(work_chunk, i):\n if args.filename == EMBEDDED_ALBUM_ART_SYMBOL:\n cover_filepath = os.path.join(tmp_dir, '%00u.%s' % (i,\n args.format.name.lower()))\n else:\n cover_filepath = os.path.join(path, args.filename)\n coroutine = sacad.search_and_download(album, artist, args.\n format, args.size, cover_filepath, size_tolerance_prct=\n args.size_tolerance_prct, amazon_tlds=args.amazon_tlds,\n no_lq_sources=args.no_lq_sources, async_loop=async_loop)\n future = asyncio.ensure_future(coroutine, loop=async_loop)\n futures[future] = path, cover_filepath, artist, album\n for future in futures:\n future.add_done_callback(update_progress)\n root_future = asyncio.gather(*futures.keys(), loop=async_loop)\n async_loop.run_until_complete(root_future)\n", "role": "user" }, { "content": "get_covers", "role": "assistant" } ]
def x(*args, **kwargs): fd, filename = tempfile.mkstemp(*args, **kwargs) os.close(fd) try: yield filename finally: os.remove(filename)
mkstemp
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*args, **kwargs):\n fd, filename = tempfile.mkstemp(*args, **kwargs)\n os.close(fd)\n try:\n yield filename\n finally:\n os.remove(filename)\n", "role": "user" }, { "content": "mkstemp", "role": "assistant" } ]
def x(tqdm_obj, logger=logging.getLogger()): assert len(logger.handlers) == 1 prev_handler = logger.handlers[0] logger.removeHandler(prev_handler) tqdm_handler = TqdmLoggingHandler(tqdm_obj) if prev_handler.formatter is not None: tqdm_handler.setFormatter(prev_handler.formatter) logger.addHandler(tqdm_handler) try: yield finally: logger.removeHandler(tqdm_handler) logger.addHandler(prev_handler)
redirect_logging
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(tqdm_obj, logger=logging.getLogger()):\n assert len(logger.handlers) == 1\n prev_handler = logger.handlers[0]\n logger.removeHandler(prev_handler)\n tqdm_handler = TqdmLoggingHandler(tqdm_obj)\n if prev_handler.formatter is not None:\n tqdm_handler.setFormatter(prev_handler.formatter)\n logger.addHandler(tqdm_handler)\n try:\n yield\n finally:\n logger.removeHandler(tqdm_handler)\n logger.addHandler(prev_handler)\n", "role": "user" }, { "content": "redirect_logging", "role": "assistant" } ]
async def search(self, album, artist): """ Search for a given album/artist and return an iterable of CoverSourceResult. """ self.logger.debug("Searching with source '%s'..." % self.__class__.__name__ ) album = self.processAlbumString(album) artist = self.processArtistString(artist) url_data = self.getSearchUrl(album, artist) if isinstance(url_data, tuple): url, post_data = url_data else: url = url_data post_data = None try: store_in_cache_callback, api_data = await self.fetchResults(url, post_data) results = await self.parseResults(api_data) except Exception as e: self.logger.warning("Search with source '%s' failed: %s %s" % (self .__class__.__name__, e.__class__.__qualname__, e)) return () else: if results: await store_in_cache_callback() futures = [] for result in filter(operator.methodcaller('needMetadataUpdate'), results): coroutine = result.updateImageMetadata() future = asyncio.ensure_future(coroutine) futures.append(future) if futures: await asyncio.wait(futures) for future in futures: future.result() results_excluded_count = 0 reference_only_count = 0 results_kept = [] for result in results: if (result.size[0] + self.size_tolerance_prct * self.target_size / 100 < self.target_size or result.size[1] + self. size_tolerance_prct * self.target_size / 100 < self.target_size or result.format is None or result.needMetadataUpdate()): if result.source_quality is CoverSourceQuality.REFERENCE: result.is_only_reference = True results_kept.append(result) reference_only_count += 1 else: results_excluded_count += 1 else: results_kept.append(result) result_kept_count = len(results_kept) - reference_only_count self.logger.info( "Got %u relevant (%u excluded) results from source '%s'" % ( result_kept_count, results_excluded_count + reference_only_count, self.__class__.__name__)) for result in itertools.filterfalse(operator.attrgetter( 'is_only_reference'), results_kept): self.logger.debug('%s %s%s %4dx%4d %s%s' % (result.__class__. __name__, '(%02d) ' % result.rank if result.rank is not None else '', result.format.name, result.size[0], result.size[1], result. urls[0], ' [x%u]' % len(result.urls) if len(result.urls) > 1 else '')) return results_kept
search
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def search(self, album, artist):\n \"\"\" Search for a given album/artist and return an iterable of CoverSourceResult. \"\"\"\n self.logger.debug(\"Searching with source '%s'...\" % self.__class__.__name__\n )\n album = self.processAlbumString(album)\n artist = self.processArtistString(artist)\n url_data = self.getSearchUrl(album, artist)\n if isinstance(url_data, tuple):\n url, post_data = url_data\n else:\n url = url_data\n post_data = None\n try:\n store_in_cache_callback, api_data = await self.fetchResults(url,\n post_data)\n results = await self.parseResults(api_data)\n except Exception as e:\n self.logger.warning(\"Search with source '%s' failed: %s %s\" % (self\n .__class__.__name__, e.__class__.__qualname__, e))\n return ()\n else:\n if results:\n await store_in_cache_callback()\n futures = []\n for result in filter(operator.methodcaller('needMetadataUpdate'), results):\n coroutine = result.updateImageMetadata()\n future = asyncio.ensure_future(coroutine)\n futures.append(future)\n if futures:\n await asyncio.wait(futures)\n for future in futures:\n future.result()\n results_excluded_count = 0\n reference_only_count = 0\n results_kept = []\n for result in results:\n if (result.size[0] + self.size_tolerance_prct * self.target_size / \n 100 < self.target_size or result.size[1] + self.\n size_tolerance_prct * self.target_size / 100 < self.target_size or\n result.format is None or result.needMetadataUpdate()):\n if result.source_quality is CoverSourceQuality.REFERENCE:\n result.is_only_reference = True\n results_kept.append(result)\n reference_only_count += 1\n else:\n results_excluded_count += 1\n else:\n results_kept.append(result)\n result_kept_count = len(results_kept) - reference_only_count\n self.logger.info(\n \"Got %u relevant (%u excluded) results from source '%s'\" % (\n result_kept_count, results_excluded_count + reference_only_count,\n self.__class__.__name__))\n for result in itertools.filterfalse(operator.attrgetter(\n 'is_only_reference'), results_kept):\n self.logger.debug('%s %s%s %4dx%4d %s%s' % (result.__class__.\n __name__, '(%02d) ' % result.rank if result.rank is not None else\n '', result.format.name, result.size[0], result.size[1], result.\n urls[0], ' [x%u]' % len(result.urls) if len(result.urls) > 1 else\n ''))\n return results_kept\n", "role": "user" }, { "content": "search", "role": "assistant" } ]
async def fetchResults(self, url, post_data=None): """ Get a (store in cache callback, search results) tuple from an URL. """ if post_data is not None: self.logger.debug("Querying URL '%s' %s..." % (url, dict(post_data))) else: self.logger.debug("Querying URL '%s'..." % url) headers = {} self.updateHttpHeaders(headers) return await self.http.query(url, post_data=post_data, headers=headers, cache=__class__.api_cache)
fetchResults
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def fetchResults(self, url, post_data=None):\n \"\"\" Get a (store in cache callback, search results) tuple from an URL. \"\"\"\n if post_data is not None:\n self.logger.debug(\"Querying URL '%s' %s...\" % (url, dict(post_data)))\n else:\n self.logger.debug(\"Querying URL '%s'...\" % url)\n headers = {}\n self.updateHttpHeaders(headers)\n return await self.http.query(url, post_data=post_data, headers=headers,\n cache=__class__.api_cache)\n", "role": "user" }, { "content": "fetchResults", "role": "assistant" } ]
async def probeUrl(self, url, response_headers=None): """ Probe URL reachability from cache or HEAD request. """ self.logger.debug("Probing URL '%s'..." % url) headers = {} self.updateHttpHeaders(headers) resp_headers = {} resp_ok = await self.http.isReachable(url, headers=headers, response_headers=resp_headers, cache=__class__.probe_cache) if response_headers is not None: response_headers.update(resp_headers) return resp_ok
probeUrl
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def probeUrl(self, url, response_headers=None):\n \"\"\" Probe URL reachability from cache or HEAD request. \"\"\"\n self.logger.debug(\"Probing URL '%s'...\" % url)\n headers = {}\n self.updateHttpHeaders(headers)\n resp_headers = {}\n resp_ok = await self.http.isReachable(url, headers=headers,\n response_headers=resp_headers, cache=__class__.probe_cache)\n if response_headers is not None:\n response_headers.update(resp_headers)\n return resp_ok\n", "role": "user" }, { "content": "probeUrl", "role": "assistant" } ]
def x(s): return ''.join(c for c in unicodedata.normalize('NFKD', s) if not unicodedata.combining(c))
unaccentuate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s):\n return ''.join(c for c in unicodedata.normalize('NFKD', s) if not\n unicodedata.combining(c))\n", "role": "user" }, { "content": "unaccentuate", "role": "assistant" } ]
def x(s, *, char_blacklist=string.punctuation): s = ''.join(c for c in s if c not in char_blacklist) return ' '.join(filter(None, s.split(' ')))
unpunctuate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s, *, char_blacklist=string.punctuation):\n s = ''.join(c for c in s if c not in char_blacklist)\n return ' '.join(filter(None, s.split(' ')))\n", "role": "user" }, { "content": "unpunctuate", "role": "assistant" } ]
def x(self, album, artist): params = collections.OrderedDict() params['search-alias'] = 'popular' params['field-artist'] = artist params['field-title'] = album params['sort'] = 'relevancerank' return __class__.assembleUrl(self.base_url, params)
getSearchUrl
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, album, artist):\n params = collections.OrderedDict()\n params['search-alias'] = 'popular'\n params['field-artist'] = artist\n params['field-title'] = album\n params['sort'] = 'relevancerank'\n return __class__.assembleUrl(self.base_url, params)\n", "role": "user" }, { "content": "getSearchUrl", "role": "assistant" } ]
async def parseResults(self, api_data): """ See CoverSource.parseResults. """ results = [] parser = lxml.etree.HTMLParser() html = lxml.etree.XML(api_data.decode('utf-8', 'ignore'), parser) for page_struct_version, result_selector in enumerate(__class__. RESULTS_SELECTORS): result_nodes = result_selector(html) if result_nodes: break for rank, result_node in enumerate(result_nodes, 1): try: img_node = __class__.IMG_SELECTORS[page_struct_version](result_node )[0] except IndexError: continue thumbnail_url = img_node.get('src') url_parts = thumbnail_url.rsplit('.', 2) img_url = '.'.join((url_parts[0], url_parts[2])) size = 500, 500 check_metadata = CoverImageMetadata.SIZE if self.target_size > size[0] and rank <= 3: product_url = __class__.PRODUCT_LINK_SELECTORS[page_struct_version ](result_node)[0].get('href') product_url_split = urllib.parse.urlsplit(product_url) if not product_url_split.scheme: product_url_query = urllib.parse.parse_qsl(product_url_split .query) product_url_query = collections.OrderedDict(product_url_query) try: product_url = product_url_query['url'] except KeyError: product_url = urllib.parse.urljoin(self.base_url, product_url) product_url_split = urllib.parse.urlsplit(product_url) product_url_query = urllib.parse.parse_qsl(product_url_split.query) product_url_query = collections.OrderedDict(product_url_query) try: del product_url_query['qid'] except KeyError: pass product_url_query = urllib.parse.urlencode(product_url_query) product_url_no_ts = urllib.parse.urlunsplit(product_url_split[: 3] + (product_url_query,) + product_url_split[4:]) store_in_cache_callback, product_page_data = (await self. fetchResults(product_url_no_ts)) product_page_html = lxml.etree.XML(product_page_data.decode( 'latin-1'), parser) try: img_node = __class__.PRODUCT_PAGE_IMG_SELECTOR( product_page_html)[0] except IndexError: pass else: better_img_url = img_node.get('data-old-hires') if better_img_url: img_url = better_img_url size_url_hint = img_url.rsplit('.', 2)[1].strip('_') assert size_url_hint.startswith('SL') size_url_hint = int(size_url_hint[2:]) size = size_url_hint, size_url_hint check_metadata = CoverImageMetadata.NONE await store_in_cache_callback() format = CoverImageFormat.JPEG results.append(AmazonCdCoverSourceResult(img_url, size, format, thumbnail_url=thumbnail_url, source=self, rank=rank, check_metadata=check_metadata)) return results
parseResults
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def parseResults(self, api_data):\n \"\"\" See CoverSource.parseResults. \"\"\"\n results = []\n parser = lxml.etree.HTMLParser()\n html = lxml.etree.XML(api_data.decode('utf-8', 'ignore'), parser)\n for page_struct_version, result_selector in enumerate(__class__.\n RESULTS_SELECTORS):\n result_nodes = result_selector(html)\n if result_nodes:\n break\n for rank, result_node in enumerate(result_nodes, 1):\n try:\n img_node = __class__.IMG_SELECTORS[page_struct_version](result_node\n )[0]\n except IndexError:\n continue\n thumbnail_url = img_node.get('src')\n url_parts = thumbnail_url.rsplit('.', 2)\n img_url = '.'.join((url_parts[0], url_parts[2]))\n size = 500, 500\n check_metadata = CoverImageMetadata.SIZE\n if self.target_size > size[0] and rank <= 3:\n product_url = __class__.PRODUCT_LINK_SELECTORS[page_struct_version\n ](result_node)[0].get('href')\n product_url_split = urllib.parse.urlsplit(product_url)\n if not product_url_split.scheme:\n product_url_query = urllib.parse.parse_qsl(product_url_split\n .query)\n product_url_query = collections.OrderedDict(product_url_query)\n try:\n product_url = product_url_query['url']\n except KeyError:\n product_url = urllib.parse.urljoin(self.base_url,\n product_url)\n product_url_split = urllib.parse.urlsplit(product_url)\n product_url_query = urllib.parse.parse_qsl(product_url_split.query)\n product_url_query = collections.OrderedDict(product_url_query)\n try:\n del product_url_query['qid']\n except KeyError:\n pass\n product_url_query = urllib.parse.urlencode(product_url_query)\n product_url_no_ts = urllib.parse.urlunsplit(product_url_split[:\n 3] + (product_url_query,) + product_url_split[4:])\n store_in_cache_callback, product_page_data = (await self.\n fetchResults(product_url_no_ts))\n product_page_html = lxml.etree.XML(product_page_data.decode(\n 'latin-1'), parser)\n try:\n img_node = __class__.PRODUCT_PAGE_IMG_SELECTOR(\n product_page_html)[0]\n except IndexError:\n pass\n else:\n better_img_url = img_node.get('data-old-hires')\n if better_img_url:\n img_url = better_img_url\n size_url_hint = img_url.rsplit('.', 2)[1].strip('_')\n assert size_url_hint.startswith('SL')\n size_url_hint = int(size_url_hint[2:])\n size = size_url_hint, size_url_hint\n check_metadata = CoverImageMetadata.NONE\n await store_in_cache_callback()\n format = CoverImageFormat.JPEG\n results.append(AmazonCdCoverSourceResult(img_url, size, format,\n thumbnail_url=thumbnail_url, source=self, rank=rank,\n check_metadata=check_metadata))\n return results\n", "role": "user" }, { "content": "parseResults", "role": "assistant" } ]
def x(obj): types = tuple([Shader, Mesh, VertexBuffer, IndexBuffer, Texture, Program, Context]) for name in dir(obj): child = getattr(obj, name) if isinstance(child, types): child.delete()
delete_all
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(obj):\n types = tuple([Shader, Mesh, VertexBuffer, IndexBuffer, Texture,\n Program, Context])\n for name in dir(obj):\n child = getattr(obj, name)\n if isinstance(child, types):\n child.delete()\n", "role": "user" }, { "content": "delete_all", "role": "assistant" } ]
def x(filename): version_checker_source = """ import sys import ctypes def get_version(library_handle): ''' Queries and returns the library version tuple or None. ''' major_value = ctypes.c_int(0) major = ctypes.pointer(major_value) minor_value = ctypes.c_int(0) minor = ctypes.pointer(minor_value) rev_value = ctypes.c_int(0) rev = ctypes.pointer(rev_value) if hasattr(library_handle, 'glfwGetVersion'): library_handle.glfwGetVersion(major, minor, rev) version = (major_value.value, minor_value.value, rev_value.value) return version else: return None try: input_func = raw_input except NameError: input_func = input filename = input_func().strip() try: library_handle = ctypes.CDLL(filename) except OSError: pass else: version = get_version(library_handle) print(version) """ args = [sys.executable, '-c', textwrap.dedent(version_checker_source)] process = subprocess.Popen(args, universal_newlines=True, stdin= subprocess.PIPE, stdout=subprocess.PIPE) out = process.communicate(_to_char_p(filename))[0] out = out.strip() if out: return eval(out) else: return None
glfw_get_version
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(filename):\n version_checker_source = \"\"\"\n import sys\n import ctypes\n\n def get_version(library_handle):\n '''\n Queries and returns the library version tuple or None.\n '''\n major_value = ctypes.c_int(0)\n major = ctypes.pointer(major_value)\n minor_value = ctypes.c_int(0)\n minor = ctypes.pointer(minor_value)\n rev_value = ctypes.c_int(0)\n rev = ctypes.pointer(rev_value)\n if hasattr(library_handle, 'glfwGetVersion'):\n library_handle.glfwGetVersion(major, minor, rev)\n version = (major_value.value,\n minor_value.value,\n rev_value.value)\n return version\n else:\n return None\n\n try:\n input_func = raw_input\n except NameError:\n input_func = input\n filename = input_func().strip()\n\n try:\n library_handle = ctypes.CDLL(filename)\n except OSError:\n pass\n else:\n version = get_version(library_handle)\n print(version)\n \"\"\"\n args = [sys.executable, '-c', textwrap.dedent(version_checker_source)]\n process = subprocess.Popen(args, universal_newlines=True, stdin=\n subprocess.PIPE, stdout=subprocess.PIPE)\n out = process.communicate(_to_char_p(filename))[0]\n out = out.strip()\n if out:\n return eval(out)\n else:\n return None\n", "role": "user" }, { "content": "glfw_get_version", "role": "assistant" } ]
def x(cbfun): global _error_callback previous_callback = _error_callback if cbfun is None: cbfun = 0 c_cbfun = _GLFWerrorfun(cbfun) _error_callback = cbfun, c_cbfun cbfun = c_cbfun _glfw.glfwSetErrorCallback(cbfun) if previous_callback is not None and previous_callback[0] != 0: return previous_callback[0]
set_error_callback
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cbfun):\n global _error_callback\n previous_callback = _error_callback\n if cbfun is None:\n cbfun = 0\n c_cbfun = _GLFWerrorfun(cbfun)\n _error_callback = cbfun, c_cbfun\n cbfun = c_cbfun\n _glfw.glfwSetErrorCallback(cbfun)\n if previous_callback is not None and previous_callback[0] != 0:\n return previous_callback[0]\n", "role": "user" }, { "content": "set_error_callback", "role": "assistant" } ]
def x(window): _glfw.glfwDestroyWindow(window) window_addr = ctypes.cast(ctypes.pointer(window), ctypes.POINTER(ctypes .c_ulong)).contents.value for callback_repository in _callback_repositories: if window_addr in callback_repository: del callback_repository[window_addr]
destroy_window
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(window):\n _glfw.glfwDestroyWindow(window)\n window_addr = ctypes.cast(ctypes.pointer(window), ctypes.POINTER(ctypes\n .c_ulong)).contents.value\n for callback_repository in _callback_repositories:\n if window_addr in callback_repository:\n del callback_repository[window_addr]\n", "role": "user" }, { "content": "destroy_window", "role": "assistant" } ]
def x(self): size = self.width, self.height bits = self.red_bits, self.green_bits, self.blue_bits return size, bits, self.refresh_rate
unwrap
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n size = self.width, self.height\n bits = self.red_bits, self.green_bits, self.blue_bits\n return size, bits, self.refresh_rate\n", "role": "user" }, { "content": "unwrap", "role": "assistant" } ]
def x(self, gammaramp): red, green, blue = gammaramp size = min(len(red), len(green), len(blue)) array_type = ctypes.c_ushort * size self.size = ctypes.c_uint(size) self.red_array = array_type() self.green_array = array_type() self.blue_array = array_type() for i in range(self.size): self.red_array[i] = int(red[i] * 65535) self.green_array[i] = int(green[i] * 65535) self.blue_array[i] = int(blue[i] * 65535) pointer_type = ctypes.POINTER(ctypes.c_ushort) self.red = ctypes.cast(self.red_array, pointer_type) self.green = ctypes.cast(self.green_array, pointer_type) self.blue = ctypes.cast(self.blue_array, pointer_type)
wrap
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, gammaramp):\n red, green, blue = gammaramp\n size = min(len(red), len(green), len(blue))\n array_type = ctypes.c_ushort * size\n self.size = ctypes.c_uint(size)\n self.red_array = array_type()\n self.green_array = array_type()\n self.blue_array = array_type()\n for i in range(self.size):\n self.red_array[i] = int(red[i] * 65535)\n self.green_array[i] = int(green[i] * 65535)\n self.blue_array[i] = int(blue[i] * 65535)\n pointer_type = ctypes.POINTER(ctypes.c_ushort)\n self.red = ctypes.cast(self.red_array, pointer_type)\n self.green = ctypes.cast(self.green_array, pointer_type)\n self.blue = ctypes.cast(self.blue_array, pointer_type)\n", "role": "user" }, { "content": "wrap", "role": "assistant" } ]
def x(self): red = [(self.red[i] / 65535.0) for i in range(self.size)] green = [(self.green[i] / 65535.0) for i in range(self.size)] blue = [(self.blue[i] / 65535.0) for i in range(self.size)] return red, green, blue
unwrap
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n red = [(self.red[i] / 65535.0) for i in range(self.size)]\n green = [(self.green[i] / 65535.0) for i in range(self.size)]\n blue = [(self.blue[i] / 65535.0) for i in range(self.size)]\n return red, green, blue\n", "role": "user" }, { "content": "unwrap", "role": "assistant" } ]
def x(value): r = (value >> 8 * 2 & 255) / 255.0 g = (value >> 8 * 1 & 255) / 255.0 b = (value >> 8 * 0 & 255) / 255.0 return r, g, b
hex_color
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(value):\n r = (value >> 8 * 2 & 255) / 255.0\n g = (value >> 8 * 1 & 255) / 255.0\n b = (value >> 8 * 0 & 255) / 255.0\n return r, g, b\n", "role": "user" }, { "content": "hex_color", "role": "assistant" } ]
def x(vector): d = sum(x * x for x in vector) ** 0.5 return tuple(x / d for x in vector)
normalize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(vector):\n d = sum(x * x for x in vector) ** 0.5\n return tuple(x / d for x in vector)\n", "role": "user" }, { "content": "normalize", "role": "assistant" } ]
def x(p1, p2): return sum((a - b) ** 2 for a, b in zip(p1, p2)) ** 0.5
distance
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(p1, p2):\n return sum((a - b) ** 2 for a, b in zip(p1, p2)) ** 0.5\n", "role": "user" }, { "content": "distance", "role": "assistant" } ]
def x(v1, v2): return v1[1] * v2[2] - v1[2] * v2[1], v1[2] * v2[0] - v1[0] * v2[2], v1[0 ] * v2[1] - v1[1] * v2[0]
cross
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(v1, v2):\n return v1[1] * v2[2] - v1[2] * v2[1], v1[2] * v2[0] - v1[0] * v2[2], v1[0\n ] * v2[1] - v1[1] * v2[0]\n", "role": "user" }, { "content": "cross", "role": "assistant" } ]
def x(v1, v2): x1, y1, z1 = v1 x2, y2, z2 = v2 return x1 * x2 + y1 * y2 + z1 * z2
dot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(v1, v2):\n x1, y1, z1 = v1\n x2, y2, z2 = v2\n return x1 * x2 + y1 * y2 + z1 * z2\n", "role": "user" }, { "content": "dot", "role": "assistant" } ]
def x(v1, v2): return tuple(a + b for a, b in zip(v1, v2))
add
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(v1, v2):\n return tuple(a + b for a, b in zip(v1, v2))\n", "role": "user" }, { "content": "add", "role": "assistant" } ]
def x(v1, v2): return tuple(a - b for a, b in zip(v1, v2))
sub
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(v1, v2):\n return tuple(a - b for a, b in zip(v1, v2))\n", "role": "user" }, { "content": "sub", "role": "assistant" } ]
def x(v1, v2, t): return add(v1, mul(sub(v2, v1), t))
interpolate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(v1, v2, t):\n return add(v1, mul(sub(v2, v1), t))\n", "role": "user" }, { "content": "interpolate", "role": "assistant" } ]
def x(a, b, c): x1, y1, z1 = a x2, y2, z2 = b x3, y3, z3 = c ab = x2 - x1, y2 - y1, z2 - z1 ac = x3 - x1, y3 - y1, z3 - z1 x, y, z = cross(ab, ac) d = (x * x + y * y + z * z) ** 0.5 return x / d, y / d, z / d
normal_from_points
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(a, b, c):\n x1, y1, z1 = a\n x2, y2, z2 = b\n x3, y3, z3 = c\n ab = x2 - x1, y2 - y1, z2 - z1\n ac = x3 - x1, y3 - y1, z3 - z1\n x, y, z = cross(ab, ac)\n d = (x * x + y * y + z * z) ** 0.5\n return x / d, y / d, z / d\n", "role": "user" }, { "content": "normal_from_points", "role": "assistant" } ]
def x(positions, normals): lookup = defaultdict(list) for position, normal in zip(positions, normals): lookup[position].append(normal) result = [] for position in positions: tx = ty = tz = 0 for x, y, z in lookup[position]: tx += x ty += y tz += z d = (tx * tx + ty * ty + tz * tz) ** 0.5 result.append((tx / d, ty / d, tz / d)) return result
smooth_normals
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(positions, normals):\n lookup = defaultdict(list)\n for position, normal in zip(positions, normals):\n lookup[position].append(normal)\n result = []\n for position in positions:\n tx = ty = tz = 0\n for x, y, z in lookup[position]:\n tx += x\n ty += y\n tz += z\n d = (tx * tx + ty * ty + tz * tz) ** 0.5\n result.append((tx / d, ty / d, tz / d))\n return result\n", "role": "user" }, { "content": "smooth_normals", "role": "assistant" } ]
def x(positions): x0, y0, z0 = x1, y1, z1 = positions[0] for x, y, z in positions: x0 = min(x0, x) y0 = min(y0, y) z0 = min(z0, z) x1 = max(x1, x) y1 = max(y1, y) z1 = max(z1, z) return (x0, y0, z0), (x1, y1, z1)
bounding_box
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(positions):\n x0, y0, z0 = x1, y1, z1 = positions[0]\n for x, y, z in positions:\n x0 = min(x0, x)\n y0 = min(y0, y)\n z0 = min(z0, z)\n x1 = max(x1, x)\n y1 = max(y1, y)\n z1 = max(z1, z)\n return (x0, y0, z0), (x1, y1, z1)\n", "role": "user" }, { "content": "bounding_box", "role": "assistant" } ]
def x(positions): (x0, y0, z0), (x1, y1, z1) = bounding_box(positions) dx = x1 - (x1 - x0) / 2.0 dy = y1 - (y1 - y0) / 2.0 dz = z1 - (z1 - z0) / 2.0 result = [] for x, y, z in positions: result.append((x - dx, y - dy, z - dz)) return result
recenter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(positions):\n (x0, y0, z0), (x1, y1, z1) = bounding_box(positions)\n dx = x1 - (x1 - x0) / 2.0\n dy = y1 - (y1 - y0) / 2.0\n dz = z1 - (z1 - z0) / 2.0\n result = []\n for x, y, z in positions:\n result.append((x - dx, y - dy, z - dz))\n return result\n", "role": "user" }, { "content": "recenter", "role": "assistant" } ]
def x(*args): result = [] for array in zip(*args): result.append(tuple(flatten(array))) return result
interleave
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*args):\n result = []\n for array in zip(*args):\n result.append(tuple(flatten(array)))\n return result\n", "role": "user" }, { "content": "interleave", "role": "assistant" } ]
def x(iterable, keyfunc=None): seen = set() for item in iterable: key = item if keyfunc is None else keyfunc(item) if key not in seen: seen.add(key) yield item
distinct
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(iterable, keyfunc=None):\n seen = set()\n for item in iterable:\n key = item if keyfunc is None else keyfunc(item)\n if key not in seen:\n seen.add(key)\n yield item\n", "role": "user" }, { "content": "distinct", "role": "assistant" } ]
def x(v1, v2, v3, o, d): eps = 1e-06 e1 = sub(v2, v1) e2 = sub(v3, v1) p = cross(d, e2) det = dot(e1, p) if abs(det) < eps: return None inv = 1.0 / det t = sub(o, v1) u = dot(t, p) * inv if u < 0 or u > 1: return None q = cross(t, e1) v = dot(d, q) * inv if v < 0 or v > 1: return None t = dot(e2, q) * inv if t > eps: return t return None
ray_triangle_intersection
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(v1, v2, v3, o, d):\n eps = 1e-06\n e1 = sub(v2, v1)\n e2 = sub(v3, v1)\n p = cross(d, e2)\n det = dot(e1, p)\n if abs(det) < eps:\n return None\n inv = 1.0 / det\n t = sub(o, v1)\n u = dot(t, p) * inv\n if u < 0 or u > 1:\n return None\n q = cross(t, e1)\n v = dot(d, q) * inv\n if v < 0 or v > 1:\n return None\n t = dot(e2, q) * inv\n if t > eps:\n return t\n return None\n", "role": "user" }, { "content": "ray_triangle_intersection", "role": "assistant" } ]
def x(fmt, data): func = struct.Struct(fmt).pack return create_string_buffer(''.join([func(x) for x in data]))
pack_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fmt, data):\n func = struct.Struct(fmt).pack\n return create_string_buffer(''.join([func(x) for x in data]))\n", "role": "user" }, { "content": "pack_list", "role": "assistant" } ]
def x(self, jquery=False): if jquery: e = JQuery(self) e.click() else: super(Clickable, self).click()
click
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, jquery=False):\n if jquery:\n e = JQuery(self)\n e.click()\n else:\n super(Clickable, self).click()\n", "role": "user" }, { "content": "click", "role": "assistant" } ]
def x(cookie, keys_map=WEB_DRIVER_COOKIE_KEYS_MAP): cookie_dict = dict() for k in keys_map.keys(): key = _to_unicode_if_str(keys_map[k]) value = _to_unicode_if_str(getattr(cookie, k)) cookie_dict[key] = value return cookie_dict
convert_cookie_to_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cookie, keys_map=WEB_DRIVER_COOKIE_KEYS_MAP):\n cookie_dict = dict()\n for k in keys_map.keys():\n key = _to_unicode_if_str(keys_map[k])\n value = _to_unicode_if_str(getattr(cookie, k))\n cookie_dict[key] = value\n return cookie_dict\n", "role": "user" }, { "content": "convert_cookie_to_dict", "role": "assistant" } ]
def x(driver, cookies): for cookie in cookies: driver.add_cookie(convert_cookie_to_dict(cookie)) return driver
add_cookies_to_web_driver
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(driver, cookies):\n for cookie in cookies:\n driver.add_cookie(convert_cookie_to_dict(cookie))\n return driver\n", "role": "user" }, { "content": "add_cookies_to_web_driver", "role": "assistant" } ]
def x(self, options, conf): self.conf = conf self.when = options.browser_closer_when
configure
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, options, conf):\n self.conf = conf\n self.when = options.browser_closer_when\n", "role": "user" }, { "content": "configure", "role": "assistant" } ]
def x(self, root): basename = os.path.basename(root) if os.path.splitext(basename)[0] != '__init__' and basename.startswith('_' ): return location = self._determine_location_for(root) if os.path.isfile(root): self._index_module(root, location) elif os.path.isdir(root) and os.path.exists(os.path.join(root, '__init__.py')): self._index_package(root, location)
index_path
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, root):\n basename = os.path.basename(root)\n if os.path.splitext(basename)[0] != '__init__' and basename.startswith('_'\n ):\n return\n location = self._determine_location_for(root)\n if os.path.isfile(root):\n self._index_module(root, location)\n elif os.path.isdir(root) and os.path.exists(os.path.join(root,\n '__init__.py')):\n self._index_package(root, location)\n", "role": "user" }, { "content": "index_path", "role": "assistant" } ]
def x(self, paths=None, name=None, refresh=False): if not paths: paths = sys.path if not name: name = 'default' self._name = name idx_dir = get_cache_dir() idx_file = os.path.join(idx_dir, name + '.json') if os.path.exists(idx_file) and not refresh: with open(idx_file) as fd: self.deserialize(fd) else: self.build_index(paths) with open(idx_file, 'w') as fd: self.serialize(fd) return self
get_or_create_index
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, paths=None, name=None, refresh=False):\n if not paths:\n paths = sys.path\n if not name:\n name = 'default'\n self._name = name\n idx_dir = get_cache_dir()\n idx_file = os.path.join(idx_dir, name + '.json')\n if os.path.exists(idx_file) and not refresh:\n with open(idx_file) as fd:\n self.deserialize(fd)\n else:\n self.build_index(paths)\n with open(idx_file, 'w') as fd:\n self.serialize(fd)\n return self\n", "role": "user" }, { "content": "get_or_create_index", "role": "assistant" } ]
def x(self, symbol): scores = [] path = [] def x(module, variable): prefix = module.split('.') if variable is not None: prefix.append(variable) seeking = symbol.split('.') new_module = [] while prefix and seeking[0] != prefix[0]: new_module.append(prefix.pop(0)) if new_module: module, variable = '.'.join(new_module), prefix[0] else: variable = None return module, variable def x(scope, scale): sub_path, score = self._score_key(scope, full_key) if score > 0.1: try: i = sub_path.index(None) sub_path, from_symbol = sub_path[:i], '.'.join(sub_path[i + 1:] ) except ValueError: from_symbol = None package_path = '.'.join(path + sub_path) package_path, from_symbol = fixup(package_path, from_symbol) scores.append((score * scale, package_path, from_symbol)) for key, subscope in scope._tree.items(): if type(subscope) is not float: path.append(key) score_walk(subscope, subscope.score * scale - 0.1) path.pop() full_key = symbol.split('.') score_walk(self, 1.0) scores.sort(reverse=True) return scores
symbol_scores
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, symbol):\n scores = []\n path = []\n\n def x(module, variable):\n prefix = module.split('.')\n if variable is not None:\n prefix.append(variable)\n seeking = symbol.split('.')\n new_module = []\n while prefix and seeking[0] != prefix[0]:\n new_module.append(prefix.pop(0))\n if new_module:\n module, variable = '.'.join(new_module), prefix[0]\n else:\n variable = None\n return module, variable\n\n def x(scope, scale):\n sub_path, score = self._score_key(scope, full_key)\n if score > 0.1:\n try:\n i = sub_path.index(None)\n sub_path, from_symbol = sub_path[:i], '.'.join(sub_path[i + 1:]\n )\n except ValueError:\n from_symbol = None\n package_path = '.'.join(path + sub_path)\n package_path, from_symbol = fixup(package_path, from_symbol)\n scores.append((score * scale, package_path, from_symbol))\n for key, subscope in scope._tree.items():\n if type(subscope) is not float:\n path.append(key)\n score_walk(subscope, subscope.score * scale - 0.1)\n path.pop()\n full_key = symbol.split('.')\n score_walk(self, 1.0)\n scores.sort(reverse=True)\n return scores\n", "role": "user" }, { "content": "symbol_scores", "role": "assistant" } ]
def x(self, path): path = path.split('.') node = self while node._parent: node = node._parent for name in path: node = node._tree.get(name, None) if node is None or type(node) is float: return None return node
find
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, path):\n path = path.split('.')\n node = self\n while node._parent:\n node = node._parent\n for name in path:\n node = node._tree.get(name, None)\n if node is None or type(node) is float:\n return None\n return node\n", "role": "user" }, { "content": "find", "role": "assistant" } ]
def x(self, path): path = path.split('.') node = self while node._parent: node = node._parent location = node.location for name in path: tree = node._tree.get(name, None) if tree is None or type(tree) is float: return location location = tree.location return location
location_for
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, path):\n path = path.split('.')\n node = self\n while node._parent:\n node = node._parent\n location = node.location\n for name in path:\n tree = node._tree.get(name, None)\n if tree is None or type(tree) is float:\n return location\n location = tree.location\n return location\n", "role": "user" }, { "content": "location_for", "role": "assistant" } ]
def x(self, option): items_list = self.get_options() for item in items_list: if item.get_attribute('value') == option: item.click() break
select_option
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, option):\n items_list = self.get_options()\n for item in items_list:\n if item.get_attribute('value') == option:\n item.click()\n break\n", "role": "user" }, { "content": "select_option", "role": "assistant" } ]
def x(self, attribute): items_list = self.get_options() return next(iter([item.get_attribute(attribute) for item in items_list if item.is_selected()]), None)
get_attribute_selected
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, attribute):\n items_list = self.get_options()\n return next(iter([item.get_attribute(attribute) for item in items_list if\n item.is_selected()]), None)\n", "role": "user" }, { "content": "get_attribute_selected", "role": "assistant" } ]
def x(self, text): xpath = './/option[normalize-space(.) = {0}]'.format(self. _escape_string(text)) opts = self.find_elements_by_xpath(xpath) matched = False for opt in opts: self._set_selected(opt) if not self.is_multiple: return matched = True if len(opts) == 0 and ' ' in text: sub_string_without_space = self._get_longest_token(text) if sub_string_without_space == '': candidates = self.get_options() else: xpath = './/option[contains(.,{0})]'.format(self._escape_string (sub_string_without_space)) candidates = self.find_elements_by_xpath(xpath) for candidate in candidates: if text == candidate.text: self._set_selected(candidate) if not self.is_multiple: return matched = True if not matched: raise NoSuchElementException( 'Could not locate element with visible text: ' + str(text))
select_by_visible_text
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, text):\n xpath = './/option[normalize-space(.) = {0}]'.format(self.\n _escape_string(text))\n opts = self.find_elements_by_xpath(xpath)\n matched = False\n for opt in opts:\n self._set_selected(opt)\n if not self.is_multiple:\n return\n matched = True\n if len(opts) == 0 and ' ' in text:\n sub_string_without_space = self._get_longest_token(text)\n if sub_string_without_space == '':\n candidates = self.get_options()\n else:\n xpath = './/option[contains(.,{0})]'.format(self._escape_string\n (sub_string_without_space))\n candidates = self.find_elements_by_xpath(xpath)\n for candidate in candidates:\n if text == candidate.text:\n self._set_selected(candidate)\n if not self.is_multiple:\n return\n matched = True\n if not matched:\n raise NoSuchElementException(\n 'Could not locate element with visible text: ' + str(text))\n", "role": "user" }, { "content": "select_by_visible_text", "role": "assistant" } ]
def x(*args, **kwargs): kwargs.setdefault('sleep_seconds', (1, None)) kwargs.setdefault('expected_exceptions', WebDriverException) kwargs.setdefault('timeout_seconds', webium.settings.wait_timeout) return wait_lib(*args, **kwargs)
wait
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*args, **kwargs):\n kwargs.setdefault('sleep_seconds', (1, None))\n kwargs.setdefault('expected_exceptions', WebDriverException)\n kwargs.setdefault('timeout_seconds', webium.settings.wait_timeout)\n return wait_lib(*args, **kwargs)\n", "role": "user" }, { "content": "wait", "role": "assistant" } ]
def x(source, filename=None): if isinstance(source, text_type) and sys.version_info[0] == 2: source = CODING_COOKIE_RE.sub('\\1', source, 1) return ast.parse(source, filename or '<unknown>')
parse_ast
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(source, filename=None):\n if isinstance(source, text_type) and sys.version_info[0] == 2:\n source = CODING_COOKIE_RE.sub('\\\\1', source, 1)\n return ast.parse(source, filename or '<unknown>')\n", "role": "user" }, { "content": "parse_ast", "role": "assistant" } ]
def x(self): unresolved = set() unreferenced = self._definitions.copy() self._collect_unresolved_and_unreferenced(set(), set(), unresolved, unreferenced, frozenset(self._definitions), start=True) return unresolved, unreferenced - Scope.ALL_BUILTINS
find_unresolved_and_unreferenced_symbols
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n unresolved = set()\n unreferenced = self._definitions.copy()\n self._collect_unresolved_and_unreferenced(set(), set(), unresolved,\n unreferenced, frozenset(self._definitions), start=True)\n return unresolved, unreferenced - Scope.ALL_BUILTINS\n", "role": "user" }, { "content": "find_unresolved_and_unreferenced_symbols", "role": "assistant" } ]
def x(key): CACHED_KEY_FILE = os.path.join(CURRENT_DIR, key) try: return json.loads(open(CACHED_KEY_FILE, 'rb').read().decode('UTF-8'))[ '_'] except (IOError, ValueError): return None
get_item
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(key):\n CACHED_KEY_FILE = os.path.join(CURRENT_DIR, key)\n try:\n return json.loads(open(CACHED_KEY_FILE, 'rb').read().decode('UTF-8'))[\n '_']\n except (IOError, ValueError):\n return None\n", "role": "user" }, { "content": "get_item", "role": "assistant" } ]
def x(key, value): CACHED_KEY_FILE = os.path.join(CURRENT_DIR, key) open(CACHED_KEY_FILE, 'wb').write(json.dumps({'_': value}).encode('UTF-8')) return value
set_item
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(key, value):\n CACHED_KEY_FILE = os.path.join(CURRENT_DIR, key)\n open(CACHED_KEY_FILE, 'wb').write(json.dumps({'_': value}).encode('UTF-8'))\n return value\n", "role": "user" }, { "content": "set_item", "role": "assistant" } ]
def x(key): CACHED_KEY_FILE = os.path.join(CURRENT_DIR, key) if os.path.isfile(CACHED_KEY_FILE): os.remove(CACHED_KEY_FILE)
delete_item
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(key):\n CACHED_KEY_FILE = os.path.join(CURRENT_DIR, key)\n if os.path.isfile(CACHED_KEY_FILE):\n os.remove(CACHED_KEY_FILE)\n", "role": "user" }, { "content": "delete_item", "role": "assistant" } ]
def x(self, data): if isinstance(data, dict) or isinstance(data, list): self._raw_data = data self._json_data = copy.deepcopy(self._raw_data) else: raise TypeError('Provided Data is not json')
parse_json_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data):\n if isinstance(data, dict) or isinstance(data, list):\n self._raw_data = data\n self._json_data = copy.deepcopy(self._raw_data)\n else:\n raise TypeError('Provided Data is not json')\n", "role": "user" }, { "content": "parse_json_data", "role": "assistant" } ]
def x(self, file_path): if file_path == '' or os.path.splitext(file_path)[1] != '.json': raise IOError('Invalid Json file') with open(file_path) as json_file: self._raw_data = json.load(json_file) self._json_data = copy.deepcopy(self._raw_data)
parse_json_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, file_path):\n if file_path == '' or os.path.splitext(file_path)[1] != '.json':\n raise IOError('Invalid Json file')\n with open(file_path) as json_file:\n self._raw_data = json.load(json_file)\n self._json_data = copy.deepcopy(self._raw_data)\n", "role": "user" }, { "content": "parse_json_file", "role": "assistant" } ]
def x(self, key, data): if key.isdigit(): return data[int(key)] if key not in data: raise KeyError('Key not exists') return data.get(key)
get_value_from_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, data):\n if key.isdigit():\n return data[int(key)]\n if key not in data:\n raise KeyError('Key not exists')\n return data.get(key)\n", "role": "user" }, { "content": "get_value_from_data", "role": "assistant" } ]
def x(self, root): leafs = root.strip(' ').split('.') for leaf in leafs: if leaf: self._json_data = self.__get_value_from_data(leaf, self._json_data) return self
at
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, root):\n leafs = root.strip(' ').split('.')\n for leaf in leafs:\n if leaf:\n self._json_data = self.__get_value_from_data(leaf, self._json_data)\n return self\n", "role": "user" }, { "content": "at", "role": "assistant" } ]
def x(self, data={}): if data and (isinstance(data, dict) or isinstance(data, list)): self._json_data = data else: self._json_data = copy.deepcopy(self._raw_data) self.__reset_queries() return self
reset
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data={}):\n if data and (isinstance(data, dict) or isinstance(data, list)):\n self._json_data = data\n else:\n self._json_data = copy.deepcopy(self._raw_data)\n self.__reset_queries()\n return self\n", "role": "user" }, { "content": "reset", "role": "assistant" } ]
def x(self, query_items): temp_index = self._current_query_index if len(self._queries) - 1 < temp_index: self._queries.append([]) self._queries[temp_index].append(query_items)
store_query
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, query_items):\n temp_index = self._current_query_index\n if len(self._queries) - 1 < temp_index:\n self._queries.append([])\n self._queries[temp_index].append(query_items)\n", "role": "user" }, { "content": "store_query", "role": "assistant" } ]
def x(self): def x(item): or_check = False for queries in self._queries: and_check = True for query in queries: and_check &= self._matcher._match(item.get(query.get('key'), None), query.get('operator'), query.get('value')) or_check |= and_check return or_check self._json_data = list(filter(lambda item: func(item), self._json_data))
execute_queries
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n\n def x(item):\n or_check = False\n for queries in self._queries:\n and_check = True\n for query in queries:\n and_check &= self._matcher._match(item.get(query.get('key'),\n None), query.get('operator'), query.get('value'))\n or_check |= and_check\n return or_check\n self._json_data = list(filter(lambda item: func(item), self._json_data))\n", "role": "user" }, { "content": "execute_queries", "role": "assistant" } ]
def x(self, key, operator, value): self.__store_query({'key': key, 'operator': operator, 'value': value}) return self
where
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, operator, value):\n self.__store_query({'key': key, 'operator': operator, 'value': value})\n return self\n", "role": "user" }, { "content": "where", "role": "assistant" } ]
def x(self, key, operator, value): if len(self._queries) > 0: self._current_query_index += 1 self.__store_query({'key': key, 'operator': operator, 'value': value}) return self
or_where
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, operator, value):\n if len(self._queries) > 0:\n self._current_query_index += 1\n self.__store_query({'key': key, 'operator': operator, 'value': value})\n return self\n", "role": "user" }, { "content": "or_where", "role": "assistant" } ]
def x(self, index): self.__prepare() return None if self.count() < math.fabs(index) else self._json_data[index]
nth
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, index):\n self.__prepare()\n return None if self.count() < math.fabs(index) else self._json_data[index]\n", "role": "user" }, { "content": "nth", "role": "assistant" } ]
def x(self, property): self.__prepare() total = 0 for i in self._json_data: total += i.get(property) return total
sum
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, property):\n self.__prepare()\n total = 0\n for i in self._json_data:\n total += i.get(property)\n return total\n", "role": "user" }, { "content": "sum", "role": "assistant" } ]
def x(self, property): self.__prepare() try: return max(self._json_data, key=lambda x: x[property]).get(property) except KeyError: raise KeyError('Key is not exists')
max
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, property):\n self.__prepare()\n try:\n return max(self._json_data, key=lambda x: x[property]).get(property)\n except KeyError:\n raise KeyError('Key is not exists')\n", "role": "user" }, { "content": "max", "role": "assistant" } ]
def x(self, property): self.__prepare() return self.sum(property) / self.count()
avg
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, property):\n self.__prepare()\n return self.sum(property) / self.count()\n", "role": "user" }, { "content": "avg", "role": "assistant" } ]
def x(self, size=0): if size == 0: raise ValueError('Invalid chunk size') self.__prepare() _new_content = [] while len(self._json_data) > 0: _new_content.append(self._json_data[0:size]) self._json_data = self._json_data[size:] self._json_data = _new_content return self._json_data
chunk
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, size=0):\n if size == 0:\n raise ValueError('Invalid chunk size')\n self.__prepare()\n _new_content = []\n while len(self._json_data) > 0:\n _new_content.append(self._json_data[0:size])\n self._json_data = self._json_data[size:]\n self._json_data = _new_content\n return self._json_data\n", "role": "user" }, { "content": "chunk", "role": "assistant" } ]
def x(self, property): self.__prepare() group_data = {} for data in self._json_data: if data[property] not in group_data: group_data[data[property]] = [] group_data[data[property]].append(data) self._json_data = group_data return self
group_by
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, property):\n self.__prepare()\n group_data = {}\n for data in self._json_data:\n if data[property] not in group_data:\n group_data[data[property]] = []\n group_data[data[property]].append(data)\n self._json_data = group_data\n return self\n", "role": "user" }, { "content": "group_by", "role": "assistant" } ]