source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(self, req, resp, origin): if self._cors_config['allow_credentials_all_origins']: self._set_allow_credentials(resp) return True if origin in self._cors_config['allow_credentials_origins_list']: self._set_allow_credentials(resp) return True credentials_regex = self._cors_config['allow_credentials_origins_regex'] if credentials_regex: if credentials_regex.match(origin): self._set_allow_credentials(resp) return True return False
process_credentials
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, req, resp, origin):\n if self._cors_config['allow_credentials_all_origins']:\n self._set_allow_credentials(resp)\n return True\n if origin in self._cors_config['allow_credentials_origins_list']:\n self._set_allow_credentials(resp)\n return True\n credentials_regex = self._cors_config['allow_credentials_origins_regex']\n if credentials_regex:\n if credentials_regex.match(origin):\n self._set_allow_credentials(resp)\n return True\n return False\n", "role": "user" }, { "content": "process_credentials", "role": "assistant" } ]
def x(self, email_messages): if not email_messages: return sent_message_count = 0 for email_message in email_messages: if self._send(email_message): sent_message_count += 1 return sent_message_count
send_messages
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, email_messages):\n if not email_messages:\n return\n sent_message_count = 0\n for email_message in email_messages:\n if self._send(email_message):\n sent_message_count += 1\n return sent_message_count\n", "role": "user" }, { "content": "send_messages", "role": "assistant" } ]
def x(self, email_message): pre_send.send(self.__class__, message=email_message) if not email_message.recipients(): return False from_email = sanitize_address(email_message.from_email, email_message. encoding) recipients = [sanitize_address(addr, email_message.encoding) for addr in email_message.recipients()] message = email_message.message().as_bytes(linesep='\r\n') try: result = self.conn.send_raw_email(Source=from_email, Destinations= recipients, RawMessage={'Data': message}) message_id = result['MessageId'] post_send.send(self.__class__, message=email_message, message_id= message_id) except ClientError: if not self.fail_silently: raise return False return True
send
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, email_message):\n pre_send.send(self.__class__, message=email_message)\n if not email_message.recipients():\n return False\n from_email = sanitize_address(email_message.from_email, email_message.\n encoding)\n recipients = [sanitize_address(addr, email_message.encoding) for addr in\n email_message.recipients()]\n message = email_message.message().as_bytes(linesep='\\r\\n')\n try:\n result = self.conn.send_raw_email(Source=from_email, Destinations=\n recipients, RawMessage={'Data': message})\n message_id = result['MessageId']\n post_send.send(self.__class__, message=email_message, message_id=\n message_id)\n except ClientError:\n if not self.fail_silently:\n raise\n return False\n return True\n", "role": "user" }, { "content": "send", "role": "assistant" } ]
def x(request, path, document_root=None, show_indexes=False, default=''): path = posixpath.normpath(unquote(path)) path = path.lstrip('/') newpath = '' for part in path.split('/'): if not part: continue drive, part = os.path.splitdrive(part) head, part = os.path.split(part) if part in (os.curdir, os.pardir): continue newpath = os.path.join(newpath, part).replace('\\', '/') if newpath and path != newpath: return HttpResponseRedirect(newpath) fullpath = os.path.join(document_root, newpath) if os.path.isdir(fullpath) and default: defaultpath = os.path.join(fullpath, default) if os.path.exists(defaultpath): fullpath = defaultpath if os.path.isdir(fullpath): if show_indexes: return directory_index(newpath, fullpath) raise Http404('Directory indexes are not allowed here.') if not os.path.exists(fullpath): raise Http404('"%s" does not exist' % fullpath) statobj = os.stat(fullpath) mimetype = mimetypes.guess_type(fullpath)[0] or 'application/octet-stream' if not was_modified_since(request.META.get('HTTP_IF_MODIFIED_SINCE'), statobj[stat.ST_MTIME], statobj[stat.ST_SIZE]): if django.VERSION > (1, 6): return HttpResponseNotModified(content_type=mimetype) else: return HttpResponseNotModified(mimetype=mimetype) contents = open(fullpath, 'rb').read() if django.VERSION > (1, 6): response = HttpResponse(contents, content_type=mimetype) else: response = HttpResponse(contents, mimetype=mimetype) response['Last-Modified'] = http_date(statobj[stat.ST_MTIME]) response['Content-Length'] = len(contents) return response
serve
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(request, path, document_root=None, show_indexes=False, default=''):\n path = posixpath.normpath(unquote(path))\n path = path.lstrip('/')\n newpath = ''\n for part in path.split('/'):\n if not part:\n continue\n drive, part = os.path.splitdrive(part)\n head, part = os.path.split(part)\n if part in (os.curdir, os.pardir):\n continue\n newpath = os.path.join(newpath, part).replace('\\\\', '/')\n if newpath and path != newpath:\n return HttpResponseRedirect(newpath)\n fullpath = os.path.join(document_root, newpath)\n if os.path.isdir(fullpath) and default:\n defaultpath = os.path.join(fullpath, default)\n if os.path.exists(defaultpath):\n fullpath = defaultpath\n if os.path.isdir(fullpath):\n if show_indexes:\n return directory_index(newpath, fullpath)\n raise Http404('Directory indexes are not allowed here.')\n if not os.path.exists(fullpath):\n raise Http404('\"%s\" does not exist' % fullpath)\n statobj = os.stat(fullpath)\n mimetype = mimetypes.guess_type(fullpath)[0] or 'application/octet-stream'\n if not was_modified_since(request.META.get('HTTP_IF_MODIFIED_SINCE'),\n statobj[stat.ST_MTIME], statobj[stat.ST_SIZE]):\n if django.VERSION > (1, 6):\n return HttpResponseNotModified(content_type=mimetype)\n else:\n return HttpResponseNotModified(mimetype=mimetype)\n contents = open(fullpath, 'rb').read()\n if django.VERSION > (1, 6):\n response = HttpResponse(contents, content_type=mimetype)\n else:\n response = HttpResponse(contents, mimetype=mimetype)\n response['Last-Modified'] = http_date(statobj[stat.ST_MTIME])\n response['Content-Length'] = len(contents)\n return response\n", "role": "user" }, { "content": "serve", "role": "assistant" } ]
def x(header=None, mtime=0, size=0): try: if header is None: raise ValueError matches = re.match('^([^;]+)(; length=([0-9]+))?$', header, re. IGNORECASE) header_mtime = parse_http_date(matches.group(1)) header_len = matches.group(3) if header_len and int(header_len) != size: raise ValueError if int(mtime) > header_mtime: raise ValueError except (AttributeError, ValueError, OverflowError): return True return False
was_modified_since
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(header=None, mtime=0, size=0):\n try:\n if header is None:\n raise ValueError\n matches = re.match('^([^;]+)(; length=([0-9]+))?$', header, re.\n IGNORECASE)\n header_mtime = parse_http_date(matches.group(1))\n header_len = matches.group(3)\n if header_len and int(header_len) != size:\n raise ValueError\n if int(mtime) > header_mtime:\n raise ValueError\n except (AttributeError, ValueError, OverflowError):\n return True\n return False\n", "role": "user" }, { "content": "was_modified_since", "role": "assistant" } ]
def x(self, obj): if not hasattr(obj, 'get_absolute_url') or not obj.get_absolute_url(): raise ImproperlyConfigured( "No URL configured. You must either set a ``get_absolute_url`` method on the %s model or override the %s view's ``get_url`` method" % (obj.__class__.__name__, self.__class__.__name__)) return obj.get_absolute_url()
get_url
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, obj):\n if not hasattr(obj, 'get_absolute_url') or not obj.get_absolute_url():\n raise ImproperlyConfigured(\n \"No URL configured. You must either set a ``get_absolute_url`` method on the %s model or override the %s view's ``get_url`` method\"\n % (obj.__class__.__name__, self.__class__.__name__))\n return obj.get_absolute_url()\n", "role": "user" }, { "content": "get_url", "role": "assistant" } ]
def x(self, obj): logger.debug('Unbuilding %s' % obj) target_path = os.path.split(self.get_build_path(obj))[0] if self.fs.exists(target_path): logger.debug('Removing {}'.format(target_path)) self.fs.removetree(target_path)
unbuild_object
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, obj):\n logger.debug('Unbuilding %s' % obj)\n target_path = os.path.split(self.get_build_path(obj))[0]\n if self.fs.exists(target_path):\n logger.debug('Removing {}'.format(target_path))\n self.fs.removetree(target_path)\n", "role": "user" }, { "content": "unbuild_object", "role": "assistant" } ]
def x(content_type_pk, obj_pk): ct = ContentType.objects.get_for_id(content_type_pk) obj = ct.get_object_for_this_type(pk=obj_pk) try: logger.info('unpublish_object task has received %s' % obj) obj.unbuild() if not getattr(settings, 'ALLOW_BAKERY_AUTO_PUBLISHING', True): logger.info( 'Not running publish command because ALLOW_BAKERY_AUTO_PUBLISHING is False' ) else: management.call_command('publish') except Exception: logger.error('Task Error: unpublish_object', exc_info=True)
unpublish_object
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(content_type_pk, obj_pk):\n ct = ContentType.objects.get_for_id(content_type_pk)\n obj = ct.get_object_for_this_type(pk=obj_pk)\n try:\n logger.info('unpublish_object task has received %s' % obj)\n obj.unbuild()\n if not getattr(settings, 'ALLOW_BAKERY_AUTO_PUBLISHING', True):\n logger.info(\n 'Not running publish command because ALLOW_BAKERY_AUTO_PUBLISHING is False'\n )\n else:\n management.call_command('publish')\n except Exception:\n logger.error('Task Error: unpublish_object', exc_info=True)\n", "role": "user" }, { "content": "unpublish_object", "role": "assistant" } ]
def x(self, target_dir): dirname = path.dirname(target_dir) if dirname: dirname = path.join(settings.BUILD_DIR, dirname) if not self.fs.exists(dirname): logger.debug('Creating directory at {}{}'.format(self.fs_name, dirname)) self.fs.makedirs(dirname)
prep_directory
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, target_dir):\n dirname = path.dirname(target_dir)\n if dirname:\n dirname = path.join(settings.BUILD_DIR, dirname)\n if not self.fs.exists(dirname):\n logger.debug('Creating directory at {}{}'.format(self.fs_name,\n dirname))\n self.fs.makedirs(dirname)\n", "role": "user" }, { "content": "prep_directory", "role": "assistant" } ]
def x(self, target_path, html): logger.debug('Building to {}{}'.format(self.fs_name, target_path)) with self.fs.open(smart_text(target_path), 'wb') as outfile: outfile.write(six.binary_type(html)) outfile.close()
write_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, target_path, html):\n logger.debug('Building to {}{}'.format(self.fs_name, target_path))\n with self.fs.open(smart_text(target_path), 'wb') as outfile:\n outfile.write(six.binary_type(html))\n outfile.close()\n", "role": "user" }, { "content": "write_file", "role": "assistant" } ]
def x(self, path): if not getattr(settings, 'BAKERY_GZIP', False): return False whitelist = getattr(settings, 'GZIP_CONTENT_TYPES', DEFAULT_GZIP_CONTENT_TYPES) return mimetypes.guess_type(path)[0] in whitelist
is_gzippable
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, path):\n if not getattr(settings, 'BAKERY_GZIP', False):\n return False\n whitelist = getattr(settings, 'GZIP_CONTENT_TYPES',\n DEFAULT_GZIP_CONTENT_TYPES)\n return mimetypes.guess_type(path)[0] in whitelist\n", "role": "user" }, { "content": "is_gzippable", "role": "assistant" } ]
def x(self, target_path, html): logger.debug('Gzipping to {}{}'.format(self.fs_name, target_path)) data_buffer = six.BytesIO() kwargs = dict(filename=path.basename(target_path), mode='wb', fileobj= data_buffer) if float(sys.version[:3]) >= 2.7: kwargs['mtime'] = 0 with gzip.GzipFile(**kwargs) as f: f.write(six.binary_type(html)) with self.fs.open(smart_text(target_path), 'wb') as outfile: outfile.write(data_buffer.getvalue()) outfile.close()
gzip_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, target_path, html):\n logger.debug('Gzipping to {}{}'.format(self.fs_name, target_path))\n data_buffer = six.BytesIO()\n kwargs = dict(filename=path.basename(target_path), mode='wb', fileobj=\n data_buffer)\n if float(sys.version[:3]) >= 2.7:\n kwargs['mtime'] = 0\n with gzip.GzipFile(**kwargs) as f:\n f.write(six.binary_type(html))\n with self.fs.open(smart_text(target_path), 'wb') as outfile:\n outfile.write(data_buffer.getvalue())\n outfile.close()\n", "role": "user" }, { "content": "gzip_file", "role": "assistant" } ]
def x(self, *args, **kwargs): if self.url: url = self.url % kwargs elif self.pattern_name: try: url = reverse(self.pattern_name, args=args, kwargs=kwargs) except NoReverseMatch: return None else: return None return url
get_redirect_url
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *args, **kwargs):\n if self.url:\n url = self.url % kwargs\n elif self.pattern_name:\n try:\n url = reverse(self.pattern_name, args=args, kwargs=kwargs)\n except NoReverseMatch:\n return None\n else:\n return None\n return url\n", "role": "user" }, { "content": "get_redirect_url", "role": "assistant" } ]
def x(self): for detail_view in self.detail_views: view = self._get_view(detail_view) view().build_object(self) self._build_extra() self._build_related()
build
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for detail_view in self.detail_views:\n view = self._get_view(detail_view)\n view().build_object(self)\n self._build_extra()\n self._build_related()\n", "role": "user" }, { "content": "build", "role": "assistant" } ]
def x(self): for detail_view in self.detail_views: view = self._get_view(detail_view) view().unbuild_object(self) self._unbuild_extra() self._build_related()
unbuild
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for detail_view in self.detail_views:\n view = self._get_view(detail_view)\n view().unbuild_object(self)\n self._unbuild_extra()\n self._build_related()\n", "role": "user" }, { "content": "unbuild", "role": "assistant" } ]
def x(self, *args, **kwargs): from bakery import tasks from django.contrib.contenttypes.models import ContentType if not kwargs.pop('publish', True): super(AutoPublishingBuildableModel, self).save(*args, **kwargs) else: try: preexisting = self.__class__.objects.get(pk=self.pk) except self.__class__.DoesNotExist: preexisting = None if not preexisting: if self.get_publication_status(): action = 'publish' else: action = None elif not self.get_publication_status( ) and preexisting.get_publication_status(): action = 'unpublish' elif self.get_publication_status(): action = 'publish' else: action = None with transaction.atomic(): super(AutoPublishingBuildableModel, self).save(*args, **kwargs) ct = ContentType.objects.get_for_model(self.__class__) if action == 'publish': tasks.publish_object.delay(ct.pk, self.pk) elif action == 'unpublish': tasks.unpublish_object.delay(ct.pk, self.pk)
save
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *args, **kwargs):\n from bakery import tasks\n from django.contrib.contenttypes.models import ContentType\n if not kwargs.pop('publish', True):\n super(AutoPublishingBuildableModel, self).save(*args, **kwargs)\n else:\n try:\n preexisting = self.__class__.objects.get(pk=self.pk)\n except self.__class__.DoesNotExist:\n preexisting = None\n if not preexisting:\n if self.get_publication_status():\n action = 'publish'\n else:\n action = None\n elif not self.get_publication_status(\n ) and preexisting.get_publication_status():\n action = 'unpublish'\n elif self.get_publication_status():\n action = 'publish'\n else:\n action = None\n with transaction.atomic():\n super(AutoPublishingBuildableModel, self).save(*args, **kwargs)\n ct = ContentType.objects.get_for_model(self.__class__)\n if action == 'publish':\n tasks.publish_object.delay(ct.pk, self.pk)\n elif action == 'unpublish':\n tasks.unpublish_object.delay(ct.pk, self.pk)\n", "role": "user" }, { "content": "save", "role": "assistant" } ]
def x(self, *args, **kwargs): from bakery import tasks from django.contrib.contenttypes.models import ContentType unpublish = kwargs.pop('unpublish', True) super(AutoPublishingBuildableModel, self).delete(*args, **kwargs) if unpublish: ct = ContentType.objects.get_for_model(self.__class__) tasks.unpublish_object.delay(ct.pk, self.pk)
delete
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *args, **kwargs):\n from bakery import tasks\n from django.contrib.contenttypes.models import ContentType\n unpublish = kwargs.pop('unpublish', True)\n super(AutoPublishingBuildableModel, self).delete(*args, **kwargs)\n if unpublish:\n ct = ContentType.objects.get_for_model(self.__class__)\n tasks.unpublish_object.delay(ct.pk, self.pk)\n", "role": "user" }, { "content": "delete", "role": "assistant" } ]
def x(self, *args, **options): logger.info('Build started') self.set_options(*args, **options) if not options.get('keep_build_dir'): self.init_build_dir() if not options.get('skip_static'): self.build_static() if not options.get('skip_media'): self.build_media() self.build_views() logger.info('Build finished')
handle
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *args, **options):\n logger.info('Build started')\n self.set_options(*args, **options)\n if not options.get('keep_build_dir'):\n self.init_build_dir()\n if not options.get('skip_static'):\n self.build_static()\n if not options.get('skip_media'):\n self.build_media()\n self.build_views()\n logger.info('Build finished')\n", "role": "user" }, { "content": "handle", "role": "assistant" } ]
def x(self, *args, **options): self.verbosity = int(options.get('verbosity', 1)) if options.get('build_dir'): self.build_dir = options.get('build_dir') settings.BUILD_DIR = self.build_dir else: if not hasattr(settings, 'BUILD_DIR'): raise CommandError(self.build_unconfig_msg) self.build_dir = settings.BUILD_DIR self.build_dir = smart_text(self.build_dir) self.static_root = smart_text(settings.STATIC_ROOT) self.media_root = smart_text(settings.MEDIA_ROOT) self.app = apps.get_app_config('bakery') self.fs = self.app.filesystem self.fs_name = self.app.filesystem_name if not self.fs.exists(self.build_dir): self.fs.makedirs(self.build_dir) if options.get('view_list'): self.view_list = options['view_list'] else: if not hasattr(settings, 'BAKERY_VIEWS'): raise CommandError(self.views_unconfig_msg) self.view_list = settings.BAKERY_VIEWS self.pooling = options.get('pooling')
set_options
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *args, **options):\n self.verbosity = int(options.get('verbosity', 1))\n if options.get('build_dir'):\n self.build_dir = options.get('build_dir')\n settings.BUILD_DIR = self.build_dir\n else:\n if not hasattr(settings, 'BUILD_DIR'):\n raise CommandError(self.build_unconfig_msg)\n self.build_dir = settings.BUILD_DIR\n self.build_dir = smart_text(self.build_dir)\n self.static_root = smart_text(settings.STATIC_ROOT)\n self.media_root = smart_text(settings.MEDIA_ROOT)\n self.app = apps.get_app_config('bakery')\n self.fs = self.app.filesystem\n self.fs_name = self.app.filesystem_name\n if not self.fs.exists(self.build_dir):\n self.fs.makedirs(self.build_dir)\n if options.get('view_list'):\n self.view_list = options['view_list']\n else:\n if not hasattr(settings, 'BAKERY_VIEWS'):\n raise CommandError(self.views_unconfig_msg)\n self.view_list = settings.BAKERY_VIEWS\n self.pooling = options.get('pooling')\n", "role": "user" }, { "content": "set_options", "role": "assistant" } ]
def x(self): logger.debug('Initializing %s' % self.build_dir) if self.verbosity > 1: self.stdout.write('Initializing build directory') if self.fs.exists(self.build_dir): self.fs.removetree(self.build_dir) self.fs.makedirs(self.build_dir)
init_build_dir
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n logger.debug('Initializing %s' % self.build_dir)\n if self.verbosity > 1:\n self.stdout.write('Initializing build directory')\n if self.fs.exists(self.build_dir):\n self.fs.removetree(self.build_dir)\n self.fs.makedirs(self.build_dir)\n", "role": "user" }, { "content": "init_build_dir", "role": "assistant" } ]
def x(self, *args, **options): logger.debug('Building static directory') if self.verbosity > 1: self.stdout.write('Building static directory') management.call_command('collectstatic', interactive=False, verbosity=0) target_dir = path.join(self.build_dir, settings.STATIC_URL.lstrip('/')) target_dir = smart_text(target_dir) if os.path.exists(self.static_root) and settings.STATIC_URL: if getattr(settings, 'BAKERY_GZIP', False): self.copytree_and_gzip(self.static_root, target_dir) else: logger.debug('Copying {}{} to {}{}'.format('osfs://', self. static_root, self.fs_name, target_dir)) copy.copy_dir('osfs:///', self.static_root, self.fs, target_dir) robots_src = path.join(target_dir, 'robots.txt') if self.fs.exists(robots_src): robots_target = path.join(self.build_dir, 'robots.txt') logger.debug('Copying {}{} to {}{}'.format(self.fs_name, robots_src, self.fs_name, robots_target)) self.fs.copy(robots_src, robots_target) favicon_src = path.join(target_dir, 'favicon.ico') if self.fs.exists(favicon_src): favicon_target = path.join(self.build_dir, 'favicon.ico') logger.debug('Copying {}{} to {}{}'.format(self.fs_name, favicon_src, self.fs_name, favicon_target)) self.fs.copy(favicon_src, favicon_target)
build_static
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *args, **options):\n logger.debug('Building static directory')\n if self.verbosity > 1:\n self.stdout.write('Building static directory')\n management.call_command('collectstatic', interactive=False, verbosity=0)\n target_dir = path.join(self.build_dir, settings.STATIC_URL.lstrip('/'))\n target_dir = smart_text(target_dir)\n if os.path.exists(self.static_root) and settings.STATIC_URL:\n if getattr(settings, 'BAKERY_GZIP', False):\n self.copytree_and_gzip(self.static_root, target_dir)\n else:\n logger.debug('Copying {}{} to {}{}'.format('osfs://', self.\n static_root, self.fs_name, target_dir))\n copy.copy_dir('osfs:///', self.static_root, self.fs, target_dir)\n robots_src = path.join(target_dir, 'robots.txt')\n if self.fs.exists(robots_src):\n robots_target = path.join(self.build_dir, 'robots.txt')\n logger.debug('Copying {}{} to {}{}'.format(self.fs_name, robots_src,\n self.fs_name, robots_target))\n self.fs.copy(robots_src, robots_target)\n favicon_src = path.join(target_dir, 'favicon.ico')\n if self.fs.exists(favicon_src):\n favicon_target = path.join(self.build_dir, 'favicon.ico')\n logger.debug('Copying {}{} to {}{}'.format(self.fs_name,\n favicon_src, self.fs_name, favicon_target))\n self.fs.copy(favicon_src, favicon_target)\n", "role": "user" }, { "content": "build_static", "role": "assistant" } ]
def x(self): logger.debug('Building media directory') if self.verbosity > 1: self.stdout.write('Building media directory') if os.path.exists(self.media_root) and settings.MEDIA_URL: target_dir = path.join(self.fs_name, self.build_dir, settings. MEDIA_URL.lstrip('/')) logger.debug('Copying {}{} to {}{}'.format('osfs://', self. media_root, self.fs_name, target_dir)) copy.copy_dir('osfs:///', smart_text(self.media_root), self.fs, smart_text(target_dir))
build_media
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n logger.debug('Building media directory')\n if self.verbosity > 1:\n self.stdout.write('Building media directory')\n if os.path.exists(self.media_root) and settings.MEDIA_URL:\n target_dir = path.join(self.fs_name, self.build_dir, settings.\n MEDIA_URL.lstrip('/'))\n logger.debug('Copying {}{} to {}{}'.format('osfs://', self.\n media_root, self.fs_name, target_dir))\n copy.copy_dir('osfs:///', smart_text(self.media_root), self.fs,\n smart_text(target_dir))\n", "role": "user" }, { "content": "build_media", "role": "assistant" } ]
def x(self): for view_str in self.view_list: logger.debug('Building %s' % view_str) if self.verbosity > 1: self.stdout.write('Building %s' % view_str) view = get_callable(view_str) self.get_view_instance(view).build_method()
build_views
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for view_str in self.view_list:\n logger.debug('Building %s' % view_str)\n if self.verbosity > 1:\n self.stdout.write('Building %s' % view_str)\n view = get_callable(view_str)\n self.get_view_instance(view).build_method()\n", "role": "user" }, { "content": "build_views", "role": "assistant" } ]
def x(self, source_dir, target_dir): build_list = [] for dirpath, dirnames, filenames in os.walk(source_dir): for f in filenames: source_path = os.path.join(dirpath, f) rel_path = os.path.relpath(dirpath, source_dir) target_path = os.path.join(target_dir, rel_path, f) build_list.append((source_path, target_path)) logger.debug('Gzipping {} files'.format(len(build_list))) if not getattr(self, 'pooling', False): [self.copyfile_and_gzip(*u) for u in build_list] else: cpu_count = multiprocessing.cpu_count() logger.debug('Pooling build on {} CPUs'.format(cpu_count)) pool = ThreadPool(processes=cpu_count) pool.map(self.pooled_copyfile_and_gzip, build_list)
copytree_and_gzip
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, source_dir, target_dir):\n build_list = []\n for dirpath, dirnames, filenames in os.walk(source_dir):\n for f in filenames:\n source_path = os.path.join(dirpath, f)\n rel_path = os.path.relpath(dirpath, source_dir)\n target_path = os.path.join(target_dir, rel_path, f)\n build_list.append((source_path, target_path))\n logger.debug('Gzipping {} files'.format(len(build_list)))\n if not getattr(self, 'pooling', False):\n [self.copyfile_and_gzip(*u) for u in build_list]\n else:\n cpu_count = multiprocessing.cpu_count()\n logger.debug('Pooling build on {} CPUs'.format(cpu_count))\n pool = ThreadPool(processes=cpu_count)\n pool.map(self.pooled_copyfile_and_gzip, build_list)\n", "role": "user" }, { "content": "copytree_and_gzip", "role": "assistant" } ]
def x(self, source_path, target_path): target_dir = path.dirname(target_path) if not self.fs.exists(target_dir): try: self.fs.makedirs(target_dir) except OSError: pass guess = mimetypes.guess_type(source_path) content_type = guess[0] encoding = guess[1] if content_type not in self.gzip_file_match: logger.debug( "Copying {}{} to {}{} because its filetype isn't on the whitelist" .format('osfs://', source_path, self.fs_name, target_path)) copy.copy_file('osfs:///', smart_text(source_path), self.fs, smart_text(target_path)) elif encoding == 'gzip': logger.debug("Copying {}{} to {}{} because it's already gzipped". format('osfs://', source_path, self.fs_name, target_path)) copy.copy_file('osfs:///', smart_text(source_path), self.fs, smart_text(target_path)) else: logger.debug('Gzipping {}{} to {}{}'.format('osfs://', source_path, self.fs_name, target_path)) with open(source_path, 'rb') as source_file: data_buffer = six.BytesIO() kwargs = dict(filename=path.basename(target_path), mode='wb', fileobj=data_buffer) if float(sys.version[:3]) >= 2.7: kwargs['mtime'] = 0 with gzip.GzipFile(**kwargs) as f: f.write(six.binary_type(source_file.read())) with self.fs.open(smart_text(target_path), 'wb') as outfile: outfile.write(data_buffer.getvalue()) outfile.close()
copyfile_and_gzip
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, source_path, target_path):\n target_dir = path.dirname(target_path)\n if not self.fs.exists(target_dir):\n try:\n self.fs.makedirs(target_dir)\n except OSError:\n pass\n guess = mimetypes.guess_type(source_path)\n content_type = guess[0]\n encoding = guess[1]\n if content_type not in self.gzip_file_match:\n logger.debug(\n \"Copying {}{} to {}{} because its filetype isn't on the whitelist\"\n .format('osfs://', source_path, self.fs_name, target_path))\n copy.copy_file('osfs:///', smart_text(source_path), self.fs,\n smart_text(target_path))\n elif encoding == 'gzip':\n logger.debug(\"Copying {}{} to {}{} because it's already gzipped\".\n format('osfs://', source_path, self.fs_name, target_path))\n copy.copy_file('osfs:///', smart_text(source_path), self.fs,\n smart_text(target_path))\n else:\n logger.debug('Gzipping {}{} to {}{}'.format('osfs://', source_path,\n self.fs_name, target_path))\n with open(source_path, 'rb') as source_file:\n data_buffer = six.BytesIO()\n kwargs = dict(filename=path.basename(target_path), mode='wb',\n fileobj=data_buffer)\n if float(sys.version[:3]) >= 2.7:\n kwargs['mtime'] = 0\n with gzip.GzipFile(**kwargs) as f:\n f.write(six.binary_type(source_file.read()))\n with self.fs.open(smart_text(target_path), 'wb') as outfile:\n outfile.write(data_buffer.getvalue())\n outfile.close()\n", "role": "user" }, { "content": "copyfile_and_gzip", "role": "assistant" } ]
def x(self, *args, **options): self.uploaded_files = 0 self.uploaded_file_list = [] self.deleted_files = 0 self.deleted_file_list = [] self.start_time = time.time() self.set_options(options) logger.debug('Connecting to s3') if self.verbosity > 2: self.stdout.write('Connecting to s3') self.s3_client, self.s3_resource = get_s3_client() logger.debug('Retriving bucket {}'.format(self.aws_bucket_name)) if self.verbosity > 2: self.stdout.write('Retriving bucket {}'.format(self.aws_bucket_name)) self.bucket = self.s3_resource.Bucket(self.aws_bucket_name) if self.force_publish and self.no_delete: self.blind_upload = True logger.debug( "Skipping object retrieval. We won't need to because we're blinding uploading everything." ) self.s3_obj_dict = {} else: self.blind_upload = False logger.debug('Retrieving objects now published in bucket') if self.verbosity > 2: self.stdout.write('Retrieving objects now published in bucket') self.s3_obj_dict = {} self.s3_obj_dict = self.get_bucket_file_list() logger.debug('Retrieving files built locally') if self.verbosity > 2: self.stdout.write('Retrieving files built locally') self.local_file_list = self.get_local_file_list() logger.debug('Syncing local files with bucket') if self.verbosity > 2: self.stdout.write('Syncing local files with bucket') self.sync_with_s3() if not self.dry_run and not self.no_delete: self.deleted_file_list = list(self.s3_obj_dict.keys()) self.deleted_files = len(self.deleted_file_list) if self.deleted_files: logger.debug('Deleting %s keys' % self.deleted_files) if self.verbosity > 0: self.stdout.write('Deleting %s keys' % self.deleted_files) self.batch_delete_s3_objects(self.deleted_file_list, self. aws_bucket_name) if not hasattr(settings, 'BAKERY_VIEWS'): raise CommandError(self.views_unconfig_msg) for view_str in settings.BAKERY_VIEWS: view = get_callable(view_str)() if hasattr(view, 'post_publish'): getattr(view, 'post_publish')(self.bucket) elapsed_time = time.time() - self.start_time msg = ( 'Publish completed, %d uploaded and %d deleted files in %.2f seconds' % (self.uploaded_files, self.deleted_files, elapsed_time)) logger.info(msg) if self.verbosity > 0: self.stdout.write(msg) if self.dry_run: logger.info( 'Publish executed with the --dry-run option. No content was changed on S3.' ) if self.verbosity > 0: self.stdout.write( 'Publish executed with the --dry-run option. No content was changed on S3.' )
handle
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *args, **options):\n self.uploaded_files = 0\n self.uploaded_file_list = []\n self.deleted_files = 0\n self.deleted_file_list = []\n self.start_time = time.time()\n self.set_options(options)\n logger.debug('Connecting to s3')\n if self.verbosity > 2:\n self.stdout.write('Connecting to s3')\n self.s3_client, self.s3_resource = get_s3_client()\n logger.debug('Retriving bucket {}'.format(self.aws_bucket_name))\n if self.verbosity > 2:\n self.stdout.write('Retriving bucket {}'.format(self.aws_bucket_name))\n self.bucket = self.s3_resource.Bucket(self.aws_bucket_name)\n if self.force_publish and self.no_delete:\n self.blind_upload = True\n logger.debug(\n \"Skipping object retrieval. We won't need to because we're blinding uploading everything.\"\n )\n self.s3_obj_dict = {}\n else:\n self.blind_upload = False\n logger.debug('Retrieving objects now published in bucket')\n if self.verbosity > 2:\n self.stdout.write('Retrieving objects now published in bucket')\n self.s3_obj_dict = {}\n self.s3_obj_dict = self.get_bucket_file_list()\n logger.debug('Retrieving files built locally')\n if self.verbosity > 2:\n self.stdout.write('Retrieving files built locally')\n self.local_file_list = self.get_local_file_list()\n logger.debug('Syncing local files with bucket')\n if self.verbosity > 2:\n self.stdout.write('Syncing local files with bucket')\n self.sync_with_s3()\n if not self.dry_run and not self.no_delete:\n self.deleted_file_list = list(self.s3_obj_dict.keys())\n self.deleted_files = len(self.deleted_file_list)\n if self.deleted_files:\n logger.debug('Deleting %s keys' % self.deleted_files)\n if self.verbosity > 0:\n self.stdout.write('Deleting %s keys' % self.deleted_files)\n self.batch_delete_s3_objects(self.deleted_file_list, self.\n aws_bucket_name)\n if not hasattr(settings, 'BAKERY_VIEWS'):\n raise CommandError(self.views_unconfig_msg)\n for view_str in settings.BAKERY_VIEWS:\n view = get_callable(view_str)()\n if hasattr(view, 'post_publish'):\n getattr(view, 'post_publish')(self.bucket)\n elapsed_time = time.time() - self.start_time\n msg = (\n 'Publish completed, %d uploaded and %d deleted files in %.2f seconds' %\n (self.uploaded_files, self.deleted_files, elapsed_time))\n logger.info(msg)\n if self.verbosity > 0:\n self.stdout.write(msg)\n if self.dry_run:\n logger.info(\n 'Publish executed with the --dry-run option. No content was changed on S3.'\n )\n if self.verbosity > 0:\n self.stdout.write(\n 'Publish executed with the --dry-run option. No content was changed on S3.'\n )\n", "role": "user" }, { "content": "handle", "role": "assistant" } ]
def x(self, options): self.verbosity = int(options.get('verbosity')) self.gzip = getattr(settings, 'BAKERY_GZIP', False) self.gzip_content_types = getattr(settings, 'GZIP_CONTENT_TYPES', DEFAULT_GZIP_CONTENT_TYPES) self.acl = getattr(settings, 'DEFAULT_ACL', self.DEFAULT_ACL) self.cache_control = getattr(settings, 'BAKERY_CACHE_CONTROL', {}) if options.get('build_dir'): if not os.path.exists(options.get('build_dir')): raise CommandError(self.build_missing_msg) self.build_dir = options.get('build_dir') else: if not hasattr(settings, 'BUILD_DIR'): raise CommandError(self.build_unconfig_msg) if not os.path.exists(settings.BUILD_DIR): raise CommandError(self.build_missing_msg) self.build_dir = settings.BUILD_DIR if options.get('aws_bucket_name'): self.aws_bucket_name = options.get('aws_bucket_name') else: if not hasattr(settings, 'AWS_BUCKET_NAME'): raise CommandError(self.bucket_unconfig_msg) self.aws_bucket_name = settings.AWS_BUCKET_NAME self.aws_bucket_prefix = options.get('aws_bucket_prefix') if options.get('force'): self.force_publish = True else: self.force_publish = False if options.get('dry_run'): self.dry_run = True if self.verbosity > 0: logger.info('Executing with the --dry-run option set.') else: self.dry_run = False self.no_delete = options.get('no_delete') self.no_pooling = options.get('no_pooling')
set_options
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, options):\n self.verbosity = int(options.get('verbosity'))\n self.gzip = getattr(settings, 'BAKERY_GZIP', False)\n self.gzip_content_types = getattr(settings, 'GZIP_CONTENT_TYPES',\n DEFAULT_GZIP_CONTENT_TYPES)\n self.acl = getattr(settings, 'DEFAULT_ACL', self.DEFAULT_ACL)\n self.cache_control = getattr(settings, 'BAKERY_CACHE_CONTROL', {})\n if options.get('build_dir'):\n if not os.path.exists(options.get('build_dir')):\n raise CommandError(self.build_missing_msg)\n self.build_dir = options.get('build_dir')\n else:\n if not hasattr(settings, 'BUILD_DIR'):\n raise CommandError(self.build_unconfig_msg)\n if not os.path.exists(settings.BUILD_DIR):\n raise CommandError(self.build_missing_msg)\n self.build_dir = settings.BUILD_DIR\n if options.get('aws_bucket_name'):\n self.aws_bucket_name = options.get('aws_bucket_name')\n else:\n if not hasattr(settings, 'AWS_BUCKET_NAME'):\n raise CommandError(self.bucket_unconfig_msg)\n self.aws_bucket_name = settings.AWS_BUCKET_NAME\n self.aws_bucket_prefix = options.get('aws_bucket_prefix')\n if options.get('force'):\n self.force_publish = True\n else:\n self.force_publish = False\n if options.get('dry_run'):\n self.dry_run = True\n if self.verbosity > 0:\n logger.info('Executing with the --dry-run option set.')\n else:\n self.dry_run = False\n self.no_delete = options.get('no_delete')\n self.no_pooling = options.get('no_pooling')\n", "role": "user" }, { "content": "set_options", "role": "assistant" } ]
def x(self): logger.debug('Retrieving bucket object list') paginator = self.s3_client.get_paginator('list_objects') options = {'Bucket': self.aws_bucket_name} if self.aws_bucket_prefix: logger.debug('Adding prefix {} to bucket list as a filter'.format( self.aws_bucket_prefix)) options['Prefix'] = self.aws_bucket_prefix page_iterator = paginator.paginate(**options) obj_dict = {} for page in page_iterator: obj_dict.update(get_bucket_page(page)) return obj_dict
get_bucket_file_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n logger.debug('Retrieving bucket object list')\n paginator = self.s3_client.get_paginator('list_objects')\n options = {'Bucket': self.aws_bucket_name}\n if self.aws_bucket_prefix:\n logger.debug('Adding prefix {} to bucket list as a filter'.format(\n self.aws_bucket_prefix))\n options['Prefix'] = self.aws_bucket_prefix\n page_iterator = paginator.paginate(**options)\n obj_dict = {}\n for page in page_iterator:\n obj_dict.update(get_bucket_page(page))\n return obj_dict\n", "role": "user" }, { "content": "get_bucket_file_list", "role": "assistant" } ]
def x(self): file_list = [] for dirpath, dirnames, filenames in os.walk(self.build_dir): for fname in filenames: local_key = os.path.join(os.path.relpath(dirpath, self. build_dir), fname) if local_key.startswith('./'): local_key = local_key[2:] file_list.append(local_key) return file_list
get_local_file_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n file_list = []\n for dirpath, dirnames, filenames in os.walk(self.build_dir):\n for fname in filenames:\n local_key = os.path.join(os.path.relpath(dirpath, self.\n build_dir), fname)\n if local_key.startswith('./'):\n local_key = local_key[2:]\n file_list.append(local_key)\n return file_list\n", "role": "user" }, { "content": "get_local_file_list", "role": "assistant" } ]
def x(self): self.update_list = [] logger.debug('Comparing {} local files with {} bucket files'.format(len (self.local_file_list), len(self.s3_obj_dict.keys()))) if self.no_pooling: [self.compare_local_file(f) for f in self.local_file_list] else: cpu_count = multiprocessing.cpu_count() logger.debug('Pooling local file comparison on {} CPUs'.format( cpu_count)) pool = ThreadPool(processes=cpu_count) pool.map(self.compare_local_file, self.local_file_list) logger.debug('Uploading {} new or updated files to bucket'.format(len( self.update_list))) if self.no_pooling: [self.upload_to_s3(*u) for u in self.update_list] else: logger.debug('Pooling s3 uploads on {} CPUs'.format(cpu_count)) pool = ThreadPool(processes=cpu_count) pool.map(self.pooled_upload_to_s3, self.update_list)
sync_with_s3
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.update_list = []\n logger.debug('Comparing {} local files with {} bucket files'.format(len\n (self.local_file_list), len(self.s3_obj_dict.keys())))\n if self.no_pooling:\n [self.compare_local_file(f) for f in self.local_file_list]\n else:\n cpu_count = multiprocessing.cpu_count()\n logger.debug('Pooling local file comparison on {} CPUs'.format(\n cpu_count))\n pool = ThreadPool(processes=cpu_count)\n pool.map(self.compare_local_file, self.local_file_list)\n logger.debug('Uploading {} new or updated files to bucket'.format(len(\n self.update_list)))\n if self.no_pooling:\n [self.upload_to_s3(*u) for u in self.update_list]\n else:\n logger.debug('Pooling s3 uploads on {} CPUs'.format(cpu_count))\n pool = ThreadPool(processes=cpu_count)\n pool.map(self.pooled_upload_to_s3, self.update_list)\n", "role": "user" }, { "content": "sync_with_s3", "role": "assistant" } ]
def x(self, filename): with open(filename, 'rb') as f: m = hashlib.md5(f.read()) return m.hexdigest()
get_md5
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, filename):\n with open(filename, 'rb') as f:\n m = hashlib.md5(f.read())\n return m.hexdigest()\n", "role": "user" }, { "content": "get_md5", "role": "assistant" } ]
def x(self, filename, chunk_size=8 * 1024 * 1024): md5s = [] with open(filename, 'rb') as fp: while True: data = fp.read(chunk_size) if not data: break md5s.append(hashlib.md5(data)) digests = b''.join(m.digest() for m in md5s) new_md5 = hashlib.md5(digests) new_etag = '"%s-%s"' % (new_md5.hexdigest(), len(md5s)) return new_etag.strip('"').strip("'")
get_multipart_md5
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, filename, chunk_size=8 * 1024 * 1024):\n md5s = []\n with open(filename, 'rb') as fp:\n while True:\n data = fp.read(chunk_size)\n if not data:\n break\n md5s.append(hashlib.md5(data))\n digests = b''.join(m.digest() for m in md5s)\n new_md5 = hashlib.md5(digests)\n new_etag = '\"%s-%s\"' % (new_md5.hexdigest(), len(md5s))\n return new_etag.strip('\"').strip(\"'\")\n", "role": "user" }, { "content": "get_multipart_md5", "role": "assistant" } ]
def x(self, file_key): file_path = os.path.join(self.build_dir, file_key) if self.force_publish: self.update_list.append((file_key, file_path)) return if file_key in self.s3_obj_dict: s3_md5 = self.s3_obj_dict[file_key].get('ETag').strip('"').strip("'") if '-' in s3_md5: local_md5 = self.get_multipart_md5(file_path) else: local_md5 = self.get_md5(file_path) if s3_md5 == local_md5: pass else: logger.debug('{} has changed'.format(file_key)) self.update_list.append((file_key, file_path)) del self.s3_obj_dict[file_key] else: logger.debug('{} has been added'.format(file_key)) self.update_list.append((file_key, file_path))
compare_local_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, file_key):\n file_path = os.path.join(self.build_dir, file_key)\n if self.force_publish:\n self.update_list.append((file_key, file_path))\n return\n if file_key in self.s3_obj_dict:\n s3_md5 = self.s3_obj_dict[file_key].get('ETag').strip('\"').strip(\"'\")\n if '-' in s3_md5:\n local_md5 = self.get_multipart_md5(file_path)\n else:\n local_md5 = self.get_md5(file_path)\n if s3_md5 == local_md5:\n pass\n else:\n logger.debug('{} has changed'.format(file_key))\n self.update_list.append((file_key, file_path))\n del self.s3_obj_dict[file_key]\n else:\n logger.debug('{} has been added'.format(file_key))\n self.update_list.append((file_key, file_path))\n", "role": "user" }, { "content": "compare_local_file", "role": "assistant" } ]
def x(self, key, filename): extra_args = {'ACL': self.acl} guess = mimetypes.guess_type(filename) content_type = guess[0] encoding = guess[1] if content_type: extra_args['ContentType'] = content_type if (self.gzip and content_type in self.gzip_content_types or encoding == 'gzip'): extra_args['ContentEncoding'] = 'gzip' if content_type in self.cache_control: extra_args['CacheControl'] = ''.join(('max-age=', str(self. cache_control[content_type]))) if not self.dry_run: logger.debug('Uploading %s' % filename) if self.verbosity > 0: self.stdout.write('Uploading %s' % filename) s3_obj = self.s3_resource.Object(self.aws_bucket_name, key) s3_obj.upload_file(filename, ExtraArgs=extra_args) self.uploaded_files += 1 self.uploaded_file_list.append(filename)
upload_to_s3
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, filename):\n extra_args = {'ACL': self.acl}\n guess = mimetypes.guess_type(filename)\n content_type = guess[0]\n encoding = guess[1]\n if content_type:\n extra_args['ContentType'] = content_type\n if (self.gzip and content_type in self.gzip_content_types or encoding ==\n 'gzip'):\n extra_args['ContentEncoding'] = 'gzip'\n if content_type in self.cache_control:\n extra_args['CacheControl'] = ''.join(('max-age=', str(self.\n cache_control[content_type])))\n if not self.dry_run:\n logger.debug('Uploading %s' % filename)\n if self.verbosity > 0:\n self.stdout.write('Uploading %s' % filename)\n s3_obj = self.s3_resource.Object(self.aws_bucket_name, key)\n s3_obj.upload_file(filename, ExtraArgs=extra_args)\n self.uploaded_files += 1\n self.uploaded_file_list.append(filename)\n", "role": "user" }, { "content": "upload_to_s3", "role": "assistant" } ]
def x(self, attname, obj, args=None, default=None): try: attr = getattr(self, attname) except AttributeError: return default if callable(attr) or args: args = args[:] if args else [] try: code = six.get_function_code(attr) except AttributeError: code = six.get_function_code(attr.__call__) if code.co_argcount == 2 + len(args): args.append(obj) return attr(*args) return attr
get_bakery_dynamic_attr
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, attname, obj, args=None, default=None):\n try:\n attr = getattr(self, attname)\n except AttributeError:\n return default\n if callable(attr) or args:\n args = args[:] if args else []\n try:\n code = six.get_function_code(attr)\n except AttributeError:\n code = six.get_function_code(attr.__call__)\n if code.co_argcount == 2 + len(args):\n args.append(obj)\n return attr(*args)\n return attr\n", "role": "user" }, { "content": "get_bakery_dynamic_attr", "role": "assistant" } ]
def x(self): year = super(BuildableYearArchiveView, self).get_year() fmt = self.get_year_format() return date(int(year), 1, 1).strftime(fmt)
get_year
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n year = super(BuildableYearArchiveView, self).get_year()\n fmt = self.get_year_format()\n return date(int(year), 1, 1).strftime(fmt)\n", "role": "user" }, { "content": "get_year", "role": "assistant" } ]
def x(self, dt): self.year = str(dt.year) logger.debug('Building %s' % self.year) self.request = self.create_request(self.get_url()) target_path = self.get_build_path() self.build_file(target_path, self.get_content())
build_year
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dt):\n self.year = str(dt.year)\n logger.debug('Building %s' % self.year)\n self.request = self.create_request(self.get_url())\n target_path = self.get_build_path()\n self.build_file(target_path, self.get_content())\n", "role": "user" }, { "content": "build_year", "role": "assistant" } ]
def x(self): qs = self.get_dated_queryset() years = self.get_date_list(qs) [self.build_year(dt) for dt in years]
build_dated_queryset
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n qs = self.get_dated_queryset()\n years = self.get_date_list(qs)\n [self.build_year(dt) for dt in years]\n", "role": "user" }, { "content": "build_dated_queryset", "role": "assistant" } ]
def x(self): year = super(BuildableMonthArchiveView, self).get_year() month = super(BuildableMonthArchiveView, self).get_month() fmt = self.get_month_format() return date(int(year), int(month), 1).strftime(fmt)
get_month
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n year = super(BuildableMonthArchiveView, self).get_year()\n month = super(BuildableMonthArchiveView, self).get_month()\n fmt = self.get_month_format()\n return date(int(year), int(month), 1).strftime(fmt)\n", "role": "user" }, { "content": "get_month", "role": "assistant" } ]
def x(self): qs = self.get_dated_queryset() months = self.get_date_list(qs) [self.build_month(dt) for dt in months]
build_dated_queryset
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n qs = self.get_dated_queryset()\n months = self.get_date_list(qs)\n [self.build_month(dt) for dt in months]\n", "role": "user" }, { "content": "build_dated_queryset", "role": "assistant" } ]
def x(self, dt): self.year = str(dt.year) self.month = str(dt.month) logger.debug('Building %s-%s' % (self.year, self.month)) target_path = os.path.split(self.get_build_path())[0] if self.fs.exists(target_path): logger.debug('Removing {}'.format(target_path)) self.fs.removetree(target_path)
unbuild_month
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dt):\n self.year = str(dt.year)\n self.month = str(dt.month)\n logger.debug('Building %s-%s' % (self.year, self.month))\n target_path = os.path.split(self.get_build_path())[0]\n if self.fs.exists(target_path):\n logger.debug('Removing {}'.format(target_path))\n self.fs.removetree(target_path)\n", "role": "user" }, { "content": "unbuild_month", "role": "assistant" } ]
def x(self): year = super(BuildableDayArchiveView, self).get_year() fmt = self.get_year_format() dt = date(int(year), 1, 1) return dt.strftime(fmt)
get_year
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n year = super(BuildableDayArchiveView, self).get_year()\n fmt = self.get_year_format()\n dt = date(int(year), 1, 1)\n return dt.strftime(fmt)\n", "role": "user" }, { "content": "get_year", "role": "assistant" } ]
def x(self): year = super(BuildableDayArchiveView, self).get_year() month = super(BuildableDayArchiveView, self).get_month() fmt = self.get_month_format() dt = date(int(year), int(month), 1) return dt.strftime(fmt)
get_month
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n year = super(BuildableDayArchiveView, self).get_year()\n month = super(BuildableDayArchiveView, self).get_month()\n fmt = self.get_month_format()\n dt = date(int(year), int(month), 1)\n return dt.strftime(fmt)\n", "role": "user" }, { "content": "get_month", "role": "assistant" } ]
def x(self): year = super(BuildableDayArchiveView, self).get_year() month = super(BuildableDayArchiveView, self).get_month() day = super(BuildableDayArchiveView, self).get_day() fmt = self.get_day_format() dt = date(int(year), int(month), int(day)) return dt.strftime(fmt)
get_day
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n year = super(BuildableDayArchiveView, self).get_year()\n month = super(BuildableDayArchiveView, self).get_month()\n day = super(BuildableDayArchiveView, self).get_day()\n fmt = self.get_day_format()\n dt = date(int(year), int(month), int(day))\n return dt.strftime(fmt)\n", "role": "user" }, { "content": "get_day", "role": "assistant" } ]
def x(self): return os.path.join('/archive', self.get_year(), self.get_month(), self .get_day())
get_url
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return os.path.join('/archive', self.get_year(), self.get_month(), self\n .get_day())\n", "role": "user" }, { "content": "get_url", "role": "assistant" } ]
def x(self): target_path = path.join(settings.BUILD_DIR, self.get_url().lstrip('/')) if not self.fs.exists(target_path): logger.debug('Creating {}'.format(target_path)) self.fs.makedirs(target_path) return os.path.join(target_path, 'index.html')
get_build_path
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n target_path = path.join(settings.BUILD_DIR, self.get_url().lstrip('/'))\n if not self.fs.exists(target_path):\n logger.debug('Creating {}'.format(target_path))\n self.fs.makedirs(target_path)\n return os.path.join(target_path, 'index.html')\n", "role": "user" }, { "content": "get_build_path", "role": "assistant" } ]
def x(self, dt): self.month = str(dt.month) self.year = str(dt.year) self.day = str(dt.day) logger.debug('Building %s-%s-%s' % (self.year, self.month, self.day)) self.request = self.create_request(self.get_url()) path = self.get_build_path() self.build_file(path, self.get_content())
build_day
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dt):\n self.month = str(dt.month)\n self.year = str(dt.year)\n self.day = str(dt.day)\n logger.debug('Building %s-%s-%s' % (self.year, self.month, self.day))\n self.request = self.create_request(self.get_url())\n path = self.get_build_path()\n self.build_file(path, self.get_content())\n", "role": "user" }, { "content": "build_day", "role": "assistant" } ]
def x(self): qs = self.get_dated_queryset() days = self.get_date_list(qs, date_type='day') [self.build_day(dt) for dt in days]
build_dated_queryset
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n qs = self.get_dated_queryset()\n days = self.get_date_list(qs, date_type='day')\n [self.build_day(dt) for dt in days]\n", "role": "user" }, { "content": "build_dated_queryset", "role": "assistant" } ]
def x(): session_kwargs = {} if hasattr(settings, 'AWS_ACCESS_KEY_ID'): session_kwargs['aws_access_key_id'] = settings.AWS_ACCESS_KEY_ID if hasattr(settings, 'AWS_SECRET_ACCESS_KEY'): session_kwargs['aws_secret_access_key' ] = settings.AWS_SECRET_ACCESS_KEY boto3.setup_default_session(**session_kwargs) s3_kwargs = {} if hasattr(settings, 'AWS_S3_ENDPOINT'): s3_kwargs['endpoint_url'] = settings.AWS_S3_ENDPOINT elif hasattr(settings, 'AWS_S3_HOST'): if hasattr(settings, 'AWS_S3_USE_SSL' ) and settings.AWS_S3_USE_SSL is False: protocol = 'http://' else: protocol = 'https://' s3_kwargs['endpoint_url'] = '{}{}'.format(protocol, settings. AWS_S3_HOST) if hasattr(settings, 'AWS_REGION'): s3_kwargs['region_name'] = settings.AWS_REGION s3_client = boto3.client('s3', **s3_kwargs) s3_resource = boto3.resource('s3', **s3_kwargs) return s3_client, s3_resource
get_s3_client
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n session_kwargs = {}\n if hasattr(settings, 'AWS_ACCESS_KEY_ID'):\n session_kwargs['aws_access_key_id'] = settings.AWS_ACCESS_KEY_ID\n if hasattr(settings, 'AWS_SECRET_ACCESS_KEY'):\n session_kwargs['aws_secret_access_key'\n ] = settings.AWS_SECRET_ACCESS_KEY\n boto3.setup_default_session(**session_kwargs)\n s3_kwargs = {}\n if hasattr(settings, 'AWS_S3_ENDPOINT'):\n s3_kwargs['endpoint_url'] = settings.AWS_S3_ENDPOINT\n elif hasattr(settings, 'AWS_S3_HOST'):\n if hasattr(settings, 'AWS_S3_USE_SSL'\n ) and settings.AWS_S3_USE_SSL is False:\n protocol = 'http://'\n else:\n protocol = 'https://'\n s3_kwargs['endpoint_url'] = '{}{}'.format(protocol, settings.\n AWS_S3_HOST)\n if hasattr(settings, 'AWS_REGION'):\n s3_kwargs['region_name'] = settings.AWS_REGION\n s3_client = boto3.client('s3', **s3_kwargs)\n s3_resource = boto3.resource('s3', **s3_kwargs)\n return s3_client, s3_resource\n", "role": "user" }, { "content": "get_s3_client", "role": "assistant" } ]
def x(page): key_list = page.get('Contents', []) logger.debug('Retrieving page with {} keys'.format(len(key_list))) return dict((k.get('Key'), k) for k in key_list)
get_bucket_page
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(page):\n key_list = page.get('Contents', [])\n logger.debug('Retrieving page with {} keys'.format(len(key_list)))\n return dict((k.get('Key'), k) for k in key_list)\n", "role": "user" }, { "content": "get_bucket_page", "role": "assistant" } ]
def x(aws_bucket_name, s3_client=None, max_keys=1000): logger.debug('Retrieving bucket object list') if not s3_client: s3_client, s3_resource = get_s3_client() obj_dict = {} paginator = s3_client.get_paginator('list_objects') page_iterator = paginator.paginate(Bucket=aws_bucket_name) for page in page_iterator: key_list = page.get('Contents', []) logger.debug('Loading page with {} keys'.format(len(key_list))) for obj in key_list: obj_dict[obj.get('Key')] = obj return obj_dict
get_all_objects_in_bucket
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(aws_bucket_name, s3_client=None, max_keys=1000):\n logger.debug('Retrieving bucket object list')\n if not s3_client:\n s3_client, s3_resource = get_s3_client()\n obj_dict = {}\n paginator = s3_client.get_paginator('list_objects')\n page_iterator = paginator.paginate(Bucket=aws_bucket_name)\n for page in page_iterator:\n key_list = page.get('Contents', [])\n logger.debug('Loading page with {} keys'.format(len(key_list)))\n for obj in key_list:\n obj_dict[obj.get('Key')] = obj\n return obj_dict\n", "role": "user" }, { "content": "get_all_objects_in_bucket", "role": "assistant" } ]
def x(keys, aws_bucket_name, chunk_size=100, s3_client=None): if s3_client is None: s3_client, s3_resource = get_s3_client() key_chunks = [] for i in range(0, len(keys), chunk_size): chunk = [] for key in list(keys)[i:i + 100]: chunk.append({'Key': key}) key_chunks.append(chunk) for chunk in key_chunks: s3_client.delete_objects(Bucket=aws_bucket_name, Delete={'Objects': chunk})
batch_delete_s3_objects
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(keys, aws_bucket_name, chunk_size=100, s3_client=None):\n if s3_client is None:\n s3_client, s3_resource = get_s3_client()\n key_chunks = []\n for i in range(0, len(keys), chunk_size):\n chunk = []\n for key in list(keys)[i:i + 100]:\n chunk.append({'Key': key})\n key_chunks.append(chunk)\n for chunk in key_chunks:\n s3_client.delete_objects(Bucket=aws_bucket_name, Delete={'Objects':\n chunk})\n", "role": "user" }, { "content": "batch_delete_s3_objects", "role": "assistant" } ]
def x(p): if isinstance(p, str): p = string(p) return regex('\\s*') >> p << regex('\\s*')
lexeme
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(p):\n if isinstance(p, str):\n p = string(p)\n return regex('\\\\s*') >> p << regex('\\\\s*')\n", "role": "user" }, { "content": "lexeme", "role": "assistant" } ]
def x(p): return lexeme(p).optional().map(lambda v: False if v is None else True)
is_present
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(p):\n return lexeme(p).optional().map(lambda v: False if v is None else True)\n", "role": "user" }, { "content": "is_present", "role": "assistant" } ]
def x(self, stream): result, _ = (self << eof).parse_partial(stream) return result
parse
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, stream):\n result, _ = (self << eof).parse_partial(stream)\n return result\n", "role": "user" }, { "content": "parse", "role": "assistant" } ]
def x(self, stream): result = self(stream, 0) if result.status: return result.value, stream[result.index:] else: raise ParseError(result.expected, stream, result.furthest)
parse_partial
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, stream):\n result = self(stream, 0)\n if result.status:\n return result.value, stream[result.index:]\n else:\n raise ParseError(result.expected, stream, result.furthest)\n", "role": "user" }, { "content": "parse_partial", "role": "assistant" } ]
def x(array_value, separators=(';', ',', ':'), **kwargs): items_sep, fields_sep, keys_sep = separators return items_sep.join(fields_sep.join(keys_sep.join(x) for x in sorted( it.items())) for it in array_value)
extract_key_values
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(array_value, separators=(';', ',', ':'), **kwargs):\n items_sep, fields_sep, keys_sep = separators\n return items_sep.join(fields_sep.join(keys_sep.join(x) for x in sorted(\n it.items())) for it in array_value)\n", "role": "user" }, { "content": "extract_key_values", "role": "assistant" } ]
def x(cls, schemafile): with open(schemafile) as f: return cls(json.load(f))
from_schemafile
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, schemafile):\n with open(schemafile) as f:\n return cls(json.load(f))\n", "role": "user" }, { "content": "from_schemafile", "role": "assistant" } ]
def x(self, name, serialize_func): if name in self._default_serialization_methods: raise ValueError("Can't replace original %s serialization method") self._serialization_methods[name] = serialize_func
register_serialization_method
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, name, serialize_func):\n if name in self._default_serialization_methods:\n raise ValueError(\"Can't replace original %s serialization method\")\n self._serialization_methods[name] = serialize_func\n", "role": "user" }, { "content": "register_serialization_method", "role": "assistant" } ]
def x(self, obj): return [self._serialize(f, obj) for f in self.fields]
flatten
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, obj):\n return [self._serialize(f, obj) for f in self.fields]\n", "role": "user" }, { "content": "flatten", "role": "assistant" } ]
def x(self, obj): return OrderedDict(zip(self.fieldnames, self.flatten(obj)))
flatten_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, obj):\n return OrderedDict(zip(self.fieldnames, self.flatten(obj)))\n", "role": "user" }, { "content": "flatten_dict", "role": "assistant" } ]
def x(self): LOGGER.debug('Connection %s closing', self.id) if self.busy: raise ConnectionBusyError(self) with self._lock: if not self.handle.closed: try: self.handle.close() except psycopg2.InterfaceError as error: LOGGER.error('Error closing socket: %s', error)
close
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n LOGGER.debug('Connection %s closing', self.id)\n if self.busy:\n raise ConnectionBusyError(self)\n with self._lock:\n if not self.handle.closed:\n try:\n self.handle.close()\n except psycopg2.InterfaceError as error:\n LOGGER.error('Error closing socket: %s', error)\n", "role": "user" }, { "content": "close", "role": "assistant" } ]
def x(self): if self.handle.isexecuting(): return True elif self.used_by is None: return False return not self.used_by() is None
busy
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.handle.isexecuting():\n return True\n elif self.used_by is None:\n return False\n return not self.used_by() is None\n", "role": "user" }, { "content": "busy", "role": "assistant" } ]
def x(self): LOGGER.debug('Connection %s freeing', self.id) if self.handle.isexecuting(): raise ConnectionBusyError(self) with self._lock: self.used_by = None LOGGER.debug('Connection %s freed', self.id)
free
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n LOGGER.debug('Connection %s freeing', self.id)\n if self.handle.isexecuting():\n raise ConnectionBusyError(self)\n with self._lock:\n self.used_by = None\n LOGGER.debug('Connection %s freed', self.id)\n", "role": "user" }, { "content": "free", "role": "assistant" } ]
def x(self, session): if self.busy: raise ConnectionBusyError(self) with self._lock: self.used_by = weakref.ref(session) LOGGER.debug('Connection %s locked', self.id)
lock
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, session):\n if self.busy:\n raise ConnectionBusyError(self)\n with self._lock:\n self.used_by = weakref.ref(session)\n LOGGER.debug('Connection %s locked', self.id)\n", "role": "user" }, { "content": "lock", "role": "assistant" } ]
def x(self, connection): if id(connection) in self.connections: raise ValueError('Connection already exists in pool') if len(self.connections) == self.max_size: LOGGER.warning('Race condition found when adding new connection') try: connection.close() except (psycopg2.Error, psycopg2.Warning) as error: LOGGER.error('Error closing the conn that cant be used: %s', error) raise PoolFullError(self) with self._lock: self.connections[id(connection)] = Connection(connection) LOGGER.debug('Pool %s added connection %s', self.id, id(connection))
add
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, connection):\n if id(connection) in self.connections:\n raise ValueError('Connection already exists in pool')\n if len(self.connections) == self.max_size:\n LOGGER.warning('Race condition found when adding new connection')\n try:\n connection.close()\n except (psycopg2.Error, psycopg2.Warning) as error:\n LOGGER.error('Error closing the conn that cant be used: %s', error)\n raise PoolFullError(self)\n with self._lock:\n self.connections[id(connection)] = Connection(connection)\n LOGGER.debug('Pool %s added connection %s', self.id, id(connection))\n", "role": "user" }, { "content": "add", "role": "assistant" } ]
def x(self): return [c for c in self.connections.values() if c.busy and not c.closed]
busy_connections
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return [c for c in self.connections.values() if c.busy and not c.closed]\n", "role": "user" }, { "content": "busy_connections", "role": "assistant" } ]
def x(self): LOGGER.debug('Cleaning the pool') for connection in [self.connections[k] for k in self.connections if self.connections[k].closed]: LOGGER.debug('Removing %s', connection.id) self.remove(connection.handle) if self.idle_duration > self.idle_ttl: self.close() LOGGER.debug('Pool %s cleaned', self.id)
clean
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n LOGGER.debug('Cleaning the pool')\n for connection in [self.connections[k] for k in self.connections if\n self.connections[k].closed]:\n LOGGER.debug('Removing %s', connection.id)\n self.remove(connection.handle)\n if self.idle_duration > self.idle_ttl:\n self.close()\n LOGGER.debug('Pool %s cleaned', self.id)\n", "role": "user" }, { "content": "clean", "role": "assistant" } ]
def x(self): for cid in list(self.connections.keys()): self.remove(self.connections[cid].handle) LOGGER.debug('Pool %s closed', self.id)
close
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for cid in list(self.connections.keys()):\n self.remove(self.connections[cid].handle)\n LOGGER.debug('Pool %s closed', self.id)\n", "role": "user" }, { "content": "close", "role": "assistant" } ]
def x(self, connection): LOGGER.debug('Pool %s freeing connection %s', self.id, id(connection)) try: self.connection_handle(connection).free() except KeyError: raise ConnectionNotFoundError(self.id, id(connection)) if self.idle_connections == list(self.connections.values()): with self._lock: self.idle_start = self.time_method() LOGGER.debug('Pool %s freed connection %s', self.id, id(connection))
free
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, connection):\n LOGGER.debug('Pool %s freeing connection %s', self.id, id(connection))\n try:\n self.connection_handle(connection).free()\n except KeyError:\n raise ConnectionNotFoundError(self.id, id(connection))\n if self.idle_connections == list(self.connections.values()):\n with self._lock:\n self.idle_start = self.time_method()\n LOGGER.debug('Pool %s freed connection %s', self.id, id(connection))\n", "role": "user" }, { "content": "free", "role": "assistant" } ]
def x(self, session): idle = self.idle_connections if idle: connection = idle.pop(0) connection.lock(session) if self.idle_start: with self._lock: self.idle_start = None return connection.handle raise NoIdleConnectionsError(self.id)
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, session):\n idle = self.idle_connections\n if idle:\n connection = idle.pop(0)\n connection.lock(session)\n if self.idle_start:\n with self._lock:\n self.idle_start = None\n return connection.handle\n raise NoIdleConnectionsError(self.id)\n", "role": "user" }, { "content": "get", "role": "assistant" } ]
def x(self): return [c for c in self.connections.values() if not c.busy and not c.closed ]
idle_connections
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return [c for c in self.connections.values() if not c.busy and not c.closed\n ]\n", "role": "user" }, { "content": "idle_connections", "role": "assistant" } ]
def x(self, connection, session): cid = id(connection) try: self.connection_handle(connection).lock(session) except KeyError: raise ConnectionNotFoundError(self.id, cid) else: if self.idle_start: with self._lock: self.idle_start = None LOGGER.debug('Pool %s locked connection %s', self.id, cid)
lock
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, connection, session):\n cid = id(connection)\n try:\n self.connection_handle(connection).lock(session)\n except KeyError:\n raise ConnectionNotFoundError(self.id, cid)\n else:\n if self.idle_start:\n with self._lock:\n self.idle_start = None\n LOGGER.debug('Pool %s locked connection %s', self.id, cid)\n", "role": "user" }, { "content": "lock", "role": "assistant" } ]
def x(self, connection): cid = id(connection) if cid not in self.connections: raise ConnectionNotFoundError(self.id, cid) self.connection_handle(connection).close() with self._lock: del self.connections[cid] LOGGER.debug('Pool %s removed connection %s', self.id, cid)
remove
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, connection):\n cid = id(connection)\n if cid not in self.connections:\n raise ConnectionNotFoundError(self.id, cid)\n self.connection_handle(connection).close()\n with self._lock:\n del self.connections[cid]\n LOGGER.debug('Pool %s removed connection %s', self.id, cid)\n", "role": "user" }, { "content": "remove", "role": "assistant" } ]
def x(self): return {'connections': {'busy': len(self.busy_connections), 'closed': len(self.closed_connections), 'executing': len(self. executing_connections), 'idle': len(self.idle_connections), 'locked': len(self.busy_connections)}, 'exceptions': sum([c. exceptions for c in self.connections.values()]), 'executions': sum( [c.executions for c in self.connections.values()]), 'full': self. is_full, 'idle': {'duration': self.idle_duration, 'ttl': self. idle_ttl}, 'max_size': self.max_size}
report
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return {'connections': {'busy': len(self.busy_connections), 'closed':\n len(self.closed_connections), 'executing': len(self.\n executing_connections), 'idle': len(self.idle_connections),\n 'locked': len(self.busy_connections)}, 'exceptions': sum([c.\n exceptions for c in self.connections.values()]), 'executions': sum(\n [c.executions for c in self.connections.values()]), 'full': self.\n is_full, 'idle': {'duration': self.idle_duration, 'ttl': self.\n idle_ttl}, 'max_size': self.max_size}\n", "role": "user" }, { "content": "report", "role": "assistant" } ]
def x(self): with self._lock: for cid in list(self.connections.keys()): if self.connections[cid].executing: raise ConnectionBusyError(cid) if self.connections[cid].locked: self.connections[cid].free() self.connections[cid].close() del self.connections[cid]
shutdown
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n with self._lock:\n for cid in list(self.connections.keys()):\n if self.connections[cid].executing:\n raise ConnectionBusyError(cid)\n if self.connections[cid].locked:\n self.connections[cid].free()\n self.connections[cid].close()\n del self.connections[cid]\n", "role": "user" }, { "content": "shutdown", "role": "assistant" } ]
def x(cls, pid, connection): with cls._lock: cls._ensure_pool_exists(pid) cls._pools[pid].add(connection)
add
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, pid, connection):\n with cls._lock:\n cls._ensure_pool_exists(pid)\n cls._pools[pid].add(connection)\n", "role": "user" }, { "content": "add", "role": "assistant" } ]
def x(cls, pid): with cls._lock: try: cls._ensure_pool_exists(pid) except KeyError: LOGGER.debug('Pool clean invoked against missing pool %s', pid) return cls._pools[pid].clean() cls._maybe_remove_pool(pid)
clean
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, pid):\n with cls._lock:\n try:\n cls._ensure_pool_exists(pid)\n except KeyError:\n LOGGER.debug('Pool clean invoked against missing pool %s', pid)\n return\n cls._pools[pid].clean()\n cls._maybe_remove_pool(pid)\n", "role": "user" }, { "content": "clean", "role": "assistant" } ]
def x(cls, pid, idle_ttl=DEFAULT_IDLE_TTL, max_size=DEFAULT_MAX_SIZE, time_method=None): if pid in cls._pools: raise KeyError('Pool %s already exists' % pid) with cls._lock: LOGGER.debug('Creating Pool: %s (%i/%i)', pid, idle_ttl, max_size) cls._pools[pid] = Pool(pid, idle_ttl, max_size, time_method)
create
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, pid, idle_ttl=DEFAULT_IDLE_TTL, max_size=DEFAULT_MAX_SIZE,\n time_method=None):\n if pid in cls._pools:\n raise KeyError('Pool %s already exists' % pid)\n with cls._lock:\n LOGGER.debug('Creating Pool: %s (%i/%i)', pid, idle_ttl, max_size)\n cls._pools[pid] = Pool(pid, idle_ttl, max_size, time_method)\n", "role": "user" }, { "content": "create", "role": "assistant" } ]
def x(cls, pid, connection): with cls._lock: LOGGER.debug('Freeing %s from pool %s', id(connection), pid) cls._ensure_pool_exists(pid) cls._pools[pid].free(connection)
free
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, pid, connection):\n with cls._lock:\n LOGGER.debug('Freeing %s from pool %s', id(connection), pid)\n cls._ensure_pool_exists(pid)\n cls._pools[pid].free(connection)\n", "role": "user" }, { "content": "free", "role": "assistant" } ]
def x(cls, pid, session): with cls._lock: cls._ensure_pool_exists(pid) return cls._pools[pid].get(session)
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, pid, session):\n with cls._lock:\n cls._ensure_pool_exists(pid)\n return cls._pools[pid].get(session)\n", "role": "user" }, { "content": "get", "role": "assistant" } ]
def x(cls, pid, connection): with cls._lock: return cls._pools[pid].connection_handle(connection)
get_connection
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, pid, connection):\n with cls._lock:\n return cls._pools[pid].connection_handle(connection)\n", "role": "user" }, { "content": "get_connection", "role": "assistant" } ]
def x(cls, pid, connection): with cls._lock: cls._ensure_pool_exists(pid) return connection in cls._pools[pid]
has_connection
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, pid, connection):\n with cls._lock:\n cls._ensure_pool_exists(pid)\n return connection in cls._pools[pid]\n", "role": "user" }, { "content": "has_connection", "role": "assistant" } ]
def x(cls, pid): with cls._lock: cls._ensure_pool_exists(pid) return bool(cls._pools[pid].idle_connections)
has_idle_connection
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, pid):\n with cls._lock:\n cls._ensure_pool_exists(pid)\n return bool(cls._pools[pid].idle_connections)\n", "role": "user" }, { "content": "has_idle_connection", "role": "assistant" } ]
def x(cls, pid): with cls._lock: cls._ensure_pool_exists(pid) return cls._pools[pid].is_full
is_full
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, pid):\n with cls._lock:\n cls._ensure_pool_exists(pid)\n return cls._pools[pid].is_full\n", "role": "user" }, { "content": "is_full", "role": "assistant" } ]
def x(cls, pid, connection, session): with cls._lock: cls._ensure_pool_exists(pid) cls._pools[pid].lock(connection, session)
lock
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, pid, connection, session):\n with cls._lock:\n cls._ensure_pool_exists(pid)\n cls._pools[pid].lock(connection, session)\n", "role": "user" }, { "content": "lock", "role": "assistant" } ]
def x(cls, pid): with cls._lock: cls._ensure_pool_exists(pid) cls._pools[pid].close() del cls._pools[pid]
remove
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, pid):\n with cls._lock:\n cls._ensure_pool_exists(pid)\n cls._pools[pid].close()\n del cls._pools[pid]\n", "role": "user" }, { "content": "remove", "role": "assistant" } ]
def x(cls, pid, connection): cls._ensure_pool_exists(pid) cls._pools[pid].remove(connection)
remove_connection
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, pid, connection):\n cls._ensure_pool_exists(pid)\n cls._pools[pid].remove(connection)\n", "role": "user" }, { "content": "remove_connection", "role": "assistant" } ]
def x(cls, pid, ttl): with cls._lock: cls._ensure_pool_exists(pid) cls._pools[pid].set_idle_ttl(ttl)
set_idle_ttl
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, pid, ttl):\n with cls._lock:\n cls._ensure_pool_exists(pid)\n cls._pools[pid].set_idle_ttl(ttl)\n", "role": "user" }, { "content": "set_idle_ttl", "role": "assistant" } ]
def x(cls, pid, size): with cls._lock: cls._ensure_pool_exists(pid) cls._pools[pid].set_max_size(size)
set_max_size
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, pid, size):\n with cls._lock:\n cls._ensure_pool_exists(pid)\n cls._pools[pid].set_max_size(size)\n", "role": "user" }, { "content": "set_max_size", "role": "assistant" } ]
def x(cls): for pid in list(cls._pools.keys()): cls._pools[pid].shutdown() LOGGER.info('Shutdown complete, all pooled connections closed')
shutdown
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls):\n for pid in list(cls._pools.keys()):\n cls._pools[pid].shutdown()\n LOGGER.info('Shutdown complete, all pooled connections closed')\n", "role": "user" }, { "content": "shutdown", "role": "assistant" } ]
def x(cls, pid): with cls._lock: cls._ensure_pool_exists(pid) return len(cls._pools[pid])
size
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, pid):\n with cls._lock:\n cls._ensure_pool_exists(pid)\n return len(cls._pools[pid])\n", "role": "user" }, { "content": "size", "role": "assistant" } ]
def x(cls): return {'timestamp': datetime.datetime.utcnow().isoformat(), 'process': os.getpid(), 'pools': dict([(i, p.report()) for i, p in cls._pools. items()])}
report
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls):\n return {'timestamp': datetime.datetime.utcnow().isoformat(), 'process':\n os.getpid(), 'pools': dict([(i, p.report()) for i, p in cls._pools.\n items()])}\n", "role": "user" }, { "content": "report", "role": "assistant" } ]
def x(cls, pid): if not len(cls._pools[pid]): del cls._pools[pid]
maybe_remove_pool
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, pid):\n if not len(cls._pools[pid]):\n del cls._pools[pid]\n", "role": "user" }, { "content": "maybe_remove_pool", "role": "assistant" } ]
def x(self, name, args=None): try: self._cursor.callproc(name, args) except psycopg2.Error as err: self._incr_exceptions() raise err finally: self._incr_executions() return results.Results(self._cursor)
callproc
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, name, args=None):\n try:\n self._cursor.callproc(name, args)\n except psycopg2.Error as err:\n self._incr_exceptions()\n raise err\n finally:\n self._incr_executions()\n return results.Results(self._cursor)\n", "role": "user" }, { "content": "callproc", "role": "assistant" } ]
def x(self): if not self._conn: raise psycopg2.InterfaceError('Connection not open') LOGGER.info('Closing connection %r in %s', self._conn, self.pid) self._pool_manager.free(self.pid, self._conn) self._pool_manager.remove_connection(self.pid, self._conn) self._conn, self._cursor = None, None
close
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if not self._conn:\n raise psycopg2.InterfaceError('Connection not open')\n LOGGER.info('Closing connection %r in %s', self._conn, self.pid)\n self._pool_manager.free(self.pid, self._conn)\n self._pool_manager.remove_connection(self.pid, self._conn)\n self._conn, self._cursor = None, None\n", "role": "user" }, { "content": "close", "role": "assistant" } ]
def x(self): return hashlib.md5(':'.join([self.__class__.__name__, self._uri]). encode('utf-8')).hexdigest()
pid
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return hashlib.md5(':'.join([self.__class__.__name__, self._uri]).\n encode('utf-8')).hexdigest()\n", "role": "user" }, { "content": "pid", "role": "assistant" } ]
def x(self, sql, parameters=None): try: self._cursor.execute(sql, parameters) except psycopg2.Error as err: self._incr_exceptions() raise err finally: self._incr_executions() return results.Results(self._cursor)
query
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, sql, parameters=None):\n try:\n self._cursor.execute(sql, parameters)\n except psycopg2.Error as err:\n self._incr_exceptions()\n raise err\n finally:\n self._incr_executions()\n return results.Results(self._cursor)\n", "role": "user" }, { "content": "query", "role": "assistant" } ]
def x(self, value=DEFAULT_ENCODING): if self._conn.encoding != value: self._conn.set_client_encoding(value)
set_encoding
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, value=DEFAULT_ENCODING):\n if self._conn.encoding != value:\n self._conn.set_client_encoding(value)\n", "role": "user" }, { "content": "set_encoding", "role": "assistant" } ]
def x(self): if self._cursor: LOGGER.debug('Closing the cursor on %s', self.pid) self._cursor.close() self._cursor = None if self._conn: LOGGER.debug('Freeing %s in the pool', self.pid) try: pool.PoolManager.instance().free(self.pid, self._conn) except pool.ConnectionNotFoundError: pass self._conn = None
cleanup
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self._cursor:\n LOGGER.debug('Closing the cursor on %s', self.pid)\n self._cursor.close()\n self._cursor = None\n if self._conn:\n LOGGER.debug('Freeing %s in the pool', self.pid)\n try:\n pool.PoolManager.instance().free(self.pid, self._conn)\n except pool.ConnectionNotFoundError:\n pass\n self._conn = None\n", "role": "user" }, { "content": "cleanup", "role": "assistant" } ]