source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(ip): if not validate_ip(ip): return None quads = ip.split('.') if len(quads) == 1: quads = quads + [0, 0, 0] elif len(quads) < 4: host = quads[-1:] quads = quads[:-1] + [0] * (4 - len(quads)) + host lngip = 0 for q in quads: lngip = lngip << 8 | int(q) return lngip
ip2long
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ip):\n if not validate_ip(ip):\n return None\n quads = ip.split('.')\n if len(quads) == 1:\n quads = quads + [0, 0, 0]\n elif len(quads) < 4:\n host = quads[-1:]\n quads = quads[:-1] + [0] * (4 - len(quads)) + host\n lngip = 0\n for q in quads:\n lngip = lngip << 8 | int(q)\n return lngip\n", "role": "user" }, { "content": "ip2long", "role": "assistant" } ]
def x(ip): if not validate_ip(ip): return None quads = ip.split('.') netw = 0 for i in range(4): netw = netw << 8 | int(len(quads) > i and quads[i] or 0) return netw
ip2network
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ip):\n if not validate_ip(ip):\n return None\n quads = ip.split('.')\n netw = 0\n for i in range(4):\n netw = netw << 8 | int(len(quads) > i and quads[i] or 0)\n return netw\n", "role": "user" }, { "content": "ip2network", "role": "assistant" } ]
def x(l): if MAX_IP < l or l < MIN_IP: raise TypeError('expected int between %d and %d inclusive' % ( MIN_IP, MAX_IP)) return '%d.%d.%d.%d' % (l >> 24 & 255, l >> 16 & 255, l >> 8 & 255, l & 255 )
long2ip
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(l):\n if MAX_IP < l or l < MIN_IP:\n raise TypeError('expected int between %d and %d inclusive' % (\n MIN_IP, MAX_IP))\n return '%d.%d.%d.%d' % (l >> 24 & 255, l >> 16 & 255, l >> 8 & 255, l & 255\n )\n", "role": "user" }, { "content": "long2ip", "role": "assistant" } ]
def x(cidr): if not validate_cidr(cidr): return None ip, prefix = cidr.split('/') prefix = int(prefix) network = ip2network(ip) return _block_from_ip_and_prefix(network, prefix)
cidr2block
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cidr):\n if not validate_cidr(cidr):\n return None\n ip, prefix = cidr.split('/')\n prefix = int(prefix)\n network = ip2network(ip)\n return _block_from_ip_and_prefix(network, prefix)\n", "role": "user" }, { "content": "cidr2block", "role": "assistant" } ]
def x(subnet): if not validate_subnet(subnet): return None ip, netmask = subnet.split('/') prefix = netmask2prefix(netmask) network = ip2network(ip) return _block_from_ip_and_prefix(network, prefix)
subnet2block
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(subnet):\n if not validate_subnet(subnet):\n return None\n ip, netmask = subnet.split('/')\n prefix = netmask2prefix(netmask)\n network = ip2network(ip)\n return _block_from_ip_and_prefix(network, prefix)\n", "role": "user" }, { "content": "subnet2block", "role": "assistant" } ]
def x(ip, prefix): shift = 32 - prefix block_start = ip >> shift << shift mask = (1 << shift) - 1 block_end = block_start | mask return long2ip(block_start), long2ip(block_end)
block_from_ip_and_prefix
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ip, prefix):\n shift = 32 - prefix\n block_start = ip >> shift << shift\n mask = (1 << shift) - 1\n block_end = block_start | mask\n return long2ip(block_start), long2ip(block_end)\n", "role": "user" }, { "content": "block_from_ip_and_prefix", "role": "assistant" } ]
def x(job, samples, config): job.fileStore.logToMaster('Downloaded shared files') file_names = ['reference', 'phase', 'mills', 'dbsnp', 'cosmic'] urls = [config.reference, config.phase, config.mills, config.dbsnp, config.cosmic] for name, url in zip(file_names, urls): if url: vars(config)[name] = job.addChildJobFn(download_url_job, url=url ).rv() job.addFollowOnJobFn(reference_preprocessing, samples, config)
download_shared_files
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(job, samples, config):\n job.fileStore.logToMaster('Downloaded shared files')\n file_names = ['reference', 'phase', 'mills', 'dbsnp', 'cosmic']\n urls = [config.reference, config.phase, config.mills, config.dbsnp,\n config.cosmic]\n for name, url in zip(file_names, urls):\n if url:\n vars(config)[name] = job.addChildJobFn(download_url_job, url=url\n ).rv()\n job.addFollowOnJobFn(reference_preprocessing, samples, config)\n", "role": "user" }, { "content": "download_shared_files", "role": "assistant" } ]
def x(job, samples, config): job.fileStore.logToMaster('Processed reference files') config.fai = job.addChildJobFn(run_samtools_faidx, config.reference).rv() config.dict = job.addChildJobFn(run_picard_create_sequence_dictionary, config.reference).rv() job.addFollowOnJobFn(map_job, download_sample, samples, config)
reference_preprocessing
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(job, samples, config):\n job.fileStore.logToMaster('Processed reference files')\n config.fai = job.addChildJobFn(run_samtools_faidx, config.reference).rv()\n config.dict = job.addChildJobFn(run_picard_create_sequence_dictionary,\n config.reference).rv()\n job.addFollowOnJobFn(map_job, download_sample, samples, config)\n", "role": "user" }, { "content": "reference_preprocessing", "role": "assistant" } ]
def x(job, sample, config): config = argparse.Namespace(**vars(config)) uuid, normal_url, tumor_url = sample job.fileStore.logToMaster('Downloaded sample: ' + uuid) config.uuid = uuid config.normal = normal_url config.tumor = tumor_url config.cores = min(config.maxCores, int(multiprocessing.cpu_count())) disk = '1G' if config.ci_test else '20G' config.normal_bam = job.addChildJobFn(download_url_job, url=config. normal, s3_key_path=config.ssec, cghub_key_path=config.gtkey, disk=disk ).rv() config.tumor_bam = job.addChildJobFn(download_url_job, url=config.tumor, s3_key_path=config.ssec, cghub_key_path=config.gtkey, disk=disk).rv() job.addFollowOnJobFn(index_bams, config)
download_sample
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(job, sample, config):\n config = argparse.Namespace(**vars(config))\n uuid, normal_url, tumor_url = sample\n job.fileStore.logToMaster('Downloaded sample: ' + uuid)\n config.uuid = uuid\n config.normal = normal_url\n config.tumor = tumor_url\n config.cores = min(config.maxCores, int(multiprocessing.cpu_count()))\n disk = '1G' if config.ci_test else '20G'\n config.normal_bam = job.addChildJobFn(download_url_job, url=config.\n normal, s3_key_path=config.ssec, cghub_key_path=config.gtkey, disk=disk\n ).rv()\n config.tumor_bam = job.addChildJobFn(download_url_job, url=config.tumor,\n s3_key_path=config.ssec, cghub_key_path=config.gtkey, disk=disk).rv()\n job.addFollowOnJobFn(index_bams, config)\n", "role": "user" }, { "content": "download_sample", "role": "assistant" } ]
def x(job, config): job.fileStore.logToMaster('Indexed sample BAMS: ' + config.uuid) disk = '1G' if config.ci_test else '20G' config.normal_bai = job.addChildJobFn(run_samtools_index, config. normal_bam, cores=1, disk=disk).rv() config.tumor_bai = job.addChildJobFn(run_samtools_index, config. tumor_bam, cores=1, disk=disk).rv() job.addFollowOnJobFn(preprocessing_declaration, config)
index_bams
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(job, config):\n job.fileStore.logToMaster('Indexed sample BAMS: ' + config.uuid)\n disk = '1G' if config.ci_test else '20G'\n config.normal_bai = job.addChildJobFn(run_samtools_index, config.\n normal_bam, cores=1, disk=disk).rv()\n config.tumor_bai = job.addChildJobFn(run_samtools_index, config.\n tumor_bam, cores=1, disk=disk).rv()\n job.addFollowOnJobFn(preprocessing_declaration, config)\n", "role": "user" }, { "content": "index_bams", "role": "assistant" } ]
def x(job, config): if config.preprocessing: job.fileStore.logToMaster('Ran preprocessing: ' + config.uuid) disk = '1G' if config.ci_test else '20G' mem = '2G' if config.ci_test else '10G' processed_normal = job.wrapJobFn(run_gatk_preprocessing, config. normal_bam, config.normal_bai, config.reference, config.dict, config.fai, config.phase, config.mills, config.dbsnp, mem, cores=1, memory=mem, disk=disk) processed_tumor = job.wrapJobFn(run_gatk_preprocessing, config. tumor_bam, config.tumor_bai, config.reference, config.dict, config.fai, config.phase, config.mills, config.dbsnp, mem, cores=1, memory=mem, disk=disk) static_workflow = job.wrapJobFn(static_workflow_declaration, config, processed_normal.rv(0), processed_normal.rv(1), processed_tumor .rv(0), processed_tumor.rv(1)) job.addChild(processed_normal) job.addChild(processed_tumor) job.addFollowOn(static_workflow) else: job.addFollowOnJobFn(static_workflow_declaration, config, config. normal_bam, config.normal_bai, config.tumor_bam, config.tumor_bai)
preprocessing_declaration
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(job, config):\n if config.preprocessing:\n job.fileStore.logToMaster('Ran preprocessing: ' + config.uuid)\n disk = '1G' if config.ci_test else '20G'\n mem = '2G' if config.ci_test else '10G'\n processed_normal = job.wrapJobFn(run_gatk_preprocessing, config.\n normal_bam, config.normal_bai, config.reference, config.dict,\n config.fai, config.phase, config.mills, config.dbsnp, mem,\n cores=1, memory=mem, disk=disk)\n processed_tumor = job.wrapJobFn(run_gatk_preprocessing, config.\n tumor_bam, config.tumor_bai, config.reference, config.dict,\n config.fai, config.phase, config.mills, config.dbsnp, mem,\n cores=1, memory=mem, disk=disk)\n static_workflow = job.wrapJobFn(static_workflow_declaration, config,\n processed_normal.rv(0), processed_normal.rv(1), processed_tumor\n .rv(0), processed_tumor.rv(1))\n job.addChild(processed_normal)\n job.addChild(processed_tumor)\n job.addFollowOn(static_workflow)\n else:\n job.addFollowOnJobFn(static_workflow_declaration, config, config.\n normal_bam, config.normal_bai, config.tumor_bam, config.tumor_bai)\n", "role": "user" }, { "content": "preprocessing_declaration", "role": "assistant" } ]
def x(job, config, normal_bam, normal_bai, tumor_bam, tumor_bai): memory = '1G' if config.ci_test else '10G' disk = '1G' if config.ci_test else '75G' mutect_results, pindel_results, muse_results = None, None, None if config.run_mutect: mutect_results = job.addChildJobFn(run_mutect, normal_bam, normal_bai, tumor_bam, tumor_bai, config.reference, config.dict, config.fai, config.cosmic, config.dbsnp, cores=1, memory=memory, disk=disk).rv() if config.run_pindel: pindel_results = job.addChildJobFn(run_pindel, normal_bam, normal_bai, tumor_bam, tumor_bai, config.reference, config.fai, cores=config.cores, memory=memory, disk=disk).rv() if config.run_muse: muse_results = job.addChildJobFn(run_muse, normal_bam, normal_bai, tumor_bam, tumor_bai, config.reference, config.dict, config.fai, config.dbsnp, cores=config.cores, memory=memory, disk=disk).rv() consolidation = job.wrapJobFn(consolidate_output, config, mutect_results, pindel_results, muse_results) job.addFollowOn(consolidation)
static_workflow_declaration
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(job, config, normal_bam, normal_bai, tumor_bam, tumor_bai):\n memory = '1G' if config.ci_test else '10G'\n disk = '1G' if config.ci_test else '75G'\n mutect_results, pindel_results, muse_results = None, None, None\n if config.run_mutect:\n mutect_results = job.addChildJobFn(run_mutect, normal_bam,\n normal_bai, tumor_bam, tumor_bai, config.reference, config.dict,\n config.fai, config.cosmic, config.dbsnp, cores=1, memory=memory,\n disk=disk).rv()\n if config.run_pindel:\n pindel_results = job.addChildJobFn(run_pindel, normal_bam,\n normal_bai, tumor_bam, tumor_bai, config.reference, config.fai,\n cores=config.cores, memory=memory, disk=disk).rv()\n if config.run_muse:\n muse_results = job.addChildJobFn(run_muse, normal_bam, normal_bai,\n tumor_bam, tumor_bai, config.reference, config.dict, config.fai,\n config.dbsnp, cores=config.cores, memory=memory, disk=disk).rv()\n consolidation = job.wrapJobFn(consolidate_output, config,\n mutect_results, pindel_results, muse_results)\n job.addFollowOn(consolidation)\n", "role": "user" }, { "content": "static_workflow_declaration", "role": "assistant" } ]
def x(job, config, mutect, pindel, muse): work_dir = job.fileStore.getLocalTempDir() mutect_tar, pindel_tar, muse_tar = None, None, None if mutect: mutect_tar = job.fileStore.readGlobalFile(mutect, os.path.join( work_dir, 'mutect.tar.gz')) if pindel: pindel_tar = job.fileStore.readGlobalFile(pindel, os.path.join( work_dir, 'pindel.tar.gz')) if muse: muse_tar = job.fileStore.readGlobalFile(muse, os.path.join(work_dir, 'muse.tar.gz')) out_tar = os.path.join(work_dir, config.uuid + '.tar.gz') tar_list = [x for x in [mutect_tar, pindel_tar, muse_tar] if x is not None] with tarfile.open(os.path.join(work_dir, out_tar), 'w:gz') as f_out: for tar in tar_list: with tarfile.open(tar, 'r') as f_in: for tarinfo in f_in: with closing(f_in.extractfile(tarinfo)) as f_in_file: if tar is mutect_tar: tarinfo.name = os.path.join(config.uuid, 'mutect', os.path.basename(tarinfo.name)) elif tar is pindel_tar: tarinfo.name = os.path.join(config.uuid, 'pindel', os.path.basename(tarinfo.name)) else: tarinfo.name = os.path.join(config.uuid, 'muse', os.path.basename(tarinfo.name)) f_out.addfile(tarinfo, fileobj=f_in_file) if urlparse(config.output_dir).scheme == 's3': job.fileStore.logToMaster('Uploading {} to S3: {}'.format(config. uuid, config.output_dir)) s3am_upload(job=job, fpath=out_tar, s3_dir=config.output_dir, num_cores=config.cores) else: job.fileStore.logToMaster('Moving {} to output dir: {}'.format( config.uuid, config.output_dir)) mkdir_p(config.output_dir) copy_files(file_paths=[out_tar], output_dir=config.output_dir)
consolidate_output
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(job, config, mutect, pindel, muse):\n work_dir = job.fileStore.getLocalTempDir()\n mutect_tar, pindel_tar, muse_tar = None, None, None\n if mutect:\n mutect_tar = job.fileStore.readGlobalFile(mutect, os.path.join(\n work_dir, 'mutect.tar.gz'))\n if pindel:\n pindel_tar = job.fileStore.readGlobalFile(pindel, os.path.join(\n work_dir, 'pindel.tar.gz'))\n if muse:\n muse_tar = job.fileStore.readGlobalFile(muse, os.path.join(work_dir,\n 'muse.tar.gz'))\n out_tar = os.path.join(work_dir, config.uuid + '.tar.gz')\n tar_list = [x for x in [mutect_tar, pindel_tar, muse_tar] if x is not None]\n with tarfile.open(os.path.join(work_dir, out_tar), 'w:gz') as f_out:\n for tar in tar_list:\n with tarfile.open(tar, 'r') as f_in:\n for tarinfo in f_in:\n with closing(f_in.extractfile(tarinfo)) as f_in_file:\n if tar is mutect_tar:\n tarinfo.name = os.path.join(config.uuid,\n 'mutect', os.path.basename(tarinfo.name))\n elif tar is pindel_tar:\n tarinfo.name = os.path.join(config.uuid,\n 'pindel', os.path.basename(tarinfo.name))\n else:\n tarinfo.name = os.path.join(config.uuid, 'muse',\n os.path.basename(tarinfo.name))\n f_out.addfile(tarinfo, fileobj=f_in_file)\n if urlparse(config.output_dir).scheme == 's3':\n job.fileStore.logToMaster('Uploading {} to S3: {}'.format(config.\n uuid, config.output_dir))\n s3am_upload(job=job, fpath=out_tar, s3_dir=config.output_dir,\n num_cores=config.cores)\n else:\n job.fileStore.logToMaster('Moving {} to output dir: {}'.format(\n config.uuid, config.output_dir))\n mkdir_p(config.output_dir)\n copy_files(file_paths=[out_tar], output_dir=config.output_dir)\n", "role": "user" }, { "content": "consolidate_output", "role": "assistant" } ]
def x(path_to_manifest): samples = [] with open(path_to_manifest, 'r') as f: for line in f.readlines(): if not line.isspace() and not line.startswith('#'): sample = line.strip().split('\t') require(len(sample) == 3, 'Bad manifest format! Expected 3 tab separated columns, got: {}' .format(sample)) uuid, normal, tumor = sample for url in [normal, tumor]: require(urlparse(url).scheme and urlparse(url), 'Invalid URL passed for {}'.format(url)) samples.append(sample) return samples
parse_manifest
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path_to_manifest):\n samples = []\n with open(path_to_manifest, 'r') as f:\n for line in f.readlines():\n if not line.isspace() and not line.startswith('#'):\n sample = line.strip().split('\\t')\n require(len(sample) == 3,\n 'Bad manifest format! Expected 3 tab separated columns, got: {}'\n .format(sample))\n uuid, normal, tumor = sample\n for url in [normal, tumor]:\n require(urlparse(url).scheme and urlparse(url),\n 'Invalid URL passed for {}'.format(url))\n samples.append(sample)\n return samples\n", "role": "user" }, { "content": "parse_manifest", "role": "assistant" } ]
def x(): parser = argparse.ArgumentParser(description=main.__doc__, formatter_class=argparse.RawTextHelpFormatter) subparsers = parser.add_subparsers(dest='command') subparsers.add_parser('generate-config', help= 'Generates an editable config in the current working directory.') subparsers.add_parser('generate-manifest', help= 'Generates an editable manifest in the current working directory.') subparsers.add_parser('generate', help= 'Generates a config and manifest in the current working directory.') parser_run = subparsers.add_parser('run', help= 'Runs the CGL exome pipeline') parser_run.add_argument('--config', default='config-toil-exome.yaml', type=str, help= """Path to the (filled in) config file, generated with "generate-config". Default value: "%(default)s\"""" ) parser_run.add_argument('--manifest', default='manifest-toil-exome.tsv', type=str, help= """Path to the (filled in) manifest file, generated with "generate-manifest". Default value: "%(default)s\"""" ) parser_run.add_argument('--normal', default=None, type=str, help= 'URL for the normal BAM. URLs can take the form: http://, ftp://, file://, s3://, and gnos://. The UUID for the sample must be given with the "--uuid" flag.' ) parser_run.add_argument('--tumor', default=None, type=str, help= 'URL for the tumor BAM. URLs can take the form: http://, ftp://, file://, s3://, and gnos://. The UUID for the sample must be given with the "--uuid" flag.' ) parser_run.add_argument('--uuid', default=None, type=str, help= 'Provide the UUID of a sample when using the"--tumor" and "--normal" option' ) if len(sys.argv) == 1: parser.print_help() sys.exit(1) Job.Runner.addToilOptions(parser_run) args = parser.parse_args() cwd = os.getcwd() if args.command == 'generate-config' or args.command == 'generate': generate_file(os.path.join(cwd, 'config-toil-exome.yaml'), generate_config) if args.command == 'generate-manifest' or args.command == 'generate': generate_file(os.path.join(cwd, 'manifest-toil-exome.tsv'), generate_manifest) elif args.command == 'run': require(os.path.exists(args.config), '{} not found. Please run "toil-rnaseq generate-config"'.format (args.config)) if args.normal or args.tumor or args.uuid: require(args.normal and args.tumor and args.uuid, '"--tumor", "--normal" and "--uuid" must all be supplied') samples = [[args.uuid, args.normal, args.tumor]] else: samples = parse_manifest(args.manifest) parsed_config = {x.replace('-', '_'): y for x, y in yaml.load(open( args.config).read()).iteritems()} config = argparse.Namespace(**parsed_config) config.maxCores = int(args.maxCores) if args.maxCores else sys.maxint if config.preprocessing: require(config.reference and config.phase and config.mills and config.dbsnp, 'Missing inputs for preprocessing, check config file.') if config.run_mutect: require(config.reference and config.dbsnp and config.cosmic, 'Missing inputs for MuTect, check config file.') if config.run_pindel: require(config.reference, 'Missing input (reference) for Pindel.') if config.run_muse: require(config.reference and config.dbsnp, 'Missing inputs for MuSe, check config file.') require(config.output_dir, 'No output location specified: {}'. format(config.output_dir)) for program in ['curl', 'docker']: require(next(which(program), None), program + ' must be installed on every node.'.format(program)) Job.Runner.startToil(Job.wrapJobFn(download_shared_files, samples, config), args)
main
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n parser = argparse.ArgumentParser(description=main.__doc__,\n formatter_class=argparse.RawTextHelpFormatter)\n subparsers = parser.add_subparsers(dest='command')\n subparsers.add_parser('generate-config', help=\n 'Generates an editable config in the current working directory.')\n subparsers.add_parser('generate-manifest', help=\n 'Generates an editable manifest in the current working directory.')\n subparsers.add_parser('generate', help=\n 'Generates a config and manifest in the current working directory.')\n parser_run = subparsers.add_parser('run', help=\n 'Runs the CGL exome pipeline')\n parser_run.add_argument('--config', default='config-toil-exome.yaml',\n type=str, help=\n \"\"\"Path to the (filled in) config file, generated with \"generate-config\". \nDefault value: \"%(default)s\\\"\"\"\"\n )\n parser_run.add_argument('--manifest', default='manifest-toil-exome.tsv',\n type=str, help=\n \"\"\"Path to the (filled in) manifest file, generated with \"generate-manifest\". \nDefault value: \"%(default)s\\\"\"\"\"\n )\n parser_run.add_argument('--normal', default=None, type=str, help=\n 'URL for the normal BAM. URLs can take the form: http://, ftp://, file://, s3://, and gnos://. The UUID for the sample must be given with the \"--uuid\" flag.'\n )\n parser_run.add_argument('--tumor', default=None, type=str, help=\n 'URL for the tumor BAM. URLs can take the form: http://, ftp://, file://, s3://, and gnos://. The UUID for the sample must be given with the \"--uuid\" flag.'\n )\n parser_run.add_argument('--uuid', default=None, type=str, help=\n 'Provide the UUID of a sample when using the\"--tumor\" and \"--normal\" option'\n )\n if len(sys.argv) == 1:\n parser.print_help()\n sys.exit(1)\n Job.Runner.addToilOptions(parser_run)\n args = parser.parse_args()\n cwd = os.getcwd()\n if args.command == 'generate-config' or args.command == 'generate':\n generate_file(os.path.join(cwd, 'config-toil-exome.yaml'),\n generate_config)\n if args.command == 'generate-manifest' or args.command == 'generate':\n generate_file(os.path.join(cwd, 'manifest-toil-exome.tsv'),\n generate_manifest)\n elif args.command == 'run':\n require(os.path.exists(args.config),\n '{} not found. Please run \"toil-rnaseq generate-config\"'.format\n (args.config))\n if args.normal or args.tumor or args.uuid:\n require(args.normal and args.tumor and args.uuid,\n '\"--tumor\", \"--normal\" and \"--uuid\" must all be supplied')\n samples = [[args.uuid, args.normal, args.tumor]]\n else:\n samples = parse_manifest(args.manifest)\n parsed_config = {x.replace('-', '_'): y for x, y in yaml.load(open(\n args.config).read()).iteritems()}\n config = argparse.Namespace(**parsed_config)\n config.maxCores = int(args.maxCores) if args.maxCores else sys.maxint\n if config.preprocessing:\n require(config.reference and config.phase and config.mills and\n config.dbsnp,\n 'Missing inputs for preprocessing, check config file.')\n if config.run_mutect:\n require(config.reference and config.dbsnp and config.cosmic,\n 'Missing inputs for MuTect, check config file.')\n if config.run_pindel:\n require(config.reference, 'Missing input (reference) for Pindel.')\n if config.run_muse:\n require(config.reference and config.dbsnp,\n 'Missing inputs for MuSe, check config file.')\n require(config.output_dir, 'No output location specified: {}'.\n format(config.output_dir))\n for program in ['curl', 'docker']:\n require(next(which(program), None), program +\n ' must be installed on every node.'.format(program))\n Job.Runner.startToil(Job.wrapJobFn(download_shared_files, samples,\n config), args)\n", "role": "user" }, { "content": "main", "role": "assistant" } ]
def x(job, inputs, samples): shared_ids = {} urls = [('amb', inputs.amb), ('ann', inputs.ann), ('bwt', inputs.bwt), ('pac', inputs.pac), ('sa', inputs.sa)] if inputs.alt: urls.append(('alt', inputs.alt)) download_ref = job.wrapJobFn(download_url_job, inputs.ref, disk='3G') job.addChild(download_ref) shared_ids['ref'] = download_ref.rv() if inputs.fai: shared_ids['fai'] = job.addChildJobFn(download_url_job, inputs.fai).rv( ) else: faidx = job.wrapJobFn(run_samtools_faidx, download_ref.rv()) shared_ids['fai'] = download_ref.addChild(faidx).rv() if all(x[1] for x in urls): for name, url in urls: shared_ids[name] = job.addChildJobFn(download_url_job, url).rv() else: job.fileStore.logToMaster('BWA index files not provided, creating now') bwa_index = job.wrapJobFn(run_bwa_index, download_ref.rv()) download_ref.addChild(bwa_index) for x, name in enumerate(['amb', 'ann', 'bwt', 'pac', 'sa']): shared_ids[name] = bwa_index.rv(x) job.addFollowOnJobFn(map_job, download_sample_and_align, samples, inputs, shared_ids)
download_reference_files
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(job, inputs, samples):\n shared_ids = {}\n urls = [('amb', inputs.amb), ('ann', inputs.ann), ('bwt', inputs.bwt),\n ('pac', inputs.pac), ('sa', inputs.sa)]\n if inputs.alt:\n urls.append(('alt', inputs.alt))\n download_ref = job.wrapJobFn(download_url_job, inputs.ref, disk='3G')\n job.addChild(download_ref)\n shared_ids['ref'] = download_ref.rv()\n if inputs.fai:\n shared_ids['fai'] = job.addChildJobFn(download_url_job, inputs.fai).rv(\n )\n else:\n faidx = job.wrapJobFn(run_samtools_faidx, download_ref.rv())\n shared_ids['fai'] = download_ref.addChild(faidx).rv()\n if all(x[1] for x in urls):\n for name, url in urls:\n shared_ids[name] = job.addChildJobFn(download_url_job, url).rv()\n else:\n job.fileStore.logToMaster('BWA index files not provided, creating now')\n bwa_index = job.wrapJobFn(run_bwa_index, download_ref.rv())\n download_ref.addChild(bwa_index)\n for x, name in enumerate(['amb', 'ann', 'bwt', 'pac', 'sa']):\n shared_ids[name] = bwa_index.rv(x)\n job.addFollowOnJobFn(map_job, download_sample_and_align, samples,\n inputs, shared_ids)\n", "role": "user" }, { "content": "download_reference_files", "role": "assistant" } ]
def x(job, sample, inputs, ids): uuid, urls = sample r1_url, r2_url = urls if len(urls) == 2 else (urls[0], None) job.fileStore.logToMaster( 'Downloaded sample: {0}. R1 {1}\nR2 {2}\nStarting BWA Run'.format( uuid, r1_url, r2_url)) ids['r1'] = job.addChildJobFn(download_url_job, r1_url, s3_key_path= inputs.ssec, disk=inputs.file_size).rv() if r2_url: ids['r2'] = job.addChildJobFn(download_url_job, r2_url, s3_key_path =inputs.ssec, disk=inputs.file_size).rv() else: ids['r2'] = None inputs.cores = min(inputs.maxCores, multiprocessing.cpu_count()) inputs.uuid = uuid config = dict(**vars(inputs)) config.update(ids) config = argparse.Namespace(**config) bam_id = job.wrapJobFn(run_bwakit, config, sort=inputs.sort, trim= inputs.trim, disk=inputs.file_size, cores=inputs.cores) job.addFollowOn(bam_id) output_name = uuid + '.bam' + str(inputs.suffix ) if inputs.suffix else uuid + '.bam' if urlparse(inputs.output_dir).scheme == 's3': bam_id.addChildJobFn(s3am_upload_job, file_id=bam_id.rv(), file_name=output_name, s3_dir=inputs.output_dir, s3_key_path= inputs.ssec, cores=inputs.cores, disk=inputs.file_size) else: mkdir_p(inputs.ouput_dir) bam_id.addChildJobFn(copy_file_job, name=output_name, file_id= bam_id.rv(), output_dir=inputs.output_dir, disk=inputs.file_size)
download_sample_and_align
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(job, sample, inputs, ids):\n uuid, urls = sample\n r1_url, r2_url = urls if len(urls) == 2 else (urls[0], None)\n job.fileStore.logToMaster(\n 'Downloaded sample: {0}. R1 {1}\\nR2 {2}\\nStarting BWA Run'.format(\n uuid, r1_url, r2_url))\n ids['r1'] = job.addChildJobFn(download_url_job, r1_url, s3_key_path=\n inputs.ssec, disk=inputs.file_size).rv()\n if r2_url:\n ids['r2'] = job.addChildJobFn(download_url_job, r2_url, s3_key_path\n =inputs.ssec, disk=inputs.file_size).rv()\n else:\n ids['r2'] = None\n inputs.cores = min(inputs.maxCores, multiprocessing.cpu_count())\n inputs.uuid = uuid\n config = dict(**vars(inputs))\n config.update(ids)\n config = argparse.Namespace(**config)\n bam_id = job.wrapJobFn(run_bwakit, config, sort=inputs.sort, trim=\n inputs.trim, disk=inputs.file_size, cores=inputs.cores)\n job.addFollowOn(bam_id)\n output_name = uuid + '.bam' + str(inputs.suffix\n ) if inputs.suffix else uuid + '.bam'\n if urlparse(inputs.output_dir).scheme == 's3':\n bam_id.addChildJobFn(s3am_upload_job, file_id=bam_id.rv(),\n file_name=output_name, s3_dir=inputs.output_dir, s3_key_path=\n inputs.ssec, cores=inputs.cores, disk=inputs.file_size)\n else:\n mkdir_p(inputs.ouput_dir)\n bam_id.addChildJobFn(copy_file_job, name=output_name, file_id=\n bam_id.rv(), output_dir=inputs.output_dir, disk=inputs.file_size)\n", "role": "user" }, { "content": "download_sample_and_align", "role": "assistant" } ]
def x(manifest_path): samples = [] with open(manifest_path, 'r') as f: for line in f: if not line.isspace() and not line.startswith('#'): sample = line.strip().split('\t') require(2 <= len(sample) <= 3, 'Bad manifest format! Expected UUID\tURL1\t[URL2] (tab separated), got: {}' .format(sample)) uuid = sample[0] urls = sample[1:] for url in urls: require(urlparse(url).scheme and urlparse(url), 'Invalid URL passed for {}'.format(url)) samples.append([uuid, urls]) return samples
parse_manifest
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(manifest_path):\n samples = []\n with open(manifest_path, 'r') as f:\n for line in f:\n if not line.isspace() and not line.startswith('#'):\n sample = line.strip().split('\\t')\n require(2 <= len(sample) <= 3,\n 'Bad manifest format! Expected UUID\\tURL1\\t[URL2] (tab separated), got: {}'\n .format(sample))\n uuid = sample[0]\n urls = sample[1:]\n for url in urls:\n require(urlparse(url).scheme and urlparse(url),\n 'Invalid URL passed for {}'.format(url))\n samples.append([uuid, urls])\n return samples\n", "role": "user" }, { "content": "parse_manifest", "role": "assistant" } ]
def x(): parser = argparse.ArgumentParser(description=main.__doc__, formatter_class=argparse.RawTextHelpFormatter) subparsers = parser.add_subparsers(dest='command') subparsers.add_parser('generate-config', help= 'Generates an editable config in the current working directory.') subparsers.add_parser('generate-manifest', help= 'Generates an editable manifest in the current working directory.') subparsers.add_parser('generate', help= 'Generates a config and manifest in the current working directory.') parser_run = subparsers.add_parser('run', help= 'Runs the BWA alignment pipeline') group = parser_run.add_mutually_exclusive_group() parser_run.add_argument('--config', default='config-toil-bwa.yaml', type=str, help= 'Path to the (filled in) config file, generated with "generate-config".' ) group.add_argument('--manifest', default='manifest-toil-bwa.tsv', type= str, help= """Path to the (filled in) manifest file, generated with "generate-manifest". Default value: "%(default)s".""" ) group.add_argument('--sample', nargs='+', action=required_length(2, 3), help= 'Space delimited sample UUID and fastq files in the format: uuid url1 [url2].' ) if len(sys.argv) == 1: parser.print_help() sys.exit(1) Job.Runner.addToilOptions(parser_run) args = parser.parse_args() cwd = os.getcwd() if args.command == 'generate-config' or args.command == 'generate': generate_file(os.path.join(cwd, 'config-toil-bwa.yaml'), generate_config) if args.command == 'generate-manifest' or args.command == 'generate': generate_file(os.path.join(cwd, 'manifest-toil-bwa.tsv'), generate_manifest) elif args.command == 'run': require(os.path.exists(args.config), '{} not found. Please run generate-config'.format(args.config)) if not args.sample: args.sample = None require(os.path.exists(args.manifest), '{} not found and no sample provided. Please run "generate-manifest"' .format(args.manifest)) parsed_config = {x.replace('-', '_'): y for x, y in yaml.load(open( args.config).read()).iteritems()} config = argparse.Namespace(**parsed_config) config.maxCores = int(args.maxCores) if args.maxCores else sys.maxint samples = [args.sample[0], args.sample[1:] ] if args.sample else parse_manifest(args.manifest) require(config.ref, 'Missing URL for reference file: {}'.format( config.ref)) require(config.output_dir, 'No output location specified: {}'. format(config.output_dir)) Job.Runner.startToil(Job.wrapJobFn(download_reference_files, config, samples), args)
main
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n parser = argparse.ArgumentParser(description=main.__doc__,\n formatter_class=argparse.RawTextHelpFormatter)\n subparsers = parser.add_subparsers(dest='command')\n subparsers.add_parser('generate-config', help=\n 'Generates an editable config in the current working directory.')\n subparsers.add_parser('generate-manifest', help=\n 'Generates an editable manifest in the current working directory.')\n subparsers.add_parser('generate', help=\n 'Generates a config and manifest in the current working directory.')\n parser_run = subparsers.add_parser('run', help=\n 'Runs the BWA alignment pipeline')\n group = parser_run.add_mutually_exclusive_group()\n parser_run.add_argument('--config', default='config-toil-bwa.yaml',\n type=str, help=\n 'Path to the (filled in) config file, generated with \"generate-config\".'\n )\n group.add_argument('--manifest', default='manifest-toil-bwa.tsv', type=\n str, help=\n \"\"\"Path to the (filled in) manifest file, generated with \"generate-manifest\". \nDefault value: \"%(default)s\".\"\"\"\n )\n group.add_argument('--sample', nargs='+', action=required_length(2, 3),\n help=\n 'Space delimited sample UUID and fastq files in the format: uuid url1 [url2].'\n )\n if len(sys.argv) == 1:\n parser.print_help()\n sys.exit(1)\n Job.Runner.addToilOptions(parser_run)\n args = parser.parse_args()\n cwd = os.getcwd()\n if args.command == 'generate-config' or args.command == 'generate':\n generate_file(os.path.join(cwd, 'config-toil-bwa.yaml'),\n generate_config)\n if args.command == 'generate-manifest' or args.command == 'generate':\n generate_file(os.path.join(cwd, 'manifest-toil-bwa.tsv'),\n generate_manifest)\n elif args.command == 'run':\n require(os.path.exists(args.config),\n '{} not found. Please run generate-config'.format(args.config))\n if not args.sample:\n args.sample = None\n require(os.path.exists(args.manifest),\n '{} not found and no sample provided. Please run \"generate-manifest\"'\n .format(args.manifest))\n parsed_config = {x.replace('-', '_'): y for x, y in yaml.load(open(\n args.config).read()).iteritems()}\n config = argparse.Namespace(**parsed_config)\n config.maxCores = int(args.maxCores) if args.maxCores else sys.maxint\n samples = [args.sample[0], args.sample[1:]\n ] if args.sample else parse_manifest(args.manifest)\n require(config.ref, 'Missing URL for reference file: {}'.format(\n config.ref))\n require(config.output_dir, 'No output location specified: {}'.\n format(config.output_dir))\n Job.Runner.startToil(Job.wrapJobFn(download_reference_files, config,\n samples), args)\n", "role": "user" }, { "content": "main", "role": "assistant" } ]
def x(address): parsed = ipv4.ip2long(address) if parsed is None: parsed = ipv6.ip2long(address) return parsed
address2long
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(address):\n parsed = ipv4.ip2long(address)\n if parsed is None:\n parsed = ipv6.ip2long(address)\n return parsed\n", "role": "user" }, { "content": "address2long", "role": "assistant" } ]
def x(self, item): item = self._cast(item) offset = item - self.startIp if offset >= 0 and offset < self._len: return offset raise ValueError('%s is not in range' % self._ipver.long2ip(item))
index
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, item):\n item = self._cast(item)\n offset = item - self.startIp\n if offset >= 0 and offset < self._len:\n return offset\n raise ValueError('%s is not in range' % self._ipver.long2ip(item))\n", "role": "user" }, { "content": "index", "role": "assistant" } ]
def x(): pid = os.fork() if pid > 0: os.waitpid(pid, 0) return True os.setsid() pid = os.fork() if pid > 0: os._exit(os.EX_OK) return False
detach_process
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n pid = os.fork()\n if pid > 0:\n os.waitpid(pid, 0)\n return True\n os.setsid()\n pid = os.fork()\n if pid > 0:\n os._exit(os.EX_OK)\n return False\n", "role": "user" }, { "content": "detach_process", "role": "assistant" } ]
def x(predicate, timeout): if timeout: if timeout is True: timeout = float('Inf') timeout = time.time() + timeout while not predicate() and time.time() < timeout: time.sleep(0.1) return predicate()
block
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(predicate, timeout):\n if timeout:\n if timeout is True:\n timeout = float('Inf')\n timeout = time.time() + timeout\n while not predicate() and time.time() < timeout:\n time.sleep(0.1)\n return predicate()\n", "role": "user" }, { "content": "block", "role": "assistant" } ]
def x(self): try: with open(self._path, 'r') as f: s = f.read().strip() if not s: return None return int(s) except IOError as e: if e.errno == errno.ENOENT: return None raise
read_pid
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n try:\n with open(self._path, 'r') as f:\n s = f.read().strip()\n if not s:\n return None\n return int(s)\n except IOError as e:\n if e.errno == errno.ENOENT:\n return None\n raise\n", "role": "user" }, { "content": "read_pid", "role": "assistant" } ]
def x(self): handles = [] for handler in self.logger.handlers: for attr in ['sock', 'socket', 'stream']: try: handle = getattr(handler, attr) if handle: handles.append(handle) break except AttributeError: continue return handles
get_logger_file_handles
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n handles = []\n for handler in self.logger.handlers:\n for attr in ['sock', 'socket', 'stream']:\n try:\n handle = getattr(handler, attr)\n if handle:\n handles.append(handle)\n break\n except AttributeError:\n continue\n return handles\n", "role": "user" }, { "content": "get_logger_file_handles", "role": "assistant" } ]
def x(self): pid = self.get_pid() if pid is None: return False try: os.kill(pid, 0) except OSError as e: if e.errno == errno.ESRCH: self.pid_file.release() return False return True
is_running
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n pid = self.get_pid()\n if pid is None:\n return False\n try:\n os.kill(pid, 0)\n except OSError as e:\n if e.errno == errno.ESRCH:\n self.pid_file.release()\n return False\n return True\n", "role": "user" }, { "content": "is_running", "role": "assistant" } ]
def x(self, s): try: return self._signal_events[int(s)] except KeyError: raise ValueError('Signal {} has not been enabled'.format(s))
get_signal_event
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, s):\n try:\n return self._signal_events[int(s)]\n except KeyError:\n raise ValueError('Signal {} has not been enabled'.format(s))\n", "role": "user" }, { "content": "get_signal_event", "role": "assistant" } ]
def x(self, s): self._get_signal_event(s) pid = self.get_pid() if not pid: raise ValueError('Daemon is not running.') os.kill(pid, s)
send_signal
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, s):\n self._get_signal_event(s)\n pid = self.get_pid()\n if not pid:\n raise ValueError('Daemon is not running.')\n os.kill(pid, s)\n", "role": "user" }, { "content": "send_signal", "role": "assistant" } ]
def x(self, block=False): self.send_signal(signal.SIGTERM) return _block(lambda : not self.is_running(), block)
stop
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, block=False):\n self.send_signal(signal.SIGTERM)\n return _block(lambda : not self.is_running(), block)\n", "role": "user" }, { "content": "stop", "role": "assistant" } ]
def x(self, block=False): pid = self.get_pid() if not pid: raise ValueError('Daemon is not running.') try: os.kill(pid, signal.SIGKILL) return _block(lambda : not self.is_running(), block) except OSError as e: if e.errno == errno.ESRCH: raise ValueError('Daemon is not running.') raise finally: self.pid_file.release()
kill
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, block=False):\n pid = self.get_pid()\n if not pid:\n raise ValueError('Daemon is not running.')\n try:\n os.kill(pid, signal.SIGKILL)\n return _block(lambda : not self.is_running(), block)\n except OSError as e:\n if e.errno == errno.ESRCH:\n raise ValueError('Daemon is not running.')\n raise\n finally:\n self.pid_file.release()\n", "role": "user" }, { "content": "kill", "role": "assistant" } ]
def x(self, block=False): pid = self.get_pid() if pid: raise ValueError('Daemon is already running at PID %d.' % pid) try: self.pid_file.acquire() finally: self.pid_file.release() self.clear_signal(signal.SIGTERM) if _detach_process(): return _block(lambda : self.is_running(), block) self._debug('Daemon has detached') def x(s, frame): self._debug('Received signal {}'.format(s)) self._signal_events[int(s)].set() def x(): try: self.pid_file.acquire() self._debug('PID file has been acquired') self._debug('Calling `run`') self.run() self._debug('`run` returned without exception') except Exception as e: self.logger.exception(e) except SystemExit: self._debug('`run` called `sys.exit`') try: self.pid_file.release() self._debug('PID file has been released') except Exception as e: self.logger.exception(e) os._exit(os.EX_OK) try: setproctitle.setproctitle(self.name) self._debug('Process title has been set') files_preserve = self.files_preserve + self._get_logger_file_handles() signal_map = {s: on_signal for s in self._signal_events} signal_map.update({signal.SIGTTIN: None, signal.SIGTTOU: None, signal.SIGTSTP: None}) with DaemonContext(detach_process=False, signal_map=signal_map, files_preserve=files_preserve): self._debug('Daemon context has been established') thread = threading.Thread(target=runner) thread.start() while thread.is_alive(): time.sleep(1) except Exception as e: self.logger.exception(e) os._exit(os.EX_OK)
start
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, block=False):\n pid = self.get_pid()\n if pid:\n raise ValueError('Daemon is already running at PID %d.' % pid)\n try:\n self.pid_file.acquire()\n finally:\n self.pid_file.release()\n self.clear_signal(signal.SIGTERM)\n if _detach_process():\n return _block(lambda : self.is_running(), block)\n self._debug('Daemon has detached')\n\n def x(s, frame):\n self._debug('Received signal {}'.format(s))\n self._signal_events[int(s)].set()\n\n def x():\n try:\n self.pid_file.acquire()\n self._debug('PID file has been acquired')\n self._debug('Calling `run`')\n self.run()\n self._debug('`run` returned without exception')\n except Exception as e:\n self.logger.exception(e)\n except SystemExit:\n self._debug('`run` called `sys.exit`')\n try:\n self.pid_file.release()\n self._debug('PID file has been released')\n except Exception as e:\n self.logger.exception(e)\n os._exit(os.EX_OK)\n try:\n setproctitle.setproctitle(self.name)\n self._debug('Process title has been set')\n files_preserve = self.files_preserve + self._get_logger_file_handles()\n signal_map = {s: on_signal for s in self._signal_events}\n signal_map.update({signal.SIGTTIN: None, signal.SIGTTOU: None,\n signal.SIGTSTP: None})\n with DaemonContext(detach_process=False, signal_map=signal_map,\n files_preserve=files_preserve):\n self._debug('Daemon context has been established')\n thread = threading.Thread(target=runner)\n thread.start()\n while thread.is_alive():\n time.sleep(1)\n except Exception as e:\n self.logger.exception(e)\n os._exit(os.EX_OK)\n", "role": "user" }, { "content": "start", "role": "assistant" } ]
def x(opts): if opts['--address'] and is_boot2docker(): raise DatacatsError('Cannot specify address on boot2docker.') return create_environment(environment_dir=opts['ENVIRONMENT_DIR'], port =opts['PORT'], create_skin=not opts['--bare'], start_web=not opts[ '--image-only'], create_sysadmin=not opts['--no-sysadmin'], site_name=opts['--site'], ckan_version=opts['--ckan'], address=opts ['--address'], log_syslog=opts['--syslog'], datapusher=not opts[ '--no-datapusher'], site_url=opts['--site-url'], interactive=opts[ '--interactive'], init_db=not opts['--no-init-db'])
create
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(opts):\n if opts['--address'] and is_boot2docker():\n raise DatacatsError('Cannot specify address on boot2docker.')\n return create_environment(environment_dir=opts['ENVIRONMENT_DIR'], port\n =opts['PORT'], create_skin=not opts['--bare'], start_web=not opts[\n '--image-only'], create_sysadmin=not opts['--no-sysadmin'],\n site_name=opts['--site'], ckan_version=opts['--ckan'], address=opts\n ['--address'], log_syslog=opts['--syslog'], datapusher=not opts[\n '--no-datapusher'], site_url=opts['--site-url'], interactive=opts[\n '--interactive'], init_db=not opts['--no-init-db'])\n", "role": "user" }, { "content": "create", "role": "assistant" } ]
def x(opts, no_install=False, quiet=False): if opts['--address'] and is_boot2docker(): raise DatacatsError('Cannot specify address on boot2docker.') environment_dir = opts['ENVIRONMENT_DIR'] port = opts['PORT'] address = opts['--address'] start_web = not opts['--image-only'] create_sysadmin = not opts['--no-sysadmin'] site_name = opts['--site'] site_url = opts['--site-url'] interactive = opts['--interactive'] init_db = not opts['--no-init-db'] environment_dir = abspath(environment_dir or '.') log_syslog = opts['--syslog'] environment = Environment.load(environment_dir, site_name) if address: environment.address = address if port: environment.port = int(port) if site_url: environment.site_url = site_url try: if environment.sites and site_name in environment.sites: raise DatacatsError('Site named {0} already exists.'.format( site_name)) making_full_environment = not environment.data_exists() if not quiet: write( 'Creating environment {0}/{1} from existing environment directory "{0}"' .format(environment.name, environment.site_name)) steps = [lambda : environment.create_directories(create_project_dir =False)] + ([environment.save, environment.create_virtualenv] if making_full_environment else []) + [environment.save_site, environment.start_supporting_containers, environment. fix_storage_permissions] for fn in steps: fn() if not quiet: write('.') if not quiet: write('\n') except: if not quiet: print raise return finish_init(environment, start_web, create_sysadmin, log_syslog= log_syslog, do_install=not no_install, quiet=quiet, site_url= site_url, interactive=interactive, init_db=init_db)
init
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(opts, no_install=False, quiet=False):\n if opts['--address'] and is_boot2docker():\n raise DatacatsError('Cannot specify address on boot2docker.')\n environment_dir = opts['ENVIRONMENT_DIR']\n port = opts['PORT']\n address = opts['--address']\n start_web = not opts['--image-only']\n create_sysadmin = not opts['--no-sysadmin']\n site_name = opts['--site']\n site_url = opts['--site-url']\n interactive = opts['--interactive']\n init_db = not opts['--no-init-db']\n environment_dir = abspath(environment_dir or '.')\n log_syslog = opts['--syslog']\n environment = Environment.load(environment_dir, site_name)\n if address:\n environment.address = address\n if port:\n environment.port = int(port)\n if site_url:\n environment.site_url = site_url\n try:\n if environment.sites and site_name in environment.sites:\n raise DatacatsError('Site named {0} already exists.'.format(\n site_name))\n making_full_environment = not environment.data_exists()\n if not quiet:\n write(\n 'Creating environment {0}/{1} from existing environment directory \"{0}\"'\n .format(environment.name, environment.site_name))\n steps = [lambda : environment.create_directories(create_project_dir\n =False)] + ([environment.save, environment.create_virtualenv] if\n making_full_environment else []) + [environment.save_site,\n environment.start_supporting_containers, environment.\n fix_storage_permissions]\n for fn in steps:\n fn()\n if not quiet:\n write('.')\n if not quiet:\n write('\\n')\n except:\n if not quiet:\n print\n raise\n return finish_init(environment, start_web, create_sysadmin, log_syslog=\n log_syslog, do_install=not no_install, quiet=quiet, site_url=\n site_url, interactive=interactive, init_db=init_db)\n", "role": "user" }, { "content": "init", "role": "assistant" } ]
def x(environment, start_web, create_sysadmin, log_syslog=False, do_install =True, quiet=False, site_url=None, interactive=False, init_db=True): if not init_db: start_web = False create_sysadmin = False if do_install: install_all(environment, False, verbose=False, quiet=quiet) if init_db: if not quiet: write('Initializing database') environment.install_postgis_sql() environment.ckan_db_init() if not quiet: write('\n') if site_url: try: site_url = site_url.format(address=environment.address, port= environment.port) environment.site_url = site_url environment.save_site(False) except (KeyError, IndexError, ValueError) as e: raise DatacatsError('Could not parse site_url: {}'.format(e)) if start_web: environment.start_ckan(log_syslog=log_syslog) if not quiet and not interactive: write('Starting web server at {0} ...\n'.format(environment. web_address())) if create_sysadmin: try: adminpw = confirm_password() environment.create_admin_set_password(adminpw) except KeyboardInterrupt: print if not start_web: environment.stop_supporting_containers()
finish_init
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(environment, start_web, create_sysadmin, log_syslog=False, do_install\n =True, quiet=False, site_url=None, interactive=False, init_db=True):\n if not init_db:\n start_web = False\n create_sysadmin = False\n if do_install:\n install_all(environment, False, verbose=False, quiet=quiet)\n if init_db:\n if not quiet:\n write('Initializing database')\n environment.install_postgis_sql()\n environment.ckan_db_init()\n if not quiet:\n write('\\n')\n if site_url:\n try:\n site_url = site_url.format(address=environment.address, port=\n environment.port)\n environment.site_url = site_url\n environment.save_site(False)\n except (KeyError, IndexError, ValueError) as e:\n raise DatacatsError('Could not parse site_url: {}'.format(e))\n if start_web:\n environment.start_ckan(log_syslog=log_syslog)\n if not quiet and not interactive:\n write('Starting web server at {0} ...\\n'.format(environment.\n web_address()))\n if create_sysadmin:\n try:\n adminpw = confirm_password()\n environment.create_admin_set_password(adminpw)\n except KeyboardInterrupt:\n print\n if not start_web:\n environment.stop_supporting_containers()\n", "role": "user" }, { "content": "finish_init", "role": "assistant" } ]
def x(self): config = self.profiledir + '/config' if not isdir(self.profiledir): makedirs(self.profiledir) cp = SafeConfigParser() cp.add_section('ssh') cp.set('ssh', 'private_key', self.ssh_private_key) cp.set('ssh', 'public_key', self.ssh_public_key) with open(config, 'w') as cfile: cp.write(cfile)
save
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n config = self.profiledir + '/config'\n if not isdir(self.profiledir):\n makedirs(self.profiledir)\n cp = SafeConfigParser()\n cp.add_section('ssh')\n cp.set('ssh', 'private_key', self.ssh_private_key)\n cp.set('ssh', 'public_key', self.ssh_public_key)\n with open(config, 'w') as cfile:\n cp.write(cfile)\n", "role": "user" }, { "content": "save", "role": "assistant" } ]
def x(self): web_command(command=['ssh-keygen', '-q', '-t', 'rsa', '-N', '', '-C', 'datacats generated {0}@{1}'.format(getuser(), gethostname()), '-f', '/output/id_rsa'], rw={self.profiledir: '/output'})
generate_ssh_key
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n web_command(command=['ssh-keygen', '-q', '-t', 'rsa', '-N', '', '-C',\n 'datacats generated {0}@{1}'.format(getuser(), gethostname()), '-f',\n '/output/id_rsa'], rw={self.profiledir: '/output'})\n", "role": "user" }, { "content": "generate_ssh_key", "role": "assistant" } ]
def x(self, environment, target_name): remote_server_command(['ssh', environment.deploy_target, 'create', target_name], environment, self, clean_up=True)
create
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, environment, target_name):\n remote_server_command(['ssh', environment.deploy_target, 'create',\n target_name], environment, self, clean_up=True)\n", "role": "user" }, { "content": "create", "role": "assistant" } ]
def x(self, environment, target_name, password): try: remote_server_command(['ssh', environment.deploy_target, 'admin_password', target_name, password], environment, self, clean_up=True) return True except WebCommandError: return False
admin_password
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, environment, target_name, password):\n try:\n remote_server_command(['ssh', environment.deploy_target,\n 'admin_password', target_name, password], environment, self,\n clean_up=True)\n return True\n except WebCommandError:\n return False\n", "role": "user" }, { "content": "admin_password", "role": "assistant" } ]
def x(self, environment, target_name, stream_output=None): try: remote_server_command(['rsync', '-lrv', '--safe-links', '--munge-links', '--delete', '--inplace', '--chmod=ugo=rwX', '--exclude=.datacats-environment', '--exclude=.git', '/project/.', environment.deploy_target + ':' + target_name], environment, self, include_project_dir=True, stream_output= stream_output, clean_up=True) except WebCommandError as e: raise DatacatsError( """Unable to deploy `{0}` to remote server for some reason: datacats was not able to copy data to the remote server""" .format((target_name,)), parent_exception=e) try: remote_server_command(['ssh', environment.deploy_target, 'install', target_name], environment, self, clean_up=True) return True except WebCommandError as e: raise DatacatsError( """Unable to deploy `{0}` to remote server for some reason: datacats copied data to the server but failed to register (or `install`) the new catalog""" .format((target_name,)), parent_exception=e)
deploy
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, environment, target_name, stream_output=None):\n try:\n remote_server_command(['rsync', '-lrv', '--safe-links',\n '--munge-links', '--delete', '--inplace', '--chmod=ugo=rwX',\n '--exclude=.datacats-environment', '--exclude=.git',\n '/project/.', environment.deploy_target + ':' + target_name],\n environment, self, include_project_dir=True, stream_output=\n stream_output, clean_up=True)\n except WebCommandError as e:\n raise DatacatsError(\n \"\"\"Unable to deploy `{0}` to remote server for some reason:\n datacats was not able to copy data to the remote server\"\"\"\n .format((target_name,)), parent_exception=e)\n try:\n remote_server_command(['ssh', environment.deploy_target, 'install',\n target_name], environment, self, clean_up=True)\n return True\n except WebCommandError as e:\n raise DatacatsError(\n \"\"\"Unable to deploy `{0}` to remote server for some reason:\ndatacats copied data to the server but failed to register\n(or `install`) the new catalog\"\"\"\n .format((target_name,)), parent_exception=e)\n", "role": "user" }, { "content": "deploy", "role": "assistant" } ]
def x(n, batch_size): b = n // batch_size if n % batch_size > 0: b += 1 return b
num_batches
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(n, batch_size):\n b = n // batch_size\n if n % batch_size > 0:\n b += 1\n return b\n", "role": "user" }, { "content": "num_batches", "role": "assistant" } ]
def x(self): if self.repeats == -1: return np.inf else: return self.length * self.repeats
num_indices_generated
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.repeats == -1:\n return np.inf\n else:\n return self.length * self.repeats\n", "role": "user" }, { "content": "num_indices_generated", "role": "assistant" } ]
def x(self, batch_size): if self.repeats == 1: for i in range(0, self.length, batch_size): yield np.arange(i, min(i + batch_size, self.length)) else: repeats = self.repeats i = 0 while True: j = i + batch_size if j <= self.length: yield np.arange(i, j) i = j elif j <= self.length * 2: if repeats != -1: repeats -= 1 if repeats == 0: if i < self.length: yield np.arange(i, self.length) break k = batch_size - (self.length - i) yield np.append(np.arange(i, self.length), np.arange(0, k), axis=0) i = k else: batch_ndx = np.arange(0) while len(batch_ndx) < batch_size: k = min(batch_size - len(batch_ndx), self.length - i) batch_ndx = np.append(batch_ndx, np.arange(i, i + k), axis=0) i += k if i >= self.length: i -= self.length if repeats != -1: repeats -= 1 if repeats == 0: break if len(batch_ndx) > 0: yield batch_ndx if repeats == 0: break
in_order_indices_batch_iterator
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, batch_size):\n if self.repeats == 1:\n for i in range(0, self.length, batch_size):\n yield np.arange(i, min(i + batch_size, self.length))\n else:\n repeats = self.repeats\n i = 0\n while True:\n j = i + batch_size\n if j <= self.length:\n yield np.arange(i, j)\n i = j\n elif j <= self.length * 2:\n if repeats != -1:\n repeats -= 1\n if repeats == 0:\n if i < self.length:\n yield np.arange(i, self.length)\n break\n k = batch_size - (self.length - i)\n yield np.append(np.arange(i, self.length), np.arange(0, k),\n axis=0)\n i = k\n else:\n batch_ndx = np.arange(0)\n while len(batch_ndx) < batch_size:\n k = min(batch_size - len(batch_ndx), self.length - i)\n batch_ndx = np.append(batch_ndx, np.arange(i, i + k),\n axis=0)\n i += k\n if i >= self.length:\n i -= self.length\n if repeats != -1:\n repeats -= 1\n if repeats == 0:\n break\n if len(batch_ndx) > 0:\n yield batch_ndx\n if repeats == 0:\n break\n", "role": "user" }, { "content": "in_order_indices_batch_iterator", "role": "assistant" } ]
def x(self, batch_size, shuffle_rng): if self.repeats == 1: indices = shuffle_rng.permutation(self.length) for i in range(0, self.length, batch_size): yield indices[i:i + batch_size] else: repeats = self.repeats indices = shuffle_rng.permutation(self.length) i = 0 while True: j = i + batch_size if j <= self.length: yield indices[i:j] i = j else: batch_ndx = np.arange(0) while len(batch_ndx) < batch_size: k = min(batch_size - len(batch_ndx), self.length - i) batch_ndx = np.append(batch_ndx, indices[i:i + k], axis=0) i += k if i >= self.length: indices = shuffle_rng.permutation(self.length) i -= self.length if repeats != -1: repeats -= 1 if repeats == 0: break if len(batch_ndx) > 0: yield batch_ndx if repeats == 0: break
shuffled_indices_batch_iterator
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, batch_size, shuffle_rng):\n if self.repeats == 1:\n indices = shuffle_rng.permutation(self.length)\n for i in range(0, self.length, batch_size):\n yield indices[i:i + batch_size]\n else:\n repeats = self.repeats\n indices = shuffle_rng.permutation(self.length)\n i = 0\n while True:\n j = i + batch_size\n if j <= self.length:\n yield indices[i:j]\n i = j\n else:\n batch_ndx = np.arange(0)\n while len(batch_ndx) < batch_size:\n k = min(batch_size - len(batch_ndx), self.length - i)\n batch_ndx = np.append(batch_ndx, indices[i:i + k], axis=0)\n i += k\n if i >= self.length:\n indices = shuffle_rng.permutation(self.length)\n i -= self.length\n if repeats != -1:\n repeats -= 1\n if repeats == 0:\n break\n if len(batch_ndx) > 0:\n yield batch_ndx\n if repeats == 0:\n break\n", "role": "user" }, { "content": "shuffled_indices_batch_iterator", "role": "assistant" } ]
def x(self, batch_size, shuffle_rng): while True: yield shuffle_rng.choice(len(self.weights), size=(batch_size,), p= self.weights)
shuffled_indices_batch_iterator
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, batch_size, shuffle_rng):\n while True:\n yield shuffle_rng.choice(len(self.weights), size=(batch_size,), p=\n self.weights)\n", "role": "user" }, { "content": "shuffled_indices_batch_iterator", "role": "assistant" } ]
def x(y): h = np.bincount(y) cls_weight = 1.0 / (h.astype(float) * len(np.nonzero(h)[0])) cls_weight[np.isnan(cls_weight)] = 0.0 sample_weight = cls_weight[y] return sample_weight
class_balancing_sample_weights
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(y):\n h = np.bincount(y)\n cls_weight = 1.0 / (h.astype(float) * len(np.nonzero(h)[0]))\n cls_weight[np.isnan(cls_weight)] = 0.0\n sample_weight = cls_weight[y]\n return sample_weight\n", "role": "user" }, { "content": "class_balancing_sample_weights", "role": "assistant" } ]
def x(self, batch_size, shuffle_rng): while True: yield shuffle_rng.choice(self.indices, size=(batch_size,), p=self. sub_weights)
shuffled_indices_batch_iterator
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, batch_size, shuffle_rng):\n while True:\n yield shuffle_rng.choice(self.indices, size=(batch_size,), p=self.\n sub_weights)\n", "role": "user" }, { "content": "shuffled_indices_batch_iterator", "role": "assistant" } ]
def x(y, indices): weights = WeightedSampler.class_balancing_sample_weights(y[indices]) return WeightedSubsetSampler(weights, indices=indices)
class_balancing_sampler
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(y, indices):\n weights = WeightedSampler.class_balancing_sample_weights(y[indices])\n return WeightedSubsetSampler(weights, indices=indices)\n", "role": "user" }, { "content": "class_balancing_sampler", "role": "assistant" } ]
def x(container, url, timeout): start = time.time() remaining = timeout while True: remaining = start + timeout - time.time() if remaining < 0: raise ServiceTimeout try: response = get(url, timeout=min(remaining, REQUEST_TIMEOUT_SECONDS) ) if 500 <= response.status_code < 600: return False return True except (ConnectionError, Timeout): pass if not inspect_container(container)['State']['Running']: return False remaining = start + timeout - time.time() delay = max(0, min(RETRY_DELAY_SECONDS, remaining)) time.sleep(delay) raise ServiceTimeout
wait_for_service_available
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(container, url, timeout):\n start = time.time()\n remaining = timeout\n while True:\n remaining = start + timeout - time.time()\n if remaining < 0:\n raise ServiceTimeout\n try:\n response = get(url, timeout=min(remaining, REQUEST_TIMEOUT_SECONDS)\n )\n if 500 <= response.status_code < 600:\n return False\n return True\n except (ConnectionError, Timeout):\n pass\n if not inspect_container(container)['State']['Running']:\n return False\n remaining = start + timeout - time.time()\n delay = max(0, min(RETRY_DELAY_SECONDS, remaining))\n time.sleep(delay)\n raise ServiceTimeout\n", "role": "user" }, { "content": "wait_for_service_available", "role": "assistant" } ]
def x(filename): if os.path.isabs(filename): return filename else: return os.path.join(get_data_dir(), filename)
get_data_path
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(filename):\n if os.path.isabs(filename):\n return filename\n else:\n return os.path.join(get_data_dir(), filename)\n", "role": "user" }, { "content": "get_data_path", "role": "assistant" } ]
def x(path, source_url): dir_path = os.path.dirname(path) if not os.path.exists(dir_path): os.makedirs(dir_path) if not os.path.exists(path): print('Downloading {} to {}'.format(source_url, path)) filename = source_url.split('/')[-1] def x(count, block_size, total_size): sys.stdout.write('\rDownloading {} {:.2%}'.format(filename, float(count * block_size) / float(total_size))) sys.stdout.flush() try: urlretrieve(source_url, path, reporthook=_progress) except: sys.stdout.write('\r') if os.path.exists(path): os.remove(path) raise sys.stdout.write('\r') return path
download
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path, source_url):\n dir_path = os.path.dirname(path)\n if not os.path.exists(dir_path):\n os.makedirs(dir_path)\n if not os.path.exists(path):\n print('Downloading {} to {}'.format(source_url, path))\n filename = source_url.split('/')[-1]\n\n def x(count, block_size, total_size):\n sys.stdout.write('\\rDownloading {} {:.2%}'.format(filename, \n float(count * block_size) / float(total_size)))\n sys.stdout.flush()\n try:\n urlretrieve(source_url, path, reporthook=_progress)\n except:\n sys.stdout.write('\\r')\n if os.path.exists(path):\n os.remove(path)\n raise\n sys.stdout.write('\\r')\n return path\n", "role": "user" }, { "content": "download", "role": "assistant" } ]
def x(path): hasher = hashlib.sha256() with open(path, 'rb') as f: for chunk in iter(lambda : f.read(10 * 1024 * 1024), b''): hasher.update(chunk) return hasher.hexdigest()
compute_sha256
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path):\n hasher = hashlib.sha256()\n with open(path, 'rb') as f:\n for chunk in iter(lambda : f.read(10 * 1024 * 1024), b''):\n hasher.update(chunk)\n return hasher.hexdigest()\n", "role": "user" }, { "content": "compute_sha256", "role": "assistant" } ]
def x(path, sha256): if not os.path.isfile(path): return False digest = compute_sha256(path) if sha256 is None: print('SHA-256 of {}:'.format(path)) print(' "{}"'.format(digest)) elif digest != sha256: return False return True
verify_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path, sha256):\n if not os.path.isfile(path):\n return False\n digest = compute_sha256(path)\n if sha256 is None:\n print('SHA-256 of {}:'.format(path))\n print(' \"{}\"'.format(digest))\n elif digest != sha256:\n return False\n return True\n", "role": "user" }, { "content": "verify_file", "role": "assistant" } ]
def x(path, source_url, sha256): if os.path.exists(path): if sha256 is None: print('The SHA-256 of {} is "{}"'.format(path, compute_sha256( path))) return path unverified_path = path + '.unverified' for i in range(_MAX_DOWNLOAD_TRIES): try: unverified_path = download(unverified_path, source_url) except Exception as e: print( 'Download of {} unsuccessful; error {}; deleting and re-trying...' .format(source_url, e)) if os.path.exists(unverified_path): os.remove(unverified_path) else: if os.path.exists(unverified_path): if verify_file(unverified_path, sha256): os.rename(unverified_path, path) return path else: print( 'Download of {} unsuccessful; verification failed; deleting and re-trying...' .format(source_url)) os.remove(unverified_path) print('Did not succeed in downloading {} (tried {} times)'.format( source_url, _MAX_DOWNLOAD_TRIES)) return None
download_and_verify
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path, source_url, sha256):\n if os.path.exists(path):\n if sha256 is None:\n print('The SHA-256 of {} is \"{}\"'.format(path, compute_sha256(\n path)))\n return path\n unverified_path = path + '.unverified'\n for i in range(_MAX_DOWNLOAD_TRIES):\n try:\n unverified_path = download(unverified_path, source_url)\n except Exception as e:\n print(\n 'Download of {} unsuccessful; error {}; deleting and re-trying...'\n .format(source_url, e))\n if os.path.exists(unverified_path):\n os.remove(unverified_path)\n else:\n if os.path.exists(unverified_path):\n if verify_file(unverified_path, sha256):\n os.rename(unverified_path, path)\n return path\n else:\n print(\n 'Download of {} unsuccessful; verification failed; deleting and re-trying...'\n .format(source_url))\n os.remove(unverified_path)\n print('Did not succeed in downloading {} (tried {} times)'.format(\n source_url, _MAX_DOWNLOAD_TRIES))\n return None\n", "role": "user" }, { "content": "download_and_verify", "role": "assistant" } ]
def x(path, source_path, sha256): if os.path.exists(path): if sha256 is None: print('The SHA-256 of {} is "{}"'.format(path, compute_sha256( path))) return path if not os.path.exists(source_path): return None unverified_path = path + '.unverified' dir_path = os.path.dirname(path) if not os.path.exists(dir_path): os.makedirs(dir_path) shutil.copy(source_path, unverified_path) if os.path.exists(unverified_path): if verify_file(unverified_path, sha256): os.rename(unverified_path, path) return path else: print('SHA verification of file {} failed'.format(source_path)) os.remove(unverified_path) return None
copy_and_verify
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path, source_path, sha256):\n if os.path.exists(path):\n if sha256 is None:\n print('The SHA-256 of {} is \"{}\"'.format(path, compute_sha256(\n path)))\n return path\n if not os.path.exists(source_path):\n return None\n unverified_path = path + '.unverified'\n dir_path = os.path.dirname(path)\n if not os.path.exists(dir_path):\n os.makedirs(dir_path)\n shutil.copy(source_path, unverified_path)\n if os.path.exists(unverified_path):\n if verify_file(unverified_path, sha256):\n os.rename(unverified_path, path)\n return path\n else:\n print('SHA verification of file {} failed'.format(source_path))\n os.remove(unverified_path)\n return None\n", "role": "user" }, { "content": "copy_and_verify", "role": "assistant" } ]
def x(name, target): setupdir = '{0}/ckanext-{1}theme'.format(target, name) extdir = setupdir + '/ckanext/{0}theme'.format(name) templatedir = extdir + '/templates/' staticdir = extdir + '/static/datacats' makedirs(templatedir + '/home/snippets') makedirs(staticdir) here = dirname(__file__) copyfile(here + '/images/chart.png', staticdir + '/chart.png') copyfile(here + '/images/datacats-footer.png', staticdir + '/datacats-footer.png') filecontents = [(setupdir + '/setup.py', SETUP_PY), (setupdir + '/.gitignore', DOT_GITIGNORE), (setupdir + '/ckanext/__init__.py', NAMESPACE_PACKAGE), (extdir + '/__init__.py', ''), (extdir + '/plugins.py', PLUGINS_PY), (templatedir + '/home/snippets/promoted.html', PROMOTED_SNIPPET), (templatedir + '/footer.html', FOOTER_HTML)] for filename, content in filecontents: with open(filename, 'w') as f: f.write(content.replace('##name##', name))
ckan_extension_template
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(name, target):\n setupdir = '{0}/ckanext-{1}theme'.format(target, name)\n extdir = setupdir + '/ckanext/{0}theme'.format(name)\n templatedir = extdir + '/templates/'\n staticdir = extdir + '/static/datacats'\n makedirs(templatedir + '/home/snippets')\n makedirs(staticdir)\n here = dirname(__file__)\n copyfile(here + '/images/chart.png', staticdir + '/chart.png')\n copyfile(here + '/images/datacats-footer.png', staticdir +\n '/datacats-footer.png')\n filecontents = [(setupdir + '/setup.py', SETUP_PY), (setupdir +\n '/.gitignore', DOT_GITIGNORE), (setupdir + '/ckanext/__init__.py',\n NAMESPACE_PACKAGE), (extdir + '/__init__.py', ''), (extdir +\n '/plugins.py', PLUGINS_PY), (templatedir +\n '/home/snippets/promoted.html', PROMOTED_SNIPPET), (templatedir +\n '/footer.html', FOOTER_HTML)]\n for filename, content in filecontents:\n with open(filename, 'w') as f:\n f.write(content.replace('##name##', name))\n", "role": "user" }, { "content": "ckan_extension_template", "role": "assistant" } ]
def x(environment, opts): environment.require_data() environment.start_supporting_containers() return environment.interactive_shell(opts['COMMAND'], detach=opts[ '--detach'])
shell
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(environment, opts):\n environment.require_data()\n environment.start_supporting_containers()\n return environment.interactive_shell(opts['COMMAND'], detach=opts[\n '--detach'])\n", "role": "user" }, { "content": "shell", "role": "assistant" } ]
def x(opts): environment = Environment.load('.') environment.require_data() environment.start_supporting_containers() if not opts['COMMAND']: opts['COMMAND'] = ['--', 'help'] assert opts['COMMAND'][0] == '--' return environment.interactive_shell(opts['COMMAND'][1:], paster=True, detach=opts['--detach'])
paster
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(opts):\n environment = Environment.load('.')\n environment.require_data()\n environment.start_supporting_containers()\n if not opts['COMMAND']:\n opts['COMMAND'] = ['--', 'help']\n assert opts['COMMAND'][0] == '--'\n return environment.interactive_shell(opts['COMMAND'][1:], paster=True,\n detach=opts['--detach'])\n", "role": "user" }, { "content": "paster", "role": "assistant" } ]
def x(site_name, sitedir, srcdir, port, address, site_url, passwords): cp = ConfigParser.SafeConfigParser() cp.read([srcdir + '/.datacats-environment']) section_name = 'site_' + site_name if not cp.has_section(section_name): cp.add_section(section_name) cp.set(section_name, 'port', str(port)) if address: cp.set(section_name, 'address', address) if site_url: cp.set(section_name, 'site_url', site_url) with open(srcdir + '/.datacats-environment', 'w') as config: cp.write(config) cp = ConfigParser.SafeConfigParser() cp.add_section('passwords') for n in sorted(passwords): cp.set('passwords', n.lower(), passwords[n]) with open(sitedir + '/passwords.ini', 'w') as config: cp.write(config)
save_new_site
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(site_name, sitedir, srcdir, port, address, site_url, passwords):\n cp = ConfigParser.SafeConfigParser()\n cp.read([srcdir + '/.datacats-environment'])\n section_name = 'site_' + site_name\n if not cp.has_section(section_name):\n cp.add_section(section_name)\n cp.set(section_name, 'port', str(port))\n if address:\n cp.set(section_name, 'address', address)\n if site_url:\n cp.set(section_name, 'site_url', site_url)\n with open(srcdir + '/.datacats-environment', 'w') as config:\n cp.write(config)\n cp = ConfigParser.SafeConfigParser()\n cp.add_section('passwords')\n for n in sorted(passwords):\n cp.set('passwords', n.lower(), passwords[n])\n with open(sitedir + '/passwords.ini', 'w') as config:\n cp.write(config)\n", "role": "user" }, { "content": "save_new_site", "role": "assistant" } ]
def x(name, datadir, srcdir, ckan_version, deploy_target=None, always_prod= False): with open(datadir + '/.version', 'w') as f: f.write('2') cp = ConfigParser.SafeConfigParser() cp.read(srcdir + '/.datacats-environment') if not cp.has_section('datacats'): cp.add_section('datacats') cp.set('datacats', 'name', name) cp.set('datacats', 'ckan_version', ckan_version) if deploy_target: if not cp.has_section('deploy'): cp.add_section('deploy') cp.set('deploy', 'target', deploy_target) if always_prod: cp.set('datacats', 'always_prod', 'true') with open(srcdir + '/.datacats-environment', 'w') as config: cp.write(config) save_srcdir_location(datadir, srcdir)
save_new_environment
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(name, datadir, srcdir, ckan_version, deploy_target=None, always_prod=\n False):\n with open(datadir + '/.version', 'w') as f:\n f.write('2')\n cp = ConfigParser.SafeConfigParser()\n cp.read(srcdir + '/.datacats-environment')\n if not cp.has_section('datacats'):\n cp.add_section('datacats')\n cp.set('datacats', 'name', name)\n cp.set('datacats', 'ckan_version', ckan_version)\n if deploy_target:\n if not cp.has_section('deploy'):\n cp.add_section('deploy')\n cp.set('deploy', 'target', deploy_target)\n if always_prod:\n cp.set('datacats', 'always_prod', 'true')\n with open(srcdir + '/.datacats-environment', 'w') as config:\n cp.write(config)\n save_srcdir_location(datadir, srcdir)\n", "role": "user" }, { "content": "save_new_environment", "role": "assistant" } ]
def x(environment_name=None, data_only=False): docker.require_images() if environment_name is None: environment_name = '.' extension_dir = 'ckan' if validate.valid_name(environment_name) and path.isdir(path.expanduser ('~/.datacats/' + environment_name)): datadir = path.expanduser('~/.datacats/' + environment_name) with open(datadir + '/project-dir') as pd: srcdir = pd.read() if not data_only and not path.exists(srcdir + '/.datacats-environment' ): raise DatacatsError( 'Environment data found but environment directory is missing. Try again from the new environment directory location or remove the environment data with "datacats purge"' ) return srcdir, extension_dir, datadir srcdir = path.abspath(environment_name) if not path.isdir(srcdir): raise DatacatsError('No environment found with that name') wd = srcdir oldwd = None while not path.exists(wd + '/.datacats-environment'): oldwd = wd wd, _ = path.split(wd) if wd == oldwd: raise DatacatsError('Environment not found in {0} or above'. format(srcdir)) srcdir = wd if oldwd: _, extension_dir = path.split(oldwd) return srcdir, extension_dir, None
find_environment_dirs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(environment_name=None, data_only=False):\n docker.require_images()\n if environment_name is None:\n environment_name = '.'\n extension_dir = 'ckan'\n if validate.valid_name(environment_name) and path.isdir(path.expanduser\n ('~/.datacats/' + environment_name)):\n datadir = path.expanduser('~/.datacats/' + environment_name)\n with open(datadir + '/project-dir') as pd:\n srcdir = pd.read()\n if not data_only and not path.exists(srcdir + '/.datacats-environment'\n ):\n raise DatacatsError(\n 'Environment data found but environment directory is missing. Try again from the new environment directory location or remove the environment data with \"datacats purge\"'\n )\n return srcdir, extension_dir, datadir\n srcdir = path.abspath(environment_name)\n if not path.isdir(srcdir):\n raise DatacatsError('No environment found with that name')\n wd = srcdir\n oldwd = None\n while not path.exists(wd + '/.datacats-environment'):\n oldwd = wd\n wd, _ = path.split(wd)\n if wd == oldwd:\n raise DatacatsError('Environment not found in {0} or above'.\n format(srcdir))\n srcdir = wd\n if oldwd:\n _, extension_dir = path.split(oldwd)\n return srcdir, extension_dir, None\n", "role": "user" }, { "content": "find_environment_dirs", "role": "assistant" } ]
def x(srcdir, datadir=None, allow_old=False): cp = ConfigParser.SafeConfigParser() try: cp.read([srcdir + '/.datacats-environment']) except ConfigParser.Error: raise DatacatsError('Error reading environment information') name = cp.get('datacats', 'name') if datadir: save_srcdir_location(datadir, srcdir) else: datadir = path.expanduser('~/.datacats/' + name) if migrate.needs_format_conversion(datadir) and not allow_old: raise DatacatsError( 'This environment uses an old format. You must migrate to the new format. To do so, use the "datacats migrate" command.' ) if migrate.is_locked(datadir): raise DatacatsError( """Migration in progress, cannot continue. If you interrupted a migration, you should attempt manual recovery or contact us by filing an issue at http://github.com/datacats/datacats. As a last resort, you could delete all your stored data and create a new environment by running "datacats purge" followed by "datacats init".""" ) ckan_version = cp.get('datacats', 'ckan_version') try: always_prod = cp.getboolean('datacats', 'always_prod') except ConfigParser.NoOptionError: always_prod = False try: extra_containers = cp.get('datacats', 'extra_containers').split(' ') except ConfigParser.NoOptionError: extra_containers = () try: deploy_target = cp.get('deploy', 'remote_server_user') + '@' + cp.get( 'deploy', 'remote_server') except (ConfigParser.NoOptionError, ConfigParser.NoSectionError): deploy_target = DEFAULT_REMOTE_SERVER_TARGET try: remote_server_key = cp.get('deploy', 'remote_server_key') except (ConfigParser.NoOptionError, ConfigParser.NoSectionError): remote_server_key = None return (datadir, name, ckan_version, always_prod, deploy_target, remote_server_key, extra_containers)
load_environment
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(srcdir, datadir=None, allow_old=False):\n cp = ConfigParser.SafeConfigParser()\n try:\n cp.read([srcdir + '/.datacats-environment'])\n except ConfigParser.Error:\n raise DatacatsError('Error reading environment information')\n name = cp.get('datacats', 'name')\n if datadir:\n save_srcdir_location(datadir, srcdir)\n else:\n datadir = path.expanduser('~/.datacats/' + name)\n if migrate.needs_format_conversion(datadir) and not allow_old:\n raise DatacatsError(\n 'This environment uses an old format. You must migrate to the new format. To do so, use the \"datacats migrate\" command.'\n )\n if migrate.is_locked(datadir):\n raise DatacatsError(\n \"\"\"Migration in progress, cannot continue.\nIf you interrupted a migration, you should attempt manual recovery or contact us by filing an issue at http://github.com/datacats/datacats.\nAs a last resort, you could delete all your stored data and create a new environment by running \"datacats purge\" followed by \"datacats init\".\"\"\"\n )\n ckan_version = cp.get('datacats', 'ckan_version')\n try:\n always_prod = cp.getboolean('datacats', 'always_prod')\n except ConfigParser.NoOptionError:\n always_prod = False\n try:\n extra_containers = cp.get('datacats', 'extra_containers').split(' ')\n except ConfigParser.NoOptionError:\n extra_containers = ()\n try:\n deploy_target = cp.get('deploy', 'remote_server_user') + '@' + cp.get(\n 'deploy', 'remote_server')\n except (ConfigParser.NoOptionError, ConfigParser.NoSectionError):\n deploy_target = DEFAULT_REMOTE_SERVER_TARGET\n try:\n remote_server_key = cp.get('deploy', 'remote_server_key')\n except (ConfigParser.NoOptionError, ConfigParser.NoSectionError):\n remote_server_key = None\n return (datadir, name, ckan_version, always_prod, deploy_target,\n remote_server_key, extra_containers)\n", "role": "user" }, { "content": "load_environment", "role": "assistant" } ]
def x(srcdir, datadir, site_name=None): if site_name is None: site_name = 'primary' if not validate.valid_name(site_name): raise DatacatsError('{} is not a valid site name.'.format(site_name)) cp = ConfigParser.SafeConfigParser() try: cp.read([srcdir + '/.datacats-environment']) except ConfigParser.Error: raise DatacatsError('Error reading environment information') site_section = 'site_' + site_name try: port = cp.getint(site_section, 'port') except (ConfigParser.NoOptionError, ConfigParser.NoSectionError): port = None try: address = cp.get(site_section, 'address') except (ConfigParser.NoOptionError, ConfigParser.NoSectionError): address = None try: site_url = cp.get(site_section, 'site_url') except (ConfigParser.NoOptionError, ConfigParser.NoSectionError): site_url = None passwords = {} cp = ConfigParser.SafeConfigParser() cp.read(datadir + '/sites/' + site_name + '/passwords.ini') try: pw_options = cp.options('passwords') except ConfigParser.NoSectionError: pw_options = [] for n in pw_options: passwords[n.upper()] = cp.get('passwords', n) return port, address, site_url, passwords
load_site
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(srcdir, datadir, site_name=None):\n if site_name is None:\n site_name = 'primary'\n if not validate.valid_name(site_name):\n raise DatacatsError('{} is not a valid site name.'.format(site_name))\n cp = ConfigParser.SafeConfigParser()\n try:\n cp.read([srcdir + '/.datacats-environment'])\n except ConfigParser.Error:\n raise DatacatsError('Error reading environment information')\n site_section = 'site_' + site_name\n try:\n port = cp.getint(site_section, 'port')\n except (ConfigParser.NoOptionError, ConfigParser.NoSectionError):\n port = None\n try:\n address = cp.get(site_section, 'address')\n except (ConfigParser.NoOptionError, ConfigParser.NoSectionError):\n address = None\n try:\n site_url = cp.get(site_section, 'site_url')\n except (ConfigParser.NoOptionError, ConfigParser.NoSectionError):\n site_url = None\n passwords = {}\n cp = ConfigParser.SafeConfigParser()\n cp.read(datadir + '/sites/' + site_name + '/passwords.ini')\n try:\n pw_options = cp.options('passwords')\n except ConfigParser.NoSectionError:\n pw_options = []\n for n in pw_options:\n passwords[n.upper()] = cp.get('passwords', n)\n return port, address, site_url, passwords\n", "role": "user" }, { "content": "load_site", "role": "assistant" } ]
def x(srcpath, site_name, ckan_version): docker.require_images() workdir, name = path.split(path.abspath(path.expanduser(srcpath))) if not validate.valid_name(name): raise DatacatsError( 'Please choose an environment name starting with a letter and including only lowercase letters and digits' ) if not path.isdir(workdir): raise DatacatsError('Parent directory for environment does not exist') datadir = path.expanduser('~/.datacats/' + name) sitedir = datadir + '/sites/' + site_name if path.isdir(datadir): with open(datadir + '/project-dir') as pd: srcdir = pd.read() else: srcdir = workdir + '/' + name if ckan_version not in SUPPORTED_PRELOADS: raise DatacatsError( """Datacats does not currently support CKAN version {}. Versions that are currently supported are: {}""" .format(ckan_version, ', '.join(SUPPORTED_PRELOADS))) preload_name = str(ckan_version) downloaded_versions = [tag for tag in docker.get_tags('datacats/ckan')] if ckan_version not in downloaded_versions: retrying_pull_image('datacats/ckan:{}'.format(preload_name)) if path.isdir(sitedir): raise DatacatsError('Site data directory {0} already exists'.format (sitedir)) if path.isdir(srcdir) and not path.isdir(datadir): raise DatacatsError( """Environment directory exists, but data directory does not. If you simply want to recreate the data directory, run "datacats init" in the environment directory.""" ) return name, datadir, srcdir
new_environment_check
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(srcpath, site_name, ckan_version):\n docker.require_images()\n workdir, name = path.split(path.abspath(path.expanduser(srcpath)))\n if not validate.valid_name(name):\n raise DatacatsError(\n 'Please choose an environment name starting with a letter and including only lowercase letters and digits'\n )\n if not path.isdir(workdir):\n raise DatacatsError('Parent directory for environment does not exist')\n datadir = path.expanduser('~/.datacats/' + name)\n sitedir = datadir + '/sites/' + site_name\n if path.isdir(datadir):\n with open(datadir + '/project-dir') as pd:\n srcdir = pd.read()\n else:\n srcdir = workdir + '/' + name\n if ckan_version not in SUPPORTED_PRELOADS:\n raise DatacatsError(\n \"\"\"Datacats does not currently support CKAN version {}.\nVersions that are currently supported are: {}\"\"\"\n .format(ckan_version, ', '.join(SUPPORTED_PRELOADS)))\n preload_name = str(ckan_version)\n downloaded_versions = [tag for tag in docker.get_tags('datacats/ckan')]\n if ckan_version not in downloaded_versions:\n retrying_pull_image('datacats/ckan:{}'.format(preload_name))\n if path.isdir(sitedir):\n raise DatacatsError('Site data directory {0} already exists'.format\n (sitedir))\n if path.isdir(srcdir) and not path.isdir(datadir):\n raise DatacatsError(\n \"\"\"Environment directory exists, but data directory does not.\nIf you simply want to recreate the data directory, run \"datacats init\" in the environment directory.\"\"\"\n )\n return name, datadir, srcdir\n", "role": "user" }, { "content": "new_environment_check", "role": "assistant" } ]
def x(datadir, sitedir, get_container_name): if any(not path.isdir(sitedir + x) for x in ('/files', '/run', '/solr')): return False if docker.is_boot2docker(): return all(docker.inspect_container(get_container_name(x)) for x in ('pgdata', 'venv')) return path.isdir(datadir + '/venv') and path.isdir(sitedir + '/postgres')
data_complete
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(datadir, sitedir, get_container_name):\n if any(not path.isdir(sitedir + x) for x in ('/files', '/run', '/solr')):\n return False\n if docker.is_boot2docker():\n return all(docker.inspect_container(get_container_name(x)) for x in\n ('pgdata', 'venv'))\n return path.isdir(datadir + '/venv') and path.isdir(sitedir + '/postgres')\n", "role": "user" }, { "content": "data_complete", "role": "assistant" } ]
def x(datadir, sitedir, srcdir=None): if not path.isdir(datadir): os.makedirs(datadir, mode=448) try: os.makedirs(sitedir, mode=448) except OSError: raise DatacatsError('Site already exists.') if not docker.is_boot2docker(): if not path.isdir(datadir + '/venv'): os.makedirs(datadir + '/venv') os.makedirs(sitedir + '/postgres') os.makedirs(sitedir + '/solr') os.makedirs(sitedir + '/files') os.makedirs(sitedir + '/run') if srcdir: os.makedirs(srcdir)
create_directories
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(datadir, sitedir, srcdir=None):\n if not path.isdir(datadir):\n os.makedirs(datadir, mode=448)\n try:\n os.makedirs(sitedir, mode=448)\n except OSError:\n raise DatacatsError('Site already exists.')\n if not docker.is_boot2docker():\n if not path.isdir(datadir + '/venv'):\n os.makedirs(datadir + '/venv')\n os.makedirs(sitedir + '/postgres')\n os.makedirs(sitedir + '/solr')\n os.makedirs(sitedir + '/files')\n os.makedirs(sitedir + '/run')\n if srcdir:\n os.makedirs(srcdir)\n", "role": "user" }, { "content": "create_directories", "role": "assistant" } ]
def x(srcdir, datadir, preload_image, get_container_name): try: if docker.is_boot2docker(): docker.data_only_container(get_container_name('venv'), [ '/usr/lib/ckan']) img_id = docker.web_command( '/bin/mv /usr/lib/ckan/ /usr/lib/ckan_original', image= preload_image, commit=True) docker.web_command(command= '/bin/cp -a /usr/lib/ckan_original/. /usr/lib/ckan/.', volumes_from=get_container_name('venv'), image=img_id) docker.remove_image(img_id) return docker.web_command(command= '/bin/cp -a /usr/lib/ckan/. /usr/lib/ckan_target/.', rw={( datadir + '/venv'): '/usr/lib/ckan_target'}, image=preload_image) finally: rw = {(datadir + '/venv'): '/usr/lib/ckan' } if not docker.is_boot2docker() else {} volumes_from = get_container_name('venv') if docker.is_boot2docker( ) else None docker.web_command(command= '/bin/chown -R --reference=/project /usr/lib/ckan', rw=rw, volumes_from=volumes_from, ro={srcdir: '/project'})
create_virtualenv
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(srcdir, datadir, preload_image, get_container_name):\n try:\n if docker.is_boot2docker():\n docker.data_only_container(get_container_name('venv'), [\n '/usr/lib/ckan'])\n img_id = docker.web_command(\n '/bin/mv /usr/lib/ckan/ /usr/lib/ckan_original', image=\n preload_image, commit=True)\n docker.web_command(command=\n '/bin/cp -a /usr/lib/ckan_original/. /usr/lib/ckan/.',\n volumes_from=get_container_name('venv'), image=img_id)\n docker.remove_image(img_id)\n return\n docker.web_command(command=\n '/bin/cp -a /usr/lib/ckan/. /usr/lib/ckan_target/.', rw={(\n datadir + '/venv'): '/usr/lib/ckan_target'}, image=preload_image)\n finally:\n rw = {(datadir + '/venv'): '/usr/lib/ckan'\n } if not docker.is_boot2docker() else {}\n volumes_from = get_container_name('venv') if docker.is_boot2docker(\n ) else None\n docker.web_command(command=\n '/bin/chown -R --reference=/project /usr/lib/ckan', rw=rw,\n volumes_from=volumes_from, ro={srcdir: '/project'})\n", "role": "user" }, { "content": "create_virtualenv", "role": "assistant" } ]
def x(srcdir, preload_image, datapusher=False): try: docker.web_command(command= '/bin/cp -a /project/ckan /project_target/ckan', rw={srcdir: '/project_target'}, image=preload_image) if datapusher: docker.web_command(command= '/bin/cp -a /project/datapusher /project_target/datapusher', rw={srcdir: '/project_target'}, image=preload_image) shutil.copy(srcdir + '/ckan/ckan/config/who.ini', srcdir) shutil.copy(srcdir + '/ckan/ckan/config/solr/schema.xml', srcdir) finally: docker.web_command(command= '/bin/chown -R --reference=/project /project', rw={srcdir: '/project'})
create_source
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(srcdir, preload_image, datapusher=False):\n try:\n docker.web_command(command=\n '/bin/cp -a /project/ckan /project_target/ckan', rw={srcdir:\n '/project_target'}, image=preload_image)\n if datapusher:\n docker.web_command(command=\n '/bin/cp -a /project/datapusher /project_target/datapusher',\n rw={srcdir: '/project_target'}, image=preload_image)\n shutil.copy(srcdir + '/ckan/ckan/config/who.ini', srcdir)\n shutil.copy(srcdir + '/ckan/ckan/config/solr/schema.xml', srcdir)\n finally:\n docker.web_command(command=\n '/bin/chown -R --reference=/project /project', rw={srcdir:\n '/project'})\n", "role": "user" }, { "content": "create_source", "role": "assistant" } ]
def x(sitedir, srcdir, passwords, get_container_name, extra_containers, log_syslog=False): if docker.is_boot2docker(): docker.data_only_container(get_container_name('pgdata'), [ '/var/lib/postgresql/data']) rw = {} volumes_from = get_container_name('pgdata') else: rw = {(sitedir + '/postgres'): '/var/lib/postgresql/data'} volumes_from = None running = set(containers_running(get_container_name)) needed = set(extra_containers).union({'postgres', 'solr'}) if not needed.issubset(running): stop_supporting_containers(get_container_name, extra_containers) docker.run_container(name=get_container_name('postgres'), image= 'datacats/postgres', environment=passwords, rw=rw, volumes_from =volumes_from, log_syslog=log_syslog) docker.run_container(name=get_container_name('solr'), image= 'datacats/solr', rw={(sitedir + '/solr'): '/var/lib/solr'}, ro= {(srcdir + '/schema.xml'): '/etc/solr/conf/schema.xml'}, log_syslog=log_syslog) for container in extra_containers: docker.run_container(name=get_container_name(container), image= EXTRA_IMAGE_MAPPING[container], ro={sitedir: '/datadir', srcdir: '/project'}, log_syslog=log_syslog)
start_supporting_containers
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sitedir, srcdir, passwords, get_container_name, extra_containers,\n log_syslog=False):\n if docker.is_boot2docker():\n docker.data_only_container(get_container_name('pgdata'), [\n '/var/lib/postgresql/data'])\n rw = {}\n volumes_from = get_container_name('pgdata')\n else:\n rw = {(sitedir + '/postgres'): '/var/lib/postgresql/data'}\n volumes_from = None\n running = set(containers_running(get_container_name))\n needed = set(extra_containers).union({'postgres', 'solr'})\n if not needed.issubset(running):\n stop_supporting_containers(get_container_name, extra_containers)\n docker.run_container(name=get_container_name('postgres'), image=\n 'datacats/postgres', environment=passwords, rw=rw, volumes_from\n =volumes_from, log_syslog=log_syslog)\n docker.run_container(name=get_container_name('solr'), image=\n 'datacats/solr', rw={(sitedir + '/solr'): '/var/lib/solr'}, ro=\n {(srcdir + '/schema.xml'): '/etc/solr/conf/schema.xml'},\n log_syslog=log_syslog)\n for container in extra_containers:\n docker.run_container(name=get_container_name(container), image=\n EXTRA_IMAGE_MAPPING[container], ro={sitedir: '/datadir',\n srcdir: '/project'}, log_syslog=log_syslog)\n", "role": "user" }, { "content": "start_supporting_containers", "role": "assistant" } ]
def x(get_container_name, extra_containers): docker.remove_container(get_container_name('postgres')) docker.remove_container(get_container_name('solr')) for container in extra_containers: docker.remove_container(get_container_name(container))
stop_supporting_containers
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(get_container_name, extra_containers):\n docker.remove_container(get_container_name('postgres'))\n docker.remove_container(get_container_name('solr'))\n for container in extra_containers:\n docker.remove_container(get_container_name(container))\n", "role": "user" }, { "content": "stop_supporting_containers", "role": "assistant" } ]
def x(get_container_name): running = [] for n in ['web', 'postgres', 'solr', 'datapusher', 'redis']: info = docker.inspect_container(get_container_name(n)) if info and not info['State']['Running']: running.append(n + '(halted)') elif info: running.append(n) return running
containers_running
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(get_container_name):\n running = []\n for n in ['web', 'postgres', 'solr', 'datapusher', 'redis']:\n info = docker.inspect_container(get_container_name(n))\n if info and not info['State']['Running']:\n running.append(n + '(halted)')\n elif info:\n running.append(n)\n return running\n", "role": "user" }, { "content": "containers_running", "role": "assistant" } ]
def x(self): if not self.sites: self.sites = task.list_sites(self.datadir) return self.sites
load_sites
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if not self.sites:\n self.sites = task.list_sites(self.datadir)\n return self.sites\n", "role": "user" }, { "content": "load_sites", "role": "assistant" } ]
def x(self, create=True): self._load_sites() if create: self.sites.append(self.site_name) task.save_new_site(self.site_name, self.sitedir, self.target, self.port, self.address, self.site_url, self.passwords)
save_site
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, create=True):\n self._load_sites()\n if create:\n self.sites.append(self.site_name)\n task.save_new_site(self.site_name, self.sitedir, self.target, self.port,\n self.address, self.site_url, self.passwords)\n", "role": "user" }, { "content": "save_site", "role": "assistant" } ]
def x(self): task.save_new_environment(self.name, self.datadir, self.target, self. ckan_version, self.deploy_target, self.always_prod)
save
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n task.save_new_environment(self.name, self.datadir, self.target, self.\n ckan_version, self.deploy_target, self.always_prod)\n", "role": "user" }, { "content": "save", "role": "assistant" } ]
def x(cls, path, ckan_version, site_name, **kwargs): if ckan_version == 'master': ckan_version = 'latest' name, datadir, srcdir = task.new_environment_check(path, site_name, ckan_version) environment = cls(name, srcdir, datadir, site_name, ckan_version, **kwargs) environment._generate_passwords() return environment
new
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, path, ckan_version, site_name, **kwargs):\n if ckan_version == 'master':\n ckan_version = 'latest'\n name, datadir, srcdir = task.new_environment_check(path, site_name,\n ckan_version)\n environment = cls(name, srcdir, datadir, site_name, ckan_version, **kwargs)\n environment._generate_passwords()\n return environment\n", "role": "user" }, { "content": "new", "role": "assistant" } ]
def x(cls, environment_name=None, site_name='primary', data_only=False, allow_old=False): srcdir, extension_dir, datadir = task.find_environment_dirs( environment_name, data_only) if datadir and data_only: return cls(environment_name, None, datadir, site_name) (datadir, name, ckan_version, always_prod, deploy_target, remote_server_key, extra_containers) = task.load_environment(srcdir, datadir, allow_old) if not allow_old: port, address, site_url, passwords = task.load_site(srcdir, datadir, site_name) else: port, address, site_url, passwords = None, None, None, None environment = cls(name, srcdir, datadir, site_name, ckan_version= ckan_version, port=port, deploy_target=deploy_target, site_url= site_url, always_prod=always_prod, address=address, extension_dir= extension_dir, remote_server_key=remote_server_key, extra_containers=extra_containers) if passwords: environment.passwords = passwords else: environment._generate_passwords() if not allow_old: environment._load_sites() return environment
load
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, environment_name=None, site_name='primary', data_only=False,\n allow_old=False):\n srcdir, extension_dir, datadir = task.find_environment_dirs(\n environment_name, data_only)\n if datadir and data_only:\n return cls(environment_name, None, datadir, site_name)\n (datadir, name, ckan_version, always_prod, deploy_target,\n remote_server_key, extra_containers) = task.load_environment(srcdir,\n datadir, allow_old)\n if not allow_old:\n port, address, site_url, passwords = task.load_site(srcdir, datadir,\n site_name)\n else:\n port, address, site_url, passwords = None, None, None, None\n environment = cls(name, srcdir, datadir, site_name, ckan_version=\n ckan_version, port=port, deploy_target=deploy_target, site_url=\n site_url, always_prod=always_prod, address=address, extension_dir=\n extension_dir, remote_server_key=remote_server_key,\n extra_containers=extra_containers)\n if passwords:\n environment.passwords = passwords\n else:\n environment._generate_passwords()\n if not allow_old:\n environment._load_sites()\n return environment\n", "role": "user" }, { "content": "load", "role": "assistant" } ]
def x(self): return task.data_complete(self.datadir, self.sitedir, self. _get_container_name)
data_complete
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return task.data_complete(self.datadir, self.sitedir, self.\n _get_container_name)\n", "role": "user" }, { "content": "data_complete", "role": "assistant" } ]
def x(self): files = task.source_missing(self.target) if files: raise DatacatsError('Missing files in source directory:\n' + '\n'. join(files)) if not self.data_exists(): raise DatacatsError('Environment datadir missing. Try "datacats init".' ) if not self.data_complete(): raise DatacatsError( 'Environment datadir damaged or volumes missing. To reset and discard all data use "datacats reset"' )
require_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n files = task.source_missing(self.target)\n if files:\n raise DatacatsError('Missing files in source directory:\\n' + '\\n'.\n join(files))\n if not self.data_exists():\n raise DatacatsError('Environment datadir missing. Try \"datacats init\".'\n )\n if not self.data_complete():\n raise DatacatsError(\n 'Environment datadir damaged or volumes missing. To reset and discard all data use \"datacats reset\"'\n )\n", "role": "user" }, { "content": "require_data", "role": "assistant" } ]
def x(self, create_project_dir=True): return task.create_directories(self.datadir, self.sitedir, self.target if create_project_dir else None)
create_directories
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, create_project_dir=True):\n return task.create_directories(self.datadir, self.sitedir, self.target if\n create_project_dir else None)\n", "role": "user" }, { "content": "create_directories", "role": "assistant" } ]
def x(self): return task.create_virtualenv(self.target, self.datadir, self. _preload_image(), self._get_container_name)
create_virtualenv
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return task.create_virtualenv(self.target, self.datadir, self.\n _preload_image(), self._get_container_name)\n", "role": "user" }, { "content": "create_virtualenv", "role": "assistant" } ]
def x(self): self.user_run_script(script=scripts.get_script_path( 'clean_virtualenv.sh'), args=[], rw_venv=True)
clean_virtualenv
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.user_run_script(script=scripts.get_script_path(\n 'clean_virtualenv.sh'), args=[], rw_venv=True)\n", "role": "user" }, { "content": "clean_virtualenv", "role": "assistant" } ]
def x(self, datapusher=True): task.create_source(self.target, self._preload_image(), datapusher)
create_source
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, datapusher=True):\n task.create_source(self.target, self._preload_image(), datapusher)\n", "role": "user" }, { "content": "create_source", "role": "assistant" } ]
def x(self, log_syslog=False): log_syslog = True if self.always_prod else log_syslog task.start_supporting_containers(self.sitedir, self.target, self. passwords, self._get_container_name, self.extra_containers, log_syslog=log_syslog)
start_supporting_containers
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, log_syslog=False):\n log_syslog = True if self.always_prod else log_syslog\n task.start_supporting_containers(self.sitedir, self.target, self.\n passwords, self._get_container_name, self.extra_containers,\n log_syslog=log_syslog)\n", "role": "user" }, { "content": "start_supporting_containers", "role": "assistant" } ]
def x(self): self.run_command(command= '/scripts/run_as_user.sh /usr/lib/ckan/bin/paster make-config ckan /project/development.ini' , rw_project=True, ro={scripts.get_script_path('run_as_user.sh'): '/scripts/run_as_user.sh'})
create_ckan_ini
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.run_command(command=\n '/scripts/run_as_user.sh /usr/lib/ckan/bin/paster make-config ckan /project/development.ini'\n , rw_project=True, ro={scripts.get_script_path('run_as_user.sh'):\n '/scripts/run_as_user.sh'})\n", "role": "user" }, { "content": "create_ckan_ini", "role": "assistant" } ]
def x(self, skin=True): command = ['/usr/lib/ckan/bin/paster', '--plugin=ckan', 'config-tool', '/project/development.ini', '-e', 'sqlalchemy.url = postgresql://<hidden>', 'ckan.datastore.read_url = postgresql://<hidden>', 'ckan.datastore.write_url = postgresql://<hidden>', 'ckan.datapusher.url = http://datapusher:8800', 'solr_url = http://solr:8080/solr', 'ckan.storage_path = /var/www/storage', 'ckan.plugins = datastore resource_proxy text_view ' + ( 'datapusher ' if exists(self.target + '/datapusher') else '') + 'recline_grid_view recline_graph_view' + (' {0}_theme'.format(self. name) if skin else ''), 'ckan.site_title = ' + self.name, 'ckan.site_logo =', 'ckan.auth.create_user_via_web = false'] self.run_command(command=command, rw_project=True)
update_ckan_ini
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, skin=True):\n command = ['/usr/lib/ckan/bin/paster', '--plugin=ckan', 'config-tool',\n '/project/development.ini', '-e',\n 'sqlalchemy.url = postgresql://<hidden>',\n 'ckan.datastore.read_url = postgresql://<hidden>',\n 'ckan.datastore.write_url = postgresql://<hidden>',\n 'ckan.datapusher.url = http://datapusher:8800',\n 'solr_url = http://solr:8080/solr',\n 'ckan.storage_path = /var/www/storage', \n 'ckan.plugins = datastore resource_proxy text_view ' + (\n 'datapusher ' if exists(self.target + '/datapusher') else '') +\n 'recline_grid_view recline_graph_view' + (' {0}_theme'.format(self.\n name) if skin else ''), 'ckan.site_title = ' + self.name,\n 'ckan.site_logo =', 'ckan.auth.create_user_via_web = false']\n self.run_command(command=command, rw_project=True)\n", "role": "user" }, { "content": "update_ckan_ini", "role": "assistant" } ]
def x(self): ckan_extension_template(self.name, self.target) self.install_package_develop('ckanext-' + self.name + 'theme')
create_install_template_skin
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n ckan_extension_template(self.name, self.target)\n self.install_package_develop('ckanext-' + self.name + 'theme')\n", "role": "user" }, { "content": "create_install_template_skin", "role": "assistant" } ]
def x(self, retry_seconds=DB_INIT_RETRY_SECONDS): started = time.time() while True: try: self.run_command( '/usr/lib/ckan/bin/paster --plugin=ckan db init -c /project/development.ini' , db_links=True, clean_up=True) break except WebCommandError: if started + retry_seconds > time.time(): raise time.sleep(DB_INIT_RETRY_DELAY)
ckan_db_init
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, retry_seconds=DB_INIT_RETRY_SECONDS):\n started = time.time()\n while True:\n try:\n self.run_command(\n '/usr/lib/ckan/bin/paster --plugin=ckan db init -c /project/development.ini'\n , db_links=True, clean_up=True)\n break\n except WebCommandError:\n if started + retry_seconds > time.time():\n raise\n time.sleep(DB_INIT_RETRY_DELAY)\n", "role": "user" }, { "content": "ckan_db_init", "role": "assistant" } ]
def x(self, production=False, log_syslog=False, paster_reload=True, interactive=False): self.stop_ckan() address = self.address or '127.0.0.1' port = self.port log_syslog = True if self.always_prod else log_syslog production = production or self.always_prod override_site_url = self.address is None and not is_boot2docker( ) and not self.site_url command = ['/scripts/web.sh', str(production), str(override_site_url), str(paster_reload)] if not production: css = self.target + '/ckan/ckan/public/base/css' if not exists(css + '/main.debug.css'): from shutil import copyfile copyfile(css + '/main.css', css + '/main.debug.css') ro = {self.target: '/project', scripts.get_script_path('datapusher.sh'): '/scripts/datapusher.sh'} if not is_boot2docker(): ro[self.datadir + '/venv'] = '/usr/lib/ckan' datapusher = self.needs_datapusher() if datapusher: run_container(self._get_container_name('datapusher'), 'datacats/web', '/scripts/datapusher.sh', ro=ro, volumes_from= self._get_container_name('venv') if is_boot2docker() else None, log_syslog=log_syslog) while True: self._create_run_ini(port, production) try: self._run_web_container(port, command, address, log_syslog= log_syslog, datapusher=datapusher, interactive=interactive) if not is_boot2docker(): self.address = address except PortAllocatedError: port = self._next_port(port) continue break
start_ckan
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, production=False, log_syslog=False, paster_reload=True,\n interactive=False):\n self.stop_ckan()\n address = self.address or '127.0.0.1'\n port = self.port\n log_syslog = True if self.always_prod else log_syslog\n production = production or self.always_prod\n override_site_url = self.address is None and not is_boot2docker(\n ) and not self.site_url\n command = ['/scripts/web.sh', str(production), str(override_site_url),\n str(paster_reload)]\n if not production:\n css = self.target + '/ckan/ckan/public/base/css'\n if not exists(css + '/main.debug.css'):\n from shutil import copyfile\n copyfile(css + '/main.css', css + '/main.debug.css')\n ro = {self.target: '/project', scripts.get_script_path('datapusher.sh'):\n '/scripts/datapusher.sh'}\n if not is_boot2docker():\n ro[self.datadir + '/venv'] = '/usr/lib/ckan'\n datapusher = self.needs_datapusher()\n if datapusher:\n run_container(self._get_container_name('datapusher'),\n 'datacats/web', '/scripts/datapusher.sh', ro=ro, volumes_from=\n self._get_container_name('venv') if is_boot2docker() else None,\n log_syslog=log_syslog)\n while True:\n self._create_run_ini(port, production)\n try:\n self._run_web_container(port, command, address, log_syslog=\n log_syslog, datapusher=datapusher, interactive=interactive)\n if not is_boot2docker():\n self.address = address\n except PortAllocatedError:\n port = self._next_port(port)\n continue\n break\n", "role": "user" }, { "content": "start_ckan", "role": "assistant" } ]
def x(self, port, production, output='development.ini', source= 'development.ini', override_site_url=True): cp = SafeConfigParser() try: cp.read([self.target + '/' + source]) except ConfigParserError: raise DatacatsError('Error reading development.ini') cp.set('DEFAULT', 'debug', 'false' if production else 'true') if self.site_url: site_url = self.site_url else: if is_boot2docker(): web_address = socket.gethostbyname(docker_host()) else: web_address = self.address site_url = 'http://{}:{}'.format(web_address, port) if override_site_url: cp.set('app:main', 'ckan.site_url', site_url) cp.set('app:main', 'sqlalchemy.url', 'postgresql://ckan:{0}@db:5432/ckan'.format(self.passwords[ 'CKAN_PASSWORD'])) cp.set('app:main', 'ckan.datastore.read_url', 'postgresql://ckan_datastore_readonly:{0}@db:5432/ckan_datastore'. format(self.passwords['DATASTORE_RO_PASSWORD'])) cp.set('app:main', 'ckan.datastore.write_url', 'postgresql://ckan_datastore_readwrite:{0}@db:5432/ckan_datastore'. format(self.passwords['DATASTORE_RW_PASSWORD'])) cp.set('app:main', 'solr_url', 'http://solr:8080/solr') cp.set('app:main', 'ckan.redis.url', 'http://redis:6379') cp.set('app:main', 'beaker.session.secret', self.passwords[ 'BEAKER_SESSION_SECRET']) if not isdir(self.sitedir + '/run'): makedirs(self.sitedir + '/run') with open(self.sitedir + '/run/' + output, 'w') as runini: cp.write(runini)
create_run_ini
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, port, production, output='development.ini', source=\n 'development.ini', override_site_url=True):\n cp = SafeConfigParser()\n try:\n cp.read([self.target + '/' + source])\n except ConfigParserError:\n raise DatacatsError('Error reading development.ini')\n cp.set('DEFAULT', 'debug', 'false' if production else 'true')\n if self.site_url:\n site_url = self.site_url\n else:\n if is_boot2docker():\n web_address = socket.gethostbyname(docker_host())\n else:\n web_address = self.address\n site_url = 'http://{}:{}'.format(web_address, port)\n if override_site_url:\n cp.set('app:main', 'ckan.site_url', site_url)\n cp.set('app:main', 'sqlalchemy.url',\n 'postgresql://ckan:{0}@db:5432/ckan'.format(self.passwords[\n 'CKAN_PASSWORD']))\n cp.set('app:main', 'ckan.datastore.read_url',\n 'postgresql://ckan_datastore_readonly:{0}@db:5432/ckan_datastore'.\n format(self.passwords['DATASTORE_RO_PASSWORD']))\n cp.set('app:main', 'ckan.datastore.write_url',\n 'postgresql://ckan_datastore_readwrite:{0}@db:5432/ckan_datastore'.\n format(self.passwords['DATASTORE_RW_PASSWORD']))\n cp.set('app:main', 'solr_url', 'http://solr:8080/solr')\n cp.set('app:main', 'ckan.redis.url', 'http://redis:6379')\n cp.set('app:main', 'beaker.session.secret', self.passwords[\n 'BEAKER_SESSION_SECRET'])\n if not isdir(self.sitedir + '/run'):\n makedirs(self.sitedir + '/run')\n with open(self.sitedir + '/run/' + output, 'w') as runini:\n cp.write(runini)\n", "role": "user" }, { "content": "create_run_ini", "role": "assistant" } ]
def x(self, port, command, address, log_syslog=False, datapusher=True, interactive=False): if is_boot2docker(): ro = {} volumes_from = self._get_container_name('venv') else: ro = {(self.datadir + '/venv'): '/usr/lib/ckan'} volumes_from = None links = {self._get_container_name('solr'): 'solr', self. _get_container_name('postgres'): 'db'} links.update({self._get_container_name(container): container for container in self.extra_containers}) if datapusher: if 'datapusher' not in self.containers_running(): raise DatacatsError(container_logs(self._get_container_name( 'datapusher'), 'all', False, False)) links[self._get_container_name('datapusher')] = 'datapusher' ro = dict({self.target: '/project/', scripts.get_script_path('web.sh'): '/scripts/web.sh', scripts.get_script_path('adjust_devini.py'): '/scripts/adjust_devini.py'}, **ro) rw = {(self.sitedir + '/files'): '/var/www/storage', (self.sitedir + '/run/development.ini'): '/project/development.ini'} try: if not interactive: run_container(name=self._get_container_name('web'), image= 'datacats/web', rw=rw, ro=ro, links=links, volumes_from= volumes_from, command=command, port_bindings={(5000): port if is_boot2docker() else (address, port)}, log_syslog=log_syslog) else: if is_boot2docker(): switches = ['--volumes-from', self._get_container_name( 'pgdata'), '--volumes-from', self._get_container_name( 'venv')] else: switches = [] switches += ['--volume={}:{}:ro'.format(vol, ro[vol]) for vol in ro ] switches += ['--volume={}:{}'.format(vol, rw[vol]) for vol in rw] links = ['--link={}:{}'.format(link, links[link]) for link in links ] args = ['docker', 'run', '-it', '--name', self. _get_container_name('web'), '-p', '{}:5000'.format(port) if is_boot2docker() else '{}:{}:5000'.format(address, port) ] + switches + links + ['datacats/web'] + command subprocess.call(args) except APIError as e: if '409' in str(e): raise DatacatsError( 'Web container already running. Please stop_web before running.' ) else: raise
run_web_container
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, port, command, address, log_syslog=False, datapusher=True,\n interactive=False):\n if is_boot2docker():\n ro = {}\n volumes_from = self._get_container_name('venv')\n else:\n ro = {(self.datadir + '/venv'): '/usr/lib/ckan'}\n volumes_from = None\n links = {self._get_container_name('solr'): 'solr', self.\n _get_container_name('postgres'): 'db'}\n links.update({self._get_container_name(container): container for\n container in self.extra_containers})\n if datapusher:\n if 'datapusher' not in self.containers_running():\n raise DatacatsError(container_logs(self._get_container_name(\n 'datapusher'), 'all', False, False))\n links[self._get_container_name('datapusher')] = 'datapusher'\n ro = dict({self.target: '/project/', scripts.get_script_path('web.sh'):\n '/scripts/web.sh', scripts.get_script_path('adjust_devini.py'):\n '/scripts/adjust_devini.py'}, **ro)\n rw = {(self.sitedir + '/files'): '/var/www/storage', (self.sitedir +\n '/run/development.ini'): '/project/development.ini'}\n try:\n if not interactive:\n run_container(name=self._get_container_name('web'), image=\n 'datacats/web', rw=rw, ro=ro, links=links, volumes_from=\n volumes_from, command=command, port_bindings={(5000): port if\n is_boot2docker() else (address, port)}, log_syslog=log_syslog)\n else:\n if is_boot2docker():\n switches = ['--volumes-from', self._get_container_name(\n 'pgdata'), '--volumes-from', self._get_container_name(\n 'venv')]\n else:\n switches = []\n switches += ['--volume={}:{}:ro'.format(vol, ro[vol]) for vol in ro\n ]\n switches += ['--volume={}:{}'.format(vol, rw[vol]) for vol in rw]\n links = ['--link={}:{}'.format(link, links[link]) for link in links\n ]\n args = ['docker', 'run', '-it', '--name', self.\n _get_container_name('web'), '-p', '{}:5000'.format(port) if\n is_boot2docker() else '{}:{}:5000'.format(address, port)\n ] + switches + links + ['datacats/web'] + command\n subprocess.call(args)\n except APIError as e:\n if '409' in str(e):\n raise DatacatsError(\n 'Web container already running. Please stop_web before running.'\n )\n else:\n raise\n", "role": "user" }, { "content": "run_web_container", "role": "assistant" } ]
def x(self): try: if not wait_for_service_available(self._get_container_name('web'), self.web_address(), WEB_START_TIMEOUT_SECONDS): raise DatacatsError('Error while starting web container:\n' + container_logs(self._get_container_name('web'), 'all', False, None)) except ServiceTimeout: raise DatacatsError('Timeout while starting web container. Logs:' + container_logs(self._get_container_name('web'), 'all', False, None) )
wait_for_web_available
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n try:\n if not wait_for_service_available(self._get_container_name('web'),\n self.web_address(), WEB_START_TIMEOUT_SECONDS):\n raise DatacatsError('Error while starting web container:\\n' +\n container_logs(self._get_container_name('web'), 'all', \n False, None))\n except ServiceTimeout:\n raise DatacatsError('Timeout while starting web container. Logs:' +\n container_logs(self._get_container_name('web'), 'all', False, None)\n )\n", "role": "user" }, { "content": "wait_for_web_available", "role": "assistant" } ]
def x(self): return 5000 + unpack('Q', sha((self.name + self.site_name).decode( 'ascii')).digest()[:8])[0] % 1000
choose_port
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return 5000 + unpack('Q', sha((self.name + self.site_name).decode(\n 'ascii')).digest()[:8])[0] % 1000\n", "role": "user" }, { "content": "choose_port", "role": "assistant" } ]
def x(self, port): port = 5000 + (port + 1) % 1000 if port == self.port: raise DatacatsError('Too many instances running') return port
next_port
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, port):\n port = 5000 + (port + 1) % 1000\n if port == self.port:\n raise DatacatsError('Too many instances running')\n return port\n", "role": "user" }, { "content": "next_port", "role": "assistant" } ]
def x(self): remove_container(self._get_container_name('web'), force=True) remove_container(self._get_container_name('datapusher'), force=True)
stop_ckan
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n remove_container(self._get_container_name('web'), force=True)\n remove_container(self._get_container_name('datapusher'), force=True)\n", "role": "user" }, { "content": "stop_ckan", "role": "assistant" } ]
def x(self): info = inspect_container(self._get_container_name('web')) if info is None: return None try: if not info['State']['Running']: return None return info['NetworkSettings']['Ports']['5000/tcp'][0]['HostPort'] except TypeError: return None
current_web_port
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n info = inspect_container(self._get_container_name('web'))\n if info is None:\n return None\n try:\n if not info['State']['Running']:\n return None\n return info['NetworkSettings']['Ports']['5000/tcp'][0]['HostPort']\n except TypeError:\n return None\n", "role": "user" }, { "content": "current_web_port", "role": "assistant" } ]
def x(self, container, error_on_exists=False): if container in self.extra_containers: if error_on_exists: raise DatacatsError('{} is already added as an extra container.' .format(container)) else: return self.extra_containers.append(container) cp = SafeConfigParser() cp.read(self.target + '/.datacats-environment') cp.set('datacats', 'extra_containers', ' '.join(self.extra_containers)) with open(self.target + '/.datacats-environment', 'w') as f: cp.write(f)
add_extra_container
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, container, error_on_exists=False):\n if container in self.extra_containers:\n if error_on_exists:\n raise DatacatsError('{} is already added as an extra container.'\n .format(container))\n else:\n return\n self.extra_containers.append(container)\n cp = SafeConfigParser()\n cp.read(self.target + '/.datacats-environment')\n cp.set('datacats', 'extra_containers', ' '.join(self.extra_containers))\n with open(self.target + '/.datacats-environment', 'w') as f:\n cp.write(f)\n", "role": "user" }, { "content": "add_extra_container", "role": "assistant" } ]
def x(self): port = self._current_web_port() address = self.address or '127.0.0.1' if port is None: return None return 'http://{0}:{1}/'.format(address if address and not is_boot2docker() else docker_host(), port)
web_address
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n port = self._current_web_port()\n address = self.address or '127.0.0.1'\n if port is None:\n return None\n return 'http://{0}:{1}/'.format(address if address and not\n is_boot2docker() else docker_host(), port)\n", "role": "user" }, { "content": "web_address", "role": "assistant" } ]
def x(self, password): with open(self.sitedir + '/run/admin.json', 'w') as out: json.dump({'name': 'admin', 'email': 'none', 'password': password, 'sysadmin': True}, out) self.user_run_script(script=scripts.get_script_path( 'update_add_admin.sh'), args=[], db_links=True, ro={(self.sitedir + '/run/admin.json'): '/input/admin.json'}) remove(self.sitedir + '/run/admin.json')
create_admin_set_password
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, password):\n with open(self.sitedir + '/run/admin.json', 'w') as out:\n json.dump({'name': 'admin', 'email': 'none', 'password': password,\n 'sysadmin': True}, out)\n self.user_run_script(script=scripts.get_script_path(\n 'update_add_admin.sh'), args=[], db_links=True, ro={(self.sitedir +\n '/run/admin.json'): '/input/admin.json'})\n remove(self.sitedir + '/run/admin.json')\n", "role": "user" }, { "content": "create_admin_set_password", "role": "assistant" } ]
def x(self, command=None, paster=False, detach=False): if not exists(self.target + '/.bash_profile'): self.create_bash_profile() if not command: command = [] use_tty = sys.stdin.isatty() and sys.stdout.isatty() background = environ.get('CIRCLECI', False) or detach if is_boot2docker(): venv_volumes = ['--volumes-from', self._get_container_name('venv')] else: venv_volumes = ['-v', self.datadir + '/venv:/usr/lib/ckan:rw'] self._create_run_ini(self.port, production=False, output='run.ini') self._create_run_ini(self.port, production=True, output='test.ini', source='ckan/test-core.ini', override_site_url=False) script = scripts.get_script_path('shell.sh') if paster: script = scripts.get_script_path('paster.sh') if command and command != ['help'] and command != ['--help']: command += ['--config=/project/development.ini'] command = [self.extension_dir] + command proxy_settings = self._proxy_settings() if proxy_settings: venv_volumes += ['-v', self.sitedir + '/run/proxy-environment:/etc/environment:ro'] links = {self._get_container_name('solr'): 'solr', self. _get_container_name('postgres'): 'db'} links.update({self._get_container_name(container): container for container in self.extra_containers}) link_params = [] for link in links: link_params.append('--link') link_params.append(link + ':' + links[link]) if 'datapusher' in self.containers_running(): link_params.append('--link') link_params.append(self._get_container_name('datapusher') + ':datapusher') return subprocess.call([DOCKER_EXE, 'run'] + (['--rm'] if not background else []) + ['-t' if use_tty else '', '-d' if detach else '-i'] + venv_volumes + ['-v', self.target + ':/project:rw', '-v', self.sitedir + '/files:/var/www/storage:rw', '-v', script + ':/scripts/shell.sh:ro', '-v', scripts.get_script_path( 'paster_cd.sh') + ':/scripts/paster_cd.sh:ro', '-v', self.sitedir + '/run/run.ini:/project/development.ini:ro', '-v', self.sitedir + '/run/test.ini:/project/ckan/test-core.ini:ro'] + link_params + [ '--hostname', self.name, 'datacats/web', '/scripts/shell.sh'] + command )
interactive_shell
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, command=None, paster=False, detach=False):\n if not exists(self.target + '/.bash_profile'):\n self.create_bash_profile()\n if not command:\n command = []\n use_tty = sys.stdin.isatty() and sys.stdout.isatty()\n background = environ.get('CIRCLECI', False) or detach\n if is_boot2docker():\n venv_volumes = ['--volumes-from', self._get_container_name('venv')]\n else:\n venv_volumes = ['-v', self.datadir + '/venv:/usr/lib/ckan:rw']\n self._create_run_ini(self.port, production=False, output='run.ini')\n self._create_run_ini(self.port, production=True, output='test.ini',\n source='ckan/test-core.ini', override_site_url=False)\n script = scripts.get_script_path('shell.sh')\n if paster:\n script = scripts.get_script_path('paster.sh')\n if command and command != ['help'] and command != ['--help']:\n command += ['--config=/project/development.ini']\n command = [self.extension_dir] + command\n proxy_settings = self._proxy_settings()\n if proxy_settings:\n venv_volumes += ['-v', self.sitedir +\n '/run/proxy-environment:/etc/environment:ro']\n links = {self._get_container_name('solr'): 'solr', self.\n _get_container_name('postgres'): 'db'}\n links.update({self._get_container_name(container): container for\n container in self.extra_containers})\n link_params = []\n for link in links:\n link_params.append('--link')\n link_params.append(link + ':' + links[link])\n if 'datapusher' in self.containers_running():\n link_params.append('--link')\n link_params.append(self._get_container_name('datapusher') +\n ':datapusher')\n return subprocess.call([DOCKER_EXE, 'run'] + (['--rm'] if not\n background else []) + ['-t' if use_tty else '', '-d' if detach else\n '-i'] + venv_volumes + ['-v', self.target + ':/project:rw', '-v', \n self.sitedir + '/files:/var/www/storage:rw', '-v', script +\n ':/scripts/shell.sh:ro', '-v', scripts.get_script_path(\n 'paster_cd.sh') + ':/scripts/paster_cd.sh:ro', '-v', self.sitedir +\n '/run/run.ini:/project/development.ini:ro', '-v', self.sitedir +\n '/run/test.ini:/project/ckan/test-core.ini:ro'] + link_params + [\n '--hostname', self.name, 'datacats/web', '/scripts/shell.sh'] + command\n )\n", "role": "user" }, { "content": "interactive_shell", "role": "assistant" } ]
def x(self, psrc, stream_output=None): package = self.target + '/' + psrc assert isdir(package), package reqname = '/requirements.txt' if not exists(package + reqname): reqname = '/pip-requirements.txt' if not exists(package + reqname): return return self.user_run_script(script=scripts.get_script_path( 'install_reqs.sh'), args=['/project/' + psrc + reqname], rw_venv= True, rw_project=True, stream_output=stream_output)
install_package_requirements
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, psrc, stream_output=None):\n package = self.target + '/' + psrc\n assert isdir(package), package\n reqname = '/requirements.txt'\n if not exists(package + reqname):\n reqname = '/pip-requirements.txt'\n if not exists(package + reqname):\n return\n return self.user_run_script(script=scripts.get_script_path(\n 'install_reqs.sh'), args=['/project/' + psrc + reqname], rw_venv=\n True, rw_project=True, stream_output=stream_output)\n", "role": "user" }, { "content": "install_package_requirements", "role": "assistant" } ]
def x(self, container, tail='all', follow=False, timestamps=False): return container_logs(self._get_container_name(container), tail, follow, timestamps)
logs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, container, tail='all', follow=False, timestamps=False):\n return container_logs(self._get_container_name(container), tail, follow,\n timestamps)\n", "role": "user" }, { "content": "logs", "role": "assistant" } ]