source_code
stringlengths 45
7.24k
| function_name
stringlengths 1
111
| messages
list |
|---|---|---|
def x(ip):
if not validate_ip(ip):
return None
quads = ip.split('.')
if len(quads) == 1:
quads = quads + [0, 0, 0]
elif len(quads) < 4:
host = quads[-1:]
quads = quads[:-1] + [0] * (4 - len(quads)) + host
lngip = 0
for q in quads:
lngip = lngip << 8 | int(q)
return lngip
|
ip2long
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ip):\n if not validate_ip(ip):\n return None\n quads = ip.split('.')\n if len(quads) == 1:\n quads = quads + [0, 0, 0]\n elif len(quads) < 4:\n host = quads[-1:]\n quads = quads[:-1] + [0] * (4 - len(quads)) + host\n lngip = 0\n for q in quads:\n lngip = lngip << 8 | int(q)\n return lngip\n",
"role": "user"
},
{
"content": "ip2long",
"role": "assistant"
}
] |
def x(ip):
if not validate_ip(ip):
return None
quads = ip.split('.')
netw = 0
for i in range(4):
netw = netw << 8 | int(len(quads) > i and quads[i] or 0)
return netw
|
ip2network
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ip):\n if not validate_ip(ip):\n return None\n quads = ip.split('.')\n netw = 0\n for i in range(4):\n netw = netw << 8 | int(len(quads) > i and quads[i] or 0)\n return netw\n",
"role": "user"
},
{
"content": "ip2network",
"role": "assistant"
}
] |
def x(l):
if MAX_IP < l or l < MIN_IP:
raise TypeError('expected int between %d and %d inclusive' % (
MIN_IP, MAX_IP))
return '%d.%d.%d.%d' % (l >> 24 & 255, l >> 16 & 255, l >> 8 & 255, l & 255
)
|
long2ip
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(l):\n if MAX_IP < l or l < MIN_IP:\n raise TypeError('expected int between %d and %d inclusive' % (\n MIN_IP, MAX_IP))\n return '%d.%d.%d.%d' % (l >> 24 & 255, l >> 16 & 255, l >> 8 & 255, l & 255\n )\n",
"role": "user"
},
{
"content": "long2ip",
"role": "assistant"
}
] |
def x(cidr):
if not validate_cidr(cidr):
return None
ip, prefix = cidr.split('/')
prefix = int(prefix)
network = ip2network(ip)
return _block_from_ip_and_prefix(network, prefix)
|
cidr2block
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cidr):\n if not validate_cidr(cidr):\n return None\n ip, prefix = cidr.split('/')\n prefix = int(prefix)\n network = ip2network(ip)\n return _block_from_ip_and_prefix(network, prefix)\n",
"role": "user"
},
{
"content": "cidr2block",
"role": "assistant"
}
] |
def x(subnet):
if not validate_subnet(subnet):
return None
ip, netmask = subnet.split('/')
prefix = netmask2prefix(netmask)
network = ip2network(ip)
return _block_from_ip_and_prefix(network, prefix)
|
subnet2block
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(subnet):\n if not validate_subnet(subnet):\n return None\n ip, netmask = subnet.split('/')\n prefix = netmask2prefix(netmask)\n network = ip2network(ip)\n return _block_from_ip_and_prefix(network, prefix)\n",
"role": "user"
},
{
"content": "subnet2block",
"role": "assistant"
}
] |
def x(ip, prefix):
shift = 32 - prefix
block_start = ip >> shift << shift
mask = (1 << shift) - 1
block_end = block_start | mask
return long2ip(block_start), long2ip(block_end)
|
block_from_ip_and_prefix
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ip, prefix):\n shift = 32 - prefix\n block_start = ip >> shift << shift\n mask = (1 << shift) - 1\n block_end = block_start | mask\n return long2ip(block_start), long2ip(block_end)\n",
"role": "user"
},
{
"content": "block_from_ip_and_prefix",
"role": "assistant"
}
] |
def x(job, samples, config):
job.fileStore.logToMaster('Downloaded shared files')
file_names = ['reference', 'phase', 'mills', 'dbsnp', 'cosmic']
urls = [config.reference, config.phase, config.mills, config.dbsnp,
config.cosmic]
for name, url in zip(file_names, urls):
if url:
vars(config)[name] = job.addChildJobFn(download_url_job, url=url
).rv()
job.addFollowOnJobFn(reference_preprocessing, samples, config)
|
download_shared_files
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(job, samples, config):\n job.fileStore.logToMaster('Downloaded shared files')\n file_names = ['reference', 'phase', 'mills', 'dbsnp', 'cosmic']\n urls = [config.reference, config.phase, config.mills, config.dbsnp,\n config.cosmic]\n for name, url in zip(file_names, urls):\n if url:\n vars(config)[name] = job.addChildJobFn(download_url_job, url=url\n ).rv()\n job.addFollowOnJobFn(reference_preprocessing, samples, config)\n",
"role": "user"
},
{
"content": "download_shared_files",
"role": "assistant"
}
] |
def x(job, samples, config):
job.fileStore.logToMaster('Processed reference files')
config.fai = job.addChildJobFn(run_samtools_faidx, config.reference).rv()
config.dict = job.addChildJobFn(run_picard_create_sequence_dictionary,
config.reference).rv()
job.addFollowOnJobFn(map_job, download_sample, samples, config)
|
reference_preprocessing
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(job, samples, config):\n job.fileStore.logToMaster('Processed reference files')\n config.fai = job.addChildJobFn(run_samtools_faidx, config.reference).rv()\n config.dict = job.addChildJobFn(run_picard_create_sequence_dictionary,\n config.reference).rv()\n job.addFollowOnJobFn(map_job, download_sample, samples, config)\n",
"role": "user"
},
{
"content": "reference_preprocessing",
"role": "assistant"
}
] |
def x(job, sample, config):
config = argparse.Namespace(**vars(config))
uuid, normal_url, tumor_url = sample
job.fileStore.logToMaster('Downloaded sample: ' + uuid)
config.uuid = uuid
config.normal = normal_url
config.tumor = tumor_url
config.cores = min(config.maxCores, int(multiprocessing.cpu_count()))
disk = '1G' if config.ci_test else '20G'
config.normal_bam = job.addChildJobFn(download_url_job, url=config.
normal, s3_key_path=config.ssec, cghub_key_path=config.gtkey, disk=disk
).rv()
config.tumor_bam = job.addChildJobFn(download_url_job, url=config.tumor,
s3_key_path=config.ssec, cghub_key_path=config.gtkey, disk=disk).rv()
job.addFollowOnJobFn(index_bams, config)
|
download_sample
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(job, sample, config):\n config = argparse.Namespace(**vars(config))\n uuid, normal_url, tumor_url = sample\n job.fileStore.logToMaster('Downloaded sample: ' + uuid)\n config.uuid = uuid\n config.normal = normal_url\n config.tumor = tumor_url\n config.cores = min(config.maxCores, int(multiprocessing.cpu_count()))\n disk = '1G' if config.ci_test else '20G'\n config.normal_bam = job.addChildJobFn(download_url_job, url=config.\n normal, s3_key_path=config.ssec, cghub_key_path=config.gtkey, disk=disk\n ).rv()\n config.tumor_bam = job.addChildJobFn(download_url_job, url=config.tumor,\n s3_key_path=config.ssec, cghub_key_path=config.gtkey, disk=disk).rv()\n job.addFollowOnJobFn(index_bams, config)\n",
"role": "user"
},
{
"content": "download_sample",
"role": "assistant"
}
] |
def x(job, config):
job.fileStore.logToMaster('Indexed sample BAMS: ' + config.uuid)
disk = '1G' if config.ci_test else '20G'
config.normal_bai = job.addChildJobFn(run_samtools_index, config.
normal_bam, cores=1, disk=disk).rv()
config.tumor_bai = job.addChildJobFn(run_samtools_index, config.
tumor_bam, cores=1, disk=disk).rv()
job.addFollowOnJobFn(preprocessing_declaration, config)
|
index_bams
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(job, config):\n job.fileStore.logToMaster('Indexed sample BAMS: ' + config.uuid)\n disk = '1G' if config.ci_test else '20G'\n config.normal_bai = job.addChildJobFn(run_samtools_index, config.\n normal_bam, cores=1, disk=disk).rv()\n config.tumor_bai = job.addChildJobFn(run_samtools_index, config.\n tumor_bam, cores=1, disk=disk).rv()\n job.addFollowOnJobFn(preprocessing_declaration, config)\n",
"role": "user"
},
{
"content": "index_bams",
"role": "assistant"
}
] |
def x(job, config):
if config.preprocessing:
job.fileStore.logToMaster('Ran preprocessing: ' + config.uuid)
disk = '1G' if config.ci_test else '20G'
mem = '2G' if config.ci_test else '10G'
processed_normal = job.wrapJobFn(run_gatk_preprocessing, config.
normal_bam, config.normal_bai, config.reference, config.dict,
config.fai, config.phase, config.mills, config.dbsnp, mem,
cores=1, memory=mem, disk=disk)
processed_tumor = job.wrapJobFn(run_gatk_preprocessing, config.
tumor_bam, config.tumor_bai, config.reference, config.dict,
config.fai, config.phase, config.mills, config.dbsnp, mem,
cores=1, memory=mem, disk=disk)
static_workflow = job.wrapJobFn(static_workflow_declaration, config,
processed_normal.rv(0), processed_normal.rv(1), processed_tumor
.rv(0), processed_tumor.rv(1))
job.addChild(processed_normal)
job.addChild(processed_tumor)
job.addFollowOn(static_workflow)
else:
job.addFollowOnJobFn(static_workflow_declaration, config, config.
normal_bam, config.normal_bai, config.tumor_bam, config.tumor_bai)
|
preprocessing_declaration
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(job, config):\n if config.preprocessing:\n job.fileStore.logToMaster('Ran preprocessing: ' + config.uuid)\n disk = '1G' if config.ci_test else '20G'\n mem = '2G' if config.ci_test else '10G'\n processed_normal = job.wrapJobFn(run_gatk_preprocessing, config.\n normal_bam, config.normal_bai, config.reference, config.dict,\n config.fai, config.phase, config.mills, config.dbsnp, mem,\n cores=1, memory=mem, disk=disk)\n processed_tumor = job.wrapJobFn(run_gatk_preprocessing, config.\n tumor_bam, config.tumor_bai, config.reference, config.dict,\n config.fai, config.phase, config.mills, config.dbsnp, mem,\n cores=1, memory=mem, disk=disk)\n static_workflow = job.wrapJobFn(static_workflow_declaration, config,\n processed_normal.rv(0), processed_normal.rv(1), processed_tumor\n .rv(0), processed_tumor.rv(1))\n job.addChild(processed_normal)\n job.addChild(processed_tumor)\n job.addFollowOn(static_workflow)\n else:\n job.addFollowOnJobFn(static_workflow_declaration, config, config.\n normal_bam, config.normal_bai, config.tumor_bam, config.tumor_bai)\n",
"role": "user"
},
{
"content": "preprocessing_declaration",
"role": "assistant"
}
] |
def x(job, config, normal_bam, normal_bai, tumor_bam, tumor_bai):
memory = '1G' if config.ci_test else '10G'
disk = '1G' if config.ci_test else '75G'
mutect_results, pindel_results, muse_results = None, None, None
if config.run_mutect:
mutect_results = job.addChildJobFn(run_mutect, normal_bam,
normal_bai, tumor_bam, tumor_bai, config.reference, config.dict,
config.fai, config.cosmic, config.dbsnp, cores=1, memory=memory,
disk=disk).rv()
if config.run_pindel:
pindel_results = job.addChildJobFn(run_pindel, normal_bam,
normal_bai, tumor_bam, tumor_bai, config.reference, config.fai,
cores=config.cores, memory=memory, disk=disk).rv()
if config.run_muse:
muse_results = job.addChildJobFn(run_muse, normal_bam, normal_bai,
tumor_bam, tumor_bai, config.reference, config.dict, config.fai,
config.dbsnp, cores=config.cores, memory=memory, disk=disk).rv()
consolidation = job.wrapJobFn(consolidate_output, config,
mutect_results, pindel_results, muse_results)
job.addFollowOn(consolidation)
|
static_workflow_declaration
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(job, config, normal_bam, normal_bai, tumor_bam, tumor_bai):\n memory = '1G' if config.ci_test else '10G'\n disk = '1G' if config.ci_test else '75G'\n mutect_results, pindel_results, muse_results = None, None, None\n if config.run_mutect:\n mutect_results = job.addChildJobFn(run_mutect, normal_bam,\n normal_bai, tumor_bam, tumor_bai, config.reference, config.dict,\n config.fai, config.cosmic, config.dbsnp, cores=1, memory=memory,\n disk=disk).rv()\n if config.run_pindel:\n pindel_results = job.addChildJobFn(run_pindel, normal_bam,\n normal_bai, tumor_bam, tumor_bai, config.reference, config.fai,\n cores=config.cores, memory=memory, disk=disk).rv()\n if config.run_muse:\n muse_results = job.addChildJobFn(run_muse, normal_bam, normal_bai,\n tumor_bam, tumor_bai, config.reference, config.dict, config.fai,\n config.dbsnp, cores=config.cores, memory=memory, disk=disk).rv()\n consolidation = job.wrapJobFn(consolidate_output, config,\n mutect_results, pindel_results, muse_results)\n job.addFollowOn(consolidation)\n",
"role": "user"
},
{
"content": "static_workflow_declaration",
"role": "assistant"
}
] |
def x(job, config, mutect, pindel, muse):
work_dir = job.fileStore.getLocalTempDir()
mutect_tar, pindel_tar, muse_tar = None, None, None
if mutect:
mutect_tar = job.fileStore.readGlobalFile(mutect, os.path.join(
work_dir, 'mutect.tar.gz'))
if pindel:
pindel_tar = job.fileStore.readGlobalFile(pindel, os.path.join(
work_dir, 'pindel.tar.gz'))
if muse:
muse_tar = job.fileStore.readGlobalFile(muse, os.path.join(work_dir,
'muse.tar.gz'))
out_tar = os.path.join(work_dir, config.uuid + '.tar.gz')
tar_list = [x for x in [mutect_tar, pindel_tar, muse_tar] if x is not None]
with tarfile.open(os.path.join(work_dir, out_tar), 'w:gz') as f_out:
for tar in tar_list:
with tarfile.open(tar, 'r') as f_in:
for tarinfo in f_in:
with closing(f_in.extractfile(tarinfo)) as f_in_file:
if tar is mutect_tar:
tarinfo.name = os.path.join(config.uuid,
'mutect', os.path.basename(tarinfo.name))
elif tar is pindel_tar:
tarinfo.name = os.path.join(config.uuid,
'pindel', os.path.basename(tarinfo.name))
else:
tarinfo.name = os.path.join(config.uuid, 'muse',
os.path.basename(tarinfo.name))
f_out.addfile(tarinfo, fileobj=f_in_file)
if urlparse(config.output_dir).scheme == 's3':
job.fileStore.logToMaster('Uploading {} to S3: {}'.format(config.
uuid, config.output_dir))
s3am_upload(job=job, fpath=out_tar, s3_dir=config.output_dir,
num_cores=config.cores)
else:
job.fileStore.logToMaster('Moving {} to output dir: {}'.format(
config.uuid, config.output_dir))
mkdir_p(config.output_dir)
copy_files(file_paths=[out_tar], output_dir=config.output_dir)
|
consolidate_output
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(job, config, mutect, pindel, muse):\n work_dir = job.fileStore.getLocalTempDir()\n mutect_tar, pindel_tar, muse_tar = None, None, None\n if mutect:\n mutect_tar = job.fileStore.readGlobalFile(mutect, os.path.join(\n work_dir, 'mutect.tar.gz'))\n if pindel:\n pindel_tar = job.fileStore.readGlobalFile(pindel, os.path.join(\n work_dir, 'pindel.tar.gz'))\n if muse:\n muse_tar = job.fileStore.readGlobalFile(muse, os.path.join(work_dir,\n 'muse.tar.gz'))\n out_tar = os.path.join(work_dir, config.uuid + '.tar.gz')\n tar_list = [x for x in [mutect_tar, pindel_tar, muse_tar] if x is not None]\n with tarfile.open(os.path.join(work_dir, out_tar), 'w:gz') as f_out:\n for tar in tar_list:\n with tarfile.open(tar, 'r') as f_in:\n for tarinfo in f_in:\n with closing(f_in.extractfile(tarinfo)) as f_in_file:\n if tar is mutect_tar:\n tarinfo.name = os.path.join(config.uuid,\n 'mutect', os.path.basename(tarinfo.name))\n elif tar is pindel_tar:\n tarinfo.name = os.path.join(config.uuid,\n 'pindel', os.path.basename(tarinfo.name))\n else:\n tarinfo.name = os.path.join(config.uuid, 'muse',\n os.path.basename(tarinfo.name))\n f_out.addfile(tarinfo, fileobj=f_in_file)\n if urlparse(config.output_dir).scheme == 's3':\n job.fileStore.logToMaster('Uploading {} to S3: {}'.format(config.\n uuid, config.output_dir))\n s3am_upload(job=job, fpath=out_tar, s3_dir=config.output_dir,\n num_cores=config.cores)\n else:\n job.fileStore.logToMaster('Moving {} to output dir: {}'.format(\n config.uuid, config.output_dir))\n mkdir_p(config.output_dir)\n copy_files(file_paths=[out_tar], output_dir=config.output_dir)\n",
"role": "user"
},
{
"content": "consolidate_output",
"role": "assistant"
}
] |
def x(path_to_manifest):
samples = []
with open(path_to_manifest, 'r') as f:
for line in f.readlines():
if not line.isspace() and not line.startswith('#'):
sample = line.strip().split('\t')
require(len(sample) == 3,
'Bad manifest format! Expected 3 tab separated columns, got: {}'
.format(sample))
uuid, normal, tumor = sample
for url in [normal, tumor]:
require(urlparse(url).scheme and urlparse(url),
'Invalid URL passed for {}'.format(url))
samples.append(sample)
return samples
|
parse_manifest
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path_to_manifest):\n samples = []\n with open(path_to_manifest, 'r') as f:\n for line in f.readlines():\n if not line.isspace() and not line.startswith('#'):\n sample = line.strip().split('\\t')\n require(len(sample) == 3,\n 'Bad manifest format! Expected 3 tab separated columns, got: {}'\n .format(sample))\n uuid, normal, tumor = sample\n for url in [normal, tumor]:\n require(urlparse(url).scheme and urlparse(url),\n 'Invalid URL passed for {}'.format(url))\n samples.append(sample)\n return samples\n",
"role": "user"
},
{
"content": "parse_manifest",
"role": "assistant"
}
] |
def x():
parser = argparse.ArgumentParser(description=main.__doc__,
formatter_class=argparse.RawTextHelpFormatter)
subparsers = parser.add_subparsers(dest='command')
subparsers.add_parser('generate-config', help=
'Generates an editable config in the current working directory.')
subparsers.add_parser('generate-manifest', help=
'Generates an editable manifest in the current working directory.')
subparsers.add_parser('generate', help=
'Generates a config and manifest in the current working directory.')
parser_run = subparsers.add_parser('run', help=
'Runs the CGL exome pipeline')
parser_run.add_argument('--config', default='config-toil-exome.yaml',
type=str, help=
"""Path to the (filled in) config file, generated with "generate-config".
Default value: "%(default)s\""""
)
parser_run.add_argument('--manifest', default='manifest-toil-exome.tsv',
type=str, help=
"""Path to the (filled in) manifest file, generated with "generate-manifest".
Default value: "%(default)s\""""
)
parser_run.add_argument('--normal', default=None, type=str, help=
'URL for the normal BAM. URLs can take the form: http://, ftp://, file://, s3://, and gnos://. The UUID for the sample must be given with the "--uuid" flag.'
)
parser_run.add_argument('--tumor', default=None, type=str, help=
'URL for the tumor BAM. URLs can take the form: http://, ftp://, file://, s3://, and gnos://. The UUID for the sample must be given with the "--uuid" flag.'
)
parser_run.add_argument('--uuid', default=None, type=str, help=
'Provide the UUID of a sample when using the"--tumor" and "--normal" option'
)
if len(sys.argv) == 1:
parser.print_help()
sys.exit(1)
Job.Runner.addToilOptions(parser_run)
args = parser.parse_args()
cwd = os.getcwd()
if args.command == 'generate-config' or args.command == 'generate':
generate_file(os.path.join(cwd, 'config-toil-exome.yaml'),
generate_config)
if args.command == 'generate-manifest' or args.command == 'generate':
generate_file(os.path.join(cwd, 'manifest-toil-exome.tsv'),
generate_manifest)
elif args.command == 'run':
require(os.path.exists(args.config),
'{} not found. Please run "toil-rnaseq generate-config"'.format
(args.config))
if args.normal or args.tumor or args.uuid:
require(args.normal and args.tumor and args.uuid,
'"--tumor", "--normal" and "--uuid" must all be supplied')
samples = [[args.uuid, args.normal, args.tumor]]
else:
samples = parse_manifest(args.manifest)
parsed_config = {x.replace('-', '_'): y for x, y in yaml.load(open(
args.config).read()).iteritems()}
config = argparse.Namespace(**parsed_config)
config.maxCores = int(args.maxCores) if args.maxCores else sys.maxint
if config.preprocessing:
require(config.reference and config.phase and config.mills and
config.dbsnp,
'Missing inputs for preprocessing, check config file.')
if config.run_mutect:
require(config.reference and config.dbsnp and config.cosmic,
'Missing inputs for MuTect, check config file.')
if config.run_pindel:
require(config.reference, 'Missing input (reference) for Pindel.')
if config.run_muse:
require(config.reference and config.dbsnp,
'Missing inputs for MuSe, check config file.')
require(config.output_dir, 'No output location specified: {}'.
format(config.output_dir))
for program in ['curl', 'docker']:
require(next(which(program), None), program +
' must be installed on every node.'.format(program))
Job.Runner.startToil(Job.wrapJobFn(download_shared_files, samples,
config), args)
|
main
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n parser = argparse.ArgumentParser(description=main.__doc__,\n formatter_class=argparse.RawTextHelpFormatter)\n subparsers = parser.add_subparsers(dest='command')\n subparsers.add_parser('generate-config', help=\n 'Generates an editable config in the current working directory.')\n subparsers.add_parser('generate-manifest', help=\n 'Generates an editable manifest in the current working directory.')\n subparsers.add_parser('generate', help=\n 'Generates a config and manifest in the current working directory.')\n parser_run = subparsers.add_parser('run', help=\n 'Runs the CGL exome pipeline')\n parser_run.add_argument('--config', default='config-toil-exome.yaml',\n type=str, help=\n \"\"\"Path to the (filled in) config file, generated with \"generate-config\". \nDefault value: \"%(default)s\\\"\"\"\"\n )\n parser_run.add_argument('--manifest', default='manifest-toil-exome.tsv',\n type=str, help=\n \"\"\"Path to the (filled in) manifest file, generated with \"generate-manifest\". \nDefault value: \"%(default)s\\\"\"\"\"\n )\n parser_run.add_argument('--normal', default=None, type=str, help=\n 'URL for the normal BAM. URLs can take the form: http://, ftp://, file://, s3://, and gnos://. The UUID for the sample must be given with the \"--uuid\" flag.'\n )\n parser_run.add_argument('--tumor', default=None, type=str, help=\n 'URL for the tumor BAM. URLs can take the form: http://, ftp://, file://, s3://, and gnos://. The UUID for the sample must be given with the \"--uuid\" flag.'\n )\n parser_run.add_argument('--uuid', default=None, type=str, help=\n 'Provide the UUID of a sample when using the\"--tumor\" and \"--normal\" option'\n )\n if len(sys.argv) == 1:\n parser.print_help()\n sys.exit(1)\n Job.Runner.addToilOptions(parser_run)\n args = parser.parse_args()\n cwd = os.getcwd()\n if args.command == 'generate-config' or args.command == 'generate':\n generate_file(os.path.join(cwd, 'config-toil-exome.yaml'),\n generate_config)\n if args.command == 'generate-manifest' or args.command == 'generate':\n generate_file(os.path.join(cwd, 'manifest-toil-exome.tsv'),\n generate_manifest)\n elif args.command == 'run':\n require(os.path.exists(args.config),\n '{} not found. Please run \"toil-rnaseq generate-config\"'.format\n (args.config))\n if args.normal or args.tumor or args.uuid:\n require(args.normal and args.tumor and args.uuid,\n '\"--tumor\", \"--normal\" and \"--uuid\" must all be supplied')\n samples = [[args.uuid, args.normal, args.tumor]]\n else:\n samples = parse_manifest(args.manifest)\n parsed_config = {x.replace('-', '_'): y for x, y in yaml.load(open(\n args.config).read()).iteritems()}\n config = argparse.Namespace(**parsed_config)\n config.maxCores = int(args.maxCores) if args.maxCores else sys.maxint\n if config.preprocessing:\n require(config.reference and config.phase and config.mills and\n config.dbsnp,\n 'Missing inputs for preprocessing, check config file.')\n if config.run_mutect:\n require(config.reference and config.dbsnp and config.cosmic,\n 'Missing inputs for MuTect, check config file.')\n if config.run_pindel:\n require(config.reference, 'Missing input (reference) for Pindel.')\n if config.run_muse:\n require(config.reference and config.dbsnp,\n 'Missing inputs for MuSe, check config file.')\n require(config.output_dir, 'No output location specified: {}'.\n format(config.output_dir))\n for program in ['curl', 'docker']:\n require(next(which(program), None), program +\n ' must be installed on every node.'.format(program))\n Job.Runner.startToil(Job.wrapJobFn(download_shared_files, samples,\n config), args)\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x(job, inputs, samples):
shared_ids = {}
urls = [('amb', inputs.amb), ('ann', inputs.ann), ('bwt', inputs.bwt),
('pac', inputs.pac), ('sa', inputs.sa)]
if inputs.alt:
urls.append(('alt', inputs.alt))
download_ref = job.wrapJobFn(download_url_job, inputs.ref, disk='3G')
job.addChild(download_ref)
shared_ids['ref'] = download_ref.rv()
if inputs.fai:
shared_ids['fai'] = job.addChildJobFn(download_url_job, inputs.fai).rv(
)
else:
faidx = job.wrapJobFn(run_samtools_faidx, download_ref.rv())
shared_ids['fai'] = download_ref.addChild(faidx).rv()
if all(x[1] for x in urls):
for name, url in urls:
shared_ids[name] = job.addChildJobFn(download_url_job, url).rv()
else:
job.fileStore.logToMaster('BWA index files not provided, creating now')
bwa_index = job.wrapJobFn(run_bwa_index, download_ref.rv())
download_ref.addChild(bwa_index)
for x, name in enumerate(['amb', 'ann', 'bwt', 'pac', 'sa']):
shared_ids[name] = bwa_index.rv(x)
job.addFollowOnJobFn(map_job, download_sample_and_align, samples,
inputs, shared_ids)
|
download_reference_files
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(job, inputs, samples):\n shared_ids = {}\n urls = [('amb', inputs.amb), ('ann', inputs.ann), ('bwt', inputs.bwt),\n ('pac', inputs.pac), ('sa', inputs.sa)]\n if inputs.alt:\n urls.append(('alt', inputs.alt))\n download_ref = job.wrapJobFn(download_url_job, inputs.ref, disk='3G')\n job.addChild(download_ref)\n shared_ids['ref'] = download_ref.rv()\n if inputs.fai:\n shared_ids['fai'] = job.addChildJobFn(download_url_job, inputs.fai).rv(\n )\n else:\n faidx = job.wrapJobFn(run_samtools_faidx, download_ref.rv())\n shared_ids['fai'] = download_ref.addChild(faidx).rv()\n if all(x[1] for x in urls):\n for name, url in urls:\n shared_ids[name] = job.addChildJobFn(download_url_job, url).rv()\n else:\n job.fileStore.logToMaster('BWA index files not provided, creating now')\n bwa_index = job.wrapJobFn(run_bwa_index, download_ref.rv())\n download_ref.addChild(bwa_index)\n for x, name in enumerate(['amb', 'ann', 'bwt', 'pac', 'sa']):\n shared_ids[name] = bwa_index.rv(x)\n job.addFollowOnJobFn(map_job, download_sample_and_align, samples,\n inputs, shared_ids)\n",
"role": "user"
},
{
"content": "download_reference_files",
"role": "assistant"
}
] |
def x(job, sample, inputs, ids):
uuid, urls = sample
r1_url, r2_url = urls if len(urls) == 2 else (urls[0], None)
job.fileStore.logToMaster(
'Downloaded sample: {0}. R1 {1}\nR2 {2}\nStarting BWA Run'.format(
uuid, r1_url, r2_url))
ids['r1'] = job.addChildJobFn(download_url_job, r1_url, s3_key_path=
inputs.ssec, disk=inputs.file_size).rv()
if r2_url:
ids['r2'] = job.addChildJobFn(download_url_job, r2_url, s3_key_path
=inputs.ssec, disk=inputs.file_size).rv()
else:
ids['r2'] = None
inputs.cores = min(inputs.maxCores, multiprocessing.cpu_count())
inputs.uuid = uuid
config = dict(**vars(inputs))
config.update(ids)
config = argparse.Namespace(**config)
bam_id = job.wrapJobFn(run_bwakit, config, sort=inputs.sort, trim=
inputs.trim, disk=inputs.file_size, cores=inputs.cores)
job.addFollowOn(bam_id)
output_name = uuid + '.bam' + str(inputs.suffix
) if inputs.suffix else uuid + '.bam'
if urlparse(inputs.output_dir).scheme == 's3':
bam_id.addChildJobFn(s3am_upload_job, file_id=bam_id.rv(),
file_name=output_name, s3_dir=inputs.output_dir, s3_key_path=
inputs.ssec, cores=inputs.cores, disk=inputs.file_size)
else:
mkdir_p(inputs.ouput_dir)
bam_id.addChildJobFn(copy_file_job, name=output_name, file_id=
bam_id.rv(), output_dir=inputs.output_dir, disk=inputs.file_size)
|
download_sample_and_align
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(job, sample, inputs, ids):\n uuid, urls = sample\n r1_url, r2_url = urls if len(urls) == 2 else (urls[0], None)\n job.fileStore.logToMaster(\n 'Downloaded sample: {0}. R1 {1}\\nR2 {2}\\nStarting BWA Run'.format(\n uuid, r1_url, r2_url))\n ids['r1'] = job.addChildJobFn(download_url_job, r1_url, s3_key_path=\n inputs.ssec, disk=inputs.file_size).rv()\n if r2_url:\n ids['r2'] = job.addChildJobFn(download_url_job, r2_url, s3_key_path\n =inputs.ssec, disk=inputs.file_size).rv()\n else:\n ids['r2'] = None\n inputs.cores = min(inputs.maxCores, multiprocessing.cpu_count())\n inputs.uuid = uuid\n config = dict(**vars(inputs))\n config.update(ids)\n config = argparse.Namespace(**config)\n bam_id = job.wrapJobFn(run_bwakit, config, sort=inputs.sort, trim=\n inputs.trim, disk=inputs.file_size, cores=inputs.cores)\n job.addFollowOn(bam_id)\n output_name = uuid + '.bam' + str(inputs.suffix\n ) if inputs.suffix else uuid + '.bam'\n if urlparse(inputs.output_dir).scheme == 's3':\n bam_id.addChildJobFn(s3am_upload_job, file_id=bam_id.rv(),\n file_name=output_name, s3_dir=inputs.output_dir, s3_key_path=\n inputs.ssec, cores=inputs.cores, disk=inputs.file_size)\n else:\n mkdir_p(inputs.ouput_dir)\n bam_id.addChildJobFn(copy_file_job, name=output_name, file_id=\n bam_id.rv(), output_dir=inputs.output_dir, disk=inputs.file_size)\n",
"role": "user"
},
{
"content": "download_sample_and_align",
"role": "assistant"
}
] |
def x(manifest_path):
samples = []
with open(manifest_path, 'r') as f:
for line in f:
if not line.isspace() and not line.startswith('#'):
sample = line.strip().split('\t')
require(2 <= len(sample) <= 3,
'Bad manifest format! Expected UUID\tURL1\t[URL2] (tab separated), got: {}'
.format(sample))
uuid = sample[0]
urls = sample[1:]
for url in urls:
require(urlparse(url).scheme and urlparse(url),
'Invalid URL passed for {}'.format(url))
samples.append([uuid, urls])
return samples
|
parse_manifest
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(manifest_path):\n samples = []\n with open(manifest_path, 'r') as f:\n for line in f:\n if not line.isspace() and not line.startswith('#'):\n sample = line.strip().split('\\t')\n require(2 <= len(sample) <= 3,\n 'Bad manifest format! Expected UUID\\tURL1\\t[URL2] (tab separated), got: {}'\n .format(sample))\n uuid = sample[0]\n urls = sample[1:]\n for url in urls:\n require(urlparse(url).scheme and urlparse(url),\n 'Invalid URL passed for {}'.format(url))\n samples.append([uuid, urls])\n return samples\n",
"role": "user"
},
{
"content": "parse_manifest",
"role": "assistant"
}
] |
def x():
parser = argparse.ArgumentParser(description=main.__doc__,
formatter_class=argparse.RawTextHelpFormatter)
subparsers = parser.add_subparsers(dest='command')
subparsers.add_parser('generate-config', help=
'Generates an editable config in the current working directory.')
subparsers.add_parser('generate-manifest', help=
'Generates an editable manifest in the current working directory.')
subparsers.add_parser('generate', help=
'Generates a config and manifest in the current working directory.')
parser_run = subparsers.add_parser('run', help=
'Runs the BWA alignment pipeline')
group = parser_run.add_mutually_exclusive_group()
parser_run.add_argument('--config', default='config-toil-bwa.yaml',
type=str, help=
'Path to the (filled in) config file, generated with "generate-config".'
)
group.add_argument('--manifest', default='manifest-toil-bwa.tsv', type=
str, help=
"""Path to the (filled in) manifest file, generated with "generate-manifest".
Default value: "%(default)s"."""
)
group.add_argument('--sample', nargs='+', action=required_length(2, 3),
help=
'Space delimited sample UUID and fastq files in the format: uuid url1 [url2].'
)
if len(sys.argv) == 1:
parser.print_help()
sys.exit(1)
Job.Runner.addToilOptions(parser_run)
args = parser.parse_args()
cwd = os.getcwd()
if args.command == 'generate-config' or args.command == 'generate':
generate_file(os.path.join(cwd, 'config-toil-bwa.yaml'),
generate_config)
if args.command == 'generate-manifest' or args.command == 'generate':
generate_file(os.path.join(cwd, 'manifest-toil-bwa.tsv'),
generate_manifest)
elif args.command == 'run':
require(os.path.exists(args.config),
'{} not found. Please run generate-config'.format(args.config))
if not args.sample:
args.sample = None
require(os.path.exists(args.manifest),
'{} not found and no sample provided. Please run "generate-manifest"'
.format(args.manifest))
parsed_config = {x.replace('-', '_'): y for x, y in yaml.load(open(
args.config).read()).iteritems()}
config = argparse.Namespace(**parsed_config)
config.maxCores = int(args.maxCores) if args.maxCores else sys.maxint
samples = [args.sample[0], args.sample[1:]
] if args.sample else parse_manifest(args.manifest)
require(config.ref, 'Missing URL for reference file: {}'.format(
config.ref))
require(config.output_dir, 'No output location specified: {}'.
format(config.output_dir))
Job.Runner.startToil(Job.wrapJobFn(download_reference_files, config,
samples), args)
|
main
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n parser = argparse.ArgumentParser(description=main.__doc__,\n formatter_class=argparse.RawTextHelpFormatter)\n subparsers = parser.add_subparsers(dest='command')\n subparsers.add_parser('generate-config', help=\n 'Generates an editable config in the current working directory.')\n subparsers.add_parser('generate-manifest', help=\n 'Generates an editable manifest in the current working directory.')\n subparsers.add_parser('generate', help=\n 'Generates a config and manifest in the current working directory.')\n parser_run = subparsers.add_parser('run', help=\n 'Runs the BWA alignment pipeline')\n group = parser_run.add_mutually_exclusive_group()\n parser_run.add_argument('--config', default='config-toil-bwa.yaml',\n type=str, help=\n 'Path to the (filled in) config file, generated with \"generate-config\".'\n )\n group.add_argument('--manifest', default='manifest-toil-bwa.tsv', type=\n str, help=\n \"\"\"Path to the (filled in) manifest file, generated with \"generate-manifest\". \nDefault value: \"%(default)s\".\"\"\"\n )\n group.add_argument('--sample', nargs='+', action=required_length(2, 3),\n help=\n 'Space delimited sample UUID and fastq files in the format: uuid url1 [url2].'\n )\n if len(sys.argv) == 1:\n parser.print_help()\n sys.exit(1)\n Job.Runner.addToilOptions(parser_run)\n args = parser.parse_args()\n cwd = os.getcwd()\n if args.command == 'generate-config' or args.command == 'generate':\n generate_file(os.path.join(cwd, 'config-toil-bwa.yaml'),\n generate_config)\n if args.command == 'generate-manifest' or args.command == 'generate':\n generate_file(os.path.join(cwd, 'manifest-toil-bwa.tsv'),\n generate_manifest)\n elif args.command == 'run':\n require(os.path.exists(args.config),\n '{} not found. Please run generate-config'.format(args.config))\n if not args.sample:\n args.sample = None\n require(os.path.exists(args.manifest),\n '{} not found and no sample provided. Please run \"generate-manifest\"'\n .format(args.manifest))\n parsed_config = {x.replace('-', '_'): y for x, y in yaml.load(open(\n args.config).read()).iteritems()}\n config = argparse.Namespace(**parsed_config)\n config.maxCores = int(args.maxCores) if args.maxCores else sys.maxint\n samples = [args.sample[0], args.sample[1:]\n ] if args.sample else parse_manifest(args.manifest)\n require(config.ref, 'Missing URL for reference file: {}'.format(\n config.ref))\n require(config.output_dir, 'No output location specified: {}'.\n format(config.output_dir))\n Job.Runner.startToil(Job.wrapJobFn(download_reference_files, config,\n samples), args)\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x(address):
parsed = ipv4.ip2long(address)
if parsed is None:
parsed = ipv6.ip2long(address)
return parsed
|
address2long
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(address):\n parsed = ipv4.ip2long(address)\n if parsed is None:\n parsed = ipv6.ip2long(address)\n return parsed\n",
"role": "user"
},
{
"content": "address2long",
"role": "assistant"
}
] |
def x(self, item):
item = self._cast(item)
offset = item - self.startIp
if offset >= 0 and offset < self._len:
return offset
raise ValueError('%s is not in range' % self._ipver.long2ip(item))
|
index
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, item):\n item = self._cast(item)\n offset = item - self.startIp\n if offset >= 0 and offset < self._len:\n return offset\n raise ValueError('%s is not in range' % self._ipver.long2ip(item))\n",
"role": "user"
},
{
"content": "index",
"role": "assistant"
}
] |
def x():
pid = os.fork()
if pid > 0:
os.waitpid(pid, 0)
return True
os.setsid()
pid = os.fork()
if pid > 0:
os._exit(os.EX_OK)
return False
|
detach_process
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n pid = os.fork()\n if pid > 0:\n os.waitpid(pid, 0)\n return True\n os.setsid()\n pid = os.fork()\n if pid > 0:\n os._exit(os.EX_OK)\n return False\n",
"role": "user"
},
{
"content": "detach_process",
"role": "assistant"
}
] |
def x(predicate, timeout):
if timeout:
if timeout is True:
timeout = float('Inf')
timeout = time.time() + timeout
while not predicate() and time.time() < timeout:
time.sleep(0.1)
return predicate()
|
block
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(predicate, timeout):\n if timeout:\n if timeout is True:\n timeout = float('Inf')\n timeout = time.time() + timeout\n while not predicate() and time.time() < timeout:\n time.sleep(0.1)\n return predicate()\n",
"role": "user"
},
{
"content": "block",
"role": "assistant"
}
] |
def x(self):
try:
with open(self._path, 'r') as f:
s = f.read().strip()
if not s:
return None
return int(s)
except IOError as e:
if e.errno == errno.ENOENT:
return None
raise
|
read_pid
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n try:\n with open(self._path, 'r') as f:\n s = f.read().strip()\n if not s:\n return None\n return int(s)\n except IOError as e:\n if e.errno == errno.ENOENT:\n return None\n raise\n",
"role": "user"
},
{
"content": "read_pid",
"role": "assistant"
}
] |
def x(self):
handles = []
for handler in self.logger.handlers:
for attr in ['sock', 'socket', 'stream']:
try:
handle = getattr(handler, attr)
if handle:
handles.append(handle)
break
except AttributeError:
continue
return handles
|
get_logger_file_handles
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n handles = []\n for handler in self.logger.handlers:\n for attr in ['sock', 'socket', 'stream']:\n try:\n handle = getattr(handler, attr)\n if handle:\n handles.append(handle)\n break\n except AttributeError:\n continue\n return handles\n",
"role": "user"
},
{
"content": "get_logger_file_handles",
"role": "assistant"
}
] |
def x(self):
pid = self.get_pid()
if pid is None:
return False
try:
os.kill(pid, 0)
except OSError as e:
if e.errno == errno.ESRCH:
self.pid_file.release()
return False
return True
|
is_running
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n pid = self.get_pid()\n if pid is None:\n return False\n try:\n os.kill(pid, 0)\n except OSError as e:\n if e.errno == errno.ESRCH:\n self.pid_file.release()\n return False\n return True\n",
"role": "user"
},
{
"content": "is_running",
"role": "assistant"
}
] |
def x(self, s):
try:
return self._signal_events[int(s)]
except KeyError:
raise ValueError('Signal {} has not been enabled'.format(s))
|
get_signal_event
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, s):\n try:\n return self._signal_events[int(s)]\n except KeyError:\n raise ValueError('Signal {} has not been enabled'.format(s))\n",
"role": "user"
},
{
"content": "get_signal_event",
"role": "assistant"
}
] |
def x(self, s):
self._get_signal_event(s)
pid = self.get_pid()
if not pid:
raise ValueError('Daemon is not running.')
os.kill(pid, s)
|
send_signal
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, s):\n self._get_signal_event(s)\n pid = self.get_pid()\n if not pid:\n raise ValueError('Daemon is not running.')\n os.kill(pid, s)\n",
"role": "user"
},
{
"content": "send_signal",
"role": "assistant"
}
] |
def x(self, block=False):
self.send_signal(signal.SIGTERM)
return _block(lambda : not self.is_running(), block)
|
stop
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, block=False):\n self.send_signal(signal.SIGTERM)\n return _block(lambda : not self.is_running(), block)\n",
"role": "user"
},
{
"content": "stop",
"role": "assistant"
}
] |
def x(self, block=False):
pid = self.get_pid()
if not pid:
raise ValueError('Daemon is not running.')
try:
os.kill(pid, signal.SIGKILL)
return _block(lambda : not self.is_running(), block)
except OSError as e:
if e.errno == errno.ESRCH:
raise ValueError('Daemon is not running.')
raise
finally:
self.pid_file.release()
|
kill
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, block=False):\n pid = self.get_pid()\n if not pid:\n raise ValueError('Daemon is not running.')\n try:\n os.kill(pid, signal.SIGKILL)\n return _block(lambda : not self.is_running(), block)\n except OSError as e:\n if e.errno == errno.ESRCH:\n raise ValueError('Daemon is not running.')\n raise\n finally:\n self.pid_file.release()\n",
"role": "user"
},
{
"content": "kill",
"role": "assistant"
}
] |
def x(self, block=False):
pid = self.get_pid()
if pid:
raise ValueError('Daemon is already running at PID %d.' % pid)
try:
self.pid_file.acquire()
finally:
self.pid_file.release()
self.clear_signal(signal.SIGTERM)
if _detach_process():
return _block(lambda : self.is_running(), block)
self._debug('Daemon has detached')
def x(s, frame):
self._debug('Received signal {}'.format(s))
self._signal_events[int(s)].set()
def x():
try:
self.pid_file.acquire()
self._debug('PID file has been acquired')
self._debug('Calling `run`')
self.run()
self._debug('`run` returned without exception')
except Exception as e:
self.logger.exception(e)
except SystemExit:
self._debug('`run` called `sys.exit`')
try:
self.pid_file.release()
self._debug('PID file has been released')
except Exception as e:
self.logger.exception(e)
os._exit(os.EX_OK)
try:
setproctitle.setproctitle(self.name)
self._debug('Process title has been set')
files_preserve = self.files_preserve + self._get_logger_file_handles()
signal_map = {s: on_signal for s in self._signal_events}
signal_map.update({signal.SIGTTIN: None, signal.SIGTTOU: None,
signal.SIGTSTP: None})
with DaemonContext(detach_process=False, signal_map=signal_map,
files_preserve=files_preserve):
self._debug('Daemon context has been established')
thread = threading.Thread(target=runner)
thread.start()
while thread.is_alive():
time.sleep(1)
except Exception as e:
self.logger.exception(e)
os._exit(os.EX_OK)
|
start
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, block=False):\n pid = self.get_pid()\n if pid:\n raise ValueError('Daemon is already running at PID %d.' % pid)\n try:\n self.pid_file.acquire()\n finally:\n self.pid_file.release()\n self.clear_signal(signal.SIGTERM)\n if _detach_process():\n return _block(lambda : self.is_running(), block)\n self._debug('Daemon has detached')\n\n def x(s, frame):\n self._debug('Received signal {}'.format(s))\n self._signal_events[int(s)].set()\n\n def x():\n try:\n self.pid_file.acquire()\n self._debug('PID file has been acquired')\n self._debug('Calling `run`')\n self.run()\n self._debug('`run` returned without exception')\n except Exception as e:\n self.logger.exception(e)\n except SystemExit:\n self._debug('`run` called `sys.exit`')\n try:\n self.pid_file.release()\n self._debug('PID file has been released')\n except Exception as e:\n self.logger.exception(e)\n os._exit(os.EX_OK)\n try:\n setproctitle.setproctitle(self.name)\n self._debug('Process title has been set')\n files_preserve = self.files_preserve + self._get_logger_file_handles()\n signal_map = {s: on_signal for s in self._signal_events}\n signal_map.update({signal.SIGTTIN: None, signal.SIGTTOU: None,\n signal.SIGTSTP: None})\n with DaemonContext(detach_process=False, signal_map=signal_map,\n files_preserve=files_preserve):\n self._debug('Daemon context has been established')\n thread = threading.Thread(target=runner)\n thread.start()\n while thread.is_alive():\n time.sleep(1)\n except Exception as e:\n self.logger.exception(e)\n os._exit(os.EX_OK)\n",
"role": "user"
},
{
"content": "start",
"role": "assistant"
}
] |
def x(opts):
if opts['--address'] and is_boot2docker():
raise DatacatsError('Cannot specify address on boot2docker.')
return create_environment(environment_dir=opts['ENVIRONMENT_DIR'], port
=opts['PORT'], create_skin=not opts['--bare'], start_web=not opts[
'--image-only'], create_sysadmin=not opts['--no-sysadmin'],
site_name=opts['--site'], ckan_version=opts['--ckan'], address=opts
['--address'], log_syslog=opts['--syslog'], datapusher=not opts[
'--no-datapusher'], site_url=opts['--site-url'], interactive=opts[
'--interactive'], init_db=not opts['--no-init-db'])
|
create
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(opts):\n if opts['--address'] and is_boot2docker():\n raise DatacatsError('Cannot specify address on boot2docker.')\n return create_environment(environment_dir=opts['ENVIRONMENT_DIR'], port\n =opts['PORT'], create_skin=not opts['--bare'], start_web=not opts[\n '--image-only'], create_sysadmin=not opts['--no-sysadmin'],\n site_name=opts['--site'], ckan_version=opts['--ckan'], address=opts\n ['--address'], log_syslog=opts['--syslog'], datapusher=not opts[\n '--no-datapusher'], site_url=opts['--site-url'], interactive=opts[\n '--interactive'], init_db=not opts['--no-init-db'])\n",
"role": "user"
},
{
"content": "create",
"role": "assistant"
}
] |
def x(opts, no_install=False, quiet=False):
if opts['--address'] and is_boot2docker():
raise DatacatsError('Cannot specify address on boot2docker.')
environment_dir = opts['ENVIRONMENT_DIR']
port = opts['PORT']
address = opts['--address']
start_web = not opts['--image-only']
create_sysadmin = not opts['--no-sysadmin']
site_name = opts['--site']
site_url = opts['--site-url']
interactive = opts['--interactive']
init_db = not opts['--no-init-db']
environment_dir = abspath(environment_dir or '.')
log_syslog = opts['--syslog']
environment = Environment.load(environment_dir, site_name)
if address:
environment.address = address
if port:
environment.port = int(port)
if site_url:
environment.site_url = site_url
try:
if environment.sites and site_name in environment.sites:
raise DatacatsError('Site named {0} already exists.'.format(
site_name))
making_full_environment = not environment.data_exists()
if not quiet:
write(
'Creating environment {0}/{1} from existing environment directory "{0}"'
.format(environment.name, environment.site_name))
steps = [lambda : environment.create_directories(create_project_dir
=False)] + ([environment.save, environment.create_virtualenv] if
making_full_environment else []) + [environment.save_site,
environment.start_supporting_containers, environment.
fix_storage_permissions]
for fn in steps:
fn()
if not quiet:
write('.')
if not quiet:
write('\n')
except:
if not quiet:
print
raise
return finish_init(environment, start_web, create_sysadmin, log_syslog=
log_syslog, do_install=not no_install, quiet=quiet, site_url=
site_url, interactive=interactive, init_db=init_db)
|
init
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(opts, no_install=False, quiet=False):\n if opts['--address'] and is_boot2docker():\n raise DatacatsError('Cannot specify address on boot2docker.')\n environment_dir = opts['ENVIRONMENT_DIR']\n port = opts['PORT']\n address = opts['--address']\n start_web = not opts['--image-only']\n create_sysadmin = not opts['--no-sysadmin']\n site_name = opts['--site']\n site_url = opts['--site-url']\n interactive = opts['--interactive']\n init_db = not opts['--no-init-db']\n environment_dir = abspath(environment_dir or '.')\n log_syslog = opts['--syslog']\n environment = Environment.load(environment_dir, site_name)\n if address:\n environment.address = address\n if port:\n environment.port = int(port)\n if site_url:\n environment.site_url = site_url\n try:\n if environment.sites and site_name in environment.sites:\n raise DatacatsError('Site named {0} already exists.'.format(\n site_name))\n making_full_environment = not environment.data_exists()\n if not quiet:\n write(\n 'Creating environment {0}/{1} from existing environment directory \"{0}\"'\n .format(environment.name, environment.site_name))\n steps = [lambda : environment.create_directories(create_project_dir\n =False)] + ([environment.save, environment.create_virtualenv] if\n making_full_environment else []) + [environment.save_site,\n environment.start_supporting_containers, environment.\n fix_storage_permissions]\n for fn in steps:\n fn()\n if not quiet:\n write('.')\n if not quiet:\n write('\\n')\n except:\n if not quiet:\n print\n raise\n return finish_init(environment, start_web, create_sysadmin, log_syslog=\n log_syslog, do_install=not no_install, quiet=quiet, site_url=\n site_url, interactive=interactive, init_db=init_db)\n",
"role": "user"
},
{
"content": "init",
"role": "assistant"
}
] |
def x(environment, start_web, create_sysadmin, log_syslog=False, do_install
=True, quiet=False, site_url=None, interactive=False, init_db=True):
if not init_db:
start_web = False
create_sysadmin = False
if do_install:
install_all(environment, False, verbose=False, quiet=quiet)
if init_db:
if not quiet:
write('Initializing database')
environment.install_postgis_sql()
environment.ckan_db_init()
if not quiet:
write('\n')
if site_url:
try:
site_url = site_url.format(address=environment.address, port=
environment.port)
environment.site_url = site_url
environment.save_site(False)
except (KeyError, IndexError, ValueError) as e:
raise DatacatsError('Could not parse site_url: {}'.format(e))
if start_web:
environment.start_ckan(log_syslog=log_syslog)
if not quiet and not interactive:
write('Starting web server at {0} ...\n'.format(environment.
web_address()))
if create_sysadmin:
try:
adminpw = confirm_password()
environment.create_admin_set_password(adminpw)
except KeyboardInterrupt:
print
if not start_web:
environment.stop_supporting_containers()
|
finish_init
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(environment, start_web, create_sysadmin, log_syslog=False, do_install\n =True, quiet=False, site_url=None, interactive=False, init_db=True):\n if not init_db:\n start_web = False\n create_sysadmin = False\n if do_install:\n install_all(environment, False, verbose=False, quiet=quiet)\n if init_db:\n if not quiet:\n write('Initializing database')\n environment.install_postgis_sql()\n environment.ckan_db_init()\n if not quiet:\n write('\\n')\n if site_url:\n try:\n site_url = site_url.format(address=environment.address, port=\n environment.port)\n environment.site_url = site_url\n environment.save_site(False)\n except (KeyError, IndexError, ValueError) as e:\n raise DatacatsError('Could not parse site_url: {}'.format(e))\n if start_web:\n environment.start_ckan(log_syslog=log_syslog)\n if not quiet and not interactive:\n write('Starting web server at {0} ...\\n'.format(environment.\n web_address()))\n if create_sysadmin:\n try:\n adminpw = confirm_password()\n environment.create_admin_set_password(adminpw)\n except KeyboardInterrupt:\n print\n if not start_web:\n environment.stop_supporting_containers()\n",
"role": "user"
},
{
"content": "finish_init",
"role": "assistant"
}
] |
def x(self):
config = self.profiledir + '/config'
if not isdir(self.profiledir):
makedirs(self.profiledir)
cp = SafeConfigParser()
cp.add_section('ssh')
cp.set('ssh', 'private_key', self.ssh_private_key)
cp.set('ssh', 'public_key', self.ssh_public_key)
with open(config, 'w') as cfile:
cp.write(cfile)
|
save
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n config = self.profiledir + '/config'\n if not isdir(self.profiledir):\n makedirs(self.profiledir)\n cp = SafeConfigParser()\n cp.add_section('ssh')\n cp.set('ssh', 'private_key', self.ssh_private_key)\n cp.set('ssh', 'public_key', self.ssh_public_key)\n with open(config, 'w') as cfile:\n cp.write(cfile)\n",
"role": "user"
},
{
"content": "save",
"role": "assistant"
}
] |
def x(self):
web_command(command=['ssh-keygen', '-q', '-t', 'rsa', '-N', '', '-C',
'datacats generated {0}@{1}'.format(getuser(), gethostname()), '-f',
'/output/id_rsa'], rw={self.profiledir: '/output'})
|
generate_ssh_key
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n web_command(command=['ssh-keygen', '-q', '-t', 'rsa', '-N', '', '-C',\n 'datacats generated {0}@{1}'.format(getuser(), gethostname()), '-f',\n '/output/id_rsa'], rw={self.profiledir: '/output'})\n",
"role": "user"
},
{
"content": "generate_ssh_key",
"role": "assistant"
}
] |
def x(self, environment, target_name):
remote_server_command(['ssh', environment.deploy_target, 'create',
target_name], environment, self, clean_up=True)
|
create
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, environment, target_name):\n remote_server_command(['ssh', environment.deploy_target, 'create',\n target_name], environment, self, clean_up=True)\n",
"role": "user"
},
{
"content": "create",
"role": "assistant"
}
] |
def x(self, environment, target_name, password):
try:
remote_server_command(['ssh', environment.deploy_target,
'admin_password', target_name, password], environment, self,
clean_up=True)
return True
except WebCommandError:
return False
|
admin_password
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, environment, target_name, password):\n try:\n remote_server_command(['ssh', environment.deploy_target,\n 'admin_password', target_name, password], environment, self,\n clean_up=True)\n return True\n except WebCommandError:\n return False\n",
"role": "user"
},
{
"content": "admin_password",
"role": "assistant"
}
] |
def x(self, environment, target_name, stream_output=None):
try:
remote_server_command(['rsync', '-lrv', '--safe-links',
'--munge-links', '--delete', '--inplace', '--chmod=ugo=rwX',
'--exclude=.datacats-environment', '--exclude=.git',
'/project/.', environment.deploy_target + ':' + target_name],
environment, self, include_project_dir=True, stream_output=
stream_output, clean_up=True)
except WebCommandError as e:
raise DatacatsError(
"""Unable to deploy `{0}` to remote server for some reason:
datacats was not able to copy data to the remote server"""
.format((target_name,)), parent_exception=e)
try:
remote_server_command(['ssh', environment.deploy_target, 'install',
target_name], environment, self, clean_up=True)
return True
except WebCommandError as e:
raise DatacatsError(
"""Unable to deploy `{0}` to remote server for some reason:
datacats copied data to the server but failed to register
(or `install`) the new catalog"""
.format((target_name,)), parent_exception=e)
|
deploy
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, environment, target_name, stream_output=None):\n try:\n remote_server_command(['rsync', '-lrv', '--safe-links',\n '--munge-links', '--delete', '--inplace', '--chmod=ugo=rwX',\n '--exclude=.datacats-environment', '--exclude=.git',\n '/project/.', environment.deploy_target + ':' + target_name],\n environment, self, include_project_dir=True, stream_output=\n stream_output, clean_up=True)\n except WebCommandError as e:\n raise DatacatsError(\n \"\"\"Unable to deploy `{0}` to remote server for some reason:\n datacats was not able to copy data to the remote server\"\"\"\n .format((target_name,)), parent_exception=e)\n try:\n remote_server_command(['ssh', environment.deploy_target, 'install',\n target_name], environment, self, clean_up=True)\n return True\n except WebCommandError as e:\n raise DatacatsError(\n \"\"\"Unable to deploy `{0}` to remote server for some reason:\ndatacats copied data to the server but failed to register\n(or `install`) the new catalog\"\"\"\n .format((target_name,)), parent_exception=e)\n",
"role": "user"
},
{
"content": "deploy",
"role": "assistant"
}
] |
def x(n, batch_size):
b = n // batch_size
if n % batch_size > 0:
b += 1
return b
|
num_batches
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(n, batch_size):\n b = n // batch_size\n if n % batch_size > 0:\n b += 1\n return b\n",
"role": "user"
},
{
"content": "num_batches",
"role": "assistant"
}
] |
def x(self):
if self.repeats == -1:
return np.inf
else:
return self.length * self.repeats
|
num_indices_generated
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.repeats == -1:\n return np.inf\n else:\n return self.length * self.repeats\n",
"role": "user"
},
{
"content": "num_indices_generated",
"role": "assistant"
}
] |
def x(self, batch_size):
if self.repeats == 1:
for i in range(0, self.length, batch_size):
yield np.arange(i, min(i + batch_size, self.length))
else:
repeats = self.repeats
i = 0
while True:
j = i + batch_size
if j <= self.length:
yield np.arange(i, j)
i = j
elif j <= self.length * 2:
if repeats != -1:
repeats -= 1
if repeats == 0:
if i < self.length:
yield np.arange(i, self.length)
break
k = batch_size - (self.length - i)
yield np.append(np.arange(i, self.length), np.arange(0, k),
axis=0)
i = k
else:
batch_ndx = np.arange(0)
while len(batch_ndx) < batch_size:
k = min(batch_size - len(batch_ndx), self.length - i)
batch_ndx = np.append(batch_ndx, np.arange(i, i + k),
axis=0)
i += k
if i >= self.length:
i -= self.length
if repeats != -1:
repeats -= 1
if repeats == 0:
break
if len(batch_ndx) > 0:
yield batch_ndx
if repeats == 0:
break
|
in_order_indices_batch_iterator
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, batch_size):\n if self.repeats == 1:\n for i in range(0, self.length, batch_size):\n yield np.arange(i, min(i + batch_size, self.length))\n else:\n repeats = self.repeats\n i = 0\n while True:\n j = i + batch_size\n if j <= self.length:\n yield np.arange(i, j)\n i = j\n elif j <= self.length * 2:\n if repeats != -1:\n repeats -= 1\n if repeats == 0:\n if i < self.length:\n yield np.arange(i, self.length)\n break\n k = batch_size - (self.length - i)\n yield np.append(np.arange(i, self.length), np.arange(0, k),\n axis=0)\n i = k\n else:\n batch_ndx = np.arange(0)\n while len(batch_ndx) < batch_size:\n k = min(batch_size - len(batch_ndx), self.length - i)\n batch_ndx = np.append(batch_ndx, np.arange(i, i + k),\n axis=0)\n i += k\n if i >= self.length:\n i -= self.length\n if repeats != -1:\n repeats -= 1\n if repeats == 0:\n break\n if len(batch_ndx) > 0:\n yield batch_ndx\n if repeats == 0:\n break\n",
"role": "user"
},
{
"content": "in_order_indices_batch_iterator",
"role": "assistant"
}
] |
def x(self, batch_size, shuffle_rng):
if self.repeats == 1:
indices = shuffle_rng.permutation(self.length)
for i in range(0, self.length, batch_size):
yield indices[i:i + batch_size]
else:
repeats = self.repeats
indices = shuffle_rng.permutation(self.length)
i = 0
while True:
j = i + batch_size
if j <= self.length:
yield indices[i:j]
i = j
else:
batch_ndx = np.arange(0)
while len(batch_ndx) < batch_size:
k = min(batch_size - len(batch_ndx), self.length - i)
batch_ndx = np.append(batch_ndx, indices[i:i + k], axis=0)
i += k
if i >= self.length:
indices = shuffle_rng.permutation(self.length)
i -= self.length
if repeats != -1:
repeats -= 1
if repeats == 0:
break
if len(batch_ndx) > 0:
yield batch_ndx
if repeats == 0:
break
|
shuffled_indices_batch_iterator
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, batch_size, shuffle_rng):\n if self.repeats == 1:\n indices = shuffle_rng.permutation(self.length)\n for i in range(0, self.length, batch_size):\n yield indices[i:i + batch_size]\n else:\n repeats = self.repeats\n indices = shuffle_rng.permutation(self.length)\n i = 0\n while True:\n j = i + batch_size\n if j <= self.length:\n yield indices[i:j]\n i = j\n else:\n batch_ndx = np.arange(0)\n while len(batch_ndx) < batch_size:\n k = min(batch_size - len(batch_ndx), self.length - i)\n batch_ndx = np.append(batch_ndx, indices[i:i + k], axis=0)\n i += k\n if i >= self.length:\n indices = shuffle_rng.permutation(self.length)\n i -= self.length\n if repeats != -1:\n repeats -= 1\n if repeats == 0:\n break\n if len(batch_ndx) > 0:\n yield batch_ndx\n if repeats == 0:\n break\n",
"role": "user"
},
{
"content": "shuffled_indices_batch_iterator",
"role": "assistant"
}
] |
def x(self, batch_size, shuffle_rng):
while True:
yield shuffle_rng.choice(len(self.weights), size=(batch_size,), p=
self.weights)
|
shuffled_indices_batch_iterator
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, batch_size, shuffle_rng):\n while True:\n yield shuffle_rng.choice(len(self.weights), size=(batch_size,), p=\n self.weights)\n",
"role": "user"
},
{
"content": "shuffled_indices_batch_iterator",
"role": "assistant"
}
] |
def x(y):
h = np.bincount(y)
cls_weight = 1.0 / (h.astype(float) * len(np.nonzero(h)[0]))
cls_weight[np.isnan(cls_weight)] = 0.0
sample_weight = cls_weight[y]
return sample_weight
|
class_balancing_sample_weights
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(y):\n h = np.bincount(y)\n cls_weight = 1.0 / (h.astype(float) * len(np.nonzero(h)[0]))\n cls_weight[np.isnan(cls_weight)] = 0.0\n sample_weight = cls_weight[y]\n return sample_weight\n",
"role": "user"
},
{
"content": "class_balancing_sample_weights",
"role": "assistant"
}
] |
def x(self, batch_size, shuffle_rng):
while True:
yield shuffle_rng.choice(self.indices, size=(batch_size,), p=self.
sub_weights)
|
shuffled_indices_batch_iterator
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, batch_size, shuffle_rng):\n while True:\n yield shuffle_rng.choice(self.indices, size=(batch_size,), p=self.\n sub_weights)\n",
"role": "user"
},
{
"content": "shuffled_indices_batch_iterator",
"role": "assistant"
}
] |
def x(y, indices):
weights = WeightedSampler.class_balancing_sample_weights(y[indices])
return WeightedSubsetSampler(weights, indices=indices)
|
class_balancing_sampler
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(y, indices):\n weights = WeightedSampler.class_balancing_sample_weights(y[indices])\n return WeightedSubsetSampler(weights, indices=indices)\n",
"role": "user"
},
{
"content": "class_balancing_sampler",
"role": "assistant"
}
] |
def x(container, url, timeout):
start = time.time()
remaining = timeout
while True:
remaining = start + timeout - time.time()
if remaining < 0:
raise ServiceTimeout
try:
response = get(url, timeout=min(remaining, REQUEST_TIMEOUT_SECONDS)
)
if 500 <= response.status_code < 600:
return False
return True
except (ConnectionError, Timeout):
pass
if not inspect_container(container)['State']['Running']:
return False
remaining = start + timeout - time.time()
delay = max(0, min(RETRY_DELAY_SECONDS, remaining))
time.sleep(delay)
raise ServiceTimeout
|
wait_for_service_available
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(container, url, timeout):\n start = time.time()\n remaining = timeout\n while True:\n remaining = start + timeout - time.time()\n if remaining < 0:\n raise ServiceTimeout\n try:\n response = get(url, timeout=min(remaining, REQUEST_TIMEOUT_SECONDS)\n )\n if 500 <= response.status_code < 600:\n return False\n return True\n except (ConnectionError, Timeout):\n pass\n if not inspect_container(container)['State']['Running']:\n return False\n remaining = start + timeout - time.time()\n delay = max(0, min(RETRY_DELAY_SECONDS, remaining))\n time.sleep(delay)\n raise ServiceTimeout\n",
"role": "user"
},
{
"content": "wait_for_service_available",
"role": "assistant"
}
] |
def x(filename):
if os.path.isabs(filename):
return filename
else:
return os.path.join(get_data_dir(), filename)
|
get_data_path
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename):\n if os.path.isabs(filename):\n return filename\n else:\n return os.path.join(get_data_dir(), filename)\n",
"role": "user"
},
{
"content": "get_data_path",
"role": "assistant"
}
] |
def x(path, source_url):
dir_path = os.path.dirname(path)
if not os.path.exists(dir_path):
os.makedirs(dir_path)
if not os.path.exists(path):
print('Downloading {} to {}'.format(source_url, path))
filename = source_url.split('/')[-1]
def x(count, block_size, total_size):
sys.stdout.write('\rDownloading {} {:.2%}'.format(filename,
float(count * block_size) / float(total_size)))
sys.stdout.flush()
try:
urlretrieve(source_url, path, reporthook=_progress)
except:
sys.stdout.write('\r')
if os.path.exists(path):
os.remove(path)
raise
sys.stdout.write('\r')
return path
|
download
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path, source_url):\n dir_path = os.path.dirname(path)\n if not os.path.exists(dir_path):\n os.makedirs(dir_path)\n if not os.path.exists(path):\n print('Downloading {} to {}'.format(source_url, path))\n filename = source_url.split('/')[-1]\n\n def x(count, block_size, total_size):\n sys.stdout.write('\\rDownloading {} {:.2%}'.format(filename, \n float(count * block_size) / float(total_size)))\n sys.stdout.flush()\n try:\n urlretrieve(source_url, path, reporthook=_progress)\n except:\n sys.stdout.write('\\r')\n if os.path.exists(path):\n os.remove(path)\n raise\n sys.stdout.write('\\r')\n return path\n",
"role": "user"
},
{
"content": "download",
"role": "assistant"
}
] |
def x(path):
hasher = hashlib.sha256()
with open(path, 'rb') as f:
for chunk in iter(lambda : f.read(10 * 1024 * 1024), b''):
hasher.update(chunk)
return hasher.hexdigest()
|
compute_sha256
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path):\n hasher = hashlib.sha256()\n with open(path, 'rb') as f:\n for chunk in iter(lambda : f.read(10 * 1024 * 1024), b''):\n hasher.update(chunk)\n return hasher.hexdigest()\n",
"role": "user"
},
{
"content": "compute_sha256",
"role": "assistant"
}
] |
def x(path, sha256):
if not os.path.isfile(path):
return False
digest = compute_sha256(path)
if sha256 is None:
print('SHA-256 of {}:'.format(path))
print(' "{}"'.format(digest))
elif digest != sha256:
return False
return True
|
verify_file
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path, sha256):\n if not os.path.isfile(path):\n return False\n digest = compute_sha256(path)\n if sha256 is None:\n print('SHA-256 of {}:'.format(path))\n print(' \"{}\"'.format(digest))\n elif digest != sha256:\n return False\n return True\n",
"role": "user"
},
{
"content": "verify_file",
"role": "assistant"
}
] |
def x(path, source_url, sha256):
if os.path.exists(path):
if sha256 is None:
print('The SHA-256 of {} is "{}"'.format(path, compute_sha256(
path)))
return path
unverified_path = path + '.unverified'
for i in range(_MAX_DOWNLOAD_TRIES):
try:
unverified_path = download(unverified_path, source_url)
except Exception as e:
print(
'Download of {} unsuccessful; error {}; deleting and re-trying...'
.format(source_url, e))
if os.path.exists(unverified_path):
os.remove(unverified_path)
else:
if os.path.exists(unverified_path):
if verify_file(unverified_path, sha256):
os.rename(unverified_path, path)
return path
else:
print(
'Download of {} unsuccessful; verification failed; deleting and re-trying...'
.format(source_url))
os.remove(unverified_path)
print('Did not succeed in downloading {} (tried {} times)'.format(
source_url, _MAX_DOWNLOAD_TRIES))
return None
|
download_and_verify
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path, source_url, sha256):\n if os.path.exists(path):\n if sha256 is None:\n print('The SHA-256 of {} is \"{}\"'.format(path, compute_sha256(\n path)))\n return path\n unverified_path = path + '.unverified'\n for i in range(_MAX_DOWNLOAD_TRIES):\n try:\n unverified_path = download(unverified_path, source_url)\n except Exception as e:\n print(\n 'Download of {} unsuccessful; error {}; deleting and re-trying...'\n .format(source_url, e))\n if os.path.exists(unverified_path):\n os.remove(unverified_path)\n else:\n if os.path.exists(unverified_path):\n if verify_file(unverified_path, sha256):\n os.rename(unverified_path, path)\n return path\n else:\n print(\n 'Download of {} unsuccessful; verification failed; deleting and re-trying...'\n .format(source_url))\n os.remove(unverified_path)\n print('Did not succeed in downloading {} (tried {} times)'.format(\n source_url, _MAX_DOWNLOAD_TRIES))\n return None\n",
"role": "user"
},
{
"content": "download_and_verify",
"role": "assistant"
}
] |
def x(path, source_path, sha256):
if os.path.exists(path):
if sha256 is None:
print('The SHA-256 of {} is "{}"'.format(path, compute_sha256(
path)))
return path
if not os.path.exists(source_path):
return None
unverified_path = path + '.unverified'
dir_path = os.path.dirname(path)
if not os.path.exists(dir_path):
os.makedirs(dir_path)
shutil.copy(source_path, unverified_path)
if os.path.exists(unverified_path):
if verify_file(unverified_path, sha256):
os.rename(unverified_path, path)
return path
else:
print('SHA verification of file {} failed'.format(source_path))
os.remove(unverified_path)
return None
|
copy_and_verify
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path, source_path, sha256):\n if os.path.exists(path):\n if sha256 is None:\n print('The SHA-256 of {} is \"{}\"'.format(path, compute_sha256(\n path)))\n return path\n if not os.path.exists(source_path):\n return None\n unverified_path = path + '.unverified'\n dir_path = os.path.dirname(path)\n if not os.path.exists(dir_path):\n os.makedirs(dir_path)\n shutil.copy(source_path, unverified_path)\n if os.path.exists(unverified_path):\n if verify_file(unverified_path, sha256):\n os.rename(unverified_path, path)\n return path\n else:\n print('SHA verification of file {} failed'.format(source_path))\n os.remove(unverified_path)\n return None\n",
"role": "user"
},
{
"content": "copy_and_verify",
"role": "assistant"
}
] |
def x(name, target):
setupdir = '{0}/ckanext-{1}theme'.format(target, name)
extdir = setupdir + '/ckanext/{0}theme'.format(name)
templatedir = extdir + '/templates/'
staticdir = extdir + '/static/datacats'
makedirs(templatedir + '/home/snippets')
makedirs(staticdir)
here = dirname(__file__)
copyfile(here + '/images/chart.png', staticdir + '/chart.png')
copyfile(here + '/images/datacats-footer.png', staticdir +
'/datacats-footer.png')
filecontents = [(setupdir + '/setup.py', SETUP_PY), (setupdir +
'/.gitignore', DOT_GITIGNORE), (setupdir + '/ckanext/__init__.py',
NAMESPACE_PACKAGE), (extdir + '/__init__.py', ''), (extdir +
'/plugins.py', PLUGINS_PY), (templatedir +
'/home/snippets/promoted.html', PROMOTED_SNIPPET), (templatedir +
'/footer.html', FOOTER_HTML)]
for filename, content in filecontents:
with open(filename, 'w') as f:
f.write(content.replace('##name##', name))
|
ckan_extension_template
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(name, target):\n setupdir = '{0}/ckanext-{1}theme'.format(target, name)\n extdir = setupdir + '/ckanext/{0}theme'.format(name)\n templatedir = extdir + '/templates/'\n staticdir = extdir + '/static/datacats'\n makedirs(templatedir + '/home/snippets')\n makedirs(staticdir)\n here = dirname(__file__)\n copyfile(here + '/images/chart.png', staticdir + '/chart.png')\n copyfile(here + '/images/datacats-footer.png', staticdir +\n '/datacats-footer.png')\n filecontents = [(setupdir + '/setup.py', SETUP_PY), (setupdir +\n '/.gitignore', DOT_GITIGNORE), (setupdir + '/ckanext/__init__.py',\n NAMESPACE_PACKAGE), (extdir + '/__init__.py', ''), (extdir +\n '/plugins.py', PLUGINS_PY), (templatedir +\n '/home/snippets/promoted.html', PROMOTED_SNIPPET), (templatedir +\n '/footer.html', FOOTER_HTML)]\n for filename, content in filecontents:\n with open(filename, 'w') as f:\n f.write(content.replace('##name##', name))\n",
"role": "user"
},
{
"content": "ckan_extension_template",
"role": "assistant"
}
] |
def x(environment, opts):
environment.require_data()
environment.start_supporting_containers()
return environment.interactive_shell(opts['COMMAND'], detach=opts[
'--detach'])
|
shell
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(environment, opts):\n environment.require_data()\n environment.start_supporting_containers()\n return environment.interactive_shell(opts['COMMAND'], detach=opts[\n '--detach'])\n",
"role": "user"
},
{
"content": "shell",
"role": "assistant"
}
] |
def x(opts):
environment = Environment.load('.')
environment.require_data()
environment.start_supporting_containers()
if not opts['COMMAND']:
opts['COMMAND'] = ['--', 'help']
assert opts['COMMAND'][0] == '--'
return environment.interactive_shell(opts['COMMAND'][1:], paster=True,
detach=opts['--detach'])
|
paster
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(opts):\n environment = Environment.load('.')\n environment.require_data()\n environment.start_supporting_containers()\n if not opts['COMMAND']:\n opts['COMMAND'] = ['--', 'help']\n assert opts['COMMAND'][0] == '--'\n return environment.interactive_shell(opts['COMMAND'][1:], paster=True,\n detach=opts['--detach'])\n",
"role": "user"
},
{
"content": "paster",
"role": "assistant"
}
] |
def x(site_name, sitedir, srcdir, port, address, site_url, passwords):
cp = ConfigParser.SafeConfigParser()
cp.read([srcdir + '/.datacats-environment'])
section_name = 'site_' + site_name
if not cp.has_section(section_name):
cp.add_section(section_name)
cp.set(section_name, 'port', str(port))
if address:
cp.set(section_name, 'address', address)
if site_url:
cp.set(section_name, 'site_url', site_url)
with open(srcdir + '/.datacats-environment', 'w') as config:
cp.write(config)
cp = ConfigParser.SafeConfigParser()
cp.add_section('passwords')
for n in sorted(passwords):
cp.set('passwords', n.lower(), passwords[n])
with open(sitedir + '/passwords.ini', 'w') as config:
cp.write(config)
|
save_new_site
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(site_name, sitedir, srcdir, port, address, site_url, passwords):\n cp = ConfigParser.SafeConfigParser()\n cp.read([srcdir + '/.datacats-environment'])\n section_name = 'site_' + site_name\n if not cp.has_section(section_name):\n cp.add_section(section_name)\n cp.set(section_name, 'port', str(port))\n if address:\n cp.set(section_name, 'address', address)\n if site_url:\n cp.set(section_name, 'site_url', site_url)\n with open(srcdir + '/.datacats-environment', 'w') as config:\n cp.write(config)\n cp = ConfigParser.SafeConfigParser()\n cp.add_section('passwords')\n for n in sorted(passwords):\n cp.set('passwords', n.lower(), passwords[n])\n with open(sitedir + '/passwords.ini', 'w') as config:\n cp.write(config)\n",
"role": "user"
},
{
"content": "save_new_site",
"role": "assistant"
}
] |
def x(name, datadir, srcdir, ckan_version, deploy_target=None, always_prod=
False):
with open(datadir + '/.version', 'w') as f:
f.write('2')
cp = ConfigParser.SafeConfigParser()
cp.read(srcdir + '/.datacats-environment')
if not cp.has_section('datacats'):
cp.add_section('datacats')
cp.set('datacats', 'name', name)
cp.set('datacats', 'ckan_version', ckan_version)
if deploy_target:
if not cp.has_section('deploy'):
cp.add_section('deploy')
cp.set('deploy', 'target', deploy_target)
if always_prod:
cp.set('datacats', 'always_prod', 'true')
with open(srcdir + '/.datacats-environment', 'w') as config:
cp.write(config)
save_srcdir_location(datadir, srcdir)
|
save_new_environment
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(name, datadir, srcdir, ckan_version, deploy_target=None, always_prod=\n False):\n with open(datadir + '/.version', 'w') as f:\n f.write('2')\n cp = ConfigParser.SafeConfigParser()\n cp.read(srcdir + '/.datacats-environment')\n if not cp.has_section('datacats'):\n cp.add_section('datacats')\n cp.set('datacats', 'name', name)\n cp.set('datacats', 'ckan_version', ckan_version)\n if deploy_target:\n if not cp.has_section('deploy'):\n cp.add_section('deploy')\n cp.set('deploy', 'target', deploy_target)\n if always_prod:\n cp.set('datacats', 'always_prod', 'true')\n with open(srcdir + '/.datacats-environment', 'w') as config:\n cp.write(config)\n save_srcdir_location(datadir, srcdir)\n",
"role": "user"
},
{
"content": "save_new_environment",
"role": "assistant"
}
] |
def x(environment_name=None, data_only=False):
docker.require_images()
if environment_name is None:
environment_name = '.'
extension_dir = 'ckan'
if validate.valid_name(environment_name) and path.isdir(path.expanduser
('~/.datacats/' + environment_name)):
datadir = path.expanduser('~/.datacats/' + environment_name)
with open(datadir + '/project-dir') as pd:
srcdir = pd.read()
if not data_only and not path.exists(srcdir + '/.datacats-environment'
):
raise DatacatsError(
'Environment data found but environment directory is missing. Try again from the new environment directory location or remove the environment data with "datacats purge"'
)
return srcdir, extension_dir, datadir
srcdir = path.abspath(environment_name)
if not path.isdir(srcdir):
raise DatacatsError('No environment found with that name')
wd = srcdir
oldwd = None
while not path.exists(wd + '/.datacats-environment'):
oldwd = wd
wd, _ = path.split(wd)
if wd == oldwd:
raise DatacatsError('Environment not found in {0} or above'.
format(srcdir))
srcdir = wd
if oldwd:
_, extension_dir = path.split(oldwd)
return srcdir, extension_dir, None
|
find_environment_dirs
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(environment_name=None, data_only=False):\n docker.require_images()\n if environment_name is None:\n environment_name = '.'\n extension_dir = 'ckan'\n if validate.valid_name(environment_name) and path.isdir(path.expanduser\n ('~/.datacats/' + environment_name)):\n datadir = path.expanduser('~/.datacats/' + environment_name)\n with open(datadir + '/project-dir') as pd:\n srcdir = pd.read()\n if not data_only and not path.exists(srcdir + '/.datacats-environment'\n ):\n raise DatacatsError(\n 'Environment data found but environment directory is missing. Try again from the new environment directory location or remove the environment data with \"datacats purge\"'\n )\n return srcdir, extension_dir, datadir\n srcdir = path.abspath(environment_name)\n if not path.isdir(srcdir):\n raise DatacatsError('No environment found with that name')\n wd = srcdir\n oldwd = None\n while not path.exists(wd + '/.datacats-environment'):\n oldwd = wd\n wd, _ = path.split(wd)\n if wd == oldwd:\n raise DatacatsError('Environment not found in {0} or above'.\n format(srcdir))\n srcdir = wd\n if oldwd:\n _, extension_dir = path.split(oldwd)\n return srcdir, extension_dir, None\n",
"role": "user"
},
{
"content": "find_environment_dirs",
"role": "assistant"
}
] |
def x(srcdir, datadir=None, allow_old=False):
cp = ConfigParser.SafeConfigParser()
try:
cp.read([srcdir + '/.datacats-environment'])
except ConfigParser.Error:
raise DatacatsError('Error reading environment information')
name = cp.get('datacats', 'name')
if datadir:
save_srcdir_location(datadir, srcdir)
else:
datadir = path.expanduser('~/.datacats/' + name)
if migrate.needs_format_conversion(datadir) and not allow_old:
raise DatacatsError(
'This environment uses an old format. You must migrate to the new format. To do so, use the "datacats migrate" command.'
)
if migrate.is_locked(datadir):
raise DatacatsError(
"""Migration in progress, cannot continue.
If you interrupted a migration, you should attempt manual recovery or contact us by filing an issue at http://github.com/datacats/datacats.
As a last resort, you could delete all your stored data and create a new environment by running "datacats purge" followed by "datacats init"."""
)
ckan_version = cp.get('datacats', 'ckan_version')
try:
always_prod = cp.getboolean('datacats', 'always_prod')
except ConfigParser.NoOptionError:
always_prod = False
try:
extra_containers = cp.get('datacats', 'extra_containers').split(' ')
except ConfigParser.NoOptionError:
extra_containers = ()
try:
deploy_target = cp.get('deploy', 'remote_server_user') + '@' + cp.get(
'deploy', 'remote_server')
except (ConfigParser.NoOptionError, ConfigParser.NoSectionError):
deploy_target = DEFAULT_REMOTE_SERVER_TARGET
try:
remote_server_key = cp.get('deploy', 'remote_server_key')
except (ConfigParser.NoOptionError, ConfigParser.NoSectionError):
remote_server_key = None
return (datadir, name, ckan_version, always_prod, deploy_target,
remote_server_key, extra_containers)
|
load_environment
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(srcdir, datadir=None, allow_old=False):\n cp = ConfigParser.SafeConfigParser()\n try:\n cp.read([srcdir + '/.datacats-environment'])\n except ConfigParser.Error:\n raise DatacatsError('Error reading environment information')\n name = cp.get('datacats', 'name')\n if datadir:\n save_srcdir_location(datadir, srcdir)\n else:\n datadir = path.expanduser('~/.datacats/' + name)\n if migrate.needs_format_conversion(datadir) and not allow_old:\n raise DatacatsError(\n 'This environment uses an old format. You must migrate to the new format. To do so, use the \"datacats migrate\" command.'\n )\n if migrate.is_locked(datadir):\n raise DatacatsError(\n \"\"\"Migration in progress, cannot continue.\nIf you interrupted a migration, you should attempt manual recovery or contact us by filing an issue at http://github.com/datacats/datacats.\nAs a last resort, you could delete all your stored data and create a new environment by running \"datacats purge\" followed by \"datacats init\".\"\"\"\n )\n ckan_version = cp.get('datacats', 'ckan_version')\n try:\n always_prod = cp.getboolean('datacats', 'always_prod')\n except ConfigParser.NoOptionError:\n always_prod = False\n try:\n extra_containers = cp.get('datacats', 'extra_containers').split(' ')\n except ConfigParser.NoOptionError:\n extra_containers = ()\n try:\n deploy_target = cp.get('deploy', 'remote_server_user') + '@' + cp.get(\n 'deploy', 'remote_server')\n except (ConfigParser.NoOptionError, ConfigParser.NoSectionError):\n deploy_target = DEFAULT_REMOTE_SERVER_TARGET\n try:\n remote_server_key = cp.get('deploy', 'remote_server_key')\n except (ConfigParser.NoOptionError, ConfigParser.NoSectionError):\n remote_server_key = None\n return (datadir, name, ckan_version, always_prod, deploy_target,\n remote_server_key, extra_containers)\n",
"role": "user"
},
{
"content": "load_environment",
"role": "assistant"
}
] |
def x(srcdir, datadir, site_name=None):
if site_name is None:
site_name = 'primary'
if not validate.valid_name(site_name):
raise DatacatsError('{} is not a valid site name.'.format(site_name))
cp = ConfigParser.SafeConfigParser()
try:
cp.read([srcdir + '/.datacats-environment'])
except ConfigParser.Error:
raise DatacatsError('Error reading environment information')
site_section = 'site_' + site_name
try:
port = cp.getint(site_section, 'port')
except (ConfigParser.NoOptionError, ConfigParser.NoSectionError):
port = None
try:
address = cp.get(site_section, 'address')
except (ConfigParser.NoOptionError, ConfigParser.NoSectionError):
address = None
try:
site_url = cp.get(site_section, 'site_url')
except (ConfigParser.NoOptionError, ConfigParser.NoSectionError):
site_url = None
passwords = {}
cp = ConfigParser.SafeConfigParser()
cp.read(datadir + '/sites/' + site_name + '/passwords.ini')
try:
pw_options = cp.options('passwords')
except ConfigParser.NoSectionError:
pw_options = []
for n in pw_options:
passwords[n.upper()] = cp.get('passwords', n)
return port, address, site_url, passwords
|
load_site
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(srcdir, datadir, site_name=None):\n if site_name is None:\n site_name = 'primary'\n if not validate.valid_name(site_name):\n raise DatacatsError('{} is not a valid site name.'.format(site_name))\n cp = ConfigParser.SafeConfigParser()\n try:\n cp.read([srcdir + '/.datacats-environment'])\n except ConfigParser.Error:\n raise DatacatsError('Error reading environment information')\n site_section = 'site_' + site_name\n try:\n port = cp.getint(site_section, 'port')\n except (ConfigParser.NoOptionError, ConfigParser.NoSectionError):\n port = None\n try:\n address = cp.get(site_section, 'address')\n except (ConfigParser.NoOptionError, ConfigParser.NoSectionError):\n address = None\n try:\n site_url = cp.get(site_section, 'site_url')\n except (ConfigParser.NoOptionError, ConfigParser.NoSectionError):\n site_url = None\n passwords = {}\n cp = ConfigParser.SafeConfigParser()\n cp.read(datadir + '/sites/' + site_name + '/passwords.ini')\n try:\n pw_options = cp.options('passwords')\n except ConfigParser.NoSectionError:\n pw_options = []\n for n in pw_options:\n passwords[n.upper()] = cp.get('passwords', n)\n return port, address, site_url, passwords\n",
"role": "user"
},
{
"content": "load_site",
"role": "assistant"
}
] |
def x(srcpath, site_name, ckan_version):
docker.require_images()
workdir, name = path.split(path.abspath(path.expanduser(srcpath)))
if not validate.valid_name(name):
raise DatacatsError(
'Please choose an environment name starting with a letter and including only lowercase letters and digits'
)
if not path.isdir(workdir):
raise DatacatsError('Parent directory for environment does not exist')
datadir = path.expanduser('~/.datacats/' + name)
sitedir = datadir + '/sites/' + site_name
if path.isdir(datadir):
with open(datadir + '/project-dir') as pd:
srcdir = pd.read()
else:
srcdir = workdir + '/' + name
if ckan_version not in SUPPORTED_PRELOADS:
raise DatacatsError(
"""Datacats does not currently support CKAN version {}.
Versions that are currently supported are: {}"""
.format(ckan_version, ', '.join(SUPPORTED_PRELOADS)))
preload_name = str(ckan_version)
downloaded_versions = [tag for tag in docker.get_tags('datacats/ckan')]
if ckan_version not in downloaded_versions:
retrying_pull_image('datacats/ckan:{}'.format(preload_name))
if path.isdir(sitedir):
raise DatacatsError('Site data directory {0} already exists'.format
(sitedir))
if path.isdir(srcdir) and not path.isdir(datadir):
raise DatacatsError(
"""Environment directory exists, but data directory does not.
If you simply want to recreate the data directory, run "datacats init" in the environment directory."""
)
return name, datadir, srcdir
|
new_environment_check
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(srcpath, site_name, ckan_version):\n docker.require_images()\n workdir, name = path.split(path.abspath(path.expanduser(srcpath)))\n if not validate.valid_name(name):\n raise DatacatsError(\n 'Please choose an environment name starting with a letter and including only lowercase letters and digits'\n )\n if not path.isdir(workdir):\n raise DatacatsError('Parent directory for environment does not exist')\n datadir = path.expanduser('~/.datacats/' + name)\n sitedir = datadir + '/sites/' + site_name\n if path.isdir(datadir):\n with open(datadir + '/project-dir') as pd:\n srcdir = pd.read()\n else:\n srcdir = workdir + '/' + name\n if ckan_version not in SUPPORTED_PRELOADS:\n raise DatacatsError(\n \"\"\"Datacats does not currently support CKAN version {}.\nVersions that are currently supported are: {}\"\"\"\n .format(ckan_version, ', '.join(SUPPORTED_PRELOADS)))\n preload_name = str(ckan_version)\n downloaded_versions = [tag for tag in docker.get_tags('datacats/ckan')]\n if ckan_version not in downloaded_versions:\n retrying_pull_image('datacats/ckan:{}'.format(preload_name))\n if path.isdir(sitedir):\n raise DatacatsError('Site data directory {0} already exists'.format\n (sitedir))\n if path.isdir(srcdir) and not path.isdir(datadir):\n raise DatacatsError(\n \"\"\"Environment directory exists, but data directory does not.\nIf you simply want to recreate the data directory, run \"datacats init\" in the environment directory.\"\"\"\n )\n return name, datadir, srcdir\n",
"role": "user"
},
{
"content": "new_environment_check",
"role": "assistant"
}
] |
def x(datadir, sitedir, get_container_name):
if any(not path.isdir(sitedir + x) for x in ('/files', '/run', '/solr')):
return False
if docker.is_boot2docker():
return all(docker.inspect_container(get_container_name(x)) for x in
('pgdata', 'venv'))
return path.isdir(datadir + '/venv') and path.isdir(sitedir + '/postgres')
|
data_complete
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(datadir, sitedir, get_container_name):\n if any(not path.isdir(sitedir + x) for x in ('/files', '/run', '/solr')):\n return False\n if docker.is_boot2docker():\n return all(docker.inspect_container(get_container_name(x)) for x in\n ('pgdata', 'venv'))\n return path.isdir(datadir + '/venv') and path.isdir(sitedir + '/postgres')\n",
"role": "user"
},
{
"content": "data_complete",
"role": "assistant"
}
] |
def x(datadir, sitedir, srcdir=None):
if not path.isdir(datadir):
os.makedirs(datadir, mode=448)
try:
os.makedirs(sitedir, mode=448)
except OSError:
raise DatacatsError('Site already exists.')
if not docker.is_boot2docker():
if not path.isdir(datadir + '/venv'):
os.makedirs(datadir + '/venv')
os.makedirs(sitedir + '/postgres')
os.makedirs(sitedir + '/solr')
os.makedirs(sitedir + '/files')
os.makedirs(sitedir + '/run')
if srcdir:
os.makedirs(srcdir)
|
create_directories
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(datadir, sitedir, srcdir=None):\n if not path.isdir(datadir):\n os.makedirs(datadir, mode=448)\n try:\n os.makedirs(sitedir, mode=448)\n except OSError:\n raise DatacatsError('Site already exists.')\n if not docker.is_boot2docker():\n if not path.isdir(datadir + '/venv'):\n os.makedirs(datadir + '/venv')\n os.makedirs(sitedir + '/postgres')\n os.makedirs(sitedir + '/solr')\n os.makedirs(sitedir + '/files')\n os.makedirs(sitedir + '/run')\n if srcdir:\n os.makedirs(srcdir)\n",
"role": "user"
},
{
"content": "create_directories",
"role": "assistant"
}
] |
def x(srcdir, datadir, preload_image, get_container_name):
try:
if docker.is_boot2docker():
docker.data_only_container(get_container_name('venv'), [
'/usr/lib/ckan'])
img_id = docker.web_command(
'/bin/mv /usr/lib/ckan/ /usr/lib/ckan_original', image=
preload_image, commit=True)
docker.web_command(command=
'/bin/cp -a /usr/lib/ckan_original/. /usr/lib/ckan/.',
volumes_from=get_container_name('venv'), image=img_id)
docker.remove_image(img_id)
return
docker.web_command(command=
'/bin/cp -a /usr/lib/ckan/. /usr/lib/ckan_target/.', rw={(
datadir + '/venv'): '/usr/lib/ckan_target'}, image=preload_image)
finally:
rw = {(datadir + '/venv'): '/usr/lib/ckan'
} if not docker.is_boot2docker() else {}
volumes_from = get_container_name('venv') if docker.is_boot2docker(
) else None
docker.web_command(command=
'/bin/chown -R --reference=/project /usr/lib/ckan', rw=rw,
volumes_from=volumes_from, ro={srcdir: '/project'})
|
create_virtualenv
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(srcdir, datadir, preload_image, get_container_name):\n try:\n if docker.is_boot2docker():\n docker.data_only_container(get_container_name('venv'), [\n '/usr/lib/ckan'])\n img_id = docker.web_command(\n '/bin/mv /usr/lib/ckan/ /usr/lib/ckan_original', image=\n preload_image, commit=True)\n docker.web_command(command=\n '/bin/cp -a /usr/lib/ckan_original/. /usr/lib/ckan/.',\n volumes_from=get_container_name('venv'), image=img_id)\n docker.remove_image(img_id)\n return\n docker.web_command(command=\n '/bin/cp -a /usr/lib/ckan/. /usr/lib/ckan_target/.', rw={(\n datadir + '/venv'): '/usr/lib/ckan_target'}, image=preload_image)\n finally:\n rw = {(datadir + '/venv'): '/usr/lib/ckan'\n } if not docker.is_boot2docker() else {}\n volumes_from = get_container_name('venv') if docker.is_boot2docker(\n ) else None\n docker.web_command(command=\n '/bin/chown -R --reference=/project /usr/lib/ckan', rw=rw,\n volumes_from=volumes_from, ro={srcdir: '/project'})\n",
"role": "user"
},
{
"content": "create_virtualenv",
"role": "assistant"
}
] |
def x(srcdir, preload_image, datapusher=False):
try:
docker.web_command(command=
'/bin/cp -a /project/ckan /project_target/ckan', rw={srcdir:
'/project_target'}, image=preload_image)
if datapusher:
docker.web_command(command=
'/bin/cp -a /project/datapusher /project_target/datapusher',
rw={srcdir: '/project_target'}, image=preload_image)
shutil.copy(srcdir + '/ckan/ckan/config/who.ini', srcdir)
shutil.copy(srcdir + '/ckan/ckan/config/solr/schema.xml', srcdir)
finally:
docker.web_command(command=
'/bin/chown -R --reference=/project /project', rw={srcdir:
'/project'})
|
create_source
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(srcdir, preload_image, datapusher=False):\n try:\n docker.web_command(command=\n '/bin/cp -a /project/ckan /project_target/ckan', rw={srcdir:\n '/project_target'}, image=preload_image)\n if datapusher:\n docker.web_command(command=\n '/bin/cp -a /project/datapusher /project_target/datapusher',\n rw={srcdir: '/project_target'}, image=preload_image)\n shutil.copy(srcdir + '/ckan/ckan/config/who.ini', srcdir)\n shutil.copy(srcdir + '/ckan/ckan/config/solr/schema.xml', srcdir)\n finally:\n docker.web_command(command=\n '/bin/chown -R --reference=/project /project', rw={srcdir:\n '/project'})\n",
"role": "user"
},
{
"content": "create_source",
"role": "assistant"
}
] |
def x(sitedir, srcdir, passwords, get_container_name, extra_containers,
log_syslog=False):
if docker.is_boot2docker():
docker.data_only_container(get_container_name('pgdata'), [
'/var/lib/postgresql/data'])
rw = {}
volumes_from = get_container_name('pgdata')
else:
rw = {(sitedir + '/postgres'): '/var/lib/postgresql/data'}
volumes_from = None
running = set(containers_running(get_container_name))
needed = set(extra_containers).union({'postgres', 'solr'})
if not needed.issubset(running):
stop_supporting_containers(get_container_name, extra_containers)
docker.run_container(name=get_container_name('postgres'), image=
'datacats/postgres', environment=passwords, rw=rw, volumes_from
=volumes_from, log_syslog=log_syslog)
docker.run_container(name=get_container_name('solr'), image=
'datacats/solr', rw={(sitedir + '/solr'): '/var/lib/solr'}, ro=
{(srcdir + '/schema.xml'): '/etc/solr/conf/schema.xml'},
log_syslog=log_syslog)
for container in extra_containers:
docker.run_container(name=get_container_name(container), image=
EXTRA_IMAGE_MAPPING[container], ro={sitedir: '/datadir',
srcdir: '/project'}, log_syslog=log_syslog)
|
start_supporting_containers
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sitedir, srcdir, passwords, get_container_name, extra_containers,\n log_syslog=False):\n if docker.is_boot2docker():\n docker.data_only_container(get_container_name('pgdata'), [\n '/var/lib/postgresql/data'])\n rw = {}\n volumes_from = get_container_name('pgdata')\n else:\n rw = {(sitedir + '/postgres'): '/var/lib/postgresql/data'}\n volumes_from = None\n running = set(containers_running(get_container_name))\n needed = set(extra_containers).union({'postgres', 'solr'})\n if not needed.issubset(running):\n stop_supporting_containers(get_container_name, extra_containers)\n docker.run_container(name=get_container_name('postgres'), image=\n 'datacats/postgres', environment=passwords, rw=rw, volumes_from\n =volumes_from, log_syslog=log_syslog)\n docker.run_container(name=get_container_name('solr'), image=\n 'datacats/solr', rw={(sitedir + '/solr'): '/var/lib/solr'}, ro=\n {(srcdir + '/schema.xml'): '/etc/solr/conf/schema.xml'},\n log_syslog=log_syslog)\n for container in extra_containers:\n docker.run_container(name=get_container_name(container), image=\n EXTRA_IMAGE_MAPPING[container], ro={sitedir: '/datadir',\n srcdir: '/project'}, log_syslog=log_syslog)\n",
"role": "user"
},
{
"content": "start_supporting_containers",
"role": "assistant"
}
] |
def x(get_container_name, extra_containers):
docker.remove_container(get_container_name('postgres'))
docker.remove_container(get_container_name('solr'))
for container in extra_containers:
docker.remove_container(get_container_name(container))
|
stop_supporting_containers
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(get_container_name, extra_containers):\n docker.remove_container(get_container_name('postgres'))\n docker.remove_container(get_container_name('solr'))\n for container in extra_containers:\n docker.remove_container(get_container_name(container))\n",
"role": "user"
},
{
"content": "stop_supporting_containers",
"role": "assistant"
}
] |
def x(get_container_name):
running = []
for n in ['web', 'postgres', 'solr', 'datapusher', 'redis']:
info = docker.inspect_container(get_container_name(n))
if info and not info['State']['Running']:
running.append(n + '(halted)')
elif info:
running.append(n)
return running
|
containers_running
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(get_container_name):\n running = []\n for n in ['web', 'postgres', 'solr', 'datapusher', 'redis']:\n info = docker.inspect_container(get_container_name(n))\n if info and not info['State']['Running']:\n running.append(n + '(halted)')\n elif info:\n running.append(n)\n return running\n",
"role": "user"
},
{
"content": "containers_running",
"role": "assistant"
}
] |
def x(self):
if not self.sites:
self.sites = task.list_sites(self.datadir)
return self.sites
|
load_sites
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if not self.sites:\n self.sites = task.list_sites(self.datadir)\n return self.sites\n",
"role": "user"
},
{
"content": "load_sites",
"role": "assistant"
}
] |
def x(self, create=True):
self._load_sites()
if create:
self.sites.append(self.site_name)
task.save_new_site(self.site_name, self.sitedir, self.target, self.port,
self.address, self.site_url, self.passwords)
|
save_site
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, create=True):\n self._load_sites()\n if create:\n self.sites.append(self.site_name)\n task.save_new_site(self.site_name, self.sitedir, self.target, self.port,\n self.address, self.site_url, self.passwords)\n",
"role": "user"
},
{
"content": "save_site",
"role": "assistant"
}
] |
def x(self):
task.save_new_environment(self.name, self.datadir, self.target, self.
ckan_version, self.deploy_target, self.always_prod)
|
save
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n task.save_new_environment(self.name, self.datadir, self.target, self.\n ckan_version, self.deploy_target, self.always_prod)\n",
"role": "user"
},
{
"content": "save",
"role": "assistant"
}
] |
def x(cls, path, ckan_version, site_name, **kwargs):
if ckan_version == 'master':
ckan_version = 'latest'
name, datadir, srcdir = task.new_environment_check(path, site_name,
ckan_version)
environment = cls(name, srcdir, datadir, site_name, ckan_version, **kwargs)
environment._generate_passwords()
return environment
|
new
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, path, ckan_version, site_name, **kwargs):\n if ckan_version == 'master':\n ckan_version = 'latest'\n name, datadir, srcdir = task.new_environment_check(path, site_name,\n ckan_version)\n environment = cls(name, srcdir, datadir, site_name, ckan_version, **kwargs)\n environment._generate_passwords()\n return environment\n",
"role": "user"
},
{
"content": "new",
"role": "assistant"
}
] |
def x(cls, environment_name=None, site_name='primary', data_only=False,
allow_old=False):
srcdir, extension_dir, datadir = task.find_environment_dirs(
environment_name, data_only)
if datadir and data_only:
return cls(environment_name, None, datadir, site_name)
(datadir, name, ckan_version, always_prod, deploy_target,
remote_server_key, extra_containers) = task.load_environment(srcdir,
datadir, allow_old)
if not allow_old:
port, address, site_url, passwords = task.load_site(srcdir, datadir,
site_name)
else:
port, address, site_url, passwords = None, None, None, None
environment = cls(name, srcdir, datadir, site_name, ckan_version=
ckan_version, port=port, deploy_target=deploy_target, site_url=
site_url, always_prod=always_prod, address=address, extension_dir=
extension_dir, remote_server_key=remote_server_key,
extra_containers=extra_containers)
if passwords:
environment.passwords = passwords
else:
environment._generate_passwords()
if not allow_old:
environment._load_sites()
return environment
|
load
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, environment_name=None, site_name='primary', data_only=False,\n allow_old=False):\n srcdir, extension_dir, datadir = task.find_environment_dirs(\n environment_name, data_only)\n if datadir and data_only:\n return cls(environment_name, None, datadir, site_name)\n (datadir, name, ckan_version, always_prod, deploy_target,\n remote_server_key, extra_containers) = task.load_environment(srcdir,\n datadir, allow_old)\n if not allow_old:\n port, address, site_url, passwords = task.load_site(srcdir, datadir,\n site_name)\n else:\n port, address, site_url, passwords = None, None, None, None\n environment = cls(name, srcdir, datadir, site_name, ckan_version=\n ckan_version, port=port, deploy_target=deploy_target, site_url=\n site_url, always_prod=always_prod, address=address, extension_dir=\n extension_dir, remote_server_key=remote_server_key,\n extra_containers=extra_containers)\n if passwords:\n environment.passwords = passwords\n else:\n environment._generate_passwords()\n if not allow_old:\n environment._load_sites()\n return environment\n",
"role": "user"
},
{
"content": "load",
"role": "assistant"
}
] |
def x(self):
return task.data_complete(self.datadir, self.sitedir, self.
_get_container_name)
|
data_complete
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return task.data_complete(self.datadir, self.sitedir, self.\n _get_container_name)\n",
"role": "user"
},
{
"content": "data_complete",
"role": "assistant"
}
] |
def x(self):
files = task.source_missing(self.target)
if files:
raise DatacatsError('Missing files in source directory:\n' + '\n'.
join(files))
if not self.data_exists():
raise DatacatsError('Environment datadir missing. Try "datacats init".'
)
if not self.data_complete():
raise DatacatsError(
'Environment datadir damaged or volumes missing. To reset and discard all data use "datacats reset"'
)
|
require_data
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n files = task.source_missing(self.target)\n if files:\n raise DatacatsError('Missing files in source directory:\\n' + '\\n'.\n join(files))\n if not self.data_exists():\n raise DatacatsError('Environment datadir missing. Try \"datacats init\".'\n )\n if not self.data_complete():\n raise DatacatsError(\n 'Environment datadir damaged or volumes missing. To reset and discard all data use \"datacats reset\"'\n )\n",
"role": "user"
},
{
"content": "require_data",
"role": "assistant"
}
] |
def x(self, create_project_dir=True):
return task.create_directories(self.datadir, self.sitedir, self.target if
create_project_dir else None)
|
create_directories
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, create_project_dir=True):\n return task.create_directories(self.datadir, self.sitedir, self.target if\n create_project_dir else None)\n",
"role": "user"
},
{
"content": "create_directories",
"role": "assistant"
}
] |
def x(self):
return task.create_virtualenv(self.target, self.datadir, self.
_preload_image(), self._get_container_name)
|
create_virtualenv
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return task.create_virtualenv(self.target, self.datadir, self.\n _preload_image(), self._get_container_name)\n",
"role": "user"
},
{
"content": "create_virtualenv",
"role": "assistant"
}
] |
def x(self):
self.user_run_script(script=scripts.get_script_path(
'clean_virtualenv.sh'), args=[], rw_venv=True)
|
clean_virtualenv
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.user_run_script(script=scripts.get_script_path(\n 'clean_virtualenv.sh'), args=[], rw_venv=True)\n",
"role": "user"
},
{
"content": "clean_virtualenv",
"role": "assistant"
}
] |
def x(self, datapusher=True):
task.create_source(self.target, self._preload_image(), datapusher)
|
create_source
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, datapusher=True):\n task.create_source(self.target, self._preload_image(), datapusher)\n",
"role": "user"
},
{
"content": "create_source",
"role": "assistant"
}
] |
def x(self, log_syslog=False):
log_syslog = True if self.always_prod else log_syslog
task.start_supporting_containers(self.sitedir, self.target, self.
passwords, self._get_container_name, self.extra_containers,
log_syslog=log_syslog)
|
start_supporting_containers
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, log_syslog=False):\n log_syslog = True if self.always_prod else log_syslog\n task.start_supporting_containers(self.sitedir, self.target, self.\n passwords, self._get_container_name, self.extra_containers,\n log_syslog=log_syslog)\n",
"role": "user"
},
{
"content": "start_supporting_containers",
"role": "assistant"
}
] |
def x(self):
self.run_command(command=
'/scripts/run_as_user.sh /usr/lib/ckan/bin/paster make-config ckan /project/development.ini'
, rw_project=True, ro={scripts.get_script_path('run_as_user.sh'):
'/scripts/run_as_user.sh'})
|
create_ckan_ini
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.run_command(command=\n '/scripts/run_as_user.sh /usr/lib/ckan/bin/paster make-config ckan /project/development.ini'\n , rw_project=True, ro={scripts.get_script_path('run_as_user.sh'):\n '/scripts/run_as_user.sh'})\n",
"role": "user"
},
{
"content": "create_ckan_ini",
"role": "assistant"
}
] |
def x(self, skin=True):
command = ['/usr/lib/ckan/bin/paster', '--plugin=ckan', 'config-tool',
'/project/development.ini', '-e',
'sqlalchemy.url = postgresql://<hidden>',
'ckan.datastore.read_url = postgresql://<hidden>',
'ckan.datastore.write_url = postgresql://<hidden>',
'ckan.datapusher.url = http://datapusher:8800',
'solr_url = http://solr:8080/solr',
'ckan.storage_path = /var/www/storage',
'ckan.plugins = datastore resource_proxy text_view ' + (
'datapusher ' if exists(self.target + '/datapusher') else '') +
'recline_grid_view recline_graph_view' + (' {0}_theme'.format(self.
name) if skin else ''), 'ckan.site_title = ' + self.name,
'ckan.site_logo =', 'ckan.auth.create_user_via_web = false']
self.run_command(command=command, rw_project=True)
|
update_ckan_ini
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, skin=True):\n command = ['/usr/lib/ckan/bin/paster', '--plugin=ckan', 'config-tool',\n '/project/development.ini', '-e',\n 'sqlalchemy.url = postgresql://<hidden>',\n 'ckan.datastore.read_url = postgresql://<hidden>',\n 'ckan.datastore.write_url = postgresql://<hidden>',\n 'ckan.datapusher.url = http://datapusher:8800',\n 'solr_url = http://solr:8080/solr',\n 'ckan.storage_path = /var/www/storage', \n 'ckan.plugins = datastore resource_proxy text_view ' + (\n 'datapusher ' if exists(self.target + '/datapusher') else '') +\n 'recline_grid_view recline_graph_view' + (' {0}_theme'.format(self.\n name) if skin else ''), 'ckan.site_title = ' + self.name,\n 'ckan.site_logo =', 'ckan.auth.create_user_via_web = false']\n self.run_command(command=command, rw_project=True)\n",
"role": "user"
},
{
"content": "update_ckan_ini",
"role": "assistant"
}
] |
def x(self):
ckan_extension_template(self.name, self.target)
self.install_package_develop('ckanext-' + self.name + 'theme')
|
create_install_template_skin
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n ckan_extension_template(self.name, self.target)\n self.install_package_develop('ckanext-' + self.name + 'theme')\n",
"role": "user"
},
{
"content": "create_install_template_skin",
"role": "assistant"
}
] |
def x(self, retry_seconds=DB_INIT_RETRY_SECONDS):
started = time.time()
while True:
try:
self.run_command(
'/usr/lib/ckan/bin/paster --plugin=ckan db init -c /project/development.ini'
, db_links=True, clean_up=True)
break
except WebCommandError:
if started + retry_seconds > time.time():
raise
time.sleep(DB_INIT_RETRY_DELAY)
|
ckan_db_init
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, retry_seconds=DB_INIT_RETRY_SECONDS):\n started = time.time()\n while True:\n try:\n self.run_command(\n '/usr/lib/ckan/bin/paster --plugin=ckan db init -c /project/development.ini'\n , db_links=True, clean_up=True)\n break\n except WebCommandError:\n if started + retry_seconds > time.time():\n raise\n time.sleep(DB_INIT_RETRY_DELAY)\n",
"role": "user"
},
{
"content": "ckan_db_init",
"role": "assistant"
}
] |
def x(self, production=False, log_syslog=False, paster_reload=True,
interactive=False):
self.stop_ckan()
address = self.address or '127.0.0.1'
port = self.port
log_syslog = True if self.always_prod else log_syslog
production = production or self.always_prod
override_site_url = self.address is None and not is_boot2docker(
) and not self.site_url
command = ['/scripts/web.sh', str(production), str(override_site_url),
str(paster_reload)]
if not production:
css = self.target + '/ckan/ckan/public/base/css'
if not exists(css + '/main.debug.css'):
from shutil import copyfile
copyfile(css + '/main.css', css + '/main.debug.css')
ro = {self.target: '/project', scripts.get_script_path('datapusher.sh'):
'/scripts/datapusher.sh'}
if not is_boot2docker():
ro[self.datadir + '/venv'] = '/usr/lib/ckan'
datapusher = self.needs_datapusher()
if datapusher:
run_container(self._get_container_name('datapusher'),
'datacats/web', '/scripts/datapusher.sh', ro=ro, volumes_from=
self._get_container_name('venv') if is_boot2docker() else None,
log_syslog=log_syslog)
while True:
self._create_run_ini(port, production)
try:
self._run_web_container(port, command, address, log_syslog=
log_syslog, datapusher=datapusher, interactive=interactive)
if not is_boot2docker():
self.address = address
except PortAllocatedError:
port = self._next_port(port)
continue
break
|
start_ckan
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, production=False, log_syslog=False, paster_reload=True,\n interactive=False):\n self.stop_ckan()\n address = self.address or '127.0.0.1'\n port = self.port\n log_syslog = True if self.always_prod else log_syslog\n production = production or self.always_prod\n override_site_url = self.address is None and not is_boot2docker(\n ) and not self.site_url\n command = ['/scripts/web.sh', str(production), str(override_site_url),\n str(paster_reload)]\n if not production:\n css = self.target + '/ckan/ckan/public/base/css'\n if not exists(css + '/main.debug.css'):\n from shutil import copyfile\n copyfile(css + '/main.css', css + '/main.debug.css')\n ro = {self.target: '/project', scripts.get_script_path('datapusher.sh'):\n '/scripts/datapusher.sh'}\n if not is_boot2docker():\n ro[self.datadir + '/venv'] = '/usr/lib/ckan'\n datapusher = self.needs_datapusher()\n if datapusher:\n run_container(self._get_container_name('datapusher'),\n 'datacats/web', '/scripts/datapusher.sh', ro=ro, volumes_from=\n self._get_container_name('venv') if is_boot2docker() else None,\n log_syslog=log_syslog)\n while True:\n self._create_run_ini(port, production)\n try:\n self._run_web_container(port, command, address, log_syslog=\n log_syslog, datapusher=datapusher, interactive=interactive)\n if not is_boot2docker():\n self.address = address\n except PortAllocatedError:\n port = self._next_port(port)\n continue\n break\n",
"role": "user"
},
{
"content": "start_ckan",
"role": "assistant"
}
] |
def x(self, port, production, output='development.ini', source=
'development.ini', override_site_url=True):
cp = SafeConfigParser()
try:
cp.read([self.target + '/' + source])
except ConfigParserError:
raise DatacatsError('Error reading development.ini')
cp.set('DEFAULT', 'debug', 'false' if production else 'true')
if self.site_url:
site_url = self.site_url
else:
if is_boot2docker():
web_address = socket.gethostbyname(docker_host())
else:
web_address = self.address
site_url = 'http://{}:{}'.format(web_address, port)
if override_site_url:
cp.set('app:main', 'ckan.site_url', site_url)
cp.set('app:main', 'sqlalchemy.url',
'postgresql://ckan:{0}@db:5432/ckan'.format(self.passwords[
'CKAN_PASSWORD']))
cp.set('app:main', 'ckan.datastore.read_url',
'postgresql://ckan_datastore_readonly:{0}@db:5432/ckan_datastore'.
format(self.passwords['DATASTORE_RO_PASSWORD']))
cp.set('app:main', 'ckan.datastore.write_url',
'postgresql://ckan_datastore_readwrite:{0}@db:5432/ckan_datastore'.
format(self.passwords['DATASTORE_RW_PASSWORD']))
cp.set('app:main', 'solr_url', 'http://solr:8080/solr')
cp.set('app:main', 'ckan.redis.url', 'http://redis:6379')
cp.set('app:main', 'beaker.session.secret', self.passwords[
'BEAKER_SESSION_SECRET'])
if not isdir(self.sitedir + '/run'):
makedirs(self.sitedir + '/run')
with open(self.sitedir + '/run/' + output, 'w') as runini:
cp.write(runini)
|
create_run_ini
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, port, production, output='development.ini', source=\n 'development.ini', override_site_url=True):\n cp = SafeConfigParser()\n try:\n cp.read([self.target + '/' + source])\n except ConfigParserError:\n raise DatacatsError('Error reading development.ini')\n cp.set('DEFAULT', 'debug', 'false' if production else 'true')\n if self.site_url:\n site_url = self.site_url\n else:\n if is_boot2docker():\n web_address = socket.gethostbyname(docker_host())\n else:\n web_address = self.address\n site_url = 'http://{}:{}'.format(web_address, port)\n if override_site_url:\n cp.set('app:main', 'ckan.site_url', site_url)\n cp.set('app:main', 'sqlalchemy.url',\n 'postgresql://ckan:{0}@db:5432/ckan'.format(self.passwords[\n 'CKAN_PASSWORD']))\n cp.set('app:main', 'ckan.datastore.read_url',\n 'postgresql://ckan_datastore_readonly:{0}@db:5432/ckan_datastore'.\n format(self.passwords['DATASTORE_RO_PASSWORD']))\n cp.set('app:main', 'ckan.datastore.write_url',\n 'postgresql://ckan_datastore_readwrite:{0}@db:5432/ckan_datastore'.\n format(self.passwords['DATASTORE_RW_PASSWORD']))\n cp.set('app:main', 'solr_url', 'http://solr:8080/solr')\n cp.set('app:main', 'ckan.redis.url', 'http://redis:6379')\n cp.set('app:main', 'beaker.session.secret', self.passwords[\n 'BEAKER_SESSION_SECRET'])\n if not isdir(self.sitedir + '/run'):\n makedirs(self.sitedir + '/run')\n with open(self.sitedir + '/run/' + output, 'w') as runini:\n cp.write(runini)\n",
"role": "user"
},
{
"content": "create_run_ini",
"role": "assistant"
}
] |
def x(self, port, command, address, log_syslog=False, datapusher=True,
interactive=False):
if is_boot2docker():
ro = {}
volumes_from = self._get_container_name('venv')
else:
ro = {(self.datadir + '/venv'): '/usr/lib/ckan'}
volumes_from = None
links = {self._get_container_name('solr'): 'solr', self.
_get_container_name('postgres'): 'db'}
links.update({self._get_container_name(container): container for
container in self.extra_containers})
if datapusher:
if 'datapusher' not in self.containers_running():
raise DatacatsError(container_logs(self._get_container_name(
'datapusher'), 'all', False, False))
links[self._get_container_name('datapusher')] = 'datapusher'
ro = dict({self.target: '/project/', scripts.get_script_path('web.sh'):
'/scripts/web.sh', scripts.get_script_path('adjust_devini.py'):
'/scripts/adjust_devini.py'}, **ro)
rw = {(self.sitedir + '/files'): '/var/www/storage', (self.sitedir +
'/run/development.ini'): '/project/development.ini'}
try:
if not interactive:
run_container(name=self._get_container_name('web'), image=
'datacats/web', rw=rw, ro=ro, links=links, volumes_from=
volumes_from, command=command, port_bindings={(5000): port if
is_boot2docker() else (address, port)}, log_syslog=log_syslog)
else:
if is_boot2docker():
switches = ['--volumes-from', self._get_container_name(
'pgdata'), '--volumes-from', self._get_container_name(
'venv')]
else:
switches = []
switches += ['--volume={}:{}:ro'.format(vol, ro[vol]) for vol in ro
]
switches += ['--volume={}:{}'.format(vol, rw[vol]) for vol in rw]
links = ['--link={}:{}'.format(link, links[link]) for link in links
]
args = ['docker', 'run', '-it', '--name', self.
_get_container_name('web'), '-p', '{}:5000'.format(port) if
is_boot2docker() else '{}:{}:5000'.format(address, port)
] + switches + links + ['datacats/web'] + command
subprocess.call(args)
except APIError as e:
if '409' in str(e):
raise DatacatsError(
'Web container already running. Please stop_web before running.'
)
else:
raise
|
run_web_container
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, port, command, address, log_syslog=False, datapusher=True,\n interactive=False):\n if is_boot2docker():\n ro = {}\n volumes_from = self._get_container_name('venv')\n else:\n ro = {(self.datadir + '/venv'): '/usr/lib/ckan'}\n volumes_from = None\n links = {self._get_container_name('solr'): 'solr', self.\n _get_container_name('postgres'): 'db'}\n links.update({self._get_container_name(container): container for\n container in self.extra_containers})\n if datapusher:\n if 'datapusher' not in self.containers_running():\n raise DatacatsError(container_logs(self._get_container_name(\n 'datapusher'), 'all', False, False))\n links[self._get_container_name('datapusher')] = 'datapusher'\n ro = dict({self.target: '/project/', scripts.get_script_path('web.sh'):\n '/scripts/web.sh', scripts.get_script_path('adjust_devini.py'):\n '/scripts/adjust_devini.py'}, **ro)\n rw = {(self.sitedir + '/files'): '/var/www/storage', (self.sitedir +\n '/run/development.ini'): '/project/development.ini'}\n try:\n if not interactive:\n run_container(name=self._get_container_name('web'), image=\n 'datacats/web', rw=rw, ro=ro, links=links, volumes_from=\n volumes_from, command=command, port_bindings={(5000): port if\n is_boot2docker() else (address, port)}, log_syslog=log_syslog)\n else:\n if is_boot2docker():\n switches = ['--volumes-from', self._get_container_name(\n 'pgdata'), '--volumes-from', self._get_container_name(\n 'venv')]\n else:\n switches = []\n switches += ['--volume={}:{}:ro'.format(vol, ro[vol]) for vol in ro\n ]\n switches += ['--volume={}:{}'.format(vol, rw[vol]) for vol in rw]\n links = ['--link={}:{}'.format(link, links[link]) for link in links\n ]\n args = ['docker', 'run', '-it', '--name', self.\n _get_container_name('web'), '-p', '{}:5000'.format(port) if\n is_boot2docker() else '{}:{}:5000'.format(address, port)\n ] + switches + links + ['datacats/web'] + command\n subprocess.call(args)\n except APIError as e:\n if '409' in str(e):\n raise DatacatsError(\n 'Web container already running. Please stop_web before running.'\n )\n else:\n raise\n",
"role": "user"
},
{
"content": "run_web_container",
"role": "assistant"
}
] |
def x(self):
try:
if not wait_for_service_available(self._get_container_name('web'),
self.web_address(), WEB_START_TIMEOUT_SECONDS):
raise DatacatsError('Error while starting web container:\n' +
container_logs(self._get_container_name('web'), 'all',
False, None))
except ServiceTimeout:
raise DatacatsError('Timeout while starting web container. Logs:' +
container_logs(self._get_container_name('web'), 'all', False, None)
)
|
wait_for_web_available
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n try:\n if not wait_for_service_available(self._get_container_name('web'),\n self.web_address(), WEB_START_TIMEOUT_SECONDS):\n raise DatacatsError('Error while starting web container:\\n' +\n container_logs(self._get_container_name('web'), 'all', \n False, None))\n except ServiceTimeout:\n raise DatacatsError('Timeout while starting web container. Logs:' +\n container_logs(self._get_container_name('web'), 'all', False, None)\n )\n",
"role": "user"
},
{
"content": "wait_for_web_available",
"role": "assistant"
}
] |
def x(self):
return 5000 + unpack('Q', sha((self.name + self.site_name).decode(
'ascii')).digest()[:8])[0] % 1000
|
choose_port
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return 5000 + unpack('Q', sha((self.name + self.site_name).decode(\n 'ascii')).digest()[:8])[0] % 1000\n",
"role": "user"
},
{
"content": "choose_port",
"role": "assistant"
}
] |
def x(self, port):
port = 5000 + (port + 1) % 1000
if port == self.port:
raise DatacatsError('Too many instances running')
return port
|
next_port
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, port):\n port = 5000 + (port + 1) % 1000\n if port == self.port:\n raise DatacatsError('Too many instances running')\n return port\n",
"role": "user"
},
{
"content": "next_port",
"role": "assistant"
}
] |
def x(self):
remove_container(self._get_container_name('web'), force=True)
remove_container(self._get_container_name('datapusher'), force=True)
|
stop_ckan
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n remove_container(self._get_container_name('web'), force=True)\n remove_container(self._get_container_name('datapusher'), force=True)\n",
"role": "user"
},
{
"content": "stop_ckan",
"role": "assistant"
}
] |
def x(self):
info = inspect_container(self._get_container_name('web'))
if info is None:
return None
try:
if not info['State']['Running']:
return None
return info['NetworkSettings']['Ports']['5000/tcp'][0]['HostPort']
except TypeError:
return None
|
current_web_port
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n info = inspect_container(self._get_container_name('web'))\n if info is None:\n return None\n try:\n if not info['State']['Running']:\n return None\n return info['NetworkSettings']['Ports']['5000/tcp'][0]['HostPort']\n except TypeError:\n return None\n",
"role": "user"
},
{
"content": "current_web_port",
"role": "assistant"
}
] |
def x(self, container, error_on_exists=False):
if container in self.extra_containers:
if error_on_exists:
raise DatacatsError('{} is already added as an extra container.'
.format(container))
else:
return
self.extra_containers.append(container)
cp = SafeConfigParser()
cp.read(self.target + '/.datacats-environment')
cp.set('datacats', 'extra_containers', ' '.join(self.extra_containers))
with open(self.target + '/.datacats-environment', 'w') as f:
cp.write(f)
|
add_extra_container
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, container, error_on_exists=False):\n if container in self.extra_containers:\n if error_on_exists:\n raise DatacatsError('{} is already added as an extra container.'\n .format(container))\n else:\n return\n self.extra_containers.append(container)\n cp = SafeConfigParser()\n cp.read(self.target + '/.datacats-environment')\n cp.set('datacats', 'extra_containers', ' '.join(self.extra_containers))\n with open(self.target + '/.datacats-environment', 'w') as f:\n cp.write(f)\n",
"role": "user"
},
{
"content": "add_extra_container",
"role": "assistant"
}
] |
def x(self):
port = self._current_web_port()
address = self.address or '127.0.0.1'
if port is None:
return None
return 'http://{0}:{1}/'.format(address if address and not
is_boot2docker() else docker_host(), port)
|
web_address
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n port = self._current_web_port()\n address = self.address or '127.0.0.1'\n if port is None:\n return None\n return 'http://{0}:{1}/'.format(address if address and not\n is_boot2docker() else docker_host(), port)\n",
"role": "user"
},
{
"content": "web_address",
"role": "assistant"
}
] |
def x(self, password):
with open(self.sitedir + '/run/admin.json', 'w') as out:
json.dump({'name': 'admin', 'email': 'none', 'password': password,
'sysadmin': True}, out)
self.user_run_script(script=scripts.get_script_path(
'update_add_admin.sh'), args=[], db_links=True, ro={(self.sitedir +
'/run/admin.json'): '/input/admin.json'})
remove(self.sitedir + '/run/admin.json')
|
create_admin_set_password
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, password):\n with open(self.sitedir + '/run/admin.json', 'w') as out:\n json.dump({'name': 'admin', 'email': 'none', 'password': password,\n 'sysadmin': True}, out)\n self.user_run_script(script=scripts.get_script_path(\n 'update_add_admin.sh'), args=[], db_links=True, ro={(self.sitedir +\n '/run/admin.json'): '/input/admin.json'})\n remove(self.sitedir + '/run/admin.json')\n",
"role": "user"
},
{
"content": "create_admin_set_password",
"role": "assistant"
}
] |
def x(self, command=None, paster=False, detach=False):
if not exists(self.target + '/.bash_profile'):
self.create_bash_profile()
if not command:
command = []
use_tty = sys.stdin.isatty() and sys.stdout.isatty()
background = environ.get('CIRCLECI', False) or detach
if is_boot2docker():
venv_volumes = ['--volumes-from', self._get_container_name('venv')]
else:
venv_volumes = ['-v', self.datadir + '/venv:/usr/lib/ckan:rw']
self._create_run_ini(self.port, production=False, output='run.ini')
self._create_run_ini(self.port, production=True, output='test.ini',
source='ckan/test-core.ini', override_site_url=False)
script = scripts.get_script_path('shell.sh')
if paster:
script = scripts.get_script_path('paster.sh')
if command and command != ['help'] and command != ['--help']:
command += ['--config=/project/development.ini']
command = [self.extension_dir] + command
proxy_settings = self._proxy_settings()
if proxy_settings:
venv_volumes += ['-v', self.sitedir +
'/run/proxy-environment:/etc/environment:ro']
links = {self._get_container_name('solr'): 'solr', self.
_get_container_name('postgres'): 'db'}
links.update({self._get_container_name(container): container for
container in self.extra_containers})
link_params = []
for link in links:
link_params.append('--link')
link_params.append(link + ':' + links[link])
if 'datapusher' in self.containers_running():
link_params.append('--link')
link_params.append(self._get_container_name('datapusher') +
':datapusher')
return subprocess.call([DOCKER_EXE, 'run'] + (['--rm'] if not
background else []) + ['-t' if use_tty else '', '-d' if detach else
'-i'] + venv_volumes + ['-v', self.target + ':/project:rw', '-v',
self.sitedir + '/files:/var/www/storage:rw', '-v', script +
':/scripts/shell.sh:ro', '-v', scripts.get_script_path(
'paster_cd.sh') + ':/scripts/paster_cd.sh:ro', '-v', self.sitedir +
'/run/run.ini:/project/development.ini:ro', '-v', self.sitedir +
'/run/test.ini:/project/ckan/test-core.ini:ro'] + link_params + [
'--hostname', self.name, 'datacats/web', '/scripts/shell.sh'] + command
)
|
interactive_shell
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, command=None, paster=False, detach=False):\n if not exists(self.target + '/.bash_profile'):\n self.create_bash_profile()\n if not command:\n command = []\n use_tty = sys.stdin.isatty() and sys.stdout.isatty()\n background = environ.get('CIRCLECI', False) or detach\n if is_boot2docker():\n venv_volumes = ['--volumes-from', self._get_container_name('venv')]\n else:\n venv_volumes = ['-v', self.datadir + '/venv:/usr/lib/ckan:rw']\n self._create_run_ini(self.port, production=False, output='run.ini')\n self._create_run_ini(self.port, production=True, output='test.ini',\n source='ckan/test-core.ini', override_site_url=False)\n script = scripts.get_script_path('shell.sh')\n if paster:\n script = scripts.get_script_path('paster.sh')\n if command and command != ['help'] and command != ['--help']:\n command += ['--config=/project/development.ini']\n command = [self.extension_dir] + command\n proxy_settings = self._proxy_settings()\n if proxy_settings:\n venv_volumes += ['-v', self.sitedir +\n '/run/proxy-environment:/etc/environment:ro']\n links = {self._get_container_name('solr'): 'solr', self.\n _get_container_name('postgres'): 'db'}\n links.update({self._get_container_name(container): container for\n container in self.extra_containers})\n link_params = []\n for link in links:\n link_params.append('--link')\n link_params.append(link + ':' + links[link])\n if 'datapusher' in self.containers_running():\n link_params.append('--link')\n link_params.append(self._get_container_name('datapusher') +\n ':datapusher')\n return subprocess.call([DOCKER_EXE, 'run'] + (['--rm'] if not\n background else []) + ['-t' if use_tty else '', '-d' if detach else\n '-i'] + venv_volumes + ['-v', self.target + ':/project:rw', '-v', \n self.sitedir + '/files:/var/www/storage:rw', '-v', script +\n ':/scripts/shell.sh:ro', '-v', scripts.get_script_path(\n 'paster_cd.sh') + ':/scripts/paster_cd.sh:ro', '-v', self.sitedir +\n '/run/run.ini:/project/development.ini:ro', '-v', self.sitedir +\n '/run/test.ini:/project/ckan/test-core.ini:ro'] + link_params + [\n '--hostname', self.name, 'datacats/web', '/scripts/shell.sh'] + command\n )\n",
"role": "user"
},
{
"content": "interactive_shell",
"role": "assistant"
}
] |
def x(self, psrc, stream_output=None):
package = self.target + '/' + psrc
assert isdir(package), package
reqname = '/requirements.txt'
if not exists(package + reqname):
reqname = '/pip-requirements.txt'
if not exists(package + reqname):
return
return self.user_run_script(script=scripts.get_script_path(
'install_reqs.sh'), args=['/project/' + psrc + reqname], rw_venv=
True, rw_project=True, stream_output=stream_output)
|
install_package_requirements
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, psrc, stream_output=None):\n package = self.target + '/' + psrc\n assert isdir(package), package\n reqname = '/requirements.txt'\n if not exists(package + reqname):\n reqname = '/pip-requirements.txt'\n if not exists(package + reqname):\n return\n return self.user_run_script(script=scripts.get_script_path(\n 'install_reqs.sh'), args=['/project/' + psrc + reqname], rw_venv=\n True, rw_project=True, stream_output=stream_output)\n",
"role": "user"
},
{
"content": "install_package_requirements",
"role": "assistant"
}
] |
def x(self, container, tail='all', follow=False, timestamps=False):
return container_logs(self._get_container_name(container), tail, follow,
timestamps)
|
logs
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, container, tail='all', follow=False, timestamps=False):\n return container_logs(self._get_container_name(container), tail, follow,\n timestamps)\n",
"role": "user"
},
{
"content": "logs",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.