source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(self, source=None, phase='build', ps=None):
from ambry.etl.pipeline import Pipeline, PartitionWriter
from ambry.dbexceptions import ConfigurationError
if source:
source = self.source(source) if isinstance(source, string_types
) else source
else:
source = None
sf, sp = self.source_pipe(source, ps) if source else (None, None)
pl = Pipeline(self, source=sp)
try:
phase_config = self.default_pipelines[phase]
except KeyError:
phase_config = None
if phase_config:
pl.configure(phase_config)
pipe_config = None
pipe_name = None
if source and source.pipeline:
pipe_name = source.pipeline
try:
pipe_config = self.metadata.pipelines[pipe_name]
except KeyError:
raise ConfigurationError(
"Pipeline '{}' declared in source '{}', but not found in metadata"
.format(source.pipeline, source.name))
else:
pipe_name, pipe_config = self._find_pipeline(source, phase)
if pipe_name:
pl.name = pipe_name
else:
pl.name = phase
pl.phase = phase
def x(pl, pipe_config):
if isinstance(pipe_config, (list, tuple)):
store, body = [], []
for pipe in pipe_config:
store.append(pipe) if isinstance(pipe, PartitionWriter
) else body.append(pipe)
pipe_config = dict(body=body, store=store)
if pipe_config:
pl.configure(pipe_config)
apply_config(pl, pipe_config)
if 'all' in self.metadata.pipelines:
apply_config(pl, self.metadata.pipelines['all'])
self.edit_pipeline(pl)
try:
pl.dest_table = source.dest_table_name
pl.source_table = source.source_table.name
pl.source_name = source.name
except AttributeError:
pl.dest_table = None
return pl
| pipeline | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, source=None, phase='build', ps=None):\n from ambry.etl.pipeline import Pipeline, PartitionWriter\n from ambry.dbexceptions import ConfigurationError\n if source:\n source = self.source(source) if isinstance(source, string_types\n ) else source\n else:\n source = None\n sf, sp = self.source_pipe(source, ps) if source else (None, None)\n pl = Pipeline(self, source=sp)\n try:\n phase_config = self.default_pipelines[phase]\n except KeyError:\n phase_config = None\n if phase_config:\n pl.configure(phase_config)\n pipe_config = None\n pipe_name = None\n if source and source.pipeline:\n pipe_name = source.pipeline\n try:\n pipe_config = self.metadata.pipelines[pipe_name]\n except KeyError:\n raise ConfigurationError(\n \"Pipeline '{}' declared in source '{}', but not found in metadata\"\n .format(source.pipeline, source.name))\n else:\n pipe_name, pipe_config = self._find_pipeline(source, phase)\n if pipe_name:\n pl.name = pipe_name\n else:\n pl.name = phase\n pl.phase = phase\n\n def x(pl, pipe_config):\n if isinstance(pipe_config, (list, tuple)):\n store, body = [], []\n for pipe in pipe_config:\n store.append(pipe) if isinstance(pipe, PartitionWriter\n ) else body.append(pipe)\n pipe_config = dict(body=body, store=store)\n if pipe_config:\n pl.configure(pipe_config)\n apply_config(pl, pipe_config)\n if 'all' in self.metadata.pipelines:\n apply_config(pl, self.metadata.pipelines['all'])\n self.edit_pipeline(pl)\n try:\n pl.dest_table = source.dest_table_name\n pl.source_table = source.source_table.name\n pl.source_name = source.name\n except AttributeError:\n pl.dest_table = None\n return pl\n",
"role": "user"
},
{
"content": "pipeline",
"role": "assistant"
}
] |
def x(self, fields):
row = self.dataset.row(fields)
for i, f in enumerate(fields):
if f == 'bstate':
row[i] = self.state
elif f == 'dstate':
row[i] = self.dstate
elif f == 'source_fs':
row[i] = self.source_fs
elif f.startswith('about'):
_, key = f.split('.')
row[i] = self.metadata.about[key]
elif f.startswith('state'):
_, key = f.split('.')
row[i] = self.buildstate.state[key]
elif f.startswith('count'):
_, key = f.split('.')
if key == 'sources':
row[i] = len(self.dataset.sources)
elif key == 'tables':
row[i] = len(self.dataset.tables)
return row
| field_row | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, fields):\n row = self.dataset.row(fields)\n for i, f in enumerate(fields):\n if f == 'bstate':\n row[i] = self.state\n elif f == 'dstate':\n row[i] = self.dstate\n elif f == 'source_fs':\n row[i] = self.source_fs\n elif f.startswith('about'):\n _, key = f.split('.')\n row[i] = self.metadata.about[key]\n elif f.startswith('state'):\n _, key = f.split('.')\n row[i] = self.buildstate.state[key]\n elif f.startswith('count'):\n _, key = f.split('.')\n if key == 'sources':\n row[i] = len(self.dataset.sources)\n elif key == 'tables':\n row[i] = len(self.dataset.tables)\n return row\n",
"role": "user"
},
{
"content": "field_row",
"role": "assistant"
}
] |
def x(self, source, ps=None):
if isinstance(source, string_types):
source = self.source(source)
source.dataset = self.dataset
source._bundle = self
iter_source, source_pipe = self._iterable_source(source, ps)
if self.limited_run:
source_pipe.limit = 500
return iter_source, source_pipe
| source_pipe | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, source, ps=None):\n if isinstance(source, string_types):\n source = self.source(source)\n source.dataset = self.dataset\n source._bundle = self\n iter_source, source_pipe = self._iterable_source(source, ps)\n if self.limited_run:\n source_pipe.limit = 500\n return iter_source, source_pipe\n",
"role": "user"
},
{
"content": "source_pipe",
"role": "assistant"
}
] |
def x(self, source, clean=False, callback=None):
from fs.zipfs import ZipOpenError
import os
from ambry_sources.sources import GoogleSource, CsvSource, TsvSource, FixedSource, ExcelSource, PartitionSource, SourceError, DelayedOpen, DelayedDownload, ShapefileSource, SocrataSource
from ambry_sources import extract_file_from_zip
spec = source.spec
cache_fs = self.library.download_cache
account_accessor = self.library.account_accessor
url_type = spec.get_urltype()
def x():
from ambry_sources.fetch import download
return download(spec.url, cache_fs, account_accessor, clean=clean,
logger=self.logger, callback=callback)
if url_type == 'file':
from fs.opener import fsopen
syspath = spec.url.replace('file://', '')
cache_path = syspath.strip('/')
cache_fs.makedir(os.path.dirname(cache_path), recursive=True,
allow_recreate=True)
if os.path.isabs(syspath):
with open(syspath) as f:
cache_fs.setcontents(cache_path, f)
else:
cache_fs.setcontents(cache_path, self.source_fs.getcontents(
syspath))
elif url_type not in ('gs', 'socrata'):
try:
cache_path, download_time = do_download()
spec.download_time = download_time
except Exception as e:
from ambry_sources.exceptions import DownloadError
raise DownloadError('Failed to download {}; {}'.format(spec.url, e)
)
else:
cache_path, download_time = None, None
if url_type == 'zip':
try:
fstor = extract_file_from_zip(cache_fs, cache_path, spec.url,
spec.file)
except ZipOpenError:
cache_fs.remove(cache_path)
cache_path, spec.download_time = do_download()
fstor = extract_file_from_zip(cache_fs, cache_path, spec.url,
spec.file)
file_type = spec.get_filetype(fstor.path)
elif url_type == 'gs':
fstor = get_gs(spec.url, spec.segment, account_accessor)
file_type = 'gs'
elif url_type == 'socrata':
spec.encoding = 'utf8'
spec.header_lines = [0]
spec.start_line = 1
url = SocrataSource.download_url(spec)
fstor = DelayedDownload(url, cache_fs)
file_type = 'socrata'
else:
fstor = DelayedOpen(cache_fs, cache_path, 'rb')
file_type = spec.get_filetype(fstor.path)
spec.filetype = file_type
TYPE_TO_SOURCE_MAP = {'gs': GoogleSource, 'csv': CsvSource, 'tsv':
TsvSource, 'fixed': FixedSource, 'txt': FixedSource, 'xls':
ExcelSource, 'xlsx': ExcelSource, 'partition': PartitionSource,
'shape': ShapefileSource, 'socrata': SocrataSource}
cls = TYPE_TO_SOURCE_MAP.get(file_type)
if cls is None:
raise SourceError(
"Failed to determine file type for source '{}'; unknown type '{}' "
.format(spec.name, file_type))
return cls(spec, fstor)
| get_source | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, source, clean=False, callback=None):\n from fs.zipfs import ZipOpenError\n import os\n from ambry_sources.sources import GoogleSource, CsvSource, TsvSource, FixedSource, ExcelSource, PartitionSource, SourceError, DelayedOpen, DelayedDownload, ShapefileSource, SocrataSource\n from ambry_sources import extract_file_from_zip\n spec = source.spec\n cache_fs = self.library.download_cache\n account_accessor = self.library.account_accessor\n url_type = spec.get_urltype()\n\n def x():\n from ambry_sources.fetch import download\n return download(spec.url, cache_fs, account_accessor, clean=clean,\n logger=self.logger, callback=callback)\n if url_type == 'file':\n from fs.opener import fsopen\n syspath = spec.url.replace('file://', '')\n cache_path = syspath.strip('/')\n cache_fs.makedir(os.path.dirname(cache_path), recursive=True,\n allow_recreate=True)\n if os.path.isabs(syspath):\n with open(syspath) as f:\n cache_fs.setcontents(cache_path, f)\n else:\n cache_fs.setcontents(cache_path, self.source_fs.getcontents(\n syspath))\n elif url_type not in ('gs', 'socrata'):\n try:\n cache_path, download_time = do_download()\n spec.download_time = download_time\n except Exception as e:\n from ambry_sources.exceptions import DownloadError\n raise DownloadError('Failed to download {}; {}'.format(spec.url, e)\n )\n else:\n cache_path, download_time = None, None\n if url_type == 'zip':\n try:\n fstor = extract_file_from_zip(cache_fs, cache_path, spec.url,\n spec.file)\n except ZipOpenError:\n cache_fs.remove(cache_path)\n cache_path, spec.download_time = do_download()\n fstor = extract_file_from_zip(cache_fs, cache_path, spec.url,\n spec.file)\n file_type = spec.get_filetype(fstor.path)\n elif url_type == 'gs':\n fstor = get_gs(spec.url, spec.segment, account_accessor)\n file_type = 'gs'\n elif url_type == 'socrata':\n spec.encoding = 'utf8'\n spec.header_lines = [0]\n spec.start_line = 1\n url = SocrataSource.download_url(spec)\n fstor = DelayedDownload(url, cache_fs)\n file_type = 'socrata'\n else:\n fstor = DelayedOpen(cache_fs, cache_path, 'rb')\n file_type = spec.get_filetype(fstor.path)\n spec.filetype = file_type\n TYPE_TO_SOURCE_MAP = {'gs': GoogleSource, 'csv': CsvSource, 'tsv':\n TsvSource, 'fixed': FixedSource, 'txt': FixedSource, 'xls':\n ExcelSource, 'xlsx': ExcelSource, 'partition': PartitionSource,\n 'shape': ShapefileSource, 'socrata': SocrataSource}\n cls = TYPE_TO_SOURCE_MAP.get(file_type)\n if cls is None:\n raise SourceError(\n \"Failed to determine file type for source '{}'; unknown type '{}' \"\n .format(spec.name, file_type))\n return cls(spec, fstor)\n",
"role": "user"
},
{
"content": "get_source",
"role": "assistant"
}
] |
def x(self):
self.buildstate.state.lasttime = time()
self.buildstate.commit()
return self.buildstate.state.error
| error_state | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.buildstate.state.lasttime = time()\n self.buildstate.commit()\n return self.buildstate.state.error\n",
"role": "user"
},
{
"content": "error_state",
"role": "assistant"
}
] |
def x(self, state):
assert state != 'build_bundle'
self.buildstate.state.current = state
self.buildstate.state[state] = time()
self.buildstate.state.lasttime = time()
self.buildstate.state.error = False
self.buildstate.state.exception = None
self.buildstate.state.exception_type = None
self.buildstate.commit()
if state in (self.STATES.NEW, self.STATES.CLEANED, self.STATES.BUILT,
self.STATES.FINALIZED, self.STATES.SOURCE):
state = state if state != self.STATES.CLEANED else self.STATES.NEW
self.dstate = state
| state | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, state):\n assert state != 'build_bundle'\n self.buildstate.state.current = state\n self.buildstate.state[state] = time()\n self.buildstate.state.lasttime = time()\n self.buildstate.state.error = False\n self.buildstate.state.exception = None\n self.buildstate.state.exception_type = None\n self.buildstate.commit()\n if state in (self.STATES.NEW, self.STATES.CLEANED, self.STATES.BUILT,\n self.STATES.FINALIZED, self.STATES.SOURCE):\n state = state if state != self.STATES.CLEANED else self.STATES.NEW\n self.dstate = state\n",
"role": "user"
},
{
"content": "state",
"role": "assistant"
}
] |
def x(self, phase, stage):
key = '{}-{}'.format(phase, stage if stage else 1)
self.buildstate.state[key] = time()
| record_stage_state | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, phase, stage):\n key = '{}-{}'.format(phase, stage if stage else 1)\n self.buildstate.state[key] = time()\n",
"role": "user"
},
{
"content": "record_stage_state",
"role": "assistant"
}
] |
def x(self, tag):
import time
self.buildstate.access.last = '{}-{}'.format(tag, time.time())
self.buildstate.commit()
| set_last_access | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tag):\n import time\n self.buildstate.access.last = '{}-{}'.format(tag, time.time())\n self.buildstate.commit()\n",
"role": "user"
},
{
"content": "set_last_access",
"role": "assistant"
}
] |
def x(self, force=False):
self.log('---- Sync In ----')
self.dstate = self.STATES.BUILDING
for path_name in self.source_fs.listdir():
f = self.build_source_files.instance_from_name(path_name)
if not f:
self.warn('Ignoring unknown file: {}'.format(path_name))
continue
if f and f.exists and (f.fs_is_newer or force):
self.log('Sync: {}'.format(f.record.path))
f.fs_to_record()
f.record_to_objects()
self.commit()
self.library.search.index_bundle(self, force=True)
| sync_in | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, force=False):\n self.log('---- Sync In ----')\n self.dstate = self.STATES.BUILDING\n for path_name in self.source_fs.listdir():\n f = self.build_source_files.instance_from_name(path_name)\n if not f:\n self.warn('Ignoring unknown file: {}'.format(path_name))\n continue\n if f and f.exists and (f.fs_is_newer or force):\n self.log('Sync: {}'.format(f.record.path))\n f.fs_to_record()\n f.record_to_objects()\n self.commit()\n self.library.search.index_bundle(self, force=True)\n",
"role": "user"
},
{
"content": "sync_in",
"role": "assistant"
}
] |
def x(self, force=False):
self.log('---- Sync Files ----')
self.dstate = self.STATES.BUILDING
for f in self.build_source_files:
if self.source_fs.exists(f.record.path):
if f.fs_is_newer or force:
self.log('Sync: {}'.format(f.record.path))
f.fs_to_record()
self.commit()
| sync_in_files | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, force=False):\n self.log('---- Sync Files ----')\n self.dstate = self.STATES.BUILDING\n for f in self.build_source_files:\n if self.source_fs.exists(f.record.path):\n if f.fs_is_newer or force:\n self.log('Sync: {}'.format(f.record.path))\n f.fs_to_record()\n self.commit()\n",
"role": "user"
},
{
"content": "sync_in_files",
"role": "assistant"
}
] |
def x(self, force=False):
self.log('---- Sync Files ----')
for f in self.build_source_files:
f.record_to_objects()
self.build_source_files.file(File.BSFILE.META).record_to_objects()
self.commit()
| sync_in_records | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, force=False):\n self.log('---- Sync Files ----')\n for f in self.build_source_files:\n f.record_to_objects()\n self.build_source_files.file(File.BSFILE.META).record_to_objects()\n self.commit()\n",
"role": "user"
},
{
"content": "sync_in_records",
"role": "assistant"
}
] |
def x(self, file_name=None, force=False):
self.log('---- Sync Out ----')
from ambry.bundle.files import BuildSourceFile
self.dstate = self.STATES.BUILDING
for f in self.build_source_files.list_records():
if (f.sync_dir() == BuildSourceFile.SYNC_DIR.RECORD_TO_FILE or f.
record.path == file_name) or force:
self.log('Sync: {}'.format(f.record.path))
f.record_to_fs()
self.commit()
| sync_out | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, file_name=None, force=False):\n self.log('---- Sync Out ----')\n from ambry.bundle.files import BuildSourceFile\n self.dstate = self.STATES.BUILDING\n for f in self.build_source_files.list_records():\n if (f.sync_dir() == BuildSourceFile.SYNC_DIR.RECORD_TO_FILE or f.\n record.path == file_name) or force:\n self.log('Sync: {}'.format(f.record.path))\n f.record_to_fs()\n self.commit()\n",
"role": "user"
},
{
"content": "sync_out",
"role": "assistant"
}
] |
def x(self):
self.dstate = self.STATES.BUILDING
self.build_source_files.record_to_objects()
| sync_objects_in | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.dstate = self.STATES.BUILDING\n self.build_source_files.record_to_objects()\n",
"role": "user"
},
{
"content": "sync_objects_in",
"role": "assistant"
}
] |
def x(self, force=False):
self.log('---- Sync Objects Out ----')
from ambry.bundle.files import BuildSourceFile
self.dstate = self.STATES.BUILDING
for f in self.build_source_files.list_records():
self.log('Sync: {}'.format(f.record.path))
f.objects_to_record()
self.commit()
| sync_objects_out | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, force=False):\n self.log('---- Sync Objects Out ----')\n from ambry.bundle.files import BuildSourceFile\n self.dstate = self.STATES.BUILDING\n for f in self.build_source_files.list_records():\n self.log('Sync: {}'.format(f.record.path))\n f.objects_to_record()\n self.commit()\n",
"role": "user"
},
{
"content": "sync_objects_out",
"role": "assistant"
}
] |
def x(self):
from ambry.orm.file import File
from ambry.bundle.files import BuildSourceFile
self.dstate = self.STATES.BUILDING
synced = 0
for fc in [File.BSFILE.BUILD, File.BSFILE.META, File.BSFILE.LIB, File.
BSFILE.TEST, File.BSFILE.DOC]:
bsf = self.build_source_files.file(fc)
if bsf.fs_is_newer:
self.log('Syncing {}'.format(bsf.file_name))
bsf.sync(BuildSourceFile.SYNC_DIR.FILE_TO_RECORD)
synced += 1
self.build_source_files.file(File.BSFILE.META).record_to_objects()
return synced
| sync_code | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry.orm.file import File\n from ambry.bundle.files import BuildSourceFile\n self.dstate = self.STATES.BUILDING\n synced = 0\n for fc in [File.BSFILE.BUILD, File.BSFILE.META, File.BSFILE.LIB, File.\n BSFILE.TEST, File.BSFILE.DOC]:\n bsf = self.build_source_files.file(fc)\n if bsf.fs_is_newer:\n self.log('Syncing {}'.format(bsf.file_name))\n bsf.sync(BuildSourceFile.SYNC_DIR.FILE_TO_RECORD)\n synced += 1\n self.build_source_files.file(File.BSFILE.META).record_to_objects()\n return synced\n",
"role": "user"
},
{
"content": "sync_code",
"role": "assistant"
}
] |
def x(self, force=False):
from ambry.orm.file import File
self.dstate = self.STATES.BUILDING
synced = 0
for fc in [File.BSFILE.SOURCES]:
bsf = self.build_source_files.file(fc)
if bsf.fs_is_newer or force:
self.log('Syncing {}'.format(bsf.file_name))
bsf.fs_to_objects()
synced += 1
return synced
| sync_sources | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, force=False):\n from ambry.orm.file import File\n self.dstate = self.STATES.BUILDING\n synced = 0\n for fc in [File.BSFILE.SOURCES]:\n bsf = self.build_source_files.file(fc)\n if bsf.fs_is_newer or force:\n self.log('Syncing {}'.format(bsf.file_name))\n bsf.fs_to_objects()\n synced += 1\n return synced\n",
"role": "user"
},
{
"content": "sync_sources",
"role": "assistant"
}
] |
def x(self):
from ambry.orm.file import File
from ambry.bundle.files import BuildSourceFile
self.dstate = self.STATES.BUILDING
synced = 0
for fc in [File.BSFILE.SCHEMA, File.BSFILE.SOURCESCHEMA]:
bsf = self.build_source_files.file(fc)
if bsf.fs_is_newer:
self.log('Syncing {}'.format(bsf.file_name))
bsf.sync(BuildSourceFile.SYNC_DIR.FILE_TO_RECORD)
synced += 1
return synced
| sync_schema | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry.orm.file import File\n from ambry.bundle.files import BuildSourceFile\n self.dstate = self.STATES.BUILDING\n synced = 0\n for fc in [File.BSFILE.SCHEMA, File.BSFILE.SOURCESCHEMA]:\n bsf = self.build_source_files.file(fc)\n if bsf.fs_is_newer:\n self.log('Syncing {}'.format(bsf.file_name))\n bsf.sync(BuildSourceFile.SYNC_DIR.FILE_TO_RECORD)\n synced += 1\n return synced\n",
"role": "user"
},
{
"content": "sync_schema",
"role": "assistant"
}
] |
def x(self):
self.commit()
self.build_source_files.schema.objects_to_record()
self.commit()
| update_schema | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.commit()\n self.build_source_files.schema.objects_to_record()\n self.commit()\n",
"role": "user"
},
{
"content": "update_schema",
"role": "assistant"
}
] |
def x(self, force=False):
if self.is_finalized and not force:
self.warn("Can't clean; bundle is finalized")
return False
self.log('---- Cleaning ----')
self.state = self.STATES.CLEANING
self.dstate = self.STATES.BUILDING
self.commit()
self.clean_sources()
self.clean_tables()
self.clean_partitions()
self.clean_build()
self.clean_files()
self.clean_ingested()
self.clean_build_state()
self.clean_progress()
self.state = self.STATES.CLEANED
self.commit()
return True
| clean | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, force=False):\n if self.is_finalized and not force:\n self.warn(\"Can't clean; bundle is finalized\")\n return False\n self.log('---- Cleaning ----')\n self.state = self.STATES.CLEANING\n self.dstate = self.STATES.BUILDING\n self.commit()\n self.clean_sources()\n self.clean_tables()\n self.clean_partitions()\n self.clean_build()\n self.clean_files()\n self.clean_ingested()\n self.clean_build_state()\n self.clean_progress()\n self.state = self.STATES.CLEANED\n self.commit()\n return True\n",
"role": "user"
},
{
"content": "clean",
"role": "assistant"
}
] |
def x(self):
if self.is_finalized:
self.warn("Can't clean; bundle is finalized")
return False
self.log('---- Cleaning ----')
self.state = self.STATES.CLEANING
self.commit()
self.clean_sources()
self.clean_tables()
self.clean_partitions()
self.clean_build()
self.clean_ingested()
self.clean_build_state()
self.state = self.STATES.CLEANED
self.commit()
self.log('---- Done Cleaning ----')
return True
| clean_except_files | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.is_finalized:\n self.warn(\"Can't clean; bundle is finalized\")\n return False\n self.log('---- Cleaning ----')\n self.state = self.STATES.CLEANING\n self.commit()\n self.clean_sources()\n self.clean_tables()\n self.clean_partitions()\n self.clean_build()\n self.clean_ingested()\n self.clean_build_state()\n self.state = self.STATES.CLEANED\n self.commit()\n self.log('---- Done Cleaning ----')\n return True\n",
"role": "user"
},
{
"content": "clean_except_files",
"role": "assistant"
}
] |
def x(self):
for src in self.dataset.sources:
src.st_id = None
src.t_id = None
self.dataset.sources[:] = []
self.dataset.source_tables[:] = []
self.dataset.st_sequence_id = 1
| clean_sources | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n for src in self.dataset.sources:\n src.st_id = None\n src.t_id = None\n self.dataset.sources[:] = []\n self.dataset.source_tables[:] = []\n self.dataset.st_sequence_id = 1\n",
"role": "user"
},
{
"content": "clean_sources",
"role": "assistant"
}
] |
def x(self):
import shutil
from ambry.orm import ColumnStat
self.session.query(ColumnStat).filter(ColumnStat.d_vid == self.dataset.vid
).delete()
self.dataset.delete_partitions()
for s in self.sources:
s.state = None
if self.build_partition_fs.exists:
try:
shutil.rmtree(self.build_partition_fs.getsyspath('/'))
except NoSysPathError:
pass
| clean_partitions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n import shutil\n from ambry.orm import ColumnStat\n self.session.query(ColumnStat).filter(ColumnStat.d_vid == self.dataset.vid\n ).delete()\n self.dataset.delete_partitions()\n for s in self.sources:\n s.state = None\n if self.build_partition_fs.exists:\n try:\n shutil.rmtree(self.build_partition_fs.getsyspath('/'))\n except NoSysPathError:\n pass\n",
"role": "user"
},
{
"content": "clean_partitions",
"role": "assistant"
}
] |
def x(self):
import shutil
if self.build_fs.exists:
try:
shutil.rmtree(self.build_fs.getsyspath('/'))
except NoSysPathError:
pass
| clean_build | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n import shutil\n if self.build_fs.exists:\n try:\n shutil.rmtree(self.build_fs.getsyspath('/'))\n except NoSysPathError:\n pass\n",
"role": "user"
},
{
"content": "clean_build",
"role": "assistant"
}
] |
def x(self):
for s in self.sources:
df = s.datafile
if df.exists and not s.is_partition:
df.remove()
s.state = s.STATES.NEW
self.commit()
| clean_ingested | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n for s in self.sources:\n df = s.datafile\n if df.exists and not s.is_partition:\n df.remove()\n s.state = s.STATES.NEW\n self.commit()\n",
"role": "user"
},
{
"content": "clean_ingested",
"role": "assistant"
}
] |
def x(self):
ds = self.dataset
ds.config.build.clean()
ds.config.process.clean()
ds.commit()
self.state = self.STATES.CLEANED
| clean_process_meta | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n ds = self.dataset\n ds.config.build.clean()\n ds.config.process.clean()\n ds.commit()\n self.state = self.STATES.CLEANED\n",
"role": "user"
},
{
"content": "clean_process_meta",
"role": "assistant"
}
] |
def x(self):
self.build_source_files.file(File.BSFILE.SOURCESCHEMA).remove()
self.build_source_files.file(File.BSFILE.SCHEMA).remove()
self.commit()
| clean_source_files | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.build_source_files.file(File.BSFILE.SOURCESCHEMA).remove()\n self.build_source_files.file(File.BSFILE.SCHEMA).remove()\n self.commit()\n",
"role": "user"
},
{
"content": "clean_source_files",
"role": "assistant"
}
] |
def x(self, sources=None, tables=None, stage=None, force=False, load_meta=False
):
from itertools import groupby
from ambry.bundle.events import TAG
from fs.errors import ResourceNotFoundError
import zlib
self.log('---- Ingesting ----')
self.dstate = self.STATES.BUILDING
self.commit()
key = lambda s: s.stage if s.stage else 1
def x(s):
import zlib
if force:
return True
try:
return s.is_processable and not s.is_ingested and not s.is_built
except (IOError, zlib.error):
s.local_datafile.remove()
return True
sources = sorted(self._resolve_sources(sources, tables, stage,
predicate=not_final_or_delete), key=key)
if not sources:
self.log('No sources left to ingest')
return
self.state = self.STATES.INGESTING
count = 0
errors = 0
self._run_events(TAG.BEFORE_INGEST, 0)
for s in self.sources:
if s.is_downloadable:
df = s.datafile
try:
info = df.info
df.close()
except (ResourceNotFoundError, zlib.error, IOError):
df.remove()
for stage, g in groupby(sources, key):
sources = [s for s in g if not_final_or_delete(s)]
if not len(sources):
continue
self._run_events(TAG.BEFORE_INGEST, stage)
stage_errors = self._ingest_sources(sources, stage, force=force)
errors += stage_errors
count += len(sources) - stage_errors
self._run_events(TAG.AFTER_INGEST, stage)
self.record_stage_state(self.STATES.INGESTING, stage)
self.state = self.STATES.INGESTED
try:
pass
finally:
self._run_events(TAG.AFTER_INGEST, 0)
self.log('Ingested {} sources'.format(count))
if load_meta:
if len(sources) == 1:
iterable_source, source_pipe = self.source_pipe(sources[0])
try:
meta = iterable_source.meta
if meta:
self.metadata.about.title = meta['title']
self.metadata.about.summary = meta['summary']
self.build_source_files.bundle_meta.objects_to_record()
except AttributeError as e:
self.warn('Failed to set metadata: {}'.format(e))
pass
else:
self.warn(
"Didn't not load meta from source. Must have exactly one soruce, got {}"
.format(len(sources)))
self.commit()
if errors == 0:
return True
else:
return False
| ingest | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, sources=None, tables=None, stage=None, force=False, load_meta=False\n ):\n from itertools import groupby\n from ambry.bundle.events import TAG\n from fs.errors import ResourceNotFoundError\n import zlib\n self.log('---- Ingesting ----')\n self.dstate = self.STATES.BUILDING\n self.commit()\n key = lambda s: s.stage if s.stage else 1\n\n def x(s):\n import zlib\n if force:\n return True\n try:\n return s.is_processable and not s.is_ingested and not s.is_built\n except (IOError, zlib.error):\n s.local_datafile.remove()\n return True\n sources = sorted(self._resolve_sources(sources, tables, stage,\n predicate=not_final_or_delete), key=key)\n if not sources:\n self.log('No sources left to ingest')\n return\n self.state = self.STATES.INGESTING\n count = 0\n errors = 0\n self._run_events(TAG.BEFORE_INGEST, 0)\n for s in self.sources:\n if s.is_downloadable:\n df = s.datafile\n try:\n info = df.info\n df.close()\n except (ResourceNotFoundError, zlib.error, IOError):\n df.remove()\n for stage, g in groupby(sources, key):\n sources = [s for s in g if not_final_or_delete(s)]\n if not len(sources):\n continue\n self._run_events(TAG.BEFORE_INGEST, stage)\n stage_errors = self._ingest_sources(sources, stage, force=force)\n errors += stage_errors\n count += len(sources) - stage_errors\n self._run_events(TAG.AFTER_INGEST, stage)\n self.record_stage_state(self.STATES.INGESTING, stage)\n self.state = self.STATES.INGESTED\n try:\n pass\n finally:\n self._run_events(TAG.AFTER_INGEST, 0)\n self.log('Ingested {} sources'.format(count))\n if load_meta:\n if len(sources) == 1:\n iterable_source, source_pipe = self.source_pipe(sources[0])\n try:\n meta = iterable_source.meta\n if meta:\n self.metadata.about.title = meta['title']\n self.metadata.about.summary = meta['summary']\n self.build_source_files.bundle_meta.objects_to_record()\n except AttributeError as e:\n self.warn('Failed to set metadata: {}'.format(e))\n pass\n else:\n self.warn(\n \"Didn't not load meta from source. Must have exactly one soruce, got {}\"\n .format(len(sources)))\n self.commit()\n if errors == 0:\n return True\n else:\n return False\n",
"role": "user"
},
{
"content": "ingest",
"role": "assistant"
}
] |
def x(self, sources, stage, force=False):
from concurrent import ingest_mp
self.state = self.STATES.INGESTING
downloadable_sources = [s for s in sources if force or s.is_processable and
not s.is_ingested and not s.is_built]
errors = 0
with self.progress.start('ingest', stage, message='Ingesting ' + ('MP' if
self.multi else 'SP'), item_total=len(sources), item_type='source',
item_count=len(downloadable_sources)) as ps:
for source in sources:
_ = source.source_table
if self.multi:
args = [(self.identity.vid, stage, source.vid, force) for
source in downloadable_sources]
pool = self.library.process_pool(limited_run=self.limited_run)
try:
result = pool.map_async(ingest_mp, args, 1)
pool.close()
pool.join()
except KeyboardInterrupt:
self.log('Got keyboard interrrupt; terminating workers')
pool.terminate()
raise
else:
for i, source in enumerate(downloadable_sources, 1):
ps.add(message='Ingesting source #{}, {}'.format(i, source.
name), source=source, state='running')
r = self._ingest_source(source, ps, force)
if not r:
errors += 1
if errors > 0:
from ambry.dbexceptions import IngestionError
raise IngestionError('Failed to ingest {} sources'.format(errors))
return errors
| ingest_sources | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, sources, stage, force=False):\n from concurrent import ingest_mp\n self.state = self.STATES.INGESTING\n downloadable_sources = [s for s in sources if force or s.is_processable and\n not s.is_ingested and not s.is_built]\n errors = 0\n with self.progress.start('ingest', stage, message='Ingesting ' + ('MP' if\n self.multi else 'SP'), item_total=len(sources), item_type='source',\n item_count=len(downloadable_sources)) as ps:\n for source in sources:\n _ = source.source_table\n if self.multi:\n args = [(self.identity.vid, stage, source.vid, force) for\n source in downloadable_sources]\n pool = self.library.process_pool(limited_run=self.limited_run)\n try:\n result = pool.map_async(ingest_mp, args, 1)\n pool.close()\n pool.join()\n except KeyboardInterrupt:\n self.log('Got keyboard interrrupt; terminating workers')\n pool.terminate()\n raise\n else:\n for i, source in enumerate(downloadable_sources, 1):\n ps.add(message='Ingesting source #{}, {}'.format(i, source.\n name), source=source, state='running')\n r = self._ingest_source(source, ps, force)\n if not r:\n errors += 1\n if errors > 0:\n from ambry.dbexceptions import IngestionError\n raise IngestionError('Failed to ingest {} sources'.format(errors))\n return errors\n",
"role": "user"
},
{
"content": "ingest_sources",
"role": "assistant"
}
] |
def x(self, source, ps, force=None):
from ambry.bundle.process import call_interval
try:
from ambry.orm.exc import NotFoundError
if not source.is_partition and source.datafile.exists:
if not source.datafile.is_finalized:
source.datafile.remove()
elif force:
source.datafile.remove()
else:
ps.update(message='Source {} already ingested, skipping'.
format(source.name), state='skipped')
return True
if source.is_partition:
try:
self.library.partition(source.ref)
except NotFoundError:
ps.update(message=
"Not Ingesting {}: referenced partition '{}' does not exist"
.format(source.name, source.ref), state='skipped')
return True
source.state = source.STATES.INGESTING
iterable_source, source_pipe = self.source_pipe(source, ps)
if not source.is_ingestible:
ps.update(message='Not an ingestiable source: {}'.format(source
.name), state='skipped', source=source)
source.state = source.STATES.NOTINGESTABLE
return True
ps.update('Ingesting {} from {}'.format(source.spec.name, source.
url or source.generator), item_type='rows', item_count=0)
@call_interval(5)
def x(i):
desc, n_records, total, rate = source.datafile.report_progress()
ps.update(message='Ingesting {}: rate: {}'.format(source.spec.
name, rate), item_count=n_records)
source.datafile.load_rows(iterable_source, callback=
ingest_progress_f, limit=500 if self.limited_run else None,
intuit_type=True, run_stats=False)
if source.datafile.meta['warnings']:
for w in source.datafile.meta['warnings']:
self.error('Ingestion error: {}'.format(w))
ps.update(message='Ingested to {}'.format(source.datafile.syspath))
ps.update(message='Updating tables and specs for {}'.format(source.
name))
source.update_spec()
if self.limited_run:
source.end_line = None
self.build_source_files.sources.objects_to_record()
ps.update(message='Ingested {}'.format(source.datafile.path), state
='done')
source.state = source.STATES.INGESTED
self.commit()
return True
except Exception as e:
import traceback
from ambry.util import qualified_class_name
ps.update(message='Source {} failed with exception: {}'.format(
source.name, e), exception_class=qualified_class_name(e),
exception_trace=str(traceback.format_exc()), state='error')
source.state = source.STATES.INGESTING + '_error'
self.commit()
return False
| ingest_source | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, source, ps, force=None):\n from ambry.bundle.process import call_interval\n try:\n from ambry.orm.exc import NotFoundError\n if not source.is_partition and source.datafile.exists:\n if not source.datafile.is_finalized:\n source.datafile.remove()\n elif force:\n source.datafile.remove()\n else:\n ps.update(message='Source {} already ingested, skipping'.\n format(source.name), state='skipped')\n return True\n if source.is_partition:\n try:\n self.library.partition(source.ref)\n except NotFoundError:\n ps.update(message=\n \"Not Ingesting {}: referenced partition '{}' does not exist\"\n .format(source.name, source.ref), state='skipped')\n return True\n source.state = source.STATES.INGESTING\n iterable_source, source_pipe = self.source_pipe(source, ps)\n if not source.is_ingestible:\n ps.update(message='Not an ingestiable source: {}'.format(source\n .name), state='skipped', source=source)\n source.state = source.STATES.NOTINGESTABLE\n return True\n ps.update('Ingesting {} from {}'.format(source.spec.name, source.\n url or source.generator), item_type='rows', item_count=0)\n\n @call_interval(5)\n def x(i):\n desc, n_records, total, rate = source.datafile.report_progress()\n ps.update(message='Ingesting {}: rate: {}'.format(source.spec.\n name, rate), item_count=n_records)\n source.datafile.load_rows(iterable_source, callback=\n ingest_progress_f, limit=500 if self.limited_run else None,\n intuit_type=True, run_stats=False)\n if source.datafile.meta['warnings']:\n for w in source.datafile.meta['warnings']:\n self.error('Ingestion error: {}'.format(w))\n ps.update(message='Ingested to {}'.format(source.datafile.syspath))\n ps.update(message='Updating tables and specs for {}'.format(source.\n name))\n source.update_spec()\n if self.limited_run:\n source.end_line = None\n self.build_source_files.sources.objects_to_record()\n ps.update(message='Ingested {}'.format(source.datafile.path), state\n ='done')\n source.state = source.STATES.INGESTED\n self.commit()\n return True\n except Exception as e:\n import traceback\n from ambry.util import qualified_class_name\n ps.update(message='Source {} failed with exception: {}'.format(\n source.name, e), exception_class=qualified_class_name(e),\n exception_trace=str(traceback.format_exc()), state='error')\n source.state = source.STATES.INGESTING + '_error'\n self.commit()\n return False\n",
"role": "user"
},
{
"content": "ingest_source",
"role": "assistant"
}
] |
def x(self, sources=None, tables=None, clean=False):
sources = self._resolve_sources(sources, tables, None, predicate=lambda
s: s.is_processable)
for source in sources:
source.update_table()
self.log("Creating source schema for '{}': Table {}, {} columns".
format(source.name, source.source_table.name, len(source.
source_table.columns)))
self.commit()
| source_schema | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, sources=None, tables=None, clean=False):\n sources = self._resolve_sources(sources, tables, None, predicate=lambda\n s: s.is_processable)\n for source in sources:\n source.update_table()\n self.log(\"Creating source schema for '{}': Table {}, {} columns\".\n format(source.name, source.source_table.name, len(source.\n source_table.columns)))\n self.commit()\n",
"role": "user"
},
{
"content": "source_schema",
"role": "assistant"
}
] |
def x(self, sources=None, tables=None, clean=False, force=False,
use_pipeline=False):
from itertools import groupby
from operator import attrgetter
from ambry.etl import Collect, Head
from ambry.orm.exc import NotFoundError
self.dstate = self.STATES.BUILDING
self.commit()
self.log('---- Schema ----')
resolved_sources = self._resolve_sources(sources, tables, predicate=lambda
s: s.is_processable)
if clean:
self.dataset.delete_tables_partitions()
self.commit()
keyfunc = attrgetter('dest_table')
for t, table_sources in groupby(sorted(resolved_sources, key=keyfunc),
keyfunc):
if use_pipeline:
for source in table_sources:
pl = self.pipeline(source)
pl.cast = [ambry.etl.CastSourceColumns]
pl.select_partition = []
pl.write = [Head, Collect]
pl.final = []
self.log_pipeline(pl)
pl.run()
pl.phase = 'build_schema'
self.log_pipeline(pl)
for h, c in zip(pl.write[Collect].headers, pl.write[Collect
].rows[1]):
c = t.add_column(name=h, datatype=type(c).__name__ if c
is not None else 'str', update_existing=True)
self.log("Populated destination table '{}' from pipeline '{}'".
format(t.name, pl.name))
else:
self.commit()
def x(source):
if source.is_partition and not source.source_table_exists:
return enumerate(source.partition.table.columns)
else:
return enumerate(source.source_table.columns)
columns = sorted(set([(i, col.dest_header, col.datatype, col.
description, col.has_codes) for source in table_sources for
i, col in source_cols(source)]))
initial_count = len(t.columns)
for pos, name, datatype, desc, has_codes in columns:
kwds = dict(name=name, datatype=datatype, description=desc,
update_existing=True)
try:
extant = t.column(name)
except NotFoundError:
extant = None
if extant is None or not extant.description:
kwds['description'] = desc
c = t.add_column(**kwds)
final_count = len(t.columns)
if final_count > initial_count:
diff = final_count - initial_count
self.log(
"Populated destination table '{}' from source table '{}' with {} columns"
.format(t.name, source.source_table.name, diff))
self.commit()
return True
| schema | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, sources=None, tables=None, clean=False, force=False,\n use_pipeline=False):\n from itertools import groupby\n from operator import attrgetter\n from ambry.etl import Collect, Head\n from ambry.orm.exc import NotFoundError\n self.dstate = self.STATES.BUILDING\n self.commit()\n self.log('---- Schema ----')\n resolved_sources = self._resolve_sources(sources, tables, predicate=lambda\n s: s.is_processable)\n if clean:\n self.dataset.delete_tables_partitions()\n self.commit()\n keyfunc = attrgetter('dest_table')\n for t, table_sources in groupby(sorted(resolved_sources, key=keyfunc),\n keyfunc):\n if use_pipeline:\n for source in table_sources:\n pl = self.pipeline(source)\n pl.cast = [ambry.etl.CastSourceColumns]\n pl.select_partition = []\n pl.write = [Head, Collect]\n pl.final = []\n self.log_pipeline(pl)\n pl.run()\n pl.phase = 'build_schema'\n self.log_pipeline(pl)\n for h, c in zip(pl.write[Collect].headers, pl.write[Collect\n ].rows[1]):\n c = t.add_column(name=h, datatype=type(c).__name__ if c\n is not None else 'str', update_existing=True)\n self.log(\"Populated destination table '{}' from pipeline '{}'\".\n format(t.name, pl.name))\n else:\n self.commit()\n\n def x(source):\n if source.is_partition and not source.source_table_exists:\n return enumerate(source.partition.table.columns)\n else:\n return enumerate(source.source_table.columns)\n columns = sorted(set([(i, col.dest_header, col.datatype, col.\n description, col.has_codes) for source in table_sources for\n i, col in source_cols(source)]))\n initial_count = len(t.columns)\n for pos, name, datatype, desc, has_codes in columns:\n kwds = dict(name=name, datatype=datatype, description=desc,\n update_existing=True)\n try:\n extant = t.column(name)\n except NotFoundError:\n extant = None\n if extant is None or not extant.description:\n kwds['description'] = desc\n c = t.add_column(**kwds)\n final_count = len(t.columns)\n if final_count > initial_count:\n diff = final_count - initial_count\n self.log(\n \"Populated destination table '{}' from source table '{}' with {} columns\"\n .format(t.name, source.source_table.name, diff))\n self.commit()\n return True\n",
"role": "user"
},
{
"content": "schema",
"role": "assistant"
}
] |
def x(self, sources):
from ambry.orm.exc import NotFoundError
for p in self.dataset.partitions:
if p.type == p.TYPE.SEGMENT:
self.log('Removing old segment partition: {}'.format(p.identity
.name))
try:
self.wrap_partition(p).local_datafile.remove()
self.session.delete(p)
except NotFoundError:
pass
for s in sources:
if s.reftype == 'partition':
continue
p = s.partition
if p:
try:
self.wrap_partition(p).local_datafile.remove()
self.session.delete(p)
except NotFoundError:
pass
if s.state in (self.STATES.BUILDING, self.STATES.BUILT):
s.state = self.STATES.INGESTED
self.commit()
| reset_build | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, sources):\n from ambry.orm.exc import NotFoundError\n for p in self.dataset.partitions:\n if p.type == p.TYPE.SEGMENT:\n self.log('Removing old segment partition: {}'.format(p.identity\n .name))\n try:\n self.wrap_partition(p).local_datafile.remove()\n self.session.delete(p)\n except NotFoundError:\n pass\n for s in sources:\n if s.reftype == 'partition':\n continue\n p = s.partition\n if p:\n try:\n self.wrap_partition(p).local_datafile.remove()\n self.session.delete(p)\n except NotFoundError:\n pass\n if s.state in (self.STATES.BUILDING, self.STATES.BUILT):\n s.state = self.STATES.INGESTED\n self.commit()\n",
"role": "user"
},
{
"content": "reset_build",
"role": "assistant"
}
] |
def x(self, sources=None, tables=None, stage=None, force=False):
from operator import attrgetter
from itertools import groupby
from .concurrent import build_mp, unify_mp
from ambry.bundle.events import TAG
self.log('==== Building ====')
self.state = self.STATES.BUILDING
self.commit()
class SourceSet(object):
"""Container for sources that can reload them after they get expired from the session"""
def x(self, bundle, v):
self.bundle = bundle
self.sources = v
self._s_vids = [s.vid for s in self.sources]
def x(self):
self.sources = [self.bundle.source(vid) for vid in self._s_vids]
def x(self):
for s in self.sources:
yield s
def x(self):
return len(self._s_vids)
self._run_events(TAG.BEFORE_BUILD, 0)
resolved_sources = SourceSet(self, self._resolve_sources(sources,
tables, stage=stage, predicate=lambda s: s.is_processable))
with self.progress.start('build', stage, item_total=len(resolved_sources)
) as ps:
if len(resolved_sources) == 0:
ps.update(message='No sources', state='skipped')
self.log('No processable sources, skipping build stage {}'.
format(stage))
return True
if not self.pre_build(force):
ps.update(message='Pre-build failed', state='skipped')
return False
if force:
self._reset_build(resolved_sources)
resolved_sources.reload()
e = [(stage, SourceSet(self, list(stage_sources))) for stage,
stage_sources in groupby(sorted(resolved_sources, key=
attrgetter('stage')), attrgetter('stage'))]
for stage, stage_sources in e:
stage_sources.reload()
for s in stage_sources:
s.state = self.STATES.WAITING
self.commit()
stage_sources.reload()
self.log('Processing {} sources, stage {} ; first 10: {}'.
format(len(stage_sources), stage, [x.name for x in
stage_sources.sources[:10]]))
self._run_events(TAG.BEFORE_BUILD, stage)
if self.multi:
try:
args = [(self.identity.vid, stage, source.vid, force) for
source in stage_sources]
pool = self.library.process_pool(limited_run=self.
limited_run)
r = pool.map_async(build_mp, args, 1)
completed_sources = r.get()
ps.add(
'Finished MP building {} sources. Starting MP coalescing'
.format(len(completed_sources)))
partition_names = [(self.identity.vid, k) for k, v in
self.collect_segment_partitions().items()]
r = pool.map_async(unify_mp, partition_names, 1)
completed_partitions = r.get()
ps.add('Finished MP coalescing {} partitions'.format(
len(completed_partitions)))
pool.close()
pool.join()
except KeyboardInterrupt:
self.log('Got keyboard interrrupt; terminating workers')
pool.terminate()
else:
for i, source in enumerate(stage_sources):
id_ = ps.add(message='Running source {}'.format(source.
name), source=source, item_count=i, state='running')
self.build_source(stage, source, ps, force=force)
ps.update(message='Finished processing source', state=
'done')
ps.get(id_).state = 'done'
self.progress.commit()
self.unify_partitions()
self._run_events(TAG.AFTER_BUILD, stage)
self.state = self.STATES.BUILT
self.commit()
self._run_events(TAG.AFTER_BUILD, 0)
self.close_session()
self.log('==== Done Building ====')
self.buildstate.commit()
self.state = self.STATES.BUILT
self.commit()
return True
| build | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, sources=None, tables=None, stage=None, force=False):\n from operator import attrgetter\n from itertools import groupby\n from .concurrent import build_mp, unify_mp\n from ambry.bundle.events import TAG\n self.log('==== Building ====')\n self.state = self.STATES.BUILDING\n self.commit()\n\n\n class SourceSet(object):\n \"\"\"Container for sources that can reload them after they get expired from the session\"\"\"\n\n def x(self, bundle, v):\n self.bundle = bundle\n self.sources = v\n self._s_vids = [s.vid for s in self.sources]\n\n def x(self):\n self.sources = [self.bundle.source(vid) for vid in self._s_vids]\n\n def x(self):\n for s in self.sources:\n yield s\n\n def x(self):\n return len(self._s_vids)\n self._run_events(TAG.BEFORE_BUILD, 0)\n resolved_sources = SourceSet(self, self._resolve_sources(sources,\n tables, stage=stage, predicate=lambda s: s.is_processable))\n with self.progress.start('build', stage, item_total=len(resolved_sources)\n ) as ps:\n if len(resolved_sources) == 0:\n ps.update(message='No sources', state='skipped')\n self.log('No processable sources, skipping build stage {}'.\n format(stage))\n return True\n if not self.pre_build(force):\n ps.update(message='Pre-build failed', state='skipped')\n return False\n if force:\n self._reset_build(resolved_sources)\n resolved_sources.reload()\n e = [(stage, SourceSet(self, list(stage_sources))) for stage,\n stage_sources in groupby(sorted(resolved_sources, key=\n attrgetter('stage')), attrgetter('stage'))]\n for stage, stage_sources in e:\n stage_sources.reload()\n for s in stage_sources:\n s.state = self.STATES.WAITING\n self.commit()\n stage_sources.reload()\n self.log('Processing {} sources, stage {} ; first 10: {}'.\n format(len(stage_sources), stage, [x.name for x in\n stage_sources.sources[:10]]))\n self._run_events(TAG.BEFORE_BUILD, stage)\n if self.multi:\n try:\n args = [(self.identity.vid, stage, source.vid, force) for\n source in stage_sources]\n pool = self.library.process_pool(limited_run=self.\n limited_run)\n r = pool.map_async(build_mp, args, 1)\n completed_sources = r.get()\n ps.add(\n 'Finished MP building {} sources. Starting MP coalescing'\n .format(len(completed_sources)))\n partition_names = [(self.identity.vid, k) for k, v in\n self.collect_segment_partitions().items()]\n r = pool.map_async(unify_mp, partition_names, 1)\n completed_partitions = r.get()\n ps.add('Finished MP coalescing {} partitions'.format(\n len(completed_partitions)))\n pool.close()\n pool.join()\n except KeyboardInterrupt:\n self.log('Got keyboard interrrupt; terminating workers')\n pool.terminate()\n else:\n for i, source in enumerate(stage_sources):\n id_ = ps.add(message='Running source {}'.format(source.\n name), source=source, item_count=i, state='running')\n self.build_source(stage, source, ps, force=force)\n ps.update(message='Finished processing source', state=\n 'done')\n ps.get(id_).state = 'done'\n self.progress.commit()\n self.unify_partitions()\n self._run_events(TAG.AFTER_BUILD, stage)\n self.state = self.STATES.BUILT\n self.commit()\n self._run_events(TAG.AFTER_BUILD, 0)\n self.close_session()\n self.log('==== Done Building ====')\n self.buildstate.commit()\n self.state = self.STATES.BUILT\n self.commit()\n return True\n",
"role": "user"
},
{
"content": "build",
"role": "assistant"
}
] |
def x(self, table, force=False):
sources = self._resolve_sources(None, [table])
for source in sources:
self.build_source(None, source, force=force)
self.unify_partitions()
| build_table | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, table, force=False):\n sources = self._resolve_sources(None, [table])\n for source in sources:\n self.build_source(None, source, force=force)\n self.unify_partitions()\n",
"role": "user"
},
{
"content": "build_table",
"role": "assistant"
}
] |
def x(self, stage, source, ps, force=False):
from ambry.bundle.process import call_interval
assert source.is_processable, source.name
if source.state == self.STATES.BUILT and not force:
ps.update(message='Source {} already built'.format(source.name),
state='skipped')
return
pl = self.pipeline(source, ps=ps)
source.state = self.STATES.BUILDING
self.log_pipeline(pl)
try:
source_name = source.name
s_vid = source.vid
ps.update(message='Running pipeline {}'.format(pl.name), s_vid=
s_vid, item_type='rows', item_count=0)
@call_interval(5)
def x(sink_pipe, rows):
n_records, rate = sink_pipe.report_progress()
if n_records > 0:
ps.update(message='Running pipeline {}: rate: {}'.format(pl
.name, rate), s_vid=s_vid, item_type='rows', item_count
=n_records)
pl.run(callback=run_progress_f)
for f in pl.final:
ps.update(message='Run final routine: {}'.format(f.__name__))
f(pl)
ps.update(message='Finished building source')
except:
self.log_pipeline(pl)
raise
self.commit()
try:
partitions = list(pl[ambry.etl.PartitionWriter].partitions)
ps.update(message='Finalizing segment partition', item_type=
'partitions', item_total=len(partitions), item_count=0)
for i, p in enumerate(partitions):
ps.update(message='Finalizing segment partition {}'.format(p.
name), item_count=i, p_vid=p.vid)
try:
p.finalize()
except AttributeError:
print(self.table(p.table_name))
raise
self.commit()
except IndexError:
self.error(
"Pipeline didn't have a PartitionWriters, won't try to finalize")
self.log_pipeline(pl)
source.state = self.STATES.BUILT
self.commit()
return source.name
| build_source | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, stage, source, ps, force=False):\n from ambry.bundle.process import call_interval\n assert source.is_processable, source.name\n if source.state == self.STATES.BUILT and not force:\n ps.update(message='Source {} already built'.format(source.name),\n state='skipped')\n return\n pl = self.pipeline(source, ps=ps)\n source.state = self.STATES.BUILDING\n self.log_pipeline(pl)\n try:\n source_name = source.name\n s_vid = source.vid\n ps.update(message='Running pipeline {}'.format(pl.name), s_vid=\n s_vid, item_type='rows', item_count=0)\n\n @call_interval(5)\n def x(sink_pipe, rows):\n n_records, rate = sink_pipe.report_progress()\n if n_records > 0:\n ps.update(message='Running pipeline {}: rate: {}'.format(pl\n .name, rate), s_vid=s_vid, item_type='rows', item_count\n =n_records)\n pl.run(callback=run_progress_f)\n for f in pl.final:\n ps.update(message='Run final routine: {}'.format(f.__name__))\n f(pl)\n ps.update(message='Finished building source')\n except:\n self.log_pipeline(pl)\n raise\n self.commit()\n try:\n partitions = list(pl[ambry.etl.PartitionWriter].partitions)\n ps.update(message='Finalizing segment partition', item_type=\n 'partitions', item_total=len(partitions), item_count=0)\n for i, p in enumerate(partitions):\n ps.update(message='Finalizing segment partition {}'.format(p.\n name), item_count=i, p_vid=p.vid)\n try:\n p.finalize()\n except AttributeError:\n print(self.table(p.table_name))\n raise\n self.commit()\n except IndexError:\n self.error(\n \"Pipeline didn't have a PartitionWriters, won't try to finalize\")\n self.log_pipeline(pl)\n source.state = self.STATES.BUILT\n self.commit()\n return source.name\n",
"role": "user"
},
{
"content": "build_source",
"role": "assistant"
}
] |
def x(self):
from collections import defaultdict
partitions = defaultdict(set)
for p in self.dataset.partitions:
if p.type == p.TYPE.SEGMENT:
name = p.identity.name
name.segment = None
partitions[name].add(p)
return partitions
| collect_segment_partitions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from collections import defaultdict\n partitions = defaultdict(set)\n for p in self.dataset.partitions:\n if p.type == p.TYPE.SEGMENT:\n name = p.identity.name\n name.segment = None\n partitions[name].add(p)\n return partitions\n",
"role": "user"
},
{
"content": "collect_segment_partitions",
"role": "assistant"
}
] |
def x(self):
partitions = self.collect_segment_partitions()
with self.progress.start('coalesce', 0, message=
'Coalescing partition segments') as ps:
for name, segments in iteritems(partitions):
ps.add(item_type='partitions', item_count=len(segments),
message='Colescing partition {}'.format(name))
self.unify_partition(name, segments, ps)
| unify_partitions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n partitions = self.collect_segment_partitions()\n with self.progress.start('coalesce', 0, message=\n 'Coalescing partition segments') as ps:\n for name, segments in iteritems(partitions):\n ps.add(item_type='partitions', item_count=len(segments),\n message='Colescing partition {}'.format(name))\n self.unify_partition(name, segments, ps)\n",
"role": "user"
},
{
"content": "unify_partitions",
"role": "assistant"
}
] |
def x(self, **kwargs):
import inspect
import dateutil.parser
import datetime
import random
from functools import partial
from ambry.valuetype.types import parse_date, parse_time, parse_datetime
import ambry.valuetype.types
import ambry.valuetype.exceptions
import ambry.valuetype.test
import ambry.valuetype
def x(f, frm):
try:
try:
f.ambry_from = frm
except AttributeError:
f.im_func.ambry_from = frm
except (TypeError, AttributeError):
pass
return f
test_env = dict(parse_date=parse_date, parse_time=parse_time,
parse_datetime=parse_datetime, partial=partial, bundle=self)
test_env.update(kwargs)
test_env.update(dateutil.parser.__dict__)
test_env.update(datetime.__dict__)
test_env.update(random.__dict__)
test_env.update(ambry.valuetype.core.__dict__)
test_env.update(ambry.valuetype.types.__dict__)
test_env.update(ambry.valuetype.exceptions.__dict__)
test_env.update(ambry.valuetype.test.__dict__)
test_env.update(ambry.valuetype.__dict__)
localvars = {}
for f_name, func in test_env.items():
if not isinstance(func, (str, tuple)):
localvars[f_name] = set_from(func, 'env')
localvars['randint'] = lambda a, b: random.randint(a, b)
if self != Bundle:
base = set(inspect.getmembers(Bundle, predicate=inspect.isfunction))
mine = set(inspect.getmembers(self.__class__, predicate=inspect.
isfunction))
localvars.update({f_name: set_from(func, 'bundle') for f_name, func in
mine - base})
base = set(inspect.getmembers(Bundle, predicate=inspect.ismethod))
mine = set(inspect.getmembers(self.__class__, predicate=inspect.
ismethod))
localvars.update({f_name: set_from(func.__get__(self), 'bundle') for
f_name, func in mine - base})
module_entries = inspect.getmembers(sys.modules['ambry.build'],
predicate=inspect.isfunction)
localvars.update({f_name: set_from(func, 'module') for f_name, func in
module_entries})
return localvars
| exec_context | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n import inspect\n import dateutil.parser\n import datetime\n import random\n from functools import partial\n from ambry.valuetype.types import parse_date, parse_time, parse_datetime\n import ambry.valuetype.types\n import ambry.valuetype.exceptions\n import ambry.valuetype.test\n import ambry.valuetype\n\n def x(f, frm):\n try:\n try:\n f.ambry_from = frm\n except AttributeError:\n f.im_func.ambry_from = frm\n except (TypeError, AttributeError):\n pass\n return f\n test_env = dict(parse_date=parse_date, parse_time=parse_time,\n parse_datetime=parse_datetime, partial=partial, bundle=self)\n test_env.update(kwargs)\n test_env.update(dateutil.parser.__dict__)\n test_env.update(datetime.__dict__)\n test_env.update(random.__dict__)\n test_env.update(ambry.valuetype.core.__dict__)\n test_env.update(ambry.valuetype.types.__dict__)\n test_env.update(ambry.valuetype.exceptions.__dict__)\n test_env.update(ambry.valuetype.test.__dict__)\n test_env.update(ambry.valuetype.__dict__)\n localvars = {}\n for f_name, func in test_env.items():\n if not isinstance(func, (str, tuple)):\n localvars[f_name] = set_from(func, 'env')\n localvars['randint'] = lambda a, b: random.randint(a, b)\n if self != Bundle:\n base = set(inspect.getmembers(Bundle, predicate=inspect.isfunction))\n mine = set(inspect.getmembers(self.__class__, predicate=inspect.\n isfunction))\n localvars.update({f_name: set_from(func, 'bundle') for f_name, func in\n mine - base})\n base = set(inspect.getmembers(Bundle, predicate=inspect.ismethod))\n mine = set(inspect.getmembers(self.__class__, predicate=inspect.\n ismethod))\n localvars.update({f_name: set_from(func.__get__(self), 'bundle') for\n f_name, func in mine - base})\n module_entries = inspect.getmembers(sys.modules['ambry.build'],\n predicate=inspect.isfunction)\n localvars.update({f_name: set_from(func, 'module') for f_name, func in\n module_entries})\n return localvars\n",
"role": "user"
},
{
"content": "exec_context",
"role": "assistant"
}
] |
def x(self):
from ambry.util.datestimes import expand_to_years
years = set()
if self.metadata.about.time:
for year in expand_to_years(self.metadata.about.time):
years.add(year)
if self.identity.btime:
for year in expand_to_years(self.identity.btime):
years.add(year)
for p in self.partitions:
years |= set(p.time_coverage)
| post_build_time_coverage | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry.util.datestimes import expand_to_years\n years = set()\n if self.metadata.about.time:\n for year in expand_to_years(self.metadata.about.time):\n years.add(year)\n if self.identity.btime:\n for year in expand_to_years(self.identity.btime):\n years.add(year)\n for p in self.partitions:\n years |= set(p.time_coverage)\n",
"role": "user"
},
{
"content": "post_build_time_coverage",
"role": "assistant"
}
] |
def x(self):
spaces = set()
grains = set()
def x(term):
places = list(self.library.search.search_identifiers(term))
if not places:
raise BuildError(
"Failed to find space identifier '{}' in full text identifier search"
.format(term))
return places[0].vid
if self.metadata.about.space:
spaces.add(resolve(self.metadata.about.space))
if self.metadata.about.grain:
grains.add(self.metadata.about.grain)
if self.identity.bspace:
spaces.add(resolve(self.identity.bspace))
for p in self.partitions.all:
if 'geo_coverage' in p.record.data:
for space in p.record.data['geo_coverage']:
spaces.add(space)
if 'geo_grain' in p.record.data:
for grain in p.record.data['geo_grain']:
grains.add(grain)
def x(g):
try:
c = GVid.get_class(g)
return b(c().summarize())
except NotASummaryName:
return g
self.metadata.coverage.geo = sorted(spaces)
self.metadata.coverage.grain = sorted(conv_grain(g) for g in grains)
self.metadata.write_to_dir()
| post_build_geo_coverage | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n spaces = set()\n grains = set()\n\n def x(term):\n places = list(self.library.search.search_identifiers(term))\n if not places:\n raise BuildError(\n \"Failed to find space identifier '{}' in full text identifier search\"\n .format(term))\n return places[0].vid\n if self.metadata.about.space:\n spaces.add(resolve(self.metadata.about.space))\n if self.metadata.about.grain:\n grains.add(self.metadata.about.grain)\n if self.identity.bspace:\n spaces.add(resolve(self.identity.bspace))\n for p in self.partitions.all:\n if 'geo_coverage' in p.record.data:\n for space in p.record.data['geo_coverage']:\n spaces.add(space)\n if 'geo_grain' in p.record.data:\n for grain in p.record.data['geo_grain']:\n grains.add(grain)\n\n def x(g):\n try:\n c = GVid.get_class(g)\n return b(c().summarize())\n except NotASummaryName:\n return g\n self.metadata.coverage.geo = sorted(spaces)\n self.metadata.coverage.grain = sorted(conv_grain(g) for g in grains)\n self.metadata.write_to_dir()\n",
"role": "user"
},
{
"content": "post_build_geo_coverage",
"role": "assistant"
}
] |
def x(self):
return (self.state == self.STATES.FINALIZED or self.state == self.
STATES.INSTALLED)
| is_finalized | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return (self.state == self.STATES.FINALIZED or self.state == self.\n STATES.INSTALLED)\n",
"role": "user"
},
{
"content": "is_finalized",
"role": "assistant"
}
] |
def x(self, tag, stage=None):
self._run_event_methods(tag, stage)
self._run_tests(tag, stage)
| run_events | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tag, stage=None):\n self._run_event_methods(tag, stage)\n self._run_tests(tag, stage)\n",
"role": "user"
},
{
"content": "run_events",
"role": "assistant"
}
] |
def x(self, tag, stage=None):
import inspect
from ambry.bundle.events import _runable_for_event
funcs = []
for func_name, f in inspect.getmembers(self, predicate=inspect.ismethod):
if _runable_for_event(f, tag, stage):
funcs.append(f)
for func in funcs:
func()
| run_event_methods | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tag, stage=None):\n import inspect\n from ambry.bundle.events import _runable_for_event\n funcs = []\n for func_name, f in inspect.getmembers(self, predicate=inspect.ismethod):\n if _runable_for_event(f, tag, stage):\n funcs.append(f)\n for func in funcs:\n func()\n",
"role": "user"
},
{
"content": "run_event_methods",
"role": "assistant"
}
] |
def x(self):
r = self.library.resolve(self.identity.vid)
return r is not None
| is_installed | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n r = self.library.resolve(self.identity.vid)\n return r is not None\n",
"role": "user"
},
{
"content": "is_installed",
"role": "assistant"
}
] |
def x(prop):
if isinstance(prop, QueryableAttribute):
prop = prop.property
assert isinstance(prop, (Column, ColumnProperty, RelationshipProperty))
_included.add(prop)
| include | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(prop):\n if isinstance(prop, QueryableAttribute):\n prop = prop.property\n assert isinstance(prop, (Column, ColumnProperty, RelationshipProperty))\n _included.add(prop)\n",
"role": "user"
},
{
"content": "include",
"role": "assistant"
}
] |
def x(prop):
if isinstance(prop, QueryableAttribute):
prop = prop.property
assert isinstance(prop, (Column, ColumnProperty, RelationshipProperty))
_excluded.add(prop)
if isinstance(prop, RelationshipProperty):
for local in prop.local_columns:
_excluded.add(local)
| exclude | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(prop):\n if isinstance(prop, QueryableAttribute):\n prop = prop.property\n assert isinstance(prop, (Column, ColumnProperty, RelationshipProperty))\n _excluded.add(prop)\n if isinstance(prop, RelationshipProperty):\n for local in prop.local_columns:\n _excluded.add(local)\n",
"role": "user"
},
{
"content": "exclude",
"role": "assistant"
}
] |
def x(source, model, cache=None):
if source is None:
return None
if cache and source in cache:
return cache[source]
db = object_session(source)
ident = identity_key(instance=source)[1]
assert ident is not None
return db.query(model).get(ident)
| reflect | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(source, model, cache=None):\n if source is None:\n return None\n if cache and source in cache:\n return cache[source]\n db = object_session(source)\n ident = identity_key(instance=source)[1]\n assert ident is not None\n return db.query(model).get(ident)\n",
"role": "user"
},
{
"content": "reflect",
"role": "assistant"
}
] |
def x(source, target, cache=None):
target_manager = manager_of_class(type(target))
column_attrs = set()
relationship_attrs = set()
relationship_columns = set()
for attr in manager_of_class(type(source)).attributes:
if attr.key not in target_manager:
continue
target_attr = target_manager[attr.key]
if isinstance(attr.property, ColumnProperty):
assert isinstance(target_attr.property, ColumnProperty)
column_attrs.add(attr)
elif isinstance(attr.property, RelationshipProperty):
assert isinstance(target_attr.property, RelationshipProperty)
relationship_attrs.add(attr)
if attr.property.direction is MANYTOONE:
relationship_columns.update(attr.property.local_columns)
for attr in column_attrs:
if _column_property_in_registry(attr.property, _excluded):
continue
elif not _column_property_in_registry(attr.property, _included
) and all(column in relationship_columns for column in attr.
property.columns):
continue
setattr(target, attr.key, getattr(source, attr.key))
for attr in relationship_attrs:
target_attr_model = target_manager[attr.key].property.argument
if not is_relation_replicatable(attr):
continue
replicate_relation(source, target, attr, target_manager[attr.key],
cache=cache)
| replicate_attributes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(source, target, cache=None):\n target_manager = manager_of_class(type(target))\n column_attrs = set()\n relationship_attrs = set()\n relationship_columns = set()\n for attr in manager_of_class(type(source)).attributes:\n if attr.key not in target_manager:\n continue\n target_attr = target_manager[attr.key]\n if isinstance(attr.property, ColumnProperty):\n assert isinstance(target_attr.property, ColumnProperty)\n column_attrs.add(attr)\n elif isinstance(attr.property, RelationshipProperty):\n assert isinstance(target_attr.property, RelationshipProperty)\n relationship_attrs.add(attr)\n if attr.property.direction is MANYTOONE:\n relationship_columns.update(attr.property.local_columns)\n for attr in column_attrs:\n if _column_property_in_registry(attr.property, _excluded):\n continue\n elif not _column_property_in_registry(attr.property, _included\n ) and all(column in relationship_columns for column in attr.\n property.columns):\n continue\n setattr(target, attr.key, getattr(source, attr.key))\n for attr in relationship_attrs:\n target_attr_model = target_manager[attr.key].property.argument\n if not is_relation_replicatable(attr):\n continue\n replicate_relation(source, target, attr, target_manager[attr.key],\n cache=cache)\n",
"role": "user"
},
{
"content": "replicate_attributes",
"role": "assistant"
}
] |
def x(source, model, cache=None):
if source is None:
return None
if cache is None:
cache = {}
elif source in cache:
return cache[source]
db = object_session(source)
cls, ident = identity_key(instance=source)
target = db.query(model).get(ident)
if target is None:
target = model()
cache[source] = target
try:
replicate_attributes(source, target, cache=cache)
except _PrimaryKeyIsNull:
return None
else:
return target
| replicate_no_merge | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(source, model, cache=None):\n if source is None:\n return None\n if cache is None:\n cache = {}\n elif source in cache:\n return cache[source]\n db = object_session(source)\n cls, ident = identity_key(instance=source)\n target = db.query(model).get(ident)\n if target is None:\n target = model()\n cache[source] = target\n try:\n replicate_attributes(source, target, cache=cache)\n except _PrimaryKeyIsNull:\n return None\n else:\n return target\n",
"role": "user"
},
{
"content": "replicate_no_merge",
"role": "assistant"
}
] |
def x(source, model, cache=None):
target = replicate_no_merge(source, model, cache=cache)
if target is not None:
db = object_session(source)
target = db.merge(target)
return target
| replicate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(source, model, cache=None):\n target = replicate_no_merge(source, model, cache=cache)\n if target is not None:\n db = object_session(source)\n target = db.merge(target)\n return target\n",
"role": "user"
},
{
"content": "replicate",
"role": "assistant"
}
] |
def x(sources, model, cache=None):
targets = [replicate_no_merge(source, model, cache=cache) for source in
sources]
return [target for target in targets if target is not None]
| replicate_filter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sources, model, cache=None):\n targets = [replicate_no_merge(source, model, cache=cache) for source in\n sources]\n return [target for target in targets if target is not None]\n",
"role": "user"
},
{
"content": "replicate_filter",
"role": "assistant"
}
] |
def x(sources, model, cache=None):
targets = [reflect(source, model, cache=cache) for source in sources]
return [target for target in targets if target is not None]
| reflect_filter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sources, model, cache=None):\n targets = [reflect(source, model, cache=cache) for source in sources]\n return [target for target in targets if target is not None]\n",
"role": "user"
},
{
"content": "reflect_filter",
"role": "assistant"
}
] |
def x(self):
from ambry.valuetype import resolve_value_type
if self.valuetype:
return resolve_value_type(self.valuetype)
else:
return resolve_value_type(self.datatype)
| valuetype_class | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry.valuetype import resolve_value_type\n if self.valuetype:\n return resolve_value_type(self.valuetype)\n else:\n return resolve_value_type(self.datatype)\n",
"role": "user"
},
{
"content": "valuetype_class",
"role": "assistant"
}
] |
def x(self):
from ambry.valuetype import resolve_value_type
if self.valuetype and resolve_value_type(self.valuetype):
return resolve_value_type(self.valuetype)._pythontype
elif self.datatype:
try:
return self.types[self.datatype][1]
except KeyError:
return resolve_value_type(self.datatype)._pythontype
else:
from ambry.exc import ConfigurationError
raise ConfigurationError(
"Can't get python_type: neither datatype of valuetype is defined")
| python_type | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry.valuetype import resolve_value_type\n if self.valuetype and resolve_value_type(self.valuetype):\n return resolve_value_type(self.valuetype)._pythontype\n elif self.datatype:\n try:\n return self.types[self.datatype][1]\n except KeyError:\n return resolve_value_type(self.datatype)._pythontype\n else:\n from ambry.exc import ConfigurationError\n raise ConfigurationError(\n \"Can't get python_type: neither datatype of valuetype is defined\")\n",
"role": "user"
},
{
"content": "python_type",
"role": "assistant"
}
] |
def x(self):
from ambry.valuetype.core import ROLE
if not self.valuetype_class:
return ''
role = self.valuetype_class.role
if role == ROLE.UNKNOWN:
vt_code = self.valuetype_class.vt_code
if len(vt_code) == 1 or vt_code[1] == '/':
return vt_code[0]
else:
return ''
return role
| role | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry.valuetype.core import ROLE\n if not self.valuetype_class:\n return ''\n role = self.valuetype_class.role\n if role == ROLE.UNKNOWN:\n vt_code = self.valuetype_class.vt_code\n if len(vt_code) == 1 or vt_code[1] == '/':\n return vt_code[0]\n else:\n return ''\n return role\n",
"role": "user"
},
{
"content": "role",
"role": "assistant"
}
] |
def x(self):
from ambry.valuetype.core import ROLE
return self.role == ROLE.DIMENSION
| is_dimension | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry.valuetype.core import ROLE\n return self.role == ROLE.DIMENSION\n",
"role": "user"
},
{
"content": "is_dimension",
"role": "assistant"
}
] |
def x(self):
from ambry.valuetype.core import ROLE
return self.role == ROLE.MEASURE
| is_measure | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry.valuetype.core import ROLE\n return self.role == ROLE.MEASURE\n",
"role": "user"
},
{
"content": "is_measure",
"role": "assistant"
}
] |
def x(self):
from ambry.valuetype.core import ROLE
return self.role == ROLE.LABEL
| is_label | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry.valuetype.core import ROLE\n return self.role == ROLE.LABEL\n",
"role": "user"
},
{
"content": "is_label",
"role": "assistant"
}
] |
def x(self):
from ambry.valuetype.core import ROLE
return self.role == ROLE.ERROR
| is_error | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry.valuetype.core import ROLE\n return self.role == ROLE.ERROR\n",
"role": "user"
},
{
"content": "is_error",
"role": "assistant"
}
] |
def x(self):
for c in self.table.columns:
if c.parent == self.name and not c.valuetype_class.is_label():
yield c
| children | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n for c in self.table.columns:\n if c.parent == self.name and not c.valuetype_class.is_label():\n yield c\n",
"role": "user"
},
{
"content": "children",
"role": "assistant"
}
] |
def x(self):
if self.valuetype_class.is_label():
return self
for c in self.table.columns:
if c.parent == self.name and c.valuetype_class.is_label():
return c
return None
| label | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.valuetype_class.is_label():\n return self\n for c in self.table.columns:\n if c.parent == self.name and c.valuetype_class.is_label():\n return c\n return None\n",
"role": "user"
},
{
"content": "label",
"role": "assistant"
}
] |
def x(self):
if self.valuetype_class.is_geoid():
return self
for c in self.table.columns:
if c.parent == self.name and c.valuetype_class.is_geoid():
return c
| geoid | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.valuetype_class.is_geoid():\n return self\n for c in self.table.columns:\n if c.parent == self.name and c.valuetype_class.is_geoid():\n return c\n",
"role": "user"
},
{
"content": "geoid",
"role": "assistant"
}
] |
def x(self, v):
if self.type_is_time():
dt = dateutil.parser.parse(v)
if self.datatype == Column.DATATYPE_TIME:
dt = dt.time()
if not isinstance(dt, self.python_type):
raise TypeError('{} was parsed to {}, expected {}'.format(v,
type(dt), self.python_type))
return dt
else:
return self.python_type(v)
| python_cast | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, v):\n if self.type_is_time():\n dt = dateutil.parser.parse(v)\n if self.datatype == Column.DATATYPE_TIME:\n dt = dt.time()\n if not isinstance(dt, self.python_type):\n raise TypeError('{} was parsed to {}, expected {}'.format(v,\n type(dt), self.python_type))\n return dt\n else:\n return self.python_type(v)\n",
"role": "user"
},
{
"content": "python_cast",
"role": "assistant"
}
] |
def x(cls, dtype):
m = {'int64': cls.DATATYPE_INTEGER64, 'float64': cls.DATATYPE_FLOAT,
'object': cls.DATATYPE_TEXT}
t = m.get(dtype.name, None)
if not t:
raise TypeError("Failed to convert numpy type: '{}' ".format(dtype.
name))
return t
| convert_numpy_type | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, dtype):\n m = {'int64': cls.DATATYPE_INTEGER64, 'float64': cls.DATATYPE_FLOAT,\n 'object': cls.DATATYPE_TEXT}\n t = m.get(dtype.name, None)\n if not t:\n raise TypeError(\"Failed to convert numpy type: '{}' \".format(dtype.\n name))\n return t\n",
"role": "user"
},
{
"content": "convert_numpy_type",
"role": "assistant"
}
] |
def x(self):
d = {p.key: getattr(self, p.key) for p in self.__mapper__.attrs if p.
key not in ('table', 'stats', '_codes', 'data')}
if not d:
raise Exception(self.__dict__)
d['schema_type'] = self.schema_type
if self.data:
for k, v in six.iteritems(self.data):
if k not in d and k not in ('table', 'stats', '_codes', 'data'):
d[k] = v
return d
| dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n d = {p.key: getattr(self, p.key) for p in self.__mapper__.attrs if p.\n key not in ('table', 'stats', '_codes', 'data')}\n if not d:\n raise Exception(self.__dict__)\n d['schema_type'] = self.schema_type\n if self.data:\n for k, v in six.iteritems(self.data):\n if k not in d and k not in ('table', 'stats', '_codes', 'data'):\n d[k] = v\n return d\n",
"role": "user"
},
{
"content": "dict",
"role": "assistant"
}
] |
def x(self):
return {k: v for k, v in six.iteritems(self.dict) if v and k != '_codes'}
| nonull_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return {k: v for k, v in six.iteritems(self.dict) if v and k != '_codes'}\n",
"role": "user"
},
{
"content": "nonull_dict",
"role": "assistant"
}
] |
def x(name):
import re
try:
return re.sub('_+', '_', re.sub('[^\\w_]', '_', name).lower()).rstrip(
'_')
except TypeError:
raise TypeError('Trying to mangle name with invalid type of: ' +
str(type(name)))
| mangle_name | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(name):\n import re\n try:\n return re.sub('_+', '_', re.sub('[^\\\\w_]', '_', name).lower()).rstrip(\n '_')\n except TypeError:\n raise TypeError('Trying to mangle name with invalid type of: ' +\n str(type(name)))\n",
"role": "user"
},
{
"content": "mangle_name",
"role": "assistant"
}
] |
def x(self):
return {c.value: (c.ikey if c.ikey else c.key) for c in self.codes}
| reverse_code_map | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return {c.value: (c.ikey if c.ikey else c.key) for c in self.codes}\n",
"role": "user"
},
{
"content": "reverse_code_map",
"role": "assistant"
}
] |
def x(self):
segments = self._expand_transform(self.transform)
if segments:
segments[0]['datatype'] = self.valuetype_class
for s in segments:
s['column'] = self
else:
segments = [self.make_xform_seg(datatype=self.valuetype_class,
column=self)]
return segments
| expanded_transform | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n segments = self._expand_transform(self.transform)\n if segments:\n segments[0]['datatype'] = self.valuetype_class\n for s in segments:\n s['column'] = self\n else:\n segments = [self.make_xform_seg(datatype=self.valuetype_class,\n column=self)]\n return segments\n",
"role": "user"
},
{
"content": "expanded_transform",
"role": "assistant"
}
] |
def x(mapper, conn, target):
if target.sequence_id is None:
from ambry.orm.exc import DatabaseError
raise DatabaseError('Must have sequence_id before insertion')
assert (target.name == 'id') == (target.sequence_id == 1), (target.name,
target.sequence_id)
Column.before_update(mapper, conn, target)
| before_insert | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(mapper, conn, target):\n if target.sequence_id is None:\n from ambry.orm.exc import DatabaseError\n raise DatabaseError('Must have sequence_id before insertion')\n assert (target.name == 'id') == (target.sequence_id == 1), (target.name,\n target.sequence_id)\n Column.before_update(mapper, conn, target)\n",
"role": "user"
},
{
"content": "before_insert",
"role": "assistant"
}
] |
def x(mapper, conn, target):
assert target.datatype or target.valuetype
target.name = Column.mangle_name(target.name)
Column.update_number(target)
| before_update | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(mapper, conn, target):\n assert target.datatype or target.valuetype\n target.name = Column.mangle_name(target.name)\n Column.update_number(target)\n",
"role": "user"
},
{
"content": "before_update",
"role": "assistant"
}
] |
def x(processProtocol, executable, args=(), env={}, path=None, uid=None,
gid=None, usePTY=0, packages=()):
env = env.copy()
pythonpath = []
for pkg in packages:
p = os.path.split(imp.find_module(pkg)[1])[0]
if p.startswith(os.path.join(sys.prefix, 'lib')):
continue
pythonpath.append(p)
pythonpath = list(set(pythonpath))
pythonpath.extend(env.get('PYTHONPATH', '').split(os.pathsep))
env['PYTHONPATH'] = os.pathsep.join(pythonpath)
return reactor.spawnProcess(processProtocol, executable, args, env,
path, uid, gid, usePTY)
| spawnProcess | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(processProtocol, executable, args=(), env={}, path=None, uid=None,\n gid=None, usePTY=0, packages=()):\n env = env.copy()\n pythonpath = []\n for pkg in packages:\n p = os.path.split(imp.find_module(pkg)[1])[0]\n if p.startswith(os.path.join(sys.prefix, 'lib')):\n continue\n pythonpath.append(p)\n pythonpath = list(set(pythonpath))\n pythonpath.extend(env.get('PYTHONPATH', '').split(os.pathsep))\n env['PYTHONPATH'] = os.pathsep.join(pythonpath)\n return reactor.spawnProcess(processProtocol, executable, args, env,\n path, uid, gid, usePTY)\n",
"role": "user"
},
{
"content": "spawnProcess",
"role": "assistant"
}
] |
def x(processProtocol, args=(), env={}, path=None, uid=None, gid=None,
usePTY=0, packages=()):
return spawnProcess(processProtocol, sys.executable, args, env, path,
uid, gid, usePTY, packages)
| spawnPythonProcess | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(processProtocol, args=(), env={}, path=None, uid=None, gid=None,\n usePTY=0, packages=()):\n return spawnProcess(processProtocol, sys.executable, args, env, path,\n uid, gid, usePTY, packages)\n",
"role": "user"
},
{
"content": "spawnPythonProcess",
"role": "assistant"
}
] |
def x(f, tag, stage):
if not hasattr(f, '__ambry_event__'):
return False
f_tag, f_stage = f.__ambry_event__
if stage is None:
stage = 0
if tag != f_tag or stage != f_stage:
return False
return True
| runable_for_event | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(f, tag, stage):\n if not hasattr(f, '__ambry_event__'):\n return False\n f_tag, f_stage = f.__ambry_event__\n if stage is None:\n stage = 0\n if tag != f_tag or stage != f_stage:\n return False\n return True\n",
"role": "user"
},
{
"content": "runable_for_event",
"role": "assistant"
}
] |
def x(import_path, prefix=None, ld=dict()):
if prefix and not import_path.startswith(prefix):
import_path = '.'.join([prefix, import_path])
log.debug('attempting to load a python object from an import path',
extra=dict(import_path=import_path, **ld))
try:
mod = importlib.import_module(import_path)
return mod
except:
pass
try:
path, obj_name = import_path.rsplit('.', 1)
except ValueError:
log_raise(
'import path needs at least 1 period in your import path. An example import path is something like: module.obj'
, dict(import_path=import_path, **ld), InvalidImportPath)
try:
mod = importlib.import_module(path)
except ImportError:
newpath = path.replace(prefix, '', 1).lstrip('.')
log.debug('Could not load import path. Trying a different one',
extra=dict(oldpath=path, newpath=newpath))
path = newpath
mod = importlib.import_module(path)
try:
obj = getattr(mod, obj_name)
except AttributeError:
log_raise(
'object does not exist in given module. Your import path is not properly defined because the given `obj_name` does not exist'
, dict(import_path=path, obj_name=obj_name, **ld),
InvalidImportPath)
return obj
| load_obj_from_path | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(import_path, prefix=None, ld=dict()):\n if prefix and not import_path.startswith(prefix):\n import_path = '.'.join([prefix, import_path])\n log.debug('attempting to load a python object from an import path',\n extra=dict(import_path=import_path, **ld))\n try:\n mod = importlib.import_module(import_path)\n return mod\n except:\n pass\n try:\n path, obj_name = import_path.rsplit('.', 1)\n except ValueError:\n log_raise(\n 'import path needs at least 1 period in your import path. An example import path is something like: module.obj'\n , dict(import_path=import_path, **ld), InvalidImportPath)\n try:\n mod = importlib.import_module(path)\n except ImportError:\n newpath = path.replace(prefix, '', 1).lstrip('.')\n log.debug('Could not load import path. Trying a different one',\n extra=dict(oldpath=path, newpath=newpath))\n path = newpath\n mod = importlib.import_module(path)\n try:\n obj = getattr(mod, obj_name)\n except AttributeError:\n log_raise(\n 'object does not exist in given module. Your import path is not properly defined because the given `obj_name` does not exist'\n , dict(import_path=path, obj_name=obj_name, **ld),\n InvalidImportPath)\n return obj\n",
"role": "user"
},
{
"content": "load_obj_from_path",
"role": "assistant"
}
] |
def x(self, artifact_quality):
allowed_values = ['NEW', 'VERIFIED', 'TESTED', 'DEPRECATED',
'BLACKLISTED', 'DELETED', 'TEMPORARY']
if artifact_quality not in allowed_values:
raise ValueError(
'Invalid value for `artifact_quality` ({0}), must be one of {1}'
.format(artifact_quality, allowed_values))
self._artifact_quality = artifact_quality
| artifact_quality | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, artifact_quality):\n allowed_values = ['NEW', 'VERIFIED', 'TESTED', 'DEPRECATED',\n 'BLACKLISTED', 'DELETED', 'TEMPORARY']\n if artifact_quality not in allowed_values:\n raise ValueError(\n 'Invalid value for `artifact_quality` ({0}), must be one of {1}'\n .format(artifact_quality, allowed_values))\n self._artifact_quality = artifact_quality\n",
"role": "user"
},
{
"content": "artifact_quality",
"role": "assistant"
}
] |
def x(connection, table):
SQL_TO_PYTHON_TYPES = {'INT': int, 'INTEGER': int, 'TINYINT': int,
'SMALLINT': int, 'MEDIUMINT': int, 'BIGINT': int,
'UNSIGNED BIG INT': int, 'INT': int, 'INT8': int, 'NUMERIC': float,
'REAL': float, 'FLOAT': float, 'DOUBLE': float, 'BOOLEAN': bool,
'CHARACTER': str, 'VARCHAR': str, 'TEXT': str}
query = "PRAGMA table_info('{}');"
result = connection.execute(query.format(table))
ret = []
for row in result:
position = row[0] + 1
name = row[1]
datatype = row[2]
try:
datatype = SQL_TO_PYTHON_TYPES[datatype]
except KeyError:
raise Exception(
'Do not know how to convert {} sql datatype to python data type.'
.format(datatype))
ret.append((name, datatype, position))
return ret
| get_sqlite_columns | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(connection, table):\n SQL_TO_PYTHON_TYPES = {'INT': int, 'INTEGER': int, 'TINYINT': int,\n 'SMALLINT': int, 'MEDIUMINT': int, 'BIGINT': int,\n 'UNSIGNED BIG INT': int, 'INT': int, 'INT8': int, 'NUMERIC': float,\n 'REAL': float, 'FLOAT': float, 'DOUBLE': float, 'BOOLEAN': bool,\n 'CHARACTER': str, 'VARCHAR': str, 'TEXT': str}\n query = \"PRAGMA table_info('{}');\"\n result = connection.execute(query.format(table))\n ret = []\n for row in result:\n position = row[0] + 1\n name = row[1]\n datatype = row[2]\n try:\n datatype = SQL_TO_PYTHON_TYPES[datatype]\n except KeyError:\n raise Exception(\n 'Do not know how to convert {} sql datatype to python data type.'\n .format(datatype))\n ret.append((name, datatype, position))\n return ret\n",
"role": "user"
},
{
"content": "get_sqlite_columns",
"role": "assistant"
}
] |
def x(self):
if self.urltype != 'partition':
return None
return self._bundle.library.partition(self.url)
| partition | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.urltype != 'partition':\n return None\n return self._bundle.library.partition(self.url)\n",
"role": "user"
},
{
"content": "partition",
"role": "assistant"
}
] |
def x(self):
from ambry_sources import MPRowsFile
if self._datafile is None:
if self.urltype == 'partition':
self._datafile = self.partition.datafile
else:
self._datafile = MPRowsFile(self._bundle.build_ingest_fs, self.name
)
return self._datafile
| datafile | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry_sources import MPRowsFile\n if self._datafile is None:\n if self.urltype == 'partition':\n self._datafile = self.partition.datafile\n else:\n self._datafile = MPRowsFile(self._bundle.build_ingest_fs, self.name\n )\n return self._datafile\n",
"role": "user"
},
{
"content": "datafile",
"role": "assistant"
}
] |
def x(self):
from ambry_sources.sources import SourceSpec
d = self.dict
d['url'] = self.url
return SourceSpec(**d)
| spec | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry_sources.sources import SourceSpec\n d = self.dict\n d['url'] = self.url\n return SourceSpec(**d)\n",
"role": "user"
},
{
"content": "spec",
"role": "assistant"
}
] |
def x(self):
from ambry.util import parse_url_to_dict
d = parse_url_to_dict(self.url)
return self._bundle.library.account(d['netloc'])
| account | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry.util import parse_url_to_dict\n d = parse_url_to_dict(self.url)\n return self._bundle.library.account(d['netloc'])\n",
"role": "user"
},
{
"content": "account",
"role": "assistant"
}
] |
def x(self, unknown_type='str'):
from ambry_sources.intuit import TypeIntuiter
st = self.source_table
if self.reftype == 'partition':
for c in self.partition.table.columns:
st.add_column(c.sequence_id, source_header=c.name, dest_header=
c.name, datatype=c.datatype, description=c.description)
elif self.datafile.exists:
with self.datafile.reader as r:
names = set()
for col in r.columns:
name = col['name']
if name in names:
name = name + '_' + str(col['pos'])
names.add(name)
c = st.column(name)
dt = col['resolved_type'] if col['resolved_type'
] != 'unknown' else unknown_type
if c:
c.datatype = TypeIntuiter.promote_type(c.datatype, col[
'resolved_type'])
else:
c = st.add_column(col['pos'], source_header=name,
dest_header=name, datatype=col['resolved_type'],
description=col['description'], has_codes=col[
'has_codes'])
| update_table | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, unknown_type='str'):\n from ambry_sources.intuit import TypeIntuiter\n st = self.source_table\n if self.reftype == 'partition':\n for c in self.partition.table.columns:\n st.add_column(c.sequence_id, source_header=c.name, dest_header=\n c.name, datatype=c.datatype, description=c.description)\n elif self.datafile.exists:\n with self.datafile.reader as r:\n names = set()\n for col in r.columns:\n name = col['name']\n if name in names:\n name = name + '_' + str(col['pos'])\n names.add(name)\n c = st.column(name)\n dt = col['resolved_type'] if col['resolved_type'\n ] != 'unknown' else unknown_type\n if c:\n c.datatype = TypeIntuiter.promote_type(c.datatype, col[\n 'resolved_type'])\n else:\n c = st.add_column(col['pos'], source_header=name,\n dest_header=name, datatype=col['resolved_type'],\n description=col['description'], has_codes=col[\n 'has_codes'])\n",
"role": "user"
},
{
"content": "update_table",
"role": "assistant"
}
] |
def x(self):
if self.datafile.exists:
with self.datafile.reader as r:
self.header_lines = r.info['header_rows']
self.comment_lines = r.info['comment_rows']
self.start_line = r.info['data_start_row']
self.end_line = r.info['data_end_row']
| update_spec | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.datafile.exists:\n with self.datafile.reader as r:\n self.header_lines = r.info['header_rows']\n self.comment_lines = r.info['comment_rows']\n self.start_line = r.info['data_start_row']\n self.end_line = r.info['data_end_row']\n",
"role": "user"
},
{
"content": "update_spec",
"role": "assistant"
}
] |
def x(self):
SKIP_KEYS = ('_source_table', '_dest_table', 'd_vid', 't_vid', 'st_id',
'dataset', 'hash', 'process_records')
return OrderedDict([(k, getattr(self, k)) for k in self.properties if k
not in SKIP_KEYS])
| dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n SKIP_KEYS = ('_source_table', '_dest_table', 'd_vid', 't_vid', 'st_id',\n 'dataset', 'hash', 'process_records')\n return OrderedDict([(k, getattr(self, k)) for k in self.properties if k\n not in SKIP_KEYS])\n",
"role": "user"
},
{
"content": "dict",
"role": "assistant"
}
] |
def x(path=None, root=None, db=None):
return load(path, root=root, db=db)
| get_runconfig | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path=None, root=None, db=None):\n return load(path, root=root, db=db)\n",
"role": "user"
},
{
"content": "get_runconfig",
"role": "assistant"
}
] |
def x(path=None, root=None, db=None, load_user=True):
config = load_config(path, load_user=load_user)
remotes = load_remotes(path, load_user=load_user)
if remotes:
if not 'remotes' in config:
config.remotes = AttrDict()
for k, v in remotes.remotes.items():
config.remotes[k] = v
accounts = load_accounts(path, load_user=load_user)
if accounts:
if not 'accounts' in config:
config.accounts = AttrDict()
for k, v in accounts.accounts.items():
config.accounts[k] = v
update_config(config)
if root:
config.library.filesystem_root = root
if db:
config.library.database = db
return config
| load | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path=None, root=None, db=None, load_user=True):\n config = load_config(path, load_user=load_user)\n remotes = load_remotes(path, load_user=load_user)\n if remotes:\n if not 'remotes' in config:\n config.remotes = AttrDict()\n for k, v in remotes.remotes.items():\n config.remotes[k] = v\n accounts = load_accounts(path, load_user=load_user)\n if accounts:\n if not 'accounts' in config:\n config.accounts = AttrDict()\n for k, v in accounts.accounts.items():\n config.accounts[k] = v\n update_config(config)\n if root:\n config.library.filesystem_root = root\n if db:\n config.library.database = db\n return config\n",
"role": "user"
},
{
"content": "load",
"role": "assistant"
}
] |
def x(file_name, extra_path=None, load_user=True):
paths = []
if extra_path is not None:
paths.append(extra_path)
if os.getenv(ENVAR.CONFIG):
paths.append(os.getenv(ENVAR.CONFIG))
if os.getenv(ENVAR.VIRT):
paths.append(os.path.join(os.getenv(ENVAR.VIRT), USER_DIR))
if load_user:
paths.append(os.path.expanduser('~/' + USER_DIR))
paths.append(ROOT_DIR)
for path in paths:
if os.path.isdir(path) and os.path.exists(os.path.join(path, file_name)
):
f = os.path.join(path, file_name)
return f
raise ConfigurationError(
"Failed to find configuration file '{}'. Looked for : {} ".format(
file_name, paths))
| find_config_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(file_name, extra_path=None, load_user=True):\n paths = []\n if extra_path is not None:\n paths.append(extra_path)\n if os.getenv(ENVAR.CONFIG):\n paths.append(os.getenv(ENVAR.CONFIG))\n if os.getenv(ENVAR.VIRT):\n paths.append(os.path.join(os.getenv(ENVAR.VIRT), USER_DIR))\n if load_user:\n paths.append(os.path.expanduser('~/' + USER_DIR))\n paths.append(ROOT_DIR)\n for path in paths:\n if os.path.isdir(path) and os.path.exists(os.path.join(path, file_name)\n ):\n f = os.path.join(path, file_name)\n return f\n raise ConfigurationError(\n \"Failed to find configuration file '{}'. Looked for : {} \".format(\n file_name, paths))\n",
"role": "user"
},
{
"content": "find_config_file",
"role": "assistant"
}
] |
def x(extra_path=None, load_user=True):
from os.path import getmtime
try:
accts_file = find_config_file(ACCOUNTS_FILE, extra_path=extra_path,
load_user=load_user)
except ConfigurationError:
accts_file = None
if accts_file is not None and os.path.exists(accts_file):
config = AttrDict()
config.update_yaml(accts_file)
if not 'accounts' in config:
config.remotes = AttrDict()
config.accounts.loaded = [accts_file, getmtime(accts_file)]
return config
else:
return None
| load_accounts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(extra_path=None, load_user=True):\n from os.path import getmtime\n try:\n accts_file = find_config_file(ACCOUNTS_FILE, extra_path=extra_path,\n load_user=load_user)\n except ConfigurationError:\n accts_file = None\n if accts_file is not None and os.path.exists(accts_file):\n config = AttrDict()\n config.update_yaml(accts_file)\n if not 'accounts' in config:\n config.remotes = AttrDict()\n config.accounts.loaded = [accts_file, getmtime(accts_file)]\n return config\n else:\n return None\n",
"role": "user"
},
{
"content": "load_accounts",
"role": "assistant"
}
] |
def x(extra_path=None, load_user=True):
from os.path import getmtime
try:
remotes_file = find_config_file(REMOTES_FILE, extra_path=extra_path,
load_user=load_user)
except ConfigurationError:
remotes_file = None
if remotes_file is not None and os.path.exists(remotes_file):
config = AttrDict()
config.update_yaml(remotes_file)
if not 'remotes' in config:
config.remotes = AttrDict()
config.remotes.loaded = [remotes_file, getmtime(remotes_file)]
return config
else:
return None
| load_remotes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(extra_path=None, load_user=True):\n from os.path import getmtime\n try:\n remotes_file = find_config_file(REMOTES_FILE, extra_path=extra_path,\n load_user=load_user)\n except ConfigurationError:\n remotes_file = None\n if remotes_file is not None and os.path.exists(remotes_file):\n config = AttrDict()\n config.update_yaml(remotes_file)\n if not 'remotes' in config:\n config.remotes = AttrDict()\n config.remotes.loaded = [remotes_file, getmtime(remotes_file)]\n return config\n else:\n return None\n",
"role": "user"
},
{
"content": "load_remotes",
"role": "assistant"
}
] |
def x(path=None, load_user=True):
from os.path import getmtime
config = AttrDict()
if not path:
path = ROOT_DIR
config_file = find_config_file(CONFIG_FILE, extra_path=path, load_user=
load_user)
if os.path.exists(config_file):
config.update_yaml(config_file)
config.loaded = [config_file, getmtime(config_file)]
else:
config = AttrDict()
config.loaded = [None, 0]
return config
| load_config | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path=None, load_user=True):\n from os.path import getmtime\n config = AttrDict()\n if not path:\n path = ROOT_DIR\n config_file = find_config_file(CONFIG_FILE, extra_path=path, load_user=\n load_user)\n if os.path.exists(config_file):\n config.update_yaml(config_file)\n config.loaded = [config_file, getmtime(config_file)]\n else:\n config = AttrDict()\n config.loaded = [None, 0]\n return config\n",
"role": "user"
},
{
"content": "load_config",
"role": "assistant"
}
] |
def x(config, use_environ=True):
from ambry.util import select_from_url
try:
_ = config.library
except KeyError:
config.library = AttrDict()
try:
_ = config.filesystem
except KeyError:
config.filesystem = AttrDict()
try:
_ = config.accounts
except KeyError:
config.accounts = AttrDict()
if not config.accounts.get('loaded'):
config.accounts.loaded = [None, 0]
try:
_ = config.accounts.password
except KeyError:
config.accounts.password = None
try:
_ = config.remotes
except KeyError:
config.remotes = AttrDict()
if not config.remotes.get('loaded'):
config.remotes.loaded = [None, 0]
if use_environ:
if os.getenv(ENVAR.DB):
config.library.database = os.getenv(ENVAR.DB)
if os.getenv(ENVAR.ROOT):
config.library.filesystem_root = os.getenv(ENVAR.ROOT)
if os.getenv(ENVAR.PASSWORD):
config.accounts.password = os.getenv(ENVAR.PASSWORD)
try:
for k, v in config.library.remotes.items():
config.remotes[k] = {'url': v}
del config.library['remotes']
except KeyError as e:
pass
try:
for k, v in config.remotes.items():
if 'url' in v:
host = select_from_url(v['url'], 'netloc')
if host in config.accounts:
config.remotes[k].update(config.accounts[host])
del config.accounts[host]
except KeyError:
pass
try:
_ = config.library.database
except KeyError:
config.library.database = 'sqlite:///{root}/library.db'
checks = ['config.library.filesystem_root']
for check in checks:
try:
_ = eval(check)
except KeyError:
raise ConfigurationError(
"Configuration is missing '{}'; loaded from {} ".format(
check, config.loaded[0]))
_, config.library.database = normalize_dsn_or_dict(config.library.database)
for k, v in filesystem_defaults.items():
if k not in config.filesystem:
config.filesystem[k] = v
config.modtime = max(config.loaded[1], config.remotes.loaded[1], config
.accounts.loaded[1])
| update_config | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(config, use_environ=True):\n from ambry.util import select_from_url\n try:\n _ = config.library\n except KeyError:\n config.library = AttrDict()\n try:\n _ = config.filesystem\n except KeyError:\n config.filesystem = AttrDict()\n try:\n _ = config.accounts\n except KeyError:\n config.accounts = AttrDict()\n if not config.accounts.get('loaded'):\n config.accounts.loaded = [None, 0]\n try:\n _ = config.accounts.password\n except KeyError:\n config.accounts.password = None\n try:\n _ = config.remotes\n except KeyError:\n config.remotes = AttrDict()\n if not config.remotes.get('loaded'):\n config.remotes.loaded = [None, 0]\n if use_environ:\n if os.getenv(ENVAR.DB):\n config.library.database = os.getenv(ENVAR.DB)\n if os.getenv(ENVAR.ROOT):\n config.library.filesystem_root = os.getenv(ENVAR.ROOT)\n if os.getenv(ENVAR.PASSWORD):\n config.accounts.password = os.getenv(ENVAR.PASSWORD)\n try:\n for k, v in config.library.remotes.items():\n config.remotes[k] = {'url': v}\n del config.library['remotes']\n except KeyError as e:\n pass\n try:\n for k, v in config.remotes.items():\n if 'url' in v:\n host = select_from_url(v['url'], 'netloc')\n if host in config.accounts:\n config.remotes[k].update(config.accounts[host])\n del config.accounts[host]\n except KeyError:\n pass\n try:\n _ = config.library.database\n except KeyError:\n config.library.database = 'sqlite:///{root}/library.db'\n checks = ['config.library.filesystem_root']\n for check in checks:\n try:\n _ = eval(check)\n except KeyError:\n raise ConfigurationError(\n \"Configuration is missing '{}'; loaded from {} \".format(\n check, config.loaded[0]))\n _, config.library.database = normalize_dsn_or_dict(config.library.database)\n for k, v in filesystem_defaults.items():\n if k not in config.filesystem:\n config.filesystem[k] = v\n config.modtime = max(config.loaded[1], config.remotes.loaded[1], config\n .accounts.loaded[1])\n",
"role": "user"
},
{
"content": "update_config",
"role": "assistant"
}
] |
def x(d):
if isinstance(d, dict):
try:
d = d.to_dict()
except AttributeError:
pass
config = d
dsn = None
elif isinstance(d, string_types):
config = None
dsn = d
else:
raise ConfigurationError(
"Can't deal with database config '{}' type '{}' ".format(d,
type(d)))
if dsn:
if dsn.startswith('sqlite') or dsn.startswith('spatialite'):
driver, path = dsn.split(':', 1)
slashes, path = path[:2], path[2:]
if slashes != '//':
raise ConfigurationError(
'Sqlite DSNs must start with at least 2 slashes')
if len(path) == 1 and path[0] == '/':
raise ConfigurationError(
"Sqlite DSNs can't have only 3 slashes in path")
if len(path) > 1 and path[0] != '/':
raise ConfigurationError(
'Sqlite DSNs with a path must have 3 or 4 slashes.')
path = path[1:]
config = dict(server=None, username=None, password=None, driver
=driver, dbname=path)
else:
d = parse_url_to_dict(dsn)
config = dict(server=d['hostname'], dbname=d['path'].strip('/'),
driver=d['scheme'], password=d.get('password', None),
username=d.get('username', None))
else:
up = d.get('username', '') or ''
if d.get('password'):
up += ':' + d.get('password', '')
if up:
up += '@'
if up and not d.get('server'):
raise ConfigurationError(
"Can't construct a DSN with a username or password without a hostname"
)
host_part = up + d.get('server', '') if d.get('server') else ''
if d.get('dbname', False):
path_part = '/' + d.get('dbname')
else:
path_part = ''
dsn = '{}://{}{}'.format(d['driver'], host_part, path_part)
return config, dsn
| normalize_dsn_or_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(d):\n if isinstance(d, dict):\n try:\n d = d.to_dict()\n except AttributeError:\n pass\n config = d\n dsn = None\n elif isinstance(d, string_types):\n config = None\n dsn = d\n else:\n raise ConfigurationError(\n \"Can't deal with database config '{}' type '{}' \".format(d,\n type(d)))\n if dsn:\n if dsn.startswith('sqlite') or dsn.startswith('spatialite'):\n driver, path = dsn.split(':', 1)\n slashes, path = path[:2], path[2:]\n if slashes != '//':\n raise ConfigurationError(\n 'Sqlite DSNs must start with at least 2 slashes')\n if len(path) == 1 and path[0] == '/':\n raise ConfigurationError(\n \"Sqlite DSNs can't have only 3 slashes in path\")\n if len(path) > 1 and path[0] != '/':\n raise ConfigurationError(\n 'Sqlite DSNs with a path must have 3 or 4 slashes.')\n path = path[1:]\n config = dict(server=None, username=None, password=None, driver\n =driver, dbname=path)\n else:\n d = parse_url_to_dict(dsn)\n config = dict(server=d['hostname'], dbname=d['path'].strip('/'),\n driver=d['scheme'], password=d.get('password', None),\n username=d.get('username', None))\n else:\n up = d.get('username', '') or ''\n if d.get('password'):\n up += ':' + d.get('password', '')\n if up:\n up += '@'\n if up and not d.get('server'):\n raise ConfigurationError(\n \"Can't construct a DSN with a username or password without a hostname\"\n )\n host_part = up + d.get('server', '') if d.get('server') else ''\n if d.get('dbname', False):\n path_part = '/' + d.get('dbname')\n else:\n path_part = ''\n dsn = '{}://{}{}'.format(d['driver'], host_part, path_part)\n return config, dsn\n",
"role": "user"
},
{
"content": "normalize_dsn_or_dict",
"role": "assistant"
}
] |
def x(self, **kwargs):
kwargs['_return_http_data_only'] = True
if kwargs.get('callback'):
return self.create_new_product_version_with_http_info(**kwargs)
else:
data = self.create_new_product_version_with_http_info(**kwargs)
return data
| create_new_product_version | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.create_new_product_version_with_http_info(**kwargs)\n else:\n data = self.create_new_product_version_with_http_info(**kwargs)\n return data\n",
"role": "user"
},
{
"content": "create_new_product_version",
"role": "assistant"
}
] |
def x(self, id, **kwargs):
kwargs['_return_http_data_only'] = True
if kwargs.get('callback'):
return self.get_build_configuration_sets_with_http_info(id, **kwargs)
else:
data = self.get_build_configuration_sets_with_http_info(id, **kwargs)
return data
| get_build_configuration_sets | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.get_build_configuration_sets_with_http_info(id, **kwargs)\n else:\n data = self.get_build_configuration_sets_with_http_info(id, **kwargs)\n return data\n",
"role": "user"
},
{
"content": "get_build_configuration_sets",
"role": "assistant"
}
] |
def x(self, id, **kwargs):
kwargs['_return_http_data_only'] = True
if kwargs.get('callback'):
return self.update_build_configuration_sets_with_http_info(id, **kwargs
)
else:
data = self.update_build_configuration_sets_with_http_info(id, **kwargs
)
return data
| update_build_configuration_sets | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.update_build_configuration_sets_with_http_info(id, **kwargs\n )\n else:\n data = self.update_build_configuration_sets_with_http_info(id, **kwargs\n )\n return data\n",
"role": "user"
},
{
"content": "update_build_configuration_sets",
"role": "assistant"
}
] |
def x(cmd, execute, echo=True):
if execute:
if echo:
print('Executing: ' + cmd)
return os.system(cmd)
else:
print(cmd)
return 0
| execute_command | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cmd, execute, echo=True):\n if execute:\n if echo:\n print('Executing: ' + cmd)\n return os.system(cmd)\n else:\n print(cmd)\n return 0\n",
"role": "user"
},
{
"content": "execute_command",
"role": "assistant"
}
] |
def x(level):
lLevel = level.lower()
unrecognized = False
if lLevel == 'debug-all':
loglevel = logging.DEBUG
elif lLevel == 'debug':
loglevel = logging.DEBUG
elif lLevel == 'info':
loglevel = logging.INFO
elif lLevel == 'warning':
loglevel = logging.WARNING
elif lLevel == 'error':
loglevel = logging.ERROR
elif lLevel == 'critical':
loglevel = logging.CRITICAL
else:
loglevel = logging.DEBUG
unrecognized = True
formatter = logging.Formatter(
'%(asctime)s %(levelname)s %(filename)s:%(lineno)d/%(funcName)s: %(message)s'
)
console = logging.StreamHandler()
console.setLevel(loglevel)
console.setFormatter(formatter)
logging.getLogger('').addHandler(console)
logging.getLogger('').setLevel(loglevel)
if lLevel != 'debug-all':
packagesWarning = ['requests.packages.urllib3', 'urllib3',
'requests_kerberos', 'jenkinsapi']
for package in packagesWarning:
logging.debug('Setting loglevel for %s to WARNING.', package)
logger = logging.getLogger(package)
logger.setLevel(logging.WARNING)
if unrecognized:
logging.warning('Unrecognized log level: %s Log level set to debug',
level)
fh = logging.FileHandler('builder.log')
fh.setLevel(loglevel)
fh.setFormatter(formatter)
logging.getLogger('').addHandler(fh)
| set_log_level | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(level):\n lLevel = level.lower()\n unrecognized = False\n if lLevel == 'debug-all':\n loglevel = logging.DEBUG\n elif lLevel == 'debug':\n loglevel = logging.DEBUG\n elif lLevel == 'info':\n loglevel = logging.INFO\n elif lLevel == 'warning':\n loglevel = logging.WARNING\n elif lLevel == 'error':\n loglevel = logging.ERROR\n elif lLevel == 'critical':\n loglevel = logging.CRITICAL\n else:\n loglevel = logging.DEBUG\n unrecognized = True\n formatter = logging.Formatter(\n '%(asctime)s %(levelname)s %(filename)s:%(lineno)d/%(funcName)s: %(message)s'\n )\n console = logging.StreamHandler()\n console.setLevel(loglevel)\n console.setFormatter(formatter)\n logging.getLogger('').addHandler(console)\n logging.getLogger('').setLevel(loglevel)\n if lLevel != 'debug-all':\n packagesWarning = ['requests.packages.urllib3', 'urllib3',\n 'requests_kerberos', 'jenkinsapi']\n for package in packagesWarning:\n logging.debug('Setting loglevel for %s to WARNING.', package)\n logger = logging.getLogger(package)\n logger.setLevel(logging.WARNING)\n if unrecognized:\n logging.warning('Unrecognized log level: %s Log level set to debug',\n level)\n fh = logging.FileHandler('builder.log')\n fh.setLevel(loglevel)\n fh.setFormatter(formatter)\n logging.getLogger('').addHandler(fh)\n",
"role": "user"
},
{
"content": "set_log_level",
"role": "assistant"
}
] |
def x(argv):
args = {}
for rarg in argv:
if re.match('^--.*', rarg):
arg = rarg.replace('--', '', 1)
fsplit = arg.split('=', 1)
if len(fsplit) != 2:
raise Exception("Command option '%s' not recognized." % rarg)
rkey, value = fsplit
ssplit = rkey.split('.', 1)
if len(ssplit) != 2 or not ssplit[1]:
raise Exception("Command option '%s' not recognized." % rarg)
section, option = ssplit
args[section] = option, value
else:
raise Exception("Command option '%s' not recognized." % rarg)
return args
| parse_conf_args | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(argv):\n args = {}\n for rarg in argv:\n if re.match('^--.*', rarg):\n arg = rarg.replace('--', '', 1)\n fsplit = arg.split('=', 1)\n if len(fsplit) != 2:\n raise Exception(\"Command option '%s' not recognized.\" % rarg)\n rkey, value = fsplit\n ssplit = rkey.split('.', 1)\n if len(ssplit) != 2 or not ssplit[1]:\n raise Exception(\"Command option '%s' not recognized.\" % rarg)\n section, option = ssplit\n args[section] = option, value\n else:\n raise Exception(\"Command option '%s' not recognized.\" % rarg)\n return args\n",
"role": "user"
},
{
"content": "parse_conf_args",
"role": "assistant"
}
] |
def x(field):
def x(f):
def x(*args):
result = f(*args)
if result is None or result == '':
raise Exception("Config option '%s' is required." % field)
else:
return result
return wrappedf
return wrap
| required | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(field):\n\n def x(f):\n\n def x(*args):\n result = f(*args)\n if result is None or result == '':\n raise Exception(\"Config option '%s' is required.\" % field)\n else:\n return result\n return wrappedf\n return wrap\n",
"role": "user"
},
{
"content": "required",
"role": "assistant"
}
] |
def x(s, delim, escape='\\', unescape=True):
ret = []
current = []
itr = iter(s)
for ch in itr:
if ch == escape:
try:
if not unescape:
current.append(escape)
current.append(next(itr))
except StopIteration:
if unescape:
current.append(escape)
elif ch == delim:
ret.append(''.join(current))
current = []
else:
current.append(ch)
ret.append(''.join(current))
return ret
| split_unescape | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(s, delim, escape='\\\\', unescape=True):\n ret = []\n current = []\n itr = iter(s)\n for ch in itr:\n if ch == escape:\n try:\n if not unescape:\n current.append(escape)\n current.append(next(itr))\n except StopIteration:\n if unescape:\n current.append(escape)\n elif ch == delim:\n ret.append(''.join(current))\n current = []\n else:\n current.append(ch)\n ret.append(''.join(current))\n return ret\n",
"role": "user"
},
{
"content": "split_unescape",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.