source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(self, source=None, phase='build', ps=None): from ambry.etl.pipeline import Pipeline, PartitionWriter from ambry.dbexceptions import ConfigurationError if source: source = self.source(source) if isinstance(source, string_types ) else source else: source = None sf, sp = self.source_pipe(source, ps) if source else (None, None) pl = Pipeline(self, source=sp) try: phase_config = self.default_pipelines[phase] except KeyError: phase_config = None if phase_config: pl.configure(phase_config) pipe_config = None pipe_name = None if source and source.pipeline: pipe_name = source.pipeline try: pipe_config = self.metadata.pipelines[pipe_name] except KeyError: raise ConfigurationError( "Pipeline '{}' declared in source '{}', but not found in metadata" .format(source.pipeline, source.name)) else: pipe_name, pipe_config = self._find_pipeline(source, phase) if pipe_name: pl.name = pipe_name else: pl.name = phase pl.phase = phase def x(pl, pipe_config): if isinstance(pipe_config, (list, tuple)): store, body = [], [] for pipe in pipe_config: store.append(pipe) if isinstance(pipe, PartitionWriter ) else body.append(pipe) pipe_config = dict(body=body, store=store) if pipe_config: pl.configure(pipe_config) apply_config(pl, pipe_config) if 'all' in self.metadata.pipelines: apply_config(pl, self.metadata.pipelines['all']) self.edit_pipeline(pl) try: pl.dest_table = source.dest_table_name pl.source_table = source.source_table.name pl.source_name = source.name except AttributeError: pl.dest_table = None return pl
pipeline
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, source=None, phase='build', ps=None):\n from ambry.etl.pipeline import Pipeline, PartitionWriter\n from ambry.dbexceptions import ConfigurationError\n if source:\n source = self.source(source) if isinstance(source, string_types\n ) else source\n else:\n source = None\n sf, sp = self.source_pipe(source, ps) if source else (None, None)\n pl = Pipeline(self, source=sp)\n try:\n phase_config = self.default_pipelines[phase]\n except KeyError:\n phase_config = None\n if phase_config:\n pl.configure(phase_config)\n pipe_config = None\n pipe_name = None\n if source and source.pipeline:\n pipe_name = source.pipeline\n try:\n pipe_config = self.metadata.pipelines[pipe_name]\n except KeyError:\n raise ConfigurationError(\n \"Pipeline '{}' declared in source '{}', but not found in metadata\"\n .format(source.pipeline, source.name))\n else:\n pipe_name, pipe_config = self._find_pipeline(source, phase)\n if pipe_name:\n pl.name = pipe_name\n else:\n pl.name = phase\n pl.phase = phase\n\n def x(pl, pipe_config):\n if isinstance(pipe_config, (list, tuple)):\n store, body = [], []\n for pipe in pipe_config:\n store.append(pipe) if isinstance(pipe, PartitionWriter\n ) else body.append(pipe)\n pipe_config = dict(body=body, store=store)\n if pipe_config:\n pl.configure(pipe_config)\n apply_config(pl, pipe_config)\n if 'all' in self.metadata.pipelines:\n apply_config(pl, self.metadata.pipelines['all'])\n self.edit_pipeline(pl)\n try:\n pl.dest_table = source.dest_table_name\n pl.source_table = source.source_table.name\n pl.source_name = source.name\n except AttributeError:\n pl.dest_table = None\n return pl\n", "role": "user" }, { "content": "pipeline", "role": "assistant" } ]
def x(self, fields): row = self.dataset.row(fields) for i, f in enumerate(fields): if f == 'bstate': row[i] = self.state elif f == 'dstate': row[i] = self.dstate elif f == 'source_fs': row[i] = self.source_fs elif f.startswith('about'): _, key = f.split('.') row[i] = self.metadata.about[key] elif f.startswith('state'): _, key = f.split('.') row[i] = self.buildstate.state[key] elif f.startswith('count'): _, key = f.split('.') if key == 'sources': row[i] = len(self.dataset.sources) elif key == 'tables': row[i] = len(self.dataset.tables) return row
field_row
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, fields):\n row = self.dataset.row(fields)\n for i, f in enumerate(fields):\n if f == 'bstate':\n row[i] = self.state\n elif f == 'dstate':\n row[i] = self.dstate\n elif f == 'source_fs':\n row[i] = self.source_fs\n elif f.startswith('about'):\n _, key = f.split('.')\n row[i] = self.metadata.about[key]\n elif f.startswith('state'):\n _, key = f.split('.')\n row[i] = self.buildstate.state[key]\n elif f.startswith('count'):\n _, key = f.split('.')\n if key == 'sources':\n row[i] = len(self.dataset.sources)\n elif key == 'tables':\n row[i] = len(self.dataset.tables)\n return row\n", "role": "user" }, { "content": "field_row", "role": "assistant" } ]
def x(self, source, ps=None): if isinstance(source, string_types): source = self.source(source) source.dataset = self.dataset source._bundle = self iter_source, source_pipe = self._iterable_source(source, ps) if self.limited_run: source_pipe.limit = 500 return iter_source, source_pipe
source_pipe
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, source, ps=None):\n if isinstance(source, string_types):\n source = self.source(source)\n source.dataset = self.dataset\n source._bundle = self\n iter_source, source_pipe = self._iterable_source(source, ps)\n if self.limited_run:\n source_pipe.limit = 500\n return iter_source, source_pipe\n", "role": "user" }, { "content": "source_pipe", "role": "assistant" } ]
def x(self, source, clean=False, callback=None): from fs.zipfs import ZipOpenError import os from ambry_sources.sources import GoogleSource, CsvSource, TsvSource, FixedSource, ExcelSource, PartitionSource, SourceError, DelayedOpen, DelayedDownload, ShapefileSource, SocrataSource from ambry_sources import extract_file_from_zip spec = source.spec cache_fs = self.library.download_cache account_accessor = self.library.account_accessor url_type = spec.get_urltype() def x(): from ambry_sources.fetch import download return download(spec.url, cache_fs, account_accessor, clean=clean, logger=self.logger, callback=callback) if url_type == 'file': from fs.opener import fsopen syspath = spec.url.replace('file://', '') cache_path = syspath.strip('/') cache_fs.makedir(os.path.dirname(cache_path), recursive=True, allow_recreate=True) if os.path.isabs(syspath): with open(syspath) as f: cache_fs.setcontents(cache_path, f) else: cache_fs.setcontents(cache_path, self.source_fs.getcontents( syspath)) elif url_type not in ('gs', 'socrata'): try: cache_path, download_time = do_download() spec.download_time = download_time except Exception as e: from ambry_sources.exceptions import DownloadError raise DownloadError('Failed to download {}; {}'.format(spec.url, e) ) else: cache_path, download_time = None, None if url_type == 'zip': try: fstor = extract_file_from_zip(cache_fs, cache_path, spec.url, spec.file) except ZipOpenError: cache_fs.remove(cache_path) cache_path, spec.download_time = do_download() fstor = extract_file_from_zip(cache_fs, cache_path, spec.url, spec.file) file_type = spec.get_filetype(fstor.path) elif url_type == 'gs': fstor = get_gs(spec.url, spec.segment, account_accessor) file_type = 'gs' elif url_type == 'socrata': spec.encoding = 'utf8' spec.header_lines = [0] spec.start_line = 1 url = SocrataSource.download_url(spec) fstor = DelayedDownload(url, cache_fs) file_type = 'socrata' else: fstor = DelayedOpen(cache_fs, cache_path, 'rb') file_type = spec.get_filetype(fstor.path) spec.filetype = file_type TYPE_TO_SOURCE_MAP = {'gs': GoogleSource, 'csv': CsvSource, 'tsv': TsvSource, 'fixed': FixedSource, 'txt': FixedSource, 'xls': ExcelSource, 'xlsx': ExcelSource, 'partition': PartitionSource, 'shape': ShapefileSource, 'socrata': SocrataSource} cls = TYPE_TO_SOURCE_MAP.get(file_type) if cls is None: raise SourceError( "Failed to determine file type for source '{}'; unknown type '{}' " .format(spec.name, file_type)) return cls(spec, fstor)
get_source
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, source, clean=False, callback=None):\n from fs.zipfs import ZipOpenError\n import os\n from ambry_sources.sources import GoogleSource, CsvSource, TsvSource, FixedSource, ExcelSource, PartitionSource, SourceError, DelayedOpen, DelayedDownload, ShapefileSource, SocrataSource\n from ambry_sources import extract_file_from_zip\n spec = source.spec\n cache_fs = self.library.download_cache\n account_accessor = self.library.account_accessor\n url_type = spec.get_urltype()\n\n def x():\n from ambry_sources.fetch import download\n return download(spec.url, cache_fs, account_accessor, clean=clean,\n logger=self.logger, callback=callback)\n if url_type == 'file':\n from fs.opener import fsopen\n syspath = spec.url.replace('file://', '')\n cache_path = syspath.strip('/')\n cache_fs.makedir(os.path.dirname(cache_path), recursive=True,\n allow_recreate=True)\n if os.path.isabs(syspath):\n with open(syspath) as f:\n cache_fs.setcontents(cache_path, f)\n else:\n cache_fs.setcontents(cache_path, self.source_fs.getcontents(\n syspath))\n elif url_type not in ('gs', 'socrata'):\n try:\n cache_path, download_time = do_download()\n spec.download_time = download_time\n except Exception as e:\n from ambry_sources.exceptions import DownloadError\n raise DownloadError('Failed to download {}; {}'.format(spec.url, e)\n )\n else:\n cache_path, download_time = None, None\n if url_type == 'zip':\n try:\n fstor = extract_file_from_zip(cache_fs, cache_path, spec.url,\n spec.file)\n except ZipOpenError:\n cache_fs.remove(cache_path)\n cache_path, spec.download_time = do_download()\n fstor = extract_file_from_zip(cache_fs, cache_path, spec.url,\n spec.file)\n file_type = spec.get_filetype(fstor.path)\n elif url_type == 'gs':\n fstor = get_gs(spec.url, spec.segment, account_accessor)\n file_type = 'gs'\n elif url_type == 'socrata':\n spec.encoding = 'utf8'\n spec.header_lines = [0]\n spec.start_line = 1\n url = SocrataSource.download_url(spec)\n fstor = DelayedDownload(url, cache_fs)\n file_type = 'socrata'\n else:\n fstor = DelayedOpen(cache_fs, cache_path, 'rb')\n file_type = spec.get_filetype(fstor.path)\n spec.filetype = file_type\n TYPE_TO_SOURCE_MAP = {'gs': GoogleSource, 'csv': CsvSource, 'tsv':\n TsvSource, 'fixed': FixedSource, 'txt': FixedSource, 'xls':\n ExcelSource, 'xlsx': ExcelSource, 'partition': PartitionSource,\n 'shape': ShapefileSource, 'socrata': SocrataSource}\n cls = TYPE_TO_SOURCE_MAP.get(file_type)\n if cls is None:\n raise SourceError(\n \"Failed to determine file type for source '{}'; unknown type '{}' \"\n .format(spec.name, file_type))\n return cls(spec, fstor)\n", "role": "user" }, { "content": "get_source", "role": "assistant" } ]
def x(self): self.buildstate.state.lasttime = time() self.buildstate.commit() return self.buildstate.state.error
error_state
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.buildstate.state.lasttime = time()\n self.buildstate.commit()\n return self.buildstate.state.error\n", "role": "user" }, { "content": "error_state", "role": "assistant" } ]
def x(self, state): assert state != 'build_bundle' self.buildstate.state.current = state self.buildstate.state[state] = time() self.buildstate.state.lasttime = time() self.buildstate.state.error = False self.buildstate.state.exception = None self.buildstate.state.exception_type = None self.buildstate.commit() if state in (self.STATES.NEW, self.STATES.CLEANED, self.STATES.BUILT, self.STATES.FINALIZED, self.STATES.SOURCE): state = state if state != self.STATES.CLEANED else self.STATES.NEW self.dstate = state
state
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, state):\n assert state != 'build_bundle'\n self.buildstate.state.current = state\n self.buildstate.state[state] = time()\n self.buildstate.state.lasttime = time()\n self.buildstate.state.error = False\n self.buildstate.state.exception = None\n self.buildstate.state.exception_type = None\n self.buildstate.commit()\n if state in (self.STATES.NEW, self.STATES.CLEANED, self.STATES.BUILT,\n self.STATES.FINALIZED, self.STATES.SOURCE):\n state = state if state != self.STATES.CLEANED else self.STATES.NEW\n self.dstate = state\n", "role": "user" }, { "content": "state", "role": "assistant" } ]
def x(self, phase, stage): key = '{}-{}'.format(phase, stage if stage else 1) self.buildstate.state[key] = time()
record_stage_state
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, phase, stage):\n key = '{}-{}'.format(phase, stage if stage else 1)\n self.buildstate.state[key] = time()\n", "role": "user" }, { "content": "record_stage_state", "role": "assistant" } ]
def x(self, tag): import time self.buildstate.access.last = '{}-{}'.format(tag, time.time()) self.buildstate.commit()
set_last_access
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tag):\n import time\n self.buildstate.access.last = '{}-{}'.format(tag, time.time())\n self.buildstate.commit()\n", "role": "user" }, { "content": "set_last_access", "role": "assistant" } ]
def x(self, force=False): self.log('---- Sync In ----') self.dstate = self.STATES.BUILDING for path_name in self.source_fs.listdir(): f = self.build_source_files.instance_from_name(path_name) if not f: self.warn('Ignoring unknown file: {}'.format(path_name)) continue if f and f.exists and (f.fs_is_newer or force): self.log('Sync: {}'.format(f.record.path)) f.fs_to_record() f.record_to_objects() self.commit() self.library.search.index_bundle(self, force=True)
sync_in
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, force=False):\n self.log('---- Sync In ----')\n self.dstate = self.STATES.BUILDING\n for path_name in self.source_fs.listdir():\n f = self.build_source_files.instance_from_name(path_name)\n if not f:\n self.warn('Ignoring unknown file: {}'.format(path_name))\n continue\n if f and f.exists and (f.fs_is_newer or force):\n self.log('Sync: {}'.format(f.record.path))\n f.fs_to_record()\n f.record_to_objects()\n self.commit()\n self.library.search.index_bundle(self, force=True)\n", "role": "user" }, { "content": "sync_in", "role": "assistant" } ]
def x(self, force=False): self.log('---- Sync Files ----') self.dstate = self.STATES.BUILDING for f in self.build_source_files: if self.source_fs.exists(f.record.path): if f.fs_is_newer or force: self.log('Sync: {}'.format(f.record.path)) f.fs_to_record() self.commit()
sync_in_files
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, force=False):\n self.log('---- Sync Files ----')\n self.dstate = self.STATES.BUILDING\n for f in self.build_source_files:\n if self.source_fs.exists(f.record.path):\n if f.fs_is_newer or force:\n self.log('Sync: {}'.format(f.record.path))\n f.fs_to_record()\n self.commit()\n", "role": "user" }, { "content": "sync_in_files", "role": "assistant" } ]
def x(self, force=False): self.log('---- Sync Files ----') for f in self.build_source_files: f.record_to_objects() self.build_source_files.file(File.BSFILE.META).record_to_objects() self.commit()
sync_in_records
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, force=False):\n self.log('---- Sync Files ----')\n for f in self.build_source_files:\n f.record_to_objects()\n self.build_source_files.file(File.BSFILE.META).record_to_objects()\n self.commit()\n", "role": "user" }, { "content": "sync_in_records", "role": "assistant" } ]
def x(self, file_name=None, force=False): self.log('---- Sync Out ----') from ambry.bundle.files import BuildSourceFile self.dstate = self.STATES.BUILDING for f in self.build_source_files.list_records(): if (f.sync_dir() == BuildSourceFile.SYNC_DIR.RECORD_TO_FILE or f. record.path == file_name) or force: self.log('Sync: {}'.format(f.record.path)) f.record_to_fs() self.commit()
sync_out
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, file_name=None, force=False):\n self.log('---- Sync Out ----')\n from ambry.bundle.files import BuildSourceFile\n self.dstate = self.STATES.BUILDING\n for f in self.build_source_files.list_records():\n if (f.sync_dir() == BuildSourceFile.SYNC_DIR.RECORD_TO_FILE or f.\n record.path == file_name) or force:\n self.log('Sync: {}'.format(f.record.path))\n f.record_to_fs()\n self.commit()\n", "role": "user" }, { "content": "sync_out", "role": "assistant" } ]
def x(self): self.dstate = self.STATES.BUILDING self.build_source_files.record_to_objects()
sync_objects_in
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.dstate = self.STATES.BUILDING\n self.build_source_files.record_to_objects()\n", "role": "user" }, { "content": "sync_objects_in", "role": "assistant" } ]
def x(self, force=False): self.log('---- Sync Objects Out ----') from ambry.bundle.files import BuildSourceFile self.dstate = self.STATES.BUILDING for f in self.build_source_files.list_records(): self.log('Sync: {}'.format(f.record.path)) f.objects_to_record() self.commit()
sync_objects_out
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, force=False):\n self.log('---- Sync Objects Out ----')\n from ambry.bundle.files import BuildSourceFile\n self.dstate = self.STATES.BUILDING\n for f in self.build_source_files.list_records():\n self.log('Sync: {}'.format(f.record.path))\n f.objects_to_record()\n self.commit()\n", "role": "user" }, { "content": "sync_objects_out", "role": "assistant" } ]
def x(self): from ambry.orm.file import File from ambry.bundle.files import BuildSourceFile self.dstate = self.STATES.BUILDING synced = 0 for fc in [File.BSFILE.BUILD, File.BSFILE.META, File.BSFILE.LIB, File. BSFILE.TEST, File.BSFILE.DOC]: bsf = self.build_source_files.file(fc) if bsf.fs_is_newer: self.log('Syncing {}'.format(bsf.file_name)) bsf.sync(BuildSourceFile.SYNC_DIR.FILE_TO_RECORD) synced += 1 self.build_source_files.file(File.BSFILE.META).record_to_objects() return synced
sync_code
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry.orm.file import File\n from ambry.bundle.files import BuildSourceFile\n self.dstate = self.STATES.BUILDING\n synced = 0\n for fc in [File.BSFILE.BUILD, File.BSFILE.META, File.BSFILE.LIB, File.\n BSFILE.TEST, File.BSFILE.DOC]:\n bsf = self.build_source_files.file(fc)\n if bsf.fs_is_newer:\n self.log('Syncing {}'.format(bsf.file_name))\n bsf.sync(BuildSourceFile.SYNC_DIR.FILE_TO_RECORD)\n synced += 1\n self.build_source_files.file(File.BSFILE.META).record_to_objects()\n return synced\n", "role": "user" }, { "content": "sync_code", "role": "assistant" } ]
def x(self, force=False): from ambry.orm.file import File self.dstate = self.STATES.BUILDING synced = 0 for fc in [File.BSFILE.SOURCES]: bsf = self.build_source_files.file(fc) if bsf.fs_is_newer or force: self.log('Syncing {}'.format(bsf.file_name)) bsf.fs_to_objects() synced += 1 return synced
sync_sources
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, force=False):\n from ambry.orm.file import File\n self.dstate = self.STATES.BUILDING\n synced = 0\n for fc in [File.BSFILE.SOURCES]:\n bsf = self.build_source_files.file(fc)\n if bsf.fs_is_newer or force:\n self.log('Syncing {}'.format(bsf.file_name))\n bsf.fs_to_objects()\n synced += 1\n return synced\n", "role": "user" }, { "content": "sync_sources", "role": "assistant" } ]
def x(self): from ambry.orm.file import File from ambry.bundle.files import BuildSourceFile self.dstate = self.STATES.BUILDING synced = 0 for fc in [File.BSFILE.SCHEMA, File.BSFILE.SOURCESCHEMA]: bsf = self.build_source_files.file(fc) if bsf.fs_is_newer: self.log('Syncing {}'.format(bsf.file_name)) bsf.sync(BuildSourceFile.SYNC_DIR.FILE_TO_RECORD) synced += 1 return synced
sync_schema
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry.orm.file import File\n from ambry.bundle.files import BuildSourceFile\n self.dstate = self.STATES.BUILDING\n synced = 0\n for fc in [File.BSFILE.SCHEMA, File.BSFILE.SOURCESCHEMA]:\n bsf = self.build_source_files.file(fc)\n if bsf.fs_is_newer:\n self.log('Syncing {}'.format(bsf.file_name))\n bsf.sync(BuildSourceFile.SYNC_DIR.FILE_TO_RECORD)\n synced += 1\n return synced\n", "role": "user" }, { "content": "sync_schema", "role": "assistant" } ]
def x(self): self.commit() self.build_source_files.schema.objects_to_record() self.commit()
update_schema
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.commit()\n self.build_source_files.schema.objects_to_record()\n self.commit()\n", "role": "user" }, { "content": "update_schema", "role": "assistant" } ]
def x(self, force=False): if self.is_finalized and not force: self.warn("Can't clean; bundle is finalized") return False self.log('---- Cleaning ----') self.state = self.STATES.CLEANING self.dstate = self.STATES.BUILDING self.commit() self.clean_sources() self.clean_tables() self.clean_partitions() self.clean_build() self.clean_files() self.clean_ingested() self.clean_build_state() self.clean_progress() self.state = self.STATES.CLEANED self.commit() return True
clean
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, force=False):\n if self.is_finalized and not force:\n self.warn(\"Can't clean; bundle is finalized\")\n return False\n self.log('---- Cleaning ----')\n self.state = self.STATES.CLEANING\n self.dstate = self.STATES.BUILDING\n self.commit()\n self.clean_sources()\n self.clean_tables()\n self.clean_partitions()\n self.clean_build()\n self.clean_files()\n self.clean_ingested()\n self.clean_build_state()\n self.clean_progress()\n self.state = self.STATES.CLEANED\n self.commit()\n return True\n", "role": "user" }, { "content": "clean", "role": "assistant" } ]
def x(self): if self.is_finalized: self.warn("Can't clean; bundle is finalized") return False self.log('---- Cleaning ----') self.state = self.STATES.CLEANING self.commit() self.clean_sources() self.clean_tables() self.clean_partitions() self.clean_build() self.clean_ingested() self.clean_build_state() self.state = self.STATES.CLEANED self.commit() self.log('---- Done Cleaning ----') return True
clean_except_files
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.is_finalized:\n self.warn(\"Can't clean; bundle is finalized\")\n return False\n self.log('---- Cleaning ----')\n self.state = self.STATES.CLEANING\n self.commit()\n self.clean_sources()\n self.clean_tables()\n self.clean_partitions()\n self.clean_build()\n self.clean_ingested()\n self.clean_build_state()\n self.state = self.STATES.CLEANED\n self.commit()\n self.log('---- Done Cleaning ----')\n return True\n", "role": "user" }, { "content": "clean_except_files", "role": "assistant" } ]
def x(self): for src in self.dataset.sources: src.st_id = None src.t_id = None self.dataset.sources[:] = [] self.dataset.source_tables[:] = [] self.dataset.st_sequence_id = 1
clean_sources
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for src in self.dataset.sources:\n src.st_id = None\n src.t_id = None\n self.dataset.sources[:] = []\n self.dataset.source_tables[:] = []\n self.dataset.st_sequence_id = 1\n", "role": "user" }, { "content": "clean_sources", "role": "assistant" } ]
def x(self): import shutil from ambry.orm import ColumnStat self.session.query(ColumnStat).filter(ColumnStat.d_vid == self.dataset.vid ).delete() self.dataset.delete_partitions() for s in self.sources: s.state = None if self.build_partition_fs.exists: try: shutil.rmtree(self.build_partition_fs.getsyspath('/')) except NoSysPathError: pass
clean_partitions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n import shutil\n from ambry.orm import ColumnStat\n self.session.query(ColumnStat).filter(ColumnStat.d_vid == self.dataset.vid\n ).delete()\n self.dataset.delete_partitions()\n for s in self.sources:\n s.state = None\n if self.build_partition_fs.exists:\n try:\n shutil.rmtree(self.build_partition_fs.getsyspath('/'))\n except NoSysPathError:\n pass\n", "role": "user" }, { "content": "clean_partitions", "role": "assistant" } ]
def x(self): import shutil if self.build_fs.exists: try: shutil.rmtree(self.build_fs.getsyspath('/')) except NoSysPathError: pass
clean_build
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n import shutil\n if self.build_fs.exists:\n try:\n shutil.rmtree(self.build_fs.getsyspath('/'))\n except NoSysPathError:\n pass\n", "role": "user" }, { "content": "clean_build", "role": "assistant" } ]
def x(self): for s in self.sources: df = s.datafile if df.exists and not s.is_partition: df.remove() s.state = s.STATES.NEW self.commit()
clean_ingested
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for s in self.sources:\n df = s.datafile\n if df.exists and not s.is_partition:\n df.remove()\n s.state = s.STATES.NEW\n self.commit()\n", "role": "user" }, { "content": "clean_ingested", "role": "assistant" } ]
def x(self): ds = self.dataset ds.config.build.clean() ds.config.process.clean() ds.commit() self.state = self.STATES.CLEANED
clean_process_meta
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n ds = self.dataset\n ds.config.build.clean()\n ds.config.process.clean()\n ds.commit()\n self.state = self.STATES.CLEANED\n", "role": "user" }, { "content": "clean_process_meta", "role": "assistant" } ]
def x(self): self.build_source_files.file(File.BSFILE.SOURCESCHEMA).remove() self.build_source_files.file(File.BSFILE.SCHEMA).remove() self.commit()
clean_source_files
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.build_source_files.file(File.BSFILE.SOURCESCHEMA).remove()\n self.build_source_files.file(File.BSFILE.SCHEMA).remove()\n self.commit()\n", "role": "user" }, { "content": "clean_source_files", "role": "assistant" } ]
def x(self, sources=None, tables=None, stage=None, force=False, load_meta=False ): from itertools import groupby from ambry.bundle.events import TAG from fs.errors import ResourceNotFoundError import zlib self.log('---- Ingesting ----') self.dstate = self.STATES.BUILDING self.commit() key = lambda s: s.stage if s.stage else 1 def x(s): import zlib if force: return True try: return s.is_processable and not s.is_ingested and not s.is_built except (IOError, zlib.error): s.local_datafile.remove() return True sources = sorted(self._resolve_sources(sources, tables, stage, predicate=not_final_or_delete), key=key) if not sources: self.log('No sources left to ingest') return self.state = self.STATES.INGESTING count = 0 errors = 0 self._run_events(TAG.BEFORE_INGEST, 0) for s in self.sources: if s.is_downloadable: df = s.datafile try: info = df.info df.close() except (ResourceNotFoundError, zlib.error, IOError): df.remove() for stage, g in groupby(sources, key): sources = [s for s in g if not_final_or_delete(s)] if not len(sources): continue self._run_events(TAG.BEFORE_INGEST, stage) stage_errors = self._ingest_sources(sources, stage, force=force) errors += stage_errors count += len(sources) - stage_errors self._run_events(TAG.AFTER_INGEST, stage) self.record_stage_state(self.STATES.INGESTING, stage) self.state = self.STATES.INGESTED try: pass finally: self._run_events(TAG.AFTER_INGEST, 0) self.log('Ingested {} sources'.format(count)) if load_meta: if len(sources) == 1: iterable_source, source_pipe = self.source_pipe(sources[0]) try: meta = iterable_source.meta if meta: self.metadata.about.title = meta['title'] self.metadata.about.summary = meta['summary'] self.build_source_files.bundle_meta.objects_to_record() except AttributeError as e: self.warn('Failed to set metadata: {}'.format(e)) pass else: self.warn( "Didn't not load meta from source. Must have exactly one soruce, got {}" .format(len(sources))) self.commit() if errors == 0: return True else: return False
ingest
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, sources=None, tables=None, stage=None, force=False, load_meta=False\n ):\n from itertools import groupby\n from ambry.bundle.events import TAG\n from fs.errors import ResourceNotFoundError\n import zlib\n self.log('---- Ingesting ----')\n self.dstate = self.STATES.BUILDING\n self.commit()\n key = lambda s: s.stage if s.stage else 1\n\n def x(s):\n import zlib\n if force:\n return True\n try:\n return s.is_processable and not s.is_ingested and not s.is_built\n except (IOError, zlib.error):\n s.local_datafile.remove()\n return True\n sources = sorted(self._resolve_sources(sources, tables, stage,\n predicate=not_final_or_delete), key=key)\n if not sources:\n self.log('No sources left to ingest')\n return\n self.state = self.STATES.INGESTING\n count = 0\n errors = 0\n self._run_events(TAG.BEFORE_INGEST, 0)\n for s in self.sources:\n if s.is_downloadable:\n df = s.datafile\n try:\n info = df.info\n df.close()\n except (ResourceNotFoundError, zlib.error, IOError):\n df.remove()\n for stage, g in groupby(sources, key):\n sources = [s for s in g if not_final_or_delete(s)]\n if not len(sources):\n continue\n self._run_events(TAG.BEFORE_INGEST, stage)\n stage_errors = self._ingest_sources(sources, stage, force=force)\n errors += stage_errors\n count += len(sources) - stage_errors\n self._run_events(TAG.AFTER_INGEST, stage)\n self.record_stage_state(self.STATES.INGESTING, stage)\n self.state = self.STATES.INGESTED\n try:\n pass\n finally:\n self._run_events(TAG.AFTER_INGEST, 0)\n self.log('Ingested {} sources'.format(count))\n if load_meta:\n if len(sources) == 1:\n iterable_source, source_pipe = self.source_pipe(sources[0])\n try:\n meta = iterable_source.meta\n if meta:\n self.metadata.about.title = meta['title']\n self.metadata.about.summary = meta['summary']\n self.build_source_files.bundle_meta.objects_to_record()\n except AttributeError as e:\n self.warn('Failed to set metadata: {}'.format(e))\n pass\n else:\n self.warn(\n \"Didn't not load meta from source. Must have exactly one soruce, got {}\"\n .format(len(sources)))\n self.commit()\n if errors == 0:\n return True\n else:\n return False\n", "role": "user" }, { "content": "ingest", "role": "assistant" } ]
def x(self, sources, stage, force=False): from concurrent import ingest_mp self.state = self.STATES.INGESTING downloadable_sources = [s for s in sources if force or s.is_processable and not s.is_ingested and not s.is_built] errors = 0 with self.progress.start('ingest', stage, message='Ingesting ' + ('MP' if self.multi else 'SP'), item_total=len(sources), item_type='source', item_count=len(downloadable_sources)) as ps: for source in sources: _ = source.source_table if self.multi: args = [(self.identity.vid, stage, source.vid, force) for source in downloadable_sources] pool = self.library.process_pool(limited_run=self.limited_run) try: result = pool.map_async(ingest_mp, args, 1) pool.close() pool.join() except KeyboardInterrupt: self.log('Got keyboard interrrupt; terminating workers') pool.terminate() raise else: for i, source in enumerate(downloadable_sources, 1): ps.add(message='Ingesting source #{}, {}'.format(i, source. name), source=source, state='running') r = self._ingest_source(source, ps, force) if not r: errors += 1 if errors > 0: from ambry.dbexceptions import IngestionError raise IngestionError('Failed to ingest {} sources'.format(errors)) return errors
ingest_sources
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, sources, stage, force=False):\n from concurrent import ingest_mp\n self.state = self.STATES.INGESTING\n downloadable_sources = [s for s in sources if force or s.is_processable and\n not s.is_ingested and not s.is_built]\n errors = 0\n with self.progress.start('ingest', stage, message='Ingesting ' + ('MP' if\n self.multi else 'SP'), item_total=len(sources), item_type='source',\n item_count=len(downloadable_sources)) as ps:\n for source in sources:\n _ = source.source_table\n if self.multi:\n args = [(self.identity.vid, stage, source.vid, force) for\n source in downloadable_sources]\n pool = self.library.process_pool(limited_run=self.limited_run)\n try:\n result = pool.map_async(ingest_mp, args, 1)\n pool.close()\n pool.join()\n except KeyboardInterrupt:\n self.log('Got keyboard interrrupt; terminating workers')\n pool.terminate()\n raise\n else:\n for i, source in enumerate(downloadable_sources, 1):\n ps.add(message='Ingesting source #{}, {}'.format(i, source.\n name), source=source, state='running')\n r = self._ingest_source(source, ps, force)\n if not r:\n errors += 1\n if errors > 0:\n from ambry.dbexceptions import IngestionError\n raise IngestionError('Failed to ingest {} sources'.format(errors))\n return errors\n", "role": "user" }, { "content": "ingest_sources", "role": "assistant" } ]
def x(self, source, ps, force=None): from ambry.bundle.process import call_interval try: from ambry.orm.exc import NotFoundError if not source.is_partition and source.datafile.exists: if not source.datafile.is_finalized: source.datafile.remove() elif force: source.datafile.remove() else: ps.update(message='Source {} already ingested, skipping'. format(source.name), state='skipped') return True if source.is_partition: try: self.library.partition(source.ref) except NotFoundError: ps.update(message= "Not Ingesting {}: referenced partition '{}' does not exist" .format(source.name, source.ref), state='skipped') return True source.state = source.STATES.INGESTING iterable_source, source_pipe = self.source_pipe(source, ps) if not source.is_ingestible: ps.update(message='Not an ingestiable source: {}'.format(source .name), state='skipped', source=source) source.state = source.STATES.NOTINGESTABLE return True ps.update('Ingesting {} from {}'.format(source.spec.name, source. url or source.generator), item_type='rows', item_count=0) @call_interval(5) def x(i): desc, n_records, total, rate = source.datafile.report_progress() ps.update(message='Ingesting {}: rate: {}'.format(source.spec. name, rate), item_count=n_records) source.datafile.load_rows(iterable_source, callback= ingest_progress_f, limit=500 if self.limited_run else None, intuit_type=True, run_stats=False) if source.datafile.meta['warnings']: for w in source.datafile.meta['warnings']: self.error('Ingestion error: {}'.format(w)) ps.update(message='Ingested to {}'.format(source.datafile.syspath)) ps.update(message='Updating tables and specs for {}'.format(source. name)) source.update_spec() if self.limited_run: source.end_line = None self.build_source_files.sources.objects_to_record() ps.update(message='Ingested {}'.format(source.datafile.path), state ='done') source.state = source.STATES.INGESTED self.commit() return True except Exception as e: import traceback from ambry.util import qualified_class_name ps.update(message='Source {} failed with exception: {}'.format( source.name, e), exception_class=qualified_class_name(e), exception_trace=str(traceback.format_exc()), state='error') source.state = source.STATES.INGESTING + '_error' self.commit() return False
ingest_source
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, source, ps, force=None):\n from ambry.bundle.process import call_interval\n try:\n from ambry.orm.exc import NotFoundError\n if not source.is_partition and source.datafile.exists:\n if not source.datafile.is_finalized:\n source.datafile.remove()\n elif force:\n source.datafile.remove()\n else:\n ps.update(message='Source {} already ingested, skipping'.\n format(source.name), state='skipped')\n return True\n if source.is_partition:\n try:\n self.library.partition(source.ref)\n except NotFoundError:\n ps.update(message=\n \"Not Ingesting {}: referenced partition '{}' does not exist\"\n .format(source.name, source.ref), state='skipped')\n return True\n source.state = source.STATES.INGESTING\n iterable_source, source_pipe = self.source_pipe(source, ps)\n if not source.is_ingestible:\n ps.update(message='Not an ingestiable source: {}'.format(source\n .name), state='skipped', source=source)\n source.state = source.STATES.NOTINGESTABLE\n return True\n ps.update('Ingesting {} from {}'.format(source.spec.name, source.\n url or source.generator), item_type='rows', item_count=0)\n\n @call_interval(5)\n def x(i):\n desc, n_records, total, rate = source.datafile.report_progress()\n ps.update(message='Ingesting {}: rate: {}'.format(source.spec.\n name, rate), item_count=n_records)\n source.datafile.load_rows(iterable_source, callback=\n ingest_progress_f, limit=500 if self.limited_run else None,\n intuit_type=True, run_stats=False)\n if source.datafile.meta['warnings']:\n for w in source.datafile.meta['warnings']:\n self.error('Ingestion error: {}'.format(w))\n ps.update(message='Ingested to {}'.format(source.datafile.syspath))\n ps.update(message='Updating tables and specs for {}'.format(source.\n name))\n source.update_spec()\n if self.limited_run:\n source.end_line = None\n self.build_source_files.sources.objects_to_record()\n ps.update(message='Ingested {}'.format(source.datafile.path), state\n ='done')\n source.state = source.STATES.INGESTED\n self.commit()\n return True\n except Exception as e:\n import traceback\n from ambry.util import qualified_class_name\n ps.update(message='Source {} failed with exception: {}'.format(\n source.name, e), exception_class=qualified_class_name(e),\n exception_trace=str(traceback.format_exc()), state='error')\n source.state = source.STATES.INGESTING + '_error'\n self.commit()\n return False\n", "role": "user" }, { "content": "ingest_source", "role": "assistant" } ]
def x(self, sources=None, tables=None, clean=False): sources = self._resolve_sources(sources, tables, None, predicate=lambda s: s.is_processable) for source in sources: source.update_table() self.log("Creating source schema for '{}': Table {}, {} columns". format(source.name, source.source_table.name, len(source. source_table.columns))) self.commit()
source_schema
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, sources=None, tables=None, clean=False):\n sources = self._resolve_sources(sources, tables, None, predicate=lambda\n s: s.is_processable)\n for source in sources:\n source.update_table()\n self.log(\"Creating source schema for '{}': Table {}, {} columns\".\n format(source.name, source.source_table.name, len(source.\n source_table.columns)))\n self.commit()\n", "role": "user" }, { "content": "source_schema", "role": "assistant" } ]
def x(self, sources=None, tables=None, clean=False, force=False, use_pipeline=False): from itertools import groupby from operator import attrgetter from ambry.etl import Collect, Head from ambry.orm.exc import NotFoundError self.dstate = self.STATES.BUILDING self.commit() self.log('---- Schema ----') resolved_sources = self._resolve_sources(sources, tables, predicate=lambda s: s.is_processable) if clean: self.dataset.delete_tables_partitions() self.commit() keyfunc = attrgetter('dest_table') for t, table_sources in groupby(sorted(resolved_sources, key=keyfunc), keyfunc): if use_pipeline: for source in table_sources: pl = self.pipeline(source) pl.cast = [ambry.etl.CastSourceColumns] pl.select_partition = [] pl.write = [Head, Collect] pl.final = [] self.log_pipeline(pl) pl.run() pl.phase = 'build_schema' self.log_pipeline(pl) for h, c in zip(pl.write[Collect].headers, pl.write[Collect ].rows[1]): c = t.add_column(name=h, datatype=type(c).__name__ if c is not None else 'str', update_existing=True) self.log("Populated destination table '{}' from pipeline '{}'". format(t.name, pl.name)) else: self.commit() def x(source): if source.is_partition and not source.source_table_exists: return enumerate(source.partition.table.columns) else: return enumerate(source.source_table.columns) columns = sorted(set([(i, col.dest_header, col.datatype, col. description, col.has_codes) for source in table_sources for i, col in source_cols(source)])) initial_count = len(t.columns) for pos, name, datatype, desc, has_codes in columns: kwds = dict(name=name, datatype=datatype, description=desc, update_existing=True) try: extant = t.column(name) except NotFoundError: extant = None if extant is None or not extant.description: kwds['description'] = desc c = t.add_column(**kwds) final_count = len(t.columns) if final_count > initial_count: diff = final_count - initial_count self.log( "Populated destination table '{}' from source table '{}' with {} columns" .format(t.name, source.source_table.name, diff)) self.commit() return True
schema
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, sources=None, tables=None, clean=False, force=False,\n use_pipeline=False):\n from itertools import groupby\n from operator import attrgetter\n from ambry.etl import Collect, Head\n from ambry.orm.exc import NotFoundError\n self.dstate = self.STATES.BUILDING\n self.commit()\n self.log('---- Schema ----')\n resolved_sources = self._resolve_sources(sources, tables, predicate=lambda\n s: s.is_processable)\n if clean:\n self.dataset.delete_tables_partitions()\n self.commit()\n keyfunc = attrgetter('dest_table')\n for t, table_sources in groupby(sorted(resolved_sources, key=keyfunc),\n keyfunc):\n if use_pipeline:\n for source in table_sources:\n pl = self.pipeline(source)\n pl.cast = [ambry.etl.CastSourceColumns]\n pl.select_partition = []\n pl.write = [Head, Collect]\n pl.final = []\n self.log_pipeline(pl)\n pl.run()\n pl.phase = 'build_schema'\n self.log_pipeline(pl)\n for h, c in zip(pl.write[Collect].headers, pl.write[Collect\n ].rows[1]):\n c = t.add_column(name=h, datatype=type(c).__name__ if c\n is not None else 'str', update_existing=True)\n self.log(\"Populated destination table '{}' from pipeline '{}'\".\n format(t.name, pl.name))\n else:\n self.commit()\n\n def x(source):\n if source.is_partition and not source.source_table_exists:\n return enumerate(source.partition.table.columns)\n else:\n return enumerate(source.source_table.columns)\n columns = sorted(set([(i, col.dest_header, col.datatype, col.\n description, col.has_codes) for source in table_sources for\n i, col in source_cols(source)]))\n initial_count = len(t.columns)\n for pos, name, datatype, desc, has_codes in columns:\n kwds = dict(name=name, datatype=datatype, description=desc,\n update_existing=True)\n try:\n extant = t.column(name)\n except NotFoundError:\n extant = None\n if extant is None or not extant.description:\n kwds['description'] = desc\n c = t.add_column(**kwds)\n final_count = len(t.columns)\n if final_count > initial_count:\n diff = final_count - initial_count\n self.log(\n \"Populated destination table '{}' from source table '{}' with {} columns\"\n .format(t.name, source.source_table.name, diff))\n self.commit()\n return True\n", "role": "user" }, { "content": "schema", "role": "assistant" } ]
def x(self, sources): from ambry.orm.exc import NotFoundError for p in self.dataset.partitions: if p.type == p.TYPE.SEGMENT: self.log('Removing old segment partition: {}'.format(p.identity .name)) try: self.wrap_partition(p).local_datafile.remove() self.session.delete(p) except NotFoundError: pass for s in sources: if s.reftype == 'partition': continue p = s.partition if p: try: self.wrap_partition(p).local_datafile.remove() self.session.delete(p) except NotFoundError: pass if s.state in (self.STATES.BUILDING, self.STATES.BUILT): s.state = self.STATES.INGESTED self.commit()
reset_build
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, sources):\n from ambry.orm.exc import NotFoundError\n for p in self.dataset.partitions:\n if p.type == p.TYPE.SEGMENT:\n self.log('Removing old segment partition: {}'.format(p.identity\n .name))\n try:\n self.wrap_partition(p).local_datafile.remove()\n self.session.delete(p)\n except NotFoundError:\n pass\n for s in sources:\n if s.reftype == 'partition':\n continue\n p = s.partition\n if p:\n try:\n self.wrap_partition(p).local_datafile.remove()\n self.session.delete(p)\n except NotFoundError:\n pass\n if s.state in (self.STATES.BUILDING, self.STATES.BUILT):\n s.state = self.STATES.INGESTED\n self.commit()\n", "role": "user" }, { "content": "reset_build", "role": "assistant" } ]
def x(self, sources=None, tables=None, stage=None, force=False): from operator import attrgetter from itertools import groupby from .concurrent import build_mp, unify_mp from ambry.bundle.events import TAG self.log('==== Building ====') self.state = self.STATES.BUILDING self.commit() class SourceSet(object): """Container for sources that can reload them after they get expired from the session""" def x(self, bundle, v): self.bundle = bundle self.sources = v self._s_vids = [s.vid for s in self.sources] def x(self): self.sources = [self.bundle.source(vid) for vid in self._s_vids] def x(self): for s in self.sources: yield s def x(self): return len(self._s_vids) self._run_events(TAG.BEFORE_BUILD, 0) resolved_sources = SourceSet(self, self._resolve_sources(sources, tables, stage=stage, predicate=lambda s: s.is_processable)) with self.progress.start('build', stage, item_total=len(resolved_sources) ) as ps: if len(resolved_sources) == 0: ps.update(message='No sources', state='skipped') self.log('No processable sources, skipping build stage {}'. format(stage)) return True if not self.pre_build(force): ps.update(message='Pre-build failed', state='skipped') return False if force: self._reset_build(resolved_sources) resolved_sources.reload() e = [(stage, SourceSet(self, list(stage_sources))) for stage, stage_sources in groupby(sorted(resolved_sources, key= attrgetter('stage')), attrgetter('stage'))] for stage, stage_sources in e: stage_sources.reload() for s in stage_sources: s.state = self.STATES.WAITING self.commit() stage_sources.reload() self.log('Processing {} sources, stage {} ; first 10: {}'. format(len(stage_sources), stage, [x.name for x in stage_sources.sources[:10]])) self._run_events(TAG.BEFORE_BUILD, stage) if self.multi: try: args = [(self.identity.vid, stage, source.vid, force) for source in stage_sources] pool = self.library.process_pool(limited_run=self. limited_run) r = pool.map_async(build_mp, args, 1) completed_sources = r.get() ps.add( 'Finished MP building {} sources. Starting MP coalescing' .format(len(completed_sources))) partition_names = [(self.identity.vid, k) for k, v in self.collect_segment_partitions().items()] r = pool.map_async(unify_mp, partition_names, 1) completed_partitions = r.get() ps.add('Finished MP coalescing {} partitions'.format( len(completed_partitions))) pool.close() pool.join() except KeyboardInterrupt: self.log('Got keyboard interrrupt; terminating workers') pool.terminate() else: for i, source in enumerate(stage_sources): id_ = ps.add(message='Running source {}'.format(source. name), source=source, item_count=i, state='running') self.build_source(stage, source, ps, force=force) ps.update(message='Finished processing source', state= 'done') ps.get(id_).state = 'done' self.progress.commit() self.unify_partitions() self._run_events(TAG.AFTER_BUILD, stage) self.state = self.STATES.BUILT self.commit() self._run_events(TAG.AFTER_BUILD, 0) self.close_session() self.log('==== Done Building ====') self.buildstate.commit() self.state = self.STATES.BUILT self.commit() return True
build
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, sources=None, tables=None, stage=None, force=False):\n from operator import attrgetter\n from itertools import groupby\n from .concurrent import build_mp, unify_mp\n from ambry.bundle.events import TAG\n self.log('==== Building ====')\n self.state = self.STATES.BUILDING\n self.commit()\n\n\n class SourceSet(object):\n \"\"\"Container for sources that can reload them after they get expired from the session\"\"\"\n\n def x(self, bundle, v):\n self.bundle = bundle\n self.sources = v\n self._s_vids = [s.vid for s in self.sources]\n\n def x(self):\n self.sources = [self.bundle.source(vid) for vid in self._s_vids]\n\n def x(self):\n for s in self.sources:\n yield s\n\n def x(self):\n return len(self._s_vids)\n self._run_events(TAG.BEFORE_BUILD, 0)\n resolved_sources = SourceSet(self, self._resolve_sources(sources,\n tables, stage=stage, predicate=lambda s: s.is_processable))\n with self.progress.start('build', stage, item_total=len(resolved_sources)\n ) as ps:\n if len(resolved_sources) == 0:\n ps.update(message='No sources', state='skipped')\n self.log('No processable sources, skipping build stage {}'.\n format(stage))\n return True\n if not self.pre_build(force):\n ps.update(message='Pre-build failed', state='skipped')\n return False\n if force:\n self._reset_build(resolved_sources)\n resolved_sources.reload()\n e = [(stage, SourceSet(self, list(stage_sources))) for stage,\n stage_sources in groupby(sorted(resolved_sources, key=\n attrgetter('stage')), attrgetter('stage'))]\n for stage, stage_sources in e:\n stage_sources.reload()\n for s in stage_sources:\n s.state = self.STATES.WAITING\n self.commit()\n stage_sources.reload()\n self.log('Processing {} sources, stage {} ; first 10: {}'.\n format(len(stage_sources), stage, [x.name for x in\n stage_sources.sources[:10]]))\n self._run_events(TAG.BEFORE_BUILD, stage)\n if self.multi:\n try:\n args = [(self.identity.vid, stage, source.vid, force) for\n source in stage_sources]\n pool = self.library.process_pool(limited_run=self.\n limited_run)\n r = pool.map_async(build_mp, args, 1)\n completed_sources = r.get()\n ps.add(\n 'Finished MP building {} sources. Starting MP coalescing'\n .format(len(completed_sources)))\n partition_names = [(self.identity.vid, k) for k, v in\n self.collect_segment_partitions().items()]\n r = pool.map_async(unify_mp, partition_names, 1)\n completed_partitions = r.get()\n ps.add('Finished MP coalescing {} partitions'.format(\n len(completed_partitions)))\n pool.close()\n pool.join()\n except KeyboardInterrupt:\n self.log('Got keyboard interrrupt; terminating workers')\n pool.terminate()\n else:\n for i, source in enumerate(stage_sources):\n id_ = ps.add(message='Running source {}'.format(source.\n name), source=source, item_count=i, state='running')\n self.build_source(stage, source, ps, force=force)\n ps.update(message='Finished processing source', state=\n 'done')\n ps.get(id_).state = 'done'\n self.progress.commit()\n self.unify_partitions()\n self._run_events(TAG.AFTER_BUILD, stage)\n self.state = self.STATES.BUILT\n self.commit()\n self._run_events(TAG.AFTER_BUILD, 0)\n self.close_session()\n self.log('==== Done Building ====')\n self.buildstate.commit()\n self.state = self.STATES.BUILT\n self.commit()\n return True\n", "role": "user" }, { "content": "build", "role": "assistant" } ]
def x(self, table, force=False): sources = self._resolve_sources(None, [table]) for source in sources: self.build_source(None, source, force=force) self.unify_partitions()
build_table
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, table, force=False):\n sources = self._resolve_sources(None, [table])\n for source in sources:\n self.build_source(None, source, force=force)\n self.unify_partitions()\n", "role": "user" }, { "content": "build_table", "role": "assistant" } ]
def x(self, stage, source, ps, force=False): from ambry.bundle.process import call_interval assert source.is_processable, source.name if source.state == self.STATES.BUILT and not force: ps.update(message='Source {} already built'.format(source.name), state='skipped') return pl = self.pipeline(source, ps=ps) source.state = self.STATES.BUILDING self.log_pipeline(pl) try: source_name = source.name s_vid = source.vid ps.update(message='Running pipeline {}'.format(pl.name), s_vid= s_vid, item_type='rows', item_count=0) @call_interval(5) def x(sink_pipe, rows): n_records, rate = sink_pipe.report_progress() if n_records > 0: ps.update(message='Running pipeline {}: rate: {}'.format(pl .name, rate), s_vid=s_vid, item_type='rows', item_count =n_records) pl.run(callback=run_progress_f) for f in pl.final: ps.update(message='Run final routine: {}'.format(f.__name__)) f(pl) ps.update(message='Finished building source') except: self.log_pipeline(pl) raise self.commit() try: partitions = list(pl[ambry.etl.PartitionWriter].partitions) ps.update(message='Finalizing segment partition', item_type= 'partitions', item_total=len(partitions), item_count=0) for i, p in enumerate(partitions): ps.update(message='Finalizing segment partition {}'.format(p. name), item_count=i, p_vid=p.vid) try: p.finalize() except AttributeError: print(self.table(p.table_name)) raise self.commit() except IndexError: self.error( "Pipeline didn't have a PartitionWriters, won't try to finalize") self.log_pipeline(pl) source.state = self.STATES.BUILT self.commit() return source.name
build_source
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, stage, source, ps, force=False):\n from ambry.bundle.process import call_interval\n assert source.is_processable, source.name\n if source.state == self.STATES.BUILT and not force:\n ps.update(message='Source {} already built'.format(source.name),\n state='skipped')\n return\n pl = self.pipeline(source, ps=ps)\n source.state = self.STATES.BUILDING\n self.log_pipeline(pl)\n try:\n source_name = source.name\n s_vid = source.vid\n ps.update(message='Running pipeline {}'.format(pl.name), s_vid=\n s_vid, item_type='rows', item_count=0)\n\n @call_interval(5)\n def x(sink_pipe, rows):\n n_records, rate = sink_pipe.report_progress()\n if n_records > 0:\n ps.update(message='Running pipeline {}: rate: {}'.format(pl\n .name, rate), s_vid=s_vid, item_type='rows', item_count\n =n_records)\n pl.run(callback=run_progress_f)\n for f in pl.final:\n ps.update(message='Run final routine: {}'.format(f.__name__))\n f(pl)\n ps.update(message='Finished building source')\n except:\n self.log_pipeline(pl)\n raise\n self.commit()\n try:\n partitions = list(pl[ambry.etl.PartitionWriter].partitions)\n ps.update(message='Finalizing segment partition', item_type=\n 'partitions', item_total=len(partitions), item_count=0)\n for i, p in enumerate(partitions):\n ps.update(message='Finalizing segment partition {}'.format(p.\n name), item_count=i, p_vid=p.vid)\n try:\n p.finalize()\n except AttributeError:\n print(self.table(p.table_name))\n raise\n self.commit()\n except IndexError:\n self.error(\n \"Pipeline didn't have a PartitionWriters, won't try to finalize\")\n self.log_pipeline(pl)\n source.state = self.STATES.BUILT\n self.commit()\n return source.name\n", "role": "user" }, { "content": "build_source", "role": "assistant" } ]
def x(self): from collections import defaultdict partitions = defaultdict(set) for p in self.dataset.partitions: if p.type == p.TYPE.SEGMENT: name = p.identity.name name.segment = None partitions[name].add(p) return partitions
collect_segment_partitions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from collections import defaultdict\n partitions = defaultdict(set)\n for p in self.dataset.partitions:\n if p.type == p.TYPE.SEGMENT:\n name = p.identity.name\n name.segment = None\n partitions[name].add(p)\n return partitions\n", "role": "user" }, { "content": "collect_segment_partitions", "role": "assistant" } ]
def x(self): partitions = self.collect_segment_partitions() with self.progress.start('coalesce', 0, message= 'Coalescing partition segments') as ps: for name, segments in iteritems(partitions): ps.add(item_type='partitions', item_count=len(segments), message='Colescing partition {}'.format(name)) self.unify_partition(name, segments, ps)
unify_partitions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n partitions = self.collect_segment_partitions()\n with self.progress.start('coalesce', 0, message=\n 'Coalescing partition segments') as ps:\n for name, segments in iteritems(partitions):\n ps.add(item_type='partitions', item_count=len(segments),\n message='Colescing partition {}'.format(name))\n self.unify_partition(name, segments, ps)\n", "role": "user" }, { "content": "unify_partitions", "role": "assistant" } ]
def x(self, **kwargs): import inspect import dateutil.parser import datetime import random from functools import partial from ambry.valuetype.types import parse_date, parse_time, parse_datetime import ambry.valuetype.types import ambry.valuetype.exceptions import ambry.valuetype.test import ambry.valuetype def x(f, frm): try: try: f.ambry_from = frm except AttributeError: f.im_func.ambry_from = frm except (TypeError, AttributeError): pass return f test_env = dict(parse_date=parse_date, parse_time=parse_time, parse_datetime=parse_datetime, partial=partial, bundle=self) test_env.update(kwargs) test_env.update(dateutil.parser.__dict__) test_env.update(datetime.__dict__) test_env.update(random.__dict__) test_env.update(ambry.valuetype.core.__dict__) test_env.update(ambry.valuetype.types.__dict__) test_env.update(ambry.valuetype.exceptions.__dict__) test_env.update(ambry.valuetype.test.__dict__) test_env.update(ambry.valuetype.__dict__) localvars = {} for f_name, func in test_env.items(): if not isinstance(func, (str, tuple)): localvars[f_name] = set_from(func, 'env') localvars['randint'] = lambda a, b: random.randint(a, b) if self != Bundle: base = set(inspect.getmembers(Bundle, predicate=inspect.isfunction)) mine = set(inspect.getmembers(self.__class__, predicate=inspect. isfunction)) localvars.update({f_name: set_from(func, 'bundle') for f_name, func in mine - base}) base = set(inspect.getmembers(Bundle, predicate=inspect.ismethod)) mine = set(inspect.getmembers(self.__class__, predicate=inspect. ismethod)) localvars.update({f_name: set_from(func.__get__(self), 'bundle') for f_name, func in mine - base}) module_entries = inspect.getmembers(sys.modules['ambry.build'], predicate=inspect.isfunction) localvars.update({f_name: set_from(func, 'module') for f_name, func in module_entries}) return localvars
exec_context
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n import inspect\n import dateutil.parser\n import datetime\n import random\n from functools import partial\n from ambry.valuetype.types import parse_date, parse_time, parse_datetime\n import ambry.valuetype.types\n import ambry.valuetype.exceptions\n import ambry.valuetype.test\n import ambry.valuetype\n\n def x(f, frm):\n try:\n try:\n f.ambry_from = frm\n except AttributeError:\n f.im_func.ambry_from = frm\n except (TypeError, AttributeError):\n pass\n return f\n test_env = dict(parse_date=parse_date, parse_time=parse_time,\n parse_datetime=parse_datetime, partial=partial, bundle=self)\n test_env.update(kwargs)\n test_env.update(dateutil.parser.__dict__)\n test_env.update(datetime.__dict__)\n test_env.update(random.__dict__)\n test_env.update(ambry.valuetype.core.__dict__)\n test_env.update(ambry.valuetype.types.__dict__)\n test_env.update(ambry.valuetype.exceptions.__dict__)\n test_env.update(ambry.valuetype.test.__dict__)\n test_env.update(ambry.valuetype.__dict__)\n localvars = {}\n for f_name, func in test_env.items():\n if not isinstance(func, (str, tuple)):\n localvars[f_name] = set_from(func, 'env')\n localvars['randint'] = lambda a, b: random.randint(a, b)\n if self != Bundle:\n base = set(inspect.getmembers(Bundle, predicate=inspect.isfunction))\n mine = set(inspect.getmembers(self.__class__, predicate=inspect.\n isfunction))\n localvars.update({f_name: set_from(func, 'bundle') for f_name, func in\n mine - base})\n base = set(inspect.getmembers(Bundle, predicate=inspect.ismethod))\n mine = set(inspect.getmembers(self.__class__, predicate=inspect.\n ismethod))\n localvars.update({f_name: set_from(func.__get__(self), 'bundle') for\n f_name, func in mine - base})\n module_entries = inspect.getmembers(sys.modules['ambry.build'],\n predicate=inspect.isfunction)\n localvars.update({f_name: set_from(func, 'module') for f_name, func in\n module_entries})\n return localvars\n", "role": "user" }, { "content": "exec_context", "role": "assistant" } ]
def x(self): from ambry.util.datestimes import expand_to_years years = set() if self.metadata.about.time: for year in expand_to_years(self.metadata.about.time): years.add(year) if self.identity.btime: for year in expand_to_years(self.identity.btime): years.add(year) for p in self.partitions: years |= set(p.time_coverage)
post_build_time_coverage
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry.util.datestimes import expand_to_years\n years = set()\n if self.metadata.about.time:\n for year in expand_to_years(self.metadata.about.time):\n years.add(year)\n if self.identity.btime:\n for year in expand_to_years(self.identity.btime):\n years.add(year)\n for p in self.partitions:\n years |= set(p.time_coverage)\n", "role": "user" }, { "content": "post_build_time_coverage", "role": "assistant" } ]
def x(self): spaces = set() grains = set() def x(term): places = list(self.library.search.search_identifiers(term)) if not places: raise BuildError( "Failed to find space identifier '{}' in full text identifier search" .format(term)) return places[0].vid if self.metadata.about.space: spaces.add(resolve(self.metadata.about.space)) if self.metadata.about.grain: grains.add(self.metadata.about.grain) if self.identity.bspace: spaces.add(resolve(self.identity.bspace)) for p in self.partitions.all: if 'geo_coverage' in p.record.data: for space in p.record.data['geo_coverage']: spaces.add(space) if 'geo_grain' in p.record.data: for grain in p.record.data['geo_grain']: grains.add(grain) def x(g): try: c = GVid.get_class(g) return b(c().summarize()) except NotASummaryName: return g self.metadata.coverage.geo = sorted(spaces) self.metadata.coverage.grain = sorted(conv_grain(g) for g in grains) self.metadata.write_to_dir()
post_build_geo_coverage
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n spaces = set()\n grains = set()\n\n def x(term):\n places = list(self.library.search.search_identifiers(term))\n if not places:\n raise BuildError(\n \"Failed to find space identifier '{}' in full text identifier search\"\n .format(term))\n return places[0].vid\n if self.metadata.about.space:\n spaces.add(resolve(self.metadata.about.space))\n if self.metadata.about.grain:\n grains.add(self.metadata.about.grain)\n if self.identity.bspace:\n spaces.add(resolve(self.identity.bspace))\n for p in self.partitions.all:\n if 'geo_coverage' in p.record.data:\n for space in p.record.data['geo_coverage']:\n spaces.add(space)\n if 'geo_grain' in p.record.data:\n for grain in p.record.data['geo_grain']:\n grains.add(grain)\n\n def x(g):\n try:\n c = GVid.get_class(g)\n return b(c().summarize())\n except NotASummaryName:\n return g\n self.metadata.coverage.geo = sorted(spaces)\n self.metadata.coverage.grain = sorted(conv_grain(g) for g in grains)\n self.metadata.write_to_dir()\n", "role": "user" }, { "content": "post_build_geo_coverage", "role": "assistant" } ]
def x(self): return (self.state == self.STATES.FINALIZED or self.state == self. STATES.INSTALLED)
is_finalized
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return (self.state == self.STATES.FINALIZED or self.state == self.\n STATES.INSTALLED)\n", "role": "user" }, { "content": "is_finalized", "role": "assistant" } ]
def x(self, tag, stage=None): self._run_event_methods(tag, stage) self._run_tests(tag, stage)
run_events
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tag, stage=None):\n self._run_event_methods(tag, stage)\n self._run_tests(tag, stage)\n", "role": "user" }, { "content": "run_events", "role": "assistant" } ]
def x(self, tag, stage=None): import inspect from ambry.bundle.events import _runable_for_event funcs = [] for func_name, f in inspect.getmembers(self, predicate=inspect.ismethod): if _runable_for_event(f, tag, stage): funcs.append(f) for func in funcs: func()
run_event_methods
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tag, stage=None):\n import inspect\n from ambry.bundle.events import _runable_for_event\n funcs = []\n for func_name, f in inspect.getmembers(self, predicate=inspect.ismethod):\n if _runable_for_event(f, tag, stage):\n funcs.append(f)\n for func in funcs:\n func()\n", "role": "user" }, { "content": "run_event_methods", "role": "assistant" } ]
def x(self): r = self.library.resolve(self.identity.vid) return r is not None
is_installed
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n r = self.library.resolve(self.identity.vid)\n return r is not None\n", "role": "user" }, { "content": "is_installed", "role": "assistant" } ]
def x(prop): if isinstance(prop, QueryableAttribute): prop = prop.property assert isinstance(prop, (Column, ColumnProperty, RelationshipProperty)) _included.add(prop)
include
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(prop):\n if isinstance(prop, QueryableAttribute):\n prop = prop.property\n assert isinstance(prop, (Column, ColumnProperty, RelationshipProperty))\n _included.add(prop)\n", "role": "user" }, { "content": "include", "role": "assistant" } ]
def x(prop): if isinstance(prop, QueryableAttribute): prop = prop.property assert isinstance(prop, (Column, ColumnProperty, RelationshipProperty)) _excluded.add(prop) if isinstance(prop, RelationshipProperty): for local in prop.local_columns: _excluded.add(local)
exclude
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(prop):\n if isinstance(prop, QueryableAttribute):\n prop = prop.property\n assert isinstance(prop, (Column, ColumnProperty, RelationshipProperty))\n _excluded.add(prop)\n if isinstance(prop, RelationshipProperty):\n for local in prop.local_columns:\n _excluded.add(local)\n", "role": "user" }, { "content": "exclude", "role": "assistant" } ]
def x(source, model, cache=None): if source is None: return None if cache and source in cache: return cache[source] db = object_session(source) ident = identity_key(instance=source)[1] assert ident is not None return db.query(model).get(ident)
reflect
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(source, model, cache=None):\n if source is None:\n return None\n if cache and source in cache:\n return cache[source]\n db = object_session(source)\n ident = identity_key(instance=source)[1]\n assert ident is not None\n return db.query(model).get(ident)\n", "role": "user" }, { "content": "reflect", "role": "assistant" } ]
def x(source, target, cache=None): target_manager = manager_of_class(type(target)) column_attrs = set() relationship_attrs = set() relationship_columns = set() for attr in manager_of_class(type(source)).attributes: if attr.key not in target_manager: continue target_attr = target_manager[attr.key] if isinstance(attr.property, ColumnProperty): assert isinstance(target_attr.property, ColumnProperty) column_attrs.add(attr) elif isinstance(attr.property, RelationshipProperty): assert isinstance(target_attr.property, RelationshipProperty) relationship_attrs.add(attr) if attr.property.direction is MANYTOONE: relationship_columns.update(attr.property.local_columns) for attr in column_attrs: if _column_property_in_registry(attr.property, _excluded): continue elif not _column_property_in_registry(attr.property, _included ) and all(column in relationship_columns for column in attr. property.columns): continue setattr(target, attr.key, getattr(source, attr.key)) for attr in relationship_attrs: target_attr_model = target_manager[attr.key].property.argument if not is_relation_replicatable(attr): continue replicate_relation(source, target, attr, target_manager[attr.key], cache=cache)
replicate_attributes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(source, target, cache=None):\n target_manager = manager_of_class(type(target))\n column_attrs = set()\n relationship_attrs = set()\n relationship_columns = set()\n for attr in manager_of_class(type(source)).attributes:\n if attr.key not in target_manager:\n continue\n target_attr = target_manager[attr.key]\n if isinstance(attr.property, ColumnProperty):\n assert isinstance(target_attr.property, ColumnProperty)\n column_attrs.add(attr)\n elif isinstance(attr.property, RelationshipProperty):\n assert isinstance(target_attr.property, RelationshipProperty)\n relationship_attrs.add(attr)\n if attr.property.direction is MANYTOONE:\n relationship_columns.update(attr.property.local_columns)\n for attr in column_attrs:\n if _column_property_in_registry(attr.property, _excluded):\n continue\n elif not _column_property_in_registry(attr.property, _included\n ) and all(column in relationship_columns for column in attr.\n property.columns):\n continue\n setattr(target, attr.key, getattr(source, attr.key))\n for attr in relationship_attrs:\n target_attr_model = target_manager[attr.key].property.argument\n if not is_relation_replicatable(attr):\n continue\n replicate_relation(source, target, attr, target_manager[attr.key],\n cache=cache)\n", "role": "user" }, { "content": "replicate_attributes", "role": "assistant" } ]
def x(source, model, cache=None): if source is None: return None if cache is None: cache = {} elif source in cache: return cache[source] db = object_session(source) cls, ident = identity_key(instance=source) target = db.query(model).get(ident) if target is None: target = model() cache[source] = target try: replicate_attributes(source, target, cache=cache) except _PrimaryKeyIsNull: return None else: return target
replicate_no_merge
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(source, model, cache=None):\n if source is None:\n return None\n if cache is None:\n cache = {}\n elif source in cache:\n return cache[source]\n db = object_session(source)\n cls, ident = identity_key(instance=source)\n target = db.query(model).get(ident)\n if target is None:\n target = model()\n cache[source] = target\n try:\n replicate_attributes(source, target, cache=cache)\n except _PrimaryKeyIsNull:\n return None\n else:\n return target\n", "role": "user" }, { "content": "replicate_no_merge", "role": "assistant" } ]
def x(source, model, cache=None): target = replicate_no_merge(source, model, cache=cache) if target is not None: db = object_session(source) target = db.merge(target) return target
replicate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(source, model, cache=None):\n target = replicate_no_merge(source, model, cache=cache)\n if target is not None:\n db = object_session(source)\n target = db.merge(target)\n return target\n", "role": "user" }, { "content": "replicate", "role": "assistant" } ]
def x(sources, model, cache=None): targets = [replicate_no_merge(source, model, cache=cache) for source in sources] return [target for target in targets if target is not None]
replicate_filter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sources, model, cache=None):\n targets = [replicate_no_merge(source, model, cache=cache) for source in\n sources]\n return [target for target in targets if target is not None]\n", "role": "user" }, { "content": "replicate_filter", "role": "assistant" } ]
def x(sources, model, cache=None): targets = [reflect(source, model, cache=cache) for source in sources] return [target for target in targets if target is not None]
reflect_filter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sources, model, cache=None):\n targets = [reflect(source, model, cache=cache) for source in sources]\n return [target for target in targets if target is not None]\n", "role": "user" }, { "content": "reflect_filter", "role": "assistant" } ]
def x(self): from ambry.valuetype import resolve_value_type if self.valuetype: return resolve_value_type(self.valuetype) else: return resolve_value_type(self.datatype)
valuetype_class
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry.valuetype import resolve_value_type\n if self.valuetype:\n return resolve_value_type(self.valuetype)\n else:\n return resolve_value_type(self.datatype)\n", "role": "user" }, { "content": "valuetype_class", "role": "assistant" } ]
def x(self): from ambry.valuetype import resolve_value_type if self.valuetype and resolve_value_type(self.valuetype): return resolve_value_type(self.valuetype)._pythontype elif self.datatype: try: return self.types[self.datatype][1] except KeyError: return resolve_value_type(self.datatype)._pythontype else: from ambry.exc import ConfigurationError raise ConfigurationError( "Can't get python_type: neither datatype of valuetype is defined")
python_type
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry.valuetype import resolve_value_type\n if self.valuetype and resolve_value_type(self.valuetype):\n return resolve_value_type(self.valuetype)._pythontype\n elif self.datatype:\n try:\n return self.types[self.datatype][1]\n except KeyError:\n return resolve_value_type(self.datatype)._pythontype\n else:\n from ambry.exc import ConfigurationError\n raise ConfigurationError(\n \"Can't get python_type: neither datatype of valuetype is defined\")\n", "role": "user" }, { "content": "python_type", "role": "assistant" } ]
def x(self): from ambry.valuetype.core import ROLE if not self.valuetype_class: return '' role = self.valuetype_class.role if role == ROLE.UNKNOWN: vt_code = self.valuetype_class.vt_code if len(vt_code) == 1 or vt_code[1] == '/': return vt_code[0] else: return '' return role
role
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry.valuetype.core import ROLE\n if not self.valuetype_class:\n return ''\n role = self.valuetype_class.role\n if role == ROLE.UNKNOWN:\n vt_code = self.valuetype_class.vt_code\n if len(vt_code) == 1 or vt_code[1] == '/':\n return vt_code[0]\n else:\n return ''\n return role\n", "role": "user" }, { "content": "role", "role": "assistant" } ]
def x(self): from ambry.valuetype.core import ROLE return self.role == ROLE.DIMENSION
is_dimension
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry.valuetype.core import ROLE\n return self.role == ROLE.DIMENSION\n", "role": "user" }, { "content": "is_dimension", "role": "assistant" } ]
def x(self): from ambry.valuetype.core import ROLE return self.role == ROLE.MEASURE
is_measure
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry.valuetype.core import ROLE\n return self.role == ROLE.MEASURE\n", "role": "user" }, { "content": "is_measure", "role": "assistant" } ]
def x(self): from ambry.valuetype.core import ROLE return self.role == ROLE.LABEL
is_label
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry.valuetype.core import ROLE\n return self.role == ROLE.LABEL\n", "role": "user" }, { "content": "is_label", "role": "assistant" } ]
def x(self): from ambry.valuetype.core import ROLE return self.role == ROLE.ERROR
is_error
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry.valuetype.core import ROLE\n return self.role == ROLE.ERROR\n", "role": "user" }, { "content": "is_error", "role": "assistant" } ]
def x(self): for c in self.table.columns: if c.parent == self.name and not c.valuetype_class.is_label(): yield c
children
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for c in self.table.columns:\n if c.parent == self.name and not c.valuetype_class.is_label():\n yield c\n", "role": "user" }, { "content": "children", "role": "assistant" } ]
def x(self): if self.valuetype_class.is_label(): return self for c in self.table.columns: if c.parent == self.name and c.valuetype_class.is_label(): return c return None
label
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.valuetype_class.is_label():\n return self\n for c in self.table.columns:\n if c.parent == self.name and c.valuetype_class.is_label():\n return c\n return None\n", "role": "user" }, { "content": "label", "role": "assistant" } ]
def x(self): if self.valuetype_class.is_geoid(): return self for c in self.table.columns: if c.parent == self.name and c.valuetype_class.is_geoid(): return c
geoid
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.valuetype_class.is_geoid():\n return self\n for c in self.table.columns:\n if c.parent == self.name and c.valuetype_class.is_geoid():\n return c\n", "role": "user" }, { "content": "geoid", "role": "assistant" } ]
def x(self, v): if self.type_is_time(): dt = dateutil.parser.parse(v) if self.datatype == Column.DATATYPE_TIME: dt = dt.time() if not isinstance(dt, self.python_type): raise TypeError('{} was parsed to {}, expected {}'.format(v, type(dt), self.python_type)) return dt else: return self.python_type(v)
python_cast
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, v):\n if self.type_is_time():\n dt = dateutil.parser.parse(v)\n if self.datatype == Column.DATATYPE_TIME:\n dt = dt.time()\n if not isinstance(dt, self.python_type):\n raise TypeError('{} was parsed to {}, expected {}'.format(v,\n type(dt), self.python_type))\n return dt\n else:\n return self.python_type(v)\n", "role": "user" }, { "content": "python_cast", "role": "assistant" } ]
def x(cls, dtype): m = {'int64': cls.DATATYPE_INTEGER64, 'float64': cls.DATATYPE_FLOAT, 'object': cls.DATATYPE_TEXT} t = m.get(dtype.name, None) if not t: raise TypeError("Failed to convert numpy type: '{}' ".format(dtype. name)) return t
convert_numpy_type
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, dtype):\n m = {'int64': cls.DATATYPE_INTEGER64, 'float64': cls.DATATYPE_FLOAT,\n 'object': cls.DATATYPE_TEXT}\n t = m.get(dtype.name, None)\n if not t:\n raise TypeError(\"Failed to convert numpy type: '{}' \".format(dtype.\n name))\n return t\n", "role": "user" }, { "content": "convert_numpy_type", "role": "assistant" } ]
def x(self): d = {p.key: getattr(self, p.key) for p in self.__mapper__.attrs if p. key not in ('table', 'stats', '_codes', 'data')} if not d: raise Exception(self.__dict__) d['schema_type'] = self.schema_type if self.data: for k, v in six.iteritems(self.data): if k not in d and k not in ('table', 'stats', '_codes', 'data'): d[k] = v return d
dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n d = {p.key: getattr(self, p.key) for p in self.__mapper__.attrs if p.\n key not in ('table', 'stats', '_codes', 'data')}\n if not d:\n raise Exception(self.__dict__)\n d['schema_type'] = self.schema_type\n if self.data:\n for k, v in six.iteritems(self.data):\n if k not in d and k not in ('table', 'stats', '_codes', 'data'):\n d[k] = v\n return d\n", "role": "user" }, { "content": "dict", "role": "assistant" } ]
def x(self): return {k: v for k, v in six.iteritems(self.dict) if v and k != '_codes'}
nonull_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return {k: v for k, v in six.iteritems(self.dict) if v and k != '_codes'}\n", "role": "user" }, { "content": "nonull_dict", "role": "assistant" } ]
def x(name): import re try: return re.sub('_+', '_', re.sub('[^\\w_]', '_', name).lower()).rstrip( '_') except TypeError: raise TypeError('Trying to mangle name with invalid type of: ' + str(type(name)))
mangle_name
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(name):\n import re\n try:\n return re.sub('_+', '_', re.sub('[^\\\\w_]', '_', name).lower()).rstrip(\n '_')\n except TypeError:\n raise TypeError('Trying to mangle name with invalid type of: ' +\n str(type(name)))\n", "role": "user" }, { "content": "mangle_name", "role": "assistant" } ]
def x(self): return {c.value: (c.ikey if c.ikey else c.key) for c in self.codes}
reverse_code_map
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return {c.value: (c.ikey if c.ikey else c.key) for c in self.codes}\n", "role": "user" }, { "content": "reverse_code_map", "role": "assistant" } ]
def x(self): segments = self._expand_transform(self.transform) if segments: segments[0]['datatype'] = self.valuetype_class for s in segments: s['column'] = self else: segments = [self.make_xform_seg(datatype=self.valuetype_class, column=self)] return segments
expanded_transform
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n segments = self._expand_transform(self.transform)\n if segments:\n segments[0]['datatype'] = self.valuetype_class\n for s in segments:\n s['column'] = self\n else:\n segments = [self.make_xform_seg(datatype=self.valuetype_class,\n column=self)]\n return segments\n", "role": "user" }, { "content": "expanded_transform", "role": "assistant" } ]
def x(mapper, conn, target): if target.sequence_id is None: from ambry.orm.exc import DatabaseError raise DatabaseError('Must have sequence_id before insertion') assert (target.name == 'id') == (target.sequence_id == 1), (target.name, target.sequence_id) Column.before_update(mapper, conn, target)
before_insert
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(mapper, conn, target):\n if target.sequence_id is None:\n from ambry.orm.exc import DatabaseError\n raise DatabaseError('Must have sequence_id before insertion')\n assert (target.name == 'id') == (target.sequence_id == 1), (target.name,\n target.sequence_id)\n Column.before_update(mapper, conn, target)\n", "role": "user" }, { "content": "before_insert", "role": "assistant" } ]
def x(mapper, conn, target): assert target.datatype or target.valuetype target.name = Column.mangle_name(target.name) Column.update_number(target)
before_update
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(mapper, conn, target):\n assert target.datatype or target.valuetype\n target.name = Column.mangle_name(target.name)\n Column.update_number(target)\n", "role": "user" }, { "content": "before_update", "role": "assistant" } ]
def x(processProtocol, executable, args=(), env={}, path=None, uid=None, gid=None, usePTY=0, packages=()): env = env.copy() pythonpath = [] for pkg in packages: p = os.path.split(imp.find_module(pkg)[1])[0] if p.startswith(os.path.join(sys.prefix, 'lib')): continue pythonpath.append(p) pythonpath = list(set(pythonpath)) pythonpath.extend(env.get('PYTHONPATH', '').split(os.pathsep)) env['PYTHONPATH'] = os.pathsep.join(pythonpath) return reactor.spawnProcess(processProtocol, executable, args, env, path, uid, gid, usePTY)
spawnProcess
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(processProtocol, executable, args=(), env={}, path=None, uid=None,\n gid=None, usePTY=0, packages=()):\n env = env.copy()\n pythonpath = []\n for pkg in packages:\n p = os.path.split(imp.find_module(pkg)[1])[0]\n if p.startswith(os.path.join(sys.prefix, 'lib')):\n continue\n pythonpath.append(p)\n pythonpath = list(set(pythonpath))\n pythonpath.extend(env.get('PYTHONPATH', '').split(os.pathsep))\n env['PYTHONPATH'] = os.pathsep.join(pythonpath)\n return reactor.spawnProcess(processProtocol, executable, args, env,\n path, uid, gid, usePTY)\n", "role": "user" }, { "content": "spawnProcess", "role": "assistant" } ]
def x(processProtocol, args=(), env={}, path=None, uid=None, gid=None, usePTY=0, packages=()): return spawnProcess(processProtocol, sys.executable, args, env, path, uid, gid, usePTY, packages)
spawnPythonProcess
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(processProtocol, args=(), env={}, path=None, uid=None, gid=None,\n usePTY=0, packages=()):\n return spawnProcess(processProtocol, sys.executable, args, env, path,\n uid, gid, usePTY, packages)\n", "role": "user" }, { "content": "spawnPythonProcess", "role": "assistant" } ]
def x(f, tag, stage): if not hasattr(f, '__ambry_event__'): return False f_tag, f_stage = f.__ambry_event__ if stage is None: stage = 0 if tag != f_tag or stage != f_stage: return False return True
runable_for_event
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(f, tag, stage):\n if not hasattr(f, '__ambry_event__'):\n return False\n f_tag, f_stage = f.__ambry_event__\n if stage is None:\n stage = 0\n if tag != f_tag or stage != f_stage:\n return False\n return True\n", "role": "user" }, { "content": "runable_for_event", "role": "assistant" } ]
def x(import_path, prefix=None, ld=dict()): if prefix and not import_path.startswith(prefix): import_path = '.'.join([prefix, import_path]) log.debug('attempting to load a python object from an import path', extra=dict(import_path=import_path, **ld)) try: mod = importlib.import_module(import_path) return mod except: pass try: path, obj_name = import_path.rsplit('.', 1) except ValueError: log_raise( 'import path needs at least 1 period in your import path. An example import path is something like: module.obj' , dict(import_path=import_path, **ld), InvalidImportPath) try: mod = importlib.import_module(path) except ImportError: newpath = path.replace(prefix, '', 1).lstrip('.') log.debug('Could not load import path. Trying a different one', extra=dict(oldpath=path, newpath=newpath)) path = newpath mod = importlib.import_module(path) try: obj = getattr(mod, obj_name) except AttributeError: log_raise( 'object does not exist in given module. Your import path is not properly defined because the given `obj_name` does not exist' , dict(import_path=path, obj_name=obj_name, **ld), InvalidImportPath) return obj
load_obj_from_path
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(import_path, prefix=None, ld=dict()):\n if prefix and not import_path.startswith(prefix):\n import_path = '.'.join([prefix, import_path])\n log.debug('attempting to load a python object from an import path',\n extra=dict(import_path=import_path, **ld))\n try:\n mod = importlib.import_module(import_path)\n return mod\n except:\n pass\n try:\n path, obj_name = import_path.rsplit('.', 1)\n except ValueError:\n log_raise(\n 'import path needs at least 1 period in your import path. An example import path is something like: module.obj'\n , dict(import_path=import_path, **ld), InvalidImportPath)\n try:\n mod = importlib.import_module(path)\n except ImportError:\n newpath = path.replace(prefix, '', 1).lstrip('.')\n log.debug('Could not load import path. Trying a different one',\n extra=dict(oldpath=path, newpath=newpath))\n path = newpath\n mod = importlib.import_module(path)\n try:\n obj = getattr(mod, obj_name)\n except AttributeError:\n log_raise(\n 'object does not exist in given module. Your import path is not properly defined because the given `obj_name` does not exist'\n , dict(import_path=path, obj_name=obj_name, **ld),\n InvalidImportPath)\n return obj\n", "role": "user" }, { "content": "load_obj_from_path", "role": "assistant" } ]
def x(self, artifact_quality): allowed_values = ['NEW', 'VERIFIED', 'TESTED', 'DEPRECATED', 'BLACKLISTED', 'DELETED', 'TEMPORARY'] if artifact_quality not in allowed_values: raise ValueError( 'Invalid value for `artifact_quality` ({0}), must be one of {1}' .format(artifact_quality, allowed_values)) self._artifact_quality = artifact_quality
artifact_quality
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, artifact_quality):\n allowed_values = ['NEW', 'VERIFIED', 'TESTED', 'DEPRECATED',\n 'BLACKLISTED', 'DELETED', 'TEMPORARY']\n if artifact_quality not in allowed_values:\n raise ValueError(\n 'Invalid value for `artifact_quality` ({0}), must be one of {1}'\n .format(artifact_quality, allowed_values))\n self._artifact_quality = artifact_quality\n", "role": "user" }, { "content": "artifact_quality", "role": "assistant" } ]
def x(connection, table): SQL_TO_PYTHON_TYPES = {'INT': int, 'INTEGER': int, 'TINYINT': int, 'SMALLINT': int, 'MEDIUMINT': int, 'BIGINT': int, 'UNSIGNED BIG INT': int, 'INT': int, 'INT8': int, 'NUMERIC': float, 'REAL': float, 'FLOAT': float, 'DOUBLE': float, 'BOOLEAN': bool, 'CHARACTER': str, 'VARCHAR': str, 'TEXT': str} query = "PRAGMA table_info('{}');" result = connection.execute(query.format(table)) ret = [] for row in result: position = row[0] + 1 name = row[1] datatype = row[2] try: datatype = SQL_TO_PYTHON_TYPES[datatype] except KeyError: raise Exception( 'Do not know how to convert {} sql datatype to python data type.' .format(datatype)) ret.append((name, datatype, position)) return ret
get_sqlite_columns
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(connection, table):\n SQL_TO_PYTHON_TYPES = {'INT': int, 'INTEGER': int, 'TINYINT': int,\n 'SMALLINT': int, 'MEDIUMINT': int, 'BIGINT': int,\n 'UNSIGNED BIG INT': int, 'INT': int, 'INT8': int, 'NUMERIC': float,\n 'REAL': float, 'FLOAT': float, 'DOUBLE': float, 'BOOLEAN': bool,\n 'CHARACTER': str, 'VARCHAR': str, 'TEXT': str}\n query = \"PRAGMA table_info('{}');\"\n result = connection.execute(query.format(table))\n ret = []\n for row in result:\n position = row[0] + 1\n name = row[1]\n datatype = row[2]\n try:\n datatype = SQL_TO_PYTHON_TYPES[datatype]\n except KeyError:\n raise Exception(\n 'Do not know how to convert {} sql datatype to python data type.'\n .format(datatype))\n ret.append((name, datatype, position))\n return ret\n", "role": "user" }, { "content": "get_sqlite_columns", "role": "assistant" } ]
def x(self): if self.urltype != 'partition': return None return self._bundle.library.partition(self.url)
partition
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.urltype != 'partition':\n return None\n return self._bundle.library.partition(self.url)\n", "role": "user" }, { "content": "partition", "role": "assistant" } ]
def x(self): from ambry_sources import MPRowsFile if self._datafile is None: if self.urltype == 'partition': self._datafile = self.partition.datafile else: self._datafile = MPRowsFile(self._bundle.build_ingest_fs, self.name ) return self._datafile
datafile
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry_sources import MPRowsFile\n if self._datafile is None:\n if self.urltype == 'partition':\n self._datafile = self.partition.datafile\n else:\n self._datafile = MPRowsFile(self._bundle.build_ingest_fs, self.name\n )\n return self._datafile\n", "role": "user" }, { "content": "datafile", "role": "assistant" } ]
def x(self): from ambry_sources.sources import SourceSpec d = self.dict d['url'] = self.url return SourceSpec(**d)
spec
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry_sources.sources import SourceSpec\n d = self.dict\n d['url'] = self.url\n return SourceSpec(**d)\n", "role": "user" }, { "content": "spec", "role": "assistant" } ]
def x(self): from ambry.util import parse_url_to_dict d = parse_url_to_dict(self.url) return self._bundle.library.account(d['netloc'])
account
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry.util import parse_url_to_dict\n d = parse_url_to_dict(self.url)\n return self._bundle.library.account(d['netloc'])\n", "role": "user" }, { "content": "account", "role": "assistant" } ]
def x(self, unknown_type='str'): from ambry_sources.intuit import TypeIntuiter st = self.source_table if self.reftype == 'partition': for c in self.partition.table.columns: st.add_column(c.sequence_id, source_header=c.name, dest_header= c.name, datatype=c.datatype, description=c.description) elif self.datafile.exists: with self.datafile.reader as r: names = set() for col in r.columns: name = col['name'] if name in names: name = name + '_' + str(col['pos']) names.add(name) c = st.column(name) dt = col['resolved_type'] if col['resolved_type' ] != 'unknown' else unknown_type if c: c.datatype = TypeIntuiter.promote_type(c.datatype, col[ 'resolved_type']) else: c = st.add_column(col['pos'], source_header=name, dest_header=name, datatype=col['resolved_type'], description=col['description'], has_codes=col[ 'has_codes'])
update_table
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, unknown_type='str'):\n from ambry_sources.intuit import TypeIntuiter\n st = self.source_table\n if self.reftype == 'partition':\n for c in self.partition.table.columns:\n st.add_column(c.sequence_id, source_header=c.name, dest_header=\n c.name, datatype=c.datatype, description=c.description)\n elif self.datafile.exists:\n with self.datafile.reader as r:\n names = set()\n for col in r.columns:\n name = col['name']\n if name in names:\n name = name + '_' + str(col['pos'])\n names.add(name)\n c = st.column(name)\n dt = col['resolved_type'] if col['resolved_type'\n ] != 'unknown' else unknown_type\n if c:\n c.datatype = TypeIntuiter.promote_type(c.datatype, col[\n 'resolved_type'])\n else:\n c = st.add_column(col['pos'], source_header=name,\n dest_header=name, datatype=col['resolved_type'],\n description=col['description'], has_codes=col[\n 'has_codes'])\n", "role": "user" }, { "content": "update_table", "role": "assistant" } ]
def x(self): if self.datafile.exists: with self.datafile.reader as r: self.header_lines = r.info['header_rows'] self.comment_lines = r.info['comment_rows'] self.start_line = r.info['data_start_row'] self.end_line = r.info['data_end_row']
update_spec
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.datafile.exists:\n with self.datafile.reader as r:\n self.header_lines = r.info['header_rows']\n self.comment_lines = r.info['comment_rows']\n self.start_line = r.info['data_start_row']\n self.end_line = r.info['data_end_row']\n", "role": "user" }, { "content": "update_spec", "role": "assistant" } ]
def x(self): SKIP_KEYS = ('_source_table', '_dest_table', 'd_vid', 't_vid', 'st_id', 'dataset', 'hash', 'process_records') return OrderedDict([(k, getattr(self, k)) for k in self.properties if k not in SKIP_KEYS])
dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n SKIP_KEYS = ('_source_table', '_dest_table', 'd_vid', 't_vid', 'st_id',\n 'dataset', 'hash', 'process_records')\n return OrderedDict([(k, getattr(self, k)) for k in self.properties if k\n not in SKIP_KEYS])\n", "role": "user" }, { "content": "dict", "role": "assistant" } ]
def x(path=None, root=None, db=None): return load(path, root=root, db=db)
get_runconfig
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path=None, root=None, db=None):\n return load(path, root=root, db=db)\n", "role": "user" }, { "content": "get_runconfig", "role": "assistant" } ]
def x(path=None, root=None, db=None, load_user=True): config = load_config(path, load_user=load_user) remotes = load_remotes(path, load_user=load_user) if remotes: if not 'remotes' in config: config.remotes = AttrDict() for k, v in remotes.remotes.items(): config.remotes[k] = v accounts = load_accounts(path, load_user=load_user) if accounts: if not 'accounts' in config: config.accounts = AttrDict() for k, v in accounts.accounts.items(): config.accounts[k] = v update_config(config) if root: config.library.filesystem_root = root if db: config.library.database = db return config
load
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path=None, root=None, db=None, load_user=True):\n config = load_config(path, load_user=load_user)\n remotes = load_remotes(path, load_user=load_user)\n if remotes:\n if not 'remotes' in config:\n config.remotes = AttrDict()\n for k, v in remotes.remotes.items():\n config.remotes[k] = v\n accounts = load_accounts(path, load_user=load_user)\n if accounts:\n if not 'accounts' in config:\n config.accounts = AttrDict()\n for k, v in accounts.accounts.items():\n config.accounts[k] = v\n update_config(config)\n if root:\n config.library.filesystem_root = root\n if db:\n config.library.database = db\n return config\n", "role": "user" }, { "content": "load", "role": "assistant" } ]
def x(file_name, extra_path=None, load_user=True): paths = [] if extra_path is not None: paths.append(extra_path) if os.getenv(ENVAR.CONFIG): paths.append(os.getenv(ENVAR.CONFIG)) if os.getenv(ENVAR.VIRT): paths.append(os.path.join(os.getenv(ENVAR.VIRT), USER_DIR)) if load_user: paths.append(os.path.expanduser('~/' + USER_DIR)) paths.append(ROOT_DIR) for path in paths: if os.path.isdir(path) and os.path.exists(os.path.join(path, file_name) ): f = os.path.join(path, file_name) return f raise ConfigurationError( "Failed to find configuration file '{}'. Looked for : {} ".format( file_name, paths))
find_config_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(file_name, extra_path=None, load_user=True):\n paths = []\n if extra_path is not None:\n paths.append(extra_path)\n if os.getenv(ENVAR.CONFIG):\n paths.append(os.getenv(ENVAR.CONFIG))\n if os.getenv(ENVAR.VIRT):\n paths.append(os.path.join(os.getenv(ENVAR.VIRT), USER_DIR))\n if load_user:\n paths.append(os.path.expanduser('~/' + USER_DIR))\n paths.append(ROOT_DIR)\n for path in paths:\n if os.path.isdir(path) and os.path.exists(os.path.join(path, file_name)\n ):\n f = os.path.join(path, file_name)\n return f\n raise ConfigurationError(\n \"Failed to find configuration file '{}'. Looked for : {} \".format(\n file_name, paths))\n", "role": "user" }, { "content": "find_config_file", "role": "assistant" } ]
def x(extra_path=None, load_user=True): from os.path import getmtime try: accts_file = find_config_file(ACCOUNTS_FILE, extra_path=extra_path, load_user=load_user) except ConfigurationError: accts_file = None if accts_file is not None and os.path.exists(accts_file): config = AttrDict() config.update_yaml(accts_file) if not 'accounts' in config: config.remotes = AttrDict() config.accounts.loaded = [accts_file, getmtime(accts_file)] return config else: return None
load_accounts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(extra_path=None, load_user=True):\n from os.path import getmtime\n try:\n accts_file = find_config_file(ACCOUNTS_FILE, extra_path=extra_path,\n load_user=load_user)\n except ConfigurationError:\n accts_file = None\n if accts_file is not None and os.path.exists(accts_file):\n config = AttrDict()\n config.update_yaml(accts_file)\n if not 'accounts' in config:\n config.remotes = AttrDict()\n config.accounts.loaded = [accts_file, getmtime(accts_file)]\n return config\n else:\n return None\n", "role": "user" }, { "content": "load_accounts", "role": "assistant" } ]
def x(extra_path=None, load_user=True): from os.path import getmtime try: remotes_file = find_config_file(REMOTES_FILE, extra_path=extra_path, load_user=load_user) except ConfigurationError: remotes_file = None if remotes_file is not None and os.path.exists(remotes_file): config = AttrDict() config.update_yaml(remotes_file) if not 'remotes' in config: config.remotes = AttrDict() config.remotes.loaded = [remotes_file, getmtime(remotes_file)] return config else: return None
load_remotes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(extra_path=None, load_user=True):\n from os.path import getmtime\n try:\n remotes_file = find_config_file(REMOTES_FILE, extra_path=extra_path,\n load_user=load_user)\n except ConfigurationError:\n remotes_file = None\n if remotes_file is not None and os.path.exists(remotes_file):\n config = AttrDict()\n config.update_yaml(remotes_file)\n if not 'remotes' in config:\n config.remotes = AttrDict()\n config.remotes.loaded = [remotes_file, getmtime(remotes_file)]\n return config\n else:\n return None\n", "role": "user" }, { "content": "load_remotes", "role": "assistant" } ]
def x(path=None, load_user=True): from os.path import getmtime config = AttrDict() if not path: path = ROOT_DIR config_file = find_config_file(CONFIG_FILE, extra_path=path, load_user= load_user) if os.path.exists(config_file): config.update_yaml(config_file) config.loaded = [config_file, getmtime(config_file)] else: config = AttrDict() config.loaded = [None, 0] return config
load_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path=None, load_user=True):\n from os.path import getmtime\n config = AttrDict()\n if not path:\n path = ROOT_DIR\n config_file = find_config_file(CONFIG_FILE, extra_path=path, load_user=\n load_user)\n if os.path.exists(config_file):\n config.update_yaml(config_file)\n config.loaded = [config_file, getmtime(config_file)]\n else:\n config = AttrDict()\n config.loaded = [None, 0]\n return config\n", "role": "user" }, { "content": "load_config", "role": "assistant" } ]
def x(config, use_environ=True): from ambry.util import select_from_url try: _ = config.library except KeyError: config.library = AttrDict() try: _ = config.filesystem except KeyError: config.filesystem = AttrDict() try: _ = config.accounts except KeyError: config.accounts = AttrDict() if not config.accounts.get('loaded'): config.accounts.loaded = [None, 0] try: _ = config.accounts.password except KeyError: config.accounts.password = None try: _ = config.remotes except KeyError: config.remotes = AttrDict() if not config.remotes.get('loaded'): config.remotes.loaded = [None, 0] if use_environ: if os.getenv(ENVAR.DB): config.library.database = os.getenv(ENVAR.DB) if os.getenv(ENVAR.ROOT): config.library.filesystem_root = os.getenv(ENVAR.ROOT) if os.getenv(ENVAR.PASSWORD): config.accounts.password = os.getenv(ENVAR.PASSWORD) try: for k, v in config.library.remotes.items(): config.remotes[k] = {'url': v} del config.library['remotes'] except KeyError as e: pass try: for k, v in config.remotes.items(): if 'url' in v: host = select_from_url(v['url'], 'netloc') if host in config.accounts: config.remotes[k].update(config.accounts[host]) del config.accounts[host] except KeyError: pass try: _ = config.library.database except KeyError: config.library.database = 'sqlite:///{root}/library.db' checks = ['config.library.filesystem_root'] for check in checks: try: _ = eval(check) except KeyError: raise ConfigurationError( "Configuration is missing '{}'; loaded from {} ".format( check, config.loaded[0])) _, config.library.database = normalize_dsn_or_dict(config.library.database) for k, v in filesystem_defaults.items(): if k not in config.filesystem: config.filesystem[k] = v config.modtime = max(config.loaded[1], config.remotes.loaded[1], config .accounts.loaded[1])
update_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(config, use_environ=True):\n from ambry.util import select_from_url\n try:\n _ = config.library\n except KeyError:\n config.library = AttrDict()\n try:\n _ = config.filesystem\n except KeyError:\n config.filesystem = AttrDict()\n try:\n _ = config.accounts\n except KeyError:\n config.accounts = AttrDict()\n if not config.accounts.get('loaded'):\n config.accounts.loaded = [None, 0]\n try:\n _ = config.accounts.password\n except KeyError:\n config.accounts.password = None\n try:\n _ = config.remotes\n except KeyError:\n config.remotes = AttrDict()\n if not config.remotes.get('loaded'):\n config.remotes.loaded = [None, 0]\n if use_environ:\n if os.getenv(ENVAR.DB):\n config.library.database = os.getenv(ENVAR.DB)\n if os.getenv(ENVAR.ROOT):\n config.library.filesystem_root = os.getenv(ENVAR.ROOT)\n if os.getenv(ENVAR.PASSWORD):\n config.accounts.password = os.getenv(ENVAR.PASSWORD)\n try:\n for k, v in config.library.remotes.items():\n config.remotes[k] = {'url': v}\n del config.library['remotes']\n except KeyError as e:\n pass\n try:\n for k, v in config.remotes.items():\n if 'url' in v:\n host = select_from_url(v['url'], 'netloc')\n if host in config.accounts:\n config.remotes[k].update(config.accounts[host])\n del config.accounts[host]\n except KeyError:\n pass\n try:\n _ = config.library.database\n except KeyError:\n config.library.database = 'sqlite:///{root}/library.db'\n checks = ['config.library.filesystem_root']\n for check in checks:\n try:\n _ = eval(check)\n except KeyError:\n raise ConfigurationError(\n \"Configuration is missing '{}'; loaded from {} \".format(\n check, config.loaded[0]))\n _, config.library.database = normalize_dsn_or_dict(config.library.database)\n for k, v in filesystem_defaults.items():\n if k not in config.filesystem:\n config.filesystem[k] = v\n config.modtime = max(config.loaded[1], config.remotes.loaded[1], config\n .accounts.loaded[1])\n", "role": "user" }, { "content": "update_config", "role": "assistant" } ]
def x(d): if isinstance(d, dict): try: d = d.to_dict() except AttributeError: pass config = d dsn = None elif isinstance(d, string_types): config = None dsn = d else: raise ConfigurationError( "Can't deal with database config '{}' type '{}' ".format(d, type(d))) if dsn: if dsn.startswith('sqlite') or dsn.startswith('spatialite'): driver, path = dsn.split(':', 1) slashes, path = path[:2], path[2:] if slashes != '//': raise ConfigurationError( 'Sqlite DSNs must start with at least 2 slashes') if len(path) == 1 and path[0] == '/': raise ConfigurationError( "Sqlite DSNs can't have only 3 slashes in path") if len(path) > 1 and path[0] != '/': raise ConfigurationError( 'Sqlite DSNs with a path must have 3 or 4 slashes.') path = path[1:] config = dict(server=None, username=None, password=None, driver =driver, dbname=path) else: d = parse_url_to_dict(dsn) config = dict(server=d['hostname'], dbname=d['path'].strip('/'), driver=d['scheme'], password=d.get('password', None), username=d.get('username', None)) else: up = d.get('username', '') or '' if d.get('password'): up += ':' + d.get('password', '') if up: up += '@' if up and not d.get('server'): raise ConfigurationError( "Can't construct a DSN with a username or password without a hostname" ) host_part = up + d.get('server', '') if d.get('server') else '' if d.get('dbname', False): path_part = '/' + d.get('dbname') else: path_part = '' dsn = '{}://{}{}'.format(d['driver'], host_part, path_part) return config, dsn
normalize_dsn_or_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(d):\n if isinstance(d, dict):\n try:\n d = d.to_dict()\n except AttributeError:\n pass\n config = d\n dsn = None\n elif isinstance(d, string_types):\n config = None\n dsn = d\n else:\n raise ConfigurationError(\n \"Can't deal with database config '{}' type '{}' \".format(d,\n type(d)))\n if dsn:\n if dsn.startswith('sqlite') or dsn.startswith('spatialite'):\n driver, path = dsn.split(':', 1)\n slashes, path = path[:2], path[2:]\n if slashes != '//':\n raise ConfigurationError(\n 'Sqlite DSNs must start with at least 2 slashes')\n if len(path) == 1 and path[0] == '/':\n raise ConfigurationError(\n \"Sqlite DSNs can't have only 3 slashes in path\")\n if len(path) > 1 and path[0] != '/':\n raise ConfigurationError(\n 'Sqlite DSNs with a path must have 3 or 4 slashes.')\n path = path[1:]\n config = dict(server=None, username=None, password=None, driver\n =driver, dbname=path)\n else:\n d = parse_url_to_dict(dsn)\n config = dict(server=d['hostname'], dbname=d['path'].strip('/'),\n driver=d['scheme'], password=d.get('password', None),\n username=d.get('username', None))\n else:\n up = d.get('username', '') or ''\n if d.get('password'):\n up += ':' + d.get('password', '')\n if up:\n up += '@'\n if up and not d.get('server'):\n raise ConfigurationError(\n \"Can't construct a DSN with a username or password without a hostname\"\n )\n host_part = up + d.get('server', '') if d.get('server') else ''\n if d.get('dbname', False):\n path_part = '/' + d.get('dbname')\n else:\n path_part = ''\n dsn = '{}://{}{}'.format(d['driver'], host_part, path_part)\n return config, dsn\n", "role": "user" }, { "content": "normalize_dsn_or_dict", "role": "assistant" } ]
def x(self, **kwargs): kwargs['_return_http_data_only'] = True if kwargs.get('callback'): return self.create_new_product_version_with_http_info(**kwargs) else: data = self.create_new_product_version_with_http_info(**kwargs) return data
create_new_product_version
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.create_new_product_version_with_http_info(**kwargs)\n else:\n data = self.create_new_product_version_with_http_info(**kwargs)\n return data\n", "role": "user" }, { "content": "create_new_product_version", "role": "assistant" } ]
def x(self, id, **kwargs): kwargs['_return_http_data_only'] = True if kwargs.get('callback'): return self.get_build_configuration_sets_with_http_info(id, **kwargs) else: data = self.get_build_configuration_sets_with_http_info(id, **kwargs) return data
get_build_configuration_sets
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.get_build_configuration_sets_with_http_info(id, **kwargs)\n else:\n data = self.get_build_configuration_sets_with_http_info(id, **kwargs)\n return data\n", "role": "user" }, { "content": "get_build_configuration_sets", "role": "assistant" } ]
def x(self, id, **kwargs): kwargs['_return_http_data_only'] = True if kwargs.get('callback'): return self.update_build_configuration_sets_with_http_info(id, **kwargs ) else: data = self.update_build_configuration_sets_with_http_info(id, **kwargs ) return data
update_build_configuration_sets
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.update_build_configuration_sets_with_http_info(id, **kwargs\n )\n else:\n data = self.update_build_configuration_sets_with_http_info(id, **kwargs\n )\n return data\n", "role": "user" }, { "content": "update_build_configuration_sets", "role": "assistant" } ]
def x(cmd, execute, echo=True): if execute: if echo: print('Executing: ' + cmd) return os.system(cmd) else: print(cmd) return 0
execute_command
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cmd, execute, echo=True):\n if execute:\n if echo:\n print('Executing: ' + cmd)\n return os.system(cmd)\n else:\n print(cmd)\n return 0\n", "role": "user" }, { "content": "execute_command", "role": "assistant" } ]
def x(level): lLevel = level.lower() unrecognized = False if lLevel == 'debug-all': loglevel = logging.DEBUG elif lLevel == 'debug': loglevel = logging.DEBUG elif lLevel == 'info': loglevel = logging.INFO elif lLevel == 'warning': loglevel = logging.WARNING elif lLevel == 'error': loglevel = logging.ERROR elif lLevel == 'critical': loglevel = logging.CRITICAL else: loglevel = logging.DEBUG unrecognized = True formatter = logging.Formatter( '%(asctime)s %(levelname)s %(filename)s:%(lineno)d/%(funcName)s: %(message)s' ) console = logging.StreamHandler() console.setLevel(loglevel) console.setFormatter(formatter) logging.getLogger('').addHandler(console) logging.getLogger('').setLevel(loglevel) if lLevel != 'debug-all': packagesWarning = ['requests.packages.urllib3', 'urllib3', 'requests_kerberos', 'jenkinsapi'] for package in packagesWarning: logging.debug('Setting loglevel for %s to WARNING.', package) logger = logging.getLogger(package) logger.setLevel(logging.WARNING) if unrecognized: logging.warning('Unrecognized log level: %s Log level set to debug', level) fh = logging.FileHandler('builder.log') fh.setLevel(loglevel) fh.setFormatter(formatter) logging.getLogger('').addHandler(fh)
set_log_level
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(level):\n lLevel = level.lower()\n unrecognized = False\n if lLevel == 'debug-all':\n loglevel = logging.DEBUG\n elif lLevel == 'debug':\n loglevel = logging.DEBUG\n elif lLevel == 'info':\n loglevel = logging.INFO\n elif lLevel == 'warning':\n loglevel = logging.WARNING\n elif lLevel == 'error':\n loglevel = logging.ERROR\n elif lLevel == 'critical':\n loglevel = logging.CRITICAL\n else:\n loglevel = logging.DEBUG\n unrecognized = True\n formatter = logging.Formatter(\n '%(asctime)s %(levelname)s %(filename)s:%(lineno)d/%(funcName)s: %(message)s'\n )\n console = logging.StreamHandler()\n console.setLevel(loglevel)\n console.setFormatter(formatter)\n logging.getLogger('').addHandler(console)\n logging.getLogger('').setLevel(loglevel)\n if lLevel != 'debug-all':\n packagesWarning = ['requests.packages.urllib3', 'urllib3',\n 'requests_kerberos', 'jenkinsapi']\n for package in packagesWarning:\n logging.debug('Setting loglevel for %s to WARNING.', package)\n logger = logging.getLogger(package)\n logger.setLevel(logging.WARNING)\n if unrecognized:\n logging.warning('Unrecognized log level: %s Log level set to debug',\n level)\n fh = logging.FileHandler('builder.log')\n fh.setLevel(loglevel)\n fh.setFormatter(formatter)\n logging.getLogger('').addHandler(fh)\n", "role": "user" }, { "content": "set_log_level", "role": "assistant" } ]
def x(argv): args = {} for rarg in argv: if re.match('^--.*', rarg): arg = rarg.replace('--', '', 1) fsplit = arg.split('=', 1) if len(fsplit) != 2: raise Exception("Command option '%s' not recognized." % rarg) rkey, value = fsplit ssplit = rkey.split('.', 1) if len(ssplit) != 2 or not ssplit[1]: raise Exception("Command option '%s' not recognized." % rarg) section, option = ssplit args[section] = option, value else: raise Exception("Command option '%s' not recognized." % rarg) return args
parse_conf_args
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(argv):\n args = {}\n for rarg in argv:\n if re.match('^--.*', rarg):\n arg = rarg.replace('--', '', 1)\n fsplit = arg.split('=', 1)\n if len(fsplit) != 2:\n raise Exception(\"Command option '%s' not recognized.\" % rarg)\n rkey, value = fsplit\n ssplit = rkey.split('.', 1)\n if len(ssplit) != 2 or not ssplit[1]:\n raise Exception(\"Command option '%s' not recognized.\" % rarg)\n section, option = ssplit\n args[section] = option, value\n else:\n raise Exception(\"Command option '%s' not recognized.\" % rarg)\n return args\n", "role": "user" }, { "content": "parse_conf_args", "role": "assistant" } ]
def x(field): def x(f): def x(*args): result = f(*args) if result is None or result == '': raise Exception("Config option '%s' is required." % field) else: return result return wrappedf return wrap
required
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(field):\n\n def x(f):\n\n def x(*args):\n result = f(*args)\n if result is None or result == '':\n raise Exception(\"Config option '%s' is required.\" % field)\n else:\n return result\n return wrappedf\n return wrap\n", "role": "user" }, { "content": "required", "role": "assistant" } ]
def x(s, delim, escape='\\', unescape=True): ret = [] current = [] itr = iter(s) for ch in itr: if ch == escape: try: if not unescape: current.append(escape) current.append(next(itr)) except StopIteration: if unescape: current.append(escape) elif ch == delim: ret.append(''.join(current)) current = [] else: current.append(ch) ret.append(''.join(current)) return ret
split_unescape
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s, delim, escape='\\\\', unescape=True):\n ret = []\n current = []\n itr = iter(s)\n for ch in itr:\n if ch == escape:\n try:\n if not unescape:\n current.append(escape)\n current.append(next(itr))\n except StopIteration:\n if unescape:\n current.append(escape)\n elif ch == delim:\n ret.append(''.join(current))\n current = []\n else:\n current.append(ch)\n ret.append(''.join(current))\n return ret\n", "role": "user" }, { "content": "split_unescape", "role": "assistant" } ]