| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | """Classes for load jobs.""" |
| |
|
| | import typing |
| | from typing import FrozenSet, List, Iterable, Optional |
| |
|
| | from google.cloud.bigquery.encryption_configuration import EncryptionConfiguration |
| | from google.cloud.bigquery.external_config import HivePartitioningOptions |
| | from google.cloud.bigquery.format_options import ParquetOptions |
| | from google.cloud.bigquery import _helpers |
| | from google.cloud.bigquery.schema import SchemaField |
| | from google.cloud.bigquery.schema import _to_schema_fields |
| | from google.cloud.bigquery.table import RangePartitioning |
| | from google.cloud.bigquery.table import TableReference |
| | from google.cloud.bigquery.table import TimePartitioning |
| | from google.cloud.bigquery.job.base import _AsyncJob |
| | from google.cloud.bigquery.job.base import _JobConfig |
| | from google.cloud.bigquery.job.base import _JobReference |
| | from google.cloud.bigquery.query import ConnectionProperty |
| |
|
| |
|
| | class ColumnNameCharacterMap: |
| | """Indicates the character map used for column names. |
| | |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#columnnamecharactermap |
| | """ |
| |
|
| | COLUMN_NAME_CHARACTER_MAP_UNSPECIFIED = "COLUMN_NAME_CHARACTER_MAP_UNSPECIFIED" |
| | """Unspecified column name character map.""" |
| |
|
| | STRICT = "STRICT" |
| | """Support flexible column name and reject invalid column names.""" |
| |
|
| | V1 = "V1" |
| | """ Support alphanumeric + underscore characters and names must start with |
| | a letter or underscore. Invalid column names will be normalized.""" |
| |
|
| | V2 = "V2" |
| | """Support flexible column name. Invalid column names will be normalized.""" |
| |
|
| |
|
| | class LoadJobConfig(_JobConfig): |
| | """Configuration options for load jobs. |
| | |
| | Set properties on the constructed configuration by using the property name |
| | as the name of a keyword argument. Values which are unset or :data:`None` |
| | use the BigQuery REST API default values. See the `BigQuery REST API |
| | reference documentation |
| | <https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad>`_ |
| | for a list of default values. |
| | |
| | Required options differ based on the |
| | :attr:`~google.cloud.bigquery.job.LoadJobConfig.source_format` value. |
| | For example, the BigQuery API's default value for |
| | :attr:`~google.cloud.bigquery.job.LoadJobConfig.source_format` is ``"CSV"``. |
| | When loading a CSV file, either |
| | :attr:`~google.cloud.bigquery.job.LoadJobConfig.schema` must be set or |
| | :attr:`~google.cloud.bigquery.job.LoadJobConfig.autodetect` must be set to |
| | :data:`True`. |
| | """ |
| |
|
| | def __init__(self, **kwargs) -> None: |
| | super(LoadJobConfig, self).__init__("load", **kwargs) |
| |
|
| | @property |
| | def allow_jagged_rows(self): |
| | """Optional[bool]: Allow missing trailing optional columns (CSV only). |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.allow_jagged_rows |
| | """ |
| | return self._get_sub_prop("allowJaggedRows") |
| |
|
| | @allow_jagged_rows.setter |
| | def allow_jagged_rows(self, value): |
| | self._set_sub_prop("allowJaggedRows", value) |
| |
|
| | @property |
| | def allow_quoted_newlines(self): |
| | """Optional[bool]: Allow quoted data containing newline characters (CSV only). |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.allow_quoted_newlines |
| | """ |
| | return self._get_sub_prop("allowQuotedNewlines") |
| |
|
| | @allow_quoted_newlines.setter |
| | def allow_quoted_newlines(self, value): |
| | self._set_sub_prop("allowQuotedNewlines", value) |
| |
|
| | @property |
| | def autodetect(self): |
| | """Optional[bool]: Automatically infer the schema from a sample of the data. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.autodetect |
| | """ |
| | return self._get_sub_prop("autodetect") |
| |
|
| | @autodetect.setter |
| | def autodetect(self, value): |
| | self._set_sub_prop("autodetect", value) |
| |
|
| | @property |
| | def clustering_fields(self): |
| | """Optional[List[str]]: Fields defining clustering for the table |
| | |
| | (Defaults to :data:`None`). |
| | |
| | Clustering fields are immutable after table creation. |
| | |
| | .. note:: |
| | |
| | BigQuery supports clustering for both partitioned and |
| | non-partitioned tables. |
| | """ |
| | prop = self._get_sub_prop("clustering") |
| | if prop is not None: |
| | return list(prop.get("fields", ())) |
| |
|
| | @clustering_fields.setter |
| | def clustering_fields(self, value): |
| | """Optional[List[str]]: Fields defining clustering for the table |
| | |
| | (Defaults to :data:`None`). |
| | """ |
| | if value is not None: |
| | self._set_sub_prop("clustering", {"fields": value}) |
| | else: |
| | self._del_sub_prop("clustering") |
| |
|
| | @property |
| | def connection_properties(self) -> List[ConnectionProperty]: |
| | """Connection properties. |
| | |
| | See |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.connection_properties |
| | |
| | .. versionadded:: 3.7.0 |
| | """ |
| | resource = self._get_sub_prop("connectionProperties", []) |
| | return [ConnectionProperty.from_api_repr(prop) for prop in resource] |
| |
|
| | @connection_properties.setter |
| | def connection_properties(self, value: Iterable[ConnectionProperty]): |
| | self._set_sub_prop( |
| | "connectionProperties", |
| | [prop.to_api_repr() for prop in value], |
| | ) |
| |
|
| | @property |
| | def create_disposition(self): |
| | """Optional[google.cloud.bigquery.job.CreateDisposition]: Specifies behavior |
| | for creating tables. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.create_disposition |
| | """ |
| | return self._get_sub_prop("createDisposition") |
| |
|
| | @create_disposition.setter |
| | def create_disposition(self, value): |
| | self._set_sub_prop("createDisposition", value) |
| |
|
| | @property |
| | def create_session(self) -> Optional[bool]: |
| | """[Preview] If :data:`True`, creates a new session, where |
| | :attr:`~google.cloud.bigquery.job.LoadJob.session_info` will contain a |
| | random server generated session id. |
| | |
| | If :data:`False`, runs load job with an existing ``session_id`` passed in |
| | :attr:`~google.cloud.bigquery.job.LoadJobConfig.connection_properties`, |
| | otherwise runs load job in non-session mode. |
| | |
| | See |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.create_session |
| | |
| | .. versionadded:: 3.7.0 |
| | """ |
| | return self._get_sub_prop("createSession") |
| |
|
| | @create_session.setter |
| | def create_session(self, value: Optional[bool]): |
| | self._set_sub_prop("createSession", value) |
| |
|
| | @property |
| | def decimal_target_types(self) -> Optional[FrozenSet[str]]: |
| | """Possible SQL data types to which the source decimal values are converted. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.decimal_target_types |
| | |
| | .. versionadded:: 2.21.0 |
| | """ |
| | prop = self._get_sub_prop("decimalTargetTypes") |
| | if prop is not None: |
| | prop = frozenset(prop) |
| | return prop |
| |
|
| | @decimal_target_types.setter |
| | def decimal_target_types(self, value: Optional[Iterable[str]]): |
| | if value is not None: |
| | self._set_sub_prop("decimalTargetTypes", list(value)) |
| | else: |
| | self._del_sub_prop("decimalTargetTypes") |
| |
|
| | @property |
| | def destination_encryption_configuration(self): |
| | """Optional[google.cloud.bigquery.encryption_configuration.EncryptionConfiguration]: Custom |
| | encryption configuration for the destination table. |
| | |
| | Custom encryption configuration (e.g., Cloud KMS keys) or :data:`None` |
| | if using default encryption. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.destination_encryption_configuration |
| | """ |
| | prop = self._get_sub_prop("destinationEncryptionConfiguration") |
| | if prop is not None: |
| | prop = EncryptionConfiguration.from_api_repr(prop) |
| | return prop |
| |
|
| | @destination_encryption_configuration.setter |
| | def destination_encryption_configuration(self, value): |
| | api_repr = value |
| | if value is not None: |
| | api_repr = value.to_api_repr() |
| | self._set_sub_prop("destinationEncryptionConfiguration", api_repr) |
| | else: |
| | self._del_sub_prop("destinationEncryptionConfiguration") |
| |
|
| | @property |
| | def destination_table_description(self): |
| | """Optional[str]: Description of the destination table. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#DestinationTableProperties.FIELDS.description |
| | """ |
| | prop = self._get_sub_prop("destinationTableProperties") |
| | if prop is not None: |
| | return prop["description"] |
| |
|
| | @destination_table_description.setter |
| | def destination_table_description(self, value): |
| | keys = [self._job_type, "destinationTableProperties", "description"] |
| | if value is not None: |
| | _helpers._set_sub_prop(self._properties, keys, value) |
| | else: |
| | _helpers._del_sub_prop(self._properties, keys) |
| |
|
| | @property |
| | def destination_table_friendly_name(self): |
| | """Optional[str]: Name given to destination table. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#DestinationTableProperties.FIELDS.friendly_name |
| | """ |
| | prop = self._get_sub_prop("destinationTableProperties") |
| | if prop is not None: |
| | return prop["friendlyName"] |
| |
|
| | @destination_table_friendly_name.setter |
| | def destination_table_friendly_name(self, value): |
| | keys = [self._job_type, "destinationTableProperties", "friendlyName"] |
| | if value is not None: |
| | _helpers._set_sub_prop(self._properties, keys, value) |
| | else: |
| | _helpers._del_sub_prop(self._properties, keys) |
| |
|
| | @property |
| | def encoding(self): |
| | """Optional[google.cloud.bigquery.job.Encoding]: The character encoding of the |
| | data. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.encoding |
| | """ |
| | return self._get_sub_prop("encoding") |
| |
|
| | @encoding.setter |
| | def encoding(self, value): |
| | self._set_sub_prop("encoding", value) |
| |
|
| | @property |
| | def field_delimiter(self): |
| | """Optional[str]: The separator for fields in a CSV file. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.field_delimiter |
| | """ |
| | return self._get_sub_prop("fieldDelimiter") |
| |
|
| | @field_delimiter.setter |
| | def field_delimiter(self, value): |
| | self._set_sub_prop("fieldDelimiter", value) |
| |
|
| | @property |
| | def hive_partitioning(self): |
| | """Optional[:class:`~.external_config.HivePartitioningOptions`]: [Beta] When set, \ |
| | it configures hive partitioning support. |
| | |
| | .. note:: |
| | **Experimental**. This feature is experimental and might change or |
| | have limited support. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.hive_partitioning_options |
| | """ |
| | prop = self._get_sub_prop("hivePartitioningOptions") |
| | if prop is None: |
| | return None |
| | return HivePartitioningOptions.from_api_repr(prop) |
| |
|
| | @hive_partitioning.setter |
| | def hive_partitioning(self, value): |
| | if value is not None: |
| | if isinstance(value, HivePartitioningOptions): |
| | value = value.to_api_repr() |
| | else: |
| | raise TypeError("Expected a HivePartitioningOptions instance or None.") |
| |
|
| | self._set_sub_prop("hivePartitioningOptions", value) |
| |
|
| | @property |
| | def ignore_unknown_values(self): |
| | """Optional[bool]: Ignore extra values not represented in the table schema. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.ignore_unknown_values |
| | """ |
| | return self._get_sub_prop("ignoreUnknownValues") |
| |
|
| | @ignore_unknown_values.setter |
| | def ignore_unknown_values(self, value): |
| | self._set_sub_prop("ignoreUnknownValues", value) |
| |
|
| | @property |
| | def json_extension(self): |
| | """Optional[str]: The extension to use for writing JSON data to BigQuery. Only supports GeoJSON currently. |
| | |
| | See: https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.json_extension |
| | |
| | """ |
| | return self._get_sub_prop("jsonExtension") |
| |
|
| | @json_extension.setter |
| | def json_extension(self, value): |
| | self._set_sub_prop("jsonExtension", value) |
| |
|
| | @property |
| | def max_bad_records(self): |
| | """Optional[int]: Number of invalid rows to ignore. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.max_bad_records |
| | """ |
| | return _helpers._int_or_none(self._get_sub_prop("maxBadRecords")) |
| |
|
| | @max_bad_records.setter |
| | def max_bad_records(self, value): |
| | self._set_sub_prop("maxBadRecords", value) |
| |
|
| | @property |
| | def null_marker(self): |
| | """Optional[str]: Represents a null value (CSV only). |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.null_marker |
| | """ |
| | return self._get_sub_prop("nullMarker") |
| |
|
| | @null_marker.setter |
| | def null_marker(self, value): |
| | self._set_sub_prop("nullMarker", value) |
| |
|
| | @property |
| | def preserve_ascii_control_characters(self): |
| | """Optional[bool]: Preserves the embedded ASCII control characters when sourceFormat is set to CSV. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.preserve_ascii_control_characters |
| | """ |
| | return self._get_sub_prop("preserveAsciiControlCharacters") |
| |
|
| | @preserve_ascii_control_characters.setter |
| | def preserve_ascii_control_characters(self, value): |
| | self._set_sub_prop("preserveAsciiControlCharacters", bool(value)) |
| |
|
| | @property |
| | def projection_fields(self) -> Optional[List[str]]: |
| | """Optional[List[str]]: If |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.source_format` is set to |
| | "DATASTORE_BACKUP", indicates which entity properties to load into |
| | BigQuery from a Cloud Datastore backup. |
| | |
| | Property names are case sensitive and must be top-level properties. If |
| | no properties are specified, BigQuery loads all properties. If any |
| | named property isn't found in the Cloud Datastore backup, an invalid |
| | error is returned in the job result. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.projection_fields |
| | """ |
| | return self._get_sub_prop("projectionFields") |
| |
|
| | @projection_fields.setter |
| | def projection_fields(self, value: Optional[List[str]]): |
| | self._set_sub_prop("projectionFields", value) |
| |
|
| | @property |
| | def quote_character(self): |
| | """Optional[str]: Character used to quote data sections (CSV only). |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.quote |
| | """ |
| | return self._get_sub_prop("quote") |
| |
|
| | @quote_character.setter |
| | def quote_character(self, value): |
| | self._set_sub_prop("quote", value) |
| |
|
| | @property |
| | def range_partitioning(self): |
| | """Optional[google.cloud.bigquery.table.RangePartitioning]: |
| | Configures range-based partitioning for destination table. |
| | |
| | .. note:: |
| | **Beta**. The integer range partitioning feature is in a |
| | pre-release state and might change or have limited support. |
| | |
| | Only specify at most one of |
| | :attr:`~google.cloud.bigquery.job.LoadJobConfig.time_partitioning` or |
| | :attr:`~google.cloud.bigquery.job.LoadJobConfig.range_partitioning`. |
| | |
| | Raises: |
| | ValueError: |
| | If the value is not |
| | :class:`~google.cloud.bigquery.table.RangePartitioning` or |
| | :data:`None`. |
| | """ |
| | resource = self._get_sub_prop("rangePartitioning") |
| | if resource is not None: |
| | return RangePartitioning(_properties=resource) |
| |
|
| | @range_partitioning.setter |
| | def range_partitioning(self, value): |
| | resource = value |
| | if isinstance(value, RangePartitioning): |
| | resource = value._properties |
| | elif value is not None: |
| | raise ValueError( |
| | "Expected value to be RangePartitioning or None, got {}.".format(value) |
| | ) |
| | self._set_sub_prop("rangePartitioning", resource) |
| |
|
| | @property |
| | def reference_file_schema_uri(self): |
| | """Optional[str]: |
| | When creating an external table, the user can provide a reference file with the |
| | table schema. This is enabled for the following formats: |
| | |
| | AVRO, PARQUET, ORC |
| | """ |
| | return self._get_sub_prop("referenceFileSchemaUri") |
| |
|
| | @reference_file_schema_uri.setter |
| | def reference_file_schema_uri(self, value): |
| | return self._set_sub_prop("referenceFileSchemaUri", value) |
| |
|
| | @property |
| | def schema(self): |
| | """Optional[Sequence[Union[ \ |
| | :class:`~google.cloud.bigquery.schema.SchemaField`, \ |
| | Mapping[str, Any] \ |
| | ]]]: Schema of the destination table. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.schema |
| | """ |
| | schema = _helpers._get_sub_prop(self._properties, ["load", "schema", "fields"]) |
| | if schema is None: |
| | return |
| | return [SchemaField.from_api_repr(field) for field in schema] |
| |
|
| | @schema.setter |
| | def schema(self, value): |
| | if value is None: |
| | self._del_sub_prop("schema") |
| | return |
| |
|
| | value = _to_schema_fields(value) |
| |
|
| | _helpers._set_sub_prop( |
| | self._properties, |
| | ["load", "schema", "fields"], |
| | [field.to_api_repr() for field in value], |
| | ) |
| |
|
| | @property |
| | def schema_update_options(self): |
| | """Optional[List[google.cloud.bigquery.job.SchemaUpdateOption]]: Specifies |
| | updates to the destination table schema to allow as a side effect of |
| | the load job. |
| | """ |
| | return self._get_sub_prop("schemaUpdateOptions") |
| |
|
| | @schema_update_options.setter |
| | def schema_update_options(self, values): |
| | self._set_sub_prop("schemaUpdateOptions", values) |
| |
|
| | @property |
| | def skip_leading_rows(self): |
| | """Optional[int]: Number of rows to skip when reading data (CSV only). |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.skip_leading_rows |
| | """ |
| | return _helpers._int_or_none(self._get_sub_prop("skipLeadingRows")) |
| |
|
| | @skip_leading_rows.setter |
| | def skip_leading_rows(self, value): |
| | self._set_sub_prop("skipLeadingRows", str(value)) |
| |
|
| | @property |
| | def source_format(self): |
| | """Optional[google.cloud.bigquery.job.SourceFormat]: File format of the data. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.source_format |
| | """ |
| | return self._get_sub_prop("sourceFormat") |
| |
|
| | @source_format.setter |
| | def source_format(self, value): |
| | self._set_sub_prop("sourceFormat", value) |
| |
|
| | @property |
| | def time_partitioning(self): |
| | """Optional[google.cloud.bigquery.table.TimePartitioning]: Specifies time-based |
| | partitioning for the destination table. |
| | |
| | Only specify at most one of |
| | :attr:`~google.cloud.bigquery.job.LoadJobConfig.time_partitioning` or |
| | :attr:`~google.cloud.bigquery.job.LoadJobConfig.range_partitioning`. |
| | """ |
| | prop = self._get_sub_prop("timePartitioning") |
| | if prop is not None: |
| | prop = TimePartitioning.from_api_repr(prop) |
| | return prop |
| |
|
| | @time_partitioning.setter |
| | def time_partitioning(self, value): |
| | api_repr = value |
| | if value is not None: |
| | api_repr = value.to_api_repr() |
| | self._set_sub_prop("timePartitioning", api_repr) |
| | else: |
| | self._del_sub_prop("timePartitioning") |
| |
|
| | @property |
| | def use_avro_logical_types(self): |
| | """Optional[bool]: For loads of Avro data, governs whether Avro logical types are |
| | converted to their corresponding BigQuery types (e.g. TIMESTAMP) rather than |
| | raw types (e.g. INTEGER). |
| | """ |
| | return self._get_sub_prop("useAvroLogicalTypes") |
| |
|
| | @use_avro_logical_types.setter |
| | def use_avro_logical_types(self, value): |
| | self._set_sub_prop("useAvroLogicalTypes", bool(value)) |
| |
|
| | @property |
| | def write_disposition(self): |
| | """Optional[google.cloud.bigquery.job.WriteDisposition]: Action that occurs if |
| | the destination table already exists. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.write_disposition |
| | """ |
| | return self._get_sub_prop("writeDisposition") |
| |
|
| | @write_disposition.setter |
| | def write_disposition(self, value): |
| | self._set_sub_prop("writeDisposition", value) |
| |
|
| | @property |
| | def parquet_options(self): |
| | """Optional[google.cloud.bigquery.format_options.ParquetOptions]: Additional |
| | properties to set if ``sourceFormat`` is set to PARQUET. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.parquet_options |
| | """ |
| | prop = self._get_sub_prop("parquetOptions") |
| | if prop is not None: |
| | prop = ParquetOptions.from_api_repr(prop) |
| | return prop |
| |
|
| | @parquet_options.setter |
| | def parquet_options(self, value): |
| | if value is not None: |
| | self._set_sub_prop("parquetOptions", value.to_api_repr()) |
| | else: |
| | self._del_sub_prop("parquetOptions") |
| |
|
| | @property |
| | def column_name_character_map(self) -> str: |
| | """Optional[google.cloud.bigquery.job.ColumnNameCharacterMap]: |
| | Character map supported for column names in CSV/Parquet loads. Defaults |
| | to STRICT and can be overridden by Project Config Service. Using this |
| | option with unsupported load formats will result in an error. |
| | |
| | See |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.column_name_character_map |
| | """ |
| | return self._get_sub_prop( |
| | "columnNameCharacterMap", |
| | ColumnNameCharacterMap.COLUMN_NAME_CHARACTER_MAP_UNSPECIFIED, |
| | ) |
| |
|
| | @column_name_character_map.setter |
| | def column_name_character_map(self, value: Optional[str]): |
| | if value is None: |
| | value = ColumnNameCharacterMap.COLUMN_NAME_CHARACTER_MAP_UNSPECIFIED |
| | self._set_sub_prop("columnNameCharacterMap", value) |
| |
|
| |
|
| | class LoadJob(_AsyncJob): |
| | """Asynchronous job for loading data into a table. |
| | |
| | Can load from Google Cloud Storage URIs or from a file. |
| | |
| | Args: |
| | job_id (str): the job's ID |
| | |
| | source_uris (Optional[Sequence[str]]): |
| | URIs of one or more data files to be loaded. See |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.source_uris |
| | for supported URI formats. Pass None for jobs that load from a file. |
| | |
| | destination (google.cloud.bigquery.table.TableReference): reference to table into which data is to be loaded. |
| | |
| | client (google.cloud.bigquery.client.Client): |
| | A client which holds credentials and project configuration |
| | for the dataset (which requires a project). |
| | """ |
| |
|
| | _JOB_TYPE = "load" |
| | _CONFIG_CLASS = LoadJobConfig |
| |
|
| | def __init__(self, job_id, source_uris, destination, client, job_config=None): |
| | super(LoadJob, self).__init__(job_id, client) |
| |
|
| | if job_config is not None: |
| | self._properties["configuration"] = job_config._properties |
| |
|
| | if source_uris is not None: |
| | _helpers._set_sub_prop( |
| | self._properties, ["configuration", "load", "sourceUris"], source_uris |
| | ) |
| |
|
| | if destination is not None: |
| | _helpers._set_sub_prop( |
| | self._properties, |
| | ["configuration", "load", "destinationTable"], |
| | destination.to_api_repr(), |
| | ) |
| |
|
| | @property |
| | def configuration(self) -> LoadJobConfig: |
| | """The configuration for this load job.""" |
| | return typing.cast(LoadJobConfig, super().configuration) |
| |
|
| | @property |
| | def destination(self): |
| | """google.cloud.bigquery.table.TableReference: table where loaded rows are written |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.destination_table |
| | """ |
| | dest_config = _helpers._get_sub_prop( |
| | self._properties, ["configuration", "load", "destinationTable"] |
| | ) |
| | return TableReference.from_api_repr(dest_config) |
| |
|
| | @property |
| | def source_uris(self): |
| | """Optional[Sequence[str]]: URIs of data files to be loaded. See |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#JobConfigurationLoad.FIELDS.source_uris |
| | for supported URI formats. None for jobs that load from a file. |
| | """ |
| | return _helpers._get_sub_prop( |
| | self._properties, ["configuration", "load", "sourceUris"] |
| | ) |
| |
|
| | @property |
| | def allow_jagged_rows(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.allow_jagged_rows`. |
| | """ |
| | return self.configuration.allow_jagged_rows |
| |
|
| | @property |
| | def allow_quoted_newlines(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.allow_quoted_newlines`. |
| | """ |
| | return self.configuration.allow_quoted_newlines |
| |
|
| | @property |
| | def autodetect(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.autodetect`. |
| | """ |
| | return self.configuration.autodetect |
| |
|
| | @property |
| | def connection_properties(self) -> List[ConnectionProperty]: |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.connection_properties`. |
| | |
| | .. versionadded:: 3.7.0 |
| | """ |
| | return self.configuration.connection_properties |
| |
|
| | @property |
| | def create_disposition(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.create_disposition`. |
| | """ |
| | return self.configuration.create_disposition |
| |
|
| | @property |
| | def create_session(self) -> Optional[bool]: |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.create_session`. |
| | |
| | .. versionadded:: 3.7.0 |
| | """ |
| | return self.configuration.create_session |
| |
|
| | @property |
| | def encoding(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.encoding`. |
| | """ |
| | return self.configuration.encoding |
| |
|
| | @property |
| | def field_delimiter(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.field_delimiter`. |
| | """ |
| | return self.configuration.field_delimiter |
| |
|
| | @property |
| | def ignore_unknown_values(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.ignore_unknown_values`. |
| | """ |
| | return self.configuration.ignore_unknown_values |
| |
|
| | @property |
| | def max_bad_records(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.max_bad_records`. |
| | """ |
| | return self.configuration.max_bad_records |
| |
|
| | @property |
| | def null_marker(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.null_marker`. |
| | """ |
| | return self.configuration.null_marker |
| |
|
| | @property |
| | def quote_character(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.quote_character`. |
| | """ |
| | return self.configuration.quote_character |
| |
|
| | @property |
| | def reference_file_schema_uri(self): |
| | """See: |
| | attr:`google.cloud.bigquery.job.LoadJobConfig.reference_file_schema_uri`. |
| | """ |
| | return self.configuration.reference_file_schema_uri |
| |
|
| | @property |
| | def skip_leading_rows(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.skip_leading_rows`. |
| | """ |
| | return self.configuration.skip_leading_rows |
| |
|
| | @property |
| | def source_format(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.source_format`. |
| | """ |
| | return self.configuration.source_format |
| |
|
| | @property |
| | def write_disposition(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.write_disposition`. |
| | """ |
| | return self.configuration.write_disposition |
| |
|
| | @property |
| | def schema(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.schema`. |
| | """ |
| | return self.configuration.schema |
| |
|
| | @property |
| | def destination_encryption_configuration(self): |
| | """google.cloud.bigquery.encryption_configuration.EncryptionConfiguration: Custom |
| | encryption configuration for the destination table. |
| | |
| | Custom encryption configuration (e.g., Cloud KMS keys) |
| | or :data:`None` if using default encryption. |
| | |
| | See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.destination_encryption_configuration`. |
| | """ |
| | return self.configuration.destination_encryption_configuration |
| |
|
| | @property |
| | def destination_table_description(self): |
| | """Optional[str] name given to destination table. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#DestinationTableProperties.FIELDS.description |
| | """ |
| | return self.configuration.destination_table_description |
| |
|
| | @property |
| | def destination_table_friendly_name(self): |
| | """Optional[str] name given to destination table. |
| | |
| | See: |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#DestinationTableProperties.FIELDS.friendly_name |
| | """ |
| | return self.configuration.destination_table_friendly_name |
| |
|
| | @property |
| | def range_partitioning(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.range_partitioning`. |
| | """ |
| | return self.configuration.range_partitioning |
| |
|
| | @property |
| | def time_partitioning(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.time_partitioning`. |
| | """ |
| | return self.configuration.time_partitioning |
| |
|
| | @property |
| | def use_avro_logical_types(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.use_avro_logical_types`. |
| | """ |
| | return self.configuration.use_avro_logical_types |
| |
|
| | @property |
| | def clustering_fields(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.clustering_fields`. |
| | """ |
| | return self.configuration.clustering_fields |
| |
|
| | @property |
| | def schema_update_options(self): |
| | """See |
| | :attr:`google.cloud.bigquery.job.LoadJobConfig.schema_update_options`. |
| | """ |
| | return self.configuration.schema_update_options |
| |
|
| | @property |
| | def input_file_bytes(self): |
| | """Count of bytes loaded from source files. |
| | |
| | Returns: |
| | Optional[int]: the count (None until set from the server). |
| | |
| | Raises: |
| | ValueError: for invalid value types. |
| | """ |
| | return _helpers._int_or_none( |
| | _helpers._get_sub_prop( |
| | self._properties, ["statistics", "load", "inputFileBytes"] |
| | ) |
| | ) |
| |
|
| | @property |
| | def input_files(self): |
| | """Count of source files. |
| | |
| | Returns: |
| | Optional[int]: the count (None until set from the server). |
| | """ |
| | return _helpers._int_or_none( |
| | _helpers._get_sub_prop( |
| | self._properties, ["statistics", "load", "inputFiles"] |
| | ) |
| | ) |
| |
|
| | @property |
| | def output_bytes(self): |
| | """Count of bytes saved to destination table. |
| | |
| | Returns: |
| | Optional[int]: the count (None until set from the server). |
| | """ |
| | return _helpers._int_or_none( |
| | _helpers._get_sub_prop( |
| | self._properties, ["statistics", "load", "outputBytes"] |
| | ) |
| | ) |
| |
|
| | @property |
| | def output_rows(self): |
| | """Count of rows saved to destination table. |
| | |
| | Returns: |
| | Optional[int]: the count (None until set from the server). |
| | """ |
| | return _helpers._int_or_none( |
| | _helpers._get_sub_prop( |
| | self._properties, ["statistics", "load", "outputRows"] |
| | ) |
| | ) |
| |
|
| | def to_api_repr(self): |
| | """Generate a resource for :meth:`_begin`.""" |
| | |
| | return { |
| | "jobReference": self._properties["jobReference"], |
| | "configuration": self._properties["configuration"], |
| | } |
| |
|
| | @classmethod |
| | def from_api_repr(cls, resource: dict, client) -> "LoadJob": |
| | """Factory: construct a job given its API representation |
| | |
| | .. note:: |
| | |
| | This method assumes that the project found in the resource matches |
| | the client's project. |
| | |
| | Args: |
| | resource (Dict): dataset job representation returned from the API |
| | |
| | client (google.cloud.bigquery.client.Client): |
| | Client which holds credentials and project |
| | configuration for the dataset. |
| | |
| | Returns: |
| | google.cloud.bigquery.job.LoadJob: Job parsed from ``resource``. |
| | """ |
| | cls._check_resource_config(resource) |
| | job_ref = _JobReference._from_api_repr(resource["jobReference"]) |
| | job = cls(job_ref, None, None, client) |
| | job._set_properties(resource) |
| | return job |
| |
|