| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | """Define API Datasets.""" |
| |
|
| | from __future__ import absolute_import |
| |
|
| | import copy |
| |
|
| | import typing |
| |
|
| | import google.cloud._helpers |
| |
|
| | from google.cloud.bigquery import _helpers |
| | from google.cloud.bigquery.model import ModelReference |
| | from google.cloud.bigquery.routine import Routine, RoutineReference |
| | from google.cloud.bigquery.table import Table, TableReference |
| | from google.cloud.bigquery.encryption_configuration import EncryptionConfiguration |
| |
|
| | from typing import Optional, List, Dict, Any, Union |
| |
|
| |
|
| | def _get_table_reference(self, table_id: str) -> TableReference: |
| | """Constructs a TableReference. |
| | |
| | Args: |
| | table_id (str): The ID of the table. |
| | |
| | Returns: |
| | google.cloud.bigquery.table.TableReference: |
| | A table reference for a table in this dataset. |
| | """ |
| | return TableReference(self, table_id) |
| |
|
| |
|
| | def _get_model_reference(self, model_id): |
| | """Constructs a ModelReference. |
| | |
| | Args: |
| | model_id (str): the ID of the model. |
| | |
| | Returns: |
| | google.cloud.bigquery.model.ModelReference: |
| | A ModelReference for a model in this dataset. |
| | """ |
| | return ModelReference.from_api_repr( |
| | {"projectId": self.project, "datasetId": self.dataset_id, "modelId": model_id} |
| | ) |
| |
|
| |
|
| | def _get_routine_reference(self, routine_id): |
| | """Constructs a RoutineReference. |
| | |
| | Args: |
| | routine_id (str): the ID of the routine. |
| | |
| | Returns: |
| | google.cloud.bigquery.routine.RoutineReference: |
| | A RoutineReference for a routine in this dataset. |
| | """ |
| | return RoutineReference.from_api_repr( |
| | { |
| | "projectId": self.project, |
| | "datasetId": self.dataset_id, |
| | "routineId": routine_id, |
| | } |
| | ) |
| |
|
| |
|
| | class DatasetReference(object): |
| | """DatasetReferences are pointers to datasets. |
| | |
| | See |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/datasets#datasetreference |
| | |
| | Args: |
| | project (str): The ID of the project |
| | dataset_id (str): The ID of the dataset |
| | |
| | Raises: |
| | ValueError: If either argument is not of type ``str``. |
| | """ |
| |
|
| | def __init__(self, project: str, dataset_id: str): |
| | if not isinstance(project, str): |
| | raise ValueError("Pass a string for project") |
| | if not isinstance(dataset_id, str): |
| | raise ValueError("Pass a string for dataset_id") |
| | self._project = project |
| | self._dataset_id = dataset_id |
| |
|
| | @property |
| | def project(self): |
| | """str: Project ID of the dataset.""" |
| | return self._project |
| |
|
| | @property |
| | def dataset_id(self): |
| | """str: Dataset ID.""" |
| | return self._dataset_id |
| |
|
| | @property |
| | def path(self): |
| | """str: URL path for the dataset based on project and dataset ID.""" |
| | return "/projects/%s/datasets/%s" % (self.project, self.dataset_id) |
| |
|
| | table = _get_table_reference |
| |
|
| | model = _get_model_reference |
| |
|
| | routine = _get_routine_reference |
| |
|
| | @classmethod |
| | def from_api_repr(cls, resource: dict) -> "DatasetReference": |
| | """Factory: construct a dataset reference given its API representation |
| | |
| | Args: |
| | resource (Dict[str, str]): |
| | Dataset reference resource representation returned from the API |
| | |
| | Returns: |
| | google.cloud.bigquery.dataset.DatasetReference: |
| | Dataset reference parsed from ``resource``. |
| | """ |
| | project = resource["projectId"] |
| | dataset_id = resource["datasetId"] |
| | return cls(project, dataset_id) |
| |
|
| | @classmethod |
| | def from_string( |
| | cls, dataset_id: str, default_project: Optional[str] = None |
| | ) -> "DatasetReference": |
| | """Construct a dataset reference from dataset ID string. |
| | |
| | Args: |
| | dataset_id (str): |
| | A dataset ID in standard SQL format. If ``default_project`` |
| | is not specified, this must include both the project ID and |
| | the dataset ID, separated by ``.``. |
| | default_project (Optional[str]): |
| | The project ID to use when ``dataset_id`` does not include a |
| | project ID. |
| | |
| | Returns: |
| | DatasetReference: |
| | Dataset reference parsed from ``dataset_id``. |
| | |
| | Examples: |
| | >>> DatasetReference.from_string('my-project-id.some_dataset') |
| | DatasetReference('my-project-id', 'some_dataset') |
| | |
| | Raises: |
| | ValueError: |
| | If ``dataset_id`` is not a fully-qualified dataset ID in |
| | standard SQL format. |
| | """ |
| | output_dataset_id = dataset_id |
| | parts = _helpers._split_id(dataset_id) |
| |
|
| | if len(parts) == 1: |
| | if default_project is not None: |
| | output_project_id = default_project |
| | else: |
| | raise ValueError( |
| | "When default_project is not set, dataset_id must be a " |
| | "fully-qualified dataset ID in standard SQL format, " |
| | 'e.g., "project.dataset_id" got {}'.format(dataset_id) |
| | ) |
| | elif len(parts) == 2: |
| | output_project_id, output_dataset_id = parts |
| | else: |
| | raise ValueError( |
| | "Too many parts in dataset_id. Expected a fully-qualified " |
| | "dataset ID in standard SQL format, " |
| | 'e.g. "project.dataset_id", got {}'.format(dataset_id) |
| | ) |
| |
|
| | return cls(output_project_id, output_dataset_id) |
| |
|
| | def to_api_repr(self) -> dict: |
| | """Construct the API resource representation of this dataset reference |
| | |
| | Returns: |
| | Dict[str, str]: dataset reference represented as an API resource |
| | """ |
| | return {"projectId": self._project, "datasetId": self._dataset_id} |
| |
|
| | def _key(self): |
| | """A tuple key that uniquely describes this field. |
| | |
| | Used to compute this instance's hashcode and evaluate equality. |
| | |
| | Returns: |
| | Tuple[str]: The contents of this :class:`.DatasetReference`. |
| | """ |
| | return (self._project, self._dataset_id) |
| |
|
| | def __eq__(self, other): |
| | if not isinstance(other, DatasetReference): |
| | return NotImplemented |
| | return self._key() == other._key() |
| |
|
| | def __ne__(self, other): |
| | return not self == other |
| |
|
| | def __hash__(self): |
| | return hash(self._key()) |
| |
|
| | def __str__(self): |
| | return f"{self.project}.{self._dataset_id}" |
| |
|
| | def __repr__(self): |
| | return "DatasetReference{}".format(self._key()) |
| |
|
| |
|
| | class AccessEntry(object): |
| | """Represents grant of an access role to an entity. |
| | |
| | An entry must have exactly one of the allowed |
| | :class:`google.cloud.bigquery.enums.EntityTypes`. If anything but ``view``, ``routine``, |
| | or ``dataset`` are set, a ``role`` is also required. ``role`` is omitted for ``view``, |
| | ``routine``, ``dataset``, because they are always read-only. |
| | |
| | See https://cloud.google.com/bigquery/docs/reference/rest/v2/datasets. |
| | |
| | Args: |
| | role: |
| | Role granted to the entity. The following string values are |
| | supported: `'READER'`, `'WRITER'`, `'OWNER'`. It may also be |
| | :data:`None` if the ``entity_type`` is ``view``, ``routine``, or ``dataset``. |
| | |
| | entity_type: |
| | Type of entity being granted the role. See |
| | :class:`google.cloud.bigquery.enums.EntityTypes` for supported types. |
| | |
| | entity_id: |
| | If the ``entity_type`` is not 'view', 'routine', or 'dataset', the |
| | ``entity_id`` is the ``str`` ID of the entity being granted the role. If |
| | the ``entity_type`` is 'view' or 'routine', the ``entity_id`` is a ``dict`` |
| | representing the view or routine from a different dataset to grant access |
| | to in the following format for views:: |
| | |
| | { |
| | 'projectId': string, |
| | 'datasetId': string, |
| | 'tableId': string |
| | } |
| | |
| | For routines:: |
| | |
| | { |
| | 'projectId': string, |
| | 'datasetId': string, |
| | 'routineId': string |
| | } |
| | |
| | If the ``entity_type`` is 'dataset', the ``entity_id`` is a ``dict`` that includes |
| | a 'dataset' field with a ``dict`` representing the dataset and a 'target_types' |
| | field with a ``str`` value of the dataset's resource type:: |
| | |
| | { |
| | 'dataset': { |
| | 'projectId': string, |
| | 'datasetId': string, |
| | }, |
| | 'target_types: 'VIEWS' |
| | } |
| | |
| | Raises: |
| | ValueError: |
| | If a ``view``, ``routine``, or ``dataset`` has ``role`` set, or a non ``view``, |
| | non ``routine``, and non ``dataset`` **does not** have a ``role`` set. |
| | |
| | Examples: |
| | >>> entry = AccessEntry('OWNER', 'userByEmail', 'user@example.com') |
| | |
| | >>> view = { |
| | ... 'projectId': 'my-project', |
| | ... 'datasetId': 'my_dataset', |
| | ... 'tableId': 'my_table' |
| | ... } |
| | >>> entry = AccessEntry(None, 'view', view) |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | role: Optional[str] = None, |
| | entity_type: Optional[str] = None, |
| | entity_id: Optional[Union[Dict[str, Any], str]] = None, |
| | ): |
| | self._properties = {} |
| | if entity_type is not None: |
| | self._properties[entity_type] = entity_id |
| | self._properties["role"] = role |
| | self._entity_type = entity_type |
| |
|
| | @property |
| | def role(self) -> Optional[str]: |
| | """The role of the entry.""" |
| | return typing.cast(Optional[str], self._properties.get("role")) |
| |
|
| | @role.setter |
| | def role(self, value): |
| | self._properties["role"] = value |
| |
|
| | @property |
| | def dataset(self) -> Optional[DatasetReference]: |
| | """API resource representation of a dataset reference.""" |
| | value = _helpers._get_sub_prop(self._properties, ["dataset", "dataset"]) |
| | return DatasetReference.from_api_repr(value) if value else None |
| |
|
| | @dataset.setter |
| | def dataset(self, value): |
| | if self.role is not None: |
| | raise ValueError( |
| | "Role must be None for a dataset. Current " "role: %r" % (self.role) |
| | ) |
| |
|
| | if isinstance(value, str): |
| | value = DatasetReference.from_string(value).to_api_repr() |
| |
|
| | if isinstance(value, (Dataset, DatasetListItem)): |
| | value = value.reference.to_api_repr() |
| |
|
| | _helpers._set_sub_prop(self._properties, ["dataset", "dataset"], value) |
| | _helpers._set_sub_prop( |
| | self._properties, |
| | ["dataset", "targetTypes"], |
| | self._properties.get("targetTypes"), |
| | ) |
| |
|
| | @property |
| | def dataset_target_types(self) -> Optional[List[str]]: |
| | """Which resources that the dataset in this entry applies to.""" |
| | return typing.cast( |
| | Optional[List[str]], |
| | _helpers._get_sub_prop(self._properties, ["dataset", "targetTypes"]), |
| | ) |
| |
|
| | @dataset_target_types.setter |
| | def dataset_target_types(self, value): |
| | self._properties.setdefault("dataset", {}) |
| | _helpers._set_sub_prop(self._properties, ["dataset", "targetTypes"], value) |
| |
|
| | @property |
| | def routine(self) -> Optional[RoutineReference]: |
| | """API resource representation of a routine reference.""" |
| | value = typing.cast(Optional[Dict], self._properties.get("routine")) |
| | return RoutineReference.from_api_repr(value) if value else None |
| |
|
| | @routine.setter |
| | def routine(self, value): |
| | if self.role is not None: |
| | raise ValueError( |
| | "Role must be None for a routine. Current " "role: %r" % (self.role) |
| | ) |
| |
|
| | if isinstance(value, str): |
| | value = RoutineReference.from_string(value).to_api_repr() |
| |
|
| | if isinstance(value, RoutineReference): |
| | value = value.to_api_repr() |
| |
|
| | if isinstance(value, Routine): |
| | value = value.reference.to_api_repr() |
| |
|
| | self._properties["routine"] = value |
| |
|
| | @property |
| | def view(self) -> Optional[TableReference]: |
| | """API resource representation of a view reference.""" |
| | value = typing.cast(Optional[Dict], self._properties.get("view")) |
| | return TableReference.from_api_repr(value) if value else None |
| |
|
| | @view.setter |
| | def view(self, value): |
| | if self.role is not None: |
| | raise ValueError( |
| | "Role must be None for a view. Current " "role: %r" % (self.role) |
| | ) |
| |
|
| | if isinstance(value, str): |
| | value = TableReference.from_string(value).to_api_repr() |
| |
|
| | if isinstance(value, TableReference): |
| | value = value.to_api_repr() |
| |
|
| | if isinstance(value, Table): |
| | value = value.reference.to_api_repr() |
| |
|
| | self._properties["view"] = value |
| |
|
| | @property |
| | def group_by_email(self) -> Optional[str]: |
| | """An email address of a Google Group to grant access to.""" |
| | return typing.cast(Optional[str], self._properties.get("groupByEmail")) |
| |
|
| | @group_by_email.setter |
| | def group_by_email(self, value): |
| | self._properties["groupByEmail"] = value |
| |
|
| | @property |
| | def user_by_email(self) -> Optional[str]: |
| | """An email address of a user to grant access to.""" |
| | return typing.cast(Optional[str], self._properties.get("userByEmail")) |
| |
|
| | @user_by_email.setter |
| | def user_by_email(self, value): |
| | self._properties["userByEmail"] = value |
| |
|
| | @property |
| | def domain(self) -> Optional[str]: |
| | """A domain to grant access to.""" |
| | return typing.cast(Optional[str], self._properties.get("domain")) |
| |
|
| | @domain.setter |
| | def domain(self, value): |
| | self._properties["domain"] = value |
| |
|
| | @property |
| | def special_group(self) -> Optional[str]: |
| | """A special group to grant access to.""" |
| | return typing.cast(Optional[str], self._properties.get("specialGroup")) |
| |
|
| | @special_group.setter |
| | def special_group(self, value): |
| | self._properties["specialGroup"] = value |
| |
|
| | @property |
| | def entity_type(self) -> Optional[str]: |
| | """The entity_type of the entry.""" |
| | return self._entity_type |
| |
|
| | @property |
| | def entity_id(self) -> Optional[Union[Dict[str, Any], str]]: |
| | """The entity_id of the entry.""" |
| | return self._properties.get(self._entity_type) if self._entity_type else None |
| |
|
| | def __eq__(self, other): |
| | if not isinstance(other, AccessEntry): |
| | return NotImplemented |
| | return self._key() == other._key() |
| |
|
| | def __ne__(self, other): |
| | return not self == other |
| |
|
| | def __repr__(self): |
| | return f"<AccessEntry: role={self.role}, {self._entity_type}={self.entity_id}>" |
| |
|
| | def _key(self): |
| | """A tuple key that uniquely describes this field. |
| | Used to compute this instance's hashcode and evaluate equality. |
| | Returns: |
| | Tuple: The contents of this :class:`~google.cloud.bigquery.dataset.AccessEntry`. |
| | """ |
| | properties = self._properties.copy() |
| | prop_tup = tuple(sorted(properties.items())) |
| | return (self.role, self._entity_type, self.entity_id, prop_tup) |
| |
|
| | def __hash__(self): |
| | return hash(self._key()) |
| |
|
| | def to_api_repr(self): |
| | """Construct the API resource representation of this access entry |
| | |
| | Returns: |
| | Dict[str, object]: Access entry represented as an API resource |
| | """ |
| | resource = copy.deepcopy(self._properties) |
| | return resource |
| |
|
| | @classmethod |
| | def from_api_repr(cls, resource: dict) -> "AccessEntry": |
| | """Factory: construct an access entry given its API representation |
| | |
| | Args: |
| | resource (Dict[str, object]): |
| | Access entry resource representation returned from the API |
| | |
| | Returns: |
| | google.cloud.bigquery.dataset.AccessEntry: |
| | Access entry parsed from ``resource``. |
| | |
| | Raises: |
| | ValueError: |
| | If the resource has more keys than ``role`` and one additional |
| | key. |
| | """ |
| | entry = resource.copy() |
| | role = entry.pop("role", None) |
| | entity_type, entity_id = entry.popitem() |
| | if len(entry) != 0: |
| | raise ValueError("Entry has unexpected keys remaining.", entry) |
| |
|
| | return cls(role, entity_type, entity_id) |
| |
|
| |
|
| | class Dataset(object): |
| | """Datasets are containers for tables. |
| | |
| | See |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/datasets#resource-dataset |
| | |
| | Args: |
| | dataset_ref (Union[google.cloud.bigquery.dataset.DatasetReference, str]): |
| | A pointer to a dataset. If ``dataset_ref`` is a string, it must |
| | include both the project ID and the dataset ID, separated by |
| | ``.``. |
| | """ |
| |
|
| | _PROPERTY_TO_API_FIELD = { |
| | "access_entries": "access", |
| | "created": "creationTime", |
| | "default_partition_expiration_ms": "defaultPartitionExpirationMs", |
| | "default_table_expiration_ms": "defaultTableExpirationMs", |
| | "friendly_name": "friendlyName", |
| | "default_encryption_configuration": "defaultEncryptionConfiguration", |
| | "is_case_insensitive": "isCaseInsensitive", |
| | "storage_billing_model": "storageBillingModel", |
| | "max_time_travel_hours": "maxTimeTravelHours", |
| | "default_rounding_mode": "defaultRoundingMode", |
| | } |
| |
|
| | def __init__(self, dataset_ref) -> None: |
| | if isinstance(dataset_ref, str): |
| | dataset_ref = DatasetReference.from_string(dataset_ref) |
| | self._properties = {"datasetReference": dataset_ref.to_api_repr(), "labels": {}} |
| |
|
| | @property |
| | def max_time_travel_hours(self): |
| | """ |
| | Optional[int]: Defines the time travel window in hours. The value can |
| | be from 48 to 168 hours (2 to 7 days), and in multiple of 24 hours |
| | (48, 72, 96, 120, 144, 168). |
| | The default value is 168 hours if this is not set. |
| | """ |
| | return self._properties.get("maxTimeTravelHours") |
| |
|
| | @max_time_travel_hours.setter |
| | def max_time_travel_hours(self, hours): |
| | if not isinstance(hours, int): |
| | raise ValueError(f"max_time_travel_hours must be an integer. Got {hours}") |
| | if hours < 2 * 24 or hours > 7 * 24: |
| | raise ValueError( |
| | "Time Travel Window should be from 48 to 168 hours (2 to 7 days)" |
| | ) |
| | if hours % 24 != 0: |
| | raise ValueError("Time Travel Window should be multiple of 24") |
| | self._properties["maxTimeTravelHours"] = hours |
| |
|
| | @property |
| | def default_rounding_mode(self): |
| | """Union[str, None]: defaultRoundingMode of the dataset as set by the user |
| | (defaults to :data:`None`). |
| | |
| | Set the value to one of ``'ROUND_HALF_AWAY_FROM_ZERO'``, ``'ROUND_HALF_EVEN'``, or |
| | ``'ROUNDING_MODE_UNSPECIFIED'``. |
| | |
| | See `default rounding mode |
| | <https://cloud.google.com/bigquery/docs/reference/rest/v2/datasets#Dataset.FIELDS.default_rounding_mode>`_ |
| | in REST API docs and `updating the default rounding model |
| | <https://cloud.google.com/bigquery/docs/updating-datasets#update_rounding_mode>`_ |
| | guide. |
| | |
| | Raises: |
| | ValueError: for invalid value types. |
| | """ |
| | return self._properties.get("defaultRoundingMode") |
| |
|
| | @default_rounding_mode.setter |
| | def default_rounding_mode(self, value): |
| | possible_values = [ |
| | "ROUNDING_MODE_UNSPECIFIED", |
| | "ROUND_HALF_AWAY_FROM_ZERO", |
| | "ROUND_HALF_EVEN", |
| | ] |
| | if not isinstance(value, str) and value is not None: |
| | raise ValueError("Pass a string, or None") |
| | if value is None: |
| | self._properties["defaultRoundingMode"] = "ROUNDING_MODE_UNSPECIFIED" |
| | if value not in possible_values and value is not None: |
| | raise ValueError( |
| | f'rounding mode needs to be one of {",".join(possible_values)}' |
| | ) |
| | if value: |
| | self._properties["defaultRoundingMode"] = value |
| |
|
| | @property |
| | def project(self): |
| | """str: Project ID of the project bound to the dataset.""" |
| | return self._properties["datasetReference"]["projectId"] |
| |
|
| | @property |
| | def path(self): |
| | """str: URL path for the dataset based on project and dataset ID.""" |
| | return "/projects/%s/datasets/%s" % (self.project, self.dataset_id) |
| |
|
| | @property |
| | def access_entries(self): |
| | """List[google.cloud.bigquery.dataset.AccessEntry]: Dataset's access |
| | entries. |
| | |
| | ``role`` augments the entity type and must be present **unless** the |
| | entity type is ``view`` or ``routine``. |
| | |
| | Raises: |
| | TypeError: If 'value' is not a sequence |
| | ValueError: |
| | If any item in the sequence is not an |
| | :class:`~google.cloud.bigquery.dataset.AccessEntry`. |
| | """ |
| | entries = self._properties.get("access", []) |
| | return [AccessEntry.from_api_repr(entry) for entry in entries] |
| |
|
| | @access_entries.setter |
| | def access_entries(self, value): |
| | if not all(isinstance(field, AccessEntry) for field in value): |
| | raise ValueError("Values must be AccessEntry instances") |
| | entries = [entry.to_api_repr() for entry in value] |
| | self._properties["access"] = entries |
| |
|
| | @property |
| | def created(self): |
| | """Union[datetime.datetime, None]: Datetime at which the dataset was |
| | created (:data:`None` until set from the server). |
| | """ |
| | creation_time = self._properties.get("creationTime") |
| | if creation_time is not None: |
| | |
| | return google.cloud._helpers._datetime_from_microseconds( |
| | 1000.0 * float(creation_time) |
| | ) |
| |
|
| | @property |
| | def dataset_id(self): |
| | """str: Dataset ID.""" |
| | return self._properties["datasetReference"]["datasetId"] |
| |
|
| | @property |
| | def full_dataset_id(self): |
| | """Union[str, None]: ID for the dataset resource (:data:`None` until |
| | set from the server) |
| | |
| | In the format ``project_id:dataset_id``. |
| | """ |
| | return self._properties.get("id") |
| |
|
| | @property |
| | def reference(self): |
| | """google.cloud.bigquery.dataset.DatasetReference: A reference to this |
| | dataset. |
| | """ |
| | return DatasetReference(self.project, self.dataset_id) |
| |
|
| | @property |
| | def etag(self): |
| | """Union[str, None]: ETag for the dataset resource (:data:`None` until |
| | set from the server). |
| | """ |
| | return self._properties.get("etag") |
| |
|
| | @property |
| | def modified(self): |
| | """Union[datetime.datetime, None]: Datetime at which the dataset was |
| | last modified (:data:`None` until set from the server). |
| | """ |
| | modified_time = self._properties.get("lastModifiedTime") |
| | if modified_time is not None: |
| | |
| | return google.cloud._helpers._datetime_from_microseconds( |
| | 1000.0 * float(modified_time) |
| | ) |
| |
|
| | @property |
| | def self_link(self): |
| | """Union[str, None]: URL for the dataset resource (:data:`None` until |
| | set from the server). |
| | """ |
| | return self._properties.get("selfLink") |
| |
|
| | @property |
| | def default_partition_expiration_ms(self): |
| | """Optional[int]: The default partition expiration for all |
| | partitioned tables in the dataset, in milliseconds. |
| | |
| | Once this property is set, all newly-created partitioned tables in |
| | the dataset will have an ``time_paritioning.expiration_ms`` property |
| | set to this value, and changing the value will only affect new |
| | tables, not existing ones. The storage in a partition will have an |
| | expiration time of its partition time plus this value. |
| | |
| | Setting this property overrides the use of |
| | ``default_table_expiration_ms`` for partitioned tables: only one of |
| | ``default_table_expiration_ms`` and |
| | ``default_partition_expiration_ms`` will be used for any new |
| | partitioned table. If you provide an explicit |
| | ``time_partitioning.expiration_ms`` when creating or updating a |
| | partitioned table, that value takes precedence over the default |
| | partition expiration time indicated by this property. |
| | """ |
| | return _helpers._int_or_none( |
| | self._properties.get("defaultPartitionExpirationMs") |
| | ) |
| |
|
| | @default_partition_expiration_ms.setter |
| | def default_partition_expiration_ms(self, value): |
| | self._properties["defaultPartitionExpirationMs"] = _helpers._str_or_none(value) |
| |
|
| | @property |
| | def default_table_expiration_ms(self): |
| | """Union[int, None]: Default expiration time for tables in the dataset |
| | (defaults to :data:`None`). |
| | |
| | Raises: |
| | ValueError: For invalid value types. |
| | """ |
| | return _helpers._int_or_none(self._properties.get("defaultTableExpirationMs")) |
| |
|
| | @default_table_expiration_ms.setter |
| | def default_table_expiration_ms(self, value): |
| | if not isinstance(value, int) and value is not None: |
| | raise ValueError("Pass an integer, or None") |
| | self._properties["defaultTableExpirationMs"] = _helpers._str_or_none(value) |
| |
|
| | @property |
| | def description(self): |
| | """Optional[str]: Description of the dataset as set by the user |
| | (defaults to :data:`None`). |
| | |
| | Raises: |
| | ValueError: for invalid value types. |
| | """ |
| | return self._properties.get("description") |
| |
|
| | @description.setter |
| | def description(self, value): |
| | if not isinstance(value, str) and value is not None: |
| | raise ValueError("Pass a string, or None") |
| | self._properties["description"] = value |
| |
|
| | @property |
| | def friendly_name(self): |
| | """Union[str, None]: Title of the dataset as set by the user |
| | (defaults to :data:`None`). |
| | |
| | Raises: |
| | ValueError: for invalid value types. |
| | """ |
| | return self._properties.get("friendlyName") |
| |
|
| | @friendly_name.setter |
| | def friendly_name(self, value): |
| | if not isinstance(value, str) and value is not None: |
| | raise ValueError("Pass a string, or None") |
| | self._properties["friendlyName"] = value |
| |
|
| | @property |
| | def location(self): |
| | """Union[str, None]: Location in which the dataset is hosted as set by |
| | the user (defaults to :data:`None`). |
| | |
| | Raises: |
| | ValueError: for invalid value types. |
| | """ |
| | return self._properties.get("location") |
| |
|
| | @location.setter |
| | def location(self, value): |
| | if not isinstance(value, str) and value is not None: |
| | raise ValueError("Pass a string, or None") |
| | self._properties["location"] = value |
| |
|
| | @property |
| | def labels(self): |
| | """Dict[str, str]: Labels for the dataset. |
| | |
| | This method always returns a dict. To change a dataset's labels, |
| | modify the dict, then call |
| | :meth:`google.cloud.bigquery.client.Client.update_dataset`. To delete |
| | a label, set its value to :data:`None` before updating. |
| | |
| | Raises: |
| | ValueError: for invalid value types. |
| | """ |
| | return self._properties.setdefault("labels", {}) |
| |
|
| | @labels.setter |
| | def labels(self, value): |
| | if not isinstance(value, dict): |
| | raise ValueError("Pass a dict") |
| | self._properties["labels"] = value |
| |
|
| | @property |
| | def default_encryption_configuration(self): |
| | """google.cloud.bigquery.encryption_configuration.EncryptionConfiguration: Custom |
| | encryption configuration for all tables in the dataset. |
| | |
| | Custom encryption configuration (e.g., Cloud KMS keys) or :data:`None` |
| | if using default encryption. |
| | |
| | See `protecting data with Cloud KMS keys |
| | <https://cloud.google.com/bigquery/docs/customer-managed-encryption>`_ |
| | in the BigQuery documentation. |
| | """ |
| | prop = self._properties.get("defaultEncryptionConfiguration") |
| | if prop: |
| | prop = EncryptionConfiguration.from_api_repr(prop) |
| | return prop |
| |
|
| | @default_encryption_configuration.setter |
| | def default_encryption_configuration(self, value): |
| | api_repr = value |
| | if value: |
| | api_repr = value.to_api_repr() |
| | self._properties["defaultEncryptionConfiguration"] = api_repr |
| |
|
| | @property |
| | def is_case_insensitive(self): |
| | """Optional[bool]: True if the dataset and its table names are case-insensitive, otherwise False. |
| | By default, this is False, which means the dataset and its table names are case-sensitive. |
| | This field does not affect routine references. |
| | |
| | Raises: |
| | ValueError: for invalid value types. |
| | """ |
| | return self._properties.get("isCaseInsensitive") or False |
| |
|
| | @is_case_insensitive.setter |
| | def is_case_insensitive(self, value): |
| | if not isinstance(value, bool) and value is not None: |
| | raise ValueError("Pass a boolean value, or None") |
| | if value is None: |
| | value = False |
| | self._properties["isCaseInsensitive"] = value |
| |
|
| | @property |
| | def storage_billing_model(self): |
| | """Union[str, None]: StorageBillingModel of the dataset as set by the user |
| | (defaults to :data:`None`). |
| | |
| | Set the value to one of ``'LOGICAL'``, ``'PHYSICAL'``, or |
| | ``'STORAGE_BILLING_MODEL_UNSPECIFIED'``. This change takes 24 hours to |
| | take effect and you must wait 14 days before you can change the storage |
| | billing model again. |
| | |
| | See `storage billing model |
| | <https://cloud.google.com/bigquery/docs/reference/rest/v2/datasets#Dataset.FIELDS.storage_billing_model>`_ |
| | in REST API docs and `updating the storage billing model |
| | <https://cloud.google.com/bigquery/docs/updating-datasets#update_storage_billing_models>`_ |
| | guide. |
| | |
| | Raises: |
| | ValueError: for invalid value types. |
| | """ |
| | return self._properties.get("storageBillingModel") |
| |
|
| | @storage_billing_model.setter |
| | def storage_billing_model(self, value): |
| | if not isinstance(value, str) and value is not None: |
| | raise ValueError( |
| | "storage_billing_model must be a string (e.g. 'LOGICAL'," |
| | " 'PHYSICAL', 'STORAGE_BILLING_MODEL_UNSPECIFIED'), or None." |
| | f" Got {repr(value)}." |
| | ) |
| | self._properties["storageBillingModel"] = value |
| |
|
| | @classmethod |
| | def from_string(cls, full_dataset_id: str) -> "Dataset": |
| | """Construct a dataset from fully-qualified dataset ID. |
| | |
| | Args: |
| | full_dataset_id (str): |
| | A fully-qualified dataset ID in standard SQL format. Must |
| | include both the project ID and the dataset ID, separated by |
| | ``.``. |
| | |
| | Returns: |
| | Dataset: Dataset parsed from ``full_dataset_id``. |
| | |
| | Examples: |
| | >>> Dataset.from_string('my-project-id.some_dataset') |
| | Dataset(DatasetReference('my-project-id', 'some_dataset')) |
| | |
| | Raises: |
| | ValueError: |
| | If ``full_dataset_id`` is not a fully-qualified dataset ID in |
| | standard SQL format. |
| | """ |
| | return cls(DatasetReference.from_string(full_dataset_id)) |
| |
|
| | @classmethod |
| | def from_api_repr(cls, resource: dict) -> "Dataset": |
| | """Factory: construct a dataset given its API representation |
| | |
| | Args: |
| | resource (Dict[str: object]): |
| | Dataset resource representation returned from the API |
| | |
| | Returns: |
| | google.cloud.bigquery.dataset.Dataset: |
| | Dataset parsed from ``resource``. |
| | """ |
| | if ( |
| | "datasetReference" not in resource |
| | or "datasetId" not in resource["datasetReference"] |
| | ): |
| | raise KeyError( |
| | "Resource lacks required identity information:" |
| | '["datasetReference"]["datasetId"]' |
| | ) |
| | project_id = resource["datasetReference"]["projectId"] |
| | dataset_id = resource["datasetReference"]["datasetId"] |
| | dataset = cls(DatasetReference(project_id, dataset_id)) |
| | dataset._properties = copy.deepcopy(resource) |
| | return dataset |
| |
|
| | def to_api_repr(self) -> dict: |
| | """Construct the API resource representation of this dataset |
| | |
| | Returns: |
| | Dict[str, object]: The dataset represented as an API resource |
| | """ |
| | return copy.deepcopy(self._properties) |
| |
|
| | def _build_resource(self, filter_fields): |
| | """Generate a resource for ``update``.""" |
| | return _helpers._build_resource_from_properties(self, filter_fields) |
| |
|
| | table = _get_table_reference |
| |
|
| | model = _get_model_reference |
| |
|
| | routine = _get_routine_reference |
| |
|
| | def __repr__(self): |
| | return "Dataset({})".format(repr(self.reference)) |
| |
|
| |
|
| | class DatasetListItem(object): |
| | """A read-only dataset resource from a list operation. |
| | |
| | For performance reasons, the BigQuery API only includes some of the |
| | dataset properties when listing datasets. Notably, |
| | :attr:`~google.cloud.bigquery.dataset.Dataset.access_entries` is missing. |
| | |
| | For a full list of the properties that the BigQuery API returns, see the |
| | `REST documentation for datasets.list |
| | <https://cloud.google.com/bigquery/docs/reference/rest/v2/datasets/list>`_. |
| | |
| | |
| | Args: |
| | resource (Dict[str, str]): |
| | A dataset-like resource object from a dataset list response. A |
| | ``datasetReference`` property is required. |
| | |
| | Raises: |
| | ValueError: |
| | If ``datasetReference`` or one of its required members is missing |
| | from ``resource``. |
| | """ |
| |
|
| | def __init__(self, resource): |
| | if "datasetReference" not in resource: |
| | raise ValueError("resource must contain a datasetReference value") |
| | if "projectId" not in resource["datasetReference"]: |
| | raise ValueError( |
| | "resource['datasetReference'] must contain a projectId value" |
| | ) |
| | if "datasetId" not in resource["datasetReference"]: |
| | raise ValueError( |
| | "resource['datasetReference'] must contain a datasetId value" |
| | ) |
| | self._properties = resource |
| |
|
| | @property |
| | def project(self): |
| | """str: Project bound to the dataset.""" |
| | return self._properties["datasetReference"]["projectId"] |
| |
|
| | @property |
| | def dataset_id(self): |
| | """str: Dataset ID.""" |
| | return self._properties["datasetReference"]["datasetId"] |
| |
|
| | @property |
| | def full_dataset_id(self): |
| | """Union[str, None]: ID for the dataset resource (:data:`None` until |
| | set from the server) |
| | |
| | In the format ``project_id:dataset_id``. |
| | """ |
| | return self._properties.get("id") |
| |
|
| | @property |
| | def friendly_name(self): |
| | """Union[str, None]: Title of the dataset as set by the user |
| | (defaults to :data:`None`). |
| | """ |
| | return self._properties.get("friendlyName") |
| |
|
| | @property |
| | def labels(self): |
| | """Dict[str, str]: Labels for the dataset.""" |
| | return self._properties.setdefault("labels", {}) |
| |
|
| | @property |
| | def reference(self): |
| | """google.cloud.bigquery.dataset.DatasetReference: A reference to this |
| | dataset. |
| | """ |
| | return DatasetReference(self.project, self.dataset_id) |
| |
|
| | table = _get_table_reference |
| |
|
| | model = _get_model_reference |
| |
|
| | routine = _get_routine_reference |
| |
|