| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| """IPython Magics |
| |
| Install ``bigquery-magics`` and call ``%load_ext bigquery_magics`` to use the |
| ``%%bigquery`` cell magic. |
| |
| See the `BigQuery Magics reference documentation |
| <https://googleapis.dev/python/bigquery-magics/latest/>`_. |
| """ |
|
|
| from __future__ import print_function |
|
|
| import re |
| import ast |
| import copy |
| import functools |
| import sys |
| import time |
| import warnings |
| from concurrent import futures |
|
|
| try: |
| import IPython |
| from IPython import display |
| from IPython.core import magic_arguments |
| except ImportError: |
| raise ImportError("This module can only be loaded in IPython.") |
|
|
| from google.api_core import client_info |
| from google.api_core import client_options |
| from google.api_core.exceptions import NotFound |
| import google.auth |
| from google.cloud import bigquery |
| import google.cloud.bigquery.dataset |
| from google.cloud.bigquery import _versions_helpers |
| from google.cloud.bigquery import exceptions |
| from google.cloud.bigquery.dbapi import _helpers |
| from google.cloud.bigquery.magics import line_arg_parser as lap |
|
|
| try: |
| import bigquery_magics |
| except ImportError: |
| bigquery_magics = None |
|
|
|
|
| IPYTHON_USER_AGENT = "ipython-{}".format(IPython.__version__) |
|
|
|
|
| class Context(object): |
| """Storage for objects to be used throughout an IPython notebook session. |
| |
| A Context object is initialized when the ``magics`` module is imported, |
| and can be found at ``google.cloud.bigquery.magics.context``. |
| """ |
|
|
| def __init__(self): |
| self._credentials = None |
| self._project = None |
| self._connection = None |
| self._default_query_job_config = bigquery.QueryJobConfig() |
| self._bigquery_client_options = client_options.ClientOptions() |
| self._bqstorage_client_options = client_options.ClientOptions() |
| self._progress_bar_type = "tqdm_notebook" |
|
|
| @property |
| def credentials(self): |
| """google.auth.credentials.Credentials: Credentials to use for queries |
| performed through IPython magics. |
| |
| Note: |
| These credentials do not need to be explicitly defined if you are |
| using Application Default Credentials. If you are not using |
| Application Default Credentials, manually construct a |
| :class:`google.auth.credentials.Credentials` object and set it as |
| the context credentials as demonstrated in the example below. See |
| `auth docs`_ for more information on obtaining credentials. |
| |
| Example: |
| Manually setting the context credentials: |
| |
| >>> from google.cloud.bigquery import magics |
| >>> from google.oauth2 import service_account |
| >>> credentials = (service_account |
| ... .Credentials.from_service_account_file( |
| ... '/path/to/key.json')) |
| >>> magics.context.credentials = credentials |
| |
| |
| .. _auth docs: http://google-auth.readthedocs.io |
| /en/latest/user-guide.html#obtaining-credentials |
| """ |
| if self._credentials is None: |
| self._credentials, _ = google.auth.default() |
| return self._credentials |
|
|
| @credentials.setter |
| def credentials(self, value): |
| self._credentials = value |
|
|
| @property |
| def project(self): |
| """str: Default project to use for queries performed through IPython |
| magics. |
| |
| Note: |
| The project does not need to be explicitly defined if you have an |
| environment default project set. If you do not have a default |
| project set in your environment, manually assign the project as |
| demonstrated in the example below. |
| |
| Example: |
| Manually setting the context project: |
| |
| >>> from google.cloud.bigquery import magics |
| >>> magics.context.project = 'my-project' |
| """ |
| if self._project is None: |
| _, self._project = google.auth.default() |
| return self._project |
|
|
| @project.setter |
| def project(self, value): |
| self._project = value |
|
|
| @property |
| def bigquery_client_options(self): |
| """google.api_core.client_options.ClientOptions: client options to be |
| used through IPython magics. |
| |
| Note:: |
| The client options do not need to be explicitly defined if no |
| special network connections are required. Normally you would be |
| using the https://bigquery.googleapis.com/ end point. |
| |
| Example: |
| Manually setting the endpoint: |
| |
| >>> from google.cloud.bigquery import magics |
| >>> client_options = {} |
| >>> client_options['api_endpoint'] = "https://some.special.url" |
| >>> magics.context.bigquery_client_options = client_options |
| """ |
| return self._bigquery_client_options |
|
|
| @bigquery_client_options.setter |
| def bigquery_client_options(self, value): |
| self._bigquery_client_options = value |
|
|
| @property |
| def bqstorage_client_options(self): |
| """google.api_core.client_options.ClientOptions: client options to be |
| used through IPython magics for the storage client. |
| |
| Note:: |
| The client options do not need to be explicitly defined if no |
| special network connections are required. Normally you would be |
| using the https://bigquerystorage.googleapis.com/ end point. |
| |
| Example: |
| Manually setting the endpoint: |
| |
| >>> from google.cloud.bigquery import magics |
| >>> client_options = {} |
| >>> client_options['api_endpoint'] = "https://some.special.url" |
| >>> magics.context.bqstorage_client_options = client_options |
| """ |
| return self._bqstorage_client_options |
|
|
| @bqstorage_client_options.setter |
| def bqstorage_client_options(self, value): |
| self._bqstorage_client_options = value |
|
|
| @property |
| def default_query_job_config(self): |
| """google.cloud.bigquery.job.QueryJobConfig: Default job |
| configuration for queries. |
| |
| The context's :class:`~google.cloud.bigquery.job.QueryJobConfig` is |
| used for queries. Some properties can be overridden with arguments to |
| the magics. |
| |
| Example: |
| Manually setting the default value for ``maximum_bytes_billed`` |
| to 100 MB: |
| |
| >>> from google.cloud.bigquery import magics |
| >>> magics.context.default_query_job_config.maximum_bytes_billed = 100000000 |
| """ |
| return self._default_query_job_config |
|
|
| @default_query_job_config.setter |
| def default_query_job_config(self, value): |
| self._default_query_job_config = value |
|
|
| @property |
| def progress_bar_type(self): |
| """str: Default progress bar type to use to display progress bar while |
| executing queries through IPython magics. |
| |
| Note:: |
| Install the ``tqdm`` package to use this feature. |
| |
| Example: |
| Manually setting the progress_bar_type: |
| |
| >>> from google.cloud.bigquery import magics |
| >>> magics.context.progress_bar_type = "tqdm_notebook" |
| """ |
| return self._progress_bar_type |
|
|
| @progress_bar_type.setter |
| def progress_bar_type(self, value): |
| self._progress_bar_type = value |
|
|
|
|
| |
| |
| |
| |
| if bigquery_magics is not None: |
| context = bigquery_magics.context |
| else: |
| context = Context() |
|
|
|
|
| def _handle_error(error, destination_var=None): |
| """Process a query execution error. |
| |
| Args: |
| error (Exception): |
| An exception that occurred during the query execution. |
| destination_var (Optional[str]): |
| The name of the IPython session variable to store the query job. |
| """ |
| if destination_var: |
| query_job = getattr(error, "query_job", None) |
|
|
| if query_job is not None: |
| IPython.get_ipython().push({destination_var: query_job}) |
| else: |
| |
| |
| print( |
| "Could not save output to variable '{}'.".format(destination_var), |
| file=sys.stderr, |
| ) |
|
|
| print("\nERROR:\n", str(error), file=sys.stderr) |
|
|
|
|
| def _run_query(client, query, job_config=None): |
| """Runs a query while printing status updates |
| |
| Args: |
| client (google.cloud.bigquery.client.Client): |
| Client to bundle configuration needed for API requests. |
| query (str): |
| SQL query to be executed. Defaults to the standard SQL dialect. |
| Use the ``job_config`` parameter to change dialects. |
| job_config (Optional[google.cloud.bigquery.job.QueryJobConfig]): |
| Extra configuration options for the job. |
| |
| Returns: |
| google.cloud.bigquery.job.QueryJob: the query job created |
| |
| Example: |
| >>> client = bigquery.Client() |
| >>> _run_query(client, "SELECT 17") |
| Executing query with job ID: bf633912-af2c-4780-b568-5d868058632b |
| Query executing: 1.66s |
| Query complete after 2.07s |
| 'bf633912-af2c-4780-b568-5d868058632b' |
| """ |
| start_time = time.perf_counter() |
| query_job = client.query(query, job_config=job_config) |
|
|
| if job_config and job_config.dry_run: |
| return query_job |
|
|
| print(f"Executing query with job ID: {query_job.job_id}") |
|
|
| while True: |
| print( |
| f"\rQuery executing: {time.perf_counter() - start_time:.2f}s".format(), |
| end="", |
| ) |
| try: |
| query_job.result(timeout=0.5) |
| break |
| except futures.TimeoutError: |
| continue |
| print(f"\nJob ID {query_job.job_id} successfully executed") |
| return query_job |
|
|
|
|
| def _create_dataset_if_necessary(client, dataset_id): |
| """Create a dataset in the current project if it doesn't exist. |
| |
| Args: |
| client (google.cloud.bigquery.client.Client): |
| Client to bundle configuration needed for API requests. |
| dataset_id (str): |
| Dataset id. |
| """ |
| dataset_reference = bigquery.dataset.DatasetReference(client.project, dataset_id) |
| try: |
| dataset = client.get_dataset(dataset_reference) |
| return |
| except NotFound: |
| pass |
| dataset = bigquery.Dataset(dataset_reference) |
| dataset.location = client.location |
| print(f"Creating dataset: {dataset_id}") |
| dataset = client.create_dataset(dataset) |
|
|
|
|
| @magic_arguments.magic_arguments() |
| @magic_arguments.argument( |
| "destination_var", |
| nargs="?", |
| help=("If provided, save the output to this variable instead of displaying it."), |
| ) |
| @magic_arguments.argument( |
| "--destination_table", |
| type=str, |
| default=None, |
| help=( |
| "If provided, save the output of the query to a new BigQuery table. " |
| "Variable should be in a format <dataset_id>.<table_id>. " |
| "If table does not exists, it will be created. " |
| "If table already exists, its data will be overwritten." |
| ), |
| ) |
| @magic_arguments.argument( |
| "--project", |
| type=str, |
| default=None, |
| help=("Project to use for executing this query. Defaults to the context project."), |
| ) |
| @magic_arguments.argument( |
| "--max_results", |
| default=None, |
| help=( |
| "Maximum number of rows in dataframe returned from executing the query." |
| "Defaults to returning all rows." |
| ), |
| ) |
| @magic_arguments.argument( |
| "--maximum_bytes_billed", |
| default=None, |
| help=( |
| "maximum_bytes_billed to use for executing this query. Defaults to " |
| "the context default_query_job_config.maximum_bytes_billed." |
| ), |
| ) |
| @magic_arguments.argument( |
| "--dry_run", |
| action="store_true", |
| default=False, |
| help=( |
| "Sets query to be a dry run to estimate costs. " |
| "Defaults to executing the query instead of dry run if this argument is not used." |
| ), |
| ) |
| @magic_arguments.argument( |
| "--use_legacy_sql", |
| action="store_true", |
| default=False, |
| help=( |
| "Sets query to use Legacy SQL instead of Standard SQL. Defaults to " |
| "Standard SQL if this argument is not used." |
| ), |
| ) |
| @magic_arguments.argument( |
| "--bigquery_api_endpoint", |
| type=str, |
| default=None, |
| help=( |
| "The desired API endpoint, e.g., bigquery.googlepis.com. Defaults to this " |
| "option's value in the context bigquery_client_options." |
| ), |
| ) |
| @magic_arguments.argument( |
| "--bqstorage_api_endpoint", |
| type=str, |
| default=None, |
| help=( |
| "The desired API endpoint, e.g., bigquerystorage.googlepis.com. Defaults to " |
| "this option's value in the context bqstorage_client_options." |
| ), |
| ) |
| @magic_arguments.argument( |
| "--no_query_cache", |
| action="store_true", |
| default=False, |
| help=("Do not use cached query results."), |
| ) |
| @magic_arguments.argument( |
| "--use_bqstorage_api", |
| action="store_true", |
| default=None, |
| help=( |
| "[Deprecated] The BigQuery Storage API is already used by default to " |
| "download large query results, and this option has no effect. " |
| "If you want to switch to the classic REST API instead, use the " |
| "--use_rest_api option." |
| ), |
| ) |
| @magic_arguments.argument( |
| "--use_rest_api", |
| action="store_true", |
| default=False, |
| help=( |
| "Use the classic REST API instead of the BigQuery Storage API to " |
| "download query results." |
| ), |
| ) |
| @magic_arguments.argument( |
| "--verbose", |
| action="store_true", |
| default=False, |
| help=( |
| "If set, print verbose output, including the query job ID and the " |
| "amount of time for the query to finish. By default, this " |
| "information will be displayed as the query runs, but will be " |
| "cleared after the query is finished." |
| ), |
| ) |
| @magic_arguments.argument( |
| "--params", |
| nargs="+", |
| default=None, |
| help=( |
| "Parameters to format the query string. If present, the --params " |
| "flag should be followed by a string representation of a dictionary " |
| "in the format {'param_name': 'param_value'} (ex. {\"num\": 17}), " |
| "or a reference to a dictionary in the same format. The dictionary " |
| "reference can be made by including a '$' before the variable " |
| "name (ex. $my_dict_var)." |
| ), |
| ) |
| @magic_arguments.argument( |
| "--progress_bar_type", |
| type=str, |
| default=None, |
| help=( |
| "Sets progress bar type to display a progress bar while executing the query." |
| "Defaults to use tqdm_notebook. Install the ``tqdm`` package to use this feature." |
| ), |
| ) |
| @magic_arguments.argument( |
| "--location", |
| type=str, |
| default=None, |
| help=( |
| "Set the location to execute query." |
| "Defaults to location set in query setting in console." |
| ), |
| ) |
| def _cell_magic(line, query): |
| """Underlying function for bigquery cell magic |
| |
| Note: |
| This function contains the underlying logic for the 'bigquery' cell |
| magic. This function is not meant to be called directly. |
| |
| Args: |
| line (str): "%%bigquery" followed by arguments as required |
| query (str): SQL query to run |
| |
| Returns: |
| pandas.DataFrame: the query results. |
| """ |
| |
| |
| try: |
| params_option_value, rest_of_args = _split_args_line(line) |
| except lap.exceptions.QueryParamsParseError as exc: |
| rebranded_error = SyntaxError( |
| "--params is not a correctly formatted JSON string or a JSON " |
| "serializable dictionary" |
| ) |
| raise rebranded_error from exc |
| except lap.exceptions.DuplicateQueryParamsError as exc: |
| rebranded_error = ValueError("Duplicate --params option.") |
| raise rebranded_error from exc |
| except lap.exceptions.ParseError as exc: |
| rebranded_error = ValueError( |
| "Unrecognized input, are option values correct? " |
| "Error details: {}".format(exc.args[0]) |
| ) |
| raise rebranded_error from exc |
|
|
| args = magic_arguments.parse_argstring(_cell_magic, rest_of_args) |
|
|
| if args.use_bqstorage_api is not None: |
| warnings.warn( |
| "Deprecated option --use_bqstorage_api, the BigQuery " |
| "Storage API is already used by default.", |
| category=DeprecationWarning, |
| ) |
| use_bqstorage_api = not args.use_rest_api |
| location = args.location |
|
|
| params = [] |
| if params_option_value: |
| |
| |
| if params_option_value.startswith("$"): |
| msg = 'Parameter expansion failed, undefined variable "{}".'.format( |
| params_option_value[1:] |
| ) |
| raise NameError(msg) |
|
|
| params = _helpers.to_query_parameters(ast.literal_eval(params_option_value), {}) |
|
|
| project = args.project or context.project |
|
|
| bigquery_client_options = copy.deepcopy(context.bigquery_client_options) |
| if args.bigquery_api_endpoint: |
| if isinstance(bigquery_client_options, dict): |
| bigquery_client_options["api_endpoint"] = args.bigquery_api_endpoint |
| else: |
| bigquery_client_options.api_endpoint = args.bigquery_api_endpoint |
|
|
| client = bigquery.Client( |
| project=project, |
| credentials=context.credentials, |
| default_query_job_config=context.default_query_job_config, |
| client_info=client_info.ClientInfo(user_agent=IPYTHON_USER_AGENT), |
| client_options=bigquery_client_options, |
| location=location, |
| ) |
| if context._connection: |
| client._connection = context._connection |
|
|
| bqstorage_client_options = copy.deepcopy(context.bqstorage_client_options) |
| if args.bqstorage_api_endpoint: |
| if isinstance(bqstorage_client_options, dict): |
| bqstorage_client_options["api_endpoint"] = args.bqstorage_api_endpoint |
| else: |
| bqstorage_client_options.api_endpoint = args.bqstorage_api_endpoint |
|
|
| bqstorage_client = _make_bqstorage_client( |
| client, |
| use_bqstorage_api, |
| bqstorage_client_options, |
| ) |
|
|
| close_transports = functools.partial(_close_transports, client, bqstorage_client) |
|
|
| try: |
| if args.max_results: |
| max_results = int(args.max_results) |
| else: |
| max_results = None |
|
|
| query = query.strip() |
|
|
| if not query: |
| error = ValueError("Query is missing.") |
| _handle_error(error, args.destination_var) |
| return |
|
|
| |
| if query.startswith("$"): |
| query_var_name = query[1:] |
|
|
| if not query_var_name: |
| missing_msg = 'Missing query variable name, empty "$" is not allowed.' |
| raise NameError(missing_msg) |
|
|
| if query_var_name.isidentifier(): |
| ip = IPython.get_ipython() |
| query = ip.user_ns.get(query_var_name, ip) |
|
|
| if query is ip: |
| raise NameError( |
| f"Unknown query, variable {query_var_name} does not exist." |
| ) |
| else: |
| if not isinstance(query, (str, bytes)): |
| raise TypeError( |
| f"Query variable {query_var_name} must be a string " |
| "or a bytes-like value." |
| ) |
|
|
| |
| |
| if not re.search(r"\s", query): |
| try: |
| rows = client.list_rows(query, max_results=max_results) |
| except Exception as ex: |
| _handle_error(ex, args.destination_var) |
| return |
|
|
| result = rows.to_dataframe( |
| bqstorage_client=bqstorage_client, |
| create_bqstorage_client=False, |
| ) |
| if args.destination_var: |
| IPython.get_ipython().push({args.destination_var: result}) |
| return |
| else: |
| return result |
|
|
| job_config = bigquery.job.QueryJobConfig() |
| job_config.query_parameters = params |
| job_config.use_legacy_sql = args.use_legacy_sql |
| job_config.dry_run = args.dry_run |
|
|
| |
| if args.no_query_cache: |
| job_config.use_query_cache = False |
|
|
| if args.destination_table: |
| split = args.destination_table.split(".") |
| if len(split) != 2: |
| raise ValueError( |
| "--destination_table should be in a <dataset_id>.<table_id> format." |
| ) |
| dataset_id, table_id = split |
| job_config.allow_large_results = True |
| dataset_ref = bigquery.dataset.DatasetReference(client.project, dataset_id) |
| destination_table_ref = dataset_ref.table(table_id) |
| job_config.destination = destination_table_ref |
| job_config.create_disposition = "CREATE_IF_NEEDED" |
| job_config.write_disposition = "WRITE_TRUNCATE" |
| _create_dataset_if_necessary(client, dataset_id) |
|
|
| if args.maximum_bytes_billed == "None": |
| job_config.maximum_bytes_billed = 0 |
| elif args.maximum_bytes_billed is not None: |
| value = int(args.maximum_bytes_billed) |
| job_config.maximum_bytes_billed = value |
|
|
| try: |
| query_job = _run_query(client, query, job_config=job_config) |
| except Exception as ex: |
| _handle_error(ex, args.destination_var) |
| return |
|
|
| if not args.verbose: |
| display.clear_output() |
|
|
| if args.dry_run and args.destination_var: |
| IPython.get_ipython().push({args.destination_var: query_job}) |
| return |
| elif args.dry_run: |
| print( |
| "Query validated. This query will process {} bytes.".format( |
| query_job.total_bytes_processed |
| ) |
| ) |
| return query_job |
|
|
| progress_bar = context.progress_bar_type or args.progress_bar_type |
|
|
| if max_results: |
| result = query_job.result(max_results=max_results).to_dataframe( |
| bqstorage_client=None, |
| create_bqstorage_client=False, |
| progress_bar_type=progress_bar, |
| ) |
| else: |
| result = query_job.to_dataframe( |
| bqstorage_client=bqstorage_client, |
| create_bqstorage_client=False, |
| progress_bar_type=progress_bar, |
| ) |
|
|
| if args.destination_var: |
| IPython.get_ipython().push({args.destination_var: result}) |
| else: |
| return result |
| finally: |
| close_transports() |
|
|
|
|
| def _split_args_line(line): |
| """Split out the --params option value from the input line arguments. |
| |
| Args: |
| line (str): The line arguments passed to the cell magic. |
| |
| Returns: |
| Tuple[str, str] |
| """ |
| lexer = lap.Lexer(line) |
| scanner = lap.Parser(lexer) |
| tree = scanner.input_line() |
|
|
| extractor = lap.QueryParamsExtractor() |
| params_option_value, rest_of_args = extractor.visit(tree) |
|
|
| return params_option_value, rest_of_args |
|
|
|
|
| def _make_bqstorage_client(client, use_bqstorage_api, client_options): |
| """Creates a BigQuery Storage client. |
| |
| Args: |
| client (:class:`~google.cloud.bigquery.client.Client`): BigQuery client. |
| use_bqstorage_api (bool): whether BigQuery Storage API is used or not. |
| client_options (:class:`google.api_core.client_options.ClientOptions`): |
| Custom options used with a new BigQuery Storage client instance |
| if one is created. |
| |
| Raises: |
| ImportError: if google-cloud-bigquery-storage is not installed, or |
| grpcio package is not installed. |
| |
| |
| Returns: |
| None: if ``use_bqstorage_api == False``, or google-cloud-bigquery-storage |
| is outdated. |
| BigQuery Storage Client: |
| """ |
| if not use_bqstorage_api: |
| return None |
|
|
| try: |
| _versions_helpers.BQ_STORAGE_VERSIONS.try_import(raise_if_error=True) |
| except exceptions.BigQueryStorageNotFoundError as err: |
| customized_error = ImportError( |
| "The default BigQuery Storage API client cannot be used, install " |
| "the missing google-cloud-bigquery-storage and pyarrow packages " |
| "to use it. Alternatively, use the classic REST API by specifying " |
| "the --use_rest_api magic option." |
| ) |
| raise customized_error from err |
| except exceptions.LegacyBigQueryStorageError: |
| pass |
|
|
| try: |
| from google.api_core.gapic_v1 import client_info as gapic_client_info |
| except ImportError as err: |
| customized_error = ImportError( |
| "Install the grpcio package to use the BigQuery Storage API." |
| ) |
| raise customized_error from err |
|
|
| return client._ensure_bqstorage_client( |
| client_options=client_options, |
| client_info=gapic_client_info.ClientInfo(user_agent=IPYTHON_USER_AGENT), |
| ) |
|
|
|
|
| def _close_transports(client, bqstorage_client): |
| """Close the given clients' underlying transport channels. |
| |
| Closing the transport is needed to release system resources, namely open |
| sockets. |
| |
| Args: |
| client (:class:`~google.cloud.bigquery.client.Client`): |
| bqstorage_client |
| (Optional[:class:`~google.cloud.bigquery_storage.BigQueryReadClient`]): |
| A client for the BigQuery Storage API. |
| |
| """ |
| client.close() |
| if bqstorage_client is not None: |
| bqstorage_client._transport.grpc_channel.close() |
|
|