hc99's picture
Add files using upload-large-folder tool
1856027 verified
# Copyright 2021 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import concurrent.futures
import datetime
import decimal
from typing import Tuple
from google.api_core import exceptions
import pytest
from google.cloud import bigquery
from google.cloud.bigquery.query import ArrayQueryParameter
from google.cloud.bigquery.query import ScalarQueryParameter
from google.cloud.bigquery.query import ScalarQueryParameterType
from google.cloud.bigquery.query import StructQueryParameter
from google.cloud.bigquery.query import StructQueryParameterType
from google.cloud.bigquery.query import RangeQueryParameter
@pytest.fixture(params=["INSERT", "QUERY"])
def query_api_method(request):
return request.param
@pytest.fixture(scope="session")
def table_with_9999_columns_10_rows(bigquery_client, project_id, dataset_id):
"""Generate a table of maximum width via CREATE TABLE AS SELECT.
The first column is named 'rowval', and has a value from 1..rowcount
Subsequent columns are named col_<N> and contain the value N*rowval, where
N is between 1 and 9999 inclusive.
"""
table_id = "many_columns"
row_count = 10
col_projections = ",".join(f"r * {n} as col_{n}" for n in range(1, 10000))
sql = f"""
CREATE TABLE `{project_id}.{dataset_id}.{table_id}`
AS
SELECT
r as rowval,
{col_projections}
FROM
UNNEST(GENERATE_ARRAY(1,{row_count},1)) as r
"""
query_job = bigquery_client.query(sql)
query_job.result()
return f"{project_id}.{dataset_id}.{table_id}"
def test_query_many_columns(
bigquery_client, table_with_9999_columns_10_rows, query_api_method
):
# Test working with the widest schema BigQuery supports, 10k columns.
query_job = bigquery_client.query(
f"SELECT * FROM `{table_with_9999_columns_10_rows}`",
api_method=query_api_method,
)
rows = list(query_job)
assert len(rows) == 10
# check field representations adhere to expected values.
for row in rows:
rowval = row["rowval"]
for column in range(1, 10000):
assert row[f"col_{column}"] == rowval * column
def test_query_w_timeout(bigquery_client, query_api_method):
job_config = bigquery.QueryJobConfig()
job_config.use_query_cache = False
query_job = bigquery_client.query(
"SELECT * FROM `bigquery-public-data.github_repos.commits`;",
location="US",
job_config=job_config,
api_method=query_api_method,
)
with pytest.raises(concurrent.futures.TimeoutError):
query_job.result(timeout=1)
# Even though the query takes >1 second, the call to getQueryResults
# should succeed.
assert not query_job.done(timeout=1)
assert bigquery_client.cancel_job(query_job) is not None
def test_query_statistics(bigquery_client, query_api_method):
"""
A system test to exercise some of the extended query statistics.
Note: We construct a query that should need at least three stages by
specifying a JOIN query. Exact plan and stats are effectively
non-deterministic, so we're largely interested in confirming values
are present.
"""
job_config = bigquery.QueryJobConfig()
job_config.use_query_cache = False
query_job = bigquery_client.query(
"""
SELECT
COUNT(1)
FROM
(
SELECT
year,
wban_number
FROM `bigquery-public-data.samples.gsod`
LIMIT 1000
) lside
INNER JOIN
(
SELECT
year,
state
FROM `bigquery-public-data.samples.natality`
LIMIT 1000
) rside
ON
lside.year = rside.year
""",
location="US",
job_config=job_config,
api_method=query_api_method,
)
# run the job to completion
query_job.result()
# Must reload job to get stats if jobs.query was used.
if query_api_method == "QUERY":
query_job.reload()
# Assert top-level stats
assert not query_job.cache_hit
assert query_job.destination is not None
assert query_job.done
assert not query_job.dry_run
assert query_job.num_dml_affected_rows is None
assert query_job.priority == "INTERACTIVE"
assert query_job.total_bytes_billed > 1
assert query_job.total_bytes_processed > 1
assert query_job.statement_type == "SELECT"
assert query_job.slot_millis > 1
# Make assertions on the shape of the query plan.
plan = query_job.query_plan
assert len(plan) >= 3
first_stage = plan[0]
assert first_stage.start is not None
assert first_stage.end is not None
assert first_stage.entry_id is not None
assert first_stage.name is not None
assert first_stage.parallel_inputs > 0
assert first_stage.completed_parallel_inputs > 0
assert first_stage.shuffle_output_bytes > 0
assert first_stage.status == "COMPLETE"
# Query plan is a digraph. Ensure it has inter-stage links,
# but not every stage has inputs.
stages_with_inputs = 0
for entry in plan:
if len(entry.input_stages) > 0:
stages_with_inputs = stages_with_inputs + 1
assert stages_with_inputs > 0
assert len(plan) > stages_with_inputs
@pytest.mark.parametrize(
("sql", "expected", "query_parameters"),
(
(
"SELECT @question",
"What is the answer to life, the universe, and everything?",
[
ScalarQueryParameter(
name="question",
type_="STRING",
value="What is the answer to life, the universe, and everything?",
)
],
),
(
"SELECT @answer",
42,
[ScalarQueryParameter(name="answer", type_="INT64", value=42)],
),
(
"SELECT @pi",
3.1415926,
[ScalarQueryParameter(name="pi", type_="FLOAT64", value=3.1415926)],
),
(
"SELECT @pi_numeric_param",
decimal.Decimal("3.141592654"),
[
ScalarQueryParameter(
name="pi_numeric_param",
type_="NUMERIC",
value=decimal.Decimal("3.141592654"),
)
],
),
(
"SELECT @bignum_param",
decimal.Decimal("-{d38}.{d38}".format(d38="9" * 38)),
[
ScalarQueryParameter(
name="bignum_param",
type_="BIGNUMERIC",
value=decimal.Decimal("-{d38}.{d38}".format(d38="9" * 38)),
)
],
),
(
"SELECT @truthy",
True,
[ScalarQueryParameter(name="truthy", type_="BOOL", value=True)],
),
(
"SELECT @beef",
b"DEADBEEF",
[ScalarQueryParameter(name="beef", type_="BYTES", value=b"DEADBEEF")],
),
(
"SELECT @naive",
datetime.datetime(2016, 12, 5, 12, 41, 9),
[
ScalarQueryParameter(
name="naive",
type_="DATETIME",
value=datetime.datetime(2016, 12, 5, 12, 41, 9),
)
],
),
(
"SELECT @naive_date",
datetime.date(2016, 12, 5),
[
ScalarQueryParameter(
name="naive_date", type_="DATE", value=datetime.date(2016, 12, 5)
)
],
),
pytest.param(
"SELECT @json",
{"alpha": "abc", "num": [1, 2, 3]},
[
ScalarQueryParameter(
name="json",
type_="JSON",
value={"alpha": "abc", "num": [1, 2, 3]},
)
],
id="scalar-json",
),
(
"SELECT @naive_time",
datetime.time(12, 41, 9, 62500),
[
ScalarQueryParameter(
name="naive_time",
type_="TIME",
value=datetime.time(12, 41, 9, 62500),
)
],
),
(
"SELECT @zoned",
datetime.datetime(2016, 12, 5, 12, 41, 9, tzinfo=datetime.timezone.utc),
[
ScalarQueryParameter(
name="zoned",
type_="TIMESTAMP",
value=datetime.datetime(
2016, 12, 5, 12, 41, 9, tzinfo=datetime.timezone.utc
),
)
],
),
(
"SELECT @array_param",
[1, 2],
[
ArrayQueryParameter(
name="array_param", array_type="INT64", values=[1, 2]
)
],
),
(
"SELECT (@hitchhiker.question, @hitchhiker.answer)",
({"_field_1": "What is the answer?", "_field_2": 42}),
[
StructQueryParameter(
"hitchhiker",
ScalarQueryParameter(
name="question",
type_="STRING",
value="What is the answer?",
),
ScalarQueryParameter(
name="answer",
type_="INT64",
value=42,
),
),
],
),
(
"SELECT "
"((@rectangle.bottom_right.x - @rectangle.top_left.x) "
"* (@rectangle.top_left.y - @rectangle.bottom_right.y))",
100,
[
StructQueryParameter(
"rectangle",
StructQueryParameter(
"top_left",
ScalarQueryParameter("x", "INT64", 12),
ScalarQueryParameter("y", "INT64", 102),
),
StructQueryParameter(
"bottom_right",
ScalarQueryParameter("x", "INT64", 22),
ScalarQueryParameter("y", "INT64", 92),
),
)
],
),
(
"SELECT ?",
[
{"name": "Phred Phlyntstone", "age": 32},
{"name": "Bharney Rhubbyl", "age": 31},
],
[
ArrayQueryParameter(
name=None,
array_type="RECORD",
values=[
StructQueryParameter(
None,
ScalarQueryParameter(
name="name", type_="STRING", value="Phred Phlyntstone"
),
ScalarQueryParameter(name="age", type_="INT64", value=32),
),
StructQueryParameter(
None,
ScalarQueryParameter(
name="name", type_="STRING", value="Bharney Rhubbyl"
),
ScalarQueryParameter(name="age", type_="INT64", value=31),
),
],
)
],
),
(
"SELECT @empty_array_param",
[],
[
ArrayQueryParameter(
name="empty_array_param",
values=[],
array_type=StructQueryParameterType(
ScalarQueryParameterType(name="foo", type_="INT64"),
ScalarQueryParameterType(name="bar", type_="STRING"),
),
)
],
),
(
"SELECT @roles",
{
"hero": {"name": "Phred Phlyntstone", "age": 32},
"sidekick": {"name": "Bharney Rhubbyl", "age": 31},
},
[
StructQueryParameter(
"roles",
StructQueryParameter(
"hero",
ScalarQueryParameter(
name="name", type_="STRING", value="Phred Phlyntstone"
),
ScalarQueryParameter(name="age", type_="INT64", value=32),
),
StructQueryParameter(
"sidekick",
ScalarQueryParameter(
name="name", type_="STRING", value="Bharney Rhubbyl"
),
ScalarQueryParameter(name="age", type_="INT64", value=31),
),
),
],
),
(
"SELECT ?",
{"friends": ["Jack", "Jill"]},
[
StructQueryParameter(
None,
ArrayQueryParameter(
name="friends", array_type="STRING", values=["Jack", "Jill"]
),
)
],
),
(
"SELECT @range_date",
{"end": None, "start": datetime.date(2016, 12, 5)},
[
RangeQueryParameter(
name="range_date",
range_element_type="DATE",
start=datetime.date(2016, 12, 5),
)
],
),
(
"SELECT @range_datetime",
{"end": None, "start": datetime.datetime(2016, 12, 5, 0, 0)},
[
RangeQueryParameter(
name="range_datetime",
range_element_type="DATETIME",
start=datetime.datetime(2016, 12, 5),
)
],
),
(
"SELECT @range_unbounded",
{"end": None, "start": None},
[
RangeQueryParameter(
name="range_unbounded",
range_element_type="DATETIME",
)
],
),
),
)
def test_query_parameters(
bigquery_client, query_api_method, sql, expected, query_parameters
):
jconfig = bigquery.QueryJobConfig()
jconfig.query_parameters = query_parameters
query_job = bigquery_client.query(
sql,
job_config=jconfig,
api_method=query_api_method,
)
rows = list(query_job.result())
assert len(rows) == 1
assert len(rows[0]) == 1
assert rows[0][0] == expected
def test_dry_run(
bigquery_client: bigquery.Client,
query_api_method: str,
scalars_table_multi_location: Tuple[str, str],
):
location, full_table_id = scalars_table_multi_location
query_config = bigquery.QueryJobConfig()
query_config.dry_run = True
query_string = f"SELECT * FROM {full_table_id}"
query_job = bigquery_client.query(
query_string,
location=location,
job_config=query_config,
api_method=query_api_method,
)
# Note: `query_job.result()` is not necessary on a dry run query. All
# necessary information is returned in the initial response.
assert query_job.dry_run is True
assert query_job.total_bytes_processed > 0
assert len(query_job.schema) > 0
def test_query_error_w_api_method_query(bigquery_client: bigquery.Client):
"""No job is returned from jobs.query if the query fails."""
with pytest.raises(exceptions.NotFound, match="not_a_real_dataset"):
bigquery_client.query(
"SELECT * FROM not_a_real_dataset.doesnt_exist", api_method="QUERY"
)
def test_query_error_w_api_method_default(bigquery_client: bigquery.Client):
"""Test that an exception is not thrown until fetching the results.
For backwards compatibility, jobs.insert is the default API method. With
jobs.insert, a failed query job is "successfully" created. An exception is
thrown when fetching the results.
"""
query_job = bigquery_client.query("SELECT * FROM not_a_real_dataset.doesnt_exist")
with pytest.raises(exceptions.NotFound, match="not_a_real_dataset"):
query_job.result()
def test_session(bigquery_client: bigquery.Client, query_api_method: str):
initial_config = bigquery.QueryJobConfig()
initial_config.create_session = True
initial_query = """
CREATE TEMPORARY TABLE numbers(id INT64)
AS
SELECT * FROM UNNEST([1, 2, 3, 4, 5]) AS id;
"""
initial_job = bigquery_client.query(
initial_query, job_config=initial_config, api_method=query_api_method
)
initial_job.result()
session_id = initial_job.session_info.session_id
assert session_id is not None
second_config = bigquery.QueryJobConfig()
second_config.connection_properties = [
bigquery.ConnectionProperty("session_id", session_id),
]
second_job = bigquery_client.query(
"SELECT COUNT(*) FROM numbers;", job_config=second_config
)
rows = list(second_job.result())
assert len(rows) == 1
assert rows[0][0] == 5