| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | """System tests for Arrow connector.""" |
| |
|
| | from typing import Optional |
| |
|
| | import pyarrow |
| | import pytest |
| |
|
| | from google.cloud import bigquery |
| | from google.cloud.bigquery import enums |
| |
|
| |
|
| | @pytest.mark.parametrize( |
| | ("max_results", "scalars_table_name"), |
| | ( |
| | (None, "scalars_table"), |
| | (10, "scalars_table"), |
| | (None, "scalars_extreme_table"), |
| | (10, "scalars_extreme_table"), |
| | ), |
| | ) |
| | def test_list_rows_nullable_scalars_dtypes( |
| | bigquery_client: bigquery.Client, |
| | scalars_table: str, |
| | scalars_extreme_table: str, |
| | max_results: Optional[int], |
| | scalars_table_name: str, |
| | ): |
| | table_id = scalars_table |
| | if scalars_table_name == "scalars_extreme_table": |
| | table_id = scalars_extreme_table |
| |
|
| | |
| | |
| | schema = [ |
| | bigquery.SchemaField("bool_col", enums.SqlTypeNames.BOOLEAN), |
| | bigquery.SchemaField("bignumeric_col", enums.SqlTypeNames.BIGNUMERIC), |
| | bigquery.SchemaField("bytes_col", enums.SqlTypeNames.BYTES), |
| | bigquery.SchemaField("date_col", enums.SqlTypeNames.DATE), |
| | bigquery.SchemaField("datetime_col", enums.SqlTypeNames.DATETIME), |
| | bigquery.SchemaField("float64_col", enums.SqlTypeNames.FLOAT64), |
| | bigquery.SchemaField("geography_col", enums.SqlTypeNames.GEOGRAPHY), |
| | bigquery.SchemaField("int64_col", enums.SqlTypeNames.INT64), |
| | bigquery.SchemaField("numeric_col", enums.SqlTypeNames.NUMERIC), |
| | bigquery.SchemaField("string_col", enums.SqlTypeNames.STRING), |
| | bigquery.SchemaField("time_col", enums.SqlTypeNames.TIME), |
| | bigquery.SchemaField("timestamp_col", enums.SqlTypeNames.TIMESTAMP), |
| | ] |
| |
|
| | arrow_table = bigquery_client.list_rows( |
| | table_id, |
| | max_results=max_results, |
| | selected_fields=schema, |
| | ).to_arrow() |
| |
|
| | schema = arrow_table.schema |
| | bignumeric_type = schema.field("bignumeric_col").type |
| | |
| | |
| | assert bignumeric_type.precision in {76, 77} |
| | assert bignumeric_type.scale == 38 |
| |
|
| | bool_type = schema.field("bool_col").type |
| | assert bool_type.equals(pyarrow.bool_()) |
| |
|
| | bytes_type = schema.field("bytes_col").type |
| | assert bytes_type.equals(pyarrow.binary()) |
| |
|
| | date_type = schema.field("date_col").type |
| | assert date_type.equals(pyarrow.date32()) |
| |
|
| | datetime_type = schema.field("datetime_col").type |
| | assert datetime_type.unit == "us" |
| | assert datetime_type.tz is None |
| |
|
| | float64_type = schema.field("float64_col").type |
| | assert float64_type.equals(pyarrow.float64()) |
| |
|
| | geography_type = schema.field("geography_col").type |
| | assert geography_type.equals(pyarrow.string()) |
| |
|
| | int64_type = schema.field("int64_col").type |
| | assert int64_type.equals(pyarrow.int64()) |
| |
|
| | numeric_type = schema.field("numeric_col").type |
| | assert numeric_type.precision == 38 |
| | assert numeric_type.scale == 9 |
| |
|
| | string_type = schema.field("string_col").type |
| | assert string_type.equals(pyarrow.string()) |
| |
|
| | time_type = schema.field("time_col").type |
| | assert time_type.equals(pyarrow.time64("us")) |
| |
|
| | timestamp_type = schema.field("timestamp_col").type |
| | assert timestamp_type.unit == "us" |
| | assert timestamp_type.tz is not None |
| |
|
| |
|
| | @pytest.mark.parametrize("do_insert", [True, False]) |
| | def test_arrow_extension_types_same_for_storage_and_REST_APIs_894( |
| | dataset_client, test_table_name, do_insert |
| | ): |
| | types = dict( |
| | astring=("STRING", "'x'"), |
| | astring9=("STRING(9)", "'x'"), |
| | abytes=("BYTES", "b'x'"), |
| | abytes9=("BYTES(9)", "b'x'"), |
| | anumeric=("NUMERIC", "42"), |
| | anumeric9=("NUMERIC(9)", "42"), |
| | anumeric92=("NUMERIC(9,2)", "42"), |
| | abignumeric=("BIGNUMERIC", "42e30"), |
| | abignumeric49=("BIGNUMERIC(37)", "42e30"), |
| | abignumeric492=("BIGNUMERIC(37,2)", "42e30"), |
| | abool=("BOOL", "true"), |
| | adate=("DATE", "'2021-09-06'"), |
| | adatetime=("DATETIME", "'2021-09-06T09:57:26'"), |
| | ageography=("GEOGRAPHY", "ST_GEOGFROMTEXT('point(0 0)')"), |
| | |
| | |
| | aint64=("INT64", "42"), |
| | afloat64=("FLOAT64", "42.0"), |
| | astruct=("STRUCT<v int64>", "struct(42)"), |
| | atime=("TIME", "'1:2:3'"), |
| | atimestamp=("TIMESTAMP", "'2021-09-06T09:57:26'"), |
| | ) |
| | columns = ", ".join(f"{k} {t[0]}" for k, t in types.items()) |
| | dataset_client.query(f"create table {test_table_name} ({columns})").result() |
| | if do_insert: |
| | names = list(types) |
| | values = ", ".join(types[name][1] for name in names) |
| | names = ", ".join(names) |
| | dataset_client.query( |
| | f"insert into {test_table_name} ({names}) values ({values})" |
| | ).result() |
| | at = dataset_client.query(f"select * from {test_table_name}").result().to_arrow() |
| | storage_api_metadata = { |
| | at.field(i).name: at.field(i).metadata for i in range(at.num_columns) |
| | } |
| | at = ( |
| | dataset_client.query(f"select * from {test_table_name}") |
| | .result() |
| | .to_arrow(create_bqstorage_client=False) |
| | ) |
| | rest_api_metadata = { |
| | at.field(i).name: at.field(i).metadata for i in range(at.num_columns) |
| | } |
| |
|
| | assert rest_api_metadata == storage_api_metadata |
| | assert rest_api_metadata["adatetime"] == { |
| | b"ARROW:extension:name": b"google:sqlType:datetime" |
| | } |
| | assert rest_api_metadata["ageography"] == { |
| | b"ARROW:extension:name": b"google:sqlType:geography", |
| | b"ARROW:extension:metadata": b'{"encoding": "WKT"}', |
| | } |
| |
|
| |
|
| | def test_list_rows_range_csv( |
| | bigquery_client: bigquery.Client, |
| | scalars_table_csv: str, |
| | ): |
| | table_id = scalars_table_csv |
| |
|
| | schema = [ |
| | bigquery.SchemaField( |
| | "range_date", enums.SqlTypeNames.RANGE, range_element_type="DATE" |
| | ), |
| | ] |
| |
|
| | arrow_table = bigquery_client.list_rows( |
| | table_id, |
| | selected_fields=schema, |
| | ).to_arrow() |
| |
|
| | schema = arrow_table.schema |
| |
|
| | expected_type = pyarrow.struct( |
| | [("start", pyarrow.date32()), ("end", pyarrow.date32())] |
| | ) |
| |
|
| | range_type = schema.field("range_date").type |
| | assert range_type == expected_type |
| |
|