| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | import functools |
| | import operator as op |
| | import unittest |
| | from unittest import mock |
| |
|
| | import pytest |
| |
|
| | import google.cloud.bigquery.table as bq_table |
| |
|
| | from google.api_core import exceptions |
| |
|
| | from tests.unit.helpers import _to_pyarrow |
| |
|
| |
|
| | class TestCursor(unittest.TestCase): |
| | @staticmethod |
| | def _get_target_class(): |
| | from google.cloud.bigquery.dbapi import Cursor |
| |
|
| | return Cursor |
| |
|
| | def _make_one(self, *args, **kw): |
| | return self._get_target_class()(*args, **kw) |
| |
|
| | def _mock_client( |
| | self, |
| | rows=None, |
| | schema=None, |
| | num_dml_affected_rows=None, |
| | dry_run_job=False, |
| | total_bytes_processed=0, |
| | total_rows=None, |
| | destination_table="test-project.test_dataset.test_table", |
| | ): |
| | from google.cloud.bigquery import client |
| |
|
| | if total_rows is None: |
| | total_rows = 0 |
| | if rows is not None: |
| | total_rows = len(rows) |
| |
|
| | table = bq_table.TableReference.from_string(destination_table) |
| | mock_client = mock.create_autospec(client.Client) |
| | mock_job = self._mock_job( |
| | total_rows=total_rows, |
| | schema=schema, |
| | num_dml_affected_rows=num_dml_affected_rows, |
| | dry_run=dry_run_job, |
| | total_bytes_processed=total_bytes_processed, |
| | rows=self._mock_rows( |
| | rows, |
| | total_rows=total_rows, |
| | schema=schema, |
| | num_dml_affected_rows=num_dml_affected_rows, |
| | table=table, |
| | ), |
| | ) |
| | mock_client.get_job.return_value = mock_job |
| | mock_client.query.return_value = mock_job |
| | mock_client.query_and_wait.return_value = self._mock_rows( |
| | rows, |
| | total_rows=total_rows, |
| | schema=schema, |
| | num_dml_affected_rows=num_dml_affected_rows, |
| | |
| | |
| | table=table if rows is not None and total_rows > len(rows) else None, |
| | ) |
| |
|
| | |
| | mock_client._ensure_bqstorage_client.return_value = None |
| |
|
| | return mock_client |
| |
|
| | def _mock_bqstorage_client(self, rows=None, stream_count=0): |
| | from google.cloud import bigquery_storage |
| |
|
| | if rows is None: |
| | rows = [] |
| |
|
| | mock_client = mock.create_autospec(bigquery_storage.BigQueryReadClient) |
| | mock_read_session = mock.MagicMock( |
| | streams=[ |
| | bigquery_storage.types.ReadStream(name="streams/stream_{}".format(i)) |
| | for i in range(stream_count) |
| | ] |
| | ) |
| |
|
| | mock_client.create_read_session.return_value = mock_read_session |
| |
|
| | mock_rows_stream = mock.MagicMock() |
| | mock_rows_stream.rows.return_value = iter(rows) |
| | mock_client.read_rows.return_value = mock_rows_stream |
| |
|
| | return mock_client |
| |
|
| | def _mock_job( |
| | self, |
| | total_rows=0, |
| | schema=None, |
| | num_dml_affected_rows=None, |
| | dry_run=False, |
| | total_bytes_processed=0, |
| | rows=None, |
| | ): |
| | from google.cloud.bigquery import job |
| |
|
| | mock_job = mock.create_autospec(job.QueryJob) |
| | mock_job.error_result = None |
| | mock_job.state = "DONE" |
| | mock_job.dry_run = dry_run |
| |
|
| | if dry_run: |
| | mock_job.result.side_effect = exceptions.NotFound |
| | mock_job.total_bytes_processed = total_bytes_processed |
| | else: |
| | mock_job.result.return_value = rows |
| | mock_job._query_results = self._mock_results( |
| | total_rows=total_rows, |
| | schema=schema, |
| | num_dml_affected_rows=num_dml_affected_rows, |
| | ) |
| | mock_job.destination.project = "P" |
| | mock_job.destination.to_bqstorage.return_value = ( |
| | "projects/P/datasets/DS/tables/T" |
| | ) |
| |
|
| | if num_dml_affected_rows is None: |
| | mock_job.statement_type = None |
| | else: |
| | mock_job.statement_type = "UPDATE" |
| |
|
| | return mock_job |
| |
|
| | def _mock_rows( |
| | self, rows, total_rows=0, schema=None, num_dml_affected_rows=None, table=None |
| | ): |
| | mock_rows = mock.create_autospec(bq_table.RowIterator, instance=True) |
| | mock_rows.__iter__.return_value = rows |
| | mock_rows._table = table |
| | mock_rows._should_use_bqstorage = functools.partial( |
| | bq_table.RowIterator._should_use_bqstorage, |
| | mock_rows, |
| | ) |
| | mock_rows._is_almost_completely_cached = functools.partial( |
| | bq_table.RowIterator._is_almost_completely_cached, |
| | mock_rows, |
| | ) |
| | mock_rows.max_results = None |
| | type(mock_rows).job_id = mock.PropertyMock(return_value="test-job-id") |
| | type(mock_rows).location = mock.PropertyMock(return_value="test-location") |
| | type(mock_rows).num_dml_affected_rows = mock.PropertyMock( |
| | return_value=num_dml_affected_rows |
| | ) |
| | type(mock_rows).total_rows = mock.PropertyMock(return_value=total_rows) |
| | type(mock_rows).schema = mock.PropertyMock(return_value=schema) |
| | return mock_rows |
| |
|
| | def _mock_results(self, total_rows=0, schema=None, num_dml_affected_rows=None): |
| | from google.cloud.bigquery import query |
| |
|
| | mock_results = mock.create_autospec(query._QueryResults) |
| | mock_results.schema = schema |
| | mock_results.num_dml_affected_rows = num_dml_affected_rows |
| | mock_results.total_rows = total_rows |
| | return mock_results |
| |
|
| | def test_ctor(self): |
| | from google.cloud.bigquery.dbapi import connect |
| | from google.cloud.bigquery.dbapi import Cursor |
| |
|
| | connection = connect(self._mock_client()) |
| | cursor = self._make_one(connection) |
| | self.assertIsInstance(cursor, Cursor) |
| | self.assertIs(cursor.connection, connection) |
| |
|
| | def test_close(self): |
| | from google.cloud.bigquery.dbapi import connect |
| |
|
| | connection = connect(self._mock_client()) |
| | cursor = connection.cursor() |
| | |
| | cursor.close() |
| |
|
| | def test_raises_error_if_closed(self): |
| | from google.cloud.bigquery.dbapi import connect |
| | from google.cloud.bigquery.dbapi.exceptions import ProgrammingError |
| |
|
| | connection = connect(self._mock_client()) |
| | cursor = connection.cursor() |
| | cursor.close() |
| |
|
| | method_names = ( |
| | "close", |
| | "execute", |
| | "executemany", |
| | "fetchall", |
| | "fetchmany", |
| | "fetchone", |
| | "setinputsizes", |
| | "setoutputsize", |
| | "__iter__", |
| | ) |
| |
|
| | for method in method_names: |
| | with self.assertRaisesRegex( |
| | ProgrammingError, r"Operating on a closed cursor\." |
| | ): |
| | getattr(cursor, method)() |
| |
|
| | def test_fetchone_wo_execute_raises_error(self): |
| | from google.cloud.bigquery import dbapi |
| |
|
| | connection = dbapi.connect(self._mock_client()) |
| | cursor = connection.cursor() |
| | self.assertRaises(dbapi.Error, cursor.fetchone) |
| |
|
| | def test_fetchone_w_row(self): |
| | from google.cloud.bigquery import dbapi |
| |
|
| | connection = dbapi.connect(self._mock_client(rows=[(1,)])) |
| | cursor = connection.cursor() |
| | cursor.execute("SELECT 1;") |
| | row = cursor.fetchone() |
| | self.assertEqual(row, (1,)) |
| | self.assertIsNone(cursor.fetchone()) |
| |
|
| | def test_fetchmany_wo_execute_raises_error(self): |
| | from google.cloud.bigquery import dbapi |
| |
|
| | connection = dbapi.connect(self._mock_client()) |
| | cursor = connection.cursor() |
| | self.assertRaises(dbapi.Error, cursor.fetchmany) |
| |
|
| | def test_fetchmany_w_row(self): |
| | from google.cloud.bigquery import dbapi |
| |
|
| | connection = dbapi.connect(self._mock_client(rows=[(1,)])) |
| | cursor = connection.cursor() |
| | cursor.execute("SELECT 1;") |
| | rows = cursor.fetchmany() |
| | self.assertEqual(len(rows), 1) |
| | self.assertEqual(rows[0], (1,)) |
| |
|
| | def test_fetchmany_w_size(self): |
| | from google.cloud.bigquery import dbapi |
| |
|
| | connection = dbapi.connect( |
| | self._mock_client(rows=[(1, 2, 3), (4, 5, 6), (7, 8, 9)]) |
| | ) |
| | cursor = connection.cursor() |
| | cursor.execute("SELECT a, b, c;") |
| | rows = cursor.fetchmany(size=2) |
| | self.assertEqual(len(rows), 2) |
| | self.assertEqual(rows[0], (1, 2, 3)) |
| | self.assertEqual(rows[1], (4, 5, 6)) |
| | second_page = cursor.fetchmany(size=2) |
| | self.assertEqual(len(second_page), 1) |
| | self.assertEqual(second_page[0], (7, 8, 9)) |
| | third_page = cursor.fetchmany(size=2) |
| | self.assertEqual(third_page, []) |
| |
|
| | def test_fetchmany_w_arraysize(self): |
| | from google.cloud.bigquery import dbapi |
| |
|
| | connection = dbapi.connect( |
| | self._mock_client(rows=[(1, 2, 3), (4, 5, 6), (7, 8, 9)]) |
| | ) |
| | cursor = connection.cursor() |
| | cursor.execute("SELECT a, b, c;") |
| | cursor.arraysize = 2 |
| | rows = cursor.fetchmany() |
| | self.assertEqual(len(rows), 2) |
| | self.assertEqual(rows[0], (1, 2, 3)) |
| | self.assertEqual(rows[1], (4, 5, 6)) |
| | second_page = cursor.fetchmany() |
| | self.assertEqual(len(second_page), 1) |
| | self.assertEqual(second_page[0], (7, 8, 9)) |
| | third_page = cursor.fetchmany() |
| | self.assertEqual(third_page, []) |
| |
|
| | def test_fetchall_wo_execute_raises_error(self): |
| | from google.cloud.bigquery import dbapi |
| |
|
| | connection = dbapi.connect(self._mock_client()) |
| | cursor = connection.cursor() |
| | self.assertRaises(dbapi.Error, cursor.fetchall) |
| |
|
| | def test_fetchall_w_row(self): |
| | from google.cloud.bigquery import dbapi |
| |
|
| | connection = dbapi.connect(self._mock_client(rows=[(1,)])) |
| | cursor = connection.cursor() |
| | cursor.execute("SELECT 1;") |
| | self.assertIsNone(cursor.description) |
| | self.assertEqual(cursor.rowcount, 1) |
| | rows = cursor.fetchall() |
| | self.assertEqual(len(rows), 1) |
| | self.assertEqual(rows[0], (1,)) |
| |
|
| | def test_fetchall_w_bqstorage_client_fetch_success(self): |
| | pytest.importorskip("google.cloud.bigquery_storage") |
| | pytest.importorskip("pyarrow") |
| | from google.cloud.bigquery import dbapi |
| |
|
| | |
| | row_data = [ |
| | bq_table.Row( |
| | [1.4, 1.1, 1.3, 1.2], {"bar": 3, "baz": 2, "foo": 1, "quux": 0} |
| | ), |
| | bq_table.Row( |
| | [2.4, 2.1, 2.3, 2.2], {"bar": 3, "baz": 2, "foo": 1, "quux": 0} |
| | ), |
| | ] |
| | bqstorage_streamed_rows = [ |
| | { |
| | "bar": _to_pyarrow(1.2), |
| | "foo": _to_pyarrow(1.1), |
| | "quux": _to_pyarrow(1.4), |
| | "baz": _to_pyarrow(1.3), |
| | }, |
| | { |
| | "bar": _to_pyarrow(2.2), |
| | "foo": _to_pyarrow(2.1), |
| | "quux": _to_pyarrow(2.4), |
| | "baz": _to_pyarrow(2.3), |
| | }, |
| | ] |
| |
|
| | mock_client = self._mock_client(rows=row_data) |
| | mock_bqstorage_client = self._mock_bqstorage_client( |
| | stream_count=1, |
| | rows=bqstorage_streamed_rows, |
| | ) |
| | mock_client._ensure_bqstorage_client.return_value = mock_bqstorage_client |
| |
|
| | connection = dbapi.connect( |
| | client=mock_client, |
| | bqstorage_client=mock_bqstorage_client, |
| | ) |
| | cursor = connection.cursor() |
| | cursor.execute("SELECT foo, bar FROM some_table") |
| |
|
| | rows = cursor.fetchall() |
| |
|
| | |
| | mock_client.list_rows.assert_not_called() |
| |
|
| | |
| | field_value = op.itemgetter(1) |
| | sorted_row_data = [sorted(row.items(), key=field_value) for row in rows] |
| | expected_row_data = [ |
| | [("foo", 1.1), ("bar", 1.2), ("baz", 1.3), ("quux", 1.4)], |
| | [("foo", 2.1), ("bar", 2.2), ("baz", 2.3), ("quux", 2.4)], |
| | ] |
| |
|
| | self.assertEqual(sorted_row_data, expected_row_data) |
| |
|
| | def test_fetchall_w_bqstorage_client_fetch_no_rows(self): |
| | pytest.importorskip("google.cloud.bigquery_storage") |
| | from google.cloud.bigquery import dbapi |
| |
|
| | mock_client = self._mock_client( |
| | rows=[], |
| | |
| | |
| | total_rows=1000, |
| | ) |
| | mock_bqstorage_client = self._mock_bqstorage_client(stream_count=0) |
| | mock_client._ensure_bqstorage_client.return_value = mock_bqstorage_client |
| |
|
| | connection = dbapi.connect( |
| | client=mock_client, |
| | bqstorage_client=mock_bqstorage_client, |
| | ) |
| | cursor = connection.cursor() |
| | cursor.execute("SELECT foo, bar FROM some_table") |
| |
|
| | rows = cursor.fetchall() |
| |
|
| | |
| | mock_client.list_rows.assert_not_called() |
| |
|
| | |
| | self.assertEqual(rows, []) |
| |
|
| | def test_fetchall_w_bqstorage_client_fetch_error_no_fallback(self): |
| | pytest.importorskip("google.cloud.bigquery_storage") |
| | from google.cloud.bigquery import dbapi |
| |
|
| | row_data = [bq_table.Row([1.1, 1.2], {"foo": 0, "bar": 1})] |
| |
|
| | def fake_ensure_bqstorage_client(bqstorage_client=None, **kwargs): |
| | return bqstorage_client |
| |
|
| | mock_client = self._mock_client( |
| | rows=row_data, |
| | |
| | |
| | total_rows=1000, |
| | ) |
| | mock_client._ensure_bqstorage_client.side_effect = fake_ensure_bqstorage_client |
| | mock_bqstorage_client = self._mock_bqstorage_client( |
| | stream_count=1, |
| | rows=row_data, |
| | ) |
| | no_access_error = exceptions.Forbidden("invalid credentials") |
| | mock_bqstorage_client.create_read_session.side_effect = no_access_error |
| |
|
| | connection = dbapi.connect( |
| | client=mock_client, |
| | bqstorage_client=mock_bqstorage_client, |
| | ) |
| | cursor = connection.cursor() |
| | cursor.execute("SELECT foo, bar FROM some_table") |
| |
|
| | with self.assertRaisesRegex(exceptions.Forbidden, "invalid credentials"): |
| | cursor.fetchall() |
| |
|
| | |
| | mock_client.list_rows.assert_not_called() |
| |
|
| | def test_fetchall_w_bqstorage_client_no_arrow_compression(self): |
| | pytest.importorskip("google.cloud.bigquery_storage") |
| | pytest.importorskip("pyarrow") |
| | from google.cloud import bigquery_storage |
| | from google.cloud.bigquery import dbapi |
| |
|
| | |
| | row_data = [bq_table.Row([1.2, 1.1], {"bar": 1, "foo": 0})] |
| | bqstorage_streamed_rows = [{"bar": _to_pyarrow(1.2), "foo": _to_pyarrow(1.1)}] |
| |
|
| | def fake_ensure_bqstorage_client(bqstorage_client=None, **kwargs): |
| | return bqstorage_client |
| |
|
| | mock_client = self._mock_client( |
| | rows=row_data, |
| | |
| | |
| | total_rows=1000, |
| | destination_table="P.DS.T", |
| | ) |
| | mock_client._ensure_bqstorage_client.side_effect = fake_ensure_bqstorage_client |
| | mock_bqstorage_client = self._mock_bqstorage_client( |
| | stream_count=1, |
| | rows=bqstorage_streamed_rows, |
| | ) |
| |
|
| | connection = dbapi.connect( |
| | client=mock_client, |
| | bqstorage_client=mock_bqstorage_client, |
| | ) |
| | cursor = connection.cursor() |
| | cursor.execute("SELECT foo, bar FROM some_table") |
| |
|
| | with mock.patch( |
| | "google.cloud.bigquery.dbapi.cursor._ARROW_COMPRESSION_SUPPORT", new=False |
| | ): |
| | rows = cursor.fetchall() |
| |
|
| | mock_client.list_rows.assert_not_called() |
| |
|
| | |
| | expected_session = bigquery_storage.ReadSession( |
| | table="projects/P/datasets/DS/tables/T", |
| | data_format=bigquery_storage.DataFormat.ARROW, |
| | ) |
| | mock_bqstorage_client.create_read_session.assert_called_once_with( |
| | parent="projects/P", read_session=expected_session, max_stream_count=1 |
| | ) |
| |
|
| | |
| | field_value = op.itemgetter(1) |
| | sorted_row_data = [sorted(row.items(), key=field_value) for row in rows] |
| | expected_row_data = [[("foo", 1.1), ("bar", 1.2)]] |
| |
|
| | self.assertEqual(sorted_row_data, expected_row_data) |
| |
|
| | def test_execute_custom_job_id(self): |
| | from google.cloud.bigquery.dbapi import connect |
| |
|
| | client = self._mock_client(rows=[], num_dml_affected_rows=0) |
| | connection = connect(client) |
| | cursor = connection.cursor() |
| | cursor.execute("SELECT 1;", job_id="foo") |
| | args, kwargs = client.query.call_args |
| | self.assertEqual(args[0], "SELECT 1;") |
| | self.assertEqual(kwargs["job_id"], "foo") |
| |
|
| | def test_execute_w_default_config(self): |
| | from google.cloud.bigquery.dbapi import connect |
| |
|
| | client = self._mock_client(rows=[], num_dml_affected_rows=0) |
| | connection = connect(client) |
| | cursor = connection.cursor() |
| |
|
| | cursor.execute("SELECT 1;", job_id="foo") |
| |
|
| | _, kwargs = client.query.call_args |
| | used_config = kwargs["job_config"] |
| | self.assertIsNone(used_config) |
| |
|
| | def test_execute_custom_job_config_wo_default_config(self): |
| | from google.cloud.bigquery.dbapi import connect |
| | from google.cloud.bigquery import job |
| |
|
| | config = job.QueryJobConfig(use_legacy_sql=True) |
| | client = self._mock_client(rows=[], num_dml_affected_rows=0) |
| | connection = connect(client) |
| | cursor = connection.cursor() |
| | cursor.execute("SELECT 1;", job_id="foo", job_config=config) |
| | args, kwargs = client.query.call_args |
| | self.assertEqual(args[0], "SELECT 1;") |
| | self.assertEqual(kwargs["job_id"], "foo") |
| | self.assertEqual(kwargs["job_config"], config) |
| |
|
| | def test_execute_custom_job_config_w_default_config(self): |
| | from google.cloud.bigquery.dbapi import connect |
| | from google.cloud.bigquery import job |
| |
|
| | client = self._mock_client(rows=[], num_dml_affected_rows=0) |
| | connection = connect(client) |
| | cursor = connection.cursor() |
| | config = job.QueryJobConfig(use_legacy_sql=True) |
| |
|
| | cursor.execute("SELECT 1;", job_id="foo", job_config=config) |
| |
|
| | _, kwargs = client.query.call_args |
| | used_config = kwargs["job_config"] |
| | expected_config = job.QueryJobConfig( |
| | use_legacy_sql=True, |
| | query_parameters=[], |
| | ) |
| | self.assertEqual(used_config._properties, expected_config._properties) |
| |
|
| | def test_execute_w_dml(self): |
| | from google.cloud.bigquery.dbapi import connect |
| |
|
| | connection = connect(self._mock_client(rows=[], num_dml_affected_rows=12)) |
| | cursor = connection.cursor() |
| | cursor.execute("DELETE FROM UserSessions WHERE user_id = 'test';") |
| | rows = cursor.fetchall() |
| | self.assertIsNone(cursor.description) |
| | self.assertEqual(cursor.rowcount, 12) |
| | self.assertEqual(rows, []) |
| |
|
| | def test_execute_w_query(self): |
| | from google.cloud.bigquery.schema import SchemaField |
| | from google.cloud.bigquery import dbapi |
| |
|
| | connection = dbapi.connect( |
| | self._mock_client( |
| | rows=[("hello", "world", 1), ("howdy", "y'all", 2)], |
| | schema=[ |
| | SchemaField("a", "STRING", mode="NULLABLE"), |
| | SchemaField("b", "STRING", mode="REQUIRED"), |
| | SchemaField("c", "INTEGER", mode="NULLABLE"), |
| | ], |
| | ) |
| | ) |
| | cursor = connection.cursor() |
| | cursor.execute("SELECT a, b, c FROM hello_world WHERE d > 3;") |
| |
|
| | |
| | self.assertEqual(len(cursor.description), 3) |
| | a_name, a_type, _, _, _, _, a_null_ok = cursor.description[0] |
| | self.assertEqual(a_name, "a") |
| | self.assertEqual(a_type, "STRING") |
| | self.assertEqual(a_type, dbapi.STRING) |
| | self.assertTrue(a_null_ok) |
| | b_name, b_type, _, _, _, _, b_null_ok = cursor.description[1] |
| | self.assertEqual(b_name, "b") |
| | self.assertEqual(b_type, "STRING") |
| | self.assertEqual(b_type, dbapi.STRING) |
| | self.assertFalse(b_null_ok) |
| | c_name, c_type, _, _, _, _, c_null_ok = cursor.description[2] |
| | self.assertEqual(c_name, "c") |
| | self.assertEqual(c_type, "INTEGER") |
| | self.assertEqual(c_type, dbapi.NUMBER) |
| | self.assertTrue(c_null_ok) |
| |
|
| | |
| | self.assertEqual(cursor.rowcount, 2) |
| | row = cursor.fetchone() |
| | self.assertEqual(row, ("hello", "world", 1)) |
| | row = cursor.fetchone() |
| | self.assertEqual(row, ("howdy", "y'all", 2)) |
| | row = cursor.fetchone() |
| | self.assertIsNone(row) |
| |
|
| | def test_execute_w_query_dry_run(self): |
| | from google.cloud.bigquery.job import QueryJobConfig |
| | from google.cloud.bigquery.schema import SchemaField |
| | from google.cloud.bigquery import dbapi |
| |
|
| | connection = dbapi.connect( |
| | self._mock_client( |
| | rows=[], |
| | schema=[ |
| | SchemaField("a", "STRING", mode="NULLABLE"), |
| | SchemaField("b", "STRING", mode="REQUIRED"), |
| | SchemaField("c", "INTEGER", mode="NULLABLE"), |
| | ], |
| | dry_run_job=True, |
| | total_bytes_processed=12345, |
| | ) |
| | ) |
| | cursor = connection.cursor() |
| |
|
| | cursor.execute( |
| | "SELECT a, b, c FROM hello_world WHERE d > 3;", |
| | job_config=QueryJobConfig(dry_run=True), |
| | ) |
| |
|
| | self.assertEqual(cursor.rowcount, 0) |
| | self.assertIsNotNone(cursor.description) |
| | rows = cursor.fetchall() |
| | self.assertEqual(list(rows), []) |
| |
|
| | def test_execute_raises_if_result_raises(self): |
| | import google.cloud.exceptions |
| |
|
| | from google.cloud.bigquery import client |
| | from google.cloud.bigquery.dbapi import connect |
| | from google.cloud.bigquery.dbapi import exceptions |
| |
|
| | client = mock.create_autospec(client.Client) |
| | client.query_and_wait.side_effect = google.cloud.exceptions.GoogleCloudError("") |
| | connection = connect(client) |
| | cursor = connection.cursor() |
| |
|
| | with self.assertRaises(exceptions.DatabaseError): |
| | cursor.execute("SELECT 1") |
| |
|
| | def test_executemany_w_dml(self): |
| | from google.cloud.bigquery.dbapi import connect |
| |
|
| | connection = connect(self._mock_client(rows=[], num_dml_affected_rows=12)) |
| | cursor = connection.cursor() |
| | cursor.executemany( |
| | "DELETE FROM UserSessions WHERE user_id = %s;", |
| | (("test",), ("anothertest",)), |
| | ) |
| | self.assertIsNone(cursor.description) |
| | self.assertEqual(cursor.rowcount, 24) |
| |
|
| | def test_executemany_empty(self): |
| | from google.cloud.bigquery.dbapi import connect |
| |
|
| | connection = connect(self._mock_client(rows=[], num_dml_affected_rows=12)) |
| | cursor = connection.cursor() |
| | cursor.executemany((), ()) |
| | self.assertIsNone(cursor.description) |
| | self.assertEqual(cursor.rowcount, -1) |
| |
|
| | def test_is_iterable(self): |
| | from google.cloud.bigquery import dbapi |
| |
|
| | connection = dbapi.connect( |
| | self._mock_client(rows=[("hello", "there", 7), ("good", "bye", -3)]) |
| | ) |
| | cursor = connection.cursor() |
| | cursor.execute("SELECT foo, bar, baz FROM hello_world WHERE baz < 42;") |
| |
|
| | rows_iter = iter(cursor) |
| |
|
| | row = next(rows_iter) |
| | self.assertEqual(row, ("hello", "there", 7)) |
| | row = next(rows_iter) |
| | self.assertEqual(row, ("good", "bye", -3)) |
| | self.assertRaises(StopIteration, next, rows_iter) |
| |
|
| | self.assertEqual( |
| | list(cursor), |
| | [], |
| | "Iterating again over the same results should produce no rows.", |
| | ) |
| |
|
| | def test_query_job_wo_execute(self): |
| | from google.cloud.bigquery import dbapi |
| |
|
| | connection = dbapi.connect(self._mock_client()) |
| | cursor = connection.cursor() |
| | self.assertIsNone(cursor.query_job) |
| |
|
| | def test_query_job_w_execute(self): |
| | from google.cloud.bigquery import dbapi, QueryJob |
| |
|
| | connection = dbapi.connect(self._mock_client()) |
| | cursor = connection.cursor() |
| | cursor.execute("SELECT 1;") |
| | self.assertIsInstance(cursor.query_job, QueryJob) |
| |
|
| | def test_query_job_w_execute_no_job(self): |
| | from google.cloud.bigquery import dbapi |
| |
|
| | connection = dbapi.connect(self._mock_client()) |
| | cursor = connection.cursor() |
| | cursor.execute("SELECT 1;") |
| |
|
| | |
| | type(cursor._query_rows).job_id = mock.PropertyMock(return_value=None) |
| |
|
| | self.assertIsNone(cursor.query_job) |
| |
|
| | def test_query_job_w_executemany(self): |
| | from google.cloud.bigquery import dbapi, QueryJob |
| |
|
| | connection = dbapi.connect(self._mock_client()) |
| | cursor = connection.cursor() |
| | cursor.executemany("SELECT %s;", (("1",), ("2",))) |
| | self.assertIsInstance(cursor.query_job, QueryJob) |
| |
|
| | def test__format_operation_w_dict(self): |
| | from google.cloud.bigquery.dbapi import cursor |
| |
|
| | parameter_types = {} |
| | formatted_operation, parameter_types = cursor._format_operation( |
| | "SELECT %(somevalue)s, %(a `weird` one:STRING)s;", |
| | {"somevalue": "hi", "a `weird` one": "world"}, |
| | ) |
| | self.assertEqual( |
| | formatted_operation, "SELECT @`somevalue`, @`a \\`weird\\` one`;" |
| | ) |
| | self.assertEqual(parameter_types, {"a `weird` one": "STRING"}) |
| |
|
| | def test__format_operation_w_wrong_dict(self): |
| | from google.cloud.bigquery import dbapi |
| | from google.cloud.bigquery.dbapi import cursor |
| |
|
| | self.assertRaises( |
| | dbapi.ProgrammingError, |
| | cursor._format_operation, |
| | "SELECT %(somevalue)s, %(othervalue)s;", |
| | {"somevalue-not-here": "hi", "othervalue": "world"}, |
| | ) |
| |
|
| | def test__format_operation_w_redundant_dict_key(self): |
| | from google.cloud.bigquery.dbapi import cursor |
| |
|
| | formatted_operation, _ = cursor._format_operation( |
| | "SELECT %(somevalue)s;", {"somevalue": "foo", "value-not-used": "bar"} |
| | ) |
| | self.assertEqual(formatted_operation, "SELECT @`somevalue`;") |
| |
|
| | def test__format_operation_w_sequence(self): |
| | from google.cloud.bigquery.dbapi import cursor |
| |
|
| | formatted_operation, _ = cursor._format_operation( |
| | "SELECT %s, %s;", ("hello", "world") |
| | ) |
| | self.assertEqual(formatted_operation, "SELECT ?, ?;") |
| |
|
| | def test__format_operation_w_too_short_sequence(self): |
| | from google.cloud.bigquery import dbapi |
| | from google.cloud.bigquery.dbapi import cursor |
| |
|
| | self.assertRaises( |
| | dbapi.ProgrammingError, |
| | cursor._format_operation, |
| | "SELECT %s, %s;", |
| | ("hello",), |
| | ) |
| |
|
| | def test__format_operation_w_too_long_sequence(self): |
| | from google.cloud.bigquery import dbapi |
| | from google.cloud.bigquery.dbapi import cursor |
| |
|
| | self.assertRaises( |
| | dbapi.ProgrammingError, |
| | cursor._format_operation, |
| | "SELECT %s, %s;", |
| | ("hello", "world", "everyone"), |
| | ) |
| |
|
| | def test__format_operation_w_empty_dict(self): |
| | from google.cloud.bigquery.dbapi import cursor |
| |
|
| | formatted_operation, _ = cursor._format_operation("SELECT '%f'", {}) |
| | self.assertEqual(formatted_operation, "SELECT '%f'") |
| |
|
| | def test__format_operation_wo_params_single_percent(self): |
| | from google.cloud.bigquery.dbapi import cursor |
| |
|
| | formatted_operation, _ = cursor._format_operation("SELECT '%'", {}) |
| | self.assertEqual(formatted_operation, "SELECT '%'") |
| |
|
| | def test__format_operation_wo_params_double_percents(self): |
| | from google.cloud.bigquery.dbapi import cursor |
| |
|
| | formatted_operation, _ = cursor._format_operation("SELECT '%%'", {}) |
| | self.assertEqual(formatted_operation, "SELECT '%'") |
| |
|
| | def test__format_operation_unescaped_percent_w_dict_param(self): |
| | from google.cloud.bigquery import dbapi |
| | from google.cloud.bigquery.dbapi import cursor |
| |
|
| | self.assertRaises( |
| | dbapi.ProgrammingError, |
| | cursor._format_operation, |
| | "SELECT %(foo)s, '100 %';", |
| | {"foo": "bar"}, |
| | ) |
| |
|
| | def test__format_operation_unescaped_percent_w_list_param(self): |
| | from google.cloud.bigquery import dbapi |
| | from google.cloud.bigquery.dbapi import cursor |
| |
|
| | self.assertRaises( |
| | dbapi.ProgrammingError, |
| | cursor._format_operation, |
| | "SELECT %s, %s, '100 %';", |
| | ["foo", "bar"], |
| | ) |
| |
|
| | def test__format_operation_no_placeholders(self): |
| | from google.cloud.bigquery import dbapi |
| | from google.cloud.bigquery.dbapi import cursor |
| |
|
| | self.assertRaises( |
| | dbapi.ProgrammingError, |
| | cursor._format_operation, |
| | "SELECT 42", |
| | ["foo", "bar"], |
| | ) |
| |
|
| |
|
| | @pytest.mark.parametrize( |
| | "inp,expect", |
| | [ |
| | ("", ("", None)), |
| | ("values(%(foo)s, %(bar)s)", ("values(%(foo)s, %(bar)s)", {})), |
| | ( |
| | "values('%%(oof:INT64)s', %(foo)s, %(bar)s)", |
| | ("values('%%(oof:INT64)s', %(foo)s, %(bar)s)", {}), |
| | ), |
| | ( |
| | "values(%(foo:INT64)s, %(bar)s)", |
| | ("values(%(foo)s, %(bar)s)", dict(foo="INT64")), |
| | ), |
| | ( |
| | "values('%%(oof:INT64)s, %(foo:INT64)s, %(foo)s)", |
| | ("values('%%(oof:INT64)s, %(foo)s, %(foo)s)", dict(foo="INT64")), |
| | ), |
| | ( |
| | "values(%(foo:INT64)s, %(foo:INT64)s)", |
| | ("values(%(foo)s, %(foo)s)", dict(foo="INT64")), |
| | ), |
| | ( |
| | "values(%(foo:INT64)s, %(bar:NUMERIC)s) 100 %", |
| | ("values(%(foo)s, %(bar)s) 100 %", dict(foo="INT64", bar="NUMERIC")), |
| | ), |
| | (" %s %()s %(:int64)s ", (" %s %s %s ", [None, None, "int64"])), |
| | (" %%s %s %()s %(:int64)s ", (" %%s %s %s %s ", [None, None, "int64"])), |
| | ( |
| | "values(%%%(foo:INT64)s, %(bar)s)", |
| | ("values(%%%(foo)s, %(bar)s)", dict(foo="INT64")), |
| | ), |
| | ( |
| | "values(%%%%(foo:INT64)s, %(bar)s)", |
| | ("values(%%%%(foo:INT64)s, %(bar)s)", dict()), |
| | ), |
| | ( |
| | "values(%%%%%(foo:INT64)s, %(bar)s)", |
| | ("values(%%%%%(foo)s, %(bar)s)", dict(foo="INT64")), |
| | ), |
| | ( |
| | "values(%%%%%(foo:struct<x_1 string, y_ int64>)s, %(bar)s)", |
| | ("values(%%%%%(foo)s, %(bar)s)", dict(foo="struct<x_1 string, y_ int64>")), |
| | ), |
| | ( |
| | "values(%%%%%(foo:struct<x string, y int64>)s, %(bar)s)", |
| | ("values(%%%%%(foo)s, %(bar)s)", dict(foo="struct<x string, y int64>")), |
| | ), |
| | ( |
| | "values(%%%%%(foo:struct<x string(10), y int64>)s, %(bar)s)", |
| | ("values(%%%%%(foo)s, %(bar)s)", dict(foo="struct<x string(10), y int64>")), |
| | ), |
| | ( |
| | "values(%(foo:struct<x string(10), y numeric(2, 9)>)s, %(bar)s)", |
| | ( |
| | "values(%(foo)s, %(bar)s)", |
| | dict(foo="struct<x string(10), y numeric(2, 9)>"), |
| | ), |
| | ), |
| | ( |
| | "values(%(foo:struct<x string(10) , y numeric(2, 9)>)s, %(bar)s)", |
| | ( |
| | "values(%(foo)s, %(bar)s)", |
| | dict(foo="struct<x string(10) , y numeric(2, 9)>"), |
| | ), |
| | ), |
| | ( |
| | "values(%(foo:string(10))s, %(bar)s)", |
| | ("values(%(foo)s, %(bar)s)", dict(foo="string(10)")), |
| | ), |
| | ], |
| | ) |
| | def test__extract_types(inp, expect): |
| | from google.cloud.bigquery.dbapi.cursor import _extract_types as et |
| |
|
| | assert et(inp) == expect |
| |
|
| |
|
| | @pytest.mark.parametrize( |
| | "match,inp", |
| | [ |
| | ( |
| | "Conflicting types for foo: numeric and int64.", |
| | " %(foo:numeric)s %(foo:int64)s ", |
| | ), |
| | (r"' %s %\(foo\)s ' mixes named and unamed parameters.", " %s %(foo)s "), |
| | (r"' %\(foo\)s %s ' mixes named and unamed parameters.", " %(foo)s %s "), |
| | ], |
| | ) |
| | def test__extract_types_fail(match, inp): |
| | from google.cloud.bigquery.dbapi.cursor import _extract_types as et |
| | from google.cloud.bigquery.dbapi import exceptions |
| |
|
| | with pytest.raises(exceptions.ProgrammingError, match=match): |
| | et(inp) |
| |
|