| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from unittest import mock |
|
|
| from ..helpers import make_connection |
|
|
| from .helpers import _Base |
| from .helpers import _make_client |
|
|
| import datetime |
|
|
|
|
| class TestCopyJobConfig(_Base): |
| JOB_TYPE = "copy" |
|
|
| @staticmethod |
| def _get_target_class(): |
| from google.cloud.bigquery.job import CopyJobConfig |
|
|
| return CopyJobConfig |
|
|
| def test_ctor_defaults(self): |
| from google.cloud.bigquery.job import OperationType |
|
|
| config = self._make_one() |
|
|
| assert config.create_disposition is None |
| assert config.write_disposition is None |
| assert config.destination_expiration_time is None |
| assert config.destination_encryption_configuration is None |
| assert config.operation_type == OperationType.OPERATION_TYPE_UNSPECIFIED |
|
|
| def test_ctor_w_properties(self): |
| from google.cloud.bigquery.job import CreateDisposition |
| from google.cloud.bigquery.job import OperationType |
| from google.cloud.bigquery.job import WriteDisposition |
|
|
| create_disposition = CreateDisposition.CREATE_NEVER |
| write_disposition = WriteDisposition.WRITE_TRUNCATE |
| snapshot_operation = OperationType.SNAPSHOT |
|
|
| today = datetime.date.today() |
| destination_expiration_time = f"{today.year + 1}-01-01T00:00:00Z" |
|
|
| config = self._get_target_class()( |
| create_disposition=create_disposition, |
| write_disposition=write_disposition, |
| operation_type=snapshot_operation, |
| destination_expiration_time=destination_expiration_time, |
| ) |
|
|
| self.assertEqual(config.create_disposition, create_disposition) |
| self.assertEqual(config.write_disposition, write_disposition) |
| self.assertEqual(config.operation_type, snapshot_operation) |
| self.assertEqual( |
| config.destination_expiration_time, destination_expiration_time |
| ) |
|
|
| def test_to_api_repr_with_encryption(self): |
| from google.cloud.bigquery.encryption_configuration import ( |
| EncryptionConfiguration, |
| ) |
|
|
| config = self._make_one() |
| config.destination_encryption_configuration = EncryptionConfiguration( |
| kms_key_name=self.KMS_KEY_NAME |
| ) |
| resource = config.to_api_repr() |
| self.assertEqual( |
| resource, |
| { |
| "copy": { |
| "destinationEncryptionConfiguration": { |
| "kmsKeyName": self.KMS_KEY_NAME |
| } |
| } |
| }, |
| ) |
|
|
| def test_to_api_repr_with_encryption_none(self): |
| config = self._make_one() |
| config.destination_encryption_configuration = None |
| resource = config.to_api_repr() |
| self.assertEqual( |
| resource, {"copy": {"destinationEncryptionConfiguration": None}} |
| ) |
|
|
| def test_operation_type_setting_none(self): |
| from google.cloud.bigquery.job import OperationType |
|
|
| config = self._make_one(operation_type=OperationType.SNAPSHOT) |
|
|
| |
| config.operation_type = None |
| assert config.operation_type == OperationType.OPERATION_TYPE_UNSPECIFIED |
|
|
| def test_operation_type_setting_non_none(self): |
| from google.cloud.bigquery.job import OperationType |
|
|
| config = self._make_one(operation_type=None) |
| config.operation_type = OperationType.RESTORE |
| assert config.operation_type == OperationType.RESTORE |
|
|
|
|
| class TestCopyJob(_Base): |
| JOB_TYPE = "copy" |
| SOURCE_TABLE = "source_table" |
| DESTINATION_TABLE = "destination_table" |
|
|
| @staticmethod |
| def _get_target_class(): |
| from google.cloud.bigquery.job import CopyJob |
|
|
| return CopyJob |
|
|
| def _make_resource(self, started=False, ended=False): |
| resource = super(TestCopyJob, self)._make_resource(started, ended) |
| config = resource["configuration"]["copy"] |
| config["sourceTables"] = [ |
| { |
| "projectId": self.PROJECT, |
| "datasetId": self.DS_ID, |
| "tableId": self.SOURCE_TABLE, |
| } |
| ] |
| config["destinationTable"] = { |
| "projectId": self.PROJECT, |
| "datasetId": self.DS_ID, |
| "tableId": self.DESTINATION_TABLE, |
| } |
|
|
| return resource |
|
|
| def _verifyResourceProperties(self, job, resource): |
| self._verifyReadonlyResourceProperties(job, resource) |
|
|
| config = resource.get("configuration", {}).get("copy") |
|
|
| table_ref = config["destinationTable"] |
| self.assertEqual(job.destination.project, table_ref["projectId"]) |
| self.assertEqual(job.destination.dataset_id, table_ref["datasetId"]) |
| self.assertEqual(job.destination.table_id, table_ref["tableId"]) |
|
|
| sources = config.get("sourceTables") |
| if sources is None: |
| sources = [config["sourceTable"]] |
| self.assertEqual(len(sources), len(job.sources)) |
| for table_ref, table in zip(sources, job.sources): |
| self.assertEqual(table.project, table_ref["projectId"]) |
| self.assertEqual(table.dataset_id, table_ref["datasetId"]) |
| self.assertEqual(table.table_id, table_ref["tableId"]) |
|
|
| if "createDisposition" in config: |
| self.assertEqual(job.create_disposition, config["createDisposition"]) |
| else: |
| self.assertIsNone(job.create_disposition) |
|
|
| if "writeDisposition" in config: |
| self.assertEqual(job.write_disposition, config["writeDisposition"]) |
| else: |
| self.assertIsNone(job.write_disposition) |
|
|
| if "destinationEncryptionConfiguration" in config: |
| self.assertIsNotNone(job.destination_encryption_configuration) |
| self.assertEqual( |
| job.destination_encryption_configuration.kms_key_name, |
| config["destinationEncryptionConfiguration"]["kmsKeyName"], |
| ) |
| else: |
| self.assertIsNone(job.destination_encryption_configuration) |
|
|
| def test_ctor(self): |
| client = _make_client(project=self.PROJECT) |
| source = self._table_ref(self.SOURCE_TABLE) |
| destination = self._table_ref(self.DESTINATION_TABLE) |
| job = self._make_one(self.JOB_ID, [source], destination, client) |
| self.assertEqual(job.destination, destination) |
| self.assertEqual(job.sources, [source]) |
| self.assertIs(job._client, client) |
| self.assertEqual(job.job_type, self.JOB_TYPE) |
| self.assertEqual(job.path, "/projects/%s/jobs/%s" % (self.PROJECT, self.JOB_ID)) |
|
|
| self._verifyInitialReadonlyProperties(job) |
|
|
| |
| self.assertIsNone(job.create_disposition) |
| self.assertIsNone(job.write_disposition) |
| self.assertIsNone(job.destination_encryption_configuration) |
|
|
| def test_from_api_repr_missing_identity(self): |
| self._setUpConstants() |
| client = _make_client(project=self.PROJECT) |
| RESOURCE = {} |
| klass = self._get_target_class() |
| with self.assertRaises(KeyError): |
| klass.from_api_repr(RESOURCE, client=client) |
|
|
| def test_from_api_repr_missing_config(self): |
| self._setUpConstants() |
| client = _make_client(project=self.PROJECT) |
| RESOURCE = { |
| "id": "%s:%s" % (self.PROJECT, self.DS_ID), |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| } |
| klass = self._get_target_class() |
| with self.assertRaises(KeyError): |
| klass.from_api_repr(RESOURCE, client=client) |
|
|
| def test_from_api_repr_bare(self): |
| self._setUpConstants() |
| client = _make_client(project=self.PROJECT) |
| RESOURCE = { |
| "id": self.JOB_ID, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "configuration": { |
| "copy": { |
| "sourceTables": [ |
| { |
| "projectId": self.PROJECT, |
| "datasetId": self.DS_ID, |
| "tableId": self.SOURCE_TABLE, |
| } |
| ], |
| "destinationTable": { |
| "projectId": self.PROJECT, |
| "datasetId": self.DS_ID, |
| "tableId": self.DESTINATION_TABLE, |
| }, |
| } |
| }, |
| } |
| klass = self._get_target_class() |
| job = klass.from_api_repr(RESOURCE, client=client) |
| self.assertIs(job._client, client) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_from_api_with_encryption(self): |
| self._setUpConstants() |
| client = _make_client(project=self.PROJECT) |
| RESOURCE = { |
| "id": self.JOB_ID, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "configuration": { |
| "copy": { |
| "sourceTables": [ |
| { |
| "projectId": self.PROJECT, |
| "datasetId": self.DS_ID, |
| "tableId": self.SOURCE_TABLE, |
| } |
| ], |
| "destinationTable": { |
| "projectId": self.PROJECT, |
| "datasetId": self.DS_ID, |
| "tableId": self.DESTINATION_TABLE, |
| }, |
| "destinationEncryptionConfiguration": { |
| "kmsKeyName": self.KMS_KEY_NAME |
| }, |
| } |
| }, |
| } |
| klass = self._get_target_class() |
| job = klass.from_api_repr(RESOURCE, client=client) |
| self.assertIs(job._client, client) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_from_api_repr_w_sourcetable(self): |
| self._setUpConstants() |
| client = _make_client(project=self.PROJECT) |
| RESOURCE = { |
| "id": self.JOB_ID, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "configuration": { |
| "copy": { |
| "sourceTable": { |
| "projectId": self.PROJECT, |
| "datasetId": self.DS_ID, |
| "tableId": self.SOURCE_TABLE, |
| }, |
| "destinationTable": { |
| "projectId": self.PROJECT, |
| "datasetId": self.DS_ID, |
| "tableId": self.DESTINATION_TABLE, |
| }, |
| } |
| }, |
| } |
| klass = self._get_target_class() |
| job = klass.from_api_repr(RESOURCE, client=client) |
| self.assertIs(job._client, client) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_from_api_repr_wo_sources(self): |
| self._setUpConstants() |
| client = _make_client(project=self.PROJECT) |
| RESOURCE = { |
| "id": self.JOB_ID, |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "configuration": { |
| "copy": { |
| "destinationTable": { |
| "projectId": self.PROJECT, |
| "datasetId": self.DS_ID, |
| "tableId": self.DESTINATION_TABLE, |
| } |
| } |
| }, |
| } |
| klass = self._get_target_class() |
| job = klass.from_api_repr(RESOURCE, client=client) |
| with self.assertRaises(KeyError): |
| _ = job.sources |
|
|
| def test_from_api_repr_w_properties(self): |
| from google.cloud.bigquery.job import CreateDisposition |
|
|
| client = _make_client(project=self.PROJECT) |
| RESOURCE = self._make_resource() |
| copy_config = RESOURCE["configuration"]["copy"] |
| copy_config["createDisposition"] = CreateDisposition.CREATE_IF_NEEDED |
| klass = self._get_target_class() |
| job = klass.from_api_repr(RESOURCE, client=client) |
| self.assertIs(job._client, client) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_begin_w_bound_client(self): |
| PATH = "/projects/%s/jobs" % (self.PROJECT,) |
| RESOURCE = self._make_resource() |
| |
| del RESOURCE["statistics"]["creationTime"] |
| del RESOURCE["etag"] |
| del RESOURCE["selfLink"] |
| del RESOURCE["user_email"] |
| conn = make_connection(RESOURCE) |
| client = _make_client(project=self.PROJECT, connection=conn) |
| source = self._table_ref(self.SOURCE_TABLE) |
| destination = self._table_ref(self.DESTINATION_TABLE) |
| job = self._make_one(self.JOB_ID, [source], destination, client) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| job._begin() |
|
|
| final_attributes.assert_called_with({"path": PATH}, client, job) |
|
|
| conn.api_request.assert_called_once_with( |
| method="POST", |
| path=PATH, |
| data={ |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "configuration": { |
| "copy": { |
| "sourceTables": [ |
| { |
| "projectId": self.PROJECT, |
| "datasetId": self.DS_ID, |
| "tableId": self.SOURCE_TABLE, |
| } |
| ], |
| "destinationTable": { |
| "projectId": self.PROJECT, |
| "datasetId": self.DS_ID, |
| "tableId": self.DESTINATION_TABLE, |
| }, |
| } |
| }, |
| }, |
| timeout=None, |
| ) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_begin_w_alternate_client(self): |
| from google.cloud.bigquery.job import CopyJobConfig |
|
|
| from google.cloud.bigquery.job import CreateDisposition |
| from google.cloud.bigquery.job import WriteDisposition |
|
|
| PATH = "/projects/%s/jobs" % (self.PROJECT,) |
| RESOURCE = self._make_resource(ended=True) |
| COPY_CONFIGURATION = { |
| "sourceTables": [ |
| { |
| "projectId": self.PROJECT, |
| "datasetId": self.DS_ID, |
| "tableId": self.SOURCE_TABLE, |
| } |
| ], |
| "destinationTable": { |
| "projectId": self.PROJECT, |
| "datasetId": self.DS_ID, |
| "tableId": self.DESTINATION_TABLE, |
| }, |
| "createDisposition": CreateDisposition.CREATE_NEVER, |
| "writeDisposition": WriteDisposition.WRITE_TRUNCATE, |
| } |
| RESOURCE["configuration"]["copy"] = COPY_CONFIGURATION |
| conn1 = make_connection() |
| client1 = _make_client(project=self.PROJECT, connection=conn1) |
| conn2 = make_connection(RESOURCE) |
| client2 = _make_client(project=self.PROJECT, connection=conn2) |
| source = self._table_ref(self.SOURCE_TABLE) |
| destination = self._table_ref(self.DESTINATION_TABLE) |
| config = CopyJobConfig() |
| config.create_disposition = CreateDisposition.CREATE_NEVER |
| config.write_disposition = WriteDisposition.WRITE_TRUNCATE |
| job = self._make_one(self.JOB_ID, [source], destination, client1, config) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| job._begin(client=client2) |
|
|
| final_attributes.assert_called_with({"path": PATH}, client2, job) |
|
|
| conn1.api_request.assert_not_called() |
| conn2.api_request.assert_called_once_with( |
| method="POST", |
| path=PATH, |
| data={ |
| "jobReference": {"projectId": self.PROJECT, "jobId": self.JOB_ID}, |
| "configuration": {"copy": COPY_CONFIGURATION}, |
| }, |
| timeout=None, |
| ) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_exists_miss_w_bound_client(self): |
| PATH = "/projects/%s/jobs/%s" % (self.PROJECT, self.JOB_ID) |
| conn = make_connection() |
| client = _make_client(project=self.PROJECT, connection=conn) |
|
|
| source = self._table_ref(self.SOURCE_TABLE) |
| destination = self._table_ref(self.DESTINATION_TABLE) |
| job = self._make_one(self.JOB_ID, [source], destination, client) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| self.assertFalse(job.exists()) |
|
|
| final_attributes.assert_called_with({"path": PATH}, client, job) |
|
|
| conn.api_request.assert_called_once_with( |
| method="GET", path=PATH, query_params={"fields": "id"}, timeout=None |
| ) |
|
|
| def test_exists_hit_w_alternate_client(self): |
| PATH = "/projects/%s/jobs/%s" % (self.PROJECT, self.JOB_ID) |
| conn1 = make_connection() |
| client1 = _make_client(project=self.PROJECT, connection=conn1) |
| conn2 = make_connection({}) |
| client2 = _make_client(project=self.PROJECT, connection=conn2) |
| source = self._table_ref(self.SOURCE_TABLE) |
| destination = self._table_ref(self.DESTINATION_TABLE) |
| job = self._make_one(self.JOB_ID, [source], destination, client1) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| self.assertTrue(job.exists(client=client2)) |
|
|
| final_attributes.assert_called_with({"path": PATH}, client2, job) |
|
|
| conn1.api_request.assert_not_called() |
| conn2.api_request.assert_called_once_with( |
| method="GET", path=PATH, query_params={"fields": "id"}, timeout=None |
| ) |
|
|
| def test_reload_w_bound_client(self): |
| from google.cloud.bigquery.retry import DEFAULT_GET_JOB_TIMEOUT |
|
|
| PATH = "/projects/%s/jobs/%s" % (self.PROJECT, self.JOB_ID) |
| RESOURCE = self._make_resource() |
| conn = make_connection(RESOURCE) |
| client = _make_client(project=self.PROJECT, connection=conn) |
| source = self._table_ref(self.SOURCE_TABLE) |
| destination = self._table_ref(self.DESTINATION_TABLE) |
| job = self._make_one(self.JOB_ID, [source], destination, client) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| job.reload() |
|
|
| final_attributes.assert_called_with( |
| { |
| "path": PATH, |
| "job_id": self.JOB_ID, |
| "location": None, |
| }, |
| client, |
| None, |
| ) |
|
|
| conn.api_request.assert_called_once_with( |
| method="GET", |
| path=PATH, |
| query_params={"projection": "full"}, |
| timeout=DEFAULT_GET_JOB_TIMEOUT, |
| ) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|
| def test_reload_w_alternate_client(self): |
| from google.cloud.bigquery.retry import DEFAULT_GET_JOB_TIMEOUT |
|
|
| PATH = "/projects/%s/jobs/%s" % (self.PROJECT, self.JOB_ID) |
| RESOURCE = self._make_resource() |
| conn1 = make_connection() |
| client1 = _make_client(project=self.PROJECT, connection=conn1) |
| conn2 = make_connection(RESOURCE) |
| client2 = _make_client(project=self.PROJECT, connection=conn2) |
| source = self._table_ref(self.SOURCE_TABLE) |
| destination = self._table_ref(self.DESTINATION_TABLE) |
| job = self._make_one(self.JOB_ID, [source], destination, client1) |
| with mock.patch( |
| "google.cloud.bigquery.opentelemetry_tracing._get_final_span_attributes" |
| ) as final_attributes: |
| job.reload(client=client2) |
|
|
| final_attributes.assert_called_with( |
| { |
| "path": PATH, |
| "job_id": self.JOB_ID, |
| "location": None, |
| }, |
| client2, |
| None, |
| ) |
|
|
| conn1.api_request.assert_not_called() |
| conn2.api_request.assert_called_once_with( |
| method="GET", |
| path=PATH, |
| query_params={"projection": "full"}, |
| timeout=DEFAULT_GET_JOB_TIMEOUT, |
| ) |
| self._verifyResourceProperties(job, RESOURCE) |
|
|