| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | import os |
| | from uuid import uuid4 |
| |
|
| | from google.api_core import exceptions |
| |
|
| | from google.cloud import aiplatform, aiplatform_v1beta1 |
| | from google.cloud import bigquery |
| | from google.cloud import storage |
| | import pytest |
| |
|
| | import helpers |
| |
|
| |
|
| | @pytest.fixture() |
| | def shared_state(): |
| | state = {} |
| | yield state |
| |
|
| |
|
| | @pytest.fixture |
| | def storage_client(): |
| | storage_client = storage.Client() |
| | return storage_client |
| |
|
| |
|
| | @pytest.fixture() |
| | def job_client(): |
| | job_client = aiplatform.gapic.JobServiceClient( |
| | client_options={"api_endpoint": "us-central1-aiplatform.googleapis.com"} |
| | ) |
| | return job_client |
| |
|
| |
|
| | @pytest.fixture() |
| | def data_labeling_job_client(): |
| | data_labeling_api_endpoint = os.getenv("DATA_LABELING_API_ENDPOINT") |
| | data_labeling_job_client = aiplatform.gapic.JobServiceClient( |
| | client_options={"api_endpoint": data_labeling_api_endpoint} |
| | ) |
| | return data_labeling_job_client |
| |
|
| |
|
| | @pytest.fixture |
| | def pipeline_client(): |
| | pipeline_client = aiplatform.gapic.PipelineServiceClient( |
| | client_options={"api_endpoint": "us-central1-aiplatform.googleapis.com"} |
| | ) |
| | return pipeline_client |
| |
|
| |
|
| | @pytest.fixture |
| | def model_client(): |
| | model_client = aiplatform.gapic.ModelServiceClient( |
| | client_options={"api_endpoint": "us-central1-aiplatform.googleapis.com"} |
| | ) |
| | yield model_client |
| |
|
| |
|
| | @pytest.fixture |
| | def endpoint_client(): |
| | endpoint_client = aiplatform.gapic.EndpointServiceClient( |
| | client_options={"api_endpoint": "us-central1-aiplatform.googleapis.com"} |
| | ) |
| | yield endpoint_client |
| |
|
| |
|
| | @pytest.fixture |
| | def dataset_client(): |
| | dataset_client = aiplatform.gapic.DatasetServiceClient( |
| | client_options={"api_endpoint": "us-central1-aiplatform.googleapis.com"} |
| | ) |
| | yield dataset_client |
| |
|
| |
|
| | @pytest.fixture |
| | def featurestore_client(): |
| | featurestore_client = aiplatform_v1beta1.FeaturestoreServiceClient( |
| | client_options={"api_endpoint": "us-central1-aiplatform.googleapis.com"} |
| | ) |
| | yield featurestore_client |
| |
|
| |
|
| | @pytest.fixture |
| | def bigquery_client(): |
| | bigquery_client = bigquery.Client( |
| | project=os.getenv("BUILD_SPECIFIC_GCLOUD_PROJECT") |
| | ) |
| | yield bigquery_client |
| |
|
| |
|
| | |
| | @pytest.fixture() |
| | def teardown_batch_prediction_job(shared_state, job_client): |
| | yield |
| |
|
| | job_client.cancel_batch_prediction_job( |
| | name=shared_state["batch_prediction_job_name"] |
| | ) |
| |
|
| | |
| | helpers.wait_for_job_state( |
| | get_job_method=job_client.get_batch_prediction_job, |
| | name=shared_state["batch_prediction_job_name"], |
| | ) |
| |
|
| | job_client.delete_batch_prediction_job( |
| | name=shared_state["batch_prediction_job_name"] |
| | ) |
| |
|
| |
|
| | @pytest.fixture() |
| | def teardown_data_labeling_job(capsys, shared_state, data_labeling_job_client): |
| | yield |
| |
|
| | assert "/" in shared_state["data_labeling_job_name"] |
| |
|
| | data_labeling_job_client.cancel_data_labeling_job( |
| | name=shared_state["data_labeling_job_name"] |
| | ) |
| |
|
| | |
| | helpers.wait_for_job_state( |
| | get_job_method=data_labeling_job_client.get_data_labeling_job, |
| | name=shared_state["data_labeling_job_name"], |
| | timeout=400, |
| | freq=10, |
| | ) |
| |
|
| | |
| | response = data_labeling_job_client.delete_data_labeling_job( |
| | name=shared_state["data_labeling_job_name"] |
| | ) |
| | print("Delete LRO:", response.operation.name) |
| | delete_data_labeling_job_response = response.result(timeout=300) |
| | print("delete_data_labeling_job_response", delete_data_labeling_job_response) |
| |
|
| | out, _ = capsys.readouterr() |
| | assert "delete_data_labeling_job_response" in out |
| |
|
| |
|
| | @pytest.fixture() |
| | def teardown_hyperparameter_tuning_job(shared_state, job_client): |
| | yield |
| |
|
| | |
| | job_client.cancel_hyperparameter_tuning_job( |
| | name=shared_state["hyperparameter_tuning_job_name"] |
| | ) |
| |
|
| | |
| | helpers.wait_for_job_state( |
| | get_job_method=job_client.get_hyperparameter_tuning_job, |
| | name=shared_state["hyperparameter_tuning_job_name"], |
| | ) |
| |
|
| | |
| | job_client.delete_hyperparameter_tuning_job( |
| | name=shared_state["hyperparameter_tuning_job_name"] |
| | ) |
| |
|
| |
|
| | @pytest.fixture() |
| | def teardown_training_pipeline(shared_state, pipeline_client): |
| | yield |
| |
|
| | try: |
| | pipeline_client.cancel_training_pipeline( |
| | name=shared_state["training_pipeline_name"] |
| | ) |
| |
|
| | |
| | timeout = shared_state["cancel_batch_prediction_job_timeout"] |
| | helpers.wait_for_job_state( |
| | get_job_method=pipeline_client.get_training_pipeline, |
| | name=shared_state["training_pipeline_name"], |
| | timeout=timeout, |
| | ) |
| |
|
| | except exceptions.FailedPrecondition: |
| | pass |
| |
|
| | finally: |
| | |
| | pipeline_client.delete_training_pipeline( |
| | name=shared_state["training_pipeline_name"] |
| | ) |
| |
|
| |
|
| | @pytest.fixture() |
| | def create_dataset(shared_state, dataset_client): |
| | def create( |
| | project, location, metadata_schema_uri, test_name="temp_import_dataset_test" |
| | ): |
| | parent = f"projects/{project}/locations/{location}" |
| | dataset = aiplatform.gapic.Dataset( |
| | display_name=f"{test_name}_{uuid4()}", |
| | metadata_schema_uri=metadata_schema_uri, |
| | ) |
| |
|
| | operation = dataset_client.create_dataset(parent=parent, dataset=dataset) |
| |
|
| | dataset = operation.result(timeout=300) |
| | shared_state["dataset_name"] = dataset.name |
| |
|
| | yield create |
| |
|
| |
|
| | @pytest.fixture() |
| | def teardown_dataset(shared_state, dataset_client): |
| | yield |
| |
|
| | |
| | dataset_client.delete_dataset(name=shared_state["dataset_name"]) |
| |
|
| |
|
| | @pytest.fixture() |
| | def teardown_featurestore(shared_state, featurestore_client): |
| | yield |
| |
|
| | |
| | force_delete_featurestore_request = { |
| | "name": shared_state["featurestore_name"], |
| | "force": True, |
| | } |
| | featurestore_client.delete_featurestore(request=force_delete_featurestore_request) |
| |
|
| |
|
| | @pytest.fixture() |
| | def teardown_entity_type(shared_state, featurestore_client): |
| | yield |
| |
|
| | |
| | force_delete_entity_type_request = { |
| | "name": shared_state["entity_type_name"], |
| | "force": True, |
| | } |
| | featurestore_client.delete_entity_type(request=force_delete_entity_type_request) |
| |
|
| |
|
| | @pytest.fixture() |
| | def teardown_feature(shared_state, featurestore_client): |
| | yield |
| |
|
| | |
| | featurestore_client.delete_feature(name=shared_state["feature_name"]) |
| |
|
| |
|
| | @pytest.fixture() |
| | def teardown_features(shared_state, featurestore_client): |
| | yield |
| |
|
| | |
| | for feature_name in shared_state["feature_names"]: |
| | featurestore_client.delete_feature(name=feature_name) |
| |
|
| |
|
| | @pytest.fixture() |
| | def teardown_batch_read_feature_values(shared_state, bigquery_client): |
| | yield |
| |
|
| | |
| | bigquery_client.delete_dataset( |
| | shared_state["destination_data_set"], delete_contents=True, not_found_ok=True |
| | ) |
| |
|
| |
|
| | @pytest.fixture() |
| | def create_endpoint(shared_state, endpoint_client): |
| | def create(project, location, test_name="temp_deploy_model_test"): |
| | parent = f"projects/{project}/locations/{location}" |
| | endpoint = aiplatform.gapic.Endpoint( |
| | display_name=f"{test_name}_{uuid4()}", |
| | ) |
| | create_endpoint_response = endpoint_client.create_endpoint( |
| | parent=parent, endpoint=endpoint |
| | ) |
| |
|
| | endpoint = create_endpoint_response.result() |
| | shared_state["endpoint_name"] = endpoint.name |
| |
|
| | yield create |
| |
|
| |
|
| | @pytest.fixture() |
| | def teardown_endpoint(shared_state, endpoint_client): |
| | yield |
| |
|
| | undeploy_model_operation = endpoint_client.undeploy_model( |
| | deployed_model_id=shared_state["deployed_model_id"], |
| | endpoint=shared_state["endpoint_name"], |
| | ) |
| | undeploy_model_operation.result() |
| |
|
| | |
| | endpoint_client.delete_endpoint(name=shared_state["endpoint_name"]) |
| |
|
| |
|
| | @pytest.fixture() |
| | def teardown_model(shared_state, model_client): |
| | yield |
| |
|
| | model_client.delete_model(name=shared_state["model_name"]) |
| |
|