| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | from __future__ import absolute_import |
| |
|
| | import json |
| | import time |
| | from contextlib import contextmanager |
| |
|
| | import boto3 |
| | import numpy as np |
| | import pandas as pd |
| | import pytest |
| | from pandas import DataFrame |
| |
|
| | from sagemaker.feature_store.feature_definition import FractionalFeatureDefinition |
| | from sagemaker.feature_store.feature_group import FeatureGroup |
| | from sagemaker.feature_store.inputs import FeatureValue, FeatureParameter |
| | from sagemaker.session import get_execution_role, Session |
| | from tests.integ.timeout import timeout |
| |
|
| | BUCKET_POLICY = { |
| | "Version": "2012-10-17", |
| | "Statement": [ |
| | { |
| | "Sid": "FeatureStoreOfflineStoreS3BucketPolicy", |
| | "Effect": "Allow", |
| | "Principal": {"Service": "sagemaker.amazonaws.com"}, |
| | "Action": ["s3:PutObject", "s3:PutObjectAcl"], |
| | "Resource": "arn:aws:s3:::{bucket_name}-{region_name}/*", |
| | "Condition": {"StringEquals": {"s3:x-amz-acl": "bucket-owner-full-control"}}, |
| | }, |
| | { |
| | "Sid": "FeatureStoreOfflineStoreS3BucketPolicy", |
| | "Effect": "Allow", |
| | "Principal": {"Service": "sagemaker.amazonaws.com"}, |
| | "Action": "s3:GetBucketAcl", |
| | "Resource": "arn:aws:s3:::{bucket_name}-{region_name}", |
| | }, |
| | ], |
| | } |
| |
|
| |
|
| | @pytest.fixture(scope="module") |
| | def region_name(feature_store_session): |
| | return feature_store_session.boto_session.region_name |
| |
|
| |
|
| | @pytest.fixture(scope="module") |
| | def role(feature_store_session): |
| | return get_execution_role(feature_store_session) |
| |
|
| |
|
| | |
| | @pytest.fixture(scope="module") |
| | def feature_store_session(): |
| | boto_session = boto3.Session(region_name="us-east-2") |
| |
|
| | sagemaker_client = boto_session.client("sagemaker") |
| | featurestore_runtime_client = boto_session.client("sagemaker-featurestore-runtime") |
| |
|
| | return Session( |
| | boto_session=boto_session, |
| | sagemaker_client=sagemaker_client, |
| | sagemaker_featurestore_runtime_client=featurestore_runtime_client, |
| | ) |
| |
|
| |
|
| | @pytest.fixture |
| | def feature_group_name(): |
| | return f"my-feature-group-{int(time.time() * 10**7)}" |
| |
|
| |
|
| | @pytest.fixture |
| | def offline_store_s3_uri(feature_store_session, region_name): |
| | bucket = f"sagemaker-test-featurestore-{region_name}-{feature_store_session.account_id()}" |
| | feature_store_session._create_s3_bucket_if_it_does_not_exist(bucket, region_name) |
| | s3 = feature_store_session.boto_session.client("s3", region_name=region_name) |
| | BUCKET_POLICY["Statement"][0]["Resource"] = f"arn:aws:s3:::{bucket}/*" |
| | BUCKET_POLICY["Statement"][1]["Resource"] = f"arn:aws:s3:::{bucket}" |
| | s3.put_bucket_policy( |
| | Bucket=f"{bucket}", |
| | Policy=json.dumps(BUCKET_POLICY), |
| | ) |
| | return f"s3://{bucket}" |
| |
|
| |
|
| | @pytest.fixture |
| | def pandas_data_frame(): |
| | df = pd.DataFrame( |
| | { |
| | "feature1": pd.Series(np.arange(10.0), dtype="float64"), |
| | "feature2": pd.Series(np.arange(10), dtype="int64"), |
| | "feature3": pd.Series(["2020-10-30T03:43:21Z"] * 10, dtype="string"), |
| | "feature4": pd.Series(np.arange(5.0), dtype="float64"), |
| | } |
| | ) |
| | return df |
| |
|
| |
|
| | @pytest.fixture |
| | def pandas_data_frame_without_string(): |
| | df = pd.DataFrame( |
| | { |
| | "feature1": pd.Series(np.arange(10), dtype="int64"), |
| | "feature2": pd.Series([3141592.6535897] * 10, dtype="float64"), |
| | } |
| | ) |
| | return df |
| |
|
| |
|
| | @pytest.fixture |
| | def record(): |
| | return [ |
| | FeatureValue(feature_name="feature1", value_as_string="10.0"), |
| | FeatureValue(feature_name="feature2", value_as_string="10"), |
| | FeatureValue(feature_name="feature3", value_as_string="2020-10-30T03:43:21Z"), |
| | ] |
| |
|
| |
|
| | @pytest.fixture |
| | def create_table_ddl(): |
| | return ( |
| | "CREATE EXTERNAL TABLE IF NOT EXISTS sagemaker_featurestore.{feature_group_name} (\n" |
| | " feature1 FLOAT\n" |
| | " feature2 INT\n" |
| | " feature3 STRING\n" |
| | " feature4 FLOAT\n" |
| | " write_time TIMESTAMP\n" |
| | " event_time TIMESTAMP\n" |
| | " is_deleted BOOLEAN\n" |
| | ")\n" |
| | "ROW FORMAT SERDE 'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'\n" |
| | " STORED AS\n" |
| | " INPUTFORMAT 'parquet.hive.DeprecatedParquetInputFormat'\n" |
| | " OUTPUTFORMAT 'parquet.hive.DeprecatedParquetOutputFormat'\n" |
| | "LOCATION '{resolved_output_s3_uri}'" |
| | ) |
| |
|
| |
|
| | def test_create_feature_store_online_only( |
| | feature_store_session, |
| | role, |
| | feature_group_name, |
| | pandas_data_frame, |
| | ): |
| | feature_group = FeatureGroup(name=feature_group_name, sagemaker_session=feature_store_session) |
| | feature_group.load_feature_definitions(data_frame=pandas_data_frame) |
| |
|
| | with cleanup_feature_group(feature_group): |
| | output = feature_group.create( |
| | s3_uri=False, |
| | record_identifier_name="feature1", |
| | event_time_feature_name="feature3", |
| | role_arn=role, |
| | enable_online_store=True, |
| | ) |
| | _wait_for_feature_group_create(feature_group) |
| |
|
| | assert output["FeatureGroupArn"].endswith(f"feature-group/{feature_group_name}") |
| |
|
| |
|
| | def test_create_feature_store( |
| | feature_store_session, |
| | role, |
| | feature_group_name, |
| | offline_store_s3_uri, |
| | pandas_data_frame, |
| | record, |
| | create_table_ddl, |
| | ): |
| | feature_group = FeatureGroup(name=feature_group_name, sagemaker_session=feature_store_session) |
| | feature_group.load_feature_definitions(data_frame=pandas_data_frame) |
| |
|
| | with cleanup_feature_group(feature_group): |
| | output = feature_group.create( |
| | s3_uri=offline_store_s3_uri, |
| | record_identifier_name="feature1", |
| | event_time_feature_name="feature3", |
| | role_arn=role, |
| | enable_online_store=True, |
| | ) |
| | _wait_for_feature_group_create(feature_group) |
| |
|
| | resolved_output_s3_uri = ( |
| | feature_group.describe() |
| | .get("OfflineStoreConfig") |
| | .get("S3StorageConfig") |
| | .get("ResolvedOutputS3Uri") |
| | ) |
| | |
| | feature_group.put_record(record=record) |
| | ingestion_manager = feature_group.ingest( |
| | data_frame=pandas_data_frame, max_workers=3, wait=False |
| | ) |
| | ingestion_manager.wait() |
| | assert 0 == len(ingestion_manager.failed_rows) |
| |
|
| | |
| | athena_query = feature_group.athena_query() |
| | df = DataFrame() |
| | with timeout(minutes=10): |
| | while df.shape[0] < 11: |
| | athena_query.run( |
| | query_string=f'SELECT * FROM "{athena_query.table_name}"', |
| | output_location=f"{offline_store_s3_uri}/query_results", |
| | ) |
| | athena_query.wait() |
| | assert "SUCCEEDED" == athena_query.get_query_execution().get("QueryExecution").get( |
| | "Status" |
| | ).get("State") |
| | df = athena_query.as_dataframe() |
| | print(f"Found {df.shape[0]} records.") |
| | time.sleep(60) |
| |
|
| | assert df.shape[0] == 11 |
| | nans = pd.isna(df.loc[df["feature1"].isin([5, 6, 7, 8, 9])]["feature4"]) |
| | for is_na in nans.items(): |
| | assert is_na |
| | assert ( |
| | create_table_ddl.format( |
| | feature_group_name=feature_group_name, |
| | region=feature_store_session.boto_session.region_name, |
| | account=feature_store_session.account_id(), |
| | resolved_output_s3_uri=resolved_output_s3_uri, |
| | ) |
| | == feature_group.as_hive_ddl() |
| | ) |
| | assert output["FeatureGroupArn"].endswith(f"feature-group/{feature_group_name}") |
| |
|
| |
|
| | def test_update_feature_group( |
| | feature_store_session, |
| | role, |
| | feature_group_name, |
| | offline_store_s3_uri, |
| | pandas_data_frame, |
| | ): |
| | feature_group = FeatureGroup(name=feature_group_name, sagemaker_session=feature_store_session) |
| | feature_group.load_feature_definitions(data_frame=pandas_data_frame) |
| |
|
| | with cleanup_feature_group(feature_group): |
| | feature_group.create( |
| | s3_uri=offline_store_s3_uri, |
| | record_identifier_name="feature1", |
| | event_time_feature_name="feature3", |
| | role_arn=role, |
| | enable_online_store=True, |
| | ) |
| | _wait_for_feature_group_create(feature_group) |
| |
|
| | new_feature_name = "new_feature" |
| | new_features = [FractionalFeatureDefinition(feature_name=new_feature_name)] |
| | feature_group.update(new_features) |
| | _wait_for_feature_group_update(feature_group) |
| | feature_definitions = feature_group.describe().get("FeatureDefinitions") |
| | assert any([True for elem in feature_definitions if new_feature_name in elem.values()]) |
| |
|
| |
|
| | def test_feature_metadata( |
| | feature_store_session, |
| | role, |
| | feature_group_name, |
| | offline_store_s3_uri, |
| | pandas_data_frame, |
| | ): |
| | feature_group = FeatureGroup(name=feature_group_name, sagemaker_session=feature_store_session) |
| | feature_group.load_feature_definitions(data_frame=pandas_data_frame) |
| |
|
| | with cleanup_feature_group(feature_group): |
| | feature_group.create( |
| | s3_uri=offline_store_s3_uri, |
| | record_identifier_name="feature1", |
| | event_time_feature_name="feature3", |
| | role_arn=role, |
| | enable_online_store=True, |
| | ) |
| | _wait_for_feature_group_create(feature_group) |
| |
|
| | parameter_additions = [ |
| | FeatureParameter(key="key1", value="value1"), |
| | FeatureParameter(key="key2", value="value2"), |
| | ] |
| | description = "test description" |
| | feature_name = "feature1" |
| | feature_group.update_feature_metadata( |
| | feature_name=feature_name, |
| | description=description, |
| | parameter_additions=parameter_additions, |
| | ) |
| | describe_feature_metadata = feature_group.describe_feature_metadata( |
| | feature_name=feature_name |
| | ) |
| | print(describe_feature_metadata) |
| | assert description == describe_feature_metadata.get("Description") |
| | assert 2 == len(describe_feature_metadata.get("Parameters")) |
| |
|
| | parameter_removals = ["key1"] |
| | feature_group.update_feature_metadata( |
| | feature_name=feature_name, parameter_removals=parameter_removals |
| | ) |
| | describe_feature_metadata = feature_group.describe_feature_metadata( |
| | feature_name=feature_name |
| | ) |
| | assert description == describe_feature_metadata.get("Description") |
| | assert 1 == len(describe_feature_metadata.get("Parameters")) |
| |
|
| |
|
| | def test_ingest_without_string_feature( |
| | feature_store_session, |
| | role, |
| | feature_group_name, |
| | offline_store_s3_uri, |
| | pandas_data_frame_without_string, |
| | ): |
| | feature_group = FeatureGroup(name=feature_group_name, sagemaker_session=feature_store_session) |
| | feature_group.load_feature_definitions(data_frame=pandas_data_frame_without_string) |
| |
|
| | with cleanup_feature_group(feature_group): |
| | output = feature_group.create( |
| | s3_uri=offline_store_s3_uri, |
| | record_identifier_name="feature1", |
| | event_time_feature_name="feature2", |
| | role_arn=role, |
| | enable_online_store=True, |
| | ) |
| | _wait_for_feature_group_create(feature_group) |
| |
|
| | ingestion_manager = feature_group.ingest( |
| | data_frame=pandas_data_frame_without_string, max_workers=3, wait=False |
| | ) |
| | ingestion_manager.wait() |
| |
|
| | assert output["FeatureGroupArn"].endswith(f"feature-group/{feature_group_name}") |
| |
|
| |
|
| | def test_ingest_multi_process( |
| | feature_store_session, |
| | role, |
| | feature_group_name, |
| | offline_store_s3_uri, |
| | pandas_data_frame, |
| | ): |
| | feature_group = FeatureGroup(name=feature_group_name, sagemaker_session=feature_store_session) |
| | feature_group.load_feature_definitions(data_frame=pandas_data_frame) |
| |
|
| | with cleanup_feature_group(feature_group): |
| | output = feature_group.create( |
| | s3_uri=offline_store_s3_uri, |
| | record_identifier_name="feature1", |
| | event_time_feature_name="feature3", |
| | role_arn=role, |
| | enable_online_store=True, |
| | ) |
| | _wait_for_feature_group_create(feature_group) |
| |
|
| | feature_group.ingest( |
| | data_frame=pandas_data_frame, max_workers=3, max_processes=2, wait=True |
| | ) |
| |
|
| | assert output["FeatureGroupArn"].endswith(f"feature-group/{feature_group_name}") |
| |
|
| |
|
| | def _wait_for_feature_group_create(feature_group: FeatureGroup): |
| | status = feature_group.describe().get("FeatureGroupStatus") |
| | while status == "Creating": |
| | print("Waiting for Feature Group Creation") |
| | time.sleep(5) |
| | status = feature_group.describe().get("FeatureGroupStatus") |
| | if status != "Created": |
| | print(feature_group.describe()) |
| | raise RuntimeError(f"Failed to create feature group {feature_group.name}") |
| | print(f"FeatureGroup {feature_group.name} successfully created.") |
| |
|
| |
|
| | def _wait_for_feature_group_update(feature_group: FeatureGroup): |
| | status = feature_group.describe().get("LastUpdateStatus").get("Status") |
| | while status == "InProgress": |
| | print("Waiting for Feature Group Update") |
| | time.sleep(5) |
| | status = feature_group.describe().get("LastUpdateStatus").get("Status") |
| | if status != "Successful": |
| | print(feature_group.describe()) |
| | raise RuntimeError(f"Failed to update feature group {feature_group.name}") |
| | print(f"FeatureGroup {feature_group.name} successfully updated.") |
| |
|
| |
|
| | @contextmanager |
| | def cleanup_feature_group(feature_group: FeatureGroup): |
| | try: |
| | yield |
| | finally: |
| | try: |
| | feature_group.delete() |
| | except Exception: |
| | raise RuntimeError(f"Failed to delete feature group with name {feature_group.name}") |
| |
|