File size: 4,567 Bytes
d8ad0fd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
import datetime
import json
import os
import platform
from importlib.metadata import version

import boto3
from pydantic import HttpUrl

from aws_lambda_powertools import Logger, Metrics, Tracer
from aws_lambda_powertools.event_handler import APIGatewayRestResolver
from aws_lambda_powertools.utilities.data_masking import DataMasking
from aws_lambda_powertools.utilities.parser import BaseModel, envelopes, event_parser
from aws_lambda_powertools.utilities.typing import LambdaContext
from aws_lambda_powertools.utilities.validation import validator

logger = Logger(service="version-track")
tracer = Tracer()  # this checks for aws-xray-sdk presence
metrics = Metrics(namespace="powertools-layer-canary", service="PowertoolsLayerCanary")
data_masker = DataMasking()
app = APIGatewayRestResolver()

layer_arn = os.getenv("POWERTOOLS_LAYER_ARN")
powertools_version = os.getenv("POWERTOOLS_VERSION")
stage = os.getenv("LAYER_PIPELINE_STAGE")
event_bus_arn = os.getenv("VERSION_TRACKING_EVENT_BUS_ARN")


# Model to check parser imports correctly, tests for pydantic
class OrderItem(BaseModel):
    order_id: int
    quantity: int
    description: str
    url: HttpUrl


# Tests for jmespath presence
@event_parser(model=OrderItem, envelope=envelopes.EventBridgeEnvelope)
def envelope_handler(event: OrderItem, context: LambdaContext):
    assert event.order_id != 1


# Tests for fastjsonschema presence
@validator(inbound_schema={}, envelope="detail")
def validator_handler(event, context: LambdaContext):
    pass


def handler(event):
    logger.info("Running checks")
    check_envs()
    verify_powertools_version()
    send_notification()
    return True


@logger.inject_lambda_context(log_event=True)
def on_event(event, context):
    request_type = event["RequestType"]
    # we handle only create events, because we recreate the canary on each run
    if request_type == "Create":
        return on_create(event)

    return "Nothing to be processed"


def on_create(event):
    props = event["ResourceProperties"]
    logger.info("create new resource with properties %s" % props)
    handler(event)


def check_envs():
    logger.info('Checking required envs ["POWERTOOLS_LAYER_ARN", "AWS_REGION", "STAGE"]')
    if not layer_arn:
        raise ValueError("POWERTOOLS_LAYER_ARN is not set. Aborting...")
    if not powertools_version:
        raise ValueError("POWERTOOLS_VERSION is not set. Aborting...")
    if not stage:
        raise ValueError("LAYER_PIPELINE_STAGE is not set. Aborting...")
    if not event_bus_arn:
        raise ValueError("VERSION_TRACKING_EVENT_BUS_ARN is not set. Aborting...")
    logger.info("All envs configured, continue...")


def verify_powertools_version() -> None:
    """
    fetches the version that we import from the Powertools for AWS Lambda (Python) layer and compares
    it with expected version set in environment variable, which we pass during deployment.
    :raise ValueError if the expected version is not the same as the version we get from the layer
    """
    logger.info("Checking Powertools for AWS Lambda (Python) version in library...")
    current_version = version("aws_lambda_powertools")
    if powertools_version != current_version:
        raise ValueError(
            f'Expected Powertools version is "{powertools_version}", but layer contains version "{current_version}"',
        )
    logger.info(f"Current Powertools version is: {current_version} [{_get_architecture()}]")


def send_notification():
    """
    sends an event to version tracking event bridge
    """
    if stage != "PROD":
        logger.info("Not sending notification to event bus, because this is not the PROD stage")
        return

    event = {
        "Time": datetime.datetime.now(),
        "Source": "powertools.layer.canary",
        "EventBusName": event_bus_arn,
        "DetailType": "deployment",
        "Detail": json.dumps(
            {
                "version": powertools_version,
                "region": os.environ["AWS_REGION"],
                "layerArn": layer_arn,
                "architecture": _get_architecture(),
            },
        ),
    }

    logger.info(f"sending notification event: {event}")

    client = boto3.client("events", region_name="eu-central-1")
    resp = client.put_events(Entries=[event])
    logger.info(resp)
    if resp["FailedEntryCount"] != 0:
        logger.error(resp)
        raise ValueError("Failed to send deployment notification to version tracking")


def _get_architecture() -> str:
    """Returns aarch64, x86_64"""
    return platform.uname()[4]