| | import json |
| | import time |
| | from contextlib import contextmanager |
| | from typing import Dict, Generator |
| |
|
| | import pytest |
| |
|
| | from aws_lambda_powertools import Metrics |
| | from aws_lambda_powertools.metrics import MetricUnit |
| | from aws_lambda_powertools.metrics import metrics as metrics_global |
| |
|
| | |
| | METRICS_VALIDATION_SLA: float = 0.002 |
| | METRICS_SERIALIZATION_SLA: float = 0.002 |
| |
|
| |
|
| | @contextmanager |
| | def timing() -> Generator: |
| | """ "Generator to quickly time operations. It can add 5ms so take that into account in elapsed time |
| | |
| | Examples |
| | -------- |
| | |
| | with timing() as t: |
| | print("something") |
| | elapsed = t() |
| | """ |
| | start = time.perf_counter() |
| | yield lambda: time.perf_counter() - start |
| |
|
| |
|
| | @pytest.fixture(scope="function", autouse=True) |
| | def reset_metric_set(): |
| | metrics = Metrics() |
| | metrics.clear_metrics() |
| | metrics_global.is_cold_start = True |
| | yield |
| |
|
| |
|
| | @pytest.fixture |
| | def namespace() -> str: |
| | return "test_namespace" |
| |
|
| |
|
| | @pytest.fixture |
| | def metric() -> Dict[str, str]: |
| | return {"name": "single_metric", "unit": MetricUnit.Count, "value": 1} |
| |
|
| |
|
| | def add_max_metrics_before_serialization(metrics_instance: Metrics): |
| | metrics_instance.add_dimension(name="test_dimension", value="test") |
| |
|
| | for i in range(99): |
| | metrics_instance.add_metric(name=f"metric_{i}", unit="Count", value=1) |
| |
|
| |
|
| | @pytest.mark.perf |
| | def test_metrics_large_operation_without_json_serialization_sla(namespace): |
| | |
| | my_metrics = Metrics(namespace=namespace) |
| |
|
| | |
| | with timing() as t: |
| | add_max_metrics_before_serialization(metrics_instance=my_metrics) |
| | my_metrics.serialize_metric_set() |
| |
|
| | |
| | elapsed = t() |
| | if elapsed > METRICS_VALIDATION_SLA: |
| | pytest.fail(f"Metric validation should be below {METRICS_VALIDATION_SLA}s: {elapsed}") |
| |
|
| |
|
| | @pytest.mark.perf |
| | def test_metrics_large_operation_and_json_serialization_sla(namespace): |
| | |
| | my_metrics = Metrics(namespace=namespace) |
| |
|
| | |
| | with timing() as t: |
| | add_max_metrics_before_serialization(metrics_instance=my_metrics) |
| | metrics = my_metrics.serialize_metric_set() |
| | print(json.dumps(metrics, separators=(",", ":"))) |
| |
|
| | |
| | elapsed = t() |
| | if elapsed > METRICS_SERIALIZATION_SLA: |
| | pytest.fail(f"Metric serialization should be below {METRICS_SERIALIZATION_SLA}s: {elapsed}") |
| |
|