Merge in RR/pyinfra from RED-6273-multi-tenant-storage to master
Squashed commit of the following:
commit 0fead1f8b59c9187330879b4e48d48355885c27c
Author: Julius Unverfehrt <julius.unverfehrt@iqser.com>
Date: Tue Mar 28 15:02:22 2023 +0200
fix typos
commit 892a803726946876f8b8cd7905a0e73c419b2fb1
Author: Matthias Bisping <matthias.bisping@axbit.com>
Date: Tue Mar 28 14:41:49 2023 +0200
Refactoring
Replace custom storage caching logic with LRU decorator
commit eafcd90260731e3360ce960571f07dee8f521327
Author: Julius Unverfehrt <julius.unverfehrt@iqser.com>
Date: Fri Mar 24 12:50:13 2023 +0100
fix bug in storage connection from endpoint
commit d0c9fb5b7d1c55ae2f90e8faa1efec9f7587c26a
Author: Julius Unverfehrt <julius.unverfehrt@iqser.com>
Date: Fri Mar 24 11:49:34 2023 +0100
add logs to PayloadProcessor
- set log messages to determine if x-tenant
storage connection is working
commit 97309fe58037b90469cf7a3de342d4749a0edfde
Author: Julius Unverfehrt <julius.unverfehrt@iqser.com>
Date: Fri Mar 24 10:41:59 2023 +0100
update PayloadProcessor
- introduce storage cache to make every unique
storage connection only once
- add functionality to pass optional processing
kwargs in queue message like the operation key to
the processing function
commit d48e8108fdc0d463c89aaa0d672061ab7dca83a0
Author: Julius Unverfehrt <julius.unverfehrt@iqser.com>
Date: Wed Mar 22 13:34:43 2023 +0100
add multi-tenant storage connection 1st iteration
- forward x-tenant-id from queue message header to
payload processor
- add functions to receive storage infos from an
endpoint or the config. This enables hashing and
caching of connections created from these infos
- add function to initialize storage connections
from storage infos
- streamline and refactor tests to make them more
readable and robust and to make it easier to add
new tests
- update payload processor with first iteration
of multi tenancy storage connection support
with connection caching and backwards compability
commit 52c047c47b98e62d0b834a9b9b6c0e2bb0db41e5
Author: Julius Unverfehrt <julius.unverfehrt@iqser.com>
Date: Tue Mar 21 15:35:57 2023 +0100
add AES/GCM cipher functions
- decrypt x-tenant storage connection strings
75 lines
2.4 KiB
Python
75 lines
2.4 KiB
Python
import gzip
|
|
import json
|
|
from operator import itemgetter
|
|
|
|
import pytest
|
|
import requests
|
|
|
|
from pyinfra.payload_processing.processor import make_payload_processor
|
|
|
|
|
|
@pytest.fixture
|
|
def target_file():
|
|
contents = {"numberOfPages": 10, "content1": "value1", "content2": "value2"}
|
|
return gzip.compress(json.dumps(contents).encode("utf-8"))
|
|
|
|
|
|
@pytest.fixture
|
|
def file_names(payload):
|
|
dossier_id, file_id, target_suffix, response_suffix = itemgetter(
|
|
"dossierId",
|
|
"fileId",
|
|
"targetFileExtension",
|
|
"responseFileExtension",
|
|
)(payload)
|
|
return f"{dossier_id}/{file_id}.{target_suffix}", f"{dossier_id}/{file_id}.{response_suffix}"
|
|
|
|
|
|
@pytest.fixture(scope="session")
|
|
def payload_processor(test_storage_config):
|
|
def file_processor_mock(json_file: dict):
|
|
return [json_file]
|
|
|
|
yield make_payload_processor(file_processor_mock, test_storage_config)
|
|
|
|
|
|
@pytest.mark.parametrize("storage_backend", ["s3"], scope="session")
|
|
@pytest.mark.parametrize("bucket_name", ["testbucket"], scope="session")
|
|
@pytest.mark.parametrize("monitoring_enabled", [True, False], scope="session")
|
|
@pytest.mark.parametrize("x_tenant_id", [None])
|
|
class TestPayloadProcessor:
|
|
def test_payload_processor_yields_correct_response_and_uploads_result(
|
|
self,
|
|
payload_processor,
|
|
storage,
|
|
bucket_name,
|
|
payload,
|
|
response_payload,
|
|
target_file,
|
|
file_names,
|
|
):
|
|
storage.clear_bucket(bucket_name)
|
|
storage.put_object(bucket_name, file_names[0], target_file)
|
|
response = payload_processor(payload)
|
|
|
|
assert response == response_payload
|
|
|
|
data_received = storage.get_object(bucket_name, file_names[1])
|
|
|
|
assert json.loads((gzip.decompress(data_received)).decode("utf-8")) == {
|
|
**payload,
|
|
"data": [json.loads(gzip.decompress(target_file).decode("utf-8"))],
|
|
}
|
|
|
|
def test_catching_of_processing_failure(self, payload_processor, storage, bucket_name, payload):
|
|
storage.clear_bucket(bucket_name)
|
|
with pytest.raises(Exception):
|
|
payload_processor(payload)
|
|
|
|
def test_prometheus_endpoint_is_available(self, test_storage_config, monitoring_enabled, storage_backend, x_tenant_id):
|
|
if monitoring_enabled:
|
|
resp = requests.get(
|
|
f"http://{test_storage_config.prometheus_host}:{test_storage_config.prometheus_port}/prometheus"
|
|
)
|
|
assert resp.status_code == 200
|