| | import pytest |
| |
|
| | from logging import info |
| | from test.testutil import env_kafka_version, random_string |
| | from threading import Event, Thread |
| | from time import time, sleep |
| |
|
| | from kafka.admin import ( |
| | ACLFilter, ACLOperation, ACLPermissionType, ResourcePattern, ResourceType, ACL, ConfigResource, ConfigResourceType) |
| | from kafka.errors import (NoError, GroupCoordinatorNotAvailableError, NonEmptyGroupError, GroupIdNotFoundError) |
| |
|
| |
|
| | @pytest.mark.skipif(env_kafka_version() < (0, 11), reason="ACL features require broker >=0.11") |
| | def test_create_describe_delete_acls(kafka_admin_client): |
| | """Tests that we can add, list and remove ACLs |
| | """ |
| |
|
| | |
| | acls, error = kafka_admin_client.describe_acls( |
| | ACLFilter( |
| | principal=None, |
| | host="*", |
| | operation=ACLOperation.ANY, |
| | permission_type=ACLPermissionType.ANY, |
| | resource_pattern=ResourcePattern(ResourceType.TOPIC, "topic") |
| | ) |
| | ) |
| |
|
| | assert error is NoError |
| | assert len(acls) == 0 |
| |
|
| | |
| | acl = ACL( |
| | principal="User:test", |
| | host="*", |
| | operation=ACLOperation.READ, |
| | permission_type=ACLPermissionType.ALLOW, |
| | resource_pattern=ResourcePattern(ResourceType.TOPIC, "topic") |
| | ) |
| | result = kafka_admin_client.create_acls([acl]) |
| |
|
| | assert len(result["failed"]) == 0 |
| | assert len(result["succeeded"]) == 1 |
| |
|
| | |
| | acl_filter = ACLFilter( |
| | principal=None, |
| | host="*", |
| | operation=ACLOperation.ANY, |
| | permission_type=ACLPermissionType.ANY, |
| | resource_pattern=ResourcePattern(ResourceType.TOPIC, "topic") |
| | ) |
| | acls, error = kafka_admin_client.describe_acls(acl_filter) |
| |
|
| | assert error is NoError |
| | assert len(acls) == 1 |
| |
|
| | |
| | delete_results = kafka_admin_client.delete_acls( |
| | [ |
| | ACLFilter( |
| | principal="User:test", |
| | host="*", |
| | operation=ACLOperation.READ, |
| | permission_type=ACLPermissionType.ALLOW, |
| | resource_pattern=ResourcePattern(ResourceType.TOPIC, "topic") |
| | ) |
| | ] |
| | ) |
| |
|
| | assert len(delete_results) == 1 |
| | assert len(delete_results[0][1]) == 1 |
| |
|
| | |
| | acls, error = kafka_admin_client.describe_acls( |
| | ACLFilter( |
| | principal="*", |
| | host="*", |
| | operation=ACLOperation.ANY, |
| | permission_type=ACLPermissionType.ANY, |
| | resource_pattern=ResourcePattern(ResourceType.TOPIC, "topic") |
| | ) |
| | ) |
| |
|
| | assert error is NoError |
| | assert len(acls) == 0 |
| |
|
| |
|
| | @pytest.mark.skipif(env_kafka_version() < (0, 11), reason="Describe config features require broker >=0.11") |
| | def test_describe_configs_broker_resource_returns_configs(kafka_admin_client): |
| | """Tests that describe config returns configs for broker |
| | """ |
| | broker_id = kafka_admin_client._client.cluster._brokers[0].nodeId |
| | configs = kafka_admin_client.describe_configs([ConfigResource(ConfigResourceType.BROKER, broker_id)]) |
| |
|
| | assert len(configs) == 1 |
| | assert configs[0].resources[0][2] == ConfigResourceType.BROKER |
| | assert configs[0].resources[0][3] == str(broker_id) |
| | assert len(configs[0].resources[0][4]) > 1 |
| |
|
| |
|
| | @pytest.mark.xfail(condition=True, |
| | reason="https://github.com/dpkp/kafka-python/issues/1929", |
| | raises=AssertionError) |
| | @pytest.mark.skipif(env_kafka_version() < (0, 11), reason="Describe config features require broker >=0.11") |
| | def test_describe_configs_topic_resource_returns_configs(topic, kafka_admin_client): |
| | """Tests that describe config returns configs for topic |
| | """ |
| | configs = kafka_admin_client.describe_configs([ConfigResource(ConfigResourceType.TOPIC, topic)]) |
| |
|
| | assert len(configs) == 1 |
| | assert configs[0].resources[0][2] == ConfigResourceType.TOPIC |
| | assert configs[0].resources[0][3] == topic |
| | assert len(configs[0].resources[0][4]) > 1 |
| |
|
| |
|
| | @pytest.mark.skipif(env_kafka_version() < (0, 11), reason="Describe config features require broker >=0.11") |
| | def test_describe_configs_mixed_resources_returns_configs(topic, kafka_admin_client): |
| | """Tests that describe config returns configs for mixed resource types (topic + broker) |
| | """ |
| | broker_id = kafka_admin_client._client.cluster._brokers[0].nodeId |
| | configs = kafka_admin_client.describe_configs([ |
| | ConfigResource(ConfigResourceType.TOPIC, topic), |
| | ConfigResource(ConfigResourceType.BROKER, broker_id)]) |
| |
|
| | assert len(configs) == 2 |
| |
|
| | for config in configs: |
| | assert (config.resources[0][2] == ConfigResourceType.TOPIC |
| | and config.resources[0][3] == topic) or \ |
| | (config.resources[0][2] == ConfigResourceType.BROKER |
| | and config.resources[0][3] == str(broker_id)) |
| | assert len(config.resources[0][4]) > 1 |
| |
|
| |
|
| | @pytest.mark.skipif(env_kafka_version() < (0, 11), reason="Describe config features require broker >=0.11") |
| | def test_describe_configs_invalid_broker_id_raises(kafka_admin_client): |
| | """Tests that describe config raises exception on non-integer broker id |
| | """ |
| | broker_id = "str" |
| |
|
| | with pytest.raises(ValueError): |
| | configs = kafka_admin_client.describe_configs([ConfigResource(ConfigResourceType.BROKER, broker_id)]) |
| |
|
| |
|
| | @pytest.mark.skipif(env_kafka_version() < (0, 11), reason='Describe consumer group requires broker >=0.11') |
| | def test_describe_consumer_group_does_not_exist(kafka_admin_client): |
| | """Tests that the describe consumer group call fails if the group coordinator is not available |
| | """ |
| | with pytest.raises(GroupCoordinatorNotAvailableError): |
| | group_description = kafka_admin_client.describe_consumer_groups(['test']) |
| |
|
| |
|
| | @pytest.mark.skipif(env_kafka_version() < (0, 11), reason='Describe consumer group requires broker >=0.11') |
| | def test_describe_consumer_group_exists(kafka_admin_client, kafka_consumer_factory, topic): |
| | """Tests that the describe consumer group call returns valid consumer group information |
| | This test takes inspiration from the test 'test_group' in test_consumer_group.py. |
| | """ |
| | consumers = {} |
| | stop = {} |
| | threads = {} |
| | random_group_id = 'test-group-' + random_string(6) |
| | group_id_list = [random_group_id, random_group_id + '_2'] |
| | generations = {group_id_list[0]: set(), group_id_list[1]: set()} |
| | def consumer_thread(i, group_id): |
| | assert i not in consumers |
| | assert i not in stop |
| | stop[i] = Event() |
| | consumers[i] = kafka_consumer_factory(group_id=group_id) |
| | while not stop[i].is_set(): |
| | consumers[i].poll(20) |
| | consumers[i].close() |
| | consumers[i] = None |
| | stop[i] = None |
| |
|
| | num_consumers = 3 |
| | for i in range(num_consumers): |
| | group_id = group_id_list[i % 2] |
| | t = Thread(target=consumer_thread, args=(i, group_id,)) |
| | t.start() |
| | threads[i] = t |
| |
|
| | try: |
| | timeout = time() + 35 |
| | while True: |
| | for c in range(num_consumers): |
| |
|
| | |
| | if c not in consumers: |
| | break |
| |
|
| | |
| | elif not consumers[c].assignment(): |
| | break |
| |
|
| | |
| | else: |
| |
|
| | info('All consumers have assignment... checking for stable group') |
| | |
| | |
| |
|
| | for consumer in consumers.values(): |
| | generations[consumer.config['group_id']].add(consumer._coordinator._generation.generation_id) |
| |
|
| | is_same_generation = any([len(consumer_generation) == 1 for consumer_generation in generations.values()]) |
| |
|
| | |
| | |
| | rejoining = any([consumer._coordinator.rejoining |
| | for consumer in list(consumers.values())]) |
| |
|
| | if not rejoining and is_same_generation: |
| | break |
| | else: |
| | sleep(1) |
| | assert time() < timeout, "timeout waiting for assignments" |
| |
|
| | info('Group stabilized; verifying assignment') |
| | output = kafka_admin_client.describe_consumer_groups(group_id_list) |
| | assert len(output) == 2 |
| | consumer_groups = set() |
| | for consumer_group in output: |
| | assert(consumer_group.group in group_id_list) |
| | if consumer_group.group == group_id_list[0]: |
| | assert(len(consumer_group.members) == 2) |
| | else: |
| | assert(len(consumer_group.members) == 1) |
| | for member in consumer_group.members: |
| | assert(member.member_metadata.subscription[0] == topic) |
| | assert(member.member_assignment.assignment[0][0] == topic) |
| | consumer_groups.add(consumer_group.group) |
| | assert(sorted(list(consumer_groups)) == group_id_list) |
| | finally: |
| | info('Shutting down %s consumers', num_consumers) |
| | for c in range(num_consumers): |
| | info('Stopping consumer %s', c) |
| | stop[c].set() |
| | threads[c].join() |
| | threads[c] = None |
| |
|
| |
|
| | @pytest.mark.skipif(env_kafka_version() < (1, 1), reason="Delete consumer groups requires broker >=1.1") |
| | def test_delete_consumergroups(kafka_admin_client, kafka_consumer_factory, send_messages): |
| | random_group_id = 'test-group-' + random_string(6) |
| | group1 = random_group_id + "_1" |
| | group2 = random_group_id + "_2" |
| | group3 = random_group_id + "_3" |
| |
|
| | send_messages(range(0, 100), partition=0) |
| | consumer1 = kafka_consumer_factory(group_id=group1) |
| | next(consumer1) |
| | consumer1.close() |
| |
|
| | consumer2 = kafka_consumer_factory(group_id=group2) |
| | next(consumer2) |
| | consumer2.close() |
| |
|
| | consumer3 = kafka_consumer_factory(group_id=group3) |
| | next(consumer3) |
| | consumer3.close() |
| |
|
| | consumergroups = {group_id for group_id, _ in kafka_admin_client.list_consumer_groups()} |
| | assert group1 in consumergroups |
| | assert group2 in consumergroups |
| | assert group3 in consumergroups |
| |
|
| | delete_results = { |
| | group_id: error |
| | for group_id, error in kafka_admin_client.delete_consumer_groups([group1, group2]) |
| | } |
| | assert delete_results[group1] == NoError |
| | assert delete_results[group2] == NoError |
| | assert group3 not in delete_results |
| |
|
| | consumergroups = {group_id for group_id, _ in kafka_admin_client.list_consumer_groups()} |
| | assert group1 not in consumergroups |
| | assert group2 not in consumergroups |
| | assert group3 in consumergroups |
| |
|
| |
|
| | @pytest.mark.skipif(env_kafka_version() < (1, 1), reason="Delete consumer groups requires broker >=1.1") |
| | def test_delete_consumergroups_with_errors(kafka_admin_client, kafka_consumer_factory, send_messages): |
| | random_group_id = 'test-group-' + random_string(6) |
| | group1 = random_group_id + "_1" |
| | group2 = random_group_id + "_2" |
| | group3 = random_group_id + "_3" |
| |
|
| | send_messages(range(0, 100), partition=0) |
| | consumer1 = kafka_consumer_factory(group_id=group1) |
| | next(consumer1) |
| | consumer1.close() |
| |
|
| | consumer2 = kafka_consumer_factory(group_id=group2) |
| | next(consumer2) |
| |
|
| | consumergroups = {group_id for group_id, _ in kafka_admin_client.list_consumer_groups()} |
| | assert group1 in consumergroups |
| | assert group2 in consumergroups |
| | assert group3 not in consumergroups |
| |
|
| | delete_results = { |
| | group_id: error |
| | for group_id, error in kafka_admin_client.delete_consumer_groups([group1, group2, group3]) |
| | } |
| |
|
| | assert delete_results[group1] == NoError |
| | assert delete_results[group2] == NonEmptyGroupError |
| | assert delete_results[group3] == GroupIdNotFoundError |
| |
|
| | consumergroups = {group_id for group_id, _ in kafka_admin_client.list_consumer_groups()} |
| | assert group1 not in consumergroups |
| | assert group2 in consumergroups |
| | assert group3 not in consumergroups |
| |
|