| | import logging |
| | import time |
| |
|
| | from mock import patch |
| | import pytest |
| | from kafka.vendor.six.moves import range |
| |
|
| | import kafka.codec |
| | from kafka.errors import UnsupportedCodecError, UnsupportedVersionError |
| | from kafka.structs import TopicPartition, OffsetAndTimestamp |
| |
|
| | from test.testutil import Timer, assert_message_count, env_kafka_version, random_string |
| |
|
| |
|
| | @pytest.mark.skipif(not env_kafka_version(), reason="No KAFKA_VERSION set") |
| | def test_kafka_version_infer(kafka_consumer_factory): |
| | consumer = kafka_consumer_factory() |
| | actual_ver_major_minor = env_kafka_version()[:2] |
| | client = consumer._client |
| | conn = list(client._conns.values())[0] |
| | inferred_ver_major_minor = conn.check_version()[:2] |
| | assert actual_ver_major_minor == inferred_ver_major_minor, \ |
| | "Was expecting inferred broker version to be %s but was %s" % (actual_ver_major_minor, inferred_ver_major_minor) |
| |
|
| |
|
| | @pytest.mark.skipif(not env_kafka_version(), reason="No KAFKA_VERSION set") |
| | def test_kafka_consumer(kafka_consumer_factory, send_messages): |
| | """Test KafkaConsumer""" |
| | consumer = kafka_consumer_factory(auto_offset_reset='earliest') |
| | send_messages(range(0, 100), partition=0) |
| | send_messages(range(0, 100), partition=1) |
| | cnt = 0 |
| | messages = {0: [], 1: []} |
| | for message in consumer: |
| | logging.debug("Consumed message %s", repr(message)) |
| | cnt += 1 |
| | messages[message.partition].append(message) |
| | if cnt >= 200: |
| | break |
| |
|
| | assert_message_count(messages[0], 100) |
| | assert_message_count(messages[1], 100) |
| |
|
| |
|
| | @pytest.mark.skipif(not env_kafka_version(), reason="No KAFKA_VERSION set") |
| | def test_kafka_consumer_unsupported_encoding( |
| | topic, kafka_producer_factory, kafka_consumer_factory): |
| | |
| | producer = kafka_producer_factory(compression_type="gzip") |
| | fut = producer.send(topic, b"simple message" * 200) |
| | fut.get(timeout=5) |
| | producer.close() |
| |
|
| | |
| | with patch.object(kafka.codec, "has_gzip") as mocked: |
| | mocked.return_value = False |
| | consumer = kafka_consumer_factory(auto_offset_reset='earliest') |
| | error_msg = "Libraries for gzip compression codec not found" |
| | with pytest.raises(UnsupportedCodecError, match=error_msg): |
| | consumer.poll(timeout_ms=2000) |
| |
|
| |
|
| | @pytest.mark.skipif(not env_kafka_version(), reason="No KAFKA_VERSION set") |
| | def test_kafka_consumer__blocking(kafka_consumer_factory, topic, send_messages): |
| | TIMEOUT_MS = 500 |
| | consumer = kafka_consumer_factory(auto_offset_reset='earliest', |
| | enable_auto_commit=False, |
| | consumer_timeout_ms=TIMEOUT_MS) |
| |
|
| | |
| | consumer.unsubscribe() |
| | consumer.assign([TopicPartition(topic, 0)]) |
| |
|
| | |
| | with Timer() as t: |
| | with pytest.raises(StopIteration): |
| | msg = next(consumer) |
| | assert t.interval >= (TIMEOUT_MS / 1000.0) |
| |
|
| | send_messages(range(0, 10)) |
| |
|
| | |
| | messages = [] |
| | with Timer() as t: |
| | for i in range(5): |
| | msg = next(consumer) |
| | messages.append(msg) |
| | assert_message_count(messages, 5) |
| | assert t.interval < (TIMEOUT_MS / 1000.0) |
| |
|
| | |
| | messages = [] |
| | with Timer() as t: |
| | with pytest.raises(StopIteration): |
| | for i in range(10): |
| | msg = next(consumer) |
| | messages.append(msg) |
| | assert_message_count(messages, 5) |
| | assert t.interval >= (TIMEOUT_MS / 1000.0) |
| |
|
| |
|
| | @pytest.mark.skipif(env_kafka_version() < (0, 8, 1), reason="Requires KAFKA_VERSION >= 0.8.1") |
| | def test_kafka_consumer__offset_commit_resume(kafka_consumer_factory, send_messages): |
| | GROUP_ID = random_string(10) |
| |
|
| | send_messages(range(0, 100), partition=0) |
| | send_messages(range(100, 200), partition=1) |
| |
|
| | |
| | consumer1 = kafka_consumer_factory( |
| | group_id=GROUP_ID, |
| | enable_auto_commit=True, |
| | auto_commit_interval_ms=100, |
| | auto_offset_reset='earliest', |
| | ) |
| | output_msgs1 = [] |
| | for _ in range(180): |
| | m = next(consumer1) |
| | output_msgs1.append(m) |
| | assert_message_count(output_msgs1, 180) |
| |
|
| | |
| | |
| | |
| | |
| | consumer1.close() |
| |
|
| | |
| | consumer2 = kafka_consumer_factory( |
| | group_id=GROUP_ID, |
| | enable_auto_commit=True, |
| | auto_commit_interval_ms=100, |
| | auto_offset_reset='earliest', |
| | ) |
| | output_msgs2 = [] |
| | for _ in range(20): |
| | m = next(consumer2) |
| | output_msgs2.append(m) |
| | assert_message_count(output_msgs2, 20) |
| |
|
| | |
| | |
| | assert_message_count(output_msgs1 + output_msgs2, 200) |
| |
|
| |
|
| | @pytest.mark.skipif(env_kafka_version() < (0, 10, 1), reason="Requires KAFKA_VERSION >= 0.10.1") |
| | def test_kafka_consumer_max_bytes_simple(kafka_consumer_factory, topic, send_messages): |
| | send_messages(range(100, 200), partition=0) |
| | send_messages(range(200, 300), partition=1) |
| |
|
| | |
| | consumer = kafka_consumer_factory( |
| | auto_offset_reset='earliest', fetch_max_bytes=300) |
| | seen_partitions = set() |
| | for i in range(90): |
| | poll_res = consumer.poll(timeout_ms=100) |
| | for partition, msgs in poll_res.items(): |
| | for msg in msgs: |
| | seen_partitions.add(partition) |
| |
|
| | |
| | assert seen_partitions == {TopicPartition(topic, 0), TopicPartition(topic, 1)} |
| |
|
| |
|
| | @pytest.mark.skipif(env_kafka_version() < (0, 10, 1), reason="Requires KAFKA_VERSION >= 0.10.1") |
| | def test_kafka_consumer_max_bytes_one_msg(kafka_consumer_factory, send_messages): |
| | |
| | |
| | send_messages(range(100, 200)) |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | group = 'test-kafka-consumer-max-bytes-one-msg-' + random_string(5) |
| | consumer = kafka_consumer_factory( |
| | group_id=group, |
| | auto_offset_reset='earliest', |
| | consumer_timeout_ms=5000, |
| | fetch_max_bytes=1) |
| |
|
| | fetched_msgs = [next(consumer) for i in range(10)] |
| | assert_message_count(fetched_msgs, 10) |
| |
|
| |
|
| | @pytest.mark.skipif(env_kafka_version() < (0, 10, 1), reason="Requires KAFKA_VERSION >= 0.10.1") |
| | def test_kafka_consumer_offsets_for_time(topic, kafka_consumer, kafka_producer): |
| | late_time = int(time.time()) * 1000 |
| | middle_time = late_time - 1000 |
| | early_time = late_time - 2000 |
| | tp = TopicPartition(topic, 0) |
| |
|
| | timeout = 10 |
| | early_msg = kafka_producer.send( |
| | topic, partition=0, value=b"first", |
| | timestamp_ms=early_time).get(timeout) |
| | late_msg = kafka_producer.send( |
| | topic, partition=0, value=b"last", |
| | timestamp_ms=late_time).get(timeout) |
| |
|
| | consumer = kafka_consumer |
| | offsets = consumer.offsets_for_times({tp: early_time}) |
| | assert len(offsets) == 1 |
| | assert offsets[tp].offset == early_msg.offset |
| | assert offsets[tp].timestamp == early_time |
| |
|
| | offsets = consumer.offsets_for_times({tp: middle_time}) |
| | assert offsets[tp].offset == late_msg.offset |
| | assert offsets[tp].timestamp == late_time |
| |
|
| | offsets = consumer.offsets_for_times({tp: late_time}) |
| | assert offsets[tp].offset == late_msg.offset |
| | assert offsets[tp].timestamp == late_time |
| |
|
| | offsets = consumer.offsets_for_times({}) |
| | assert offsets == {} |
| |
|
| | |
| |
|
| | offsets = consumer.offsets_for_times({tp: 0}) |
| | assert offsets[tp].offset == early_msg.offset |
| | assert offsets[tp].timestamp == early_time |
| |
|
| | offsets = consumer.offsets_for_times({tp: 9999999999999}) |
| | assert offsets[tp] is None |
| |
|
| | |
| |
|
| | offsets = consumer.beginning_offsets([tp]) |
| | assert offsets == {tp: early_msg.offset} |
| | offsets = consumer.end_offsets([tp]) |
| | assert offsets == {tp: late_msg.offset + 1} |
| |
|
| |
|
| | @pytest.mark.skipif(env_kafka_version() < (0, 10, 1), reason="Requires KAFKA_VERSION >= 0.10.1") |
| | def test_kafka_consumer_offsets_search_many_partitions(kafka_consumer, kafka_producer, topic): |
| | tp0 = TopicPartition(topic, 0) |
| | tp1 = TopicPartition(topic, 1) |
| |
|
| | send_time = int(time.time() * 1000) |
| | timeout = 10 |
| | p0msg = kafka_producer.send( |
| | topic, partition=0, value=b"XXX", |
| | timestamp_ms=send_time).get(timeout) |
| | p1msg = kafka_producer.send( |
| | topic, partition=1, value=b"XXX", |
| | timestamp_ms=send_time).get(timeout) |
| |
|
| | consumer = kafka_consumer |
| | offsets = consumer.offsets_for_times({ |
| | tp0: send_time, |
| | tp1: send_time |
| | }) |
| |
|
| | assert offsets == { |
| | tp0: OffsetAndTimestamp(p0msg.offset, send_time), |
| | tp1: OffsetAndTimestamp(p1msg.offset, send_time) |
| | } |
| |
|
| | offsets = consumer.beginning_offsets([tp0, tp1]) |
| | assert offsets == { |
| | tp0: p0msg.offset, |
| | tp1: p1msg.offset |
| | } |
| |
|
| | offsets = consumer.end_offsets([tp0, tp1]) |
| | assert offsets == { |
| | tp0: p0msg.offset + 1, |
| | tp1: p1msg.offset + 1 |
| | } |
| |
|
| |
|
| | @pytest.mark.skipif(env_kafka_version() >= (0, 10, 1), reason="Requires KAFKA_VERSION < 0.10.1") |
| | def test_kafka_consumer_offsets_for_time_old(kafka_consumer, topic): |
| | consumer = kafka_consumer |
| | tp = TopicPartition(topic, 0) |
| |
|
| | with pytest.raises(UnsupportedVersionError): |
| | consumer.offsets_for_times({tp: int(time.time())}) |
| |
|
| |
|
| | @pytest.mark.skipif(env_kafka_version() < (0, 10, 1), reason="Requires KAFKA_VERSION >= 0.10.1") |
| | def test_kafka_consumer_offsets_for_times_errors(kafka_consumer_factory, topic): |
| | consumer = kafka_consumer_factory(fetch_max_wait_ms=200, |
| | request_timeout_ms=500) |
| | tp = TopicPartition(topic, 0) |
| | bad_tp = TopicPartition(topic, 100) |
| |
|
| | with pytest.raises(ValueError): |
| | consumer.offsets_for_times({tp: -1}) |
| |
|
| | assert consumer.offsets_for_times({bad_tp: 0}) == {bad_tp: None} |
| |
|