| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | from tempfile import TemporaryDirectory |
| | from typing import Optional |
| |
|
| | import pytest |
| | from pydantic import Field |
| |
|
| | from pyiceberg.avro.file import AvroFile |
| | from pyiceberg.avro.reader import ( |
| | DecimalReader, |
| | DefaultReader, |
| | DoubleReader, |
| | FloatReader, |
| | IntegerReader, |
| | MapReader, |
| | OptionReader, |
| | StringReader, |
| | StructReader, |
| | ) |
| | from pyiceberg.avro.resolver import resolve_reader, resolve_writer |
| | from pyiceberg.avro.writer import ( |
| | BinaryWriter, |
| | DefaultWriter, |
| | DoubleWriter, |
| | IntegerWriter, |
| | ListWriter, |
| | MapWriter, |
| | OptionWriter, |
| | StringWriter, |
| | StructWriter, |
| | ) |
| | from pyiceberg.exceptions import ResolveError |
| | from pyiceberg.io.pyarrow import PyArrowFileIO |
| | from pyiceberg.manifest import MANIFEST_ENTRY_SCHEMAS |
| | from pyiceberg.schema import Schema |
| | from pyiceberg.typedef import Record |
| | from pyiceberg.types import ( |
| | BinaryType, |
| | BooleanType, |
| | DecimalType, |
| | DoubleType, |
| | FloatType, |
| | IntegerType, |
| | ListType, |
| | LongType, |
| | MapType, |
| | NestedField, |
| | StringType, |
| | StructType, |
| | ) |
| |
|
| |
|
| | def test_resolver() -> None: |
| | write_schema = Schema( |
| | NestedField(1, "id", LongType()), |
| | NestedField(2, "data", StringType()), |
| | NestedField( |
| | 3, |
| | "location", |
| | StructType( |
| | NestedField(4, "lat", DoubleType()), |
| | NestedField(5, "long", DoubleType()), |
| | ), |
| | ), |
| | NestedField(6, "preferences", MapType(7, StringType(), 8, StringType())), |
| | schema_id=1, |
| | ) |
| |
|
| | location_struct = StructType( |
| | NestedField(4, "lat", DoubleType()), |
| | NestedField(5, "long", DoubleType()), |
| | ) |
| | read_schema = Schema( |
| | NestedField( |
| | 3, |
| | "location", |
| | location_struct, |
| | ), |
| | NestedField(1, "id", LongType(), required=False), |
| | NestedField(6, "preferences", MapType(7, StringType(), 8, StringType())), |
| | schema_id=1, |
| | ) |
| | read_tree = resolve_reader(write_schema, read_schema) |
| |
|
| | assert read_tree == StructReader( |
| | ( |
| | (1, OptionReader(option=IntegerReader())), |
| | (None, OptionReader(option=StringReader())), |
| | ( |
| | 0, |
| | OptionReader( |
| | option=StructReader( |
| | ( |
| | (0, OptionReader(option=DoubleReader())), |
| | (1, OptionReader(option=DoubleReader())), |
| | ), |
| | Record, |
| | location_struct, |
| | ), |
| | ), |
| | ), |
| | (2, OptionReader(option=MapReader(StringReader(), StringReader()))), |
| | ), |
| | Record, |
| | read_schema.as_struct(), |
| | ) |
| |
|
| |
|
| | def test_resolver_new_required_field() -> None: |
| | write_schema = Schema( |
| | NestedField(1, "id", LongType()), |
| | schema_id=1, |
| | ) |
| | read_schema = Schema( |
| | NestedField(1, "id", LongType()), |
| | NestedField(2, "data", StringType(), required=True), |
| | schema_id=1, |
| | ) |
| |
|
| | with pytest.raises(ResolveError) as exc_info: |
| | resolve_reader(write_schema, read_schema) |
| |
|
| | assert "2: data: required string is non-optional, and not part of the file schema" in str(exc_info.value) |
| |
|
| |
|
| | def test_resolver_invalid_evolution() -> None: |
| | write_schema = Schema( |
| | NestedField(1, "id", LongType()), |
| | schema_id=1, |
| | ) |
| | read_schema = Schema( |
| | NestedField(1, "id", DoubleType()), |
| | schema_id=1, |
| | ) |
| |
|
| | with pytest.raises(ResolveError) as exc_info: |
| | resolve_reader(write_schema, read_schema) |
| |
|
| | assert "Cannot promote long to double" in str(exc_info.value) |
| |
|
| |
|
| | def test_resolver_promotion_string_to_binary() -> None: |
| | write_schema = Schema( |
| | NestedField(1, "id", StringType()), |
| | schema_id=1, |
| | ) |
| | read_schema = Schema( |
| | NestedField(1, "id", BinaryType()), |
| | schema_id=1, |
| | ) |
| | resolve_reader(write_schema, read_schema) |
| |
|
| |
|
| | def test_resolver_promotion_binary_to_string() -> None: |
| | write_schema = Schema( |
| | NestedField(1, "id", BinaryType()), |
| | schema_id=1, |
| | ) |
| | read_schema = Schema( |
| | NestedField(1, "id", StringType()), |
| | schema_id=1, |
| | ) |
| | resolve_reader(write_schema, read_schema) |
| |
|
| |
|
| | def test_resolver_change_type() -> None: |
| | write_schema = Schema( |
| | NestedField(1, "properties", ListType(2, StringType())), |
| | schema_id=1, |
| | ) |
| | read_schema = Schema( |
| | NestedField(1, "properties", MapType(2, StringType(), 3, StringType())), |
| | schema_id=1, |
| | ) |
| |
|
| | with pytest.raises(ResolveError) as exc_info: |
| | resolve_reader(write_schema, read_schema) |
| |
|
| | assert "File/read schema are not aligned for list, got map<string, string>" in str(exc_info.value) |
| |
|
| |
|
| | def test_resolve_int_to_long() -> None: |
| | assert resolve_reader(IntegerType(), LongType()) == IntegerReader() |
| |
|
| |
|
| | def test_resolve_float_to_double() -> None: |
| | |
| | assert resolve_reader(FloatType(), DoubleType()) == FloatReader() |
| |
|
| |
|
| | def test_resolve_decimal_to_decimal() -> None: |
| | |
| | assert resolve_reader(DecimalType(19, 25), DecimalType(22, 25)) == DecimalReader(19, 25) |
| |
|
| |
|
| | def test_struct_not_aligned() -> None: |
| | with pytest.raises(ResolveError): |
| | assert resolve_reader(StructType(), StringType()) |
| |
|
| |
|
| | def test_map_not_aligned() -> None: |
| | with pytest.raises(ResolveError): |
| | assert resolve_reader(MapType(1, StringType(), 2, IntegerType()), StringType()) |
| |
|
| |
|
| | def test_primitive_not_aligned() -> None: |
| | with pytest.raises(ResolveError): |
| | assert resolve_reader(IntegerType(), MapType(1, StringType(), 2, IntegerType())) |
| |
|
| |
|
| | def test_integer_not_aligned() -> None: |
| | with pytest.raises(ResolveError): |
| | assert resolve_reader(IntegerType(), StringType()) |
| |
|
| |
|
| | def test_float_not_aligned() -> None: |
| | with pytest.raises(ResolveError): |
| | assert resolve_reader(FloatType(), StringType()) |
| |
|
| |
|
| | def test_string_not_aligned() -> None: |
| | with pytest.raises(ResolveError): |
| | assert resolve_reader(StringType(), FloatType()) |
| |
|
| |
|
| | def test_binary_not_aligned() -> None: |
| | with pytest.raises(ResolveError): |
| | assert resolve_reader(BinaryType(), FloatType()) |
| |
|
| |
|
| | def test_decimal_not_aligned() -> None: |
| | with pytest.raises(ResolveError): |
| | assert resolve_reader(DecimalType(22, 19), StringType()) |
| |
|
| |
|
| | def test_resolve_decimal_to_decimal_reduce_precision() -> None: |
| | |
| | with pytest.raises(ResolveError) as exc_info: |
| | _ = resolve_reader(DecimalType(19, 25), DecimalType(10, 25)) == DecimalReader(22, 25) |
| |
|
| | assert "Cannot reduce precision from decimal(19, 25) to decimal(10, 25)" in str(exc_info.value) |
| |
|
| |
|
| | def test_column_assignment() -> None: |
| | int_schema = { |
| | "type": "record", |
| | "name": "ints", |
| | "fields": [ |
| | {"name": "a", "type": "int", "field-id": 1}, |
| | {"name": "b", "type": "int", "field-id": 2}, |
| | {"name": "c", "type": "int", "field-id": 3}, |
| | ], |
| | } |
| |
|
| | from fastavro import parse_schema, writer |
| |
|
| | parsed_schema = parse_schema(int_schema) |
| |
|
| | int_records = [ |
| | { |
| | "a": 1, |
| | "b": 2, |
| | "c": 3, |
| | } |
| | ] |
| |
|
| | with TemporaryDirectory() as tmpdir: |
| | tmp_avro_file = tmpdir + "/manifest.avro" |
| | with open(tmp_avro_file, "wb") as out: |
| | writer(out, parsed_schema, int_records) |
| |
|
| | class Ints(Record): |
| | c: int = Field() |
| | d: Optional[int] = Field() |
| |
|
| | MANIFEST_ENTRY_SCHEMA = Schema( |
| | NestedField(3, "c", IntegerType(), required=True), |
| | NestedField(4, "d", IntegerType(), required=False), |
| | ) |
| |
|
| | with AvroFile[Ints](PyArrowFileIO().new_input(tmp_avro_file), MANIFEST_ENTRY_SCHEMA, {-1: Ints}) as reader: |
| | records = list(reader) |
| |
|
| | assert repr(records) == "[Ints[c=3, d=None]]" |
| |
|
| |
|
| | def test_resolver_initial_value() -> None: |
| | write_schema = Schema( |
| | NestedField(1, "name", StringType()), |
| | schema_id=1, |
| | ) |
| | read_schema = Schema( |
| | NestedField(2, "something", StringType(), required=False, initial_default="vo"), |
| | schema_id=2, |
| | ) |
| |
|
| | assert resolve_reader(write_schema, read_schema) == StructReader( |
| | ( |
| | (None, OptionReader(option=StringReader())), |
| | (0, DefaultReader("vo")), |
| | ), |
| | Record, |
| | read_schema.as_struct(), |
| | ) |
| |
|
| |
|
| | def test_resolve_writer() -> None: |
| | actual = resolve_writer(record_schema=MANIFEST_ENTRY_SCHEMAS[2], file_schema=MANIFEST_ENTRY_SCHEMAS[1]) |
| | expected = StructWriter(( |
| | (0, IntegerWriter()), |
| | (1, IntegerWriter()), |
| | ( |
| | 4, |
| | StructWriter(( |
| | (1, StringWriter()), |
| | (2, StringWriter()), |
| | (3, StructWriter(())), |
| | (4, IntegerWriter()), |
| | (5, IntegerWriter()), |
| | (None, DefaultWriter(writer=IntegerWriter(), value=67108864)), |
| | (6, OptionWriter(option=MapWriter(key_writer=IntegerWriter(), value_writer=IntegerWriter()))), |
| | (7, OptionWriter(option=MapWriter(key_writer=IntegerWriter(), value_writer=IntegerWriter()))), |
| | (8, OptionWriter(option=MapWriter(key_writer=IntegerWriter(), value_writer=IntegerWriter()))), |
| | (9, OptionWriter(option=MapWriter(key_writer=IntegerWriter(), value_writer=IntegerWriter()))), |
| | (10, OptionWriter(option=MapWriter(key_writer=IntegerWriter(), value_writer=BinaryWriter()))), |
| | (11, OptionWriter(option=MapWriter(key_writer=IntegerWriter(), value_writer=BinaryWriter()))), |
| | (12, OptionWriter(option=BinaryWriter())), |
| | (13, OptionWriter(option=ListWriter(element_writer=IntegerWriter()))), |
| | (15, OptionWriter(option=IntegerWriter())), |
| | )), |
| | ), |
| | )) |
| |
|
| | assert actual == expected |
| |
|
| |
|
| | def test_resolve_writer_promotion() -> None: |
| | with pytest.raises(ResolveError) as exc_info: |
| | _ = resolve_writer( |
| | record_schema=Schema(NestedField(field_id=1, name="floating", type=DoubleType(), required=True)), |
| | file_schema=Schema(NestedField(field_id=1, name="floating", type=FloatType(), required=True)), |
| | ) |
| |
|
| | assert "Cannot promote double to float" in str(exc_info.value) |
| |
|
| |
|
| | def test_writer_ordering() -> None: |
| | actual = resolve_writer( |
| | record_schema=Schema( |
| | NestedField(field_id=1, name="str", type=StringType(), required=True), |
| | NestedField(field_id=2, name="dbl", type=DoubleType(), required=True), |
| | ), |
| | file_schema=Schema( |
| | NestedField(field_id=2, name="dbl", type=DoubleType(), required=True), |
| | NestedField(field_id=1, name="str", type=StringType(), required=True), |
| | ), |
| | ) |
| |
|
| | expected = StructWriter(((1, DoubleWriter()), (0, StringWriter()))) |
| |
|
| | assert actual == expected |
| |
|
| |
|
| | def test_writer_one_more_field() -> None: |
| | actual = resolve_writer( |
| | record_schema=Schema( |
| | NestedField(field_id=3, name="bool", type=BooleanType(), required=True), |
| | NestedField(field_id=1, name="str", type=StringType(), required=True), |
| | NestedField(field_id=2, name="dbl", type=DoubleType(), required=True), |
| | ), |
| | file_schema=Schema( |
| | NestedField(field_id=2, name="dbl", type=DoubleType(), required=True), |
| | NestedField(field_id=1, name="str", type=StringType(), required=True), |
| | ), |
| | ) |
| |
|
| | expected = StructWriter(((2, DoubleWriter()), (1, StringWriter()))) |
| |
|
| | assert actual == expected |
| |
|
| |
|
| | def test_writer_missing_optional_in_read_schema() -> None: |
| | actual = resolve_writer( |
| | record_schema=Schema(), |
| | file_schema=Schema( |
| | NestedField(field_id=1, name="str", type=StringType(), required=False), |
| | ), |
| | ) |
| |
|
| | expected = StructWriter(field_writers=((None, OptionWriter(option=StringWriter())),)) |
| |
|
| | assert actual == expected |
| |
|