| |
|
|
| import struct |
| from functools import partial |
| from io import BufferedIOBase |
| from typing import Any, Dict, Iterator, List, NamedTuple, Optional, Tuple, Union, cast |
|
|
| import numpy as np |
|
|
|
|
| try: |
| from math import prod |
| except ImportError: |
| import operator |
| from functools import reduce |
|
|
| def prod(xs): |
| return reduce(operator.mul, xs, 1) |
|
|
|
|
| U = Union[bytes, bytearray, str] |
| TFRecordFeatureSpec = Tuple[Tuple[int, ...], Union[str, np.dtype]] |
| TFRecordExampleSpec = Dict[str, TFRecordFeatureSpec] |
|
|
| |
| |
| |
| |
| TFRecordBinaryData = Union[str, List[str], List[List[str]], List[List[List[Any]]]] |
| TFRecordExampleFeature = Union[np.ndarray, List[np.ndarray], TFRecordBinaryData] |
| TFRecordExample = Dict[str, TFRecordExampleFeature] |
|
|
|
|
| class SequenceExampleSpec(NamedTuple): |
| context: TFRecordExampleSpec |
| feature_lists: TFRecordExampleSpec |
|
|
|
|
| def iterate_tfrecord_file(data: BufferedIOBase) -> Iterator[memoryview]: |
| length_bytes = bytearray(8) |
| crc_bytes = bytearray(4) |
| data_bytes = bytearray(1024) |
|
|
| while True: |
| bytes_read = data.readinto(length_bytes) |
| if bytes_read == 0: |
| break |
| elif bytes_read != 8: |
| raise RuntimeError("Invalid tfrecord file: failed to read the record size.") |
| if data.readinto(crc_bytes) != 4: |
| raise RuntimeError("Invalid tfrecord file: failed to read the start token.") |
| (length,) = struct.unpack("<Q", length_bytes) |
| if length > len(data_bytes): |
| data_bytes = data_bytes.zfill(int(length * 1.5)) |
| data_bytes_view = memoryview(data_bytes)[:length] |
| if data.readinto(data_bytes_view) != length: |
| raise RuntimeError("Invalid tfrecord file: failed to read the record.") |
| if data.readinto(crc_bytes) != 4: |
| raise RuntimeError("Invalid tfrecord file: failed to read the end token.") |
|
|
| |
| yield data_bytes_view |
|
|
|
|
| def process_feature(feature) -> np.ndarray: |
| |
| |
| field = feature.ListFields()[0] |
| inferred_typename, value = field[0].name, field[1].value |
| if inferred_typename == "bytes_list": |
| pass |
| elif inferred_typename == "float_list": |
| value = np.array(value, dtype=np.float32) |
| elif inferred_typename == "int64_list": |
| value = np.array(value, dtype=np.int64) |
| return value |
|
|
|
|
| def _reshape_list(value, shape): |
| |
| flat_list = [] |
|
|
| def flatten(value): |
| if isinstance(value, (str, bytes)): |
| flat_list.append(value) |
| else: |
| for x in value: |
| flatten(x) |
|
|
| flatten(value) |
|
|
| |
| common_divisor = prod(x for x in shape if x != -1) |
| if sum(1 for x in shape if x == -1) > 1: |
| raise RuntimeError("Shape can contain at most one dynamic dimension (-1).") |
| if len(flat_list) % max(common_divisor, 1) != 0: |
| raise RuntimeError(f"Cannot reshape {len(flat_list)} values into shape {shape}") |
| shape = [x if x != -1 else (len(flat_list) // common_divisor) for x in shape] |
|
|
| |
| def _reshape(value, shape): |
| if len(shape) == 0: |
| assert len(value) == 1 |
| return value[0] |
| elif len(shape) == 1: |
| assert len(value) == shape[0] |
| return value |
| dim_size = len(value) // shape[0] |
| return [_reshape(value[i * dim_size : (i + 1) * dim_size], shape[1:]) for i in range(dim_size)] |
|
|
| return _reshape(flat_list, shape) |
|
|
|
|
| def _apply_feature_spec(value, feature_spec): |
| if isinstance(value, np.ndarray): |
| if feature_spec is not None: |
| shape, dtype = feature_spec |
| if isinstance(dtype, (str, np.dtype)): |
| if shape: |
| value = value.reshape(shape) |
| value = value.astype(dtype) |
| elif shape: |
| |
| value = _reshape_list(value, shape) |
| return value |
|
|
|
|
| def _parse_tfrecord_features(features, spec: Optional[TFRecordExampleSpec]) -> Dict[str, np.ndarray]: |
| result = {} |
| features = features.feature |
| for key in features.keys(): |
| if spec is not None and key not in spec: |
| continue |
| feature_spec = None if spec is None else spec[key] |
| feature = features[key] |
| result[key] = _apply_feature_spec(process_feature(feature), feature_spec) |
| return result |
|
|
|
|
| def parse_tfrecord_sequence_example(example, spec: Optional[TFRecordExampleSpec]) -> TFRecordExample: |
| |
| result = cast(TFRecordExample, _parse_tfrecord_features(example.context, spec)) |
|
|
| |
| feature_lists_keys = None if spec is None else set(spec.keys()) - set(result.keys()) |
| features = example.feature_lists.feature_list |
| for key in features.keys(): |
| if feature_lists_keys is not None and key not in feature_lists_keys: |
| continue |
| feature_spec = None if spec is None else spec[key] |
| feature = features[key].feature |
| if key in result: |
| raise RuntimeError( |
| "TFRecord example's key {key} is contained in both the context and feature lists. This is not supported." |
| ) |
|
|
| value: Union[np.ndarray, List[Any]] = list(map(partial(process_feature), feature)) |
|
|
| |
| if feature_spec is not None and isinstance(feature_spec[1], (str, np.dtype)): |
| value = np.stack(cast(List[np.ndarray], value), 0) |
| value = _apply_feature_spec(value, feature_spec) |
| result[key] = value |
| if spec is not None and len(result.keys()) != len(spec.keys()): |
| raise RuntimeError(f"Example is missing some required keys: {sorted(result.keys())} != {sorted(spec.keys())}") |
| return result |
|
|