| import os |
| import json |
| import jsonlines |
| import torch |
| import math |
| import random |
| import cv2 |
|
|
| from tqdm import tqdm |
| from collections import OrderedDict |
|
|
| from PIL import Image |
| from PIL import ImageFile |
| ImageFile.LOAD_TRUNCATED_IMAGES = True |
|
|
| import numpy as np |
| import subprocess |
| from torch.utils.data import Dataset, DataLoader |
| from torchvision import transforms |
| from torchvision.transforms.functional import InterpolationMode |
| from torchvision.transforms import functional as F |
|
|
|
|
| class ImageTextDataset(Dataset): |
| """ |
| Usage: |
| The dataset class for image-text pairs, used for image generation training |
| It supports multi-aspect ratio training |
| params: |
| anno_file: The annotation file list |
| add_normalize: whether to normalize the input image pixel to [-1, 1], default: True |
| ratios: The aspect ratios during training, format: width / height |
| sizes: The resoultion of training images, format: (width, height) |
| """ |
| def __init__( |
| self, anno_file, add_normalize=True, |
| ratios=[1/1, 3/5, 5/3], |
| sizes=[(1024, 1024), (768, 1280), (1280, 768)], |
| crop_mode='random', p_random_ratio=0.0, |
| ): |
| |
| super().__init__() |
| |
| self.image_annos = [] |
| if not isinstance(anno_file, list): |
| anno_file = [anno_file] |
|
|
| for anno_file_ in anno_file: |
| print(f"Load image annotation files from {anno_file_}") |
| with jsonlines.open(anno_file_, 'r') as reader: |
| for item in reader: |
| self.image_annos.append(item) |
|
|
| print(f"Totally Remained {len(self.image_annos)} images") |
|
|
| transform_list = [ |
| transforms.ToTensor(), |
| ] |
| |
| if add_normalize: |
| transform_list.append(transforms.Normalize(mean=(0.5, 0.5, 0.5), std=(0.5, 0.5, 0.5))) |
| |
| self.transform = transforms.Compose(transform_list) |
|
|
| print(f"Transform List is {transform_list}") |
|
|
| assert crop_mode in ['center', 'random'] |
| self.crop_mode = crop_mode |
| self.ratios = ratios |
| self.sizes = sizes |
| self.p_random_ratio = p_random_ratio |
|
|
| def get_closest_size(self, x): |
| if self.p_random_ratio > 0 and np.random.rand() < self.p_random_ratio: |
| best_size_idx = np.random.randint(len(self.ratios)) |
| else: |
| w, h = x.width, x.height |
| best_size_idx = np.argmin([abs(w/h-r) for r in self.ratios]) |
| return self.sizes[best_size_idx] |
|
|
| def get_resize_size(self, orig_size, tgt_size): |
| if (tgt_size[1]/tgt_size[0] - 1) * (orig_size[1]/orig_size[0] - 1) >= 0: |
| alt_min = int(math.ceil(max(tgt_size)*min(orig_size)/max(orig_size))) |
| resize_size = max(alt_min, min(tgt_size)) |
| else: |
| alt_max = int(math.ceil(min(tgt_size)*max(orig_size)/min(orig_size))) |
| resize_size = max(alt_max, max(tgt_size)) |
| return resize_size |
|
|
| def __len__(self): |
| return len(self.image_annos) |
|
|
| def __getitem__(self, index): |
| image_anno = self.image_annos[index] |
|
|
| try: |
| img = Image.open(image_anno['image']).convert("RGB") |
| text = image_anno['text'] |
|
|
| assert isinstance(text, str), "Text should be str" |
|
|
| size = self.get_closest_size(img) |
| resize_size = self.get_resize_size((img.width, img.height), size) |
|
|
| img = transforms.functional.resize(img, resize_size, interpolation=transforms.InterpolationMode.BICUBIC, antialias=True) |
| |
| if self.crop_mode == 'center': |
| img = transforms.functional.center_crop(img, (size[1], size[0])) |
| elif self.crop_mode == 'random': |
| img = transforms.RandomCrop((size[1], size[0]))(img) |
| else: |
| img = transforms.functional.center_crop(img, (size[1], size[0])) |
|
|
| image_tensor = self.transform(img) |
| |
| return { |
| "video": image_tensor, |
| "text" : text, |
| "identifier": 'image', |
| } |
| |
| except Exception as e: |
| print(f'Load Image Error with {e}') |
| return self.__getitem__(random.randint(0, self.__len__() - 1)) |
|
|
|
|
| class LengthGroupedVideoTextDataset(Dataset): |
| """ |
| Usage: |
| The dataset class for video-text pairs, used for video generation training |
| It groups the video with the same frames together |
| Now only supporting fixed resolution during training |
| params: |
| anno_file: The annotation file list |
| max_frames: The maximum temporal lengths (This is the vae latent temporal length) 16 => (16 - 1) * 8 + 1 = 121 frames |
| load_vae_latent: Loading the pre-extracted vae latents during training, we recommend to extract the latents in advance |
| to reduce the time cost per batch |
| load_text_fea: Loading the pre-extracted text features during training, we recommend to extract the prompt textual features |
| in advance, since the T5 encoder will cost many GPU memories |
| """ |
| |
| def __init__(self, anno_file, max_frames=16, resolution='384p', load_vae_latent=True, load_text_fea=True): |
| super().__init__() |
|
|
| self.video_annos = [] |
| self.max_frames = max_frames |
| self.load_vae_latent = load_vae_latent |
| self.load_text_fea = load_text_fea |
| self.resolution = resolution |
|
|
| assert load_vae_latent, "Now only support loading vae latents, we will support to directly load video frames in the future" |
|
|
| if not isinstance(anno_file, list): |
| anno_file = [anno_file] |
|
|
| for anno_file_ in anno_file: |
| with jsonlines.open(anno_file_, 'r') as reader: |
| for item in tqdm(reader): |
| self.video_annos.append(item) |
| |
| print(f"Totally Remained {len(self.video_annos)} videos") |
|
|
| def __len__(self): |
| return len(self.video_annos) |
|
|
| def __getitem__(self, index): |
| try: |
| video_anno = self.video_annos[index] |
| text = video_anno['text'] |
| latent_path = video_anno['latent'] |
| latent = torch.load(latent_path, map_location='cpu', weights_only=False) |
|
|
| |
| if self.resolution == '384p': |
| assert latent.shape[-1] == 640 // 8 |
| assert latent.shape[-2] == 384 // 8 |
| else: |
| assert self.resolution == '768p' |
| assert latent.shape[-1] == 1280 // 8 |
| assert latent.shape[-2] == 768 // 8 |
|
|
| cur_temp = latent.shape[2] |
| cur_temp = min(cur_temp, self.max_frames) |
|
|
| video_latent = latent[:,:,:cur_temp].float() |
| assert video_latent.shape[1] == 16 |
|
|
| if self.load_text_fea: |
| text_fea_path = video_anno['text_fea'] |
| text_fea = torch.load(text_fea_path, map_location='cpu') |
| return { |
| 'video': video_latent, |
| 'prompt_embed': text_fea['prompt_embed'], |
| 'prompt_attention_mask': text_fea['prompt_attention_mask'], |
| 'pooled_prompt_embed': text_fea['pooled_prompt_embed'], |
| "identifier": 'video', |
| } |
|
|
| else: |
| return { |
| 'video': video_latent, |
| 'text': text, |
| "identifier": 'video', |
| } |
|
|
| except Exception as e: |
| print(f'Load Video Error with {e}') |
| return self.__getitem__(random.randint(0, self.__len__() - 1)) |
|
|
|
|
| class VideoFrameProcessor: |
| |
| def __init__(self, resolution=256, num_frames=24, add_normalize=True, sample_fps=24): |
| |
| image_size = resolution |
|
|
| transform_list = [ |
| transforms.Resize(image_size, interpolation=InterpolationMode.BICUBIC, antialias=True), |
| transforms.CenterCrop(image_size), |
| ] |
| |
| if add_normalize: |
| transform_list.append(transforms.Normalize(mean=(0.5, 0.5, 0.5), std=(0.5, 0.5, 0.5))) |
|
|
| print(f"Transform List is {transform_list}") |
| self.num_frames = num_frames |
| self.transform = transforms.Compose(transform_list) |
| self.sample_fps = sample_fps |
|
|
| def __call__(self, video_path): |
| try: |
| video_capture = cv2.VideoCapture(video_path) |
| fps = video_capture.get(cv2.CAP_PROP_FPS) |
| frames = [] |
|
|
| while True: |
| flag, frame = video_capture.read() |
| if not flag: |
| break |
|
|
| frame = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB) |
| frame = torch.from_numpy(frame) |
| frame = frame.permute(2, 0, 1) |
| frames.append(frame) |
|
|
| video_capture.release() |
| sample_fps = self.sample_fps |
| interval = max(int(fps / sample_fps), 1) |
| frames = frames[::interval] |
|
|
| if len(frames) < self.num_frames: |
| num_frame_to_pack = self.num_frames - len(frames) |
| recurrent_num = num_frame_to_pack // len(frames) |
| frames = frames + recurrent_num * frames + frames[:(num_frame_to_pack % len(frames))] |
| assert len(frames) >= self.num_frames, f'{len(frames)}' |
|
|
| start_indexs = list(range(0, max(0, len(frames) - self.num_frames + 1))) |
| start_index = random.choice(start_indexs) |
|
|
| filtered_frames = frames[start_index : start_index+self.num_frames] |
| assert len(filtered_frames) == self.num_frames, f"The sampled frames should equals to {self.num_frames}" |
|
|
| filtered_frames = torch.stack(filtered_frames).float() / 255 |
| filtered_frames = self.transform(filtered_frames) |
| filtered_frames = filtered_frames.permute(1, 0, 2, 3) |
|
|
| return filtered_frames, None |
| |
| except Exception as e: |
| print(f"Load video: {video_path} Error, Exception {e}") |
| return None, None |
|
|
|
|
| class VideoDataset(Dataset): |
| def __init__(self, anno_file, resolution=256, max_frames=6, add_normalize=True): |
| super().__init__() |
|
|
| self.video_annos = [] |
| self.max_frames = max_frames |
|
|
| if not isinstance(anno_file, list): |
| anno_file = [anno_file] |
|
|
| print(f"The training video clip frame number is {max_frames} ") |
|
|
| for anno_file_ in anno_file: |
| print(f"Load annotation file from {anno_file_}") |
|
|
| with jsonlines.open(anno_file_, 'r') as reader: |
| for item in tqdm(reader): |
| self.video_annos.append(item) |
| |
| print(f"Totally Remained {len(self.video_annos)} videos") |
| |
| self.video_processor = VideoFrameProcessor(resolution, max_frames, add_normalize) |
|
|
| def __len__(self): |
| return len(self.video_annos) |
|
|
| def __getitem__(self, index): |
| video_anno = self.video_annos[index] |
| video_path = video_anno['video'] |
|
|
| try: |
| video_tensors, video_frames = self.video_processor(video_path) |
|
|
| assert video_tensors.shape[1] == self.max_frames |
| |
| return { |
| "video": video_tensors, |
| "identifier": 'video', |
| } |
|
|
| except Exception as e: |
| print('Loading Video Error with {e}') |
| return self.__getitem__(random.randint(0, self.__len__() - 1)) |
|
|
|
|
| class ImageDataset(Dataset): |
| def __init__(self, anno_file, resolution=256, max_frames=8, add_normalize=True): |
| super().__init__() |
|
|
| self.image_annos = [] |
| self.max_frames = max_frames |
| image_paths = [] |
|
|
| if not isinstance(anno_file, list): |
| anno_file = [anno_file] |
|
|
| for anno_file_ in anno_file: |
| print(f"Load annotation file from {anno_file_}") |
| with jsonlines.open(anno_file_, 'r') as reader: |
| for item in tqdm(reader): |
| image_paths.append(item['image']) |
| |
| print(f"Totally Remained {len(image_paths)} images") |
|
|
| |
| for idx in range(0, len(image_paths), self.max_frames): |
| image_path_shard = image_paths[idx : idx + self.max_frames] |
| if len(image_path_shard) < self.max_frames: |
| image_path_shard = image_path_shard + image_paths[:self.max_frames - len(image_path_shard)] |
| assert len(image_path_shard) == self.max_frames |
| self.image_annos.append(image_path_shard) |
|
|
| image_size = resolution |
| transform_list = [ |
| transforms.Resize(image_size, interpolation=InterpolationMode.BICUBIC, antialias=True), |
| transforms.CenterCrop(image_size), |
| transforms.ToTensor(), |
| ] |
| if add_normalize: |
| transform_list.append(transforms.Normalize(mean=(0.5, 0.5, 0.5), std=(0.5, 0.5, 0.5))) |
|
|
| print(f"Transform List is {transform_list}") |
| self.transform = transforms.Compose(transform_list) |
|
|
| def __len__(self): |
| return len(self.image_annos) |
|
|
| def __getitem__(self, index): |
| image_paths = self.image_annos[index] |
|
|
| try: |
| packed_pil_frames = [Image.open(image_path).convert("RGB") for image_path in image_paths] |
| filtered_frames = [self.transform(frame) for frame in packed_pil_frames] |
| filtered_frames = torch.stack(filtered_frames) |
| filtered_frames = filtered_frames.permute(1, 0, 2, 3) |
| |
| return { |
| "video": filtered_frames, |
| "identifier": 'image', |
| } |
|
|
| except Exception as e: |
| print(f'Load Images Error with {e}') |
| return self.__getitem__(random.randint(0, self.__len__() - 1)) |