| | import math |
| | import bisect |
| |
|
| | import imgaug |
| | import numpy as np |
| |
|
| | import torch |
| | import torch.distributed as dist |
| | from torch.utils.data import Sampler, ConcatDataset, BatchSampler |
| |
|
| | from concern.config import Configurable, State |
| |
|
| |
|
| | def default_worker_init_fn(worker_id): |
| | np.random.seed(worker_id) |
| | imgaug.seed(worker_id) |
| |
|
| |
|
| | class DataLoader(Configurable, torch.utils.data.DataLoader): |
| | dataset = State() |
| | batch_size = State(default=256) |
| | num_workers = State(default=10) |
| | is_train = State(default=True) |
| | collect_fn = State(default=None) |
| | drop_last = State(default=True) |
| | shuffle = State() |
| |
|
| | def __init__(self, **kwargs): |
| | self.load_all(**kwargs) |
| | if self.collect_fn is None: |
| | self.collect_fn = torch.utils.data.dataloader.default_collate |
| | cmd = kwargs.get('cmd', {}) |
| | self.is_train = cmd['is_train'] |
| | if 'batch_size' in cmd: |
| | self.batch_size = cmd['batch_size'] |
| | if self.shuffle is None: |
| | self.shuffle = self.is_train |
| | self.num_workers = cmd.get('num_workers', self.num_workers) |
| |
|
| | if cmd.get('distributed'): |
| | sampler = DistributedSampler( |
| | self.dataset, shuffle=self.shuffle, |
| | num_replicas=cmd['num_gpus']) |
| | batch_sampler = BatchSampler( |
| | sampler, self.batch_size//cmd['num_gpus'], False) |
| | torch.utils.data.DataLoader.__init__( |
| | self, self.dataset, batch_sampler=batch_sampler, |
| | num_workers=self.num_workers, pin_memory=False, |
| | drop_last=self.drop_last, collate_fn=self.collect_fn, |
| | worker_init_fn=default_worker_init_fn) |
| | else: |
| | torch.utils.data.DataLoader.__init__( |
| | self, self.dataset, |
| | batch_size=self.batch_size, num_workers=self.num_workers, |
| | drop_last=self.drop_last, shuffle=self.shuffle, |
| | pin_memory=True, collate_fn=self.collect_fn, |
| | worker_init_fn=default_worker_init_fn) |
| | self.collect_fn = str(self.collect_fn) |
| |
|
| |
|
| | class SuccessiveRandomSampler(Sampler): |
| | '''Random Sampler that yields sorted data in successive ranges. |
| | Args: |
| | dataset: Dataset used for sampling. |
| | ''' |
| | def __init__(self, dataset): |
| | self.dataset = dataset |
| | self.epoch = 0 |
| |
|
| | def __iter__(self): |
| | if self.shuffle: |
| | |
| | g = torch.Generator() |
| | g.manual_seed(self.epoch) |
| | indices = torch.randperm(len(self.dataset)).tolist() |
| | else: |
| | indices = torch.arange(len(self.dataset)).tolist() |
| |
|
| | |
| | indices += indices[: (self.total_size - len(indices))] |
| | assert len(indices) == self.total_size |
| |
|
| | |
| | offset = self.num_samples * self.rank |
| | indices = indices[offset: offset + self.num_samples] |
| | assert len(indices) == self.num_samples |
| |
|
| | return iter(indices) |
| |
|
| | def __len__(self): |
| | return len(self.dataset) |
| |
|
| | def set_epoch(self, epoch): |
| | self.epoch = epoch |
| |
|
| |
|
| | class DistributedSampler(Sampler): |
| | """Sampler that restricts data loading to a subset of the dataset. |
| | It is especially useful in conjunction with |
| | :class:`torch.nn.parallel.DistributedDataParallel`. In such case, each |
| | process can pass a DistributedSampler instance as a DataLoader sampler, |
| | and load a subset of the original dataset that is exclusive to it. |
| | .. note:: |
| | Dataset is assumed to be of constant size. |
| | Arguments: |
| | dataset: Dataset used for sampling. |
| | num_replicas (optional): Number of processes participating in |
| | distributed training. |
| | rank (optional): Rank of the current process within num_replicas. |
| | """ |
| |
|
| | def __init__(self, dataset, num_replicas=None, rank=None, shuffle=True): |
| | if num_replicas is None: |
| | if not dist.is_available(): |
| | raise RuntimeError( |
| | "Requires distributed package to be available") |
| | num_replicas = dist.get_world_size() |
| | if rank is None: |
| | if not dist.is_available(): |
| | raise RuntimeError( |
| | "Requires distributed package to be available") |
| | rank = dist.get_rank() |
| | self.dataset = dataset |
| | self.num_replicas = num_replicas |
| | self.rank = rank |
| | self.epoch = 0 |
| | self.num_samples = int( |
| | math.ceil(len(self.dataset) * 1.0 / self.num_replicas)) |
| | self.total_size = self.num_samples * self.num_replicas |
| | self.shuffle = shuffle |
| |
|
| | def __iter__(self): |
| | if self.shuffle: |
| | |
| | g = torch.Generator() |
| | g.manual_seed(self.epoch) |
| | indices = torch.randperm(len(self.dataset)).tolist() |
| | else: |
| | indices = torch.arange(len(self.dataset)).tolist() |
| |
|
| | |
| | indices += indices[: (self.total_size - len(indices))] |
| | assert len(indices) == self.total_size |
| |
|
| | |
| | offset = self.num_samples * self.rank |
| | indices = indices[offset: offset + self.num_samples] |
| | assert len(indices) == self.num_samples |
| |
|
| | return iter(indices) |
| |
|
| | def __len__(self): |
| | return self.num_samples |
| |
|
| | def set_epoch(self, epoch): |
| | self.epoch = epoch |
| |
|
| |
|
| | class InfiniteOrderedSampler(Sampler): |
| | def __init__(self, data_source, limit_size): |
| | self.data_source = data_source |
| | self.limit_size = limit_size |
| |
|
| | def __iter__(self): |
| | n = len(self.data_source) |
| |
|
| | def wrapper(): |
| | cnt = 0 |
| | while cnt < self.limit_size: |
| | if cnt % n == 0: |
| | idx = torch.randperm(n).tolist() |
| | yield idx[cnt % n] |
| | cnt += 1 |
| | return wrapper() |
| |
|
| | def __len__(self): |
| | return self.limit_size |
| |
|
| |
|
| | class InfiniteDataLoader(Configurable, torch.utils.data.DataLoader): |
| | dataset = State() |
| | batch_size = State(default=256) |
| | num_workers = State(default=10) |
| | limit_size = State(default=2 ** 31) |
| |
|
| | def __init__(self, **kwargs): |
| | self.load_all(**kwargs) |
| |
|
| | cmd = kwargs['cmd'] |
| | if 'batch_size' in cmd: |
| | self.batch_size = cmd['batch_size'] |
| |
|
| | sampler = InfiniteOrderedSampler(self.dataset, self.limit_size) |
| |
|
| | torch.utils.data.DataLoader.__init__( |
| | self, self.dataset, |
| | batch_size=self.batch_size, num_workers=self.num_workers, |
| | sampler=sampler, worker_init_fn=default_worker_init_fn, |
| | ) |
| |
|
| |
|
| | class RandomSampleSampler(Sampler): |
| | def __init__(self, data_source, weights=None, size=2 ** 31): |
| | self.data_source = data_source |
| | if weights is None: |
| | self.probabilities = np.full(len(data_source), 1 / len(data_source)) |
| | else: |
| | self.probabilities = np.array(weights) / np.sum(weights) |
| | self.cum_prob = np.cumsum(self.probabilities) |
| | self.size = size |
| |
|
| | def __iter__(self): |
| | def wrapper(): |
| | for i in range(self.size): |
| | yield bisect.bisect(self.cum_prob, torch.rand(1)[0], hi=len(self.data_source) - 1) |
| | return wrapper() |
| |
|
| | def __len__(self): |
| | return self.size |
| |
|
| |
|
| | class RandomSampleDataLoader(Configurable, torch.utils.data.DataLoader): |
| | datasets = State() |
| | weights = State() |
| | batch_size = State(default=256) |
| | num_workers = State(default=10) |
| | size = State(default=2 ** 31) |
| |
|
| | def __init__(self, **kwargs): |
| | self.load_all(**kwargs) |
| |
|
| | cmd = kwargs['cmd'] |
| | if 'batch_size' in cmd: |
| | self.batch_size = cmd['batch_size'] |
| |
|
| | probs = [] |
| | for dataset, weight in zip(self.datasets, self.weights): |
| | probs.append(np.full(len(dataset), weight / len(dataset))) |
| |
|
| | dataset = ConcatDataset(self.datasets) |
| | probs = np.concatenate(probs) |
| | assert(len(dataset) == len(probs)) |
| |
|
| | sampler = RandomSampleSampler(dataset, probs, self.size) |
| |
|
| | torch.utils.data.DataLoader.__init__( |
| | self, dataset, |
| | batch_size=self.batch_size, num_workers=self.num_workers, |
| | sampler=sampler, worker_init_fn=default_worker_init_fn, |
| | ) |
| |
|