|
import os |
|
import numpy as np |
|
import cv2 |
|
import albumentations |
|
from PIL import Image |
|
from torch.utils.data import Dataset |
|
|
|
from taming.data.sflckr import SegmentationBase |
|
|
|
|
|
class Examples(SegmentationBase): |
|
def __init__(self, size=256, random_crop=False, interpolation="bicubic"): |
|
super().__init__(data_csv="data/ade20k_examples.txt", |
|
data_root="data/ade20k_images", |
|
segmentation_root="data/ade20k_segmentations", |
|
size=size, random_crop=random_crop, |
|
interpolation=interpolation, |
|
n_labels=151, shift_segmentation=False) |
|
|
|
|
|
|
|
class ADE20kBase(Dataset): |
|
def __init__(self, config=None, size=None, random_crop=False, interpolation="bicubic", crop_size=None): |
|
self.split = self.get_split() |
|
self.n_labels = 151 |
|
self.data_csv = {"train": "data/ade20k_train.txt", |
|
"validation": "data/ade20k_test.txt"}[self.split] |
|
self.data_root = "data/ade20k_root" |
|
with open(os.path.join(self.data_root, "sceneCategories.txt"), "r") as f: |
|
self.scene_categories = f.read().splitlines() |
|
self.scene_categories = dict(line.split() for line in self.scene_categories) |
|
with open(self.data_csv, "r") as f: |
|
self.image_paths = f.read().splitlines() |
|
self._length = len(self.image_paths) |
|
self.labels = { |
|
"relative_file_path_": [l for l in self.image_paths], |
|
"file_path_": [os.path.join(self.data_root, "images", l) |
|
for l in self.image_paths], |
|
"relative_segmentation_path_": [l.replace(".jpg", ".png") |
|
for l in self.image_paths], |
|
"segmentation_path_": [os.path.join(self.data_root, "annotations", |
|
l.replace(".jpg", ".png")) |
|
for l in self.image_paths], |
|
"scene_category": [self.scene_categories[l.split("/")[1].replace(".jpg", "")] |
|
for l in self.image_paths], |
|
} |
|
|
|
size = None if size is not None and size<=0 else size |
|
self.size = size |
|
if crop_size is None: |
|
self.crop_size = size if size is not None else None |
|
else: |
|
self.crop_size = crop_size |
|
if self.size is not None: |
|
self.interpolation = interpolation |
|
self.interpolation = { |
|
"nearest": cv2.INTER_NEAREST, |
|
"bilinear": cv2.INTER_LINEAR, |
|
"bicubic": cv2.INTER_CUBIC, |
|
"area": cv2.INTER_AREA, |
|
"lanczos": cv2.INTER_LANCZOS4}[self.interpolation] |
|
self.image_rescaler = albumentations.SmallestMaxSize(max_size=self.size, |
|
interpolation=self.interpolation) |
|
self.segmentation_rescaler = albumentations.SmallestMaxSize(max_size=self.size, |
|
interpolation=cv2.INTER_NEAREST) |
|
|
|
if crop_size is not None: |
|
self.center_crop = not random_crop |
|
if self.center_crop: |
|
self.cropper = albumentations.CenterCrop(height=self.crop_size, width=self.crop_size) |
|
else: |
|
self.cropper = albumentations.RandomCrop(height=self.crop_size, width=self.crop_size) |
|
self.preprocessor = self.cropper |
|
|
|
def __len__(self): |
|
return self._length |
|
|
|
def __getitem__(self, i): |
|
example = dict((k, self.labels[k][i]) for k in self.labels) |
|
image = Image.open(example["file_path_"]) |
|
if not image.mode == "RGB": |
|
image = image.convert("RGB") |
|
image = np.array(image).astype(np.uint8) |
|
if self.size is not None: |
|
image = self.image_rescaler(image=image)["image"] |
|
segmentation = Image.open(example["segmentation_path_"]) |
|
segmentation = np.array(segmentation).astype(np.uint8) |
|
if self.size is not None: |
|
segmentation = self.segmentation_rescaler(image=segmentation)["image"] |
|
if self.size is not None: |
|
processed = self.preprocessor(image=image, mask=segmentation) |
|
else: |
|
processed = {"image": image, "mask": segmentation} |
|
example["image"] = (processed["image"]/127.5 - 1.0).astype(np.float32) |
|
segmentation = processed["mask"] |
|
onehot = np.eye(self.n_labels)[segmentation] |
|
example["segmentation"] = onehot |
|
return example |
|
|
|
|
|
class ADE20kTrain(ADE20kBase): |
|
|
|
def __init__(self, config=None, size=None, random_crop=True, interpolation="bicubic", crop_size=None): |
|
super().__init__(config=config, size=size, random_crop=random_crop, |
|
interpolation=interpolation, crop_size=crop_size) |
|
|
|
def get_split(self): |
|
return "train" |
|
|
|
|
|
class ADE20kValidation(ADE20kBase): |
|
def get_split(self): |
|
return "validation" |
|
|
|
|
|
if __name__ == "__main__": |
|
dset = ADE20kValidation() |
|
ex = dset[0] |
|
for k in ["image", "scene_category", "segmentation"]: |
|
print(type(ex[k])) |
|
try: |
|
print(ex[k].shape) |
|
except: |
|
print(ex[k]) |
|
|