| | from PIL import Image |
| | import numpy as np |
| | import torch |
| |
|
| |
|
| | def tensor_to_pil(img_tensor, batch_index=0): |
| | |
| | img_tensor = img_tensor[batch_index].unsqueeze(0) |
| | i = 255. * img_tensor.cpu().numpy() |
| | img = Image.fromarray(np.clip(i, 0, 255).astype(np.uint8).squeeze()) |
| | return img |
| |
|
| |
|
| | def batch_tensor_to_pil(img_tensor): |
| | |
| | return [tensor_to_pil(img_tensor, i) for i in range(img_tensor.shape[0])] |
| |
|
| |
|
| | def pil_to_tensor(image): |
| | |
| | image = np.array(image).astype(np.float32) / 255.0 |
| | image = torch.from_numpy(image).unsqueeze(0) |
| | if len(image.shape) == 3: |
| | image = image.unsqueeze(-1) |
| | return image |
| |
|
| |
|
| | def batched_pil_to_tensor(images): |
| | |
| | return torch.cat([pil_to_tensor(image) for image in images], dim=0) |
| |
|