AlexNet_CNN_Visualisation / dataset_creation.py
makiisthebes's picture
Upload 18 files
4ec6f12 verified
raw
history blame contribute delete
No virus
2.9 kB
# Dataset creation for usage from the videos based on the video preprocessing required.
from torch.utils.data import DataLoader
from torchvision.io import read_image
from torch.utils.data import Dataset
from torchvision.transforms import v2
from torchvision import transforms
from torchvision import datasets
from PIL import Image
import pandas as pd
import idx2numpy, os
import torch
# Dataset creation,
# Loading from a custom dataset
IMAGE_DIMS = 224
normal_transforms = v2.Compose([
v2.Resize(size=(IMAGE_DIMS, IMAGE_DIMS)),
# convert to rgb from greyscale.
# v2.RandomHorizontalFlip(p=0.5),
v2.ToDtype(torch.float32), # , scale=True),
# v2.RandomPerspective(distortion_scale=0.6, p=0.4),
# v2.GaussianBlur(kernel_size=(5, 11), sigma=(0.1, 0.2)),
v2.RandomRotation(degrees=(-15, 15)),
# v2.RandomAffine(degrees=(-15, 15)),
# v2.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]),
transforms.Normalize((0.13066047430038452,), (0.30810782313346863,)),
])
class CustomImageDataset(Dataset):
"""
This class must inherit from the torch.utils.data.Dataset class.
And contina functions __init__, __len__, and __getitem__.
"""
def __init__(self, annotations_file, img_dir, transform=None, target_transform=None):
self.img_labels = pd.read_csv(annotations_file)
self.img_dir = img_dir
self.transform = transform
self.target_transform = target_transform
def __len__(self):
return len(self.img_labels)
def __getitem__(self, idx):
"""Get the image and label at the index idx."""
img_path = os.path.join(self.img_dir, self.img_labels.iloc[idx, 0])
Image.open(img_path).convert("RGB").save(img_path)
image = read_image(img_path)
label = self.img_labels.iloc[idx, 1]
if self.transform:
image = self.transform(image)
if self.target_transform:
label = self.target_transform(label)
return image, label
train_data = CustomImageDataset("./dataset/root/labels.csv", "./dataset/root/train/", transform=normal_transforms)
# Create a DataLoader, so we can iterate through the dataset in batches.
#train_loader = DataLoader(train_data, batch_size=64, shuffle=True, )
# Testing the dataloader.
# for i, (images, labels) in enumerate(train_loader):
# print(i, images.shape, labels.shape)
train_size = int(0.8 * len(train_data))
test_size = len(train_data) - train_size
train_dataset, test_dataset = torch.utils.data.random_split(train_data, [train_size, test_size])
# Create DataLoader for train and test sets
train_loader = DataLoader(train_dataset, batch_size=32, shuffle=True)
test_loader = DataLoader(test_dataset, batch_size=32, shuffle=False)
print("Data loader and Test Loaders are ready to be used.")
# Create first stage labels,
# movement stage labels
# final stage labels.