From 551f9b0e21c0f889cebf205129e685eac70057ab Mon Sep 17 00:00:00 2001 From: CodingTil <36734749+CodingTil@users.noreply.github.com> Date: Mon, 30 Oct 2023 12:43:41 +0100 Subject: [PATCH 1/2] WIP on fusion --- eiuie/fusion_model.py | 294 +++++++++++++++++++++++++++++++++++++++++ eiuie/main.py | 10 +- eiuie/pixel_dataset.py | 36 +++++ pyproject.toml | 1 + 4 files changed, 339 insertions(+), 2 deletions(-) create mode 100644 eiuie/fusion_model.py create mode 100644 eiuie/pixel_dataset.py diff --git a/eiuie/fusion_model.py b/eiuie/fusion_model.py new file mode 100644 index 0000000..a206679 --- /dev/null +++ b/eiuie/fusion_model.py @@ -0,0 +1,294 @@ +from typing import Optional +import os + +import numpy as np +import torch +import torch.nn as nn +import torch.optim as optim +import torch.cuda as cuda +from torch.utils.data import Dataset, DataLoader, random_split + +import base_model as bm +import unsharp_masking as um +import retinex as rtx +import homomorphic_filtering as hf +import pixel_dataset as pxds + + +CHECKPOINT_DIRECTORY = "data/checkpoints" + + +class FusionNet(nn.Module): + def __init__(self, dropout_rate=0.5): + super(FusionNet, self).__init__() + + self.model = nn.Sequential( + nn.Linear(12, 12), + nn.ReLU(), + nn.Dropout(dropout_rate), + nn.Linear(12, 9), + nn.ReLU(), + nn.Dropout(dropout_rate), + nn.Linear(9, 6), + nn.ReLU(), + nn.Dropout(dropout_rate), + nn.Linear(6, 3), + nn.ReLU(), + nn.Linear(3, 3), + nn.Sigmoid(), + ) + + def forward(self, x): + return self.model(x) + + +class EarlyStopping: + def __init__( + self, patience=5, verbose=False, delta=0, path="checkpoint.pt", trace_func=print + ): + """ + Args: + patience (int): How long to wait after last time validation loss improved. + Default: 5 + verbose (bool): If True, prints a message for each validation loss improvement. + Default: False + delta (float): Minimum change in the monitored quantity to qualify as an improvement. + Default: 0 + path (str): Path for the checkpoint to be saved to. + Default: 'checkpoint.pt' + trace_func (function): trace print function. + Default: print + """ + self.patience = patience + self.verbose = verbose + self.counter = 0 + self.best_score = None + self.early_stop = False + self.val_loss_min = np.Inf + self.delta = delta + self.path = path + self.trace_func = trace_func + + def __call__(self, val_loss, model): + score = -val_loss + + if self.best_score is None: + self.best_score = score + self.save_checkpoint(val_loss, model) + elif score < self.best_score + self.delta: + self.counter += 1 + self.trace_func( + f"EarlyStopping counter: {self.counter} out of {self.patience}" + ) + if self.counter >= self.patience: + self.early_stop = True + else: + self.best_score = score + self.save_checkpoint(val_loss, model) + self.counter = 0 + + def save_checkpoint(self, val_loss, model): + """Saves model when validation loss decreases.""" + if self.verbose: + self.trace_func( + f"Validation loss decreased ({self.val_loss_min:.6f} --> {val_loss:.6f}). Saving model ..." + ) + torch.save(model.state_dict(), self.path) + self.val_loss_min = val_loss + + +class FusionModel(bm.BaseModel): + """FusionModel""" + + unsharp_masking: um.UnsharpMasking + homomorphic_filtering: hf.HomomorphicFiltering + retinex: rtx.Retinex + device: torch.device + net: FusionNet + optimizer: optim.Optimizer + criterion: nn.Module + start_epoch: int = 0 + + def __init__( + self, + checkpoint_path: Optional[str] = None, + *, + unsharp_masking: um.UnsharpMasking = um.UnsharpMasking(), + homomorphic_filtering: hf.HomomorphicFiltering = hf.HomomorphicFiltering(), + retinex: rtx.Retinex = rtx.Retinex(), + ): + """ + Parameters + ---------- + checkpoint + Checkpoint for loading model for inference / resume training. + """ + self.unsharp_masking = unsharp_masking + self.homomorphic_filtering = homomorphic_filtering + self.retinex = retinex + + # Check for GPU availability + self.device = torch.device("cuda" if cuda.is_available() else "cpu") + + # Neural Network Model + self.net = FusionNet() + self.optimizer = optim.Adam(self.net.parameters()) + self.criterion = nn.MSELoss() # assuming regression task + self.start_epoch = 0 + + if checkpoint_path: + self.load_checkpoint(checkpoint_path) + else: + latest_checkpoint = self._get_latest_checkpoint() + if latest_checkpoint: + self.load_checkpoint(latest_checkpoint) + + def save_checkpoint(self, epoch: int, checkpoint_path: str): + torch.save( + { + "epoch": epoch, + "model_state_dict": self.net.state_dict(), + "optimizer_state_dict": self.optimizer.state_dict(), + }, + f"{CHECKPOINT_DIRECTORY}/{checkpoint_path}", + ) + + def load_checkpoint(self, checkpoint_path: str): + checkpoint = torch.load(f"{CHECKPOINT_DIRECTORY}/{checkpoint_path}") + self.net.load_state_dict(checkpoint["model_state_dict"]) + self.optimizer.load_state_dict(checkpoint["optimizer_state_dict"]) + self.start_epoch = checkpoint["epoch"] + + def _get_latest_checkpoint(self) -> Optional[str]: + """ + Returns the path to the latest checkpoint from the CHECKPOINT_DIRECTORY. + + Returns + ------- + Optional[str] + Path to the latest checkpoint file or None if no checkpoint found. + """ + checkpoint_files = [ + f for f in os.listdir(CHECKPOINT_DIRECTORY) if "checkpoint_epoch_" in f + ] + if not checkpoint_files: + return None + + # Sort based on epoch number + checkpoint_files.sort(key=lambda x: int(x.split("_")[-1].split(".")[0])) + + return checkpoint_files[-1] # Return the latest checkpoint + + @property + def name(self) -> str: + """ + Name of the model. + + Returns + ------- + str + Name of the model. + """ + return "fusion_model" + + def process_image(self, image: np.ndarray) -> np.ndarray: + """ + Process image using the model. + + Parameters + ---------- + image : np.ndarray + Image to be processed. + + Returns + ------- + np.ndarray + Processed image. + """ + original = bm.BGR2HSI(image) + um_imagge = self.unsharp_masking.process_image(image) + um_image = bm.BGR2HSI(um_imagge) + hf_image = self.homomorphic_filtering.process_image(image) + hf_image = bm.BGR2HSI(hf_image) + rtx_image = self.retinex.process_image(image) + rtx_image = bm.BGR2HSI(rtx_image) + + dimensions = image.shape + assert dimensions == um_imagge.shape == hf_image.shape == rtx_image.shape + + # Use numpy functions for efficient concatenation + # Reshape each processed image into (-1, 3), essentially unrolling them + original = original.reshape(-1, 3) + um_image = um_image.reshape(-1, 3) + hf_image = hf_image.reshape(-1, 3) + rtx_image = rtx_image.reshape(-1, 3) + + # Concatenate them along the horizontal axis (axis=1) + all_inputs = np.hstack([original, um_image, hf_image, rtx_image]) + + # Convert to tensor and move to device + all_inputs = torch.tensor(all_inputs, dtype=torch.float32).to(self.device) + + # Model inference + outputs = self.net(all_inputs).cpu().detach().numpy() + + # Reshape outputs back to the original image shape + fused_image = outputs.reshape(dimensions[0], dimensions[1], 3) + fused_image = bm.HSI2BGR(fused_image) + + return fused_image + + def train_model( + self, + dataset: Dataset = pxds.PixelDataset(), + total_epochs=100, + patience=5, + train_ratio=0.8, + ): + # Splitting dataset into training and validation subsets + train_size = int(train_ratio * len(dataset)) + val_size = len(dataset) - train_size + train_dataset, val_dataset = random_split(dataset, [train_size, val_size]) + + train_loader = DataLoader(train_dataset, batch_size=32, shuffle=True) + val_loader = DataLoader(val_dataset, batch_size=32) + + early_stopping = EarlyStopping( + patience=patience, + verbose=True, + path=f"{CHECKPOINT_DIRECTORY}/best_model.pt", + ) + + self.net.train() + for epoch in range(self.start_epoch, total_epochs): + for inputs, targets in train_loader: + inputs, targets = inputs.to(self.device), targets.to(self.device) + self.optimizer.zero_grad() + outputs = self.net(inputs) + loss = self.criterion(outputs, targets) + loss.backward() + self.optimizer.step() + # After training, check validation loss + val_loss = self.validate(val_loader) + + early_stopping(val_loss, self.net) + + if early_stopping.early_stop: + print("Early stopping") + break + + # Save checkpoint after every epoch + self.save_checkpoint(epoch, f"checkpoint_epoch_{epoch}.pth") + + def validate(self, val_loader): + self.net.eval() + total_val_loss = 0 + with torch.no_grad(): + for inputs, targets in val_loader: + inputs, targets = inputs.to(self.device), targets.to(self.device) + outputs = self.net(inputs) + loss = self.criterion(outputs, targets) + total_val_loss += loss.item() + + average_val_loss = total_val_loss / len(val_loader) + return average_val_loss diff --git a/eiuie/main.py b/eiuie/main.py index 3f4c283..c48374f 100644 --- a/eiuie/main.py +++ b/eiuie/main.py @@ -7,6 +7,7 @@ import unsharp_masking import retinex import homomorphic_filtering +import fusion_model def main(): @@ -15,7 +16,7 @@ def main(): parser.add_argument( "command", type=str, - choices=["single", "batch_process"], + choices=["single", "batch_process", "train"], help="Command to run", ) @@ -24,7 +25,7 @@ def main(): "--method", type=str, default="unsharp_masking", - choices=["unsharp_masking", "retinex", "homomorphic_filtering"], + choices=["unsharp_masking", "retinex", "homomorphic_filtering", "fusion_model"], help="Filter method to use", ) @@ -45,6 +46,8 @@ def main(): method = retinex.Retinex() case "homomorphic_filtering": method = homomorphic_filtering.HomomorphicFiltering() + case "fusion_model": + method = fusion_model.FusionModel() case _: raise ValueError(f"Unknown method: {args.method}") @@ -57,6 +60,9 @@ def main(): cv2.waitKey() case "batch_process": bp.batch_process_dataset() + case "train": + method = fusion_model.FusionModel() + method.train_model() case _: raise ValueError(f"Unknown command: {args.command}") diff --git a/eiuie/pixel_dataset.py b/eiuie/pixel_dataset.py new file mode 100644 index 0000000..50ce954 --- /dev/null +++ b/eiuie/pixel_dataset.py @@ -0,0 +1,36 @@ +from typing import Tuple + +import torch +from torch.utils.data import Dataset +import pandas as pd + +TSV_FILE = "data/pixel_dataset.tsv" + + +class PixelDataset(Dataset): + """ + PixelDataset class. + + Attributes + ---------- + df: pd.DataFrame + Dataframe. + """ + + df: pd.DataFrame + + def __init__(self): + self.df = pd.read_table(TSV_FILE, header=None) + self.df = self.df.astype(float) + + def __len__(self) -> int: + return len(self.df) + + def __getitem__(self, idx) -> Tuple[torch.Tensor, torch.Tensor]: + row = self.df.iloc[idx].values + + # Splitting the 15 values into two tensors: first 12 and last 3. + input_tensor = torch.tensor(row[:12], dtype=torch.float32) + output_tensor = torch.tensor(row[12:], dtype=torch.float32) + + return input_tensor, output_tensor diff --git a/pyproject.toml b/pyproject.toml index 5d50bc7..a59c99f 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -14,6 +14,7 @@ requires = [ "tdqm", "mpire", "numpy", + "pandas", "opencv-python", ] From cb920a0195e3625464644f66ec6aea01b7bc0601 Mon Sep 17 00:00:00 2001 From: CodingTil <36734749+CodingTil@users.noreply.github.com> Date: Mon, 30 Oct 2023 12:45:32 +0100 Subject: [PATCH 2/2] format --- eiuie/consolidate_dataset.py | 1 - 1 file changed, 1 deletion(-) diff --git a/eiuie/consolidate_dataset.py b/eiuie/consolidate_dataset.py index 7de28ae..2b10aa8 100644 --- a/eiuie/consolidate_dataset.py +++ b/eiuie/consolidate_dataset.py @@ -50,4 +50,3 @@ def image_to_pandas(source_path) -> pd.DataFrame: } df = pd.DataFrame(data) return df -