| import os
|
| from typing import Tuple, Dict
|
| import torch
|
| import torch.nn as nn
|
| import torch.nn.functional as F
|
| import numpy as np
|
| from PIL import Image
|
| import timm
|
|
|
|
|
| class XceptionModel:
|
|
|
|
|
| CLASS_NAMES = ["Auto Rickshaws", "Bikes", "Cars", "Motorcycles", "Planes", "Ships", "Trains"]
|
|
|
| def __init__(self, model_dir: str, model_file: str = "best_model_finetuned_full.pt"):
|
| self.model_dir = model_dir
|
| self.model_file = model_file
|
| self.device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| self.model = None
|
| self.inference_transform = None
|
| self.class_names = self.CLASS_NAMES
|
|
|
| print(f"[Xception] Using device: {self.device}")
|
| print(f"[Xception] Classes: {self.class_names}")
|
| self._load_model()
|
|
|
| def _load_model(self):
|
| try:
|
| model_path = os.path.join(self.model_dir, self.model_file)
|
|
|
| if not os.path.exists(model_path):
|
| raise FileNotFoundError(f"Model file not found: {model_path}")
|
|
|
|
|
| torch._dynamo.config.suppress_errors = True
|
| torch._dynamo.reset()
|
|
|
|
|
| checkpoint = torch.load(model_path, map_location=self.device, weights_only=False)
|
|
|
| num_classes = len(self.CLASS_NAMES)
|
|
|
| if isinstance(checkpoint, dict) and not hasattr(checkpoint, "forward"):
|
|
|
| model = timm.create_model("xception", pretrained=False, num_classes=num_classes)
|
| in_features = model.get_classifier().in_features
|
| model.fc = nn.Sequential(
|
| nn.Linear(in_features, 512),
|
| nn.ReLU(),
|
| nn.Dropout(0.5),
|
| nn.Linear(512, num_classes),
|
| )
|
|
|
| state_dict = checkpoint
|
| if any(k.startswith("_orig_mod.") for k in state_dict.keys()):
|
| state_dict = {k.replace("_orig_mod.", ""): v for k, v in state_dict.items()}
|
|
|
| model.load_state_dict(state_dict)
|
| else:
|
|
|
| model = checkpoint
|
| if hasattr(model, "_orig_mod"):
|
| model = model._orig_mod
|
|
|
|
|
| self.model = model.to(self.device).eval()
|
|
|
|
|
| data_config = timm.data.resolve_model_data_config(self.model)
|
| self.inference_transform = timm.data.create_transform(**data_config, is_training=False)
|
|
|
| print(f"[Xception] Model loaded successfully from {model_path}")
|
|
|
| except Exception as e:
|
| print(f"[Xception] Error loading model: {e}")
|
| raise
|
|
|
| def _preprocess_image(self, img: Image.Image) -> torch.Tensor:
|
| img = img.convert("RGB")
|
| tensor = self.inference_transform(img).unsqueeze(0).to(self.device)
|
| return tensor
|
|
|
| def predict(self, image: Image.Image) -> Tuple[str, float, Dict[str, float]]:
|
| if image is None:
|
| return "No image provided", 0.0, {}
|
|
|
| try:
|
|
|
| if not isinstance(image, Image.Image):
|
| image = Image.fromarray(image)
|
|
|
|
|
| inputs = self._preprocess_image(image)
|
|
|
|
|
| with torch.no_grad():
|
| outputs = self.model(inputs)
|
| probs = F.softmax(outputs, dim=-1).cpu().numpy()[0]
|
|
|
|
|
| class_idx = int(np.argmax(probs))
|
| confidence = float(probs[class_idx])
|
| prob_dict = {self.class_names[i]: float(probs[i]) for i in range(len(self.class_names))}
|
|
|
| return self.class_names[class_idx], confidence, prob_dict
|
|
|
| except Exception as e:
|
| print(f"[Xception] Error during prediction: {e}")
|
| raise
|
|
|