This project focuses on enhancing remote sensing image quality and improving object detection accuracy in aerial imagery. Using the DOTA v1 dataset, low-resolution images are simulated and restored with Super-Resolution Generative Adversarial Network (SRGAN) to enhance sharpness and fine details.
Object detection is performed with YOLO-v11 and YOLO-v11 OBB (Oriented Bounding Box), enabling accurate detection of objects at varying orientations. The method identifies eight object classes and demonstrates significant improvements in image quality and detection accuracy through the integration of super-resolution and oriented bounding boxes.
Comparative results on low-resolution vs. super-resolved images highlight the effectiveness of this approach for applications in remote sensing and aerial surveillance.
DOTA v1 dataset can be found here.
Link to download SRGAN checkpoint train on DOTA v1 dataset: link
Link to download YOLOv11 and YOLOv11-OBB models: link
We developed a Streamlit-based GUI for interacting with the super-resolution and object detection pipeline. Choose a few test images from the DOTA V1 dataset for the GUI demo. To see how the GUI works, run the following code:
streamlit run GUI.py
The GUI offers the following features:
1. Image Upload & Model Selection: Users can upload images, select object detection models (YOLOv11 or YOLOv11-OBB), and super-resolution models (SRGAN or SRResNet) to observe differences.
2. Super-Resolution Results: Displays super-resolved images with PSNR and SSIM metrics compared to the high-resolution ground truth.
3. Object Detection on Super-Resolved Images: Shows detected objects with bounding boxes and class labels using the chosen YOLO model.
i) Number of detections (with confidence ≥ 0.5).
ii) Average confidence scores for detections.
This GUI allows users to visually assess how SRGAN improves object detection on real-world test data.