Skip to content
/ GOIS Public

Enhancing Tiny Object Detection Without Fine-Tuning: Guided Object Inference Slicing Framework with Latest YOLO Models and RT-DETR Transformer

License

Notifications You must be signed in to change notification settings

MMUZAMMUL/GOIS

Repository files navigation

MIT License - All rights reserved to the author. This project may be used for study and educational purposes, but **redistribution, redevelopment, or use of the code for personal or commercial purposes is strictly prohibited without the author's written consent.

🎥 Watch Live Demo (YouTube) | 🎥 Watch Live Demo (Bilibili)

🚀 Enhancing Tiny Object Detection Using Guided Object Inference Slicing (GOIS): An Efficient Dynamic Adaptive Framework for Fine-Tuned and Non-Fine-Tuned Deep Learning Models

Guided-Object Inference Slicing (GOIS) Innovatory Framework with Several Open source code Deployed on Google Colab/Gradio Live/Huggingface
🔬 Research by: Muhammad Muzammul, Xuewei Li, Xi Li
📄 Under Review in Neurocomputing
Contact: [email protected]

📌 Citation

**MUZAMMUL, MUHAMMAD and LI, Xuewei and Li, Xi**,  
*"Enhancing Tiny Object Detection Without Fine Tuning: Dynamic Adaptive Guided Object Inference Slicing Framework with Latest YOLO Models and RT-DETR Transformer."*  
Available at [SSRN](https://ssrn.com/abstract=5092517) or via DOI: [10.2139/ssrn.5092517](http://dx.doi.org/10.2139/ssrn.5092517)

📥 Quick Start

Step Command
1️⃣ Clone Repo git clone https://github.com/MMUZAMMUL/GOIS.git && cd GOIS
2️⃣ Download Data Follow Dataset Instructions or Download 15% Dataset
3️⃣ Download Models cd Models && python download_models.py
4️⃣ Generate Ground Truth python scripts/generate_ground_truth.py --annotations_folder "<annotations_path>" --images_folder "<images_path>" --output_coco_path "./data/ground_truth/ground_truth_coco.json"
5️⃣ Full Inference (FI-Det) python scripts/full_inference.py --images_folder "<path>" --model_path "Models/yolo11n.pt" --model_type "YOLO" --output_base_path "./data/FI_Predictions"
6️⃣ GOIS Inference python scripts/gois_inference.py --images_folder "<path>" --model_path "Models/yolo11n.pt" --model_type "YOLO" --output_base_path "./data/gois_Predictions"
7️⃣ Evaluate FI-Det python scripts/evaluate_prediction.py --ground_truth_path "./data/ground_truth/ground_truth_coco.json" --predictions_path "./data/FI_Predictions/full_inference.json" --iou_type bbox
8️⃣ Evaluate GOIS-Det python scripts/evaluate_prediction.py --ground_truth_path "./data/ground_truth/ground_truth_coco.json" --predictions_path "./data/gois_Predictions/gois_inference.json" --iou_type bbox
9️⃣ Compare Results python scripts/calculate_results.py --ground_truth_path "./data/ground_truth/ground_truth_coco.json" --full_inference_path "./data/FI_Predictions/full_inference.json" --gois_inference_path "./data/gois_Predictions/gois_inference.json"
🔟 Upscale Metrics python scripts/evaluate_upscaling.py --ground_truth_path "./data/ground_truth/ground_truth_coco.json" --full_inference_path "./data/FI_Predictions/full_inference.json" --gois_inference_path "./data/gois_Predictions/gois_inference.json"

📊 Test GOIS Benchmarks & Gradio Live Deployement

📂 GOIS Benchmarks Repository
🎥 Watch Live Demo (YouTube) | 🎥 Watch Live Demo (Bilibili)

🔑 MIT License - Study & Educational Use Only
📧 Contact: Author Email

🚀 GOIS Live Deployed Applications on Gradio ✅

Explore the GOIS-Det vs. FI-Det benchmark results through live interactive applications on Gradio. These applications provide detailed comparisons using graphs, tables, and output images, demonstrating the effectiveness of GOIS-Det in tiny object detection.

🔥 Live Benchmark Tests different categories

Test Function Description Live Test
1️⃣ Single Image Analysis (GOIS-Det vs. FI-Det) Perform a single image test to visualize graphs and results, comparing FI-Det vs. GOIS-Det. View detection metrics such as the number of detections and class diversity. Outputs include pie charts, bar charts, and two comparative images that highlight the significance of GOIS. Open in Colab
2️⃣ Multiple Images Analysis (GOIS-Det vs. FI-Det) Upload multiple images simultaneously and compare GOIS-Det and FI-Det outputs. A table of detection metrics is generated to clearly evaluate the improvements achieved by GOIS. Open in Colab
3️⃣ Video Analysis (GOIS-Det vs. FI-Det) Perform a video-based evaluation of GOIS-Det vs. FI-Det. The application generates a table comparing the number of detections and detected classes, providing insights into GOIS's effectiveness. Open in Colab
4️⃣ Metrics Evaluation & Results Graphs (GOIS-Det vs. FI-Det) Compare key detection metrics, including AP, AR, mAP, and F1-score for FI-Det and GOIS-Det. View graphs, tables, percentage improvements, and output images to assess GOIS's impact on detection performance. Open in Colab

📌 Instructions:

  1. Click on any "Open in Colab" button above to launch the interactive notebook.
  2. Follow the instructions in the notebook to test GOIS-Det vs. FI-Det.
  3. Evaluate detection performance using provided visualizations and metrics.

🚀 GOIS Live Deployed Applications on Hugging Face ✅ Hugging Face

Experience Guided Object Inference Slicing (GOIS) across images, videos, and live cameras with configurable parameters. Evaluate real-time small object detection and compare against full-image inference (FI-Det).

📂 Compatible Datasets: VisDrone, UAV Surveillance (100-150ft), Pedestrian & Tiny Object Detection, Geo-Sciences
🖥️ Applied Models: YOLO11, YOLOv10, YOLOv9, YOLOv8, YOLOv6, YOLOv5, RT-DETR-L, YOLOv8s-Worldv2

GOIS incorporates a two-stage hierarchical slicing strategy, dynamically adjusting coarse-to-fine slicing and overlap rates to optimize tiny object detection while reducing false positives. These live applications allow users to test GOIS against full-image inference, analyze occlusion handling, boundary artifacts, and false positive reductions, while adjusting key parameters.

Test Function Description 🔗 Test Link
GOIS vs. Full-Image Detection Evaluates dynamic slicing vs. full-image inference (FI-Det) across images, identifying missed objects and reducing false positives. Hugging Face GOIS Live Image Processing
Video Detection (Single Stage) Tests frame-wise GOIS slicing to improve small object detection, mitigating occlusion issues. Hugging Face GOIS Video Inference (Single Stage)
Advanced Video Detection (Two Stage) Uses coarse-to-fine GOIS slicing based on object density to dynamically adjust slicing strategies and eliminate boundary artifacts. Hugging Face GOIS Video Inference (Two Stage)
Live Camera Detection (FI vs. GOIS) Compares full-frame inference vs. GOIS slicing in real-time, highlighting differences in object localization and accuracy. Hugging Face GOIS Live Camera Test
Live Camera Advanced Detection Demonstrates adaptive slicing based on object density, improving small object retrieval while maintaining efficiency. Hugging Face GOIS Live Camera Advanced

🔹 How to Use

1️⃣ Click a Test Link → 2️⃣ Upload Image/Video → 3️⃣ Adjust Parameters (Slice Size, Overlap, NMS) → 4️⃣ Compare FI vs. GOIS Results → 5️⃣ Analyze Performance in Real-Time


📌 Google Colab Live Test Links for GOIS Google Colab

To validate ✅ the Guided Object Inference Slicing (GOIS) framework, the following Google Colab notebooks provide real-time inference and analysis. These tests allow users to compare GOIS vs. Full-Image Detection (FI-Det) across different datasets and models.

📂 Compatible Datasets: VisDrone, UAV Surveillance (100-150ft), Pedestrian & Tiny Object Detection, Geo-Sciences
🖥️ Applied Models: YOLO11, YOLOv10, YOLOv9, YOLOv8, YOLOv6, YOLOv5, RT-DETR-L, YOLOv8s-Worldv2

GOIS differs from traditional slicing methods (SAHI, ASAHI) by dynamically adjusting slicing parameters based on object density rather than static window sizes. These notebooks enable comparative testing, allowing users to experiment with slicing sizes, overlap rates, and NMS thresholds, addressing key performance trade-offs.

Test Function Description Colab Link
GOIS vs. SAHI/ASAHI (Proposed Method) Compares GOIS dynamic slicing vs. static slicing (SAHI, ASAHI-like), analyzing boundary artifacts and false positive rates. Google Colab📌 GOIS vs. SAHI/ASAHI
GOIS - Single Image Inference Runs GOIS on a single image, adjusting slicing parameters and overlap rates. Google Colab📌 GOIS Single Image Test
GOIS vs. FI-Det (Single Image) Side-by-side visual comparison of GOIS vs. FI-Det, addressing occlusion and small object visibility. Google Colab📌 GOIS vs. FI-Det (Single Image)
GOIS vs. FI-Det (Multiple Images) Processes multiple images to compare detection consistency across datasets. Google Colab📌 GOIS vs. FI-Det (Multiple Images)
Detection Count & Metrics Comparison Evaluates object count, area coverage, and false positive reduction rates. Google Colab📌 GOIS vs. FI-Det (Metrics Test)
Slice Size Optimization - Speed Test Tests how different slicing sizes and overlap settings impact speed vs. accuracy. Google Colab📌 GOIS Optimized Speed Test
GOIS - 81 Parameter Combinations Test Tests 81 slicing, overlap, and NMS variations for optimal performance. Google Colab📌 GOIS 81 Combinations Test
GOIS - Three Best Slicing Configurations Evaluates three optimized GOIS slicing setups based on empirical results:
C1: 512px/128px (0.1 overlap, NMS 0.3)
C2: 640px/256px (0.2 overlap, NMS 0.4)
C3: 768px/384px (0.3 overlap, NMS 0.5). These configurations were determined as optimal trade-offs between accuracy, false positive reduction, and computational efficiency.
Google Colab📌 GOIS Ideal Slicing Test

🛠 How to Use

1️⃣ Open any Colab link → 2️⃣ Run the notebook → 3️⃣ Upload images or use datasets → 4️⃣ Adjust GOIS parameters (slice size, overlap, NMS) → 5️⃣ Compare FI vs. GOIS results


📊 GOIS Benchmark Results - Performance Comparison Across Datasets

The following tables present benchmark evaluations of the Guided Object Inference Slicing (GOIS) framework, comparing Full Inference (FI-Det) vs. GOIS-Det across different datasets and model configurations.

GOIS integrates a two-stage hierarchical slicing strategy, dynamically adjusting slice size, overlap rate, and NMS thresholds to optimize detection performance. These results highlight improvements in small object detection, reduction of boundary artifacts, and comparisons with existing slicing methods like SAHI and ASAHI.

Test/Part Dataset & Setup Description Benchmark Link
Part 1 Without Fine-Tuning - 15% Dataset (970 Images) - VisDrone2019Train Evaluates FI-Det vs. GOIS-Det on a small dataset subset. The table presents AP and AR metrics for seven models, comparing detection performance with and without GOIS enhancements. The percentage improvement achieved by GOIS is included for each model. 📌 Section 1 - GOIS Benchmarks
Part 2 Fine-Tuned Models (10 Epochs) - Full Dataset (6,471 Images) - VisDrone2019Train GOIS performance is tested after 10 epochs of fine-tuning. The impact of GOIS slicing parameters (coarse-fine slice size, overlap rate, NMS filtering) is analyzed. The table provides detailed AP and AR metrics for five models, highlighting GOIS's ability to improve small object recall while managing computational efficiency. 📌 Section 2 - GOIS Benchmarks
Part 3 Without Fine-Tuning - Five Models - Full Dataset (6,471 Images) - VisDrone2019Train Evaluates GOIS on a large-scale dataset without fine-tuning, highlighting its robust generalization ability. Comparative results for five models (YOLO11, YOLOv10, YOLOv9, YOLOv8, YOLOv5) include FI-Det, GOIS-Det, and % improvement achieved by GOIS. This setup assesses GOIS’s impact on both small and large object detection. 📌 Section 3 - GOIS Benchmarks
Part 4 General Analysis - Pretrained Weights on VisDrone, xView, MS COCO GOIS's adaptability is tested across multiple datasets and model architectures. This section evaluates pretrained YOLO and transformer-based detectors (e.g., RT-DETR-L) to measure cross-domain effectiveness, computational trade-offs, and improvements in occlusion handling. Key focus: Can GOIS be applied universally? 📌 Section 4,5 - GOIS Benchmarks
Part 5 Comparative Analysis - SAHI vs. ASAHI vs. GOIS A quantitative and qualitative comparison between GOIS and other slicing frameworks (SAHI, ASAHI) across VisDrone2019 and xView datasets. This section examines: 1️⃣ Boundary artifact reduction, 2️⃣ False positive minimization, and 3️⃣ Effectiveness of dynamic slicing in handling occlusion issues. Detailed benchmark tables are included. 📌 Section 4,5 - GOIS vs. SAHI/ASAHI Benchmarks

🔍 Key Improvements in GOIS Over Other Methods

Dynamic Slicing Optimization: Unlike static SAHI/ASAHI methods, GOIS adjusts slice sizes and overlap rates based on object density, reducing redundant processing.
Occlusion Handling & Boundary Artifact Reduction: GOIS minimizes false detections and truncated object artifacts by dynamically refining inference slices.
Scalability Across Models & Datasets: Successfully applied to YOLO models, RT-DETR, and various datasets, proving its universal applicability.
Performance Gains in Small Object Detection: GOIS consistently improves AP-Small and AR-Small metrics, as validated on VisDrone and xView datasets.

📌 For additional benchmark results and evaluation scripts, visit: GOIS Benchmarks Repository

Cite This Work

If you use GOIS in your research, please consider citing our paper:

About

Enhancing Tiny Object Detection Without Fine-Tuning: Guided Object Inference Slicing Framework with Latest YOLO Models and RT-DETR Transformer

Topics

Resources

License

Code of conduct

Security policy

Stars

Watchers

Forks

Languages