Skip to content

Latest commit

 

History

History
159 lines (132 loc) · 5.69 KB

README.md

File metadata and controls

159 lines (132 loc) · 5.69 KB

COHI-O365

We present the most diverse in number of images/labels/classes fisheye synthetic dataset with source codes and models. As well as a benchmarking testing real dataset.

COHI-O365 is a benchmark testing dataset for object detection in hemispherical/fisheye for field of view invariant applications. It contains 1,000 real fisheye images of 74 classes sampled from the Objects365 dataset with 20.798 object instances per image on average. The images were captured using a hemispherical camera ELP-USB8MP02G-L180 with the 2,448 by 3,264 pixel resolution and manually annotated with standard axis-aligned bounding boxes afterward. The samples of raw images from the COHI-O365 dataset are shown below

The names of sampled classes and the number of bounding boxes for each class are presented in the next figure (to be inserted).

RMFV365 dataset

To train object detection models for the COHI-O365 dataset, the Fisheye versions of Objects 365 dataset was generated by applying a non-linear mapping to obtain fisheye-looking images. We called it the RMFV365 dataset, comprising 5.1 million images that encompass a broad spectrum of perspectives and distortions. This dataset serves as a valuable resource for training the model in generalized object recognition tasks. The samples of raw and annotated images are illustrated below.

Download the datasets

The COHI-O365 dataset

One can access the COHI-O365 dataset using the following Google Drive link: COHI-O365

The RMFV365 dataset

The RMFV365 dataset can be downloaded with a link: (Link is coming!!!). Alternatively it can be generated from Objects365 dataset using the python scripts in transformations directory for image transformation and coordinates directory for coordinate mapping

Requirements

For data preprocessing

  • numpy
  • PIL
  • pandas

For object detection

We used YOLOv7 to train and evaluate object detection models. All needed information can be found on their official GitHub page YOLOv7.

Pre-trained models

We trained the YOLOv7 model with 36.9 M parameters on three datasets, namely Objects365, RMFV365, and a variant of RMFV365 codenamed RMFV365-v1 and evaluated the performance of models with our benchmark testing dataset - COHI-O365.

  • YOLOv7-0: trained on the Objects365 dataset
  • YOLOv7-T1: trained on Objects365 dataset and fisheye images transformed using lens and camera independent fisheye transformation with parameter n = 4, codenamed RMFV365-v1
  • YOLOv7-T2: trained on RMFV365

Results

mAP50 results are summarized in the table below.

S/N Model Test Results (%)
Objects365 RMFV365-v1 RMFV365 COHI-365
mAP50 mAP50:95 mAP50 mAP50:95 mAP50 mAP50:95 mAP50 mAP50:95
1 FPN 35.5 22.5 N/A N/A N/A N/A N/A N/A
2 RetinaNet 27.3 18.7 N/A N/A N/A N/A N/A N/A
3 YOLOv5m 27.3 18.8 22.6 14.1 18.7 10.1 40.4 28.0
4 YOLOv7-0 34.97 24.57 29.1 18.3 24.2 13.0 47.5 33.5
5 YOLOv7-T1 34.3 24.0 32.7 22.7 32.0 22.0 49.1 34.6
6 YOLOv7-T2 34 23.1 32.9 23 33 22.8 49.9 34.9

Table: Objects recognition results on Objects365, RMFV365-v1, RMFV365, and COHI-365 Testing Sets

Pre-trained model weights can be downloaded using a Google Drive link: (Link is Coming)

Citation

.........................