MaskAL is an active learning framework that automatically selects the most-informative images for training Mask R-CNN. By using MaskAL, it is possible to reduce the number of image annotations, without negatively affecting the performance of Mask R-CNN. Generally speaking, MaskAL involves the following steps:
- Train Mask R-CNN on a small initial subset of a bigger dataset
- Use the trained Mask R-CNN algorithm to make predictions on the unlabelled images of the remaining dataset
- Select the most-informative images with a sampling algorithm
- Annotate the most-informative images, and then retrain Mask R-CNN on the most informative-images
- Repeat step 2-4 for a specified number of sampling iterations
The figure below shows the performance improvement of MaskAL on our dataset. By using MaskAL, the performance of Mask R-CNN improved more quickly and therefore 1400 annotations could be saved (see the black dashed line):
Linux/Ubuntu: INSTALL.md
Windows: INSTALL_Windows.md
Split the dataset in a training set, validation set and a test set. It is not required to annotate every image in the training set, because MaskAL will select the most-informative images automatically.
- From the training set, a smaller initial dataset is randomly sampled (the dataset size can be specified in the maskAL.yaml file). The images that do not have an annotation are placed in the annotate subfolder inside the image folder. You first need to annotate these images with LabelMe (json), V7-Darwin (json), Supervisely (json) or CVAT (xml) (when using CVAT, export the annotations to LabelMe 3.0 format). Refer to our annotation procedure: ANNOTATION.md
- Step 1 is repeated for the validation set and the test set (the file locations can be specified in the maskAL.yaml file).
- After the first training iteration of Mask R-CNN, the sampling algorithm selects the most-informative images (its size can be specified in the maskAL.yaml file).
- The most-informative images that don't have an annotation, are placed in the annotate subfolder. Annotate these images with LabelMe (json), V7-Darwin (json), Supervisely (json) or CVAT (xml) (when using CVAT, export the annotations to LabelMe 3.0 format).
- OPTIONAL: it is possible to use the trained Mask R-CNN model to auto-annotate the unlabelled images to further reduce annotation time. Activate auto_annotate in the maskAL.yaml file, and specify the export_format (currently supported formats: 'labelme', 'cvat', 'darwin', 'supervisely').
- Step 3-5 are repeated for several training iterations. The number of iterations (loops) can be specified in the maskAL.yaml file.
Please note that MaskAL does not work with the default COCO json-files of detectron2. These json-files contain all annotations that are completed before the training starts. Because MaskAL involves an iterative train and annotation procedure, the default COCO json-files lack the desired format.
Open a terminal (Ctrl+Alt+T):
(base) user@computer:~$ cd maskal
(base) user@computer:~/maskal$ conda activate maskAL
(maskAL) user@computer:~/maskal$ python maskAL.py --config maskAL.yaml
Change the following settings in the maskAL.yaml file:
Setting | Description |
---|---|
weightsroot | The file directory where the weight-files are stored |
resultsroot | The file directory where the result-files are stored |
dataroot | The root directory where all image-files are stored |
initial_train_dir | When use_initial_train_dir is activated: the file directory where the initial training images and annotations are stored |
traindir | The file directory where the training images and annotations are stored |
valdir | The file directory where the validation images and annotations are stored |
testdir | The file directory where the test images and annotations are stored |
use_initial_train_dir | Set this to True when you want to start the active-learning from an initial training dataset. When False, the initial dataset of size initial_datasize is randomly sampled from the traindir |
network_config | The Mask R-CNN configuration-file (.yaml) file (see the folder './configs') |
pretrained_weights | The pretrained weights to start the active-learning. Either specify the network_config (.yaml) or a custom weights-file (.pth or .pkl) |
cuda_visible_devices | The identifiers of the CUDA device(s) you want to use for training and sampling (in string format, for example: '0,1') |
classes | The names of the classes in the image annotations |
learning_rate | The learning-rate to train Mask R-CNN (default value: 0.01) |
confidence_threshold | Confidence-threshold for the image analysis with Mask R-CNN (default value: 0.5) |
nms_threshold | Non-maximum suppression threshold for the image analysis with Mask R-CNN (default value: 0.3) |
initial_datasize | The size of the initial dataset to start the active learning (when use_initial_train_dir is False) |
pool_size | The number of most-informative images that are selected from the traindir |
loops | The number of sampling iterations |
auto_annotate | Set this to True when you want to auto-annotate the unlabelled images |
export_format | When auto_annotate is activated: specify the export-format of the annotations (currently supported formats: 'labelme', 'cvat', 'darwin', 'supervisely') |
supervisely_meta_json | When supervisely auto_annotate is activated: specify the file location of the meta.json for supervisely export |
Description of the other settings in the maskAL.yaml file: MISC_SETTINGS.md
Please refer to the folder active_learning/config for more setting-files.
Use a trained Mask R-CNN algorithm to auto-annotate unlabelled images: auto_annotate.py
Argument | Description |
---|---|
--img_dir | The file directory where the unlabelled images are stored |
--network_config | Configuration of the backbone of the network |
--classes | The names of the classes of the annotated instances |
--conf_thres | Confidence threshold of the CNN to do the image analysis |
--nms_thres | Non-maximum suppression threshold of the CNN to do the image analysis |
--weights_file | Weight-file (.pth) of the trained CNN |
--export_format | Specifiy the export-format of the annotations (currently supported formats: 'labelme', 'cvat', 'darwin', 'supervisely') |
--supervisely_meta_json | The file location of the meta.json for supervisely export |
Example syntax (auto_annotate.py):
python auto_annotate.py --img_dir datasets/train --network_config COCO-InstanceSegmentation/mask_rcnn_X_101_32x8d_FPN_3x.yaml --classes healthy damaged matured cateye headrot --conf_thres 0.5 --nms_thres 0.2 --weights_file weights/broccoli/model_final.pth --export_format supervisely --supervisely_meta_json datasets/meta.json
See our research article on arXiv and Computers and Electronics in Agriculture for more information:
@misc{blok2021active,
title = {Active learning with MaskAL reduces annotation effort for training Mask R-CNN},
author = {Blok, Pieter M. and Kootstra, Gert and Elghor, Hakim Elchaoui and Diallo, Boubacar and van Evert, Frits K. and van Henten, Eldert J.},
year = {2021},
doi = {10.48550/ARXIV.2112.06586},
url = {https://arxiv.org/abs/2112.06586},
publisher = {arXiv},
copyright = {Creative Commons Attribution Share Alike 4.0 International},
keywords = {Computer Vision and Pattern Recognition (cs.CV), FOS: Computer and information sciences, FOS: Computer and information sciences}
}
@article{blok2022106917,
title = {Active learning with MaskAL reduces annotation effort for training Mask R-CNN on a broccoli dataset with visually similar classes},
author = {Pieter M. Blok and Gert Kootstra and Hakim Elchaoui Elghor and Boubacar Diallo and Frits K. {van Evert} and Eldert J. {van Henten}},
journal = {Computers and Electronics in Agriculture},
volume = {197},
pages = {106917},
year = {2022},
issn = {0168-1699},
doi = {10.1016/j.compag.2022.106917},
url = {https://www.sciencedirect.com/science/article/pii/S0168169922002344}
}
Our software was forked from Detectron2 (https://github.com/facebookresearch/detectron2). As such, the software will be released under the Apache 2.0 license.
The uncertainty calculation methods were inspired by the research of Doug Morrison:
https://nikosuenderhauf.github.io/roboticvisionchallenges/assets/papers/CVPR19/rvc_4.pdf
Two software methods were inspired by the work of RovelMan:
https://github.com/RovelMan/active-learning-framework
MaskAL uses the Bayesian Active Learning (BaaL) software:
https://github.com/ElementAI/baal
MaskAL is developed and maintained by Pieter Blok.