Abstract
Deep learning has shown potential in domains where large-scale annotated datasets are available. However, manual annotation is expensive, time-consuming, and tedious. Pixel-level annotations are particularly costly for semantic segmentation in images with dense irregular patterns of object instances, such as in plant images. In this work, we propose a method for developing high-performing deep learning models for semantic segmentation of wheat heads utilizing little manual annotation. We simulate a computationally-annotated dataset using a few annotated images, a short unannotated video clip of a wheat field, and several video clips from fields with no wheat. This dataset is then used to train a customized U-Net model for wheat head segmentation. Considering the distribution shift between the simulated and real data, we apply three domain adaptation steps to gradually bridge the domain gap. Only using two annotated images, we achieved a Dice score of 0.89 on the internal test set, i.e., images extracted from the wheat field video. The model trained using only two annotated images was evaluated on a diverse external dataset collected from 18 different domains across five countries and achieved a Dice score of 0.73. To further expose the model to images from different growth stages and environmental conditions, we incorporated two annotated images from each of the 18 domains and further fine-tuned the model. This resulted in improving the Dice score to 0.91. These promising results highlight the utility of the proposed approach in the absence of large-annotated datasets. Although the utility of the proposed method is shown on a wheat head dataset, it can be extended to other segmentation tasks with similar characteristics of irregularly repeating patterns of object instances.
Competing Interest Statement
The authors have declared no competing interest.
Footnotes
The affiliation of Gholam Hassan Shirdel should be Mathematics Department, Faculty of Sciences, University of Qom, Qom, Iran. It was stated incorrectly in the first version.