Impact of Synthetic Images in the Training of Neural Networks for Airborne Vessel Segmentation
Francisco Matilde, Gonçalo Cruz, Diogo Silva
Automation scripts for Blender and other utilities used in the development of this work can be found in the official Github repository.
Synthetic images have appeared as a possible solution to the scarcity of real images for training neural networks. We investigate their use in the training of ship detectors in aerial imagery. We create a synthetic dataset through 2 distinct methods: modeling and rendering with Blender, and a learning-based approach with GauGAN. YOLACT++ was chosen as the detector and its performance was evaluated when synthetic images were included in its training. The results suggest that synthetic augmentations can improve a model’s performance, under certain conditions. Adding synthetic data to large training sets degraded performance on the target domain, but significantly improved it on small datasets.
Images from Seagull dataset [1] were used for the test set. These are aerial images of vessels, which represent our target scenario. Our setup was under the assumption that we didn’t have enough data for our target scenario. To adress this, we simulated our target scenario with Blender and also used real images from a related domain, i.e. satellite imagery from Airbus Ship Detection Challenge [2]. We also used GauGAN to produce synthetic images from the related domain. We didn’t use generated GauGAN images for the target domain in the training set, as we had very little images and the generated images were fairly similar.
Our Blender generation process aimed to extend the methodology adopted in the production of the MarSyn dataset [3], by automating many tasks.