Object detection is a central theme for many Artificial Intelligence (AI) applications such as autonomous vehicles, surveillance etc. The algorithms providing this capability rely on training data being available in corresponding sensor mode. Having co-located data from multiple sensor modes enhances the detection confidence, but the availability of training data in desired sensor mode is not always readily available, which slows down progress. In this paper, we investigate the ability to translate images from one sensor mode to another, on a single fixed camera dataset, using conditional Generative Adversarial Network (cGAN). Specifically, images are transferred from Electro-Optical (EO) to Infra-Red (IR) images and vice-versa using cGAN models, which are generative models that learn the data distribution in a minimax game setting. To investigate the usability of such transferred images, we apply object detection algorithm on ground truth and transferred images and compare their performance. The results indicate that transferred images match closely to real images and object detection has good performance on transferred images, especially when the object size is large.
|