Abstract
2D radars are efficient sensors used for e.g. coastal or shipborne surveillance. However, the recorded data contains echoes from all its surroundings, without any discrimination of land, sea or occluded terrain, which degrades the performance of target detectors and trackers. We assume that a complete 360 radar scan can be used as an image and thereby exploit its spatial information with a multi-scale feature-connected convolutional autoencoder to perform image-based radar segmentation. Our method is compared against the reimplementation of a temporal-based classifier when using unfiltered radar data. The conducted experiments display that our framework can overcome the noise problems inherit in 2D radar data and discriminate the different surfaces by outperforming the temporal-based implementation with a 20% increase in mean pixel-wise accuracy, with a mAP of 67%, and a mean IoU of 58.67%. This is a promising approach towards the application of deep learning for segmentation of radar-based images.
Original language | English |
---|---|
Title of host publication | Proceedings of AVSS 2018 - 2018 15th IEEE International Conference on Advanced Video and Signal-Based Surveillance |
Publisher | IEEE |
Publication date | 14 Feb 2019 |
Article number | 8639357 |
ISBN (Print) | 978-1-5386-9295-0 |
ISBN (Electronic) | 978-1-5386-9294-3 |
DOIs | |
Publication status | Published - 14 Feb 2019 |
Event | 15th IEEE International Conference on Advanced Video and Signal-Based Surveillance, AVSS 2018 - Auckland, New Zealand Duration: 27 Nov 2018 → 30 Nov 2018 |
Conference
Conference | 15th IEEE International Conference on Advanced Video and Signal-Based Surveillance, AVSS 2018 |
---|---|
Country/Territory | New Zealand |
City | Auckland |
Period | 27/11/2018 → 30/11/2018 |