Bio-inspired enhancement for optical detection of drones using convolutional neural networks

Conference Paper (2023)
Author(s)

S. Luesutthiviboon (TU Delft - Control & Simulation, TU Delft - Reflection & Lifestyle)

Guido C.H.E.de de Croon (TU Delft - Control & Simulation)

A. Altena (TU Delft - Aircraft Noise and Climate Effects)

M Snellen (TU Delft - Control & Operations)

Mark Voskuijl (TU Delft - Flight Performance and Propulsion, Netherlands Defence Academy)

Research Group
Reflection & Lifestyle
DOI related publication
https://doi.org/10.1117/12.2673788
More Info
expand_more
Publication Year
2023
Language
English
Research Group
Reflection & Lifestyle
ISBN (electronic)
9781510667136
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Threats posed by drones urge defence sectors worldwide to develop drone detection systems. Visible-light and infrared cameras complement other sensors in detecting and identifying drones. Application of Convolutional Neural Networks (CNNs), such as the You Only Look Once (YOLO) algorithm, are known to help detect drones in video footage captured by the cameras quickly, and to robustly differentiate drones from other flying objects such as birds, thus avoiding false positives. However, using still video frames for training the CNN may lead to low drone-background contrast when it is flying in front of clutter, and omission of useful temporal data such as the flight trajectory. This deteriorates the drone detection performance, especially when the distance to the target increases. This work proposes to pre-process the video frames using a Bio-Inspired Vision (BIV) model of insects, and to concatenate the pre-processed video frame with the still frame as input for the CNN. The BIV model uses information from preceding frames to enhance the moving target-to-background contrast and embody the target’s recent trajectory in the input frames. An open benchmark dataset containing infrared videos of small drones (< 25 kg) and other flying objects is used to train and test the proposed methodology. Results show that, at a high sensor-to-target distance, the YOLO algorithms trained on BIV-processed frames and concatenation of the BIV-processed frames with still frames increase the Average Precision (AP) to 0.92 and 0.88, respectively, compared to 0.83 when it is trained on still frames alone.

Files

127420F.pdf
(pdf | 4.14 Mb)
License info not available