The Internet of Things (IoT), with smart sensors, collects and generates big data streams for a wide range of applications. One of the important applications in this regard is video analytics which includes object detection. It has been considered as an important research area particularly after the development of deep neural networks. We demonstrate the applications, effectiveness, and efficiency of the convolutional neural network algorithms, i.e., Faster-RCNN and Mask-RCNN, to facilitate video analytics in the IoT domain, for overhead view multiple object detection and segmentation. We used the Faster-RCNN and Mask-RCNN models trained on the frontal view data set. To evaluate the performance of both algorithms, we used a newly recorded overhead view data set containing images of different objects having variation in field of view, background, illumination condition, poses, scales, sizes, angles, height, aspect ratio, and camera resolutions. Although the overhead view appearance of an object is significantly different as compared to a frontal view, even then the experimental results show the potential of the deep learning models by achieving the promising results. For Faster-RCNN, we achieved a true-positive rate (TPR) of 94% with a false-positive rate (FPR) of 0.4% for the overhead view images of persons, while for other objects the maximum obtained TPR is 92%. The Mask-RCNN model produced TPR of 93% with FPR of 0.5% for person images and maximum TPR of 92% for other objects. Furthermore, the detailed discussion is made on output results which highlights the challenges and possible future directions.

Exploring Deep Learning Models for Overhead View Multiple Object Detection / Ahmed, I.; Din, S.; Jeon, G.; Piccialli, F.. - In: IEEE INTERNET OF THINGS JOURNAL. - ISSN 2327-4662. - 7:7(2020), pp. 5737-5744. [10.1109/JIOT.2019.2951365]

Exploring Deep Learning Models for Overhead View Multiple Object Detection

Piccialli F.
2020

Abstract

The Internet of Things (IoT), with smart sensors, collects and generates big data streams for a wide range of applications. One of the important applications in this regard is video analytics which includes object detection. It has been considered as an important research area particularly after the development of deep neural networks. We demonstrate the applications, effectiveness, and efficiency of the convolutional neural network algorithms, i.e., Faster-RCNN and Mask-RCNN, to facilitate video analytics in the IoT domain, for overhead view multiple object detection and segmentation. We used the Faster-RCNN and Mask-RCNN models trained on the frontal view data set. To evaluate the performance of both algorithms, we used a newly recorded overhead view data set containing images of different objects having variation in field of view, background, illumination condition, poses, scales, sizes, angles, height, aspect ratio, and camera resolutions. Although the overhead view appearance of an object is significantly different as compared to a frontal view, even then the experimental results show the potential of the deep learning models by achieving the promising results. For Faster-RCNN, we achieved a true-positive rate (TPR) of 94% with a false-positive rate (FPR) of 0.4% for the overhead view images of persons, while for other objects the maximum obtained TPR is 92%. The Mask-RCNN model produced TPR of 93% with FPR of 0.5% for person images and maximum TPR of 92% for other objects. Furthermore, the detailed discussion is made on output results which highlights the challenges and possible future directions.
2020
Exploring Deep Learning Models for Overhead View Multiple Object Detection / Ahmed, I.; Din, S.; Jeon, G.; Piccialli, F.. - In: IEEE INTERNET OF THINGS JOURNAL. - ISSN 2327-4662. - 7:7(2020), pp. 5737-5744. [10.1109/JIOT.2019.2951365]
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11588/821057
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 68
  • ???jsp.display-item.citation.isi??? 69
social impact