Please use this identifier to cite or link to this item: https://dspace.iiti.ac.in/handle/123456789/6561
Full metadata record
DC FieldValueLanguage
dc.contributor.authorTanveer, M.en_US
dc.date.accessioned2022-03-17T01:00:00Z-
dc.date.accessioned2022-03-21T10:49:49Z-
dc.date.available2022-03-17T01:00:00Z-
dc.date.available2022-03-21T10:49:49Z-
dc.date.issued2021-
dc.identifier.citationShahid, M., Virtusio, J. J., Wu, Y. -., Chen, Y. -., Tanveer, M., Muhammad, K., & Hua, K. -. (2022). Spatio-temporal self-attention network for fire detection and segmentation in video surveillance. IEEE Access, 10, 1259-1275. doi:10.1109/ACCESS.2021.3132787en_US
dc.identifier.issn2169-3536-
dc.identifier.otherEID(2-s2.0-85120895391)-
dc.identifier.urihttps://doi.org/10.1109/ACCESS.2021.3132787-
dc.identifier.urihttps://dspace.iiti.ac.in/handle/123456789/6561-
dc.description.abstractConvolutional Neural Network (CNN) based approaches are popular for various image/video related tasks due to their state-of-the-art performance. However, for problems like object detection and segmentation, CNNs still suffer from objects with arbitrary shapes or sizes, occlusions, and varying viewpoints. This problem makes it mostly unsuitable for fire detection and segmentation since flames can have an unpredictable scale and shape. In this paper, we propose a method that detects and segments fire-regions with special considerations of their arbitrary sizes and shapes. Specifically, our approach uses a self-attention mechanism to augment spatial characteristics with temporal features, allowing the network to reduce its reliance on spatial factors like shape or size and take advantage of robust spatial-temporal dependencies. As a whole, our pipeline has two stages: In the first stage, we take out region proposals using Spatial-Temporal features, and in the second stage, we classify whether each region proposal is flame or not. Due to the scarcity of generous fire datasets, we adopt a transfer learning strategy to pre-train our classifier with the ImageNet dataset. Additionally, our Spatial-Temporal Network only requires semi-supervision, where it only needs one ground-truth segmentation mask per frame-sequence input. The experimental results of our proposed method significantly outperform the state-of-the-art fire detection with a 2 ~ 4% relative enhancement in F1-score for large scale fires and a nearly ~ 60% relative improvement for small fires at a very early stage. Authoren_US
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineers Inc.en_US
dc.sourceIEEE Accessen_US
dc.subjectClassification (of information)en_US
dc.subjectDisaster preventionen_US
dc.subjectFire detectorsen_US
dc.subjectFiresen_US
dc.subjectImage segmentationen_US
dc.subjectNeural networksen_US
dc.subjectObject detectionen_US
dc.subjectObject recognitionen_US
dc.subjectSecurity systemsen_US
dc.subjectConvolutional neural networken_US
dc.subjectEarly detectionen_US
dc.subjectFire detectionen_US
dc.subjectFire segmentationsen_US
dc.subjectSemi-superviseden_US
dc.subjectSmall-sized fireen_US
dc.subjectSpatial temporalsen_US
dc.subjectSpatio-temporalen_US
dc.subjectVideo fire segmentationen_US
dc.subjectVideo surveillanceen_US
dc.subjectDisastersen_US
dc.titleSpatio-Temporal Self-Attention Network for Fire Detection and Segmentation in Video Surveillanceen_US
dc.typeJournal Articleen_US
dc.rights.licenseAll Open Access, Gold-
Appears in Collections:Department of Mathematics

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Altmetric Badge: