Volume 3, Issue 1, ICCK Transactions on Sensing, Communication, and Control
Volume 3, Issue 1, 2026
Submit Manuscript Edit a Special Issue
Article QR Code
Article QR Code
Scan the QR code for reading
Popular articles
ICCK Transactions on Sensing, Communication, and Control, Volume 3, Issue 1, 2026: 15-26

Free to Read | Research Article | 13 February 2026
Intelligent Fire Recognition for Surveillance Control Using Cascaded Multi-Scale Attention Framework
1 Department of Computer Science, Coventry University, Coventry CV1 5FB, United Kingdom
2 BRAINS Institute Peshawar, Peshawar 25000, Pakistan
* Corresponding Author: Faryal Zahoor, [email protected]
ARK: ark:/57805/tscc.2025.862776
Received: 17 December 2025, Accepted: 12 January 2026, Published: 13 February 2026  
Abstract
Fire incidents cause devastating environmental damage and human casualties, necessitating robust automated detection systems. Existing fire recognition methods struggle with visual ambiguities, illumination variations, and computational constraints, while current attention mechanisms lack hierarchical integration for comprehensive feature refinement. We propose a cascaded multi-attention architecture that combines Multi-Scale Strip Attention (MSSA), Optimized Spatial Attention (OSA), and the Convolutional Block Attention Module (CBAM) to enhance fire detection. MSSA employs three-scale orthogonal strip pooling to capture fire patterns across varying spatial extents through horizontal and vertical feature decomposition. OSA employs optimized convolutions rather than conventional kernels, reducing computational cost while maintaining spatial localization accuracy. CBAM applies sequential channel- and spatial-level attention for comprehensive feature recalibration. Operating on the EfficientNetB7 backbone, the cascaded design progressively refines representations through complementary mechanisms. Systematic ablation studies validate individual module contributions, while Grad-CAM visualizations confirm precise localization of fire regions. Extensive experiments on the FD and BoWFire datasets demonstrate superior performance compared to state-of-the-art approaches.

Graphical Abstract
Intelligent Fire Recognition for Surveillance Control Using Cascaded Multi-Scale Attention Framework

Keywords
fire detection
strip pooling
surveillance monitoring
intelligent sensing
safety control
visual recognition

Data Availability Statement
Data will be made available on request.

Funding
This work was supported without any funding.

Conflicts of Interest
The authors declare no conflicts of interest.

AI Use Statement
The authors declare that no generative AI was used in the preparation of this manuscript.

Ethical Approval and Consent to Participate
Not applicable.

References
  1. Urza, A. K., Hanberry, B. B., & Jain, T. B. (2023). Landscape-scale fuel treatment effectiveness: lessons learned from wildland fire case studies in forests of the western United States and Great Lakes region. Fire Ecology, 19(1), 1–12.
    [CrossRef]   [Google Scholar]
  2. Tan, C. & Feng, Z. (2023). Mapping Forest Fire Risk Zones Using Machine Learning Algorithms in Hunan Province, China. Sustainability, 15(7), 6292.
    [CrossRef]   [Google Scholar]
  3. U.S. Fire Administration. (2025). Residential fire estimate summaries (2014–2023). U.S. Department of Homeland Security, Federal Emergency Management Agency. Retrieved from https://www.usfa.fema.gov/statistics/residential-fires/
    [Google Scholar]
  4. Keith, D. A., Allen, S. P., Gallagher, R. V., Mackenzie, B. D., Auld, T. D., Barrett, S., ... & Tozer, M. G. (2022). Fire‐related threats and transformational change in Australian ecosystems. Global Ecology and Biogeography, 31(10), 2070-2084.
    [CrossRef]   [Google Scholar]
  5. Elidolu, G., Akyuz, E., Arslan, O., & Arslanoğlu, Y. (2022). Quantitative failure analysis for static electricity-related explosion and fire accidents on tanker vessels under fuzzy bow-tie CREAM approach. Engineering Failure Analysis, 131, 105917.
    [CrossRef]   [Google Scholar]
  6. Swain, D. L., Abatzoglou, J. T., Kolden, C., Shive, K., Kalashnikov, D. A., Singh, D., & Smith, E. (2023). Climate change is narrowing and shifting prescribed fire windows in western United States. Communications Earth & Environment, 4(1), 340.
    [CrossRef]   [Google Scholar]
  7. Ahmed, I., & Ledger, K. (2023). Lessons from the 2019/2020 ‘black summer bushfires’ in Australia. International journal of disaster risk reduction, 96, 103947.
    [CrossRef]   [Google Scholar]
  8. Richards, L., Brew, N., & Smith, L. (2020, March 12). 2019–20 Australian bushfires—frequently asked questions: a quick guide. Parliament of Australia. Retrieved from https://www.aph.gov.au/About_Parliament/Parliamentary_departments/Parliamentary_Library/Research/Quick_Guides/2019-20/AustralianBushfires
    [Google Scholar]
  9. Ghali, R., & Akhloufi, M. A. (2024). Deep learning approach for wildland fire recognition using RGB and thermal infrared aerial image. Fire, 7(10), 343.
    [CrossRef]   [Google Scholar]
  10. Celik, T., Ozkaramanli, H., & Demirel, H. (2007, April). Fire pixel classification using fuzzy logic and statistical color model. In 2007 IEEE International Conference on Acoustics, Speech and Signal Processing-ICASSP'07 (Vol. 1, pp. I-1205). IEEE.
    [CrossRef]   [Google Scholar]
  11. Khan, T., Khan, Z. A., & Choi, C. (2025). Enhancing real-time fire detection: An effective multi-attention network and a fire benchmark. Neural Computing and Applications, 37(18), 11693-11707.
    [CrossRef]   [Google Scholar]
  12. Celik, T., Demirel, H., Ozkaramanli, H., & Uyguroglu, M. (2007). Fire detection using statistical color model in video sequences. Journal of Visual Communication and Image Representation, 18(2), 176-185.
    [CrossRef]   [Google Scholar]
  13. Celik, T., & Demirel, H. (2009). Fire detection in video sequences using a generic color model. Fire safety journal, 44(2), 147-158.
    [CrossRef]   [Google Scholar]
  14. Yar, H., Khan, Z. A., Ullah, F. U. M., Ullah, W., & Baik, S. W. (2023). A modified YOLOv5 architecture for efficient fire detection in smart cities. Expert Systems with Applications, 231, 120465.
    [CrossRef]   [Google Scholar]
  15. Yar, H., Hussain, T., Agarwal, M., Khan, Z. A., Gupta, S. K., & Baik, S. W. (2022). Optimized dual fire attention network and medium-scale fire classification benchmark. IEEE Transactions on Image Processing, 31, 6331-6343.
    [CrossRef]   [Google Scholar]
  16. Tsalera, E., Papadakis, A., Voyiatzis, I., & Samarakou, M. (2023). CNN-based, contextualized, real-time fire detection in computational resource-constrained environments. Energy Reports, 9, 247-257.
    [CrossRef]   [Google Scholar]
  17. Mao, W., Wang, W., Dou, Z., & Li, Y. (2018). Fire recognition based on multi-channel convolutional neural network. Fire technology, 54(2), 531-554.
    [CrossRef]   [Google Scholar]
  18. Woo, S., Park, J., Lee, J.-Y., & Kweon, I. S. (2018). Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV) (pp. 3–19).
    [CrossRef]   [Google Scholar]
  19. Chen, T. H., Wu, P. H., & Chiou, Y. C. (2004, October). An early fire-detection method based on image processing. In 2004 International Conference on Image Processing, 2004. ICIP'04. (Vol. 3, pp. 1707-1710). IEEE.
    [CrossRef]   [Google Scholar]
  20. Angayarkkani, K., & Radhakrishnan, N. (2009). Efficient forest fire detection system: a spatial data mining and image processing based approach. International Journal of Computer Science and Network Security, 9(3), 100-107.
    [Google Scholar]
  21. Chino, D. Y., Avalhais, L. P., Rodrigues, J. F., & Traina, A. J. (2015, August). Bowfire: detection of fire in still images by integrating pixel color and texture analysis. In 2015 28th SIBGRAPI conference on graphics, patterns and images (pp. 95-102). IEEE.
    [CrossRef]   [Google Scholar]
  22. Hussain, T., Ullah, F. U. M., Khan, S. U., Ullah, A., Haroon, U., Muhammad, K., ... & de Albuquerque, V. H. C. (2022). Deep learning assists surveillance experts: Toward video data prioritization. IEEE Transactions on Industrial Informatics, 19(7), 7946-7956.
    [CrossRef]   [Google Scholar]
  23. Cheng, G., Chen, X., Wang, C., Li, X., Xian, B., & Yu, H. (2024). Visual fire detection using deep learning: A survey. Neurocomputing, 596, 127975.
    [CrossRef]   [Google Scholar]
  24. Yar, H., Khan, Z. A., Hussain, T., & Baik, S. W. (2024). A modified vision transformer architecture with scratch learning capabilities for effective fire detection. Expert Systems with Applications, 252, 123935.
    [CrossRef]   [Google Scholar]
  25. Khan, Z. A., Hussain, T., Ullah, F. U. M., Gupta, S. K., Lee, M. Y., & Baik, S. W. (2022). Randomly initialized CNN with densely connected stacked autoencoder for efficient fire detection. Engineering Applications of Artificial Intelligence, 116, 105403.
    [CrossRef]   [Google Scholar]
  26. Lee, W., Kim, S., Lee, Y. T., Lee, H. W., & Choi, M. (2017, January). Deep neural networks for wild fire detection with unmanned aerial vehicle. In 2017 IEEE international conference on consumer electronics (ICCE) (pp. 252-253). IEEE.
    [CrossRef]   [Google Scholar]
  27. Sharma, J., Granmo, O. C., Goodwin, M., & Fidje, J. T. (2017, August). Deep convolutional neural networks for fire detection in images. In International conference on engineering applications of neural networks (pp. 183-193). Cham: Springer International Publishing.
    [CrossRef]   [Google Scholar]
  28. Dunnings, A. J., & Breckon, T. P. (2018, October). Experimentally defined convolutional neural network architecture variants for non-temporal real-time fire detection. In 2018 25th IEEE international conference on image processing (ICIP) (pp. 1558-1562). IEEE.
    [CrossRef]   [Google Scholar]
  29. Ye, S., Feng, X., Zhang, T., Ma, X., Lin, S., Li, Z., ... & Wang, Y. (2019). Progressive dnn compression: A key to achieve ultra-high weight pruning and quantization rates using admm. arXiv preprint arXiv:1903.09769.
    [Google Scholar]
  30. Carreira-Perpinan, M. A., & Idelbayev, Y. (2018, June). `` Learning-Compression'' Algorithms for Neural Net Pruning. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 8532-8541). IEEE.
    [CrossRef]   [Google Scholar]
  31. Rastegari, M., Ordonez, V., Redmon, J., & Farhadi, A. (2016, September). Xnor-net: Imagenet classification using binary convolutional neural networks. In European conference on computer vision (pp. 525-542). Cham: Springer International Publishing.
    [CrossRef]   [Google Scholar]
  32. McDonnell, M. D. (2018). Training wide residual networks for deployment using a single bit for each weight. arXiv preprint arXiv:1802.08530.
    [Google Scholar]
  33. Vanhoucke, V., Senior, A., & Mao, M. Z. (2011, December). Improving the speed of neural networks on CPUs. In Proc. deep learning and unsupervised feature learning NIPS workshop (Vol. 1, No. 2011, p. 4).
    [Google Scholar]
  34. Wang, Z., Wang, Z., Zhang, H., & Guo, X. (2017, July). A novel fire detection approach based on CNN-SVM using tensorflow. In International conference on intelligent computing (pp. 682-693). Cham: Springer International Publishing.
    [CrossRef]   [Google Scholar]
  35. Wu, X., Lu, X., & Leung, H. (2017, October). An adaptive threshold deep learning method for fire and smoke detection. In 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC) (pp. 1954-1959). IEEE.
    [CrossRef]   [Google Scholar]
  36. Maksymiv, O., Rak, T., & Peleshko, D. (2017, February). Real-time fire detection method combining AdaBoost, LBP and convolutional neural network in video sequence. In 2017 14th international conference the experience of designing and application of CAD Systems in microelectronics (CADSM) (pp. 351-353). IEEE.
    [CrossRef]   [Google Scholar]
  37. Hu, C., Tang, P., Jin, W., He, Z., & Li, W. (2018, July). Real-time fire detection based on deep convolutional long-recurrent networks and optical flow method. In 2018 37th Chinese control conference (CCC) (pp. 9061-9066). IEEE.
    [CrossRef]   [Google Scholar]
  38. Shen, C., Qi, G. J., Jiang, R., Jin, Z., Yong, H., Chen, Y., & Hua, X. S. (2018). Sharp attention network via adaptive sampling for person re-identification. IEEE Transactions on Circuits and Systems for Video Technology, 29(10), 3016-3027.
    [CrossRef]   [Google Scholar]
  39. Ullah, W., Ullah, A., Hussain, T., Khan, Z. A., & Baik, S. W. (2021). An efficient anomaly recognition framework using an attention residual LSTM in surveillance videos. Sensors, 21(8), 2811.
    [CrossRef]   [Google Scholar]
  40. Majid, S., Alenezi, F., Masood, S., Ahmad, M., Gündüz, E. S., & Polat, K. (2022). Attention based CNN model for fire detection and localization in real-world images. Expert Systems with Applications, 189, 116114.
    [CrossRef]   [Google Scholar]
  41. Sun, H., & Yao, T. (2025). Multi-Scale Construction Site Fire Detection Algorithm with Integrated Attention Mechanism. Fire, 8(7), 257.
    [CrossRef]   [Google Scholar]
  42. Woo, S., Park, J., Lee, J. Y., & Kweon, I. S. (2018, September). CBAM: Convolutional Block Attention Module. In European Conference on Computer Vision (pp. 3-19). Cham: Springer International Publishing.
    [CrossRef]   [Google Scholar]
  43. Tan, M., & Le, Q. (2019, May). Efficientnet: Rethinking model scaling for convolutional neural networks. In International conference on machine learning (pp. 6105-6114). PMLR.
    [Google Scholar]
  44. Li, S., Yan, Q., & Liu, P. (2020). An efficient fire detection method based on multiscale feature extraction, implicit deep supervision and channel attention mechanism. IEEE Transactions on Image Processing, 29, 8467-8475.
    [CrossRef]   [Google Scholar]
  45. Yao, J., Lei, J., Zhou, J., & Liu, C. (2025). FG-YOLO: an improved YOLOv8 algorithm for real-time fire and smoke detection. Signal, Image and Video Processing, 19(5), 346.
    [CrossRef]   [Google Scholar]
  46. Foggia, P., Saggese, A., & Vento, M. (2015). Real-time fire detection for video-surveillance applications using a combination of experts based on color, shape, and motion. IEEE TRANSACTIONS on circuits and systems for video technology, 25(9), 1545-1556.
    [CrossRef]   [Google Scholar]
  47. Zhang, D., Han, S., Zhao, J., Zhang, Z., Qu, C., Ke, Y., & Chen, X. (2009, April). Image based forest fire detection using dynamic characteristics with artificial neural networks. In 2009 international joint conference on artificial intelligence (pp. 290-293). IEEE.
    [CrossRef]   [Google Scholar]

Cite This Article
APA Style
Khan, I. M., & Zahoor, F. (2026). Intelligent Fire Recognition for Surveillance Control Using Cascaded Multi-Scale Attention Framework. ICCK Transactions on Sensing, Communication, and Control, 3(1), 15–26. https://doi.org/10.62762/TSCC.2025.862776
Export Citation
RIS Format
Compatible with EndNote, Zotero, Mendeley, and other reference managers
RIS format data for reference managers
TY  - JOUR
AU  - Khan, Ikram Majeed
AU  - Zahoor, Faryal
PY  - 2026
DA  - 2026/02/13
TI  - Intelligent Fire Recognition for Surveillance Control Using Cascaded Multi-Scale Attention Framework
JO  - ICCK Transactions on Sensing, Communication, and Control
T2  - ICCK Transactions on Sensing, Communication, and Control
JF  - ICCK Transactions on Sensing, Communication, and Control
VL  - 3
IS  - 1
SP  - 15
EP  - 26
DO  - 10.62762/TSCC.2025.862776
UR  - https://www.icck.org/article/abs/TSCC.2025.862776
KW  - fire detection
KW  - strip pooling
KW  - surveillance monitoring
KW  - intelligent sensing
KW  - safety control
KW  - visual recognition
AB  - Fire incidents cause devastating environmental damage and human casualties, necessitating robust automated detection systems. Existing fire recognition methods struggle with visual ambiguities, illumination variations, and computational constraints, while current attention mechanisms lack hierarchical integration for comprehensive feature refinement. We propose a cascaded multi-attention architecture that combines Multi-Scale Strip Attention (MSSA), Optimized Spatial Attention (OSA), and the Convolutional Block Attention Module (CBAM) to enhance fire detection. MSSA employs three-scale orthogonal strip pooling to capture fire patterns across varying spatial extents through horizontal and vertical feature decomposition. OSA employs optimized convolutions rather than conventional kernels, reducing computational cost while maintaining spatial localization accuracy. CBAM applies sequential channel- and spatial-level attention for comprehensive feature recalibration. Operating on the EfficientNetB7 backbone, the cascaded design progressively refines representations through complementary mechanisms. Systematic ablation studies validate individual module contributions, while Grad-CAM visualizations confirm precise localization of fire regions. Extensive experiments on the FD and BoWFire datasets demonstrate superior performance compared to state-of-the-art approaches.
SN  - 3068-9287
PB  - Institute of Central Computation and Knowledge
LA  - English
ER  - 
BibTeX Format
Compatible with LaTeX, BibTeX, and other reference managers
BibTeX format data for LaTeX and reference managers
@article{Khan2026Intelligen,
  author = {Ikram Majeed Khan and Faryal Zahoor},
  title = {Intelligent Fire Recognition for Surveillance Control Using Cascaded Multi-Scale Attention Framework},
  journal = {ICCK Transactions on Sensing, Communication, and Control},
  year = {2026},
  volume = {3},
  number = {1},
  pages = {15-26},
  doi = {10.62762/TSCC.2025.862776},
  url = {https://www.icck.org/article/abs/TSCC.2025.862776},
  abstract = {Fire incidents cause devastating environmental damage and human casualties, necessitating robust automated detection systems. Existing fire recognition methods struggle with visual ambiguities, illumination variations, and computational constraints, while current attention mechanisms lack hierarchical integration for comprehensive feature refinement. We propose a cascaded multi-attention architecture that combines Multi-Scale Strip Attention (MSSA), Optimized Spatial Attention (OSA), and the Convolutional Block Attention Module (CBAM) to enhance fire detection. MSSA employs three-scale orthogonal strip pooling to capture fire patterns across varying spatial extents through horizontal and vertical feature decomposition. OSA employs optimized convolutions rather than conventional kernels, reducing computational cost while maintaining spatial localization accuracy. CBAM applies sequential channel- and spatial-level attention for comprehensive feature recalibration. Operating on the EfficientNetB7 backbone, the cascaded design progressively refines representations through complementary mechanisms. Systematic ablation studies validate individual module contributions, while Grad-CAM visualizations confirm precise localization of fire regions. Extensive experiments on the FD and BoWFire datasets demonstrate superior performance compared to state-of-the-art approaches.},
  keywords = {fire detection, strip pooling, surveillance monitoring, intelligent sensing, safety control, visual recognition},
  issn = {3068-9287},
  publisher = {Institute of Central Computation and Knowledge}
}

Article Metrics
Citations:

Crossref

0

Scopus

0

Web of Science

0
Article Access Statistics:
Views: 9
PDF Downloads: 3

Publisher's Note
ICCK stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and Permissions
Institute of Central Computation and Knowledge (ICCK) or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
ICCK Transactions on Sensing, Communication, and Control

ICCK Transactions on Sensing, Communication, and Control

ISSN: 3068-9287 (Online) | ISSN: 3068-9279 (Print)

Email: [email protected]

Portico

Portico

All published articles are preserved here permanently:
https://www.portico.org/publishers/icck/