Nanomaterial-Enhanced Robotic System for Real-Time Crowd, Violence, Safety Detection Using YOLOv8 in oil and chemical Industry

Authors

  • Dr.Kamaleshwar T Dept. Of Computer Science and Engineering Vel Tech Rangarajan Dr. Sagunthala R&D Institute of Science and Technology Chennai, India Author https://orcid.org/0000-0003-4612-1046
  • Lokesh Kumar Paila Dept. Of Computer Science and Engineering Vel Tech Rangarajan Dr. Sagunthala R&D Institute of Science and Technology Chennai, India Author https://orcid.org/0009-0001-3101-7783
  • Sneja Godi Dept. Of Computer Science and Engineering Vel Tech Rangarajan Dr. Sagunthala R&D Institute of Science and Technology Chennai, India Author https://orcid.org/0009-0004-6410-7875
  • Christo Ananth Faculty of Artificial Intelligence and Digital Technologies, Samarkand State University,Uzbekistan Author https://orcid.org/0000-0001-6979-584X
  • V.H. Abdullayev Azerbaijan State Oil and İndustry University, Baku, Azerbaijan Author
  • R.G. Abaszade Azerbaijan State Oil and İndustry University, Baku, Azerbaijan Author

DOI:

https://doi.org/10.62486/978-9915-9851-0-7_202633

Keywords:

YOLOv8, Real-Time Surveillance, Crowd and violence detection, oil and chemical Industry, Computer Vision, Robotic Systems, Nanomaterials in robotics

Abstract

Quick and accurate recognition of suspicious and abnormal density variations and violent acts is necessary to ensure the safety of the population in high-density and sensitive areas. This publication introduces a nanomaterial-crowded robotic surveillance system, which combines the improved detectives of small-scale materials with the YOLOv8 deep learning framework in terms of detecting crowds and violence in real time. The nanomaterial-based sensors enhance the responsiveness of the robotic system and make it sensitive to imaging, noise-sensitive when acquiring signals and more adaptable to its environmental capacity. Live video streams are targeted by the YOLOv8 processor and suspicious activities are detected with high accuracy, and the robotic platform will automatically cause audible alarms and send email notifications to the relevant authorities. The proposed system will overcome the drawbacks of manual monitoring, where there is increased accuracy in detection, faster reaction, and scalability of the operations. Experimental testing of various real-world systems proves that the integration of nanomaterial-based sensor technology with the contemporary computer vision can greatly improve the reliability of the system, and it is a proactive and efficient solution to the automation of the public safety and also safety enviroment in oil and chemical Industry.

References

[1] Agnihotri, A. (2021). DeepFake Detection using Deep Neural Networks (Doctoral dissertation, Dublin, National College of Ireland).

[2] Abdullah, M., Ahmad, M., & Han, D. (2020, January). Facial expression recognition in videos: A CNN-LSTM-based model for video classification. In 2020 International Conference on Electronics, Information, and Communication (ICEIC) (app. 1–3). IEEE.

[3] Ghadekar, P., Adsare, T., Agrawal, N., Dharmik, T., Patil, A., & Zod, S. (2023, September). Ensemble learning approach for anomaly detection in crowd scene classification. In IEEE Fifth International Conference on Advances in Electronics, Computers and Communications (ICAECC) (pp. 01–11). IEEE.

[4] Helode, A., Yadav, A., Verma, V. P., & Srinivasa, K. G. (2024, June). Fusion of Machine Learning and Deep Learning: A Hybrid Approach for Deepfake Detection. In 2024 15th International Conference on Computing Communication and Networking Technologies (ICCCNT) (pp. 1-6). IEEE.

[5] Mihanpour, A., Rashti, M. J., & Alavi, S. E. (2020, April). Human action recognition in video using DB-LSTM and ResNet. In 6th International Conference on Web Research (ICWR) (pp. 133–138). IEEE.

[6] Qazi, N., & Ahmed, I. (2024, July). Enhancing Authenticity Verification with Transfer Learning and Ensemble Techniques in Facial Feature-Based Deepfake Detection. In 2024 14th International Conference on Pattern Recognition Systems (ICPRS) (pp. 1-6). IEEE.

[7] Rajeev, A., & Raviraj, P. (2024). Performance evaluation of deep learning models for detecting deep fakes. International Journal of Systematic Innovation, 8(1), 49-62.

[8] Raza, A., Munir, K., & Almutairi, M. (2022). A novel deep learning approach for deepfake image detection. Applied Sciences, 12(19), 9820.

[9] Saikia, P., Dholaria, D., Yadav, P., Patel, V., & Roy, M. (2022, July). A hybrid CNN-LSTM model for video deepfake detection by leveraging optical flow features. In 2022 International Joint Conference on Neural Networks (IJCNN) (pp. 1-7). IEEE.

[10] Sánchez-Caballero, A., Fuentes-Jiménez, D., & Losada-Gutiérrez, C. (2022, October). Real-time human action recognition using raw depth video-based recurrent neural networks. Multimedia Tools and Applications, 82(11), 16213–16235.

[11] Singh, R. P., Sree, N. H., Reddy, K. L. S. P., & Jashwanth, K. (2024). Convergence of Deep Learning and Forensic Methodologies Using Self-attention Integrated EfficientNet Model for Deep Fake Detection. SN Computer Science, 5(8), 1-12.

[12] Jayaraj, R., Pushpalatha, A., Sangeetha, K., Kamaleshwar, T., Udhaya Shree, S., & Damodaran, D. (2024). Intrusion detection based on phishing detection with machine learning. Measurement: Sensors, 31, Article 101003. https://doi.org/10.1016/j.measen.2023.101003.

[13] Z. Tong, Y. Chen, Z. Xu, and R. Yu, “Wise-IoU: Bounding box regression loss with dynamic focusing mechanism,” arXiv preprint arXiv:2301.10051, 2023.

[14] C. Liu, K. Wang, Q. Li, F. Zhao, and H. Ma, “Powerful-IoU: More straightforward and faster bounding box regression loss with a nonmonotonic focusing mechanism,” Neural Networks, vol. 170, pp. 276–284, 2024.

[15] D. Misra, T. Nalamada, A. U. Arasanipalai, and Q. Hou, “Rotate to Attend: Convolutional Triplet Attention Module,” in Proc. IEEE/CVF Winter Conf. Applications of Computer Vision (WACV), 2021, pp. 3139–3148.

[16] S. Woo, J. Park, J. Lee, and I. Kweon, “CBAM: Convolutional block attention module,” in Proc. Eur. Conf. Computer Vision (ECCV), 2018, pp. 3–19.

[17] L. Yang, R. Zhang, Y. Li, and X. Xie, “SimAM: A Simple, Parameter-Free Attention Module for Convolutional Neural Networks,” in Proc. Int. Conf. Machine Learning, 2021, pp. 11863–11874.

[18] H. Liu, R. Liu, X. Fan, and D. Huang, “Polarized self-attention: Towards high-quality pixel-wise regression,” Neurocomputing, vol. 506, pp. 158–167, 2022.

[19] L. Zhu, X. Wang, Z. Ke, and R. Lau, “Biformer: Vision transformer with bi-level routing attention,” in Proc. IEEE/CVF Conf. Computer Vision and Pattern Recognition (CVPR), 2023, pp. 10323–10333.

[20] X. Li, H. Xu, and J. Yang, “Spatial group-wise enhance: Improving semantic feature learning in convolutional networks,” arXiv preprint arXiv:1905.09646, 2019.

[21] A. Vaswani et al., “Attention Is All You Need,” in Advances in Neural Information Processing Systems, vol. 30, 2017.

[22] J. Yang, C. Li, X. Dai, and J. Gao, “Focal modulation networks,” Advances in Neural Information Processing Systems, pp. 4203–4217, 2022.

[23] H. Wang, P. Guo, P. Zhou, and L. Xie, “MLCA-AVSR: Multi-Layer Cross Attention Fusion based Audio-Visual Speech Recognition,” in Proc. IEEE Int. Conf. Acoustics, Speech and Signal Processing (ICASSP), 2024, pp. 8150–8154.

[24] Y. Li, Q. Hou, Z. Zheng, M. Cheng, and J. Yang, “Large selective kernel network for remote sensing object detection,” in Proc. IEEE/CVF Int. Conf. Computer Vision (ICCV), 2023, pp. 16794–16805.

[25] R. Azad et al., “Beyond self-attention: Deformable large kernel attention for medical image segmentation,” in Proc. IEEE Winter Conf. Applications of Computer Vision (WACV), 2024, pp. 1287–1297.

[26] X. Li, W. Wang, H. Xu, and J. Yang, “Selective kernel networks,” in Proc. IEEE/CVF Conf. Computer Vision and Pattern Recognition (CVPR), 2019, pp. 510–519.

[27] Chen Y, Kalantidis Y, Li J, Yan S, Feng J (2018) A²-nets: Double attention networks. Advances in neural information processing systems 31.

[28] Li Y, Yao T, Pan Y, Mei T (2022) Contextual Transformer Networks for Visual Recognition. IEEE T Pattern Anal 45: 1489–1500.

[29] Lee Y, Park J (2020) Centermask: Real-time anchor-free instance segmentation. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 13906–13915. https://doi.org/10.1109/CVPR42600.2020.01392

[30] Xia Z, Pan X, Song S, Li LE, Huang G (2022) Vision transformer with deformable attention. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 4794–4803. https://doi.org/10.1109/CVPR52688.2022.00475

[31] Ouyang D, He S, Zhang G, Luo M, Guo H, Zhan J, et al. (2023) Efficient multi-scale attention module with cross-spatial learning. ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 1–5. IEEE. https://doi.org/10.1109/ICASSP49357.2023.10069516

[32] Hu L, Li Y (2021) Micro-YOLO: Exploring Efficient Methods to Compress CNN based Object Detection Model. In ICAART (2), 151–158. https://doi.org/10.5220/0010234410151058

[33] Li C, Li L, Jiang H, Weng K, Geng Y, Li L, et al. (2022) YOLOv6: A single-stage object detection framework for industrial applications. arXiv preprint arXiv:2209.02976.

[34] Carion N, Massa F, Synnaeve G, Usunier N, Kirillov A, Zagoruyko S (2020) End-to-End Object Detection with Transformers. European conference on computer vision, 213–229. Cham: Springer International Publishing. https://doi.org/10.1007/978-3-030-58452-8_13

[35] Duan K, Bai S, Xie L, Qi H, Huang Q, Tian Q (2019) Centernet: Keypoint triplets for object detection. Proceedings of the IEEE/CVF international conference on computer vision, 6569–6578. https://doi.org/10.1109/ICCV.2019.00667

Downloads

Published

2026-01-01

How to Cite

1.
T D, Kumar Paila L, Godi S, Ananth C, Abdullayev V, Abaszade R. Nanomaterial-Enhanced Robotic System for Real-Time Crowd, Violence, Safety Detection Using YOLOv8 in oil and chemical Industry. Superintelligence Series [Internet]. 2026 Jan. 1 [cited 2026 Jan. 14];3:33. Available from: https://sis.southam.pub/index.php/sis/article/view/33