An Approach for Improving Accuracy and Optimizing Resource Usage for Violence Detection in Surveillance Cameras in IoT systems

Hoang-Tu Vo, Phuc Pham Tien, Nhon Nguyen Thien, Kheo Chau Mui

Abstract


Smart farming that uses information and communication technology is developed as a critical technique to address the challenges related to agricultural production, environmental effects (climate change), food security, and supply chain. The recent statistics reveal that the world's population has been increased significantly, which is expected to reach 7.7 billion. It is essential to achieve a significant rise in food output to meet the requirement of such a massive growth of population. However, due to the natural conditions and a variety of plant illnesses, food productivity and farms are reduced. In order to diagnose food diseases in farming, new technologies like the Internet of Things and artificial intelligence are now essential. To this end, the research paper introduces a novel artificial intelligence model represented by a twelvelayer deep convolution neural network to identify and classify plant image diseases. 38 distinct types of plant leaf photos are used for training and testing the suggested model, which are obtained by adjusting different parameters such as (a) hyperparameters; (b) coevolutionary layers; (c) and pooling layers in number. The proposed model consists of an extractor and classifier of functions. The first section involves three phases, i.e., it consists of two convolution layers and a maximum pooling layer for each phase. The second section consists of three levels: flattening, hidden, and output layers. The proposed model is compared with LeNet, VGG16, AlexNet, and Inception v3, which are considered state-of-the-art pre-trained models. The results demonstrate that the accuracy of LeNet, VGG16, AlexNet, and Inception v3 is given as 89%, 93%, 96.11%, and 97.6%, respectively. The findings provided in this research show that the suggested model outperforms state-of-the-art models in terms of training speed and computing time. Also, the results show that the proposed model achieves a considerable improvement in terms of accuracy and the mean square error compared to the state-of-the-art methods. In particular, The outcomes demonstrate that the suggested model achieves a mean square error and prediction accuracy of 98.76% and 0.0580, respectively. The results also depict that the proposed model is more reliable, allows fast convergence time in obtaining the results, and requires only a small number of trained parameters to identify the plant diseases accurately.

Keywords


Violence detection; Surveillance camera; fine-tune model; BiLSTM; Deep learning

References


Dario Bacchini and Concetta Esposito. Growing up in violent contexts: differential effects of community, family, and school violence on child adjustment. Children and peace: From research to action, pages 157–171, 2020.

James A Mercy, Susan D Hillis, Alexander Butchart, Mark A Bellis, Catherine L Ward, Xiangming Fang, and Mark L Rosenberg. Interpersonal violence: global impact and paths to prevention. Injury prevention and environmental health. 3rd edition, 2017.

Linda L Dahlberg and Etienne G Krug. Violence a global public health problem. Ciencia & Saude Coletiva, 11(2):277–292, 2006.

Duarte Duque, Henrique Santos, and Paulo Cortez. Prediction of abnormal behaviors for intelligent video surveillance systems. In 2007 IEEE Symposium on Computational Intelligence and Data Mining, pages 362–367. IEEE, 2007.

Jong Sun Kim, Dong Hae Yeom, and Young Hoon Joo. Fast and robust algorithm of tracking multiple moving objects for intelligent video surveillance systems. IEEE Transactions on Consumer Electronics, 57(3):1165–1170, 2011.

Chao Huang, Zhihao Wu, Jie Wen, Yong Xu, Qiuping Jiang, and Yaowei Wang. Abnormal event detection using deep contrastive learning for intelligent video surveillance system. IEEE Transactions on Industrial Informatics, 18(8):5171–5179, 2021.

Bangpeng Yao, Xiaoye Jiang, Aditya Khosla, Andy Lai Lin, Leonidas Guibas, and Li Fei-Fei. Human action recognition by learning bases of action attributes and parts. In 2011 International conference on computer vision, pages 1331–1338. IEEE, 2011

Josephine Sullivan and Stefan Carlsson. Recognizing and tracking human action. In Computer Vision—ECCV 2002: 7th European Conference on Computer Vision Copenhagen, Denmark, May 28–31, 2002 Proceedings, Part I 7, pages 629–644. Springer, 2002.

Muhammad Attique Khan, Kashif Javed, Sajid Ali Khan, Tanzila Saba, Usman Habib, Junaid Ali Khan, and Aaqif Afzaal Abbasi. Human action recognition using fusion of multiview and deep features: an application to video surveillance. Multimedia tools and applications, pages 1–27, 2020.

Mark Sandler, Andrew Howard, Menglong Zhu, Andrey Zhmoginov, and Liang-Chieh Chen. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4510–4520, 2018.

Mike Schuster and Kuldip K Paliwal. Bidirectional recurrent neural networks. IEEE transactions on Signal Processing, 45(11):2673–2681, 1997.

Enrique Bermejo Nievas, Oscar Deniz Suarez, Gloria Bueno Garcıa, and Rahul Sukthankar. Violence detection in video using computer vision techniques. In Computer Analysis of Images and Patterns: 14th International Conference, CAIP 2011, Seville, Spain, August 29-31, 2011, Proceedings, Part II 14, pages 332–339. Springer, 2011.

Liang-Hua Chen, Hsi-Wen Hsu, Li-Yun Wang, and Chih-Wen Su. Violence detection in movies. In 2011 Eighth International Conference Computer Graphics, Imaging and Visualization, pages 119–124. IEEE, 2011.

Yuan Gao, Hong Liu, Xiaohu Sun, Can Wang, and Yi Liu. Violence detection using oriented violent flows. Image and vision computing, 48:37–41, 2016.

Tao Zhang, Zhijie Yang, Wenjing Jia, Baoqing Yang, Jie Yang, and Xiangjian He. A new method for violence detection in surveillance scenes. Multimedia Tools and Applications, 75:7327–7349, 2016.

Romas Vijeikis, Vidas Raudonis, and Gintaras Dervinis. Efficient violence detection in surveillance. Sensors, 22(6):2216, 2022.

Samee Ullah Khan, Ijaz Ul Haq, Seungmin Rho, Sung Wook Baik, and Mi Young Lee. Cover the violence: A novel deep-learning-based approach towards violence-detection in movies. Applied Sciences, 9(22):4963, 2019.

Javad Mahmoodi and Afsane Salajeghe. A classification method based on optical flow for violence detection. Expert systems with applications, 127:121–127, 2019.

Fath U Min Ullah, Amin Ullah, Khan Muhammad, Ijaz Ul Haq, and Sung Wook Baik. Violence detection using spatiotemporal features with 3d convolutional neural network. Sensors, 19(11):2472, 2019.

Shakil Ahmed Sumon, Raihan Goni, Niyaz Bin Hashem, Tanzil Shahria, and Rashedur M Rahman. Violence detection by pretrained modules with different deep learning approaches. Vietnam Journal of Computer Science, 7(01):19–40, 2020.

Rohit Halder and Rajdeep Chatterjee. Cnn-bilstm model for violence detection in smart surveillance. SN Computer science, 1(4):201, 2020.

Mujtaba Asad, Jie Yang, Jiang He, Pourya Shamsolmoali, and Xiangjian He. Multi-frame feature-fusion-based model for violence detection. The Visual Computer, 37:1415–1431, 2021.

Manan Sharma and Rishabh Baghel. Video surveillance for violence detection using deep learning. In Advances in Data Science and Management: Proceedings of ICDSM 2019, pages 411–420. Springer, 2020.

Simone Accattoli, Paolo Sernani, Nicola Falcionelli, Dagmawi Neway Mekuria, and Aldo Franco Dragoni. Violence detection in videos by combining 3d convolutional neural networks and support vector machines. Applied Artificial Intelligence, 34(4):329–344, 2020.

Real life violence situations dataset, available online: https://www.kaggle.com/datasets/mohamedmustafa/real-lifeviolencesituations-dataset.

Mohamed Mostafa Soliman, Mohamed Hussein Kamal, Mina Abd ElMassih Nashed, Youssef Mohamed Mostafa, Bassel Safwat Chawky, and Dina Khattab. Violence recognition from videos using deep learning techniques. In 2019 Ninth International Conference on Intelligent Computing and Information Systems (ICICIS), pages 80–85, 2019.

Andrew G Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, and Hartwig Adam. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861, 2017.

Gao Huang, Zhuang Liu, Laurens Van Der Maaten, and Kilian Q Weinberger. Densely connected convolutional networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4700–4708, 2017.

Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016

Fernando J Rend´on-Segador, Juan A ´Alvarez-Garc´ıa, Fernando Enr´ıquez, and Oscar Deniz. Violencenet: Dense multi-head self-attention with bidirectional convolutional lstm for detecting violence. Electronics, 10(13):1601, 2021

Swathikiran Sudhakaran and Oswald Lanz. Learning to detect violent videos using convolutional long short-term memory. In 2017 14th IEEE international conference on advanced video and signal based surveillance (AVSS), pages 1–6. IEEE, 2017.

S¸ eymanur Aktı, G¨ozde Ays¸e Tataro˘glu, and Hazım Kemal Ekenel. Vision-based fight detection from surveillance cameras. In 2019 Ninth International Conference on Image Processing Theory, Tools and Applications (IPTA), pages 1–6. IEEE, 2019.

Ji Li, Xinghao Jiang, Tanfeng Sun, and Ke Xu. Efficient violence detection using 3d convolutional neural networks. In 2019 16th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), pages 1–8. IEEE, 2019.


Full Text: PDF

Refbacks

  • There are currently no refbacks.


 

Indonesian Journal of Electrical Engineering and Informatics (IJEEI)
ISSN 2089-3272

Creative Commons Licence

This work is licensed under a Creative Commons Attribution 4.0 International License.

web analytics
View IJEEI Stats

Error. Page cannot be displayed. Please contact your service provider for more details. (3)