RECOGNITION OF HUMAN FACES IN VIDEO CONFERENCE APPLICATIONS USING THE CNN PIPELINE
The COVID-19 pandemic has forced daily face-to-face activities to be carried out online using video conferencing applications. To record participant participation in meetings using a video conference application, an online form application is used. However, participants sometimes do not see this and are often missed due to the large number of incoming chats. Therefore, the use of face detection for attendance using a combination of CNN to detect all the faces in a video conference using YOLO Face and CNN to recognize the owner of a face using Smaller VGG in a pipeline will make it easier to recognize participants who are present at the video conference. The results of the Smaller VGG training are obtained, namely the loss value of 0.059, the accuracy value is 0.995, the recall value is 0.994, the precision value is 0.996. Meanwhile, for the validation phase of the model, the loss value is 0.497, the accuracy value is 0.979, the recall value is 0.979 and the precision value is 0.981. In terms of training duration, the smaller VGG has a duration of 4 minutes and 16 seconds. The Smaller VGG model was combined with YOLO to create a CNN pipeline and was successful in recognizing the faces of video conference participants
K. B. Pranav and J. Manikandan, “Design and Evaluation of a Real-Time Face Recognition System using Convolutional Neural Networks,” Procedia Comput. Sci., vol. 171, no. 2019, pp. 1651–1659, 2020, doi: 10.1016/j.procs.2020.04.177.
B. Tej Chinimilli, A. Anjali, A. Kotturi, V. Reddy Kaipu, and J. Varma Mandapati, “Face Recognition based Attendance System using Haar Cascade and Local Binary Pattern Histogram Algorithm,” Proc. 4th Int. Conf. Trends Electron. Informatics, ICOEI 2020, no. Icoei, pp. 701–704, 2020, doi: 10.1109/ICOEI48184.2020.9143046.
H. Yang and X. Han, “Face recognition attendance system based on real-time video processing,” IEEE Access, vol. 8, pp. 159143–159150, 2020, doi: 10.1109/ACCESS.2020.3007205.
U. Jayaraman, P. Gupta, S. Gupta, G. Arora, and K. Tiwari, “Recent development in face recognition,” Neurocomputing, vol. 408, pp. 231–245, 2020, doi: 10.1016/j.neucom.2019.08.110.
B. Li and D. Lima, “Facial expression recognition via ResNet-50,” Int. J. Cogn. Comput. Eng., vol. 2, no. February, pp. 57–64, 2021, doi: 10.1016/j.ijcce.2021.02.002.
Z. Chen and P. H. Ho, “Global-connected network with generalized ReLU activation,” Pattern Recognit., vol. 96, 2019, doi: 10.1016/j.patcog.2019.07.006.
S. Khan, M. H. Javed, E. Ahmed, S. A. A. Shah, and S. U. Ali, “Facial recognition using convolutional neural networks and implementation on smart glasses,” 2019 Int. Conf. Inf. Sci. Commun. Technol. ICISCT 2019, pp. 1–6, 2019, doi: 10.1109/CISCT.2019.8777442.
Y. Tian, “Artificial Intelligence Image Recognition Method Based on Convolutional Neural Network Algorithm,” IEEE Access, vol. 8, pp. 125731–125744, 2020, doi: 10.1109/ACCESS.2020.3006097.
J. Tang, Q. Su, B. Su, S. Fong, W. Cao, and X. Gong, “Parallel ensemble learning of convolutional neural networks and local binary patterns for face recognition,” Comput. Methods Programs Biomed., vol. 197, p. 105622, 2020, doi: 10.1016/j.cmpb.2020.105622.
A. Bochkovskiy, C. Y. Wang, and H. Y. M. Liao, “YOLOv4: Optimal Speed and Accuracy of Object Detection,” arXiv, 2020.
W. Chen, H. Huang, S. Peng, C. Zhou, and C. Zhang, “YOLO-face: a real-time face detector,” Vis. Comput., 2020, doi: 10.1007/s00371-020-01831-7.
J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You Only Look Once: Unified, Real-Time Object Detection.” Accessed: May 18, 2021. [Online]. Available: http://pjreddie.com/yolo/
A. A. Wahid, “Analisis Metode Waterfall Untuk Pengembangan Sistem Informasi,” J. Ilmu-ilmu Inform. dan Manaj. STMIK, pp. 1–5, 2020.
D. C. Khrisne and I. A. Suyadnya, “Indonesian Herbs and Spices Recognition using Smaller VGGNet-like Network,” 2018 Int. Conf. Smart Green Technol. Electr. Inf. Syst., vol. 4, pp. 221–224, 2018.
A. Verma, P. Singh, and J. S. Rani Alex, “Modified Convolutional Neural Network Architecture Analysis for Facial Emotion Recognition,” Int. Conf. Syst. Signals, Image Process., vol. 2019-June, pp. 169–173, 2019, doi: 10.1109/IWSSIP.2019.8787215.
Q. Shi, W. Li, F. Zhang, W. Hu, X. Sun, and L. Gao, “Deep CNN with Multi-Scale Rotation Invariance Features for Ship Classification,” IEEE Access, vol. 6, pp. 38656–38668, 2018, doi: 10.1109/ACCESS.2018.2853620.
Copyright (c) 2022 Evan Tanuwijaya, Reinaldo Lewis Lordianto, Reiner Anggriawan Jasin
This work is licensed under a Creative Commons Attribution 4.0 International License.