Figure 16 Window for displaying some personalized information for a selected period
Source: created by authors
Discussion and conclusions
The paper examined the possibility of expanding the basic functionality of the Zoom online conference service by adding automatic identity verification and monitoring the activities of conference participants. As a result of the research, a Zoom conference application with additional features based on the Zoom Video SDK was developed.
The following issues were considered in this paper:
- the authors examine and compare various services for video conferences, study their functions and characteristics; investigate the relevance of adding functionality for monitoring the activities of participants; analyze what functionality is lacking in online conferences for use in the educational process, where there are a large number of participants and individual conferences are combined into a series of related events (lectures, workshops, seminars);
- the popularity of online conference services and technical capabilities for expanding the basic functionality is considered; it is recognized that Zoom is the most popular service and has sufficiently wide technical capabilities for adding functionality due to the existence of a large number of SDKs;
- the technical features of creating applications for the Zoom conference were studied; the capabilities of the Zoom Meeting SDK and the Zoom Video SDK were compared; it was decided to use the Video SDK because of the richer functionality for working with video streams and other Zoom events, as well as the lack of restrictions for designing the user interface;
- Computer Vision methods and libraries for solving the problems of face verification and emotional state recognition based on video stream analysis were investigated; it was decided to use the SSD model from the OpenCV library for face detection and the DeepFace library for verification and emotion recognition, where the ArcFace model was chosen to obtain face embeddings;
- the architecture of the application for identity verification and monitoring of participants' activities was designed, and the prototype application was implemented.
As a result of testing the developed application, it was shown that Zoom Video SDK meets the requirements of the task set for this work. We would like to draw special attention to the fact that Zoom Video SDK provides video streams of sufficient resolution for the selected models and libraries to be able to perform face detection, face verification, and emotion analysis.
The proposed extension of the Zoom online conference functionality to monitor activities and verify the participant's identity can significantly improve the quality of online events in general, and especially in the educational process:
- ensure the authenticity of participants, preventing possible fraud or misidentification. This feature is particularly important in an academic context where instructors need to know the real identities of a video conference participants, and ensure that grades are assigned to the correct students;
- use video stream analysis for compliance with the rules during exams will improve the fairness and transparency of the educational process;
- capturing the emotional state, the "raising hand" activities will allow teachers to understand how students respond to materials or teaching, teachers will be able to analyze and compare the effectiveness of different teaching strategies, adjust their approach to maximize student engagement and understanding of the material; will help in planning future events and understanding which topics or formats are most attractive to the audience;
- monitoring will help to collect data (dataset) that will allow:
- to investigate and identify the impact of the following factors on academic performance: learning activity (use of chat, "raising hands", use of emojis, etc.); presence in classes with an enabled or disabled camera (tracking the time of connection and disconnection from the conference, turning the camera on and off); emotional state (analyzing the video stream to determine the emotional state of students);
- to identify problems with students' learning promptly in order to have time to correct the situation (signs of problems can be detected in absence from the class for a certain period, presence with a disabled camera, presence with an enabled camera but having negative emotions, not using chat, "raising hand" and other tools:
- to plan future online events and understand which topics or formats attract the audience the most.
The developed application runs on the Windows platform. In the future, it would be desirable to develop interfaces that will allow users to connect from other platforms. It is also reasonable to study quantitative assessment of the quality of the verification module to allow us to formulate the requirements more clearly for the properties of the student's face in the frame, for example, requirements for its illumination, position, and size. Research on counteracting cheating of the verification module would also be useful (anti-spoofing).
In general, monitoring participants’ activities in online conferences is important to ensure the high quality and effectiveness of any type of event. The developed application is especially useful in the educational process. It will help the teacher both during the lesson and will allow to collect statistics for the semester. The collected monitoring information can be useful for further research on the interaction between student behavior in the lessons and their final level of knowledge. The processed information can help teachers better understand students' needs and adapt their teaching methods and resources. All this will contribute to improving the quality of distance education.
Our SYOTSS’s department of AI research and development conducted the work on implementing advanced monitoring features for Zoom meetings. SYTOSS specializes in custom software development solutions. We can help you integrate additional functionalities, implement innovations, optimize your current processes and create a secure and productive software environment. Contact us today to discuss your specific needs!
Acknowledgements
The authors are grateful to SYTOSS s.r.o., Bratislava, Slovakia, represented by the CEO Oleksiy Matikaynen, for the equipment provided for the research, as well as to employees for participating in experiments.
The work is funded by the EU NextGenerationEU through the Recovery and Resilience Plan for Slovakia under project No. 09I03-03-V01-00115.
Authors
Olena YAKOVLEVA, Marián KOVÁČ, Vadim ARDASOV, Ivan YEREMENKO
Department of AI Research and Development, SYTOSS Ltd
Citation
In case of citing, mentioning, quoting, or using other ways any part of this work, please give a reference to: Yakovleva, O., Kovač, M., Ardasov, V. & Yeremenko, I. (2023). Study on adding functionality to the Zoom online conference system for monitoring the participant activities. Public Administration and Regional Development, 19(1), pp. 158–184.
URL: https://www.vsemba.sk/portals/0/Subory/vedecky%20casopis%2001%20-%202023.pdf
References
[1] Impact of distant teaching during Сovid-19 pandemic on civic and financial literacy / Marcel Lincényi, Matej Mindár, 2022. DOI 10.9770/jesi.2022.10.1(5). In: Entrepreneurship and Sustainability Issues / Marcel Lincényi, Matej Mindár. — ISSN 2345-2082, Roč. 10, no. 1 (2022), pp. 92-106.
[2] Education in times of SARS-COV-2 Pandemic why we did not close our schools / Ľubomír Nebeský, Michal Fabuš, 2021. In: Conference Proceedings of the 2nd Online International Scientific Conference "Economics, Politics and Management in times of change", November 19th, 2021, Hungary. - Gödöllö: Hungarian University of Agriculture and Life Sciences, 2021. — ISBN 978-80-89654-83-3.
[3] Brandl, R. 2023. Retrieved from https://www.emailtooltester.com/en/blog/video-conferencing-market-share.
[4] Zoom meeting SDK. 2023. Retrieved from https://developers.zoom.us/docs/meeting-sdk.
[5] Zoom Video SDK. 2023. Retrieved from https://developers.zoom.us/docs/video-sdk.
[6] Kovtunenko, A., Yakovleva, O., Liubchenko, V., & Yanholenko, O. 2020. Research of the joint use of mathematical morphology and convolutional neural networks for the solution of the price tag recognition problem. Bulletin of National Technical University "KhPI". Series: System Analysis, Control and Information Technologies, 1 (3), 24-31. doi:10.20998/2079-0023.2020.01.05
[7] Yakovleva, O., Kovtunenko, A., Liubchenko, V., Honcharenko, V., & Kobylin, O. 2023. Face Detection for Video Surveillance-based Security System (COLINS-2023). In CEUR Workshop Proceedings (Vol. 3403). pp. 69-86.
[8] K. Zhang, Z. Zhang, Z. Li, Y. 2016. Qiao, Joint face detection and alignment using multitask cascaded convolutional networks, IEEE Signal Processing Letters 23(10) 1499-1503. doi:10.1109/lsp.2016.2603342.
[9] S. Zhang, X. Zhu, Z. Lei, H. Shi, X. Wang, S. Z. Li. 2017. FaceBoxes: A CPU real-time face detector with high accuracy, in: 2017 IEEE International Joint Conference on Biometrics (IJCB), pp. 1-9. doi:10.1109/btas.2017.8272675.
[10] S. Ren, K. He, R. Girshick, J. Sun. 2016. Faster R-CNN: Towards real-time object detection with region proposal networks, IEEE Transactions on Pattern Analysis and Machine Intelligence 39(6). 1137-1149. doi:10.1109/tpami.2016.2577031.
[11] J. Deng, J. Guo, E. Ververas, I. Kotsia, S. Zafeiriou. 2020. Retinaface: Single-shot multi-level face localization in the wild, in: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 5203-5212. doi:10.1109/cvpr42600.2020.00525.
[12] Y. Xu, W. Yan, G. Yang, J. Luo, T. Li, J. He. 2020 Centerface: Joint face detection and alignment using face as point, Scientific Programming. 1-8. doi:10.1155/2020/7845384.
[13] J. Guo, J. Deng, A. Lattas, S. Zafeiriou. 2021. Sample and computation redistribution for efficient face detection. Retrieved from https://arxiv.org/abs/2105.04714.
[14] Liu, Y., Liu, R., Wang, S., Yan, D., Peng, B., & Zhang, T. 2022. Video face detection based on improved SSD model and target tracking algorithm. Journal of Web Engineering. Retrieved from https://doi.org/10.13052/jwe1540-9589.21218/.
[15] Taigman, Y., Yang, M., Ranzato, M., & Wolf, L. 2014. Deepface: Closing the gap to human-level performance in face verification. 2014 IEEE Conference on Computer Vision and Pattern Recognition. Retrieved from https://doi.org/10.1109/cvpr.2014.220.
[16] Deepface 0.0.79. 2023. Retrieved from https://pypi.org/project/deepface/.
[17] Labeled Faces in the Wild. 2007. Retrieved from http://vis-www.cs.umass.edu/lfw/.
[18] Serengil, S. 2022. Face recognition with facebook deepface in Keras. Retrieved from https://sefiks.com/2020/02/17/face-recognition-with-facebook-deepface-in-keras/ Labeled faces in the wild home.
[19] Goodfellow, I. J., Erhan, D., Carrier, P. L., Courville, A., Mirza, M., Hamner, B., ... Bengio, Y. 2013. Challenges in representation learning: A report on three machine learning contests. Neural Information Processing, 117-124. doi:10.1007/978-3-642-42051-1_16
[20] Challenges in Representation Learning: Facial Expression Recognition Challenge. (2013). Retrieved from https://www.kaggle.com/c/challenges-in-representation-learning-facial-expression-recognition-challenge/data.
[21] EDPS TechDispatch: Facial Emotion Recognition (FER). 2021. Issue 1. pp. 1-5. doi:10.2804/519064. Retrieved from https://edps.europa.eu/system/files/2021-05/21-05-26_techdispatch-facial-emotion-recognition_ref_en.pdf.
[22] Facial Expression Recognition (FER) on FER2013 (2023) Papers with code - fer2013 benchmark Retrieved from https://paperswithcode.com/sota/facial-expression-recognition-on-fer2013/.
[23] Serengil, S. 2021. Facial expression recognition with Keras. Retrieved from https://sefiks.com/2018/01/01/facial-expression-recognition-with-keras/.