Real-time video sentiment analysis through the use of snapshots
| dc.contributor.advisor | Nyirenda, Juwa | |
| dc.contributor.author | Ramma, Sudiptee | |
| dc.date.accessioned | 2025-09-19T09:54:23Z | |
| dc.date.available | 2025-09-19T09:54:23Z | |
| dc.date.issued | 2025 | |
| dc.date.updated | 2025-09-19T09:51:43Z | |
| dc.description.abstract | There are many types of emotions that one can experience and they usually have a direct impact on a person's behaviour. Emotions can be conveyed in several ways such as gestures/body movement, words or facial expressions and this dissertation we aim to distinguish the emotional state of a person based on their facial expressions. Several approaches have been devised in this regard by various past researchers within the computer vision field but unfortunately, despite the similarities in the adopted techniques for the facial and emotion detection processes, there still exist some discrepancies regarding their performances when applied to different images or video streams. As such, the goal of this dissertation is to develop a program that can analyse a real-time video stream and take in each of the frames as an image snapshot which can be in turn processed to efficiently identify faces and recognise a person's emotion based on their facial expressions. Two scenarios, namely Frontal only, and Profile and Frontal, each with their datasets were accounted for in this research. The first dataset (Frontal) consists only of users who are facing forward and the second one (Profile and Frontal) consists of users who are facing forward as well as sideways. Convolutional Neural Network (CNN) models were constructed for each of the given datasets on both the augmented and non-augmented versions of these datasets to obtain the best possible model for each scenario before applying such model to a real-time video stream. In both scenarios, the augmented models outperformed the non-augmented models when tested on unseen static image data and when such a model was applied to a real-time video stream with the help of the OpenCV library and the relevant Haar Cascade classifiers, required for the face detection process (depending on which scenario), fairly accurate results were obtained when each frame within the video stream were converted into an image snapshot before classification. The code for this dissertation can be found here: https://github.com/Drish19/Facial-Emotion-Recognition. | |
| dc.identifier.apacitation | Ramma, S. (2025). <i>Real-time video sentiment analysis through the use of snapshots</i>. (). University of Cape Town ,Faculty of Science ,Department of Statistical Sciences. Retrieved from http://hdl.handle.net/11427/41869 | en_ZA |
| dc.identifier.chicagocitation | Ramma, Sudiptee. <i>"Real-time video sentiment analysis through the use of snapshots."</i> ., University of Cape Town ,Faculty of Science ,Department of Statistical Sciences, 2025. http://hdl.handle.net/11427/41869 | en_ZA |
| dc.identifier.citation | Ramma, S. 2025. Real-time video sentiment analysis through the use of snapshots. . University of Cape Town ,Faculty of Science ,Department of Statistical Sciences. http://hdl.handle.net/11427/41869 | en_ZA |
| dc.identifier.ris | TY - Thesis / Dissertation AU - Ramma, Sudiptee AB - There are many types of emotions that one can experience and they usually have a direct impact on a person's behaviour. Emotions can be conveyed in several ways such as gestures/body movement, words or facial expressions and this dissertation we aim to distinguish the emotional state of a person based on their facial expressions. Several approaches have been devised in this regard by various past researchers within the computer vision field but unfortunately, despite the similarities in the adopted techniques for the facial and emotion detection processes, there still exist some discrepancies regarding their performances when applied to different images or video streams. As such, the goal of this dissertation is to develop a program that can analyse a real-time video stream and take in each of the frames as an image snapshot which can be in turn processed to efficiently identify faces and recognise a person's emotion based on their facial expressions. Two scenarios, namely Frontal only, and Profile and Frontal, each with their datasets were accounted for in this research. The first dataset (Frontal) consists only of users who are facing forward and the second one (Profile and Frontal) consists of users who are facing forward as well as sideways. Convolutional Neural Network (CNN) models were constructed for each of the given datasets on both the augmented and non-augmented versions of these datasets to obtain the best possible model for each scenario before applying such model to a real-time video stream. In both scenarios, the augmented models outperformed the non-augmented models when tested on unseen static image data and when such a model was applied to a real-time video stream with the help of the OpenCV library and the relevant Haar Cascade classifiers, required for the face detection process (depending on which scenario), fairly accurate results were obtained when each frame within the video stream were converted into an image snapshot before classification. The code for this dissertation can be found here: https://github.com/Drish19/Facial-Emotion-Recognition. DA - 2025 DB - OpenUCT DP - University of Cape Town KW - Emotion recognition KW - Facial emotion recognition KW - Face detection KW - Emotion classification KW - Convolutional Neural Network (CNN) KW - OpenCV LK - https://open.uct.ac.za PB - University of Cape Town PY - 2025 T1 - Real-time video sentiment analysis through the use of snapshots TI - Real-time video sentiment analysis through the use of snapshots UR - http://hdl.handle.net/11427/41869 ER - | en_ZA |
| dc.identifier.uri | http://hdl.handle.net/11427/41869 | |
| dc.identifier.vancouvercitation | Ramma S. Real-time video sentiment analysis through the use of snapshots. []. University of Cape Town ,Faculty of Science ,Department of Statistical Sciences, 2025 [cited yyyy month dd]. Available from: http://hdl.handle.net/11427/41869 | en_ZA |
| dc.language.iso | en | |
| dc.language.rfc3066 | eng | |
| dc.publisher.department | Department of Statistical Sciences | |
| dc.publisher.faculty | Faculty of Science | |
| dc.publisher.institution | University of Cape Town | |
| dc.subject | Emotion recognition | |
| dc.subject | Facial emotion recognition | |
| dc.subject | Face detection | |
| dc.subject | Emotion classification | |
| dc.subject | Convolutional Neural Network (CNN) | |
| dc.subject | OpenCV | |
| dc.title | Real-time video sentiment analysis through the use of snapshots | |
| dc.type | Thesis / Dissertation | |
| dc.type.qualificationlevel | Masters | |
| dc.type.qualificationlevel | MSc |