Recognition of Emotions using Energy Based Bimodal Information Fusion and Correlation
Multi-sensor information fusion is a rapidly developing research area which forms the backbone of numerous essential technologies such as intelligent robotic control, sensor networks, video and image processing and many more. In this paper, we have developed a novel technique to analyze and correlate human emotions expressed in voice tone & facial expression. Audio and video streams captured to populate audio and video bimodal data sets to sense the expressed emotions in voice tone and facial expression respectively. An energy based mapping is being done to overcome the inherent heterogeneity of the recorded bi-modal signal. The fusion process uses sampled and mapped energy signal of both modalities’s data stream and further recognize the overall emotional component using Support Vector Machine (SVM) classifier with the accuracy 93.06%.
|Year of Publication||
International Journal of Interactive Multimedia and Artificial Intelligence
Special Issue on Multisensor User Tracking and Analytics to Improve Education and other Application Fields
|Number of Pages||