In the last decade, there has been a growing interest in emotion analysis research, which has been applied in several areas of computer science.
Many authors have contributed to the development of emotion recognition algorithms, considering textual or non verbal data as input, such as facial expressions, gestures or, in the case of multi-modal emotion recognition, a combination of them. In this paper, we describe a method to detect emotions from gestures using the skeletal data obtained from Kinect-like devices as input, as well as a textual description of their meaning.
The experimental results show that the correlation existing between body movements and spoken user sentence(s) can be used to reveal user’s emotions from gestures.
This is one of the scientific articles published by one or more synbrAIn collaborators and data scientists.
If you are interested in learning more, read the entire article here.
