Multimodal sentiment analysis combining language, audio, visual, and physiological signals (IMAGE)
Caption
The multimodal neural network is used to predict user sentiment from multimodal features such as text, audio, and visual data. In a new study, researchers from Japan account for physiological signals in sentiment estimation while talking with the system, greatly improving the system’s performance.
Credit
Shogo Okada from JAIST.
Usage Restrictions
none
License
Original content