The present invention is a method and
system to provide an automatic measurement of people's responses to dynamic digital media, based on changes in their facial expressions and attention to specific content. First, the method detects and tracks faces from the audience. It then localizes each of the faces and facial features to extract emotion-sensitive features of the face by applying emotion-sensitive feature filters, to determine the facial
muscle actions of the face based on the extracted emotion-sensitive features. The changes in facial
muscle actions are then converted to the changes in affective state, called an emotion trajectory. On the other hand, the method also estimates eye
gaze based on extracted eye images and three-dimensional facial
pose of the face based on localized facial images. The
gaze direction of the person, is estimated based on the estimated eye
gaze and the three-dimensional facial
pose of the person. The gaze target on the media display is then estimated based on the estimated gaze direction and the position of the person. Finally, the response of the person to the dynamic digital
media content is determined by analyzing the emotion trajectory in relation to the time and screen positions of the specific digital media sub-content that the person is watching.