TY - GEN
T1 - Segment and recognize expression phase by fusion of motion area and neutral divergence features
AU - Chen, Shizhi
AU - Tian, Yingli
AU - Liu, Qingshan
AU - Metaxas, Dimitris N.
PY - 2011
Y1 - 2011
N2 - An expression can be approximated by a sequence of temporal segments called neutral, onset, offset and apex. However, it is not easy to accurately detect such temporal segments only based on facial features. Some researchers try to temporally segment expression phases with the help of body gesture analysis. The problem of this approach is that the expression temporal phases from face and gesture channels are not synchronized. Additionally, most previous work adopted facial key points tracking or body tracking to extract motion information, which is unreliable in practice due to illumination variations and occlusions. In this paper, we present a novel algorithm to overcome the above issues, in which two simple and robust features are designed to describe face and gesture information, i.e., motion area and neutral divergence features. Both features do not depend on motion tracking, and they can be easily calculated too. Moreover, it is different from previous work in that we integrate face and body gesture together in modeling the temporal dynamics through a single channel of sensorial source, so it avoids the unsynchronized issue between face and gesture channels. Extensive experimental results demonstrate the effectiveness of the proposed algorithm.
AB - An expression can be approximated by a sequence of temporal segments called neutral, onset, offset and apex. However, it is not easy to accurately detect such temporal segments only based on facial features. Some researchers try to temporally segment expression phases with the help of body gesture analysis. The problem of this approach is that the expression temporal phases from face and gesture channels are not synchronized. Additionally, most previous work adopted facial key points tracking or body tracking to extract motion information, which is unreliable in practice due to illumination variations and occlusions. In this paper, we present a novel algorithm to overcome the above issues, in which two simple and robust features are designed to describe face and gesture information, i.e., motion area and neutral divergence features. Both features do not depend on motion tracking, and they can be easily calculated too. Moreover, it is different from previous work in that we integrate face and body gesture together in modeling the temporal dynamics through a single channel of sensorial source, so it avoids the unsynchronized issue between face and gesture channels. Extensive experimental results demonstrate the effectiveness of the proposed algorithm.
KW - motion area
KW - neutral divergence
KW - temporal segment
UR - http://www.scopus.com/inward/record.url?scp=79958703466&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=79958703466&partnerID=8YFLogxK
U2 - 10.1109/FG.2011.5771419
DO - 10.1109/FG.2011.5771419
M3 - Conference contribution
AN - SCOPUS:79958703466
SN - 9781424491407
T3 - 2011 IEEE International Conference on Automatic Face and Gesture Recognition and Workshops, FG 2011
SP - 330
EP - 335
BT - 2011 IEEE International Conference on Automatic Face and Gesture Recognition and Workshops, FG 2011
T2 - 2011 IEEE International Conference on Automatic Face and Gesture Recognition and Workshops, FG 2011
Y2 - 21 March 2011 through 25 March 2011
ER -