TY - JOUR
T1 - Representative segment-based emotion analysis and classification with automatic respiration signal segmentation
AU - Wu, Chi Keng
AU - Chung, Pau Choo Julia
AU - Wang, Chi Jen
N1 - Funding Information:
This work was supported by the National Science Council of Taiwan under Grant NSC98-2218-E-006-004.
PY - 2012
Y1 - 2012
N2 - Human respiration has been reported as having great relevance to the emotional state. However, the respiration signal obtained using traditional biosensors not only reflects changing emotions and emotion intensities, but also contains motion artifacts. The resulting information ambiguity limits the practical application of the respiration signal as a means of reliable affective state appraisal. Thus, the present study proposes a method for extracting representative Emotion Elicited Segments (EESs) from the respiration signal such that the affective state of the individual can be more reliably determined. The EES extraction process involves the combination of the following procedures, namely, 1) Mutual Information-Based Emotion Relevance Feature Ranking based on the Dynamic Time Warping Distance (MIDTW), and 2) Constraint-based Elicited Segment Density (CESD) analysis. Due to the innate regularity of the respiration signal, the respiration pattern under a particular emotional state would be relatively quasihomogeneous over time. Accordingly, a parameter-free Respiration quasi-Homogeneity Segmentation (RHS) algorithm is proposed for partitioning the respiration signal into quasi-homogenous segments from which the EESs can then be extracted. The experimental results obtained for five prototypical emotions (i.e., "love," "sadness," "joy," "anger," and "fear") show that the proposed segmentation/extraction methodology enables the EESs to be reliably identified and yields an average classification rate of 88 percent.
AB - Human respiration has been reported as having great relevance to the emotional state. However, the respiration signal obtained using traditional biosensors not only reflects changing emotions and emotion intensities, but also contains motion artifacts. The resulting information ambiguity limits the practical application of the respiration signal as a means of reliable affective state appraisal. Thus, the present study proposes a method for extracting representative Emotion Elicited Segments (EESs) from the respiration signal such that the affective state of the individual can be more reliably determined. The EES extraction process involves the combination of the following procedures, namely, 1) Mutual Information-Based Emotion Relevance Feature Ranking based on the Dynamic Time Warping Distance (MIDTW), and 2) Constraint-based Elicited Segment Density (CESD) analysis. Due to the innate regularity of the respiration signal, the respiration pattern under a particular emotional state would be relatively quasihomogeneous over time. Accordingly, a parameter-free Respiration quasi-Homogeneity Segmentation (RHS) algorithm is proposed for partitioning the respiration signal into quasi-homogenous segments from which the EESs can then be extracted. The experimental results obtained for five prototypical emotions (i.e., "love," "sadness," "joy," "anger," and "fear") show that the proposed segmentation/extraction methodology enables the EESs to be reliably identified and yields an average classification rate of 88 percent.
UR - http://www.scopus.com/inward/record.url?scp=84872249747&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84872249747&partnerID=8YFLogxK
U2 - 10.1109/T-AFFC.2012.14
DO - 10.1109/T-AFFC.2012.14
M3 - Article
AN - SCOPUS:84872249747
SN - 1949-3045
VL - 3
SP - 482
EP - 495
JO - IEEE Transactions on Affective Computing
JF - IEEE Transactions on Affective Computing
IS - 4
M1 - 6212432
ER -