您的位置: 专家智库 > >

国家自然科学基金(s61272211)

作品数:3 被引量:3H指数:1
发文基金:国家自然科学基金中国博士后科学基金更多>>
相关领域:电子电信社会学自动化与计算机技术经济管理更多>>

文献类型

  • 3篇中文期刊文章

领域

  • 2篇电子电信
  • 1篇经济管理
  • 1篇自动化与计算...
  • 1篇社会学

主题

  • 1篇NEURAL...
  • 1篇UNSUPE...
  • 1篇VALENC...
  • 1篇DIMENS...
  • 1篇FEATUR...
  • 1篇FUNCTI...
  • 1篇FUSION
  • 1篇LEARNI...
  • 1篇PARALA...
  • 1篇ORDINA...
  • 1篇PARA
  • 1篇AROUSA...
  • 1篇AFFECT...
  • 1篇RANKIN...
  • 1篇SPEAKE...

传媒

  • 2篇Fronti...
  • 1篇Journa...

年份

  • 1篇2018
  • 1篇2015
  • 1篇2013
3 条 记 录,以下是 1-3
排序方式:
Speech emotion recognition with unsupervised feature learning被引量:1
2015年
Emotion-based features are critical for achieving high performance in a speech emotion recognition(SER) system. In general, it is difficult to develop these features due to the ambiguity of the ground-truth. In this paper, we apply several unsupervised feature learning algorithms(including K-means clustering, the sparse auto-encoder, and sparse restricted Boltzmann machines), which have promise for learning task-related features by using unlabeled data, to speech emotion recognition. We then evaluate the performance of the proposed approach and present a detailed analysis of the effect of two important factors in the model setup, the content window size and the number of hidden layer nodes. Experimental results show that larger content windows and more hidden nodes contribute to higher performance. We also show that the two-layer network cannot explicitly improve performance compared to a single-layer network.
Zheng-wei HUANGWen-tao XUEQi-rong MAO
Affective rating ranking based on face images in arousal-valence dimensional space
2018年
In dimensional affect recognition, the machine learning methods, which are used to model and predict affect, are mostly classification and regression. However, the annotation in the dimensional affect space usually takes the form of a continuous real value which has an ordinal property. The aforementioned methods do not focus on taking advantage of this important information. Therefore, we propose an affective rating ranking framework for affect recognition based on face images in the valence and arousal dimensional space. Our approach can appropriately use the ordinal information among affective ratings which are generated by discretizing continuous annotations.Specifically, we first train a series of basic cost-sensitive binary classifiers, each of which uses all samples relabeled according to the comparison results between corresponding ratings and a given rank of a binary classifier. We obtain the final affective ratings by aggregating the outputs of binary classifiers. By comparing the experimental results with the baseline and deep learning based classification and regression methods on the benchmarking database of the AVEC 2015 Challenge and the selected subset of SEMAINE database, we find that our ordinal ranking method is effective in both arousal and valence dimensions.
Guo-peng XUHai-tang LUFei-fei ZHANGQi-rong MAO
关键词:VALENCEAROUSAL
Speaker-independent speech emotion recognition by fusion of functional and accompanying paralanguage features被引量:2
2013年
Functional paralanguage includes considerable emotion information, and it is insensitive to speaker changes. To improve the emotion recognition accuracy under the condition of speaker-independence, a fusion method combining the functional paralanguage features with the accompanying paralanguage features is proposed for the speaker-independent speech emotion recognition. Using this method, the functional paralanguages, such as laughter, cry, and sigh, are used to assist speech emotion recognition. The contributions of our work are threefold. First, one emotional speech database including six kinds of functional paralanguage and six typical emotions were recorded by our research group. Second, the functional paralanguage is put forward to recognize the speech emotions combined with the accompanying paralanguage features. Third, a fusion algorithm based on confidences and probabilities is proposed to combine the functional paralanguage features with the accompanying paralanguage features for speech emotion recognition. We evaluate the usefulness of the functional paralanguage features and the fusion algorithm in terms of precision, recall, and F1-measurement on the emotional speech database recorded by our research group. The overall recognition accuracy achieved for six emotions is over 67% in the speaker-independent condition using the functional paralanguage features.
Qi-rong MAOXiao-lei ZHAOZheng-wei HUANGYong-zhao ZHAN
关键词:SPEAKER-INDEPENDENT
共1页<1>
聚类工具0