Multimodal integration for meeting group action segmentation and recognition

MACHINE LEARNING FOR MULTIMODAL INTERACTION(2005)

引用 35|浏览0
暂无评分
摘要
We address the problem of segmentation and recognition of sequences of multimodal human interactions in meetings. These interactions can be seen as a rough structure of a meeting, and can be used either as input for a meeting browser or as a first step towards a higher semantic analysis of the meeting. A common lexicon of multimodal group meeting actions, a shared meeting data set, and a common evaluation procedure enable us to compare the different approaches. We compare three different multimodal feature sets and our modelling infrastructures: a higher semantic feature approach, multi-layer HMMs, a multi-stream DBN, as well as a multi-stream mixed-state DBN for disturbed data.
更多
查看译文
关键词
multimodal integration,shared meeting data,common lexicon,multimodal human interaction,higher semantic analysis,disturbed data,different approach,meeting group action segmentation,different multimodal feature set,multimodal group meeting action,common evaluation procedure,meeting browser,human interaction,vision,group action
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要