Semantic video annotation by mining association patterns from visual and speech features

Vincent S. Tseng, Ja Hwung Su, Jhih Hong Huang, Chih Jen Chen

研究成果: Conference contribution同行評審

2 引文 斯高帕斯(Scopus)

摘要

In this paper, we propose a novel approach for semantic video annotation through integrating visual features and speech features. By employing statistics and association patterns, the relations between video shots and human concept can be discovered effectively to conceptualize videos. In other words, the utilization of high-level rules can effectively complement the insufficiency of statistics-based methods in dealing with broad and complex keyword identification in video annotation. Empirical evaluations on NIST TRECVID video datasets reveal that our proposed approach can enhance the annotation accuracy substantially.

原文English
主出版物標題Advances in Knowledge Discovery and Data Mining - 12th Pacific-Asia Conference, PAKDD 2008, Proceedings
頁面1035-1041
頁數7
DOIs
出版狀態Published - 2008
事件12th Pacific-Asia Conference on Knowledge Discovery and Data Mining, PAKDD 2008 - Osaka, 日本
持續時間: 20 5月 200823 5月 2008

出版系列

名字Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
5012 LNAI
ISSN(列印)0302-9743
ISSN(電子)1611-3349

Conference

Conference12th Pacific-Asia Conference on Knowledge Discovery and Data Mining, PAKDD 2008
國家/地區日本
城市Osaka
期間20/05/0823/05/08

指紋

深入研究「Semantic video annotation by mining association patterns from visual and speech features」主題。共同形成了獨特的指紋。

引用此