TY - JOUR
T1 - Video content summarization and augmentation based on structural semantic processing and social network analysis
AU - Chen, Bo Wei
AU - Wang, Jhing Fa
AU - Wang, Jia Ching
N1 - Funding Information:
This work was supported in part by the National Science Council of the Republic of China under Grant NSC97-2221-E-006-249-MY3.
PY - 2010
Y1 - 2010
N2 - Video summarization techniques have been proposed for years to offer people comprehensive understanding of a whole story on video. However, although these traditional methods give brief summaries for users, they still do not provide concept-organized or structural views. Besides, the knowledge they offer to users is often limited to existing videos. In this study, we present a structural video content summarization that utilizes the four kinds of entities, “who,” “what,” “where,” and “when,” to establish the framework of the video contents. Relevant media associated with each entity in the online resource are also analyzed to enrich existing contents. With the above-mentioned information, the structure of the story and its complementary knowledge can be built up according to the entities. Therefore, users can not only browse the video efficiently but also focus on what they are interested in. In order to construct the fundamental system, we employ the maximum entropy criterion to integrate visual and text features extracted from video frames and speech transcripts, generating high-level concept entities. Shots are linked together based on their contents. After constructing the relational graph, we exploit the graph entropy model to detect meaningful shots and relations. The social network analysis based on the Markov clustering algorithm is performed to explore relevant information online. The results demonstrate that our system can achieve excellent performance and information coverage.
AB - Video summarization techniques have been proposed for years to offer people comprehensive understanding of a whole story on video. However, although these traditional methods give brief summaries for users, they still do not provide concept-organized or structural views. Besides, the knowledge they offer to users is often limited to existing videos. In this study, we present a structural video content summarization that utilizes the four kinds of entities, “who,” “what,” “where,” and “when,” to establish the framework of the video contents. Relevant media associated with each entity in the online resource are also analyzed to enrich existing contents. With the above-mentioned information, the structure of the story and its complementary knowledge can be built up according to the entities. Therefore, users can not only browse the video efficiently but also focus on what they are interested in. In order to construct the fundamental system, we employ the maximum entropy criterion to integrate visual and text features extracted from video frames and speech transcripts, generating high-level concept entities. Shots are linked together based on their contents. After constructing the relational graph, we exploit the graph entropy model to detect meaningful shots and relations. The social network analysis based on the Markov clustering algorithm is performed to explore relevant information online. The results demonstrate that our system can achieve excellent performance and information coverage.
KW - Content augmentation
KW - Graph clustering
KW - Social network analysis
KW - Structural contents
UR - http://www.scopus.com/inward/record.url?scp=77958112122&partnerID=8YFLogxK
U2 - 10.1080/02533839.2010.9671663
DO - 10.1080/02533839.2010.9671663
M3 - 期刊論文
AN - SCOPUS:77958112122
SN - 0253-3839
VL - 33
SP - 737
EP - 750
JO - Journal of the Chinese Institute of Engineers, Transactions of the Chinese Institute of Engineers,Series A/Chung-kuo Kung Ch'eng Hsuch K'an
JF - Journal of the Chinese Institute of Engineers, Transactions of the Chinese Institute of Engineers,Series A/Chung-kuo Kung Ch'eng Hsuch K'an
IS - 5
ER -