VSUM: Summarizing from videos

Yu Chyeh Wu, Yue Shi Lee, Chia Hui Chang

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

4 Scopus citations

Abstract

Summarization on produced type of video data (like news or movies) is to find important segments that contain rich information. Users could obtain the important messages by reading summaries rather than full documents. The researches in this area could be divided into two parts: (1) Image Processing (IP) perspective, and (2) NLP (Nature Language Processing) perspective. The former put emphasis on the detection of key frames, while the later focused on the extraction of important concepts. This paper proposes a video summarization system, VSUM. VSUM first identifies all caption words, and then adopts a technique to find the important segments. An external thesaurus is also used in VSUM to enhance the summary extraction process. The experimental results show that VSUM could perform well even if the accuracy of OCR (Optical Character Recognition) is not sophisticating.

Original languageEnglish
Title of host publicationProceedings - IEEE Sixth International Symposium on Multimedia Software Engineering, MSE 2004
Pages302-309
Number of pages8
StatePublished - 2004
EventProceedings - IEEE Sixth International Symposium on Multimedia Software Engineering, MSE 2004 - Miami, FL, United States
Duration: 13 Dec 200415 Dec 2004

Publication series

NameProceedings - IEEE Sixth International Symposium on Multimedia Software Engineering, MSE 2004

Conference

ConferenceProceedings - IEEE Sixth International Symposium on Multimedia Software Engineering, MSE 2004
Country/TerritoryUnited States
CityMiami, FL
Period13/12/0415/12/04

Fingerprint

Dive into the research topics of 'VSUM: Summarizing from videos'. Together they form a unique fingerprint.

Cite this