Author:
Gyo Chung Min,Wang Taehyung (George),Sheu Phillip C.‐Y.
Abstract
PurposeVideo summarisation is one of the most active fields in content‐based video retrieval research. A new video summarisation scheme is proposed by this paper based on socially generated temporal tags.Design/methodology/approachTo capture users' collaborative tagging activities the proposed scheme maintains video bookmarks, which contain some temporal or positional information about videos, such as relative time codes or byte offsets. For each video all the video bookmarks collected from users are then statistically analysed in order to extract some meaningful key frames (the video equivalent of keywords), which collectively constitute the summary of the video.FindingsCompared with traditional video summarisation methods that use low‐level audio‐visual features, the proposed method is based on users' high‐level collaborative activities, and thus can produce semantically more important summaries than existing methods.Research limitations/implicationsIt is assumed that the video frames around the bookmarks inserted by users are informative and representative, and therefore can be used as good sources for summarising videos.Originality/valueFolksonomy, commonly called collaborative tagging, is a Web 2.0 method for users to freely annotate shared information resources with keywords. It has mostly been used for collaboratively tagging photos (Flickr), web site bookmarks (Del.icio.us), or blog posts (Technorati), but has never been applied to the field of automatic video summarisation. It is believed that this is the first attempt to utilise users' high‐level collaborative tagging activities, instead of low‐level audio‐visual features, for video summarisation.
Subject
Library and Information Sciences,Computer Science Applications,Information Systems
Reference34 articles.
1. Abdollahian, G. and Delp, E.J. (2009), “User generated video annotation using geo‐tagged image databases”, Proceedings of the IEEE International Conference on Multimedia and Expo, IEEE, Los Alamitos, CA, pp. 610‐13.
2. Ames, M. and Naaman, M. (2007), “Why we tag: motivations for annotation in mobile and online media”, Proceedings of the Conference on Human Factors in Computing Systems (CHI 2007), ACM Press, New York, NY, pp. 971‐80.
3. Angus, E., Thelwall, M. and Stuart, D. (2008), “General patterns of tag usage among university groups in Flickr”, Online Information Review, Vol. 32 No. 1, pp. 89‐101.
4. Avila, S., Luz, A., Araujo, A.A. and Cord, M. (2008), “VSUMM: an approach for automatic video summarization and quantitative evaluation”, Proceedings of XXI Brazilian Symposium on Computer Graphics and Image Processing, IEEE, Los Alamitos, CA, pp. 103‐10.
5. Barbieri, M., Agnihotri, L. and Dimitrova, N. (2003), “Video summarization: methods and landscape”, Proceedings of the Society of Photographic Instrumentation Engineers Internet Multimedia Management Systems IV Conference, Vol. 5242, SPIE, Bellingham, WA, pp. 1‐13.
Cited by
7 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献