Paper
6 October 1997 Multiscale content extraction and representation for video indexing
Ahmet Mufit Ferman, A. Murat Tekalp
Author Affiliations +
Proceedings Volume 3229, Multimedia Storage and Archiving Systems II; (1997) https://doi.org/10.1117/12.290352
Event: Voice, Video, and Data Communications, 1997, Dallas, TX, United States
Abstract
This paper presents a general multiscale framework for extraction and representation of video content. The approach exploits the inherent multiscale nature of many TV and film productions to delineate an input stream effectively and to construct consistent scenes reliably. The method first utilizes basic signal processing techniques, and unsupervised clustering to determine shot boundaries in the video sequence. Similarity comparison using shot representative histograms and clustering to determine shot boundaries in the video sequence. Similarity comparison using shot representative histograms and clustering is then carried out within each shot to automatically select representative key frames. Finally, a model that takes into account the filmic structure of the input stream is discussed and developed to efficiently merge individual shots into coherent, meaningful segments, i.e. scenes.
© (1997) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Ahmet Mufit Ferman and A. Murat Tekalp "Multiscale content extraction and representation for video indexing", Proc. SPIE 3229, Multimedia Storage and Archiving Systems II, (6 October 1997); https://doi.org/10.1117/12.290352
Lens.org Logo
CITATIONS
Cited by 44 scholarly publications.
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Video

Visualization

Data processing

Video processing

Digital filtering

Motion analysis

Semantic video

Back to Top