Publication Type
Journal Article
Version
acceptedVersion
Publication Date
11-2010
Abstract
This paper considers the mining and localization of near-duplicate segments at arbitrary positions of partial near-duplicate videos in a corpus. Temporal network is proposed to model the visual-temporal consistency between video sequence by embedding temporal constraints as directed edges in the network. Partial alignment is then achieved through network flow programming. To handle multiple alignments, we consider two properties of network structure: conciseness and divisibility, to ensure that the mining is efficient and effective. Frame-level matching is further integrated in the temporal network for alignment verification. This results in an iterative alignment-verification procedure to fine tune the localization of near-duplicate segments. The scalability of frame-level matching is enhanced by exploring visual keyword matching algorithms. We demonstrate the proposed work for mining partial alignments from two months of broadcast videos and across six TV sources.
Keywords
Keyword matching, Partial near-duplicate, temporal graph
Discipline
Graphics and Human Computer Interfaces | OS and Networks
Research Areas
Intelligent Systems and Optimization
Publication
IEEE Transactions on Circuits and Systems for Video Technology
Volume
20
Issue
11
First Page
1486
Last Page
1498
ISSN
1051-8215
Identifier
10.1109/TCSVT.2010.2077531
Publisher
Institute of Electrical and Electronics Engineers
Citation
1
Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 International License.