Copy Detection Using Graphical Model

Author(s):  
Shikui Wei ◽  
Yao Zhao ◽  
Zhenfeng Zhu

With the growing popularity of video sharing websites and editing tools, it is easy for people to involve the video content from different sources into their own work, which raises the copyright problem. Content-based video copy detection attempts to track the usage of the copyright-protected video content by using video analysis techniques, which deals with not only whether a copy occurs in a query video stream but also where the copy is located and where the copy is originated from. While a lot of work has addressed the problem with good performance, less effort has been made to consider the copy detection problem in the case of a continuous query stream, for which precise temporal localization and some complex video transformations like frame insertion and video editing need to be handled. In this chapter, the authors attack the problem by employing the graphical model to facilitate the frame fusion based video copy detection approach. The key idea is to convert frame fusion problem into graph model decoding problem with the temporal consistency constraint and three relaxed constraints. This work employs the HMM model to perform frame fusion and propose a Viterbi-like algorithm to speedup frame fusion process.

Author(s):  
Shikui Wei ◽  
Yao Zhao ◽  
Ce Zhu ◽  
Changsheng Xu ◽  
Zhenfeng Zhu

Author(s):  
Chanjal C

Predicting the relevance between two given videos with respect to their visual content is a key component for content-based video recommendation and retrieval. The application is in video recommendation, video annotation, Category or near-duplicate video retrieval, video copy detection and so on. In order to estimate video relevance previous works utilize textual content of videos and lead to poor performance. The proposed method is feature re-learning for video relevance prediction. This work focus on the visual contents to predict the relevance between two videos. A given feature is projected into a new space by an affine transformation. Different from previous works this use a standard triplet ranking loss that optimize the projection process by a novel negative-enhanced triplet ranking loss. In order to generate more training data, propose a data augmentation strategy which works directly on video features. The multi-level augmentation strategy works for video features, which benefits the feature relearning. The proposed augmentation strategy can be flexibly used for frame-level or video-level features. The loss function that consider the absolute similarity of positive pairs and supervise the feature re-learning process and a new formula for video relevance computation.


2016 ◽  
pp. 8-13
Author(s):  
Daniel Reynolds ◽  
Richard A. Messner

Video copy detection is the process of comparing and analyzing videos to extract a measure of their similarity in order to determine if they are copies, modified versions, or completely different videos. With video frame sizes increasing rapidly, it is important to allow for a data reduction process to take place in order to achieve fast video comparisons. Further, detecting video streaming and storage of legal and illegal video data necessitates the fast and efficient implementation of video copy detection algorithms. In this paper some commonly used algorithms for video copy detection are implemented with the Log-Polar transformation being used as a pre-processing step to reduce the frame size prior to signature calculation. Two global based algorithms were chosen to validate the use of Log-Polar as an acceptable data reduction stage. The results of this research demonstrate that the addition of this pre-processing step significantly reduces the computation time of the overall video copy detection process while not significantly affecting the detection accuracy of the algorithm used for the detection process.


Sign in / Sign up

Export Citation Format

Share Document