Content-Based Video Retrieval Using Integration of Curvelet Transform and Simple Linear Iterative Clustering

Author(s):  
Reddy Mounika Bommisetty ◽  
Ashish Khare ◽  
Manish Khare ◽  
P. Palanisamy

Video is a rich information source containing both audio and visual information along with motion information embedded in it. Applications such as e-learning, live TV, video on demand, traffic monitoring, etc. need an efficient video retrieval strategy. Content-based video retrieval and superpixel segmentation are two diverse application areas of computer vision. In this work, we are presenting an algorithm for content-based video retrieval with help of Integration of Curvelet transform and Simple Linear Iterative Clustering (ICTSLIC) algorithm. Proposed algorithm consists of two steps: off line processing and online processing. In offline processing, keyframes of the database videos are extracted by employing features: Pearson Correlation Coefficient (PCC) and color moments (CM) and on the extracted keyframes superpixel generation algorithm ICTSLIC is applied. The superpixels generated by applying ICTSLIC on keyframes are used to represent database videos. On other side, in online processing, ICTSLIC superpixel segmentation is applied on query frame and the superpixels generated by segmentation are used to represent query frame. Then videos similar to query frame are retrieved through matching done by calculation of Euclidean distance between superpixels of query frame and database keyframes. Results of the proposed method are irrespective of query frame features such as camera motion, object’s pose, orientation and motion due to the incorporation of ICTSLIC superpixels as base feature for matching and retrieval purpose. The proposed method is tested on the dataset comprising of different categories of video clips such as animations, serials, personal interviews, news, movies and songs which is publicly available. For evaluation, the proposed method randomly picks frames from database videos, instead of selecting keyframes as query frames. Experiments were conducted on the developed dataset and the performance is assessed with different parameters Precision, Recall, Jaccard Index, Accuracy and Specificity. The experimental results shown that the proposed method is performing better than the other state-of-art methods.

2021 ◽  
Author(s):  
ElMehdi SAOUDI ◽  
Said Jai Andaloussi

Abstract With the rapid growth of the volume of video data and the development of multimedia technologies, it has become necessary to have the ability to accurately and quickly browse and search through information stored in large multimedia databases. For this purpose, content-based video retrieval ( CBVR ) has become an active area of research over the last decade. In this paper, We propose a content-based video retrieval system providing similar videos from a large multimedia data-set based on a query video. The approach uses vector motion-based signatures to describe the visual content and uses machine learning techniques to extract key-frames for rapid browsing and efficient video indexing. We have implemented the proposed approach on both, single machine and real-time distributed cluster to evaluate the real-time performance aspect, especially when the number and size of videos are large. Experiments are performed using various benchmark action and activity recognition data-sets and the results reveal the effectiveness of the proposed method in both accuracy and processing time compared to state-of-the-art methods.


2013 ◽  
Vol 64 (3) ◽  
pp. 35-38 ◽  
Author(s):  
Sudeep D.Thepade ◽  
Krishnasagar Subhedarpage ◽  
Ankur A. Mali ◽  
Tushar S. Vaidya

Author(s):  
Sumaya Hamad ◽  
Ahmeed Suliman Farhan ◽  
Doaa Yaseen Khudhur

A content based video retrieval (CBVR)framework is built in this paper.  One of the essential features of video retrieval process and CBVR is a color value. The discrete cosine transform (DCT) is used to extract a query video features to compare with the video features stored in our database. Average result of 0.6475 was obtained by using the DCT after implementing it to the database we created and collected, and on all categories. This technique was applied on our database of video, Check 100 database videos, 5 videos in each category.


Sign in / Sign up

Export Citation Format

Share Document