Boosting Temporal Binary Coding for Large-Scale Video Search

2021 ◽  
Vol 23 ◽  
pp. 353-364
Author(s):  
Yan Wu ◽  
Xianglong Liu ◽  
Haotong Qin ◽  
Ke Xia ◽  
Sheng Hu ◽  
...  
Author(s):  
Ke Xia ◽  
Yuqing Ma ◽  
Xianglong Liu ◽  
Yadong Mu ◽  
Li Liu

Author(s):  
Loris Sauter ◽  
Mahnaz Amiri Parian ◽  
Ralph Gasser ◽  
Silvan Heller ◽  
Luca Rossetto ◽  
...  

2010 ◽  
Vol 21 (8) ◽  
pp. 771-772 ◽  
Author(s):  
Meng Wang ◽  
Nicu Sebe ◽  
Tao Mei ◽  
Jia Li ◽  
Kiyoharu Aizawa
Keyword(s):  

Author(s):  
Ke Wang ◽  
Xin Geng

Label Distribution Learning (LDL) is a general learning paradigm in machine learning, which includes both single-label learning (SLL) and multi-label learning (MLL) as its special cases. Recently, many LDL algorithms have been proposed to handle different application tasks such as facial age estimation, head pose estimation and visual sentiment distributions prediction. However, the training time complexity of most existing LDL algorithms is too high, which makes them unapplicable to large-scale LDL. In this paper, we propose a novel LDL method to address this issue, termed Discrete Binary Coding based Label Distribution Learning (DBC-LDL). Specifically, we design an efficiently discrete coding framework to learn binary codes for instances. Furthermore, both the pair-wise semantic similarities and the original label distributions are integrated into this framework to learn highly discriminative binary codes. In addition, a fast approximate nearest neighbor (ANN) search strategy is utilized to predict label distributions for testing instances. Experimental results on five real-world datasets demonstrate its superior performance over several state-of-the-art LDL methods with the lower time cost.


2021 ◽  
Vol 7 (5) ◽  
pp. 76
Author(s):  
Giuseppe Amato ◽  
Paolo Bolettieri ◽  
Fabio Carrara ◽  
Franca Debole ◽  
Fabrizio Falchi ◽  
...  

This paper describes in detail VISIONE, a video search system that allows users to search for videos using textual keywords, the occurrence of objects and their spatial relationships, the occurrence of colors and their spatial relationships, and image similarity. These modalities can be combined together to express complex queries and meet users’ needs. The peculiarity of our approach is that we encode all information extracted from the keyframes, such as visual deep features, tags, color and object locations, using a convenient textual encoding that is indexed in a single text retrieval engine. This offers great flexibility when results corresponding to various parts of the query (visual, text and locations) need to be merged. In addition, we report an extensive analysis of the retrieval performance of the system, using the query logs generated during the Video Browser Showdown (VBS) 2019 competition. This allowed us to fine-tune the system by choosing the optimal parameters and strategies from those we tested.


Author(s):  
Qiang Fu ◽  
Xu Han ◽  
Xianglong Liu ◽  
Jingkuan Song ◽  
Cheng Deng

Building multiple hash tables has been proven a successful technique for indexing massive databases, which can guarantee a desired level of overall performance. However, existing hash based multi-indexing methods suffer from the heavy redundancy, without strong table complementarity and effective hash code learning. To address the problems, this paper proposes a complementary binary quantization (CBQ) method to jointly learning multiple hash tables. It exploits the power of incomplete binary coding based on prototypes to align the original space and the Hamming space, and further utilizes the nature of multi-indexing search to jointly reduce the quantization loss based on the prototype based hash function. Our alternating optimization adaptively discovers the complementary prototype sets and the corresponding code sets of a varying size in an efficient way, which together robustly approximate the data relations. Our method can be naturally generalized to the product space for long hash codes. Extensive experiments carried out on two popular large-scale tasks including Euclidean and semantic nearest neighbor search demonstrate that the proposed CBQ method enjoys the strong table complementarity and significantly outperforms the state-of-the-art, with up to 57.76\% performance gains relatively.


2021 ◽  
Author(s):  
Aozhu Chen ◽  
Fan Hu ◽  
Zihan Wang ◽  
Fangming Zhou ◽  
Xirong Li

Sign in / Sign up

Export Citation Format

Share Document