automatic image annotation
Recently Published Documents


TOTAL DOCUMENTS

382
(FIVE YEARS 38)

H-INDEX

29
(FIVE YEARS 2)

IEEE Access ◽  
2022 ◽  
pp. 1-1
Author(s):  
Myasar Mundher Adnan ◽  
Mohd Shafry Mohd Rahim ◽  
AR Khan ◽  
Tanzila Saba ◽  
Suliman Mohamed Fati ◽  
...  

PLoS ONE ◽  
2021 ◽  
Vol 16 (12) ◽  
pp. e0260758
Author(s):  
Zhiqiang Chen ◽  
Leelavathi Rajamanickam ◽  
Jianfang Cao ◽  
Aidi Zhao ◽  
Xiaohui Hu

This study aims to solve the overfitting problem caused by insufficient labeled images in the automatic image annotation field. We propose a transfer learning model called CNN-2L that incorporates the label localization strategy described in this study. The model consists of an InceptionV3 network pretrained on the ImageNet dataset and a label localization algorithm. First, the pretrained InceptionV3 network extracts features from the target dataset that are used to train a specific classifier and fine-tune the entire network to obtain an optimal model. Then, the obtained model is used to derive the probabilities of the predicted labels. For this purpose, we introduce a squeeze and excitation (SE) module into the network architecture that augments the useful feature information, inhibits useless feature information, and conducts feature reweighting. Next, we perform label localization to obtain the label probabilities and determine the final label set for each image. During this process, the number of labels must be determined. The optimal K value is obtained experimentally and used to determine the number of predicted labels, thereby solving the empty label set problem that occurs when the predicted label values of images are below a fixed threshold. Experiments on the Corel5k multilabel image dataset verify that CNN-2L improves the labeling precision by 18% and 15% compared with the traditional multiple-Bernoulli relevance model (MBRM) and joint equal contribution (JEC) algorithms, respectively, and it improves the recall by 6% compared with JEC. Additionally, it improves the precision by 20% and 11% compared with the deep learning methods Weight-KNN and adaptive hypergraph learning (AHL), respectively. Although CNN-2L fails to improve the recall compared with the semantic extension model (SEM), it improves the comprehensive index of the F1 value by 1%. The experimental results reveal that the proposed transfer learning model based on a label localization strategy is effective for automatic image annotation and substantially boosts the multilabel image annotation performance.


2021 ◽  
Vol 11 (21) ◽  
pp. 10176
Author(s):  
Ramla Bensaci ◽  
Belal Khaldi ◽  
Oussama Aiadi ◽  
Ayoub Benchabana

Automatic image annotation is an active field of research in which a set of annotations are automatically assigned to images based on their content. In literature, some works opted for handcrafted features and manual approaches of linking concepts to images, whereas some others involved convolutional neural networks (CNNs) as black boxes to solve the problem without external interference. In this work, we introduce a hybrid approach that combines the advantages of both CNN and the conventional concept-to-image assignment approaches. J-image segmentation (JSEG) is firstly used to segment the image into a set of homogeneous regions, then a CNN is employed to produce a rich feature descriptor per area, and then, vector of locally aggregated descriptors (VLAD) is applied to the extracted features to generate compact and unified descriptors. Thereafter, the not too deep clustering (N2D clustering) algorithm is performed to define local manifolds constituting the feature space, and finally, the semantic relatedness is calculated for both image–concept and concept–concept using KNN regression to better grasp the meaning of concepts and how they relate. Through a comprehensive experimental evaluation, our method has indicated a superiority over a wide range of recent related works by yielding F1 scores of 58.89% and 80.24% with the datasets Corel 5k and MSRC v2, respectively. Additionally, it demonstrated a relatively high capacity of learning more concepts with higher accuracy, which results in N+ of 212 and 22 with the datasets Corel-5k and MSRC v2, respectively.


Coral Reefs ◽  
2021 ◽  
Author(s):  
Elena Bollati ◽  
Cecilia D’Angelo ◽  
David I. Kline ◽  
B. Greg Mitchell ◽  
Jörg Wiedenmann

AbstractBenthic surveys are a key component of monitoring and conservation efforts for coral reefs worldwide. While traditional image-based surveys rely on manual annotation of photographs to characterise benthic composition, automatic image annotation based on computer vision is becoming increasingly common. However, accurate classification of some benthic groups from reflectance images presents a challenge to local ecologists and computers alike. Most coral reef organisms produce one or a combination of fluorescent pigments, such as Green Fluorescent Protein (GFP)-like proteins found in corals, chlorophyll-a found in all photosynthetic organisms, and phycobiliproteins found in red macroalgae, crustose coralline algae (CCA) and cyanobacteria. Building on the potential of these pigments as a target for automatic image annotation, we developed a novel imaging method based on off-the-shelf components to improve classification of coral and other biotic substrates using a multi-excitation fluorescence (MEF) imaging system. We used RGB cameras to image the fluorescence emission of coral and algal pigments stimulated by narrow-waveband blue and green light, and then combined the information into three-channel pseudocolour images. Using a set of a priori rules defined by the relative pixel intensity produced in different channels, the method achieved successful classification of organisms into three categories based on the dominant fluorescent pigment expressed, facilitating discrimination of traditionally problematic groups. This work provides a conceptual foundation for future technological developments that will improve the cost, accuracy and speed of coral reef surveys.


Author(s):  
Zhixin Li ◽  
Lan Lin ◽  
Canlong Zhang ◽  
Huifang Ma ◽  
Weizhong Zhao ◽  
...  

To learn a well-performed image annotation model, a large number of labeled samples are usually required. Although the unlabeled samples are readily available and abundant, it is a difficult task for humans to annotate large numbers of images manually. In this article, we propose a novel semi-supervised approach based on adaptive weighted fusion for automatic image annotation that can simultaneously utilize the labeled data and unlabeled data to improve the annotation performance. At first, two different classifiers, constructed based on support vector machine and covolutional neural network, respectively, are trained by different features extracted from the labeled data. Therefore, these two classifiers are independently represented as different feature views. Then, the corresponding features of unlabeled images are extracted and input into these two classifiers, and the semantic annotation of images can be obtained respectively. At the same time, the confidence of corresponding image annotation can be measured by an adaptive weighted fusion strategy. After that, the images and its semantic annotations with high confidence are submitted to the classifiers for retraining until a certain stop condition is reached. As a result, we can obtain a strong classifier that can make full use of unlabeled data. Finally, we conduct experiments on four datasets, namely, Corel 5K, IAPR TC12, ESP Game, and NUS-WIDE. In addition, we measure the performance of our approach with standard criteria, including precision, recall, F-measure, N+, and mAP. The experimental results show that our approach has superior performance and outperforms many state-of-the-art approaches.


IEEE Access ◽  
2021 ◽  
Vol 9 ◽  
pp. 135742-135754
Author(s):  
Houjie Li ◽  
Wei Li ◽  
Hongda Zhang ◽  
Xin He ◽  
Mingxiao Zheng ◽  
...  

IEEE Access ◽  
2021 ◽  
pp. 1-1
Author(s):  
Myasar Mundher Adnan ◽  
Mohd Shafry Mohd Rahim ◽  
Amjad Rehman ◽  
Zahid Mehmood ◽  
Tanzila Saba ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document