8,477 results on '"Automatic image annotation"'
Search Results
2. Central Attention with Multi-Graphs for Image Annotation.
- Author
-
Liu, Baodi, Liu, Yan, Shao, Qianqian, and Liu, Weifeng
- Abstract
In recent decades, the development of multimedia and computer vision has sparked significant interest among researchers in the field of automatic image annotation. However, much of the research has primarily focused on using a single graph for annotating images in semi-supervised learning. Conversely, numerous approaches have explored the integration of multi-view or image segmentation techniques to create multiple graph structures. Yet, relying solely on a single graph proves to be challenging, as it struggles to capture the complete manifold of structural information. Furthermore, the computational complexity of building multiple graph structures based on multi-view or image segmentation is substantial and time-consuming. To address these issues, we propose a novel method called "Central Attention with Multi-graphs for Image Annotation." Our approach emphasizes the critical role of the central image region in the annotation process. Remarkably, we demonstrate that impressive performance can be achieved by leveraging just two graph structures, composed of central and overall features, in semi-supervised learning. To validate the effectiveness of our proposed method, we conducted a series of experiments on benchmark datasets, including Corel5K, ESPGame, and IAPRTC12. These experiments provide empirical evidence of our method’s capabilities. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. Image description using tags latent concepts in convolutional neural networks
- Author
-
Barati, Alireza, Farsi, Hassan, and Mohamadzadeh, Sajad
- Published
- 2024
- Full Text
- View/download PDF
4. Automated Image Annotation With Novel Features Based on Deep ResNet50-SLT
- Author
-
Myasar Mundher Adnan, Mohd Shafry Mohd Rahim, Amjad Rehman Khan, Ahmed Alkhayyat, Faten S. Alamri, Tanzila Saba, and Saeed Ali Bahaj
- Subjects
Automatic image annotation ,deep learning ,features extraction ,digital learning ,Slantlet transform ,technological development ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Due to their vast size, the growing number of digital images found in personal archives and on websites has become unmanageable, making it challenging to retrieve images from these large databases accurately. While these collections are popular due to their convenience, they often need to be equipped with proper indexing information, making it difficult for users to find what they need. One of the most significant challenges in computer vision and multimedia is image annotation, which involves labeling images with descriptive keywords. However, computers need to possess the capability to understand the essence of images in the same way that humans do, and people can only identify images based on their visual attributes rather than their deeper semantic meaning. Therefore, image annotation requires keywords to effectively communicate the contents of an image to a computer system. However, raw pixels in an image need to provide more information to generate semantic concepts, making image annotation a complex task. Unlike text annotation, where the dictionary linking words to semantics is well established, image annotation lacks a clear definition of “words” or “sentences” that can be associated with the meaning of the image, known as the semantic gap. To address this challenge, this study aimed to characterize image content meaningfully to make information retrieval easier. An improved automatic image annotation (AIA) system was proposed to bridge the semantic gap between low-level computer features and human interpretation of images by assigning one or multiple labels to images. The proposed AIA system can convert raw image pixels into semantic-level concepts, providing a clearer representation of the image content. The study combined the ResNet50 and slantlet transform with word2vec and principal component analysis with t-distributed stochastic neighbor embedding to balance precision and recall. This allowed the researchers to determine the optimal model for the proposed ResNet50-SLT AIA framework. A Word2vec model with ResNet50-SLT was used with principal component analysis and t-distributed stochastic neighbor embedding to improve IA prediction accuracy. The distributed representation approach involved encoding and storing information about image features. The proposed AIA system utilized seq2seq to generate sentences depending on feature vectors. The system was implemented on the most popular datasets (Flickr8k, Corel-5k, ESP-Game). The results showed that the newly developed AIA scheme overcame the computational time complexity associated with most existing image annotation models during the training phase for large datasets. The performance evaluation of the AIA scheme showed its excellent flexibility of annotation, improved accuracy, and reduced computational costs, thus outperforming the existing state-of-the-art methods. In conclusion, this AIA framework can provide immense benefits in accurately selecting and extracting image features and easily retrieving images from large databases. The extracted features can effectively be used to represent the image, thus accelerating the annotation process and minimizing the computational complexity.
- Published
- 2023
- Full Text
- View/download PDF
5. Evaluating the use of Instagram images color histograms and hashtags sets for automatic image annotation
- Author
-
Stamatios Giannoulakis, Nicolas Tsapatsoulis, and Constantinos Djouvas
- Subjects
Instagram hashtags ,Instagram images ,histogram ,Bhattacharyya distance ,word embedding ,automatic image annotation ,Information technology ,T58.5-58.64 - Abstract
Color similarity has been a key feature for content-based image retrieval by contemporary search engines, such as Google. In this study, we compare the visual content information of images, obtained through color histograms, with their corresponding hashtag sets in the case of Instagram posts. In previous studies, we had concluded that less than 25% of Instagram hashtags are related to the actual visual content of the image they accompany. Thus, the use of Instagram images' corresponding hashtags for automatic image annotation is questionable. In this study, we are answering this question through the computational comparison of images' low-level characteristics with the semantic and syntactic information of their corresponding hashtags. The main conclusion of our study on 26 different subjects (concepts) is that color histograms and filtered hashtag sets, although related, should be better seen as a complementary source for image retrieval and automatic image annotation.
- Published
- 2023
- Full Text
- View/download PDF
6. 基于 CenterNet 的半监督起落架自动标注.
- Author
-
方 伟, 汤 淼, 闫文君, and 张婷婷
- Subjects
MODEL airplanes ,SUPERVISED learning ,ANNOTATIONS ,GEARING machinery - Abstract
Copyright of Journal of Ordnance Equipment Engineering is the property of Chongqing University of Technology and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2023
- Full Text
- View/download PDF
7. Automatic Image Annotation Using Quantization Reweighting Function and Graph Neural Networks
- Author
-
Lotfi, Fariba, Jamzad, Mansour, Beigy, Hamid, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Hacid, Hakim, editor, Aldwairi, Monther, editor, Bouadjenek, Mohamed Reda, editor, Petrocchi, Marinella, editor, Faci, Noura, editor, Outay, Fatma, editor, Beheshti, Amin, editor, Thamsen, Lauritz, editor, and Dong, Hai, editor
- Published
- 2022
- Full Text
- View/download PDF
8. Suggesting an Integration System for Image Annotation.
- Author
-
Ghostan Khatchatoorian, Artin and Jamzad, Mansour
- Subjects
SYSTEM integration ,IMAGING systems ,TAGS (Metadata) ,ANNOTATIONS ,VIRTUAL reality ,SUPPLY & demand - Abstract
The number of digital images uploaded in the virtual world is rapidly growing every day. Therefore, an automatic image annotation system that can retrieve information from these images seems to be in high demand. One of the challenges in this field is the imbalanced data sets and the difficulty of successfully learning tags from them. Even if a nearly balanced data set exists for image annotation, it is unlikely to find a single learner, which could learn all tags with the same accuracy. In this paper, we suggest a novel integration system that selects an elite group of models from all existing annotation models and then combines them to take the best advantage of each model's learning technique. The proposed system studies the data sets of selected models without the need for direct access to those data sets. As this algorithm is independent of the annotation models or data sets, it could be used to combine the currently available annotation models and those developed in future, along with their data sets and learning models. We believe the proposed approach has the potential of becoming an integrated ground for automatic image annotation models. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
9. Deep learning-based automatic annotation and online classification of remote multimedia images.
- Author
-
Kang, Sucheng
- Subjects
DEEP learning ,ARTIFICIAL neural networks ,CONVOLUTIONAL neural networks ,FEATURE extraction ,AUTOMATIC classification ,MACHINE learning - Abstract
In this paper, based on in-depth analysis of remote multimedia images, the automatic annotation and classification of graphics are tested and analyzed by algorithms of deep learning. To reduce the time of remote multimedia image labeling and online classification, and improve efficiency, we study the use of deep learning methods to automate annotation and online classification of remote multimedia images. An image is re-labeling algorithm based on modeling the correlation of hidden feature dimensions is proposed to improve the effect of hidden feature models by modeling the correlation between hid feature dimensions. The algorithm constructs the correlation between each pair of dimensions in the hidden features through the outer product operation to form a two-dimensional interactive graph. The information in the interaction graph is refined layer by layer by using the ability of the convolutional neural network to model local features, and finally, a representation of the correlation of all dimensions in the hidden features is formed to realize the re-labeling of social images. The experimental results show that this method can utilize the hidden feature information more effectively and improve the image re-labeling results. The light-weight feature extraction network significantly reduces the number of model parameters at the expense of a small amount of detection accuracy, and the FPN pyramid structure enhances the feature characterization ability of the feature extraction network. The performance is close to that of the Yolo algorithm. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
10. DeepAIA: An Automatic Image Annotation Model Based on Generative Adversarial Networks and Transfer Learning
- Author
-
Abeer Alshehri, Mounira Taileb, and Reem Alotaibi
- Subjects
Automatic image annotation ,convolutional neural network ,generative adversarial network ,transfer learning ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Automatic image annotation (AIA) has been adopted in different applications such as image retrieval and classification. Deep Learning is used in AIA to extract image features and then convert these features into text descriptions and labels. However, conventional AIA models that employ deep learning methods suffer from various shortcomings, such as poor annotation performance. This work proposes an AIA model based on convolutional neural networks (CNNs), generative adversarial networks (GANs), and transfer learning. GANs have attracted a lot of interest because of its ability to generate data without explicitly using probability density. Thus, it has proven its usefulness in image annotation and image augmentation. In this work, an Auxiliary classifier-GAN (ACGAN) has been used, where the discriminator predicts the class of an image rather than taking it as a given input; therefore, the stabilization of the training stage is ensured, and the generation of high-quality images is provided. Transfer learning is also used to enhance the performance of the classification. The proposed model outperforms the best state-of-the-art models in terms of MiAP, F-measure and error rate using ImageClef, ESPGame and IAPR-TC12 datasets.
- Published
- 2022
- Full Text
- View/download PDF
11. An Improved Automatic Image Annotation Approach Using Convolutional Neural Network-Slantlet Transform
- Author
-
Myasar Mundher Adnan, Mohd Shafry Mohd Rahim, Amjad Rehman Khan, Tanzila Saba, Suliman Mohamed Fati, and Saeed Ali Bahaj
- Subjects
Automatic image annotation ,deep learning ,features extraction ,slantlet transform ,technological development ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Every day, websites and personal archives generate an increasing number of photographs. The extent of these archives is unfathomable. The ease of usage of these enormous digital image collections contributes to their popularity. However, not all of these databases provide appropriate indexing data. As a result, it’s tough to find information that the user is interested in. Thus, in order to find information about an image, it is necessary to classify its content in a meaningful way. Image annotation is one of the most difficult issues in computer vision and multimedia research. The objective is to convert an image into a single or numerous labels. This necessitates a grasp of the visual content of an image. The necessity for unambiguous information to build semantic-level concepts from raw image pixels is one of the challenges of image annotation. Unlike text annotation, where a dictionary links words to their meaning, raw picture pixels are insufficient to construct semantic-level notions directly. A simple syntax, on the other hand, is well specified for combining letters to form words and words to form sentences. The automatic feature extraction for automatic annotation was the emphasis of this paper. And they employed a deep learning convolutional neural network to build and improve image coding and annotation capabilities. Performance of the suggested technique on the Corel-5K, ESP-Game, and IAPRTC-12 datasets. Finally, experimental findings on three data sets were used to demonstrate the usefulness of this model for image annotation.
- Published
- 2022
- Full Text
- View/download PDF
12. Design of ISAR Image Annotation System Based on Deep Learning
- Author
-
Li, Bingning, Zhang, Chi, Pei, Wei, Shen, Liang, Kacprzyk, Janusz, Series Editor, Pal, Nikhil R., Advisory Editor, Bello Perez, Rafael, Advisory Editor, Corchado, Emilio S., Advisory Editor, Hagras, Hani, Advisory Editor, Kóczy, László T., Advisory Editor, Kreinovich, Vladik, Advisory Editor, Lin, Chin-Teng, Advisory Editor, Lu, Jie, Advisory Editor, Melin, Patricia, Advisory Editor, Nedjah, Nadia, Advisory Editor, Nguyen, Ngoc Thanh, Advisory Editor, Wang, Jun, Advisory Editor, Liu, Qi, editor, Liu, Xiaodong, editor, Shen, Tao, editor, and Qiu, Xuesong, editor
- Published
- 2021
- Full Text
- View/download PDF
13. A Modality Converting Approach for Image Annotation to Overcome the Inconsistent Labels in Training Data
- Author
-
Suzuki, Tokinori, Ikeda, Daisuke, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Del Bimbo, Alberto, editor, Cucchiara, Rita, editor, Sclaroff, Stan, editor, Farinella, Giovanni Maria, editor, Mei, Tao, editor, Bertini, Marco, editor, Escalante, Hugo Jair, editor, and Vezzani, Roberto, editor
- Published
- 2021
- Full Text
- View/download PDF
14. Automatic Image Annotation: A Review of Recent Advances and Literature
- Author
-
Rajesh, K. V. N., Lalitha Bhaskari, D., Howlett, Robert J., Series Editor, Jain, Lakhmi C., Series Editor, Satapathy, Suresh Chandra, editor, Bhateja, Vikrant, editor, Mohanty, J. R., editor, and Udgata, Siba K., editor
- Published
- 2020
- Full Text
- View/download PDF
15. Two Efficient Image Bag Generators for Multi-instance Multi-label Learning
- Author
-
Bhagat, P. K., Choudhary, Prakash, Singh, Kh Manglem, Barbosa, Simone Diniz Junqueira, Editorial Board Member, Filipe, Joaquim, Editorial Board Member, Ghosh, Ashish, Editorial Board Member, Kotenko, Igor, Editorial Board Member, Zhou, Lizhu, Editorial Board Member, Nain, Neeta, editor, Vipparthi, Santosh Kumar, editor, and Raman, Balasubramanian, editor
- Published
- 2020
- Full Text
- View/download PDF
16. A Tags Mining Approach for Automatic Image Annotation Using Neighbor Images Tree
- Author
-
Maihami, Vafa, Tsihrintzis, George A., Series Editor, Virvou, Maria, Series Editor, Jain, Lakhmi C., Series Editor, Mallick, Pradeep Kumar, editor, Pattnaik, Prasant Kumar, editor, Panda, Amiya Ranjan, editor, and Balas, Valentina Emilia, editor
- Published
- 2020
- Full Text
- View/download PDF
17. Automatic Image Annotation by Sequentially Learning From Multi-Level Semantic Neighborhoods
- Author
-
Houjie Li, Wei Li, Hongda Zhang, Xin He, Mingxiao Zheng, and Haiyu Song
- Subjects
Automatic image annotation ,semantic gap ,nearest neighbor ,weak-labeling ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Automatic image annotation is a key technology in image understanding and pattern recognition, and is becoming increasingly important in order to annotate large-scale images. In the past decade, the nearest neighbor model-based AIA (Automatic image annotation) methods have been proved to be the most successful in all classical models. This model has four major challenges including semantic gap, label-imbalance, wider range labels, and weak-labeling. In this paper, we propose a novel annotation model based on three-pass KNN (k-Nearest Neighbor) to address the aforementioned challenges. The key idea is to identify appropriate neighbors at each pass KNN. In the first pass KNN, we identify the several most relevant categories based on label feature rather than visual feature as traditional models. In the second pass KNN, we determine the relevant images based on multi-modal (visual and textual label) embedding features. As the test image has not been annotated with any label, we propose a pre-annotation strategory before image annotation to improve the semantic level. In the third pass KNN, we capture relevant labels from semantically and visually similar images and propagate them to the given unlabeled image. In contrast with traditional nearest neighbor based methods, our method can inherently alleviate the problems of semantic gap, label-imbalance, and wider range labels. In addition, to alleviate the issue of weak-labeling, we propose label refinement for training images. Extensive experiments on three classical benchmark datasets and MS-COCO demonstrate that the proposed method significantly outperforms the state-of-the-art in terms of per-label and per-image metrics.
- Published
- 2021
- Full Text
- View/download PDF
18. Deep Convolutional Neural Network with KNN Regression for Automatic Image Annotation.
- Author
-
Bensaci, Ramla, Khaldi, Belal, Aiadi, Oussama, and Benchabana, Ayoub
- Subjects
CONVOLUTIONAL neural networks ,PROBLEM solving ,PREHENSION (Physiology) ,ANNOTATIONS ,ALGORITHMS - Abstract
Automatic image annotation is an active field of research in which a set of annotations are automatically assigned to images based on their content. In literature, some works opted for handcrafted features and manual approaches of linking concepts to images, whereas some others involved convolutional neural networks (CNNs) as black boxes to solve the problem without external interference. In this work, we introduce a hybrid approach that combines the advantages of both CNN and the conventional concept-to-image assignment approaches. J-image segmentation (JSEG) is firstly used to segment the image into a set of homogeneous regions, then a CNN is employed to produce a rich feature descriptor per area, and then, vector of locally aggregated descriptors (VLAD) is applied to the extracted features to generate compact and unified descriptors. Thereafter, the not too deep clustering (N2D clustering) algorithm is performed to define local manifolds constituting the feature space, and finally, the semantic relatedness is calculated for both image–concept and concept–concept using KNN regression to better grasp the meaning of concepts and how they relate. Through a comprehensive experimental evaluation, our method has indicated a superiority over a wide range of recent related works by yielding F1 scores of 58.89% and 80.24% with the datasets Corel 5k and MSRC v2, respectively. Additionally, it demonstrated a relatively high capacity of learning more concepts with higher accuracy, which results in N+ of 212 and 22 with the datasets Corel 5k and MSRC v2, respectively. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
19. A Weighted Topic Model Learned From Local Semantic Space for Automatic Image Annotation
- Author
-
Haiyu Song, Pengjie Wang, Jian Yun, Wei Li, Bo Xue, and Gang Wu
- Subjects
Automatic image annotation ,image retrieval ,probabilistic latent semantic analysis ,topic model ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Automatic image annotation plays a significant role in image understanding, retrieval, classification, and indexing. Today, it is becoming increasingly important in order to annotate large-scale social media images from content-sharing websites and social networks. These social images are usually annotated by user-provided low-quality tags. The topic model is considered as a promising method to describe these weak-labeling images by learning latent representations of training samples. The recent annotation methods based on topic models have two shortcomings. First, they are difficult to scale to a large-scale image dataset. Second, they can not be used to online image repository because of continuous addition of new images and new tags. In this paper, we propose a novel annotation method based on topic model, namely local learning-based probabilistic latent semantic analysis (LL-PLSA), to solve the above problems. The key idea is to train a weighted topic model for a given test image on its semantic neighborhood consisting of a fixed number of semantically and visually similar images. This method can scale to a large-scale image database, as training samples involved in modeling are a few nearest neighbors rather than the entire database. Moreover, this proposed topic model, online customized for the test image, naturally addresses the issue of continuous addition of new images and new tags in a database. Extensive experiments on three benchmark datasets demonstrate that the proposed method significantly outperforms the state-of-the-art especially in terms of overall metrics.
- Published
- 2020
- Full Text
- View/download PDF
20. An Efficient and Effective Model Based on Mean Positive Examples for Social Image Annotation
- Author
-
Haiyu Song, Jian Yun, Houjie Li, Mingxiao Zheng, Jinxing Yao, Hailin Lv, and Anqi Fang
- Subjects
Automatic image annotation ,social image ,tag refinement ,semantic gap ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Nowadays, with the rapid growth of imaging and social network, huge volumes of image data are produced and shared on social media. Social image annotation has been an important and challenging task in the fields of computer vision and machine learning, which can facilitate large-scale image retrieval, indexing, and management. The four most challenges of social image annotation are semantic gap, tag refinement, label-imbalance, and annotation efficiency. To address these issues, we propose an efficient and effective annotation method based on the Mean of Positive Examples (MPE) corresponding to each label. First, we refine user-provided noisy tags by our proposed local smoothing process, and consider the refined tags as key features in contrast to the previous methods that consider them as side information, which significantly improves annotation performance. Second, we propose a weighted trans-media similarity measure method that fuses all modality information in identifying proper neighbors, which promotes the semantic level and eases image annotation. Third, our MPE model gives equal importance to all labels, thus, improving the annotation performance of infrequent labels without sacrificing that of frequent labels. Fourth, our MPE model can dramatically decrease space-time overheads, since the time cost of annotating an image is unaffected by the size of the training image dataset, but relying on the size of label vocabulary. Therefore, our proposed method can be applied to real-world large-scale online social image repositories. Extensive experiments on both benchmark datasets demonstrate the effectiveness and efficiency of our MPE model.
- Published
- 2020
- Full Text
- View/download PDF
21. Tags Re-ranking Using Multi-level Features in Automatic Image Annotation
- Author
-
Forogh Ahmadi and Vafa Maihami
- Subjects
automatic image annotation ,low level feature ,tag ranking ,neighbor voting ,Technology (General) ,T1-995 ,Science - Abstract
Automatic image annotation is a process in which computer systems automatically assign the textual tags related with visual content to a query image. In most cases, inappropriate tags generated by the users as well as the images without any tags among the challenges available in this field have a negative effect on the query's result. In this paper, a new method is presented for automatic image annotation with the aim at improving the obtained tags, as well as reducing the effect of unrelated tags. In the proposed method, first, the initial tags are determined by extracting the low-level features of the image and using neighbor voting method. Afterwards, each initial tag is assigned by a degree based on the neighbor image features of the query image. Finally, they will be ranked based on summing the degrees of each tag and the best tags will be selected by removing the unrelated tags. The experiments conducted on the proposed method using the NUS-WIDE dataset and the commonly used evaluation metrics demonstrate the effectiveness of the proposed system compared to the previous works.
- Published
- 2019
22. Automatic Image Annotation Using Improved Wasserstein Generative Adversarial Networks.
- Author
-
Jian Liu and Weisheng Wu
- Subjects
GENERATIVE adversarial networks ,DEEP learning ,ARTIFICIAL neural networks ,CONVOLUTIONAL neural networks ,PROBLEM solving - Abstract
In an image annotation model based on deep learning, the number of neurons in its output layer is proportional to the vocabulary of the annotation, i.e., the model structure changes with a change in the vocabulary, thereby reducing the accuracy of image annotation. To solve this problem, in this study a new annotation model combining the improved Wasserstein generative adversarial network (GAN) and word2vec was proposed. First, the tagged vocabulary was mapped to a fixed multidimensional word vector by word2vec. Second, a neural network model (GAN-IW) was constructed by using the generated confrontation network. It was observed that the number of neurons in the output layer was equal to the dimension of the multidimensional word vector and no longer relevant to the vocabulary. Finally, the model was tested for the Corel 5K and IAPRTC-12 image annotation datasets. Compared to the convolutional neural network regression method, the model accuracy, the recall rate, and the F1 value increased by 16%, 6%, and 9%, respectively, when the model was tested on the Corel 5K dataset. Compared to the two-pass K-nearest neighbor models, our model accuracy, recall rate, and F1 value were increased by 8%, 6%, and 4%, respectively, when the model was tested on the IAPRTC-12 dataset. The experimental results showed that the GAN-IW model can solve the problem of change in the number of output neurons with a change in the vocabulary and the number of labels annotated with each image is adaptive, making the results of model annotation more in line with the actual image annotation. [ABSTRACT FROM AUTHOR]
- Published
- 2021
23. Automatic Image Annotation
- Author
-
Hervé, Nicolas, Boujemaa, Nozha, Oria, Vincent, Section Editor, Satoh, Shin'ichi, Section Editor, Liu, Ling, editor, and Özsu, M. Tamer, editor
- Published
- 2018
- Full Text
- View/download PDF
24. Kernel Based Approaches for Context Based Image Annotatıon
- Author
-
Nair, L. Swati, Manjusha, R., Parameswaran, Latha, Tavares, João Manuel R.S., Series Editor, Jorge, Renato Natal, Series Editor, Hemanth, D. Jude, editor, and Smys, S., editor
- Published
- 2018
- Full Text
- View/download PDF
25. A Content-Based Visual Information Retrieval Approach for Automated Image Annotation
- Author
-
Senthil, Karthik, Arun, Abhi, Sowmya, Kamath S., Kacprzyk, Janusz, Series editor, Pal, Nikhil R., Advisory editor, Bello Perez, Rafael, Advisory editor, Corchado, Emilio S., Advisory editor, Hagras, Hani, Advisory editor, Kóczy, László T., Series editor, Kreinovich, Vladik, Advisory editor, Lin, Chin-Teng, Advisory editor, Lu, Jie, Advisory editor, Melin, Patricia, Advisory editor, Nedjah, Nadia, Advisory editor, Nguyen, Ngoc Thanh, Advisory editor, Wang, Jun, Advisory editor, Sa, Pankaj Kumar, editor, Sahoo, Manmath Narayan, editor, Murugappan, M., editor, Wu, Yulei, editor, and Majhi, Banshidhar, editor
- Published
- 2018
- Full Text
- View/download PDF
26. Image Annotation Using a Semantic Hierarchy
- Author
-
Bouzaieni, Abdessalem, Tabbone, Salvatore, Hutchison, David, Series Editor, Kanade, Takeo, Series Editor, Kittler, Josef, Series Editor, Kleinberg, Jon M., Series Editor, Mattern, Friedemann, Series Editor, Mitchell, John C., Series Editor, Naor, Moni, Series Editor, Pandu Rangan, C., Series Editor, Steffen, Bernhard, Series Editor, Terzopoulos, Demetri, Series Editor, Tygar, Doug, Series Editor, Weikum, Gerhard, Series Editor, Bai, Xiao, editor, Hancock, Edwin R., editor, Ho, Tin Kam, editor, Wilson, Richard C., editor, Biggio, Battista, editor, and Robles-Kelly, Antonio, editor
- Published
- 2018
- Full Text
- View/download PDF
27. Review: Automatic Image Annotation for Semantic Image Retrieval
- Author
-
Abioui, Hasna, Idarrou, Ali, Bouzit, Ali, Mammass, Driss, Hutchison, David, Series Editor, Kanade, Takeo, Series Editor, Kittler, Josef, Series Editor, Kleinberg, Jon M., Series Editor, Mattern, Friedemann, Series Editor, Mitchell, John C., Series Editor, Naor, Moni, Series Editor, Pandu Rangan, C., Series Editor, Steffen, Bernhard, Series Editor, Terzopoulos, Demetri, Series Editor, Tygar, Doug, Series Editor, Weikum, Gerhard, Series Editor, Mansouri, Alamin, editor, El Moataz, Abderrahim, editor, Nouboud, Fathallah, editor, and Mammass, Driss, editor
- Published
- 2018
- Full Text
- View/download PDF
28. ENCAPSULATION OF IMAGE METADATA FOR EASE OF RETRIEVAL AND MOBILITY
- Author
-
Nancy WOODS and Charles ROBERT
- Subjects
automatic image annotation ,image tagging ,metadata ,Information technology ,T58.5-58.64 ,Electronic computers. Computer science ,QA75.5-76.95 - Abstract
Increasing proliferation of images due to multimedia capabilities of hand-held devices has resulted in loss of source information resulting from inherent mobility. These images are cumbersome to search out once stored away from their original source because they drop their descriptive data. This work, developed a model to encapsulate descriptive metadata into the Exif section of image header for effective retrieval and mobility. The resulting metadata used for retrieval purposes was mobile, searchable and non-obstructive.
- Published
- 2019
- Full Text
- View/download PDF
29. Automatic image annotation based on an improved nearest neighbor technique with tag semantic extension model.
- Author
-
Wei, Wei, Wu, Qiong, Chen, Deng, Zhang, Yanduo, Liu, Wei, Duan, Gonghao, and Luo, Xu
- Subjects
CONVOLUTIONAL neural networks ,TAGS (Metadata) ,NEAREST neighbor analysis (Statistics) ,PROBLEM solving ,ANNOTATIONS ,DIGITAL image correlation ,LABELS - Abstract
Nearest Neighbor method (KNN) is a typical method to solve the problem of automatic image annotation (AIA). However, traditional AIA methods based on KNN only consider the relationships among images and labels. In this paper, we propose an improved KNN image annotation method based on a tag semantic extension model (TSEM). Our approach uses the convolutional neural network (CNN) to extract image features and predicts image tags automatically via nearest features. Different from existing work, the proposed method considers correlations among images, correlations between images and labels and those among labels. Additionally, a label quantity prediction (LQP) model is proposed to predict the number of tags, which further improves the tag prediction accuracy. Comparison experiments were performed on three typical image datasets Corel5k, ESP game and laprtc12. Experimental results show that the average F1 of our model is 0.427, which outperforms the state-of-the-art KNN image annotation methods. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
30. Evaluating the use of Instagram images color histograms and hashtags sets for automatic image annotation
- Author
-
Giannoulakis, Stamatios, Tsapatsoulis, Nicolas, and Djouvas, Constantinos
- Subjects
Instagram hashtags ,Instagram images ,Computer and Information Sciences ,Histogram ,Word embedding ,Automatic image annotation ,Bhattacharyya distance ,Natural Sciences - Abstract
Color similarity has been a key feature for content-based image retrieval by contemporary search engines, such as Google. In this study, we compare the visual content information of images, obtained through color histograms, with their corresponding hashtag sets in the case of Instagram posts. In previous studies, we had concluded that less than 25% of Instagram hashtags are related to the actual visual content of the image they accompany. Thus, the use of Instagram images' corresponding hashtags for automatic image annotation is questionable. In this study, we are answering this question through the computational comparison of images' low-level characteristics with the semantic and syntactic information of their corresponding hashtags. The main conclusion of our study on 26 different subjects (concepts) is that color histograms and filtered hashtag sets, although related, should be better seen as a complementary source for image retrieval and automatic image annotation.
- Published
- 2023
31. Deep Convolutional Neural Network with KNN Regression for Automatic Image Annotation
- Author
-
Ramla Bensaci, Belal Khaldi, Oussama Aiadi, and Ayoub Benchabana
- Subjects
automatic image annotation ,image segmentation ,region annotation ,image content understanding ,Technology ,Engineering (General). Civil engineering (General) ,TA1-2040 ,Biology (General) ,QH301-705.5 ,Physics ,QC1-999 ,Chemistry ,QD1-999 - Abstract
Automatic image annotation is an active field of research in which a set of annotations are automatically assigned to images based on their content. In literature, some works opted for handcrafted features and manual approaches of linking concepts to images, whereas some others involved convolutional neural networks (CNNs) as black boxes to solve the problem without external interference. In this work, we introduce a hybrid approach that combines the advantages of both CNN and the conventional concept-to-image assignment approaches. J-image segmentation (JSEG) is firstly used to segment the image into a set of homogeneous regions, then a CNN is employed to produce a rich feature descriptor per area, and then, vector of locally aggregated descriptors (VLAD) is applied to the extracted features to generate compact and unified descriptors. Thereafter, the not too deep clustering (N2D clustering) algorithm is performed to define local manifolds constituting the feature space, and finally, the semantic relatedness is calculated for both image–concept and concept–concept using KNN regression to better grasp the meaning of concepts and how they relate. Through a comprehensive experimental evaluation, our method has indicated a superiority over a wide range of recent related works by yielding F1 scores of 58.89% and 80.24% with the datasets Corel 5k and MSRC v2, respectively. Additionally, it demonstrated a relatively high capacity of learning more concepts with higher accuracy, which results in N+ of 212 and 22 with the datasets Corel 5k and MSRC v2, respectively.
- Published
- 2021
- Full Text
- View/download PDF
32. Automatic image annotation via category labels.
- Author
-
Zhang, Weifeng, Hu, Hua, Hu, Haiyang, and Yu, Jing
- Subjects
ARTIFICIAL neural networks ,LABELS ,IMAGE representation ,PERFORMANCE standards ,ANNOTATIONS - Abstract
Automatic image annotation aims to assign relevant keywords to images and has become a research focus. Although many techniques have been proposed to solve this problem in the last decade, giving promissing performance on standard datasets, we propose a novel automatic image annotation technique in this paper. Our method uses a label transfer mechanism to automatically recommend those promising tags to each image by using the category information of images. As image representation is one of the key technique in image annotation, we use sparse coding based spatial pyramid matching and deep convolutional neural networks to model image features. And metric learning technique is further used to combine these features to achieve more effective image representation in this paper. Experimental results illustrate that the proposed method get similar or better results than the state-of-the-art methods on three standard image datasets. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
33. Multi-modal multi-concept-based deep neural network for automatic image annotation.
- Author
-
Xu, Haijiao, Huang, Changqin, Huang, Xiaodi, and Huang, Muxiong
- Subjects
CONCEPT learning ,COMPUTER vision ,ANNOTATIONS ,IMAGE ,SEMANTICS - Abstract
Automatic Image Annotation (AIA) remains as a challenge in computer vision with real-world applications, due to the semantic gap between high-level semantic concepts and low-level visual appearances. Contextual tags attached to visual images and context semantics among semantic concepts can provide further semantic information to bridge this gap. In order to effectively capture these semantic correlations, we present a novel approach called Multi-modal Multi-concept-based Deep Neural Network (M2-DNN) in this study, which models the correlations of visual images, contextual tags, and multi-concept semantics. Unlike traditional AIA methods, our M2-DNN approach takes into account not only single-concept context semantics, but also multi-concept context semantics with abstract scenes. In our model, a multi-concept such as { "plane" , "buildings" } is viewed as one holistic scene concept for concept learning. Specifically, we first construct a multi-modal Deep Neural Network (DNN) as a concept classifier for visual images and contextual tags, and then employ it to annotate unlabeled images. Second, real-world databases commonly include many difficult concepts that are hard to be recognized, such as concepts with similar appearances, concepts with abstract scenes, and rare concepts. To effectively recognize them, we utilize multi-concept semantics inference and multi-modal correlation learning to refine semantic annotations. Finally, we estimate the most relevant labels for each of unlabeled images through a new strategy of label decision. The results of our comprehensive experiments on two publicly available datasets have shown that our method performs favourably compared with several other state-of-the-art methods. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
34. Automatic image annotation using model fusion and multi-label selection algorithm.
- Author
-
Wang, Liqin, Zhang, Aofan, Wang, Peng, and Dong, Yongfeng
- Subjects
- *
ARTIFICIAL neural networks , *DIGITAL image correlation , *ANNOTATIONS , *IMAGE representation , *LABELS - Abstract
Automatic Image Annotation (AIA) aims to provide a semantic description for the content of image by assigning a set of textual labels. The recent approaches mainly focus on the improvement of single model and neglect the potential advantages of different models. In order to make full use of the advantages of different annotation models, Dual Model based on Multi-Label Selection Algorithm(DM-SA) is proposed in this research which combines a discriminative model with a nearest-neighbor-based model. The algorithm takes consideration of the advantages of each model, thus provides better annotation performance. A deep Convolutional Neural Network (CNN) is used to obtain visual representation of images first, then a discriminative model, CNN with Label Smoothing (CNN-LS), and a nearest-neighbor-based model, 2PKNN with Canonical Correlation Analysis (2PKNN-CCA) generate candidate label set respectively. Finally, a multi-label selection algorithm based on inverse document frequency is adopted to assign the final labels from two candidate label sets. Experimental results based on Corel5K and IAPRTC-12 datasets show that the proposed method can achieve state-of-the-art performance for average recall, 0.52 and 0.42 on Corel5K and IAPRTC-12 respectively. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
35. Unsupervised Clustering of Natural Images in Automatic Image Annotation Systems
- Author
-
Favorskaya, Margarita, Jain, Lakhmi C., Proskurin, Alexander, Kacprzyk, Janusz, Series editor, Jain, Lakhmi C., Series editor, Kountchev, Roumen, editor, and Nakamatsu, Kazumi, editor
- Published
- 2016
- Full Text
- View/download PDF
36. Automatic Image Annotation Based on Semi-supervised Probabilistic CCA
- Author
-
Zhang, Bo, Ma, Gang, Yang, Xi, Shi, Zhongzhi, Hao, Jie, Rannenberg, Kai, Editor-in-chief, Sakarovitch, Jacques, Series editor, Goedicke, Michael, Series editor, Tatnall, Arthur, Series editor, Neuhold, Erich J., Series editor, Pras, Aiko, Series editor, Tröltzsch, Fredi, Series editor, Pries-Heje, Jan, Series editor, Whitehouse, Diane, Series editor, Reis, Ricardo, Series editor, Furnell, Steven, Series editor, Furbach, Ulrich, Series editor, Gulliksen, Jan, Series editor, Rauterberg, Matthias, Series editor, Shi, Zhongzhi, editor, Vadera, Sunil, editor, and Li, Gang, editor
- Published
- 2016
- Full Text
- View/download PDF
37. Deep-LIFT: Deep Label-Specific Feature Learning for Image Annotation
- Author
-
Changqing Zhang, Huazhu Fu, Qinghua Hu, Joey Tianyi Zhou, Junbing Li, and Shuyin Xia
- Subjects
Lift (data mining) ,business.industry ,Computer science ,Pattern recognition ,Computer Science Applications ,Image (mathematics) ,Human-Computer Interaction ,ComputingMethodologies_PATTERNRECOGNITION ,Automatic image annotation ,Control and Systems Engineering ,Feature (computer vision) ,Benchmark (computing) ,Graph (abstract data type) ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,Feature learning ,Algorithms ,Data Curation ,Software ,Information Systems ,Interpretability - Abstract
Image annotation aims to jointly predict multiple tags for an image. Although significant progress has been achieved, existing approaches usually overlook aligning specific labels and their corresponding regions due to the weak supervised information (i.e., ``bag of labels'' for regions), thus failing to explicitly exploit the discrimination from different classes. In this article, we propose the deep label-specific feature (Deep-LIFT) learning model to build the explicit and exact correspondence between the label and the local visual region, which improves the effectiveness of feature learning and enhances the interpretability of the model itself. Deep-LIFT extracts features for each label by aligning each label and its region. Specifically, Deep-LIFTs are achieved through learning multiple correlation maps between image convolutional features and label embeddings. Moreover, we construct two variant graph convolutional networks (GCNs) to further capture the interdependency among labels. Empirical studies on benchmark datasets validate that the proposed model achieves superior performance on multilabel classification over other existing state-of-the-art methods.
- Published
- 2022
38. Co-occurrence patterns based fruit quality detection for hierarchical fruit image annotation
- Author
-
S. Sangita B. Nemade and Shefali Sonavane
- Subjects
Subcategory ,General Computer Science ,Computer science ,business.industry ,Co-occurrence ,020206 networking & telecommunications ,Pattern recognition ,02 engineering and technology ,Support vector machine ,Annotation ,Automatic image annotation ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,F1 score ,business ,Classifier (UML) ,Data objects - Abstract
Automatic image annotation is a method of assigning caption to images that provide some convenient way to index, retrieve and handle a large amount of data objects. It focuses on recent agricultural automation applications; it finds potential in classification along with contextual labeling of the involved objects or detailing based on its statistical properties on fruit categories. However, producing hierarchical labels provide details of a particular fruit subcategory. This paper proposes fruit annotation in a broad sense along with its hierarchical features that can be narrowed down to inherit, further achieving fruit classification into binary or multiple classes indicating subcategories of that fruit. The fruit objects within images are measured to its actual size in the required units. The classification is also used for identifying true color, texture, size, deep features and shape based on the ratio of major to minor axis helpful for fruit gradations. The co-occurrence patterns are obtained based on the visual features of the selected fruit. This is useful for finding the fruit quality categories and combined properties that are used to form the co-occurrence patterns. These patterns are further used by the classifier for fruit annotation. The evaluation of the performance is carried out using the F1 score, accuracy, precision, recall and G-measure. The results show that the co-occurrence pattern with SVM provides an overall accuracy of 97.3% and 97.2% for grape and mango fruit subcategories. The comparative results are obtained to cross-check with the subjective evaluation of gradation validated by local farmers.
- Published
- 2022
39. Automatic Image Annotation for Description of Urban and Outdoor Scenes
- Author
-
Cruz-Perez, Claudia, Starostenko, Oleg, Alarcon-Aquino, Vicente, Rodriguez-Asomoza, Jorge, Sobh, Tarek, editor, and Elleithy, Khaled, editor
- Published
- 2015
- Full Text
- View/download PDF
40. A Hybrid Model Based on Mutual Information and Support Vector Machine for Automatic Image Annotation
- Author
-
Jin, Cong, Liu, Jinan, Guo, Jinglei, Kacprzyk, Janusz, Series editor, Silhavy, Radek, editor, Senkerik, Roman, editor, Oplatkova, Zuzana Kominkova, editor, Prokopova, Zdenka, editor, and Silhavy, Petr, editor
- Published
- 2015
- Full Text
- View/download PDF
41. A Multi-feature Fusion Method for Automatic Multi-label Image Annotation with Weighted Histogram Integral and Closure Regions Counting
- Author
-
Xia, Sen, Chen, Peng, Zhang, Jun, Li, Xiao-Ping, Wang, Bing, Goebel, Randy, Series editor, Tanaka, Yuzuru, Series editor, Wahlster, Wolfgang, Series editor, Huang, De-Shuang, editor, and Han, Kyungsook, editor
- Published
- 2015
- Full Text
- View/download PDF
42. Image Annotation Based on Multi-view Learning
- Author
-
Shi, Zhe, Zhu, Songhao, Sun, Chengjian, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Weikum, Gerhard, Series editor, and Zhang, Yu-Jin, editor
- Published
- 2015
- Full Text
- View/download PDF
43. Automatic Image Annotation Based on Multi-Auxiliary Information
- Author
-
Pengyu Zhang, Zhihua Wei, Yunyi Li, and Cairong Zhao
- Subjects
Automatic image annotation ,extended tag set multi-auxiliary information ,tag refinement ,visibility ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
This paper introduces an automatic image annotation framework based on multi-auxiliary information which aims at improving the annotation performance. We propose three novel ideas in the framework of annotation: 1) multi-information extraction: besides various visual features, tag co-occurrence, and user interest vector are added to enrich the multi-auxiliary information; 2) initial labeling: based on the traditional term frequency-inverse document frequency model-we utilize the visibility of words and extended tag set to enhance the result of initial labeling and propose a more efficient model, TF-IDF, visibility and extended tag set model; and 3) tag refinement: by considering multi-auxiliary information, including multi-visual content, tag co-occurrence, and user interest similarity, we propose the multi-information alllabels model for tag refinement. The tag refinement process is formalized as an optimization problem by adjusting confidence score set by the initial labeling model. Experimental results demonstrate that, compared with the state-of-the-art methods, our method achieves the best performance on MIR-Flickr data sets, outperforming the second best by 2%.
- Published
- 2017
- Full Text
- View/download PDF
44. Design and Implementation of a Vision- and Grating-Sensor-Based Intelligent Unmanned Settlement System
- Author
-
Hong-Bo Zhang, Qing Lei, Li-Jia Dong, Ji-Xiang Du, and Zhou Yizhong
- Subjects
Automatic image annotation ,Invoice ,Computer science ,Event (computing) ,business.industry ,Computer vision ,Workload ,Artificial intelligence ,Architecture ,business ,Facial recognition system ,Mobile device ,Convolutional neural network - Abstract
In this paper, a new vision- and grating-sensor-based intelligent unmanned settlement (IUS) system is proposed for convenience stores to automatically recognize the shopping behavior of customers, record their identities, and generate invoices. First, we design a new IUS architecture, which includes a shelf module and exit module. To achieve automatic settlement for each customer, a shopping event detection method is proposed. In this method, a vision-based human pose estimation algorithm is used to detect a human form standing in front of a shelf. The hand actions of each customer are detected by a grating sensor, and an image recognition method based on a convolutional neural network (CNN) is applied to recognize the items in the hands of customers. To reduce the image annotation workload, we propose a semisupervised training method for the recognition network. Based on hand action detection and item recognition, a shopping event recognition method is designed for the system, and a facial image of the customer corresponding to each shopping behavior is captured. Finally, each detected shopping event is added to the invoice of the corresponding customer via a facial recognition method. To verify the effectiveness of the proposed IUS system, we have built a handheld item image dataset and a shopping event dataset for an unmanned convenience store. The experimental results show that the proposed system can accurately recognize shopping behaviors and generate invoices.
- Published
- 2022
45. Learning cross space mapping via DNN using large scale click-through logs
- Author
-
Kuiyuan Yang, Yalong Bai, Wei Yu, Yong Rui, and Hongxun Yao
- Subjects
FOS: Computer and information sciences ,Similarity (geometry) ,Artificial neural network ,Contextual image classification ,Computer science ,business.industry ,Computer Vision and Pattern Recognition (cs.CV) ,Feature extraction ,Computer Science - Computer Vision and Pattern Recognition ,Pattern recognition ,Computer Science Applications ,Visualization ,Automatic image annotation ,Signal Processing ,Media Technology ,Visual Word ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,Image retrieval - Abstract
The gap between low-level visual signals and high-level semantics has been progressively bridged by continuous development of deep neural network (DNN). With recent progress of DNN, almost all image classification tasks have achieved new records of accuracy. To extend the ability of DNN to image retrieval tasks, we proposed a unified DNN model for image-query similarity calculation by simultaneously modeling image and query in one network. The unified DNN is named the cross space mapping (CSM) model, which contains two parts, a convolutional part and a query-embedding part. The image and query are mapped to a common vector space via these two parts respectively, and image-query similarity is naturally defined as an inner product of their mappings in the space. To ensure good generalization ability of the DNN, we learn weights of the DNN from a large number of click-through logs which consists of 23 million clicked image-query pairs between 1 million images and 11.7 million queries. Both the qualitative results and quantitative results on an image retrieval evaluation task with 1000 queries demonstrate the superiority of the proposed method., Accepted by IEEE Transactions on Multimedia 2015
- Published
- 2023
46. Computer Image Generation
- Author
-
Jon Peddie
- Subjects
Real-time computer graphics ,Digital image ,Automatic image annotation ,Computer science ,Computer image generation ,Computer graphics (images) ,Digital image processing ,Image processing ,Image-based modeling and rendering - Published
- 2023
47. Two Approaches for Mobile Phone Image Insignia Identification
- Author
-
Mishra, Nitin, Kopparapu, Sunil Kumar, Kacprzyk, Janusz, Series editor, Thampi, Sabu M., editor, Gelbukh, Alexander, editor, and Mukhopadhyay, Jayanta, editor
- Published
- 2014
- Full Text
- View/download PDF
48. Chained ensemble classifier for image annotation.
- Author
-
Marin-Castro, Heidy M., Hernandez-Resendiz, Jaciel D., Escalante-Balderas, Hugo J., Pellegrin, Luis, and Tello-Leal, Edgar
- Subjects
INFORMATION organization ,ANNOTATIONS ,MACHINE learning ,IMAGE ,DESCRIPTOR systems - Abstract
Image annotation is the task of assigning keywords or identifiers to images, holistically or in specific regions. These keywords serve as descriptors of high-level semantics to facilitate retrieval and organization of visual information. It plays an important role in content-based image understanding, as well as in areas such as object recognition in robotics, content-based image searching and knowledge extraction. Automatic image annotation is usually approached by means of supervised classification, where a set of previously annotated images is required to train a learning algorithm that later predicts the labels for new images. This paper proposes a novel ensemble classifier for the supervised image annotation task inspired in chain classifiers. In the proposed approach a chain of individual classifiers is build, where each classifier is trained by using a different modality. In addition, the input space of models in the chain is augmented with the output of the preceding model in the sequence. Each model in the chain deals with the same classification problem, making the proposed method an ensemble model build from multimodal data. To the best of our knowledge, chain classifiers have not been used in this particular setting. Experimental results in a challenging image collection show that the proposed method is able to obtain an f −value superior to 0.5, outperforming related work. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
49. Semi-supervised dual low-rank feature mapping for multi-label image annotation.
- Author
-
Wang, Xiaoying, Feng, Songhe, and Lang, Congyan
- Subjects
LABELS ,ANNOTATIONS ,MATHEMATICAL regularization ,LEARNING problems ,IMAGE ,TRAINING needs - Abstract
Automatic image annotation as a typical multi-label learning problem, has gained extensive attention in recent years owing to its application in image semantic understanding and relevant disciplines. Nevertheless, existing annotation methods share the same challenge that labels annotated on the training images are usually incomplete and unclean, while the need for adequate training data is costly and unrealistic. Being aware of this, we propose a dual low-rank regularized multi-label learning model under a graph regularized semi-supervised learning framework, which can effectively capture the label correlations in the learned feature space, and enforce the label matrix be self-recovered in label space as well. To be specific, the proposed approach firstly puts forward a label matrix refinement approach, by introducing a label coefficient matrix to build a linear self-recovery model. Then, graph Laplacian regularization is introduced to make use of a large number of unlabeled images by enforcing the local geometric structure on both labeled and unlabeled images. Lastly, we exploit dual trace norm regularization on both feature mapping matrix and self-recovery coefficient matrix to capture the correlations among different labels in both feature space and label space, and control the model complexity as well. Empirical studies on four real-world image datasets demonstrate the effectiveness and efficiency of the proposed framework. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
50. The Deep Features and Attention Mechanism-Based Method to Dish Healthcare Under Social IoT Systems: An Empirical Study With a Hand-Deep Local–Global Net
- Author
-
Min Cao, Yuyu Yin, Kaili Xu, Junsheng Xiao, Honghao Gao, and Qiang Xu
- Subjects
business.industry ,Computer science ,Taste (sociology) ,media_common.quotation_subject ,Mechanism based ,Data science ,Human-Computer Interaction ,Empirical research ,Automatic image annotation ,Modeling and Simulation ,Health care ,Key (cryptography) ,Feature (machine learning) ,business ,Internet of Things ,Social Sciences (miscellaneous) ,media_common - Abstract
Many mobile apps of social Internet of Things (sIOT) systems can help us record and share daily events, such as health and sport events. In fact, healthy diet recognition is an important and challenging problem in dish health assessment. Via the collection and monitoring of data pertaining to our daily diet, we can work in collaborative ways to achieve dish image annotation based on sIOT systems to enhance deep features. To this end, this article proposes a deep feature and attention mechanism-based method for dish health assessment, which aims to apply a hand-deep local-global net (HDLGN) for dish image recognition. Then, food taste is used as health guidance for people who want to lose weight or follow doctors' advice. First, the local attention mechanism is introduced to identify key areas of the dish image. Second, ingredient and handcrafted color features are extracted to learn deep features. Subsequently, we combine local and global attention mechanisms to return the dish taste as the recognition result. Finally, experiments show that our proposed method can effectively improve the accuracy of taste recognition.
- Published
- 2022
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.