6 results on '"crowdsourced annotations"'
Search Results
2. Structured Prediction of Music Mood with Twin Gaussian Processes
- Author
-
Chapaneri, Santosh, Jayaswal, Deepak, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Weikum, Gerhard, Series editor, Shankar, B. Uma, editor, Ghosh, Kuntal, editor, Mandal, Deba Prasad, editor, Ray, Shubhra Sankar, editor, Zhang, David, editor, and Pal, Sankar K., editor
- Published
- 2017
- Full Text
- View/download PDF
3. Structured Gaussian Process Regression of Music Mood.
- Author
-
Chapaneri, Santosh, Jayaswal, Deepak, Ghosh, Kuntal, and Mitra, Sushmita
- Subjects
- *
KRIGING , *ENVIRONMENTAL music , *GAUSSIAN mixture models , *RECOMMENDER systems , *SINGLES (Sound recordings) , *ACOUSTIC models , *GAUSSIAN processes - Abstract
Modeling the music mood has wide applications in music categorization, retrieval, and recommendation systems; however, it is challenging to computationally model the affective content of music due to its subjective nature. In this work, a structured regression framework is proposed to model the valence and arousal mood dimensions of music using a single regression model at a linear computational cost. To tackle the subjectivity phenomena, a confidence-interval based estimated consensus is computed by modeling the behavior of various annotators (e.g. biased, adversarial) and is shown to perform better than using the average annotation values. For a compact feature representation of music clips, variational Bayesian inference is used to learn the Gaussian mixture model representation of acoustic features and chord-related features are used to improve the valence estimation by probing the chord progressions between chroma frames. The dimensionality of features is further reduced using an adaptive version of kernel PCA. Using an efficient implementation of twin Gaussian process for structured regression, the proposed work achieves a significant improvement in R2 for arousal and valence dimensions relative to state-of-the-art techniques on two benchmark datasets for music mood estimation. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
4. Solution to overcome the sparsity issue of annotated data in medical domain
- Author
-
Appan K. Pujitha and Jayanthi Sivaswamy
- Subjects
learning (artificial intelligence) ,image colour analysis ,neural nets ,image classification ,image segmentation ,medical image processing ,diseases ,annotated data ,medical domain ,machine learning ,developing computer ,diagnosis algorithms ,CAD ,good performance ,medical data ,image level ,data-driven approaches ,deep learning ,data augmentation ,popular solution ,synthetic image generation ,crowdsourced annotations ,interest markings ,pixel-level markings ,generative adversarial network-based solution ,severity level ,crowdsourced region ,synthetically generated data ,colour fundus images ,processed/refined crowdsourced data/synthetic images ,detection performance ,Computational linguistics. Natural language processing ,P98-98.5 ,Computer software ,QA76.75-76.765 - Abstract
Annotations are critical for machine learning and developing computer aided diagnosis (CAD) algorithms. Good performance of CAD is critical to their adoption, which generally rely on training with a wide variety of annotated data. However, a vast amount of medical data is either unlabeled or annotated only at the image-level. This poses a problem for exploring data driven approaches like deep learning for CAD. In this paper, we propose a novel crowdsourcing and synthetic image generation for training deep neural net-based lesion detection. The noisy nature of crowdsourced annotations is overcome by assigning a reliability factor for crowd subjects based on their performance and requiring region of interest markings from the crowd. A generative adversarial network-based solution is proposed to generate synthetic images with lesions to control the overall severity level of the disease. We demonstrate the reliability of the crowdsourced annotations and synthetic images by presenting a solution for training the deep neural network (DNN) with data drawn from a heterogeneous mixture of annotations. Experimental results obtained for hard exudate detection from retinal images show that training with refined crowdsourced data/synthetic images is effective as detection performance in terms of sensitivity improves by 25%/27% over training with just expert-markings.
- Published
- 2018
- Full Text
- View/download PDF
5. Analysing comparative soft biometrics from crowdsourced annotations.
- Author
-
Martinho‐Corbishley, Daniel, Nixon, Mark S., and Carter, John N.
- Abstract
Soft biometrics enable human description and identification from low‐quality surveillance footage. This study premises the design, collection and analysis of a novel crowdsourced dataset of comparative soft biometric body annotations, obtained from a richly diverse set of human annotators. The authors annotate 100 subject images to provide a coherent, in‐depth appraisal of the collected annotations and inferred relative labels. The dataset includes gender as a comparative trait and the authors find that comparative labels characteristically contain additional discriminative information over traditional categorical annotations. Using the authors' pragmatic dataset, semantic recognition is performed by inferring relative biometric signatures using a RankSVM algorithm. This demonstrates a practical scenario, reproducing responses from a video surveillance operator searching for an individual. The approach can reliably return the correct match in the top 7% of results with ten comparisons, or top 13% of results using just five sets of subject comparisons. [ABSTRACT FROM AUTHOR]
- Published
- 2016
- Full Text
- View/download PDF
6. Unconstrained Aerial Scene Recognition with Deep Neural Networks and a New Dataset
- Author
-
Pu Jin, Xiao Xiang Zhu, Lichao Mou, and Yuansheng Hua
- Subjects
010504 meteorology & atmospheric sciences ,Noise measurement ,Computer science ,business.industry ,Deep learning ,Convolutional neural network (CNN) ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,0211 other engineering and technologies ,Pattern recognition ,02 engineering and technology ,01 natural sciences ,Convolutional neural network ,multi-scene recognition in single images ,Image (mathematics) ,large-scale aerial image dataset ,Task analysis ,Benchmark (computing) ,Deep neural networks ,crowdsourced annotations ,Artificial intelligence ,Focus (optics) ,business ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences - Abstract
Aerial scene recognition is a fundamental research problem in interpreting high-resolution aerial imagery. Over the past few years, most studies focus on classifying an image into one scene category, while in real-world scenarios, it is more often that a single image contains multiple scenes. Therefore, in this paper, we investigate a more practical yet underexplored task-multi-scene recognition in single images. To this end, we create a large-scale dataset, called Mul-tiScene dataset, composed of 100,000 unconstrained images each with multiple labels from 36 different scenes. Among these images, 14,000 of them are manually interpreted and assigned ground-truth labels, while the remaining images are provided with crowdsourced labels, which are generated from low-cost but noisy OpenStreetMap (OSM) data. By doing so, our dataset allows two branches of studies: 1) developing novel CNNs for multi-scene recognition and 2) learning with noisy labels. We experiment with extensive baseline models on our dataset to offer a benchmark for multi-scene recognition in single images. Aiming to expedite further researches, we will make our dataset and pre-trained models available11https://github.com/Hua-YS/Multi-Scene-Recognition.
- Published
- 2021
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.