378 results on '"Automatic annotation"'
Search Results
2. The Browser-Based GLAUx Treebank Infrastructure: Framework, Functionality, and Future.
- Author
-
Keersmaekers, Alek, Pietowski, Frédéric, Van Hal, Toon, and Depauw, Mark
- Subjects
CORPORA ,GREEK language ,DATABASE searching ,SQL ,SYNTAX (Grammar) - Abstract
This paper presents the browser-based treebank infrastructure of GLAUx (the Greek Language AUtomated). This linguistic annotation project now has its integrated and user-friendly platform for exploring this data. After discussing the size and types of texts included in the GLAUx corpus, the contribution succinctly surveys the types of linguistic annotation covered by the project (morphology, lemmatization, and syntax). The emphasis of the contribution is on a description of the underlying SQL database structure and the search architecture. Infrastructure-related challenges faced by the GLAUx project are also discussed. Finally, the paper concludes with a discussion of future steps for the project, including additional functionality and expansion of the corpus. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. Fine mapping of RNA isoform diversity using an innovative targeted long-read RNA sequencing protocol with novel dedicated bioinformatics pipeline
- Author
-
Camille Aucouturier, Nicolas Soirat, Laurent Castéra, Denis Bertrand, Alexandre Atkinson, Thibaut Lavolé, Nicolas Goardon, Céline Quesnelle, Julien Levilly, Sosthène Barbachou, Angelina Legros, Olivier Caron, Louise Crivelli, Philippe Denizeau, Pascaline Berthet, Agathe Ricou, Flavie Boulouard, Dominique Vaur, Sophie Krieger, and Raphael Leman
- Subjects
RNA splicing ,Long read sequencing ,HBOC ,Isoform assembly ,Automatic annotation ,Biotechnology ,TP248.13-248.65 ,Genetics ,QH426-470 - Abstract
Abstract Background Solving the structure of mRNA transcripts is a major challenge for both research and molecular diagnostic purposes. Current approaches based on short-read RNA sequencing and RT-PCR techniques cannot fully explore the complexity of transcript structure. The emergence of third-generation long-read sequencing addresses this problem by solving this sequence directly. However, genes with low expression levels are difficult to study with the whole transcriptome sequencing approach. To fix this technical limitation, we propose a novel method to capture transcripts of a gene panel using a targeted enrichment approach suitable for Pacific Biosciences and Oxford Nanopore Technologies platforms. Results We designed a set of probes to capture transcripts of a panel of genes involved in hereditary breast and ovarian cancer syndrome. We present SOSTAR (iSofOrmS annoTAtoR), a versatile pipeline to assemble, quantify and annotate isoforms from long read sequencing using a new tool specially designed for this application. The significant enrichment of transcripts by our capture protocol, together with the SOSTAR annotation, allowed the identification of 1,231 unique transcripts within the gene panel from the eight patients sequenced. The structure of these transcripts was annotated with a resolution of one base relative to a reference transcript. All major alternative splicing events of the BRCA1 and BRCA2 genes described in the literature were found. Complex splicing events such as pseudoexons were correctly annotated. SOSTAR enabled the identification of abnormal transcripts in the positive controls. In addition, a case of unexplained inheritance in a family with a history of breast and ovarian cancer was solved by identifying an SVA retrotransposon in intron 13 of the BRCA1 gene. Conclusions We have validated a new protocol for the enrichment of transcripts of interest using probes adapted to the ONT and PacBio platforms. This protocol allows a complete description of the alternative structures of transcripts, the estimation of their expression and the identification of aberrant transcripts in a single experiment. This proof-of-concept opens new possibilities for RNA structure exploration in both research and molecular diagnostics.
- Published
- 2024
- Full Text
- View/download PDF
4. 电磁大数据自动化标注补全算法.
- Author
-
王 娜, 杨君子, and 邵怀宗
- Subjects
LOW-rank matrices ,MEAN square algorithms ,STANDARD deviations ,SPARSE matrices ,PROBLEM solving - Abstract
Copyright of Telecommunication Engineering is the property of Telecommunication Engineering and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2024
- Full Text
- View/download PDF
5. Efficient Real-Time Droplet Tracking in Crop-Spraying Systems.
- Author
-
Huynh, Truong Nhut, Burgers, Travis, and Nguyen, Kim-Doang
- Subjects
SUSTAINABILITY ,ENVIRONMENTAL responsibility ,PEST control ,AGRICULTURAL innovations ,SPRAY droplet drift - Abstract
Spray systems in agriculture serve essential roles in the precision application of pesticides, fertilizers, and water, contributing to effective pest control, nutrient management, and irrigation. These systems enhance efficiency, reduce labor, and promote environmentally friendly practices by minimizing chemical waste and runoff. The efficacy of a spray is largely determined by the characteristics of its droplets, including their size and velocity. These parameters are not only pivotal in assessing spray retention, i.e., how much of the spray adheres to crops versus becoming environmental runoff, but also in understanding spray drift dynamics. This study introduces a real-time deep learning-based approach for droplet detection and tracking which significantly improves the accuracy and efficiency of measuring these droplet properties. Our methodology leverages advanced AI techniques to overcome the limitations of previous tracking frameworks, employing three novel deep learning-based tracking methods. These methods are adept at handling challenges such as droplet occlusion and varying velocities, ensuring precise tracking in real-time potentially on mobile platforms. The use of a high-speed camera operating at 2000 frames per second coupled with innovative automatic annotation tools enables the creation of a large and accurately labeled droplet dataset for training and evaluation. The core of our framework lies in the ability to track droplets across frames, associating them temporally despite changes in appearance or occlusions. We utilize metrics including Multiple Object Tracking Accuracy (MOTA) and Multiple Object Tracking Precision (MOTP) to quantify the tracking algorithm's performance. Our approach is set to pave the way for innovations in agricultural spraying systems, offering a more efficient, accurate, and environmentally responsible method of applying sprays and representing a significant step toward sustainable agricultural practices. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
6. Crop Growth Analysis Using Automatic Annotations and Transfer Learning in Multi-Date Aerial Images and Ortho-Mosaics.
- Author
-
Rana, Shubham, Gerbino, Salvatore, Akbari Sekehravani, Ehsan, Russo, Mario Brandon, and Carillo, Petronia
- Subjects
- *
CROP growth , *SUSTAINABILITY , *CROP management , *IMAGE analysis , *RESOURCE allocation - Abstract
Growth monitoring of crops is a crucial aspect of precision agriculture, essential for optimal yield prediction and resource allocation. Traditional crop growth monitoring methods are labor-intensive and prone to errors. This study introduces an automated segmentation pipeline utilizing multi-date aerial images and ortho-mosaics to monitor the growth of cauliflower crops (Brassica Oleracea var. Botrytis) using an object-based image analysis approach. The methodology employs YOLOv8, a Grounding Detection Transformer with Improved Denoising Anchor Boxes (DINO), and the Segment Anything Model (SAM) for automatic annotation and segmentation. The YOLOv8 model was trained using aerial image datasets, which then facilitated the training of the Grounded Segment Anything Model framework. This approach generated automatic annotations and segmentation masks, classifying crop rows for temporal monitoring and growth estimation. The study's findings utilized a multi-modal monitoring approach to highlight the efficiency of this automated system in providing accurate crop growth analysis, promoting informed decision-making in crop management and sustainable agricultural practices. The results indicate consistent and comparable growth patterns between aerial images and ortho-mosaics, with significant periods of rapid expansion and minor fluctuations over time. The results also indicated a correlation between the time and method of observation which paves a future possibility of integration of such techniques aimed at increasing the accuracy in crop growth monitoring based on automatically derived temporal crop row segmentation masks. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
7. ScoreREM: A user-friendly Matlab-GUI for rapid eye movement (REM) sleep microstructure (Phasic/Tonic) annotation and quantification
- Author
-
Kiran K G Ravindran and Derk-Jan Dijk
- Subjects
Sleep scoring ,Automatic annotation ,Graphical user interface (GUI) ,Polysomnography (PSG) ,Rapid eye movement (REM) ,MATLAB ,Computer software ,QA76.75-76.765 - Abstract
Rapid eye movement (REM) sleep plays a crucial role in brain functions such as memory consolidation and mood regulation. Alterations in REM sleep and REM sleep behaviour disorders are early biomarkers of neurodegenerative disorders. The microstructure of REM sleep consists of two states: phasic and tonic REM sleep. Quantification of phasic and tonic REM sleep is increasingly pursued. Here we introduce, ScoreREM, an open-source user-friendly MATLAB graphical user interface (GUI) designed for rapid annotation and quantification of Phasic and Tonic REM sleep. Due to its intuitive functionalities and vast array of potential application areas, ScoreREM will serve as an indispensable tool for rapid and accurate REM microstructure quantification in clinical studies.
- Published
- 2025
- Full Text
- View/download PDF
8. Satellite Image Cloud Automatic Annotator with Uncertainty Estimation.
- Author
-
Gao, Yijiang, Shao, Yang, Jiang, Rui, Yang, Xubing, and Zhang, Li
- Subjects
- *
MACHINE learning , *REMOTE sensing , *DEEP learning , *IMAGE analysis , *LANDSAT satellites , *REMOTE-sensing images - Abstract
In satellite imagery, clouds obstruct the ground information, directly impacting various downstream applications. Thus, cloud annotation/cloud detection serves as the initial preprocessing step in remote sensing image analysis. Recently, deep learning methods have significantly improved in the field of cloud detection, but training these methods necessitates abundant annotated data, which requires experts with professional domain knowledge. Moreover, the influx of remote sensing data from new satellites has further led to an increase in the cost of cloud annotation. To address the dependence on labeled datasets and professional domain knowledge, this paper proposes an automatic cloud annotation method for satellite remote sensing images, CloudAUE. Unlike traditional approaches, CloudAUE does not rely on labeled training datasets and can be operated by users without domain expertise. To handle the irregular shapes of clouds, CloudAUE firstly employs a convex hull algorithm for selecting cloud and non-cloud regions by polygons. When selecting convex hulls, the cloud region is first selected, and points at the edges of the cloud region are sequentially selected as polygon vertices to form a polygon that includes the cloud region. Then, the same selection is performed on non-cloud regions. Subsequently, the fast KD-Tree algorithm is used for pixel classification. Finally, an uncertainty method is proposed to evaluate the quality of annotation. When the confidence value of the image exceeds a preset threshold, the annotation process terminates and achieves satisfactory results. When the value falls below the threshold, the image needs to undergo a subsequent round of annotation. Through experiments on two labeled datasets, HRC and Landsat 8, CloudAUE demonstrates comparable or superior accuracy to deep learning algorithms, and requires only one to two annotations to obtain ideal results. An unlabeled self-built Google Earth dataset is utilized to validate the effectiveness and generalizability of CloudAUE. To show the extension capabilities in various fields, CloudAUE also achieves desirable results on a forest fire dataset. Finally, some suggestions are provided to improve annotation performance and reduce the number of annotations. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
9. 面向遥感图像目标感知的群目标检测框架.
- Author
-
张, 鸿伟, 金, 磊, 邹, 学超, 方, 宇强, 尹, 璐, 赵, 健, and 兴, 军亮
- Subjects
OPTICAL remote sensing ,REMOTE sensing ,RECONNAISSANCE operations ,AEROSPACE technology ,VISIBLE spectra - Abstract
Copyright of Journal of Remote Sensing is the property of Editorial Office of Journal of Remote Sensing & Science Publishing Co. and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2024
- Full Text
- View/download PDF
10. Initial Steps Required for Archaeological Image Processing
- Author
-
Ben Salah, Marwa, Yengui, Ameni, Neji, Mahmoud, Kacprzyk, Janusz, Series Editor, Gomide, Fernando, Advisory Editor, Kaynak, Okyay, Advisory Editor, Liu, Derong, Advisory Editor, Pedrycz, Witold, Advisory Editor, Polycarpou, Marios M., Advisory Editor, Rudas, Imre J., Advisory Editor, Wang, Jun, Advisory Editor, Elhadj, Yahya Mohamed, editor, Nanne, Mohamedade Farouk, editor, Koubaa, Anis, editor, Meziane, Farid, editor, and Deriche, Mohamed, editor
- Published
- 2024
- Full Text
- View/download PDF
11. Investigating Low-Cost LLM Annotation for Spoken Dialogue Understanding Datasets
- Author
-
Druart, Lucas, Vielzeuf, Valentin, Estève, Yannick, Goos, Gerhard, Series Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Nöth, Elmar, editor, Horák, Aleš, editor, and Sojka, Petr, editor
- Published
- 2024
- Full Text
- View/download PDF
12. A Diverse Environment Coal Gangue Image Segmentation Model Combining Improved U-Net and Semi-supervised Automatic Annotation
- Author
-
Liu, Xiuhua, Zhu, Wenbo, Zhu, Zhengjun, Luo, Lufeng, Zhang, Yunzhi, Lu, Qinghua, Filipe, Joaquim, Editorial Board Member, Ghosh, Ashish, Editorial Board Member, Prates, Raquel Oliveira, Editorial Board Member, Zhou, Lizhu, Editorial Board Member, Sun, Fuchun, editor, Meng, Qinghu, editor, Fu, Zhumu, editor, and Fang, Bin, editor
- Published
- 2024
- Full Text
- View/download PDF
13. Automatic de-identification of French electronic health records: a cost-effective approach exploiting distant supervision and deep learning models
- Author
-
Mohamed El Azzouzi, Gouenou Coatrieux, Reda Bellafqira, Denis Delamarre, Christine Riou, Naima Oubenali, Sandie Cabon, Marc Cuggia, and Guillaume Bouzillé
- Subjects
Clinical de-identification ,Distant supervision ,Automatic annotation ,Named entity recognition ,Word representations ,Deep learning ,Computer applications to medicine. Medical informatics ,R858-859.7 - Abstract
Abstract Background Electronic health records (EHRs) contain valuable information for clinical research; however, the sensitive nature of healthcare data presents security and confidentiality challenges. De-identification is therefore essential to protect personal data in EHRs and comply with government regulations. Named entity recognition (NER) methods have been proposed to remove personal identifiers, with deep learning-based models achieving better performance. However, manual annotation of training data is time-consuming and expensive. The aim of this study was to develop an automatic de-identification pipeline for all kinds of clinical documents based on a distant supervised method to significantly reduce the cost of manual annotations and to facilitate the transfer of the de-identification pipeline to other clinical centers. Methods We proposed an automated annotation process for French clinical de-identification, exploiting data from the eHOP clinical data warehouse (CDW) of the CHU de Rennes and national knowledge bases, as well as other features. In addition, this paper proposes an assisted data annotation solution using the Prodigy annotation tool. This approach aims to reduce the cost required to create a reference corpus for the evaluation of state-of-the-art NER models. Finally, we evaluated and compared the effectiveness of different NER methods. Results A French de-identification dataset was developed in this work, based on EHRs provided by the eHOP CDW at Rennes University Hospital, France. The dataset was rich in terms of personal information, and the distribution of entities was quite similar in the training and test datasets. We evaluated a Bi-LSTM + CRF sequence labeling architecture, combined with Flair + FastText word embeddings, on a test set of manually annotated clinical reports. The model outperformed the other tested models with a significant F1 score of 96,96%, demonstrating the effectiveness of our automatic approach for deidentifying sensitive information. Conclusions This study provides an automatic de-identification pipeline for clinical notes, which can facilitate the reuse of EHRs for secondary purposes such as clinical research. Our study highlights the importance of using advanced NLP techniques for effective de-identification, as well as the need for innovative solutions such as distant supervision to overcome the challenge of limited annotated data in the medical domain.
- Published
- 2024
- Full Text
- View/download PDF
14. A comprehensive automatic labeling and repair strategy for cracks and peeling conditions of literary murals in ancient buildings.
- Author
-
Du, Pei
- Subjects
- *
MURAL art , *SELF-organizing maps , *GENETIC algorithms , *CULTURAL property , *GENE mapping , *REPAIRING - Abstract
To protect the historical and cultural heritage, the application of self-organizing mapping networks and genetic algorithms in the restoration of ancient architectural murals is studied. The results show that the average repair time for different types of mural paintings is less than 60 seconds, and the shortest repair time is only 17.81 seconds. The evaluation effect of the research model is good, and the comprehensive efficiency evaluation of the mural restoration work is improved by about 40.42%. The repair system has excellent performance, and the algorithm has high feasibility and effectiveness. The impact of restoring murals is substantial, and the extent of restoration is highly consequential for the restoration of ancient architectural murals. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
15. RafanoSet: Dataset of raw, manually, and automatically annotated Raphanus Raphanistrum weed images for object detection and segmentation.
- Author
-
Shubham Rana, Salvatore Gerbino, Domenico Barretta, Petronia Carillo, Mariano Crimaldi, Valerio Cirillo, Albino Maggio, and Fabrizio Sarghini
- Subjects
Weed segmentation ,Automatic annotation ,Multispectral ,Segment anything model ,Grounding DINO ,Computer applications to medicine. Medical informatics ,R858-859.7 ,Science (General) ,Q1-390 - Abstract
The rationale for this data article is to provide resources which could facilitate the studies focussed over weed detection and segmentation in precision farming using computer vision. We have curated Multispectral (MS) images over crop fields of Triticum Aestivum containing heterogenous mix of Raphanus raphanistrum in both uniform and random crop spacing. This dataset is designed to facilitate weed detection and segmentation based on manual and automatically annotated Raphanus raphanistrum, commonly known as wild radish. The dataset is publicly available through the Zenodo data library and provides annotated pixel-level information that is crucial for registration and segmentation purposes. The dataset consists of 85 original MS images captured over 17 scenes covering various spectra including Blue, Green, Red, NIR (Near-Infrared), and RedEdge. Each image has a dimension of 1280 × 960 pixels and serves as the basis for the specific weed detection and segmentation. Manual annotations were performed using Visual Geometry Group Image Annotator (VIA) and the results were saved in Common Objects in Context (COCO) segmentation format. To facilitate this resource-intensive task of annotation, a Grounding DINO + Segment Anything Model (SAM) was trained with this manually annotated data to obtain automated Visual Object Classes Extended Markup Language (PASCAL VOC) annotations for 80 MS images. The dataset emphasizes quality control, validating both the 'manual'' and 'automated'' repositories by extracting and evaluating binary masks. The codes used for these processes are accessible to ensure transparency and reproducibility. This dataset is the first-of-its-kind public resource providing manual and automatically annotated weed information over close-ranged MS images in heterogenous agriculture environment. Researchers and practitioners in the fields of precision agriculture and computer vision can use this dataset to improve MS image registration and segmentation at close range photogrammetry with a focus on wild radish. The dataset not only helps with intra-subject registration to improve segmentation accuracy, but also provides valuable spectral information for training and refining machine learning models.
- Published
- 2024
- Full Text
- View/download PDF
16. GobhiSet: Dataset of raw, manually, and automatically annotated RGB images across phenology of Brassica oleracea var. Botrytis
- Author
-
Shubham Rana, Mariano Crimaldi, Domenico Barretta, Petronia Carillo, Valerio Cirillo, Albino Maggio, Fabrizio Sarghini, and Salvatore Gerbino
- Subjects
Brassica oleracea ,Manual annotation ,Automatic annotation ,Segment anything model ,Grounding DINO ,Computer applications to medicine. Medical informatics ,R858-859.7 ,Science (General) ,Q1-390 - Abstract
This research introduces an extensive dataset of unprocessed aerial RGB images and orthomosaics of Brassica oleracea crops, captured via a DJI Phantom 4. The dataset, publicly accessible, comprises 244 raw RGB images, acquired over six distinct dates in October and November of 2020 as well as 6 orthomosaics from an experimental farm located in Portici, Italy. The images, uniformly distributed across crop spaces, have undergone both manual and automatic annotations, to facilitate the detection, segmentation, and growth modelling of crops. Manual annotations were performed using bounding boxes via the Visual Geometry Group Image Annotator (VIA) and exported in the Common Objects in Context (COCO) segmentation format. The automated annotations were generated using a framework of Grounding DINO + Segment Anything Model (SAM) facilitated by YOLOv8x-seg pretrained weights obtained after training manually annotated images dated 8 October, 21 October, and 29 October 2020. The automated annotations were archived in Pascal Visual Object Classes (PASCAL VOC) format. Seven classes, designated as Row 1 through Row 7, have been identified for crop labelling. Additional attributes such as individual crop ID and the repetitiveness of individual crop specimens are delineated in the Comma Separated Values (CSV) version of the manual annotation. This dataset not only furnishes annotation information but also assists in the refinement of various machine learning models, thereby contributing significantly to the field of smart agriculture. The transparency and reproducibility of the processes are ensured by making the utilized codes accessible. This research marks a significant stride in leveraging technology for vision-based crop growth monitoring.
- Published
- 2024
- Full Text
- View/download PDF
17. A step-by-step method for cultural annotation by LLMs
- Author
-
Edgar Dubourg, Valentin Thouzeau, and Nicolas Baumard
- Subjects
automatic annotation ,human cultures ,large language models ,annotation loop ,tutorial ,Electronic computers. Computer science ,QA75.5-76.95 - Abstract
Building on the growing body of research highlighting the capabilities of Large Language Models (LLMs) like Generative Pre-trained Transformers (GPT), this paper presents a structured pipeline for the annotation of cultural (big) data through such LLMs, offering a detailed methodology for leveraging GPT’s computational abilities. Our approach provides researchers across various fields with a method for efficient and scalable analysis of cultural phenomena, showcasing the potential of LLMs in the empirical study of human cultures. LLMs proficiency in processing and interpreting complex data finds relevance in tasks such as annotating descriptions of non-industrial societies, measuring the importance of specific themes in stories, or evaluating psychological constructs in texts across societies or historical periods. These applications demonstrate the model’s versatility in serving disciplines like cultural anthropology, cultural psychology, cultural history, and cultural sciences at large.
- Published
- 2024
- Full Text
- View/download PDF
18. WeVoTe: A Weighted Voting Technique for Automatic Sentiment Annotation of Moroccan Dialect Comments
- Author
-
Yassir Matrane, Faouzia Benabbou, and Zouheir Banou
- Subjects
Sentiment analysis ,Arabic dialect ,automatic annotation ,labeling technique ,machine learning ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Sentiment analysis represents the systematic procedure of independently discerning polarity inherent in a textual document. A multitude of sectors can derive substantial advantages from this specialized domain. Conducting sentiment analysis (SA) involves various phases, with the initial step being the annotation process, which is often time-consuming and laborious. Within this framework, there exists a notable scarcity of existing research works. The complexity of this task becomes more difficult when analyzing texts in ‘Darija’, a form of the Moroccan dialect (MD). In our research endeavors, we introduced a novel automatic annotation methodology designed explicitly for sentiment analysis within the Moroccan dialect. A pivotal aspect of our contribution is the refinement of the stacking approach, utilizing a weighted voting technique for enhanced predictive accuracy. Our advanced method starts with the training of various neural network models across six unique MD datasets. The selection of these neural network architectures was underpinned by a comprehensive grid search procedure. Conclusively, it was discerned that models predicated on Recurrent Neural Networks (RNNs) outperformed others. Subsequent to this, we deployed an augmented stacking model, grounded in the aforementioned weighted voting technique. This model leverages the predictions generated by the neural networks as inputs. It then employs the mode of these inputs as an output, which feeds directly into a meta-classifier, which in turn produces the coefficients. These coefficients are then multiplicatively combined with the initial neural network predictions to derive the finale outputs. To evaluate the efficiency of our proposed methodology in annotating the six datasets, each dataset was isolated as a test while the remaining five served as training sets. Consequently, within the set of six datasets, the annotation results of three datasets have outperformed the established standards, attaining agreement rate percentages of 87.54% for MSAC, 91.25% for FB, 85.10% for MSDA, and 83.60% for MSTD, all of which represent new achievements in the literature.
- Published
- 2024
- Full Text
- View/download PDF
19. A Method for Target Detection Based on Synthetic Samples of Digital Twins
- Author
-
Zhe Dong, Yue Yang, Anqi Wang, and Tianxu Wu
- Subjects
Digital twins ,coordinates transformation ,automatic annotation ,synthetic samples ,target detection ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Target detection technology in the field of machine vision plays a vital role in industrial production and manufacturing. In industrial production, productivity can be improved by accurate target detection. To implement this technology, many enterprises must manually clean and label a huge dataset. Meanwhile, it is a huge challenge for enterprises to obtain the dataset because of enterprise data privacy and security constraints. This paper proposes a method for rapidly generating synthetic samples based on digital twins to address this challenge. First, the virtual environment is utilized to replicate the real detecting environment, generating a variety of sample photos. The three-dimensional coordinates of the target object are then extracted in the virtual scene. Subsequently, an annotation method is designed for synthetic samples obtained from the virtual scene, utilizing principles of three-dimensional coordinate transformation and perspective coordinate transformation. This method efficiently produces numerous labeled samples with diverse annotations. Ultimately, the model performs detection tasks in the actual world using the synthetic samples as training data. The experimental results show that the synthetic samples created by this method based on digital twins can substitute real samples and effectively identify target objects during actual detection tasks. This paper proposes a unique strategy for synthetic samples that reduces sample collection costs and privacy risks, thereby addressing the limitations of machine vision detection technology induced by sample limitations.
- Published
- 2024
- Full Text
- View/download PDF
20. Efficient Real-Time Droplet Tracking in Crop-Spraying Systems
- Author
-
Truong Nhut Huynh, Travis Burgers, and Kim-Doang Nguyen
- Subjects
droplet tracking ,crop-spraying systems ,deep learning ,real-time detection ,high-speed camera ,automatic annotation ,Agriculture (General) ,S1-972 - Abstract
Spray systems in agriculture serve essential roles in the precision application of pesticides, fertilizers, and water, contributing to effective pest control, nutrient management, and irrigation. These systems enhance efficiency, reduce labor, and promote environmentally friendly practices by minimizing chemical waste and runoff. The efficacy of a spray is largely determined by the characteristics of its droplets, including their size and velocity. These parameters are not only pivotal in assessing spray retention, i.e., how much of the spray adheres to crops versus becoming environmental runoff, but also in understanding spray drift dynamics. This study introduces a real-time deep learning-based approach for droplet detection and tracking which significantly improves the accuracy and efficiency of measuring these droplet properties. Our methodology leverages advanced AI techniques to overcome the limitations of previous tracking frameworks, employing three novel deep learning-based tracking methods. These methods are adept at handling challenges such as droplet occlusion and varying velocities, ensuring precise tracking in real-time potentially on mobile platforms. The use of a high-speed camera operating at 2000 frames per second coupled with innovative automatic annotation tools enables the creation of a large and accurately labeled droplet dataset for training and evaluation. The core of our framework lies in the ability to track droplets across frames, associating them temporally despite changes in appearance or occlusions. We utilize metrics including Multiple Object Tracking Accuracy (MOTA) and Multiple Object Tracking Precision (MOTP) to quantify the tracking algorithm’s performance. Our approach is set to pave the way for innovations in agricultural spraying systems, offering a more efficient, accurate, and environmentally responsible method of applying sprays and representing a significant step toward sustainable agricultural practices.
- Published
- 2024
- Full Text
- View/download PDF
21. Automatic de-identification of French electronic health records: a cost-effective approach exploiting distant supervision and deep learning models.
- Author
-
Azzouzi, Mohamed El, Coatrieux, Gouenou, Bellafqira, Reda, Delamarre, Denis, Riou, Christine, Oubenali, Naima, Cabon, Sandie, Cuggia, Marc, and Bouzillé, Guillaume
- Subjects
ELECTRONIC health records ,DEEP learning ,FRENCH language ,KNOWLEDGE base ,DATA warehousing ,PERSONALLY identifiable information - Abstract
Background: Electronic health records (EHRs) contain valuable information for clinical research; however, the sensitive nature of healthcare data presents security and confidentiality challenges. De-identification is therefore essential to protect personal data in EHRs and comply with government regulations. Named entity recognition (NER) methods have been proposed to remove personal identifiers, with deep learning-based models achieving better performance. However, manual annotation of training data is time-consuming and expensive. The aim of this study was to develop an automatic de-identification pipeline for all kinds of clinical documents based on a distant supervised method to significantly reduce the cost of manual annotations and to facilitate the transfer of the de-identification pipeline to other clinical centers. Methods: We proposed an automated annotation process for French clinical de-identification, exploiting data from the eHOP clinical data warehouse (CDW) of the CHU de Rennes and national knowledge bases, as well as other features. In addition, this paper proposes an assisted data annotation solution using the Prodigy annotation tool. This approach aims to reduce the cost required to create a reference corpus for the evaluation of state-of-the-art NER models. Finally, we evaluated and compared the effectiveness of different NER methods. Results: A French de-identification dataset was developed in this work, based on EHRs provided by the eHOP CDW at Rennes University Hospital, France. The dataset was rich in terms of personal information, and the distribution of entities was quite similar in the training and test datasets. We evaluated a Bi-LSTM + CRF sequence labeling architecture, combined with Flair + FastText word embeddings, on a test set of manually annotated clinical reports. The model outperformed the other tested models with a significant F1 score of 96,96%, demonstrating the effectiveness of our automatic approach for deidentifying sensitive information. Conclusions: This study provides an automatic de-identification pipeline for clinical notes, which can facilitate the reuse of EHRs for secondary purposes such as clinical research. Our study highlights the importance of using advanced NLP techniques for effective de-identification, as well as the need for innovative solutions such as distant supervision to overcome the challenge of limited annotated data in the medical domain. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
22. INTELLIGENT VEHICLE INSPECTION TOOL DESIGN BASED ON FREEMAN CHAIN CODE FOR AUTOMATIC ANNOTATION OF 3D MODELS.
- Author
-
SHAN CHENG and GUOQIANG TAN
- Subjects
BACK propagation ,AUTOMOTIVE engineering ,ANNOTATIONS ,CLASSIFICATION ,AUTONOMOUS vehicles - Abstract
Autonomous vehicle are more and more widely used in daily life, and the requirements for their safety performance are higher and higher. As a tool for testing auto parts, intelligent inspection tools are crucial to the guarantee of automobile quality. However, traditional fixture design relies on manual drawing, which is inefficient and prone to errors. To solve this problem, this research uses Freeman chain code to determine the annotation object, uses case clustering method to annotate, and uses error back propagation algorithm to realize case knowledge classification learning, and designs intelligent vehicle inspection tool design technology based on Freeman chain code 3D automatic annotation method. The experimental results show that the geometric feature matching results are correct, and the difference in feature comparison results is significant, with a high accuracy rate. Meanwhile, the geometric similarity annotation method has a high accuracy rate, taking only 3 minutes to complete the annotation, which is 7 minutes longer than traditional manual annotation. The error backpropagation algorithm can accurately achieve feature classification, and the design time of size chain inspection tool deformation design is reduced by 214min compared to manual reverse deformation design, significantly improving design efficiency. In summary, the proposed design method for automotive inspection tools can achieve automatic model annotation, improve design efficiency, and reduce design time. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
23. Evaluating the Use of Generative LLMs for Intralingual Diachronic Translation of Middle-Polish Texts into Contemporary Polish
- Author
-
Klamra, Cezary, Kryńska, Katarzyna, Ogrodniczuk, Maciej, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Goh, Dion H., editor, Chen, Shu-Jiun, editor, and Tuarob, Suppawong, editor
- Published
- 2023
- Full Text
- View/download PDF
24. CLARIN-Emo: Training Emotion Recognition Models Using Human Annotation and ChatGPT
- Author
-
Koptyra, Bartłomiej, Ngo, Anh, Radliński, Łukasz, Kocoń, Jan, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Mikyška, Jiří, editor, de Mulatier, Clélia, editor, Paszynski, Maciej, editor, Krzhizhanovskaya, Valeria V., editor, Dongarra, Jack J., editor, and Sloot, Peter M.A., editor
- Published
- 2023
- Full Text
- View/download PDF
25. Facilitation of Language Acquisition by Non-Native Speakers through the Construction of a Japanese Cultural and Educational Corpus
- Author
-
Chen Yamin
- Subjects
corpus ,machine learning ,automatic annotation ,language acquisition. ,97p10 ,Mathematics ,QA1-939 - Abstract
With the diversification of the world landscape and the accelerating process of higher education, Japanese is being valued as an important language, yet there are differences in the formal expression of the language due to different cultural backgrounds. The creation of a corpus of Japanese language culture and education is explored in this article, which contributes to language acquisition by non-native speakers. After collecting relevant information about Japanese language culture and education, the article takes it as a corpus and develops a corpus annotation system and specification further modifies the model and algorithm through machine learning, and finally realizes the construction of the system. By testing the usefulness and effectiveness of the Japanese language culture education corpus proposed in this paper, the paper concludes that in the analysis of the number and rate of bias of “了 (have done)” in subjects with different levels of Japanese language proficiency, the average rate of bias of the beginner level group is the highest at 24.65%, and the average rate of bias of the intermediate level group is the lowest at 14.6%. The “over-marking” bias rate of students’ use of “着(be doing)” reveals that the bias rate of students decreases by 0.64 from the beginner stage to the advanced stage, which indicates that the Japanese cultural education corpus proposed in this paper has a good effect on improving and facilitating the language acquisition of non-native speakers. It shows that the Japanese cultural education corpus proposed in this paper has a positive effect on improving and promoting language acquisition for non-native speakers.
- Published
- 2024
- Full Text
- View/download PDF
26. Research on Automatic Annotation Method of Korean Language under Data Driving and Fusion
- Author
-
Xiang Tianyu, Gao Li, and Liu Wenming
- Subjects
seq2seq model ,bidirectional gru model ,hidden markov model ,semi-supervised learning ,seneff model ,automatic annotation ,62-07 ,Mathematics ,QA1-939 - Abstract
In the quest to streamline Korean text and speech annotation, this research introduces innovative automatic annotation methods that promise to revolutionize efficiency and technical prowess in constructing Korean annotation datasets. By leveraging the sophisticated Seq2Seq architecture with BERT and bidirectional GRU models, we significantly enhance the model’s ability to grasp contextual nuances, ensuring precise text annotations. The speech annotation frontier benefits from a novel amalgamation of the Hidden Markov Model’s forced alignment and semi-supervised learning, perfected with Seneff auditory features for meticulous phonological consonant boundary detection. Empirical validation across diverse datasets showcases our methodology’s superiority, achieving a remarkable 96.01% accuracy in text annotation and setting a new benchmark for phonological boundary detection at a 14.5ms minimum distance threshold. Our approach outperforms traditional algorithms, marking a pivotal step forward in Korean automatic annotation.
- Published
- 2024
- Full Text
- View/download PDF
27. Bird's Eye View Semantic Segmentation based on Improved Transformer for Automatic Annotation.
- Author
-
Tianjiao Liang, Weiguo Pan, Hong Bao, Xinyue Fan, and Han Li
- Subjects
ANNOTATIONS ,FEATURE extraction ,POINT cloud ,AUTONOMOUS vehicles ,AUTOMOBILE license plates - Abstract
High-definition (HD) maps can provide precise road information that enables an autonomous driving system to effectively navigate a vehicle. Recent research has focused on leveraging semantic segmentation to achieve automatic annotation of HD maps. However, the existing methods suffer from low recognition accuracy in automatic driving scenarios, leading to inefficient annotation processes. In this paper, we propose a novel semantic segmentation method for automatic HD map annotation. Our approach introduces a new encoder, known as the convolutional transformer hybrid encoder, to enhance the model's feature extraction capabilities. Additionally, we propose a multi-level fusion module that enables the model to aggregate different levels of detail and semantic information. Furthermore, we present a novel decoupled boundary joint decoder to improve the model's ability to handle the boundary between categories. To evaluate our method, we conducted experiments using the Bird's Eye View point cloud images dataset and Cityscapes dataset. Comparative analysis against stateof- the-art methods demonstrates that our model achieves the highest performance. Specifically, our model achieves an mIoU of 56.26%, surpassing the results of SegFormer with an mIoU of 1.47%. This innovative promises to significantly enhance the efficiency of HD map automatic annotation. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
28. A Multilabel Learning-Based Automatic Annotation Method for Semantic Roles in English Text
- Author
-
Li Lei and Hao Wang
- Subjects
Multi-label learning ,semantic comprehension ,automatic annotation ,deep neural networks ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
With the increasing amount of textual information in the Internet, smart semantic comprehension is a practical demand. Among, automatic annotation for semantic roles remains the fundamental part for effective semantic comprehension. Although machine learning-based methods had received much attention in recent years, they mostly divided each sentences into separable parts for calculation. To deal with such challenge, this paper introduces multilabel learning to propose a novel automatic annotation method for semantic roles in English text. In the semantic representation of words, the method uses convolutional neural networks to extract local feature information of words from the character level. Such design can alleviate the problem of inconspicuous semantic features caused by random initialization of unregistered words. Secondly, in the process of implication recognition, by combining the interactive attention mechanism to construct a capsule for each implication relation separately, the recognition of the final implication relation is completed in the way of categorical learning. At last, some experiments are conducted on real-world data to verify the proposed method with being compared with several typical relevant methods. The obtained results show that the proposal achieves better Macro-F1 results on eight datasets compared to seven algorithms. Besides, the proposal also performs better than others in the sensitivity testing, as its performance can remain stable with the increase of noise input. In summary, the proposal can achieve good results and show strong capability in semantic role labeling tasks.
- Published
- 2023
- Full Text
- View/download PDF
29. Explicit Aspect Annotation via Transfer and Active Learning.
- Author
-
Maroua, Boudabous and Anna, Pappa
- Subjects
ACTIVE learning ,TRANSFER of training ,ANNOTATIONS ,LEARNING strategies ,ELECTRONIC equipment ,DEEP learning - Abstract
We present a semi-supervised annotation process for identifying and labelling explicit aspects of an initially unlabelled corpus. Firstly, we employ cross-domain learning to pre-annotate the initial data, deliberately excluding domain-related input features to ensure effective learning transfer. Then, we apply an active learning strategy to enhance the pre-annotation performance and enrich the learning data. We adjust the strategy to sequence labeling and address class imbalance. We evaluate this process using two unlabelled datasets in French, consisting of user opinions on beauty products and electronic devices, respectively. The results show an improved F1-score achieved by increasing and correcting 30% of the training dataset. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
30. Early Experiments on Automatic Annotation of Portuguese Medieval Texts
- Author
-
Bico, Maria Inës, Baptista, Jorge, Batista, Fernando, Cardeira, Esperança, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Silvello, Gianmaria, editor, Corcho, Oscar, editor, Manghi, Paolo, editor, Di Nunzio, Giorgio Maria, editor, Golub, Koraljka, editor, Ferro, Nicola, editor, and Poggi, Antonella, editor
- Published
- 2022
- Full Text
- View/download PDF
31. Named Entity Extractors for New Domains by Transfer Learning with Automatically Annotated Data
- Author
-
Matos, Emanuel, Rodrigues, Mário, Teixeira, António, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Woeginger, Gerhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Pinheiro, Vládia, editor, Gamallo, Pablo, editor, Amaro, Raquel, editor, Scarton, Carolina, editor, Batista, Fernando, editor, Silva, Diego, editor, Magro, Catarina, editor, and Pinto, Hugo, editor
- Published
- 2022
- Full Text
- View/download PDF
32. Automatic annotation of error types for grammatical error correction
- Author
-
Bryant, Christopher Jack and Briscoe, Edward
- Subjects
Natural Language Processing ,Grammatical Error Correction ,Automatic Annotation - Abstract
Grammatical Error Correction (GEC) is the task of automatically detecting and correcting grammatical errors in text. Although previous work has focused on developing systems that target specific error types, the current state of the art uses machine translation to correct all error types simultaneously. A significant disadvantage of this approach is that machine translation does not produce annotated output and so error type information is lost. This means we can only evaluate a system in terms of overall performance and cannot carry out a more detailed analysis of different aspects of system performance. In this thesis, I develop a system to automatically annotate parallel original and corrected sentence pairs with explicit edits and error types. In particular, I first extend the Damerau- Levenshtein alignment algorithm to make use of linguistic information when aligning parallel sentences, and supplement this alignment with a set of merging rules to handle multi-token edits. The output from this algorithm surpasses other edit extraction approaches in terms of approximating human edit annotations and is the current state of the art. Having extracted the edits, I next classify them according to a new rule-based error type framework that depends only on automatically obtained linguistic properties of the data, such as part-of-speech tags. This framework was inspired by existing frameworks, and human judges rated the appropriateness of the predicted error types as 'Good' (85%) or 'Acceptable' (10%) in a random sample of 200 edits. The whole system is called the ERRor ANnotation Toolkit (ERRANT) and is the first toolkit capable of automatically annotating parallel sentences with error types. I demonstrate the value of ERRANT by applying it to the system output produced by the participants of the CoNLL-2014 shared task, and carry out a detailed error type analysis of system performance for the first time. I also develop a simple language model based approach to GEC, that does not require annotated training data, and show how it can be improved using ERRANT error types.
- Published
- 2019
- Full Text
- View/download PDF
33. Hodnoty slovesných morfologických kategorií v korpusu SYN2020 — atribut verbtag
- Author
-
Tomáš Jelínek, Vladimír Petkevič, and Hana Skoumalová
- Subjects
verbtag attribute ,morphology of czech verbs ,morphological categories and values ,automatic annotation ,syn2020 corpus ,Philology. Linguistics ,P1-1091 - Abstract
The paper describes the verbtag attribute, which allows a user to search, in the SYN2020 corpus (and also subsequent corpora, SYNv9 and SYNv10) of contemporary Czech, for all values of morphological categories of verbs, i.e., not only those contained in the tag attribute, but also those related mainly to multi-word participial verb predicates, which are prevalent in Czech. The verbtag attribute contains information indicating whether the verb (co-)forming the verbal meaning is either auxiliary or autosemantic, as well as information about the verb mode, diathesis, person, number and tense. The annotation applies both to verb predicates expressed in a single word (e.g., the 1st person indicative present tense: Čtu rád detektivní příběhy. ‘I like to read detective stories.’) and (especially) to verb predicates expressed in multiple words (e.g., the present conditional of the 1st person singular: Pak bych mu s chutí nabídla výhodnou smlouvu. ‘Then I would gladly offer him a good deal.’). The authors introduce the motivation and the concept of the verbtag annotation, describe relevant morphological categories and their values in detail, and show, via examples, how various multiword structures expressing verbal meaning are annotated in the verbtag attribute. They also offer users a guide to the whole issue of verbal morphosyntax manifested in the verbtag attribute and possibilities for efficient search for and retrieval of morphological/morphosyntactic data. The paper shows which multiple verb complexes are simple in terms of annotation, but also identifies more complex cases (e.g., coordination of participles) which are not easy to automatically annotate, and/or whose annotation is unclear in terms of an adequate theoretical approach. The authors also present the method used for annotating multiword verbal complexes and its current success rate.
- Published
- 2022
- Full Text
- View/download PDF
34. An event-based automatic annotation method for datasets of interpersonal relation extraction.
- Author
-
Li, Fangfang, Chen, Guikai, and Liu, Xiyao
- Subjects
INTERPERSONAL relations ,ANNOTATIONS ,GENERALIZATION - Abstract
Distant supervision (DS) has been widely used in automatic annotation of interpersonal relation datasets recently. However, DS suffers from high cost of constructing knowledge sets, low generalization ability and cannot effectively handle missing or adaptive relationships. To address these issues, we propose a novel event-based automatic annotation method for interpersonal datasets, namely, event distant supervision (EDS). In this method, We design an event-based set to replace the knowledge set, which reduces the cost for reconstructing the knowledge set and offers considerably higher generalization ability. Moreover, we design an event alignment label-annotation mechanism and a scoring mechanism to reduce the possibility of inaccurate and incomplete annotation. Experiments demonstrate that the annotation performance of EDS is significantly superior to that of DS in terms of cost, generalization ability, and effectiveness, particularly for annotating entity pairs with adaptive relationships. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
35. Combining semantic and linguistic representations for media recommendation.
- Author
-
Harrando, Ismail and Troncy, Raphael
- Subjects
- *
RECOMMENDER systems , *KNOWLEDGE graphs , *DATA mining , *PODCASTING , *ACQUISITION of data , *METADATA - Abstract
Content-based recommendation systems offer the possibility of promoting media (e.g., posts, videos, podcasts) to users based solely on a representation of the content (i.e., without using any user-related data such as views or interactions between users and items). In this work, we study the potential of using different textual representations (based on the content of the media) and semantic representations (created from a knowledge graph of media metadata). We also show that using off-the-shelf automatic annotation tools from the Information Extraction literature, we can improve recommendation performance, without any extra cost of training, data collection or annotation. We first evaluate multiple textual content representations on two tasks of recommendation: user-specific, which is performed by suggesting new items to the user given a history of interactions, and item-based, which is based solely on content relatedness, and is rarely investigated in the literature of recommender systems. We compare how using automatically extracted content (via ASR) compares to using human-written summaries. We then derive a semantic content representation by combining manually created metadata and automatically extracted annotations and we show that Knowledge Graphs, through their embeddings, constitute a great modality to seamlessly integrate extracted knowledge to legacy metadata and can be used to provide good content recommendations. We finally study how combining both semantic and textual representations can lead to superior performance on both recommendation tasks. Our code is available at https://github.com/D2KLab/ka-recsys to support experiment reproducibility. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
36. NaNG-ST: A natural neighborhood graph-based self-training method for semi-supervised classification.
- Author
-
Li, Junnan
- Subjects
- *
DATA distribution , *CLASSIFICATION - Abstract
The self-training method has been favored by scholars in semi-supervised classification. One of the greatest challenges in self-training methods is finding high-confidence unlabeled samples at each iteration. While multiple variations of self-training methods are developed, the strategy of finding high-confidence unlabeled samples heavlies on parameters and only utilizes labeled data alone, which makes the self-training method limited by the distribution of labeled data. To solve the above issues, a novel natural neighborhood graph-based self-training method (NaNG-ST) is proposed. In NaNG-ST, a parameter-free natural neighborhood graph (NaNG) is first constructed. The NaNG roughly reveals the real data distribution by exploiting unlabeled and labeled data. Based on NaNG, homogeneous and heterogeneous edges are defined to divide unlabeled samples into three cases. After that, homogeneous and heterogeneous edges can use the revealed distribution rather than labeled data alone to help NaNG-ST fast and effectively find confident unlabeled samples without any parameters. Besides, they also help NaNG-ST not to be limited by the distribution of initial labeled data. When a few initial labeled data can not roughly represent the distribution of the entire data, the NaNG helps NaNG-ST restore the real data distribution better. Intensive experiments on real-world data sets prove that NaNG-ST outperforms 7 popular semi-supervised self-taught approaches in terms of classification accuracy, mean F -measure and required running time. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
37. FormulaNet: A Benchmark Dataset for Mathematical Formula Detection
- Author
-
Felix M. Schmitt-Koopmann, Elaine M. Huang, Hans-Peter Hutter, Thilo Stadelmann, and Alireza Darvishy
- Subjects
Automatic annotation ,dataset ,document analysis ,deep learning ,mathematical formula detection ,page object detection ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
One unsolved sub-task of document analysis is mathematical formula detection (MFD). Research by ourselves and others has shown that existing MFD datasets with inline and display formula labels are small and have insufficient labeling quality. There is therefore an urgent need for datasets with better quality labeling for future research in the MFD field, as they have a high impact on the performance of the models trained on them. We present an advanced labeling pipeline and a new dataset called FormulaNet in this paper. At over 45k pages, we believe that FormulaNet is the largest MFD dataset with inline formula labels. Our experiments demonstrate substantially improved labeling quality for inline and display formulae detection over existing datasets. Additionally, we provide a math formula detection baseline for FormulaNet with an mAP of 0.754. Our dataset is intended to help address the MFD task and may enable the development of new applications, such as making mathematical formulae accessible in PDFs for visually impaired screen reader users.
- Published
- 2022
- Full Text
- View/download PDF
38. Video object segmentation for automatic image annotation of ethernet connectors with environment mapping and 3D projection.
- Author
-
Danta, Marrone, Dreyer, Pedro, Bezerra, Daniel, Reis, Gabriel, Souza, Ricardo, Lins, Silvia, Kelner, Judith, and Sadok, Djamel
- Subjects
GRAPHICAL projection ,ARTIFICIAL neural networks ,CONVOLUTIONAL neural networks ,MAP projection ,IMAGE segmentation ,OBJECT tracking (Computer vision) - Abstract
The creation of a dataset is time-consuming and sometimes discourages researchers from pursuing their goals. To overcome this problem, we present and discuss two solutions adopted for the automation of this process. Both optimize valuable user time and resources and use video object segmentation with object tracking and 3D projection. In our scenario, we acquire images from a moving robotic arm and, for each approach, generate distinct annotated datasets. We evaluated the precision of the annotations by comparing these with a manually annotated dataset. As a complementary test to assess the quality of the generated datasets and to achieve a generalization of our contribution, we tested detection and classification problems. In both tests, we rely on solutions with Convolution Neural Network and Deep Learning. For detection support, we used YOLO and obtained for the projection dataset an F1-Score, accuracy, and mAP values of 0.846, 0.924, and 0.875, respectively. Concerning the tracking dataset, we achieved an F1-Score of 0.861, an accuracy of 0.932, whereas mAP reached 0.894. For the classification, we adopted the two metrics accuracy and F1-Score, and used the known networks VGG, DenseNet, MobileNet, Inception, and ResNet. The VGG architecture outperformed the others for both projection and tracking datasets. It reached an accuracy and F1-score of 0.997 and 0.993, respectively. Similarly, for the tracking dataset, it achieved an accuracy of 0.991 and an F1-Score of 0.981. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
39. Automatic Annotation Performance of TextBlob and VADER on Covid Vaccination Dataset.
- Author
-
Alenzi, Badriya Murdhi, Khan, Muhammad Badruddin, Hasanat, Mozaherul Hoque Abul, Saudagar, Abdul Khader Jilani, AlKhathami, Mohammed, and AlTameem, Abdullah
- Subjects
COVID-19 vaccines ,ANNOTATIONS ,SENTIMENT analysis ,TASK analysis - Abstract
With the recent boom in the corpus size of sentiment analysis tasks, automatic annotation is poised to be a necessary alternative to manual annotation for generating ground truth dataset labels. This article aims to investigate and validate the performance of two widely used lexicon-based automatic annotation approaches, TextBlob and Valence Aware Dictionary and Sentiment Reasoner (VADER), by comparing them with manual annotation. The dataset of 5402 Arabic tweets was annotated manually, containing 3124 positive tweets, 1463 negative tweets, and 815 neutral tweets. The tweets were translated into English so that TextBlob and VADER could be used for their annotation. TextBlob and VADER automatically classified the tweets to positive, negative, and neutral sentiments and compared them with manual annotation. This study shows that automatic annotation cannot be trusted as the gold standard for annotation. In addition, the study discussed many drawbacks and limitations of automatic annotation using lexicon-based algorithms. The highest level of accuracies of 75% and 70% were achieved by TextBlob and VADER, respectively. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
40. Investigation of the reliability of semi-automatic annotation by the Geri time-lapse system.
- Author
-
Vandame, Jessica, Fossard, Camille, Filali, Meryem, Benammar, Achraf, Ranga, Stéphanie, Pirtea, Paul, Racowsky, Catherine, Ayoubi, Jean-Marc, and Poulain, Marine
- Subjects
- *
HUMAN embryos , *ANNOTATIONS , *HUMAN phenotype , *OVUM - Abstract
What is the reliability of Geri® Assess 2.0 software time-lapse technology for annotating kinetic events and identifying abnormal phenotypes in preimplantation human embryos? Embryos were annotated using Assess 2.0 for the appearance and fading of pronuclei, and for progression to the 2-, 3-, 4-, 5- and 6-cell stages and to three blastocyst stages. Identification of reverse cleavage and direct cleavage phenotypes was also recorded. Manual annotation was undertaken after these events in a blinded fashion. Embryo scores were compared between Assess 2.0 and manual annotation. A total of 513 oocytes from 34 women were included. Detection rates for Assess 2.0 versus manual annotation among the 10 kinetic events and including direct cleavage and reverse cleavage ranged between 0% and 94.4%. The percentage of discordant pairs was significantly different for all 12 events analysed (P -value range 0.036 to <0.0001). The sensitivity of Assess 2.0 ranged from 68.2% to 94.4% and specificity ranged from 63.8% to 97.3%. Assess 2.0 called for verification by the embryologist for at least one event in 55.2% of oocytes assessed. Of the 297 embryos scored by manual annotation, Assess 2.0 assigned the same score for only 125 (42.1%), although after manual corrections, concordance with manual annotation scores was raised to 66.0%. The results reveal striking differences between Assess 2.0 and manual annotation for kinetic annotations. Failure of Assess 2.0 to detect direct cleavage events and the low detection rate of reverse cleavage are further limitations. These collective findings highlight the importance of validating time-lapse annotation software before clinical implementation. Manual verification of Assess 2.0 outputs remains essential for accurate data interpretation. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
41. Event-Cap – Event Ranking and Transformer-based Video Captioning
- Author
-
Cederqvist, Gabriel, Gustafsson, Henrik, Cederqvist, Gabriel, and Gustafsson, Henrik
- Abstract
In the field of video surveillance, vast amounts of data are gathered each day. To be able to identify what occurred during a recorded session, a human annotator has to go through the footage and annotate the different events. This is a tedious and expensive process that takes up a large amount of time. With the rise of machine learning and in particular deep learning, the field of both image and video captioning has seen large improvements. Contrastive Language-Image Pretraining is capable of efficiently learning a multimodal space, thus able to merge the understanding of text and images. This enables visual features to be extracted and processed into text describing the visual content. This thesis presents a system for extracting and ranking important events from surveillance videos as well as a way of automatically generating a description of the event. By utilizing the pre-trained models X-CLIP and GPT-2 to extract visual information from the videos and process it into text, a video captioning model was created that requires very little training. Additionally, the ranking system was implemented to extract important parts in video, utilizing anomaly detection as well as polynomial regression. Captions were evaluated using the metrics BLEU, METEOR, ROUGE and CIDEr, and the model receives scores comparable to other video captioning models. Additionally, captions were evaluated by experts in the field of video surveillance, who rated them on accuracy, reaching up to 62.9%, and semantic quality, reaching 99.2%. Furthermore the ranking system was also evaluated by the experts, where they agree with the ranking system 78% of the time., Inom videoövervakning samlas stora mängder data in varje dag. För att kunna identifiera vad som händer i en inspelad övervakningsvideo så måste en människa gå igenom och annotera de olika händelserna. Detta är en långsam och dyr process som tar upp mycket tid. Under de senaste åren har det setts en enorm ökning av användandet av olika maskininlärningsmodeller. Djupinlärningsmodeller har fått stor framgång när det kommer till att generera korrekt och trovärdig text. De har också använts för att generera beskrivningar för både bilder och video. Contrastive Language-Image Pre-training har gjort det möjligt att träna en multimodal rymd som kombinerar förståelsen av text och bild. Detta gör det möjligt att extrahera visuell information och skapa textbeskrivningar. Denna master uppsatts beskriver ett system som kan extrahera och ranka viktiga händelser i en övervakningsvideo samt ett automatiskt sätt att generera beskrivningar till dessa. Genom att använda de förtränade modellerna X-CLIP och GPT-2 för att extrahera visuell information och textgenerering, har en videobeskrivningsmodell skapats som endast behöver en liten mängd träning. Dessutom har ett rankingsystem implementerats för att extrahera de viktiga delarna i en video genom att använda anomalidetektion och polynomregression. Video beskrivningarna utvärderades med måtten BLEU, METOER, ROUGE och CIDEr, där modellerna får resultat i klass med andra videobeskrivningsmodeller. Fortsättningsvis utvärderades beskrivningarna också av experter inom videoövervakningsområdet där de fick besvara hur bra beskrivningarna var i måtten: beskrivningsprecision som uppnådde 62.9% och semantisk kvalité som uppnådde 99.2%. Ranknignssystemet utvärderades också av experterna. Deras åsikter överensstämde till 78% med rankningssystemet.
- Published
- 2024
42. Machine Learning for Automatic Annotation and Recognition of Demographic Characteristics in Facial Images
- Author
-
Gustavsson Roth, Ludvig, Rimér Högberg, Camilla, Gustavsson Roth, Ludvig, and Rimér Högberg, Camilla
- Abstract
Recent increase in widespread use of facial recognition technologies have accelerated the utilization of demographic information, as extracted from facial features, yet it is accompanied by ethical concerns. It is therefore crucial, for ethical reasons, to ensure that algorithms like face recognition algorithms employed in legal proceedings are equitable and thoroughly documented across diverse populations. Accurate classification of demographic traits are therefore essential for enabling a comprehensive understanding of other algorithms. This thesis explores how classical machine learning algorithms compare to deep-learning models in predicting sex, age and skin color, concluding that the more compute-heavy deep-learning models, where the best performing models achieved an MCC of 0.99, 0.48 and 0.85 for sex, age and skin color respectively, significantly outperform their classical machine learning counterparts which achieved an MCC of 0.57, 0.22 and 0.54 at best. Once establishing that the deep-learning models are superior, further methods such as semi-supervised learning, a multi-characteristic classifier, sex-specific age classifiers and using tightly cropped facial images instead of upper-body images were employed to try and improve the deep-learning results. Throughout all deep-learning experiments the state of the art vision transformer and convolutional neural network were compared. Whilst the different architectures performed remarkably alike, a slight edge was seen for the convolutional neural network. The results further show that using cropped facial images generally improve the model performance and that more specialized models achieve modest improvements as compared to their less specialized counterparts. Semi-supervised learning showed potential in slightly improving the models further. The predictive performances achieved in this thesis indicate that the deep-learning models can reliably predict demographic features close to, or surpassing, a human.
- Published
- 2024
43. Using Satellite Images and Deep Learning to Detect Water Hidden Under the Vegetation : A cross-modal knowledge distillation-based method to reduce manual annotation work
- Author
-
Cristofoli, Ezio and Cristofoli, Ezio
- Abstract
Detecting water under vegetation is critical to tracking the status of geological ecosystems like wetlands. Researchers use different methods to estimate water presence, avoiding costly on-site measurements. Optical satellite imagery allows the automatic delineation of water using the concept of the Normalised Difference Water Index (NDWI). Still, optical imagery is subject to visibility conditions and cannot detect water under the vegetation, a typical situation for wetlands. Synthetic Aperture Radar (SAR) imagery works under all visibility conditions. It can detect water under vegetation but requires deep network algorithms to segment water presence, and manual annotation work is required to train the deep models. This project uses DEEPAQUA, a cross-modal knowledge distillation method, to eliminate the manual annotation needed to extract water presence from SAR imagery with deep neural networks. In this method, a deep student model (e.g., UNET) is trained to segment water in SAR imagery. The student model uses the NDWI algorithm as the non-parametric, cross-modal teacher. The key prerequisite is that NDWI works on the optical imagery taken from the exact location and simultaneously as the SAR. Three different deep architectures are tested in this project: UNET, SegNet, and UNET++, and the Otsu method is used as the baseline. Experiments on imagery from Swedish wetlands in 2020-2022 show that cross-modal distillation consistently achieved better segmentation performances across architectures than the baseline. Additionally, the UNET family of algorithms performed better than SegNet with a confidence of 95%. The UNET++ model achieved the highest Intersection Over Union (IOU) performance. However, no statistical evidence emerged that UNET++ performs better than UNET, with a confidence of 95%. In conclusion, this project shows that cross-modal knowledge distillation works well across architectures and removes tedious and expensive manual work hours when detecting wate, Att upptäcka vatten under vegetation är avgörande för att hålla koll på statusen på geologiska ekosystem som våtmarker. Forskare använder olika metoder för att uppskatta vattennärvaro vilket undviker kostsamma mätningar på plats. Optiska satellitbilder tillåter automatisk avgränsning av vatten med hjälp av konceptet Normalised Difference Water Index (NDWI). Optiska bilder fortfarande beroende av siktförhållanden och kan inte upptäcka vatten under vegetationen, en typisk situation för våtmarker. Synthetic Aperture Radar (SAR)-bilder fungerar under alla siktförhållanden. Den kan detektera vatten under vegetation men kräver djupa nätverksalgoritmer för att segmentera vattennärvaro, och manuellt anteckningsarbete krävs för att träna de djupa modellerna. Detta projekt använder DEEPAQUA, en cross-modal kunskapsdestillationsmetod, för att eliminera det manuella annoteringsarbete som behövs för att extrahera vattennärvaro från SAR-bilder med djupa neurala nätverk. I denna metod tränas en djup studentmodell (t.ex. UNET) att segmentera vatten i SAR-bilder semantiskt. Elevmodellen använder NDWI, som fungerar på de optiska bilderna tagna från den exakta platsen och samtidigt som SAR, som den icke-parametriska, cross-modal lärarmodellen. Tre olika djupa arkitekturer testas i detta examensarbete: UNET, SegNet och UNET++, och Otsu-metoden används som baslinje. Experiment på bilder tagna på svenska våtmarker 2020-2022 visar att cross-modal destillation konsekvent uppnådde bättre segmenteringsprestanda över olika arkitekturer jämfört med baslinjen. Dessutom presterade UNET-familjen av algoritmer bättre än SegNet med en konfidens på 95%. UNET++-modellen uppnådde högsta prestanda för Intersection Over Union (IOU). Det framkom dock inga statistiska bevis för att UNET++ presterar bättre än UNET, med en konfidens på 95%. Sammanfattningsvis visar detta projekt att cross-modal kunskapsdestillation fungerar bra över olika arkitekturer och tar bort tidskrävande och kostsamma manuella arbetst
- Published
- 2024
44. Automated methods for cell type annotation on scRNA-seq data
- Author
-
Giovanni Pasquini, Jesus Eduardo Rojo Arias, Patrick Schäfer, and Volker Busskamp
- Subjects
scRNA-seq ,Cell type ,Cell state ,Automatic annotation ,Biotechnology ,TP248.13-248.65 - Abstract
The advent of single-cell sequencing started a new era of transcriptomic and genomic research, advancing our knowledge of the cellular heterogeneity and dynamics. Cell type annotation is a crucial step in analyzing single-cell RNA sequencing data, yet manual annotation is time-consuming and partially subjective. As an alternative, tools have been developed for automatic cell type identification. Different strategies have emerged to ultimately associate gene expression profiles of single cells with a cell type either by using curated marker gene databases, correlating reference expression data, or transferring labels by supervised classification. In this review, we present an overview of the available tools and the underlying approaches to perform automated cell type annotations on scRNA-seq data.
- Published
- 2021
- Full Text
- View/download PDF
45. Automatic tool to annotate smile intensities in conversational face-to-face interactions.
- Author
-
Rauzy, Stéphane and Amoyal, Mary
- Subjects
FACIAL expression ,SMILING ,VIDEO recording ,LAUGHTER ,STATISTICAL models - Abstract
This study presents an automatic tool that allows to trace smile intensities along a video record of conversational face-to-face interactions. The processed output proposes a sequence of adjusted time intervals labeled following the Smiling Intensity Scale (Gironzetti, Attardo, and Pickering, 2016), a 5 levels scale varying from neutral facial expression to laughing smile. The underlying statistical model of this tool is trained on a manually annotated corpus of conversations featuring spontaneous facial expressions. This model will be detailed in this study. This tool can be used with benefits for annotating smile in interactions. The results are twofold. First, the evaluation reveals an observed agreement of 68% between manual and automatic annotations. Second, manually correcting the labels and interval boundaries of the automatic outputs reduces by a factor 10 the annotation time as compared with the time spent for manually annotating smile intensities without pretreatment. Our annotation engine makes use of the state-of-the-art toolbox OpenFace for tracking the face and for measuring the intensities of the facial Action Units of interest all along the video. The documentation and the scripts of our tool, the SMAD software, are available to download at the HMAD open source project URL page https://github.com/srauzy/HMAD (last access 31 July 2023). [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
46. Beat-Level Interpretation of Intra-Patient Paradigm Based on Object Detection
- Author
-
Man Kang, Xue-Feng Wang, Jing Xiao, He Tian, and Tian-Ling Ren
- Subjects
object detection ,ECG ,beat-level classification ,deep learning ,automatic annotation ,Diseases of the circulatory (Cardiovascular) system ,RC666-701 - Abstract
Electrocardiogram (ECG), as a product that can most directly reflect the electrical activity of the heart, has become the most common clinical technique used for the analysis of cardiac abnormalities. However, it is a heavy and tedious burden for doctors to analyze a large amount of ECG data from the long-term monitoring system. The realization of automatic ECG analysis is of great significance. This work proposes a beat-level interpretation method based on the automatic annotation algorithm and object detector, which abandons the previous mode of separate R peak detection and heartbeat classification. The ground truth of the QRS complex is automatically annotated and also regarded as the object the model can learn like category information. The object detector unifies the localization and classification tasks, achieving an end-to-end optimization as well as decoupling the high dependence on the R peak. Compared with most advanced methods, this work shows superior performance. For the interpretation of 12 heartbeat types in the MIT-BIH dataset, the average accuracy is 99.60%, the average sensitivity is 97.56%, and the average specificity is 99.78%. This method can be used as a clinical auxiliary tool to help doctors diagnose arrhythmia after receiving large-scale database training.
- Published
- 2022
- Full Text
- View/download PDF
47. Automatic Summarization Generation Technology of Network Document Based on Knowledge Graph
- Author
-
Wu, Yuezhong, Chen, Rongrong, Li, Changyun, Chen, Shuhong, Zou, Wenjun, Akan, Ozgur, Editorial Board Member, Bellavista, Paolo, Editorial Board Member, Cao, Jiannong, Editorial Board Member, Coulson, Geoffrey, Editorial Board Member, Dressler, Falko, Editorial Board Member, Ferrari, Domenico, Editorial Board Member, Gerla, Mario, Editorial Board Member, Kobayashi, Hisashi, Editorial Board Member, Palazzo, Sergio, Editorial Board Member, Sahni, Sartaj, Editorial Board Member, Shen, Xuemin (Sherman), Editorial Board Member, Stan, Mircea, Editorial Board Member, Xiaohua, Jia, Editorial Board Member, Zomaya, Albert Y., Editorial Board Member, Liu, Shuai, editor, and Yang, Gelan, editor
- Published
- 2019
- Full Text
- View/download PDF
48. MAM: Transfer Learning for Fully Automatic Video Annotation and Specialized Detector Creation
- Author
-
Fuhl, Wolfgang, Castner, Nora, Zhuang, Lin, Holzer, Markus, Rosenstiel, Wolfgang, Kasneci, Enkelejda, Hutchison, David, Series Editor, Kanade, Takeo, Series Editor, Kittler, Josef, Series Editor, Kleinberg, Jon M., Series Editor, Mattern, Friedemann, Series Editor, Mitchell, John C., Series Editor, Naor, Moni, Series Editor, Pandu Rangan, C., Series Editor, Steffen, Bernhard, Series Editor, Terzopoulos, Demetri, Series Editor, Tygar, Doug, Series Editor, Leal-Taixé, Laura, editor, and Roth, Stefan, editor
- Published
- 2019
- Full Text
- View/download PDF
49. AUTOMATING THE DATASET GENERATION AND ANNOTATION FOR A DEEP LEARNING BASED ROBOT TRAJECTORY ADJUSTMENT APPLICATION FOR WELDING PROCESSES IN THE AUTOMOTIVE INDUSTRY.
- Author
-
SLIM WERDA, Mohamed, AL SAIFY, Theodor, KOUISS, Khalid, and GABER, Jaafar
- Subjects
DEEP learning ,AUTOMOBILE industry ,WELDING ,ARTIFICIAL intelligence ,COMPUTERS ,WELDED joints - Abstract
Industrial companies are more and more interested in the use of artificial intelligence (AI) in the control and monitoring of their processes. They try to take advantage of the power of this technology in order to increase the level of automation and to build smarter machines with new capabilities of self-adaptation and self-control. Especially, the automotive industry, with their high requirements in productivity and diversity management, are eager to adapt AI concepts to their processes. However, the training of Deep Learning (DL) models requires an important effort of data preparation, providing a dataset of all possible configurations. Indeed, this dataset must be collected and then annotated. Considering the fact that automotive industry deals with a huge number of references and that it often and quickly needs to modify their products, it is very difficult, if not impossible, to gather sufficient datasets for each produced reference and to have the time to train DL models in the plants with the traditional methods. This paper presents an innovative methodology to prepare the dataset by creating virtual images instead of collecting real ones and then automatically annotating them. It will demonstrate that this method will reduce the efforts and the time of the preparation of the dataset significantly. The paper will also present how this method was deployed for the quality control of welding operations in the automotive industry. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
50. Hodnoty slovesných morfologických kategorií v korpusu SYN2020 -- atribut verbtag.
- Author
-
Jelínek, Tomáš, Petkevič, Vladimír, and Skoumalová, Hana
- Subjects
VERBS ,MORPHOSYNTAX ,DISEASE susceptibility ,CORPORA ,DETECTIVES ,GENE ontology - Abstract
The paper describes the verbtag attribute, which allows a user to search, in the SYN2020 corpus (and also subsequent corpora, SYNv9 and SYNv10) of contemporary Czech, for all values of morphological categories of verbs, i.e., not only those contained in the tag attribute, but also those related mainly to multi-word participial verb predicates, which are prevalent in Czech. The verbtag attribute contains information indicating whether the verb (co-)forming the verbal meaning is either auxiliary or autosemantic, as well as information about the verb mode, diathesis, person, number and tense. The annotation applies both to verb predicates expressed in a single word (e.g., the 1st person indicative present tense: Čtu rád detektivní příběhy. 'I like to read detective stories.') and (especially) to verb predicates expressed in multiple words (e.g., the present conditional of the 1st person singular: Pak bych mu s chutí nabídla výhodnou smlouvu. 'Then I would gladly offer him a good deal.'). The authors introduce the motivation and the concept of the verbtag annotation, describe relevant morphological categories and their values in detail, and show, via examples, how various multiword structures expressing verbal meaning are annotated in the verbtag attribute. They also offer users a guide to the whole issue of verbal morphosyntax manifested in the verbtag attribute and possibilities for efficient search for and retrieval of morphological/morphosyntactic data. The paper shows which multiple verb complexes are simple in terms of annotation, but also identifies more complex cases (e.g., coordination of participles) which are not easy to automatically annotate, and/or whose annotation is unclear in terms of an adequate theoretical approach. The authors also present the method used for annotating multiword verbal complexes and its current success rate. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.