16,743 results on '"ground truth"'
Search Results
2. Toward Blockchain-Based Crowdsourcing for Machine Learning Ground Truth
- Author
-
Alzahrani, Asma, Alahmadi, Dimah, Alharbi, Nesreen, Kacprzyk, Janusz, Series Editor, and Hamdan, Allam, editor
- Published
- 2025
- Full Text
- View/download PDF
3. Approaches for Benchmarking Single-Cell Gene Regulatory Network Methods.
- Author
-
Karamveer and Uzun, Yasin
- Subjects
- *
GENETIC models , *CELL differentiation , *GENE expression , *EPIGENOMICS , *GENOMICS , *GENE regulatory networks - Abstract
Gene regulatory networks are powerful tools for modeling genetic interactions that control the expression of genes driving cell differentiation, and single-cell sequencing offers a unique opportunity to build these networks with high-resolution genomic data. There are many proposed computational methods to build these networks using single-cell data, and different approaches are used to benchmark these methods. However, a comprehensive discussion specifically focusing on benchmarking approaches is missing. In this article, we lay the GRN terminology, present an overview of common gold-standard studies and data sets, and define the performance metrics for benchmarking network construction methodologies. We also point out the advantages and limitations of different benchmarking approaches, suggest alternative ground truth data sets that can be used for benchmarking, and specify additional considerations in this context. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. Ant: a process aware annotation software for regulatory compliance.
- Author
-
Gyory, Raphaël, Restrepo Amariles, David, Lewkowicz, Gregory, and Bersini, Hugues
- Subjects
ANNOTATIONS & citations (Law) ,REGULATORY compliance ,MACHINE learning ,AUDITING ,ENGINEERING law ,COMPUTER software - Abstract
Accurate data annotation is essential to successfully implementing machine learning (ML) for regulatory compliance. Annotations allow organizations to train supervised ML algorithms and to adapt and audit the software they buy. The lack of annotation tools focused on regulatory data is slowing the adoption of established ML methodologies and process models, such as CRISP-DM, in various legal domains, including in regulatory compliance. This article introduces Ant, an open-source annotation software for regulatory compliance. Ant is designed to adapt to complex organizational processes and enable compliance experts to be in control of ML projects. By drawing on Business Process Modeling (BPM), we show that Ant can contribute to lift major technical bottlenecks to effectively implement regulatory compliance through software, such as the access to multiple sources of heterogeneous data and the integration of process complexities in the ML pipeline. We provide empirical data to validate the performance of Ant, illustrate its potential to speed up the adoption of ML in regulatory compliance, and highlight its limitations. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. The evolution of raw data archiving and the growth of its importance in crystallography
- Author
-
John R. Helliwell, James R. Hester, Loes M. J. Kroon-Batenburg, Brian McMahon, and Selina L. S. Storm
- Subjects
raw data measuring hardware ,raw data archive hardware ,raw data processing software ,raw data policies at photon and neutron facilities ,ground truth ,Crystallography ,QD901-999 - Abstract
The hardware for data archiving has expanded capacities for digital storage enormously in the past decade or more. The IUCr evaluated the costs and benefits of this within an official working group which advised that raw data archiving would allow ground truth reproducibility in published studies. Consultations of the IUCr's Commissions ensued via a newly constituted standing advisory committee, the Committee on Data. At all stages, the IUCr financed workshops to facilitate community discussions and possible methods of raw data archiving implementation. The recent launch of the IUCrData journal's Raw Data Letters is a milestone in the implementation of raw data archiving beyond the currently published studies: it includes diffraction patterns that have not been fully interpreted, if at all. The IUCr 75th Congress in Melbourne included a workshop on raw data reuse, discussing the successes and ongoing challenges of raw data reuse. This article charts the efforts of the IUCr to facilitate discussions and plans relating to raw data archiving and reuse within the various communities of crystallography, diffraction and scattering.
- Published
- 2024
- Full Text
- View/download PDF
6. Frame-Based Change Detection Using Histogram and Threshold to Separate Moving Objects from Dynamic Background.
- Author
-
Jasim, Hala A., George, Loay E., and Abd-almajied, Mohammed I.
- Subjects
- *
HISTOGRAMS , *CAMERAS , *ALGORITHMS - Abstract
Detecting and subtracting the Motion objects from backgrounds is one of the most important areas. The development of cameras and their widespread use in most areas of security, surveillance, and others made face this problem. The difficulty of this area is unstable in the classification of the pixels (foreground or background). This paper proposed a suggested background subtraction algorithm based on the histogram. The classification threshold is adaptively calculated according to many tests. The performance of the proposed algorithms was compared with state-of-the-art methods in complex dynamic scenes. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
7. Automated Classification of Physiologic, Glaucomatous, and Glaucoma-Suspected Optic Discs Using Machine Learning.
- Author
-
Diener, Raphael, Renz, Alexander W., Eckhard, Florian, Segbert, Helmar, Eter, Nicole, Malcherek, Arnim, and Biermann, Julia
- Subjects
- *
OPTIC disc , *MACHINE learning , *VISUAL fields , *NERVE fibers , *CLASSIFICATION , *DIABETIC retinopathy - Abstract
In order to generate a machine learning algorithm (MLA) that can support ophthalmologists with the diagnosis of glaucoma, a carefully selected dataset that is based on clinically confirmed glaucoma patients as well as borderline cases (e.g., patients with suspected glaucoma) is required. The clinical annotation of datasets is usually performed at the expense of the data volume, which results in poorer algorithm performance. This study aimed to evaluate the application of an MLA for the automated classification of physiological optic discs (PODs), glaucomatous optic discs (GODs), and glaucoma-suspected optic discs (GSODs). Annotation of the data to the three groups was based on the diagnosis made in clinical practice by a glaucoma specialist. Color fundus photographs and 14 types of metadata (including visual field testing, retinal nerve fiber layer thickness, and cup–disc ratio) of 1168 eyes from 584 patients (POD = 321, GOD = 336, GSOD = 310) were used for the study. Machine learning (ML) was performed in the first step with the color fundus photographs only and in the second step with the images and metadata. Sensitivity, specificity, and accuracy of the classification of GSOD vs. GOD and POD vs. GOD were evaluated. Classification of GOD vs. GSOD and GOD vs. POD performed in the first step had AUCs of 0.84 and 0.88, respectively. By combining the images and metadata, the AUCs increased to 0.92 and 0.99, respectively. By combining images and metadata, excellent performance of the MLA can be achieved despite having only a small amount of data, thus supporting ophthalmologists with glaucoma diagnosis. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
8. Unmanned air/ground vehicle survey following a radiological dispersal event.
- Author
-
Munsie, Timothy, Beckman, Blake, Fawkes, Ross, Shippen, Alan B., Fairbrother, Blaine, and Green, Anna Rae
- Subjects
REMOTELY piloted vehicles ,DRONE aircraft ,AUTONOMOUS vehicles ,AERIAL surveys - Abstract
This paper presents the method and results of surveying a dispersed radioactive field using an unmanned air vehicle (UAV) and an unmanned ground vehicle (UGV). A 35 GBq of 140 ${}^{140}$La material was distributed in a specific geometric l‐polygon pattern measuring a 120 m × $\times $ 20 m longitudinal rectangle and an 80 m × $\times $ 10 m transverse rectangle. Two methods were used to determine the amount and distribution of the lanthanum over the polygon including 20 plywood coupons distributed over the area and a UGV equipped with a Kromek GR1® driving over the area. The aerial survey was conducted using an unmanned aerial vehicle UAV carrying the Kromek GR1® while flying a traditional grid pattern and circular pattern at different elevations and speeds over the area. The data collected by the UAV were further postprocessed using N‐Visage, a 3D radiation modeling method developed by Createc, to create a model of the ground activity. This model was compared with the reference data collected on the ground by the UGV and was found to be in agreement with 11%. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
9. Artificial Intelligence in Computed Tomography Image Reconstruction
- Author
-
Seeram, Euclid, Kanade, Vijay, Seeram, Euclid, and Kanade, Vijay
- Published
- 2024
- Full Text
- View/download PDF
10. Training Decisions: Ground-Truthing the Interesting
- Author
-
Hind, Sam and Hind, Sam
- Published
- 2024
- Full Text
- View/download PDF
11. Ground Truth from Multiple Manually Marked Images to Evaluate Blood Vessel Segmentation
- Author
-
Tariq, Nazish, Tang, Michael Chi Seng, Ibrahim, Haidi, Siang, Teoh Soo, Embong, Zunaina, Hamid, Aini Ismafairus Abd, Zainon, Rafidah, Angrisani, Leopoldo, Series Editor, Arteaga, Marco, Series Editor, Chakraborty, Samarjit, Series Editor, Chen, Jiming, Series Editor, Chen, Shanben, Series Editor, Chen, Tan Kay, Series Editor, Dillmann, Rüdiger, Series Editor, Duan, Haibin, Series Editor, Ferrari, Gianluigi, Series Editor, Ferre, Manuel, Series Editor, Jabbari, Faryar, Series Editor, Jia, Limin, Series Editor, Kacprzyk, Janusz, Series Editor, Khamis, Alaa, Series Editor, Kroeger, Torsten, Series Editor, Li, Yong, Series Editor, Liang, Qilian, Series Editor, Martín, Ferran, Series Editor, Ming, Tan Cher, Series Editor, Minker, Wolfgang, Series Editor, Misra, Pradeep, Series Editor, Mukhopadhyay, Subhas, Series Editor, Ning, Cun-Zheng, Series Editor, Nishida, Toyoaki, Series Editor, Oneto, Luca, Series Editor, Panigrahi, Bijaya Ketan, Series Editor, Pascucci, Federica, Series Editor, Qin, Yong, Series Editor, Seng, Gan Woon, Series Editor, Speidel, Joachim, Series Editor, Veiga, Germano, Series Editor, Wu, Haitao, Series Editor, Zamboni, Walter, Series Editor, Zhang, Junjie James, Series Editor, Tan, Kay Chen, Series Editor, Ahmad, Nur Syazreen, editor, Mohamad-Saleh, Junita, editor, and Teh, Jiashen, editor
- Published
- 2024
- Full Text
- View/download PDF
12. Developing a Traffic Analysis Suite for Modified Packet Capture File
- Author
-
Morozova, O. P., Orlova, M. A., Naumov, N. A., Abrosimov, L. I., Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Vishnevskiy, Vladimir M., editor, Samouylov, Konstantin E., editor, and Kozyrev, Dmitry V., editor
- Published
- 2024
- Full Text
- View/download PDF
13. Ground Truth Crossfield Guided Mesher-Native Box Imprinting for Automotive Crash Analysis
- Author
-
Mukherjee, Nilanjan, Barth, Timothy J., Series Editor, Griebel, Michael, Series Editor, Keyes, David E., Series Editor, Nieminen, Risto M., Series Editor, Roose, Dirk, Series Editor, Schlick, Tamar, Series Editor, Ruiz-Gironés, Eloi, editor, Sevilla, Rubén, editor, and Moxey, David, editor
- Published
- 2024
- Full Text
- View/download PDF
14. Performance Assessment and Dataset Description
- Author
-
Paul, Arati, Chaki, Nabendu, Paul, Arati, and Chaki, Nabendu
- Published
- 2024
- Full Text
- View/download PDF
15. Synergizing Deep Learning-Enabled Preprocessing and Human–AI Integration for Efficient Automatic Ground Truth Generation.
- Author
-
Collazo, Christopher, Vargas, Ian, Cara, Brendon, Weinheimer, Carla J., Grabau, Ryan P., Goldgof, Dmitry, Hall, Lawrence, Wickline, Samuel A., and Pan, Hua
- Subjects
- *
SUPERVISED learning , *DEEP learning , *CONVOLUTIONAL neural networks , *ACTIVE learning , *IMAGE analysis - Abstract
The progress of incorporating deep learning in the field of medical image interpretation has been greatly hindered due to the tremendous cost and time associated with generating ground truth for supervised machine learning, alongside concerns about the inconsistent quality of images acquired. Active learning offers a potential solution to these problems of expanding dataset ground truth by algorithmically choosing the most informative samples for ground truth labeling. Still, this effort incurs the costs of human labeling, which needs minimization. Furthermore, automatic labeling approaches employing active learning often exhibit overfitting tendencies while selecting samples closely aligned with the training set distribution and excluding out-of-distribution samples, which could potentially improve the model's effectiveness. We propose that the majority of out-of-distribution instances can be attributed to inconsistent cross images. Since the FDA approved the first whole-slide image system for medical diagnosis in 2017, whole-slide images have provided enriched critical information to advance the field of automated histopathology. Here, we exemplify the benefits of a novel deep learning strategy that utilizes high-resolution whole-slide microscopic images. We quantitatively assess and visually highlight the inconsistencies within the whole-slide image dataset employed in this study. Accordingly, we introduce a deep learning-based preprocessing algorithm designed to normalize unknown samples to the training set distribution, effectively mitigating the overfitting issue. Consequently, our approach significantly increases the amount of automatic region-of-interest ground truth labeling on high-resolution whole-slide images using active deep learning. We accept 92% of the automatic labels generated for our unlabeled data cohort, expanding the labeled dataset by 845%. Additionally, we demonstrate expert time savings of 96% relative to manual expert ground-truth labeling. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
16. A systematic review of the validity of Criteria-based Content Analysis in child sexual abuse cases and other field studies.
- Author
-
Sporer, Siegfried Ludwig and Masip, Jaume
- Abstract
Criteria-based Content Analysis (CBCA) has been primarily employed to assess the credibility of child sexual abuse (CSA) allegations. However, several studies on the validity of CBCA have focused on autobiographical events other than CSA. Because of the differences between real cases and the laboratory, we focused specifically on CBCA field studies on both CSA and other areas of application. We formally assessed several ground-truth criteria (and other methodological aspects) in a pool of 36 field studies. Seven archival studies (six of which were on CSA) and seven quasi-experiments (none of which was on CSA) were found to be either methodologically sound (12 studies) or acceptable with reservations (two studies), and were therefore included. We describe the paradigm and methods used in each study. Across studies, most CBCA criteria significantly differed between truthful and deceptive accounts, with similar medium to large effect sizes for the methodologically sound quasi-experiments and archival CSA studies. Our review shows that CBCA criteria may discriminate in domains other than CSA. The implications for the real-world usage of CBCA are discussed. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
17. WebLabel: OpenLABEL-compliant multi-sensor labelling.
- Author
-
Urbieta, Itziar, Mujika, Andoni, Piérola, Gonzalo, Irigoyen, Eider, Nieto, Marcos, Loyo, Estibaliz, and Aginako, Naiara
- Abstract
Annotated datasets have become crucial for training Machine Learning (ML) models for developing Autonomous Vehicles (AVs) and their functions. Generating these datasets usually involves a complex coordination of automation and manual effort. Moreover, most available labelling tools focus on specific media types (e.g., images or video). Consequently, they cannot perform complex labelling tasks for multi-sensor setups. Recently, ASAM published OpenLABEL, a standard designed to specify an annotation format flexible enough to support the development of automated driving features and to guarantee interoperability among different systems and providers. In this work, we present WebLabel, the first multipurpose web application tool for labelling complex multi-sensor data that is fully compliant with OpenLABEL 1.0. The proposed work analyses several labelling use cases demonstrating the standard's benefits and the application's flexibility to cover various heterogeneous requirements: image labelling, multi-view video object annotation, point-cloud view-based labelling for 3D geometries and action recognition. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
18. Comparison of ASI-PRISMA Data, DLR-EnMAP Data, and Field Spectrometer Measurements on "Sale 'e Porcus", a Salty Pond (Sardinia, Italy).
- Author
-
Musacchio, Massimo, Silvestri, Malvina, Romaniello, Vito, Casu, Marco, Buongiorno, Maria Fabrizia, and Melis, Maria Teresa
- Subjects
- *
ENVIRONMENTAL mapping , *SPECTROMETERS , *CARTOGRAPHY software , *PONDS , *RADIANCE , *REFLECTANCE - Abstract
A comparison between the ASI-PRISMA (Agenzia Spaziale Italiana-PRecursore IperSpettrale della Missione Applicativa) DLR-EnMAP (German Aerospace Center—Environmental Mapping and Analysis Program) data and field spectrometer measurements has been performed. The test site, located at the "Sale 'e Porcus" pond (hereafter SPp) in Western Sardinia, Italy, offers particularly homogenous characteristics, making it an ideal location not only for experimentation but also for calibration purposes. Three remote-sensed data acquisitions have been performed by these agencies (ASI and DLR) starting on 14 July 2023 and continuing until 22 July 2023. The DLR-EnMAP data acquired on 22 July overestimates both that of the ASI-PRISMA and the 14 July DLR-EnMAP radiance in the VNIR region, while all the datasets are close to each other, up to 2500 nm, for all considered days. The average absolute mean difference between the reflectance values estimated by the ASI-PRISMA and DLR-EnMAP, in the test area, is around 0.015, despite the small difference in their time of acquisition (8 days); their maximum relative difference value occurs at about 2100 nm. In this study, we investigate the relationship between the averaged ground truth value of reflectance, acquired by means of a portable ASD FieldSpec spectoradiometer, characterizing the test site and the EO reflectance data derived from the official datasets. FieldSpec measurements confirm the quality of both the ASI-PRISMA and DLR-EnMAP's reflectance estimations. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
19. Development and use of a co-produced short mood survey to collect ground truth in digital footprints research
- Author
-
Nina Di Cara, Oliver Davis, and Claire Haworth
- Subjects
EMA ,experience sampling ,ground truth ,mood ,measurement ,Demography. Population. Vital events ,HB848-3697 - Abstract
Introduction & Background To use digital footprint data for mental health and well-being research we often need to collect concurrent, high-quality measures of ground truth. Delivering frequent surveys to participants using an ecological momentary assessment (EMA) methodology is one way to collect such data. However, existing surveys tend to be long, not focused on momentary states or rely on rating images which are not platform agnostic. Here we present a five-item test-based survey designed with participants and validated for use in EMA studies to collect data about momentary changes in mood. We describe its methodological development and how it has been used to investigate music listening on Spotify as a digital footprint of mood. Objectives & Approach The survey is based on the circumplex model of affect. It was co-produced with a participant advisory group (N=5), who gave feedback on the length, content and delivery of the survey. It was then piloted in a group of N=98 participants to assess statistical validity, and congruence with the 20-item Positive and Negative Affect Schedule (PANAS). Following this it was delivered in a wider sample (N=150) four times a day over a two-week period using an EMA app on participant’s phones. Relevance to Digital Footprints EMA is an increasingly popular method for collecting ground truth to support the interpretation of digital footprint data. This newly developed and tested mood survey offers an opportunity to reduce participant burden for collecting mood data in EMA studies which will support the collection of high quality and high time-resolution ground truth for digital footprints research. Results Together with participants we selected four emotions across the axes of arousal and valence, as well as rumination which participants considered important in their music listening behaviors. Factor analysis of pilot data showed that the questions represented two factors of positive and negative affect. The ratings on a 0-10 scale of the emotions ‘cheerful’ and ‘relaxed’ explained 44% of the variance in positive affect, and ratings of ‘worried’, ‘sad’ and ‘frustrated’ explained 40% of the variance in negative affect. Delivery of the questionnaire in a wider student sample (N=150) four times per day for two weeks allowed for the opportunity to assess typical response rates in a realistic EMA setting. On average participants completed 3 out of the 4 surveys a day. Conclusions & Implications The co-created, short mood survey for the collection of ground truth in digital footprint studies was validated across two independent samples, and shown to allow for good response rates in a two week study. Future testing on wider samples will provide opportunities to validate the survey and assess its effectiveness across demographic groups and different sample types.
- Published
- 2024
- Full Text
- View/download PDF
20. Reichsanzeiger-GT: An OCR ground truth dataset based on the historical newspaper 'Deutscher Reichsanzeiger und Preußischer Staatsanzeiger' (German Imperial Gazette and Prussian Official Gazette) (1819–1945)
- Author
-
Thomas Schmidt, Jan Kamlah, and Stefan Weil
- Subjects
OCR ,Text recognition ,Ground truth ,Historical newspapers ,Computer applications to medicine. Medical informatics ,R858-859.7 ,Science (General) ,Q1-390 - Abstract
Reichsanzeiger-GT is a ground truth dataset for OCR training and evaluation based on the historical German newspaper “Deutscher Reichsanzeiger und Preußischer Staatsanzeiger” (German Imperial Gazette and Prussian Official Gazette), which was published from 1819 to 1945 and printed mostly in the typeface Fraktur (Black Letter). The dataset consists of 101 newspaper pages for the years 1820–1939, that cover a wide variety of topics, page layouts (lists, tables, and advertisements) as well as different typefaces. Using the transcription software Transkribus and the open-source OCR engine Tesseract we automatically created and manually corrected layout segmentations and transcriptions for each page, resulting in 65,563 text regions, 412 table regions, 119,429 text lines and 490,679 words. By applying transcription guidelines that preserve the printing conditions, the dataset contains language and printing specific phenomena like the historical use of glyphs like long s (ſ), rotunda r (ꝛ), and historical currency symbols (M, ₰) among others. The dataset is provided in two variants in PAGE XML format. The first one contains ground truth data with table regions transformed to text regions for easier processing. The second variant preserves all table regions. Researchers can reuse this dataset to train new or finetune existing text recognition or layout segmentation models. The dataset can also be used to evaluate the accuracy of existing OCR models. Using specific, community driven transcription guidelines our dataset is easily interoperable and reusable with other datasets based on the same transcription level.
- Published
- 2024
- Full Text
- View/download PDF
21. A Comparative Study of State-of-the-Art Deep Learning Models for Semantic Segmentation of Pores in Scanning Electron Microscope Images of Activated Carbon
- Author
-
Bishwas Pokharel, Deep Shankar Pandey, Anjuli Sapkota, Bhimraj Yadav, Vasanta Gurung, Mandira Pradhananga Adhikari, Lok Nath Regmi, and Nanda Bikram Adhikari
- Subjects
Semantic segmentation ,SEM images ,activated carbon ,ground truth ,Adam optimizer ,intersection over union ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Accurate measurement of the microspores, mesopores, and macropores on the surface of the activated carbon is essential due to its direct influence on the material’s adsorption capacity, surface area, and overall performance in various applications like water purification, air filtration, and gas separation. Traditionally, Scanning Electron Microscopy (SEM) images of activated carbons are collected and manually annotated by a human expert to differentiate and measure different pores on the surface. However, manual analysis of such surfaces is costly, time-consuming, and resource-intensive, as it requires expert supervision. In this paper, we propose an automatic deep-learning-based solution to address this challenge of activated carbon surface segmentation. Our deep-learning approach optimizes pore analysis by reducing time and resources, eliminating human subjectivity, and effectively adapting to diverse pore structures and imaging conditions. We introduce a novel SEM image segmentation dataset for activated carbon, comprising 128 images that capture the variability in pore sizes, structures, and imaging artifacts. Challenges encountered during dataset creation, irregularities in pore structures, and the presence of impurities were addressed to ensure robust model performance. We then evaluate the state-of-the-art deep learning models on the novel semantic segmentation task that shows promising results. Notably, DeepLabV3Plus, DeepLabV3, and FPN emerge as the most promising models based on semantic segmentation test results, with DeepLabV3Plus achieving the highest test Dice coefficient of 68.68%. Finally, we outline the key research challenges and discuss potential research directions to address these challenges.
- Published
- 2024
- Full Text
- View/download PDF
22. P‐116: A Demoiré Method for Display Test Using CNN Model with Pixel Shift.
- Author
-
Tang, Hao, Zhang, Wuxing, and Xu, Gang
- Subjects
MASS production ,MACHINE learning ,ARTIFICIAL intelligence ,PIXELS ,TEST methods ,DEEP learning - Abstract
A deep learning based demoiré algorithm is applied to display test images to remove moiré patterns. The ground truth is generated by pixel shift (PS) method implemented by shifting display panels, instead of camera sensor, a low‐cost alternative to expensive PS cameras for mass production (MP). [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
23. AI and mental health: evaluating supervised machine learning models trained on diagnostic classifications
- Author
-
van Oosterzee, Anna
- Published
- 2024
- Full Text
- View/download PDF
24. Automatic caries detection in bitewing radiographs—Part II: experimental comparison.
- Author
-
Tichý, Antonín, Kunt, Lukáš, Nagyová, Valéria, and Kybic, Jan
- Abstract
Objective: The objective of this study was to compare the detection of caries in bitewing radiographs by multiple dentists with an automatic method and to evaluate the detection performance in the absence of a reliable ground truth. Materials and methods: Four experts and three novices marked caries using bounding boxes in 100 bitewing radiographs. The same dataset was processed by an automatic object detection deep learning method. All annotators were compared in terms of the number of errors and intersection over union (IoU) using pairwise comparisons, with respect to the consensus standard, and with respect to the annotator of the training dataset of the automatic method. Results: The number of lesions marked by experts in 100 images varied between 241 and 425. Pairwise comparisons showed that the automatic method outperformed all dentists except the original annotator in the mean number of errors, while being among the best in terms of IoU. With respect to a consensus standard, the performance of the automatic method was best in terms of the number of errors and slightly below average in terms of IoU. Compared with the original annotator, the automatic method had the highest IoU and only one expert made fewer errors. Conclusions: The automatic method consistently outperformed novices and performed as well as highly experienced dentists. Clinical significance: The consensus in caries detection between experts is low. An automatic method based on deep learning can improve both the accuracy and repeatability of caries detection, providing a useful second opinion even for very experienced dentists. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
25. Hong Kong UrbanNav: An Open-Source Multisensory Dataset for Benchmarking Urban Navigation Algorithms.
- Author
-
Li-Ta Hsu, Feng Huang, Hoi-Fung Ng, Guohao Zhang, Yihan Zhong, Xiwei Bai, and Weisong Wen
- Subjects
- *
GLOBAL Positioning System , *OPTICAL radar , *LIDAR , *VISUAL odometry , *RAILROAD tunnels - Abstract
Accurate positioning in urban canyons remains a challenging problem. To facilitate the research and development of reliable and precise positioning methods using multiple sensors in urban canyons, we built a multisensory dataset, UrbanNav, collected in diverse, challenging urban scenarios in Hong Kong. The dataset provides multi-sensor data, including data from multi-frequency global navigation satellite system (GNSS) receivers, an inertial measurement unit (IMU), multiple light detection and ranging (lidar) units, and cameras. Meanwhile, the ground truth of the positioning (with centimeter-level accuracy) is postprocessed by commercial software from NovAtel using an integrated GNSS real-time kinematic and fiber optics gyroscope inertial system. In this paper, the sensor systems, spatial and temporal calibration, data formats, and scenario descriptions are presented in detail. Meanwhile, the benchmark performance of several existing positioning methods is provided as a baseline. Based on the evaluations, we conclude that GNSS can provide satisfactory results in a middle-class urban canyon if an appropriate receiver and algorithms are applied. Both visual and lidar odometry are satisfactory in deep urban canyons, whereas tunnels are still a major challenge. Multisensory integration with the aid of an IMU is a promising solution for achieving seamless positioning in cities. The dataset in its entirety can be found on GitHub at https://github.com/ IPNL-POLYU/UrbanNavDataset. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
26. Evaluating data linkage algorithms with perfect synthetic ground truth
- Author
-
Dalton, Thomas Stanley, Kirby, Graham N. C., and Dearle, Alan
- Subjects
data linkage ,record linkage ,evaluation ,synthetic data ,ground truth ,synthetic ground truth ,gold standard ,linkage evaluation - Abstract
Data linkage algorithms join datasets by identifying commonalities between them. The ability to evaluate the efficacy of different algorithms is a challenging problem that is often overlooked. If incorrect links are made or links are missed by a linkage algorithm then conclusions based on its linkage may be unfounded. Evaluating linkage quality is particularly challenging in domains where datasets are large and the number of links is low. Example domains include historical population data, bibliographic data, and administrative data. In these domains the evaluation of linkage quality is not well understood. A common approach to evaluating linkage quality is the use of metrics, most commonly precision, recall, and F-measure. These metrics indicate how often links are missed or false links are made. To calculate a metric, datasets are used where the true links and non-links are known. The linkage algorithm attempts to link the datasets and the constructed set of links is compared with the set of true links. In these domains we can rarely have confidence that the evaluation datasets contain all the true links and that no false links have been included. If such errors exist in the evaluation datasets, the calculated metrics may not truly reflect the performance of the linkage algorithm. This presents issues when making comparisons between linkage algorithms. To rigorously evaluate the efficacy of linkage algorithms, it is necessary to objectively measure an algorithm's linkage quality with a range of different configuration parameters and datasets. These many datasets must be of appropriate scale and have ground truth which denotes all true links and non-links. Evaluating algorithms using shared standardised datasets enables objective comparisons between linkage algorithms. To facilitate objective linkage evaluation, a set of standardised datasets need to be shared and widely adopted. This thesis establishes an approach for the construction of synthetic datasets that can be used to evaluate linkage algorithms. This thesis addresses the following research questions: • What are appropriate approaches to the evaluation of linkage algorithms? • Is it feasible to synthesise realistic evaluation data? • Is synthetic evaluation data with perfect ground truth useful for evaluation? • How should synthesised data be statistically validated for correctness? • How should sets of synthesised data be used to evaluate linkage? • How can the evaluation of linkage algorithms be effectively communicated? This thesis makes a number of contributions, most notably a framework for the comprehensive evaluation of data linkage algorithms, thus significantly improving the comparability of linkage algorithms, especially in domains lacking evaluation data. The thesis demonstrates these techniques within the population reconstruction domain. Integral to the evaluation framework, approaches to synthesis and statistical validation of evaluation datasets have been investigated, resulting in a simulation model able to create many, characteristically varied, large-scale datasets.
- Published
- 2022
- Full Text
- View/download PDF
27. Tamil handwritten palm leaf manuscript dataset (THPLMD)
- Author
-
I. Jailingeswari and S. Gopinathan
- Subjects
Tamil palm leaf ,Enhancement ,Binarization ,Segmentation ,Ground truth ,Otsu ,Computer applications to medicine. Medical informatics ,R858-859.7 ,Science (General) ,Q1-390 - Abstract
Most palm leaf manuscripts are generally accessible in deteriorated condition, including cracks, discoloration, moisture and humidity, and insects bite. Such a manuscript is considered challenging in the research field. We captured deteriorated Tamil palm leaves around 262 dataset samples are ‘Naladiyar(27)’,’ Tholkappiyam(221)’, and’ Thirikadugam(14)’ which are genned up mortal health, discipline, authoritative text on Tamil grammar. We contribute the high-quality raw dataset with the aid of a Nikon camera, pre-enhance samples by editing software tool, and applied the Otsu threshold to deliver the ground images through binarization as readily accessible content presenting a highly time-consuming task to play a vital role in Machine/Deep/ Transfer learning, AI, and ANN
- Published
- 2024
- Full Text
- View/download PDF
28. Exploring Data Provenance in Handwritten Text Recognition Infrastructure: Sharing and Reusing Ground Truth Data, Referencing Models, and Acknowledging Contributions. Starting the Conversation on How We Could Get It Done
- Author
-
C. Annemieke Romein, Tobias Hodel, Femke Gordijn, Joris J. van Zundert, Alix Chagué, Milan van Lange, Helle Strandgaard Jensen, Andy Stauder, Jake Purcell, Melissa M. Terras, Pauline van den Heuvel, Carlijn Keijzer, Achim Rabus, Chantal Sitaram, Aakriti Bhatia, Katrien Depuydt, Mary Aderonke Afolabi-Adeolu, Anastasiia Anikina, Elisa Bastianello, Lukas Vincent Benzinger, Arno Bosse, David Brown, Ash Charlton, André Nilsson Dannevig, Klaas van Gelder, Sabine C.P.J. Go, Marcus J.C. Goh, Silvia Gstrein, Sewa Hasan, Stefan von der Heide, Maximilian Hindermann, Dorothee Huff, Ineke Huysman, Ali Idris, Liesbeth Keijzer, Simon Kemper, Sanne Koenders, Erika Kuijpers, Lisette Rønsig Larsen, Sven Lepa, Tommy O. Link, Annelies van Nispen, Joe Nockels, Laura M. van Noort, Joost Johannes Oosterhuis, Vivien Popken, María Estrella Puertollano, Joosep J. Puusaag, Ahmed Sheta, Lex Stoop, Ebba Strutzenbladh, Nicoline van der Sijs, Jan Paul van der Spek, Barry Benaissa Trouw, Geertrui Van Synghel, Vladimir Vučković, Heleen Wilbrink, Sonia Weiss, David Joseph Wrisley, and Riet Zweistra
- Subjects
automatic text recognition ,handwritten text recognition ,data publication ,open data ,data provenance ,data curation ,ground truth ,sharing ,History of scholarship and learning. The humanities ,AZ20-999 ,Bibliography. Library science. Information resources - Abstract
This paper discusses best practices for sharing and reusing Ground Truth in Handwritten Text Recognition infrastructures, as well as ways to reference and acknowledge contributions to the creation and enrichment of data within these systems. We discuss how one can place Ground Truth data in a repository and, subsequently, inform others through HTR-United. Furthermore, we want to suggest appropriate citation methods for ATR data, models, and contributions made by volunteers. Moreover, when using digitised sources (digital facsimiles), it becomes increasingly important to distinguish between the physical object and the digital collection. These topics all relate to the proper acknowledgement of labour put into digitising, transcribing, and sharing Ground Truth HTR data. This also points to broader issues surrounding the use of machine learning in archival and library contexts, and how the community should begin to acknowledge and record both contributions and data provenance.
- Published
- 2024
- Full Text
- View/download PDF
29. IS AI GROUND TRUTH REALLY TRUE? THE DANGERS OF TRAINING AND EVALUATING AI TOOLS BASED ON EXPERTS' KNOW-WHAT.
- Author
-
Lebovitz, Sarah, Levina, Natalia, and Lifshitz-Assaf, Hila
- Abstract
Organizational decision-makers need to evaluate AI tools in light of increasing claims that such tools outperform human experts. Yet, measuring the quality of knowledge work is challenging, raising the question of how to evaluate AI performance in such contexts. We investigate this question through a field study of a major U.S. hospital, observing how managers evaluated five different machine-learning (ML) based AI tools. Each tool reported high performance according to standard AI accuracy measures, which were based on ground truth labels provided by qualified experts. Trying these tools out in practice, however, revealed that none of them met expectations. Searching for explanations, managers began confronting the high uncertainty of experts' know-what knowledge captured in ground truth labels used to train and validate ML models. In practice, experts address this uncertainty by drawing on rich know-how practices, which were not incorporated into these ML-based tools. Discovering the disconnect between AI's know-what and experts' know-how enabled managers to better understand the risks and benefits of each tool. This study shows dangers of treating ground truth labels used in ML models objectively when the underlying knowledge is uncertain. We outline implications of our study for developing, training, and evaluating AI for knowledge work. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
30. SESAME System
- Author
-
Hamada, Yohei, Shigenaga, Yukihisa, Takahashi, Hidenori, Osaki, Mitsuru, editor, Tsuji, Nobuyuki, editor, Kato, Tsuyoshi, editor, and Sulaiman, Albertus, editor
- Published
- 2023
- Full Text
- View/download PDF
31. Street Naming in Malta as a Geo-Cultural and Political Exercise as Seen from Local Sources
- Author
-
Schembri, John A., Gauci, Ritienne, Koutsopoulos, Kostis C., Series Editor, Miguel González, Rafael De, Series Editor, Schmeinck, Daniela, Series Editor, and O’Reilly, Gerry, editor
- Published
- 2023
- Full Text
- View/download PDF
32. Insect Image Semantic Segmentation and Identification Using UNET and DeepLab V3+
- Author
-
Bose, Kunal, Shubham, Kumar, Tiwari, Vivek, Patel, Kuldip Singh, Kacprzyk, Janusz, Series Editor, Gomide, Fernando, Advisory Editor, Kaynak, Okyay, Advisory Editor, Liu, Derong, Advisory Editor, Pedrycz, Witold, Advisory Editor, Polycarpou, Marios M., Advisory Editor, Rudas, Imre J., Advisory Editor, Wang, Jun, Advisory Editor, Tuba, Milan, editor, Akashe, Shyam, editor, and Joshi, Amit, editor
- Published
- 2023
- Full Text
- View/download PDF
33. Design, Implementation, and Evaluation of an External Pose-Tracking System for Underwater Cameras
- Author
-
Winkel, Birger, Nakath, David, Woelk, Felix, and Köser, Kevin
- Published
- 2024
- Full Text
- View/download PDF
34. Distributed Visual-Based Ground Truth System For Mobile Robotics
- Author
-
Sergey Sorokumov, Sergey Glazunov, and Konstantin Chaika
- Subjects
ground truth ,mobile robots ,calibration ,perspective ,apriltag ,opencv ,visualization ,Telecommunication ,TK5101-6720 - Abstract
The paper considers ground truth systems for mobile robots, which are gaining popularity due to the development of artificial intelligence algorithms and autonomous robot control systems. In the course of the work, we compared the methods of localization by visualization criteria during the experiment, localization in the coordinates of the test polygon, and visualization of the data on the layout of the polygon. As a result of the analysis, the advantages and disadvantages of the systems under consideration were identified, on the basis of which the solution method for the system under development was determined. The implemented distributed ground truth system uses affine transformations to calculate the position of the robot. Introduced a new criterion that describes the cost of accuracy per square.
- Published
- 2023
- Full Text
- View/download PDF
35. Automated Classification of Physiologic, Glaucomatous, and Glaucoma-Suspected Optic Discs Using Machine Learning
- Author
-
Raphael Diener, Alexander W. Renz, Florian Eckhard, Helmar Segbert, Nicole Eter, Arnim Malcherek, and Julia Biermann
- Subjects
machine learning ,glaucoma ,glaucoma suspects ,data annotation ,ground truth ,Medicine (General) ,R5-920 - Abstract
In order to generate a machine learning algorithm (MLA) that can support ophthalmologists with the diagnosis of glaucoma, a carefully selected dataset that is based on clinically confirmed glaucoma patients as well as borderline cases (e.g., patients with suspected glaucoma) is required. The clinical annotation of datasets is usually performed at the expense of the data volume, which results in poorer algorithm performance. This study aimed to evaluate the application of an MLA for the automated classification of physiological optic discs (PODs), glaucomatous optic discs (GODs), and glaucoma-suspected optic discs (GSODs). Annotation of the data to the three groups was based on the diagnosis made in clinical practice by a glaucoma specialist. Color fundus photographs and 14 types of metadata (including visual field testing, retinal nerve fiber layer thickness, and cup–disc ratio) of 1168 eyes from 584 patients (POD = 321, GOD = 336, GSOD = 310) were used for the study. Machine learning (ML) was performed in the first step with the color fundus photographs only and in the second step with the images and metadata. Sensitivity, specificity, and accuracy of the classification of GSOD vs. GOD and POD vs. GOD were evaluated. Classification of GOD vs. GSOD and GOD vs. POD performed in the first step had AUCs of 0.84 and 0.88, respectively. By combining the images and metadata, the AUCs increased to 0.92 and 0.99, respectively. By combining images and metadata, excellent performance of the MLA can be achieved despite having only a small amount of data, thus supporting ophthalmologists with glaucoma diagnosis.
- Published
- 2024
- Full Text
- View/download PDF
36. Synergizing Deep Learning-Enabled Preprocessing and Human–AI Integration for Efficient Automatic Ground Truth Generation
- Author
-
Christopher Collazo, Ian Vargas, Brendon Cara, Carla J. Weinheimer, Ryan P. Grabau, Dmitry Goldgof, Lawrence Hall, Samuel A. Wickline, and Hua Pan
- Subjects
machine learning ,active deep learning ,ground truth ,convolutional neural network ,deep learning-based preprocessing ,Technology ,Biology (General) ,QH301-705.5 - Abstract
The progress of incorporating deep learning in the field of medical image interpretation has been greatly hindered due to the tremendous cost and time associated with generating ground truth for supervised machine learning, alongside concerns about the inconsistent quality of images acquired. Active learning offers a potential solution to these problems of expanding dataset ground truth by algorithmically choosing the most informative samples for ground truth labeling. Still, this effort incurs the costs of human labeling, which needs minimization. Furthermore, automatic labeling approaches employing active learning often exhibit overfitting tendencies while selecting samples closely aligned with the training set distribution and excluding out-of-distribution samples, which could potentially improve the model’s effectiveness. We propose that the majority of out-of-distribution instances can be attributed to inconsistent cross images. Since the FDA approved the first whole-slide image system for medical diagnosis in 2017, whole-slide images have provided enriched critical information to advance the field of automated histopathology. Here, we exemplify the benefits of a novel deep learning strategy that utilizes high-resolution whole-slide microscopic images. We quantitatively assess and visually highlight the inconsistencies within the whole-slide image dataset employed in this study. Accordingly, we introduce a deep learning-based preprocessing algorithm designed to normalize unknown samples to the training set distribution, effectively mitigating the overfitting issue. Consequently, our approach significantly increases the amount of automatic region-of-interest ground truth labeling on high-resolution whole-slide images using active deep learning. We accept 92% of the automatic labels generated for our unlabeled data cohort, expanding the labeled dataset by 845%. Additionally, we demonstrate expert time savings of 96% relative to manual expert ground-truth labeling.
- Published
- 2024
- Full Text
- View/download PDF
37. Comparison of ASI-PRISMA Data, DLR-EnMAP Data, and Field Spectrometer Measurements on 'Sale ‘e Porcus', a Salty Pond (Sardinia, Italy)
- Author
-
Massimo Musacchio, Malvina Silvestri, Vito Romaniello, Marco Casu, Maria Fabrizia Buongiorno, and Maria Teresa Melis
- Subjects
ASI-PRISMA ,DLR-EnMAP ,ground truth ,surface reflectance ,Science - Abstract
A comparison between the ASI-PRISMA (Agenzia Spaziale Italiana-PRecursore IperSpettrale della Missione Applicativa) DLR-EnMAP (German Aerospace Center—Environmental Mapping and Analysis Program) data and field spectrometer measurements has been performed. The test site, located at the “Sale ‘e Porcus” pond (hereafter SPp) in Western Sardinia, Italy, offers particularly homogenous characteristics, making it an ideal location not only for experimentation but also for calibration purposes. Three remote-sensed data acquisitions have been performed by these agencies (ASI and DLR) starting on 14 July 2023 and continuing until 22 July 2023. The DLR-EnMAP data acquired on 22 July overestimates both that of the ASI-PRISMA and the 14 July DLR-EnMAP radiance in the VNIR region, while all the datasets are close to each other, up to 2500 nm, for all considered days. The average absolute mean difference between the reflectance values estimated by the ASI-PRISMA and DLR-EnMAP, in the test area, is around 0.015, despite the small difference in their time of acquisition (8 days); their maximum relative difference value occurs at about 2100 nm. In this study, we investigate the relationship between the averaged ground truth value of reflectance, acquired by means of a portable ASD FieldSpec spectoradiometer, characterizing the test site and the EO reflectance data derived from the official datasets. FieldSpec measurements confirm the quality of both the ASI-PRISMA and DLR-EnMAP’s reflectance estimations.
- Published
- 2024
- Full Text
- View/download PDF
38. Deriving Ground Truth Labels for Regression Problems Using Annotator Precision †.
- Author
-
Johnston, Benjamin and de Chazal, Philip
- Subjects
MACHINE learning ,CROWDSOURCING - Abstract
When training machine learning models with practical applications, a quality ground truth dataset is critical. Unlike in classification problems, there is currently no effective method for determining a single ground truth value or landmark from a set of annotations in regression problems. We propose a novel method for deriving ground truth labels in regression problems that considers the performance and precision of individual annotators when identifying each label separately. In contrast to the commonly accepted method of computing the global mean, our method does not assume each annotator to be equally capable of completing the specified task, but rather ensures that higher-performing annotators have a greater contribution to the final result. The ground truth selection method described within this paper provides a means of improving the quality of input data for machine learning model development by removing lower-quality labels. In this study, we objectively demonstrate the improved performance by applying the method to a simulated dataset where a canonical ground truth position can be known, as well as to a sample of data collected from crowd-sourced labels. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
39. Seeing beyond the spikes: reconstructing the complete spatiotemporal membrane potential distribution from paired intra‐ and extracellular recordings.
- Author
-
Meszéna, Domokos, Barlay, Anna, Boldog, Péter, Furuglyás, Kristóf, Cserpán, Dorottya, Wittner, Lucia, Ulbert, István, and Somogyvári, Zoltán
- Subjects
- *
MEMBRANE potential , *ACTION potentials , *PYRAMIDAL neurons , *ELECTRIC potential , *DENSITY currents , *SPATIAL resolution - Abstract
Although electrophysiologists have been recording intracellular neural activity routinely ever since the ground‐breaking work of Hodgkin and Huxley, and extracellular multichannel electrodes have also been used frequently and extensively, a practical experimental method to track changes in membrane potential along a complete single neuron is still lacking. Instead of obtaining multiple intracellular measurements on the same neuron, we propose an alternative method by combining single‐channel somatic patch‐clamp and multichannel extracellular potential recordings. In this work, we show that it is possible to reconstruct the complete spatiotemporal distribution of the membrane potential of a single neuron with the spatial resolution of an extracellular probe during action potential generation. Moreover, the reconstruction of the membrane potential allows us to distinguish between the two major but previously hidden components of the current source density (CSD) distribution: the resistive and the capacitive currents. This distinction provides a clue to the clear interpretation of the CSD analysis, because the resistive component corresponds to transmembrane ionic currents (all the synaptic, voltage‐sensitive and passive currents), whereas capacitive currents are considered to be the main contributors of counter‐currents. We validate our model‐based reconstruction approach on simulations and demonstrate its application to experimental data obtained in vitro via paired extracellular and intracellular recordings from a single pyramidal cell of the rat hippocampus. In perspective, the estimation of the spatial distribution of resistive membrane currents makes it possible to distiguish between active and passive sinks and sources of the CSD map and the localization of the synaptic input currents, which make the neuron fire. Key points: A new computational method is introduced to calculate the unbiased current source density distribution on a single neuron with known morphology.The relationship between extracellular and intracellular electric potential is determined via mathematical formalism, and a new reconstruction method is applied to reveal the full spatiotemporal distribution of the membrane potential and the resistive and capacitive current components.The new reconstruction method was validated on simulations.Simultaneous and colocalized whole‐cell patch‐clamp and multichannel silicon probe recordings were performed from the same pyramidal neuron in the rat hippocampal CA1 region, in vitro.The method was applied in experimental measurements and returned precise and distinctive characteristics of various intracellular phenomena, such as action potential generation, signal back‐propagation and the initial dendritic depolarization preceding the somatic action potential. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
40. Truth Discovery from Conflicting Data: A Survey.
- Author
-
FANG Xiu, WANG Kang, SUN Guohao, SI Suxin, and LYU Hang
- Subjects
INTERNET ,ARTIFICIAL intelligence ,IMAGE storage & retrieval systems ,AUTOMATION ,INFORMATION retrieval - Abstract
With the rocketing progress of the Internet, it is easier for people to get information about the objects that they are interested in. However, this information usually has conflicts. In order to resolve conflicts and get the true information, truth discovery has been proposed and received widespread attention. Many algorithms have been proposed to adapt to different scenarios. This paper aims to investigate these algorithms and summarize them from the perspective of algorithm models and specific concepts. Some classic datasets and evaluation metrics are given in this paper. Some future directions for readers are also provided to better understand the field of truth discovery. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
41. Reliability of whole mount radical prostatectomy histopathology as the ground truth for artificial intelligence assisted prostate imaging.
- Author
-
Jager, Auke, Postema, Arnoud W., van der Linden, Hans, Nooijen, Peet T.G.A., Bekers, Elise, Kweldam, Charlotte F., Daures, Gautier, Zwart, Wim, Mischi, M., Beerlage, Harrie P., and Oddens, Jorg R.
- Abstract
The development of artificial intelligence–based imaging techniques for prostate cancer (PCa) detection and diagnosis requires a reliable ground truth, which is generally based on histopathology from radical prostatectomy specimens. This study proposes a comprehensive protocol for the annotation of prostatectomy pathology slides. To evaluate the reliability of the protocol, interobserver variability was assessed between five pathologists, who annotated ten radical prostatectomy specimens consisting of 74 whole mount pathology slides. Interobserver variability was assessed for both the localization and grading of PCa. The results indicate excellent overall agreement on the localization of PCa (Gleason pattern ≥ 3) and clinically significant PCa (Gleason pattern ≥ 4), with Dice similarity coefficients (DSC) of 0.91 and 0.88, respectively. On a per-slide level, agreement for primary and secondary Gleason pattern was almost perfect and substantial, with Fleiss Kappa of.819 (95% CI.659–.980) and.726 (95% CI.573–.878), respectively. Agreement on International Society of Urological Pathology Grade Group was evaluated for the index lesions and showed agreement in 70% of cases, with a mean DSC of 0.92 for all index lesions. These findings show that a standardized protocol for prostatectomy pathology annotation provides reliable data on PCa localization and grading, with relatively high levels of interobserver agreement. More complicated tissue characterization, such as the presence of cribriform growth and intraductal carcinoma, remains a source of interobserver variability and should be treated with care when used in ground truth datasets. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
42. ViF-GTAD: A new automotive dataset with ground truth for ADAS/AD development, testing, and validation.
- Author
-
Haas, Sarah, Solmaz, Selim, Reckenzaun, Jakob, and Genser, Simon
- Subjects
- *
GEOGRAPHICAL perception , *METADATA , *DRIVER assistance systems , *TRAFFIC safety , *SYSTEMS development - Abstract
A new dataset for automated driving, which is the subject matter of this paper, identifies and addresses a gap in existing similar perception datasets. While most state-of-the-art perception datasets primarily focus on the provision of various onboard sensor measurements along with the semantic information under various driving conditions, the provided information is often insufficient since the object list and position data provided include unknown and time-varying errors. The current paper and the associated dataset describe the first publicly available perception measurement data that include not only the onboard sensor information from the camera, Lidar, and radar with semantically classified objects but also the high-precision ground-truth position measurements enabled by the accurate RTK-assisted GPS localization systems available on both the ego vehicle and the dynamic target objects. This paper provides insight on the capturing of the data, explicitly explaining the metadata structure and the content, as well as the potential application examples where it has been, and can potentially be, applied and implemented in relation to automated driving and environmental perception systems development, testing, and validation. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
43. A New Region-Based Minimal Path Selection Algorithm for Crack Detection and Ground Truth Labeling Exploiting Gabor Filters.
- Author
-
de León, Gonzalo, Fiorentini, Nicholas, Leandri, Pietro, and Losa, Massimo
- Subjects
- *
GABOR filters , *ROCK deformation , *ALGORITHMS , *DRUG labeling , *DEEP learning , *JUDGMENT (Psychology) , *IMAGE segmentation - Abstract
Cracks are fractures or breaks that occur in materials such as concrete, metals, rocks, and other solids. Various methods are used to detect and monitor cracks; among many of them, image-based methodologies allow fast identification of the distress and easy quantification of the percentage of cracks in the scene. Two main categories can be identified: classical and deep learning approaches. In the last decade, the tendency has moved towards the use of the latter. Even though they have proven their outstanding predicting performance, they suffer some drawbacks: a "black-box" nature leaves the user blind and without the possibility of modifying any parameters, a huge amount of labeled data is generally needed, a process that requires expert judgment is always required, and, finally, they tend to be time-consuming. Accordingly, the present study details the methodology for a new algorithm for crack segmentation based on the theory of minimal path selection combined with a region-based approach obtained through the segmentation of texture features extracted using Gabor filters. A pre-processing step is described, enabling the equalization of brightness and shadows, which results in better detection of local minima. These local minimal are constrained by a minimum distance between adjacent points, enabling a better coverage of the cracks. Afterward, a region-based segmentation technique is introduced to determine two areas that are used to determine threshold values used for rejection. This step is critical to generalize the algorithm to images presenting close-up scenes or wide cracks. Finally, a geometrical thresholding step is presented, allowing the exclusion of rounded areas and small isolated cracks. The results showed a very competitive F1-score (0.839), close to state-of-the-art values achieved with deep learning techniques. The main advantage of this approach is the transparency of the workflow, contrary to what happens with deep learning frameworks. In the proposed approach, no prior information is required; however, the statistical parameters may have to be adjusted to the particular case and requirements of the situation. The proposed algorithm results in a useful tool for researchers and practitioners needing to validate their results against some reference or needing labeled data for their models. Moreover, the current study could establish the grounds to standardize the procedure for crack segmentation with a lower human bias and faster results. The direct application of the methodology to images obtained with any low-cost sensor makes the proposed algorithm an operational support tool for authorities needing crack detection systems in order to monitor and evaluate the current state of the infrastructures, such as roads, tunnels, or bridges. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
44. RiceMapEngine: A Google Earth Engine-Based Web Application for Fast Paddy Rice Mapping
- Author
-
Zhiqi Yu, Liping Di, Sravan Shrestha, Chen Zhang, Liying Guo, Faisal Qamar, and Timothy J. Mayer
- Subjects
Google Earth engine (GEE) ,ground truth ,land use land cover ,rice mapping ,Ocean engineering ,TC1501-1800 ,Geophysics. Cosmic physics ,QC801-809 - Abstract
Mapping rice area is a critical resource planning task in many South Asia countries where rice is the primary crop. Remote-sensing-based methods typically rely on domain knowledge, such as crop calendar and crop phenology, and supervised classification with ground truth samples. Applying such methods on Google Earth engine (GEE) has been proven effective, especially at large scale owing to the comprehensive and up-to-date data catalog and massive server-side processing power. However, writing scripts through the code editor requires users to program in JavaScript and understand GEE application programming interface (API), which can be challenging for many researchers. Thus, this article presents a GEE-based web application that aims to eliminate the programming requirements for data selection, preprocessing, and visualizations so that users can easily produce rice maps and refine ground truth collections through intuitive graphical user interfaces (GUI). This software includes three submodule apps, namely the ground truth collection app, threshold-based rice mapping app, and classification-based rice mapping app. Users can customize data processing flow using GUI designed with Bootstrap, and the backend server uses GEE Python API, and a Google service account for authentication to execute the workflow on Google cloud servers. The experiment shows that both the overall accuracy and Kappa scores of the mapping result are higher than 0.9, which suggests that RiceMapEngine can significantly reduce the complexity and time costs it takes to produce the accurate rice area maps and meet the demands of real-world stakeholders.
- Published
- 2023
- Full Text
- View/download PDF
45. Ant Colony Optimization with BrainSeg3D Protocol for Multiple Sclerosis Lesion Detection
- Author
-
Bouzidi, Dalenda, Ghozzi, Fahmi, Fakhfakh, Ahmed, Goos, Gerhard, Founding Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Aloulou, Hamdi, editor, Abdulrazak, Bessam, editor, de Marassé-Enouf, Antoine, editor, and Mokhtari, Mounir, editor
- Published
- 2022
- Full Text
- View/download PDF
46. Artificially Intelligent Solutions: Detection, Debunking, and Fact-Checking
- Author
-
Rubin, Victoria L. and Rubin, Victoria L.
- Published
- 2022
- Full Text
- View/download PDF
47. Geospatial Applications in Inventory of Horticulture Plantations
- Author
-
Ravishankar, H. M., Trivedi, Shivam, Subramoniam, S. Rama, Ahamed, J. Mohammed, Nagashree, T. R., Manjula, V. B., Hebbar, R., Jha, C. S., Dadhwal, V. K., Singh, V. P., Editor-in-Chief, Berndtsson, R., Editorial Board Member, Rodrigues, L. N., Editorial Board Member, Sarma, Arup Kumar, Editorial Board Member, Sherif, M. M., Editorial Board Member, Sivakumar, B., Editorial Board Member, Zhang, Q., Editorial Board Member, Jha, Chandra Shekhar, editor, Pandey, Ashish, editor, Chowdary, V.M., editor, and Singh, Vijay, editor
- Published
- 2022
- Full Text
- View/download PDF
48. Visualization Approach to Presentation of New Referral Dataset for Maritime Zone Video Surveillance in Various Weather Conditions
- Author
-
Vujović, Igor, Petković, Miro, Kuzmanić, Ivica, Šoda, Joško, Öchsner, Andreas, Series Editor, da Silva, Lucas F. M., Series Editor, and Altenbach, Holm, Series Editor
- Published
- 2022
- Full Text
- View/download PDF
49. Data Preparation for Artificial Intelligence
- Author
-
de Araujo, Aline L., Hardell, Cailin, Koszek, Wojciech A., Wu, Jie, Willemink, Martin J., Schoepf, U. Joseph, Series Editor, De Cecco, Carlo N., editor, van Assen, Marly, editor, and Leiner, Tim, editor
- Published
- 2022
- Full Text
- View/download PDF
50. Development and validation of deep learning-based automatic brain segmentation for East Asians: a comparison with Freesurfer.
- Author
-
Chung-Man Moon, Yun Young Lee, Ki-Eun Hyeong, Woong Yoon, Byung Hyun Baek, Suk-Hee Heo, Sang-Soo Shin, and Seul Kee Kim
- Subjects
EAST Asians ,INTRACLASS correlation ,PEARSON correlation (Statistics) ,CONVOLUTIONAL neural networks ,MAGNETIC resonance imaging - Abstract
Purpose: To develop and validate deep learning-based automatic brain segmentation for East Asians with comparison to data for healthy controls from Freesurfer based on a ground truth. Methods: A total of 30 healthy participants were enrolled and underwent T1- weighted magnetic resonance imaging (MRI) using a 3-tesla MRI system. Our Neuro I software was developed based on a three-dimensional convolutional neural networks (CNNs)-based, deep-learning algorithm, which was trained using data for 776 healthy Koreans with normal cognition. Dice coefficient (D) was calculated for each brain segment and compared with control data by paired t-test. The inter-method reliability was assessed by intraclass correlation coefficient (ICC) and effect size. Pearson correlation analysis was applied to assess the relationship between D values for each method and participant ages. Results: The D values obtained from Freesurfer (ver6.0) were significantly lower than those from Neuro I. The histogram of the Freesurfer results showed remarkable differences in the distribution of D values from Neuro I. Overall, D values obtained by Freesurfer and Neuro I showed positive correlations, but the slopes and intercepts were significantly different. It was showed the largest effect sizes ranged 1.07–3.22, and ICC also showed significantly poor to moderate correlations between the two methods (0.498 ≤ ICC ≤ 0.688). For Neuro I, D values resulted in reduced residuals when fitting data to a line of best fit, and indicated consistent values corresponding to each age, even in young and older adults. Conclusion: Freesurfer and Neuro I were not equivalent when compared to a ground truth, where Neuro I exhibited higher performance. We suggest that Neuro I is a useful alternative for the assessment of the brain volume. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.