113,557 results on '"Electronic data processing"'
Search Results
2. Enhancing ferroelectric characterization at nanoscale: A comprehensive approach for data processing in spectroscopic piezoresponse force microscopy.
- Author
-
Valloire, H., Quéméré, P., Vaxelaire, N., Kuentz, H., Le Rhun, G., and Borowik, Ł.
- Subjects
- *
PIEZORESPONSE force microscopy , *NANOMECHANICS , *ELECTRONIC data processing , *FERROELECTRIC thin films , *POTASSIUM niobate , *MACHINE learning , *HYSTERESIS loop - Abstract
Switching Spectroscopy Piezoresponse Force Microscopy (SSPFM) stands out as a powerful method for probing ferroelectric properties within materials subjected to incremental polarization induced by an external electric field. However, the dense data processing linked to this technique is a critical factor influencing the quality of obtained results. Furthermore, meticulous exploration of various artifacts, such as electrostatics, which may considerably influence the signal, is a key factor in obtaining quantitative results. In this paper, we present a global methodology for SSPFM data processing, accessible in open-source with a user-friendly Python application called PySSPFM. A ferroelectric thin film sample of potassium sodium niobate has been probed to illustrate the different aspects of our methodology. Our approach enables the reconstruction of hysteresis nano-loops by determining the PR as a function of applied electric field. These hysteresis loops are then fitted to extract characteristic parameters that serve as measures of the ferroelectric properties of the sample. Various artifact decorrelation methods are employed to enhance measurement accuracy, and additional material properties can be assessed. Performing this procedure on a grid of points across the surface of the sample enables the creation of spatial maps. Furthermore, different techniques have been proposed to facilitate post-treatment analysis, incorporating algorithms for machine learning (K-means), phase separation, and mapping cross correlation, among others. Additionally, PySSPFM enables a more in-depth investigation of the material by studying the nanomechanical properties during poling, through the measurement of the resonance properties of the cantilever–tip–sample surface system. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. Unmanned Aerial Vehicle (UAV) path planning and control assisted by Augmented Reality (AR): the case of indoor drones.
- Author
-
Mourtzis, Dimitris, Angelopoulos, John, and Panopoulos, Nikos
- Subjects
DRONE aircraft ,AUGMENTED reality ,ENGINEERING design ,INDUSTRY 4.0 ,ELECTRONIC data processing - Abstract
Following the recent advances in Industry 4.0 and the upcoming Industry 5.0, the use of multiple UAVs for indoor tasks has risen, particularly in real-time remote monitoring, wireless coverage, and remote sensing. As a result, UAVs can be viewed as proactive problem solvers and can support Internet of Things (IoT) platforms by collecting and monitoring data cost-effectively and efficiently, leading to better decision-making. Moreover, sophisticated drone operations require specialised software and data processing abilities. However, the utilisation of drones has been mainly focused on outdoor environments, thus creating a literature gap regarding indoor navigation and operation. Therefore, the design and development of a method for remote planning and control of drones based on the utilisation of AR is presented in this paper. The proposed method is based on the utilisation of drones for remote monitoring. The suggested approach involves engineers designing a sequence of actions and transmitting them wirelessly to the drone, eliminating the need for human intervention. Thus, the proposed method contributes towards enabling engineers visualise the drone path with the use of Augmented Reality and provides the flexibility of adding multiple way points. The applicability of the developed framework is tested in a laboratory-based machine shop. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. Deep learning-based data processing method for transient thermoreflectance measurements.
- Author
-
Mao, Yali, Zhou, Shaojie, Tang, Weiyuan, Wu, Mei, Zhang, Haochen, Sun, Haiding, and Yuan, Chao
- Subjects
- *
DEEP learning , *ELECTRONIC data processing , *THERMOPHYSICAL properties , *GLOBAL optimization - Abstract
Pump–probe thermoreflectance has been commonly applied for characterizing the thermal properties of materials. Generally, a reliable and efficient non-linear fitting process is often implemented to extract unknown thermal parameters during the pump–probe thermoreflectance characterizations. However, when it comes to processing large amounts of data acquired from similar structural samples, non-linear fitting process appears to be very time-consuming and labor-intensive to search for the best fitting for every testing curve. Herein, we propose to apply deep learning (DL) approach to nanosecond transient thermoreflectance technique for high-throughput experimental data processing. We first investigated the effect of training set parameters (density and bounds) on the predictive performance of the DL model, providing a guidance to optimize the DL model. Then, the DL model is further verified in the measurement of the bulk sapphire, SiC, diamond samples, and GaN-based multilayer structures, demonstrating its capability of analyzing the results with high accuracy. Compared to the conventional non-linear fitting method (such as Global Optimization), the computation time of the new model is 1000 times lower. Such a data-driven DL model enables the faster inference and stronger fitting capabilities and is particularly efficient and effective in processing data acquired from wafer-level measurements with similar material structures. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. Empirical issues concerning studies of firm entry.
- Author
-
Coad, Alex, Kato, Masatoshi, and Srhoj, Stjepan
- Subjects
VALUE (Economics) ,ENDOWMENTS ,ELECTRONIC data processing - Abstract
We discuss that entry can be considered from various levels of analysis: entrepreneur-level, firm-level, and also at higher levels of aggregation, such as the industry-level and country-level. We also formulate a list of six challenges for econometric studies of firm entry, highlighting the data sources, typical empirical setups, potential sources of bias, and appropriate econometric techniques. While progress can be made with sophisticated econometric estimators, a pressing need for entry studies concerns detailed data on the gestation process, entry modes, and the value of resource endowments and knowledge endowments. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
6. Co-Developing Programs and Their Proof of Correctness.
- Author
-
Chapman, Roderick, Dross, Claire, Matthews, Stuart, and Moy, Yannick
- Subjects
- *
SPARK (Computer program language) , *PROGRAMMING languages , *ELECTRONIC data processing , *COMPUTER programming , *COMPUTER software development , *COMPUTER software correctness - Abstract
The article focuses on the auto-active approach for co-developing programs and their proof of correctness, specifically the open source SPARK technology. The authors discuss the key design and technological choices for SPARK, which made it successful within the industry, and explore the possible future of SPARK and other analyzers of the same family.
- Published
- 2024
- Full Text
- View/download PDF
7. Trends in Data Curation: Competencies and Tools.
- Author
-
Murillo, Angela P. and Yoon, Ayoung
- Subjects
- *
DATA curation , *INFORMATION science , *LIBRARY science , *TEACHING aids , *ELECTRONIC data processing - Abstract
Library and information science has led data curation research and education for the last two decades, providing data curation education, professional development programs, and a robust professional opportunity. To keep current with the latest trends in competencies and tools needed to conduct data curation activities, this research conducted a systematic literature review of literature that captures competencies and tools to build a framework of current trends in data curation work that educators can utilize to ensure up‐to‐date educational materials for the next generation of data curation professionals. This poster presents the preliminary findings of this data curation competencies and tools analysis. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
8. Thematic Trends in Data Curation Literature.
- Author
-
Murillo, Angela P. and Yoon, Ayoung
- Subjects
- *
DATA curation , *TECHNOLOGICAL innovations , *BIBLIOMETRICS , *DATA analysis , *ELECTRONIC data processing - Abstract
The field of data curation is rapidly changing due to new developments in technologies and techniques for conducting data work. As the field of data curation evolves, researchers, practitioners, and educators need to be able to respond to these developments. One way to understand trends in a field is by examining published literature. This study first gathered data curation literature through a modified systematic literature review with the framing question, 'What competencies, skill sets, and proficiencies are needed to conduct data curation activities?'. These literatures were then analyzed using bibliometric analysis, visual analysis of the citation data, and topic modeling to understand trends in the data curation field. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
9. Investigating Co‐Authorship Networks of Academic and Industry Researchers in Artificial Intelligence.
- Author
-
Liang, Lizhen
- Subjects
- *
ARTIFICIAL intelligence , *AUTHORSHIP collaboration , *ELECTRONIC data processing , *SOCIAL capital , *EIGENVECTORS , *SOCIAL network analysis - Abstract
Research teams from the industry, especially big technology companies, have been pushing impactful research work in the field of artificial intelligence (AI), changing the prospects of the field and the careers of many researchers. Research teams from big technology companies usually possess more data, bigger computing infrastructure, and research talent, granting them the advantages in advancing AI research. While most previous work focuses on investigating the advantages the industry has in the field of AI, and how their research publication is different from those published by academic teams, few research has been done to investigate whether working as an industry researcher is beneficial at the individual level. In this work, by analyzing co‐authorship networks of researchers published in AI conferences, we investigate whether working in the industry gives researchers advantages in "intangible" forms, such as social capital, represented by the collaborative relationships they gained or maintained. Our result shows that the many advantages industry researchers possess correlate with the social capital they have, measured by degree centrality, eigenvector centrality, betweenness centrality, and effective size. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
10. Discursive AI Infrastructures: Envisioned and Overlooked Museum Futures.
- Author
-
Kist, Cassandra
- Subjects
- *
ARTIFICIAL intelligence , *MUSEUMS , *CRITICAL thinking , *ELECTRONIC data processing , *DISCOURSE - Abstract
Prompted by recent innovations, artificial intelligence (AI) is increasingly being discussed across the museum sector regarding its implications for institutional roles and practices. However, AI in particular, is an ambiguous term, a 'black box' which is capable of containing and reflecting numerous values and ideals (Crawford, 2021). This paper positions discourse around AI as a 'discursive' infrastructure, capable of not only embodying ideals but also shaping and justifying certain institutional practices and roles. This paper thematically analyses 115 pieces of grey literature produced and shared by professional governance bodies in the museum sector from 1995–2023, mainly across Canada, the United Kingdom, and the United States. In doing so, it identifies four preliminary themes encompassing shifts in discourse over time which give shape to a contemporary discursive infrastructure. This prompts timely critical reflections of museum professionals and stakeholders on both imagined and overlooked public roles, responsibilities, and practices of the museum in relation to AI. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
11. Exploratory Search in Digital Humanities: A Study of Visual Keyword/Result Linking.
- Author
-
Hoeber, Orland, Harvey, Morgan, Momeni, Milad, Pirmoradi, Abbas, and Gleeson, David
- Subjects
- *
DIGITAL humanities , *DIGITAL libraries , *INFORMATION retrieval , *ACADEMIC libraries , *ELECTRONIC data processing - Abstract
While searching within digital humanities collections is an important aspect of digital humanities research, the search features provided are usually more suited to lookup search than exploratory search. This limits the ability of digital humanities scholars to undertake complex search scenarios. Drawing upon recent studies on supporting exploratory search in academic digital libraries, we implemented two visual keyword/result linking approaches for searching within the Europeana collection; one that keeps the keywords linked to the search results and another that aggregates the keywords over the search result set. Using a controlled laboratory study, we assessed these approaches in comparison to the existing Europeana search mechanisms. We found that both visual keyword/result linking approaches were improvements over the baseline, with some differences between the new approaches that were dependent on the stage of the exploratory search process. This work illustrates the value of providing advanced search functionality within digital humanities collections to support exploratory search processes, and the need for further design and study of digital humanities search tools that support complex search scenarios. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
12. Emulsion stability and level of preference for chicken nuggets of broiler meat stored at cold temperatures with the application of Bay leaves (Syzygium polyanthum).
- Author
-
Sari, A., Hatta, W., and Maruddin, F.
- Subjects
- *
CHICKEN as food , *COLD (Temperature) , *MEAT quality , *BROILER chickens , *ELECTRONIC data processing , *COLD storage - Abstract
Bay leaves are plants as natural preservatives, and cooking spices have a distinctive aroma and taste. Bay leaves are very beneficial for health and food. However, meat is prone to decay and damage due to relatively low shelf life and improper storage. The research objective was to examine the effect of soaking broiler meat in bay leaf solution stored in the refrigerator on the pH value of the meat, emulsion stability, and the preferences of the resulting nuggets. Broiler meat is soaked in 3 concentrations of bay leaves, namely 0, 15, and 30%, and stored in the refrigerator. Measurements of meat quality and nugget production were carried out on days 0, 2, and 4 of storage. This study used a 3x3 factorial, completely randomized design. ANOVA data processing method and Duncan's advanced test. The results showed that soaking the meat in bay leaf solution lowered the pH value of the meat (P<0.01). The lowest pH value at a concentration of 15% compared without immersion. Storage time does not affect the pH value. Soaking meat in bay leaf solution at a concentration of 15% decreased the stability of the emulsion (P<0.01) and increased with the length of time the meat was stored (P<0.01). The level of preference for nuggets of 15% bay leaf solution was higher, but lower at a concentration of 30% (P<0.01). Concentration 15% bay leaf solution increased the preference of the nuggets, while storing the meat for up to 4 days increased the stability of the emulsion. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
13. Navigating the convergence: A comprehensive review of synergies and interplay between Mobile Cloud Computing, Edge Computing, and Fog Computing.
- Author
-
Dhawan, Sunil
- Subjects
- *
REAL-time computing , *EDGE computing , *PROCESS capability , *ELECTRONIC data processing , *CLOUD computing - Abstract
In today's emerging world of technology, the combination of Mobile Cloud Computing (MCC) Edge Computing (EC) and Fog Computing (FC) is pouring innovation and change. These interrelated technologies provide a mechanism for addressing the challenges brought by the growing use of devices, apps and services. Mobile Cloud Computing serves as a link between cloud systems and mobile devices overpowering the limitations these devices face. Edge Computing takes computational resources closer to users for real time data processing while Fog Computing seamlessly blends localized processing with cloud capabilities. This study probes into how MCC, EC and FC work discovering where they interconnect and enhance each other. It discusses how these technologies improve resource management, streamline data processing and boost service delivery to enrich user experiences and enable applications. Acknowledging the importance of MCC, EC and FC are crucial for researchers and professionals to influence their combined strength in expanding computing efficiency and innovation while addressing issues like data security in Fog Computing. The paper will be structured in specific sections focusing on the principles, architecture, applications, challenges and future directions of MCC, EC and FC to provide a comprehensive understanding of their interactions. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
14. Vigilance 360: Realtime surveillance system.
- Author
-
Raut, Vedant, Yashwante, Avishkar, and Mundada, G. S.
- Subjects
- *
SECURITY systems , *COMPUTER vision , *COLLEGE environment , *ELECTRONIC data processing , *AUTOMATION - Abstract
The "Vigilance 360" system represents a pivotal advancement in security measures for both educational institutions and diverse organizations. This survey paper examines the impact of implementing this system, which prioritizes the automation of real-time event detection at entrance gates to preemptively mitigate potential threats. By doing so, it establishes a fortified and secure campus environment, ensuring the safety of all stakeholders involved. The system leverages state-of-the-art computer vision techniques and deep learning models to attain a remarkable level of accuracy and precision in threat identification. Moreover, it excels in rapid data processing, enabling seamless automation of real-time performance. This paper highlights the significance of Vigilance 360 in revolutionizing security practices, emphasizing its potential to significantly reduce response times and enhance overall security readiness. The integration of cutting-edge technologies demonstrates a forward-thinking approach in safeguarding educational and organizational spaces, underscoring the potential for widespread adoption in similar contexts. It provides a thorough understanding of its capabilities, showcasing its effectiveness as an automized proactive security solution. The findings presented herein underscore the critical importance of embracing such advanced systems in fortifying security measures, ultimately creating safer environments for all stakeholders involved. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
15. A Ramanujan subspace and dynamic time warping and adaptive singular value decomposition combined denoising method for low signal‐to‐noise ratio surface microseismic monitoring data in hydraulic fracturing.
- Author
-
Wang, Xu‐Lin, Zhang, Jian‐Zhong, and Huang, Zhong‐Lai
- Subjects
- *
HILBERT-Huang transform , *HYDRAULIC fracturing , *BANDPASS filters , *ELECTRONIC data processing , *DATA quality , *RANDOM noise theory , *SINGULAR value decomposition - Abstract
Surface microseismic monitoring is widely used in hydraulic fracturing. Real‐time monitoring data collected during fracturing can be used to perform surface‐microseismic localization, which aids in assessing the effects of fracturing and provides guidance for the process. The accuracy of localization critically depends on the quality of monitoring data. However, the signal‐to‐noise ratio of the data is often low due to strong coherent and random noise, making denoising essential for processing surface monitoring data. To suppress noise more effectively, this paper introduces a novel denoising method that integrates the Ramanujan subspace with dynamic time warping and adaptive singular value decomposition. The new method consists of two steps: First, a Ramanujan subspace is constructed to suppress periodic noise. Then, dynamic time warping and adaptive singular value decomposition are applied to eliminate remaining coherent and random noise. The method has been evaluated using both synthetic and field data, and its performance is compared with traditional microseismic denoising techniques, including bandpass filtering and empirical mode decomposition. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
16. Design and development of STCF offline software.
- Author
-
Ai, Xiaocong, Huang, Xingtao, Li, Teng, Qi, Binbin, and Qin, Xiaoshuai
- Subjects
- *
ELECTRONIC data processing , *COLLIDERS (Nuclear physics) , *DIGITIZATION , *DATA analysis , *LUMINOSITY - Abstract
The Super Tau-Charm Facility (STCF) is a next-generation positron–electron collider experiment, designed to study various physics topics in the tau-charm energy region. The designed peak luminosity of STCF is 0.5×1035cm−2s−1, which is almost two orders of magnitude higher than the current tau-charm factory, BESIII. To implement the offline data processing tasks, as well as tackling the great challenge posed by the huge data volume, the offline software of STCF (OSCAR) is designed to implement the detector simulation, digitization, calibration, reconstruction as well as to provide a common platform for data analysis. This paper presents the status and progress of the OSCAR system, including the design, implementation as well as the preliminary performance of the core software, detector simulation, digitization and background mixing and reconstruction algorithms for all sub-detectors. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
17. An optimized method for dose–effect prediction of traditional Chinese medicine based on 1D-ResCNN-PLS.
- Author
-
Xiong, Wangping, Pan, Jiasong, Liu, Zhaoyang, Du, Jianqiang, Zhu, Yimin, Luo, Jigen, Yang, Ming, and Zhou, Xian
- Subjects
- *
CONVOLUTIONAL neural networks , *CHINESE medicine , *LEAST squares , *ELECTRONIC data processing - Abstract
AbstractWe introduce a one-dimensional (1D) residual convolutional neural network with Partial Least Squares (1D-ResCNN-PLS) to solve the covariance and nonlinearity problems in traditional Chinese medicine dose–effect relationship data. The model combines a 1D convolutional layer with a residual block to extract nonlinear features and employs PLS for prediction. Tested on the Ma Xing Shi Gan Decoction datasets, the model significantly outperformed conventional models, achieving high accuracies, sensitivities, specificities, and AUC values, with considerable reductions in mean square error. Our results confirm its effectiveness in nonlinear data processing and demonstrate potential for broader application across public datasets. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
18. Blockchain for Edge Computing in Smart Environments: Use Cases, Issues, and Challenges.
- Author
-
Prabadevi, B., Deepa, N., Sudhagara Rajan, S., and Srivastava, Gautam
- Subjects
- *
DIGITAL transformation , *ELECTRONIC data processing , *EDGE computing , *CENOZOIC Era , *DIGITAL technology , *BLOCKCHAINS - Abstract
The Cenozoic era is the digital age where people, things, and any device with network capabilities can communicate with each other, and the Internet of Things (IoT) paves the way for it. Almost all domains are adopting IoT from smart home appliances, smart healthcare, smart transportation, Industrial IoT and many others. As the adoption of IoT increases, the accretion of data also grows. Furthermore, digital transformations have led to more security vulnerabilities, resulting in data breaches and cyber-attacks. One of the most prominent issues in smart environments is a delay in data processing while all IoT smart environments store their data in the cloud and retrieve them for every transaction. With increased data accumulations on the cloud, most smart applications face unprecedented delays. Thus, data security and low-latency response time are mandatory to deploy a robust IoT-based smart environment. Blockchain, a decentralized and immutable distributed ledger technology, is an essential candidate for ensuring secured data transactions, but it has a variety of challenges in accommodating resource-constrained IoT devices. Edge computing brings data storage and computation closer to the network’s edge and can be integrated with blockchain for low-latency response time in data processing. Integrating blockchain with edge computing will ensure faster and more secure data transactions, thus reducing the computational and communicational overhead concerning resource allocation, data transaction and decision-making. This paper discusses the seamless integration of blockchain and edge computing in IoT environments, various use cases, notable blockchain-enabled edge computing architectures in the literature, secured data transaction frameworks, opportunities, research challenges, and future directions. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
19. Enhancing the data processing speed of a deep-learning-based three-dimensional single molecule localization algorithm (FD-DeepLoc) with a combination of feature compression and pipeline programming.
- Author
-
Guo, Shuhao, Lin, Jiaxun, Zhang, Yingjun, and Huang, Zhen-Li
- Subjects
- *
REAL-time computing , *IMAGE processing , *ONLINE algorithms , *ELECTRONIC data processing , *SINGLE molecules , *DEEP learning , *SADDLEPOINT approximations - Abstract
Three-dimensional (3D) single molecule localization microscopy (SMLM) plays an important role in biomedical applications, but its data processing is very complicated. Deep learning is a potential tool to solve this problem. As the state of art 3D super-resolution localization algorithm based on deep learning, FD-DeepLoc algorithm reported recently still has a gap with the expected goal of online image processing, even though it has greatly improved the data processing throughput. In this paper, a new algorithm Lite-FD-DeepLoc is developed on the basis of FD-DeepLoc algorithm to meet the online image processing requirements of 3D SMLM. This new algorithm uses the feature compression method to reduce the parameters of the model, and combines it with pipeline programming to accelerate the inference process of the deep learning model. The simulated data processing results show that the image processing speed of Lite-FD-DeepLoc is about twice as fast as that of FD-DeepLoc with a slight decrease in localization accuracy, which can realize real-time processing of 256×256 pixels size images. The results of biological experimental data processing imply that Lite-FD-DeepLoc can successfully analyze the data based on astigmatism and saddle point engineering, and the global resolution of the reconstructed image is equivalent to or even better than FD-DeepLoc algorithm. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
20. A unified longitudinal trajectory dataset for automated vehicle.
- Author
-
Zhou, Hang, Ma, Ke, Liang, Shixiao, Li, Xiaopeng, and Qu, Xiaobo
- Subjects
DATA scrubbing ,AUTONOMOUS vehicles ,MOTOR vehicle driving ,RESEARCH personnel ,ELECTRONIC data processing - Abstract
Automated Vehicles (AVs) promise significant advances in transportation. Critical to these improvements is understanding AVs' longitudinal behavior, relying heavily on real-world trajectory data. Existing open-source trajectory datasets of AV, however, often fall short in refinement, reliability, and completeness, hindering effective performance metrics analysis and model development. This study addresses these challenges by creating a Unified longitudinal trajectory dataset for AVs (Ultra-AV) to analyze their microscopic longitudinal driving behaviors. This dataset compiles data from 14 distinct sources, encompassing various AV types, test sites, and experiment scenarios. We established a three-step data processing: 1. extraction of longitudinal trajectory data, 2. general data cleaning, and 3. data-specific cleaning to obtain the longitudinal trajectory data and car-following trajectory data. The validity of the processed data is affirmed through performance evaluations across safety, mobility, stability, and sustainability, along with an analysis of the relationships between variables in car-following models. Our work not only furnishes researchers with standardized data and metrics for longitudinal AV behavior studies but also sets guidelines for data collection and model development. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
21. Speedy Component Resolution Using Spatially Encoded Diffusion NMR Data.
- Author
-
Lorandel, Benjamin, Rocha, Hugo, Cazimajou, Oksana, Mishra, Rituraj, Bernard, Aurélie, Bowyer, Paul, Nilsson, Mathias, and Dumez, Jean‐Nicolas
- Subjects
- *
MULTIVARIATE analysis , *LEAST squares , *DIFFUSION coefficients , *ELECTRONIC data processing , *DATA analysis - Abstract
ABSTRACT Diffusion‐ordered NMR spectroscopy (DOSY) is a powerful tool to analyse mixtures. Spatially encoded (SPEN) DOSY enables recording a full DOSY dataset in just one scan by performing spatial parallelisation of the gradient dimension. The simplest and most widely used approach to processing DOSY data is to fit each peak in the spectrum with a single or multiple exponential decay. However, when there is peak overlap, and/or when the diffusion decays of the contributing components are too similar, this method has limitations. Multivariate analysis of DOSY data, which is an attractive alternative, consists of decomposing the experimental data, into compound‐specific diffusion decays and 1D NMR spectra. Multivariate analysis has been very successfully used for conventional DOSY data, but its use for SPEN DOSY data has only recently been reported. Here, we present a comparison, for SPEN DOSY data, of two widely used algorithms, SCORE and OUTSCORE, that aim at unmixing the spectra of overlapped species through a least square fit or a cross‐talk minimisation, respectively. Data processing was performed with the General NMR Analysis Toolbox (GNAT), with custom‐written code elements that now expands the capabilities, and makes it possible to import and process SPEN DOSY data. This comparison is demonstrated on three different two‐component mixtures, each with different characteristics in terms of signal overlap, diffusion coefficient similarity, and component concentration. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
22. Enhancing the implementation and integration of mHealth interventions in resource-limited settings: a scoping review.
- Author
-
Tumuhimbise, Wilson, Theuring, Stefanie, Kaggwa, Fred, Atukunda, Esther C., Rubaihayo, John, Atwine, Daniel, Sekandi, Juliet N., and Musiimenta, Angella
- Subjects
- *
RESOURCE-limited settings , *MONETARY incentives , *MOBILE health , *UNFUNDED mandates , *ELECTRONIC data processing - Abstract
Background: Although mobile health (mHealth) interventions have shown promise in improving health outcomes, most of them rarely translate to scale. Prevailing mHealth studies are largely small-sized, short-term and donor-funded pilot studies with limited evidence on their effectiveness. To facilitate scale-up, several frameworks have been proposed to enhance the generic implementation of health interventions. However, there is a lack of a specific focus on the implementation and integration of mHealth interventions in routine care in low-resource settings. Our scoping review aimed to synthesize and develop a framework that could guide the implementation and integration of mHealth interventions. Methods: We searched the PubMed, Google Scholar, and ScienceDirect databases for published theories, models, and frameworks related to the implementation and integration of clinical interventions from 1st January 2000 to 31st December 2023. The data processing was guided by a scoping review methodology proposed by Arksey and O'Malley. Studies were included if they were i) peer-reviewed and published between 2000 and 2023, ii) explicitly described a framework for clinical intervention implementation and integration, or iii) available in full text and published in English. We integrated different domains and constructs from the reviewed frameworks to develop a new framework for implementing and integrating mHealth interventions. Results: We identified eight eligible papers with eight frameworks composed of 102 implementation domains. None of the identified frameworks were specific to the integration of mHealth interventions in low-resource settings. Two constructs (skill impartation and intervention awareness) related to the training domain, four constructs (technical and logistical support, identifying committed staff, supervision, and redesigning) from the restructuring domain, two constructs (monetary incentives and nonmonetary incentives) from the incentivize domain, two constructs (organizational mandates and government mandates) from the mandate domain and two constructs (collaboration and routine workflows) from the integrate domain. Therefore, a new framework that outlines five main domains—train, restructure, incentivize, mandate, and integrate (TRIMI)—in relation to the integration and implementation of mHealth interventions in low-resource settings emerged. Conclusion: The TRIMI framework presents a realistic and realizable solution for the implementation and integration deficits of mHealth interventions in low-resource settings. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
23. Earthquake prediction using satellite data: Advances and ahead challenges.
- Author
-
Akhoondzadeh, Mehdi
- Subjects
- *
ARTIFICIAL intelligence , *ELECTRONIC data processing , *EMERGENCY management , *SEISMIC event location , *CHEMICAL precursors - Abstract
• Using remote sensing data, the statistical studies related to earthquake precursors increased significantly. • Most of them emphasized the existence of abnormal behaviors in precursors. • An attempt has been made to address the advances and challenges in the way of earthquake prediction. Due to the considerable loss of life and damages caused by powerful Earthquakes, along with the strengthening of structures and disaster management systems, many researches are being conducted associated with the possibility of creating Earthquake warning systems. However, no reliable scientific report on successful Earthquake prediction has been published so far. Although there are a couple of published scientific papers indicating the detection of unusual changes in physical and chemical parameters obtained from ground stations, but due to the limitations of in-situ measurement in terms of number, location, time and cost, there was no noticeable progress in Earthquake warning. After the advent of remote sensing satellites, the number of statistical studies related to Earthquake precursors increased significantly, and most of them emphasized the existence of abnormal behaviors in physical and chemical precursors around the time (about 1–30 days before) and locations of powerful Earthquakes. In this study, an attempt has been made to address the advances and challenges in the way of Earthquake prediction using satellite data and provide a perspective of the future of these researches. It should be noted that the increase in the number of satellites designed and launched specially for Earthquake studies, the variety of available Earthquake precursors (multi-precursor analysis), the development of classic and intelligent anomaly detection and predictor algorithms (multi-method analysis), the provision of intelligent systems for the integration of various precursors (fusion and decision systems), the creation of cloud storing and processing data services (Google Earth Engine, Giovanni, etc.) and the development of intelligent user interfaces for public, have made researchers more hopeful for the appearance of low-uncertainty Earthquake warning systems in the near future. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
24. DC_OCEAN: an open-source algorithm for identification of duplicates in ocean databases.
- Author
-
Xinyi Song, Zhetao Tan, Locarnini, Ricardo, Simoncelli, Simona, Cowley, Rebecca, Kizu, Shoichi, Boyer, Tim, Reseghetti, Franco, Castelao, Guilherme, Gouretski, Viktor, and Lijing Cheng
- Subjects
ROBUST statistics ,PRINCIPAL components analysis ,DATABASES ,DATA management ,ELECTRONIC data processing ,METADATA - Abstract
A high-quality hydrographic observational database is essential for ocean and climate studies and operational applications. Because there are numerous global and regional ocean databases, duplicate data continues to be an issue in data management, data processing and database merging, posing a challenge on effectively and accurately using oceanographic data to derive robust statistics and reliable data products. This study aims to provide algorithms to identify the duplicates and assign labels to them. We propose first a set of criteria to define the duplicate data; and second, an open-source and semi-automatic system to detect duplicate data and erroneous metadata. This system includes several algorithms for automatic checks using statistical methods (such as Principal Component Analysis and entropy weighting) and an additional expert (manual) check. The robustness of the system is then evaluated with a subset of the World Ocean Database (WOD18) with over 600,000 in-situ temperature and salinity profiles. This system is an open-source Python package (named DC_OCEAN) allowing users to effectively use the software. Users can customize their settings. The application result from the WOD18 subset also forms a benchmark dataset, which is available to support future studies on duplicate checks, metadata error identification, and machine learning applications. This duplicate checking system will be incorporated into the International Quality-controlled Ocean Database (IQuOD) data quality control system to guarantee the uniqueness of ocean observation data in this product. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
25. Mining and evaluation of adverse event signals for capmatinib based on the FAERS database.
- Author
-
Xinnan Chen, Ying Jiang, Haohao Zhu, and Man Tian
- Subjects
DATABASES ,DRUG labeling ,VOCAL cords ,DATABASE searching ,ELECTRONIC data processing ,INNER ear - Abstract
Objective: To conduct a comprehensive data analysis based on the FDA's Adverse Event Reporting System (FAERS) to mine possible adverse event (AE) signals of Capmatinib, providing valuable references for its clinical application. Methods: Capmatinib was the primary suspected drug in the search of FAERS database from the second quarter of 2020 to the fourth quarter of 2023. Data processing, screening, and classification were performed using methods such as the Reporting Odds Ratio (ROR), Proportional Reporting Ratio (PRR), Bayesian Confidence Propagation Neural Network (BCPNN), and Multi-item Gamma Poisson Shrinker (MGPS). Results: A total of 1,991 AE reports directly related to Capmatinib were screened, identifying 269 Preferred Terms (PTs) involving 26 System Organ Classes (SOCs). Besides the AEs recorded in the drug label (such as edema, nausea, fatigue, and dyspnea), the study unearthed other high-risk AEs not listed in the label, including Renal and urinary disorders, Vocal cord paralysis, and Ear and labyrinth disorders. Among these, renal and urinary disorders, and ear and labyrinth disorders had a higher frequency and intensity of signals, suggesting that their mechanisms of occurrence could be a future research direction. Conclusion: This study uncovered new potential AEs of Capmatinib based on the FAERS database, providing reference for its safe clinical use. Special attention should be given to the occurrence of ear and labyrinth disorders and renal and urinary disorders, primarily presenting as pseudo-acute kidney injury, during treatment. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
26. One‐dimensional deep learning inversion of marine controlled‐source electromagnetic data.
- Author
-
Li, Pan, Du, Zhijun, Li, Yuguo, and Wang, Jianhua
- Subjects
- *
CONVOLUTIONAL neural networks , *RECURRENT neural networks , *MACHINE learning , *ELECTRONIC data processing , *ELECTRICAL resistivity , *DEEP learning - Abstract
This paper explores the application of machine learning techniques, specifically deep learning, to the inverse problem of marine controlled‐source electromagnetic data. A novel approach is proposed that combines the convolutional neural network and recurrent neural network architectures to reconstruct layered electrical resistivity variation beneath the seafloor from marine controlled‐source electromagnetic data. The approach leverages the strengths of both convolutional neural network and recurrent neural network, where convolutional neural network is used for recognizing and classifying features in the data, and recurrent neural network is used to capture the contextual information in the sequential data. We have built a large synthetic dataset based on one‐dimensional forward modelling of a large number of resistivity models with different levels of electromagnetic structural complexity. The combined learning of convolutional neural network and recurrent neural network is used to construct the mapping relationship between the marine controlled‐source electromagnetic data and the resistivity model. The trained network is then used to predict the distribution of resistivity in the model by feeding it with marine controlled‐source electromagnetic responses. The accuracy of the proposed approach is examined using several synthetic scenarios and applied to a field dataset. We explore the sensitivity of deep learning inversion to different electromagnetic responses produced by resistive targets distributed at different depths and with varying levels of noise. Results from both numerical simulations and field data processing consistently demonstrate that deep learning inversions reliably reconstruct the subsurface resistivity structures. Moreover, the proposed method significantly improves the efficiency of electromagnetic inversion and offers significant performance advantages over traditional electromagnetic inversion methods. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
27. A hybrid transformer with domain adaptation using interpretability techniques for the application to the detection of risk situations.
- Author
-
Mallick, Rupayan, Benois-Pineau, Jenny, Zemmari, Akka, Guerda, Kamel, Mansencal, Boris, Amieva, Helene, and Middleton, Laura
- Subjects
ARTIFICIAL neural networks ,CONVOLUTIONAL neural networks ,ELECTRONIC data processing ,TIME series analysis - Abstract
Multimedia approaches are strongly required in multi-modal data processing for the detection and recognition of specific events in the data. Hybrid architectures with time series and image/video inputs in the framework of twin CNNs have shown increased performances compared to mono-modal approaches. Pre-trained models have been used in transfer learning to fine-tune the last few layers in the network. This often leads to distribution shifts in the domain. In a real-world scenario, the distribution shifts between the source and target domains can yield poor classification results. With interpretable techniques used in deep neural networks, important features can be highlighted not only for trained models but also reinforced in the training process. Hence the initialization of the target domain model can be performed with improved weights. During data transfer between datasets, the dimensions of the data are also different. We propose a method for model transfer with the adaptation of data dimension and improved initialization with interpretability approaches. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
28. Deprivation and NHS General Ophthalmic Service sight testing activity in England in 2022–2023.
- Author
-
Harper, Robert A., Hooper, Jeremy, Parkins, David J., Fenerty, Cecilia H., Roach, James, and Bowen, Michael
- Subjects
- *
ELECTRONIC data processing , *VISION testing , *ODDS ratio , *PRIMARY care , *POSTAL codes - Abstract
Purpose Methods Results Conclusion Socioeconomic deprivation is associated with an increased incidence of sight‐loss. To inform potential developments in eyecare, General Ophthalmic Service (GOS) sight‐testing activity was explored in relation to deprivation for GOS contractors submitting National Health Service (NHS) claims in England.Data on NHS sight‐test claims for the financial year 2022–2023 were sought from NHS England (NHSE), including number of sight‐tests by GOS contractors, their unique Organisation Data Service codes and postcodes and age‐bands of patients accessing sight‐testing. Deprivation scores were assigned to contractor practices using the Index of Multiple Deprivation (IMD) and the average number of sight‐tests for all contractors within each IMD decile calculated, allowing rate of sight‐testing per 1000 population per decile of deprivation to be estimated using Office of National Statistics (ONS) Lower Layer Super Output Area mid‐year population estimates. Inequality was examined using the Odds Ratio (OR) and slope and relative index of inequality measures (SSI and RII).Overall, 12.94 million NHS sight‐tests were provided by 5622 GOS contractors in England in 2022–2023. Most affluent decile GOS contractors undertook an average ~2200 NHS sight‐tests, while in the most deprived decile, average NHS sight‐tests per contractor was ~1100. Rate of sight‐testing per 1000 population in the most deprived decile was one quarter of that in the most affluent, with an OR of 5.29 (95% CI 5.27–5.30), indicating those in the most affluent areas were ~five times more likely to access NHS sight‐tests. Overall, SII and RII were 333.5 (95% CI 333.52–333.53) and 6.4 (95% CI 6.39–6.40), respectively, findings reflective of substantial inequality in uptake.There remains substantial unwarranted variation in uptake of NHS sight‐testing, with those in more affluent areas accessing sight‐testing substantially more than those in more deprived areas. Strategies are required to facilitate primary care optometry to provide more equitable access to eyecare. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
29. CoGTEx: Unscaled system-level coexpression estimation from GTEx data forecast novel functional gene partners.
- Author
-
Cortes-Guzman, Miguel-Angel and Treviño, Víctor
- Subjects
- *
GENE expression , *COMPUTER workstation clusters , *GENETIC transcription regulation , *HUMAN genes , *ELECTRONIC data processing - Abstract
Motivation: Coexpression estimations are helpful for analysis of pathways, cofactors, regulators, targets, and human health and disease. Ideally, coexpression estimations should consider as many diverse cell types as possible and consider that available data is not uniform across tissues. Importantly, the coexpression estimations accessible today are performed on a "tissue level", which is based on cell type standardized formulations. Little or no attention is paid to overall gene expression levels. The tissue-level estimation assumes that variance expression levels are more important than mean expression levels. Here, we challenge this assumption by estimating a coexpression calculation at the "system level", which is estimated without standardization by tissue, and show that it provides valuable information. We made available a resource to view, download, and analyze both, tissue- and system-level coexpression estimations from GTEx human data. Methods: GTEx v8 expression data was globally normalized, batch-processed, and filtered. Then, PCA, clustering, and tSNE stringent procedures were applied to generate 42 distinct and curated tissue clusters. Coexpression was estimated from these 42 tissue clusters computing the correlation of 33,445 genes by sampling 70 samples per tissue cluster to avoid tissue overrepresentation. This process was repeated 20 times, extracting the minimum value provided as a robust estimation. Three metrics were calculated (Pearson, Spearman, and G-statistic) in two data processing modes, at the system-level (TPM scale) and tissue levels (z-score scale). Results: We first validate our tissue-level estimations compared with other databases. Then, by specific analyses in several examples and literature validations of predictions, we show that system-level coexpression estimation differs from tissue-level estimations and that both contain valuable information reflected in biological pathways. We also show that coexpression estimations are associated to transcriptional regulation. Finally, we present CoGTEx, a valuable resource for viewing and analyzing coexpressed genes in human adult tissues from GTEx v8 data. We introduce our web resource to list, view and explore the coexpressed genes from GTEx data. Conclusion: We conclude that system-level coexpression is a novel and interesting coexpression metric capable of generating plausible predictions and biological hypotheses; and that CoGTEx is a valuable resource to view, compare, and download system- and tissue- level coexpression estimations from GTEx data. Availability: The web resource is available at http://bioinformatics.mx/cogtex. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
30. Optimal experiment design for inverse problems via selective normalization and zero-shift times.
- Author
-
Chassain, Clément, Kusiak, Andrzej, Krause, Kevin, and Battaglia, Jean-Luc
- Subjects
- *
INVERSE problems , *PARAMETER identification , *PARAMETER estimation , *SIGNAL processing , *ELECTRONIC data processing , *TIKHONOV regularization - Abstract
Inverse problems are commonly used in many fields as they enable the estimation of parameters that cannot be experimentally measured. However, the complex nature of inverse problems requires a strong background in data and signal processing. Moreover, ill-posed problems yield parameters that have a strong linear dependence on the problem. The ill-posed nature of these problems lead to many errors in numerical computations that can make parameter identification nearly impossible. In this paper, a new data processing tool is proposed to maximize the sensitivity of the model to the parameters of interest, while reducing the correlation between them. The effectiveness of the toll is demonstrated through a given inverse problem example using Periodically Pulsed Photothermal Radiometry (PPTR). [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
31. Signal reconstruction with infrared thermographic data in the application of rail defect detection.
- Author
-
Ranting Cui, Chaojun Wei, Keping Zhang, Yuning Wu, and Xuan Zhu
- Subjects
- *
SIGNAL reconstruction , *NONDESTRUCTIVE testing , *IRON & steel plates , *THERMOGRAPHY , *ELECTRONIC data processing - Abstract
Thermographic techniques are widely used to characterise material damage, porosity and moisture in structures. Due to the thermographic signal reconstruction (TSR) technique, active thermography has made significant progress. In this study, the TSR technique was applied to identify a rubber layer that mimics an internal defect, which is attached to the back of the steel plate, after a 1.0 min thermal stimulation. The peaks of the first and second derivative logarithmic curves correspond to the important time point revealing the position of the attached flaw. Furthermore, a new post-processing technique for infrared thermographic data was proposed by exploiting polynomial orders of the fitting curves to produce a synthetic image, which can highlight the flaw with distinctive contrast. Finally, the upgraded TSR technique was applied to image a natural shelling defect in a rail sample with an excellent contrast. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
32. How structural biology has changed our understanding of icosahedral viruses.
- Author
-
Comas-Garcia, Mauricio
- Subjects
- *
HEPATITIS B , *ARTIFICIAL intelligence , *VIRION , *ELECTRONIC data processing , *ALPHAVIRUSES - Abstract
Cryo-electron microscopy and tomography have allowed us to unveil the remarkable structure of icosahedral viruses. However, in the past few years, the idea that these viruses must have perfectly symmetric virions, but in some cases, it might not be true. This has opened the door to challenging paradigms in structural virology and raised new questions about the biological implications of "unusual" or "defective" symmetries and structures. Also, the continual improvement of these technologies, coupled with more rigorous sample purification protocols, improvements in data processing, and the use of artificial intelligence, has allowed solving the structure of sub-viral particles in highly heterogeneous samples and finding novel symmetries or structural defects. In this review, I initially analyzed the case of the symmetry and composition of hepatitis B virus-produced spherical sub-viral particles. Then, I focused on Alphaviruses as an example of "imperfect" icosahedrons and analyzed how structural biology has changed our understanding of the Alphavirus assembly and some biological implications arising from these discoveries. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
33. Honorarkürzung bei fehlender Anbindung eines Vertragsarztes an Telematikinfrastruktur.
- Subjects
- *
ELECTRONIC data processing , *TELEMATICS , *GENERAL Data Protection Regulation, 2016 , *PHYSICIANS , *CONTRACTS - Abstract
The article deals with the fee reduction of a contract doctor due to a lack of connection to the telematics infrastructure. The obligation to connect to the TI is considered appropriate and not as a disproportionate restriction of the medical professional freedom. The data processing during the execution of the insured person master data comparison by contract doctors complies with the requirements of the GDPR. A revision against the fee reduction was dismissed as unfounded, and the differential cost regulation according to § 106b para. 2a SGB V applies only to uneconomical prescriptions. [Extracted from the article]
- Published
- 2024
- Full Text
- View/download PDF
34. A two-stage seismic data denoising network based on deep learning.
- Author
-
Zhang, Yan, Zhang, Chi, and Song, Liwei
- Subjects
- *
CONVOLUTIONAL neural networks , *SIGNAL-to-noise ratio , *ELECTRONIC data processing , *PROBLEM solving , *NOISE - Abstract
Seismic data with a high signal-to-noise ratio is beneficial in the inversion and interpretation. Thus, denoising is an indispensable step in the seismic data processing. Traditional denoising methods based on prior knowledge are susceptible to the influence of the hypothesis model and parameters. In contrast, deep learning-based denoising methods can extract deep features from the data autonomously and generate a sophisticated denoising model through adaptive learning. However, these methods generally learn a specific model for each noise level, which results in poor representation ability and suboptimal denoising efficacy when applied to seismic data with different noise levels. To address this issue, we propose a denoising method based on a two-stage convolutional neural network (TSCNN). The TSCNN comprises an estimation subnet (ES) and a denoising subnet (DS). The ES employs a multilayer CNN to estimate noise levels, and the DS performs noise suppression on noisy seismic data based on the ES estimation of the noise distribution. In addition, attention mechanisms are implemented in the proposed network to efficiently extract noise information hidden in complex backgrounds. The TSCNN also adopts the L1 loss function to enhance the generalization ability and denoising outcome of the model, and a residual learning scheme is utilized to solve the problem of network degradations. Experimental results demonstrate that the proposed method can preserve event features more accurately and outperforms existing methods in terms of signal-to-noise ratio and generalization ability. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
35. Sangerbox 2: Enhanced functionalities and update for a comprehensive clinical bioinformatics data analysis platform.
- Author
-
Chen, Di, Xu, Lixia, Xing, Huiwu, Shen, Weitao, Song, Ziguang, Li, Hongjiang, Zhu, Xuqiang, Li, Xueyuan, Wu, Lixin, Jiao, Henan, Li, Shuang, Yan, Jing, He, Yuting, and Yan, Dongming
- Subjects
- *
PATTERN recognition systems , *SUPPORT vector machines , *COMPUTER performance , *RANDOM forest algorithms , *ELECTRONIC data processing - Abstract
In recent years, development in high‐throughput sequencing technologies has experienced an increasing application of statistics, pattern recognition, and machine learning in bioinformatics analyses. SangeBox platform to meet different scientific demands. The new version of Sangs is a widely used tool among many researchers, which encourages us to continuously improve the plerBox 2 (http://vip.sangerbox.com) and extends and optimizes the functions of interactive graphics and analysis of clinical bioinformatics data. We introduced novel analytical tools such as random forests and support vector machines, as well as corresponding plotting functions. At the same time, we also optimized the performance of the platform and fixed known problems to allow users to perform data analyses more quickly and efficiently. SangerBox 2 improved the speed of analysis, reduced resource required for computer performance, and provided more analysis methods, greatly promoting the research efficiency. Highlights: SangerBox 2.0 has expanded its features by adding new machine learning tools, including random forest and support vector machine (SVM), with improved plotting capabilities.The platform's performance and visualization tools have been significantly optimized, including the introduction of interactive adjustments for heatmaps.SangerBox 2.0 stands out in bioinformatics analysis due to its enhanced multifunctionality, user‐friendliness, and superior performance compared to other platforms. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
36. An Iteratively Reweighted Importance Kernel Bayesian Filtering Approach for High-Dimensional Data Processing.
- Author
-
Liu, Xin
- Subjects
- *
BAYESIAN field theory , *ELECTRONIC data processing - Abstract
This paper proposes an iteratively re-weighted importance kernel Bayes filter (IRe-KBF) method for handling high-dimensional or complex data in Bayesian filtering problems. This innovative approach incorporates importance weights and an iterative re-weighting scheme inspired by iteratively re-weighted Least Squares (IRLS) to enhance the robustness and accuracy of Bayesian inference. The proposed method does not require explicit specification of prior and likelihood distributions; instead, it learns the kernel mean representations from training data. Experimental results demonstrate the superior performance of this method over traditional KBF methods on high-dimensional datasets. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
37. Returnees and innovation: evidence from Chinese publicly listed firms.
- Author
-
Qiao, Yibo, Ascani, Andrea, Breschi, Stefano, and Morrison, Andrea
- Subjects
- *
ELECTRONIC data processing , *SKILLED labor , *INNOVATIONS in business , *EMPLOYMENT in foreign countries ,ECONOMIC conditions in China - Abstract
As the Chinese economy shifts from factor-driven to innovation-driven growth, Chinese firms are increasingly lacking highly skilled talents. In this context, attracting high-skill returnees might represent an effective strategy to access knowledge. In this paper, we investigate the relationship between high-skill returnees and innovation of Chinese publicly listed firms. We construct a unique dataset of 2,499 firms over the period 2002–16 by combining three different data sources (i.e. Chinese Research Data Services Platform, China Stock Market & Accounting Research Database, and LinkedIn). Our results show that different typologies of returnees (employees, technologists, and managers) with different experiences abroad (work vs study) may bring back different skills and impact differently on firm innovation. Our main findings show that (1) returnee employees and technologists are positively associated with firm's patenting; (2) returnees' overseas work experience matters more than study experience; and (3) the positive role of returnees is subject to contingencies related to firm characteristics such as ownership, location, and size. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
38. PointCloud-At: Point Cloud Convolutional Neural Networks with Attention for 3D Data Processing.
- Author
-
Umar, Saidu and Taherkhani, Aboozar
- Subjects
- *
MACHINE learning , *CONVOLUTIONAL neural networks , *POINT cloud , *ELECTRONIC data processing , *POINT processes - Abstract
The rapid growth in technologies for 3D sensors has made point cloud data increasingly available in different applications such as autonomous driving, robotics, and virtual and augmented reality. This raises a growing need for deep learning methods to process the data. Point clouds are difficult to be used directly as inputs in several deep learning techniques. The difficulty is raised by the unstructured and unordered nature of the point cloud data. So, machine learning models built for images or videos cannot be used directly on point cloud data. Although the research in the field of point clouds has gained high attention and different methods have been developed over the decade, very few research works directly with point cloud data, and most of them convert the point cloud data into 2D images or voxels by performing some pre-processing that causes information loss. Methods that directly work on point clouds are in the early stage and this affects the performance and accuracy of the models. Advanced techniques in classical convolutional neural networks, such as the attention mechanism, need to be transferred to the methods directly working with point clouds. In this research, an attention mechanism is proposed to be added to deep convolutional neural networks that process point clouds directly. The attention module was proposed based on specific pooling operations which are designed to be applied directly to point clouds to extract vital information from the point clouds. Segmentation of the ShapeNet dataset was performed to evaluate the method. The mean intersection over union (mIoU) score of the proposed framework was increased after applying the attention method compared to a base state-of-the-art framework that does not have the attention mechanism. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
39. DDE KG Editor: A data service system for knowledge graph construction in geoscience.
- Author
-
Hou, Chengbin, Liu, Kaichuang, Wang, Tianheng, Shi, Shunzhong, Li, Yan, Zhu, Yunqiang, Hu, Xiumian, Wang, Chengshan, Zhou, Chenghu, and Lv, Hairong
- Subjects
- *
KNOWLEDGE graphs , *ELECTRONIC data processing , *DATA mining , *COMPUTER science , *EARTH scientists - Abstract
Deep‐time Digital Earth (DDE) is an innovative international big science program, focusing on scientific propositions of earth evolution, changing Earth Science by coordinating global geoscience data, and sharing global geoscience knowledge. To facilitate the DDE program with recent advances in computer science, the geoscience knowledge graph plays a key role in organizing the data and knowledge of multiple geoscience subjects into Knowledge Graphs (KGs), which enables the calculation and inference over geoscience KGs for data mining and knowledge discovery. However, the construction of geoscience KGs is challenging. Though there have been some construction tools, they commonly lack collaborative editing and peer review for building high‐quality large‐scale geoscience professional KGs. To this end, a data service system or tool, DDE KG Editor, is developed to construct geoscience KGs. Specifically, it comes with several distinctive features such as collaborative editing, peer review, contribution records, intelligent assistance, and discussion forums. Currently, global geoscientists have contributed over 60,000 ontologies for 22 subjects. The stability, scalability, and intelligence of the system are regularly improving as a public online platform to better serve the DDE program. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
40. CAMELS‐SE: Long‐term hydroclimatic observations (1961–2020) across 50 catchments in Sweden as a resource for modelling, education, and collaboration.
- Author
-
Teutschbein, Claudia
- Subjects
- *
ELECTRONIC data processing , *WATER management , *STREAMFLOW , *GEOLOGICAL surveys , *RESEARCH personnel , *WATERSHEDS - Abstract
This paper introduces a community‐accessible dataset comprising daily hydroclimatic variables (precipitation, temperature, and streamflow) observed in 50 catchments in Sweden (median size of 1019 km2). The dataset covers a 60‐year period (1961–2020) and includes information on geographical location, landcover, soil classes, hydrologic signatures, and regulation for each catchment. Data were collected from various sources, such as the Swedish Meteorological and Hydrological Institute, the Swedish Geological Survey, and several Copernicus products provided by the European Environment Agency. The compiled, spatially‐matched, and processed data are publicly available online through the Swedish National Data Service (https://snd.se/en), contributing a new region to the collection of existing CAMELS (Catchment Attributes and Meteorology for Large‐sample Studies) datasets. The CAMELS‐SE dataset spans a wide range of hydroclimatic, topographic, and environmental catchment properties, making it a valuable resource for researchers and practitioners to study hydrological processes, climate dynamics, environmental impacts, and sustainable water management strategies in Nordic regions. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
41. Online data service for geologic formations (Lexicons) of China, India, Vietnam and Thailand with one‐click visualizations onto East Asia plate reconstructions.
- Author
-
Du, Wen, Mishra, Suyash, Ogg, James G., Qian, Yuzheng, Chang, Sabrina, Oberoi, Karan, Ault, Aaron, Zahirovic, Sabin, Hou, Hongfei, Raju, D. S. N., Mamallapalli, O'Neil, Ogg, Gabriele M., Li, Haipeng, Scotese, Christopher R., and Dong, Bui
- Subjects
- *
GEOLOGICAL formations , *ELECTRONIC data processing , *GEOLOGICAL time scales , *MAP projection , *PLATE tectonics ,GONDWANA (Continent) - Abstract
Paleogeography is the merger of sediment and volcanic facies, depositional settings, tectonic plate movements, topography, climate patterns and ecosystems through time. The construction of paleogeographic maps on tectonic plate reconstruction models requires a team effort to compile databases, data sharing standards and map projection methods. Two goals of the Paleogeography Working Group of the International Union of Geological Sciences (IUGS) program for Deep‐Time Digital Earth (DDE) are: (1) to interlink online national lexicons for all sedimentary and volcanic formations, and develop online ones for nations that currently lack these; (2) to target specific regions and intervals for testing/showcasing paleogeography output from the merger of these databases. Following those goals, we developed and applied new cloud‐based lexicon data services and interactive visualization techniques to regions in East Asia. This has been a successful collaboration among computer engineers and plate modellers and has involved stratigraphers in India (ONGC), China (Chengdu Univ. Tech., and Chinese Acad. Geol. Sci.), United States (Paleomap Project, and Purdue Univ.), Australia (GPlates visualization team) and Vietnam (Vietnam Nat. Univ.). Independent online lexicons with map‐based and stratigraphy‐based user interfaces have been developed (as of the date of this submission in March 2022) for all Proterozoic to Quaternary formations on the Indian Plate (over 800) and Vietnam (over 200), the majority of the Devonian through Neogene of China (ca. 2000) and partially for Thailand. A multi‐database search system returns all geologic formations of a desired geologic time from these four independent databases. With one click, users can plot the regional extent of one or of all of those regional formations on different plate reconstruction models of that desired age, and these polygons are filled with the lithologic facies pattern. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
42. High-Precision Instance Segmentation Detection of Micrometer-Scale Primary Carbonitrides in Nickel-Based Superalloys for Industrial Applications.
- Author
-
Zhang, Jie, Zheng, Haibin, Zeng, Chengwei, and Gu, Changlong
- Subjects
- *
HEAT resistant alloys , *DEEP learning , *ELECTRONIC data processing , *ALLOYS , *HOMOGENEITY - Abstract
In industrial production, the identification and characterization of micron-sized second phases, such as carbonitrides in alloys, hold significant importance for optimizing alloy compositions and processes. However, conventional methods based on threshold segmentation suffer from drawbacks, including low accuracy, inefficiency, and subjectivity. Addressing these limitations, this study introduced a carbonitride instance segmentation model tailored for various nickel-based superalloys. The model enhanced the YOLOv8n network structure by integrating the SPDConv module and the P2 small target detection layer, thereby augmenting feature fusion capability and small target detection performance. Experimental findings demonstrated notable improvements: the mAP50 (Box) value increased from 0.676 to 0.828, and the mAP50 (Mask) value from 0.471 to 0.644 for the enhanced YOLOv8n model. The proposed model for carbonitride detection surpassed traditional threshold segmentation methods, meeting requirements for precise, rapid, and batch-automated detection in industrial settings. Furthermore, to assess the carbonitride distribution homogeneity, a method for quantifying dispersion uniformity was proposed and integrated into a data processing framework for seamless automation from prediction to analysis. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
43. On the Capabilities of the IREA-CNR Airborne SAR Infrastructure.
- Author
-
Esposito, Carmen, Natale, Antonio, Lanari, Riccardo, Berardino, Paolo, and Perna, Stefano
- Subjects
- *
INFORMATION technology , *DATA warehousing , *ELECTRONIC data processing , *EMERGENCY management , *PRODUCT attributes - Abstract
In this work, the airborne Synthetic Aperture Radar (SAR) infrastructure developed at the Institute for Electromagnetic Sensing of the Environment (IREA) of the National Research Council of Italy (CNR) is described. This infrastructure allows IREA-CNR to plan and execute airborne SAR campaigns and to process the acquired data with a twofold aim. On one hand, the aim is to develop research activities; on the other hand, the aim is to support the emergency prevention and management activities of the Department of Civil Protection of the Italian Presidency of the Council of Ministers, for which IREA-CNR serves as National Centre of Competence. Such infrastructure consists of a flight segment and a ground segment that include a multi-frequency airborne SAR sensor based on the Frequency-Modulated Continuous Wave (FMCW) technology and operating in the X- and L-bands, an Information Technology (IT) platform for data storage and processing and an airborne SAR data processing chain. In this work, the technical aspects related to the flight and ground segments of the infrastructure are presented. Moreover, a discussion on the response times and characteristics of the final products that can be achieved with the infrastructure is provided with the aim of showing its capabilities to support the monitoring activities required in a possible emergency scenario. In particular, as a case study, the acquisition and subsequent interferometric processing of airborne SAR data relevant to the Stromboli volcanic area in the Sicily region, southern Italy, are presented [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
44. NDP-FD6:一种 IPv6 网络 NDP 洪泛行为多分类检测框架.
- Author
-
夏文豪, 张连成, 郭毅, 张宏涛, and 林斌“
- Subjects
- *
MACHINE learning , *TRANSFORMER models , *ELECTRONIC data processing , *ACQUISITION of data , *DEEP learning , *DENIAL of service attacks - Abstract
Current researches on NDP flooding behavior detection mainly focus on detecting RA flooding and NS flooding behaviors, and there is insufficient flooding detection for other messages of the NDP protocol. Moreover, traditional threshold rule detection methods suffer from poor dynamics and low accuracy, while most of the Al-based detection methods can only perform binary classification detection, and there are still challenges in performing multi-classification detection. In short, there is a lack of corresponding research in multi-classification flooding detection of all messages of NDP protocol. Therefore, this paper proposed a multi-classification detection framework for NDP protocol flooding behaviors, and proposed a flooding behavior detection method for NDP protocol based on time interval characteristics. The framework constructed the first multiclassification dataset for NDP flooding detection through the processes of traffic collection and data processing, it compared and used 5 machine learning and 5 deep learning algorithms to train the detection model. The experimental results show that the detection accuracy of the XGBOOST algorithm in machine learning can reach 99.18%, and the detection accuracy of the Transformer algorithm in deep learning can reach 98.45%. Compared with the existing detection methods, the accuracy is higher. Meanwhile, the detection framework can detect 9 types of flooding behaviors for all 5 types of messages of NDP protocol and classify the flooding behaviors into multiple types. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
45. 复杂事件处理中多聚合查询共享方法.
- Author
-
董攀攀, 苏航, and 高红雨
- Subjects
- *
DATA structures , *CONTINUOUS processing , *STREAMING technology , *COST structure , *ELECTRONIC data processing - Abstract
Complex event processing technology is a streaming data processing technique that detects specific event sequences or performs statistics on matching events in continuously flowing data. When processing trend aggregation queries with Kleene operators, caching intermediate results is necessary to match an indefinite number of event sequences, thus requiring significant resources from the query system. Exploiting opportunities for sharing among multiple queries, generating shared plans to guide query processing can effectively enhance the processing efficiency of trend aggregation queries. However, existing methods for handling aggregate queries do not dynamically adjust the generated shared plans in real-time to support continuous and efficient processing of complex event detection in response to changes in the execution environment. Addressing these issues, this paper proposed a dynamically updatable method for multiple aggregate query sharing to support the continuous and efficient processing of real-time changing complex event detection. By introducing the shared graph data structure and cost model, this method achieved real-time adjustment of the generated shared plans. And by using the online incremental aggregation execution sharing method, it further enhanced the processing efficiency of event trend aggregation queries with Kleene operators. This paper conducted experiments on both real and simulated datasets, compared the method with other approaches for handling aggregate queries. The results of the experiments indicate that the method effectively reduces query latency and improves the overall processing performance of queries. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
46. LiveLattice: Real‐time visualisation of tilted light‐sheet microscopy data using a memory‐efficient transformation algorithm.
- Author
-
Wang, Zichen, Hakozaki, Hiroyuki, McMahon, Gillian, Medina‐Carbonero, Marta, and Schöneberg, Johannes
- Subjects
- *
FLUORESCENCE microscopy , *SUBSTRATES (Materials science) , *MICROSCOPY , *ACQUISITION of data , *ELECTRONIC data processing - Abstract
Light‐sheet fluorescence microscopy (LSFM), a prominent fluorescence microscopy technique, offers enhanced temporal resolution for imaging biological samples in four dimensions (4D;
x ,y ,z , time). Some of the most recent implementations, including inverted selective plane illumination microscopy (iSPIM) and lattice light‐sheet microscopy (LLSM), move the sample substrate at an oblique angle relative to the detection objective's optical axis. Data from such tilted‐sample‐scan LSFMs require subsequent deskewing and rotation for proper visualisation and analysis. Such data preprocessing operations currently demand substantial memory allocation and pose significant computational challenges for large 4D dataset. The consequence is prolonged data preprocessing time compared to data acquisition time, which limits the ability for live‐viewing the data as it is being captured by the microscope. To enable the fast preprocessing of large light‐sheet microscopy datasets without significant hardware demand, we have developed WH‐Transform, a memory‐efficient transformation algorithm for deskewing and rotating the raw dataset, significantly reducing memory usage and the run time by more than 10‐fold for large image stacks. Benchmarked against the conventional method and existing software, our approach demonstrates linear runtime compared to the cubic and quadratic runtime of the other approaches. Preprocessing a raw 3D volume of 2 GB (512 × 1536 × 600 pixels) can be accomplished in 3 s using a GPU with 24 GB of memory on a single workstation. Applied to 4D LLSM datasets of human hepatocytes, lung organoid tissue and brain organoid tissue, our method provided rapid and accurate preprocessing within seconds. Importantly, such preprocessing speeds now allow visualisation of the raw microscope data stream in real time, significantly improving the usability of LLSM in biology. In summary, this advancement holds transformative potential for light‐sheet microscopy, enabling real‐time, on‐the‐fly data preprocessing, visualisation, and analysis on standard workstations, thereby revolutionising biological imaging applications for LLSM and similar microscopes. [ABSTRACT FROM AUTHOR]- Published
- 2024
- Full Text
- View/download PDF
47. FullProfAPP: a graphical user interface for the streamlined automation of powder diffraction data analysis.
- Author
-
Arcelus, Oier, Rodríguez-Carvajal, Juan, Katcho, Nebil A., Reynaud, Marine, Black, Ashley P., Chatzogiannakis, Dimitrios, Frontera, Carlos, Serrano-Sevillano, Jon, Ismail, Maha, Carrasco, Javier, Fauth, Francois, Palacin, M. Rosa, and Casas-Cabanas, Montse
- Subjects
- *
GRAPHICAL user interfaces , *RIETVELD refinement , *DIFFRACTION patterns , *ELECTRONIC data processing , *SOFTWARE development tools - Abstract
FullProfAPP is a software tool for data processing, refinement and visualization of large collections of powder diffraction patterns. Featuring an intuitive graphical user interface, it seamlessly facilitates a variety of tasks. These include conducting full‐profile phase searches, sequential and high‐throughput Rietveld refinements, and managing background (and peak) detection. FullProfAPP also provides convenient interaction with crystallographic databases and supports the visualization and export of high‐quality pixel and vector graphics depicting the refinement results, among other functionalities. FullProfAPP wraps around the refinement program FullProf [Rodríguez‐Carvajal (1993), Physica B, 192, 55–69] and offers the flexibility of user‐defined workflows by accessing and editing FullProf's input files and triggering its execution as necessary. FullProfAPP is distributed as open‐source software and is presently available for Windows and Linux operating systems. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
48. Evaluating the single-frequency static precise point positioning accuracies from multi-constellation GNSS observations at an Indian low-latitude station.
- Author
-
Aginiparthi, Anantha Srinivas, Vankadara, Ram Kumar, Mokkapati, Ravi Kumar, and Panda, Sampad Kumar
- Subjects
- *
GLOBAL Positioning System , *KALMAN filtering , *ELECTRONIC data processing , *IONOSPHERE , *CONSTELLATIONS - Abstract
Multiple constellations and their combinations in the global navigation satellite systems (GNSS) provide a great opportunity for single-station precise point positioning (PPP) models. The PPP models are of more importance for GNSS users as they are cost-effective with reasonable accuracy. There are abundant models in the market that use different data processing techniques based on the location and constellations used. In this study, we used the precise point positioning (PPP) software Net_Diff to verify the positioning accuracy at a low latitude Indian location using individual global satellite constellations (GPS, GLONASS, Galileo, BeiDou) and their combinations with GPS fixed. The ionospheric correction models such as GIM and Klobuchar are applied in the post-processing to determine the positioning accuracy. The Kalman filter method is applied to model the input data along with including the noise to derive the position solution. The results revealed that the GPS showed constant residual error for both quiet and disturbed days. In terms of single constellation Galileo produced less residuals in WGS 84 and ECEF coordinate systems whereas in the multi-constellation combination GPS, Galileo and GLONASS resulted in residuals of lesser magnitudes compared to the other combinations considered in this study. The RMS and STD of the residuals confirmed the high precision for Galileo and low precision for BeiDou constellations at the location. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
49. An alternative macro‐Raman mapping algorithm, adapting the numbers of accumulations.
- Author
-
Vandenabeele, Peter
- Subjects
- *
RADIANT intensity , *TIME measurements , *WATERCOLOR painting , *ELECTRONIC data processing , *ALGORITHMS - Abstract
Macro‐Raman mapping is an approach that is well suited to study the distribution of molecules over a surface of typically several cm2, such as the pigment distribution on a painting. During such an experiment, thousands Raman spectra are recorded in a predefined array of points. As there can be a huge difference in Raman sensitivity between different molecules, an alternative approach is proposed, where measurement times are adapted according to the quality of the Raman scatterer. Therefore, an iterative process is implemented, involving recording short accumulations, followed by a quick evaluation. As soon as the accumulated spectrum reaches a quality criterium, measurement of the next point is started, while in the other case further accumulations are added. The influence of using multiple accumulations opposite to a single measurement for a prolonged period has been assessed by comparing the overall measurement time and the spectral intensity of the main Raman band (520.5 cm−1) of silicon. Another prerequisite for this algorithm to be successful is the need for a quick and reliable evaluation of the spectral quality. This was examined by applying this function on the accumulations of Raman spectra of 49 paint samples from a colour chart. Finally, the variable‐accumulation algorithm is applied during a macro‐Raman mapping experiment of a watercolour painting on paper. Despite a significant gain in overall measuring time, this novel approach yields a qualitative data set that can be used for creating clear Raman maps, using different data processing methods. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
50. ASV vs OTUs clustering: Effects on alpha, beta, and gamma diversities in microbiome metabarcoding studies.
- Author
-
Fasolo, Andrea, Deb, Saptarathi, Stevanato, Piergiorgio, Concheri, Giuseppe, and Squartini, Andrea
- Subjects
- *
BIOINDICATORS , *ENVIRONMENTAL sciences , *ELECTRONIC data processing , *SPECIES diversity , *MICROBIAL communities - Abstract
In microbial community sequencing, involving bacterial ribosomal 16S rDNA or fungal ITS, the targeted genes are the basis for taxonomical assignment. The traditional bioinformatical procedure has for decades made use of a clustering protocol by which sequences are pooled into packages of shared percent identity, typically at 97%, to yield Operational Technical Units (OTUs). Progress in the data processing methods has however led to the possibility of minimizing technical sequencers errors, which were the main reason for the OTU choice, and to analyze instead the exact Amplicon Sequence Variants (ASV) which is a choice yielding much less agglomerated reads. We have tested the two procedures on the same 16S metabarcoded bacterial amplicons dataset encompassing a series of samples from 17 adjacent habitats, taken across a 700 meter-long transect of different ecological conditions unfolding in a gradient spanning from cropland, through meadows, forest and all successional transitions up to the seashore, within the same coastal area. This design allowed to scan a high biodiversity basin and to measure alpha, beta and gamma diversity of the area, to verify the effect of the bioinformatics on the same data as concerns the values of ten different ecological indexes and other parameters. Two levels of progressive OTUs clustering, (99% and 97%) were compared with the ASV data. The results showed that the OTUs clustering proportionally led to a marked underestimation of the ecological indicators values for species diversity and to a distorted behaviour of the dominance and evenness indexes with respect to the direct use of the ASV data. Multivariate ordination analyses resulted also sensitive in terms of tree topology and coherence. Overall, data support the view that reference-based OTU clustering carries several misleading disadvantageous biases, including the risk of missing novel taxa which are yet unreferenced in databases. Since its alternatives as de novo clustering have on the other hand drawbacks due to heavier computational demand and results comparability, especially for environmental studies which contain several yet uncharacterized species, the direct ASV based analysis, at least for prokaryotes, appears to warrant significand advantages in comparison to OTU clustering at every level of percent identity cutoff. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.