21,265 results
Search Results
2. Web-based diagnostic platform for microorganism-induced deterioration on paper-based cultural relics with iterative training from human feedback
- Author
-
Chenshu Liu, Songbin Ben, Chongwen Liu, Xianchao Li, Qingxia Meng, Yilin Hao, Qian Jiao, and Pinyi Yang
- Subjects
Paper-based cultural relics ,Conservation ,Computer vision ,Deep learning ,Strain classification ,Web application ,Fine Arts ,Analytical chemistry ,QD71-142 - Abstract
Abstract Purpose Paper-based artifacts hold significant cultural and social values. However, paper is intrinsically fragile to microorganisms, such as mold, due to its cellulose composition, which can serve as a microorganisms’ nutrient source. Mold not only can damage papers’ structural integrity and pose significant challenges to conservation works but also may subject individuals attending the contaminated artifacts to health risks. Current approaches for strain identification usually require extensive training, prolonged time for analysis, expensive operation costs, and higher risks of secondary damage due to sampling. Thus, in current conservation practices with mold-contaminated artifacts, little pre-screening or strain identification was performed before mold removal, and the cleaning techniques are usually broad-spectrum rather than strain-specific. With deep learning showing promising applications across various domains, this study investigated the feasibility of using a convolutional neural network (CNN) for fast in-situ recognition and classification of mold on paper. Methods Molds were first non-invasively sampled from ancient Xuan Paper-based Chinese books from the Qing and Ming dynasties. Strains were identified using molecular biology methods and the four most prevalent strains were inoculated on Xuan paper to create mockups for image collection. Microscopic images of the molds as well as their stains situated on paper were collected using a compound microscope and commercial microscope lens for cell phone cameras, which were then used for training CNN models with a transfer learning scheme to perform the classification of mold. To enable involvement and contribution from the research community, a web interface that actuates the process while providing interactive features for users to learn about the information of the classified strain was constructed. Moreover, a feedback functionality in the web interface was embedded for catching potential classification errors, adding additional training images, or introducing new strains, all to refine the generalizability and robustness of the model. Results & Conclusion In the study, we have constructed a suite of high-confidence classification CNN models for the diagnostic process for mold contamination in conservation. At the same time, a web interface was constructed that allows recurrently refining the model with human feedback through engaging the research community. Overall, the proposed framework opens new avenues for effective and timely identification of mold, thus enabling proactive and targeted mold remediation strategies in conservation.
- Published
- 2024
- Full Text
- View/download PDF
3. Computer vision digitization of smartphone images of anesthesia paper health records from low-middle income countries
- Author
-
Folks, Ryan D., Naik, Bhiken I., Brown, Donald E., and Durieux, Marcel E.
- Published
- 2024
- Full Text
- View/download PDF
4. Computer vision digitization of smartphone images of anesthesia paper health records from low-middle income countries
- Author
-
Ryan D. Folks, Bhiken I. Naik, Donald E. Brown, and Marcel E. Durieux
- Subjects
Computer vision ,Computer extraction of time series ,Document analysis ,Computer applications to medicine. Medical informatics ,R858-859.7 ,Biology (General) ,QH301-705.5 - Abstract
Abstract Background In low-middle income countries, healthcare providers primarily use paper health records for capturing data. Paper health records are utilized predominately due to the prohibitive cost of acquisition and maintenance of automated data capture devices and electronic medical records. Data recorded on paper health records is not easily accessible in a digital format to healthcare providers. The lack of real time accessible digital data limits healthcare providers, researchers, and quality improvement champions to leverage data to improve patient outcomes. In this project, we demonstrate the novel use of computer vision software to digitize handwritten intraoperative data elements from smartphone photographs of paper anesthesia charts from the University Teaching Hospital of Kigali. We specifically report our approach to digitize checkbox data, symbol-denoted systolic and diastolic blood pressure, and physiological data. Methods We implemented approaches for removing perspective distortions from smartphone photographs, removing shadows, and improving image readability through morphological operations. YOLOv8 models were used to deconstruct the anesthesia paper chart into specific data sections. Handwritten blood pressure symbols and physiological data were identified, and values were assigned using deep neural networks. Our work builds upon the contributions of previous research by improving upon their methods, updating the deep learning models to newer architectures, as well as consolidating them into a single piece of software. Results The model for extracting the sections of the anesthesia paper chart achieved an average box precision of 0.99, an average box recall of 0.99, and an mAP0.5-95 of 0.97. Our software digitizes checkbox data with greater than 99% accuracy and digitizes blood pressure data with a mean average error of 1.0 and 1.36 mmHg for systolic and diastolic blood pressure respectively. Overall accuracy for physiological data which includes oxygen saturation, inspired oxygen concentration and end tidal carbon dioxide concentration was 85.2%. Conclusions We demonstrate that under normal photography conditions we can digitize checkbox, blood pressure and physiological data to within human accuracy when provided legible handwriting. Our contributions provide improved access to digital data to healthcare practitioners in low-middle income countries.
- Published
- 2024
- Full Text
- View/download PDF
5. NSTU-BDTAKA: An open dataset for Bangladeshi paper currency detection and recognition
- Author
-
Md. Jubayar Alam Rafi, Mohammad Rony, and Nazia Majadi
- Subjects
Computer vision ,Deep learning ,Image analysis ,Taka detection ,Taka recognition ,YOLOv5 model ,Computer applications to medicine. Medical informatics ,R858-859.7 ,Science (General) ,Q1-390 - Abstract
One of the most popular and well-established forms of payment in use today is paper money. Handling paper money might be challenging for those with vision impairments. Assistive technology has been reinventing itself throughout time to better serve the elderly and disabled people. To detect paper currency and extract other useful information from them, image processing techniques and other advanced technologies, such as Artificial Intelligence, Deep Learning, etc., can be used. In this paper, we present a meticulously curated and comprehensive dataset named ‘NSTU-BDTAKA’ tailored for the simultaneous detection and recognition of a specific object of cultural significance - the Bangladeshi paper currency (in Bengali it is called ‘Taka’). This research aims to facilitate the development and evaluation of models for both taka detection and recognition tasks, offering a rich resource for researchers and practitioners alike. The dataset is divided into two distinct components: (i) taka detection, and (ii) taka recognition. The taka detection subset comprises 3,111 high-resolution images, each meticulously annotated with rectangular bounding boxes that encompass instances of the taka. These annotations serve as ground truth for training and validating object detection models, and we adopt the state-of-the-art YOLOv5 architecture for this purpose. In the taka recognition subset, the dataset has been extended to include a vast collection of 28,875 images, each showcasing various instances of the taka captured in diverse contexts and environments. The recognition dataset is designed to address the nuanced task of taka recognition providing researchers with a comprehensive set of images to train, validate, and test recognition models. This subset encompasses challenges such as variations in lighting, scale, orientation, and occlusion, further enhancing the robustness of developed recognition algorithms. The dataset NSTU-BDTAKA not only serves as a benchmark for taka detection and recognition but also fosters advancements in object detection and recognition methods that can be extrapolated to other cultural artifacts and objects. We envision that the dataset will catalyze research efforts in the field of computer vision, enabling the development of more accurate, robust, and efficient models for both detection and recognition tasks.
- Published
- 2024
- Full Text
- View/download PDF
6. Facilitating free travel in the Schengen area—A position paper by the European Association for Biometrics
- Author
-
Christoph Busch, Farzin Deravi, Dinusha Frings, Els Kindt, Ralph Lessmann, Alexander Nouak, Jean Salomon, Mateus Achcar, Fernando Alonso‐Fernandez, Daniel Bachenheimer, David Bethell, Josef Bigun, Matthew Brawley, Guido Brockmann, Enrique Cabello, Patrizio Campisi, Aleksandrs Cepilovs, Miles Clee, Mickey Cohen, Christian Croll, Andrzej Czyżewski, Bernadette Dorizzi, Martin Drahansky, Pawel Drozdowski, Catherine Fankhauser, Julian Fierrez, Marta Gomez‐Barrero, Georg Hasse, Richard Guest, Ekaterina Komleva, Sebastien Marcel, Gian Luca Marcialis, Laurent Mercier, Emilio Mordini, Stefance Mouille, Pavlina Navratilova, Javier Ortega‐Garcia, Dijana Petrovska, Norman Poh, Istvan Racz, Ramachandra Raghavendra, Christian Rathgeb, Christophe Remillet, Uwe Seidel, Luuk Spreeuwers, Brage Strand, Sirra Toivonen, and Andreas Uhl
- Subjects
biometrics (access control) ,biometric template protection ,biometric applications ,computer vision ,data privacy ,image analysis for biometrics ,Electronic computers. Computer science ,QA75.5-76.95 - Abstract
Abstract Due to migration, terror‐threats and the viral pandemic, various EU member states have re‐established internal border control or even closed their borders. European Association for Biometrics (EAB), a non‐profit organisation, solicited the views of its members on ways which biometric technologies and services may be used to help with re‐establishing open borders within the Schengen area while at the same time mitigating any adverse effects. From the responses received, this position paper was composed to identify ideas to re‐establish free travel between the member states in the Schengen area. The paper covers the contending needs for security, open borders and fundamental rights as well as legal constraints that any technological solution must consider. A range of specific technologies for direct biometric recognition alongside complementary measures are outlined. The interrelated issues of ethical and societal considerations are also highlighted. Provided a holistic approach is adopted, it may be possible to reach a more optimal trade‐off with regards to open borders while maintaining a high‐level of security and protection of fundamental rights. European Association for Biometrics and its members can play an important role in fostering a shared understanding of security and mobility challenges and their solutions.
- Published
- 2023
- Full Text
- View/download PDF
7. Deep Learning for Automated Visual Inspection in Manufacturing and Maintenance: A Survey of Open- Access Papers
- Author
-
Nils Hütten, Miguel Alves Gomes, Florian Hölken, Karlo Andricevic, Richard Meyes, and Tobias Meisen
- Subjects
automated visual inspection ,industrial applications ,deep learning ,computer vision ,convolutional neural network ,vision transformer ,Technology ,Applied mathematics. Quantitative methods ,T57-57.97 - Abstract
Quality assessment in industrial applications is often carried out through visual inspection, usually performed or supported by human domain experts. However, the manual visual inspection of processes and products is error-prone and expensive. It is therefore not surprising that the automation of visual inspection in manufacturing and maintenance is heavily researched and discussed. The use of artificial intelligence as an approach to visual inspection in industrial applications has been considered for decades. Recent successes, driven by advances in deep learning, present a possible paradigm shift and have the potential to facilitate automated visual inspection, even under complex environmental conditions. For this reason, we explore the question of to what extent deep learning is already being used in the field of automated visual inspection and which potential improvements to the state of the art could be realized utilizing concepts from academic research. By conducting an extensive review of the openly accessible literature, we provide an overview of proposed and in-use deep-learning models presented in recent years. Our survey consists of 196 open-access publications, of which 31.7% are manufacturing use cases and 68.3% are maintenance use cases. Furthermore, the survey also shows that the majority of the models currently in use are based on convolutional neural networks, the current de facto standard for image classification, object recognition, or object segmentation tasks. Nevertheless, we see the emergence of vision transformer models that seem to outperform convolutional neural networks but require more resources, which also opens up new research opportunities for the future. Another finding is that in 97% of the publications, the authors use supervised learning techniques to train their models. However, with the median dataset size consisting of 2500 samples, deep-learning models cannot be trained from scratch, so it would be beneficial to use other training paradigms, such as self-supervised learning. In addition, we identified a gap of approximately three years between approaches from deep-learning-based computer vision being published and their introduction in industrial visual inspection applications. Based on our findings, we additionally discuss potential future developments in the area of automated visual inspection.
- Published
- 2024
- Full Text
- View/download PDF
8. Deep Learning for 3D Reconstruction, Augmentation, and Registration: A Review Paper.
- Author
-
Vinodkumar, Prasoon Kumar, Karabulut, Dogus, Avots, Egils, Ozcinar, Cagri, and Anbarjafari, Gholamreza
- Subjects
- *
DEEP learning , *COMPUTER vision , *GRAPH neural networks , *ARTIFICIAL intelligence , *MACHINE learning , *GENERATIVE adversarial networks - Abstract
The research groups in computer vision, graphics, and machine learning have dedicated a substantial amount of attention to the areas of 3D object reconstruction, augmentation, and registration. Deep learning is the predominant method used in artificial intelligence for addressing computer vision challenges. However, deep learning on three-dimensional data presents distinct obstacles and is now in its nascent phase. There have been significant advancements in deep learning specifically for three-dimensional data, offering a range of ways to address these issues. This study offers a comprehensive examination of the latest advancements in deep learning methodologies. We examine many benchmark models for the tasks of 3D object registration, augmentation, and reconstruction. We thoroughly analyse their architectures, advantages, and constraints. In summary, this report provides a comprehensive overview of recent advancements in three-dimensional deep learning and highlights unresolved research areas that will need to be addressed in the future. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
9. Editorial for the Special Issue on "Feature Papers in Section AI in Imaging".
- Author
-
Fernández-Caballero, Antonio
- Subjects
GENERATIVE artificial intelligence ,COMPUTER vision ,ARTIFICIAL intelligence ,INTELLIGENT agents ,COMPUTER graphics ,DEEP learning ,EXPERT systems - Published
- 2024
- Full Text
- View/download PDF
10. Method and Installation for Efficient Automatic Defect Inspection of Manufactured Paper Bowls.
- Author
-
Yu, Shaoyong, Lee, Yang-Han, Chen, Cheng-Wen, Gao, Peng, Xu, Zhigang, Chen, Shunyi, and Yang, Cheng-Fu
- Subjects
COMPUTER vision ,AUTOMATIC optical inspection ,SEMICONDUCTOR detectors ,IMAGE processing ,PIXELS ,SENSOR arrays ,LIGHT sources - Abstract
Various techniques were combined to optimize an optical inspection system designed to automatically inspect defects in manufactured paper bowls. A self-assembled system was utilized to capture images of defects on the bowls. The system employed an image sensor with a multi-pixel array that combined a complementary metal-oxide semiconductor and a photo detector. A combined ring light served as the light source, while an infrared (IR) LED matrix panel was used to provide constant IR light to highlight the outer edges of the objects being inspected. The techniques employed in this study to enhance defect inspections on produced paper bowls included Gaussian filtering, Sobel operators, binarization, and connected components. Captured images were processed using these technologies. Once the non-contact inspection system's machine vision method was completed, defects on the produced paper bowls were inspected using the system developed in this study. Three inspection methods were used in this study: internal inspection, external inspection, and bottom inspection. All three methods were able to inspect surface features of produced paper bowls, including dirt, burrs, holes, and uneven thickness. The results of our study showed that the average time required for machine vision inspections of each paper bowl was significantly less than the time required for manual inspection. Therefore, the investigated machine vision system is an efficient method for inspecting defects in fabricated paper bowls. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
11. Valmet to Supply IQ Web Inspection System to Yueyang Forest & Paper Co., Ltd.
- Subjects
- *
PAPER industry , *PRODUCT quality , *LED lighting , *COMPUTER vision , *INFORMATION retrieval - Published
- 2024
12. Research and Evaluation on an Optical Automatic Detection System for the Defects of the Manufactured Paper Cups.
- Author
-
Wang, Ping, Lee, Yang-Han, Tseng, Hsien-Wei, and Yang, Cheng-Fu
- Subjects
- *
COMPUTER vision , *SURFACE defects , *LIGHT sources , *IMAGE sensors , *IMAGE processing ,RESEARCH evaluation - Abstract
In this paper, the paper cups were used as the research objects, and the machine vision detection technology was combined with different image processing techniques to investigate a non-contact optical automatic detection system to identify the defects of the manufactured paper cups. The combined ring light was used as the light source, an infrared (IR) LED matrix panel was used to provide the IR light to constantly highlight the outer edges of the detected objects, and a multi-grid pixel array was used as the image sensor. The image processing techniques, including the Gaussian filter, Sobel operator, Binarization process, and connected component, were used to enhance the inspection and recognition of the defects existing in the produced paper cups. There were three different detection processes for paper cups, which were divided into internal, external, and bottom image acquisition processes. The present study demonstrated that all the detection processes could clearly detect the surface defect features of the manufactured paper cups, such as dirt, burrs, holes, and uneven thickness. Our study also revealed that the average time for the investigated Automatic Optical Detection to detect the defects on the paper cups was only 0.3 s. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
13. Facilitating free travel in the Schengen area—A position paper by the European Association for Biometrics.
- Author
-
Busch, Christoph, Deravi, Farzin, Frings, Dinusha, Kindt, Els, Lessmann, Ralph, Nouak, Alexander, Salomon, Jean, Achcar, Mateus, Alonso‐Fernandez, Fernando, Bachenheimer, Daniel, Bethell, David, Bigun, Josef, Brawley, Matthew, Brockmann, Guido, Cabello, Enrique, Campisi, Patrizio, Cepilovs, Aleksandrs, Clee, Miles, Cohen, Mickey, and Croll, Christian
- Subjects
- *
BIOMETRY , *DATA privacy , *BORDER security , *NONPROFIT organizations , *CIVIL rights - Abstract
Due to migration, terror‐threats and the viral pandemic, various EU member states have re‐established internal border control or even closed their borders. European Association for Biometrics (EAB), a non‐profit organisation, solicited the views of its members on ways which biometric technologies and services may be used to help with re‐establishing open borders within the Schengen area while at the same time mitigating any adverse effects. From the responses received, this position paper was composed to identify ideas to re‐establish free travel between the member states in the Schengen area. The paper covers the contending needs for security, open borders and fundamental rights as well as legal constraints that any technological solution must consider. A range of specific technologies for direct biometric recognition alongside complementary measures are outlined. The interrelated issues of ethical and societal considerations are also highlighted. Provided a holistic approach is adopted, it may be possible to reach a more optimal trade‐off with regards to open borders while maintaining a high‐level of security and protection of fundamental rights. European Association for Biometrics and its members can play an important role in fostering a shared understanding of security and mobility challenges and their solutions. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
14. VIBRANT-WALK: An algorithm to detect plagiarism of figures in academic papers.
- Author
-
Parmar, Shashank and Jain, Bhavya
- Subjects
- *
PLAGIARISM , *COMPUTER algorithms , *ALGORITHMS , *COMPUTER vision , *RANDOM walks - Abstract
Detecting plagiarism in academic papers is crucial for maintaining academic integrity, preserving the originality of published work, and safeguarding intellectual property. While existing applications excel at text plagiarism detection, they fall short when it comes to image plagiarism. This paper introduces a novel algorithm, named "VIBRANT-WALK," designed to detect image plagiarism in academic manuscripts. The challenge of identifying plagiarized images is formidable, requiring a unique approach. Traditional Computer Vision algorithms, proficient in image similarity tasks, face limitations in determining whether an image has been previously used in an article. To address this, the proposed algorithm leverages a repository of all published article pages, focusing on absolute identicality rather than image similarity. The algorithm comprises two stages. In the first stage, a "Vibrancy Matrix" is created through image preprocessing, aiding in contour determination. The second stage involves pixel-by-pixel comparison with images from published manuscripts. To enhance efficiency, the algorithm initiates comparisons from the pixel with the highest score in the Vibrancy Matrix, followed by pixel comparisons through random walks, significantly reducing complexity. To conduct the study, a custom dataset was compiled from 69 research articles, capturing snapshots of each page and figure. Overall, we present 485 unique test cases where we can test the accuracy and efficiency of the algorithm. The lack of publicly available datasets necessitated this approach. The proposed algorithm outperformed the existing models and algorithms in this field by achieving an overall accuracy of 94.8% on the collated dataset, identifying 460 instances of plagiarism out of the 485 test cases. The algorithm also demonstrated a 100% accuracy rate in avoiding false positives. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
15. Special issue on intelligent systems: ISMIS 2022 selected papers.
- Author
-
Ceci, Michelangelo, Flesca, Sergio, Manco, Giuseppe, and Masciari, Elio
- Subjects
MACHINE learning ,ARTIFICIAL intelligence ,DECISION support systems ,KNOWLEDGE representation (Information theory) ,COMPUTER vision ,DEEP learning - Abstract
This document is a special issue of the Journal of Intelligent Information Systems, focusing on the selected papers from the International Symposium on Methodologies for Intelligent Systems (ISMIS 2022). The symposium, held in Cosenza, Italy, showcased research on various topics related to artificial intelligence, including decision support, knowledge representation, machine learning, computer vision, and more. The special issue includes eleven papers that have undergone rigorous peer-reviewing and cover a wide range of research topics, such as deep learning, anomaly detection, malware detection, sentiment classification, and healthcare professionals' burnout. The authors express their gratitude to the contributors and reviewers for their valuable contributions. [Extracted from the article]
- Published
- 2024
- Full Text
- View/download PDF
16. PaperNet: A Dataset and Benchmark for Fine-Grained Paper Classification.
- Author
-
Yue, Tan, Li, Yong, Shi, Xuzhao, Qin, Jiedong, Fan, Zijiao, and Hu, Zonghai
- Subjects
NATURAL language processing ,COMPUTER vision ,VISUAL fields ,CLASSIFICATION - Abstract
Document classification is an important area in Natural Language Processing (NLP). Because a huge amount of scientific papers have been published at an accelerating rate, it is beneficial to carry out intelligent paper classifications, especially fine-grained classification for researchers. However, a public scientific paper dataset for fine-grained classification is still lacking, so the existing document classification methods have not been put to the test. To fill this vacancy, we designed and collected the PaperNet-Dataset that consists of multi-modal data (texts and figures). PaperNet 1.0 version contains hierarchical categories of papers in the fields of computer vision (CV) and NLP, 2 coarse-grained and 20 fine-grained (7 in CV and 13 in NLP). We ran current mainstream models on the PaperNet-Dataset, along with a multi-modal method that we propose. Interestingly, none of these methods reaches an accuracy of 80% in fine-grained classification, showing plenty of room for improvement. We hope that PaperNet-Dataset will inspire more work in this challenging area. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
17. Watching plants grow – a position paper on computer vision and Arabidopsis thaliana
- Author
-
Jonathan Bell and Hannah M. Dee
- Subjects
computer vision ,Arabidopsis thaliana ,image analysis ,image segmentation ,image sequences ,Computer applications to medicine. Medical informatics ,R858-859.7 ,Computer software ,QA76.75-76.765 - Abstract
The authors present a comprehensive overview of image processing and analysis work done to support research into the model flowering plant Arabidopsis thaliana. Beside the plant's importance in biological research, using image analysis to obtain experimental measurements of it is an interesting vision problem in its own right, involving the segmentation and analysis of sequences of images of objects whose shape varies between individual specimens and also changes over time. While useful measurements can be obtained by segmenting a whole plant from the background, they suggest that the increased range and precision of measurements made available by leaf‐level segmentation makes this a problem well worth solving. A variety of approaches have been tried by biologists as well as computer vision researchers. This is an interdisciplinary area and the computer vision community has an important contribution to make. They suggest that there is a need for publicly available datasets with ground truth annotations to enable the evaluation of new approaches and to support the building of training data for modern data‐driven computer vision approaches, which are those most likely to result in the kind of fully automated systems that will be of use to biologists.
- Published
- 2017
- Full Text
- View/download PDF
18. AI Machine Vision based Oven White Paper Color Classification and Label Position Real-time Monitoring System to Check Direction.
- Author
-
Hee-Chul Kim, Youn-Saup Yoon, and Yong-Mo Kim
- Subjects
COMPUTER vision ,DEEP learning ,JOB classification ,MANUFACTURING process automation ,ARTIFICIAL intelligence ,COLOR image processing - Abstract
We develop a vision system for batch inspection by oven white paper model color by manufacturing a machine vision system for the oven manufacturing automation process. In the vision system, white paper object detection (spring), color clustering, and histogram extraction are performed. In addition, for the automated process of home appliances, we intend to develop an automatic mold combination detection algorithm that inspects the label position and direction (angle/coordinate) using deep learning. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
19. Intelligent computer vision system for segregating recyclable waste papers
- Author
-
Rahman, Mohammad Osiur, Hussain, Aini, Scavino, Edgar, Basri, Hassan, and Hannan, M.A.
- Subjects
- *
PAPER recycling , *COMPUTER vision , *IMAGE processing , *ARTIFICIAL intelligence , *SORTING (Electronic computers) , *MACHINE learning , *PATTERN recognition systems , *PAPER chemicals - Abstract
Abstract: This article explores the application of image processing techniques in recyclable waste paper sorting. In recycling, waste papers are segregated into various grades as they are subjected to different recycling processes. Highly sorted paper streams facilitate high quality end products and save processing chemicals and energy. From 1932 to 2009, different mechanical and optical paper sorting methods have been developed to fill the paper sorting demand. Still, in many countries including Malaysia, waste papers are sorted into different grades using a manual sorting system. Because of inadequate throughput and some major drawbacks of mechanical paper sorting systems, the popularity of optical paper sorting systems has increased. Automated paper sorting systems offer significant advantages over human inspection in terms of worker fatigue, throughput, speed, and accuracy. This research attempts to develop a smart vision sensing system that is able to separate the different grades of paper using first-order features. To construct a template database, a statistical approach with intra-class and inter-class variation techniques are applied to the feature selection process. Finally, the K-nearest neighbor (KNN) algorithm is applied for paper object grade identification. The remarkable achievement obtained with the method is the accurate identification and dynamic sorting of all grades of papers using simple image processing techniques. [Copyright &y& Elsevier]
- Published
- 2011
- Full Text
- View/download PDF
20. Enhancing IoT Network Security: Unveiling the Power of Self-Supervised Learning against DDoS Attacks.
- Author
-
Almaraz-Rivera, Josue Genaro, Cantoral-Ceballos, Jose Antonio, and Botero, Juan Felipe
- Subjects
SUPERVISED learning ,DENIAL of service attacks ,COMPUTER network security ,INTRUSION detection systems (Computer security) ,COMPUTER network traffic ,INTERNET of things ,ELECTRONIC paper - Abstract
The Internet of Things (IoT), projected to exceed 30 billion active device connections globally by 2025, presents an expansive attack surface. The frequent collection and dissemination of confidential data on these devices exposes them to significant security risks, including user information theft and denial-of-service attacks. This paper introduces a smart, network-based Intrusion Detection System (IDS) designed to protect IoT networks from distributed denial-of-service attacks. Our methodology involves generating synthetic images from flow-level traffic data of the Bot-IoT and the LATAM-DDoS-IoT datasets and conducting experiments within both supervised and self-supervised learning paradigms. Self-supervised learning is identified in the state of the art as a promising solution to replace the need for massive amounts of manually labeled data, as well as providing robust generalization. Our results showcase that self-supervised learning surpassed supervised learning in terms of classification performance for certain tests. Specifically, it exceeded the F1 score of supervised learning for attack detection by 4.83% and by 14.61% in accuracy for the multiclass task of protocol classification. Drawing from extensive ablation studies presented in our research, we recommend an optimal training framework for upcoming contrastive learning experiments that emphasize visual representations in the cybersecurity realm. This training approach has enabled us to highlight the broader applicability of self-supervised learning, which, in some instances, outperformed supervised learning transferability by over 5% in precision and nearly 1% in F1 score. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
21. Intelligent Noncontact Structural Displacement Detection Method Based on Computer Vision and Deep Learning.
- Author
-
Liu, Hongbo, Zhang, Fan, Ma, Rui, Wang, Longxuan, Chen, Zhihua, Zhang, Qian, and Guo, Liulu
- Subjects
DISPLACEMENT (Psychology) ,COMPUTER vision ,EUCLIDEAN distance ,COMPUTER simulation ,DEEP learning ,BAMBOO - Abstract
Accurate identification of structural displacements is important for structural state assessment and performance evaluation. This paper proposes a real-time structural displacement detection model based on computer vision and deep learning. The model consists of three stages: identification, tracking, and displacement resolution. First, the displacement target is identified and tracked by the improved YOLO v7 algorithm and the improved DeepSORT algorithm. Then, the Euclidean distance method based on inverse perspective mapping (IPM-ED) is proposed for the analytical conversion of the displacement. Next, the accuracy and effectiveness of this displacement detection model are evaluated through four groups of bamboo axial compression tests. A comparative analysis is conducted between the IPM-ED displacement analysis method and the commonly used ED displacement analysis method. Finally, the robustness of this method is tested by using a cable breakage test of a cable dome structure as an application case. The research results demonstrate that the maximum average error of the four groups of bamboo displacement tests is only 3.10 mm, and the maximum relative error of peak displacement is only 6.54 mm. The RMSE basically stays around 3.5 mm. The maximum displacement error in the application case is only 4.91 mm, with a maximum MAPE of 4.94%. In addition, the error percentage under the IPM-ED algorithm is basically within 5%, while the error percentage of the ED algorithm is more than 10%. The method in this paper achieves efficient and intelligent identification of structural displacements in a non-contact manner. The proposed method is suitable for environments where the contact displacement sensor is easily affected by vibration, the site is complex and requires additional displacement sensor fixing equipment, the displacement sensor with super-high structure is unsafe to deploy, and the contact displacement sensor in narrow space is inconvenient to deploy, so it has broad application prospects. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
22. Irregular Rotation Deformation from Paper Scanning: An Investigation.
- Author
-
Nasrudin, Mohammad Faidzul, Wahdan, Omar M., and Omar, Khairuddin
- Subjects
SCANNING systems ,ARTIFICIAL intelligence ,DIGITAL images ,COMPUTER vision ,COMPUTER science ,DIGITAL computer simulation - Abstract
Abstract: Image acquisition has great influence on the performance of any computer vision application. Different methods can be utilized to acquire the digital image of a paper, whilst scanning scheme is among the most attractive methods. This attractiveness is because of the fewer types of potential deformations and the low cost of the scanning devices, e.g. flatbed scanners. However, paper is commonly placed imperfectly on the scanner. This slight rotation is not usually based on a pivot around the paper''s geometrical center (the well known regular rotation) but instead it is based on a pivot placed at the corner of the paper. Thus, the result is a digital image that is deformed with an “irregular rotation”. The characteristic of this deformation phenomenon is currently unknown to computer vision scientists. In this paper we provide an extensive investigation of this deformation. In addition, a new set of equations that sway and measure the transformation is proposed. Our investigation leads to the conclusion that the “irregular rotation” phenomenon produces a shear transformation. Furthermore, the experimental results confirm the theoretical findings. [Copyright &y& Elsevier]
- Published
- 2012
- Full Text
- View/download PDF
23. Quality assurance strategies for machine learning applications in big data analytics: an overview.
- Author
-
Ogrizović, Mihajlo, Drašković, Dražen, and Bojić, Dragan
- Subjects
MACHINE learning ,NATURAL language processing ,COMPUTER vision ,ARTIFICIAL intelligence ,DATA analytics ,DEEP learning - Abstract
Machine learning (ML) models have gained significant attention in a variety of applications, from computer vision to natural language processing, and are almost always based on big data. There are a growing number of applications and products with built-in machine learning models, and this is the area where software engineering, artificial intelligence and data science meet. The requirement for a system to operate in a real-world environment poses many challenges, such as how to design for wrong predictions the model may make; How to assure safety and security despite possible mistakes; which qualities matter beyond a model's prediction accuracy; How can we identify and measure important quality requirements, including learning and inference latency, scalability, explainability, fairness, privacy, robustness, and safety. It has become crucial to test thoroughly these models to assess their capabilities and potential errors. Existing software testing methods have been adapted and refined to discover faults in machine learning and deep learning models. This paper covers a taxonomy, a methodologically uniform presentation of all presented solutions to the aforementioned issues, as well as conclusions about possible future development trends. The main contributions of this paper are a classification that closely follows the structure of the ML-pipeline, a precisely defined role of each team member within that pipeline, an overview of trends and challenges in the combination of ML and big data analytics, with uses in the domains of industry and education. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
24. DPNet: Scene text detection based on dual perspective CNN-transformer.
- Author
-
Li, Yuan
- Subjects
MACHINE learning ,COMPUTER vision ,VISUAL fields ,TRANSFORMER models ,CONTEXTUAL learning ,DEEP learning - Abstract
With the continuous advancement of deep learning, research in scene text detection has evolved significantly. However, complex backgrounds and various text forms complicate the task of detecting text from images. CNN is a deep learning algorithm that automatically extracts features through convolution operation. In the task of scene text detection, it can capture local text features in images, but it lacks global attributes. In recent years, inspired by the application of transformers in the field of computer vision, it can capture the global information of images and describe them intuitively. Therefore, this paper proposes scene text detection based on dual perspective CNN-transformer. The channel enhanced self-attention module (CESAM) and spatial enhanced self-attention module (SESAM) proposed in this paper are integrated into the traditional ResNet backbone network. This integration effectively facilitates the learning of global contextual information and positional relationships of text, thereby alleviating the challenge of detecting small target text. Furthermore, this paper introduces a feature decoder designed to refine the effective text information within the feature map and enhance the perception of detailed information. Experiments show that the method proposed in this paper significantly improves the robustness of the model for different types of text detection. Compared to the baseline, it achieves performance improvements of 2.51% (83.81 vs. 81.3) on the Total-Text dataset, 1.87% (86.07 vs. 84.2) on the ICDAR 2015 dataset, and 3.63% (86.72 vs. 83.09) on the MSRA-TD500 dataset, while also demonstrating better visual effects. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
25. INTRODUCTION TO THE SPECIAL ISSUE ON NEXT GENERATION PERVASIVE RECONFIGURABLE COMPUTING FOR HIGH PERFORMANCE REAL TIME APPLICATIONS.
- Author
-
VENKATESAN, C., YU-DONG ZHANG, CHOW CHEE ONN, and AND YONG SHI
- Subjects
MACHINE learning ,REINFORCEMENT learning ,HIGH performance computing ,COMPUTER vision ,ARTIFICIAL intelligence ,PARSING (Computer grammar) ,DEEP learning - Abstract
This document introduces a special issue of the journal "Scalable Computing: Practice & Experience" focused on next-generation pervasive reconfigurable computing for high-performance real-time applications. The authors discuss the importance of adaptable platforms for real-time tasks and highlight the benefits of reconfigurable computing in accelerating applications like image processing and machine learning. The special issue aims to explore recent advancements in this field and includes research papers on topics such as network security, malware detection, software reliability prediction, and optimization algorithms for wing design. The papers cover a range of computer science and technology topics, showcasing advancements and their potential impact on various computing domains. [Extracted from the article]
- Published
- 2024
- Full Text
- View/download PDF
26. Pantograph Slider Detection Architecture and Solution Based on Deep Learning.
- Author
-
Guo, Qichang, Tang, Anjie, and Yuan, Jiabin
- Subjects
IMAGE processing ,TECHNICAL specifications ,COMPUTER vision ,DEEP learning ,PANTOGRAPH - Abstract
Railway transportation has been integrated into people's lives. According to the "Notice on the release of the General Technical Specification of High-speed Railway Power Supply Safety Testing (6C System) System" issued by the National Railway Administration of China in 2012, it is required to install pantograph and slide monitoring devices in high-speed railway stations, station throats and the inlet and exit lines of high-speed railway sections, and it is required to detect the damage of the slider with high precision. It can be seen that the good condition of the pantograph slider is very important for the normal operation of the railway system. As a part of providing power for high-speed rail and subway, the pantograph must be paid attention to in railway transportation to ensure its integrity. The wear of the pantograph is mainly due to the contact power supply between the slide block and the long wire during high-speed operation, which inevitably produces scratches, resulting in depressions on the upper surface of the pantograph slide block. During long-term use, because the depression is too deep, there is a risk of fracture. Therefore, it is necessary to monitor the slider regularly and replace the slider with serious wear. At present, most of the traditional methods use automation technology or simple computer vision technology for detection, which is inefficient. Therefore, this paper introduces computer vision and deep learning technology into pantograph slide wear detection. Specifically, this paper mainly studies the wear detection of the pantograph slider based on deep learning and the main purpose is to improve the detection accuracy and improve the effect of segmentation. From a methodological perspective, this paper employs a linear array camera to enhance the quality of the data sets. Additionally, it integrates an attention mechanism to improve segmentation performance. Furthermore, this study introduces a novel image stitching method to address issues related to incomplete images, thereby providing a comprehensive solution. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
27. Guest Editorial: Advanced image restoration and enhancement in the wild.
- Author
-
Wang, Longguang, Li, Juncheng, Yokoya, Naoto, Timofte, Radu, and Guo, Yulan
- Subjects
IMAGE intensifiers ,IMAGE reconstruction ,COMPUTER vision ,SCHOLARSHIPS ,COMPUTER engineering ,IMAGE denoising ,DEEP learning ,VIDEO compression - Abstract
This document is a guest editorial from the journal IET Computer Vision, discussing the topic of advanced image restoration and enhancement. The editorial highlights the challenges faced in this field, such as the complexity of degradation models for real-world low-quality images and the difficulty of acquiring paired data. It also introduces a special issue of the journal that includes five accepted papers, which focus on video reconstruction and image super-resolution. The editorial concludes by providing brief summaries of each accepted paper. The guest editors of the special issue are also mentioned, along with their research interests and affiliations. [Extracted from the article]
- Published
- 2024
- Full Text
- View/download PDF
28. The people behind the papers – Thomas Naert and Soeren Lienkamp.
- Subjects
- *
DEEP learning , *POLYCYSTIC kidney disease , *CYSTIC kidney disease , *DEVELOPMENTAL biology , *COMPUTER vision - Published
- 2021
- Full Text
- View/download PDF
29. Development and testing of an image transformer for explainable autonomous driving systems
- Author
-
Dong, Jiqian, Chen, Sikai, Miralinaghi, Mohammad, Chen, Tiantian, and Labi, Samuel
- Published
- 2022
- Full Text
- View/download PDF
30. The Global Publication Output in Augmented Reality Research: A Scientometric Assessment for 1992-2019.
- Author
-
Gupta, B. M. and Dhawan, S. M.
- Subjects
AUGMENTED reality ,DATA visualization ,COMPUTER graphics ,COMPUTER science ,VIRTUAL reality ,CITATION indexes - Abstract
This paper describes global research in the field of augmented reality (22078) as indexed in Scopus database during 1992-2019, using a series of bibliometric indicators. The augmented reality (AR) research registered high 54.23% growth, averaged citation impact of 8.90 citations per paper. Nearly 1% of global output in the subject (226 papers) registered high-end citations (100+) per paper. The top 15 countries accounted for 87.05% of global publications output in the subject. The USA is in leadership position for its highest publications productivity (19.25% global share). The U.K. leads the world on relative citation index (2.05). International collaboration has been a major driver of AR research pursuits; between 11.89% and 44.04% of national share of top 15 countries in AR research appeared as international collaborative publications. AR research productivity by application types was the largest across sectors, such as education, industry and medical. Computer science has emerged as the most popular areas in AR research pursuits. Technical University of Munich, Germany and Osaka University, Japan have been the most productive organizations and Nara Institute of S&T, Japan (66.55 and 7.48) and Imperial College, London, U.K. (57.14 and 6.42) have been the most impactful organizations. M. Billinghurst and N. Navab have been the most productive authors and S. Feiner and B. MacIntyre have been the most impactful authors. IEEE Transactions on Visualization & Computer Graphics, Multimedia Tools & Applications and Virtual Reality topped the list of most productive journals. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
31. The emergence and evolution of urban AI.
- Author
-
Batty, Michael
- Subjects
ARTIFICIAL intelligence ,DEEP learning ,COMPUTER vision - Abstract
The fourth paper which is about "Emotional AI and Crime" takes the argument into the key area of how good or bad are AI techniques that are designed for facial and related recognition. To an extent, the focus on AI is wider than what might find in any discussion of AI in the narrower technical field for context is all important to see urban AI in context. Artificial intelligence (AI) emerged alongside the development of the digital computer more than 80 years ago during the second world war. [Extracted from the article]
- Published
- 2023
- Full Text
- View/download PDF
32. SIFusion: Lightweight infrared and visible image fusion based on semantic injection.
- Author
-
Qian, Song, Yang, Liwei, Xue, Yan, and Li, Ping
- Subjects
IMAGE fusion ,COMPUTER vision ,INFRARED imaging ,VISIBLE spectra ,INFORMATION processing - Abstract
The objective of image fusion is to integrate complementary features from source images to better cater to the needs of human and machine vision. However, existing image fusion algorithms predominantly focus on enhancing the visual appeal of the fused image for human perception, often neglecting their impact on subsequent high-level visual tasks, particularly the processing of semantic information. Moreover, these fusion methods that incorporate downstream tasks tend to be overly complex and computationally intensive, which is not conducive to practical applications. To address these issues, a lightweight infrared and visible light image fusion method known as SIFusion, which is based on semantic injection, is proposed in this paper. This method employs a semantic-aware branch to extract semantic feature information, and then integrates these features into the fused features through a Semantic Injection Module (SIM) to meet the semantic requirements of high-level visual tasks. Furthermore, to simplify the complexity of the fusion network, this method introduces an Edge Convolution Module (ECB) based on structural reparameterization technology to enhance the representational capacity of the encoder and decoder. Extensive experimental comparisons demonstrate that the proposed method performs excellently in terms of visual appeal and advanced semantics, providing satisfactory fusion results for subsequent high-level visual tasks even in challenging scenarios. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
33. Semantic Segmentation Model-Based Boundary Line Recognition Method for Wheat Harvesting.
- Author
-
Wang, Qian, Qin, Wuchang, Liu, Mengnan, Zhao, Junjie, Zhu, Qingzhen, and Yin, Yanxin
- Subjects
GEOGRAPHIC boundaries ,COMPUTER vision ,COMBINES (Agricultural machinery) ,AUTONOMOUS vehicles ,WHEAT - Abstract
The wheat harvesting boundary line is vital reference information for the path tracking of an autonomously driving combine harvester. However, unfavorable factors, such as a complex light environment, tree shade, weeds, and wheat stubble color interference in the field, make it challenging to identify the wheat harvest boundary line accurately and quickly. Therefore, this paper proposes a harvest boundary line recognition model for wheat harvesting based on the MV3_DeepLabV3+ network framework, which can quickly and accurately complete the identification in complex environments. The model uses the lightweight MobileNetV3_Large as the backbone network and the LeakyReLU activation function to avoid the neural death problem. Depth-separable convolution is introduced into Atrous Spatial Pyramid Pooling (ASPP) to reduce the complexity of network parameters. The cubic B-spline curve-fitting method extracts the wheat harvesting boundary line. A prototype harvester for wheat harvesting boundary recognition was built, and field tests were conducted. The test results show that the wheat harvest boundary line recognition model proposed in this paper achieves a segmentation accuracy of 98.04% for unharvested wheat regions in complex environments, with an IoU of 95.02%. When the combine harvester travels at 0~1.5 m/s, the normal speed for operation, the average processing time and pixel error for a single image are 0.15 s and 7.3 pixels, respectively. This method could achieve high recognition accuracy and fast recognition speed. This paper provides a practical reference for the autonomous harvesting operation of a combine harvester. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
34. Improved YOLOv5 Network for High-Precision Three-Dimensional Positioning and Attitude Measurement of Container Spreaders in Automated Quayside Cranes.
- Author
-
Zhang, Yujie, Song, Yangchen, Zheng, Luocheng, Postolache, Octavian, Mi, Chao, and Shen, Yang
- Subjects
COMPUTER vision ,CRANES (Machinery) ,CONTAINERS ,ATTITUDE (Psychology) ,CAMERAS - Abstract
For automated quayside container cranes, accurate measurement of the three-dimensional positioning and attitude of the container spreader is crucial for the safe and efficient transfer of containers. This paper proposes a high-precision measurement method for the spreader's three-dimensional position and rotational angles based on a single vertically mounted fixed-focus visual camera. Firstly, an image preprocessing method is proposed for complex port environments. The improved YOLOv5 network, enhanced with an attention mechanism, increases the detection accuracy of the spreader's keypoints and the container lock holes. Combined with image morphological processing methods, the three-dimensional position and rotational angle changes of the spreader are measured. Compared to traditional detection methods, the single-camera-based method for three-dimensional positioning and attitude measurement of the spreader employed in this paper achieves higher detection accuracy for spreader keypoints and lock holes in experiments and improves the operational speed of single operations in actual tests, making it a feasible measurement approach. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
35. Vision sensor based residual vibration suppression strategy of non-deformable object for robot-assisted assembly operation with gripper flexibility
- Author
-
Jalendra, Chetan, Rout, B.K., and Marathe, Amol
- Published
- 2022
- Full Text
- View/download PDF
36. Combining Multiple Cues for Visual Madlibs Question Answering.
- Author
-
Tommasi, Tatiana, Mallya, Arun, Plummer, Bryan, Lazebnik, Svetlana, Berg, Alexander C., and Berg, Tamara L.
- Subjects
- *
PAPER , *VISUAL analytics , *REASONING , *INTEGRATION (Theory of knowledge) , *COMPUTER vision - Abstract
This paper presents an approach for answering fill-in-the-blank multiple choice questions from the Visual Madlibs dataset. Instead of generic and commonly used representations trained on the ImageNet classification task, our approach employs a combination of networks trained for specialized tasks such as scene recognition, person activity classification, and attribute prediction. We also present a method for localizing phrases from candidate answers in order to provide spatial support for feature extraction. We map each of these features, together with candidate answers, to a joint embedding space through normalized canonical correlation analysis (nCCA). Finally, we solve an optimization problem to learn to combine scores from nCCA models trained on multiple cues to select the best answer. Extensive experimental results show a significant improvement over the previous state of the art and confirm that answering questions from a wide range of types benefits from examining a variety of image cues and carefully choosing the spatial support for feature extraction. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
37. Showthrough and Strikethrough print defect detection using histogram equalization based computer vision method.
- Author
-
Saha, Jayeeta, Naskar, Shilpi, and Maiti, Sayanti
- Subjects
IMAGE enhancement (Imaging systems) ,HISTOGRAMS ,PICTURES ,COMPUTER vision ,THRESHOLDING algorithms ,PIXELS - Abstract
This paper presents a comparatively simple approach for showthrough and strikethrough print defect detection using computer vision method. Showthrough and strikethrough are common printing problem and are typically functions of a paper's opacity. Under normal lighting condition the visibility of printing on the reverse side of printed paper is termed as showthrough whereas the penetration of ink to the other side is termed as strikethrough. Moreover the intensity of showthrough pixel is extremely low thus it is difficult to identify the showthrough pixel from the printed area. On the other hand strikethrough is the result of penetration of ink through paper and depends on the absorbent nature of paper. Comparatively the intensity of the strikethrough pixel is higher than that of the showthrough but due to similar intensity of the ink of the printed pixel and strikethrough pixel, both overlapped with each other in the foreground of the image. These print defects can degrade the image quality as well as print production. In this study, the detection of these two print defects achieved using histogram equalization technique, to enhance the contrast between foreground and back ground pixels. A global thresholding algorithm was applied on a histogram equalized image to segment the printed area from the background of the image. Pixels in the background which are considered as showthrough and strike through pixels are identified by image subtraction. The pictorial representations of the results show the remarkable potential of the proposed technique which can be possible alternative of present subjective measures of showthrough and strikethrough. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
38. Effects on the Ocular Surface from Reading on Different Smartphone Screens: A Prospective Randomized Controlled Study.
- Author
-
Yuan, Kelan, Zhu, Haiping, Mou, Yujie, Wu, Yaying, He, Jingliang, Huang, Xiaodan, and Jin, Xiuming
- Subjects
- *
ELECTRONIC paper , *SMARTPHONES , *MEIBOMIAN glands , *COMPUTER vision , *LIGHT emitting diodes , *CORNEA , *BENZALKONIUM chloride - Abstract
The purpose of this study was to investigate the influence of smartphone reading on the ocular surface and to compare the various effects of different screens and light conditions on the ocular surface. One hundred nineteen volunteers were randomly divided into: light + organic light‐emitting diode (OLED), light + electronic ink (eINK), dark + OLED, and dark + eINK. Ocular surface examinations, including noninvasive break‐up time (NIBUT), noninvasive keratograph tear meniscus height (NIKTMH), ocular redness, fluorescein break‐up time (FBUT), corneal fluorescein staining, meibomian gland assessment, Schirmer I Test, and blinking frequency, were performed before and after a reading task. Symptoms were evaluated using the Ocular Surface Disease Index (OSDI) and Computer Vision Syndrome Questionnaire (CVS‐Q). NIBUT and FBUT were decreased statistically significantly after participants read on an OLED screen for 2 hours compared with the baseline in light and dark environments, whereas no statistically significant decrease was observed on an eINK screen. NIKTMH was statistically significantly decreased after reading on an OLED screen in light and dark settings, and the eINK screen had a lesser effect on NIKTMH. An obvious increase in the ocular redness, OSDI and CVS‐Q scores was observed after reading on an OLED screen, whereas the eINK screen had a lesser effect on these indicators. Blink rate increased gradually in OLED subgroups during the reading task, whereas no statistically significant difference was observed in the eINK subgroups. Our research suggested that reading on an OLED screen can cause ocular surface disorder and obvious subjective discomfort, whereas reading on an eINK screen can minimize ocular surface disorder in both dark and light environments. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
39. Watching plants grow – a position paper on computer vision and Arabidopsis thaliana.
- Author
-
Bell, Jonathan and Dee, Hannah M.
- Subjects
- *
IMAGE processing , *COMPUTER vision , *ARABIDOPSIS thaliana , *IMAGE analysis , *ANGIOSPERMS - Abstract
The authors present a comprehensive overview of image processing and analysis work done to support research into the model flowering plant Arabidopsis thaliana. Beside the plant's importance in biological research, using image analysis to obtain experimental measurements of it is an interesting vision problem in its own right, involving the segmentation and analysis of sequences of images of objects whose shape varies between individual specimens and also changes over time. While useful measurements can be obtained by segmenting a whole plant from the background, they suggest that the increased range and precision of measurements made available by leaf-level segmentation makes this a problem well worth solving. A variety of approaches have been tried by biologists as well as computer vision researchers. This is an interdisciplinary area and the computer vision community has an important contribution to make. They suggest that there is a need for publicly available datasets with ground truth annotations to enable the evaluation of new approaches and to support the building of training data for modern data-driven computer vision approaches, which are those most likely to result in the kind of fully automated systems that will be of use to biologists. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
40. Research on defect detection of bottle cap interior based on low-angle and large divergence angle vision system.
- Author
-
Chen, Bowen, Li, Chen, Yuan, Pi, Yan, Yujie, and Yin, Yongjing
- Subjects
COLOR space ,COMPUTER vision ,ANGLES ,BOTTLES ,INDUSTRIAL capacity - Abstract
During the machine vision inspection of the inner section of bottle caps within pharmaceutical packaging, the unique conca bottom and convex side walls often create obstructions to the illumination. Consequently, this results in challenges such as irregular background and diminished feature contrast in the image, ultimately leading to the misidentification of defects. As a solution, a vision system characterized by a Low-Angle and Large Divergence Angle (LALDA) is presented in this paper. Using the large divergence angle of LED, combined with low-angle illumination, a uniform image of the side wall region with bright-field characteristics and a uniform image of inner circle region at the bottom with dark-field characteristics are obtained, thus solving the problems of light being obscured and brightness overexposure of the background. Based on the imaging characteristics of LALDA, a multi-channel segmentation (MCS) algorithm is designed. The HSV color space has been transformed, and the image is automatically segmented into multiple sub-regions by mutual calculation of different channels. Further, image homogenization and enhancement are used to eliminate fluctuations in the background and to enhance the contrast of defects. In addition, a variety of defect extraction methods are designed based on the imaging characteristics of different sub-regions, which can avoid the problem of over-segmentation in detection. In this paper, the LALDA is applied to the defect detection inside the cap of capsule medicine bottle, the detection speed is better than 400 pcs/min and the detection accuracy is better than 95%, which can meet the actual production line capacity and detection requirements. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
41. A study on the application of convolutional neural networks for the maintenance of railway tracks.
- Author
-
Pappaterra, Mauro José, Pappaterra, María Lucía, and Flammini, Francesco
- Subjects
CONVOLUTIONAL neural networks ,COMPUTER vision ,LITERATURE reviews ,JOINT use of railroad facilities ,INTELLIGENT sensors ,BALLAST (Railroads) - Abstract
This paper provides an overview of the applications of Convolutional Neural Networks (CNN) in the railway maintenance industry. Our research covers specifically the subdomain of railway track maintenance. In this study, we have analyzed the state-of-the-art of CNNs applied to railway track maintenance by conducting an extensive literature review, summarizing different tasks and problems related to the topic and presenting solutions based on CNNs with a special emphasis on the data used to create these models. The results of our research show different applications of CNNs within the scope, including the detection of defects in the surface of railway rails and railway track components, such as fasteners, joints, sleepers, switches and crossings, as well as the recognition of track components, and the continuous monitoring of railway tracks. The architecture of CNNs is fitting to learning spatial hierarchies of features directly from the input data, making them of great use for Computer Vision and other applications related to the topic at hand. The implementation of IoT devices and smart sensors aid the collection of real-time data which can be used to feed powerful CNN models to recognize patterns and identify complex events related to the maintenance of railway tracks. This and more insights are discussed in detail within the contents of this paper. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
42. Hierarchical Spatial Sum–Product Networks for Action Recognition in Still Images.
- Author
-
Wang, Jinghua and Wang, Gang
- Subjects
- *
PAPER construction , *PAPER textiles , *DATA analysis , *SPATIAL analysis (Statistics) , *BIG data - Abstract
Recognizing actions from still images has been popularly studied recently. In this paper, we model an action class as a flexible number of spatial configurations of body parts by proposing a new spatial sum–product network (SPN). First, we discover a set of parts in image collections via unsupervised learning. Then, our new spatial SPN is applied to model the spatial relationship and also the high-order correlations of parts. To learn robust networks, we further develop a hierarchical spatial SPN method, which models pairwise spatial relationship between parts inside subimages and models the correlation of subimages via extra layers of SPN. Our method is shown to be effective on two benchmark data sets. [ABSTRACT FROM PUBLISHER]
- Published
- 2018
- Full Text
- View/download PDF
43. Automatic image-based detection and inspection of paper fibres for grasping.
- Author
-
Hirvonen, Juha and Kallio, Pasi
- Subjects
- *
COMPUTER vision , *COMPUTER algorithms , *IMAGE segmentation , *COMPARATIVE studies , *COMPUTER science - Abstract
An automatic computer vision algorithm that detects individual paper fibres from an image, assesses the possibility of grasping the detected fibres with microgrippers and detects the suitable grasping points is presented. The goal of the algorithm is to enable automatic fibre manipulation for mechanical characterisation, which has traditionally been slow manual work. The algorithm classifies the objects in images based on their morphology, and detects the proper grasp points from the individual fibres by applying given geometrical constraints. The authors test the ability of the algorithm to detect the individual fibres with 35 images containing more than 500 fibres in total, and also compare the graspability analysis and the calculated grasp points with the results of an experienced human operator with 15 images that contain a total of almost 200 fibres. The detection results are outstanding, with fewer than 1% of fibres missed. The graspability analysis gives sensitivity of 0.83 and specificity of 0.92, and the average distance between the grasp points of the human and the algorithm is 220 µm. Also, the choices made by the algorithm are much more consistent than the human choices. [ABSTRACT FROM AUTHOR]
- Published
- 2015
- Full Text
- View/download PDF
44. IJAIT Topics and Papers in the Third Decade.
- Author
-
Liu, Alan
- Subjects
- *
DECISION support systems , *COMPUTER vision , *ARTIFICIAL intelligence , *TRAVELING salesman problem - Published
- 2022
- Full Text
- View/download PDF
45. A lightweight dual-attention network for tomato leaf disease identification.
- Author
-
Enxu Zhang, Ning Zhang, Fei Li, and Cheng Lv
- Subjects
MACHINE learning ,COMPUTER vision ,RICE diseases & pests ,IMAGE recognition (Computer vision) ,PLANT diseases ,DIGITAL-to-analog converters ,DEEP learning - Abstract
Tomato disease image recognition plays a crucial role in agricultural production. Today, whilemachine vision methods based on deep learning have achieved some success indiseaserecognition, theystill faceseveralchallenges. These include issues such as imbalanced datasets, unclear disease features, small inter-class differences, and large intra-class variations. To address these challenges, this paper proposes a method for classifying and recognizing tomato leaf diseases based on machine vision. First, to enhance the disease feature details in images, a piecewise linear transformation method is used for image enhancement, and oversampling is employed to expand the dataset, compensating for the imbalanced dataset. Next, this paper introduces a convolutional block with a dual attention mechanismcalled DAC Block, which is used to construct a lightweight model named LDAMNet. The DAC Block innovatively uses Hybrid Channel Attention (HCA) and Coordinate Attention (CSA) to process channel information and spatial information of input images respectively, enhancing the model's feature extraction capabilities. Additionally, this paper proposes a Robust Cross-Entropy (RCE) loss function that is robust tonoisylabels, aimedat reducing the impact of noisy labels on the LDAM Net model during training. Experimental results show that this method achieves an average recognition accuracy of 98.71% on the tomato disease dataset, effectively retaining disease information in images and capturing disease areas. Furthermore, the method also demonstrates strong recognition capabilities on rice crop disease datasets, indicating good generalization performance and the ability to function effectively in disease recognition across different crops. The research findings of this paper provide new ideas and methods for the field of crop disease recognition. However, future research needs to further optimize the model's structure and computational efficiency, and validate its application effects in more practical scenarios. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
46. Detecting Steam Leakage in Nuclear Power Systems Based on the Improved Background Subtraction Method.
- Author
-
Liu, Jie, Huang, Yanping, Zhang, Minglu, Zhou, Suting, Nie, Changhua, Li, Minggang, and Zhang, Lin
- Subjects
LEAK detection ,NUCLEAR energy ,COMPUTER vision ,PRESSURE gages ,GRAYSCALE model - Abstract
As a key system in nuclear power plants, nuclear power systems contain high-temperature, high-pressure water media. A steam leak, if it occurs, can at minimum cause system functional loss and at worst lead to casualties. Therefore, it is urgent to carry out steam leakage detection work for high-temperature, high-pressure loop systems. Currently, steam leaks are primarily detected through visual monitoring and pressure gauges. However, if there is a minor leak under high system pressure, the slight decrease in pressure may not be enough to alert the operators, leading to a delay in detecting the steam leak. Thus, this detection method has certain drawbacks. In view of these issues, this paper introduces computer vision technology to monitor the high-temperature, high-pressure loop system and proposes the use of an improved background subtraction method to detect steam leaks in the loop system. The results show the following advantages of this method: (1) It can effectively identify steam leaks at an early stage; (2) it overcomes the difficulty of determining the threshold value for the binarization of grayscale images in traditional background subtraction methods; (3) it eliminates the noise impact brought by the binarization of grayscale images in existing improved background subtraction methods. The introduction of this method provides a new approach for detecting steam leaks in high-temperature, high-pressure loop systems and can be effectively applied in engineering fields. It also offers reference value for the detection of high-temperature, high-pressure media leaks in other fields. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
47. Frequency roughness analysis in image processing and game design.
- Author
-
Li, Jiaqi
- Subjects
IMAGE processing ,IMAGE analysis ,PATTERN recognition systems ,COMPUTER vision ,IMAGE intensifiers ,DIGITAL image processing ,CONCEPT mapping - Abstract
With the continuous progress of science and technology, image processing techniques have been used increasingly in recent years. Image processing plays an indispensable role in the fields of computer vision, artificial intelligence, pattern recognition, and related fields. Improvements in basic algorithms and the development of new algorithms have resulted in considerable innovation and progress. This paper is devoted to finding new game applications in a branch of image processing. It introduces an analysis model proposed by the author and discusses the relationship between roughness in the frequency domain and visual image interpretation. By using the concept of roughness, we separated the image features into meaningful information and residual information and analysed the image in the frequency domain. The results were compared with those of traditional image processing methods. The starting point is the visual identification of a feature based on human interpretation. The image information was separated into meaningful features and the residual component to reduce the redundancy of the model. This allowed for a sparse representation of the feature information in the image. By analysing the meaningful features and residual components of an image separately, we established a relationship between the results and the original images. Parameters such as texture, morphology, and the degree of blurring were considered and we developed a parameter called "frequency roughness". The algorithm incorporates the concepts of frequency and roughness and the roughness is determined in the frequency domain. The frequency roughness algorithm successfully separated the rough features in the frequency domain and calculated the residual value in an image. This model provided more accurate image processing results than comparable methods. This paper includes an analysis and game applications of the proposed model for de-blurring, image enhancement, recognition, and other image processing tasks. Some game applications were successful, whereas others require further investigation. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
48. Pose estimation algorithm based on point pair features using PointNet + +.
- Author
-
Chen, Yifan, Li, Zhenjian, Li, Qingdang, and Zhang, Mingyue
- Subjects
MACHINE learning ,POINT cloud ,COMPUTER vision ,FIX-point estimation ,BASE pairs ,DEEP learning - Abstract
This study proposes an innovative deep learning algorithm for pose estimation based on point clouds, aimed at addressing the challenges of pose estimation for objects affected by the environment. Previous research on using deep learning for pose estimation has primarily been conducted using RGB-D data. This paper introduces an algorithm that utilizes point cloud data for deep learning-based pose computation. The algorithm builds upon previous work by integrating PointNet + + technology and the classical Point Pair Features algorithm, achieving accurate pose estimation for objects across different scene scales. Additionally, an adaptive parameter-density clustering method suitable for point clouds is introduced, effectively segmenting clusters in varying point cloud density environments. This resolves the complex issue of parameter determination for density clustering in different point cloud environments and enhances the robustness of clustering. Furthermore, the LineMod dataset is transformed into a point cloud dataset, and experiments are conducted on the transformed dataset to achieve promising results with our algorithm. Finally, experiments under both strong and weak lighting conditions demonstrate the algorithm's robustness. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
49. A review of machine vision pose measurement.
- Author
-
Wang Xiaoxiao, Ng Seng Beng, Rahmat, Rahmita Wirza O. K., and Sulaima, Puteri Suhaiza
- Subjects
COMPUTER vision ,VISUAL fields ,MACHINE performance ,MACHINE learning ,RESEARCH personnel - Abstract
This review paper provides a comprehensive overview of machine vision pose measurement algorithms. The paper focuses on the state-of-the-art algorithms and their applications. The paper is structured as follows: the introduction in provides a brief overview of the field of machine vision pose measurement. Describes the commonly used algorithms for machine vision pose measurement. Discusses the factors that affect the accuracy and reliability of machine vision pose measurement algorithms. Summarizes the paper and provides future research directions. The paper highlights the need for more robust and accurate algorithms that can handle varying lighting conditions and occlusion. It also suggests that the integration of machine learning techniques may improve the performance of machine vision pose measurement algorithms. Overall, this review paper provides a comprehensive overview of machine vision pose measurement algorithms, their applications, and the factors that affect their accuracy and reliability. It provides a valuable resource for researchers and practitioners working in the field of computer vision. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
50. Flexible thin parts multi‐target positioning method of multi‐level feature fusion.
- Author
-
Deng, Yaohua, Liu, Xiali, Yang, Kenan, and Li, Zehang
- Subjects
IMAGE fusion ,COMPUTER vision ,IMAGE recognition (Computer vision) ,NICKEL-plating ,GAUSSIAN processes - Abstract
In new energy battery manufacturing, machine vision is widely used in automated assembly scenarios for key parts. To improve the accuracy and real‐time multi‐target positioning recognition of flexible thin parts, this paper proposes a multi‐level feature fusion template matching algorithm based on the Gaussian pyramid. Firstly, the algorithm constructs a Gaussian pyramid by multi‐scale image construction. Secondly, considering the image features of each layer of the pyramid, this paper uses the grey‐based Fast Normalized Matching algorithm to obtain coarse positioning coordinates on the upper layer, and the improved Linemod‐2D algorithm is applied to the bottom layer image to get accurate positioning coordinates. Finally, the positioning coordinates returned from each layer are fused to obtain the final positioning coordinate. The experimental results show that the proposed algorithm achieves excellent performance in nickel plate positioning and recognition. It exhibits satisfactory performance in nickel sheet localization and recognition. In terms of angular error, repeat accuracy, and matching speed, it competes favourably with Halcon, VisionMaster, and SCISmart. Its positioning error closely approximates that of Halcon, effectively meeting the practical production demands for high‐speed feeding and high‐precision positioning. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.