98 results on '"Image map"'
Search Results
2. Robust color image hashing using convolutional stacked denoising auto-encoders for image authentication
- Author
-
Madhumita Paul, Arnab Jyoti Thakuria, Fazal Ahmed Talukdar, and Ram Kumar Karsh
- Subjects
0209 industrial biotechnology ,Authentication ,Color image ,Computer science ,business.industry ,Noise reduction ,Image map ,Hash function ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,Image (mathematics) ,020901 industrial engineering & automation ,Artificial Intelligence ,Robustness (computer science) ,Distortion ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,business ,Software - Abstract
Image authentication based on image hashing has gained large attention in recent years. However, limited work has been done in color image hashing. Also, most of the existing methods are unable to detect tampering, if the composite rotation–scaling–translation (RST) distortion and tampering in a color image occur simultaneously. In this paper, an image hashing technique has been proposed based on convolutional stacked denoising auto-encoders (CSDAEs). In addition, a blind geometric correction approach is used to correct the composite RST distortion in the image. An input image is hierarchically mapped to a lower-dimensional hash code via CSDAEs, which have been trained for content-preserving operations (CPOs). An image map is generated from the hash via the decoder. The tampered area has been localized, by comparing the image map of hash codes from the reference image and the received image. The experimental results show that the proposed method is robust against most of the CPOs, especially to composite RST, a better trade-off between robustness and discrimination, and can localize the tampered regions. The receiver operating characteristics show that the proposed model is better than some of the state-of-the-art methods.
- Published
- 2021
3. Digital Analysis of Geo-Referenced Concrete Scanning Electron Microscope (SEM) Images
- Author
-
Elly Nur Myaisara Maizul and Mohd Sanusi S. Ahamad
- Subjects
Materials science ,geo-referenced image ,Scanning electron microscope ,Image map ,image mapping ,0211 other engineering and technologies ,Digital analysis ,02 engineering and technology ,General Medicine ,scanning electron microscope ,021001 nanoscience & nanotechnology ,lcsh:Environmental engineering ,021105 building & construction ,Digital image analysis ,digital image analysis ,lcsh:TA170-171 ,0210 nano-technology ,Business management ,Remote sensing - Abstract
The microstructural evaluation of complex cementitious materials has been made possible by the microscopic imaging tools such as Scanning Electron Microscope (SEM) and X-Ray Microanalysis. Particularly, the application of concrete SEM imaging and digital image analysis have become common in the analysis and mapping of concrete technology. In this study, six samples of two-dimensional (2D) SEM images were spatially resampled to produce Geo-referenced SEM sample images. Subsequently, they were analyzed and the intensity histogram plot was produced to facilitate visual interpretation. The consecutive digital image analysis performed was the enhancement and noise removal process using two filtering methods i.e. median and adaptive box filter. The filtered resampled images, then undergone the unsupervised K-Means classification process to collectively separate each individual pixel corresponds to the spectral data. By spatial segmentation of K-Means algorithms, the cluster groups generated were carefully reviewed before proceeding to the final analysis. From the resulting data, the mapping of the spatial distribution of k-cluster and the quantification of micro-cracks (voids) were performed. The results of the SEM images (1st - 4th sample) showed a higher percentage of k-cluster data indicating a good correlation with the major elemental composition of EDX analysis, namely Oxide (O), Silicon (Si) and Carbon (C). Meanwhile, the subjective visual assessment of the image (5th and 6th sample) has confirmed the micro-crack developments on the concrete SEM images upon which the crack density was 3.02 % and 1.30 %, respectively.
- Published
- 2020
4. Sensor-Level Mosaic of Multistrip KOMPSAT-3 Level 1R Products
- Author
-
Changno Lee and Jaehong Oh
- Subjects
Technology ,010504 meteorology & atmospheric sciences ,QH301-705.5 ,Computer science ,QC1-999 ,Image map ,0211 other engineering and technologies ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Field of view ,02 engineering and technology ,Shuttle Radar Topography Mission ,01 natural sciences ,KOMPSAT-3A ,strip ,discrepancy ,Focal length ,General Materials Science ,Biology (General) ,QD1-999 ,Instrumentation ,Image resolution ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences ,Remote sensing ,Fluid Flow and Transfer Processes ,Physics ,Process Chemistry and Technology ,matching ,General Engineering ,Engineering (General). Civil engineering (General) ,Computer Science Applications ,RPCs ,Chemistry ,Line (geometry) ,Outlier ,sensor modeling ,Satellite ,mosaic ,TA1-2040 - Abstract
High-resolution satellite images such as KOMPSAT-3 data provide detailed geospatial information over interest areas that are evenly located in an inaccessible area. The high-resolution satellite cameras are designed with a long focal length and a narrow field of view to increase spatial resolution. Thus, images show relatively narrow swath widths (10–15 km) compared to dozens or hundreds of kilometers in mid-/low-resolution satellite data. Therefore, users often face obstacles to orthorectify and mosaic a bundle of delivered images to create a complete image map. With a single mosaicked image at the sensor level delivered only with radiometric correction, users can process and manage simplified data more efficiently. Thus, we propose sensor-level mosaicking to generate a seamless image product with geometric accuracy to meet mapping requirements. Among adjacent image data with some overlaps, one image is the reference, whereas the others are projected using the sensor model information with shuttle radar topography mission. In the overlapped area, the geometric discrepancy between the data is modeled in spline along the image line based on image matching with outlier removals. The new sensor model information for the mosaicked image is generated by extending that of the reference image. Three strips of KOMPSAT-3 data were tested for the experiment. The data showed that irregular image discrepancies between the adjacent data were observed along the image line. This indicated that the proposed method successfully identified and removed these discrepancies. Additionally, sensor modeling information of the resulted mosaic could be improved by using the averaging effects of input data.
- Published
- 2021
- Full Text
- View/download PDF
5. Adjusted JPEG Quantization Tables in Support of GPS Maps
- Author
-
Yair Wiseman
- Subjects
0303 health sciences ,Computer science ,Image quality ,business.industry ,Communication ,Image map ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Data_CODINGANDINFORMATIONTHEORY ,02 engineering and technology ,computer.file_format ,021001 nanoscience & nanotechnology ,Table (information) ,JPEG ,Industrial and Manufacturing Engineering ,03 medical and health sciences ,Media Technology ,Chrominance ,Global Positioning System ,0210 nano-technology ,Quantization (image processing) ,business ,computer ,Algorithm ,030304 developmental biology ,Image compression - Abstract
Image quality manipulating in JPEG is done by quantization tables. JPEG has two quantization tables – one table for the luminance information and one table for the chrominance information. These quantization tables have been designed in support of images with few sharp changes; however, typically most GPS image maps have many sharp changes and as a result, the images are not optimally compressed. The designers of the quantization tables have presumed that sharp changes in the colors will rarely occur. Therefore, they divide the values that represent sharp changes in the frequency space by large numbers and divide other values by smaller numbers. As a result, when there are sharp changes in an image, the proportional allocation for each kind of data in the compressed image is inappropriate and results in an inefficient compression. In this paper the standard quantization tables have been modified as to handle the different kinds of GPS image map data appropriately. Consequently, the experimental results show that images with sharp changes are compressed more efficiently when making use of the new quantization tables.
- Published
- 2021
6. Exploring BCI Control in Smart Environments: Intention Recognition Via EEG Representation Enhancement Learning
- Author
-
Sen Wang, Xiaowei Zhao, Hao Shen, Weitong Chen, Guodong Long, Lin Yue, and Robert J. Boots
- Subjects
General Computer Science ,Computer science ,business.industry ,Interface (computing) ,Image map ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Pattern recognition ,02 engineering and technology ,03 medical and health sciences ,0302 clinical medicine ,Motor imagery ,Feature (computer vision) ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Smart environment ,Artificial Intelligence & Image Processing ,Noise (video) ,Artificial intelligence ,Representation (mathematics) ,business ,030217 neurology & neurosurgery ,0801 Artificial Intelligence and Image Processing, 0806 Information Systems ,Brain–computer interface - Abstract
The brain–computer interface (BCI) control technology that utilizes motor imagery to perform the desired action instead of manual operation will be widely used in smart environments. However, most of the research lacks robust feature representation of multi-channel EEG series, resulting in low intention recognition accuracy. This article proposes an EEG2Image based Denoised-ConvNets (called EID) to enhance feature representation of the intention recognition task. Specifically, we perform signal decomposition, slicing, and image mapping to decrease the noise from the irrelevant frequency bands. After that, we construct the Denoised-ConvNets structure to learn the colorspace and spatial variations of image objects without cropping new training images precisely. Toward further utilizing the color and spatial transformation layers, the colorspace and colored area of image objects have been enhanced and enlarged, respectively. In the multi-classification scenario, extensive experiments on publicly available EEG datasets confirm that the proposed method has better performance than state-of-the-art methods.
- Published
- 2021
7. Achieving a reversible lower dimensionality transformation for picture archiving and communication system in healthcare
- Author
-
Ab Rouf Khan, Shoaib Amin Banday, and Mohammad Khalid Pandit
- Subjects
Computer science ,Image map ,020208 electrical & electronic engineering ,02 engineering and technology ,Communications system ,Picture archiving and communication system ,Computer engineering ,Transmission (telecommunications) ,Image texture ,0202 electrical engineering, electronic engineering, information engineering ,Visual communication ,Electrical and Electronic Engineering ,Texture mapping ,Image compression ,Data compression - Abstract
With the progression of picture archiving and communication systems (PACSs) over the past decade, it has become imperative that such systems be optimised in security, storage, and transmission aspects. The work presented in this Letter shows a framework for medical image compression and secure image transmission for PACSs. The work aims to achieve a lower dimensionality of input medical image signified by a high-compression ratio, a secure image transmission that can withstand adversarial attacks and provide a reversible reconstruction with minimal error. The authors illustrate that sinusoid modulated Gaussian texture maps, multi-level chaotic maps, and high-frequency image maps can be efficiently fused and utilised in a deep learning architecture. The overall analysis depicts promising results with regard to the capability of image compression, security, and transmission. The proposed framework will be a potential candidate for use in PACSs, which effectively is the backbone of the current healthcare paradigm.
- Published
- 2020
8. Key Parameter Identification and Defective Wafer Detection of Semiconductor Manufacturing Processes Using Image Processing Techniques
- Author
-
Fei He, Chih-Hung Jen, Jui-Yu Huang, Du-Ming Tsai, and Shu-Kai S. Fan
- Subjects
0209 industrial biotechnology ,Computer science ,Semiconductor device fabrication ,Image map ,Process (computing) ,Image processing ,02 engineering and technology ,Condensed Matter Physics ,computer.software_genre ,Industrial and Manufacturing Engineering ,Fault detection and isolation ,Electronic, Optical and Magnetic Materials ,020901 industrial engineering & automation ,Key (cryptography) ,Wafer ,Data mining ,Electrical and Electronic Engineering ,computer ,Advanced process control - Abstract
The semiconductor industry has become fully automated during the manufacturing process and abundant process parameters are collected online by sensors for fault detection and classification purposes. Analyzing process parameters and identifying a smaller set of key parameters that have crucial influence on wafer quality will bring great benefits in stabilizing the manufacturing process and enhancing productive yield. Typically, this type of the parameter set is called the “raw trace data.” This paper considers image processing techniques as a novel approach for analyzing and visualizing the raw trace data. First, the 1-D time series data of a wafer batch was transformed into a 2-D image. Fisher’s criterion ratios of the labeled good and defective wafer image maps are computed to identify the key parameters. The key parameters identified by the proposed image processing technique are consistent with the technical experience of the process engineers. Furthermore, the texture analysis technique with 2-D Fourier transform is utilized to analyze the images of the key parameters to detect defective wafers. The proposed key parameter identification and wafer classification method proves to be a viable solution under the paradigm of advanced process control practice for semiconductor manufacturing.
- Published
- 2019
9. Deictic Image Mapping: An Abstraction for Learning Pose Invariant Manipulation Policies
- Author
-
Marcus Gualtieri, Robert W. Platt, and Colin Kohler
- Subjects
0209 industrial biotechnology ,business.industry ,Computer science ,Image map ,Robotics ,02 engineering and technology ,General Medicine ,Deixis ,020901 industrial engineering & automation ,0202 electrical engineering, electronic engineering, information engineering ,Reinforcement learning ,020201 artificial intelligence & image processing ,Artificial intelligence ,Invariant (mathematics) ,business - Abstract
In applications of deep reinforcement learning to robotics, it is often the case that we want to learn pose invariant policies: policies that are invariant to changes in the position and orientation of objects in the world. For example, consider a pegin-hole insertion task. If the agent learns to insert a peg into one hole, we would like that policy to generalize to holes presented in different poses. Unfortunately, this is a challenge using conventional methods. This paper proposes a novel state and action abstraction that is invariant to pose shifts called deictic image maps that can be used with deep reinforcement learning. We provide broad conditions under which optimal abstract policies are optimal for the underlying system. Finally, we show that the method can help solve challenging robotic manipulation problems.
- Published
- 2019
10. Compressed Sensing Image Mapping Spectrometer
- Author
-
Xiaoming Ding
- Subjects
Diffraction ,reconstruction ,General Computer Science ,Computer science ,Image map ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Imaging spectrometer ,02 engineering and technology ,01 natural sciences ,010309 optics ,Data cube ,0103 physical sciences ,General Materials Science ,Computer vision ,Image resolution ,compressed sensing ,Snapshot imaging spectrometer ,business.industry ,Detector ,General Engineering ,Image mapping spectrometer ,021001 nanoscience & nanotechnology ,datacube ,Compressed sensing ,lcsh:Electrical engineering. Electronics. Nuclear engineering ,Artificial intelligence ,0210 nano-technology ,business ,lcsh:TK1-9971 - Abstract
This paper presents a novel snapshot imaging spectrometer based on the image mapping and compressed sensing concept named Compressed Sensing Image Mapping Spectrometer (CSIMS). The operation principle is to slice the input image to different directions and encode the strip pieces before dispersion by a prism. The detector obtains the mixture spatial-spectral data simultaneously. The datacube is reconstructed by the compressed sensing algorithm and combining all the pieces together. The mathematical model of CSIMS is established to describe the light wave propagation through the entire system based on the scalar diffraction theory. The simulations are conducted to prove the effectiveness of the CSIMS principle, and the results show that the reconstructed datacube reveals higher spatial resolution and more accurate spectral curves than that of the relative snapshot imaging spectrometer based on compressed sensing.
- Published
- 2019
11. An Intelligent Photographing Guidance System Based on Compositional Deep Features and Intepretable Machine Learning Model
- Author
-
Chin-Shyurng Fahn, Meng-Luen Wu, and Sheng-Kuei Tsau
- Subjects
0209 industrial biotechnology ,Computer science ,business.industry ,media_common.quotation_subject ,Deep learning ,Image map ,Photography ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,Machine learning ,computer.software_genre ,Image (mathematics) ,03 medical and health sciences ,Tree (data structure) ,020901 industrial engineering & automation ,0302 clinical medicine ,030220 oncology & carcinogenesis ,Quality (business) ,Artificial intelligence ,business ,Guidance system ,Classifier (UML) ,computer ,media_common - Abstract
Photography is the activity of recording precious moments which are often difficult to make up afterwards. Therefore, taking the correct picture under proper guidance assistance is important. Although there are many factors that can determine a good photo, in general, photos that do not follow the composition rules usually look bad that make the viewer feel uncomfortable. Acting as a solution, in this paper, we propose an intelligent photographing guidance system using machine learning. The guidance is based on a tree-based interpretable machine learning model that can give reasons for decisions. There are two categories of features for guidance, which are traditional image features and deep features. Traditional features include prominent lines and image maps, such as saliency map and sharpness map, each of which exists in a multi-scale Gaussian pyramid. Deep features are extracted during the establishment of a CNN-based image composition classifier. We use these two categories of features as inputs for the interpretable machine learning model to establish a feasible photographing guidance system. The guidance system references our composition classifier with precision rate of 94.8%, and recall rate of 95.0% where the comprising tree-based interpretable model is capable of guiding camera users to alter image contents for obtaining better aesthetical compositions to take photos of good quality.
- Published
- 2021
12. Effect of Neighbourhood Size in Entropy Mapping of Ultrasound Images
- Author
-
Dinesh Kumbhare, Ryan G. L. Koh, Margaret Kirkwood, and Michael Behr
- Subjects
Pixel ,business.industry ,Image map ,Entropy ,0206 medical engineering ,Ultrasound ,Healthy subjects ,Pattern recognition ,02 engineering and technology ,020601 biomedical engineering ,Grayscale ,03 medical and health sciences ,0302 clinical medicine ,Image Processing, Computer-Assisted ,Entropy (information theory) ,Humans ,Artificial intelligence ,business ,Statistical Effect Size ,030217 neurology & neurosurgery ,Randomness ,Mathematics ,Ultrasonography - Abstract
Image filtering is a technique that can create additional visual representations of the original image. Entropy filtering is a specific application that can be used to highlight randomness of pixel grayscale intensities within an image. These image map created from filtering are based on the number of surrounding neighbourhood of pixels considered. However, there is no standard procedure for determining the correct "neighbourhood size" to use. We investigated the effects of neighbourhood size on the entropy calculation and provide a standardized approach for determining an appropriate neighbourhood size in entropy filtering in a musculoskeletal application. Ten healthy subjects showing no symptoms related to neuromuscular disease were recruited and ultrasound images of their trapezius muscle were acquired. The muscle regions in the images were manually isolated and regions of interest with varying neighbourhood sizes (increasing by 2 pixels) from 3x3 to 61X61 pixels were extracted. The entropy, relative signal entropy over noise entropy, statistical effect size as well as the percentage change of the effect size and instantaneous slope of the effect size was examined. The analysis showed that a neighbourhood size within the range of 21-25 pixels provides the maximum amount of information gained and coincides with a percentage change of the effect size of less than 5% and instantaneous slopes < 0.05.
- Published
- 2020
13. UCTGAN: Diverse Image Inpainting Based on Unsupervised Cross-Space Translation
- Author
-
Haibo Chen, Qihang Mo, Dongming Lu, Wei Xing, Zhizhong Wang, Sihuan Lin, Zhiwen Zuo, and Lei Zhao
- Subjects
Image quality ,business.industry ,Computer science ,Image map ,Inpainting ,Pattern recognition ,02 engineering and technology ,010501 environmental sciences ,Translation (geometry) ,01 natural sciences ,Image (mathematics) ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,Layer (object-oriented design) ,business ,Projection (set theory) ,Image restoration ,0105 earth and related environmental sciences - Abstract
Although existing image inpainting approaches have been able to produce visually realistic and semantically correct results, they produce only one result for each masked input. In order to produce multiple and diverse reasonable solutions, we present Unsupervised Cross-space Translation Generative Adversarial Network (called UCTGAN) which mainly consists of three network modules: conditional encoder module, manifold projection module and generation module. The manifold projection module and the generation module are combined to learn one-to-one image mapping between two spaces in an unsupervised way by projecting instance image space and conditional completion image space into common low-dimensional manifold space, which can greatly improve the diversity of the repaired samples. For understanding of global information, we also introduce a new cross semantic attention layer that exploits the long-range dependencies between the known parts and the completed parts, which can improve realism and appearance consistency of repaired samples. Extensive experiments on various datasets such as CelebA-HQ, Places2, Paris Street View and ImageNet clearly demonstrate that our method not only generates diverse inpainting solutions from the same image to be repaired, but also has high image quality.
- Published
- 2020
14. Trident Dehazing Network
- Author
-
Jing Liu, Yuan Xie, Haiyan Wu, Yanyun Qu, and Lizhuang Ma
- Subjects
Haze ,business.industry ,Computer science ,Image map ,Feature extraction ,020207 software engineering ,02 engineering and technology ,Iterative reconstruction ,Feature (computer vision) ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,business - Abstract
Most existing dehazing methods are not robust to nonhomogeneous haze. Meanwhile, the information of dense haze region is usually unknown and hard to estimate, leading to blurry in dehaze result for those regions. Focusing on these two issues, we propose a novel coarse-to-fine model, namely Trident Dehazing Network (TDN), to learn the hazy to hazy- free image mapping with automatic haze density recognition. In detail, TDN is composed of three sub-nets: the Encoder-Decoder Net (EDN) is the main net of TDN to reconstruct the coarse hazy-free feature; the Detail Refinement sub-Net (DRN) helps to refine the high frequency details that was easily lost in the pooling layers in the encoder; and the Haze Density Map Generation sub-Net (HDMGN) can automatically distinguish the thick haze region with thin one, to prevent over-dehazing or under-dehazing in regions of different haze density. Moreover, we propose a frequency domain loss function to make supervision of different frequency band more uniform. Extensive experimental results on synthetic and real datasets demonstrate that our proposed TDN outperforms the state-of-the-arts with better fidelity and perceptual, generalizing well on both dense haze and nonhomogeneous haze scene. Our method won the first place in NTIRE2020 nonhomogeneous dehazing challenge.
- Published
- 2020
15. The Concave n-Square Salient Wood Image-based Quality Assessment
- Author
-
Esa Prakasa, Iwan Muhammad Erwin, and Risnandar
- Subjects
Image quality ,Computer science ,business.industry ,Image map ,Feature extraction ,020206 networking & telecommunications ,Pattern recognition ,02 engineering and technology ,Convolutional neural network ,Square (algebra) ,Image (mathematics) ,Salient ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,Transform coding - Abstract
We make an offer of a state-of-the-art method of the deep salient wood image-based quality assessment (DS-WIQA) for no-reference image appraisal. We explore a five-layer deep convolutional neural network (DCNN) for the salient wood image map. The DS-WIQA uses the concave n-square method. The outcomes allow that DS-WIQA model has a greater achievement on Zenodo and Lignoindo datasets, respectively. We appraise a salient wood image map by extracting in small wood image patches. The DS-WIQA has an admirable performance of other recent methods on Zenodo and Lignoindo datasets, respectively. DS-WIQA outdoes other recent techniques by 14.29% and 19.96% more advanced than other techniques with respect to SROCC and LCC measurement, respectively. DS-WIQA shows up to be more significant than the other DCNN methods.
- Published
- 2020
16. Encoding Temporal Information For Automatic Depression Recognition From Facial Analysis
- Author
-
Miguel Bordallo López, Eric Granger, and Wheidima Carneiro de Melo
- Subjects
Expression Recognition ,Computer science ,Image map ,Pooling ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,Overfitting ,Depression Detection ,SDG 3 - Good Health and Well-being ,Encoding (memory) ,0202 electrical engineering, electronic engineering, information engineering ,Two-stream Model ,Affective computing ,Facial expression ,Artificial neural network ,business.industry ,Deep learning ,020206 networking & telecommunications ,Pattern recognition ,Affective Computing ,Temporal Pooling ,020201 artificial intelligence & image processing ,Artificial intelligence ,business - Abstract
Depression is a mental illness that may be harmful to an individual’s health. Using deep learning models to recognize the facial expressions of individuals captured in videos has shown promising results for automatic depression detection. Typically, depression levels are recognized using 2D-Convolutional Neural Networks (CNNs) that are trained to extract static features from video frames, which impairs the capture of dynamic spatio-temporal relations. As an alternative, 3D-CNNs may be employed to extract spatiotemporal features from short video clips, although the risk of overfitting increases due to the limited availability of labeled depression video data. To address these issues, we propose a novel temporal pooling method to capture and encode the spatio-temporal dynamic of video clips into an image map. This approach allows fine-tuning a pre-trained 2D CNN to model facial variations, and thereby improving the training process and model accuracy. Our proposed method is based on two-stream model that performs late fusion of appearance and dynamic information. Extensive experiments on two benchmark AVEC datasets indicate that the proposed method is efficient and outperforms the state-of-the-art schemes.
- Published
- 2020
17. Accurate Spatial Mapping of Social Media Data with Physical Locations
- Author
-
Panote Siriaraya, Yukiko Kawai, Chonho Lee, Shinji Shimojo, Mohit Mittal, and Takashi Yoshikawa
- Subjects
Social network ,Computer science ,business.industry ,Image map ,0202 electrical engineering, electronic engineering, information engineering ,Spatial mapping ,020206 networking & telecommunications ,020201 artificial intelligence & image processing ,Social media ,02 engineering and technology ,business ,Data science - Abstract
In recent years, an evolutionary change occurred in the digital world when various social networking sites became popular. This led to a big increase of users who share their activities in digital form. A huge amount of digital information has become available, providing researchers with a unique insight into the behavior and activities of entire population. The use of Geo-tagged social media data became an emerging trend to represent user actions and behaviour at specific geographical locations. However, the inaccuracies of Geo-tagged data from social media often limits the utility of this data source in micro scale analysis (at the street or place of interest level (POI)). Our study supports how social media data could be matched accurately with specific physical locations. More specifically, our investigation includes Geo-tweet data and image data from flickr to map accurately physical location using machine leaning and deep learning techniques. In this paper, a preliminary discussion of this work is provided, using Geo-tagged data from Twitter and Open StreetMap in cities such as San Francisco and London as well as for image mapping using Flickr data in cities such as San Francisco and Kyoto.
- Published
- 2019
18. A study on the Image Mapping of the Exhibition Environment
- Author
-
Jae-Young Lee and Jun-Sik Kwon
- Subjects
Exhibition ,020204 information systems ,media_common.quotation_subject ,Image map ,0502 economics and business ,05 social sciences ,0202 electrical engineering, electronic engineering, information engineering ,050211 marketing ,02 engineering and technology ,General Medicine ,Art ,Visual arts ,media_common - Published
- 2018
19. Advancement of close range photogrammetry with a portable panoramic image mapping system (PPIMS)
- Author
-
Yung Chuan Chen and Yi Hsing Tseng
- Subjects
Computer science ,Calibration (statistics) ,business.industry ,Image map ,010401 analytical chemistry ,0211 other engineering and technologies ,Bundle adjustment ,02 engineering and technology ,01 natural sciences ,0104 chemical sciences ,Computer Science Applications ,Close range photogrammetry ,Earth and Planetary Sciences (miscellaneous) ,Computer vision ,Artificial intelligence ,Computers in Earth Sciences ,business ,Engineering (miscellaneous) ,021101 geological & geomatics engineering ,Mobile mapping - Published
- 2018
20. 3D skeleton based action recognition by video-domain translation-scale invariant mapping and multi-scale dilated CNN
- Author
-
Yucheng Chen, Mingyi He, Yuchao Dai, Bo Li, and Xuelian Cheng
- Subjects
Contextual image classification ,Computer Networks and Communications ,business.industry ,Computer science ,Image map ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,020207 software engineering ,Pattern recognition ,02 engineering and technology ,Scale invariance ,Convolutional neural network ,Discriminative model ,Hardware and Architecture ,Robustness (computer science) ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,RGB color model ,020201 artificial intelligence & image processing ,Artificial intelligence ,Invariant (mathematics) ,business ,Software - Abstract
In this paper, we present an image classification approach to action recognition with 3D skeleton videos. First, we propose a video domain translation-scale invariant image mapping, which transforms the 3D skeleton videos to color images, namely skeleton images. Second, a multi-scale dilated convolutional neural network (CNN) is designed for the classification of the skeleton images. Our multi-scale dilated CNN model could effectively improve the frequency adaptiveness and exploit the discriminative temporal-spatial cues for the skeleton images. Even though the skeleton images are very different from natural images, we show that the fine-tuning strategy still works well. Furthermore, we propose different kinds of data augmentation strategies to improve the generalization and robustness of our method. Experimental results on popular benchmark datasets such as NTU RGB + D, UTD-MHAD, MSRC-12 and G3D demonstrate the superiority of our approach, which outperforms the state-of-the-art methods by a large margin.
- Published
- 2018
21. Blind image sharpness assessment based on local contrast map statistics
- Author
-
Goran Gvozden, Sonja Grgic, and Mislav Grgic
- Subjects
media_common.quotation_subject ,Image map ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Image processing ,02 engineering and technology ,Image (mathematics) ,Reduction (complexity) ,Wavelet ,Statistics ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,Contrast (vision) ,Computer vision ,Electrical and Electronic Engineering ,Mathematics ,media_common ,Pixel ,No-reference ,Image quality assessment ,Contrast ,Percentile ,Dynamic range ,business.industry ,Estimator ,020206 networking & telecommunications ,Computer Science::Computer Vision and Pattern Recognition ,Signal Processing ,020201 artificial intelligence & image processing ,Computer Vision and Pattern Recognition ,Artificial intelligence ,business - Abstract
This paper presents a fast blind image sharpness/blurriness assessment model (BISHARP) which operates in spatial and transform domain. The proposed model generates local contrast image maps by computing the root-mean-squared values for each image pixel within a defined size of local neighborhood. The resulting local contrast maps are then transformed into the wavelet domain where the reduction of high frequency content is evaluated in the presence of varying blur strengths. It was found that percentile values computed from sorted, level-shifted, high-frequency wavelet coefficients can serve as reliable image sharpness/blurriness estimators. Furthermore, it was found that higher dynamic range of contrast maps significantly improves model performance. The results of validation performed on seven image databases showed a very high correlation with perceptual scores. Due to low computational requirements the proposed model can be easily utilized in real-world image processing applications.
- Published
- 2018
22. Road Scene Content Analysis for Driver Assistance and Autonomous Driving
- Author
-
Mehmet Celenk and Melih Altun
- Subjects
050210 logistics & transportation ,Engineering ,Visual perception ,business.industry ,Machine vision ,Mechanical Engineering ,Image map ,05 social sciences ,Feature extraction ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Image processing ,02 engineering and technology ,Sensor fusion ,Object detection ,Computer Science Applications ,0502 economics and business ,Automotive Engineering ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Segmentation ,Computer vision ,Artificial intelligence ,business ,Simulation - Abstract
This paper aims to develop a vision-based driver assistance system for scene awareness using video frames obtained from a dashboard camera. A saliency image map is devised with features pertinent to the driving scene. This saliency map mimics the human contour and motion sensitive visual perception by extracting spatial, spectral, and temporal information from the input frames and applying entropy driven image-context-feature data fusion. The resultant fusion output comprises high-level descriptors for still segment boundaries and non-stationary object appearance. Following the segmentation and foreground object detection stage, an adaptive maximum likelihood classifier selects road surface regions. The proposed scene driven vision system improves the driver’s situational awareness by enabling adaptive road surface classification. As experimental results demonstrate, context-aware low-level to high-level information fusion based on human vision model produces superior segmentation, tracking, and classification results that lead to high- level abstraction of driving scene.
- Published
- 2017
23. EdgeGAN: One-way mapping generative adversarial network based on the edge information for unpaired training set
- Author
-
Qiaokang Liang, Youcheng Lei, Zhengwei Li, Yaonan Wang, Wei Sun, Dan Zhang, and Yijie Li
- Subjects
Structure (mathematical logic) ,Computer science ,business.industry ,Image map ,020207 software engineering ,Pattern recognition ,02 engineering and technology ,Edge detection ,Image conversion ,Consistency (database systems) ,Signal Processing ,Content (measure theory) ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,Image translation ,020201 artificial intelligence & image processing ,Computer Vision and Pattern Recognition ,Artificial intelligence ,Enhanced Data Rates for GSM Evolution ,Electrical and Electronic Engineering ,business - Abstract
Image conversion has attracted mounting attention due to its practical applications. This paper proposes a lightweight network structure that can implement unpaired training sets to complete one-way image mapping, based on the generative adversarial network (GAN) and a fixed-parameter edge detection convolution kernel. Compared with the cycle consistent adversarial network (CycleGAN), the proposed network features simpler structure, fewer parameters (only 37.48% of the parameters in CycleGAN), and less training cost (only 35.47% of the GPU memory usage and 17.67% of the single iteration time in CycleGAN). Remarkably, the cyclic consistency becomes not mandatory for ensuring the consistency of the content before and after image mapping. This network has achieved significant processing effects in some image translation tasks, and its effectiveness and validity have been well demonstrated through typical experiments. In the quantitative classification evaluation based on VGG-16, the algorithm proposed in this paper has achieved superior performance.
- Published
- 2021
24. CaffePresso
- Author
-
Siddhartha, Gopalakrishna Hegde, and Nachiket Kapre
- Subjects
Class (computer programming) ,Artificial neural network ,Computer science ,business.industry ,Deep learning ,Image map ,020208 electrical & electronic engineering ,02 engineering and technology ,020202 computer hardware & architecture ,Computer architecture ,Hardware and Architecture ,0202 electrical engineering, electronic engineering, information engineering ,Use case ,Artificial intelligence ,business ,Field-programmable gate array ,Software ,Digital signal processing ,Parametric statistics - Abstract
Auto-tuning and parametric implementation of deep learning kernels allow off-the-shelf accelerator-based embedded platforms to deliver high-performance and energy-efficient mappings of the inference phase of lightweight neural networks. Low-complexity classifiers are characterized by operations on small image maps with two to three deep layers and few class labels. For these use cases, we consider a range of embedded systems with 20W power budgets such as the Xilinx ZC706 (FPGA), NVIDIA Jetson TX1 (GPU), TI Keystone II (DSP), and Adapteva Parallella (RISC+NoC). In CaffePresso, we combine auto-tuning of the implementation parameters, and platform-specific constraints deliver optimized solutions for each input ConvNet specification.
- Published
- 2017
25. Global-local feature attention network with reranking strategy for image caption generation
- Author
-
Yao-wen Chen, Si-ya Xie, Jie Wu, and Xin-bao Shi
- Subjects
Closed captioning ,Computer science ,Image map ,Context (language use) ,02 engineering and technology ,010501 environmental sciences ,Machine learning ,computer.software_genre ,01 natural sciences ,Image (mathematics) ,0202 electrical engineering, electronic engineering, information engineering ,Electrical and Electronic Engineering ,0105 earth and related environmental sciences ,business.industry ,Pattern recognition ,Condensed Matter Physics ,Atomic and Molecular Physics, and Optics ,Electronic, Optical and Magnetic Materials ,Recurrent neural network ,Feature (computer vision) ,Salient ,Benchmark (computing) ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,computer - Abstract
In this paper, a novel framework, named as global-local feature attention network with reranking strategy (GLAN-RS), is presented for image captioning task. Rather than only adopting unitary visual information in the classical models, GLAN-RS explores the attention mechanism to capture local convolutional salient image maps. Furthermore, we adopt reranking strategy to adjust the priority of the candidate captions and select the best one. The proposed model is verified using the Microsoft Common Objects in Context (MSCOCO) benchmark dataset across seven standard evaluation metrics. Experimental results show that GLAN-RS significantly outperforms the state-of-the-art approaches, such as multimodal recurrent neural network (MRNN) and Google NIC, which gets an improvement of 20% in terms of BLEU4 score and 13 points in terms of CIDER score.
- Published
- 2017
26. PERFORMANCE EVALUATION OF THREE DIFFERENT HIGH RESOLUTION SATELLITE IMAGES IN SEMI-AUTOMATIC URBAN ILLEGAL BUILDING DETECTION
- Author
-
Parichehr Hanachi, M. R. Delavar, and N. Khalilimoghadama
- Subjects
lcsh:Applied optics. Photonics ,education.field_of_study ,010504 meteorology & atmospheric sciences ,lcsh:T ,Image map ,Population ,0211 other engineering and technologies ,lcsh:TA1501-1820 ,City map ,02 engineering and technology ,lcsh:Technology ,01 natural sciences ,Metropolitan area ,Geography ,Megacity ,lcsh:TA1-2040 ,Satellite ,Semi automatic ,lcsh:Engineering (General). Civil engineering (General) ,education ,Cartography ,Change detection ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences ,Remote sensing - Abstract
The problem of overcrowding of mega cities has been bolded in recent years. To meet the need of housing this increased population, which is of great importance in mega cities, a huge number of buildings are constructed annually. With the ever-increasing trend of building constructions, we are faced with the growing trend of building infractions and illegal buildings (IBs). Acquiring multi-temporal satellite images and using change detection techniques is one of the proper methods of IB monitoring. Using the type of satellite images with different spatial and spectral resolutions has always been an issue in efficient detection of the building changes. In this research, three bi-temporal high-resolution satellite images of IRS-P5, GeoEye-1 and QuickBird sensors acquired from the west of metropolitan area of Tehran, capital of Iran, in addition to city maps and municipality property database were used to detect the under construction buildings with improved performance and accuracy. Furthermore, determining the employed bi-temporal satellite images to provide better performance and accuracy in the case of IB detection is the other purpose of this research. The Kappa coefficients of 70 %, 64 %, and 68 % were obtained for producing change image maps using GeoEye-1, IRS-P5, and QuickBird satellite images, respectively. In addition, the overall accuracies of 100 %, 6 %, and 83 % were achieved for IB detection using the satellite images, respectively. These accuracies substantiate the fact that the GeoEye-1 satellite images had the best performance among the employed images in producing change image map and detecting the IBs.
- Published
- 2017
27. A diagnostic imaging approach for online characterization of multi-impact in aircraft composite structures based on a scanning spatial-wavenumber filter of guided wave
- Author
-
Shenfang Yuan, Lei Qiu, Yuanqiang Ren, and Zhongqing Su
- Subjects
Engineering ,Guided wave testing ,business.industry ,Mechanical Engineering ,Acoustics ,Image map ,Aerospace Engineering ,02 engineering and technology ,Filter (signal processing) ,021001 nanoscience & nanotechnology ,01 natural sciences ,Signal ,Computer Science Applications ,Sensor array ,Control and Systems Engineering ,Feature (computer vision) ,0103 physical sciences ,Signal Processing ,Electronic engineering ,Wavenumber ,Structural health monitoring ,0210 nano-technology ,business ,010301 acoustics ,Civil and Structural Engineering - Abstract
Monitoring of impact and multi-impact in particular in aircraft composite structures has been an intensive research topic in the field of guided-wave-based structural health monitoring (SHM). Compared with the majority of existing methods such as those using signal features in the time-, frequency- or joint time-frequency domain, the approach based on spatial-wavenumber filter of guided wave shows superb advantage in effectively distinguishing particular wave modes and identifying their propagation direction relative to the sensor array. However, there exist two major issues when conducting online characterization of multi-impact event. Firstly, the spatial-wavenumber filter should be realized in the situation that the wavenumber of high spatial resolution of the complicated multi-impact signal cannot be measured or modeled. Secondly, it’s difficult to identify the multiple impacts and realize multi-impact localization due to the overlapping of wavenumbers. To address these issues, a scanning spatial-wavenumber filter based diagnostic imaging method for online characterization of multi-impact event is proposed to conduct multi-impact imaging and localization in this paper. The principle of the scanning filter for multi-impact is developed first to conduct spatial-wavenumber filtering and to achieve wavenumber-time imaging of the multiple impacts. Then, a feature identification method of multi-impact based on eigenvalue decomposition and wavenumber searching is presented to estimate the number of impacts and calculate the wavenumber of the multi-impact signal, and an image mapping method is proposed as well to convert the wavenumber-time image to an angle-distance image to distinguish and locate the multiple impacts. A series of multi-impact events are applied to a carbon fiber laminate plate to validate the proposed methods. The validation results show that the localization of the multiple impacts are well achieved.
- Published
- 2017
28. CAMERA CALIBRATION ACCURACY AT DIFFERENT UAV FLYING HEIGHTS
- Author
-
Ahmad Razali Yusoff, Mohd Farid Mohd Ariff, Khairulnizam M. Idris, Albert K. Chong, and Zulkepli Majid
- Subjects
lcsh:Applied optics. Photonics ,Accuracy and precision ,business.product_category ,010504 meteorology & atmospheric sciences ,Image map ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,0211 other engineering and technologies ,Bundle adjustment ,02 engineering and technology ,lcsh:Technology ,01 natural sciences ,Camera auto-calibration ,Calibration ,Computer vision ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences ,Digital camera ,Remote sensing ,lcsh:T ,business.industry ,lcsh:TA1501-1820 ,Geography ,lcsh:TA1-2040 ,Artificial intelligence ,lcsh:Engineering (General). Civil engineering (General) ,business ,Focus (optics) ,Camera resectioning - Abstract
Unmanned Aerial Vehicles (UAVs) can be used to acquire highly accurate data in deformation survey, whereby low-cost digital cameras are commonly used in the UAV mapping. Thus, camera calibration is considered important in obtaining high-accuracy UAV mapping using low-cost digital cameras. The main focus of this study was to calibrate the UAV camera at different camera distances and check the measurement accuracy. The scope of this study included camera calibration in the laboratory and on the field, and the UAV image mapping accuracy assessment used calibration parameters of different camera distances. The camera distances used for the image calibration acquisition and mapping accuracy assessment were 1.5 metres in the laboratory, and 15 and 25 metres on the field using a Sony NEX6 digital camera. A large calibration field and a portable calibration frame were used as the tools for the camera calibration and for checking the accuracy of the measurement at different camera distances. Bundle adjustment concept was applied in Australis software to perform the camera calibration and accuracy assessment. The results showed that the camera distance at 25 metres is the optimum object distance as this is the best accuracy obtained from the laboratory as well as outdoor mapping. In conclusion, the camera calibration at several camera distances should be applied to acquire better accuracy in mapping and the best camera parameter for the UAV image mapping should be selected for highly accurate mapping measurement.
- Published
- 2017
29. A review of research works on VGI understanding and image map design
- Author
-
Christian E. Murphy, Liqiu Meng, Linfang Ding, and Jian Yang
- Subjects
Volunteered geographic information ,0209 industrial biotechnology ,020901 industrial engineering & automation ,Computer science ,Image map ,0211 other engineering and technologies ,Earth and Planetary Sciences (miscellaneous) ,Library science ,02 engineering and technology ,Computers in Earth Sciences ,021101 geological & geomatics engineering ,Earth-Surface Processes ,Visual arts - Abstract
This paper reviews more than 10 recent doctoral theses and master’s theses accomplished at the Chair of Cartography, Technical University of Munich (TUM). These research works address two persistent challenges involved in the digital mapmaking process: understanding complex input data and improving cartographic design to keep up with the changing roles and increasing demands of users on immediately usable maps. At first, a dedicated computational approach based on Conditional Random Fields (CRF) for data enhancement of Volunteered Geographic Information (VGI) is introduced and demonstrated with test data from trajectories of floating taxis and OpenStreetMap (OSM). It is followed by a summarized overview of visual analytical approaches for event and behavior discovery and their implementations on various VGI data sources including trajectories of floating taxis, twitter messages and trajectories of football matches. Finally, a concept of image map with a multilayered visual hierarchy is proposed and demonstrated with a set of attention-guided design strategies. Keywords: Probabilistic Graphical Model, visual analytics, event and behavior discovery, attention-guided design Der vorliegende Beitrag gibt einen Uberblick uber die Forschungsinhalte von mehr als 10 aktuellen Dissertationen und Masterarbeiten, die am Lehrstuhl fur Kartographie der Technischen Universitat Munchen durchgefuhrt wurden. Diese Arbeiten befassen sich mit zwei dauerhaften Herausforderun- gen in der digitalen Kartographie: Verstehen von komplexen Eingabedaten und Verbesserung der Gestaltungsverfahren, um mit den sich verandernden Rollen und den stetig wachsenden Anforderungen der Kartennutzer mitzuhalten. Zunachst wird ein auf Probabilistischen Graphischen Modellen (PGM) basierter Ansatz zur Datenanreicherung der VGI vorgestellt und anhand Taxifahrtrouten und Strasennetz von OpenStreetMap (OSM) demonstriert. Es folgt eine Darstellung von visuellen analy- tischen Ansatzen sowie deren Umsetzung zur Entdeckung des Events und des Verhaltens von Personen in verschiedenen VGI-Datenquellen wie z.B. Trajektorien der Floating Taxis, Twitternachrichten und Trajektorien des Fusballspiels. Schlieslich wird das Konzept der Bildkarte, die in der Regel aus Kartensymbolen im Vordergrund und einem Rasterbild im Hintergrund besteht, in Frage gestellt. Eine Reihe von Aufmerk- samkeitsgetriebenen Gestaltungsstrategien zeigen, wie sich Bildkarten mit mehreren visuellen Ebenen generieren lassen. Schlusselworter: Probabilistische Graphische Modelle, Visuelle Analytik, Entdeckung des Events und Verhaltens, Aufmerksamkeits-betriebene Gestaltung
- Published
- 2017
30. Bundle Adjustment of Spherical Images Acquired with a Portable Panoramic Image Mapping System (PPIMS)
- Author
-
Yi Hsing Tseng, Yung Chuan Chen, and Kuan Ying Lin
- Subjects
010504 meteorology & atmospheric sciences ,Panorama ,business.industry ,Orientation (computer vision) ,Image map ,Triangulation (computer vision) ,Bundle adjustment ,02 engineering and technology ,01 natural sciences ,Collinearity equation ,Photogrammetry ,Geography ,Computer graphics (images) ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,Computers in Earth Sciences ,business ,0105 earth and related environmental sciences ,Mobile mapping - Abstract
Thanks to the development of mobile mapping technologies, close-range photogrammetry ( crp ) has advanced to be an efficient mapping method for a variety of applications. A compact crp system equipped with multiple cameras and a gps receiver is one of those advanced portable mapping systems. A portable panoramic image mapping system ( ppms ) was specially designed to capture panoramic images with eight cameras and to obtain the position of image station with a gps receiver. A ppims can be considered as a panoramic crp system. The coordinates of an object point can be determined by the intersection of panoramic image points. For the implementation, we propose a new concept of photogrammetry by using panoramic images. Eight images captured by ppims forms a spherical panorama image ( spi ). Instead of using the original images, ppims spi s are then used for photogrammetric triangulation and mapping. Under this circumstance, one spi is formed for each station, and it is associated with only one set of exterior orientation ( eo ) parameters. Traditional collinearity equations are not applicable to spi triangulation and mapping. Therefore, a novel bundle adjustment algorithm is proposed to solve eo of multi-station spi s. Because ppims spi s are not ideal spi s, a correction scheme was also developed to correct the imperfect geometry of ppims spi . Two test studies were performed for the data collected at a campus test field of National Cheng Kung University ( ncku ) and at a historical site of Tainan. Both cases demonstrate the feasibility of spi bundle adjustment and applying corrections for ppims spi s necessary for effective for bundle adjustment. Furthermore, the experiment's results also confirm that spi s can replace original images for ppims triangulation.
- Published
- 2016
31. Image Mapping Detection of Green Areas Using Speed Up Robust Features
- Author
-
Ika Widiastuti, Faisal Lutfi Afriansyah, Nurul Zainal Fanani, Niyalatul Muna, and Fendik Eko Purnomo
- Subjects
010302 applied physics ,Minimum mean square error ,Speedup ,Pixel ,Aerial photos ,business.industry ,Computer science ,Image map ,Feature extraction ,020206 networking & telecommunications ,02 engineering and technology ,01 natural sciences ,Aerial photography ,0103 physical sciences ,0202 electrical engineering, electronic engineering, information engineering ,Computer vision ,Artificial intelligence ,business ,Image based - Abstract
Development of mapping and remote sensing to detection of green areas in a wide range can do aerial photography using drones. The aerial photo in question is a small format aerial photo using a camera. The image produced from aerial photographs is still fragmented into separate parts. Therefore, it is necessary to merge each sequential image. Merging is done by detecting the mapping of the area by sewing each image based on the point of similarity in pixels. The method applied with the search for similar features uses the Speeded Up Robust Features (SURF). The results obtained to see the level of similarity in the feature mapping area so that the merger into one detected area does not require a long time. The SURF method is applied, giving the results of the number of images that correspond to the Minimum Mean Square Error (MSE) level of 0.0246. The results obtained are the level of similarity at matched point 32 gives a panoramic view approaching the mapping according to the green area of the aerial photo.
- Published
- 2019
32. Writing a Systematic Review for Publication in a Health-Related Degree Program
- Author
-
Clemens Scott Kruse
- Subjects
020205 medical informatics ,Process (engineering) ,Image map ,Computer applications to medicine. Medical informatics ,R858-859.7 ,02 engineering and technology ,Health administration ,health information management ,03 medical and health sciences ,Viewpoint ,0302 clinical medicine ,systematic review ,0202 electrical engineering, electronic engineering, information engineering ,030212 general & internal medicine ,Publication ,Protocol (science) ,Medical education ,business.industry ,4. Education ,General Medicine ,Schedule (workplace) ,Systematic review ,Medicine ,Learning Management ,business ,Psychology - Abstract
Background The protocol in this manuscript was designed to help graduate students publish. It is the result of a challenge from our provost in 2013. I developed this protocol over the last 6 years and have exercised the protocol for the last 5 years. The current version of the protocol has remained mostly static for the last 2 years—only small changes have been made to the process. Objective The objective of this protocol is to enable students to learn a valuable skill of conducting a systematic review and to write the review in a way that can be published. I have designed the protocol to fit into the schedule of a traditional semester, but also used it in compressed semesters. Methods An image map was created in HTML 5.0 and imported into a learning management system. It augments traditional instruction by providing references to published articles, examples, and previously recorded instructional videos. Students use the image map outside the classroom after traditional instruction. The image map helps students create manuscripts that follow established practice and are reported in accordance with the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA), and whose authorship follows guidelines by the International Committee of Medical Journal Editors. Results Since its inception, this protocol has helped 77 students publish 27 systematic reviews in nine journals worldwide. Some manuscripts take multiple years to progress through multiple review processes at multiple journals submitted in sequence. Two other professors in the School of Health Administration have used this protocol in their classes. Conclusions So far, this method has helped 51% of graduate students who used it in my graduate courses publish articles (with more manuscripts under consideration whose numbers have remained uncounted in this sum). I wish success to others who might use this protocol.
- Published
- 2019
33. Multicenter Computer-Aided Diagnosis for Lymph Nodes Using Unsupervised Domain-Adaptation Networks Based on Cross-Domain Confounding Representations
- Author
-
Lingyun Jiang, Dapeng Shi, Ruoxi Qin, Jian Chen, Bin Yan, Jinjin Hai, Huike Zhang, Kai Qiao, and Junling Xu
- Subjects
Article Subject ,Computer science ,Image map ,Computer applications to medicine. Medical informatics ,R858-859.7 ,Information Storage and Retrieval ,02 engineering and technology ,General Biochemistry, Genetics and Molecular Biology ,030218 nuclear medicine & medical imaging ,Domain (software engineering) ,Image (mathematics) ,03 medical and health sciences ,0302 clinical medicine ,Deep Learning ,Discriminative model ,0202 electrical engineering, electronic engineering, information engineering ,Humans ,Diagnosis, Computer-Assisted ,Adaptation (computer science) ,General Immunology and Microbiology ,business.industry ,Applied Mathematics ,Deep learning ,Computational Biology ,Pattern recognition ,General Medicine ,Effective domain ,Feature (computer vision) ,Modeling and Simulation ,Radiographic Image Interpretation, Computer-Assisted ,020201 artificial intelligence & image processing ,Artificial intelligence ,Lymph Nodes ,Neural Networks, Computer ,business ,Tomography, X-Ray Computed ,Algorithms ,Unsupervised Machine Learning ,Research Article - Abstract
To achieve the robust high-performance computer-aided diagnosis systems for lymph nodes, CT images may be typically collected from multicenter data, which cause the isolated performance of the model based on different data source centers. The variability adaptation problem of lymph node data which is related to the problem of domain adaptation in deep learning differs from the general domain adaptation problem because of the typically larger CT image size and more complex data distributions. Therefore, domain adaptation for this problem needs to consider the shared feature representation and even the conditioning information of each domain so that the adaptation network can capture significant discriminative representations in a domain-invariant space. This paper extracts domain-invariant features based on a cross-domain confounding representation and proposes a cycle-consistency learning framework to encourage the network to preserve class-conditioning information through cross-domain image translations. Compared with the performance of different domain adaptation methods, the accurate rate of our method achieves at least 4.4% points higher under multicenter lymph node data. The pixel-level cross-domain image mapping and the semantic-level cycle consistency provided a stable confounding representation with class-conditioning information to achieve effective domain adaptation under complex feature distribution.
- Published
- 2019
34. Visual Navigation of Large Image Graphs
- Author
-
Konstantin Schall, Nico Hezel, Klaus Jung, and Kai Uwe Barthel
- Subjects
Web browser ,Computer science ,business.industry ,Nearest neighbor search ,Image map ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,Visual navigation ,Graph ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,business ,Image retrieval - Abstract
It is impossible to inspect or get an overview of image collections with millions of images. Users often start “exploring” images with a keyword or a similarity search. Both lead to long unstructured lists of result images. In this demo we present a graph-based system for visually exploring and navigating continuously changing sets of millions of images with a web browser. Subsets of images are successively retrieved from a image similarity graph and displayed as a visually sorted 2D image map, which can be zoomed and dragged to explore images from related concepts.
- Published
- 2019
35. Injective State-Image Mapping facilitates Visual Adversarial Imitation Learning
- Author
-
Ryuki Tachibana, Subhajit Chaudhury, Asim Munawar, and Daiki Kimura
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,Computer science ,Computer Vision and Pattern Recognition (cs.CV) ,media_common.quotation_subject ,Image map ,Autonomous agent ,Feature extraction ,Computer Science - Computer Vision and Pattern Recognition ,Machine Learning (stat.ML) ,02 engineering and technology ,010501 environmental sciences ,01 natural sciences ,Motion (physics) ,Machine Learning (cs.LG) ,Statistics - Machine Learning ,0202 electrical engineering, electronic engineering, information engineering ,Reinforcement learning ,0105 earth and related environmental sciences ,media_common ,business.industry ,Deep learning ,020207 software engineering ,Video processing ,Artificial intelligence ,Imitation ,business - Abstract
The growing use of virtual autonomous agents in applications like games and entertainment demands better control policies for natural-looking movements and actions. Unlike the conventional approach of hard-coding motion routines, we propose a deep learning method for obtaining control policies by directly mimicking raw video demonstrations. Previous methods in this domain rely on extracting low-dimensional features from expert videos followed by a separate hand-crafted reward estimation step. We propose an imitation learning framework that reduces the dependence on hand-engineered reward functions by jointly learning the feature extraction and reward estimation steps using Generative Adversarial Networks (GANs). Our main contribution in this paper is to show that under injective mapping between low-level joint state (angles and velocities) trajectories and corresponding raw video stream, performing adversarial imitation learning on video demonstrations is equivalent to learning from the state trajectories. Experimental results show that the proposed adversarial learning method from raw videos produces a similar performance to state-of-the-art imitation learning techniques while frequently outperforming existing hand-crafted video imitation methods. Furthermore, we show that our method can learn action policies by imitating video demonstrations on YouTube with similar performance to learned agents from true reward signals. Please see the supplementary video submission at https://ibm.biz/BdzzNA., Updated the paper to match with version accepted at IEEE MMSP 2019
- Published
- 2019
36. Study on 3S Technology Applied in Law Enforcement of Satellite Image
- Author
-
ZhiShan Cheng and Lili Yin
- Subjects
Land use ,Computer science ,business.industry ,Image map ,020208 electrical & electronic engineering ,Land management ,Law enforcement ,020206 networking & telecommunications ,02 engineering and technology ,0202 electrical engineering, electronic engineering, information engineering ,Data analysis ,Global Positioning System ,Satellite ,business ,Telecommunications ,Pace - Abstract
With the rapid development of Chinese economy, the pace of urban construction has been accelerated obviously. At the same time, illegal land use and illegal construction are becoming more and more serious. At present, using satellite remote sensing technology to monitor land use dynamics is the main task of Law Enforcement Inspection of Satellite Land Image. 3S technology plays an important role in satellite film law enforcement and inspection. High-definition image maps acquired by RS in different periods are the basis of satellite film law enforcement. GPS can accurately locate dynamic patches. GIS technology can be used to construct basic geographic information database and spatial analysis of data. Therefore, taking 3S as the technical core, building a land digital monitoring platform greatly improves the efficiency of satellite film law enforcement and inspection.
- Published
- 2019
37. Change Detection in Remote Sensing Images Based on Image Mapping and a Deep Capsule Network
- Author
-
Licheng Jiao, Xiong Yunta, Yang Hui, Wenping Ma, Yue Wu, and Xiangrong Zhang
- Subjects
Synthetic aperture radar ,Computer science ,Image map ,Feature vector ,Science ,image mapping ,0211 other engineering and technologies ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,optical image ,0202 electrical engineering, electronic engineering, information engineering ,synthetic aperture radar image ,change detection ,021101 geological & geomatics engineering ,Remote sensing ,Image fusion ,Pixel ,capsule network ,Euclidean distance ,heterogeneous image ,Computer Science::Computer Vision and Pattern Recognition ,General Earth and Planetary Sciences ,020201 artificial intelligence & image processing ,Noise (video) ,Change detection - Abstract
Homogeneous image change detection research has been well developed, and many methods have been proposed. However, change detection between heterogeneous images is challenging since heterogeneous images are in different domains. Therefore, direct heterogeneous image comparison in the way that we do it is difficult. In this paper, a method for heterogeneous synthetic aperture radar (SAR) image and optical image change detection is proposed, which is based on a pixel-level mapping method and a capsule network with a deep structure. The mapping method proposed transforms an image from one feature space to another feature space. Then, the images can be compared directly in a similarly transformed space. In the mapping process, some image blocks in unchanged areas are selected, and these blocks are only a small part of the image. Then, the weighted parameters are acquired by calculating the Euclidean distances between the pixel to be transformed and the pixels in these blocks. The Euclidean distance calculated according to the weighted coordinates is taken as the pixel gray value in another feature space. The other image is transformed in a similar manner. In the transformed feature space, these images are compared, and the fusion of the two different images is achieved. The two experimental images are input to a capsule network, which has a deep structure. The image fusion result is taken as the training labels. The training samples are selected according to the ratio of the center pixel label and its neighboring pixels’ labels. The capsule network can improve the detection result and suppress noise. Experiments on remote sensing datasets show the final detection results, and the proposed method obtains a satisfactory performance.
- Published
- 2019
- Full Text
- View/download PDF
38. Assessment of tissue-specific multifactor effects in environmental –omics studies of heterogeneous biological samples: Combining hyperspectral image information and chemometrics
- Author
-
Demetrio Raldúa, Mónica Marro, Eva Prats, Pablo Loza-Alvarez, Romà Tauler, Víctor Olmos, Anna de Juan, Benjamin Piña, European Research Council, Ministerio de Economía y Competitividad (España), Raldúa, Demetrio, Prats, Eva, Piña, Benjamin, Tauler, Romà, Raldúa, Demetrio [0000-0001-5256-1641], Prats, Eva [0000-0001-7838-2027], Piña, Benjamin [0000-0001-9216-2768], and Tauler, Romà [0000-0001-8559-9670]
- Subjects
Hyperspectral imaging ,Image map ,Population ,02 engineering and technology ,Environment ,01 natural sciences ,ANOVA-Simultaneous Component Analysis (ASCA) ,Analytical Chemistry ,Chemometrics ,Component analysis ,Fingerprint ,Image Processing, Computer-Assisted ,Animals ,Least-Squares Analysis ,education ,Zebrafish ,education.field_of_study ,Spectral signature ,Pixel ,Chemistry ,business.industry ,010401 analytical chemistry ,Computational Biology ,Pattern recognition ,Multivariate Curve Resolution-Alternating Least Squares (MCR-ALS) ,021001 nanoscience & nanotechnology ,0104 chemical sciences ,Molecular Imaging ,Phenotype ,Organ Specificity ,Multivariate Analysis ,Raman spectroscopy ,Artificial intelligence ,0210 nano-technology ,business ,Environmental –omics - Abstract
The use of hyperspectral imaging techniques in biological studies has increased in the recent years. Hyperspectral images (HSI) provide chemical information and preserve the morphology and original structure of heterogeneous biological samples, which can be potentially useful in environmental –omics studies when effects due to several factors, e.g., contaminant exposure, phenotype,… at a specific tissue level need to be investigated. Yet, no available strategies exist to exploit adequately this kind of information. This work offers a novel chemometric strategy to pass from the raw image information to useful knowledge in terms of statistical assessment of the multifactor effects of interest in –omic studies. To do so, unmixing of the hyperspectral image measurement is carried out to provide tissue-specific information. Afterwards, several specific ANOVA-Simultaneous Component Analysis (ASCA) models are generated to properly assess and interpret the diverse effect of the factors of interest on the spectral fingerprints of the different tissues characterized. The unmixing step is performed by Multivariate Curve Resolution-Alternating Least Squares (MCR-ALS) on multisets of biological images related to each studied condition and provides reliable HSI spectral signatures and related image maps for each specific tissue in the regions imaged. The variability associated with these signatures within a population is obtained through an MCR-based resampling step on representative pixel subsets of the images analyzed. All spectral fingerprints obtained for a particular tissue in the different conditions studied are used to obtain the related ASCA model that will help to assess the significance of the factors studied on the tissue and, if relevant, to describe the associated fingerprint modifications. The potential of the approach is assessed in a real case of study linked to the investigation of the effect of exposure time to chlorpyrifos-oxon (CPO) on ocular tissues of different phenotypes of zebrafish larvae from Raman HSI of eye cryosections. The study allowed the characterization of melanin, crystalline and internal eye tissue and the phenotype, exposure time and the interaction of the two factors were found to be significant in the changes found in all kind of tissues. Factor-related changes in the spectral fingerprint were described and interpreted per each kind of tissue characterized. © 2018 Elsevier B.V., The research leading to these results has received funding from the European Research Council under the European Union's Seventh Framework Programme ( FP/2007-2013 ) / ERC Grant Agreement n. 32073 (CHEMAGEB project). The authors of this work belong to the network of recognized research groups by the Catalan government (2017 SGR 753 / 2017 SGR 902) and acknowledge the support of the Spanish government through project CTQ2015-66254-C2-2-P , and from NATO (SfP project MD.SFPP 984777 ). ICFO acknowledges financial support from the Spanish Ministry of Economy and Competitiveness through the “Severo Ochoa” program for Centres of Excellence in R&D ( SEV-2015-0522 ), from Fundación Cellex , Fundación Mig-Puig and from Generalitat de Catalunya through the CERCA program", Laserlab-Europe ( EU-H2020 654148 ), Fundació la Marató de TV3 (Molecular imaging of the retina in patients with Multiple Sclerosis by Raman Spectroscopy, 20142030 ), and the National Institutes of Health (NIH, grant 5R21CA187890-02 ). Appendix A
- Published
- 2019
39. Automated Image Mapping and Quantification of Microstructure Heterogeneity in Additive Manufactured Ti6Al4V
- Author
-
Phil Prangnell, Alistair Ho, Alec Davis, Alphons A. Antonysamy, and Hao Zhao
- Subjects
010302 applied physics ,Microscope ,Materials science ,Mechanical Engineering ,Image map ,Detector ,02 engineering and technology ,Function (mathematics) ,021001 nanoscience & nanotechnology ,Condensed Matter Physics ,Microstructure ,01 natural sciences ,law.invention ,Transformation (function) ,Mechanics of Materials ,law ,0103 physical sciences ,Thermal ,Range (statistics) ,General Materials Science ,0210 nano-technology ,Biological system - Abstract
In Additive Manufacturing AM, each volume of material experiences a complex thermal history due to both short-range effects, from the repeated overlap of the thermal field from each heat source pass, and long-range variation in the thermal boundary conditions, related to the part geometry and build height. With an α + β alloy, like Ti64, this can lead to significant local variation in the transformation microstructure, which can contribute to heterogeneity in the mechanical properties of a component. In order to better understand the transformation microstructure variability in AM parts, an automated microstructure analysis tool has been developed, and tested against independently measured data, that can accurately map the inter-lamellar spacing of the α phase and spheroidicity of the β phase, at both high resolution and over large distances. The approach used was based on automated batch image analysis of thousands of image tiles obtained using a mapping function in a high-resolution SEM with a scanning stage. Within a practical operating range of drift in the microscope parameters (e.g. working distance, detector contrast) the errors in the measurements were found to be minimal (
- Published
- 2019
40. Learning to detect genuine versus posed pain from facial expressions using residual generative adversarial networks
- Author
-
Mohammad Tavakolian, Abdenour Hadid, and Carlos Guillermo Bermudez Cruces
- Subjects
Facial expression ,Discriminator ,Computer science ,business.industry ,Image map ,Pooling ,image sequences ,Pattern recognition ,02 engineering and technology ,010501 environmental sciences ,Residual ,ENCODE ,01 natural sciences ,neural nets ,Face (geometry) ,emotion recognition ,0202 electrical engineering, electronic engineering, information engineering ,Benchmark (computing) ,020201 artificial intelligence & image processing ,learning (artificial intelligence) ,Artificial intelligence ,business ,video signal processing ,0105 earth and related environmental sciences ,face recognition - Abstract
We present a novel approach based on Residual Generative Adversarial Network (R-GAN) to discriminate genuine pain expression from posed pain expression by magnifying the subtle changes in the face. In addition to the adversarial task, the discriminator network in R-GAN estimates the intensity level of the pain. Moreover, we propose a novel Weighted Spatiotemporal Pooling (WSP) to capture and encode the appearance and dynamic of a given video sequence into an image map. In this way, we are able to transform any video into an image map embedding subtle variations in the facial appearance and dynamics. This allows using any pre-trained model on still images for video analysis. Our extensive experiments show that our proposed framework achieves promising results compared to state-of-the-art approaches on three benchmark databases, i.e., UNBC-McMaster Shoulder Pain, BioVid Head Pain, and STOIC.
- Published
- 2019
41. Attention Enriched Deep Learning Model for Breast Tumor Segmentation in Ultrasound Images
- Author
-
Min Xian, Aleksandar Vakanski, and Phoebe E. Freer
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,Acoustics and Ultrasonics ,Computer science ,Image map ,Biophysics ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Breast Neoplasms ,Machine Learning (stat.ML) ,02 engineering and technology ,Article ,030218 nuclear medicine & medical imaging ,Machine Learning (cs.LG) ,03 medical and health sciences ,0302 clinical medicine ,Deep Learning ,Robustness (computer science) ,Statistics - Machine Learning ,Prior probability ,0202 electrical engineering, electronic engineering, information engineering ,medicine ,FOS: Electrical engineering, electronic engineering, information engineering ,Humans ,Radiology, Nuclear Medicine and imaging ,Segmentation ,Attention ,Breast ultrasound ,Radiological and Ultrasound Technology ,medicine.diagnostic_test ,business.industry ,Deep learning ,Image and Video Processing (eess.IV) ,Pattern recognition ,Models, Theoretical ,Electrical Engineering and Systems Science - Image and Video Processing ,Image Enhancement ,Salient ,Domain knowledge ,020201 artificial intelligence & image processing ,Female ,Artificial intelligence ,Ultrasonography, Mammary ,business - Abstract
Incorporating human domain knowledge for breast tumor diagnosis is challenging, since shape, boundary, curvature, intensity, or other common medical priors vary significantly across patients and cannot be employed. This work proposes a new approach for integrating visual saliency into a deep learning model for breast tumor segmentation in ultrasound images. Visual saliency refers to image maps containing regions that are more likely to attract radiologists visual attention. The proposed approach introduces attention blocks into a U-Net architecture, and learns feature representations that prioritize spatial regions with high saliency levels. The validation results demonstrate increased accuracy for tumor segmentation relative to models without salient attention layers. The approach achieved a Dice similarity coefficient of 90.5 percent on a dataset of 510 images. The salient attention model has potential to enhance accuracy and robustness in processing medical images of other organs, by providing a means to incorporate task-specific knowledge into deep learning architectures., Comment: 16 pages, 5 figures
- Published
- 2019
- Full Text
- View/download PDF
42. Path-Invariant Map Networks
- Author
-
Xiaowei Zhou, Qixing Huang, Zaiwei Zhang, Lemeng Wu, and Zhenxiao Liang
- Subjects
FOS: Computer and information sciences ,Artificial neural network ,Computer science ,business.industry ,Computer Vision and Pattern Recognition (cs.CV) ,Image map ,Deep learning ,Computer Science - Computer Vision and Pattern Recognition ,020207 software engineering ,02 engineering and technology ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Segmentation ,Artificial intelligence ,Invariant (mathematics) ,business ,Algorithm ,Time complexity ,Parametric statistics - Abstract
Optimizing a network of maps among a collection of objects/domains (or map synchronization) is a central problem across computer vision and many other relevant fields. Compared to optimizing pairwise maps in isolation, the benefit of map synchronization is that there are natural constraints among a map network that can improve the quality of individual maps. While such self-supervision constraints are well-understood for undirected map networks (e.g., the cycle-consistency constraint), they are under-explored for directed map networks, which naturally arise when maps are given by parametric maps (e.g., a feed-forward neural network). In this paper, we study a natural self-supervision constraint for directed map networks called path-invariance, which enforces that composite maps along different paths between a fixed pair of source and target domains are identical. We introduce path-invariance bases for efficient encoding of the path-invariance constraint and present an algorithm that outputs a path-variance basis with polynomial time and space complexities. We demonstrate the effectiveness of our approach on optimizing object correspondences, estimating dense image maps via neural networks, and semantic segmentation of 3D scenes via map networks of diverse 3D representations. In particular, for 3D semantic segmentation, our approach only requires 8% labeled data from ScanNet to achieve the same performance as training a single 3D segmentation network with 30% to 100% labeled data.
- Published
- 2018
43. Geometrical Correction of Side-scan Sonar Images
- Author
-
Hugo Guterman and Tal Sheffer
- Subjects
010302 applied physics ,Side-scan sonar ,business.industry ,Computer science ,Image map ,02 engineering and technology ,Slant range ,Iterative reconstruction ,021001 nanoscience & nanotechnology ,01 natural sciences ,Sonar ,Distortion ,0103 physical sciences ,Computer vision ,Noise (video) ,Artificial intelligence ,Underwater ,0210 nano-technology ,business - Abstract
The underwater environment makes object-detection missions difficult. Side-scan Sonar (SSS) has been found to be suitable for seabed scanning missions, however the sonar images acquired from SSS often suffer from considerable noise and geometrical distortion, which changes the understanding of the texture, size, and shape of seabed objects. In order to identify seabed objects, it is thus vital to reconstruct the actual shape by reducing distortion. This paper proposes a process for correcting and reconstructing the sonar image map that utilizes intensity normalization, slant range correction, yaw and pitch correction, and speed and location correction. This is done using navigation and inertial data acquired by the autonomous underwater vehicle sensors.
- Published
- 2018
44. A new bearing fault diagnosis method based on signal-to-image mapping and convolutional neural network
- Author
-
Xiaohui Gu, Shaopu Yang, Yongqiang Liu, Wenpeng Liu, Jing Zhao, and Qiang Li
- Subjects
Artificial neural network ,Computer science ,business.industry ,Applied Mathematics ,Image map ,Deep learning ,020208 electrical & electronic engineering ,010401 analytical chemistry ,SIGNAL (programming language) ,Feature extraction ,Process (computing) ,Pattern recognition ,02 engineering and technology ,Condensed Matter Physics ,Fault (power engineering) ,01 natural sciences ,Convolutional neural network ,0104 chemical sciences ,0202 electrical engineering, electronic engineering, information engineering ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,Instrumentation - Abstract
Fault diagnosis is important to ensure the safety and efficience of mechanical equipment. In recent years, data-driven fault diagnosis methods have received extensive attention and research from many scholars. Different from the traditional fault feature extraction methods that rely on expert experience, this paper proposes a feature extraction method based on deep learning (DL). In order to meet the needs of neural networks for the amount of data, data augmentation is emploied to increase the amount of original data. Then, a novel signal-to-image mapping (STIM) is proposed to convert the one-dimensional vibration signals into two-dimensional grey images, which greatly reduce the human involvement. Finally, a convolutional neural network (CNN) model is established to extract fault features from grey images and realize fault classification. The learning process of the CNN model is analyzed and two different bearing experiment datasets are used to verify the effectiveness of the proposed method.
- Published
- 2021
45. Characterization system based on image mapping for field emission devices
- Author
-
Francisco J. Ramirez-Fernandez, Michel O. S. Dantas, Elisabete Galeazzo, Henrique E. M. Peres, Débora A.C. Silva, and Maycon M. Kopelvski
- Subjects
Data processing ,Virtual instrumentation ,business.industry ,Computer science ,Applied Mathematics ,Image map ,Electrical engineering ,Video camera ,02 engineering and technology ,010402 general chemistry ,021001 nanoscience & nanotechnology ,Condensed Matter Physics ,01 natural sciences ,0104 chemical sciences ,Characterization (materials science) ,law.invention ,Field electron emission ,law ,Measuring instrument ,Electronic engineering ,Electronics ,Electrical and Electronic Engineering ,0210 nano-technology ,business ,Instrumentation - Abstract
Field Emission devices (FE) have been proposed as efficient electron sources for several applications such as electron microscopy and vacuum sensors. Evidently, characterization methods applied during development phase of FE devices are crucial to evaluate aspects related with their working stability, homogeneity, and efficiency. However, the traditional methods provide only overall information about such characteristics, which difficult to improve the performance of these devices and their integration with electronics. To overcome this problem, this work presents an alternative system to characterize FE devices through electron emission imaging in real-time. The proposed system acquires I - V features of FE devices, while a video camera captures the emission image from a phosphor screen. Virtual instrumentation based on LabVIEW manages the whole system including measurement instruments, image capture, and data processing. As a result, histograms, 3D maps, and other FE analyses provide information about emitting characteristics of selected regions of interest. The main contribution of this work is to offer an important tool for the analyses of electron emission, by the association of captured images with the localized emission current. The extracted information from our system can efficiently support the characterization and the development of FE devices.
- Published
- 2016
46. A parametric intensity-based 3D image registration method for magnetic resonance imaging
- Author
-
Huajun Song and Peihua Qiu
- Subjects
business.industry ,Image map ,Geometric transformation ,Image registration ,02 engineering and technology ,Translation (geometry) ,030218 nuclear medicine & medical imaging ,03 medical and health sciences ,0302 clinical medicine ,Signal Processing ,Line (geometry) ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,Affine transformation ,Electrical and Electronic Engineering ,business ,Rotation (mathematics) ,Transformation geometry ,Mathematics - Abstract
Image registration (IR) aims to map one image to another of a same scene. With rapid progress in image acquisition technologies, 3D IR becomes an important problem in magnetic resonance imaging (MRI) and other applications. In the literature, however, most IR methods are for 2D images and there are only a limited number of 3D methods available. Because 3D images have much complicated structure than their 2D counterparts, 3D IR is not just a simple generalization of the 2D IR problem. In this paper, we develop a 3D IR method that can handle cases with affine geometric transformations well. By its definition, an affine transformation maps a line to a line, and it includes rotation, translation, and scaling as special cases. In practice, most geometric transformations involved in IR problems are affine transformations. Therefore, our proposed method can find many IR applications. It is shown that this method works well in various cases, including cases when the data size of a 3D image is reduced for different reasons. This latter property makes it attractive for many 3D IR applications, since 3D images are often big in data size and it is natural to reduce their size for fast computation.
- Published
- 2016
47. A novel method of photographic geological logging based on parallel image sequence in small tunnel
- Author
-
Yan-lan He, Ming-fei Wu, Hao Li, Yufeng Mao, and Biao Yang
- Subjects
Database ,Computer science ,Settlement (structural) ,Image map ,Logging ,0211 other engineering and technologies ,Metals and Alloys ,General Engineering ,InformationSystems_DATABASEMANAGEMENT ,02 engineering and technology ,computer.software_genre ,Image correction ,Sketch ,020501 mining & metallurgy ,0205 materials engineering ,Mining engineering ,Image sequence ,Image acquisition ,Informatization ,computer ,ComputingMethodologies_COMPUTERGRAPHICS ,021101 geological & geomatics engineering - Abstract
Small tunnels such as engineering geological exploratory tunnels and mine roadways are generally narrow, which make the existing photographic geological logging technique inapplicable. Therefore, geological logging of exploratory tunnels has always been taking the method of manual sketch work which has low efficiency and poor informatization degree of products, and it is a technical issue requiring urgent settlement for geological logging of small tunnels. This paper proposes and studies novel methods of photographic geological logging suitable for small tunnels, including image acquisition, image orientation control, image geometric correction, unfolded image map generation and geological attitude measurement, etc. Experiments show that the method can meet the precision requirement of geological logging. The novel method helps to realize the fast acquisition and processing of image-based geological logging data for small tunnels, and the forms of logging result are more abundant and more applicable to informatized management and application of geological logging data.
- Published
- 2016
48. Novel Image Metrics for Retrieval of the Lateral Resolution in Line Scan-Based 2D LA-ICPMS Imaging via an Experimental-Modeling Approach
- Author
-
Andrei Izmer, Frank Vanhaecke, Johannes T. van Elteren, and Vid Simon Šelih
- Subjects
Chemistry ,business.industry ,Image map ,medicine.medical_treatment ,010401 analytical chemistry ,02 engineering and technology ,021001 nanoscience & nanotechnology ,Ablation ,01 natural sciences ,Sample (graphics) ,0104 chemical sciences ,Analytical Chemistry ,Convolution ,Optics ,Sampling (signal processing) ,Optical transfer function ,Image noise ,medicine ,0210 nano-technology ,business ,Image resolution - Abstract
The quality of elemental image maps obtained via line scan-based LA-ICPMS is a function of the temporal response of the entire system, governed by the design of the system and mapping and acquisition conditions used, next to the characteristics of the sample. To quantify image degradation, ablation targets with periodic gratings are required for the construction of a modulation transfer function (MTF) and subsequent determination of the lateral resolution as a function of image noise and contrast. Since such ablation targets, with suitable matrix composition, are not readily available, computer-generated periodic gratings were virtually ablated via a computational process based on a two-step discrete-time convolution procedure using empirical/experimental input data. This experimental-modeling procedure simulates LA-ICPMS imaging based on two consecutive processes, viz., LA sampling (via ablation crater profiles [ACP]) and aerosol washout/transfer/ICPMS measurement (via single pulse responses [SPR]). By random selection of experimental SPRs from a large database for each individual pulse during the simulation, the convolution procedure simulates an accurate elemental image map of the periodic gratings with realistic (proportional or flicker) noise. This facilitates indirect retrieval of the experimental lateral resolution for the matrix targeted without performing actual line scanning on periodic gratings.
- Published
- 2016
49. Underwater 3D Reconstruction for Underwater Construction Robot Based on 2D Multibeam Imaging Sonar
- Author
-
Seung-Joon Choi and Young-eun Song
- Subjects
010505 oceanography ,business.industry ,Image map ,3D reconstruction ,Elevation ,02 engineering and technology ,Remotely operated underwater vehicle ,01 natural sciences ,Sonar ,Geography ,0202 electrical engineering, electronic engineering, information engineering ,Reflection (physics) ,Synthetic aperture sonar ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,Underwater ,business ,0105 earth and related environmental sciences ,Remote sensing - Abstract
This paper presents an underwater structure 3D reconstruction method using a 2D multibeam imaging sonar. Compared with other underwater environmental recognition sensors, the 2D multibeam imaging sonar offers high resolution images in water with a high turbidity level by showing the reflection intensity data in real-time. With such advantages, almost all underwater applications, including ROVs, have applied this 2D multibeam imaging sonar. However, the elevation data are missing in sonar images, which causes difficulties with correctly understanding the underwater topography. To solve this problem, this paper concentrates on the physical relationship between the sonar image and the scene topography to find the elevation information. First, the modeling of the sonar reflection intensity data is studied using the distances and angles of the sonar beams and underwater objects. Second, the elevation data are determined based on parameters like the reflection intensity and shadow length. Then, the elevation information is applied to the 3D underwater reconstruction. This paper evaluates the presented real-time 3D reconstruction method using real underwater environments. Experimental results are shown to appraise the performance of the method. Additionally, with the utilization of ROVs, the contour and texture image mapping results from the obtained 3D reconstruction results are presented as applications.
- Published
- 2016
50. THE CARTOGRAPHIC CONCEPT OF THE IMAGE MAP
- Author
-
L. Belka and Vit Vozenilek
- Subjects
lcsh:Applied optics. Photonics ,media_common.quotation_subject ,Image map ,0211 other engineering and technologies ,Locator map ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,010502 geochemistry & geophysics ,01 natural sciences ,lcsh:Technology ,Terminology ,Projection (mathematics) ,Component (UML) ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences ,media_common ,ComputingMethodologies_COMPUTERGRAPHICS ,Information retrieval ,lcsh:T ,lcsh:TA1501-1820 ,Visualization ,Symbol ,Thematic map ,Geography ,lcsh:TA1-2040 ,lcsh:Engineering (General). Civil engineering (General) ,Cartography - Abstract
Image maps have become very popular and frequently produced cartographical outputs during recent years. However, the unambiguous terminology, definitions, content and appearance specification have not been widely researched. The paper deals with the new definition of image map, its components delineation, and basic classification. The authors understand the image map as a special map portraying geographic space in a particular cartographical projection and map scale, where its content consists of two basic components – image and symbol components. Image component is represented by remote sensing image(s), while symbol component is represented by cartographical symbols. An image map has to have three essential attributes: cartographical projection, map scale and symbol component by means of map language. The authors also present aspects of topographic and thematic image maps.
- Published
- 2016
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.