502 results on '"Jing-Ming Guo"'
Search Results
202. High-capacity data hiding in halftone images using minimal-error bit searching and least-mean square filter.
- Author
-
Soo-Chang Pei and Jing-Ming Guo
- Published
- 2006
- Full Text
- View/download PDF
203. Novel robust watermarking technique in dithering halftone images.
- Author
-
Soo-Chang Pei, Jing-Ming Guo, and Hua Lee
- Published
- 2005
- Full Text
- View/download PDF
204. A Light-Weight CNN for Object Detection with Sparse Model and Knowledge Distillation
- Author
-
Jing-Ming Guo, Jr-Sheng Yang, Sankarasrinivasan Seshathiri, and Hung-Wei Wu
- Subjects
Computer Networks and Communications ,Hardware and Architecture ,Control and Systems Engineering ,CNN ,object detection ,sparse model ,knowledge distillation ,student–teacher model ,Signal Processing ,Electrical and Electronic Engineering - Abstract
This study details the development of a lightweight and high performance model, targeting real-time object detection. Several designed features were integrated into the proposed framework to accomplish a light weight, rapid execution, and optimal performance in object detection. Foremost, a sparse and lightweight structure was chosen as the network’s backbone, and feature fusion was performed using modified feature pyramid networks. Recent learning strategies in data augmentation, mixed precision training, and network sparsity were incorporated to substantially enhance the generalization for the lightweight model and boost the detection accuracy. Moreover, knowledge distillation was applied to tackle dropping issues, and a student–teacher learning mechanism was also integrated to ensure the best performance. The model was comprehensively tested using the MS-COCO 2017 dataset, and the experimental results clearly demonstrated that the proposed model could obtain a high detection performance in comparison to state-of-the-art methods, and required minimal computational resources, making it feasible for many real-time deployments.
- Published
- 2022
- Full Text
- View/download PDF
205. Incorporating color feature on LBP-based image retrieval.
- Author
-
Jing-Ming Guo and Heri Prasetyo
- Published
- 2015
- Full Text
- View/download PDF
206. Two-Stage Cascaded CNN Model for 3D Mitochondria EM Segmentation
- Author
-
Jing-Ming Guo, Sankarasrinivasan Seshathiri, Jia-Hao Liu, and Wei-Wen Hsu
- Subjects
3D CNN ,electron microscopy ,image analysis ,Computer Networks and Communications ,Hardware and Architecture ,Control and Systems Engineering ,Signal Processing ,Electrical and Electronic Engineering ,mitochondria segmentation - Abstract
Mitochondria are the organelles that generate energy for the cells. Many studies have suggested that mitochondrial dysfunction or impairment may be related to cancer and other neurodegenerative disorders such as Alzheimer’s and Parkinson’s diseases. Therefore, morphologically detailed alterations in mitochondria and 3D reconstruction of mitochondria are highly demanded research problems in the performance of clinical diagnosis. Nevertheless, manual mitochondria segmentation over 3D electron microscopy volumes is not a trivial task. This study proposes a two-stage cascaded CNN architecture to achieve automated 3D mitochondria segmentation, combining the merits of top-down and bottom-up approaches. For top-down approaches, the segmentation is conducted on objects’ localization so that the delineations of objects’ contours can be more precise. However, the combinations of 2D segmentation from the top-down approaches are inadequate to perform proper 3D segmentation without the information on connectivity among frames. On the other hand, the bottom-up approach finds coherent groups of pixels and takes the information of 3D connectivity into account in segmentation to avoid the drawbacks of the 2D top-down approach. However, many small areas that share similar pixel properties with mitochondria become false positives due to insufficient information on objects’ localization. In the proposed method, the detection of mitochondria is carried out with multi-slice fusion in the first stage, forming the segmentation cues. Subsequently, the second stage is to perform 3D CNN segmentation that learns the pixel properties and the information of 3D connectivity under the supervision of cues from the detection stage. Experimental results show that the proposed structure alleviates the problems in both the top-down and bottom-up approaches, which significantly accomplishes better performance in segmentation and expedites clinical analysis.
- Published
- 2023
- Full Text
- View/download PDF
207. Hybrid pixel-based data hiding and block-based watermarking for error-diffused halftone images.
- Author
-
Soo-Chang Pei and Jing-Ming Guo
- Published
- 2003
- Full Text
- View/download PDF
208. Improved Block Truncation Coding Based on the Void-and-Cluster Dithering Approach.
- Author
-
Jing-Ming Guo and Ming-Feng Wu
- Published
- 2009
- Full Text
- View/download PDF
209. Watermarking in dot-diffusion halftones using adaptive class-matrix and error diffusion
- Author
-
Jing-Ming Guo and Sankarasrinivasan Seshathiri
- Subjects
Error diffusion ,Class (set theory) ,Matrix (mathematics) ,Information Systems and Management ,Computer Networks and Communications ,Computer science ,Computer Science::Multimedia ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Statistical physics ,Electrical and Electronic Engineering ,Diffusion (business) ,Digital watermarking ,Information Systems - Abstract
Digital halftoning deals with transforming a gray or color image into its binary version which is useful in printing applications. Dot diffusion is one of the prominent halftone methods which can yield superior image quality with parallel processing capabilities. In this paper, a rapid watermarking algorithm is proposed for dot-diffusion halftone images using adaptive class-matrix selection and modified error diffusion kernels. To process the image using an adaptive class matrix, the processing order of the class matrix is reversed and transposed, and for error diffusion the coefficients are replaced with different weights. For decoding, an effective strategy is proposed based on a correlation analysis and halftone statistics. The proposed strategy can successfully embed and decode the binary watermark from a single dot-diffused halftone image. From the experimental results, the proposed method is found to be effective in terms of good decoding accuracy, imperceptibility and robustness against various printed distortions.
- Published
- 2019
- Full Text
- View/download PDF
210. Vehicle verification using Generalized Gaussian Distribution feature descriptor.
- Author
-
Jing-Ming Guo and Heri Prasetyo
- Published
- 2014
- Full Text
- View/download PDF
211. AWGN-Based Image Denoiser using Convolutional Vision Transformer
- Author
-
Heri Prasetyo, Alim Wicaksono Hari Prayuda, and Jing-Ming Guo
- Subjects
Noise measurement ,Computer science ,business.industry ,Noise reduction ,Noise map ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Pattern recognition ,Residual ,Image (mathematics) ,Noise ,symbols.namesake ,Additive white Gaussian noise ,Computer Science::Computer Vision and Pattern Recognition ,symbols ,Artificial intelligence ,business ,Transformer (machine learning model) - Abstract
This paper presents a new method for denoising an image corrupted with Additive White Gaussian Noise (AWGN). We adopt the effectiveness of Convolutional Vision Transformer (CvT) to suppress the occurred noise on an image. The proposed method exploits the residual learning approach in order to estimate and reduce the noise on a noisy image. Herein, the model is trained in the end-to-end manner to capture the relation between the noisy image and its noise map. The experimental results show the effectiveness of the proposed method in terms of subjective and objective measurements.
- Published
- 2021
- Full Text
- View/download PDF
212. Security attacks on the wavelet transform and singular value decomposition image watermarking.
- Author
-
Jing-Ming Guo and Heri Prasetyo
- Published
- 2013
- Full Text
- View/download PDF
213. Nighttime image dehazing based on improved erosion dark channel and multi-scale clipping limit histogram equalization
- Author
-
Chia-Hsiang Lin and Jing-Ming Guo
- Subjects
symbols.namesake ,Channel (digital image) ,Transmission (telecommunications) ,Clipping (photography) ,Scale (ratio) ,Computer science ,Gaussian ,symbols ,Limit (mathematics) ,Erosion (morphology) ,Algorithm ,Histogram equalization - Abstract
In this paper, a new night dehazing strategy is proposed by modifying the transmission and atmospheric light conditions. To suppress the local light sources, a Gaussian based atmospheric light model is developed along with erosion filters to improve the traditional atmospheric light factors. A similar erosion operator is also utilized for transmission to refine, and the model is combined with a multi-scale clipping limit of ICLAHE to obtain an optimal nighttime dehazing effect. The experimental results confirm that the proposed strategy performs better in comparison to state-of-the-art techniques and is computationally less intensive.
- Published
- 2021
- Full Text
- View/download PDF
214. Data hiding in halftone images with noise-balanced error diffusion.
- Author
-
Soo-Chang Pei and Jing-Ming Guo
- Published
- 2003
- Full Text
- View/download PDF
215. Deep Concatenated Residual Networks for Improving Quality of Halftoning-Based BTC Decoded Image
- Author
-
Heri Prasetyo, Jing-Ming Guo, Alim Wicaksono Hari Prayuda, and Chih-Hsien Hsia
- Subjects
Computer science ,02 engineering and technology ,Iterative reconstruction ,Residual ,lcsh:Computer applications to medicine. Medical informatics ,Convolutional neural network ,lcsh:QA75.5-76.95 ,Article ,Image (mathematics) ,convolutional neural networks ,0202 electrical engineering, electronic engineering, information engineering ,halftoning ,Radiology, Nuclear Medicine and imaging ,lcsh:Photography ,Electrical and Electronic Engineering ,business.industry ,Deep learning ,deep learning ,020207 software engineering ,Pattern recognition ,lcsh:TR1-1050 ,image reconstruction ,Computer Graphics and Computer-Aided Design ,Block Truncation Coding ,block truncation coding ,Noise ,lcsh:R858-859.7 ,020201 artificial intelligence & image processing ,Computer Vision and Pattern Recognition ,Artificial intelligence ,lcsh:Electronic computers. Computer science ,business ,residual learning ,Image compression - Abstract
This paper presents a simple technique for improving the quality of the halftoning-based block truncation coding (H-BTC) decoded image. The H-BTC is an image compression technique inspired from typical block truncation coding (BTC). The H-BTC yields a better decoded image compared to that of the classical BTC scheme under human visual observation. However, the impulsive noise commonly appears on the H-BTC decoded image. It induces an unpleasant feeling while one observes this decoded image. Thus, the proposed method presented in this paper aims to suppress the occurring impulsive noise by exploiting a deep learning approach. This process can be regarded as an ill-posed inverse imaging problem, in which the solution candidates of a given problem can be extremely huge and undetermined. The proposed method utilizes the convolutional neural networks (CNN) and residual learning frameworks to solve the aforementioned problem. These frameworks effectively reduce the impulsive noise occurrence, and at the same time, it improves the quality of H-BTC decoded images. The experimental results show the effectiveness of the proposed method in terms of subjective and objective measurements.
- Published
- 2020
216. DDSnet: A Deep Document Segmentation with Hybrid Blocks Architecture Network
- Author
-
Jing-Ming Guo, Hao-Hsuan Lee, and Li-Ying Chang
- Subjects
Structure (mathematical logic) ,business.industry ,Computer science ,Deep learning ,02 engineering and technology ,Semantics ,computer.software_genre ,01 natural sciences ,010309 optics ,Market segmentation ,0103 physical sciences ,0202 electrical engineering, electronic engineering, information engineering ,Task analysis ,Benchmark (computing) ,020201 artificial intelligence & image processing ,Segmentation ,Data mining ,Artificial intelligence ,business ,computer ,Block (data storage) - Abstract
In recent years, the development of document segmentation technology is gaining more and more attention in the area of semantic segmentation, which plays an important role in the task of understanding the structure of documents. Although this demand applying deep learning approaches has undergone continuing advancement, the document segmentation systems still suffer from low accuracy rate. This paper presents a new high-performance document segmentation algorithm, namely Deep Document Segmentation Network (DDSnet), which incorporates advanced end-to-end deep learning methods for segmenting four different types of document features, including background, texts, tables, and figures. To overcome the small receptive fields, the atrous residual block is proposed, which is efficiently boosted by the adoption of multi-branches structure. For the better fine-grained output, the proposed atrous convolution residual block is conducted to achieve high accuracy. Moreover, this paper also releases the brand-new large-scale database, namely PPSD2019, for document segmentation that provides a pixel-level database for another benchmark. As documented in the experimental results, the proposed document segmentation method achieves a superior segmentation rate than that of the former competitive schemes. As a result, the proposed method and database can be considered as a very competitive candidate for the document segmentation applications.
- Published
- 2020
- Full Text
- View/download PDF
217. Pyramid M-Shaped Network for Ordered Dithering Block Truncation Coding Image Restoration
- Author
-
Heri Prasetyo, H. P. Alim Wicaksono, Jing-Ming Guo, and Muhammad Farhan Ichlasul Amal
- Subjects
Ordered dithering ,Computer science ,business.industry ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Wavelet transform ,Pattern recognition ,Iterative reconstruction ,Convolutional neural network ,Block Truncation Coding ,Wavelet ,Computer Science::Computer Vision and Pattern Recognition ,Pyramid ,Pyramid (image processing) ,Artificial intelligence ,business ,Image restoration - Abstract
This paper presents a novel deep learning-based technique for reconstructing the Ordered Dithering Block Truncation Coding (ODBTC) decoded image. The proposed method inherits the effectiveness of wavelets transform and Convolutional Neural Networks (CNN). The proposed technique employs the two-dimensional Decimated Wavelet Transform (DWT) to decompose the ODBTC image into low and high frequency sub-bands over various resolution. These image sub-bands are progressively reconstructed with the Pyramid M-Shaped CNN consisting multiple input and output. This scheme considers the ODBTC decoded image as noisy image in Which the information of image can reconstructed by modifying the wavelets sub-bands using residual learning strategy. The produced sub-bands can be combined and transformed back to improve the quality of ODBTC reconstructed image. As documented in experimental results, the proposed method yields promising result for the ODBTC image reconstruction.
- Published
- 2020
- Full Text
- View/download PDF
218. Deep Learning based Inverse Halftoning via Stationary Wavelet Domain
- Author
-
Jing-Ming Guo, Heri Prasetyo, and H. P. Alim Wicaksono
- Subjects
Halftone ,Pixel ,Computer science ,business.industry ,Deep learning ,05 social sciences ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,050209 industrial relations ,Pattern recognition ,Iterative reconstruction ,Convolutional neural network ,Image (mathematics) ,Wavelet ,Distortion ,0502 economics and business ,Artificial intelligence ,business ,050203 business & management ,ComputingMethodologies_COMPUTERGRAPHICS - Abstract
Inverse halftoning aims to reconstructs the continuous-tone image, i.e. 255 discrete color levels from its halftone image. The main challenge of inverse halftoning task is on removing the noisy dots over flat areas and, at the same time, restoring the image structures and texture details. To tackle this problem, we propose a novel approach to reconstruct the halftone image in wavelet domain, namely Subbands Aware Inverse Halftoning Network (SAIHN). The proposed framework consists of two sub networks, i.e. the first networks adjust the wavelet sub-bands for image reconstruction, whereas, the second networks refines the quality of reconstructed image by correcting the distortion in pixel domain. The experimental results confirm that the proposed method can reconstruct the halftone image with high quality measured in term of objectively and subjectively assessments. Furthermore, the proposed method demonstrates its superiority on less learnable parameters and more efficient computational complexity in comparison to that of the former existing deep learning-based reconstruction method.
- Published
- 2020
- Full Text
- View/download PDF
219. Efficient Finger Vein Technology Based on Fast Binary Robust Independent Elementary Feature Combined with Multi-Image Quality Assessment Verification
- Author
-
Li-Ying Chang, Chih-Hsien Hsia, Chong-Sheng Wu, and Jing-Ming Guo
- Subjects
021110 strategic, defence & security studies ,Biometrics ,Computer science ,Quality assessment ,business.industry ,0211 other engineering and technologies ,Multi-image ,Binary number ,Pattern recognition ,02 engineering and technology ,Fingerprint recognition ,Finger vein recognition ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,Invariant (mathematics) ,business - Abstract
Recent years, fingerprint recognition has a high market share in biometrics, but humans' hands are often covered with oil and sweat that they secrete, rainwater or dirt, which often affect the accuracy of fingerprint recognition. This work proposed a finger vein identification technology for low response time that can be implemented in a cost-efficient embedded system using the binary robust invariant elementary feature. In this method allows the features matching with binary robust invariant elementary feature and verification with multi-quality assessment process. In experimental results, that the EER performance are 0.13% and 0.69%, using homemade and public (FVUSM) datasets when the data were collected with training and testing. The method is very suitable for real-time finger vein recognition applications.
- Published
- 2020
- Full Text
- View/download PDF
220. Autoencoder-based Image Companding
- Author
-
Heri Prasetyo, H. P. Alim Wicaksono, and Jing-Ming Guo
- Subjects
Computer science ,business.industry ,Deep learning ,Process (computing) ,Pattern recognition ,02 engineering and technology ,Residual ,Autoencoder ,Convolutional neural network ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,Encoder ,Companding ,High dynamic range - Abstract
This paper presents a deep learning-based method for effective image companding. The autoencoder inherits the effectiveness of Convolutional Neural Networks (CNN) and residual learning framework to transform High Dynamic Range (HDR) images to Low Dynamic Range (LDR) and its reverse process. Since, the image companding task involves the nondifferentiable operation, thus the encoder and decoder networks are alternately trained using an iterative approach. The experimental result clearly reveals that the proposed method yields a promising result.
- Published
- 2020
- Full Text
- View/download PDF
221. Inverse Halftoning With Context Driven Prediction.
- Author
-
Jing-Ming Guo, Yun-Fu Liu, Jen-Ho Chen, and Jiann-Der Lee
- Published
- 2014
- Full Text
- View/download PDF
222. Improved Beta Chaotic Image Encryption for Multiple Secret Sharing
- Author
-
Heri Prasetyo, Dwi Riyono, and Jing-Ming Guo
- Subjects
Scheme (programming language) ,General Computer Science ,Computer science ,Chaotic ,Chinese remainder theorem ,02 engineering and technology ,Encryption ,Secret sharing ,Image (mathematics) ,Set (abstract data type) ,Beta chaotic ,0202 electrical engineering, electronic engineering, information engineering ,General Materials Science ,Security level ,computer.programming_language ,business.industry ,General Engineering ,eXclusive-OR ,020206 networking & telecommunications ,image encryption ,secret sharing ,020201 artificial intelligence & image processing ,Confusion and diffusion ,lcsh:Electrical engineering. Electronics. Nuclear engineering ,business ,computer ,Algorithm ,lcsh:TK1-9971 - Abstract
This paper proposes a new simple image encryption technique using the Beta chaotic map for performing the confusion and diffusion of input plain image. This image encryption is further applied to the multiple secret sharing (MSS). The former existing scheme in MSS employs the Chinese remainder theorem (CRT) and eXclusive- OR (XOR) operations to convert the secret image to become a set of shared images. The former scheme yields a good result while the number of secret images is even. However, it has slight limitation while the number of secret images is odd. This paper aims to overcome this problem by utilizing the proposed image encryption along with three new MSS approaches: 1) incorporating an additional random images; 2) performing $k$ image encryption; and 3) using two different masking coefficients. As documented in the experimental section, the proposed method performs well in the MSS task. It solves the problem in former scheme on dealing with odd number of secret images. In addition, the proposed method is superior compared with the other competing schemes. The proposed image encryption also improves the security level in the MSS system.
- Published
- 2018
223. Fusion of Deep Learning and Compressed Domain Features for Content-Based Image Retrieval
- Author
-
Jing-Ming Guo, Peizhong Liu, Chi-Yi Wu, and Danlin Cai
- Subjects
Computer science ,Feature extraction ,02 engineering and technology ,Content-based image retrieval ,Convolutional neural network ,Text mining ,Image texture ,Histogram ,0202 electrical engineering, electronic engineering, information engineering ,Computer vision ,Visual Word ,Image retrieval ,business.industry ,Deep learning ,Dimensionality reduction ,Vector quantization ,Codebook ,020207 software engineering ,Pattern recognition ,computer.file_format ,Computer Graphics and Computer-Aided Design ,Block Truncation Coding ,Bitmap ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,computer ,Software - Abstract
This paper presents an effective image retrieval method by combining high-level features from convolutional neural network (CNN) model and low-level features from dot-diffused block truncation coding (DDBTC). The low-level features, e.g., texture and color, are constructed by vector quantization -indexed histogram from DDBTC bitmap, maximum, and minimum quantizers. Conversely, high-level features from CNN can effectively capture human perception. With the fusion of the DDBTC and CNN features, the extended deep learning two-layer codebook features is generated using the proposed two-layer codebook, dimension reduction, and similarity reweighting to improve the overall retrieval rate. Two metrics, average precision rate and average recall rate (ARR), are employed to examine various data sets. As documented in the experimental results, the proposed schemes can achieve superior performance compared with the state-of-the-art methods with either low-or high-level features in terms of the retrieval rate. Thus, it can be a strong candidate for various image retrieval related applications.
- Published
- 2017
- Full Text
- View/download PDF
224. Ocular Recognition for Blinking Eyes
- Author
-
Chen-Chieh Yao, Peizhong Liu, Jing-Ming Guo, Szu-Han Tseng, Jiann-Der Lee, Daxin Zhu, and KokSheik Wong
- Subjects
Support Vector Machine ,genetic structures ,Computer science ,Feature extraction ,Iris recognition ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,Facial recognition system ,Image Processing, Computer-Assisted ,0202 electrical engineering, electronic engineering, information engineering ,Feature (machine learning) ,Humans ,Three-dimensional face recognition ,Computer vision ,Flexibility (engineering) ,Blinking ,business.industry ,020208 electrical & electronic engineering ,Reproducibility of Results ,Eye movement ,Computer Graphics and Computer-Aided Design ,eye diseases ,Support vector machine ,Biometric Identification ,020201 artificial intelligence & image processing ,sense organs ,Artificial intelligence ,business ,Algorithms ,Software - Abstract
Ocular recognition is expected to provide a higher flexibility in handling practical applications as oppose to the iris recognition, which only works for the ideal open-eye case. However, the accuracy of the recent efforts is still far from satisfactory at uncontrollable conditions, such as eye blinking which implies any poses of eyes. To address these issues, the skin texture, eyelids, and additional geometrical features are employed. In addition, to achieve higher accuracy, sequential forward floating selection is utilized to select the best feature combinations. Finally, the non-linear support vector machine is applied for identification purpose. Experimental results demonstrate that the proposed algorithm achieves the best accuracy for both open eye and blinking eye scenarios. As a result, it offers greater flexibility for the prospective subjects during recognition as well as higher reliability for security.
- Published
- 2017
- Full Text
- View/download PDF
225. Multimedia Classification Using Bipolar Relation Graphs
- Author
-
Lingling An, Yun-Fu Liu, and Jing-Ming Guo
- Subjects
Theoretical computer science ,Computer science ,business.industry ,Feature extraction ,Probabilistic logic ,02 engineering and technology ,Overfitting ,Machine learning ,computer.software_genre ,Graph ,Computer Science Applications ,020204 information systems ,Signal Processing ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,Leverage (statistics) ,020201 artificial intelligence & image processing ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,Classifier (UML) ,computer - Abstract
Recent studies on category relations have shown the promising progress in addressing classification problems. Existing works independently consider the known relation and classifier optimization, and thus restrain the room for performance improvement. In this work, a new loss function is proposed to leverage the underlining relations among categories and classifiers. In addition, the bipolar relation (BR) graph is employed to formulate a general form for diverse relations. This bipolar graph is automatically learnt for reliving the constraints which may happen during the cost minimization. Extensive experiments on three benchmarks with various hypotheses and graphs demonstrate that our method can offer a significant performance improvement by jointly learning from both BR graph and hypothesis, in particular on a small training dataset scenario that suffers from severe overfitting problem.
- Published
- 2017
- Full Text
- View/download PDF
226. Contrast Enhancement Using Stratified Parametric-Oriented Histogram Equalization
- Author
-
Jie-Cyun Yu, Yun-Fu Liu, and Jing-Ming Guo
- Subjects
business.industry ,media_common.quotation_subject ,Cumulative distribution function ,Histogram matching ,020206 networking & telecommunications ,02 engineering and technology ,Histogram ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,Contrast (vision) ,020201 artificial intelligence & image processing ,Adaptive histogram equalization ,Computer vision ,Artificial intelligence ,Electrical and Electronic Engineering ,Visual artifact ,business ,Algorithm ,Histogram equalization ,Parametric statistics ,media_common ,Mathematics - Abstract
A contrast enhancement method termed stratified parametric-oriented histogram equalization (SPOHE) is proposed to effectively provide a regional enhanced effect without visual artifacts, e.g., halo or blocking artifacts, which is normally incurred in the former simplified enhancement methods. First, the stratified sampling theory is applied to uniformly sample the original image through many divided strata with the size defined by the two parameters ( $\alpha $ , $\beta $ ). Second, the required statistical information is efficiently derived through the integral image concept. Finally, the corrected SPOHE is also proposed to further improve the contrast with limited tradeoff computations. The experimental results demonstrate that the proposed scheme yields a cumulative distribution function similar to the actual one for an accurate contrast enhancement performance while significantly reducing the computational complexity. Moreover, compared with the former speed-oriented methods, good contrast and artifact-free results can be achieved simultaneously.
- Published
- 2017
- Full Text
- View/download PDF
227. Atrial Fibrillation Detection in Spectrogram Based on Convolution Neural Networks
- Author
-
Chih-Hsien Hsia, Zong-Hui Wang, Li-Ying Chang, Chiao-Chun Yang, and Jing-Ming Guo
- Subjects
Artificial neural network ,Heart disease ,Computer science ,business.industry ,Speech recognition ,Deep learning ,Atrial fibrillation ,medicine.disease ,Residual neural network ,Convolution ,Task (project management) ,medicine ,Spectrogram ,cardiovascular diseases ,Artificial intelligence ,business - Abstract
Nowadays, the computerized electrocardiogram (ECG) interpretation is the best available tool for the detection of heart diseases. The symptoms of atrial fibrillation, one of the heart disease, are the most challenging task of heart disease that relies on the cardiologist to read the ECG. However, this task involves a time-consuming process, leading to fatigue-induced medical errors. In this paper, a novel atrial fibrillation diagnostic algorithm based on deep learning architecture is proposed which incorporates the spectrogram to further improve the accuracy performance. As shown in the experimental results, the proposed method on PhysioNet /CinC Challenge 2017, which contains 8528 of single-lead ECG recordings, achieves the 10-fold cross-validation set performance of 78%.
- Published
- 2019
- Full Text
- View/download PDF
228. Reconstruction of Multitone BTC Images using Conditional Generative Adversarial Nets
- Author
-
S. Sankarasrinivasan and Jing-Ming Guo
- Subjects
Halftone ,Ordered dithering ,Image quality ,Computer science ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,020206 networking & telecommunications ,02 engineering and technology ,Iterative reconstruction ,Block Truncation Coding ,0202 electrical engineering, electronic engineering, information engineering ,Image translation ,020201 artificial intelligence & image processing ,Dither ,Algorithm ,Block (data storage) - Abstract
Multitone Block Truncation Coding (MT-BTC) image is the superior version of halftone based BTC compression methods. The MT-BTC images are developed based on ordered dithering method which utilize multitone dithering mask for image construction. During the transformation, the original image is processed in a block-wise manner and is replaced in terms of the maximum, minimum and their intermediate values of the respective block. In comparison with standard compressions techniques, MT-BTC possess very unique representation and suffers from inherent halftone noises. In this paper, a simplified version of image to image translation architecture is developed based on cGAN's. To begin with, a MT-BTC database is developed using the latest multitone approach, and it comprise of around 10,000 images. Further, the proposed cGAN's model is optimized to perform with minimal layers and reduced parameters. The PatchGAN discriminator is adjusted to judge over the patch size of 64×64 which has good impact over quality improvements. From the comprehensive performance evaluation, it has been validated that the proposed approach can achieve consistent and improved reconstruction quality.
- Published
- 2019
- Full Text
- View/download PDF
229. Improving EDBTC Image Quality Using Stationary and Decimated Wavelet Transform
- Author
-
Chih-Hsien Hsia, Jing-Ming Guo, and Heri Prasetyo
- Subjects
Discrete wavelet transform ,Image quality ,business.industry ,Computer science ,Stationary wavelet transform ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Wavelet transform ,020206 networking & telecommunications ,Pattern recognition ,Data_CODINGANDINFORMATIONTHEORY ,02 engineering and technology ,Iterative reconstruction ,Block Truncation Coding ,Wavelet ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,Noise (video) ,business - Abstract
This paper presents a new technique for improving the quality of Error Difussion Block Truncation Coding (EDBTC) decoded image. The proposed method exploits the usability of wavelet transform on decomposing the EDBTC decoded image into low and high frequency subbands. This new technique employs the two dimensional Discrete Wavelet Transform (DWT) and two dimensional Stationary Wavelet Transform (SWT). This scheme considers the EDBTC decoded image as noisy image in which the occurence of noise can be minimized by modifying the wavelet high frequency. As documented in experimental results, the proposed method yields promising result for the EDBTC image reconstruction.
- Published
- 2019
- Full Text
- View/download PDF
230. Friendly and Progressive Visual Secret Sharing with Lossless Reconstruction
- Author
-
Heri Prasetyo, Jing-Ming Guo, and Chih-Hsien Hsia
- Subjects
TheoryofComputation_MISCELLANEOUS ,Lossless compression ,Color image ,business.industry ,Computer science ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,020207 software engineering ,Usability ,02 engineering and technology ,Lossy compression ,Secret sharing ,Grayscale ,Image (mathematics) ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,business - Abstract
This paper presents a new simple and efficient technique for secret sharing. This technique inherits the effectiveness of Progressive Visual Secret Sharing (PVSS), in which the quality of recovered secret image is improved if more shared images are stacked together using the eXclusive-OR (XOR) operation. The proposed method also extends the usability of PVSS approach into the Friendly Visual Secret Sharing (PVSS). The content of shared image is converted into more friendly appearances rather than in the noise-like shared image. The proposed method offers promising result for grayscale and color image. It also produces recovered secret image in the lossless condition. The proposed method can be regarded as a good candidate for secret sharing.
- Published
- 2019
- Full Text
- View/download PDF
231. Image Semantic Segmentation With Edge and Feature Level Attenuators
- Author
-
Herleeyandi Markoni and Jing-Ming Guo
- Subjects
Computer science ,business.industry ,Feature extraction ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,Image segmentation ,Iterative reconstruction ,Object (computer science) ,Feature (computer vision) ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Segmentation ,Computer vision ,Artificial intelligence ,Enhanced Data Rates for GSM Evolution ,business ,Encoder - Abstract
Image segmentation is one of the popular techniques for vision application, i.e., retrieving object information from an image. Segmentation not only provides the class and location of an object, but the associated contour as well. Recent advances of the segmentation deploy the encoder and decoder architectures with skip connection, and also utilize the edge information to obtain the best contour. This work focuses on maintaining the information flow from skip connection, and also identifies suited feature from the bottom layer. The feature selector can be also deployed in edge information for yielding segmentation contour. Designed filters are also employed to analyze the best feature for the reconstruction purpose. Experiment results show that the proposed scheme can improvise the performance of the simple ENet to achieve a superior IoU from 51.3% to 57.9%.
- Published
- 2019
- Full Text
- View/download PDF
232. A Hybrid Facial Expression Recognition System Based on Recurrent Neural Network
- Author
-
Po-Cheng Huang, Jing-Ming Guo, and Li-Ying Chang
- Subjects
Facial expression ,Landmark ,business.industry ,Computer science ,Deep learning ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,020206 networking & telecommunications ,Pattern recognition ,02 engineering and technology ,stomatognathic diseases ,Recurrent neural network ,Facial expression recognition ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,Temporal information - Abstract
Facial expression recognition (FER) is an important and challenging problem for automatic inspection of surveillance videos. In recent years, with the progress of hardware and the evolution of deep learning technology, it is possible to change the way of tackling facial expression recognition. In this paper, we propose a sequence-based facial expression recognition framework for differentiating facial expression. The proposed framework is extended to a frame-to-sequence approach by exploiting temporal information with gated recurrent units. In addition, facial landmark points and facial action unit are also used as input features to train our network which can represent facial regions and its components effectively. Based on this, we build a robust facial expression system and is evaluated using two publicly available databases. The experimental results show that despite the uncontrolled factors in the videos, the proposed deep learning-based solution is consistent in achieving promising performance compared to that of the former schemes.
- Published
- 2019
- Full Text
- View/download PDF
233. Clustered-Dot Screen Design for Digital Multitoning
- Author
-
Jing-Ming Guo and Yun-Fu Liu
- Subjects
Pixel ,Noise measurement ,Computer science ,business.industry ,Image quality ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,020206 networking & telecommunications ,02 engineering and technology ,Computer Graphics and Computer-Aided Design ,Rendering (computer graphics) ,Visualization ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,business ,Software - Abstract
Digital multitoning is an extension of halftoning for rendering more than two tones at each pixel for higher image quality. Although a lot of effort has been put in generating dispersed dots previously, the blue-noise feature can hardly be achieved for those printers utilizing the electrophotography (EP) process to avoid the physically unstable isolated dots. To overcome this issue, Chandu et al. proposed a screening method for yielding green-noise dot clusters, yet noisy multitone texture was accompanied. This degrades the visual quality and the stability of tone rendering. In this paper, a significantly improved homogeneity of clustered dots can be achieved by the proposed screening method based upon the new inter-iterative clustered-dot direct multi-bit search algorithm. Compared with the former approaches, the inter-iteration design leads to less error by the updated initial multitone patterns. As demonstrated in the experimental results, both of the high homogenous multitone texture and less noisy perception at all absorptance levels are offered in contrast to the former Chandu et al. ’s results. The high-quality output proves it as a very competitive candidate for EP printers, e.g., laser printers.
- Published
- 2016
- Full Text
- View/download PDF
234. Hiding Multitone Watermarks in Halftone Images.
- Author
-
Jing-Ming Guo and Yun-Fu Liu
- Published
- 2010
- Full Text
- View/download PDF
235. Multi-Person Pose Estimation via Multi-Layer Fractal Network and Joints Kinship Pattern
- Author
-
Yongzhao Du, Zhitong Xu, Yanmin Luo, Jing-Ming Guo, and Peizhong Liu
- Subjects
Matching (graph theory) ,Computer science ,business.industry ,Process (computing) ,Pattern recognition ,Computer Graphics and Computer-Aided Design ,Convolutional neural network ,Image (mathematics) ,Fractal ,Feature (computer vision) ,Pattern matching ,Artificial intelligence ,business ,Pose ,Software - Abstract
We propose an effective method to boost the accuracy of multi-person pose estimation in images. Initially, the three-layer fractal network was constructed to regress multi-person joints location heatmap that can help to enhance an image region with receptive field and capture more joints local-contextual feature information, thereby producing keypoints heatmap intermediate prediction to optimize human body joints regression results. Subsequently, the hierarchical bi-directional inference algorithm was proposed to calculate the degree of relatedness (call it Kinship) for adjacent joints, and it combines the Kinship between adjacent joints with the spatial constraints, which we refer to as joints kinship pattern matching mechanism, to determine the best matched joints pair. We iterate the above-mentioned joints matching process layer by layer until all joints are assigned to a corresponding individual. Comprehensive experiments demonstrate that the proposed approach outperforms the state-of-the-art schemes and achieves about 1% and 0.6% increase in mAP on MPII multi-person subset and MSCOCO 2016 keypoints challenge.
- Published
- 2018
236. Error Diffused Halftohe Classification Using Stochastic Geometrical Features
- Author
-
Jing-Ming Guo and S. Sankarasrinivasan
- Subjects
0209 industrial biotechnology ,Halftone ,Computer science ,Feature vector ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,Iterative reconstruction ,Grayscale ,020901 industrial engineering & automation ,Kernel (image processing) ,Computer Science::Multimedia ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Algorithm ,ComputingMethodologies_COMPUTERGRAPHICS - Abstract
Digital halftoning is a strategy to convert the color or gray scale image into a printable format. Among several halftoning techniques, error diffusion is one of the conventional and widely adopted method in many printing devices. In further, halftone classification is very important to obtain the perfect reconstruction of binary printed images. The paper attempts to solve this issue by exploiting the intrinsic similarity between the stochastic geometry and halftoning. Feature vectors are constructed using the point process statistic parameters such as directional distribution function and radially averaged power spectral density. Extreme learning machine model is developed and eight varieties of error diffusion halftone images are considered for classification. From the results, it has been validated that the proposed scheme yields better accuracy of 97.5% and it is faster than the existing approaches.
- Published
- 2018
- Full Text
- View/download PDF
237. Tone-Replacement Error Diffusion for Multitoning
- Author
-
Yun-Fu Liu, Guo-Hong Lai, Jing-Ming Guo, Jiann-Der Lee, and Jia-Yu Chang
- Subjects
Halftone ,Similarity (geometry) ,Visual perception ,business.industry ,Computer science ,Image quality ,Iterative reconstruction ,Computer Graphics and Computer-Aided Design ,Image (mathematics) ,Error diffusion ,Tone (musical instrument) ,Computer vision ,Artificial intelligence ,business ,Software - Abstract
Error diffusion is an efficient halftone method for mainly being applied on printers. The promising high image quality and processing efficiency endorse it as a popular and competitive candidate in halftoning and multitoning applications. The multitoning is an extension of halftoning, adopting more than two-tone levels for the improvement of the similarity between an original image and the converted image. Yet, the banding effect, indicating the areas with discontinuous tone level, disturbs the visual perception, and thus seriously degrades image quality. To solve the banding effect, the tone-replacement strategy is proposed in this paper. As documented in the experimental results, excellent tone-similarity as that of the original image and promising reconstructed dot-distribution can be provided simultaneously. Comparing with the former banding-free methods, the apparent improvements/features suggest that the proposed method can be a very competitive candidate for multitoning applications.
- Published
- 2015
- Full Text
- View/download PDF
238. Dot-Diffused Halftoning With Improved Homogeneity
- Author
-
Jing-Ming Guo and Yun-Fu Liu
- Subjects
FOS: Computer and information sciences ,Halftone ,Ordered dithering ,Pixel ,Computer science ,Iterative method ,Computer Science - Information Theory ,Information Theory (cs.IT) ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Spectral density ,Computer Graphics and Computer-Aided Design ,Multimedia (cs.MM) ,Sampling (signal processing) ,Parallel processing (DSP implementation) ,Colors of noise ,Algorithm ,Computer Science - Multimedia ,Software - Abstract
Compared to the error diffusion, dot diffusion provides an additional pixel-level parallelism for digital halftoning. However, even though its periodic and blocking artifacts had been eased by previous works, it was still far from satisfactory in terms of the blue noise spectrum perspective. In this work, we strengthen the relationship among the pixel locations of the same processing order by an iterative halftoning method, and the results demonstrate a significant improvement. Moreover, a new approach of deriving the averaged power spectrum density (APSD) is proposed to avoid the regular sampling of the well-known Bartlett's procedure which inaccurately presents the halftone periodicity of certain halftoning techniques with parallelism. As a result, the proposed dot diffusion is substantially superior to the state-of-the-art parallel halftoning methods in terms of visual quality and artifact-free property, and competitive runtime to the theoretical fastest ordered dithering is offered simultaneously., Comment: Accepted to IEEE Trans. on Image Processing
- Published
- 2015
- Full Text
- View/download PDF
239. Vehicle Verification Using Features From Curvelet Transform and Generalized Gaussian Distribution Modeling
- Author
-
Heri Prasetyo, Jing-Ming Guo, Mahmoud E. Farfoura, and Hua Lee
- Subjects
Laplace transform ,business.industry ,Mechanical Engineering ,Gaussian ,Feature extraction ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Statistical parameter ,Pattern recognition ,Object detection ,Computer Science Applications ,symbols.namesake ,Computer Science::Computer Vision and Pattern Recognition ,Computer Science::Multimedia ,Automotive Engineering ,symbols ,Artificial intelligence ,Marginal distribution ,business ,Classifier (UML) ,Generalized normal distribution ,Mathematics - Abstract
This paper presents a new feature descriptor for vehicle verification. The object detection scheme generates the vehicle hypothesis (candidate) that requires subsequent confirmation in the vehicle verification stage with specific feature descriptors. In the procedure of vehicle verification, an image descriptor is generated from the statistical parameter of the curvelet-transformed (CT) subbands. The marginal distribution of CT output is a heavy-tailed bell-shaped function, which can be approximated as Gaussian, Laplace, and generalized Gaussian distribution (GGD) with high accuracy. The maximum likelihood estimation (MLE) produces the distribution parameters of each CT subband for the generation of the image feature descriptor. The classifier then assigns a class label for the vehicle hypothesis based on this descriptor information. As documented in the experimental results, this feature descriptor is effective and outperforms the existing methods in the vehicle verification tasks.
- Published
- 2015
- Full Text
- View/download PDF
240. Content-Based Image Retrieval Using Error Diffusion Block Truncation Coding Features
- Author
-
Jen-Ho Chen, Jing-Ming Guo, and Heri Prasetyo
- Subjects
Color histogram ,Image quality ,Computer science ,Feature extraction ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Top-hat transform ,Image processing ,Content-based image retrieval ,Fractal ,Image texture ,Histogram ,Media Technology ,Feature descriptor ,Computer vision ,Visual Word ,Electrical and Electronic Engineering ,Image warping ,Image retrieval ,Image gradient ,Feature detection (computer vision) ,business.industry ,Color image ,Binary image ,Quantization (signal processing) ,Vector quantization ,Pattern recognition ,Block Truncation Coding ,Automatic image annotation ,Feature (computer vision) ,Artificial intelligence ,business ,Image histogram ,Image compression ,Color Cell Compression - Abstract
This paper presents a new approach to index color images using the features extracted from the error diffusion block truncation coding (EDBTC). The EDBTC produces two color quantizers and a bitmap image, which are further processed using vector quantization (VQ) to generate the image feature descriptor. Herein two features are introduced, namely, color histogram feature (CHF) and bit pattern histogram feature (BHF), to measure the similarity between a query image and the target image in database. The CHF and BHF are computed from the VQ-indexed color quantizer and VQ-indexed bitmap image, respectively. The distance computed from CHF and BHF can be utilized to measure the similarity between two images. As documented in the experimental result, the proposed indexing method outperforms the former block truncation coding based image indexing and the other existing image retrieval schemes with natural and textural data sets. Thus, the proposed EDBTC is not only examined with good capability for image compression but also offers an effective way to index images for the content-based image retrieval system.
- Published
- 2015
- Full Text
- View/download PDF
241. Multi-mode Halftoning Using Stochastic Clustered-Dot Screen
- Author
-
Shih-Chieh Lin, Jing-Ming Guo, and Yun-Fu Liu
- Subjects
Artifact (error) ,Halftone ,Computer science ,business.industry ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Stacking ,Boundary (topology) ,Pattern recognition ,Object (computer science) ,Image (mathematics) ,Constraint (information theory) ,Quality (physics) ,Artificial intelligence ,business ,ComputingMethodologies_COMPUTERGRAPHICS - Abstract
The conventional dual-mode halftoning methods achieve high quality halftone patterns by easing smooth artifact and preserving image details. However, the boundaries between low- and high-frequency image regions still present undesired texture in some particular cases, which significantly degrades the visual quality. In this work, a multi-mode halftoning method is proposed to deal with the object map artifact and boundary artifact simultaneously. In addition, the new absorptance-frequency stacking constraint is also employed to solve the noisy textures of the halftone outputs. As documented in the experimental results, high quality halftone outputs can be obtained, proving that the proposed method can be a competitive candidate for electrophotographic printers.
- Published
- 2018
- Full Text
- View/download PDF
242. An Efficient Fusion-Based Defogging
- Author
-
Jin-yu Syue, Vincent R. Radzicki, Jing-Ming Guo, and Hua Lee
- Subjects
Haze ,Channel (digital image) ,business.industry ,Computer science ,Flicker ,020206 networking & telecommunications ,02 engineering and technology ,Computer Graphics and Computer-Aided Design ,Light scattering ,Weighting ,Transmission (telecommunications) ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Algorithm design ,Computer vision ,Artificial intelligence ,business ,Visibility ,Focus (optics) ,Software ,Image restoration - Abstract
Degradation in visibility is often introduced to images captured in poor weather conditions, such as fog or haze. To overcome this problem, conventional approaches focus mainly on the enhancement of the overall image contrast. However, because of the unspecified light-source distribution or unsuitable mathematical constraints of the cost functions, it is often difficult to achieve quality results. In this paper, a fusion-based transmission estimation method is introduced to adaptively combine two different transmission models. Specifically, the new fusion weighting scheme and the atmospheric light computed from the Gaussian-based dark channel method improve the estimation of the locations of the light sources. To reduce the flickering effect introduced during the process of frame-based dehazing, a flicker-free module is formulated to alleviate the impacts. The systematic assessments show that this approach is capable of achieving superior defogging and dehazing performance, compared with superior defogging and dehazing performance, compared with the state-of-the-art methods, both quantitatively and qualitatively.
- Published
- 2017
243. Sample Space Dimensionality Refinement for Symmetrical Object Detection
- Author
-
Chih-Hsien Hsia, Jing-Ming Guo, Hua Lee, Sheng-Yao Su, and Yun-Fu Liu
- Subjects
Computer Networks and Communications ,business.industry ,Computer science ,Dimensionality reduction ,Pedestrian detection ,Feature extraction ,Pattern recognition ,Object detection ,Reduction (complexity) ,Pattern recognition (psychology) ,Artificial intelligence ,Safety, Risk, Reliability and Quality ,Face detection ,business ,Curse of dimensionality - Abstract
Formerly, dimensionality reduction techniques are effective ways for extracting statistical significance of features from their original dimensions. However, the dimensionality reduction also induces an additional complexity burden which may encumber the real efficiency. In this paper, a technique is proposed for the reduction of the dimension of samples rather than the features in the former schemes, and it is able to additionally reduce the computational complexity of the applied systems during the reduction process. This method effectively reduces the redundancies of a sample, in particular for those objects which possess partially symmetric property, such as human face, pedestrian, and license plate. As demonstrated in the experiments, based upon the premises of faster speeds in training and detection by a factor of 4.06 and 1.24, respectively, similar accuracies to the ones without considering the proposed method are achieved. The performance verifies that the proposed technique can offer competitive practical values in pattern recognition related fields.
- Published
- 2014
- Full Text
- View/download PDF
244. Security analyses of the watermarking scheme based on redundant discrete wavelet transform and singular value decomposition
- Author
-
Jing-Ming Guo and Heri Prasetyo
- Subjects
Discrete wavelet transform ,Theoretical computer science ,business.industry ,Data_MISCELLANEOUS ,ComputingMilieux_LEGALASPECTSOFCOMPUTING ,Watermark ,Pattern recognition ,Image processing ,Grayscale ,Singular value ,Redundancy (information theory) ,Singular value decomposition ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,Digital watermarking ,Mathematics - Abstract
This study analyzes the recent image watermarking schemes based on redundant discrete wavelet transform (RDWT) and singular value decomposition (SVD), and shows that in fact they are insecure and cannot be used for protecting the rightful ownership. The RDWT-SVD watermarking directly embeds a grayscale watermark image of the same size with the host image into the singular value matrix of the RDWT-transformed host image, then produces the left and right orthogonal matrices as side information which is later used in the watermark extraction stage. The RDWT-SVD approach enjoys the advantage of the RDWT redundancy to achieve a high embedding capacity, and preserves the watermark imperceptibility by exploiting the SVD stability properties. It is claimed that RDWT-SVD watermarking is robust against several common image processing and geometrical attacks, yet a fundamental flaw in the RDWT-SVD scheme is found, which leads to severe the false positive issue. Three vulnerable attacks should be considered in the RDWT-SVD scheme: (1) An attacker can easily claim the owner watermarked image; (2) the owner has the ambiguity because of the wrong side information usage, and (3) the owner can extract the correct watermark from arbitrary image. Thus, it is important to highlight these attacks when implementing the RDWT-SVD watermarking scheme.
- Published
- 2014
- Full Text
- View/download PDF
245. Fast Background Subtraction Based on a Multilayer Codebook Model for Moving Object Detection
- Author
-
Min-Hsiung Shih, Chih-Hsien Hsia, Jing-Yu Wu, Yun-Fu Liu, Jing-Ming Guo, and Cheng-Hsin Chang
- Subjects
Background subtraction ,Contextual image classification ,Pixel ,business.industry ,Computer science ,Feature extraction ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Codebook ,Pattern recognition ,Object detection ,Video tracking ,Media Technology ,Computer vision ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,Block (data storage) - Abstract
Moving object detection is an important and fundamental step for intelligent video surveillance systems because it provides a focus of attention for post-processing. A multilayer codebook-based background subtraction (MCBS) model is proposed for video sequences to detect moving objects. Combining the multilayer block-based strategy and the adaptive feature extraction from blocks of various sizes, the proposed method can remove most of the nonstationary (dynamic) background and significantly increase the processing efficiency. Moreover, the pixel-based classification is adopted for refining the results from the block-based background subtraction, which can further classify pixels as foreground, shadows, and highlights. As a result, the proposed scheme can provide a high precision and efficient processing speed to meet the requirements of real-time moving object detection.
- Published
- 2013
- Full Text
- View/download PDF
246. Reversible Data Hiding Scheme with High Embedding Capacity Using Semi-Indicator-Free Strategy
- Author
-
Jing-Ming Guo, Yaw-Hwang Chiou, and Jiann-Der Lee
- Subjects
Scheme (programming language) ,Theoretical computer science ,Article Subject ,Computer science ,General Mathematics ,Data_CODINGANDINFORMATIONTHEORY ,ENCODE ,Image (mathematics) ,Histogram ,Computer vision ,Mathematics ,computer.programming_language ,Lossless compression ,business.industry ,lcsh:Mathematics ,Vector quantization ,General Engineering ,lcsh:QA1-939 ,lcsh:TA1-2040 ,Information hiding ,Bit rate ,Embedding ,Artificial intelligence ,Low bit rate ,business ,lcsh:Engineering (General). Civil engineering (General) ,Algorithm ,computer - Abstract
This work presents a novel reversible data-hiding scheme which embeds secret data into a side matched Vector Quantization (SMVQ)-compressed image, and achieves lossless reconstruction of a Vector Quantization (VQ)-compressed image. The rather random distributed histogram of a VQ-compressed image can be re-located to locations close to zero by SMVQ prediction. Thus, fewer bits can be used to encode SMVQ indices with very small values, and no indicator is required to encode these indices, which yields extra hiding space to hide secret data. Consequently, high embedding capacity and low bit rate scenarios can be deposited. Experimental results demonstrate the effectiveness and reversibility of the proposed scheme. Moreover, in terms of the embedding rate, the bit rate, and the embedding capacity, experimental results show that the performance of the proposed scheme is better than those of the former data hiding schemes for VQ-based and VQ/SMVQ-based compressed images.
- Published
- 2013
- Full Text
- View/download PDF
247. Adaptive block truncation coding image compression technique using optimized dot diffusion
- Author
-
Yu Cheng, Jing-Ming Guo, and Yun-Fu Liu
- Subjects
Artifact (error) ,Computer science ,Image quality ,020208 electrical & electronic engineering ,Real-time computing ,Image processing ,02 engineering and technology ,Impulse noise ,Blocking (statistics) ,Block Truncation Coding ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Algorithm ,Image compression - Abstract
Block truncation coding (BTC) has been considered as a highly efficient compression technique for decades, but the blocking artifact is its main issue. The halftoning-based BTC has significantly eased this issue, yet an apparent impulse noise artifact is accompanied. In this study, an improved BTC, termed adaptive dot-diffused BTC (ADBTC), is proposed to further improve the visual quality. Also, this method provides an additional flexibility on the compression ratios determination in contrast to the former fixed and few number of configuration possibilities. As documented in the experimental results, the proposed method achieves the superior image quality regarding the five various objective IQA methods. As a result, it is a very competitive approach for the needs of both high frame rate and high-resolution image compression.
- Published
- 2016
- Full Text
- View/download PDF
248. High Capacity Data Hiding for Error-Diffused Block Truncation Coding
- Author
-
Yun-Fu Liu and Jing-Ming Guo
- Subjects
Image quality ,business.industry ,Watermark ,computer.file_format ,Computer Graphics and Computer-Aided Design ,Block Truncation Coding ,Robustness (computer science) ,Information hiding ,Bitmap ,Computer vision ,Artificial intelligence ,business ,Algorithm ,computer ,Software ,Color Cell Compression ,Data compression ,Mathematics - Abstract
Block truncation coding (BTC) is an efficient compression technique with extremely low computational complexity. However, the blocking and false contour effects are two major deficiencies in BTC which cause severe perceptual artifacts. The former scheme, error-diffused BTC (EDBTC), can significantly improve the above issues through the visual low-pass compensation on the bitmap, which thus widens its possible application market, yet the corresponding security issue may limit its value. In this paper, a method namely complementary hiding EDBTC is developed to cope the above issue. This paper is managed by firstly discussing when a single watermark is embedded, and then multiple watermarks are employed to test the limitation of the proposed scheme. Herein, an adaptive external bias factor is employed to control the watermark embedding, and this factor also affects the image quality and robustness simultaneously. Experimental results demonstrate that the proposed method only requires an extremely small external bias factor to carry watermarks, which enables a high capacity scenario without significantly damaging image quality.
- Published
- 2012
- Full Text
- View/download PDF
249. Impact of the Lips for Biometrics
- Author
-
Jing-Ming Guo, Yun-Fu Liu, and Chao-Yu Lin
- Subjects
Databases, Factual ,Biometrics ,business.industry ,Computer science ,Feature extraction ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Frame rate ,Computer Graphics and Computer-Aided Design ,Grayscale ,Facial recognition system ,Lip ,Support vector machine ,Biometric Identification ,Pattern recognition (psychology) ,Humans ,Computer vision ,Artificial intelligence ,business ,Rotation (mathematics) ,Algorithms ,Software - Abstract
In this paper, the impact of the lips for identity recognition is investigated. In fact, it is a challenging issue for identity recognition solely by the lips. In the first stage of the proposed system, a fast box filtering is proposed to generate a noise-free source with high processing efficiency. Afterward, five various mouth corners are detected through the proposed system, in which it is also able to resist shadow, beard, and rotation problems. For the feature extraction, two geometric ratios and ten parabolic-related parameters are adopted for further recognition through the support vector machine. Experimental results demonstrate that, when the number of subjects is fewer or equal to 29, the correct accept rate (CAR) is greater than 98%, and the false accept rate (FAR) is smaller than 0.066%. (CAR95.02%, FAR0.095% # Subjects ≤ 57). Moreover, the processing speed of the overall system achieves 34.43 frames per second, which meets the real-time requirement. Thus, the proposed system can be an effective candidate for facial biometrics applications when other facial organs are covered or when it is applied for an access control system.
- Published
- 2012
- Full Text
- View/download PDF
250. Improved Hand Tracking System
- Author
-
Yun-Fu Liu, Jing-Ming Guo, Hoang-Son Nguyen, and Che-Hao Chang
- Subjects
Foreground detection ,Computer science ,business.industry ,Feature extraction ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Pattern recognition ,Tracking system ,Image segmentation ,Object detection ,Feature (computer vision) ,Video tracking ,Media Technology ,Segmentation ,Computer vision ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,Pose - Abstract
This paper presents an improved hand tracking system using pixel-based hierarchical-feature AdaBoosting (PBHFA), skin color segmentation, and codebook (CB) background cancellation. The proposed PBH feature significantly reduces the training time by a factor of at least 1440 compared to the traditional Haar-like feature. Moreover, lower computation and high tracking accuracy are also provided simultaneously. Yet, one of the disadvantages of the PBHFA is the false positive which is the consequence of the appearance of complex background in positive samples. To effectively reduce the false positive rate, the skin color segmentation and the foreground detection by applying the CB model are catered for rejecting all of the candidates which are not hand targets. As documented in the experimental results, the proposed system can achieve promising results, and thus it can be considered as an effective candidate in handling practical applications which require hand postures.
- Published
- 2012
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.