1,430 results on '"Chrominance"'
Search Results
2. PVO Based Reversible Secret Data Hiding Technique in YCbCr Color Space
- Author
-
Kumar, Neeraj, Singh, Dinesh Kumar, Awasthi, Shashank, Angrisani, Leopoldo, Series Editor, Arteaga, Marco, Series Editor, Panigrahi, Bijaya Ketan, Series Editor, Chakraborty, Samarjit, Series Editor, Chen, Jiming, Series Editor, Chen, Shanben, Series Editor, Chen, Tan Kay, Series Editor, Dillmann, Rüdiger, Series Editor, Duan, Haibin, Series Editor, Ferrari, Gianluigi, Series Editor, Ferre, Manuel, Series Editor, Hirche, Sandra, Series Editor, Jabbari, Faryar, Series Editor, Jia, Limin, Series Editor, Kacprzyk, Janusz, Series Editor, Khamis, Alaa, Series Editor, Kroeger, Torsten, Series Editor, Li, Yong, Series Editor, Liang, Qilian, Series Editor, Martín, Ferran, Series Editor, Ming, Tan Cher, Series Editor, Minker, Wolfgang, Series Editor, Misra, Pradeep, Series Editor, Möller, Sebastian, Series Editor, Mukhopadhyay, Subhas, Series Editor, Ning, Cun-Zheng, Series Editor, Nishida, Toyoaki, Series Editor, Pascucci, Federica, Series Editor, Qin, Yong, Series Editor, Seng, Gan Woon, Series Editor, Speidel, Joachim, Series Editor, Veiga, Germano, Series Editor, Wu, Haitao, Series Editor, Zamboni, Walter, Series Editor, Zhang, Junjie James, Series Editor, Sanyal, Goutam, editor, Travieso-González, Carlos M., editor, Awasthi, Shashank, editor, Pinto, Carla M.A., editor, and Purushothama, B. R., editor
- Published
- 2022
- Full Text
- View/download PDF
3. Color Texture Analysis: A Survey
- Author
-
Anne Humeau-Heurtier
- Subjects
Chrominance ,classification ,color texture ,feature extraction ,image processing ,image synthesis ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
In the field of image processing, texture features and color are fundamental visual cue with complementary roles. They are used in many applications and in a large variety of areas such as quality control, content-based image retrieval, remote sensing, industrial inspection, surface inspection, object recognition, and medical image analysis. For this purpose, a large number of algorithms have been proposed for texture feature extraction. Some of them are dedicated to gray-scale images while others aim at processing both color and texture. It has been shown that, for many cases, the use of color improves the performance of gray level texture classification. This paper provides a comprehensive survey of the texture feature extraction methods that consider both texture and color information. We propose a categorization of these methods into seven classes, two of them being very recent. For each method, we present the concept, the advantages and drawbacks, and we give examples of application.
- Published
- 2022
- Full Text
- View/download PDF
4. Blind Quality Assessment of PFA-Affected Images Based on Chromatic Eigenvalue Ratio
- Author
-
Karthik, Kannan, Malik, Parveen, Kacprzyk, Janusz, Series Editor, Pal, Nikhil R., Advisory Editor, Bello Perez, Rafael, Advisory Editor, Corchado, Emilio S., Advisory Editor, Hagras, Hani, Advisory Editor, Kóczy, László T., Advisory Editor, Kreinovich, Vladik, Advisory Editor, Lin, Chin-Teng, Advisory Editor, Lu, Jie, Advisory Editor, Melin, Patricia, Advisory Editor, Nedjah, Nadia, Advisory Editor, Nguyen, Ngoc Thanh, Advisory Editor, Wang, Jun, Advisory Editor, Chaudhuri, Bidyut B., editor, Nakagawa, Masaki, editor, Khanna, Pritee, editor, and Kumar, Sanjeev, editor
- Published
- 2020
- Full Text
- View/download PDF
5. Passive Image Forgery Detection Based on the Demosaicing Algorithm and JPEG Compression
- Author
-
Esteban Alejandro Armas Vega, Edgar Gonzalez Fernandez, Ana Lucila Sandoval Orozco, and Luis Javier Garcia Villalba
- Subjects
Blind technique ,chrominance ,copy-move ,digital image ,forensics analysis ,forgery detection ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Multimedia files play an important role in everyday life. Today, the majority of the population owns state-of-the-art cameras integrated into their mobile devices. Technological development not only facilitates the generation of multimedia content, but also the intentional manipulation of it, and this is where forensic techniques of detecting manipulation on images and videos take on great importance. Although historically there has been confidence in the integrity of images, the advance of technology has begun to erode this confidence. This work proposes a digital image authentication method based on the quadratic mean error of the Color Filter Array interpolation pattern estimated from the analysed image. For the evaluation of the proposed method, experiments were carried out with public databases of forged images that are widely developed for research purposes. The results of the experiments demonstrate the efficiency of the proposed method.
- Published
- 2020
- Full Text
- View/download PDF
6. Image splicing detection technique based on Illumination-Reflectance model and LBP.
- Author
-
Niyishaka, Patrick and Bhagvati, Chakravarthy
- Subjects
NAIVE Bayes classification ,FISHER discriminant analysis ,SUPPORT vector machines ,MACHINE learning ,DEEP learning - Abstract
A Copy-create digital image forgery is image tampering that merges two or more areas of images from different sources into one composite image; it is also known as image splicing. Excellent forgeries are so tricky that they are not noticeable to the naked eye and don't reveal traces of tampering to traditional image tamper detection techniques. To tackle this image splicing detection problem, machines learning-based techniques are used to instantly discriminate between the authentic and forged image. Numerous image forgery detection methods to detect and localize spliced areas in the composite image have been proposed. However, the existing methods with high detection accuracy are computationally expensive since most of them are based on hybrid feature set or rely on the complex deep learning models, which are very expensive to train, run on expensive GPUs, and require a very large amount of data to perform better. In this paper, we propose a simple and computationally efficient image splicing forgery detection that considers a trade-off between performance and the cost to the users. Our method involves the following steps: first, luminance and chrominance are found from the input image; second, illumination is estimated from Luminance using Illumination-Reflectance model; third, Local Binary Patterns normalized histogram for illumination and Chrominance is computed and used as the feature vector for classification using the following machine learning algorithms: Support Vector Machine, Linear Discriminant Analysis, Logistic Regression, K-Nearest Neighbors, Decision Tree, and Naive Bayes. Extensive experiments on the public dataset CASIA v2.0 show that the new algorithm is computationally efficient and effective for image splicing tampering detection. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
7. Dental Shade Matching Method Based on Hue, Saturation, Value Color Model with Machine Learning and Fuzzy Decision.
- Abstract
Color information is an important indicator of color matching. It is recommended to use hue (H) and saturation (S) to improve the accuracy of color analysis. The proposed method for dental shade matching in this study is based on the hue, saturation, value (HSV) color model. To evaluate the performance of the proposed method in matching dental shades, peak signalto- noise ratio (PSNR), structural similarity index (SSIM), composite peak signal-to-noise ratio (CPSNR), and S-CIELAB (Special International Commission on Illumination, L* for lightness, a* from green to red, and b* from blue to yellow) were utilized. To further improve the performance of the proposed method, dental image samples were multiplied by the weighted coefficients derived by training the model using machine learning to reduce errors. Thus, the PSNR of 97.64% was enhanced to 99.93% when applied with the proposed fuzzy decision model. Results show that the proposed method based on the new fuzzy decision technology is effective and has an accuracy of 99.78%, which is a significant improvement of previous results. The new fuzzy decision is a method that combines the HSV color model, PSNR(H), PSNR(S), and SSIM information, which are used for the first time in research on tooth color matching. Results show that the proposed method performs better than previous methods. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
8. 草莓品质光谱检测方法比较研究.
- Author
-
宁晓峰, 刘娜, 陈永亮, 田素博, and 宫元娟
- Abstract
Copyright of Journal of Shenyang Agricultural University is the property of Journal of Shenyang Agricultural University Editorial Department and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2020
- Full Text
- View/download PDF
9. 成纱号数与捻系数对色纺纱色差影响的研究.
- Author
-
程璐, 葛梦嘉, 曹吉强, and 夏鑫
- Subjects
MANUFACTURING processes ,FACTOR analysis ,SPUN yarns ,YARN ,COLORING matter ,FIBERS - Abstract
Copyright of Cotton Textile Technology is the property of Cotton Textile Technology Editorial Office and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
- Published
- 2020
10. A Secure Image Steganography Technique to Hide Multiple Secret Images
- Author
-
Hemalatha, S., Dinesh Acharya, U., Renuka, A., Kamath, Priya R., Chaki, Nabendu, editor, Meghanathan, Natarajan, editor, and Nagamalai, Dhinaharan, editor
- Published
- 2013
- Full Text
- View/download PDF
11. Enhancement Layer Coding for Chroma Sub-Sampled Screen Content Video
- Author
-
Andre Kaup, Alexander Gehlert, Benjamin Prestele, and Andreas Heindel
- Subjects
Lossless compression ,business.industry ,Computer science ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,YCbCr ,Media Technology ,Chrominance ,RGB color model ,Codec ,Computer vision ,Artificial intelligence ,Electrical and Electronic Engineering ,Layer (object-oriented design) ,Joint (audio engineering) ,business ,Coding (social sciences) - Abstract
Prevalent video codec implementations deployed in the field often solely support chrominance sub-sampled video data represented by the YCbCr 4:2:0 format. For certain applications like screen sharing, however, chroma sub-sampling leads to disturbing artifacts, especially for text or graphics with thin lines. It is desirable to reduce these artifacts while maintaining compatibility with all user devices. For this reason, an enhancement layer coding framework for chroma-format scalable coding with specific focus on screen content is proposed in this paper. Based on an analysis of screen content data characteristics, the enhancement layer codec is optimized specifically for this content class, is of low algorithmic complexity, and applicable with any image or video codec for base layer compression in a joint coding system. The system is intentionally not designed as a general purpose lossless YCbCr 4:4:4 coding scheme, but instead chooses to close a quality gap that prevalent video codecs do not address. Experimental analysis reveals an average BD-PSNR gain in the RGB domain of 1.0 dB comparing the proposed two-layer scalable coding approach to single layer compression using base layer coding only. In relation to simulcast coding, an average BDR RGB difference of 8.1% is observed.
- Published
- 2022
- Full Text
- View/download PDF
12. Tensor Product and Tensor-Singular Value Decomposition Based Multi-Exposure Fusion of Images
- Author
-
Yang Song, Haiyong Xu, Mei Yu, Gangyi Jiang, Zhongjie Zhu, Yongqiang Bai, and Huifang Sun
- Subjects
Image fusion ,Computer science ,business.industry ,Pattern recognition ,Luminance ,Computer Science Applications ,Term (time) ,Tensor product ,Feature (computer vision) ,Tensor (intrinsic definition) ,Signal Processing ,Singular value decomposition ,Media Technology ,Chrominance ,Artificial intelligence ,Electrical and Electronic Engineering ,business - Abstract
Considering multidimensional structure of the multi-exposure images,a new Tensor product and Tensorsingular value decomposition based Multi-Exposure image Fusion (TT-MEF) method is proposed. The main innovation of this work is to explore a new feature representation of multi-exposure images in the new tensor domain and design the fusion strategy on this basis. Specifically,the luminance and the chrominance channels are fused separately to maintain color consistency. For the luminance fusion,the luminance channel of multi-exposure images is divided into two parts,that is,de-mean term and mean term. The de-mean term is represented as a tensor to extract the feature. Then,the tensor product and tensor-singular value decomposition (T-SVD) are used to design a tensor feature extractor. Furthermore,a fusion strategy of the de-mean term is presented according to the visual saliency model,and a fusion strategy of the mean term is defined by the local and the global visual weights to control counterpoise between the local and global luminance. For the chrominance fusion,a new fusion strategy is also designed by the tensor product and T-SVD,similar to the luminance fusion. Finally,the fused image is obtained by combining the luminance and chrominance fusion. Experimental results show that the proposed TT-MEF method generally outperforms the existing state-of-the-art in terms of subjective visual quality and objective evaluation.
- Published
- 2022
- Full Text
- View/download PDF
13. Objective Quality Assessment of Lenslet Light Field Image Based on Focus Stack
- Author
-
Ping An, Xinpeng Huang, Liquan Shen, Bin Wang, Chunli Meng, and Chao Yang
- Subjects
business.industry ,Computer science ,Image quality ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Lenslet ,Computer Science Applications ,Phase congruency ,Feature (computer vision) ,Signal Processing ,Media Technology ,Chrominance ,Computer vision ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,Focus (optics) ,Spatial analysis ,Light field - Abstract
The huge amount of complex scene information recorded by light field imaging has the prospect of immersive media applications. Compression and reconstruction algorithms are crucial for the transmission, storage, and display of such massive data. Most of the existing quality evaluation indexes do not take an effective account of light field characteristics. To accurately evaluate the distortions caused by compression and reconstruction algorithms, it is necessary to construct an image evaluation index that reflects the angular-spatial characteristic of the light field. This work proposes a full reference light field image quality evaluation index, which attempts to extract less information from the focus stack to accurately evaluate the entire light field quality. The proposed framework includes three specific steps. Firstly, we construct a key refocused images extraction framework by the maximal spatial information contrast and the minimal angular information variation. Specifically, the gradient and phase congruency operators are used in the extraction framework. Secondly, a novel light field quality evaluation index is built based on the angular-spatial characteristic of the key refocused images. In detail, the features used in the key refocused images extraction framework and the chrominance feature are combined to construct the union feature. Then the similarity of the union feature is pooled by the relevant visual saliency map to get the predicted score. Finally, the overall quality of the light field is measured by applying the proposed index to the key refocused images. The high efficiency and precision of the proposed method are shown by extensive comparison experiments.
- Published
- 2022
- Full Text
- View/download PDF
14. EMFusion: An unsupervised enhanced medical image fusion network
- Author
-
Han Xu and Jiayi Ma
- Subjects
Image fusion ,Fusion ,Computer science ,business.industry ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,020206 networking & telecommunications ,Pattern recognition ,02 engineering and technology ,Constraint (information theory) ,Modal ,Hardware and Architecture ,Distortion ,Signal Processing ,0202 electrical engineering, electronic engineering, information engineering ,Chrominance ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,Encoder ,Software ,Information Systems - Abstract
Existing image fusion methods always use the same representations for different modal medical images. Otherwise, they solve the fusion problem by subjectively defining characteristics to be preserved. However, it leads to the distortion of unique information and restricts the fusion performance. To address the limitations, this paper proposes an unsupervised enhanced medical image fusion network. We perform both surface-level and deep-level constraints for enhanced information preservation. The surface-level constraint is based on the saliency and abundance measurement to preserve the subjectively defined and intuitive characteristics. In the deep-level constraint, the unique information is objectively defined based on the unique channels of a pre-trained encoder. Moreover, in our method, the chrominance information of fusion results is also enhanced. It is because we use the high-quality details in structural images (e.g., MRI) to alleviate the mosaic in functional images (e.g., PET, SPECT). Both qualitative and quantitative experiments demonstrate the superiority of our method over the state-of-the-art fusion methods.
- Published
- 2021
- Full Text
- View/download PDF
15. Image Enhancement for Remote Photoplethysmography in a Low-Light Environment
- Author
-
Lin Xi, Jianhua Wang, Xingming Wu, Weihai Chen, and Changchen Zhao
- Subjects
FOS: Computer and information sciences ,Motion compensation ,Video capture ,Remote patient monitoring ,Computer science ,business.industry ,Computer Vision and Pattern Recognition (cs.CV) ,0206 medical engineering ,Image and Video Processing (eess.IV) ,Computer Science - Computer Vision and Pattern Recognition ,02 engineering and technology ,Electrical Engineering and Systems Science - Image and Video Processing ,020601 biomedical engineering ,Signal ,Facial recognition system ,Independent component analysis ,0202 electrical engineering, electronic engineering, information engineering ,Chrominance ,FOS: Electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,Face detection ,business - Abstract
With the improvement of sensor technology and significant algorithmic advances, the accuracy of remote heart rate monitoring technology has been significantly improved. Despite of the significant algorithmic advances, the performance of rPPG algorithm can degrade in the long-term, high-intensity continuous work occurred in evenings or insufficient light environments. One of the main challenges is that the lost facial details and low contrast cause the failure of detection and tracking. Also, insufficient lighting in video capturing hurts the quality of physiological signal. In this paper, we collect a large-scale dataset that was designed for remote heart rate estimation recorded with various illumination variations to evaluate the performance of the rPPG algorithm (Green, ICA, and POS). We also propose a low-light enhancement solution (technical solution) for remote heart rate estimation under the low-light condition. Using collected dataset, we found 1) face detection algorithm cannot detect faces in video captured in low light conditions; 2) A decrease in the amplitude of the pulsatile signal will lead to the noise signal to be in the dominant position; and 3) the chrominance-based method suffers from the limitation in the assumption about skin-tone will not hold, and Green and ICA method receive less influence than POS in dark illuminance environment. The proposed solution for rPPG process is effective to detect and improve the signal-to-noise ratio and precision of the pulsatile signal., Comment: Accepted by FG2020
- Published
- 2023
- Full Text
- View/download PDF
16. Measure a Subjective Video Quality Via a Neural Network
- Author
-
El Khattabi, Hasnaa, Tamtaoui, Ahmed, Aboutajdine, Driss, Cherifi, Hocine, editor, Zain, Jasni Mohamad, editor, and El-Qawasmeh, Eyas, editor
- Published
- 2011
- Full Text
- View/download PDF
17. An efficient image processing methodology based on fuzzy decision for dental shade matching.
- Author
-
Lin, Ting-Lan, Chuang, Chien-Hui, Chen, Shih-Lun, Lin, Nung-Hsiang, Miaou, Shaou-Gang, Lin, Szu-Yin, Chen, Chiung-An, Liu, Hui-Wen, Villaverde, Jocelyn Flores, and Hsieh, Wen-Hsiang
- Subjects
- *
IMAGE processing , *AGE , *ANALYSIS of colors , *DENTAL care - Abstract
For dentists, it is very important to determine the color of the denture. Shade selection in dental practice is an important and difficult task. In the dental shade matching process, the shade selection will be affected by the observer's physiological conditions such as age, mood, fatigue, and so on. These will make a difference on the judgement between the matching shade and the actual teeth color. In the past, dentists use shade tabs as a reference basis to match the teeth in the intra-oral environment. In this paper, an efficient color analysis methodology based on image processing and fuzzy decision techniques is proposed for dental shade matching. Since the color information is a very important index for the shade matching, the proposed methodology used the chrominance values Cb and Cr to increase the accuracy of color analysis. In order to improve the performance of the proposed methodology, three formulas, such as PSNR value of Cb, PSNR value of Cr, and S-CIELAB value, were selected by a fuzzy decision model. As shown in the results, the proposed efficient methodology based on fuzzy decision techniques improved at least 1.92 % in average accuracy and 0.59 in average score from the PSNR (Cb) and PSNR (Cr) in this work. In addition, the average values of the accuracies and scores in this work are 92.31% and 98.74, respectively, which are much better than the previous studies. To summarized, this work is the first study that applied fuzzy decision with the PSNR (Cb), PSNR (Cr) and S-CLIELAB information for dental shade matching. The results showed that the proposed methodology performs better than the previous work and other methods. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
18. Pixel-Wise Based Digital Watermarking in YC b C r Color Space
- Author
-
Surachat, Komwit, Amornraksa, Thumrongrat, Hutchison, David, editor, Kanade, Takeo, editor, Kittler, Josef, editor, Kleinberg, Jon M., editor, Mattern, Friedemann, editor, Mitchell, John C., editor, Naor, Moni, editor, Nierstrasz, Oscar, editor, Pandu Rangan, C., editor, Steffen, Bernhard, editor, Sudan, Madhu, editor, Terzopoulos, Demetri, editor, Tygar, Doug, editor, Vardi, Moshe Y., editor, Weikum, Gerhard, editor, Muneesawang, Paisarn, editor, Wu, Feng, editor, Kumazawa, Itsuo, editor, Roeksabutr, Athikom, editor, Liao, Mark, editor, and Tang, Xiaoou, editor
- Published
- 2009
- Full Text
- View/download PDF
19. Optimal keyframe selection-based lossless video-watermarking technique using IGSA in LWT domain for copyright protection
- Author
-
Himanshu Mittal, Roop Singh, and Raju Pal
- Subjects
Lossless compression ,Channel (digital image) ,Computer science ,business.industry ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Wavelet transform ,Watermark ,General Medicine ,Video processing ,Robustness (computer science) ,Chrominance ,Computer vision ,Artificial intelligence ,business ,Digital watermarking - Abstract
Video piracy is a challenging issue in the modern world. Approximately $$90\%$$ 90 % of newly released films were illegally distributed around the world via the Internet. To overcome this issue, video watermarking is an effective process that integrates a logo in video frames as a watermark. Therefore, this paper presents an efficient lossless video-watermarking scheme based on optimal keyframe selection using an intelligent gravitational search algorithm in linear wavelet transform. This technique obtains color motion and motionless frames from the cover video by the histogram difference method. One-level linear wavelet transform is performed on the chrominance channel of motion frames and a low-frequency sub-band LL opts for watermark embedding. The performance of the proposed technique has been evaluated against 12 video processing attacks in terms of imperceptibility and robustness. Experiments demonstrate that the proposed technique outperforms five state-of-the-art schemes on the considered attacks.
- Published
- 2021
- Full Text
- View/download PDF
20. Deep Demosaicking with Luminance and Chrominance Estimations
- Author
-
Taishi Iriyama, Hisashi Aomori, Tsuyoshi Otake, and Masatoshi Sato
- Subjects
Demosaicing ,business.industry ,Computer science ,Chrominance ,Computer vision ,Artificial intelligence ,business ,Luminance - Published
- 2021
- Full Text
- View/download PDF
21. A novel deep learning method for detection and classification of plant diseases
- Author
-
Waleed Albattah, Momina Masood, Ali Javed, Saleh Albahli, and Marriam Nawaz
- Subjects
Computer science ,business.industry ,Process (engineering) ,Deep learning ,media_common.quotation_subject ,Computational intelligence ,General Medicine ,Machine learning ,computer.software_genre ,Plant disease ,Identification (information) ,Chrominance ,Quality (business) ,Noise (video) ,Artificial intelligence ,business ,computer ,media_common - Abstract
The agricultural production rate plays a pivotal role in the economic development of a country. However, plant diseases are the most significant impediment to the production and quality of food. The identification of plant diseases at an early stage is crucial for global health and wellbeing. The traditional diagnosis process involves visual assessment of an individual plant by a pathologist through on-site visits. However, manual examination for crop diseases is restricted because of less accuracy and the small accessibility of human resources. To tackle such issues, there is a demand to design automated approaches capable of efficiently detecting and categorizing numerous plant diseases. Precise identification and classification of plant diseases is a tedious job due because of the occurrence of low-intensity information in the image background and foreground, the huge color resemblance in the healthy and diseased plant areas, the occurrence of noise in the samples, and changes in the position, chrominance, structure, and size of plant leaves. To tackle the above-mentioned problems, we have introduced a robust plant disease classification system by introducing a Custom CenterNet framework with DenseNet-77 as a base network. The presented method follows three steps. In the first step, annotations are developed to get the region of interest. Secondly, an improved CenterNet is introduced in which DenseNet-77 is proposed for deep keypoints extraction. Finally, the one-stage detector CenterNet is used to detect and categorize several plant diseases. To conduct the performance analysis, we have used the PlantVillage Kaggle database, which is the standard dataset for plant diseases and challenges in terms of intensity variations, color changes, and differences found in the shapes and sizes of leaves. Both the qualitative and quantitative analysis confirms that the presented method is more proficient and reliable to identify and classify plant diseases than other latest approaches.
- Published
- 2021
- Full Text
- View/download PDF
22. An Efficient Algorithm for Luminance Optimization in Chroma Downsampling
- Author
-
Bang-Hao Liu, Kun-Hu Jiang, and Ting-Lan Lin
- Subjects
Upsampling ,Pixel ,Computational complexity theory ,Image quality ,Media Technology ,Chrominance ,RGB color model ,Electrical and Electronic Engineering ,Chroma subsampling ,Algorithm ,Luminance ,Mathematics - Abstract
The classical chroma subsampling involves downsampling chrominance components (U and V) while maintaining the luminance component (Y). Recently, a study has attempted to change Y for the chroma-subsampling process, and improved results in the reconstructed image quality are obtained. However, computational complexity remains an issue because the study examines all the candidates in the determined range for minimal pixel distortion for the modified Y; the complexity can be high if the range is large. In this study, we reduced the candidate number to seven at most, which is mathematically optimized. Boundary points for in-the-range red, green, and blue (RGB) values are first decided, followed by the determination of the intervals that concatenate the entire curve (at most seven intervals). Each interval belongs to one of the seven sub-cases of different linear combinations of individual curves. The optimal solution (modified Y) for each sub-case is derived, which is in highly efficient form. Compared with the existing method, the proposed fast method ensures that the image quality is preserved, while the number of search candidates is reduced by 61.30%–69.19% on average, and the computational time of the process is reduced by 31.39%–58.97%, thereby demonstrating an efficient performance.
- Published
- 2021
- Full Text
- View/download PDF
23. Gradient-Based Intraprediction Fusion for Video Coding
- Author
-
Mohsen Abdoli, Thomas Guionnet, Gosala Kulupana, Saverio Blasi, and Mickaël Raulet
- Subjects
Reference software ,Computer science ,business.industry ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Single-mode optical fiber ,Luminance ,Computer Science Applications ,Hardware and Architecture ,Signal Processing ,Media Technology ,Chrominance ,Codec ,business ,Encoder ,Software ,Decoding methods ,Computer hardware ,Coding (social sciences) - Abstract
Video coding is more important than ever due to emergence of larger resolutions, new formats requiring higher bandwidths, prevalence of new video services, etc. Consequently, new codecs are designed to minimize transmission rates of such contents without sacrificing quality. This article proposes an intraprediction solution for future video codecs that save the bit rates of intramode signaling for luminance and chrominance blocks. In doing so, modes are implicitly derived at the decoder side through a gradient-based analysis. In contrast to state-of-the-art intraprediction that uses a single mode for a luma block, the proposed method offers a combination of multiple derived modes. Furthermore, the chrominance components Cb/Cr are allowed to have their separate modes, instead of sharing one mode. The implementation of the proposed method in the Versatile Video Coding reference software (VTM) provides 0.72% Bjontegaard-delta rate saving, with run-time complexity of 110% and 104% at the encoder and decoder sides, respectively.
- Published
- 2021
- Full Text
- View/download PDF
24. Frame-Wise CNN-Based Filtering for Intra-Frame Quality Enhancement of HEVC Videos
- Author
-
Hongyue Huang, Ionut Schiopu, Adrian Munteanu, Electronics and Informatics, and Faculty of Engineering
- Subjects
Artificial neural network ,Computer science ,business.industry ,Deep learning ,Frame (networking) ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,Deep-learning ,Convolutional neural network ,Intra-frame ,video coding ,quality enhancement ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,Chrominance ,020201 artificial intelligence & image processing ,Segmentation ,Computer vision ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,Block (data storage) - Abstract
The paper proposes a novel frame-wise filtering method based on Convolutional Neural Networks (CNNs) for enhancing the quality of HEVC decoded videos. A novel deep neural network architecture is proposed for post-filtering the entire intra-coded videos. A novel scheme utilizing frame-size patches is employed for training the network. The proposed method filters the luminance channel separately from the pair of chrominance channels. A novel patch generation paradigm is proposed where, for each color channel, the corresponding mode map is generated based on the HEVC intra-prediction mode index and block segmentation. The proposed CNN-based filtering method is an alternative to the traditional HEVC built-in in-loop filtering module for intra-coded frames. Experimental results on standard test sequences show that the proposed method outperforms the HEVC standard with average BD-rate savings of 11.1% and an average BD-PSNR improvement of 0.602 dB. The average relative improvement in $\Delta $ PSNR is around 105% at $QP = 42$ and around 85% at $QP = 32$ compared with state-of-the-art machine-learning-based methods.
- Published
- 2021
- Full Text
- View/download PDF
25. Projection Invariant Feature and Visual Saliency-Based Stereoscopic Omnidirectional Image Quality Assessment
- Author
-
Yang Zhou, Yo-Sung Ho, Yun Zhang, Na Li, Xu Wang, and Xuemei Zhou
- Subjects
Monocular ,Image quality ,business.industry ,Computer science ,Distortion (optics) ,Scale-invariant feature transform ,020206 networking & telecommunications ,Stereoscopy ,02 engineering and technology ,law.invention ,law ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,Chrominance ,Computer vision ,Artificial intelligence ,Electrical and Electronic Engineering ,Invariant (mathematics) ,Projection (set theory) ,business - Abstract
In this article, we propose a quality assessment model-based on the projection invariant feature and the visual saliency for Stereoscopic Omnidirectional Images (SOIs). Firstly, the projection invariant monocular and binocular features of SOI are derived from the Scale-Invariant Feature Transform (SIFT) points to tackle the inconsistency between the stretched projection formats and the viewports. Secondly, the visual saliency model, which combines the chrominance and contrast perceptual factors, is used to facilitate the prediction accuracy. Thirdly, according to the characteristics of the panoramic image, we generate the weight map and utilize it as a location prior, which can be adapted to different projection formats. Finally, the proposed SOI quality assessment model fuses the projection invariant features, visual saliency, and location prior. Experimental results on both the NingBo University SOI Database (NBU-SOID) and Stereoscopic OmnidirectionaL Image quality assessment Database (SOLID) demonstrate the proposed metric on equi-rectangular projection format outperforms the state-of-the-art schemes, the pearson linear correlation coefficient and spearman rank order correlation coefficient performance are 0.933 and 0.933 on SOLID, and 0.907 and 0.910 on NBU-SOID, respectively. Meanwhile, the proposed algorithm is extended to another five representative projection formats and achieves superior performance.
- Published
- 2021
- Full Text
- View/download PDF
26. PulseGAN: Learning to Generate Realistic Pulse Waveforms in Remote Photoplethysmography
- Author
-
Rencheng Song, Juan Cheng, Huan Chen, Chang Li, Yu Liu, and Xun Chen
- Subjects
Computer science ,Noise reduction ,0206 medical engineering ,Feature extraction ,02 engineering and technology ,Signal ,Standard deviation ,Health Information Management ,Heart Rate ,Photoplethysmogram ,FOS: Electrical engineering, electronic engineering, information engineering ,0202 electrical engineering, electronic engineering, information engineering ,Humans ,Electrical and Electronic Engineering ,Photoplethysmography ,Signal processing ,business.industry ,Image and Video Processing (eess.IV) ,Signal Processing, Computer-Assisted ,Pattern recognition ,Electrical Engineering and Systems Science - Image and Video Processing ,020601 biomedical engineering ,Computer Science Applications ,Face ,Chrominance ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,Algorithms ,Biotechnology ,Interbeat interval - Abstract
Remote photoplethysmography (rPPG) is a non-contact technique for measuring cardiac signals from facial videos. High-quality rPPG pulse signals are urgently demanded in many fields, such as health monitoring and emotion recognition. However, most of the existing rPPG methods can only be used to get average heart rate (HR) values due to the limitation of inaccurate pulse signals. In this paper, a new framework based on generative adversarial network, called PulseGAN, is introduced to generate realistic rPPG pulse signals through denoising the chrominance signals. Considering that the cardiac signal is quasi-periodic and has apparent time-frequency characteristics, the error losses defined in time and spectrum domains are both employed with the adversarial loss to enforce the model generating accurate pulse waveforms as its reference. The proposed framework is tested on the public UBFC-RPPG database in both within-database and cross-database configurations. The results show that the PulseGAN framework can effectively improve the waveform quality, thereby enhancing the accuracy of HR, the heart rate variability (HRV) and the interbeat interval (IBI). The proposed method achieves the best performance compared to the denoising autoencoder (DAE) and CHROM, with the mean absolute error of AVNN (the average of all normal-to-normal intervals) improving 20.85% and 41.19%, and the mean absolute error of SDNN (the standard deviation of all NN intervals) improving 20.28% and 37.53%, respectively, in the cross-database test. This framework can be easily extended to other existing deep learning based rPPG methods, which is expected to expand the application scope of rPPG techniques., 10 pages, 11 figures
- Published
- 2021
- Full Text
- View/download PDF
27. Deep Template-Based Watermarking
- Author
-
Zehua Ma, Qidong Huang, Han Fang, Nenghai Yu, Weiming Zhang, Jie Zhang, and Dongdong Chen
- Subjects
Artificial neural network ,business.industry ,Computer science ,Feature extraction ,Pattern recognition ,02 engineering and technology ,Robustness (computer science) ,Human visual system model ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,Chrominance ,Embedding ,020201 artificial intelligence & image processing ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,Feature learning ,Digital watermarking - Abstract
Traditional watermarking algorithms have been extensively studied. As an important type of watermarking schemes, template-based approaches maintain a very high embedding rate. In such scheme, the message is often represented by some dedicatedly designed templates, and then the message embedding process is carried out by additive operation with the templates and the host image. To resist potential distortions, these templates often need to contain some special statistical features so that they can be successfully recovered at the extracting side. But in existing methods, most of these features are handcrafted and too simple, thus making them not robust enough to resist serious distortions unless very strong and obvious templates are used. Inspired by the powerful feature learning capacity of deep neural network, we propose the first deep template-based watermarking algorithm in this paper. Specifically, at the embedding side, we first design two new templates for message embedding and locating, which is achieved by leveraging the special properties of human visual system, i.e. , insensitivity to specific chrominance components, the proximity principle and the oblique effect. At the extracting side, we propose a novel two-stage deep neural network, which consists of an auxiliary enhancing sub-network and a classification sub-network. Thanks to the power of deep neural networks, our method achieves both digital editing resilience and camera shooting resilience based on typical application scenarios. Through extensive experiments, we demonstrate that the proposed method can achieve much better robustness than existing methods while guaranteeing the original visual quality.
- Published
- 2021
- Full Text
- View/download PDF
28. An improved faster-RCNN model for handwritten character recognition
- Author
-
Ali Javed, Aun Irtaza, Marriam Nawaz, and Saleh Albahli
- Subjects
Multidisciplinary ,business.industry ,Computer science ,Deep learning ,010102 general mathematics ,Pattern recognition ,01 natural sciences ,Convolutional neural network ,Task (project management) ,Numeral system ,Region of interest ,Distortion ,Chrominance ,Artificial intelligence ,0101 mathematics ,business ,MNIST database - Abstract
Existing techniques for hand-written digit recognition (HDR) rely heavily on the hand-coded key points and requires prior knowledge. Training an efficient HDR network with these preconditions is a complicated task. Recently, work on HDR is mainly focused on deep learning (DL) approaches and has exhibited remarkable results. However, effective detection and classification of numerals is still a challenging task due to people’s varying writing styles and the presence of blurring, distortion, light and size variations in the input sample. To cope with these limitations, we present an effective and efficient HDR system, introducing a customized faster regional convolutional neural network (Faster-RCNN). This approach comprises three main steps. Initially, we develop annotations to obtain the region of interest. Then, an improved Faster-RCNN is employed in which DenseNet-41 is introduced to compute the deep features. Finally, the regressor and classification layer is used to localize and classify the digits into ten classes. The performance of the proposed method is analyzed on the standard MNIST database, which is diverse in terms of changes in lighting conditions, chrominance, shape and size of digits, and the occurrence of blurring and noise effects, etc. Additionally, we have also evaluated our technique over a cross-dataset scenario to prove its efficacy. Experimental evaluations demonstrate that the approach is more competent and able to accurately detect and classify numerals than other recent methods.
- Published
- 2021
- Full Text
- View/download PDF
29. CWEA: A Digital Video Copyright Protection Scheme.
- Author
-
Ali, Jabir and Ghrera, Satya Prakash
- Subjects
DIGITAL video ,COPYRIGHT ,INTELLECTUAL property ,ALGORITHMS ,LUMINANCE (Video) - Abstract
Digital Watermarking is an important method of protecting the intellectual property and copyright of the digital media. In this paper, a new digital video watermarking algorithm Color Watermark Embedding Algorithm (CWEA) is proposed. CWEA has two important parts. First, YCbCr color format is used to insert the variable size watermark. Second, embedding of detail coefficients of LUMINANCE (Y-luminance) of the watermark into the detail coefficients of CHROMINANCE (Cb and Cr- chrominance) of identical frames (I-Frames) of digital video. Data is inserted into the detail coefficients in an adaptive manner based on the energy of high frequency. A number of tests have been executed for many video-frame manipulations and attacks. All these tests are also performed on CWEA and it provides good results. In this paper, non-blind and semi-blind watermarking systems are used where the non-blind watermarking mechanism has been proved to be robust, imperceptible and efficient to protect the copyright of H.264 and MPEG-4 coded video within the video retrieval system. [ABSTRACT FROM AUTHOR]
- Published
- 2018
30. Fast Fusion-Based Dehazing With Histogram Modification and Improved Atmospheric Illumination Prior
- Author
-
Jian Qiu, Xuegui Zhu, Hongjiang Zeng, Fushuo Huo, and Qifeng Liu
- Subjects
Channel (digital image) ,Pixel ,business.industry ,Computer science ,media_common.quotation_subject ,010401 analytical chemistry ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Color balance ,Color space ,01 natural sciences ,0104 chemical sciences ,Histogram ,Chrominance ,Contrast (vision) ,Computer vision ,Artificial intelligence ,Electrical and Electronic Engineering ,Visual artifact ,business ,Instrumentation ,media_common - Abstract
Haze can seriously affect the visible and visual quality of outdoor optical sensor systems, e.g., driving assistance, remote sensing, and video surveillance. Single image dehazing is an intractable problem due to its ill-posed nature. The main idea of the paper is combining multi-scale fusion strategy and prior knowledge, thereby presenting balanced image contrast enhancement and intrinsic color preservation, efficiently. The atmospheric illumination prior (AIP) has been proved that haze mainly degrades the contrast of the luminance channel rather than chrominance channels in YCrCb colorspace. To this end, we firstly identify and remove the color veil (unbalanced color channel) with the white balance algorithm, to reduce the influence of unbalanced color channels neglected by the AIP. Considering the new observation that hazy regions exhibit low contrast with high-intensity pixels, the dense and mild haze are enhanced by a set of histogram modification techniques, respectively. Then, with the derived inputs, multi-scale fusion based on Laplacian decomposition strategy is proposed to blend visual contrast only in the luminance channel. Without relying on complex enhancement algorithms and only dealing with one channel, the proposed method is attractive for real-time applications. Moreover, The proposed method can be directly applied to the video sequences frame by frame, alleviating visual artifacts. The simulation results show that our method is comparative to and even better than the more complex state-of-the-art techniques.
- Published
- 2021
- Full Text
- View/download PDF
31. An optimization-based approach to gamma correction parameter estimation for low-light image enhancement
- Author
-
Inho Jeong and Chul Lee
- Subjects
Logarithm ,Computer Networks and Communications ,Computer science ,Estimation theory ,Color image ,Astrophysics::High Energy Astrophysical Phenomena ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,020207 software engineering ,02 engineering and technology ,Luminance ,Hardware and Architecture ,Gamma correction ,Convex optimization ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,Chrominance ,Algorithm ,Software - Abstract
We propose an efficient low-light image enhancement algorithm based on an optimization-based approach for gamma correction parameter estimation. We first separate an input color image into the luminance and chrominance channels, and then normalize the luminance channel using the logarithmic function to make it consistent with the human perception. Then, we divide the luminance image into dark and bright regions, and estimate the optimal gamma correction parameter for each region independently. Specifically, based on the statistical properties of the input image, we formulate a convex optimization problem that maximizes the image contrast subject to the constraint on the gamma value. By efficiently solving the optimization problems using the convex optimization theories, we obtain the optimal gamma parameter for each region. Finally, we obtain an enhanced image by merging the independently enhanced dark and bright regions with the optimal gamma parameters. Experimental results on real-world images demonstrate that the proposed algorithm can provide higher enhancement performance than state-of-the-art algorithms in terms of both subjective and objective evaluations, while providing a substantial improvement in speed.
- Published
- 2021
- Full Text
- View/download PDF
32. Image Compression Based on Arithmetic Coding Algorithm
- Author
-
Ruaa Ibrahim Yousif and Nassir H. Salman
- Subjects
General Computer Science ,Image quality ,Computer science ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,020206 networking & telecommunications ,YCbCr ,Data_CODINGANDINFORMATIONTHEORY ,02 engineering and technology ,General Chemistry ,Lossy compression ,General Biochemistry, Genetics and Molecular Biology ,Arithmetic coding ,Compression ratio ,0202 electrical engineering, electronic engineering, information engineering ,Chrominance ,Discrete cosine transform ,020201 artificial intelligence & image processing ,Algorithm ,Image compression - Abstract
The past years have seen a rapid development in the area of image compression techniques, mainly due to the need of fast and efficient techniques for storage and transmission of data among individuals. Compression is the process of representing the data in a compact form rather than in its original or incompact form. In this paper, integer implementation of Arithmetic Coding (AC) and Discreet Cosine Transform (DCT) were applied to colored images. The DCT was applied using the YCbCr color model. The transformed image was then quantized with the standard quantization tables for luminance and chrominance. The quantized coefficients were scanned by zigzag scan and the output was encoded using AC. The results showed a decent compression ratio with high image quality.
- Published
- 2021
- Full Text
- View/download PDF
33. Determination of ‘Hass’ avocado ripeness during storage by a smartphone camera using artificial neural network and support vector regression
- Author
-
Byeong-Hyo Cho, Kento Koyama, and Shigenobu Koseki
- Subjects
Artificial neural network ,biology ,Computer science ,business.industry ,General Chemical Engineering ,010401 analytical chemistry ,Hass avocado ,Pattern recognition ,04 agricultural and veterinary sciences ,Color space ,biology.organism_classification ,Ripeness ,040401 food science ,01 natural sciences ,Industrial and Manufacturing Engineering ,0104 chemical sciences ,Support vector machine ,0404 agricultural biotechnology ,Chrominance ,RGB color model ,Hass ,Artificial intelligence ,Safety, Risk, Reliability and Quality ,business ,Food Science - Abstract
Avocado undergoes quality transformation during storage, which needs to be managed in order to prevent quantity losses. A machine vision system devised with a smartphone camera was used to capture ‘Hass’ avocado images. Color features in L*a*b* and YUV (YUV color space is defined in terms of one luminance (Y) and two chrominance components (U and Y)) were extracted from the RGB images. Artificial Neural Network (ANN) and Support Vector Regression (SVR) were used compared for firmness estimation using the L*a*b* and YUV color features. The results indicated the ANN model is more accurate and robust than the SVR model for estimating ‘Hass’ avocado firmness with R2, RMSE, and RPD of 0.94, 0.38, 4.03 respectively for the model testing data set. It was concluded that the machine vision system devised with a smartphone camera and ANN model could be a low-cost tool for the determination of ripeness of ‘Hass’ avocado during harvest, storage, and distribution.
- Published
- 2021
- Full Text
- View/download PDF
34. Luminance approximated vector quantization algorithm to retain better image quality of the decompressed image
- Author
-
Abul Hasnat, Dibyendu Barman, and Bandana Barman
- Subjects
Computer Networks and Communications ,Structural similarity ,Image quality ,Computer science ,Color image ,Quantization (signal processing) ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Vector quantization ,020207 software engineering ,Data_CODINGANDINFORMATIONTHEORY ,02 engineering and technology ,Luminance ,Color model ,Hardware and Architecture ,Computer Science::Computer Vision and Pattern Recognition ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,Chrominance ,Quantization (image processing) ,Algorithm ,Software ,Block (data storage) - Abstract
Some compressed images using Vector Quantization algorithm suffers from blocking artifacts which degrades the visual appeal of the image. Present study proposes a hybrid vector quantization method applicable on de-correlated color model. As luminance channel carries image information and loss of image information results in degradation of the visual appeal of an image, so aim of this study is focused on retaining more image information during compression process. For luminance channel compression, a new four level quantization based compression method is developed. Luminance channel is partitioned into smaller blocks. Then for each block, four level quantization is applied which local to the current block only. This results many level luminance value effectively for the whole image. It helps to retain better information. Chrominance channels are compressed using conventional Vector Quantization. This hybrid compression method improves visual quality of the decompressed image reasonably compared to VQ. The proposed method is applied on many standard images found in literature and images of UCIDv.2 color image database. Results are analyzed in terms of Peak Signal to Noise Ratio, Structure Similarity Index and space requirement reduction for compressed image using the method. Experimental results show that proposed method retains better quality of image in terms of PSNR and SSIM than Vector Quantization and Modified Vector Quantization. This method reduces storage space requirement for the compressed images in the range of 84% to 89%.
- Published
- 2021
- Full Text
- View/download PDF
35. YuvConv: Multi-Scale Non-Uniform Convolution Structure Based on YUV Color Model
- Author
-
Xixi Yuan, Zhanchuan Cai, and Youqing Xiao
- Subjects
business.industry ,Computer science ,Deep learning ,Feature extraction ,Pattern recognition ,Luminance ,Convolutional neural network ,Computer Science Applications ,Convolution ,Digital image ,Feature (computer vision) ,Signal Processing ,Media Technology ,Chrominance ,Artificial intelligence ,Electrical and Electronic Engineering ,business - Abstract
Since digital images are able to be encoded through the luminance-bandwidth-chrominance (YUV) mode, and the contribution of luminance information is greater than that of chrominance information for human visual perception, it can be inferred that the appropriate reduction of chrominance information in convolutional neural network does not disturb image object recognition. In this paper, we propose a new multi-scale non-uniform convolution called YuvConv, wherein the output feature map of the convolutional layer is regarded as an image. First, the output channels in the new convolution are divided into three kinds of components: Y, U, and V tensors. Then, the tensor Y is used to process luminance information, which is high-resolution and occupies more output channels. Next, the tensors U and V are low-resolution and use fewer channels to process chrominance information. Finally, the adjacent tensors (Y-U, Y-U-V, and U-V) are fused as the output of YuvConv. Experimental results indicate that the use of the YuvConv instead of the standard convolution can improve the performance of deep learning tasks, and it can also reduce memory consumption and computation cost.
- Published
- 2021
- Full Text
- View/download PDF
36. Joint Luminance and Chrominance Learning for Underwater Image Enhancement
- Author
-
Zhenhua Hao, Risheng Liu, Xinwei Xue, Yi Wang, and Long Ma
- Subjects
Computer science ,business.industry ,Applied Mathematics ,Color correction ,Deep learning ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Process (computing) ,020206 networking & telecommunications ,02 engineering and technology ,Iterative reconstruction ,Interference (wave propagation) ,Luminance ,Distortion ,Signal Processing ,0202 electrical engineering, electronic engineering, information engineering ,Chrominance ,Computer vision ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,ComputingMethodologies_COMPUTERGRAPHICS - Abstract
Recently, learning-based works have been widely-investigated to enhance underwater images. However, interactions between various degradation factors (e.g., color distortion and haze effects) inevitably cause negative interference during the inference phase. Thus, these works cannot fully remove degraded factors. To address this problem, we propose a novel Joint Luminance and Chrominance Learning Network (JLCL-Net). Concretely, we reformulate the task as luminance reconstruction (for haze removal), and chrominance correction (for color correction) sub-tasks by separating the luminance and chrominance (i.e., color appearance) of the underwater images. In this way, we successfully realize the disentanglement in degraded factors to avoid introducing interference. We specify the reconstruction by integrating the atmospheric scattering model, which endows the adaptive dehazing ability over different scenarios. The correction learns to compensate for color by a simple network to reverse the color attenuation process. To this end, we obtain our JLCL-Net. To better train it, we design a new multi-stage cross-space training strategy, which progressively updates the network parameters to enlarge the network potentiality. Extensive evaluations are presented to fully verify our superiority against other methods.
- Published
- 2021
- Full Text
- View/download PDF
37. Two-Branch Deep Neural Network for Underwater Image Enhancement in HSV Color Space
- Author
-
Runmin Cong, Wei Gao, Feng Shao, Junkang Hu, and Qiuping Jiang
- Subjects
Artificial neural network ,Computer science ,Underwater vision ,business.industry ,Applied Mathematics ,media_common.quotation_subject ,HSL and HSV ,Convolutional neural network ,RGB color space ,Signal Processing ,Chrominance ,Contrast (vision) ,Computer vision ,Artificial intelligence ,Electrical and Electronic Engineering ,Underwater ,business ,media_common - Abstract
Due to the influence of light absorption and scattering, underwater images usually suffer from quality deteriorations such as color cast and reduced contrast. The diverse quality degradations not only dissatisfy the user expectation but also lead to a significant performance drop in many underwater vision applications. This letter proposes a novel two-branch deep neural network for underwater image enhancement (UIE), which is capable of separately removing color cast and enhancing image contrast by fully leveraging useful properties of the HSV color space in disentangling chrominance and intensity. Specifically, the input underwater image is first converted into the HSV color space and disentangled into HS and V channels to serve as the input of the two branches, respectively. Then, the color cast removal branch enhances the H and S channels with a generative adversarial network architecture while the contrast enhancement branch enhances the V channel via a traditional convolutional neural network. The enhanced channels by the two branches are merged and converted back into RGB color space to obtain the final enhanced result. Experimental results demonstrate that, compared with state-of-the-art UIE methods, our method can produce much more visually pleasing enhanced results.
- Published
- 2021
- Full Text
- View/download PDF
38. Real-Time Deep Image Retouching Based on Learnt Semantics Dependent Global Transforms
- Author
-
Qifan Gao and Xiaolin Wu
- Subjects
Artificial neural network ,Pixel ,Computer science ,business.industry ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,Tone mapping ,Computer Graphics and Computer-Aided Design ,Convolutional neural network ,Robustness (computer science) ,0202 electrical engineering, electronic engineering, information engineering ,Chrominance ,020201 artificial intelligence & image processing ,Computer vision ,Affine transformation ,Artificial intelligence ,business ,Software ,Image restoration - Abstract
Although artists' actions in photo retouching appear to be highly nonlinear in nature and very difficult to characterize analytically, we find that the net effects of interactively editing a mundane image to a desired appearance can be modeled, in most cases, by a parametric monotonically non-decreasing global tone mapping function in the luminance axis and by a global affine transform in the chrominance plane that are weighted by saliency. This allows us to simplify the machine learning problem of mimicking artists in photo retouching to constructing a deep artful image transform (DAIT) using convolutional neural networks (CNN). The CNN design of DAIT aims to learn the image-dependent parameters of the luminance tone mapping function and the affine chrominance transform, rather than learning the end-to-end pixel level mapping as in the mainstream methods of image restoration and enhancement. The proposed DAIT approach reduces the computation complexity of the neural network by two orders of magnitude, which also, as a side benefit, improves the robustness and generalization capability at the inference stage. The high throughput and robustness of DAIT lend itself readily to real-time video enhancement as well after a simple temporal processing. Experiments and a Turing-type test are conducted to evaluate the proposed method and its competitors.
- Published
- 2021
- Full Text
- View/download PDF
39. Variational Single Image Dehazing for Enhanced Visualization
- Author
-
Tieyong Zeng, Faming Fang, Guixu Zhang, Tingting Wang, and Yang Wang
- Subjects
Haze ,Channel (digital image) ,Computer science ,business.industry ,02 engineering and technology ,Color space ,Luminance ,Computer Science Applications ,Visualization ,Signal Processing ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,Chrominance ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,Electrical and Electronic Engineering ,Focus (optics) ,business ,Image restoration - Abstract
In this paper, we investigate the challenging task of removing haze from a single natural image. The analysis on the haze formation model shows that the atmospheric veil has much less relevance to chrominance than luminance, which motivates us to neglect the haze in the chrominance channel and concentrate on the luminance channel in the dehazing process. Besides, the experimental study illustrates that the YUV color space is most suitable for image dehazing. Accordingly, a variational model is proposed in the Y channel of the YUV color space by combining the reformulation of the haze model and the two effective priors. As we mainly focus on the Y channel, most of the chrominance information of the image is preserved after dehazing. The numerical procedure based on the alternating direction method of multipliers (ADMM) scheme is presented to obtain the optimal solution. Extensive experimental results on real-world hazy images and synthetic dataset demonstrate clearly that our method can unveil the details and recover vivid color information, which is competitive among many existing dehazing algorithms. Further experiments show that our model also can be applied for image enhancement.
- Published
- 2020
- Full Text
- View/download PDF
40. Video coding and processing: A survey
- Author
-
Si Liu, Yonghao Wang, Hongguo Zhao, and Yunxia Liu
- Subjects
business.industry ,Computer science ,Cognitive Neuroscience ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Computer Science Applications ,Redundancy (information theory) ,Artificial Intelligence ,Video encoding ,Human visual system model ,Chrominance ,Redundancy (engineering) ,Computer vision ,Artificial intelligence ,business ,Coding (social sciences) - Abstract
Vision is the main way for people to perceive and recognize the world. In this paper, four categories of the redundant information of video encoding, spatial redundancy, time redundancy, visual redundancy and statistical redundancy, are first introduced. Spatial redundancy is redundant information existing in static frames, time redundancy is the luminance and chrominance correlation between adjacent frames in the video sequence, visual redundancy is information that the human visual system cannot perceive, statistical redundancy is the expression redundancy of the information source symbols. Then the performance assessment of video coding are introduced. And the future development of video coding and conclusion are finally discussed in this paper.
- Published
- 2020
- Full Text
- View/download PDF
41. Multi-scale retinex-based contrast enhancement method for preserving the naturalness of color image
- Author
-
Yang Chuanying, Ma Shaoying, and Shi Bao
- Subjects
Color constancy ,Computer science ,business.industry ,Color image ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Image processing ,02 engineering and technology ,01 natural sciences ,Atomic and Molecular Physics, and Optics ,010309 optics ,020210 optoelectronics & photonics ,Naturalness ,Gamma correction ,0103 physical sciences ,0202 electrical engineering, electronic engineering, information engineering ,Chrominance ,Entropy (information theory) ,Computer vision ,Artificial intelligence ,business ,Hue - Abstract
For images with insufficient visibility, image processing is required. Retinex theory is often implemented for image contrast enhancement for images with this characteristic. This paper proposes a multi-scale retinex-based contrast enhancement method using the illumination component; this method could preserve the naturalness of color images. In the proposed method, illumination was first modified using an illumination modification factor; second, the image was enhanced via adaptive gamma correction. Finally, through the combination of the illumination components of the input image and the adaptive gamma correction image, we ensured the visibility and the naturalness of the output image. To confirm the effectiveness of the proposed method, we compared it with existing contrast enhancement methods. For the experiments, we employed discrete entropy, lightness order error, and mean chrominance error to perform the numerical evaluation. The results indicated that our method was better than a majority of existing methods. Moreover, with regard to the visual evaluation, the naturalness of the image obtained via the proposed method was superior to that of images obtained using other methods.
- Published
- 2020
- Full Text
- View/download PDF
42. Person re-identification using prioritized chromatic texture (PCT) with deep learning
- Author
-
K. Jayapriya, N. Ani Brown Mary, and I. Jeena Jacob
- Subjects
Pixel ,Computer Networks and Communications ,business.industry ,Color vision ,Computer science ,Deep learning ,Feature extraction ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,020207 software engineering ,Pattern recognition ,02 engineering and technology ,HSL and HSV ,Convolutional neural network ,Hardware and Architecture ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,Chrominance ,Chromatic scale ,Artificial intelligence ,business ,Software - Abstract
Person re-identification (re-ID) helps to identify a person’s attention in different cameras. But this is not an easy task, due to distance, illumination and lack of dataset. Nowadays, this field attracts many researchers because of its varied applications. Here, the information of both local texture and global color representations are concatenated with an original raw image. This concatenated information is gathered by finding the maximum value of chrominance in terms of HSV, texture in terms of Scale Invariant Local Ternary Pattern (SILTP) for each pixel and original raw image. SILTP is well known for its illumination invariant texture description. Convolutional Neural Network (CNN) is used in the proposed work to extract the features from the concatenated information. The proposed Prioritized Chromatic Texture Image (PCTimg) is concatenated with original raw image and fed into CNN. Here, finally a six dimensionalfeature is fed into CNN to extract the deep features. Cross-view Quadratic Discriminant Analysis (XQDA) similarity metric algorithm is employed to re-identify a person.Multiscale Retinex algorithm is used for pre-processing the images.To address the challenges in terms of view point deflection, a sliding window is formed for describing local details of a person in the SILTP feature extraction phase. The HSV helps to incorporate the human color perception. The triplet loss function is used to learn the similarity and the dissimilarity of the training images. The performance analysis of the proposed work is improvedwhen compared to the existingworks.
- Published
- 2020
- Full Text
- View/download PDF
43. A Multi-Attribute Blind Quality Evaluator for Tone-Mapped Images
- Author
-
Peter Schelkens and Saeed Mahmoudpour
- Subjects
Brightness ,business.industry ,Computer science ,Dynamic range ,Image quality ,Feature extraction ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,Tone mapping ,Luminance ,Computer Science Applications ,Visualization ,High-dynamic-range imaging ,Signal Processing ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,Chrominance ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,Electrical and Electronic Engineering ,business ,High dynamic range - Abstract
High dynamic range (HDR) imaging enables capturing a wide range of luminance levels existing in real-world scenes. While HDR capturing devices become widespread in the market, the display technology is yet limited in representing full luminance ranges and standard low dynamic range (LDR) displays are currently more prevalent. To visualize the HDR content on traditional displays, tone mapping (TM) operators are introduced that convert HDR content into LDR. The dynamic range compression and different processing steps during TM can lead to loss of scene details, as well as luminance and chrominance changes. Such signal deviations will affect image naturalness and consequently disturb the visual quality of experience. Therefore, research into objective methods for quality evaluation of tone-mapped images has received attention in recent years. In this paper, we proposed a completely blind image quality evaluator for tone-mapped images based on a multi-attribute feature extraction scheme. Due to the diversity of TM distortions, various image characteristics are taken into account to develop an effective metric. The features are designed by considering spectral and spatial entropy, detection probability of visual information, image exposure, sharpness, and color properties. The quality-relevant features are then fed into a machine-learning regression framework to pool a quality score. The validation tests on two benchmark datasets reveal the superior performance of the proposed approach compared to the competing metrics.
- Published
- 2020
- Full Text
- View/download PDF
44. Multi-exposure image fusion based on tensor decomposition
- Author
-
Shengcong Wu, Ting Luo, Yang Song, and Haiyong Xu
- Subjects
Image fusion ,Computer Networks and Communications ,business.industry ,Computer science ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,020207 software engineering ,YCbCr ,Pattern recognition ,02 engineering and technology ,Luminance ,Higher-order singular value decomposition ,Hardware and Architecture ,Feature (computer vision) ,Computer Science::Computer Vision and Pattern Recognition ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,Chrominance ,RGB color model ,Fusion rules ,Artificial intelligence ,business ,Software ,ComputingMethodologies_COMPUTERGRAPHICS - Abstract
In this paper, a multi-exposure image fusion (MEF) method is proposed based on tensor decomposition and saliency model. The main innovation of the proposed method is to explore a tensor domain for MEF and define the fusion rules based on tensor feature of higher order singular value decomposition (HOSVD) and saliency. Specifically, RGB images are converted to YCbCr images to maintain the stability of color information. For luminance channels, luminance patches of luminance images are constructed 3-order sub-tensors, and HOSVD is used to extract features of sub-tensors. Then, the sum of absolute coefficients (SAC) of weight coefficients are defined. Meanwhile, considering the impact of saliency on visual perception, visual saliency maps (VSMs) is used to evaluate luminance patches quality and guide the fusion rules to define the rule of fusion. For chrominance channels, VSMs of the chrominance channels is used to define fused rule. The experimental results show that the fused image with more texture details and saturated color is successfully generated by proposed method.
- Published
- 2020
- Full Text
- View/download PDF
45. Cross-Component Prediction in HEVC
- Author
-
Mischa Siekmann, Marta Karczewicz, Wei Pu, Joel Sole, Detlev Marpe, Woo-Shik Kim, Jianle Chen, Tung Nguyen, Ali Khairat, and Publica
- Subjects
business.industry ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,YCbCr ,02 engineering and technology ,Coding tree unit ,Color depth ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,Chrominance ,RGB color model ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,Electrical and Electronic Engineering ,Multiview Video Coding ,business ,Context-adaptive binary arithmetic coding ,Mathematics ,Context-adaptive variable-length coding - Abstract
Video coding in the YCbCr color space has been widely used, since it is efficient for compression, but it can result in color distortion due to conversion error. Meanwhile, coding in the RGB color space maintains high color fidelity, having the drawback of a substantial bitrate increase with respect to YCbCr coding. Cross-component prediction (CCP) efficiently compresses video content by decorrelating color components while keeping high color fidelity. In this scheme, the chroma residual signal is predicted from the luma residual signal inside the coding loop. This paper gives a description of the CCP scheme from several points of view, from theoretical background to practical implementation. The proposed CCP scheme has been evaluated in standardization communities and adopted into H.265/High Efficiency Video Coding (HEVC) Range Extensions. The experimental results show significant coding performance improvements for both natural and screen content video, while the quality of all color components is maintained. The average coding gains for natural video are 17% and 5% bitrate reduction in the case of intra coding and 11% and 4% in the case of inter coding for RGB and YCbCr coding, respectively, while the average increment of encoding and decoding times in the HEVC reference software implementation are 10% and 4%, respectively.
- Published
- 2020
- Full Text
- View/download PDF
46. Selective Image Encryption Based on DCT, Hybrid Shift Coding and Randomly Generated Secret Key
- Author
-
Enas Kh. Hassan, Loay E. George, Faisel G. Mohammed, and Sajaa G. Mohammed
- Subjects
General Computer Science ,Computer science ,business.industry ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,020207 software engineering ,YCbCr ,Data_CODINGANDINFORMATIONTHEORY ,010103 numerical & computational mathematics ,02 engineering and technology ,General Chemistry ,Encryption ,01 natural sciences ,Peak signal-to-noise ratio ,General Biochemistry, Genetics and Molecular Biology ,Run-length encoding ,0202 electrical engineering, electronic engineering, information engineering ,Chrominance ,Discrete cosine transform ,RGB color model ,0101 mathematics ,business ,Algorithm ,Data compression - Abstract
Most of today’s techniques encrypt all of the image data, which consumes a tremendous amount of time and computational payload. This work introduces a selective image encryption technique that encrypts predetermined bulks of the original image data in order to reduce the encryption/decryption time and thecomputational complexity of processing the huge image data. This technique is applying a compression algorithm based on Discrete Cosine Transform (DCT). Two approaches are implemented based on color space conversion as a preprocessing for the compression phases YCbCr and RGB, where the resultant compressed sequence is selectively encrypted using randomly generated combined secret key.The results showed a significant reduction in image quality degradation when applying the system based on YCbCr over RGB, where the compression ratio was raised in some of the tested images to 50% for the same Peak Signal to Noise Ratio (PSNR). The usage of 1-D DCT reduced the transform time by 47:1 times comparedto the same transform using 2-D DCT. The values of the adaptive scalar quantization parameters were reduced to the half for the luminance (Y band) to preserve the visual quality, while the chrominance (Cb and Cr bands) were quantized by the predetermined quantization parameters. In the hybrid encoder horizontal zigzag,block scanning was applied to scan the image. The Detailed Coefficient (DC) coefficients are highly correlated in this arrangement- where DC are losslessly compressed by Differential Pulse Coding Modulation (DPCM) and theAccumulative Coefficients (AC) are compressed using Run Length Encoding (RLE). As a consequence, for the compression algorithm, the compression gain obtained was up to 95%. Three arrays are resulted from each band (DC coefficients, AC values, and AC runs), where the cipher is applied to some or all of those bulksselectively. This reduces the encryption decryption time significantly, where encrypting the DC coefficients provided the second best randomness and the least encryption/decryption time recorded (3 10-3 sec.) for the entire image. Although the compression algorithm consumes time but it is more efficient than the savedencryption time.
- Published
- 2020
- Full Text
- View/download PDF
47. A 3D Image Quality Assessment Method Based on Vector Information and SVD of Quaternion Matrix under Cloud Computing Environment
- Author
-
Lan Zhang, Kaixuan Lu, and Xingang Liu
- Subjects
Pixel ,Computer Networks and Communications ,Color image ,Computer science ,business.industry ,Image quality ,020208 electrical & electronic engineering ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,Computer Science Applications ,Hardware and Architecture ,Depth map ,Distortion ,Human visual system model ,0202 electrical engineering, electronic engineering, information engineering ,Chrominance ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,business ,Quaternion ,Software ,Information Systems - Abstract
With the increasing demands of end-users to the visual perception in three-dimension (3D) image, quality assessment for 3D imageis dominantly required as the feedback information for multimedia transmission systems. In this paper, a novel full-reference quality assessment method by considering the depth and integral color information of 3D image under cloud computing environment is proposed. Based on the property of the depth information in 3D image, the depth map is firstly separated into different planes according to the perception of human visual system (HVS). Then, after express the image pixels of every separated plane through quaternions, the structural and energy information are separated by quaternion singular value decomposition (QSVD). The distortion of structural and energy in every plane are calculated in various formulas respectively. The final result is calculated in terms of the global score, which synthesizes the structural and energy distortion scores in every individual depth plane. It should be pointed out that the chrominance information is employed in our mechanism to evaluate the color image quality because of its useful characteristic for 3D color image quality assessment, and its spatial correlation is used for calculating structural distortion through vector cross-product. Our experimental results confirm that the proposed method has achieves better performance under cloud computing environments compared with other existing 3D image quality assessment methods.
- Published
- 2020
- Full Text
- View/download PDF
48. ROBUST COLOR IMAGE WATERMARKING SCHEMES IN THE WAVELET DOMAIN
- Author
-
Dejey and R.S. Rajesh
- Subjects
DWT ,Luminance ,Chrominance ,Attacks ,Collusion ,Telecommunication ,TK5101-6720 ,Computer applications to medicine. Medical informatics ,R858-859.7 - Abstract
In this paper, two approaches for color image watermarking in the wavelet domain are proposed. The first approach utilizes only the chrominance content of the color image for watermarking after performing a single level DWT (Discrete Wavelet Transform) decomposition and hence named as DWTC whereas the second approach utilizes both the luminance and chrominance content for watermarking and hence termed as DWTLC. Watermark which is a logo is scrambled before embedding to increase its robustness. Watermarking is done on the approximation band coefficients chosen randomly by a function in L*a*b* space. Both the proposed approaches result in watermarked images of high quality. The robustness of both the approaches is verified by the conduct of various attacks on the watermarked images and is compared with an existing SCDFT based approach and a DWT based approach. Experimental results show that DWTC is robust to non geometric attacks like filtering, blurring, sharpening, histogram equalization, JPEG compression and to geometric attacks like additive noise and scaling. DWTLC watermarking also shows robustness to all the above attacks. Further, both the approaches resist collusion attack even with a minimum number of colluders and with increasing number of colluders there is much distortion in the visual quality of the colluded images.
- Published
- 2010
49. A robust GAN-generated face detection method based on dual-color spaces and an improved Xception
- Author
-
Xin Liu, Beijing Chen, Yun-Qing Shi, Guoying Zhao, and Yuhui Zheng
- Subjects
Computer science ,business.industry ,generative adversarial network ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Xception ,YCbCr ,HSL and HSV ,Color space ,color space ,generated face ,Gamma correction ,Robustness (computer science) ,Media Technology ,Chrominance ,RGB color model ,Computer vision ,Artificial intelligence ,Electrical and Electronic Engineering ,Face detection ,business - Abstract
In recent years, generative adversarial networks (GANs) have been widely used to generate realistic fake face images, which can easily deceive human beings. To detect these images, some methods have been proposed. However, their detection performance will be degraded greatly when the testing samples are post-processed. In this paper, some experimental studies on detecting post-processed GAN-generated face images find that (a) both the luminance component and chrominance components play an important role, and (b) the RGB and YCbCr color spaces achieve better performance than the HSV and Lab color spaces. Therefore, to enhance the robustness, both the luminance component and chrominance components of dual-color spaces (RGB and YCbCr) are considered to utilize color information effectively. In addition, the convolutional block attention module and multilayer feature aggregation module are introduced into the Xception model to enhance its feature representation power and aggregate multilayer features, respectively. Finally, a robust dual-stream network is designed by integrating dual-color spaces RGB and YCbCr and using an improved Xception model. Experimental results demonstrate that our method outperforms some existing methods, especially in its robustness against different types of post-processing operations, such as JPEG compression, Gaussian blurring, gamma correction, and median filtering.
- Published
- 2022
50. Inter layer up-sampling filtering scheme applied in SVC
- Author
-
WANG Zhang, LIU Jian, and YAN Guo-ping
- Subjects
scalable video coding ,up-sampling filtering ,luminance ,chrominance ,human vision system ,Telecommunication ,TK5101-6720 - Abstract
An efficient inter layer up-sampling filtering algorithm in scalable video coding (SVC) was proposed. Based on different component-sensitive of the human vision system,scheme assigns up-sampling filters depending on the sensi-tivity priority of the image component in order to increase the coding efficiency was proposed. Particularly,a 6-tap filter is assigned for luminance component and a bilinear filter is for chrominance components. The experiment results show that our proposed scheme maintains the performances of coded bit-rate and PSNR value without any noticeable loss,and provides significant reduction in computational complexity. As a result,it can be applied in the inter layer spatial inter-polation of SVC.
- Published
- 2008
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.