94 results on '"Duan, Junbo"'
Search Results
2. Comparative study of non-convex penalties and related algorithms in compressed sensing
- Author
-
Xu, Fanding, Duan, Junbo, and Liu, Wenyu
- Published
- 2023
- Full Text
- View/download PDF
3. LoRA-TV: read depth profile-based clustering of tumor cells in single-cell sequencing.
- Author
-
Duan, Junbo, Zhao, Xinrui, and Wu, Xiaoming
- Subjects
- *
DEPTH profiling , *ROBUST optimization - Abstract
Single-cell sequencing has revolutionized our ability to dissect the heterogeneity within tumor populations. In this study, we present LoRA-TV (Low Rank Approximation with Total Variation), a novel method for clustering tumor cells based on the read depth profiles derived from single-cell sequencing data. Traditional analysis pipelines process read depth profiles of each cell individually. By aggregating shared genomic signatures distributed among individual cells using low-rank optimization and robust smoothing, the proposed method enhances clustering performance. Results from analyses of both simulated and real data demonstrate its effectiveness compared with state-of-the-art alternatives, as supported by improvements in the adjusted Rand index and computational efficiency. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. On the core segmentation algorithms of copy number variation detection tools.
- Author
-
Zhang, Yibo, Liu, Wenyu, and Duan, Junbo
- Subjects
DNA copy number variations ,SHOTGUN sequencing ,NUCLEOTIDE sequencing ,MARKOV processes ,DISTRIBUTION (Probability theory) ,SHOTGUNS - Abstract
Shotgun sequencing is a high-throughput method used to detect copy number variants (CNVs). Although there are numerous CNV detection tools based on shotgun sequencing, their quality varies significantly, leading to performance discrepancies. Therefore, we conducted a comprehensive analysis of next-generation sequencing-based CNV detection tools over the past decade. Our findings revealed that the majority of mainstream tools employ similar detection rationale: calculates the so-called read depth signal from aligned sequencing reads and then segments the signal by utilizing either circular binary segmentation (CBS) or hidden Markov model (HMM). Hence, we compared the performance of those two core segmentation algorithms in CNV detection, considering varying sequencing depths, segment lengths and complex types of CNVs. To ensure a fair comparison, we designed a parametrical model using mainstream statistical distributions, which allows for pre-excluding bias correction such as guanine-cytosine (GC) content during the preprocessing step. The results indicate the following key points: (1) Under ideal conditions, CBS demonstrates high precision, while HMM exhibits a high recall rate. (2) For practical conditions, HMM is advantageous at lower sequencing depths, while CBS is more competitive in detecting small variant segments compared to HMM. (3) In case involving complex CNVs resembling real sequencing, HMM demonstrates more robustness compared with CBS. (4) When facing large-scale sequencing data, HMM costs less time compared with the CBS, while their memory usage is approximately equal. This can provide an important guidance and reference for researchers to develop new tools for CNV detection. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. Effects of Smoothed BRDF on Multi-angle Albedo Retrievals
- Author
-
Shuai Yanmin, Yang Jian, Duan Junbo, Tuerhanjiang Latipa, Wang Chongyang, and Ma Yu
- Subjects
Environmental sciences ,GE1-350 - Abstract
BRDF as the intrinsic feature of surface targets, is an important parameter required by albedo inversion from multi-angle observations, especially for satellite data suit with less directional measurements. Several studies have shown up to introduce BRDF priori knowledge into albedo retrievals at different scale by spatial or temporal smoothing. Thus, it is necessary to further understand what’s the influence induced by BRDF smoothing on albedo retrieval. This work investigated effects of smoothed BRDF on albedo magnitude through case studies over North America region using operational MCD43A&C BRDF products respectively smoothed in spatial and temporal scales. Our results show that BRDF of seasonal DBF samples smoothed from daily to monthly can lead to apparent relative difference to smoothed values of 10.97%, 9.42%, 8.24% and detectable absolute differences of 0.0172, 0.0095 and 0.0035 on related albedo respectively at Near Infrared, Short Wave and Visible broadband. The spatial smoothing of BRDF from 500m to 5600m results in relative differences to smoothed values of 17.38%, 14.38%, 27.23% and absolute differences of 0.0250, 0.0139, 0.0052 for the inversed albedo at above three broadbands.
- Published
- 2022
- Full Text
- View/download PDF
6. Comparative study of whole exome sequencing-based copy number variation detection tools
- Author
-
Zhao, Lanling, Liu, Han, Yuan, Xiguo, Gao, Kun, and Duan, Junbo
- Published
- 2020
- Full Text
- View/download PDF
7. Trends of non-melanoma skin cancer incidence in Hong Kong and projection up to 2030 based on changing demographics.
- Author
-
Xu, Qingqiang, Wang, Xiaoyan, Bai, Yan, Zheng, Yan, Duan, Junbo, Du, Jianqiang, and Wu, Xiaoming
- Subjects
SKIN cancer ,COHORT analysis ,AGE groups ,BAYESIAN analysis ,POPULATION aging ,CONFIDENCE intervals - Abstract
To assess the trends in non-melanoma skin cancer (NMSC) incidence in Hong Kong from 1990 to 2019 and the associations of age, calendar period, and birth cohort, to make projections to 2030, and to examine the drivers of NMSC incidence. We assessed the age, calendar period, and birth cohort effects of NMSC incidence in Hong Kong between 1990 and 2019 using an age-period-cohort model. Using Bayesian age-period-cohort analysis with integrated nested Laplace approximations, we projected the incidence of NMSC in Hong Kong to 2030. From 1990 to 2019, the age-standardized incidence rate of NMSC increased from 6.7 per 100,000 population to 8.6 per 100,000 population in men and from 5.4 per 100,000 to 5.9 per 100,000 population in women, among the 19,568 patients in the study (9812 male patients [50.14%]). The annual net drift was 2.00% (95% confidence interval [CI]: 1.50–2.50%) for men and 1.53% (95% CI: 0.95–2.11%) for women. Local drifts increased for both sexes above the 35–39-year age group. The period and cohort risk of developing NMSC tended to rise but slowed gradually in the most recent period and post-1975 birth cohort. From 2019 to 2030, it is projected that the number of newly diagnosed NMSC cases in Hong Kong will increase from 564 to 829 in men and from 517 to 863 in women. Population aging, population growth, and epidemiologic changes contributed to the increase in incident NMSCs, with population aging being the most significant contributor. The slowing of the period and cohort effects suggests that the rising incidence of NMSC is partly attributable to increased awareness and diagnosis. The increasing prevalence of NMSC among the elderly and an aging population will significantly impact the clinical workload associated with NMSC for the foreseeable future. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
8. Generalized LASSO with under-determined regularization matrices
- Author
-
Duan, Junbo, Soussen, Charles, Brie, David, Idier, Jérôme, Wan, Mingxi, and Wang, Yu-Ping
- Published
- 2016
- Full Text
- View/download PDF
9. HOPS: A Fast Algorithm for Segmenting Piecewise Polynomials of Arbitrary Orders
- Author
-
Qing Wang, Duan Junbo, and Yu-Ping Wang
- Subjects
General Computer Science ,Piecewise polynomials ,Computer science ,segmentation ,General Engineering ,Fast algorithm ,TK1-9971 ,ComputingMethodologies_SYMBOLICANDALGEBRAICMANIPULATION ,curve fitting ,Piecewise ,Computer Science::Programming Languages ,sparse modeling ,General Materials Science ,breakpoint detection ,Electrical engineering. Electronics. Nuclear engineering ,Algorithm - Abstract
The segmentation of piecewise polynomial signals arises in a variety of scientific and engineering fields. When a signal is modeled as a piecewise polynomial, the key then becomes the detection of breakpoints followed by curve fitting and parameter estimation. This paper proposes HOPS, a fast High-Order Polynomial Segmenter, which is based on $\ell _{0}$ -penalized least-square regression. While the least-squares regression ensures fitting fidelity, the $\ell _{0}$ penalty takes the number of breakpoints into account. We show that dynamic programming can be applied to find the optimal solution to this problem and that a pruning strategy and matrix factorization can be utilized to accelerate the execution speed. Finally, we provide some illustrative examples, and compare the proposed method with state-of-the-art alternatives.
- Published
- 2021
10. Laser-induced photoexcited audible sound effect based on reticular 2-bromo-2-methylpropionic acid modified Fe3O4 nanoparticle aggregates.
- Author
-
Li, Yan, Zhu, Hongrui, Duan, Junbo, Wu, Youshen, and Wu, Daocheng
- Published
- 2022
- Full Text
- View/download PDF
11. Subtyping glioblastoma by combining miRNA and mRNA expression data using compressed sensing-based approach
- Author
-
Tang, Wenlong, Duan, Junbo, Zhang, Ji-Gang, and Wang, Yu-Ping
- Published
- 2013
- Full Text
- View/download PDF
12. The Description and Application of BRDF Based on Shape Vectors for Typical Landcovers.
- Author
-
Yang, Jian, Huang, Jiapeng, Fan, Hongdong, Duan, Junbo, and Ma, Xianwei
- Abstract
As the inherent attribute of land cover, anisotropy leads to the heterogeneity of directional reflection; meanwhile, it creates the opportunity for retrieving characteristics of land surface based on multi-angle observations. BRDF (Bidirectional Reflectance Distribution Function) is the theoretical expression of anisotropy and describes the reflectance in terms of incident-view geometry. Prior BRDF knowledge is used to achieve the multi-angle retrieval for earth observation systems with a narrow FOV (Field of View). Shape indicators are a feasible way to capture the characteristics of BRDF or to build an a priori database of BRDF. However, existing shape indicators based on the ratio of reflectance or the weight of scattering effects are too rough to describe the BRDF's shape. Thus, it is necessary to propose new shape vectors to satisfy the demand. We selected six typical land covers from MODIS-MCD12 on the homogeneous underlayers as the study sites in North America. The daily BRDF is retrieved by MODIS-BRDF parameters and the RossThick-LiSparseR model. When the SZA (Solar Zenith Angle) is set at 45°, seven directions (−70°, −45°, −20°, 0°, 20°, 45°, and 70°) including edge spot, zenith spot, hot spot and approximate dark spot of the BRDF principal plane were selected to construct two vectors by the change rate of reflectance and angle formulation: Partial Anisotropic Vector (PAV) and Angular Effect Vector (AEV). Then, we assessed the effectiveness of PAV and AEV compared with ANIX (Anisotropic Index), ANIF (Anisotropic Factor) and AFX (Anisotropic Flat Index) by two typical BRDF shapes. The representativeness of PAV and AEV for the original BRDF was also assessed by cosine similarity and error transfer function. Lastly, the application of hot spot components in AEV for land cover classification, the monitoring of land cover in mining areas and the adjustment effect by NDVI (Normalized Difference Vegetation Index) were investigated. The results show that (1) the shape vectors have good representativeness compared with original BRDF. The representativeness of PAV assessed by cosine similarity is 0.980, 0.979 and 0.969, and the representativeness of AEV assessed by error transfer function is 0.987, 0.991 and 0.994 in the three MODIS broadbands of Near Infrared (NIR, 0.7–5.0 µm), Short Wave (SW, 0.3–5.0 µm) and Visible (VIS, 0.3–0.7 µm). (2) Some components of shape vectors have high correlation with AFX. The correlation coefficient between hot spot components in AEV and AFX is 0.936, 0.945 and 0.863, respectively, in NIR, SW and VIS bands. (3) The shape vectors show potentiality for land cover classification and the monitoring of land cover in mining areas. The correlation coefficients of hot spot components in AEV for MODIS-pixels with the same types (0.557, 0.561, 0.527) are significantly higher than MODIS-pixels with various types (0.069, 0.055, 0.051) in NIR, SW and VIS bands. The coefficients of variation for hot spot components are significantly higher after land reclamation (0.0071, 0.0099) than before land reclamation (0.0020, 0.0028). (4) The correlation between NDVI and the BRDF shapes is poor in three MODIS broad bands. The correlation coefficients between NDVI and the BRDF shapes in three temporal scales of annual, seasonal and monthly phases are only 0.134, 0.063 and 0.038 (NIR), 0.199, 0.185 and 0.165 (SW), and 0.323, 0.320 and 0.337 (VIS), on average. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
13. Sparse representation based biomarker selection for schizophrenia with integrated analysis of fMRI and SNPs
- Author
-
Cao, Hongbao, Duan, Junbo, Lin, Dongdong, Shugart, Yin Yao, Calhoun, Vince, and Wang, Yu-Ping
- Published
- 2014
- Full Text
- View/download PDF
14. Accurate detection of atrial fibrillation events with R-R intervals from ECG signals.
- Author
-
Duan, Junbo, Wang, Qing, Zhang, Bo, Liu, Chen, Li, Chenrui, and Wang, Lei
- Subjects
- *
ATRIAL fibrillation , *ATRIAL arrhythmias , *ELECTROCARDIOGRAPHY , *SUPPORT vector machines , *STATISTICS , *VENTRICULAR arrhythmia - Abstract
Atrial fibrillation (AF) is a typical category of arrhythmia. Clinical diagnosis of AF is based on the detection of abnormal R-R intervals (RRIs) with an electrocardiogram (ECG). Previous studies considered this detection problem as a classification problem and focused on extracting a number of features. In this study we demonstrate that instead of using any specific numerical characteristic as the input feature, the probability density of RRIs from ECG conserves comprehensive statistical information; hence, is a natural and efficient input feature for AF detection. Incorporated with a support vector machine as the classifier, results on the MIT-BIH database indicates that the proposed method is a simple and accurate approach for AF detection in terms of accuracy, sensitivity, and specificity. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
15. Impact of BRDF Spatiotemporal Smoothing on Land Surface Albedo Estimation.
- Author
-
Yang, Jian, Shuai, Yanmin, Duan, Junbo, Xie, Donghui, Zhang, Qingling, and Zhao, Ruishan
- Subjects
ALBEDO ,SURFACE of the earth ,SOLAR radiation ,TERRESTRIAL radiation - Abstract
Surface albedo, as a key parameter determining the partition of solar radiation at the Earth's surface, has been developed into a satellite-based product from various Earth observation systems to serve numerous global or regional applications. Studies point out that apparent uncertainty can be introduced into albedo retrieval without consideration of surface anisotropy, which is a challenge to albedo estimation especially from observations with fewer angular samplings. Researchers have begun to introduce smoothed anisotropy prior knowledge into albedo estimation to improve the inversion efficiency, or for the scenario of observations with signal or poor angular sampling. Thus, it is necessary to further understand the potential influence of smoothed anisotropy features adopted in albedo estimation. We investigated the albedo variation induced by BRDF smoothing at both temporal and spatial scales over six typical landscapes in North America using MODIS standard anisotropy products with high quality BRDF inversed from multi-angle observations in 500 m and 5.6 km spatial resolutions. Components of selected typical landscapes were assessed with the confidence of the MCD12 land cover product and 30 m CDL (cropland data layer) classification maps followed by an evaluation of spatial heterogeneity in 30 m scale through the semi-variogram model. High quality BRDF of MODIS standard anisotropy products were smoothed in multi-temporal scales of 8 days, 16 days, and 32 days, and in multi-spatial scales from 500 m to 5.6 km. The induced relative and absolute albedo differences were estimated using the RossThick-LiSparseR model and BRDFs smoothed before and after spatiotemporal smoothing. Our results show that albedo estimated using BRDFs smoothed temporally from daily to monthly over each scenario exhibits relative differences of 11.3%, 12.5%, and 27.2% and detectable absolute differences of 0.025, 0.012, and 0.013, respectively, in MODIS near-infrared (0.7–5.0 µm), short-wave (0.3–5.0 µm), and visible (0.3–0.7 µm) broad bands. When BRDFs of investigated landscapes are smoothed from 500 m to 5.6 km, variations of estimated albedo can achieve up to 36.5%, 37.1%, and 94.7% on relative difference and absolute difference of 0.037, 0.024, and 0.018, respectively, in near-infrared (0.7–5.0 µm), short wave (0.3–5.0 µm), and visible (0.3–0.7 µm) broad bands. In addition, albedo differences caused by temporal smoothing show apparent seasonal characteristic that the differences are significantly higher in spring and summer than those in autumn and winter, while albedo differences induced by spatial smoothing exhibit a noticeable relationship with sill values of a fitted semi-variogram marked by a correlation coefficient of 0.8876. Both relative and absolute albedo differences induced by BRDF smoothing are demonstrated to be captured, thus, it is necessary to avoid the smoothing process in quantitative remote sensing communities, especially when immediate anisotropy retrievals are available at the required spatiotemporal scale. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
16. ERINS: Novel Sequence Insertion Detection by Constructing an Extended Reference.
- Author
-
Yuan, Xiguo, Xu, Xiangyan, Zhao, Haiyong, and Duan, Junbo
- Abstract
Next generation sequencing technology has led to the development of methods for the detection of novel sequence insertions (nsINS). Multiple signatures from short reads are usually extracted to improve nsINS detection performance. However, characterization of nsINSs larger than the mean insert size is still challenging. This article presents a new method, ERINS, to detect nsINS contents and genotypes of full spectrum range size. It integrates the features of structural variations and mapping states of split reads to find nsINS breakpoints, and then adopts a left-most mapping strategy to infer nsINS content by iteratively extending the standard reference at each breakpoint. Finally, it realigns all reads to the extended reference and infers nsINS genotypes through statistical testing on read counts. We test and validate the performance of ERINS on simulation and real sequencing datasets. The simulation experimental results demonstrate that it outperforms several peer methods with respect to sensitivity and precision. The real data application indicates that ERINS obtains high consistent results with those of previously reported and detects nsINSs over 200 base pairs that many other methods fail. In conclusion, ERINS can be used as a supplement to existing tools and will become a routine approach for characterizing nsINSs. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
17. Fitting of Atomic Force Microscopy Force Curves with a Sparse Representation Model.
- Author
-
Wang, Qing, Hu, Nan, and Duan, Junbo
- Subjects
CURVE fitting ,PARAMETER estimation ,ATOMIC force microscopy ,PARAMETRIC modeling - Abstract
Atomic force microscopy (AFM) is a high-resolution scanning technology, and the measured data are a set of force curves, which can be fitted with a piecewise curve model and be analyzed further. Most methods usually follow a two-step strategy: first, the discontinuities (or breakpoints) are detected as the boundaries of two consecutive pieces; second, each piece separated by the discontinuities is fitted with a parametric model, such as the well-known worm-like chain (WLC) model. The disadvantage of this method is that the fitting (the second step) accuracy depends largely on the discontinuity detection (the first step) accuracy. In this study, a sparse representation model is proposed to jointly detect discontinuities and fit curves. The proposed model fits the curve with a linear combination of parametric functions, and the estimation of the parameters in the model can be formulated as an optimization problem with ℓ 0 -norm constraint. The performance of the proposed model is demonstrated by the fitting of AFM retraction force curves with the WLC model. Results shows that the proposed method can segment the force curve and estimate the parameter jointly with better accuracy, and hence, it is promising for automatic AFM force curve processing. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
18. CONDEL: Detecting Copy Number Variation and Genotyping Deletion Zygosity from Single Tumor Samples Using Sequence Data.
- Author
-
Yuan, Xiguo, Bai, Jun, Zhang, Junying, Yang, Liying, Duan, Junbo, Li, Yaoyao, and Gao, Meihong
- Abstract
Characterizing copy number variations (CNVs) from sequenced genomes is a both feasible and cost-effective way to search for driver genes in cancer diagnosis. A number of existing algorithms for CNV detection only explored part of the features underlying sequence data and copy number structures, resulting in limited performance. Here, we describe CONDEL, a method for detecting CNVs from single tumor samples using high-throughput sequence data. CONDEL utilizes a novel statistic in combination with a peel-off scheme to assess the statistical significance of genome bins, and adopts a Bayesian approach to infer copy number gains, losses, and deletion zygosity based on statistical mixture models. We compare CONDEL to six peer methods on a large number of simulation datasets, showing improved performance in terms of true positive and false positive rates, and further validate CONDEL on three real datasets derived from the 1000 Genomes Project and the EGA archive. CONDEL obtained higher consistent results in comparison with other three single sample-based methods, and exclusively identified a number of CNVs that were previously associated with cancers. We conclude that CONDEL is a powerful tool for detecting copy number variations on single tumor samples even if these are sequenced at low-coverage. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
19. Homotopy based algorithms for L0-regularized least-squares
- Author
-
Soussen, Charles, Idier, Jérome, Duan, Junbo, Brie, David, Centre de Recherche en Automatique de Nancy (CRAN), Centre National de la Recherche Scientifique (CNRS)-Université de Lorraine (UL), Institut de Recherche en Communications et en Cybernétique de Nantes (IRCCyN), Mines Nantes (Mines Nantes)-École Centrale de Nantes (ECN)-Ecole Polytechnique de l'Université de Nantes (EPUN), Université de Nantes (UN)-Université de Nantes (UN)-PRES Université Nantes Angers Le Mans (UNAM)-Centre National de la Recherche Scientifique (CNRS), Department of Biomedical Engineering, Xi'an Jiaotong University (Xjtu), CRAN - IRCCyN - Xi'an Jiaotong University, and Université de Lorraine (UL)-Centre National de la Recherche Scientifique (CNRS)
- Subjects
L0-regularized least-squares ,orthogonal least squares ,L0-regularization path ,stepwise algorithms ,L0-homotopy ,sparse signal estimation ,model order selection ,continuation ,[SPI.SIGNAL]Engineering Sciences [physics]/Signal and Image processing ,L1-homotopy - Abstract
International audience; Sparse signal restoration is usually formulated as the minimization of a quadratic cost function ||y-Ax||_2^2, where A is a dictionary and x is an unknown sparse vector. It is well-known that imposing an L0 constraint leads to an NP-hard minimization problem. The convex relaxation approach has received considerable attention, where the L0-norm is replaced by the L1-norm. Among the many efficient L1 solvers, the homotopy algorithm minimizes ||y-Ax||_2^2+lambda ||x||_1 with respect to x for a continuum of lambda's. It is inspired by the piecewise regularity of the L1-regularization path, also referred to as the homotopy path. In this paper, we address the minimization problem ||y-Ax||_2^2+lambda ||x||_0 for a continuum of lambda's and propose two heuristic search algorithms for L0-homotopy. Continuation Single Best Replacement is a forward-backward greedy strategy extending the Single Best Replacement algorithm, previously proposed for L0-minimization at a given lambda. The adaptive search of the lambda-values is inspired by L1-homotopy. L0 Regularization Path Descent is a more complex algorithm exploiting the structural properties of the L0-regularization path, which is piecewise constant with respect to lambda. Both algorithms are empirically evaluated for difficult inverse problems involving ill-conditioned dictionaries. Finally, we show that they can be easily coupled with usual methods of model order selection.
- Published
- 2015
20. Detection of False-Positive Deletions from the Database of Genomic Variants.
- Author
-
Duan, Junbo, Liu, Han, Zhao, Lanling, Yuan, Xiguo, Wang, Yu-Ping, and Wan, Mingxi
- Subjects
- *
CHROMOSOMES , *DATABASES , *DOCUMENTATION , *GENETIC polymorphisms , *PAPER chromatography , *SEQUENCE analysis - Abstract
Next generation sequencing is an emerging technology that has been widely used in the detection of genomic variants. However, since its depth of coverage, a main signature used for variant calling, is affected greatly by biases such as GC content and mappability, some callings are false positives. In this study, we utilized paired-end read mapping, another signature that is not affected by the aforementioned biases, to detect false-positive deletions in the database of genomic variants. We first identified 1923 suspicious variants that may be false positives and then conducted validation studies on each suspicious variant, which detected 583 false-positive deletions. Finally we analysed the distribution of these false positives by chromosome, sample, and size. Hopefully, incorrect documentation and annotations in downstream studies can be avoided by correcting these false positives in public repositories. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
21. The comparison and analysis of the surface drought monitoring model based on Xinjiang.
- Author
-
Shuai, Yanmin, Duan, Junbo, Yang, Jian, Shao, Congying, and Ma, Xianwei
- Published
- 2023
- Full Text
- View/download PDF
22. Restoration and separation of piecewise polynomial signals. Application to Atomic Force Microscopy
- Author
-
Duan, Junbo, Centre de Recherche en Automatique de Nancy (CRAN), Université Henri Poincaré - Nancy 1 (UHP)-Institut National Polytechnique de Lorraine (INPL)-Centre National de la Recherche Scientifique (CNRS), Université Henri Poincaré - Nancy 1, David Brie (david.brie@cran.uhp-nancy.fr), and Duan, Junbo
- Subjects
discontinuities ,Atomic Force Microscopy (AFM) ,physical models for force curves ,Orthogonal Least Squares (OLS) ,pseudo-norme ℓ0 ,microscopie de force atomique (AFM) ,Approximation parcimonieuse ,Sparse approximation ,discontinuités ,algorithmes de type ajout-retrait ,mélanges avec retard ,sparse source separation ,modèles physiques de courbes de force ,ℓ0 pseudo-norm ,forward-backward greedy algorithms ,séparation de sources parcimonieuses ,imagerie force-volume ,continuation ,force-volume imaging ,[SPI.SIGNAL]Engineering Sciences [physics]/Signal and Image processing ,[SPI.SIGNAL] Engineering Sciences [physics]/Signal and Image processing ,delayed mixtures - Abstract
This thesis handles several inverse problems occurring in sparse signal processing. The main contributions include the conception of algorithms dedicated to the restoration and the separation of sparse signals, and their application to force curve approximation in Atomic Force Microscopy (AFM), where the notion of sparsity is related to the number of discontinuity points in the signal (jumps, change of slope, change of curvature). In the signal processing viewpoint, we propose sub-optimal algorithms dedicated to the sparse signal approximation problem based on the ℓ0 pseudo-norm : the Single Best Replacement algorithm (SBR) is an iterative "forward-backward" algorithm inspired from existing Bernoulli-Gaussian signal restoration algorithms. The Continuation Single Best Replacement algorithm (CSBR) is an extension providing approximations at various sparsity levels. We also address the problem of sparse source separation from delayed mixtures. The proposed algorithm is based on the prior application of CSBR on every mixture followed by a matching procedure which attributes a label for each peak occurring in each mixture. Atomic Force Microscopy (AFM) is a recent technology enabling to measure interaction forces between nano-objects. The force-curve analysis relies on piecewise parametric models. We address the detection of the regions of interest (the pieces) where each model holds and the subsequent estimation of physical parameters (elasticity, adhesion forces, topography, etc.) in each region by least-squares optimization. We finally propose an alternative approach in which a force curve is modeled as a mixture of delayed sparse sources. The research of the source signals and the delays from a force-volume image is done based on a large number of mixtures since there are as many mixtures as the number of image pixels., Cette thèse s'inscrit dans le domaine des problèmes inverses en traitement du signal. Elle est consacrée à la conception d'algorithmes de restauration et de séparation de signaux parcimonieux et à leur application à l'approximation de courbes de forces en microscopie de force atomique (AFM), où la notion de parcimonie est liée au nombre de points de discontinuité dans le signal (sauts, changements de pente, changements de courbure). Du point de vue méthodologique, des algorithmes sous-optimaux sont proposés pour le problème de l'approximation parcimonieuse basée sur la pseudo-norme ℓ0 : l'algorithme Single Best Replacement (SBR) est un algorithme itératif de type « ajout-retrait » inspiré d'algorithmes existants pour la restauration de signaux Bernoulli-Gaussiens. L'algorithme Continuation Single Best Replacement (CSBR) est un algorithme permettant de fournir des approximations à des degrés de parcimonie variables. Nous proposons aussi un algorithme de séparation de sources parcimonieuses à partir de mélanges avec retards, basé sur l'application préalable de l'algorithme CSBR sur chacun des mélanges, puis sur une procédure d'appariement des pics présents dans les différents mélanges. La microscopie de force atomique est une technologie récente permettant de mesurer des forces d'interaction entre nano-objets. L'analyse de courbes de forces repose sur des modèles paramétriques par morceaux. Nous proposons un algorithme permettant de détecter les régions d'intérêt (les morceaux) où chaque modèle s'applique puis d'estimer par moindres carrés les paramètres physiques (élasticité, force d'adhésion, topographie, etc.) dans chaque région. Nous proposons finalement une autre approche qui modélise une courbe de force comme un mélange de signaux sources parcimonieux retardées. La recherche des signaux sources dans une image force-volume s'effectue à partir d'un grand nombre de mélanges car il y autant de mélanges que de pixels dans l'image.
- Published
- 2010
23. Homotopy based algorithms for $\ell_0$-regularized least-squares
- Author
-
Soussen, Charles, Idier, J��r��me, Duan, Junbo, and Brie, David
- Subjects
FOS: Computer and information sciences ,Statistics::Machine Learning ,Computer Science - Learning ,FOS: Mathematics ,Computer Science - Numerical Analysis ,Numerical Analysis (math.NA) ,Machine Learning (cs.LG) - Abstract
Sparse signal restoration is usually formulated as the minimization of a quadratic cost function $\|y-Ax\|_2^2$, where A is a dictionary and x is an unknown sparse vector. It is well-known that imposing an $\ell_0$ constraint leads to an NP-hard minimization problem. The convex relaxation approach has received considerable attention, where the $\ell_0$-norm is replaced by the $\ell_1$-norm. Among the many efficient $\ell_1$ solvers, the homotopy algorithm minimizes $\|y-Ax\|_2^2+\lambda\|x\|_1$ with respect to x for a continuum of $\lambda$'s. It is inspired by the piecewise regularity of the $\ell_1$-regularization path, also referred to as the homotopy path. In this paper, we address the minimization problem $\|y-Ax\|_2^2+\lambda\|x\|_0$ for a continuum of $\lambda$'s and propose two heuristic search algorithms for $\ell_0$-homotopy. Continuation Single Best Replacement is a forward-backward greedy strategy extending the Single Best Replacement algorithm, previously proposed for $\ell_0$-minimization at a given $\lambda$. The adaptive search of the $\lambda$-values is inspired by $\ell_1$-homotopy. $\ell_0$ Regularization Path Descent is a more complex algorithm exploiting the structural properties of the $\ell_0$-regularization path, which is piecewise constant with respect to $\lambda$. Both algorithms are empirically evaluated for difficult inverse problems involving ill-conditioned dictionaries. Finally, we show that they can be easily coupled with usual methods of model order selection., Comment: 38 pages
- Published
- 2014
24. Restauration et séparation de signaux polynomiaux par morceaux. Application à la microscopie de force atomique
- Author
-
Duan, Junbo, Centre de Recherche en Automatique de Nancy (CRAN), Université Henri Poincaré - Nancy 1 (UHP)-Institut National Polytechnique de Lorraine (INPL)-Centre National de la Recherche Scientifique (CNRS), Université Henri Poincaré - Nancy 1, and David Brie (david.brie@cran.uhp-nancy.fr)
- Subjects
discontinuities ,Atomic Force Microscopy (AFM) ,physical models for force curves ,Orthogonal Least Squares (OLS) ,pseudo-norme ℓ0 ,microscopie de force atomique (AFM) ,Approximation parcimonieuse ,Sparse approximation ,discontinuités ,algorithmes de type ajout-retrait ,mélanges avec retard ,sparse source separation ,modèles physiques de courbes de force ,ℓ0 pseudo-norm ,forward-backward greedy algorithms ,séparation de sources parcimonieuses ,imagerie force-volume ,continuation ,force-volume imaging ,[SPI.SIGNAL]Engineering Sciences [physics]/Signal and Image processing ,delayed mixtures - Abstract
This thesis handles several inverse problems occurring in sparse signal processing. The main contributions include the conception of algorithms dedicated to the restoration and the separation of sparse signals, and their application to force curve approximation in Atomic Force Microscopy (AFM), where the notion of sparsity is related to the number of discontinuity points in the signal (jumps, change of slope, change of curvature). In the signal processing viewpoint, we propose sub-optimal algorithms dedicated to the sparse signal approximation problem based on the ℓ0 pseudo-norm : the Single Best Replacement algorithm (SBR) is an iterative "forward-backward" algorithm inspired from existing Bernoulli-Gaussian signal restoration algorithms. The Continuation Single Best Replacement algorithm (CSBR) is an extension providing approximations at various sparsity levels. We also address the problem of sparse source separation from delayed mixtures. The proposed algorithm is based on the prior application of CSBR on every mixture followed by a matching procedure which attributes a label for each peak occurring in each mixture. Atomic Force Microscopy (AFM) is a recent technology enabling to measure interaction forces between nano-objects. The force-curve analysis relies on piecewise parametric models. We address the detection of the regions of interest (the pieces) where each model holds and the subsequent estimation of physical parameters (elasticity, adhesion forces, topography, etc.) in each region by least-squares optimization. We finally propose an alternative approach in which a force curve is modeled as a mixture of delayed sparse sources. The research of the source signals and the delays from a force-volume image is done based on a large number of mixtures since there are as many mixtures as the number of image pixels.; Cette thèse s'inscrit dans le domaine des problèmes inverses en traitement du signal. Elle est consacrée à la conception d'algorithmes de restauration et de séparation de signaux parcimonieux et à leur application à l'approximation de courbes de forces en microscopie de force atomique (AFM), où la notion de parcimonie est liée au nombre de points de discontinuité dans le signal (sauts, changements de pente, changements de courbure). Du point de vue méthodologique, des algorithmes sous-optimaux sont proposés pour le problème de l'approximation parcimonieuse basée sur la pseudo-norme ℓ0 : l'algorithme Single Best Replacement (SBR) est un algorithme itératif de type « ajout-retrait » inspiré d'algorithmes existants pour la restauration de signaux Bernoulli-Gaussiens. L'algorithme Continuation Single Best Replacement (CSBR) est un algorithme permettant de fournir des approximations à des degrés de parcimonie variables. Nous proposons aussi un algorithme de séparation de sources parcimonieuses à partir de mélanges avec retards, basé sur l'application préalable de l'algorithme CSBR sur chacun des mélanges, puis sur une procédure d'appariement des pics présents dans les différents mélanges. La microscopie de force atomique est une technologie récente permettant de mesurer des forces d'interaction entre nano-objets. L'analyse de courbes de forces repose sur des modèles paramétriques par morceaux. Nous proposons un algorithme permettant de détecter les régions d'intérêt (les morceaux) où chaque modèle s'applique puis d'estimer par moindres carrés les paramètres physiques (élasticité, force d'adhésion, topographie, etc.) dans chaque région. Nous proposons finalement une autre approche qui modélise une courbe de force comme un mélange de signaux sources parcimonieux retardées. La recherche des signaux sources dans une image force-volume s'effectue à partir d'un grand nombre de mélanges car il y autant de mélanges que de pixels dans l'image.
- Published
- 2010
25. Force curve segmentation by piecewise polynomial approximation: mathematical formulation and complete structure of the algorithm
- Author
-
Soussen, Charles, Duan, Junbo, Brie, David, Polyakov, Pavel, Francius, Gregory, Duval, Jérôme, Centre de Recherche en Automatique de Nancy (CRAN), Université Henri Poincaré - Nancy 1 (UHP)-Institut National Polytechnique de Lorraine (INPL)-Centre National de la Recherche Scientifique (CNRS), Laboratoire de Chimie Physique et Microbiologie pour l'Environnement (LCPME), Université de Lorraine (UL)-Centre National de la Recherche Scientifique (CNRS)-Institut de Chimie du CNRS (INC), Laboratoire Environnement et Minéralurgie (LEM), Institut National Polytechnique de Lorraine (INPL)-Centre National de la Recherche Scientifique (CNRS), and Soussen, Charles
- Subjects
[SPI.AUTO] Engineering Sciences [physics]/Automatic ,[SPI.AUTO]Engineering Sciences [physics]/Automatic - Abstract
In this technical report, we give a detailed formulation of the force curve segmentation problem based on the fitting of the curve by a piecewise polynomial. The segmentation algorithm is a discrete search in which the unknowns are the positions Dj, j=1,...,k of the discontinuity points. We first formulate the problem as the minimization of a least-square cost function with respect to D=[D1,...,Dk]\in R^k. Then, we describe the structure of the proposed optimization algorithm including the update of the list of discontinuity positions D when their number k increases.
- Published
- 2010
26. A continuation approach to estimate a solution path of mixed L2-L0 minimization problems
- Author
-
Duan, Junbo, Soussen, Charles, Brie, David, Idier, Jérôme, Centre de Recherche en Automatique de Nancy (CRAN), Université Henri Poincaré - Nancy 1 (UHP)-Institut National Polytechnique de Lorraine (INPL)-Centre National de la Recherche Scientifique (CNRS), Institut de Recherche en Communications et en Cybernétique de Nantes (IRCCyN), Mines Nantes (Mines Nantes)-École Centrale de Nantes (ECN)-Ecole Polytechnique de l'Université de Nantes (EPUN), and Université de Nantes (UN)-Université de Nantes (UN)-PRES Université Nantes Angers Le Mans (UNAM)-Centre National de la Recherche Scientifique (CNRS)
- Subjects
[SPI.AUTO]Engineering Sciences [physics]/Automatic - Abstract
International audience; The approximation of a signal using a limited number of dictionary elements is stated as an L0-constrained or an L0-penalized least-square problem. We first give the working assumptions and then propose the heuristic Single Best Replacement (SBR) algorithm for the penalized problem. It is inspired by the Single Most Likely Replacement (SMLR) algorithm, initially proposed in the context of Bernoulli-Gaussian deconvolution. Then, we extend the SBR algorithm to a continuation version estimating a whole solution path, i.e., a series of solutions depending on the level of sparsity. The continuation algorithm, up to a slight adaptation, also provides an estimate of a solution path of the L0-constrained problem. The effectiveness of this approach is illustrated on a sparse signal deconvolution problem.
- Published
- 2009
27. On the properties of the solution path of the constrained and penalized L2-L0 problems
- Author
-
Duan, Junbo, Soussen, Charles, Brie, David, Idier, Jérôme, Centre de Recherche en Automatique de Nancy (CRAN), Université Henri Poincaré - Nancy 1 (UHP)-Institut National Polytechnique de Lorraine (INPL)-Centre National de la Recherche Scientifique (CNRS), Institut de Recherche en Communications et en Cybernétique de Nantes (IRCCyN), Mines Nantes (Mines Nantes)-École Centrale de Nantes (ECN)-Ecole Polytechnique de l'Université de Nantes (EPUN), and Université de Nantes (UN)-Université de Nantes (UN)-PRES Université Nantes Angers Le Mans (UNAM)-Centre National de la Recherche Scientifique (CNRS)
- Subjects
Sparse approximation ,solution path ,L0-penalized least-square problem ,L0-norm ,L0-constrained least-square problem ,[SPI.AUTO]Engineering Sciences [physics]/Automatic - Abstract
12 pages; Technical report on the properties of the L0-constrained least-square minimization problem and the L0-penalized least-square minimization problem: domain of optimization, notion of solution path, properties of the "penalized" solution path...
- Published
- 2009
28. Sparse representation of a force spectrum and a force-volume image
- Author
-
Duan, Junbo, Soussen, Charles, Brie, David, Idier, Jérôme, Gaboriaud, Fabien, Centre de Recherche en Automatique de Nancy (CRAN), Université Henri Poincaré - Nancy 1 (UHP)-Institut National Polytechnique de Lorraine (INPL)-Centre National de la Recherche Scientifique (CNRS), Institut de Recherche en Communications et en Cybernétique de Nantes (IRCCyN), Mines Nantes (Mines Nantes)-École Centrale de Nantes (ECN)-Ecole Polytechnique de l'Université de Nantes (EPUN), Université de Nantes (UN)-Université de Nantes (UN)-PRES Université Nantes Angers Le Mans (UNAM)-Centre National de la Recherche Scientifique (CNRS), Laboratoire de Chimie Physique et Microbiologie pour l'Environnement (LCPME), and Université de Lorraine (UL)-Centre National de la Recherche Scientifique (CNRS)-Institut de Chimie du CNRS (INC)
- Subjects
[SPI.AUTO]Engineering Sciences [physics]/Automatic - Abstract
We present a new method for analyzing AFM data, relying on the concept of sparse representation. A signal is called sparse if it can be described by a very limited set of parameters. For AFM spectra, the sparse description is based on the discontinuities (jumps and derivatives) embedded in the data. The communication aims at presenting a very recent signal processing method to analyze sparse signals (denoising and extraction of patterns). The method is introduced and some results on real data are displayed, showing that patterns can be extracted in a very accurate fashion.
- Published
- 2008
29. The Next Generation Sequencing and Applications in Clinical Research.
- Author
-
Duan, Junbo, Fu, Xiaoying, Zhang, Jigang, Wang, Yu-Ping, and Deng, Hong-Wen
- Published
- 2016
- Full Text
- View/download PDF
30. Pulse-Inversion Subharmonic Ultrafast Active Cavitation Imaging in Tissue Using Fast Eigenspace-Based Adaptive Beamforming and Cavitation Deconvolution.
- Author
-
Bai, Chen, Xu, Shanshan, Duan, Junbo, Jing, Bowen, Yang, Miao, and Wan, Mingxi
- Subjects
BEAMFORMING ,SIGNAL resolution ,SIGNAL processing ,ANALYSIS of covariance ,CHEMICAL decomposition ,SUBHARMONIC functions ,COMPUTATIONAL complexity - Abstract
Pulse-inversion subharmonic (PISH) imaging can display information relating to pure cavitation bubbles while excluding that of tissue. Although plane-wave-based ultrafast active cavitation imaging (UACI) can monitor the transient activities of cavitation bubbles, its resolution and cavitation-to-tissue ratio (CTR) are barely satisfactory but can be significantly improved by introducing eigenspace-based (ESB) adaptive beamforming. PISH and UACI are a natural combination for imaging of pure cavitation activity in tissue; however, it raises two problems: 1) the ESB beamforming is hard to implement in real time due to the enormous amount of computation associated with the covariance matrix inversion and eigendecomposition and 2) the narrowband characteristic of the subharmonic filter will incur a drastic degradation in resolution. Thus, in order to jointly address these two problems, we propose a new PISH–UACI method using novel fast ESB (F-ESB) beamforming and cavitation deconvolution for nonlinear signals. This method greatly reduces the computational complexity by using F-ESB beamforming through dimensionality reduction based on principal component analysis, while maintaining the high quality of ESB beamforming. The degraded resolution is recovered using cavitation deconvolution through a modified convolution model and compressive deconvolution. Both simulations and in vitro experiments were performed to verify the effectiveness of the proposed method. Compared with the ESB-based PISH–UACI, the entire computation of our proposed approach was reduced by 99%, while the axial resolution gain and CTR were increased by 3 times and 2 dB, respectively, confirming that satisfactory performance can be obtained for monitoring pure cavitation bubbles in tissue erosion. [ABSTRACT FROM PUBLISHER]
- Published
- 2017
- Full Text
- View/download PDF
31. Increasing Axial Resolution of Ultrasonic Imaging With a Joint Sparse Representation Model.
- Author
-
Duan, Junbo, Zhong, Hui, Jing, Bowen, Zhang, Siyuan, and Wan, Mingxi
- Subjects
- *
ULTRASONIC imaging , *ACOUSTIC pulses , *ACOUSTIC transducers , *BANDWIDTHS , *DECONVOLUTION (Mathematics) - Abstract
The axial resolution of ultrasonic imaging is confined by the temporal width of acoustic pulse generated by the transducer, which has a limited bandwidth. Deconvolution can eliminate this effect and, therefore, improve the resolution. However, most ultrasonic imaging methods perform deconvolution scan line by scan line, and therefore the information embedded within the neighbor scan lines is unexplored, especially for those materials with layered structures such as blood vessels. In this paper, a joint sparse representation model is proposed to increase the axial resolution of ultrasonic imaging. The proposed model combines the sparse deconvolution along the axial direction with a sparsity-favoring constraint along the lateral direction. Since the constraint explores the information embedded within neighbor scan lines by connecting nearby pixels in the ultrasound image, the axial resolution of the image improves after deconvolution. The results on simulated data showed that the proposed method can increase resolution and discover layered structure. Moreover, the results on real data showed that the proposed method can measure carotid intima-media thickness automatically with good quality ( $0.56\pm 0.03$ versus $0.60\pm 0.06$ mm manually). [ABSTRACT FROM PUBLISHER]
- Published
- 2016
- Full Text
- View/download PDF
32. An optimal method to segment piecewise poisson distributed signals with application to sequencing data.
- Author
-
Duan, Junbo, Soussen, Charles, Brie, David, Idier, Jerome, Wang, Yu-Ping, and Wan, Mingxi
- Published
- 2015
- Full Text
- View/download PDF
33. Homotopy Based Algorithms for \ell \scriptscriptstyle 0-Regularized Least-Squares.
- Author
-
Soussen, Charles, Idier, Jerome, Duan, Junbo, and Brie, David
- Subjects
HOMOTOPY theory ,COST functions ,LEAST squares ,ALGORITHMS ,ENCYCLOPEDIAS & dictionaries ,RANDOM noise theory - Abstract
Sparse signal restoration is usually formulated as the minimization of a quadratic cost function \Vert \mbi y - \mbi A \mbi x \Vert2^{2} where \mbi A is a dictionary and \mbi x is an unknown sparse vector. It is well-known that imposing an \ell 0 constraint leads to an NP-hard minimization problem. The convex relaxation approach has received considerable attention, where the \ell 0-norm is replaced by the \ell 1-norm. Among the many effective \ell 1 solvers, the homotopy algorithm minimizes \Vert \mbi y - \mbi A \mbi x \Vert2^{2}+\lambda \Vert \mbi x \Vert 1 with respect to \mbi x for a continuum of \lambda ’s. It is inspired by the piecewise regularity of the \ell 1-regularization path, also referred to as the homotopy path. In this paper, we address the minimization problem \Vert \mbi y - \mbi A \mbi x \Vert2^{2}+\lambda \Vert \mbi x \Vert 0 for a continuum of \lambda ’s and propose two heuristic search algorithms for \ell 0-homotopy. Continuation Single Best Replacement is a forward–backward greedy strategy extending the Single Best Replacement algorithm, previously proposed for \ell 0-minimization at a given \lambda . The adaptive search of the \lambda -values is inspired by \ell 1-homotopy. \ell 0 Regularization Path Descent is a more complex algorithm exploiting the structural properties of the \ell 0-regularization path, which is piecewise constant with respect to \lambda . Both algorithms are empirically evaluated for difficult inverse problems involving ill-conditioned dictionaries. Finally, we show that they can be easily coupled with usual methods of model order selection. [ABSTRACT FROM PUBLISHER]
- Published
- 2015
- Full Text
- View/download PDF
34. Modeling exome sequencing data with generalized Gaussian distribution with application to copy number variation detection.
- Author
-
Duan, Junbo, Wan, Mingxi, Deng, Hong-Wen, and Wang, Yu-Ping
- Published
- 2013
- Full Text
- View/download PDF
35. Sparse representation based biomarker selection for schizophrenia with integrated analysis of fMRI and SNP data.
- Author
-
Cao, Hongbao, Duan, Junbo, Lin, Dongdong, Calhoun, Vince, and Wang, Yu-Ping
- Published
- 2013
- Full Text
- View/download PDF
36. Bio marker identification for diagnosis of schizophrenia with integrated analysis of fMRI and SNPs.
- Author
-
Cao, Hongbao, Lin, Dongdong, Duan, Junbo, Wang, Yu-Ping, and Calhoun, Vince
- Abstract
It is important to identify significant biomarkers such as SNPs for medical diagnosis and treatment. However, the size of a biological sample is usually far less than the number of measurements, which makes the problem more challenging. To overcome this difficulty, we propose a sparse representation based variable selection (SRVS) approach. A simulated data set was first tested to demonstrate the advantages and properties of the proposed method. Then, we applied the algorithm to a joint analysis of 759075 SNPs and 153594 functional magnetic resonance imaging (fMRJ) voxels in 208 subjects (92 cases/116 controls) to identify significant biomarkers for schizophrenia (SZ). When compared with previous studies, our proposed method located 20 genes out of the top 45 SZ genes that are publicly reported We also detected some interesting functional brain regions from the fMRI study. In addition, a leave one out (LOO) cross-validation was performed and the results were compared with that of a previously reported method, which showed that our method gave significantly higher classification accuracy. In addition, the identification accuracy with integrative analysis is much better than that of using single type of data, suggesting that integrative analysis may lead to better diagnostic accuracy by combining complementary SNP and fMRI data. [ABSTRACT FROM PUBLISHER]
- Published
- 2012
- Full Text
- View/download PDF
37. Detection of common copy number variation with application to population clustering from next generation sequencing data.
- Author
-
Duan, Junbo, Zhang, Ji-Gang, Deng, Hong-Wen, and Wang, Yu-Ping
- Abstract
Copy number variation (CNV) is a structural variation in human genome that has been associated with many complex diseases. In this paper we present a method to detect common copy number variation from next generation sequencing data. First, copy number variations are detected from each individual sample, which is formulated as a total variation penalized least square problem. Second, the common copy number discovery from multiple samples is obtained using source separation techniques such as the non-negative matrix factorization (NMF). Finally, the method is applied to population clustering. The results on real data analysis show that two family trio with different ancestries can be clustered into two ethnic groups based on their common CNVs, demonstrating the potential of the proposed method for application to population genetics. [ABSTRACT FROM PUBLISHER]
- Published
- 2012
- Full Text
- View/download PDF
38. Common Copy Number Variation Detection From Multiple Sequenced Samples.
- Author
-
Duan, Junbo, Deng, Hong-Wen, and Wang, Yu-Ping
- Subjects
- *
GENOMICS , *GENETIC disorders , *SOMATIC cells , *ONCOLOGY , *GENE expression - Abstract
Common copy number variations (CNVs) refid="ref1"/ are small regions of genomic variations at the same loci across multiple samples, which can be detected with high resolution from next-generation sequencing (NGS) technique. Multiple sequencing data samples are often available from genomic studies; examples include sequences from multiple platforms and sequences from multiple individuals. By integrating complementary information from multiple data samples, detection power can be potentially improved. However, most of current CNV detection methods often process an individual sequence sample, or two samples in an abnormal versus matched normal study; researches on detecting common CNVs across multiple samples have been very limited but are much needed. In this paper, we propose a novel method to detect common CNVs from multiple sequencing samples by exploiting the concurrency of genomic variations in read depth signals derived from multiple NGS data. We use a penalized sparse regression model to fit multiple read depth profiles, based on which common CNV identification is formulated as a change-point detection problem. Finally, we validate the proposed method on both simulation and real data, showing that it can give both higher detection power and better break point estimation over several published CNV detection methods. [ABSTRACT FROM PUBLISHER]
- Published
- 2014
- Full Text
- View/download PDF
39. Comparative Studies of Copy Number Variation Detection Methods for Next-Generation Sequencing Technologies.
- Author
-
Duan, Junbo, Zhang, Ji-Gang, Deng, Hong-Wen, and Wang, Yu-Ping
- Subjects
- *
NUCLEOTIDE sequence , *COMPARATIVE studies , *FLUORESCENCE in situ hybridization , *DISEASE susceptibility , *ALGORITHMS , *COMPUTER simulation , *BIOENGINEERING , *PARAMETER estimation - Abstract
Copy number variation (CNV) has played an important role in studies of susceptibility or resistance to complex diseases. Traditional methods such as fluorescence in situ hybridization (FISH) and array comparative genomic hybridization (aCGH) suffer from low resolution of genomic regions. Following the emergence of next generation sequencing (NGS) technologies, CNV detection methods based on the short read data have recently been developed. However, due to the relatively young age of the procedures, their performance is not fully understood. To help investigators choose suitable methods to detect CNVs, comparative studies are needed. We compared six publicly available CNV detection methods: CNV-seq, FREEC, readDepth, CNVnator, SegSeq and event-wise testing (EWT). They are evaluated both on simulated and real data with different experiment settings. The receiver operating characteristic (ROC) curve is employed to demonstrate the detection performance in terms of sensitivity and specificity, box plot is employed to compare their performances in terms of breakpoint and copy number estimation, Venn diagram is employed to show the consistency among these methods, and F-score is employed to show the overlapping quality of detected CNVs. The computational demands are also studied. The results of our work provide a comprehensive evaluation on the performances of the selected CNV detection methods, which will help biological investigators choose the best possible method. [ABSTRACT FROM AUTHOR]
- Published
- 2013
- Full Text
- View/download PDF
40. Sparse Representation Based Clustering for Integrated Analysis of Gene Copy Number Variation and Gene Expression Data.
- Author
-
Cao, Hongbao, Duan, Junbo, Lin, Dongdong, and Wang, Yu-Ping
- Subjects
GENOMICS ,GENETICS ,GENE expression ,CELL lines ,GENETIC regulation ,CELL culture - Abstract
Integrated analysis of multiple types of genomic data has received increasing attention in recently years, due to the rapid development of new genetic techniques and the strong demand for improvement of the reliability of these techniques. In this work, we proposed a sparse representation based clustering (SRC) method for joint analysis of gene expression and copy number data with the purpose to select significant genes/variables for identification of genes susceptible to a disease. Different from traditional gene selections methods, the proposed SRC model employs information of multifeatures and clusters the data into multi-groups, and then selects significant genes/variables in a particular group. By using joint features extracted from both types of data, the proposed SRC method provides an efficient approach to integrate different types of genomic measurements for comprehensive analysis. Our method has been tested on both breast cancer cell lines and breast tumors data. In addition, simulated data sets were used to test the robustness of the method to several factors such as noise, data sizes and data types. Experiments showed that our proposed method can effectively identify genes/variables with interesting characteristics, e.g., genes/variables with large variations across all genes, and genes/variables that are statistically significant in both measurements with strong correlations. The proposed method can be applicable to a wide variety of biological problems where joint analysis of biological measurements is a common challenge. [ABSTRACT FROM AUTHOR]
- Published
- 2012
41. A COMPRESSED SENSING BASED APPROACH FOR SUBTYPING OF LEUKEMIA FROM GENE EXPRESSION DATA.
- Author
-
TANG, WENLONG, CAO, HONGBAO, DUAN, JUNBO, and WANG, YU-PING
- Subjects
LEUKEMIA diagnosis ,GENE expression ,GENOMICS ,CLASSIFICATION ,STATISTICS ,SIGNAL processing ,DATA analysis - Abstract
With the development of genomic techniques, the demand for new methods that can handle high-throughput genome-wide data effectively is becoming stronger than ever before. Compressed sensing (CS) is an emerging approach in statistics and signal processing. With the CS theory, a signal can be uniquely reconstructed or approximated from its sparse representations, which can therefore better distinguish different types of signals. However, the application of CS approach to genome-wide data analysis has been rarely investigated. We propose a novel CS-based approach for genomic data classification and test its performance in the subtyping of leukemia through gene expression analysis. The detection of subtypes of cancers such as leukemia according to different genetic markups is significant, which holds promise for the individualization of therapies and improvement of treatments. In our work, four statistical features were employed to select significant genes for the classification. With our selected genes out of 7,129 ones, the proposed CS method achieved a classification accuracy of 97.4% when evaluated with the cross validation and 94.3% when evaluated with another independent data set. The robustness of the method to noise was also tested, giving good performance. Therefore, this work demonstrates that the CS method can effectively detect subtypes of leukemia, implying improved accuracy of diagnosis of leukemia. [ABSTRACT FROM AUTHOR]
- Published
- 2011
- Full Text
- View/download PDF
42. From Bernoulli–Gaussian Deconvolution to Sparse Signal Restoration.
- Author
-
Soussen, Charles, Idier, Jérôme, Brie, David, and Duan, Junbo
- Subjects
DIGITAL signal processing ,MATHEMATICAL optimization ,LEAST squares ,ALGORITHMS ,APPROXIMATION theory ,INVERSE problems ,GAUSSIAN processes - Abstract
Formulated as a least square problem under an \ell_0 constraint, sparse signal restoration is a discrete optimization problem, known to be NP complete. Classical algorithms include, by increasing cost and efficiency, matching pursuit (MP), orthogonal matching pursuit (OMP), orthogonal least squares (OLS), stepwise regression algorithms and the exhaustive search. We revisit the single most likely replacement (SMLR) algorithm, developed in the mid-1980s for Bernoulli–Gaussian signal restoration. We show that the formulation of sparse signal restoration as a limit case of Bernoulli–Gaussian signal restoration leads to an \ell_0-penalized least square minimization problem, to which SMLR can be straightforwardly adapted. The resulting algorithm, called single best replacement (SBR), can be interpreted as a forward–backward extension of OLS sharing similarities with stepwise regression algorithms. Some structural properties of SBR are put forward. A fast and stable implementation is proposed. The approach is illustrated on two inverse problems involving highly correlated dictionaries. We show that SBR is very competitive with popular sparse algorithms in terms of tradeoff between accuracy and computation time. [ABSTRACT FROM PUBLISHER]
- Published
- 2011
- Full Text
- View/download PDF
43. Automated Force Volume Image Processing for Biological Samples.
- Author
-
Polyakov, Pavel, Soussen, Charles, Duan, Junbo, Duval, Jèrôme F. L., Brie, David, and Francius, Grégory
- Subjects
ATOMIC force microscopy ,BIOLOGICAL systems ,QUANTITATIVE research ,CRITICAL point (Thermodynamics) ,REGRESSION analysis ,COMPUTER algorithms ,SPECTRUM analysis ,DIGITAL image processing ,NANOELECTROMECHANICAL systems - Abstract
Atomic force microscopy (AFM) has now become a powerful technique for investigating on a molecular level, surface forces, nanomechanical properties of deformable particles, biomolecular interactions, kinetics, and dynamic processes. This paper specifically focuses on the analysis of AFM force curves collected on biological systems, in particular, bacteria. The goal is to provide fully automated tools to achieve theoretical interpretation of force curves on the basis of adequate, available physical models. In this respect, we propose two algorithms, one for the processing of approach force curves and another for the quantitative analysis of retraction force curves. In the former, electrostatic interactions prior to contact between AFM probe and bacterium are accounted for and mechanical interactions operating after contact are described in terms of Hertz- Hooke formalism. Retraction force curves are analyzed on the basis of the Freely Jointed Chain model. For both algorithms, the quantitative reconstruction of force curves is based on the robust detection of critical points (jumps, changes of slope or changes of curvature) which mark the transitions between the various relevant interactions taking place between the AFM tip and the studied sample during approach and retraction. Once the key regions of separation distance and indentation are detected, the physical parameters describing the relevant interactions operating in these regions are extracted making use of regression procedure for fitting experiments to theory. The flexibility, accuracy and strength of the algorithms are illustrated with the processing of two force-volume images, which collect a large set of approach and retraction curves measured on a single biological surface. For each force-volume image, several maps are generated, representing the spatial distribution of the searched physical parameters as estimated for each pixel of the force-volume image. [ABSTRACT FROM AUTHOR]
- Published
- 2011
- Full Text
- View/download PDF
44. Laser-induced photoexcited audible sound effect based on reticular 2-bromo-2-methylpropionic acid modified Fe 3 O 4 nanoparticle aggregates.
- Author
-
Li Y, Zhu H, Duan J, Wu Y, and Wu D
- Abstract
Reticular 2-bromo-2-methylpropionic acid (BMPA) modified Fe
3 O4 nanoparticle aggregates with novel acoustic properties, namely the photoexcited audible sound (PEAS) effect, were prepared by a laser-induced irradiation method. Their morphology was observed by Lorentz transmission electron microscopy. Their chemical structure, crystal composition, and magnetic properties were analyzed using infrared spectroscopy, X-ray diffraction, and a magnetic property measurement instrument, respectively. It is found that the nanoparticle aggregates appeared reticular, with the size of the BMPA modified Fe3 O4 nanoparticles being 5.5 ± 0.4 nm. The saturation magnetization values of the BMPA modified Fe3 O4 nanoparticles and associated aggregates were 59.99 and 63.51 emu g-1 , respectively. The reticular BMPA modified nanoparticle aggregates can produce strong PEAS signals under very weak laser irradiation with great stability and repeatability. The emitted PEAS signals possessed strong specificity, suitable decay time and a large amount of information under a very weak laser power and can be detected by the human ear without any special detection equipment. Subsequently, a heat transfer model was constructed for the simulation of the possible mechanism of the PEAS effect using COMSOL software. The simulation results showed that the aggregates have a fast heat transfer rate with the temperature increasing to 480 K in only 0.25 s and 600 K in 5 s, respectively, meeting the requirements of the vapor explosion mechanism. Therefore, we realized that the possible mechanism of the PEAS effect of the reticular BMPA modified Fe3 O4 nanoparticle aggregates is laser-induced fast heat transfer and vapor explosion in situ , resulting in the observed audible sound phenomenon. This novel PEAS effect has potential for application in materials science, biomedical engineering and other fields.- Published
- 2022
- Full Text
- View/download PDF
45. APAview: A web-based platform for alternative polyadenylation analyses in hematological cancers.
- Author
-
Hu X, Song J, Chyr J, Wan J, Wang X, Du J, Duan J, Zhang H, Zhou X, and Wu X
- Abstract
Background: Hematologic malignancies, such as acute promyelocytic leukemia (APL) and acute myeloid leukemia (AML), are cancers that start in blood-forming tissues and can affect the blood, bone marrow, and lymph nodes. They are often caused by genetic and molecular alterations such as mutations and gene expression changes. Alternative polyadenylation (APA) is a post-transcriptional process that regulates gene expression, and dysregulation of APA contributes to hematological malignancies. RNA-sequencing-based bioinformatic methods can identify APA sites and quantify APA usages as molecular indexes to study APA roles in disease development, diagnosis, and treatment. Unfortunately, APA data pre-processing, analysis, and visualization are time-consuming, inconsistent, and laborious. A comprehensive, user-friendly tool will greatly simplify processes for APA feature screening and mining. Results: Here, we present APAview, a web-based platform to explore APA features in hematological cancers and perform APA statistical analysis. APAview server runs on Python3 with a Flask framework and a Jinja2 templating engine. For visualization, APAview client is built on Bootstrap and Plotly. Multimodal data, such as APA quantified by QAPA/DaPars, gene expression data, and clinical information, can be uploaded to APAview and analyzed interactively. Correlation, survival, and differential analyses among user-defined groups can be performed via the web interface. Using APAview, we explored APA features in two hematological cancers, APL and AML. APAview can also be applied to other diseases by uploading different experimental data., Competing Interests: The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest., (Copyright © 2022 Hu, Song, Chyr, Wan, Wang, Du, Duan, Zhang, Zhou and Wu.)
- Published
- 2022
- Full Text
- View/download PDF
46. CNV_IFTV: An Isolation Forest and Total Variation-Based Detection of CNVs from Short-Read Sequencing Data.
- Author
-
Yuan X, Yu J, Xi J, Yang L, Shang J, Li Z, and Duan J
- Subjects
- Databases, Genetic, Decision Trees, Genome, Human genetics, High-Throughput Nucleotide Sequencing, Humans, Algorithms, Computational Biology methods, DNA Copy Number Variations genetics, Models, Statistical
- Abstract
Accurate detection of copy number variations (CNVs) from short-read sequencing data is challenging due to the uneven distribution of reads and the unbalanced amplitudes of gains and losses. The direct use of read depths to measure CNVs tends to limit performance. Thus, robust computational approaches equipped with appropriate statistics are required to detect CNV regions and boundaries. This study proposes a new method called CNV_IFTV to address this need. CNV_IFTV assigns an anomaly score to each genome bin through a collection of isolation trees. The trees are trained based on isolation forest algorithm through conducting subsampling from measured read depths. With the anomaly scores, CNV_IFTV uses a total variation model to smooth adjacent bins, leading to a denoised score profile. Finally, a statistical model is established to test the denoised scores for calling CNVs. CNV_IFTV is tested on both simulated and real data in comparison to several peer methods. The results indicate that the proposed method outperforms the peer methods. CNV_IFTV is a reliable tool for detecting CNVs from short-read sequencing data even for low-level coverage and tumor purity. The detection results on tumor samples can aid to evaluate known cancer genes and to predict target drugs for disease diagnosis.
- Published
- 2021
- Full Text
- View/download PDF
47. SVSR: A Program to Simulate Structural Variations and Generate Sequencing Reads for Multiple Platforms.
- Author
-
Yuan X, Gao M, Bai J, and Duan J
- Subjects
- Algorithms, Genome, Human genetics, Humans, INDEL Mutation genetics, Polymorphism, Single Nucleotide genetics, Sequence Analysis, DNA methods, Genomic Structural Variation genetics, Genomics methods, High-Throughput Nucleotide Sequencing methods, Software
- Abstract
Structural variation accounts for a major fraction of mutations in the human genome and confers susceptibility to complex diseases. Next generation sequencing along with the rapid development of computational methods provides a cost-effective procedure to detect such variations. Simulation of structural variations and sequencing reads with real characteristics is essential for benchmarking the computational methods. Here, we develop a new program, SVSR, to simulate five types of structural variations (indels, tandem duplication, CNVs, inversions, and translocations) and SNPs for the human genome and to generate sequencing reads with features from popular platforms (Illumina, SOLiD, 454, and Ion Torrent). We adopt a selection model trained from real data to predict copy number states, starting from the first site of a particular genome to the end. Furthermore, we utilize references of microbial genomes to produce insertion fragments and design probabilistic models to imitate inversions and translocations. Moreover, we create platform-specific errors and base quality profiles to generate normal, tumor, or normal-tumor mixture reads. Experimental results show that SVSR could capture more features that are realistic and generate datasets with satisfactory quality scores. SVSR is able to evaluate the performance of structural variation detection methods and guide the development of new computational methods.
- Published
- 2020
- Full Text
- View/download PDF
48. Modeling Ramp-hold Indentation Measurements based on Kelvin-Voigt Fractional Derivative Model.
- Author
-
Zhang H, Zhang Q, Ruan L, Duan J, Wan M, Insana MF, Zhang H, Zhang Q, Ruan L, Duan J, Wan M, and Insana MF
- Abstract
Interpretation of experimental data from micro- and nano-scale indentation testing is highly dependent on the constitutive model selected to relate measurements to mechanical properties. The Kelvin-Voigt Fractional Derivative model (KVFD) offers a compact set of viscoelastic features appropriate for characterizing soft biological materials. This paper provides a set of KVFD solutions for converting indentation testing data acquired for different geometries and scales into viscoelastic properties of soft materials. These solutions, which are mostly in closed-form, apply to ramp-hold relaxation, load-unload and ramp-load creep-testing protocols. We report on applications of these model solutions to macro- and nano-indentation testing of hydrogels, gastric cancer cells and ex vivo breast tissue samples using an Atomic Force Microscope (AFM). We also applied KVFD models to clinical ultrasonic breast data using a compression plate as required for elasticity imaging. Together the results show that KVFD models fit a broad range of experimental data with a correlation coefficient typically R
2 > 0.99. For hydrogel samples, estimation of KVFD model parameters from test data using spherical indentation versus plate compression as well as ramp relaxation versus load-unload compression all agree within one standard deviation. Results from measurements made using macro- and nano-scale indentation agree in trend. For gastric cell and ex vivo breast tissue measurements, KVFD moduli are, respectively, 1/3 - 1/2 and 1/6 of the elasticity modulus found from the Sneddon model. In vivo breast tissue measurements yield model parameters consistent with literature results. The consistency of results found for a broad range of experimental parameters suggest the KVFD model is a reliable tool for exploring intrinsic features of the cell/tissue microenvironments.- Published
- 2018
- Full Text
- View/download PDF
49. A Sparse Model Based Detection of Copy Number Variations From Exome Sequencing Data.
- Author
-
Duan J, Wan M, Deng HW, and Wang YP
- Subjects
- Algorithms, Humans, Reproducibility of Results, Computational Biology methods, DNA Copy Number Variations genetics, Exome genetics, Sequence Analysis, DNA methods
- Abstract
Goal: Whole-exome sequencing provides a more cost-effective way than whole-genome sequencing for detecting genetic variants, such as copy number variations (CNVs). Although a number of approaches have been proposed to detect CNVs from whole-genome sequencing, a direct adoption of these approaches to whole-exome sequencing will often fail because exons are separately located along a genome. Therefore, an appropriate method is needed to target the specific features of exome sequencing data., Methods: In this paper, a novel sparse model based method is proposed to discover CNVs from multiple exome sequencing data. First, exome sequencing data are represented with a penalized matrix approximation, and technical variability and random sequencing errors are assumed to follow a generalized Gaussian distribution. Second, an iteratively reweighted least squares algorithm is used to estimate the solution., Results: The method is tested and validated on both synthetic and real data, and compared with other approaches including CoNIFER, XHMM, and cn.MOPS. The test demonstrates that the proposed method outperform other approaches., Conclusion: The proposed sparse model can detect CNVs from exome sequencing data with high power and precision. Significance: Sparse model can target the specific features of exome sequencing data. The software codes are freely available at http://www.tulane.edu/ wyp/software/Exon_CNV.m.
- Published
- 2016
- Full Text
- View/download PDF
50. An optimal method to segment piecewise poisson distributed signals with application to sequencing data.
- Author
-
Duan J, Soussen C, Brie D, Idier J, Wang YP, and Wan M
- Subjects
- Animals, DNA analysis, DNA genetics, DNA Copy Number Variations, Genome, Humans, Normal Distribution, Poisson Distribution, Sequence Analysis, DNA, Algorithms, High-Throughput Nucleotide Sequencing methods
- Abstract
To analyze the next generation sequencing data, the so-called read depth signal is often segmented with standard segmentation tools. However, these tools usually assume the signal to be a piecewise constant signal and contaminated with zero mean Gaussian noise, and therefore modeling error occurs. This paper models the read depth signal with piecewise Poisson distribution, which is more appropriate to the next generation sequencing mechanism. Based on the proposed model, an opti- mal dynamic programming algorithm with parallel computing is proposed to segment the piecewise signal, and furthermore detect the copy number variation.
- Published
- 2015
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.