95 results on '"Liquet, B"'
Search Results
2. In-situ measurements of energetic depth-limited wave loading
- Author
-
Poncet, P.A., Liquet, B., Larroque, B., D’Amico, D., Sous, D., and Abadie, S.
- Published
- 2022
- Full Text
- View/download PDF
3. Robust Estimation Procedure for Autoregressive Models with Heterogeneity
- Author
-
Callens, A., Wang, Y.-G., Fu, L., and Liquet, B.
- Published
- 2021
- Full Text
- View/download PDF
4. ClustOfVar: An R Package for the Clustering of Variables
- Author
-
Chavent, M., Kuentz, V., Liquet, B., and Saracco, L.
- Subjects
Statistics - Computation - Abstract
Clustering of variables is as a way to arrange variables into homogeneous clusters, i.e., groups of variables which are strongly related to each other and thus bring the same information. These approaches can then be useful for dimension reduction and variable selection. Several specific methods have been developed for the clustering of numerical variables. However concerning qualitative variables or mixtures of quantitative and qualitative variables, far fewer methods have been proposed. The R package ClustOfVar was specifically developed for this purpose. The homogeneity criterion of a cluster is defined as the sum of correlation ratios (for qualitative variables) and squared correlations (for quantitative variables) to a synthetic quantitative variable, summarizing "as good as possible" the variables in the cluster. This synthetic variable is the first principal component obtained with the PCAMIX method. Two algorithms for the clustering of variables are proposed: iterative relocation algorithm and ascendant hierarchical clustering. We also propose a bootstrap approach in order to determine suitable numbers of clusters. We illustrate the methodologies and the associated package on small datasets.
- Published
- 2011
5. Bootstrap Choice of Estimators in Parametric and Semiparametric Families: An Extension of EIC
- Author
-
Liquet, B., Sakarovitch, C., and Commenges, D.
- Published
- 2003
6. Community evaluation of glycoproteomics informatics solutions reveals high-performance search strategies for serum glycopeptide analysis (vol 18, pg 1304, 2021)
- Author
-
Kawahara, R, Chernykh, A, Alagesan, K, Bern, M, Cao, W, Chalkley, RJ, Cheng, K, Choo, MS, Edwards, N, Goldman, R, Hoffmann, M, Hu, Y, Huang, Y, Kim, JY, Kletter, D, Liquet, B, Liu, M, Mechref, Y, Meng, B, Neelamegham, S, Nguyen-Khuong, T, Nilsson, J, Pap, A, Park, GW, Parker, BL, Pegg, CL, Penninger, JM, Phung, TK, Pioch, M, Rapp, E, Sakalli, E, Sanda, M, Schulz, BL, Scott, NE, Sofronov, G, Stadlmann, J, Vakhrushev, SY, Woo, CM, Wu, H-Y, Yang, P, Ying, W, Zhang, H, Zhang, Y, Zhao, J, Zaia, J, Haslam, SM, Palmisano, G, Yoo, JS, Larson, G, Khoo, K-H, Medzihradszky, KF, Kolarich, D, Packer, NH, Thaysen-Andersen, M, Kawahara, R, Chernykh, A, Alagesan, K, Bern, M, Cao, W, Chalkley, RJ, Cheng, K, Choo, MS, Edwards, N, Goldman, R, Hoffmann, M, Hu, Y, Huang, Y, Kim, JY, Kletter, D, Liquet, B, Liu, M, Mechref, Y, Meng, B, Neelamegham, S, Nguyen-Khuong, T, Nilsson, J, Pap, A, Park, GW, Parker, BL, Pegg, CL, Penninger, JM, Phung, TK, Pioch, M, Rapp, E, Sakalli, E, Sanda, M, Schulz, BL, Scott, NE, Sofronov, G, Stadlmann, J, Vakhrushev, SY, Woo, CM, Wu, H-Y, Yang, P, Ying, W, Zhang, H, Zhang, Y, Zhao, J, Zaia, J, Haslam, SM, Palmisano, G, Yoo, JS, Larson, G, Khoo, K-H, Medzihradszky, KF, Kolarich, D, Packer, NH, and Thaysen-Andersen, M
- Published
- 2022
7. Central subspaces review: methods and applications
- Author
-
Rodrigues, SA, Huggins, R, Liquet, B, Rodrigues, SA, Huggins, R, and Liquet, B
- Published
- 2022
8. Influence du type d’orthèse de genou sur l’évolution clinique postopératoire d’une chirurgie du ligament croisé antérieur chez le sportif
- Author
-
Laboute, E., Liquet, B., Savalli, L., Puig, P., and Trouve, P.
- Published
- 2011
- Full Text
- View/download PDF
9. Analyse de la pléiotropie dans les GWAS à l’aide de méthodes bayésiennes prenant en compte la structure de groupe de variables
- Author
-
Baghfalaki, T., primary, Sugier, P., additional, Truong, T., additional, Pettitt, A., additional, Mengersen, K., additional, and Liquet, B., additional
- Published
- 2021
- Full Text
- View/download PDF
10. An appraisal of respiratory system compliance in mechanically ventilated covid-19 patients
- Author
-
Li Bassi, G, Suen, JY, Dalton, HJ, White, N, Shrapnel, S, Fanning, JP, Liquet, B, Hinton, S, Vuorinen, A, Booth, G, Millar, JE, Forsyth, S, Panigada, M, Laffey, J, Brodie, D, Fan, E, Torres, A, Chiumello, D, Corley, A, Elhazmi, A, Hodgson, C, Ichiba, S, Luna, C, Murthy, S, Nichol, A, Ng, PY, Ogino, M, Pesenti, A, Huynh, TT, Fraser, JF, Li Bassi, G, Suen, JY, Dalton, HJ, White, N, Shrapnel, S, Fanning, JP, Liquet, B, Hinton, S, Vuorinen, A, Booth, G, Millar, JE, Forsyth, S, Panigada, M, Laffey, J, Brodie, D, Fan, E, Torres, A, Chiumello, D, Corley, A, Elhazmi, A, Hodgson, C, Ichiba, S, Luna, C, Murthy, S, Nichol, A, Ng, PY, Ogino, M, Pesenti, A, Huynh, TT, and Fraser, JF
- Abstract
BACKGROUND: Heterogeneous respiratory system static compliance (CRS) values and levels of hypoxemia in patients with novel coronavirus disease (COVID-19) requiring mechanical ventilation have been reported in previous small-case series or studies conducted at a national level. METHODS: We designed a retrospective observational cohort study with rapid data gathering from the international COVID-19 Critical Care Consortium study to comprehensively describe CRS-calculated as: tidal volume/[airway plateau pressure-positive end-expiratory pressure (PEEP)]-and its association with ventilatory management and outcomes of COVID-19 patients on mechanical ventilation (MV), admitted to intensive care units (ICU) worldwide. RESULTS: We studied 745 patients from 22 countries, who required admission to the ICU and MV from January 14 to December 31, 2020, and presented at least one value of CRS within the first seven days of MV. Median (IQR) age was 62 (52-71), patients were predominantly males (68%) and from Europe/North and South America (88%). CRS, within 48 h from endotracheal intubation, was available in 649 patients and was neither associated with the duration from onset of symptoms to commencement of MV (p = 0.417) nor with PaO2/FiO2 (p = 0.100). Females presented lower CRS than males (95% CI of CRS difference between females-males: - 11.8 to - 7.4 mL/cmH2O p < 0.001), and although females presented higher body mass index (BMI), association of BMI with CRS was marginal (p = 0.139). Ventilatory management varied across CRS range, resulting in a significant association between CRS and driving pressure (estimated decrease - 0.31 cmH2O/L per mL/cmH20 of CRS, 95% CI - 0.48 to - 0.14, p < 0.001). Overall, 28-day ICU mortality, accounting for the competing risk of being discharged within the period, was 35.6% (SE 1.7). Cox proportional hazard analysis demonstrated that CRS (+ 10 mL/cm H2O) was only associated with being discharge from the ICU within 28 days (HR 1.14, 95% CI 1.02-1.
- Published
- 2021
11. A new method for integrated analysis applied to gene expression and cytokines secretion in response to LIPO-5 vaccine in HIV-negative volunteers
- Author
-
Thiebaut R, Liquet B, Hocini H, Hue S, Richert L, Raimbault M, Lê Cao K, and Levy Y
- Subjects
Immunologic diseases. Allergy ,RC581-607 - Published
- 2012
- Full Text
- View/download PDF
12. Robust Estimation Procedure for Autoregressive Models with Heterogeneity
- Author
-
Callens, A., primary, Wang, Y.-G., additional, Fu, L., additional, and Liquet, B., additional
- Published
- 2020
- Full Text
- View/download PDF
13. Analysis of the pleiotropy between breast cancer and thyroid cancer
- Author
-
Lucotte, E., primary, Sugier, P., additional, Lefranc, A., additional, Boland, A., additional, Deleuze, J., additional, Ostroumovae, E., additional, Boutron, M., additional, de Vathaire, F., additional, Guénel, P., additional, Liquet, B., additional, and Truong, T., additional
- Published
- 2020
- Full Text
- View/download PDF
14. Mesures in-situ d'impacts de vagues sur une digue composite
- Author
-
Poncet, P.A., Abadie, S., Larroque, B., Liquet, B., and Sous, D.
- Subjects
aération ,impact ,vague ,digue ,mesure - Abstract
Afin d’étudier les pressions d’impact in-situ, la digue de l’Artha à Saint-Jean de Luz a été équipée de deux capteurs de pression haute-fréquence disposés l’un au dessus de l’autre sur la partie la plus raide du musoir ouest. Ces deux capteurs ont ainsi enregistré le signal de pression à 10 kHz par tranches de 10 minutes toutes les heures de janvier à avril 2016. Par ailleurs le champs de vagues au large est mesuré par une bouée directionnelle, le niveau d’eau est obtenu au marégraphe de Socoa et le vent est mesuré au sémaphore de Socoa. Les mesures de pression sont en accord avec des études similaires précédentes([1], [2]), mettant en évidence des pics de pressions significativement plus faibles que ceux obtenus en canal [3] ou par simulation numérique. La structure de la digue est à l’origine de phénomènes qui sont en général susceptibles de minimiser les impacts par rapport au cas idéalisé. Un des objectifs de cette étude est aussi d’identifier les conditions qui génèrent les impacts les plus destructeurs. Une première analyse statistique de l’influence des facteurs environnementaux a été réalisé. Une étude du signal de pression brut montre que la digue est soumise à différents types d’impact. La grande majorité des impacts sont relativement lents et leurs intensités en partie contrôlée par la hauteur de vague. Mais des impacts intenses et rapides ont aussi été mesurés dans des conditions de houle et de vent relativement calme. L’intensité de ces impacts ne peut être expliquée uniquement par la pression hydrostatique. Des effets potentiel de compression de poche d’air sont étudiés. La prochaine campagne de mesure qui emploiera plus d’une vingtaine de capteurs permettra peut être de détecter plus d’impacts de ce type et de mieux comprendre leur dynamique.
- Published
- 2018
- Full Text
- View/download PDF
15. Pre-diagnostic blood immune markers, incidence and progression of B-cell lymphoma and multiple myeloma; univariate and functionally-informed multivariate analyses
- Author
-
Vermeulen, R, Saberi Hosnijeh, F, Bodinier, B, Portengen, L, Liquet, B, Garrido Manriquez, J, Lokhorst, H, Bergdahl, I, Kyrtopoulos, S, Johansson, A-S, Georgiadis, P, Melin, B, Palli, D, Krogh, V, Panico, S, Sacerdote, C, Tumino, R, Vineis, P, Castagne, RS, Chadeau, M, and Cancer Research UK
- Subjects
mixed-effect modeling ,Science & Technology ,BONE-MARROW ,prospective cohort ,FACTOR-ALPHA ,lymphoma ,multivariate models ,SERUM-LEVELS ,time to diagnosis ,OVARIAN-CANCER ,TRANSFORMING-GROWTH-FACTOR ,multiple myeloma ,CYTOKINE LEVELS ,EnviroGenoMarkers Consortium Consortium members ,Oncology ,POOR-PROGNOSIS ,cytokine ,INDEPENDENT PREDICTOR ,Oncology & Carcinogenesis ,NON-HODGKIN-LYMPHOMA ,Life Sciences & Biomedicine ,SOLUBLE CD30 ,1112 Oncology And Carcinogenesis - Abstract
Recent prospective studies have shown that dysregulation of the immune system may precede the development of B-cell lymphomas (BCL) in immunocompetent individuals. However, to date, the studies were restricted to a few immune markers, which were considered separately. Using a nested case-control study within two European prospective cohorts, we measured plasma levels of 28 immune markers in samples collected a median of 6 years prior to diagnosis (range, 2.01-15.97) in 268 incident cases of BCL (including multiple myeloma) and matched controls. Linear mixed models, and Partial Least Square analyses were used to analyze the association between levels of immune marker and the incidence of BCL and its main histological subtypes, and to investigate potential biomarkers predictive of the time to diagnosis. Linear mixed modelIrrespective of the model, our analyses identified associations linking blood lower immune markerslevels of and BCL incidence. In particular, we identified growth factors, and within that family, fibroblast growth factor-2 (FGF-2 , p=7.2x10 -4 ), ) and transforming growth factor alpha (TGF-α , p=6.5x10 -5 ) and BCL incidence. Analyses stratified by histological subtypes identified inverse associations for MM subtype including FGF-2 (p=7.8x10 -7 ), TGF-α (p=4.08x10 -5 ), fractalkine (p=1.12x10 -3 ), monocyte chemotactic protein-3 (p=1.36x10 -4 ), macrophage inflammatory protein 1-alpha (p=4.6x10 -4 ), and vascular endothelial growth factor (p=4.23x10 -5 ). , and vascular endothelial growth factor (VEGF), to be consistently (and inversely) associated with MM incidence. Our results also provide d marginal support for already reported associations between chemokines and diffuse large B-Cell lymphoma (DLBCL) , and cytokines and chronic lymphocytic leukemia (CLL) . Case-only analyses showed that GM-CSF levels were consistently higher closer to diagnosis, which provides further evidence of its role in tumor progression. In conclusion, our study suggests a role of growth-factors in the incidence of MM, and of chemokine and cytokine regulation in DLBCL and CLL.
- Published
- 2018
16. Pleiotropic mapping for genome-wide association studies using group variable selection
- Author
-
Liquet, B., primary
- Published
- 2019
- Full Text
- View/download PDF
17. PLS for Big Data: A unified parallel algorithm for regularised group PLS
- Author
-
Lafaye De Micheaux, P, Liquet, B, Sutton, M, Lafaye De Micheaux, P, Liquet, B, and Sutton, M
- Abstract
Partial Least Squares (PLS) methods have been heavily exploited to analyse the association between two blocks of data. These powerful approaches can be applied to data sets where the number of variables is greater than the number of observations and in the presence of high collinearity between variables. Different sparse versions of PLS have been developed to integrate multiple data sets while simultaneously selecting the contributing variables. Sparse modeling is a key factor in obtaining better estimators and identifying associations between multiple data sets. The cornerstone of the sparse PLS methods is the link between the singular value decomposition (SVD) of a matrix (constructed from deflated versions of the original data) and least squares minimization in linear regression. We review four popular PLS methods for two blocks of data. A unified algorithm is proposed to perform all four types of PLS including their regularised versions. We present various approaches to decrease the computation time and show how the whole procedure can be scalable to big data sets. The bigsgPLS R package implements our unified algorithm and is available at https://github.com/matt-sutton/bigsgPLS.
- Published
- 2019
18. A multivariate approach to investigate the combined biological effects of multiple exposures
- Author
-
Chadeau, M, Jain, P, Vineis, P, Liquet, B, Vlaanderen, J, Bodinier, B, Van Veldhoven, C, Kogevinas, M, Athersuch, TJ, Font-Ribera, L, Villanueva, C, Vermeulen, R, One Health Chemisch, dIRAS RA-2, University of Torino and CPO-Piemonte, Università degli studi di Torino (UNITO), Laboratoire de Mathématiques et de leurs Applications [Pau] (LMAP), Université de Pau et des Pays de l'Adour (UPPA)-Centre National de la Recherche Scientifique (CNRS), Centre International de Recherche contre le Cancer - International Agency for Research on Cancer (CIRC - IARC), Organisation Mondiale de la Santé / World Health Organization Office (OMS / WHO), CIBER de Epidemiología y Salud Pública (CIBERESP), Center for Research in Environmental Epidemiology (CREAL), Universitat Pompeu Fabra [Barcelona] (UPF)-Catalunya ministerio de salud, Division of Environmental Epidemiology, Utrecht University [Utrecht]-Institute of Risk Assessment Sciences, Department of Epidemiology and Biostatistics, MRC-HPA Centre for Environment and Health, School of Public Health, Imperial College London, Commission of the European Communities, and Cancer Research UK
- Subjects
1604 Human Geography ,Epidemiology ,Medi ambient ,Environment ,[STAT]Statistics [stat] ,Multiple exposures ,Exposome ,1117 Public Health And Health Services ,Multivariate response ,OMICs data ,Epidemiologia ,Multi-level sparse PLS models ,ComputingMilieux_MISCELLANEOUS - Abstract
Epidemiological studies provide evidence that environmental exposures may affect health through complex mixtures. Formal investigation of the effect of exposure mixtures is usually achieved by modelling interactions, which relies on strong assumptions relating to the identity and the number of the exposures involved in such interactions, and on the order and parametric form of these interactions. These hypotheses become difficult to formulate and justify in an exposome context, where influential exposures are numerous and heterogeneous. To capture both the complexity of the exposome and its possibly pleiotropic effects, models handling multivariate predictors and responses, such as partial least squares (PLS) algorithms, can prove useful. As an illustrative example, we applied PLS models to data from a study investigating the inflammatory response (blood concentration of 13 immune markers) to the exposure to four disinfection by-products (one brominated and three chlorinated compounds), while swimming in a pool. To accommodate the multiple observations per participant (n=60; before and after the swim), we adopted a multilevel extension of PLS algorithms, including sparse PLS models shrinking loadings coefficients of unimportant predictors (exposures) and/or responses (protein levels). Despite the strong correlation among co-occurring exposures, our approach identified a subset of exposures (n=3/4) affecting the exhaled levels of 8 (out of 13) immune markers. PLS algorithms can easily scale to high-dimensional exposures and responses, and prove useful for exposome research to identify sparse sets of exposures jointly affecting a set of (selected) biological markers. Our descriptive work may guide these extensions for higher dimensional data.
- Published
- 2018
- Full Text
- View/download PDF
19. Bayesian Variable Selection Regression of Multivariate Responses for Group Data
- Author
-
Liquet, B., primary, Mengersen, K., additional, Pettitt, A. N., additional, and Sutton, M., additional
- Published
- 2017
- Full Text
- View/download PDF
20. A Unified Parallel Algorithm for Regularized Group PLS Scalable to Big Data
- Author
-
Micheaux, PLD, Liquet, B, Sutton, M, Micheaux, PLD, Liquet, B, and Sutton, M
- Abstract
Partial Least Squares (PLS) methods have been heavily exploited to analysethe association between two blocs of data. These powerful approaches can beapplied to data sets where the number of variables is greater than the numberof observations and in presence of high collinearity between variables.Different sparse versions of PLS have been developed to integrate multiple datasets while simultaneously selecting the contributing variables. Sparsemodelling is a key factor in obtaining better estimators and identifyingassociations between multiple data sets. The cornerstone of the sparsityversion of PLS methods is the link between the SVD of a matrix (constructedfrom deflated versions of the original matrices of data) and least squaresminimisation in linear regression. We present here an accurate description ofthe most popular PLS methods, alongside their mathematical proofs. A unifiedalgorithm is proposed to perform all four types of PLS including theirregularised versions. Various approaches to decrease the computation time areoffered, and we show how the whole procedure can be scalable to big data sets.
- Published
- 2017
21. Classification de variables avec la méthode PCAMIX
- Author
-
CHAVENT, Marie, KUENTZ, V., LIQUET, B., and SARACCO, Jérôme
- Published
- 2011
22. Approche de classification par la méthode PCAMIX
- Author
-
CHAVENT, Marie, KUENTZ, V., LIQUET, B., and SARACCO, Jérôme
- Published
- 2011
23. ClustOfVar : un package R pour la classification de variables
- Author
-
CHAVENT, Marie, KUENTZ, V., LIQUET, B., and SARACCO, Jérôme
- Published
- 2011
24. Rôle de l’interféron-α dans l’activation immune chronique des patients VIH avec une charge virale indétectable sous traitement : une approche par équations structurelles
- Author
-
Picat, M.-Q., primary, Pellegrin, I., additional, Bitard, J., additional, Wittkop, L., additional, Proust-Lima, C., additional, Liquet, B., additional, Moreau, J.-F., additional, and Thiébaut, R., additional
- Published
- 2014
- Full Text
- View/download PDF
25. A novel approach for biomarker selection and the integration of repeated measures experiments from two assays
- Author
-
Liquet, B, Le Cao, K-A, Hocini, H, Thiebaut, R, Liquet, B, Le Cao, K-A, Hocini, H, and Thiebaut, R
- Abstract
BACKGROUND: High throughput 'omics' experiments are usually designed to compare changes observed between different conditions (or interventions) and to identify biomarkers capable of characterizing each condition. We consider the complex structure of repeated measurements from different assays where different conditions are applied on the same subjects. RESULTS: We propose a two-step analysis combining a multilevel approach and a multivariate approach to reveal separately the effects of conditions within subjects from the biological variation between subjects. The approach is extended to two-factor designs and to the integration of two matched data sets. It allows internal variable selection to highlight genes able to discriminate the net condition effect within subjects. A simulation study was performed to demonstrate the good performance of the multilevel multivariate approach compared to a classical multivariate method. The multilevel multivariate approach outperformed the classical multivariate approach with respect to the classification error rate and the selection of relevant genes. The approach was applied to an HIV-vaccine trial evaluating the response with gene expression and cytokine secretion. The discriminant multilevel analysis selected a relevant subset of genes while the integrative multilevel analysis highlighted clusters of genes and cytokines that were highly correlated across the samples. CONCLUSIONS: Our combined multilevel multivariate approach may help in finding signatures of vaccine effect and allows for a better understanding of immunological mechanisms activated by the intervention. The integrative analysis revealed clusters of genes, that were associated with cytokine secretion. These clusters can be seen as gene signatures to predict future cytokine response. The approach is implemented in the R package mixOmics (http://cran.r-project.org/) with associated tutorials to perform the analysis(a).
- Published
- 2012
26. Investigating trial and treatment heterogeneity in an individual patient data meta-analysis of survival data by means of the penalized maximum likelihood approach
- Author
-
Rondeau, V., primary, Michiels, S., additional, Liquet, B., additional, and Pignon, J. P., additional
- Published
- 2008
- Full Text
- View/download PDF
27. Correction of the P-value after multiple coding of an explanatory variable in logistic regression.
- Author
-
Liquet, Benoit, Commenges, Daniel, Liquet, B, and Commenges, D
- Published
- 2001
- Full Text
- View/download PDF
28. Best Subset Solution Path for Linear Dimension Reduction Models Using Continuous Optimization.
- Author
-
Liquet B, Moka S, and Muller S
- Subjects
- Least-Squares Analysis, Algorithms, Linear Models, Models, Statistical, Principal Component Analysis, Biometry methods
- Abstract
The selection of best variables is a challenging problem in supervised and unsupervised learning, especially in high-dimensional contexts where the number of variables is usually much larger than the number of observations. In this paper, we focus on two multivariate statistical methods: principal components analysis and partial least squares. Both approaches are popular linear dimension-reduction methods with numerous applications in several fields including in genomics, biology, environmental science, and engineering. In particular, these approaches build principal components, new variables that are combinations of all the original variables. A main drawback of principal components is the difficulty to interpret them when the number of variables is large. To define principal components from the most relevant variables, we propose to cast the best subset solution path method into principal component analysis and partial least square frameworks. We offer a new alternative by exploiting a continuous optimization algorithm for best subset solution path. Empirical studies show the efficacy of our approach for providing the best subset solution path. The usage of our algorithm is further exposed through the analysis of two real data sets. The first data set is analyzed using the principle component analysis while the analysis of the second data set is based on partial least square framework., (© 2024 Wiley‐VCH GmbH.)
- Published
- 2025
- Full Text
- View/download PDF
29. Mixture Cure Semiparametric Accelerated Failure Time Models With Partly Interval-Censored Data.
- Author
-
Li I, Ma J, and Liquet B
- Subjects
- Humans, Survival Analysis, Time Factors, Proportional Hazards Models, Likelihood Functions, Melanoma drug therapy, Biometry methods, Models, Statistical
- Abstract
In practical survival analysis, the situation of no event for a patient can arise even after a long period of waiting time, which means a portion of the population may never experience the event of interest. Under this circumstance, one remedy is to adopt a mixture cure Cox model to analyze the survival data. However, if there clearly exhibits an acceleration (or deceleration) factor among their survival times, then an accelerated failure time (AFT) model will be preferred, leading to a mixture cure AFT model. In this paper, we consider a penalized likelihood method to estimate the mixture cure semiparametric AFT models, where the unknown baseline hazard is approximated using Gaussian basis functions. We allow partly interval-censored survival data which can include event times and left-, right-, and interval-censoring times. The penalty function helps to achieve a smooth estimate of the baseline hazard function. We will also provide asymptotic properties to the estimates so that inferences can be made on regression parameters and hazard-related quantities. Simulation studies are conducted to evaluate the model performance, which includes a comparative study with an existing method from the smcure R package. The results show that our proposed penalized likelihood method has acceptable performance in general and produces less bias when faced with the identifiability issue compared to smcure. To illustrate the application of our method, a real case study involving melanoma recurrence is conducted and reported. Our model is implemented in our R package aftQnp which is available from https://github.com/Isabellee4555/aftQnP., (© 2024 Wiley‐VCH GmbH.)
- Published
- 2024
- Full Text
- View/download PDF
30. Deep learning-based hyperspectral image correction and unmixing for brain tumor surgery.
- Author
-
Black D, Gill J, Xie A, Liquet B, Di Ieva A, Stummer W, and Suero Molina E
- Abstract
Hyperspectral imaging for fluorescence-guided brain tumor resection improves visualization of tissue differences, which can ameliorate patient outcomes. However, current methods do not effectively correct for heterogeneous optical and geometric tissue properties, leading to less accurate results. We propose two deep learning models for correction and unmixing that can capture these effects. While one is trained with protoporphyrin IX (PpIX) concentration labels, the other is semi-supervised. The models were evaluated on phantom and pig brain data with known PpIX concentration; the supervised and semi-supervised models achieved Pearson correlation coefficients (phantom, pig brain) between known and computed PpIX concentrations of (0.997, 0.990) and (0.98, 0.91), respectively. The classical approach achieved (0.93, 0.82). The semi-supervised approach also generalizes better to human data, achieving a 36% lower false-positive rate for PpIX detection and giving qualitatively more realistic results than existing methods. These results show promise for using deep learning to improve hyperspectral fluorescence-guided neurosurgery., Competing Interests: E.S.M. received research support from Carl Zeiss Meditec AG. W.S. has received speaker and consultant fees from SBI ALA Pharma, medac, Carl Zeiss Meditec AG, NXDC, and research support from Zeiss., (© 2024 The Authors.)
- Published
- 2024
- Full Text
- View/download PDF
31. Spectral library and method for sparse unmixing of hyperspectral images in fluorescence guided resection of brain tumors.
- Author
-
Black D, Liquet B, Di Ieva A, Stummer W, and Suero Molina E
- Abstract
Through spectral unmixing, hyperspectral imaging (HSI) in fluorescence-guided brain tumor surgery has enabled the detection and classification of tumor regions invisible to the human eye. Prior unmixing work has focused on determining a minimal set of viable fluorophore spectra known to be present in the brain and effectively reconstructing human data without overfitting. With these endmembers, non-negative least squares regression (NNLS) was commonly used to compute the abundances. However, HSI images are heterogeneous, so one small set of endmember spectra may not fit all pixels well. Additionally, NNLS is the maximum likelihood estimator only if the measurement is normally distributed, and it does not enforce sparsity, which leads to overfitting and unphysical results. In this paper, we analyzed 555666 HSI fluorescence spectra from 891 ex vivo measurements of patients with various brain tumors to show that a Poisson distribution indeed models the measured data 82% better than a Gaussian in terms of the Kullback-Leibler divergence, and that the endmember abundance vectors are sparse. With this knowledge, we introduce (1) a library of 9 endmember spectra, including PpIX (620 nm and 634 nm photostates), NADH, FAD, flavins, lipofuscin, melanin, elastin, and collagen, (2) a sparse, non-negative Poisson regression algorithm to perform physics-informed unmixing with this library without overfitting, and (3) a highly realistic spectral measurement simulation with known endmember abundances. The new unmixing method was then tested on the human and simulated data and compared to four other candidate methods. It outperforms previous methods with 25% lower error in the computed abundances on the simulated data than NNLS, lower reconstruction error on human data, better sparsity, and 31 times faster runtime than state-of-the-art Poisson regression. This method and library of endmember spectra can enable more accurate spectral unmixing to aid the surgeon better during brain tumor resection., Competing Interests: Eric Suero Molina received research support from Carl Zeiss Meditec AG. Walter Stummer has received speaker and consultant fees from SBI ALA Pharma, medac, Carl Zeiss Meditec AG, and NXDC, as well as research support from Zeiss. None of the other authors have any competing interests., (© 2024 Optica Publishing Group.)
- Published
- 2024
- Full Text
- View/download PDF
32. Navigating Mathematical Basics: A Primer for Deep Learning in Science.
- Author
-
Liquet B, Moka S, and Nazarathy Y
- Subjects
- Humans, Algorithms, Mathematics education, Neural Networks, Computer, Science education, Deep Learning
- Abstract
We present a gentle introduction to elementary mathematical notation with the focus of communicating deep learning principles. This is a "math crash course" aimed at quickly enabling scientists with understanding of the building blocks used in many equations, formulas, and algorithms that describe deep learning. While this short presentation cannot replace solid mathematical knowledge that needs multiple courses and years to solidify, our aim is to allow nonmathematical readers to overcome hurdles of reading texts that also use such mathematical notation. We describe a few basic deep learning models using mathematical notation before we unpack the meaning of the notation. In particular, this text includes an informal introduction to summations, sets, functions, vectors, matrices, gradients, and a few more objects that are often used to describe deep learning. While this is a mathematical crash course, our presentation is kept in the context of deep learning and machine learning models including the sigmoid model, the softmax model, and fully connected feedforward deep neural networks. We also hint at basic mathematical objects appearing in neural networks for images and text data., (© 2024. The Author(s), under exclusive license to Springer Nature Switzerland AG.)
- Published
- 2024
- Full Text
- View/download PDF
33. Investigation of common genetic risk factors between thyroid traits and breast cancer.
- Author
-
Lucotte EA, Asgari Y, Sugier PE, Karimi M, Domenighetti C, Lesueur F, Boland-Augé A, Ostroumova E, de Vathaire F, Zidane M, Guénel P, Deleuze JF, Boutron-Ruault MC, Severi G, Liquet B, and Truong T
- Subjects
- Humans, Female, Thyrotropin genetics, Thyroxine genetics, Risk Factors, Genetic Risk Score, Thyroid Gland, Breast Neoplasms genetics
- Abstract
Breast cancer (BC) risk is suspected to be linked to thyroid disorders, however observational studies exploring the association between BC and thyroid disorders gave conflicting results. We proposed an alternative approach by investigating the shared genetic risk factors between BC and several thyroid traits. We report a positive genetic correlation between BC and thyroxine (FT4) levels (corr = 0.13, p-value = 2.0 × 10-4) and a negative genetic correlation between BC and thyroid-stimulating hormone (TSH) levels (corr = -0.09, p-value = 0.03). These associations are more striking when restricting the analysis to estrogen receptor-positive BC. Moreover, the polygenic risk scores (PRS) for FT4 and hyperthyroidism are positively associated to BC risk (OR = 1.07, 95%CI: 1.00-1.13, p-value = 2.8 × 10-2 and OR = 1.04, 95%CI: 1.00-1.08, p-value = 3.8 × 10-2, respectively), while the PRS for TSH is inversely associated to BC risk (OR = 0.93, 95%CI: 0.89-0.97, p-value = 2.0 × 10-3). Using the PLACO method, we detected 49 loci associated to both BC and thyroid traits (p-value < 5 × 10-8), in the vicinity of 130 genes. An additional colocalization and gene-set enrichment analyses showed a convincing causal role for a known pleiotropic locus at 2q35 and revealed an additional one at 8q22.1 associated to both BC and thyroid cancer. We also found two new pleiotropic loci at 14q32.33 and 17q21.31 that were associated to both TSH levels and BC risk. Enrichment analyses and evidence of regulatory signals also highlighted brain tissues and immune system as candidates for obtaining associations between BC and TSH levels. Overall, our study sheds light on the complex interplay between BC and thyroid traits and provides evidence of shared genetic risk between those conditions., (© The Author(s) 2023. Published by Oxford University Press. All rights reserved. For Permissions, please email: journals.permissions@oup.com.)
- Published
- 2023
- Full Text
- View/download PDF
34. GCPBayes pipeline: a tool for exploring pleiotropy at the gene level.
- Author
-
Asgari Y, Sugier PE, Baghfalaki T, Lucotte E, Karimi M, Sedki M, Ngo A, Liquet B, and Truong T
- Abstract
Cross-phenotype association using gene-set analysis can help to detect pleiotropic genes and inform about common mechanisms between diseases. Although there are an increasing number of statistical methods for exploring pleiotropy, there is a lack of proper pipelines to apply gene-set analysis in this context and using genome-scale data in a reasonable running time. We designed a user-friendly pipeline to perform cross-phenotype gene-set analysis between two traits using GCPBayes, a method developed by our team. All analyses could be performed automatically by calling for different scripts in a simple way (using a Shiny app, Bash or R script). A Shiny application was also developed to create different plots to visualize outputs from GCPBayes. Finally, a comprehensive and step-by-step tutorial on how to use the pipeline is provided in our group's GitHub page. We illustrated the application on publicly available GWAS (genome-wide association studies) summary statistics data to identify breast cancer and ovarian cancer susceptibility genes. We have shown that the GCPBayes pipeline could extract pleiotropic genes previously mentioned in the literature, while it also provided new pleiotropic genes and regions that are worthwhile for further investigation. We have also provided some recommendations about parameter selection for decreasing computational time of GCPBayes on genome-scale data., (© The Author(s) 2023. Published by Oxford University Press on behalf of NAR Genomics and Bioinformatics.)
- Published
- 2023
- Full Text
- View/download PDF
35. Understanding links between water-quality variables and nitrate concentration in freshwater streams using high frequency sensor data.
- Author
-
Kermorvant C, Liquet B, Litt G, Mengersen K, Peterson EE, Hyndman RJ, Jones JB Jr, and Leigh C
- Subjects
- Fresh Water, Water, Data Accuracy, Nitrates, Rivers
- Abstract
Real-time monitoring using in-situ sensors is becoming a common approach for measuring water-quality within watersheds. High-frequency measurements produce big datasets that present opportunities to conduct new analyses for improved understanding of water-quality dynamics and more effective management of rivers and streams. Of primary importance is enhancing knowledge of the relationships between nitrate, one of the most reactive forms of inorganic nitrogen in the aquatic environment, and other water-quality variables. We analysed high-frequency water-quality data from in-situ sensors deployed in three sites from different watersheds and climate zones within the National Ecological Observatory Network, USA. We used generalised additive mixed models to explain the nonlinear relationships at each site between nitrate concentration and conductivity, turbidity, dissolved oxygen, water temperature, and elevation. Temporal auto-correlation was modelled with an auto-regressive-moving-average (ARIMA) model and we examined the relative importance of the explanatory variables. Total deviance explained by the models was high for all sites (99%). Although variable importance and the smooth regression parameters differed among sites, the models explaining the most variation in nitrate contained the same explanatory variables. This study demonstrates that building a model for nitrate using the same set of explanatory water-quality variables is achievable, even for sites with vastly different environmental and climatic characteristics. Applying such models will assist managers to select cost-effective water-quality variables to monitor when the goals are to gain a spatial and temporal in-depth understanding of nitrate dynamics and adapt management plans accordingly., Competing Interests: The authors have declared that no competing interests exist., (Copyright: © 2023 Kermorvant et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.)
- Published
- 2023
- Full Text
- View/download PDF
36. SMOTE-CD: SMOTE for compositional data.
- Author
-
Nguyen T, Mengersen K, Sous D, and Liquet B
- Subjects
- Entropy, Minority Groups, Neural Networks, Computer, Acclimatization
- Abstract
Compositional data are a special kind of data, represented as a proportion carrying relative information. Although this type of data is widely spread, no solution exists to deal with the cases where the classes are not well balanced. After describing compositional data imbalance, this paper proposes an adaptation of the original Synthetic Minority Oversampling TEchnique (SMOTE) to deal with compositional data imbalance. The new approach, called SMOTE for Compositional Data (SMOTE-CD), generates synthetic examples by computing a linear combination of selected existing data points, using compositional data operations. The performance of the SMOTE-CD is tested with three different regressors (Gradient Boosting tree, Neural Networks, Dirichlet regressor) applied to two real datasets and to synthetic generated data, and the performance is evaluated using accuracy, cross-entropy, F1-score, R2 score and RMSE. The results show improvements across all metrics, but the impact of oversampling on performance varies depending on the model and the data. In some cases, oversampling may lead to a decrease in performance for the majority class. However, for the real data, the best performance across all models is achieved when oversampling is used. Notably, the F1-score is consistently increased with oversampling. Unlike the original technique, the performance is not improved when combining oversampling of the minority classes and undersampling of the majority class. The Python package smote-cd implements the method and is available online., Competing Interests: The authors have declared that no competing interests exist., (Copyright: © 2023 Nguyen et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.)
- Published
- 2023
- Full Text
- View/download PDF
37. Leveraging pleiotropic association using sparse group variable selection in genomics data.
- Author
-
Sutton M, Sugier PE, Truong T, and Liquet B
- Subjects
- Algorithms, Humans, Phenotype, Polymorphism, Single Nucleotide, Genome-Wide Association Study, Genomics methods
- Abstract
Background: Genome-wide association studies (GWAS) have identified genetic variants associated with multiple complex diseases. We can leverage this phenomenon, known as pleiotropy, to integrate multiple data sources in a joint analysis. Often integrating additional information such as gene pathway knowledge can improve statistical efficiency and biological interpretation. In this article, we propose statistical methods which incorporate both gene pathway and pleiotropy knowledge to increase statistical power and identify important risk variants affecting multiple traits., Methods: We propose novel feature selection methods for the group variable selection in multi-task regression problem. We develop penalised likelihood methods exploiting different penalties to induce structured sparsity at a gene (or pathway) and SNP level across all studies. We implement an alternating direction method of multipliers (ADMM) algorithm for our penalised regression methods. The performance of our approaches are compared to a subset based meta analysis approach on simulated data sets. A bootstrap sampling strategy is provided to explore the stability of the penalised methods., Results: Our methods are applied to identify potential pleiotropy in an application considering the joint analysis of thyroid and breast cancers. The methods were able to detect eleven potential pleiotropic SNPs and six pathways. A simulation study found that our method was able to detect more true signals than a popular competing method while retaining a similar false discovery rate., Conclusion: We developed feature selection methods for jointly analysing multiple logistic regression tasks where prior grouping knowledge is available. Our method performed well on both simulation studies and when applied to a real data analysis of multiple cancers., (© 2021. The Author(s).)
- Published
- 2022
- Full Text
- View/download PDF
38. Author Correction: Community evaluation of glycoproteomics informatics solutions reveals high-performance search strategies for serum glycopeptide analysis.
- Author
-
Kawahara R, Chernykh A, Alagesan K, Bern M, Cao W, Chalkley RJ, Cheng K, Choo MS, Edwards N, Goldman R, Hoffmann M, Hu Y, Huang Y, Kim JY, Kletter D, Liquet B, Liu M, Mechref Y, Meng B, Neelamegham S, Nguyen-Khuong T, Nilsson J, Pap A, Park GW, Parker BL, Pegg CL, Penninger JM, Phung TK, Pioch M, Rapp E, Sakalli E, Sanda M, Schulz BL, Scott NE, Sofronov G, Stadlmann J, Vakhrushev SY, Woo CM, Wu HY, Yang P, Ying W, Zhang H, Zhang Y, Zhao J, Zaia J, Haslam SM, Palmisano G, Yoo JS, Larson G, Khoo KH, Medzihradszky KF, Kolarich D, Packer NH, and Thaysen-Andersen M
- Published
- 2022
- Full Text
- View/download PDF
39. Reconstructing Missing and Anomalous Data Collected from High-Frequency In-Situ Sensors in Fresh Waters.
- Author
-
Kermorvant C, Liquet B, Litt G, Jones JB, Mengersen K, Peterson EE, Hyndman RJ, and Leigh C
- Subjects
- Environmental Monitoring, Fresh Water, Nitrogen Oxides, Rivers, Ecosystem, Water Quality
- Abstract
In situ sensors that collect high-frequency data are used increasingly to monitor aquatic environments. These sensors are prone to technical errors, resulting in unrecorded observations and/or anomalous values that are subsequently removed and create gaps in time series data. We present a framework based on generalized additive and auto-regressive models to recover these missing data. To mimic sporadically missing (i) single observations and (ii) periods of contiguous observations, we randomly removed (i) point data and (ii) day- and week-long sequences of data from a two-year time series of nitrate concentration data collected from Arikaree River, USA, where synoptically collected water temperature, turbidity, conductance, elevation, and dissolved oxygen data were available. In 72% of cases with missing point data, predicted values were within the sensor precision interval of the original value, although predictive ability declined when sequences of missing data occurred. Precision also depended on the availability of other water quality covariates. When covariates were available, even a sudden, event-based peak in nitrate concentration was reconstructed well. By providing a promising method for accurate prediction of missing data, the utility and confidence in summary statistics and statistical trends will increase, thereby assisting the effective monitoring and management of fresh waters and other at-risk ecosystems.
- Published
- 2021
- Full Text
- View/download PDF
40. Community evaluation of glycoproteomics informatics solutions reveals high-performance search strategies for serum glycopeptide analysis.
- Author
-
Kawahara R, Chernykh A, Alagesan K, Bern M, Cao W, Chalkley RJ, Cheng K, Choo MS, Edwards N, Goldman R, Hoffmann M, Hu Y, Huang Y, Kim JY, Kletter D, Liquet B, Liu M, Mechref Y, Meng B, Neelamegham S, Nguyen-Khuong T, Nilsson J, Pap A, Park GW, Parker BL, Pegg CL, Penninger JM, Phung TK, Pioch M, Rapp E, Sakalli E, Sanda M, Schulz BL, Scott NE, Sofronov G, Stadlmann J, Vakhrushev SY, Woo CM, Wu HY, Yang P, Ying W, Zhang H, Zhang Y, Zhao J, Zaia J, Haslam SM, Palmisano G, Yoo JS, Larson G, Khoo KH, Medzihradszky KF, Kolarich D, Packer NH, and Thaysen-Andersen M
- Subjects
- Glycosylation, Humans, Proteome metabolism, Tandem Mass Spectrometry, Glycopeptides blood, Glycoproteins blood, Informatics methods, Proteome analysis, Proteomics methods, Research Personnel statistics & numerical data, Software
- Abstract
Glycoproteomics is a powerful yet analytically challenging research tool. Software packages aiding the interpretation of complex glycopeptide tandem mass spectra have appeared, but their relative performance remains untested. Conducted through the HUPO Human Glycoproteomics Initiative, this community study, comprising both developers and users of glycoproteomics software, evaluates solutions for system-wide glycopeptide analysis. The same mass spectrometrybased glycoproteomics datasets from human serum were shared with participants and the relative team performance for N- and O-glycopeptide data analysis was comprehensively established by orthogonal performance tests. Although the results were variable, several high-performance glycoproteomics informatics strategies were identified. Deep analysis of the data revealed key performance-associated search parameters and led to recommendations for improved 'high-coverage' and 'high-accuracy' glycoproteomics search solutions. This study concludes that diverse software packages for comprehensive glycopeptide data analysis exist, points to several high-performance search strategies and specifies key variables that will guide future software developments and assist informatics decision-making in glycoproteomics., (© 2021. The Author(s).)
- Published
- 2021
- Full Text
- View/download PDF
41. An appraisal of respiratory system compliance in mechanically ventilated covid-19 patients.
- Author
-
Li Bassi G, Suen JY, Dalton HJ, White N, Shrapnel S, Fanning JP, Liquet B, Hinton S, Vuorinen A, Booth G, Millar JE, Forsyth S, Panigada M, Laffey J, Brodie D, Fan E, Torres A, Chiumello D, Corley A, Elhazmi A, Hodgson C, Ichiba S, Luna C, Murthy S, Nichol A, Ng PY, Ogino M, Pesenti A, Trieu HT, and Fraser JF
- Subjects
- Adult, Cohort Studies, Critical Care methods, Europe, Female, Humans, Intensive Care Units, Male, Middle Aged, Retrospective Studies, Severity of Illness Index, COVID-19 complications, COVID-19 therapy, Lung Compliance physiology, Respiration, Artificial methods, Respiratory Distress Syndrome etiology, Respiratory Distress Syndrome therapy
- Abstract
Background: Heterogeneous respiratory system static compliance (C
RS ) values and levels of hypoxemia in patients with novel coronavirus disease (COVID-19) requiring mechanical ventilation have been reported in previous small-case series or studies conducted at a national level., Methods: We designed a retrospective observational cohort study with rapid data gathering from the international COVID-19 Critical Care Consortium study to comprehensively describe CRS -calculated as: tidal volume/[airway plateau pressure-positive end-expiratory pressure (PEEP)]-and its association with ventilatory management and outcomes of COVID-19 patients on mechanical ventilation (MV), admitted to intensive care units (ICU) worldwide., Results: We studied 745 patients from 22 countries, who required admission to the ICU and MV from January 14 to December 31, 2020, and presented at least one value of CRS within the first seven days of MV. Median (IQR) age was 62 (52-71), patients were predominantly males (68%) and from Europe/North and South America (88%). CRS , within 48 h from endotracheal intubation, was available in 649 patients and was neither associated with the duration from onset of symptoms to commencement of MV (p = 0.417) nor with PaO2 /FiO2 (p = 0.100). Females presented lower CRS than males (95% CI of CRS difference between females-males: - 11.8 to - 7.4 mL/cmH2 O p < 0.001), and although females presented higher body mass index (BMI), association of BMI with CRS was marginal (p = 0.139). Ventilatory management varied across CRS range, resulting in a significant association between CRS and driving pressure (estimated decrease - 0.31 cmH2 O/L per mL/cmH2 0 of CRS , 95% CI - 0.48 to - 0.14, p < 0.001). Overall, 28-day ICU mortality, accounting for the competing risk of being discharged within the period, was 35.6% (SE 1.7). Cox proportional hazard analysis demonstrated that CRS (+ 10 mL/cm H2 O) was only associated with being discharge from the ICU within 28 days (HR 1.14, 95% CI 1.02-1.28, p = 0.018)., Conclusions: This multicentre report provides a comprehensive account of CRS in COVID-19 patients on MV. CRS measured within 48 h from commencement of MV has marginal predictive value for 28-day mortality, but was associated with being discharged from ICU within the same period. Trial documentation: Available at https://www.covid-critical.com/study ., Trial Registration: ACTRN12620000421932.- Published
- 2021
- Full Text
- View/download PDF
42. Bayesian meta-analysis models for cross cancer genomic investigation of pleiotropic effects using group structure.
- Author
-
Baghfalaki T, Sugier PE, Truong T, Pettitt AN, Mengersen K, and Liquet B
- Subjects
- Bayes Theorem, Genomics, Group Structure, Humans, Models, Genetic, Polymorphism, Single Nucleotide, Genome-Wide Association Study, Neoplasms
- Abstract
An increasing number of genome-wide association studies (GWAS) summary statistics is made available to the scientific community. Exploiting these results from multiple phenotypes would permit identification of novel pleiotropic associations. In addition, incorporating prior biological information in GWAS such as group structure information (gene or pathway) has shown some success in classical GWAS approaches. However, this has not been widely explored in the context of pleiotropy. We propose a Bayesian meta-analysis approach (termed GCPBayes) that uses summary-level GWAS data across multiple phenotypes to detect pleiotropy at both group-level (gene or pathway) and within group (eg, at the SNP level). We consider both continuous and Dirac spike and slab priors for group selection. We also use a Bayesian sparse group selection approach with hierarchical spike and slab priors that enables us to select important variables both at the group level and within group. GCPBayes uses a Bayesian statistical framework based on Markov chain Monte Carlo (MCMC) Gibbs sampling. It can be applied to multiple types of phenotypes for studies with overlapping or nonoverlapping subjects, and takes into account heterogeneity in the effect size and allows for the opposite direction of the genetic effects across traits. Simulations show that the proposed methods outperform benchmark approaches such as ASSET and CPBayes in the ability to retrieve pleiotropic associations at both SNP and gene-levels. To illustrate the GCPBayes method, we investigate the shared genetic effects between thyroid cancer and breast cancer in candidate pathways., (© 2020 John Wiley & Sons, Ltd.)
- Published
- 2021
- Full Text
- View/download PDF
43. Penalized partial least squares for pleiotropy.
- Author
-
Broc C, Truong T, and Liquet B
- Subjects
- Least-Squares Analysis, Phenotype, Genome-Wide Association Study, Polymorphism, Single Nucleotide
- Abstract
Background: The increasing number of genome-wide association studies (GWAS) has revealed several loci that are associated to multiple distinct phenotypes, suggesting the existence of pleiotropic effects. Highlighting these cross-phenotype genetic associations could help to identify and understand common biological mechanisms underlying some diseases. Common approaches test the association between genetic variants and multiple traits at the SNP level. In this paper, we propose a novel gene- and a pathway-level approach in the case where several independent GWAS on independent traits are available. The method is based on a generalization of the sparse group Partial Least Squares (sgPLS) to take into account groups of variables, and a Lasso penalization that links all independent data sets. This method, called joint-sgPLS, is able to convincingly detect signal at the variable level and at the group level., Results: Our method has the advantage to propose a global readable model while coping with the architecture of data. It can outperform traditional methods and provides a wider insight in terms of a priori information. We compared the performance of the proposed method to other benchmark methods on simulated data and gave an example of application on real data with the aim to highlight common susceptibility variants to breast and thyroid cancers., Conclusion: The joint-sgPLS shows interesting properties for detecting a signal. As an extension of the PLS, the method is suited for data with a large number of variables. The choice of Lasso penalization copes with architectures of groups of variables and observations sets. Furthermore, although the method has been applied to a genetic study, its formulation is adapted to any data with high number of variables and an exposed a priori architecture in other application fields.
- Published
- 2021
- Full Text
- View/download PDF
44. Estimation of semi-Markov multi-state models: a comparison of the sojourn times and transition intensities approaches.
- Author
-
Asanjarani A, Liquet B, and Nazarathy Y
- Subjects
- Markov Chains, Probability, Survival Analysis, Reproducibility of Results
- Abstract
Semi-Markov models are widely used for survival analysis and reliability analysis. In general, there are two competing parameterizations and each entails its own interpretation and inference properties. On the one hand, a semi-Markov process can be defined based on the distribution of sojourn times, often via hazard rates, together with transition probabilities of an embedded Markov chain. On the other hand, intensity transition functions may be used, often referred to as the hazard rates of the semi-Markov process. We summarize and contrast these two parameterizations both from a probabilistic and an inference perspective, and we highlight relationships between the two approaches. In general, the intensity transition based approach allows the likelihood to be split into likelihoods of two-state models having fewer parameters, allowing efficient computation and usage of many survival analysis tools. Nevertheless, in certain cases the sojourn time based approach is natural and has been exploited extensively in applications. In contrasting the two approaches and contemporary relevant R packages used for inference, we use two real datasets highlighting the probabilistic and inference properties of each approach. This analysis is accompanied by an R vignette., (© 2020 Azam Asanjarani et al., published by De Gruyter, Berlin/Boston.)
- Published
- 2021
- Full Text
- View/download PDF
45. Classification algorithm for high-dimensional protein markers in time-course data.
- Author
-
Vishwakarma GK, Bhattacharjee A, Banerjee S, and Liquet B
- Subjects
- Bayes Theorem, Biomarkers, Humans, Proportional Hazards Models, Algorithms, Medical Oncology
- Abstract
Identification of biomarkers is an emerging area in oncology. In this article, we develop an efficient statistical procedure for the classification of protein markers according to their effect on cancer progression. A high-dimensional time-course dataset of protein markers for 80 patients motivates us for developing the model. The threshold value is formulated as a level of a marker having maximum impact on cancer progression. The classification algorithm technique for high-dimensional time-course data is developed and the algorithm is validated by comparing random components using both proportional hazard and accelerated failure time frailty models. The study elucidates the application of two separate joint modeling techniques using auto regressive-type model and mixed effect model for time-course data and proportional hazard model for survival data with proper utilization of Bayesian methodology. Also, a prognostic score is developed on the basis of few selected genes with application on patients. This study facilitates to identify relevant biomarkers from a set of markers., (© 2020 John Wiley & Sons Ltd.)
- Published
- 2020
- Full Text
- View/download PDF
46. Design and rationale of the COVID-19 Critical Care Consortium international, multicentre, observational study.
- Author
-
Li Bassi G, Suen J, Barnett AG, Corley A, Millar J, Fanning J, Lye I, Colombo S, Wildi K, Livingstone S, Abbate G, Hinton S, Liquet B, Shrapnel S, Dalton H, and Fraser JF
- Subjects
- Humans, Evidence-Based Medicine, Global Health, Observational Studies as Topic, Outcome Assessment, Health Care, Pandemics, Pragmatic Clinical Trials as Topic, SARS-CoV-2, Multicenter Studies as Topic, COVID-19 mortality, COVID-19 therapy, Intensive Care Units statistics & numerical data, Registries
- Abstract
Introduction: There is a paucity of data that can be used to guide the management of critically ill patients with COVID-19. In response, a research and data-sharing collaborative-The COVID-19 Critical Care Consortium-has been assembled to harness the cumulative experience of intensive care units (ICUs) worldwide. The resulting observational study provides a platform to rapidly disseminate detailed data and insights crucial to improving outcomes., Methods and Analysis: This is an international, multicentre, observational study of patients with confirmed or suspected SARS-CoV-2 infection admitted to ICUs. This is an evolving, open-ended study that commenced on 1 January 2020 and currently includes >350 sites in over 48 countries. The study enrols patients at the time of ICU admission and follows them to the time of death, hospital discharge or 28 days post-ICU admission, whichever occurs last. Key data, collected via an electronic case report form devised in collaboration with the International Severe Acute Respiratory and Emerging Infection Consortium/Short Period Incidence Study of Severe Acute Respiratory Illness networks, include: patient demographic data and risk factors, clinical features, severity of illness and respiratory failure, need for non-invasive and/or mechanical ventilation and/or extracorporeal membrane oxygenation and associated complications, as well as data on adjunctive therapies., Ethics and Dissemination: Local principal investigators will ensure that the study adheres to all relevant national regulations, and that the necessary approvals are in place before a site may contribute data. In jurisdictions where a waiver of consent is deemed insufficient, prospective, representative or retrospective consent will be obtained, as appropriate. A web-based dashboard has been developed to provide relevant data and descriptive statistics to international collaborators in real-time. It is anticipated that, following study completion, all de-identified data will be made open access., Trial Registration Number: ACTRN12620000421932 (http://anzctr.org.au/ACTRN12620000421932.aspx)., Competing Interests: Competing interests: GLB and JFF received research funds, through their affiliated institution, from Fisher & Paykel for studies related to high-flow oxygen therapy., (© Author(s) (or their employer(s)) 2020. Re-use permitted under CC BY-NC. No commercial re-use. See rights and permissions. Published by BMJ.)
- Published
- 2020
- Full Text
- View/download PDF
47. Detecting Technical Anomalies in High-Frequency Water-Quality Data Using Artificial Neural Networks.
- Author
-
Rodriguez-Perez J, Leigh C, Liquet B, Kermorvant C, Peterson E, Sous D, and Mengersen K
- Subjects
- Bayes Theorem, Water Quality, Neural Networks, Computer, Water
- Abstract
Anomaly detection (AD) in high-volume environmental data requires one to tackle a series of challenges associated with the typical low frequency of anomalous events, the broad-range of possible anomaly types, and local nonstationary environmental conditions, suggesting the need for flexible statistical methods that are able to cope with unbalanced high-volume data problems. Here, we aimed to detect anomalies caused by technical errors in water-quality (turbidity and conductivity) data collected by automated in situ sensors deployed in contrasting riverine and estuarine environments. We first applied a range of artificial neural networks that differed in both learning method and hyperparameter values, then calibrated models using a Bayesian multiobjective optimization procedure, and selected and evaluated the "best" model for each water-quality variable, environment, and anomaly type. We found that semi-supervised classification was better able to detect sudden spikes, sudden shifts, and small sudden spikes, whereas supervised classification had higher accuracy for predicting long-term anomalies associated with drifts and periods of otherwise unexplained high variability.
- Published
- 2020
- Full Text
- View/download PDF
48. Forecasting intensifying disturbance effects on coral reefs.
- Author
-
Vercelloni J, Liquet B, Kennedy EV, González-Rivero M, Caley MJ, Peterson EE, Puotinen M, Hoegh-Guldberg O, and Mengersen K
- Subjects
- Animals, Coral Reefs, Ecosystem, Anthozoa, Cyclonic Storms
- Abstract
Anticipating future changes of an ecosystem's dynamics requires knowledge of how its key communities respond to current environmental regimes. The Great Barrier Reef (GBR) is under threat, with rapid changes of its reef-building hard coral (HC) community structure already evident across broad spatial scales. While several underlying relationships between HC and multiple disturbances have been documented, responses of other benthic communities to disturbances are not well understood. Here we used statistical modelling to explore the effects of broad-scale climate-related disturbances on benthic communities to predict their structure under scenarios of increasing disturbance frequency. We parameterized a multivariate model using the composition of benthic communities estimated by 145,000 observations from the northern GBR between 2012 and 2017. During this time, surveyed reefs were variously impacted by two tropical cyclones and two heat stress events that resulted in extensive HC mortality. This unprecedented sequence of disturbances was used to estimate the effects of discrete versus interacting disturbances on the compositional structure of HC, soft corals (SC) and algae. Discrete disturbances increased the prevalence of algae relative to HC while the interaction between cyclones and heat stress was the main driver of the increase in SC relative to algae and HC. Predictions from disturbance scenarios included relative increases in algae versus SC that varied by the frequency and types of disturbance interactions. However, high uncertainty of compositional changes in the presence of several disturbances shows that responses of algae and SC to the decline in HC needs further research. Better understanding of the effects of multiple disturbances on benthic communities as a whole is essential for predicting the future status of coral reefs and managing them in the light of new environmental regimes. The approach we develop here opens new opportunities for reaching this goal., (© 2020 John Wiley & Sons Ltd.)
- Published
- 2020
- Full Text
- View/download PDF
49. CPMCGLM: an R package for p-value adjustment when looking for an optimal transformation of a single explanatory variable in generalized linear models.
- Author
-
Liquet B and Riou J
- Subjects
- Cholesterol, HDL blood, Dementia blood, Female, Humans, Male, Reproducibility of Results, Algorithms, Biometry methods, Computational Biology methods, Linear Models
- Abstract
Background: In medical research, explanatory continuous variables are frequently transformed or converted into categorical variables. If the coding is unknown, many tests can be used to identify the "optimal" transformation. This common process, involving the problems of multiple testing, requires a correction of the significance level. Liquet and Commenges proposed an asymptotic correction of significance level in the context of generalized linear models (GLM) (Liquet and Commenges, Stat Probab Lett 71:33-38, 2005). This procedure has been developed for dichotomous and Box-Cox transformations. Furthermore, Liquet and Riou suggested the use of resampling methods to estimate the significance level for transformations into categorical variables with more than two levels (Liquet and Riou, BMC Med Res Methodol 13:75, 2013)., Results: CPMCGLM provides to users both methods of p-value adjustment. Futhermore, they are available for a large set of transformations. This paper aims to provide insight the user an overview of the methodological context, and explain in detail the use of the CPMCGLM R package through its application to a real epidemiological dataset., Conclusion: We present here the CPMCGLMR package providing efficient methods for the correction of type-I error rate in the context of generalized linear models. This is the first and the only available package in R providing such methods applied to this context. This package is designed to help researchers, who work principally in the field of biostatistics and epidemiology, to analyze their data in the context of optimal cutoff point determination.
- Published
- 2019
- Full Text
- View/download PDF
50. Age at menarche and the risk of operative delivery.
- Author
-
Chong HP, Frøen JF, Richardson S, Liquet B, Charnock-Jones DS, and Smith GCS
- Subjects
- Adult, Age Factors, Cesarean Section statistics & numerical data, Extraction, Obstetrical statistics & numerical data, Female, Humans, Infant, Newborn, Male, Norway epidemiology, Obstetrical Forceps, Pregnancy, Risk Factors, Term Birth, Vacuum Extraction, Obstetrical statistics & numerical data, Young Adult, Delivery, Obstetric methods, Delivery, Obstetric statistics & numerical data, Menarche physiology, Obstetric Labor Complications epidemiology, Obstetric Labor Complications surgery
- Abstract
Objectives: We sought to evaluate the impact of later menarche on the risk of operative delivery., Population: We studied 38,069 eligible women (first labors at term with a singleton infant in a cephalic presentation) from the Norwegian Mothers and Child Cohort Study. The main exposures were the age at menarche and the duration of the interval between menarche and the first birth., Methods: Poisson's regression with a robust variance estimator., Main Outcome Measures: Operative delivery, defined as emergency cesarean or assisted vaginal delivery (ventouse extraction or forceps)., Results: A 5 year increase in age at menarche was associated with a reduced risk of operative delivery (risk ratio [RR] 0.84, 95%CI 0.78, 0.89; p < .001). Adjustment for the age at first birth slightly strengthened the association (RR 0.79, 95%CI 0.74, 0.84; p < .001). However, the association was lost following adjustment for the menarche to birth interval (RR 0.99, 95%CI 0.93, 1.06; p = .81). A 5 years increase in menarche to birth interval was associated with an increased risk of operative delivery (RR 1.26, 95%CI 1.23, 1.28; p < .001). This was not materially affected by adjustment for an extensive series of maternal characteristics (RR 1.23, 95%CI 1.20, 1.25; p < .001)., Conclusions: Later menarche reduces the risk of an operative first birth through shortening the menarche to birth interval. This observation is consistent with the hypothesis that the pattern and/or duration of prepregnancy exposure of the uterus to estrogen and progesterone contributes to uterine aging.
- Published
- 2019
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.