14 results on '"Koen Ilja Neijenhuijs"'
Search Results
2. Cross-cultural adaptation and validation of the Dutch language version of the Pictorial Fear of Activity Scale - Cervical
- Author
-
Maaike Kragting, Koen Ilja Neijenhuijs, Lennard Voogt, Annelies Pool-Goudzwaard, Michel W. Coppieters, AMS - Rehabilitation & Development, Neuromechanics, APH - Mental Health, Clinical Psychology, AMS - Musculoskeletal Health, and Physiotherapy, Human Physiology and Anatomy
- Subjects
Cross-Cultural Comparison ,lcsh:Diseases of the musculoskeletal system ,SDG 16 - Peace ,Psychometrics ,artikel tijdschrift ,Validity ,03 medical and health sciences ,0302 clinical medicine ,Rheumatology ,Cronbach's alpha ,Surveys and Questionnaires ,Humans ,Medicine ,Orthopedics and Sports Medicine ,030212 general & internal medicine ,Musculoskeletal health ,Whiplash Injuries ,Language ,Face validity ,Neck pain ,business.industry ,SDG 16 - Peace, Justice and Strong Institutions ,Discriminant validity ,Reproducibility of Results ,Construct validity ,Fear ,Reliability ,Confirmatory factor analysis ,Justice and Strong Institutions ,Psychometric properties ,Convergent validity ,Scale (social sciences) ,lcsh:RC925-935 ,medicine.symptom ,business ,Fear of movement ,030217 neurology & neurosurgery ,Research Article ,Clinical psychology - Abstract
Background The Pictorial Fear of Activity Scale-Cervical (PFActS-C) is a reliable and valid instrument to assess fear of movement in people with whiplash associated disorders. It is not available in Dutch and has not been evaluated in other neck pain populations. This study aimed to systematically translate the PFActS-C into Dutch and evaluate the psychometric properties of this Dutch Language Version (DLV) in people with non-specific neck pain. Methods The PFActS-C was translated according to international guidelines. Internal consistency, test-retest reliability, floor and ceiling effects, face validity and construct validity (convergent and discriminant validity by hypotheses testing and structural validity by confirmatory and exploratory factor-analyses) of the PFActS-C-DLV were tested in 125 people with non-specific neck pain. Results The PFActS-C-DLV showed good to excellent internal consistency (Cronbach’s alpha: 0.98) and stability over time (ICC: 0.90 [95%CI: 0.82–0.93). Four out of five a priori formulated hypotheses regarding related (convergent validity) and unrelated (discriminant validity) constructs were confirmed. However, the confirmatory factor analysis could not confirm the expected 1-factor solution. Furthermore, the exploratory factor analyses revealed that also a higher factor solution would not lead to a good fit of the model. Conclusions The PFActS-C-DLV is a reliable region-specific instrument for people with non-specific neck pain. The construct validity was supported, based on hypotheses testing. However, factor analyses could not confirm a 1-factor solution, so the underlying construct of the PFActS-C-DLV remains unclear. Given the PFActS-C’s photographic format, we believe these findings also have relevance for the original English version.
- Published
- 2020
3. Symptom clusters among cancer survivors: what can machine learning techniques tell us?
- Author
-
Carel F. W. Peeters, Koen Ilja Neijenhuijs, Irma M. Verdonck-de Leeuw, Pim Cuijpers, Henk van Weert, Clinical Psychology, APH - Mental Health, World Health Organization (WHO) Collaborating Center, APH - Global Health, Epidemiology and Data Science, CCA - Cancer Treatment and quality of life, APH - Personalized Medicine, Otolaryngology / Head & Neck Surgery, General practice, ACS - Heart failure & arrhythmias, and APH - Quality of Care
- Subjects
Medicine (General) ,Symptom clusters ,Epidemiology ,Health Informatics ,Prom ,Machine learning ,computer.software_genre ,Wiskundige en Statistische Methoden - Biometris ,R5-920 ,Quality of life ,Cancer Survivors ,SDG 3 - Good Health and Well-being ,Neoplasms ,medicine ,Humans ,Cluster analysis ,Mathematical and Statistical Methods - Biometris ,Cancer ,Framingham Risk Score ,business.industry ,Self-Management ,Research ,Targeted interventions ,Syndrome ,medicine.disease ,Oncology ,Quality of Life ,Patient-reported outcome ,Artificial intelligence ,business ,Psychosocial ,computer - Abstract
Purpose Knowledge regarding symptom clusters may inform targeted interventions. The current study investigated symptom clusters among cancer survivors, using machine learning techniques on a large data set. Methods Data consisted of self-reports of cancer survivors who used a fully automated online application ‘Oncokompas’ that supports them in their self-management. This is done by 1) monitoring their symptoms through patient reported outcome measures (PROMs); and 2) providing a personalized overview of supportive care options tailored to their scores, aiming to reduce symptom burden and improve health-related quality of life. In the present study, data on 26 generic symptoms (physical and psychosocial) were used. Results of the PROM of each symptom are presented to the user as a no well-being risk, moderate well-being risk, or high well-being risk score. Data of 1032 cancer survivors were analysed using Hierarchical Density-Based Spatial Clustering of Applications with Noise (HDBSCAN) on high risk scores and moderate-to-high risk scores separately. Results When analyzing the high risk scores, seven clusters were extracted: one main cluster which contained most frequently occurring physical and psychosocial symptoms, and six subclusters with different combinations of these symptoms. When analyzing moderate-to-high risk scores, three clusters were extracted: two main clusters were identified, which separated physical symptoms (and their consequences) and psycho-social symptoms, and one subcluster with only body weight issues. Conclusion There appears to be an inherent difference on the co-occurrence of symptoms dependent on symptom severity. Among survivors with high risk scores, the data showed a clustering of more connections between physical and psycho-social symptoms in separate subclusters. Among survivors with moderate-to-high risk scores, we observed less connections in the clustering between physical and psycho-social symptoms.
- Published
- 2021
4. The Female Sexual Function Index (FSFI)-A Systematic Review of Measurement Properties
- Author
-
Neil K. Aaronson, Karen Holtmaat, Irma M. Verdonck-de Leeuw, Mogens Groenvold, Koen Ilja Neijenhuijs, Caroline B. Terwee, Nienke Hooghiemstra, Pim Cuijpers, and Bernhard Holzner
- Subjects
Psychometrics ,Patient Reported Outcome Measure ,Urology ,Endocrinology, Diabetes and Metabolism ,media_common.quotation_subject ,Female sexual dysfunction ,030232 urology & nephrology ,MEDLINE ,Sample (statistics) ,Validity ,03 medical and health sciences ,0302 clinical medicine ,Endocrinology ,Female Sexual Function Index ,Outcome Assessment, Health Care ,Criterion validity ,medicine ,Humans ,Quality (business) ,Measurement invariance ,Patient Reported Outcome Measures ,Sexual Dysfunctions, Psychological ,Reliability (statistics) ,media_common ,030219 obstetrics & reproductive medicine ,SDG 5 - Gender Equality ,Reproducibility of Results ,Construct validity ,medicine.disease ,Reliability ,Psychiatry and Mental health ,Reproductive Medicine ,Female Sexual Dysfunction ,Female ,Psychology ,Clinical psychology - Abstract
Introduction The Female Sexual Function Index (FSFI) is a patient-reported outcome measure measuring female sexual dysfunction. The FSFI-19 was developed with 6 theoretical subscales in 2000. In 2010, a shortened version became available (FSFI-6). Aim To investigate the measurement properties of the FSFI-19 and FSFI-6. Methods A systematic search was performed of Embase, Medline, and Web of Science for studies that investigated measurement properties of the FSFI-19 or FSFI-6 up to April 2018. Data were extracted and analyzed according to COnsensus-based Standards for the selection of health Measurement INstruments (COSMIN) guidelines. Evidence was categorized into sufficient, insufficient, inconsistent, or indeterminate, and quality of evidence as very high, high, moderate, or low. Main Outcome Measures The Main Outcome Measure is the evidence of a measurement property, and the quality of evidence based on the COSMIN guidelines. Results 83 studies were included. Concerning the FSFI-19, the evidence for internal consistency was sufficient and of moderate quality. The evidence for reliability was sufficient but of low quality. The evidence for criterion validity was sufficient and of high quality. The evidence for structural validity was inconsistent of low quality. The evidence for construct validity was inconsistent of moderate quality. Concerning the FSFI-6, the evidence for criterion validity was sufficient of moderate quality. The evidence for internal consistency was rated as indeterminate. The evidence for reliability was inconsistent of low quality. The evidence for construct validity was inconsistent of very low quality. No information was available on structural validity of the FSFI-6, and measurement error, responsiveness, and cross-cultural validity of both FSFI-6 and FSFI-19. Clinical Implications Conflicting and lack of evidence for some of the measurement properties of the FSFI-19 and FSFI-6 indicates the importance of further research on the validity of these patient-reported outcome measures. We advise researchers who use the FSFI-19 to perform confirmatory factor analyses and report the factor structure found in their sample. Regardless of these concerns, the FSFI-19 and FSFI-6 have strong criterion validity. Pragmatically, they are good screening tools for the current definition of female sexual dysfunction. Strength & Limitation A strong point of the review is the use of predefined guidelines. A limitation is the use of a precise rather than a sensitive search filter. Conclusions The FSFI requires more research on structural validity (FSFI-19 and FSFI-6), reliability (FSFI-6), construct validity (FSFI-19), measurement error (FSFI-19 and FSFI-6), and responsiveness (FSFI-19 and FSFI-6). Further corroboration of measurement invariance (both across cultures and across subpopulations) in the factor structure of the FSFI-19 is necessary, as well as tests for the unidimensionality of the FSFI-6. Neijenhuijs KI, Hooghiemstra N, Holtmaat K, et al. The Female Sexual Function Index (FSFI)—A Systematic Review of Measurement Properties. J Sex Med 2019;16:640–660.
- Published
- 2019
5. A systematic review of the measurement properties of the Body Image Scale (BIS) in cancer patients
- Author
-
Cornelia F. van Uden-Kraan, Heleen C. Melissant, Mogens Groenvold, Koen Ilja Neijenhuijs, Femke Jansen, Caroline B. Terwee, Pim Cuijpers, Bernhard Holzner, Irma M. Verdonck-de Leeuw, and Neil K. Aaronson
- Subjects
medicine.medical_specialty ,MEDLINE ,Prom ,PsycINFO ,Review Article ,03 medical and health sciences ,0302 clinical medicine ,Quality of life ,SDG 3 - Good Health and Well-being ,Neoplasms ,Medicine ,Humans ,Medical physics ,030212 general & internal medicine ,Patient Reported Outcome Measures ,Medical diagnosis ,Reliability (statistics) ,PROM ,Statistical hypothesis testing ,Cancer ,business.industry ,Construct validity ,Body Dysmorphic Disorders ,Body image ,Oncology ,030220 oncology & carcinogenesis ,Systematic review ,Quality of Life ,business ,Body Image Scale ,Measurement properties - Abstract
Introduction Body image is acknowledged as an important aspect of health-related quality of life in cancer patients. The Body Image Scale (BIS) is a patient-reported outcome measure (PROM) to evaluate body image in cancer patients. The aim of this study was to systematically review measurement properties of the BIS among cancer patients. Methods A search in Embase, MEDLINE, PsycINFO, and Web of Science was performed to identify studies that investigated measurement properties of the BIS (Prospero ID 42017057237). Study quality was assessed (excellent, good, fair, poor), and data were extracted and analyzed according to the COnsensus-based Standards for the selection of health Measurement INstruments (COSMIN) methodology on structural validity, internal consistency, reliability, measurement error, hypothesis testing for construct validity, and responsiveness. Evidence was categorized into sufficient, insufficient, inconsistent, or indeterminate. Results Nine studies were included. Evidence was sufficient for structural validity (one factor solution), internal consistency (α = 0.86–0.96), and reliability (r > 0.70); indeterminate for measurement error (information on minimal important change lacked) and responsiveness (increasing body image disturbance in only one study); and inconsistent for hypothesis testing (conflicting results). Quality of the evidence was moderate to low. No studies reported on cross-cultural validity. Conclusion The BIS is a PROM with good structural validity, internal consistency, and test-retest reliability, but good quality studies on the other measurement properties are needed to optimize evidence. It is recommended to include a wider variety of cancer diagnoses and treatment modalities in these future studies. Electronic supplementary material The online version of this article (10.1007/s00520-018-4145-x) contains supplementary material, which is available to authorized users.
- Published
- 2018
6. Measuring health-related quality of life in colorectal cancer patients: systematic review of measurement properties of the EORTC QLQ-CR29
- Author
-
Neil K. Aaronson, Pim Cuijpers, Femke Jansen, Mogens Groenvold, Koen Ilja Neijenhuijs, Lonneke V. van de Poll-Franse, Caroline B. Terwee, Anja van der Hout, Cornelia F. van Uden-Kraan, Bernhard Holzner, Irma M. Verdonck-de Leeuw, Klinische Psychologie (Psychologie, FMG), and Medical and Clinical Psychology
- Subjects
Quality of Life/psychology ,Psychometrics ,Colorectal cancer ,IMPACT ,Health-related quality of life ,MEDLINE ,EUROPEAN-ORGANIZATION ,QUESTIONNAIRE ,Review Article ,PsycINFO ,VALIDATION ,03 medical and health sciences ,0302 clinical medicine ,SDG 3 - Good Health and Well-being ,Quality of life ,EORTC QLQ-CR29 ,Surveys and Questionnaires ,medicine ,RESECTION SYNDROME SCORE ,Humans ,COSMIN ,030212 general & internal medicine ,EHEALTH APPLICATION ,Measurement property ,Reliability (statistics) ,Statistical hypothesis testing ,OUTCOMES ,Patient-reported outcome measure (PROM) ,business.industry ,Nursing research ,Construct validity ,Reproducibility of Results ,medicine.disease ,3. Good health ,Oncology ,030220 oncology & carcinogenesis ,Quality of Life ,Psychometrics/methods ,Systematic review ,CHINESE VERSION ,Colorectal Neoplasms ,business ,METHODOLOGICAL QUALITY ,Clinical psychology ,Colorectal Neoplasms/psychology - Abstract
Introduction The EORTC QLQ-CR29 is a patient-reported outcome measure to evaluate health-related quality of life among colorectal cancer patients in research and clinical practice. The aim of this systematic review was to investigate whether the initial positive results regarding the measurement properties of the QLQ-CR29 are confirmed in subsequent studies. Methods A systematic search of Embase, Medline, PsycINFO, and Web of Science was conducted to identify studies investigating the measurement properties of the QLQ-CR29 published up to January 2019. For the 11 included studies, data were extracted, methodological quality was assessed, results were synthesized, and evidence was graded according to the COnsensus-based Standards for the selection of health Measurement INstruments (COSMIN) methodology on the measurement properties: structural validity, internal consistency, reliability, measurement error, construct validity (hypothesis testing, including known-group comparison, convergent and divergent validity), cross-cultural validity, and responsiveness. Results Internal consistency was rated as “sufficient,” with low evidence. Reliability was rated as “insufficient,” with moderate evidence. Construct validity (hypothesis testing; known-group comparison, convergent and divergent validity) was rated as “inconsistent,” with moderate evidence. Structural validity, measurement error, and responsiveness were rated as “indeterminate” and could therefore not be graded. Conclusion This review indicates that current evidence supporting the measurement properties of the QLQ-CR29 is limited. Additionally, better quality research is needed, taking into account the COSMIN methodology. Electronic supplementary material The online version of this article (10.1007/s00520-019-04764-7) contains supplementary material, which is available to authorized users.
- Published
- 2019
7. The International Index of Erectile Function (IIEF)—A Systematic Review of Measurement Properties
- Author
-
Neil K. Aaronson, Bernhard Holzner, Koen Ilja Neijenhuijs, Irma M. Verdonck-de Leeuw, Caroline B. Terwee, Pim Cuijpers, and Karen Holtmaat
- Subjects
Male ,medicine.medical_specialty ,Index (economics) ,Psychometrics ,media_common.quotation_subject ,Urology ,Endocrinology, Diabetes and Metabolism ,030232 urology & nephrology ,MEDLINE ,Scientific literature ,Validity ,03 medical and health sciences ,0302 clinical medicine ,Endocrinology ,Erectile Dysfunction ,Surveys and Questionnaires ,Criterion validity ,Humans ,Medicine ,COSMIN ,Quality (business) ,International Index of Erectile Function ,Patient Reported Outcome Measures ,Reliability (statistics) ,Statistical hypothesis testing ,media_common ,030219 obstetrics & reproductive medicine ,business.industry ,Penile Erection ,Reproducibility of Results ,Construct validity ,Erectile function ,medicine.disease ,Reliability ,Measurement Properties ,Psychiatry and Mental health ,Erectile dysfunction ,Reproductive Medicine ,Physical therapy ,Psychology ,business ,Clinical psychology - Abstract
Introduction The International Index of Erectile Function (IIEF) is a patient-reported outcome measure to evaluate erectile dysfunction and other sexual problems in men. Aim To perform a systematic review of the measurement properties of the 15-item patient-reported outcome measure (IIEF-15) and the shortened 5-item version (IIEF-5). Methods A systematic search of scientific literature up to April 2018 was performed. Data were extracted and analyzed according to COnsensus-based Standards for the selection of health Measurement INstruments (COSMIN) guidelines for structural validity, internal consistency, reliability, measurement error, hypothesis testing for construct validity, and responsiveness. Evidence of measurement properties was categorized into sufficient, insufficient, inconsistent, or indeterminate, and quality of evidence as very high, high, moderate, or low. Results 40 studies were included. The evidence for criterion validity (of the Erectile Function subscale), and responsiveness of the IIEF-15 was sufficient (high quality), but inconsistent (moderate quality) for structural validity, internal consistency, construct validity, and test-retest reliability. Evidence for structural validity, test-retest reliability, construct validity, and criterion validity of the IIEF-5 was sufficient (moderate quality) but indeterminate for internal consistency, measurement error, and responsiveness. Clinical Implications Lack of evidence for and evidence not supporting some of the measurement properties of the IIEF-15 and IIEF-5 shows the importance of further research on the validity of these questionnaires in clinical research and clinical practice. Strengths & Limitations A strength of the current review is the use of predefined guidelines (COSMIN). A limitation of this review is the use of a precise rather than a sensitive search filter regarding measurement properties to identify studies to be included. Conclusion The IIEF requires more research on structural validity (IIEF-15), internal consistency (IIEF-15 and IIEF-5), construct validity (IIEF-15), measurement error (IIEF-15 and IIEF-5), and responsiveness (IIEF-5). The most pressing matter for future research is determining the unidimensionality of the IIEF-5 and the exact factor structure of the IIEF-15.
- Published
- 2019
8. Cost and Effectiveness of Blended Versus Standard Cognitive Behavioral Therapy for Outpatients With Depression in Routine Specialized Mental Health Care: Pilot Randomized Controlled Trial
- Author
-
Patricia van Oppen, Jenneke Elize Wiersma, Heleen Riper, Joran Lokkerbol, Filip Smit, Jeroen Ruwaard, Koen Ilja Neijenhuijs, Lisa Kooistra, Psychiatry, APH - Mental Health, Amsterdam Neuroscience - Compulsivity, Impulsivity & Attention, Epidemiology and Data Science, and APH - Global Health
- Subjects
Adult ,Male ,medicine.medical_specialty ,Cost effectiveness ,medicine.medical_treatment ,Cost-Benefit Analysis ,Health Informatics ,Pilot Projects ,behavioral disciplines and activities ,law.invention ,blended cognitive behavioral therapy ,03 medical and health sciences ,0302 clinical medicine ,Randomized controlled trial ,law ,Health care ,mental disorders ,Outpatients ,medicine ,Humans ,030212 general & internal medicine ,cost-effectiveness ,Depression (differential diagnoses) ,Original Paper ,Cognitive Behavioral Therapy ,business.industry ,Depression ,Absolute risk reduction ,030227 psychiatry ,Cognitive behavioral therapy ,Clinical trial ,Mental Health ,Treatment Outcome ,specialized mental health care ,randomized controlled trial ,Physical therapy ,Female ,Quality-Adjusted Life Years ,business ,Psychopathology - Abstract
Background Cognitive behavioral therapy (CBT) is an effective treatment, but access is often restricted due to costs and limited availability of trained therapists. Blending online and face-to-face CBT for depression might improve cost-effectiveness and treatment availability. Objective This pilot study aimed to examine the costs and effectiveness of blended CBT compared with standard CBT for depressed patients in specialized mental health care to guide further research and development of blended CBT. Methods Patients were randomly allocated to blended CBT (n=53) or standard CBT (n=49). Blended CBT consisted of 10 weekly face-to-face sessions and 9 Web-based sessions. Standard CBT consisted of 15 to 20 weekly face-to-face sessions. At baseline and 10, 20, and 30 weeks after start of treatment, self-assessed depression severity, quality-adjusted life-years (QALYs), and costs were measured. Clinicians, blinded to treatment allocation, assessed psychopathology at all time points. Data were analyzed using linear mixed models. Uncertainty intervals around cost and effect estimates were estimated with 5000 Monte Carlo simulations. Results Blended CBT treatment duration was mean 19.0 (SD 12.6) weeks versus mean 33.2 (SD 23.0) weeks in standard CBT (P Conclusions This pilot study shows that blended CBT might be a promising way to engage depressed patients in specialized mental health care. Compared with standard CBT, blended CBT was not considered cost-effective from a societal perspective but had an acceptable probability of being cost-effective from the health care provider perspective. Results should be carefully interpreted due to the small sample size. Further research in larger replication studies focused on optimizing the clinical effects of blended CBT and its budget impact is warranted. Trial Registration Netherlands Trial Register NTR4650; https://www.trialregister.nl/trial/4408 International Registered Report Identifier (IRRID) RR2-10.1186/s12888-014-0290-z
- Published
- 2019
9. Translation of the eHealth Impact Questionnaire for a Population of Dutch Electronic Health Users: Validation Study (Preprint)
- Author
-
Koen Ilja Neijenhuijs, Anja van der Hout, Evalien Veldhuijzen, Gwendolijne G M Scholten-Peeters, Cornelia F van Uden-Kraan, Pim Cuijpers, and Irma M Verdonck-de Leeuw
- Abstract
BACKGROUND The eHealth Impact Questionnaire (eHIQ) provides a standardized method to measure attitudes of electronic health (eHealth) users toward eHealth. It has previously been validated in a population of eHealth users in the United Kingdom and consists of 2 parts and 5 subscales. Part 1 measures attitudes toward eHealth in general and consists of the subscales attitudes towards online health information (5 items) and attitudes towards sharing health experiences online (6 items). Part 2 measures the attitude toward a particular eHealth application and consists of the subscales confidence and identification (9 items), information and presentation (8 items), and understand and motivation (9 items). OBJECTIVE This study aimed to translate and validate the eHIQ in a Dutch population of eHealth users. METHODS The eHIQ was translated and validated in accordance with the COnsensus-based Standards for the selection of health status Measurement INstruments criteria. The validation comprised 3 study samples, with a total of 1287 participants. Structural validity was assessed using confirmatory factor analyses and exploratory factor analyses (EFAs; all 3 samples). Internal consistency was assessed using hierarchical omega (all 3 samples). Test-retest reliability was assessed after 2 weeks, using 2-way intraclass correlation coefficients (sample 1). Measurement error was assessed by calculating the smallest detectable change (sample 1). Convergent and divergent validity were assessed using correlations with the remaining measures (all 3 samples). A graded response model was fit, and item information curves were plotted to describe the information provided by items across item trait levels (all 3 samples). RESULTS The original factor structure showed a bad fit in all 3 study samples. EFAs showed a good fit for a modified factor structure in the first study sample. This factor structure was subsequently tested in samples 2 and 3 and showed acceptable to good fits. Internal consistency, test-retest reliability, convergent validity, and divergent validity were acceptable to good for both the original as the modified factor structure, except for test-retest reliability of one of the original subscales and the 2 derivative subscales in the modified factor structure. The graded response model showed that some items underperformed in both the original and modified factor structure. CONCLUSIONS The Dutch version of the eHIQ (eHIQ-NL) shows a different factor structure compared with the original English version. Part 1 of the eHIQ-NL consists of 3 subscales: attitudes towards online health information (5 items), comfort with sharing health experiences online (3 items), and usefulness of sharing health experiences online (3 items). Part 2 of the eHIQ-NL consists of 3 subscales: motivation and confidence to act (10 items), information and presentation (13 items), and identification (3 items).
- Published
- 2019
10. Many Labs 2: Investigating Variation in Replicability Across Samples and Settings
- Author
-
Jesse Chandler, Charles R. Ebersole, Marek A. Vranka, Sharon Coen, David Sirlopú, David Torres, Eva E. Chen, Joshua C. Gandi, Anthony J. Nelson, Robbie C. M. van Aert, Eva G. T. Green, Jeroen Stouten, Elsie Ong, Cheryl Alyssa Welch, Anja Eller, Jen-Ho Chang, Catherine Verniers, Liane Young, Canay Doğulu, Joshua M. Tybur, Jon Grahe, Michelangelo Vianello, German Kurapov, Ingrid P. J. Voermans, Maaike J. de Bruijn, Karin C.H.J. Smolders, William E. Davis, Felix D. Schönbrodt, Stephanie Szeto, Katherine S. Corker, Paul G. Curran, Fernando Mena-pacheco, Julie A. Woodzicka, Leander De Schutter, Ángel Gómez, Elisa Maria Galliani, Mike Friedman, Huajian Cai, Agata Sobkow, Joy E. Losee, Kevin Durrheim, Manuela Thomae, Jennifer A. Coleman, Tanuka Ghoshal, Serdar Karabati, Vanessa Smith-Castro, Konrad Bocian, Hans IJzerman, Mark Aveyard, Janko Međedović, Marta Wrońska, Mark Verschoor, Ho Phi Huynh, Félix Neto, Joshua A. Hicks, Alexandra Vázquez, John E. Edlund, Anna van 't Veer, Zubairu K. Dagona, Thierry Devos, Oskar K. Sundfelt, Philipp Spachtholz, Tripat Gill, Maciej Sekerdej, Jennifer A. Joy-Gaba, Masanori Oikawa, Aaron Ocampo, Sean C. Murphy, Haruna Karick, Kathleen Schmidt, Goran Knežević, Ivan Grahek, Nicolas Kervyn, Grant Packard, Fred Hasselman, Byron G. Adams, Satia A. Marotta, Aaron L. Wichman, Olga Bialobrzeska, Manini Srivastava, Mark J. Brandt, Nikolette P. Lipsey, Carmel A. Levitan, Ewa Szumowska, Gábor Orosz, Michael J. Bernstein, Morgan Conway, Timo Gnambs, Koen Ilja Neijenhuijs, John M. Zelenski, Lysandra Podesta, Esther Maassen, Sinan Alper, Ilker Dalgar, Chris N. H. Street, Alexander K. Saeri, Anna Dalla Rosa, Reginald B. Adams, Ljiljana B. Lazarević, Marcel A.L.M. van Assen, Brian Collisson, Mayowa T. Babalola, Daniel Lakens, Cheryl L. Carmichael, Carrie Kovacs, Yarrow Dunham, Yoel Inbar, Patricio Saavedra, Natalia Frankowska, Austin Lee Nichols, Evans Dami Binan, Samuel Lincoln Bezerra Lins, Marije van der Hulst, Jamie Kurtz, Morgan J. Tear, Brad Pinter, Jeffrey R. Huntsinger, Åse Innes-Ker, Steffen R. Giessner, Kakul Hai, Boban Petrović, Elizabeth L. Haines, Francisco Ceric, Andriy Myachykov, Wendy L. Morris, Jordan Axt, Marieke de Vries, Lacy E. Krueger, Erika Salomon, Heather Barry Kappes, Anna Cabak Rédei, Armand Chatard, Matthew Haigh, Haruka Oikawa, Nick Neave, Monique Pollmann, Daniel R. Berry, Abraham M. Rutchick, Fiery Cushman, David C. Cicero, Brian A. Nosek, Ronaldo Pilati, Zeng Zhijia, Jakub Traczyk, Mihály Berkics, Michael P. Hall, Luis Diego Vega, Nerisa Dozo, Roberto González, Paul A. M. Van Lange, William Jiménez-Leal, Leigh Ann Vaughn, Winnee Cheong, Rolando Pérez-Sánchez, Rishtee Batra, Michael A. Smith, Andrew C. W. Tang, Štěpán Bahník, Małgorzata Osowiecka, Gabrielle Pogge, Lisa A. Williams, Alexander S. English, Robert Busching, Melissa-Sue John, Taciano L. Milfont, Fanny Cambier, Alejandro Vásquez Echeverría, Carolyn Finck, Richard A. Klein, Anna Kende, Norbert K. Tanzer, Victor N. Keller, Marie E. Heffernan, Neil A. Lewis, Jordan Theriault, Walter Sowden, Troy G. Steiner, Kristin Nicole Dukes, Adrienn Ujhelyi, Susan L. O'Donnell, Angela T. Maitner, Michael Wood, Jesse Graham, Winfrida Malingumu, Petr Houdek, Katarzyna Cantarero, Miguel-Ángel Freyre, Jeanine L. M. Skorinko, Roza G. Kamiloğlu, Robyn K. Mallett, Sociale Psychologie (Psychologie, FMG), Amsterdam Interdisciplinary Centre for Emotion (AICE, Psychology, FMG), Laboratoire Inter-universitaire de Psychologie : Personnalité, Cognition, Changement Social (LIP-PC2S ), Université Savoie Mont Blanc (USMB [Université de Savoie] [Université de Chambéry])-Université Grenoble Alpes [2016-2019] (UGA [2016-2019]), Florida International University [Miami] (FIU), Massachusetts Institute of Technology (MIT), Universidad Nacional de Educación a Distancia (UNED), EDF (EDF), Biozentrum [Basel, Suisse], University of Basel (Unibas), Tilburg University [Netherlands], Performance Engineering Laboratory [Dublin] (PEL), Dublin City University [Dublin] (DCU), Department of Rheumatology, Ghent University Hospital, Laboratoire de Psychologie Sociale et Cognitive (LAPSCO), Université Clermont Auvergne [2017-2020] (UCA [2017-2020])-Centre National de la Recherche Scientifique (CNRS), School of Information Technology [Kharagpur], Indian Institute of Technology Kharagpur (IIT Kharagpur), Human Technology Interaction, UCL - SSH/LouRIM - Louvain Research Institute in Management and Organizations, Tilburg University, Social Psychology, IBBA, A-LAB, Faculdade de Psicologia e de Ciências da Educação, Department of Social Psychology, Department of Methodology and Statistics, Department Communication and Cognition, Language, Communication and Cognition, Department of Business-Society Management, Department of Organisation and Personnel Management, Erasmus MC other, Research Methods and Techniques, and IJzerman, Hans
- Subjects
Sampling effects ,H Social Sciences (General) ,Culture ,Open materials ,Learning and Plasticity ,[SHS.PSY]Humanities and Social Sciences/Psychology ,Replication ,Situational effects ,050109 social psychology ,Context (language use) ,Social psychology ,Preregistered ,050105 experimental psychology ,[SHS.PSY] Humanities and Social Sciences/Psychology ,social psychology, cognitive psychology, replication, culture, individual differences, sampling effects, situational effects, meta-analysis, Registered Report, open data, open materials, preregistered ,[STAT.ML]Statistics [stat]/Machine Learning [stat.ML] ,Statistical significance ,Cognitive psychology ,Statistics ,Replication (statistics) ,Medicine and Health Sciences ,0501 psychology and cognitive sciences ,General Psychology ,ComputingMilieux_MISCELLANEOUS ,[STAT.ME] Statistics [stat]/Methodology [stat.ME] ,Behaviour Change and Well-being ,Data Science ,05 social sciences ,Open data ,[STAT.ML] Statistics [stat]/Machine Learning [stat.ML] ,C800 ,Meta-analysis ,Registered Report ,Individual differences ,Developmental Psychopathology ,[STAT.ME]Statistics [stat]/Methodology [stat.ME] - Abstract
We conducted preregistered replications of 28 classic and contemporary published findings, with protocols that were peer reviewed in advance, to examine variation in effect magnitudes across samples and settings. Each protocol was administered to approximately half of 125 samples that comprised 15,305 participants from 36 countries and territories. Using the conventional criterion of statistical significance (p < .05), we found that 15 (54%) of the replications provided evidence of a statistically significant effect in the same direction as the original finding. With a strict significance criterion (p < .0001), 14 (50%) of the replications still provided such evidence, a reflection of the extremely highpowered design. Seven (25%) of the replications yielded effect sizes larger than the original ones, and 21 (75%) yielded effect sizes smaller than the original ones. The median comparable Cohen’s ds were 0.60 for the original findings and 0.15 for the replications. The effect sizes were small (< 0.20) in 16 of the replications (57%), and 9 effects (32%) were in the direction opposite the direction of the original effect. Across settings, the Q statistic indicated significant heterogeneity in 11 (39%) of the replication effects, and most of those were among the findings with the largest overall effect sizes; only 1 effect that was near zero in the aggregate showed significant heterogeneity according to this measure. Only 1 effect had a tau value greater than .20, an indication of moderate heterogeneity. Eight others had tau values near or slightly above .10, an indication of slight heterogeneity. Moderation tests indicated that very little heterogeneity was attributable to the order in which the tasks were performed or whether the tasks were administered in lab versus online. Exploratory comparisons revealed little heterogeneity between Western, educated, industrialized, rich, and democratic (WEIRD) cultures and less WEIRD cultures (i.e., cultures with relatively high and low WEIRDness scores, respectively). Cumulatively, variability in the observed effect sizes was attributable more to the effect being studied than to the sample or setting in which it was studied. UCR::Vicerrectoría de Investigación::Unidades de Investigación::Ciencias Sociales::Instituto de Investigaciones Psicológicas (IIP)
- Published
- 2018
11. A systematic review of the measurement properties of the European Organisation for Research and Treatment of Cancer In-patient Satisfaction with Care Questionnaire, the EORTC IN-PATSAT32
- Author
-
Femke Jansen, Anne Brédart, Neil K. Aaronson, Caroline B. Terwee, Mogens Groenvold, Koen Ilja Neijenhuijs, Bernhard Holzner, Irma M. Verdonck-de Leeuw, and Pim Cuijpers
- Subjects
medicine.medical_specialty ,Psychometrics ,MEDLINE ,Review Article ,PsycINFO ,Validity ,03 medical and health sciences ,0302 clinical medicine ,Patient satisfaction ,SDG 3 - Good Health and Well-being ,Neoplasms ,Surveys and Questionnaires ,Health care ,Criterion validity ,Humans ,Medicine ,Medical physics ,Patient Reported Outcome Measures ,030212 general & internal medicine ,Qualitative Research ,Reliability (statistics) ,PROM ,Psychometry ,Cancer ,business.industry ,Construct validity ,Patient-reported outcome measure ,Reliability ,IN-PATSAT32 ,Europe ,EORTC ,Oncology ,030220 oncology & carcinogenesis ,Systematic review ,business - Abstract
Purpose The EORTC IN-PATSAT32 is a patient-reported outcome measure (PROM) to assess cancer patients’ satisfaction with in-patient health care. The aim of this study was to investigate whether the initial good measurement properties of the IN-PATSAT32 are confirmed in new studies. Methods Within the scope of a larger systematic review study (Prospero ID 42017057237), a systematic search was performed of Embase, Medline, PsycINFO, and Web of Science for studies that investigated measurement properties of the IN-PATSAT32 up to July 2017. Study quality was assessed, data were extracted, and synthesized according to the COnsensus-based Standards for the selection of health Measurement INstruments (COSMIN) methodology. Results Nine studies were included in this review. The evidence on reliability and construct validity were rated as sufficient and of the quality of the evidence as moderate. The evidence on structural validity was rated as insufficient and of low quality. The evidence on internal consistency was indeterminate. Measurement error, responsiveness, criterion validity, and cross-cultural validity were not reported in the included studies. Measurement error could be calculated for two studies and was judged indeterminate. Conclusion In summary, the IN-PATSAT32 performs as expected with respect to reliability and construct validity. No firm conclusions can be made yet whether the IN-PATSAT32 also performs as well with respect to structural validity and internal consistency. Further research on these measurement properties of the PROM is therefore needed as well as on measurement error, responsiveness, criterion validity, and cross-cultural validity. For future studies, it is recommended to take the COSMIN methodology into account. Electronic supplementary material The online version of this article (10.1007/s00520-018-4243-9) contains supplementary material, which is available to authorized users.
- Published
- 2018
12. A young scientists’ perspective on DBS
- Author
-
Daphne van de Bult, Sebastian Arts, Koen Ilja Neijenhuijs, Hannah Andringa, Leon de Bruin, Emiel Wanningen, Laura Klockenbusch, R.P.J.M. Dings, Rowan P. Sommers, Pim Haselager, Clinical Psychology, EMGO+ - Mental Health, CLUE+, and Epistemology and Metaphysics
- Subjects
Medical education ,Deep brain stimulation ,Health Policy ,medicine.medical_treatment ,Communication ,Deep Brain Stimulation ,Perspective (graphical) ,Neuroscience (miscellaneous) ,Repository ,Center for Cognition, Culture and Language (CCCL) ,Perception, Action and Control [DI-BCB_DCC_Theme 2] ,Cognitive artificial intelligence ,Brief Communication ,Psychiatry and Mental health ,Plea ,Neurology ,medicine ,Psychology (miscellaneous) ,Small-N studies ,Psychology ,Set (psychology) ,Neuroscience - Abstract
Contains fulltext : 150585.pdf (Publisher’s version ) (Open Access) Our think tank tasked by the Dutch Health Council, consisting of Radboud University Nijmegen Honours Academy students with various backgrounds, investigated the implications of Deep Brain Stimulation (DBS) for psychiatric patients. During this investigation, a number of methodological, ethical and societal difficulties were identified. We consider these difficulties to be a reflection of a still fragmented field of research that can be overcome with improved organization and communication. To this effect, we suggest that it would be useful to found a centralized DBS organization. Such an organization makes it possible to 1) set up and maintain a repository, 2) facilitate DBS studies with a larger sample size, 3) improve communication amongst researchers, clinicians and ethical committees, and 4) improve communication between DBS experts and the public at large. 4 p.
- Published
- 2015
13. Response to Comment on 'Estimating the reproducibility of psychological science'
- Author
-
Stanka A. Fitneva, Felix Henninger, Franziska Plessow, Fred Hasselman, Jeffrey R. Spies, Carmel A. Levitan, Frank A. Bosco, James A. Grange, Anna Fedor, Felix Cheung, Jesse Chandler, Kathleen Schmidt, Marije van der Hulst, Vivien Estel, Michael C. Frank, Jason M. Prenoveau, Brian A. Nosek, Ashley A. Ricker, Wolf Vanpaemel, Michael Barnett-Cowan, Cody D. Christopherson, Christopher R. Chartier, Johannes M. Meixner, Joshua K. Hartshorne, Nicolás Della Penna, Calvin K. Lai, Štěpán Bahník, Marcel A.L.M. van Assen, Gustav Nilsonne, Marcus R. Munafò, Jeremy K. Miller, Gavin Brent Sullivan, Kai J. Jonas, Katherine Moore, Stefan Stieger, Michelangelo Vianello, Robbie C. M. van Aert, Edward Cremata, Andreas Cordes, Kellylynn Zuni, Koen Ilja Neijenhuijs, Christopher J. Anderson, Martin Voracek, Nina Strohminger, Obstetrics & Gynecology, Social Networks, Solidarity and Inequality, Leerstoel Buskens, Clinical Psychology, EMGO+ - Mental Health, and Department of Methodology and Statistics
- Subjects
Psychological science ,Open science ,media_common.quotation_subject ,MEDLINE ,Behavioral Research ,Psychology ,Publishing ,Research ,Pessimism ,Reproducibility Project ,Research Support ,Brain and Behaviour ,050105 experimental psychology ,03 medical and health sciences ,0302 clinical medicine ,Replication (statistics) ,Journal Article ,0501 psychology and cognitive sciences ,Non-U.S. Gov't ,media_common ,Reproducibility ,Multidisciplinary ,Research Support, Non-U.S. Gov't ,Tobacco and Alcohol ,Comment ,05 social sciences ,Causal inference ,REPLICATION ,Developmental Psychopathology ,030217 neurology & neurosurgery ,Cognitive psychology - Abstract
Contains fulltext : 157278.pdf (Publisher’s version ) (Closed access) Gilbert et al. conclude that evidence from the Open Science Collaboration’s Reproducibility Project: Psychology indicates high reproducibility, given the study methodology. Their very optimistic assessment is limited by statistical misconceptions and by causal inferences from selectively interpreted, correlational data. Using the Reproducibility Project: Psychology data, both optimistic and pessimistic conclusions about reproducibility are possible, and neither are yet warranted. 2 p.
- Published
- 2016
14. Estimating the reproducibility of psychological science
- Author
-
Yoram K. Kunkels, Dylan Selterman, Denise J. Humphries, Kristina G. Brown, David G. Dobolyi, David J. Johnson, Mark A. Roebke, Andy T. Woods, John Hodsoll, Marije van der Hulst, Alexander A. Aarts, Kim Kelso, Erin C. Westgate, James A. Grange, Jesse Chandler, Jenelle Feather, Annick Bosch, Olivia Devitt, Benjamin T. Brown, Megan M. Kyc, Štěpán Bahník, Alissa Melinger, Michael Conn, Rebecca S. Frazier, Marc Jekel, Sara Bowman, Michael J. Wood, Erica Baranski, Sining Wu, Milan Valášek, Anna E. Van't Veer, Jeanine L. M. Skorinko, Joeri Wissink, Sara Steegen, Michael C. Pitts, Douglas Gazarian, Steve N.H. Tsang, Matthew W. Kirkhart, Jennifer S. Beer, Nathali Immelman, Elizabeth Chagnon, Robbie C. M. van Aert, Maya B. Mathur, Magnus Johannesson, Joshua D. Foster, Frank J. Farach, Gandalf Nicolas, Ian S. Penton-Voak, Rebecca M. Goldberg, Sarah L. Thomas, Kathleen Schmidt, Stephanie C. Lin, Linda Cillessen, Belén Fernández-Castilla, Taru Flagan, René Schlegelmilch, Joanneke Weerdmeester, Cyril Pernet, Andreas Cordes, Onur Sahin, Jolanda J. Kossakowski, Samuel Shaki, David Santos, Sabine Scholz, Jeremy R. Gray, Frank Renkewitz, Key Jung Lee, Gillian M. Sandstrom, Marie K. Deserno, Melissa Vazquez, Ed Cremata, Rebecca Saxe, Manuela Thomae, Johannes M. Meixner, Emma Heikensten, Sylvia Eboigbe, Carmel A. Levitan, Natalia Vélez, James G. Field, Riet van Bork, Vivien Estel, Michèle B. Nuijten, Lin Lin, Kate M. Johnson, Bobby Den Bezemer, Jennifer A. Joy-Gaba, Francis Tuerlinckx, Frits Traets, Ilse Luteijn, Christopher R. Chartier, Denise C. Marigold, Denny Borsboom, Elizabeth Gilbert, Jeff Galak, Shannon P. Callahan, E. J. Masicampo, Thomas Talhelm, Chris H.J. Hartgerink, Patrick T. Goodbourn, Stephanie M. Müller, Taylor Nervi, Marcus Möschl, Katherine Moore, Wolf Vanpaemel, Seung K. Kim, Elizabeth Bartmess, Heather N. Mainard, Martin Voracek, Gea Hoogendoorn, Sean P. Mackinnon, Ryan Donohue, Kate A. Ratliff, Jin X. Goh, Anastasia E. Rigney, Andreas Glöckner, Marieke Vermue, Angela S. Attwood, Michelle A. DeGaetano, Nick Spencer, Heather Bentley, Nina Strohminger, Geneva T. Dodson, R. Nathan Pipitone, Hayley M. D. Cleary, Matt Motyl, Amanda L. Forest, Marcus R. Munafò, Marcel Zeelenberg, Susann Fiedler, Ann Calhoun-Sauls, Mallorie Miller, Anondah R. Saide, Ljiljana B. Lazarević, Hilmar Brohmer, Mallory C. Kidwell, Pranjal H. Mehta, Jessie Gorges, Russ Clay, Jeffrey R. Spies, Joanna E. Anderson, Johnny van Doorn, Ashley A. Ricker, Elizabeth W. Dunn, Erin L Braswell, Jamie DeCoster, Larissa Seibel, Matthias Lippold, Lutz Ostkamp, William B. Simpson, Cathy On-Ying Hung, Carina Sonnleitner, Emily M. Wright, Laura Dewitte, Koen Ilja Neijenhuijs, Tim Kuhlmann, Job Krijnen, Leah Beyan, Jesse Graham, Andrew M Rivers, Sacha Epskamp, Aamir Laique, Christopher J. Anderson, Peter Raymond Attridge, Eric-Jan Wagenmakers, Agnieszka Slowik, Michael C. Frank, Bryan Gorges, Alejandro Vásquez Echeverría, Gina Vuu, Giulio Costantini, Eskil Forsell, Michelangelo Vianello, Don van den Bergh, Anna Fedor, Courtney K. Soderberg, M. Brent Donnellan, Kayleigh E Easey, Shauna Gordon-McKeon, Raoul Bell, William J. Johnston, Brian A. Nosek, Ashlee Welsh, Melissa Lewis, Anna Dreber, Simon Columbus, Frank A. Bosco, Pia Tio, Joshua K. Hartshorne, Lars Goellner, Elisa Maria Galliani, Etienne P. Le Bel, Kellylynn Zuni, Olivia Perna, Kristi M. Lemm, Marco Perugini, Anniek M. te Dorsthorst, Hedderik van Rijn, Timothy M. Errington, Bennett Kleinberg, Vanessa C. Irsik, Frank Jäkel, Timothy Hayes, Mark Verschoor, Mark D. Cloud, Bethany Lassetter, Justin Goss, Paul J. Turchan, Gavin Brent Sullivan, Darren Loureiro, Jo Embley, Robert S. Ryan, Jovita Brüning, Jan Crusius, Joel S. Snyder, Larissa Gabrielle Johnson, Nicolás Delia Penna, Grace Binion, Calvin K. Lai, Gustav Nilsonne, Heather M. Fuchs, Angela Rachael Dorrough, Michelle Dugas, Johanna Cohoon, Minha Lee, Robert Krause, David Reinhard, Goran Knežević, Jason M. Prenoveau, Kristin A. Lane, Stanka A. Fitneva, Rima-Maria Rahal, Mathijs Van De Ven, Anup Gampa, Marcel A.L.M. van Assen, Jordan Axt, Felix Henninger, Misha Pavel, Daniel Lakens, Jeremy K. Miller, Sara García, Leslie Cramblet Alvarez, Colleen Osborne, Kai J. Jonas, Taylor Holubar, Stefan Stieger, Heather Barry Kappes, Felix Cheung, Daan R. van Renswoude, Catherine Olsson, Roel van Dooren, Tylar Martinez, Megan Tapia, Philip A. Gable, Cody D. Christopherson, Franziska Plessow, Roger Giner-Sorolla, Abraham M. Rutchick, Michael Barnett-Cowan, Mark J. Brandt, Rebecca A. Dore, Michael May, H. Colleen Sinclair, Georg Jahn, Daniel P. Martin, Fred Hasselman, Casey Eggleston, Nicole Mechin, Joshua J. Matacotta, Molly Babel, Franziska Maria Kolorz, Social & Organizational Psychology, IBBA, Clinical Psychology, EMGO+ - Mental Health, Social Networks, Solidarity and Inequality, Department of Social Psychology, Department of Methodology and Statistics, Aarts, A, Anderson, J, Anderson, C, Attridge, P, Attwood, A, Axt, J, Babel, M, Bahník, Š, Baranski, E, Barnett Cowan, M, Bartmess, E, Beer, J, Bell, R, Bentley, H, Beyan, L, Binion, G, Borsboom, D, Bosch, A, Bosco, F, Bowman, S, Brandt, M, Braswell, E, Brohmer, H, Brown, B, Brown, K, Brüning, J, Calhoun Sauls, A, Callahan, S, Chagnon, E, Chandler, J, Chartier, C, Cheung, C, Cd, Cillessen, L, Clay, R, Cleary, H, Cloud, M, Cohn, M, Cohoon, J, Columbus, S, Cordes, A, Costantini, G, Cramblet Alvarez, L, Cremata, E, Crusius, J, Decoster, J, Degaetano, M, Della Penna, N, den Bezemer, B, Deserno, M, Devitt, O, Dewitte, L, Dobolyi, D, Dodson, G, Donnellan, M, Donohue, R, Dore, R, Dorrough, A, Dreber, A, Dugas, M, Dunn, E, Easey, K, Eboigbe, S, Eggleston, C, Embley, J, Epskamp, S, Errington, T, Estel, V, Farach, F, Feather, J, Fedor, A, Fernández Castilla, B, Fiedler, S, Field, J, Fitneva, S, Flagan, T, Forest, A, Forsell, E, Foster, J, Frank, M, Frazier, R, Fuchs, H, Gable, P, Galak, J, Galliani, E, Gampa, A, Garcia, S, Gazarian, D, Gilbert, E, Giner Sorolla, R, Glöckner, A, Goellner, L, Goh, J, Goldberg, R, Goodbourn, P, Gordon McKeon, S, Gorges, B, Gorges, J, Goss, J, Graham, J, Grange, J, Gray, J, Hartgerink, C, Hartshorne, J, Hasselman, F, Hayes, T, Heikensten, E, Henninger, F, Hodsoll, J, Holubar, T, Hoogendoorn, G, Humphries, D, Hung, C, Immelman, N, Irsik, V, Jahn, G, Jäkel, F, Jekel, M, Johannesson, M, Johnson, L, Johnson, D, Johnson, K, Johnston, W, Jonas, K, Joy Gaba, J, Kappes, H, Kelso, K, Kidwell, M, Kim, S, Kirkhart, M, Kleinberg, B, Kneževic, G, Kolorz, F, Kossakowski, J, Krause, R, Krijnen, J, Kuhlmann, T, Kunkels, Y, Kyc, M, Lai, C, Laique, A, Lakens, D, Lane, K, Lassetter, B, Lazarevic, L, Lebel, E, Lee, K, Lee, M, Lemm, K, Levitan, C, Lewis, M, Lin, L, Lin, S, Lippold, M, Loureiro, D, Luteijn, I, Mackinnon, S, Mainard, H, Marigold, D, Martin, D, Martinez, T, Masicampo, E, Matacotta, J, Mathur, M, May, M, Mechin, N, Mehta, P, Meixner, J, Melinger, A, Miller, J, Miller, M, Moore, K, Möschl, M, Motyl, M, Müller, S, Munafo, M, Neijenhuijs, K, Nervi, T, Nicolas, G, Nilsonne, G, Nosek, B, Nuijten, M, Olsson, C, Osborne, C, Ostkamp, L, Pavel, M, Penton Voak, I, Perna, O, Pernet, C, Perugini, M, Pipitone, N, Pitts, M, Plessow, F, Prenoveau, J, Rahal, R, Ratliff, K, Reinhard, D, Renkewitz, F, Ricker, A, Rigney, A, Rivers, A, Roebke, M, Rutchick, A, Ryan, R, Sahin, O, Saide, A, Sandstrom, G, Santos, D, Saxe, R, Schlegelmilch, R, Schmidt, K, Scholz, S, Seibel, L, Selterman, D, Shaki, S, Simpson, E, Sinclair, H, Skorinko, J, Slowik, A, Snyder, J, Soderberg, C, Sonnleitner, C, Spencer, N, Spies, J, Steegen, S, Stieger, S, Strohminger, N, Sullivan, G, Talhelm, T, Tapia, M, te Dorsthorst, A, Thomae, M, Thomas, S, Tio, P, Traets, F, Tsang, S, Tuerlinckx, F, Turchan, P, Valášek, M, van 't Veer, A, Van Aert, R, van Assen, M, van Bork, R, van de Ven, M, van den Bergh, D, van der Hulst, M, van Dooren, R, van Doorn, J, van Renswoude, D, van Rijn, H, Vanpaemel, W, Vásquez Echeverría, A, Vazquez, M, Velez, N, Vermue, M, Verschoor, M, Vianello, M, Voracek, M, Vuu, G, Wagenmakers, E, Weerdmeester, J, Welsh, A, Westgate, E, Wissink, J, Wood, M, Woods, A, Wright, E, Wu, S, Zeelenberg, M, Zuni, K, Sociology/ICS, Experimental Psychology, Human Technology Interaction, Sociale Psychologie (Psychologie, FMG), Ontwikkelingspsychologie (Psychologie, FMG), and Brein en Cognitie (Psychologie, FMG)
- Subjects
Research design ,Department Psychologie ,BF Psychology ,media_common.quotation_subject ,POWER ,Learning and Plasticity ,Reproducibility Project ,Q1 ,Experimental Psychopathology and Treatment ,Replication (statistics) ,Statistics ,TRUTH ,Psychology ,General ,Mathematics ,media_common ,Selection bias ,Replication crisis ,Behaviour Change and Well-being ,Multidisciplinary ,PUBLICATION ,Publication bias ,Reproducibility ,Confidence interval ,INCENTIVES ,PREVALENCE ,Meta-analysis ,REPLICABILITY ,REPLICATION ,Developmental Psychopathology ,FALSE - Abstract
IntroductionReproducibility is a defining feature of science, but the extent to which it characterizes current research is unknown. Scientific claims should not gain credence because of the status or authority of their originator but by the replicability of their supporting evidence. Even research of exemplary quality may have irreproducible empirical findings because of random or systematic error.RationaleThere is concern about the rate and predictors of reproducibility, but limited evidence. Potentially problematic practices include selective reporting, selective analysis, and insufficient specification of the conditions necessary or sufficient to obtain the results. Direct replication is the attempt to recreate the conditions believed sufficient for obtaining a previously observed finding and is the means of establishing reproducibility of a finding with new data. We conducted a large-scale, collaborative effort to obtain an initial estimate of the reproducibility of psychological science.ResultsWe conducted replications of 100 experimental and correlational studies published in three psychology journals using high-powered designs and original materials when available. There is no single standard for evaluating replication success. Here, we evaluated reproducibility using significance and P values, effect sizes, subjective assessments of replication teams, and meta-analysis of effect sizes. The mean effect size (r) of the replication effects (Mr = 0.197, SD = 0.257) was half the magnitude of the mean effect size of the original effects (Mr = 0.403, SD = 0.188), representing a substantial decline. Ninety-seven percent of original studies had significant results (P < .05). Thirty-six percent of replications had significant results; 47% of original effect sizes were in the 95% confidence interval of the replication effect size; 39% of effects were subjectively rated to have replicated the original result; and if no bias in original results is assumed, combining original and replication results left 68% with statistically significant effects. Correlational tests suggest that replication success was better predicted by the strength of original evidence than by characteristics of the original and replication teams.ConclusionNo single indicator sufficiently describes replication success, and the five indicators examined here are not the only ways to evaluate reproducibility. Nonetheless, collectively these results offer a clear conclusion: A large portion of replications produced weaker evidence for the original findings despite using materials provided by the original authors, review in advance for methodological fidelity, and high statistical power to detect the original effect sizes. Moreover, correlational evidence is consistent with the conclusion that variation in the strength of initial evidence (such as original P value) was more predictive of replication success than variation in the characteristics of the teams conducting the research (such as experience and expertise). The latter factors certainly can influence replication success, but they did not appear to do so here. Reproducibility is not well understood because the incentives for individual scientists prioritize novelty over replication. Innovation is the engine of discovery and is vital for a productive, effective scientific enterprise. However, innovative ideas become old news fast. Journal reviewers and editors may dismiss a new test of a published idea as unoriginal. The claim that “we already know this” belies the uncertainty of scientific evidence. Innovation points out paths that are possible; replication points out paths that are likely; progress relies on both. Replication can increase certainty when findings are reproduced and promote innovation when they are not. This project provides accumulating evidence for many findings in psychological research and suggests that there is still more work to do to verify whether we know what we think we know.
- Published
- 2015
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.