115 results on '"Dinnes J"'
Search Results
2. Electronic and animal noses for detecting SARS-CoV-2 infection (Protocol)
- Author
-
Leeflang, MMG, Bell, Katy J.L., Deeks, JJ, Dinnes, J, Doust, J, Korevaar, DA, Lord, SJ, and Spijker, R
- Subjects
SARs ,1107 Immunology ,animal ,1117 Public Health and Health Services - Abstract
This is a protocol for a Cochrane Review (diagnostic). The objectives are as follows: 1. To assess the diagnostic test accuracy of eNoses to screen for severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) infection in public places, such as airports. 2. To assess the diagnostic test accuracy of sniffer animals, and more specifically dogs, to screen for SARS-CoV-2 infection in public places, such as airports. 3. To assess the diagnostic test accuracy of eNoses for SARS-CoV-2 infection or COVID-19 in symptomatic people presenting in the community, or in secondary care. 4. To assess the diagnostic test accuracy of sniffer animals, and more specifically dogs, for SARS-CoV-2 infection or COVID-19 in symptomatic people presenting in the community, or in secondary care.
- Published
- 2021
3. Rapid, point-of-care antigen and molecular-based tests for diagnosis of SARS-CoV-2 infection
- Author
-
Dinnes, J, Deeks, JJ, Berhane, S, Taylor, M, Adriano, A, Davenport, C, Dittrich, S, Emperador, D, Takwoingi, Y, Cunningham, J, Beese, S, Domen, J, Dretzke, J, Ferrante di Ruffano, L, Harris, IM, Price, MJ, Taylor-Phillips, S, Hooft, L, Leeflang, MM, McInnes, MD, Spijker, R, Van den Bruel, A, Epidemiology and Data Science, APH - Methodology, APH - Personalized Medicine, and Group, Cochrane COVID-19 Diagnostic Test Accuracy
- Subjects
Cohort Studies ,COVID-19 Testing ,0302 clinical medicine ,Diagnosis ,wc_505 ,False positive paradox ,Medicine ,Pharmacology (medical) ,030212 general & internal medicine ,Child ,Asymptomatic Infections ,Antigens, Viral ,False Negative Reactions ,wa_105 ,Infectious disease ,education.field_of_study ,Reference Standards ,Molecular Diagnostic Techniques ,Rapid antigen test ,COVID-19 Nucleic Acid Testing ,Predictive value of tests ,Meta-analysis ,Sample collection ,wb_141 ,medicine.symptom ,Coronavirus Infections ,Adult ,medicine.medical_specialty ,Point-of-Care Systems ,Pneumonia, Viral ,Population ,Sensitivity and Specificity ,Asymptomatic ,COVID-19 Serological Testing ,Betacoronavirus ,03 medical and health sciences ,Bias ,Predictive Value of Tests ,Internal medicine ,Humans ,False Positive Reactions ,education ,Pandemics ,qw_573 ,Clinical Laboratory Techniques ,SARS-CoV-2 ,business.industry ,COVID-19 ,Confidence interval ,Cochrane COVID-19 Diagnostic Test Accuracy Group ,business ,030217 neurology & neurosurgery - Abstract
Background Accurate rapid diagnostic tests for SARS‐CoV‐2 infection could contribute to clinical and public health strategies to manage the COVID‐19 pandemic. Point‐of‐care antigen and molecular tests to detect current infection could increase access to testing and early confirmation of cases, and expediate clinical and public health management decisions that may reduce transmission. Objectives To assess the diagnostic accuracy of point‐of‐care antigen and molecular‐based tests for diagnosis of SARS‐CoV‐2 infection. We consider accuracy separately in symptomatic and asymptomatic population groups. Search methods Electronic searches of the Cochrane COVID‐19 Study Register and the COVID‐19 Living Evidence Database from the University of Bern (which includes daily updates from PubMed and Embase and preprints from medRxiv and bioRxiv) were undertaken on 30 Sept 2020. We checked repositories of COVID‐19 publications and included independent evaluations from national reference laboratories, the Foundation for Innovative New Diagnostics and the Diagnostics Global Health website to 16 Nov 2020. We did not apply language restrictions. Selection criteria We included studies of people with either suspected SARS‐CoV‐2 infection, known SARS‐CoV‐2 infection or known absence of infection, or those who were being screened for infection. We included test accuracy studies of any design that evaluated commercially produced, rapid antigen or molecular tests suitable for a point‐of‐care setting (minimal equipment, sample preparation, and biosafety requirements, with results within two hours of sample collection). We included all reference standards that define the presence or absence of SARS‐CoV‐2 (including reverse transcription polymerase chain reaction (RT‐PCR) tests and established diagnostic criteria). Data collection and analysis Studies were screened independently in duplicate with disagreements resolved by discussion with a third author. Study characteristics were extracted by one author and checked by a second; extraction of study results and assessments of risk of bias and applicability (made using the QUADAS‐2 tool) were undertaken independently in duplicate. We present sensitivity and specificity with 95% confidence intervals (CIs) for each test and pooled data using the bivariate model separately for antigen and molecular‐based tests. We tabulated results by test manufacturer and compliance with manufacturer instructions for use and according to symptom status. Main results Seventy‐eight study cohorts were included (described in 64 study reports, including 20 pre‐prints), reporting results for 24,087 samples (7,415 with confirmed SARS‐CoV‐2). Studies were mainly from Europe (n = 39) or North America (n = 20), and evaluated 16 antigen and five molecular assays. We considered risk of bias to be high in 29 (37%) studies because of participant selection; in 66 (85%) because of weaknesses in the reference standard for absence of infection; and in 29 (37%) for participant flow and timing. Studies of antigen tests were of a higher methodological quality compared to studies of molecular tests, particularly regarding the risk of bias for participant selection and the index test. Characteristics of participants in 35 (45%) studies differed from those in whom the test was intended to be used and the delivery of the index test in 39 (50%) studies differed from the way in which the test was intended to be used. Nearly all studies (97%) defined the presence or absence of SARS‐CoV‐2 based on a single RT‐PCR result, and none included participants meeting case definitions for probable COVID‐19. Antigen tests Forty‐eight studies reported 58 evaluations of antigen tests. Estimates of sensitivity varied considerably between studies. There were differences between symptomatic (72.0%, 95% CI 63.7% to 79.0%; 37 evaluations; 15530 samples, 4410 cases) and asymptomatic participants (58.1%, 95% CI 40.2% to 74.1%; 12 evaluations; 1581 samples, 295 cases). Average sensitivity was higher in the first week after symptom onset (78.3%, 95% CI 71.1% to 84.1%; 26 evaluations; 5769 samples, 2320 cases) than in the second week of symptoms (51.0%, 95% CI 40.8% to 61.0%; 22 evaluations; 935 samples, 692 cases). Sensitivity was high in those with cycle threshold (Ct) values on PCR ≤25 (94.5%, 95% CI 91.0% to 96.7%; 36 evaluations; 2613 cases) compared to those with Ct values >25 (40.7%, 95% CI 31.8% to 50.3%; 36 evaluations; 2632 cases). Sensitivity varied between brands. Using data from instructions for use (IFU) compliant evaluations in symptomatic participants, summary sensitivities ranged from 34.1% (95% CI 29.7% to 38.8%; Coris Bioconcept) to 88.1% (95% CI 84.2% to 91.1%; SD Biosensor STANDARD Q). Average specificities were high in symptomatic and asymptomatic participants, and for most brands (overall summary specificity 99.6%, 95% CI 99.0% to 99.8%). At 5% prevalence using data for the most sensitive assays in symptomatic people (SD Biosensor STANDARD Q and Abbott Panbio), positive predictive values (PPVs) of 84% to 90% mean that between 1 in 10 and 1 in 6 positive results will be a false positive, and between 1 in 4 and 1 in 8 cases will be missed. At 0.5% prevalence applying the same tests in asymptomatic people would result in PPVs of 11% to 28% meaning that between 7 in 10 and 9 in 10 positive results will be false positives, and between 1 in 2 and 1 in 3 cases will be missed. No studies assessed the accuracy of repeated lateral flow testing or self‐testing. Rapid molecular assays Thirty studies reported 33 evaluations of five different rapid molecular tests. Sensitivities varied according to test brand. Most of the data relate to the ID NOW and Xpert Xpress assays. Using data from evaluations following the manufacturer’s instructions for use, the average sensitivity of ID NOW was 73.0% (95% CI 66.8% to 78.4%) and average specificity 99.7% (95% CI 98.7% to 99.9%; 4 evaluations; 812 samples, 222 cases). For Xpert Xpress, the average sensitivity was 100% (95% CI 88.1% to 100%) and average specificity 97.2% (95% CI 89.4% to 99.3%; 2 evaluations; 100 samples, 29 cases). Insufficient data were available to investigate the effect of symptom status or time after symptom onset. Authors' conclusions Antigen tests vary in sensitivity. In people with signs and symptoms of COVID‐19, sensitivities are highest in the first week of illness when viral loads are higher. The assays shown to meet appropriate criteria, such as WHO's priority target product profiles for COVID‐19 diagnostics (‘acceptable’ sensitivity ≥ 80% and specificity ≥ 97%), can be considered as a replacement for laboratory‐based RT‐PCR when immediate decisions about patient care must be made, or where RT‐PCR cannot be delivered in a timely manner. Positive predictive values suggest that confirmatory testing of those with positive results may be considered in low prevalence settings. Due to the variable sensitivity of antigen tests, people who test negative may still be infected. Evidence for testing in asymptomatic cohorts was limited. Test accuracy studies cannot adequately assess the ability of antigen tests to differentiate those who are infectious and require isolation from those who pose no risk, as there is no reference standard for infectiousness. A small number of molecular tests showed high accuracy and may be suitable alternatives to RT‐PCR. However, further evaluations of the tests in settings as they are intended to be used are required to fully establish performance in practice. Several important studies in asymptomatic individuals have been reported since the close of our search and will be incorporated at the next update of this review. Comparative studies of antigen tests in their intended use settings and according to test operator (including self‐testing) are required., Plain language summary How accurate are rapid tests for diagnosing COVID‐19? What are rapid point‐of‐care tests for COVID‐19? Rapid point‐of‐care tests aim to confirm or rule out COVID‐19 infection in people with or without COVID‐19 symptoms. They: ‐ are portable, so they can be used wherever the patient is (at the point of care); ‐ are easy to perform, with a minimum amount of extra equipment or complicated preparation steps; ‐ are less expensive than standard laboratory tests; ‐ do not require a specialist operator or setting; and ‐ provide results ‘while you wait’. We were interested in two types of commercially available, rapid point‐of‐care tests: antigen and molecular tests. Antigen tests identify proteins on the virus; they come in disposable plastic cassettes, similar to pregnancy tests. Rapid molecular tests detect the virus’s genetic material in a similar way to laboratory methods, but using smaller devices that are easy to transport or to set up outside of a specialist laboratory. Both test nose or throat samples. Why is this question important? People with suspected COVID‐19 need to know quickly whether they are infected, so that they can self‐isolate, receive treatment, and inform close contacts. Currently, COVID‐19 infection is confirmed by a laboratory test called RT‐PCR, which uses specialist equipment and often takes at least 24 hours to produce a result. Rapid point‐of‐care tests could open access to testing for many more people, with and without symptoms, potentially in locations other than healthcare settings. If they are accurate, faster diagnosis could allow people to take appropriate action more quickly, with the potential to reduce the spread of COVID‐19. What did we want to find out? We wanted to know whether commercially available, rapid point‐of‐care antigen and molecular tests are accurate enough to diagnose COVID‐19 infection reliably, and to find out if accuracy differs in people with and without symptoms. What did we do? We looked for studies that measured the accuracy of any commercially produced, rapid antigen or molecular point‐of‐care test, in people tested for COVID‐19 using RT‐PCR. People could be tested in hospital or the community. Studies could test people with or without symptoms. Tests had to use minimal equipment, be performed safely without risking infection from the sample, and have results available within two hours of the sample being collected. What we found We included 64 studies in the review. They investigated a total of 24,087 nose or throat samples; COVID‐19 was confirmed in 7415 of these samples. Studies investigated 16 different antigen tests and five different molecular tests. They took place mainly in Europe and North America. Main results Antigen tests In people with confirmed COVID‐19, antigen tests correctly identified COVID‐19 infection in an average of 72% of people with symptoms, compared to 58% of people without symptoms. Tests were most accurate when used in the first week after symptoms first developed (an average of 78% of confirmed cases had positive antigen tests). This is likely to be because people have the most virus in their system in the first days after they are infected. In people who did not have COVID‐19, antigen tests correctly ruled out infection in 99.5% of people with symptoms and 98.9% of people without symptoms. Different brands of tests varied in accuracy. Pooled results for one test (SD Biosensor STANDARD Q) met World Health Organization (WHO) standards as ‘acceptable’ for confirming and ruling out COVID‐19 in people with signs and symptoms of COVID‐19. Two more tests met the WHO acceptable standards (Abbott Panbio and BIONOTE NowCheck) in at least one study. Using summary results for SD Biosensor STANDARD Q, if 1000 people with symptoms had the antigen test, and 50 (5%) of them really had COVID‐19: ‐ 53 people would test positive for COVID‐19. Of these, 9 people (17%) would not have COVID‐19 (false positive result). ‐ 947 people would test negative for COVID‐19. Of these, 6 people (0.6%) would actually have COVID‐19 (false negative result). In people with no symptoms of COVID‐19 the number of confirmed cases is expected to be much lower than in people with symptoms. Using summary results for SD Biosensor STANDARD Q in a bigger population of 10,000 people with no symptoms, where 50 (0.5%) of them really had COVID‐19: ‐ 125 people would test positive for COVID‐19. Of these, 90 people (72%) would not have COVID‐19 (false positive result). ‐ 9,875 people would test negative for COVID‐19. Of these, 15 people (0.2%) would actually have COVID‐19 (false negative result). Molecular tests Although overall results for diagnosing and ruling out COVID‐19 were good (95.1% of infections correctly diagnosed and 99% correctly ruled out), 69% of the studies used the tests in laboratories instead of at the point‐of‐care and few studies followed test manufacturer instructions. Most of the data relate to the ID NOW and Xpert Xpress tests. We noted a large difference in COVID‐19 detection between the two tests, but we cannot be certain about whether results will remain the same in a real world setting. We could not investigate differences in people with or without symptoms, nor time from when symptoms first showed because the studies did not provide enough information about their participants. How reliable were the results of the studies? In general, studies that assessed antigen tests used more rigorous methods than those that assessed molecular tests, particularly when selecting participants and performing the tests. Sometimes studies did not perform the test on the people for whom it was intended and did not follow the manufacturers’ instructions for using the test. Sometimes the tests were not carried out at the point‐of‐care. Nearly all the studies (97%) relied on a single negative RT‐PCR result as evidence of no COVID‐19 infection. Results from different test brands varied, and few studies directly compared one test brand with another. Finally, not all studies gave enough information about their participants for us to judge how long they had had symptoms, or even whether or not they had symptoms. What does this mean? Some antigen tests are accurate enough to replace RT‐PCR when used in people with symptoms. This would be most useful when quick decisions are needed about patient care, or if RT‐PCR is not available. Antigen tests may be most useful to identify outbreaks, or to select people with symptoms for further testing with PCR, allowing self‐isolation or contact tracing and reducing the burden on laboratory services. People who receive a negative antigen test result may still be infected. Several point‐of‐care molecular tests show very high accuracy and potential for use, but more evidence of their performance when evaluated in real life settings is required. We need more evidence on rapid testing in people without symptoms, on the accuracy of repeated testing, testing in non‐healthcare settings such as schools (including self‐testing), and direct comparisons of test brands, with testers following manufacturers’ instructions. How up‐to‐date is this review? This review updates our previous review and includes evidence published up to 30 September 2020.
- Published
- 2020
- Full Text
- View/download PDF
4. Antibody tests for identification of current and past infection with SARS-CoV-2
- Author
-
Fox, T, Geppert, J, Dinnes, J, Scandrett, K, Bigio, J, Sulis, G, Hettiarachchi, D, Mathangasinghe, Y, Weeratunga, P, Wickramasinghe, D, Bergman, Hanna, Buckly, Brian, Probyn, Katrin, Sguassero, Yanina, Davenport, Clare, Cunningham, Jane, Dittrich, Sabine, Emperador, Devy, Hooft, Lotty, Leeflang, Mariska, McInnes, Matthew, Spijker, René, Struyf, Thomas, Van den Bruel, Ann, Verbakel, Jan, Takwoingi, Yemisi, Taylor-Phillips, Sian, Deeks, Jonathan, Cochrane COVID-19 Diagnostic Test Accuracy Group, Epidemiology and Data Science, APH - Methodology, and APH - Personalized Medicine
- Subjects
medicine.medical_specialty ,COVID-19 Vaccines ,media_common.quotation_subject ,Pneumonia, Viral ,Logistic regression ,Antibodies, Viral ,Asymptomatic ,Sensitivity and Specificity ,Serology ,03 medical and health sciences ,Betacoronavirus ,0302 clinical medicine ,Antibody Specificity ,Seroepidemiologic Studies ,Internal medicine ,Medicine ,Seroprevalence ,Humans ,False Positive Reactions ,Serologic Tests ,Pharmacology (medical) ,030212 general & internal medicine ,False Negative Reactions ,Pandemics ,Selection Bias ,media_common ,Selection bias ,business.industry ,Reverse Transcriptase Polymerase Chain Reaction ,SARS-CoV-2 ,COVID-19 ,Reference Standards ,Confidence interval ,Immunoglobulin A ,Immunoglobulin M ,Sample size determination ,Immunoglobulin G ,medicine.symptom ,business ,Coronavirus Infections ,030217 neurology & neurosurgery - Abstract
Background The diagnostic challenges associated with the COVID‐19 pandemic resulted in rapid development of diagnostic test methods for detecting SARS‐CoV‐2 infection. Serology tests to detect the presence of antibodies to SARS‐CoV‐2 enable detection of past infection and may detect cases of SARS‐CoV‐2 infection that were missed by earlier diagnostic tests. Understanding the diagnostic accuracy of serology tests for SARS‐CoV‐2 infection may enable development of effective diagnostic and management pathways, inform public health management decisions and understanding of SARS‐CoV‐2 epidemiology. Objectives To assess the accuracy of antibody tests, firstly, to determine if a person presenting in the community, or in primary or secondary care has current SARS‐CoV‐2 infection according to time after onset of infection and, secondly, to determine if a person has previously been infected with SARS‐CoV‐2. Sources of heterogeneity investigated included: timing of test, test method, SARS‐CoV‐2 antigen used, test brand, and reference standard for non‐SARS‐CoV‐2 cases. Search methods The COVID‐19 Open Access Project living evidence database from the University of Bern (which includes daily updates from PubMed and Embase and preprints from medRxiv and bioRxiv) was searched on 30 September 2020. We included additional publications from the Evidence for Policy and Practice Information and Co‐ordinating Centre (EPPI‐Centre) ‘COVID‐19: Living map of the evidence’ and the Norwegian Institute of Public Health ’NIPH systematic and living map on COVID‐19 evidence’. We did not apply language restrictions. Selection criteria We included test accuracy studies of any design that evaluated commercially produced serology tests, targeting IgG, IgM, IgA alone, or in combination. Studies must have provided data for sensitivity, that could be allocated to a predefined time period after onset of symptoms, or after a positive RT‐PCR test. Small studies with fewer than 25 SARS‐CoV‐2 infection cases were excluded. We included any reference standard to define the presence or absence of SARS‐CoV‐2 (including reverse transcription polymerase chain reaction tests (RT‐PCR), clinical diagnostic criteria, and pre‐pandemic samples). Data collection and analysis We use standard screening procedures with three reviewers. Quality assessment (using the QUADAS‐2 tool) and numeric study results were extracted independently by two people. Other study characteristics were extracted by one reviewer and checked by a second. We present sensitivity and specificity with 95% confidence intervals (CIs) for each test and, for meta‐analysis, we fitted univariate random‐effects logistic regression models for sensitivity by eligible time period and for specificity by reference standard group. Heterogeneity was investigated by including indicator variables in the random‐effects logistic regression models. We tabulated results by test manufacturer and summarised results for tests that were evaluated in 200 or more samples and that met a modification of UK Medicines and Healthcare products Regulatory Agency (MHRA) target performance criteria. Main results We included 178 separate studies (described in 177 study reports, with 45 as pre‐prints) providing 527 test evaluations. The studies included 64,688 samples including 25,724 from people with confirmed SARS‐CoV‐2; most compared the accuracy of two or more assays (102/178, 57%). Participants with confirmed SARS‐CoV‐2 infection were most commonly hospital inpatients (78/178, 44%), and pre‐pandemic samples were used by 45% (81/178) to estimate specificity. Over two‐thirds of studies recruited participants based on known SARS‐CoV‐2 infection status (123/178, 69%). All studies were conducted prior to the introduction of SARS‐CoV‐2 vaccines and present data for naturally acquired antibody responses. Seventy‐nine percent (141/178) of studies reported sensitivity by week after symptom onset and 66% (117/178) for convalescent phase infection. Studies evaluated enzyme‐linked immunosorbent assays (ELISA) (165/527; 31%), chemiluminescent assays (CLIA) (167/527; 32%) or lateral flow assays (LFA) (188/527; 36%). Risk of bias was high because of participant selection (172, 97%); application and interpretation of the index test (35, 20%); weaknesses in the reference standard (38, 21%); and issues related to participant flow and timing (148, 82%). We judged that there were high concerns about the applicability of the evidence related to participants in 170 (96%) studies, and about the applicability of the reference standard in 162 (91%) studies. Average sensitivities for current SARS‐CoV‐2 infection increased by week after onset for all target antibodies. Average sensitivity for the combination of either IgG or IgM was 41.1% in week one (95% CI 38.1 to 44.2; 103 evaluations; 3881 samples, 1593 cases), 74.9% in week two (95% CI 72.4 to 77.3; 96 evaluations, 3948 samples, 2904 cases) and 88.0% by week three after onset of symptoms (95% CI 86.3 to 89.5; 103 evaluations, 2929 samples, 2571 cases). Average sensitivity during the convalescent phase of infection (up to a maximum of 100 days since onset of symptoms, where reported) was 89.8% for IgG (95% CI 88.5 to 90.9; 253 evaluations, 16,846 samples, 14,183 cases), 92.9% for IgG or IgM combined (95% CI 91.0 to 94.4; 108 evaluations, 3571 samples, 3206 cases) and 94.3% for total antibodies (95% CI 92.8 to 95.5; 58 evaluations, 7063 samples, 6652 cases). Average sensitivities for IgM alone followed a similar pattern but were of a lower test accuracy in every time slot. Average specificities were consistently high and precise, particularly for pre‐pandemic samples which provide the least biased estimates of specificity (ranging from 98.6% for IgM to 99.8% for total antibodies). Subgroup analyses suggested small differences in sensitivity and specificity by test technology however heterogeneity in study results, timing of sample collection, and smaller sample numbers in some groups made comparisons difficult. For IgG, CLIAs were the most sensitive (convalescent‐phase infection) and specific (pre‐pandemic samples) compared to both ELISAs and LFAs (P < 0.001 for differences across test methods). The antigen(s) used (whether from the Spike‐protein or nucleocapsid) appeared to have some effect on average sensitivity in the first weeks after onset but there was no clear evidence of an effect during convalescent‐phase infection. Investigations of test performance by brand showed considerable variation in sensitivity between tests, and in results between studies evaluating the same test. For tests that were evaluated in 200 or more samples, the lower bound of the 95% CI for sensitivity was 90% or more for only a small number of tests (IgG, n = 5; IgG or IgM, n = 1; total antibodies, n = 4). More test brands met the MHRA minimum criteria for specificity of 98% or above (IgG, n = 16; IgG or IgM, n = 5; total antibodies, n = 7). Seven assays met the specified criteria for both sensitivity and specificity. In a low‐prevalence (2%) setting, where antibody testing is used to diagnose COVID‐19 in people with symptoms but who have had a negative PCR test, we would anticipate that 1 (1 to 2) case would be missed and 8 (5 to 15) would be falsely positive in 1000 people undergoing IgG or IgM testing in week three after onset of SARS‐CoV‐2 infection. In a seroprevalence survey, where prevalence of prior infection is 50%, we would anticipate that 51 (46 to 58) cases would be missed and 6 (5 to 7) would be falsely positive in 1000 people having IgG tests during the convalescent phase (21 to 100 days post‐symptom onset or post‐positive PCR) of SARS‐CoV‐2 infection. Authors' conclusions Some antibody tests could be a useful diagnostic tool for those in whom molecular‐ or antigen‐based tests have failed to detect the SARS‐CoV‐2 virus, including in those with ongoing symptoms of acute infection (from week three onwards) or those presenting with post‐acute sequelae of COVID‐19. However, antibody tests have an increasing likelihood of detecting an immune response to infection as time since onset of infection progresses and have demonstrated adequate performance for detection of prior infection for sero‐epidemiological purposes. The applicability of results for detection of vaccination‐induced antibodies is uncertain. Plain language summary What is the diagnostic accuracy of antibody tests for the detection of infection with the COVID‐19 virus? Background COVID‐19 is an infectious disease caused by the SARS‐CoV‐2 virus that spreads easily between people in a similar way to the common cold or ‘flu’. Most people with COVID‐19 have a mild‐to‐moderate respiratory illness, and some may have no symptoms (asymptomatic infection). Others experience severe symptoms and need specialist treatment and intensive care. In response to COVID‐19 infection, the immune system develops proteins called antibodies that can attack the virus as it circulates in their blood. People who have been vaccinated against COVID‐19 also produce these antibodies against the virus. Tests are available to detect antibodies in peoples' blood, which may indicate that they currently have COVID‐19 or have had it previously, or it may indicate that they have been vaccinated (although this group was not the focus of this review). Why are accurate tests important? Accurate testing allows identification of people who need to isolate themselves to prevent the spread of infection, or who might need treatment for their infection. Failure of diagnostic tests to detect infection with COVID‐19 when it is present (a false negative result) may delay treatment and risk further spread of infection to others. Incorrect diagnosis of COVID‐19 when it is not present (a false positive result) may lead to unnecessary further testing, treatment, and isolation of the person and close contacts. Accurate identification of people who have previously had COVID‐19 is important in measuring disease spread and assessing the success of public health interventions. To determine the accuracy of an antibody test in identifying COVID‐19, test results are compared in people known to have (or have had) COVID‐19 and in people known not to have (or have had) COVID‐19. The criteria used to determine whether people are known or not known to have COVID‐19 is called the ‘reference standard’. Many studies use a test called reverse transcriptase polymerase chain reaction (RT‐PCR) as the reference standard, with samples taken from the nose and throat. Additional tests that can be used include measuring symptoms, like coughing or high temperature, or ‘imaging’ tests like chest X‐rays. People known not to have COVID‐19 are sometimes identified from stored blood samples taken before COVID‐19 existed, or from patients with symptoms confirmed to be caused by other diseases. What did the review study? We wanted to find out whether antibody tests: ‐ are able to diagnose infection in people with or without symptoms of COVID‐19, and ‐ can be used to find out if someone has already had COVID‐19. The studies we included in our review looked at three types of antibodies. Most commonly, antibody tests measure two types known as IgG and IgM, but some tests only measure a single type of antibody or different combinations of the three types of antibodies (IgA, IgG, IgM). What did we do? We looked for studies that measured the diagnostic accuracy of antibody tests to detect current or past COVID‐19 infection and compared them with reference standard criteria. Since there are many antibody tests available, we included studies assessing any antibody test compared with any reference standard. People could be tested in hospital or in the community. The people tested may have been confirmed to have, or not to have, COVID‐19 infection, or they may be suspected of having COVID‐19. Study characteristics We found 178 relevant studies. Studies took place in Europe (94), Asia (45), North America (35), Australia (2), and South America (2). Seventy‐eight studies included people who were in hospital with suspected or confirmed COVID‐19 infection and 14 studies included people in community settings. Several studies included people from multiple settings (35) or did not report where the participants were recruited from (39). One hundred and forty‐one studies included recent infection cases (mainly week 1 to week 3 after onset of symptoms), and many also included people tested later (from day 21 onwards after infection) (117). Main results In participants that had COVID‐19 and were tested one week after symptoms developed, antibody tests detected only 27% to 41% of infections. In week 2 after first symptoms, 64% to 79% of infections were detected, rising to 78% to 88% in week 3. Tests that specifically detected IgG or IgM antibodies were the most accurate and, when testing people from 21 days after first symptoms, they detected 93% of people with COVID‐19. Tests gave false positive results for 1% of those without COVID‐19. Below we illustrate results for two different scenarios. If 1000 people were tested for IgG or IgM antibodies during the third week after onset of symptoms and only 20 (2%) of them actually had COVID‐19: ‐ 26 people would test positive. Of these, 8 people (31%) would not have COVID‐19 (false positive result). ‐ 974 people would test negative. Of these, 2 people (0.2%) would actually have COVID‐19 (false negative result). If 1000 people with no symptoms for COVID‐19 were tested for IgG antibodies and 500 (50%) of them had previously had COVID‐19 infection more than 21 days previously: ‐ 455 people would test positive. Of these, 6 people (1%) would not have been infected (false positive result). ‐ 545 people would test negative. Of these, 51 (9%) would actually have had a prior COVID‐19 infection (false negative result). How reliable were the results of the studies of this review? We have limited confidence in the evidence for several reasons. The number of samples contributed by studies for each week post‐symptom onset was often small, and there were sometimes problems with how studies were conducted. Participants included in the studies were often hospital patients who were more likely to have experienced severe symptoms of COVID‐19. The accuracy of antibody tests for detecting COVID‐19 in these patients may be different from the accuracy of the tests in people with mild or moderate symptoms. It is not possible to identify by how much the test results would differ in other populations. Who do the results of this review apply to? A high percentage of participants were in hospital with COVID‐19, so were likely to have more severe disease than people with COVID‐19 who were not hospitalised. Only a small number of studies assessed these tests in people with no symptoms. The results of the review may therefore be more applicable to those with severe disease than people with mild symptoms. Studies frequently did not report whether participants had symptoms at the time samples were taken for testing making it difficult to fully separate test results for early‐phase infection as opposed to later‐phase infections. The studies in our review assessed several test methods across a global population, therefore it is likely that test results would be similar in most areas of the world. What are the implications of this review? The review shows that antibody tests could have a useful role in detecting if someone has had COVID‐19, but the timing of test use is important. Some antibody tests may help to confirm COVID‐19 infection in people who have had symptoms for more than two weeks but who have been unable to confirm their infection using other methods. This is particularly useful if they are experiencing potentially serious symptoms that may be due to COVID‐19 as they may require specific treatment. Antibody tests may also be useful to determine how many people have had a previous COVID‐19 infection. We could not be certain about how well the tests work for people who have milder disease or no symptoms, or for detecting antibodies resulting from vaccination. How up‐to‐date is this review? This review updates our previous review. The evidence is up‐to‐date to September 2020. [Infectious Diseases] Abstract - Background The diagnostic challenges associated with the COVID‐19 pandemic resulted in rapid development of diagnostic test methods for detecting SARS‐CoV‐2 infection. Serology tests to detect the presence of antibodies to SARS‐CoV‐2 enable detection of past infection and may detect cases of SARS‐CoV‐2 infection that were missed by earlier diagnostic tests. Understanding the diagnostic accuracy of serology tests for SARS‐CoV‐2 infection may enable development of effective diagnostic and management pathways, inform public health management decisions and understanding of SARS‐CoV‐2 epidemiology. Objectives To assess the accuracy of antibody tests, firstly, to determine if a person presenting in the community, or in primary or secondary care has current SARS‐CoV‐2 infection according to time after onset of infection and, secondly, to determine if a person has previously been infected with SARS‐CoV‐2. Sources of heterogeneity investigated included: timing of test, test method, SARS‐CoV‐2 antigen used, test brand, and reference standard for non‐SARS‐CoV‐2 cases. Search methods The COVID‐19 Open Access Project living evidence database from the University of Bern (which includes daily updates from PubMed and Embase and preprints from medRxiv and bioRxiv) was searched on 30 September 2020. We included additional publications from the Evidence for Policy and Practice Information and Co‐ordinating Centre (EPPI‐Centre) ‘COVID‐19: Living map of the evidence’ and the Norwegian Institute of Public Health ’NIPH systematic and living map on COVID‐19 evidence’. We did not apply language restrictions. Selection criteria We included test accuracy studies of any design that evaluated commercially produced serology tests, targeting IgG, IgM, IgA alone, or in combination. Studies must have provided data for sensitivity, that could be allocated to a predefined time period after onset of symptoms, or after a positive RT‐PCR test. Small studies with fewer than 25 SARS‐CoV‐2 infection cases were excluded. We included any reference standard to define the presence or absence of SARS‐CoV‐2 (including reverse transcription polymerase chain reaction tests (RT‐PCR), clinical diagnostic criteria, and pre‐pandemic samples). Data collection and analysis We use standard screening procedures with three reviewers. Quality assessment (using the QUADAS‐2 tool) and numeric study results were extracted independently by two people. Other study characteristics were extracted by one reviewer and checked by a second. We present sensitivity and specificity with 95% confidence intervals (CIs) for each test and, for meta‐analysis, we fitted univariate random‐effects logistic regression models for sensitivity by eligible time period and for specificity by reference standard group. Heterogeneity was investigated by including indicator variables in the random‐effects logistic regression models. We tabulated results by test manufacturer and summarised results for tests that were evaluated in 200 or more samples and that met a modification of UK Medicines and Healthcare products Regulatory Agency (MHRA) target performance criteria. Main results We included 178 separate studies (described in 177 study reports, with 45 as pre‐prints) providing 527 test evaluations. The studies included 64,688 samples including 25,724 from people with confirmed SARS‐CoV‐2; most compared the accuracy of two or more assays (102/178, 57%). Participants with confirmed SARS‐CoV‐2 infection were most commonly hospital inpatients (78/178, 44%), and pre‐pandemic samples were used by 45% (81/178) to estimate specificity. Over two‐thirds of studies recruited participants based on known SARS‐CoV‐2 infection status (123/178, 69%). All studies were conducted prior to the introduction of SARS‐CoV‐2 vaccines and present data for naturally acquired antibody responses. Seventy‐nine percent (141/178) of studies reported sensitivity by week after symptom onset and 66% (117/178) for convalescent phase infection. Studies evaluated enzyme‐linked immunosorbent assays (ELISA) (165/527; 31%), chemiluminescent assays (CLIA) (167/527; 32%) or lateral flow assays (LFA) (188/527; 36%). Risk of bias was high because of participant selection (172, 97%); application and interpretation of the index test (35, 20%); weaknesses in the reference standard (38, 21%); and issues related to participant flow and timing (148, 82%). We judged that there were high concerns about the applicability of the evidence related to participants in 170 (96%) studies, and about the applicability of the reference standard in 162 (91%) studies. Average sensitivities for current SARS‐CoV‐2 infection increased by week after onset for all target antibodies. Average sensitivity for the combination of either IgG or IgM was 41.1% in week one (95% CI 38.1 to 44.2; 103 evaluations; 3881 samples, 1593 cases), 74.9% in week two (95% CI 72.4 to 77.3; 96 evaluations, 3948 samples, 2904 cases) and 88.0% by week three after onset of symptoms (95% CI 86.3 to 89.5; 103 evaluations, 2929 samples, 2571 cases). Average sensitivity during the convalescent phase of infection (up to a maximum of 100 days since onset of symptoms, where reported) was 89.8% for IgG (95% CI 88.5 to 90.9; 253 evaluations, 16,846 samples, 14,183 cases), 92.9% for IgG or IgM combined (95% CI 91.0 to 94.4; 108 evaluations, 3571 samples, 3206 cases) and 94.3% for total antibodies (95% CI 92.8 to 95.5; 58 evaluations, 7063 samples, 6652 cases). Average sensitivities for IgM alone followed a similar pattern but were of a lower test accuracy in every time slot. Average specificities were consistently high and precise, particularly for pre‐pandemic samples which provide the least biased estimates of specificity (ranging from 98.6% for IgM to 99.8% for total antibodies). Subgroup analyses suggested small differences in sensitivity and specificity by test technology however heterogeneity in study results, timing of sample collection, and smaller sample numbers in some groups made comparisons difficult. For IgG, CLIAs were the most sensitive (convalescent‐phase infection) and specific (pre‐pandemic samples) compared to both ELISAs and LFAs (P < 0.001 for differences across test methods). The antigen(s) used (whether from the Spike‐protein or nucleocapsid) appeared to have some effect on average sensitivity in the first weeks after onset but there was no clear evidence of an effect during convalescent‐phase infection. Investigations of test performance by brand showed considerable variation in sensitivity between tests, and in results between studies evaluating the same test. For tests that were evaluated in 200 or more samples, the lower bound of the 95% CI for sensitivity was 90% or more for only a small number of tests (IgG, n = 5; IgG or IgM, n = 1; total antibodies, n = 4). More test brands met the MHRA minimum criteria for specificity of 98% or above (IgG, n = 16; IgG or IgM, n = 5; total antibodies, n = 7). Seven assays met the specified criteria for both sensitivity and specificity. In a low‐prevalence (2%) setting, where antibody testing is used to diagnose COVID‐19 in people with symptoms but who have had a negative PCR test, we would anticipate that 1 (1 to 2) case would be missed and 8 (5 to 15) would be falsely positive in 1000 people undergoing IgG or IgM testing in week three after onset of SARS‐CoV‐2 infection. In a seroprevalence survey, where prevalence of prior infection is 50%, we would anticipate that 51 (46 to 58) cases would be missed and 6 (5 to 7) would be falsely positive in 1000 people having IgG tests during the convalescent phase (21 to 100 days post‐symptom onset or post‐positive PCR) of SARS‐CoV‐2 infection. Authors' conclusions Some antibody tests could be a useful diagnostic tool for those in whom molecular‐ or antigen‐based tests have failed to detect the SARS‐CoV‐2 virus, including in those with ongoing symptoms of acute infection (from week three onwards) or those presenting with post‐acute sequelae of COVID‐19. However, antibody tests have an increasing likelihood of detecting an immune response to infection as time since onset of infection progresses and have demonstrated adequate performance for detection of prior infection for sero‐epidemiological purposes. The applicability of results for detection of vaccination‐induced antibodies is uncertain. Plain language summary What is the diagnostic accuracy of antibody tests for the detection of infection with the COVID‐19 virus? Background COVID‐19 is an infectious disease caused by the SARS‐CoV‐2 virus that spreads easily between people in a similar way to the common cold or ‘flu’. Most people with COVID‐19 have a mild‐to‐moderate respiratory illness, and some may have no symptoms (asymptomatic infection). Others experience severe symptoms and need specialist treatment and intensive care. In response to COVID‐19 infection, the immune system develops proteins called antibodies that can attack the virus as it circulates in their blood. People who have been vaccinated against COVID‐19 also produce these antibodies against the virus. Tests are available to detect antibodies in peoples' blood, which may indicate that they currently have COVID‐19 or have had it previously, or it may indicate that they have been vaccinated (although this group was not the focus of this review). Why are accurate tests important? Accurate testing allows identification of people who need to isolate themselves to prevent the spread of infection, or who might need treatment for their infection. Failure of diagnostic tests to detect infection with COVID‐19 when it is present (a false negative result) may delay treatment and risk further spread of infection to others. Incorrect diagnosis of COVID‐19 when it is not present (a false positive result) may lead to unnecessary further testing, treatment, and isolation of the person and close contacts. Accurate identification of people who have previously had COVID‐19 is important in measuring disease spread and assessing the success of public health interventions. To determine the accuracy of an antibody test in identifying COVID‐19, test results are compared in people known to have (or have had) COVID‐19 and in people known not to have (or have had) COVID‐19. The criteria used to determine whether people are known or not known to have COVID‐19 is called the ‘reference standard’. Many studies use a test called reverse transcriptase polymerase chain reaction (RT‐PCR) as the reference standard, with samples taken from the nose and throat. Additional tests that can be used include measuring symptoms, like coughing or high temperature, or ‘imaging’ tests like chest X‐rays. People known not to have COVID‐19 are sometimes identified from stored blood samples taken before COVID‐19 existed, or from patients with symptoms confirmed to be caused by other diseases. What did the review study? We wanted to find out whether antibody tests: ‐ are able to diagnose infection in people with or without symptoms of COVID‐19, and ‐ can be used to find out if someone has already had COVID‐19. The studies we included in our review looked at three types of antibodies. Most commonly, antibody tests measure two types known as IgG and IgM, but some tests only measure a single type of antibody or different combinations of the three types of antibodies (IgA, IgG, IgM). What did we do? We looked for studies that measured the diagnostic accuracy of antibody tests to detect current or past COVID‐19 infection and compared them with reference standard criteria. Since there are many antibody tests available, we included studies assessing any antibody test compared with any reference standard. People could be tested in hospital or in the community. The people tested may have been confirmed to have, or not to have, COVID‐19 infection, or they may be suspected of having COVID‐19. Study characteristics We found 178 relevant studies. Studies took place in Europe (94), Asia (45), North America (35), Australia (2), and South America (2). Seventy‐eight studies included people who were in hospital with suspected or confirmed COVID‐19 infection and 14 studies included people in community settings. Several studies included people from multiple settings (35) or did not report where the participants were recruited from (39). One hundred and forty‐one studies included recent infection cases (mainly week 1 to week 3 after onset of symptoms), and many also included people tested later (from day 21 onwards after infection) (117). Main results In participants that had COVID‐19 and were tested one week after symptoms developed, antibody tests detected only 27% to 41% of infections. In week 2 after first symptoms, 64% to 79% of infections were detected, rising to 78% to 88% in week 3. Tests that specifically detected IgG or IgM antibodies were the most accurate and, when testing people from 21 days after first symptoms, they detected 93% of people with COVID‐19. Tests gave false positive results for 1% of those without COVID‐19. Below we illustrate results for two different scenarios. If 1000 people were tested for IgG or IgM antibodies during the third week after onset of symptoms and only 20 (2%) of them actually had COVID‐19: ‐ 26 people would test positive. Of these, 8 people (31%) would not have COVID‐19 (false positive result). ‐ 974 people would test negative. Of these, 2 people (0.2%) would actually have COVID‐19 (false negative result). If 1000 people with no symptoms for COVID‐19 were tested for IgG antibodies and 500 (50%) of them had previously had COVID‐19 infection more than 21 days previously: ‐ 455 people would test positive. Of these, 6 people (1%) would not have been infected (false positive result). ‐ 545 people would test negative. Of these, 51 (9%) would actually have had a prior COVID‐19 infection (false negative result). How reliable were the results of the studies of this review? We have limited confidence in the evidence for several reasons. The number of samples contributed by studies for each week post‐symptom onset was often small, and there were sometimes problems with how studies were conducted. Participants included in the studies were often hospital patients who were more likely to have experienced severe symptoms of COVID‐19. The accuracy of antibody tests for detecting COVID‐19 in these patients may be different from the accuracy of the tests in people with mild or moderate symptoms. It is not possible to identify by how much the test results would differ in other populations. Who do the results of this review apply to? A high percentage of participants were in hospital with COVID‐19, so were likely to have more severe disease than people with COVID‐19 who were not hospitalised. Only a small number of studies assessed these tests in people with no symptoms. The results of the review may therefore be more applicable to those with severe disease than people with mild symptoms. Studies frequently did not report whether participants had symptoms at the time samples were taken for testing making it difficult to fully separate test results for early‐phase infection as opposed to later‐phase infections. The studies in our review assessed several test methods across a global population, therefore it is likely that test results would be similar in most areas of the world. What are the implications of this review? The review shows that antibody tests could have a useful role in detecting if someone has had COVID‐19, but the timing of test use is important. Some antibody tests may help to confirm COVID‐19 infection in people who have had symptoms for more than two weeks but who have been unable to confirm their infection using other methods. This is particularly useful if they are experiencing potentially serious symptoms that may be due to COVID‐19 as they may require specific treatment. Antibody tests may also be useful to determine how many people have had a previous COVID‐19 infection. We could not be certain about how well the tests work for people who have milder disease or no symptoms, or for detecting antibodies resulting from vaccination. How up‐to‐date is this review? This review updates our previous review. The evidence is up‐to‐date to September 2020. ispartof: Cochrane Database of Systematic Reviews vol:11 issue:11 ispartof: location:England status: Published online
- Published
- 2020
- Full Text
- View/download PDF
5. Systematic review of imaging tests to predict the development of rheumatoid arthritis in people with unclassified arthritis
- Author
-
de Pablo, P, primary, Dinnes, J, additional, Berhane, S, additional, Osman, A, additional, Lim, Z, additional, Coombe, A, additional, Raza, K, additional, Filer, A, additional, and Deeks, JJ, additional
- Published
- 2021
- Full Text
- View/download PDF
6. Total body photography for the diagnosis of cutaneous melanoma in adults: a systematic review and meta‐analysis*
- Author
-
Ji‐Xu, A., primary, Dinnes, J., additional, and Matin, R.N., additional
- Published
- 2021
- Full Text
- View/download PDF
7. Sensitivity and specificity of SkinVision are likely to have been overestimated
- Author
-
Deeks, J.J., primary, Dinnes, J., additional, and Williams, H.C., additional
- Published
- 2020
- Full Text
- View/download PDF
8. Algorithm based smartphone apps to assess risk of skin cancer in adults: systematic review of diagnostic accuracy studies
- Author
-
Freeman, K, Dinnes, J, Chuchu, N, Takwoingi, Y, Bayliss, SE, Matin, RN, Jain, A, Walter, FM, Williams, HC, Deeks, JJ, Freeman, K, Dinnes, J, Chuchu, N, Takwoingi, Y, Bayliss, SE, Matin, RN, Jain, A, Walter, FM, Williams, HC, and Deeks, JJ
- Abstract
OBJECTIVE: To examine the validity and findings of studies that examine the accuracy of algorithm based smartphone applications ("apps") to assess risk of skin cancer in suspicious skin lesions. DESIGN: Systematic review of diagnostic accuracy studies. DATA SOURCES: Cochrane Central Register of Controlled Trials, MEDLINE, Embase, CINAHL, CPCI, Zetoc, Science Citation Index, and online trial registers (from database inception to 10 April 2019). ELIGIBILITY CRITERIA FOR SELECTING STUDIES: Studies of any design that evaluated algorithm based smartphone apps to assess images of skin lesions suspicious for skin cancer. Reference standards included histological diagnosis or follow-up, and expert recommendation for further investigation or intervention. Two authors independently extracted data and assessed validity using QUADAS-2 (Quality Assessment of Diagnostic Accuracy Studies 2 tool). Estimates of sensitivity and specificity were reported for each app. RESULTS: Nine studies that evaluated six different identifiable smartphone apps were included. Six verified results by using histology or follow-up (n=725 lesions), and three verified results by using expert recommendations (n=407 lesions). Studies were small and of poor methodological quality, with selective recruitment, high rates of unevaluable images, and differential verification. Lesion selection and image acquisition were performed by clinicians rather than smartphone users. Two CE (Conformit Europenne) marked apps are available for download. SkinScan was evaluated in a single study (n=15, five melanomas) with 0% sensitivity and 100% specificity for the detection of melanoma. SkinVision was evaluated in two studies (n=252, 61 malignant or premalignant lesions) and achieved a sensitivity of 80% (95% confidence interval 63% to 92%) and a specificity of 78% (67% to 87%) for the detection of malignant or premalignant lesions. Accuracy of the SkinVision app verified against expert recommendations was poor (three studie
- Published
- 2020
9. Methods for the evaluation of biomarkers in patients with kidney and liver diseases: multicentre research programme including ELUCIDATE RCT
- Author
-
Selby, PJ, Banks, RE, Gregory, W, Hewison, J, Rosenberg, W, Altman, DG, Deeks, JJ, McCabe, C, Parkes, J, Sturgeon, C, Thompson, D, Twiddy, M, Bestall, J, Bedlington, J, Hale, T, Dinnes, J, Jones, M, Lewington, A, Messenger, MP, Napp, V, Sitch, A, Tanwar, S, Vasudev, NS, Baxter, P, Bell, S, Cairns, DA, Calder, N, Corrigan, N, Del Galdo, F, Heudtlass, P, Hornigold, N, Hulme, C, Hutchinson, M, Lippiatt, C, Livingstone, T, Longo, R, Potton, M, Roberts, S, Sim, S, Trainor, S, Welberry Smith, M, Neuberger, J, Thorburn, D, Richardson, P, Christie, J, Sheerin, N, McKane, W, Gibbs, P, Edwards, A, Soomro, N, Adeyoju, A, Stewart, GD, and Hrouda, D
- Abstract
Background: Protein biomarkers with associations with the activity and outcomes of diseases are being identified by modern proteomic technologies. They may be simple, accessible, cheap and safe tests that can inform diagnosis, prognosis, treatment selection, monitoring of disease activity and therapy and may substitute for complex, invasive and expensive tests. However, their potential is not yet being realised. Design and methods: The study consisted of three workstreams to create a framework for research: workstream 1, methodology – to define current practice and explore methodology innovations for biomarkers for monitoring disease; workstream 2, clinical translation – to create a framework of research practice, high-quality samples and related clinical data to evaluate the validity and clinical utility of protein biomarkers; and workstream 3, the ELF to Uncover Cirrhosis as an Indication for Diagnosis and Action for Treatable Event (ELUCIDATE) randomised controlled trial (RCT) – an exemplar RCT of an established test, the ADVIA Centaur® Enhanced Liver Fibrosis (ELF) test (Siemens Healthcare Diagnostics Ltd, Camberley, UK) [consisting of a panel of three markers – (1) serum hyaluronic acid, (2) amino-terminal propeptide of type III procollagen and (3) tissue inhibitor of metalloproteinase 1], for liver cirrhosis to determine its impact on diagnostic timing and the management of cirrhosis and the process of care and improving outcomes. Results: The methodology workstream evaluated the quality of recommendations for using prostate-specific antigen to monitor patients, systematically reviewed RCTs of monitoring strategies and reviewed the monitoring biomarker literature and how monitoring can have an impact on outcomes. Simulation studies were conducted to evaluate monitoring and improve the merits of health care. The monitoring biomarker literature is modest and robust conclusions are infrequent. We recommend improvements in research practice. Patients strongly endorsed the need for robust and conclusive research in this area. The clinical translation workstream focused on analytical and clinical validity. Cohorts were established for renal cell carcinoma (RCC) and renal transplantation (RT), with samples and patient data from multiple centres, as a rapid-access resource to evaluate the validity of biomarkers. Candidate biomarkers for RCC and RT were identified from the literature and their quality was evaluated and selected biomarkers were prioritised. The duration of follow-up was a limitation but biomarkers were identified that may be taken forward for clinical utility. In the third workstream, the ELUCIDATE trial registered 1303 patients and randomised 878 patients out of a target of 1000. The trial started late and recruited slowly initially but ultimately recruited with good statistical power to answer the key questions. ELF monitoring altered the patient process of care and may show benefits from the early introduction of interventions with further follow-up. The ELUCIDATE trial was an ‘exemplar’ trial that has demonstrated the challenges of evaluating biomarker strategies in ‘end-to-end’ RCTs and will inform future study designs. Conclusions: The limitations in the programme were principally that, during the collection and curation of the cohorts of patients with RCC and RT, the pace of discovery of new biomarkers in commercial and non-commercial research was slower than anticipated and so conclusive evaluations using the cohorts are few; however, access to the cohorts will be sustained for future new biomarkers. The ELUCIDATE trial was slow to start and recruit to, with a late surge of recruitment, and so final conclusions about the impact of the ELF test on long-term outcomes await further follow-up. The findings from the three workstreams were used to synthesise a strategy and framework for future biomarker evaluations incorporating innovations in study design, health economics and health informatics.
- Published
- 2018
10. Effectiveness and cost-effectiveness of double reading of mammograms in breast cancer screening: findings of a systematic review
- Author
-
Dinnes, J., Moss, S., Melia, J., Blanks, R., Song, F., and Kleijnen, J.
- Published
- 2001
- Full Text
- View/download PDF
11. STARD for Abstracts : Essential items for reporting diagnostic accuracy studies in journal or conference abstracts
- Author
-
Cohen, Jf, Korevaar, Da, Gatsonis, Ca, Glasziou, Pp, Hooft, L, Moher, D, Reitsma, Jb, de Vet HC, Bossuyt, Pm, STARD Group: Alonzo, T, Altman, Dg, Azuara-Blanco, A, Bachmann, L, Blume, J, Boutron, I, Bruns, D, Büller, H, Buntinx, F, Byron, S, Chang, S, Cohen, J, Cooper, R, de Groot, J, de Vet HCW, Deeks, J, Dendukuri, N, Dinnes, J, Fleming, K, Glasziou, Pg, Golub, Rm, Guyatt, G, Heneghan, C, Hilden, J, Horvath, R, Hunink, M, Hyde, C, Ioannidis, J, Irwig, L, Janes, H, Kleijnen, J, Knottnerus, A, Kressel, Hy, Lange, S, Leeflang, M, Lijmer, Jg, Lord, S, Lumbreras, B, Macaskill, P, Magid, E, Mallett, S, Mcinnes, M, Mcneil, B, Mcqueen, M, Moons, K, Morris, K, Mustafa, R, Obuchowski, N, Ochodo, E, Onderdonk, A, Overbeke, J, Pai, N, Peeling, R, Pepe, M, Petersen, S, Price, C, Ravaud, P, Rennie, D, Rifai, N, Rutjes, A, Schunemann, H, Simel, D, Simera, I, Smidt, N, Steyerberg, E, Straus, S, Summerskill, W, Takwoingi, Y, Thompson, M, van den Bruel, A, van Maanen, H, Vickers, A, Virgili, G, Walter, S, Weber, W, Westwood, M, Whiting, P, Wilczynski, N, Ziegler, A., Epidemiology and Data Science, APH - Methodology, Epidemiology, Radiology & Nuclear Medicine, Erasmus MC other, Erasmus School of Health Policy & Management, Public Health, APH - Personalized Medicine, and Other departments
- Subjects
Medicine(all) ,medicine.medical_specialty ,Information retrieval ,business.industry ,MEDLINE ,Diagnostic accuracy ,General Medicine ,Executive committee ,030218 nuclear medicine & medical imaging ,03 medical and health sciences ,0302 clinical medicine ,Completion rate ,Medicine ,Medical physics ,030212 general & internal medicine ,business ,Web based survey - Abstract
Many abstracts of diagnostic accuracy studies are currently insufficiently informative. We extended the STARD (Standards for Reporting Diagnostic Accuracy) statement by developing a list of essential items that authors should consider when reporting diagnostic accuracy studies in journal or conference abstracts. After a literature review of published guidance for reporting biomedical studies, we identified 39 items potentially relevant to report in an abstract. We then selected essential items through a two round web based survey among the 85 members of the STARD Group, followed by discussions within an executive committee. Seventy three STARD Group members responded (86%), with 100% completion rate. STARD for Abstracts is a list of 11 quintessential items, to be reported in every abstract of a diagnostic accuracy study. We provide examples of complete reporting, and developed template text for writing informative abstracts.
- Published
- 2017
- Full Text
- View/download PDF
12. Erratum to:Methods for evaluating medical tests and biomarkers
- Author
-
Gopalakrishna, G, Langendam, M, Scholten, R, Bossuyt, P, Leeflang, M, Noel-Storr, A, Thomas, J, Marshall, I, Wallace, B, Whiting, P, Davenport, C, GopalaKrishna, G, De Salis, I, Mallett, S, Wolff, R, Riley, R, Westwood, M, Kleinen, J, Collins, G, Reitsma, H, Moons, K, Zapf, A, Hoyer, A, Kramer, K, Kuss, O, Ensor, J, Deeks, JJ, Martin, EC, Riley, RD, Rücker, G, Steinhauser, S, Schumacher, M, Snell, K, Willis, B, Debray, T, Deeks, J, Di Ruffano, LF, Taylor-Phillips, S, Hyde, C, Taylor, SA, Batnagar, G, STREAMLINE COLON Investigators, STREAMLINE LUNG Investigators, METRIC Investigators, Seedat, F, Clarke, A, Byron, S, Nixon, F, Albrow, R, Walker, T, Deakin, C, Zhelev, Z, Hunt, H, Yang, Y, Abel, L, Buchanan, J, Fanshawe, T, Shinkins, B, Wynants, L, Verbakel, J, Van Huffel, S, Timmerman, D, Van Calster, B, Zwinderman, A, Oke, J, O'Sullivan, J, Perera, R, Nicholson, B, Bromley, HL, Roberts, TE, Francis, A, Petrie, D, Mann, GB, Malottki, K, Smith, H, Billingham, L, Sitch, A, Gerke, O, Holm-Vilstrup, M, Segtnan, EA, Halekoh, U, Høilund-Carlsen, PF, Francq, BG, Dinnes, J, Parkes, J, Gregory, W, Hewison, J, Altman, D, Rosenberg, W, Selby, P, Asselineau, J, Perez, P, Paye, A, Bessede, E, Proust-Lima, C, Naaktgeboren, C, De Groot, J, Rutjes, A, Reitsma, J, Ogundimu, E, Cook, J, Le Manach, Y, Vergouwe, Y, Pajouheshnia, R, Groenwold, R, Peelen, L, Nieboer, D, De Cock, B, Pencina, MJ, Steyerberg, EW, Cooper, J, Parsons, N, Stinton, C, Smith, S, Dickens, A, Jordan, R, Enocson, A, Fitzmaurice, D, Adab, P, Boachie, C, Vidmar, G, Freeman, K, Connock, M, Court, R, Moons, C, Harris, J, Mumford, A, Plummer, Z, Lee, K, Reeves, B, Rogers, C, Verheyden, V, Angelini, GD, Murphy, GJ, Huddy, J, Ni, M, Good, K, Cooke, G, Hanna, G, Ma, J, Moons, KGMC, De Groot, JAH, Altman, DG, Reitsma, JB, Collins, GS, Moons, KGM, Kamarudin, AN, Kolamunnage-Dona, R, Cox, T, Borsci, S, Pérez, T, Pardo, MC, Candela-Toha, A, Muriel, A, Zamora, J, Sanghera, S, Mohiuddin, S, Martin, R, Donovan, J, Coast, J, Seo, MK, Cairns, J, Mitchell, E, Smith, A, Wright, J, Hall, P, Messenger, M, Calder, N, Wickramasekera, N, Vinall-Collier, K, Lewington, A, Damen, J, Cairns, D, Hutchinson, M, Sturgeon, C, Mitchel, L, Kift, R, Christakoudi, S, Rungall, M, Mobillo, P, Montero, R, Tsui, T-L, Kon, SP, Tucker, B, Sacks, S, Farmer, C, Strom, T, Chowdhury, P, Rebollo-Mesa, I, Hernandez-Fuentes, M, Damen, JAAG, Debray, TPA, Heus, P, Hooft, L, Scholten, RJPM, Schuit, E, Tzoulaki, I, Lassale, CM, Siontis, GCM, Chiocchia, V, Roberts, C, Schlüssel, MM, Gerry, S, Black, JA, Van der Schouw, YT, Peelen, LM, Spence, G, McCartney, D, Van den Bruel, A, Lasserson, D, Hayward, G, Vach, W, De Jong, A, Burggraaff, C, Hoekstra, O, Zijlstra, J, De Vet, H, Graziadio, S, Allen, J, Johnston, L, O'Leary, R, Power, M, Johnson, L, Waters, R, Simpson, J, Fanshawe, TR, Phillips, P, Plumb, A, Helbren, E, Halligan, S, Gale, A, Sekula, P, Sauerbrei, W, Forman, JR, Dutton, SJ, Takwoingi, Y, Hensor, EM, Nichols, TE, Kempf, E, Porcher, R, De Beyer, J, Hopewell, S, Dennis, J, Shields, B, Jones, A, Henley, W, Pearson, E, Hattersley, A, MASTERMIND consortium, Scheibler, F, Rummer, A, Sturtz, S, Großelfinger, R, Banister, K, Ramsay, C, Azuara-Blanco, A, Burr, J, Kumarasamy, M, Bourne, R, Uchegbu, I, Murphy, J, Carter, A, Marti, J, Eatock, J, Robotham, J, Dudareva, M, Gilchrist, M, Holmes, A, Monaghan, P, Lord, S, StJohn, A, Sandberg, S, Cobbaert, C, Lennartz, L, Verhagen-Kamerbeek, W, Ebert, C, Horvath, A, Test Evaluation Working Group of the European Federation of Clinical Chemistry and Laboratory Medicine, Jenniskens, K, Peters, J, Grigore, B, Ukoumunne, O, Levis, B, Benedetti, A, Levis, AW, Ioannidis, JPA, Shrier, I, Cuijpers, P, Gilbody, S, Kloda, LA, McMillan, D, Patten, SB, Steele, RJ, Ziegelstein, RC, Bombardier, CH, Osório, FDL, Fann, JR, Gjerdingen, D, Lamers, F, Lotrakul, M, Loureiro, SR, Löwe, B, Shaaban, J, Stafford, L, Van Weert, HCPM, Whooley, MA, Williams, LS, Wittkampf, KA, Yeung, AS, Thombs, BD, Cooper, C, Nieto, T, Smith, C, Tucker, O, Dretzke, J, Beggs, A, Rai, N, Bayliss, S, Stevens, S, Mallet, S, Sundar, S, Hall, E, Porta, N, Estelles, DL, De Bono, J, CTC-STOP protocol development group, and National Institute for Health Research
- Subjects
medicine.medical_specialty ,Astrophysics::High Energy Astrophysical Phenomena ,MEDLINE ,030204 cardiovascular system & hematology ,BTC (Bristol Trials Centre) ,MASTERMIND consortium ,03 medical and health sciences ,0302 clinical medicine ,medicine ,030212 general & internal medicine ,Intensive care medicine ,CTC-STOP protocol development group ,lcsh:R5-920 ,business.industry ,Test Evaluation Working Group of the European Federation of Clinical Chemistry and Laboratory Medicine ,Published Erratum ,STREAMLINE COLON Investigators ,3. Good health ,STREAMLINE LUNG Investigators ,Centre for Surgical Research ,Family medicine ,METRIC Investigators ,High Energy Physics::Experiment ,Erratum ,business ,lcsh:Medicine (General) - Abstract
[This corrects the article DOI: 10.1186/s41512-016-0001-y.].
- Published
- 2017
- Full Text
- View/download PDF
13. Establishing the use of total body photography among UK dermatologists.
- Author
-
Ji‐Xu, A., Dinnes, J., and Matin, R. N.
- Subjects
- *
DERMATOLOGISTS , *MELANOMA , *PHOTOGRAPHY , *SKIN cancer - Abstract
Although 38% of respondents (26 of 69) reported patients using smartphone apps (including MySkinSelfie and Miiskin) to monitor their lesions, only 19% (13 of 69) recommended apps to monitor skin lesions to their patients. Total body photography (TBP) is increasingly used to monitor skin lesions in individuals at high risk of melanoma. [Extracted from the article]
- Published
- 2022
- Full Text
- View/download PDF
14. STARD 2015 : an updated list of essential items for reporting diagnostic accuracy studies
- Author
-
Bossuyt, Pm, Reitsma, Jb, Bruns, De, Gatsonis, Ca, Glasziou, Pp, Irwig, L, Lijmer, Jg, Moher, D, Rennie, D, de Vet HCW, Kressel, Hy, Rifai, N, Golub, Rm, Altman, Dg, Hooft, L, Korevaar, Da, Cohen JF [Contributors: Alonzo, T, Azuara-Blanco, A, Bachmann, L, Blume, J, Boutron, I, Bruns, D, Büller, H, Buntinx, F, Byron, S, Chang, S, Cohen, Jf, Cooper, R, de Groot, J, Deeks, J, Dendukuri, N, Dinnes, J, Fleming, K, Guyatt, G, Heneghan, C, Hilden, J, Horvath, R, Hunink, M, Hyde, C, Ioannidis, J, Janes, H, Kleijnen, J, Knottnerus, A, Lange, S, Leeflang, M, Lord, S, Lumbreras, B, Macaskill, P, Magid, E, Mallett, S, Mcinnes, M, Mcneil, B, Mcqueen, M, Moons, K, Morris, K, Mustafa, R, Obuchowski, N, Ochodo, E, Onderdonk, A, Overbeke, J, Pai, N, Peeling, R, Pepe, M, Petersen, S, Price, C, Ravaud, P, Rutjes, A, Schunemann, H, Simel, D, Simera, I, Smidt, N, Steyerberg, E, Straus, S, Summerskill, W, Takwoingi, Y, Thompson, M, van de Bruel, A, van Maanen, H, Vickers, A, Virgili, G, Walter, S, Weber, W, Westwood, M, Whiting, P, Wilczynski, N, Ziegler, A, APH - Amsterdam Public Health, 10 Public Health & Methodologie, Other departments, Epidemiology and Data Science, ACS - Amsterdam Cardiovascular Sciences, Vascular Medicine, and EMGO - Musculoskeletal health
- Subjects
Quality Control ,Research design ,PRIMARY OUTCOMES ,medicine.medical_specialty ,Computer science ,RANDOMIZED CONTROLLED-TRIALS ,Clinical Biochemistry ,MEDLINE ,Diagnostic accuracy ,Disclosure ,GUIDELINES ,Research Support ,Data accuracy ,Terminology as Topic ,Journal Article ,Humans ,Research Methods & Reporting ,Medicine ,Radiology, Nuclear Medicine and imaging ,Medical physics ,Non-U.S. Gov't ,Reference standards ,Diagnostic Techniques and Procedures ,Bias (Epidemiology) ,UTILITY ,Diagnostic Tests, Routine ,Information Dissemination ,business.industry ,STATEMENT ,Research Support, Non-U.S. Gov't ,Biochemistry (medical) ,Reproducibility of Results ,Diagnostic test ,General Medicine ,Reference Standards ,Data Accuracy ,TRANSPARENT ,Critical appraisal ,EQUATOR ,BIAS ,Research Design ,Practice Guidelines as Topic ,TESTS ,business - Abstract
Incomplete reporting has been identified as a major source of avoidable waste in biomedical research. Essential information is often not provided in study reports, impeding the identification, critical appraisal, and replication of studies. To improve the quality of reporting of diagnostic accuracy studies, the Standards for Reporting of Diagnostic Accuracy Studies (STARD) statement was developed. Here we present STARD 2015, an updated list of 30 essential items that should be included in every report of a diagnostic accuracy study. This update incorporates recent evidence about sources of bias and variability in diagnostic accuracy and is intended to facilitate the use of STARD. As such, STARD 2015 may help to improve completeness and transparency in reporting of diagnostic accuracy studies.
- Published
- 2015
- Full Text
- View/download PDF
15. Erratum to: Methods for evaluating medical tests and biomarkers.
- Author
-
Gopalakrishna, G, Langendam, M, Scholten, R, Bossuyt, P, Leeflang, M, Noel-Storr, A, Thomas, J, Marshall, I, Wallace, B, Whiting, P, Davenport, C, GopalaKrishna, G, de Salis, I, Mallett, S, Wolff, R, Riley, R, Westwood, M, Kleinen, J, Collins, G, Reitsma, H, Moons, K, Zapf, A, Hoyer, A, Kramer, K, Kuss, O, Ensor, J, Deeks, JJ, Martin, EC, Riley, RD, Rücker, G, Steinhauser, S, Schumacher, M, Snell, K, Willis, B, Debray, T, Deeks, J, di Ruffano, LF, Taylor-Phillips, S, Hyde, C, Taylor, SA, Batnagar, G, STREAMLINE COLON Investigators, STREAMLINE LUNG Investigators, METRIC Investigators, Di Ruffano, LF, Seedat, F, Clarke, A, Byron, S, Nixon, F, Albrow, R, Walker, T, Deakin, C, Zhelev, Z, Hunt, H, Yang, Y, Abel, L, Buchanan, J, Fanshawe, T, Shinkins, B, Wynants, L, Verbakel, J, Van Huffel, S, Timmerman, D, Van Calster, B, Zwinderman, A, Oke, J, O'Sullivan, J, Perera, R, Nicholson, B, Bromley, HL, Roberts, TE, Francis, A, Petrie, D, Mann, GB, Malottki, K, Smith, H, Billingham, L, Sitch, A, Gerke, O, Holm-Vilstrup, M, Segtnan, EA, Halekoh, U, Høilund-Carlsen, PF, Francq, BG, Dinnes, J, Parkes, J, Gregory, W, Hewison, J, Altman, D, Rosenberg, W, Selby, P, Asselineau, J, Perez, P, Paye, A, Bessede, E, Proust-Lima, C, Naaktgeboren, C, de Groot, J, Rutjes, A, Reitsma, J, Ogundimu, E, Cook, J, Le Manach, Y, Vergouwe, Y, Pajouheshnia, R, Groenwold, R, Peelen, L, Nieboer, D, De Cock, B, Pencina, MJ, Steyerberg, EW, Cooper, J, Parsons, N, Stinton, C, Smith, S, Dickens, A, Jordan, R, Enocson, A, Fitzmaurice, D, Adab, P, Boachie, C, Vidmar, G, Freeman, K, Connock, M, Court, R, Moons, C, Harris, J, Mumford, A, Plummer, Z, Lee, K, Reeves, B, Rogers, C, Verheyden, V, Angelini, GD, Murphy, GJ, Huddy, J, Ni, M, Good, K, Cooke, G, Hanna, G, Ma, J, Moons, KGMC, de Groot, JAH, Altman, DG, Reitsma, JB, Collins, GS, Moons, KGM, Kamarudin, AN, Kolamunnage-Dona, R, Cox, T, Borsci, S, Pérez, T, Pardo, MC, Candela-Toha, A, Muriel, A, Zamora, J, Sanghera, S, Mohiuddin, S, Martin, R, Donovan, J, Coast, J, Seo, MK, Cairns, J, Mitchell, E, Smith, A, Wright, J, Hall, P, Messenger, M, Calder, N, Wickramasekera, N, Vinall-Collier, K, Lewington, A, Damen, J, Cairns, D, Hutchinson, M, Sturgeon, C, Mitchel, L, Kift, R, Christakoudi, S, Rungall, M, Mobillo, P, Montero, R, Tsui, T-L, Kon, SP, Tucker, B, Sacks, S, Farmer, C, Strom, T, Chowdhury, P, Rebollo-Mesa, I, Hernandez-Fuentes, M, Damen, JAAG, Debray, TPA, Heus, P, Hooft, L, Scholten, RJPM, Schuit, E, Tzoulaki, I, Lassale, CM, Siontis, GCM, Chiocchia, V, Roberts, C, Schlüssel, MM, Gerry, S, Black, JA, van der Schouw, YT, Peelen, LM, Spence, G, McCartney, D, van den Bruel, A, Lasserson, D, Hayward, G, Vach, W, de Jong, A, Burggraaff, C, Hoekstra, O, Zijlstra, J, de Vet, H, Graziadio, S, Allen, J, Johnston, L, O'Leary, R, Power, M, Johnson, L, Waters, R, Simpson, J, Fanshawe, TR, Phillips, P, Plumb, A, Helbren, E, Halligan, S, Gale, A, Sekula, P, Sauerbrei, W, Forman, JR, Dutton, SJ, Takwoingi, Y, Hensor, EM, Nichols, TE, Kempf, E, Porcher, R, de Beyer, J, Hopewell, S, Dennis, J, Shields, B, Jones, A, Henley, W, Pearson, E, Hattersley, A, MASTERMIND consortium, Scheibler, F, Rummer, A, Sturtz, S, Großelfinger, R, Banister, K, Ramsay, C, Azuara-Blanco, A, Burr, J, Kumarasamy, M, Bourne, R, Uchegbu, I, Murphy, J, Carter, A, Marti, J, Eatock, J, Robotham, J, Dudareva, M, Gilchrist, M, Holmes, A, Monaghan, P, Lord, S, StJohn, A, Sandberg, S, Cobbaert, C, Lennartz, L, Verhagen-Kamerbeek, W, Ebert, C, Horvath, A, Test Evaluation Working Group of the European Federation of Clinical Chemistry and Laboratory Medicine, Jenniskens, K, Peters, J, Grigore, B, Ukoumunne, O, Levis, B, Benedetti, A, Levis, AW, Ioannidis, JPA, Shrier, I, Cuijpers, P, Gilbody, S, Kloda, LA, McMillan, D, Patten, SB, Steele, RJ, Ziegelstein, RC, Bombardier, CH, Osório, FDL, Fann, JR, Gjerdingen, D, Lamers, F, Lotrakul, M, Loureiro, SR, Löwe, B, Shaaban, J, Stafford, L, van Weert, HCPM, Whooley, MA, Williams, LS, Wittkampf, KA, Yeung, AS, Thombs, BD, Cooper, C, Nieto, T, Smith, C, Tucker, O, Dretzke, J, Beggs, A, Rai, N, Bayliss, S, Stevens, S, Mallet, S, Sundar, S, Hall, E, Porta, N, Estelles, DL, de Bono, J, CTC-STOP protocol development group, Gopalakrishna, G, Langendam, M, Scholten, R, Bossuyt, P, Leeflang, M, Noel-Storr, A, Thomas, J, Marshall, I, Wallace, B, Whiting, P, Davenport, C, GopalaKrishna, G, de Salis, I, Mallett, S, Wolff, R, Riley, R, Westwood, M, Kleinen, J, Collins, G, Reitsma, H, Moons, K, Zapf, A, Hoyer, A, Kramer, K, Kuss, O, Ensor, J, Deeks, JJ, Martin, EC, Riley, RD, Rücker, G, Steinhauser, S, Schumacher, M, Snell, K, Willis, B, Debray, T, Deeks, J, di Ruffano, LF, Taylor-Phillips, S, Hyde, C, Taylor, SA, Batnagar, G, STREAMLINE COLON Investigators, STREAMLINE LUNG Investigators, METRIC Investigators, Di Ruffano, LF, Seedat, F, Clarke, A, Byron, S, Nixon, F, Albrow, R, Walker, T, Deakin, C, Zhelev, Z, Hunt, H, Yang, Y, Abel, L, Buchanan, J, Fanshawe, T, Shinkins, B, Wynants, L, Verbakel, J, Van Huffel, S, Timmerman, D, Van Calster, B, Zwinderman, A, Oke, J, O'Sullivan, J, Perera, R, Nicholson, B, Bromley, HL, Roberts, TE, Francis, A, Petrie, D, Mann, GB, Malottki, K, Smith, H, Billingham, L, Sitch, A, Gerke, O, Holm-Vilstrup, M, Segtnan, EA, Halekoh, U, Høilund-Carlsen, PF, Francq, BG, Dinnes, J, Parkes, J, Gregory, W, Hewison, J, Altman, D, Rosenberg, W, Selby, P, Asselineau, J, Perez, P, Paye, A, Bessede, E, Proust-Lima, C, Naaktgeboren, C, de Groot, J, Rutjes, A, Reitsma, J, Ogundimu, E, Cook, J, Le Manach, Y, Vergouwe, Y, Pajouheshnia, R, Groenwold, R, Peelen, L, Nieboer, D, De Cock, B, Pencina, MJ, Steyerberg, EW, Cooper, J, Parsons, N, Stinton, C, Smith, S, Dickens, A, Jordan, R, Enocson, A, Fitzmaurice, D, Adab, P, Boachie, C, Vidmar, G, Freeman, K, Connock, M, Court, R, Moons, C, Harris, J, Mumford, A, Plummer, Z, Lee, K, Reeves, B, Rogers, C, Verheyden, V, Angelini, GD, Murphy, GJ, Huddy, J, Ni, M, Good, K, Cooke, G, Hanna, G, Ma, J, Moons, KGMC, de Groot, JAH, Altman, DG, Reitsma, JB, Collins, GS, Moons, KGM, Kamarudin, AN, Kolamunnage-Dona, R, Cox, T, Borsci, S, Pérez, T, Pardo, MC, Candela-Toha, A, Muriel, A, Zamora, J, Sanghera, S, Mohiuddin, S, Martin, R, Donovan, J, Coast, J, Seo, MK, Cairns, J, Mitchell, E, Smith, A, Wright, J, Hall, P, Messenger, M, Calder, N, Wickramasekera, N, Vinall-Collier, K, Lewington, A, Damen, J, Cairns, D, Hutchinson, M, Sturgeon, C, Mitchel, L, Kift, R, Christakoudi, S, Rungall, M, Mobillo, P, Montero, R, Tsui, T-L, Kon, SP, Tucker, B, Sacks, S, Farmer, C, Strom, T, Chowdhury, P, Rebollo-Mesa, I, Hernandez-Fuentes, M, Damen, JAAG, Debray, TPA, Heus, P, Hooft, L, Scholten, RJPM, Schuit, E, Tzoulaki, I, Lassale, CM, Siontis, GCM, Chiocchia, V, Roberts, C, Schlüssel, MM, Gerry, S, Black, JA, van der Schouw, YT, Peelen, LM, Spence, G, McCartney, D, van den Bruel, A, Lasserson, D, Hayward, G, Vach, W, de Jong, A, Burggraaff, C, Hoekstra, O, Zijlstra, J, de Vet, H, Graziadio, S, Allen, J, Johnston, L, O'Leary, R, Power, M, Johnson, L, Waters, R, Simpson, J, Fanshawe, TR, Phillips, P, Plumb, A, Helbren, E, Halligan, S, Gale, A, Sekula, P, Sauerbrei, W, Forman, JR, Dutton, SJ, Takwoingi, Y, Hensor, EM, Nichols, TE, Kempf, E, Porcher, R, de Beyer, J, Hopewell, S, Dennis, J, Shields, B, Jones, A, Henley, W, Pearson, E, Hattersley, A, MASTERMIND consortium, Scheibler, F, Rummer, A, Sturtz, S, Großelfinger, R, Banister, K, Ramsay, C, Azuara-Blanco, A, Burr, J, Kumarasamy, M, Bourne, R, Uchegbu, I, Murphy, J, Carter, A, Marti, J, Eatock, J, Robotham, J, Dudareva, M, Gilchrist, M, Holmes, A, Monaghan, P, Lord, S, StJohn, A, Sandberg, S, Cobbaert, C, Lennartz, L, Verhagen-Kamerbeek, W, Ebert, C, Horvath, A, Test Evaluation Working Group of the European Federation of Clinical Chemistry and Laboratory Medicine, Jenniskens, K, Peters, J, Grigore, B, Ukoumunne, O, Levis, B, Benedetti, A, Levis, AW, Ioannidis, JPA, Shrier, I, Cuijpers, P, Gilbody, S, Kloda, LA, McMillan, D, Patten, SB, Steele, RJ, Ziegelstein, RC, Bombardier, CH, Osório, FDL, Fann, JR, Gjerdingen, D, Lamers, F, Lotrakul, M, Loureiro, SR, Löwe, B, Shaaban, J, Stafford, L, van Weert, HCPM, Whooley, MA, Williams, LS, Wittkampf, KA, Yeung, AS, Thombs, BD, Cooper, C, Nieto, T, Smith, C, Tucker, O, Dretzke, J, Beggs, A, Rai, N, Bayliss, S, Stevens, S, Mallet, S, Sundar, S, Hall, E, Porta, N, Estelles, DL, de Bono, J, and CTC-STOP protocol development group
- Abstract
[This corrects the article DOI: 10.1186/s41512-016-0001-y.].
- Published
- 2017
16. Development and validation of methods for assessing the quality of diagnostic accuracy studies
- Author
-
Whiting, P, Rutjes, A, Dinnes, J, Reitsma, J, Bossuyt, Pm, and Kleijnen, J
- Published
- 2004
17. Evaluating non-randomised intervention studies
- Author
-
Jonathan Deeks, Dinnes, J., D Amico, R., Sowden, A. J., Sakarovitch, C., Song, F., Petticrew, M., Altman, D. G., International Stroke Trial Collaborative Group, and European Carotid Surgery Trial Collaborative Group
- Abstract
OBJECTIVES: To consider methods and related evidence for evaluating bias in non-randomised intervention studies. DATA SOURCES: Systematic reviews and methodological papers were identified from a search of electronic databases; handsearches of key medical journals and contact with experts working in the field. New empirical studies were conducted using data from two large randomised clinical trials. METHODS: Three systematic reviews and new empirical investigations were conducted. The reviews considered, in regard to non-randomised studies, (1) the existing evidence of bias, (2) the content of quality assessment tools, (3) the ways that study quality has been assessed and addressed. (4) The empirical investigations were conducted generating non-randomised studies from two large, multicentre randomised controlled trials (RCTs) and selectively resampling trial participants according to allocated treatment, centre and period. RESULTS: In the systematic reviews, eight studies compared results of randomised and non-randomised studies across multiple interventions using meta-epidemiological techniques. A total of 194 tools were identified that could be or had been used to assess non-randomised studies. Sixty tools covered at least five of six pre-specified internal validity domains. Fourteen tools covered three of four core items of particular importance for non-randomised studies. Six tools were thought suitable for use in systematic reviews. Of 511 systematic reviews that included non-randomised studies, only 169 (33%) assessed study quality. Sixty-nine reviews investigated the impact of quality on study results in a quantitative manner. The new empirical studies estimated the bias associated with non-random allocation and found that the bias could lead to consistent over- or underestimations of treatment effects, also the bias increased variation in results for both historical and concurrent controls, owing to haphazard differences in case-mix between groups. The biases were large enough to lead studies falsely to conclude significant findings of benefit or harm. Four strategies for case-mix adjustment were evaluated: none adequately adjusted for bias in historically and concurrently controlled studies. Logistic regression on average increased bias. Propensity score methods performed better, but were not satisfactory in most situations. Detailed investigation revealed that adequate adjustment can only be achieved in the unrealistic situation when selection depends on a single factor. CONCLUSIONS: Results of non-randomised studies sometimes, but not always, differ from results of randomised studies of the same intervention. Non-randomised studies may still give seriously misleading results when treated and control groups appear similar in key prognostic factors. Standard methods of case-mix adjustment do not guarantee removal of bias. Residual confounding may be high even when good prognostic data are available, and in some situations adjusted results may appear more biased than unadjusted results. Although many quality assessment tools exist and have been used for appraising non-randomised studies, most omit key quality domains. Healthcare policies based upon non-randomised studies or systematic reviews of non-randomised studies may need re-evaluation if the uncertainty in the true evidence base was not fully appreciated when policies were made. The inability of case-mix adjustment methods to compensate for selection bias and our inability to identify non-randomised studies that are free of selection bias indicate that non-randomised studies should only be undertaken when RCTs are infeasible or unethical. Recommendations for further research include: applying the resampling methodology in other clinical areas to ascertain whether the biases described are typical; developing or refining existing quality assessment tools for non-randomised studies; investigating how quality assessments of non-randomised studies can be incorporated into reviews and the implications of individual quality features for interpretation of a review's results; examination of the reasons for the apparent failure of case-mix adjustment methods; and further evaluation of the role of the propensity score.
- Published
- 2003
18. Evaluating non-randomised intervention studies
- Author
-
Deeks, Jj, Dinnes, J, D'Amico, Roberto, Sowden, Aj, Sakarovitch, C, Song, F, Petticrew, M, Altman, Dg, INTERNATIONAL STROKE TRIAL COLLABORATIVE GROUP, European, Carotid, and SURGERY TRIAL COLLABORATIVE GROUP
- Subjects
regression model ,Non-randomised study ,reliability estimate ,efficacy ,comparative study ,multivariate models ,performance - Published
- 2003
19. Randomised controlled trials for policy interventions: a review of reviews and meta-regression.
- Author
-
Oliver, S, Bagnall, A, Thomas, J, Shepherd, J, Sowden, A, White, I, Dinnes, J, Rees, R, Colquitt, J, Oliver, K, Garrett, Z, Oliver, S, Bagnall, A, Thomas, J, Shepherd, J, Sowden, A, White, I, Dinnes, J, Rees, R, Colquitt, J, Oliver, K, and Garrett, Z
- Abstract
To determine whether randomised controlled trials (RCTs) lead to the same effect size and variance as non-randomised studies (NRSs) of similar policy interventions, and whether these findings can be explained by other factors associated with the interventions or their evaluation.
- Published
- 2010
20. Randomised controlled trials for policy interventions: a review of reviews and meta-regression
- Author
-
Oliver, S, primary, Bagnall, AM, additional, Thomas, J, additional, Shepherd, J, additional, Sowden, A, additional, White, I, additional, Dinnes, J, additional, Rees, R, additional, Colquitt, J, additional, Oliver, K, additional, and Garrett, Z, additional
- Published
- 2010
- Full Text
- View/download PDF
21. A systematic review of rapid diagnostic tests for the detection of tuberculosis infection
- Author
-
Dinnes, J, primary, Deeks, J, additional, Kunst, H, additional, Gibson, A, additional, Cummins, E, additional, Waugh, N, additional, Drobniewski, F, additional, and Lalvani, A, additional
- Published
- 2007
- Full Text
- View/download PDF
22. A methodological review of how heterogeneity has been examined in systematic reviews of diagnostic test accuracy
- Author
-
Dinnes, J, primary, Deeks, J, additional, Kirby, J, additional, and Roderick, P, additional
- Published
- 2005
- Full Text
- View/download PDF
23. Clinical effectiveness and cost-effectiveness of drotrecogin alfa (activated) (Xigris®) for the treatment of severe sepsis in adults: a systematic review and economic evaluation
- Author
-
Green, C, primary, Dinnes, J, additional, Takeda, A, additional, Shepherd, J, additional, Hartwell, D, additional, Cave, C, additional, Payne, E, additional, and Cuthbertson, B, additional
- Published
- 2005
- Full Text
- View/download PDF
24. Development and validation of methods for assessing the quality of diagnostic accuracy studies
- Author
-
Whiting, P, primary, Rutjes, A, additional, Dinnes, J, additional, Reitsma, J, additional, Bossuyt, P, additional, and Kleijnen, J, additional
- Published
- 2004
- Full Text
- View/download PDF
25. The effectiveness of diagnostic tests for the assessment of shoulder pain due to soft tissue disorders: a systematic review
- Author
-
Dinnes, J, primary, Loveman, E, additional, McIntyre, L, additional, and Waugh, N, additional
- Published
- 2003
- Full Text
- View/download PDF
26. Evaluating non-randomised intervention studies
- Author
-
Deeks, J, primary, Dinnes, J, additional, D'Amico, R, additional, Sowden, A, additional, Sakarovitch, C, additional, Song, F, additional, Petticrew, M, additional, and Altman, D, additional
- Published
- 2003
- Full Text
- View/download PDF
27. A rapid and systematic review of the effectiveness of temozolomide for the treatment of recurrent malignant glioma
- Author
-
Dinnes, J, primary, Cave, C, additional, Huang, S, additional, and Milne, R, additional
- Published
- 2002
- Full Text
- View/download PDF
28. The effectiveness and cost-effectiveness of temozolomide for the treatment of recurrent malignant glioma: a rapid and systematic review
- Author
-
Dinnes, J., primary, Cave, C., additional, Huang, S., additional, Major, K., additional, and Milne, R., additional
- Published
- 2001
- Full Text
- View/download PDF
29. Cardiac rehabilitation
- Author
-
Dinnes, J., primary, Kleijnen, J., additional, Leitner, M., additional, and Thompson, D., additional
- Published
- 1999
- Full Text
- View/download PDF
30. Systematic reviews to evaluate diagnostic tests
- Author
-
Khan, K. S., Dinnes, J., and Kleijnen, J.
- Published
- 2001
- Full Text
- View/download PDF
31. Cardiac rehabilitation.
- Author
-
Dinnes, J, Kleijnen, J, Leitner, M, and Thompson, D
- Published
- 1999
32. Cochrane systematic review of diagnostic accuracy of dermoscopy in comparison to visual inspection for the diagnosis of melanoma
- Author
-
Matin, R. N., Chuchu, N., Dinnes, J., Jonathan Deeks, Di Ruffano, L. Ferrante, Thomson, D. R., Wong, K. Y., Aldridge, R. B., Abbott, R., Fawzy, M., Bayliss, S. E., Grainge, M. J., Takwoingi, Y., Davenport, C., Godfrey, K., Walter, F. M., and Williams, H.
- Abstract
Early detection of melanoma is essential to improve survival.The additional value of dermoscopy over and above visualinspection (VI) of a suspicious skin lesion is critical to under-stand its contribution to the diagnosis of melanoma. ACochrane systematic review of the diagnostic accuracy of der-moscopy for detection of melanoma in adults was undertakenfor (i) in-person diagnosis and (ii) diagnosis based on dermo-scopic images, and to compare its accuracy with VI alone. Acomprehensive search of 10 databases up to August 2016identified studies of any design evaluating dermoscopy inadults with lesions suspicious for melanoma, compared withhistology or clinical follow-up. Two reviewers independentlyextracted data and quality assessment (using QUADAS-2). Theaccuracy was estimated using hierarchical summary ROCmethods; sensitivities and specificities were estimated forselected points on the summary receiver operating characteris-tic (SROC) curve. Overall, 106 publications were included.The detection of melanoma or intraepidermal melanocyticvariants was analysed for 27 in-person (23 487 lesions; 1737melanomas) and 60 image-based (13 475 lesions; 2851 mela-nomas) datasets. In-person dermoscopy was more accuratethan image-based interpretation [relative diagnostic odds ratio(RDOR) 4.5; 95% CI 2.3–8.5, P < 0.0001]. Dermoscopy wasmore accurate than VI alone; RDORs (i) 4.8 (95% CI: 3.1–7.4; P < 0.0001) for in-person and (ii) 5.6 (95% CI: 3.7–8.5;P < 0.0001) for image-based evaluations. Predicted increasesin sensitivity were (i) 16% (92% vs. 76%) and (ii) 34% (81%vs. 47%) at a fixed specificity of 80%. Use of a publishedalgorithm to assist dermoscopy had no significant impact onaccuracy. The accuracy was significantly higher for experi-enced observers compared with less experienced. Dermoscopyis a valuable tool to support VI of suspicious skin lesions todetect melanoma, particularly in referred populations and for
33. What can available randomised controlled trials evaluating monitoring strategies tell us about the design and analysis of future trials?
- Author
-
Dinnes Jacqueline, Hewison Jenny, Altman Doug, and Deeks Jon
- Subjects
Medicine (General) ,R5-920 - Published
- 2011
- Full Text
- View/download PDF
34. MSR66 Development and Validation of Risk Prediction Tools for Pressure Injury Occurrence: An Umbrella Review.
- Author
-
Hillier, B, Scandrett, K, Coombe, A, Hernandez-Boussard, T., Steyerberg, E, Takwoingi, Y, Velickovic, V, and Dinnes, J
- Published
- 2024
- Full Text
- View/download PDF
35. Risk prediction tools for pressure injury occurrence: an umbrella review of systematic reviews reporting model development and validation methods.
- Author
-
Hillier B, Scandrett K, Coombe A, Hernandez-Boussard T, Steyerberg E, Takwoingi Y, Velickovic V, and Dinnes J
- Abstract
Background: Pressure injuries (PIs) place a substantial burden on healthcare systems worldwide. Risk stratification of those who are at risk of developing PIs allows preventive interventions to be focused on patients who are at the highest risk. The considerable number of risk assessment scales and prediction models available underscores the need for a thorough evaluation of their development, validation, and clinical utility. Our objectives were to identify and describe available risk prediction tools for PI occurrence, their content and the development and validation methods used., Methods: The umbrella review was conducted according to Cochrane guidance. MEDLINE, Embase, CINAHL, EPISTEMONIKOS, Google Scholar, and reference lists were searched to identify relevant systematic reviews. The risk of bias was assessed using adapted AMSTAR-2 criteria. Results were described narratively. All included reviews contributed to building a comprehensive list of risk prediction tools., Results: We identified 32 eligible systematic reviews only seven of which described the development and validation of risk prediction tools for PI. Nineteen reviews assessed the prognostic accuracy of the tools and 11 assessed clinical effectiveness. Of the seven reviews reporting model development and validation, six included only machine learning models. Two reviews included external validations of models, although only one review reported any details on external validation methods or results. This was also the only review to report measures of both discrimination and calibration. Five reviews presented measures of discrimination, such as the area under the curve (AUC), sensitivities, specificities, F1 scores, and G-means. For the four reviews that assessed the risk of bias assessment using the PROBAST tool, all models but one were found to be at high or unclear risk of bias., Conclusions: Available tools do not meet current standards for the development or reporting of risk prediction models. The majority of tools have not been externally validated. Standardised and rigorous approaches to risk prediction model development and validation are needed., Trial Registration: The protocol was registered on the Open Science Framework ( https://osf.io/tepyk )., Competing Interests: Declarations. Ethics approval and consent to participate.: Not applicable. Consent for publication: Not applicable. Competing interests: The authors of this manuscript have the following competing interests: VV is an employee of Paul Hartmann AG; ES and THB received consultancy fees from Paul Hartmann AG. VV, ES and THB were not involved in data curation, screening, data extraction, analysis of results or writing of the original draft. These roles were conducted independently by authors at the University of Birmingham. All other authors received no personal funding or personal compensation from Paul Hartmann AG and have declared that no competing interests exist., (© 2024. The Author(s).)
- Published
- 2025
- Full Text
- View/download PDF
36. Optimising research investment by simulating and evaluating monitoring strategies to inform a trial: a simulation of liver fibrosis monitoring.
- Author
-
Sitch AJ, Dinnes J, Hewison J, Gregory W, Parkes J, and Deeks JJ
- Subjects
- Humans, Randomized Controlled Trials as Topic methods, Liver Cirrhosis diagnosis, Disease Progression, Computer Simulation, Biomarkers analysis
- Abstract
Background: The aim of the study was to investigate the development of evidence-based monitoring strategies in a population with progressive or recurrent disease. A simulation study of monitoring strategies using a new biomarker (ELF) for the detection of liver cirrhosis in people with known liver fibrosis was undertaken alongside a randomised controlled trial (ELUCIDATE)., Methods: Existing data and expert opinion were used to estimate the progression of disease and the performance of repeat testing with ELF. Knowledge of the true disease status in addition to the observed test results for a cohort of simulated patients allowed various monitoring strategies to be implemented, evaluated and validated against trial data., Results: Several monitoring strategies ranging in complexity were successfully modelled and compared regarding the timing of detection of disease, the duration of monitoring, and the predictive value of a positive test result. The results of sensitivity analysis showed the importance of accurate data to inform the simulation. Results of the simulation were similar to those from the trial., Conclusion: Monitoring data can be simulated and strategies compared given adequate knowledge of disease progression and test performance. Such exercises should be carried out to ensure optimal strategies are evaluated in trials thus reducing research waste. Monitoring data can be generated and monitoring strategies can be assessed if data is available on the monitoring test performance and the test variability. This work highlights the data necessary and the general method for evaluating the performance of monitoring strategies, allowing appropriate strategies to be selected for evaluation. Modelling work should be conducted prior to full scale investigation of monitoring strategies, allowing optimal monitoring strategies to be assessed., Competing Interests: Declarations. Ethics approval and consent to participate: Not applicable. Consent for publication: Not applicable. Competing interests: Julie Parkes has been paid for providing lectures by Siemens on the topic of the ELF marker. Julie Parkes reported receiving support from the Speakers’ Bureau for Siemens Healthcare Diagnostics, outside the submitted work, and is married to William Rosenberg. William Rosenberg was an inventor of the enhanced liver fibrosis (ELF) test when he was an employee of the University of Southampton. His rights were transferred to Siemens Healthcare Diagnostics Ltd by the University of Southampton. He does not receive any payment in relation to sales of the test by the manufacturer Siemens Healthcare Diagnostics. He has received grant support and speaker fees from Siemens Healthcare Diagnostics and is a director of iQur Ltd (Southampton, UK), a company that provides ELF testing. In the context of this NIHR-funded study, all ELF testing provided by iQur Ltd was performed on a not-for-profit, cost-recovery basis. During the period of the study Jonathan J Deeks was a panel member of the HTA Commissioning Board. Jenny Hewison was a panel member of the National Institute for Health Research (NIHR) Clinical Trials Unit Standing Advisory Committee and Subpanel Chair, NIHR Programme Grants for Applied Research. Walter Gregory was the ELUCIDATE Trial Director and Principal Statistician. Alice Sitch and Jacqueline Dinnes report no competing interests., (© 2024. The Author(s).)
- Published
- 2024
- Full Text
- View/download PDF
37. The effect of sample site and collection procedure on identification of SARS-CoV-2 infection.
- Author
-
Davenport C, Arevalo-Rodriguez I, Mateos-Haro M, Berhane S, Dinnes J, Spijker R, Buitrago-Garcia D, Ciapponi A, Takwoingi Y, Deeks JJ, Emperador D, Leeflang MMG, and Van den Bruel A
- Subjects
- Humans, Oropharynx virology, Viral Load, COVID-19 Nucleic Acid Testing methods, COVID-19 Testing methods, Pharynx virology, Nasal Cavity virology, Nose virology, COVID-19 diagnosis, Specimen Handling methods, SARS-CoV-2 isolation & purification, Nasopharynx virology, Sensitivity and Specificity
- Abstract
Background: Sample collection is a key driver of accuracy in the diagnosis of SARS-CoV-2 infection. Viral load may vary at different anatomical sampling sites and accuracy may be compromised by difficulties obtaining specimens and the expertise of the person taking the sample. It is important to optimise sampling accuracy within cost, safety and accessibility constraints., Objectives: To compare the sensitivity of different sampling collection sites and methods for the detection of current SARS-CoV-2 infection with any molecular or antigen-based test., Search Methods: Electronic searches of the Cochrane COVID-19 Study Register and the COVID-19 Living Evidence Database from the University of Bern (which includes daily updates from PubMed and Embase and preprints from medRxiv and bioRxiv) were undertaken on 22 February 2022. We included independent evaluations from national reference laboratories, FIND and the Diagnostics Global Health website. We did not apply language restrictions., Selection Criteria: We included studies of symptomatic or asymptomatic people with suspected SARS-CoV-2 infection undergoing testing. We included studies of any design that compared results from different sample types (anatomical location, operator, collection device) collected from the same participant within a 24-hour period., Data Collection and Analysis: Within a sample pair, we defined a reference sample and an index sample collected from the same participant within the same clinical encounter (within 24 hours). Where the sample comparison was different anatomical sites, the reference standard was defined as a nasopharyngeal or combined naso/oropharyngeal sample collected into the same sample container and the index sample as the alternative anatomical site. Where the sample comparison was concerned with differences in the sample collection method from the same site, we defined the reference sample as that closest to standard practice for that sample type. Where the sample pair comparison was concerned with differences in personnel collecting the sample, the more skilled or experienced operator was considered the reference sample. Two review authors independently assessed the risk of bias and applicability concerns using the QUADAS-2 and QUADAS-C checklists, tailored to this review. We present estimates of the difference in the sensitivity (reference sample (%) minus index sample sensitivity (%)) in a pair and as an average across studies for each index sampling method using forest plots and tables. We examined heterogeneity between studies according to population (age, symptom status) and index sample (time post-symptom onset, operator expertise, use of transport medium) characteristics., Main Results: This review includes 106 studies reporting 154 evaluations and 60,523 sample pair comparisons, of which 11,045 had SARS-CoV-2 infection. Ninety evaluations were of saliva samples, 37 nasal, seven oropharyngeal, six gargle, six oral and four combined nasal/oropharyngeal samples. Four evaluations were of the effect of operator expertise on the accuracy of three different sample types. The majority of included evaluations (146) used molecular tests, of which 140 used RT-PCR (reverse transcription polymerase chain reaction). Eight evaluations were of nasal samples used with Ag-RDTs (rapid antigen tests). The majority of studies were conducted in Europe (35/106, 33%) or the USA (27%) and conducted in dedicated COVID-19 testing clinics or in ambulatory hospital settings (53%). Targeted screening or contact tracing accounted for only 4% of evaluations. Where reported, the majority of evaluations were of adults (91/154, 59%), 28 (18%) were in mixed populations with only seven (4%) in children. The median prevalence of confirmed SARS-CoV-2 was 23% (interquartile (IQR) 13%-40%). Risk of bias and applicability assessment were hampered by poor reporting in 77% and 65% of included studies, respectively. Risk of bias was low across all domains in only 3% of evaluations due to inappropriate inclusion or exclusion criteria, unclear recruitment, lack of blinding, nonrandomised sampling order or differences in testing kit within a sample pair. Sixty-eight percent of evaluation cohorts were judged as being at high or unclear applicability concern either due to inflation of the prevalence of SARS-CoV-2 infection in study populations by selectively including individuals with confirmed PCR-positive samples or because there was insufficient detail to allow replication of sample collection. When used with RT-PCR • There was no evidence of a difference in sensitivity between gargle and nasopharyngeal samples (on average -1 percentage points, 95% CI -5 to +2, based on 6 evaluations, 2138 sample pairs, of which 389 had SARS-CoV-2). • There was no evidence of a difference in sensitivity between saliva collection from the deep throat and nasopharyngeal samples (on average +10 percentage points, 95% CI -1 to +21, based on 2192 sample pairs, of which 730 had SARS-CoV-2). • There was evidence that saliva collection using spitting, drooling or salivating was on average -12 percentage points less sensitive (95% CI -16 to -8, based on 27,253 sample pairs, of which 4636 had SARS-CoV-2) compared to nasopharyngeal samples. We did not find any evidence of a difference in the sensitivity of saliva collected using spitting, drooling or salivating (sensitivity difference: range from -13 percentage points (spit) to -21 percentage points (salivate)). • Nasal samples (anterior and mid-turbinate collection combined) were, on average, 12 percentage points less sensitive compared to nasopharyngeal samples (95% CI -17 to -7), based on 9291 sample pairs, of which 1485 had SARS-CoV-2. We did not find any evidence of a difference in sensitivity between nasal samples collected from the mid-turbinates (3942 sample pairs) or from the anterior nares (8272 sample pairs). • There was evidence that oropharyngeal samples were, on average, 17 percentage points less sensitive than nasopharyngeal samples (95% CI -29 to -5), based on seven evaluations, 2522 sample pairs, of which 511 had SARS-CoV-2. A much smaller volume of evidence was available for combined nasal/oropharyngeal samples and oral samples. Age, symptom status and use of transport media do not appear to affect the sensitivity of saliva samples and nasal samples. When used with Ag-RDTs • There was no evidence of a difference in sensitivity between nasal samples compared to nasopharyngeal samples (sensitivity, on average, 0 percentage points -0.2 to +0.2, based on 3688 sample pairs, of which 535 had SARS-CoV-2)., Authors' Conclusions: When used with RT-PCR, there is no evidence for a difference in sensitivity of self-collected gargle or deep-throat saliva samples compared to nasopharyngeal samples collected by healthcare workers when used with RT-PCR. Use of these alternative, self-collected sample types has the potential to reduce cost and discomfort and improve the safety of sampling by reducing risk of transmission from aerosol spread which occurs as a result of coughing and gagging during the nasopharyngeal or oropharyngeal sample collection procedure. This may, in turn, improve access to and uptake of testing. Other types of saliva, nasal, oral and oropharyngeal samples are, on average, less sensitive compared to healthcare worker-collected nasopharyngeal samples, and it is unlikely that sensitivities of this magnitude would be acceptable for confirmation of SARS-CoV-2 infection with RT-PCR. When used with Ag-RDTs, there is no evidence of a difference in sensitivity between nasal samples and healthcare worker-collected nasopharyngeal samples for detecting SARS-CoV-2. The implications of this for self-testing are unclear as evaluations did not report whether nasal samples were self-collected or collected by healthcare workers. Further research is needed in asymptomatic individuals, children and in Ag-RDTs, and to investigate the effect of operator expertise on accuracy. Quality assessment of the evidence base underpinning these conclusions was restricted by poor reporting. There is a need for further high-quality studies, adhering to reporting standards for test accuracy studies., (Copyright © 2024 The Authors. Cochrane Database of Systematic Reviews published by John Wiley & Sons, Ltd. on behalf of The Cochrane Collaboration.)
- Published
- 2024
- Full Text
- View/download PDF
38. Laboratory-based molecular test alternatives to RT-PCR for the diagnosis of SARS-CoV-2 infection.
- Author
-
Arevalo-Rodriguez I, Mateos-Haro M, Dinnes J, Ciapponi A, Davenport C, Buitrago-Garcia D, Bennouna-Dalero T, Roqué-Figuls M, Van den Bruel A, von Eije KJ, Emperador D, Hooft L, Spijker R, Leeflang MM, Takwoingi Y, and Deeks JJ
- Subjects
- Humans, Bias, False Negative Reactions, False Positive Reactions, Pandemics, Real-Time Polymerase Chain Reaction methods, Reverse Transcriptase Polymerase Chain Reaction methods, Reverse Transcriptase Polymerase Chain Reaction standards, COVID-19 diagnosis, COVID-19 Nucleic Acid Testing methods, RNA, Viral analysis, SARS-CoV-2 genetics, SARS-CoV-2 isolation & purification, Sensitivity and Specificity
- Abstract
Background: Diagnosing people with a SARS-CoV-2 infection played a critical role in managing the COVID-19 pandemic and remains a priority for the transition to long-term management of COVID-19. Initial shortages of extraction and reverse transcription polymerase chain reaction (RT-PCR) reagents impaired the desired upscaling of testing in many countries, which led to the search for alternatives to RNA extraction/purification and RT-PCR testing. Reference standard methods for diagnosing the presence of SARS-CoV-2 infection rely primarily on real-time reverse transcription-polymerase chain reaction (RT-PCR). Alternatives to RT-PCR could, if sufficiently accurate, have a positive impact by expanding the range of diagnostic tools available for the timely identification of people infected by SARS-CoV-2, access to testing and the use of resources., Objectives: To assess the diagnostic accuracy of alternative (to RT-PCR assays) laboratory-based molecular tests for diagnosing SARS-CoV-2 infection., Search Methods: We searched the COVID-19 Open Access Project living evidence database from the University of Bern until 30 September 2020 and the WHO COVID-19 Research Database until 31 October 2022. We did not apply language restrictions., Selection Criteria: We included studies of people with suspected or known SARS-CoV-2 infection, or where tests were used to screen for infection, and studies evaluating commercially developed laboratory-based molecular tests for the diagnosis of SARS-CoV-2 infection considered as alternatives to RT-PCR testing. We also included all reference standards to define the presence or absence of SARS-CoV-2, including RT-PCR tests and established clinical diagnostic criteria., Data Collection and Analysis: Two authors independently screened studies and resolved disagreements by discussing them with a third author. Two authors independently extracted data and assessed the risk of bias and applicability of the studies using the QUADAS-2 tool. We presented sensitivity and specificity, with 95% confidence intervals (CIs), for each test using paired forest plots and summarised results using average sensitivity and specificity using a bivariate random-effects meta-analysis. We illustrated the findings per index test category and assay brand compared to the WHO's acceptable sensitivity and specificity threshold for diagnosing SARS-CoV-2 infection using nucleic acid tests., Main Results: We included data from 64 studies reporting 94 cohorts of participants and 105 index test evaluations, with 74,753 samples and 7517 confirmed SARS-CoV-2 cases. We did not identify any published or preprint reports of accuracy for a considerable number of commercially produced NAAT assays. Most cohorts were judged at unclear or high risk of bias in more than three QUADAS-2 domains. Around half of the cohorts were considered at high risk of selection bias because of recruitment based on COVID status. Three quarters of 94 cohorts were at high risk of bias in the reference standard domain because of reliance on a single RT-PCR result to determine the absence of SARS-CoV-2 infection or were at unclear risk of bias due to a lack of clarity about the time interval between the index test assessment and the reference standard, the number of missing results, or the absence of a participant flow diagram. For index tests categories with four or more evaluations and when summary estimations were possible, we found that: a) For RT-PCR assays designed to omit/adapt RNA extraction/purification, the average sensitivity was 95.1% (95% CI 91.1% to 97.3%), and the average specificity was 99.7% (95% CI 98.5% to 99.9%; based on 27 evaluations, 2834 samples and 1178 SARS-CoV-2 cases); b) For RT-LAMP assays, the average sensitivity was 88.4% (95% CI 83.1% to 92.2%), and the average specificity was 99.7% (95% CI 98.7% to 99.9%; 24 evaluations, 29,496 samples and 2255 SARS-CoV-2 cases); c) for TMA assays, the average sensitivity was 97.6% (95% CI 95.2% to 98.8%), and the average specificity was 99.4% (95% CI 94.9% to 99.9%; 14 evaluations, 2196 samples and 942 SARS-CoV-2 cases); d) for digital PCR assays, the average sensitivity was 98.5% (95% CI 95.2% to 99.5%), and the average specificity was 91.4% (95% CI 60.4% to 98.7%; five evaluations, 703 samples and 354 SARS-CoV-2 cases); e) for RT-LAMP assays omitting/adapting RNA extraction, the average sensitivity was 73.1% (95% CI 58.4% to 84%), and the average specificity was 100% (95% CI 98% to 100%; 24 evaluations, 14,342 samples and 1502 SARS-CoV-2 cases). Only two index test categories fulfil the WHO-acceptable sensitivity and specificity requirements for SARS-CoV-2 nucleic acid tests: RT-PCR assays designed to omit/adapt RNA extraction/purification and TMA assays. In addition, WHO-acceptable performance criteria were met for two assays out of 35 when tests were used according to manufacturer instructions. At 5% prevalence using a cohort of 1000 people suspected of SARS-CoV-2 infection, the positive predictive value of RT-PCR assays omitting/adapting RNA extraction/purification will be 94%, with three in 51 positive results being false positives, and around two missed cases. For TMA assays, the positive predictive value of RT-PCR assays will be 89%, with 6 in 55 positive results being false positives, and around one missed case., Authors' Conclusions: Alternative laboratory-based molecular tests aim to enhance testing capacity in different ways, such as reducing the time, steps and resources needed to obtain valid results. Several index test technologies with these potential advantages have not been evaluated or have been assessed by only a few studies of limited methodological quality, so the performance of these kits was undetermined. Only two index test categories with enough evaluations for meta-analysis fulfil the WHO set of acceptable accuracy standards for SARS-CoV-2 nucleic acid tests: RT-PCR assays designed to omit/adapt RNA extraction/purification and TMA assays. These assays might prove to be suitable alternatives to RT-PCR for identifying people infected by SARS-CoV-2, especially when the alternative would be not having access to testing. However, these findings need to be interpreted and used with caution because of several limitations in the evidence, including reliance on retrospective samples without information about the symptom status of participants and the timing of assessment. No extrapolation of found accuracy data for these two alternatives to any test brands using the same techniques can be made as, for both groups, one test brand with high accuracy was overrepresented with 21/26 and 12/14 included studies, respectively. Although we used a comprehensive search and had broad eligibility criteria to include a wide range of tests that could be alternatives to RT-PCR methods, further research is needed to assess the performance of alternative COVID-19 tests and their role in pandemic management., (Copyright © 2024 The Cochrane Collaboration. Published by John Wiley & Sons, Ltd.)
- Published
- 2024
- Full Text
- View/download PDF
39. Accuracy of routine laboratory tests to predict mortality and deterioration to severe or critical COVID-19 in people with SARS-CoV-2.
- Author
-
De Rop L, Bos DA, Stegeman I, Holtman G, Ochodo EA, Spijker R, Otieno JA, Alkhlaileh F, Deeks JJ, Dinnes J, Van den Bruel A, McInnes MD, Leeflang MM, and Verbakel JY
- Subjects
- Humans, Biomarkers blood, Prognosis, Clinical Deterioration, Bias, Pandemics, Sensitivity and Specificity, Severity of Illness Index, COVID-19 Testing methods, COVID-19 mortality, COVID-19 blood, COVID-19 diagnosis, SARS-CoV-2, C-Reactive Protein analysis
- Abstract
Background: Identifying patients with COVID-19 disease who will deteriorate can be useful to assess whether they should receive intensive care, or whether they can be treated in a less intensive way or through outpatient care. In clinical care, routine laboratory markers, such as C-reactive protein, are used to assess a person's health status., Objectives: To assess the accuracy of routine blood-based laboratory tests to predict mortality and deterioration to severe or critical (from mild or moderate) COVID-19 in people with SARS-CoV-2., Search Methods: On 25 August 2022, we searched the Cochrane COVID-19 Study Register, encompassing searches of various databases such as MEDLINE via PubMed, CENTRAL, Embase, medRxiv, and ClinicalTrials.gov. We did not apply any language restrictions., Selection Criteria: We included studies of all designs that produced estimates of prognostic accuracy in participants who presented to outpatient services, or were admitted to general hospital wards with confirmed SARS-CoV-2 infection, and studies that were based on serum banks of samples from people. All routine blood-based laboratory tests performed during the first encounter were included. We included any reference standard used to define deterioration to severe or critical disease that was provided by the authors., Data Collection and Analysis: Two review authors independently extracted data from each included study, and independently assessed the methodological quality using the Quality Assessment of Prognostic Accuracy Studies tool. As studies reported different thresholds for the same test, we used the Hierarchical Summary Receiver Operator Curve model for meta-analyses to estimate summary curves in SAS 9.4. We estimated the sensitivity at points on the SROC curves that corresponded to the median and interquartile range boundaries of specificities in the included studies. Direct and indirect comparisons were exclusively conducted for biomarkers with an estimated sensitivity and 95% CI of ≥ 50% at a specificity of ≥ 50%. The relative diagnostic odds ratio was calculated as a summary of the relative accuracy of these biomarkers., Main Results: We identified a total of 64 studies, including 71,170 participants, of which 8169 participants died, and 4031 participants deteriorated to severe/critical condition. The studies assessed 53 different laboratory tests. For some tests, both increases and decreases relative to the normal range were included. There was important heterogeneity between tests and their cut-off values. None of the included studies had a low risk of bias or low concern for applicability for all domains. None of the tests included in this review demonstrated high sensitivity or specificity, or both. The five tests with summary sensitivity and specificity above 50% were: C-reactive protein increase, neutrophil-to-lymphocyte ratio increase, lymphocyte count decrease, d-dimer increase, and lactate dehydrogenase increase. Inflammation For mortality, summary sensitivity of a C-reactive protein increase was 76% (95% CI 73% to 79%) at median specificity, 59% (low-certainty evidence). For deterioration, summary sensitivity was 78% (95% CI 67% to 86%) at median specificity, 72% (very low-certainty evidence). For the combined outcome of mortality or deterioration, or both, summary sensitivity was 70% (95% CI 49% to 85%) at median specificity, 60% (very low-certainty evidence). For mortality, summary sensitivity of an increase in neutrophil-to-lymphocyte ratio was 69% (95% CI 66% to 72%) at median specificity, 63% (very low-certainty evidence). For deterioration, summary sensitivity was 75% (95% CI 59% to 87%) at median specificity, 71% (very low-certainty evidence). For mortality, summary sensitivity of a decrease in lymphocyte count was 67% (95% CI 56% to 77%) at median specificity, 61% (very low-certainty evidence). For deterioration, summary sensitivity of a decrease in lymphocyte count was 69% (95% CI 60% to 76%) at median specificity, 67% (very low-certainty evidence). For the combined outcome, summary sensitivity was 83% (95% CI 67% to 92%) at median specificity, 29% (very low-certainty evidence). For mortality, summary sensitivity of a lactate dehydrogenase increase was 82% (95% CI 66% to 91%) at median specificity, 60% (very low-certainty evidence). For deterioration, summary sensitivity of a lactate dehydrogenase increase was 79% (95% CI 76% to 82%) at median specificity, 66% (low-certainty evidence). For the combined outcome, summary sensitivity was 69% (95% CI 51% to 82%) at median specificity, 62% (very low-certainty evidence). Hypercoagulability For mortality, summary sensitivity of a d-dimer increase was 70% (95% CI 64% to 76%) at median specificity of 56% (very low-certainty evidence). For deterioration, summary sensitivity was 65% (95% CI 56% to 74%) at median specificity of 63% (very low-certainty evidence). For the combined outcome, summary sensitivity was 65% (95% CI 52% to 76%) at median specificity of 54% (very low-certainty evidence). To predict mortality, neutrophil-to-lymphocyte ratio increase had higher accuracy compared to d-dimer increase (RDOR (diagnostic Odds Ratio) 2.05, 95% CI 1.30 to 3.24), C-reactive protein increase (RDOR 2.64, 95% CI 2.09 to 3.33), and lymphocyte count decrease (RDOR 2.63, 95% CI 1.55 to 4.46). D-dimer increase had higher accuracy compared to lymphocyte count decrease (RDOR 1.49, 95% CI 1.23 to 1.80), C-reactive protein increase (RDOR 1.31, 95% CI 1.03 to 1.65), and lactate dehydrogenase increase (RDOR 1.42, 95% CI 1.05 to 1.90). Additionally, lactate dehydrogenase increase had higher accuracy compared to lymphocyte count decrease (RDOR 1.30, 95% CI 1.13 to 1.49). To predict deterioration to severe disease, C-reactive protein increase had higher accuracy compared to d-dimer increase (RDOR 1.76, 95% CI 1.25 to 2.50). The neutrophil-to-lymphocyte ratio increase had higher accuracy compared to d-dimer increase (RDOR 2.77, 95% CI 1.58 to 4.84). Lastly, lymphocyte count decrease had higher accuracy compared to d-dimer increase (RDOR 2.10, 95% CI 1.44 to 3.07) and lactate dehydrogenase increase (RDOR 2.22, 95% CI 1.52 to 3.26)., Authors' Conclusions: Laboratory tests, associated with hypercoagulability and hyperinflammatory response, were better at predicting severe disease and mortality in patients with SARS-CoV-2 compared to other laboratory tests. However, to safely rule out severe disease, tests should have high sensitivity (> 90%), and none of the identified laboratory tests met this criterion. In clinical practice, a more comprehensive assessment of a patient's health status is usually required by, for example, incorporating these laboratory tests into clinical prediction rules together with clinical symptoms, radiological findings, and patient's characteristics., (Copyright © 2024 The Authors. Cochrane Database of Systematic Reviews published by John Wiley & Sons, Ltd. on behalf of The Cochrane Collaboration.)
- Published
- 2024
- Full Text
- View/download PDF
40. Target Product Profile for a Machine Learning-Automated Retinal Imaging Analysis Software for Use in English Diabetic Eye Screening: Protocol for a Mixed Methods Study.
- Author
-
Macdonald T, Dinnes J, Maniatopoulos G, Taylor-Phillips S, Shinkins B, Hogg J, Dunbar JK, Solebo AL, Sutton H, Attwood J, Pogose M, Given-Wilson R, Greaves F, Macrae C, Pearson R, Bamford D, Tufail A, Liu X, and Denniston AK
- Abstract
Background: Diabetic eye screening (DES) represents a significant opportunity for the application of machine learning (ML) technologies, which may improve clinical and service outcomes. However, successful integration of ML into DES requires careful product development, evaluation, and implementation. Target product profiles (TPPs) summarize the requirements necessary for successful implementation so these can guide product development and evaluation., Objective: This study aims to produce a TPP for an ML-automated retinal imaging analysis software (ML-ARIAS) system for use in DES in England., Methods: This work will consist of 3 phases. Phase 1 will establish the characteristics to be addressed in the TPP. A list of candidate characteristics will be generated from the following sources: an overview of systematic reviews of diagnostic test TPPs; a systematic review of digital health TPPs; and the National Institute for Health and Care Excellence's Evidence Standards Framework for Digital Health Technologies. The list of characteristics will be refined and validated by a study advisory group (SAG) made up of representatives from key stakeholders in DES. This includes people with diabetes; health care professionals; health care managers and leaders; and regulators and policy makers. In phase 2, specifications for these characteristics will be drafted following a series of semistructured interviews with participants from these stakeholder groups. Data collected from these interviews will be analyzed using the shortlist of characteristics as a framework, after which specifications will be drafted to create a draft TPP. Following approval by the SAG, in phase 3, the draft will enter an internet-based Delphi consensus study with participants sought from the groups previously identified, as well as ML-ARIAS developers, to ensure feasibility. Participants will be invited to score characteristic and specification pairs on a scale from "definitely exclude" to "definitely include," and suggest edits. The document will be iterated between rounds based on participants' feedback. Feedback on the draft document will be sought from a group of ML-ARIAS developers before its final contents are agreed upon in an in-person consensus meeting. At this meeting, representatives from the stakeholder groups previously identified (minus ML-ARIAS developers, to avoid bias) will be presented with the Delphi results and feedback of the user group and asked to agree on the final contents by vote., Results: Phase 1 was completed in November 2023. Phase 2 is underway and expected to finish in March 2024. Phase 3 is expected to be complete in July 2024., Conclusions: The multistakeholder development of a TPP for an ML-ARIAS for use in DES in England will help developers produce tools that serve the needs of patients, health care providers, and their staff. The TPP development process will also provide methods and a template to produce similar documents in other disease areas., International Registered Report Identifier (irrid): DERR1-10.2196/50568., (©Trystan Macdonald, Jacqueline Dinnes, Gregory Maniatopoulos, Sian Taylor-Phillips, Bethany Shinkins, Jeffry Hogg, John Kevin Dunbar, Ameenat Lola Solebo, Hannah Sutton, John Attwood, Michael Pogose, Rosalind Given-Wilson, Felix Greaves, Carl Macrae, Russell Pearson, Daniel Bamford, Adnan Tufail, Xiaoxuan Liu, Alastair K Denniston. Originally published in JMIR Research Protocols (https://www.researchprotocols.org), 27.03.2024.)
- Published
- 2024
- Full Text
- View/download PDF
41. Accuracy of package inserts of SARS-CoV-2 rapid antigen tests: a secondary analysis of manufacturer versus systematic review data.
- Author
-
Bigio J, MacLean EL, Das R, Sulis G, Kohli M, Berhane S, Dinnes J, Deeks JJ, Brümmer LE, Denkinger CM, and Pai M
- Subjects
- Humans, Pandemics, Product Labeling, Sensitivity and Specificity, Systematic Reviews as Topic, COVID-19 diagnosis, SARS-CoV-2
- Abstract
Background: Rapid antigen tests (RATs) were crucial during the COVID-19 pandemic. Information provided by the test manufacturer in product package inserts, also known as instructions for use (IFUs), is often the only data available to clinicians, public health professionals, and individuals on the diagnostic accuracy of these tests. We aimed to assess whether manufacturer IFU accuracy data aligned with evidence from independent research., Methods: We searched company websites for package inserts for RATs that were included in the July 2022 update of the Cochrane meta-analysis of SARS-CoV-2 RATs, which served as a benchmark for research evidence. We fitted bivariate hierarchical models to obtain absolute differences in sensitivity and specificity between IFU and Cochrane Review estimates for each test, as well as overall combined differences., Findings: We found 22 (100%) of 22 IFUs of the RATs included in the Cochrane Review. IFUs for 12 (55%) of 22 RATs reported statistically significantly higher sensitivity estimates than the Cochrane Review, and none reported lower estimates. The mean difference between IFU and Cochrane Review sensitivity estimates across tests was 12·0% (95% CI 7·5-16·6). IFUs in three (14%) of 22 diagnostic tests had significantly higher specificity estimates than the Cochrane Review and two (9%) of 22 had lower estimates. The mean difference between IFU and Cochrane Review specificity estimates across tests was 0·3% (95% CI 0·1-0·5). If 100 people with SARS-CoV-2 infection were tested with each of the tests in this study, on average 12 fewer people would be correctly diagnosed than is suggested by the package inserts., Interpretation: Health professionals and the public should be aware that package inserts for SARS-CoV-2 RATs might provide an overly optimistic picture of the sensitivity of a test. Regulatory bodies should strengthen their requirements for the reporting of diagnostic accuracy data in package inserts and policy makers should demand independent validation data for decision making., Funding: None., Competing Interests: Declaration of interests JB has received consulting fees from the New Diagnostics Working Group of the Stop TB Partnership. SB, JD, and JJD are supported by the National Institute for Health and Care Research (NIHR) Birmingham Biomedical Research Centre. This Article presents independent research supported by the NIHR Birmingham Biomedical Research Centre at the University Hospitals Birmingham National Health Services (NHS) Foundation Trust and the University of Birmingham. The views expressed are those of the authors and not necessarily those of the NHS, the NIHR, or the Department of Health and Social Care. CMD has received research grants from the US National Institutes of Health, German Ministry of Education and Research, German Alliance for Global Health Research, United States Agency for International Development, FIND, German Center for Infection Research, UNAIDS, WHO, and Roche. CMD also declares a payment from Roche Diagnostics that she accepted as German law requires a manufacturer to pay for the use of data for regulatory purposes. Data was generated as part of an independent evaluation by CMD and team. CMD is an academic editor of PLoS Medicine and sits on the WHO technical advisory group on tuberculosis diagnostics. MP holds a Canada Research Chair Award from the Canadian Institutes of Health Research and serves as an adviser to the following non-profit agencies in global health: Bill & Melinda Gates Foundation; Foundation for Innovative New Diagnostics, WHO, and Stop TB Partnership. All other authors declare no competing interests., (Copyright © 2023 The Author(s). Published by Elsevier Ltd. This is an Open Access article under the CC BY-NC-ND 4.0 license. Published by Elsevier Ltd.. All rights reserved.)
- Published
- 2023
- Full Text
- View/download PDF
42. Diagnosis of suspicious pigmented lesions in specialist settings with artificial intelligence.
- Author
-
Matin RN and Dinnes J
- Subjects
- Humans, Algorithms, Artificial Intelligence, Skin Neoplasms diagnosis
- Published
- 2023
- Full Text
- View/download PDF
43. Key issues when considering adopting a skin cancer diagnostic tool that uses artificial intelligence.
- Author
-
Kelly L, Coote L, Dinnes J, Fleming C, Holmes H, and Matin RN
- Subjects
- Humans, Artificial Intelligence, Skin Neoplasms diagnosis
- Abstract
Competing Interests: Conflicts of interest R.N.M. is Chair of the British Association of Dermatologists Artificial Intelligence Working Party Group.
- Published
- 2023
- Full Text
- View/download PDF
44. Rapid antigen-based and rapid molecular tests for the detection of SARS-CoV-2: a rapid review with network meta-analysis of diagnostic test accuracy studies.
- Author
-
Veroniki AA, Tricco AC, Watt J, Tsokani S, Khan PA, Soobiah C, Negm A, Doherty-Kirby A, Taylor P, Lunny C, McGowan J, Little J, Mallon P, Moher D, Wong S, Dinnes J, Takwoingi Y, Saxinger L, Chan A, Isaranuwatchai W, Lander B, Meyers A, Poliquin G, and Straus SE
- Subjects
- Humans, Network Meta-Analysis, Bias, Diagnostic Tests, Routine, Sensitivity and Specificity, COVID-19 Testing, SARS-CoV-2 genetics, COVID-19 diagnosis
- Abstract
Background: The global spread of COVID-19 created an explosion in rapid tests with results in < 1 hour, but their relative performance characteristics are not fully understood yet. Our aim was to determine the most sensitive and specific rapid test for the diagnosis of SARS-CoV-2., Methods: Design: Rapid review and diagnostic test accuracy network meta-analysis (DTA-NMA)., Eligibility Criteria: Randomized controlled trials (RCTs) and observational studies assessing rapid antigen and/or rapid molecular test(s) to detect SARS-CoV-2 in participants of any age, suspected or not with SARS-CoV-2 infection., Information Sources: Embase, MEDLINE, and Cochrane Central Register of Controlled Trials, up to September 12, 2021., Outcome Measures: Sensitivity and specificity of rapid antigen and molecular tests suitable for detecting SARS-CoV-2. Data extraction and risk of bias assessment: Screening of literature search results was conducted by one reviewer; data abstraction was completed by one reviewer and independently verified by a second reviewer. Risk of bias was not assessed in the included studies., Data Synthesis: Random-effects meta-analysis and DTA-NMA., Results: We included 93 studies (reported in 88 articles) relating to 36 rapid antigen tests in 104,961 participants and 23 rapid molecular tests in 10,449 participants. Overall, rapid antigen tests had a sensitivity of 0.75 (95% confidence interval 0.70-0.79) and specificity of 0.99 (0.98-0.99). Rapid antigen test sensitivity was higher when nasal or combined samples (e.g., combinations of nose, throat, mouth, or saliva samples) were used, but lower when nasopharyngeal samples were used, and in those classified as asymptomatic at the time of testing. Rapid molecular tests may result in fewer false negatives than rapid antigen tests (sensitivity: 0.93, 0.88-0.96; specificity: 0.98, 0.97-0.99). The tests with the highest sensitivity and specificity estimates were the Xpert Xpress rapid molecular test by Cepheid (sensitivity: 0.99, 0.83-1.00; specificity: 0.97, 0.69-1.00) among the 23 commercial rapid molecular tests and the COVID-VIRO test by AAZ-LMB (sensitivity: 0.93, 0.48-0.99; specificity: 0.98, 0.44-1.00) among the 36 rapid antigen tests we examined., Conclusions: Rapid molecular tests were associated with both high sensitivity and specificity, while rapid antigen tests were mainly associated with high specificity, according to the minimum performance requirements by WHO and Health Canada. Our rapid review was limited to English, peer-reviewed published results of commercial tests, and study risk of bias was not assessed. A full systematic review is required., Review Registration: PROSPERO CRD42021289712., (© 2023. The Author(s).)
- Published
- 2023
- Full Text
- View/download PDF
45. Routine laboratory testing to determine if a patient has COVID-19.
- Author
-
Stegeman I, Ochodo EA, Guleid F, Holtman GA, Yang B, Davenport C, Deeks JJ, Dinnes J, Dittrich S, Emperador D, Hoo L, Spijker R, Takwoingi Y, Van den Bruel A, Wang J, Langendam M, Verbakel JY, and Leeflang MMG
- Subjects
- Humans, SARS-CoV-2, COVID-19 Testing, Blood Coagulation Tests, COVID-19 diagnosis
- Published
- 2022
46. Antibody tests for identification of current and past infection with SARS-CoV-2.
- Author
-
Fox T, Geppert J, Dinnes J, Scandrett K, Bigio J, Sulis G, Hettiarachchi D, Mathangasinghe Y, Weeratunga P, Wickramasinghe D, Bergman H, Buckley BS, Probyn K, Sguassero Y, Davenport C, Cunningham J, Dittrich S, Emperador D, Hooft L, Leeflang MM, McInnes MD, Spijker R, Struyf T, Van den Bruel A, Verbakel JY, Takwoingi Y, Taylor-Phillips S, and Deeks JJ
- Subjects
- Humans, Antibodies, Viral, Immunoglobulin G, COVID-19 Vaccines, Pandemics, Seroepidemiologic Studies, Immunoglobulin M, SARS-CoV-2, COVID-19 diagnosis, COVID-19 epidemiology
- Abstract
Background: The diagnostic challenges associated with the COVID-19 pandemic resulted in rapid development of diagnostic test methods for detecting SARS-CoV-2 infection. Serology tests to detect the presence of antibodies to SARS-CoV-2 enable detection of past infection and may detect cases of SARS-CoV-2 infection that were missed by earlier diagnostic tests. Understanding the diagnostic accuracy of serology tests for SARS-CoV-2 infection may enable development of effective diagnostic and management pathways, inform public health management decisions and understanding of SARS-CoV-2 epidemiology., Objectives: To assess the accuracy of antibody tests, firstly, to determine if a person presenting in the community, or in primary or secondary care has current SARS-CoV-2 infection according to time after onset of infection and, secondly, to determine if a person has previously been infected with SARS-CoV-2. Sources of heterogeneity investigated included: timing of test, test method, SARS-CoV-2 antigen used, test brand, and reference standard for non-SARS-CoV-2 cases., Search Methods: The COVID-19 Open Access Project living evidence database from the University of Bern (which includes daily updates from PubMed and Embase and preprints from medRxiv and bioRxiv) was searched on 30 September 2020. We included additional publications from the Evidence for Policy and Practice Information and Co-ordinating Centre (EPPI-Centre) 'COVID-19: Living map of the evidence' and the Norwegian Institute of Public Health 'NIPH systematic and living map on COVID-19 evidence'. We did not apply language restrictions., Selection Criteria: We included test accuracy studies of any design that evaluated commercially produced serology tests, targeting IgG, IgM, IgA alone, or in combination. Studies must have provided data for sensitivity, that could be allocated to a predefined time period after onset of symptoms, or after a positive RT-PCR test. Small studies with fewer than 25 SARS-CoV-2 infection cases were excluded. We included any reference standard to define the presence or absence of SARS-CoV-2 (including reverse transcription polymerase chain reaction tests (RT-PCR), clinical diagnostic criteria, and pre-pandemic samples)., Data Collection and Analysis: We use standard screening procedures with three reviewers. Quality assessment (using the QUADAS-2 tool) and numeric study results were extracted independently by two people. Other study characteristics were extracted by one reviewer and checked by a second. We present sensitivity and specificity with 95% confidence intervals (CIs) for each test and, for meta-analysis, we fitted univariate random-effects logistic regression models for sensitivity by eligible time period and for specificity by reference standard group. Heterogeneity was investigated by including indicator variables in the random-effects logistic regression models. We tabulated results by test manufacturer and summarised results for tests that were evaluated in 200 or more samples and that met a modification of UK Medicines and Healthcare products Regulatory Agency (MHRA) target performance criteria., Main Results: We included 178 separate studies (described in 177 study reports, with 45 as pre-prints) providing 527 test evaluations. The studies included 64,688 samples including 25,724 from people with confirmed SARS-CoV-2; most compared the accuracy of two or more assays (102/178, 57%). Participants with confirmed SARS-CoV-2 infection were most commonly hospital inpatients (78/178, 44%), and pre-pandemic samples were used by 45% (81/178) to estimate specificity. Over two-thirds of studies recruited participants based on known SARS-CoV-2 infection status (123/178, 69%). All studies were conducted prior to the introduction of SARS-CoV-2 vaccines and present data for naturally acquired antibody responses. Seventy-nine percent (141/178) of studies reported sensitivity by week after symptom onset and 66% (117/178) for convalescent phase infection. Studies evaluated enzyme-linked immunosorbent assays (ELISA) (165/527; 31%), chemiluminescent assays (CLIA) (167/527; 32%) or lateral flow assays (LFA) (188/527; 36%). Risk of bias was high because of participant selection (172, 97%); application and interpretation of the index test (35, 20%); weaknesses in the reference standard (38, 21%); and issues related to participant flow and timing (148, 82%). We judged that there were high concerns about the applicability of the evidence related to participants in 170 (96%) studies, and about the applicability of the reference standard in 162 (91%) studies. Average sensitivities for current SARS-CoV-2 infection increased by week after onset for all target antibodies. Average sensitivity for the combination of either IgG or IgM was 41.1% in week one (95% CI 38.1 to 44.2; 103 evaluations; 3881 samples, 1593 cases), 74.9% in week two (95% CI 72.4 to 77.3; 96 evaluations, 3948 samples, 2904 cases) and 88.0% by week three after onset of symptoms (95% CI 86.3 to 89.5; 103 evaluations, 2929 samples, 2571 cases). Average sensitivity during the convalescent phase of infection (up to a maximum of 100 days since onset of symptoms, where reported) was 89.8% for IgG (95% CI 88.5 to 90.9; 253 evaluations, 16,846 samples, 14,183 cases), 92.9% for IgG or IgM combined (95% CI 91.0 to 94.4; 108 evaluations, 3571 samples, 3206 cases) and 94.3% for total antibodies (95% CI 92.8 to 95.5; 58 evaluations, 7063 samples, 6652 cases). Average sensitivities for IgM alone followed a similar pattern but were of a lower test accuracy in every time slot. Average specificities were consistently high and precise, particularly for pre-pandemic samples which provide the least biased estimates of specificity (ranging from 98.6% for IgM to 99.8% for total antibodies). Subgroup analyses suggested small differences in sensitivity and specificity by test technology however heterogeneity in study results, timing of sample collection, and smaller sample numbers in some groups made comparisons difficult. For IgG, CLIAs were the most sensitive (convalescent-phase infection) and specific (pre-pandemic samples) compared to both ELISAs and LFAs (P < 0.001 for differences across test methods). The antigen(s) used (whether from the Spike-protein or nucleocapsid) appeared to have some effect on average sensitivity in the first weeks after onset but there was no clear evidence of an effect during convalescent-phase infection. Investigations of test performance by brand showed considerable variation in sensitivity between tests, and in results between studies evaluating the same test. For tests that were evaluated in 200 or more samples, the lower bound of the 95% CI for sensitivity was 90% or more for only a small number of tests (IgG, n = 5; IgG or IgM, n = 1; total antibodies, n = 4). More test brands met the MHRA minimum criteria for specificity of 98% or above (IgG, n = 16; IgG or IgM, n = 5; total antibodies, n = 7). Seven assays met the specified criteria for both sensitivity and specificity. In a low-prevalence (2%) setting, where antibody testing is used to diagnose COVID-19 in people with symptoms but who have had a negative PCR test, we would anticipate that 1 (1 to 2) case would be missed and 8 (5 to 15) would be falsely positive in 1000 people undergoing IgG or IgM testing in week three after onset of SARS-CoV-2 infection. In a seroprevalence survey, where prevalence of prior infection is 50%, we would anticipate that 51 (46 to 58) cases would be missed and 6 (5 to 7) would be falsely positive in 1000 people having IgG tests during the convalescent phase (21 to 100 days post-symptom onset or post-positive PCR) of SARS-CoV-2 infection., Authors' Conclusions: Some antibody tests could be a useful diagnostic tool for those in whom molecular- or antigen-based tests have failed to detect the SARS-CoV-2 virus, including in those with ongoing symptoms of acute infection (from week three onwards) or those presenting with post-acute sequelae of COVID-19. However, antibody tests have an increasing likelihood of detecting an immune response to infection as time since onset of infection progresses and have demonstrated adequate performance for detection of prior infection for sero-epidemiological purposes. The applicability of results for detection of vaccination-induced antibodies is uncertain., (Copyright © 2022 The Authors. Cochrane Database of Systematic Reviews published by John Wiley & Sons, Ltd. on behalf of The Cochrane Collaboration.)
- Published
- 2022
- Full Text
- View/download PDF
47. TOMAS-R: A template to identify and plan analysis for clinically important variation and multiplicity in diagnostic test accuracy systematic reviews.
- Author
-
Mallett S, Dinnes J, Takwoingi Y, and de Ruffano LF
- Abstract
The Cochrane Handbook for Systematic Reviews of Diagnostic Test Accuracy (DTA) provides guidance on important aspects of conducting a test accuracy systematic review. In this paper we present TOMAS-R (Template of Multiplicity and Analysis in Systematic Reviews), a structured template to use in conjunction with current Cochrane DTA guidance, to help identify complexities in the review question and to assist planning of data extraction and analysis when clinically important variation and multiplicity is present. Examples of clinically important variation and multiplicity could include differences in participants, index tests and test methods, target conditions and reference standards used to define them, study design and methodological quality. Our TOMAS-R template goes beyond the broad topic headings in current guidance that are sources of potential variation and multiplicity, by providing prompts for common sources of heterogeneity encountered from our experience of authoring over 100 reviews. We provide examples from two reviews to assist users. The TOMAS-R template adds value by supplementing available guidance for DTA reviews by providing a tool to facilitate discussions between methodologists, clinicians, statisticians and patient/public team members to identify the full breadth of review question complexities early in the process. The use of a structured set of prompting questions at the important stage of writing the protocol ensures clinical relevance as a main focus of the review, while allowing identification of key clinical components for data extraction and later analysis thereby facilitating a more efficient review process., (© 2022. The Author(s).)
- Published
- 2022
- Full Text
- View/download PDF
48. Rapid, point-of-care antigen tests for diagnosis of SARS-CoV-2 infection.
- Author
-
Dinnes J, Sharma P, Berhane S, van Wyk SS, Nyaaba N, Domen J, Taylor M, Cunningham J, Davenport C, Dittrich S, Emperador D, Hooft L, Leeflang MM, McInnes MD, Spijker R, Verbakel JY, Takwoingi Y, Taylor-Phillips S, Van den Bruel A, and Deeks JJ
- Subjects
- COVID-19 Testing, Humans, Pandemics, Point-of-Care Systems, SARS-CoV-2, Sensitivity and Specificity, COVID-19 diagnosis
- Abstract
Background: Accurate rapid diagnostic tests for SARS-CoV-2 infection would be a useful tool to help manage the COVID-19 pandemic. Testing strategies that use rapid antigen tests to detect current infection have the potential to increase access to testing, speed detection of infection, and inform clinical and public health management decisions to reduce transmission. This is the second update of this review, which was first published in 2020., Objectives: To assess the diagnostic accuracy of rapid, point-of-care antigen tests for diagnosis of SARS-CoV-2 infection. We consider accuracy separately in symptomatic and asymptomatic population groups. Sources of heterogeneity investigated included setting and indication for testing, assay format, sample site, viral load, age, timing of test, and study design., Search Methods: We searched the COVID-19 Open Access Project living evidence database from the University of Bern (which includes daily updates from PubMed and Embase and preprints from medRxiv and bioRxiv) on 08 March 2021. We included independent evaluations from national reference laboratories, FIND and the Diagnostics Global Health website. We did not apply language restrictions., Selection Criteria: We included studies of people with either suspected SARS-CoV-2 infection, known SARS-CoV-2 infection or known absence of infection, or those who were being screened for infection. We included test accuracy studies of any design that evaluated commercially produced, rapid antigen tests. We included evaluations of single applications of a test (one test result reported per person) and evaluations of serial testing (repeated antigen testing over time). Reference standards for presence or absence of infection were any laboratory-based molecular test (primarily reverse transcription polymerase chain reaction (RT-PCR)) or pre-pandemic respiratory sample., Data Collection and Analysis: We used standard screening procedures with three people. Two people independently carried out quality assessment (using the QUADAS-2 tool) and extracted study results. Other study characteristics were extracted by one review author and checked by a second. We present sensitivity and specificity with 95% confidence intervals (CIs) for each test, and pooled data using the bivariate model. We investigated heterogeneity by including indicator variables in the random-effects logistic regression models. We tabulated results by test manufacturer and compliance with manufacturer instructions for use and according to symptom status., Main Results: We included 155 study cohorts (described in 166 study reports, with 24 as preprints). The main results relate to 152 evaluations of single test applications including 100,462 unique samples (16,822 with confirmed SARS-CoV-2). Studies were mainly conducted in Europe (101/152, 66%), and evaluated 49 different commercial antigen assays. Only 23 studies compared two or more brands of test. Risk of bias was high because of participant selection (40, 26%); interpretation of the index test (6, 4%); weaknesses in the reference standard for absence of infection (119, 78%); and participant flow and timing 41 (27%). Characteristics of participants (45, 30%) and index test delivery (47, 31%) differed from the way in which and in whom the test was intended to be used. Nearly all studies (91%) used a single RT-PCR result to define presence or absence of infection. The 152 studies of single test applications reported 228 evaluations of antigen tests. Estimates of sensitivity varied considerably between studies, with consistently high specificities. Average sensitivity was higher in symptomatic (73.0%, 95% CI 69.3% to 76.4%; 109 evaluations; 50,574 samples, 11,662 cases) compared to asymptomatic participants (54.7%, 95% CI 47.7% to 61.6%; 50 evaluations; 40,956 samples, 2641 cases). Average sensitivity was higher in the first week after symptom onset (80.9%, 95% CI 76.9% to 84.4%; 30 evaluations, 2408 cases) than in the second week of symptoms (53.8%, 95% CI 48.0% to 59.6%; 40 evaluations, 1119 cases). For those who were asymptomatic at the time of testing, sensitivity was higher when an epidemiological exposure to SARS-CoV-2 was suspected (64.3%, 95% CI 54.6% to 73.0%; 16 evaluations; 7677 samples, 703 cases) compared to where COVID-19 testing was reported to be widely available to anyone on presentation for testing (49.6%, 95% CI 42.1% to 57.1%; 26 evaluations; 31,904 samples, 1758 cases). Average specificity was similarly high for symptomatic (99.1%) or asymptomatic (99.7%) participants. We observed a steady decline in summary sensitivities as measures of sample viral load decreased. Sensitivity varied between brands. When tests were used according to manufacturer instructions, average sensitivities by brand ranged from 34.3% to 91.3% in symptomatic participants (20 assays with eligible data) and from 28.6% to 77.8% for asymptomatic participants (12 assays). For symptomatic participants, summary sensitivities for seven assays were 80% or more (meeting acceptable criteria set by the World Health Organization (WHO)). The WHO acceptable performance criterion of 97% specificity was met by 17 of 20 assays when tests were used according to manufacturer instructions, 12 of which demonstrated specificities above 99%. For asymptomatic participants the sensitivities of only two assays approached but did not meet WHO acceptable performance standards in one study each; specificities for asymptomatic participants were in a similar range to those observed for symptomatic people. At 5% prevalence using summary data in symptomatic people during the first week after symptom onset, the positive predictive value (PPV) of 89% means that 1 in 10 positive results will be a false positive, and around 1 in 5 cases will be missed. At 0.5% prevalence using summary data for asymptomatic people, where testing was widely available and where epidemiological exposure to COVID-19 was suspected, resulting PPVs would be 38% to 52%, meaning that between 2 in 5 and 1 in 2 positive results will be false positives, and between 1 in 2 and 1 in 3 cases will be missed., Authors' Conclusions: Antigen tests vary in sensitivity. In people with signs and symptoms of COVID-19, sensitivities are highest in the first week of illness when viral loads are higher. Assays that meet appropriate performance standards, such as those set by WHO, could replace laboratory-based RT-PCR when immediate decisions about patient care must be made, or where RT-PCR cannot be delivered in a timely manner. However, they are more suitable for use as triage to RT-PCR testing. The variable sensitivity of antigen tests means that people who test negative may still be infected. Many commercially available rapid antigen tests have not been evaluated in independent validation studies. Evidence for testing in asymptomatic cohorts has increased, however sensitivity is lower and there is a paucity of evidence for testing in different settings. Questions remain about the use of antigen test-based repeat testing strategies. Further research is needed to evaluate the effectiveness of screening programmes at reducing transmission of infection, whether mass screening or targeted approaches including schools, healthcare setting and traveller screening., (Copyright © 2022 The Authors. Cochrane Database of Systematic Reviews published by John Wiley & Sons, Ltd. on behalf of The Cochrane Collaboration.)
- Published
- 2022
- Full Text
- View/download PDF
49. COVID-19 rapid antigen testing strategies must be evaluated in intended use settings.
- Author
-
Dinnes J and Davenport C
- Abstract
Competing Interests: The authors declare no conflict of interest.
- Published
- 2022
- Full Text
- View/download PDF
50. Signs and symptoms to determine if a patient presenting in primary care or hospital outpatient settings has COVID-19.
- Author
-
Struyf T, Deeks JJ, Dinnes J, Takwoingi Y, Davenport C, Leeflang MM, Spijker R, Hooft L, Emperador D, Domen J, Tans A, Janssens S, Wickramasinghe D, Lannoy V, Horn SRA, and Van den Bruel A
- Subjects
- Aged, Anosmia diagnosis, Anosmia etiology, Artificial Intelligence, COVID-19 Testing, Child, Cough etiology, Dyspnea, Fatigue etiology, Fever diagnosis, Fever etiology, Hospitals, Humans, Outpatients, Primary Health Care, Prospective Studies, SARS-CoV-2, Sensitivity and Specificity, Ageusia complications, COVID-19 diagnosis, COVID-19 epidemiology, Pharyngitis
- Abstract
Background: COVID-19 illness is highly variable, ranging from infection with no symptoms through to pneumonia and life-threatening consequences. Symptoms such as fever, cough, or loss of sense of smell (anosmia) or taste (ageusia), can help flag early on if the disease is present. Such information could be used either to rule out COVID-19 disease, or to identify people who need to go for COVID-19 diagnostic tests. This is the second update of this review, which was first published in 2020., Objectives: To assess the diagnostic accuracy of signs and symptoms to determine if a person presenting in primary care or to hospital outpatient settings, such as the emergency department or dedicated COVID-19 clinics, has COVID-19., Search Methods: We undertook electronic searches up to 10 June 2021 in the University of Bern living search database. In addition, we checked repositories of COVID-19 publications. We used artificial intelligence text analysis to conduct an initial classification of documents. We did not apply any language restrictions., Selection Criteria: Studies were eligible if they included people with clinically suspected COVID-19, or recruited known cases with COVID-19 and also controls without COVID-19 from a single-gate cohort. Studies were eligible when they recruited people presenting to primary care or hospital outpatient settings. Studies that included people who contracted SARS-CoV-2 infection while admitted to hospital were not eligible. The minimum eligible sample size of studies was 10 participants. All signs and symptoms were eligible for this review, including individual signs and symptoms or combinations. We accepted a range of reference standards., Data Collection and Analysis: Pairs of review authors independently selected all studies, at both title and abstract, and full-text stage. They resolved any disagreements by discussion with a third review author. Two review authors independently extracted data and assessed risk of bias using the QUADAS-2 checklist, and resolved disagreements by discussion with a third review author. Analyses were restricted to prospective studies only. We presented sensitivity and specificity in paired forest plots, in receiver operating characteristic (ROC) space and in dumbbell plots. We estimated summary parameters using a bivariate random-effects meta-analysis whenever five or more primary prospective studies were available, and whenever heterogeneity across studies was deemed acceptable., Main Results: We identified 90 studies; for this update we focused on the results of 42 prospective studies with 52,608 participants. Prevalence of COVID-19 disease varied from 3.7% to 60.6% with a median of 27.4%. Thirty-five studies were set in emergency departments or outpatient test centres (46,878 participants), three in primary care settings (1230 participants), two in a mixed population of in- and outpatients in a paediatric hospital setting (493 participants), and two overlapping studies in nursing homes (4007 participants). The studies did not clearly distinguish mild COVID-19 disease from COVID-19 pneumonia, so we present the results for both conditions together. Twelve studies had a high risk of bias for selection of participants because they used a high level of preselection to decide whether reverse transcription polymerase chain reaction (RT-PCR) testing was needed, or because they enrolled a non-consecutive sample, or because they excluded individuals while they were part of the study base. We rated 36 of the 42 studies as high risk of bias for the index tests because there was little or no detail on how, by whom and when, the symptoms were measured. For most studies, eligibility for testing was dependent on the local case definition and testing criteria that were in effect at the time of the study, meaning most people who were included in studies had already been referred to health services based on the symptoms that we are evaluating in this review. The applicability of the results of this review iteration improved in comparison with the previous reviews. This version has more studies of people presenting to ambulatory settings, which is where the majority of assessments for COVID-19 take place. Only three studies presented any data on children separately, and only one focused specifically on older adults. We found data on 96 symptoms or combinations of signs and symptoms. Evidence on individual signs as diagnostic tests was rarely reported, so this review reports mainly on the diagnostic value of symptoms. Results were highly variable across studies. Most had very low sensitivity and high specificity. RT-PCR was the most often used reference standard (40/42 studies). Only cough (11 studies) had a summary sensitivity above 50% (62.4%, 95% CI 50.6% to 72.9%)); its specificity was low (45.4%, 95% CI 33.5% to 57.9%)). Presence of fever had a sensitivity of 37.6% (95% CI 23.4% to 54.3%) and a specificity of 75.2% (95% CI 56.3% to 87.8%). The summary positive likelihood ratio of cough was 1.14 (95% CI 1.04 to 1.25) and that of fever 1.52 (95% CI 1.10 to 2.10). Sore throat had a summary positive likelihood ratio of 0.814 (95% CI 0.714 to 0.929), which means that its presence increases the probability of having an infectious disease other than COVID-19. Dyspnoea (12 studies) and fatigue (8 studies) had a sensitivity of 23.3% (95% CI 16.4% to 31.9%) and 40.2% (95% CI 19.4% to 65.1%) respectively. Their specificity was 75.7% (95% CI 65.2% to 83.9%) and 73.6% (95% CI 48.4% to 89.3%). The summary positive likelihood ratio of dyspnoea was 0.96 (95% CI 0.83 to 1.11) and that of fatigue 1.52 (95% CI 1.21 to 1.91), which means that the presence of fatigue slightly increases the probability of having COVID-19. Anosmia alone (7 studies), ageusia alone (5 studies), and anosmia or ageusia (6 studies) had summary sensitivities below 50% but summary specificities over 90%. Anosmia had a summary sensitivity of 26.4% (95% CI 13.8% to 44.6%) and a specificity of 94.2% (95% CI 90.6% to 96.5%). Ageusia had a summary sensitivity of 23.2% (95% CI 10.6% to 43.3%) and a specificity of 92.6% (95% CI 83.1% to 97.0%). Anosmia or ageusia had a summary sensitivity of 39.2% (95% CI 26.5% to 53.6%) and a specificity of 92.1% (95% CI 84.5% to 96.2%). The summary positive likelihood ratios of anosmia alone and anosmia or ageusia were 4.55 (95% CI 3.46 to 5.97) and 4.99 (95% CI 3.22 to 7.75) respectively, which is just below our arbitrary definition of a 'red flag', that is, a positive likelihood ratio of at least 5. The summary positive likelihood ratio of ageusia alone was 3.14 (95% CI 1.79 to 5.51). Twenty-four studies assessed combinations of different signs and symptoms, mostly combining olfactory symptoms. By combining symptoms with other information such as contact or travel history, age, gender, and a local recent case detection rate, some multivariable prediction scores reached a sensitivity as high as 90%., Authors' Conclusions: Most individual symptoms included in this review have poor diagnostic accuracy. Neither absence nor presence of symptoms are accurate enough to rule in or rule out the disease. The presence of anosmia or ageusia may be useful as a red flag for the presence of COVID-19. The presence of cough also supports further testing. There is currently no evidence to support further testing with PCR in any individuals presenting only with upper respiratory symptoms such as sore throat, coryza or rhinorrhoea. Combinations of symptoms with other readily available information such as contact or travel history, or the local recent case detection rate may prove more useful and should be further investigated in an unselected population presenting to primary care or hospital outpatient settings. The diagnostic accuracy of symptoms for COVID-19 is moderate to low and any testing strategy using symptoms as selection mechanism will result in both large numbers of missed cases and large numbers of people requiring testing. Which one of these is minimised, is determined by the goal of COVID-19 testing strategies, that is, controlling the epidemic by isolating every possible case versus identifying those with clinically important disease so that they can be monitored or treated to optimise their prognosis. The former will require a testing strategy that uses very few symptoms as entry criterion for testing, the latter could focus on more specific symptoms such as fever and anosmia., (Copyright © 2022 The Authors. Cochrane Database of Systematic Reviews published by John Wiley & Sons, Ltd. on behalf of The Cochrane Collaboration.)
- Published
- 2022
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.