In this issue of KT Update Dr. Marcel Dijkers shares information about duplication of evidence within systematic reviews. How can authors ensure that multiple sources do not include duplicates of the evidence? This article describes the problem and some possible solutions for deduplication.
Duplicate Publications and Systematic Reviews:
Problems and Proposals
Marcel Dijkers, PhD, FACRM
Icahn School of Medicine at Mount Sinai
Department of Rehabilitation Medicine
[ Download PDF version 194kb ]
“Double dipped? What…what…what are you talking about?” George Costanza
The Popularity of Systematic Reviews
Disability and rehabilitation researchers increasingly use systematic reviews (SRs) to summarize the evidence that exists with regard to a particular issue, such as the effectiveness of an intervention, the quality of an outcome measure, or the prognosis after a sudden-onset disorder. The National Rehabilitation Information Center, the library of the National Institute on Disability, Independent Living, and Rehabilitation Research (NIDILRR), which collects the articles, reports, curricula, guides, and other publications and products of the research projects funded by NIDILRR, contains the following numbers of documents submitted in recent years that have “systematic review” in their title or abstract (search conducted July 23, 2017):
2010: 78
2011: 85
2012: 90
2013: 110
2014: 142
2015: 122
2016: 117
2017: 49
A total of 793 documents have been published over 7.5 years, for an average of 105 per year. Some of these documents may not be SRs, but papers reporting on the methodology of SRs, or otherwise referring to SRs. A scan of the titles of the documents published in 2017 revealed that 41 were SRs (with or without meta-analysis), three described SR methodology, and five were “other.”
- Footnote 1
-
1 David, L. (Creator), Seinfeld, J. (Creator), Mehlman, P. (Writer), & Cherones, T. (Director). (1993). The implant [Television series episode, season 4, episode 19]. In L. David, A. Scheinman, G. Shapiro, & H. West (Executive Producers), Seinfeld. Beverly Hills, CA: West/Shapiro.
The Problem of Duplicate Publication
SRs aim to collect all the relevant evidence (although it is common that non-English publications are excluded, and other convenience exclusions are made), carefully assess the quality of the study that generated each piece of evidence, and then combine the findings across studies to answer the question(s) that led to the review. Combining can be done quantitatively (producing a meta-analysis) or qualitatively. In both approaches, weighting of the studies–based on sample size, study quality, or other factors–may play a role. It goes without saying that each study should be included once only, which means that each unit of analysis (most commonly, a patient, client, or research subject) is counted only once.
Using multiple bibliographic databases (some say at least three should be consulted) to find potential studies is proper SR procedure, which may be supplemented by ancestor and descendant searches, hand searching of key journals, searches of the gray literature, contacts with authors, and other means of finding published and unpublished reports (Task Force on Systematic Review and Guidelines, 2013). This searching often leads to multiple inclusions of the same paper, which need to be deduplicated, a stage typically shown in the CONSORT flow diagram depicting the various steps in the identification of publications to be extracted (Schulz, Altman, Moher, & CONSORT Group, 2010). A reference manager such as Endnote or RefWorks is typically used to do this (Qi et al., 2013). Given the poor performance of reference managers, which require much double-checking of the results by the SR authors and deletion of duplicates that slipped through, other approaches have been developed (Rathbone, Carter, Hoffmann, & Glasziou, 2015).
Reference managers and these other approaches serve only to detect duplicate publication titles. A second step is needed to determine whether there is duplicate publication of the same material in manifold papers. There are legitimate reasons for conveying study results in multiple papers, for instance, follow-up outcomes are reported after the initial outcomes have been published, or a different analytical method (preferably, one not available when a first paper was written) is used to have another look at the sample or a subgroup (Jamrozik, 2004). These legitimate additional papers generally refer to the earlier publication(s) that have taken place so that readers (and SR authors) can easily find all reports resulting from a single study. However, publish or perish and similar pressures on researchers have led to many less legitimate follow-on reports, ranging from pure duplicates (although possibly worded in a slightly different language) to multiple publications each addressing a different outcome (the “salami science” approach to padding one’s résumé). One study (Abby, Massey, Galandiuk, & Polk, 1994) found that one first author had 83 publications that “expressed the same theme 83 different ways” (p. 107); in three of these, the Methods, Results, tables, and figures were virtually identical.
If the outcomes for one study sample are reported twice without detection by an SR author, this may have various impacts on the findings of the SR. In the case of intervention studies, duplicate entry in the evidence table of a study with a low effect size may bring down the pooled effect size. In the case of studies with a high effect size, the opposite may happen. Even double counting a study with an effect size that is about average for the entire set of studies may have a deleterious consequence: the confidence interval around the pooled effect size is reduced, potentially changing the interpretation of the clinical significance of the intervention. Similar problems may be the result of “double dipping” in SRs of a diagnostic measure, prognosis, or economic impact. Given the strong lure of “p < .05” for authors and editors, the chance of “significant” studies with duplicate reports being found is greater. As the Cochrane Handbook for Systematic Reviews of Interventions states, “Studies with significant results are more likely to lead to multiple publications and presentations… which makes it more likely that they will be located and included in a meta-analysis” (Higgins & Green, 2011, section 10.2.2.1).
Frequency and Types of Duplicate Publication
This issue has been on the radar of SR authors and SR methodologists for a long time (Bailey, 2002; Gøtzsche, 1989; Tramèr, Reynolds, Moore, & McQuay, 1997). von Elm, Poglia, Walder, and Tramèr (2004) analyzed 141 SRs in the anesthesia and analgesia domain, and found that 56 of these had to address duplicate publication issues (although 14 had not mentioned it in their report but responded positively to the authors’ queries). These 56 systematic reviews included 1,131 main articles and excluded 103 duplicates that originated from 78 main articles. “Sixty articles were published twice, 13 three times, 3 four times, and 2 five times” (von Elm et al., 2004, p. 974). The authors analyzed the papers involved, and identified six duplication patterns:
- The sample and outcomes reported were the same (n = 21 pairs). In three quarters of these instances, the secondary article did not refer to the prior publication; 29% were translations.
- Two or more prior articles were “assembled to produce yet another article” (n = 16) (von Elm et al., 2004, p. 977).
- The same sample was used but different outcomes were reported (n = 24).
- A larger sample (additional cases) was used, but identical outcomes were described (n = 11).
- A smaller sample (subgroup of the cases in a large trial) and identical outcomes were reported (n = 11).
- Different samples and different outcomes were conveyed (n = 20). “In pattern [6] duplicates, both samples and outcomes were different from the main article. Confirmation of duplication was only possible through contact with the original authors” (von Elm et al., 2004, p. 977).
von Elm et al. (2004) concluded, “Duplication goes beyond simple copying. Six distinct duplication patterns were identified after comparing study samples and outcomes of duplicates and corresponding main articles. Authorship was an unreliable criterion” (p. 974).
Suggestions for Systematic Review Authors
The key phrase here for prospective SR authors is, “Authorship was an unreliable criterion.” The authors of the Cochrane Handbook (Higgins & Green, 2011, section 10.2.2.1) state: “There are examples where two articles reporting the same trial do not share a single common author…. Thus, it may be difficult or impossible for review authors to determine whether two papers represent duplicate publications of one study or two separate studies without contacting the authors, which may result in biasing a meta-analysis of this data.” These authors also assert that it can be difficult to detect duplicate publication, and detective work may be required on the part of SR authors (Higgins & Green, 2011, section 7.2.2). According to the Handbook, some of the most useful criteria for comparing reports are:
- Names of authors (most duplicate reports have authors in common, although that is not always the case);
- Location and setting of the study (particularly if institutions, such as hospitals, are named);
- Unambiguous details of the interventions (e.g., dose, frequency);
- The number of participants and the baseline data reported for them; and
- The date and duration of the study, including recruitment periods and follow-up periods.
(I suggest that the following can be added: grant number and granting agency; name of uncommon outcome measures). Higgins and Green (2011) indicate that where uncertainties remain after considering these and other factors, it may be necessary to correspond with the authors of the reports.
Despite of the commonness of duplicate publication and the widespread knowledge of its existence, the issue is not always addressed in handbooks for SR authors. For instance, the Joanna Briggs Institute (JBI) Reviewers’ Manual is “designed to provide authors with a comprehensive guide to conducting JBI systematic reviews. It describes in detail the process of planning, undertaking and writing up a systematic review of qualitative, quantitative, economic, text and opinion based evidence” (Joanna Briggs Institute, 2014, p. 9). Even so, the JBI Reviewer’s Manual describes title deduplication using a reference manager, but it does not explain how to address duplicate publication.
The following eight steps are recommended to identify duplicate publication:
Step 1. Create an alphabetical list of all authors (not just the primary ones) of all candidate studies, with all coauthors for each author provided on the same line. Table 1 shows an example, with the entries for two hypothetical papers, Cleary, Jones and Carron (2012) and Jones, Williams and Smith (2015).
Table 1. Example of Alphabetical Author List
First Author |
Coauthors |
Cleary |
Clarron, Jones |
Clarron |
Cleary, Jones |
Jones |
Cleary, Clarron |
Jones |
Williams, Smith |
Smith |
Jones, Williams |
Williams |
Jones, Smith |
This type of table reveals that Jones may have published the same material twice, once as a coauthor of Cleary, and once as first author with the assistance of Smith and Williams.
Step 2. Analyze this list to identify authors who appear on two or more publications, as first author or coauthor.
Step 3. Obtain the full texts of the papers involved.
Step 4. Analyze the references the authors make to earlier, later, or parallel papers by any of the team members or associates.
Step 5. If there is no citing of such reports, analyze in all two, three, four, and so on, suspected papers (i.e., with at least one shared author) what is written about the sample, baseline demographic and clinical data, the intervention (if applicable), and key outcome measures. Determine whether there is any double reporting and, if so, of which von Elm pattern.
Step 6. If there is no clear-cut evidence of an overlap in subjects and no clear-cut evidence of a lack of overlap, write to the authors for clarification.
Step 7: Based on the results of this detective work, in the SR use the best report (von Elm pattern 1); a combination of the findings in two or more reports (pattern 3); the later paper with more subjects (pattern 4); the original paper, unless the subsample analysis has more useful data (pattern 5); or a combination of the articles (pattern 2). With a response from the authors, pattern 6 duplication instances presumably can be reassigned to patterns 1–5. No guidance can be given when there is no response from any of the (co)authors involved; SR authors should take their best guess, but it would not be wrong for them to add a comment such as: “Report x and report y possibly describe the same sample; however, this could not be confirmed.”
Step 8: Steps 1–7 are predicated on overlapping authorships. Just in case there are multiple reports on a study that do not share a single author, at least one SR author should read all papers still being considered as potential sources of evidence. This SR author should be armed with a high level of suspicion to ferret out remaining duplications. Any suspicious twins or triplets should be submitted to the step 5, and as necessary, step 6 actions.
Some Personal Experience
Two current projects have brought the importance of, and difficulty of, detecting duplicate publication to my attention. The first project is an SR of the frequency of, reasons for, costs of, and impacts of rehospitalization of individuals with spinal cord injury (SCI) after discharge from initial inpatient rehabilitation. A key set of papers are those based on data from the NIDILRR-supported SCI Model Systems, as deposited in the SCI National Data Base (NDB). The question “In the past year, have you been hospitalized? (and if so, for what and how long?)” has been part of the SCI NDB since 2000, and in that period there have been 11 publications by varying author teams. Here, determining that these publications are related is not difficult because each publication refers to the SCI NDB as the source of the data.
The difficult issue is what to do with the papers. Generally, the most recent publication tends to be the largest because it contains more follow-ups for more patients (Year 1, 5, 10, 15, and so on, after initial rehabilitation discharge). However, specifically what is analyzed and how it may vary from one report to the next. (Were multiple readmissions within a year all considered? Added up to get a cumulative length of stay?) Data from multiple publications may be of relevance, even if there is overlap of patients and follow-up years.
This issue occurs with all large and longitudinal databases to which geographically dispersed researchers contribute new cases as well as follow-ups on previously registered cases, all of whom have a right to analyze the entire data set. Within the NIDILRR-supported SCI, traumatic burn injury (TBI) and burn model systems, there is some control on wild growth of duplicative analyses: authors or author groups considering an analysis need to inform their colleagues of their plans. In the TBI Model Systems, this process is known as the notification system. All insiders and outsiders wanting to use the database need to inform all current project directors of the nature of their analysis plans. The project directors can object to an analysis because another group is already working on the same or an overlapping question, or has published a paper in recent years.
The second project colleagues and I are finishing is a systematic review of systematic reviews–an overview of SRs–of the use of exoskeletons with patient with neurological disorders (mostly, stroke and SCI). Being familiar with the SCI literature (Bryce, Dijkers, & Kozlowski, 2015), I had a high level of suspicion, which was borne out by a careful comparison of the primary study reports. Most studies in this research area are small case series, and the authors tend to provide a table listing demographic and clinical information for each subject. A comparison of this information suggested that it was highly unlikely that, for example, “Jones & Williams, 2015” studied seven subjects who did not overlap at all with the eight subjects examined by “Williams & Jones, 2016” (Dijkers, Akers, Galen, Patzer, & Vu, 2016). Therefore, one question to be addressed in our SR of SRs is: Do the secondary studies (i.e. the SRs) indicate whether they have searched for duplication of cases in the primary studies, and what amount of duplication does exist, with or without the SR authors having looked for it?
To address this question, we created a list of all authors of all primary studies referenced by all SRs (step 1 above), sorted them into author groups, and obtained the full text of each publication. We analyzed these following steps 2–6. In some instances, it was clear that there was no duplication. In other situations, duplication was mathematically very likely, to say the least. We are following up with at least one author of all suspected studies to confirm our misgivings.
We found 13 SRs, each of which was based on three to 26 out of a total of 98 different primary studies. Only three of the SRs mentioned issues of double dipping in the primary studies–and one of these three did not exclude the duplicates.
Recommendations
What is the bottom line for SR authors and users? First, duplicate reporting exists in disability and rehabilitation studies as much as in other areas of scholarly activity. Authors report the same study in two or more publications, for legitimate reasons (addition of significant new numbers of cases to a model system database) and illegitimate ones (republication of essentially the same data in a new paper to cope with “publish or perish”). It behooves the authors of SRs to take all reasonable efforts to find related papers, delete the complete duplicates (von Elm’s pattern 1), and carefully consider how to handle all others (patterns 2–6). SR authors should avoid double dipping, and in doing so will escape offering misleading if not completely wrong conclusions and advice.
The following recommendations are made for the various entities involved in the SR enterprise:
- For authors of reporting guidelines for primary studies (e.g., CONSORT): Require that authors state what earlier reports on a study exist, and how they relate to the present paper (http://www.consort-statement.org).
- For authors of reporting guidelines for SRs and other review papers (e.g., PRISMA): Require that authors describe how they searched for and handled duplicate publications (http://www.prisma-statement.org).
- For authors of primary studies: Describe clearly what earlier reports on a study exist and how they relate to the current paper.
- For authors of SRs and other review papers: Explain how duplicate publications were searched for and how evidence from duplicate and overlapping studies (if identified) was handled in the evidence synthesis.
- For journal editors:
- Demand that the authors of primary studies report on parallel publications, and that they submit a copy of these papers to allow editors and peer reviewers to determine whether the manuscript newly submitted really produces novel information that expands our knowledge base.
- Require that the authors of SRs and similar manuscripts describe how steps 1–8 were handled in data extraction and synthesis, of qualitative and quantitative information.
- Ask peer reviewers (who are presumably selected for their expertise in the area addressed by the paper they are requested to review) whether they are familiar with similar papers already in the literature.
- For collective owners of databases (like the SCI NDB): Create a mechanism by which newly proposed data analyses are made known to all parties with a proprietary interest.
With these steps, illegitimate doubling of reports can be minimized in SRs, and legitimate duplication can be more easily identified by SR authors. Until these measures are implemented across all areas of rehabilitation and disability research, reader beware: The systematic review you are reading and considering for implementation in your practice could have a serious flaw.
References:
Abby, M., Massey, M. D., Galandiuk, S., & Polk, H. C. Jr. (1994). Peer review is an effective screening process to evaluate medical manuscripts. JAMA, 272(2), 105–107.
Bailey, B. J. (2002). Duplicate publication in the field of otolaryngology-head and neck surgery. Otolaryngology-Head and Neck Surgery, 126(3), 211–216.
Bryce, T. N., Dijkers, M. P., & Kozlowski, A. J. (2015). Framework for assessment of the usability of lower-extremity robotic exoskeletal orthoses. American Journal of Physical Medicine & Rehabilitation, 94(11), 1000-1014.
Dijkers, M. P., Akers, K. G., Galen, S. S., Patzer, D. E., & Vu, P. T. (2016). Letter to the editor regarding "Clinical effectiveness and safety of powered exoskeleton-assisted walking in patients with spinal cord injury: systematic review with meta-analysis." Medical Devices: Evidence and Research, 9, 419–421.
Gøtzsche, P. C. (1989). Multiple publication of reports of drug trials. European Journal of Clinical Pharmacology, 36(5), 429–432.
Higgins, J. P. T., & Green, S. (Eds.). (2011). Cochrane handbook for systematic reviews of interventions (Version 5.1.0). London, UK: The Cochrane Collaboration. Retrieved from http://handbook-5-1.cochrane.org/
Jamrozik, K. (2004). Of sausages and salami. Australian and New Zealand Journal of Public Health, 28(1), 5–6.
JBI. (2014). Joanna Briggs Institute reviewers’ manual: 2014 edition. Adelaide, Australia: Author. Retrieved from https://wiki.jbi.global/display/MANUAL
Qi, X., Yang, M., Ren, W., Jia, J., Wang, J., Han, G., & Fan, D. (2013). Find duplicates among the PubMed, EMBASE, and Cochrane Library Databases in systematic review. PLoS One, 8(8), e71838. doi:10.1371/journal.pone.0071838
Rathbone, J., Carter, M., Hoffmann, T., & Glasziou, P. (2015). Better duplicate detection for systematic reviewers: Evaluation of Systematic Review Assistant-Deduplication Module. Systematic Review, 4(6). doi:10.1186/2046-4053-4-6.
Schulz, K. F., Altman, D. G., Moher, D., & CONSORT Group. (2010). CONSORT 2010 Statement: Updated guidelines for reporting parallel group randomised trials. Journal of Clinical Epidemiology, 63(8), 834–840.
Task Force on Systematic Review and Guidelines. (2013). Assessing the quality and applicability of systematic reviews (AQASR). Austin, TX: SEDL, Center on Knowledge Translation for Disability and Rehabilitation Research.
Tramèr, M. R., Reynolds, D. J., Moore, R. A., & McQuay, H. J. (1997). Impact of covert duplicate publication on meta-analysis: a case study. BMJ, 315(7109), 635–640.
von Elm, E., Poglia, G., Walder, B., & Tramèr, M. R. (2004). Different patterns of duplicate publication: An analysis of articles used in systematic reviews. JAMA, 291(8), 974–980.