Plagiarism and Duplicate Publication: Policies and Detection in Science
Academic publishing runs on trust — the assumption that what appears in a journal represents original work, honestly attributed. Plagiarism and duplicate publication are the two most common ways that trust breaks down, and the consequences reach far beyond the authors involved. This page covers how both violations are defined, how detection systems work, the scenarios that routinely catch researchers off guard, and where the genuinely difficult judgment calls lie.
Definition and scope
Plagiarism in scientific publishing means presenting someone else's text, ideas, data, or images as one's own without appropriate attribution. The Committee on Publication Ethics (COPE), which publishes freely available guidelines at publicationethics.org, treats plagiarism as a spectrum: verbatim copying of another author's prose sits at one end; paraphrasing without citation sits somewhere in the middle; and unattributed re-use of one's own previously published work — called self-plagiarism or text recycling — occupies its own contested category.
Duplicate publication is a distinct but related problem. It refers to publishing the same study, or substantially the same data, in more than one journal without disclosure. The Office of Research Integrity (ORI) at the U.S. Department of Health and Human Services (ori.hhs.gov) includes duplicate publication in its definition of research misconduct when the intent is to inflate a researcher's publication record or mislead the scientific record.
The scope is not trivial. A 2015 analysis published in PNAS estimated that the majority of retracted papers in biomedical literature involved either fraud or suspected fraud, with plagiarism and duplicate publication among the leading documented causes tracked in the Retraction Watch database (retractionwatch.com).
How it works
Detection now relies primarily on automated similarity-checking software. iThenticate, developed by Turnitin, is the dominant tool in academic publishing — used by journals across Elsevier, Springer Nature, Wiley, and the American Chemical Society, among others. The software compares submitted manuscripts against published literature, preprint servers, and a proprietary cross-publisher database called CrossCheck, maintained through Crossref.
The output is a similarity score expressed as a percentage. A score of 20% is not automatically alarming — boilerplate methods sections, standard chemical nomenclature, and correctly quoted passages inflate the number without indicating misconduct. Editors are trained to look at which sections generate matches, not just the headline figure. A methods section at 40% similarity may be unremarkable; an introduction at 40% warrants scrutiny.
For image manipulation — a growing concern in cell biology and materials science — journals increasingly use tools like ImageTwin or manual forensic review. The Journal of Cell Biology pioneered systematic image screening in the early 2000s and found problems in roughly 1% of accepted manuscripts, a figure that prompted widespread adoption of pre-publication screening across the life sciences.
Common scenarios
The situations that produce misconduct findings are rarely cartoonish acts of deliberate theft. The four patterns that appear most often in COPE case consultations:
- Salami slicing — splitting a single study into two or more papers to maximize publication count, with overlapping data published in different journals without cross-disclosure.
- Duplicate submission — submitting the same manuscript to two journals simultaneously, which most journals prohibit explicitly in their author guidelines.
- Translated self-plagiarism — publishing a paper in English that substantially reproduces a paper previously published by the same author in another language, without acknowledgment.
- Mosaic plagiarism — stitching together sentences from multiple sources, slightly rephrased, rather than copying a single source verbatim. Similarity software catches this less reliably than block copying.
Text recycling in methods sections sits in genuinely murky territory. Describing a validated protocol using the same words used to describe it previously is arguably accurate reporting. The distinction that COPE draws — and that journals increasingly codify in policy — is whether the recycled content is disclosed and whether it constitutes the core intellectual contribution of the new paper.
The research ethics and publication standards landscape has formalized around these distinctions, with most major publishers now requiring authors to declare prior related publications at submission.
Decision boundaries
Not every similarity flag is misconduct, and not every undisclosed duplicate is equally serious. The factors that push a case from policy violation toward formal misconduct investigation include:
- Intent — Was the duplication accidental (two conference papers merged without noticing overlap) or deliberate (submitting identical work to game a promotion threshold)?
- Materiality — Does the overlap affect the scientific conclusions, or is it confined to background prose?
- Disclosure opportunity — Did the submission process provide a clear mechanism to disclose related work, and did the author ignore it?
The contrast between plagiarism and duplicate publication matters here. Plagiarism harms the original author by stealing attribution. Duplicate publication primarily harms the scientific record by creating false signal — a study that appears to replicate a finding but is actually the same study cited twice. Both are serious; they damage scientific publishing in different ways and typically trigger different institutional responses.
Understanding how these violations sit within the broader ecosystem of publishing integrity is part of what the Scientific Journal Authority covers across topics from retractions and corrections in science to the peer review process. The policies governing plagiarism do not exist in isolation — they connect directly to how editors handle manuscript submission and how preprint servers interact with peer-reviewed journals in ways that create new disclosure questions.
References
- Committee on Publication Ethics (COPE) — Guidelines and Case Consultations
- Office of Research Integrity (ORI), U.S. Department of Health and Human Services
- Crossref — CrossCheck / iThenticate Program
- Retraction Watch Database — Center for Scientific Integrity
- COPE Text Recycling Guidelines