Journal of Economic Surveys
ARTICLE

THE MEANING OF FAILED REPLICATIONS: A REVIEW AND PROPOSAL

Michael A. Clemens,

Corresponding Author

Center for Global Development and IZA

Corresponding author contact email: mclemens@cgdev.org; Tel: +1-202-416-4000.Search for more papers by this author
First published: 26 December 2015
Citations: 93
Get access to the full version of this article. View access options below.
Institutional Login
Loading institution options...
Log in to Wiley Online Library

If you have previously obtained access with your personal account, please log in.

Purchase Instant Access
    • View the article PDF and any associated supplements and figures for a period of 48 hours.
    • Article can not be printed.
    • Article can not be downloaded.
    • Article can not be redistributed.
    • Unlimited viewing of the article PDF and any associated supplements and figures.
    • Article can not be printed.
    • Article can not be downloaded.
    • Article can not be redistributed.
    • Unlimited viewing of the article/chapter PDF and any associated supplements and figures.
    • Article/chapter can be printed.
    • Article/chapter can be downloaded.
    • Article/chapter can not be redistributed.

Abstract

The welcome rise of replication tests in economics has not been accompanied by a consensus standard for determining what constitutes a replication. A discrepant replication, in current usage of the term, can signal anything from an unremarkable disagreement over methods to scientific incompetence or misconduct. This paper proposes a standard for classifying one study as a replication of some other study. It is a standard that places the burden of proof on a study to demonstrate that it should have obtained identical results to the original, a conservative standard that is already used implicitly by many researchers. It contrasts this standard with decades of unsuccessful attempts to harmonize terminology, and argues that many prominent results described as replication tests should not be described as such. Adopting a conservative standard like this one can improve incentives for researchers, encouraging more and better replication tests.

The full text of this article hosted at iucr.org is unavailable due to technical difficulties.