Skip to main content
Printable page generated Thursday, 27 November 2025, 11:19 AM
Use 'Print preview' to check the number of pages and printer settings.
Print functionality varies between browsers.
Unless otherwise stated, copyright © 2025 The Open University, all rights reserved.
Printable page generated Thursday, 27 November 2025, 11:19 AM

Integrity: Challenging questionable research practices

Introduction

  

The images shows five hands making the ‘thumbs up’ sign, and five hands making the ‘thumbs down’ sign

  

In Week 1, we broke down the idea of open research into three key facets that make research open: transparency, integrity and accessibility. We’re now going to take a deeper dive into integrity: how trustworthy a study is.

This week, you will learn how to recognise and avoid questionable research practices. You will discover why it is often important to be able to replicate other researchers’ findings, and how to go about doing this. You will experience an important test for the integrity of a piece of research: its replicability.

  

Replicability

There may be cases where you don’t expect to get the same results if you conduct the same study again. For example, if a study is based around a specific political event, it may be difficult or even impossible to replicate. But in many other types of investigation, we would expect to be able to get the same results when we run the same study again.

In Week 2, we talked about reproducibility – being able to get the same results when conducting the same analyses on the same data as the original study. Replicability is similar, in that it’s about getting the same results as the original study when running the same analyses, however, the difference is that now these analyses are run on new data. So, replication means conducting the same study again, and seeing if you get the same results.

Replication studies are deliberate attempts to do this. But what does the ‘same study’ mean? There are always going to be differences between the original study and the replication study. Replication studies vary on a spectrum from ‘direct’ to ‘conceptual’. Direct replications try to stay as close to the original study as possible, whereas conceptual replications purposefully vary some aspects to better understand the underlying phenomenon. Here are some examples, from most direct (the first) to most conceptual (the last):

  • A researcher makes a surprising finding in their research. To test whether they should rely on this result, they conduct a replication immediately after, using all the same materials and the same participant pool.
  • A researcher wants to replicate a study they’ve read about. The study is much older (from the 1990s), when open materials were not common. They only have the methods described in the original short paper to refer to, so they interpret these as best they can.
  • A researcher wants to replicate a study they’ve read about. They don’t think the original study was well-designed, but they think the hypothesis is interesting so they design a new study testing the same hypothesis but in a different way.

  

Now let’s dig deeper into the process of designing a replication study.

Replication studies

In the next video, psychologist Priya Silverstein talks about their first forays into conducting a replication study, and lessons learned. As you watch the video, think about what Priya’s results tell us about the process of running a good replication study.

Download this video clip.Video player: Video 2: Conducting a replication study
Copy this transcript to the clipboard
Print this transcript
Show transcript|Hide transcript
Video 2: Conducting a replication study
Interactive feature not available in single page view (see it in standard view).

Use this box to write your comments on what Priya advises.

Allow about 10 minutes for this.

To use this interactive functionality a free OU account is required. Sign in or register.
Interactive feature not available in single page view (see it in standard view).

When you are ready, press 'reveal' to see our comments.

Discussion

Priya suggests getting in touch with the authors of the original study and asking for more detail than a published paper provides. Using a larger sample size increases confidence that your findings do (or don’t) support those of the previous study. Priya also recommends submitting a registered report, to increase your chances of getting published.

Limits to replication

There are fields and methodologies where the value of replication is hotly debated. For instance:

  • Some argue that replication should be encouraged in qualitative research, whereas others argue that there are still open questions about whether replication is possible, desirable, or even aligned with the fundamental principles of qualitative research.
  • Economics has had a long history with replication studies, but not under this name. In economics, replication often takes place as ‘robustness checks’, where researchers test if their results hold when they use different datasets.
  • Research in the humanities is primarily interpretive and context-specific, focusing on understanding human experiences, cultures, texts, and historical events. This interpretive nature makes exact replication more challenging.

  

It is important to think carefully about whether replication makes sense for your field and methodology.

If you are working in a field where replication is important, and if your study replicates the one you are trying to replicate, you can be pretty confident about the result.

But what does it mean if, like Priya’s first attempts, your study does not replicate? One explanation could be that the original result was a ‘false positive’, and so the failed replication is a ‘true negative’. Another explanation is that the replication result was a ‘false negative’, and that the original study was a ‘true positive’. It’s also possible that differences between the two studies are responsible for the different results.

  

Activity 1:

Allow about 20 minutes

This activity relates to our examples of typical direct and conceptual replication studies. By way of reminder:

  • A researcher finds a surprising finding in their research. To test whether they should rely on this result, they conduct a replication immediately after, using all the same materials and the same participant pool. This is a direct replication.
  • A researcher wants to replicate a study they’ve read about. They don’t think the original study was well-designed, but they think the hypothesis is interesting, so they design a new study testing the same hypothesis but in a different way. This is a conceptual replication.

Now imagine these two researchers both carry out their studies. List the reasons why each of these two researchers may not replicate the original result.

To use this interactive functionality a free OU account is required. Sign in or register.
Interactive feature not available in single page view (see it in standard view).

When you are ready, press 'reveal' to see our comments.

Discussion

You might have listed:

  • The original result was a false positive
  • The replication result is a false negative
  • There are important differences between the original study and the replication study:
    • a.These could be small changes that researchers didn’t think should be important but that turned out to be (e.g.: which brand of a specific chemical was used).
    • b.It could be that the replication researcher didn’t realise these were differences because there wasn’t enough detail in the original paper to be able to work out how everything had been done.
    • c.The replication researcher might know they’re making a change from the original protocol, but approve this change because theoretically it shouldn’t make a difference to the result.

The replication crisis

This section will highlight some of the issues around replication in quantitative research. Replication is possible in qualitative research, and many qualitative researchers see the value of replication. So if you are a qualitative researcher, this section is still relevant to you. It will allow you to explore key issues faced by quantitative colleagues, learn how to read quantitative research papers more critically, and think about whether these issues could also apply to qualitative research, albeit manifested differently.

If we consider relatively direct replications, using the same materials as the original authors but conducted by different researchers, what percentage of published results do you imagine would replicate? It would be tempting to think that most published research findings are true, and therefore that a replication of a published research finding would be pretty likely to find the same result. However, researchers have found these percentages of findings could not be replicated:

  • Psychology: up to 60%
  • Cancer biology: up to 55%
  • Economics: up to 40%
  • Philosophy: up to 30%

  

The number of studies that could not be replicated was much higher than expected in certain fields, which has led some to refer to this as a ‘replication crisis'.

Why is it that so many quantitative studies cannot be replicated? It’s complicated!

Previously, you learned the three classifications of failed replications: the original result was a false positive, the replication result was a false negative, or differences between the two studies could have been responsible for the different results. However, these three interpretations are not all as likely as each other. There are ways to try and work out which of these are most likely.

The original result being a false positive is more likely than you would think. Researchers often do not publish all the research that they do. As a researcher, there is an incentive to publish papers in ‘high impact’ journals (journals that are regarded highly in the researcher’s discipline, and that publish papers that receive a high number of citations). Historically, it has been harder to publish negative (null) results than positive (statistically significant) results, as journals have prioritised headline-grabbing results that confirm popular or contemporary positions. This has been the case for all journals, but especially high-impact ones.

This means that researchers have an incentive to get positive results in their research and can feel disappointed, stressed, and even ashamed if they don’t get a significant result. This can entice them to turn to questionable research practices, to increase the likelihood of a false positive result.

Questionable research practices

Yes, you read the end of the previous section correctly. There are questionable research practices that researchers may feel pressurised to use. Here are some examples:

  • P-hacking: in quantitative research, p-hacking means exploiting techniques that increase the likelihood of obtaining a statistically significant result, for example by performing multiple analyses, or stopping data collection once a significant p-value is reached.
  • Selective reporting: when results from research are deliberately not fully or accurately reported, in order to suppress negative or undesirable findings. For example, researchers might run two analyses but only report the one with significant findings, or be selective in what results are included in a report aimed at particular audiences.
  • HARK-ing: is a shortening of ‘hypothesising after the results are known’. This is when researchers write their papers as if they had hypotheses that they then went on to test in their study, when really they made up the hypothesis after seeing their results, to pick one that best fit.
  • Post-hoc justifications: means stating, after the fact, justifications for decisions made during the research project. For example, if the researcher only managed to recruit women for a study after trying to recruit all genders, but claimed in the paper that this was intentional.

  

Although pressures to publish can sometimes be seen as barriers to transparency, the benefits of writing transparently can also be seen as a positive incentive, as the next section shows.

Writing transparently

When writing manuscripts, researchers should aim to be as transparent as possible, being honest about what happened in the study, how it was conducted, and when and why decisions were made. By using questionable research practices, researchers make it more likely that they get a false positive result, which can partially explain low replicability rates.

In the video, Priya introduced another important consideration for evaluating replication results: sample size (the number of samples in your study, e.g. participants). Smaller sample sizes make it more likely to get both a false positive and a false negative result. This is because smaller sample sizes provide less information about the population you are studying, which increases the variability and uncertainty in your results. With a small sample, the random variation (or 'noise') can more easily overshadow the true effect you are trying to measure. This means you might detect an effect that isn’t really there, a false positive, or miss an effect that actually exists, a false negative.

For instance, imagine trying to judge the average height of a population by looking at just a few individuals. Your estimate is more likely to be off compared to measuring a larger group, because you may happen to have either a very tall or very short person in your sample. So, if you have an original study with a small sample size and a (well-designed) replication with a large sample size, you could be more confident in the result of the replication than the result of the original study.

  

Activity 2:

What not to do!

Allow about 30 minutes

So far, you have considered good and bad writing practices. With these in mind, have a go at this ‘hack your way to scientific glory’ activity. First, choose a political party: Republican [UK equivalent: Conservative] or Democrat [UK equivalent: Labour]. Then predict whether the party has a positive or negative impact on the economy. When you have done that, change aspects of the research (e.g. participant inclusion criteria and how you’re measuring your dependent variable) and see whether you can find a significant result (p < 0.05) in your predicted direction.

The reason this is an example of ‘what not to do’ is because when you first choose a political party and predict whether they will have a positive or negative impact on the economy, you are forming a hypothesis. But, if you then play around with the data until you get the result that you wanted, and only stop when you do, then you are fixing the result.

The activity involves various questionable research practices, such as p-hacking, HARK-ing, and selective reporting. However, there is a way to do different analyses on the same data without any of these being a problem. If instead of deciding on a hypotheses first then confirming it, you were to conduct purely exploratory research (without a hypothesis) you could be transparent about all of the different ways you looked at the data and how the results differed when you tried different things. This could even lead to people conducting their own future studies to confirm your exploratory results!

When reading an academic paper, it’s important to read with a critical mindset and feel free to disagree with the methodological or analysis strategy, the interpretation of the results, or the conclusions drawn. Although we know that there are rare instances of outright fraud in science, we would expect that the researchers are truthfully describing what happened in the study, how it was conducted, and when and why decisions were made.

Generalisability

You have learned that replication studies vary on a spectrum from ‘direct’ to ‘conceptual’. However, most replication studies have some differences from the original study, even if these weren’t intentional. Consider one of the examples from before, where a researcher was replicating a paper from the 1990s. The materials they create will be different from the original materials, and if what they’re studying is context-dependent, a lot might have changed since then.

For example, a study on internet usage habits conducted in the 1990s would yield very different results if replicated today, due to the dramatic changes in technology and how people use the internet. Similarly, a study examining public attitudes toward mental health in the 1990s might produce different findings now because societal awareness and acceptance of mental health issues have evolved significantly over the past few decades.

For this reason, some consider that most replication studies are actually generalisability studies. Generalisability means whether a particular result generalises beyond the specific participants and conditions of the study to broader groups of samples, settings, methods, or measures. For example, if we’re interested in public attitudes to mental health, it wouldn’t make sense for us to only ask people aged 50-60, or only men, or only those living in cities. It’s possible that any of these characteristics could affect people’s opinions on mental health, meaning the results would be biased and not representative of the full population.

Without generalisability studies, it might be possible that the theoretical explanation for why the finding occurred might be incorrect. For example, there could even be a mistake in the design of the study that biased the results. For instance, imagine a biological study investigating the effects of a new drug using a specific strain of lab mice. If this particular strain has a unique genetic mutation that makes it respond differently to the drug compared to other strains, the study’s results might not generalize to other mice or to humans. This could lead to an incorrect conclusion about the drug’s overall effectiveness and safety.

Researchers wishing to be transparent when writing their papers should declare possible ‘Constraints on generality’ in the discussion section. This could take the form of a statement that identifies and justifies the target populations for the reported findings, and other considerations the authors think would be necessary for replicating their result. This could help other researchers to sample from the same populations when conducting a direct replication, or to test the boundaries of generalisability when conducting a conceptual replication.

Studying generalisability

So you think your research has potential do good in the world, but don’t know how widely it can be applied? There are lots of different ways to study generalisability:

  • Systematic reviews: these look at how an outcome varies in the published literature across samples, settings, measures and methods (meta-analyses do this statistically). This can be done without conducting any new studies.
  • Comparative studies: comparing results from different populations using the same (adapted) materials can show where there may be similarities and differences.
  • Big team science: when researchers from around the world conduct the same study and pool their results, they can look at various factors affecting the presence or size of the effect they’re interested in.
    • For example, the first ManyGoats project is examining goat responses to different human attentional states, and will be testing a diverse range of goats in different living conditions.

  

Activity 3:

Allow about 10 minutes

Think about when a study in your field would or wouldn’t generalise, and make a few notes as to why this might be the case.

To use this interactive functionality a free OU account is required. Sign in or register.
Interactive feature not available in single page view (see it in standard view).

When you are ready, press 'reveal' to see our comments.

Discussion

There are lots of reasons why a study may or may not generalise. Imagine a study evaluating a new therapy for depression in a university clinic with primarily urban-based participants. While the therapy showed significant improvement in depressive symptoms over ten weeks among a diverse sample, including college students and middle-aged adults of various ethnicities recruited through local health centres and university channels, its applicability to other populations and settings may be limited. Factors such as regional differences in mental health resources, demographic diversity beyond the studied age groups, and recruitment biases could affect the therapy's effectiveness in rural or suburban areas and among older adults or adolescents.

Quiz

The image shows an abstract pattern which reminds you of a brain or a maze.

Throughout the course, we offer you self-test quizzes to help you test your understanding of the course concepts. These quizzes are there to help you consolidate your knowledge. This one tackles key ideas in replication and the principle of generalisability. It is important to answer the questions carefully, then read the feedback, whether you got the answer right or not.

Answer the questions on the following pages:

  Question 1

Guest users do not have permission to interact with embedded questions.
Interactive feature not available in single page view (see it in standard view).

  Question 2

Guest users do not have permission to interact with embedded questions.
Interactive feature not available in single page view (see it in standard view).

  Question 3

Guest users do not have permission to interact with embedded questions.
Interactive feature not available in single page view (see it in standard view).

  Question 4

Guest users do not have permission to interact with embedded questions.
Interactive feature not available in single page view (see it in standard view).

  Question 5

Guest users do not have permission to interact with embedded questions.
Interactive feature not available in single page view (see it in standard view).

Summary

This week you learned about an important aspect of integrity: replicability. Replicability relates to whether or not a study ‘replicates’, i.e. whether or not, when you repeat the study with new data, you get the same result. You learned some reasons why replicability may be low in many fields, and how differences between studies may sometimes contribute to this. You also learned about the importance of generalisability in research. Next week, you’ll learn techniques which can support both the integrity and the transparency of your research.

References

Camerer, C, Dreber, A, Forsell, E, Ho, T, Huber, J, Johannesson, M, Kirchler, M, Almenberg, J, Altmejd, A, Chan, T, Heikensten, E, Holzmeister, F, Imai, T, Isaksson, S, Nave, G, Pfeiffer, T, Razen, M, Wu, H (2016): Evaluating replicability of laboratory experiments in Economics. Science, 3 Mar 2016, Vol 351, Issue 6280, pp. 1433-1436.
Available at: https://doi.org/ 10.1126/ science.aaf0918

Cova, F, Strickland, B, Abatista, A, (2021): Estimating the reproducibility of experimental philosophy. Review of Philosophy and Psychology 12(1):9-44
Available at: https://www.researchgate.net/ publication/ 325216701_Estimating_the_Reproducibility_of_Experimental_Philosophy

Ebersole CR, Mathur MB, Baranski E, et al (2020): Many labs 5: Testing pre-data-collection peer review as an intervention to increase replicability. Advances in Methods and Practices in Psychological Science. 2020;3(3):309-33
Available at: https://doi.org/ 10.1177/ 2515245920958687

Ebersole C. R., Atherton O. E., Belanger A. L., Skulborstad H. M., Allen J. M., Banks J. B., … Nosek B. A. (2016): Many Labs 3: Evaluating participant pool quality across the academic semester via replication. Journal of Experimental Social Psychology, 67, 68–82.
Available at: https://doi.org/ 10.1016/ j.jesp.2015.10.012

Errington, T, Mathur, M, Soderberg, C, Denis, A, Perfito, N, Iorns, E, Nosek, B, (2021): Investigating the replicability of preclinical cancer biology, eLife.
Available at: https://doi.org/ 10.7554/ eLife.71601

Fifethirtyeight.com: Hack your way to scientific glory (website)
Available at: https://projects.fivethirtyeight.com/ p-hacking/

FORRT (2024): Lesson plan 8: open data and qualitative research (lesson template with a CC-By Attribution 4.0 licence).
Available at: https://osf.io/ nyfqx

Hofstede, G. (1980): Culture's consequences: international differences in work-related values. Beverly Hills, CA: Sage Publications.
Available at: https://books.google.co.uk/ books/ about/ Culture_s_Consequences.html?id=Cayp_Um4O9gC&redir_esc=y

Klein, R. A., Ratliff, K. A., Vianello, M., Adams, R. B., Jr., Bahník, Š., Bernstein, M. J., Bocian, K., Brandt, M. J., Brooks, B., Brumbaugh, C. C., Cemalcilar, Z., Chandler, J., Cheong, W., Davis, W. E., Devos, T., Eisner, M., Frankowska, N., Furrow, D., Galliani, E. M., . . . Nosek, B. A. (2014). Investigating variation in replicability: A ‘many labs’ replication project.Social Psychology, 45(3), 142–152.
Available at: https://doi.org/ 10.1027/ 1864-9335/ a000178

Open Science Collaboration (2015): Estimating the reproducibility of psychological science Science, 349 aac4716.
Available at: https://doi.org/ 10.1126/ science.aac4716

Silverstein, P., Gliga, T., Westermann, G., Parise, E.: Probing communication induced biases in preverbal infants: two replication attempts of Yoon, Johnson and CsibraInfant Behaviour and Development, 55, 77-87.
Available at: https://www.sciencedirect.com/ science/ article/ pii/ S0163638318301474?via%3Dihub

Simons, D. J., Shoda, Y., & Lindsay, D. S. (2017): Constraints on Generality (COG): A proposed addition to all empirical papers. Perspectives on Psychological Science, 12(6), 1123-1128.
Available at: https://doi.org/ 10.1177/ 1745691617708630

UNICEF Innocenti (2022): Evidence and Gap Map Research Briefs: UNICEF Strategic Plan 2018–2021 Goal Areas

Available at: https://www.unicef.org/ innocenti/ reports/ evidence-and-gap-map-research-briefs

 

Click here to move on to the next week

Glossary

Big team science
A research project in which researchers from around the world conduct the same study and pool their results.
Conceptual replications
A type of replication study which aims to vary some aspect of the original study, in order to better understand the underlying phenomenon.
Constraints on generality
A statement identifying populations sampled in the study and potential limits to the samples and methods, enabling others to assess the extent to which results can be generalised.
Direct replications
A type of replication study which aims to stay as close to the original study as possible.
False positive
An error that occurs when a researcher believes that there is a genuine effect or difference when there is not (e.g. a person has a positive Covid test although they do not have Covid).
False negative
An error that occurs when a researcher believes that there is no effect or difference, when actually there is (e.g. a person has a negative Covid test although they do have Covid).
Generalisability
The extent to which the findings of a study can be generalised to other situations, beyond the specific participants and conditions of the study.
HARK-ing
Researchers are HARK-ing if they write papers as if they had a hypothesis they wanted to test in their study, whereas in reality, they made up the hypothesis after seeing the results.
P-hacking
In quantitative research, exploiting techniques that increase the likelihood of obtaining a statistically significant result.
Post-hoc justifications
Researchers write up justifications for their actions after a study – these justifications were not planned or decided before the study happened.
Reproducibility
A study is reproducible if you are able to get the same results when conducting the same analyses on the same data as the original study.
Replicability
A study is replicable if you are able to conduct the same study again, generate new data, and still get the same results as the original study.
Selective reporting
Researchers are selective reporting if their results are deliberately not fully or accurately reported, in order to suppress negative or undesirable findings.
Systematic review
A structured literature review, which analyses existing research evidence according to a fixed set of criteria, then synthesises what the research evidence shows.