Skip to main content

About this free course

Download this course

Share this free course

Expert evidence and forensic science in the courtroom
Expert evidence and forensic science in the courtroom

Start this free course now. Just create an account and sign in. Enrol and complete the course for a free statement of participation or digital badge if available.

5.2 Partial solutions

If you struggled a little with the figures in the previous section, you are not alone! Yet scientific tests can get even more complicated than that. For example, we assumed that the blood test referred to in the previous section was completely reliable. But in real life, scientific tests commonly suffer from two types of problem:

  • false positives (also known as a ‘Type 1’ error): this is where the test reports something of interest when there is actually nothing.
  • false negatives (also known as a ‘Type 2’ error): this is where the test reports nothing of interest when there is actually something.

In addition, the accuracy of a test depends on how common the thing you are interested in (for example, blood type, DNA genotype) is in the population – that is, the base rate.

Described image
Figure 15 Cadaver dog

An illustration of this is ‘cadaver dogs’ – dogs that use their sense of smell to indicate whether there has been a dead body at a location. This evidence might be used if there is no longer a body at a suspected crime scene, but the police believe there might have been. For example, in 2013, in the USA, a man was convicted of murdering his wife based in part on the evidence of a cadaver dog, even where there was no body (Ward, 2013).

Investigators are interested in responses from cadaver dogs when:

  • a dog correctly indicates that there has been a body and there actually has been one; or
  • a dog correctly indicates that there has been no body when there has not been one.

Problematic responses are where:

  • a dog wrongly indicates that there has been a body but there has not been one (a false positive); and
  • a dog wrongly indicates that there has not been a body but there has been one (a false negative).

Table 1 shows findings from a study on cadaver dog response rates by Jackson et al. (2015, p. 78). As you can see, cadaver dogs are generally quite accurate, but not all the time.

Table 1 Cadaver dog response rate
Dog’s response Scent actually present Scent actually absent
Woof! A body! 224 11
No body 4 115

Here, the number 11 represents a false positive, and 4 represents a false negative.

In this next activity, you will be given a chance to try to work out how often cadaver dogs accurately identify a body. A word of reassurance: very few people get this right!

Activity 6 Cadaver detector dogs

Timing: Allow about 30 minutes

Quite often, these sorts of figures will be presented by experts as percentages. This was what you encountered in the previous activity.

Here, the probability of a body if the dog signals is 0.95 (224/224 + 11) or 95%, and the probability of a body if the dog does not signal is 0.03 (4/ 4 + 115) or 3%. The calculation becomes even more complicated because the base rate in Jackson, Aitken and Roberts’, research was derived in a laboratory where they chose to provide scents from cadavers around two-thirds of the time (or roughly 66%). That seems quite high. In the real world, crime scenes with traces of bodies are likely to be rarer than that. Let us assume that in the real world the base rate is about 1 in 10 of suspected murder scenes (or 10%).

So, if a dog in the real world indicates a body, what is the probability there really has been a body?

This is really difficult to calculate, even for experts, so just do your best. Select your answer from the options below. Once you have made your selection, remember to click ‘Save response’. The results of the poll will be displayed in the discussion below.

Active content not displayed. This content requires JavaScript to be enabled.
Interactive feature not available in single page view (see it in standard view).

Comment

The results of the poll are displayed below. You may need to refresh your browser to load the results.

Active content not displayed. This content requires JavaScript to be enabled.
Interactive feature not available in single page view (see it in standard view).

We know that even experts make mistakes with such calculations most of the time (see Gigerenzer, 1996, and Hoffrage and Gigerenzer, 1998). People seem to struggle with probabilities. However, Gigerenzer and others have suggested that presenting the numbers as ‘natural frequencies’ makes it a bit easier. That effectively means starting with an obvious number of cases.

This time, imagine 1,000 criminal cases where there might have been a body. Of these, based on the base rate assumption of 1 in 10, there will have been a body in 100 cases and no body in 900 cases. Where there has been a body, the dogs will detect it in 95 of those cases (as we worked out they would spot it in 95 per cent of cases and miss it in 5 per cent). If there is not a cadaver, the dogs will nonetheless wrongly indicate there has been a cadaver in 27 of those 900 cases (as we know they will wrongly indicate in 3 per cent or 3 out of every 100 cases). Figure 16 presents this same information visually.

Described image
Figure 16 One thousand criminal cases

Now try answering the question again:

If a dog indicates a body, what is the probability that there really has been a body? Hopefully this is a little easier. Again, remember to click ‘Save response’ after you make your selection.

Active content not displayed. This content requires JavaScript to be enabled.
Interactive feature not available in single page view (see it in standard view).

Comment

The results of the poll are displayed below. You may need to refresh your browser to load the results.

Active content not displayed. This content requires JavaScript to be enabled.
Interactive feature not available in single page view (see it in standard view).

The correct answer is 78%. You can work it out because out of 1,000 cases, a dog will indicate 122 times (95 + 27). Of these, 95 will be correct but 27 will be false positives. So the right answer is 0.78 (95/122) or 78%. Another way of seeing this is by looking at the ratio of the coloured dots in Figure 16. The pink and green dots represent the times that the dog indicates. Pink is a false positive and green is a correct identification. The ratio of the green dots out of all the coloured dots is 78%.

You may still have made a mistake the second time, but the research shows that far fewer mistakes are made when the statistics are presented as natural frequencies. The argument is that experts should present test results in natural frequencies to prevent the types of avoidable errors that happen with expert evidence.