Читать книгу Statistics and the Evaluation of Evidence for Forensic Scientists - Franco Taroni - Страница 21
1.3.3 Glass Fragments
ОглавлениеSection 1.3.2 discussed an example of the interpretation of the evidence of DNA profiling. Consider now an example concerning glass fragments and the measurement of the refractive index of these.
Example 1.2 As before, consider the investigation of a crime. A window has been broken during the commission of the crime. A PoI is found with fragments of glass on their clothing, similar in refractive index to the broken window. Several fragments are taken for investigation and their refractive index measurements taken.
Note that there is a difference here from Example 1.1, where it was assumed that the crime stain had come from the criminal and been transferred to the crime scene. In Example 1.2 glass is transferred from the crime scene to the criminal. Glass on the PoI need not have come from the scene of the crime; it may have come from elsewhere and by perfectly innocent means. This is an asymmetry associated with this kind of scenario. The evidence is known as transfer evidence, as discussed in Section 1.1, because evidence (e.g. blood or glass fragments) has been transferred from the criminal to the scene or vice versa. Transfer from the criminal to the scene has to be considered differently from evidence transferred from the scene to the criminal. A full discussion of this is given in Chapters 5 and 6 .
Comparison in Example 1.2 has to be made between the two sets of fragments on the basis of their refractive index measurements. The evidential value of the outcome of this comparison has to be assessed. Notice that it is assumed that none of the fragments has any distinctive features and comparison is based only on the refractive index measurements.
Methods for evaluating such evidence were discussed in many papers in the late 1970s and early 1980s Evett (1977, 1978), Evett and Lambert (1982, 1984, 1985), Grove (1981, 1984), Lindley (1977c), Seheult (1978), and Shafer (1982). These methods will be described as appropriate in Chapters 3 and 7. Knowledge‐based computer systems have been developed. See Curran and Hicks (2009) and Curran (2009) for a review of practices in the forensic evaluation of glass and DNA evidence. As an aside, sophisticated systems have been developed to deal with DNA, notably DNA mixtures complexities (i.e. number of donors, peaks heights). Examples are presented and evaluated in Bright et al. (2016), Alladio et al. (2018), and Bleka et al. (2019).
Evett (1977) gave an example of the sort of problem that may be considered and developed a procedure for evaluating the evidence that mimicked the interpretative thinking of the forensic scientist of the time. The case is an imaginary one. Five fragments from a suspect are to be compared with 10 fragments from a window broken at the scene of a crime. The values of the refractive index measurements are given in Table 1.2. The procedure developed by Evett is a two‐stage one. It is described here briefly. It is a rather arbitrary and hybrid procedure. While it follows the thinking of the forensic scientist, there are interpretative problems, which are described here, in attempting to provide due value to the evidence. An alternative approach that overcomes these problems is described in Chapter 7 .
Table 1.2 Refractive index measurements.
Measurements from the window | 1.518 44 | 1.518 48 | 1.518 44 | 1.518 50 | 1.518 40 |
1.518 48 | 1.518 46 | 1.518 46 | 1.518 44 | 1.518 48 | |
Measurements from the PoI | 1.518 48 | 1.518 50 | 1.518 48 | 1.518 44 | 1.518 46 |
The first stage is known as the comparison stage. The two sets of measurements are compared. The comparison takes the form of the calculation of a statistic, , say. This statistic provides a measure of the difference, known as a standardised difference, between the two sets of measurements that takes account of the natural variation there is in the refractive index measurements of glass fragments from within the same window. If the absolute value of is less than (or equal to) some pre‐specified value, known as a threshold value, then the two sets of fragments are deemed to be similar and the second stage is implemented. If the absolute value of is greater than the threshold value, then the two sets of fragments are deemed to be dissimilar. The two sets of fragments are then deemed to have come from different sources and the second stage is not implemented. (Note the use here of the word statistic, which in this context can be thought of simply as a function of the observations.) A classic example of such an approach is the use of the Student ‐test or the modified Welch test for the comparison of means (Welch 1937; Walsh et al. 1996; Curran et al. 2000).
The second stage is known as the significance stage. This stage attempts to determine the significance of the finding from the first stage that the two sets of fragments were similar. The significance is determined by calculating the probability of the result that the two sets of fragments were found to be similar, under the assumption that the two sets had come from different sources. If this probability is very low then this assumption is deemed to be false. The fragments are then assumed to come from the same source, an assumption that places the PoI at the crime scene. This assumption says nothing about how the fragments came to be associated with the PoI. This may have occurred in an innocent manner. See Section 5.3.2 for further discussion of this point in the context of activity level propositions.
The procedure can be criticised on two points. First, in the comparison stage the threshold provides a qualitative step that may provide very different outcomes for two different pairs of observations. One pair of sets of fragments may provide a value of , which is just below the threshold, whereas the other pair may provide a value of just above the threshold. The first pair will proceed to the significance stage, the second stage will not. Yet, the two pairs may have measurements, which are close together. The difference in the consequences is greater than the difference in the measurements merits (such an approach is called a fall‐off‐the‐cliff effect; see Evett (1991) who attributed this term to Ken Smalldon. Criticisms have been developed in Robertson and Vignaux (1995b). They wrote:
This sudden change in decision due to crossing a particular line is likened to falling off a cliff, one moment you are safe, the next dead. In fact, rather than a cliff we have merely a steep slope. Other things being equal, the more similar the samples the stronger the evidence that they had a common origin, and the less similar the samples the stronger the evidence that they came from different sources. (p. 118)
A related problem is that of cut‐off where a decision is taken dependent on whether a statistic is above or below a certain value (see Section 7.7.5).
A better approach, as suggested in the quotation from Robertson and Vignaux (1995b) above and that is described in Section 7.3, provides a measure of the value of the evidence that decreases as the distance between the two sets of measurements increases, subject, as explained later, to the rarity or otherwise of the measurements.
The second criticism is that the result is difficult to interpret. Because of the effect of the comparison stage, the result is not just simply the probability of the evidence, assuming the two sets of fragments came from different sources. A reasonable interpretation, as will be explained in Section 2.4, of the value of the evidence is the effect that it has on the odds in favour of the true guilt of the PoI. In the two‐stage approach this effect is difficult to measure. The first stage discards certain sets of measurements, which may have come from the same source and does not discard other sets of measurements which may have come from different sources. The second stage calculates a probability, not of the evidence but of that part of the evidence for which was not greater than the threshold value, assuming the two sets came from different sources. It is necessary, as is seen later, to compare this probability with the probability of the same result, assuming the two sets came from the same source. There is also an implication in the determination of the probability in the significance stage that a small probability for the evidence, assuming the two sets came from different sources, means that there is a large probability that the two sets came from the same source. This implication is unfounded; see Section 2.5.1.
A review of the two‐stage approach and the development of a Bayesian approach is provided by Curran et al. (2000) and Curran and Hicks (2009).
As with DNA profiling, there are problems associated with the definition of a suitable population from which probability distributions for refractive measurements may be obtained; see, for example, Walsh and Buckleton (1986).
These examples have been introduced to provide a framework within which the evaluation of evidence may be considered. In order to evaluate evidence, something about which there is much uncertainty, it is necessary to establish a suitable terminology and to have some method for the assessment of uncertainty. First, some terminology will be introduced followed by a method for the measurement of uncertainty. This method is probability. The role of uncertainty, as represented by probability, in the assessment of the value of scientific evidence will form the basis of the rest of this chapter. A commentary on so‐called knowledge management, of which this is one part, has been given by Evett (1993b, 2015).