Draft - 10/11/2001

 

 

      CHAPTER FOUR

 

    A RIDE ON THE BLUE BUS

 

 

7.1       Betty Smith and the Blue Bus Problem

On January 6, 1941, Betty Smith was driving in her automobile from Dorchester to Winthrop, Massachusetts.  As she entered Winthrop at about one o’clock in the morning, she was crowded off the road by a bus, and as a consequence collided with a parked car.  Smith was injured in the accident, and sued Rapid Transit, Inc. in the Superior Court of the Commonwealth of Massachusetts.

 


Because the accident occurred in the middle of the night, because the bus that forced her off the road did not stop, and because she was preoccupied with trying to avoid the accident, Betty Smith did not see any of the identifying marks on the bus.  All she could testify to at the trial was that the vehicle that forced her off the road was a bus, and that it was “a great, big, long wide affair.”  Smith was, however, able to establish at the trial that the Rapid Transit Company had been licensed by the City of Winthrop to operate buses on the very route on which the accident had occurred, that Rapid Transit’s buses left Winthrop Highlands on the thirty-minute trip to Maverick Square, Boston, at 12:10 A.M., 12:45 A.M., 1:15 A.M., and 2:15 A.M., that this route included the Main Street location of the accident, and that no other company was licensed to operate its buses on this route. 

 

Despite having offered this evidence to prove the proposition that the bus that caused the accident was operated by Rapid Transit, the trial court refused to let the jury even consider the case.  The judge ruled that Betty Smith could not, as a matter of law, recover against Rapid Transit, because there was no direct evidence that the bus that hit Smith was one of Rapid Transit’s buses.  This ruling was upheld by the Massachusetts Supreme Judicial Court,[1] which noted that Rapid Transit’s exclusive franchise “did not preclude private or chartered buses from using this street; the bus in question could very well have been operated by someone other than the defendant.”  The court acknowledged that this was unlikely and that “perhaps the mathematical chances somewhat favor the proposition that a bus of the defendant caused the accident.”  But “this was not enough,” the court said, concluding that the mathematical probability that the bus in question was the defendant’s bus was not the type of “direct” evidence that could lead a jury to have an “actual belief” in the proposition that this was one of Rapid Transit’s buses.

 


Had this been a criminal case, the ruling would strike us as unexceptionable.  After all, the evidence that the bus belonged to Rapid Transit was hardly of the quality and quantity that would establish “beyond a reasonable doubt” that this was one of Rapid Transit’s buses, especially given the small but hardly inconceivable possibility that the accident was caused by a private bus or by a chartered bus.  But Smith’s lawsuit was a civil case and not a criminal one, and thus the required standard of proof was not that of “beyond a reasonable doubt,” but was such that Betty Smith needed to establish her case only “by a preponderance of the evidence.”   And as the equivalent phrase in English law, “by a balance of the probabilities,” indicates, we ordinarily understand the preponderance of the evidence to be the equivalent of just over a 50 percent likelihood that the proposition asserted is true.[2]  Whatever the possibility might have been of a private bus or a chartered bus having caused the accident, no one claimed that the probability of such an occurrence was anything approaching 50 percent,[3] and thus there seemed to be no reasonably denying that the evidence presented by Smith established to a probability considerably greater than .5 that this was Rapid Transit’s bus.  But if that was the case, then why did not Betty Smith win?



Why not indeed?   Smith’s case is hardly unique,[4] and the Supreme Judicial Court’s ruling is generally in line with the law as it was then, and as it is now.[5]  Yet it still seems odd that if the plaintiff is only required to prove her case to a probability of .51 (to put it roughly), then statistical evidence that would do so is thought by itself to be insufficient, or so the courts routinely conclude.  Indeed, it seems so odd to so many people that Smith’s case has become a staple of academic teaching of evidence law in law schools, and the centerpiece of much of academic writing about what has come to be called the problem of “naked statistical evidence.”[6]  Commonly, the problem is made analytically crisper when presented as a hypothetical version of the Smith case that has come to be known as the Blue Bus Problem: Suppose that it is late at night, under one version of the Problem, and an individual’s car is hit by a bus.  This individual cannot identify the bus, but she can establish that it is a blue bus, and she can prove as well that 80 percent of the blue buses in the city are operated by the Blue Bus Company, that 20 percent are operated by the Red Bus Company, and that there are no buses in the vicinity except those operated by one or the other of these two companies.  Moreover, each of the other elements of the case -- negligence, causation, and, especially, the fact and the extent of the injury -- is either stipulated or established to a virtual certainty.  In these circumstances can the plaintiff recover in civil litigation against the Blue Bus Company, or, if not (as the overwhelming majority of American courts would conclude), then why not?  Or, in a variation of the Blue Bus Problem even closer to Betty Smith’s case, the plaintiff’s car is hit by a bus late at night and all she knows about the offending vehicle is that it was a bus.  80 percent of the buses in town are operated by the Blue Bus Company.  Can the plaintiff win a lawsuit against the Blue Bus Company on that evidence alone, assuming, as in the previous examples, that there is nothing in dispute about the issues of causation, negligence, or injury?

 

7.2       The Generality of Statistics and the Statistics of Generality


Scholars have been debating the Blue Bus Problem for decades,[7] sometimes in the highly technical language of mathematical statistics, and sometimes in a more commonsense way.  Some scholars have defended the legal system’s skepticism about statistical evidence, often, like Laurence Tribe and later by Charles Nesson, pointing to the way in which explicit acknowledgment of the probability of error might, even if accurate, undermine confidence in the legal system.  Other scholars urge increased acceptance of statistical evidence, arguing, among other things, that a legal rule should not be premised on keeping jurors and the public in the dark about the actual nature and consequences of legal decisions.   At this stage of the debate, however, it is not my intention here to enter the fray, but rather to draw attention to the way in which the debate about naked statistical evidence links more closely than the literature recognizes with the seemingly different questions about the role of generality in decision-making.

 

Recall our discussion of pit bull regulation in Chapter Three.  The issue about the use of the generalization “pit bull,” which gathered up all of the individual pit bulls, each with its individual characteristics, under the single category of pit bulls, was that one attribute of the category -- a tendency towards dangerous aggressiveness -- was not necessary an attribute of each member of the category.  The generalization about the dangerousness of pit bulls is not spurious -- the evidence does appear to establish that dangerousness exists in the class of pit bulls to a greater degree than it does in the class of all dogs, and to a greater degree than it does in almost all of the sub-classes that we call breeds  -- but there is still no disputing that many, probably even most, pit bulls are not dangerous at all.

 


Similarly with the Blue Bus Company.  If the relevant attribute is ownership of a particular bus, as in the latter version of the Blue Bus Problem, and if the Blue Bus Company owns 80 percent of the buses, then the Blue Bus Company possesses the attribute of ownership of this particular bus to a higher probability than does any another possible defendant, just as pit bulls possess the attribute of dangerousness to a higher probability than dogs simpliciter or than most other breeds.  Moreover, the Blue Bus Company possesses the attribute of ownership of the bus in question to a probability seemingly sufficient to justify liability in a civil lawsuit.  If we were to hypothesize that 80 percent of the vicious dogs were pit bulls, then we could conclude, absent further information, that an attack by an otherwise unidentified vicious dog was 80 percent likely to have been an attack by a pit bull.  Similarly, if 80 percent of the buses are owned by the Blue Bus Company, then we could conclude, absent further information, that an accident caused by an otherwise unidentified bus is 80 percent likely to have been an accident caused by a Blue Bus Company bus. 

 


Casting the problem in this way brings to mind another famous hypothetical case, this one offered by the British philosopher L. Jonathan Cohen.  In what he labels The Paradox of the Gatecrasher,[8] Cohen hypothesizes a rodeo that charges for admission.  During the event the organizers of the rodeo count the spectators, and they discover that there are 1000 spectators in attendance.  But at the conclusion of the rodeo, the organizers count the tickets collected at the ticket booth, and it turns out that there are only 499 tickets in the ticket box.  The mathematical corollary of this, of course, is that 501 of the 1000 spectators at the rodeo were gatecrashers.  So now suppose that the organizers of the rodeo bring a lawsuit against one -- any one -- of the 1000 spectators for fraudulent entrance.  No one saw this particular person enter fraudulently, and there is no other evidence connecting this particular individual to a fraudulent entry.  Yet still, absent any other evidence, there is a .501 probability that this person (or any of the other 999 spectators) was a gatecrasher. Why, then, cannot the statistical evidence by itself be sufficient to warrant a verdict, at least under the preponderance standard in civil litigation, in favor of the rodeo organizers?  Cohen maintains that such a verdict would be profoundly unjust, and for him the paradox consists not in the unwillingness of the courts to award judgment to the rodeo organizers, but in the fact that courts would not in fact award damages in such a case, and would be correct in refusing to do so, but would still hold to the belief that the standard of proof in civil cases is a preponderance of the evidence, a standard that the .501 likelihood of gatecrashing by any one of the spectators appears to satisfy.

 


To repeat, my aim is not to “solve” either the Paradox of the Gatecrasher or the Blue Bus Problem.  It is, however, to identify the way in which these problems are best seen as variants on the larger problem of generality in decision-making.  One way of framing the issue would be to understand the problem of generality as the problem of attempting to determine when in making decisions we should use and when we should not use statistically reliable but non-universal indicators.  In this sense the problem of generality is “really” the problem of statistical inference, and thus the Blue Bus and Gatecrasher problems, which appear to be problems of statistical inference, resemble the problem of generality, because all are problems relating to the wisdom or justice of using non-universal but non-spurious statistical indicators.

 

Alternatively, and preferably, both the Blue Bus Problem and the Paradox of the Gatecrasher, which are typically presented as problems of statistical inference, are fundamentally problems about the use of generalizations.  In each the issue is, at bottom, not a problem of statistics but instead a problem of the extent to which we can use, at least for purposes of awarding damages in civil litigation, generalizations about rodeo spectators (most but not all entered fraudulently) and a generalization about the Blue Bus Company (owns most of the buses in this city).  Framed in this way, therefore, it is the problem of generality and generalization that is primary, and the problem of statistical inference is but another way of describing what is at its core an issue about generalization.   

 


This way of framing the issue becomes clearer when we see that what the Massachusetts Supreme Judicial Court in Smith saw as the problem was not a problem of statistics at all.  Rather, the court, although it did employ the potentially confusing language of “mathematical chances” and “probability,” was primarily focused on what it saw as the difference between so-called “direct” or “actual” evidence, on the one hand, and the kind of evidence that is based on the characteristics of the class of which the putative perpetrator is a member, on the other.[9]  This is even clearer in the Paradox of the Gatecrasher.  Again, the fact that “statistics,” in the numerical sense of that word, might have been part of the hypothetical rodeo organizers’ case is largely beside the point.  When the organizers bring a case against a particular individual, they are basing that case on the attribution of non-spurious class characteristics -- non-payment of the admission charge – to an individual member of the class.  In doing so, the organizers relied on the same process of generalization that Plato’s training master relied on in attributing the characteristics of “the herd” to each of its members, that pit bull ordinances rely on in attributing the characteristics of the class of pit bulls to each individual pit bull, that insurance companies rely on in attributing the characteristics of the class of teen-age male drivers to each teen-age male driver, and that many of us rely on in attributing the honesty of the class of dealers in used automobiles to each dealer in used automobiles.  In all of these cases, the process, in the final analysis, is the process of basing decisions for all members of a class on non-spurious but non-universal characteristics of the class taken as a whole.  This is the process of generalization, and this is the process of which the problem of so-called statistical evidence is but one component.

 


The argument in the immediately preceding paragraph notwithstanding, it may not be overly important whether it is statistical inference that is primary and generality and generalization that is secondary, or generality and generalization that are primary and statistical inference that is secondary.  What is important, however, is that we can appreciate that the seemingly diverse, at least in a number of literatures and in common understanding as well, issues of generality and statistical evidence are in fact remarkably similar, and it may well be that the resources that would enable us to understand and negotiate the problem of generality may be the same resources that could be used to understand and negotiate problems about the use of statistical evidence in civil and criminal trials.  And once we understand this, there remains more to be said about these problems of statistical evidence and the light they shed on the issue of generality.

 

7.3            Probabilistic Inference in an All or Nothing World


The Blue Bus and Gatecrasher problems are in an important way artefacts of the all-or-nothing way in which most aspects of most modern legal systems operate.  In much of non-legal life people can operationalize their uncertainty by acting in accordance with the principle of expected value, taking the value of an uncertain outcome as the product of the value of some set of consequences multiplied by the probability that those consequences will come to pass.   Just as any wager of less than ten dollars is a good one if you are betting on a ten percent chance of winning one hundred dollars, so too do we act in similar ways in much of our daily life.  We invest less in risky investments than in more certain ones, we make shorter commitments when we are unsure of the value of what we are committing to than when we have greater confidence, we calculate how much insurance to buy based on expected value (just as the insurance company does in determining how much to charge us), we plan travel times based on factoring in the probability of delays, we calculate expected fines in deciding whether it is worth it to engage in minor illegalities such as overtime parking, and the expression to “hedge one’s bets” is applicable in much more of our lives than the occasional trip to the racetrack.  In these cases, and many more, an imprecise but serviceable conception of expected value guides many of our daily decisions.

 

Not so, however, in almost all of the law.  To the statistician the Paradox of the Gatecrasher may be no paradox at all.   If there is .51 probability that any given spectator entered fraudulently, and if the purchase price of a ticket is $1.00, then the statistician sees the easy solution -- the rodeo organizers recover 51 cents against each of the 1000 spectators. In this way the rodeo organizers do not recover more than their fair share of the proceeds, and each spectator is liable only to the extent of the likelihood that he or she entered without purchasing a ticket.  And so too with the Blue Bus Problem.   If there is a .80 chance that the bus that plainly negligently caused an indisputable $1000 worth of damages to, say, Betty Smith, is a bus owned and operated by the Blue Bus Company, then the principle of expected value would indicate that Smith should recover $800 against the Blue Bus Company. 

 


The law, however, does not operate in this way.  Perhaps oddly to the statistician, the law would give Smith all of her damages if she proved her case to a .51 probability, and nothing if she proved it to a .49 probability.  And it would give her not a dollar more of she proved her case to a .90 probability than if she proved it to a .51 probability.  With rare exceptions,[10] the expected value of a plaintiff’s claim, by which the extent of the plaintiff’s proof would be multiplied by the extent of the plaintiff’s damages, is not a principle of advanced legal systems.[11] These systems, we see throughout the world, are all or nothing affairs.

 

In the context of a criminal case, our intuitions confirm the approach of the law.  If there is a .70 chance that the defendant is the one who committed an aggravated assault, and if the penalty for aggravated assault is 10 years imprisonment, few of us, and not even the statisticians, would be comfortable in this case imposing a sentence of seven years based on the principle of expected value.  And that is perhaps because of the strength of the Blackstonian maxim that “it is better that ten guilty persons escape, than that one innocent suffer.”[12]  The value we place on liberty, and thus the graveness of the error of denying liberty to the innocent, makes us uncomfortable with imprisoning those who are .30 likely to have done nothing wrong, and thus the principle of expected value is properly a stranger to the criminal law.


In civil cases, however, the aversion to expected value verdicts seems less justifiable.  After all, the plaintiff in a typical negligence case is claiming to have been injured through someone else’s fault while doing nothing wrong.  In such a case, it is not clear why erroneously denying recovery to a worthy plaintiff is any less harmful an error than erroneously awarding recovery against a non-negligent defendant.  To put it differently, we assume that erroneous denials of liability and erroneous impositions of liability are equally regrettable.  The false positive is no worse than the false negative.  And if this is so, if the Type I and Type II errors, to use the statistician’s language, are equivalent, then it is by no means clear that the aversion to expected value verdicts in criminal cases ought to be extended to civil cases.

 

The law, however, does not agree, and continues to be pervasively and perhaps perversely an all or nothing decision process.  As a result, the statistician’s easy solution to the Paradox of the Gatecrasher and the Blue Bus Problems is dismissed, at least by the legal system, as too easy by half, and these hypothetical problems, as well as the real cases they distill, continue to be problematic.  Consequently, it is plausible to suppose that the difficulties presented by the Blue Bus Problem, the Paradox of the Gatecrasher, and other real and imagined examples are largely the products of the all-or-nothing character of most of legal decision-making.    

 



Once we see the relationship between the paradoxes of the law of evidence and the all-or-nothing nature of legal decision, however, however, we can understand the larger problem of generality in a new light. For if the problem of statistical inference in the law of evidence is, as we have seen, but an instantiation of the problem of generality, then the problems created by an all-or-nothing legal system parallel the problems created by the all-or-nothing parts of many other dimensions of our decisional lives.  In numerous instances in which we employ probabilistically sound but non-universal generalizations in ordinary decision-making, it is because the nature of the decision makes an expected value decision impossible or, at the very least, impractical.  If I am looking for a pet, it is not possible for me to have a pit bull for one day out of seven and a golden retriever for the other six, even if this were the right result were I to conclude that all other things were equal, but that a pit bull was six times as likely to behave aggressively as a golden retriever.  Similarly, tax officials rarely conduct partial audits (even though some are more thorough than others), customs officials rarely conduct partial inspections, police officers cannot conduct partial stops, airlines do not believe that they can deal with the problem of pilots 10 percent more likely to cause an accident by having them fly 10 percent fewer flights, and hockey referees who are 75 percent sure that a player has committed a high-sticking infraction do not have the option of sending the offender to the penalty box for 90 seconds rather than the designated two minutes for that offense.   In these and many more examples, what looks at first to be the special all-or-nothing quality of legal system may be replicated in non-legal decision-making.  In more cases than we or the statisticians might suppose,  non-legal decision-makers often understand themselves to be making all-or-nothing decisions (do I hire this person as a babysitter or not) in which the expected value approach is just not available.  The use of generalizations, therefore, appears to be a product not only of the frequent need to use generalizations as a time- and effort-saving heuristic in circumstances in which individual determinations would likely be too costly or too prone to the errors of discretion, but also of the fact that expected value decision-making is considerably more of a stranger to everyday decisional life than we may at first have fully appreciated.

 

7.4            Individuality and Reliability

If the nature of most of legal and more than we thought of non-legal decision-making requires us to engage in all-or-nothing decision-making, and if the nature of all-or-nothing decision-making pushes us towards what seem to many people to be unjust outcomes, then one way of understanding the instinct behind the Smith rule is as a desire to minimize the number of erroneous outcomes inevitably generated by all-or-nothing decision procedures.   Perhaps the insistence on so-called “direct” or “actual” evidence, as the court in Smith put it, is explained by a reluctance to have the legal system forced into accepting the 20 percent error rate that giving Betty Smith 100 percent of her damages on an 80 percent chance of Rapid Transit’s liability would entail.

 


Yet if this kind of error minimization is the goal, then it is hard to see how a supposed requirement of “direct” or “actual” evidence serves it.  Initially, we can ask what the Massachusetts Supreme Judicial Court in Smith might have meant by the terms “direct” and “actual.”  Presumably the court had in mind evidence that comes from a perception of a witness, with that very witness then testifying to that perception in court.[13]  Typically this would be a visual perception -- an eyewitness -- although there can also be perceptions by any of the other senses -- hearing, smelling, tasting, and touching.  But apart from sensory perception testified to under oath by the perceiver in court, it is difficult to see what the court could have meant by the terms “direct” and “actual.” 

 


If “direct” and “actual” refer to perceptual evidence testified to by the perceiver, then we must consider the reliability of this evidence as compared to the allegedly indirect or “non-actual” evidence offered in Smith and similar cases.   Consider, therefore, a variant on a hypothetical example given by Daniel Shaviro.[14]  Suppose Betty Smith had testified that she saw what looked like the words “Rapid Transit” written in red letters on the side of the blue bus that hit her.  But then suppose that on cross-examination she was forced to acknowledge that it was a foggy and rainy night, that the eyeglasses she always wears were knocked from her head by the impact of the accident, that she had first reported her observation of the words “Rapid Transit” not to the police officer who came upon the scene of the accident but only later after having consulted with an attorney, and that she saw the words only as the bus was heading away from her, at an angle to her direct vision, at a speed of no less than thirty miles per hour, and at a distance of no less than 300 feet.  Yet despite all of these reasons to doubt the accuracy of the hypothetical Smith’s observation of the words “Rapid Transit,” and despite the fact that it might be reasonable to place the probable accuracy of her observation of the words “Rapid Transit” at well less than .80, it remains the case that the very court that refused to let the real case go the jury, even on a probability well above .80 that the bus in question was a Rapid Transit bus, would almost certainly have let the fuzzy observation case go to the jury on a probability well below .80 that the bus in question was a Rapid Transit bus, reasoning that these issues are for the jury and for the jury alone to determine.

 

Part of this anomaly is explained by a widespread but empirically unsupported faith in eyewitness identification.   Although there persists an aura of credibility historically attached to eyewitness accounts, a raft of serious psychological research has established that much of this historical faith in eyewitness testimony lacks much of an empirical foundation.  People often see what they want to see or see what they think they are expected to see or see what they are positively reinforced in seeing; their perceptions are filtered through their own biases, prejudices, and preconceptions; they simply forget or misremember what they saw; and they are afflicted with a host of other cognitive deficiencies make eyewitness testimony less reliable than the conventional wisdom would suppose.[15]   If the preference for direct or actual evidence is based on a preference for perception over inference, then almost all of what we know about the deficiencies of human perception cast doubt on such a preference.

 


These doubts about perceptual abilities are exacerbated by the tendency of people not only to overweight perception as an empirical matter, but also to ignore what the statisticians call “base rates” and thus make logical as well as empirical errors.  Consider an example made famous by Amos Tversky and Daniel Kahneman, and one that bears a close resemblance to the Blue Bus Problem.[16]  Suppose that the Green Cab Company owns and operates taxis that are green in color, and the Blue Cab Company owns and operates blue taxis.  85% of the taxis in town are the green taxis of the Green Cab Company, and the other 15% are the blue taxis of the Blue Cab Company.  As in the Smith case, suppose that a car is sideswiped or run off the road by a taxi, and a witness, is 100% certain, presumably from the light on top of the cab, that the “guilty” car is a taxi, and is confident, but not certain, that the guilty taxi was blue.  Suppose that the witness is 80% confident that the taxi was blue and thus that it was a taxi of the Blue Cab Company.

 

On these facts, most people would conclude, with the witness, that it was the Blue Cab Company that should be held liable, but in fact this gets it exactly wrong.  The conclusion that the taxi was probably blue because the witness said so to a moderately high degree of confidence ignores the base rate distribution of taxis, and thus ignores the fact that the witness’s .2 likelihood of error must be applied to this distribution and not to a factually inaccurate presumed even distribution.  Thus, the number of cases, on these probabilities, in which a witness said the cab was blue when it was green is somewhat higher than the number of cases in which a witness said the cab was green when it was blue.  On these probabilities, in fact, the probability of the cab having been green is .59 despite the fact that the witness was .80 certain that it was blue.[17]       


The prevalence of ignoring the base rate, combined with the prevalence of overestimating the reliability of eyewitness testimony (which may be a contributing factor in people’s willingness to ignore the base rate in cases like these), makes the legal system’s prevailing skepticism about statistical evidence even more puzzling.  As the above examples of fuzzy or otherwise uncertain observations are designed to illustrate, it could very well be the case that the kind of evidence commonly thought to be direct or non-statistical is less reliable than the kind of evidence often thought to be indirect or statistical.  Or, to translate this into the language of generality, it may often be the case that the inferences to be drawn from non-spurious but non-universal generalizations are empirically stronger than the inferences to be drawn from decision-making approaches that seemingly do not rely on generalizations. 

 


The potential empirical superiority of decision-making by generalization may not be all there is to the matter.  As we will explore in subsequent chapters, people may think that there is a moral imperative in maximal individuation in decision-making even if the actual practices of such individuation are less reliable than the alternative.[18]  But at the very least the preference for individuation, of which Betty Smith’s case is but one example, cannot plausibly be seen as resting on some overall greater accuracy of non-generalized decision-making.

The possibility that relying on generalizations known ex ante to be imperfect might still be empirically superior to relying on allegedly direct or individualized assessments replicates the debates about the virtues and vices of rules and rule-based decision-making.  As prescriptive generalizations, rules necessarily entail the possibility that their strict application will produce sub-optimal outcomes in some cases, where sub-optimality is measured by reference to the outcome that would have been produced by accurate application of the background justification lying behind the rule.[19]  To take a hoary example from the world of legal philosophy,[20] if in order to prevent noise in the park (the background justification) we prohibit all vehicles from entering the park (the rule), we then produce a sub-optimal result whenever we exclude non-noisy vehicles (bicycles and electric cars) and whenever we fail to exclude noisy non-vehicles (musical instruments and loud radios). 

 


The inevitable sub-optimality of rules, however, is premised on a supposition about the accuracy of individualized decision-making, an accuracy that often does not exist, especially when there are reasons of bias and mistake, among other things, to distrust the reliability of the individualized decision.  If there were grounds to believe that enforcement officers would make numerous mistakes in trying to determine which instrumentalities were noisy and which not, then in practice the sub-optimal rule could very well produce fewer errors than the theoretically optimal individualized assessment.

 


The debate about statistical evidence, therefore, like the debates about rules, highlights the empirical dimensions of those debates, and highlights as well the often erroneous empirical underpinnings of the aversion to generalization.  If that aversion is based on an unwillingness to accept the mistakes that decision-making by generalization necessarily entails, then the aversion must presuppose that the actual human beings who make more individualized decisions would in practice make fewer mistakes than those made in relying on the generalization.  As the comparison of the record of unreliability of eyewitness testimony with the greater reliability of at least some statistical generalizations shows, however, this presupposition is often simply false.  If there is something that is problematic across the board about relying on generalizations, it cannot be that there is good reason to believe that such reliance is necessarily or even typically likely to produce more errors than the alternative.[21]   

 

7.5       The Non-Individualized Nature of Individualized Evidence                     


The objection to preferring so-called “direct” or “actual” evidence to other sorts of evidence, however, is not only an empirical one.  Rather, the objection rests as well on understanding that the avoidance of generalizations is, with few or no qualifications, simply not possible at all.  Put differently, even those decisions that appear initially to be maximally individual, that appear to be “direct” or “actual,” in the words of the Massachusetts Supreme Judicial Court in the Smith case, may turn out to rely more on generalizations than many people suppose.

 


Because most readers of this book are not visually impaired, it may be easier to understand the issue by use of an example involving direct but non-visual perception.  Suppose, therefore, that there were a totally blind passenger in our hypothetical Betty Smith’s car.  And suppose as well that the Blue Bus Company owns all of the buses in the city, and indeed all of the buses in the county and surrounding counties.  Because the possibility of buses owned by others is so minuscule, the defendant Blue Bus Company is willing to concede that if Betty Smith’s car was crowded off the road by a bus then it was crowded off the road by one of the Blue Bus Company’s buses   But that Betty Smith was crowded off the road by a bus rather than a car, truck, or piece of construction equipment, however, is something that the Blue Bus Company is not willing to concede.  Taking the position that Betty Smith’s alleged visual observation of a bus was a fabrication (the Blue Bus Company being wealthy and well-insured), the Blue Bus Company attempts at trial to cast doubt on the part of her story maintaining that it was a bus that crowded her off the road.   In order to counter this strategy, Betty Smith’s lawyer calls to the witness stand Smith’s blind passenger, Walter Wilson.  Wilson then testifies that he heard the sound of the vehicle approaching the car, that the vehicle approached Smith’s car to a distance of no more than two feet, and that the vehicle was definitely a bus.  On cross-examination by the Blue Bus Company’s lawyer, Wilson testifies to his previous experience with perceiving the sounds of vehicles and inferring their size, nature, and distance from the sounds.  Betty Smith’s lawyer, in further support of Walter Wilson’s testimony, then introduces two expert witnesses who bolster Wilson’s account by reporting that laboratory experiments bear out the ability of blind people to determine the proximity and nature of vehicles on the basis of hearing alone, which is just what Wilson claimed to have done.

 

There is, of course, nothing more or less “direct” or “actual” or “real” about Wilson’s primary aural sensory perceptions than about Smith’s primary visual ones.  Yet in considering what to make of Wilson’s perceptions, we would naturally think that the validity of these perceptions depends on a process of generalization and non-certain inference.  Wilson has perceived certain sounds in the past and they have turned out to be buses.  He has perceived distances in the past and they have turned out to be accurate.  And so on.  As a result, Wilson’s inference from this sound to this conclusion (it is a bus at this distance) is an inference based on most but not necessarily all sounds of this type having turned out in the past to be buses.  This is a non-spurious but non-universal generalization -- most but not all of sounds like this are buses – that undergirds what appears to be a direct and thus individualized perception.

 


Although less obvious to those of us who are sighted, the process of making visual observations from what philosophers refer to as “sense-data” is conceptually no different in the case of visual observations than it is in the case of aural ones.  And as the studies of the unreliability of eyewitness identification indicate, there may not be much of an empirical difference as well, no matter how hard it is for those of us who are sighted to confront the possibility that, more often than we think, we should simply not believe our eyes.  As a result, acknowledging the way in which seemingly direct observation involves a process of inference and generalization enables us to appreciate that even the processes that initially appear to us to be “direct,” “actual,” or individualized turn out to rely far more on generalizations from past experience than is often appreciated.  Once we see that all evidence is in the final analysis probabilistic, the distinction between the probabilistic and the “direct,” “actual,” or “real” emerges as even more of an anomaly.

 

Not only are individualized assessment still premised on probabilities and generalizations, but such individualized assessments are also always only partially individualized, omitting numerous dimensions of the particular case that might under other circumstances or other rules be relevant.  Let us go back to the real Smith case, and assume that what the Supreme Judicial Court was looking for was testimony by Betty Smith that she actually saw the words “Rapid Transit” on the side of the bus that crowded her off the road.  But even if this evidence had been forthcoming, Smith would not have been permitted, under well-accepted principles of tort law and evidence law, from testifying to how much she needed the money from a recovery against Rapid Transit, to how easily Rapid Transit or its insurer could have afforded to pay the judgment, to how exemplary a life she had lived in the past, to how many times Rapid Transit had been found liable for the negligence of one of its bus drivers, or to the positive effect that even a mistaken judgment for Smith would improve bus safety in the Town of Winthrop.  Yet in a truly particularist account of the events – act-consequentialism, for example, applied to the legal system – none of these genuinely “real” facts would be deemed irrelevant, and all of them would be components of a full individualized consideration of the case.


So what are we to make of the fact that Betty Smith would not have been allowed to testify to some number of facts that a truly individualized determination might have allowed into consideration?  Most significantly, the acceptance of the exclusion of these facts demonstrates from a different angle that most of our so-called individualized determinations are not as individualized as we suppose.  Moreover, the exclusion of these facts is itself something that occurs by virtue of the operation of a rule, and that consequently operates by virtue of a generalization.  We exclude evidence of the plaintiff’s poverty, the defendant’s wealth, the existence or terms of insurance coverage, and the defendant’s past negligent acts, among others, because it has been determined at some earlier time that these facts would as a rule not promote justice, or not further efficiency, or whatever.  But because these are rules, we exclude the evidence even in the face of a showing in the particular case that admission of this evidence might serve justice, or might foster efficiency, or might promote some other goal that can be seen as one of the background justifications lying behind the exclusionary[22] rules.

 


Now it could be the case that one or the other parties would argue that the exclusionary rules should be overridden in a particular case, and the exclusion wrought by an exclusionary rule is best thought of in presumptive rather than absolute terms.[23]  Nevertheless, the fact that every piece of unadmitted evidence is typically unadmitted, whether consciously or not, by virtue of a rule that is itself based on a generalization about the usual or probable, but not universal, irrelevance of the excluded fact reinforces the claim here that the idea of decision-making in a totally individualized or particularistic way is essentially impossible.

 


That all seemingly particular or individualized decisions turn out to have important dimensions of generality is not totally to deny the logical distinction between the particular and the general.   Although pressing against this distinction has a distinguished philosophical provenance,[24] there is no need for us here to examine the deepest questions of metaphysics and philosophical logic bearing on the nature and existence of the distinction between the particular and the general, or the relationship between particulars and universals.[25]  For our purposes, the commonsense distinction between a thing and a group of things will suffice, and my only point here, although an important one, is that many of the things we perceive as particular objects or particular observations turn out to depend on the kinds of generalizations that, even if not on the same metaphysical status as true universals, are much the stuff of ordinary reasoning.  And even if this is true, it is still not to deny that there are important differences in degree between the more and the less particular and the more and the less general.  Nevertheless, however, once we understand these differences, at least as they typically confront us in ordinary decision-making, as differences of degree more than differences in kind, we become properly skeptical of a widespread but still mistaken view that the particular has some sort of natural epistemological or moral primacy over the general.

 

It turns out, therefore, that the Supreme Judicial Court’s unwillingness to allow a jury to consider Betty Smith’s case against the Rapid Transit company is a product of two significant mistakes – an over-confidence in the empirical reliability and even the very directness of direct evidence, and an under-appreciation of the essential continuity between so-called indirect or statistical evidence and evidence that on its face appears to be more individualized and thus less statistical.  The Supreme Judicial Court’s skepticism about a “mathematical” case, therefore, even if it was correct that this was a mathematical case, is, as we have seen, not so much a skepticism about mathematical or statistical evidence but a skepticism about resting legal decisions on non-spurious but non-universal generalizations.

 


Seen in this way, the Supreme Judicial Court’s skepticism is of a piece with the skepticism of Plato and Aristotle about relying to heavily on what they called “laws” and with the inflammatory slogans of the pit bull sympathizers.  In all of these cases, the preference for particulars is seen as a moral imperative.  But if particularism is itself reliant on generalizations, and if particularized decisions provide no guarantee of greater reliability, then the foundations for the preference for particularism are shakier than they often appear.  The consequences of this shakiness are apparent in the hypothetical Blue Bus Problem, but the consequences of the weak foundations of an a priori preference for particularism will emerge even more clearly when in ensuing Chapters we take on a range of morally salient controversies regarding discrimination and equality.     

 

  

 

   

 

          



[1]Smith v. Rapid Transit, 317 Mass. 469, 58 N.E.2d 754 (1945).

[2]See McCormick’s Handbook of the Law of Evidence (St. Paul, Minnesota: West Publishing Company, 2d ed., 1972), §339, at p. 794; Ronald J. Allen, “Burdens of Proof, Uncertainty, and Ambiguity in Modern Legal Discourse,” Harvard Journal of Law and Public Policy, vol. 17 (1994), pp. 627-40; James Brook, “Inevitable Errors: The Preponderance of the Evidence Standard in Civil Litigation,” Tulsa Law Journal, vol. 18 (1982), pp. 79-104; Bruce Hay and Kathryn Spier, “Burdens of Proof in Civil Litigation: An Economic Perspective,” Journal of Legal Studies, vol. 26 (1997), pp. 413-42.

[3]James Brook argues that the Smith case actually stands for less than this, maintaining that because Mrs. Smith presented no evidence whatsoever as to the likelihood (or lack thereof) of buses other than Rapid Transit’s buses this was essentially a case in which the plaintiff offered no proof at all on an essential element of the case.  James Brook, “The Use of Statistical Evidence of Identification in Civil Litigation: Well-Worn Hypotheticals, Real Cases, and Controversy,”St. Louis University Law Journal, vol. 29 (1985), pp. 293-352, at pp. 301-03.  But the entire tenor of the court’s opinion makes clear that statistical evidence on this score would have availed Mrs. Smith not at all, and the case seems more plausibly supportive, as generations of scholars of the law of evidence have taken it to be, of the proposition that statistical evidence of identification is insufficient as a matter of law. 

[4]See Guenther v. Armstrong Rubber Company, 406 F.2d 1315 (3d Cir. 1969); Sawyer v. United States, 148 F. Supp. 877 (M.D. Georgia 1956); Curtis v. United States, 117 F. Supp. 912 (N.D. New York, 1953); Sargent v. Massachusetts Accident Company, 29 N.E.2d 825 (Massachusetts 1940); Lampe v. Franklin American Trust Company, 96 S.W.2d 710 (Missouri, 1936); Day v. Boston & Maine Railroad, 96 Me. 207, 52 A. 771 (1902).

[5]One exception is Kaminsky v. Hertz Corporation, 288 N.W.2d 426 (Michigan Court of Appeals 1979), in which the court allowed a jury to conclude that a vehicle with Hertz markings was owned by the Hertz Corporation on the basis of proof that, although some so-marked vehicles were owned by licensees, approximately ninety percent of the so-marked vehicles were directly owned by Hertz.

[6]David Kaye, “The Limits of the Preponderance of the Evidence Standard: Justifiably Naked Statistical Evidence and Multiple Causation,” American Bar Foundation Research Journal, vol. 1982, pp. 487-522.

[7]See, for example, Ronald J. Allen, “On the Significance of Batting Averages and Strikeout Totals: A Clarification of the ‘Naked Statistical Evidence’ Debate, the Meaning of “Evidence,’ and the Requirement of Proof Beyond a Reasonable Doubt,” Tulane Law Review, vol. 65 (1991), pp. 1093-1110; James Brook, op. cit.note 3; Craig Callen. “Adjudication and the Appearance of Statistical Evidence,” Tulane Law Review, vol. 65 (1991), pp. 457-83; Charles Nesson, “The Evidence or the Event? On Judicial Proof and the Acceptability of Verdicts,” Harvard Law Review, vol. 98 (1985), pp. 1357-92, at pp. 1378-85; Michael J. Saks and Robert F. Kidd, “Human Information Processing and Adjudication: Trial by Heuristics,” Law and Society Review, vol. 15 (1980), pp. 140-73; Daniel Shaviro, “Statistical-Probability Evidence and the Appearance of Justice,” Harvard Law Review, vol. 103 (1989), pp. 530-554; Judith Jarvis Thomson, “Liability and Individualized Evidence,” in Rights, Restitution, and Risk: Essays in Moral Theory (Cambridge, Massachusetts: Harvard University Press, 1986), pp. 225-50; Laurence H. Tribe, “Trial by Mathematics: Precision and Ritual in the Legal Process,” Harvard Law Review, vol. 84 (1971), pp. 1329-93, at pp.1340-41 (Tribe’s article is the actual source for the Blue Bus hypothetical).

[8]L. Jonathan Cohen, The Probable and the Provable (Oxford: Clarendon Press, 1977), pp. 74-81.  For a sampling of the debates spawned by Cohen’s hypothetical example, see Brook, op. cit. Note 3; Richard Eggleston, “The Probability Debate,” Criminal Law Review, vol. 1980, pp. 678-91; David Kaye, “The Paradox of the Gatecrasher and Other Stories,” Arizona State Law Journal, vol. 1979, pp. 101-43; David Kaye, “Paradoxes, Gedanken Experiments and the Burden of Proof: A Response to Dr. Cohen’s Reply,” Arizona State Law Review, vol. 1981, pp. 635-48; Glanville Ll. Williams, “The Mathematics of Proof - I,” Criminal Law Review, vol. 1979, pp. 297-312; Glanville Ll. Williams, “The Mathematics of Proof - II,” Criminal Law Review, vol. 1979, pp. 340-54; Glanville Ll. Williams, “A Short Rejoinder,” Criminal Law Review, vol. 1980, pp. 103-08.

[9]One of the earliest court cases to address the problem distinguished between probabilistic and “real” evidence:

Quantitative probability, however, is only the greater chance.  It is not proof, nor even

probative evidence, of the proposition to be proved.  That in one throw of dice there is a quantitative probability, or greater chance, that a less number of spots than six will fall is no evidence, whatever, that in a given throw such was the actual result.  Without something more, the actual result of the throw would still be utterly unknown.  The slightest real evidence that sixes did in fact fall uppermost would outweigh all the probability otherwise.

Day v. Boston & Maine Railroad, 96 Maine at 207, 52 A. at 774, as quoted in Thomson, op. cit. note 7, at 234 note 13.

[10]Among the exceptions is the occasional case that distributes liability in “mass tort” according to the market share of various possible defendants.  Most noteworthy is the widely-discussed Sindell v. Abbott Laboratories, 26 Cal. 3d 588, 163 Cal. Rptr. 132, 607 P.2d 924 (1980), distributing liability among eleven manufacturers of the pharmaceutical diethylstillbesterol (DES), a drug designed to prevent miscarriages but which, although relatively ineffective at lessening the risk of miscarriage, did increase the risk of cancer for the daiughters of women who took the drug.  The case is helpfully discussed in, inter alia, Glen O. Robinson, “Multiple Causation in Tort Law: Reflections on the DES Cases,” Virginia Law Review, vol. 68 (1982), pp. 000-000; Judith Jarvis Thomson, “Remarks on Causation and Liability,” in Rights, Restitution, and Risk, op. cit. note 7, pp. 192-224.   Yet although many people find some intuitive appeal in allocating 80 percent of the damages to Defendant A, with 80 percent of the market, and 20 percent of the damages to Defendant B, with 20 percent of the market, at least in a case brought by 1000 identically situated plaintiffs, the intuitions seem to change, perhaps irrationally and perhaps not, when a single plaintiff in a single case against Defendant A alone is allowed to collect 80 percent of her damages against Defendant A.

[11]There are some who advocate moving the legal system more in line with expected value principles.  See Michael Abramowicz, “A Compromise Approach to Compromise Verdicts,” California Law Review, vol. 89 (2001), pp. 231-313; John E. Coons, “Approaches to Court Imposed Compromise - The Uses of Doubt and Reason,” Northwestern University Law Review, vol. 58 (1964), pp. 750-93; John E. Coons, “Compromise as Precise Justice,” California Law Review, vol. 68 (1980), pp. 250-73; David Rosenberg, “The Causal Connection in Mass Exposure Cases: A ‘Public Law’ Vision of the Tort System,” Harvard Law Review, vol. 97 (1984), pp. 849-929; Steven Shavell, “Uncertainty over Causation and the Determination of Civil Liability,” Journal of Law and Economics, vol. 28 (1985), pp. 587-612.  There are others, however, who continue to demur.  See David Kaye, op. cit. note 8; Charles Nesson, op. cit. note 10.   

[12]William Blackstone, Commentaries on the Laws of England, vol. 4, p. 358 (London: 1769).

[13]Indeed, that would explain why we have come to refer to those who testify in court, many of whom (like expert witnesses) have not witnessed anything, as “witnesses.”

[14]Shaviro, op. cit. note 8.

[15]For an excellent survey of the literature, see Daniel Schachter, Searching for Memory (New York: Basic Books, 1996).  See also Elizabeth Loftus, Eyewitness Identification (1979); Gary L. Wells and Elizabeth Loftus, eds., Eyewitness Testimony: Psychological Perspectives (1984).  Good examples of the studies include Elizabeth Loftus, Julie Feldman, and Richard Dashiell, “The Reality of Illusory Memories,” in Daniel Schachter et al., eds., Memory Distortion: How Minds, Brains, and Societies Reconstruct the Past (Cambridge, Massachusetts: Harvard University Press, 1995); Elizabeth Loftus, David Miller, and Helen Burns, “Semantic Integration of Verbal Information into a Visual Memory,” Journal of Experimental Psychology: Human Learning and Memory, vol. 4 (1978), pp. 19-31; Gary L. Wells and Amy L. Bradford, “‘Good, You Identified the Suspect’: Feedback to Eyewitnesses Distorts Their Reports of the Witnessing Experience,” Journal of Applied Psychology, vol. 83 (1998), pp. 360-72. 

[16]Amos Tversky and Daniel Kahneman, “Judgment Under Uncertainty: Heuristics and Biases,” Science, vol. 185 (1974), pp. 1124-31.

[17]These numbers can be worked out by the application of Bayes’ Rule, such that the probability of the guilty cab being a blue cab given a witness testifying that it was a blue cab is .41, so that the probability of the cab being a green cab given that the witness testified it was a blue cab is .59.  For the calculations, see Ian Hacking, An Introduction to Probability and Inductive Logic Cambridge: Cambridge University Press, 2001), pp. 72-73. 

[18]Judith Thomson, op. cit. note 7, appears to come close to this view in resting her support for requiring individualized evidence on the way in which a person who claims to know something is offering a guarantee of its truth, even when the guarantee is as likely to be unfounded as a warrant coming solely from aggregate probabilities.  Yet insofar as her argument is grounded on there being an important difference between “I know x” and “I believe to a high probability that x,” then her argument may only be a highly sophisticated version of the view that probabilistic version is for mysterious reasons inferior to other sorts of evidence.

[19]For my full exploration of these issues, see Frederick Schauer, Playing By the Rules: A Philosophical Analysis of Rule-Based Decision-Making in Law and in Life (Oxford: Clarendon Press, 1991); Frederick Schauer, “On the Supposed Defeasibility of Legal Rules,” in M.D.A. Freeman, ed., Current Legal Problems 1998 (volume 51) (Oxford: Oxford University Press, 1998), pp. 223-40; Frederick Schauer, “Prescriptions in Three Dimensions,” Iowa Law Review, vol. 82 (1997), pp. 911-22; Frederick Schauer, “The Practice and Problems of Plain Meaning,” Vanderbilt Law Review, vol. 45 (1992), pp. 715-41. 

[20]The “no vehicles in the park” example originates in H.L.A. Hart, “Positivism and the Separation of Law and Morals,” Harvard Law Review, vol. 71 (1958), pp. 593-629, at p. 607.

[21]Much of normative philosophy is properly focused on ideal conditions and presuppositions of strict compliance, see most notably John Rawls, A Theory of Justice (Cambridge, Massachusetts: Harvard University Press, 1971), pp. 8-9, 142-45, 245-46, and thus it might be objected here that my focus on mistaken individuators, however real, is theoretically uninteresting.  But if we are in the realm of ideal theory, we would also be able to stipulate that our generalizations would be universal, and not as imprecise as real generalizations usually are.  The very fact that we are considering generalizations whose very imprecision is in the realm of non-ideal theory authorizes, and indeed demands, that we consider the alternative to generalizations in the realm of non-ideal theory as well. 

[22]I use the word “exclusionary” here to connect with Joseph Raz’s important insight that rules operate through the use of exclusionary reasons that exclude other reasons from being part of a decision-making process.  Joseph Raz, The Authority of Law: Essays on Law and Morality (Oxford: Clarendon Press, 1979); Joseph Raz, Practical Reason and Norms (Princeton: Princeton University Press, [reprinted] 1990).  For an important commentary, see Larry Alexander, “Law and Exclusionary Reasons,” Philosophical Topics, vol. 18 (1990), pp. 5-22.

[23]See Frederick Schauer, Playing By the Rules, op. cit., pp. 88-91.

[24]For example, Frank Plumpton Ramsey, The Foundations of Mathematics and Other Essays (London: Kegan Paul, 1931).

[25]Among the milestones in the modern literature are Willard Van Orman Quine, Word and Object (Cambridge, Massachusetts: MIT Press, 1960); W.V. Quine, “On the Individuation of Attributes” and “Predicates, Terms and Classes,” in Theories and Things (Cambridge, Massachusetts: Harvard University Press, 1981), pp. 100-12; 164-72; Bertrand Russell, “The World of Universals,” in The Problems of Philosophy (Oxford: Oxford University Press, 1959), pp. 91-100; P.F. Strawson, “Particular and General,” in Logico-Linguistic Papers (London: Methuen, 1971), pp. 28-52.