Frequency format hypothesis
This article has multiple issues. Please help improve it or discuss these issues on the talk page. (Learn how and when to remove these template messages)(Learn how and when to remove this template message)
The frequency format hypothesis is the idea that the brain understands and processes information better when presented in frequency formats rather than a numerical or probability format. Thus according to the hypothesis, presenting information as 1 in 5 people rather than 20% leads to better comprehension. The idea was proposed by German scientist Gerd Gigerenzer, after compilation and comparison of data collected between 1976–1997.
- 1 Origin
- 2 Supporting arguments
- 3 Refuting arguments
- 4 References
Certain information about one's experience is often stored in the memory using an implicit encoding process. Where did you sit last time in class? Do you say the word hello or charisma more? People are very good at answering such questions without actively thinking about it or not knowing how they got that information in the first place. This was the observation that lead to Hasher and Zacks' 1979 study on frequency.
Through their research work, Hasher and Zacks found out that information about frequency is stored without the intention of the person. Also, training and feedback does not increase ability to encode frequency. Frequency information was also found to be continually registered in the memory, regardless of age, ability or motivation. The ability to encode frequency also does not decrease with old age, depression or multiple task requirements. They called this characteristic of the frequency encoding as automatic encoding.
Another important evidence for the hypothesis came through the study of infants. In one study, 40 newborn infants were tested for their ability to discriminate between 2 dots versus 3 dots and 4 dots versus 6 dots. Even though infants were able to make the discrimination between 2 versus 3 dots, they were not able to distinguish between 4 versus 6 dots. The tested new born infants were only 21 hours to 144 hours old.
Similarly in another study, to test whether infants could recognize numerical correspondences, Starkey et al. designed a series of experiments in which 6 to 8 month old infants were shown pairs of either a display of two objects or a display of three objects. While the displays were still visible, infants heard either two or three drumbeats. Measurement of looking time revealed that the infants looked significantly longer toward the display that matched the number of sounds.
The contigency rule
Later on, Barbara A. Spellmen from University of Texas describes the performance of humans in determining cause and effects as the contingency rule ΔP, defined as
ΔP = P(E|C) - P(E|~C)
where P(E|C) is the probability of the effect given the presence of the proposed cause and P(E|~C) is the probability of the effect given the absence of the proposed cause. Suppose we wish to evaluate the performance of a fertilizer. If the plants bloomed 15 out of 20 times when the fertilizer was used, and only 5 out of 20 plants bloomed in the absence of the fertilizer. In this case
P(E|C) = 15/20 = 0.75 P(E|~C)= 5/20 = 0.25 ΔP = P(E|C) - P(E|~C) ΔP = 0.75 - 0.25 = 0.50
The ΔP value as a result is always bound between -1 and 1. Even though the contingency rule is a good model of what humans do in predicting one event causation of another, when it comes to predicting outcomes of events with multiple causes, there exists a large deviation from the contingency rule called the cue-interaction-effect.
In 1993 Baker Mercer and his team used video games to demonstrate this effect. Each test subject is given the task of helping a tank travel across a mine field using a button that sometimes worked correctly in camouflaging and sometimes did not. As a second cause a spotter plane, a friend or an enemy would sometimes fly over the tank. After 40 trials, the test subjects were asked to evaluate the effectiveness of the camouflage and the plane in helping the tank through the minefield. They were asked to give it a number between -100 to 100.
Mathematically, there are two contigency values possible for the plane: the plane was either irrelevant to tank's success, then ΔP=0(.5/0 condition) and the plane was relevant to the plane's success, ΔP=1 (.5/1 condition). Even though the ΔP for the camouflage in either condition is 0.5, the test subjects evaluated the ΔP of camouflage to be much higher in the .5/0 condition than in the .5/1 condition. The results are shown in table below.
|Condition||ΔPplane||ΔPcamouflage||Camouflage rating given|
In each case, the test subjects are very good in noticing when two events occur together. When the plane is relevant to the camouflage success, they mark the camouflage success high and when the plane doesn't affect the camouflage's success, they mark the camouflage's success value low.
Several experiments have been performed that shows that ordinary and sometimes skilled people make basic probabilistic fallacies, especially in the case of Bayesian inference quizzes. Gigerenzer claims that the observed errors are consistent with the way we acquired mathematical abilities during the course of human evolution. Gigerenzer argues that the problem with these quizzes is the way the information is presented. During these quizzes the information is presented in percentages. Gigerenzer argues that presenting information in frequency format would help in solving these puzzles accurately. He argues that evolutionary the brain physiologically evolved to understand frequency information better than probability information. Thus if the Bayesian quizzes were asked in frequency format, then test subjects would be better at it. Gigerenzer calls this idea the frequency format hypothesis in his published paper titled "The psychology of good judgment: frequency formats and simple algorithms".
Gigerenzer argued that from an evolutionary point of view, a frequency method was easier and more communicable compared to conveying information in probability format. He argues that probability and percentages are rather recent forms of representation as opposed to frequency. The first known existence of a representative form of percentages is in the seventeenth century. He also argues that more information is given in the case of frequency representation. For instance, instead of conveying data as 50 out of 100, using the frequency form, as opposed to saying 50%, using the probability format, gives the users more information about the sample size. This can in turn make the data and results more reliable and more appealing.
An explanation given as to why people choose encounter frequency is that in the case of frequencies, the subjects are given vivid descriptions, while with probabilities only a dry number is given to the subject. Therefore, in the case of frequency, subjects are given more recall cues. This could in turn mean that the frequency encounters are remembered by the brain more often than in the case of probability numbers. Thus this might be a reason why people in general intuitively choose frequency encountered choices rather than probability based choices.
Yet another explanation offered by the authors is the fact that in the case of frequency, people often come across them multiple times and have a sequential input, compared to a probability value, which is given in one time. From John Medina’s Brain Rules, sequential input can lead to a stronger memory than a onetime input. This can be a primary reason why humans choose frequency encounters over probability.
Another rationale provided in justifying the frequency format hypothesis is that using frequencies makes it easier to keep track and update a database of events. For example, if an event happened 3 out of 6 times, the probability format would store this as 50%, whereas in frequency format it is stored as 3 out of 6. Now imagine that the event does not happen this time. The frequency format can be updated to 3 out of 7. However, for the probability format updating is extremely harder.
Frequency representation can also be helpful in keeping track of classes and statistical information. Picture a scenario where every 500 out of 1000 people die due to lung cancer. However, 40 of those 1000 were smokers and 20 out of the 40 had a genetic condition predisposed to possible lung cancer. Such class division and information storage can only be done using frequency format, since a number .05% probability of having lung cancer does not give any information or allow to calculate such information.
Ease of comparison
Critics of the frequency format hypothesis argue that probability formats allow for much easier comparison than frequency format representation of data. In some cases, using frequency formats actually does allow for easy comparison. If team A wins 19 of its 29 games, and another team B wins 10 of its 29 games, one can clearly see that team A is much better than team B. However comparison in frequency format is not always this clear and easy.If team A won 19 out of its 29 games, comparing this team with team B that won 6 out of its 11 games becomes much harder in frequency format. But, in the probability format, one could say since 65.6%(19/29) is greater than 54.5%, one could much easily compare the two.
Tooby and Cosmides had argued that frequency representation helps update data easier each time one gets new data. However this involves updating both numbers. Referring back to the example of teams, if team A won its 31st game, note that both the number of games won(20->21) and the number of games played(30->31) has to be updated. In the case of probability the only number to be updated is the single percentage number. Also, this number could be updated over the course of 10 games instead of updating each game, which cannot be done in the case of frequency format.
- Hasher, L.; Zacks, R. (1984). "Automatic processing of fundamental information: the case of frequency of occurrence". The American Psychologist. 39 (12): 1372–1388. doi:10.1037/0003-066x.39.12.1372. PMID 6395744.
- Hasher, Lynn; Zacks, Rose T. (1979). "Automatic and effortful processes in memory". Journal of Experimental Psychology: General. 108 (3): 356–388. doi:10.1037/0096-3418.104.22.1686.
- Hasher, L.; Chromiak, W. (1977). "The processing of frequency information: An automatic mechanism?". Journal of Verbal Learning and Verbal Behavior. 16: 173–184. doi:10.1016/s0022-5371(77)80045-5.
- Antell, S. E.; Keating, D. P. (1983). "Perception of numerical invariance in neonates". Child Development. 54 (3): 695–701. doi:10.2307/1130057.
- Starkey, P.; Spelke, E.; Gelman, R. (1990). "Numerical abstraction by human infants". Cognition. 36 (2): 97–127. doi:10.1016/0010-0277(90)90001-z.
- Spellman, B. A. (1996). "Acting as intuitive scientists: Contingency judgments are made while controlling for alternative potential causes". Psychological Science. 7 (6): 337–342. doi:10.1111/j.1467-9280.1996.tb00385.x.
- Baker, A.G.; Mercier, Pierre; Vallée-Tourangeau, Frédéric; Frank, Robert; Pan, Maria (1993). "Selective Associations and Causality Judgments: Presence of a Strong Causal Factor May Reduce Judgments of a Weaker One". Journal of Experimental Psychology: Learning, Memory, and Cognition. 19 (2): 414–432. doi:10.1037/0278-7322.214.171.1244.
- A.G. Baker, Robin A. Murphy, Associative and Normative Models of Causal Induction: Reacting to Versus Understanding Cause, In: David R. Shanks, Douglas L. Medin and Keith J. Holyoak, Editor(s), Psychology of Learning and Motivation, Academic Press, 1996, Volume 34, Pages 1-45, ISSN 0079-7421, ISBN 978-0-12-543334-1, doi:10.1016/S0079-7421(08)60557-5
- Sloman, S. A.; Over, D.; Slovak, L.; Stibel, J. M. (2003). "Frequency illusions and other fallacies". Organizational Behavior and Human Decision Processes. 91 (2): 296–309. doi:10.1016/s0749-5978(03)00021-9.
- Birnbaum, M. H.; Mellers, B. A. (1983). "Bayesian inference: Combining base rates with opinions of sources who vary in credibility". Journal of Personality and Social Psychology. 45 (4): 792–804. doi:10.1037/0022-35126.96.36.1992.
- Murphy, G. L.; Ross, B. H. (2010). "Uncertainty in category-based induction: When do people integrate across categories?". Journal of Experimental Psychology: Learning, Memory, and Cognition. 36 (2): 263–276. doi:10.1037/a0018685.
- Sirota, M.; Juanchich, M. (2011). "ROLE OF NUMERACY AND COGNITIVE REFLECTION IN BAYESIAN REASONING WITH NATURAL FREQUENCIES". Studia Psychologica. 53 (2): 151–161.
- Gigerenzer, G (1996). "The psychology of good judgment. frequency formats and simple algorithms". Medical Decision Making. 16 (3): 273–280. doi:10.1177/0272989X9601600312.
- Gigerenzer, G. (2002). Calculated risks, how to know when numbers deceive you. (p. 310). New York: Simon & Schuster.
- Daston, L.; Gigerenzer, G. (1989). "The Problem of Irrationality". Science. 244 (4908): 1094–1094. doi:10.1126/science.244.4908.1094.
- Reyna, V. F.; Brainerd, C. J. (2008). "Numeracy, ratio bias, and denominator neglect in judgments of risk and probability". Learning and Individual Differences. 18 (1): 89–107. doi:10.1016/j.lindif.2007.03.011.
- Hacking, I. (1986). The emergence of probability, a philosophical study of early ideas about probability, induction and statistical inference. London: Cambridge Univ Pr.
- Obrecht, N. A.; Chapman, G. B.; Gelman, R. (2009). "An encounter frequency account of how experience affects likelihood estimation". Memory & Cognition. 37 (5): 632–643. doi:10.3758/mc.37.5.632.
- Medina, J. (2010). Brain rules, 12 principles for surviving and thriving at work, home, and school. Seattle, WA: Pear Pr.
- Cosmides, L.; Tooby, J. (1996). "Are humans good intuitive statisticians after all? Rethinking some conclusions from the literature on judgment under uncertainty". Cognition. 58: 1–73. doi:10.1016/0010-0277(95)00664-8.