Bayesian search theory

From Wikipedia, the free encyclopedia
Jump to: navigation, search

Bayesian search theory is the application of Bayesian statistics to the search for lost objects. It has been used several times to find lost sea vessels, for example the USS Scorpion. It also played a key role in the recovery of the flight recorders in the Air France Flight 447 disaster of 2009.

Procedure[edit]

The usual procedure is as follows:

  1. Formulate as many reasonable hypotheses as possible about what may have happened to the object.
  2. For each hypothesis, construct a probability density function for the location of the object.
  3. Construct a function giving the probability of actually finding an object in location X when searching there if it really is in location X. In an ocean search, this is usually a function of water depth — in shallow water chances of finding an object are good if the search is in the right place. In deep water chances are reduced.
  4. Combine the above information coherently to produce an overall probability density map. (Usually this simply means multiplying the two functions together.) This gives the probability of finding the object by looking in location X, for all possible locations X. (This can be visualized as a contour map of probability.)
  5. Construct a search path which starts at the point of highest probability and 'scans' over high probability areas, then intermediate probabilities, and finally low probability areas.
  6. Revise all the probabilities continuously during the search. For example, if the hypotheses for location X imply the likely disintegration of the object and the search at location X has yielded no fragments, then the probability that the object is somewhere around there is greatly reduced (though not usually to zero) while the probabilities of its being at other locations is correspondingly increased. The revision process is done by applying Bayes' theorem.

In other words, first search where it most probably will be found, then search where finding it is less probable, then search where the probability is even less (but still possible due to limitations on fuel, range, water currents, etc.), until insufficient hope of locating the object at acceptable cost remains.

The advantages of the Bayesian method are that all information available is used coherently (i.e., in a "leak-proof" manner) and the method automatically produces estimates of the cost for a given success probability. That is, even before the start of searching, one can say, hypothetically, "there is a 65% chance of finding it in a 5-day search. That probability will rise to 90% after a 10-day search and 97% after 15 days" or a similar statement. Thus the economic viability of the search can be estimated before committing resources to a search.

Apart from the USS Scorpion, other vessels located by Bayesian search theory include the MV Derbyshire, the largest British vessel ever lost at sea, and the SS Central America. It also proved successful in the search for a lost hydrogen bomb following the 1966 Palomares B-52 crash in Spain, and the recovery in the Atlantic Ocean of the crashed Air France Flight 447.

Bayesian search theory is incorporated into the CASP (Computer Assisted Search Program) mission planning software used by the United States Coast Guard for search and rescue. This program was later adapted for inland search by adding terrain and ground cover factors for use by the United States Air Force and Civil Air Patrol.

Mathematics[edit]

Suppose a grid square has a probability p of containing the wreck and that the probability of successfully detecting the wreck if it is there is q. If the square is searched and no wreck is found, then, by Bayes' theorem, the revised probability of the wreck being in the square is given by

  p' = \frac{p(1-q)}{(1-p)+p(1-q)} = p \frac{1-q}{1-pq} < p.

For each other grid square, if its prior probability is r, its posterior probability is given by

 r' = r \frac{1}{1- pq} > r.


Optimal Distribution of Search Effort[edit]

The classical book on this subject, based on probabilistic information, by Lawrence D. Stone, won the 1975 Frederick W. Lanchester Prize by the Operations Research Society of America.

USS Scorpion[edit]

In May 1968, the U.S. Navy's nuclear submarine USS Scorpion (SSN-589) failed to arrive as expected at her home port of Norfolk, Virginia. The command officers of the U.S. Navy were nearly convinced that the vessel had been lost off the Eastern Seaboard, but an extensive search there failed to discover the remains of the Scorpion.

Then, a Navy deep-water expert, John P. Craven, suggested that the USS Scorpion had sunk elsewhere. Craven organised a search southwest of the Azores based on a controversial approximate triangulation by hydrophones. He was allocated only a single ship, the Mizar, and he took advice from a firm of consultant mathematicians in order to maximise his resources. A Bayesian search methodology was adopted. Experienced submarine commanders were interviewed to construct hypotheses about what could have caused the loss of the Scorpion.

The sea area was divided up into grid squares and a probability assigned to each square, under each of the hypotheses, to give a number of probability grids, one for each hypothesis. These were then added together to produce an overall probability grid. The probability attached to each square was then the probability that the wreck was in that square. A second grid was constructed with probabilities that represented the probability of successfully finding the wreck if that square were to be searched and the wreck were to be actually there. This was a known function of water depth. The result of combining this grid with the previous grid is a grid which gives the probability of finding the wreck in each grid square of the sea if it were to be searched.


Optimal distribution of Search Effort[edit]

The classical book on this subject The Theory of Optimal Search (Operations Research Society of America, 1975) by Lawrence D. Stone won the 1975 Lancaster Prize by the American Operations Research Society.

Searching in Boxes[edit]

Assume that a stationary object is hidden in one of n boxes. For each box i there are three known parameters: the cost of a single search at location i, c_i; the probability of finding the object by a single search at location i if it is at this location, a_i; and the probability that the object is at location i, p_i. A searcher looks for the object. He knows the a priori probabilities at the beginning and updates them by Bayes’ law after each (unsuccessful) attempt. The problem of finding the object in minimal expected cost is a classical problem solved by David Blackwell. Surprisingly, the optimal policy is easy to describe: at each stage look into the location which maximizes p_i × a_i / c_i. This is actually a special case of Gittins index.

See also[edit]

References[edit]

  • Stone, Lawrence D., The Theory of Optimal Search, published by the Operations Research Society of America, 1975
  • Stone, Lawrence D., In Search of Air France Flight 447. Institute of Operations Research and the Management Sciences, 2011. https://www.informs.org/ORMS-Today/Public-Articles/August-Volume-38-Number-4/In-Search-of-Air-France-Flight-447
  • Iida, Koji., Studies on the Optimal Search Plan, Vol. 70, Lecture Notes in Statistics, Springer-Verlag, 1992.
  • De Groot, Morris H., Optimal Statistical Decisions, Wiley Classics Library, 2004.
  • Richardson, Henry R; and Stone, Lawrence D. Operations Analysis during the underwater search for Scorpion. Naval Research Logistics Quarterly, June 1971, Vol. 18, Number 2. Office of Naval Research.
  • Stone, Lawrence D. Search for the SS Central America: Mathematical Treasure Hunting. Technical Report, Metron Inc. Reston, Virginia.
  • Koopman, B.O. Search and Screening, Operations Research Evaluation Group Report 56, Center for Naval Analyses, Alexandria, Virginia. 1946.
  • Richardson, Henry R; and Discenza, J.H. The United States Coast Guard computer-assisted search planning system (CASP). Naval Research Logistics Quarterly. Vol. 27 number 4. pp. 659–680. 1980.
  • Ross, Sheldon M., An Introduction to Stochastic Dynamic Programming, Academic Press. 1983.