Jump to content

Logical depth

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by 71.238.24.162 (talk) at 03:31, 4 December 2019 (Added formal definition from Bennett's paper.). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Logical depth is a measure of complexity for individual strings devised by Charles H. Bennett based on the computational complexity of an algorithm that can recreate a given piece of information. It differs from Kolmogorov complexity in that it considers the computation time of the algorithm with nearly minimal length, rather than the length of the minimal algorithm.

Formally, in the context of some universal computer the logical depth of a string to significance level is given by the running time of the fastest program that produces and is no more than longer than the minimal program.

See also

References

  • Bennett, Charles H. (1988), "Logical Depth and Physical Complexity", in Herken, Rolf (ed.), The Universal Turing Machine: a Half-Century Survey, Oxford U. Press, pp. 227–257, CiteSeerX 10.1.1.70.4331
  • Craig, Edward (1998), "Computability and Information, Section 6: Logical depth", Routledge Encyclopedia of Philosophy, Vol. 10: Index, Taylor & Francis, p. 481, ISBN 9780415073103