Jump to content

Directed information

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by Me, Myself, and I are Here (talk | contribs) at 21:17, 20 November 2016 (top: comma). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Directed information, , is a measure of information theory and it measures the amount of information that flows from the process to , where denotes the vector and denotes . The term "directed information" was coined by James Massey and is defined as

,

where is the conditional mutual information.

Note that if , directed information becomes mutual information . Directed information has many applications in problems where causality plays an important role such as capacity of channel with feedback,[1][2] capacity of discrete memoryless networks with feedback,[3] gambling with causal side information,[4] and compression with causal side information.[5]

References

  1. ^ Massey, James (1990). "Causality, Feedback And Directed Information" (ISITA). {{cite journal}}: Cite journal requires |journal= (help)
  2. ^ Permuter, Haim Henry; Weissman, Tsachy; Goldsmith, Andrea J. (February 2009). "Finite State Channels With Time-Invariant Deterministic Feedback". IEEE Transactions on Information Theory. 55 (2): 644–662. doi:10.1109/TIT.2008.2009849.
  3. ^ Kramer, G. (January 2003). "Capacity results for the discrete memoryless network". IEEE Transactions on Information Theory. 49 (1): 4–21. doi:10.1109/TIT.2002.806135.
  4. ^ Permuter, Haim H.; Kim, Young-Han; Weissman, Tsachy (June 2011). "Interpretations of Directed Information in Portfolio Theory, Data Compression, and Hypothesis Testing". IEEE Transactions on Information Theory. 57 (6): 3248–3259. doi:10.1109/TIT.2011.2136270.
  5. ^ Simeone, Osvaldo; Permuter, Haim Henri (June 2013). "Source Coding When the Side Information May Be Delayed". IEEE Transactions on Information Theory. 59 (6): 3607–3618. doi:10.1109/TIT.2013.2248192.