Directed information
Appearance
This article has multiple issues. Please help improve it or discuss these issues on the talk page. (Learn how and when to remove these messages)
|
Directed information, , is a measure of information theory and it measures the amount of information that flows from the process to , where denotes the vector and denotes . The term "directed information" was coined by James Massey and is defined as
- ,
where is the conditional mutual information.
Note that if , directed information becomes mutual information . Directed information has many applications in problems where causality plays an important role such as capacity of channel with feedback,[1][2] capacity of discrete memoryless networks with feedback,[3] gambling with causal side information,[4] and compression with causal side information.[5]
References
- ^ Massey, James (1990). "Causality, Feedback And Directed Information" (ISITA).
{{cite journal}}
: Cite journal requires|journal=
(help) - ^ Permuter, Haim Henry; Weissman, Tsachy; Goldsmith, Andrea J. (February 2009). "Finite State Channels With Time-Invariant Deterministic Feedback". IEEE Transactions on Information Theory. 55 (2): 644–662. doi:10.1109/TIT.2008.2009849.
- ^ Kramer, G. (January 2003). "Capacity results for the discrete memoryless network". IEEE Transactions on Information Theory. 49 (1): 4–21. doi:10.1109/TIT.2002.806135.
- ^ Permuter, Haim H.; Kim, Young-Han; Weissman, Tsachy (June 2011). "Interpretations of Directed Information in Portfolio Theory, Data Compression, and Hypothesis Testing". IEEE Transactions on Information Theory. 57 (6): 3248–3259. doi:10.1109/TIT.2011.2136270.
- ^ Simeone, Osvaldo; Permuter, Haim Henri (June 2013). "Source Coding When the Side Information May Be Delayed". IEEE Transactions on Information Theory. 59 (6): 3607–3618. doi:10.1109/TIT.2013.2248192.