Graph entropy

In information theory, the graph entropy is a measure of the information rate achievable by communicating symbols over a channel in which certain pairs of values may be confused.[1] This measure, first introduced by Körner in the 1970s,[2][3] has since also proven itself useful in other settings, including combinatorics.[4]

Definition

Let be an undirected graph. The graph entropy of , denoted is defined as

where is chosen uniformly from , ranges over independent sets of G, the joint distribution of and is such that with probability one, and is the mutual information of and .[5]

Properties

  • Monotonicity. If is a subgraph of on the same vertex set, then .
  • Subadditivity. Given two graphs and on the same set of vertices, the graph union satisfies .
  • Arithmetic mean of disjoint unions. Let be a sequence of graphs on disjoint sets of vertices, with vertices, respectively. Then .

Additionally, simple formulas exist for certain families classes of graphs.

Example

Here, we use properties of graph entropy to provide a simple proof that a complete graph on vertices cannot be expressed as the union of fewer than bipartite graphs.

Proof By monotonicity, no bipartite graph can have graph entropy greater than that of a complete bipartite graph, which is bounded by . Thus, by sub-additivity, the union of bipartite graphs cannot have entropy greater than . Now let be a complete graph on vertices. By the properties listed above, . Therefore, the union of fewer than bipartite graphs cannot have the same entropy as , so cannot be expressed as such a union.

General References

  • Matthias Dehmer; Frank Emmert-Streib; Zengqiang Chen; Xueliang Li; Yongtang Shi (25 July 2016). Mathematical Foundations and Applications of Graph Entropy. Wiley. ISBN 978-3-527-69325-2.

Notes

  1. Matthias Dehmer; Abbe Mowshowitz; Frank Emmert-Streib (21 June 2013). Advances in Network Complexity. John Wiley & Sons. pp. 186–. ISBN 978-3-527-67048-2.
  2. Körner, János (1973). "Coding of an information source having ambiguous alphabet and the entropy of graphs". 6th Prague conference on information theory: 411–425.
  3. Niels da Vitoria Lobo; Takis Kasparis; Michael Georgiopoulos (24 November 2008). Structural, Syntactic, and Statistical Pattern Recognition: Joint IAPR International Workshop, SSPR & SPR 2008, Orlando, USA, December 4-6, 2008. Proceedings. Springer Science & Business Media. pp. 237–. ISBN 978-3-540-89688-3.
  4. Bernadette Bouchon; Lorenza Saitta; Ronald R. Yager (8 June 1988). Uncertainty and Intelligent Systems: 2nd International Conference on Information Processing and Management of Uncertainty in Knowledge Based Systems IPMU '88. Urbino, Italy, July 4-7, 1988. Proceedings. Springer Science & Business Media. pp. 112–. ISBN 978-3-540-19402-6.
  5. G. Simonyi, "Perfect graphs and graph entropy. An updated survey," Perfect Graphs, John Wiley and Sons (2001) pp. 293-328, Definition 2”
This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.