Comparing splits using information theory

To understand the information-based metrics implemented in ‘TreeDist’, it is useful to recall some basic concepts of information theory.

For an introduction, see MacKay (2003) or an introductory video to the clustering information distance:

Introduction to the Clustering Info Distance

Splits

Each internal edge in a tree represents a split that divides its leaves into two partitions. Intuitively, some splits are more instructive than others. For example, the fact that mammals and reptiles represent two separate groups is profound enough that it is worth teaching to schoolchildren; much less information is represented by a split that identifies two species of bat as more closely related to one another than to any other mammal or reptile.

Quantifying information

How can we formalize the intuition that some splits contain more information than others? More generally, how can we quantify an amount of information?

Information is usually measured in bits. One bit is the amount of information generated by tossing a fair coin: to record the outcome of a coin toss, I must record either a H or a T, and with each of the two symbols equally likely, there is no way to compress the results of multiple tosses.

The Shannon (1948) information content of an outcome x is defined to be h(x) = −log2P(x), which simplifies to log2n when all n outcomes are equally likely. Thus, the outcome of a fair coin toss delivers log22 = 1 bit of information; the outcome of rolling a fair six-sided die contains log26 ≈ 2.58 bits of information; and the outcome of selecting at random one of the 105 unrooted binary six-leaf trees is log2105 ≈ 6.71 bits.

Unlikely outcomes are more surprising, and thus contain more information than likely outcomes. The information content of rolling a twelve on two fair six-sided dice is $-\log_2{\frac{1}{36}} \approx 5.16\textrm{ bits}$, whereas a seven, which could be produced by six of the 36 possible rolls (1 & 6, 2 & 5, …), is less surprising, and thus contains less information: $-\log_2{\frac{6}{36}} \approx 2.58\textrm{ bits}$. An additional 2.58 bits of information would be required to establish which of the six possible rolls produced the seven.

Application to splits

The split S1= AB|CDEF is found in 15 of the 105 six-leaf trees; as such, the probability that a randomly drawn tree contains S1 is $P(S_1) = \frac{15}{105}$, and the information content $h(S_1) = -\log_2{\frac{15}{105}} \approx 2.81\textrm{ bits}$. Steel & Penny (2006) dub this quantity the phylogenetic information content.

Likewise, the split S2= ABC|DEF occurs in nine of the 105 six-leaf trees, so $h(S_2) = -\log_2{\frac{9}{105}} \approx 3.54\textrm{ bits}$. Three six-leaf trees contain both splits, so in combination the splits deliver $h(S_1,S_2) = -\log_2{\frac{3}{105}} \approx 5.13\textrm{ bits}$ of information.

Because h(S1, S2) < h(S1) + h(S2), some of the information in S1 is also present in S2. The information in common between S1 and S2 is hshared(S1, S2) = h(S1) + h(S2) − h(S1, S2) ≈ 1.22 bits. The information unique to S1 and S2 is hdifferent(S1, S2) = 2h(S1, S2) − h(S1) − h(S2) ≈ 3.91 bits.

These quantities can be calculated using functions in the ‘TreeTools’ package.

library("TreeTools", quietly = TRUE)
library("TreeDist")
treesMatchingSplit <- c(
  AB.CDEF = TreesMatchingSplit(2, 4),
  ABC.DEF = TreesMatchingSplit(3, 3)
)
treesMatchingSplit
## AB.CDEF ABC.DEF 
##      15       9
proportionMatchingSplit <- treesMatchingSplit / NUnrooted(6)
proportionMatchingSplit
##    AB.CDEF    ABC.DEF 
## 0.14285714 0.08571429
splitInformation <- -log2(proportionMatchingSplit)
splitInformation
##  AB.CDEF  ABC.DEF 
## 2.807355 3.544321
treesMatchingBoth <- TreesConsistentWithTwoSplits(6, 2, 3)
combinedInformation <- -log2(treesMatchingBoth / NUnrooted(6))

sharedInformation <- sum(splitInformation) - combinedInformation
sharedInformation
## [1] 1.222392
# Or more concisely:
SplitSharedInformation(n = 6, 2, 3)
## [1] 1.222392

Entropy

Entropy is the average information content of each outcome, weighted by its probability: ∑−plog2(p). Where all n outcomes are equiprobable, this simplifies to log2n.

Consider a case in which Jane rolls a dice, and makes two true statements about the outcome x:

S1: “Is the roll even?”.

  • Two equally-possible outcomes: yes or no
  • Entropy: H(S1) = log22 = 1 bit.

S2: “Is the roll greater than 3?”

  • Two equally-possible outcomes: yes or no
  • Entropy: H(S2) = log22 = 1 bit.

The joint entropy of S1 and S2 is the entropy of the association matrix that considers each possible outcome:

  S1 : x odd S1 : x even
S2 : x ≤ 3 $x \in {1, 3}; p = \frac{2}{6}$ $x = 2; p = \frac{1}{6}$
S2 : x > 3 $x = 5; p = \frac{1}{6}$ $x \in {4, 6}; p = \frac{2}{6}$

$\begin{aligned} H(S_1, S_2) = \frac{2}{3}\log_2{\frac{2}{3}} + \frac{1}{3}\log_2{\frac{1}{3}} + \frac{1}{3}\log_2{\frac{1}{3}} + \frac{2}{3}\log_2{\frac{2}{3}} \approx 1.84 \textrm{ bits} \end{aligned}$

Note that this less than the log26 ≈ 2.58 bits we require to determine the exact value of the roll: knowledge of S1 and S2 is not guaranteed to be sufficient to unambiguously identify x.

The mutual information between S1 and S2 describes how much knowledge of S1 reduces our uncertainty in S2 (or vice versa). So if we learn that S1 is ‘even’, we become a little more confident that S2 is ‘greater than three’.

The mutual information I(S1; S2), denoted in blue below, corresponds to the sum of the individual entropies, minus the joint entropy:

If two statements have high mutual information, then once you have heard one statement, you already have a good idea what the outcome of the other statement will be, and thus learn little new on hearing it.

The entropy distance, also termed the variation of information (Meila, 2007), corresponds to the information that S1 and S2 do not have in common (denoted below in yellow):

The higher the entropy distance, the harder it is to predict the outcome of one statement from the other; the maximum entropy distance occurs when the two statements are entirely independent.

Application to splits

A split divides leaves into two partitions. If we arbitrarily label these partitions ‘A’ and ‘B’, and select a leaf at random, we can view the partition label associated with the leaf. If 60/100 leaves belong to partition ‘A’, and 40/100 to ‘B’, then the a leaf drawn at random has a 40% chance of bearing the label ‘A’; the split has an entropy of $-\frac{60}{100}\log_2{\frac{60}{100}}-\frac{40}{100}\log_2{\frac{40}{100}} \approx 0.97\textrm{ bits}$.

Now consider a different split, perhaps in a different tree, that assigns 50 leaves from ‘A’ to a partition ‘C’, leaving the remaining 10 leaves from ‘A’, along with the 40 from ‘B’, in partition ‘D’. This split has $-\frac{50}{100}\log_2{\frac{50}{100}}-\frac{50}{100}\log_2{\frac{50}{100}} = 1\textrm{ bit}$ of entropy.
Put these together, and a randomly selected leaf may now bear one of three possible labellings:

  • ‘A’ and ‘C’: 50 leaves
  • ‘A’ and ‘D’: 10 leaves
  • ‘B’ and ‘D’: 40 leaves.

The two splits thus have a joint entropy of $-\frac{50}{100}\log_2{\frac{50}{100}} -\frac{10}{100}\log_2{\frac{10}{100}} -\frac{40}{100}\log_2{\frac{40}{100}} \approx 1.36\textrm{ bits} < 0.97 + 1$.

The joint entropy is less than the sum of the individual entropies because the two splits contain some mutual information: for instance, if a leaf bears the label ‘B’, we can be certain that it will also bear the label ‘D’. The more similar the splits are, and the more they agree in their division of leaves, the more mutual information they will exhibit. I term this the clustering information, in contradistinction to the concept of phylogenetic information discussed above.

More formally, let split S divides n leaves into two partitions A and B. The probability that a randomly chosen leaf x is in partition k is $P(x \in k) = \frac{|k|}{n}$. S thus corresponds to a random variable with entropy $H(S) = -\frac{|A|}{n} \log_2{\frac{|A|}{n}} - \frac{|B|}{n}\log_2{\frac{|B|}{n}}$ (Meila, 2007). The joint entropy of two splits, S1 and S2, corresponds to the entropy of the association matrix of probabilities that a randomly selected leaf belongs to each pair of partitions:

  S1 : x ∈ A1 S1 : x ∈ B1
S2 : x ∈ A2 $P(A_1,A_2) = \frac{|A_1 \cap A_2|}{n}$ $P(B_1,A_2) = \frac{|B_1 \cap A_2|}{n}$
S2 : x ∈ B2 $P(A_1,B_2) = \frac{|A_1 \cap B_2|}{n}$ $P(B_1,B_2) = \frac{|B_1 \cap B_2|}{n}$

H(S1, S2) = P(A1, A2)log2P(A1, A2) + P(B1, A2)log2P(B1, A2)

+P(A1, B2)log2P(A1, B2) + P(B1, B2)log2P(B1, B2)

These values can then be substituted into the definitions of mutual information and entropy distance given above.

As S1 and S2 become more different, the disposition of S1 gives less information about the configuration of S2, and the mutual information decreases accordingly.

References

MacKay, D. J. C. (2003). Information theory, inference, and learning algorithms. Cambridge: Cambridge University Press. Retrieved from https://www.inference.org.uk/itprnn/book.pdf
Meila, M. (2007). Comparing clusterings—an information based distance. Journal of Multivariate Analysis, 98(5), 873–895. doi: 10.1016/j.jmva.2006.11.013
Shannon, C. E. (1948). A mathematical theory of communication. Bell System Technical Journal, 27, 379–423, 623–656.
Steel, M. A., & Penny, D. (2006). Maximum parsimony and the phylogenetic information in multistate characters. In V. A. Albert (Ed.), Parsimony, phylogeny, and genomics (pp. 163–178). Oxford: Oxford University Press.