Question: How to calculate the joint entropy H(X,Y,Z) where X, Y, Z are position of MSA
0
sivanc710 wrote:

Hi to all!

I want to perform a mutual information (MI) among three or more position of one multiple sequence alignment (MSA). So far, I know that

MI(XY) = sum( p(X,Y) * log ( p(X, Y) / ( p(X) * p(Y) ) ) )

But the MI among three position is:

MI(X,Y,Z) = MI(X,Y) + MI(X,Z) + MI(Y,Z) - [H(X) + H(Y) + H(Z) - H(X,Y,Z)]

where H(X) is calculated as:

H(X) = sum( p(X) * log p(X) )

The problem here is that I don't know how to calculate the joint entropy H(X,Y,Z). Someone knows how to calculate it? Also, if someone knows how to extend the MI among more than three position I will be grateful for the info.

Thanks to all! =)

entropy mutual information rna • 2.5k views
modified 4.8 years ago by Brian Bushnell17k • written 4.8 years ago by sivanc710

It's in Wikipedia: http://en.wikipedia.org/wiki/Joint_entropy.

0
Brian Bushnell17k wrote:

I don't really understand why this matters.  Why can you not take the minimum of the individual entropies of the different sequences?  They are all equally valid.  So, if two of them have very complex sequences, and the third is poly-A, well, you're not doing anything useful.  I posit that the information content of a multiple sequence alignment is constrained by the most uninformative sequence you decide to include.