Dear All,
I'll be very appreciative of your help to deal with the following. Given a RxCxLxS table A, I merge either two rows, two columns, two layers or two strata according to merge criteria. Resulting table B, say, is still a four-way table. I would like to assess information loss relative to move from A to B. I claim Information Loss = Entropy(A) - Entropy(B). Freitas (http://www.ppgia.pucpr.br/~alex/thesis.html) defined Information loss as Info_After_Merge - Info_Before_Merge where Info_After_Merge = -SUMj(C+j/C++)log(C+j/C++) and Info_Before_Merge = -SUMi(Ci+/C++)SUMj(Cij/Ci+)log(Cij/Ci+). This turns out to be, in two-way contingency table, Column(entropy) - Column(entropy given row). (According to definitions & formulae in Statistical Decomposition Analysis, Theil, 1972).
What is the rational for considering Column(entropy after merge) vs. Column(entropy given row before merge)? I would go for either Column(entropy after merge) - Column(entropy before merge) or Column(entropy given row after merge) - Column(entropy given row before merge). I might ignore that Column(entropy given row) = Column(entropy) though! Most importantly, how can this be extended to a four-way table?
Next, in case of two-way table Theil defines the expected mutual information as the amount to be subtracted from the total marginal entropy, Column(entropy) + Row(entropy), in order to obtain the joint entropy (or coentropy). Applied to 3-way XYZ table this definition should give XYY(coentropy) = XY(marginal entropy) + XZ(marginal entropy) + YZ(marginal entropy) + Column(entropy) + Row(entropy) + Layer(entropy) - XYZ(mutual information). Virtanen and Astola (http://www.uwasa.fi/~itv/publicat/entropy.html) established I(XYZ) = H(XYZ)-H(XY)-H(YZ)-H(ZX)+H(X)+H(Y)+H(Z) or
H(XYZ) = I(XYZ)+H(XY)+H(YZ)+H(ZX)-H(X)-H(Y)-H(Z). i.e. mutual information is not subtracted; it is added instead as H(XYZ)=XYZ(coentropy), H(XY)=XY(marginal entropy), H(XZ)=XZ(marginal entropy), H(YZ)=YZ(marginal entropy), H(Y) =Column(entropy), H(X)=Row(entropy), H(Z)=Layer(entropy), and I(XYZ) = XYZ(mutual information). The formula
I(XYZ) = -SUMiSUMjSUMk(Pijk*log(Pij.*Pi.k*P.jk/Pijk*Pi..*P.j.*P..k)) (in Virtanen and Astola) leads to I(XYZ) = H(XY)+H(YZ)+H(ZX)-H(XYZ)-H(X)-H(Y)-H(Z). Here mutual information and one-dimensional entropies are subtracted from the total two-dimensional marginal entropy. Question is what the right answer.
Generalizing the formula I(XYZ) = -SUMiSUMjSUMk(Pijk*log(Pij.*Pi.k*P.jk/Pijk*Pi..*P.j.*P..k)) to four-way case, gives
I(WXYZ) = -SUMiSUMjSUMkSUMm(Pijkm*log(Pijk.*Pij.m*Pi.km*P.jkm/Pijkm*Pi...*P.j..*P..k.*P...m) whose expansion does not involve two-dimensional marginal entropies. Could someone gives the right formula?
What would be good recent references to consult for entropy concept applied to multiway tables?
Thanks in advance for your valuable help and suggestions,
Ray Haraf.
|