noowutah
- 56
- 3
It is well-known that with known marginal probabilities a_{i} and
b_{j} the joint probability distribution maximizing the entropy
H(P)=-\sum_{i=1}^{m}\sum_{j=1}^{n}p_{ij}\log{}p_{ij}
is p_{ij}=a_{i}b_{j}
For m=3 and n=3, a=(0.2,0.3,0.5), b=(0.1,0.6,0.3), for example,
\begin{equation}<br /> \label{eq:r1}<br /> P=\left(<br /> \begin{array}{rrr}<br /> 0.02 & 0.12 & 0.06 \\<br /> 0.03 & 0.18 & 0.09 \\<br /> 0.05 & 0.30 & 0.15<br /> \end{array}<br /> \right)<br /> \end{equation}
Now, I have a problem where the joint probability distribution is
constrained (much like a random walk where the path from one node to
another is blocked). For example, let m, n, a, and b be as above with
the constraint that
\begin{equation}<br /> \label{eq:r2}<br /> P=\left(<br /> \begin{array}{rrr}<br /> p_{11} & 0 & p_{13} \\<br /> p_{21} & p_{22} & p_{23} \\<br /> p_{31} & p_{32} & p_{33} \\<br /> \end{array}<br /> \right)<br /> \end{equation}
Because a and b are known, it suffices to find out any 2x2 matrix
contained in the 3x3 matrix, for example (x,y,z the variables by which
w_{1}, w_{2}, v_{1}, v_{2}, and sigma are expressible, given a and b)
\begin{equation}<br /> \label{eq:r3}<br /> P=\left(<br /> \begin{array}{rrr}<br /> x & 0 & w_{1} \\<br /> y & z & w_{2} \\<br /> v_{1} & v_{2} & \sigma \\<br /> \end{array}<br /> \right)<br /> \end{equation}
I use this to write out the entropy and differentiate with respect to
x, y, and z to find out that the maximum will be where
\begin{equation}<br /> \label{eq:r4}<br /> \det\left(<br /> \begin{array}{ll}<br /> x & w_{1} \\<br /> v_{1} & \sigma<br /> \end{array}<br /> \right)=0<br /> \end{equation}
\begin{equation}<br /> \label{eq:r5}<br /> \det\left(<br /> \begin{array}{ll}<br /> y & w_{2} \\<br /> v_{1} & \sigma<br /> \end{array}<br /> \right)=0<br /> \end{equation}
\begin{equation}<br /> \label{eq:r6}<br /> \det\left(<br /> \begin{array}{ll}<br /> z & w_{2} \\<br /> v_{2} & \sigma<br /> \end{array}<br /> \right)=0<br /> \end{equation}
This is a system of 3 non-linear equations which are awkward to solve
algebraically. In the end, I am interested to know which property a
and b need to have to ensure that P is a proper probability
distribution (i.e. no negative elements). For now, however, I'd be
thrilled if anybody could give me a hint how I could find the
solutions for x, y, and z algebraically without these non-linear
equations.
The numeric solution for this is
\begin{equation}<br /> \label{eq:r7}<br /> P=\left(<br /> \begin{array}{rrr}<br /> 0.093 & 0.000 & 0.107 \\<br /> 0.001 & 0.113 & 0.186 \\<br /> 0.006 & 0.487 & 0.007<br /> \end{array}<br /> \right)<br /> \end{equation}
But I have definitely seen systems where all solutions were complex
and some of the probabilities ended up <0, especially when m and n are
larger than 3 and there are more than one zero constraint in the joint
probability matrix.
b_{j} the joint probability distribution maximizing the entropy
H(P)=-\sum_{i=1}^{m}\sum_{j=1}^{n}p_{ij}\log{}p_{ij}
is p_{ij}=a_{i}b_{j}
For m=3 and n=3, a=(0.2,0.3,0.5), b=(0.1,0.6,0.3), for example,
\begin{equation}<br /> \label{eq:r1}<br /> P=\left(<br /> \begin{array}{rrr}<br /> 0.02 & 0.12 & 0.06 \\<br /> 0.03 & 0.18 & 0.09 \\<br /> 0.05 & 0.30 & 0.15<br /> \end{array}<br /> \right)<br /> \end{equation}
Now, I have a problem where the joint probability distribution is
constrained (much like a random walk where the path from one node to
another is blocked). For example, let m, n, a, and b be as above with
the constraint that
\begin{equation}<br /> \label{eq:r2}<br /> P=\left(<br /> \begin{array}{rrr}<br /> p_{11} & 0 & p_{13} \\<br /> p_{21} & p_{22} & p_{23} \\<br /> p_{31} & p_{32} & p_{33} \\<br /> \end{array}<br /> \right)<br /> \end{equation}
Because a and b are known, it suffices to find out any 2x2 matrix
contained in the 3x3 matrix, for example (x,y,z the variables by which
w_{1}, w_{2}, v_{1}, v_{2}, and sigma are expressible, given a and b)
\begin{equation}<br /> \label{eq:r3}<br /> P=\left(<br /> \begin{array}{rrr}<br /> x & 0 & w_{1} \\<br /> y & z & w_{2} \\<br /> v_{1} & v_{2} & \sigma \\<br /> \end{array}<br /> \right)<br /> \end{equation}
I use this to write out the entropy and differentiate with respect to
x, y, and z to find out that the maximum will be where
\begin{equation}<br /> \label{eq:r4}<br /> \det\left(<br /> \begin{array}{ll}<br /> x & w_{1} \\<br /> v_{1} & \sigma<br /> \end{array}<br /> \right)=0<br /> \end{equation}
\begin{equation}<br /> \label{eq:r5}<br /> \det\left(<br /> \begin{array}{ll}<br /> y & w_{2} \\<br /> v_{1} & \sigma<br /> \end{array}<br /> \right)=0<br /> \end{equation}
\begin{equation}<br /> \label{eq:r6}<br /> \det\left(<br /> \begin{array}{ll}<br /> z & w_{2} \\<br /> v_{2} & \sigma<br /> \end{array}<br /> \right)=0<br /> \end{equation}
This is a system of 3 non-linear equations which are awkward to solve
algebraically. In the end, I am interested to know which property a
and b need to have to ensure that P is a proper probability
distribution (i.e. no negative elements). For now, however, I'd be
thrilled if anybody could give me a hint how I could find the
solutions for x, y, and z algebraically without these non-linear
equations.
The numeric solution for this is
\begin{equation}<br /> \label{eq:r7}<br /> P=\left(<br /> \begin{array}{rrr}<br /> 0.093 & 0.000 & 0.107 \\<br /> 0.001 & 0.113 & 0.186 \\<br /> 0.006 & 0.487 & 0.007<br /> \end{array}<br /> \right)<br /> \end{equation}
But I have definitely seen systems where all solutions were complex
and some of the probabilities ended up <0, especially when m and n are
larger than 3 and there are more than one zero constraint in the joint
probability matrix.
Last edited by a moderator: