The very top of the page is marked "ROUGH DRAFT - BEWARE suggestions <email address elided>". That certainly is true with regard to pages 21 and 22. The row and column labels are missing on matrix A on page 21, and what he said about the eigenvalues on page 22 is inconsistent.
Let me try to reconstruct that. This is a guess. A "ROUGH DRAFT", if you will. The matrix A is an occurrence matrix. Each element indicates how many times a given word occurs in a given document. With labels attached, here's my guess as to that matrix:
<br />
\begin{array}{lrrrrr}<br />
& \text{#1} & \text{#2} & \text{#3} & \text{#4} & \text{#5} \\<br />
\text{doctor} & 2 & 0 & 8 & 6 & 0 \\<br />
\text{car} & 1 & 6 & 0 & 1 & 7 \\<br />
\text{hospital} & 5 & 0 & 7 & 4 & 0 \\<br />
\text{nurse} & 7 & 0 & 8 & 5 & 0 \\<br />
\text{wheel} & 0 & 10 & 0 & 0 & 7<br />
\end{array}
The rows represent different words whose occurrence we want to study; the columns represent documents in which the words appear. I've made a guess as to which word corresponds to which row. I've just labeled the columns #1, #2, etc. The names of those documents isn't relevant here. (Obviously it is in an analysis of those documents.)
Some obvious things that jump out, just looking at the matrix:
- Doctors, nurses, and hospitals are deeply connected with one another. You can see this in documents #1, #3, and #4.
- Cars and wheels are deeply connected with one another. Two of the documents, #2 and #5, only mention cars and wheels.
- A bit weaker, nurses and cars are connected when doctors are absent. Documents #1 and #4 mention cars, document #3 doesn't. Document #1 barely mentions doctors but does mention cars. Document #3 mentions doctors a lot but doesn't mention cars at all.
The following link
computes an SVD of that array, courtesy WolframAlpha.
Note that except for some sign changes, the U matrix in that document and the U matrix from WolframAlpha are the same. Negate an eigenvector and you still have an eigenvector.
The author of the paper completely messed up the interpretation of the matrix. The first column vector corresponds to the largest singular value, so it's the most significant eingenvector. That column says that doctors, nurses, and hospitals go hand in hand. A document that contains one of those words is most likely to contain all three. Cars and wheels are a bit irrelevant in this first eigenvector. The second eigenvector says that some documents contain the words car and wheel and none of the other words. The third eigenvector represents documents that don't contain the words doctor or wheel but do contain the words nurse, and to a lesser extent contain the words car and hospital. The fourth and fifth eigenvectors? Now we're in the noise. Those eigenvectors are small, particularly for a 5x5 matrix.