- #1
- 236
- 40
For example, does this always hold true?
M_ab = v_a × w_b
If not, where does it break down?
M_ab = v_a × w_b
If not, where does it break down?
Let's say there are a column vector ##A = \begin{bmatrix}a \\ b\end{bmatrix}## and row vector ##B = \begin{bmatrix}c & d\end{bmatrix}## that have
##AB = \begin{bmatrix}ac & ad \\ bc & bd\end{bmatrix} = \begin{bmatrix}0 & 1 \\ 1 & 0\end{bmatrix}##.
Is this possible, knowing that ##xy = 0## for real numbers ##x,y## implies that either ##x=0## or ##y=0## ?
It can always be done by a linear combination of those where ##\operatorname{rank}M## is the minimal length of it.For example, does this always hold true?
M_ab = v_a × w_b
If not, where does it break down?
For example, does this always hold true?
M_ab = v_a × w_b
If not, where does it break down?
Sure, they are rank one matrices.Also, I guess the determinant of this kind of matrices (if they're square) is always zero, as the columns are multiples of each other. This is a very limiting property for a matrix.
The question gets interesting, if we ask for the minimal length of linear combinations of ##x\otimes y \otimes z## to represent a given bilinear mapping, e.g. matrix multiplication. If we define the matrix exponent ## \omega := \min\{\,\gamma\,|\,(A;B) \longmapsto A\cdot B = \sum_{i}^Rx_i(A) \otimes y_i(B) \otimes Z_i \,\wedge \, R=O(n^\gamma)\,\} ## then ##2\leq \omega \leq 2.3727## and we do not know how close we can come to the lower bound.