Given certain matrix [tex]A\in\mathbb{R}^{n\times m},[/tex]the rank d approximation L with the same number of rows/column as A, minimizing the Frobenius norm of the difference [tex]||A-L||[/tex] is matrix obtained by singular value decomposition of A, with only d dominant singular values (the rest is simply set to zero).(adsbygoogle = window.adsbygoogle || []).push({});

However, I often encounter the minimization of the adapted norm, such as various kinds of normalization on the norm, ie.

[tex]i) ||A-K||^2[/tex]

[tex]ii) \left(\frac{||A-K||}{||A||}\right)^{1/2}[/tex]

and I'm not sure if the solution L from the above non-squared Frobenius norm coincides with the normalized Frobenius norm solution from i) and ii).

Isn't it the case that K should be L, but appropriately scaled for i) and/or ii)?

**Physics Forums - The Fusion of Science and Community**

# Squared norms: difference or notational convenience

Know someone interested in this topic? Share a link to this question via email,
Google+,
Twitter, or
Facebook

Have something to add?

- Similar discussions for: Squared norms: difference or notational convenience

Loading...

**Physics Forums - The Fusion of Science and Community**