Quadratic forms show up in many places. In physics, energy is often a quadratic form. For example, the kinetic energy of a rigid body is
<br />
T = \frac{1}{2}\mathbf{\omega^T I \omega}<br />
where \mathbf{\omega} is the angular velocity vector (3x1) and \mathbf{I} is the tensor of inertia (just think of it as a 3x3 matrix). Often times we want to find the principle axes, which simply means finding a rotation that makes \mathbf{I} diagonal (this is where your eigenvectors matter!), so that the quadratic form becomes a simple sum of squares.
Another place they show up is in optimization. Consider a twice-differentiable function of N variables f(\mathbf{x}), where \mathbf{x} is the Nx1 vector of variables. If we want to find a local maximum and minimum, these will occur at a location (call it \mathbf{x=x_0}) where the first derivatives are zero,
<br />
\left. \nabla f(x) \right|_{\mathbf{x=x_0}} = \mathbf{0}.<br />
This equation represents N scalar equations. The first two terms of the Taylor expansion of f about \mathbf{x=x_0} is then
<br />
f(\mathbf{x}) \approx f(\mathbf{x_0}) + \frac{1}{2}(\mathbf{x-x_0})^T \mathbf{H(\mathbf{x_0})}(\mathbf{x-x_0})<br />
where each element of \mathbf{H} (called the Hessian) is simply a second derivative evaluated at \mathbf{x=x_0}:
<br />
H_{ij}(\mathbf{x_0}) = \left. \frac{\partial^2 f}{\partial x_i \partial x_j} \right|_{\mathbf{x=x_0}}<br />
If the quadratic form is positive definite (all eigenvalues are positive), then \mathbf{x=x_0} is a local minima, if it is negative definite (all eigenvalues are negative) then it is a local maxima.
They show up in more places as well. So be rest assured that learning quadratic forms is useful!
jason