samalkhaiat said:
3) Given a generalized function [itex]f \in \mathcal{D}^{\prime}(\mathbb{R}^{n})[/itex] and a good function [itex]a \in C^{\infty}(\mathbb{R}^{n})[/itex], how do you know that the Leibniz rule holds for the differentiation of the product [itex]af[/itex]? You don’t. But, if you use the definition (1), you can show that [tex]\partial^{m}\left( af \right) = \sum_{k \leq m} \begin{pmatrix} k \\ m \end{pmatrix} \ \partial^{k}a \ \partial^{m - k}f .[/tex]
To properly answer this, first let's set up the proper definitions so everything checks out latter.
Let [itex]V[/itex] be a vector space over the scalar field [itex]F[/itex] which can be either [itex]\mathbb{R}[/itex] or [itex]\mathbb{C}[/itex]. Given two elements of [itex]V[/itex] such as [itex]u, v \in V[/itex] be elements of [itex]V[/itex], then we define the inner product as map
$$(.,.): V \times V \to F$$
$$ u, v \to (u, v) \in F$$
such that it satisfies the following properties:
1. Conjugate symmetry:
$$(u, v) = (u, v)^{*}$$
where the notation means that given [itex]z \in \mathbb{C};[/itex] then [itex]z^{*}[/itex] is just the complex-conjugate of [itex]z[/itex].
2. Linearity in the first argument:
$$\alpha, \beta \in F; \quad u, v, b \in V; \quad (\alpha u + \beta v, b) = \alpha(u, b) + \beta(v, b)$$
3. Positive definite:
$$(x, x) \ge 0 \quad \forall x \in V \setminus \{0\}$$
Now, let [itex]\mathcal{D}(\mathbb{R}^n)[/itex] be the vector space over the field [itex]\mathbb{R}[/itex] (if it was over the field [itex]\mathbb{C}[/itex], then we'd say [itex]\mathcal{D}(\mathbb{C}^n)[/itex]) such that its elements are "nice" functions [itex]\varphi :\mathbb{R}^n \to \mathbb{R}[/itex]. We say that a function is "nice" if:
- It's smooth
- It has compact support
And finally, given [itex]f, g \in \mathcal{D}(\mathbb{R}^n)[/itex], we take the our inner product to be:
$$(f, g) := \displaystyle\int d^n x \bigg( f(x) g(x) \bigg)$$
With all this in place, we can finally define a "distribution". If [itex]V = \mathcal{D}(\mathbb{R}^n)[/itex] and [itex]V^{*}[/itex] is its dual space, then [itex]T_f \in V^{*}[/itex] is a linear form such that if [itex]f, \phi \in V[/itex] are nice functions, then [itex]\quad T_f[\phi] = (f, \phi) \in \mathbb{R}[/itex]. We call [itex]T_f[/itex] a "distribution", but usually we abuse the notation by calling [itex]f[/itex] itself the distribution. Because of that, it is desirable to define the derivative of a distribution such that [itex]\partial (T_f) = T_{\partial f}[/itex]. Using integration by parts and the fact that the nice functions have compact support:
$$(\partial f, \phi) = \displaystyle\int d^n x \bigg( \partial f(x) \phi(x) \bigg) = \cancel{\bigg[ f(x) \phi(x) \bigg]_{-\infty}^{\infty}} - \displaystyle\int d^n x \bigg( f(x) \partial \phi(x) \bigg) = -(f, \partial \phi)$$
$$\implies T_{\partial f}[\phi] = -T_{f}[\partial \phi]$$
From now on we'll abuse the notation too and refer to the distribution [itex]T_f[/itex] just as [itex]f[/itex].
Now, before proving the statement, I would like to point out that there's an error in the binomial number that you've written there, the [itex]m[/itex] and [itex]k[/itex] are swapped. I think the correct formula is:
$$\partial^{m}\left( af \right) = \sum_{k \leq m} \begin{pmatrix} m \\ k \end{pmatrix} \ \partial^{k}a \ \partial^{m - k}f = \sum_{k = 0}^m \begin{pmatrix} m \\ k \end{pmatrix} \ \partial^{k}a \ \partial^{m - k}f$$
based on what I see on wikipedia
here.
Anyway, let's begin with the proof:
We'll first prove that the distribution [itex]f[/itex] satisfies the Leibniz rule as usual [itex]\partial(af) = (\partial a) f + a(\partial f)[/itex] . Then the proof for the given formula for a function [itex]a[/itex] and a functional [itex]f[/itex] is the same as it would be for two functions.
Let [itex]f(x), \varphi(x) \in V[/itex] be nice functions and [itex]f[/itex] be a distribution. Also let [itex]\varphi(x) = a(x) \phi(x)[/itex] where [itex]a, \phi \in V[/itex]. We'll understand that when we write [itex]f(x)[/itex] we refer to the function and when we write [itex]f[/itex] or [itex]f[.][/itex] we refer to the distribution.
$$f[\varphi] = (f(x), \phi(x)) = \displaystyle\int d^n x \bigg( f(x) \varphi(x) \bigg) = \displaystyle\int d^n x \bigg( a(x) f(x) \phi(x) \bigg) = $$
$$ = \displaystyle\int d^n x \bigg( [a(x) f(x)] \phi(x) \bigg) = (af)[\phi] \equiv g[\phi]$$
where we have defined the distribution [itex]g[/itex] as [itex]af[/itex] and the function [itex]g(x) := a(x)f(x)[/itex]. Now
$$(\partial f, \varphi) = \displaystyle\int d^n x \bigg( \partial f(x) \varphi(x) \bigg) = -\displaystyle\int d^n x \bigg( f(x) \partial\{ a(x) \phi(x) \} \bigg) = $$
$$= -\displaystyle\int d^n x \bigg( \partial a(x) f(x) \phi(x) \bigg) -\displaystyle\int d^n x \bigg( g(x) \partial \phi(x) \bigg) = -(\partial a(x) f(x), \phi(x)) -(g(x), \partial \phi(x))$$
$$(\partial f, \varphi) = \displaystyle\int d^n x \bigg( \partial f(x) \varphi(x) \bigg) = \displaystyle\int d^n x \bigg( [a(x) \partial f(x)] \phi(x) \bigg) = (a(x)\partial f(x), \phi(x))$$
$$(a\partial f)[\phi] = -(\partial af)[\phi] - g[\partial \phi] = -(\partial af)[\phi] + (\partial g)[\phi] \implies$$
$$\implies \bigg( \partial(af) \bigg) = \bigg( (\partial a) f \bigg) + \bigg(a(\partial f) \bigg)$$
and so we've proven that distributions obey the Leibniz rule.
Q.E.D.
And so, because distributions behave the same way under differentiation as functions do, you can copy-paste the proof for the formula
$$\partial^{m}\left( af \right) = \sum_{k \leq m} \begin{pmatrix} m \\ k \end{pmatrix} \ \partial^{k}a \ \partial^{m - k}f = \sum_{k = 0}^m \begin{pmatrix} m \\ k \end{pmatrix} \ \partial^{k}a \ \partial^{m - k}f$$
for [itex]a[/itex] and [itex]f[/itex] being functions (which is trivial) into the one where they are distributions; thus proving the formula for distributions as well.
You can check that proof
here.