Dirac Delta Function: Explanation & Usage

dreamLord
Messages
203
Reaction score
7
I know this probably belongs in one of the math sections, but I did not quite know where to put it, so I put it in here since I am studying Electrodynamics from Griffiths, and in the first chapter he talks about Dirac Delta function.

From what I've gathered, Dirac Delta function is 0 for x\neq0, and ∞ for x = 0.

Now he assumes any function f(x), and says that the product f(x)*\delta(x) = 0 for x\neq0. Fine, got that.

Now he goes on to say that the above statement can also be written as f(0)*\delta(x) = 0. My question is - we could also have written it as f(29.5)*\delta(x) = 0 for x\neq0, right? So then why did we choose f(0)?
 
Physics news on Phys.org
First of all it is very important to understand that \delta is not a function but a distribution. It is defined as a linear form on an appropriate space of functions, e.g., the infinitely many times differentiable functions with compact support or rapidly falling functions (Schwartz space). It is defined as
\int_{\mathbb{R}} \mathrm{d} x f(x) \delta(x)=f(0).
This is not 0.

Sometimes you can simplify equations by the formal setting f(x) \delta(x)=f(0) \delta(x). Strictly speaking that's not correct, because you cannot integrate the \delta distribution over the test function which is constant, because this function does not belong to the test-function space, where the \delta distribution is defined.
 
I'm afraid your first few lines were completely lost on me! Is there any way you can dumb it down a bit?

Also, the equation that you wrote ; is this the definition of the Dirac Delta function? Or the fact that it is 0 when x is not zero, and infinity when x is 0. Which one defines it? Or are they the same thing.
 
Griffths focuses on f(0)δ(x) because f(0) is the only value of f that matters with the dirac delta. So basically f(0)δ(x) will behave the same way as f(x)δ(x).

This can be best understood under an integral sign, which is the only place the dirac delta function is precisely defined. You have (edit: you can take these two properties are the definition, but the exact mathematical definition is a bit more complicated):

\int \delta (x) dx = 1 and \int f(x) \delta (x) dx = f(0)

So you can write \int f(0) \delta (x) dx = f(0) \int \delta(x)dx = f(0)*1 = f(0) which is the same as for f(x)δ(x)

On the other hand, \int f(1) \delta(x) dx = f(1)\int \delta(x)dx = f(1), which is not correct.
 
Last edited:
The vanhees71 definition is right. The property "δ=0 for x ≠ 0, δ=∞ fo x = 0" is just an intuitive description of sharply peaked function, which is valid picture of δ only in some situations. For example, it is valid for charge density distribution of point-like charged particle. However, when solving for the Green function of the Schroedinger equation, such description of δ is incorrect, while the integral property above is valid.
 
I don't get the impression that the OP is comfortable talking about distributions or Green's functions. If he were, I doubt he would be having trouble with the dirac delta function.
 
dreamLord said:
I'm afraid your first few lines were completely lost on me! Is there any way you can dumb it down a bit?

Also, the equation that you wrote ; is this the definition of the Dirac Delta function? Or the fact that it is 0 when x is not zero, and infinity when x is 0. Which one defines it? Or are they the same thing.
The point is that many introductory physics books confuse their readers with unprecise definitions of what a distribution is. Griffiths seems to be another example. I don't know his E&M book very well besides from discussions here in the forum.

Objects like the Dirac \delta are socalled distributions. They are defined as mappings from a function space (containing a certain set of functions, called test functions) to the (real or complex) numbers. They can only be defined in a manner that makes sense under an integral, where they are multiplied with a test function, and for the Dirac \delta distribution this definition reads
\int_{\mathbb{R}} \mathrm{d} x \delta(x) f(x)=f(0).
It's the value of the test function at the argument 0.

It is quite obvious that \delta(x) cannot be a function in the usual sense, because you won't find any function with the above given property. However you can define the \delta distribution as kind of limit, the socalled weak limit. The idea is to define functions which are sharply peaked around 0 with the integral normalized to 1. The most simple example is the "box function",
\delta_{\epsilon}(x)=\begin{cases}<br /> 1/(2 \epsilon) &amp; \text{for} \quad x \in (-\epsilon,\epsilon) \\<br /> 0 &amp; \text{elsewhere}.<br /> \end{cases}
The test function should have "nice properties" to make things convenient. They should still form a vector space of functions, i.e., with two functions also their sum and the product with a constant should belong to this function space. A very convenient choice is Schwartz's space of rapidly falling smooth functions, i.e., they are arbitrarily many times differentiable and fall off at infinity faster than any polynomial.

Now we check the integral
I_{\epsilon}=\int_{\mathbb{R}} \mathrm{d} x \delta_{\epsilon}(x) f(x) = \frac{1}{2 \epsilon} \int_{-\epsilon}^{\epsilon} \mathrm{d} x f(x).
Now according to the mean-value problem for integrals over continuous functions, there is a value \xi \in [-\epsilon,\epsilon] such that
I_{\epsilon}=f(\xi).
Now, since f is continuous and we let \epsilon \rightarrow 0^+, you get
\lim_{\epsilon \rightarrow 0^+} I_{\epsilon}=f(0).
This means that in the sense of a weak limit you may write
\lim_{\epsilon \rightarrow 0^+} \delta_{\epsilon}(x)=\delta(x).
"Weak limit" means that you have to follow the above given procedure: You first have to take an integral with a test function and then take the limit. This is the crucial point.

If you now look what happens in this example, Griffiths sloppy definition makes some sense, but one has to keep in mind the proper meaning in the above given sense. Obviously our functions \delta_{\epsilon} are concentrated around 0 and become the larger the smaller \epsilon gets, when taking the limit \epsilon \rightarrow 0^+. At the same time the interval, where our function is different from 0 shrinks, and the construction is such that the total area under the graph (which here is a rectangle) stays constant 1 for all \epsilon. In this sense you may charcterize the \delta distribution as cited by Griffiths. To avoid confusion, however, it's mandatory to learn about the proper definition of distributions (also known as "generalized functions").

The best book for the physicist I know of is

M. J. Lighthill, Introduction to Fourier Analysis and Generalised Functions, Cambridge University Press (1959)

That it treats the distributions together with Fourier series and Fourier integrals is no disadvantage since this you'll need anyway when studying electrodynamics.
 
dreamLord said:
I know this probably belongs in one of the math sections, but I did not quite know where to put it,
Topology & analysis is the right place for it. I'm moving it there. Edit: I also changed "Diract" to "Dirac" in the title.
 
Last edited:
Yes Griffiths explanation is horrible, here is the idea without technicalities.

In finite calculus we define the delta function so that

$$a_0=\sum_{k=-\infty}^\infty \delta_k a_k$$

That a handy thing to do, it let's us write function evaluation as a sum.
We would like to do the same thing in infinitesimal calculus

$$\mathop{f}(0)=\int_{-\infty}^\infty \! \mathop{\delta}(x) \, \mathop{f}(x) \,\mathop{dx}$$
we ignore that the delta function does not exist as a function

now we adopt as equality f=g if
$$\int_{-\infty}^\infty \! (\mathop{f}(x) - \mathop{g}(x)) \,\mathop{dx}=0$$

in this sense
$$\mathop{\delta}(x) \, \mathop{f}(x)=\mathop{\delta}(x) \, \mathop{f}(0)$$
since clearly
$$\int_{-\infty}^\infty \! (\mathop{\delta}(x) \, \mathop{f}(x) - \mathop{\delta}(x) \, \mathop{f}(0)) \,\mathop{dx}=\int_{-\infty}^\infty \! ( \mathop{\delta}(x) \, ( \mathop{f}(x) - \mathop{f}(0))) \,\mathop{dx}=( \mathop{f}(0) -\mathop{f}(0) )=0$$

For some purposes we probably want to adopt as equality f=g if
$$\int_{a}^b \! (\mathop{f}(x) - \mathop{g}(x)) \,\mathop{dx}=0$$
for all a and b
 
Last edited:
  • #10
Things are becoming a little clearer now, though I am still fairly lost. Thank you for the amazing posts, vanhees, DimReg, Jano and lurflurf. I will need to read this thread a couple more times before I am ready to frame my doubts regarding your posts.
 
  • #11
dreamLord said:
Now he goes on to say that the above statement can also be written as f(0)*\delta(x) = 0. My question is - we could also have written it as f(29.5)*\delta(x) = 0 for x\neq0, right? So then why did we choose f(0)?

Do you know about the Riemann–Stieltjes integral?
By convention the spike is at x=0. Since δ(x) purpose is to evaluate f(x) near x=0 it does not care what f does away from zero much like

$$\lim_{x \rightarrow 0} \mathop{f}(x)$$
 
  • #12
No lurflurf, I do not know what that integral is.

By the way, an immediate question regarding your post (#9) ; how did you proceed in the second last step? That is :
∫(δ(x)(f(x)−f(0)))dx=(f(0)−f(0))=0

Thanks for telling me the purpose of the delta function - I did not understand why Griffiths brought it up in the first place!
 
  • #13
dreamLord said:
No lurflurf, I do not know what that integral is.

By the way, an immediate question regarding your post (#9) ; how did you proceed in the second last step? That is :
∫(δ(x)(f(x)−f(0)))dx=(f(0)−f(0))=0

Thanks for telling me the purpose of the delta function - I did not understand why Griffiths brought it up in the first place!

I showed the algebraic steps required in my first reply. Basically, f(0) is a constant, and integrals are linear, so:

\int(\delta(x)(f(x) - f(0)))dx = \int \delta(x) f(x)dx - \int \delta(x) f(0) dx = \int \delta(x) f(x) dx - f(0) \int \delta(x) dx = f(0) - f(0)

Where in the last step I used ∫f(x)δ(x)dx = f(0) for the first term and ∫δ(x)dx = 1 for the second term
 
Last edited:
  • #14
One thing that I think should be mentioned is that when ##\delta## is defined as a function that takes test functions to numbers, the definition can be written as ##\delta(f)=f(0)## for all test functions f. The notation ##\delta(f)## is far more natural than ##\int \delta(x)f(x)dx##. The reason that the latter is used must be that distributions were invented to make sense of expressions like ##\int \delta(x)f(x)dx##, which were already used in non-rigorous calculations.

So ##\int\delta(x)f(x)dx## isn't an integral of the product of a distribution and a function. It's just a notation that means ##\delta(f)##.

For each real number x, define ##\delta_x## by ##\delta_x(f)=f(x)## for all test functions f. Define the notation ##\int f(x)\delta(x-y)dx## to mean ##\delta_y(f)##. This ensures that ##\int f(x)\delta(x-y)dx=f(y)##.
 
  • #15
Thanks DimReg, I understand the step now.

Fredrik ; so does that mean that if I take f(x) = 2x - 5, then δ(f) = f(0) = -5 ?
Also, in your last 2 lines, why did you change your definition from δ(f) = f(0) to δ(f) = f(x)?

By the way, thanks for moving the thread to the correct section and also for fixing the typo!
 
  • #16
vanhees71 said:
Now, since f is continuous and we let \epsilon \rightarrow 0^+, you get
\lim_{\epsilon \rightarrow 0^+} I_{\epsilon}=f(0).
This means that in the sense of a weak limit you may write
\lim_{\epsilon \rightarrow 0^+} \delta_{\epsilon}(x)=\delta(x).
"Weak limit" means that you have to follow the above given procedure: You first have to take an integral with a test function and then take the limit. This is the crucial point.

You lost me in this specific paragraph. Why is epsilon approaching 0 from the + side? And if it is, how does the next equation follow?
 
  • #17
dreamLord said:
Fredrik ; so does that mean that if I take f(x) = 2x - 5, then δ(f) = f(0) = -5 ?
Yes.

dreamLord said:
Also, in your last 2 lines, why did you change your definition from δ(f) = f(0) to δ(f) = f(x)?
I didn't, I defined infinitely many new distributions, one for each real number. Only one of them (##\delta_0##) is equal to ##\delta##.
 
  • #18
So it is also true that δ(f) = f(1) = -3 ? If I take f(x) = 2x - 5.
 
  • #19
dreamLord said:
So it is also true that δ(f) = f(1) = -3 ? If I take f(x) = 2x - 5.
No, by my definitions ##\delta(f)=\delta_0(f)=f(0)=-5##, but ##\delta_1(f)=f(1)=-3##.

I don't know if anyone else uses this notation by the way. I just think it's a good way to make sense of expressions of the form ##\int f(x)\delta(x-y)dx## where y is a real number.
 
  • #20
I have never seen Fredrik's notation, and I can't really make any sense of it. The Dirac delta is never equal to anything besides 0 or infinity. In fact I often use the identity
\delta(f(x))=\sum_{\{\tilde{x}|f(\tilde{x})=0\}}\frac{\delta(x-\tilde{x})}{\left | \frac{df}{dx}|_\tilde{x}\right|}

If you're having trouble reading that, it just says to find the Dirac delta where the argument is a function, find all the zeros of the function, then form the sum of dirac deltas, one located at each zero, and each one devided by the absolute value of the function's derivative at that zero.

That's actually a rigorous statement that follows from the most common definition of the dirac delta function:
<br /> \delta(x):=\lim_{\alpha\rightarrow\infty}\sqrt{\frac{\alpha}{\pi}}e^{-\alpha x^2}<br />

This definition works better than the limit of rectangular functions since you can find the derivative with this one.
 
Last edited:
  • #21
I can't quite understand what delta-not and delta-one are, Fredrik (apologies, I can't use LaTex currently). Can you explain what they stand for?

Jolb ; why do we need to find the zeroes of the function? I thought the delta function was valid for all x?

I have never encountered such a vague and confusing topic in maths so far - which probably means I haven't done much, but either way, I am thoroughly confused. I'm not even sure I know why we need the delta function.
 
  • #22
The reason you need to find the zeros of the function in the argument of the Dirac delta is because the Dirac delta only "fires" when its argument is zero. Whenever the Dirac delta's argument is nonzero, the Dirac delta is equal to zero, and does nothing interesting. When its argument is zero, the Dirac delta does interesting things.

To explain this and the OP in a dumbed-down way, there's a great mnemonic to help with this. Sometimes people call the Dirac delta the "sampling function." If you have any function f(x) and you want to somehow pull out its value at a point x', you can get it by "sampling" it with the Dirac delta:

f(x') = ∫δ(x-x') f(x) dx
 
  • #23
By argument equaling zero, you mean the function that is multiplied with it - like f(x), should be 0 right? If that is the case, then why do we have expressions like the one in post #2 by vanhees? How are they relevant? Under the integral, we don't have f(x) = 0, which it ought to be for the Dirac function to be 'interesting' as you put it.
 
  • #24
dreamLord said:
I'm not even sure I know why we need the delta function.
Most physics books at the undergrad level will thoroughly butcher the definition and unfortunately the rigorous formulation requires some advanced mathematics (distribution theory). For now, can you at least see the physical motivations for it? Recall Griffiths' motivation, which is the apparent vanishing divergence of the Coulomb field at all points in space even when there is a localized point charge which should technically contribute to the divergence via Gauss's law.
 
  • #25
Yes, I understood how the divergence was vanishing everywhere except at r = 0. Does that mean that the divergence of Electric Field is a Dirac Delta function?

Also, Wannabe, aren't you an undergrad ? How are you so goddamn knowledgeable!
 
  • #26
dreamLord said:
By argument equaling zero, you mean the function that is multiplied with it - like f(x), should be 0 right? If that is the case, then why do we have expressions like the one in post #2 by vanhees? How are they relevant? Under the integral, we don't have f(x) = 0, which it ought to be for the Dirac function to be 'interesting' as you put it.

No... the "argument" of a function is what you stick into it, not what you multiply it with.

So if we have the expression
f(p)
then f is the function and p is its argument.

So δ(f(x)) is the Dirac delta with the argument f(x). This is completely different from things like δ(x)f(x). The latter is what appears in the sampling equation f(x') = ∫δ(x-x') f(x) dx.
 
  • #27
Sorry, I thought δ(f(x)) = ∫f(x)δ(x)dx - which I see makes no sense.
 
  • #28
dreamLord said:
Yes, I understood how the divergence was vanishing everywhere except at r = 0. Does that mean that the divergence of Electric Field is a Dirac Delta function?
Yes, at least for the case I mentioned above. Can you see why intuitively? Recall that ##\nabla \cdot E = \frac{\rho}{\epsilon_0}##. Now for the Coulomb field the source is just a single point charge ##Q## at say ##r = 0##. How in the world are we going to represent the charge density of this thing-it's localized to a point! Well what we want to do is somehow find a mathematical quantity that can represent nothing in space at every point except one-and at this one point there will be a sudden spike to represent the presence of that point charge.

This might help as well: https://www.physicsforums.com/showthread.php?t=695129&highlight=current+wire

For now I would personally just focus on how the dirac delta function is used in electromagnetism (because it's used quite a lot) and what it tries to model physically in terms of charge and current distributions. There's no need to make things any more complicated at this level by going into all the rigorous mathematics behind this.

dreamLord said:
Also, Wannabe, aren't you an undergrad ?
Yessir.
 
  • #29
Yes Wannabe, I understood that part. Post #2 in that link was helpful. I will re-read this thread and Griffiths once again tomorrow, and make what I can of it. Thank you guys.
 
  • #30
I have never encountered such a vague and confusing topic in maths so far - which probably means I haven't done much, but either way, I am thoroughly confused. I'm not even sure I know why we need the delta function.

In E&M we often use something called a Green's function to solve tricky PDES. This is a very high level technique that you don't learn about until graduate level E&M. The delta function is central to this technique, and Griffiths is introducing the delta function now to give you some exposure to delta function and hopefully spare you some pain when you take Jackson E&M.

I'm going outline the technique below to illustrate why delta functions are important. When learning advanced math I find it instructive to study physical problems where the math is applicable. If its above you don't worry about it.

A common problem in electrostatics is to solve for the potential \phi(\vec x) (and thus the Electric Field) of a given charge distribution \rho(\vec x).

This problem amounts to solving Poisson's equation:
\nabla^2 \phi(\vec x) = \frac{ \rho(\vec x)}{\epsilon} subject to certain boundary conditions.

One trick to solve this equation is to use Green Functions.

We start by solving a modified equation:
\nabla^2 G(\vec x, \vec x&#039;) = \delta(\vec x - \vec x&#039;)

Here G is a Green's function and \vec x&#039; is a dummy variable.

Now we can use G to solve the original equation by noting:
\int d^3x&#039; \nabla^2 G(\vec x, \vec x&#039;) \frac{ \rho(\vec x&#039;)}{\epsilon} = \int d^3x&#039; \delta(\vec x - \vec x&#039;)\frac{ \rho(\vec x&#039;)}{\epsilon}

Using the properties of the delta function this becomes:
\int d^3x&#039; \nabla^2 G(\vec x, \vec x&#039;) \frac{ \rho(\vec x&#039;)}{\epsilon} =\frac{ \rho(\vec x)}{\epsilon}
Using Poisson's equation we can then equate:
\int d^3x&#039; \nabla^2 G(\vec x, \vec x&#039;) \frac{ \rho(\vec x&#039;)}{\epsilon} = \nabla^2 \phi(\vec x).
And finally we pull \nabla^2 outside the integral because it does not depend on x'
\nabla^2 \int d^3x&#039; G(\vec x, \vec x&#039;) \frac{ \rho(\vec x&#039;)}{\epsilon} = \nabla^2 \phi(\vec x).
Or
\int d^3x&#039; G(\vec x, \vec x&#039;) \frac{ \rho(\vec x&#039;)}{\epsilon} = \phi(\vec x).


What we have done is split up the task into two steps.
1) Solve \nabla^2 G(\vec x, \vec x&#039;) = \delta(\vec x - \vec x&#039;) for G.
2) After solving for G we integrate \int d^3x&#039; G(\vec x, \vec x&#039;) \frac{ \rho(\vec x&#039;)}{\epsilon} giving us \phi(\vec x).

This works because its often easier to solve the PDE for G than it is to solve for \phi(\vec x).

Also once we solve the PDE for G for a given geometry and boundary conditions we can use the same G to solve for \phi(\vec x) for a bunch of different charge distributions. This saves us a lot of work, because integrating a function is many times easier than solving a PDE.
 
  • #31
dreamLord said:
I can't quite understand what delta-not and delta-one are, Fredrik (apologies, I can't use LaTex currently). Can you explain what they stand for?
I defined them in post #14. Not sure what more I can say, unless you explain what issues you're having with the definition.

dreamLord said:
I have never encountered such a vague and confusing topic in maths so far
Nothing is butchered quite as badly by mediocre physics text as the Dirac delta and tensors. Not sure which is worse. I remember being a lot more frustrated about the tensors actually, so maybe that's worse.
 
  • #32
That usage of the Dirac function is quite beautiful wolfman, I like it very much. Thanks!
 
  • #33
Jolb said:
I have never seen Fredrik's notation, and I can't really make any sense of it. The Dirac delta is never equal to anything besides 0 or infinity.
The second sentence here suggests that you're thinking of ##\delta## as a function that takes numbers as input. It's not defined that way in any rigorous treatments. It's defined either as a distribution (a function that takes "nice enough" functions to real numbers) or as a measure (a function that take subsets of ℝ to non-negative extended real numbers that we can think of as the "sizes" of those sets).
 
  • #34
Fredrik said:
I remember being a lot more frustrated about the tensors actually, so maybe that's worse.
Lol you should see how tensors are defined in Melvin Schwartz' book on Electromagnetism. You will cry. It was some definition in terms of rotations that I had never even seen before, and I first skimmed this book after having done Wald for some time so I was like what the hell is this?!
 
  • #35
Fredrik said:
The second sentence here suggests that you're thinking of ##\delta## as a function that takes numbers as input. It's not defined that way in any rigorous treatments. It's defined either as a distribution (a function that takes "nice enough" functions to real numbers) or as a measure (a function that take subsets of ℝ to non-negative extended real numbers that we can think of as the "sizes" of those sets).

You're right, but this technicality couldn't be any more irrelevant. I gave a definition for the dirac as the limit of functions, and in almost any expression with a Dirac delta distribution, you could approximate the expression to arbitrary accuracy using a function by picking a big enough \alpha in the definition I gave. The subtleties of what you're talking about won't matter at all when the OP is using Griffiths; in fact Griffiths himself says you can treat it like a function in his book.

On the other hand, your definitions are very confusing because you use the notation δ(f) to mean something completely different from δ(f(x)), which is highly nonstandard (at least for people at the level of Griffiths. Maybe it makes sense if you've taken distribution theory, but I'm sure the OP hasn't.) That's what caused the OP to get confused:
Sorry, I thought δ(f(x)) = ∫f(x)δ(x)dx - which I see makes no sense.
 
  • #36
Jolb said:
You're right, but this technicality couldn't be any more irrelevant.
It's relevant to people who want to understand the posts on page 1 of this thread better.

Jolb said:
On the other hand, your definitions are very confusing because you use the notation δ(f) to mean something completely different from δ(f(x)), which is highly nonstandard (at least for people at the level of Griffiths. Maybe it makes sense if you've taken distribution theory, but I'm sure the OP hasn't.) That's what caused the OP to get confused:
I don't see how it can be that confusing when the definition is that simple and stated that clearly. I guess it has something to do with the fact that people at that level are for some reason used to thinking of f(x) as a function. Of course, f is the function, and f(x) is its value at x. My notation should make sense to people who understand that distinction. It's certainly not "highly non-standard".

Since ##\delta##, as I defined it, takes functions to numbers, it doesn't make sense to try to use the number f(x) as input.
 
Last edited:
  • #37
Fredrik said:
It's relevant to people who want to understand the posts on page 1 of this thread better.
And the posts on page 1 you're talking about obviously did more to confuse the OP than to help him.

I guess it has something to do with the fact that people at that level are for some reason used to thinking of f(x) as a function. Of course, f is the function, and f(x) is its value at x. My notation should make sense to people who understand that distinction. It's certainly not "highly non-standard".
Well your notation δ(f) doesn't appear in Griffiths EM or any other undergraduate physics text I've ever used. However, you commonly see δ(f(x)) [I'm not sure if it's in Griffiths EM but it's certainly in Griffiths QM] and you commonly see f(x) abbreviated as f. And, honestly, stressing the distinction between f and f(x) reeks of a mathematician. Most physicists would say f(x) is "a function," and even though they know it's really the value of the function f at the point x, they do identify f(x) with f since confusing the two things is almost impossible for a physicist. Confusion would only arise in more advanced contexts. Similarly, physicists often call δ(x) the "Dirac delta function." This less rigorous but much less unnecessarily fussy approach is the one Griffiths takes, along with most other undergrad physics book authors.

I don't see how it can be that confusing when the definition is that simple and stated that clearly.
Let me try to retrace your definitions to demonstrate where they lead to confusion.

Fredrik said:
For each real number x, define ##\delta_x## by ##\delta_x(f)=f(x)## for all test functions f. Define the notation ##\int f(x)\delta(x-y)dx## to mean ##\delta_y(f)##. This ensures that ##\int f(x)\delta(x-y)dx=f(y)##.

And later you say.
Fredrik said:
Only one of them (##\delta_0##) is equal to ##\delta##.
It looks to me like you're defining the symbol δ, but nowhere does it say it's the Dirac delta... But as best I can figure out you're saying that δ0 is the Dirac delta. (?)

If that were the case then the definition you give for the Dirac delta is:
##\int f(x)\delta(x)dx =: \delta(f)##
So you have a δ(x) in your definition of δ(f)... what is the definition of δ(x)? You give a definition in terms of sampling I guess, but the point is that δ(x) is actually the Dirac delta (at least in the way Dirac and most physicists use it), not δ(f).

-----
Side note: Also, if I'm guessing right that you say δ0 is the Dirac delta, this seems to imply that only δ(x) ever has anything to do with the Dirac delta, never δ(x-x0)
 
Last edited:
  • #38
Jolb said:
Well your notation δ(f) doesn't appear in Griffiths EM or any other undergraduate physics text I've ever used.
That's because they don't cover the definition of ##\delta##, so this is not an argument for using a different notation in the definition.

Jolb said:
It looks to me like you're defining the symbol δ, but nowhere does it say it's the Dirac delta... But as best I can figure out you're saying that δ0 is the Dirac delta. (?)
I defined ##\delta## at the start of that post:
Fredrik said:
One thing that I think should be mentioned is that when ##\delta## is defined as a function that takes test functions to numbers, the definition can be written as ##\delta(f)=f(0)## for all test functions f.
I think it's 100% clear from the context that the delta I'm referring to is the one that's the topic of this thread. This distribution is what we need to make sense of the notation ##\int f(x)\delta(x)dx##.

You quoted something else at the end of the post. I defined a distribution ##\delta_x## for each real number x. By the definition of ##\delta##, we have ##\delta(f)=f(0)## for all test functions f. By the definition of ##\delta_0##, we have ##\delta_0(f)=f(0)## for all test functions f. So for all test functions f, we have ##\delta(f)=f(0)=\delta_0(f)##. This is why I said that ##\delta=\delta_0##. They have the same domain and the same value at each point in the domain.

These ##\delta_x## distributions (the set ##\{\delta_x|x\in\mathbb R\})## are what we need to make sense of the notation ##\int f(x)\delta(x-y)dx## where y≠0.

Jolb said:
If that were the case then the definition you give for the Dirac delta is:
##\int f(x)\delta(x)dx =: \delta(f)##
Nooooo...That equality defines the notation ##\int f(x)\delta(x)dx## by saying that it's equal to the previously defined ##\delta(f)##. The expression ##\delta(x)## is left undefined, just like the ##\partial y## in the definition of ##\partial f/\partial y##. The entire expression ##\int f(x)\delta(x)dx## is defined to be equal to ##\delta(f)##.

Recall that the notation ##\int f(x)\delta(x)dx## was already in use by physicists before anyone knew how to make sense of it. The point of the definition I'm talking about is that it can be used to assign a meaning to the notation ##\int f(x)\delta(x)dx##.

Jolb said:
So you have a δ(x) in your definition of δ(f)... what is the definition of δ(x)? You give a definition in terms of sampling I guess, but the point is that δ(x) is actually the Dirac delta (at least in the way Dirac and most physicists use it), not δ(f).
Some of this is based on a misunderstanding of what I was defining, but you have a point about the symbol ##\delta## being used in two places. This should however not be a problem once you understand that we're defining the entire expression ##\int f(x)\delta(x)dx##, not the component parts of it. I suppose I could have used a different notation, like ##\Delta## or ##D_\delta##, for the delta distribution, but I doubt that it would have made much of a difference.

Jolb said:
Side note: Also, if I'm guessing right that you say δ0 is the Dirac delta, this seems to imply that only δ(x) ever has anything to do with the Dirac delta, never δ(x-x0)
I'm not sure what the standard terminology is, but it makes sense to me to call each ##\delta_x## with ##x\in\mathbb R## a delta distribution, and to call ##\delta_0## (also denoted by ##\delta##) the delta distribution.
 
Last edited:
  • #39
Fredrik said:
That's because they don't cover the definition of ##\delta##, so this is not an argument for using a different notation in the definition.
The definition I gave is in many common undergraduate textbooks. Take for example Shankar's Principles of Quantum Mechanics, 2nd Edition, page 61, equation 1.10.19. I'll just restate it here: \delta(x):=\lim_{\alpha\rightarrow\infty}\sqrt{\frac{\alpha}{\pi}}e^{-\alpha x^2}
I think it's 100% clear from the context that the delta I'm referring to is the one that's the topic of this thread. This distribution is what we need to make sense of the notation ##\int f(x)\delta(x)dx##.
Absolutely not. I think it's 100% clear in this thread and in Griffiths that δ(x) is the topic, not your δ(f), which you use to define a big expression containing δ(x), and you leave δ(x) undefined.
These ##\delta_x## distributions (the set ##\{\delta_x|x\in\mathbb R\})## are what we need to make sense of the notation ##\int f(x)\delta(x-y)dx## where y≠0.
As I said before, if you use the definition I gave/Shankar gives, you can get arbitrarily close approximations to any expression (as far as Griffiths EM is concerned) with a δ-distribution by replacing it with the function in my/Shankar's definition, provided you pick a large enough \alpha. I think this makes perfect sense as a way to interpret δ(x) and the expression ##\int f(x)\delta(x-y)dx##. We don't need fussy and confusing distribution theory.

The rest of your issues are just correcting my confusion of what you had written. Seriously, it is confusing. This is what happens when you go overboard with math, even when you don't accidentally reuse a symbol.

Recall that the notation ##\int f(x)\delta(x)dx## was already in use by physicists before anyone knew how to make sense of it.
Well those physicists (Dirac, Fermi, Feynman, etc.) were able to use it correctly without the full glory of distribution theory. So I think they knew adequately how to make sense of it. That's what's important here, not the formalism.

I'm not sure what the standard terminology is, but it makes sense to me to call each ##\delta_x## with ##x\in\mathbb R## a delta distribution, and to call ##\delta_0## (also denoted by ##\delta##) the delta distribution.

To reiterate: δ(f) is not the Dirac delta. δ(x) is a Dirac delta. Two different things. Everything of the form δ(x-x0) is also a Dirac delta.

This is the way physicists including Griffiths talk about the Dirac delta.
 
Last edited:
  • #40
Jolb said:
The definition I gave is in many common undergraduate textbooks. Take for example Shankar's Principles of Quantum Mechanics, 2nd Edition, page 61, equation 1.10.19.
I don't consider that a definition. It's a description of how to think about a concept that's left undefined.

Jolb said:
Absolutely not. I think it's 100% clear in this post and in Griffiths that δ(x) is the topic, not your δ(f), which is defined in terms of δ(x), which you leave undefined.
You are 100% wrong about this. My sentence said "...when δ is defined as a function that takes test functions to numbers..." It's impossible to interpret that as anything but the Dirac delta, and it's impossible to think that I'm talking about some other version of it than the one that takes test functions to numbers, because I stated explicitly that that's the version I'm talking about.

Jolb said:
We don't need fussy and confusing distribution theory.
If you mean that you don't need it for this type of problem, you're probably right. I haven't really thought about it, but it's not needed for most practical issues. If you mean that it's never needed, your view is pretty naive. If you mean that it wasn't needed in this thread when I posted it, you are wrong, because I only posted to clarify a few points abouts distributions that had already been made in this thread.

Jolb said:
Seriously, it is confusing. This is what happens when you go overboard with math, even when you don't accidentally reuse a symbol.
Seriously, a definition is not going overboard.

Jolb said:
To reiterate: δ(f) is not the Dirac delta. δ(x) is a Dirac delta. Two different things. Everything of the form δ(x-x0) is also a Dirac delta.
I can see how one might prefer a non-rigorous presentation of the idea, but it doesn't make sense to argue that this is the right way to do it, while the rigorous definition is the wrong way.
 
  • #41
Your method is not wrong, it is just more confusing and goes overboard with fussy mathematics in spite of there being a simpler definition which is perfectly adequate for any calculation in Griffiths EM. What you wrote would be a lot less confusing if you make the correction you suggested: keep a separate symbol for δ(f) so it doesn't get confused with δ(f(x))--if you had done that I probably wouldn't have complained. But even if you had done that, it would have been unnecessarily rigorous for a Griffiths EM student.

Indeed, if we go beyond Griffiths EM into, say, peer-reviewed Math journals, then in some cases it will turn out that my method is wrong and yours is right. But you have to go to quite a high level to find where my method breaks down--such a high level is not something a Griffiths EM student needs to be worrying about.

For the record, there should be an implicit "in the context of Griffiths EM" along with everything I said up to here: certainly there are places where distribution theory is needed, and it is an important aspect of mathematics. But my definition works fine for Griffiths EM and pretty much anything else you do as a physics undergrad (and most physics grad classes too).

Edit: Maybe you don't need to go to "quite a high level" just to find pathological examples. But these pathological examples certainly won't show up in the level of Griffiths EM, and they probably won't pose any real problems until you reach a high level.
 
Last edited:
  • #42
Jolb said:
The definition I gave is in many common undergraduate textbooks. Take for example Shankar's Principles of Quantum Mechanics, 2nd Edition, page 61, equation 1.10.19. I'll just restate it here: \delta(x):=\lim_{\alpha\rightarrow\infty}\sqrt{\frac{\alpha}{\pi}}e^{-\alpha x^2}

I need to point out that that definition is only valid as a weak limit (i.e. in the weak topology of functionals). If you don't explain what a weak limit is, it's a very bad definition.
 
  • #43
The Riemann–Stieltjes integral is a good way to handle these simple matters without the theory of distributions. The Riemann–Stieltjes integral is not in most first year calculus books, so you might need a second year book. We consider sums like

$$\int \! \mathop{f}(x) \, \mathop{dg}=\lim \sum \mathop{f}(x^*_k)(\mathop{g}(x_{k-a}-\mathop{g}(x_k))$$
where for the Riemann integral g(x)=x

Then

$$\int \! \delta (x) \, \mathop{f}(x) \, \mathop{d x}=\int \! \mathop{f}(x) \, \mathop{d u}=\mathop{f}(0)$$
where
u=0 x<0;1x>0
u'=δ

This all arises when we want to model something as concentrated. For example something that happens instantly or substance that is only at one point or surface instead of a region in space.
 
  • #44
lurflurf said:
The Riemann–Stieltjes integral is a good way to handle these simple matters without the theory of distributions. The Riemann–Stieltjes integral is not in most first year calculus books, so you might need a second year book. <snip>

The R-S integral is also good because there is a geometric interpretation. See
Gregory L. Bullock, A geometric interpretation of the Riemann-Stieltjes integral,
The American Mathematical Monthly 95 (1988), no. 5, 448–455.
 
  • #45
pwsnafu said:
I need to point out that that definition is only valid as a weak limit (i.e. in the weak topology of functionals). If you don't explain what a weak limit is, it's a very bad definition.

Well I didn't go into the formalism of the topology of functionals, because I don't think that would help for a Griffiths EM student. What I said a few times is that "you can get arbitrarily close approximations to any expression with a δ (in Griffiths) by substituting it with my definition, picking a large enough \alpha." This is actually the same thing Shankar says in the section I referenced. I think that in the context of Griffiths, that's perfectly unambiguous and well-defined and is equivalent to whatever fancy mathematical terminology you prefer.
 
Back
Top