Convolution of densities and distributions

In summary, the conversation discusses the definitions and distinctions between density, measure, and distribution in the context of probability theory and convolution. The conversation also touches on the use of these terms in different situations and their relationships to each other. The conclusion is that these terms can be used interchangeably in certain cases, but have specific definitions and implications in others.
  • #1
Kreizhn
743
1
Hello everyone,

I have a quick theoretical question regarding probability. If you answer, I would appreciate it if you would be as precise as possible about terminology.

Here is the problem: I'm working on some physics problems that do probability in abstract spaces and the author freely moves between calling some poorly defined function f a density, a measure, and a distribution. From my knowledge of measure theory, these are all very different things, though are all inter-related.

In particular, the author talks about having two objects, say [itex] x_1 [/itex] sampled from [itex] f_1 [/itex] and [itex] x_2 [/itex] sampled from [itex] f_2 [/itex]. He wants to calculate the joint sampling distribution (?) of [itex] x = x_1 \cdot x_2 [/itex] which is defined via convolution

[tex] (f_1 \star f_2)(x) = \int f_1(x \cdot x_2^{-1}) f_2(x_2) d x_2 [/tex]

Where I hope it's clear that we're using multiplicative notation rather than the classical additive notation.

What it comes down to is that I'm trying to figure out what the author really means when talking about f. Is f a density or a distribution?

http://en.wikipedia.org/wiki/Convolution#Applications" says that densities obey convolution.

It would seem to me that this must be a density since if it were a distribution we would need to integrate over "sections" or at least non-zero measure sets.

Anyway, I really just want an answer as to whether densities, distributions, or both obey convolution. Thanks.
 
Last edited by a moderator:
Mathematics news on Phys.org
  • #2
Convolution makes sense for both densities and distribution functions. If [tex] X_1, X_2 [/tex] have densities [tex] f_1, f_2 [/tex] the density of [tex] X_1 + X_2 [/tex]
is the convolution of [tex] f_1, f_2[/tex]. The same is true for the distribution functions.
 
  • #3
Hey statdad, I really appreciate the reply.

Perhaps you could answer something else then that's been bugging me.

Theoretically, a distribution is the measure of the preimage of a random variable right? That is, let [itex] (\Omega, \Sigma, \mu) [/itex] is a probability space with real valued measure [itex] \mu [/itex] and let [itex] (\mathbb R, \nu) [/itex] be real space under the Lebesgue measure. X is a random variable if [itex] X: \Omega \to \mathbb R [/itex] is measurable. Then the distribution on X is [itex] f:\mathcal L_\nu(\mathbb R) \to \mathbb R [/itex] given by [itex] f(S) = \mu(X^{-1}(S)) [/itex] where [itex] \mathcal L_\nu(\mathbb R) [/itex] is the set of Lebesgue measurable real sets.

In this case, how can we define a convolution on distributions? Namely, if [itex] f_1, f_2 [/itex] are distributions, then

[tex] (f_1\star f_2)(x) = \int_{\Omega} f_1(y) f_2(x-y) d\mu(y) [/tex]
(where I hope that you'll forgive the fact that I've reverted back to additive notation: it just seems easier for this discussion).

However, we're integrating over singletons in the [itex] \sigma-[/itex]algebra and hence the convolution will be zero. Does this make sense? Am I missing something? Should this instead be

[tex] (f_1 \star f_2)(x) = \int_{\Sigma} f_1(Y) f_2(X\setminus_Y) d \mu(Y) [/tex]
 
  • #4
you seem to be using [tex] f_1, f_2 [/tex] to represent measureable functions rather than the measures generated by the distributions. If so, there isn't anything out of sorts with your notation - it is simply the convolution of two functions.

It is possible to define the convolution of the associated measures. Look at the discussion that begins near the bottom of page 1 of the pdf at this link:

http://www.galaxy.gmu.edu/stats/syllabi/it971/Lecture11.pdf
 
Last edited by a moderator:
  • #5
Sounds good. But I wonder if this makes sense: it is taken almost verbatim from a paper I'm reading

Let G denote the compact Lie group U(D) [unitary matrices of dimension D] with elements [itex] g \in G [/itex]. Let [itex] \mathcal F [/itex] denote the set of probability measures over the group. Suppose [itex] g_1 [/itex] and [itex] g_2 [/itex] are drawn at random from the measures [itex] f_1 \in \mathcal F[/itex] and [itex] f_2 \in \mathcal F [/itex] respectively. The composed element [itex] g = g_1 \cdot g_2 [/itex], where [itex] \cdot [/itex] denotes the standard group multiplication, is then distributed according to

[tex] f(g) = (f_1 \star f_2)(g) = \int d\mu(h) f_1(gh^{-1}) f_2(h) [/tex]

where [itex] \star [/itex] denotes the convolution product and [itex] d\mu(g) [/itex] denotes the Haar measure on G.

What is f here? It doesn't seem to me like it is actually a measure; at best a distribution? The author then also later notes that

[tex] \int d\mu(g) f^{\star m}(g) = 1, \qquad \forall m \in \mathbb N [/tex]
where

[tex]f \in \mathcal F, f^{\star m} = \underbrace{f\star \cdots \star f }_{m \text{ times }} [/tex].

Doesn't that imply it's a density? The author refers to f in this context as a measure, a distribution and a density, so I'm very confused.

The paper is http://link.aps.org/doi/10.1103/PhysRevA.72.060302"
 
Last edited by a moderator:
  • #6
Distribution density and measure can all mean the same thing. A density is something that you integrate to find the probability that you're inside of a given region. But a measure, as long as the measure of the whole space is 1, is the same thing. You can integrate it over a region to find the probability you are in that region as long as you interpret the measure as a density. A probability distribution sometimes refers to the cumulative distribution, but if it's being used interchangeably with density then it's probably not.
 
  • #7
Why would you integrate a measure to find the probability of being in a region? Let A be our region: if [itex] \mu [/itex] is the probability measure, the probability of being in A is just [itex] \mu(A) [/itex]. By definition, the domain of a measure is the associated sigma-algebra, while the domain of a density is the underlying set; how can these be interpreted as the same thing?

Also, a distribution must have a random variable associated to it, while a probability measure need not. As a matter of fact, distributions need not have an associated density. A distribution is another measure on the space, and if one is lucky enough that it is absolutely continuous and the domain/codomain of the random variable are both sigma-finite, we can define that density using the Radon-Nikodym theorem.

Am I missing something here? These words all have very precise mathematical meanings and while they can all be related to each other in special circumstances, they are certainly different things.
 
  • #8
"But a measure, as long as the measure of the whole space is 1, is the same thing. You can integrate it over a region to find the probability you are in that region as long as you interpret the measure as a density"

No, densities are not equivalent to the measure.

density -> distribution function <--> measure.

Consider a VERY simple case, the density

[tex] f(x) = 1, 0 \le x \le 1[/tex]

You integrate this to find probability:

[tex]
P(X <.5) = \int_0^{.5} f(x) \, dx = \int_0^{.5} 1 \,dx = .5
[/tex]


The associated distribution function is defined in terms of the anti-derivative of f.

[tex]
F(x) = \int_0^x f(t) \, dt = x \quad 0 \le x \le 1
[/tex]

and [itex] F(x) = 0 [/itex] for [itex] x < 0 [/itex], [itex] F(x) = 1 [/itex] for [itex] x > 1 [/itex].

Both the density and distribution function are point functions. The associated measure is a set function:

[tex]
\mu(\{t \mid t \le x\}) = F(x)
[/tex]

In general, if [itex] B [/itex] is a Borel subset of the real line, and [itex] F [/itex] is the distribution function of a random variable [itex] X [/itex], the associated probability measure is

[tex]
\mu(B) = \int_B \, dF
[/tex]

(Lebesque-Stieltjes integral). If [itex] F [/itex] happens to be absolutely continuous with respect to Lebesque measure there is a density.

"Also, a distribution must have a random variable associated to it, while a probability measure need not."
Semantics. It is true that a probability measure is simply a special type of measure (one that assigns mass 1 to the underlying space) but, if you are going through the both of referring to "probability measure", you typically have a random variable in mind.
 
  • #9
Thanks again statdad, you have proven to be most helpful.

I wonder though, does the situation not change in the event that we're not considering random variables on a real codomain?

I believe in theory it is possible to consider a two probability spaces [itex] S_i=(\Omega_i, \Sigma_i, \mu_i), i\in\left\{1,2\right\} [/itex]. In this case, I think we really do require an explicit random variable, represented by a measurable function [itex] X:S_1 \to S_2 [/itex]. In this case the distribution is not a point-function because there isn't necessarily a well-defined order on [itex] S_2 [/itex] to make it a cumulative function. So the probability distribution on X, denoted [itex] P_X [/itex] is given by [itex] P_X(B) = P(X^{-1}(B)) [/itex] which is also a measure.

density -> distribution function <--> measure.

I'm not so certain about the later dual implication, even in the real case. Doesn't one need certain conditions for a Lebesgue-Stieltjes measure to exist? More precisely, I think one requires that a distribution be bounded, non-decreasing, left continuous, and vanishes toward negative infinity to guarantee the existence of the LS measure (and hence the LS integral).
 

1. What is a convolution of densities and distributions?

A convolution of densities and distributions is a mathematical operation that combines two probability distributions to produce a new distribution. It is often used to model the behavior of two random variables that are related to each other.

2. How is a convolution calculated?

The convolution of two densities or distributions is calculated by taking the integral of the product of the two functions over all possible values. This can be represented mathematically as:

(f * g)(x) = ∫f(y)g(x-y)dy

where f and g are the two probability distributions, x is the value of the new distribution, and y is the variable of integration.

3. What is the significance of convolution in statistics and data analysis?

Convolution is an important tool in statistics and data analysis as it allows us to combine two distributions and understand the relationship between them. It is commonly used in signal processing, image processing, and machine learning to analyze data and make predictions.

4. Can the convolution of two distributions be visualized?

Yes, the convolution of two distributions can be visualized using a convolution plot, which shows the resulting distribution as a smooth curve. It can also be represented as a histogram, where the values of the new distribution are calculated from the overlapping points of the two original distributions.

5. What are some real-world applications of convolution of densities and distributions?

Convolution has many practical applications, such as in finance for modeling stock prices, in epidemiology for analyzing disease spread, and in physics for modeling particle collisions. It is also used in image and audio processing to enhance and filter signals.

Similar threads

Replies
3
Views
725
Replies
11
Views
1K
Replies
3
Views
209
  • General Math
Replies
20
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
5
Views
473
Replies
1
Views
831
Replies
3
Views
931
  • Set Theory, Logic, Probability, Statistics
Replies
4
Views
914
  • Set Theory, Logic, Probability, Statistics
Replies
12
Views
3K
  • General Math
Replies
6
Views
2K
Back
Top