craigi said:
I'm talking about randomly sampling a value for an unbounded real number and by unbounded I mean that it can take any value. I'm not specifying an "a priori" probability density distribution but the important point is that it doesn't converge to zero as we approach +/- infinity.
Just for future reference: when you are trying to describe something where you don't know the "correct" words, you are just learning, it is important to be specific. i.e. as
what goes to infinity? As
what doesn't converge? You were vague there.
People trying to help you will be aware of many more possibilities than you are so you need to help narrow it down.
While you have not specified any particular probability density distribution you are specifying what kind you are interested in. i.e. It is a probability
density, it is continuous (because you wanted any real number value to be possible), and, it appears, that it is non-zero for all real number values.
So you have pdf: ##p(x): x\in (-\infty, \infty)## and $$P(a<x<b)=\int_a^b p(x)\;\text{d}x$$
That was pretty much what I was working from when I was trying to understand what you wanted.
I doubt that any such problem is covered in introductory statistics or stochastics processes texts, but if you can prove me wrong then you've found exactly what I'm looking for.
So noted - the main problem is trying to describe to me what you are thinking about without actually knowing the "right" words - this means there will be a certain about of cross-talk until we are on the same page.
Regarding your observations:
1.1) I understand your point but I don't think we can say the probability of sampling something that we have actually sampled is zero. If we have sampled it, there exists a chance however small that we could sample it again. It's infinitesmial, but non-zero.
This is simply incorrect - a probability density function can only tell you the probability of getting a value that is between two other values. This material is covered in introductory probability and stats courses.
Since you insisted on a pdf as above, then the probability that x=a (some specific real number) is given by: $$P(x=a)=\int_a^a p(x)\;\text{d}x=0$$
In practice, a real life measurement always has some uncertainty - i.e. there are a range of possible values that will give the same measurement (eg. from rounding off).
I know it seems logical that, if you got some value off a random sample, then there must be a non-zero probability of getting it again, but that is not how probability, or measurement, works.
There is a non-zero probability of getting
very close to that value - but the probability of getting
exactly that value is zero. You may get so close to the initial value that your equipment is unable to tell the difference.
This is where I started to think you may need some theory about how measurements happen rather than theories about how probability density functions may be set up or modeled. However - it is very unusual for a real measurement to be able to take on any real number value at all ... there is always some sort of restriction determined by the type of thing being measured.
1.2) Perhaps I should have specified that I'm talking about a situation where all values are possible and I'm just offering a symmetry argument, in that if one value has non-zero probability then all values have non-zero probability.
Still does not follow - but if you restrict yourself to those situations where the distribution s non-zero for all real numbers then the point is moot.
But you cannot deduce the type of distribution from a single measured value.
2) It just a sampling scheme to generate an infinite number of samples. I guess we can call it a unidirectional random walk in each direction.
A statistical sample typically has more than one value in it - known as a data point. You would typically want more than one sample to investigate how a specific random variable is distributed. Thus it is not clear what you mean by that statement ... that's OK: this is just me trying to teach you how to talk to statisticians.
Anyway - a random walk could be used to generate a single set of N>0 random numbers ... a single sample, size N ... as follows:
let ##X = \{x_n\}## where each ##x_n=\sum_{i=1}^M s_i## is the result of m>0 random steps from the origin. The ith step size ##s\in \{s_i\}\sim \text{H}(-1,1)## is distributed according to a top hat defined as: $$X\sim \text{H}(a,b) \implies p(x)=\begin{cases} \frac{1}{b-a} &: a<x<b\\ 0 &: \text{else}
\end{cases}$$ ... this would be a stochastic process.
3) I'm just saying that any 2 numbers randomly sampled from all real numbers will differ by an infinite amount.
Why would that be so?
If you have two real numbers, then they must differ by a finite amount.
craigi said:
I think I've found the answer.
Measure Theory seems to be what I'm looking for.
Fair enough. This post basically pre-empts all of above.
It looks to me like you are shaky on the foundations of probability and statistics though - you are certainly writing as if you are not used to the topics - but that may be due to uncertainty about what you are asking about too.
I wrote all that so you may have better luck with questions here in the future.
Have fun ;)