- 22,169
- 3,328
Math_QED said:We now see: X~B(9,1/2)
How do you know this?
Math_QED said:We now see: X~B(9,1/2)
micromass said:How do you know this?
Math_QED said:We only need 9 guesses to know the result. As you said earlier, 9 out of 10 is impossible. Is it correct?
micromass said:I don't see at all how you take into account that I know 5 are pepsi and 5 are coca cola. In your formulation, it is entirely possibly that I say 8 are pepsi.
micromass said:How did you find these formulas?
Math_QED said:I supposed that the person who has to judge knows there are 5 pepsi's and 5 cola's.
micromass said:I don't think you did. In either case, your answer is incorrect.
PeroK said:They all reduce to variations on ##\sum \frac{1}{n(n+1)}##. For example:
##r = 3##
##\sum_{n = k}^{\infty} \frac{F_{n}}{n} = \frac{1}{k} + \frac{1}{k+1} \frac{k-2}{k} + \frac{1}{k+2} \frac{k-2}{k} \frac{k-1}{k+1} + \frac{1}{k+3} \frac{k-2}{k} \frac{k-1}{k+1} \frac{k}{k+2} \dots##
## = \frac{1}{k} + \frac{1}{k+1} \frac{k-2}{k} + (k-1)(k-2)[\frac{1}{k(k+1)(k+2)} + \frac{1}{(k+1)(k+2)(k+3)} \dots]##
## = \frac{1}{k} + \frac{1}{k+1} \frac{k-2}{k} + (k-1)(k-2)[\frac{1}{2k(k+1)}]## (Summed using partial fractions)
## = \frac{1}{2}##
micromass said:Is it possible to derive a closed form for the ##F_n##?
micromass said:4. Unstable particles are emitted from a source and decay at a distance xx, a real number that has an exponential distribution with characteristic length λ\lambda. Decay events can be observed only if they occur in a window extending from x=1x=1 to x=20x=20. We observe 66 decays at locations {2,5,12,13,13,16}\{2,5,12,13,13,16\}. What is λ\lambda?
Ygggdrasil said:Seems like the approach to take here would bemaximum likelihood estimation.
Do you want us to work this out by hand, or would it be sufficient to post code? (Or does knowing this much disqualify me from answering)
Ok.micromass said:Posting code and the answer would be enough.
P = @(x,lambda) exppdf(x,lambda)/(expcdf(20,lambda)-expcdf(1,lambda));
data = [2 5 12 13 13 16];
est = mle(data,'pdf',P,'start',mean(data),'lower',0)
thephystudent said:For #9, can you confirm whether all 26 letters plus space are allowed in principle in the new language/in the random text generation?
And, if random, do yo mean 'pulled from a uniform distribution' or 'I just hit my keyboard'. If it is not random, can we assume the original text was in English?
Is there a letter-by letter correspondance, word-by-word correspondance or something more delocalized like a 'fourier transform' or so?
micromass said:No, you cannot assume that the "random text" comes from a uniform distribution (which I think is very clearly not the case if you look at the data). And you cannot assume the original text was English.
PeroK said:##E(n) = \frac{(k-1)(r-1)}{r-2}##
Another try on 9).micromass said:That's an interesting analysis. I'll be waiting for more input before I spill the beans on this one.
Charles Link said:I would like to respond to #2. I did a simple statistical test. There are many tests that could be used, but I decided to look at the first head in a sequence and check the number of times it is also followed by a head. Assign a "1" to a head and a 0 to a tail. The upper one, I counted 17 first heads, and 14 of them were followed by a head. For a binomial distribution, ## \mu=Np ## and ## \sigma^2=Npq ## This gives for N=17 and p=1/2 and q=1/2, the mean is 8.5 and ## \sigma=2.1 ##. 14 is nearly 3 ## \sigma ## from the mean: ## z=(14-8.5)/2.1 ##. For the second group with the same test, I counted N=29 such occurrences and 13 subsequent heads. For this case, the mean is 14.5 and ## \sigma=2.7 ## . On this second row, the sample resulted in differing from the mean by much less than 1 ## \sigma ##. Clearly the top row is the one invented by a human and the bottom row is from a coin toss experiment. ( Note: I would like to doublecheck my counting-it was a little hard to read-I think my tallies of 17 and 14, and 29 and 13 were nearly correct). I see @PeroK made a similar "qualitative" observation in post 11.
Correction: I've made a mistake and had used a wrong denominator in one line. (#words instead of #letters) So the picture is a different one concerning the occurrences of letters. However, my next step would be attempts to decode the text, now that I have some clues for setting up a bijection of the alphabet. Since this is forbidden, my last hypothesis isfresh_42 said:Another try on 9)...
Hopefully he gives me credit for it already, but if not, a similar test could be done observing the two coin flips after a head first appears. In that case, a simple observation is that there are often two heads following it. I didn't compute it numerically yet, but it looks as if that might give about 4 ## \sigma ## or 5 ## \sigma ## from the mean for the top sequence, and even 3 ## \sigma ## is very unlikely to happen by chance...editing...this second test of the two subsequent coin tosses (if I computed it correctly) gave just over ## 2 \sigma ## from the mean for the first set (and not ## 4 \sigma ## or ## 5 \sigma ##), but it still is an indicator that it was likely man-made.PeroK said:Yes, in the first sequence there seems to be a correlation between the outcome of each toss and the previous one (it's more likely to be the same). For once I was more interested in the psychology than the maths. The normal tendency, I believe, is to do the opposite: try to even things up by favouring a Head afer a Tail and vice versa. If sequence 1 is human generated, then whoever did it appears to have known this and over compensated by favouring another Head after a Head and another Tail after a Tail. Or, perhaps, it was @micromass doing this deliberately!
Charles Link said:Hopefully he gives me credit for it already, but if not, a similar test could be done observing the two coin flips after a head first appears. In that case, a simple observation is that there are often two heads following it. I didn't compute it numerically yet, but it looks as if that might give about 4 ## \sigma ## or 5 ## \sigma ## from the mean for the top sequence, and even 3 ## \sigma ## is very unlikely to happen by chance.
fresh_42 said:@micromass Thank you. I now have a nice little mini Enigma for Caesars, random Letter encryption, word and letter counting, decoding attempts ... guess I'll have to add the RSA feature to finish it.
I don't know whether this is good or bad news for my little playground. As far as I know you the chances are high you will put some noise on the channel and I'm not sure whether I want to deal with error correcting mechanisms.micromass said:I might make a cryptography challenge later![]()

fresh_42 said:I'm not sure whether I want to deal with error correcting mechanisms.![]()
I have a more definitive statistical test for #2. The item that separates the first sequence from the second is the number of changes of state that take place (from the previous flip). In 100 coin tosses, there are ## N=99 ## possible changes of state. It's a simple binomial test with ## \mu=Np ## and ## \sigma^2=Npq ##. (with ##p=q=1/2 ##). This gives ## \mu=49.5 ## and ##\sigma=5.0 ##. For the first sequence I counted k=33 changes of state. This gives a ## z=(49.5-33)/5.0 ## that is slightly greater than 3. (result was more than ## 3 \sigma ## from the mean.) For the second sequence I still need to tally it, but I want to post this before someone else beats me to it. ...editing... for the second sequence I counted k=50 (approximately) transitions. (my eyes aren't real steady so the count may be off by one or two.) This second case is well within ## 1 \sigma ## of the mean and is the kind of result to be expected from a random coin toss.micromass said:I'll let you guys continue to think of this a bit more. I'll give the answer in a few days time.