A couple of things.
1.) I don't think this is a
B thread. A lot of things in probability are easy to state and not easy to solve. This is one of them. Problems of k length runs are not treated in any of the first course undergrad probability books that I'm aware of.
2.) rather than looking at the probability of having at least one 'success' after n tosses, let's look at the complement -- i.e. the probability of never having seen a run of ##k## tails after ##n## tosses.
Let me suggest partitioning the events as
##H##
##TH##
##TTH##
##\vdots##
##\underbrace{T...T}_{\text{k-1 times}}H##
where there are ##k## of these in total. Calculate the probabilities of each event as ##p_0, p_1, ..., p_{k-1}## respectively
It is assumed that ##n\geq k## and then try working with the below matrix.
##\mathbf A^r \mathbf u = \begin{bmatrix}
p_0 & p_1 & p_2 & \dots &p_{k-1}\\
1 & 0 & 0 & \dots & 0\\
0& 1& 0 & \dots & 0\\
\vdots & \vdots & \ddots & \vdots & \vdots\\
0 & 0 & 0 & 1 & 0
\end{bmatrix}^r \begin{bmatrix}
u_{k-1}\\
u_{k-2}\\
\vdots\\
u_{1}\\
u_{0}
\end{bmatrix}##
This chain is used to give you the probability of not having seen a run after a desired number of tosses. eg ##\mathbf e_1^T \mathbf A^r \mathbf u## with ##\mathbf e_1## as the first standard basis vector, and setting ##r = 1## gives probability of no run after ##k## tosses, ##r=2## gives probability of no run after ##k+1## tosses, and so on.
By inspection we see that
##
\begin{bmatrix}
u_{k-1}\\
u_{k-2}\\
\vdots\\
u_{1}\\
u_{0}
\end{bmatrix} =
\begin{bmatrix}
1\\
1\\
\vdots\\
1\\
1
\end{bmatrix}##
i.e. ##\mathbf u = \mathbf 1##
In words: the above can be interpreted as "last step analysis" and the way to interpret this is that after 0, 1, 2, ...,and even ##k-1## tosses, we haven't seen a run yet with probability 1. These are the initial conditions of the recurrence that are implied by how the problem is set up.
Then if we look at the total probability of no run after say ##k## tosses, it is because we must have seen an event from the above collection. I.e. we must have not yet had a run of ##k## tails at time ##k-j## and then an event of length ##j## from the above -- and we sum over ##j=1, 2,...,n##. If you take a
careful look at the lack of 'overlaps' in the events, you'll see they are mutually exclusive.
A renewal theory take:
Technically you don't need renewal theory to interpret the above argument though it helps clarify that we're looking at a defective renewal process and the first row of ##\mathbf A## gives the (defective) probability distribution of time until next renewal (where a renewal is defined as the event that a ##H## occurs and a run of k ##T##'s has not yet occurred). Note that this view point offers two different perspectives -- ##\mathbf e_1^T \big(\mathbf A^r \mathbf 1\big)## is the basic recurrence argument from above and we are in effect using ##\mathbf A## as an expectations operator. On the other hand if we use associativity and look at ##\big(\mathbf e_1^T \mathbf A^r\big) \mathbf 1## we can interpret this a queuing model where we start at the 'top' of the queue and after ##r## trials we use the ones vector to sum up everything that still exists in the queue (but since the process is defective a decent chunk will have 'fallen off the face of the earth' -- and this is precisely the probability of at least one run of k tails)
- - - - -
The above matrix represents a residual life chain from renewal theory. It is also similar to the companion matrix. It also should look somewhat fibonacci matrix-ish.
A nice way to have some fun and get the conclusion OP was asking about, is consider that
##p_0 + p_1 + p_2 + ... +p_{k-1}\lt 1##
(how do we know this? easy way: consider that if this was an infinite serries it would be the geometric distribution and sum to one)
but we may (perhaps numerically) solve for some ##\theta\gt 1## such that
##\theta p_0 + \theta^2 p_1 + \theta^3 p_2 + ... + \theta^k p_{k-1}= 1##
and make use of exponential tilting. See e.g. the last 2 pages of here, which shows (with a typo or two) how to apply exponential tilting to a fibonacci matrix:
http://galton.uchicago.edu/~lalley/Courses/Summer/Renewal2.pdf
- - - -
conclusion:
in the case of ##k=2## you'll see that ##\theta = -1 + \sqrt{5}##
This should look familiar in terms of Fibonacci-- i.e. you have the dominant root, but times two.
and in general, because you have a fair coin (homogenous probability between heads and tails) and are talking about a pattern that is a 'pure' run of ##k ## tails (homogenous pattern)
you'll see that
##\theta p_0 + \theta^2 p_1 + \theta^3 p_2 + ... + \theta^k p_{k-1}= 1##
is equivalent to
##\big(\frac{\theta}{2}\big) + \big(\frac{\theta}{2}\big)^2 + \big(\frac{\theta}{2}\big)^3 + ... + \big(\frac{\theta}{2}\big)^k= 1##
so if you solve for ##\big(\frac{\theta}{2}\big)## then yes, this is equivalent to the dominant root of a k-step fibonnaci.
The advantage of the above setup is it rapidly generalizes to arbitrary patterns and coins that are not necessarily fair (or even objects that aren't coins, say for looking at patterns in dice or random letters, etc.)
It is possible that there's a more direct combinatorial argument that works for this special case of a fair coin and a run of k tails, though I don't know it, and I have a strong preference for arguments that are probabilistic in nature as opposed to combinatorial manipulations.
- - - -
a more direct close: you can, with some care, also skip the 'proper' exponential tilting by setting ##\theta := 2## in which case your first row ##\mathbf A_{\text{tilted}}## becomes all ones and you recover the k step fibonnaci matrix exactly. Maybe this is what OP wants here, though you lose some nice probabilistic linkages and exponential tilting becomes just an algebraic hack. But yes, this would seem to tie directly in to
Chris Miller said:
The probability P of k consecutive tails occurring in n coin tosses is 1 - (1 / F) where F is element n+2 in the k-step Fibonacci series divided by 2n.
Though I think there is a typo and it should read ##1 -F##
Qualitatively ##\frac{1}{F}## can't be right since probabilities are bounded in 0 and 1, and ##\big(\frac{\text{fibonacci of n}}{2^n}\big)^{-1} =\frac{2^n}{\text{fibonacci of n}}## grows much larger than ##1## -- e.g. consider the roots for a length 2 fibonanci -- they both have magnitude of less than 2.