1. Random
  2. 15. Markov Processes
  3. 1
  4. 2
  5. 3
  6. 4
  7. 5
  8. 6
  9. 7
  10. 8
  11. 9
  12. 10
  13. 11
  14. 12
  15. 13
  16. 14
  17. 15
  18. 16
  19. 17
  20. 18
  21. 19
  22. 20
  23. 21
  24. 22
  25. 23

23. Continuous-Time Branching Chains

Basic Theory

Introduction

Generically, suppose that we have a system of particles that can generate or split into other particles of the same type. Here are some typical examples:

We assume that the lifetime of each particle is exponentially distributed with parameter α(0,), and at the end of its life, is replaced by a random number of new particles that we will refer to as children of the original particle. The number of children N of a particle has probability density function f on N. The particles act independently, so in addition to being identically distributed, the lifetimes and the number of children are independent from particle to particle. Finally, we assume that f(1)=0, so that a particle cannot simply die and be replaced by a single new particle. Let μ and σ2 denote the mean and variance of the number of offspring of a single particle. So μ=E(N)=n=0nf(n),σ2=var(N)=n=0(nμ)2f(n) We assume that μ is finite and so σ2 makes sense. In our study of discrete-time Markov chains, we studied branching chains in terms of generational time. Here we want to study the model in real time.

Let Xt denote the number of particles at time t[0,). Then X={Xt:t[0,)} is a continuous-time Markov chain on N, known as a branching chain. The exponential parameter function λ and jump transition matrix Q are given by

  1. λ(x)=αx for xN
  2. Q(x,x+k1)=f(k) for xN+ and kN.
Details:

That X is a continuous-time Markov chain follows from the assumptions and the basic structure of continuous-time Markov chains. In turns out that the assumption that μ< implies that X is regular, so that τn as n, where τn is the time of the nth jump for nN+.

  1. Starting with x particles, the time of the first state change is the minimum of x independent variables, each exponentially distributed with parameter α. As we know, this minimum is also exponentially distributed with parameter αx.
  2. Starting in state xN+, the next state will be x+k1 for kN, if the particle dies and leaves k children in her place. This happens with probability f(k).

Of course 0 is an absorbing state, since this state means extinction with no particles. (Note that λ(0)=0 and so by default, Q(0,0)=1.) So with a branching chain, there are essentially two types of behavior: population extinction or population explosion.

For the branching chain X={Xt:t[0,)} one of the following events occurs with probability 1:

  1. Extinction: Xt=0 for some t[0,) and hence Xs=0 for all st.
  2. Explosion: Xt as t.
Details:

If f(0)>0 then all states lead to the absorbing state 0 and hence the set of positive staties N+ is transient. With probability 1, the jump chain Y visits a transient state only finitely many times, so with probability 1 either Yn=0 for some nN or Yn as n. If f(0)=0 then Yn is strictly increasing in n, since f(1)=0 by assumption. Hence with probability 1, Yn as n.

Without the assumption that μ<, explosion can actually occur in finite time. On the other hand, the assumption that f(1)=0 is for convenience. Without this assumption, X would still be a continuous-time Markov chain, but as discussed in the Introduction, the exponential parameter function would be λ(x)=αf(1)x for xN and the jump transition matrix would be Q(x,x+k1)=f(k)1f(1),xN+,k{0,2,3,}

Because all particles act identically and independently, the branching chain starting with xN+ particles is essentially x independent copies of the branching chain starting with 1 particle. In many ways, this is the fundamental insight into branching chains, and in particular, means that we can often condition on X(0)=1.

Generator and Transition Matrices

As usual, we will let P={Pt:t[0,)} denote the semigroup of transition matrices of X, so that Pt(x,y)=P(Xt=yX=x) for (x,y)N2. Similarly, G denotes the infinitesimal generator matrix of X.

The infinitesimal generator G is given by G(x,x)=αx,xNG(x,x+k1)=αxf(k),xN+,kN

Details:

This follows immediately from the exponential parameter function and the jump transition matrix in [1].

The Kolmogorov backward equation is ddtPt(x,y)=αxPt(x,x)+αxk=0f(k)Pt(x+k1,y),(x,y)N2

Details:

The backward equation is ddtPt=GPt, so the result follows from [3].

Unlike some of our other continuous-time models, the jump chain Y governed by Q is not the discrete-time version of the model. That is, Y is not a discrete-time branching chain, since in discrete time, the index n represents the nth generation, whereas here it represent the nth time that a particle reproduces. However, there are lots of discrete-time branching chain embedded in the continuous-time chain.

Fix t(0,) and define Zt={Xnt:nN}. Then Zt is a discrete-time branching chain with offspring probability density function ft given by ft(x)=Pt(1,x) for xN.

Details:

In general, we know that sampling a (homogeneous) continuous-time Markov chain at multiples of a fixed t(0,), results in a (homogeneous) discrete-time Markov chain. For Zt to be a branching chain, we just need to note that Pt(x,y)=ftx(y),(x,y)N2 where ftx is the convolution power of ft of order x. This is a consequence of the fundamental fact that Xt given X0=x has the same distribution as the sum of x independent copies of Xt given X0=1. Recall that the PDF of a sum of independent variables is the convolution of the individual PDFs.

Probability Generating Functions

As in the discrete case, probability generating functions are an important analytic tool for continuous-time branching chains.

For t[0,) let Φt denote the probability generating function of Xt given X0=1 Φt(r)=E(rXtX0=1)=x=0rxPt(1,x) Let Ψ denote the probability generating function of N Ψ(r)=E(rN)=n=0rnf(n)The generating functions are defined (the series are absolutely convergent) at least for r(1,1].

The collection of generating functions Φ={Φt:t[0,)} gives the same information as the collection of probability density functions {Pt(1,):t[0,)}. With the fundamental insight that the branching process starting with one particle determines the branching process in general, Φ actually determines the transition semigroup P={Pt:t[0,)}.

For t[0,) and xN, the probability generating function of Xt given X0=x is Φtx: y=0ryPt(x,y)=[Φt(r)]x

Details:

Again, given X0=x, the number of particles Xt at time t has the same distribution as the sum of x independent copies of Xt given X0=1. Recall that the PGF of a sum of independent variables is the product of the PGFs of the variables.

Note that Φt is the generating function of the offspring distribution for the embedded discrete-time branching chain Zt={Xnt:nN} for t(0,). On the other hand, Ψ is the generating function of the offspring distribution for the continuous-time chain. So our main goal in this discussion is to see how Φ is built from Ψ. Because P is a semigroup under matrix multiplication, and because the particles act identically and independently, Φ is a semigroup under composition.

Φs+t=ΦsΦt for s,t[0,).

Details:

Using the semigroup property (the Chapman-Kolmogorov equations) and [7] we have Φs+t(r)=y=0ryPs+t(1,y)=y=0ryx=0Ps(1,x)Pt(x,y)=x=0Ps(1,x)y=0ryPt(x,y)=x=0Ps(1,x)[Φt(r)]x=Φs[Φt(r)]

Note also that Φ0(r)=E(rX0X0=1)=r for all rR. This also follows from the semigroup property: Φ0=Φ0Φ0. The fundamental relationship between the collection of generating functions Φ and the generating function Ψ is given in the following theorem:

The mapping tΦt satisfies the differential equation ddtΦt=α(ΨΦtΦt)

Details:

Using the Kolmogorov backward equation in [4] we have ddtΦt(r)=x=0rxddtPt(1,x)=x=0rxGPt(1,x) Using the generator in [2], GPt(1,x)=y=0G(1,y)Pt(y,x)=αPt(1,x)+k=0αf(k)Pt(k,x),xN Substituting and using [7] gives ddtΦt(r)=x=0rx[αPt(1,x)+k=0αf(k)Pt(k,x)]=αx=0rxPt(1,x)+αx=0k=0rxf(k)Pt(k,x)=αΦt(r)+αk=0f(k)x=0rxPt(k,x)=αΨt(r)+αk=0f(k)[Φt(r)]k=αΦt(r)+αΨ[Φt(r)]

This differential equation, along with the initial condition Φ0(r)=r for all rR determines the collection of generating functions Φ. In fact, an implicit solution for Φt(r) is given by the integral equation rΦt(r)1Ψ(u)udu=αt Another relationship is given in the following theorem. Here, Φt refers to the derivative of the generating function Φt with respect to its argument, of course (so r, not t).

For t[0,), Φt=ΨΦtΦtΨΦ0

Details:

From the semigroup property, we have Φt+s(r)=Φt[Φs(r)] for s,t[0,). Differentiating with respect to s and using the chain rule along with [9] gives ddsΦt+s(r)=Φt[Φs(r)]ddsΦs(r)=Φt[Φs(r)]α[Ψ(Φs(r))Φs(r)] Evaluating at s=0 and using the condition Φ0(r)=r we have ddtΦt(r)=Φt(r)α[Ψ(r)r] Using the [9] once again gives α[Ψ(Φt(r))Φt(r)]=Φt(r)α[Ψ(r)r] Solving for Φt(r) gives the result.

Moments

In this discussion, we wil study the mean and variance of the number of particles at time t[0,). Let mt=E(XtX0=1),vt=var(XtX0=1),t[0,) so that mt and vt are the mean and variance, starting with a single particle. As always with a branching process, it suffices to consider a single particle:

For t[0,) and xN,

  1. E(XtX0=x)=xmt
  2. var(XtX0=x)=xvt
Details:

Once again, the distribution of Xt given X0=x is the same as the distribution of the sum of x independent copies of Xt given X0=1. Recall that the mean of a sum of variables is the sum of the individual means, and the variance of the sum of independent variables is the sum of the individual variances.

Recall also that μ and σ2 are the the mean and variance of the number of offspring of a particle. Here is the connection between the means:

mt=eα(μ1)t for t[0,).

  1. If μ<1 then mt0 as t. This is extinction in the mean.
  2. If μ>1 then mt as t. This is explosion in the mean.
  3. If μ=1 then mt=1 for all t[0,). This is stability in the mean.
Details:

From the proof of [10], ddtΦt(r)=αΦt(r)[Ψ(r)r] Differentiating with respect to r, interchanging the order of integration on the left, and using the product rule on the right gives ddtΦt(r)=αΦt(r)[Ψ(r)r]+αΦt(r)[Ψ(r)1] Now let r=1 and recall that Φ(1)=1. We get ddtΦt(1)=αΦt(1)[Ψ(1)1] From the basic theory of probability generating functions, mt=Φt(1) and similarly, μ=Ψ(1). Hence we have ddtmt=α(μ1)mt Of course we have the initial condition m0=1.

This result is intuitively very appealing. As a function of time, the expected number of particles either grows or decays exponentially, depending on whether the expected number of offspring of a particle is greater or less than one. The connection between the variances is more complicated. We assume that σ2<.

If μ1 then vt=[σ2μ1+(μ1)][e2α(μ1)teα(μ1)t],t[0,) If μ=1 then vt=ασ2t.

  1. If μ<1 then vt0 as t
  2. If μ1 then vt as t
Details:

Probability generating functions are naturally connected to factorial moments, so it's best to work with these. Thus, let wt=E[Xt(Xt1)X0=1] for t[0,) and let δ=E[N(N1)]. These are the factorial moments of order 2. In the proof of the last theorem we showed that ddtΦt(r)=αΦt(r)[Ψ(r)r]+αΦt(r)[Ψ(r)1] Differentiating with respect to r again gives ddtΦt(r)=αΦt(r)[Ψ(r)r]+2αΦt(r)[Ψ(r)1]+αΦt(r)Ψ(r) Now substitute r=1. Recall that Φt(1)=wt, Φt(1)=mt=eα(μ1)t, Ψ(1)=δ, Ψ(1)=μ, and Ψ(1)=1. We get the differential equation ddtwt=2α(μ1)wt+αδeα(μ1)t with the initial condition w0=0.

Suppose that μ1. Then using standard methods for a linear, first order differential equations with constant coefficients and an exponential forcing function, the solution is wt=δμ1[e2α(μ1)eα(μ1)t] But δ=σ2+μ2μ, and similarly wt=vt+mt2mt with mt=eα(μ1)t. Substitution and some algebra then gives the result.

Suppose now that μ=1. Then also mt=1 for all t[0,) and so δ=σ2 and vt=wt. The differential equation above reduces simply to ddtvt=ασ2 with initial condition v0=0 so trivially vt=ασ2t. Finally, in the context of part (b), note that if μ=1 we must have σ2>0 since we have assumed that f(1)=0.

If μ<1 so that mt0 as t and we have extinction in the mean, then vt0 as t also. If μ>1 so that mt as t and we have explosion in the mean, then vt as t also. We would expect these results. On the other hand, if μ=1 so that mt=1 for all t[0,) and we have stability in the mean, then vt grows linearly in t. This gives some insight into what to expect next when we consider the probability of extinction.

The Probability of Extinction

As shown in [2], there are two types of behavior for a branching process, either population extinction or population explosion. In this discussion, we study the extinction probability, starting as usual with a single particle: q=P(Xt=0 for some t(0,)X0=1)=limtP(Xt=0X0=1) Need we say it? The extinction probability starting with an arbitrary number of particles is easy.

For xN, P(Xt=0 for some t(0,)X0=x)=limtP(Xt=0X0=x)=qx

Details:

Given X0=x, extinction has occurred by time t if and only if extinction has occurred by time t for each of the x independent branching chains formed from the descendents of the x initial particles.

We can easily relate extinction for the continuous-time branching chain X to extinction for any of the embedded discrete-time branching chains.

If extinction occurs for X then extinction occurs for Zt for every t(0,). Conversely if extinction occurs for Zt for some t(0,) then extinction occurs for Zt for every t(0,) and extinction occurs for X. Hence q is the minimum solution in (0,1] of the equation Φt(r)=r for every t(0,).

Details:

The statements about the extinction event follow immediately from the fact that 0 is absorbing, so that if Xt=0 for some t(0,) then Xs=0 for every s[t,). The result for the extinction probability q follows from the theory of discrete-time branching chains.

So whether or not extinction is certain depends on the critical parameter μ.

The extinction probability q and the mean of the offspring distribution μ are related as follows:

  1. If μ1 then q=1, so extinction is certain.
  2. If μ>1 then 0<q<1, so there is a positive probability of extinction and a positive probability of explosion.
Details:

These results follow from the corresponding results for discrete-time branching chains. Fix t(0,) and recall that mt is the mean of the offspring distribution for the discrete-time chain Zt={Xnt:nN}. From [12],

  1. If μ1 then mt1.
  2. If μ>1 then mt>1.

It would be nice to have an equation for q in terms of the offspring probability generating function Ψ. This is also easy

The probability of extinction q is the minimum solution in (0,1] of the equation Ψ(r)=r.

Details:

From [16], Φt(q)=1 for every t(0,). Substituting r=q in exercose [9], we have ddtΦt(q)=0 and hence Ψ(q)=q. As in the theory of discrete branching chains, the equation Ψ(r)=r has only the solution 1 in (0, 1] if μ=Ψ(1)1 or there are two solutions q(0,1) and 1 if μ>0. In both cases, q is the smaller solution.

Special Models

We now turn our attention to a number of special branching chains that are important in applications or lead to interesting insights. We will use the notation established above, so that α is the parameter of the exponential lifetime of a particle, Q is the transition matrix of the jump chain, G is the infinitesimal generator matrix, and Pt is the transition matrix at time t[0,). Similarly, mt=E(XtX0=x), vt=var(XtX0=x), and Φt are the mean, variance, and generating function of the number of particles at time t[0,), starting with a single particle. As always, be sure to try these exercises yourself before looking at the proofs and solutions.

The Pure Death Branching Chain

First we consider the branching chain in which each particle simply dies without offspring. Sadly for these particles, extinction is inevitable, but this case is still a good place to start because the analysis is simple and lead to explicit formulas. Thus, suppose that X={Xt:t[0,)} is a branching process with lifetime parameter α(0,) and offspring probability density function f with f(0)=1.

The transition matrix of the jump chain and the generator matrix are given by

  1. Q(0,0)=1 and Q(x,x1)=1 for xN+
  2. G(x,x)=αx for xN and G(x,x1)=αx for xN+

The time-varying functions are more interesting.

Let t[0,). Then

  1. mt=eαt
  2. vt=eαte2αt
  3. Φt(r)=1(1r)eαt for rR
  4. Given X0=x the distribution of Xt is binomial with trial parameter x and success parameter eαt. Pt(x,y)=(xy)eαty(1eαt)xy,xN,y{0,1,,x}
Details:

All of these results follow from the general methods above, with μ=σ=0 and Ψ(r)=1 for rR. But it's helpful to give direct proofs. Given X0=1, let τ be the time until the first transition, which is simply the lifetime of the particle. So τ has the exponential distribution with parameter α. For t[0,), Xt is an indicator random variable (taking just values 0 and 1) with P(Xt=1X0=1)=P(τ>tX0=1)=eαt Part (a), (b), and (c) are standard results for an indicator variable. For part (d), given X0=x, each of the x particles, independently, is still alive at time t with probability eαt. Hence the number of particles still alive has the binomial distribution with parameters x and eαt.

In particular, note that Pt(x,0)=(1eαt)x1 as t. that is, the probability of extinction by time t increases to 1 exponentially fast. Since we have an explicit formula for the transition matrices, we can find an explicit formula for the potential matrices as well. The result uses the beta function B.

For β(0,) the potential matrix Uβ is given by Uβ(x,y)=1α(xy)B(y+β/α,xy+1),xN,y{0,1,,x} For β=0, the potential matrix U is given by

  1. U(x,0)= for xN
  2. U(x,y)=1/αy for x,yN+ and xy.
Details:

Suppose that β>0 and that x,yN with xy. By definition Uβ(x,y)=0eβtPt(x,y)dt=0eβt(xy)eαty(1eαt)xydt Substitute u=eαt so that du=αeαtdt or equivalently dt=du/αu. After some algebra, the result is Uβ(x,y)=1α(xy)01uy+β/α1(1u)xydu By definition, the last integral is B(y+β/α,xy+1).

  1. For xN, U(x,0)=0(1eαt)xdd=
  2. For x,yN+ with xy, the derivation above and properties of the beta function give U(x,y)=1α(xy)B(y,xy+1)=1α(xy)(y1)!(xy)!x!=1αy

We could argue the results for the potential U directly. Recall that U(x,y) is the expected time spent in state y starting in state x. Since 0 is absorbing and all states lead to 0, U(x,0)= for xN. If x,yN+ and xy, then x leads to y with probability 1. Once in state y the time spent in y has an exponential distribution with parameter λ(y)=αy, and so the mean is 1/αy. Of course, when the chain leaves y, it never returns.

Recall that βUβ is a transition probability matrix for β>0, and in fact βUβ(x,) is the probability density function of XT given X0=x where T is independent of X has the exponential distribution with parameter β. For the next result, recall the ascending power notation a[k]=a(a+1)(a+k1),aR,kN

For β>0 and xN+, the function βUβ(x,) is the beta-binomial probability density function with parameters x, β/α, and 1. βUβ(x,y)=(xy)(β/α)[y]1[xy](1+β/α)[x],xN,y{0,1,x}

Details:

From [20] and properties of the beta function. βUβ(x,y)=βα(xy)B(y+β/α,xy+1),xN,y{0,1,,x} But from properties of the beta function, B(y+β/α,xy+1)=B(β/α,1)(β/α)[y]1[xy](1+β/α)[x]=αβ(β/α)[y]1[xy](1+β/α)[x] Substituting gives the result

The Yule Process

Next we consider the pure birth branching chain in which each particle, at the end of its life, is replaced by 2 new particles. Equivalently, we can think of particles that never die, but each particle gives birth to a new particle at a constant rate. This chain could serve as the model for an unconstrained nuclear reaction, and is known as the Yule process, named for George Yule. So specifically, let X={Xt:t[0,)} be the branching chain with exponential parameter α(0,) and offspring probability density function given by f(2)=1. Explosion is inevitable, starting with at least one particle, but other properties of the Yule process are interesting. in particular, there are fascinating parallels with the pure death branching chain in . Since 0 is an isolated, absorbing state, we will sometimes restrict our attention to positive states.

The transition matrix of the jump chain and the generator matrix are given by

  1. Q(0,0)=1 and Q(x,x+1)=1 for xN+
  2. G(x,x)=αx for xN and G(x,x+1)=αx for xN+

Since the Yule process is a pure birth process and the birth rate in state xN is αx, the process is also called the linear birth chain. As with the pure death process, we can give the distribution of Xt specifically.

Let t[0,). Then

  1. mt=eαt
  2. vt=e2αteαt
  3. Φt(r)=reαt1r+reαt for |r|<11eαt
  4. Given X0=x, Xt has the negative binomial distribution on N+ with stopping parameter x and success parameter eαt. Pt(x,y)=(y1x1)exαt(1eαt)yx,xN+,y{x,x+1,}
Details:

Proof from the general results: Parts (a) and (b) follow from the general moment results above, with μ=2 and σ2=0. For part (c), note that Ψ(r)=r2 for rR, so the integral equation for Φt is rΦt(r)1u2u=αt From partial fractions, 1u2u=1u11u, so the result follows by standard integration and algebra. We recognize Φt as the probability generating function of the geometric distribution on N+ with success parameter eαt, so for part (d) we use our standard argument. Given X0=xN+, Xt has the same distribution as the sum of x independent copies of Xt given X0=1, and so this is the distribution of the sum of x independent variables each with the geometric distribution on N+ with parameter eαt. But this is the negative binomial distribution on N+ with parameters x and eαt.

Direct proof: As usual, let τ0=0 and let τn denote the time of the nth transition (birth) for nN+. Given X0=1, the population is n at time τn1. So the random interval τnτn1 (the time until the next birth) has the exponential distribution with parameter αn and these intervals are independent as n varies. From a result in the section on the exponential distribution, it follows that τn=k=1n(τkτk1) has distribution function given by P(τntX0=1)=(1eαt)n,t[0,) Curiously, this is also the distribution function of the maximum of n independent variables, each with the exponential distribution with rate α. Hence P(XtnX0=1)=P(τn1tX0=1)=(1eαt)n1,nN+ and therefore P(Xt=nX0=1)=P(XtnX0=1)P(Xtn+1X0=1)=(1eαt)n1eαt,nN+ So given X0=1, Xt has the geometric distribution with parameter eαt. The other results then follow easily.

Recall that the negative binomial distribution with parameters kN+ and p(0,1) governs the trial number of the kth success in a sequence of Bernoulli Trials with success parameter p. So the occurrence of this distribution in the Yule process suggests such an interpretation. However this interpretation is not nearly as obvious as with the binomial distribution in the pure death branching chain. Next we give the potential matrices.

For β[0,) the potential matrix Uβ is given by Uβ(x,y)=1α(y1x1)B(x+β/α,yx+1),xN+,y{x,x+1,} If β>0, the function βUβ(x,) is the beta-negative binomial probability density function with parameters x, β/α, and 1: βUβ(x,y)=(y1x1)(β/α)[x]1[yx](1+β/α)[y],xN,y{x,x+1,}

Details:

The proof is very similar to the one in [23]. Suppose that β0 and that x,yN+ with yx. By definition Uβ(x,y)=0eβtPt(x,y)dt=0eβt(y1x1)eαtx(1eαt)yxdt Substitute u=eαt so that du=αeαtdt or equivalently dt=du/αu. After some algebra, the result is Uβ(x,y)=1α(y1x1)01ux+β/α1(1u)yxdu By definition, the last integral is B(x+β/α,yx+1).

If we think of the Yule process in terms of particles that never die, but each particle gives birth to a new particle at rate α, then we can study the age of the particles at a given time. As usual, we can start with a single, new particle at time 0. So to set up the notation, let X={Xt:t[0,)} be the Yule branching chain with birth rate α(0,), and assume that X0=1. Let τ0=0 and for nN+, let τn denote the time of the nth transition (birth).

For t[0,), let At denote the total age of the particles at time t. Then At=n=0Xt1(tτn),t[0,) The random process A={At:t[0,)} is the age process.

Details:

Note that there have been Xt1 births in the interval [0,t]. For n{0,1,,Xt1}, the age at time t of the particle born at time τn is tτn.

Here is another expression for the age process.

Again, let A={At:t[0,)} be the age process for the Yule chain starting with a single particle. Then At=0tXsds,t[0,)

Details:

Suppose that Xt=k+1 where kN, so that τkt<τk+1. Note that Xs=n for τn1s<τn and n{1,2,,k}, while Xs=k+1 for τkst. Hence 0tXsds=n=1kn(τnτn1)+(k+1)(tτk)=(k+1)tn=0kτn From [25], At=n=0k(tτn)=(k+1)tn=0kτn

With the last representation, we can easily find the expected total age at time t.

Again, let A={At:t[0,)} be the age process for the Yule chain starting with a single particle. Then E(At)=eαt1α,t[0,)

Details:

We can interchange the expected value and the integral by Fubini's theorem. So using [23], E(At)=E(0tXsds)=0tE(Xs)ds=0teαsds=eαt1α

The General Birth-Death Branching Chain

Next we consider the continuous-time branching chain in which each particle, at the end of its life, leaves either no children or two children. At each transition, the number of particles either increases by 1 or decreases by 1, and so such a branching chain is also a continuous-time birth-death chain. Specifically, let X={Xt:t[0,)} be a continuous-time branching chain with lifetime parameter α(0,) and offspring probability density function f given by f(0)=1p, f(2)=p, where p[0,1]. When p=0 we have the pure death chain in [18], and when p=1 we have the Yule process in [22]. We have already studied these, so the interesting case is when p(0,1) so that both extinction and explosion are possible.

The transition matrix of the jump chain and the generator matrix are given by

  1. Q(0,0)=1, and Q(x,x1)=1p, Q(x,x+1)=p for xN+
  2. G(x,x)=αx for xN, and G(x,x1)=α(1p)x, G(x,x+1)=αpx for xN+

As mentioned earlier, X is also a continuous-time birth-death chain on N, with 0 absorbing. In state xN+, the birth rate is αpx and the death rate is α(1p)x. The moment functions are given next.

For t[0,),

  1. mt=eα(2p1)t
  2. If p12, vt=[4p(1p)2p1+(2p1)][e2α(2p1)teα(2p1)t] If p=12, vt=4αp(1p)t.
Details:

These results follow from the general formulas above for mt and vt, since μ=2p and σ2=4p(1p).

The next result gives the generating function of the offspring distribution and the extinction probability.

For the birth-death branching chain,

  1. Ψ(r)=pr2+(1p) for rR.
  2. q=1 if 0<p12 and q=1pp if 12<p<1.
Details:
  1. By definition, Ψ(r)=(1p)r0+pr2.
  2. The equation Ψ(r)=r factors into (r1)[pr(1p)]=0. The extinction probability q is the smaller solution in (0,1].
Graphs of rΨ(r) and rr when p=13
Graphs
Graphs of rΨ(r) and rr when p=23
Graphs

For t[0,), the generating function Φt is given by Φt(r)=pr(1p)+(1p)(1r)eα(2p1)tpr(1p)+p(1r)eα(2p1)t,if p1/2Φt(r)=2r+(1r)αt2+(1r)αt,if p=12

Details:

The integral equation for Φt is rΦt(r)dupu2+(1p)u=αt The denominator in the integral factors into (u1)[pu(1p)]. If p12, use partial fractions, standard integration, and some algebra. If p=12 the factoring is 12(u1)2 and partial fractions is not necessary. Again, use standard integration and algebra.