The second kind of stochastic process we will consider is a process X(t) where the time t is a real-valued deterministic variable and at each time t, the sample space S of the random variable X(t) is a (finite or infinite) discrete set. In such a case, a state X of the entire process is described as a sequence of jump times () and a sequence of values of X (elements of the common discrete state space)
. These describe a "path" of the system through S given by
for 0<t<
,
for
<t<
, etc... Because X moves by a sequence of jumps, these processes are often called "jump" processes.
Thus, to describe the stochastic process, we need, for each x in S, a continuous random variable that describes how long it will be before the process (starting at X=x) will jump, and a set of discrete random variables
that describe the probabilities of jumping from x to other states when the jump occurs. We will let
be the probability that the state starting at x will jump to y (when it jumps), so
We also define the function to be the probability that the process starting at X=x has already jumped at time t, i.e.,
All this is defined so that
(this is an independence assumption: where we jump to is independent of when we jump.
Instead of writing , we will begin to write
(...).
We will make the usual stationarity and Markov assumptions as follows: We assume first that
It will be convenient to define the quantity to be the probability that X(t)=y, given that X(0)=0, i.e.,
Then the Markov assumption is
The three assumptions A1, A2, A3 will enable us to derive a (possibly infinite) system of differential equations for the probability functions for x and y in S. First, the stationarity assumption A1 tells us that
This means that the probability of having to wait at least one minute more before the process jumps is independent of how long we have waited already, i.e., we can take any time to be t=0 and the rules governing the process will not change. (From a group-theoretic point of view this means that the process is invariant under time translation). Using the formula for conditional probability (), the fact that the event
is a subset of the event
, we translate this equation into
This implies that the function 1- is a multiplicative function, i.e., it must be an exponential function. It is decreasing since
represents the probability that something happens before time t, and
=0, so we must have that 1-
=
for some positive number
. In other words,
so the two ways of computing agree.
The next consequence of our stationarity and Morkov assumptions are the Chapman-Kolmorgorov equations in this setting:
They say that to calculate the probability of going from state x to state y in time t+s, one can sum up all the ways of going from x to y via any z at the intermediate time t. This is an obvious consistency equation, but it will prove very useful later on.
Now we can begin the calculations that will produce our system of differential equations fod . The first step is to express
based on the time of the first jump from x. The result we intend to prove is the following:
As intimidating as this formula looks, it can be understood in a fairly intuitive way. We take it a term at a time:
To account for all of the ways of going from x to y in time t, we need to add up these probabilities over all possible times s of the first jump (this gives an integral with respect to ds in the limit as ds goes to 0), and over all possible results z different from x of the first jump. This reasoning yields the second term of the formula.
We can make a change of variables in the integral term of the formula: let s be replaced by t-s. Note that this will transform ds into -ds, but it will also reverse the limits of integration. The result of this change of variables is (after taking out factors from the sum and integral that depend neither upon z nor upon s) :
Now is a continuous function of t because it is defined as the product of a continuous function with a constant plus a constant times the integral of a bounded function of t (the integrand is bounded on any bounded t interval). But once
is continuous, we can differentiate both sides of the last equation with respect to t and show that
is also differentiable.
We take this derivative (and take note of the fact that when we differentiate the first factor we just get a constant times what we started with) using the fundamental theorem of calculus to arrive at:
A special case of this differential equation occurs when t=0. Note if we know that X(0)=x, then it must be true that . This observation enables us to write:
One last bit of notation! Let . So
and
if
. (Note that this implies that
so the sum of the as y ranges over all of S is 0. the constants
are called the infinitesimal parameters of the stochastic process.
Given this notation, we can rewrite our equation for given above as follows:
for all t>0. This system of differential equations is called the backward equation of the process.
Another system of differential equations we can derive is the forward equation of the process. To do this, we differentiate both sides of the Chapman Kolmogorov equation with respect to s and get
If we set s=0 in this equation and use the definition of , we get the forward equation:
Next time: We will use the forward equation to show the relationship between the exponential and Poisson distributions, and examine other examples of jump processes.