We have characterized recurrence in terms of return times. However, if a Markov chain is positive recurrent, what proportion of time does it spend in each state?
Continue reading “Stationary Distributions — Recap”Markov Chains Recurrence
We quickly recap the recurrence properties of Markov chains. The aim is to give a short review not a full rigorous treatment.
Continue reading “Markov Chains Recurrence”A discrete-time single-server queue
Many queues can be modeled as discrete-time Markov chains. Here, the state is the length of the queue, and fluctuations in arrivals and service give the random evolution from one time period to the next. Here is a diagram of a queue.

Figure: arrivals occur from the left and are stored in a buffer. Jobs or customers are represented as rectangles. A server is represented by a circle, which processes the jobs, and they then depart on the right-hand side.
Continue reading “A discrete-time single-server queue”Markov chains and matrices — A Quick Summary
Probabilities for a Markov chain can be expressed in terms of the probability vector and the transition matrix
.
For example, for and a function
,
So essentially, multiplication to the left gives probabilities and multiplication to the right gives expectations.
Notation explanation. If the above is unclear, some notation may need explaining. For , we multiply
copies of the matrix
together.1 Interpreting
as a row vector,
is a row vector, and
is the
-th component of that row vector.
Similarly, we can think of the function as a column vector
. Then
is a column vector under matrix multiplication, and
denotes its
-th component.
Throughout these notes, and
.
1 We use for transpose, not
.
Discrete-time Markov chains — A Quick Look
Markov chains are the primary stochastic process used in queueing theory. The main elements of a Markov chain are its state and the independent randomness used to generate future states. This creates a very clean and natural form of dependence for which many properties are known. We will cover a few important properties of Markov chains, including recurrence, transience, and the existence of stationary distributions.
Continue reading “Discrete-time Markov chains — A Quick Look”Zero-Order Stochastic Optimization: Keifer-Wolfowitz
We want to optimize the expected value of some random function. This is the problem we solved with Stochastic Gradient Descent. However, we assume that we no longer have access to unbiased estimate of the gradient. We only can obtain estimates of the function itself. In this case we can apply the Kiefer-Wolfowitz procedure.
The idea here is to replace the random gradient estimate used in stochastic gradient descent with a finite difference. If the increments used for these finite differences are sufficiently small, then over time convergence can be achieved. The approximation error for the finite difference has some impact on the rate of convergence.
Continue reading “Zero-Order Stochastic Optimization: Keifer-Wolfowitz”
The Law of Large Numbers and Central Limit Theorem
Let’s explain why the normal distribution is so important.
(This is a section in the notes here.)
Continue reading “The Law of Large Numbers and Central Limit Theorem”
Continuous Probability Distributions
We consider distributions that have a continuous range of values. Discrete probability distributions where defined by a probability mass function. Analogously continuous probability distributions are defined by a probability density function.
(This is a section in the notes here.)
Discrete Probability Distributions
There are some probability distributions that occur frequently. This is because they either have a particularly natural or simple construction. Or they arise as the limit of some simpler distribution. Here we cover
- Bernoulli random variables
- Binomial distribution
- Geometric distribution
- Poisson distribution.
(This is a section in the notes here.)
Random Variables and Expectation
Often we are interested in the magnitude of an outcome as well as its probability. E.g. in a gambling game amount you win or loss is as important as the probability each outcome.
(This is a section in the notes here.)