Simulation and the Monte Carlo Method
()
About this ebook
This accessible new edition explores the major topics in Monte Carlo simulation that have arisen over the past 30 years and presents a sound foundation for problem solving
Simulation and the Monte Carlo Method, Third Edition reflects the latest developments in the field and presents a fully updated and comprehensive account of the state-of-the-art theory, methods and applications that have emerged in Monte Carlo simulation since the publication of the classic First Edition over more than a quarter of a century ago. While maintaining its accessible and intuitive approach, this revised edition features a wealth of up-to-date information that facilitates a deeper understanding of problem solving across a wide array of subject areas, such as engineering, statistics, computer science, mathematics, and the physical and life sciences. The book begins with a modernized introduction that addresses the basic concepts of probability, Markov processes, and convex optimization. Subsequent chapters discuss the dramatic changes that have occurred in the field of the Monte Carlo method, with coverage of many modern topics including: Markov Chain Monte Carlo, variance reduction techniques such as importance (re-)sampling, and the transform likelihood ratio method, the score function method for sensitivity analysis, the stochastic approximation method and the stochastic counter-part method for Monte Carlo optimization, the cross-entropy method for rare events estimation and combinatorial optimization, and application of Monte Carlo techniques for counting problems. An extensive range of exercises is provided at the end of each chapter, as well as a generous sampling of applied examples.
The Third Edition features a new chapter on the highly versatile splitting method, with applications to rare-event estimation, counting, sampling, and optimization. A second new chapter introduces the stochastic enumeration method, which is a new fast sequential Monte Carlo method for tree search. In addition, the Third Edition features new material on:
• Random number generation, including multiple-recursive generators and the Mersenne Twister
• Simulation of Gaussian processes, Brownian motion, and diffusion processes
• Multilevel Monte Carlo method
• New enhancements of the cross-entropy (CE) method, including the “improved” CE method, which uses sampling from the zero-variance distribution to find the optimal importance sampling parameters
• Over 100 algorithms in modern pseudo code with flow control
• Over 25 new exercises
Simulation and the Monte Carlo Method, Third Edition is an excellent text for upper-undergraduate and beginning graduate courses in stochastic simulation and Monte Carlo techniques. The book also serves as a valuable reference for professionals who would like to achieve a more formal understanding of the Monte Carlo method.
Reuven Y. Rubinstein, DSc, was Professor Emeritus in the Faculty of Industrial Engineering and Management at Technion-Israel Institute of Technology. He served as a consultant at numerous large-scale organizations, such as IBM, Motorola, and NEC. The author of over 100 articles and six books, Dr. Rubinstein was also the inventor of the popular score-function method in simulation analysis and generic cross-entropy methods for combinatorial optimization and counting.
Dirk P. Kroese, PhD, is a Professor of Mathematics and Statistics in the School of Mathematics and Physics of The University of Queensland, Australia. He has published over 100 articles and four books in a wide range of areas in applied probability and statistics, including Monte Carlo methods, cross-entropy, randomized algorithms, tele-traffic c theory, reliability, computational statistics, applied probability, and stochastic modeling.
Related to Simulation and the Monte Carlo Method
Titles in the series (100)
Linear Statistical Inference and its Applications Rating: 0 out of 5 stars0 ratingsMeasuring Agreement: Models, Methods, and Applications Rating: 0 out of 5 stars0 ratingsTime Series Analysis: Nonstationary and Noninvertible Distribution Theory Rating: 0 out of 5 stars0 ratingsAspects of Multivariate Statistical Theory Rating: 0 out of 5 stars0 ratingsStatistics and Causality: Methods for Applied Empirical Research Rating: 0 out of 5 stars0 ratingsForecasting with Univariate Box - Jenkins Models: Concepts and Cases Rating: 0 out of 5 stars0 ratingsNonparametric Finance Rating: 0 out of 5 stars0 ratingsSurvey Measurement and Process Quality Rating: 0 out of 5 stars0 ratingsNonlinear Statistical Models Rating: 0 out of 5 stars0 ratingsMeasurement Errors in Surveys Rating: 0 out of 5 stars0 ratingsTheory of Probability: A critical introductory treatment Rating: 0 out of 5 stars0 ratingsFlowgraph Models for Multistate Time-to-Event Data Rating: 0 out of 5 stars0 ratingsPeriodically Correlated Random Sequences: Spectral Theory and Practice Rating: 0 out of 5 stars0 ratingsApplications of Statistics to Industrial Experimentation Rating: 3 out of 5 stars3/5Probability and Conditional Expectation: Fundamentals for the Empirical Sciences Rating: 0 out of 5 stars0 ratingsRobust Correlation: Theory and Applications Rating: 0 out of 5 stars0 ratingsTheory of Ridge Regression Estimation with Applications Rating: 0 out of 5 stars0 ratingsComputation for the Analysis of Designed Experiments Rating: 0 out of 5 stars0 ratingsBusiness Survey Methods Rating: 0 out of 5 stars0 ratingsFundamental Statistical Inference: A Computational Approach Rating: 0 out of 5 stars0 ratingsSequential Stochastic Optimization Rating: 0 out of 5 stars0 ratingsTime Series Analysis with Long Memory in View Rating: 0 out of 5 stars0 ratingsStatistical Design and Analysis of Experiments: With Applications to Engineering and Science Rating: 0 out of 5 stars0 ratingsA Course in Time Series Analysis Rating: 3 out of 5 stars3/5The Statistical Analysis of Failure Time Data Rating: 0 out of 5 stars0 ratingsSensitivity Analysis in Linear Regression Rating: 0 out of 5 stars0 ratingsPreparing for the Worst: Incorporating Downside Risk in Stock Market Investments Rating: 4 out of 5 stars4/5Experiments with Mixtures: Designs, Models, and the Analysis of Mixture Data Rating: 5 out of 5 stars5/5Methods for Statistical Data Analysis of Multivariate Observations Rating: 0 out of 5 stars0 ratingsApplied Spatial Statistics for Public Health Data Rating: 0 out of 5 stars0 ratings
Related ebooks
Probability and Conditional Expectation: Fundamentals for the Empirical Sciences Rating: 0 out of 5 stars0 ratingsQuantile Regression: Theory and Applications Rating: 0 out of 5 stars0 ratingsStatistical Inference: A Short Course Rating: 4 out of 5 stars4/5Computational Intelligence and Pattern Analysis in Biology Informatics Rating: 0 out of 5 stars0 ratingsThe Numerate Leader: How to Pull Game-Changing Insights from Statistical Data Rating: 0 out of 5 stars0 ratingsBayesian Estimation and Tracking: A Practical Guide Rating: 0 out of 5 stars0 ratingsBayesian Inference in the Social Sciences Rating: 0 out of 5 stars0 ratingsHandbook of Monte Carlo Methods Rating: 3 out of 5 stars3/5Measuring Agreement: Models, Methods, and Applications Rating: 0 out of 5 stars0 ratingsHandbook of Volatility Models and Their Applications Rating: 5 out of 5 stars5/5BIM in Small-Scale Sustainable Design Rating: 5 out of 5 stars5/5Cost Estimation: Methods and Tools Rating: 5 out of 5 stars5/5Introduction to Linear Regression Analysis Rating: 3 out of 5 stars3/5Applied Bayesian Modelling Rating: 0 out of 5 stars0 ratingsSPSS Data Analysis for Univariate, Bivariate, and Multivariate Statistics Rating: 0 out of 5 stars0 ratingsModern Industrial Statistics: with applications in R, MINITAB and JMP Rating: 0 out of 5 stars0 ratingsThe Failure of Risk Management: Why It's Broken and How to Fix It Rating: 0 out of 5 stars0 ratingsBusiness Statistics For Dummies Rating: 5 out of 5 stars5/5Complex Surveys: A Guide to Analysis Using R Rating: 0 out of 5 stars0 ratingsUnderstanding Biostatistics Rating: 0 out of 5 stars0 ratingsApplied Econometrics Using the SAS System Rating: 0 out of 5 stars0 ratingsAnalyzing and Modeling Spatial and Temporal Dynamics of Infectious Diseases Rating: 0 out of 5 stars0 ratingsHandbook of High-Frequency Trading and Modeling in Finance Rating: 0 out of 5 stars0 ratingsPrediction Revisited: The Importance of Observation Rating: 0 out of 5 stars0 ratingsMultiple Imputation and its Application Rating: 0 out of 5 stars0 ratingsQuantitative Portfolio Management: The Art and Science of Statistical Arbitrage Rating: 0 out of 5 stars0 ratingsQuantitative Assessments of Distributed Systems: Methodologies and Techniques Rating: 0 out of 5 stars0 ratingsWavelet Neural Networks: With Applications in Financial Engineering, Chaos, and Classification Rating: 0 out of 5 stars0 ratingsProbability, Statistics, and Stochastic Processes Rating: 0 out of 5 stars0 ratings
Mathematics For You
Quantum Physics for Beginners Rating: 4 out of 5 stars4/5Geometry For Dummies Rating: 5 out of 5 stars5/5Mental Math Secrets - How To Be a Human Calculator Rating: 5 out of 5 stars5/5Practice Makes Perfect Algebra II Review and Workbook, Second Edition Rating: 4 out of 5 stars4/5Algebra - The Very Basics Rating: 5 out of 5 stars5/5Algebra I Workbook For Dummies Rating: 3 out of 5 stars3/5Basic Math & Pre-Algebra For Dummies Rating: 4 out of 5 stars4/5See Ya Later Calculator: Simple Math Tricks You Can Do in Your Head Rating: 4 out of 5 stars4/5The Thirteen Books of the Elements, Vol. 1 Rating: 0 out of 5 stars0 ratingsThe Little Book of Mathematical Principles, Theories & Things Rating: 3 out of 5 stars3/5Game Theory: A Simple Introduction Rating: 4 out of 5 stars4/5Calculus Made Easy Rating: 4 out of 5 stars4/5The Everything Guide to Algebra: A Step-by-Step Guide to the Basics of Algebra - in Plain English! Rating: 4 out of 5 stars4/5Painless Geometry Rating: 4 out of 5 stars4/5Relativity: The special and the general theory Rating: 5 out of 5 stars5/5Precalculus: A Self-Teaching Guide Rating: 4 out of 5 stars4/5Is God a Mathematician? Rating: 4 out of 5 stars4/5The Golden Ratio: The Divine Beauty of Mathematics Rating: 5 out of 5 stars5/5Real Estate by the Numbers: A Complete Reference Guide to Deal Analysis Rating: 0 out of 5 stars0 ratingsThe Math of Life and Death: 7 Mathematical Principles That Shape Our Lives Rating: 4 out of 5 stars4/5Introducing Game Theory: A Graphic Guide Rating: 4 out of 5 stars4/5Standard Deviations: Flawed Assumptions, Tortured Data, and Other Ways to Lie with Statistics Rating: 4 out of 5 stars4/5Flatland Rating: 4 out of 5 stars4/5Summary of The Black Swan: by Nassim Nicholas Taleb | Includes Analysis Rating: 5 out of 5 stars5/5My Best Mathematical and Logic Puzzles Rating: 5 out of 5 stars5/5
Reviews for Simulation and the Monte Carlo Method
0 ratings0 reviews
Book preview
Simulation and the Monte Carlo Method - Reuven Y. Rubinstein
DPK
CHAPTER 1
PRELIMINARIES
1.1 INTRODUCTION
The purpose of this chapter is to review some basic facts from probability, information theory, and optimization. In particular, Sections 1.2–1.11 summarize the main points from probability theory. Sections 1.12–1.14 describe various fundamental stochastic processes, such as Poisson, Markov, and Gaussian processes. Elements of information theory are given in Section 1.15, and Section 1.16 concludes with an outline of convex optimization theory.
1.2 RANDOM EXPERIMENTS
The basic notion in probability theory is that of a random experiment: an experiment whose outcome cannot be determined in advance. The most fundamental example is the experiment where a fair coin is tossed a number of times. For simplicity suppose that the coin is tossed three times. The sample space, has eight possible outcomes.
where, for example, HTH means that the first toss is heads, the second tails, and the third heads.
Subsets of the sample space are called eventsthat the third toss is heads is
occurs , called the union , called the intersection , called the complement that have no outcomes in common, that is, their intersection is empty, are called disjoint events. The main step is to specify the probability of each event.
Definition 1.2.1 (Probability) A probability of disjoint events
(1.1)
Equation (1.1) is referred to as the sum rule of probability. It states that if an event can happen in a number of different ways, but not simultaneously, the probability of that event is simply the sum of the probabilities of the comprising events.
, the sum rule implies that
,
(1.2)
. More generally, if a random experiment has finitely many and equally likely outcomes, the probability is always of the form (1.2). In that case the calculation of probabilities reduces to counting.
1.3 CONDITIONAL PROBABILITY AND INDEPENDENCE
. This leads to the definition of the conditional probability :
(1.3)
.
Rewriting . This can be generalized easily to the product rule ,
(1.4)
.
is a partition and hence, by the definition of conditional probability, we have the law of total probability:
(1.5)
Combining this with the definition of conditional probability gives Bayes’ rule:
(1.6)
are said to be independent , an alternative definition of independence is
. We can extend this definition to arbitrarily many events.
Definition 1.3.1 (Independence) , are said to be independent ,
Remark 1.3.1 such that certain events are independent.
■ EXAMPLE 1.1
are tails is
1.4 RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS
may not always be convenient or necessary. In practice, we are only interested in certain observations (i.e., numerical measurements) in the experiment. We incorporate these into our modeling process via the introduction of random variables, .
■ EXAMPLE 1.2
. The probability distribution is given by the binomial formula
(1.7)
Namely, by such events.
, and so on — is completely specified by the cumulative distribution function (cdf), defined by
is said to have a discrete is called the probability mass function — but see Remark 1.4.1.
■ EXAMPLE 1.3
is given by
, either (1, 3), (2, 3), (3, 3), (3, 2), or (3, 1) has to be thrown, each of which happens with probability 1/36.
is said to have a continuous ,
(1.8)
is called the probability density function . Note that in the continuous case the cdf is given by
in the sense that
Remark 1.4.1 (Probability Density) Note that we have deliberately used the same , for both pmf and pdf. This is because the pmf and pdf play very similar roles and can, in more advanced probability theory, both be viewed as particular instances of the general notion of probability densityin both the discrete and continuous case the pdf or (probability) density (function).
1.5 SOME IMPORTANT DISTRIBUTIONS
.
Table 1.1: Commonly used continuous distributions.
Table 1.2: Commonly used discrete distributions.
1.6 EXPECTATION
It is often useful to consider different kinds of numerical characteristics of a random variable. One such quantity is the expectation, which measures the mean value of the distribution.
Definition 1.6.1 (Expectation) . The expectation ), is defined by
Another useful quantity is the variance, which measures the spread or dispersion of the distribution.
Definition 1.6.2 (Variance) The variance ), is defined by
The square root of the variance is called the standard deviation. Table 1.3 lists the expectations and variances for some well-known distributions.
Table 1.3: Expectations and variances for some well-known distributions.
, we can write
from which follows the Markov inequality: ,
(1.9)
, we have
(1.10)
This is called the Chebyshev inequality; then, by the Markov inequality (1.9) and the definition of the variance,
, so that (1.10) follows.
1.7 JOINT DISTRIBUTIONS
Often a random experiment is described by more than one random variable. The theory for multiple random variables is similar to that for a single random variable.
be random variables describing some random experiment. We can accumulate these into a random vector of random variables is called a stochastic processis called the parameter set or index set or [1, 10]). The set of possible values for the stochastic process is called the state space.
is specified by the joint cdf
The joint pdf is such that
is found as
. Then the conditional pdf is given by
The corresponding conditional expectation is (in the continuous case)
. It can be shown (see, for example, , that is,
(1.11)
are said to be independent. More precisely:
Definition 1.7.1 (Independent Random Variables) are called independent ,
(discrete or continuous) are independent if and only if
(1.12)
are the marginal pdfs.
■ EXAMPLE 1.4 Bernoulli Sequence
means tails (or failure). Also, let
is called a Bernoulli sequence or Bernoulli process elements. We now have
. Compare this with Example 1.2.
Remark 1.7.1 An infinite that are independent and identically distributed, . We will use this abbreviation throughout this book.
is a weighted average of all values that this function can take. Specifically, in the continuous case,
As a direct consequence of the definitions of expectation and independence, we have
(1.13)
are constants. Similarly, for independent random variables, we have
The covariance , respectively, is defined as
This is a measure for the amount of linear dependency between the variables. A scaled version of the covariance is given by the correlation coefficient,
. It can be shown that the correlation coefficient always lies between −1 and 1; see Problem 1.13.
For easy reference, Table 1.4 lists some important properties of the variance and covariance. The proofs follow directly from the definitions of covariance and variance and the properties of the expectation.
Table 1.4: Properties of variance and covariance.
As a consequence of properties 2 and 7, for any sequence of independent
(1.14)
.
, it is convenient to write the expectations and covariances in vector notation.
Definition 1.7.2 (Expectation Vector and Covariance Matrix) , we define the expectation vector as the vector of expectations
The covariance matrix element is
If we define the expectation of a vector (matrix) to be the vector (matrix) of expectations, then we can write
and
in the one-dimensional case.
Remark 1.7.2 is symmetric. In fact (see Problem 1.16), it is positive semidefinite, ,
1.8 FUNCTIONS OF RANDOM VARIABLES
are measurements of a random experiment. Often we are only interested in certain functions of the measurements rather than the individual measurements. Here are some examples.
■ EXAMPLE 1.5
. Thus, in general,
(1.15)
■ EXAMPLE 1.6
we first write
. Differentiating with respect to z now gives
(1.16)
in the first equation needs to be replaced with its negative value.
■ EXAMPLE 1.7 Order Statistics
. In many applications one is interested in the distribution of the order statistics follows from
Similarly,
! times the joint density of the unordered sample and zero elsewhere.
1.8.1 Linear Transformations
, is called a linear transformation. Now consider a random , and let
. Let us first see how the expectation vector and covariance matrix are transformed.
Theorem 1.8.1 If has an expectation vector and covariance matrix , then the expectation vector and covariance matrix of are given by
(1.17)
and
(1.18)
Proof: and
□
? Consider . Then,
Figure 1.1: Linear transformation.
Now recall from linear algebra (e.g., . Thus,
, we obtain
(1.19)
1.8.2 General Transformations
, written out as
is the matrix of Jacobi , that is,
. Then, as in the linear case,
Hence we have the transformation rule
(1.20)
Remark 1.8.1 .
1.9 TRANSFORMS
Many calculations and manipulations involving probability distributions are facilitated by the use of transforms. Two typical examples are the probability generating function , defined by
and the Laplace transform , by
All transforms share an important uniqueness property: two distributions are the same if and only if their respective transforms are the same.
■ EXAMPLE 1.8
; then its probability generating function is given by
(1.21)
is given by
.
■ EXAMPLE 1.9
is given by
is
.
1.10 JOINTLY NORMAL RANDOM VARIABLES
It is helpful to view normally distributed random variables as simple transformations of standard normal given by
. Then, by has density
. This procedure is called standardization.
is given by
(1.22)
Consider the affine transformation (i.e., a linear transformation plus a constant vector)
(1.23)
. Any random vector of the form (1.23) is said to have a jointly normal or multivariate normal is given by
We have
, so that
, and therefore
(1.24)
Note that this formula is very similar to that of the one-dimensional case.
, there exists a unique lower triangular matrix
(1.25)
. This matrix can be obtained efficiently via the Cholesky square root method; see Section A.1 of the Appendix.
1.11 LIMIT THEOREMS
We briefly discuss two of the main results in probability: the law of large numbers and the central limit theorem. Both are associated with sums of independent random variables.
.
. Here is the more precise statement.
Theorem 1.11.1 (Strong Law of Large Numbers) If are iid with expectation , then
is large. The more precise statement is given next.
Theorem 1.11.2 (Central Limit Theorem) If are iid with expectation and variance , then for all ,
where is the cdf of the standard normal distribution.
. To see the central limit theorem in action, consider distribution. We clearly see convergence to a bell-shaped curve, characteristic of the normal distribution.
Figure 1.2: Illustration of the central limit theorem for (left) the uniform distribution and (right) the exponential distribution.
(1.26)
. As a rule of thumb, this normal approximation to the binomial distribution are larger than 5.
.
1.12 POISSON PROCESSES
is given in Figure 1.3.
.
when the two intervals do not intersect. Such considerations lead to the following definition:
Definition 1.12.1 (Poisson Process) is called a Poisson process with rate if
The numbers of points in nonoverlapping intervals are independent.
.
is the rate .
is
(1.27)
. As a consequence, we have
distribution; see Problem 1.17. Thus
(1.28)
-distributed random variables. This corresponds with the second important characterization of a Poisson process:
An arrival counting process is a Poisson process with rate if and only if the interarrival times are independent and -distributed random variables.
Poisson and Bernoulli processes are akin, and much can be learned about Poisson processes via the following Bernoulli approximation. Let is called the Bernoulli approximation .
As an example of the usefulness of this interpretation, we now demonstrate that the Poisson property (b) in Definition 1.12.1 follows basically from the independence ). Hence,
(1.29)
Equation (1.29) follows from the Poisson approximation to the binomial distribution; see Problem 1.22.
.
1.13 MARKOV PROCESSES
, is called a Markov process ,
(1.30)
. That is, in order to predict future states, we only need to know the present one. Property (1.30) is called the Markov property.
can take), Markov processes come in many different forms. A Markov process with a discrete index set is called a Markov chain) is called a Markov jump process.
1.13.1 Markov Chains
. In this case the Markov property (1.30) is
(1.31)
. We restrict ourselves to Markov chains for which the conditional probabilities
(1.32)
. Such chains are called time-homogeneous. The probabilities in (1.32) are called the (one-step) transition probabilities is called the initial distribution . Namely, we have by the product rule (1.4) and the Markov property (1.30),
is countable, we can arrange the one-step transition probabilities in an array. This array is called the (one-step) transition matrix has the form
Note that the elements in every row are positive and sum up to unity.
is through its transition graph.
■ EXAMPLE 1.10 Random Walk on the Integers
defined by
is called a random walk on the integers. The corresponding transition graph is given in .
.
, define the row vector
the distribution vector, or simply the distribution, of the initial distribution -step probabilities can be found simply by matrix multiplication.
Theorem 1.13.1 The distribution of at time is given by
(1.33)
for all . (Here denotes the identity matrix.)
Proof: The proof is by induction. Equality by definition.
. We have
But ,
. This completes the induction step, and thus the theorem is proved. □
.
1.13.2 Classification of States
leads to communicate into equivalence classes , the Markov chain is said to be irreducibleis called a closed is called an absorbing is closed. For example, in the transition graph depicted in is the only closed set: the Markov chain cannot escape from it. If state 1 were missing, state 2 would be absorbing. In Example 1.10 the Markov chain is irreducible since all states communicate.
Figure 1.5: A transition graph with three equivalence classes.
is called a recurrent is called transient. A recurrent state is called positive recurrent ; otherwise, it is called null recurrent. Finally, a state is said to be periodic, with period ; otherwise, it is called aperiodic. For example, in Figure 1.5 states 1 and 2 are recurrent, and the other states are transient. All these states are aperiodic. The states of the random walk of Example 1.10 are periodic with period 2.
recurrent (transient). Thus, in an irreducible Markov chain, one state being recurrent implies that all other states are also recurrent. And if one state is transient, then so are all the others.
1.13.3 Limiting Behavior
. It can be shown (see, for example, -step probabilities converge to a constant that does not depend on the initial state. More specifically,
(1.34)
form the limiting distribution .
Theorem 1.13.2 For an irreducible, aperiodic Markov chain with transition matrix , if the limiting distribution exists, then it is uniquely determined by the solution of
(1.35)
with and . Conversely, if there exists a positive row vector satisfying (1.35) and summing up to 1, then is the limiting distribution of the Markov chain. Moreover, in that case, for all and all states are positive recurrent.
Proof: is finite, the result is simply a consequence of -th unit vector, we have
, we obtain
sum up to unity. We omit the proof of the converse statement. □
■ EXAMPLE 1.11 Random Walk on the Positive Integers
This is a slightly different random walk than the one in with transition matrix
.
becomes
as
, leads to null-recurrent states.)
. Then, combining (and is given by this