Probability Distributions: With Truncated, Log and Bivariate Extensions
()
About this ebook
Related to Probability Distributions
Related ebooks
Introduction to Statistics in Metrology Rating: 0 out of 5 stars0 ratingsApplications of Variational Inequalities in Stochastic Control Rating: 2 out of 5 stars2/5Random Forests with R Rating: 0 out of 5 stars0 ratingsLog-Linear Modeling: Concepts, Interpretation, and Application Rating: 0 out of 5 stars0 ratingsForest Biometrics Rating: 0 out of 5 stars0 ratingsLinear and Generalized Linear Mixed Models and Their Applications Rating: 0 out of 5 stars0 ratingsStatistical Inference in Financial and Insurance Mathematics with R Rating: 0 out of 5 stars0 ratingsCoding Ockham's Razor Rating: 0 out of 5 stars0 ratingsStatistical Methods for Overdispersed Count Data Rating: 0 out of 5 stars0 ratingsData Treatment in Environmental Sciences Rating: 0 out of 5 stars0 ratingsMultivariate Statistical Simulation: A Guide to Selecting and Generating Continuous Multivariate Distributions Rating: 0 out of 5 stars0 ratingsAn Introduction to Stochastic Orders Rating: 0 out of 5 stars0 ratingsLectures on Dynamics of Stochastic Systems Rating: 0 out of 5 stars0 ratingsApplications of Statistics to Industrial Experimentation Rating: 3 out of 5 stars3/5Information Theory and Statistics Rating: 0 out of 5 stars0 ratingsDistributed Parameter Control Systems: Theory and Application Rating: 0 out of 5 stars0 ratingsSchaum's Outline of Elements of Statistics I: Descriptive Statistics and Probability Rating: 0 out of 5 stars0 ratingsStatistics of Directional Data Rating: 5 out of 5 stars5/5The Gradient Test: Another Likelihood-Based Test Rating: 0 out of 5 stars0 ratingsModel Averaging Rating: 0 out of 5 stars0 ratingsStatistics Rating: 4 out of 5 stars4/5Elements of Copula Modeling with R Rating: 0 out of 5 stars0 ratingsComplex Surveys: A Guide to Analysis Using R Rating: 0 out of 5 stars0 ratingsAdvances in Domain Adaptation Theory Rating: 0 out of 5 stars0 ratingsBusiness Statistics I Essentials Rating: 5 out of 5 stars5/5Experimental Statistics Rating: 4 out of 5 stars4/5Sample Size Methodology Rating: 1 out of 5 stars1/5Sensitivity Analysis in Linear Regression Rating: 0 out of 5 stars0 ratingsStatistical Methods for Hospital Monitoring with R Rating: 0 out of 5 stars0 ratingsIntroductory Statistics Rating: 0 out of 5 stars0 ratings
Mathematics For You
My Best Mathematical and Logic Puzzles Rating: 5 out of 5 stars5/5Quantum Physics for Beginners Rating: 4 out of 5 stars4/5Calculus Made Easy Rating: 4 out of 5 stars4/5Algebra - The Very Basics Rating: 5 out of 5 stars5/5Standard Deviations: Flawed Assumptions, Tortured Data, and Other Ways to Lie with Statistics Rating: 4 out of 5 stars4/5The Thirteen Books of the Elements, Vol. 1 Rating: 0 out of 5 stars0 ratingsReal Estate by the Numbers: A Complete Reference Guide to Deal Analysis Rating: 0 out of 5 stars0 ratingsThe Everything Guide to Algebra: A Step-by-Step Guide to the Basics of Algebra - in Plain English! Rating: 4 out of 5 stars4/5Game Theory: A Simple Introduction Rating: 4 out of 5 stars4/5Alan Turing: The Enigma: The Book That Inspired the Film The Imitation Game - Updated Edition Rating: 4 out of 5 stars4/5Mental Math Secrets - How To Be a Human Calculator Rating: 5 out of 5 stars5/5Basic Math & Pre-Algebra For Dummies Rating: 4 out of 5 stars4/5The Little Book of Mathematical Principles, Theories & Things Rating: 3 out of 5 stars3/5Flatland Rating: 4 out of 5 stars4/5Algebra I For Dummies Rating: 4 out of 5 stars4/5The Everything Everyday Math Book: From Tipping to Taxes, All the Real-World, Everyday Math Skills You Need Rating: 5 out of 5 stars5/5Logicomix: An epic search for truth Rating: 4 out of 5 stars4/5The Math of Life and Death: 7 Mathematical Principles That Shape Our Lives Rating: 4 out of 5 stars4/5Is God a Mathematician? Rating: 4 out of 5 stars4/5Basic Math Notes Rating: 5 out of 5 stars5/5Algebra I Workbook For Dummies Rating: 3 out of 5 stars3/5The Golden Ratio: The Divine Beauty of Mathematics Rating: 5 out of 5 stars5/5Relativity: The special and the general theory Rating: 5 out of 5 stars5/5See Ya Later Calculator: Simple Math Tricks You Can Do in Your Head Rating: 4 out of 5 stars4/5A Mind for Numbers | Summary Rating: 4 out of 5 stars4/5ACT Math & Science Prep: Includes 500+ Practice Questions Rating: 3 out of 5 stars3/5
Reviews for Probability Distributions
0 ratings0 reviews
Book preview
Probability Distributions - Nick T. Thomopoulos
© Springer International Publishing AG, part of Springer Nature 2018
Nick T. ThomopoulosProbability Distributions https://doi.org/10.1007/978-3-319-76042-1_1
1. Continuous Distributions
Nick T. Thomopoulos¹
(1)
Stuart School of Business, Illinois Institute of Technology, Chicago, IL, USA
1.1 Introduction
A variable, x, is continuous when x can be any number between two limits. For example, a scale measures a boy at 150 pounds; and assuming the scale is correct within one-half pound, the boy’s actual weight is a continuous variable that could fall anywhere from 149.5 to 150.5 pounds. The variable, x, is a continuous random variable when a mathematical function, called the probability density defines the shape along the admissible range. The density is always zero or larger and the positive area below the density equals one. Each unique continuous random variable is defined by a probability density that flows over the admissible range. Eight of the common continuous distributions are described in the chapter. For each of these, the range of the variable is stated, along with the probability density, and the associated parameters. Also described is the cumulative probability distribution that is needed by an analyst to measure the probability of the x falling in a sub-range of the admissible region. Some of the distributions do not have closed-form solutions, and thereby, quantitative methods are needed to measure the cumulative probability. Sample data is used to estimate the parameter values. Examples are included to demonstrate the features and use of each distribution. The distributions described in this chapter are the following: continuous uniform , exponential , Erlang , gamma , beta , Weibull , normal and lognormal . The continuous uniform occurs when all values between limits a to b are equally likely. The normal density is symmetrical and bell shaped. The exponential happens when the most likely value is at x = 0, and the density tails down in a relative way as x increases. The density of the Erlang has many shapes that range between the exponential and the normal. The shape of the gamma density varies from exponential-like to one where the mode (most likely) and the density skews to the right. The beta has many shapes: uniform, ramp down, ramp up, bathtub-like, normal-like, and all shapes that skew to the right and in the same manner they skew to the left. The Weibull density varies from exponential-like to shapes that skew to the right. The lognormal density peaks near zero and skews far to the right.
Law and Kelton [1]; Hasting and Peacock [2]; and Hines et al. [3] present thorough descriptions on the properties of the common continuous probability distributions.
1.2 Sample Data Statistics
When n sample data, (x1, …, xn), are collected, various statistical measures can be computed as described below:
$$ \mathrm{x}(1)=\mathrm{minimum}\ \mathrm{of}\ \left({\mathrm{x}}_1,\dots, {\mathrm{x}}_{\mathrm{n}}\right) $$$$ \mathrm{x}\left(\mathrm{n}\right)=\mathrm{maximum}\ \mathrm{of}\ \left({\mathrm{x}}_1,\dots, {\mathrm{x}}_{\mathrm{n}}\right) $$$$ \overline{x}=\mathrm{average} $$$$ \mathrm{s}=\mathrm{standard}\ \mathrm{deviation} $$$$ \kern1.4em \operatorname{cov}=\mathrm{s}/\overline{x}=\mathrm{coefficient}\ \mathrm{of}\ \mathrm{variation} $$$$ \uptau ={\mathrm{s}}^2/\overline{x}=\mathrm{lexis}\ \mathrm{ratio} $$1.3 Notation
The statistical notation used in this book is the following:
$$ {\displaystyle \begin{array}{l}\mathrm{E}\left(\mathrm{x}\right)=\mathrm{expected}\ \mathrm{value}\ \mathrm{of}\ \mathrm{x}\\ {}\mathrm{V}\left(\mathrm{x}\right)=\mathrm{variance}\ \mathrm{of}\ \mathrm{x}\\ {}\upmu =\mathrm{mean}\\ {}{\upsigma}^2=\mathrm{variance}\\ {}\upsigma =\mathrm{standard}\ \mathrm{deviation}\end{array}} $$Example 1.1
Suppose an experiment yields n = 10 sample data values as follows: [24, 27, 19, 14, 32, 28, 35, 29, 25, 33]. The statistical measures from this data are listed below.
$$ \mathrm{x}(1)=\min =14 $$$$ \mathrm{x}(10)=\max =35 $$$$ \overline{x}=26.6 $$$$ \mathrm{s}=6.44 $$$$ \operatorname{cov}=0.24 $$$$ \tau =1.56 $$1.4 Parameter Estimating Methods
Two popular methods have been developed to estimate the parameters of a distribution from sample data . One is called the maximum-likelihood-estimate method (MLE) that is mathematically formulated to find the parameter estimate that gives the most likely fit with the sample data. The other method is called the method-of-moments (MoM) that substitutes the statistical measures like ( $$ \overline{x} $$ , s) into their mathematical counterparts [μ, σ] and applies algebra to find the estimates of the parameters.
1.5 Transforming Variables
While analyzing sample data , it is sometimes useful to convert a variable x to another variable, x′, where x′ ranges from zero to one; or where x′ is zero or larger. More discussion is below.
Transform Data to (0,1)
A way to convert a variable from x to x` so that x′ lies between 0 and 1 is described here. Recall the summary statistics of the variable x as listed earlier. For convenience in notation, let a′ = x(1) for the minimum, and b′ = x(n) for the maximum. When x, with average $$ \overline{x} $$ and standard deviation s , is converted to x′ by the relation:
$$ {\mathrm{x}}^{\prime }=\left(\mathrm{x}\hbox{--} {\mathrm{a}}^{\prime}\right)/\left({\mathrm{b}}^{\prime}\hbox{--} {\mathrm{a}}^{\prime}\right) $$the range on x` becomes (0,1). The converted sample average and standard deviation are listed below:
$$ {\displaystyle \begin{array}{l}{\overline{x}}^{\prime }=\left(\overline{x}\hbox{--} {\mathrm{a}}^{\prime}\right)/\left({\mathrm{b}}^{\prime}\hbox{--} {\mathrm{a}}^{\prime}\right)\\ {}{\mathrm{s}}^{\prime }=\mathrm{s}/\left({\mathrm{b}}^{\prime}\hbox{--} {\mathrm{a}}^{\prime}\right)\end{array}} $$respectively, and the coefficient of variation of x` is:
$$ \mathrm{co}{\mathrm{v}}^{\prime }={\mathrm{s}}^{`}/{\overline{x}}^{\prime } $$When x lies in the (0,1) range, the cov is sometimes useful to identify the distribution that best fits sample data .
Transform Data to (x ≥ 0)
A way to convert a variable, x to, x`, where x′ ≥ 0 is given here. The summary statistics described earlier of the variable x are used again with a′ = x(1) for the minimum. When x is converted to x′ by the relation:
$$ {\mathrm{x}}^{\prime }=\left(\mathrm{x}\hbox{--} {\mathrm{a}}^{\prime}\right) $$the range of x′ becomes zero and larger. The corresponding sample average and standard deviation become:
$$ {\overline{x}}^{\prime }=\left(\overline{x}\hbox{--} {\mathrm{a}}^{\prime}\right) $$$$ {\mathrm{s}}^{\prime }=\mathrm{s} $$respectively. Finally, the coefficient of variation is:
$$ \mathrm{co}{\mathrm{v}}^{\prime }={\mathrm{s}}^{`}/{\overline{x}}^{\prime } $$1.6 Continuous Random Variables
A continuous random variable, x, can take on any value in a range that spans between limits a and b. Note where the low limit , a, could be minus infinity; and the high limit , b, could be plus infinity. An example is the amount of rainwater found is a five-gallon bucket after a rainfall. A probability density function, f(x), defines how the probability varies along the range, where the sum of the area within the admissible region sums to one. Below defines the probability density function, f(x), and the cumulative distribution function, F(x):
$$ {\displaystyle \begin{array}{ll}\mathrm{f}\left(\mathrm{x}\right)\ge 0& \mathrm{a}\le \mathrm{x}\le \mathrm{b}\\ {}\mathrm{F}\left(\mathrm{x}\right)={\int}_a^xf(w) dw& \mathrm{a}\le \mathrm{x}\le \mathrm{b}\end{array}} $$This chapter describes some of the common continuous probability distributions and their properties. The random variable of each is denoted as x, and below is a list of the distributions with their designations and parameters.
Of particular interest with each distribution is the coefficient of variation (cov) and its range of values that apply. When sample data is available, the sample cov can be measured and compared to each distribution’s cov range to help narrow the choice of the distribution that applies.
1.7 Continuous Uniform
A variable, x, follows a continuous uniform probability distribution, Cu(a,b), when it has two parameters a and b, where x can fall equally likely anywhere from a to b. See Fig. 1.1. The probability density , and the cumulative distribution function of x are below:
../images/464026_1_En_1_Chapter/464026_1_En_1_Fig1_HTML.gifFig. 1.1
The continuous uniform distribution
$$ {\displaystyle \begin{array}{ll}\mathrm{f}\left(\mathrm{x}\right)=1/\left(\mathrm{b}-\mathrm{a}\right)& \mathrm{a}\le \mathrm{x}\le \mathrm{b}\\ {}\mathrm{F}\left(\mathrm{x}\right)=\left(\mathrm{x}\hbox{--} \mathrm{a}\right)/\left(\mathrm{b}-\mathrm{a}\right)& \mathrm{a}\le \mathrm{x}\le \mathrm{b}\end{array}} $$The expected value, variance , and standard deviation of x are listed below:
$$ \mathrm{E}\left(\mathrm{x}\right)=\upmu =\left(\mathrm{b}+\mathrm{a}\right)/2 $$$$ \mathrm{V}\left(\mathrm{x}\right)={\upsigma}^2={\left(\mathrm{b}-\mathrm{a}\right)}^2/12 $$$$ \upsigma =\left(\mathrm{b}\hbox{--} \mathrm{a}\right)/\sqrt{12} $$Coefficient of Variation
Note when the low limit is set to zero, (a = 0):
$$ \upmu =\mathrm{b}/2 $$$$ \upsigma =\mathrm{b}/\sqrt{12} $$$$ \operatorname{cov}=2/\sqrt{12}=0.577 $$Parameter Estimates
When sample data , (x1, …, xn), is available, the parameters (a,b) are estimated as shown below by either the maximum-likelihood estimate (MLE) method, or by the method-of-moment estimate .
From MLE, the estimates of the two parameters are the following:
$$ \widehat{a}=\mathrm{x}(1)=\min\ \left({\mathrm{x}}_1,\dots, {\mathrm{x}}_{\mathrm{n}}\right) $$$$ \widehat{b}=\mathrm{x}\left(\mathrm{n}\right)=\max\ \left({\mathrm{x}}_1,\dots, {\mathrm{x}}_{\mathrm{n}}\right) $$The method-of-moment way uses the two equations: μ = (b + a)/2, and σ = (b – a)/ $$ \sqrt{12} $$ , to estimate the parameters (a,b). as below:
$$ \widehat{a}=\overline{x}-\sqrt{12}\mathrm{s}/2 $$$$ \widehat{b}=\overline{x}+\sqrt{12}\mathrm{s}/2 $$Recall, $$ \overline{x} $$ is the sample average and s is the sample standard deviation .
Example 1.2
Suppose a continuous uniform variable x has min = 0 and max = 1, yielding: x ~ CU(0,1). Some statistics are below:
$$ \mathrm{f}\left(\mathrm{x}\right)=1\kern1em 0\le \mathrm{x}\le 1 $$$$ \mathrm{F}\left(\mathrm{x}\right)=\mathrm{x}\kern1em 0\le \mathrm{x}\le 1 $$$$ \upmu =0.5 $$$$ {\upsigma}^2=1/12=0.083 $$$$ \upsigma =\sqrt{1/12}=0.289 $$$$ \operatorname{cov}=0.289/0.500=0.578 $$The probability of x less or equal to 0.45, say, is:
$$ \mathrm{P}\left(\mathrm{x}\le 0.45\right)=\mathrm{F}(0.45)=0.45. $$Example 1.3
The yield strength on a copper tube was measured at 70.23 from a device with accuracy of ± 0.40, evenly distributed. Hence, the true yield strength, denoted as x, follows a continuous uniform distribution with parameters:
$$ \mathrm{a}=70.23\hbox{--} 0.40=69.83 $$$$ \mathrm{b}=70.23+0.40=70.63 $$The probability density becomes:
$$ \mathrm{f}\left(\mathrm{x}\right)=1/0.80\kern1em 69.83\le \mathrm{x}\le 70.63 $$and the cumulative distribution is:
$$ \mathrm{F}\left(\mathrm{x}\right)=\left(\mathrm{x}\hbox{--} 69.83\right)/0.80\kern1.25em 69.83\le \mathrm{x}\le 70.63 $$The mean , variance and standard deviation are the following:
$$ \upmu =70.23 $$$$ {\upsigma}^2={(0.80)}^2/12=0.053 $$$$ \upsigma =\sqrt{0.53}=0.231 $$The probability that the true yield strength is below 70, say, becomes:
$$ \mathrm{F}(70.00)=\left(70.00\hbox{--} 69.83\right)/0.80=0.212 $$Note, the cov is 0.231/70.23 = 0.003
But when x is converted to x` = x – a, the mean , standard deviation , and coefficient of variation become:
$$ {\displaystyle \begin{array}{c}{\upmu}^{`}=\left(70.23\hbox{--} 69.83\right)=0.40\\ {}{\upsigma}^{`}=0.231\\ {}{\operatorname{cov}}^{`}=0.231/0.40=0.577\end{array}} $$Example 1.4
An experiment yields the following ten sample data entries: (12.7, 11.4, 15.3, 20.5, 13.6, 17.4, 15.6, 14.9, 19.7, 18.3). The analyst assumes the data comes from a continuous uniform distribution and seeks to estimate the parameters , (a, b). To accomplish, the following statistics are measured :
$$ \mathrm{x}(1)=\min =11.4 $$$$ \mathrm{x}\left(\mathrm{n}\right)=\max =20.5 $$$$ \overline{x}=15.93 $$$$ \mathrm{s}=3.00 $$The two methods of estimating the parameters (a,b) are applied. The MLE estimates are the following:
$$ \widehat{a}=\min =11.4 $$$$ \widehat{b}=\max =20.5 $$The method-of-moment estimates become:
$$ \widehat{a}=15.93-\sqrt{12}\times 3.00/2=10.73 $$$$ \widehat{b}=15.93+\sqrt{12}\times 3.00/2=21.13 $$Note, when x` = (x – a):
$$ {\overline{x}}^{`}=\left(15.93\hbox{--} 11.4\right)=4.53 $$$$ {\mathrm{s}}^{`}=\mathrm{s}=3.00 $$and
$$ \operatorname{cov}=\mathrm{s}/{\overline{x}}^{`}=3.00/4.53=0.662 $$which is reasonably close to the continuous uniform value of 0.577.
1.8 Exponential
The exponential distribution, Ex(θ), is used in many areas of science and is the primary distribution that applies in queuing theory to represent the time between arrivals and the time to service a unit. The variable, x, has its peak at x = 0 and a density that continually decreases as x increases. See Fig. 1.2 where θ = 1. The density has one parameter, θ, and is defined as below:
../images/464026_1_En_1_Chapter/464026_1_En_1_Fig2_HTML.gifFig. 1.2
The Exponential Distribution when μ = 1.0
$$ \mathrm{f}\left(\mathrm{x}\right)=\uptheta {\mathrm{e}}^{-\uptheta \mathrm{x}}\kern1em \mathrm{for}\;\mathrm{x}\ge 0 $$The cumulative probability distribution becomes,
$$ \mathrm{F}\left(\mathrm{x}\right)=1-{\mathrm{e}}^{-\uptheta \mathrm{x}}\kern1em \mathrm{for}\kern0.5em \mathrm{x}\ge 0 $$The mean , variance , and standard deviation of x are listed below:
$$ {\displaystyle \begin{array}{l}\upmu =1/\uptheta \\ {}{\upsigma}^2=1/{\uptheta}^2\\ {}\upsigma =1/\uptheta \end{array}} $$Since μ = σ, the coefficient-of-variation becomes:
$$ \operatorname{cov}=1.00 $$The median, x0.50, occurs when F(x) = 0.50; and thereby,
$$ \mathrm{F}\left({\mathrm{x}}_{0.50}\right)=0.50=1-{\mathrm{e}}^{-\uptheta {\mathrm{x}}_{0.50}} $$Solving for x0.50, yields:
$$ {\mathrm{x}}_{0.50}=-\ln \left(1\hbox{--} 0.50\right)/\uptheta =0.693/\uptheta =0.693\upmu $$where ln = the natural logarithm.
Parameter Estimate
When a sample of size n yields sample data (x1, …, xn) and an average, $$ \overline{x} $$ , the estimate of θ becomes:
$$ \widehat{\theta}=1/\widehat{x} $$