Random variables app
For a given sample space δ of some experiment, a random variable is any rule tha associates a number with each outcome in δ.dom
Any random variable whose only possible values are 0 and 1 is called Bermoulli random variable.ide
Two Types of Random Variablesui
The probability distribution of X says how the total probability of 1 is distributed among the various possible X values.this
The probability distribution or probability mass function of a discrete rv is defined for every number x by p(x)=P(X=x)=P(all s in δ:X(s)=x):lua
In words, for every possible value x of the random variable, the function specifies the probability of observing that value when the experiment is performed.spa
Suppose p(x) depends on a quantity that can be assigned any one of a number of possible values, with each different value determining a different probability distribution. Such a quantity is called a parameter of the distribution. The collection of all probability distributions for different values of the parameter is called a family of probability distribution.rest
The cumulative distribution function F(x) of a discrete rv variable X with pmf p(x) is defined for every number x byorm
F(x) = P(X≤x) = ∑p(y)ci
For any number x, F(x) is the probability that the observed value of X will be at most x.
It is often helpful to think of a pmf as specifying a mathematical model for a discrete population. Once we have such a mathematical model for a population, we will use it to compute values of population characteristics and make inferences about such characteristics.
When computing the expections of X, the population size is irrelevant as long as the pmf is given. The average or mean value of X is then a weighted average of the possible values, where the weights are the probabilities of those values.
Let X be a discrete rv with set of possible values D and pmf p(x). The expected value or mean value of X, denoted by E(X) or μx, is E(X)=μx=∑x*p(x)
If the rv X has set of possible values D and pmf p(x), then the expected value of any function h(x), denoted by E[h(x)] or ch(x) is computed by E[h(X)] = ∑h(x)*p(x)
According to this proposition, E[h(x)] is computed in the same way that E(x) itself is, expect that h(x) is substituted in place of x.
The h(X) function of interest is quite frequently a linear function aX+b. In this case, E[h(x)] is easily computed from E(X).
The expected value of a linear function equals the linear function evaluated at the expected value E(X).
PROPOSITION:E(aX+b) = a* E(X) + b
We will use variance of X to measure the amount of variability in the distribution of X.
Let X have pmf p(x) and expected value μ. Then the variance of X, denoted by V(X) or σx2 or just σ2, is
V(X) = ∑(x-μ)2*p(x) = E[(X-μ)2]
The number of arithmetic operations necessary to compute σ2 can be reduced by using an alternative computing formula.
V(X) = σ2 = E(X2) - [E(X)]2
The variance of h(x) is the expected value of the squared difference between h(x) and its expected value.
An experiment for which conditions 1-4 are satisfied is called a binomial experiment:
Given a binomial experiment consisting of n trials, the binomial random variable X associate with this experiment is defined as X = the number of S's among the n trails
For X ~ Bin(n,p), the cdf will be denoted by
P(X≤x) = B(x;n,p) = ∑b(y;n,p) x= 0, 1,...,n
PROPOSITION: If X ~ Bin(n,p), then E(X)=np, V(X)=np(1-p)=npq, and σx = (npq)1/2 where q = 1-p.
The hypergeometric distribution is the exact probability model for the number of S's in the sample.
The binomial rv X is the number of S's when the number n of trials is fixed, whereas the negative binomial distribution arises from fixing the number of S's and leting the number of trials be random.
The assumptions leading to the hypergeometric distribution are as follows:
If X is the number of S's in a completely random sample of size n drawn from a population consisting of M S's and (N-M) F's, then the probability distribution of X, called the hypergeometric distribution.
The mean and variance of the hypergeometric rv X having pmf h(x;n,M,N) are
E(X) = n * M/N
V(X) = ((N-n)/(N-1))*n*M/N*(1-M/N)
The means of the binomial and hypergeometric rv's are equal, whereas the variances of the two rv's differ by the factor (N-n)/(N-1), often called the finite population correction factor.
Let the population size, N, and number of population S's, M, get large with the ratio M/N approaching p. Then h(x;n,M,N) approaches b(x;n,p), so for n/N small, the two are approximately equal provided that p is not too near either 0 or 1.
The negative binomial rv and distribution are based on an experiment satisfying the following conditions:
The random variable of interest is X = the number of failures that precede the rth success;
X is called a negative binomial random variable because, in contrast to the binomial rv, the number of successes is fixed and the number of trials is random.
A random variable X is said to have a Poisson distribution if the pmf of X is p(x;λ) = e-λλx/x! x = 0,1,2,... for some λ>0
PROPOSITION: Suppose that in a binomial pmf b(x;n,p), we let n→∞ and p→0 in such a way that np approaches a value λ>0. Then b(x;n,p)→p(x;λ).
PROPOSITION: If X has a Poisson distribution with parameter λ, then E(X) = V(X) = λ.