Some basic knowledge about probability theory

probability space

Random experiment is the basic concept of probability theory. The result of experiment cannot be expressed in exact language, but it has the following three characteristics: ( 1)
It can be repeated under the same conditions;
Know all possible results of the experiment;
(3) Before each trial, it is not possible to determine which result will appear.

The set of all possible results of a random experiment is called the sample space or basic event space of this experiment , denoted as Ω \OmegaOhOh \OmegaThe element e in Ω is called sample pointorbasic event,Ω \OmegaA subset event A of Ω is called an event, and the sample spaceΩ \OmegaΩ is calledan inevitable event, and the empty set∅ \emptyset is calledan impossible event.

Definition 1 : Let Ω \OmegaΩ is a set, F isΩ \OmegaA family of sets consisting of some subsets of Ω . If
(1)Ω ∈ F \Omega \in FOhF
(2)若 A ∈ F A \in F AF ,则A ~ = Ω \ A ∈ F \width{A} = \Omega \backslash A \in FA =Oh \ AF
(3)若 A n ∈ F , n = 1 , 2 , . . . A_{n} \in F,n=1,2,... AnF,n=1,2,. . . ,则∪ n = 1 ∞ A n ∈ F \cup^{\infty}_{n=1}A_{n} \in Fn=1AnF
is calledσ − \sigma^{-}p Algebra (Borel fields). (Ω, F) (\Omega, F)( Ω , F ) is a measurable space, and the elements in F are called events.
It is easy to know from the definition:
(4)∅ ∈ F \emptyset \in FF
(5)若A,B ∈ \in F,则 A \ B ∈ F A \backslash B \in F A\BF
(6)若 A i ∈ F , n = 1 , 2 , . . . A_{i} \in F,n=1,2,... AiF,n=1,2,. _ _ __i=1Ai ∩ i = 1 ∞ A i \cap^{\infty}_{i=1}A_{i} i=1Ai ∩ i = 1 ∞ A i ∈ F \cap^{\infty}_{i=1}A_{i} \in F i=1AiF

Random Variables and Their Distributions

Random variables are the main research objects of probability theory, and the statistical laws of random variables are described by distribution functions.

Definition 4 : Let ( Ω , F , P ) (\Omega,F,P)( Oh ,F,P ) is the probability space,X = X ( e ) X=X(e)X=X ( e ) is defined inΩ \OmegaA real function on Ω , if for any real numberx , { e : X ( e ) ≤ x } x,\{e: X(e) \leq x\}x,{ and X ( and )x } , thenX ( e ) X(e)X ( e ) is a random variableon F, abbreviated as random variableXXX,称
F ( x ) = P ( e : X ( e ) ≤ x ) , ∞ < x < ∞ F(x)=P(e:X(e) \leq x), \infty < x<\infty F(x)=P(e:X ( and )x),<x< is the distribution function
of the random variable.

The distribution function has the following properties:
(1) F ( x ) F(x)F ( x ) is a non-decreasing function, that is, whenx 1 < x 2 x_1 < x_2x1<x2, there is F ( x 1 ) ≤ F ( x 2 ) F(x_1) \leq F(x_2)F(x1)F(x2).
(2) F ( − ∞ ) = lim ⁡ x → − ∞ F ( x ) = 0 , F ( ∞ ) = lim ⁡ x → ∞ F ( x ) = 1 F(- \infty) =\lim _{x \to -\infty}F(x)=0, F(\infty)=\lim _{x \to \infty}F(x)=1 F()=limxF(x)=0,F()=limxF(x)=1.
(3) F ( x ) F(x) F ( x ) is right continuous, that is,F ( x + 0 ) = F ( x ) F(x+0)=F(x)F(x+0)=F(x).

The probability distribution of a discrete shorthand variable X is described by a distribution column :
pk = P ( X = xk ) , k = 1 , 2 , . . . p_k=P(X=x_k),k=1,2,...pk=P(X=xk),k=1,2,. . .Its
distribution function
F ( x ) = ∑ xk ≤ xpk F(x)=\sum_{x_k \leq x}p_kF(x)=xkxpk
Probability distribution of continuous random variable X uses probability density f ( x ) f(x)f ( x ) description, its distribution function
F ( x ) = ∫ − ∞ xf ( t ) dt F(x)= \int_{- \infty}^{x}f(t)dtF(x)=xf(t)dt

Common random variable distribution tables are as follows:

insert image description here
Definition 5 : Let (Ω, F, P) (\Omega, F, P)( Ω , F , P ) is a probability space,X = X ( e ) = ( X 1 ( e ) , . . . X n ( e ) ) X=X(e)=(X_1(e),... X_n(e))X=X ( and )=(X1(e),...Xn( e ) ) is defined inΩ \OmegaOn Ω in n-dimensional spaceR n R^{n}RA vector function of values ​​in n , if for any x = ( x 1 , x 2 , . . . xn ) ∈ R n , { e : X 1 ( e ) ≤ x 1 , X 2 ≤ x 2 , . . . , X n ( e ) ≤ xn } ∈ F x=(x_1,x_2,...x_n) \in R^{n}, \{ e:X_1(e)\leq x_1,X_2 \leq x_2,.. .,X_n(e)\leq x_n\} \in Fx=(x1,x2,...xn)Rn,{ e:X1(e)x1,X2x2,...,Xn(e)xn}F , thenX = X ( e ) X=X(e)X=X ( e ) is ann-dimensional random variableorn-dimensional random vector, which is called
F ( x ) = F ( x 1 , x 2 , . . . xn ) = P ( e : X 1 ( e ) ≤ x 1 , X 2 ≤ x 2 , . . . , X n ( e ) ≤ xn ) , x − ( x 1 , x 2 , . . . xn ) ∈ R n F(x)=F(x_1,x_2,...x_n) =P( e:X_1(e)\leq x_1,X_2 \leq x_2,...,X_n(e)\leq x_n), x-(x_1,x_2,...x_n) \in R^{n}F(x)=F(x1,x2,...xn)=P(e:X1(e)x1,X2x2,...,Xn(e)xn),x(x1,x2,...xn)Rn
X = ( X 1 , X 2 , . . . X n ) X=(X_1,X_2,...X_n) X=(X1,X2,...Xn) jointdistribution function.

Numerical properties of random variables

The probability distribution of a random variable is completely described by its distribution function, but how to determine the distribution function is quite troublesome. In practical problems, sometimes we only need to know some eigenvalues ​​of random variables.
Definition 7 : Let the distribution function of random variable X be F ( x ) F(x)F(x),若 ∫ − ∞ ∞ ∣ x ∣ d F ( x ) < ∞ \int_{- \infty}^{\infty}|x|dF(x)<\infty xdF(x)< , then
EX = ∫ − ∞ ∞ xd F ( x ) EX=\int_{- \infty}^{\infty}xdF(x)EX=x d F ( x ) is the numerical expectationormeanof
X. The integral on the right-hand side of the above equation is called the Lebesgue-Stieltjes integral. If X is a discrete random variable, the distribution ispk = P ( X = xk ) , k = 1 , 2 , . . . p_k=P(X=x_k),k=1,2,...

pk=P(X=xk),k=1,2,. . .
Then
EX = ∑ k = 1 ∞ xkpk EX=\sum_{k=1}^{\infty} x_kp_kEX=k=1xkpk
If X is a continuous random variable, the probability density is f ( x ) f(x)f(x),则
E X = ∫ − ∞ ∞ x f ( x ) d x EX=\int_{- \infty}^{\infty}xf(x)dx EX=x f ( x ) d x
The mathematical expectation of a random variable is the average of the values ​​of the random variable according to the probability.

Definition 8 : Let X be a random variable, if EX 2 < ∞ EX^2<\inftyEX2< , abbreviatedDX = E [ ( X − EX ) 2 ] DX=E[(X-EX)^2]DX=E [ ( XEX)2 ]is the variance of X.
The mathematical expectation of a random variable is the degree to which the value of the random variable deviates from the mean.

Definition 9 : Suppose X, Y are random variables, EX 2 < ∞ EX^2<\inftyEX2<, E Y 2 < ∞ EY^2<\infty EY2< ,base
BXY = E [ ( X − EX ) ( Y − EY ) ] B_{XY}=E[(X-EX)(Y-EY)]BXY=E [ ( XEX)(YE Y ) ]
is the covariance of X and Y, which means
ρ XY = BXYDXDY \rho_{XY}=\frac {B_{XY}}{\sqrt{DX}\sqrt{DY}}rXY=DX DY BXY
ρ XY \rho_{XY}rXYIf it is 0, it means that X and Y are not correlated, and the correlation coefficient indicates the degree of linear correlation between X and Y.

Several properties that need attention:
(1) E ( a X + b Y ) = a EX + b EYE(aX+bY)=aEX+bEYE ( to X+bY)=a E X+b E Y , where a and b are constants;
(2) If X and Y are independent, thenE [ XY ] = EXEYE[XY]=EXEYAnd [ XY ] _=E X E Y ;
(3) If X and Y are independent, thenD ( a X + b Y ) = a 2 DX + b 2 DYD(aX+bY)=a^{2}DX+b^{2}DYD(aX+bY)=a2DX+b2 DY, where a, b are constants

conditional expectation

Let X and Y be discrete random variables. For a given y, if P { Y = y } > 0 P\{Y=y\}>0P { AND=y}>0则称
P { X = x ∣ Y = y } = P { X = x , Y = y } P { Y = y } P\{X=x|Y=y\}=\frac{P\{X =x,Y=y\}}{P\{Y=y\}}P{ X=xY=y}=P { Y = Y }P{ X=x,Y=y}
For a given Y = y Y=yY=conditional probabilityof X at y

给定Y = y Y = yY=y ,the conditional distribution functionis:
F ( x ∣ y ) = P { x ≤ x ∣ Y = y } , x ∈ RF(x|y)=P\{x \leq x| Y = y\} , x \in RF(xy)=P{ xxY=y},xR

And given Y = y Y=yY=y ,the conditional expectationis:
E [ X ∣ Y = y ] = ∫ xd F ( x ∣ y ) = ∑ xx P { X = x ∣ Y = y } E[X|Y=y]= \int xdF(x|y)=\sum_{x}xP\{ X=x|Y=y\}E[XY=y]=xdF(xy)=xxP{ X=xY=y}

If X and Y are continuous random variables, their joint probability density is f ( x , y ) f(x,y)f(x,y ) , then f Y ( y ) > 0 f_Y(y)>0for everythingfY(y)>0by y,byY = y Y=yY=y ,the conditional probability densityis defined as:
f ( x ∣ y ) = f ( x , y ) f Y ( y ) f(x|y)=\frac{f(x,y)}{f_Y(y )}f(xy)=fY(y)f(x,y)

给定Y = y Y = yY=y ,the conditional distribution functionof
X is F ( x ∣ y ) = P { X ≤ x ∣ Y = y } = ∫ − ∞ xf ( u ∣ y ) du F(x|y)=P\{X\leq x| Y=y\}=\int^{x}_{-\infty}f(u|y)duF(xy)=P{ XxY=y}=xf ( u y ) d u

给定Y = y Y = yY=y ,the conditional expectationis defined as
E [ X ∣ Y = y ] = ∫ xd F ( x ∣ y ) = ∫ xf ( x ∣ y ) E[X|Y=y]= \int xdF(x|y )=\int xf(x|y)E[XY=y]=xdF(xy)=xf(xy)

From this it can be seen that the definition is now exactly as in the unconditional case, except that probability is the conditional probability of the event {Y=y}.

E[X|Y=y] is a function of y, where y is a possible value of Y. If the mean value of X is fully considered under the condition of known Y, it is necessary to replace y with Y, then E[X|Y=y] is a function of random variable Y, which is also a random variable, called X under Y conditional expectations .

Properties :
If random variables X and Y are expected to exist, then
EX = E [ E ( X ∣ Y ) ] = ∫ E [ X ∣ Y = y ] d FY ( y ) EX=E[E(X|Y)]= \int E[X|Y=y]dF_Y(y)EX=E [ E ( X Y ) ]=E[XY=y]dFY( y )
If Y is a discrete random variable, the formula can be expressed as
EX = ∑ y E [ X ∣ Y = y ] P { Y = y } EX=\sum _{y}E[X|Y=y ]P\{Y=y\}EX=yE[XY=and ] P { AND=y }
If Y is a continuous random variable with probability densityf ( y ) f(y)f(y)
E X = ∫ − ∞ ∞ E [ X ∣ Y = y ] f ( y ) d y EX=\int_{- \infty}^{\infty} E[X|Y=y]f(y)dy EX=E[XY=y ] f ( y ) d y

Guess you like

Origin blog.csdn.net/daijingxin/article/details/123920094