Learning in Information - A First Look at the Door

Chapter 1 Content

Information entropy: a measure of the uncertainty of the information selection of the source
Signal: the carrier of the message, representing the physical quantity of the
message Source encoder : data compression, improve effectiveness
Channel coding : add error correction elements, improve reliability, anti-interference Ability Enhancement
Variance Formula :
D(X)=E{[xE(X) 2 ]}=E(X 2 )-E 2 (X)

Chapter 2 Content

Self-Information: I(x i )=-log p(x i )
The definition of self-information reflects the uncertainty of time itself

Small-probability events contain large uncertainties and have a large amount of self-information, while high-probability events are the opposite.

Joint self-information:
I(x i y j )=-log p(x i y j )
If the two events are independent, then I(x i y j )=I(x i )+I(y j )


Conditional Self-Information: Conditional Self-Information I(x i |y j )=-log p(x i |y j )
I(x i y j )=I( xi |y under a given condition of an event j )+I(y j )
I(x 1 x 2 …x n )=I(x 1 )+I(x 2 |x 1 )+…I(x n |x 1 x 2 …x n-1 )
That is, the occurrence of x i will help reduce or increase the uncertainty of y j
itself. Vera graph
insert image description here
Mutual information: the occurrence of event one gives the amount of information about event two
I ( x i ; y j ) = p ( x i ∣ y j ) p ( x i ) = l o g 1 p ( x i ) − l o g 1 p ( x i ∣ y j ) = I ( x i ) − I ( x i ∣ y j ) I(x_i;y_j)=\frac{p(x_i|y_j)}{p(x_i)}=log\frac{1}{p(x_i)}-log\frac{1}{p(x_i|y_j)}=I(x_i)-I(x_i|y_j) I(xi;yj)=p(xi)p(xiyj)=logp(xi)1logp(xiyj)1=I(xi)I(xiyj) indicates the uncertaintyi
when yjoccurs
The amount of mutual information can be positive or negative

The average self-information of a discrete set, also known as information entropy
H ( X ) = − ∑ i = 1 qp ( xi ) logp ( xi ) H(X)=-\sum_{i=1}^{q}p(x_i )log p(x_i)H(X)=i=1qp(xi)logp(xi)

Entropy is symmetrical, the order of each component is changed, and the entropy is constant.
Entropy is non-negative, and the entropy of the determined field is the smallest. If it is 0,
entropy has scalability, and the increase of a minimal probability event, the entropy value can be regarded as constant.
Entropy has an additive value. Properties: H(XY)=H(X)+H(Y|X)
entropy has extreme value, maximum entropy theorem, the entropy of isoprobable field is the largest, that is, H(X)=log n
entropy has upward convexity, H n (p 1 ,p 2 ,…p q ) is a strictly upward convex function of the probability distribution (p 1 ,p 2 ,…p q )

Guess you like

Origin blog.csdn.net/yyfloveqcw/article/details/124237221