[Statistical Learning|Book Reading] Chapter 4 Naive Bayes p47-p54

train of thought

Naive Bayesian method is a classification method based on Bayesian theorem and the independent assumption of feature conditions. For a given training data set, firstly learn the joint probability distribution of input/output based on the independent assumptions of the characteristic feature conditions; then based on this model, for a given input, use Bayes theorem to find the output y with the largest posterior probability.

Naive Bayesian Learning and Classification

Naive Bayesian learning strategy:
y = f ( x ) = arg max ⁡ ck P ( Y = ck ) ∏ j P ( X ( j ) = x ( j ) ∣ Y = ck ) y=f(x)= arg\max_{c_k} P(Y=c_k)\prod_{j}P(X^{(j)}=x^{(j)}|Y=c_k)y=f(x)=argckmaxP ( Y)=ck)jP(X(j)=x(j)Y=ck)

Posterior probability maximization:
f ( x ) = arg max ⁡ ck P ( ck | X = x ) f(x)=arg\max_{c_k} P(c_k|X=x)f(x)=argckmaxP(ckX=x)

Parameter Estimation Algorithm of Naive Bayes Method

maximum likelihood estimation

insert image description here

Bayesian estimation

insert image description here

Guess you like

Origin blog.csdn.net/m0_52427832/article/details/127046683