This article has participated in the "New Talent Creation Ceremony" event
foreword
Gradient ascent principal component analysis is also a commonly used operation for dimensionality reduction and noise reduction. It transforms the coordinate axis (actually the mapping of the data) to maximize the direct variance of the data, that is, it is easier to divide
Analysis of purpose and principle
The principal component feature vector is obtained. In the gradient ascent principal component analysis, the original data can be mapped to the coordinate axis with greater variance (some books or videos also say moving the coordinate axis), and the mapping method can be multiplied by a The unit vector of the mapping direction.
preprocessing
In order to facilitate the operation, the data must be preprocessed first. The preprocessing here is mean zeroing, that is, subtracting the mean from all the data.
Formula derivation and explanation
Target
What we want to get is the data with the largest variance, namely: Because we have preprocessed the data. So our goal becomes:
Mapping relationship derivation
Suppose xi is to be mapped to the new axis 在这里我们假设向量w为单位向量 则使之变为:
将映射关系带入
带入映射关系后我们的目标变为了
梯度上升
构建一个函数
现在我们构建一个函数
求导
下面我们开始对f(w)求导 我们的问题就变成了使用上述式子进行梯度上升,接下来我们使用梯度上升的方式优化w,优化得到的w即为可以将数据转换到方差更大的坐标轴上的主成分