Principal Component Analysis ---- PRML读书笔记

    To summarize, principal component analysis involves evaluating the mean x and the covariance matrix S 

of the data set and then finding the M eigenvectors of S corresponding to the M largest eigenvalues. If we

plan to project our data onto the first M principal compents, then we only need to find the first M eigenvalues

and eigenvectors.

    PCA can be defined as the orthogonal projection of the data onto a lower dimensional linear space, known as

the principal subspace, such that the variance of the projected data is maximized. Equivalently, it can be defined 

as the linear projection that minimizes the average projection cost, defined as the mean squared distance between

the data points and their projections.

    Consider a data set of observations {xn} where n = 1,...,N, and xn is a Euclidean variable with dimensionality D.

Our goal is to project the data onto a space having dimensionality M < D while maximizing the variance of the projected

data. 

    The general solution to the minimization of J for arbitrary D and arbitrary M < D is obtained by choosing the {ui} to be

eigenvectors of the covariance matrix given by Suiiui. where i=1,...,D, and as usual the eigenvectors {ui} are chosen to 

be orthonormal.

posted @   东宫得臣  阅读(147)  评论(0编辑  收藏  举报
编辑推荐:
· AI与.NET技术实操系列:基于图像分类模型对图像进行分类
· go语言实现终端里的倒计时
· 如何编写易于单元测试的代码
· 10年+ .NET Coder 心语,封装的思维:从隐藏、稳定开始理解其本质意义
· .NET Core 中如何实现缓存的预热?
阅读排行:
· 分享一个免费、快速、无限量使用的满血 DeepSeek R1 模型,支持深度思考和联网搜索!
· 基于 Docker 搭建 FRP 内网穿透开源项目(很简单哒)
· 25岁的心里话
· ollama系列01:轻松3步本地部署deepseek,普通电脑可用
· 按钮权限的设计及实现
点击右上角即可分享
微信分享提示