Manifold Interpretation of PCA and Linear Auto-Encoders
目标是寻找
x 在子空间的一个映射,并保存尽量多的信息
令编码器为
h 是x 的一个低维模拟
解码器为
因为解码器和编码器都是线性的,那么最小化重构误差就是
也就是
对于PCA,
最优的重构误差
其中
x∈RD h∈Rd λi 是协方差矩阵的特征值.
如果协方差的秩是
ICA
Independent Component Analysis
独立成分分析
Herault and Ans, 1984; Jutten and Herault, 1991; Comon, 1994; Hyv¨arinen,1999; Hyv¨arinen et al., 2001
和概率PCA,特征分析相似,它也满足线性特征模型的条件
- sample real-valued factors
h∼P(h) - sample the real-valued observable variables
x=Wh+b+mnoise
其中不同的一点是它不假设先验分布是高斯分布,它只假设是参数化的,例如
P(h)=∏iP(hi)
假如假定隐藏变量是非高斯分布的,那么就可以重现他们。这也是ICA的目的。
Sparse Coding as a Generative Model
一个比较有趣的非高斯分布模型-分布是稀疏的
Student_t prior is
Greedy Layerwise Unsupervised Pre-Training
Greedy
- 不同层没有一起统合起来训练,可能会得到局部最优解Layerwise
- 每次只训练一层,训练第K层的时候保持前面的层保持不变Unsupervised
- 每一层都是无监督学习Pre-Training
- 它只是算法的第一步
Transfer Learning and Domain Adaptation
目标是抽取和利用数据集A的信息来应用到数据集B
譬如,不同的领域的具体评价不同(电影,音乐,书籍的评价),但有些地方是相同的。所以叫 Domain Adaptation
两个例子
Extreme forms of transfer learning
- one-shot learning
- zero-shot learning
- zero-data learning
Manifold Interpretation of PCA and Linear Auto-Encoders
标签(空格分隔): 深度学习 个人兴趣
Look for projections of
x into a subspace that preserves as much as information as possible aboutx
Let the encoder be
h is a low-dimensional representation ofx
Decoder
With liner encoder and decoder, minimizing reconstruction error
means that
and the rows of
In the case of PCA, the rows of
the optimal reconstruction error
Where
D is the dimension ofx d is the dimension ofh λi are the eigenvalues of the convariance.
If the covariance has rank
ICA
Independent Component Analysis
Herault and Ans, 1984; Jutten and Herault, 1991; Comon, 1994; Hyv¨arinen,1999; Hyv¨arinen et al., 2001
Like probabilistic PCA and factor analysis, it also fits the linear factor model of Eqs.
- sample real-valued factors
h∼P(h) - sample the real-valued observable variables
x=Wh+b+mnoise
What is particular about ICA is that unlike PCA and factor analysis it does not assume that the prior is Gaussian. It only assumes that it is factorized, i.e.
P(h)=∏iP(hi)
In this case, if we assume that the latent variables arenon-Gaussian
, then we can recover them, and this is whatICA
is trying to achieve.
Sparse Coding as a Generative Model
A particularly interesting form of non-Gaussianity arises with distributions that are sparse.
Student_t prior is
Greedy Layerwise Unsupervised Pre-Training
Greedy
- the different layers are not jointly trained with respect to a global training objective, which could make the procedure sub-optimalLayerwise
- it proceeds one layer at a time, training the k-layer while keeping the previous ones fixed.Unsupervised
- each layer is trained with an unsupervised representation learning algorithm.Pre-Training
- it should be only a first step before a joint training algorithm is applied to fine-tune all the layers together with respect to a criterion of interest
Transfer Learning and Domain Adaptation
The objective is to take advantage of data from a first setting to extract information that may be useful when learning or even directly making predictions in the second setting.
Two examples
Extreme forms of transfer learning
- one-shot learning
- zero-shot learning
- zero-data learning