Manifold Interpretation of PCA and Linear Auto-Encoders
目標是尋找
x 在子空間的一個映射,並保存儘量多的信息
令編碼器爲
h 是x 的一個低維模擬
解碼器爲
因爲解碼器和編碼器都是線性的,那麼最小化重構誤差就是
也就是
對於PCA,
最優的重構誤差
其中
x∈RD h∈Rd λi 是協方差矩陣的特徵值.
如果協方差的秩是
ICA
Independent Component Analysis
獨立成分分析
Herault and Ans, 1984; Jutten and Herault, 1991; Comon, 1994; Hyv¨arinen,1999; Hyv¨arinen et al., 2001
和概率PCA,特徵分析相似,它也滿足線性特徵模型的條件
- sample real-valued factors
h∼P(h) - sample the real-valued observable variables
x=Wh+b+mnoise
其中不同的一點是它不假設先驗分佈是高斯分佈,它只假設是參數化的,例如
P(h)=∏iP(hi)
假如假定隱藏變量是非高斯分佈的,那麼就可以重現他們。這也是ICA的目的。
Sparse Coding as a Generative Model
一個比較有趣的非高斯分佈模型-分佈是稀疏的
Student_t prior is
Greedy Layerwise Unsupervised Pre-Training
Greedy
- 不同層沒有一起統合起來訓練,可能會得到局部最優解Layerwise
- 每次只訓練一層,訓練第K層的時候保持前面的層保持不變Unsupervised
- 每一層都是無監督學習Pre-Training
- 它只是算法的第一步
Transfer Learning and Domain Adaptation
目標是抽取和利用數據集A的信息來應用到數據集B
譬如,不同的領域的具體評價不同(電影,音樂,書籍的評價),但有些地方是相同的。所以叫 Domain Adaptation
兩個例子
Extreme forms of transfer learning
- one-shot learning
- zero-shot learning
- zero-data learning
Manifold Interpretation of PCA and Linear Auto-Encoders
標籤(空格分隔): 深度學習 個人興趣
Look for projections of
x into a subspace that preserves as much as information as possible aboutx
Let the encoder be
h is a low-dimensional representation ofx
Decoder
With liner encoder and decoder, minimizing reconstruction error
means that
and the rows of
In the case of PCA, the rows of
the optimal reconstruction error
Where
D is the dimension ofx d is the dimension ofh λi are the eigenvalues of the convariance.
If the covariance has rank
ICA
Independent Component Analysis
Herault and Ans, 1984; Jutten and Herault, 1991; Comon, 1994; Hyv¨arinen,1999; Hyv¨arinen et al., 2001
Like probabilistic PCA and factor analysis, it also fits the linear factor model of Eqs.
- sample real-valued factors
h∼P(h) - sample the real-valued observable variables
x=Wh+b+mnoise
What is particular about ICA is that unlike PCA and factor analysis it does not assume that the prior is Gaussian. It only assumes that it is factorized, i.e.
P(h)=∏iP(hi)
In this case, if we assume that the latent variables arenon-Gaussian
, then we can recover them, and this is whatICA
is trying to achieve.
Sparse Coding as a Generative Model
A particularly interesting form of non-Gaussianity arises with distributions that are sparse.
Student_t prior is
Greedy Layerwise Unsupervised Pre-Training
Greedy
- the different layers are not jointly trained with respect to a global training objective, which could make the procedure sub-optimalLayerwise
- it proceeds one layer at a time, training the k-layer while keeping the previous ones fixed.Unsupervised
- each layer is trained with an unsupervised representation learning algorithm.Pre-Training
- it should be only a first step before a joint training algorithm is applied to fine-tune all the layers together with respect to a criterion of interest
Transfer Learning and Domain Adaptation
The objective is to take advantage of data from a first setting to extract information that may be useful when learning or even directly making predictions in the second setting.
Two examples
Extreme forms of transfer learning
- one-shot learning
- zero-shot learning
- zero-data learning