Nonlinear Dimension Reduction Presenter Xingwei Yang The powerpoint
- Slides: 20
Nonlinear Dimension Reduction Presenter: Xingwei Yang The powerpoint is organized from: 1. Ronald R. Coifman et al. (Yale University) 2. Jieping Ye, (Arizona State University)
Motivation Linear projections will not detect the pattern.
Nonlinear PCA using Kernels n Traditional PCA applies linear transformation ¨ May not be effective for nonlinear data n Solution: apply nonlinear transformation to potentially very highdimensional space. n Computational efficiency: apply the kernel trick. ¨ Require PCA can be rewritten in terms of dot product. More on kernels later
Nonlinear PCA using Kernels Rewrite PCA in terms of dot product The covariance matrix S can be written as Let v be The eigenvector of S corresponding to nonzero eigenvalue Eigenvectors of S lie in the space spanned by all data points.
Nonlinear PCA using Kernels The covariance matrix can be written in matrix form: Any benefits?
Nonlinear PCA using Kernels Next consider the feature space: The (i, j)-th entry of Apply the kernel trick: K is called the kernel matrix. is
Nonlinear PCA using Kernels n Projection of a test point x onto v: Explicit mapping is not required here.
8/14 Diffusion distance and Diffusion map • A symmetric matrix Ms can be derived from M as • M and Ms has same N eigenvalues, f : left eigenvector of M y : right eigenvector of M • Under random walk representation of the graph M e : time step
Diffusion distance and Diffusion map • e has the dual representation (time step and kernel width). • If one starts random walk from location xi , the probability of landing in location y after r time steps is given by where ei is a row vector with all zeros except that ith position = 1. • For large e, all points in the graph are connected (Mi, j >0) and the eigenvalues of M
Diffusion distance and Diffusion map • One can show that regardless of starting point xi Left eigenvector of M with eigenvalue l 0=1 with • Eigenvector f 0(x) has the dual representation : 1. Stationary probability distribution on the curve, i. e. , the probability of landing at location x after taking infinite steps of random walk (independent of the start location). 2. It is the density estimate at location x.
Diffusion distance • For any finite time r, • yk and fk are the right and left eigenvectors of graph Laplacian M. • is the kth eigenvalue of M r (arranged in descending order). • Given the definition of random walk, we denote Diffusion distance as a distance measure at time t between two pmfs as with empirical choice w(y)=1/f 0(y).
Diffusion Map • Diffusion distance : • Diffusion map : Mapping between original space and first k eigenvectors as Relationship : • This relationship justifies using Euclidean distance in diffusion map space for spectral clustering. • Since , it is justified to stop at appropriate k with a negligible error of order O(lk+1/lk)t).
Example: Hourglass
Example: Image imbedding
Example: Lip image
Shape description
Dimension Reduction of Shape space
Dimension Reduction of Shape space
Dimension Reduction of Shape space
References Unsupervised Learning of Shape Manifolds (BMVC 2007) n Diffusion Maps(Appl. Comput. Harmon. Anal. 21 (2006)) n Geometric diffusions for the analysis of data from sensor networks (Current Opinion in Neurobiology 2005) n
- Xingwei yang
- Réduction de dimension
- Hát kết hợp bộ gõ cơ thể
- Frameset trong html5
- Bổ thể
- Tỉ lệ cơ thể trẻ em
- Voi kéo gỗ như thế nào
- Glasgow thang điểm
- Chúa yêu trần thế
- Môn thể thao bắt đầu bằng từ đua
- Thế nào là hệ số cao nhất
- Các châu lục và đại dương trên thế giới
- Công của trọng lực
- Trời xanh đây là của chúng ta thể thơ
- Cách giải mật thư tọa độ
- Làm thế nào để 102-1=99
- Phản ứng thế ankan
- Các châu lục và đại dương trên thế giới
- Thể thơ truyền thống
- Quá trình desamine hóa có thể tạo ra
- Một số thể thơ truyền thống