Isomap
Isomap is a nonlinear dimensionality reduction method. It is one of several widely used low-dimensional embedding methods.[1] Isomap is used for computing a quasi-isometric, low-dimensional embedding of a set of high-dimensional data points. The algorithm provides a simple method for estimating the intrinsic geometry of a data manifold based on a rough estimate of each data point’s neighbors on the manifold. Isomap is highly efficient and generally applicable to a broad range of data sources and dimensionalities.
Introduction
Isomap is one representative of isometric mapping methods, and extends metric multidimensional scaling (MDS) by incorporating the geodesic distances imposed by a weighted graph. To be specific, the classical scaling of metric MDS performs low-dimensional embedding based on the pairwise distance between data points, which is generally measured using straight-line Euclidean distance. Isomap is distinguished by its use of the geodesic distance induced by a neighborhood graph embedded in the classical scaling. This is done to incorporate manifold structure in the resulting embedding. Isomap defines the geodesic distance to be the sum of edge weights along the shortest path between two nodes (computed using Dijkstra's algorithm, for example). The top n eigenvectors of the geodesic distance matrix, represent the coordinates in the new n-dimensional Euclidean space.
Algorithm
A very high-level description of Isomap algorithm is given below.
- Determine the neighbors of each point.
- All points in some fixed radius.
- K nearest neighbors.
- Construct a neighborhood graph.
- Each point is connected to other if it is a K nearest neighbor.
- Edge length equal to Euclidean distance.
- Compute shortest path between two nodes.
- Compute lower-dimensional embedding.
Extensions of ISOMAP
- LandMark ISOMAP (L-ISOMAP): Landmark-Isomap is a variant of Isomap which is faster than Isomap. However, the accuracy of the manifold is compromised by a marginal factor. In this algorithm, n << N landmark points are used out of the total N data points and an nxN matrix of the geodesic distance between each data point to the landmark points is computed. Landmark-MDS (LMDS) is then applied on the matrix to find a Euclidean embedding of all the data points.[2]
- C Isomap : C-Isomap involves magnifying the regions of high density and shrink the regions of low density of data points in the manifold. Edge weights that are maximized in Multi-Dimensional Scaling(MDS) are modified, with everything else remaining unaffected.[2]
- Parallel Transport Unfolding : Replaces the Dijkstra path-based geodesic distance estimates with parallel transport based approximations instead, improving robustness to irregularity and voids in the sampling.[3]
Possible issues
The connectivity of each data point in the neighborhood graph is defined as its nearest k Euclidean neighbors in the high-dimensional space. This step is vulnerable to "short-circuit errors" if k is too large with respect to the manifold structure or if noise in the data moves the points slightly off the manifold.[4] Even a single short-circuit error can alter many entries in the geodesic distance matrix, which in turn can lead to a drastically different (and incorrect) low-dimensional embedding. Conversely, if k is too small, the neighborhood graph may become too sparse to approximate geodesic paths accurately. But improvements have been made to this algorithm to make it work better for sparse and noisy data sets.[5]
Relationship with other methods
Following the connection between the classical scaling and PCA, metric MDS can be interpreted as kernel PCA. In a similar manner, the geodesic distance matrix in Isomap can be viewed as a kernel matrix. The doubly centered geodesic distance matrix K in Isomap is of the form
- [math]\displaystyle{ K = -\frac{1}{2} HD^2 H\, }[/math]
where [math]\displaystyle{ D^2 = D^2_{ij}:=(D_{ij})^2 }[/math] is the elementwise square of the geodesic distance matrix D = [Dij], H is the centering matrix, given by
- [math]\displaystyle{ H = I_n-\frac{1}{N} e_N e^T_N, \quad\text{where }e_N= [1\ \dots\ 1]^T \in \mathbb{R}^N. }[/math]
However, the kernel matrix K is not always positive semidefinite. The main idea for kernel Isomap is to make this K as a Mercer kernel matrix (that is positive semidefinite) using a constant-shifting method, in order to relate it to kernel PCA such that the generalization property naturally emerges .[6]
See also
References
- ↑ 1.0 1.1 Tenenbaum, Joshua B.; Silva, Vin de; Langford, John C. (22 December 2000). "A Global Geometric Framework for Nonlinear Dimensionality Reduction". Science 290 (5500): 2319–2323. doi:10.1126/science.290.5500.2319.
- ↑ 2.0 2.1 Silva, Vin; Tenenbaum, Joshua (2002). "Global Versus Local Methods in Nonlinear Dimensionality Reduction". Advances in Neural Information Processing Systems (MIT Press) 15. https://proceedings.neurips.cc/paper/2002/hash/5d6646aad9bcc0be55b2c82f69750387-Abstract.html.
- ↑ Budninskiy, Max; Yin, Gloria; Feng, Leman; Tong, Yiying; Desbrun, Mathieu (2019). "Parallel Transport Unfolding: A Connection-Based Manifold Learning Approach" (in en). SIAM Journal on Applied Algebra and Geometry 3 (2): 266–291. doi:10.1137/18M1196133. ISSN 2470-6566. https://epubs.siam.org/doi/10.1137/18M1196133.
- ↑ M. Balasubramanian, E. L. Schwartz, The Isomap Algorithm and Topological Stability. Science 4 January 2002: Vol. 295 no. 5552 p. 7
- ↑ A. Saxena, A. Gupta and A. Mukerjee. Non-linear dimensionality reduction by locally linear Isomaps, . Lecture Notes in Computer Science, 3316:1038–1043, 2004.
- ↑ H. Choi, S. Choi, Robust Kernel Isomap, Pattern Recognition, Vol. 40, No. 3, pp. 853-862, 2007
External links
Original source: https://en.wikipedia.org/wiki/Isomap.
Read more |