Difference between revisions of "Orange: Manifold Learning"
Onnowpurbo (talk | contribs) (Created page with "Sumber: https://docs.biolab.si//3/visual-programming/widgets/unsupervised/manifoldlearning.html Nonlinear dimensionality reduction. Inputs Data: input dataset Output...") |
Onnowpurbo (talk | contribs) |
||
Line 15: | Line 15: | ||
Manifold Learning is a technique which finds a non-linear manifold within the higher-dimensional space. The widget then outputs new coordinates which correspond to a two-dimensional space. Such data can be later visualized with Scatter Plot or other visualization widgets. | Manifold Learning is a technique which finds a non-linear manifold within the higher-dimensional space. The widget then outputs new coordinates which correspond to a two-dimensional space. Such data can be later visualized with Scatter Plot or other visualization widgets. | ||
− | + | [[File:Manifold-learning-stamped.png|center|200px|thumb]] | |
+ | |||
Method for manifold learning: | Method for manifold learning: | ||
− | |||
t-SNE | t-SNE | ||
− | |||
MDS, see also MDS widget | MDS, see also MDS widget | ||
− | |||
Isomap | Isomap | ||
− | |||
Locally Linear Embedding | Locally Linear Embedding | ||
− | |||
Spectral Embedding | Spectral Embedding | ||
− | |||
Set parameters for the method: | Set parameters for the method: | ||
− | |||
t-SNE (distance measures): | t-SNE (distance measures): | ||
− | |||
Euclidean distance | Euclidean distance | ||
− | |||
Manhattan | Manhattan | ||
− | |||
Chebyshev | Chebyshev | ||
− | |||
Jaccard | Jaccard | ||
− | |||
Mahalanobis | Mahalanobis | ||
− | |||
Cosine | Cosine | ||
− | |||
MDS (iterations and initialization): | MDS (iterations and initialization): | ||
− | |||
max iterations: maximum number of optimization interactions | max iterations: maximum number of optimization interactions | ||
− | |||
initialization: method for initialization of the algorithm (PCA or random) | initialization: method for initialization of the algorithm (PCA or random) | ||
− | |||
Isomap: | Isomap: | ||
− | |||
number of neighbors | number of neighbors | ||
− | |||
Locally Linear Embedding: | Locally Linear Embedding: | ||
− | |||
method: | method: | ||
− | |||
standard | standard | ||
− | |||
modified | modified | ||
− | |||
hessian eigenmap | hessian eigenmap | ||
− | |||
local | local | ||
− | |||
number of neighbors | number of neighbors | ||
− | |||
max iterations | max iterations | ||
− | |||
Spectral Embedding: | Spectral Embedding: | ||
− | |||
affinity: | affinity: | ||
− | |||
nearest neighbors | nearest neighbors | ||
− | |||
RFB kernel | RFB kernel | ||
− | |||
Output: the number of reduced features (components). | Output: the number of reduced features (components). | ||
− | |||
If Apply automatically is ticked, changes will be propagated automatically. Alternatively, click Apply. | If Apply automatically is ticked, changes will be propagated automatically. Alternatively, click Apply. | ||
− | |||
Produce a report. | Produce a report. | ||
Manifold Learning widget produces different embeddings for high-dimensional data. | Manifold Learning widget produces different embeddings for high-dimensional data. | ||
− | + | [[File:Collage-manifold.png|center|200px|thumb]] | |
From left to right, top to bottom: t-SNE, MDS, Isomap, Locally Linear Embedding and Spectral Embedding. | From left to right, top to bottom: t-SNE, MDS, Isomap, Locally Linear Embedding and Spectral Embedding. | ||
− | + | ||
+ | ==Contoh== | ||
Manifold Learning widget transforms high-dimensional data into a lower dimensional approximation. This makes it great for visualizing datasets with many features. We used voting.tab to map 16-dimensional data onto a 2D graph. Then we used Scatter Plot to plot the embeddings. | Manifold Learning widget transforms high-dimensional data into a lower dimensional approximation. This makes it great for visualizing datasets with many features. We used voting.tab to map 16-dimensional data onto a 2D graph. Then we used Scatter Plot to plot the embeddings. | ||
− | + | [[File:Manifold-learning-example.png|center|200px|thumb]] | |
− | |||
Revision as of 09:22, 24 January 2020
Sumber: https://docs.biolab.si//3/visual-programming/widgets/unsupervised/manifoldlearning.html
Nonlinear dimensionality reduction.
Inputs
Data: input dataset
Outputs
Transformed Data: dataset with reduced coordinates
Manifold Learning is a technique which finds a non-linear manifold within the higher-dimensional space. The widget then outputs new coordinates which correspond to a two-dimensional space. Such data can be later visualized with Scatter Plot or other visualization widgets.
Method for manifold learning: t-SNE MDS, see also MDS widget Isomap Locally Linear Embedding Spectral Embedding Set parameters for the method: t-SNE (distance measures): Euclidean distance Manhattan Chebyshev Jaccard Mahalanobis Cosine MDS (iterations and initialization): max iterations: maximum number of optimization interactions initialization: method for initialization of the algorithm (PCA or random) Isomap: number of neighbors Locally Linear Embedding: method: standard modified hessian eigenmap local number of neighbors max iterations Spectral Embedding: affinity: nearest neighbors RFB kernel Output: the number of reduced features (components). If Apply automatically is ticked, changes will be propagated automatically. Alternatively, click Apply. Produce a report.
Manifold Learning widget produces different embeddings for high-dimensional data.
From left to right, top to bottom: t-SNE, MDS, Isomap, Locally Linear Embedding and Spectral Embedding.
Contoh
Manifold Learning widget transforms high-dimensional data into a lower dimensional approximation. This makes it great for visualizing datasets with many features. We used voting.tab to map 16-dimensional data onto a 2D graph. Then we used Scatter Plot to plot the embeddings.