Conférencier invité

Mathieu Desbrun

Mathieu Desbrun

Inria Saclay - on leave Caltech

 

Dimensionality reduction via geometry processing

A common and oft-observed assumption for high-dimensional datasets is that they sample (possibly with added noise) a low-dimensional manifold embedded in a high-dimensional space. In this situation, Non-linear Dimensional Reduction (NLDR) offers to find a low-dimensional embedding of the data (i.e., unfold/unroll the original dataset in R^d for a small value of d) that creates little to no distortion so as to reduce the  computational complexity of subsequent data processing.

Given the clear geometric premises of this task, we revisit a series of  well-known NLDR tools (such as ISOMAP, Locally-Linear
Embeddings, etc)  from a geometric processing standpoint, and introduce two novel  geometric manifold learning methods, Parallel Transport Unfolding (PTU) and Spectral Affine Kernel Embedding (SAKE). These contributions leverage basic geometry processing notions such as parallel transport  or generalized barycentric coordinates. We show that tapping into discrete differential geometry significantly improves the robustness of  NLDR compared to existing manifold learning algorithms. Somehow surprisingly, we also demonstrate that we can trivially derive from NLDR the first eigen-based shape editing technique, hinting at  the generality of our approach.

Personnes connectées : 13 Vie privée
Chargement...