Ambition

Math Goal
By the end of the summer I want to have read all of the following papers and have an outline of my own original contribution.
Foundational & General Approaches
- 
A Riemannian Framework for Optimal Transport
- Authors: M. Arjovsky, A. Doucet, et al.
 - Develops a geometry-aware framework for OT over manifolds, making PCA-like generalizations more natural in curved spaces.
 
 - 
Learning Generative Models with Sinkhorn Divergences
- Authors: Genevay et al.
 - While not PCA-specific, this introduces latent representations learned with OT-based losses—relevant for embedding/latent space OT.
 
 
PCA-Like and Subspace Methods with OT
- 
Wasserstein Principal Geodesic Analysis
- Authors: Seguy, Cuturi
 - Generalizes PCA to the Wasserstein space using geodesic analysis, important for structured data like distributions.
 
 - 
Wasserstein Principal Component Analysis: Sparse Optimal Transport Based Dimensionality Reduction
- Authors: Wang et al.
 - An explicit method for Wasserstein PCA, adapted for probability distributions rather than Euclidean vectors.
 
 - 
A New Formulation of Principal Geodesic Analysis in the Wasserstein Space
- Authors: Bigot et al.
 - Focuses on computing principal components in Wasserstein space more efficiently, relevant for shape and distributional data.
 
 
Embedded Manifolds & Latent OT
- 
Learning Optimal Transport Maps using Generative Adversarial Networks
- Introduces learning OT in latent/embedded spaces, allowing for manifold-constrained transport.
 
 - 
Learning Optimal Transport for Domain Adaptation
- Authors: Damodaran et al.
 - Uses PCA for latent dimensionality reduction, then applies OT—an instance of embedded OT.
 
 - 
- Proposes OT over non-Euclidean geometries (e.g., spheres), and matching distributions on these curved spaces.
 
 
Autoencoders, Latent Space + OT
- 
Sliced-Wasserstein Autoencoders
- Uses Sliced-Wasserstein distance in a latent representation learning setting.
 
 - 
Autoencoding Probabilistic PCA with Optimal Transport
- Combines PCA, probabilistic models, and OT regularization.
 
 
Other Notable Contributions
- 
Wasserstein Dictionary Learning
- Extends PCA to the OT setting via sparse dictionary learning over probability distributions.
 
 - 
Sliced-Wasserstein Flow: Nonparametric Generative Modeling via Optimal Transport and Diffusions
- Uses diffusion + OT to generate data in embedded spaces, related to low-dimensional transport learning.
 
 
Summary of Concepts
| Concept | Relation to OT | 
| ------------------------- | ----------------------------------------- | 
| Wasserstein PCA | PCA in probability/measure space | 
| Geodesic PCA (PGA) | PCA generalized to curved OT geometry | 
| Latent OT / Embedded OT | OT after PCA or in learned subspaces | 
| Sliced Wasserstein | Efficient approximation of OT for high-D | 
| Manifold OT | Optimal transport on Riemannian manifolds | 
| Autoencoding + OT | Embedding generation with OT-based loss |