Back to Search
Start Over
MANGO: Disentangled Image Transformation Manifolds with Grouped Operators
- Publication Year :
- 2024
-
Abstract
- Learning semantically meaningful image transformations (i.e. rotation, thickness, blur) directly from examples can be a challenging task. Recently, the Manifold Autoencoder (MAE) proposed using a set of Lie group operators to learn image transformations directly from examples. However, this approach has limitations, as the learned operators are not guaranteed to be disentangled and the training routine is prohibitively expensive when scaling up the model. To address these limitations, we propose MANGO (transformation Manifolds with Grouped Operators) for learning disentangled operators that describe image transformations in distinct latent subspaces. Moreover, our approach allows practitioners the ability to define which transformations they aim to model, thus improving the semantic meaning of the learned operators. Through our experiments, we demonstrate that MANGO enables composition of image transformations and introduces a one-phase training routine that leads to a 100x speedup over prior works.<br />Comment: Submitted to IEEE ICASSP 2025. This work has been submitted to the IEEE for possible publication
Details
- Database :
- arXiv
- Publication Type :
- Report
- Accession number :
- edsarx.2409.09542
- Document Type :
- Working Paper