Disentangling Patterns and Transformations from One Sequence of Images with Shape-invariant Lie Group Transformer

03/21/2022
by   T. Takada, et al.
0

An effective way to model the complex real world is to view the world as a composition of basic components of objects and transformations. Although humans through development understand the compositionality of the real world, it is extremely difficult to equip robots with such a learning mechanism. In recent years, there has been significant research on autonomously learning representations of the world using the deep learning; however, most studies have taken a statistical approach, which requires a large number of training data. Contrary to such existing methods, we take a novel algebraic approach for representation learning based on a simpler and more intuitive formulation that the observed world is the combination of multiple independent patterns and transformations that are invariant to the shape of patterns. Since the shape of patterns can be viewed as the invariant features against symmetric transformations such as translation or rotation, we can expect that the patterns can naturally be extracted by expressing transformations with symmetric Lie group transformers and attempting to reconstruct the scene with them. Based on this idea, we propose a model that disentangles the scenes into the minimum number of basic components of patterns and Lie transformations from only one sequence of images, by introducing the learnable shape-invariant Lie group transformers as transformation components. Experiments show that given one sequence of images in which two objects are moving independently, the proposed model can discover the hidden distinct objects and multiple shape-invariant transformations that constitute the scenes.

READ FULL TEXT

page 1

page 4

page 5

page 6

research
07/04/2023

Learning Lie Group Symmetry Transformations with Neural Networks

The problem of detecting and quantifying the presence of symmetries in d...
research
10/19/2020

Learning to Reconstruct and Segment 3D Objects

To endow machines with the ability to perceive the real-world in a three...
research
11/16/2019

AETv2: AutoEncoding Transformations for Self-Supervised Representation Learning by Minimizing Geodesic Distances in Lie Groups

Self-supervised learning by predicting transformations has demonstrated ...
research
05/31/2017

Naturally Combined Shape-Color Moment Invariants under Affine Transformations

We proposed a kind of naturally combined shape-color affine moment invar...
research
07/12/2018

HyperNets and their application to learning spatial transformations

In this paper we propose a conceptual framework for higher-order artific...
research
05/17/2021

Construction of Diffeomorphisms with Prescribed Jacobian Determinant and Curl

The variational principle (VP) is designed to generate non-folding grids...
research
03/30/2021

Beltrami Signature: A Novel Invariant 2D Shape Representation for Object Classification

There is a growing interest in shape analysis in recent years and in thi...

Please sign up or login with your details

Forgot password? Click here to reset