Multi-Tensor Network Representation for High-Order Tensor Completion

09/09/2021
by   Chang Nie, et al.
14

This work studies the problem of high-dimensional data (referred to tensors) completion from partially observed samplings. We consider that a tensor is a superposition of multiple low-rank components. In particular, each component can be represented as multilinear connections over several latent factors and naturally mapped to a specific tensor network (TN) topology. In this paper, we propose a fundamental tensor decomposition (TD) framework: Multi-Tensor Network Representation (MTNR), which can be regarded as a linear combination of a range of TD models, e.g., CANDECOMP/PARAFAC (CP) decomposition, Tensor Train (TT), and Tensor Ring (TR). Specifically, MTNR represents a high-order tensor as the addition of multiple TN models, and the topology of each TN is automatically generated instead of manually pre-designed. For the optimization phase, an adaptive topology learning (ATL) algorithm is presented to obtain latent factors of each TN based on a rank incremental strategy and a projection error measurement strategy. In addition, we theoretically establish the fundamental multilinear operations for the tensors with TN representation, and reveal the structural transformation of MTNR to a single TN. Finally, MTNR is applied to a typical task, tensor completion, and two effective algorithms are proposed for the exact recovery of incomplete data based on the Alternating Least Squares (ALS) scheme and Alternating Direction Method of Multiplier (ADMM) framework. Extensive numerical experiments on synthetic data and real-world datasets demonstrate the effectiveness of MTNR compared with the start-of-the-art methods.

READ FULL TEXT

page 1

page 12

page 14

page 16

page 17

research
05/22/2018

Rank Minimization on Tensor Ring: A New Paradigm in Scalable Tensor Decomposition and Completion

In low-rank tensor completion tasks, due to the underlying multiple larg...
research
07/27/2020

Additive Tensor Decomposition Considering Structural Data Information

Tensor data with rich structural information becomes increasingly import...
research
02/18/2018

Inductive Framework for Multi-Aspect Streaming Tensor Completion with Side Information

Low-rank tensor completion is a well-studied problem and has application...
research
03/04/2023

Tensorized LSSVMs for Multitask Regression

Multitask learning (MTL) can utilize the relatedness between multiple ta...
research
02/16/2023

A weighted subspace exponential kernel for support tensor machines

High-dimensional data in the form of tensors are challenging for kernel ...
research
08/30/2023

Low-Rank Multitask Learning based on Tensorized SVMs and LSSVMs

Multitask learning (MTL) leverages task-relatedness to enhance performan...
research
10/26/2016

Tensor Decompositions for Identifying Directed Graph Topologies and Tracking Dynamic Networks

Directed networks are pervasive both in nature and engineered systems, o...

Please sign up or login with your details

Forgot password? Click here to reset