Abstract
Using the orthogonal rank of the tensor, a new estimation method for the upper bounds on the nuclear norms is presented and some new tight upper bounds on the nuclear norms are established. Taking into account the structure information of the tensor, an important factor affecting the upper bounds is discussed and some corresponding properties related to the nuclear norms are given. Meanwhile, some new results of the upper bounds on the nuclear norms are obtained.
Keywords: Tensor, Nuclear norm, Orthogonal rank, Upper bound
Introduction
A tensor is a multidimensional array that can provide a natural and convenient way for representing the multidimensional data such as discrete forms of multivariate functions, images, video sequences, and so on [1, 2, 10, 14]. With the successful and widespread use of the matrix nuclear norm (a sum of the singular values) in the information recovery, the research on the nuclear norm of the tensor (see Definition 2.3 in Sect. 2) has been a hot topic in both the theory and applications [5–8].
A natural problem is how to compute the nuclear norm of a tensor. Unfortunately, compared with the matrix nuclear norm, the nuclear norm of a tensor is closely related to the number field, and the computation of the tensor nuclear norm is NP-hard [5]. Thus, exploring some simple polynomial-time computable upper bounds on the nuclear norm is very important.
Relating to the nuclear norm of a tensor, Friedland and Lim established the following upper bound through the Frobenius norm of this tensor in [5].
Theorem 1.1
([5])
Let . Then
In [8], Hu established a tighter upper bound.
Theorem 1.2
(Lemma 5.1 in [8])
Let . Then
| 1 |
Furthermore, Hu established another upper bound on the nuclear norm of a tensor through the nuclear norms of its unfolding matrices.
Theorem 1.3
(Theorem 5.2 in [8])
Let . Then
| 2 |
In this paper, we present some new upper bounds on the nuclear norms through using the orthogonal rank of the tensor [9, 12]. Furthermore, taking into account the structure information of the tensor, some new results of the upper bounds on the nuclear norms are obtained.
Since the spectral norm and nuclear norm of a tensor are closely related to the number field [6], for the sake of simplicity, we always assume that the tensors discussed are nonzero real tensors, and unless mentioned otherwise, we just discuss the spectral and the nuclear norm over the real field. Some corresponding notations are as follows: a tensor is denoted by the calligraphic letter (e.g., ), while the scalar is denoted by the plain letter, and matrices and vectors are denoted by bold letters (e.g., X and x).
The rest of the paper is organized as follows. In Sect. 2, we recall some definitions and related results which are needed for the subsequent sections. In Sect. 3, we present the upper bounds on the nuclear norms of general tensors and discuss the factor affecting the upper bounds. Finally, some conclusions are made in Sect. 4.
Notations and preliminaries
This section is devoted to reviewing some conceptions and results related to tensors, which are needed for the following sections.
Firstly, we discuss the unfolding matrix or matrix representation of a tensor.
Let . Then, by organizing several indexes of and the remaining indexes of as the row index and the column index, respectively, the tensor can be reshaped into a matrix form [13]. Especially, if the number of row index is equal to one, then we get the mode-d matricization , the columns of which are the mode-d fibers of the tensor (obtained by fixing every coordinate except one, ), arranged in a cyclic ordering; see [3] for details. Contrary to the operation above, a matrix can also be reshaped into a tensor by using the opposite operation.
In what follows, we review some definitions of the tensor norms.
Definition 2.1
([3])
Let , the Frobenius norm or Hilbert–Schmidt norm of the tensor is defined as
Definition 2.2
([6])
Let “∘” denote the outer product operation. The spectral norm of is defined as
Furthermore, is equal to the Frobenius norm of the best rank-one approximation of the tensor .
Similar to the matrix case, the nuclear norm can be defined through the dual norm of the spectral norm.
Definition 2.3
([6])
Let , the nuclear norm of is defined as the dual norm of the spectral norm. That is,
| 3 |
For the nuclear norm defined by (3), it can be shown that
Another important concept related to the matrix and the tensor is the mode-d multiplication.
Definition 2.4
([3])
Let . Then the mode-d multiplication of by the matrix is defined by
It should be mentioned that the mode-d multiplication is also available for .
Furthermore, let
If for all the matrices are orthogonal matrices ( is an identity matrix), then is called a multi-linear orthogonal transformation of the tensor .
Finally in this section, we introduce the tool used in the paper for the estimation of the upper bounds.
Definition 2.5
([9])
The orthogonal rank of is defined as the smallest number R such that
| 4 |
where () are rank-one tensors such that , () for and .
The decomposition of given by (4) is also called the orthogonal decomposition of the tensor .
For the orthogonal rank of a tensor, the following conclusion is true.
Theorem 2.1
([11])
Let . Then, for any , it holds
where denotes the orthogonal rank of .
Noting the fact that indices relabeling does not change the tensor orthogonal rank, then by Theorem 2.1, we have that, for any , it holds
| 5 |
Especially, for the third order tensor, the following result was established in [11].
Lemma 2.1
([11])
Let . Then, for any , the following holds:
Upper bounds of the nuclear norm
In this section, we discuss the upper bounds on the nuclear norm of a tensor. Meanwhile, some properties and polynomial-time computable bounds related to the nuclear norm will be given.
Upper bounds given by the Frobenius norm
In this subsection, we use the orthogonal rank of a general tensor to establish the upper bounds on the nuclear norm through the Frobenius norm of this tensor.
Theorem 3.1
Let . Suppose that
Then
| 6 |
Proof
Let be an arbitrary nonzero tensor and the orthogonal rank of be . Suppose that
is the orthogonal rank decomposition of the tensor , where .
Then, according to the properties of the orthogonal rank decomposition and the best rank-one approximation, we have
| 7 |
and
| 8 |
Without loss of generality, suppose that
Then it follows from (7) and (8) that
| 9 |
Thus, according to the arbitrariness of and (9), we get
Noting the definition of the nuclear norm (Definition 2.3), the conclusion is established. □
Remark 3.1
Comparing the upper bound given by (6) with the upper bound given by (1), which is obtained in [8], the new upper bound given by (6) is tighter.
Actually, it follows from (5) that the upper bound given by (6) improves the upper bound given by (1).
More specifically, we present a simple example to show that the upper bound given by Theorem 3.1 not only can be tighter than the upper bound given by (1) but also a sharp bound.
Example 3.1
Let
By Theorem 3.1 and Lemma 2.1, we get
This means that the upper bound given by (3.1) is tighter than the upper bound given by (1).
Furthermore, by a simple computation, we get . Then it follows from the definition of the nuclear norm that
Thus, it holds
Actually,
| 10 |
is a nuclear decomposition of , where , , , , and . Since
then, according to the sufficient and necessary conditions of the nuclear norm decomposition obtained in [6], we get that (10) is a nuclear decomposition of .
This also means that the upper bound given by Theorem 3.1 is a sharp upper bound of the nuclear norm.
Upper bounds given by nuclear norms of the unfolding matrices of a tensor
In this subsection, we present a new way to establish the upper bounds on the nuclear norm of a tensor through the nuclear norms of the unfolding matrices of this tensor.
Theorem 3.2
Let . Suppose that
Then
| 11 |
Proof
Let the singular value decomposition of the matrix be
| 12 |
where , , , , , and .
Then equality (12) can be expressed as the following form:
| 13 |
where are obtained by reordering the vector into th order tensor with a certain order, . Suppose that the orthogonal rank decomposition of is
Then, by taking the expression of into the right-hand side of (13), we get
| 14 |
Noting that
| 15 |
where , , then it follows from the definition of the nuclear norm and (14) that
□
Remark 3.2
Comparing the upper bound given by (11) with the upper bound given by (2), which is obtained in [8], the new upper bound given by (11) is smaller.
Actually, it follows from inequality (5) that the upper bound given by (11) improves the upper bound given by (2).
Similar to the discussion of Hu [8], the upper bounds can also be obtained by other unfolding ways and further improved by considering the multi-linear ranks of a tensor (ranks of the unfolding matrices).
Corollary 3.1
Let and , . Then
Proof
According to the conditions of the corollary and the higher order singular value decomposition of the tensor [3], the tensor can be expressed as
where and satisfying that is an identity matrix for all .
By the definition of the tensor nuclear norm (Definition 2.3), one can easily verify the following conclusions:
| 16 |
and
| 17 |
It follows from Theorem 3.2 and (17) that
Noting (16), we get
Similarly, we have
and
Thus, the conclusion is obtained. □
Factors affecting the upper bounds on the nuclear norm and further results
In this subsection, we discuss the factors affecting the nuclear norm of a tensor. Especially, we focus on the structure analysis of a tensor. Based on the discussion, some new upper bounds on the nuclear norms of tensors are presented.
Firstly, we give a simple example to illustrate that the nuclear norm of a tensor is closely related to the structure of this tensor.
Example 3.2
Let
| 18 |
Similar to the discussion of Example 3.1, since
and
then, according to the sufficient and necessary conditions of the nuclear norm decomposition obtained in [6], we get
It is well known that the nuclear norm of a tensor is closely related to the number field [6]. Actually, the tensor can be expressed as the following form:
Let
Then it holds
| 19 |
For the sake of convenience, let . Then, using the same method as above, we have
Thus . Since all three matrices () are unitary matrices, based on the invariance of the Frobenius norm of a tensor under the multi-linear orthogonal transformations, we get
Noting the structures of the tensors given by (18) and (19), the above derivation process implies that the nuclear norm of a tensor is closely related to the structure of this tensor. In what follows, we discuss the block diagonal tensor, which can be illustrated by Fig. 1. Furthermore, the block diagonal tensor can be expressed by using the direct sum operation “⊕” [4], which is defined as follows:
Figure 1.

The block diagonal tensor with three diagonal blocks
Let and , then the direct sum of and is an order-D tensor defined by
Based on the discussion above, we present some properties of the spectral norm and nuclear norm of the tensor.
Lemma 3.1
Let , , and
Then
| 20 |
Proof
According to the definition of the spectral norm of a tensor (Definition 2.3), it is easy to get
Thus, the rest of the proof just needs to show
| 21 |
Firstly, we consider the case of the third order tensors.
Suppose that (),
and is the best rank-one approximation of , where , , , , (), and . Then the following matrix
is a block diagonal matrix. It follows
Hence, inequality (21) is proved. This also implies that equality (20) is true for the third order tensors.
Secondly, for the case of higher order tensors with order larger than or equal to four, the same result can be established by the recursive method.
In all, the conclusion is true. □
Then, based on Lemma 3.1, the following two results related to the nuclear norms of tensors can be established.
Lemma 3.2
Let be a zero tensor, and . Then
Proof
Suppose that
and
where for all , are rank-one tensors with . Then
where all 0s denote zero vectors with suitable dimensions. This implies
Furthermore, assume that
where and .
Then, by Lemma 3.1, we have
It follows from Definition 2.5 that
Thus it holds .
Using the same method, the equality can be proved. □
Lemma 3.3
Let , , and
Then
Proof
We just need to prove the case of . For the general case, the conclusion can be obtained in a recursive way.
Let
where and are both zero tensors.
Then, by using Lemma 3.2, we get
| 22 |
Suppose that
Then, by Lemma 3.1, we get
Thus, according to Definition 2.5, we have
| 23 |
Combined (22) with (23), the results can be obtained. □
Based on the fact that the nuclear norm of a tensor is also kept invariant under the multi-linear orthogonal transformation, we get the following result.
Corollary 3.2
Let . If the tensor admits a diagonal structure under the multi-linear orthogonal transformations, then
where () are the diagonal elements and .
This case presented by Corollary 3.2 is consistent with the definition of the nuclear norm of the matrix case, and in this case, the nuclear norm of the tensor can be accurately calculated.
Taking into account the structure information of the tensor, some new results of the upper bounds on the nuclear norms can be obtained. For the convenience of comparison, we just present the upper bounds on the nuclear norms of tensors through the dimensions of the tensors, without considering the orthogonal rank of the tensors.
Theorem 3.3
Let and L be the maximum number of diagonal blocks that the tensor can attain under the multi-linear orthogonal transformations. Suppose that the size of each diagonal block is and
Then it holds
| 24 |
Proof
Assume that
where
and () are orthogonal matrices, , .
Then it follows from the invariance of the Frobenius norm of a tensor under the multi-linear orthogonal transformation that
| 25 |
Furthermore, since the nuclear norm of a tensor is also kept invariant under the multi-linear orthogonal transformation, we get
Hence, by Lemma 3.3 and (25), we get
□
Without loss of generality, suppose that
Since
it is easy to get
Thus, the upper bound given by (24) improves (1). Theorem 3.3 also shows that the upper bound on the nuclear norm can be improved by using the structural information.
Similarly, the following upper bound can also be obtained.
Theorem 3.4
Let , and L be the maximum number of diagonal blocks that the tensor can attain under the multi-linear orthogonal transformations. Suppose that the size of each diagonal block is , and
and
Then it holds
| 26 |
Proof
Similar to the proof of Theorem 3.3, assume
where
and () are orthogonal matrices.
Then it holds
□
Theorem 3.4 implies that the upper bound given by nuclear norms of the unfolding matrices is more closely related to the structure. For the sake of clarity, we give a simple example to illustrate.
Example 3.3
Let be defined in Example 3.2, and
Then, by Theorem 1.2, we get
It follows from Theorem 3.4 that
There has been a marked improvement in the upper bounds on the nuclear norm.
Conclusions
In this paper, we provide a new estimation method for the upper bounds on the nuclear norms and obtain some new upper bounds related to the nuclear norms. Meanwhile, it is found that the upper bounds on the nuclear norms are not only related to the dimensions of the tensor but also to the structure of the tensor. Taking into consideration the structure information of the tensor, the upper bounds on the nuclear norms can be improved.
Authors’ contributions
All three authors contributed equally to this work. All authors read and approved the final manuscript.
Funding
This research work was supported by the Natural Science Foundation of China (NSFC) (Nos: 11401286, 11671318, 11401472), Natural Science Foundation of Shaanxi Province (No: 2014JM1029), and Scientific Research Foundation of Liaocheng University.
Competing interests
All three authors declare that they have no competing interests.
Footnotes
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Contributor Information
Xu Kong, Email: xu.kong@hotmail.com.
Jicheng Li, Email: jcli@mail.xjtu.edu.cn.
Xiaolong Wang, Email: xlwang@nwpu.edu.cn.
References
- 1.Che M.L., Cichocki A., Wei Y.M. Neural networks for computing best rank-one approximations of tensors and its applications. Neurocomputing. 2017;267:114–133. doi: 10.1016/j.neucom.2017.04.058. [DOI] [Google Scholar]
- 2.Cichocki A., Mandicand D., Pan A.-H., Caiafa C., Zhou G., Zhao Q., De Lathauwer L. Tensor decompositions for signal processing applications: from two-way to multiway component analysis. IEEE Signal Process. Mag. 2015;32(2):145–163. doi: 10.1109/MSP.2013.2297439. [DOI] [Google Scholar]
- 3.De Lathauwer L., De Moor B., Vandewalle J. A multilinear singular value decomposition. SIAM J. Matrix Anal. Appl. 2000;21(4):1253–1278. doi: 10.1137/S0895479896305696. [DOI] [Google Scholar]
- 4.De Silva V., Lim L.H. Tensor rank and the ill-posedness of the best low-rank approximation problem. SIAM J. Matrix Anal. Appl. 2008;30(3):1084–1127. doi: 10.1137/06066518X. [DOI] [Google Scholar]
- 5.Friedland S., Lim L.-H. The computational complexity of duality. SIAM J. Optim. 2016;26(4):2378–2393. doi: 10.1137/16M105887X. [DOI] [Google Scholar]
- 6.Friedland S., Lim L.H. Nuclear norm of higher-order tensors. Math. Comput. 2018;87(311):1255–1281. doi: 10.1090/mcom/3239. [DOI] [Google Scholar]
- 7.Gandy S., Recht B., Yamada I. Tensor completion and low-n-rank tensor recovery via convex optimization. Inverse Probl. 2011;27(2):025010. doi: 10.1088/0266-5611/27/2/025010. [DOI] [Google Scholar]
- 8.Hu S.L. Relations of the nuclear norm of a tensor and its matrix flattenings. Linear Algebra Appl. 2015;478:188–199. doi: 10.1016/j.laa.2015.04.003. [DOI] [Google Scholar]
- 9.Kolda T.G., Bader B.W. Orthogonal tensor decomposition. SIAM J. Matrix Anal. Appl. 2001;23:243–255. doi: 10.1137/S0895479800368354. [DOI] [Google Scholar]
- 10.Kolda T.G., Bader B.W. Tensor decompositions and applications. SIAM Rev. 2009;51(3):455–500. doi: 10.1137/07070111X. [DOI] [Google Scholar]
- 11.Kong X., Meng D.Y. The bounds for the best rank-1 approximation ratio of a finite dimensional tensor space. Pac. J. Optim. 2015;11:323–337. [Google Scholar]
- 12.Li Z., Nakatsukasa Y., Soma T., Uschmajew A. On orthogonal tensors and best rank-one approximation ratio. SIAM J. Matrix Anal. Appl. 2018;39(1):400–425. doi: 10.1137/17M1144349. [DOI] [Google Scholar]
- 13.Oseledets I.V., Tyrtyshnikov E.E. Breaking the curse of dimensionality, or how to use SVD in many dimensions. SIAM J. Sci. Comput. 2009;31(5):3744–3759. doi: 10.1137/090748330. [DOI] [Google Scholar]
- 14.Qi L., Luo Z. Tensor Analysis: Spectral Theory and Special Tensor. Philadelphia: SIAM; 2017. [Google Scholar]
