can be seen as a K-dimensional cube of data, in which each dimension correspond to a different factor of the input data space. With this definition, the j-mode vector of the K-th order tensor is a vector obtained from elements of by varying only one its index n j while keeping all other indices fixed. Further, if from the tensor the matrix T (j) is created, where
(1)
then columns of T (j) are j-mode vectors of . Also, T (j) is a matrix representation of the tensor , called a j-mode tensor flattening. The j-th index becomes a row index of T (j), whereas its column index is a product of all the rest K-1 indices. An analysis of sufficient computer representations of (1) are discussed in many publications, for instance in [5].
Further important concept is a p-mode product of a tensor with a matrix. A result of this operation is the tensor whose elements are obtained based on the following scheme:
(2)
As was shown, the p-mode product can be equivalently represented in terms of the flattened versions of the tensors T (p) and S (p). That is, if the following holds
(3)
then we have the following
(4)
In some computations, it is more efficient to represent the tensor and matrix product given in (2) in an equivalent representation based on the p-mode tensor flattening and the Kronecker product. That is,
(5)
can be equivalently represented as follows
(6)
where ⊗ denotes the Kronecker product between the matrices.
Equipped with the above concepts, the best rank-(R 1 , R 2 , …, R K ) decomposition of a tensor can be defined as a problem of computying a tensor , which is characteristic of the ranks , , …, , and which as close as possible approximates to the input tensor [11, 12], that is the following functional should be minimized:
(7)
where ||.|| F denotes the Frobenius norm . It can be shown that the approximated tensor conveys as much of the “energy”, in the sense of the squared entries of a tensor, as the original tensor , under the requested rank constraints. A value of E is called the reconstruction error. Figure 1 depicts the best rank-(R 1 , R 2 , R 3 ) decomposition of a 3D tensor . However, contrary to the rank definition of the matrices, there are different rank definitions for tensors. For more discussion see [5, 11].
Fig. 1
Schematic representation of the best rank-(R 1 , R 2 , R 3 ) decomposition of a 3D tensor
It can be also easily observed that the assumed rank conditions mean that the approximation tensor can be decomposed as follows
(8)
Each of the matrices , , …, and in (8) has orthonormal columns. The number of columns for S i is given by R i .
The core tensor is of dimensions R 1 , R 2 , …, R K . It can be computed from the original tensor as follows
(9)
Summarizing, to find the best rank-(R 1 , R 2 , …, R K ) approximation of it is sufficient to determine only a set of S i in (8), and then is computed from Eq. (9).
Further analysis is constrained exclusively to 3D tensors, such as the one shown in Fig. 1. It can be seen that this decomposition leads to a significant data reduction. The compression C ratio can be expressed as follows:
(10)
As alluded to previously, the only control parameters of the method are the ranks R 1 , R 2 , and R 3 . A trade-off can be achieved between the compression ratio C in (10) with respect to the approximation error expressed in Eq. (7). This influences also pattern recognition accuracy, as will be discussed.
2.2 Pattern Classification with the Best Rank Tensor Decomposition
The already described, the subspace obtained after the best rank decomposition can be used to generate specific features of an image X, which can be then used for pattern recognition [16]. The features are obtained by projecting the image X of dimensions N 1 ×N 2 into the space spanned by the two matrices S 1 and S 2 in accordance with (9). However, at first the pattern X needs to be represented in an equivalent tensor form which is of dimensions N 1 × N 2 × 1. Then, the feature tensor of dimensions R 1 × R 2 × 1 is obtained by projecting onto the space spanned by S 1 and S 2, as follows