Tutorial 1: Tensor Contractions
Please look at the tutorials readme page if you have not done so already. In this tutorial you will learn the basics of manipulating tensors and contracting tensor networks efficiently, including:

Initialization of tensors

Diagrammatic notation for tensors and tensor networks

Manipulation of tensors via 'permute' and 'reshape' functions

Binary tensor contractions and computational costs

Use of 'ncon' routine to contract networks
Code examples:
T1.1: Diagrammatic notation
For our present purpose, tensors are simply multidimensional arrays of (real or complex) numbers. Below are some useful ways to initialize tensors:
It is convenient to represent tensor networks using a diagrammatic notation, where individual tensors are represented as a solid shape with a number of 'legs' that corresponds to the rank of the tensor. Each leg is here labelled with a dummy index (usually a Latin letter: i, j, k, lâ€¦) necessary to relate the equation to the diagram. Some examples are presented below.
Note: both MATLAB and Julia use 1based indexing, such that a ddim index takes values [1,2,...,d] as depicted in the figures below, while Python uses 0based indexing such that the index takes values [0,1,...,d1].

vector

matrix

order3 tensor
Fig.1.1(a)
The diagrammatic tensor notation is useful for describing networks comprised of multiple tensors. An index shared by two tensors denotes a contraction (or summation) over this index. Examples:
Fig.1.1(b)
Fig.1.1(c)
Notice that example in Fig.1.1(b) is equivalent to a matrix multiplication between matrices A and B, while Fig.1.1(c) produces a rank3 tensor D via the contraction of a network with three tensors. Even in this relatively the simple example, we see that the diagrammatic notation is already easier to interpret than the corresponding index equation. In practice, once we have established a convention for index ordering, we can omit labeling the diagrams with dummy indices which further enhances their clarity.
T1.2: Permute and reshape operations
Fundamental to the manipulation of tensors are the 'permute' and â€˜reshapeâ€™ functions. Permute allows the index ordering of a tensor to be changed (but does not change the number of indices). The reshape function which allows a collection of tensor indices to be combined into a single larger index (or viceversa), thus can change the number of indices but not the total dimension. In equations, we sometimes bracket subscript indices to denote the combining of indices, while diagrammatically this may be indicated by drawing two tensor indices coming together to form a single index. Often tensor indices of larger dimension are depicted by thicker lines!
Fig.1.2(a): Permute
Fig.1.2(b): Reshape
Technical notes:
â€‹

The tensor reshape behaves differently in MATLAB/Julia versus Python due to a difference in convention. Both MATLAB and Julia use columnmajor order for storing matrices and tensors, such that a dbyd matrix Bij is stored as a length d^2 vector vk, with k = i + (j1)Ã—d. In contrast, Python uses rowmajor order such that a dbyd matrix Bij is stored as a vector vk, with k = iÃ—d + j. Fortunately this difference in convention does not often have significant consequences in terms of writing tensor network codes, since the choice of convention is not so important so long as it is consistently applied.
â€‹

The permute function reorders the storage of the elements of a tensor in computer memory, thus incurs some (often nonnegligible) computational cost. In contrast, the reshape function leaves the elements of a tensor unchanged in memory, instead only changing the metadata for how the tensor is to be interpreted (and thus incurs negligible cost).
T1.3: Binary tensor contractions
The usefulness of permute and reshape functions is that they allow a contraction between a pair of tensors (which we call a binary tensor contraction) to be recast as a matrix multiplication. Although the computational cost (measured in number of scalar multiplications) is the same both ways, it is usually preferable to recast as multiplication as modern hardware performs vectorized operations much faster than when using the equivalent FOR loop. The steps for doing this are outlined below:
Method:
Fig.1.3(a): Given:
Evaluate tensor C:
(i) Permute the A and B tensors such that the indices to be contracted becoming the trailing and leading indices respectively (with consistent order between both tensors)
â€‹
(ii) Reshape tensors into matrices (between indices to be contracted versus the remaining indices)
â€‹
(iii) Do the matrix multiplication
â€‹
(iv) Reshape C back into a tensor, perform final permutation if desired index ordering is different from current ordering
T1.4: Contraction costs
The computational cost of multiplying a d1byd2 dimensional matrix A with a d2byd3 dimensional matrix B is: cost(AÃ—B) = d1âˆ™d2âˆ™d3. Given the equivalence with matrix multiplication, this is also the cost of a binary tensor contraction (where each dimension d1, d2, d3 may now result as the product of several tensor indices from the reshapes).
Another way of computing the cost of contracting A and B is to take the product of the total dimensions, denoted dim(A) and dim(B), of each tensor divided by the total dimension of the contracted indices, denoted dim(Aâˆ©B). Examples are given below:
Fig.1.4(a):
Fig.1.4(b): (assume ddim indices)
Broadly speaking, there are two approaches that could be taken to contract a network containing N>2 tensors: (i) in a single step as a direct summation over all internal indices of the network or (ii) as a sequence of N1 binary contractions. In practice we prefer the latter option, which is either computationally cheaper or an equivalent cost as the former option. Examples:
Fig.1.4(c), which represents a product of three matrices, illustrates that it is more efficient (in terms of the total number of scalar multiplications) to evaluate the network as a sequence of binary contractions than as a single summation over all internal indices.
Fig.1.4(d):
Fig.1.4(d) illustrates that the total cost of contracting a tensor network can depend on the sequence of binary contractions used; here the optimal sequence depends on whether D is larger than d.
T1.5: Contraction of tensor networks
Given a tensor network composed of N tensors, there are two distinct steps needed to contract the network efficiently:

determine the optimal sequence of the (N1) binary tensor contractions,

evaluate each of the binary contractions in turn as a matrix multiplication by taking the proper tensor permutes and reshapes.
Notes: determining the optimal contraction sequence:
â€‹

Usually we refer to the â€˜optimalâ€™ sequence at that which minimizes the number of scalar multiplications, but one could also seek to minimize the size of intermediate tensors used in the contraction (if the calculation was memory limited). Often, though not always, these two criteria will coincide.
â€‹

Given a tensor network with only a few tensors it is often easy to find the optimal sequence â€˜manuallyâ€™ through inspection of the network. For more complicated networks with a large number of tensors it may be necessary to employ an automated search algorithm such as this.
Fig.1.5(a): Optimal contraction sequence (assume ddim indices):
Once the optimal contraction sequence has been determined, a network can be evaluated by implementing each of the binary contractions in turn. However, using â€˜reshapeâ€™ and â€˜permuteâ€™ commands for each binary tensor contraction, although computationally effective, has two significant drawbacks: (i) it results in lengthy code that is error prone and difficult to check and (ii) it does not allow for the contraction sequence to be easily changed (as, in general, the entire code for the contraction would need to be rewritten to accommodate a different ordering).
Network contractor â€˜nconâ€™:
The â€˜nconâ€™ function is a useful tool to lessen the programming effort required to implement a tensor network contraction. This function works by automatically performing a desired sequence of permutes, reshapes and matrix multiplications required to evaluate a tensor network. The â€˜nconâ€™ code and detailed instructions for its usage can be found here, or alternatively the code is also presented on the example code page. The first step in using â€˜nconâ€™ to evaluate a network is to make a labelled diagram of the network such that:â€‹
â€‹

Each internal index is labelled with a unique positive integer (typically sequential integers starting from 1, although this is not necessary).
â€‹

External indices of the diagram (if there are any) are labelled with sequential negative integers [1,2,3,â€¦] which denote the desired index order on the final tensor (with 1 as the first index, 2 as the second etc).
â€‹
Following this, the â€˜nconâ€™ routine is called as follows,
â€‹
OutputTensor = ncon(TensorArray, IndexArray, ContOrder),
â€‹
with input arguments defined:

TensorArray: 1D cell array containing the tensors comprising the network
â€‹

IndexArray: 1D cell array of vectors, where the kth element is a vector of the integer labels from the diagram on the kth tensor from â€˜TensorArrayâ€™ (ordered following the corresponding index order on this tensor).
â€‹

ContOrder: a vector containing the positive integer labels from the diagram, used to specify order in which â€˜nconâ€™ contracts the indices. Note that â€˜ContOrderâ€™ is an optional input that can be omitted if desired, in which case â€˜nconâ€™ will contract in ascending order of index labels.
â€‹
Fig.1.5(b): Example network contraction with 'ncon'
Index orders (clockwise from 6 o'clock):
Network contraction:
Solution: Labelled network:
Code:
Notes on ncon
â€‹

If a pair of tensors is connected via multiple indices then 'ncon' will perform the contraction as a single multiplication (as opposed to contracting each index sequentially).
â€‹

Can be used to evaluate partial traces (see example below).
â€‹

Can be used to combine disjoint tensors into a single tensor (see example below).
Fig.1.5(c): Partial trace
Fig.1.5(d): Disjoint networks
Problem Set 1:
Pb.1(a):
Pb.1(a):

find the optimal contraction sequence for this network (assume all indices are of equal dimension d).
â€‹

what is the leading order computational cost of this contraction? Express your answer as a power of d.
Pb.1(b): Index orders (clockwise from 6 o'clock):
Network contraction:
Pb.1(b)
Initialize rank3 random tensors A, B, C (assuming all indices are dimension d = 20). Write code to evaluate the network contraction (using the specified index orders) in three different ways:
â€‹

As a single summation â€‹â€‹over all internal indices using FOR loops.

As a sequence of binary contractions implemented using 'permute' and 'reshape'.

Using the 'ncon' routine.
â€‹
Check that all three approaches produce the same output tensor D, and compare their respective computation times.