# Basics#

This page describes the basic ways to set up a contraction and an optimizer and defines some common terms.
The focus of `cotengra`

is *exact contraction*. That is, taking a collection of tensors with indices described by a tensor network or einsum equation and then:

finding the best sequence of pairwise contractions that reduces them to a single output tensor

performing this contraction using a mix of

`tensordot`

and potentially`einsum`

calls

Note

`cotengra`

doesn’t involve itself with building, modifying, simplifying or decomposing tensors and tensor networks etc.

The minimal information you need to describe such a contraction is:

the index labels for each tensor

the output index labels

the size of each index

Here’s a very small example of such information involving 4 tensors:

```
%config InlineBackend.figure_formats = ['svg']
import cotengra as ctg
inputs = [
('a', 'b', 'x'),
('b', 'c', 'd'),
('c', 'e', 'y'),
('e', 'a', 'd'),
]
output = ('x', 'y')
size_dict = {'x': 2, 'y': 3, 'a': 4, 'b': 5, 'c': 6, 'd': 7, 'e': 8}
```

This is equivalent to describing an einsum equation and array shapes (such as for `numpy.einsum`

or `opt_einsum.contract`

):

```
import numpy as np
eq = 'abx,bcd,cey,ead->xy'
shapes = [(4, 5, 2), (5, 6, 7), (6, 8, 3), (8, 4, 7)]
arrays = [np.ones(s) for s in shapes]
```

The actual names of indices here are only relevant for defining the *geometry*
within this contraction.

Note

For readability, generally a single unicode character is used per index.
See `opt_einsum.get_symbol(i)`

if you need to generate a usable set of these.

Each index can be thought of as an * edge* in a tensor network,
with each tensor being a

*. If every index appears exactly twice in either the inputs or output, then the underlying geometry is described as a*

**node***, otherwise it is a*

**graph***. You can visualize an input contraction with:*

**hypergraph**```
ctg.HyperGraph(inputs, output, size_dict).plot()
```

Usually one of these representations is very easy to produce, or a library like `quimb`

will do it for you. In any case, the next step is to create an *optimizer*.

`HyperOptimizer`

#

The main driver is the `HyperOptimizer`

class, which optimizes a single contraction:

```
opt = ctg.HyperOptimizer()
```

The most flexible way to use this is to call the `search`

method which directly returns a
`ContractionTree`

. This is a * rooted
binary tree*:

```
# direct use
tree = opt.search(inputs, output, size_dict)
tree
```

```
<ContractionTree(N=4, branches=3, complete=True)>
```

The tree (which also has the mathematical name, ‘carving decomposition’) contains all the crucial information about costs and sizes of intermediate tensors:

```
tree.contraction_width(), tree.contraction_cost()
```

```
(8.39231742277876, 4656.0)
```

the

, $W$, is $log_2$ the size of the largest intermediate tensor**contraction width**the

, $C$, is the total number of scalar multiplications**contraction cost**

The tree can be used to perform the actual contraction too:

```
tree.contract(arrays)
```

```
array([[6720., 6720., 6720.],
[6720., 6720., 6720.]])
```

A tree combined with a specific traversal ordering is known as a * path*:

```
path = tree.get_path()
path
```

```
((0, 1), (1, 2), (0, 1))
```

Such paths can be supplied to `opt_einsum`

and `quimb`

functions, *or*
you can supply the `HyperOptimizer`

itself, in which case it will first run a search and then pass on a path.

### With `quimb`

#

```
import quimb.tensor as qtn
tn = qtn.TensorNetwork([
qtn.Tensor(array, inds)
for array, inds in zip(arrays, inputs)
])
tn.contract(..., optimize=ctg.HyperOptimizer())
```

```
Tensor(shape=(2, 3), inds=('x', 'y'), tags=oset([]))
```

Note for non-hyper graphs `quimb`

will figure out the output indices for you,
else you will need to supply `output_inds`

. `quimb`

also knows how to return
the `ContractionTree`

directly with:

```
tn.contraction_tree(path)
```

```
<ContractionTree(N=4, branches=3, complete=True)>
```

And many other methods and algorithms take a `optimize=path_or_optimizer`

option.

### With `opt_einsum`

#

You can supply a `path`

or `HyperOptimizer`

to all the functions of `opt_einsum`

which take an `optimize`

kwarg:

```
import opt_einsum as oe
oe.contract(eq, *arrays, optimize=path)
```

```
array([[6720., 6720., 6720.],
[6720., 6720., 6720.]])
```

Hint

For convenience `cotengra`

also registers a few presets which can be used like `optimize='hyper'`

, these can also be created.

A single `HyperOptimizer`

instance can only be used for a single contraction - everytime you supply it, as long as the contraction matches, it will simply continue it’s search.

`ReusableHyperOptimizer`

#

Often, instead you want a single optimizer with maybe customized settings to use for many different contractions - the answer is to use a `ReusableHyperOptimizer`

.
Everytime this is supplied to a *new* contraction it runs a search and stores the resulting path. The next time it sees this contraction it simply returns this cached path.

```
opt = ctg.ReusableHyperOptimizer(progbar=True)
```

```
opt.search(inputs, output, size_dict)
```

```
log2[SIZE]: 8.39 log10[FLOPs]: 3.97: 100%|██████████████████████| 128/128 [00:01<00:00, 84.50it/s]
```

```
<ContractionTree(N=4, branches=3, complete=True)>
```

```
opt.search(inputs, output, size_dict)
```

```
<ContractionTree(N=4, branches=3, complete=True)>
```

Note how the second call didn’t display a progress bar as it used the cached result.

Hint

The contractions are not cached using full (hyper) graph isomoprhism, which would not be scalable. Instead, the inputs have to be in the same order to produce the same hash key.

### Disk persistence#

If you want to persist contraction paths across python sessions (i.e. don’t want to explicitly save the `tree`

or `path`

objects yourself), then you can supply the `directory`

kwarg:

```
opt = ctg.ReusableHyperOptimizer(directory='cotengra_cache')
```

```
opt.search(inputs, output, size_dict)
```

```
<ContractionTree(N=4, branches=3, complete=True)>
```

The directory contains a single pickled file per contraction it has seen:

```
!ls cotengra_cache/
```

```
0489b9ab17a455bebcf3cc86c8ea5642f518ae95
```

```
# clean it up for now
!rm -rf cotengra_cache/
```

## What next?#

The ‘Advanced’ page, describes how to customize a

`HyperOptimizer`

in detail, for example to include subtree reconfiguration and dynamic slicingThe ‘Visualization’ page details many functions for plotting the contraction, tree and optimizer trials

The ‘Contraction’ page contains more information about actually performing the contraction, for example using a GPU

The ‘Tree Surgery’ page describes the design of the central

`ContractionTree`

object and ways to manipulate it

## Quick-start#

If you just want to get going, the following illustrate some common customizations of the optimizers.

### High quality sliced optimizer#

The following is an example of a high quality optimizer you might use to search for a single contraction, where you are memory bound to width $W=30$ and thus need to use slicing:

```
opt_hq_W30 = ctg.HyperOptimizer(
# do extra runs
max_repeats=1024,
# use dynamic slicing to target a width of 30
slicing_reconf_opts={'target_size': 2**30},
# use the nevergrad space searcher - good with large trial budget
optlib='nevergrad',
# terminate search if no change for 128 trials
max_time='equil:128',
# show live progress
progbar=True,
)
```

Everytime you supply this optimizer instance it will

*continue*its search, so you can interactively run it until you are happy or it seems to have converged.While a few hundred runs is usually sufficient when no slicing is needed, very large contractions requiring heavy slicing might benefit from a few thousand runs.

### Economical optimizer#

The following is an example of a reusable optimizer that is cheap to run and
requires no extra depedencies (i.e. `kahypar`

or a Bayesian optimizer),
but will still yield much better results than simple algorithms.
Useful if you have many smallish contractions:

```
opt_eco = ctg.ReusableHyperOptimizer(
# just do a few runs
max_repeats=32,
# only use the basic greedy optimizer ...
methods=['greedy'],
# ... but pair it with reconfiguration
reconf_opts={},
# just uniformly sample the space
optlib='random',
# terminate search if contraction is cheap
max_time='rate:1e6',
# account for both flops and write - usually wise for practical performance
minimize='combo-64',
# persist paths found in here
directory='cotengra_cache_eco',
)
```

```
# clean up the cache for now
!rm -rf cotengra_cache_eco
```