# Constrained tuning

(Redirected from CTE)

Constrained tunings are tuning optimization techniques using the constraint of some purely tuned intervals (i.e. eigenmonzos). CTE tuning (constrained Tenney-Euclidean tuning) is the most typical instance. It has a more sophisticated variant, CTWE tuning (constrained Tenney-Weil-Euclidean tuning), aka KE tuning (Kees-Euclidean tuning). These two tunings will be the focus of this article. Otherwise normed tunings can be defined and computed analogously.

All constrained tunings are standard temperament optimization problems. Specifically, as TE tuning can be viewed as a least squares problem, CTE tuning can be viewed as an equality-constrained least squares problem.

The most common subject of constraint is the octave, which is assumed unless specified otherwise. For higher-rank temperaments, it may make sense to add multiple constraints, such as a pure-2.3 constrained tuning. For a rank-r temperament, specifying m eigenmonzos will yield r - m degrees of freedom to be optimized.

## Definition

Given a temperament mapping V and the JIP J, we specify a weight and a skew, represented by transformation matrices W and X, respectively, and a p-norm. Suppose the tuning is constrained by the eigenmonzo list MC. The goal is to find the generator list G by

Minimize

$\displaystyle \lVert GV_{WX} - J_{WX} \rVert_p$

subject to

$\displaystyle (GV - J)M_{\rm C} = O$

where (·)WX denotes the weight-skew transformation, found by

\displaystyle \begin{align} V_{WX} &= VWX \\ J_{WX} &= JWX \end{align}

The problem is feasible if

1. rank (MC) ≤ rank (V), and
2. The subspaces of MC and N (V) are linearly independent.

## Computation

As a standard optimization problem, numerous algorithms exist to solve it, such as sequential quadratic programming, to name one. Flora Canou's tuning optimizer is such an implementation in Python. Note: it depends on Scipy.

Code
# © 2020-2022 Flora Canou | Version 0.22.1

import warnings
import numpy as np
from scipy import optimize, linalg
np.set_printoptions (suppress = True, linewidth = 256, precision = 4)

PRIME_LIST = [2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61]
SCALAR = 1200 #could be in octave, but for precision reason

def get_subgroup (main, subgroup):
if subgroup is None:
subgroup = PRIME_LIST[:main.shape]
elif main.shape != len (subgroup):
warnings.warn ("dimension does not match. Casting to the smaller dimension. ")
dim = min (main.shape, len (subgroup))
main = main[:, :dim]
subgroup = subgroup[:dim]
return main, subgroup

def get_weight (subgroup, wtype = "tenney", wamount = 1):
if wtype == "tenney":
weight_vec = np.reciprocal (np.log2 (np.array (subgroup, dtype = float)))
elif wtype == "wilson" or wtype == "benedetti":
weight_vec = np.reciprocal (np.array (subgroup, dtype = float))
elif wtype == "equilateral":
weight_vec = np.ones (len (subgroup))
elif wtype == "frobenius":
warnings.warn ("\"frobenius\" is deprecated. Use \"equilateral\" instead. ")
weight_vec = np.ones (len (subgroup))
else:
warnings.warn ("weighter type not supported, using default (\"tenney\")")
return get_weight (subgroup, wtype = "tenney", wamount = wamount)
return np.diag (weight_vec**wamount)

def get_skew (subgroup, skew = 0, order = 2):
if skew == 0:
return np.eye (len (subgroup))
elif skew == np.inf:
return np.eye (len (subgroup)) - np.ones ((len (subgroup), len (subgroup)))/len(subgroup)
elif order == 2:
r = skew/(len (subgroup)*skew**2 + 1)
return np.append (
np.eye (len (subgroup)) - skew*r*np.ones ((len (subgroup), len (subgroup))),
r*np.ones ((len (subgroup), 1)), axis = 1)
else:
raise NotImplementedError ("Weil skew only works with Euclidean norm as of now.")

def weightskewed (main, subgroup, wtype = "tenney", wamount = 1, skew = 0, order = 2):
return main @ get_weight (subgroup, wtype, wamount) @ get_skew (subgroup, skew, order)

def error (gen, map, jip, order):
return linalg.norm (gen @ map - jip, ord = order)

def optimizer_main (map, subgroup = None, wtype = "tenney", wamount = 1, skew = 0, order = 2,
cons_monzo_list = None, des_monzo = None, show = True):
map, subgroup = get_subgroup (np.array (map), subgroup)

jip = np.log2 (subgroup)*SCALAR
map_wx = weightskewed (map, subgroup, wtype, wamount, skew, order)
jip_wx = weightskewed (jip, subgroup, wtype, wamount, skew, order)
if order == 2 and cons_monzo_list is None: #te with no constraints, simply use lstsq for better performance
res = linalg.lstsq (map_wx.T, jip_wx)
gen = res
print ("L2 tuning without constraints, solved using lstsq. ")
else:
gen0 = [SCALAR]*map.shape #initial guess
cons = () if cons_monzo_list is None else {'type': 'eq', 'fun': lambda gen: (gen @ map - jip) @ cons_monzo_list}
res = optimize.minimize (error, gen0, args = (map_wx, jip_wx, order), method = "SLSQP",
options = {'ftol': 1e-9}, constraints = cons)
print (res.message)
if res.success:
gen = res.x
else:
raise ValueError ("infeasible optimization problem. ")

if not des_monzo is None:
if np.array (des_monzo).ndim > 1 and np.array (des_monzo).shape != 1:
raise IndexError ("only one destretch target is allowed. ")
elif (tempered_size := gen @ map @ des_monzo) == 0:
raise ZeroDivisionError ("destretch target is in the nullspace. ")
else:
gen *= (jip @ des_monzo)/tempered_size

tuning_map = gen @ map
mistuning_map = tuning_map - jip

if show:
print (f"Generators: {gen} (¢)",
f"Tuning map: {tuning_map} (¢)",
f"Mistuning map: {mistuning_map} (¢)", sep = "\n")

return gen, tuning_map, mistuning_map

optimiser_main = optimizer_main


Constraints can be added from the parameter cons_monzo_list of the optimizer_main function. For example, to find the CTE tuning for septimal meantone, you type:

map = np.array ([[1, 0, -4, -13], [0, 1, 4, 10]])
optimizer_main (map, cons_monzo_list = np.transpose ( + *(map.shape - 1)))


You should get:

Optimization terminated successfully.
Generators: [1200.     1896.9521] (¢)
Tuning map: [1200.     1896.9521 2787.8085 3369.5214] (¢)
Mistuning map: [ 0.     -5.0029  1.4948  0.6955] (¢)


Analytical solutions exist for Euclidean (L2) tunings, see Constrained tuning/Analytical solution to constrained Euclidean tunings.

For CTE in particular, it can be solved in the method of Lagrange multiplier. The solution is given by

$\begin{bmatrix} G^{\mathsf T} \\ \Lambda^{\mathsf T} \end{bmatrix} = \begin{bmatrix} V_{WX} V_{WX}^{\mathsf T} & VM \\ (VM)^{\mathsf T} & O \end{bmatrix}^{-1} \begin{bmatrix} V_{WX} J_{WX}^{\mathsf T}\\ (JM)^{\mathsf T} \end{bmatrix}$

which is almost an analytical solution. Notice we introduced the vector of lagrange multipliers Λ, with length equal to the number of constraints. The lagrange multipliers have no concrete meaning for the resulting tuning, so they can be discarded.

## CTE tuning vs CTWE tuning

Consider the fact that TE tuning does not treat divisive ratios as more important than multiplicative ratios – 5/3 and 15/1 are taken as equally important, for example. To address that, a skew on the space may be introduced, resulting in TWE tuning. Constraining the equave to pure on top of TWE gives CTWE aka KE tuning.

POTE tuning works as a quick approximation to CTWE. As POTE destretches the equave, it keeps the angle in the tuning space unchanged, and thus sacrifices multiplicative ratios for divisive ratios. On the contrary, CTE sticks to the original design book of TE as its result remains TE optimal.

These tunings can be very different from each other. Take 7-limit meantone as an example. The POTE tuning map is a little bit flatter than quarter-comma meantone, with all the primes tuned flat:

$\langle \begin{matrix} 1200.000 & 1896.495 & 2785.980 & 3364.949 \end{matrix} ]$

The CTWE tuning map is a little bit sharper than quarter-comma meantone, with 5 tuned sharp and 3 and 7 flat:

$\langle \begin{matrix} 1200.000 & 1896.656 & 2786.625 & 3366.562 \end{matrix} ]$

The CTE tuning map is even sharper, with 3 tuned flat and 5 and 7 sharp.

$\langle \begin{matrix} 1200.000 & 1896.952 & 2787.809 & 3369.521 \end{matrix} ]$

## Special constraint

The special eigenmonzo WXj, where j is the all-ones monzo, has the effect of removing the weight-skewed tuning bias. This eigenmonzo is actually proportional to the monzo of the extra dimension introduced by the skew. In other words, it forces the extra dimension to be pure, and therefore, the skew will have no effect with this constrained tuning.

It can be regarded as a distinct optimum. In the case of Tenney weighting, it is the TOCTE tuning (Tenney ones constrained Tenney-Euclidean tuning).

### For equal temperaments

Since the one degree of freedom of equal temperaments is determined by the constraint, optimization is not involved. It is thus reduced to TOC tuning (Tenney ones constrained tuning). This constrained tuning demonstrates interesting properties.

The step size g can be found by

$\displaystyle g = 1/\operatorname {mean} (V_{WX})$

The edo number n can be found by

$\displaystyle n = 1/g = \operatorname {mean} (V_{WX})$

Unlike TE or TOP, the optimal edo number space in TOC is linear with respect to V. That is, if V = αV1 + βV2, then

\displaystyle \begin{align} n &= \operatorname {mean} (VWX) \\ &= \operatorname {mean} ((\alpha V_1 + \beta V_2)WX) \\ &= \operatorname {mean} (\alpha V_1 WX) + \operatorname {mean} (\beta V_2 WX) \\ &= \alpha n_1 + \beta n_2 \end{align}

As a result, the relative error space is also linear with respect to V.

For example, the relative errors of 12ettoc5 (12et in 5-limit TOC) is

$\displaystyle E_\text {r, 12} = \langle \begin{matrix} -1.55\% & -4.42\% & +10.08\% \end{matrix} ]$

That of 19ettoc5 is

$\displaystyle E_\text {r, 19} = \langle \begin{matrix} +4.08\% & -4.97\% & -2.19\% \end{matrix} ]$

As 31 = 12 + 19, the relative errors of 31ettoc5 is

\displaystyle \begin{align} E_\text {r, 31} &= E_\text {r, 12} + E_\text {r, 19} \\ &= \langle \begin{matrix} +2.52\% & -9.38\% & +7.88\% \end{matrix} ] \end{align}