Tenney–Euclidean temperament measures: Difference between revisions

The wedgies should be weighted too
Undo revision 203571 by VectorGraphics (talk). Readers aren't obligated to see your poor-tasted operation name
Tag: Undo
 
(14 intermediate revisions by 2 users not shown)
Line 17: Line 17:
As these metrics are mainly used to rank temperaments within the same [[rank]] and [[just intonation subgroup]], it does not matter much which scheme is used, because they are equivalent up to a scaling factor, so they will rank temperaments identically. As a result, it is somewhat common to equivocate between the various choices of scaling factor, and treat the entire thing as "the" Tenney–Euclidean norm, so that we are really only concerned with the results of these metrics up to that equivalence.
As these metrics are mainly used to rank temperaments within the same [[rank]] and [[just intonation subgroup]], it does not matter much which scheme is used, because they are equivalent up to a scaling factor, so they will rank temperaments identically. As a result, it is somewhat common to equivocate between the various choices of scaling factor, and treat the entire thing as "the" Tenney–Euclidean norm, so that we are really only concerned with the results of these metrics up to that equivalence.


Graham Breed's original definitions from his ''primerr.pdf'' paper tend to use the third definition, as do parts of his [http://x31eq.com/temper/ temperament finder], although other scaling and normalization methods are sometimes used as well.
Graham Breed's original definitions from his ''primerr.pdf'' paper tend to use the third definition, as do parts of his [https://x31eq.com/temper/ temperament finder], although other scaling and normalization methods are sometimes used as well.


It is also possible to normalize the metrics to allow us to meaningfully compare temperaments across subgroups and even ranks. [[Sintel]]'s scheme in 2023 is the first attempt at this goal.  
It is also possible to normalize the metrics to allow us to meaningfully compare temperaments across subgroups and even ranks. [[Sintel]]'s scheme in 2023 is the first attempt at this goal<ref name="sintel">Sintel. [https://github.com/Sin-tel/temper/blob/c0d5c36e3c189f64860f4aea288ff3ff3bc34982/lib_temper/temper.py "Collection of functions for dealing with regular temperaments"], Temperament Calculator.</ref>.  


== TE complexity ==
== TE complexity ==
Given a [[wedgie]] ''M'', that is a canonically reduced ''r''-val correspondng to a temperament of rank ''r'', the norm ‖''M''‖ is a measure of the [[complexity]] of ''M''; that is, how many notes in some sort of weighted average it takes to get to intervals. For 1-vals, for instance, it is approximately equal to the number of scale steps it takes to reach an octave. We may call it '''Tenney–Euclidean complexity''', or '''TE complexity''' since it can be defined in terms of the [[Tenney–Euclidean metrics|Tenney–Euclidean norm]].  
{{Todo|rework|inline=1|text=Explain without wedgies}}
 
Given a [[wedgie]] ''M'', that is a canonically reduced ''r''-val correspondng to a temperament of rank ''r'', the norm ‖''M''‖ is a measure of the complexity of ''M''; that is, how many notes in some sort of weighted average it takes to get to intervals. For 1-vals, for instance, it is approximately equal to the number of scale steps it takes to reach an octave.  


Let us define the val weighting matrix ''W'' to be the {{w|diagonal matrix}} with values 1, 1/log<sub>2</sub>3, 1/log<sub>2</sub>5 … 1/log<sub>2</sub>''p'' along the diagonal. For the prime basis {{nowrap|''Q'' {{=}} {{val| 2 3 5 … ''p'' }} }},  
Let us define the val weighting matrix ''W'' to be the {{w|diagonal matrix}} with values 1, 1/log<sub>2</sub>3, 1/log<sub>2</sub>5 … 1/log<sub>2</sub>''p'' along the diagonal. For the prime basis {{nowrap|''Q'' {{=}} {{val| 2 3 5 … ''p'' }} }},  
Line 28: Line 30:
$$ W = \operatorname {diag} (1/\log_2 (Q)) $$
$$ W = \operatorname {diag} (1/\log_2 (Q)) $$


If ''V'' is the mapping matrix of a temperament, then ''V<sub>W</sub>'' {{=}} ''VW'' is the mapping matrix in the weighted space, its rows being the weighted vals ('''v'''<sub>''w''</sub>)<sub>''i''</sub>.  
If ''V'' is the mapping matrix of a temperament, then ''V<sub>W</sub>'' {{=}} ''VW'' is the mapping matrix in the weighted space, its rows being the weighted vals (''v''<sub>''w''</sub>)<sub>''i''</sub>.  


Our first complexity measure of a temperament is given by the ''L''<sup>2</sup> norm of the Tenney-weighted wedgie ''M''<sub>''W''</sub>, which can in turn be obtained from the Tenney-weighted mapping matrix ''V''<sub>''W''</sub>. This complexity can be easily computed either from the wedgie or from the mapping matrix, using the {{w|Gramian matrix|Gramian}}:  
Our first complexity measure of a temperament is given by the ''L''<sup>2</sup> norm of the Tenney-weighted wedgie ''M''<sub>''W''</sub>, which can in turn be obtained from the Tenney-weighted mapping matrix ''V''<sub>''W''</sub>. This complexity can be easily computed either from the wedgie or from the mapping matrix, using the {{w|Gramian matrix|Gramian}}:  


$$ \norm{M_W}_2 = \sqrt {\abs{V_W V_W^\mathsf{T}}} $$
$$ \norm{M_W}_2 = \sqrt {\det(V_W V_W^\mathsf{T})} $$


where {{!}}''A''{{!}} denotes the determinant of ''A'', and ''A''{{t}} denotes the transpose of ''A''.  
where det(·) denotes the determinant, and {{t}} denotes the transpose.  


We denote the RMS norm of ''M'' as ‖''M''‖<sub>RMS</sub>. In Graham Breed's paper<ref name="primerr">[http://x31eq.com/temper/primerr.pdf ''Prime Based Error and Complexity Measures''], often referred to as ''primerr.pdf''</ref>, an RMS norm is proposed as
Graham Breed and [[Gene Ward Smith]] have proposed different RMS norms. Let us denote the RMS norm of ''M'' as ‖''M''‖<sub>RMS</sub>. In Graham's paper<ref name="primerr">Graham Breed. [http://x31eq.com/temper/primerr.pdf ''Prime Based Error and Complexity Measures''], often referred to as ''primerr.pdf''.</ref>, an RMS norm is proposed as


$$ \norm{M_W}_\text{RMS} = \sqrt {\abs{\frac {V_W V_W^\mathsf{T}}{n}}} = \frac {\norm{M}_2}{\sqrt {n^r}} $$
$$ \norm{M_W}_\text{RMS} = \sqrt {\det \left( \frac {V_W V_W^\mathsf{T}}{n} \right)} = \frac {\norm{M_W}_2}{\sqrt {n^r}} $$


where ''n'' is the number of primes up to the prime limit ''p'', and ''r'' is the rank of the temperament, which equals the number of vals wedged together to compute the wedgie.  
where ''n'' is the number of primes up to the prime limit ''p'', and ''r'' is the rank of the temperament. Thus ''n''<sup>''r''</sup> is the number of permutations of ''n'' things taken ''r'' at a time with repetition, which equals the number of entries of the wedgie in its full tensor form.  


: '''Note''': that is the definition used by Graham Breed's temperament finder.  
: '''Note''': that is the definition used by Graham Breed's temperament finder.  


[[Gene Ward Smith]] has recognized that TE complexity can be interpreted as the RMS norm of the wedgie. That defines another RMS norm,
Gene Ward Smith's RMS norm is given as
 
$$ \norm{M_W}_\text{RMS'} = \sqrt {\frac{\det(V_W V_W^\mathsf{T})}{C(n, r)}} = \frac {\norm{M_W}_2}{\sqrt {C(n, r)}} $$
 
where {{nowrap|C(''n'', ''r'')}} is the number of combinations of ''n'' things taken ''r'' at a time without repetition, which equals the number of entries of the wedgie in the usual, compressed form.  


$$ \norm{M_W}_\text{RMS}' = \sqrt {\frac{\abs{V_W V_W^\mathsf{T}}}{C(n, r)}} = \frac {\norm{M}_2}{\sqrt {C(n, r)}} $$
We may also note {{nowrap| det(''V''<sub>''W''</sub>''V''<sub>''W''</sub>{{t}}) {{=}} det(''VW''<sup>2</sup>''V''{{t}}) }}. This may be related to the [[Tenney–Euclidean metrics|TE tuning projection matrix]] ''P''<sub>''W''</sub>, which is ''V''<sub>''W''</sub>{{t}}(''V''<sub>''W''</sub>''V''<sub>''W''</sub>{{t}}){{inv}}''V''<sub>''W''</sub>, and the corresponding matrix for unweighted monzos {{nowrap|''P'' {{=}} ''V''{{t}}(''VW''<sup>2</sup>''V''{{t}}){{inv}}''V''}}.


where {{nowrap|C(''n'', ''r'')}} is the number of combinations of ''n'' things taken ''r'' at a time, which equals the number of entries of the wedgie.  
Sintel has defined a complexity measure that serves as an intermediate step for his badness metric<ref name="sintel"/>, which we will get to later. To obtain this complexity, we normalize the Tenney-weighting matrix ''W'' to ''U'' such that {{nowrap| det(''U'') {{=}} 1 }}, and then take the ''L''<sup>2</sup> norm of ''M''<sub>''U''</sub>. It can be shown that


: '''Note''': that is the definition currently used throughout the wiki, unless stated otherwise.
$$ U = W / \det(W)^{1/n} $$


We may also note {{nowrap|{{!}}''V''<sub>''W''</sub>''V''<sub>''W''</sub>{{t}}{{!}} {{=}} {{!}}''VW''<sup>2</sup>''V''{{t}}{{!}}}}. This may be related to the [[Tenney–Euclidean metrics|TE tuning projection matrix]] ''P''<sub>''W''</sub>, which is ''V''<sub>''W''</sub>{{t}}(''V''<sub>''W''</sub>''V''<sub>''W''</sub>{{t}}){{inv}}''V''<sub>''W''</sub>, and the corresponding matrix for unweighted monzos {{nowrap|''P'' {{=}} ''V''{{t}}(''VW''<sup>2</sup>''V''{{t}}){{inv}}''V''}}.
and so the complexity is
 
$$ \norm{M_U}_2 = \norm{M_W}_2 / \det(W)^{r/n} $$


== TE error ==
== TE error ==
We can consider '''TE error''' to be a weighted average of the error of each [[prime harmonic]]s in [[TE tuning]], that is, a weighted average of the [[error map]] in TE tuning. TE error may be expressed in any logarithmic [[interval size unit]]s such as [[cent]]s or [[octave]]s.  
We can consider TE error to be a weighted average of the error of each [[prime harmonic]]s in [[TE tuning]], that is, a weighted average of the [[error map]] in the tuning where it is minimized. In this regard, TE error may be expressed in any logarithmic [[interval size unit]]s such as [[cent]]s or [[octave]]s.  
 
As with complexity, we may simply define the TE error as the ''L''<sup>2</sup> norm of the weighted TE error map. If {{nowrap| ''T''<sub>''W''</sub> {{=}} ''TW'' }} is the weighted TE tuning map and {{nowrap| ''J''<sub>''W''</sub> {{=}} ''JW'' {{=}} {{val| 1 1 … 1 }} }} is the weighted just tuning map, then the TE error ''E'' is given by
 
$$
\begin{align}
E &= \norm{T_W - J_W}_2 \\
&= \norm{J_W(V_W^+ V_W - I) }_2 \\
&= \sqrt{J_W(V_W^+ V_W - I)(V_W^+ V_W - I)^\mathsf{T} J_W^\mathsf{T}}
\end{align}
$$


By Graham Breed's definition<ref name="primerr"/>, TE error may be accessed via [[Tenney–Euclidean tuning|TE tuning map]]. If ''T''<sub>''W''</sub> is the Tenney-weighted tuning map, then the TE error ''G'' can be found by
where <sup>+</sup> denotes the [[pseudoinverse]].
 
Often, it is desirable to know the average of errors instead of the sum, which corresponds to Graham Breed's definition<ref name="primerr"/>. This error figure, ''G'', can be found by


$$
$$
\begin{align}
\begin{align}
G &= \norm{T_W - J_W}_\text{RMS} \\
G &= \norm{T_W - J_W}_\text{RMS} \\
&= \norm{J_W(V_W^+ V_W - I) }_\text{RMS} \\
&= E / \sqrt{n}
&= \sqrt{J_W(V_W^+ V_W - I)(V_W^+ V_W - I)^\mathsf{T} J_W^\mathsf{T}/n}
\end{align}
\end{align}
$$
$$
If ''T''<sub>''W''</sub> is denominated in cents, then ''J''<sub>''W''</sub> should be also, so that {{nowrap|''J''<sub>''W''</sub> {{=}} {{val| 1200 1200 … 1200 }}}}. Here {{nowrap|''T''<sub>''W''</sub> − ''J''<sub>''W''</sub>}} is the list of weighted errors of each prime harmonic.


: '''Note''': that is the definition used by Graham Breed's temperament finder.  
: '''Note''': that is the definition used by Graham Breed's temperament finder.  


By Gene Ward Smith's definition, the TE error is derived from the relationship of TE simple badness and TE complexity. We denote this definition of TE error ''Ψ''.  
Gene Ward Smith defines the TE error as the ratio ‖''M''<sub>''W''</sub> ∧ ''J''<sub>''W''</sub>‖/‖''M''<sub>''W''</sub>‖, derived from the relationship of TE simple badness and TE complexity. See the next section. We denote this definition of TE error ''Ψ''. From {{nowrap|‖''M''<sub>''W''</sub> ∧ ''J''<sub>''W''</sub>‖/‖''M''<sub>''W''</sub>‖}} we can extract a coefficient {{nowrap| sqrt(''C''(''n'', ''r'' + 1)/''C''(''n'', ''r'')) {{=}} sqrt((''n'' − ''r'')/(''r'' + 1)) }}, which relates ''Ψ'' with ''E'' as follows:


From the ratio {{nowrap|(‖''J''<sub>''W''</sub> ∧ ''M''‖ / ‖''M''‖)<sup>2</sup>}} we obtain {{nowrap|{{sfrac|''C''(''n'', ''r'' + 1)|''n''⋅''C''(''n'', ''r'')}} {{=}} {{sfrac|''n'' − ''r''|''n''(''r'' + 1)}}}}. If we take the ratio of this for rank 1 with this for rank ''r'', the ''n'' cancels, and we get {{nowrap|{{sfrac|''n'' − 1|2}} · {{sfrac|''r'' + 1|''n'' − ''r''}} {{=}} {{sfrac|(''r'' + 1)(''n'' − 1)|2(''n'' − ''r'')}}}}. It follows that dividing TE error by the square root of this ratio gives a constant of proportionality such that if Ψ is the TE error of a rank-''r'' temperament then
$$ \Psi = \sqrt{\frac{r + 1}{n - r}} E $$


$$ \psi = \sqrt{\frac{2(n-r)}{(r+1)(n-1)}} \Psi $$
Also, if we set the rank ''r'' to 1, we get {{nowrap| (''n'' − 1)/2 }}. It follows that dividing TE error by this value gives a constant of proportionality such that


is an '''adjusted error''' which makes the error of a rank ''r'' temperament correspond to the errors of the edo vals which support it; so that requiring the edo val error to be less than {{nowrap|(1 + ''ε'')ψ}} for any positive ''ε'' results in an infinite set of vals supporting the temperament.
$$ \psi = \sqrt{\frac{2}{n - 1}} E $$


''Ψ'', ''ψ'', and ''G'' error can be related as follows:
gives another error, called the ''adjusted error'', which makes the error of a rank-''r'' temperament correspond to the errors of the edo vals which support it; so that requiring the edo val error to be less than {{nowrap|(1 + ''ε'')''ψ''}} for any positive ''ε'' results in an infinite set of vals supporting the temperament.


$$ G = \sqrt{\frac{n-1}{2n}} \psi = \sqrt{\frac{n-r}{(r+1)n}} \Psi $$
''G'' and ''ψ'' error both have the advantage that higher-rank temperament error corresponds directly to rank-1 error, but the RMS normalization has the further advantage that in the rank-1 case, {{nowrap| ''G'' {{=}} sin ''θ'' }} octaves, where ''θ'' is the angle between ''J''<sub>''W''</sub> and the val in question.  
 
''G'' and ''ψ'' error both have the advantage that higher-rank temperament error corresponds directly to rank-1 error, but the RMS normalization has the further advantage that in the rank-1 case, {{nowrap|''G'' {{=}} sin ''θ''}}, where ''θ'' is the angle between ''J''<sub>''W''</sub> and the val in question. Multiplying by 1200 to obtain a result in cents leads to 1200&nbsp;sin(''θ''), the TE error in cents.


== TE simple badness ==
== TE simple badness ==
The '''TE simple badness''' of a temperament, which we may also call the '''relative error''' of a temperament, may be considered error relativized to the complexity of the temperament. It is error proportional to the complexity, or size, of the multival; in particular for a 1-val, it is (weighted) error compared to the size of a step.  
The '''TE simple badness''' of a temperament, which we may also call the '''relative error''' of a temperament, may be considered error relativized to the complexity of the temperament. It is error proportional to the complexity, or size, of the multival; in particular for a 1-val, it is (weighted) error compared to the size of a step.  


Gene Ward Smith defines the simple badness of ''M'' as {{nowrap|‖''J''<sub>''W''</sub> ∧ ''M''<sub>''W''</sub>‖<sub>RMS</sub>}}, where {{nowrap|''J''<sub>''W''</sub> {{=}} {{val| 1 1 … 1 }}}} is the JIP in weighted coordinates. Once again, if we have a list of vectors we may use a Gramian to compute it. First we note that {{nowrap|''a''<sub>''i''</sub> {{=}} ''J''<sub>''W''</sub>·('''v'''<sub>''w''</sub>)<sub>''i''</sub>/''n''}} is the mean value of the entries of ('''v'''<sub>''w''</sub>)<sub>''i''</sub>. Then note that {{nowrap|''J''<sub>''W''</sub> ∧ (('''v'''<sub>''w''</sub>)<sub>1</sub> − ''a''<sub>1</sub>''J''<sub>''W''</sub>) ∧ (('''v'''<sub>''w''</sub>)<sub>2</sub> − ''a''<sub>2</sub>''J''<sub>''W''</sub>) ∧ … ∧ (('''v'''<sub>''w''</sub>)<sub>''r''</sub> − ''a''<sub>''r''</sub>''J''<sub>''W''</sub>) {{=}} ''J''<sub>''W''</sub> ∧ ('''v'''<sub>''w''</sub>)<sub>1</sub> ∧ ('''v'''<sub>''w''</sub>)<sub>2</sub> ∧ … ∧ ('''v'''<sub>''w''</sub>)<sub>''r''</sub>}}, since wedge products with more than one term ''J''<sub>''W''</sub> are zero. The Gram matrix of the vectors ''J''<sub>''W''</sub> and {{nowrap|('''v'''<sub>''w''</sub>)<sub>1</sub> − ''a''<sub>''i''</sub>''J''<sub>''W''</sub>}} will have ''n'' as the {{nowrap|(1, 1)}} entry, and 0's in the rest of the first row and column. Hence we obtain:
In general, if ''C'' is the complexity and ''E'' is the error of a temperament, then TE simple badness ''B'' is found by


$$ \norm{ J_W \wedge M_W }'_\text {RMS} = \sqrt{\frac{n}{C(n, r + 1)}} \abs{(\vec{v_w})_i \cdot (\vec{v_w})_j - n a_i a_j} $$
$$ B = C \cdot E $$


A perhaps simpler way to view this is to start with a mapping matrix ''V''<sub>''W''</sub> and add an extra row ''J''<sub>''W''</sub> corresponding to the JIP; we will label this matrix ''V''<sub>''J''</sub>. Then the simple badness is:
Gene Ward Smith defines the simple badness of ''M'' as {{nowrap|‖''M''<sub>''W''</sub> ∧ ''J''<sub>''W''</sub>‖<sub>RMS</sub>}}. A perhaps simpler way to view this is to start with a mapping matrix ''V''<sub>''W''</sub> and add an extra row ''J''<sub>''W''</sub> corresponding to the just tuning map; we will label this matrix ''''<sub>''W''</sub>. Then the simple badness is:


$$ \norm{ J_W \wedge M_W }'_\text {RMS} = \sqrt{\frac{n}{C(n, r + 1)}} \abs{V_J V_J^\mathsf{T}} $$
$$ \norm{ M_W \wedge J_W }_\text {RMS'} = \sqrt{\frac{\det(\tilde V_W \tilde V_W^\mathsf{T})}{C(n, r + 1)}} $$


So that we can basically view the simple badness as the TE complexity of the "pseudo-temperament" formed by adding the JIP to the mapping matrix as if it were another val.
So that we can basically view the simple badness as the TE complexity of the "pseudo-temperament" formed by adding the JIP to the mapping matrix as if it were another val.


Graham Breed defines the simple badness slightly differently, again equivalent to a choice of scaling. This is skipped here because, by that definition, it is easier to find TE complexity and TE error first and multiply them together to get the simple badness.
Graham Breed defines the simple badness slightly differently, again equivalent to a choice of scaling, skipped here because it is derived from the general formula.
 
Sintel has likewise given a simple badness as
 
$$ \norm{ M_U \wedge J_U }_2 = \sqrt{\det(\tilde V_U \tilde V_U^\mathsf{T})} $$
 
where {{nowrap| ''J''<sub>''U''</sub> {{=}} ''J''<sub>''W''</sub>/det(''W'')<sup>1/''n''</sup> }} is the ''U''-weighted just tuning map.  


=== Reduction to the span of a comma ===
=== Reduction to the span of a comma ===
Line 105: Line 126:
This relationship also holds if TOP is used rather than TE, as the TOP damage associated with tempering out some comma ''n''/''d'' is log(''n''/''d'')/(''nd''), and if we multiply by the complexity ''nd'', we simply get log(''n''/''d'') as our result.
This relationship also holds if TOP is used rather than TE, as the TOP damage associated with tempering out some comma ''n''/''d'' is log(''n''/''d'')/(''nd''), and if we multiply by the complexity ''nd'', we simply get log(''n''/''d'') as our result.


=== TE logflat badness ===
== TE logflat badness ==
Some consider the simple badness to be a sort of badness which heavily favors complex temperaments. The '''logflat badness''' is developed to address that. If we define S(''A'') to be the simple badness (relative error) of a temperament ''A'', and C(''A'') to be the complexity of ''A'', then the logflat badness is defined by the formula
Some consider the simple badness to be a sort of badness which favors complex temperaments. The '''logflat badness''' (called ''Dirichlet coefficients'' in Sintel's scheme), is developed to address that. If we define ''B'' to be the simple badness (relative error) of a temperament, and ''C'' to be the complexity, then the logflat badness ''L'' is defined by the formula
 
$$ L = B \cdot C^{r/(n - r)} $$


<math>\displaystyle
The exponent is chosen such that if we set a cutoff margin for logflat badness, there are still infinite numbers of new temperaments appearing as complexity goes up, at a lower rate which is approximately logarithmic in terms of complexity.
S(A)C(A)^{r/(n - r)} \\
= \norm{ J_W \wedge M_W } \norm{M_W}^{r/(n - r)}
</math>


If we set a cutoff margin for logflat badness, there are still infinite numbers of new temperaments appearing as complexity goes up, at a lower rate which is approximately logarithmic in terms of complexity.
In Graham's and Gene's derivations,
 
$$ L = \norm{ M_W \wedge J_W } \norm{M_W}^{r/(n - r)} $$
 
In Sintel's derivation,  
 
$$ L = \norm{ M_U \wedge J_U } \norm{M_U}^{r/(n - r)} / \norm{J_U} $$
 
Notice the extra factor 1/‖''J''<sub>''U''</sub>‖, which is to say we divide it by the norm of the just tuning map. For comparison, Gene's derivation does not have this factor, whereas with Tenney weights, whether this factor is omitted or not has no effects on Graham's derivation since ‖''J''<sub>''W''</sub>‖<sub>RMS</sub> is unity.  


== Examples ==
== Examples ==
The different definitions yield different results, but they are related to each other by a factor derived only from the rank and limit. A meaningful comparison of temperaments in the same rank and limit can be provided by picking any one of them.  
While the different definitions yield different results, they are related to each other by a factor derived only from the rank and subgroup. A meaningful comparison of temperaments in the same rank and subgroup is provided by picking any one of them. Here, we consider septimal [[magic]] and [[meantone]], as follows.  


Here is a demonstration from [[7-limit]] [[magic]] and [[meantone]], comparing each of the definitions.  
{| class="wikitable center-all left-1"
 
|+ style="font-size: 105%;" | ''L''<sup>2</sup> norm
{| class="wikitable center-all"
|-
|+ style="font-size: 105%;" | 7-limit magic (left) vs. meantone (right) in TE temperament measures
! Temperament
! Complexity
! Error (¢)
! Simple badness
|-
| Septimal meantone
| 5.400
| 2.763
| 12.435×10<sup>−3</sup>
|-
| Septimal magic
| 7.195
| 2.149
| 12.882×10<sup>−3</sup>
|}
{| class="wikitable center-all left-1"
|+ style="font-size: 105%;" | Breed's RMS norm
|-
! Temperament
! Complexity
! Error (¢)
! Simple badness
|-
| Septimal meantone
| 1.350
| 1.382
| 1.554×10<sup>−3</sup>
|-
|-
!
| Septimal magic
! TE complexity
| 1.799
! TE error (¢)
| 1.074
! TE simple badness
| 1.610×10<sup>−3</sup>
|}
{| class="wikitable center-all left-1"
|+ style="font-size: 105%;" | Smith's RMS norm
|-
|-
! {{W|Norm (mathematics) #Euclidean norm|Standard ''L''<sup>2</sup> norm}}
! Temperament
| 7.195 : 5.400
! Complexity
| 2.149 : 2.763
! Error (¢)
| 12.882×10<sup>−3</sup> : 12.435×10<sup>−3</sup>
! Simple badness
|-
|-
! Breed's RMS norm
| Septimal meantone
| 1.799 : 1.350
| 2.204
| 1.074 : 1.382
| 3.384
| 1.610×10<sup>−3</sup> : 1.554×10<sup>−3</sup>
| 6.218×10<sup>−3</sup>
|-
|-
! Smith's RMS norm
| Septimal magic
| 2.937 : 2.204
| 2.937
| 2.631 : 3.384
| 2.631
| 6.441×10<sup>−3</sup> : 6.218×10<sup>−3</sup>
| 6.441×10<sup>−3</sup>
|}
|}