# Harmonic Entropy

**Harmonic Entropy**, sometimes abbreviated as "HE", is a simple model to quantify the extent to which musical chords exhibit various psychoacoustic effects, lumped together in a single construct called psychoacoustic **concordance**. It was invented by Paul Erlich and developed extensively on the Yahoo! tuning and harmonic_entropy lists. Various later contributions to the model have been made by Steve Martin, Mike Battaglia, Keenan Pepper, and others.

## Contents

- 1 Background
- 2 Basic Model: Shannon Entropy
- 3 Harmonic Rényi Entropy
- 4 Extending HE to [math]N=\infty[/math]: zeta-HE
- 4.1 Background: Unnormalized Entropy
- 4.2 Derivation
- 4.3 Examples
- 4.4 Equivalence of exp-UHE and UHE for [math]a \leq 2[/math]
- 4.5 Why not Normalized HE?
- 4.6 Interpretation as a New Free Parameter: the Weighting Exponent
- 4.7 Equivalence of the Weighting Exponent and [math]a[/math] for Generalized Normal Distributions
- 4.8 Reduced Rationals Only

- 5 To Do
- 6 References

# Background

The general workings of the human auditory system lead to a plethora of well-documented and sonically interesting phenomena that can occur when a musical chord is played:

- The perception of partial
**timbral fusion**of the chord into one complex sound - The appearance of a
**virtual fundamental**pitch in the bass - Critical band effects, such as timbral
**beatlessness**, compared to mistunings of the chord in the surrounding area - The appearance of a quick fluttering effect sometimes known as
**periodicity buzz**

There has been much research specifically on the musical implications critical band effects in the literature (e.g. Sethares's work), which are perhaps the psychoacoustic phenomena that readers are most familiar with. However, the modern xenharmonic community has displayed immense interest in exploring the other effects mentioned above as well, which have proven extremely important to the development of modern xenharmonic music.

These effects sometimes behave differently, and do not always appear strictly in tandem with one another. For instance, Paul Erlich has noted that most models for beatlessness measure 10:12:15 and 4:5:6 as being identical, whereas the latter exhibits more timbral fusion and a more salient virtual fundamental than the former.

However, suppose we want to come up with a combined measure for how often effects such as the above tend to occur. It is then useful to note that:

- In general, effects such as these tend to appear most strongly for those chords with large subsets that correspond to simple chunks of the harmonic series
- In general, the effects produced exhibit some degree of tolerance for mistuning

This enables us to speak of a general notion of the psychoacoustic **concordance** of an interval - the degree to which effects such as the above will appear when an arbitrary musical chord is played. Additionally, chords which are very inharmonic often exhibit a quality known as psychoacoustic **discordance**.

While psychoacoustic concordance is not a feature universal to all styles of music, it has been utilized significantly in Western music in the study of intonation. For instance, flexible-pitch ensembles operating within 12-EDO, such as barbershop quartets and string ensembles, will often adjust intonationally from the underlying 12-EDO reference to maximize the concordance of individual chords. Indeed, the entire history of Western tuning theory -- from meantone temperament, to the various Baroque well-temperaments, to 12-EDO itself, to the modern theory of regular temperament -- can be seen as an attempt to reason mathematically about how to generate manageable tuning systems that will maximize concordance and minimize discordance.

The Harmonic Entropy model is a simple way of quantifying how much an arbitrary chord will exhibit psychoacoustic concordance.

Concordance has often been confused with actual musical consonance, an unfortunate fact made more common by the psychoacoustics literature under the unfortunate name **sensory consonance**, most often used to refer to phenomena related to roughness and beatlessness specifically. This is not to be confused with the more familiar construct of tonal stability, typically just called "consonance" in Western common practice music theory and sometimes clarified as "musical consonance" in the music cognition literature. To make matters worse, the literature has also at times referred to concordance -- and not tonal stability -- as **tonal consonance**, often referring to phenomena related to virtual pitch integration, creating a complete terminological mess. As a result, the term "consonance" has been completely avoided in this article.

# Basic Model: Shannon Entropy

The original Harmonic Entropy model limited itself to working with dyads. More recently, work by Steve Martin and others has extended this basic idea to higher-cardinality chords. This article will concern itself with dyads, as the dyadic case is still the most well-developed, and many of the ideas extend naturally to larger chords without need for much exposition.

The general idea of Harmonic Entropy is to first develop a discrete probability distribution quantifying how strongly an arbitrary incoming dyad "matches" every element in a set of basis rational intervals, and then seeing how evenly distributed the resulting probabilities are. If the distribution for some dyad is spread out very evenly, such that there is no clear "victor" basis interval that dominates the distribution, the dyad is considered to be more discordant; on the other extreme, if the distribution tends to concentrate on one or a small set of dyads, the dyad is considered to be more concordant.

A clear mathematical way of quantifying this "dispersion" is via the Shannon entropy of the probability distribution, which can be thought of as describing the "uncertainty" in the distribution. A distribution which has a very high probability of picking one outcome has low entropy and is not very uncertain, whereas a distribution which has the probability spread out on many outcomes is highly uncertain and has a high entropy.

## Definitions

To formalize our notion of Shannon entropy, we will first describe the random variable [math]J[/math], representing the set of JI "basis" intervals that our incoming interval is being "matched" to, and the parameter [math]C[/math], representing the "cents" of the incoming interval being played. For example, the interval [math]C[/math] would take values such as "400 cents," and the interval [math]J[/math] would take values in the set of basis ratios, such as "5/4" or "9/7."

So for example, if we want to express the probability that the incoming dyad "400 cents" is perceived as the JI basis interval "5/4," we would write that as the conditional probability [math]\displaystyle \newcommand{\cent}{\text{¢}}[/math] [math]\displaystyle P(J=5/4|C=400\cent)[/math]

Or, in general, if we want to write the conditional probability that some incoming dyad of [math]c[/math] cents is perceived as the JI basis interval [math]j[/math], we would write that as

[math]\displaystyle P(J=j|C=c)[/math]

which notationally, we will often abbreviate as

[math]\displaystyle P(j|c)[/math]

Note that at this point, we haven't yet specified what the particular probability distribution is. There are different ways to do this, which are described in more detail below. Generally, most approaches involve each JI interval's probability being assigned based on how close it is to [math]c[/math] (closer dyads are given a larger probability), and how simple it is (simple dyads are given a higher probability, if distance is the same).

A noteworthy point is that we generally do not assume any probability distribution on [math]C[/math]. This reflects that we do not make any assumptions at all about which notes or intervals are likely to be played to begin with. In other words, we are treating [math]C[/math] more as a "parameter" rather than as a random variable.

Once we have decided on a probability distribution, we can finally evaluate the Shannon entropy. For a random variable [math]X[/math], the Shannon entropy is defined as:

[math]\displaystyle H(X) = -\sum_{x \in X} P(x) \log_b P(x)[/math]

where the different [math]x[/math] are taken from the sample space of [math]X[/math], and [math]b[/math] is the base of the log. Different choices of [math]b[/math] simply change the units in which entropy is given, the most common values being 2 and e, denoting "bits" and "nats". We will omit the base going forward, for simplicity.

In our case, we want to find the entropy of the random variable [math]J[/math] of JI intervals, given a particular choice of incoming dyad in cents. The corresponding quantity that we want is:

[math]\displaystyle H(J|c) = -\sum_{j \in J} P(j|c) \log P(j|c)[/math]

Note that above, the summation is only taken on the [math]j[/math] from the sample space of [math]J[/math] (i.e. the set of JI basis intervals), whereas the parameter [math]c[/math] is treated as constant within the summation (and is taken as the free parameter to the function).

Since the parameter [math]c[/math] is the free parameter, sometimes the above is notated as

[math]\displaystyle \text{HE}(c) = H(J|c)[/math]

which makes more explicit that [math]c[/math] is the argument to the harmonic entropy function, which is equal to the entropy of [math]J[/math], conditioned on the incoming dyad of [math]c[/math] cents.

## Probability Distributions

In order to systematically assign a probability distribution to this dyad, we first start by defining a **spreading function**, denoted by [math]S(x)[/math], that dictates how the dyad is "smeared" out in log-frequency space, representing how the auditory system allows for some tolerance for mistuning. The typical choice that we will assume here for a spreading function is a Gaussian distribution, with mean centered around the incoming dyad, and standard deviation typically taken as a free parameter in the system and denoted as [math]s[/math].

A fairly typical choice of settings for a basic dyadic HE model would be:

- The basis set is all those rationals bounded by some maximum Tenney height, with the bound typically notated as [math]N[/math] and set to at least 10,000.
- The spreading function is typically a Gaussian distribution with a frequency deviation of 1% either way, or about s=~17 cents.

Other spreading functions have also been explored, such as the use of the heavy-tailed Laplace distribution, sometimes described as the "Vos function" in Paul's writings. These two functions are part of the Generalized normal distribution family, which has a parameter not only for the variance but for the kurtosis. However, for simplicity, we will assume the Gaussian distribution as the spreading function for the remainder of this article, so that the spreading function for an incoming dyad [math]c[/math] can be written as follows:

[math]\displaystyle S(x-c) = \frac{1}{s\sqrt{2\pi}} e^{-\frac{(x-c)^2}{2s^2}}[/math]

where the notation [math]S(x-c)[/math] is chosen to make clear that we are translating [math]S(x)[/math] to be centered around the incoming dyad [math]c[/math], which is now the mean of the Gaussian.

We assume here that the variable [math]x[/math] is a dummy variable representing cents, and will adopt this convention for the remainder of the article.

In this notation, [math]s[/math] becomes the standard deviation of the Gaussian, being an ASCII-friendly version of the more familiar symbol [math]\sigma[/math] for representing the standard deviation. Note that in previous expositions on Harmonic Entropy, [math]s[/math] was sometimes given in units representing a percentage of linear-frequency deviation; we allow [math]s[/math] to stand for cents here to simplify the notation. To convert from a percentage to cents, the formula [math]\text{cents} = 1200\log_2(1+\text{percentage})[/math] can be used.

It is also common to use as a basis set all those rationals bounded by some maximum Weil height, with a typical cutoff for [math]N[/math] set to at least 100. This has sometimes been referred to as seeding HE with the "Farey sequence of order [math]N[/math]" and its reciprocals, so references in Paul's work to "Farey series HE" vs "Tenney series HE" are sometimes seen.

Lastly, the set of rationals is often chosen to be only those "reduced" rationals within the cutoff, such that [math]n/d[/math] is in the set only if [math]n[/math] and [math]d[/math] are coprime. HE can also be formulated with unreduced rationals as well. Both methods tend to give similar results. In Paul's work, reduced rationals are most common, although the use of unreduced rationals may be useful in extending HE to the case where [math]N=\infty[/math].

Given a spreading function and set of basis rationals, there are two different procedures commonly used to assign probabilities to each rational. The first, the **domain-integral approach**, works for arbitrary nowhere dense sets of rationals without any further free parameters. The second, the **simple weighted approach**, has nice mathematical properties which sometimes make it easier to compute and which may lead to generalizations to infinite sets of rationals which are sometimes dense in the reals. It is conjectured that there are certain important limiting situations where the two converge; both are described in detail below.

### Domain-Integral Probabilities

For discrete sets of JI basis ratios, the log-frequency spectrum can be divided up into **domains** assigned to each ratio. Each ratio is assigned a domain with lower bound equal to the mediant of itself and its nearest lower neighbor, and likewise with upper bound equal to the mediant of itself and its nearest upper neighbor. If no such neighbor exists, [math]\pm \infty[/math] is used instead. Mathematically, this can be represented via the following expression:

[math]\displaystyle P(j|c) = \int_{\cent(j_l)}^{\cent(j_u)} S(x-c) dx[/math]

where [math]S(x-c)[/math] is the spreading function associated with c, [math]j_l[/math] and [math]j_u[/math] are the domain lower and upper bounds associated with JI basis ratio [math]j[/math], and [math]\cent(f) = 1200\log_2(f)[/math], or the "cents" function converting frequency ratios to cents. Typically, [math]j_l[/math] is set equal to the mediant of [math]j[/math] and its nearest lower neighbor (if it exists), or [math]-\infty[/math] if not; likewise with [math]j_u[/math] and its nearest upper neighbor.

This process can be summarized by the following picture, taken from William Sethares' paper on Harmonic Entropy:

Note the difference in terminology here - in this example, the [math]f_{j+n}[/math] are the basis ratios, the [math]r_{j+n}[/math] are the domains for each basis ratio, and the bounds for each domain are the mediants between each [math]f_{j+n}[/math] and its nearest neighbor. The probability assigned to each basis ratio is then the area under the spreading function curve for each ratio's domain. The entropy of this probability distribution is then the Harmonic Entropy for that dyad.

In the case where the set of basis rationals consists of a finite set bounded by Tenney or Weil height, the resulting set of widths is conjectured to have interesting mathematical properties, leading to mathematically nice conceptual simplifications of the model. These simplifications are explained below.

### Simple Weighted Probabilities

It has been noted empirically by Paul Erlich that, given all those rationals with Tenney height under some cutoff [math]N[/math] as a basis set, that the domain widths for rationals sufficiently far from the cutoff seem to be proportional to [math]\frac{1}{\sqrt{nd}}[/math].

While it's still an open conjecture that this pattern holds for arbitrarily large [math]N[/math], the assumption is sometimes made that this is the case, and hence that for these basis ratio sets, [math]\frac{1}{\sqrt{nd}}[/math] "approximations" to the width are sufficient to estimate domain-integral Harmonic Entropy.

This modifies the expression for the probabilities [math]P(j|c)[/math] as follows, noting that for now the "probabilities" won't sum to 1:

[math]\displaystyle Q(j|c) = \frac{S(\cent(j)-c)}{\sqrt{j_n \cdot j_d}}[/math]

where the [math]Q[/math] notation now represents that these "probabilities" are unnormalized, and [math]j_n[/math] and [math]j_d[/math] are the numerator and denominator, respectively, of JI basis ratio [math]j[/math]. Again, the set of basis rationals here is assumed to be all of those rationals of Tenney Height ≤ [math]N[/math] for some [math]N[/math].

A similar observation for the use of Weil-bounded subsets of the rationals suggests domain widths of [math]\frac{1}{\max(n,d)}[/math], yielding instead the following formula:

[math]\displaystyle Q(j|c) = \frac{S(\cent(j)-c)}{\max(j_n, j_d)}[/math]

where this time the set of basis rationals is assumed to be all of those of Weil Height ≤ [math]N[/math] for some [math]N[/math].

In both cases, the general approach is the same: the value of the spreading function, taken at the value of [math]\cent(j)[/math], is divided by some sort of "weighting" (or sometimes, "complexity") function representing how much weight is given to that rational number. While the two weighting functions considered thus far were derived empirically by observing the asymptotic behavior of various height-bounded subsets of the rationals, we can generalize this for arbitrary basis sets of rationals and arbitrary weights as follows:

[math]\displaystyle Q(j|c) = \frac{S(\cent(j)-c)}{\|j\|}[/math]

where [math]\|j\|[/math] denotes a weighting function that maps from rational numbers to non-negative reals.

As these "probabilities" don't sum to 1, the result is not a probability distribution at all, invalidating the use of the Shannon Entropy. To rectify this, the distribution is normalized so that the probabilities do sum to 1:

[math]\displaystyle P(j|c) = \frac{Q(j|c)}{\sum_{j \in J} Q(j|c)}[/math]

which is equal to the unnormalized probability, divided by the sum of all unnormalized probabilities. This definition of [math]P(j|c)[/math] is then used directly to compute the entropy.

This approach to assigning probabilities to basis rationals is useful because it hypothetically makes it possible to consider the HE of sets of rationals which are dense in the reals, or even the entire set of positive rationals, although the best way to do this is a subject of ongoing research.

## Examples

In all of these examples, the x-axis represents the width in cents of the dyad, and the y-axis represents *discordance* rather than concordance, measured in nats of Shannon entropy.

### s=17, N<10000, sqrt(n*d) weights

This uses as a spreading function the Gaussian distribution with [math]s=~17\cent[/math] (or a lin-frequency deviation of 1%). The basis set is all rationals of Tenney height less than 10,000. This uses the simple weighted approach, and the weighting function is [math]\sqrt{nd}[/math]:

### s=17, N<100, max(n,d) weights

This example uses the same spreading function and standard deviation, but this time the basis set is all rationals of Weil height less than 100. The weighting function here is [math]\max(n,d)[/math]:

### s=17, N<10000, sqrt(n*d) vs mediant-to-mediant weights

The following image (from Paul Erlich) compares the domain-integral and simple weighted approaches by overlaying the two curves on top of each other. In both cases, the spreading function is again a Gaussian with s=~17 cents, and the basis set is all those rationals with Tenney height ≤ 10000. It can be seen that the curves are extremely similar, and that the locations of the minima and maxima are largely preserved:

# Harmonic Rényi Entropy

An extension to the base Harmonic Entropy model, proposed by Mike Battaglia, is to generalize the use of Shannon entropy by replacing it instead with Rényi entropy, a q-analog of Shannon's original entropy. This can be thought of as adding a second parameter, called [math]a[/math], to the model, reflecting how "intelligent" the brain's "decoding" process is when determining the most likely JI interpretation of an ambiguous interval.

## Definitions and Background

The **Harmonic Rényi Entropy of order a** of an incoming dyad can be defined as follows:

[math]\displaystyle \text{HE}_a(c) = H_a(J|c) = \frac{1}{1-a} \log \sum_{j \in J} P(j|c)^a[/math]

Being a q-analog, it is noteworthy that Rényi entropy converges to Shannon entropy in the limit as [math]a \to 1[/math], a fact which can be verified using L'Hôpital's rule as found here.

The Rényi entropy has found use in cryptography as a measure of the strength of a cryptographic code in the face of an intelligent attacker, an application for which Shannon entropy has long been known to be insufficient as described in this paper and this RFC. More precisely, the Rényi entropy of order [math]\infty[/math], also called the **min-entropy**, is used to measure the strength of the randomness used to define a cryptographic secret against a "worst-case" attacker who has complete knowledge of the probability distribution from which cryptographic secrets are drawn.

In a musical context, by considering the incoming dyad as analogous to a cryptographic code which is attempting to be "cracked" by an intelligent auditory system, we can consider that the analogous "worst-case attacker" would be a "best-case auditory system" which has complete awareness of the probability distribution for any incoming dyad. This analogy would view such an auditory system as actively attempting to choose the most probable rational, rather than drawing a rational at random weighted by the distribution.

The use of [math]a=∞[/math] min-entropy would reflect this view. In contrast, the use of [math]a=1[/math] Shannon entropy reflects a much "dumber" process which performs no such analysis and perhaps doesn't even seek to "choose" any sort of "victor" rational at all. As the parameter a interpolates between these two options, it can be interpreted as the extent to which the rational-matching process for incoming dyads is considered to be "intelligent" and "active" in this way.

Some psychoacoustic effects naturally fit into this paradigm, such as the virtual pitch integration process, which actually does attempt to find a single victor when matching incoming chords with chunks of the harmonic series. Other psychoacoustic effects, such as that of beatlessness, may instead be better viewed as "dumb" processes whereby nothing in particular is being "chosen," but where a more uniform distribution of matching rational numbers for a dyad simply generates a more discordant sonic effect. Different values of a can differentiate between the predominance given to these two types of effect in the overall construct of psychoacoustic concordance.

Certain values of [math]a[/math] reduce to simpler expressions and have special names, as given in the examples below.

## Examples

### a=0: Harmonic Hartley Entropy

[math]\displaystyle H_0(J|c) = \log |J|[/math]

where [math]|J|[/math] is the cardinality of the set of basis rationals. This assumes, in essence, an "infinitely dumb" auditory system which can do no better than picking a rational number from a uniform distribution completely at random. All dyads have the same Harmonic Hartley Entropy. The Hartley Entropy is sometimes called the "max-entropy," and is useful mainly as an upper bound on the other forms of entropy: all Rényi Entropies are always guaranteed to be less than the Hartley Entropy.

*Harmonic Hartley Entropy (a=0) with the basis set all rationals with Tenney height ≤ 10000. Note that the choice of spreading function makes no difference in the end result at all.*

### a=1: Harmonic Shannon Entropy (Harmonic Entropy)

[math]\displaystyle H_1(J|c) = -\sum_{j \in J} P(j|c) \log P(j|c)[/math]

This is Paul's original Harmonic Entropy. Within the cryptographic analogy, this can be thought of as an auditory system which simply selects a rational at random from the incoming distribution, weighted via the distribution itself.

*Harmonic Shannon Entropy (a=1) with the basis set all rationals with Tenney height ≤ 10000, spreading function a Gaussian distribution with s=1% (~17 cents), and [math]\sqrt{nd}[/math] weighting.*

### a=2: Harmonic Collision Entropy

[math]\displaystyle H_2(J|c) = -\log \sum_{j \in J} P(j|c)^2 = -\log (J_1 = J_2|c)[/math]

where [math]J_1[/math] and [math]J_2[/math] are two independent and identically distributed random variables of JI basis ratios, conditioned on the same incoming dyad [math]c[/math], and the collision entropy is the same as the negative log of the probability that the two JI variables produce the same outcome.

*Harmonic Collision Entropy (a=2) with the basis set all rationals with Tenney height ≤ 10000, spreading function a Gaussian distribution with s=1% (~17 cents), and [math]\sqrt{nd}[/math] weighting.*

### a=∞: Harmonic Min-Entropy

[math]\displaystyle H_\infty(J|c) = -\log \max_{j \in J} P(j|c)[/math]

This is the min-entropy, which simply takes the negative log of the largest probability in the distribution. This can be thought of as representing the "strength" of the incoming dyad from being "deciphered" by a "best-case" auditory system. The name "min-entropy" reflects that the [math]a=\infty[/math] case is guaranteed to be a lower bound among all Rényi entropies.

*Harmonic Rényi Entropy with a=7, with the high value of a being chosen to approximate min-entropy (a=*∞*). The basis set is still all rationals with Tenney height ≤ 10000, the spreading function a Gaussian distribution with s=1% (~17 cents), and the weighting function [math]\sqrt{nd}[/math].*

## Convolution-Based Expression For Quickly Computing Rényi Entropy

Below is given an derivation that expresses Harmonic Rényi Entropy in terms of two simpler functions, each of which is a convolution product and hence can be computed quickly using the Fast Fourier Transform.

The below derivation depends on the use of simple weighted probabilities, although it may be possible to extend to domain-integral probabilities instead.

### Preliminaries

The Harmonic Rényi Entropy is defined as

[math]\displaystyle \text{HE}_a(c) = H_a(J|c) = \frac{1}{1-a} \log \sum_{j \in J} P(j|c)^a[/math]

As before, we can write [math]P(j|c)[/math] as follows:

[math]\displaystyle P(j|c) = \frac{Q(j|c)}{\sum_{j \in J} Q(j|c)}[/math]

where [math]Q(j|c)[/math] is the "unnormalized" probability, and the denominator above is the sum of these unnormalized probabilities, so that all of the [math]P(j|c)[/math] sum to 1.

To simplify notation, we first rewrite the denominator as a "normalization" function:

[math]\displaystyle \psi(c) = \sum_{j \in J} Q(j|c)[/math]

and putting back into the original equation, we get

[math]\displaystyle H_a(J|c) = \frac{1}{1-a} \log \left( \sum_{j \in J} \left( \frac{Q(j|c)}{\psi(c)} \right)^a \right)[/math]

Since [math]\psi(c)[/math] is the same for each basis ratio, we can pull it out of the summation to obtain:

[math]\displaystyle H_a(J|c) = \frac{1}{1-a} \log \left( \frac{\sum_{j \in J} Q(j|c)^a}{\psi(c)^a} \right)[/math]

To simplify notation further, we can also rewrite the numerator, the sum of "raw" (unnormalized) pseudo-probabilities, as a function:

[math]\displaystyle \rho_a(c) = \sum_{j \in J} Q(j|c)^a[/math]

Finally, we put this all together to obtain a simplified version of the Harmonic Rényi Entropy equation:

[math]\displaystyle \text{HE}_a(c) = H_a(J|c) = \frac{1}{1-a} \log \left( \frac{\rho_a(c)}{\psi(c)^a} \right)[/math]

We thus reduce the term inside the logarithm to the quotient of the functions [math]\rho_a(c)[/math] and [math]\psi(c)[/math]. Our aim is now to express each of these two functions in terms of a convolution product.

### Convolution product for [math]\psi(c)[/math]

[math]\displaystyle \psi(c)[/math], the normalization function, is written as follows:

[math]\displaystyle \psi(c) = \sum_{j \in J} Q(j|c)[/math]

Again, [math]Q(j|c)[/math] is defined as follows:

[math]\displaystyle Q(j|c) = \frac{S(\cent(j)-c)}{\|j\|}[/math]

We can rewrite the above equation as a convolution with a delta distribution:

[math]\displaystyle Q(j|c) = \left(S \ast \frac{\delta_{-\cent(j)}}{\|j\|}\right)(-c)[/math]

Putting this back into the original summation, we obtain

[math]\displaystyle \psi(c) = \sum_{j \in J} \left(S \ast \frac{\delta_{-\cent(j)}}{\|j\|}\right)(-c)[/math]

We note that the left factor in the convolution product is always the same [math]S(-c)[/math], which is not dependent on [math]j[/math] in any way. Since convolution distributes over addition, we can factor the [math]S[/math] out of the summation to obtain

[math]\displaystyle \psi(c) = \left[S \ast \left(\sum_{j \in J} \frac{\delta_{-\cent(j)}}{\|j\|}\right)\right](-c)[/math]

We can clean up this notation by defining the auxiliary distribution K:

[math]\displaystyle K(c) = \sum_{j \in J} \frac{\delta_{-\cent(j)}}{\|j\|}[/math]

Which leaves us with the final expression:

[math]\displaystyle \psi(c) = \left[S \ast K\right](-c)[/math]

### Convolution product for [math]\rho_a(c)[/math]

The derivation for [math]\rho_a(c)[/math] proceeds similarly. Recall the function is written as follows:

[math]\displaystyle \rho_a(c) = \sum_{j \in J} Q(j|c)^a[/math]

The expression for each [math]Q(j|c)^a[/math] is:

[math]\displaystyle Q(j|c)^a = \frac{S(\cent(j)-c)^a}{\|j\|^a}[/math]

We can again express this as a convolution, this time of the function [math]S^a(-c)[/math], meaning the spreading function S taken to the a'th power, and a delta distribution:

[math]\displaystyle Q(j|c)^a = \left(S^a \ast \frac{\delta_{-\cent(j)}}{\|j\|^a}\right)(-c)[/math]

Putting this back into the original summation and factoring as before, we obtain

[math]\displaystyle \rho_a(c) = \left[S^a \ast \left(\sum_{j \in J} \frac{\delta_{-\cent(j)}}{\|j\|^a}\right)\right](-c)[/math]

And again we clean up notation by defining the auxiliary distribution

[math]\displaystyle K^a(c) = \sum_{j \in J} \frac{\delta_{-\cent(j)}}{\|j\|^a}[/math]

so that

[math]\displaystyle \rho_a(c) = \left[S^a \ast K^a\right](-c)[/math]

We have now succeeded in representing [math]\rho_a(c)[/math] as a convolution.

Note that the function [math]K^a(c)[/math] involves a slight abuse of notation, as it is not literally [math]K(c)[/math] taken to the [math]a[/math]'th power (as the square of the delta distribution is undefined). Rather, we are simply taking the weights of each delta distribution in the summation to the [math]a[/math]'th power.

### Round-up

Taking all of this, we can rewrite the original expression for Harmonic Rényi Entropy as follows:

[math]\displaystyle \text{HE}_a(c) = H_a(J|c) = \frac{1}{1-a} \log \left( \frac{\left[S^a \ast K^a\right](-c)}{\left[S \ast K\right]^a(-c)} \right)[/math]

where the expression

[math]\displaystyle \left[S \ast K\right]^a(-c)[/math]

represents the convolution of [math]S[/math] and </math>K, taken to the [math]a[/math]'th power, and flipped backwards. Note that if [math]S(x)[/math] is a symmetrical (even) spreading function, and if for each ratio [math]n/d[/math] in [math]J[/math], if the inverse [math]d/n[/math] is also in [math]J[/math], then the above convolution will also be symmetrical, and we also have

[math]\displaystyle \left[S \ast K\right]^a(-c) = \left[S \ast K\right]^a(c)[/math]

We have succeeded in representing Harmonic Rényi Entropy in simple terms of two convolution products, each of which can be computed in [math]O(N log N)[/math] time.

# Extending HE to [math]N=\infty[/math]: zeta-HE

All of the models described above involve a finite set of rational numbers, bounded by some weighting function, and where the weighting is less than some max value [math]N[/math].

It so happens that we are more or less able to analytically continue this definition to the situation where [math]N=\infty[/math]. More precisely, we are able to analytically continue the exponential of HE, which yields the same relative interval rankings as standard HE.

The only technical caveat is that we use the HE of the "unnormalized" probability distribution. However, in the large limit of [math]N[/math], this appears to agree closely with the usual HE. We go into more detail below about this.

Our basic approach is: rather than weighting intervals by [math](nd)^{0.5}[/math], we choose a different exponent, such as [math](nd)^2[/math]. For an exponent which is large enough (we will show that it must be greater than 1), HE does indeed converge as [math]N \to \infty[/math], and we show that this yields an expression related to the Riemann Zeta function. We can then use the analytic continuation of the zeta function to obtain an analytically continued curve for the [math](nd)^{0.5}[/math] weighting, which we then show empirically does indeed appear to be what HE converges on for large values of [math]N[/math].

In short, what we will show is that the Fourier Transform of Harmonic Shannon Entropy is given by

[math]|\zeta(0.5+it)|^2 \cdot \overline {\phi(t)}[/math]

where [math]\phi(t)[/math] is the characteristic function of the spreading distribution and [math]\overline {\phi(t)}[/math] is complex conjugation. Below we also give an expression for the Renyi entropy for arbitrary choice of the parameter [math]a[/math].

This enables us to speak cognizantly of the harmonic entropy of an interval as measured against *all* rational numbers.

## Background: Unnormalized Entropy

Our derivation only analytically continues the entropy function for the "unnormalized" set of probabilities, which we previously wrote as [math]Q(j|c)[/math]. For this definition to be philosophically perfect, we would want to analytically continue the entropy function for the normalized sense of probabilities, previously written as [math]P(j|c)[/math].

However, in practice, the "unnormalized entropy" appears to be an extremely good approximation to the normalized entropy for large values of [math]N[/math]. The resulting curve has the same minima and maxima as HE, the same general shape, and for all intents and purposes looks exactly like HE, just shifted on the y-axis.

Here are some examples for different values of [math]s[/math]. All of these are Shannon HE ([math]a=1[/math]), using [math]\sqrt{nd}[/math] weights, with unreduced rationals (more on this below), with the bound that [math]nd \lt 1000000[/math], just with different values of [math]s[/math]. All have been scaled so that the minimum entropy is 0, and the maximum entropy is 1:

As you can see, the unnormalized version is extremely close to a linear function of the normalized one. A similar situation holds for larger values of [math]a[/math]. The Pearson correlation coefficient of "rho" is also given, and is typically very close to 1 - for example, for [math]s=1%[/math], it's equal to 0.99922. The correlation also seems to get better with increasing values of [math]N[/math], such that the correlation for N=1,000,000 (shown above) is much better than the one for N=10,000 (not pictured).

In the above examples, note that there are slightly adjusted values of [math]s[/math] (usually by less than a cent) between the normalized and unnormalized comparisons for each plot. For example, in the plot for [math]s=1%[/math], corresponding to 17.2264 cents, we compare to a slightly adjusted UHE of 16.4764 cents. This is because, empirically, sometimes a very slight adjustment corresponds to a better correlation coefficient, suggesting that the UHE may be equivalent to the HE with a miniscule adjustment in the value of [math]s[/math].

It would be nice to show the exact relationship of unnormalized entropy to the normalized entropy in the limit of large [math]N[/math], and whether the two converge to be exactly equal (perhaps given some miniscule adjustment in [math]s[/math] or [math]a[/math]). However, we will leave this for future research, as well as the question of how to do an exact derivation of normalized HE.

For now, we will start with a derivation of the unnormalized entropy for [math]N=\infty[/math], as an interesting function worthy of study in its own right - not only because it looks exactly like HE, but because it leads to an expression for unnormalized HE in terms of the Riemann Zeta function.

## Derivation

For now, our derivation is limited to the case of [math]\sqrt{nd}[/math] Tenney-weighted rationals, although it may be possible to derive a similar result for [math]\max(n,d)[/math] weighting as well.

Additionally, because it simplifies the derivation, we will use *unreduced rationals* in our basis set, meaning that we will even allow unreduced fractions such as [math]4/2[/math] so long as [math]nd\lt N[/math] for our bound N. The use of HE with unreduced rationals has previously been studied by Paul and shown to be not that much different than HE with reduced ones. However, we will easily show later unreduced and reduced rationals converge to the same thing in the limit as [math]N \to \infty[/math], up to a constant multiplicative scaling.

### Definition of the Unnormalized Harmonic Rényi Entropy

Let's start by recalling the original definition for Harmonic Rényi Entropy, using simple weighted probabilities:

[math]\displaystyle \text{HE}_a(c) = \frac{1}{1-a} \log \sum_{j \in J} P(j|c)^a[/math]

Remember also that the definition of [math]P(j|c)[/math] is as follows:

[math]\displaystyle P(j|c) = \frac{Q(j|c)}{\sum_{j \in J} Q(j|c)}[/math]

where the [math]Q(j|c)[/math] is the "unnormalized probability" - the raw value of the spreading function, evaluated at the ratio in question, divided by the ratio's weighting. The above equation tells us that the normalized probability is equal to the unnormalized probability, divided by the sum of all unnormalized probabilities.

Putting the two together, we get

[math]\displaystyle \text{HE}_a(c) = \frac{1}{1-a} \log \sum_{j \in J} \left( \frac{Q(j|c)}{\sum_{j \in J} Q(j|c)} \right)^a[/math]

Now, for us to define the unnormalized HE, we simply take the standard Rényi entropy equation, and replace the normalized probabilities with unnormalized ones, yielding

[math]\displaystyle \text{UHE}_a(c) = \frac{1}{1-a} \log \sum_{j \in J} Q(j|c)^a[/math]

Using our convolution theorem from before, we can express the above as

[math]\displaystyle \text{UHE}_a(c) = \frac{1}{1-a} \log \left( S^a \ast K^a \right)(-c)[/math]

where, as before, [math]S^a[/math] is our spreading function, taken to the [math]a[/math]'th power, and [math]K^a[/math] is our convolution kernel, with the weights on the delta functions taken to the [math]a[/math]'th power as described previously.

Note that if [math]S[/math] is symmetric, as in the case of the Gaussian or Laplace distributions, then the inverted argument of [math](-c)[/math] on the end is redundant, and can be replaced by [math](c)[/math].

Lastly, it so happens that it will be much easier to understand our analytic continuation if we look at the exponential of the UHE times [math](1-a)[/math], rather than the UHE itself. The reasons for this will become clear later. If we do so, we get

[math]\displaystyle \exp((1-a) \text{UHE}_a(c)) = \left( S^a \ast K^a \right)(-c)[/math]

Note that this function is simply a monotonic transformation of the original, and so preserves the exact same concordance ranking on all intervals.

### Analytic Continuation of the Convolution Kernel

The definition for [math]K[/math] is:

[math]\displaystyle K(c) = \sum_{j \in J} \frac{\delta_{-\cent(j)}}{\|j\|}[/math]

where [math]\|j\|[/math] represents the weighting of the JI basis ratio [math]j[/math]. In the particular case of Tenney weighting, we get:

[math]\displaystyle K(c) = \sum_{j \in J} \frac{\delta_{-\cent(j)}}{(j_n \cdot j_d)^{0.5}}[/math]

where [math]j_n[/math] and [math]j_d[/math] are the numerator and denominator of [math]j[/math], respectively.

Although it may seem odd, we can take the Fourier transform of the above to obtain the following expression:

[math]\displaystyle \mathcal{F}\left\{K(c)\right\}(t) = \sum_{j \in J} \frac{e^{i t \cent(j)}}{(j_n \cdot j_d)^{0.5}}[/math]

Furthermore, for simplicity, we can change the units, so that rather than the argument being given in cents, it is given in "natural" units of "nepers", a technique often used by Martin Gough in his work on Logarithmic approximants. The representation of any interval in nepers is given by simply taking is natural logarithm. Doing so, by defining the change of variables [math]c = \frac{1200}{\log(2)}n[/math], we obtain

[math]\displaystyle \mathcal{F}\left\{K(n)\right\}(t) = \sum_{j \in J} \frac{e^{i t \log (j_n/j_d)}}{(j_n \cdot j_d)^{0.5}}[/math]

We can treat the presence of the logarithm within the exponential function as changing the base of the exponential, so that we get

[math]\displaystyle \mathcal{F}\left\{K(n)\right\}(t) = \sum_{j \in J} \frac{(j_n/j_d)^{i t}}{(j_n \cdot j_d)^{0.5}}[/math]

We can also factor each term in the summation to obtain

[math]\displaystyle \mathcal{F}\left\{K(n)\right\}(t) = \sum_{j \in J} \left[ \frac{{j_n}^{i t}}{j_n^{0.5}} \cdot \frac{{j_d}^{-i t}}{j_d^{0.5}} \right][/math]

which we can rewrite as

[math]\displaystyle \mathcal{F}\left\{K(n)\right\}(t) = \sum_{j \in J} \left[ \frac{1}{{j_n}^{0.5 -i t}} \cdot \frac{1}{{j_d}^{0.5 + i t}} \right][/math]

Now, we note our summation is currently written simply as [math]\sum_{j \in J}[/math]. For a Tenney height weighting, we typically bound by [math]\sqrt{nd} \lt N[/math] for some [math]N[/math]. However, although it is unusual, for the sake of simplifying the derivation, we will bound by [math]\max(n,d) \lt N[/math] instead, despite the use of Tenney height for our weighting. This will not end up being much of a problem, as the two will converge on the same result anyway.

Bounding by [math]\max(n,d) \lt N[/math] is the same as specifying that [math]j_n \lt N[/math] and [math]j_d \lt N[/math]. Doing so, we get

[math]\displaystyle \mathcal{F}\left\{K(n)\right\}(t) = \sum_{1\lt j_n, j_d\lt N} \left[ \frac{1}{{j_n}^{0.5 -i t}} \cdot \frac{1}{{j_d}^{0.5 + i t}} \right][/math]

We can now factor the above product to obtain:

[math]\displaystyle \mathcal{F}\left\{K(n)\right\}(t) = \left[ \sum_{j_n=1}^N \frac{1}{{j_n}^{0.5 -i t}} \right] \cdot \left[ \sum_{j_d=1}^N\frac{1}{{j_d}^{0.5 + i t}} \right][/math]

Now, we can see that as [math]N \to \infty[/math] above, the summations do not converge. However, incredibly enough, each of the above expressions has a very well-known analytic continuation, which is the Riemann zeta function.

To perform the analytic continuation, we temporarily change the [math]0.5[/math] in the denominator to some other weight [math]w\gt 1[/math]. This is equivalent to changing our original [math]\sqrt{nd}[/math] weighting to some other exponent, such as [math](nd)^2[/math] or [math](nd)^{1.5}[/math]. Doing this causes both of the summations above to converge, so that we obtain

[math]\displaystyle \mathcal{F}\left\{K(n)\right\}(t) = \left[ \sum_{j_n=1}^\infty \frac{1}{{j_n}^{w -i t}} \right] \cdot \left[ \sum_{j_d=1}^\infty\frac{1}{{j_d}^{w + i t}} \right][/math]

It has been well known for more than a century that both of these summations converge to the Riemann zeta function, so that we get

[math]\displaystyle \mathcal{F}\left\{K(n)\right\}(t) = \zeta(w-i t) \cdot \zeta(w+i t)[/math]

Rewriting as a function of a complex variable [math]z = w + i t[/math], and noting that the zeta function obeys the property that [math]\zeta(\overline z)=\overline{\zeta(z)}[/math], where [math]\overline s[/math] represents complex conjugation, we get

[math]\displaystyle \mathcal{F}\left\{K(n)\right\}(t) = \overline{\zeta(z)} \cdot \zeta(z) = |\zeta(z)|^2[/math]

And we have now obtained a very interesting result: if we had instead gone with something like [math](nd)^2[/math] weighting on rationals, rather than [math]\sqrt{nd}[/math], that our HE setup *would* have converged as [math]N \to \infty[/math], and our original HE convolution kernel would have been the Fourier transform of a particular vertical "slice" of the Riemann zeta function where [math]\Re(z) = 2[/math].

Furthermore, although the above series doesn't converge for [math]w = 0.5[/math], we can simply use the analytic continuation of the Riemann zeta function to obtain a meaningful function at that point, so that our original convolution kernel can be written as

[math]\displaystyle K(n) = \mathcal{F}^{-1}\left\{| \zeta(0.5+ t) |^2\right\}(n)[/math]

which is the inverse Fourier transform of the squared absolute value of the Riemann zeta function, taken at the critical line.

Lastly, to do some cleanup, we previously went with [math]\max(n,d)\lt N[/math] bounds, rather than [math]\sqrt{nd}\lt N[/math] bounds, despite using [math]\sqrt{nd}[/math] weighting on ratios. However, it is easy to show above that regardless of which bounds you use, both choices converge to the same function when [math]w \gt 1[/math] in the limit as [math]N \gt \infty[/math]. Since these series agree on this right half-plane of the Riemann zeta function, they share the same analytic continuation, so that we get the same result, despite using our technique to simplify the derivation. A good explanation fo this can be found on StackExchange here.

It is likewise easy to show that the function [math]K^a(n)[/math], taken from the numerator of our original Harmonic Rényi Entropy convolution expression, can be expressed as

[math]\displaystyle K^a(n) = \mathcal{F}^{-1}\left\{|\zeta(0.5a+ t) |^2\right\}(n)[/math]

so that the choice of [math]a[/math] simply changes our choice of vertical slice of the Riemann zeta function.

### Analytic Continuation of Unnormalized Harmonic Rényi Entropy

We can put this back into our equation for the Unnormalized Harmonic Rényi Entropy. To do so, we will continue with our change of units from cents to nepers, corresponding to a change of our variable from [math]c[/math] to [math]n[/math]. We will likewise assume the spreading probability distribution [math]S[/math] has been scaled to reflect the new choice of units.

Our original equation was

[math]\displaystyle \exp((1-a) \text{UHE}_a(n)) = \left( S^a \ast K^a \right)(-n)[/math]

Using our expression for [math]K^a[/math] as [math]N \to \infty[/math], we get

[math]\displaystyle \exp((1-a) \text{UHE}_a(n)) = \left( S^a \ast \mathcal{F}^{-1}\left\{|\zeta(0.5a+ t)|^2\right\} \right)(-n)[/math]

To simplify this, we will define an auxiliary notation for the zeta function as follows:

[math]\zeta_w(t) = \zeta(w + i t)[/math]

yielding the simplified expression:

[math]\displaystyle \exp((1-a) \text{UHE}_a(n)) = \left( S^a \ast \mathcal{F}^{-1}\left\{|\zeta_{0.5a}|^2\right\} \right)(-n)[/math]

We can simplify the expression of the above if we likewise take the Fourier transform of [math]S[/math]. If we do, we obtain the characteristic function of the distribution, which is typically denoted by [math]\phi(t)[/math]. We will use the following definitions:

[math]\displaystyle \phi(t) = \mathcal{F}\left\{S(n)\right\}(t)[/math]

[math]\displaystyle \phi_a(t) = \mathcal{F}\left\{S(n)^a\right\}(t)[/math]

Doing so, and noting that convolution becomes multiplication in the Fourier domain, we get

[math]\displaystyle \exp((1-a) \text{UHE}_a(n)) = \mathcal{F}^{-1}\left\{\phi_a \cdot |\zeta_{0.5a}|^2\right\}(-n)[/math]

Lastly, we note that for any function [math]f(x)[/math], we have [math]\mathcal{F}\left\{f(-x)\right\} = \mathcal{F}\left\{\overline {f(x)} \right\}[/math], where [math]\overline x[/math] is complex conjugation. For simplicity's sake, we can this write as [math]\mathcal{F}\left\{\overline f\right\}[/math]. Putting that all together, we get

[math]\displaystyle \exp((1-a) \text{UHE}_a(n)) = \mathcal{F}^{-1}\left\{\overline \phi_a \cdot |\zeta_{0.5a}|^2\right\}[/math]

where we can drop the overline on [math]|\zeta_{0.5a}|^2[/math] because it is purely real, and its complex conjugate is itself.

## Examples

It is very easy to see empirically that our expression does seem to converge be the thing that UHE converges on in the limit of large N. Here are some examples for different values of s and a, showing that as N increases it converges on the our analytically continued "zeta HE."

In all these examples, the left plot is a series of plots of UHE at different N, with the zeta HE being a slightly thicker line in the background. The right plot is the largest N plotted against zeta HE, being almost a perfect line.

Note also that we don't plot the UHE directly, but rather [math]\exp((1-a)UHE)[/math], as described previously. The units have also been converted back to cents. Each HE function has been scaled so that the minimum entropy is 0 and the maximum entropy is 1.

Lastly, note that, due to our multiplication of UHE by [math](1-a)[/math] above, the UHE would be flipped upside down relative to what we're used to, with higher values corresponding to more concordant intervals. In the pictures below we have flipped it back upside down, for consistency with the earlier pictures.

### s=0.5%, a=1.00001

### s=1%, a=1.00001

### s=1.5%, a=1.00001

Note that in all these plots, the value of [math]a[/math] is chosen to be [math]1.00001[/math] rather than exactly [math]1[/math], so as to avoid that [math](1-a)[/math] term becoming 0. Similar results are seen for other choices of [math]a[/math]:

### s=1%, a=2.2

Note that you have to be careful if you choose to compute the analytically continued [math]a=2[/math] numerically: this corresponds to the vertical slice of zeta function at [math]\Re(z) = 1[/math], where there is a pole.

## Equivalence of exp-UHE and UHE for [math]a \leq 2[/math]

Let's go back to our original convolution expression for finite-[math]N[/math] UHE:

[math]\displaystyle \text{UHE}_a(c) = \frac{1}{1-a} \log \left( S^a \ast K^a \right)(-c)[/math]

We will also, for now, use the notation

[math]U(c) = \exp((1-a) \text{UHE}_a(c)) = \left( S^a \ast K^a \right)(-c)[/math]

to get

[math]\displaystyle \text{UHE}_a(c) = \frac{1}{1-a} \log U(c)[/math]

Now let's consider the auxiliary function [math]\tilde{U}(c) = U(c) - U(0)[/math], which gives us a "shifted" version of the UHE which has the UHE of 1/1 normalized to 0. Then we can re-express the UHE as follows:

[math]\displaystyle \text{UHE}_a(c) = \frac{1}{1-a} \log \left(U(0) + \tilde{U}(c) \right)[/math]

We can expand the above into a Taylor series as follows:

[math]\displaystyle \text{UHE}_a(c) = \frac{1}{1-a} \left(\log(U(0)) + \frac{\tilde{U}(c)}{U(0)} - \frac{\tilde{U}(c)^2}{U(0)^2} + \frac{\tilde{U}(c)^3}{U(0)^3} - ...\right)[/math]

Now, suppose we only care about the behavior of this function up to a constant vertical shift and scaling. Then we can drop the [math]\frac{1}{1-a}[/math] term, since it's just a constant scaling, and also get rid of the [math]\log(U(0))[/math] term, which simply subtracts a constant offset. This leaves us with

[math]\displaystyle \text{UHE}_a(c) \sim \left(\frac{\tilde{U}(c)}{U(0)} - \frac{\tilde{U}(c)^2}{U(0)^2} + \frac{\tilde{U}(c)^3}{U(0)^3} - ...\right)[/math]

where [math]\sim[/math] denotes the two sides are now "equivalent" up to a constant shifting and scaling.

Finally, we can go one step further and multiply the above by the constant [math]U(0)[/math]. Doing so, we get

[math]\displaystyle \text{UHE}_a(c) \sim \left(\tilde{U}(c) - \frac{\tilde{U}(c)^2}{U(0)} + \frac{\tilde{U}(c)^3}{U(0)} - ...\right)[/math]

And we now have a function that is equivalent to our original, up to a constant shift and scaling, but which is fairly easy to analyze asymptotically.

Now, for the Shannon entropy, we know that the above doesn't converge as [math]N \to \infty[/math]. Furthermore, after our above analysis, we have seen that it doesn't converge for [math]a \leq 2[/math], corresponding to the region of the Riemann zeta function where [math]\Re(s) \lt 1[/math].

However, in the case of [math]a \leq 2[/math], we note empirically that while the function diverges, it diverges in a certain "uniform" sense. That is, as [math]N[/math] increases, a constant vertical offset is added to [math]U(c)[/math], so that the function blows up to infinity. However, if this vertical offset is corrected for, for example by subtracting U(0), the resulting curve doesn't seem to grow at all, but rather shrinks in height slightly until it seems to converge. We can see this from the following plot:

In other words, we can see that as [math]N[/math] increases, the growth rate of [math]U(0)[/math] dwarfs that of [math]\tilde{U}(c)[/math], which does not seem to grow at all.

Assuming this is true (which we will not prove here, but which seems self-evident given all of the other results), we can go back to our last expression for exp-UHE:

[math]\displaystyle \text{UHE}_a(c) \sim \left(\tilde{U}(c) - \frac{\tilde{U}(c)^2}{U(0)} + \frac{\tilde{U}(c)^3}{U(0)} - ...\right)[/math]

and see that as [math]N \to \infty[/math], the [math]U(0)[/math] terms blow up, whereas the [math]\tilde{U}(c)[/math] terms do not. As a result, all of the terms with [math]U(0)[/math] in the denominator disappear, and we are left with:

[math]\displaystyle \text{UHE}_a(c) \sim \tilde{U}(c)[/math]

Putting this together from our prior realization that [math]\text{UHE}_a(c) \sim \log \tilde{U}(c)[/math], we get

before, we get

[math]\displaystyle \tilde{U}(c) \sim \log \tilde{U}(c)[/math]

additionally, noting that [math]\log \tilde{U}(c) \sim \frac{1}{1-a} \log \tilde{U}(c) = \log \tilde{U}(c)^{\frac{1}{1-a}} \sim \tilde{U}(c)^{\frac{1}{1-a}} = \exp(\text{UHE_a}(c))[/math], we get

[math]\displaystyle \text{UHE}_a(c) \sim \exp(\text{UHE_a}(c))[/math]

as [math]N \to \infty[/math], for [math]a \leq 2[/math].

We note that this proof is entirely dependent on the growth rate of [math]\tilde{U}(c)[/math] being dwarfed by that of [math]U(0)[/math]. This is a fairly weak conjecture to make, given that empirical evidence suggests something much stronger - that not only does it grow more slowly, but that it seems to converge. In particular, it seems to converge on our analytic continuation from before. However, a strict proof of any of these things would be nice.

Note again that this does not hold for [math]a \gt 2[/math], where the graph does display a very large difference between UHE and exp-UHE.

## Why not Normalized HE?

On the surface, everything we did with the convolution theorem, and subsequent analytic continuation, should appear to work for normalized HE as well. For example, let's review our result for the exp of unnormalized HE:

[math]\displaystyle \exp((1-a) \text{UHE}_a(n)) = \mathcal{F}^{-1}\left\{\overline \phi_a \cdot |\zeta_{0.5a}|^2\right\}[/math]

Our original expression for normalized HE was

[math]\displaystyle \exp((1-a) \text{HE}_a(n)) = \frac{\left[S^a \ast K^a\right](-c)}{\left[S \ast K\right]^a(-c)}[/math]

Naively, you might expect we could simply apply the same analytic continuation technique to the numerator and denominator. If we do so, we would get

[math]\displaystyle \exp((1-a) \text{HE}_a(n)) = \frac{\mathcal{F}^{-1}\left\{\overline \phi_a \cdot |\zeta_{0.5a}|^2\right\}}{\mathcal{F}^{-1}\left\{\overline \phi \cdot |\zeta_{0.5}|^2\right\}^a}[/math]

However, to see why this doesn't work, let's compare the analytically continued version of the denominator (i.e. the normalization term) with the finite versions. Look at the following picture, which is a plot of [math]\mathcal{F}^{-1}\left\{\overline \phi \cdot |\zeta_{0.5}|^2\right\}^a[/math]:

This picture shows how the denominator changes as [math]N[/math] increases: you can see that in general, the function is shifted upward, increasing without bound. The thin plots reflect this for N=1000, 5000, 10000, 50000, and 100000, where you can see them increasing.

You will note that the denominator also looks exactly like unnormalized HE, just upside down. Normalized HE is the quotient of two functions that both look like this, which are slightly different. This quotient produces the usual HE curve, which is flipped upside down relative to the denominator, and which also increases without bound. That all these functions increase without bound is just another way to state that these things generally don't converge as [math]N \to \infty[/math].

However, look at what happens with our analytic continuation, which is given by the thicker blue line at the bottom. Despite our sequence of finite-[math]N[/math] denominator terms increasing on the y-axis, the analytically continued version suddenly "snaps" back to zero. Although the curve shape is roughly the same, the vertical offset is almost completely eliminated when the analytic continuation is done.

The problem here is that the original HE function was the quotient of two very large, strictly positive functions - the numerator and denominator. However, performing the analytic continuation on each separately has caused both to "snap" back to zero, so that the denominator, while retaining the same shape, now has points where it touches the x-axis. As a result, the quotient of the two will have poles where the denominator is zero.

The resulting quotient of analytically continued functions looks like this, and does not remotely resemble HE:

Those "spikes" are poles where the denominator is zero.

The problem is that we're really stretching the boundaries of complex analysis with this. With unnormalized HE, we were able to analytically continue the Fourier transform of exp-UHE to obtain a concrete expression in terms of the Riemann zeta function. While complex analysis makes no guarantees on the behavior of the Fourier transform of the analytic continuation of a holomorphic function, we did see the result seemed to converge on exp-UHE in the limit of large [math]N[/math] when transforming back from the Fourier domain, confirming empirically that our analytically continued expression seemed to make sense.

But in the case of "normalized HE," we analytically continued the Fourier transforms of the numerator and denominator, separately, transformed both out of the Fourier domain, and then took the quotient. Complex analysis *really* makes no guarantee on the behavior of the quotient of two Fourier transforms of the analytic continuations of holomorphic functions, and in this case the behavior is very strange. A different approach to analytically continuing the expression would be required.

This same principle explains why we plotted the exp of UHE, rather than UHE itself. Were we to take the log of finite UHE, we would be taking the log of a strictly positive function. However, the analytically continued exp-UHE snaps back to the x-axis, so that there are points where the function is zero or even negative. Taking the log of the analytically continued exp-UHE would yield a complex-valued function where it is negative, due to this snapping effect. However, looking at exp-UHE directly has no such problem.

Finally, it is noteworthy that for [math]a\gt 2[/math], we end up looking at slices of the zeta function for which [math]\Re(z)\gt 1[/math]. This is where our original unnormalized HE function should converge as [math]N \to \infty[/math], corresponding to the region where the Riemann zeta function Dirichlet series converges. For these values of [math]a[/math], the exp-UHE *is* positive. So, we can take the log again and look at the usual UHE. This can be useful for plotting, since exp-UHE tends to "flatten" out the curve for high values of [math]a[/math], whereas taking the log accentuates the minima and maxima (and more closely resembles the usual HRE).

## Interpretation as a New Free Parameter: the Weighting Exponent

In our original derivation of the analytic continuation, we temporarily changed the weighting for rationals from [math](nd)^{0.5}[/math] to some other [math](nd)^w[/math], with [math]w \gt 1[/math], for the sake of obtaining a series that converges. We then changed the exponent back to [math]0.5[/math].

This can be thought of as giving us another free parameter to HE, in addition to [math]s[/math] and [math]a[/math]: the exponent for the weighting for each rational. That is, although Paul originally derived the [math](nd)^{0.5}[/math] exponent empirically by studying the behavior of mediant-to-mediant HE for Tenney-bounded rationals, there is no reason we can't simply that exponent to something else. As shown before, so long as that exponent is greater than 1, unnormalized HE will converge in the limit as [math]N -\gt \infty[/math], and will converge to the same thing whether we are bounding [math]nd \lt N[/math], [math]\max(n,d) \lt N[/math], or anything else (see again here). We can then analytically continue to the case where [math]w \lt 1[/math].

If we add this as a third parameter, called [math]w[/math] we can modify our definition of exp-UHE as follows:

[math]\displaystyle \exp((1-a) \text{UHE}_{a,w}(n)) = \mathcal{F}^{-1}\left\{\overline \phi_a \cdot |\zeta_{w a}|^2\right\}[/math]

So that our vertical slice of the zeta function is given by $\Re(z) = w\cdot \a$.

## Equivalence of the Weighting Exponent and [math]a[/math] for Generalized Normal Distributions

We get a very interesting result if our spreading distribution is a generalized normal distribution, which a family that encompasses both the Gaussian and the Laplace distributions (sometimes referred to as the "Vos curve" in Paul's work).

Let's go back to our three-parameter definition of exp-UHE above:

[math]\displaystyle \exp((1-a) \text{UHE}_{a,w}(n)) = \mathcal{F}^{-1}\left\{\overline \phi_a \cdot |\zeta_{w a}|^2\right\}[/math]

We can see that, in a sense, the need for both [math]a[/math] and [math]w[/math] is almost redundant. Their product specifies the vertical slice of the zeta function. If you set [math]w=0.5[/math] and [math]a=1[/math], corresponding to the Shannon entropy with [math]\sqrt{nd}[/math] weighting, you get the same vertical slice as if you set [math]w=0.25[/math] and [math]a=2[/math], corresponding to the collision entropy with [math]^4\sqrt{nd}[/math] weighting: in both cases this is the critical line of the zeta function.

The only reason that these expressions are different is due to the [math]\phi_a[/math] above. We had previously defined that as:

[math]\displaystyle \phi_a(t) = \mathcal{F}\left\{S(n)^a\right\}(t)[/math]

or, the Fourier transform of the spreading distribution, raised to the power of [math]a[/math]. So if you hold the product [math]w a[/math] as constant, but change the balance of [math]w[/math] and [math]a[/math], you will indeed get different results, simply because only the choice of [math]a[/math] changes the [math]\phi_a[/math].

However, we get a very neat result if we are using the generalized normal distribution. In that case, if we take the generalized normal distribution to a power [math]a[/math], we get another instance of the same generalized normal distribution. The difference is, the variance will be divided by [math]a^{\frac{1}{\beta}}[/math], where [math]\beta[/math] is the shape parameter for the distribution (a value of 1 is the Laplace distribution, a value of 2 is the Gaussian distribution, etc). The whole distribution will also no longer have an integral of 1, since we have also raised the scaling coefficient to a power, but this won't change anything, as it just corresponds to a uniform scaling of the end result.

In practice, what this means is that if you are using one of the above distributions, and you change [math]a[/math], this is *equivalent* to changing the weighting exponent [math]w[/math], and tweaking the standard deviation [math]s[/math] according to the above equation.

This gives us a very nice interpretation of our [math]a[/math] coefficient from HRE: it basically represents the weighting exponent on the rationals, with a corresponding adjustment to the standard deviation. The collision entropy [math]a=2[/math] with the standard weighting [math]\sqrt{nd}[/math] is totally equivalent to the Shannon entropy [math]a=1[/math] with the weighting [math]nd[/math] on the rationals, so long as the value of [math]s[/math] is adjusted according to the equation above. However, it should be noted that this definition only holds for the "unnormalized HRE" given above.

## Reduced Rationals Only

In our derivation, we assumed the use of unreduced rationals. It turns out that with a minor adjustment, the same model gives us reduced rationals, up to a constant multiplicative scaling. Let's go back to our analytic continuation of the convolution kernel, for some arbitrary weighting:

[math]\displaystyle \mathcal{F}\left\{K(n)\right\}(t) = \sum_{j \in J} \frac{e^{i t \log (j_n/j_d)}}{(j_n \cdot j_d)^{w}}[/math]

Now, suppose we want to analytically continue this so that the set [math]J[/math] is the set of all reduced rational numbers. We can first do so by starting again with unreduced rationals, but expressing each rational not as [math]\frac{n}{d}[/math], but rather as [math]\frac{n}{d} \cdot \frac{c}{c}[/math], where [math]n'[/math] and [math]d'[/math] are coprime, and [math]c[/math] is the gcd of both. For example, we would express [math]\frac{6}{4}[/math] as [math]\frac{3}{2} \cdot \frac{2}{2}[/math]. Doing so, and assuming that we denote the set of unreduced rationals by [math]\mathbb{U}[/math], we get the following equivalent expression of the same convolution kernel above:

[math]\displaystyle \mathcal{F}\left\{K(n)\right\}(t) = \sum_{j \in \mathbb{U}} \frac{e^{i t \log (\frac{j_c j_{n'}}{j_c j_{d'}})}}{(j_c j_{n'} \cdot j_c j_{d'})^{w}} = |\zeta(w+i t)|^2[/math]

where the last equality is what we proved before. Note that this is the same exact function as before, just written such that the GCD of the unreduced fraction has an explicit term.

The [math]\frac{j_c}{j_c}[/math] in the log in the numerator cancels out, but in the denominator we have an extra factor of [math]{j_c}^2[/math] to contend with. This yields

[math]\displaystyle |\zeta(w+i t)|^2 = \sum_{j \in \mathbb{U}} \frac{e^{i t \log (\frac{j_{n'}}{j_{d'}})}}{({j_c}^2 \cdot j_{n'} j_{d'})^{w}} = \sum_{j \in \mathbb{U}} \left[ \frac{1}{{j_c}^{2w}} \cdot \frac{e^{i t \log (\frac{j_{n'}}{j_{d'}})}}{(j_{n'} j_{d'})^{w}} \right][/math]

Now, assuming we have [math]w\gt 1[/math] and everything is absolutely convergent, we can factor this into a product of series as follows:

[math]\displaystyle |\zeta(w+i t)|^2 = \left[ \sum_{j_c \in \mathbb{N}^+} \frac{1}{{j_c}^{2w}} \right] \cdot \left[ \sum_{j \in \mathbb{Q}} \frac{e^{i t \log (\frac{j_{n'}}{j_{d'}})}}{(j_{n'} j_{d'})^{w}} \right][/math]

where the left summation now has [math]j_c \in \mathbb{N}^+[/math], the set of strictly positive rational numbers, and the right summation now has [math]j \in \mathbb{Q}[/math] the set of reduced rationals. Note again that the product above yields all unreduced rationals, thanks to the [math]j_c[/math].

Now, note that that left series is, itself, just another Dirichlet series that converges to the zeta function. We have

[math]\displaystyle |\zeta(w+i t)|^2 = \zeta(2w) \cdot \left[ \sum_{j \in \mathbb{Q}} \frac{e^{i t \log (\frac{j_{n'}}{j_{d'}})}}{(j_{n'} j_{d'})^{w}} \right][/math]

and now we are done. The right series is the thing that we want, representing the Fourier transform of the convolution kernel where only reduced fractions are allowed. To get that, we simply divide the whole thing by [math]\zeta(2w)[/math]:

[math]\displaystyle \frac{|\zeta(w+i t)|^2}{\zeta(2w)} = \sum_{j \in \mathbb{Q}} \frac{e^{i t \log (\frac{j_{n'}}{j_{d'}})}}{(j_{n'} j_{d'})^{w}}[/math]

This function then becomes our new [math]\mathcal{F}\left\{K(n)\right\}[/math].

However, you will note that [math]\zeta(2w)[/math] is a constant not depending at all on [math]t[/math]. As a result, the reduced rational kernel is exactly equal to the unreduced rational kernel, times a constant depending only on [math]w[/math]. This means that when we take the inverse Fourier transform and convolve, the result for exp-UHE will likewise be identical, scaled only by a constant.

As a result, we have shown that we get the same exact results for reduced and unreduced rationals, differing only by a multiplicative scaling.

Lastly, you will note that for the special value [math]w=0.5[/math], corresponding to the usual [math]\sqrt{nd}[/math] weighting, we end up dividing by the term [math]\zeta(1)[/math]. This is the only pole in the zeta function, so we wind up dividing by infinity, making the entire function zero, as pointed out by Martin Gough. However, as we can get arbitrarily close to [math]w=0.5[/math] and still exhibit the behavior that the unreduced and reduced functions are scaled versions of one another, we can simply use the unreduced version of exp-UHE for [math]w=0.5[/math] and consider it equivalent to reduced exp-UHE in the limit.

# To Do

There are a number of things that need to be added to this article. Below are listed some for reference:

- 3HE, both for finite HE and for [math]N -\gt \infty[/math]
- write-up of fast computation for infinite zeta-UHE, perhaps with a zeta table
- addition of many more pictures

# References

Harmonic entropy (TonalSoft encyclopedia)