Skip to main content
Wiley Open Access Collection logoLink to Wiley Open Access Collection
. 2025 Oct 3;1553(1):363–378. doi: 10.1111/nyas.70037

Hidden assumptions of integer ratio analyses in bioacoustics and music

Yannick Jadoul 1,, Tommaso Tufarelli 2, Chloé Coissac 1, Marco Gamba 3, Andrea Ravignani 1,4,5
PMCID: PMC12645274  PMID: 41043082

Abstract

Rhythm is ubiquitous in human culture and in nature, but hard to capture in all its complexity. A key dimension of rhythm, integer ratio categories occur when the relationship between temporal intervals can be expressed as small‐integer ratios. Recent work has found integer ratio categories in most human musical cultures and some animal species' vocalizations or behavioral displays. But biological systems are noisy, and empirically measured intervals rarely form an exact small‐integer ratio. Here, we mathematically assess whether a leading integer ratio analysis method makes valid statistical and biological assumptions. In particular, we (1) make the temporal properties of empirical ratios explicit, both in general and for the typical use in the literature; (2) show how the choice of ratio formula affects the probability distribution of rhythm ratios and ensuing statistical results; (3) guide the reader to carefully consider the assumptions and null hypotheses of the statistical analysis; and (4) present a comprehensive methodology to statistically test integer ratios for any null hypothesis of choice. Our observations have implications for both past and future research in music cognition and animal behavior: They suggest how to interpret past findings and provide tools to choose the correct null hypotheses in future empirical work.

Keywords: animal behavior, beat, categorical rhythm, meter, statistical assumptions, timing, vocalization

INTRODUCTION: INTEGER RATIOS AS A MEASURE OF RHYTHMIC STRUCTURE

Quantifying temporal patterns is a crucial step toward a better understanding of behavioral rhythm. One common pattern in human music are intervals whose durations conform to small‐integer ratios; for example, the durations of a quarter note and an eight note are related in a 2:1 ratio. 1 Small‐integer ratios are ubiquitous across individuals and cultures, and appear to be a striking universal feature of human music and rhythm. 2 , 3 , 4 Humans both perceive and produce small‐integer ratios in temporal sequences; a few hypotheses exist about the underlying mechanisms. 5 , 6 Investigating the presence of rhythmic categories conforming to small‐integer ratios is one way of measuring temporal structure in behavior, and perhaps even inferring internal mechanisms of the biological agents producing those behaviors. 7 , 8 , 9

In search for the evolutionary origins of this fundamental building block of rhythm perception and production, recent studies have started quantifying the degree to which small‐integer ratios occur across animals. Small‐integer ratios are not only found in human music, 1 , 3 , 4 but also in a broad range of nonhuman animal behaviors: Many species show high rates of isochrony (1:1 ratio) in their vocal displays (e.g., several gibbon species, 10 , 11 , 12 orangutans, 13 , 14 rock hyraxes, 15 zebra finches 1 ). Several other species show different rhythmic categories (e.g., thrush nightingales, 1 indri, 8 , 16 Australian pied butcherbirds 17 ). In other studies, while no clear rhythmic categories are found in vocalizations, 18 rhythmic categories are present in nonvocal displays (e.g., harbor seals' flipper slapping, 19 chimpanzee drumming 20 ) or behavior (e.g., horse gait 21 , 22 ).

Here, we dissect a recent approach for quantifying and testing the presence of rhythmic categories at small‐integer ratios in empirical data, pioneered in nonhuman animals by Roeske et al. 1 In the following section (Section “Background”), we first shortly outline this approach. Next, we present a summary of our findings using minimal mathematical jargon and formulas (Section “Summary of results”). The unconcerned reader can read the section containing an example analysis (Section “An example”) right after this summary, skipping the three middle sections where we go into full detail about the mathematical and statistical results. To guide the interpretation of the presented equations, Table 1 provides a summary of the notation used. In the subsequent section (Section “An example”), we demonstrate the presented formulas with a fully worked‐out example based on previous empirical data. Finally, we present our results and takeaway message for the empirical scientist intending to apply the approach, refraining from technical details (Section “Conclusion and takeaway message”).

TABLE 1.

Overview of the mathematical notation used in derivations and formulas below.

Notation
m:n
integer ratio between two intervals
ik
empirical interval in a sequence
qk
empirical fraction between two interval durations in a sequence
rk
empirical rhythm ratio in a sequence
rm:n
rhythm ratio of a perfect integer ratio
i1, i2 any two intervals
q
fraction between two interval durations
r
standard rhythm ratio of two intervals
s
any alternative rhythm ratio of two intervals
F
any transformation function from i1,i2 to s
f
any transformation function from q to s
pI
probability density functions of intervals
pQ, pR, pS probability density functions of q, r, s
PQ, PR, PS cumulative probability functions of q, r, s
w
width of an rk ratios bin
w^I,u,v
normalization factor for a bin [u,v] for intervals distributed as pI
c
auxiliary variable representing a scaling factor
t
auxiliary variable of integration
λ
rate of a Poisson point process
a, b minimum and maximum of a uniform interval distribution pI
μ, σ logarithm of location and scale of a log‐normal distribution

BACKGROUND: HOW INTEGER RATIOS ARE TYPICALLY MEASURED

How to investigate the presence of integer ratios in empirical temporal sequences is not a trivial problem. Any measurement of a biological system produces noisy data. In the temporal domain, such noise will make it challenging to detect small‐integer ratios precisely. Hence, we need a methodological framework that allows us to statistically assess whether the observed ratios are concentrated around small‐integer ratios. The approach presented here starts from an observation of a potentially rhythmic temporal sequence (e.g., the notes in a piece of music or the onset of calls in an animal display; see Figure 1). To test for rhythmicity in such a sequence, we measure the intervals between events, and investigate whether the ratios between subsequent intervals approximate a small‐integer ratio more often than expected. This then allows us to determine whether or not those rhythmic patterns are prominent aspects of that temporal sequence.

FIGURE 1.

FIGURE 1

The rhythmic analysis of an audio recording starts by extracting the events and intervals. For example, these events could be the onset of syllables in a thrush nightingale's song 27 (A), the onset of the musical notes played by a trumpet 28 , 29 (B), or the drumbeats of a dundun 28 , 29 (C). The rhythm ratio formula (D, Equation 1) is applied to each pair of adjacent intervals in a temporal sequence (A– C), resulting in a distribution of rhythm ratios (D). The region around a specific integer ratio is split into several parts: an on‐integer ratio bin and two off‐integer ratio bins (dark and light shading). The fraction of rhythm ratios rk contained in each bin is then normalized by the total number of ratios and by the total bin width. The resulting values are empirical estimates of the probability density per bin (e.g., 1:1 and 3:1, E, F), which are then compared using a statistical test. Note that we have combined both off‐integer ratio bins into a single normalized bin count, following De Gregorio et al. 8 and others. Other studies keep both normalized bin counts separated. 1 In the context of the current study, the difference between these approaches is irrelevant.

Below, we examine and generalize the approach by Roeske et al., 1 which analyzes the ratio between adjacent intervals in temporal sequences. Limiting the analysis to only adjacent intervals has the practical advantage of ignoring a slow drift in tempo throughout a sequence. Moreover, while such analysis may fail to explicitly capture other rhythmic patterns, the presence of small‐integer ratios implies nonrandom relationships between adjacent intervals. Measuring ratios between adjacent intervals also captures the local rhythmic structure found in human perception and production. 3

The rhythm ratio rk quantifies the relationship between two adjacent intervals as the ratio of one interval duration to their joint duration. 1 More precisely, given a sequence of temporal intervals1 i1,i2,,ik,,in>0, rk is calculated as follows:

rk=ikik+ik+1 (1)

The resulting ratio rk has several straightforward and enticing properties:

  • 1.

    Only the relative duration of intervals matters:

    For example, two adjacent intervals of 1 and 2 s result in a rhythm ratio of rk=1s1s+2s=13=0.33; two much shorter intervals of 25 and 50 ms with the same relative durations have the exact same ratio: 25ms25ms+50ms=13=0.33

  • 2.

    Ratios rk are bounded between 0 and 1:

    For example, for two intervals of extremely different durations, 1 ms and 1 s (i.e., 1000 ms), the rk ratio is almost 0 (1ms1ms+1000ms=0.000999) or almost 1 (1000ms1ms+1000ms=0.999000), depending on the intervals' order.

  • 3.

    Inverse integer ratios, m:n and n:m, sum to 1 and are symmetric around 1:1:

    For example, the rhythm ratio corresponding to integer ratio 1:2 is 11+2=0.33, and the rhythm ratio corresponding to integer ratio 2:1 is rk=22+1=0.66=111+2; so, integer ratios 1:2 and 2:1 have rk values that add up to 1 and are symmetric around 0.5 (which itself corresponds to 1:1, or isochrony).

Note that for a simpler formula, such as the direct fraction of two intervals qk=ik+1ik (e.g., Ref. 2), properties 2 and 3 do not hold: If one uses this formula to quantify ratios, the q values for integer ratios 2:1, 3:1, and 4:1 lie very close together between 0.25 and 0.5, while the inverse integer ratios 1:2, 1:3, and 1:4 are spread out between 2 and 4. Conversely, the corresponding rk values lie equally close to each other; that is, the respective rk ranges, from 0.66… to 0.8 and from 0.2 to 0.33…, are equal in size. However, there could be other, unexplored ratio measures which might feature these same or other convenient properties (see below, Section “Summary of results”).

After calculating the rhythm ratios in empirical interval sequences (Figures 1A–C), the resulting distribution of rhythm ratios can be inspected for clusters around small‐integer ratios (i.e., peaks in the density distribution; see Figure 1D). The approach taken by recent research is to count the number of rhythm ratios falling into several bins around small‐integer ratios (see Figures 1D–F). In short, to statistically test the presence of an integer ratio peak in the data, all points within a region of interest around that ratio are assigned to one of two groups: on‐integer ratio and off‐integer ratio values; the former data points fall in the vicinity of the integer ratio, while the latter away from it (for details, see Refs. 1, 8).

These counts in different bins around an integer ratio can then be compared using a statistical method, through, for example, bootstrapping, 1 a Wilcoxon signed‐rank test, 8 , 12 or generalized linear mixed models. 10 , 13 Crucially, in most cases, the proposed on‐ and off‐integer bins 1 do not have the same width; hence, their respective counts are normalized by dividing each count by its bin's width before statistically comparing them. If a bin of width w contains m rhythm ratio values out of a total of N values, this normalized count equals mN·w. 1 In other words, one divides the relative proportion of observations in that bin by the width of the bin. This normalization intends to allow for a fair comparison between proportions of observations that fall into broad bins, which by pure chance are likely to contain many observations, with those that fall into narrow bins.

Roughly speaking, regardless of the statistical test one decides to employ, the test's null hypothesis will be that there is no difference between the number of on‐peak ratios and the number of off‐peak ratios. A significant result associated with a higher number of on‐peak rk values for a certain small‐integer ratio will, therefore, indicate a discrepancy sufficient to suggest that a particular rhythmic category is present. As we demonstrate below, this view is actually simplistic: Slight variations in the calculation of rk and the bins' normalization may correspond to nontrivial deviations in the resulting statistical analysis.

This approach has proven convenient and practical to investigate the presence of integer ratios in temporal sequence data. Three main methodological questions with practical implications seem to have remained unexplored:

  • Many formulas could capture the ratio between two intervals. What are the mathematical properties of the rhythm ratio rk formula? (Section “Mathematical properties of rhythm ratios”)

  • Statistical tests make mathematical assumptions about the data and its distribution. How does the choice of formula affect the probability distribution of rhythm ratios and the ensuing statistical results? (Section “Statistical inference of rhythm ratios”)

  • Assumptions which do not match reality or inappropriate null hypotheses might render a statistical result useless. Are there alternative ways of transforming intervals to visualize and test integer ratios, making different assumptions or offering different advantages? (Section “Choosing a different null hypothesis”)

SUMMARY OF RESULTS: THE IMPORTANCE OF THE RIGHT FORMULA AND NULL HYPOTHESIS

Many formulas could capture the ratio between two intervals. What are the mathematical properties of the rhythm ratio

rk

formula?

The rk formula presented above (Equation 1) is only one way of measuring the relationship between two intervals, as there are many other formulas that could quantify the relationship between ik and ik+1 in a single value. One property of the rk formula is nevertheless particularly noteworthy: The resulting rk value is not affected by the overall tempo of a temporal sequence. This is usually a desirable property when investigating rhythmic patterns and the presence of small‐integer ratios. Further analysis shows that there are indeed infinite alternative ratio formulas with this same property; for any such formula to be tempo‐independent, it can only depend on the fraction of the two intervals.

However, the choice of a ratio formula also influences the resulting distribution of the ratios. For example—as noted in Section “Background”—using the straightforward fraction of two intervals (ik+1ik) introduces asymmetry. Consequently, the chosen rhythm ratio formula plays a crucial role in the visualization and analysis of a temporal sequence.

We can investigate the mathematical properties of the rk ratio by examining different random processes and the rk values they generate. Our derivations show how the random distribution of a sequence's interval durations (ik) determines the resulting distribution of rk values (see Figure 2 and Equation 7). One mathematical process for generating temporal events and intervals provides particular insight into the rk formula: the Poisson point process. The expected distribution of rk ratios resulting from such a highly random process is a uniform distribution; that is, for a Poisson process, independent of its rate parameter, the rk values will equally likely fall anywhere between 0 and 1. As probability distributions and statistics are strongly connected, the choice of ratio formula affects any subsequent statistical inference.

FIGURE 2.

FIGURE 2

Four example probability distributions of interval durations ik (left column) all result in differently shaped probability distributions of rhythm ratios rk (middle column). The theoretically derived probability density functions (full lines) match the distributions of rk in a randomly sampled sequence of 1000 intervals (histograms). A half‐normal, a uniform, and a log‐normal distribution of intervals all result in nonuniform distributions of rk with a higher density around rk=0.5 (ratio 1:1). Normalizing bin counts by bin width results significantly more often in a higher normalized count of on‐1:1 ratio rk (right column, 1000 sequences of 1000 intervals, Wilcoxon signed‐rank test; half‐normal distribution: T=209379, p<105; uniform distribution: T=515, p<10163; log‐normal distribution: T=98, p<10164). For an exponential distribution of intervals (i.e., a Poisson process), the normalized on‐ and off‐ratio counts for 1:1 do not differ significantly (Wilcoxon signed‐rank test; T=235643, p0.11).

Statistical tests make mathematical assumptions about the data and its distribution. How does the choice of formula affect the probability distribution of rhythm ratios and the ensuing statistical results?

A key feature of statistical analyses is the assumption of a null hypothesis, formally specifying the statistical properties of the data if the studied effect were not present. Most studies performing statistical tests on the rhythm ratios rk count the number of ratios that fall close to a small‐integer ratio. As described earlier (Section “Background”), these counts are then normalized by bin width and statistically compared. Crucially, the count of a bin that is wider will be divided by a higher normalization factor: This means that under the null hypothesis, the number of rk values in each bin is proportional to a bin's width. This null hypothesis, resulting from the normalization by a bin's width, is equivalent to assuming that rk values are randomly and uniformly distributed between 0 and 1.

This brings us to a key result: Previous studies have implicitly compared their empirical rk distributions against a specific null hypothesis. This implicit null hypothesis is that the data has been generated by a Poisson process. Any significant results based on normalized bin counts (following Refs. 1, 8) statistically show that the data was not generated by a random Poisson process. Even if this null hypothesis was not made explicit, statistical significance shows that the animal vocalizations feature rhythmic categories different from one particular type of random timing. One important caveat follows, as a Poisson process generates highly random temporal sequences: This implicitly assumed null hypothesis represents a relatively low bar for testing the presence of rhythm categories, and is not necessarily an indication of extraordinary rhythmic capabilities. As Figure 2 shows, any limit on the possible range of interval durations produced by an animal (e.g., because of biomechanical constraints) will also cause rk values to be significantly different from a uniform distribution.

Assumptions which do not match reality or inappropriate null hypotheses might render a statistical result useless. Are there alternative ways of transforming intervals to visualize and test integer ratios, making different assumptions or offering different advantages?

During statistical analysis, the distribution of observed rhythm ratios is usually compared to a random baseline distribution, representing the null hypothesis. As we show, this baseline is often a uniform distribution of ratios, which corresponds to a Poisson process and an exponential distribution of interval durations (Figure 2). However, in certain cases, one might want to test the empirical data against a different null hypothesis. For example, if previous observations or theoretical considerations make it likely that an animal's produced interval durations fall between a minimum and a maximum value, a uniform distribution over this range can be a natural null hypothesis. This case corresponds to the simulated null distributions and Kolmogorov–Smirnov tests in a handful of previous studies. 8 , 15 , 17 , 18 Similarly, when a behavior is hypothesized to produce log‐normally distributed intervals, one might want to test whether certain integer ratios occur more frequently than expected under a log‐normal distribution of intervals.

To test empirical data against a different null hypothesis, there are two options: The first option is to use a different rhythm ratio formula (instead of rk; Equation 1), which would result in a uniform distribution of rhythm ratios for the chosen baseline distribution of intervals (see Equations 12 and 13). The second, often more practical option, is to weight the data according to its expected probability. For example, following the mathematical derivations we present below, we can calculate the probability that a rhythm ratio rk falls between 0.4 and 0.5 or between 0.3 and 0.4 if the interval durations are uniformly distributed. To statistically test whether rhythm ratios fall more than expected by chance in one of these two bins, the counts should be normalized by the expected probability of each bin (i.e., the area under the null distribution's curve), instead of by the bin widths. A comparison between these normalized bin counts will correctly take into account the underlying null hypothesis of the intervals' distribution. Our worked‐out example (Section “An example”) demonstrates the latter approach by applying it to the same empirical data analyzed by Roeske et al. 1

MATHEMATICAL PROPERTIES OF RHYTHM RATIOS

Scale‐invariant rhythm ratio formulas

The three properties of rk mentioned above are straightforward to show mathematically:

  • 1.

    Only the relative duration of intervals matters:

    Scaling ik,ik+1 by a factor c>0, the resulting rhythm ratio remains the same, as rk=c·ikc·ik+c·ik+1=ikik+ik+1=rk.

  • 2.

    Ratios rk are bounded between 0 and 1:

    For any interval durations ik,ik+1>0, the value of rk is bounded by 0=00+ik+1<rk<ikik+0=1.

  • 3.

    Inverse integer ratios, m:n and n:m, sum to 1 and are symmetric around 1:1:

    Two equally sized intervals will have a ratio r1:1=11+1=12=0.5.

    Two inverse ratio rhythms, rm:n and rn:m, will add up to 1, as rn:m+rm:n=nn+m+mm+n=1. It follows that those two ratios will be symmetrically distributed around 0.5, as 12rn:m+rm:n=12=r1:1.

Reasonably, the underlying tempo should not affect how the relative relationship between intervals is quantified; in other words, we require any rhythm ratio formula to be scale‐invariant.

Consider any rhythm ratio transformation function F(i1,i2), with interval durations i1,i2>0. If we impose the scale invariance condition, then for any c>0,

F(ci1,ci2)=F(i1,i2). (2)

We can show that F must only depend on the ratio q=i2i1, that is, F(i2,i1)=f(q) for some function f. As proof, choose c=1i1, such that ci1=1 and ci2=i2i1q. Then, F(i1,i2)=F(ci1,ci2)=F(1,q), such that it is easy to identify

f(q)=F(1,q). (3)

In conclusion, we can reduce any scale‐invariant quantification of relative rhythm (between two intervals) to a single‐variable function f(q) that is only dependent on the fraction q=i2i1. Conversely, any function f(q) of q=i2i1 is scale‐invariant by construction, since the ratio does not change when both intervals are scaled by the same constant: ci2ci1=i2i1=q.

The rhythm ratio rk (Equation 1; Ref. 1) fulfills this property, the explicit form of the corresponding f being f(q)=11+q=i1i1+i2. This scale‐invariance neatly fits with the observation that a sequence of n intervals ik has only n1 rk values. This difference is a single degree of freedom, lost when calculating a temporal sequence's rk values, and directly corresponds to the average tempo—or equivalently, the mean ik value.

The probability distribution of rhythm ratios

In a perfectly isochronous sequence, all intervals are the same size (i.e., ratio 1:1). In this case, all the sequence's qk and rk values will be equal, since qk=11=1 and rk=11+1=0.5. Similarly, for other temporal sequences built using only a few exact interval durations, the resulting distribution of qk and rk will contain a limited number of unique values. However, a more interesting case occurs when intervals can vary continuously and are randomly sampled from a certain probability distribution.

A first step toward understanding the behavior of rhythm ratios is calculating the probability distribution of the ratio q. Assuming two intervals i1,i2>0 distributed according to a joint probability density pI(i1,i2), we can calculate the resulting probability density function of the ratio q, pQ(q). To find the two‐dimensional probability distribution over i1 and q=i2i1, we replace variable i2 and write it in function of q, setting i2=qi1: pI(i1,q)=i1pI(i1,qi1). Subsequently, to calculate the marginal probability distribution over only q, we rename i1 to t, and integrate over all possible values of t. Performing this change of variables from (i1,i2)(t,qt), and integrating over t, simplifies to the following equation 23 , 24 :

pQ(q)=0tpI(t,qt)dt. (4)

Additionally, if i1 and i2 are statistically independent, pI(i1,i2)=pI(i1)pI(i2), and Equation (4) reduces to

pQ(q)=0tpI(t)pI(qt)dt. (5)

Given pQ(q), we can also determine the probability density pS of any rhythm ratio s given its transformation f(q)=s, if f has an inverse f1. If so, another change of variables in the probability density function, qs=f(q), shows that

pS(s)=pQf1(s)ddsf1(s). (6)

Altogether, for the rhythm ratio r=i1i1+i2, where f(q)=11+q, f1(r)=1rr, the probability density function pR(r) works out to be

pR(r)=1r20tpI(t)pIt1rrdt. (7)

So, given a probability distribution of interval durations ik, with density pI, Equation (7) describes the resulting probability density of resulting ratios rk.

Rhythm ratios of a Poisson point process

One relevant stochastic process to consider for generating a temporal sequence is a Poisson point process. A homogeneous Poisson process with rate λ generates events in time which are completely independent from each other and are separated by exponentially distributed intervals; that is,

pI(i)=λeλi (8)

where 1/λ is the average waiting time for an event to occur.

A Poisson point process is often considered a “highly random” way of generating a temporal sequence. The events in a Poisson process happen independently from each other, and the process is “memoryless” in the sense that the likelihood of an event only depends on the system's current state and not its past history. Poisson processes are routinely used to mathematically model, for example, the radioactive decay of individual atoms.

Following Equations (5) and (7), it can be shown that the probability distributions of the ratios q and r resulting from two independent and exponentially distributed intervals equal

pQ(q)=1(1+q)2 (9)

and

pR(r)=1. (10)

In other words, independently of a constant rate λ, the rhythm ratios rk=ikik+ik+1 (as proposed in Ref. 1) of intervals generated by a Poisson process follow a uniform distribution on [0,1]. The Poisson process's rate λ does not appear in the equations pQ(q) and pR(r) because its value corresponds to the average tempo. As shown earlier, q, r, and any other scale‐invariant rhythm ratio capture the rhythmic relationship between two interval durations independently of the underlying tempo.

This is an attractive property of the rk rhythm ratio, since a Poisson point process produces highly random temporal sequences and is thus a good baseline to which other sequences can be compared. One could even argue that in many ways, it is a maximally random process for temporal sequences, as the exponential distribution is the maximum entropy distribution on [0,) with a fixed mean of 1λ. In plainer terms, this means that a Poisson process generates the least predictable sequence of intervals (for intervals with unbounded duration and a fixed mean). Correspondingly, the uniform distribution of rk ratios is the distribution which achieves maximum entropy on a bounded interval (such as [0,1]). This connection between a Poisson process and a uniform distribution of rk values, both maximum entropy distributions, provides a strong theoretical justification for the use of the rk rhythm ratio formula.

STATISTICAL INFERENCE OF RHYTHM RATIOS

As shown above, the widely used rhythm ratio rk=ikik+ik+1 results in a uniform distribution over [0,1] for a Poisson point process (i.e., when ik and ik+1 follow an exponential distribution with the same rate λ; see exponential distribution of intervals in Figure 2). This mathematical link has consequences for statistical analyses based on this ratio rk.

Many statistical approaches compare an empirical sample of ratios rk to a uniform distribution, implicitly or explicitly. Any such statistical analysis implicitly assumes the null hypothesis that the data was generated by a homogeneous Poisson point process. This holds, in particular, for the typical approach where calculated rhythm ratios are binned and those counts are normalized by the bin widths (see Section “Background” and Figure 1; e.g., Refs. 1, 8). Normalizing bin counts by their width implies a baseline distribution where the expected number of points in each bin is directly proportional to its width. In other words, this division normalizes bin counts with respect to a uniform distribution over [0,1]. So, a statistical comparison between normalized bin counts de facto tests the implicit null hypothesis that the (binned) empirical rhythm ratios have been generated by a Poisson process. This goes for most studies which have relied on the rk ratio formula: Almost all of past research has tested, without explicit mention of it, whether the empirical data is clustered around certain small‐integer ratios significantly more than a Poisson point process.

Many studies have assumed a Poisson point process as the null hypothesis. This is an important nuance, since a Poisson process produces highly random rhythmic sequences with no strict limitation on the possible interval durations. Therefore, a Poisson process represents a relatively weak null hypothesis to be rejected. Depending on the rhythmic sequences or animal displays being investigated, studies might want to test a stricter null hypothesis, perhaps tailored to the species, behavior, or experimental task at hand. For example, intervals ik sampled independently from a uniform distribution, subject to a minimum and maximum interval duration, will result in ratios rk which do not follow a uniform distribution (see simulated null distributions and Kolmogorov–Smirnov tests in Refs. 8, 15, 17, 18). In fact, further analyses and computational simulations show that this distribution has a peak around integer ratio 1:1, causing the null hypothesis to be rejected more often than by chance (see Figure 2). This is a correct rejection; that is, the intervals were not generated by a Poisson process. However, this null hypothesis might not be the one authors should be trying to reject, since the resulting peak at 1:1 is still due to a random process (albeit a different one than a Poisson process).

CHOOSING A DIFFERENT NULL HYPOTHESIS

Alternative null hypotheses

As illustrated above, the use of the rhythm ratio rk=ikik+ik+1 often goes hand in hand with one key assumption: In statistical tests, empirical data is compared to a Poisson process as null hypothesis. However, depending on the research question and its context, a Poisson process might not be the most appropriate choice. For example, a biological system may feature physiological constraints such as a minimum duration between two movements or a maximum lung capacity. These constraints limit the range of the interval distribution by imposing minimum and maximum durations, while a Poisson process has a nonzero probability of generating extremely small or extremely large intervals. Temporal sequences generated according to such additional constraints are, mathematically speaking, less random than a Poisson process, and will exhibit different statistical properties.

Using the above mathematical insights, we can, however, test different types of “nonrandomness” in rhythmic patterns by changing the null hypothesis. There are two ways of testing empirical data against a chosen null hypothesis: implementing a matching rhythm ratio transformation (Subsection “Rescaling the rhythm ratio”), or weighting the data correspondingly (Subsection “Adjusting the normalization constant”).

For the first option, ratio rescaling, we can derive a new rhythm ratio formula, s=f(q), which transforms the desired null distribution of intervals pI(i) into a uniform distribution of ratios pS(s)=1. For example, Figure 2 demonstrates how a uniform distribution of intervals results in a nonuniform distribution of rk values. However, an appropriate choice of rescaled rhythm ratio formula s=f(q) effectively “flattens” the rhythm ratio distribution; any subsequent statistical analysis can again correctly assume a uniform distribution for the rescaled ratio s, even though the null hypothesis is now different from a Poisson process.

For the second option, data normalization, we stick to the rk rhythm ratio formula and instead adapt the statistical analysis. The issue with a nonuniform null distribution is that not all observed ratio values are equally likely. For example, given the previously mentioned uniform distribution of intervals (Figure 2), there is a higher chance of getting an rk value close to 0.5 than to 0.2. So, the expected number of rk values within a bin is not directly proportional to the width of the interval anymore; that is, the normalization by bin width 1 underestimates—and thus undercorrects—the expected number of points in a bin such as [0.5,0.6]. Below, we show how to calculate the correct normalization factor. Note that this second option is most often the most straightforward one, and will be easier to practically apply in future research.

Rescaling the rhythm ratio

Based on Equations (5) and (6), we can calculate the resulting probability distribution pS(s) of any rescaling s=f(q) of the rhythm ratio q, for any null distribution of independently sampled intervals pI(i). Conversely, we can start from a given interval distribution pI(i) and a desired distribution of rhythm ratios. For example, a convenient choice is to impose that the variable s (our new rhythm ratio quantification) is uniformly distributed over [0,1], that is, pS(s)=1 for 0s1, pS(s)=0 otherwise. From these constraints, we can determine the corresponding ratio transformation function f, which achieves this goal.

To do so, we calculate the cumulative probability distribution of both distributions, PQ(q)=0qpQ(x)dx and PS(s)=0qpS(x)dx, and combine them to get the two resulting transformation functions2, f+(q)=PS1PQ(q) and f(q)=PS11PQ(q). Since we have chosen pS(s)=1 for s[0,1], we simply have PS(s)=s, in turn implying PS1(t)=t. For independently sampled intervals with density pI(i), the cumulative distribution function of the ratio is

PQ(q)=0qpQ(x)dx=0q0tpI(t)pI(xt)dtdx. (11)

Consequently, the two possible ratio transformations resulting in a uniform probability density over [0,1] of ratios are

f+(q)=PQ(q)=0qpQ(x)dx (12)

and

f(q)=1PQ(q)=10qpQ(x)dx, (13)

where pQ and PQ are calculated as in Equations (5) and (11).

It can be helpful to note how these general formulas reduce to what we already know for a Poisson null hypothesis. For the specific case of a Poisson point process, with pI(i)=λeλi (Equation 8), we know that pQ(q)=1(1+q)2 (Equation 9). The two corresponding rhythm ratio formulas for a uniform ratio distribution over [0,1] (cfr. Equations 12 and 13) are

f+(q)=q1+q (14)

and

f(q)=11+q. (15)

Note that Equation (15) results in the familiar rk=f(qk)=11+qk=ikik+ik+1 which indeed results in a uniform distribution of rhythm ratios.3

Adjusting the normalization constant

Finding a custom ratio transformation by solving Equations (12) and (13) can be nontrivial, particularly if one wants to find an explicit formula. As an alternative approach to the ratio rescaling of the previous section, it is possible to pick any rhythm ratio (such as rk), and change the normalization or weight of the data points during the statistical analysis. For example, in the common binning approach (Section “Background”), we can normalize the bin counts differently. Note that the expected number of rhythm ratios that fall within a bin, under a specific null hypothesis, is proportional to the probability that a single rhythm ratio value falls inside this bin. So, instead of a division by the corresponding bin width, counts should be normalized by the total probability mass within a specific bin.

More formally, assume a null hypothesis distribution of interval durations pI. To statistically test whether the number of rhythm ratios in two bins differs significantly with respect to this null hypothesis, the count of rhythm ratios s in a bin b=[u,v] should be divided by a normalization factor w^I,u,v:

w^I,u,v=uvpS(s)ds, (16)

where pS(s) is the resulting distribution of ratios s (following Equation 6). This normalization constant replaces the bin width in the normalization formula. For a bin [u,v], which contains m out of N observed rhythm ratios, the normalized bin count is mN·w^I,u,v.

Note that Equation (16) simplifies to vu (i.e., the width of bin [u,v]) for a uniform distribution pS=1. Therefore, under the Poisson process null hypothesis, this normalization factor equals the bin width, as commonly used in previous work (e.g., Refs. 1, 8). This confirms again, from a different perspective, how the commonly used normalization by bin width implicitly assumes a Poisson process as null hypothesis—or more generally, it assumes that the distribution pS(s) is flat.

If the integral in Equation (16) is hard or impossible to solve analytically, it can be approximated numerically. In most cases, a simple Monte Carlo simulation suffices to approximate w^I,u,v: Sample many pairs of (pseudo)random values from the null hypothesis distribution pI, calculate the resulting ratios (e.g., following the common rk formula), count how many randomly sampled ratios fall within the [u,v] bin, and calculate the corresponding fraction of the total number of samples; the resulting number is an estimate of the normalization factor w^I,u,v (see Supporting Information, Algorithm S1 for pseudocode and an example implementation in Python and R in Code fragments S1 and S2). For a large enough number of samples, the resulting value will approximate the normalization constant w^I,u,v.

If we apply this new normalization to the previous sample of uniformly distributed intervals, we can change the null hypothesis and test for statistically significant differences with respect to this uniform distribution. The high rate of statistically significant results we saw before is strongly reduced, confirming our derivation of the correct normalization factor. Both ways of adapting the null hypothesis, a rescaled rhythm ratio and a corrected normalization factor, are demonstrated in Figure 3 for a uniform null distribution of intervals.

FIGURE 3.

FIGURE 3

For a Poisson process null hypothesis, a uniform distribution of ratios results does not result in a uniform distribution of rk rhythm ratios (top row). In fact, a statistical comparison between 1000 sequences of 1000 intervals shows a significant different between on‐ and off‐ratio bins for the 1:1 integer ratio (top right; Wilcoxon signed‐rank test; T=515, p<10163; see Figure 2). A different rhythm ratio formula results in a uniform distribution of the rescaled rhythm ratio sk (middle row; for details, see Supporting Information, Section S1). Given this rhythm ratio formula, there is no significant difference between the on‐ and off‐ratio bins (Wilcoxon signed‐rank test; T=244399, p0.52). Alternatively, the bin count normalization of the original rk ratios can be changed (bottom row). If we do not divide by the bin width w, but instead by the area under the probability density curve within a bin (w^I,u,v; Equation 16), statistical comparison is also nonsignificant (Wilcoxon signed‐rank test; T=238765, p0.21). Both nonsignificant results demonstrate how these approaches explicitly change the statistical null hypothesis.

AN EXAMPLE: TESTING ALTERNATIVE NULL HYPOTHESES ON BIRDSONG AND MUSIC DATA

Based on the above results, we now present a fully worked‐out example. We use four of the datasets originally analyzed by Roeske et al. 1 and test the empirical rk rhythm ratios against two new null hypotheses: We test which small‐integer ratios appear more often than chance if the intervals were (1) randomly sampled from a uniform distribution or (2) randomly sampled from a log‐normal distribution. We start by deriving the respective null distributions of rk values, and then calculate the area under the curve in each of the on‐ and off‐integer ratio bins. Once we have obtained these, we can correctly use them as normalization factors for the empirical data's bin counts.

Uniformly distributed intervals

A practically relevant null hypothesis is a uniform distribution of interval durations. Several studies have simulated the resulting ratio distribution and compared their observed distribution using a Kolmogorov–Smirnov test. 8 , 15 , 17 , 18 Such an approach statistically compares the two distributions globally, but does not test the presence of peaks around specific integer ratios. As shown in Figure 2, if intervals are uniformly distributed, the resulting rk rhythm ratio distribution is more concentrated around 1:1 (rk=0.5) than around other integer ratios. Consequently, the default normalization by bin width would underestimate the null distribution's expected bin count in the on‐1:1 ratio bin, and thus overestimate the empirical concentration around the 1:1 ratio.

With the results and equations presented above, we can deduce the exact ratio distribution if intervals are uniformly distributed. Afterward, we calculate the normalization factor to correctly test the bins around 1:1 against a uniform distribution. The full derivation of these formulas can be found in the Supporting Information (Section S1) and serves as an example of how to concretely apply our mathematical findings in other contexts. The unconcerned reader can skip the details of the following equations and continue reading the next subsection.

As intervals are distributed according to a uniform distribution over [a,b], the probability density of the rhythm ratio r (following Equation 7; shown in the top middle pane in Figure 3) is

pR(r)=12(ba)2b1r2ar2ifaa+br<1212(ba)2br2a1r2if12rba+b0otherwise (17)

To determine the total probability on an interval [u,v] (i.e., the area under the probability density function between u and v), we then need to calculate PR(v)PR(u), where PR(r) is the cumulative distribution function:

PR(r)=0if0r<aa+b12(ba)2b21r+a2r(a+b)2ifaa+br<12112(ba)2b2r+a21r(a+b)2if12r<ba+b1ifaa+br1 (18)

The new normalization factor for a bin [u,v] can now be calculated as PR(v)PR(u). As a concrete example, assume a uniform distribution of interval durations between 0.25 and 1.75 (as in Figures 2 and 3): The resulting normalization factors for the on‐ratio and off‐ratio bins around the 1:1 ratio (as used by Roeske et al. 1 and others) are higher than the corresponding bin's widths:

w^U[0.25,1.75],0.44,0.5=w^U[0.25,1.75],0.5,0.550.13264, (19)
w^U[0.25,1.75],0.4,0.44=w^U[0.25,1.75],0.55,0.60.08727. (20)

Note how Equations (17) and (18) are specified in terms of a and b. However, the absolute values of a and b do not influence the distribution of r; only the relative durations of the maximum and the minimum interval—that is, the fraction ba—changes the shape of the ratio distribution.4 For example, uniformly distributed intervals between 10 and 50 ms result in the same rhythm ratio distribution as intervals between 3 and 15 s; in both cases, the maximum interval is five times the minimum (i.e., ba=5), and the rhythm ratios ignore differences in average tempo. On the contrary, uniformly distributed intervals between 10 and 40 ms will result in a rhythm ratio distribution with a shape different from the one above.

Log‐normally distributed intervals

The second alternative null hypothesis we consider below is a log‐normal distribution. Akin to a common normal distribution, a log‐normal distribution naturally models a process producing interval durations which are concentrated around a single value with added random noise. Crucially, the log‐normal distribution has two theoretical advantages over the normal distribution: Random sampling from a log normal distribution will always produce a positive value, matching the property that measured interval durations cannot be negative. Moreover, many aspects of perception, including timing, are known to scale logarithmically, 25 , 26 and can thus be connected to a log‐normal distribution.

We now derive the null distribution of rk values when interval durations are randomly sampled from a log‐normal distribution. Once again, the unconcerned reader can skip the following equations and continue reading the following subsection.

A log‐normal distribution is characterized by two parameters, μ and σ, respectively the mean and standard deviation in log‐space. Under this parametrization, the null distribution of interval durations is:

pI(i)=1iσ2πe(ln(i)μ)22σ2 (21)

The quotient of two log‐normally distributed variables itself also follows a log‐normal distribution, resulting in the following null probability density and cumulative distribution functions of the rhythm ratio r:

pR(r)=1r(1r)2σπe(ln(1r)ln(r))24σ2 (22)

and

PR(r)=Φln(r)ln(1r)σ2=121+erfln(r)ln(1r)2σ (23)

The corresponding normalization factor, w^I,u,v, can be calculated by subtracting PR(v)PR(u). Taking, for example, σ=0.3, as in Figure 2, the normalization factors are again larger than the bins' widths:

w^LN(μ,0.3),0.44,0.5=w^LN(μ,0.3),0.5,0.550.20054, (24)
w^LN(μ,0.3),0.4,0.44=w^LN(μ,0.3),0.55,0.60.12985. (25)

Finally, note that the log‐normal distribution's scale parameter μ does not feature in Equations (22) and (23). Similar to the λ parameter for a Poisson process and to the absolute scale of the a and b parameters of a uniform distribution, μ corresponds to the average interval duration and thus to the underlying tempo. As a consequence, the μ parameter cancels out in the calculations of pR(r).

Rhythm ratios in birdsong and music

Based on the normalization factors derived above, we reanalyzed four of the datasets originally presented, 1 and explored how the alternative null hypotheses (i.e., uniformly or log‐normally distributed interval durations) influence the results. All data is openly available, and we refer to the previous study for exact details on recordings, onset annotations, and temporal intervals. 1 , 27 , 28 We selected both songbird datasets, consisting of intervals extracted from recordings of thrush nightingale (Luscinia luscinia) and zebra finch (Taeniopygia guttata) songs. We also chose the Cuban salsa and Malian jembe as two distinct examples to test the presence of small‐integer ratios in human music (part of the Interpersonal Entrainment in Music Performance corpus 29 , 30 ).

Starting from these datasets, we largely followed the same method of analysis as previously described 1 : We calculated the rk rhythm ratio between all pairs of adjacent intervals that had a longer “cycle duration” (i.e., ik+ik+1) than the thresholds determined in this previous study. To shortly explain these thresholds, the original analysis 1 found that interval pairs with a short combined duration mostly featured isochrony (dubbed “unimodal rhythms”) and that most small‐interval ratios occur in pairs of intervals with a longer duration (“flexible rhythms”). We followed this approach and focused our analysis on the flexible rhythms, using the same thresholds. One minor difference to the previous study is that we also removed outliers by only using interval durations inside the 1st to 99th percentile of the data. This did not visibly affect the distribution of the rk values, but removed some very long intervals, which were likely not part of the rhythmic structure and allowed us to better choose a fitting null hypothesis (see below).

We sorted the obtained rk values into the same bins as previous work; for example, the on‐integer ratio bin for the 2:1 ratio (i.e., around 13) contains all rk values between 13.25 and 12.75, while the two off‐integer ratio bins range from 13.5 to 13.25 and from 12.75 to 12.5. Following the previous approach, 1 we also combined the corresponding bin counts for the 1:2 and 2:1 ratios, and for the 1:3 and 3:1 ratios.

A key step where our analyses differ from previous studies is the normalization of these bin counts. For our demonstration here, we chose three different null hypotheses against which to test the empirical rk distributions, which all resulted in a different normalization factor for each bin:

  • The first normalization is the one used in previous work (e.g., Refs. 1, 8). We divided the relative bin count (i.e., the number of rk values in a bin, divided by the overall total number of rk values) by the bins' width. This normalizes the empirical data with respect to the null hypothesis that intervals were generated by a Poisson process; that is, the null distribution of the ik values is an exponential distribution.

  • For the second normalization, we determined the expected distribution of rk values for interval durations uniformly sampled between the minimum and maximum observed duration (see Subsection “Uniformly distributed intervals”).

  • The third normalization assumed a log‐normal distribution of interval durations. We fitted a log‐normal distribution to the observed intervals, and calculated the expected rk distribution in this case (see Subsection “Log‐normally distributed intervals”).

The rk distributions matching each of the three null hypotheses are shown in Figures 4 and 5. For these cases, it was possible to derive the exact formulas to calculate the bins' normalization constant. Note, however, that we could have also approximated these constants numerically (see Subsection “Adjusting the normalization constant” for details). We provide an example of how to do so in Python and in R in the Supporting Information (Code fragments S1 and S2).

FIGURE 4.

FIGURE 4

Our reanalysis of the thrush nightingale (A) and zebra finch (B) datasets 27 demonstrates the influence of the chosen null hypothesis on the statistical results. We compare three selected null hypotheses: exponentially distributed (i.e., a Poisson process; blue), uniformly distributed (green), and log‐normally distributed interval durations (red). The density plots present the estimated probability density of the observed rk ratios, and the expected distribution under each of the three null hypotheses. The bar charts present the bin frequencies, normalized according to each null hypothesis. The symbols on top (* and n.s.) show whether two adjacent on‐/off‐integer ratio bins contain a significantly different number of rk values. The statistical results right above each bar show whether a bin's normalized count is statistically significant from the null distribution.

FIGURE 5.

FIGURE 5

Our reanalysis of the Cuban salsa (A) and Malian jembe (B) datasets 28 , 29 shows some potentially interesting differences between the different null hypotheses. The plots can be read and interpreted analogously to Figure 5.

We performed the same bootstrapping procedure as described in detail by Roeske et al. 1 In short, we resampled each rk distribution with replacement 1000 times in order to estimate the confidence intervals of each normalized bin count. Accordingly, when the Bonferroni‐adjusted confidence intervals of an on‐ and off‐integer ratio bin did not overlap, we considered the two normalized bin counts to be statistically significantly different. In addition to comparing on‐ and off‐integer ratio bins, we also tested whether each bin in itself contained significantly more or less observations than expected by chance. To do so, we tested whether the confidence intervals of the normalized bin frequencies contained the value 1, the normalized bin count expected under the null hypothesis. These results indicate whether a bin contains significantly more or less rk values than the null hypothesis. The results of these statistical tests are shown in Figures 4 and 5; a table with all results is included as Supporting Information (Tables S1–S4), as well as the Python code used to perform the analysis.

Figure 4 shows the results of our approach for the two songbird corpora. Figure 5 shows the reanalysis of the Cuban salsa and Malian jembe datasets. One general trend emerges from these four examples: The uniform distribution (in blue) is typically a worse fit for the rk distributions than the two other null distributions. As a reminder, the two other null distributions (in green and red) are also achieved by sampling intervals randomly and independently; that is, none of the null hypotheses have a specific mechanism that fine‐tunes pairs of intervals to conform to small‐integer ratios. As a result of this closer fit of the uniform and log‐normal null hypotheses, most normalized bin counts end up much closer to 1 than for the Poisson process null hypothesis. These observations confirm our previous assertion that a Poisson process and its associated uniform rk distribution is a relatively low bar to statistically reject. Zooming out, our results also demonstrate how the commonly used statistical approach is suitable for evaluating any hypothesis‐driven interval relation, be it small‐ or large‐integer ratios, or even noninteger ratios: The corresponding rk null distributions and normalization factors can be derived mathematically or approximated numerically.

A full interpretation of our reanalysis would involve a critical assessment of which null hypotheses match the original research question, and would go beyond the scope of the current manuscript. We can, however, point out a couple of differences between the different analyses in Figures 4 and 5. Note, for example, how the zebra finches' significant difference between the on‐ and off‐1:1 ratio bins disappears when tested against a null hypothesis of uniformly distributed intervals (Figure 4). This indicates that while the zebra finches produce more rk ratios close to 1:1, this difference might would not be unexpected if the birds would uniformly randomly sample intervals. On the contrary, the on‐integer ratio bins around the 1:2 and 1:3 ratios in the Cuban salsa rk distribution do contain significantly more observed rhythm ratios than the corresponding off‐ratio bins, under the null hypotheses of uniformly or log‐normally randomly distributed interval duration (Figure 5). An a‐priori, appropriately chosen null hypothesis will not necessarily result in fewer significant results, but allows for a more fine‐tuned statistical analysis.

Finally, we want to stress that these examples are meant to demonstrate our methodological findings regarding the often implicitly assumed null hypothesis in previous studies. They illustrate how to adapt an existing approach to a new, explicitly chosen null hypothesis. The rest of our approach follows the previous study as closely as possible (i.e., Ref. 1). However, we have presented no arguments in favor or against following this exact approach. Other studies have approached the analyses from a slightly different angle and have chosen different statistical tests; their results are equally valid within their own statistical context. Additionally, we intentionally have not discussed the choice of on‐ and off‐integer bins: Our current findings have not provided any clear insights into the appropriate way to select boundaries for these bins. Altogether, the choice of null distribution is orthogonal to most of the other choices made when testing the presence of small‐integer ratios in an empirical dataset. Future research should investigate these choices for binning the rk values and potentially propose alternative ways of selecting a region of interest. Until then, our results provide a theoretical foundation and detailed interpretation of the typical methodology. Our results also highlight the importance of explicitly considering the choice of a null hypothesis, and demonstrate how this choice can be incorporated into any existing statistical approach.

CONCLUSION AND TAKEAWAY MESSAGE

The choice of a rhythm ratio formula has a strong effect on the analysis of temporal sequences: If we calculate the rhythm ratios of the same sequence with two different formulas, the resulting ratio distribution can be vastly different. The commonly used rhythm ratio rk formula, introduced by Roeske et al., 1 results in a uniform distribution of ratios when applied to exponentially distributed intervals. Such an exponential distribution of intervals occurs between events generated by a homogeneous Poisson point process, a mathematical description of a highly random probabilistic temporal process, which allows for events of arbitrarily short or long duration.

The mathematical relationship between a Poisson process, the rhythm ratio rk formula (Equation 1), and a uniform distribution of ratios means that a Poisson process often implicitly becomes the null hypothesis in statistical analyses. In several studies (e.g., Refs. 1, 8, 10, 13), an underlying assumption of the used statistical tests is a comparison to this uniform null distribution. A Poisson process may be a good default choice for a maximally random, arrhythmic baseline. We do, however, believe it is important to be aware of this implicit choice of null hypothesis.

For the case of biological organisms, a Poisson process is perhaps unrealistic: There is no limit on the duration of intervals, as interval durations can get very close to zero or very large. In other words, the null hypothesis from a Poisson process often sets the bar too low, and risks overinterpreting an animal's rhythmic capabilities. We, therefore, advice that future research makes the choice of null hypothesis explicit. Doing so (1) forces the researcher to consider whether this choice is appropriate, and (2) allows that researcher to adapt the statistical analysis to the underlying scientific question.

Consequently, our findings also strengthen the case for the use of the rk rhythm ratio as originally presented. 1 The rk ratio is not an arbitrary or convenient choice, but represents a fundamental mathematical link to the random generation of temporal sequences. Given the intrinsic random properties of a Poisson process, the rhythm ratio formula is an appropriate default quantification of the relation between two intervals. Moreover, we also show how a sample of observed rk rhythm ratios can be weighted to statically test it against a more specific null hypothesis. As such, the rk ratio is as good as any other ratio to statistically analyze empirical data, but comes with the added benefit of being simple to calculate, symmetric around 1:1, and bounded between 0 and 1.

In conclusion, our takeaway message to future research into rhythmic categories and integer ratios is twofold. On the one hand, we recommend using the rhythm ratio rk to analyze the relationship between intervals and believe it is a good default choice. On the other hand, it is essential to interpret any such results in the context of a Poisson process and carefully consider which other null hypothesis might be more appropriate to a particular research question.

AUTHOR CONTRIBUTIONS

Yannick Jadoul: Conceptualization; methodology; software; formal analysis; investigation; writing—original draft; writing—review and editing; visualization. Tommaso Tufarelli: Methodology; formal analysis; investigation; writing—original draft; writing—review and editing. Chloé Coissac: Software; investigation; writing—review and editing. Marco Gamba: Writing—review and editing. Andrea Ravignani: Conceptualization; methodology; investigation; writing —original draft; writing—review and editing; supervision; funding acquisition.

CONFLICT OF INTEREST STATEMENT

The authors have no competing interests to declare.

Supporting information

Supplementary material

NYAS-1553-363-s001.zip (231.5KB, zip)

ACKNOWLEDGMENTS

The authors thank Jelle van der Werff, Dunia Giomo, and Vesta Eleuteri for their valuable feedback on the manuscript, Bart de Boer for additional mathematical input, and Christopher Fallaize for the book suggestions. The authors also thank three anonymous reviewers for their constructive feedback and suggestions. YJ, CC, and AR are funded by the European Union (ERC, TOHR, 101041885) and are also supported by the HFSP research grant RGP0019/2022.

Open access publishing facilitated by Universita degli Studi di Roma La Sapienza, as part of the Wiley ‐ CRUI‐CARE agreement.

Jadoul, Y. , Tufarelli, T. , Coissac, C. , Gamba, M. , & Ravignani, A. (2025). Hidden assumptions of integer ratio analyses in bioacoustics and music. Ann NY Acad Sci., 1553, 363–378. 10.1111/nyas.70037

Footnotes

1

In some studies (e.g., Ref. 8), intervals are denoted as tk. To avoid potential confusion with points in time, we here stick to the ik notation.

2

The reason two resulting probability functions exist, f+ and f, is that they correspond to one monotonically increasing and one monotonically decreasing function. f+ is monotonically increasing and will map large values of q=i2i1 (i.e., when i2>i1) to large values of s, and a small q to a small s value. f do the inverse: small values of q will be mapped to large values of s, and as q increases, f(q)=s will decrease.

3

Also note that rk=ikik+ik+1 corresponds to the monotonically decreasing transformation function f(q)=11+q: For example, when qk=ik+1ik increases above 1, rk=ikik+ik+1 decreases below 0.5.

4

The interested reader can easily verify this by multiplying numerators and denominators of Equations (17) and (18) by 1a2, naming ba=c, and noticing that all occurrences of a and b can be replaced by c.

REFERENCES

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary material

NYAS-1553-363-s001.zip (231.5KB, zip)

Articles from Annals of the New York Academy of Sciences are provided here courtesy of Wiley

RESOURCES