Estimating noise ceilings for the Pearson correlation coefficient

In this post, I derive a commonly used split-half estimator for correlational noise ceilings. This involves a detour through classical test theory and the Spearman-Brown prediction formula.


Many scientific experiments have the following setup: choose some condition XXX \in \mathcal{X}, then produce a real-valued measurement YRY \in \mathbb{R} that is the average of multiple “reps” taken at condition XX. In my graduate school lab, the canonical example was choosing an image XX, then measuring the rep-averaged firing rate YY of a particular neuron (or multiunit) in a monkey’s visual cortex.

This procedure had randomness: if we repeated this procedure using the same image XX, we would almost surely record a different rep-averaged firing rate YY from that neuron. Because of this, there was a fundamental limit to how well we could predict that neuron’s firing rate based on the image that was shown. In other words, there was a noise ceiling.

Getting an estimate of the noise ceiling is important if one wants to build models Y^=f(X)\hat{Y} = f(X): it lets one distinguish when the model is genuinely failing as opposed to hitting irreducible randomness in the data.

Interestingly, this fairly universal situation can be addressed using an approach originally developed by the psychologists Charles Spearman and William Brown to deal with the rather specific problem that human responses to mental tests are noisy (Brown, 1910; Spearman, 1910). This approach is particularly popular in systems neuroscience, where one often sees Spearman and Brown name-dropped in methods sections – e.g. (Issa & DiCarlo, 2012; Ratan Murty et al., 2021; Lahner et al., 2024).

Conceptually, there are two parts to this blog post: presenting a definition of a noise ceiling, then presenting an estimator for it. The short version is this:

Motivating a definition of a noise ceiling

To motivate this post’s working definition of a noise ceiling, I’ll first assume a particular modeling goal: that a model Y^=f(X)\hat{Y} = f(X) is to be evaluated in terms of its population Pearson correlation coefficient (PCC) over some distribution (X,Y)P(X, Y) \sim P. This is a probabilistic frame, where experiments consist of draws from a joint distribution over conditions XX and measurements YY.

There are ways in which this modeling goal might have limitations for your use case. For example, the PCC is invariant to rescaling and offsets of the model’s predictions. If you care about identifying a model ff that spits out predictions Y^\hat Y in the same units as your measurements YY, rather than merely correlating positively, you might be better served using a goodness-of-fit metric like mean-squared-error.

But given that models are evaluated using the PCC, a natural noise ceiling definition is:

η:=supfCorr(f(X),Y) \boxed{ \eta_{\star} := \operatorname{sup}_f \operatorname{Corr}(f(X), Y) }

In words, η\eta_{\star} is the highest PCC that any model could hope to achieve with YY.

At first glance, this might seem like an intractable definition: it’s not as if we can loop over all possible models ff, calculate their corresponding population PCCs, then select the highest one; there are an infinite number of models. But it turns out we can simplify this formula, and write η\eta_{\star} purely in terms of moments of PP.

Deriving the noise ceiling of the Pearson correlation coefficient

The derivation below is a kind of tedious and mechanical calculation. But the core idea is intuitive: the highest possible Pearson correlation coefficient (PCC) is achieved by a model f(X)f_\star(X) that is a copy of the reproducible “signal” s(X):=E[YX]s(X) := \mathbb{E}[Y | X]. So if we can calculate the PCC achieved by this model, we have our noise ceiling η\eta_\star.

First, by definition of the population Pearson correlation coefficient Corr\operatorname{Corr}, η:=supfCorr(f(X),Y)=supf(Cov(f(X),Y)Var(f(X))Var(Y)) \begin{aligned} \eta_{\star} &:= \operatorname{sup}_f \operatorname{Corr}(f(X), Y) \\ &= \operatorname{sup}_f \left( \frac{\operatorname{Cov}(f(X), Y)}{\sqrt{\operatorname{Var}{(f(X))}} \sqrt{\operatorname{Var}(Y)}} \right) \end{aligned}

It is easy to show the term inside supf\operatorname{sup}_f is invariant to scaling and offsets of f(X)f(X), so WLOG we shall consider functions g(X)g(X) where E[g(X)]=0\mathbb{E}[g(X)] = 0 and Var(g(X))=1\operatorname{Var}(g(X))=1. Thus, η=supg:E[g(X)]=0,Var(g(X))=1(Cov(g(X),Y)Var(g(X))Var(Y))=supg:E[g(X)]=0,Var(g(X))=1(Cov(g(X),Y)Var(Y)) \begin{aligned} \eta_{\star} &= \operatorname{sup}_{g: \mathbb{E}[g(X)]=0, \operatorname{Var}(g(X))=1} \left( \frac{\operatorname{Cov}(g(X), Y)}{\sqrt{\operatorname{Var}{(g(X))}} \sqrt{\operatorname{Var}(Y)}} \right) \\ &= \operatorname{sup}_{g: \mathbb{E}[g(X)]=0, \operatorname{Var}(g(X))=1} \left( \frac{\operatorname{Cov}(g(X), Y)}{\sqrt{\operatorname{Var}(Y)}} \right) \end{aligned}

By an elementary property of covariance and the fact E[g(X)]=0\mathbb{E}[g(X)] =0, Cov(g(X),Y)=E[g(X)Y]E[g(X)]E[Y]=E[g(X)Y] \begin{aligned} \operatorname{Cov}(g(X), Y) &= \mathbb{E}[g(X)Y] - \mathbb{E}[g(X)]\mathbb{E}[Y] \\ &= \mathbb{E}[g(X)Y] \end{aligned}

By the law of total covariance,

Cov(g(X),Y)=E[Cov(g(X),YX)]+Cov(E[g(X)X],E[YX]) \begin{aligned} \operatorname{Cov}(g(X), Y) &= \mathbb{E}[\operatorname{Cov}(g(X), Y | X)] + \operatorname{Cov}(\mathbb{E}[g(X) | X], \mathbb{E}[Y | X]) \end{aligned}

As g(X)g(X) has no variance conditional on XX, this simplifies to Cov(g(X),Y)=Cov(g(X),E[YX]) \begin{aligned} \operatorname{Cov}(g(X), Y) &= \operatorname{Cov}(g(X), \mathbb{E}[Y | X]) \end{aligned}

To simplify notation, let s(X):=E[YX]s(X):=\mathbb{E}[Y|X]. Substituting,

η=supg:E[g(X)]=0,Var(g(X))=1(Cov(g(X),s(X))Var(Y)) \begin{aligned} \eta_{\star} &= \operatorname{sup}_{g: \mathbb{E}[g(X)]=0, \operatorname{Var}(g(X))=1} \left( \frac{\operatorname{Cov}(g(X), s(X))}{\sqrt{\operatorname{Var}(Y)}} \right) \end{aligned}

For any random variables AA and BB, we have the covariance inequality, by Cauchy-Schwarz:

Cov(A,B)Var(A)Var(B) |\operatorname{Cov}(A, B)| \leq \sqrt{\operatorname{Var}(A) \operatorname{Var}(B)}

Moreover, we know equality is achieved for the positive upper bound iff one variable is an affine function of the other with positive slope. Thus, a function of the form g(X):=αs(X)+βg_{\star}(X) := \alpha s(X) + \beta with α>0\alpha > 0 will achieve the upper bound of η\eta_\star. Recall we restricted ourselves to functions gg where g(X)g(X) has mean zero and unit variance, so we know the upper bound η\eta_\star is realized by

g(X)=s(X)E[s(X)]Var(s(X)) g_\star(X) = \frac{s(X) - \mathbb{E}[s(X)]}{\sqrt{\operatorname{Var}(s(X))}}

Substituting,

η=Cov(g(X),s(X))Var(Y)=Var(g(X))Var(s(X))Var(Y) \begin{aligned} \eta_{\star} &= \frac{\operatorname{Cov}(g_\star(X), s(X))}{\sqrt{\operatorname{Var}(Y)}} \\ &= \frac{\sqrt{\operatorname{Var}(g_\star(X))}\sqrt{\operatorname{Var}(s(X))}}{\sqrt{\operatorname{Var}(Y)}} \end{aligned}

We know Var(g(X))=1\operatorname{Var}(g_\star(X))=1, and we used the notation s(X):=E[YX]s(X):=\mathbb{E}[Y|X]. Simplifying,

η=Var(E[YX])Var(Y) \boxed{ \eta_{\star} = \sqrt{\frac{\operatorname{Var}(\mathbb{E}[Y|X])}{\operatorname{Var}(Y)}} }

In words: the noise ceiling for the Pearson correlation coefficient is the square root of the signal-to-total-variance ratio.

An estimator for the noise ceiling η\eta_{\star} using split-half reliability

In practice, we cannot ever know the value of the population parameter η\eta_{\star}, as we don’t have direct access to the true values of Var(E[YX])\operatorname{Var}(\mathbb{E}[Y|X]) or Var(Y)\operatorname{Var}(Y). In a real experiment, we often only have access to nn samples (X1,Y1),(X2,Y2),...,(Xn,Yn)(X_1, Y_1), (X_2, Y_2), ..., (X_n, Y_n) drawn from PP.

Here, I’ll go over a strategy which is available when each of your measurements YiY_i is itself an average of mm repeated measurements taken at XiX_i, or reps. Notationally, I’ll use RijR_{ij} to denote the jjth repeated measurement taken at XiX_i. So:

Yi=1mj=1mRij Y_i = \frac{1}{m}\sum_{j=1}^{m} R_{ij}

If your data look like this, then you can use the split-half method to estimate the noise ceiling η\eta_\star. The basic move is to split the reps at each condition into two disjoint halves, average the reps within each half, and then ask how correlated those two half-datasets are across conditions.

Let’s write down the notation. For a random split bb, let Ai(b)A_i^{(b)} and Bi(b)B_i^{(b)} denote the two sets of rep indices for condition XiX_i. These sets are disjoint, each has size m/2m/2, and together they contain the rep indices {1,...,m}\{1, ..., m\}. For that split, define the two half-averaged measurements:

Yi(b,1):=1Ai(b)jAi(b)RijYi(b,2):=1Bi(b)jBi(b)Rij \begin{aligned} Y_i^{(b, 1)} &:= \frac{1}{|A_i^{(b)}|}\sum_{j \in A_i^{(b)}} R_{ij} \\ Y_i^{(b, 2)} &:= \frac{1}{|B_i^{(b)}|}\sum_{j \in B_i^{(b)}} R_{ij} \end{aligned}

Now stack these half-averaged measurements across the nn sampled experimental conditions:

Y(b,1):=(Y1(b,1),Y2(b,1),...,Yn(b,1))Y(b,2):=(Y1(b,2),Y2(b,2),...,Yn(b,2)) \begin{aligned} \mathbf{Y}^{(b, 1)} &:= (Y_1^{(b, 1)}, Y_2^{(b, 1)}, ..., Y_n^{(b, 1)}) \\ \mathbf{Y}^{(b, 2)} &:= (Y_1^{(b, 2)}, Y_2^{(b, 2)}, ..., Y_n^{(b, 2)}) \end{aligned}

The split-half correlation for split bb is then just the sample Pearson correlation between these two length-nn vectors:

r^half(b):=r^(Y(b,1),Y(b,2)) \hat{r}^{(b)}_\text{half} := \hat{r}(\mathbf{Y}^{(b, 1)}, \mathbf{Y}^{(b, 2)})

Now, before going any further, I’ll just present the estimator for the noise ceiling η\eta_\star without explanation:

η^=1Bb=1B2r^half(b)1+r^half(b) \boxed{ \hat \eta_\star = \frac{1}{B}\sum_{b=1}^{B} \sqrt{\frac{2 \hat{r}^{(b)}_\text{half}}{1 + \hat{r}^{(b)}_\text{half}}} }

Where BB is the number of iterations, with larger BB being better. In the last section of this post, I’ll describe how this formula was derived.

Where the estimator comes from

To understand where the estimator η^\hat \eta_\star comes from, we must take a brief detour into the land of classical test theory, before returning back to the more general setting.

I’ll organize the derivation of the estimator into four steps:

1. Reliability

In classical test theory, the objects of study are tests. A test consists of a set of mm items, each of which yields a real-valued score when administered to a person. These mm individual item scores are averaged together to result in an overall test score which reflects something about an individual’s underlying psychological construct, like his or her ability to perform multiplication.

Suppose you administered a test to a population of people, and you observed variation in their test scores. A key methodological concern that classical test theory addresses is: how much of the raw test score variation is due to true variation in psychological constructs, as opposed to random noise inherent to the test taking process?

To measure this, classical test theory introduced a real-valued quantity called reliability, denoted by the symbol ρ\rho:

ρ:=Var(E[Test ScorePerson])Var(Test Score) \rho := \frac{\operatorname{Var}(\mathbb{E}[\text{Test Score} | \text{Person}])}{\operatorname{Var}(\text{Test Score})}

This should look suspiciously similar to the definition of the noise ceiling for the Pearson correlation coefficient (PCC), η\eta_\star, from before. Indeed, if we translate things into the notation we’ve been using:

Then we see that reliability is equal to the square of the noise ceiling:

ρ=Var(E[YX])Var(Y)=η2 \rho = \frac{\operatorname{Var}(\mathbb{E}[Y | X])}{\operatorname{Var}(Y)} = \eta_\star^2

Equivalently, the noise ceiling for the Pearson correlation coefficient is the square root of the reliability. η=ρ \boxed{ \eta_\star = \sqrt{\rho} }

It’s a little mind-bending to understand what a noise ceiling means here, if you’re more used to thinking about noise ceilings for neurons (as I am). In the classical test taking setting, the noise ceiling η\eta_\star indicates the highest PCC a model that maps people XX to test scores YY could achieve.

It’s also worth pausing to note that the terminology can be confusing: the symbol ρ\rho is often used to denote both reliability and population Pearson correlation coefficients. In this post, I use Corr\operatorname{Corr} for the population PCC, and ρ\rho for reliability.

2. Deriving the Spearman-Brown prediction formula

Now we’re equipped with the needed vocabulary to present the Spearman-Brown prediction formula. This formula predicts the reliability ρ\rho of a test after increasing the number of test items by a factor of kk. The key assumption in this formula is that the test items are parallel: conditioned on a test taker, the item scores each have the same expected value, and the item score residuals are independent with the same variance.

The Spearman-Brown prediction formula is:

ρk=kρ1+(k1)ρ \rho_k = \frac{k \rho}{1 + (k-1) \rho}

To derive the formula, one starts by calculating the reliability of the original, base-length test YY. Recall that XX denotes the person taking the test.

ρ=Var(E[YX])Var(Y) \rho = \frac{\operatorname{Var}(\mathbb{E}[Y|X])}{\operatorname{Var}(Y)}

Next, recall that a test score YY is the average of mm items:

Y:=1mj=1mRj Y := \frac{1}{m} \sum_{j=1}^{m} R_j

A key assumption is that item scores RjR_j have the same expected value conditional on the test taker XX. That is, there exists some function ss where s(X)=E[RjX],js(X) = \mathbb{E}[R_j | X], \forall j. We will also let Ej:=Rjs(X)E_j := R_j - s(X). Substituting,

Y=1mj=1m(s(X)+Ej)=s(X)+1mj=1mEj \begin{aligned} Y &= \frac{1}{m} \sum_{j=1}^{m}(s(X) + E_j) \\ &= s(X) + \frac{1}{m} \sum_{j=1}^{m} E_j \end{aligned}

Calculating the variance of YY,

Var(Y)=Var(s(X)+1mj=1mEj) \begin{aligned} \operatorname{Var}(Y) &= \operatorname{Var}\left( s(X) + \frac{1}{m} \sum_{j=1}^{m} E_j\right) \\ \end{aligned}

By assumption, all residuals EjE_j are independent of s(X)s(X). Thus,

Var(Y)=Var(s(X))+Var(1mj=1mEj) \begin{aligned} \operatorname{Var}(Y) &= \operatorname{Var}(s(X)) +\operatorname{Var}\left( \frac{1}{m} \sum_{j=1}^{m} E_j\right) \\ \end{aligned}

By assumption, the residuals EjE_j are independent across items and have common variance Var(E)\operatorname{Var}(E):

Var(Y)=Var(s(X))+1mVar(E) \begin{aligned} \operatorname{Var}(Y) &= \operatorname{Var}(s(X)) + \frac{1}{m} \operatorname{Var}(E) \\ \end{aligned}

Substituting into the reliability formula,

ρ=Var(E[YX])Var(Y)=Var(s(X))Var(s(X))+1mVar(E) \begin{aligned} \rho &= \frac{\operatorname{Var}(\mathbb{E}[Y|X])}{\operatorname{Var}(Y)} \\ &= \frac{\operatorname{Var}(s(X))}{\operatorname{Var}(s(X)) + \frac{1}{m} \operatorname{Var}(E)} \end{aligned}

Suppose you increased the number of items by a factor of kk. Denote this reliability by ρk\rho_k:

ρk=Var(s(X))Var(s(X))+1kmVar(E) \begin{aligned} \rho_k &= \frac{\operatorname{Var}(s(X))}{\operatorname{Var}(s(X)) + \frac{1}{km} \operatorname{Var}(E)} \end{aligned}

One can rewrite ρk\rho_k in terms of ρ\rho, using simple algebraic manipulations. First, substitute Var(s(X))=ρ(Var(s(X))+1mVar(E))\operatorname{Var}(s(X)) = \rho \left(\operatorname{Var}(s(X)) + \frac{1}{m} \operatorname{Var}(E)\right)

ρk=ρ(Var(s(X))+1mVar(E))ρ(Var(s(X))+1mVar(E))+1kmVar(E)=kρ(Var(s(X))+1mVar(E))kρ(Var(s(X))+1mVar(E))+1mVar(E)=kρ(Var(s(X))+1mVar(E))kρ(Var(s(X))+1mVar(E))+1mVar(E)+Var(s(X))Var(s(X))=kρkρ+1Var(s(X))Var(s(X))+1mVar(E)=kρkρ+1ρ=kρ1+(k1)ρ  \begin{aligned} \rho_k &= \frac{\rho \left(\operatorname{Var}(s(X)) + \frac{1}{m}\operatorname{Var}(E)\right)} {\rho \left(\operatorname{Var}(s(X)) + \frac{1}{m}\operatorname{Var}(E)\right) + \frac{1}{km}\operatorname{Var}(E)} \\ &= \frac{k\rho \left(\operatorname{Var}(s(X)) + \frac{1}{m}\operatorname{Var}(E)\right)} {k\rho \left(\operatorname{Var}(s(X)) + \frac{1}{m}\operatorname{Var}(E)\right) + \frac{1}{m}\operatorname{Var}(E)} \\ &= \frac{k\rho \left(\operatorname{Var}(s(X)) + \frac{1}{m}\operatorname{Var}(E)\right)} {k\rho \left(\operatorname{Var}(s(X)) + \frac{1}{m}\operatorname{Var}(E)\right) + \frac{1}{m}\operatorname{Var}(E) + \operatorname{Var}(s(X)) - \operatorname{Var}(s(X))} \\ &= \frac{k\rho} {k\rho + 1 - \frac{\operatorname{Var}(s(X))}{\operatorname{Var}(s(X)) + \frac{1}{m}\operatorname{Var}(E)}} \\ &= \frac{k\rho}{k\rho + 1 - \rho} \\ &= \frac{k\rho}{1 + (k-1)\rho} ~\square \\ \end{aligned}

3. The relationship of split-half correlations to the noise ceiling η\eta_\star

One can show that the population Pearson correlation coefficient between two split halves Y(1)Y^{(1)} and Y(2)Y^{(2)} of an experiment with mm reps is equal to the reliability ρhalf\rho_{\text{half}} of a hypothetical experiment with m/2m/2 reps:

Corr(Y(1),Y(2))=Cov(Y(1),Y(2))Var(Y(1))Var(Y(2)) \begin{aligned} \operatorname{Corr}(Y^{(1)}, Y^{(2)}) = \frac{\operatorname{Cov}(Y^{(1)}, Y^{(2)})}{\sqrt{\operatorname{Var}(Y^{(1)})\operatorname{Var}(Y^{(2)})}} \end{aligned}

As Y(1)Y^{(1)} and Y(2)Y^{(2)} are conditionally i.i.d. on XX, let Var(Y(h)):=Var(Y(1))=Var(Y(2))\operatorname{Var}(Y^{(h)}) := \operatorname{Var}(Y^{(1)})=\operatorname{Var}(Y^{(2)}). Moreover, they have the same means conditional on XX, so through the law of total covariance this implies Cov(Y(1),Y(2))=Var(E[Y(h)X])\operatorname{Cov}(Y^{(1)}, Y^{(2)}) = \operatorname{Var}(\mathbb{E}[Y^{(h)}|X]). So,

Corr(Y(1),Y(2))=Var(E[Y(h)X])Var(Y(h)) \begin{aligned} \operatorname{Corr}(Y^{(1)}, Y^{(2)}) = \frac{\operatorname{Var}(\mathbb{E}[Y^{(h)}|X])}{\operatorname{Var}(Y^{(h)})} \end{aligned}

By definition, this is equal to the reliability of an experiment with m/2m/2 reps. So,

ρhalf=Corr(Y(1),Y(2)) \rho_\text{half} = \operatorname{Corr}(Y^{(1)}, Y^{(2)})

Now, to obtain the reliability ρfull\rho_\text{full} of the “full” experiment with mm reps from the reliability ρhalf\rho_\text{half} of the experiment with m/2m/2 reps, we simply apply the k=2k=2 case of the Spearman-Brown formula,

ρfull=2ρhalf1+ρhalf=2Corr(Y(1),Y(2))1+Corr(Y(1),Y(2)) \rho_\text{full} = \frac{2\rho_\text{half}}{1 + \rho_\text{half}} = \frac{2 \operatorname{Corr}(Y^{(1)}, Y^{(2)})}{1+\operatorname{Corr}(Y^{(1)}, Y^{(2)})}

Finally, we know from above that η=ρfull\eta_\star = \sqrt{\rho_\text{full}}. Thus,

η=2Corr(Y(1),Y(2))1+Corr(Y(1),Y(2)) \boxed{ \eta_\star = \sqrt{\frac{2 \operatorname{Corr}(Y^{(1)}, Y^{(2)})}{1+\operatorname{Corr}(Y^{(1)}, Y^{(2)})}} }

4. Assembling the estimator

We are now equipped to assemble an estimator η^\hat \eta_\star. The guiding idea is the plug-in principle, where we will “plug in” sample-based estimates for the population parameters in the formula for η\eta_\star.

Namely, we use the sample Pearson correlation coefficient r^\hat r as an estimator for Corr\operatorname{Corr}. Given random split halves Y(b,1)\mathbf{Y}^{(b, 1)} and Y(b,2)\mathbf{Y}^{(b, 2)}, we have the estimator:

r^half(b)=r^(Y(b,1),Y(b,2)) \hat{r}^{(b)}_\text{half} = \hat{r}(\mathbf{Y}^{(b, 1)}, \mathbf{Y}^{(b, 2)})

We then simply plug this in to the population formula for η\eta_\star to obtain the estimator:

η^(b)=2r^half(b)1+r^half(b). \begin{aligned} \hat{\eta}^{(b)}_\star &= \sqrt{ \frac{2 \hat{r}^{(b)}_\text{half}}{1 + \hat{r}^{(b)}_\text{half}} }. \end{aligned}

In practice, this procedure is repeated over BB random splits, and the resulting estimates are averaged. This averaging reduces the variation in our final estimate, as the choice of split bb is random. This leads to the final estimator for the noise ceiling:

η^=1Bb=1B2r^half(b)1+r^half(b) \boxed{ \hat \eta_\star = \frac{1}{B}\sum_{b=1}^{B} \sqrt{\frac{2 \hat{r}^{(b)}_\text{half}}{1 + \hat{r}^{(b)}_\text{half}}} }

In general, the estimator is biased because sample correlations, the Spearman-Brown correction, and the square root are all finite-sample and nonlinear operations. Moreover, it can behave badly if split-half correlations are near zero or are negative.

The estimator is more defensible as the number of sampled conditions and number of sampled reps grow. To give an arbitrary reference point, folks in the lab would regularly use this formula to estimate noise ceilings for visual neuron spike rates when they had datasets consisting of ~100s of images, with ~50 reps each.

In the derivation above, we assumed every condition had the same number of reps mm, and that each split half contained m/2m/2 reps. In the more realistic scenario where rep counts vary across conditions, the above derivation no longer applies exactly. People often still use this estimator heuristically, especially when the rep counts are similar or all reasonably large.

Closing thoughts

I was first taught about the estimator η^\hat \eta_\star as a first year graduate student from my PhD advisor and senior lab colleagues, who passed it down almost like a piece of lab folklore. It was pretty mysterious to me how a method developed to analyze psychological survey data would end up becoming a workhorse in analyzing spike rates from visually driven neurons. It felt like I was learning some sort of esoteric statistical spell, not taught in the classroom.

Indeed, I had never even encountered the notion of a noise ceiling in any of the introductory statistics courses I had taken. When it came to the study of Pearson correlation coefficients, the focus was usually on conducting null hypothesis testing against H0:Corr(f(X),Y)=0H_0: \operatorname{Corr}(f(X), Y) = 0. The focus on noise ceilings provided an interesting complement to this frame, in that it encourages one to start thinking about H0:Corr(f(X),Y)=ηH_0: \operatorname{Corr}(f(X), Y) = \eta_\star.

As a final note, it’s important to note that the split-half estimator in this post is not the only possible estimator of η\eta_\star. Since η\eta_\star is defined in terms of Var(E[YX])\operatorname{Var}(\mathbb{E}[Y|X]) and Var(Y)\operatorname{Var}(Y), one could instead try to estimate these variance components directly, instead of going through the ceremony of generating split-halves. I actually suspect that’s a better method, in some sense. Maybe that’s a topic for a future blog post. This one ended up being much longer than I thought!


  1. Brown, W. (1910). Some experimental results in the correlation of mental abilities 1. British Journal of Psychology, 1904-1920, 3(3), 296–322.
  2. Spearman, C. (1910). Correlation calculated from faulty data. British Journal of Psychology, 3(3), 271.
  3. Issa, E. B., & DiCarlo, J. J. (2012). Precedence of the eye region in neural processing of faces. Journal of Neuroscience, 32(47), 16666–16682.
  4. Ratan Murty, N. A., Bashivan, P., Abate, A., DiCarlo, J. J., & Kanwisher, N. (2021). Computational models of category-selective brain regions enable high-throughput tests of selectivity. Nature Communications, 12(1), 5540.
  5. Lahner, B., Dwivedi, K., Iamshchinina, P., Graumann, M., Lascelles, A., Roig, G., Gifford, A. T., Pan, B., Jin, S. Y., Ratan Murty, N. A., & others. (2024). Modeling short visual events through the BOLD moments video fMRI dataset and metadata. Nature Communications, 15(1), 6241.