I discuss proposals for a method that estimates how much predictive information additional degrees of precision in forecasts add and at which point additional precision is just noise, and investigate these proposals with empirical forecasting data. I furthermore describe desirable criteria such functions ought to fulfill. I conclude that methods based on rounding probabilities are hot flaming garbage, methods based on rounding odds or log-odds are regular garbage, and methods based on applying noise to log-odds are ok but sort of slow. Finding the maximal informative precision turns out to be tricky.

Epistemic Status

Maybe not just reinventing the wheel, but the whole bicycle.


Say we have a set of resolved forecasts and can display them on a calibration plot.

We can grade the forecasts according to some proper scoring rule, e.g. the Brier score or the logarithmic scoring rule, maybe even broken up by calibration and resolution.

But we can also ask the question: how fine-grained are the predictions of our forecaster? I.e., at which level of precision can we assume that the additional information is just noise?

Overprecise Omar

Take, for example, a hypothetical forecaster Omar who always gives their forecasts with 5 decimal digits of precision, such as forecasting a "24.566% probability of North Korea testing an ICBM in the year 2022", even though if we look at their calibration plot (of sufficiently many forecasts), we see that they are pretty much random in any given interval of length 0.1 (i.e., their forecast with 15% and a forecast of 5% can be expected to resolve to the same outcome with equal probability). This means that 4 of the 5 decimal digits of precision are likely just noise!

Omar would be behaving absurdly; misleading their audience into believing they had spent much more time on their forecasts than they actually had (or, more likely, into correctly leading the audience into believing that there was something epistemically sketchy going on). It is certainly useful to use probability resilience, and not imprecision, to communicate uncertainty, but there is an adequate & finite limit to precision.

I believe something similar is going on when people encounter others putting probabilities on claims: It appears like an attempt at claiming undue quantitativeness (quantitativity?) in their reasoning, and at making the listener fall prey to precision bias, as well as an implicit claim at scientific rigour. However, not all precision in judgmental forecasting is false precision: At some point, if remove digits of precision, the forecasts will become worse in expectation.

But how might we confront our forecaster Omar from above? How might we estimate the level of degrees of precision after which their forecasts gave no more additional information?

Definitions

Ideally we'd want to find a number that tells us, for a given set of (resolved) forecasts, the precision that those predictions display: Any additional digits added to the probability beyond this precision would just be noise.

Let us call this number the precision of a set of forecasts.

Let be a dataset of forecasts and resolutions .

Then is simply a function that takes in such a dataset of forecasts and produces a real number , so for example for the forecasts and outcomes of some forecaster, or team of forecasters.

We call the precision function.

Bits, Not Probabilities, for Precision

It is natural to assume that returns a probability: after all, the input dataset has probabilities, and when talking about Omar's calibration plot I was explicitely calling out the loss of accuracy in probability intervals shorter than 0.1.

Furthermore, Friedman et al. 2018 also talk about precision in terms of probabilities (or rather "bins of probabilities"), we are all used to probabilities, probabilities are friends.

But this doesn't stand up to scrutiny: If we accept this, assuming we use probability "bins" or "buckets" of size 5%, then 99.99999% and 96% are as similar to each other as 51% and 54.99999%. But the readers familiar with the formulation of probability in log-odds form will surely balk at this equivalence: 99.99999% is a beast of a probability, an invocation only uttered in situations of extreme certainty, while 96%, 51% and 54.99999% (modulo false precision) are everyday probabilities, plebeian even.

And, in terms of precision, 54.99999% stands out like a sore thumb: while 99.99999% is supremely confident, it is not overprecise, since rounding up to 100% would be foolish; but with 54.99999%, there is no good reason we can't just round to 55%.

So precision should not be investigated on probabilities. Instead I claim that it should be calculated in log-odds space (which has an advantage over the odds form by being symmetric), where one moves in bits instead of probabilities. Since we want to make a statement how much we can move the probabilities around until the proper scoring rule we apply starts giving worse results, it is only natural to express the precision in bits as well. (Which can't be converted linearly into probabilities: moving from 50% to 75% or from 80% to 40% is one bit, but similarly moving from ~99.954% to ~99.977% is also a change of one bit). Instead one uses the logit function, and the logistic function to return to probabilities.

Algorithms!

The assumption of expressing precision in bits naturally leads to two different algorithms.

These algorithms follow a common pattern:

  1. Input: a dataset of forecasts and resolutions.
  2. Set a perturbation parameter to a minimum value.
  3. Repeat:
    1. Potentially do something with the probabilities, for example by moving to log-odds.
    2. Perturb the result, using as a parameter for the perturbation function so that .
    3. Reverse the things done in step 1 (if necessary).
    4. Calculate the score of .
    5. If the perturbed score is significantly worse than the unperturbed score, then is the correct level of precision, stop. This is the output of .
    6. Else increase .

Log-Odds Rounding

A technique explored by @tenthkrige explores rounding in odds-form: The probabilities are converted into odds-form and there rounded to the nearest multiple of the perturbation parameter .

Tenthkrige value

Here, the precision of the community predictions looks like it is somewhere between 0.5 and 1, and the precision of the Metaculus prediction is around 0.25 and 0.5, but it's hard to tell without access to the measurements.

Log-odds rounding is pretty similar to odds-rounding.

Implementation

Once a probability has been converted into log-odds form , then rounding with a perturbation is simply .

Scoring the forecasts using the logarithmic scoring rule, one can then write this in a couple of lines of python code:

def logodds_rounded_score(forecasts, perturbation=1):
    p=forecasts[1]
    lo=logit(p)
    rounded_lo=perturbation*np.round(lo/perturbation)
    rounded_probs=logistic(rounded_lo)
    return np.mean(logscore(forecasts[0], rounded_probs))

Full code for all algorithms is here.

Testing with Toy Data

One can then check whether everything is working nicely by applying the method to a couple of small (n=4,4,4,100,100,2000) toy datasets.

The smallest three datasets, all of n=4, are d1 (red in the chart), d2 (blue in the chart) and d3 (green in the chart).

They are

d1=np.array([[1,0.8],[0,0.4],[0,0.65],[1,0.99]]).T
d2=np.array([[1,0.8],[0,0.4],[0,0.65],[1,0.9]]).T
d3=np.array([[0,0.8],[1,0.4],[1,0.65],[0,0.9]]).T

One notices that d2 is just a slightly less resolute & precise d1, and d3 is pretty bad.

Rounding log-odds of forecasts

Problems

Now this is not… great, and certainly quite different from the data by tenthkrige. I'm pretty sure this isn't a bug in my implementation or due to the switch from odds to log-odds, but a deeper problem with the method of rounding for perturbation.

There are two forces that make these charts so weird:

  1. If you decrease precision by rounding, you can actually make a probability better by moving it closer to 0%/100%. If you have one forecast with probability and the outcome . Without rounding, this has a log-score of ~-0.29. Setting and rounding in log-odds space and transforming back gives , with an improved log-score of ~-0.26. When one sees the weird zig-zag pattern for log-score, I believe that this is the reason. This is likely only an issue in smaller datasets: In larger ones, these individual random improving/worsening effects cancel each other out, and one can see the underlying trend of worsening (as is already visible with the purple plot for n=2000, and to a lesser extent brown & orange). Still, I count this as a major problem with rounding-based methods. Friedman et al. 2018 note the same: "Coarsening probability estimates could actually improve predictive accuracy if this prevents foreign policy analysts from making their judgments too extreme".

  2. Rounding very strongly rounds everything to 50%, so with strong enough rounding every dataset has the same score. This has some counter-intuitive implications: If you are worse than chance, perturbing your probabilities more and more leads you reliably to a better score (in the case of the log score, ). One can also see this in the plot above: All but one datasets end up with approximately the same log score. This isn't an property that kills rounding, since we in theory only care about the point where the perturbed score starts diverging from the unperturbed score, but it is still undesirable.

Advantages

That said, the method has one big advantage: It is quite fast, running ~instantaneously for even big datasets on my laptop.

Noise in Log-Odds Space

But one can also examine a different paradigm: Applying noise to the predictions. In our framework, this is concretely operationalized by repeatedly ( times) projecting the probabilities into log-odds space, applying some noise with width to them, and then calculating the resulting probabilities and scoring them, finally taking the mean of the scores.

There are some free parameters here, especially around the exact type of noise to use (normal? beta? uniform?), as well as the number of samples.

I have decided to use uniform noise, but for no special mathematical reason, and adjust the number of samples by the size of the dataset (with small datasets , less with bigger datasets).

def noised_score(forecasts, perturbation=1, samples=100):
    o=forecasts[0]
    p=forecasts[1]
    pert_scores=[]
    for i in range(0,samples):
    noised=logistic(logit(p)+np.random.default_rng(seed=time.monotonic_ns()).uniform(-perturbation/2,perturbation/2,len(p)))
         pert_scores.append(logscore(o,noised))
    return np.mean(pert_scores)

This gives a much nicer looking plot, with samples:

Applying noise to forecasts

The plots are falling ~monotonously, with a worse score for increasing noise, as expected. The score for d2 drops more quickly than the one for d1, maybe because d2 is less precise than d1? I'm not sure.

Advantages and Problems

Noising log-odds has a bunch of advantages: As we increase the perturbation, the score falls ~monotonically (which I conjecture to always be true in the limit of infinitely many samples), and doesn't converge to a specific value as rounding-based methods do.

This can be seen when comparing increasing perturbation with the dataset:

Precision comparison of some synthetic forecasting datasets

The disadvantage lies in the runtime: Taking many samples makes the method slow, but a small number of samples is too noisy to reliably detect when the the quality of forecasts starts dropping off. I think this is less of a problem with bigger datasets, but in the worst case I'd have to do a bunch of numpy-magic to optimize this further (or rewrite the code in a faster programming language, prototype in C here).

Finding Divergence

If the unperturbed dataset has a score , then we also need to find a value for with so that , with (in the case of scoring rules where lower scores are worse) or even that for all positive and it holds that . That is, as we increase the perturbation more and more, the score becomes worse and worse.

The easiest way to do this is to iterate through values for , and once the difference between and is too big, return the corresponding . This method has disadvantages (the least-noteworthy difference between and is probably an arbitrary constant, simply iterating requires a bunch of compute and only finds an upper bound on a which is "too much"), but shines by virtue of being bog-simple and easy to implement. A more sophisticated method can use binary search to zero in on a that just crosses the threshold.

def linsearch_precision(forecasts, samples=100, low=10e-4, high=10, change=1.05, mindiff=0.01, mode='round'):
    clean_score=pert_score=np.mean(logscore(forecasts[0], forecasts[1]))
    p=low
    while np.abs(clean_score-pert_score)<mindiff and p<high:
        if mode=='round':
             pert_score=logodds_rounded_score(forecasts, perturbation=p)
        elif mode=='noise':
             pert_score=noised_score(forecasts, perturbation=p, samples=samples)
        p=p*change
    return p

Unfortunately, this method doesn't really give reliable results for small sample sizes:

>>> precisions_1=[linsearch_precision(d1, change=1.05, mode='noise', samples=1000) for i in range(0, 5)]
>>> precisions_1
[1.2406168179333095, 1.302647658829975, 1.1815398266031518, 1.302647658829975, 1.302647658829975]
>>> np.mean(precisions_1)
1.2660199242052772
>>> np.var(precisions_1)
0.002361395506068271
>>> precisions_6=[linsearch_precision(d6, change=1.05, mode='noise', samples=200) for i in range(0, 5)]
>>> precisions_6
[1.302647658829975, 1.302647658829975, 1.302647658829975, 1.302647658829975, 1.302647658829975]
>>> np.mean(precisions_6)
1.302647658829975
>>> np.var(precisions_6)
0.0

It seems like smaller datasets need higher sample sizes to adequately assess the precision using noise-based methods.

But this does tell us that bits, and bits.

The code can be changed to be faster, using binary search:

def binsearch_precision(forecasts, samples=100, low=10e-4, high=10, mindiff=0.01, mode='round', minstep=10e-4):
    clean_score=np.mean(logscore(forecasts[0], forecasts[1]))
    while high-low>minstep:
        mid=(high+low)/2
        if mode=='round':
            pert_score=logodds_rounded_score(forecasts, perturbation=mid)
        elif mode=='noise':
            pert_score=noised_score(forecasts, perturbation=mid, samples=samples)
        if np.abs(clean_score-pert_score)<mindiff:
            low=mid
        else:
            high=mid
    return mid

We can take advantage of the fact that we're not dealing the the indices of arrays here, so one can just divide by two as desired.

>>> bin_precisions_1=[binsearch_precision(d1, mode='noise', samples=1000) for i in range(0, 5)]
>>> bin_precisions_1
[1.2002208862304686, 1.2612499389648437, 1.2685734252929686, 1.3686610717773438, 1.2136472778320313]
>>> np.var(bin_precisions_1)
0.0035147788876175893
>>> bin_precisions_6=[binsearch_precision(d6, mode='noise', samples=200) for i in range(0, 5)]
>>> bin_precisions_6
[1.1916768188476563, 1.2160884399414063, 1.2173090209960937, 1.2124266967773438, 1.207544372558594]
>>> np.mean(bin_precisions_6)
1.2090090698242189
>>> np.var(bin_precisions_6)
8.664782133936837e-05

Now we can check whether the two methods give the ~same results:

>>> np.mean(precisions_1)
1.2660199242052772
>>> np.mean(bin_precisions_1)
1.2624705200195312
>>> np.mean(precisions_6)
1.302647658829975
>>> np.mean(bin_precisions_6)
1.2090090698242189

I suspect that this is simply a problem of small sample sizes, though: Increasing the sample sizes by 5× doesn't change problem at all:

>>> bin_precisions_6=[binsearch_precision(d6, mode='noise', samples=500) for i in
range(0, 10)]
>>> np.mean(bin_precisions_6)
1.2027841064453124

I'm kind of puzzled at this result, and not sure what to make of it. Maybe binary search biases the results downwards, while exponential search biases them upwards? We can check by changing the stepsize to linear:

>>> precisions_6=[linsearch_precision(d6, change=0.01, mode='noise', stepmode='lin', samples=200) for i in range(0, 5)]
>>> precisions_6
[1.2010000000000007, 1.2110000000000007, 1.2110000000000007, 1.2010000000000007, 1.2210000000000008]
>>> np.mean(precisions_6)
1.209000000000001
>>> np.var(precisions_6)
5.60000000000001e-05

This, prima facie, resolves our dilemma, and indicates that linear steps are better than exponential steps, and if speed is a problem, then binary search is better.

More Ideas

Rewriting the code to use noisy binary search, since the comparisons of scores are not reliable, might be a cool project.

But perhaps the entire idea of using a fixed value for divergence is flawed? The divergence point needs to depend on the perturbation parameter. I think I should instead assume there is a perfectly calibrated, infinitely big dataset . How much does its score drop if we perturb by ? If , perturbed, drops by half (or whatever) as much as the score of then we have found our point.

A more sophisticated technique could try to estimate the elbow point of the declining curve of scores, but as far as I know there is no reliable method for doing so, nor is there a mathematical framework for this.

Precision of Forecasting Datasets

Equipped with some shaky heuristics about forecasting precision, one can now try to estimate the precision of different forecasting datasets (especially interesting should be comparing different platforms).

I will be doing so using the library iqisa.

import numpy as np
import pandas as pd
import iqisa as iqs
import iqisa.gjp as gjp

gjpmarket_fcasts=gjp.load_markets()
gjpsurvey_fcasts=gjp.load_surveys()
gjpmarket=gjpmarket_fcasts[['probability', 'answer_option', 'outcome']].dropna()
gjpsurvey=gjpsurvey_fcasts[['probability', 'answer_option', 'outcome']].dropna()
gjpmarket=np.array([np.int64(gjpmarket['answer_option']==gjpmarket['outcome']), gjpmarket['probability']])
gjpsurvey=np.array([np.int64(gjpsurvey['answer_option']==gjpsurvey['outcome']), gjpsurvey['probability']])

Now we can use the algorithms for estimating precision on the dataset:

import algorithms as prc

>>> prc.binsearch_precision(gjpsurvey, mode='noise', samples=10)
1.5017044067382812
>>> prc.binsearch_precision(gjpmarket, mode='noise', samples=10)
1.279558654785156

Apparently the markets were more precise than the survey forecasts. Interesting. Now let's see Paul Allen's Metaculus' precision.

import iqisa.metaculus as metaculus
metaculus_fcasts=metaculus.load_public_binary()
met_fcasts=met_fcasts_fcasts[['probability', 'answer_option', 'outcome']].dropna()
met_fcasts=np.array([np.int64(met_fcasts['answer_option']==met_fcasts['outcome']), met_fcasts['probability']])

>>> prc.binsearch_precision(met_fcasts, mode='noise', samples=10)
1.1941179809570313

Harsh words from Metaculus here… the precision is higher here, to be clear: The smallest perturbation for which the score becomes noticeably worse is smaller.

Same with PredictionBook:

import iqisa.predictionbook as predictionbook
pred_fcasts=predictionbook.load()
pred_fcasts=pred_fcasts[['probability', 'answer_option', 'outcome']].dropna()
pred_fcasts=np.array([np.int64(pred_fcasts['answer_option']==pred_fcasts['outcome']), pred_fcasts['probability']])

>>> prc.binsearch_precision(pred_fcasts, mode='noise', samples=10)
1.3198378295898436

The difference is quite small but noticeable, and the purple and the green plot overlap a bunch: Metaculus and PredictionBook aren't so different (?):

Precision comparison of different forecasting platforms

The precision tracks the unperturbed score to a high degree. I'm not sure whether this is an airtight relation or just commonly the case, perhaps someone will look into this.

Usage

Given the precision of some forecaster or forecasting platform, one can much easier perform sensitivity analysis (especially after correcting for miscalibration), as it allows to create a distribution over what a probability could reasonably mean, given some resolved past data from the same source.

Knowing the precision of some source of forecasts can also be nice to stave off criticism of false precision: One can now answer "No, you're wrong: I'm being wholly correct at reporting my forecasts at 1.2 bits of precision!". Surely no incredulous stares with that.

Conclusion

When evaluating the precision of a set of forecasts, all sources I know agree that perturbing the forecasts in some way and then observing how the forecasts worsen with more perturbation is the correct way to go about evaluating the precision of forecasts.

However, the existing attempts are based on rounding: Moving probabilities, odds or log-odds that are close together to the same number.

I believe that these approaches are not good: For small datasets they produce large oscillations in the score, not smooth declines, and they improve the scores of worse-than-random forecast datasets.

Instead, a better way to estimate the precision is to apply noise to the forecasts and track how the score worsens. This has the advantage of producing monotonically declining scores.

Testing the method of noising forecasts on real-world datasets shows that they have similar precisions.

Appendix A: Conditions for a Precision Evaluation Function

Using a precision function , perturbation function , proper scoring rule and noise . These are very much me thinking out loud.

  1. (Emptiness) If , is undefined.
  2. (Loneliness) If , : We are generally suspicious of any single forecast.
    1. More generally, if contains an so that there is no other prediction with a probability within , then . Yes, even if the set of forecasts is "dense" and non-random in other places.
  3. (Edginess) If , then it should hold for an (but close to 0): .
    1. More generally, if we have only zeros in the left half and ones in the right half, with , and a sufficiently small , it should hold that is for and for .
  4. (Perfection) For a sufficiently large , and a sufficiently small , and being 1 with probability and 0 with probability , it should hold that : If we have lots of datapoints, all perfectly calibrated, the score is nearly 0 at all precisions.
  5. (Anti-Uniformity) In expectation, if we sample every uniformly from with replacement, .
  6. (Monotonicity) With , (smaller precision shouldn't lead to a greater score, since if you're uncalibrated at a precision of 10%, you're not going to be suddenly calibrated at a precision of 5%).
    1. We can't just do this by multiplying the result with the precision, since that would violate condition 4.
    2. This gets violated if one uses a proper scoring rule that is always negative, but if we flip the score this can still trivially hold.

But what should be done about a calibration plot that looks like this?

A lopsided calibration plot: Linear and ascending up to 0.5, and then linearly descending to 0

There are two ways of arguing what, morally, the precision of the forecasts is:

  • The argument for having a score of for every precision (assuming a large ) is that inferring the correct way to make forecasts from this calibration plot is trivial: With forecasts of probability , re-assign a probability .
  • The argument against giving a score of is that extending this rule would mean that at every kind of correction on the plot is valid, but there is no clear cutoff point that prevents us from applying this to individual predictions ("If you predict 99% instead of 43%, and 1% instead of 13%, and 1% instead of 23%, and […], then you achieve perfect resolution and calibration.")

Appendix B: Further Idea Sketches for Algorithms

  • Algorithms for quantifying the precision of calibration plots
    • Input: A list of n forecasts and their resolutions
    • First idea:
      • For i=2, n
        • Segment the forecasts into i different segments, ordered by probability
        • Calculate average outcome
        • For two adjacent segments, calculate the slope for those values
        • Append the mean of all slopes of adjacent segments to the array output
      • Return output
    • Second idea:
      • For i=2, n
        • Segment the forecasts into i different segments, ordered by probability
        • Re-scale each segment to give probabilities from 0 to 1
        • Use a proper scoring rule? Idk I haven't thought this through
    • Third idea:
      • Something like the first idea, but with a sliding window
    • Fourth idea:
      • Average linear regression of all subsequences with length
  • Start with perfect predictor, the level of noise at which its Brier score is equal to the dataset
  • Additional ideas:
    • Multiply score with the average number of datapoints inside the given precision
New Comment
5 comments, sorted by Click to highlight new comments since:

I don't understand what question you're trying to answer here. e.g., I could imagine questions of the form "If I see someone forecast 39.845%, should I treat this nearly the same as if they had forecast 40%?" Most mathematical ways of dealing with forecasts do (e.g. scoring rules, formulas for aggregating multiple people's forecasts, decision rules based on EV). But that doesn't seem like what you're interested in here.

Also not sure if/how these graphs differ from what they'd look like if they were based on forecasts which were perfectly calibrated with arbitrary precision.

Also, the description of Omar sounds impossible. If (e.g.) his 25% forecasts come true more often than his 5% forecasts, then his 15% forecasts must differ from at least one of those two (though perhaps you could have too little data to be able to tell).

methods based on rounding probabilities are hot flaming garbage

I think this depends a lot on what you're interested in, i.e. what scoring rules you use. Someone who runs the same analysis with Brier instead of log-scores might disagree.

More generally, I'm not convinced it makes sense to think of "precision" as a constant, let alone a universal one, since it depends on

  • the scoring rule in question: Imagine a set of forecasts that's awfully calibrated on values <1% and >99%, but perfectly calibrated on values between 1% and 99%. With the log-score, this will probably get a bad precision value, while with Brier this would give a great one. 
  • someone's calibration, as you point out with your final calibration plot.

I believe that these approaches are not good: For small datasets they produce large oscillations in the score, not smooth declines, and they improve the scores of worse-than-random forecast datasets.

I don't think it's very counterintuitive/undesirable for (what, in practice, is essentially) noise to make worse-than-random forecasts better. As a matter of fact, this also happens if you replace log-scores with Brier in your analysis with random noise instead of rounding.

Also, regarding oscillations: I don't think properties of "precision" obtained from small datasets are too important, for similar reasons why I usually don't pay a lot of attention to calibration plots obtained from a handful of forecasts.

As we increase the perturbation, the score falls ~monotonically (which I conjecture to always be true in the limit of infinitely many samples)

This conjecture is true and should easily generalise to more general 1-parameter families of centered, symmetric distributions admitting suitable couplings (e.g. additive N(0,\sigma^2) noise in log-odds space) using the fact that log(sigmoid(x+y))+log(sigmoid(x-y)) is decreasing in y for all log-odds x and all positive y (QED).
(NB: This fails when replacing log-scores with Brier.)

Rounding very strongly rounds everything to 50%, so with strong enough rounding every dataset has the same score.

I could make a similar argument for the noise-based version, if I chose to use Brier (or any other scoring rule S that depends only on |p-outcome| and converges to finite values as p tends towards 0 and 1): With sufficiently strong noise, every forecast becomes ≈0% and ≈100% with equal probability, so the expected score in the "large noise limit" converges to (S(0, outcome) + S(1, outcome))/2.

Hmmm, I’m still thinking about this. I’m kinda unconvinced that you even need an algorithm-heavy approach here. Let’s say that you want to apply logit, add some small amount of noise, apply logistic, then score. Consider the function on R^n defined as (score function) composed with (coordinate-wise logistic function). We care about the expected value of this function with respect to the probability measure induced by our noise. For very small noise, you can approximate this function by its power series expansion. For example, if we’re adding iid Gaussian noise, then look at the second order approximation. Then in the limit as the standard deviation of the noise goes to zero, the expected value of the change is some constant (something something Gaussian integral) times the Laplacian of our function on R^n times the square of the standard deviation. Thus the Laplacian is very related to this precision we care about (it basically determines it for small noise). For most reasonable scoring functions, the Laplacian should have a closed-form solution. I think that gets you out of having to simulate anything. Let me know if I messed anything up! Cheers!

Awesome post! I'm very ignorant of the precision-estimation literature so I'm going to be asking dumb questions here.

First of all, I feel like a precision function should take some kind of "acceptable loss" parameter. From what I gather, to specify the precision you need some threshold in your algorithm(s) for how much accuracy loss you're willing to tolerate. 

More fundamentally, though, I'm trying to understand what exactly we want to measure. The list of desired properties of a precision function feel somewhat pulled out of thin air, and I'd feel more comfortable with a philosophical understanding of where these properties come from. So let's say we have a set  of possible states/trajectories of the world, the world provides us with some evidence , and we're interested in  for some event . Maybe reality has some fixed  out there, but we're not privy to that, so we're forced to use some "hyperprior" (am I using that word right?) on probability measures over . After conditioning on , we get some probability distribution on , which participants in a prediction market will take the expected value of as their answer. The precision is trying to quantify something like the standard deviation of this probability distribution on values of , right?

P.S. This is entirely a skill issue on my part but I'm not sure what symbols you're using for precision function and perturbation function. Detexify was of no use. Feel free to enlighten me!

If my interpretation of precision function is correct then I guess my main concern is this: how are we reaching inside the minds of the predictors to see what their distribution on  is? Like, imagine we have an urn with black and red marbles in it and we have a prediction market on the probability that a uniformly randomly chosen marble will be red. Let's say that two people participated in this prediction market: Alice and Bob. Alice estimated there to be a 0.3269230769 (or approximately 17/52) chance of the marble being red because she saw the marbles being put in and there were 17 red marbles and 52 marbles total. Bob estimated there to be a 0.3269230769 chance of the marble being red because he felt like it. Bob is clearly providing false precision while Alice is providing entirely justified precision. However, no matter which way the urn draw goes, the input tuple (0.3269230769, 0) or (0.3269230769, 1) will be the same for both participants and thus the precision returned by any precision function will be the same. This feels to me like a fundamental disconnect between what we want to measure and what we are measuring. Am I mistaken in my understanding? Thanks!