Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Clarification of assumptions required for dimension scaling results in paper #70

Closed
matt-graham opened this issue Jul 8, 2021 · 0 comments · Fixed by #78
Closed

Clarification of assumptions required for dimension scaling results in paper #70

matt-graham opened this issue Jul 8, 2021 · 0 comments · Fixed by #78

Comments

@matt-graham
Copy link

Submitting as part of JOSS review openjournals/joss-reviews#3397

First, the Bayesian hierarchical methods implemented in `PyEI` rest on modern probabilistic programming tooling [@salvatier2016probabilistic] and gradient-based MCMC methods such as the No U-Turn Sampler (NUTS) [@hoffman2014no]. Using NUTS where possible should allow for faster convergence than existing implementations that rest primarily on Metropolis-Hastings and Gibbs sampling steps. Consider effective sample size, which is a measure of how the variance of the mean of drawn samples compare to the variance of i.i.d. samples from the posterior distribution (or, very roughly, how “effective” the samples are for computing the posterior mean, compared to i.i.d. samples) [@BDA3]. In Metropolis-Hastings, the number of evaluations of the log-posterior required for a given effective sample size scales linearly with the dimensionality of the parameter space, while in Hamiltonian Monte Carlo approaches such as NUTS, the number of required evaluations of the gradient of the log-posterior scales only as the fourth root of the dimension [@neal2011mcmc]. Reasonable scaling with the dimensionality of the parameter space is important in ecological inference, as that dimensionality is large when there are many precincts.

The claim regarding the scaling with dimension for Metropolis–Hastings versus Hamiltonian Monte Carlo (in the sentence in lines 65–69 in proof PDF) should ideally have a clarification that the result requires assumptions on the target distribution (in particular that it is a product of independent factors in each dimension) to hold. To avoid having to introduce technical details this could just simply be adding 'Under certain assumptions on the target posterior distribution,' at the beginning of the sentence.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

1 participant