-
Notifications
You must be signed in to change notification settings - Fork 105
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: Parallelised folding in Gemini #550
Changes from all commits
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,365 @@ | ||
|
||
#include "gemini.hpp" | ||
#include "barretenberg/common/thread.hpp" | ||
|
||
#include <memory> | ||
#include <vector> | ||
#include <bit> | ||
|
||
/** | ||
* @brief Protocol for opening several multi-linear polynomials at the same point. | ||
* | ||
* | ||
* m = number of variables | ||
* n = 2ᵐ | ||
* u = (u₀,...,uₘ₋₁) | ||
* f₀, …, fₖ₋₁ = multilinear polynomials, | ||
* g₀, …, gₕ₋₁ = shifted multilinear polynomial, | ||
* Each gⱼ is the left-shift of some f↺ᵢ, and gⱼ points to the same memory location as fᵢ. | ||
* v₀, …, vₖ₋₁, v↺₀, …, v↺ₕ₋₁ = multilinear evalutions s.t. fⱼ(u) = vⱼ, and gⱼ(u) = f↺ⱼ(u) = v↺ⱼ | ||
* | ||
* We use a challenge ρ to create a random linear combination of all fⱼ, | ||
* and actually define A₀ = F + G↺, where | ||
* F = ∑ⱼ ρʲ fⱼ | ||
* G = ∑ⱼ ρᵏ⁺ʲ gⱼ, | ||
* G↺ = is the shift of G | ||
* where fⱼ is normal, and gⱼ is shifted. | ||
* The evaluations are also batched, and | ||
* v = ∑ ρʲ⋅vⱼ + ∑ ρᵏ⁺ʲ⋅v↺ⱼ = F(u) + G↺(u) | ||
* | ||
* The prover then creates the folded polynomials A₀, ..., Aₘ₋₁, | ||
* and opens them at different points, as univariates. | ||
* | ||
* We open A₀ as univariate at r and -r. | ||
* Since A₀ = F + G↺, but the verifier only has commitments to the gⱼs, | ||
* we need to partially evaluate A₀ at both evaluation points. | ||
* As univariate, we have | ||
* A₀(X) = F(X) + G↺(X) = F(X) + G(X)/X | ||
* So we define | ||
* - A₀₊(X) = F(X) + G(X)/r | ||
* - A₀₋(X) = F(X) − G(X)/r | ||
* So that A₀₊(r) = A₀(r) and A₀₋(-r) = A₀(-r). | ||
* The verifier is able to computed the simulated commitments to A₀₊(X) and A₀₋(X) | ||
* since they are linear-combinations of the commitments [fⱼ] and [gⱼ]. | ||
*/ | ||
namespace proof_system::honk::pcs::gemini { | ||
|
||
/** | ||
* @brief Computes d-1 fold polynomials Fold_i, i = 1, ..., d-1 | ||
* | ||
* @param mle_opening_point multilinear opening point 'u' | ||
* @param batched_unshifted F(X) = ∑ⱼ ρʲ fⱼ(X) | ||
* @param batched_to_be_shifted G(X) = ∑ⱼ ρᵏ⁺ʲ gⱼ(X) | ||
* @return std::vector<Polynomial> | ||
*/ | ||
template <typename Params> | ||
std::vector<typename barretenberg::Polynomial<typename Params::Fr>> MultilinearReductionScheme< | ||
Params>::compute_fold_polynomials(std::span<const Fr> mle_opening_point, | ||
Polynomial&& batched_unshifted, | ||
Polynomial&& batched_to_be_shifted) | ||
{ | ||
|
||
using Fr = typename Params::Fr; | ||
using Polynomial = barretenberg::Polynomial<Fr>; | ||
|
||
const size_t num_variables = mle_opening_point.size(); // m | ||
|
||
const size_t num_threads = get_num_cpus_pow2(); | ||
constexpr size_t efficient_operations_per_thread = 64; // A guess of the number of operation for which there | ||
// would be a point in sending them to a separate thread | ||
|
||
// Allocate space for m+1 Fold polynomials | ||
// | ||
// The first two are populated here with the batched unshifted and to-be-shifted polynomial respectively. | ||
// They will eventually contain the full batched polynomial A₀ partially evaluated at the challenges r,-r. | ||
// This function populates the other m-1 polynomials with the foldings of A₀. | ||
std::vector<Polynomial> fold_polynomials; | ||
fold_polynomials.reserve(num_variables + 1); | ||
|
||
// F(X) = ∑ⱼ ρʲ fⱼ(X) and G(X) = ∑ⱼ ρᵏ⁺ʲ gⱼ(X) | ||
Polynomial& batched_F = fold_polynomials.emplace_back(std::move(batched_unshifted)); | ||
Polynomial& batched_G = fold_polynomials.emplace_back(std::move(batched_to_be_shifted)); | ||
constexpr size_t offset_to_folded = 2; // Offset because of F an G | ||
// A₀(X) = F(X) + G↺(X) = F(X) + G(X)/X. | ||
Polynomial A_0(batched_F); | ||
A_0 += batched_G.shifted(); | ||
|
||
// Allocate everything before parallel computation | ||
for (size_t l = 0; l < num_variables - 1; ++l) { | ||
// size of the previous polynomial/2 | ||
const size_t n_l = 1 << (num_variables - l - 1); | ||
|
||
// A_l_fold = Aₗ₊₁(X) = (1-uₗ)⋅even(Aₗ)(X) + uₗ⋅odd(Aₗ)(X) | ||
fold_polynomials.emplace_back(Polynomial(n_l)); | ||
} | ||
|
||
// A_l = Aₗ(X) is the polynomial being folded | ||
// in the first iteration, we take the batched polynomial | ||
// in the next iteration, it is the previously folded one | ||
auto A_l = A_0.data(); | ||
for (size_t l = 0; l < num_variables - 1; ++l) { | ||
// size of the previous polynomial/2 | ||
const size_t n_l = 1 << (num_variables - l - 1); | ||
|
||
// Use as many threads as it is useful so that 1 thread doesn't process 1 element, but make sure that there is | ||
// at least 1 | ||
size_t num_used_threads = std::min(n_l / efficient_operations_per_thread, num_threads); | ||
num_used_threads = num_used_threads ? num_used_threads : 1; | ||
size_t chunk_size = n_l / num_used_threads; | ||
size_t last_chunk_size = (n_l % chunk_size) ? (n_l % num_used_threads) : chunk_size; | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I think this robustness is nice to have but I wonder if in practice we can always assume (otherwise, enforce) that num_threads is a power of 2 and thus divides There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Lots of 6-core laptops out there) |
||
|
||
// Openning point is the same for all | ||
const Fr u_l = mle_opening_point[l]; | ||
|
||
// A_l_fold = Aₗ₊₁(X) = (1-uₗ)⋅even(Aₗ)(X) + uₗ⋅odd(Aₗ)(X) | ||
auto A_l_fold = fold_polynomials[l + offset_to_folded].data(); | ||
|
||
parallel_for(num_used_threads, [&](size_t i) { | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. As far as I can tell, There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I suppose since There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Yes, that's why I get the value of the number of available threads first and then compute num_used_threads based on that |
||
size_t current_chunk_size = (i == (num_used_threads - 1)) ? last_chunk_size : chunk_size; | ||
for (std::ptrdiff_t j = (std::ptrdiff_t)(i * chunk_size); | ||
j < (std::ptrdiff_t)((i * chunk_size) + current_chunk_size); | ||
j++) { | ||
// fold(Aₗ)[j] = (1-uₗ)⋅even(Aₗ)[j] + uₗ⋅odd(Aₗ)[j] | ||
// = (1-uₗ)⋅Aₗ[2j] + uₗ⋅Aₗ[2j+1] | ||
// = Aₗ₊₁[j] | ||
A_l_fold[j] = A_l[j << 1] + u_l * (A_l[(j << 1) + 1] - A_l[j << 1]); | ||
} | ||
}); | ||
// set Aₗ₊₁ = Aₗ for the next iteration | ||
A_l = A_l_fold; | ||
} | ||
|
||
return fold_polynomials; | ||
}; | ||
|
||
/** | ||
* @brief Computes/aggragates d+1 Fold polynomials and their opening pairs (challenge, evaluation) | ||
* | ||
* @details This function assumes that, upon input, last d-1 entries in fold_polynomials are Fold_i. | ||
* The first two entries are assumed to be, respectively, the batched unshifted and batched to-be-shifted | ||
* polynomials F(X) = ∑ⱼ ρʲfⱼ(X) and G(X) = ∑ⱼ ρᵏ⁺ʲ gⱼ(X). This function completes the computation | ||
* of the first two Fold polynomials as F + G/r and F - G/r. It then evaluates each of the d+1 | ||
* fold polynomials at, respectively, the points r, rₗ = r^{2ˡ} for l = 0, 1, ..., d-1. | ||
* | ||
* @param mle_opening_point u = (u₀,...,uₘ₋₁) is the MLE opening point | ||
* @param fold_polynomials vector of polynomials whose first two elements are F(X) = ∑ⱼ ρʲfⱼ(X) | ||
* and G(X) = ∑ⱼ ρᵏ⁺ʲ gⱼ(X), and the next d-1 elements are Fold_i, i = 1, ..., d-1. | ||
* @param r_challenge univariate opening challenge | ||
*/ | ||
template <typename Params> | ||
ProverOutput<Params> MultilinearReductionScheme<Params>::compute_fold_polynomial_evaluations( | ||
std::span<const Fr> mle_opening_point, std::vector<Polynomial>&& fold_polynomials, const Fr& r_challenge) | ||
{ | ||
|
||
using Fr = typename Params::Fr; | ||
using Polynomial = barretenberg::Polynomial<Fr>; | ||
|
||
const size_t num_variables = mle_opening_point.size(); // m | ||
|
||
Polynomial& batched_F = fold_polynomials[0]; // F(X) = ∑ⱼ ρʲ fⱼ(X) | ||
Polynomial& batched_G = fold_polynomials[1]; // G(X) = ∑ⱼ ρᵏ⁺ʲ gⱼ(X) | ||
|
||
// Compute univariate opening queries rₗ = r^{2ˡ} for l = 0, 1, ..., m-1 | ||
std::vector<Fr> r_squares = squares_of_r(r_challenge, num_variables); | ||
|
||
// Compute G/r | ||
Fr r_inv = r_challenge.invert(); | ||
batched_G *= r_inv; | ||
|
||
// Construct A₀₊ = F + G/r and A₀₋ = F - G/r in place in fold_polynomials | ||
Polynomial tmp = batched_F; | ||
Polynomial& A_0_pos = fold_polynomials[0]; | ||
|
||
// A₀₊(X) = F(X) + G(X)/r, s.t. A₀₊(r) = A₀(r) | ||
A_0_pos += batched_G; | ||
|
||
// Perform a swap so that tmp = G(X)/r and A_0_neg = F(X) | ||
std::swap(tmp, batched_G); | ||
Polynomial& A_0_neg = fold_polynomials[1]; | ||
|
||
// A₀₋(X) = F(X) - G(X)/r, s.t. A₀₋(-r) = A₀(-r) | ||
A_0_neg -= tmp; | ||
|
||
std::vector<OpeningPair<Params>> fold_poly_opening_pairs; | ||
fold_poly_opening_pairs.reserve(num_variables + 1); | ||
|
||
// Compute first opening pair {r, A₀(r)} | ||
fold_poly_opening_pairs.emplace_back(OpeningPair<Params>{ r_challenge, fold_polynomials[0].evaluate(r_challenge) }); | ||
|
||
// Compute the remaining m opening pairs {−r^{2ˡ}, Aₗ(−r^{2ˡ})}, l = 0, ..., m-1. | ||
for (size_t l = 0; l < num_variables; ++l) { | ||
fold_poly_opening_pairs.emplace_back( | ||
OpeningPair<Params>{ -r_squares[l], fold_polynomials[l + 1].evaluate(-r_squares[l]) }); | ||
} | ||
|
||
return { fold_poly_opening_pairs, std::move(fold_polynomials) }; | ||
}; | ||
|
||
/** | ||
* @brief Checks that all MLE evaluations vⱼ contained in the list of m MLE opening claims | ||
* is correct, and returns univariate polynomial opening claims to be checked later | ||
* | ||
* @param mle_opening_point the MLE evaluation point u | ||
* @param batched_evaluation batched evaluation from multivariate evals at the point u | ||
* @param batched_f batched commitment to unshifted polynomials | ||
* @param batched_g batched commitment to to-be-shifted polynomials | ||
* @param proof commitments to the m-1 folded polynomials, and alleged evaluations. | ||
* @param transcript | ||
* @return Fold polynomial opening claims: (r, A₀(r), C₀₊), (-r, A₀(-r), C₀₋), and | ||
* (Cⱼ, Aⱼ(-r^{2ʲ}), -r^{2}), j = [1, ..., m-1] | ||
*/ | ||
|
||
template <typename Params> | ||
std::vector<OpeningClaim<Params>> MultilinearReductionScheme<Params>::reduce_verify( | ||
std::span<const Fr> mle_opening_point, /* u */ | ||
const Fr batched_evaluation, /* all */ | ||
GroupElement& batched_f, /* unshifted */ | ||
GroupElement& batched_g, /* to-be-shifted */ | ||
VerifierTranscript<Fr>& transcript) | ||
{ | ||
|
||
using Fr = typename Params::Fr; | ||
using Commitment = typename Params::Commitment; | ||
const size_t num_variables = mle_opening_point.size(); | ||
|
||
// Get polynomials Fold_i, i = 1,...,m-1 from transcript | ||
std::vector<Commitment> commitments; | ||
commitments.reserve(num_variables - 1); | ||
for (size_t i = 0; i < num_variables - 1; ++i) { | ||
auto commitment = transcript.template receive_from_prover<Commitment>("Gemini:FOLD_" + std::to_string(i + 1)); | ||
commitments.emplace_back(commitment); | ||
} | ||
|
||
// compute vector of powers of random evaluation point r | ||
const Fr r = transcript.get_challenge("Gemini:r"); | ||
std::vector<Fr> r_squares = squares_of_r(r, num_variables); | ||
|
||
// Get evaluations a_i, i = 0,...,m-1 from transcript | ||
std::vector<Fr> evaluations; | ||
evaluations.reserve(num_variables); | ||
for (size_t i = 0; i < num_variables; ++i) { | ||
auto eval = transcript.template receive_from_prover<Fr>("Gemini:a_" + std::to_string(i)); | ||
evaluations.emplace_back(eval); | ||
} | ||
|
||
// Compute evaluation A₀(r) | ||
auto a_0_pos = compute_eval_pos(batched_evaluation, mle_opening_point, r_squares, evaluations); | ||
|
||
// C₀_r_pos = ∑ⱼ ρʲ⋅[fⱼ] + r⁻¹⋅∑ⱼ ρᵏ⁺ʲ [gⱼ] | ||
// C₀_r_pos = ∑ⱼ ρʲ⋅[fⱼ] - r⁻¹⋅∑ⱼ ρᵏ⁺ʲ [gⱼ] | ||
auto [c0_r_pos, c0_r_neg] = compute_simulated_commitments(batched_f, batched_g, r); | ||
|
||
std::vector<OpeningClaim<Params>> fold_polynomial_opening_claims; | ||
fold_polynomial_opening_claims.reserve(num_variables + 1); | ||
|
||
// ( [A₀₊], r, A₀(r) ) | ||
fold_polynomial_opening_claims.emplace_back(OpeningClaim<Params>{ { r, a_0_pos }, c0_r_pos }); | ||
// ( [A₀₋], -r, A₀(-r) ) | ||
fold_polynomial_opening_claims.emplace_back(OpeningClaim<Params>{ { -r, evaluations[0] }, c0_r_neg }); | ||
for (size_t l = 0; l < num_variables - 1; ++l) { | ||
// ([A₀₋], −r^{2ˡ}, Aₗ(−r^{2ˡ}) ) | ||
fold_polynomial_opening_claims.emplace_back( | ||
OpeningClaim<Params>{ { -r_squares[l + 1], evaluations[l + 1] }, commitments[l] }); | ||
} | ||
|
||
return fold_polynomial_opening_claims; | ||
}; | ||
|
||
template <typename Params> | ||
std::vector<typename Params::Fr> MultilinearReductionScheme<Params>::powers_of_rho(const Fr rho, | ||
const size_t num_powers) | ||
{ | ||
|
||
using Fr = typename Params::Fr; | ||
std::vector<Fr> rhos = { Fr(1), rho }; | ||
rhos.reserve(num_powers); | ||
for (size_t j = 2; j < num_powers; j++) { | ||
rhos.emplace_back(rhos[j - 1] * rho); | ||
} | ||
return rhos; | ||
}; | ||
/** | ||
* @brief Compute the expected evaluation of the univariate commitment to the batched polynomial. | ||
* | ||
* @param batched_mle_eval The evaluation of the folded polynomials | ||
* @param mle_vars MLE opening point u | ||
* @param r_squares squares of r, r², ..., r^{2ᵐ⁻¹} | ||
* @param fold_polynomial_evals series of Aᵢ₋₁(−r^{2ⁱ⁻¹}) | ||
* @return evaluation A₀(r) | ||
*/ | ||
template <typename Params> | ||
typename Params::Fr MultilinearReductionScheme<Params>::compute_eval_pos(const Fr batched_mle_eval, | ||
std::span<const Fr> mle_vars, | ||
std::span<const Fr> r_squares, | ||
std::span<const Fr> fold_polynomial_evals) | ||
{ | ||
using Fr = typename Params::Fr; | ||
const size_t num_variables = mle_vars.size(); | ||
|
||
const auto& evals = fold_polynomial_evals; | ||
|
||
// Initialize eval_pos with batched MLE eval v = ∑ⱼ ρʲ vⱼ + ∑ⱼ ρᵏ⁺ʲ v↺ⱼ | ||
Fr eval_pos = batched_mle_eval; | ||
for (size_t l = num_variables; l != 0; --l) { | ||
const Fr r = r_squares[l - 1]; // = rₗ₋₁ = r^{2ˡ⁻¹} | ||
const Fr eval_neg = evals[l - 1]; // = Aₗ₋₁(−r^{2ˡ⁻¹}) | ||
const Fr u = mle_vars[l - 1]; // = uₗ₋₁ | ||
|
||
// The folding property ensures that | ||
// Aₗ₋₁(r^{2ˡ⁻¹}) + Aₗ₋₁(−r^{2ˡ⁻¹}) Aₗ₋₁(r^{2ˡ⁻¹}) - Aₗ₋₁(−r^{2ˡ⁻¹}) | ||
// Aₗ(r^{2ˡ}) = (1-uₗ₋₁) ----------------------------- + uₗ₋₁ ----------------------------- | ||
// 2 2r^{2ˡ⁻¹} | ||
// We solve the above equation in Aₗ₋₁(r^{2ˡ⁻¹}), using the previously computed Aₗ(r^{2ˡ}) in eval_pos | ||
// and using Aₗ₋₁(−r^{2ˡ⁻¹}) sent by the prover in the proof. | ||
eval_pos = ((r * eval_pos * 2) - eval_neg * (r * (Fr(1) - u) - u)) / (r * (Fr(1) - u) + u); | ||
} | ||
|
||
return eval_pos; // return A₀(r) | ||
}; | ||
|
||
/** | ||
* @brief Compute squares of folding challenge r | ||
* | ||
* @tparam Params | ||
* @param r | ||
* @param num_squares The number of foldings | ||
* @return std::vector<typename Params::Fr> | ||
*/ | ||
template <typename Params> | ||
std::vector<typename Params::Fr> MultilinearReductionScheme<Params>::squares_of_r(const Fr r, const size_t num_squares) | ||
{ | ||
std::vector<Fr> squares = { r }; | ||
squares.reserve(num_squares); | ||
for (size_t j = 1; j < num_squares; j++) { | ||
squares.emplace_back(squares[j - 1].sqr()); | ||
} | ||
return squares; | ||
}; | ||
|
||
/** | ||
* @brief Computes two commitments to A₀ partially evaluated in r and -r. | ||
* | ||
* @param batched_f batched commitment to non-shifted polynomials | ||
* @param batched_g batched commitment to to-be-shifted polynomials | ||
* @param r evaluation point at which we have partially evaluated A₀ at r and -r. | ||
* @return std::pair<Commitment, Commitment> c0_r_pos, c0_r_neg | ||
*/ | ||
template <typename Params> | ||
std::pair<typename Params::GroupElement, typename Params::GroupElement> MultilinearReductionScheme< | ||
Params>::compute_simulated_commitments(GroupElement& batched_f, GroupElement& batched_g, Fr r) | ||
{ | ||
// C₀ᵣ₊ = [F] + r⁻¹⋅[G] | ||
GroupElement C0_r_pos = batched_f; | ||
// C₀ᵣ₋ = [F] - r⁻¹⋅[G] | ||
GroupElement C0_r_neg = batched_f; | ||
Fr r_inv = r.invert(); | ||
if (!batched_g.is_point_at_infinity()) { | ||
batched_g *= r_inv; | ||
C0_r_pos += batched_g; | ||
C0_r_neg -= batched_g; | ||
} | ||
return { C0_r_pos, C0_r_neg }; | ||
}; | ||
template class MultilinearReductionScheme<proof_system::honk::pcs::kzg::Params>; | ||
template class MultilinearReductionScheme<proof_system::honk::pcs::ipa::Params>; | ||
}; // namespace proof_system::honk::pcs::gemini |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We can always play with this later but I'm just curious whether this is motivated by some experimentation or if its really just a guess
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
It's a guess. I think it would be helpful to actually check what parallelize_for is useful for. I'll add an issue to concretely evaluate the size of computation for which it is efficient.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Issue: #553