-
Notifications
You must be signed in to change notification settings - Fork 6
Description
Migrating this comment rom @procacci in #9 here for better tracking:
My feelings on this, future (bCDs), and on past (CB8) host-guest
challenges are that the focus is too much on force fields while being
somehow unfaithful to the original commitment of the challenge. In my
understanding, that commitment was aimed at identifying the most
appropriate methodology in dealing with the SAMPLing of systems with
disparate time scales/complex collective variables and hence in
producing sufficiently reliable (and reproducible) binding free
energies. WP6 is a mostly rigid host that does not involve any
serious sampling issue. The very same can be said for the guest
molecules. This is at variance with the reality of ligand-protein
systems where both the host (the binding pocket) and the guest
(drug-like compounds with up to 9 rotatable bonds) are characterized
by a complex conformational landscape. Besides, the issue of the
protonation state of heavily charged WP6 is an extra complication that
can affect severely the prediction. If WP6 is in part protonated at ph
7.4, then the BFEs of the guest cations (12 out of 13) computed
assuming a fully deprotonated state for WP6 are expected to be
overestimated. On top of all this, electrostatic interactions have
certainly a prevailing role in shaping the affinity. So, despite all
these modelization challenges, it was quite a surprise for me to see
that all the MD-based methodologies performed decently, whether they
used openFF GAFF or AMOEBA.I do think that testing FF is important but I also think that the
SAMPLing issue in BFE determination is still far from being solved for
protein-ligand systems, even in the GPU era. My hope is that in the
future challenges these two crucial issues (FF and SAMPLing) are
addressed in separate specialized sessions: i) test the FF (and any
other physical methods) for rigid host and guest challenges ii) test
MD-based methodologies for protein-ligand systems, requesting for all
participants to use only one (well-established) FF. The common set of
topological/parameter files for the most popular MD codes could be
provided by the organizers with little work.Using a fancy and expensive FF (e.g. AMOEBA or QM-M) with a poor
sampling methodology in a system where sampling is important makes no
real sense in my view. By the same token, using a fancy sampling
methodology with a poor FF in systems where sampling is not an issue
makes no sense as well. I am quite convinced that disentangling the FF
issue (systematic errors) from the sampling issue (reproducibility)
will allow us to move forward more quickly.Piero Procacci
I'll try and take this point-by-point for the record:
My feelings on this, future (bCDs), and on past (CB8) host-guest
challenges are that the focus is too much on force fields while being
somehow unfaithful to the original commitment of the challenge. In my
understanding, that commitment was aimed at identifying the most
appropriate methodology in dealing with the SAMPLing of systems with
disparate time scales/complex collective variables and hence in
producing sufficiently reliable (and reproducible) binding free
energies.
Honestly, the SAMPL challenges are not staged with a particular purpose in mind other than to (a) allow fair assessment of the state of the art in prospective challenges, and (b) to drive progress in the field. Whether that progress comes from force fields or from sampling methods is dependent on participation/innovation, not on the structure of the challenges necessarily.
That said, the SAMPL6 "SAMPLing" challenge did have a clear focus specifically on sampling methods. One thing on my list has been to stage another "SAMPLing" challenge as I see it important to focus some amount of attention directly on sampling issues (as opposed to overall accuracy, which can be the result of many factors) but I haven't had the bandwidth to orchestrate another such challenge recently, as it requires a great deal of engagement with the community. Are you interested in helping out with organizing one?
WP6 is a mostly rigid host that does not involve any
serious sampling issue. The very same can be said for the guest
molecules. This is at variance with the reality of ligand-protein
systems where both the host (the binding pocket) and the guest
(drug-like compounds with up to 9 rotatable bonds) are characterized
by a complex conformational landscape.
This seems like it's an argument for NOT using WP6 for a SAMPLing challenge, but not necessarily for not including WP6 in SAMPL, unless I'm missing something.
Besides, the issue of the
protonation state of heavily charged WP6 is an extra complication that
can affect severely the prediction. If WP6 is in part protonated at ph
7.4, then the BFEs of the guest cations (12 out of 13) computed
assuming a fully deprotonated state for WP6 are expected to be
overestimated.
I did just add some experimental data on WP6 protonation states in #11
I do think that testing FF is important but I also think that the
SAMPLing issue in BFE determination is still far from being solved for
protein-ligand systems, even in the GPU era. My hope is that in the
future challenges these two crucial issues (FF and SAMPLing) are
addressed in separate specialized sessions: i) test the FF (and any
other physical methods) for rigid host and guest challenges ii) test
MD-based methodologies for protein-ligand systems, requesting for all
participants to use only one (well-established) FF. The common set of
topological/parameter files for the most popular MD codes could be
provided by the organizers with little work.
Completely isolating FF from sampling is tricky, in my experience, since even "simple" systems often have some amount of sampling problems. However, as noted, one can pose a "SAMPLing" challenge (as in SAMPL6/as you suggest) to specifically focus on sampling issues. Note, however, such challenges do not necessarily require new data as in such cases the goal is to get the "right" answer (the force field gold standard) efficiently rather than to get the most accurate result.
It is probably time for another such challenge; let me know if you're interested in lending a hand with organization.
Using a fancy and expensive FF (e.g. AMOEBA or QM-M) with a poor
sampling methodology in a system where sampling is important makes no
real sense in my view. By the same token, using a fancy sampling
methodology with a poor FF in systems where sampling is not an issue
makes no sense as well. I am quite convinced that disentangling the FF
issue (systematic errors) from the sampling issue (reproducibility)
will allow us to move forward more quickly.
In general I agree with you that one needs to address both sampling and force field quality and if sampling is the limiting factor, force field quality won't fix it. That said, my role in running these challenge is... to make them available to the community and let them use the challenges to drive progress, not to decide who gets to use which methods and how. I would note that over the years, far fewer pure QM methods have participated in SAMPL host-guest challenges and far more methods have begun incorporating extensive sampling, as this has proven critical for decent performance. Yes, we still see more expensive force fields used -- but there are far fewer methods now which just do expensive QM calculations on host-guest complexes.
Anyway, let me know if you're interested in helping orchestrate another round of SAMPLing challenges, and what system(s) you think would be best suited for those.