-
Notifications
You must be signed in to change notification settings - Fork 13
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
COMPETITION: A Predictive Model For Series Four #421
Comments
Deadline updated to Jan 16th 2017 as per comment here. |
What kind of in silico model is requested here? A classification model (e.g. 'active' vs 'inactive' above/below a particular potency)? Potency predictions against parasite (in uMol, as directed by 'Column B' as found here? Thanks! |
Activity against PfATP4 according to Kiaran Kirk's ion regulation assay. On 17 Nov 2016 5:45 am, "spadavec" notifications@github.com wrote:
|
@mattodd Thank you for the response! I want to clarify the purpose of this competition a little further, so I hope its ok if I ask a few other questions:
My apologies if I'm asking questions that have been answered elsewhere, but I dug around for a bit and couldn't find anything to these points :) |
No problem at all. Well, it'd be great to understand this PfATP4 target a little better, but the immediate practical concern for us is indeed the triage. We would love to be able to predict better which of OSM's future Series 4 compounds will be actives. So developing a model around Series 4 is the main objective. A model that achieves this may or may not be effective in predicting structures based on other chemotypes. If the mechanism of action is the same, we might expect it to predict those other structures. If, however, PfATP4 has multiple binding sites, then it won't. I have my money on the latter. Accuracy: This is not set. I'm not sure, to be honest - others should feel free to chime in with what is desirable/reasonable. From a pragmatic standpoint I'd like to know whether a proposed structure is predicted to be potent at the 200 nM level (good) or 5 micromolar level (bad). If it's possible to be confident in finer-grained predictions (compound X will be 200 nM, compound Y will be better than that) then great. I had assumed relatively low-definition predictions. You can use whatever data you like. We'd tried to collate the most relevant and useful data into one sheet - i.e. all the compounds for which we have ion regulation data and all the compounds we've made in OSM so far, as well as lit compounds that have been assayed. Be aware that other antimalarials identified in major screens (like the ones you cite) may have different mechanisms of action. They could of course be assumed to be PfATP4-inactives, but I don't know how well that will hold up. The Malaria Box screen mentioned above was of a subset of 400 known antimalarial compounds, and many more than might be expected were active in the "PfATP4" ( more properly, "ion regulation") assay. |
2016 paper from Elizabeth Winzeler may be of interest - homology model of related target in yeast (ScPma1p) and docking with KAE609 and dihydroisoquinolines. |
Thanks again for the clarification! If I'm understanding things properly, we want a in silico model for PfATP4, as it seems to correlate well with whole cell assays/potency well, and is the putative target for the lead series we have--is that correct? In my initial tests, it actually seems to me that the PfATP4 assay doesn't correlate well with whole cell assay data. If you look at all the compounds which are "inactive' in the PfATP4 assay, the average EC50 value is ~1.1 microM, with a standard deviation of 3.08 microM. Conversely, the average EC50 of "active" compounds in the PfATP4 assay is ~0.66 microM, with a standard deviation of 0.56 microM. It appears to me that there isn't a strong correlation between the two assays, or rather, that the PfATP4 assay doesn't have a good discriminatory power (the active EC50 falls well within 1 std. dev. of the inactives EC50). I should not that the data I'm using for this analysis is in the spreadsheet @mattodd compiled here, notably the "Ion Regulation Activity" and "Potency vs. Parasite" columns. If I am looking at this properly, then I think any model based on this data will have a poor discriminatory power. Instead, I believe it would make more sense to actually base the model off the the whole cell potency instead, as it has a lot more data associated with and will lead to better predictions. I would be more than happy to create 2 models (1 based off PfATP4 assay data, and one based off whole cell potency data), if the competition will allow it. My apologies if I'm not understanding the connection properly! |
Hi @spadavec sorry for the delay. For the other question, I think you may be on to something. The "MBox" entries in the spreadsheet are for a set of molecules that is structurally diverse. There are many molecules in there that are potent against the parasite and which have a mechanism of action that is unrelated to PfATP4. This may be skewing things. I'm not sure how best to factor that in. (FYI: Data on a new set of 20 compounds (or so) should be coming this week with potency and PfATP4 data within Series 4, boosting what we can work with in the short term.) |
Submission for competition: http://malaria.ourexperiment.org/in_silico_pfatp4_mo I split it into 2 parts; the first is just using the data that was marked with relevant Pfal assay activity. The second is using all available whole cell data, given that (in my opinion) the Pfal assay doesn't seem to correlate at all with the whole cell potency. |
My submission, using EC50 data and 3D alignment of ligands/fields to predicted protein binding site on PfATPase4. Predict OSM-S364, OSM-S365, OSM-S375 to be most active from Test Set C. |
Hi @spadavec and @holeung - your entries are duly noted, thank you. We were awaiting these final data on relevant and interesting compounds - both frontrunners and newly-synthesised compounds with some more structural variation. The dataset is now, finally, complete. Please accept my apologies for the delay in reaching this point. We will extend the deadline for the competition until probably the end of March, but will clarify this early next week. You can leave your entries as they are or modify them. As indicated in the rules, models will be tested against the Mystery Dataset (which already exists) after the end of the submission period. |
@mattodd Thanks for the update! I've used the new data in my model and posted a new entry in my notebook: |
Hi everybody!
Any answer, feedback and suggestion will be appreciated. |
@gcincilla (Anyone please correct me if I am wrong!)
The "activity against parasite" column is the whole cell assay EC50 value, while the "Ion regulation Activity" column are the results of the biochemical assay against PfATP4, which is the putative target. The intent (correct me if I'm wrong), as to have a QSAR model for the PfATP4 assay data, but you could, in theory, use the whole cell assay data to help.
Either a quantitative or qualitative model is allowed. @mattodd mentioned earlier that a classification model would be allowed and would hope that the cuttoff would enable separation of nM from uM compounds. If you go the regression route, then there isn't a cutoff for accuracy, it just needs to be better than everyone else's and actually help move the project forward (or at least have potential to?).
You are supposed to train your model on compounds which are not marked "B" or "C" in the Ion Regulation Test Set (e.g. A or Lit, etc). The "test sets" come in two forms; first, the quality of your model will be judged against compounds labeled "B" or "C" in the Ion Regulation Test Set column. Then, after the competition date closes, we will be given another set of "secret" compounds (not yet posted) for us to evaluate the potency. |
@spadavec, thank you very much for your reply and explanation. Now everything is clearer. |
Hi @gcincilla - I hope this competition is of interest! Thanks @spadavec for these clarifications. You beat me to it. Yes, the language we used about "test sets" is a little ambiguous. In my mind we would be able to see whether the model is a good one by its ability to predict activity vs PfATP4 for the molecules for which we already have data. There's no restriction on how you do this. But the final test will be vs the mystery set which we'll reveal after the competition closes (which will have structures and PfATP4 assay data in it, as well as whole parasite potency I believe). Models submitted at the time of the competition close need to be run vs that set. Note one thing (which I hope is clear in the above). There appear to be a number of chemotypes hitting PfATP4. We'd love to know how this is possible. It is possible that a model is developed which is predictive for OSM's Series 4 but not predictive for other chemotypes. Or vice versa. Both these situations are interesting, as is the case that one model works for all chemotypes. Since we don't understand the target well, it's as well not to be too prescriptive here. On the one hand we really need to be much more predictive for Series 4 (that, pragmatically, is OSM's priority) but we'd also love to understand how this target can be impacted by so many different structures, even if we're not going to make those in the short term. |
@mattodd and @spadavec, thank you for your replies. I decided to participate in the competition trying to build a good model for PfATP4, hoping this can helps moving research forward. I would have a couple of additional questions for you guys:
|
EC50
As far as I understand it, only 1 model was needed. I made a second model more out of morbid curiosity than anything, and it might eventually get to the point where we want to predict compounds against whole cell, not just Series 4 against PfATP. I am only submitting modelB as my submission to the competition.
In my notebook, I mention that my 'pEC50' is log(EC50) (base 10), hence the negative values. Sorry for the confusion! |
Hi guys - just quickly. Am so sorry for the delay in responding to recent posts and for not yet posting a roundup/info about what happens next. Will get to this in the coming few days. Hang on... |
Hi everyone, I've constructed a Google Sheet with a summary of entries and notebook information ready for the judges and anyone following this competition. Feel free to let me know if you'd like to edit/add anything that could be useful. https://docs.google.com/spreadsheets/d/1pY6sYXIw66jnzUO3CoP8HceYdDjLRvwg5_pLkBY1Wek/edit?usp=sharing |
@alintheopen, thank you for the summary you prepared. In my case it seems the 3 references about QSAR modeling I followed are missing in the references column. They are listed at the end of this notebook entry. |
Just finalising this right now, in fact @gcincilla . I'm hoping in the next day or so, but we're just clarifying how public we can make the dataset at the moment. |
Hi all - so sorry for the delay. The test dataset you need in order to evaluate the models consists of 400 compounds - it’s the MMV Pathogen Box, available here. This has been screened for activity in the Kiaran Kirk ion regulation assay by Adelaide Dennis. We are not, at this stage, making the identity of the hits public; by maintaining the confidentiality of the data set we can continue to use it as a test set for further iterations of any of the models, or indeed alternative models. Instead we will ask all of the submitters of the models to download the spreadsheet of Pathogen Box compounds, run it through their own submitted models and to generate an output for the judges. Ideally we would have outputs that can be easily compared. Can we come to an agreement as to what that would look like? A number? A diagram? The outputs can be uploaded to this Github post, or elsewhere. The judges can then compare those outputs with the actual activities obtained by Kiaran’s team and determine performance of the models. We will have two technical judges who will oversee this evaluation phase, to ensure that everything is done fairly. Can each submitter please “like” this post to indicate their agreement to this way forward and that the dataset is downloadable and suitable? |
An active list of say 50-100 (number to be decided) would be good. It would then be easy to compare with the test set data. Judging should not only be based on hit-rate % but also scaffold variation of correctly predicted active compounds. If results could be submitted in a standard format that would make judging easier too. Probably a csv file of smiles would be the most useful |
@mattodd Just to clarify--are we tasked with ranking/classifying all of the compounds in the "MASTER SHEET" sheet (of which there seems to be ~400 compounds) in the file you attached? I am also fine with creating a report on a subset of the compounds in the sheet, but would like to know what reporting method is preferred. Given that I went with a regression approach, I can both rank-order compounds or provide all compounds that are beyond a certain micoM cutoff, whichever is best for the group! |
Sounds good. I request that the number of compounds to be evaluated be not too large, maybe ~20. My methods are not automated. I like to look and evaluate everything visually. I suggest rank-order as output as that is easy to evaluate and actionable. |
So: a rank order in .csv format that includes SMILES. Everyone OK with that? The dataset contains experimental assay data for all 400 compounds. Some actives, some inactives. I guess the ideal situation is that the model predicts something for every compound, then ranks them. But is that likely to cause problems, based on what you've been saying above? |
Hi everybody,
Thank you in advance |
Hi all, @gcincilla - sorry for the delay, I was in transit to a conference in Boulder where it's snowing, despite our being in the middle of May. I'll consult with the judges on your questions and post back here. No immediate deadline, but each submitter of a model should submit the output of the model. If this does not happen in the next week, I'll chase people manually, to ensure that everyone is aware we're at this stage of the competition. There is an unresolved issue about throughput - whether people can analyse compounds in batch or not - this needs resolving. Is the analysis of all 400 compounds going to be a problem, and should we do something about this if so, e.g. apply the models to a subset of the full Pathogen Box? Please let me know - I think we'd all made the assumption that the models would be able to take a number of molecules as inputs at the same time, in batch. |
Hi everybody,
Molecules are reported in rank order with the most probable to be active at the top so that rank-based evaluation (e.g. AUC) are facilitated. @mattodd, thank you for your answer. Regarding to your question about throughput: I was surprised about the nature of the hidden test set. Meaning that I was expecting a smaller dataset focused on series-4-like compounds while the provided hidden test set is a quite large, general and diverse set of compounds. Having said this the model we developed (as I think all the other models developed by other participants using machine learning techniques) is high-throughput. Meaning it can be be applied in batch on large set of compounds. @mattodd, one last thing: in the Google Sheet with the summary for the judges they didn’t report our references in the specific reference column and I think these can be useful to understand the method we applied. Could you please add those? I added a comment with the references in the specific cell. |
Yes, I can still use 3D-based methods on a dataset of 400 compounds. It will just take more time for the computations and manual analysis of each prediction. |
Hello All, I have attached a csv spreadsheet (google drive link below) with the ranked 400 molecules for PfATP4 activity for the final competition evaluation. The rankings are based on an ensemble of Neural Network models trained on molecular information obtained from the EDragon freeware available here: http://www.vcclab.org/lab/edragon/ This software converts molecular sdf records (smiles) into 1666 pharmacophore fields. These fields were normalized and submitted to a conventional (dense) neural network for training and analysis. The NNs were trained on the 3 reported PfATP4 activity classes; [ACTIVE, PARTIAL, INACTIVE]. The "SCORE" field in the spreadsheet is an average of the probability of each molecule being in the ACTIVE class across all the trained NN models. If anyone is interested, I also have NN sensitivity results available for the 1666 pharmacophore fields (the relative importance of each field in the trained NN models). https://docs.google.com/spreadsheets/d/1SfTbZ4aktv9CHmsxzF40xTZ3kZog_Xvi9LzNr7PUq0c/edit?usp=sharing |
That's great, thanks. I'll take a look at the submitted lists tomorrow. I'm hoping that we might have all models submitted by Monday and the judges can begin to evaluate/compare. Is that timeline reasonable? |
Good with me. |
My latest notebook entry contains my csv file containing the ranking of the 400 MMV box compounds. This is in notebook "MMV400 QSAR modeling". This is based on 3D homology modeling of PfATP4 and field alignment of compounds in Cresset Forge. http://malaria.ourexperiment.org/mmv400_qsar_modelin/15893/post.html |
Hi everybody, Here is my submission to the competition for the hidden set. I have used CDK to calculate molecular descriptors and eXtreme Gradient Boosting (xgboost) to predict the class of the compounds. I will post the source code I used after the competition. The list of compounds is ranked by probability of Active class but I have also included the probabilities for Partial and Inactive. Six of the compounds were predicted to be Active. |
Hey everyone, Here are my model predictions for the MMV dataset. I have both classified and predicted the EC50 activity of the MMV molecules. |
Hi everybody, is there any news on the competition front? |
We've had a tough time getting everyone together - 3 different time zones. I'm taking a new approach of talking to judges in series and have just got the logistical ball rolling. Initial analysis of all entries has been done, and this just needs verification by others. Won't be much longer, sorry for the delay again. |
Will this likely be completed before the Series4 paper comes out? |
Yes! Judging all finished, and just waiting on one bit of formal clearance before announcing winner(s). |
**UPDATE: Competition Deadline Extended to 31st March 2017. All necessary data can be found here (contains all the relevant compounds, with ATP4 activities and potencies).
See YouTube video here
This** competition is to develop a computational model that predicts which molecules will block the malaria parasite's ion pump, PfATP4.
PfATP4 is an important target for the development of new drugs for malaria. We are providing a dataset of actives and inactives. The challenge is to use the data to develop a model that allows us to (better) design compounds that will be active against that target.
The competition soft-launched on 19th August 2016. It will close on 31st March 2017. This competition is part of Open Source Malaria, meaning everything needs to adhere to the Six Laws.
Details are as follows.
Outline We need a predictive model for PfATP4.
PfATP4 is a sodium pump found in the membrane of the malaria parasite. A number of promising antimalarial compounds, with distinct and diverse chemical structures, have been found to be active in an ion regulation assay, which was developed by Kiaran Kirk's lab at the ANU. A number of publications have indicated that this pump is an important new target for malaria medicines. It seems that PfATP4 active compounds disrupt the pump and cause the rapid influx of Na+ into the parasite, leading to its demise. The structure of PfATP4 is not known. Simulations, based on docking of PfATP4 actives, have used a homology model developed by Joseph DeRisi's laboratory. The OSM consortium needs a predictive model for (potency vs PfATP4 activity) to assist in the design and synthesis of new Series 4 compounds and of course to help others working on other compound series.
The first attempt
@murrayfold had a quick, informal attempt (here and here) at the development of a pharmacophore model using known actives and inactives from the MMV Malaria Box. At the time Kiaran Kirk's paper was under embargo but Murray has since written up his work. This initial attempt was unsuccessful (i.e. not predictive - see image below, where the "P Model predictions" correlate poorly with what was found in the ion regulation assay) possibly because the model did not allow for overlapping binding sites or take into consideration compound chirality.
The Competition
We need a predictive in silico model. The best model will win the prize.
How will it work?
OSM will provide:
PfATP4-PNAS2014.pdb.txt
Submission Rules
Open Electronic Notebooks (ELN) such as Labtrove or LabArchives can be useful places to post data and work collaboratively. For example, Ho Leung Ng's ELN can be viewed and commented on here. Please note that LabTrove authors are not alerted when a comment is added to an entry so GitHub is a useful place to tag others.
How will entries be assessed?
NB the 'closed dataset' of compounds may not be publicly available when the winner is announced. All models will be tested against this 'closed dataset' and all details, including the source of this 'closed dataset' will be revealed as soon as the data has been published.
What's the prize
$500
...and the opportunity to contribute to our understanding of a new class of antimalarials
...and authorship on a resulting peer-reviewed publication arising from the OSM consortium
What if none of the models are any good?
Good question. If none of the models prove to be predictive, then it may not be possible to announce a 'winner'. All data will be collated and published at least in the form of a blog, if not a paper...and then we will try again.
Deadline for Entries
30th October 2016, 23:59:59 AEST (Sydney Time)
*A 'valid' entry is one that stands up to the rigour expected from published in silico models. Judges are entitled to use discretion in the case of unconventional entrants, for example those from people with no formal training such as high school students.
An open consultation on how best to run this competition was conducted in Issue #412
Initial competition discussion was started in Issue #417
The text was updated successfully, but these errors were encountered: