In the high throughput virtual screening workshop of Schrodinger, you can pick how many (%) compounds to allow through to the next higher precision screen - ie after HTVS, allow 10%, then after SP allow 10%, then after XP show 10% of hits. These percentages are all variable by the user.
My question is, has anyone experimented with changing these filters and do they affect the quality of the final percentage of hits received after a high-throughput screen? I am not a computational chemist so am quite inexperienced in this area. I have performed a test run on 10k compounds with the filters as :
HTS-10% SP-10% XP-10% (the preset settings),
HTS-20% SP-10% XP-25% and
HTS-10% SP-20% XP-25%.
I am not worried about increasing computation time, only about whether relaxing the earlier filters allows through a greater number of low-quality hits which may displace good quality hits from the top % of hits shown at the end. This will be significant after scaling up to ca. 1M compounds where we want to dedicate time to as few hits as possible, of the highest possible quality. Surely, in theory, for the highest quality results one should allow as many compounds as possible through to the later, more reliable simulations? However, I suspect that if too many "bad hits" are permitted by the earlier screens, they may outrank good ones in the SP screen and result in promising compounds not making it to XP. As I understand it SP is a slightly more rigorous version of HTS, but XP is significantly improved in terms of the algorithms' reliability. https://www.schrodinger.com/kb/1013
We are trying to find a good balance between accepting more compounds and keeping the top hits as a reliable set of good quality hits which is not too large. It is hard to quantitatively assess "quality of the top 10 results" besides visual inspection. Does anyone have experience with this? Any advice is greatly appreciated. Thanks