How to speed up protein screening #481
Replies: 7 comments
-
Hi Gabriel, The inference speed depends not only on sequence lengths but also on other factors like the size of MSA. Did you compare AlphaFold and AlphaPulldown run time on the exact same input, including the same input sequence alignments and templates, same number of recycles, and so on? Best, |
Beta Was this translation helpful? Give feedback.
-
Hi @gabrielpan147 |
Beta Was this translation helpful? Give feedback.
-
Hello all, Thank you for your reply! I haven't tested the official AF on the same input yet, and I agree this is what I need to test. I am also wondering if we focus on AlphaPullDown, do you have any tips& suggestions for accelerating the inferencing speed? |
Beta Was this translation helpful? Give feedback.
-
well, you can limit MSA depth, reduce number of cycles, or play with some other params in config.py, but it's always a trade off between speed and quality and I don't think there is a trick to just accelerate inference without affecting accuracy. Otherwise, it would be well known by now :) we recently found out that conversion to modelCIF might take a while for large PAEs, so maybe turn off this too. |
Beta Was this translation helpful? Give feedback.
-
Exactly as Dima said, unfortunately speed always sacrifices sensitivity and accuracy so it depends on your biological question, e.g. whether you want to find as many interactions as possible in your system, or you are fine with missing some. In addition to the parameters Dima listed above, you can also adjust --num_predictions_per_model=1 and only run one model (e.g.--model_names=model_2_multimer_v3). @DimaMolod, by the way, since this question comes back, maybe we could have some subpage in the doc listing all these settings for the fastest speed, with a warning about accuracy? (we should also occasionally test that AP is as fast as AF with the same input). |
Beta Was this translation helpful? Give feedback.
-
yes, I like the idea: maybe instead of the current exhaustive manual with all possible functionality we should split it by use cases, e.g.
|
Beta Was this translation helpful? Give feedback.
-
Yes, we need to restructure the front page along these lines, let's discuss that separately. |
Beta Was this translation helpful? Give feedback.
-
Hello,
Thank you for developing such a great tool! I am currently doing protein screening based on the pulldown mode. We have several A100 GPU on a slurm based cluster. However, I just found the inferencing speed of the tool is slow: for a 827 residues , the prediction time on a single A100 card was ~150s, significantly slower than the Alphafold suggest prediction speed ~60-90 s(which also inference on A100).
I just followed your installation tutorial but I'm not sure if I configure everything correct. I'm wondering if you have parameter such as "global_config.subbatch_size" to increase the batch size or speed it up? Could you give me some suggestions?
Thanks,
Gabriel
Beta Was this translation helpful? Give feedback.
All reactions