-
Notifications
You must be signed in to change notification settings - Fork 89
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
Result Problem #19
Comments
Hello, thanks for your questions.
I hope this answers your questions. |
Hi, thank you very much for your reply. The first reply answered my question very well, but for the second question, I still want to know more about it, I wonder if it is convenient. Observing the files generated by raw and those generated by processed, it can be found that processed only uses the first line of data of each processing result in the raw folder. I haven't seen the processing process yet, and I am confused about it. When visualizing the data, I found that these points are very discrete. Using openbabel for post-processing, I can’t generate new molecules. I don’t know where the problem is, can you answer it? |
Hello, what was the result of the 'Diversity' indicator when you evaluate it? The following are the results of evaluating the molecules in raw:
|
Hello @pearl-rabbit, the diversity is usually zero when only a single molecule is generated for each protein pocket because we compute this value per target (and afterwards mean and standard deviation across all targets). |
Hi @arneschneuing , thank you for your timely response. I set n_samples=100 and batch_ size=60 in test.py. If I want to generate multiple molecules for a protein pocket, should I set the 'n_samples' value higher? |
Hi, A cutoff value of 1 [Å] seems very low. It is less than a typical bond length. Maybe you should consider a higher threshold to create sufficient edges. |
@PikaQiu520521, I'm sorry but I don't really understand the question.. What do you mean by 'discrete points'? Could you provide an example? |
I will try training again. Is it appropriate to set the cutoff to 3 (my server may not be able to support larger cutoff values)?
'filePath' is the path of the test set sampling results, storing 100 sdf files, each containing the coordinates and atomic types of molecules. (Due to some issues with the server where I store the files, I am unable to provide an example of sdf here temporarily.) |
A cutoff of 3.0 still seems rather low but it might work because information is also propagated through several layers of message passing. However, I haven't tried this value myself yet and can therefore not say if the scores will be similar to the ones from the paper. Your code for reading molecules looks fine to me. Could you please check how many items your print(len(pocket_list))
print([len(x) for x in pocket_list]) |
The output is:
If there is no problem, it may be due to the value of 'cutoff'. |
It's hard to tell. The cutoff value could cause your molecules to be less realistic but I'm still surprised by the |
I didn't carefully analyze the reason, just replaced the original method of calculating diversity with the code in the original annotation section, and obtained a non zero value. By the way, in the text, regarding ca_only model doesn't have a cutoff set, but it's too big for me , I set a limit on the cutoff and achieved effective results. Does this value only affect the number of edges when generating the graph? |
Yes, it determines which atoms (nodes) are connected in the graph that the neural network processes. |
Hello, thank you for your previous question. Unfortunately, I was not able to fully replicate the process, but I used the same calculation method to evaluate the ligands generated by the model. I have two more questions now. 1. The generated ligand files have "row" and "processed" versions, as well as another folder containing paths and scores. I am not sure what they specifically mean. Could you provide some explanation? 2. The generated ligands do not have the bond relationships between atoms, so they are actually discrete points. How do you think about this issue?
The text was updated successfully, but these errors were encountered: