-
Hi @rusty1s, I'm doing a link prediction referring to your rgcn_link_pred.py example. I'm a little bit confused about how the negative sampling implemented should be implemented. In this example, it provides the In addition, according to the negative_sampling function used in this sample, it might even sample the negative edges which are already in the Thank you! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 2 replies
-
Hi, and thanks for your interest. You are right that Related to your second issue: Yes, this is a trade-off regarding runtime efficiency and correctness. In general, we do not see a decrease in performance when sampling a positive edge as negative during training once in a while. |
Beta Was this translation helpful? Give feedback.
Hi, and thanks for your interest. You are right that
negative_sampling
might sample positive edges contained in the validation and test set. However, we are not allowed to acknowledge their existence during training to prevent any data leakage. Although this seems to be counter-intuitive at first glance, it is to be expected that such a noise in the learning signal is averaged out during optimization.Related to your second issue: Yes, this is a trade-off regarding runtime efficiency and correctness. In general, we do not see a decrease in performance when sampling a positive edge as negative during training once in a while.