-
Notifications
You must be signed in to change notification settings - Fork 73
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Non-Equivariant BatchNorm #16
Comments
Hi, thanks for your interest in our work. First, the layer normalization is defined as Actually the layernorm layers were added to the model to stabilize the training process. We did not notice that Second, for the linear layer for the initial vector embedding, it may not be equivarant to the translation operation. This is easy to solve by translating all the center of mass of pocket to the origin of the coordinate. Besides, we speculated that this non-equivariant layer may have minor effect because the substraction of two vectors remained unaffected and equivariant. The model had a chance to learn from the equivariant parts,such as adapted the weights to focus on the substraction of two vector features. Overall, these two layers had impact on the equivariance of the model. The solutions are easy as discussed above, and the influence on the model performance might not be large. Thanks again for your careful reading and pointing out this problem. |
Hi, nice paper published. Retarding to your E(3) model, I think your model might be not translational equivariant. I have similar questions with @oriondollar. Do you have any ideas to solve this issue? Looking forward to your reply. Thank you. |
Hi there,
I've been playing with your codebase to see how equivariant features propagate through different layer types and I think there might be an error in your code. The AttentionInteractionBlockVN normalizes the vector representation with a standard nn.LayerNorm layer which breaks the equivariance of the vector representations inside the encoder. Was this intended? I'm not sure how much of an effect it will have on the rest of the model as the ligand and pocket are jointly encoded. Similarly, a standard nn.Linear layer is used to embed the initial atomic vector representation which also breaks the initial maintenance of equivariance between the atomic coordinates and the machine learned embeddings.
The text was updated successfully, but these errors were encountered: