You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello, I’m interested in adapting the GBT for my research. In the paper, it’s mentioned that "the AWMA-based transformer module selectively removes components of the attention weight matrix with smaller singular values." However, I haven’t been able to locate the specific part of the code where this modification to the weight matrix occurs.
Additionally, from my understanding of the implementation, it seems that the LRR loss is applied directly to the raw node features. I’d appreciate it if you could confirm whether I’m interpreting this correctly or if there’s a part of the code I might have missed.
Thank you very much for your time and assistance!
The text was updated successfully, but these errors were encountered:
Hello, I’m interested in adapting the GBT for my research. In the paper, it’s mentioned that "the AWMA-based transformer module selectively removes components of the attention weight matrix with smaller singular values." However, I haven’t been able to locate the specific part of the code where this modification to the weight matrix occurs.
Additionally, from my understanding of the implementation, it seems that the LRR loss is applied directly to the raw node features. I’d appreciate it if you could confirm whether I’m interpreting this correctly or if there’s a part of the code I might have missed.
Thank you very much for your time and assistance!
The text was updated successfully, but these errors were encountered: