Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

GBT Model #1

Open
gururgg opened this issue Oct 31, 2024 · 0 comments
Open

GBT Model #1

gururgg opened this issue Oct 31, 2024 · 0 comments

Comments

@gururgg
Copy link

gururgg commented Oct 31, 2024

Hello, I’m interested in adapting the GBT for my research. In the paper, it’s mentioned that "the AWMA-based transformer module selectively removes components of the attention weight matrix with smaller singular values." However, I haven’t been able to locate the specific part of the code where this modification to the weight matrix occurs.

Additionally, from my understanding of the implementation, it seems that the LRR loss is applied directly to the raw node features. I’d appreciate it if you could confirm whether I’m interpreting this correctly or if there’s a part of the code I might have missed.

Thank you very much for your time and assistance!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant