-
Notifications
You must be signed in to change notification settings - Fork 43
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Some question about loss function #8
Comments
The question is occured in single step forecast |
The batch_x and batch_y are not the same. |
When use_tvm=False, we implement Pyraformer by adding an attention mask to the ordinary multi-head attention. Therefore, the setting of use_tvm does not affect the results, but does affect speed and memory usage. |
In your code, batch_x is the same with batch_y. When calculating the MSE, pred and ture represent the 168 steps with shifting one step and the 168 steps with original data, so there are some question about data misalignment, and it is mean that the calculation of MSE is not alignment with original data and predition data with one step difference.
Hope your ansser sincerely!
The text was updated successfully, but these errors were encountered: