Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

_forward_step_fn does not always return two values so eval.py breaks if is_pipe_parallel is false #1320

Open
markNZed opened this issue Nov 12, 2024 · 2 comments
Labels
bug Something isn't working

Comments

@markNZed
Copy link
Contributor

This call to _forward_step_fn expects two values returned

_, logits = self._forward_step_fn(model=self.model, data_iterator=inps)

The forward_step can return three values

return loss, outputs, metrics

I guess I am seeing this because I have is_pipe_parallel false and that is uncommon. Maybe there needs to be an option not to return metrics.

@markNZed markNZed added the bug Something isn't working label Nov 12, 2024
@markNZed
Copy link
Contributor Author

There are several "fixes" in https://github.com/markNZed/gpt-neox/tree/pipe_parallel_size_1 which might be related to this. I have not had the time to prepare PR but if someone who knows the code base just looks at the changes there I guess they will quickly see many easy to fix issues.

@iPRET
Copy link

iPRET commented Nov 14, 2024

Can confirm I've run into this issue multiple times aswell, even with pipe parallel size >1.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants