-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unexpected behavior for GGML_MEAN #1005
Comments
Hi, is this issue still up for grabs? If it is, may I take this up? |
Most operations already operate on a per-row basis. If you want a full reduction, in most cases you can do that by creating an 1D view (at worst you might have to make it contiguous). It might make more sense to standardize on that rather than duplicating the ops between a global and per-row version. |
I am not aware of anyone working on this at the moment. If desirable we can talk for a few minutes in order to explain the general structure of the code.
As of right now a per-tensor mean is not used for anything so performance would be completely irrelevant anyways. I think |
It might make more sense to remove |
I would also be fine with having only a single op for summation and defining reductive GGML ops to be per-row by default (but I think that we should then document this since the default in e.g. NumPy is per tensor). As far as I can tell there are currently |
So, these could be the tasks for this issue?
|
That seems to cover it. There are C and CUDA implementations for |
I have played around a little bit with CUDA but I might reach out if I feel stuck. |
The usual behavior for the "mean" operation in numerical frameworks is a reduction of a tensor to a single value. However, in GGML this operation instead calculates the mean per row. This is I think unexpected behavior and the naming is also inconsistent with
GGML_SUM
(per tensor) andGGML_SUM_ROWS
(per row).The text was updated successfully, but these errors were encountered: