-
Notifications
You must be signed in to change notification settings - Fork 56
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
llama llminference result value is strange #461
Comments
The output from llama tflite is strange, have you ran this script https://github.com/google-ai-edge/ai-edge-torch/blob/main/ai_edge_torch/generative/examples/llama/verify.py to validate if the re-authored model is producing reasonable output? |
fail. In the code(verify.py), it use checkpoint file from online.
I don't know why it failed... |
hi, could you update your ai-edge-torch nightly version by this command: pip install ai-edge-torch-nightly |
I already have installed ai-edge-torch nightly version.
|
Description of the bug:
I converted Llama 3.2 1B model using example. ( i have some error with latest ai-edge-torch, so i converted using 5a93316 version)
And I did the llm interference test according to the guide, but the result value is strange.
Is there anyone who has had an experience with llama convert? If so, is this result normal?
Actual vs expected behavior:
This is gemma2 example.
llama tflite
Any other information you'd like to share?
No response
The text was updated successfully, but these errors were encountered: