Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Assertion error when running EAGLE GPU example #12535

Open
kyang-06 opened this issue Dec 12, 2024 · 3 comments
Open

Assertion error when running EAGLE GPU example #12535

kyang-06 opened this issue Dec 12, 2024 · 3 comments
Assignees

Comments

@kyang-06
Copy link

kyang-06 commented Dec 12, 2024

Hi Team!

I followed this docs to run EAGLE, but encountered assertion error as shown in the picture.

# EAGLE - Speculative Sampling using IPEX-LLM on Intel GPUs

image

I can confirmed that error dumped at the ipdb breakpoint: function layer_outputs = decoder_layer(...).

The execute script is

USE_XETLA=OFF SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=1 SYCL_CACHE_PERSISTENT=1 python -m evaluation.gen_ea_answer_llama2chat_e2_ipex_optimize \ 
--ea-model-path yuhuili/EAGLE-llama2-chat-7B \
--base-model-path meta-llama/Llama-2-7b-chat-hf \
--model-id e2-llama-2-7b-chat --enable-ipex-llm

My conda env is as below, also tried ipex-llm==2.1.0 which does not work:

bigdl-core-xe                    2.5.0b20231210
bigdl-core-xe-21                 2.6.0b20241212
bigdl-core-xe-addons-21          2.6.0b20241212
bigdl-core-xe-batch-21           2.6.0b20241212
bigdl-core-xe-esimd              2.5.0b20231210
bigdl-llm                        2.5.0b20231210
bitsandbytes                     0.41.2.post2
eagle-llm                        1.2.1                /data1/kyyx/github/LLM/speculative_decoding/EAGLE
huggingface-hub                  0.23.4
intel-cmplr-lib-ur               2024.2.0
intel-extension-for-pytorch      2.1.10+xpu
intel-extension-for-transformers 1.2.1
intel-openmp                     2024.2.0
ipex-llm                         2.2.0b20241212
torch                            2.1.0a0+cxx11.abi
torch-grammar                    0.3.3
torchvision                      0.16.0a0+cxx11.abi

As for eagle-llm package, I've tried install branch main or v1, both of them raise the same error.

@jenniew
Copy link
Contributor

jenniew commented Dec 13, 2024

@kyang-06 Can you try earlier version of Eagle2 because recently there are some changes may affect the example code. You can try the older version like Jun 30 commit of main branch to see if there is any issue.

@kyang-06
Copy link
Author

Jun 30 commit bffca4358a6615f9717ce82114d9a69ca9416091 still threw the error. VRAM uses 4.5GB, far away from OOM.

@kyang-06
Copy link
Author

@jenniew Any update?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants