You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
(qwen) C:\Users\admin\Code\LLM\ipex-llm\python\llm\dev\benchmark\all-in-one>python run.py
C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\transformers\deepspeed.py:23: FutureWarning: transformers.deepspeed module is deprecated and will be removed in a future version. Please import deepspeed modules directly from transformers.integrations
warnings.warn(
C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torchvision\io\image.py:13: UserWarning: Failed to load image Python extension: 'Could not find module 'C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torchvision\image.pyd' (or one of its dependencies). Try using the full path with constructor syntax.'If you don't plan on using image functionality from torchvision.io, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have libjpeg or libpng installed before building torchvision from source?
warn(
2025-02-25 14:36:22,290 - INFO - intel_extension_for_pytorch auto imported
Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████| 4/4 [00:00<00:00, 14.85it/s]
2025-02-25 14:36:23,500 - INFO - Converting the current model to sym_int4 format......
C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\init.py:412: UserWarning: Initializing zero-element tensors is a no-op
warnings.warn("Initializing zero-element tensors is a no-op")
Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
loading of model costs 6.762704000000667s and 6.6796875GB
<class 'transformers.models.qwen2.modeling_qwen2.Qwen2ForCausalLM'>
C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\transformers\generation\configuration_utils.py:392: UserWarning: do_sample is set to False. However, temperature is set to 0.7 -- this flag is only used in sample-based generation modes. You should set do_sample=True or unset temperature.
warnings.warn(
C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\transformers\generation\configuration_utils.py:397: UserWarning: do_sample is set to False. However, top_p is set to 0.8 -- this flag is only used in sample-based generation modes. You should set do_sample=True or unset top_p.
warnings.warn(
C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\transformers\generation\configuration_utils.py:407: UserWarning: do_sample is set to False. However, top_k is set to 20 -- this flag is only used in sample-based generation modes. You should set do_sample=True or unset top_k.
warnings.warn(
model generate cost: 1.5775966000001063
461 U.S. 238 (1983) OLIM ET AL. v. WAKINEKONA No. 81-1705 Argued: November 1, 1982 --- Decided: May 16, 1983
model generate cost: 0.5141620999966108
461 U.S. 238 (1983) OLIM ET AL. v. WAKINEKONA No. 81-1705 Argued: November 1, 1982 --- Decided: May 16, 1983
model generate cost: 0.5217821000005642
461 U.S. 238 (1983) OLIM ET AL. v. WAKINEKONA No. 81-1705 Argued: November 1, 1982 --- Decided: May 16, 1983
model generate cost: 0.49074429999745917
461 U.S. 238 (1983) OLIM ET AL. v. WAKINEKONA No. 81-1705 Argued: November 1, 1982 --- Decided: May 16, 1983
onednn_verbose,info,oneDNN v3.3.0 (commit 887fb044ccd6308ed1780a3863c2c6f5772c94b3)
onednn_verbose,info,cpu,runtime:threadpool,nthr:12
onednn_verbose,info,cpu,isa:Intel AVX2 with Intel DL Boost
onednn_verbose,info,gpu,runtime:DPC++
onednn_verbose,info,gpu,engine,0,backend:Level Zero,name:Intel(R) Arc(TM) A770 Graphics,driver_version:1.3.31896,binary_kernels:enabled
onednn_verbose,info,graph,backend,0:dnnl_backend
onednn_verbose,info,experimental features are enabled
onednn_verbose,info,use batch_normalization stats one pass is enabled
onednn_verbose,primitive,info,template:operation,engine,primitive,implementation,prop_kind,memory_descriptors,attributes,auxiliary,problem_desc,exec_time
onednn_verbose,graph,info,template:operation,engine,partition_id,partition_kind,op_names,data_formats,logical_tensors,fpmath_mode,backend,exec_time
onednn_verbose,common,error,level_zero,errcode 1879048196
Traceback (most recent call last):
File "C:\Users\admin\Code\LLM\ipex-llm\python\llm\dev\benchmark\all-in-one\run.py", line 1301, in run_transformer_int4_gpu_win
output_ids = model.generate(input_ids, do_sample=False,
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\utils_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\ipex_llm\utils\benchmark_util_4_29.py", line 1563, in generate
return self.greedy_search(
^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\ipex_llm\utils\benchmark_util_4_29.py", line 2385, in greedy_search
outputs = self(
^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\ipex_llm\utils\benchmark_util_4_29.py", line 533, in call
return self.model(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\ipex_llm\transformers\models\qwen2.py", line 147, in qwen2_causal_lm_forward
outputs = self.model(
^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\ipex_llm\transformers\models\qwen2.py", line 107, in qwen2_model_forward
return Qwen2Model.forward(
^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\transformers\models\qwen2\modeling_qwen2.py", line 1058, in forward
layer_outputs = decoder_layer(
^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\transformers\models\qwen2\modeling_qwen2.py", line 773, in forward
hidden_states, self_attn_weights, present_key_value = self.self_attn(
^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\ipex_llm\transformers\models\qwen2.py", line 251, in qwen2_attention_forward
qkv = self.qkv_proj(hidden_states)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\ipex_llm\transformers\low_bit_linear.py", line 664, in forward
result = xe_linear.forward_new(x_2d, w, self.qtype, self.out_len)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
RuntimeError: could not create a primitive
onednn_verbose,common,error,level_zero,errcode 1879048196
Traceback (most recent call last):
File "C:\Users\admin\Code\LLM\ipex-llm\python\llm\dev\benchmark\all-in-one\run.py", line 1301, in run_transformer_int4_gpu_win
output_ids = model.generate(input_ids, do_sample=False,
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\utils_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\ipex_llm\utils\benchmark_util_4_29.py", line 1563, in generate
return self.greedy_search(
^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\ipex_llm\utils\benchmark_util_4_29.py", line 2385, in greedy_search
outputs = self(
^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\ipex_llm\utils\benchmark_util_4_29.py", line 533, in call
return self.model(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\ipex_llm\transformers\models\qwen2.py", line 147, in qwen2_causal_lm_forward
outputs = self.model(
^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\ipex_llm\transformers\models\qwen2.py", line 107, in qwen2_model_forward
return Qwen2Model.forward(
^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\transformers\models\qwen2\modeling_qwen2.py", line 1058, in forward
layer_outputs = decoder_layer(
^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\transformers\models\qwen2\modeling_qwen2.py", line 773, in forward
hidden_states, self_attn_weights, present_key_value = self.self_attn(
^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\ipex_llm\transformers\models\qwen2.py", line 251, in qwen2_attention_forward
qkv = self.qkv_proj(hidden_states)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\ipex_llm\transformers\low_bit_linear.py", line 664, in forward
result = xe_linear.forward_new(x_2d, w, self.qtype, self.out_len)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
RuntimeError: could not create a primitive
(qwen) C:\Users\admin\Code\LLM\ipex-llm\python\llm\dev\benchmark\all-in-one>python run.py
C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\transformers\deepspeed.py:23: FutureWarning: transformers.deepspeed module is deprecated and will be removed in a future version. Please import deepspeed modules directly from transformers.integrations
warnings.warn(
C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torchvision\io\image.py:13: UserWarning: Failed to load image Python extension: 'Could not find module 'C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torchvision\image.pyd' (or one of its dependencies). Try using the full path with constructor syntax.'If you don't plan on using image functionality from
torchvision.io
, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you havelibjpeg
orlibpng
installed before buildingtorchvision
from source?warn(
2025-02-25 14:36:22,290 - INFO - intel_extension_for_pytorch auto imported
Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████| 4/4 [00:00<00:00, 14.85it/s]
2025-02-25 14:36:23,500 - INFO - Converting the current model to sym_int4 format......
C:\Users\admin\anaconda3\envs\qwen\Lib\site-packages\torch\nn\init.py:412: UserWarning: Initializing zero-element tensors is a no-op
warnings.warn("Initializing zero-element tensors is a no-op")
Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
`
<title>英特尔® 驱动程序和支持助理 - 详细报告</title>英特尔® 驱动程序和支持助理
详细报告
上次扫描: 2025-02-25 14:32
计算机
操作系统
处理器
显卡
音频
联网和 I/O
内存
存储
英特尔软件
The text was updated successfully, but these errors were encountered: