diff --git a/docs/docs/integrations/llms/ipex_llm_gpu.ipynb b/docs/docs/integrations/llms/ipex_llm_gpu.ipynb index 470f5022992ff..040882b8ea922 100644 --- a/docs/docs/integrations/llms/ipex_llm_gpu.ipynb +++ b/docs/docs/integrations/llms/ipex_llm_gpu.ipynb @@ -146,7 +146,12 @@ "\n", "llm = IpexLLM.from_model_id(\n", " model_id=\"lmsys/vicuna-7b-v1.5\",\n", - " model_kwargs={\"temperature\": 0, \"max_length\": 64, \"trust_remote_code\": True, \"device\":\"xpu\"},\n", + " model_kwargs={\n", + " \"temperature\": 0,\n", + " \"max_length\": 64,\n", + " \"trust_remote_code\": True,\n", + " \"device\": \"xpu\",\n", + " },\n", ")" ] }, @@ -213,7 +218,12 @@ " model_id=saved_lowbit_model_path,\n", " tokenizer_id=\"lmsys/vicuna-7b-v1.5\",\n", " # tokenizer_name=saved_lowbit_model_path, # copy the tokenizers to saved path if you want to use it this way\n", - " model_kwargs={\"temperature\": 0, \"max_length\": 64, \"trust_remote_code\": True, \"device\":\"xpu\"},\n", + " model_kwargs={\n", + " \"temperature\": 0,\n", + " \"max_length\": 64,\n", + " \"trust_remote_code\": True,\n", + " \"device\": \"xpu\",\n", + " },\n", ")" ] },