From ceb296aba20f1d3224c093ddc8124d964bedb53f Mon Sep 17 00:00:00 2001 From: Lorenzo Toniazzi Date: Sun, 11 Aug 2024 00:46:12 +0100 Subject: [PATCH] New base prompt --- tests/test_lora_conversion_and_inference.sh | 7 +++---- 1 file changed, 3 insertions(+), 4 deletions(-) diff --git a/tests/test_lora_conversion_and_inference.sh b/tests/test_lora_conversion_and_inference.sh index 2beb7b3c294a6d..004d1ee99c9e63 100755 --- a/tests/test_lora_conversion_and_inference.sh +++ b/tests/test_lora_conversion_and_inference.sh @@ -18,7 +18,6 @@ results=() run_conversion_and_inference_lora() { local model_name=$1 local size_matrix=$2 - local bos_token=$3 # Convert safetensors to gguf echo "Running convert_hf_to_gguf.py for $model_name with size $size_matrix..." @@ -40,7 +39,7 @@ run_conversion_and_inference_lora() { # Run inference echo "Running llama-cli without lora for $model_name with size $size_matrix..." OUTPUT_BASE=$(llama-cli -m $MODELS_REPO/$model_name/size=$size_matrix/base/Base-F32.gguf \ - -p "When forty winters shall besiege" -n 50 --seed 42) + -p "Look again at that dot." -n 50 --seed 42) echo "Running llama-cli with lora for $model_name with size $size_matrix..." OUTPUT_LORA_HOT=$(llama-cli -m $MODELS_REPO/$model_name/size=$size_matrix/base/Base-F32.gguf \ @@ -65,8 +64,8 @@ run_conversion_and_inference_lora() { # Array of parameters to iterate over declare -a params=( - "Gemma2ForCausalLM 64 " - "LlamaForCausalLM 64 <|begin_of_text|>" + "Gemma2ForCausalLM 64" + "LlamaForCausalLM 64" ) # Loop through each set of parameters