Skip to content

Commit

Permalink
New base prompt
Browse files Browse the repository at this point in the history
  • Loading branch information
Lorenzo Toniazzi committed Aug 10, 2024
1 parent 943baac commit ceb296a
Showing 1 changed file with 3 additions and 4 deletions.
7 changes: 3 additions & 4 deletions tests/test_lora_conversion_and_inference.sh
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,6 @@ results=()
run_conversion_and_inference_lora() {
local model_name=$1
local size_matrix=$2
local bos_token=$3

# Convert safetensors to gguf
echo "Running convert_hf_to_gguf.py for $model_name with size $size_matrix..."
Expand All @@ -40,7 +39,7 @@ run_conversion_and_inference_lora() {
# Run inference
echo "Running llama-cli without lora for $model_name with size $size_matrix..."
OUTPUT_BASE=$(llama-cli -m $MODELS_REPO/$model_name/size=$size_matrix/base/Base-F32.gguf \
-p "When forty winters shall besiege" -n 50 --seed 42)
-p "Look again at that dot." -n 50 --seed 42)

echo "Running llama-cli with lora for $model_name with size $size_matrix..."
OUTPUT_LORA_HOT=$(llama-cli -m $MODELS_REPO/$model_name/size=$size_matrix/base/Base-F32.gguf \
Expand All @@ -65,8 +64,8 @@ run_conversion_and_inference_lora() {

# Array of parameters to iterate over
declare -a params=(
"Gemma2ForCausalLM 64 <bos>"
"LlamaForCausalLM 64 <|begin_of_text|>"
"Gemma2ForCausalLM 64"
"LlamaForCausalLM 64"
)

# Loop through each set of parameters
Expand Down

0 comments on commit ceb296a

Please sign in to comment.