From acc274e806599b9d86dd0059b3976d8210530ab1 Mon Sep 17 00:00:00 2001 From: Daniel Bevenius Date: Sun, 23 Jun 2024 15:39:45 +0200 Subject: [PATCH] Fix typo in llama_set_embeddings comment (#8077) --- llama.h | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama.h b/llama.h index 05d8b092b42a4e..53e06d9db52733 100644 --- a/llama.h +++ b/llama.h @@ -786,7 +786,7 @@ extern "C" { // Get the number of threads used for prompt and batch processing (multiple token). LLAMA_API uint32_t llama_n_threads_batch(struct llama_context * ctx); - // Set whether the model is in embeddings model or not + // Set whether the model is in embeddings mode or not // If true, embeddings will be returned but logits will not LLAMA_API void llama_set_embeddings(struct llama_context * ctx, bool embeddings);