From fb984b930ed1b8c0dae1e167c45c03a88511e52a Mon Sep 17 00:00:00 2001 From: Kai Huang Date: Mon, 27 Jan 2025 15:43:22 +0800 Subject: [PATCH] remove useless --- .../LLM/CPP_Examples/llama-cli-npu.cpp | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/python/llm/example/NPU/HF-Transformers-AutoModels/LLM/CPP_Examples/llama-cli-npu.cpp b/python/llm/example/NPU/HF-Transformers-AutoModels/LLM/CPP_Examples/llama-cli-npu.cpp index fe4cc329255..6783b6ae47f 100644 --- a/python/llm/example/NPU/HF-Transformers-AutoModels/LLM/CPP_Examples/llama-cli-npu.cpp +++ b/python/llm/example/NPU/HF-Transformers-AutoModels/LLM/CPP_Examples/llama-cli-npu.cpp @@ -55,7 +55,6 @@ static void print_usage(int, char ** argv) { vector gguf_tokenize(std::string prompt, gguf_tokenizer_params tok_params) { - int n_tokens = prompt.length() + 2 * tok_params.add_bos; // TODO: no need std::vector ids = llama_tokenize(tok_params.ctx, prompt, tok_params.add_bos, tok_params.parse_special); return ids; @@ -189,7 +188,7 @@ int main(int argc, char ** argv) { exit(1); } - // TODO: rename npu_outfile, it can be a default place where IRs are placed + // TODO: rename npu_outfile, it may be a cache folder together with config and blobs if (params.npu_outfile == "NPU_MODEL") { fprintf(stderr , "\033[31m%s: error: Please provide npu model output dir with -o \033[0m\n" , __func__); exit(1);