From 7de7c17f761c68ab1bded5c55e60c752d9822b7f Mon Sep 17 00:00:00 2001 From: Francis Couture-Harpin Date: Thu, 25 Jul 2024 22:58:20 -0400 Subject: [PATCH] llama : give up fix for uint64_t format type by casting to size_t Otherwise some platforms use "%lu" and others "%llu" for uint64_t, and I'm not sure how to handle that. --- src/llama.cpp | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/llama.cpp b/src/llama.cpp index a8d473309d8a33..0c8a86b5374b3e 100644 --- a/src/llama.cpp +++ b/src/llama.cpp @@ -17767,7 +17767,7 @@ struct llama_data_read_context { read_to(&v_size_el_ref, sizeof(v_size_el_ref)); const size_t v_size_el = ggml_type_size(kv_self.v_l[il]->type); if (v_size_el != v_size_el_ref) { - LLAMA_LOG_ERROR("%s: mismatched value element size (%zu != %llu, layer %d)\n", __func__, v_size_el, v_size_el_ref, il); + LLAMA_LOG_ERROR("%s: mismatched value element size (%zu != %zu, layer %d)\n", __func__, v_size_el, (size_t) v_size_el_ref, il); return false; }