From 9cd22177d0495c676203bbc97e5dc05bac35c9a4 Mon Sep 17 00:00:00 2001 From: Lizonghang <870644199@qq.com> Date: Wed, 27 Nov 2024 21:34:45 +0400 Subject: [PATCH] remove arg test_file --- common/common.cpp | 2 +- include/llama.h | 3 +-- 2 files changed, 2 insertions(+), 3 deletions(-) diff --git a/common/common.cpp b/common/common.cpp index 1a62d260..586a4991 100644 --- a/common/common.cpp +++ b/common/common.cpp @@ -896,7 +896,7 @@ struct llama_init_result llama_init_from_gpt_params(gpt_params & params) { device_info dev_info; dev_info.rank = params.rank; - llama_profile_device(&dev_info, model, ml, params.model.c_str(), params.cpuparams.n_threads); + llama_profile_device(&dev_info, model, ml, params.cpuparams.n_threads); // create llama context struct llama_context_params cparams = llama_context_params_from_gpt_params(params); diff --git a/include/llama.h b/include/llama.h index b75640b4..783a9669 100644 --- a/include/llama.h +++ b/include/llama.h @@ -413,8 +413,7 @@ extern "C" { LLAMA_API void llama_profile_device( struct device_info * dev_info, struct llama_model * model, - struct llama_model_loader * ml, - const char * test_file, + struct llama_model_loader * ml, int n_threads); LLAMA_API ggml_backend_buffer_type_t llama_dev_buffer_type(struct llama_model * model, int device);