diff --git a/common/common.cpp b/common/common.cpp index 1a62d260..586a4991 100644 --- a/common/common.cpp +++ b/common/common.cpp @@ -896,7 +896,7 @@ struct llama_init_result llama_init_from_gpt_params(gpt_params & params) { device_info dev_info; dev_info.rank = params.rank; - llama_profile_device(&dev_info, model, ml, params.model.c_str(), params.cpuparams.n_threads); + llama_profile_device(&dev_info, model, ml, params.cpuparams.n_threads); // create llama context struct llama_context_params cparams = llama_context_params_from_gpt_params(params); diff --git a/include/llama.h b/include/llama.h index b75640b4..783a9669 100644 --- a/include/llama.h +++ b/include/llama.h @@ -413,8 +413,7 @@ extern "C" { LLAMA_API void llama_profile_device( struct device_info * dev_info, struct llama_model * model, - struct llama_model_loader * ml, - const char * test_file, + struct llama_model_loader * ml, int n_threads); LLAMA_API ggml_backend_buffer_type_t llama_dev_buffer_type(struct llama_model * model, int device);