Skip to content

Commit 7e79bfa

Browse files
committed
add cpu threads
1 parent 29e240e commit 7e79bfa

File tree

1 file changed

+5
-3
lines changed

1 file changed

+5
-3
lines changed

controllers/llamaCPP.cc

+5-3
Original file line numberDiff line numberDiff line change
@@ -339,7 +339,6 @@ void llamaCPP::loadModel(
339339

340340
gpt_params params;
341341

342-
params.cont_batching = false;
343342
// By default will setting based on number of handlers
344343
int drogon_thread = drogon::app().getThreadNum();
345344
LOG_INFO << "Drogon thread is:" << drogon_thread;
@@ -351,8 +350,11 @@ void llamaCPP::loadModel(
351350
// Check if n_parallel exists in jsonBody, if not, set to drogon_thread
352351

353352
params.n_parallel = (*jsonBody).get("n_parallel", drogon_thread).asInt();
354-
355-
params.cont_batching = (*jsonBody)["cont_batching"].asBool();
353+
params.n_threads =
354+
(*jsonBody)
355+
.get("cpu_threads", std::thread::hardware_concurrency())
356+
.asInt();
357+
params.cont_batching = (*jsonBody).get("cont_batching", false).asBool();
356358

357359
this->user_prompt = (*jsonBody).get("user_prompt", "USER: ").asString();
358360
this->ai_prompt = (*jsonBody).get("ai_prompt", "ASSISTANT: ").asString();

0 commit comments

Comments
 (0)