Skip to content

Commit 2602b4f

Browse files
committed
enable cache prompt
1 parent 7c0be9f commit 2602b4f

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

controllers/llamaCPP.cc

+1-1
Original file line numberDiff line numberDiff line change
@@ -164,7 +164,7 @@ void llamaCPP::chatCompletion(
164164
data["frequency_penalty"] =
165165
(*jsonBody).get("frequency_penalty", 0).asFloat();
166166
data["presence_penalty"] = (*jsonBody).get("presence_penalty", 0).asFloat();
167-
167+
data["cache_prompt"] = true;
168168
const Json::Value &messages = (*jsonBody)["messages"];
169169
for (const auto &message : messages) {
170170
std::string input_role = message["role"].asString();

0 commit comments

Comments
 (0)