From 581be66cbc643105514f744246c7257ab364cb72 Mon Sep 17 00:00:00 2001 From: rickard Date: Sat, 23 Nov 2024 19:13:10 +0100 Subject: [PATCH] revert back end --- llm_config.yaml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llm_config.yaml b/llm_config.yaml index f8b09602..d2fc7e5e 100644 --- a/llm_config.yaml +++ b/llm_config.yaml @@ -1,6 +1,6 @@ WORD_LIMIT: 200 # max number of words the model is encoraged to generate. not a hard limit SHORT_WORD_LIMIT: 25 # max number of words when asked to write something short. not a hard limit -BACKEND: "llama_cpp" # valid options: "openai", "llama_cpp", "kobold_cpp". if using ooba, use and modify openai template +BACKEND: "kobold_cpp" # valid options: "openai", "llama_cpp", "kobold_cpp". if using ooba, use and modify openai template MEMORY_SIZE: 512 UNLIMITED_REACTS: False DIALOGUE_TEMPLATE: '{{"response":"may be both dialogue and action.", "sentiment":"sentiment based on response", "give":"if any physical item of {character2}s is given as part of the dialogue. Or nothing."}}'