18
18
import io .github .ollama4j .models .chat .OllamaChatRequest ;
19
19
import io .github .ollama4j .models .chat .OllamaChatRequestBuilder ;
20
20
import io .github .ollama4j .models .chat .OllamaChatResult ;
21
+ import io .github .ollama4j .models .response .LibraryModelTag ;
21
22
import io .github .ollama4j .models .response .Model ;
22
23
import io .github .ollama4j .types .OllamaModelType ;
23
24
import io .twentysixty .ollama .hologram .chatbot .jms .OllamaProducer ;
@@ -95,7 +96,7 @@ public String getChatResponse(List<OllamaChatMessage> messages) throws OllamaBas
95
96
OllamaAPI ollamaAPI = new OllamaAPI (serviceUrl );
96
97
ollamaAPI .setRequestTimeoutSeconds (timeout );
97
98
98
- OllamaChatRequestBuilder builder = OllamaChatRequestBuilder .getInstance (OllamaModelType . LLAMA2 );
99
+ OllamaChatRequestBuilder builder = OllamaChatRequestBuilder .getInstance ("llama3.2:1b" );
99
100
OllamaChatRequest requestModel = builder .withMessages (messages ).build ();
100
101
OllamaChatResult chatResult = ollamaAPI .chat (requestModel );
101
102
@@ -129,7 +130,7 @@ private void checkModels() {
129
130
try {
130
131
models = ollamaAPI .listModels ();
131
132
for (Model m : models ) {
132
- if (m .getName ().startsWith ("llama2: " )) {
133
+ if (m .getName ().startsWith ("llama3.2:1b " )) {
133
134
modelLoaded = true ;
134
135
}
135
136
logger .info (m .getName ());
@@ -142,7 +143,7 @@ private void checkModels() {
142
143
if (!modelLoaded ) {
143
144
logger .info ("loading model..." );
144
145
try {
145
- ollamaAPI .pullModel (OllamaModelType . LLAMA2 );
146
+ ollamaAPI .pullModel ("llama3.2:1b" );
146
147
} catch (Exception e ) {
147
148
logger .error ("" , e );
148
149
}
0 commit comments