@@ -50,8 +50,8 @@ async def get_next_action(model, messages, objective, session_id):
50
50
return "coming soon"
51
51
if model == "gemini-pro-vision" :
52
52
return call_gemini_pro_vision (messages , objective ), None
53
- if model == "llava" :
54
- operation = call_ollama_llava (messages )
53
+ if model == "llava" or model == "llava:13b" or "bakllava" or "llava-llama3" :
54
+ operation = call_ollama_llava (messages , model )
55
55
return operation , None
56
56
if model == "claude-3" :
57
57
operation = await call_claude_3_with_ocr (messages , objective , model )
@@ -558,9 +558,11 @@ async def call_gpt_4o_labeled(messages, objective, model):
558
558
return call_gpt_4o (messages )
559
559
560
560
561
- def call_ollama_llava (messages ):
561
+ def call_ollama_llava (messages , model ):
562
+ if model == "" :
563
+ model = "llava"
562
564
if config .verbose :
563
- print ("[call_ollama_llava]" )
565
+ print (f "[call_ollama_llava] model { model } " )
564
566
time .sleep (1 )
565
567
try :
566
568
model = config .initialize_ollama ()
@@ -590,8 +592,8 @@ def call_ollama_llava(messages):
590
592
}
591
593
messages .append (vision_message )
592
594
593
- response = model .chat (
594
- model = "llava" ,
595
+ response = ollama .chat (
596
+ model = model ,
595
597
messages = messages ,
596
598
)
597
599
@@ -633,7 +635,7 @@ def call_ollama_llava(messages):
633
635
)
634
636
if config .verbose :
635
637
traceback .print_exc ()
636
- return call_ollama_llava (messages )
638
+ return call_ollama_llava (messages , model )
637
639
638
640
639
641
async def call_claude_3_with_ocr (messages , objective , model ):
0 commit comments