Skip to content

Commit c9b2f95

Browse files
committed
fix: new FunctionCallType
1 parent 1fae043 commit c9b2f95

File tree

3 files changed

+8
-8
lines changed

3 files changed

+8
-8
lines changed

Cargo.lock

-1
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

src/conversation/mod.rs

+4-4
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ use crate::{
1111
};
1212
use actix_web_lab::sse::Sender;
1313
pub use data::*;
14-
use openai_api_rs::v1::chat_completion::FinishReason;
14+
use openai_api_rs::v1::chat_completion::{FinishReason, FunctionCallType};
1515
use openai_api_rs::v1::{
1616
api::Client,
1717
chat_completion::{
@@ -93,7 +93,7 @@ impl<D: RepositoryEmbeddingsDB, M: EmbeddingsModel> Conversation<D, M> {
9393
#[allow(unused_labels)]
9494
'conversation: loop {
9595
//Generate a request with the message history and functions
96-
let request = generate_completion_request(self.messages.clone(), "auto");
96+
let request = generate_completion_request(self.messages.clone(), FunctionCallType::Auto);
9797

9898
match self.send_request(request) {
9999
Ok(response) => {
@@ -199,7 +199,7 @@ impl<D: RepositoryEmbeddingsDB, M: EmbeddingsModel> Conversation<D, M> {
199199
//Generate a request with the message history and no functions
200200
let request = generate_completion_request(
201201
self.messages.clone(),
202-
"none",
202+
FunctionCallType::None,
203203
);
204204
emit(&self.sender, QueryEvent::GenerateResponse(None))
205205
.await;
@@ -261,7 +261,7 @@ fn sanitize_query(query: &str) -> Result<String> {
261261
content: sanitize_query_prompt(query),
262262
};
263263
let client = Client::new(env::var("OPENAI_API_KEY")?);
264-
let request = generate_completion_request(vec![message], "none");
264+
let request = generate_completion_request(vec![message], FunctionCallType::None);
265265
let response = client.chat_completion(request)?;
266266
if let FinishReason::stop = response.choices[0].finish_reason {
267267
let sanitized_query = response.choices[0]

src/conversation/prompts.rs

+4-3
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
use openai_api_rs::v1::chat_completion::{
22
ChatCompletionMessage, ChatCompletionRequest, Function as F, FunctionParameters,
3-
JSONSchemaDefine, JSONSchemaType,
3+
JSONSchemaDefine, JSONSchemaType, FunctionCallType,
44
};
55
use std::collections::HashMap;
66

@@ -15,13 +15,14 @@ use crate::{
1515
// https://bloop.ai/
1616
pub fn generate_completion_request(
1717
messages: Vec<ChatCompletionMessage>,
18-
function_call: &str,
18+
function_call: FunctionCallType,
1919
) -> ChatCompletionRequest {
20+
2021
ChatCompletionRequest {
2122
model: CHAT_COMPLETION_MODEL.into(),
2223
messages,
2324
functions: Some(functions()),
24-
function_call: Some(function_call.into()),
25+
function_call: Some(function_call),
2526
temperature: Some(CHAT_COMPLETION_TEMPERATURE),
2627
top_p: None,
2728
n: None,

0 commit comments

Comments
 (0)