Skip to content

Commit e244751

Browse files
committed
[*] formatting fixes
1 parent de4d97c commit e244751

File tree

5 files changed

+11
-5
lines changed

5 files changed

+11
-5
lines changed

aisuite/async_client.py

+3-1
Original file line numberDiff line numberDiff line change
@@ -51,4 +51,6 @@ async def create(self, model: str, messages: list, **kwargs):
5151
raise ValueError(f"Could not load provider for '{provider_key}'.")
5252

5353
# Delegate the chat completion to the correct provider's async implementation
54-
return await provider.chat_completions_create_async(model_name, messages, **kwargs)
54+
return await provider.chat_completions_create_async(
55+
model_name, messages, **kwargs
56+
)

aisuite/provider.py

+1
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,7 @@ def chat_completions_create_async(self, model, messages):
2323
"""Abstract method for async chat completion calls, to be implemented by each provider."""
2424
pass
2525

26+
2627
class ProviderFactory:
2728
"""Factory to dynamically load provider instances based on naming conventions."""
2829

tests/client/test_client.py

+4-2
Original file line numberDiff line numberDiff line change
@@ -176,7 +176,7 @@ def test_invalid_model_format_in_create(monkeypatch):
176176
"aisuite.providers.fireworks_provider.FireworksProvider.chat_completions_create_async",
177177
"fireworks",
178178
"fireworks-model",
179-
)
179+
),
180180
],
181181
)
182182
@pytest.mark.asyncio
@@ -194,7 +194,9 @@ async def test_async_client_chat_completions(
194194
]
195195

196196
model_str = f"{provider}:{model}"
197-
model_response = await client.chat.completions.create(model_str, messages=messages)
197+
model_response = await client.chat.completions.create(
198+
model_str, messages=messages
199+
)
198200
assert model_response == expected_response
199201

200202

tests/client/test_prerelease.py

+2-1
Original file line numberDiff line numberDiff line change
@@ -37,9 +37,10 @@ def get_test_async_models() -> List[str]:
3737
return [
3838
"anthropic:claude-3-5-sonnet-20240620",
3939
"mistral:open-mistral-7b",
40-
"openai:gpt-3.5-turbo"
40+
"openai:gpt-3.5-turbo",
4141
]
4242

43+
4344
def get_test_messages() -> List[Dict[str, str]]:
4445
"""Return the test messages to send to each model."""
4546
return [

tests/providers/test_mistral_provider.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -74,4 +74,4 @@ async def test_mistral_provider_async():
7474
temperature=chosen_temperature,
7575
)
7676

77-
assert response.choices[0].message.content == response_text_content
77+
assert response.choices[0].message.content == response_text_content

0 commit comments

Comments
 (0)