|
39 | 39 | providers,
|
40 | 40 | telemetry,
|
41 | 41 | vector_io,
|
| 42 | + benchmarks, |
42 | 43 | eval_tasks,
|
43 | 44 | toolgroups,
|
44 | 45 | vector_dbs,
|
@@ -94,6 +95,7 @@ class LlamaStackClient(SyncAPIClient):
|
94 | 95 | scoring: scoring.ScoringResource
|
95 | 96 | scoring_functions: scoring_functions.ScoringFunctionsResource
|
96 | 97 | eval_tasks: eval_tasks.EvalTasksResource
|
| 98 | + benchmarks: benchmarks.BenchmarksResource |
97 | 99 | with_raw_response: LlamaStackClientWithRawResponse
|
98 | 100 | with_streaming_response: LlamaStackClientWithStreamedResponse
|
99 | 101 |
|
@@ -176,6 +178,7 @@ def __init__(
|
176 | 178 | self.scoring = scoring.ScoringResource(self)
|
177 | 179 | self.scoring_functions = scoring_functions.ScoringFunctionsResource(self)
|
178 | 180 | self.eval_tasks = eval_tasks.EvalTasksResource(self)
|
| 181 | + self.benchmarks = benchmarks.BenchmarksResource(self) |
179 | 182 | self.with_raw_response = LlamaStackClientWithRawResponse(self)
|
180 | 183 | self.with_streaming_response = LlamaStackClientWithStreamedResponse(self)
|
181 | 184 |
|
@@ -310,6 +313,7 @@ class AsyncLlamaStackClient(AsyncAPIClient):
|
310 | 313 | scoring: scoring.AsyncScoringResource
|
311 | 314 | scoring_functions: scoring_functions.AsyncScoringFunctionsResource
|
312 | 315 | eval_tasks: eval_tasks.AsyncEvalTasksResource
|
| 316 | + benchmarks: benchmarks.AsyncBenchmarksResource |
313 | 317 | with_raw_response: AsyncLlamaStackClientWithRawResponse
|
314 | 318 | with_streaming_response: AsyncLlamaStackClientWithStreamedResponse
|
315 | 319 |
|
@@ -392,6 +396,7 @@ def __init__(
|
392 | 396 | self.scoring = scoring.AsyncScoringResource(self)
|
393 | 397 | self.scoring_functions = scoring_functions.AsyncScoringFunctionsResource(self)
|
394 | 398 | self.eval_tasks = eval_tasks.AsyncEvalTasksResource(self)
|
| 399 | + self.benchmarks = benchmarks.AsyncBenchmarksResource(self) |
395 | 400 | self.with_raw_response = AsyncLlamaStackClientWithRawResponse(self)
|
396 | 401 | self.with_streaming_response = AsyncLlamaStackClientWithStreamedResponse(self)
|
397 | 402 |
|
@@ -529,6 +534,7 @@ def __init__(self, client: LlamaStackClient) -> None:
|
529 | 534 | self.scoring = scoring.ScoringResourceWithRawResponse(client.scoring)
|
530 | 535 | self.scoring_functions = scoring_functions.ScoringFunctionsResourceWithRawResponse(client.scoring_functions)
|
531 | 536 | self.eval_tasks = eval_tasks.EvalTasksResourceWithRawResponse(client.eval_tasks)
|
| 537 | + self.benchmarks = benchmarks.BenchmarksResourceWithRawResponse(client.benchmarks) |
532 | 538 |
|
533 | 539 |
|
534 | 540 | class AsyncLlamaStackClientWithRawResponse:
|
@@ -560,6 +566,7 @@ def __init__(self, client: AsyncLlamaStackClient) -> None:
|
560 | 566 | client.scoring_functions
|
561 | 567 | )
|
562 | 568 | self.eval_tasks = eval_tasks.AsyncEvalTasksResourceWithRawResponse(client.eval_tasks)
|
| 569 | + self.benchmarks = benchmarks.AsyncBenchmarksResourceWithRawResponse(client.benchmarks) |
563 | 570 |
|
564 | 571 |
|
565 | 572 | class LlamaStackClientWithStreamedResponse:
|
@@ -591,6 +598,7 @@ def __init__(self, client: LlamaStackClient) -> None:
|
591 | 598 | client.scoring_functions
|
592 | 599 | )
|
593 | 600 | self.eval_tasks = eval_tasks.EvalTasksResourceWithStreamingResponse(client.eval_tasks)
|
| 601 | + self.benchmarks = benchmarks.BenchmarksResourceWithStreamingResponse(client.benchmarks) |
594 | 602 |
|
595 | 603 |
|
596 | 604 | class AsyncLlamaStackClientWithStreamedResponse:
|
@@ -624,6 +632,7 @@ def __init__(self, client: AsyncLlamaStackClient) -> None:
|
624 | 632 | client.scoring_functions
|
625 | 633 | )
|
626 | 634 | self.eval_tasks = eval_tasks.AsyncEvalTasksResourceWithStreamingResponse(client.eval_tasks)
|
| 635 | + self.benchmarks = benchmarks.AsyncBenchmarksResourceWithStreamingResponse(client.benchmarks) |
627 | 636 |
|
628 | 637 |
|
629 | 638 | Client = LlamaStackClient
|
|
0 commit comments