|
39 | 39 | providers, |
40 | 40 | telemetry, |
41 | 41 | vector_io, |
| 42 | + benchmarks, |
42 | 43 | eval_tasks, |
43 | 44 | toolgroups, |
44 | 45 | vector_dbs, |
@@ -94,6 +95,7 @@ class LlamaStackClient(SyncAPIClient): |
94 | 95 | scoring: scoring.ScoringResource |
95 | 96 | scoring_functions: scoring_functions.ScoringFunctionsResource |
96 | 97 | eval_tasks: eval_tasks.EvalTasksResource |
| 98 | + benchmarks: benchmarks.BenchmarksResource |
97 | 99 | with_raw_response: LlamaStackClientWithRawResponse |
98 | 100 | with_streaming_response: LlamaStackClientWithStreamedResponse |
99 | 101 |
|
@@ -176,6 +178,7 @@ def __init__( |
176 | 178 | self.scoring = scoring.ScoringResource(self) |
177 | 179 | self.scoring_functions = scoring_functions.ScoringFunctionsResource(self) |
178 | 180 | self.eval_tasks = eval_tasks.EvalTasksResource(self) |
| 181 | + self.benchmarks = benchmarks.BenchmarksResource(self) |
179 | 182 | self.with_raw_response = LlamaStackClientWithRawResponse(self) |
180 | 183 | self.with_streaming_response = LlamaStackClientWithStreamedResponse(self) |
181 | 184 |
|
@@ -310,6 +313,7 @@ class AsyncLlamaStackClient(AsyncAPIClient): |
310 | 313 | scoring: scoring.AsyncScoringResource |
311 | 314 | scoring_functions: scoring_functions.AsyncScoringFunctionsResource |
312 | 315 | eval_tasks: eval_tasks.AsyncEvalTasksResource |
| 316 | + benchmarks: benchmarks.AsyncBenchmarksResource |
313 | 317 | with_raw_response: AsyncLlamaStackClientWithRawResponse |
314 | 318 | with_streaming_response: AsyncLlamaStackClientWithStreamedResponse |
315 | 319 |
|
@@ -392,6 +396,7 @@ def __init__( |
392 | 396 | self.scoring = scoring.AsyncScoringResource(self) |
393 | 397 | self.scoring_functions = scoring_functions.AsyncScoringFunctionsResource(self) |
394 | 398 | self.eval_tasks = eval_tasks.AsyncEvalTasksResource(self) |
| 399 | + self.benchmarks = benchmarks.AsyncBenchmarksResource(self) |
395 | 400 | self.with_raw_response = AsyncLlamaStackClientWithRawResponse(self) |
396 | 401 | self.with_streaming_response = AsyncLlamaStackClientWithStreamedResponse(self) |
397 | 402 |
|
@@ -529,6 +534,7 @@ def __init__(self, client: LlamaStackClient) -> None: |
529 | 534 | self.scoring = scoring.ScoringResourceWithRawResponse(client.scoring) |
530 | 535 | self.scoring_functions = scoring_functions.ScoringFunctionsResourceWithRawResponse(client.scoring_functions) |
531 | 536 | self.eval_tasks = eval_tasks.EvalTasksResourceWithRawResponse(client.eval_tasks) |
| 537 | + self.benchmarks = benchmarks.BenchmarksResourceWithRawResponse(client.benchmarks) |
532 | 538 |
|
533 | 539 |
|
534 | 540 | class AsyncLlamaStackClientWithRawResponse: |
@@ -560,6 +566,7 @@ def __init__(self, client: AsyncLlamaStackClient) -> None: |
560 | 566 | client.scoring_functions |
561 | 567 | ) |
562 | 568 | self.eval_tasks = eval_tasks.AsyncEvalTasksResourceWithRawResponse(client.eval_tasks) |
| 569 | + self.benchmarks = benchmarks.AsyncBenchmarksResourceWithRawResponse(client.benchmarks) |
563 | 570 |
|
564 | 571 |
|
565 | 572 | class LlamaStackClientWithStreamedResponse: |
@@ -591,6 +598,7 @@ def __init__(self, client: LlamaStackClient) -> None: |
591 | 598 | client.scoring_functions |
592 | 599 | ) |
593 | 600 | self.eval_tasks = eval_tasks.EvalTasksResourceWithStreamingResponse(client.eval_tasks) |
| 601 | + self.benchmarks = benchmarks.BenchmarksResourceWithStreamingResponse(client.benchmarks) |
594 | 602 |
|
595 | 603 |
|
596 | 604 | class AsyncLlamaStackClientWithStreamedResponse: |
@@ -624,6 +632,7 @@ def __init__(self, client: AsyncLlamaStackClient) -> None: |
624 | 632 | client.scoring_functions |
625 | 633 | ) |
626 | 634 | self.eval_tasks = eval_tasks.AsyncEvalTasksResourceWithStreamingResponse(client.eval_tasks) |
| 635 | + self.benchmarks = benchmarks.AsyncBenchmarksResourceWithStreamingResponse(client.benchmarks) |
627 | 636 |
|
628 | 637 |
|
629 | 638 | Client = LlamaStackClient |
|
0 commit comments