diff --git a/README.md b/README.md index d1c0f19..422f5bc 100644 --- a/README.md +++ b/README.md @@ -79,13 +79,19 @@ import java.util.List; InferencePipelineDataStreamParams params = InferencePipelineDataStreamParams.builder() .inferencePipelineId("182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e") .config(InferencePipelineDataStreamParams.Config.ofLlmData(InferencePipelineDataStreamParams.Config.LlmData.builder() - .outputColumnName("output") - .costColumnName("cost") .inputVariableNames(List.of("user_query")) + .outputColumnName("output") .numOfTokenColumnName("tokens") + .costColumnName("cost") .timestampColumnName("timestamp") .build())) - .row(List.of(InferencePipelineDataStreamParams.Row.builder().build())) + .row(List.of(InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("what is the meaning of life?")) + .putAdditionalProperty("output", JsonValue.from("42")) + .putAdditionalProperty("tokens", JsonValue.from(7)) + .putAdditionalProperty("cost", JsonValue.from(0.02)) + .putAdditionalProperty("timestamp", JsonValue.from(1610000000)) + .build())) .build(); InferencePipelineDataStreamResponse response = client.inferencePipelines().data().stream(params); ``` diff --git a/openlayer-java-core/src/test/kotlin/com/openlayer/api/models/InferencePipelineDataStreamParamsTest.kt b/openlayer-java-core/src/test/kotlin/com/openlayer/api/models/InferencePipelineDataStreamParamsTest.kt index de558c5..5a260fa 100644 --- a/openlayer-java-core/src/test/kotlin/com/openlayer/api/models/InferencePipelineDataStreamParamsTest.kt +++ b/openlayer-java-core/src/test/kotlin/com/openlayer/api/models/InferencePipelineDataStreamParamsTest.kt @@ -21,10 +21,10 @@ class InferencePipelineDataStreamParamsTest { .costColumnName("cost") .groundTruthColumnName("ground_truth") .inferenceIdColumnName("id") - .inputVariableNames(listOf("string")) + .inputVariableNames(listOf("user_query")) .latencyColumnName("latency") .metadata(JsonValue.from(mapOf())) - .numOfTokenColumnName("num_tokens") + .numOfTokenColumnName("tokens") .prompt( listOf( InferencePipelineDataStreamParams.Config.LlmData.Prompt.builder() @@ -38,7 +38,17 @@ class InferencePipelineDataStreamParamsTest { .build() ) ) - .rows(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .rows( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) .build() } @@ -55,10 +65,10 @@ class InferencePipelineDataStreamParamsTest { .costColumnName("cost") .groundTruthColumnName("ground_truth") .inferenceIdColumnName("id") - .inputVariableNames(listOf("string")) + .inputVariableNames(listOf("user_query")) .latencyColumnName("latency") .metadata(JsonValue.from(mapOf())) - .numOfTokenColumnName("num_tokens") + .numOfTokenColumnName("tokens") .prompt( listOf( InferencePipelineDataStreamParams.Config.LlmData.Prompt @@ -73,7 +83,17 @@ class InferencePipelineDataStreamParamsTest { .build() ) ) - .rows(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .rows( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) .build() val body = params.getBody() assertThat(body).isNotNull @@ -86,10 +106,10 @@ class InferencePipelineDataStreamParamsTest { .costColumnName("cost") .groundTruthColumnName("ground_truth") .inferenceIdColumnName("id") - .inputVariableNames(listOf("string")) + .inputVariableNames(listOf("user_query")) .latencyColumnName("latency") .metadata(JsonValue.from(mapOf())) - .numOfTokenColumnName("num_tokens") + .numOfTokenColumnName("tokens") .prompt( listOf( InferencePipelineDataStreamParams.Config.LlmData.Prompt.builder() @@ -104,7 +124,17 @@ class InferencePipelineDataStreamParamsTest { ) ) assertThat(body.rows()) - .isEqualTo(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .isEqualTo( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) } @Test @@ -119,7 +149,17 @@ class InferencePipelineDataStreamParamsTest { .build() ) ) - .rows(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .rows( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) .build() val body = params.getBody() assertThat(body).isNotNull @@ -132,7 +172,17 @@ class InferencePipelineDataStreamParamsTest { ) ) assertThat(body.rows()) - .isEqualTo(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .isEqualTo( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) } @Test @@ -147,7 +197,17 @@ class InferencePipelineDataStreamParamsTest { .build() ) ) - .rows(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .rows( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) .build() assertThat(params).isNotNull // path param "inferencePipelineId" diff --git a/openlayer-java-core/src/test/kotlin/com/openlayer/api/models/ProjectCommitCreateParamsTest.kt b/openlayer-java-core/src/test/kotlin/com/openlayer/api/models/ProjectCommitCreateParamsTest.kt index 5eefc1e..288453b 100644 --- a/openlayer-java-core/src/test/kotlin/com/openlayer/api/models/ProjectCommitCreateParamsTest.kt +++ b/openlayer-java-core/src/test/kotlin/com/openlayer/api/models/ProjectCommitCreateParamsTest.kt @@ -93,8 +93,12 @@ class ProjectCommitCreateParamsTest { ProjectCommitCreateParams.Commit.builder() .id("3fa85f64-5717-4562-b3fc-2c963f66afa6") .authorId("589ece63-49a2-41b4-98e1-10547761d4b0") + .fileSize(1024L) .message("Updated the prompt.") + .mlModelId("3fa85f64-5717-4562-b3fc-2c963f66afa6") .storageUri("s3://...") + .trainingDatasetId("3fa85f64-5717-4562-b3fc-2c963f66afa6") + .validationDatasetId("3fa85f64-5717-4562-b3fc-2c963f66afa6") .build() ) .storageUri("s3://...") @@ -106,8 +110,12 @@ class ProjectCommitCreateParamsTest { ProjectCommitCreateParams.Commit.builder() .id("3fa85f64-5717-4562-b3fc-2c963f66afa6") .authorId("589ece63-49a2-41b4-98e1-10547761d4b0") + .fileSize(1024L) .message("Updated the prompt.") + .mlModelId("3fa85f64-5717-4562-b3fc-2c963f66afa6") .storageUri("s3://...") + .trainingDatasetId("3fa85f64-5717-4562-b3fc-2c963f66afa6") + .validationDatasetId("3fa85f64-5717-4562-b3fc-2c963f66afa6") .build() ) assertThat(body.storageUri()).isEqualTo("s3://...") @@ -122,8 +130,12 @@ class ProjectCommitCreateParamsTest { ProjectCommitCreateParams.Commit.builder() .id("3fa85f64-5717-4562-b3fc-2c963f66afa6") .authorId("589ece63-49a2-41b4-98e1-10547761d4b0") + .fileSize(1024L) .message("Updated the prompt.") + .mlModelId("3fa85f64-5717-4562-b3fc-2c963f66afa6") .storageUri("s3://...") + .trainingDatasetId("3fa85f64-5717-4562-b3fc-2c963f66afa6") + .validationDatasetId("3fa85f64-5717-4562-b3fc-2c963f66afa6") .build() ) .storageUri("s3://...") diff --git a/openlayer-java-core/src/test/kotlin/com/openlayer/api/services/ErrorHandlingTest.kt b/openlayer-java-core/src/test/kotlin/com/openlayer/api/services/ErrorHandlingTest.kt index f52754e..8a726e7 100644 --- a/openlayer-java-core/src/test/kotlin/com/openlayer/api/services/ErrorHandlingTest.kt +++ b/openlayer-java-core/src/test/kotlin/com/openlayer/api/services/ErrorHandlingTest.kt @@ -67,10 +67,10 @@ class ErrorHandlingTest { .costColumnName("cost") .groundTruthColumnName("ground_truth") .inferenceIdColumnName("id") - .inputVariableNames(listOf("string")) + .inputVariableNames(listOf("user_query")) .latencyColumnName("latency") .metadata(JsonValue.from(mapOf())) - .numOfTokenColumnName("num_tokens") + .numOfTokenColumnName("tokens") .prompt( listOf( InferencePipelineDataStreamParams.Config.LlmData.Prompt @@ -85,7 +85,17 @@ class ErrorHandlingTest { .build() ) ) - .rows(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .rows( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) .build() val expected = @@ -111,10 +121,10 @@ class ErrorHandlingTest { .costColumnName("cost") .groundTruthColumnName("ground_truth") .inferenceIdColumnName("id") - .inputVariableNames(listOf("string")) + .inputVariableNames(listOf("user_query")) .latencyColumnName("latency") .metadata(JsonValue.from(mapOf())) - .numOfTokenColumnName("num_tokens") + .numOfTokenColumnName("tokens") .prompt( listOf( InferencePipelineDataStreamParams.Config.LlmData.Prompt @@ -129,7 +139,17 @@ class ErrorHandlingTest { .build() ) ) - .rows(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .rows( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) .build() stubFor( @@ -156,10 +176,10 @@ class ErrorHandlingTest { .costColumnName("cost") .groundTruthColumnName("ground_truth") .inferenceIdColumnName("id") - .inputVariableNames(listOf("string")) + .inputVariableNames(listOf("user_query")) .latencyColumnName("latency") .metadata(JsonValue.from(mapOf())) - .numOfTokenColumnName("num_tokens") + .numOfTokenColumnName("tokens") .prompt( listOf( InferencePipelineDataStreamParams.Config.LlmData.Prompt @@ -174,7 +194,17 @@ class ErrorHandlingTest { .build() ) ) - .rows(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .rows( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) .build() stubFor( @@ -201,10 +231,10 @@ class ErrorHandlingTest { .costColumnName("cost") .groundTruthColumnName("ground_truth") .inferenceIdColumnName("id") - .inputVariableNames(listOf("string")) + .inputVariableNames(listOf("user_query")) .latencyColumnName("latency") .metadata(JsonValue.from(mapOf())) - .numOfTokenColumnName("num_tokens") + .numOfTokenColumnName("tokens") .prompt( listOf( InferencePipelineDataStreamParams.Config.LlmData.Prompt @@ -219,7 +249,17 @@ class ErrorHandlingTest { .build() ) ) - .rows(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .rows( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) .build() stubFor( @@ -250,10 +290,10 @@ class ErrorHandlingTest { .costColumnName("cost") .groundTruthColumnName("ground_truth") .inferenceIdColumnName("id") - .inputVariableNames(listOf("string")) + .inputVariableNames(listOf("user_query")) .latencyColumnName("latency") .metadata(JsonValue.from(mapOf())) - .numOfTokenColumnName("num_tokens") + .numOfTokenColumnName("tokens") .prompt( listOf( InferencePipelineDataStreamParams.Config.LlmData.Prompt @@ -268,7 +308,17 @@ class ErrorHandlingTest { .build() ) ) - .rows(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .rows( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) .build() stubFor( @@ -295,10 +345,10 @@ class ErrorHandlingTest { .costColumnName("cost") .groundTruthColumnName("ground_truth") .inferenceIdColumnName("id") - .inputVariableNames(listOf("string")) + .inputVariableNames(listOf("user_query")) .latencyColumnName("latency") .metadata(JsonValue.from(mapOf())) - .numOfTokenColumnName("num_tokens") + .numOfTokenColumnName("tokens") .prompt( listOf( InferencePipelineDataStreamParams.Config.LlmData.Prompt @@ -313,7 +363,17 @@ class ErrorHandlingTest { .build() ) ) - .rows(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .rows( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) .build() stubFor( @@ -344,10 +404,10 @@ class ErrorHandlingTest { .costColumnName("cost") .groundTruthColumnName("ground_truth") .inferenceIdColumnName("id") - .inputVariableNames(listOf("string")) + .inputVariableNames(listOf("user_query")) .latencyColumnName("latency") .metadata(JsonValue.from(mapOf())) - .numOfTokenColumnName("num_tokens") + .numOfTokenColumnName("tokens") .prompt( listOf( InferencePipelineDataStreamParams.Config.LlmData.Prompt @@ -362,7 +422,17 @@ class ErrorHandlingTest { .build() ) ) - .rows(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .rows( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) .build() stubFor( @@ -389,10 +459,10 @@ class ErrorHandlingTest { .costColumnName("cost") .groundTruthColumnName("ground_truth") .inferenceIdColumnName("id") - .inputVariableNames(listOf("string")) + .inputVariableNames(listOf("user_query")) .latencyColumnName("latency") .metadata(JsonValue.from(mapOf())) - .numOfTokenColumnName("num_tokens") + .numOfTokenColumnName("tokens") .prompt( listOf( InferencePipelineDataStreamParams.Config.LlmData.Prompt @@ -407,7 +477,17 @@ class ErrorHandlingTest { .build() ) ) - .rows(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .rows( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) .build() stubFor( @@ -438,10 +518,10 @@ class ErrorHandlingTest { .costColumnName("cost") .groundTruthColumnName("ground_truth") .inferenceIdColumnName("id") - .inputVariableNames(listOf("string")) + .inputVariableNames(listOf("user_query")) .latencyColumnName("latency") .metadata(JsonValue.from(mapOf())) - .numOfTokenColumnName("num_tokens") + .numOfTokenColumnName("tokens") .prompt( listOf( InferencePipelineDataStreamParams.Config.LlmData.Prompt @@ -456,7 +536,17 @@ class ErrorHandlingTest { .build() ) ) - .rows(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .rows( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) .build() stubFor( @@ -488,10 +578,10 @@ class ErrorHandlingTest { .costColumnName("cost") .groundTruthColumnName("ground_truth") .inferenceIdColumnName("id") - .inputVariableNames(listOf("string")) + .inputVariableNames(listOf("user_query")) .latencyColumnName("latency") .metadata(JsonValue.from(mapOf())) - .numOfTokenColumnName("num_tokens") + .numOfTokenColumnName("tokens") .prompt( listOf( InferencePipelineDataStreamParams.Config.LlmData.Prompt @@ -506,7 +596,17 @@ class ErrorHandlingTest { .build() ) ) - .rows(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .rows( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) .build() stubFor(post(anyUrl()).willReturn(status(200).withBody("Not JSON"))) @@ -532,10 +632,10 @@ class ErrorHandlingTest { .costColumnName("cost") .groundTruthColumnName("ground_truth") .inferenceIdColumnName("id") - .inputVariableNames(listOf("string")) + .inputVariableNames(listOf("user_query")) .latencyColumnName("latency") .metadata(JsonValue.from(mapOf())) - .numOfTokenColumnName("num_tokens") + .numOfTokenColumnName("tokens") .prompt( listOf( InferencePipelineDataStreamParams.Config.LlmData.Prompt @@ -550,7 +650,17 @@ class ErrorHandlingTest { .build() ) ) - .rows(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .rows( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) .build() stubFor(post(anyUrl()).willReturn(status(400).withBody("Not JSON"))) diff --git a/openlayer-java-core/src/test/kotlin/com/openlayer/api/services/ServiceParamsTest.kt b/openlayer-java-core/src/test/kotlin/com/openlayer/api/services/ServiceParamsTest.kt index 4378175..6c1529a 100644 --- a/openlayer-java-core/src/test/kotlin/com/openlayer/api/services/ServiceParamsTest.kt +++ b/openlayer-java-core/src/test/kotlin/com/openlayer/api/services/ServiceParamsTest.kt @@ -65,10 +65,10 @@ class ServiceParamsTest { .costColumnName("cost") .groundTruthColumnName("ground_truth") .inferenceIdColumnName("id") - .inputVariableNames(listOf("string")) + .inputVariableNames(listOf("user_query")) .latencyColumnName("latency") .metadata(JsonValue.from(mapOf())) - .numOfTokenColumnName("num_tokens") + .numOfTokenColumnName("tokens") .prompt( listOf( InferencePipelineDataStreamParams.Config.LlmData.Prompt @@ -83,7 +83,17 @@ class ServiceParamsTest { .build() ) ) - .rows(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .rows( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) .additionalHeaders(additionalHeaders) .additionalBodyProperties(additionalBodyProperties) .additionalQueryParams(additionalQueryParams) diff --git a/openlayer-java-core/src/test/kotlin/com/openlayer/api/services/blocking/inferencePipelines/DataServiceTest.kt b/openlayer-java-core/src/test/kotlin/com/openlayer/api/services/blocking/inferencePipelines/DataServiceTest.kt index 0b70b54..b7c8d52 100644 --- a/openlayer-java-core/src/test/kotlin/com/openlayer/api/services/blocking/inferencePipelines/DataServiceTest.kt +++ b/openlayer-java-core/src/test/kotlin/com/openlayer/api/services/blocking/inferencePipelines/DataServiceTest.kt @@ -32,10 +32,10 @@ class DataServiceTest { .costColumnName("cost") .groundTruthColumnName("ground_truth") .inferenceIdColumnName("id") - .inputVariableNames(listOf("string")) + .inputVariableNames(listOf("user_query")) .latencyColumnName("latency") .metadata(JsonValue.from(mapOf())) - .numOfTokenColumnName("num_tokens") + .numOfTokenColumnName("tokens") .prompt( listOf( InferencePipelineDataStreamParams.Config.LlmData.Prompt @@ -50,7 +50,17 @@ class DataServiceTest { .build() ) ) - .rows(listOf(InferencePipelineDataStreamParams.Row.builder().build())) + .rows( + listOf( + InferencePipelineDataStreamParams.Row.builder() + .putAdditionalProperty("user_query", JsonValue.from("bar")) + .putAdditionalProperty("output", JsonValue.from("bar")) + .putAdditionalProperty("tokens", JsonValue.from("bar")) + .putAdditionalProperty("cost", JsonValue.from("bar")) + .putAdditionalProperty("timestamp", JsonValue.from("bar")) + .build() + ) + ) .build() ) println(inferencePipelineDataStreamResponse)