diff --git a/docs/destinations/forwarding-events/snowbridge/3-X-X-upgrade-guide/index.md b/docs/destinations/forwarding-events/snowbridge/3-X-X-upgrade-guide/index.md new file mode 100644 index 0000000000..312f838b12 --- /dev/null +++ b/docs/destinations/forwarding-events/snowbridge/3-X-X-upgrade-guide/index.md @@ -0,0 +1,42 @@ +--- +title: "3.X.X upgrade guide" +sidebar_position: 400 +--- + +## Breaking Changes + +The below breaking changes were made in version 3.0.0. All other functionality is backwards compatible. + +### Lua support removed + +Support for Lua transformations has been removed. If you are running a Lua transformation, you can port the logic to [Javascript](/docs/destinations/forwarding-events/snowbridge/configuration/transformations/custom-scripts/javascript-configuration/index.md) or [JQ](/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/jq.md). + +## HTTP target: non-JSON data no longer supported + +We never intended to support non-JSON data, but prior to version 3.0.0, the request body was simply populated with whatever bytes were found in the message data, regardless of whether it is valid JSON. + +From version 3.0.0 onwards, only valid JSON will work, otherwise the message will be considered invalid and sent to the failure target. + +## HTTP target: request batching + +Many HTTP APIs allow sending several events in a single request by putting them into a JSON array. Since version 3.0.0, if the Snowbridge source provides data in batches, the HTTP target will batch events in this way. + +As a consequence, even when the source provides events in a single event batch, it will now be placed into an array of one element. For example, prior to version 3.0.0, a request body might look like this: + +``` +{"foo": "bar"} +``` + +But it will now look like this: + +``` +[{"foo": "bar"}] +``` + +As of version 3.0.0, the SQS source provides events in batches of up to ten, and the Kinesis, Kafka, and Pubsub and Stdin sources provide events in single-event batches. This behavior will likely change in a future version. + +You can preserve the previous behavior and ensure that requests are always single-event non-array objects, even with a batching source. To do so, set `request_max_messages` to 1, and provide this template (as long as your data is valid JSON): + +```go reference +https://github.com/snowplow/snowbridge/blob/master/assets/docs/configuration/targets/http-template-unwrap-example.file +``` diff --git a/docs/destinations/forwarding-events/snowbridge/concepts/failure-model/index.md b/docs/destinations/forwarding-events/snowbridge/concepts/failure-model/index.md index 7796312ccf..e83b72ecd3 100644 --- a/docs/destinations/forwarding-events/snowbridge/concepts/failure-model/index.md +++ b/docs/destinations/forwarding-events/snowbridge/concepts/failure-model/index.md @@ -18,11 +18,15 @@ There are several different failures that Snowbridge may hit. ### Target failure -This is where a request to the destination technology fails or is rejected - for example a http 400 response is received. When Snowbridge hits this failure, it will retry 5 times. If all 5 attempts fail, it will be reported as a 'MsgFailed' for monitoring purposes, and will proceed without acking the failed Messages. As long as the source's acking model allows for it, these will be re-processed through Snowbridge again. +This is where a request to the destination technology fails or is rejected - for example a HTTP 400 response is received. -Note that this means failures on the receiving end (eg. if an endpoint is unavailable), mean Snowbridge will continue to attempt to process the data until the issue is fixed. +Retry behavior for target failures is determined by the retry configuration. You can find details of this in the [configuration section](/docs/destinations/forwarding-events/snowbridge/configuration/retries/index.md). -As of Snowbridge 2.4.2, the kinesis target does not treat kinesis write throughput exceptions as this type of failure. Rather it has an in-built backoff and retry, which will persist until each event in the batch is either successful, or fails for a different reason. +As of Snowbridge 2.4.2, the Kinesis target does not treat kinesis write throughput exceptions as this type of failure. Rather it has an in-built backoff and retry, which will persist until each event in the batch is either successful, or fails for a different reason. + +Before version 3.0.0, Snowbridge treated every kind of target failure the same - it would retry 5 times. If all 5 attempts failed, it would proceed without acking the failed Messages. As long as the source's acking model allows for it, these would be re-processed through Snowbridge again. + +Each target failure attempt will be reported as a 'MsgFailed' for monitoring purposes. ### Oversized data @@ -34,6 +38,8 @@ Writes of oversized messages to the failure target will be recorded with 'Oversi In the unlikely event that Snowbridge encounters data which is invalid for the target destination (for example empty data is invalid for pubsub), it will create a [generic error failed event](/docs/fundamentals/failed-events/index.md#generic-error), emit it to the failure target, and ack the original message. +As of version 3.0.0, the HTTP target may produce 'invalid' type failures. This occurs when: the a POST request body cannot be formed; the templating feature's attempts to template data result in an error; or the response conforms to a response rules configuration which specifies that the failure is to be treated as invalid. You can find more details in the [configuration section](/docs/destinations/forwarding-events/snowbridge/configuration/targets/http/index.md). + Transformation failures are also treated as invalid, as described below. Writes of invalid messages to the failure target will be recorded with 'InvalidMsg' statistics in monitoring. Any failure to write to the failure target will cause a [fatal failure](#fatal-failure). @@ -54,6 +60,6 @@ Firstly, if it hits an error in retrieving data from the source stream, it will Secondly, as described above, where there are failures it will attempt to reprocess the data if it can, and where failures aren't recoverable it will attempt to handle that via a failure target. Normally, even reaching this point is rare. -In the very unlikely event that Snowbridge reaches this point and cannot write to a failure target, the app will crash. Should this happen, and the app is re-deployed, it will begin processing data from the last acked message. Note that the likely impact of this is duplicated sends to the target, but not data loss. +In the very unlikely event that Snowbridge reaches this point and cannot write to a failure target, the app will crash. Should this happen, and the app is re-deployed, it will begin processing data from the last acked message. Note that the likely impact of this is duplicated sends to the target, but not data loss. -Of course, if you experience crashes or other issues that are not explained by the above, please log an issue detailing the bahaviour. \ No newline at end of file +Of course, if you experience crashes or other issues that are not explained by the above, please log an issue detailing the behavior. diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/monitoring/index.md b/docs/destinations/forwarding-events/snowbridge/configuration/monitoring/index.md index 8e5e1c13ad..fd10242515 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/monitoring/index.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/monitoring/index.md @@ -23,7 +23,8 @@ https://github.com/snowplow/snowbridge/blob/master/assets/docs/configuration/mon ```hcl reference https://github.com/snowplow/snowbridge/blob/master/assets/docs/configuration/monitoring/sentry-example.hcl ``` -### StatsD stats receiver + +### StatsD stats receiver configuration ```hcl reference https://github.com/snowplow/snowbridge/blob/master/assets/docs/configuration/monitoring/statsd-example.hcl @@ -33,10 +34,10 @@ Snowbridge sends the following metrics to statsd: | Metric | Definitions | |--------------------------|---------------------------------------------------------------------------------------------------------------------------------------------------------| -| `target_success` | Events successfully sent to the target. | -| `target_failed` | Events which failed to reach the target, after 5 retries. Will be retried later. | +| `target_success` | Events successfully sent to the target. | +| `target_failed` | Events which failed to reach the target, and will be handled by the retry config. Retries which fail are also counted. | | `message_filtered` | Events filtered out via transformation. | -| `failure_target_success` | Events we could not send to the target, which are not retryable, successfully sent to the failure target. | +| `failure_target_success` | Events we could not send to the target, which are not retryable, successfully sent to the failure target. | | `failure_target_failed` | Events we could not send to the target, which are not retryable, which we failed to send to the failure target. In this scenario, Snowbridge will crash. | | `min_processing_latency` | Min time between entering Snowbridge and write to target/failure target. | | `max_processing_latency` | Max time between entering Snowbridge and write to target/failure target. | diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/retries/index.md b/docs/destinations/forwarding-events/snowbridge/configuration/retries/index.md new file mode 100644 index 0000000000..88f2ca7d14 --- /dev/null +++ b/docs/destinations/forwarding-events/snowbridge/configuration/retries/index.md @@ -0,0 +1,26 @@ +--- +title: "Retry behavior (beta)" +description: "Configure retry behaviour." +--- + +:::note +This feature was added in version 3.0.0 + +This feature is in beta status because we may make breaking changes in future versions. +::: + +This feature allows you to configure the retry behavior when the target encounters a failure in sending the data. There are two types of failure you can define: + +A **transient failure** is a failure which we expect to succeed again on retry. For example, some temporary network error, or when we encounter throttling. Typically, you would configure a short backoff for this type of failure. When we encounter a transient failure, we keep processing the rest of the data as normal, under the expectation that everything is operating as normal. The failed data is retried after a backoff. + +A **setup failure** is one we don't expect to be immediately resolved, for example an incorrect address, or an invalid API Key. Typically, you would configure a long backoff for this type of failure, under the assumption that the issue needs to be fixed with either a configuration change or a change to the target itself (e.g. permissions need to be granted). Setup errors will be retried 5 times before the app crashes. + +As of version 3.0.0, only the http target can be configured to return setup errors, via the response rules feature - see [the http target configuration section](/docs/destinations/forwarding-events/snowbridge/configuration/targets/http/index.md). For all other targets, all errors returned will be considered transient, and behavior can be configured using the `transient` block of the retry configuration. + +Retries will be attempted with an exponential backoff. In other words, on each subsequent failure, the backoff time will double. You can configure transient failures to be retried indefinitely by setting `max_attempts` to 0. + +## Configuration options + +```hcl reference +https://github.com/snowplow/snowbridge/blob/master/assets/docs/configuration/retry-example.hcl +``` diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/sources/kafka.md b/docs/destinations/forwarding-events/snowbridge/configuration/sources/kafka.md index acec615ee6..9a3c9f3614 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/sources/kafka.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/sources/kafka.md @@ -1,5 +1,5 @@ --- -title: "Kafka Source" +title: "Kafka" description: "Read data from a Kafka topic." --- diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/sources/kinesis.md b/docs/destinations/forwarding-events/snowbridge/configuration/sources/kinesis.md index cbd496105b..e59e80fcf3 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/sources/kinesis.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/sources/kinesis.md @@ -1,5 +1,5 @@ --- -title: "Kinesis Source" +title: "Kinesis" description: "Read data from a Kinesis stream." --- diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/sources/pubsub.md b/docs/destinations/forwarding-events/snowbridge/configuration/sources/pubsub.md index b70d744a68..fa4fa9295e 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/sources/pubsub.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/sources/pubsub.md @@ -1,5 +1,5 @@ --- -title: "PubSub Source" +title: "PubSub" description: "Read data from a PubSub topic." --- diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/sources/sqs.md b/docs/destinations/forwarding-events/snowbridge/configuration/sources/sqs.md index e4a4c9060d..816c9ac8f9 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/sources/sqs.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/sources/sqs.md @@ -1,5 +1,5 @@ --- -title: "SQS Source" +title: "SQS" description: "Read data from an SQS queue." --- diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/sources/stdin.md b/docs/destinations/forwarding-events/snowbridge/configuration/sources/stdin.md index bbb99aa12d..b9961ab34a 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/sources/stdin.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/sources/stdin.md @@ -1,13 +1,14 @@ --- -title: "Stdin Source" +title: "Stdin" description: "Read data from stdin." +sidebar_position: 1 --- Stdin source is the default, and has one optional configuration to set the concurrency. Stdin source simply treats stdin as the input. -## Configuration +## Configuration options Here is an example of the minimum required configuration: diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/targets/eventhub.md b/docs/destinations/forwarding-events/snowbridge/configuration/targets/eventhub.md index e535124c2d..cbc747c324 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/targets/eventhub.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/targets/eventhub.md @@ -1,5 +1,5 @@ --- -title: "EventHub Target" +title: "EventHub" description: "Write data to an EventHub." --- diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/targets/http/index.md b/docs/destinations/forwarding-events/snowbridge/configuration/targets/http/index.md index 033fec3676..08e8983406 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/targets/http/index.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/targets/http/index.md @@ -1,22 +1,84 @@ --- -title: "HTTP Target" +title: "HTTP" description: "Send data over HTTP." --- +:::note +Version 3.0.0 makes breaking changes to the http target. Details on migrating can be found [in the migration guide](/docs/destinations/forwarding-events/snowbridge/3-X-X-upgrade-guide/index.md) +::: + ## Basic Authentication -Where basicauth is used, it may be configured using the `basic_auth_username` and `basic_auth_password` options. Where an authorisation header is used, it may be set via the `headers` option. +Where basicauth is used, it may be configured using the `basic_auth_username` and `basic_auth_password` options. Where an authorization header is used, it may be set via the `headers` option. -we recommend using environment variables for sensitive values - which can be done via HCL's native `env.MY_ENV_VAR` format (as seen below). +We recommend using environment variables for sensitive values - which can be done via HCL's native `env.MY_ENV_VAR` format (as seen below). TLS may be configured by providing the `key_file`, `cert_file` and `ca_file` options with paths to the relevant TLS files. ## OAuth2 -Snowbridge supports sending authorized requests to OAuth2 - compliant HTTP targets. This can be enabled in by setting `oauth2_client_id `, `oauth2_client_secret`, `oauth2_refresh_token` (these 3 are long-lived credentials used to generate short-lived bearer access tokens) and `oauth2_token_url`(which is the URL of authorization server providing access tokens). +Snowbridge supports sending authorized requests to OAuth2-compliant HTTP targets. This can be enabled by setting `oauth2_client_id`, `oauth2_client_secret`, `oauth2_refresh_token` (these three are long-lived credentials used to generate short-lived bearer access tokens), and `oauth2_token_url` (which is the URL of the authorization server providing access tokens). Like in the case of basic authentication, we recommend using environment variables for sensitive values. +## Dynamic headers + +:::note +This feature was added in version 2.3.0 +::: + +When enabled, this feature attaches a header to the data according to what your transformation provides in the `HTTPHeaders` field of `engineProtocol`. Data is batched independently per each dynamic header value before requests are sent. + +## Request templating + +:::note +This feature was added in version 3.0.0 +::: + +This feature allows you to provide a [Golang text template](https://pkg.go.dev/text/template) to construct a request body from a batch of data. This feature should be useful in constructing requests to send to an API, for example. + +Input data must be valid JSON, any message that fails to be marshaled to JSON will be treated as invalid and sent to the failure target. Equally, if an attempt to template a batch of data results in an error, then all messages in the batch will be considered invalid and sent to the failure target. + +Where the dynamic headers feature is enabled, data is split into batches according to the provided header value, and the templater will operate on each batch separately. + +### Helper functions + +In addition to all base functions available in the Go text/template package, the following custom functions are available for convenience: + +`prettyPrint` - Because the input to the templater is a Go data structure, simply providing a reference to an object field won't output it as JSON. `prettyPrint` converts the data to prettified JSON. Use it wherever you expect a JSON object in the output. This is compatible with any data type, but it shouldn't be necessary if the data is not an object. + +`env` - Allows you to set and refer to an env var in your template. Use it when your request body must contain sensitive data, for example an API key. + +### Template example + +The following example provides an API key via environment variable, and iterates the batch to provide JSON-formatted data one by one into a new key, inserting a comma before all but the first event. + +```hcl reference +https://github.com/snowplow/snowbridge/blob/master/assets/docs/configuration/targets/http-template-full-example.file +``` + +## Response rules (beta) + +:::note +This feature was added in version 3.0.0 + +This feature is in beta status because we may make breaking changes in future versions. +::: + +Response rules allow you to configure how the app deals with failures in sending the data. You can configure a response code and an optional string match on the response body to determine how a failure response is handled. Response codes between 200 and 299 are considered successful, and are not handled by this feature. + +There are three categories of failure: + +`invalid` means that the data is considered incompatible with the target for some reason. For example, you may have defined a mapping for a given API, but the event happens to have null data for a required field. In this instance, retrying the data won't fix the issue, so you would configure an invalid response rule, identifying which responses indicate this scenario. + +Data that matches an invalid response rule is sent to the failure target. + +`setup` means that this error is not retryable, but is something which can only be resolved by a change in configuration or a change to the target. An example of this is an authentication failure - retrying will fix the issue, the resolution is to grant the appropriate permissions, or provide the correct API key. + +Data that matches a setup response rule is handled by a retry as determined in the `setup` configuration block of [retry configuration](/docs/destinations/forwarding-events/snowbridge/configuration/retries/index.md). + +`transient` errors are everything else - we assume that the issue is temporary and retrying will resolve the problem. An example of this is being throttled by an API because too much data is being sent at once. There is no explicit configuration for transient - rather, anything that is not configured as one of the other types is considered transient. + ## Configuration options Here is an example of the minimum required configuration: diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/targets/kafka.md b/docs/destinations/forwarding-events/snowbridge/configuration/targets/kafka.md index d9ca318f0e..604cfc3ebe 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/targets/kafka.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/targets/kafka.md @@ -1,5 +1,5 @@ --- -title: "Kafka Target" +title: "Kafka" description: "Write data to a Kafka topic." --- diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/targets/kinesis.md b/docs/destinations/forwarding-events/snowbridge/configuration/targets/kinesis.md index 6e6f4c7079..7a1b18b304 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/targets/kinesis.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/targets/kinesis.md @@ -1,5 +1,5 @@ --- -title: "Kinesis Target" +title: "Kinesis" description: "Write data to a Kinesis stream." --- diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/targets/pubsub.md b/docs/destinations/forwarding-events/snowbridge/configuration/targets/pubsub.md index e002572f23..e50ab5302d 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/targets/pubsub.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/targets/pubsub.md @@ -1,5 +1,5 @@ --- -title: "PubSub Target" +title: "PubSub" description: "Write data to a Pubsub topic." --- diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/targets/sqs.md b/docs/destinations/forwarding-events/snowbridge/configuration/targets/sqs.md index 8e90020d57..ae45bbe7f3 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/targets/sqs.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/targets/sqs.md @@ -1,5 +1,5 @@ --- -title: "SQS Target" +title: "SQS" description: "Write data to an SQS queue." --- diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/targets/stdout.md b/docs/destinations/forwarding-events/snowbridge/configuration/targets/stdout.md index ebb07189e4..853526f21a 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/targets/stdout.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/targets/stdout.md @@ -1,10 +1,12 @@ --- -title: "Stdout Target" +title: "Stdout" description: "Write data to an SQS queue." +sidebar_position: 1 --- Stdout target doesn't have any configurable options - when configured it simply outputs the messages to stdout. -## Configuration + +## Configuration options Here is an example of the configuration: diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/telemetry/index.md b/docs/destinations/forwarding-events/snowbridge/configuration/telemetry/index.md index 93a34ca749..20eaa80c13 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/telemetry/index.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/telemetry/index.md @@ -8,7 +8,7 @@ sidebar_position: 500 You can read about our telemetry principles [here](/docs/get-started/snowplow-community-edition/telemetry/index.md). -## Configuration via file: +## Configuration options Enabling telemetry: diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/base64Decode.md b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/base64Decode.md index 19c2f4d3ad..0dc2fd17c1 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/base64Decode.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/base64Decode.md @@ -8,7 +8,7 @@ This transformation base64 decodes the message's data from a base64 byte array, `base64Decode` has no options. -Example: +## Configuration options ```hcl reference https://github.com/snowplow/snowbridge/blob/master/assets/docs/configuration/transformations/builtin/base64Decode-minimal-example.hcl diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/base64Encode.md b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/base64Encode.md index e7f3306064..d8923bed76 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/base64Encode.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/base64Encode.md @@ -8,7 +8,7 @@ This transformation base64 encodes the message's data to a base 64 byte array. `base64Encode` has no options. -Example: +## Configuration options ```hcl reference https://github.com/snowplow/snowbridge/blob/master/assets/docs/configuration/transformations/builtin/base64Encode-minimal-example.hcl diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/jq.md b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/jq.md new file mode 100644 index 0000000000..27c55e5141 --- /dev/null +++ b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/jq.md @@ -0,0 +1,41 @@ +# jq + +:::note +This transformation was added in version 3.0.0. +::: + +```mdx-code-block +import JQDescriptionSharedBlock from "./reusable/_jqDescription.md" + + +``` + +`jq` runs a jq command on the message data, and outputs the result of the command. While jq supports multi-element results, commands must output only a single element - this single element can be an array data type. + +If the provided jq command results in an error, the message will be considered invalid, and will be sent to the failure target. + +The minimal example here returns the input data as a single element array, and the full example maps the data to a new data structure. + +The jq transformation will remove any keys with null values from the data. + +## Configuration options + +Minimal configuration: + +```hcl reference +https://github.com/snowplow/snowbridge/blob/master/assets/docs/configuration/transformations/builtin/jq-minimal-example.hcl +``` + +Every configuration option: + +```hcl reference +https://github.com/snowplow/snowbridge/blob/master/assets/docs/configuration/transformations/builtin/jq-full-example.hcl +``` + +## Helper functions + +```mdx-code-block +import JQHelpersSharedBlock from "./reusable/_jqHelpers.md" + + +``` diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/jqFilter.md b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/jqFilter.md new file mode 100644 index 0000000000..3510bc02cd --- /dev/null +++ b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/jqFilter.md @@ -0,0 +1,39 @@ +# jqFilter + +:::note +This transformation was added in version 3.0.0. +::: + +```mdx-code-block +import JQDescriptionSharedBlock from "./reusable/_jqDescription.md" + + +``` + +`jqFilter` filters messages based on the output of a jq command which is run against the data. The provided command must return a boolean result. `false` filters the message out, `true` keeps it. + +If the provided jq command returns a non-boolean value error, or results in an error, then the message will be considered invalid, and will be sent to the failure target. + +## Configuration options + +This example filters out all data that doesn't have an `app_id` key. + +Minimal configuration: + +```hcl reference +https://github.com/snowplow/snowbridge/blob/master/assets/docs/configuration/transformations/builtin/jqFilter-minimal-example.hcl +``` + +Every configuration option: + +```hcl reference +https://github.com/snowplow/snowbridge/blob/master/assets/docs/configuration/transformations/builtin/jqFilter-full-example.hcl +``` + +## Helper Functions + +```mdx-code-block +import JQHelpersSharedBlock from "./reusable/_jqHelpers.md" + + +``` diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/reusable/_jqDescription.md b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/reusable/_jqDescription.md new file mode 100644 index 0000000000..ad3548ef87 --- /dev/null +++ b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/reusable/_jqDescription.md @@ -0,0 +1,5 @@ +[jq](https://github.com/jqlang/jq) is a lightweight and flexible command-line JSON processor. Snowbridge's jq features utilise the [gojq](https://github.com/itchyny/gojq) package, which is a pure Go implementation of jq. jq is Turing complete, so these features allow you to configure arbitrary logic dealing with JSON data structures. + +jq supports formatting values, mathematical operations, boolean comparisons, regex matches, and many more useful features. To get started with jq command, see the [tutorial](https://jqlang.github.io/jq/tutorial/), and [full reference manual](https://jqlang.github.io/jq/manual/). [This open-source jq playground tool](https://jqplay.org/) may also be helpful. + +For most use cases, you are unlikely to encounter them, but note that there are [some small differences](https://github.com/itchyny/gojq?tab=readme-ov-file#difference-to-jq) between jq and gojq. diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/reusable/_jqHelpers.md b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/reusable/_jqHelpers.md new file mode 100644 index 0000000000..e0cb9c63de --- /dev/null +++ b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/reusable/_jqHelpers.md @@ -0,0 +1,13 @@ +In addition to the native functions available in the jq language, the following helper functions are available for use in a jq query: + +* `epoch` converts a Go `time.Time` timestamp to an epoch timestamp in seconds, as integer type. jq's native timestamp-based functions expect integer input, but the Snowplow Analytics SDK provides base level timestamps as `time.Time`. This function can be chained with jq native functions to get past this limitation. For example: + +``` +{ foo: .collector_tstamp | epoch | todateiso8601 } +``` + +* `epochMillis` converts a Go `time.Time` timestamp to an epoch timestamp in milliseconds, as unsigned integer type. Because of how integers are handled in Go, unsigned integers aren't compatible with jq's native timestamp functions, so the `epoch` function truncates to seconds, and the `epochMillis` function exists in case milliseconds are needed. This function cannot be chained with native jq functions, but where milliseconds matter for a value, use this function. + +``` +{ foo: .collector_tstamp | epochMillis } +``` \ No newline at end of file diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedFilter.md b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedFilter.md index 9cea971a8e..22e4f5b8dc 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedFilter.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedFilter.md @@ -8,6 +8,8 @@ Filters can be used in one of two ways, which is determined by the `filter_actio This example filters out all data whose `platform` value does not match either `web` or `mobile`. +## Configuration options + Minimal configuration: ```hcl reference diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedFilterContext.md b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedFilterContext.md index c23629bfb3..31cc42f310 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedFilterContext.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedFilterContext.md @@ -2,7 +2,7 @@ `spEnrichedFilterContext`: Specific to Snowplow data. Filters messages based on a regex match against a field in a context. -This transformation is for use on fileds from contexts. +This transformation is for use on fields from contexts. Note that if the same context is present in the data more than once, one instance of a match is enough for the regex condition to be considered a match — and the message to be kept. @@ -14,6 +14,8 @@ Filters can be used in one of two ways, which is determined by the `filter_actio The below example keeps messages which contain `prod` in the `environment` field of the `contexts_com_acme_env_context_1` context. Note that the `contexts_com_acme_env_context_1` context is attached more than once, if _any_ of the values at `dev` don't match `environment`, the message will be kept. +## Configuration options + Minimal configuration: ```hcl reference diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedFilterUnstructEvent.md b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedFilterUnstructEvent.md index 2191a36688..f0e5759c7c 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedFilterUnstructEvent.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedFilterUnstructEvent.md @@ -12,6 +12,8 @@ The path to the field to match against must be provided as a jsonpath (dot notat Filters can be used in one of two ways, which is determined by the `filter_action` option. `filter_action` determines the behavior of the app when the regex provided evaluates to `true`. If it's set to `"keep"`, the app will complete the remaining transformations and send the message to the destination (unless a subsequent filter determines otherwise). If it's set to `"drop"`, the message will be acked and discarded, without continuing to the next transformation or target. +## Configuration options + This example keeps all events whose `add_to_cart` event data at the `sku` field matches `test-data`. Minimal configuration: diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedSetPk.md b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedSetPk.md index 4dee477b8b..98b7594176 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedSetPk.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedSetPk.md @@ -2,9 +2,10 @@ `spEnrichedSetPk`: Specific to Snowplow data. Sets the message's destination partition key to an atomic field from a Snowplow Enriched tsv string. The input data must be a valid Snowplow enriched TSV. -`SpEnrichedSetPk` only takes one option — the field to use for the partition key. -Example: +## Configuration options + +`SpEnrichedSetPk` only takes one option — the field to use for the partition key. ```hcl reference https://github.com/snowplow/snowbridge/blob/master/assets/docs/configuration/transformations/snowplow-builtin/spEnrichedSetPk-minimal-example.hcl diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedToJson.md b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedToJson.md index 31f7a3c041..4d20c59b73 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedToJson.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spEnrichedToJson.md @@ -4,7 +4,7 @@ `spEnrichedToJson` has no options. -Example: +## Configuration options ```hcl reference https://github.com/snowplow/snowbridge/blob/master/assets/docs/configuration/transformations/snowplow-builtin/spEnrichedToJson-minimal-example.hcl diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spGtmssPreview.md b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spGtmssPreview.md index 40ccfd5e5e..8a8835beae 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spGtmssPreview.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/builtin/spGtmssPreview.md @@ -8,10 +8,20 @@ This transformation was added in version 2.3.0 Only one preview mode context should be sent at a time. -There are no configuration optons for this transformation. +:::note +As of version 3.0.0: + +Invalid preview headers sent to GTM SS can result in requests failing, which may be problematic. There is insufficient information available about the values to allow us to confidently validate them, but we do two things to avoid this problem. + +First, we validate to ensure that the value is a valid base64 string. Second, we compare the age of the event (based on `collector_tstamp`) to ensure it is under a configurable timeout age. If either of these conditions fail, we treat the message as invalid, and output to the failure target. +::: -Example: +## Configuration options ```hcl reference https://github.com/snowplow/snowbridge/blob/master/assets/docs/configuration/transformations/snowplow-builtin/spGtmssPreview-minimal-example.hcl ``` + +```hcl reference +https://github.com/snowplow/snowbridge/blob/master/assets/docs/configuration/transformations/snowplow-builtin/spGtmssPreview-full-example.hcl +``` \ No newline at end of file diff --git a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/custom-scripts/javascript-configuration/index.md b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/custom-scripts/javascript-configuration/index.md index 23e217844b..b9ee0f107a 100644 --- a/docs/destinations/forwarding-events/snowbridge/configuration/transformations/custom-scripts/javascript-configuration/index.md +++ b/docs/destinations/forwarding-events/snowbridge/configuration/transformations/custom-scripts/javascript-configuration/index.md @@ -16,13 +16,15 @@ If a script errors or times out, a [transformation failure](/docs/destinations/f Scripts must be available to the runtime of the application at the path provided in the `script_path` configuration option. For docker, this means mounting the script to the container and providing that path. -Here is an example of a minimal configuration for the custom Javascript script transformation: +## Configuration options + +Minimal Configuration: ```hcl reference https://github.com/snowplow/snowbridge/blob/master/assets/docs/configuration/transformations/custom-scripts/js-configuration-minimal-example.hcl ``` -Here is an example of every configuration option: +Every configuration option: ```hcl reference https://github.com/snowplow/snowbridge/blob/master/assets/docs/configuration/transformations/custom-scripts/js-configuration-full-example.hcl diff --git a/docs/destinations/forwarding-events/snowbridge/testing/index.md b/docs/destinations/forwarding-events/snowbridge/testing/index.md index 57b487c571..e6bc54954d 100644 --- a/docs/destinations/forwarding-events/snowbridge/testing/index.md +++ b/docs/destinations/forwarding-events/snowbridge/testing/index.md @@ -36,13 +36,13 @@ Point some test environment tracking to `localhost:9090`, and your events should You can run Snowbridge locally via Docker: { -`docker run --env SNOWBRIDGE_CONFIG_FILE="" --env ACCEPT_LIMITED_USE_LICENSE=yes snowplow/snowbridge:${versions.snowbridge}` +`docker run --env ACCEPT_LIMITED_USE_LICENSE=yes snowplow/snowbridge:${versions.snowbridge}` } The default configuration for Snowbridge uses the `stdin` source and the `stdout` target. So, to test sending data through with no transformations, we can run the following command (where `data.tsv` is a file with Snowplow events in TSV format): { -`cat data.tsv | docker run --env SNOWBRIDGE_CONFIG_FILE="" --env ACCEPT_LIMITED_USE_LICENSE=yes -i snowplow/snowbridge:${versions.snowbridge}` +`cat data.tsv | docker run --env ACCEPT_LIMITED_USE_LICENSE=yes -i snowplow/snowbridge:${versions.snowbridge}` } This will print the data to the terminal, along with logs. @@ -106,4 +106,4 @@ transform { ## Further testing -You can use the above method to test all aspects of the app from a local environment too, including sources, targets, failure targets, metrics endpoints etc. In some cases, you'll need to ensure that the local envionment has access to any required resources and can authenticate (e.g. connecting from a laptop to a cloud account/local mock of cloud resources, or setting up a local metrics server for testing). Once that’s done, provide Snowbridge with an hcl file configuring it to connect to those resources, and run it the same way as in the examples above. +You can use the above method to test all aspects of the app from a local environment too, including sources, targets, failure targets, metrics endpoints etc. In some cases, you'll need to ensure that the local environment has access to any required resources and can authenticate (e.g. connecting from a laptop to a cloud account/local mock of cloud resources, or setting up a local metrics server for testing). Once that’s done, provide Snowbridge with an hcl file configuring it to connect to those resources, and run it the same way as in the examples above. diff --git a/src/componentVersions.js b/src/componentVersions.js index 4328aacbcd..5d9b6e2d9a 100644 --- a/src/componentVersions.js +++ b/src/componentVersions.js @@ -26,7 +26,7 @@ export const versions = { enrich: '5.2.0', sqs2kinesis: '1.0.4', dataflowRunner: '0.7.5', - snowbridge: '2.4.2', + snowbridge: '3.0.1', // Loaders bqLoader: '1.7.1',