Skip to content

Commit aae1af8

Browse files
authored
Merge pull request #332 from marklogic/feature/docs-tweak
Doc tweaks
2 parents ee9a185 + bca2c75 commit aae1af8

File tree

5 files changed

+7
-6
lines changed

5 files changed

+7
-6
lines changed

docs/copy.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,7 @@ of `--collections`.
8484

8585
## Building a RAG data pipeline
8686

87-
[Retrieval-augmented generation](https://en.wikipedia.org/wiki/Retrieval-augmented_generation), or RAG, with MarkLogic depends on preparing data so that the most relevant
87+
[Retrieval-augmented generation](https://www.progress.com/marklogic/solutions/generative-ai), or RAG, with MarkLogic depends on preparing data so that the most relevant
8888
chunks of text for a user's question can be sent to a Large Language Model, or LLM. Starting with release 1.2.0, Flux
8989
supports the construction of a data pipeline by splitting the text in a document into chunks and adding a vector
9090
embedding to each chunk while copying data. Please see [the guide on splitting text](import/splitting.md) and

docs/import/common-import-features.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -109,7 +109,7 @@ The following shows an example of each option:
109109

110110
## Building a RAG data pipeline
111111

112-
[Retrieval-augmented generation](https://en.wikipedia.org/wiki/Retrieval-augmented_generation), or RAG, with MarkLogic depends on preparing data so that the most relevant
112+
[Retrieval-augmented generation](https://www.progress.com/marklogic/solutions/generative-ai), or RAG, with MarkLogic depends on preparing data so that the most relevant
113113
chunks of text for a user's question can be sent to a Large Language Model, or LLM. Starting with release 1.2.0, Flux
114114
supports the construction of a data pipeline by splitting the text in a document into chunks and adding a vector
115115
embedding to each chunk while importing data. Please see [the guide on splitting text](splitting.md) and

docs/import/embedder/embedder.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ associated with a "chunk" of text that is usually, but not necessarily, produced
1111
[support for splitting text](../splitting.md). Flux can add embeddings during
1212
any import operation and also when [copying documents](../../copy.md). Adding embeddings to documents is a critical
1313
part of creating a data pipeline in support of
14-
[retrieval-augmented generation, or RAG](https://en.wikipedia.org/wiki/Retrieval-augmented_generation),
14+
[retrieval-augmented generation, or RAG](https://www.progress.com/marklogic/solutions/generative-ai),
1515
use cases that utilize MarkLogic's support for
1616
[vector queries](https://docs.marklogic.com/12.0/guide/release-notes/en/new-features-in-marklogic-12-0-ea1/native-vector-support.html).
1717

docs/import/splitting.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ nav_order: 6
88
Flux supports splitting the text in documents into chunks of configurable size, either written to the source document
99
or to separate "sidecar" documents containing one or more chunks. Flux can split text during any import operation and also
1010
when [copying documents](../copy.md). Splitting text is often a critical part of creating a data pipeline in support
11-
of [retrieval-augmented generation, or RAG](https://en.wikipedia.org/wiki/Retrieval-augmented_generation), use cases with MarkLogic.
11+
of [retrieval-augmented generation, or RAG](https://www.progress.com/marklogic/solutions/generative-ai), use cases with MarkLogic.
1212

1313
## Table of contents
1414
{: .no_toc .text-delta }
@@ -321,7 +321,7 @@ via the `--transform` option for your import command or `--output-transform` opt
321321
By default, Flux will create each XML sidecar document using the following structure:
322322

323323
```
324-
<root>
324+
<root xmlns="http://marklogic.com/appservices/model">
325325
<source-uri>The URI of the source document</source-uri>
326326
<chunks>
327327
<chunk><text>The first chunk</text></chunk>

docs/index.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,6 +12,7 @@ With Flux, you can automate common data movement use cases including:
1212

1313
- Importing rows from an RDBMS.
1414
- Importing JSON, XML, CSV, Parquet and other file types from a local filesystem or S3.
15+
- Implementing a data pipeline for a [RAG solution with MarkLogic](https://www.progress.com/marklogic/solutions/generative-ai).
1516
- Copying data from one MarkLogic database to another database.
1617
- Reprocessing data in MarkLogic via custom code.
1718
- Exporting data to an RDBMS, a local filesystem, or S3.
@@ -25,7 +26,7 @@ Flux has the following system requirements:
2526
Java 21 should work but has not been thoroughly tested yet. Java 23 will not yet work.
2627

2728
Earlier versions of MarkLogic 9 and 10 will support any features not involving Optic queries.
28-
Additionally, the latest version of MarkLogic 11 is recommended if possible.
29+
Additionally, the latest version of MarkLogic 11 or 12 is recommended if possible.
2930

3031
Flux is built on top of [Apache Spark](https://spark.apache.org/), but you do not need to know anything about Spark
3132
to use Flux. If you are already making use of Spark for other use cases, see the

0 commit comments

Comments
 (0)