-
Notifications
You must be signed in to change notification settings - Fork 12
/
Copy pathindex.js
274 lines (240 loc) · 10.8 KB
/
index.js
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
import { dirname } from 'node:path'
import { fileURLToPath } from 'node:url'
import { parsers } from '@rdfjs/formats-common'
import rdf from '@zazuko/env'
import { sparqlSerializeQuadStream, sparqlSupportedTypes, sparqlGetRewriteConfiguration } from 'trifid-core'
import { defaultConfiguration } from './lib/config.js'
import { getAcceptHeader } from './lib/headers.js'
import { checkDatasetBaseUrl } from './lib/base.js'
import { createEntityRenderer } from './renderer/entity.js'
import { createMetadataProvider } from './renderer/metadata.js'
const currentDir = dirname(fileURLToPath(import.meta.url))
const DEFAULT_ENDPOINT_NAME = 'default'
const replaceIriInQuery = (query, iri) => {
return query.split('{{iri}}').join(iri)
}
const fixContentTypeHeader = (contentType) => {
return contentType.split(';')[0].trim().toLocaleLowerCase()
}
const factory = async (trifid) => {
const { render, logger, config, query } = trifid
const mergedConfig = { ...defaultConfiguration, ...config }
const entityRenderer = createEntityRenderer({ options: config, logger, query })
const metadataProvider = createMetadataProvider({ options: config })
const { path, ignorePaths, rewrite: rewriteConfigValue, datasetBaseUrl: datasetBaseUrlValue, allowEndpointSwitch: allowEndpointSwitchConfigValue } = config
const allowEndpointSwitch = `${allowEndpointSwitchConfigValue}` === 'true'
const entityTemplatePath = path || `${currentDir}/views/render.hbs`
const datasetBaseUrls = checkDatasetBaseUrl(logger, datasetBaseUrlValue)
/**
* Map of dataset base URLs with their rewrite configuration.
* @type {Map<string, { rewrite: boolean, replaceIri: (iri: string) => string, iriOrigin: (iri: string) => string, datasetBaseUrl: string }>}
*/
const dbu = new Map()
if (datasetBaseUrls.length === 0) {
dbu.set('default', { rewrite: rewriteConfigValue, replaceIri: (iri) => iri, iriOrigin: (iri) => iri, datasetBaseUrl: '' })
logger.debug('No datasetBaseUrl provided, no rewriting will be done')
}
datasetBaseUrls.forEach((value) => {
const rewriteConfig = sparqlGetRewriteConfiguration(rewriteConfigValue, value)
// Just to have all the fields explicitly defined
const { rewrite: rewriteValue, replaceIri, iriOrigin, datasetBaseUrl } = rewriteConfig
dbu.set(value, { rewrite: rewriteValue, replaceIri, iriOrigin, datasetBaseUrl })
logger.debug(`Rewriting is ${rewriteValue ? 'enabled' : 'disabled'} for '${value}' dataset base URL`)
})
const additionalRewritesConfig = config.additionalRewrites || []
if (!Array.isArray(additionalRewritesConfig)) {
throw new Error('The `additionalRewrites` configuration must be an array')
}
const additionalRewrites = additionalRewritesConfig.map((value) => {
if (typeof value === 'string') {
return { find: value }
}
if (!value || !value.find) {
throw new Error('Each additional rewrite must have a `find` value')
}
return value
})
// If `ignorePaths` is not provided or invalid, we configure some defaults values
let ignoredPaths = ignorePaths
if (!ignorePaths || !Array.isArray(ignorePaths)) {
ignoredPaths = ['/query']
}
return {
defaultConfiguration: async () => {
return {
methods: ['GET'],
paths: ['/*'],
}
},
routeHandler: async () => {
/**
* Route handler.
* @param {import('fastify').FastifyRequest & { session: Map<string, any> }} request Request.
* @param {import('fastify').FastifyReply} reply Reply.
*/
const handler = async (request, reply) => {
const currentPath = request.url.split('?')[0]
// Check if it is a path that needs to be ignored (check of type is already done at the load of the plugin)
if (ignoredPaths.includes(currentPath)) {
reply.callNotFound()
return reply
}
// Get the endpoint name from the query parameter or from the cookie (if allowed)
const savedEndpointName = request.cookies.endpointName || DEFAULT_ENDPOINT_NAME
let endpointName = request.query.endpoint || savedEndpointName
endpointName = endpointName.replace(/[^a-z0-9-]/gi, '')
const i18nValue = `${request.session.get('currentLanguage') || 'en'}`.replace(/[^a-z0-9-]/gi, '')
// To avoid any languge issues, we will forward the i18n cookie to the SPARQL endpoint + add the endpointName cookie
const queryHeaders = {
cookie: `i18n=${i18nValue}${allowEndpointSwitch && endpointName !== DEFAULT_ENDPOINT_NAME ? `; endpointName=${endpointName}` : ''}`,
}
// Get the expected format from the Accept header or from the `format` query parameter
const acceptHeader = getAcceptHeader(request)
// Generate the IRI we expect
let requestPort = ''
if (request.port) {
requestPort = `:${request.port}`
}
const fullUrl = `${request.protocol}://${request.hostname}${requestPort}${request.url}`
const iriUrl = new URL(fullUrl)
iriUrl.search = ''
iriUrl.searchParams.forEach((_value, key) => iriUrl.searchParams.delete(key))
const iriUrlString = iriUrl.toString()
let iri
let iriOrigin
let replaceIri
let rewriteValue
let datasetBaseUrl
let isContainer
for (const [_key, value] of dbu) {
if (iri !== undefined) {
break
}
const tmpIri = value.replaceIri(iriUrlString)
const tmpIsContainer = mergedConfig.resourceNoSlash && tmpIri.endsWith('/')
logger.debug(`IRI value: ${tmpIri}${value.rewriteValue ? ' (rewritten)' : ''} - is container: ${tmpIsContainer ? 'true' : 'false'}`)
// Check if the IRI exists in the dataset ; if so, use it for the rest of the process
const askQuery = tmpIsContainer ? mergedConfig.containerExistsQuery : mergedConfig.resourceExistsQuery
const exists = await query(replaceIriInQuery(askQuery, tmpIri), { ask: true, headers: queryHeaders })
if (exists) {
iriOrigin = value.iriOrigin
replaceIri = value.replaceIri
rewriteValue = value.rewrite
datasetBaseUrl = value.datasetBaseUrl
isContainer = tmpIsContainer
iri = tmpIri
logger.debug(`IRI found: ${iri}`)
break
}
}
const rewriteResponse = rewriteValue
? [
...additionalRewrites.map(({ find, replace }) => {
if (!replace) {
return { find, replace: iriOrigin(iriUrlString) }
} else {
return { find, replace }
}
}),
{ find: datasetBaseUrl, replace: iriOrigin(iriUrlString) },
]
: []
// If the IRI is not found, we return a 404
if (!iri) {
reply.callNotFound()
return reply
}
try {
// Check if there is a redirect for the IRI
if (mergedConfig.followRedirects) {
const redirect = await query(replaceIriInQuery(mergedConfig.redirectQuery, iri), {
ask: false,
select: true, // Force the parsing of the response
rewriteResponse,
headers: queryHeaders,
})
if (redirect.length > 0) {
const entityRedirect = redirect[0]
const { responseCode, location } = entityRedirect
if (responseCode && location && responseCode.value && location.value) {
logger.debug(`Redirecting <${iri}> to <${location.value}> (HTTP ${responseCode.value})`)
reply.status(parseInt(responseCode.value, 10)).redirect(location.value)
return reply
} else {
logger.warn('Redirect query did not return the expected results')
}
}
}
// Get the entity from the dataset
const describeQuery = isContainer ? mergedConfig.containerGraphQuery : mergedConfig.resourceGraphQuery
const entity = await query(replaceIriInQuery(describeQuery, iri), {
ask: false,
rewriteResponse,
headers: {
...queryHeaders,
accept: 'application/n-quads',
},
})
const entityContentType = entity.contentType || 'application/n-triples'
const entityStream = entity.response
if (!entityStream) {
reply.callNotFound()
return reply
}
// Make sure the Content-Type is lower case and without parameters (e.g. charset)
const fixedContentType = fixContentTypeHeader(entityContentType)
const quadStream = parsers.import(fixedContentType, entityStream)
if (sparqlSupportedTypes.includes(acceptHeader)) {
const serialized = await sparqlSerializeQuadStream(quadStream, acceptHeader)
reply.type(acceptHeader).send(serialized)
return reply
}
const dataset = await rdf.dataset().import(quadStream)
if (mergedConfig.enableSchemaUrlRedirect && acceptHeader === 'text/html') {
const disabledSchemaUrlRedirect =
request.headers['x-disable-schema-url-redirect'] === 'true' ||
request.query.disableSchemaUrlRedirect === 'true'
// Get all triples that have a schema:URL property with value of type xsd:anyURI
const urls = []
dataset.match(iriUrlString, rdf.ns.schema.URL)
.filter(({ object }) => ['xsd:anyURI', rdf.ns.xsd.anyURI.value].includes(object?.datatype?.value))
.map(({ object }) => urls.push(object.value))
if (!disabledSchemaUrlRedirect && urls.length > 0) {
const redirectUrl = urls[0]
logger.debug(`Redirecting to ${redirectUrl}`)
reply.redirect(redirectUrl)
return reply
}
}
const { entityHtml, entityLabel, entityUrl } = await entityRenderer(
request,
{
dataset,
rewriteResponse,
replaceIri,
headers: queryHeaders,
entityRoot: rewriteValue ? iri.replace(datasetBaseUrl, iriOrigin(iriUrlString)) : iri,
},
)
const metadata = await metadataProvider(request, { dataset })
const jsonldSerialized = await sparqlSerializeQuadStream(dataset.toStream(), 'application/ld+json')
reply.type('text/html').send(await render(request, entityTemplatePath, {
dataset: entityHtml,
entityLabel,
entityUrl,
metadata,
jsonld: jsonldSerialized,
config,
}))
} catch (e) {
logger.error(e)
reply.callNotFound()
return reply
}
return reply
}
return handler
},
}
}
export default factory