Commit 343d4db0 authored by Nick's avatar Nick

feat: algolia search engine

parent 99a7d5d7
...@@ -40,6 +40,7 @@ ...@@ -40,6 +40,7 @@
}, },
"dependencies": { "dependencies": {
"@bugsnag/js": "5.2.0", "@bugsnag/js": "5.2.0",
"algoliasearch": "3.32.1",
"apollo-fetch": "0.7.0", "apollo-fetch": "0.7.0",
"apollo-server": "2.3.3", "apollo-server": "2.3.3",
"apollo-server-express": "2.3.3", "apollo-server-express": "2.3.3",
......
...@@ -4,7 +4,7 @@ description: Algolia is a powerful search-as-a-service solution, made easy to us ...@@ -4,7 +4,7 @@ description: Algolia is a powerful search-as-a-service solution, made easy to us
author: requarks.io author: requarks.io
logo: https://static.requarks.io/logo/algolia.svg logo: https://static.requarks.io/logo/algolia.svg
website: https://www.algolia.com/ website: https://www.algolia.com/
isAvailable: false isAvailable: true
props: props:
appId: appId:
type: String type: String
......
module.exports = { const _ = require('lodash')
activate() { const algoliasearch = require('algoliasearch')
const { pipeline, Transform } = require('stream')
}, /* global WIKI */
deactivate() {
module.exports = {
async activate() {
// not used
}, },
query() { async deactivate() {
// not used
}, },
created() { /**
* INIT
*/
async init() {
WIKI.logger.info(`(SEARCH/ALGOLIA) Initializing...`)
this.client = algoliasearch(this.config.appId, this.config.apiKey)
this.index = this.client.initIndex(this.config.indexName)
// -> Create Search Index
WIKI.logger.info(`(SEARCH/ALGOLIA) Setting index configuration...`)
await this.index.setSettings({
searchableAttributes: [
'title',
'description',
'content'
],
attributesToRetrieve: [
'locale',
'path',
'title',
'description'
],
advancedSyntax: true
})
WIKI.logger.info(`(SEARCH/ALGOLIA) Initialization completed.`)
}, },
updated() { /**
* QUERY
*
* @param {String} q Query
* @param {Object} opts Additional options
*/
async query(q, opts) {
try {
const results = await this.index.search({
query: q,
hitsPerPage: 50
})
return {
results: _.map(results.hits, r => ({
id: r.objectID,
locale: r.locale,
path: r.path,
title: r.title,
description: r.description
})),
suggestions: [],
totalHits: results.nbHits
}
} catch (err) {
WIKI.logger.warn('Search Engine Error:')
WIKI.logger.warn(err)
}
}, },
deleted() { /**
* CREATE
*
* @param {Object} page Page to create
*/
async created(page) {
await this.index.addObject({
objectID: page.hash,
locale: page.localeCode,
path: page.path,
title: page.title,
description: page.description,
content: page.content
})
}, },
renamed() { /**
* UPDATE
*
* @param {Object} page Page to update
*/
async updated(page) {
await this.index.partialUpdateObject({
objectID: page.hash,
title: page.title,
description: page.description,
content: page.content
})
},
/**
* DELETE
*
* @param {Object} page Page to delete
*/
async deleted(page) {
await this.index.deleteObject(page.hash)
}, },
rebuild() { /**
* RENAME
*
* @param {Object} page Page to rename
*/
async renamed(page) {
await this.index.deleteObject(page.sourceHash)
await this.index.addObject({
objectID: page.destinationHash,
locale: page.localeCode,
path: page.destinationPath,
title: page.title,
description: page.description,
content: page.content
})
},
/**
* REBUILD INDEX
*/
async rebuild() {
WIKI.logger.info(`(SEARCH/ALGOLIA) Rebuilding Index...`)
await this.index.clearIndex()
const MAX_DOCUMENT_BYTES = 10 * Math.pow(2, 10) // 10 KB
const MAX_INDEXING_BYTES = 10 * Math.pow(2, 20) - Buffer.from('[').byteLength - Buffer.from(']').byteLength // 10 MB
const MAX_INDEXING_COUNT = 1000
const COMMA_BYTES = Buffer.from(',').byteLength
let chunks = []
let bytes = 0
const processDocument = async (cb, doc) => {
try {
if (doc) {
const docBytes = Buffer.from(JSON.stringify(doc)).byteLength
// -> Document too large
if (docBytes >= MAX_DOCUMENT_BYTES) {
throw new Error('Document exceeds maximum size allowed by Algolia.')
}
// -> Current batch exceeds size hard limit, flush
if (docBytes + COMMA_BYTES + bytes >= MAX_INDEXING_BYTES) {
await flushBuffer()
}
if (chunks.length > 0) {
bytes += COMMA_BYTES
}
bytes += docBytes
chunks.push(doc)
// -> Current batch exceeds count soft limit, flush
if (chunks.length >= MAX_INDEXING_COUNT) {
await flushBuffer()
}
} else {
// -> End of stream, flush
await flushBuffer()
}
cb()
} catch (err) {
cb(err)
}
}
const flushBuffer = async () => {
WIKI.logger.info(`(SEARCH/ALGOLIA) Sending batch of ${chunks.length}...`)
try {
await this.index.addObjects(
_.map(chunks, doc => ({
objectID: doc.id,
locale: doc.locale,
path: doc.path,
title: doc.title,
description: doc.description,
content: doc.content
}))
)
} catch (err) {
WIKI.logger.warn('(SEARCH/ALGOLIA) Failed to send batch to Algolia: ', err)
}
chunks.length = 0
bytes = 0
}
await pipeline(
WIKI.models.knex.column({ id: 'hash' }, 'path', { locale: 'localeCode' }, 'title', 'description', 'content').select().from('pages').where({
isPublished: true,
isPrivate: false
}).stream(),
new Transform({
objectMode: true,
transform: async (chunk, enc, cb) => processDocument(cb, chunk),
flush: async (cb) => processDocument(cb)
})
)
WIKI.logger.info(`(SEARCH/ALGOLIA) Index rebuilt successfully.`)
} }
} }
...@@ -2,6 +2,8 @@ const _ = require('lodash') ...@@ -2,6 +2,8 @@ const _ = require('lodash')
const AWS = require('aws-sdk') const AWS = require('aws-sdk')
const { pipeline, Transform } = require('stream') const { pipeline, Transform } = require('stream')
/* global WIKI */
module.exports = { module.exports = {
async activate() { async activate() {
// not used // not used
...@@ -110,12 +112,12 @@ module.exports = { ...@@ -110,12 +112,12 @@ module.exports = {
rebuildIndex = true rebuildIndex = true
} }
//-> Define suggester // -> Define suggester
const suggesters = await this.client.describeSuggesters({ const suggesters = await this.client.describeSuggesters({
DomainName: this.config.domain, DomainName: this.config.domain,
SuggesterNames: ['default_suggester'] SuggesterNames: ['default_suggester']
}).promise() }).promise()
if(_.get(suggesters, 'Suggesters', []).length < 1) { if (_.get(suggesters, 'Suggesters', []).length < 1) {
WIKI.logger.info(`(SEARCH/AWS) Defining Suggester...`) WIKI.logger.info(`(SEARCH/AWS) Defining Suggester...`)
await this.client.defineSuggester({ await this.client.defineSuggester({
DomainName: this.config.domain, DomainName: this.config.domain,
...@@ -323,7 +325,7 @@ module.exports = { ...@@ -323,7 +325,7 @@ module.exports = {
const flushBuffer = async () => { const flushBuffer = async () => {
WIKI.logger.info(`(SEARCH/AWS) Sending batch of ${chunks.length}...`) WIKI.logger.info(`(SEARCH/AWS) Sending batch of ${chunks.length}...`)
try { try {
const resp = await this.clientDomain.uploadDocuments({ await this.clientDomain.uploadDocuments({
contentType: 'application/json', contentType: 'application/json',
documents: JSON.stringify(_.map(chunks, doc => ({ documents: JSON.stringify(_.map(chunks, doc => ({
type: 'add', type: 'add',
...@@ -351,8 +353,8 @@ module.exports = { ...@@ -351,8 +353,8 @@ module.exports = {
}).stream(), }).stream(),
new Transform({ new Transform({
objectMode: true, objectMode: true,
transform: async (chunk, enc, cb) => await processDocument(cb, chunk), transform: async (chunk, enc, cb) => processDocument(cb, chunk),
flush: async (cb) => await processDocument(cb) flush: async (cb) => processDocument(cb)
}) })
) )
...@@ -364,4 +366,3 @@ module.exports = { ...@@ -364,4 +366,3 @@ module.exports = {
WIKI.logger.info(`(SEARCH/AWS) Index rebuilt successfully.`) WIKI.logger.info(`(SEARCH/AWS) Index rebuilt successfully.`)
} }
} }
...@@ -3,6 +3,8 @@ const { SearchService, QueryType } = require('azure-search-client') ...@@ -3,6 +3,8 @@ const { SearchService, QueryType } = require('azure-search-client')
const request = require('request-promise') const request = require('request-promise')
const { pipeline } = require('stream') const { pipeline } = require('stream')
/* global WIKI */
module.exports = { module.exports = {
async activate() { async activate() {
// not used // not used
...@@ -20,7 +22,7 @@ module.exports = { ...@@ -20,7 +22,7 @@ module.exports = {
// -> Create Search Index // -> Create Search Index
const indexes = await this.client.indexes.list() const indexes = await this.client.indexes.list()
if (!_.find(_.get(indexes, 'result.value', []), ['name', this.config.indexName])) { if (!_.find(_.get(indexes, 'result.value', []), ['name', this.config.indexName])) {
WIKI.logger.info(`(SEARCH/AWS) Creating index...`) WIKI.logger.info(`(SEARCH/AZURE) Creating index...`)
await this.client.indexes.create({ await this.client.indexes.create({
name: this.config.indexName, name: this.config.indexName,
fields: [ fields: [
...@@ -74,7 +76,7 @@ module.exports = { ...@@ -74,7 +76,7 @@ module.exports = {
searchMode: 'analyzingInfixMatching', searchMode: 'analyzingInfixMatching',
sourceFields: ['title', 'description', 'content'] sourceFields: ['title', 'description', 'content']
} }
], ]
}) })
} }
WIKI.logger.info(`(SEARCH/AZURE) Initialization completed.`) WIKI.logger.info(`(SEARCH/AZURE) Initialization completed.`)
......
This diff was suppressed by a .gitattributes entry.
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment