|
|
@ -0,0 +1,109 @@ |
|
|
|
const Model = require('objection').Model |
|
|
|
const path = require('path') |
|
|
|
const fs = require('fs-extra') |
|
|
|
const _ = require('lodash') |
|
|
|
const yaml = require('js-yaml') |
|
|
|
const commonHelper = require('../helpers/common') |
|
|
|
|
|
|
|
/* global WIKI */ |
|
|
|
|
|
|
|
/** |
|
|
|
* SearchEngine model |
|
|
|
*/ |
|
|
|
module.exports = class SearchEngine extends Model { |
|
|
|
static get tableName() { return 'searchEngines' } |
|
|
|
|
|
|
|
static get jsonSchema () { |
|
|
|
return { |
|
|
|
type: 'object', |
|
|
|
required: ['key', 'isEnabled'], |
|
|
|
|
|
|
|
properties: { |
|
|
|
id: {type: 'integer'}, |
|
|
|
key: {type: 'string'}, |
|
|
|
isEnabled: {type: 'boolean'}, |
|
|
|
level: {type: 'string'}, |
|
|
|
config: {type: 'object'} |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
static async getSearchEngines() { |
|
|
|
return WIKI.models.searchEngines.query() |
|
|
|
} |
|
|
|
|
|
|
|
static async refreshSearchEnginesFromDisk() { |
|
|
|
let trx |
|
|
|
try { |
|
|
|
const dbSearchEngines = await WIKI.models.searchEngines.query() |
|
|
|
|
|
|
|
// -> Fetch definitions from disk
|
|
|
|
const searchEnginesDirs = await fs.readdir(path.join(WIKI.SERVERPATH, 'modules/search')) |
|
|
|
let diskSearchEngines = [] |
|
|
|
for (let dir of searchEnginesDirs) { |
|
|
|
const def = await fs.readFile(path.join(WIKI.SERVERPATH, 'modules/search', dir, 'definition.yml'), 'utf8') |
|
|
|
diskSearchEngines.push(yaml.safeLoad(def)) |
|
|
|
} |
|
|
|
WIKI.data.searchEngines = diskSearchEngines.map(searchEngine => ({ |
|
|
|
...searchEngine, |
|
|
|
props: commonHelper.parseModuleProps(searchEngine.props) |
|
|
|
})) |
|
|
|
|
|
|
|
// -> Insert new searchEngines
|
|
|
|
let newSearchEngines = [] |
|
|
|
for (let searchEngine of WIKI.data.searchEngines) { |
|
|
|
if (!_.some(dbSearchEngines, ['key', searchEngine.key])) { |
|
|
|
newSearchEngines.push({ |
|
|
|
key: searchEngine.key, |
|
|
|
isEnabled: false, |
|
|
|
config: _.transform(searchEngine.props, (result, value, key) => { |
|
|
|
_.set(result, key, value.default) |
|
|
|
return result |
|
|
|
}, {}) |
|
|
|
}) |
|
|
|
} else { |
|
|
|
const searchEngineConfig = _.get(_.find(dbSearchEngines, ['key', searchEngine.key]), 'config', {}) |
|
|
|
await WIKI.models.searchEngines.query().patch({ |
|
|
|
config: _.transform(searchEngine.props, (result, value, key) => { |
|
|
|
if (!_.has(result, key)) { |
|
|
|
_.set(result, key, value.default) |
|
|
|
} |
|
|
|
return result |
|
|
|
}, searchEngineConfig) |
|
|
|
}).where('key', searchEngine.key) |
|
|
|
} |
|
|
|
} |
|
|
|
if (newSearchEngines.length > 0) { |
|
|
|
trx = await WIKI.models.Objection.transaction.start(WIKI.models.knex) |
|
|
|
for (let searchEngine of newSearchEngines) { |
|
|
|
await WIKI.models.searchEngines.query(trx).insert(searchEngine) |
|
|
|
} |
|
|
|
await trx.commit() |
|
|
|
WIKI.logger.info(`Loaded ${newSearchEngines.length} new search engines: [ OK ]`) |
|
|
|
} else { |
|
|
|
WIKI.logger.info(`No new search engines found: [ SKIPPED ]`) |
|
|
|
} |
|
|
|
} catch (err) { |
|
|
|
WIKI.logger.error(`Failed to scan or load new search engines: [ FAILED ]`) |
|
|
|
WIKI.logger.error(err) |
|
|
|
if (trx) { |
|
|
|
trx.rollback() |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
static async pageEvent({ event, page }) { |
|
|
|
const searchEngines = await WIKI.models.storage.query().where('isEnabled', true) |
|
|
|
if (searchEngines && searchEngines.length > 0) { |
|
|
|
_.forEach(searchEngines, logger => { |
|
|
|
WIKI.queue.job.syncStorage.add({ |
|
|
|
event, |
|
|
|
logger, |
|
|
|
page |
|
|
|
}, { |
|
|
|
removeOnComplete: true |
|
|
|
}) |
|
|
|
}) |
|
|
|
} |
|
|
|
} |
|
|
|
} |