mirror of
https://github.com/laurent22/joplin.git
synced 2024-12-24 10:27:10 +02:00
516 lines
17 KiB
JavaScript
516 lines
17 KiB
JavaScript
/* eslint-disable no-unused-vars */
|
|
/* eslint prefer-const: 0*/
|
|
|
|
|
|
const time = require('@joplinapp/lib/time').default;
|
|
const { fileContentEqual, setupDatabase, setupDatabaseAndSynchronizer, asyncTest, db, synchronizer, fileApi, sleep, clearDatabase, switchClient, syncTargetId, objectsEqual, checkThrowAsync, mockDate, restoreDate } = require('./test-utils.js');
|
|
const SearchEngine = require('@joplinapp/lib/services/searchengine/SearchEngine');
|
|
const Note = require('@joplinapp/lib/models/Note');
|
|
const ItemChange = require('@joplinapp/lib/models/ItemChange');
|
|
const Setting = require('@joplinapp/lib/models/Setting').default;
|
|
|
|
process.on('unhandledRejection', (reason, p) => {
|
|
console.log('Unhandled Rejection at: Promise', p, 'reason:', reason);
|
|
});
|
|
|
|
let engine = null;
|
|
|
|
|
|
const IDF = (N, n) => Math.max(Math.log((N - n + 0.5) / (n + 0.5)), 0);
|
|
|
|
const frequency = (word, string) => {
|
|
const re = new RegExp(`\\b(${word})\\b`, 'g');
|
|
return (string.match(re) || []).length;
|
|
};
|
|
|
|
const calculateScore = (searchString, notes) => {
|
|
const K1 = 1.2;
|
|
const B = 0.75;
|
|
|
|
const freqTitle = notes.map(note => frequency(searchString, note.title));
|
|
const notesWithWord = freqTitle.filter(count => count !== 0).length;
|
|
const numTokens = notes.map(note => note.title.split(' ').length);
|
|
const avgTokens = Math.round(numTokens.reduce((a, b) => a + b, 0) / notes.length);
|
|
|
|
const msSinceEpoch = Math.round(new Date().getTime());
|
|
const msPerDay = 86400000;
|
|
const weightForDaysSinceLastUpdate = (row) => {
|
|
// BM25 weights typically range 0-10, and last updated date should weight similarly, though prioritizing recency logarithmically.
|
|
// An alpha of 200 ensures matches in the last week will show up front (11.59) and often so for matches within 2 weeks (5.99),
|
|
// but is much less of a factor at 30 days (2.84) or very little after 90 days (0.95), focusing mostly on content at that point.
|
|
if (!row.user_updated_time) {
|
|
return 0;
|
|
}
|
|
|
|
const alpha = 200;
|
|
const daysSinceLastUpdate = (msSinceEpoch - row.user_updated_time) / msPerDay;
|
|
return alpha * Math.log(1 + 1 / Math.max(daysSinceLastUpdate, 0.5));
|
|
};
|
|
|
|
let titleBM25WeightedByLastUpdate = new Array(notes.length).fill(-1);
|
|
if (avgTokens != 0) {
|
|
for (let i = 0; i < notes.length; i++) {
|
|
titleBM25WeightedByLastUpdate[i] = IDF(notes.length, notesWithWord) * ((freqTitle[i] * (K1 + 1)) / (freqTitle[i] + K1 * (1 - B + B * (numTokens[i] / avgTokens))));
|
|
titleBM25WeightedByLastUpdate[i] += weightForDaysSinceLastUpdate(notes[i]);
|
|
}
|
|
}
|
|
|
|
const scores = [];
|
|
for (let i = 0; i < notes.length; i++) {
|
|
if (freqTitle[i]) scores.push(titleBM25WeightedByLastUpdate[i]);
|
|
}
|
|
|
|
scores.sort().reverse();
|
|
return scores;
|
|
};
|
|
|
|
describe('services_SearchEngine', function() {
|
|
|
|
beforeEach(async (done) => {
|
|
await setupDatabaseAndSynchronizer(1);
|
|
await switchClient(1);
|
|
|
|
engine = new SearchEngine();
|
|
engine.setDb(db());
|
|
|
|
done();
|
|
});
|
|
|
|
it('should keep the content and FTS table in sync', asyncTest(async () => {
|
|
let rows, n1, n2, n3;
|
|
|
|
n1 = await Note.save({ title: 'a' });
|
|
n2 = await Note.save({ title: 'b' });
|
|
await engine.syncTables();
|
|
rows = await engine.search('a');
|
|
expect(rows.length).toBe(1);
|
|
expect(rows[0].title).toBe('a');
|
|
|
|
await Note.delete(n1.id);
|
|
await engine.syncTables();
|
|
rows = await engine.search('a');
|
|
expect(rows.length).toBe(0);
|
|
rows = await engine.search('b');
|
|
expect(rows[0].title).toBe('b');
|
|
|
|
await Note.save({ id: n2.id, title: 'c' });
|
|
await engine.syncTables();
|
|
rows = await engine.search('b');
|
|
expect(rows.length).toBe(0);
|
|
rows = await engine.search('c');
|
|
expect(rows[0].title).toBe('c');
|
|
|
|
await Note.save({ id: n2.id, encryption_applied: 1 });
|
|
await engine.syncTables();
|
|
rows = await engine.search('c');
|
|
expect(rows.length).toBe(0);
|
|
|
|
await Note.save({ id: n2.id, encryption_applied: 0 });
|
|
await engine.syncTables();
|
|
rows = await engine.search('c');
|
|
expect(rows.length).toBe(1);
|
|
}));
|
|
|
|
it('should, after initial indexing, save the last change ID', asyncTest(async () => {
|
|
const n1 = await Note.save({ title: 'abcd efgh' }); // 3
|
|
const n2 = await Note.save({ title: 'abcd aaaaa abcd abcd' }); // 1
|
|
|
|
expect(Setting.value('searchEngine.initialIndexingDone')).toBe(false);
|
|
|
|
await ItemChange.waitForAllSaved();
|
|
const lastChangeId = await ItemChange.lastChangeId();
|
|
|
|
await engine.syncTables();
|
|
|
|
expect(Setting.value('searchEngine.lastProcessedChangeId')).toBe(lastChangeId);
|
|
expect(Setting.value('searchEngine.initialIndexingDone')).toBe(true);
|
|
}));
|
|
|
|
|
|
it('should order search results by relevance BM25', asyncTest(async () => {
|
|
// BM25 is based on term frequency - inverse document frequency
|
|
// The tf–idf value increases proportionally to the number of times a word appears in the document
|
|
// and is offset by the number of documents in the corpus that contain the word, which helps to adjust
|
|
// for the fact that some words appear more frequently in general.
|
|
|
|
// BM25 returns weight zero for search term which occurs in more than half the notes.
|
|
// So terms that are abundant in all notes to have zero relevance w.r.t BM25.
|
|
|
|
const n1 = await Note.save({ title: 'abcd efgh' }); // 3
|
|
const n2 = await Note.save({ title: 'abcd efgh abcd abcd' }); // 1
|
|
const n3 = await Note.save({ title: 'abcd aaaaa bbbb eeee abcd' }); // 2
|
|
const n4 = await Note.save({ title: 'xyz xyz' });
|
|
const n5 = await Note.save({ title: 'xyz xyz xyz xyz' });
|
|
const n6 = await Note.save({ title: 'xyz xyz xyz xyz xyz xyz' });
|
|
const n7 = await Note.save({ title: 'xyz xyz xyz xyz xyz xyz' });
|
|
const n8 = await Note.save({ title: 'xyz xyz xyz xyz xyz xyz xyz xyz' });
|
|
|
|
await engine.syncTables();
|
|
let rows = await engine.search('abcd');
|
|
|
|
expect(rows[0].id).toBe(n2.id);
|
|
expect(rows[1].id).toBe(n3.id);
|
|
expect(rows[2].id).toBe(n1.id);
|
|
|
|
rows = await engine.search('abcd efgh');
|
|
expect(rows[0].id).toBe(n1.id); // shorter note; also 'efgh' is more rare than 'abcd'.
|
|
expect(rows[1].id).toBe(n2.id);
|
|
}));
|
|
|
|
it('should correctly weigh notes using BM25 and user_updated_time', asyncTest(async () => {
|
|
await mockDate(2020, 9, 30, 50);
|
|
const noteData = [
|
|
{
|
|
title: 'abc test2 test2',
|
|
updated_time: 1601425064756,
|
|
user_updated_time: 1601425064756,
|
|
created_time: 1601425064756,
|
|
user_created_time: 1601425064756,
|
|
},
|
|
{
|
|
title: 'foo foo',
|
|
updated_time: 1601425064758,
|
|
user_updated_time: 1601425064758,
|
|
created_time: 1601425064758,
|
|
user_created_time: 1601425064758,
|
|
},
|
|
{
|
|
title: 'dead beef',
|
|
updated_time: 1601425064760,
|
|
user_updated_time: 1601425064760,
|
|
created_time: 1601425064760,
|
|
user_created_time: 1601425064760,
|
|
},
|
|
{
|
|
title: 'test2 bar',
|
|
updated_time: 1601425064761,
|
|
user_updated_time: 1601425064761,
|
|
created_time: 1601425064761,
|
|
user_created_time: 1601425064761,
|
|
},
|
|
{
|
|
title: 'blah blah abc',
|
|
updated_time: 1601425064763,
|
|
user_updated_time: 1601425064763,
|
|
created_time: 1601425064763,
|
|
user_created_time: 1601425064763,
|
|
},
|
|
];
|
|
|
|
const n0 = await Note.save(noteData[0], { autoTimestamp: false });
|
|
const n1 = await Note.save(noteData[1], { autoTimestamp: false });
|
|
const n2 = await Note.save(noteData[2], { autoTimestamp: false });
|
|
const n3 = await Note.save(noteData[3], { autoTimestamp: false });
|
|
const n4 = await Note.save(noteData[4], { autoTimestamp: false });
|
|
restoreDate();
|
|
await engine.syncTables();
|
|
await mockDate(2020, 9, 30, 50);
|
|
|
|
let searchString = 'abc';
|
|
let scores = calculateScore(searchString, noteData);
|
|
let rows = await engine.search(searchString);
|
|
|
|
expect(rows[0].weight).toEqual(scores[0]);
|
|
expect(rows[1].weight).toEqual(scores[1]);
|
|
|
|
// console.log(rows);
|
|
// console.log(scores);
|
|
|
|
searchString = 'test2';
|
|
scores = calculateScore(searchString, noteData);
|
|
rows = await engine.search(searchString);
|
|
|
|
// console.log(rows);
|
|
// console.log(scores);
|
|
|
|
expect(rows[0].weight).toEqual(scores[0]);
|
|
expect(rows[1].weight).toEqual(scores[1]);
|
|
|
|
searchString = 'foo';
|
|
scores = calculateScore(searchString, noteData);
|
|
rows = await engine.search(searchString);
|
|
|
|
// console.log(rows);
|
|
// console.log(scores);
|
|
|
|
expect(rows[0].weight).toEqual(scores[0]);
|
|
await restoreDate();
|
|
}));
|
|
|
|
it('should tell where the results are found', asyncTest(async () => {
|
|
const notes = [
|
|
await Note.save({ title: 'abcd efgh', body: 'abcd' }),
|
|
await Note.save({ title: 'abcd' }),
|
|
await Note.save({ title: 'efgh', body: 'abcd' }),
|
|
];
|
|
|
|
await engine.syncTables();
|
|
|
|
const testCases = [
|
|
['abcd', ['title', 'body'], ['title'], ['body']],
|
|
['efgh', ['title'], [], ['title']],
|
|
];
|
|
|
|
for (const testCase of testCases) {
|
|
const rows = await engine.search(testCase[0]);
|
|
|
|
for (let i = 0; i < notes.length; i++) {
|
|
const row = rows.find(row => row.id === notes[i].id);
|
|
const actual = row ? row.fields.sort().join(',') : '';
|
|
const expected = testCase[i + 1].sort().join(',');
|
|
expect(expected).toBe(actual);
|
|
}
|
|
}
|
|
}));
|
|
|
|
it('should order search results by relevance (last updated first)', asyncTest(async () => {
|
|
let rows;
|
|
|
|
const n1 = await Note.save({ title: 'abcd' });
|
|
await sleep(0.1);
|
|
const n2 = await Note.save({ title: 'abcd' });
|
|
await sleep(0.1);
|
|
const n3 = await Note.save({ title: 'abcd' });
|
|
await sleep(0.1);
|
|
|
|
await engine.syncTables();
|
|
rows = await engine.search('abcd');
|
|
|
|
expect(rows[0].id).toBe(n3.id);
|
|
expect(rows[1].id).toBe(n2.id);
|
|
expect(rows[2].id).toBe(n1.id);
|
|
|
|
await Note.save({ id: n1.id, title: 'abcd' });
|
|
|
|
await engine.syncTables();
|
|
rows = await engine.search('abcd');
|
|
expect(rows[0].id).toBe(n1.id);
|
|
expect(rows[1].id).toBe(n3.id);
|
|
expect(rows[2].id).toBe(n2.id);
|
|
}));
|
|
|
|
it('should order search results by relevance (completed to-dos last)', asyncTest(async () => {
|
|
let rows;
|
|
|
|
const n1 = await Note.save({ title: 'abcd', is_todo: 1 });
|
|
await sleep(0.1);
|
|
const n2 = await Note.save({ title: 'abcd', is_todo: 1 });
|
|
await sleep(0.1);
|
|
const n3 = await Note.save({ title: 'abcd', is_todo: 1 });
|
|
await sleep(0.1);
|
|
|
|
await engine.syncTables();
|
|
rows = await engine.search('abcd');
|
|
|
|
expect(rows[0].id).toBe(n3.id);
|
|
expect(rows[1].id).toBe(n2.id);
|
|
expect(rows[2].id).toBe(n1.id);
|
|
|
|
await Note.save({ id: n3.id, todo_completed: Date.now() });
|
|
|
|
await engine.syncTables();
|
|
rows = await engine.search('abcd');
|
|
expect(rows[0].id).toBe(n2.id);
|
|
expect(rows[1].id).toBe(n1.id);
|
|
expect(rows[2].id).toBe(n3.id);
|
|
}));
|
|
|
|
it('should supports various query types', asyncTest(async () => {
|
|
let rows;
|
|
|
|
const n1 = await Note.save({ title: 'abcd efgh ijkl', body: 'aaaa bbbb' });
|
|
const n2 = await Note.save({ title: 'iiii efgh bbbb', body: 'aaaa bbbb' });
|
|
const n3 = await Note.save({ title: 'Агентство Рейтер' });
|
|
const n4 = await Note.save({ title: 'Dog' });
|
|
const n5 = await Note.save({ title: 'СООБЩИЛО' });
|
|
|
|
await engine.syncTables();
|
|
|
|
rows = await engine.search('abcd ijkl');
|
|
expect(rows.length).toBe(1);
|
|
|
|
rows = await engine.search('"abcd ijkl"');
|
|
expect(rows.length).toBe(0);
|
|
|
|
rows = await engine.search('"abcd efgh"');
|
|
expect(rows.length).toBe(1);
|
|
|
|
rows = await engine.search('title:abcd');
|
|
expect(rows.length).toBe(1);
|
|
|
|
rows = await engine.search('title:efgh');
|
|
expect(rows.length).toBe(2);
|
|
|
|
rows = await engine.search('body:abcd');
|
|
expect(rows.length).toBe(0);
|
|
|
|
rows = await engine.search('body:bbbb');
|
|
expect(rows.length).toBe(2);
|
|
|
|
rows = await engine.search('body:bbbb iiii');
|
|
expect(rows.length).toBe(1);
|
|
|
|
rows = await engine.search('Рейтер');
|
|
expect(rows.length).toBe(1);
|
|
|
|
rows = await engine.search('рейтер');
|
|
expect(rows.length).toBe(1);
|
|
|
|
rows = await engine.search('Dog');
|
|
expect(rows.length).toBe(1);
|
|
|
|
rows = await engine.search('dog');
|
|
expect(rows.length).toBe(1);
|
|
|
|
rows = await engine.search('сообщило');
|
|
expect(rows.length).toBe(1);
|
|
}));
|
|
|
|
it('should support queries with or without accents', asyncTest(async () => {
|
|
let rows;
|
|
const n1 = await Note.save({ title: 'père noël' });
|
|
|
|
await engine.syncTables();
|
|
|
|
expect((await engine.search('père')).length).toBe(1);
|
|
expect((await engine.search('pere')).length).toBe(1);
|
|
expect((await engine.search('noe*')).length).toBe(1);
|
|
expect((await engine.search('noë*')).length).toBe(1);
|
|
}));
|
|
|
|
it('should support queries with Chinese characters', asyncTest(async () => {
|
|
let rows;
|
|
const n1 = await Note.save({ title: '我是法国人', body: '中文测试' });
|
|
|
|
await engine.syncTables();
|
|
|
|
expect((await engine.search('我')).length).toBe(1);
|
|
expect((await engine.search('法国人')).length).toBe(1);
|
|
expect((await engine.search('法国人*'))[0].fields.sort()).toEqual(['body', 'title']); // usually assume that keyword was matched in body
|
|
expect((await engine.search('测试')).length).toBe(1);
|
|
expect((await engine.search('测试'))[0].fields).toEqual(['body']);
|
|
expect((await engine.search('测试*'))[0].fields).toEqual(['body']);
|
|
}));
|
|
|
|
it('should support queries with Japanese characters', asyncTest(async () => {
|
|
let rows;
|
|
const n1 = await Note.save({ title: '私は日本語を話すことができません', body: 'テスト' });
|
|
|
|
await engine.syncTables();
|
|
|
|
expect((await engine.search('日本')).length).toBe(1);
|
|
expect((await engine.search('できません')).length).toBe(1);
|
|
expect((await engine.search('できません*'))[0].fields.sort()).toEqual(['body', 'title']); // usually assume that keyword was matched in body
|
|
expect((await engine.search('テスト'))[0].fields.sort()).toEqual(['body']);
|
|
|
|
}));
|
|
|
|
it('should support queries with Korean characters', asyncTest(async () => {
|
|
let rows;
|
|
const n1 = await Note.save({ title: '이것은 한국말이다' });
|
|
|
|
await engine.syncTables();
|
|
|
|
expect((await engine.search('이것은')).length).toBe(1);
|
|
expect((await engine.search('말')).length).toBe(1);
|
|
}));
|
|
|
|
it('should support queries with Thai characters', asyncTest(async () => {
|
|
let rows;
|
|
const n1 = await Note.save({ title: 'นี่คือคนไทย' });
|
|
|
|
await engine.syncTables();
|
|
|
|
expect((await engine.search('นี่คือค')).length).toBe(1);
|
|
expect((await engine.search('ไทย')).length).toBe(1);
|
|
}));
|
|
|
|
it('should support field restricted queries with Chinese characters', asyncTest(async () => {
|
|
let rows;
|
|
const n1 = await Note.save({ title: '你好', body: '我是法国人' });
|
|
|
|
await engine.syncTables();
|
|
|
|
expect((await engine.search('title:你好*')).length).toBe(1);
|
|
expect((await engine.search('title:你好*'))[0].fields).toEqual(['title']);
|
|
expect((await engine.search('body:法国人')).length).toBe(1);
|
|
expect((await engine.search('body:法国人'))[0].fields).toEqual(['body']);
|
|
expect((await engine.search('body:你好')).length).toBe(0);
|
|
expect((await engine.search('title:你好 body:法国人')).length).toBe(1);
|
|
expect((await engine.search('title:你好 body:法国人'))[0].fields.sort()).toEqual(['body', 'title']);
|
|
expect((await engine.search('title:你好 body:bla')).length).toBe(0);
|
|
expect((await engine.search('title:你好 我是')).length).toBe(1);
|
|
expect((await engine.search('title:你好 我是'))[0].fields.sort()).toEqual(['body', 'title']);
|
|
expect((await engine.search('title:bla 我是')).length).toBe(0);
|
|
|
|
// For non-alpha char, only the first field is looked at, the following ones are ignored
|
|
expect((await engine.search('title:你好 title:hello')).length).toBe(1);
|
|
}));
|
|
|
|
it('should parse normal query strings', asyncTest(async () => {
|
|
let rows;
|
|
|
|
const testCases = [
|
|
['abcd efgh', { _: ['abcd', 'efgh'] }],
|
|
['abcd efgh', { _: ['abcd', 'efgh'] }],
|
|
['title:abcd efgh', { _: ['efgh'], title: ['abcd'] }],
|
|
['title:abcd', { title: ['abcd'] }],
|
|
['"abcd efgh"', { _: ['abcd efgh'] }],
|
|
['title:abcd title:efgh', { title: ['abcd', 'efgh'] }],
|
|
];
|
|
|
|
for (let i = 0; i < testCases.length; i++) {
|
|
const t = testCases[i];
|
|
const input = t[0];
|
|
const expected = t[1];
|
|
const actual = await engine.parseQuery(input);
|
|
|
|
const _Values = actual.terms._ ? actual.terms._.map(v => v.value) : undefined;
|
|
const titleValues = actual.terms.title ? actual.terms.title.map(v => v.value) : undefined;
|
|
const bodyValues = actual.terms.body ? actual.terms.body.map(v => v.value) : undefined;
|
|
|
|
expect(JSON.stringify(_Values)).toBe(JSON.stringify(expected._), `Test case (_) ${i}`);
|
|
expect(JSON.stringify(titleValues)).toBe(JSON.stringify(expected.title), `Test case (title) ${i}`);
|
|
expect(JSON.stringify(bodyValues)).toBe(JSON.stringify(expected.body), `Test case (body) ${i}`);
|
|
}
|
|
}));
|
|
|
|
it('should handle queries with special characters', asyncTest(async () => {
|
|
let rows;
|
|
|
|
const testCases = [
|
|
// "-" is considered a word delimiter so it is stripped off
|
|
// when indexing the notes. "did-not-match" is translated to
|
|
// three word "did", "not", "match"
|
|
['did-not-match', 'did not match'],
|
|
['did-not-match', '"did-not-match"'],
|
|
['does match', 'does match'],
|
|
];
|
|
|
|
for (let i = 0; i < testCases.length; i++) {
|
|
const t = testCases[i];
|
|
const content = t[0];
|
|
const query = t[1];
|
|
|
|
const n = await Note.save({ title: content });
|
|
await engine.syncTables();
|
|
rows = await engine.search(query);
|
|
expect(rows.length).toBe(1);
|
|
|
|
|
|
await Note.delete(n.id);
|
|
}
|
|
}));
|
|
|
|
it('should allow using basic search', asyncTest(async () => {
|
|
const n1 = await Note.save({ title: '- [ ] abcd' });
|
|
const n2 = await Note.save({ title: '[ ] abcd' });
|
|
|
|
await engine.syncTables();
|
|
|
|
expect((await engine.search('"- [ ]"', { searchType: SearchEngine.SEARCH_TYPE_FTS })).length).toBe(0);
|
|
expect((await engine.search('"- [ ]"', { searchType: SearchEngine.SEARCH_TYPE_BASIC })).length).toBe(1);
|
|
expect((await engine.search('"[ ]"', { searchType: SearchEngine.SEARCH_TYPE_BASIC })).length).toBe(2);
|
|
}));
|
|
});
|