1
0
mirror of https://github.com/laurent22/joplin.git synced 2025-04-01 21:24:45 +02:00

Server: Added task to compress changes older than 6 months

This commit is contained in:
Laurent Cozic 2021-11-23 16:25:36 +00:00
parent 799fe81449
commit 75f729620e
6 changed files with 26 additions and 15 deletions

View File

@ -6,14 +6,14 @@ interface Argv {
ttl: number; ttl: number;
} }
export default class DeleteOldChangesCommand extends BaseCommand { export default class CompressOldChangesCommand extends BaseCommand {
public command() { public command() {
return 'delete-old-changes'; return 'compress-old-changes';
} }
public description() { public description() {
return 'deletes old changes'; return 'compresses old changes by discarding consecutive updates';
} }
public options(): Record<string, Options> { public options(): Record<string, Options> {
@ -26,7 +26,7 @@ export default class DeleteOldChangesCommand extends BaseCommand {
} }
public async run(argv: Argv, runContext: RunContext): Promise<void> { public async run(argv: Argv, runContext: RunContext): Promise<void> {
await runContext.models.change().deleteOldChanges(argv.ttl ? argv.ttl * Day : null); await runContext.models.change().compressOldChanges(argv.ttl ? argv.ttl * Day : null);
} }
} }

View File

@ -227,13 +227,13 @@ describe('ChangeModel', function() {
expect(await models().change().count()).toBe(7); expect(await models().change().count()).toBe(7);
// Shouldn't do anything initially because it only deletes old changes. // Shouldn't do anything initially because it only deletes old changes.
await models().change().deleteOldChanges(); await models().change().compressOldChanges();
expect(await models().change().count()).toBe(7); expect(await models().change().count()).toBe(7);
// 180 days after T4, it should delete all U1 updates events except for // 180 days after T4, it should delete all U1 updates events except for
// the last one // the last one
jest.setSystemTime(new Date(t4 + changeTtl).getTime()); jest.setSystemTime(new Date(t4 + changeTtl).getTime());
await models().change().deleteOldChanges(); await models().change().compressOldChanges();
expect(await models().change().count()).toBe(5); expect(await models().change().count()).toBe(5);
{ {
const updateChange = (await models().change().all()).find(c => c.item_id === note1.id && c.type === ChangeType.Update); const updateChange = (await models().change().all()).find(c => c.item_id === note1.id && c.type === ChangeType.Update);
@ -247,13 +247,13 @@ describe('ChangeModel', function() {
// there's only one note 2 change that is older than 90 days at this // there's only one note 2 change that is older than 90 days at this
// point. // point.
jest.setSystemTime(new Date(t5 + changeTtl).getTime()); jest.setSystemTime(new Date(t5 + changeTtl).getTime());
await models().change().deleteOldChanges(); await models().change().compressOldChanges();
expect(await models().change().count()).toBe(5); expect(await models().change().count()).toBe(5);
// After T6, more than 90 days later - now the change at T5 should be // After T6, more than 90 days later - now the change at T5 should be
// deleted, keeping only the change at T6. // deleted, keeping only the change at T6.
jest.setSystemTime(new Date(t6 + changeTtl).getTime()); jest.setSystemTime(new Date(t6 + changeTtl).getTime());
await models().change().deleteOldChanges(); await models().change().compressOldChanges();
expect(await models().change().count()).toBe(4); expect(await models().change().count()).toBe(4);
{ {
const updateChange = (await models().change().all()).find(c => c.item_id === note2.id && c.type === ChangeType.Update); const updateChange = (await models().change().all()).find(c => c.item_id === note2.id && c.type === ChangeType.Update);

View File

@ -309,7 +309,9 @@ export default class ChangeModel extends BaseModel<Change> {
return output; return output;
} }
public async deleteOldChanges(ttl: number = null) { // See spec for complete documentation:
// https://joplinapp.org/spec/server_delta_sync/#regarding-the-deletion-of-old-change-events
public async compressOldChanges(ttl: number = null) {
ttl = ttl === null ? defaultChangeTtl : ttl; ttl = ttl === null ? defaultChangeTtl : ttl;
const cutOffDate = Date.now() - ttl; const cutOffDate = Date.now() - ttl;
const limit = 1000; const limit = 1000;
@ -324,7 +326,7 @@ export default class ChangeModel extends BaseModel<Change> {
let error: Error = null; let error: Error = null;
let totalDeletedCount = 0; let totalDeletedCount = 0;
logger.info(`deleteOldChanges: Processing changes older than: ${formatDateTime(cutOffDate)} (${cutOffDate})`); logger.info(`compressOldChanges: Processing changes older than: ${formatDateTime(cutOffDate)} (${cutOffDate})`);
while (true) { while (true) {
// First get all the UPDATE changes before the specified date, and // First get all the UPDATE changes before the specified date, and
@ -373,14 +375,14 @@ export default class ChangeModel extends BaseModel<Change> {
totalDeletedCount += deletedCount; totalDeletedCount += deletedCount;
doneItemIds.push(row.item_id); doneItemIds.push(row.item_id);
} }
}, 'ChangeModel::deleteOldChanges'); }, 'ChangeModel::compressOldChanges');
logger.info(`deleteOldChanges: Processed: ${doneItemIds.length} items. Deleted: ${totalDeletedCount} changes.`); logger.info(`compressOldChanges: Processed: ${doneItemIds.length} items. Deleted: ${totalDeletedCount} changes.`);
if (error) throw error; if (error) throw error;
} }
logger.info(`deleteOldChanges: Finished processing. Done ${doneItemIds.length} items. Deleted: ${totalDeletedCount} changes.`); logger.info(`compressOldChanges: Finished processing. Done ${doneItemIds.length} items. Deleted: ${totalDeletedCount} changes.`);
} }
public async save(change: Change, options: SaveOptions = {}): Promise<Change> { public async save(change: Change, options: SaveOptions = {}): Promise<Change> {

View File

@ -13,6 +13,7 @@ export enum TaskId {
HandleBetaUserEmails = 4, HandleBetaUserEmails = 4,
HandleFailedPaymentSubscriptions = 5, HandleFailedPaymentSubscriptions = 5,
DeleteExpiredSessions = 6, DeleteExpiredSessions = 6,
CompressOldChanges = 7,
} }
export enum RunType { export enum RunType {

View File

@ -1,7 +1,7 @@
import yargs = require('yargs'); import yargs = require('yargs');
import BaseCommand from '../commands/BaseCommand'; import BaseCommand from '../commands/BaseCommand';
import DbCommand from '../commands/DbCommand'; import DbCommand from '../commands/DbCommand';
import DeleteOldChangesCommand from '../commands/DeleteOldChangesCommand'; import CompressOldChangesCommand from '../commands/CompressOldChangesCommand';
import StorageCommand from '../commands/StorageCommand'; import StorageCommand from '../commands/StorageCommand';
import MigrateCommand from '../commands/MigrateCommand'; import MigrateCommand from '../commands/MigrateCommand';
@ -16,7 +16,7 @@ export default async function setupCommands(): Promise<Commands> {
const commands: BaseCommand[] = [ const commands: BaseCommand[] = [
new MigrateCommand(), new MigrateCommand(),
new DbCommand(), new DbCommand(),
new DeleteOldChangesCommand(), new CompressOldChangesCommand(),
new StorageCommand(), new StorageCommand(),
]; ];

View File

@ -12,6 +12,7 @@ export default function(env: Env, models: Models, config: Config): TaskService {
schedule: '0 */6 * * *', schedule: '0 */6 * * *',
run: (models: Models) => models.token().deleteExpiredTokens(), run: (models: Models) => models.token().deleteExpiredTokens(),
}, },
{ {
id: TaskId.UpdateTotalSizes, id: TaskId.UpdateTotalSizes,
description: 'Update total sizes', description: 'Update total sizes',
@ -19,6 +20,13 @@ export default function(env: Env, models: Models, config: Config): TaskService {
run: (models: Models) => models.item().updateTotalSizes(), run: (models: Models) => models.item().updateTotalSizes(),
}, },
{
id: TaskId.CompressOldChanges,
description: 'Compress old changes',
schedule: '0 0 */2 * *',
run: (models: Models) => models.change().compressOldChanges(),
},
// Need to do it relatively frequently so that if the user fixes // Need to do it relatively frequently so that if the user fixes
// whatever was causing the oversized account, they can get it // whatever was causing the oversized account, they can get it
// re-enabled quickly. Also it's done on minute 30 because it depends on // re-enabled quickly. Also it's done on minute 30 because it depends on