import Logger from './Logger'; import LockHandler, { LockType } from './services/synchronizer/LockHandler'; import Setting from './models/Setting'; import shim from './shim'; import MigrationHandler from './services/synchronizer/MigrationHandler'; import eventManager from './eventManager'; import { _ } from './locale'; import BaseItem from './models/BaseItem'; import Folder from './models/Folder'; import Note from './models/Note'; import Resource from './models/Resource'; import ItemChange from './models/ItemChange'; import ResourceLocalState from './models/ResourceLocalState'; import MasterKey from './models/MasterKey'; import BaseModel, { ModelType } from './BaseModel'; import time from './time'; import ResourceService from './services/ResourceService'; import EncryptionService from './services/e2ee/EncryptionService'; import JoplinError from './JoplinError'; import ShareService from './services/share/ShareService'; import TaskQueue from './TaskQueue'; import ItemUploader from './services/synchronizer/ItemUploader'; import { FileApi } from './file-api'; import JoplinDatabase from './JoplinDatabase'; import { fetchSyncInfo, getActiveMasterKey, localSyncInfo, mergeSyncInfos, saveLocalSyncInfo, SyncInfo, syncInfoEquals, uploadSyncInfo } from './services/synchronizer/syncInfoUtils'; import { getMasterPassword, setupAndDisableEncryption, setupAndEnableEncryption } from './services/e2ee/utils'; import { generateKeyPair } from './services/e2ee/ppk'; const { sprintf } = require('sprintf-js'); const { Dirnames } = require('./services/synchronizer/utils/types'); const logger = Logger.create('Synchronizer'); interface RemoteItem { id: string; path?: string; type_?: number; isDeleted?: boolean; // This the time when the file was created on the server. It is used for // example for the locking mechanim or any file that's not an actual Joplin // item. updated_time?: number; // This is the time that corresponds to the actual Joplin item updated_time // value. A note is always uploaded with a delay so the server updated_time // value will always be ahead. However for synchronising we need to know the // exact Joplin item updated_time value. jop_updated_time?: number; } function isCannotSyncError(error: any): boolean { if (!error) return false; if (['rejectedByTarget', 'fileNotFound'].indexOf(error.code) >= 0) return true; // If the request times out we give up too because sometimes it's due to the // file being large or some other connection issues, and we don't want that // file to block the sync process. The user can choose to retry later on. // // message: "network timeout at: ..... // name: "FetchError" // type: "request-timeout" if (error.type === 'request-timeout' || error.message.includes('network timeout')) return true; return false; } export default class Synchronizer { public static verboseMode: boolean = true; private db_: JoplinDatabase; private api_: FileApi; private appType_: string; private logger_: Logger = new Logger(); private state_: string = 'idle'; private cancelling_: boolean = false; public maxResourceSize_: number = null; private downloadQueue_: any = null; private clientId_: string; private lockHandler_: LockHandler; private migrationHandler_: MigrationHandler; private encryptionService_: EncryptionService = null; private resourceService_: ResourceService = null; private syncTargetIsLocked_: boolean = false; private shareService_: ShareService = null; // Debug flags are used to test certain hard-to-test conditions // such as cancelling in the middle of a loop. public testingHooks_: string[] = []; private onProgress_: Function; private progressReport_: any = {}; public dispatch: Function; public constructor(db: JoplinDatabase, api: FileApi, appType: string) { this.db_ = db; this.api_ = api; this.appType_ = appType; this.clientId_ = Setting.value('clientId'); this.onProgress_ = function() {}; this.progressReport_ = {}; this.dispatch = function() {}; this.apiCall = this.apiCall.bind(this); } state() { return this.state_; } db() { return this.db_; } api() { return this.api_; } clientId() { return this.clientId_; } setLogger(l: Logger) { this.logger_ = l; } logger() { return this.logger_; } lockHandler() { if (this.lockHandler_) return this.lockHandler_; this.lockHandler_ = new LockHandler(this.api()); return this.lockHandler_; } migrationHandler() { if (this.migrationHandler_) return this.migrationHandler_; this.migrationHandler_ = new MigrationHandler(this.api(), this.db(), this.lockHandler(), this.appType_, this.clientId_); return this.migrationHandler_; } maxResourceSize() { if (this.maxResourceSize_ !== null) return this.maxResourceSize_; return this.appType_ === 'mobile' ? 100 * 1000 * 1000 : Infinity; } public setShareService(v: ShareService) { this.shareService_ = v; } public setEncryptionService(v: any) { this.encryptionService_ = v; } encryptionService() { return this.encryptionService_; } public setResourceService(v: ResourceService) { this.resourceService_ = v; } protected resourceService(): ResourceService { return this.resourceService_; } async waitForSyncToFinish() { if (this.state() === 'idle') return; while (true) { await time.sleep(1); if (this.state() === 'idle') return; } } private static reportHasErrors(report: any): boolean { return !!report && !!report.errors && !!report.errors.length; } static reportToLines(report: any) { const lines = []; if (report.createLocal) lines.push(_('Created local items: %d.', report.createLocal)); if (report.updateLocal) lines.push(_('Updated local items: %d.', report.updateLocal)); if (report.createRemote) lines.push(_('Created remote items: %d.', report.createRemote)); if (report.updateRemote) lines.push(_('Updated remote items: %d.', report.updateRemote)); if (report.deleteLocal) lines.push(_('Deleted local items: %d.', report.deleteLocal)); if (report.deleteRemote) lines.push(_('Deleted remote items: %d.', report.deleteRemote)); if (report.fetchingTotal && report.fetchingProcessed) lines.push(_('Fetched items: %d/%d.', report.fetchingProcessed, report.fetchingTotal)); if (report.cancelling && !report.completedTime) lines.push(_('Cancelling...')); if (report.completedTime) lines.push(_('Completed: %s (%s)', time.formatMsToLocal(report.completedTime), `${Math.round((report.completedTime - report.startTime) / 1000)}s`)); if (this.reportHasErrors(report)) lines.push(_('Last error: %s', report.errors[report.errors.length - 1].toString().substr(0, 500))); return lines; } logSyncOperation(action: any, local: any = null, remote: RemoteItem = null, message: string = null, actionCount: number = 1) { const line = ['Sync']; line.push(action); if (message) line.push(message); let type = local && local.type_ ? local.type_ : null; if (!type) type = remote && remote.type_ ? remote.type_ : null; if (type) line.push(BaseItem.modelTypeToClassName(type)); if (local) { const s = []; s.push(local.id); line.push(`(Local ${s.join(', ')})`); } if (remote) { const s = []; s.push(remote.id ? remote.id : remote.path); line.push(`(Remote ${s.join(', ')})`); } if (Synchronizer.verboseMode) { logger.info(line.join(': ')); } else { logger.debug(line.join(': ')); } if (!this.progressReport_[action]) this.progressReport_[action] = 0; this.progressReport_[action] += actionCount; this.progressReport_.state = this.state(); this.onProgress_(this.progressReport_); // Make sure we only send a **copy** of the report since it // is mutated within this class. Should probably use a lib // for this but for now this simple fix will do. const reportCopy: any = {}; for (const n in this.progressReport_) reportCopy[n] = this.progressReport_[n]; if (reportCopy.errors) reportCopy.errors = this.progressReport_.errors.slice(); this.dispatch({ type: 'SYNC_REPORT_UPDATE', report: reportCopy }); } async logSyncSummary(report: any) { logger.info('Operations completed: '); for (const n in report) { if (!report.hasOwnProperty(n)) continue; if (n == 'errors') continue; if (n == 'starting') continue; if (n == 'finished') continue; if (n == 'state') continue; if (n == 'startTime') continue; if (n == 'completedTime') continue; logger.info(`${n}: ${report[n] ? report[n] : '-'}`); } const folderCount = await Folder.count(); const noteCount = await Note.count(); const resourceCount = await Resource.count(); logger.info(`Total folders: ${folderCount}`); logger.info(`Total notes: ${noteCount}`); logger.info(`Total resources: ${resourceCount}`); if (Synchronizer.reportHasErrors(report)) { logger.warn('There was some errors:'); for (let i = 0; i < report.errors.length; i++) { const e = report.errors[i]; logger.warn(e); } } } async cancel() { if (this.cancelling_ || this.state() == 'idle') return; // Stop queue but don't set it to null as it may be used to // retrieve the last few downloads. if (this.downloadQueue_) this.downloadQueue_.stop(); this.logSyncOperation('cancelling', null, null, ''); this.cancelling_ = true; return new Promise((resolve) => { const iid = shim.setInterval(() => { if (this.state() == 'idle') { shim.clearInterval(iid); resolve(null); } }, 100); }); } cancelling() { return this.cancelling_; } logLastRequests() { const lastRequests = this.api().lastRequests(); if (!lastRequests || !lastRequests.length) return; for (const r of lastRequests) { const timestamp = time.unixMsToLocalHms(r.timestamp); logger.info(`Req ${timestamp}: ${r.request}`); logger.info(`Res ${timestamp}: ${r.response}`); } } static stateToLabel(state: string) { if (state === 'idle') return _('Idle'); if (state === 'in_progress') return _('In progress'); return state; } isFullSync(steps: string[]) { return steps.includes('update_remote') && steps.includes('delete_remote') && steps.includes('delta'); } async lockErrorStatus_() { const hasActiveExclusiveLock = await this.lockHandler().hasActiveLock(LockType.Exclusive); if (hasActiveExclusiveLock) return 'hasExclusiveLock'; const hasActiveSyncLock = await this.lockHandler().hasActiveLock(LockType.Sync, this.appType_, this.clientId_); if (!hasActiveSyncLock) return 'syncLockGone'; return ''; } private async setPpkIfNotExist(localInfo: SyncInfo, remoteInfo: SyncInfo) { if (localInfo.ppk || remoteInfo.ppk) return localInfo; const password = getMasterPassword(false); if (!password) return localInfo; localInfo.ppk = await generateKeyPair(this.encryptionService(), password); return localInfo; } private async apiCall(fnName: string, ...args: any[]) { if (this.syncTargetIsLocked_) throw new JoplinError('Sync target is locked - aborting API call', 'lockError'); try { const output = await (this.api() as any)[fnName](...args); return output; } catch (error) { const lockStatus = await this.lockErrorStatus_(); // When there's an error due to a lock, we re-wrap the error and change the error code so that error handling // does not do special processing on the original error. For example, if a resource could not be downloaded, // don't mark it as a "cannotSyncItem" since we don't know that. if (lockStatus) { throw new JoplinError(`Sync target lock error: ${lockStatus}. Original error was: ${error.message}`, 'lockError'); } else { throw error; } } } // Synchronisation is done in three major steps: // // 1. UPLOAD: Send to the sync target the items that have changed since the last sync. // 2. DELETE_REMOTE: Delete on the sync target, the items that have been deleted locally. // 3. DELTA: Find on the sync target the items that have been modified or deleted and apply the changes locally. public async start(options: any = null) { if (!options) options = {}; if (this.state() != 'idle') { const error: any = new Error(sprintf('Synchronisation is already in progress. State: %s', this.state())); error.code = 'alreadyStarted'; throw error; } this.state_ = 'in_progress'; this.onProgress_ = options.onProgress ? options.onProgress : function() {}; this.progressReport_ = { errors: [] }; const lastContext = options.context ? options.context : {}; const syncSteps = options.syncSteps ? options.syncSteps : ['update_remote', 'delete_remote', 'delta']; // The default is to log errors, but when testing it's convenient to be able to catch and verify errors const throwOnError = options.throwOnError === true; const syncTargetId = this.api().syncTargetId(); this.syncTargetIsLocked_ = false; this.cancelling_ = false; // const masterKeysBefore = await MasterKey.count(); // let hasAutoEnabledEncryption = false; const synchronizationId = time.unixMs().toString(); const outputContext = Object.assign({}, lastContext); this.progressReport_.startTime = time.unixMs(); this.dispatch({ type: 'SYNC_STARTED' }); eventManager.emit('syncStart'); this.logSyncOperation('starting', null, null, `Starting synchronisation to target ${syncTargetId}... supportsAccurateTimestamp = ${this.api().supportsAccurateTimestamp}; supportsMultiPut = ${this.api().supportsMultiPut} [${synchronizationId}]`); const handleCannotSyncItem = async (ItemClass: any, syncTargetId: any, item: any, cannotSyncReason: string, itemLocation: any = null) => { await ItemClass.saveSyncDisabled(syncTargetId, item, cannotSyncReason, itemLocation); this.dispatch({ type: 'SYNC_HAS_DISABLED_SYNC_ITEMS' }); }; const resourceRemotePath = (resourceId: string) => { return `${Dirnames.Resources}/${resourceId}`; }; // We index resources and apply the "is_shared" flag before syncing // because it's going to affect what's sent encrypted, and what's sent // plain text. try { if (this.resourceService()) { logger.info('Indexing resources...'); await this.resourceService().indexNoteResources(); } } catch (error) { logger.error('Error indexing resources:', error); } // Before synchronising make sure all share_id properties are set // correctly so as to share/unshare the right items. await Folder.updateAllShareIds(); const itemUploader = new ItemUploader(this.api(), this.apiCall); let errorToThrow = null; let syncLock = null; try { await this.api().initialize(); this.api().setTempDirName(Dirnames.Temp); try { let remoteInfo = await fetchSyncInfo(this.api()); logger.info('Sync target remote info:', remoteInfo); if (!remoteInfo.version) { logger.info('Sync target is new - setting it up...'); await this.migrationHandler().upgrade(Setting.value('syncVersion')); remoteInfo = await fetchSyncInfo(this.api()); } logger.info('Sync target is already setup - checking it...'); await this.migrationHandler().checkCanSync(remoteInfo); let localInfo = await localSyncInfo(); logger.info('Sync target local info:', localInfo); localInfo = await this.setPpkIfNotExist(localInfo, remoteInfo); // console.info('LOCAL', localInfo); // console.info('REMOTE', remoteInfo); if (!syncInfoEquals(localInfo, remoteInfo)) { const newInfo = mergeSyncInfos(localInfo, remoteInfo); const previousE2EE = localInfo.e2ee; logger.info('Sync target info differs between local and remote - merging infos: ', newInfo.toObject()); await this.lockHandler().acquireLock(LockType.Exclusive, this.appType_, this.clientId_, { clearExistingSyncLocksFromTheSameClient: true }); await uploadSyncInfo(this.api(), newInfo); await saveLocalSyncInfo(newInfo); await this.lockHandler().releaseLock(LockType.Exclusive, this.appType_, this.clientId_); // console.info('NEW', newInfo); if (newInfo.e2ee !== previousE2EE) { if (newInfo.e2ee) { const mk = getActiveMasterKey(newInfo); await setupAndEnableEncryption(this.encryptionService(), mk); } else { await setupAndDisableEncryption(this.encryptionService()); } } } else { // Set it to remote anyway so that timestamps are the same // Note: that's probably not needed anymore? // await uploadSyncInfo(this.api(), remoteInfo); } } catch (error) { if (error.code === 'outdatedSyncTarget') { Setting.setValue('sync.upgradeState', Setting.SYNC_UPGRADE_STATE_SHOULD_DO); } throw error; } syncLock = await this.lockHandler().acquireLock(LockType.Sync, this.appType_, this.clientId_); this.lockHandler().startAutoLockRefresh(syncLock, (error: any) => { logger.warn('Could not refresh lock - cancelling sync. Error was:', error); this.syncTargetIsLocked_ = true; void this.cancel(); }); // ======================================================================== // 2. DELETE_REMOTE // ------------------------------------------------------------------------ // Delete the remote items that have been deleted locally. // ======================================================================== if (syncSteps.indexOf('delete_remote') >= 0) { const deletedItems = await BaseItem.deletedItems(syncTargetId); for (let i = 0; i < deletedItems.length; i++) { if (this.cancelling()) break; const item = deletedItems[i]; const path = BaseItem.systemPath(item.item_id); this.logSyncOperation('deleteRemote', null, { id: item.item_id }, 'local has been deleted'); await this.apiCall('delete', path); if (item.item_type === BaseModel.TYPE_RESOURCE) { const remoteContentPath = resourceRemotePath(item.item_id); await this.apiCall('delete', remoteContentPath); } await BaseItem.remoteDeletedItem(syncTargetId, item.item_id); } } // DELETE_REMOTE STEP // ======================================================================== // 1. UPLOAD // ------------------------------------------------------------------------ // First, find all the items that have been changed since the // last sync and apply the changes to remote. // ======================================================================== if (syncSteps.indexOf('update_remote') >= 0) { const donePaths: string[] = []; const completeItemProcessing = (path: string) => { donePaths.push(path); }; while (true) { if (this.cancelling()) break; const result = await BaseItem.itemsThatNeedSync(syncTargetId); const locals = result.items; await itemUploader.preUploadItems(result.items.filter((it: any) => result.neverSyncedItemIds.includes(it.id))); for (let i = 0; i < locals.length; i++) { if (this.cancelling()) break; let local = locals[i]; const ItemClass = BaseItem.itemClass(local); const path = BaseItem.systemPath(local); // Safety check to avoid infinite loops. // - In fact this error is possible if the item is marked for sync (via sync_time or force_sync) while synchronisation is in // progress. In that case exit anyway to be sure we aren't in a loop and the item will be re-synced next time. // - It can also happen if the item is directly modified in the sync target, and set with an update_time in the future. In that case, // the local sync_time will be updated to Date.now() but on the next loop it will see that the remote item still has a date ahead // and will see a conflict. There's currently no automatic fix for this - the remote item on the sync target must be fixed manually // (by setting an updated_time less than current time). if (donePaths.indexOf(path) >= 0) throw new JoplinError(sprintf('Processing a path that has already been done: %s. sync_time was not updated? Remote item has an updated_time in the future?', path), 'processingPathTwice'); const remote: RemoteItem = result.neverSyncedItemIds.includes(local.id) ? null : await this.apiCall('stat', path); let action = null; let reason = ''; let remoteContent = null; const getConflictType = (conflictedItem: any) => { if (conflictedItem.type_ === BaseModel.TYPE_NOTE) return 'noteConflict'; if (conflictedItem.type_ === BaseModel.TYPE_RESOURCE) return 'resourceConflict'; return 'itemConflict'; }; if (!remote) { if (!local.sync_time) { action = 'createRemote'; reason = 'remote does not exist, and local is new and has never been synced'; } else { // Note or item was modified after having been deleted remotely // "itemConflict" is for all the items except the notes, which are dealt with in a special way action = getConflictType(local); reason = 'remote has been deleted, but local has changes'; } } else { // Note: in order to know the real updated_time value, we need to load the content. In theory we could // rely on the file timestamp (in remote.updated_time) but in practice it's not accurate enough and // can lead to conflicts (for example when the file timestamp is slightly ahead of it's real // updated_time). updated_time is set and managed by clients so it's always accurate. // Same situation below for updateLocal. // // This is a bit inefficient because if the resulting action is "updateRemote" we don't need the whole // content, but for now that will do since being reliable is the priority. // // Note: assuming a particular sync target is guaranteed to have accurate timestamps, the driver maybe // could expose this with a accurateTimestamps() method that returns "true". In that case, the test // could be done using the file timestamp and the potentially unnecessary content loading could be skipped. // OneDrive does not appear to have accurate timestamps as lastModifiedDateTime would occasionally be // a few seconds ahead of what it was set with setTimestamp() try { remoteContent = await this.apiCall('get', path); } catch (error) { if (error.code === 'rejectedByTarget') { this.progressReport_.errors.push(error); logger.warn(`Rejected by target: ${path}: ${error.message}`); completeItemProcessing(path); continue; } else { throw error; } } if (!remoteContent) throw new Error(`Got metadata for path but could not fetch content: ${path}`); remoteContent = await BaseItem.unserialize(remoteContent); if (remoteContent.updated_time > local.sync_time) { // Since, in this loop, we are only dealing with items that require sync, if the // remote has been modified after the sync time, it means both items have been // modified and so there's a conflict. action = getConflictType(local); reason = 'both remote and local have changes'; } else { action = 'updateRemote'; reason = 'local has changes'; } } // We no longer upload Master Keys however we keep them // in the database for extra safety. In a future // version, once it's confirmed that the new E2EE system // works well, we can delete them. if (local.type_ === ModelType.MasterKey) action = null; this.logSyncOperation(action, local, remote, reason); if (local.type_ == BaseModel.TYPE_RESOURCE && (action == 'createRemote' || action === 'updateRemote')) { const localState = await Resource.localState(local.id); if (localState.fetch_status !== Resource.FETCH_STATUS_DONE) { // This condition normally shouldn't happen // because the normal cases are as follow: // // - User creates a resource locally - in that // case the fetch status is DONE, so we cannot // end up here. // // - User fetches a new resource metadata, but // not the blob - in that case fetch status is // IDLE. However in that case, we cannot end // up in this place either, because the action // cannot be createRemote (because the // resource has not been created locally) or // updateRemote (because a resouce cannot be // modified locally unless the blob is present // too). // // Possibly the only case we can end up here is // if a resource metadata has been downloaded, // but not the blob yet. Then the sync target is // switched to a different one. In that case, we // can have a fetch status IDLE, with an // "updateRemote" action, if the timestamp of // the server resource is before the timestamp // of the local resource. // // In that case we can't do much so we mark the // resource as "cannot sync". Otherwise it will // throw the error "Processing a path that has // already been done" on the next loop, and sync // will never finish because we'll always end up // here. logger.info(`Need to upload a resource, but blob is not present: ${path}`); await handleCannotSyncItem(ItemClass, syncTargetId, local, 'Trying to upload resource, but only metadata is present.'); action = null; } else { try { const remoteContentPath = resourceRemotePath(local.id); const result = await Resource.fullPathForSyncUpload(local); const resource = result.resource; local = resource as any; const localResourceContentPath = result.path; if (resource.size >= 10 * 1000 * 1000) { logger.warn(`Uploading a large resource (resourceId: ${local.id}, size:${resource.size} bytes) which may tie up the sync process.`); } await this.apiCall('put', remoteContentPath, null, { path: localResourceContentPath, source: 'file', shareId: resource.share_id }); } catch (error) { if (isCannotSyncError(error)) { await handleCannotSyncItem(ItemClass, syncTargetId, local, error.message); action = null; } else { throw error; } } } } if (action == 'createRemote' || action == 'updateRemote') { let canSync = true; try { if (this.testingHooks_.indexOf('notesRejectedByTarget') >= 0 && local.type_ === BaseModel.TYPE_NOTE) throw new JoplinError('Testing rejectedByTarget', 'rejectedByTarget'); await itemUploader.serializeAndUploadItem(ItemClass, path, local); } catch (error) { if (error && error.code === 'rejectedByTarget') { await handleCannotSyncItem(ItemClass, syncTargetId, local, error.message); canSync = false; } else { throw error; } } // Note: Currently, we set sync_time to update_time, which should work fine given that the resolution is the millisecond. // In theory though, this could happen: // // 1. t0: Editor: Note is modified // 2. t0: Sync: Found that note was modified so start uploading it // 3. t0: Editor: Note is modified again // 4. t1: Sync: Note has finished uploading, set sync_time to t0 // // Later any attempt to sync will not detect that note was modified in (3) (within the same millisecond as it was being uploaded) // because sync_time will be t0 too. // // The solution would be to use something like an etag (a simple counter incremented on every change) to make sure each // change is uniquely identified. Leaving it like this for now. if (canSync) { // 2018-01-21: Setting timestamp is not needed because the delta() logic doesn't rely // on it (instead it uses a more reliable `context` object) and the itemsThatNeedSync loop // above also doesn't use it because it fetches the whole remote object and read the // more reliable 'updated_time' property. Basically remote.updated_time is deprecated. await ItemClass.saveSyncTime(syncTargetId, local, local.updated_time); } } else if (action == 'itemConflict') { // ------------------------------------------------------------------------------ // For non-note conflicts, we take the remote version (i.e. the version that was // synced first) and overwrite the local content. // ------------------------------------------------------------------------------ if (remote) { local = remoteContent; const syncTimeQueries = BaseItem.updateSyncTimeQueries(syncTargetId, local, time.unixMs()); await ItemClass.save(local, { autoTimestamp: false, changeSource: ItemChange.SOURCE_SYNC, nextQueries: syncTimeQueries }); } else { await ItemClass.delete(local.id, { changeSource: ItemChange.SOURCE_SYNC }); } } else if (action == 'noteConflict') { // ------------------------------------------------------------------------------ // First find out if the conflict matters. For example, if the conflict is on the title or body // we want to preserve all the changes. If it's on todo_completed it doesn't really matter // so in this case we just take the remote content. // ------------------------------------------------------------------------------ let mustHandleConflict = true; if (remoteContent) { mustHandleConflict = Note.mustHandleConflict(local, remoteContent); } // ------------------------------------------------------------------------------ // Create a duplicate of local note into Conflicts folder // (to preserve the user's changes) // ------------------------------------------------------------------------------ if (mustHandleConflict) { await Note.createConflictNote(local, ItemChange.SOURCE_SYNC); } } else if (action == 'resourceConflict') { // ------------------------------------------------------------------------------ // Unlike notes we always handle the conflict for resources // ------------------------------------------------------------------------------ await Resource.createConflictResourceNote(local); if (remote) { // The local content we have is no longer valid and should be re-downloaded await Resource.setLocalState(local.id, { fetch_status: Resource.FETCH_STATUS_IDLE, }); } } if (['noteConflict', 'resourceConflict'].includes(action)) { // ------------------------------------------------------------------------------ // For note and resource conflicts, the creation of the conflict item is done // differently. However the way the local content is handled is the same. // Either copy the remote content to local or, if the remote content has // been deleted, delete the local content. // ------------------------------------------------------------------------------ if (remote) { local = remoteContent; const syncTimeQueries = BaseItem.updateSyncTimeQueries(syncTargetId, local, time.unixMs()); await ItemClass.save(local, { autoTimestamp: false, changeSource: ItemChange.SOURCE_SYNC, nextQueries: syncTimeQueries }); if (local.encryption_applied) this.dispatch({ type: 'SYNC_GOT_ENCRYPTED_ITEM' }); } else { // Remote no longer exists (note deleted) so delete local one too await ItemClass.delete(local.id, { changeSource: ItemChange.SOURCE_SYNC }); } } completeItemProcessing(path); } if (!result.hasMore) break; } } // UPLOAD STEP // ------------------------------------------------------------------------ // 3. DELTA // ------------------------------------------------------------------------ // Loop through all the remote items, find those that // have been created or updated, and apply the changes to local. // ------------------------------------------------------------------------ if (this.downloadQueue_) await this.downloadQueue_.stop(); this.downloadQueue_ = new TaskQueue('syncDownload'); this.downloadQueue_.logger_ = logger; if (syncSteps.indexOf('delta') >= 0) { // At this point all the local items that have changed have been pushed to remote // or handled as conflicts, so no conflict is possible after this. let context = null; let newDeltaContext = null; const localFoldersToDelete = []; let hasCancelled = false; if (lastContext.delta) context = lastContext.delta; while (true) { if (this.cancelling() || hasCancelled) break; const listResult: any = await this.apiCall('delta', '', { context: context, // allItemIdsHandler() provides a way for drivers that don't have a delta API to // still provide delta functionality by comparing the items they have to the items // the client has. Very inefficient but that's the only possible workaround. // It's a function so that it is only called if the driver needs these IDs. For // drivers with a delta functionality it's a noop. allItemIdsHandler: async () => { return BaseItem.syncedItemIds(syncTargetId); }, wipeOutFailSafe: Setting.value('sync.wipeOutFailSafe'), logger: logger, }); const remotes: RemoteItem[] = listResult.items; this.logSyncOperation('fetchingTotal', null, null, 'Fetching delta items from sync target', remotes.length); const remoteIds = remotes.map(r => BaseItem.pathToId(r.path)); const locals = await BaseItem.loadItemsByIds(remoteIds); for (const remote of remotes) { if (this.cancelling()) break; let needsToDownload = true; if (this.api().supportsAccurateTimestamp) { const local = locals.find(l => l.id === BaseItem.pathToId(remote.path)); if (local && local.updated_time === remote.jop_updated_time) needsToDownload = false; } if (needsToDownload) { this.downloadQueue_.push(remote.path, async () => { return this.apiCall('get', remote.path); }); } } for (let i = 0; i < remotes.length; i++) { if (this.cancelling() || this.testingHooks_.indexOf('cancelDeltaLoop2') >= 0) { hasCancelled = true; break; } this.logSyncOperation('fetchingProcessed', null, null, 'Processing fetched item'); const remote = remotes[i]; if (!BaseItem.isSystemPath(remote.path)) continue; // The delta API might return things like the .sync, .resource or the root folder const loadContent = async () => { const task = await this.downloadQueue_.waitForResult(path); if (task.error) throw task.error; if (!task.result) return null; return await BaseItem.unserialize(task.result); }; const path = remote.path; const remoteId = BaseItem.pathToId(path); let action = null; let reason = ''; let local = locals.find(l => l.id === remoteId); let ItemClass = null; let content = null; try { if (!local) { if (remote.isDeleted !== true) { action = 'createLocal'; reason = 'remote exists but local does not'; content = await loadContent(); ItemClass = content ? BaseItem.itemClass(content) : null; } } else { ItemClass = BaseItem.itemClass(local); local = ItemClass.filter(local); if (remote.isDeleted) { action = 'deleteLocal'; reason = 'remote has been deleted'; } else { if (this.api().supportsAccurateTimestamp && remote.jop_updated_time === local.updated_time) { // Nothing to do, and no need to fetch the content } else { content = await loadContent(); if (content && content.updated_time > local.updated_time) { action = 'updateLocal'; reason = 'remote is more recent than local'; } } } } } catch (error) { if (error.code === 'rejectedByTarget') { this.progressReport_.errors.push(error); logger.warn(`Rejected by target: ${path}: ${error.message}`); action = null; } else { error.message = `On file ${path}: ${error.message}`; throw error; } } if (this.testingHooks_.indexOf('skipRevisions') >= 0 && content && content.type_ === BaseModel.TYPE_REVISION) action = null; if (!action) continue; this.logSyncOperation(action, local, remote, reason); if (action == 'createLocal' || action == 'updateLocal') { if (content === null) { logger.warn(`Remote has been deleted between now and the delta() call? In that case it will be handled during the next sync: ${path}`); continue; } content = ItemClass.filter(content); // 2017-12-03: This was added because the new user_updated_time and user_created_time properties were added // to the items. However changing the database is not enough since remote items that haven't been synced yet // will not have these properties and, since they are required, it would cause a problem. So this check // if they are present and, if not, set them to a reasonable default. // Let's leave these two lines for 6 months, by which time all the clients should have been synced. if (!content.user_updated_time) content.user_updated_time = content.updated_time; if (!content.user_created_time) content.user_created_time = content.created_time; const options: any = { autoTimestamp: false, nextQueries: BaseItem.updateSyncTimeQueries(syncTargetId, content, time.unixMs()), changeSource: ItemChange.SOURCE_SYNC, }; if (action == 'createLocal') options.isNew = true; if (action == 'updateLocal') options.oldItem = local; const creatingOrUpdatingResource = content.type_ == BaseModel.TYPE_RESOURCE && (action == 'createLocal' || action == 'updateLocal'); if (creatingOrUpdatingResource) { if (content.size >= this.maxResourceSize()) { await handleCannotSyncItem(ItemClass, syncTargetId, content, `File "${content.title}" is larger than allowed ${this.maxResourceSize()} bytes. Beyond this limit, the mobile app would crash.`, BaseItem.SYNC_ITEM_LOCATION_REMOTE); continue; } await ResourceLocalState.save({ resource_id: content.id, fetch_status: Resource.FETCH_STATUS_IDLE }); } if (content.type_ === ModelType.MasterKey) { // Special case for master keys - if we download // one, we only add it to the store if it's not // already there. That can happen for example if // the new E2EE migration was processed at a // time a master key was still on the sync // target. In that case, info.json would not // have it. // // If info.json already has the key we shouldn't // update because the most up to date keys // should always be in info.json now. const existingMasterKey = await MasterKey.load(content.id); if (!existingMasterKey) { logger.info(`Downloaded a master key that was not in info.json - adding it to the store. ID: ${content.id}`); await MasterKey.save(content); } } else { await ItemClass.save(content, options); } if (creatingOrUpdatingResource) this.dispatch({ type: 'SYNC_CREATED_OR_UPDATED_RESOURCE', id: content.id }); // if (!hasAutoEnabledEncryption && content.type_ === BaseModel.TYPE_MASTER_KEY && !masterKeysBefore) { // hasAutoEnabledEncryption = true; // logger.info('One master key was downloaded and none was previously available: automatically enabling encryption'); // logger.info('Using master key: ', content.id); // await this.encryptionService().enableEncryption(content); // await this.encryptionService().loadMasterKeysFromSettings(); // logger.info('Encryption has been enabled with downloaded master key as active key. However, note that no password was initially supplied. It will need to be provided by user.'); // } if (content.encryption_applied) this.dispatch({ type: 'SYNC_GOT_ENCRYPTED_ITEM' }); } else if (action == 'deleteLocal') { if (local.type_ == BaseModel.TYPE_FOLDER) { localFoldersToDelete.push(local); continue; } const ItemClass = BaseItem.itemClass(local.type_); await ItemClass.delete(local.id, { trackDeleted: false, changeSource: ItemChange.SOURCE_SYNC }); } } // If user has cancelled, don't record the new context (2) so that synchronisation // can start again from the previous context (1) next time. It is ok if some items // have been synced between (1) and (2) because the loop above will handle the same // items being synced twice as an update. If the local and remote items are identical // the update will simply be skipped. if (!hasCancelled) { if (options.saveContextHandler) { const deltaToSave = Object.assign({}, listResult.context); // Remove these two variables because they can be large and can be rebuilt // the next time the sync is started. delete deltaToSave.statsCache; delete deltaToSave.statIdsCache; options.saveContextHandler({ delta: deltaToSave }); } if (!listResult.hasMore) { newDeltaContext = listResult.context; break; } context = listResult.context; } } outputContext.delta = newDeltaContext ? newDeltaContext : lastContext.delta; // ------------------------------------------------------------------------ // Delete the folders that have been collected in the loop above. // Folders are always deleted last, and only if they are empty. // If they are not empty it's considered a conflict since whatever deleted // them should have deleted their content too. In that case, all its notes // are marked as "is_conflict". // ------------------------------------------------------------------------ if (!this.cancelling()) { for (let i = 0; i < localFoldersToDelete.length; i++) { const item = localFoldersToDelete[i]; const noteIds = await Folder.noteIds(item.id); if (noteIds.length) { // CONFLICT await Folder.markNotesAsConflict(item.id); } await Folder.delete(item.id, { deleteChildren: false, changeSource: ItemChange.SOURCE_SYNC, trackDeleted: false }); } } if (!this.cancelling()) { await BaseItem.deleteOrphanSyncItems(); } } // DELTA STEP } catch (error) { if (throwOnError) { errorToThrow = error; } else if (error && ['cannotEncryptEncrypted', 'noActiveMasterKey', 'processingPathTwice', 'failSafe', 'lockError', 'outdatedSyncTarget'].indexOf(error.code) >= 0) { // Only log an info statement for this since this is a common condition that is reported // in the application, and needs to be resolved by the user. // Or it's a temporary issue that will be resolved on next sync. logger.info(error.message); if (error.code === 'failSafe' || error.code === 'lockError') { // Get the message to display on UI, but not in testing to avoid poluting stdout if (!shim.isTestingEnv()) this.progressReport_.errors.push(error.message); this.logLastRequests(); } } else if (error.code === 'unknownItemType') { this.progressReport_.errors.push(_('Unknown item type downloaded - please upgrade Joplin to the latest version')); logger.error(error); } else { logger.error(error); if (error.details) logger.error('Details:', error.details); // Don't save to the report errors that are due to things like temporary network errors or timeout. if (!shim.fetchRequestCanBeRetried(error)) { this.progressReport_.errors.push(error); this.logLastRequests(); } } } if (syncLock) { this.lockHandler().stopAutoLockRefresh(syncLock); await this.lockHandler().releaseLock(LockType.Sync, this.appType_, this.clientId_); } this.syncTargetIsLocked_ = false; if (this.cancelling()) { logger.info('Synchronisation was cancelled.'); this.cancelling_ = false; } // After syncing, we run the share service maintenance, which is going // to fetch share invitations, if any. if (this.shareService_) { try { await this.shareService_.maintenance(); } catch (error) { logger.error('Could not run share service maintenance:', error); } } this.progressReport_.completedTime = time.unixMs(); this.logSyncOperation('finished', null, null, `Synchronisation finished [${synchronizationId}]`); await this.logSyncSummary(this.progressReport_); eventManager.emit('syncComplete', { withErrors: Synchronizer.reportHasErrors(this.progressReport_), }); this.onProgress_ = function() {}; this.progressReport_ = {}; this.dispatch({ type: 'SYNC_COMPLETED', isFullSync: this.isFullSync(syncSteps) }); this.state_ = 'idle'; if (errorToThrow) throw errorToThrow; return outputContext; } }