Stackedit/src/services/syncSvc.js

822 lines
30 KiB
JavaScript
Raw Normal View History

2017-07-28 07:40:24 +00:00
import localDbSvc from './localDbSvc';
import store from '../store';
import utils from './utils';
2017-08-25 10:37:46 +00:00
import diffUtils from './diffUtils';
2017-12-11 00:53:46 +00:00
import networkSvc from './networkSvc';
2018-04-27 14:37:05 +00:00
import providerRegistry from './providers/common/providerRegistry';
2017-12-10 23:49:20 +00:00
import googleDriveAppDataProvider from './providers/googleDriveAppDataProvider';
2018-01-24 07:31:54 +00:00
import './providers/couchdbWorkspaceProvider';
2018-04-27 14:37:05 +00:00
import './providers/githubWorkspaceProvider';
import './providers/googleDriveWorkspaceProvider';
2018-03-12 00:45:54 +00:00
import tempFileSvc from './tempFileSvc';
2018-05-04 18:07:28 +00:00
import fileSvc from './fileSvc';
2017-08-15 10:43:26 +00:00
2018-04-27 14:37:05 +00:00
const minAutoSyncEvery = 60 * 1000; // 60 sec
2017-08-15 10:43:26 +00:00
const inactivityThreshold = 3 * 1000; // 3 sec
2017-08-17 23:10:35 +00:00
const restartSyncAfter = 30 * 1000; // 30 sec
2018-04-27 14:37:05 +00:00
const restartContentSyncAfter = 500; // Restart if an authorize window pops up
2018-04-08 10:18:56 +00:00
const maxContentHistory = 20;
2017-08-25 10:37:46 +00:00
2018-04-27 14:37:05 +00:00
const LAST_SEEN = 0;
const LAST_MERGED = 1;
const LAST_SENT = 2;
2018-01-04 20:19:10 +00:00
let actionProvider;
let workspaceProvider;
2017-12-10 23:49:20 +00:00
/**
* Use a lock in the local storage to prevent multiple windows concurrency.
*/
let lastSyncActivity;
2017-12-11 00:53:46 +00:00
const getLastStoredSyncActivity = () =>
parseInt(localStorage.getItem(store.getters['workspace/lastSyncActivityKey']), 10) || 0;
2017-12-10 23:49:20 +00:00
/**
* Return true if workspace sync is possible.
*/
2017-12-17 15:08:52 +00:00
const isWorkspaceSyncPossible = () => !!store.getters['workspace/syncToken'];
2017-12-10 23:49:20 +00:00
/**
* Return true if file has at least one explicit sync location.
*/
2017-08-25 10:37:46 +00:00
const hasCurrentFileSyncLocations = () => !!store.getters['syncLocation/current'].length;
2017-12-10 23:49:20 +00:00
/**
* Return true if we are online and we have something to sync.
*/
2017-08-25 10:37:46 +00:00
const isSyncPossible = () => !store.state.offline &&
2017-12-10 23:49:20 +00:00
(isWorkspaceSyncPossible() || hasCurrentFileSyncLocations());
2017-08-15 10:43:26 +00:00
2017-12-10 23:49:20 +00:00
/**
* Return true if we are the many window, ie we have the lastSyncActivity lock.
*/
2017-08-15 10:43:26 +00:00
function isSyncWindow() {
2017-10-07 11:22:24 +00:00
const storedLastSyncActivity = getLastStoredSyncActivity();
2017-08-15 10:43:26 +00:00
return lastSyncActivity === storedLastSyncActivity ||
Date.now() > inactivityThreshold + storedLastSyncActivity;
}
2017-12-10 23:49:20 +00:00
/**
2017-12-11 00:53:46 +00:00
* Return true if auto sync can start, ie if lastSyncActivity is old enough.
2017-12-10 23:49:20 +00:00
*/
2017-08-17 23:10:35 +00:00
function isAutoSyncReady() {
2018-05-06 00:46:33 +00:00
let { autoSyncEvery } = store.getters['data/computedSettings'];
2017-12-17 15:08:52 +00:00
if (autoSyncEvery < minAutoSyncEvery) {
autoSyncEvery = minAutoSyncEvery;
}
return Date.now() > autoSyncEvery + getLastStoredSyncActivity();
2017-08-15 10:43:26 +00:00
}
2017-12-10 23:49:20 +00:00
/**
* Update the lastSyncActivity, assuming we have the lock.
*/
2017-08-15 10:43:26 +00:00
function setLastSyncActivity() {
const currentDate = Date.now();
lastSyncActivity = currentDate;
2017-12-11 00:53:46 +00:00
localStorage.setItem(store.getters['workspace/lastSyncActivityKey'], currentDate);
2017-08-15 10:43:26 +00:00
}
2018-04-27 14:37:05 +00:00
/**
* Upgrade hashes if syncedContent is from an old version
*/
function upgradeSyncedContent(syncedContent) {
if (syncedContent.v) {
return syncedContent;
}
const hashUpgrades = {};
const historyData = {};
const syncHistory = {};
Object.entries(syncedContent.historyData).forEach(([hash, content]) => {
const newContent = utils.addItemHash(content);
historyData[newContent.hash] = newContent;
hashUpgrades[hash] = newContent.hash;
});
Object.entries(syncedContent.syncHistory).forEach(([id, hashEntries]) => {
syncHistory[id] = hashEntries.map(hash => hashUpgrades[hash]);
});
return {
...syncedContent,
historyData,
syncHistory,
v: 1,
};
}
2017-12-10 23:49:20 +00:00
/**
* Clean a syncedContent.
*/
2017-08-25 10:37:46 +00:00
function cleanSyncedContent(syncedContent) {
// Clean syncHistory from removed syncLocations
Object.keys(syncedContent.syncHistory).forEach((syncLocationId) => {
if (syncLocationId !== 'main' && !store.state.syncLocation.itemMap[syncLocationId]) {
delete syncedContent.syncHistory[syncLocationId];
}
});
2018-05-06 00:46:33 +00:00
const allSyncLocationHashSet = new Set([]
.concat(...Object.keys(syncedContent.syncHistory)
.map(id => syncedContent.syncHistory[id])));
2017-08-25 10:37:46 +00:00
// Clean historyData from unused contents
2018-05-06 00:46:33 +00:00
Object.keys(syncedContent.historyData)
.map(hash => parseInt(hash, 10))
.forEach((hash) => {
if (!allSyncLocationHashSet.has(hash)) {
delete syncedContent.historyData[hash];
}
});
2017-08-25 10:37:46 +00:00
}
2017-12-10 23:49:20 +00:00
/**
* Apply changes retrieved from the main provider. Update sync data accordingly.
*/
2017-08-17 23:10:35 +00:00
function applyChanges(changes) {
const storeItemMap = { ...store.getters.allItemMap };
const syncData = { ...store.getters['data/syncData'] };
2017-12-23 18:25:14 +00:00
let saveSyncData = false;
2017-08-17 23:10:35 +00:00
changes.forEach((change) => {
2017-12-17 15:08:52 +00:00
const existingSyncData = syncData[change.syncDataId];
2017-10-05 07:18:21 +00:00
const existingItem = existingSyncData && storeItemMap[existingSyncData.itemId];
2017-12-23 18:25:14 +00:00
if (!change.item && existingSyncData) {
// Item was removed
2018-04-27 14:37:05 +00:00
if (syncData[change.syncDataId]) {
delete syncData[change.syncDataId];
saveSyncData = true;
}
2017-10-05 07:18:21 +00:00
if (existingItem) {
// Remove object from the store
store.commit(`${existingItem.type}/deleteItem`, existingItem.id);
delete storeItemMap[existingItem.id];
}
2017-12-23 18:25:14 +00:00
} else if (change.item && change.item.hash) {
// Item was modifed
2018-04-27 14:37:05 +00:00
if ((existingSyncData || {}).hash !== change.syncData.hash) {
syncData[change.syncDataId] = change.syncData;
saveSyncData = true;
}
2017-12-23 18:25:14 +00:00
if (
// If no sync data or existing one is different
(existingSyncData || {}).hash !== change.item.hash
// And no existing item or existing item is different
&& (existingItem || {}).hash !== change.item.hash
// And item is not content nor data, which will be merged later
&& change.item.type !== 'content' && change.item.type !== 'data'
) {
store.commit(`${change.item.type}/setItem`, change.item);
storeItemMap[change.item.id] = change.item;
}
2017-08-17 23:10:35 +00:00
}
});
2017-12-23 18:25:14 +00:00
if (saveSyncData) {
2017-08-17 23:10:35 +00:00
store.dispatch('data/setSyncData', syncData);
}
}
2017-12-10 23:49:20 +00:00
/**
* Create a sync location by uploading the current file content.
*/
2017-09-23 19:01:50 +00:00
function createSyncLocation(syncLocation) {
syncLocation.id = utils.uid();
const currentFile = store.getters['file/current'];
const fileId = currentFile.id;
syncLocation.fileId = fileId;
// Use deepCopy to freeze item
const content = utils.deepCopy(store.getters['content/current']);
2018-05-06 00:46:33 +00:00
store.dispatch(
'queue/enqueue',
2017-09-23 19:01:50 +00:00
() => {
const provider = providerRegistry.providers[syncLocation.providerId];
const token = provider.getToken(syncLocation);
return provider.uploadContent(token, {
...content,
history: [content.hash],
}, syncLocation)
2017-09-27 20:27:12 +00:00
.then(syncLocationToStore => localDbSvc.loadSyncedContent(fileId)
2017-09-23 19:01:50 +00:00
.then(() => {
2018-05-06 00:46:33 +00:00
const newSyncedContent = utils.deepCopy(upgradeSyncedContent(store.state.syncedContent.itemMap[`${fileId}/syncedContent`]));
2017-09-23 19:01:50 +00:00
const newSyncHistoryItem = [];
newSyncedContent.syncHistory[syncLocation.id] = newSyncHistoryItem;
2018-04-08 10:18:56 +00:00
newSyncHistoryItem[LAST_SEEN] = content.hash;
2017-09-23 19:01:50 +00:00
newSyncHistoryItem[LAST_SENT] = content.hash;
newSyncedContent.historyData[content.hash] = content;
store.commit('syncedContent/patchItem', newSyncedContent);
store.commit('syncLocation/setItem', syncLocationToStore);
store.dispatch('notification/info', `A new synchronized location was added to "${currentFile.name}".`);
}));
2018-05-06 00:46:33 +00:00
},
);
2017-09-23 19:01:50 +00:00
}
2018-04-27 14:37:05 +00:00
// Prevent from sending new data too long after old data has been fetched
const tooLateChecker = (timeout) => {
const tooLateAfter = Date.now() + timeout;
return cb => (res) => {
if (tooLateAfter < Date.now()) {
throw new Error('TOO_LATE');
}
return cb(res);
};
};
class SyncContext {
2018-03-12 00:45:54 +00:00
restart = false;
attempted = {};
}
2018-05-06 00:46:33 +00:00
/**
2017-12-10 23:49:20 +00:00
* Sync one file with all its locations.
*/
function syncFile(fileId, syncContext = new SyncContext()) {
2018-03-12 00:45:54 +00:00
syncContext.attempted[`${fileId}/content`] = true;
2017-09-27 20:27:12 +00:00
return localDbSvc.loadSyncedContent(fileId)
.then(() => localDbSvc.loadItem(`${fileId}/content`)
.catch(() => {})) // Item may not exist if content has not been downloaded yet
2017-08-25 10:37:46 +00:00
.then(() => {
const getFile = () => store.state.file.itemMap[fileId];
2017-08-25 10:37:46 +00:00
const getContent = () => store.state.content.itemMap[`${fileId}/content`];
2018-05-06 00:46:33 +00:00
const getSyncedContent = () => upgradeSyncedContent(store.state.syncedContent.itemMap[`${fileId}/syncedContent`]);
2017-08-25 10:37:46 +00:00
const getSyncHistoryItem = syncLocationId => getSyncedContent().syncHistory[syncLocationId];
2018-03-12 00:45:54 +00:00
const isTempFile = () => {
2017-11-15 08:12:56 +00:00
if (store.getters['data/syncDataByItemId'][`${fileId}/content`]) {
2018-03-12 00:45:54 +00:00
// If file has already been synced, it's not a temp file
2017-11-15 08:12:56 +00:00
return false;
}
const file = getFile();
const content = getContent();
2017-11-15 08:12:56 +00:00
if (!file || !content) {
return false;
}
2018-03-12 00:45:54 +00:00
if (file.parentId === 'temp') {
return true;
}
const locations = [
2018-04-27 14:37:05 +00:00
...store.getters['syncLocation/filteredGroupedByFileId'][fileId] || [],
...store.getters['publishLocation/filteredGroupedByFileId'][fileId] || [],
2018-03-12 00:45:54 +00:00
];
if (locations.length) {
// If file has explicit sync/publish locations, it's not a temp file
return false;
}
// Return true if it's a welcome file that has no discussion
2018-05-06 00:46:33 +00:00
const { welcomeFileHashes } = store.getters['data/localSettings'];
2017-11-15 08:12:56 +00:00
const hash = utils.hash(content.text);
const hasDiscussions = Object.keys(content.discussions).length;
return file.name === 'Welcome file' && welcomeFileHashes[hash] && !hasDiscussions;
};
2018-03-12 00:45:54 +00:00
if (isTempFile()) {
return null;
}
const attemptedLocations = {};
2017-08-25 10:37:46 +00:00
const syncOneContentLocation = () => {
const syncLocations = [
2018-04-27 14:37:05 +00:00
...store.getters['syncLocation/filteredGroupedByFileId'][fileId] || [],
2017-08-25 10:37:46 +00:00
];
2017-12-10 23:49:20 +00:00
if (isWorkspaceSyncPossible()) {
2018-01-04 20:19:10 +00:00
syncLocations.unshift({ id: 'main', providerId: workspaceProvider.id, fileId });
2017-08-25 10:37:46 +00:00
}
let result;
syncLocations.some((syncLocation) => {
const provider = providerRegistry.providers[syncLocation.providerId];
if (
2018-03-12 00:45:54 +00:00
// Skip if it has been attempted already
!attemptedLocations[syncLocation.id] &&
// Skip temp file
!isTempFile()
2017-09-23 19:01:50 +00:00
) {
2018-03-12 00:45:54 +00:00
attemptedLocations[syncLocation.id] = true;
2017-11-15 08:12:56 +00:00
const token = provider && provider.getToken(syncLocation);
result = token && store.dispatch('queue/doWithLocation', {
2017-09-23 19:01:50 +00:00
location: syncLocation,
promise: provider.downloadContent(token, syncLocation)
.then((serverContent = null) => {
const syncedContent = getSyncedContent();
const syncHistoryItem = getSyncHistoryItem(syncLocation.id);
let mergedContent = (() => {
const clientContent = utils.deepCopy(getContent());
2017-09-27 20:27:12 +00:00
if (!clientContent) {
return utils.deepCopy(serverContent);
}
2017-09-23 19:01:50 +00:00
if (!serverContent) {
// Sync location has not been created yet
return clientContent;
}
if (serverContent.hash === clientContent.hash) {
// Server and client contents are synced
return clientContent;
}
if (syncedContent.historyData[serverContent.hash]) {
// Server content has not changed or has already been merged
return clientContent;
}
// Perform a merge with last merged content if any, or a simple fusion otherwise
let lastMergedContent;
serverContent.history.some((hash) => {
lastMergedContent = syncedContent.historyData[hash];
return lastMergedContent;
});
if (!lastMergedContent && syncHistoryItem) {
lastMergedContent = syncedContent.historyData[syncHistoryItem[LAST_MERGED]];
}
return diffUtils.mergeContent(serverContent, clientContent, lastMergedContent);
})();
2017-09-27 20:27:12 +00:00
if (!mergedContent) {
return null;
}
// Update or set content in store
store.commit('content/setItem', {
2017-09-23 19:01:50 +00:00
id: `${fileId}/content`,
text: utils.sanitizeText(mergedContent.text),
properties: utils.sanitizeText(mergedContent.properties),
discussions: mergedContent.discussions,
comments: mergedContent.comments,
2017-09-23 19:01:50 +00:00
});
2017-09-26 22:54:26 +00:00
// Retrieve content with new `hash` and freeze it
2017-09-23 19:01:50 +00:00
mergedContent = utils.deepCopy(getContent());
// Make merged content history
const mergedContentHistory = serverContent ? serverContent.history.slice() : [];
let skipUpload = true;
if (mergedContentHistory[0] !== mergedContent.hash) {
// Put merged content hash at the beginning of history
mergedContentHistory.unshift(mergedContent.hash);
// Server content is either out of sync or its history is incomplete, do upload
skipUpload = false;
2017-08-25 10:37:46 +00:00
}
2018-04-08 10:18:56 +00:00
if (syncHistoryItem
&& syncHistoryItem[LAST_SENT] != null
&& syncHistoryItem[LAST_SENT] !== mergedContent.hash
) {
2017-09-23 19:01:50 +00:00
// Clean up by removing the hash we've previously added
2018-04-08 10:18:56 +00:00
const idx = mergedContentHistory.lastIndexOf(syncHistoryItem[LAST_SENT]);
2017-09-23 19:01:50 +00:00
if (idx !== -1) {
mergedContentHistory.splice(idx, 1);
}
2017-08-25 10:37:46 +00:00
}
2017-09-23 19:01:50 +00:00
// Store last sent if it's in the server history,
// and merged content which will be sent if different
const newSyncedContent = utils.deepCopy(syncedContent);
const newSyncHistoryItem = newSyncedContent.syncHistory[syncLocation.id] || [];
newSyncedContent.syncHistory[syncLocation.id] = newSyncHistoryItem;
2018-04-08 10:18:56 +00:00
if (serverContent &&
(serverContent.hash === newSyncHistoryItem[LAST_SEEN] ||
serverContent.history.indexOf(newSyncHistoryItem[LAST_SEEN]) !== -1)
2017-09-23 19:01:50 +00:00
) {
2018-04-08 10:18:56 +00:00
// That's the 2nd time we've seen this content, trust it for future merges
newSyncHistoryItem[LAST_MERGED] = newSyncHistoryItem[LAST_SEEN];
2017-08-25 10:37:46 +00:00
}
2018-04-08 10:18:56 +00:00
newSyncHistoryItem[LAST_MERGED] = newSyncHistoryItem[LAST_MERGED] || null;
newSyncHistoryItem[LAST_SEEN] = mergedContent.hash;
newSyncHistoryItem[LAST_SENT] = skipUpload ? null : mergedContent.hash;
2017-09-23 19:01:50 +00:00
newSyncedContent.historyData[mergedContent.hash] = mergedContent;
// Clean synced content from unused revisions
cleanSyncedContent(newSyncedContent);
// Store synced content
store.commit('syncedContent/patchItem', newSyncedContent);
if (skipUpload) {
// Server content and merged content are equal, skip content upload
return null;
2017-08-25 10:37:46 +00:00
}
2017-09-23 19:01:50 +00:00
// Upload merged content
return provider.uploadContent(token, {
...mergedContent,
2018-04-08 10:18:56 +00:00
history: mergedContentHistory.slice(0, maxContentHistory),
2018-04-27 14:37:05 +00:00
}, syncLocation, tooLateChecker(restartContentSyncAfter))
2017-09-23 19:01:50 +00:00
.then((syncLocationToStore) => {
// Replace sync location if modified
if (utils.serializeObject(syncLocation) !==
utils.serializeObject(syncLocationToStore)
) {
store.commit('syncLocation/patchItem', syncLocationToStore);
}
2017-09-27 20:27:12 +00:00
// If content was just created, restart sync to create the file as well
2018-01-04 20:19:10 +00:00
if (provider === workspaceProvider &&
!store.getters['data/syncDataByItemId'][fileId]
) {
syncContext.restart = true;
2017-09-27 20:27:12 +00:00
}
2017-09-23 19:01:50 +00:00
});
})
.catch((err) => {
2018-01-30 07:36:33 +00:00
if (store.state.offline || (err && err.message === 'TOO_LATE')) {
2017-09-23 19:01:50 +00:00
throw err;
2017-08-25 10:37:46 +00:00
}
2017-09-23 19:01:50 +00:00
console.error(err); // eslint-disable-line no-console
store.dispatch('notification/error', err);
}),
})
2018-05-06 00:46:33 +00:00
.then(() => syncOneContentLocation());
2017-08-25 10:37:46 +00:00
}
return result;
});
return result;
};
return syncOneContentLocation();
})
.then(
() => localDbSvc.unloadContents(),
err => localDbSvc.unloadContents()
.then(() => {
throw err;
2018-05-06 00:46:33 +00:00
}),
)
.catch((err) => {
if (err && err.message === 'TOO_LATE') {
// Restart sync
return syncFile(fileId, syncContext);
}
throw err;
});
2017-08-25 10:37:46 +00:00
}
2017-12-10 23:49:20 +00:00
/**
2017-12-17 15:08:52 +00:00
* Sync a data item, typically settings, workspaces and templates.
2017-12-10 23:49:20 +00:00
*/
2017-09-26 22:54:26 +00:00
function syncDataItem(dataId) {
2017-12-17 15:08:52 +00:00
const getItem = () => store.state.data.itemMap[dataId]
|| store.state.data.lsItemMap[dataId];
const item = getItem();
2017-09-26 22:54:26 +00:00
const syncData = store.getters['data/syncDataByItemId'][dataId];
// Sync if item hash and syncData hash are inconsistent
if (syncData && item && item.hash === syncData.hash) {
return null;
}
2017-12-17 15:08:52 +00:00
2018-01-04 20:19:10 +00:00
return workspaceProvider.downloadData(dataId)
2017-09-26 22:54:26 +00:00
.then((serverItem = null) => {
const dataSyncData = store.getters['data/dataSyncData'][dataId];
let mergedItem = (() => {
2017-12-17 15:08:52 +00:00
const clientItem = utils.deepCopy(getItem());
2017-09-27 20:27:12 +00:00
if (!clientItem) {
return serverItem;
}
2017-09-26 22:54:26 +00:00
if (!serverItem) {
return clientItem;
}
if (!dataSyncData) {
return serverItem;
}
if (dataSyncData.hash !== serverItem.hash) {
// Server version has changed
if (dataSyncData.hash !== clientItem.hash && typeof clientItem.data === 'object') {
// Client version has changed as well, merge data objects
return {
...clientItem,
data: diffUtils.mergeObjects(serverItem.data, clientItem.data),
};
}
return serverItem;
}
return clientItem;
})();
2017-09-27 20:27:12 +00:00
if (!mergedItem) {
return null;
}
2017-09-26 22:54:26 +00:00
// Update item in store
store.commit('data/setItem', {
id: dataId,
...mergedItem,
});
// Retrieve item with new `hash` and freeze it
2017-12-17 15:08:52 +00:00
mergedItem = utils.deepCopy(getItem());
2017-09-26 22:54:26 +00:00
return Promise.resolve()
.then(() => {
if (serverItem && serverItem.hash === mergedItem.hash) {
return null;
}
2018-04-27 14:37:05 +00:00
return workspaceProvider.uploadData(mergedItem, tooLateChecker(restartContentSyncAfter));
2017-09-26 22:54:26 +00:00
})
2018-04-27 14:37:05 +00:00
.then(() => store.dispatch('data/patchDataSyncData', {
[dataId]: utils.deepCopy(store.getters['data/syncDataByItemId'][dataId]),
}));
2017-09-26 22:54:26 +00:00
});
}
2017-12-10 23:49:20 +00:00
/**
* Sync the whole workspace with the main provider and the current file explicit locations.
*/
2017-12-23 18:25:14 +00:00
function syncWorkspace() {
const workspace = store.getters['workspace/currentWorkspace'];
const syncContext = new SyncContext();
2017-12-17 15:08:52 +00:00
return Promise.resolve()
.then(() => {
// Store the sub in the DB since it's not safely stored in the token
2017-12-23 18:25:14 +00:00
const syncToken = store.getters['workspace/syncToken'];
2017-12-17 15:08:52 +00:00
const localSettings = store.getters['data/localSettings'];
if (!localSettings.syncSub) {
store.dispatch('data/patchLocalSettings', {
2017-12-23 18:25:14 +00:00
syncSub: syncToken.sub,
2017-12-17 15:08:52 +00:00
});
} else if (localSettings.syncSub !== syncToken.sub) {
throw new Error('Synchronization failed due to token inconsistency.');
}
})
2018-01-04 20:19:10 +00:00
.then(() => workspaceProvider.getChanges())
2017-08-15 10:43:26 +00:00
.then((changes) => {
2017-08-17 23:10:35 +00:00
// Apply changes
applyChanges(changes);
2018-04-27 14:37:05 +00:00
if (workspaceProvider.onChangesApplied) {
workspaceProvider.onChangesApplied();
}
2017-08-17 23:10:35 +00:00
// Prevent from sending items too long after changes have been retrieved
2018-04-27 14:37:05 +00:00
const ifNotTooLate = tooLateChecker(restartSyncAfter);
2017-08-17 23:10:35 +00:00
// Called until no item to save
const saveNextItem = ifNotTooLate(() => {
2017-08-25 10:37:46 +00:00
const storeItemMap = {
...store.state.file.itemMap,
...store.state.folder.itemMap,
...store.state.syncLocation.itemMap,
2017-09-23 19:01:50 +00:00
...store.state.publishLocation.itemMap,
2017-09-26 22:54:26 +00:00
// Deal with contents and data later
2017-08-25 10:37:46 +00:00
};
2017-08-17 23:10:35 +00:00
const syncDataByItemId = store.getters['data/syncDataByItemId'];
2017-12-23 18:25:14 +00:00
let promise;
2017-12-10 23:49:20 +00:00
Object.entries(storeItemMap).some(([id, item]) => {
2017-08-17 23:10:35 +00:00
const existingSyncData = syncDataByItemId[id];
2017-12-23 18:25:14 +00:00
if ((!existingSyncData || existingSyncData.hash !== item.hash)
// Add file/folder if parent has been added
&& (!storeItemMap[item.parentId] || syncDataByItemId[item.parentId])
// Add file if content has been added
&& (item.type !== 'file' || syncDataByItemId[`${id}/content`])
2017-09-27 20:27:12 +00:00
) {
2018-04-27 14:37:05 +00:00
promise = workspaceProvider
.saveSimpleItem(
// Use deepCopy to freeze objects
utils.deepCopy(item),
utils.deepCopy(existingSyncData),
ifNotTooLate,
)
2017-08-17 23:10:35 +00:00
.then(resultSyncData => store.dispatch('data/patchSyncData', {
[resultSyncData.id]: resultSyncData,
}))
.then(() => saveNextItem());
}
2017-12-23 18:25:14 +00:00
return promise;
2017-08-15 10:43:26 +00:00
});
2017-12-23 18:25:14 +00:00
return promise;
2017-08-15 10:43:26 +00:00
});
2017-08-17 23:10:35 +00:00
// Called until no item to remove
const removeNextItem = ifNotTooLate(() => {
2017-08-25 10:37:46 +00:00
const storeItemMap = {
...store.state.file.itemMap,
...store.state.folder.itemMap,
...store.state.syncLocation.itemMap,
2017-09-23 19:01:50 +00:00
...store.state.publishLocation.itemMap,
...store.state.content.itemMap,
2017-08-25 10:37:46 +00:00
};
2017-08-17 23:10:35 +00:00
const syncData = store.getters['data/syncData'];
2017-12-23 18:25:14 +00:00
let promise;
2017-12-10 23:49:20 +00:00
Object.entries(syncData).some(([, existingSyncData]) => {
2017-08-25 10:37:46 +00:00
if (!storeItemMap[existingSyncData.itemId] &&
2017-12-17 15:08:52 +00:00
// We don't want to delete data items, especially on first sync
existingSyncData.type !== 'data' &&
2017-08-25 10:37:46 +00:00
// Remove content only if file has been removed
(existingSyncData.type !== 'content' || !storeItemMap[existingSyncData.itemId.split('/')[0]])
) {
2017-08-17 23:10:35 +00:00
// Use deepCopy to freeze objects
const syncDataToRemove = utils.deepCopy(existingSyncData);
2018-01-04 20:19:10 +00:00
promise = workspaceProvider
2017-12-23 18:25:14 +00:00
.removeItem(syncDataToRemove, ifNotTooLate)
2017-08-17 23:10:35 +00:00
.then(() => {
const syncDataCopy = { ...store.getters['data/syncData'] };
delete syncDataCopy[syncDataToRemove.id];
store.dispatch('data/setSyncData', syncDataCopy);
})
.then(() => removeNextItem());
}
2017-12-23 18:25:14 +00:00
return promise;
2017-08-17 23:10:35 +00:00
});
2017-12-23 18:25:14 +00:00
return promise;
2017-08-17 23:10:35 +00:00
});
2017-08-25 10:37:46 +00:00
const getOneFileIdToSync = () => {
2017-09-27 20:27:12 +00:00
const contentIds = [...new Set([
...Object.keys(localDbSvc.hashMap.content),
...store.getters['file/items'].map(file => `${file.id}/content`),
])];
2017-08-25 10:37:46 +00:00
let fileId;
2017-09-27 20:27:12 +00:00
contentIds.some((contentId) => {
2017-08-25 10:37:46 +00:00
// Get content hash from itemMap or from localDbSvc if not loaded
const loadedContent = store.state.content.itemMap[contentId];
const hash = loadedContent ? loadedContent.hash : localDbSvc.hashMap.content[contentId];
const syncData = store.getters['data/syncDataByItemId'][contentId];
if (
2018-01-30 07:36:33 +00:00
// Sync if content syncing was not attempted yet
2018-03-12 00:45:54 +00:00
!syncContext.attempted[contentId] &&
2018-01-30 07:36:33 +00:00
// And if syncData does not exist or if content hash and syncData hash are inconsistent
(!syncData || syncData.hash !== hash)
) {
2017-08-25 10:37:46 +00:00
[fileId] = contentId.split('/');
2017-08-17 23:10:35 +00:00
}
2017-08-25 10:37:46 +00:00
return fileId;
});
return fileId;
};
const syncNextFile = () => {
2017-08-25 10:37:46 +00:00
const fileId = getOneFileIdToSync();
2017-09-27 20:27:12 +00:00
if (!fileId) {
return null;
2017-09-27 20:27:12 +00:00
}
return syncFile(fileId, syncContext)
.then(() => syncNextFile());
2017-08-25 10:37:46 +00:00
};
2017-08-17 23:10:35 +00:00
2018-05-06 00:46:33 +00:00
const onSyncEnd = () => Promise.resolve(workspaceProvider.onSyncEnd
&& workspaceProvider.onSyncEnd());
2018-04-27 14:37:05 +00:00
2017-08-17 23:10:35 +00:00
return Promise.resolve()
.then(() => saveNextItem())
.then(() => removeNextItem())
2017-12-17 15:08:52 +00:00
// Sync settings only in the main workspace
.then(() => workspace.id === 'main' && syncDataItem('settings'))
// Sync workspaces only in the main workspace
.then(() => workspace.id === 'main' && syncDataItem('workspaces'))
2017-09-26 22:54:26 +00:00
.then(() => syncDataItem('templates'))
2017-08-25 10:37:46 +00:00
.then(() => {
2017-09-27 20:27:12 +00:00
const currentFileId = store.getters['file/current'].id;
2017-09-26 22:54:26 +00:00
if (currentFileId) {
2017-08-25 10:37:46 +00:00
// Sync current file first
return syncFile(currentFileId, syncContext)
.then(() => syncNextFile());
2017-08-25 10:37:46 +00:00
}
return syncNextFile();
})
2018-04-27 14:37:05 +00:00
.then(
() => onSyncEnd(),
err => onSyncEnd().then(() => {
throw err;
}, () => {
throw err;
2018-05-06 00:46:33 +00:00
}),
)
2017-09-27 20:27:12 +00:00
.then(
() => {
if (syncContext.restart) {
2017-09-27 20:27:12 +00:00
// Restart sync
2017-12-23 18:25:14 +00:00
return syncWorkspace();
2017-09-27 20:27:12 +00:00
}
return null;
},
(err) => {
if (err && err.message === 'TOO_LATE') {
// Restart sync
2017-12-23 18:25:14 +00:00
return syncWorkspace();
2017-09-27 20:27:12 +00:00
}
throw err;
2018-05-06 00:46:33 +00:00
},
);
2017-08-15 10:43:26 +00:00
});
}
2017-12-10 23:49:20 +00:00
/**
* Enqueue a sync task, if possible.
*/
2017-08-15 10:43:26 +00:00
function requestSync() {
2018-03-12 00:45:54 +00:00
// No sync in light mode
if (store.state.light) {
return;
}
2017-08-15 10:43:26 +00:00
store.dispatch('queue/enqueueSyncRequest', () => new Promise((resolve, reject) => {
let intervalId;
const attempt = () => {
// Only start syncing when these conditions are met
2017-12-11 00:53:46 +00:00
if (networkSvc.isUserActive() && isSyncWindow()) {
2017-08-15 10:43:26 +00:00
clearInterval(intervalId);
2017-08-25 10:37:46 +00:00
if (!isSyncPossible()) {
2017-08-15 10:43:26 +00:00
// Cancel sync
2018-05-06 00:46:33 +00:00
reject(new Error('Sync not possible.'));
2017-08-25 10:37:46 +00:00
return;
2017-08-15 10:43:26 +00:00
}
2017-08-25 10:37:46 +00:00
2017-10-07 11:22:24 +00:00
// Determine if we have to clean files
const fileHashesToClean = {};
if (getLastStoredSyncActivity() + utils.cleanTrashAfter < Date.now()) {
// Last synchronization happened 7 days ago
const syncDataByItemId = store.getters['data/syncDataByItemId'];
store.getters['file/items'].forEach((file) => {
// If file is in the trash and has not been modified since it was last synced
const syncData = syncDataByItemId[file.id];
if (syncData && file.parentId === 'trash' && file.hash === syncData.hash) {
fileHashesToClean[file.id] = file.hash;
}
});
}
2017-08-25 10:37:46 +00:00
// Call setLastSyncActivity periodically
intervalId = utils.setInterval(() => setLastSyncActivity(), 1000);
setLastSyncActivity();
const cleaner = cb => (res) => {
clearInterval(intervalId);
cb(res);
};
2017-10-07 11:22:24 +00:00
2017-08-25 10:37:46 +00:00
Promise.resolve()
.then(() => {
2017-12-10 23:49:20 +00:00
if (isWorkspaceSyncPossible()) {
2017-12-23 18:25:14 +00:00
return syncWorkspace();
2017-08-25 10:37:46 +00:00
}
if (hasCurrentFileSyncLocations()) {
2017-12-17 15:08:52 +00:00
// Only sync current file if workspace sync is unavailable.
// We could also sync files that are out-of-sync but it would
// require to load all the syncedContent objects from the DB.
2017-08-25 10:37:46 +00:00
return syncFile(store.getters['file/current'].id);
}
return null;
})
2017-10-07 11:22:24 +00:00
.then(() => {
// Clean files
2017-12-10 23:49:20 +00:00
Object.entries(fileHashesToClean).forEach(([fileId, fileHash]) => {
2017-10-07 11:22:24 +00:00
const file = store.state.file.itemMap[fileId];
2017-12-10 23:49:20 +00:00
if (file && file.hash === fileHash) {
2018-05-04 18:07:28 +00:00
fileSvc.deleteFile(fileId);
2017-10-07 11:22:24 +00:00
}
});
})
2017-08-25 10:37:46 +00:00
.then(cleaner(resolve), cleaner(reject));
2017-08-15 10:43:26 +00:00
}
};
intervalId = utils.setInterval(() => attempt(), 1000);
attempt();
}));
}
2017-12-10 23:49:20 +00:00
export default {
init() {
2018-01-04 20:19:10 +00:00
return Promise.resolve()
.then(() => {
// Load workspaces and tokens from localStorage
localDbSvc.syncLocalStorage();
2017-07-28 07:40:24 +00:00
2018-01-04 20:19:10 +00:00
// Try to find a suitable action provider
actionProvider = providerRegistry.providers[utils.queryParams.providerId];
return actionProvider && actionProvider.initAction && actionProvider.initAction();
})
.then(() => {
// Try to find a suitable workspace sync provider
workspaceProvider = providerRegistry.providers[utils.queryParams.providerId];
if (!workspaceProvider || !workspaceProvider.initWorkspace) {
workspaceProvider = googleDriveAppDataProvider;
}
return workspaceProvider.initWorkspace();
})
2017-12-17 15:08:52 +00:00
.then(workspace => store.dispatch('workspace/setCurrentWorkspaceId', workspace.id))
2017-12-10 23:49:20 +00:00
.then(() => localDbSvc.init())
2018-01-04 20:19:10 +00:00
.then(() => {
// Try to find a suitable action provider
actionProvider = providerRegistry.providers[utils.queryParams.providerId] || actionProvider;
return actionProvider && actionProvider.performAction && actionProvider.performAction()
.then(newSyncLocation => newSyncLocation && this.createSyncLocation(newSyncLocation));
})
2018-03-12 00:45:54 +00:00
.then(() => tempFileSvc.init())
2017-12-10 23:49:20 +00:00
.then(() => {
2018-03-12 00:45:54 +00:00
if (!store.state.light) {
// Sync periodically
utils.setInterval(() => {
if (isSyncPossible()
&& networkSvc.isUserActive()
&& isSyncWindow()
&& isAutoSyncReady()
) {
requestSync();
}
}, 1000);
2017-08-19 10:24:08 +00:00
2018-03-12 00:45:54 +00:00
// Unload contents from memory periodically
utils.setInterval(() => {
// Wait for sync and publish to finish
if (store.state.queue.isEmpty) {
localDbSvc.unloadContents();
}
}, 5000);
}
2017-12-10 23:49:20 +00:00
});
},
2017-08-25 10:37:46 +00:00
isSyncPossible,
2017-08-15 10:43:26 +00:00
requestSync,
2017-09-23 19:01:50 +00:00
createSyncLocation,
2017-08-15 10:43:26 +00:00
};