mirror of
https://github.com/gristlabs/grist-core.git
synced 2024-10-27 20:44:07 +00:00
This check should be unnecessary for stores with strong consistency guarantees (virtually everywhere now). --------- Co-authored-by: Florent FAYOLLE <florent.fayolle@beta.gouv.fr>
This commit is contained in:
parent
66f025c7df
commit
1e3e076820
@ -291,6 +291,7 @@ COOKIE_MAX_AGE | session cookie max age, defaults to 90 days; can be set to
|
|||||||
HOME_PORT | port number to listen on for REST API server; if set to "share", add API endpoints to regular grist port.
|
HOME_PORT | port number to listen on for REST API server; if set to "share", add API endpoints to regular grist port.
|
||||||
PORT | port number to listen on for Grist server
|
PORT | port number to listen on for Grist server
|
||||||
REDIS_URL | optional redis server for browser sessions and db query caching
|
REDIS_URL | optional redis server for browser sessions and db query caching
|
||||||
|
GRIST_SKIP_REDIS_CHECKSUM_MISMATCH | Experimental. If set, only warn if the checksum in Redis differs with the one in your S3 backend storage. You may turn it on if your backend storage implements the [read-after-write consistency](https://aws.amazon.com/fr/blogs/aws/amazon-s3-update-strong-read-after-write-consistency/). Defaults to false.
|
||||||
GRIST_SNAPSHOT_TIME_CAP | optional. Define the caps for tracking buckets. Usage: {"hour": 25, "day": 32, "isoWeek": 12, "month": 96, "year": 1000}
|
GRIST_SNAPSHOT_TIME_CAP | optional. Define the caps for tracking buckets. Usage: {"hour": 25, "day": 32, "isoWeek": 12, "month": 96, "year": 1000}
|
||||||
GRIST_SNAPSHOT_KEEP | optional. Number of recent snapshots to retain unconditionally for a document, regardless of when they were made
|
GRIST_SNAPSHOT_KEEP | optional. Number of recent snapshots to retain unconditionally for a document, regardless of when they were made
|
||||||
GRIST_PROMCLIENT_PORT | optional. If set, serve the Prometheus metrics on the specified port number. ⚠️ Be sure to use a port which is not publicly exposed ⚠️.
|
GRIST_PROMCLIENT_PORT | optional. If set, serve the Prometheus metrics on the specified port number. ⚠️ Be sure to use a port which is not publicly exposed ⚠️.
|
||||||
|
@ -1,6 +1,8 @@
|
|||||||
import {ObjMetadata, ObjSnapshot, ObjSnapshotWithMetadata} from 'app/common/DocSnapshot';
|
import {ObjMetadata, ObjSnapshot, ObjSnapshotWithMetadata} from 'app/common/DocSnapshot';
|
||||||
|
import {isAffirmative} from 'app/common/gutil';
|
||||||
import log from 'app/server/lib/log';
|
import log from 'app/server/lib/log';
|
||||||
import {createTmpDir} from 'app/server/lib/uploads';
|
import {createTmpDir} from 'app/server/lib/uploads';
|
||||||
|
|
||||||
import {delay} from 'bluebird';
|
import {delay} from 'bluebird';
|
||||||
import * as fse from 'fs-extra';
|
import * as fse from 'fs-extra';
|
||||||
import * as path from 'path';
|
import * as path from 'path';
|
||||||
@ -226,15 +228,29 @@ export class ChecksummedExternalStorage implements ExternalStorage {
|
|||||||
const expectedChecksum = await this._options.sharedHash.load(fromKey);
|
const expectedChecksum = await this._options.sharedHash.load(fromKey);
|
||||||
// Let null docMD5s pass. Otherwise we get stuck if redis is cleared.
|
// Let null docMD5s pass. Otherwise we get stuck if redis is cleared.
|
||||||
// Otherwise, make sure what we've got matches what we expect to get.
|
// Otherwise, make sure what we've got matches what we expect to get.
|
||||||
// S3 is eventually consistent - if you overwrite an object in it, and then read from it,
|
// AWS S3 was eventually consistent, but now has stronger guarantees:
|
||||||
// you may get an old version for some time.
|
// https://aws.amazon.com/blogs/aws/amazon-s3-update-strong-read-after-write-consistency/
|
||||||
|
//
|
||||||
|
// Previous to this change, if you overwrote an object in it,
|
||||||
|
// and then read from it, you may have got an old version for some time.
|
||||||
|
// We are confident this should not be the case anymore, though this has to be studied carefully.
|
||||||
// If a snapshotId was specified, we can skip this check.
|
// If a snapshotId was specified, we can skip this check.
|
||||||
if (expectedChecksum && expectedChecksum !== checksum) {
|
if (expectedChecksum && expectedChecksum !== checksum) {
|
||||||
log.error("ext %s download: data for %s has wrong checksum: %s (expected %s)",
|
const message = `ext ${this.label} download: data for ${fromKey} has wrong checksum:` +
|
||||||
this.label, fromKey, checksum, expectedChecksum);
|
` ${checksum} (expected ${expectedChecksum})`;
|
||||||
|
|
||||||
|
// If GRIST_SKIP_REDIS_CHECKSUM_MISMATCH is set, issue a warning only and continue,
|
||||||
|
// rather than issuing an error and failing.
|
||||||
|
// This flag is experimental and should be removed once we are
|
||||||
|
// confident that the checksums verification is useless.
|
||||||
|
if (isAffirmative(process.env.GRIST_SKIP_REDIS_CHECKSUM_MISMATCH)) {
|
||||||
|
log.warn(message);
|
||||||
|
} else {
|
||||||
|
log.error(message);
|
||||||
return undefined;
|
return undefined;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// If successful, rename the temporary file to its proper name. The destination should NOT
|
// If successful, rename the temporary file to its proper name. The destination should NOT
|
||||||
// exist in this case, and this should fail if it does.
|
// exist in this case, and this should fail if it does.
|
||||||
|
@ -25,7 +25,7 @@ import {createClient, RedisClient} from 'redis';
|
|||||||
import * as sinon from 'sinon';
|
import * as sinon from 'sinon';
|
||||||
import {createInitialDb, removeConnection, setUpDB} from 'test/gen-server/seed';
|
import {createInitialDb, removeConnection, setUpDB} from 'test/gen-server/seed';
|
||||||
import {createTmpDir, getGlobalPluginManager} from 'test/server/docTools';
|
import {createTmpDir, getGlobalPluginManager} from 'test/server/docTools';
|
||||||
import {setTmpLogLevel, useFixtureDoc} from 'test/server/testUtils';
|
import {EnvironmentSnapshot, setTmpLogLevel, useFixtureDoc} from 'test/server/testUtils';
|
||||||
import {waitForIt} from 'test/server/wait';
|
import {waitForIt} from 'test/server/wait';
|
||||||
import uuidv4 from "uuid/v4";
|
import uuidv4 from "uuid/v4";
|
||||||
|
|
||||||
@ -273,6 +273,17 @@ class TestStore {
|
|||||||
private _externalStorageCreate: (purpose: 'doc'|'meta', extraPrefix: string) => ExternalStorage|undefined) {
|
private _externalStorageCreate: (purpose: 'doc'|'meta', extraPrefix: string) => ExternalStorage|undefined) {
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public async run<T>(fn: () => Promise<T>): Promise<T> {
|
||||||
|
await this.begin();
|
||||||
|
let result;
|
||||||
|
try {
|
||||||
|
result = await fn();
|
||||||
|
} finally {
|
||||||
|
await this.end();
|
||||||
|
}
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
// Simulates doc worker startup.
|
// Simulates doc worker startup.
|
||||||
public async begin() {
|
public async begin() {
|
||||||
await this.end();
|
await this.end();
|
||||||
@ -366,6 +377,7 @@ describe('HostedStorageManager', function() {
|
|||||||
describe(storage, function() {
|
describe(storage, function() {
|
||||||
|
|
||||||
const sandbox = sinon.createSandbox();
|
const sandbox = sinon.createSandbox();
|
||||||
|
let oldEnv: EnvironmentSnapshot;
|
||||||
|
|
||||||
const workerId = 'dw17';
|
const workerId = 'dw17';
|
||||||
let cli: RedisClient;
|
let cli: RedisClient;
|
||||||
@ -376,6 +388,7 @@ describe('HostedStorageManager', function() {
|
|||||||
before(async function() {
|
before(async function() {
|
||||||
if (!process.env.TEST_REDIS_URL) { this.skip(); return; }
|
if (!process.env.TEST_REDIS_URL) { this.skip(); return; }
|
||||||
cli = createClient(process.env.TEST_REDIS_URL);
|
cli = createClient(process.env.TEST_REDIS_URL);
|
||||||
|
oldEnv = new EnvironmentSnapshot();
|
||||||
await cli.flushdbAsync();
|
await cli.flushdbAsync();
|
||||||
workers = new DocWorkerMap([cli]);
|
workers = new DocWorkerMap([cli]);
|
||||||
await workers.addWorker({
|
await workers.addWorker({
|
||||||
@ -439,6 +452,7 @@ describe('HostedStorageManager', function() {
|
|||||||
});
|
});
|
||||||
|
|
||||||
afterEach(async function() {
|
afterEach(async function() {
|
||||||
|
oldEnv.restore();
|
||||||
sandbox.restore();
|
sandbox.restore();
|
||||||
if (store) {
|
if (store) {
|
||||||
await store.end();
|
await store.end();
|
||||||
@ -468,57 +482,67 @@ describe('HostedStorageManager', function() {
|
|||||||
assert.equal(await getRedisChecksum(docId), 'null');
|
assert.equal(await getRedisChecksum(docId), 'null');
|
||||||
|
|
||||||
// Create an empty document when checksum in redis is 'null'.
|
// Create an empty document when checksum in redis is 'null'.
|
||||||
await store.begin();
|
const checksum = await store.run(async () => {
|
||||||
await store.docManager.fetchDoc(docSession, docId);
|
await store.docManager.fetchDoc(docSession, docId);
|
||||||
assert(await store.waitForUpdates());
|
assert(await store.waitForUpdates());
|
||||||
const checksum = await getRedisChecksum(docId);
|
const checksum = await getRedisChecksum(docId);
|
||||||
assert.notEqual(checksum, 'null');
|
assert.notEqual(checksum, 'null');
|
||||||
await store.end();
|
return checksum;
|
||||||
|
});
|
||||||
|
|
||||||
// Check if we nobble the expected checksum then fetch eventually errors.
|
// Check what happens when we nobble the expected checksum.
|
||||||
await setRedisChecksum(docId, 'nobble');
|
await setRedisChecksum(docId, 'nobble');
|
||||||
await store.removeAll();
|
await store.removeAll();
|
||||||
await store.begin();
|
|
||||||
|
// With GRIST_SKIP_REDIS_CHECKSUM_MISMATCH set, the fetch should work
|
||||||
|
process.env.GRIST_SKIP_REDIS_CHECKSUM_MISMATCH = 'true';
|
||||||
|
await store.run(async () => {
|
||||||
|
await assert.isFulfilled(store.docManager.fetchDoc(docSession, docId));
|
||||||
|
});
|
||||||
|
|
||||||
|
// By default, the fetch should eventually errors.
|
||||||
|
delete process.env.GRIST_SKIP_REDIS_CHECKSUM_MISMATCH;
|
||||||
|
await store.run(async () => {
|
||||||
await assert.isRejected(store.docManager.fetchDoc(docSession, docId),
|
await assert.isRejected(store.docManager.fetchDoc(docSession, docId),
|
||||||
/operation failed to become consistent/);
|
/operation failed to become consistent/);
|
||||||
await store.end();
|
});
|
||||||
|
|
||||||
// Check we get the document back on fresh start if checksum is correct.
|
// Check we get the document back on fresh start if checksum is correct.
|
||||||
await setRedisChecksum(docId, checksum);
|
await setRedisChecksum(docId, checksum);
|
||||||
await store.removeAll();
|
await store.removeAll();
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
await store.docManager.fetchDoc(docSession, docId);
|
await store.docManager.fetchDoc(docSession, docId);
|
||||||
await store.end();
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
it('can save modifications', async function() {
|
it('can save modifications', async function() {
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
|
|
||||||
await workers.assignDocWorker('Hello');
|
await workers.assignDocWorker('Hello');
|
||||||
await useFixtureDoc('Hello.grist', store.storageManager);
|
await useFixtureDoc('Hello.grist', store.storageManager);
|
||||||
|
|
||||||
await workers.assignDocWorker('Hello2');
|
await workers.assignDocWorker('Hello2');
|
||||||
|
|
||||||
let doc = await store.docManager.fetchDoc(docSession, 'Hello');
|
const doc = await store.docManager.fetchDoc(docSession, 'Hello');
|
||||||
let doc2 = await store.docManager.fetchDoc(docSession, 'Hello2');
|
const doc2 = await store.docManager.fetchDoc(docSession, 'Hello2');
|
||||||
await doc.docStorage.exec("update Table1 set A = 'magic_word' where id = 1");
|
await doc.docStorage.exec("update Table1 set A = 'magic_word' where id = 1");
|
||||||
await doc2.docStorage.exec("insert into Table1(id) values(42)");
|
await doc2.docStorage.exec("insert into Table1(id) values(42)");
|
||||||
await store.end();
|
return { doc, doc2 };
|
||||||
|
});
|
||||||
|
|
||||||
await store.removeAll();
|
await store.removeAll();
|
||||||
await store.begin();
|
|
||||||
doc = await store.docManager.fetchDoc(docSession, 'Hello');
|
await store.run(async () => {
|
||||||
|
const doc = await store.docManager.fetchDoc(docSession, 'Hello');
|
||||||
let result = await doc.docStorage.get("select A from Table1 where id = 1");
|
let result = await doc.docStorage.get("select A from Table1 where id = 1");
|
||||||
assert.equal(result!.A, 'magic_word');
|
assert.equal(result!.A, 'magic_word');
|
||||||
doc2 = await store.docManager.fetchDoc(docSession, 'Hello2');
|
const doc2 = await store.docManager.fetchDoc(docSession, 'Hello2');
|
||||||
result = await doc2.docStorage.get("select id from Table1");
|
result = await doc2.docStorage.get("select id from Table1");
|
||||||
assert.equal(result!.id, 42);
|
assert.equal(result!.id, 42);
|
||||||
await store.end();
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
it('can save modifications with interfering backup file', async function() {
|
it('can save modifications with interfering backup file', async function() {
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
|
|
||||||
// There was a bug where if a corrupt/truncated backup file was created, all future
|
// There was a bug where if a corrupt/truncated backup file was created, all future
|
||||||
// backups would fail. This tickles the condition and makes sure backups now succeed.
|
// backups would fail. This tickles the condition and makes sure backups now succeed.
|
||||||
await fse.writeFile(path.join(tmpDir, 'Hello.grist-backup'), 'not a sqlite file');
|
await fse.writeFile(path.join(tmpDir, 'Hello.grist-backup'), 'not a sqlite file');
|
||||||
@ -526,36 +550,37 @@ describe('HostedStorageManager', function() {
|
|||||||
await workers.assignDocWorker('Hello');
|
await workers.assignDocWorker('Hello');
|
||||||
await useFixtureDoc('Hello.grist', store.storageManager);
|
await useFixtureDoc('Hello.grist', store.storageManager);
|
||||||
|
|
||||||
let doc = await store.docManager.fetchDoc(docSession, 'Hello');
|
const doc = await store.docManager.fetchDoc(docSession, 'Hello');
|
||||||
await doc.docStorage.exec("update Table1 set A = 'magic_word2' where id = 1");
|
await doc.docStorage.exec("update Table1 set A = 'magic_word2' where id = 1");
|
||||||
await store.end(); // S3 push will happen prior to this returning.
|
});
|
||||||
|
|
||||||
|
// S3 should have happened after store.run()
|
||||||
|
|
||||||
await store.removeAll();
|
await store.removeAll();
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
doc = await store.docManager.fetchDoc(docSession, 'Hello');
|
const doc = await store.docManager.fetchDoc(docSession, 'Hello');
|
||||||
const result = await doc.docStorage.get("select A from Table1 where id = 1");
|
const result = await doc.docStorage.get("select A from Table1 where id = 1");
|
||||||
assert.equal(result!.A, 'magic_word2');
|
assert.equal(result!.A, 'magic_word2');
|
||||||
await store.end();
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
it('survives if there is a doc marked dirty that turns out to be clean', async function() {
|
it('survives if there is a doc marked dirty that turns out to be clean', async function() {
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
|
|
||||||
await workers.assignDocWorker('Hello');
|
await workers.assignDocWorker('Hello');
|
||||||
await useFixtureDoc('Hello.grist', store.storageManager);
|
await useFixtureDoc('Hello.grist', store.storageManager);
|
||||||
|
|
||||||
let doc = await store.docManager.fetchDoc(docSession, 'Hello');
|
const doc = await store.docManager.fetchDoc(docSession, 'Hello');
|
||||||
await doc.docStorage.exec("update Table1 set A = 'magic_word' where id = 1");
|
await doc.docStorage.exec("update Table1 set A = 'magic_word' where id = 1");
|
||||||
await store.end();
|
});
|
||||||
|
|
||||||
await store.removeAll();
|
await store.removeAll();
|
||||||
|
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
doc = await store.docManager.fetchDoc(docSession, 'Hello');
|
const doc = await store.docManager.fetchDoc(docSession, 'Hello');
|
||||||
const result = await doc.docStorage.get("select A from Table1 where id = 1");
|
const result = await doc.docStorage.get("select A from Table1 where id = 1");
|
||||||
assert.equal(result!.A, 'magic_word');
|
assert.equal(result!.A, 'magic_word');
|
||||||
store.docManager.markAsChanged(doc);
|
store.docManager.markAsChanged(doc);
|
||||||
await store.end();
|
});
|
||||||
|
|
||||||
// The real test is whether this test manages to complete.
|
// The real test is whether this test manages to complete.
|
||||||
});
|
});
|
||||||
@ -564,15 +589,15 @@ describe('HostedStorageManager', function() {
|
|||||||
await workers.assignDocWorker('Hello');
|
await workers.assignDocWorker('Hello');
|
||||||
|
|
||||||
// put a doc in s3
|
// put a doc in s3
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
await useFixtureDoc('Hello.grist', store.storageManager);
|
await useFixtureDoc('Hello.grist', store.storageManager);
|
||||||
let doc = await store.docManager.fetchDoc(docSession, 'Hello');
|
const doc = await store.docManager.fetchDoc(docSession, 'Hello');
|
||||||
await doc.docStorage.exec("update Table1 set A = 'parallel' where id = 1");
|
await doc.docStorage.exec("update Table1 set A = 'parallel' where id = 1");
|
||||||
await store.end();
|
});
|
||||||
|
|
||||||
// now open it many times in parallel
|
// now open it many times in parallel
|
||||||
await store.removeAll();
|
await store.removeAll();
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
const docs = Promise.all([
|
const docs = Promise.all([
|
||||||
store.docManager.fetchDoc(docSession, 'Hello'),
|
store.docManager.fetchDoc(docSession, 'Hello'),
|
||||||
store.docManager.fetchDoc(docSession, 'Hello'),
|
store.docManager.fetchDoc(docSession, 'Hello'),
|
||||||
@ -580,15 +605,15 @@ describe('HostedStorageManager', function() {
|
|||||||
store.docManager.fetchDoc(docSession, 'Hello'),
|
store.docManager.fetchDoc(docSession, 'Hello'),
|
||||||
]);
|
]);
|
||||||
await assert.isFulfilled(docs);
|
await assert.isFulfilled(docs);
|
||||||
doc = (await docs)[0];
|
const doc = (await docs)[0];
|
||||||
const result = await doc.docStorage.get("select A from Table1 where id = 1");
|
const result = await doc.docStorage.get("select A from Table1 where id = 1");
|
||||||
assert.equal(result!.A, 'parallel');
|
assert.equal(result!.A, 'parallel');
|
||||||
await store.end();
|
});
|
||||||
|
|
||||||
// To be sure we are checking something, let's call prepareLocalDoc directly
|
// To be sure we are checking something, let's call prepareLocalDoc directly
|
||||||
// on storage manager and make sure it fails.
|
// on storage manager and make sure it fails.
|
||||||
await store.removeAll();
|
await store.removeAll();
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
const preps = Promise.all([
|
const preps = Promise.all([
|
||||||
store.storageManager.prepareLocalDoc('Hello'),
|
store.storageManager.prepareLocalDoc('Hello'),
|
||||||
store.storageManager.prepareLocalDoc('Hello'),
|
store.storageManager.prepareLocalDoc('Hello'),
|
||||||
@ -596,7 +621,7 @@ describe('HostedStorageManager', function() {
|
|||||||
store.storageManager.prepareLocalDoc('Hello')
|
store.storageManager.prepareLocalDoc('Hello')
|
||||||
]);
|
]);
|
||||||
await assert.isRejected(preps, /in parallel/);
|
await assert.isRejected(preps, /in parallel/);
|
||||||
await store.end();
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
it ('can delete a document', async function() {
|
it ('can delete a document', async function() {
|
||||||
@ -604,17 +629,17 @@ describe('HostedStorageManager', function() {
|
|||||||
await workers.assignDocWorker(docId);
|
await workers.assignDocWorker(docId);
|
||||||
|
|
||||||
// Create a document
|
// Create a document
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
let doc = await store.docManager.fetchDoc(docSession, docId);
|
const doc = await store.docManager.fetchDoc(docSession, docId);
|
||||||
await doc.docStorage.exec("insert into Table1(id) values(42)");
|
await doc.docStorage.exec("insert into Table1(id) values(42)");
|
||||||
await store.end();
|
});
|
||||||
|
|
||||||
const docPath = store.getDocPath(docId);
|
const docPath = store.getDocPath(docId);
|
||||||
const ext = store.storageManager.testGetExternalStorage();
|
const ext = store.storageManager.testGetExternalStorage();
|
||||||
|
|
||||||
// Check that the document exists on filesystem and in external store.
|
// Check that the document exists on filesystem and in external store.
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
doc = await store.docManager.fetchDoc(docSession, docId);
|
const doc = await store.docManager.fetchDoc(docSession, docId);
|
||||||
assert.equal(await fse.pathExists(docPath), true);
|
assert.equal(await fse.pathExists(docPath), true);
|
||||||
assert.equal(await fse.pathExists(docPath + '-hash-doc'), true);
|
assert.equal(await fse.pathExists(docPath + '-hash-doc'), true);
|
||||||
await waitForIt(async () => assert.equal(await ext.exists(docId), true), 20000);
|
await waitForIt(async () => assert.equal(await ext.exists(docId), true), 20000);
|
||||||
@ -626,7 +651,7 @@ describe('HostedStorageManager', function() {
|
|||||||
assert.equal(await fse.pathExists(docPath + '-hash-doc'), false);
|
assert.equal(await fse.pathExists(docPath + '-hash-doc'), false);
|
||||||
assert.equal(await getRedisChecksum(docId), DELETED_TOKEN);
|
assert.equal(await getRedisChecksum(docId), DELETED_TOKEN);
|
||||||
await waitForIt(async () => assert.equal(await ext.exists(docId), false), 20000);
|
await waitForIt(async () => assert.equal(await ext.exists(docId), false), 20000);
|
||||||
await store.end();
|
});
|
||||||
|
|
||||||
// As far as the underlying storage is concerned it should be
|
// As far as the underlying storage is concerned it should be
|
||||||
// possible to recreate a doc with the same id after deletion.
|
// possible to recreate a doc with the same id after deletion.
|
||||||
@ -634,24 +659,23 @@ describe('HostedStorageManager', function() {
|
|||||||
// document it must exist in the db - however we'll need to watch
|
// document it must exist in the db - however we'll need to watch
|
||||||
// out for caching.
|
// out for caching.
|
||||||
// TODO: it could be worth tweaking fetchDoc so creation is explicit.
|
// TODO: it could be worth tweaking fetchDoc so creation is explicit.
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
doc = await store.docManager.fetchDoc(docSession, docId);
|
const doc = await store.docManager.fetchDoc(docSession, docId);
|
||||||
await doc.docStorage.exec("insert into Table1(id) values(42)");
|
await doc.docStorage.exec("insert into Table1(id) values(42)");
|
||||||
await store.end();
|
});
|
||||||
|
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
doc = await store.docManager.fetchDoc(docSession, docId);
|
await store.docManager.fetchDoc(docSession, docId);
|
||||||
assert.equal(await fse.pathExists(docPath), true);
|
assert.equal(await fse.pathExists(docPath), true);
|
||||||
assert.equal(await fse.pathExists(docPath + '-hash-doc'), true);
|
assert.equal(await fse.pathExists(docPath + '-hash-doc'), true);
|
||||||
await store.end();
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
it('individual document close is orderly', async function() {
|
it('individual document close is orderly', async function() {
|
||||||
const docId = `create-${uuidv4()}`;
|
const docId = `create-${uuidv4()}`;
|
||||||
await workers.assignDocWorker(docId);
|
await workers.assignDocWorker(docId);
|
||||||
|
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
|
|
||||||
let doc = await store.docManager.fetchDoc(docSession, docId);
|
let doc = await store.docManager.fetchDoc(docSession, docId);
|
||||||
await store.closeDoc(doc);
|
await store.closeDoc(doc);
|
||||||
const checksum1 = await getRedisChecksum(docId);
|
const checksum1 = await getRedisChecksum(docId);
|
||||||
@ -681,8 +705,7 @@ describe('HostedStorageManager', function() {
|
|||||||
const checksum3 = await getRedisChecksum(docId);
|
const checksum3 = await getRedisChecksum(docId);
|
||||||
assert.notEqual(checksum2, checksum3);
|
assert.notEqual(checksum2, checksum3);
|
||||||
await asyncClose;
|
await asyncClose;
|
||||||
|
});
|
||||||
await store.end();
|
|
||||||
});
|
});
|
||||||
|
|
||||||
// Viewing a document should not mark it as changed (unless a document-level migration
|
// Viewing a document should not mark it as changed (unless a document-level migration
|
||||||
@ -691,8 +714,7 @@ describe('HostedStorageManager', function() {
|
|||||||
const docId = `create-${uuidv4()}`;
|
const docId = `create-${uuidv4()}`;
|
||||||
await workers.assignDocWorker(docId);
|
await workers.assignDocWorker(docId);
|
||||||
|
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
|
|
||||||
const markAsChanged: {callCount: number} = store.storageManager.markAsChanged as any;
|
const markAsChanged: {callCount: number} = store.storageManager.markAsChanged as any;
|
||||||
|
|
||||||
const changesInitial = markAsChanged.callCount;
|
const changesInitial = markAsChanged.callCount;
|
||||||
@ -707,8 +729,7 @@ describe('HostedStorageManager', function() {
|
|||||||
await store.closeDoc(doc);
|
await store.closeDoc(doc);
|
||||||
const changesAfterViewing = markAsChanged.callCount;
|
const changesAfterViewing = markAsChanged.callCount;
|
||||||
assert.equal(changesAfterViewing, changesAfterCreation);
|
assert.equal(changesAfterViewing, changesAfterCreation);
|
||||||
|
});
|
||||||
await store.end();
|
|
||||||
});
|
});
|
||||||
|
|
||||||
it('can fork documents', async function() {
|
it('can fork documents', async function() {
|
||||||
@ -717,35 +738,35 @@ describe('HostedStorageManager', function() {
|
|||||||
await workers.assignDocWorker(docId);
|
await workers.assignDocWorker(docId);
|
||||||
await workers.assignDocWorker(forkId);
|
await workers.assignDocWorker(forkId);
|
||||||
|
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
await useFixtureDoc('Hello.grist', store.storageManager, `${docId}.grist`);
|
await useFixtureDoc('Hello.grist', store.storageManager, `${docId}.grist`);
|
||||||
let doc = await store.docManager.fetchDoc(docSession, docId);
|
const doc = await store.docManager.fetchDoc(docSession, docId);
|
||||||
await doc.docStorage.exec("update Table1 set A = 'trunk' where id = 1");
|
await doc.docStorage.exec("update Table1 set A = 'trunk' where id = 1");
|
||||||
await store.end();
|
});
|
||||||
|
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
await store.docManager.storageManager.prepareFork(docId, forkId);
|
await store.docManager.storageManager.prepareFork(docId, forkId);
|
||||||
doc = await store.docManager.fetchDoc(docSession, forkId);
|
const doc = await store.docManager.fetchDoc(docSession, forkId);
|
||||||
assert.equal('trunk', (await doc.docStorage.get("select A from Table1 where id = 1"))!.A);
|
assert.equal('trunk', (await doc.docStorage.get("select A from Table1 where id = 1"))!.A);
|
||||||
await doc.docStorage.exec("update Table1 set A = 'fork' where id = 1");
|
await doc.docStorage.exec("update Table1 set A = 'fork' where id = 1");
|
||||||
await store.end();
|
});
|
||||||
|
|
||||||
await store.removeAll();
|
await store.removeAll();
|
||||||
|
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
doc = await store.docManager.fetchDoc(docSession, docId);
|
let doc = await store.docManager.fetchDoc(docSession, docId);
|
||||||
assert.equal('trunk', (await doc.docStorage.get("select A from Table1 where id = 1"))!.A);
|
assert.equal('trunk', (await doc.docStorage.get("select A from Table1 where id = 1"))!.A);
|
||||||
doc = await store.docManager.fetchDoc(docSession, forkId);
|
doc = await store.docManager.fetchDoc(docSession, forkId);
|
||||||
assert.equal('fork', (await doc.docStorage.get("select A from Table1 where id = 1"))!.A);
|
assert.equal('fork', (await doc.docStorage.get("select A from Table1 where id = 1"))!.A);
|
||||||
await store.end();
|
});
|
||||||
|
|
||||||
// Check that the trunk can be replaced by a fork
|
// Check that the trunk can be replaced by a fork
|
||||||
await store.removeAll();
|
await store.removeAll();
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
await store.storageManager.replace(docId, {sourceDocId: forkId});
|
await store.storageManager.replace(docId, {sourceDocId: forkId});
|
||||||
doc = await store.docManager.fetchDoc(docSession, docId);
|
const doc = await store.docManager.fetchDoc(docSession, docId);
|
||||||
assert.equal('fork', (await doc.docStorage.get("select A from Table1 where id = 1"))!.A);
|
assert.equal('fork', (await doc.docStorage.get("select A from Table1 where id = 1"))!.A);
|
||||||
await store.end();
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
it('can persist a fork with no modifications', async function() {
|
it('can persist a fork with no modifications', async function() {
|
||||||
@ -755,16 +776,16 @@ describe('HostedStorageManager', function() {
|
|||||||
await workers.assignDocWorker(forkId);
|
await workers.assignDocWorker(forkId);
|
||||||
|
|
||||||
// Create a document.
|
// Create a document.
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
await useFixtureDoc('Hello.grist', store.storageManager, `${docId}.grist`);
|
await useFixtureDoc('Hello.grist', store.storageManager, `${docId}.grist`);
|
||||||
let doc = await store.docManager.fetchDoc(docSession, docId);
|
const doc = await store.docManager.fetchDoc(docSession, docId);
|
||||||
await doc.docStorage.exec("update Table1 set A = 'trunk' where id = 1");
|
await doc.docStorage.exec("update Table1 set A = 'trunk' where id = 1");
|
||||||
await store.end();
|
});
|
||||||
|
|
||||||
// Create a fork with no modifications.
|
// Create a fork with no modifications.
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
await store.docManager.storageManager.prepareFork(docId, forkId);
|
await store.docManager.storageManager.prepareFork(docId, forkId);
|
||||||
await store.end();
|
});
|
||||||
await store.waitForUpdates();
|
await store.waitForUpdates();
|
||||||
await store.removeAll();
|
await store.removeAll();
|
||||||
|
|
||||||
@ -772,10 +793,10 @@ describe('HostedStorageManager', function() {
|
|||||||
await fse.remove(store.getDocPath(docId));
|
await fse.remove(store.getDocPath(docId));
|
||||||
|
|
||||||
// Make sure opening the fork works as expected.
|
// Make sure opening the fork works as expected.
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
doc = await store.docManager.fetchDoc(docSession, forkId);
|
const doc = await store.docManager.fetchDoc(docSession, forkId);
|
||||||
assert.equal('trunk', (await doc.docStorage.get("select A from Table1 where id = 1"))!.A);
|
assert.equal('trunk', (await doc.docStorage.get("select A from Table1 where id = 1"))!.A);
|
||||||
await store.end();
|
});
|
||||||
await store.removeAll();
|
await store.removeAll();
|
||||||
});
|
});
|
||||||
|
|
||||||
@ -792,35 +813,36 @@ describe('HostedStorageManager', function() {
|
|||||||
await workers.assignDocWorker(forkId2);
|
await workers.assignDocWorker(forkId2);
|
||||||
await workers.assignDocWorker(forkId3);
|
await workers.assignDocWorker(forkId3);
|
||||||
|
|
||||||
await store.begin();
|
const doc = await store.run(async () => {
|
||||||
await useFixtureDoc('Hello.grist', store.storageManager, `${docId}.grist`);
|
await useFixtureDoc('Hello.grist', store.storageManager, `${docId}.grist`);
|
||||||
let doc = await store.docManager.fetchDoc(docSession, docId);
|
const doc = await store.docManager.fetchDoc(docSession, docId);
|
||||||
await doc.waitForInitialization();
|
await doc.waitForInitialization();
|
||||||
for (let i = 0; i < forks; i++) {
|
for (let i = 0; i < forks; i++) {
|
||||||
await doc.docStorage.exec(`update Table1 set A = 'v${i}' where id = 1`);
|
await doc.docStorage.exec(`update Table1 set A = 'v${i}' where id = 1`);
|
||||||
await doc.testKeepOpen();
|
await doc.testKeepOpen();
|
||||||
await store.waitForUpdates();
|
await store.waitForUpdates();
|
||||||
}
|
}
|
||||||
await store.end();
|
return doc;
|
||||||
|
});
|
||||||
|
|
||||||
const {snapshots} = await store.storageManager.getSnapshots(doc.docName);
|
const {snapshots} = await store.storageManager.getSnapshots(doc.docName);
|
||||||
assert.isAtLeast(snapshots.length, forks + 1); // May be 1 greater depending on how long
|
assert.isAtLeast(snapshots.length, forks + 1); // May be 1 greater depending on how long
|
||||||
// it takes to run initial migrations.
|
// it takes to run initial migrations.
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
for (let i = forks - 1; i >= 0; i--) {
|
for (let i = forks - 1; i >= 0; i--) {
|
||||||
const snapshot = snapshots.shift()!;
|
const snapshot = snapshots.shift()!;
|
||||||
const forkId = snapshot.docId;
|
const forkId = snapshot.docId;
|
||||||
await workers.assignDocWorker(forkId);
|
await workers.assignDocWorker(forkId);
|
||||||
doc = await store.docManager.fetchDoc(docSession, forkId);
|
const doc = await store.docManager.fetchDoc(docSession, forkId);
|
||||||
assert.equal(`v${i}`, (await doc.docStorage.get("select A from Table1 where id = 1"))!.A);
|
assert.equal(`v${i}`, (await doc.docStorage.get("select A from Table1 where id = 1"))!.A);
|
||||||
}
|
}
|
||||||
await store.end();
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
it('can access snapshots with old schema versions', async function() {
|
it('can access snapshots with old schema versions', async function() {
|
||||||
const snapshotId = `World~v=1`;
|
const snapshotId = `World~v=1`;
|
||||||
await workers.assignDocWorker(snapshotId);
|
await workers.assignDocWorker(snapshotId);
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
// Pretend we have a snapshot of World-v33.grist and fetch/load it.
|
// Pretend we have a snapshot of World-v33.grist and fetch/load it.
|
||||||
await useFixtureDoc('World-v33.grist', store.storageManager, `${snapshotId}.grist`);
|
await useFixtureDoc('World-v33.grist', store.storageManager, `${snapshotId}.grist`);
|
||||||
const doc = await store.docManager.fetchDoc(docSession, snapshotId);
|
const doc = await store.docManager.fetchDoc(docSession, snapshotId);
|
||||||
@ -839,14 +861,14 @@ describe('HostedStorageManager', function() {
|
|||||||
/Snapshots cannot be replaced/);
|
/Snapshots cannot be replaced/);
|
||||||
await assert.isRejected(doc.applyUserActions(docSession, [['AddTable', 'NewTable', [{id: 'A'}]]]),
|
await assert.isRejected(doc.applyUserActions(docSession, [['AddTable', 'NewTable', [{id: 'A'}]]]),
|
||||||
/pyCall is not available in snapshots/);
|
/pyCall is not available in snapshots/);
|
||||||
await store.end();
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
it('can prune snapshots', async function() {
|
it('can prune snapshots', async function() {
|
||||||
const versions = 8;
|
const versions = 8;
|
||||||
|
|
||||||
const docId = `create-${uuidv4()}`;
|
const docId = `create-${uuidv4()}`;
|
||||||
await store.begin();
|
const doc = await store.run(async () => {
|
||||||
await useFixtureDoc('Hello.grist', store.storageManager, `${docId}.grist`);
|
await useFixtureDoc('Hello.grist', store.storageManager, `${docId}.grist`);
|
||||||
const doc = await store.docManager.fetchDoc(docSession, docId);
|
const doc = await store.docManager.fetchDoc(docSession, docId);
|
||||||
for (let i = 0; i < versions; i++) {
|
for (let i = 0; i < versions; i++) {
|
||||||
@ -855,7 +877,8 @@ describe('HostedStorageManager', function() {
|
|||||||
await store.waitForUpdates();
|
await store.waitForUpdates();
|
||||||
}
|
}
|
||||||
await store.storageManager.testWaitForPrunes();
|
await store.storageManager.testWaitForPrunes();
|
||||||
await store.end();
|
return doc;
|
||||||
|
});
|
||||||
await waitForIt(async () => {
|
await waitForIt(async () => {
|
||||||
const {snapshots} = await store.storageManager.getSnapshots(doc.docName);
|
const {snapshots} = await store.storageManager.getSnapshots(doc.docName);
|
||||||
// Should be keeping at least five, and then maybe 1 more if the hour changed
|
// Should be keeping at least five, and then maybe 1 more if the hour changed
|
||||||
@ -878,20 +901,20 @@ describe('HostedStorageManager', function() {
|
|||||||
|
|
||||||
// Create a series of versions of a document, and fetch them sequentially
|
// Create a series of versions of a document, and fetch them sequentially
|
||||||
// so that they are potentially available as stale values.
|
// so that they are potentially available as stale values.
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
await useFixtureDoc('Hello.grist', store.storageManager, `${docId}.grist`);
|
await useFixtureDoc('Hello.grist', store.storageManager, `${docId}.grist`);
|
||||||
let doc = await store.docManager.fetchDoc(docSession, docId);
|
await store.docManager.fetchDoc(docSession, docId);
|
||||||
await store.end();
|
});
|
||||||
for (let i = 0; i < 3; i++) {
|
for (let i = 0; i < 3; i++) {
|
||||||
await store.removeAll();
|
await store.removeAll();
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
doc = await store.docManager.fetchDoc(docSession, docId);
|
const doc = await store.docManager.fetchDoc(docSession, docId);
|
||||||
if (i > 0) {
|
if (i > 0) {
|
||||||
const prev = await doc.docStorage.get("select A from Table1 where id = 1");
|
const prev = await doc.docStorage.get("select A from Table1 where id = 1");
|
||||||
assert.equal(prev!.A, `magic_word${i - 1}`);
|
assert.equal(prev!.A, `magic_word${i - 1}`);
|
||||||
}
|
}
|
||||||
await doc.docStorage.exec(`update Table1 set A = 'magic_word${i}' where id = 1`);
|
await doc.docStorage.exec(`update Table1 set A = 'magic_word${i}' where id = 1`);
|
||||||
await store.end();
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
// Wipe all checksums and make sure (1) we don't get any errors and (2) the
|
// Wipe all checksums and make sure (1) we don't get any errors and (2) the
|
||||||
@ -903,10 +926,10 @@ describe('HostedStorageManager', function() {
|
|||||||
// Optionally wipe all local files.
|
// Optionally wipe all local files.
|
||||||
await store.removeAll();
|
await store.removeAll();
|
||||||
}
|
}
|
||||||
await store.begin();
|
await store.run(async () => {
|
||||||
doc = await store.docManager.fetchDoc(docSession, docId);
|
const doc = await store.docManager.fetchDoc(docSession, docId);
|
||||||
result = (await doc.docStorage.get("select A from Table1 where id = 1"))?.A;
|
result = (await doc.docStorage.get("select A from Table1 where id = 1"))?.A;
|
||||||
await store.end();
|
});
|
||||||
if (result !== 'magic_word2') {
|
if (result !== 'magic_word2') {
|
||||||
throw new Error(`inconsistent result: ${result}`);
|
throw new Error(`inconsistent result: ${result}`);
|
||||||
}
|
}
|
||||||
@ -917,7 +940,7 @@ describe('HostedStorageManager', function() {
|
|||||||
|
|
||||||
it('can access metadata', async function() {
|
it('can access metadata', async function() {
|
||||||
const docId = `create-${uuidv4()}`;
|
const docId = `create-${uuidv4()}`;
|
||||||
await store.begin();
|
const { tz, h, doc } = await store.run(async () => {
|
||||||
// Use a doc that's up-to-date on storage migrations, but needs a python schema migration.
|
// Use a doc that's up-to-date on storage migrations, but needs a python schema migration.
|
||||||
await useFixtureDoc('BlobMigrationV8.grist', store.storageManager, `${docId}.grist`);
|
await useFixtureDoc('BlobMigrationV8.grist', store.storageManager, `${docId}.grist`);
|
||||||
const doc = await store.docManager.fetchDoc(docSession, docId);
|
const doc = await store.docManager.fetchDoc(docSession, docId);
|
||||||
@ -926,7 +949,8 @@ describe('HostedStorageManager', function() {
|
|||||||
const tz = rec.tableData[3].timezone[0];
|
const tz = rec.tableData[3].timezone[0];
|
||||||
const h = (await doc.getRecentStates(makeExceptionalDocSession('system')))[0].h;
|
const h = (await doc.getRecentStates(makeExceptionalDocSession('system')))[0].h;
|
||||||
await store.docManager.makeBackup(doc, 'hello');
|
await store.docManager.makeBackup(doc, 'hello');
|
||||||
await store.end();
|
return { tz, h, doc };
|
||||||
|
});
|
||||||
const {snapshots} = await store.storageManager.getSnapshots(doc.docName);
|
const {snapshots} = await store.storageManager.getSnapshots(doc.docName);
|
||||||
assert.equal(snapshots[0]?.metadata?.label, 'hello');
|
assert.equal(snapshots[0]?.metadata?.label, 'hello');
|
||||||
// There can be extra snapshots, depending on timing.
|
// There can be extra snapshots, depending on timing.
|
||||||
|
Loading…
Reference in New Issue
Block a user