mirror of
https://github.com/gristlabs/grist-core.git
synced 2024-10-27 20:44:07 +00:00
66643a5e6b
This adds a new `GRIST_SANDBOX_FLAVOR=pyodide` option where the version of Python used for the data engine is wasm, and so can be run by node like the rest of the back end. It still runs as a separate process. There are a few small version changes made to packages to avoid various awkwardnesses present in the current versions. All existing tests pass. This is very experimental. To use, you'll need something with a bash shell and make. First do: ``` cd sandbox/pyodide make setup # README.md and Makefile have details cd .. ``` Then running Grist as: ``` GRIST_SANDBOX_FLAVOR=pyodide yarn start ``` should work. Adding a formula with content: ``` import sys; return sys.version ``` should return a different Python version than other sandboxes. The motivation for this work is to have a form of sandboxing that will work on Windows for Grist Electron (for Linux we have gvisor/runsc, for Mac we have sandbox-exec, but I haven't found anything comparable for Windows). It also brings a back-end-free version of Grist a bit closer, for use-cases where that would make sense - such as serving a report (in the form of a Grist document) on a static site.
118 lines
3.6 KiB
JavaScript
118 lines
3.6 KiB
JavaScript
const path = require('path');
|
|
const fs = require('fs');
|
|
const fetch = require('node-fetch');
|
|
|
|
async function listLibs(src) {
|
|
const txt = fs.readFileSync(path.join(__dirname, '..', 'requirements3.txt'), 'utf8');
|
|
const libs = {};
|
|
for (const line of txt.split(/\r?\n/)) {
|
|
const raw = line.split('#')[0];
|
|
if (!raw.includes('==')) { continue; }
|
|
const [name, version] = line.split('==');
|
|
libs[name] = version;
|
|
}
|
|
const hits = [];
|
|
const misses = [];
|
|
const toLoad = [];
|
|
const material = fs.readdirSync(src);
|
|
for (const [lib, version] of Object.entries(libs)) {
|
|
const nlib = lib.replace(/-/g, '_');
|
|
const info = {
|
|
name: lib,
|
|
standardName: nlib,
|
|
version: version,
|
|
}
|
|
try {
|
|
const found = material.filter(m => m.startsWith(`${nlib}-${version}-`));
|
|
if (found.length !== 1) {
|
|
throw new Error('did not find 1');
|
|
}
|
|
const fname = found[0];
|
|
info.fullName = path.join(src, fname);
|
|
info.fileName = fname;
|
|
toLoad.push(info);
|
|
hits.push(lib);
|
|
} catch (e) {
|
|
misses.push(info);
|
|
}
|
|
}
|
|
return {
|
|
available: toLoad,
|
|
misses,
|
|
};
|
|
}
|
|
exports.listLibs = listLibs;
|
|
|
|
async function findOnDisk(src, dest) {
|
|
console.log(`Organizing packages on disk`, {src, dest});
|
|
fs.mkdirSync(dest, {recursive: true});
|
|
let libs = (await listLibs(src));
|
|
for (const lib of libs.available) {
|
|
fs.copyFileSync(lib.fullName, path.join(dest, lib.fileName));
|
|
fs.writeFileSync(path.join(dest, `${lib.name}-${lib.version}.json`),
|
|
JSON.stringify({
|
|
name: lib.name,
|
|
version: lib.version,
|
|
fileName: lib.fileName,
|
|
}, null, 2));
|
|
console.log("Copied", {
|
|
content: path.join(dest, lib.fileName),
|
|
meta: path.join(dest, `${lib.name}-${lib.version}.json`),
|
|
});
|
|
}
|
|
libs = await listLibs(dest);
|
|
console.log(`Cached`, {libs: libs.available.map(lib => lib.name)});
|
|
console.log(`Missing`, {libs: libs.misses.map(lib => lib.name)});
|
|
}
|
|
|
|
async function findOnNet(src, dest) {
|
|
console.log(`Caching packages on disk`, {src, dest});
|
|
fs.mkdirSync(dest, {recursive: true});
|
|
let libs = await listLibs(dest);
|
|
console.log(`Cached`, {libs: libs.available.map(lib => lib.name)});
|
|
for (const lib of libs.misses) {
|
|
console.log('Fetching', lib);
|
|
const url = new URL(src);
|
|
url.pathname = url.pathname + lib.name + '-' + lib.version + '.json';
|
|
const result = await fetch(url.href);
|
|
if (result.status === 200) {
|
|
const data = await result.json();
|
|
const url2 = new URL(src);
|
|
url2.pathname = url2.pathname + data.fileName;
|
|
const result2 = await fetch(url2.href);
|
|
if (result2.status === 200) {
|
|
fs.writeFileSync(path.join(dest, `${lib.name}-${lib.version}.json`),
|
|
JSON.stringify(data, null, 2));
|
|
fs.writeFileSync(path.join(dest, data.fileName),
|
|
await result2.buffer());
|
|
} else {
|
|
console.error("No payload available", {lib});
|
|
}
|
|
} else {
|
|
console.error("No metadata available", {lib});
|
|
}
|
|
}
|
|
libs = await listLibs(dest);
|
|
console.log(`Missing`, {libs: libs.misses.map(lib => lib.name)});
|
|
}
|
|
|
|
async function main(src, dest) {
|
|
if (!src) {
|
|
console.error('please supply a source');
|
|
process.exit(1);
|
|
}
|
|
if (!dest) {
|
|
console.error('please supply a destination');
|
|
process.exit(1);
|
|
}
|
|
if (src.startsWith('http:') || src.startsWith('https:')) {
|
|
await findOnNet(src, dest);
|
|
return;
|
|
}
|
|
await findOnDisk(src, dest);
|
|
}
|
|
|
|
if (require.main === module) {
|
|
main(...process.argv.slice(2)).catch(e => console.error(e));
|
|
}
|