212 lines
5 KiB
JavaScript
212 lines
5 KiB
JavaScript
const PARSER_PACKAGE = "./wheel/lrparsers-0.7.9-py3-none-any.whl"
|
|
|
|
|
|
// Load the whole pyodide thingy.
|
|
importScripts("pyodide/pyodide.js");
|
|
|
|
function data_to_js(thing) {
|
|
if (thing.toJs) {
|
|
pyproxies = [];
|
|
thing = thing.toJs({
|
|
pyproxies,
|
|
dict_converter: Object.fromEntries,
|
|
create_pyproxies: false,
|
|
});
|
|
for (p of pyproxies) {
|
|
p.destroy();
|
|
}
|
|
}
|
|
return thing;
|
|
}
|
|
|
|
const dingus_module = {
|
|
post_grammar_status: function (message) {
|
|
console.log("Grammar Status:", message);
|
|
postMessage({kind: "grammar", status: "loading", message});
|
|
},
|
|
|
|
post_grammar_loaded: function (name) {
|
|
const msg = "Grammar '" + name + "' loaded";
|
|
console.log(msg);
|
|
postMessage({kind: "grammar", status: "ok", message: msg});
|
|
},
|
|
|
|
post_grammar_error: function(errors) {
|
|
errors = data_to_js(errors);
|
|
console.log("Grammar Error:", errors);
|
|
postMessage({
|
|
kind:"grammar",
|
|
status: "error",
|
|
message: "An error occurred loading the grammar",
|
|
errors: errors,
|
|
});
|
|
},
|
|
|
|
post_doc_parse: function(tree, errors) {
|
|
tree = data_to_js(tree);
|
|
errors = data_to_js(errors);
|
|
|
|
console.log("Doc parse:", tree, errors);
|
|
postMessage({
|
|
kind: "input",
|
|
status: "ok",
|
|
message: "Parsed",
|
|
errors: errors,
|
|
tree: tree,
|
|
});
|
|
},
|
|
|
|
post_doc_error: function(errors) {
|
|
errors = data_to_js(errors);
|
|
console.log("Doc Error:", errors);
|
|
postMessage({
|
|
kind:"input",
|
|
status: "error",
|
|
message: "An error occurred parsing the document",
|
|
errors: errors,
|
|
});
|
|
},
|
|
};
|
|
|
|
async function setup_python() {
|
|
dingus_module.post_grammar_status("Loading python....");
|
|
const pyodide = await loadPyodide({
|
|
packages: ["micropip"],
|
|
});
|
|
pyodide.setStdout({ batched: (msg) => console.log(msg) }); // TODO: I know this is an option above.
|
|
|
|
// TODO: Do I actually want micropip? Probably not?
|
|
dingus_module.post_grammar_status("Installing parser package...");
|
|
const micropip = pyodide.pyimport("micropip");
|
|
await micropip.install(PARSER_PACKAGE);
|
|
|
|
dingus_module.post_grammar_status("Configuring dingus...");
|
|
pyodide.registerJsModule("dingus", dingus_module);
|
|
|
|
pyodide.runPython(`
|
|
import traceback
|
|
|
|
import dingus
|
|
import parser
|
|
import parser.runtime as runtime
|
|
import pyodide.code
|
|
import pyodide.ffi as ffi
|
|
|
|
GRAMMAR = None
|
|
PARSE_TABLE = None
|
|
LEXER = None
|
|
|
|
def eval_grammar(code):
|
|
global GRAMMAR
|
|
global PARSE_TABLE
|
|
global LEXER
|
|
|
|
try:
|
|
dingus.post_grammar_status("Evaluating grammar...")
|
|
print("Hey?")
|
|
grammar_globals={}
|
|
pyodide.code.eval_code(code, globals=grammar_globals)
|
|
|
|
grammar = None
|
|
for key, value in grammar_globals.items():
|
|
if isinstance(value, type) and issubclass(value, parser.Grammar) and value is not parser.Grammar:
|
|
value = value()
|
|
|
|
if isinstance(value, parser.Grammar):
|
|
if grammar is None:
|
|
grammar = value
|
|
else:
|
|
raise Exception("More than one Grammar found in the file")
|
|
|
|
if grammar is None:
|
|
raise Exception("No grammar definition, define or instantiate a class that inherits from parser.Grammar")
|
|
|
|
GRAMMAR = grammar
|
|
|
|
dingus.post_grammar_status("Building parse table...")
|
|
PARSE_TABLE = grammar.build_table()
|
|
|
|
dingus.post_grammar_status("Building lexer...")
|
|
LEXER = grammar.compile_lexer()
|
|
|
|
dingus.post_grammar_loaded(grammar.name)
|
|
|
|
except Exception as e:
|
|
ohno = traceback.format_exc()
|
|
print(f"grammar: {ohno}")
|
|
dingus.post_grammar_error(ohno.splitlines())
|
|
|
|
def tree_to_js(tree):
|
|
if tree is None:
|
|
return None
|
|
|
|
elif isinstance(tree, runtime.Tree):
|
|
return {
|
|
"kind": "tree",
|
|
"name": tree.name,
|
|
"start": tree.start,
|
|
"end": tree.end,
|
|
"children": [tree_to_js(child) for child in tree.children],
|
|
}
|
|
|
|
else:
|
|
return {
|
|
"kind": "token",
|
|
"name": tree.kind,
|
|
"start": tree.start,
|
|
"end": tree.end,
|
|
}
|
|
|
|
def eval_document(code):
|
|
global PARSE_TABLE
|
|
global LEXER
|
|
|
|
try:
|
|
tree, errors = runtime.parse(PARSE_TABLE, LEXER, code)
|
|
dingus.post_doc_parse(tree_to_js(tree), errors)
|
|
except Exception as e:
|
|
ohno = traceback.format_exc()
|
|
print(f"doc: {ohno}")
|
|
dingus.post_doc_error(ohno.splitlines())
|
|
`);
|
|
|
|
dingus_module.post_grammar_status("Ready.");
|
|
self.pyodide = pyodide;
|
|
return pyodide;
|
|
}
|
|
const pyodide_promise = setup_python();
|
|
|
|
|
|
async function load_grammar_module(code) {
|
|
const pyodide = self.pyodide;
|
|
|
|
// console.log("Running...");
|
|
const my_fn = pyodide.globals.get("eval_grammar");
|
|
my_fn(code);
|
|
my_fn.destroy();
|
|
}
|
|
|
|
async function parse_document(code) {
|
|
const pyodide = self.pyodide;
|
|
|
|
// console.log("Running...");
|
|
const my_fn = pyodide.globals.get("eval_document");
|
|
my_fn(code);
|
|
my_fn.destroy();
|
|
}
|
|
|
|
self.onmessage = async function(event) {
|
|
await pyodide_promise;
|
|
|
|
try {
|
|
const { kind, data } = event.data;
|
|
if (kind === "grammar") {
|
|
await load_grammar_module(data);
|
|
} else if (kind === "input") {
|
|
await parse_document(data);
|
|
}
|
|
} catch (e) {
|
|
console.log("INTERNAL ERROR: ", e.message);
|
|
postMessage({error: e.message});
|
|
}
|
|
};
|