"use strict"; Object.defineProperty(exports, "__esModule", { value: true }); exports.parse = parse; var _helperCodeFrame = require("@webassemblyjs/helper-code-frame"); var t = _interopRequireWildcard(require("@webassemblyjs/ast")); var _numberLiterals = require("./number-literals"); var _stringLiterals = require("./string-literals"); var _tokenizer = require("./tokenizer"); function _interopRequireWildcard(obj) { if (obj && obj.__esModule) { return obj; } else { var newObj = {}; if (obj != null) { for (var key in obj) { if (Object.prototype.hasOwnProperty.call(obj, key)) { var desc = Object.defineProperty && Object.getOwnPropertyDescriptor ? Object.getOwnPropertyDescriptor(obj, key) : {}; if (desc.get || desc.set) { Object.defineProperty(newObj, key, desc); } else { newObj[key] = obj[key]; } } } } newObj.default = obj; return newObj; } } function _typeof(obj) { if (typeof Symbol === "function" && typeof Symbol.iterator === "symbol") { _typeof = function _typeof(obj) { return typeof obj; }; } else { _typeof = function _typeof(obj) { return obj && typeof Symbol === "function" && obj.constructor === Symbol && obj !== Symbol.prototype ? "symbol" : typeof obj; }; } return _typeof(obj); } function _toConsumableArray(arr) { if (Array.isArray(arr)) { for (var i = 0, arr2 = new Array(arr.length); i < arr.length; i++) { arr2[i] = arr[i]; } return arr2; } else { return Array.from(arr); } } function hasPlugin(name) { if (name !== "wast") throw new Error("unknow plugin"); return true; } function isKeyword(token, id) { return token.type === _tokenizer.tokens.keyword && token.value === id; } function tokenToString(token) { if (token.type === "keyword") { return "keyword (".concat(token.value, ")"); } return token.type; } function identifierFromToken(token) { var _token$loc = token.loc, end = _token$loc.end, start = _token$loc.start; return t.withLoc(t.identifier(token.value), end, start); } function parse(tokensList, source) { var current = 0; var getUniqueName = t.getUniqueNameGenerator(); var state = { registredExportedElements: [] }; // But this time we're going to use recursion instead of a `while` loop. So we // define a `walk` function. function walk() { var token = tokensList[current]; function eatToken() { token = tokensList[++current]; } function getEndLoc() { var currentToken = token; if (typeof currentToken === "undefined") { var lastToken = tokensList[tokensList.length - 1]; currentToken = lastToken; } return currentToken.loc.end; } function getStartLoc() { return token.loc.start; } function eatTokenOfType(type) { if (token.type !== type) { throw new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "Assertion error: expected token of type " + type + ", given " + tokenToString(token)); } eatToken(); } function parseExportIndex(token) { if (token.type === _tokenizer.tokens.identifier) { var index = identifierFromToken(token); eatToken(); return index; } else if (token.type === _tokenizer.tokens.number) { var _index = t.numberLiteralFromRaw(token.value); eatToken(); return _index; } else { throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "unknown export index" + ", given " + tokenToString(token)); }(); } } function lookaheadAndCheck() { var len = arguments.length; for (var i = 0; i < len; i++) { var tokenAhead = tokensList[current + i]; var expectedToken = i < 0 || arguments.length <= i ? undefined : arguments[i]; if (tokenAhead.type === "keyword") { if (isKeyword(tokenAhead, expectedToken) === false) { return false; } } else if (expectedToken !== tokenAhead.type) { return false; } } return true; } // TODO(sven): there is probably a better way to do this // can refactor it if it get out of hands function maybeIgnoreComment() { if (typeof token === "undefined") { // Ignore return; } while (token.type === _tokenizer.tokens.comment) { eatToken(); if (typeof token === "undefined") { // Hit the end break; } } } /** * Parses a memory instruction * * WAST: * * memory: ( memory ? ) * ( memory ? ( export ) <...> ) * ( memory ? ( import ) ) * ( memory ? ( export )* ( data * ) * memory_sig: ? * */ function parseMemory() { var id = t.identifier(getUniqueName("memory")); var limits = t.limit(0); if (token.type === _tokenizer.tokens.string || token.type === _tokenizer.tokens.identifier) { id = t.identifier(token.value); eatToken(); } else { id = t.withRaw(id, ""); // preserve anonymous } /** * Maybe data */ if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.data)) { eatToken(); // ( eatToken(); // data // TODO(sven): do something with the data collected here var stringInitializer = token.value; eatTokenOfType(_tokenizer.tokens.string); // Update limits accordingly limits = t.limit(stringInitializer.length); eatTokenOfType(_tokenizer.tokens.closeParen); } /** * Maybe export */ if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.export)) { eatToken(); // ( eatToken(); // export if (token.type !== _tokenizer.tokens.string) { throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Expected string in export" + ", given " + tokenToString(token)); }(); } var _name = token.value; eatToken(); state.registredExportedElements.push({ exportType: "Memory", name: _name, id: id }); eatTokenOfType(_tokenizer.tokens.closeParen); } /** * Memory signature */ if (token.type === _tokenizer.tokens.number) { limits = t.limit((0, _numberLiterals.parse32I)(token.value)); eatToken(); if (token.type === _tokenizer.tokens.number) { limits.max = (0, _numberLiterals.parse32I)(token.value); eatToken(); } } return t.memory(limits, id); } /** * Parses a data section * https://webassembly.github.io/spec/core/text/modules.html#data-segments * * WAST: * * data: ( data ? ) */ function parseData() { // optional memory index var memidx = 0; if (token.type === _tokenizer.tokens.number) { memidx = token.value; eatTokenOfType(_tokenizer.tokens.number); // . } eatTokenOfType(_tokenizer.tokens.openParen); var offset; if (token.type === _tokenizer.tokens.valtype) { eatTokenOfType(_tokenizer.tokens.valtype); // i32 eatTokenOfType(_tokenizer.tokens.dot); // . if (token.value !== "const") { throw new Error("constant expression required"); } eatTokenOfType(_tokenizer.tokens.name); // const var numberLiteral = t.numberLiteralFromRaw(token.value, "i32"); offset = t.objectInstruction("const", "i32", [numberLiteral]); eatToken(); eatTokenOfType(_tokenizer.tokens.closeParen); } else { eatTokenOfType(_tokenizer.tokens.name); // get_global var _numberLiteral = t.numberLiteralFromRaw(token.value, "i32"); offset = t.instruction("get_global", [_numberLiteral]); eatToken(); eatTokenOfType(_tokenizer.tokens.closeParen); } var byteArray = (0, _stringLiterals.parseString)(token.value); eatToken(); // "string" return t.data(t.memIndexLiteral(memidx), offset, t.byteArray(byteArray)); } /** * Parses a table instruction * * WAST: * * table: ( table ? ) * ( table ? ( export ) <...> ) * ( table ? ( import ) ) * ( table ? ( export )* ( elem * ) ) * * table_type: ? * elem_type: anyfunc * * elem: ( elem ? (offset * ) * ) * ( elem ? * ) */ function parseTable() { var name = t.identifier(getUniqueName("table")); var limit = t.limit(0); var elemIndices = []; var elemType = "anyfunc"; if (token.type === _tokenizer.tokens.string || token.type === _tokenizer.tokens.identifier) { name = identifierFromToken(token); eatToken(); } else { name = t.withRaw(name, ""); // preserve anonymous } while (token.type !== _tokenizer.tokens.closeParen) { /** * Maybe export */ if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.elem)) { eatToken(); // ( eatToken(); // elem while (token.type === _tokenizer.tokens.identifier) { elemIndices.push(t.identifier(token.value)); eatToken(); } eatTokenOfType(_tokenizer.tokens.closeParen); } else if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.export)) { eatToken(); // ( eatToken(); // export if (token.type !== _tokenizer.tokens.string) { throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Expected string in export" + ", given " + tokenToString(token)); }(); } var exportName = token.value; eatToken(); state.registredExportedElements.push({ exportType: "Table", name: exportName, id: name }); eatTokenOfType(_tokenizer.tokens.closeParen); } else if (isKeyword(token, _tokenizer.keywords.anyfunc)) { // It's the default value, we can ignore it eatToken(); // anyfunc } else if (token.type === _tokenizer.tokens.number) { /** * Table type */ var min = parseInt(token.value); eatToken(); if (token.type === _tokenizer.tokens.number) { var max = parseInt(token.value); eatToken(); limit = t.limit(min, max); } else { limit = t.limit(min); } eatToken(); } else { throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token" + ", given " + tokenToString(token)); }(); } } if (elemIndices.length > 0) { return t.table(elemType, limit, name, elemIndices); } else { return t.table(elemType, limit, name); } } /** * Parses an import statement * * WAST: * * import: ( import ) * imkind: ( func ? ) * ( global ? ) * ( table ? ) * ( memory ? ) * * global_sig: | ( mut ) */ function parseImport() { if (token.type !== _tokenizer.tokens.string) { throw new Error("Expected a string, " + token.type + " given."); } var moduleName = token.value; eatToken(); if (token.type !== _tokenizer.tokens.string) { throw new Error("Expected a string, " + token.type + " given."); } var name = token.value; eatToken(); eatTokenOfType(_tokenizer.tokens.openParen); var descr; if (isKeyword(token, _tokenizer.keywords.func)) { eatToken(); // keyword var fnParams = []; var fnResult = []; var typeRef; var fnName = t.identifier(getUniqueName("func")); if (token.type === _tokenizer.tokens.identifier) { fnName = identifierFromToken(token); eatToken(); } while (token.type === _tokenizer.tokens.openParen) { eatToken(); if (lookaheadAndCheck(_tokenizer.keywords.type) === true) { eatToken(); typeRef = parseTypeReference(); } else if (lookaheadAndCheck(_tokenizer.keywords.param) === true) { eatToken(); fnParams.push.apply(fnParams, _toConsumableArray(parseFuncParam())); } else if (lookaheadAndCheck(_tokenizer.keywords.result) === true) { eatToken(); fnResult.push.apply(fnResult, _toConsumableArray(parseFuncResult())); } else { throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in import of type" + ", given " + tokenToString(token)); }(); } eatTokenOfType(_tokenizer.tokens.closeParen); } if (typeof fnName === "undefined") { throw new Error("Imported function must have a name"); } descr = t.funcImportDescr(fnName, typeRef !== undefined ? typeRef : t.signature(fnParams, fnResult)); } else if (isKeyword(token, _tokenizer.keywords.global)) { eatToken(); // keyword if (token.type === _tokenizer.tokens.openParen) { eatToken(); // ( eatTokenOfType(_tokenizer.tokens.keyword); // mut keyword var valtype = token.value; eatToken(); descr = t.globalType(valtype, "var"); eatTokenOfType(_tokenizer.tokens.closeParen); } else { var _valtype = token.value; eatTokenOfType(_tokenizer.tokens.valtype); descr = t.globalType(_valtype, "const"); } } else if (isKeyword(token, _tokenizer.keywords.memory) === true) { eatToken(); // Keyword descr = parseMemory(); } else if (isKeyword(token, _tokenizer.keywords.table) === true) { eatToken(); // Keyword descr = parseTable(); } else { throw new Error("Unsupported import type: " + tokenToString(token)); } eatTokenOfType(_tokenizer.tokens.closeParen); return t.moduleImport(moduleName, name, descr); } /** * Parses a block instruction * * WAST: * * expr: ( block ? * ) * instr: block ? * end ? * block_sig : ( result * )* * */ function parseBlock() { var label = t.identifier(getUniqueName("block")); var blockResult = null; var instr = []; if (token.type === _tokenizer.tokens.identifier) { label = identifierFromToken(token); eatToken(); } else { label = t.withRaw(label, ""); // preserve anonymous } while (token.type === _tokenizer.tokens.openParen) { eatToken(); if (lookaheadAndCheck(_tokenizer.keywords.result) === true) { eatToken(); blockResult = token.value; eatToken(); } else if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword ) { // Instruction instr.push(parseFuncInstr()); } else { throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in block body of type" + ", given " + tokenToString(token)); }(); } maybeIgnoreComment(); eatTokenOfType(_tokenizer.tokens.closeParen); } return t.blockInstruction(label, instr, blockResult); } /** * Parses a if instruction * * WAST: * * expr: * ( if ? ( then * ) ( else * )? ) * ( if ? + ( then * ) ( else * )? ) * * instr: * if ? * end ? * if ? * else ? * end ? * * block_sig : ( result * )* * */ function parseIf() { var blockResult = null; var label = t.identifier(getUniqueName("if")); var testInstrs = []; var consequent = []; var alternate = []; if (token.type === _tokenizer.tokens.identifier) { label = identifierFromToken(token); eatToken(); } else { label = t.withRaw(label, ""); // preserve anonymous } while (token.type === _tokenizer.tokens.openParen) { eatToken(); // ( /** * Block signature */ if (isKeyword(token, _tokenizer.keywords.result) === true) { eatToken(); blockResult = token.value; eatTokenOfType(_tokenizer.tokens.valtype); eatTokenOfType(_tokenizer.tokens.closeParen); continue; } /** * Then */ if (isKeyword(token, _tokenizer.keywords.then) === true) { eatToken(); // then while (token.type === _tokenizer.tokens.openParen) { eatToken(); // Instruction if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword ) { consequent.push(parseFuncInstr()); } else { throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in consequent body of type" + ", given " + tokenToString(token)); }(); } eatTokenOfType(_tokenizer.tokens.closeParen); } eatTokenOfType(_tokenizer.tokens.closeParen); continue; } /** * Alternate */ if (isKeyword(token, _tokenizer.keywords.else)) { eatToken(); // else while (token.type === _tokenizer.tokens.openParen) { eatToken(); // Instruction if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword ) { alternate.push(parseFuncInstr()); } else { throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in alternate body of type" + ", given " + tokenToString(token)); }(); } eatTokenOfType(_tokenizer.tokens.closeParen); } eatTokenOfType(_tokenizer.tokens.closeParen); continue; } /** * Test instruction */ if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword ) { testInstrs.push(parseFuncInstr()); eatTokenOfType(_tokenizer.tokens.closeParen); continue; } throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in if body" + ", given " + tokenToString(token)); }(); } return t.ifInstruction(label, testInstrs, blockResult, consequent, alternate); } /** * Parses a loop instruction * * WAT: * * blockinstr :: 'loop' I:label rt:resulttype (in:instr*) 'end' id? * * WAST: * * instr :: loop ? * end ? * expr :: ( loop ? * ) * block_sig :: ( result * )* * */ function parseLoop() { var label = t.identifier(getUniqueName("loop")); var blockResult; var instr = []; if (token.type === _tokenizer.tokens.identifier) { label = identifierFromToken(token); eatToken(); } else { label = t.withRaw(label, ""); // preserve anonymous } while (token.type === _tokenizer.tokens.openParen) { eatToken(); if (lookaheadAndCheck(_tokenizer.keywords.result) === true) { eatToken(); blockResult = token.value; eatToken(); } else if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword ) { // Instruction instr.push(parseFuncInstr()); } else { throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in loop body" + ", given " + tokenToString(token)); }(); } eatTokenOfType(_tokenizer.tokens.closeParen); } return t.loopInstruction(label, blockResult, instr); } function parseCallIndirect() { var typeRef; var params = []; var results = []; var instrs = []; while (token.type !== _tokenizer.tokens.closeParen) { if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.type)) { eatToken(); // ( eatToken(); // type typeRef = parseTypeReference(); } else if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.param)) { eatToken(); // ( eatToken(); // param /** * Params can be empty: * (params)` */ if (token.type !== _tokenizer.tokens.closeParen) { params.push.apply(params, _toConsumableArray(parseFuncParam())); } } else if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.result)) { eatToken(); // ( eatToken(); // result /** * Results can be empty: * (result)` */ if (token.type !== _tokenizer.tokens.closeParen) { results.push.apply(results, _toConsumableArray(parseFuncResult())); } } else { eatTokenOfType(_tokenizer.tokens.openParen); instrs.push(parseFuncInstr()); } eatTokenOfType(_tokenizer.tokens.closeParen); } return t.callIndirectInstruction(typeRef !== undefined ? typeRef : t.signature(params, results), instrs); } /** * Parses an export instruction * * WAT: * * export: ( export ) * exkind: ( func ) * ( global ) * ( table ) * ( memory ) * var: | * */ function parseExport() { if (token.type !== _tokenizer.tokens.string) { throw new Error("Expected string after export, got: " + token.type); } var name = token.value; eatToken(); var moduleExportDescr = parseModuleExportDescr(); return t.moduleExport(name, moduleExportDescr); } function parseModuleExportDescr() { var startLoc = getStartLoc(); var type = ""; var index; eatTokenOfType(_tokenizer.tokens.openParen); while (token.type !== _tokenizer.tokens.closeParen) { if (isKeyword(token, _tokenizer.keywords.func)) { type = "Func"; eatToken(); index = parseExportIndex(token); } else if (isKeyword(token, _tokenizer.keywords.table)) { type = "Table"; eatToken(); index = parseExportIndex(token); } else if (isKeyword(token, _tokenizer.keywords.global)) { type = "Global"; eatToken(); index = parseExportIndex(token); } else if (isKeyword(token, _tokenizer.keywords.memory)) { type = "Memory"; eatToken(); index = parseExportIndex(token); } eatToken(); } if (type === "") { throw new Error("Unknown export type"); } if (index === undefined) { throw new Error("Exported function must have a name"); } var node = t.moduleExportDescr(type, index); var endLoc = getEndLoc(); eatTokenOfType(_tokenizer.tokens.closeParen); return t.withLoc(node, endLoc, startLoc); } function parseModule() { var name = null; var isBinary = false; var isQuote = false; var moduleFields = []; if (token.type === _tokenizer.tokens.identifier) { name = token.value; eatToken(); } if (hasPlugin("wast") && token.type === _tokenizer.tokens.name && token.value === "binary") { eatToken(); isBinary = true; } if (hasPlugin("wast") && token.type === _tokenizer.tokens.name && token.value === "quote") { eatToken(); isQuote = true; } if (isBinary === true) { var blob = []; while (token.type === _tokenizer.tokens.string) { blob.push(token.value); eatToken(); maybeIgnoreComment(); } eatTokenOfType(_tokenizer.tokens.closeParen); return t.binaryModule(name, blob); } if (isQuote === true) { var string = []; while (token.type === _tokenizer.tokens.string) { string.push(token.value); eatToken(); } eatTokenOfType(_tokenizer.tokens.closeParen); return t.quoteModule(name, string); } while (token.type !== _tokenizer.tokens.closeParen) { moduleFields.push(walk()); if (state.registredExportedElements.length > 0) { state.registredExportedElements.forEach(function (decl) { moduleFields.push(t.moduleExport(decl.name, t.moduleExportDescr(decl.exportType, decl.id))); }); state.registredExportedElements = []; } token = tokensList[current]; } eatTokenOfType(_tokenizer.tokens.closeParen); return t.module(name, moduleFields); } /** * Parses the arguments of an instruction */ function parseFuncInstrArguments(signature) { var args = []; var namedArgs = {}; var signaturePtr = 0; while (token.type === _tokenizer.tokens.name || isKeyword(token, _tokenizer.keywords.offset)) { var key = token.value; eatToken(); eatTokenOfType(_tokenizer.tokens.equal); var value = void 0; if (token.type === _tokenizer.tokens.number) { value = t.numberLiteralFromRaw(token.value); } else { throw new Error("Unexpected type for argument: " + token.type); } namedArgs[key] = value; eatToken(); } // $FlowIgnore var signatureLength = signature.vector ? Infinity : signature.length; while (token.type !== _tokenizer.tokens.closeParen && ( // $FlowIgnore token.type === _tokenizer.tokens.openParen || signaturePtr < signatureLength)) { if (token.type === _tokenizer.tokens.identifier) { args.push(t.identifier(token.value)); eatToken(); } else if (token.type === _tokenizer.tokens.valtype) { // Handle locals args.push(t.valtypeLiteral(token.value)); eatToken(); } else if (token.type === _tokenizer.tokens.string) { args.push(t.stringLiteral(token.value)); eatToken(); } else if (token.type === _tokenizer.tokens.number) { args.push( // TODO(sven): refactor the type signature handling // https://github.com/xtuc/webassemblyjs/pull/129 is a good start t.numberLiteralFromRaw(token.value, // $FlowIgnore signature[signaturePtr] || "f64")); // $FlowIgnore if (!signature.vector) { ++signaturePtr; } eatToken(); } else if (token.type === _tokenizer.tokens.openParen) { /** * Maybe some nested instructions */ eatToken(); // Instruction if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword ) { // $FlowIgnore args.push(parseFuncInstr()); } else { throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in nested instruction" + ", given " + tokenToString(token)); }(); } if (token.type === _tokenizer.tokens.closeParen) { eatToken(); } } else { throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in instruction argument" + ", given " + tokenToString(token)); }(); } } return { args: args, namedArgs: namedArgs }; } /** * Parses an instruction * * WAT: * * instr :: plaininst * blockinstr * * blockinstr :: 'block' I:label rt:resulttype (in:instr*) 'end' id? * 'loop' I:label rt:resulttype (in:instr*) 'end' id? * 'if' I:label rt:resulttype (in:instr*) 'else' id? (in2:intr*) 'end' id? * * plaininst :: 'unreachable' * 'nop' * 'br' l:labelidx * 'br_if' l:labelidx * 'br_table' l*:vec(labelidx) ln:labelidx * 'return' * 'call' x:funcidx * 'call_indirect' x, I:typeuse * * WAST: * * instr: * * * block ? * end ? * loop ? * end ? * if ? * end ? * if ? * else ? * end ? * * expr: * ( ) * ( + ) * ( block ? * ) * ( loop ? * ) * ( if ? ( then * ) ( else * )? ) * ( if ? + ( then * ) ( else * )? ) * * op: * unreachable * nop * br * br_if * br_table + * return * call * call_indirect * drop * select * get_local * set_local * tee_local * get_global * set_global * .load((8|16|32)_)? ? ? * .store(8|16|32)? ? ? * current_memory * grow_memory * .const * . * . * . * . * ./ * * func_type: ( type )? * * */ function parseFuncInstr() { var startLoc = getStartLoc(); maybeIgnoreComment(); /** * A simple instruction */ if (token.type === _tokenizer.tokens.name || token.type === _tokenizer.tokens.valtype) { var _name2 = token.value; var object; eatToken(); if (token.type === _tokenizer.tokens.dot) { object = _name2; eatToken(); if (token.type !== _tokenizer.tokens.name) { throw new TypeError("Unknown token: " + token.type + ", name expected"); } _name2 = token.value; eatToken(); } if (token.type === _tokenizer.tokens.closeParen) { var _endLoc = token.loc.end; if (typeof object === "undefined") { return t.withLoc(t.instruction(_name2), _endLoc, startLoc); } else { return t.withLoc(t.objectInstruction(_name2, object, []), _endLoc, startLoc); } } var signature = t.signatureForOpcode(object || "", _name2); var _parseFuncInstrArgume = parseFuncInstrArguments(signature), _args = _parseFuncInstrArgume.args, _namedArgs = _parseFuncInstrArgume.namedArgs; var endLoc = token.loc.end; if (typeof object === "undefined") { return t.withLoc(t.instruction(_name2, _args, _namedArgs), endLoc, startLoc); } else { return t.withLoc(t.objectInstruction(_name2, object, _args, _namedArgs), endLoc, startLoc); } } else if (isKeyword(token, _tokenizer.keywords.loop)) { /** * Else a instruction with a keyword (loop or block) */ eatToken(); // keyword return parseLoop(); } else if (isKeyword(token, _tokenizer.keywords.block)) { eatToken(); // keyword return parseBlock(); } else if (isKeyword(token, _tokenizer.keywords.call_indirect)) { eatToken(); // keyword return parseCallIndirect(); } else if (isKeyword(token, _tokenizer.keywords.call)) { eatToken(); // keyword var index; if (token.type === _tokenizer.tokens.identifier) { index = identifierFromToken(token); eatToken(); } else if (token.type === _tokenizer.tokens.number) { index = t.indexLiteral(token.value); eatToken(); } var instrArgs = []; // Nested instruction while (token.type === _tokenizer.tokens.openParen) { eatToken(); instrArgs.push(parseFuncInstr()); eatTokenOfType(_tokenizer.tokens.closeParen); } if (typeof index === "undefined") { throw new Error("Missing argument in call instruciton"); } if (instrArgs.length > 0) { return t.callInstruction(index, instrArgs); } else { return t.callInstruction(index); } } else if (isKeyword(token, _tokenizer.keywords.if)) { eatToken(); // Keyword return parseIf(); } else if (isKeyword(token, _tokenizer.keywords.module) && hasPlugin("wast")) { eatToken(); // In WAST you can have a module as an instruction's argument // we will cast it into a instruction to not break the flow // $FlowIgnore var module = parseModule(); return module; } else { throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected instruction in function body" + ", given " + tokenToString(token)); }(); } } /* * Parses a function * * WAT: * * functype :: ( 'func' t1:vec(param) t2:vec(result) ) * param :: ( 'param' id? t:valtype ) * result :: ( 'result' t:valtype ) * * WAST: * * func :: ( func ? * * ) * ( func ? ( export ) <...> ) * ( func ? ( import ) ) * func_sig :: ( type )? * * * param :: ( param * ) | ( param ) * result :: ( result * ) * local :: ( local * ) | ( local ) * */ function parseFunc() { var fnName = t.identifier(getUniqueName("func")); var typeRef; var fnBody = []; var fnParams = []; var fnResult = []; // name if (token.type === _tokenizer.tokens.identifier) { fnName = identifierFromToken(token); eatToken(); } else { fnName = t.withRaw(fnName, ""); // preserve anonymous } maybeIgnoreComment(); while (token.type === _tokenizer.tokens.openParen || token.type === _tokenizer.tokens.name || token.type === _tokenizer.tokens.valtype) { // Instructions without parens if (token.type === _tokenizer.tokens.name || token.type === _tokenizer.tokens.valtype) { fnBody.push(parseFuncInstr()); continue; } eatToken(); if (lookaheadAndCheck(_tokenizer.keywords.param) === true) { eatToken(); fnParams.push.apply(fnParams, _toConsumableArray(parseFuncParam())); } else if (lookaheadAndCheck(_tokenizer.keywords.result) === true) { eatToken(); fnResult.push.apply(fnResult, _toConsumableArray(parseFuncResult())); } else if (lookaheadAndCheck(_tokenizer.keywords.export) === true) { eatToken(); parseFuncExport(fnName); } else if (lookaheadAndCheck(_tokenizer.keywords.type) === true) { eatToken(); typeRef = parseTypeReference(); } else if (lookaheadAndCheck(_tokenizer.tokens.name) === true || lookaheadAndCheck(_tokenizer.tokens.valtype) === true || token.type === "keyword" // is any keyword ) { // Instruction fnBody.push(parseFuncInstr()); } else { throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in func body" + ", given " + tokenToString(token)); }(); } eatTokenOfType(_tokenizer.tokens.closeParen); } return t.func(fnName, typeRef !== undefined ? typeRef : t.signature(fnParams, fnResult), fnBody); } /** * Parses shorthand export in func * * export :: ( export ) */ function parseFuncExport(funcId) { if (token.type !== _tokenizer.tokens.string) { throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Function export expected a string" + ", given " + tokenToString(token)); }(); } var name = token.value; eatToken(); /** * Func export shorthand, we trait it as a syntaxic sugar. * A export ModuleField will be added later. * * We give the anonymous function a generated name and export it. */ var id = t.identifier(funcId.value); state.registredExportedElements.push({ exportType: "Func", name: name, id: id }); } /** * Parses a type instruction * * WAST: * * typedef: ( type ? ( func * * ) ) */ function parseType() { var id; var params = []; var result = []; if (token.type === _tokenizer.tokens.identifier) { id = identifierFromToken(token); eatToken(); } if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.func)) { eatToken(); // ( eatToken(); // func if (token.type === _tokenizer.tokens.closeParen) { eatToken(); // function with an empty signature, we can abort here return t.typeInstruction(id, t.signature([], [])); } if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.param)) { eatToken(); // ( eatToken(); // param params = parseFuncParam(); eatTokenOfType(_tokenizer.tokens.closeParen); } if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.result)) { eatToken(); // ( eatToken(); // result result = parseFuncResult(); eatTokenOfType(_tokenizer.tokens.closeParen); } eatTokenOfType(_tokenizer.tokens.closeParen); } return t.typeInstruction(id, t.signature(params, result)); } /** * Parses a function result * * WAST: * * result :: ( result * ) */ function parseFuncResult() { var results = []; while (token.type !== _tokenizer.tokens.closeParen) { if (token.type !== _tokenizer.tokens.valtype) { throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unexpected token in func result" + ", given " + tokenToString(token)); }(); } var valtype = token.value; eatToken(); results.push(valtype); } return results; } /** * Parses a type reference * */ function parseTypeReference() { var ref; if (token.type === _tokenizer.tokens.identifier) { ref = identifierFromToken(token); eatToken(); } else if (token.type === _tokenizer.tokens.number) { ref = t.numberLiteralFromRaw(token.value); eatToken(); } return ref; } /** * Parses a global instruction * * WAST: * * global: ( global ? * ) * ( global ? ( export ) <...> ) * ( global ? ( import ) ) * * global_sig: | ( mut ) * */ function parseGlobal() { var name = t.identifier(getUniqueName("global")); var type; // Keep informations in case of a shorthand import var importing = null; maybeIgnoreComment(); if (token.type === _tokenizer.tokens.identifier) { name = identifierFromToken(token); eatToken(); } else { name = t.withRaw(name, ""); // preserve anonymous } /** * maybe export */ if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.export)) { eatToken(); // ( eatToken(); // export var exportName = token.value; eatTokenOfType(_tokenizer.tokens.string); state.registredExportedElements.push({ exportType: "Global", name: exportName, id: name }); eatTokenOfType(_tokenizer.tokens.closeParen); } /** * maybe import */ if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.import)) { eatToken(); // ( eatToken(); // import var moduleName = token.value; eatTokenOfType(_tokenizer.tokens.string); var _name3 = token.value; eatTokenOfType(_tokenizer.tokens.string); importing = { module: moduleName, name: _name3, descr: undefined }; eatTokenOfType(_tokenizer.tokens.closeParen); } /** * global_sig */ if (token.type === _tokenizer.tokens.valtype) { type = t.globalType(token.value, "const"); eatToken(); } else if (token.type === _tokenizer.tokens.openParen) { eatToken(); // ( if (isKeyword(token, _tokenizer.keywords.mut) === false) { throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unsupported global type, expected mut" + ", given " + tokenToString(token)); }(); } eatToken(); // mut type = t.globalType(token.value, "var"); eatToken(); eatTokenOfType(_tokenizer.tokens.closeParen); } if (type === undefined) { throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Could not determine global type" + ", given " + tokenToString(token)); }(); } maybeIgnoreComment(); var init = []; if (importing != null) { importing.descr = type; init.push(t.moduleImport(importing.module, importing.name, importing.descr)); } /** * instr* */ while (token.type === _tokenizer.tokens.openParen) { eatToken(); init.push(parseFuncInstr()); eatTokenOfType(_tokenizer.tokens.closeParen); } return t.global(type, init, name); } /** * Parses a function param * * WAST: * * param :: ( param * ) | ( param ) */ function parseFuncParam() { var params = []; var id; var valtype; if (token.type === _tokenizer.tokens.identifier) { id = token.value; eatToken(); } if (token.type === _tokenizer.tokens.valtype) { valtype = token.value; eatToken(); params.push({ id: id, valtype: valtype }); /** * Shorthand notation for multiple anonymous parameters * @see https://webassembly.github.io/spec/core/text/types.html#function-types * @see https://github.com/xtuc/webassemblyjs/issues/6 */ if (id === undefined) { while (token.type === _tokenizer.tokens.valtype) { valtype = token.value; eatToken(); params.push({ id: undefined, valtype: valtype }); } } } else {// ignore } return params; } /** * Parses an element segments instruction * * WAST: * * elem: ( elem ? (offset * ) * ) * ( elem ? * ) * * var: | */ function parseElem() { var tableIndex = t.indexLiteral(0); var offset = []; var funcs = []; if (token.type === _tokenizer.tokens.identifier) { tableIndex = identifierFromToken(token); eatToken(); } if (token.type === _tokenizer.tokens.number) { tableIndex = t.indexLiteral(token.value); eatToken(); } while (token.type !== _tokenizer.tokens.closeParen) { if (lookaheadAndCheck(_tokenizer.tokens.openParen, _tokenizer.keywords.offset)) { eatToken(); // ( eatToken(); // offset while (token.type !== _tokenizer.tokens.closeParen) { eatTokenOfType(_tokenizer.tokens.openParen); offset.push(parseFuncInstr()); eatTokenOfType(_tokenizer.tokens.closeParen); } eatTokenOfType(_tokenizer.tokens.closeParen); } else if (token.type === _tokenizer.tokens.identifier) { funcs.push(t.identifier(token.value)); eatToken(); } else if (token.type === _tokenizer.tokens.number) { funcs.push(t.indexLiteral(token.value)); eatToken(); } else if (token.type === _tokenizer.tokens.openParen) { eatToken(); // ( offset.push(parseFuncInstr()); eatTokenOfType(_tokenizer.tokens.closeParen); } else { throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unsupported token in elem" + ", given " + tokenToString(token)); }(); } } return t.elem(tableIndex, offset, funcs); } /** * Parses the start instruction in a module * * WAST: * * start: ( start ) * var: | * * WAT: * start ::= ‘(’ ‘start’ x:funcidx ‘)’ */ function parseStart() { if (token.type === _tokenizer.tokens.identifier) { var index = identifierFromToken(token); eatToken(); return t.start(index); } if (token.type === _tokenizer.tokens.number) { var _index2 = t.indexLiteral(token.value); eatToken(); return t.start(_index2); } throw new Error("Unknown start, token: " + tokenToString(token)); } if (token.type === _tokenizer.tokens.openParen) { eatToken(); var startLoc = getStartLoc(); if (isKeyword(token, _tokenizer.keywords.export)) { eatToken(); var node = parseExport(); var _endLoc2 = getEndLoc(); return t.withLoc(node, _endLoc2, startLoc); } if (isKeyword(token, _tokenizer.keywords.loop)) { eatToken(); var _node = parseLoop(); var _endLoc3 = getEndLoc(); return t.withLoc(_node, _endLoc3, startLoc); } if (isKeyword(token, _tokenizer.keywords.func)) { eatToken(); var _node2 = parseFunc(); var _endLoc4 = getEndLoc(); maybeIgnoreComment(); eatTokenOfType(_tokenizer.tokens.closeParen); return t.withLoc(_node2, _endLoc4, startLoc); } if (isKeyword(token, _tokenizer.keywords.module)) { eatToken(); var _node3 = parseModule(); var _endLoc5 = getEndLoc(); return t.withLoc(_node3, _endLoc5, startLoc); } if (isKeyword(token, _tokenizer.keywords.import)) { eatToken(); var _node4 = parseImport(); var _endLoc6 = getEndLoc(); eatTokenOfType(_tokenizer.tokens.closeParen); return t.withLoc(_node4, _endLoc6, startLoc); } if (isKeyword(token, _tokenizer.keywords.block)) { eatToken(); var _node5 = parseBlock(); var _endLoc7 = getEndLoc(); eatTokenOfType(_tokenizer.tokens.closeParen); return t.withLoc(_node5, _endLoc7, startLoc); } if (isKeyword(token, _tokenizer.keywords.memory)) { eatToken(); var _node6 = parseMemory(); var _endLoc8 = getEndLoc(); eatTokenOfType(_tokenizer.tokens.closeParen); return t.withLoc(_node6, _endLoc8, startLoc); } if (isKeyword(token, _tokenizer.keywords.data)) { eatToken(); var _node7 = parseData(); var _endLoc9 = getEndLoc(); eatTokenOfType(_tokenizer.tokens.closeParen); return t.withLoc(_node7, _endLoc9, startLoc); } if (isKeyword(token, _tokenizer.keywords.table)) { eatToken(); var _node8 = parseTable(); var _endLoc10 = getEndLoc(); eatTokenOfType(_tokenizer.tokens.closeParen); return t.withLoc(_node8, _endLoc10, startLoc); } if (isKeyword(token, _tokenizer.keywords.global)) { eatToken(); var _node9 = parseGlobal(); var _endLoc11 = getEndLoc(); eatTokenOfType(_tokenizer.tokens.closeParen); return t.withLoc(_node9, _endLoc11, startLoc); } if (isKeyword(token, _tokenizer.keywords.type)) { eatToken(); var _node10 = parseType(); var _endLoc12 = getEndLoc(); eatTokenOfType(_tokenizer.tokens.closeParen); return t.withLoc(_node10, _endLoc12, startLoc); } if (isKeyword(token, _tokenizer.keywords.start)) { eatToken(); var _node11 = parseStart(); var _endLoc13 = getEndLoc(); eatTokenOfType(_tokenizer.tokens.closeParen); return t.withLoc(_node11, _endLoc13, startLoc); } if (isKeyword(token, _tokenizer.keywords.elem)) { eatToken(); var _node12 = parseElem(); var _endLoc14 = getEndLoc(); eatTokenOfType(_tokenizer.tokens.closeParen); return t.withLoc(_node12, _endLoc14, startLoc); } var instruction = parseFuncInstr(); var endLoc = getEndLoc(); maybeIgnoreComment(); if (_typeof(instruction) === "object") { if (typeof token !== "undefined") { eatTokenOfType(_tokenizer.tokens.closeParen); } return t.withLoc(instruction, endLoc, startLoc); } } if (token.type === _tokenizer.tokens.comment) { var _startLoc = getStartLoc(); var builder = token.opts.type === "leading" ? t.leadingComment : t.blockComment; var _node13 = builder(token.value); eatToken(); // comment var _endLoc15 = getEndLoc(); return t.withLoc(_node13, _endLoc15, _startLoc); } throw function () { return new Error("\n" + (0, _helperCodeFrame.codeFrameFromSource)(source, token.loc) + "\n" + "Unknown token" + ", given " + tokenToString(token)); }(); } var body = []; while (current < tokensList.length) { body.push(walk()); } return t.program(body); }