Skip to content

Commit

Permalink
Move entire codebase to ESM (inikulin#351)
Browse files Browse the repository at this point in the history
Co-authored-by: Felix Böhm <[email protected]>
  • Loading branch information
2 people authored and jmbpwtw committed Feb 16, 2023
1 parent b59fed9 commit 0f0b235
Show file tree
Hide file tree
Showing 83 changed files with 4,057 additions and 2,598 deletions.
20 changes: 0 additions & 20 deletions .eslintrc.js

This file was deleted.

21 changes: 21 additions & 0 deletions .eslintrc.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,21 @@
{
"env": {
"es2020": true,
"node": true,
"mocha": true
},
"extends": ["eslint:recommended", "prettier"],
"plugins": ["prettier"],
"rules": {
"prettier/prettier": "error",
"no-console": "error",
"curly": ["error", "all"],
"prefer-arrow-callback": "error",
"one-var": ["error", "never"],
"no-var": "error",
"prefer-const": "error"
},
"parserOptions": {
"sourceType": "module"
}
}
9 changes: 0 additions & 9 deletions bench/.eslintrc.js

This file was deleted.

6 changes: 6 additions & 0 deletions bench/.eslintrc.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,6 @@
{
"extends": ["../.eslintrc.json"],
"rules": {
"no-console": "off"
}
}
8 changes: 3 additions & 5 deletions bench/memory/named-entity-data.js
Original file line number Diff line number Diff line change
@@ -1,13 +1,11 @@
'use strict';

const format = require('human-format');
import format from 'human-format';

main();

function main() {
async function main() {
const before = process.memoryUsage().rss;

require('../../packages/parse5/lib/tokenizer/named-entity-data');
await import('../../packages/parse5/lib/tokenizer/named-entity-data.js');

const after = process.memoryUsage().rss;

Expand Down
20 changes: 9 additions & 11 deletions bench/memory/sax-parser.js
Original file line number Diff line number Diff line change
@@ -1,10 +1,8 @@
'use strict';

const fs = require('fs');
const format = require('human-format');
const promisifyEvent = require('promisify-event');
const memwatch = require('node-memwatch');
const SAXParser = require('../../packages/parse5-sax-parser/lib');
import * as fs from 'fs';
import format from 'human-format';
import promisifyEvent from 'promisify-event';
import memwatch from '@airbnb/node-memwatch';
import SAXParser from '../../packages/parse5-sax-parser/lib/index.js';

main();

Expand All @@ -16,21 +14,21 @@ async function main() {
const heapDiffMeasurement = new memwatch.HeapDiff();
let heapDiff = null;

memwatch.on('stats', stats => {
memwatch.on('stats', (stats) => {
maxMemUsage = Math.max(maxMemUsage, stats['current_base']);
});

startDate = new Date();

const parserPromise = parse().then(dataSize => {
const parserPromise = parse().then((dataSize) => {
parsedDataSize = dataSize;
endDate = new Date();
heapDiff = heapDiffMeasurement.end();
});

await Promise.all([
parserPromise,
promisifyEvent(memwatch, 'stats') // NOTE: we need at least one `stats` result
promisifyEvent(memwatch, 'stats'), // NOTE: we need at least one `stats` result
]);

printResults(parsedDataSize, startDate, endDate, heapDiff, maxMemUsage);
Expand All @@ -57,7 +55,7 @@ function getDuration(startDate, endDate) {
const scale = new format.Scale({
seconds: 1,
minutes: 60,
hours: 3600
hours: 3600,
});

return format((endDate - startDate) / 1000, { scale: scale });
Expand Down
3 changes: 2 additions & 1 deletion bench/package.json
Original file line number Diff line number Diff line change
@@ -1,13 +1,14 @@
{
"name": "parse5-benchmarks",
"type": "module",
"version": "1.0.0",
"description": "parse5 regression benchmarks",
"author": "Ivan Nikulin <[email protected]>",
"license": "MIT",
"dependencies": {
"benchmark": "^2.1.4",
"human-format": "^0.7.0",
"node-memwatch": "^1.0.1",
"@airbnb/node-memwatch": "^2.0.0",
"parse5": "*",
"promisify-event": "^1.0.0"
}
Expand Down
75 changes: 38 additions & 37 deletions bench/perf/index.js
Original file line number Diff line number Diff line change
@@ -1,37 +1,38 @@
'use strict';

const { join } = require('path');
const { readFileSync, createReadStream, readdirSync } = require('fs');
const Benchmark = require('benchmark');
const { loadTreeConstructionTestData } = require('../../test/utils/generate-parsing-tests');
const loadSAXParserTestData = require('../../test/utils/load-sax-parser-test-data');
const { treeAdapters, WritableStreamStub } = require('../../test/utils/common');
import { readFileSync, createReadStream, readdirSync } from 'fs';
import Benchmark from 'benchmark';
import { loadTreeConstructionTestData } from '../../test/utils/generate-parsing-tests.js';
import { loadSAXParserTestData } from '../../test/utils/load-sax-parser-test-data.js';
import { treeAdapters, WritableStreamStub } from '../../test/utils/common.js';
import * as parse5 from '../../packages/parse5/lib/index.js';
import { ParserStream as parse5Stream } from '../../packages/parse5-parser-stream/lib/index.js';
import * as parse5Upstream from 'parse5';

const hugePagePath = new URL('../../test/data/huge-page/huge-page.html', import.meta.url);
const treeConstructionPath = new URL('../../test/data/html5lib-tests/tree-construction', import.meta.url);
const saxPath = new URL('../../test/data/sax/', import.meta.url);

//HACK: https://github.com/bestiejs/benchmark.js/issues/51
/* global workingCopy, WorkingCopyParserStream, upstreamParser, hugePage, microTests, runMicro, runPages, files */
global.workingCopy = require('../../packages/parse5/lib');
global.WorkingCopyParserStream = require('../../packages/parse5-parser-stream/lib');
global.upstreamParser = require('parse5');
global.workingCopy = parse5;
global.WorkingCopyParserStream = parse5Stream;
global.upstreamParser = parse5Upstream;

// Huge page data
global.hugePage = readFileSync(join(__dirname, '../../test/data/huge-page/huge-page.html')).toString();
global.hugePage = readFileSync(hugePagePath).toString();

// Micro data
global.microTests = loadTreeConstructionTestData(
[join(__dirname, '../../test/data/html5lib-tests/tree-construction')],
treeAdapters.default
)
global.microTests = loadTreeConstructionTestData([treeConstructionPath], treeAdapters.default)
.filter(
test =>
(test) =>
//NOTE: this test caused stack overflow in parse5 v1.x
test.input !== '<button><p><button>'
)
.map(test => ({
.map((test) => ({
html: test.input,
fragmentContext: test.fragmentContext
fragmentContext: test.fragmentContext,
}));

global.runMicro = function(parser) {
global.runMicro = function (parser) {
for (const test of microTests) {
if (test.fragmentContext) {
parser.parseFragment(test.fragmentContext, test.html);
Expand All @@ -42,22 +43,22 @@ global.runMicro = function(parser) {
};

// Pages data
const pages = loadSAXParserTestData().map(test => test.src);
const pages = loadSAXParserTestData().map((test) => test.src);

global.runPages = function(parser) {
global.runPages = function (parser) {
for (let j = 0; j < pages.length; j++) {
parser.parse(pages[j]);
}
};

// Stream data
global.files = readdirSync(join(__dirname, '../../test/data/sax')).map(dirName =>
join(__dirname, '../../test/data/sax', dirName, 'src.html')
);
global.files = readdirSync(saxPath).map((dirName) => {
return new URL(`./${dirName}/src.html`, saxPath).pathname;
});

// Utils
function getHz(suite, testName) {
return suite.filter(t => t.name === testName)[0].hz;
return suite.filter((t) => t.name === testName)[0].hz;
}

function runBench({ name, workingCopyFn, upstreamFn, defer = false }) {
Expand All @@ -67,7 +68,7 @@ function runBench({ name, workingCopyFn, upstreamFn, defer = false }) {
.add('Working copy', workingCopyFn, { defer })
.add('Upstream', upstreamFn, { defer })
.on('start', () => console.log(name))
.on('cycle', event => console.log(String(event.target)))
.on('cycle', (event) => console.log(String(event.target)))
.on('complete', () => {
const workingCopyHz = getHz(suite, 'Working copy');
const upstreamHz = getHz(suite, 'Upstream');
Expand All @@ -85,28 +86,28 @@ function runBench({ name, workingCopyFn, upstreamFn, defer = false }) {
runBench({
name: 'parse5 regression benchmark - MICRO',
workingCopyFn: () => runMicro(workingCopy),
upstreamFn: () => runMicro(upstreamParser)
upstreamFn: () => runMicro(upstreamParser),
});

runBench({
name: 'parse5 regression benchmark - HUGE',
workingCopyFn: () => workingCopy.parse(hugePage),
upstreamFn: () => upstreamParser.parse(hugePage)
upstreamFn: () => upstreamParser.parse(hugePage),
});

runBench({
name: 'parse5 regression benchmark - PAGES',
workingCopyFn: () => runPages(workingCopy),
upstreamFn: () => runPages(upstreamParser)
upstreamFn: () => runPages(upstreamParser),
});

runBench({
name: 'parse5 regression benchmark - STREAM',
defer: true,
workingCopyFn: async deferred => {
workingCopyFn: async (deferred) => {
const parsePromises = files.map(
fileName =>
new Promise(resolve => {
(fileName) =>
new Promise((resolve) => {
const stream = createReadStream(fileName, 'utf8');
const parserStream = new WorkingCopyParserStream();

Expand All @@ -118,10 +119,10 @@ runBench({
await Promise.all(parsePromises);
deferred.resolve();
},
upstreamFn: async deferred => {
upstreamFn: async (deferred) => {
const parsePromises = files.map(
fileName =>
new Promise(resolve => {
(fileName) =>
new Promise((resolve) => {
const stream = createReadStream(fileName, 'utf8');
const writable = new WritableStreamStub();

Expand All @@ -136,5 +137,5 @@ runBench({

await Promise.all(parsePromises);
deferred.resolve();
}
},
});
Loading

0 comments on commit 0f0b235

Please sign in to comment.