%PDF- <> %âãÏÓ endobj 2 0 obj <> endobj 3 0 obj <>/ExtGState<>/ProcSet[/PDF/Text/ImageB/ImageC/ImageI] >>/Annots[ 28 0 R 29 0 R] /MediaBox[ 0 0 595.5 842.25] /Contents 4 0 R/Group<>/Tabs/S>> endobj ºaâÚÎΞ-ÌE1ÍØÄ÷{òò2ÿ ÛÖ^ÔÀá TÎ{¦?§®¥kuµùÕ5sLOšuY>endobj 2 0 obj<>endobj 2 0 obj<>endobj 2 0 obj<>endobj 2 0 obj<> endobj 2 0 obj<>endobj 2 0 obj<>es 3 0 R>> endobj 2 0 obj<> ox[ 0.000000 0.000000 609.600000 935.600000]/Fi endobj 3 0 obj<> endobj 7 1 obj<>/ProcSet[/PDF/Text/ImageB/ImageC/ImageI]>>/Subtype/Form>> stream
var stream = require('stream'); var JsonParse = require('../jsonparse'); var test = require('tape'); test('can handle large tokens without running out of memory', function (t) { var parser = new JsonParse(); var chunkSize = 1024; var chunks = 1024 * 200; // 200mb var quote = Buffer.from ? Buffer.from('"') : new Buffer('"'); t.plan(1); parser.onToken = function (type, value) { t.equal(value.length, chunkSize * chunks, 'token should be size of input json'); t.end(); }; parser.write(quote); for (var i = 0; i < chunks; ++i) { var buf = Buffer.alloc ? Buffer.alloc(chunkSize) : new Buffer(chunkSize); buf.fill('a'); parser.write(buf); } parser.write(quote); });