async_hooks: add missing async_hooks destroys in AsyncReset
This adds missing async_hooks destroy calls for sockets (in _http_agent.js) and HTTP parsers. We need to emit a destroy in AsyncWrap#AsyncReset before assigning a new async_id when the instance has already been in use and is being recycled, because in that case, we have already emitted an init for the "old" async_id. This also removes a duplicated init call for HTTP parser: Each time a new parser was created, AsyncReset was being called via the C++ Parser class constructor (super constructor AsyncWrap) and also via Parser::Reinitialize. PR-URL: https://github.com/nodejs/node/pull/23272 Fixes: https://github.com/nodejs/node/issues/19859 Reviewed-By: Matteo Collina <matteo.collina@gmail.com> Reviewed-By: Anna Henningsen <anna@addaleax.net> Reviewed-By: James M Snell <jasnell@gmail.com>
This commit is contained in:
parent
45c70b0ce7
commit
eb9748d222
@ -25,7 +25,7 @@ function main({ len, n }) {
|
|||||||
bench.start();
|
bench.start();
|
||||||
for (var i = 0; i < n; i++) {
|
for (var i = 0; i < n; i++) {
|
||||||
parser.execute(header, 0, header.length);
|
parser.execute(header, 0, header.length);
|
||||||
parser.reinitialize(REQUEST);
|
parser.reinitialize(REQUEST, i > 0);
|
||||||
}
|
}
|
||||||
bench.end(n);
|
bench.end(n);
|
||||||
}
|
}
|
||||||
|
@ -9,7 +9,7 @@ const bench = common.createBenchmark(main, {
|
|||||||
});
|
});
|
||||||
|
|
||||||
function main({ n }) {
|
function main({ n }) {
|
||||||
const FreeList = require('internal/freelist');
|
const { FreeList } = require('internal/freelist');
|
||||||
const poolSize = 1000;
|
const poolSize = 1000;
|
||||||
const list = new FreeList('test', poolSize, Object);
|
const list = new FreeList('test', poolSize, Object);
|
||||||
var j;
|
var j;
|
||||||
|
@ -167,7 +167,7 @@ Agent.prototype.addRequest = function addRequest(req, options, port/* legacy */,
|
|||||||
var socket = this.freeSockets[name].shift();
|
var socket = this.freeSockets[name].shift();
|
||||||
// Guard against an uninitialized or user supplied Socket.
|
// Guard against an uninitialized or user supplied Socket.
|
||||||
if (socket._handle && typeof socket._handle.asyncReset === 'function') {
|
if (socket._handle && typeof socket._handle.asyncReset === 'function') {
|
||||||
// Assign the handle a new asyncId and run any init() hooks.
|
// Assign the handle a new asyncId and run any destroy()/init() hooks.
|
||||||
socket._handle.asyncReset();
|
socket._handle.asyncReset();
|
||||||
socket[async_id_symbol] = socket._handle.getAsyncId();
|
socket[async_id_symbol] = socket._handle.getAsyncId();
|
||||||
}
|
}
|
||||||
|
@ -47,6 +47,7 @@ const {
|
|||||||
ERR_UNESCAPED_CHARACTERS
|
ERR_UNESCAPED_CHARACTERS
|
||||||
} = require('internal/errors').codes;
|
} = require('internal/errors').codes;
|
||||||
const { validateTimerDuration } = require('internal/timers');
|
const { validateTimerDuration } = require('internal/timers');
|
||||||
|
const is_reused_symbol = require('internal/freelist').symbols.is_reused_symbol;
|
||||||
|
|
||||||
const INVALID_PATH_REGEX = /[^\u0021-\u00ff]/;
|
const INVALID_PATH_REGEX = /[^\u0021-\u00ff]/;
|
||||||
|
|
||||||
@ -631,7 +632,7 @@ function tickOnSocket(req, socket) {
|
|||||||
var parser = parsers.alloc();
|
var parser = parsers.alloc();
|
||||||
req.socket = socket;
|
req.socket = socket;
|
||||||
req.connection = socket;
|
req.connection = socket;
|
||||||
parser.reinitialize(HTTPParser.RESPONSE);
|
parser.reinitialize(HTTPParser.RESPONSE, parser[is_reused_symbol]);
|
||||||
parser.socket = socket;
|
parser.socket = socket;
|
||||||
parser.outgoing = req;
|
parser.outgoing = req;
|
||||||
req.parser = parser;
|
req.parser = parser;
|
||||||
|
@ -23,7 +23,7 @@
|
|||||||
|
|
||||||
const { methods, HTTPParser } = internalBinding('http_parser');
|
const { methods, HTTPParser } = internalBinding('http_parser');
|
||||||
|
|
||||||
const FreeList = require('internal/freelist');
|
const { FreeList } = require('internal/freelist');
|
||||||
const { ondrain } = require('internal/http');
|
const { ondrain } = require('internal/http');
|
||||||
const incoming = require('_http_incoming');
|
const incoming = require('_http_incoming');
|
||||||
const {
|
const {
|
||||||
|
@ -42,6 +42,7 @@ const {
|
|||||||
defaultTriggerAsyncIdScope,
|
defaultTriggerAsyncIdScope,
|
||||||
getOrSetAsyncId
|
getOrSetAsyncId
|
||||||
} = require('internal/async_hooks');
|
} = require('internal/async_hooks');
|
||||||
|
const is_reused_symbol = require('internal/freelist').symbols.is_reused_symbol;
|
||||||
const { IncomingMessage } = require('_http_incoming');
|
const { IncomingMessage } = require('_http_incoming');
|
||||||
const {
|
const {
|
||||||
ERR_HTTP_HEADERS_SENT,
|
ERR_HTTP_HEADERS_SENT,
|
||||||
@ -338,7 +339,7 @@ function connectionListenerInternal(server, socket) {
|
|||||||
socket.on('timeout', socketOnTimeout);
|
socket.on('timeout', socketOnTimeout);
|
||||||
|
|
||||||
var parser = parsers.alloc();
|
var parser = parsers.alloc();
|
||||||
parser.reinitialize(HTTPParser.REQUEST);
|
parser.reinitialize(HTTPParser.REQUEST, parser[is_reused_symbol]);
|
||||||
parser.socket = socket;
|
parser.socket = socket;
|
||||||
socket.parser = parser;
|
socket.parser = parser;
|
||||||
|
|
||||||
|
@ -1,5 +1,7 @@
|
|||||||
'use strict';
|
'use strict';
|
||||||
|
|
||||||
|
const is_reused_symbol = Symbol('isReused');
|
||||||
|
|
||||||
class FreeList {
|
class FreeList {
|
||||||
constructor(name, max, ctor) {
|
constructor(name, max, ctor) {
|
||||||
this.name = name;
|
this.name = name;
|
||||||
@ -9,9 +11,15 @@ class FreeList {
|
|||||||
}
|
}
|
||||||
|
|
||||||
alloc() {
|
alloc() {
|
||||||
return this.list.length ?
|
let item;
|
||||||
this.list.pop() :
|
if (this.list.length > 0) {
|
||||||
this.ctor.apply(this, arguments);
|
item = this.list.pop();
|
||||||
|
item[is_reused_symbol] = true;
|
||||||
|
} else {
|
||||||
|
item = this.ctor.apply(this, arguments);
|
||||||
|
item[is_reused_symbol] = false;
|
||||||
|
}
|
||||||
|
return item;
|
||||||
}
|
}
|
||||||
|
|
||||||
free(obj) {
|
free(obj) {
|
||||||
@ -23,4 +31,9 @@ class FreeList {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
module.exports = FreeList;
|
module.exports = {
|
||||||
|
FreeList,
|
||||||
|
symbols: {
|
||||||
|
is_reused_symbol
|
||||||
|
}
|
||||||
|
};
|
||||||
|
@ -563,6 +563,7 @@ AsyncWrap::AsyncWrap(Environment* env,
|
|||||||
CHECK_NE(provider, PROVIDER_NONE);
|
CHECK_NE(provider, PROVIDER_NONE);
|
||||||
CHECK_GE(object->InternalFieldCount(), 1);
|
CHECK_GE(object->InternalFieldCount(), 1);
|
||||||
|
|
||||||
|
async_id_ = -1;
|
||||||
// Use AsyncReset() call to execute the init() callbacks.
|
// Use AsyncReset() call to execute the init() callbacks.
|
||||||
AsyncReset(execution_async_id, silent);
|
AsyncReset(execution_async_id, silent);
|
||||||
}
|
}
|
||||||
@ -606,6 +607,14 @@ void AsyncWrap::EmitDestroy(Environment* env, double async_id) {
|
|||||||
// and reused over their lifetime. This way a new uid can be assigned when
|
// and reused over their lifetime. This way a new uid can be assigned when
|
||||||
// the resource is pulled out of the pool and put back into use.
|
// the resource is pulled out of the pool and put back into use.
|
||||||
void AsyncWrap::AsyncReset(double execution_async_id, bool silent) {
|
void AsyncWrap::AsyncReset(double execution_async_id, bool silent) {
|
||||||
|
if (async_id_ != -1) {
|
||||||
|
// This instance was in use before, we have already emitted an init with
|
||||||
|
// its previous async_id and need to emit a matching destroy for that
|
||||||
|
// before generating a new async_id.
|
||||||
|
EmitDestroy(env(), async_id_);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Now we can assign a new async_id_ to this instance.
|
||||||
async_id_ =
|
async_id_ =
|
||||||
execution_async_id == -1 ? env()->new_async_id() : execution_async_id;
|
execution_async_id == -1 ? env()->new_async_id() : execution_async_id;
|
||||||
trigger_async_id_ = env()->get_default_trigger_async_id();
|
trigger_async_id_ = env()->get_default_trigger_async_id();
|
||||||
|
@ -465,6 +465,8 @@ class Parser : public AsyncWrap, public StreamListener {
|
|||||||
Environment* env = Environment::GetCurrent(args);
|
Environment* env = Environment::GetCurrent(args);
|
||||||
|
|
||||||
CHECK(args[0]->IsInt32());
|
CHECK(args[0]->IsInt32());
|
||||||
|
CHECK(args[1]->IsBoolean());
|
||||||
|
bool isReused = args[1]->IsTrue();
|
||||||
http_parser_type type =
|
http_parser_type type =
|
||||||
static_cast<http_parser_type>(args[0].As<Int32>()->Value());
|
static_cast<http_parser_type>(args[0].As<Int32>()->Value());
|
||||||
|
|
||||||
@ -473,8 +475,12 @@ class Parser : public AsyncWrap, public StreamListener {
|
|||||||
ASSIGN_OR_RETURN_UNWRAP(&parser, args.Holder());
|
ASSIGN_OR_RETURN_UNWRAP(&parser, args.Holder());
|
||||||
// Should always be called from the same context.
|
// Should always be called from the same context.
|
||||||
CHECK_EQ(env, parser->env());
|
CHECK_EQ(env, parser->env());
|
||||||
// The parser is being reused. Reset the async id and call init() callbacks.
|
// This parser has either just been created or it is being reused.
|
||||||
|
// We must only call AsyncReset for the latter case, because AsyncReset has
|
||||||
|
// already been called via the constructor for the former case.
|
||||||
|
if (isReused) {
|
||||||
parser->AsyncReset();
|
parser->AsyncReset();
|
||||||
|
}
|
||||||
parser->Init(type);
|
parser->Init(type);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -38,20 +38,14 @@ process.on('exit', function() {
|
|||||||
{ type: 'HTTPPARSER',
|
{ type: 'HTTPPARSER',
|
||||||
id: 'httpparser:1',
|
id: 'httpparser:1',
|
||||||
triggerAsyncId: 'tcpserver:1' },
|
triggerAsyncId: 'tcpserver:1' },
|
||||||
{ type: 'HTTPPARSER',
|
|
||||||
id: 'httpparser:2',
|
|
||||||
triggerAsyncId: 'tcpserver:1' },
|
|
||||||
{ type: 'TCPWRAP', id: 'tcp:2', triggerAsyncId: 'tcpserver:1' },
|
{ type: 'TCPWRAP', id: 'tcp:2', triggerAsyncId: 'tcpserver:1' },
|
||||||
{ type: 'Timeout', id: 'timeout:1', triggerAsyncId: 'tcp:2' },
|
{ type: 'Timeout', id: 'timeout:1', triggerAsyncId: 'tcp:2' },
|
||||||
{ type: 'HTTPPARSER',
|
{ type: 'HTTPPARSER',
|
||||||
id: 'httpparser:3',
|
id: 'httpparser:2',
|
||||||
triggerAsyncId: 'tcp:2' },
|
|
||||||
{ type: 'HTTPPARSER',
|
|
||||||
id: 'httpparser:4',
|
|
||||||
triggerAsyncId: 'tcp:2' },
|
triggerAsyncId: 'tcp:2' },
|
||||||
{ type: 'Timeout',
|
{ type: 'Timeout',
|
||||||
id: 'timeout:2',
|
id: 'timeout:2',
|
||||||
triggerAsyncId: 'httpparser:4' },
|
triggerAsyncId: 'httpparser:2' },
|
||||||
{ type: 'SHUTDOWNWRAP',
|
{ type: 'SHUTDOWNWRAP',
|
||||||
id: 'shutdown:1',
|
id: 'shutdown:1',
|
||||||
triggerAsyncId: 'tcp:2' } ]
|
triggerAsyncId: 'tcp:2' } ]
|
||||||
|
84
test/parallel/test-async-hooks-http-agent-destroy.js
Normal file
84
test/parallel/test-async-hooks-http-agent-destroy.js
Normal file
@ -0,0 +1,84 @@
|
|||||||
|
'use strict';
|
||||||
|
// Flags: --expose-internals
|
||||||
|
const common = require('../common');
|
||||||
|
const assert = require('assert');
|
||||||
|
const { async_id_symbol } = require('internal/async_hooks').symbols;
|
||||||
|
const async_hooks = require('async_hooks');
|
||||||
|
const http = require('http');
|
||||||
|
|
||||||
|
// Regression test for https://github.com/nodejs/node/issues/19859
|
||||||
|
// Checks that an http.Agent emits a destroy for the old asyncId before calling
|
||||||
|
// asyncReset()s when reusing a socket handle. The setup is nearly identical to
|
||||||
|
// parallel/test-async-hooks-http-agent (which focuses on the assertion that
|
||||||
|
// a fresh asyncId is assigned to the net.Socket instance).
|
||||||
|
|
||||||
|
const destroyedIds = new Set();
|
||||||
|
async_hooks.createHook({
|
||||||
|
destroy: common.mustCallAtLeast((asyncId) => {
|
||||||
|
destroyedIds.add(asyncId);
|
||||||
|
}, 1)
|
||||||
|
}).enable();
|
||||||
|
|
||||||
|
// Make sure a single socket is transparently reused for 2 requests.
|
||||||
|
const agent = new http.Agent({
|
||||||
|
keepAlive: true,
|
||||||
|
keepAliveMsecs: Infinity,
|
||||||
|
maxSockets: 1
|
||||||
|
});
|
||||||
|
|
||||||
|
const server = http.createServer(common.mustCall((req, res) => {
|
||||||
|
req.once('data', common.mustCallAtLeast(() => {
|
||||||
|
res.writeHead(200, { 'Content-Type': 'text/plain' });
|
||||||
|
res.write('foo');
|
||||||
|
}));
|
||||||
|
req.on('end', common.mustCall(() => {
|
||||||
|
res.end('bar');
|
||||||
|
}));
|
||||||
|
}, 2)).listen(0, common.mustCall(() => {
|
||||||
|
const port = server.address().port;
|
||||||
|
const payload = 'hello world';
|
||||||
|
|
||||||
|
// First request. This is useless except for adding a socket to the
|
||||||
|
// agent’s pool for reuse.
|
||||||
|
const r1 = http.request({
|
||||||
|
agent, port, method: 'POST'
|
||||||
|
}, common.mustCall((res) => {
|
||||||
|
// Remember which socket we used.
|
||||||
|
const socket = res.socket;
|
||||||
|
const asyncIdAtFirstRequest = socket[async_id_symbol];
|
||||||
|
assert.ok(asyncIdAtFirstRequest > 0, `${asyncIdAtFirstRequest} > 0`);
|
||||||
|
// Check that request and response share their socket.
|
||||||
|
assert.strictEqual(r1.socket, socket);
|
||||||
|
|
||||||
|
res.on('data', common.mustCallAtLeast(() => {}));
|
||||||
|
res.on('end', common.mustCall(() => {
|
||||||
|
// setImmediate() to give the agent time to register the freed socket.
|
||||||
|
setImmediate(common.mustCall(() => {
|
||||||
|
// The socket is free for reuse now.
|
||||||
|
assert.strictEqual(socket[async_id_symbol], -1);
|
||||||
|
|
||||||
|
// second request:
|
||||||
|
const r2 = http.request({
|
||||||
|
agent, port, method: 'POST'
|
||||||
|
}, common.mustCall((res) => {
|
||||||
|
assert.ok(destroyedIds.has(asyncIdAtFirstRequest));
|
||||||
|
|
||||||
|
// Empty payload, to hit the “right” code path.
|
||||||
|
r2.end('');
|
||||||
|
|
||||||
|
res.on('data', common.mustCallAtLeast(() => {}));
|
||||||
|
res.on('end', common.mustCall(() => {
|
||||||
|
// Clean up to let the event loop stop.
|
||||||
|
server.close();
|
||||||
|
agent.destroy();
|
||||||
|
}));
|
||||||
|
}));
|
||||||
|
|
||||||
|
// Schedule a payload to be written immediately, but do not end the
|
||||||
|
// request just yet.
|
||||||
|
r2.write(payload);
|
||||||
|
}));
|
||||||
|
}));
|
||||||
|
}));
|
||||||
|
r1.end(payload);
|
||||||
|
}));
|
61
test/parallel/test-async-hooks-http-parser-destroy.js
Normal file
61
test/parallel/test-async-hooks-http-parser-destroy.js
Normal file
@ -0,0 +1,61 @@
|
|||||||
|
'use strict';
|
||||||
|
const common = require('../common');
|
||||||
|
const Countdown = require('../common/countdown');
|
||||||
|
const assert = require('assert');
|
||||||
|
const async_hooks = require('async_hooks');
|
||||||
|
const http = require('http');
|
||||||
|
|
||||||
|
// Regression test for https://github.com/nodejs/node/issues/19859.
|
||||||
|
// Checks that matching destroys are emitted when creating new/reusing old http
|
||||||
|
// parser instances.
|
||||||
|
|
||||||
|
const N = 50;
|
||||||
|
const KEEP_ALIVE = 100;
|
||||||
|
|
||||||
|
const createdIds = [];
|
||||||
|
const destroyedIds = [];
|
||||||
|
async_hooks.createHook({
|
||||||
|
init: common.mustCallAtLeast((asyncId, type) => {
|
||||||
|
if (type === 'HTTPPARSER') {
|
||||||
|
createdIds.push(asyncId);
|
||||||
|
}
|
||||||
|
}, N),
|
||||||
|
destroy: (asyncId) => {
|
||||||
|
destroyedIds.push(asyncId);
|
||||||
|
}
|
||||||
|
}).enable();
|
||||||
|
|
||||||
|
const server = http.createServer(function(req, res) {
|
||||||
|
res.end('Hello');
|
||||||
|
});
|
||||||
|
|
||||||
|
const keepAliveAgent = new http.Agent({
|
||||||
|
keepAlive: true,
|
||||||
|
keepAliveMsecs: KEEP_ALIVE,
|
||||||
|
});
|
||||||
|
|
||||||
|
const countdown = new Countdown(N, () => {
|
||||||
|
server.close(() => {
|
||||||
|
// give the server sockets time to close (which will also free their
|
||||||
|
// associated parser objects) after the server has been closed.
|
||||||
|
setTimeout(() => {
|
||||||
|
createdIds.forEach((createdAsyncId) => {
|
||||||
|
assert.ok(destroyedIds.indexOf(createdAsyncId) >= 0);
|
||||||
|
});
|
||||||
|
}, KEEP_ALIVE * 2);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
server.listen(0, function() {
|
||||||
|
for (let i = 0; i < N; ++i) {
|
||||||
|
(function makeRequest() {
|
||||||
|
http.get({
|
||||||
|
port: server.address().port,
|
||||||
|
agent: keepAliveAgent
|
||||||
|
}, function(res) {
|
||||||
|
countdown.dec();
|
||||||
|
res.resume();
|
||||||
|
});
|
||||||
|
})();
|
||||||
|
}
|
||||||
|
});
|
@ -4,28 +4,27 @@
|
|||||||
|
|
||||||
require('../common');
|
require('../common');
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const FreeList = require('internal/freelist');
|
const { FreeList } = require('internal/freelist');
|
||||||
|
|
||||||
assert.strictEqual(typeof FreeList, 'function');
|
assert.strictEqual(typeof FreeList, 'function');
|
||||||
|
|
||||||
const flist1 = new FreeList('flist1', 3, String);
|
const flist1 = new FreeList('flist1', 3, Object);
|
||||||
|
|
||||||
// Allocating when empty, should not change the list size
|
// Allocating when empty, should not change the list size
|
||||||
const result = flist1.alloc('test');
|
const result = flist1.alloc();
|
||||||
assert.strictEqual(typeof result, 'string');
|
assert.strictEqual(typeof result, 'object');
|
||||||
assert.strictEqual(result, 'test');
|
|
||||||
assert.strictEqual(flist1.list.length, 0);
|
assert.strictEqual(flist1.list.length, 0);
|
||||||
|
|
||||||
// Exhaust the free list
|
// Exhaust the free list
|
||||||
assert(flist1.free('test1'));
|
assert(flist1.free({ id: 'test1' }));
|
||||||
assert(flist1.free('test2'));
|
assert(flist1.free({ id: 'test2' }));
|
||||||
assert(flist1.free('test3'));
|
assert(flist1.free({ id: 'test3' }));
|
||||||
|
|
||||||
// Now it should not return 'true', as max length is exceeded
|
// Now it should not return 'true', as max length is exceeded
|
||||||
assert.strictEqual(flist1.free('test4'), false);
|
assert.strictEqual(flist1.free({ id: 'test4' }), false);
|
||||||
assert.strictEqual(flist1.free('test5'), false);
|
assert.strictEqual(flist1.free({ id: 'test5' }), false);
|
||||||
|
|
||||||
// At this point 'alloc' should just return the stored values
|
// At this point 'alloc' should just return the stored values
|
||||||
assert.strictEqual(flist1.alloc(), 'test3');
|
assert.strictEqual(flist1.alloc().id, 'test3');
|
||||||
assert.strictEqual(flist1.alloc(), 'test2');
|
assert.strictEqual(flist1.alloc().id, 'test2');
|
||||||
assert.strictEqual(flist1.alloc(), 'test1');
|
assert.strictEqual(flist1.alloc().id, 'test1');
|
||||||
|
@ -98,7 +98,7 @@ function expectBody(expected) {
|
|||||||
throw new Error('hello world');
|
throw new Error('hello world');
|
||||||
};
|
};
|
||||||
|
|
||||||
parser.reinitialize(HTTPParser.REQUEST);
|
parser.reinitialize(HTTPParser.REQUEST, false);
|
||||||
|
|
||||||
assert.throws(
|
assert.throws(
|
||||||
() => { parser.execute(request, 0, request.length); },
|
() => { parser.execute(request, 0, request.length); },
|
||||||
@ -558,7 +558,7 @@ function expectBody(expected) {
|
|||||||
parser[kOnBody] = expectBody('ping');
|
parser[kOnBody] = expectBody('ping');
|
||||||
parser.execute(req1, 0, req1.length);
|
parser.execute(req1, 0, req1.length);
|
||||||
|
|
||||||
parser.reinitialize(REQUEST);
|
parser.reinitialize(REQUEST, false);
|
||||||
parser[kOnBody] = expectBody('pong');
|
parser[kOnBody] = expectBody('pong');
|
||||||
parser[kOnHeadersComplete] = onHeadersComplete2;
|
parser[kOnHeadersComplete] = onHeadersComplete2;
|
||||||
parser.execute(req2, 0, req2.length);
|
parser.execute(req2, 0, req2.length);
|
||||||
|
@ -7,5 +7,5 @@ const config = process.binding('config');
|
|||||||
|
|
||||||
console.log(config, process.argv);
|
console.log(config, process.argv);
|
||||||
|
|
||||||
assert.strictEqual(typeof require('internal/freelist'), 'function');
|
assert.strictEqual(typeof require('internal/freelist').FreeList, 'function');
|
||||||
assert.strictEqual(config.exposeInternals, true);
|
assert.strictEqual(config.exposeInternals, true);
|
||||||
|
@ -7,6 +7,7 @@ const common = require('../common');
|
|||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const httpCommon = require('_http_common');
|
const httpCommon = require('_http_common');
|
||||||
const { internalBinding } = require('internal/test/binding');
|
const { internalBinding } = require('internal/test/binding');
|
||||||
|
const is_reused_symbol = require('internal/freelist').symbols.is_reused_symbol;
|
||||||
const { HTTPParser } = internalBinding('http_parser');
|
const { HTTPParser } = internalBinding('http_parser');
|
||||||
const net = require('net');
|
const net = require('net');
|
||||||
|
|
||||||
@ -25,7 +26,7 @@ function execAndClose() {
|
|||||||
process.stdout.write('.');
|
process.stdout.write('.');
|
||||||
|
|
||||||
const parser = parsers.pop();
|
const parser = parsers.pop();
|
||||||
parser.reinitialize(HTTPParser.RESPONSE);
|
parser.reinitialize(HTTPParser.RESPONSE, parser[is_reused_symbol]);
|
||||||
|
|
||||||
const socket = net.connect(common.PORT);
|
const socket = net.connect(common.PORT);
|
||||||
socket.on('error', (e) => {
|
socket.on('error', (e) => {
|
||||||
|
Loading…
x
Reference in New Issue
Block a user