-
Notifications
You must be signed in to change notification settings - Fork 551
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
fix: increased memory in finalization first appearing in v6.16.0 #3445
Conversation
Holding a reference to the stream in the finalization registry causes a memory leak, even when consuming the body.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
@@ -526,7 +527,7 @@ function fromInnerResponse (innerResponse, guard) { | |||
setHeadersGuard(response[kHeaders], guard) | |||
|
|||
if (hasFinalizationRegistry && innerResponse.body?.stream) { | |||
registry.register(response, innerResponse.body.stream) | |||
registry.register(response, new WeakRef(innerResponse.body.stream)) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
registry.register(target, "some value");
If target is reclaimed, your cleanup callback may be called at some point with the held value you provided for it ("some value" in the above). The held value can be any value you like: a primitive or an object, even undefined. If the held value is an object, the registry keeps a strong reference to it (so it can pass it to your cleanup callback later).
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
if no one merges this before you can, adding this as a comment in the code itself would be appreciated
Thanks for sending a fix! Were you able to reproduce without Next.js? I have a been using this quite a bit and this is the first time I saw this problem. I really think we need a test to prevent this from regressing. You can find a test to copy in #3199. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
lgtm
In the future, please don't land PRs without tests or clear reproduction on undici. If we do, let's give those a bit more scrutiny. |
A few notes. I've verified that:
|
@snyamathi how were you able to verify that this fixes Next.js issue? How can I tell Next to use the patched undici instead of the global one? |
* fix: memory leak Holding a reference to the stream in the finalization registry causes a memory leak, even when consuming the body. * docs: add comment explaining the strong reference * typo
@mcollina I built everything from source (node+undici) when testing, but you could also
where foo.js is globalThis.fetch = require('undici').fetch; It's difficult to tell what is different with the RSC setup since both Next and React patch the
It's hard to separate React/RSC from Next and test them independently (see notes here). However, I'm leaning toward this issue being more to do with React's patch vs Next's patch Maybe someone from the React team (cc @sebmarkbage) would be able to chime in? There is some code in React's It's very hard to disable React's patch since next bundles its own copy of React within its own dist package https://unpkg.com/[email protected]/dist/compiled/next-server/app-page.runtime.prod.js I was able to edit the above file locally to disable React's patch (basically bail here) and that also seems to alleviate the memory issue. So I think while the fix here is still prudent since it reduces memory pressure with the only downside being the creating/dereference of the WeakRef, the React patch is likely what's causing the memory retention. |
@snyamathi based on your comment, I found that cloned responses can prevent a Response from being collected: diff --git a/test/fetch/fire-and-forget.js b/test/fetch/fire-and-forget.js
index d8090885..0a5925db 100644
--- a/test/fetch/fire-and-forget.js
+++ b/test/fetch/fire-and-forget.js
@@ -38,7 +38,7 @@ test('does not need the body to be consumed to continue', { timeout: 180_000, sk
gc(true)
const array = new Array(batch)
for (let i = 0; i < batch; i++) {
- array[i] = fetch(url).catch(() => {})
+ array[i] = fetch(url).then(r => r.clone()).catch(() => {})
}
await Promise.all(array)
await sleep(delay) running with
|
That sort of makes sense since cloning the response will create a second copy of the body. If you were to consume (or cancel) the body of the clone, I wonder if the memory usage would not increase in the same way. But then again - the clone should have been garbage collected at some point. And this cloning seems like exactly what React is doing. |
commenting out following line prevents the memory leak. Line 273 in d63afeb
|
Actually, that's still an issue considering I tested on the latest commit to undici. I have no idea what the RSC/next patch is doing to cause the fixed issue.
Sure, but it'll break cloning. var response = await fetch('...')
var clone = response.clone()
assert.deepStrictEqual(await response.text(), await clone.text()) Note that the response we cloned must still have a readable body. We need tee to make the original body accessible (otherwise using a body mixin on a cloned response would lock the original response's body). I'm pretty sure the clone issue would be resolved by handling it in the same way we currently do for 'normal' responses. |
var A = await fetch('...') // -> A is readable
var B = response.clone() // -> A, B are readable
await B.text() // -> A is readable; B is locked (and vice-versa) |
Of course it will break by commenting it out. I was just investigating if it is something trivial. But the .tee() call makes absolutely sense, and if you clone, you are responsible for it. |
More context and a test added in #3450 showing a minimal reproduction (https://github.com/snyamathi/undici-mem-leak) |
Will this be backported to Node v20? |
@pepijn-vanvlaanderen looks like it will: nodejs/node#54274 (comment) |
Holding a reference to the stream in the finalization registry causes a memory leak, even when consuming the body.
This relates to...
https://github.com/orgs/nodejs/discussions/54248
and
vercel/next.js#68636
Rationale
I believe we're maintaining a strong reference to
innerResponse.body.stream
even when the body is consumed.Please see the above issues where many people are seeing a memory leak when Node is updated from 20.15.1 to 20.16.0
I was able to reproduce the issue (vercel/next.js#68636 (comment)) and use that to test in order to find the commit where things started breaking.
I'm not an expert on memory leaks, but this seems to make sense and fixes the issue.
It seems like this was introduced in 63b7794
Changes
Use a WeakRef instead - if the stream is already GC'd then we have nothing to worry about - otherwise we can do the cleanup as usual.
I'm testing this by building Node + Undici locally and then running load testing on https://github.com/snyamathi/68636/tree/simple
When load testing using a build of the prior commit (c0dc3dd) everything is fine. When using the commit 63b7794 the memory leak can be observed.
Adding this fix to the main branch fixes the issue.
The chart above shows the memory usage before (main) and after the fix (my branch)
Features
N/A
Bug Fixes
https://github.com/orgs/nodejs/discussions/54248
Breaking Changes and Deprecations
N/A
Status