Files
bun.sh/test/js/node/http/node-fetch.test.js
Jarred Sumner 536dc8653b Fix request body streaming in node-fetch wrapper. (#22458)
### What does this PR do?

Fix request body streaming in node-fetch wrapper.

### How did you verify your code works?

Added a test that previously failed

---------

Co-authored-by: autofix-ci[bot] <114827586+autofix-ci[bot]@users.noreply.github.com>
2025-09-06 22:40:41 -07:00

161 lines
4.6 KiB
JavaScript

import * as vercelFetch from "@vercel/fetch";
import * as iso from "isomorphic-fetch";
import fetch2, { fetch, Headers, Request, Response } from "node-fetch";
import * as stream from "stream";
import { afterEach, expect, test } from "bun:test";
const originalResponse = globalThis.Response;
const originalRequest = globalThis.Request;
const originalHeaders = globalThis.Headers;
afterEach(() => {
globalThis.Response = originalResponse;
globalThis.Request = originalRequest;
globalThis.Headers = originalHeaders;
globalThis.fetch = Bun.fetch;
});
test("node-fetch", () => {
expect(Response.prototype).toBeInstanceOf(globalThis.Response);
expect(Request.prototype).toBeInstanceOf(globalThis.Request);
expect(Headers.prototype).toBeInstanceOf(globalThis.Headers);
expect(fetch2.default).toBe(fetch2);
expect(fetch2.Response).toBe(Response);
});
test("node-fetch Headers.raw()", () => {
const headers = new Headers({ "a": "1" });
headers.append("Set-Cookie", "b=1");
headers.append("Set-Cookie", "c=1");
expect(headers.raw()).toEqual({
"set-cookie": ["b=1", "c=1"],
"a": ["1"],
});
});
for (const [impl, name] of [
[fetch, "node-fetch.fetch"],
[fetch2, "node-fetch.default"],
[fetch2.default, "node-fetch.default.default"],
[iso.fetch, "isomorphic-fetch.fetch"],
[iso.default.fetch, "isomorphic-fetch.default.fetch"],
[iso.default, "isomorphic-fetch.default"],
[vercelFetch.default(fetch), "@vercel/fetch.default"],
]) {
test(name + " fetches", async () => {
using server = Bun.serve({
port: 0,
fetch(req, server) {
server.stop();
return new Response("it works");
},
});
expect(await impl("http://" + server.hostname + ":" + server.port)).toBeInstanceOf(globalThis.Response);
});
}
test("node-fetch uses node streams instead of web streams", async () => {
using server = Bun.serve({
port: 0,
async fetch(req, server) {
const body = await req.text();
expect(body).toBe("the input text");
return new Response("hello world");
},
});
{
const result = await fetch2("http://" + server.hostname + ":" + server.port, {
body: new stream.Readable({
read() {
this.push("the input text");
this.push(null);
},
}),
method: "POST",
});
expect(result.body).toBeInstanceOf(stream.Readable);
expect(result.body === result.body).toBe(true); // cached lazy getter
const headersJSON = result.headers.toJSON();
for (const key of Object.keys(headersJSON)) {
const value = headersJSON[key];
headersJSON[key] = Array.isArray(value) ? value : [value];
}
expect(result.headers.raw()).toEqual(headersJSON);
const chunks = [];
for await (const chunk of result.body) {
chunks.push(chunk);
}
expect(Buffer.concat(chunks).toString()).toBe("hello world");
}
});
test("node-fetch request body streams properly", async () => {
let responseResolve;
const responsePromise = new Promise(resolve => {
responseResolve = resolve;
});
let receivedChunks = [];
let requestBodyComplete = false;
using server = Bun.serve({
port: 0,
async fetch(req, server) {
const reader = req.body.getReader();
// Read first chunk
const { value: firstChunk } = await reader.read();
receivedChunks.push(firstChunk);
// Signal that response can be sent
responseResolve();
// Continue reading remaining chunks
let result;
while (!(result = await reader.read()).done) {
receivedChunks.push(result.value);
}
requestBodyComplete = true;
return new Response("response sent");
},
});
const requestBody = new stream.Readable({
read() {
// Will be controlled manually
},
});
// Start the fetch request
const fetchPromise = fetch2(server.url.href, {
body: requestBody,
method: "POST",
});
// Send first chunk
requestBody.push("first chunk");
// Wait for response to be available (server has read first chunk)
await responsePromise;
// Response is available, but request body should still be streaming
expect(requestBodyComplete).toBe(false);
// Send more data after response is available
requestBody.push("second chunk");
requestBody.push("third chunk");
requestBody.push(null); // End the stream
// Now wait for the fetch to complete
const result = await fetchPromise;
expect(await result.text()).toBe("response sent");
// Verify all chunks were received
const allData = Buffer.concat(receivedChunks).toString();
expect(allData).toBe("first chunksecond chunkthird chunk");
expect(requestBodyComplete).toBe(true);
});