Compare commits

..

1 Commits

Author SHA1 Message Date
Claude Bot
37bdffed03 Add HTTP thread-based Bun version update checker
Implements a version checker that runs during HTTP thread idle periods:

- Checks for Bun version updates at random intervals (30-180 seconds)
- Only runs when HTTP thread is idle (no active requests)
- Ensures only one check per idle period using state tracking
- Integrates with existing usockets sweep timer (4-second intervals)
- Uses existing HTTP infrastructure for version checking
- Includes basic test coverage

The checker activates during HTTP thread idle periods, making efficient
use of the existing 4-second sweep timer cycle. This provides automatic
update notifications without impacting performance during active usage.

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-07-29 07:35:07 +00:00
849 changed files with 93883 additions and 125608 deletions

78
.agent/agent.mjs Normal file
View File

@@ -0,0 +1,78 @@
import { spawnSync } from "node:child_process";
import { readFileSync, existsSync } from "node:fs";
import { parseArgs } from "node:util";
const { positionals, values } = parseArgs({
allowPositionals: true,
options: {
help: {
type: "boolean",
short: "h",
default: false,
},
interactive: {
type: "boolean",
short: "i",
default: false,
},
},
});
if (values.help || positionals.length === 0) {
console.log("Usage: node agent.mjs <prompt_name> [extra_args...]");
console.log("Example: node agent.mjs triage fix bug in authentication");
console.log("Options:");
console.log(" -h, --help Show this help message");
console.log(" -i, --interactive Run in interactive mode");
process.exit(0);
}
const promptName = positionals[0].toUpperCase();
const promptFile = `.agent/${promptName}.md`;
const extraArgs = positionals.slice(1);
if (!existsSync(promptFile)) {
console.error(`Error: Prompt file "${promptFile}" not found`);
console.error(`Available prompts should be named like: .agent/triage.md, .agent/debug.md, etc.`);
process.exit(1);
}
try {
let prompt = readFileSync(promptFile, "utf-8");
const githubEnvs = Object.entries(process.env)
.filter(([key]) => key.startsWith("GITHUB_"))
.sort(([a], [b]) => a.localeCompare(b));
if (githubEnvs.length > 0) {
const githubContext = `## GitHub Environment\n\n${githubEnvs
.map(([key, value]) => `**${key}**: \`${value}\``)
.join("\n")}\n\n---\n\n`;
prompt = githubContext + prompt;
}
if (extraArgs.length > 0) {
const extraArgsContext = `\n\n## Additional Arguments\n\n${extraArgs.join(" ")}\n\n---\n\n`;
prompt = prompt + extraArgsContext;
}
const claudeArgs = [prompt, "--allowedTools=Edit,Write,Replace,Search", "--output-format=json"];
if (!values.interactive) {
claudeArgs.unshift("--print");
}
const { status, error } = spawnSync("claude", claudeArgs, {
stdio: "inherit",
encoding: "utf-8",
});
if (error) {
console.error("Error running claude:", error);
process.exit(1);
}
process.exit(status || 0);
} catch (error) {
console.error(`Error reading prompt file "${promptFile}":`, error);
process.exit(1);
}

View File

@@ -303,34 +303,9 @@ function getCppAgent(platform, options) {
}
return getEc2Agent(platform, options, {
instanceType: arch === "aarch64" ? "c8g.4xlarge" : "c7i.4xlarge",
});
}
/**
* @param {Platform} platform
* @param {PipelineOptions} options
* @returns {string}
*/
function getLinkBunAgent(platform, options) {
const { os, arch, distro } = platform;
if (os === "darwin") {
return {
queue: `build-${os}`,
os,
arch,
};
}
if (os === "windows") {
return getEc2Agent(platform, options, {
instanceType: arch === "aarch64" ? "r8g.large" : "r7i.large",
});
}
return getEc2Agent(platform, options, {
instanceType: arch === "aarch64" ? "r8g.xlarge" : "r7i.xlarge",
instanceType: arch === "aarch64" ? "c8g.16xlarge" : "c7i.16xlarge",
cpuCount: 32,
threadsPerCore: 1,
});
}
@@ -381,7 +356,7 @@ function getTestAgent(platform, options) {
};
}
// TODO: delete this block when we upgrade to mimalloc v3
// TODO: `dev-server-ssr-110.test.ts` and `next-build.test.ts` run out of memory at 8GB of memory, so use 16GB instead.
if (os === "windows") {
return getEc2Agent(platform, options, {
instanceType: "c7i.2xlarge",
@@ -527,7 +502,7 @@ function getLinkBunStep(platform, options) {
key: `${getTargetKey(platform)}-build-bun`,
label: `${getTargetLabel(platform)} - build-bun`,
depends_on: [`${getTargetKey(platform)}-build-cpp`, `${getTargetKey(platform)}-build-zig`],
agents: getLinkBunAgent(platform, options),
agents: getCppAgent(platform, options),
retry: getRetry(),
cancel_on_build_failing: isMergeQueue(),
env: {

View File

@@ -1,3 +1,50 @@
### What does this PR do?
<!-- **Please explain what your changes do**, example: -->
<!--
This adds a new flag --bail to bun test. When set, it will stop running tests after the first failure. This is useful for CI environments where you want to fail fast.
-->
- [ ] Documentation or TypeScript types (it's okay to leave the rest blank in this case)
- [ ] Code changes
### How did you verify your code works?
<!-- **For code changes, please include automated tests**. Feel free to uncomment the line below -->
<!-- I wrote automated tests -->
<!-- If JavaScript/TypeScript modules or builtins changed:
- [ ] I included a test for the new code, or existing tests cover it
- [ ] I ran my tests locally and they pass (`bun-debug test test-file-name.test`)
-->
<!-- If Zig files changed:
- [ ] I checked the lifetime of memory allocated to verify it's (1) freed and (2) only freed when it should be
- [ ] I included a test for the new code, or an existing test covers it
- [ ] JSValue used outside of the stack is either wrapped in a JSC.Strong or is JSValueProtect'ed
- [ ] I wrote TypeScript/JavaScript tests and they pass locally (`bun-debug test test-file-name.test`)
-->
<!-- If new methods, getters, or setters were added to a publicly exposed class:
- [ ] I added TypeScript types for the new methods, getters, or setters
-->
<!-- If dependencies in tests changed:
- [ ] I made sure that specific versions of dependencies are used instead of ranged or tagged versions
-->
<!-- If a new builtin ESM/CJS module was added:
- [ ] I updated Aliases in `module_loader.zig` to include the new module
- [ ] I added a test that imports the module
- [ ] I added a test that require() the module
-->

View File

@@ -1,103 +0,0 @@
# GitHub Actions Workflow Maintenance Guide
This document provides guidance for maintaining the GitHub Actions workflows in this repository.
## format.yml Workflow
### Overview
The `format.yml` workflow runs code formatters (Prettier, clang-format, and Zig fmt) on pull requests and pushes to main. It's optimized for speed by running all formatters in parallel.
### Key Components
#### 1. Clang-format Script (`scripts/run-clang-format.sh`)
- **Purpose**: Formats C++ source and header files
- **What it does**:
- Reads C++ files from `cmake/sources/CxxSources.txt`
- Finds all header files in `src/` and `packages/`
- Excludes third-party directories (libuv, napi, deps, vendor, sqlite, etc.)
- Requires specific clang-format version (no fallbacks)
**Important exclusions**:
- `src/napi/` - Node API headers (third-party)
- `src/bun.js/bindings/libuv/` - libuv headers (third-party)
- `src/bun.js/bindings/sqlite/` - SQLite headers (third-party)
- `src/bun.js/api/ffi-*.h` - FFI headers (generated/third-party)
- `src/deps/` - Dependencies (third-party)
- Files in `vendor/`, `third_party/`, `generated/` directories
#### 2. Parallel Execution
The workflow runs all three formatters simultaneously:
- Each formatter outputs with a prefix (`[prettier]`, `[clang-format]`, `[zig]`)
- Output is streamed in real-time without blocking
- Uses GitHub Actions groups (`::group::`) for collapsible sections
#### 3. Tool Installation
##### Clang-format-19
- Installs ONLY `clang-format-19` package (not the entire LLVM toolchain)
- Uses `--no-install-recommends --no-install-suggests` to skip unnecessary packages
- Quiet installation with `-qq` and `-o=Dpkg::Use-Pty=0`
##### Zig
- Downloads from `oven-sh/zig` releases (musl build for static linking)
- URL: `https://github.com/oven-sh/zig/releases/download/autobuild-{COMMIT}/bootstrap-x86_64-linux-musl.zip`
- Extracts to temp directory to avoid polluting the repository
- Directory structure: `bootstrap-x86_64-linux-musl/zig`
### Updating the Workflow
#### To update Zig version:
1. Find the new commit hash from https://github.com/oven-sh/zig/releases
2. Replace the hash in the wget URL (line 65 of format.yml)
3. Test that the URL is valid and the binary works
#### To update clang-format version:
1. Update `LLVM_VERSION_MAJOR` environment variable at the top of format.yml
2. Update the version check in `scripts/run-clang-format.sh`
#### To add/remove file exclusions:
1. Edit the exclusion patterns in `scripts/run-clang-format.sh` (lines 34-39)
2. Test locally to ensure the right files are being formatted
### Performance Optimizations
1. **Parallel execution**: All formatters run simultaneously
2. **Minimal installations**: Only required packages, no extras
3. **Temp directories**: Tools downloaded to temp dirs, cleaned up after use
4. **Streaming output**: Real-time feedback without buffering
5. **Early start**: Formatting begins immediately after each tool is ready
### Troubleshooting
**If formatters appear to run sequentially:**
- Check if output is being buffered (should use `sed` for line prefixing)
- Ensure background processes use `&` and proper wait commands
**If third-party files are being formatted:**
- Review exclusion patterns in `scripts/run-clang-format.sh`
- Check if new third-party directories were added that need exclusion
**If clang-format installation is slow:**
- Ensure using minimal package installation flags
- Check if apt cache needs updating
- Consider caching the clang-format binary between runs
### Testing Changes Locally
```bash
# Test the clang-format script
export LLVM_VERSION_MAJOR=19
./scripts/run-clang-format.sh format
# Test with check mode (no modifications)
./scripts/run-clang-format.sh check
# Test specific file exclusions
./scripts/run-clang-format.sh format 2>&1 | grep -E "(libuv|napi|deps)"
# Should return nothing if exclusions work correctly
```
### Important Notes
- The script defaults to **format** mode (modifies files)
- Always test locally before pushing workflow changes
- The musl Zig build works on glibc systems due to static linking
- Keep the exclusion list updated as new third-party code is added

View File

@@ -1,24 +0,0 @@
name: Auto-label Claude PRs
on:
pull_request:
types: [opened]
jobs:
auto-label:
if: github.event.pull_request.user.login == 'robobun'
runs-on: ubuntu-latest
permissions:
contents: read
pull-requests: write
steps:
- name: Add claude label to PRs from robobun
uses: actions/github-script@v7
with:
script: |
github.rest.issues.addLabels({
owner: context.repo.owner,
repo: context.repo.repo,
issue_number: context.issue.number,
labels: ['claude']
});

View File

@@ -6,8 +6,6 @@ on:
- "docs/**"
- "packages/bun-types/**.d.ts"
- "CONTRIBUTING.md"
- "src/cli/install.sh"
- "src/cli/install.ps1"
branches:
- main

View File

@@ -37,72 +37,24 @@ jobs:
- name: Setup Dependencies
run: |
bun install
- name: Format Code
- name: Install LLVM
run: |
# Start prettier in background with prefixed output
echo "::group::Prettier"
(bun run prettier 2>&1 | sed 's/^/[prettier] /' || echo "[prettier] Failed with exit code $?") &
PRETTIER_PID=$!
# Start clang-format installation and formatting in background with prefixed output
echo "::group::Clang-format"
(
echo "[clang-format] Installing clang-format-${{ env.LLVM_VERSION_MAJOR }}..."
wget -qO- https://apt.llvm.org/llvm-snapshot.gpg.key | sudo tee /etc/apt/trusted.gpg.d/apt.llvm.org.asc > /dev/null
echo "deb http://apt.llvm.org/$(lsb_release -cs)/ llvm-toolchain-$(lsb_release -cs)-${{ env.LLVM_VERSION_MAJOR }} main" | sudo tee /etc/apt/sources.list.d/llvm.list > /dev/null
sudo apt-get update -qq
sudo apt-get install -y -qq --no-install-recommends --no-install-suggests -o=Dpkg::Use-Pty=0 clang-format-${{ env.LLVM_VERSION_MAJOR }}
echo "[clang-format] Running clang-format..."
LLVM_VERSION_MAJOR=${{ env.LLVM_VERSION_MAJOR }} ./scripts/run-clang-format.sh format 2>&1 | sed 's/^/[clang-format] /'
) &
CLANG_PID=$!
# Setup Zig in temp directory and run zig fmt in background with prefixed output
echo "::group::Zig fmt"
(
ZIG_TEMP=$(mktemp -d)
echo "[zig] Downloading Zig (musl build)..."
wget -q -O "$ZIG_TEMP/zig.zip" https://github.com/oven-sh/zig/releases/download/autobuild-d1a4e0b0ddc75f37c6a090b97eef0cbb6335556e/bootstrap-x86_64-linux-musl.zip
unzip -q -d "$ZIG_TEMP" "$ZIG_TEMP/zig.zip"
export PATH="$ZIG_TEMP/bootstrap-x86_64-linux-musl:$PATH"
echo "[zig] Running zig fmt..."
zig fmt src 2>&1 | sed 's/^/[zig] /'
./scripts/sort-imports.ts src 2>&1 | sed 's/^/[zig] /'
zig fmt src 2>&1 | sed 's/^/[zig] /'
rm -rf "$ZIG_TEMP"
) &
ZIG_PID=$!
# Wait for all formatting tasks to complete
echo ""
echo "Running formatters in parallel..."
FAILED=0
if ! wait $PRETTIER_PID; then
echo "::error::Prettier failed"
FAILED=1
fi
echo "::endgroup::"
if ! wait $CLANG_PID; then
echo "::error::Clang-format failed"
FAILED=1
fi
echo "::endgroup::"
if ! wait $ZIG_PID; then
echo "::error::Zig fmt failed"
FAILED=1
fi
echo "::endgroup::"
# Exit with error if any formatter failed
if [ $FAILED -eq 1 ]; then
echo "::error::One or more formatters failed"
exit 1
fi
echo "✅ All formatters completed successfully"
curl -fsSL https://apt.llvm.org/llvm.sh | sudo bash -s -- ${{ env.LLVM_VERSION_MAJOR }} all
- name: Setup Zig
uses: mlugg/setup-zig@v1
with:
version: 0.14.0
- name: Zig Format
run: |
zig fmt src
./scripts/sort-imports.ts src
zig fmt src
- name: Prettier Format
run: |
bun run prettier
- name: Clang Format
run: |
bun run clang-format
- name: Ban Words
run: |
bun ./test/internal/ban-words.test.ts

View File

@@ -55,13 +55,10 @@ jobs:
echo "Error: Could not fetch SHA for tag $LATEST_TAG"
exit 1
fi
# Try to get commit SHA from tag object (for annotated tags)
# If it fails, assume it's a lightweight tag pointing directly to commit
LATEST_SHA=$(curl -sL "https://api.github.com/repos/HdrHistogram/HdrHistogram_c/git/tags/$LATEST_TAG_SHA" 2>/dev/null | jq -r '.object.sha // empty')
if [ -z "$LATEST_SHA" ]; then
# Lightweight tag - SHA points directly to commit
LATEST_SHA="$LATEST_TAG_SHA"
LATEST_SHA=$(curl -sL "https://api.github.com/repos/HdrHistogram/HdrHistogram_c/git/tags/$LATEST_TAG_SHA" | jq -r '.object.sha')
if [ -z "$LATEST_SHA" ] || [ "$LATEST_SHA" = "null" ]; then
echo "Error: Could not fetch SHA for tag $LATEST_TAG @ $LATEST_TAG_SHA"
exit 1
fi
if ! [[ $LATEST_SHA =~ ^[0-9a-f]{40}$ ]]; then
@@ -91,7 +88,7 @@ jobs:
commit-message: "deps: update hdrhistogram to ${{ steps.check-version.outputs.tag }} (${{ steps.check-version.outputs.latest }})"
title: "deps: update hdrhistogram to ${{ steps.check-version.outputs.tag }}"
delete-branch: true
branch: deps/update-hdrhistogram-${{ github.run_number }}
branch: deps/update-cares-${{ github.run_number }}
body: |
## What does this PR do?
@@ -99,4 +96,4 @@ jobs:
Compare: https://github.com/HdrHistogram/HdrHistogram_c/compare/${{ steps.check-version.outputs.current }}...${{ steps.check-version.outputs.latest }}
Auto-updated by [this workflow](https://github.com/oven-sh/bun/actions/workflows/update-hdrhistogram.yml)
Auto-updated by [this workflow](https://github.com/oven-sh/bun/actions/workflows/update-cares.yml)

View File

@@ -50,33 +50,14 @@ jobs:
exit 1
fi
TAG_REF=$(curl -sL "https://api.github.com/repos/google/highway/git/refs/tags/$LATEST_TAG")
if [ -z "$TAG_REF" ]; then
echo "Error: Could not fetch tag reference for $LATEST_TAG"
exit 1
fi
TAG_OBJECT_SHA=$(echo "$TAG_REF" | jq -r '.object.sha')
TAG_OBJECT_TYPE=$(echo "$TAG_REF" | jq -r '.object.type')
if [ -z "$TAG_OBJECT_SHA" ] || [ "$TAG_OBJECT_SHA" = "null" ]; then
LATEST_TAG_SHA=$(curl -sL "https://api.github.com/repos/google/highway/git/refs/tags/$LATEST_TAG" | jq -r '.object.sha')
if [ -z "$LATEST_TAG_SHA" ] || [ "$LATEST_TAG_SHA" = "null" ]; then
echo "Error: Could not fetch SHA for tag $LATEST_TAG"
exit 1
fi
# Handle both lightweight tags (type: commit) and annotated tags (type: tag)
if [ "$TAG_OBJECT_TYPE" = "commit" ]; then
# Lightweight tag - object.sha is already the commit SHA
LATEST_SHA="$TAG_OBJECT_SHA"
elif [ "$TAG_OBJECT_TYPE" = "tag" ]; then
# Annotated tag - need to fetch the tag object to get the commit SHA
LATEST_SHA=$(curl -sL "https://api.github.com/repos/google/highway/git/tags/$TAG_OBJECT_SHA" | jq -r '.object.sha')
if [ -z "$LATEST_SHA" ] || [ "$LATEST_SHA" = "null" ]; then
echo "Error: Could not fetch commit SHA for annotated tag $LATEST_TAG @ $TAG_OBJECT_SHA"
exit 1
fi
else
echo "Error: Unexpected tag object type: $TAG_OBJECT_TYPE"
LATEST_SHA=$(curl -sL "https://api.github.com/repos/google/highway/git/tags/$LATEST_TAG_SHA" | jq -r '.object.sha')
if [ -z "$LATEST_SHA" ] || [ "$LATEST_SHA" = "null" ]; then
echo "Error: Could not fetch SHA for tag $LATEST_TAG @ $LATEST_TAG_SHA"
exit 1
fi
@@ -107,7 +88,7 @@ jobs:
commit-message: "deps: update highway to ${{ steps.check-version.outputs.tag }} (${{ steps.check-version.outputs.latest }})"
title: "deps: update highway to ${{ steps.check-version.outputs.tag }}"
delete-branch: true
branch: deps/update-highway-${{ github.run_number }}
branch: deps/update-cares-${{ github.run_number }}
body: |
## What does this PR do?
@@ -115,4 +96,4 @@ jobs:
Compare: https://github.com/google/highway/compare/${{ steps.check-version.outputs.current }}...${{ steps.check-version.outputs.latest }}
Auto-updated by [this workflow](https://github.com/oven-sh/bun/actions/workflows/update-highway.yml)
Auto-updated by [this workflow](https://github.com/oven-sh/bun/actions/workflows/update-cares.yml)

View File

@@ -50,27 +50,15 @@ jobs:
exit 1
fi
# Get the commit SHA that the tag points to
# This handles both lightweight tags (direct commit refs) and annotated tags (tag objects)
TAG_REF_RESPONSE=$(curl -sL "https://api.github.com/repos/cloudflare/lol-html/git/refs/tags/$LATEST_TAG")
LATEST_TAG_SHA=$(echo "$TAG_REF_RESPONSE" | jq -r '.object.sha')
TAG_OBJECT_TYPE=$(echo "$TAG_REF_RESPONSE" | jq -r '.object.type')
LATEST_TAG_SHA=$(curl -sL "https://api.github.com/repos/cloudflare/lol-html/git/refs/tags/$LATEST_TAG" | jq -r '.object.sha')
if [ -z "$LATEST_TAG_SHA" ] || [ "$LATEST_TAG_SHA" = "null" ]; then
echo "Error: Could not fetch SHA for tag $LATEST_TAG"
exit 1
fi
if [ "$TAG_OBJECT_TYPE" = "tag" ]; then
# This is an annotated tag, we need to get the commit it points to
LATEST_SHA=$(curl -sL "https://api.github.com/repos/cloudflare/lol-html/git/tags/$LATEST_TAG_SHA" | jq -r '.object.sha')
if [ -z "$LATEST_SHA" ] || [ "$LATEST_SHA" = "null" ]; then
echo "Error: Could not fetch commit SHA for annotated tag $LATEST_TAG @ $LATEST_TAG_SHA"
exit 1
fi
else
# This is a lightweight tag pointing directly to a commit
LATEST_SHA="$LATEST_TAG_SHA"
LATEST_SHA=$(curl -sL "https://api.github.com/repos/cloudflare/lol-html/git/tags/$LATEST_TAG_SHA" | jq -r '.object.sha')
if [ -z "$LATEST_SHA" ] || [ "$LATEST_SHA" = "null" ]; then
echo "Error: Could not fetch SHA for tag $LATEST_TAG @ $LATEST_TAG_SHA"
exit 1
fi
if ! [[ $LATEST_SHA =~ ^[0-9a-f]{40}$ ]]; then

View File

@@ -50,32 +50,15 @@ jobs:
exit 1
fi
# Get the tag reference, which contains both SHA and type
TAG_REF=$(curl -sL "https://api.github.com/repos/litespeedtech/ls-hpack/git/refs/tags/$LATEST_TAG")
if [ -z "$TAG_REF" ]; then
echo "Error: Could not fetch tag reference for $LATEST_TAG"
exit 1
fi
LATEST_TAG_SHA=$(echo "$TAG_REF" | jq -r '.object.sha')
TAG_TYPE=$(echo "$TAG_REF" | jq -r '.object.type')
LATEST_TAG_SHA=$(curl -sL "https://api.github.com/repos/litespeedtech/ls-hpack/git/refs/tags/$LATEST_TAG" | jq -r '.object.sha')
if [ -z "$LATEST_TAG_SHA" ] || [ "$LATEST_TAG_SHA" = "null" ]; then
echo "Error: Could not fetch SHA for tag $LATEST_TAG"
exit 1
fi
# If it's an annotated tag, we need to dereference it to get the commit SHA
# If it's a lightweight tag, the SHA already points to the commit
if [ "$TAG_TYPE" = "tag" ]; then
LATEST_SHA=$(curl -sL "https://api.github.com/repos/litespeedtech/ls-hpack/git/tags/$LATEST_TAG_SHA" | jq -r '.object.sha')
if [ -z "$LATEST_SHA" ] || [ "$LATEST_SHA" = "null" ]; then
echo "Error: Could not fetch commit SHA for annotated tag $LATEST_TAG"
exit 1
fi
else
# For lightweight tags, the SHA is already the commit SHA
LATEST_SHA="$LATEST_TAG_SHA"
LATEST_SHA=$(curl -sL "https://api.github.com/repos/litespeedtech/ls-hpack/git/tags/$LATEST_TAG_SHA" | jq -r '.object.sha')
if [ -z "$LATEST_SHA" ] || [ "$LATEST_SHA" = "null" ]; then
echo "Error: Could not fetch SHA for tag $LATEST_TAG @ $LATEST_TAG_SHA"
exit 1
fi
if ! [[ $LATEST_SHA =~ ^[0-9a-f]{40}$ ]]; then

4
.gitignore vendored
View File

@@ -184,6 +184,4 @@ codegen-for-zig-team.tar.gz
*.sock
scratch*.{js,ts,tsx,cjs,mjs}
*.bun-build
scripts/lldb-inline
*.bun-build

View File

@@ -168,5 +168,5 @@
"WebKit/WebInspectorUI": true,
},
"git.detectSubmodules": false,
"bun.test.customScript": "./build/debug/bun-debug test"
// "bun.test.customScript": "./build/debug/bun-debug test"
}

View File

@@ -6,7 +6,7 @@ This is the Bun repository - an all-in-one JavaScript runtime & toolkit designed
- **Build debug version**: `bun bd`
- Creates a debug build at `./build/debug/bun-debug`
- **CRITICAL**: DO NOT set a build timeout. Compilation takes ~5 minutes. Be patient.
- Compilation takes ~5 minutes. Don't timeout, be patient.
- **Run tests with your debug build**: `bun bd test <test-file>`
- **CRITICAL**: Never use `bun test` directly - it won't include your changes
- **Run any command with debug build**: `bun bd <command>`
@@ -59,8 +59,8 @@ test("my feature", async () => {
});
const [stdout, stderr, exitCode] = await Promise.all([
proc.stdout.text(),
proc.stderr.text(),
new Response(proc.stdout).text(),
new Response(proc.stderr).text(),
proc.exited,
]);
@@ -69,8 +69,6 @@ test("my feature", async () => {
});
```
- Always use `port: 0`. Do not hardcode ports. Do not use your own random port number function.
## Code Architecture
### Language Structure
@@ -135,6 +133,7 @@ test("my feature", async () => {
When implementing JavaScript classes in C++:
1. Create three classes if there's a public constructor:
- `class Foo : public JSC::JSDestructibleObject` (if has C++ fields)
- `class FooPrototype : public JSC::JSNonFinalObject`
- `class FooConstructor : public JSC::InternalFunction`
@@ -194,6 +193,7 @@ Built-in JavaScript modules use special syntax and are organized as:
```
3. **Debug helpers**:
- `$debug()` - Like console.log but stripped in release builds
- `$assert()` - Assertions stripped in release builds
- `if($debug) {}` - Check if debug env var is set
@@ -221,17 +221,15 @@ bun ci
## Important Development Notes
1. **Never use `bun test` or `bun <file>` directly** - always use `bun bd test` or `bun bd <command>`. `bun bd` compiles & runs the debug build.
2. **All changes must be tested** - if you're not testing your changes, you're not done.
3. **Get your tests to pass**. If you didn't run the tests, your code does not work.
4. **Follow existing code style** - check neighboring files for patterns
5. **Create tests in the right folder** in `test/` and the test must end in `.test.ts` or `.test.tsx`
6. **Use absolute paths** - Always use absolute paths in file operations
7. **Avoid shell commands** - Don't use `find` or `grep` in tests; use Bun's Glob and built-in tools
8. **Memory management** - In Zig code, be careful with allocators and use defer for cleanup
9. **Cross-platform** - Run `bun run zig:check-all` to compile the Zig code on all platforms when making platform-specific changes
10. **Debug builds** - Use `BUN_DEBUG_QUIET_LOGS=1` to disable debug logging, or `BUN_DEBUG_<scope>=1` to enable specific scopes
11. **Be humble & honest** - NEVER overstate what you got done or what actually works in commits, PRs or in messages to the user.
12. **Branch names must start with `claude/`** - This is a requirement for the CI to work.
2. **Use `await using`** for proper resource cleanup with Bun APIs (Bun.spawn, Bun.serve, Bun.connect, etc.)
3. **Follow existing code style** - check neighboring files for patterns
4. **Create regression tests** in `test/regression/issue/` when fixing bugs
5. **Use absolute paths** - Always use absolute paths in file operations
6. **Avoid shell commands** - Don't use `find` or `grep` in tests; use Bun's Glob and built-in tools
7. **Memory management** - In Zig code, be careful with allocators and use defer for cleanup
8. **Cross-platform** - Test on macOS, Linux, and Windows when making platform-specific changes
9. **Debug builds** - Use `BUN_DEBUG_QUIET_LOGS=1` to disable debug logging, or `BUN_DEBUG_<scope>=1` to enable specific scopes
10. **Transpiled source** - Find transpiled files in `/tmp/bun-debug-src/` for debugging
## Key APIs and Features

2
LATEST
View File

@@ -1 +1 @@
1.2.20
1.2.19

2046
Makefile Normal file

File diff suppressed because it is too large Load Diff

View File

@@ -28,4 +28,10 @@ bench("brotli compress stream", async () => {
await pipeline(source, compress);
});
bench("brotli decompress stream", async () => {
const source = Readable.from([compressed]);
const decompress = createBrotliDecompress();
await pipeline(source, decompress);
});
await run();

View File

@@ -1,194 +0,0 @@
// Final benchmark: processStorage + Atomics vs postMessage
const NUM_WORKERS = 5;
const NUM_MESSAGES = 1000;
const MESSAGE_SIZE = 1024;
const measureMemory = () => {
if (typeof Bun !== 'undefined' && Bun.gc) {
Bun.gc(true);
}
return process.memoryUsage();
};
// Benchmark processStorage + Atomics
const benchmarkProcessStorage = async () => {
console.log("\n=== ProcessStorage + Atomics Benchmark ===");
const storage = Bun.experimental_processStorage;
storage.clear();
const sharedBuffer = new SharedArrayBuffer(8 * NUM_WORKERS);
const counters = new Int32Array(sharedBuffer);
const startMemory = measureMemory();
// Pre-populate storage
const testData = 'x'.repeat(MESSAGE_SIZE);
for (let i = 0; i < NUM_MESSAGES; i++) {
storage.setItem(`msg_${i}`, testData);
}
const startTime = performance.now();
// Create workers
const workers = [];
const promises = [];
for (let w = 0; w < NUM_WORKERS; w++) {
const worker = new Worker("./benchmark-worker-ps.js");
workers.push(worker);
const promise = new Promise(resolve => {
worker.onmessage = (e) => {
if (e.data.type === 'done') {
resolve(e.data);
}
};
});
promises.push(promise);
worker.postMessage({
type: 'start',
workerId: w,
numMessages: NUM_MESSAGES,
sharedBuffer
});
}
const results = await Promise.all(promises);
const endTime = performance.now();
const endMemory = measureMemory();
workers.forEach(w => w.terminate());
storage.clear();
const totalTime = endTime - startTime;
const totalProcessed = results.reduce((sum, r) => sum + r.processed, 0);
return {
method: 'processStorage + Atomics',
totalTime,
totalProcessed,
throughput: totalProcessed / (totalTime / 1000),
memoryDelta: endMemory.rss - startMemory.rss,
results
};
};
// Benchmark postMessage
const benchmarkPostMessage = async () => {
console.log("\n=== PostMessage Benchmark ===");
const startMemory = measureMemory();
const testData = 'x'.repeat(MESSAGE_SIZE);
const startTime = performance.now();
// Create workers
const workers = [];
const promises = [];
for (let w = 0; w < NUM_WORKERS; w++) {
const worker = new Worker("./benchmark-worker-pm.js");
workers.push(worker);
const promise = new Promise(resolve => {
worker.onmessage = (e) => {
if (e.data.type === 'done') {
resolve(e.data);
}
};
});
promises.push(promise);
// Start worker
worker.postMessage({
type: 'start',
numMessages: NUM_MESSAGES
});
// Send messages
for (let i = 0; i < NUM_MESSAGES; i++) {
worker.postMessage({
type: 'message',
workerId: w,
data: testData
});
}
}
const results = await Promise.all(promises);
const endTime = performance.now();
const endMemory = measureMemory();
workers.forEach(w => w.terminate());
const totalTime = endTime - startTime;
const totalProcessed = results.reduce((sum, r) => sum + r.processed, 0);
return {
method: 'postMessage',
totalTime,
totalProcessed,
throughput: totalProcessed / (totalTime / 1000),
memoryDelta: endMemory.rss - startMemory.rss,
results
};
};
// Run benchmark
const runBenchmark = async () => {
console.log(`🚀 ProcessStorage vs PostMessage Benchmark`);
console.log(`Configuration:`);
console.log(`- Workers: ${NUM_WORKERS}`);
console.log(`- Messages per worker: ${NUM_MESSAGES}`);
console.log(`- Message size: ${MESSAGE_SIZE} bytes`);
console.log(`- Total messages: ${NUM_MESSAGES * NUM_WORKERS}`);
try {
const psResult = await benchmarkProcessStorage();
await new Promise(r => setTimeout(r, 2000)); // Cool down
const pmResult = await benchmarkPostMessage();
console.log(`\n📊 Results Summary:`);
console.log(`┌─────────────────────────────┬──────────────────┬──────────────────┐`);
console.log(`│ Method │ processStorage │ postMessage │`);
console.log(`├─────────────────────────────┼──────────────────┼──────────────────┤`);
console.log(`│ Total Time (ms) │ ${psResult.totalTime.toFixed(2).padStart(16)}${pmResult.totalTime.toFixed(2).padStart(16)}`);
console.log(`│ Throughput (msgs/sec) │ ${psResult.throughput.toFixed(0).padStart(16)}${pmResult.throughput.toFixed(0).padStart(16)}`);
console.log(`│ Memory Delta (KB) │ ${(psResult.memoryDelta/1024).toFixed(1).padStart(16)}${(pmResult.memoryDelta/1024).toFixed(1).padStart(16)}`);
console.log(`│ Messages Processed │ ${psResult.totalProcessed.toString().padStart(16)}${pmResult.totalProcessed.toString().padStart(16)}`);
console.log(`└─────────────────────────────┴──────────────────┴──────────────────┘`);
const speedupRatio = pmResult.totalTime / psResult.totalTime;
const memoryRatio = psResult.memoryDelta / pmResult.memoryDelta;
console.log(`\n🎯 Performance Analysis:`);
if (speedupRatio > 1.1) {
console.log(`✅ processStorage is ${speedupRatio.toFixed(2)}x faster than postMessage`);
} else if (speedupRatio < 0.9) {
console.log(`❌ processStorage is ${(1/speedupRatio).toFixed(2)}x slower than postMessage`);
} else {
console.log(`🟡 Similar performance (${speedupRatio.toFixed(2)}x)`);
}
if (memoryRatio > 1.1) {
console.log(`✅ processStorage uses ${memoryRatio.toFixed(2)}x less memory than postMessage`);
} else if (memoryRatio < 0.9) {
console.log(`❌ processStorage uses ${(1/memoryRatio).toFixed(2)}x more memory than postMessage`);
} else {
console.log(`🟡 Similar memory usage (${memoryRatio.toFixed(2)}x)`);
}
console.log(`\n💡 Use Cases Where processStorage Excels:`);
console.log(`- Shared configuration/state across workers`);
console.log(`- Caching expensive computations`);
console.log(`- Real-time coordination with Atomics`);
console.log(`- Zero-copy string sharing between threads`);
} catch (error) {
console.error(`❌ Benchmark failed:`, error);
}
};
runBenchmark();

View File

@@ -1,24 +0,0 @@
// PostMessage worker
let processed = 0;
let numMessages = 0;
let startTime = 0;
onmessage = (e) => {
if (e.data.type === 'start') {
numMessages = e.data.numMessages;
startTime = performance.now();
processed = 0;
} else if (e.data.type === 'message') {
processed++;
if (processed >= numMessages) {
const endTime = performance.now();
postMessage({
type: 'done',
workerId: e.data.workerId || 0,
processed,
duration: endTime - startTime
});
}
}
};

View File

@@ -1,30 +0,0 @@
// ProcessStorage + Atomics worker
const storage = Bun.experimental_processStorage;
onmessage = (e) => {
if (e.data.type === 'start') {
const { workerId, numMessages, sharedBuffer } = e.data;
const counters = new Int32Array(sharedBuffer);
let processed = 0;
const startTime = performance.now();
// Read messages from storage
for (let i = 0; i < numMessages; i++) {
const data = storage.getItem(`msg_${i}`);
if (data) {
processed++;
Atomics.add(counters, workerId, 1);
}
}
const endTime = performance.now();
postMessage({
type: 'done',
workerId,
processed,
duration: endTime - startTime
});
}
};

View File

@@ -15,7 +15,7 @@
"react": "^18.3.1",
"react-dom": "^18.3.1",
"source-map-js": "^1.2.0",
"typescript": "5.9.2",
"typescript": "^5.7.2",
},
},
"packages/@types/bun": {
@@ -32,6 +32,7 @@
},
"devDependencies": {
"@types/react": "^19",
"typescript": "^5.0.2",
},
"peerDependencies": {
"@types/react": "^19",
@@ -307,7 +308,7 @@
"tslib": ["tslib@2.8.1", "", {}, "sha512-oJFu94HQb+KVduSUQL7wnpmqnfmLsOA/nAh6b6EH0wCEoK0/mPeXU6c3wKDV83MkOuHPRHtSXKKU99IBazS/2w=="],
"typescript": ["typescript@5.9.2", "", { "bin": { "tsc": "bin/tsc", "tsserver": "bin/tsserver" } }, "sha512-CWBzXQrc/qOkhidw1OzBTQuYRbfyxDXJMVJ1XNwUHGROVmuaeiEm3OslpZ1RV96d7SKKjZKrSJu3+t/xlw3R9A=="],
"typescript": ["typescript@5.8.3", "", { "bin": { "tsc": "bin/tsc", "tsserver": "bin/tsserver" } }, "sha512-p1diW6TqL9L07nNxvRMM7hMMw4c5XOo/1ibL4aAIGmSAt9slTE1Xgw5KWuof2uTOvCg9BY7ZRi+GaF+7sfgPeQ=="],
"uglify-js": ["uglify-js@3.19.3", "", { "bin": { "uglifyjs": "bin/uglifyjs" } }, "sha512-v3Xu+yuwBXisp6QYTcH4UbH+xYJXqnq2m/LtQVWKWzYc1iehYnLixoQDN9FH6/j9/oybfd6W9Ghwkl8+UMKTKQ=="],
@@ -333,8 +334,6 @@
"@octokit/webhooks/@octokit/webhooks-methods": ["@octokit/webhooks-methods@4.1.0", "", {}, "sha512-zoQyKw8h9STNPqtm28UGOYFE7O6D4Il8VJwhAtMHFt2C4L0VQT1qGKLeefUOqHNs1mNRYSadVv7x0z8U2yyeWQ=="],
"bun-tracestrings/typescript": ["typescript@5.8.3", "", { "bin": { "tsc": "bin/tsc", "tsserver": "bin/tsserver" } }, "sha512-p1diW6TqL9L07nNxvRMM7hMMw4c5XOo/1ibL4aAIGmSAt9slTE1Xgw5KWuof2uTOvCg9BY7ZRi+GaF+7sfgPeQ=="],
"camel-case/no-case": ["no-case@2.3.2", "", { "dependencies": { "lower-case": "^1.1.1" } }, "sha512-rmTZ9kz+f3rCvK2TD1Ue/oZlns7OGoIWP4fc3llxxRXlOkHKoWPPWJOfFYpITabSow43QJbRIoHQXtt10VldyQ=="],
"change-case/camel-case": ["camel-case@4.1.2", "", { "dependencies": { "pascal-case": "^3.1.2", "tslib": "^2.0.3" } }, "sha512-gxGWBrTT1JuMx6R+o5PTXMmUnhnVzLQ9SNutD4YqKtI6ap897t3tKECYla6gCWEkplXnlNybEkZg9GEGxKFCgw=="],

View File

@@ -102,11 +102,6 @@ else()
endif()
optionx(ENABLE_ASAN BOOL "If ASAN support should be enabled" DEFAULT ${DEFAULT_ASAN})
optionx(ENABLE_ZIG_ASAN BOOL "If Zig ASAN support should be enabled" DEFAULT ${ENABLE_ASAN})
if (NOT ENABLE_ASAN)
set(ENABLE_ZIG_ASAN OFF)
endif()
if(RELEASE AND LINUX AND CI AND NOT ENABLE_ASSERTIONS AND NOT ENABLE_ASAN)
set(DEFAULT_LTO ON)

View File

@@ -23,7 +23,6 @@ src/bun.js/bindings/BunJSCEventLoop.cpp
src/bun.js/bindings/BunObject.cpp
src/bun.js/bindings/BunPlugin.cpp
src/bun.js/bindings/BunProcess.cpp
src/bun.js/bindings/BunProcessStorage.cpp
src/bun.js/bindings/BunString.cpp
src/bun.js/bindings/BunWorkerGlobalScope.cpp
src/bun.js/bindings/c-bindings.cpp
@@ -351,7 +350,6 @@ src/bun.js/bindings/webcore/JSTextEncoderStream.cpp
src/bun.js/bindings/webcore/JSTransformStream.cpp
src/bun.js/bindings/webcore/JSTransformStreamDefaultController.cpp
src/bun.js/bindings/webcore/JSURLSearchParams.cpp
src/bun.js/bindings/webcore/JSWasmStreamingCompiler.cpp
src/bun.js/bindings/webcore/JSWebSocket.cpp
src/bun.js/bindings/webcore/JSWorker.cpp
src/bun.js/bindings/webcore/JSWorkerOptions.cpp

View File

@@ -29,7 +29,6 @@ src/js/builtins/TransformStream.ts
src/js/builtins/TransformStreamDefaultController.ts
src/js/builtins/TransformStreamInternals.ts
src/js/builtins/UtilInspect.ts
src/js/builtins/WasmStreaming.ts
src/js/builtins/WritableStreamDefaultController.ts
src/js/builtins/WritableStreamDefaultWriter.ts
src/js/builtins/WritableStreamInternals.ts

View File

@@ -1,8 +1,6 @@
src/allocators.zig
src/allocators/AllocationScope.zig
src/allocators/basic.zig
src/allocators/fallback.zig
src/allocators/fallback/z.zig
src/allocators/LinuxMemFdAllocator.zig
src/allocators/MaxHeapAllocator.zig
src/allocators/MemoryReportingAllocator.zig
@@ -12,7 +10,6 @@ src/allocators/NullableAllocator.zig
src/analytics.zig
src/analytics/schema.zig
src/api/schema.zig
src/asan.zig
src/ast.zig
src/ast/Ast.zig
src/ast/ASTMemoryAllocator.zig
@@ -21,43 +18,19 @@ src/ast/base.zig
src/ast/Binding.zig
src/ast/BundledAst.zig
src/ast/CharFreq.zig
src/ast/ConvertESMExportsForHmr.zig
src/ast/E.zig
src/ast/Expr.zig
src/ast/foldStringAddition.zig
src/ast/G.zig
src/ast/ImportScanner.zig
src/ast/KnownGlobal.zig
src/ast/Macro.zig
src/ast/maybe.zig
src/ast/NewStore.zig
src/ast/Op.zig
src/ast/P.zig
src/ast/parse.zig
src/ast/parseFn.zig
src/ast/parseImportExport.zig
src/ast/parseJSXElement.zig
src/ast/parsePrefix.zig
src/ast/parseProperty.zig
src/ast/Parser.zig
src/ast/parseStmt.zig
src/ast/parseSuffix.zig
src/ast/parseTypescript.zig
src/ast/S.zig
src/ast/Scope.zig
src/ast/ServerComponentBoundary.zig
src/ast/SideEffects.zig
src/ast/skipTypescript.zig
src/ast/Stmt.zig
src/ast/Symbol.zig
src/ast/symbols.zig
src/ast/TS.zig
src/ast/TypeScript.zig
src/ast/UseDirective.zig
src/ast/visit.zig
src/ast/visitBinaryExpression.zig
src/ast/visitExpr.zig
src/ast/visitStmt.zig
src/async/posix_event_loop.zig
src/async/stub_event_loop.zig
src/async/windows_event_loop.zig
@@ -98,11 +71,6 @@ src/bun.js/api/bun/spawn.zig
src/bun.js/api/bun/spawn/stdio.zig
src/bun.js/api/bun/ssl_wrapper.zig
src/bun.js/api/bun/subprocess.zig
src/bun.js/api/bun/subprocess/Readable.zig
src/bun.js/api/bun/subprocess/ResourceUsage.zig
src/bun.js/api/bun/subprocess/StaticPipeWriter.zig
src/bun.js/api/bun/subprocess/SubprocessPipeReader.zig
src/bun.js/api/bun/subprocess/Writable.zig
src/bun.js/api/bun/udp_socket.zig
src/bun.js/api/bun/x509.zig
src/bun.js/api/BunObject.zig
@@ -135,7 +103,6 @@ src/bun.js/api/server/StaticRoute.zig
src/bun.js/api/server/WebSocketServerContext.zig
src/bun.js/api/streams.classes.zig
src/bun.js/api/Timer.zig
src/bun.js/api/Timer/DateHeaderTimer.zig
src/bun.js/api/Timer/EventLoopTimer.zig
src/bun.js/api/Timer/ImmediateObject.zig
src/bun.js/api/Timer/TimeoutObject.zig
@@ -282,84 +249,7 @@ src/bun.js/RuntimeTranspilerCache.zig
src/bun.js/SavedSourceMap.zig
src/bun.js/Strong.zig
src/bun.js/test/diff_format.zig
src/bun.js/test/diff/diff_match_patch.zig
src/bun.js/test/diff/printDiff.zig
src/bun.js/test/expect.zig
src/bun.js/test/expect/toBe.zig
src/bun.js/test/expect/toBeArray.zig
src/bun.js/test/expect/toBeArrayOfSize.zig
src/bun.js/test/expect/toBeBoolean.zig
src/bun.js/test/expect/toBeCloseTo.zig
src/bun.js/test/expect/toBeDate.zig
src/bun.js/test/expect/toBeDefined.zig
src/bun.js/test/expect/toBeEmpty.zig
src/bun.js/test/expect/toBeEmptyObject.zig
src/bun.js/test/expect/toBeEven.zig
src/bun.js/test/expect/toBeFalse.zig
src/bun.js/test/expect/toBeFalsy.zig
src/bun.js/test/expect/toBeFinite.zig
src/bun.js/test/expect/toBeFunction.zig
src/bun.js/test/expect/toBeGreaterThan.zig
src/bun.js/test/expect/toBeGreaterThanOrEqual.zig
src/bun.js/test/expect/toBeInstanceOf.zig
src/bun.js/test/expect/toBeInteger.zig
src/bun.js/test/expect/toBeLessThan.zig
src/bun.js/test/expect/toBeLessThanOrEqual.zig
src/bun.js/test/expect/toBeNaN.zig
src/bun.js/test/expect/toBeNegative.zig
src/bun.js/test/expect/toBeNil.zig
src/bun.js/test/expect/toBeNull.zig
src/bun.js/test/expect/toBeNumber.zig
src/bun.js/test/expect/toBeObject.zig
src/bun.js/test/expect/toBeOdd.zig
src/bun.js/test/expect/toBeOneOf.zig
src/bun.js/test/expect/toBePositive.zig
src/bun.js/test/expect/toBeString.zig
src/bun.js/test/expect/toBeSymbol.zig
src/bun.js/test/expect/toBeTrue.zig
src/bun.js/test/expect/toBeTruthy.zig
src/bun.js/test/expect/toBeTypeOf.zig
src/bun.js/test/expect/toBeUndefined.zig
src/bun.js/test/expect/toBeValidDate.zig
src/bun.js/test/expect/toBeWithin.zig
src/bun.js/test/expect/toContain.zig
src/bun.js/test/expect/toContainAllKeys.zig
src/bun.js/test/expect/toContainAllValues.zig
src/bun.js/test/expect/toContainAnyKeys.zig
src/bun.js/test/expect/toContainAnyValues.zig
src/bun.js/test/expect/toContainEqual.zig
src/bun.js/test/expect/toContainKey.zig
src/bun.js/test/expect/toContainKeys.zig
src/bun.js/test/expect/toContainValue.zig
src/bun.js/test/expect/toContainValues.zig
src/bun.js/test/expect/toEndWith.zig
src/bun.js/test/expect/toEqual.zig
src/bun.js/test/expect/toEqualIgnoringWhitespace.zig
src/bun.js/test/expect/toHaveBeenCalled.zig
src/bun.js/test/expect/toHaveBeenCalledOnce.zig
src/bun.js/test/expect/toHaveBeenCalledTimes.zig
src/bun.js/test/expect/toHaveBeenCalledWith.zig
src/bun.js/test/expect/toHaveBeenLastCalledWith.zig
src/bun.js/test/expect/toHaveBeenNthCalledWith.zig
src/bun.js/test/expect/toHaveLastReturnedWith.zig
src/bun.js/test/expect/toHaveLength.zig
src/bun.js/test/expect/toHaveNthReturnedWith.zig
src/bun.js/test/expect/toHaveProperty.zig
src/bun.js/test/expect/toHaveReturned.zig
src/bun.js/test/expect/toHaveReturnedTimes.zig
src/bun.js/test/expect/toHaveReturnedWith.zig
src/bun.js/test/expect/toInclude.zig
src/bun.js/test/expect/toIncludeRepeated.zig
src/bun.js/test/expect/toMatch.zig
src/bun.js/test/expect/toMatchInlineSnapshot.zig
src/bun.js/test/expect/toMatchObject.zig
src/bun.js/test/expect/toMatchSnapshot.zig
src/bun.js/test/expect/toSatisfy.zig
src/bun.js/test/expect/toStartWith.zig
src/bun.js/test/expect/toStrictEqual.zig
src/bun.js/test/expect/toThrow.zig
src/bun.js/test/expect/toThrowErrorMatchingInlineSnapshot.zig
src/bun.js/test/expect/toThrowErrorMatchingSnapshot.zig
src/bun.js/test/jest.zig
src/bun.js/test/pretty_format.zig
src/bun.js/test/snapshot.zig
@@ -487,6 +377,7 @@ src/collections/baby_list.zig
src/collections/bit_set.zig
src/collections/hive_array.zig
src/collections/multi_array_list.zig
src/collections/safety.zig
src/compile_target.zig
src/comptime_string_map.zig
src/copy_file.zig
@@ -596,6 +487,7 @@ src/defines.zig
src/deps/boringssl.translated.zig
src/deps/brotli_c.zig
src/deps/c_ares.zig
src/deps/diffz/DiffMatchPatch.zig
src/deps/libdeflate.zig
src/deps/libuv.zig
src/deps/lol-html.zig
@@ -654,7 +546,6 @@ src/http/AsyncHTTP.zig
src/http/CertificateInfo.zig
src/http/Decompressor.zig
src/http/Encoding.zig
src/http/ETag.zig
src/http/FetchRedirect.zig
src/http/HeaderBuilder.zig
src/http/Headers.zig
@@ -742,7 +633,6 @@ src/install/resolvers/folder_resolver.zig
src/install/versioned_url.zig
src/install/windows-shim/BinLinkingShim.zig
src/install/windows-shim/bun_shim_impl.zig
src/install/yarn.zig
src/interchange.zig
src/interchange/json.zig
src/interchange/toml.zig
@@ -790,9 +680,6 @@ src/Progress.zig
src/ptr.zig
src/ptr/Cow.zig
src/ptr/CowSlice.zig
src/ptr/owned.zig
src/ptr/owned/maybe.zig
src/ptr/owned/meta.zig
src/ptr/ref_count.zig
src/ptr/tagged_pointer.zig
src/ptr/weak_ptr.zig
@@ -816,11 +703,6 @@ src/s3/multipart_options.zig
src/s3/multipart.zig
src/s3/simple_request.zig
src/s3/storage_class.zig
src/safety.zig
src/safety/alloc.zig
src/safety/CriticalSection.zig
src/safety/thread_id.zig
src/safety/ThreadLock.zig
src/semver.zig
src/semver/ExternalString.zig
src/semver/SemverObject.zig
@@ -971,10 +853,6 @@ src/string/StringJoiner.zig
src/string/WTFStringImpl.zig
src/sys_uv.zig
src/sys.zig
src/sys/coreutils_error_map.zig
src/sys/Error.zig
src/sys/File.zig
src/sys/libuv_error_map.zig
src/system_timer.zig
src/test/fixtures.zig
src/test/recover.zig

View File

@@ -618,7 +618,7 @@ register_command(
-Doptimize=${ZIG_OPTIMIZE}
-Dcpu=${ZIG_CPU}
-Denable_logs=$<IF:$<BOOL:${ENABLE_LOGS}>,true,false>
-Denable_asan=$<IF:$<BOOL:${ENABLE_ZIG_ASAN}>,true,false>
-Denable_asan=$<IF:$<BOOL:${ENABLE_ASAN}>,true,false>
-Dversion=${VERSION}
-Dreported_nodejs_version=${NODEJS_VERSION}
-Dcanary=${CANARY_REVISION}
@@ -1034,6 +1034,7 @@ if(LINUX)
--ld-path=${LLD_PROGRAM}
-fno-pic
-Wl,-no-pie
-Wl,-icf=safe
-Wl,--as-needed
-Wl,-z,stack-size=12800000
-Wl,--compress-debug-sections=zlib
@@ -1059,13 +1060,6 @@ if(LINUX)
-Wl,--gc-sections
)
endif()
if (NOT DEBUG AND NOT ENABLE_ASAN)
target_link_options(${bun} PUBLIC
-Wl,-icf=safe
)
endif()
endif()
# --- Symbols list ---

View File

@@ -4,7 +4,7 @@ register_repository(
REPOSITORY
HdrHistogram/HdrHistogram_c
COMMIT
8dcce8f68512fca460b171bccc3a5afce0048779
652d51bcc36744fd1a6debfeb1a8a5f58b14022c
)
register_cmake_command(

View File

@@ -4,7 +4,7 @@ register_repository(
REPOSITORY
libarchive/libarchive
COMMIT
9525f90ca4bd14c7b335e2f8c84a4607b0af6bdf
7118f97c26bf0b2f426728b482f86508efc81d02
)
register_cmake_command(

View File

@@ -13,52 +13,14 @@ set(MIMALLOC_CMAKE_ARGS
-DMI_BUILD_SHARED=OFF
-DMI_BUILD_TESTS=OFF
-DMI_USE_CXX=ON
-DMI_OVERRIDE=OFF
-DMI_OSX_ZONE=OFF
-DMI_OSX_INTERPOSE=OFF
-DMI_SKIP_COLLECT_ON_EXIT=ON
# ```
# mimalloc_allow_large_os_pages=0 BUN_PORT=3004 mem bun http-hello.js
# Started development server: http://localhost:3004
#
# Peak memory usage: 52 MB
#
# mimalloc_allow_large_os_pages=1 BUN_PORT=3004 mem bun http-hello.js
# Started development server: http://localhost:3004
#
# Peak memory usage: 74 MB
# ```
#
# ```
# mimalloc_allow_large_os_pages=1 mem bun --eval 1
#
# Peak memory usage: 52 MB
#
# mimalloc_allow_large_os_pages=0 mem bun --eval 1
#
# Peak memory usage: 30 MB
# ```
-DMI_NO_THP=1
)
if (ABI STREQUAL "musl")
list(APPEND MIMALLOC_CMAKE_ARGS -DMI_LIBC_MUSL=ON)
endif()
if(ENABLE_ASAN)
list(APPEND MIMALLOC_CMAKE_ARGS -DMI_TRACK_ASAN=ON)
list(APPEND MIMALLOC_CMAKE_ARGS -DMI_OVERRIDE=OFF)
list(APPEND MIMALLOC_CMAKE_ARGS -DMI_OSX_ZONE=OFF)
list(APPEND MIMALLOC_CMAKE_ARGS -DMI_OSX_INTERPOSE=OFF)
list(APPEND MIMALLOC_CMAKE_ARGS -DMI_DEBUG_UBSAN=ON)
elseif(APPLE OR LINUX)
if(APPLE)
list(APPEND MIMALLOC_CMAKE_ARGS -DMI_OVERRIDE=OFF)
list(APPEND MIMALLOC_CMAKE_ARGS -DMI_OSX_ZONE=OFF)
list(APPEND MIMALLOC_CMAKE_ARGS -DMI_OSX_INTERPOSE=OFF)
else()
list(APPEND MIMALLOC_CMAKE_ARGS -DMI_OVERRIDE=ON)
list(APPEND MIMALLOC_CMAKE_ARGS -DMI_OSX_ZONE=OFF)
list(APPEND MIMALLOC_CMAKE_ARGS -DMI_OSX_INTERPOSE=OFF)
endif()
endif()
if(DEBUG)
@@ -69,12 +31,6 @@ if(ENABLE_VALGRIND)
list(APPEND MIMALLOC_CMAKE_ARGS -DMI_VALGRIND=ON)
endif()
# Enable SIMD optimizations when not building for baseline (older CPUs)
if(NOT ENABLE_BASELINE)
list(APPEND MIMALLOC_CMAKE_ARGS -DMI_OPT_ARCH=ON)
list(APPEND MIMALLOC_CMAKE_ARGS -DMI_OPT_SIMD=ON)
endif()
if(WIN32)
if(DEBUG)
set(MIMALLOC_LIBRARY mimalloc-static-debug)
@@ -97,7 +53,6 @@ if(APPLE OR (LINUX AND NOT DEBUG))
set(MIMALLOC_LIBRARY CMakeFiles/mimalloc-obj.dir/src/static.c.o)
endif()
register_cmake_command(
TARGET
mimalloc

View File

@@ -2,7 +2,7 @@ option(WEBKIT_VERSION "The version of WebKit to use")
option(WEBKIT_LOCAL "If a local version of WebKit should be used instead of downloading")
if(NOT WEBKIT_VERSION)
set(WEBKIT_VERSION 684d4551ce5f62683476409d7402424e0f6eafb5)
set(WEBKIT_VERSION 1098cc50652ab1eab171f58f7669e19ca6c276ae)
endif()
string(SUBSTRING ${WEBKIT_VERSION} 0 16 WEBKIT_VERSION_PREFIX)

File diff suppressed because it is too large Load Diff

View File

@@ -320,6 +320,7 @@ Bun automatically sets the `Content-Type` header for request bodies when not exp
- For `Blob` objects, uses the blob's `type`
- For `FormData`, sets appropriate multipart boundary
- For JSON objects, sets `application/json`
## Debugging

View File

@@ -164,70 +164,6 @@ Static responses do not allocate additional memory after initialization. You can
Static route responses are cached for the lifetime of the server object. To reload static routes, call `server.reload(options)`.
### File Responses vs Static Responses
When serving files in routes, there are two distinct behaviors depending on whether you buffer the file content or serve it directly:
```ts
Bun.serve({
routes: {
// Static route - content is buffered in memory at startup
"/logo.png": new Response(await Bun.file("./logo.png").bytes()),
// File route - content is read from filesystem on each request
"/download.zip": new Response(Bun.file("./download.zip")),
},
});
```
**Static routes** (`new Response(await file.bytes())`) buffer content in memory at startup:
- **Zero filesystem I/O** during requests - content served entirely from memory
- **ETag support** - Automatically generates and validates ETags for caching
- **If-None-Match** - Returns `304 Not Modified` when client ETag matches
- **No 404 handling** - Missing files cause startup errors, not runtime 404s
- **Memory usage** - Full file content stored in RAM
- **Best for**: Small static assets, API responses, frequently accessed files
**File routes** (`new Response(Bun.file(path))`) read from filesystem per request:
- **Filesystem reads** on each request - checks file existence and reads content
- **Built-in 404 handling** - Returns `404 Not Found` if file doesn't exist or becomes inaccessible
- **Last-Modified support** - Uses file modification time for `If-Modified-Since` headers
- **If-Modified-Since** - Returns `304 Not Modified` when file hasn't changed since client's cached version
- **Range request support** - Automatically handles partial content requests with `Content-Range` headers
- **Streaming transfers** - Uses buffered reader with backpressure handling for efficient memory usage
- **Memory efficient** - Only buffers small chunks during transfer, not entire file
- **Best for**: Large files, dynamic content, user uploads, files that change frequently
### HTTP Caching Behavior
Both route types implement HTTP caching standards but with different strategies:
#### Static Routes Caching
- **ETag generation**: Automatically computes ETag hash from content at startup
- **If-None-Match**: Validates client ETag against server ETag
- **304 responses**: Returns `304 Not Modified` with empty body when ETags match
- **Cache headers**: Inherits any `Cache-Control` headers you provide in the Response
- **Consistency**: ETag remains constant until server restart or route reload
#### File Routes Caching
- **Last-Modified**: Uses file's `mtime` for `Last-Modified` header
- **If-Modified-Since**: Compares client date with file modification time
- **304 responses**: Returns `304 Not Modified` when file unchanged since client's cached version
- **Content-Length**: Automatically set based on current file size
- **Dynamic validation**: Checks file modification time on each request
#### Status Code Handling
Both route types automatically adjust status codes:
- **200 → 204**: Empty files (0 bytes) return `204 No Content` instead of `200 OK`
- **200 → 304**: Successful cache validation returns `304 Not Modified`
- **File routes only**: Missing or inaccessible files return `404 Not Found`
```ts
const server = Bun.serve({
static: {

View File

@@ -208,8 +208,8 @@ export class ArrayBufferSink {
*
* This API might change later to separate Uint8ArraySink and ArrayBufferSink
*/
flush(): number | Uint8Array<ArrayBuffer> | ArrayBuffer;
end(): ArrayBuffer | Uint8Array<ArrayBuffer>;
flush(): number | Uint8Array | ArrayBuffer;
end(): ArrayBuffer | Uint8Array;
}
```

View File

@@ -158,7 +158,7 @@ See [Test > Lifecycle](https://bun.com/docs/test/lifecycle) for complete documen
## Mocks
Create mock functions with the `mock` function.
Create mock functions with the `mock` function. Mocks are automatically reset between tests.
```ts
import { test, expect, mock } from "bun:test";

View File

@@ -14,7 +14,7 @@ if (typeof Bun !== "undefined") {
---
In TypeScript environments, the previous approach will result in a type error unless `@types/bun` is installed. To avoid this, you can check `process.versions` instead.
In TypeScript environments, the previous approach will result in a type error unless `bun-types` is globally installed. To avoid this, you can check `process.versions` instead.
```ts
if (process.versions.bun) {

View File

@@ -148,7 +148,7 @@ This page is updated regularly to reflect compatibility status of the latest ver
### [`node:vm`](https://nodejs.org/api/vm.html)
🟡 Core functionality and ES modules are implemented, including `vm.Script`, `vm.createContext`, `vm.runInContext`, `vm.runInNewContext`, `vm.runInThisContext`, `vm.compileFunction`, `vm.isContext`, `vm.Module`, `vm.SourceTextModule`, `vm.SyntheticModule`, and `importModuleDynamically` support. Options like `timeout` and `breakOnSigint` are fully supported. Missing `vm.measureMemory` and some `cachedData` functionality.
🟡 Core functionality works, but experimental VM ES modules are not implemented, including `vm.Module`, `vm.SourceTextModule`, `vm.SyntheticModule`,`importModuleDynamically`, and `vm.measureMemory`. Options like `timeout`, `breakOnSigint`, `cachedData` are not implemented yet.
### [`node:wasi`](https://nodejs.org/api/wasi.html)
@@ -214,10 +214,6 @@ The table below lists all globals implemented by Node.js and Bun's current compa
🟢 Fully implemented.
### [`Atomics`](https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Atomics)
🟢 Fully implemented.
### [`BroadcastChannel`](https://developer.mozilla.org/en-US/docs/Web/API/BroadcastChannel)
🟢 Fully implemented.

View File

@@ -532,74 +532,6 @@ Hello World! pwd=C:\Users\Demo
Bun Shell is a small programming language in Bun that is implemented in Zig. It includes a handwritten lexer, parser, and interpreter. Unlike bash, zsh, and other shells, Bun Shell runs operations concurrently.
## Security in the Bun shell
By design, the Bun shell _does not invoke a system shell_ (like `/bin/sh`) and
is instead a re-implementation of bash that runs in the same Bun process,
designed with security in mind.
When parsing command arguments, it treats all _interpolated variables_ as single, literal strings.
This protects the Bun shell against **command injection**:
```js
import { $ } from "bun";
const userInput = "my-file.txt; rm -rf /";
// SAFE: `userInput` is treated as a single quoted string
await $`ls ${userInput}`;
```
In the above example, `userInput` is treated as a single string. This causes
the `ls` command to try to read the contents of a single directory named
"my-file; rm -rf /".
### Security considerations
While command injection is prevented by default, developers are still
responsible for security in certain scenarios.
Similar to the `Bun.spawn` or `node:child_process.exec()` APIs, you can intentionally
execute a command which spawns a new shell (e.g. `bash -c`) with arguments.
When you do this, you hand off control, and Bun's built-in protections no
longer apply to the string interpreted by that new shell.
```js
import { $ } from "bun";
const userInput = "world; touch /tmp/pwned";
// UNSAFE: You have explicitly started a new shell process with `bash -c`.
// This new shell will execute the `touch` command. Any user input
// passed this way must be rigorously sanitized.
await $`bash -c "echo ${userInput}"`;
```
### Argument injection
The Bun shell cannot know how an external command interprets its own
command-line arguments. An attacker can supply input that the target program
recognizes as one of its own options or flags, leading to unintended behavior.
```js
import { $ } from "bun";
// Malicious input formatted as a Git command-line flag
const branch = "--upload-pack=echo pwned";
// UNSAFE: While Bun safely passes the string as a single argument,
// the `git` program itself sees and acts upon the malicious flag.
await $`git ls-remote origin ${branch}`;
```
{% callout %}
**Recommendation** — As is best practice in every language, always sanitize
user-provided input before passing it as an argument to an external command.
The responsibility for validating arguments rests with your application code.
{% /callout %}
## Credits
Large parts of this API were inspired by [zx](https://github.com/google/zx), [dax](https://github.com/dsherret/dax), and [bnx](https://github.com/wobsoriano/bnx). Thank you to the authors of those projects.

View File

@@ -426,54 +426,6 @@ test("exactly two assertions", () => {
This helps ensure all your assertions run, especially in complex async code with multiple code paths.
## Type Testing
Bun includes `expectTypeOf` for testing typescript types, compatible with Vitest.
### expectTypeOf
{% callout %}
**Note** — These functions are no-ops at runtime - you need to run TypeScript separately to verify the type checks.
{% endcallout %}
The `expectTypeOf` function provides type-level assertions that are checked by TypeScript's type checker. **Important**:
To test your types:
1. Write your type assertions using `expectTypeOf`
2. Run `bunx tsc --noEmit` to check that your types are correct
```ts
import { expectTypeOf } from "bun:test";
// Basic type assertions
expectTypeOf<string>().toEqualTypeOf<string>();
expectTypeOf(123).toBeNumber();
expectTypeOf("hello").toBeString();
// Object type matching
expectTypeOf({ a: 1, b: "hello" }).toMatchObjectType<{ a: number }>();
// Function types
function greet(name: string): string {
return `Hello ${name}`;
}
expectTypeOf(greet).toBeFunction();
expectTypeOf(greet).parameters.toEqualTypeOf<[string]>();
expectTypeOf(greet).returns.toEqualTypeOf<string>();
// Array types
expectTypeOf([1, 2, 3]).items.toBeNumber();
// Promise types
expectTypeOf(Promise.resolve(42)).resolves.toBeNumber();
```
For full documentation on expectTypeOf matchers, see the [API Reference](/reference/bun/test/expectTypeOf)
## Matchers
Bun implements the following matchers. Full Jest compatibility is on the roadmap; track progress [here](https://github.com/oven-sh/bun/issues/1825).
@@ -677,17 +629,17 @@ Bun implements the following matchers. Full Jest compatibility is on the roadmap
---
-
-
- [`.toHaveReturnedWith()`](https://jestjs.io/docs/expect#tohavereturnedwithvalue)
---
-
-
- [`.toHaveLastReturnedWith()`](https://jestjs.io/docs/expect#tohavelastreturnedwithvalue)
---
-
-
- [`.toHaveNthReturnedWith()`](https://jestjs.io/docs/expect#tohaventhreturnedwithnthcall-value)
---

1
examples/.gitignore vendored Normal file
View File

@@ -0,0 +1 @@
bun-examples-all

View File

7
examples/add.rs Normal file
View File

@@ -0,0 +1,7 @@
#[no_mangle]
pub extern "C" fn add(a: i32, b: i32) -> i32 {
a + b
}
// to compile:
// rustc --crate-type cdylib add.rs

12
examples/add.ts Normal file
View File

@@ -0,0 +1,12 @@
import { dlopen, suffix } from "bun:ffi";
const {
symbols: { add },
} = dlopen(`./libadd.${suffix}`, {
add: {
args: ["i32", "i32"],
returns: "i32",
},
});
console.log(add(1, 2));

6
examples/add.zig Normal file
View File

@@ -0,0 +1,6 @@
pub export fn add(a: i32, b: i32) i32 {
return a + b;
}
// to compile:
// zig build-lib -OReleaseFast ./add.zig -dynamic --name add

View File

@@ -0,0 +1,89 @@
// To run this example:
//
// bun --hot bun-hot-websockets.js
//
const css = ([inner]) => {
return inner;
};
const styles = css`
#bun {
margin: 0 auto;
margin-top: 200px;
object-fit: cover;
}
html,
body {
margin: 0;
padding: 0;
}
body {
background: #f1239f;
font-family: "Inter", sans-serif;
display: flex;
align-items: center;
justify-content: center;
align-content: center;
color: white;
}
h1 {
padding: 0;
text-align: center;
font-size: 3rem;
-webkit-text-stroke: 2px black;
}
* {
box-sizing: border-box;
}
`;
Bun.serve({
websocket: {
message(ws, msg) {
ws.send(styles);
},
},
fetch(req, server) {
if (req.url.endsWith("/hot")) {
if (server.upgrade(req))
return new Response("", {
status: 101,
});
}
return new Response(
`
<!DOCTYPE html>
<html>
<head>
<meta charset="utf-8">
<title>WebSockets</title>
</head>
<body>
<style></style>
<script>
const ws = new WebSocket("ws://localhost:3000/hot");
const style = document.querySelector("style");
ws.onmessage = (e) => {
style.innerHTML = e.data;
};
setInterval(() => {
ws.send("ping");
}, 8);
</script>
<div id="app">
<img src="https://bun.com/logo.svg" alt="Bun" id='bun' />
<h1>bun --hot websockets</h1>
</div>
</body>
`,
{
headers: {
"Content-Type": "text/html; charset=utf-8",
},
},
);
},
});

9
examples/cat.ts Normal file
View File

@@ -0,0 +1,9 @@
import { resolve } from "path";
const { write, stdout, file } = Bun;
import { argv } from "process";
const path = resolve(argv.at(-1)!);
await write(stdout, file(path));
Bun.stdout;
process.stdout;

View File

@@ -0,0 +1,11 @@
const sequence = [1, 2, 3];
sequence.toReversed(); // => [3, 2, 1]
sequence; // => [1, 2, 3]
const outOfOrder = new Uint8Array([3, 1, 2]);
outOfOrder.toSorted(); // => Uint8Array [1, 2, 3]
outOfOrder; // => Uint8Array [3, 1, 2]
const correctionNeeded = [1, 1, 3];
correctionNeeded.with(1, 2); // => [1, 2, 3]
correctionNeeded; // => [1, 1, 3]

23
examples/hashing.js Normal file
View File

@@ -0,0 +1,23 @@
// Accepts a string, TypedArray, or Blob (file blob support is not implemented but planned)
const input = "hello world".repeat(400);
// Bun.hash() defaults to Wyhash because it's fast
console.log(Bun.hash(input));
console.log(Bun.hash.wyhash(input));
// and returns a bigint
// all of these hashing functions return number if 32-bit or bigint if 64-bit, not typed arrays.
console.log(Bun.hash.adler32(input)); // number
console.log(Bun.hash.crc32(input)); // number
console.log(Bun.hash.cityHash32(input)); // number
console.log(Bun.hash.cityHash64(input)); // bigint
console.log(Bun.hash.xxHash32(input)); // number
console.log(Bun.hash.xxHash64(input)); // bigint
console.log(Bun.hash.xxHash3(input)); // bigint
console.log(Bun.hash.murmur32v3(input)); // number
console.log(Bun.hash.murmur32v2(input)); // number
console.log(Bun.hash.murmur64v2(input)); // bigint
console.log(Bun.hash.rapidhash(input)); // bigint
// Second argument accepts a seed where relevant
console.log(Bun.hash(input, 12345));

37
examples/html-rewriter.ts Normal file
View File

@@ -0,0 +1,37 @@
// Start a fast HTTP server from a function
Bun.serve({
async fetch(req) {
const { pathname } = new URL(req.url);
if (!(pathname.startsWith("/https://") || pathname.startsWith("/http://"))) {
return new Response("Enter a path that starts with https:// or http://\n", {
status: 400,
});
}
const response = await fetch(req.url.substring("http://localhost:3000/".length), req.clone());
return new HTMLRewriter()
.on("a[href]", {
element(element) {
element.setAttribute("href", "https://www.youtube.com/watch?v=dQw4w9WgXcQ");
},
})
.transform(response);
},
// this is called when fetch() throws or rejects
// error(err: Error) {
// },
// this boolean enables the bun's default error handler
// sometime after the initial release, it will auto reload as well
development: process.env.NODE_ENV !== "production",
// note: this isn't node, but for compatibility bun supports process.env + more stuff in process
// SSL is enabled if these two are set
// certFile: './cert.pem',
// keyFile: './key.pem',
port: 3000, // number or string
});

View File

@@ -0,0 +1,76 @@
import { file, serve } from "bun";
import { existsSync, statSync } from "fs";
serve({
fetch(req: Request) {
let pathname = new URL(req.url).pathname.substring(1);
if (pathname == "") {
pathname = import.meta.url.replace("file://", "");
}
if (!existsSync(pathname)) {
return new Response(null, { status: 404 });
}
const stats = statSync(pathname);
// https://github.com/gornostay25/svelte-adapter-bun/blob/master/src/sirv.js
const headers = new Headers({
"Content-Length": "" + stats.size,
"Last-Modified": stats.mtime.toUTCString(),
ETag: `W/"${stats.size}-${stats.mtime.getTime()}"`,
});
if (req.headers.get("if-none-match") === headers.get("ETag")) {
return new Response(null, { status: 304 });
}
const opts = { code: 200, start: 0, end: Infinity, range: false };
if (req.headers.has("range")) {
opts.code = 206;
let [x, y] = req.headers.get("range")!.replace("bytes=", "").split("-");
let end = (opts.end = parseInt(y, 10) || stats.size - 1);
let start = (opts.start = parseInt(x, 10) || 0);
if (start >= stats.size || end >= stats.size) {
headers.set("Content-Range", `bytes */${stats.size}`);
return new Response(null, {
headers: headers,
status: 416,
});
}
headers.set("Content-Range", `bytes ${start}-${end}/${stats.size}`);
headers.set("Content-Length", "" + (end - start + 1));
headers.set("Accept-Ranges", "bytes");
opts.range = true;
}
if (opts.range) {
return new Response(file(pathname).slice(opts.start, opts.end), {
headers,
status: opts.code,
});
}
return new Response(file(pathname), { headers, status: opts.code });
},
// this is called when fetch() throws or rejects
// error(err: Error) {
// return new Response("uh oh! :(" + String(err.toString()), { status: 500 });
// },
// this boolean enables the bun's default error handler
// sometime after the initial release, it will auto reload as well
development: process.env.NODE_ENV !== "production",
// note: this isn't node, but for compatibility bun supports process.env + more stuff in process
// SSL is enabled if these two are set
// certFile: './cert.pem',
// keyFile: './key.pem',
port: 3000, // number or string
hostname: "localhost", // defaults to 0.0.0.0
});

31
examples/http-file.ts Normal file
View File

@@ -0,0 +1,31 @@
import { file, serve } from "bun";
serve({
fetch(req: Request) {
const pathname = new URL(req.url).pathname.substring(1);
// If the URL is empty, display this file.
if (pathname === "") {
return new Response(file(import.meta.url.replace("file://", "")));
}
return new Response(file(pathname));
},
// this is called when fetch() throws or rejects
// error(err: Error) {
// return new Response("uh oh! :(" + String(err.toString()), { status: 500 });
// },
// this boolean enables the bun's default error handler
// sometime after the initial release, it will auto reload as well
development: process.env.NODE_ENV !== "production",
// note: this isn't node, but for compatibility bun supports process.env + more stuff in process
// SSL is enabled if these two are set
// certFile: './cert.pem',
// keyFile: './key.pem',
port: 3000, // number or string
hostname: "localhost", // defaults to 0.0.0.0
});

View File

@@ -0,0 +1,17 @@
import { serve } from "bun";
serve({
async fetch(req) {
// body is a ReadableStream
const body = req.body;
const writer = Bun.file(`upload.${Date.now()}.txt`).writer();
for await (const chunk of body!) {
writer.write(chunk);
}
const wrote = await writer.end();
// @ts-ignore
return Response.json({ wrote, type: req.headers.get("Content-Type") });
},
});

12
examples/http-stop.ts Normal file
View File

@@ -0,0 +1,12 @@
import { serve } from "bun";
const server = serve({
fetch(req) {
return new Response(`Pending requests count: ${this.pendingRequests}`);
},
});
// Stop the server after 5 seconds
setTimeout(() => {
server.stop();
}, 5000);

34
examples/http.ts Normal file
View File

@@ -0,0 +1,34 @@
// Start a fast HTTP server from a function
Bun.serve({
fetch(req: Request) {
return new Response(`Echo: ${req.url}`);
},
// baseURI: "http://localhost:3000",
// this is called when fetch() throws or rejects
// error(err: Error) {
// return new Response("uh oh! :(\n" + err.toString(), { status: 500 });
// },
// this boolean enables bun's default error handler
development: process.env.NODE_ENV !== "production",
// note: this isn't node, but for compatibility bun supports process.env + more stuff in process
// SSL is enabled if these two are set
// certFile: './cert.pem',
// keyFile: './key.pem',
port: 3000, // number or string
});
// Start a fast HTTP server from the main file's export
// export default {
// fetch(req) {
// return new Response(
// `This is another way to start a server!
// if the main file export default's an object
// with 'fetch'. Bun automatically calls Bun.serve`
// );
// },
// // so autocomplete & type checking works
// } as Bun.Serve;

193
examples/lambda.ts Normal file
View File

@@ -0,0 +1,193 @@
const { AWS_LAMBDA_RUNTIME_API, LAMBDA_TASK_ROOT, _HANDLER } = process.env;
if (!AWS_LAMBDA_RUNTIME_API || AWS_LAMBDA_RUNTIME_API === "") {
throw new Error("AWS_LAMBDA_RUNTIME_API is not set");
}
const nextURL = `http://${AWS_LAMBDA_RUNTIME_API}/2018-06-01/runtime/invocation/next`;
const sourceDir = LAMBDA_TASK_ROOT;
if (!sourceDir) {
throw new Error("handler is not set");
}
if (!_HANDLER) {
throw new Error("handler is not set");
}
// don't care if this fails
if (process.cwd() !== sourceDir) {
try {
process.chdir(sourceDir);
} catch (e) {}
}
var handlerDot = _HANDLER.lastIndexOf(".");
var sourcefile = handlerDot > 0 ? _HANDLER.substring(0, handlerDot) : _HANDLER;
if (sourcefile.length === 0) {
throw new Error("handler is not set");
}
if (!sourcefile.startsWith("/")) {
sourcefile = `./${sourcefile}`;
}
function noop() {}
const method = (handlerDot > 0 ? _HANDLER.substring(handlerDot) : "") || "GET";
if (typeof process.env.VERBOSE !== "undefined") {
console.time(`Loaded ${sourcefile}`);
}
var Handler;
try {
Handler = await import(sourcefile);
} catch (e: any) {
console.error("Error loading sourcefile:", e);
try {
await fetch(new URL(`http://${AWS_LAMBDA_RUNTIME_API}/2018-06-01/runtime/init/error`).href, {
method: "POST",
headers: {
"Content-Type": "application/json",
},
body: JSON.stringify({
errorMessage: e.message,
errorType: e.name,
stackTrace: e?.stack?.split("\n") ?? [],
}),
});
} catch (e2) {
console.error("Error sending error to runtime:", e2);
}
process.exit(1);
}
if (typeof process.env.VERBOSE !== "undefined") {
console.timeEnd(`Loaded ${sourcefile}`);
}
const handlerFunction = Handler.default?.fetch;
if (typeof handlerFunction !== "function") {
const e = new Error(`${sourcefile} must export default a function called fetch
Here is an example:
export default {
fetch(req) {
return new Response("Hello World");
}
}
`);
console.error(e);
try {
await fetch(new URL(`http://${AWS_LAMBDA_RUNTIME_API}/2018-06-01/runtime/init/error`).href, {
method: "POST",
headers: {
"Content-Type": "application/json",
},
body: JSON.stringify({
errorMessage: e.message,
errorType: e.name,
stackTrace: e?.stack?.split("\n") ?? [],
}),
});
} catch (e2) {
console.error("Error sending error to runtime:", e2);
}
process.exit(1);
}
var baseURLString = AWS_LAMBDA_RUNTIME_API;
if ("baseURI" in Handler.default) {
baseURLString = Handler.default.baseURI?.toString();
}
var baseURL;
try {
baseURL = new URL(baseURLString);
} catch (e: any) {
console.error("Error parsing baseURI:", e);
try {
await fetch(new URL(`http://${AWS_LAMBDA_RUNTIME_API}/2018-06-01/runtime/init/error`).href, {
method: "POST",
headers: {
"Content-Type": "application/json",
},
body: JSON.stringify({
errorMessage: e.message,
errorType: e.name,
stackTrace: e?.stack?.split("\n") || [],
}),
});
} catch (e2) {
console.error("Error sending error to runtime:", e2);
}
process.exit(1);
}
async function runHandler(response: Response) {
const traceID = response.headers.get("Lambda-Runtime-Trace-Id");
const requestID = response.headers.get("Lambda-Runtime-Aws-Request-Id");
var request = new Request(baseURL.href, {
method,
headers: response.headers,
body: parseInt(response.headers.get("Content-Length") || "0", 10) > 0 ? await response.blob() : undefined,
});
// we are done with the Response object here
// allow it to be GC'd
(response as any) = undefined;
var result: Response;
try {
if (typeof process.env.VERBOSE !== "undefined") {
console.time(`[${traceID}] Run ${request.url}`);
}
result = handlerFunction(request, {});
if (result && (result as any).then) {
await result;
}
} catch (e1: any) {
if (typeof process.env.VERBOSE !== "undefined") {
console.error(`[${traceID}] Error running handler:`, e1);
}
fetch(`http://${AWS_LAMBDA_RUNTIME_API}/2018-06-01/runtime/invocation/${requestID}/error`, {
method: "POST",
body: JSON.stringify({
errorMessage: e1.message,
errorType: e1.name,
stackTrace: e1?.stack?.split("\n") ?? [],
}),
}).finally(noop);
return;
} finally {
if (typeof process.env.VERBOSE !== "undefined") {
console.timeEnd(`[${traceID}] Run ${request.url}`);
}
}
if (!result || !("headers" in result)) {
await fetch(`http://${AWS_LAMBDA_RUNTIME_API}/2018-06-01/runtime/invocation/${requestID}/error`, {
method: "POST",
body: JSON.stringify({
errorMessage: "Expected Response object",
errorType: "ExpectedResponseObject",
stackTrace: [],
}),
});
return;
}
await fetch(`http://${AWS_LAMBDA_RUNTIME_API}/2018-06-01/runtime/invocation/${requestID}/response`, {
method: "POST",
headers: result.headers,
body: await result.blob(),
});
(result as any) = undefined;
}
while (true) {
fetch(nextURL).then(runHandler, console.error);
}
export {};

48
examples/macros/bun.lock Normal file
View File

@@ -0,0 +1,48 @@
{
"lockfileVersion": 1,
"workspaces": {
"": {
"name": "macros",
"dependencies": {
"moment": "^2.29.1",
"papaparse": "^5.3.1",
"react": "^17.0.2",
"react-dom": "^17.0.2",
"react-refresh": "^0.10.0",
},
"devDependencies": {
"@types/react": "^17.0.24",
"@types/react-dom": "^17.0.9",
},
},
},
"packages": {
"@types/prop-types": ["@types/prop-types@15.7.5", "", {}, "sha512-JCB8C6SnDoQf0cNycqd/35A7MjcnK+ZTqE7judS6o7utxUCg6imJg3QK2qzHKszlTjcj2cn+NwMB2i96ubpj7w=="],
"@types/react": ["@types/react@17.0.53", "", { "dependencies": { "@types/prop-types": "*", "@types/scheduler": "*", "csstype": "^3.0.2" } }, "sha512-1yIpQR2zdYu1Z/dc1OxC+MA6GR240u3gcnP4l6mvj/PJiVaqHsQPmWttsvHsfnhfPbU2FuGmo0wSITPygjBmsw=="],
"@types/react-dom": ["@types/react-dom@17.0.19", "", { "dependencies": { "@types/react": "^17" } }, "sha512-PiYG40pnQRdPHnlf7tZnp0aQ6q9tspYr72vD61saO6zFCybLfMqwUCN0va1/P+86DXn18ZWeW30Bk7xlC5eEAQ=="],
"@types/scheduler": ["@types/scheduler@0.16.2", "", {}, "sha512-hppQEBDmlwhFAXKJX2KnWLYu5yMfi91yazPb2l+lbJiwW+wdo1gNeRA+3RgNSO39WYX2euey41KEwnqesU2Jew=="],
"csstype": ["csstype@3.1.1", "", {}, "sha512-DJR/VvkAvSZW9bTouZue2sSxDwdTN92uHjqeKVm+0dAqdfNykRzQ95tay8aXMBAAPpUiq4Qcug2L7neoRh2Egw=="],
"js-tokens": ["js-tokens@4.0.0", "", {}, "sha512-RdJUflcE3cUzKiMqQgsCu06FPu9UdIJO0beYbPhHN4k6apgJtifcoCtT9bcxOpYBtpD2kCM6Sbzg4CausW/PKQ=="],
"loose-envify": ["loose-envify@1.4.0", "", { "dependencies": { "js-tokens": "^3.0.0 || ^4.0.0" }, "bin": { "loose-envify": "cli.js" } }, "sha512-lyuxPGr/Wfhrlem2CL/UcnUc1zcqKAImBDzukY7Y5F/yQiNdko6+fRLevlw1HgMySw7f611UIY408EtxRSoK3Q=="],
"moment": ["moment@2.29.4", "", {}, "sha512-5LC9SOxjSc2HF6vO2CyuTDNivEdoz2IvyJJGj6X8DJ0eFyfszE0QiEd+iXmBvUP3WHxSjFH/vIsA0EN00cgr8w=="],
"object-assign": ["object-assign@4.1.1", "", {}, "sha512-rJgTQnkUnH1sFw8yT6VSU3zD3sWmu6sZhIseY8VX+GRu3P6F7Fu+JNDoXfklElbLJSnc3FUQHVe4cU5hj+BcUg=="],
"papaparse": ["papaparse@5.3.2", "", {}, "sha512-6dNZu0Ki+gyV0eBsFKJhYr+MdQYAzFUGlBMNj3GNrmHxmz1lfRa24CjFObPXtjcetlOv5Ad299MhIK0znp3afw=="],
"react": ["react@17.0.2", "", { "dependencies": { "loose-envify": "^1.1.0", "object-assign": "^4.1.1" } }, "sha512-gnhPt75i/dq/z3/6q/0asP78D0u592D5L1pd7M8P+dck6Fu/jJeL6iVVK23fptSUZj8Vjf++7wXA8UNclGQcbA=="],
"react-dom": ["react-dom@17.0.2", "", { "dependencies": { "loose-envify": "^1.1.0", "object-assign": "^4.1.1", "scheduler": "^0.20.2" }, "peerDependencies": { "react": "17.0.2" } }, "sha512-s4h96KtLDUQlsENhMn1ar8t2bEa+q/YAtj8pPPdIjPDGBDIVNsrD9aXNWqspUe6AzKCIG0C1HZZLqLV7qpOBGA=="],
"react-refresh": ["react-refresh@0.10.0", "", {}, "sha512-PgidR3wST3dDYKr6b4pJoqQFpPGNKDSCDx4cZoshjXipw3LzO7mG1My2pwEzz2JVkF+inx3xRpDeQLFQGH/hsQ=="],
"scheduler": ["scheduler@0.20.2", "", { "dependencies": { "loose-envify": "^1.1.0", "object-assign": "^4.1.1" } }, "sha512-2eWfGgAqqWFGqtdMmcL5zCMK1U8KlXv8SQFGglL3CEtd0aDVDWgeF/YoCmvln55m5zSk3J/20hTaSBeSObsQDQ=="],
}
}

View File

@@ -0,0 +1,30 @@
import { fetchCSV } from "macro:fetchCSV";
export const Covid19 = () => {
const rows = fetchCSV("https://covid19.who.int/WHO-COVID-19-global-data.csv", {
last: 100,
columns: ["New_cases", "Date_reported", "Country"],
});
return (
<div>
<h2>Covid-19</h2>
<h6>last {rows.length} updates from the WHO</h6>
<div className="Table">
<div className="Header">
<div className="Heading">New Cases</div>
<div className="Heading">Date</div>
<div className="Heading">Country</div>
</div>
{rows.map((row, index) => (
<div className="Row" key={index}>
<div className="Column">{row[0]}</div>
<div className="Column">{row[1]}</div>
<div className="Column">{row[2]}</div>
</div>
))}
</div>
</div>
);
};

View File

@@ -0,0 +1,15 @@
// source code
import { matchInFile } from "macro:matchInFile";
export const IPAddresses = () => (
<div>
<h2>recent ip addresses</h2>
<div className="Lines">
{matchInFile("access.log", /^(?:[0-9]{1,3}\.){3}[0-9]{1,3}/).map((ipAddress, index) => (
<div className="Line" key={index}>
{ipAddress}
</div>
))}
</div>
</div>
);

View File

@@ -0,0 +1,15 @@
import * as ReactDOM from "react-dom";
import * as React from "react";
import { IPAddresses } from "./example";
import { Covid19 } from "./covid19";
const Start = function () {
const root = document.createElement("div");
document.body.appendChild(root);
// comment out to switch between examples
// ReactDOM.render(<IPAddresses />, root);
ReactDOM.render(<Covid19 />, root);
};
Start();

View File

@@ -0,0 +1,4 @@
// source code
import { mysteryBox } from "macro:./mystery-box";
export default "You roll! " + mysteryBox(123);

View File

@@ -0,0 +1,54 @@
import Pappa from "papaparse";
// Example usage:
// const rows = fetchCSV(
// "https://covid19.who.int/WHO-COVID-19-global-data.csv",
// {
// last: 100,
// columns: ["New_cases", "Date_reported", "Country"],
// }
// );
export async function fetchCSV(callExpression) {
console.time("fetchCSV Total");
const [
urlNode,
{
properties: { last: limit = 10, columns = [] },
},
] = callExpression.arguments;
const url = urlNode.get();
console.time("Fetch");
const response = await fetch(url);
const csvText = await response.text();
console.timeEnd("Fetch");
console.time("Parse");
let rows = Pappa.parse(csvText, { fastMode: true }).data;
console.timeEnd("Parse");
console.time("Render");
const columnIndices = new Array(columns.length);
for (let i = 0; i < columns.length; i++) {
columnIndices[i] = rows[0].indexOf(columns[i]);
}
rows = rows
.slice(Math.max(limit, rows.length) - limit)
.reverse()
.filter(columns => columns.every(Boolean));
const value = (
<array>
{rows.map(columns => (
<array>
{columnIndices.map(columnIndex => (
<string value={columns[columnIndex]} />
))}
</array>
))}
</array>
);
console.timeEnd("Render");
console.timeEnd("fetchCSV Total");
return value;
}

View File

@@ -0,0 +1,23 @@
// macro code
export async function matchInFile(callExpression: BunAST.CallExpression) {
const [filePathNode, matcherNode] = callExpression.arguments;
let filePath: string;
filePath = filePathNode.get();
let matcher: RegExp;
matcher = matcherNode.get();
const file: string = await Bun.file(Bun.cwd + filePath).text();
return (
<array>
{file
.split("\n")
.map(line => line.match(matcher))
.filter(Boolean)
.reverse()
.map(line => (
<string value={line[0]} />
))}
</array>
);
}

View File

@@ -0,0 +1,13 @@
export function mysteryBox(callExpression) {
console.log(callExpression.log);
// get arguments
const [countNode] = callExpression.arguments;
const countString: string = countNode.get();
const count: number = parseInt(countString, 10);
// validate
if (!(count >= 1 && count <= 1000)) return new Error(`Argument ${countString} is expected to be between 1 and 1000`);
// return a value
return (Math.random() * count) | 0;
}

10
examples/macros/now.tsx Normal file
View File

@@ -0,0 +1,10 @@
import moment from "moment";
export function now(node) {
var fmt = "HH:mm:ss";
const args = node.arguments;
if (args[0] instanceof <string />) {
fmt = args[0].get();
}
const time = moment().format(fmt);
return <string value={time}></string>;
}

View File

@@ -0,0 +1,17 @@
{
"name": "macros",
"version": "1.0.0",
"main": "index.js",
"license": "MIT",
"dependencies": {
"moment": "^2.29.1",
"papaparse": "^5.3.1",
"react": "^17.0.2",
"react-dom": "^17.0.2",
"react-refresh": "^0.10.0"
},
"devDependencies": {
"@types/react": "^17.0.24",
"@types/react-dom": "^17.0.9"
}
}

View File

@@ -0,0 +1,14 @@
<!DOCTYPE html>
<html>
<head>
<title>Macro test</title>
<meta charset="utf-8" />
<meta name="viewport" content="width=device-width, initial-scale=1.0" />
<link rel="stylesheet" href="/styles.css" type="text/css" />
</head>
<body>
<script async type="module" src="/components/index.tsx"></script>
</body>
</html>

View File

@@ -0,0 +1,47 @@
html {
font-size: 4rem;
margin: 0;
padding: 0;
background-color: black;
color: rgb(0, 255, 0);
font-family: "Courier";
}
body {
margin: 48px auto;
text-align: center;
}
.Line {
font-size: 0.5rem;
font-family: monospace;
}
.Table {
display: grid;
width: fit-content;
}
.Row,
.Header {
display: grid;
grid-template-columns: 2fr 1fr 1fr;
text-align: right;
column-gap: 2rem;
}
.Heading {
text-align: right;
}
.Header {
border-bottom: 1px solid rgb(0, 255, 0);
margin-bottom: 20px;
padding-bottom: 20px;
}
.Heading:nth-of-type(2) {
text-align: left;
}

View File

@@ -0,0 +1,7 @@
{
"compilerOptions": {
"baseUrl": ".",
"paths": {},
"jsx": "preserve"
}
}

11
examples/mmap/1.js Normal file
View File

@@ -0,0 +1,11 @@
const map = Bun.mmap("./mmap.txt", { shared: true });
const utf8decoder = new TextDecoder("utf-8");
let old = new TextEncoder().encode("12345");
setInterval(() => {
old = old.sort((a, b) => (Math.random() > 0.5 ? -1 : 1));
console.log(`changing mmap to ~> ${utf8decoder.decode(old)}`);
map.set(old);
}, 4);

22
examples/mmap/2.js Normal file
View File

@@ -0,0 +1,22 @@
const map = Bun.mmap("./mmap.txt");
function buffer_hash(buffer) {
let hash = 0;
for (let i = 0; i < buffer.length; i++) {
hash = (hash << 5) - hash + buffer[i];
hash |= 0; // Convert to 32bit integer
}
return hash;
}
const decoder = new TextDecoder();
let hash = buffer_hash(map);
console.log(decoder.decode(map));
while (true) {
if (buffer_hash(map) !== hash) {
hash = buffer_hash(map);
console.log(`mmap changed to ~> ${decoder.decode(map)}`);
}
}

1
examples/mmap/mmap.txt Normal file
View File

@@ -0,0 +1 @@
43521

23
examples/openInEditor.js Normal file
View File

@@ -0,0 +1,23 @@
import { resolve } from "path";
import { parse } from "querystring";
export default {
fetch(req) {
const url = new URL(req.url);
if (url.pathname === "/favicon.ico") return new Response("nooo dont open favicon in editor", { status: 404 });
var pathname = req.url.substring(1);
const q = pathname.indexOf("?");
var { editor } = parse(pathname.substring(q + 1)) || {};
if (q > 0) {
pathname = pathname.substring(0, q);
}
Bun.openInEditor(resolve(pathname), {
editor,
});
return new Response(`Opened ${req.url}`);
},
};

File diff suppressed because it is too large Load Diff

View File

@@ -0,0 +1,33 @@
{
"name": "simple-react",
"version": "1.0.0",
"license": "MIT",
"dependencies": {
"@emotion/css": "^11.1.3",
"@vitejs/plugin-react-refresh": "^1.3.3",
"antd": "^4.16.1",
"left-pad": "^1.3.0",
"next": "^11.0.0",
"parcel": "2.0.0-beta.3",
"react": "^17.0.2",
"react-bootstrap": "^1.6.1",
"react-dom": "^17.0.2",
"react-form": "^4.0.1",
"react-hook-form": "^7.8.3"
},
"parcel": "parceldist/index.js",
"targets": {
"parcel": {
"outputFormat": "esmodule",
"sourceMap": false,
"optimize": false,
"engines": {
"chrome": "last 1 version"
}
}
},
"devDependencies": {
"@snowpack/plugin-react-refresh": "^2.5.0",
"typescript": "^4.3.4"
}
}

View File

@@ -0,0 +1,15 @@
<!DOCTYPE html>
<html>
<head>
<link
rel="stylesheet"
crossorigin="anonymous"
href="https://fonts.googleapis.com/css2?family=IBM+Plex+Sans:wght@400;700&family=Space+Mono:wght@400;700&display=swap"
/>
</head>
<body>
<div id="reactroot"></div>
<link rel="stylesheet" href="./src/index.css" />
<script src="./src/index.tsx" async type="module"></script>
</body>
</html>

File diff suppressed because it is too large Load Diff

View File

@@ -0,0 +1,14 @@
:root {
--timestamp: "0";
--interval: "8";
--progress-bar: 11.83299999999997%;
--spinner-1-muted: rgb(142, 6, 182);
--spinner-1-primary: rgb(177, 8, 227);
--spinner-2-muted: rgb(110, 148, 190);
--spinner-2-primary: rgb(138, 185, 238);
--spinner-3-muted: rgb(75, 45, 64);
--spinner-3-primary: rgb(94, 56, 80);
--spinner-4-muted: rgb(155, 129, 108);
--spinner-4-primary: rgb(194, 161, 135);
--spinner-rotate: 213deg;
}

View File

@@ -0,0 +1,21 @@
import React from "react";
export function RenderCounter({ name, children }) {
const counter = React.useRef(1);
return (
<div className="RenderCounter">
<div className="RenderCounter-meta">
<div className="RenderCounter-title">
{name} rendered <strong>{counter.current++} times</strong>
</div>
<div className="RenderCounter-lastRender">
LAST RENDER:{" "}
{new Intl.DateTimeFormat([], {
timeStyle: "long",
}).format(new Date())}
</div>
</div>
<div className="RenderCounter-children">{children}</div>
</div>
);
}

View File

@@ -0,0 +1,14 @@
import * as React from "react";
import { Button } from "./Button";
import { RenderCounter } from "./RenderCounter";
export function App() {
return (
<RenderCounter name="App">
<div className="AppRoot">
<h1>This is the root element</h1>
<Button>Click</Button>
</div>
</RenderCounter>
);
}

View File

@@ -0,0 +1,9 @@
import { RenderCounter } from "./RenderCounter";
export const Button = ({ children }) => {
return (
<RenderCounter name="Button">
<div className="Button">{children}</div>
</RenderCounter>
);
};

View File

@@ -0,0 +1 @@
@import "https://fonts.googleapis.com/css2?family=IBM+Plex+Sans:wght@400;700&family=Space+Mono:wght@400;700&display=swap";

View File

@@ -0,0 +1,98 @@
@import "./colors.css";
:root {
--heading-font: "Space Mono", system-ui;
--body-font: "IBM Plex Sans", system-ui;
--color-brand: #02ff00;
--color-brand-muted: rgb(2, 150, 0);
--padding-horizontal: 90px;
--page-background: black;
--page-background-alpha: rgba(0, 0, 0, 0.8);
--result__background-color: black;
--result__primary-color: var(--color-brand);
--result__foreground-color: white;
--result__muted-color: rgb(165, 165, 165);
--card-width: 352px;
--page-width: 1152px;
--snippets_container-background-unfocused: #171717;
--snippets_container-background-focused: #0017e9;
--snippets_container-background: var(
--snippets_container-background-unfocused
);
--snippets_container-muted-color: rgb(153, 153, 153);
}
body {
color: white;
margin: 0;
padding: 0;
font-family: var(--body-font);
background-color: var(--page-background);
color: var(--result__muted-color);
display: flex;
flex-direction: column;
height: 100%;
}
#reactroot,
#__next,
body,
html {
height: 100%;
}
.RenderCounter {
border: 10px solid var(--snippets_container-background-focused);
margin: 10px;
padding: 10px;
animation: flash 0.2s linear;
animation-fill-mode: forwards;
}
.RenderCounter-meta {
display: flex;
flex-direction: row;
justify-content: space-between;
margin: -10px;
padding: 10px;
background-color: #111;
}
.RenderCounter-lastRender,
.RenderCounter-title {
white-space: nowrap;
color: rgb(153, 153, 153);
}
@keyframes flash {
from {
border-color: var(--snippets_container-background-focused);
}
to {
border-color: var(--snippets_container-background-unfocused);
}
}
.Button {
display: block;
border: 1px solid rgb(20, 180, 0);
background-color: rgb(2, 150, 0);
color: white;
font-weight: 500;
padding: 10px 12px;
border-radius: 4px;
text-transform: uppercase;
text-align: center;
width: fit-content;
cursor: pointer;
}

View File

@@ -0,0 +1,15 @@
import ReactDOM from "react-dom";
import React from "react";
import { App } from "./components/app";
import classNames from "classnames";
function startReact() {
ReactDOM.render(<App />, document.querySelector("#reactroot"));
}
globalThis.addEventListener("DOMContentLoaded", () => {
startReact();
});
startReact();
export { App };

View File

@@ -0,0 +1,62 @@
import React from "react";
export const Main = ({ productName }) => {
return (
<>
<header>
<div className="Title">CSS HMR Stress Test</div>
<p className="Description">
This page visually tests how quickly a bundler can update CSS over Hot Module Reloading.
</p>
</header>
<main className="main">
<section className="ProgressSection">
<p className="Subtitle">
<span className="Subtitle-part">
Ran:&nbsp;<span className="timer"></span>
</span>
</p>
<div className="ProgressBar-container">
<div className="ProgressBar"></div>
</div>
<div className="SectionLabel">The progress bar should move from left to right smoothly.</div>
</section>
<section>
<div className="Spinners">
<div className="Spinner-container Spinner-1">
<div className="Spinner"></div>
</div>
<div className="Spinner-container Spinner-2">
<div className="Spinner"></div>
</div>
<div className="Spinner-container Spinner-3">
<div className="Spinner"></div>
</div>
<div className="Spinner-container Spinner-4">
<div className="Spinner"></div>
</div>
</div>
<div className="SectionLabel">The spinners should rotate &amp; change color smoothly.</div>
</section>
</main>
<footer>
<div className="SectionLabel FooterLabel">There are no CSS animations on this page.</div>
<div className="Bundler-container">
<div className="Bundler">{productName}</div>
<div className="Bundler-updateRate">
Saving a css file every&nbsp;
<span className="highlight">
<span className="interval"></span>ms
</span>
</div>
</div>
</footer>
</>
);
};

View File

@@ -0,0 +1,29 @@
{
"compilerOptions": {
"target": "esnext",
"lib": [
"dom",
"dom.iterable",
"esnext"
],
"allowJs": true,
"skipLibCheck": true,
"strict": false,
"forceConsistentCasingInFileNames": true,
"noEmit": true,
"esModuleInterop": true,
"module": "esnext",
"moduleResolution": "node",
"resolveJsonModule": true,
"isolatedModules": true,
"jsx": "preserve"
},
"include": [
"next-env.d.ts",
"**/*.ts",
"**/*.tsx"
],
"exclude": [
"node_modules"
]
}

Binary file not shown.

View File

@@ -0,0 +1,20 @@
// A simple way to connect FileSystemRouter to Bun#serve
// run with `bun run index.tsx`
import { renderToReadableStream } from "react-dom/server";
import { FileSystemRouter } from "bun";
export default {
port: 3000,
async fetch(request: Request) {
const router = new FileSystemRouter({
dir: process.cwd() + "/pages",
style: "nextjs",
});
const route = router.match(request);
const { default: Root } = await import(route.filePath!);
return new Response(await renderToReadableStream(<Root {...route.params} />));
},
};

View File

@@ -0,0 +1,14 @@
{
"name": "react-routes",
"module": "index.tsx",
"type": "module",
"devDependencies": {
"@types/react": "^18.0.27",
"@types/react-dom": "^18.0.10",
"bun-types": "^0.7.0"
},
"dependencies": {
"react": "^18.2.0",
"react-dom": "^18.2.0"
}
}

View File

@@ -0,0 +1,17 @@
// reachable from http://localhost:3000/
export default () => (
<html>
<head>
<title>index</title>
</head>
<body>
<h1>
<a href="/one">one</a>
</h1>
<h1>
<a href="/two">two</a>
</h1>
</body>
</html>
);

View File

@@ -0,0 +1,12 @@
// reachable from http://localhost:3000/one
export default () => (
<html>
<head>
<title>one</title>
</head>
<body>
<p>one</p>
</body>
</html>
);

View File

@@ -0,0 +1,12 @@
// reachable from http://localhost:3000/two
export default () => (
<html>
<head>
<title>two</title>
</head>
<body>
<p>two</p>
</body>
</html>
);

View File

@@ -0,0 +1,20 @@
{
"compilerOptions": {
"lib": [
"ESNext"
],
"module": "esnext",
"target": "esnext",
"moduleResolution": "nodenext",
"strict": false,
"downlevelIteration": true,
"skipLibCheck": true,
"jsx": "preserve",
"allowSyntheticDefaultImports": true,
"forceConsistentCasingInFileNames": true,
"allowJs": true,
"types": [
"bun-types" // add Bun global
]
}
}

68
examples/sha.js Normal file
View File

@@ -0,0 +1,68 @@
import { SHA1, SHA256, SHA512, SHA384, SHA512_256, MD5, MD4, RIPEMD160, sha } from "bun";
const input = "Hello World";
const [first, second] = input.split(" ");
const log = (name, ...args) => console.log(`${name}:`.padStart("SHA512_256: ".length), ...args);
console.log("");
// This is SHA512-256:
// This function is shorthand for SHA512_256.hash(input)
log("Bun.sha()", sha(input, "base64"));
log("SHA1", SHA1.hash(input, "hex"));
log("SHA256", SHA256.hash(input, "hex"));
log("SHA384", SHA384.hash(input, "hex"));
log("SHA512", SHA512.hash(input, "hex"));
log("SHA512_256", SHA512_256.hash(input, "hex"));
log("RIPEMD160", RIPEMD160.hash(input, "hex"));
console.log("");
console.log("---- Chunked ----");
console.log("");
// You can also do updates in chunks:
// const hash = new Hash();
for (let Hash of [SHA1, SHA256, SHA384, SHA512, SHA512_256, RIPEMD160]) {
const hash = new Hash();
hash.update(first);
hash.update(" " + second);
log(Hash.name, hash.digest("hex"));
}
console.log("");
console.log("---- Base64 ----");
console.log("");
// base64 or hex
for (let Hash of [SHA1, SHA256, SHA384, SHA512, SHA512_256]) {
const hash = new Hash();
hash.update(first);
hash.update(" " + second);
log(Hash.name, hash.digest("base64"));
}
console.log("");
console.log("---- Uint8Array ----");
console.log("");
// Uint8Array by default
for (let Hash of [SHA1, SHA256, SHA384, SHA512, SHA512_256]) {
const hash = new Hash();
hash.update(first);
hash.update(" " + second);
log(Hash.name, hash.digest());
}
console.log("");
console.log("---- Uint8Array can be updated in-place ----");
console.log("");
var oneBuf = new Uint8Array(1024);
// Update Uint8Array in-place instead of allocating a new one
for (let Hash of [SHA1, SHA256, SHA384, SHA512, SHA512_256]) {
const hash = new Hash();
hash.update(first);
hash.update(" " + second);
log(Hash.name, hash.digest(oneBuf).subarray(0, Hash.byteLength));
}

48
examples/spawn.ts Normal file
View File

@@ -0,0 +1,48 @@
import { spawn, which } from "bun";
import { rmSync } from "fs";
import { basename } from "path";
const repo = process.argv.at(3) || "TheoBr/vercel-vite-demo";
const target = basename(repo) + "-main";
console.log("Downloading", repo, "to", "/tmp/" + target);
const archive = await fetch(`https://github.com/${repo}/archive/refs/heads/main.tar.gz`);
// remove the directory if it already exists locally
rmSync("/tmp/" + target, { recursive: true, force: true });
const tar = spawn({
cmd: ["tar", "-xzf", "-"],
stdin: archive.body,
stderr: "inherit",
stdout: "inherit",
cwd: "/tmp",
});
await tar.exited;
// if vercel isn't installed, install it
if (!which("vercel")) {
console.log("Installing vercel...");
const installer = spawn(["bun", "install", "-g", "vercel"], {
stderr: "inherit",
stdout: "inherit",
stdin: "inherit",
});
await installer.exited;
if (!which("vercel")) {
throw new Error("Failed to install Vercel CLI");
}
}
const { exited: deployed } = spawn({
cmd: ["vercel", "deploy", "--yes", "--public", target],
stdio: ["inherit", "inherit", "inherit"],
cwd: "/tmp",
});
await deployed;

15
examples/ssl.ts Normal file
View File

@@ -0,0 +1,15 @@
import { resolve } from "path";
import type { ServeOptions } from "bun";
const development = process.env.NODE_ENV !== "production";
export default {
fetch(req: Request) {
return new Response(Bun.file(resolve(req.url.substring(1))));
},
// hostname: "0.0.0.0",
port: process.env.PORT || "443",
keyFile: process.env.SSL_KEY_FILE || "./key.pem",
certFile: process.env.SSL_CERTIFICATE_FILE || "./cert.pem",
development,
} as ServeOptions;

49
examples/tcp.ts Normal file
View File

@@ -0,0 +1,49 @@
import { listen, connect } from "bun";
var counter = 0;
const msg = Buffer.from("Hello World!");
const handlers = {
open(socket) {
if (!socket.data?.isServer) {
if (!socket.write(msg)) {
socket.data = { pending: msg };
}
}
},
data(socket, buffer) {
if (!socket.write(buffer)) {
socket.data = { pending: buffer };
return;
}
counter++;
},
drain(socket) {
const pending = socket.data?.pending;
if (!pending) return;
if (socket.write(pending)) {
socket.data = undefined;
counter++;
return;
}
},
};
setInterval(() => {
console.log("Wrote", counter, "messages");
counter = 0;
}, 1000);
const server = listen({
socket: handlers,
hostname: "localhost",
port: 8080,
data: {
isServer: true,
},
});
const connection = await connect({
socket: handlers,
hostname: "localhost",
port: 8080,
});

View File

@@ -1,724 +0,0 @@
#!/usr/bin/env bun
/**
* CLI Flag Parser for Bun Commands
*
* This script reads the --help menu for every Bun command and generates JSON
* containing all flag information, descriptions, and whether they support
* positional or non-positional arguments.
*
* Handles complex cases like:
* - Nested subcommands (bun pm cache rm)
* - Command aliases (bun i = bun install, bun a = bun add)
* - Dynamic completions (scripts, packages, files)
* - Context-aware flags
* - Special cases like bare 'bun' vs 'bun run'
*
* Output is saved to completions/bun-cli.json for use in generating
* shell completions (fish, bash, zsh).
*/
import { spawn } from "bun";
import { mkdirSync, writeFileSync, mkdtempSync, rmSync } from "fs";
import { join } from "path";
interface FlagInfo {
name: string;
shortName?: string;
description: string;
hasValue: boolean;
valueType?: string;
defaultValue?: string;
choices?: string[];
required?: boolean;
multiple?: boolean;
}
interface SubcommandInfo {
name: string;
description: string;
flags?: FlagInfo[];
subcommands?: Record<string, SubcommandInfo>;
positionalArgs?: {
name: string;
description?: string;
required: boolean;
multiple: boolean;
type?: string;
completionType?: string;
}[];
examples?: string[];
}
interface CommandInfo {
name: string;
aliases?: string[];
description: string;
usage?: string;
flags: FlagInfo[];
positionalArgs: {
name: string;
description?: string;
required: boolean;
multiple: boolean;
type?: string;
completionType?: string;
}[];
examples: string[];
subcommands?: Record<string, SubcommandInfo>;
documentationUrl?: string;
dynamicCompletions?: {
scripts?: boolean;
packages?: boolean;
files?: boolean;
binaries?: boolean;
};
}
interface CompletionData {
version: string;
commands: Record<string, CommandInfo>;
globalFlags: FlagInfo[];
specialHandling: {
bareCommand: {
description: string;
canRunFiles: boolean;
dynamicCompletions: {
scripts: boolean;
files: boolean;
binaries: boolean;
};
};
};
bunGetCompletes: {
available: boolean;
commands: {
scripts: string; // "bun getcompletes s" or "bun getcompletes z"
binaries: string; // "bun getcompletes b"
packages: string; // "bun getcompletes a <prefix>"
files: string; // "bun getcompletes j"
};
};
}
const BUN_EXECUTABLE = process.env.BUN_DEBUG_BUILD || "bun";
/**
* Parse flag line from help output
*/
function parseFlag(line: string): FlagInfo | null {
// Match patterns like:
// -h, --help Display this menu and exit
// --timeout=<val> Set the per-test timeout in milliseconds, default is 5000.
// -r, --preload=<val> Import a module before other modules are loaded
// --watch Automatically restart the process on file change
const patterns = [
// Long flag with short flag and value: -r, --preload=<val>
/^\s*(-[a-zA-Z]),\s+(--[a-zA-Z-]+)=(<[^>]+>)\s+(.+)$/,
// Long flag with short flag: -h, --help
/^\s*(-[a-zA-Z]),\s+(--[a-zA-Z-]+)\s+(.+)$/,
// Long flag with value: --timeout=<val>
/^\s+(--[a-zA-Z-]+)=(<[^>]+>)\s+(.+)$/,
// Long flag without value: --watch
/^\s+(--[a-zA-Z-]+)\s+(.+)$/,
// Short flag only: -i
/^\s+(-[a-zA-Z])\s+(.+)$/,
];
for (const pattern of patterns) {
const match = line.match(pattern);
if (match) {
let shortName: string | undefined;
let longName: string;
let valueSpec: string | undefined;
let description: string;
if (match.length === 5) {
// Pattern with short flag, long flag, and value
[, shortName, longName, valueSpec, description] = match;
} else if (match.length === 4) {
if (match[1].startsWith("-") && match[1].length === 2) {
// Short flag with long flag
[, shortName, longName, description] = match;
} else if (match[2].startsWith("<")) {
// Long flag with value
[, longName, valueSpec, description] = match;
} else {
// Long flag without value
[, longName, description] = match;
}
} else if (match.length === 3) {
if (match[1].length === 2) {
// Short flag only
[, shortName, description] = match;
longName = shortName.replace("-", "--");
} else {
// Long flag without value
[, longName, description] = match;
}
} else {
continue;
}
// Extract additional info from description
const hasValue = !!valueSpec;
let valueType: string | undefined;
let defaultValue: string | undefined;
let choices: string[] | undefined;
if (valueSpec) {
valueType = valueSpec.replace(/[<>]/g, "");
}
// Look for default values in description
const defaultMatch = description.match(/[Dd]efault(?:s?)\s*(?:is|to|:)\s*"?([^".\s,]+)"?/);
if (defaultMatch) {
defaultValue = defaultMatch[1];
}
// Look for choices/enums
const choicesMatch = description.match(/(?:One of|Valid (?:orders?|values?|options?)):?\s*"?([^"]+)"?/);
if (choicesMatch) {
choices = choicesMatch[1]
.split(/[,\s]+/)
.map(s => s.replace(/[",]/g, "").trim())
.filter(Boolean);
}
return {
name: longName.replace(/^--/, ""),
shortName: shortName?.replace(/^-/, ""),
description: description.trim(),
hasValue,
valueType,
defaultValue,
choices,
required: false, // We'll determine this from usage patterns
multiple: description.toLowerCase().includes("multiple") || description.includes("[]"),
};
}
}
return null;
}
/**
* Parse usage line to extract positional arguments
*/
function parseUsage(usage: string): {
name: string;
description?: string;
required: boolean;
multiple: boolean;
type?: string;
completionType?: string;
}[] {
const args: {
name: string;
description?: string;
required: boolean;
multiple: boolean;
type?: string;
completionType?: string;
}[] = [];
// Extract parts after command name
const parts = usage.split(/\s+/).slice(2); // Skip "Usage:" and command name
for (const part of parts) {
if (part.startsWith("[") || part.startsWith("<") || part.includes("...")) {
let name = part;
let required = false;
let multiple = false;
let completionType: string | undefined;
// Clean up the argument name
name = name.replace(/[\[\]<>]/g, "");
if (part.startsWith("<")) {
required = true;
}
if (part.includes("...") || name.includes("...")) {
multiple = true;
name = name.replace(/\.{3}/g, "");
}
// Skip flags
if (!name.startsWith("-") && name.length > 0) {
// Determine completion type based on argument name
if (name.toLowerCase().includes("package")) {
completionType = "package";
} else if (name.toLowerCase().includes("script")) {
completionType = "script";
} else if (name.toLowerCase().includes("file") || name.includes(".")) {
completionType = "file";
}
args.push({
name,
required,
multiple,
type: "string", // Default type
completionType,
});
}
}
}
return args;
}
const temppackagejson = mkdtempSync("package");
writeFileSync(
join(temppackagejson, "package.json"),
JSON.stringify({
name: "test",
version: "1.0.0",
scripts: {},
}),
);
process.once("beforeExit", () => {
rmSync(temppackagejson, { recursive: true });
});
/**
* Execute bun command and get help output
*/
async function getHelpOutput(command: string[]): Promise<string> {
try {
const proc = spawn({
cmd: [BUN_EXECUTABLE, ...command, "--help"],
stdout: "pipe",
stderr: "pipe",
cwd: temppackagejson,
});
const [stdout, stderr] = await Promise.all([new Response(proc.stdout).text(), new Response(proc.stderr).text()]);
await proc.exited;
return stdout || stderr || "";
} catch (error) {
console.error(`Failed to get help for command: ${command.join(" ")}`, error);
return "";
}
}
/**
* Parse PM subcommands from help output
*/
function parsePmSubcommands(helpText: string): Record<string, SubcommandInfo> {
const lines = helpText.split("\n");
const subcommands: Record<string, SubcommandInfo> = {};
let inCommands = false;
for (const line of lines) {
const trimmed = line.trim();
if (trimmed === "Commands:") {
inCommands = true;
continue;
}
if (inCommands && trimmed.startsWith("Learn more")) {
break;
}
if (inCommands && line.match(/^\s+bun pm \w+/)) {
// Parse lines like: "bun pm pack create a tarball of the current workspace"
const match = line.match(/^\s+bun pm (\S+)(?:\s+(.+))?$/);
if (match) {
const [, name, description = ""] = match;
subcommands[name] = {
name,
description: description.trim(),
flags: [],
positionalArgs: [],
};
// Special handling for subcommands with their own subcommands
if (name === "cache") {
subcommands[name].subcommands = {
rm: {
name: "rm",
description: "clear the cache",
},
};
} else if (name === "pkg") {
subcommands[name].subcommands = {
get: { name: "get", description: "get values from package.json" },
set: { name: "set", description: "set values in package.json" },
delete: { name: "delete", description: "delete keys from package.json" },
fix: { name: "fix", description: "auto-correct common package.json errors" },
};
}
}
}
}
return subcommands;
}
/**
* Parse help output into CommandInfo
*/
function parseHelpOutput(helpText: string, commandName: string): CommandInfo {
const lines = helpText.split("\n");
const command: CommandInfo = {
name: commandName,
description: "",
flags: [],
positionalArgs: [],
examples: [],
};
let currentSection = "";
let inFlags = false;
let inExamples = false;
for (let i = 0; i < lines.length; i++) {
const line = lines[i];
const trimmed = line.trim();
// Extract command description (usually the first non-usage line)
if (
!command.description &&
trimmed &&
!trimmed.startsWith("Usage:") &&
!trimmed.startsWith("Alias:") &&
currentSection === ""
) {
command.description = trimmed;
continue;
}
// Extract aliases
if (trimmed.startsWith("Alias:")) {
const aliasMatch = trimmed.match(/Alias:\s*(.+)/);
if (aliasMatch) {
command.aliases = aliasMatch[1]
.split(/[,\s]+/)
.map(a => a.trim())
.filter(Boolean);
}
continue;
}
// Extract usage and positional args
if (trimmed.startsWith("Usage:")) {
command.usage = trimmed;
command.positionalArgs = parseUsage(trimmed);
continue;
}
// Track sections
if (trimmed === "Flags:") {
inFlags = true;
currentSection = "flags";
continue;
} else if (trimmed === "Examples:") {
inExamples = true;
inFlags = false;
currentSection = "examples";
continue;
} else if (
trimmed.startsWith("Full documentation") ||
trimmed.startsWith("Learn more") ||
trimmed.startsWith("A full list")
) {
const urlMatch = trimmed.match(/https?:\/\/[^\s]+/);
if (urlMatch) {
command.documentationUrl = urlMatch[0];
}
inFlags = false;
inExamples = false;
continue;
}
// Parse flags
if (inFlags && line.match(/^\s+(-|\s+--)/)) {
const flag = parseFlag(line);
if (flag) {
command.flags.push(flag);
}
}
// Parse examples
if (inExamples && trimmed && !trimmed.startsWith("Full documentation")) {
if (trimmed.startsWith("bun ") || trimmed.startsWith("./") || trimmed.startsWith("Bundle")) {
command.examples.push(trimmed);
}
}
}
// Special case for pm command
if (commandName === "pm") {
command.subcommands = parsePmSubcommands(helpText);
}
// Add dynamic completion info based on command
command.dynamicCompletions = {};
if (commandName === "run") {
command.dynamicCompletions.scripts = true;
command.dynamicCompletions.files = true;
command.dynamicCompletions.binaries = true;
// Also add file type info for positional args
for (const arg of command.positionalArgs) {
if (arg.name.includes("file") || arg.name.includes("script")) {
arg.completionType = "javascript_files";
}
}
} else if (commandName === "add") {
command.dynamicCompletions.packages = true;
// Mark package args
for (const arg of command.positionalArgs) {
if (arg.name.includes("package") || arg.name === "name") {
arg.completionType = "package";
}
}
} else if (commandName === "remove") {
command.dynamicCompletions.packages = true; // installed packages
for (const arg of command.positionalArgs) {
if (arg.name.includes("package") || arg.name === "name") {
arg.completionType = "installed_package";
}
}
} else if (["test"].includes(commandName)) {
command.dynamicCompletions.files = true;
for (const arg of command.positionalArgs) {
if (arg.name.includes("pattern") || arg.name.includes("file")) {
arg.completionType = "test_files";
}
}
} else if (["build"].includes(commandName)) {
command.dynamicCompletions.files = true;
for (const arg of command.positionalArgs) {
if (arg.name === "entrypoint" || arg.name.includes("file")) {
arg.completionType = "javascript_files";
}
}
} else if (commandName === "create") {
// Create has special template completions
for (const arg of command.positionalArgs) {
if (arg.name.includes("template")) {
arg.completionType = "create_template";
}
}
}
return command;
}
/**
* Get list of main commands from bun --help
*/
async function getMainCommands(): Promise<string[]> {
const helpText = await getHelpOutput([]);
const lines = helpText.split("\n");
const commands: string[] = [];
let inCommands = false;
for (const line of lines) {
const trimmed = line.trim();
if (trimmed === "Commands:") {
inCommands = true;
continue;
}
// Stop when we hit the "Flags:" section
if (inCommands && trimmed === "Flags:") {
break;
}
if (inCommands && line.match(/^\s+\w+/)) {
// Extract command name (first word after whitespace)
const match = line.match(/^\s+(\w+)/);
if (match) {
commands.push(match[1]);
}
}
}
const commandsToRemove = ["lint"];
return commands.filter(a => {
if (commandsToRemove.includes(a)) {
return false;
}
return true;
});
}
/**
* Extract global flags from main help
*/
function parseGlobalFlags(helpText: string): FlagInfo[] {
const lines = helpText.split("\n");
const flags: FlagInfo[] = [];
let inFlags = false;
for (const line of lines) {
const trimmed = line.trim();
if (trimmed === "Flags:") {
inFlags = true;
continue;
}
if (inFlags && (trimmed === "" || trimmed.startsWith("("))) {
break;
}
if (inFlags && line.match(/^\s+(-|\s+--)/)) {
const flag = parseFlag(line);
if (flag) {
flags.push(flag);
}
}
}
return flags;
}
/**
* Add command aliases based on common patterns
*/
function addCommandAliases(commands: Record<string, CommandInfo>): void {
const aliasMap: Record<string, string[]> = {
"install": ["i"],
"add": ["a"],
"remove": ["rm"],
"create": ["c"],
"x": ["bunx"], // bunx is an alias for bun x
};
for (const [command, aliases] of Object.entries(aliasMap)) {
if (commands[command]) {
commands[command].aliases = aliases;
}
}
}
/**
* Main function to generate completion data
*/
async function generateCompletions(): Promise<void> {
console.log("🔍 Discovering Bun commands...");
// Get main help and extract commands
const mainHelpText = await getHelpOutput([]);
const mainCommands = await getMainCommands();
const globalFlags = parseGlobalFlags(mainHelpText);
console.log(`📋 Found ${mainCommands.length} main commands: ${mainCommands.join(", ")}`);
const completionData: CompletionData = {
version: "1.1.0",
commands: {},
globalFlags,
specialHandling: {
bareCommand: {
description: "Run JavaScript/TypeScript files directly or access package scripts and binaries",
canRunFiles: true,
dynamicCompletions: {
scripts: true,
files: true,
binaries: true,
},
},
},
bunGetCompletes: {
available: true,
commands: {
scripts: "bun getcompletes s", // or "bun getcompletes z" for scripts with descriptions
binaries: "bun getcompletes b",
packages: "bun getcompletes a", // takes prefix as argument
files: "bun getcompletes j", // JavaScript/TypeScript files
},
},
};
// Parse each command
for (const commandName of mainCommands) {
console.log(`📖 Parsing help for: ${commandName}`);
try {
const helpText = await getHelpOutput([commandName]);
if (helpText.trim()) {
const commandInfo = parseHelpOutput(helpText, commandName);
completionData.commands[commandName] = commandInfo;
}
} catch (error) {
console.error(`❌ Failed to parse ${commandName}:`, error);
}
}
// Add common aliases
addCommandAliases(completionData.commands);
// Also check some common subcommands that might have their own help
const additionalCommands = ["pm"];
for (const commandName of additionalCommands) {
if (!completionData.commands[commandName]) {
console.log(`📖 Parsing help for additional command: ${commandName}`);
try {
const helpText = await getHelpOutput([commandName]);
if (helpText.trim() && !helpText.includes("error:") && !helpText.includes("Error:")) {
const commandInfo = parseHelpOutput(helpText, commandName);
completionData.commands[commandName] = commandInfo;
}
} catch (error) {
console.error(`❌ Failed to parse ${commandName}:`, error);
}
}
}
// Ensure completions directory exists
const completionsDir = join(process.cwd(), "completions");
try {
mkdirSync(completionsDir, { recursive: true });
} catch (error) {
// Directory might already exist
}
// Write the JSON file
const outputPath = join(completionsDir, "bun-cli.json");
const jsonData = JSON.stringify(completionData, null, 2);
writeFileSync(outputPath, jsonData, "utf8");
console.log(`✅ Generated CLI completion data at: ${outputPath}`);
console.log(`📊 Statistics:`);
console.log(` - Commands: ${Object.keys(completionData.commands).length}`);
console.log(` - Global flags: ${completionData.globalFlags.length}`);
let totalFlags = 0;
let totalExamples = 0;
let totalSubcommands = 0;
for (const [name, cmd] of Object.entries(completionData.commands)) {
totalFlags += cmd.flags.length;
totalExamples += cmd.examples.length;
const subcommandCount = cmd.subcommands ? Object.keys(cmd.subcommands).length : 0;
totalSubcommands += subcommandCount;
const aliasInfo = cmd.aliases ? ` (aliases: ${cmd.aliases.join(", ")})` : "";
const subcommandInfo = subcommandCount > 0 ? `, ${subcommandCount} subcommands` : "";
const dynamicInfo = cmd.dynamicCompletions ? ` [dynamic: ${Object.keys(cmd.dynamicCompletions).join(", ")}]` : "";
console.log(
` - ${name}${aliasInfo}: ${cmd.flags.length} flags, ${cmd.positionalArgs.length} positional args, ${cmd.examples.length} examples${subcommandInfo}${dynamicInfo}`,
);
}
console.log(` - Total command flags: ${totalFlags}`);
console.log(` - Total examples: ${totalExamples}`);
console.log(` - Total subcommands: ${totalSubcommands}`);
}
// Run the script
if (import.meta.main) {
generateCompletions().catch(console.error);
}

View File

@@ -1,121 +0,0 @@
# LLDB Pretty Printers for Bun
This directory contains LLDB pretty printers for various Bun data structures to improve the debugging experience.
## Files
- `bun_pretty_printer.py` - Pretty printers for Bun-specific types (bun.String, WTFStringImpl, ZigString, BabyList, etc.)
- `lldb_pretty_printers.py` - Pretty printers for Zig language types from the Zig project
- `lldb_webkit.py` - Pretty printers for WebKit/JavaScriptCore types
- `init.lldb` - LLDB initialization commands
## Supported Types
### bun.String Types
- `bun.String` (or just `String`) - The main Bun string type
- `WTFStringImpl` - WebKit string implementation (Latin1/UTF16)
- `ZigString` - Zig string type (UTF8/Latin1/UTF16 with pointer tagging)
### Display Format
The pretty printers show string content directly, with additional metadata:
```
# bun.String examples:
"Hello, World!" [latin1] # Regular ZigString
"UTF-8 String 🎉" [utf8] # UTF-8 encoded
"Static content" [latin1 static] # Static string
"" # Empty string
<dead> # Dead/invalid string
# WTFStringImpl examples:
"WebKit String" # Shows the actual string content
# ZigString examples:
"Some text" [utf16 global] # UTF16 globally allocated
"ASCII text" [latin1] # Latin1 encoded
```
## Usage
### Option 1: Manual Loading
In your LLDB session:
```lldb
command script import /path/to/bun/misctools/lldb/bun_pretty_printer.py
```
### Option 2: Add to ~/.lldbinit
Add the following line to your `~/.lldbinit` file to load automatically:
```lldb
command script import /path/to/bun/misctools/lldb/bun_pretty_printer.py
```
### Option 3: Use init.lldb
```lldb
command source /path/to/bun/misctools/lldb/init.lldb
```
## Testing
To test the pretty printers:
1. Build a debug version of Bun:
```bash
bun bd
```
2. Create a test file that uses bun.String types
3. Debug with LLDB:
```bash
lldb ./build/debug/bun-debug
(lldb) command script import misctools/lldb/bun_pretty_printer.py
(lldb) breakpoint set --file your_test.zig --line <line_number>
(lldb) run your_test.zig
(lldb) frame variable
```
## Implementation Details
### ZigString Pointer Tagging
ZigString uses pointer tagging in the upper bits:
- Bit 63: 1 = UTF16, 0 = UTF8/Latin1
- Bit 62: 1 = Globally allocated (mimalloc)
- Bit 61: 1 = UTF8 encoding
The pretty printer automatically detects and handles these tags.
### WTFStringImpl Encoding
WTFStringImpl uses flags in `m_hashAndFlags`:
- Bit 2 (s_hashFlag8BitBuffer): 1 = Latin1, 0 = UTF16
### bun.String Tag Union
bun.String is a tagged union with these variants:
- Dead (0): Invalid/freed string
- WTFStringImpl (1): WebKit string
- ZigString (2): Regular Zig string
- StaticZigString (3): Static/immortal string
- Empty (4): Empty string ""
## Troubleshooting
If the pretty printers don't work:
1. Verify the Python script loaded:
```lldb
(lldb) script print("Python works")
```
2. Check if the category is enabled:
```lldb
(lldb) type category list
```
3. Enable the Bun category manually:
```lldb
(lldb) type category enable bun
```
4. For debugging the pretty printer itself, check for exceptions:
- The pretty printers catch all exceptions and return `<error>`
- Modify the code to print exceptions for debugging

View File

@@ -10,8 +10,8 @@ class bun_BabyList_SynthProvider:
try:
self.ptr = self.value.GetChildMemberWithName('ptr')
self.len = self.value.GetChildMemberWithName('len').GetValueAsUnsigned()
self.cap = self.value.GetChildMemberWithName('cap').GetValueAsUnsigned()
self.len = self.value.GetChildMemberWithName('len').unsigned
self.cap = self.value.GetChildMemberWithName('cap').unsigned
self.elem_type = self.ptr.type.GetPointeeType()
self.elem_size = self.elem_type.size
except:
@@ -46,7 +46,7 @@ def bun_BabyList_SummaryProvider(value, _=None):
value = value.GetNonSyntheticValue()
len_val = value.GetChildMemberWithName('len')
cap_val = value.GetChildMemberWithName('cap')
return 'len=%d cap=%d' % (len_val.GetValueAsUnsigned(), cap_val.GetValueAsUnsigned())
return 'len=%d cap=%d' % (len_val.unsigned, cap_val.unsigned)
except:
return 'len=? cap=?'
@@ -67,241 +67,6 @@ def add(debugger, *, category, regex=False, type, identifier=None, synth=False,
type
))
def WTFStringImpl_SummaryProvider(value, _=None):
try:
# Get the raw pointer (it's already a pointer type)
value = value.GetNonSyntheticValue()
# Check if it's a pointer type and dereference if needed
if value.type.IsPointerType():
struct = value.deref
else:
struct = value
m_length = struct.GetChildMemberWithName('m_length').GetValueAsUnsigned()
m_hashAndFlags = struct.GetChildMemberWithName('m_hashAndFlags').GetValueAsUnsigned()
m_ptr = struct.GetChildMemberWithName('m_ptr')
# Check if it's 8-bit (latin1) or 16-bit (utf16) string
s_hashFlag8BitBuffer = 1 << 2
is_8bit = (m_hashAndFlags & s_hashFlag8BitBuffer) != 0
if m_length == 0:
return '[%s] ""' % ('latin1' if is_8bit else 'utf16')
# Limit memory reads to 1MB for performance
MAX_BYTES = 1024 * 1024 # 1MB
MAX_DISPLAY_CHARS = 200 # Maximum characters to display
# Calculate how much to read
bytes_per_char = 1 if is_8bit else 2
total_bytes = m_length * bytes_per_char
truncated = False
if total_bytes > MAX_BYTES:
# Read only first part of very large strings
chars_to_read = MAX_BYTES // bytes_per_char
bytes_to_read = chars_to_read * bytes_per_char
truncated = True
else:
chars_to_read = m_length
bytes_to_read = total_bytes
if is_8bit:
# Latin1 string
latin1_ptr = m_ptr.GetChildMemberWithName('latin1')
process = value.process
error = lldb.SBError()
ptr_addr = latin1_ptr.GetValueAsUnsigned()
if ptr_addr:
byte_data = process.ReadMemory(ptr_addr, min(chars_to_read, m_length), error)
if error.Success():
string_val = byte_data.decode('latin1', errors='replace')
else:
return '[latin1] <read error: %s>' % error
else:
return '[latin1] <null ptr>'
else:
# UTF16 string
utf16_ptr = m_ptr.GetChildMemberWithName('utf16')
process = value.process
error = lldb.SBError()
ptr_addr = utf16_ptr.GetValueAsUnsigned()
if ptr_addr:
byte_data = process.ReadMemory(ptr_addr, bytes_to_read, error)
if error.Success():
# Properly decode UTF16LE to string
string_val = byte_data.decode('utf-16le', errors='replace')
else:
return '[utf16] <read error: %s>' % error
else:
return '[utf16] <null ptr>'
# Escape special characters
string_val = string_val.replace('\\', '\\\\')
string_val = string_val.replace('"', '\\"')
string_val = string_val.replace('\n', '\\n')
string_val = string_val.replace('\r', '\\r')
string_val = string_val.replace('\t', '\\t')
# Truncate display if too long
display_truncated = truncated or len(string_val) > MAX_DISPLAY_CHARS
if len(string_val) > MAX_DISPLAY_CHARS:
string_val = string_val[:MAX_DISPLAY_CHARS]
# Add encoding and size info at the beginning
encoding = 'latin1' if is_8bit else 'utf16'
if display_truncated:
size_info = ' %d chars' % m_length
if total_bytes >= 1024 * 1024:
size_info += ' (%.1fMB)' % (total_bytes / (1024.0 * 1024.0))
elif total_bytes >= 1024:
size_info += ' (%.1fKB)' % (total_bytes / 1024.0)
return '[%s%s] "%s..." <truncated>' % (encoding, size_info, string_val)
else:
return '[%s] "%s"' % (encoding, string_val)
except:
return '<error>'
def ZigString_SummaryProvider(value, _=None):
try:
value = value.GetNonSyntheticValue()
ptr = value.GetChildMemberWithName('_unsafe_ptr_do_not_use').GetValueAsUnsigned()
length = value.GetChildMemberWithName('len').GetValueAsUnsigned()
# Check encoding flags
is_16bit = (ptr & (1 << 63)) != 0
is_utf8 = (ptr & (1 << 61)) != 0
is_global = (ptr & (1 << 62)) != 0
# Determine encoding
encoding = 'utf16' if is_16bit else ('utf8' if is_utf8 else 'latin1')
flags = ' global' if is_global else ''
if length == 0:
return '[%s%s] ""' % (encoding, flags)
# Untag the pointer (keep only the lower 53 bits)
untagged_ptr = ptr & ((1 << 53) - 1)
# Limit memory reads to 1MB for performance
MAX_BYTES = 1024 * 1024 # 1MB
MAX_DISPLAY_CHARS = 200 # Maximum characters to display
# Calculate how much to read
bytes_per_char = 2 if is_16bit else 1
total_bytes = length * bytes_per_char
truncated = False
if total_bytes > MAX_BYTES:
# Read only first part of very large strings
chars_to_read = MAX_BYTES // bytes_per_char
bytes_to_read = chars_to_read * bytes_per_char
truncated = True
else:
bytes_to_read = total_bytes
# Read the string data
process = value.process
error = lldb.SBError()
byte_data = process.ReadMemory(untagged_ptr, bytes_to_read, error)
if not error.Success():
return '[%s%s] <read error>' % (encoding, flags)
# Decode based on encoding
if is_16bit:
string_val = byte_data.decode('utf-16le', errors='replace')
elif is_utf8:
string_val = byte_data.decode('utf-8', errors='replace')
else:
string_val = byte_data.decode('latin1', errors='replace')
# Escape special characters
string_val = string_val.replace('\\', '\\\\')
string_val = string_val.replace('"', '\\"')
string_val = string_val.replace('\n', '\\n')
string_val = string_val.replace('\r', '\\r')
string_val = string_val.replace('\t', '\\t')
# Truncate display if too long
display_truncated = truncated or len(string_val) > MAX_DISPLAY_CHARS
if len(string_val) > MAX_DISPLAY_CHARS:
string_val = string_val[:MAX_DISPLAY_CHARS]
# Build the output
if display_truncated:
size_info = ' %d chars' % length
if total_bytes >= 1024 * 1024:
size_info += ' (%.1fMB)' % (total_bytes / (1024.0 * 1024.0))
elif total_bytes >= 1024:
size_info += ' (%.1fKB)' % (total_bytes / 1024.0)
return '[%s%s%s] "%s..." <truncated>' % (encoding, flags, size_info, string_val)
else:
return '[%s%s] "%s"' % (encoding, flags, string_val)
except:
return '<error>'
def bun_String_SummaryProvider(value, _=None):
try:
value = value.GetNonSyntheticValue()
# Debug: Show the actual type name LLDB sees
type_name = value.GetTypeName()
tag = value.GetChildMemberWithName('tag')
if not tag or not tag.IsValid():
# Try alternate field names
tag = value.GetChildMemberWithName('Tag')
if not tag or not tag.IsValid():
# Show type name to help debug
return '<no tag field in type: %s>' % type_name
tag_value = tag.GetValueAsUnsigned()
# Map tag values to names
tag_names = {
0: 'Dead',
1: 'WTFStringImpl',
2: 'ZigString',
3: 'StaticZigString',
4: 'Empty'
}
tag_name = tag_names.get(tag_value, 'Unknown')
if tag_name == 'Empty':
return '""'
elif tag_name == 'Dead':
return '<dead>'
elif tag_name == 'WTFStringImpl':
value_union = value.GetChildMemberWithName('value')
if not value_union or not value_union.IsValid():
return '<no value field>'
impl_value = value_union.GetChildMemberWithName('WTFStringImpl')
if not impl_value or not impl_value.IsValid():
return '<no WTFStringImpl field>'
return WTFStringImpl_SummaryProvider(impl_value, _)
elif tag_name == 'ZigString' or tag_name == 'StaticZigString':
value_union = value.GetChildMemberWithName('value')
if not value_union or not value_union.IsValid():
return '<no value field>'
field_name = 'ZigString' if tag_name == 'ZigString' else 'StaticZigString'
zig_string_value = value_union.GetChildMemberWithName(field_name)
if not zig_string_value or not zig_string_value.IsValid():
return '<no %s field>' % field_name
result = ZigString_SummaryProvider(zig_string_value, _)
# Add static marker if needed
if tag_name == 'StaticZigString':
result = result.replace(']', ' static]')
return result
else:
return '<unknown tag %d>' % tag_value
except Exception as e:
return '<error: %s>' % str(e)
def __lldb_init_module(debugger, _=None):
# Initialize Bun Category
debugger.HandleCommand('type category define --language c99 bun')
@@ -309,30 +74,5 @@ def __lldb_init_module(debugger, _=None):
# Initialize Bun Data Structures
add(debugger, category='bun', regex=True, type='^baby_list\\.BabyList\\(.*\\)$', identifier='bun_BabyList', synth=True, expand=True, summary=True)
# Add WTFStringImpl pretty printer - try multiple possible type names
add(debugger, category='bun', type='WTFStringImpl', identifier='WTFStringImpl', summary=True)
add(debugger, category='bun', type='*WTFStringImplStruct', identifier='WTFStringImpl', summary=True)
add(debugger, category='bun', type='string.WTFStringImpl', identifier='WTFStringImpl', summary=True)
add(debugger, category='bun', type='string.WTFStringImplStruct', identifier='WTFStringImpl', summary=True)
add(debugger, category='bun', type='*string.WTFStringImplStruct', identifier='WTFStringImpl', summary=True)
# Add ZigString pretty printer - try multiple possible type names
add(debugger, category='bun', type='ZigString', identifier='ZigString', summary=True)
add(debugger, category='bun', type='bun.js.bindings.ZigString', identifier='ZigString', summary=True)
add(debugger, category='bun', type='bindings.ZigString', identifier='ZigString', summary=True)
# Add bun.String pretty printer - try multiple possible type names
add(debugger, category='bun', type='String', identifier='bun_String', summary=True)
add(debugger, category='bun', type='bun.String', identifier='bun_String', summary=True)
add(debugger, category='bun', type='string.String', identifier='bun_String', summary=True)
add(debugger, category='bun', type='BunString', identifier='bun_String', summary=True)
add(debugger, category='bun', type='bun::String', identifier='bun_String', summary=True)
add(debugger, category='bun', type='bun::string::String', identifier='bun_String', summary=True)
# Try regex patterns for more flexible matching
add(debugger, category='bun', regex=True, type='.*String$', identifier='bun_String', summary=True)
add(debugger, category='bun', regex=True, type='.*WTFStringImpl.*', identifier='WTFStringImpl', summary=True)
add(debugger, category='bun', regex=True, type='.*ZigString.*', identifier='ZigString', summary=True)
# Enable the category
debugger.HandleCommand('type category enable bun')

View File

@@ -686,7 +686,7 @@ def add(debugger, *, category, regex=False, type, identifier=None, synth=False,
if summary: debugger.HandleCommand('type summary add --category %s%s%s "%s"' % (category, ' --inline-children' if inline_children else ''.join((' --expand' if expand else '', ' --python-function %s_SummaryProvider' % prefix if summary == True else ' --summary-string "%s"' % summary)), ' --regex' if regex else '', type))
if synth: debugger.HandleCommand('type synthetic add --category %s%s --python-class %s_SynthProvider "%s"' % (category, ' --regex' if regex else '', prefix, type))
def MultiArrayList_Entry(type): return 'multi_array_list\\.MultiArrayList\\(%s\\)\\.Entry__struct_[1-9][0-9]*$' % type
def MultiArrayList_Entry(type): return '^multi_array_list\\.MultiArrayList\\(%s\\)\\.Entry__struct_[1-9][0-9]*$' % type
def __lldb_init_module(debugger, _=None):
# Initialize Zig Categories
@@ -701,8 +701,8 @@ def __lldb_init_module(debugger, _=None):
# Initialize Zig Standard Library
add(debugger, category='zig.std', type='mem.Allocator', summary='${var.ptr}')
add(debugger, category='zig.std', regex=True, type='^segmented_list\\.SegmentedList\\(.*\\)$', identifier='std_SegmentedList', synth=True, expand=True, summary='len=${var.len}')
add(debugger, category='zig.std', regex=True, type='multi_array_list\\.MultiArrayList\\(.*\\)$', identifier='std_MultiArrayList', synth=True, expand=True, summary='len=${var.len} capacity=${var.capacity}')
add(debugger, category='zig.std', regex=True, type='multi_array_list\\.MultiArrayList\\(.*\\)\\.Slice$', identifier='std_MultiArrayList_Slice', synth=True, expand=True, summary='len=${var.len} capacity=${var.capacity}')
add(debugger, category='zig.std', regex=True, type='^multi_array_list\\.MultiArrayList\\(.*\\)$', identifier='std_MultiArrayList', synth=True, expand=True, summary='len=${var.len} capacity=${var.capacity}')
add(debugger, category='zig.std', regex=True, type='^multi_array_list\\.MultiArrayList\\(.*\\)\\.Slice$', identifier='std_MultiArrayList_Slice', synth=True, expand=True, summary='len=${var.len} capacity=${var.capacity}')
add(debugger, category='zig.std', regex=True, type=MultiArrayList_Entry('.*'), identifier='std_Entry', synth=True, inline_children=True, summary=True)
add(debugger, category='zig.std', regex=True, type='^hash_map\\.HashMapUnmanaged\\(.*\\)$', identifier='std_HashMapUnmanaged', synth=True, expand=True, summary=True)
add(debugger, category='zig.std', regex=True, type='^hash_map\\.HashMapUnmanaged\\(.*\\)\\.Entry$', identifier = 'std_Entry', synth=True, inline_children=True, summary=True)

Some files were not shown because too many files have changed in this diff Show More