Files
bun.sh/misctools/compression.zig
pfg 05d0475c6c Update to zig 0.15.2 (#24204)
Fixes ENG-21287

Build times, from `bun run build && echo '//' >> src/main.zig && time
bun run build`

|Platform|0.14.1|0.15.2|Speedup|
|-|-|-|-|
|macos debug asan|126.90s|106.27s|1.19x|
|macos debug noasan|60.62s|50.85s|1.19x|
|linux debug asan|292.77s|241.45s|1.21x|
|linux debug noasan|146.58s|130.94s|1.12x|
|linux debug use_llvm=false|n/a|78.27s|1.87x|
|windows debug asan|177.13s|142.55s|1.24x|

Runtime performance:

- next build memory usage may have gone up by 5%. Otherwise seems the
same. Some code with writers may have gotten slower, especially one
instance of a counting writer and a few instances of unbuffered writers
that now have vtable overhead.
- File size reduced by 800kb (from 100.2mb to 99.4mb)

Improvements:

- `@export` hack is no longer needed for watch
- native x86_64 backend for linux builds faster. to use it, set use_llvm
false and no_link_obj false. also set `ASAN_OPTIONS=detect_leaks=0`
otherwise it will spam the output with tens of thousands of lines of
debug info errors. may need to use the zig lldb fork for debugging.
- zig test-obj, which we will be able to use for zig unit tests

Still an issue:

- false 'dependency loop' errors remain in watch mode
- watch mode crashes observed

Follow-up:

- [ ] search `comptime Writer: type` and `comptime W: type` and remove
- [ ] remove format_mode in our zig fork
- [ ] remove deprecated.zig autoFormatLabelFallback
- [ ] remove deprecated.zig autoFormatLabel
- [ ] remove deprecated.BufferedWriter and BufferedReader
- [ ] remove override_no_export_cpp_apis as it is no longer needed
- [ ] css Parser(W) -> Parser, and remove all the comptime writer: type
params
- [ ] remove deprecated writer fully

Files that add lines:

```
649     src/deprecated.zig
167     scripts/pack-codegen-for-zig-team.ts
54      scripts/cleartrace-impl.js
46      scripts/cleartrace.ts
43      src/windows.zig
18      src/fs.zig
17      src/bun.js/ConsoleObject.zig
16      src/output.zig
12      src/bun.js/test/debug.zig
12      src/bun.js/node/node_fs.zig
8       src/env_loader.zig
7       src/css/printer.zig
7       src/cli/init_command.zig
7       src/bun.js/node.zig
6       src/string/escapeRegExp.zig
6       src/install/PnpmMatcher.zig
5       src/bun.js/webcore/Blob.zig
4       src/crash_handler.zig
4       src/bun.zig
3       src/install/lockfile/bun.lock.zig
3       src/cli/update_interactive_command.zig
3       src/cli/pack_command.zig
3       build.zig
2       src/Progress.zig
2       src/install/lockfile/lockfile_json_stringify_for_debugging.zig
2       src/css/small_list.zig
2       src/bun.js/webcore/prompt.zig
1       test/internal/ban-words.test.ts
1       test/internal/ban-limits.json
1       src/watcher/WatcherTrace.zig
1       src/transpiler.zig
1       src/shell/builtin/cp.zig
1       src/js_printer.zig
1       src/io/PipeReader.zig
1       src/install/bin.zig
1       src/css/selectors/selector.zig
1       src/cli/run_command.zig
1       src/bun.js/RuntimeTranspilerStore.zig
1       src/bun.js/bindings/JSRef.zig
1       src/bake/DevServer.zig
```

Files that remove lines:

```
-1      src/test/recover.zig
-1      src/sql/postgres/SocketMonitor.zig
-1      src/sql/mysql/MySQLRequestQueue.zig
-1      src/sourcemap/CodeCoverage.zig
-1      src/css/values/color_js.zig
-1      src/compile_target.zig
-1      src/bundler/linker_context/convertStmtsForChunk.zig
-1      src/bundler/bundle_v2.zig
-1      src/bun.js/webcore/blob/read_file.zig
-1      src/ast/base.zig
-2      src/sql/postgres/protocol/ArrayList.zig
-2      src/shell/builtin/mkdir.zig
-2      src/install/PackageManager/patchPackage.zig
-2      src/install/PackageManager/PackageManagerDirectories.zig
-2      src/fmt.zig
-2      src/css/declaration.zig
-2      src/css/css_parser.zig
-2      src/collections/baby_list.zig
-2      src/bun.js/bindings/ZigStackFrame.zig
-2      src/ast/E.zig
-3      src/StandaloneModuleGraph.zig
-3      src/deps/picohttp.zig
-3      src/deps/libuv.zig
-3      src/btjs.zig
-4      src/threading/Futex.zig
-4      src/shell/builtin/touch.zig
-4      src/meta.zig
-4      src/install/lockfile.zig
-4      src/css/selectors/parser.zig
-5      src/shell/interpreter.zig
-5      src/css/error.zig
-5      src/bun.js/web_worker.zig
-5      src/bun.js.zig
-6      src/cli/test_command.zig
-6      src/bun.js/VirtualMachine.zig
-6      src/bun.js/uuid.zig
-6      src/bun.js/bindings/JSValue.zig
-9      src/bun.js/test/pretty_format.zig
-9      src/bun.js/api/BunObject.zig
-14     src/install/install_binding.zig
-14     src/fd.zig
-14     src/bun.js/node/path.zig
-14     scripts/pack-codegen-for-zig-team.sh
-17     src/bun.js/test/diff_format.zig
```

`git diff --numstat origin/main...HEAD | awk '{ print ($1-$2)"\t"$3 }' |
sort -rn`

---------

Co-authored-by: autofix-ci[bot] <114827586+autofix-ci[bot]@users.noreply.github.com>
Co-authored-by: Dylan Conway <dylan.conway567@gmail.com>
Co-authored-by: Meghan Denny <meghan@bun.com>
Co-authored-by: tayor.fish <contact@taylor.fish>
2025-11-10 14:38:26 -08:00

294 lines
12 KiB
Zig

/// Demo app testing the macOS libcompression bindings.
const std = @import("std");
const CompressionFramework = struct {
var handle: ?*anyopaque = null;
pub fn load() !void {
handle = std.posix.darwin.dlopen("libcompression.dylib", 1);
if (handle == null)
return error.@"failed to load Compression.framework";
compression_encode_scratch_buffer_size = @alignCast(@ptrCast(std.c.dlsym(handle, "compression_encode_scratch_buffer_size").?));
compression_encode_buffer = @alignCast(@ptrCast(std.c.dlsym(handle, "compression_encode_buffer").?));
compression_decode_scratch_buffer_size = @alignCast(@ptrCast(std.c.dlsym(handle, "compression_decode_scratch_buffer_size").?));
compression_decode_buffer = @alignCast(@ptrCast(std.c.dlsym(handle, "compression_decode_buffer").?));
compression_stream_init = @alignCast(@ptrCast(std.c.dlsym(handle, "compression_stream_init").?));
compression_stream_process = @alignCast(@ptrCast(std.c.dlsym(handle, "compression_stream_process").?));
compression_stream_destroy = @alignCast(@ptrCast(std.c.dlsym(handle, "compression_stream_destroy").?));
}
pub const compression_algorithm = enum(c_uint) {
LZ4 = 256,
ZLIB = 0x205,
LZMA = 774,
LZ4_RAW = 257,
BROTLI = 2818,
LZFSE = 2049,
LZBITMAP = 1794,
pub fn fromName(name: []const u8) ?compression_algorithm {
if (std.mem.endsWith(u8, name, ".br")) {
return .BROTLI;
} else if (std.mem.endsWith(u8, name, ".lz4")) {
return .LZ4;
} else if (std.mem.endsWith(u8, name, ".lzma")) {
return .LZMA;
} else if (std.mem.endsWith(u8, name, ".lzfse")) {
return .LZFSE;
} else if (std.mem.endsWith(u8, name, ".zlib") or std.mem.endsWith(u8, name, ".gz")) {
return .ZLIB;
} else {
return null;
}
}
};
const compression_encode_scratch_buffer_size_type = fn (algorithm: compression_algorithm) callconv(.c) usize;
const compression_encode_buffer_type = fn (noalias dst_buffer: [*]u8, dst_size: usize, noalias src_buffer: ?[*]const u8, src_size: usize, noalias scratch_buffer: ?*anyopaque, algorithm: compression_algorithm) callconv(.c) usize;
const compression_decode_scratch_buffer_size_type = fn (algorithm: compression_algorithm) callconv(.c) usize;
const compression_decode_buffer_type = fn (noalias dst_buffer: [*]u8, dst_size: usize, noalias src_buffer: ?[*]const u8, src_size: usize, noalias scratch_buffer: ?*anyopaque, algorithm: compression_algorithm) callconv(.c) usize;
const compression_stream_init_type = fn (stream: *compression_stream, operation: compression_stream_operation, algorithm: compression_algorithm) callconv(.c) compression_status;
const compression_stream_process_type = fn (stream: *compression_stream, flags: c_int) callconv(.c) compression_status;
const compression_stream_destroy_type = fn (stream: *compression_stream) callconv(.c) compression_status;
var compression_encode_scratch_buffer_size: *const compression_encode_scratch_buffer_size_type = undefined;
var compression_encode_buffer: *const compression_encode_buffer_type = undefined;
var compression_decode_scratch_buffer_size: *const compression_decode_scratch_buffer_size_type = undefined;
var compression_decode_buffer: *const compression_decode_buffer_type = undefined;
var compression_stream_init: *const compression_stream_init_type = undefined;
var compression_stream_process: *const compression_stream_process_type = undefined;
var compression_stream_destroy: *const compression_stream_destroy_type = undefined;
pub const compression_stream = extern struct {
dst_ptr: ?[*]u8 = null,
dst_size: usize = 0,
src_ptr: ?[*]const u8 = null,
src_size: usize = 0,
state: ?*anyopaque = null,
pub fn init(src: []const u8, operation: compression_stream_operation, algorithm: compression_algorithm) !compression_stream {
var stream = compression_stream{
.src_ptr = src.ptr,
.src_size = src.len,
.dst_ptr = null,
.dst_size = 0,
};
switch (compression_stream_init(&stream, operation, algorithm)) {
.OK => {},
.ERROR => return error.@"failed to initialize compression stream",
.END => return error.@"compression stream init returned END",
}
return stream;
}
pub fn process(stream: *compression_stream, data: []const u8, is_done: bool, comptime Iterator: type, iter: *Iterator) !StreamResult {
stream.src_ptr = data.ptr;
stream.src_size = data.len;
const initial_dest = try iter.wrote(0);
stream.dst_ptr = initial_dest.ptr;
stream.dst_size = initial_dest.len;
var total_written: usize = 0;
while (true) {
var flags: c_int = 0;
if (stream.src_size == 0 and is_done) {
flags = COMPRESSION_STREAM_FINALIZE;
} else if (stream.src_size == 0) {
return .{
.progress = .{
.read = data.len - stream.src_size,
.wrote = total_written,
},
};
}
const prev_size = stream.dst_size;
const rc = compression_stream_process(stream, flags);
const wrote = prev_size - stream.dst_size;
switch (rc) {
.OK => {
const new_buffer = try iter.wrote(wrote);
stream.dst_ptr = new_buffer.ptr;
stream.dst_size = new_buffer.len;
total_written += wrote;
},
.END => {
_ = try iter.wrote(wrote);
total_written += wrote;
return .{
.done = .{
.read = data.len - stream.src_size,
.wrote = total_written,
},
};
},
.ERROR => {
return .{
.err = .{
.err = error.@"failed to process compression stream",
.read = data.len - stream.src_size,
.wrote = total_written,
},
};
},
}
}
}
};
pub const COMPRESSION_STREAM_ENCODE: c_int = 0;
pub const COMPRESSION_STREAM_DECODE: c_int = 1;
pub const compression_stream_operation = enum(c_uint) {
ENCODE = 0,
DECODE = 1,
};
pub const COMPRESSION_STREAM_FINALIZE: c_int = 1;
pub const compression_stream_flags = c_uint;
pub const compression_status = enum(c_int) {
OK = 0,
ERROR = -1,
END = 1,
};
const StreamResult = union(enum) {
done: struct {
read: usize = 0,
wrote: usize = 0,
},
err: struct {
read: usize = 0,
wrote: usize = 0,
err: anyerror,
},
progress: struct {
read: usize = 0,
wrote: usize = 0,
},
};
pub fn compress(data: []const u8, algorithm: compression_algorithm, is_done: bool, writer: anytype) !StreamResult {
var scratch_buffer: [64 * 1024]u8 = undefined;
const scratch_buffer_size = compression_encode_scratch_buffer_size(algorithm);
if (scratch_buffer_size >= scratch_buffer.len) {
std.debug.panic("scratch buffer size is too small {d}", .{scratch_buffer_size});
}
var stream = try compression_stream.init(data, .ENCODE, algorithm);
defer _ = compression_stream_destroy(&stream);
const Iterator = struct {
writer: @TypeOf(writer),
scratch_buffer: []u8,
pub fn wrote(this: *@This(), w: usize) ![]u8 {
try this.writer.writeAll(this.scratch_buffer[0..w]);
return this.scratch_buffer;
}
};
var iter = Iterator{
.writer = writer,
.scratch_buffer = &scratch_buffer,
};
return try stream.process(data, is_done, Iterator, &iter);
}
pub fn decompress(data: []const u8, algorithm: compression_algorithm, is_done: bool, writer: anytype) !StreamResult {
var scratch_buffer: [64 * 1024]u8 = undefined;
const scratch_buffer_size = compression_decode_scratch_buffer_size(algorithm);
if (scratch_buffer_size >= scratch_buffer.len) {
std.debug.panic("scratch buffer size is too small {d}", .{scratch_buffer_size});
}
var stream = try compression_stream.init(data, .DECODE, algorithm);
defer _ = compression_stream_destroy(&stream);
const Iterator = struct {
writer: @TypeOf(writer),
scratch_buffer: []u8,
pub fn wrote(this: *@This(), w: usize) ![]u8 {
try this.writer.writeAll(this.scratch_buffer[0..w]);
return this.scratch_buffer;
}
};
var iter = Iterator{
.writer = writer,
.scratch_buffer = &scratch_buffer,
};
return try stream.process(data, is_done, Iterator, &iter);
}
};
pub fn main() anyerror!void {
try CompressionFramework.load();
var args = std.process.args();
const argv0 = args.next() orelse "";
const first = args.next() orelse "";
const second = args.next() orelse "";
var algorithm: ?CompressionFramework.compression_algorithm = null;
var operation: ?CompressionFramework.compression_stream_operation = null;
if (CompressionFramework.compression_algorithm.fromName(first)) |a| {
algorithm = a;
operation = .DECODE;
} else if (CompressionFramework.compression_algorithm.fromName(second)) |o| {
algorithm = o;
operation = .ENCODE;
}
if (algorithm == null or operation == null) {
try std.io.getStdErr().writer().print("to compress: {s} ./file ./out.{{br,gz,lz4,lzfse}}\nto decompress: {s} ./out.{{br,gz,lz4,lzfse}} ./out\n", .{ argv0, argv0 });
std.posix.exit(1);
}
var output_file: std.fs.File = undefined;
var input_file: std.fs.File = undefined;
if (second.len == 0) {
output_file = std.io.getStdOut();
} else {
output_file = try std.fs.cwd().createFile(second, .{
.truncate = true,
});
}
if (first.len == 0) {
input_file = std.io.getStdIn();
} else {
input_file = try std.fs.cwd().openFile(first, .{});
}
var writer = std.io.BufferedWriter(64 * 1024, @TypeOf(output_file.writer())){
.unbuffered_writer = output_file.writer(),
};
const input_bytes = try input_file.readToEndAlloc(std.heap.c_allocator, std.math.maxInt(usize));
if (operation == .ENCODE) {
switch (try CompressionFramework.compress(input_bytes, algorithm.?, true, writer.writer())) {
.err => |err| {
return err.err;
},
else => {},
}
} else {
switch (try CompressionFramework.decompress(input_bytes, algorithm.?, true, writer.writer())) {
.err => |err| {
return err.err;
},
else => {},
}
}
try writer.flush();
}