mirror of
https://github.com/oven-sh/bun
synced 2026-02-02 15:08:46 +00:00
Fixes ENG-21287
Build times, from `bun run build && echo '//' >> src/main.zig && time
bun run build`
|Platform|0.14.1|0.15.2|Speedup|
|-|-|-|-|
|macos debug asan|126.90s|106.27s|1.19x|
|macos debug noasan|60.62s|50.85s|1.19x|
|linux debug asan|292.77s|241.45s|1.21x|
|linux debug noasan|146.58s|130.94s|1.12x|
|linux debug use_llvm=false|n/a|78.27s|1.87x|
|windows debug asan|177.13s|142.55s|1.24x|
Runtime performance:
- next build memory usage may have gone up by 5%. Otherwise seems the
same. Some code with writers may have gotten slower, especially one
instance of a counting writer and a few instances of unbuffered writers
that now have vtable overhead.
- File size reduced by 800kb (from 100.2mb to 99.4mb)
Improvements:
- `@export` hack is no longer needed for watch
- native x86_64 backend for linux builds faster. to use it, set use_llvm
false and no_link_obj false. also set `ASAN_OPTIONS=detect_leaks=0`
otherwise it will spam the output with tens of thousands of lines of
debug info errors. may need to use the zig lldb fork for debugging.
- zig test-obj, which we will be able to use for zig unit tests
Still an issue:
- false 'dependency loop' errors remain in watch mode
- watch mode crashes observed
Follow-up:
- [ ] search `comptime Writer: type` and `comptime W: type` and remove
- [ ] remove format_mode in our zig fork
- [ ] remove deprecated.zig autoFormatLabelFallback
- [ ] remove deprecated.zig autoFormatLabel
- [ ] remove deprecated.BufferedWriter and BufferedReader
- [ ] remove override_no_export_cpp_apis as it is no longer needed
- [ ] css Parser(W) -> Parser, and remove all the comptime writer: type
params
- [ ] remove deprecated writer fully
Files that add lines:
```
649 src/deprecated.zig
167 scripts/pack-codegen-for-zig-team.ts
54 scripts/cleartrace-impl.js
46 scripts/cleartrace.ts
43 src/windows.zig
18 src/fs.zig
17 src/bun.js/ConsoleObject.zig
16 src/output.zig
12 src/bun.js/test/debug.zig
12 src/bun.js/node/node_fs.zig
8 src/env_loader.zig
7 src/css/printer.zig
7 src/cli/init_command.zig
7 src/bun.js/node.zig
6 src/string/escapeRegExp.zig
6 src/install/PnpmMatcher.zig
5 src/bun.js/webcore/Blob.zig
4 src/crash_handler.zig
4 src/bun.zig
3 src/install/lockfile/bun.lock.zig
3 src/cli/update_interactive_command.zig
3 src/cli/pack_command.zig
3 build.zig
2 src/Progress.zig
2 src/install/lockfile/lockfile_json_stringify_for_debugging.zig
2 src/css/small_list.zig
2 src/bun.js/webcore/prompt.zig
1 test/internal/ban-words.test.ts
1 test/internal/ban-limits.json
1 src/watcher/WatcherTrace.zig
1 src/transpiler.zig
1 src/shell/builtin/cp.zig
1 src/js_printer.zig
1 src/io/PipeReader.zig
1 src/install/bin.zig
1 src/css/selectors/selector.zig
1 src/cli/run_command.zig
1 src/bun.js/RuntimeTranspilerStore.zig
1 src/bun.js/bindings/JSRef.zig
1 src/bake/DevServer.zig
```
Files that remove lines:
```
-1 src/test/recover.zig
-1 src/sql/postgres/SocketMonitor.zig
-1 src/sql/mysql/MySQLRequestQueue.zig
-1 src/sourcemap/CodeCoverage.zig
-1 src/css/values/color_js.zig
-1 src/compile_target.zig
-1 src/bundler/linker_context/convertStmtsForChunk.zig
-1 src/bundler/bundle_v2.zig
-1 src/bun.js/webcore/blob/read_file.zig
-1 src/ast/base.zig
-2 src/sql/postgres/protocol/ArrayList.zig
-2 src/shell/builtin/mkdir.zig
-2 src/install/PackageManager/patchPackage.zig
-2 src/install/PackageManager/PackageManagerDirectories.zig
-2 src/fmt.zig
-2 src/css/declaration.zig
-2 src/css/css_parser.zig
-2 src/collections/baby_list.zig
-2 src/bun.js/bindings/ZigStackFrame.zig
-2 src/ast/E.zig
-3 src/StandaloneModuleGraph.zig
-3 src/deps/picohttp.zig
-3 src/deps/libuv.zig
-3 src/btjs.zig
-4 src/threading/Futex.zig
-4 src/shell/builtin/touch.zig
-4 src/meta.zig
-4 src/install/lockfile.zig
-4 src/css/selectors/parser.zig
-5 src/shell/interpreter.zig
-5 src/css/error.zig
-5 src/bun.js/web_worker.zig
-5 src/bun.js.zig
-6 src/cli/test_command.zig
-6 src/bun.js/VirtualMachine.zig
-6 src/bun.js/uuid.zig
-6 src/bun.js/bindings/JSValue.zig
-9 src/bun.js/test/pretty_format.zig
-9 src/bun.js/api/BunObject.zig
-14 src/install/install_binding.zig
-14 src/fd.zig
-14 src/bun.js/node/path.zig
-14 scripts/pack-codegen-for-zig-team.sh
-17 src/bun.js/test/diff_format.zig
```
`git diff --numstat origin/main...HEAD | awk '{ print ($1-$2)"\t"$3 }' |
sort -rn`
---------
Co-authored-by: autofix-ci[bot] <114827586+autofix-ci[bot]@users.noreply.github.com>
Co-authored-by: Dylan Conway <dylan.conway567@gmail.com>
Co-authored-by: Meghan Denny <meghan@bun.com>
Co-authored-by: tayor.fish <contact@taylor.fish>
294 lines
12 KiB
Zig
294 lines
12 KiB
Zig
/// Demo app testing the macOS libcompression bindings.
|
|
const std = @import("std");
|
|
const CompressionFramework = struct {
|
|
var handle: ?*anyopaque = null;
|
|
pub fn load() !void {
|
|
handle = std.posix.darwin.dlopen("libcompression.dylib", 1);
|
|
|
|
if (handle == null)
|
|
return error.@"failed to load Compression.framework";
|
|
|
|
compression_encode_scratch_buffer_size = @alignCast(@ptrCast(std.c.dlsym(handle, "compression_encode_scratch_buffer_size").?));
|
|
compression_encode_buffer = @alignCast(@ptrCast(std.c.dlsym(handle, "compression_encode_buffer").?));
|
|
compression_decode_scratch_buffer_size = @alignCast(@ptrCast(std.c.dlsym(handle, "compression_decode_scratch_buffer_size").?));
|
|
compression_decode_buffer = @alignCast(@ptrCast(std.c.dlsym(handle, "compression_decode_buffer").?));
|
|
compression_stream_init = @alignCast(@ptrCast(std.c.dlsym(handle, "compression_stream_init").?));
|
|
compression_stream_process = @alignCast(@ptrCast(std.c.dlsym(handle, "compression_stream_process").?));
|
|
compression_stream_destroy = @alignCast(@ptrCast(std.c.dlsym(handle, "compression_stream_destroy").?));
|
|
}
|
|
|
|
pub const compression_algorithm = enum(c_uint) {
|
|
LZ4 = 256,
|
|
ZLIB = 0x205,
|
|
LZMA = 774,
|
|
LZ4_RAW = 257,
|
|
BROTLI = 2818,
|
|
LZFSE = 2049,
|
|
LZBITMAP = 1794,
|
|
|
|
pub fn fromName(name: []const u8) ?compression_algorithm {
|
|
if (std.mem.endsWith(u8, name, ".br")) {
|
|
return .BROTLI;
|
|
} else if (std.mem.endsWith(u8, name, ".lz4")) {
|
|
return .LZ4;
|
|
} else if (std.mem.endsWith(u8, name, ".lzma")) {
|
|
return .LZMA;
|
|
} else if (std.mem.endsWith(u8, name, ".lzfse")) {
|
|
return .LZFSE;
|
|
} else if (std.mem.endsWith(u8, name, ".zlib") or std.mem.endsWith(u8, name, ".gz")) {
|
|
return .ZLIB;
|
|
} else {
|
|
return null;
|
|
}
|
|
}
|
|
};
|
|
const compression_encode_scratch_buffer_size_type = fn (algorithm: compression_algorithm) callconv(.c) usize;
|
|
const compression_encode_buffer_type = fn (noalias dst_buffer: [*]u8, dst_size: usize, noalias src_buffer: ?[*]const u8, src_size: usize, noalias scratch_buffer: ?*anyopaque, algorithm: compression_algorithm) callconv(.c) usize;
|
|
const compression_decode_scratch_buffer_size_type = fn (algorithm: compression_algorithm) callconv(.c) usize;
|
|
const compression_decode_buffer_type = fn (noalias dst_buffer: [*]u8, dst_size: usize, noalias src_buffer: ?[*]const u8, src_size: usize, noalias scratch_buffer: ?*anyopaque, algorithm: compression_algorithm) callconv(.c) usize;
|
|
|
|
const compression_stream_init_type = fn (stream: *compression_stream, operation: compression_stream_operation, algorithm: compression_algorithm) callconv(.c) compression_status;
|
|
const compression_stream_process_type = fn (stream: *compression_stream, flags: c_int) callconv(.c) compression_status;
|
|
const compression_stream_destroy_type = fn (stream: *compression_stream) callconv(.c) compression_status;
|
|
|
|
var compression_encode_scratch_buffer_size: *const compression_encode_scratch_buffer_size_type = undefined;
|
|
var compression_encode_buffer: *const compression_encode_buffer_type = undefined;
|
|
var compression_decode_scratch_buffer_size: *const compression_decode_scratch_buffer_size_type = undefined;
|
|
var compression_decode_buffer: *const compression_decode_buffer_type = undefined;
|
|
|
|
var compression_stream_init: *const compression_stream_init_type = undefined;
|
|
var compression_stream_process: *const compression_stream_process_type = undefined;
|
|
var compression_stream_destroy: *const compression_stream_destroy_type = undefined;
|
|
pub const compression_stream = extern struct {
|
|
dst_ptr: ?[*]u8 = null,
|
|
dst_size: usize = 0,
|
|
src_ptr: ?[*]const u8 = null,
|
|
src_size: usize = 0,
|
|
state: ?*anyopaque = null,
|
|
|
|
pub fn init(src: []const u8, operation: compression_stream_operation, algorithm: compression_algorithm) !compression_stream {
|
|
var stream = compression_stream{
|
|
.src_ptr = src.ptr,
|
|
.src_size = src.len,
|
|
.dst_ptr = null,
|
|
.dst_size = 0,
|
|
};
|
|
|
|
switch (compression_stream_init(&stream, operation, algorithm)) {
|
|
.OK => {},
|
|
.ERROR => return error.@"failed to initialize compression stream",
|
|
.END => return error.@"compression stream init returned END",
|
|
}
|
|
|
|
return stream;
|
|
}
|
|
|
|
pub fn process(stream: *compression_stream, data: []const u8, is_done: bool, comptime Iterator: type, iter: *Iterator) !StreamResult {
|
|
stream.src_ptr = data.ptr;
|
|
stream.src_size = data.len;
|
|
|
|
const initial_dest = try iter.wrote(0);
|
|
stream.dst_ptr = initial_dest.ptr;
|
|
stream.dst_size = initial_dest.len;
|
|
|
|
var total_written: usize = 0;
|
|
while (true) {
|
|
var flags: c_int = 0;
|
|
if (stream.src_size == 0 and is_done) {
|
|
flags = COMPRESSION_STREAM_FINALIZE;
|
|
} else if (stream.src_size == 0) {
|
|
return .{
|
|
.progress = .{
|
|
.read = data.len - stream.src_size,
|
|
.wrote = total_written,
|
|
},
|
|
};
|
|
}
|
|
|
|
const prev_size = stream.dst_size;
|
|
const rc = compression_stream_process(stream, flags);
|
|
const wrote = prev_size - stream.dst_size;
|
|
switch (rc) {
|
|
.OK => {
|
|
const new_buffer = try iter.wrote(wrote);
|
|
stream.dst_ptr = new_buffer.ptr;
|
|
stream.dst_size = new_buffer.len;
|
|
total_written += wrote;
|
|
},
|
|
.END => {
|
|
_ = try iter.wrote(wrote);
|
|
total_written += wrote;
|
|
|
|
return .{
|
|
.done = .{
|
|
.read = data.len - stream.src_size,
|
|
.wrote = total_written,
|
|
},
|
|
};
|
|
},
|
|
.ERROR => {
|
|
return .{
|
|
.err = .{
|
|
.err = error.@"failed to process compression stream",
|
|
.read = data.len - stream.src_size,
|
|
.wrote = total_written,
|
|
},
|
|
};
|
|
},
|
|
}
|
|
}
|
|
}
|
|
};
|
|
pub const COMPRESSION_STREAM_ENCODE: c_int = 0;
|
|
pub const COMPRESSION_STREAM_DECODE: c_int = 1;
|
|
pub const compression_stream_operation = enum(c_uint) {
|
|
ENCODE = 0,
|
|
DECODE = 1,
|
|
};
|
|
pub const COMPRESSION_STREAM_FINALIZE: c_int = 1;
|
|
pub const compression_stream_flags = c_uint;
|
|
pub const compression_status = enum(c_int) {
|
|
OK = 0,
|
|
ERROR = -1,
|
|
END = 1,
|
|
};
|
|
|
|
const StreamResult = union(enum) {
|
|
done: struct {
|
|
read: usize = 0,
|
|
wrote: usize = 0,
|
|
},
|
|
err: struct {
|
|
read: usize = 0,
|
|
wrote: usize = 0,
|
|
err: anyerror,
|
|
},
|
|
progress: struct {
|
|
read: usize = 0,
|
|
wrote: usize = 0,
|
|
},
|
|
};
|
|
|
|
pub fn compress(data: []const u8, algorithm: compression_algorithm, is_done: bool, writer: anytype) !StreamResult {
|
|
var scratch_buffer: [64 * 1024]u8 = undefined;
|
|
|
|
const scratch_buffer_size = compression_encode_scratch_buffer_size(algorithm);
|
|
if (scratch_buffer_size >= scratch_buffer.len) {
|
|
std.debug.panic("scratch buffer size is too small {d}", .{scratch_buffer_size});
|
|
}
|
|
|
|
var stream = try compression_stream.init(data, .ENCODE, algorithm);
|
|
|
|
defer _ = compression_stream_destroy(&stream);
|
|
const Iterator = struct {
|
|
writer: @TypeOf(writer),
|
|
scratch_buffer: []u8,
|
|
pub fn wrote(this: *@This(), w: usize) ![]u8 {
|
|
try this.writer.writeAll(this.scratch_buffer[0..w]);
|
|
return this.scratch_buffer;
|
|
}
|
|
};
|
|
|
|
var iter = Iterator{
|
|
.writer = writer,
|
|
.scratch_buffer = &scratch_buffer,
|
|
};
|
|
|
|
return try stream.process(data, is_done, Iterator, &iter);
|
|
}
|
|
|
|
pub fn decompress(data: []const u8, algorithm: compression_algorithm, is_done: bool, writer: anytype) !StreamResult {
|
|
var scratch_buffer: [64 * 1024]u8 = undefined;
|
|
|
|
const scratch_buffer_size = compression_decode_scratch_buffer_size(algorithm);
|
|
if (scratch_buffer_size >= scratch_buffer.len) {
|
|
std.debug.panic("scratch buffer size is too small {d}", .{scratch_buffer_size});
|
|
}
|
|
|
|
var stream = try compression_stream.init(data, .DECODE, algorithm);
|
|
defer _ = compression_stream_destroy(&stream);
|
|
|
|
const Iterator = struct {
|
|
writer: @TypeOf(writer),
|
|
scratch_buffer: []u8,
|
|
pub fn wrote(this: *@This(), w: usize) ![]u8 {
|
|
try this.writer.writeAll(this.scratch_buffer[0..w]);
|
|
return this.scratch_buffer;
|
|
}
|
|
};
|
|
|
|
var iter = Iterator{
|
|
.writer = writer,
|
|
.scratch_buffer = &scratch_buffer,
|
|
};
|
|
|
|
return try stream.process(data, is_done, Iterator, &iter);
|
|
}
|
|
};
|
|
|
|
pub fn main() anyerror!void {
|
|
try CompressionFramework.load();
|
|
|
|
var args = std.process.args();
|
|
const argv0 = args.next() orelse "";
|
|
|
|
const first = args.next() orelse "";
|
|
const second = args.next() orelse "";
|
|
var algorithm: ?CompressionFramework.compression_algorithm = null;
|
|
var operation: ?CompressionFramework.compression_stream_operation = null;
|
|
|
|
if (CompressionFramework.compression_algorithm.fromName(first)) |a| {
|
|
algorithm = a;
|
|
operation = .DECODE;
|
|
} else if (CompressionFramework.compression_algorithm.fromName(second)) |o| {
|
|
algorithm = o;
|
|
operation = .ENCODE;
|
|
}
|
|
|
|
if (algorithm == null or operation == null) {
|
|
try std.io.getStdErr().writer().print("to compress: {s} ./file ./out.{{br,gz,lz4,lzfse}}\nto decompress: {s} ./out.{{br,gz,lz4,lzfse}} ./out\n", .{ argv0, argv0 });
|
|
std.posix.exit(1);
|
|
}
|
|
|
|
var output_file: std.fs.File = undefined;
|
|
var input_file: std.fs.File = undefined;
|
|
|
|
if (second.len == 0) {
|
|
output_file = std.io.getStdOut();
|
|
} else {
|
|
output_file = try std.fs.cwd().createFile(second, .{
|
|
.truncate = true,
|
|
});
|
|
}
|
|
|
|
if (first.len == 0) {
|
|
input_file = std.io.getStdIn();
|
|
} else {
|
|
input_file = try std.fs.cwd().openFile(first, .{});
|
|
}
|
|
|
|
var writer = std.io.BufferedWriter(64 * 1024, @TypeOf(output_file.writer())){
|
|
.unbuffered_writer = output_file.writer(),
|
|
};
|
|
|
|
const input_bytes = try input_file.readToEndAlloc(std.heap.c_allocator, std.math.maxInt(usize));
|
|
|
|
if (operation == .ENCODE) {
|
|
switch (try CompressionFramework.compress(input_bytes, algorithm.?, true, writer.writer())) {
|
|
.err => |err| {
|
|
return err.err;
|
|
},
|
|
else => {},
|
|
}
|
|
} else {
|
|
switch (try CompressionFramework.decompress(input_bytes, algorithm.?, true, writer.writer())) {
|
|
.err => |err| {
|
|
return err.err;
|
|
},
|
|
else => {},
|
|
}
|
|
}
|
|
|
|
try writer.flush();
|
|
}
|