diff --git a/ci/srht/freebsd_script b/ci/srht/freebsd_script index fb4a5d357e..6c8fd50a07 100755 --- a/ci/srht/freebsd_script +++ b/ci/srht/freebsd_script @@ -3,13 +3,6 @@ set -x set -e -# The following line can be removed as soon as FreeBSD fixes -# their packaging glitch. If not fixed by March 15, 2020 -# there is something wrong. Should be fixed much sooner. -# note: this will cause some complaints when running -# pkg commands but let's ignore them. -sudo rm /usr/local/etc/pkg/repos/FreeBSD.conf - sudo pkg update -fq sudo pkg install -y cmake py27-s3cmd wget curl jq diff --git a/doc/docgen.zig b/doc/docgen.zig index 319d9e0035..4d2625f54f 100644 --- a/doc/docgen.zig +++ b/doc/docgen.zig @@ -40,12 +40,9 @@ pub fn main() !void { var out_file = try fs.cwd().createFile(out_file_name, .{}); defer out_file.close(); - var file_in_stream = in_file.inStream(); + const input_file_bytes = try in_file.inStream().readAllAlloc(allocator, max_doc_file_size); - const input_file_bytes = try file_in_stream.stream.readAllAlloc(allocator, max_doc_file_size); - - var file_out_stream = out_file.outStream(); - var buffered_out_stream = io.BufferedOutStream(fs.File.WriteError).init(&file_out_stream.stream); + var buffered_out_stream = io.bufferedOutStream(out_file.outStream()); var tokenizer = Tokenizer.init(in_file_name, input_file_bytes); var toc = try genToc(allocator, &tokenizer); @@ -53,7 +50,7 @@ pub fn main() !void { try fs.cwd().makePath(tmp_dir_name); defer fs.deleteTree(tmp_dir_name) catch {}; - try genHtml(allocator, &tokenizer, &toc, &buffered_out_stream.stream, zig_exe); + try genHtml(allocator, &tokenizer, &toc, buffered_out_stream.outStream(), zig_exe); try buffered_out_stream.flush(); } @@ -327,8 +324,7 @@ fn genToc(allocator: *mem.Allocator, tokenizer: *Tokenizer) !Toc { var toc_buf = try std.Buffer.initSize(allocator, 0); defer toc_buf.deinit(); - var toc_buf_adapter = io.BufferOutStream.init(&toc_buf); - var toc = &toc_buf_adapter.stream; + var toc = toc_buf.outStream(); var nodes = std.ArrayList(Node).init(allocator); defer nodes.deinit(); @@ -342,7 +338,7 @@ fn genToc(allocator: *mem.Allocator, tokenizer: *Tokenizer) !Toc { if (header_stack_size != 0) { return parseError(tokenizer, token, "unbalanced headers", .{}); } - try toc.write(" \n"); + try toc.writeAll(" \n"); break; }, Token.Id.Content => { @@ -407,7 +403,7 @@ fn genToc(allocator: *mem.Allocator, tokenizer: *Tokenizer) !Toc { if (last_columns) |n| { try toc.print("