Merge pull request #24493 from ziglang/readSliceShort

std.Io.Reader: fix readSliceShort with smaller buffer than Reader
This commit is contained in:
Andrew Kelley 2025-07-17 18:35:38 +02:00 committed by GitHub
commit a8dc32e4ec
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
3 changed files with 43 additions and 66 deletions

View File

@ -593,48 +593,29 @@ pub fn readSliceAll(r: *Reader, buffer: []u8) Error!void {
/// See also:
/// * `readSliceAll`
pub fn readSliceShort(r: *Reader, buffer: []u8) ShortError!usize {
const in_buffer = r.buffer[r.seek..r.end];
const copy_len = @min(buffer.len, in_buffer.len);
@memcpy(buffer[0..copy_len], in_buffer[0..copy_len]);
if (buffer.len - copy_len == 0) {
r.seek += copy_len;
return buffer.len;
}
var i: usize = copy_len;
r.end = 0;
r.seek = 0;
var i: usize = 0;
while (true) {
const buffer_contents = r.buffer[r.seek..r.end];
const dest = buffer[i..];
const copy_len = @min(dest.len, buffer_contents.len);
@memcpy(dest[0..copy_len], buffer_contents[0..copy_len]);
if (dest.len - copy_len == 0) {
@branchHint(.likely);
r.seek += copy_len;
return buffer.len;
}
i += copy_len;
r.end = 0;
r.seek = 0;
const remaining = buffer[i..];
var wrapper: Writer.VectorWrapper = .{
.it = .{
.first = remaining,
.last = r.buffer,
},
.writer = .{
.buffer = if (remaining.len >= r.buffer.len) remaining else r.buffer,
.vtable = Writer.VectorWrapper.vtable,
},
};
const n = r.vtable.stream(r, &wrapper.writer, .unlimited) catch |err| switch (err) {
error.WriteFailed => {
if (!wrapper.used) {
assert(r.seek == 0);
r.seek = remaining.len;
r.end = wrapper.writer.end;
@memcpy(remaining, r.buffer[0..remaining.len]);
}
return buffer.len;
},
const new_remaining_len = readVecInner(r, &.{}, remaining, remaining.len) catch |err| switch (err) {
error.EndOfStream => return i,
error.ReadFailed => return error.ReadFailed,
};
if (n < remaining.len) {
i += n;
continue;
}
r.end = n - remaining.len;
return buffer.len;
if (new_remaining_len == 0) return buffer.len;
i += remaining.len - new_remaining_len;
}
return buffer.len;
}
/// Fill `buffer` with the next `buffer.len` bytes from the stream, advancing
@ -1640,6 +1621,19 @@ test readSliceShort {
try testing.expectEqual(0, try r.readSliceShort(&buf));
}
test "readSliceShort with smaller buffer than Reader" {
var reader_buf: [15]u8 = undefined;
const str = "This is a test";
var one_byte_stream: testing.Reader = .init(&reader_buf, &.{
.{ .buffer = str },
});
one_byte_stream.artificial_limit = .limited(1);
var buf: [14]u8 = undefined;
try testing.expectEqual(14, try one_byte_stream.interface.readSliceShort(&buf));
try testing.expectEqualStrings(str, &buf);
}
test readVec {
var r: Reader = .fixed(std.ascii.letters);
var flat_buffer: [52]u8 = undefined;
@ -1702,33 +1696,13 @@ fn failingDiscard(r: *Reader, limit: Limit) Error!usize {
}
test "readAlloc when the backing reader provides one byte at a time" {
const OneByteReader = struct {
str: []const u8,
i: usize,
reader: Reader,
fn stream(r: *Reader, w: *Writer, limit: Limit) StreamError!usize {
assert(@intFromEnum(limit) >= 1);
const self: *@This() = @fieldParentPtr("reader", r);
if (self.str.len - self.i == 0) return error.EndOfStream;
try w.writeByte(self.str[self.i]);
self.i += 1;
return 1;
}
};
const str = "This is a test";
var tiny_buffer: [1]u8 = undefined;
var one_byte_stream: OneByteReader = .{
.str = str,
.i = 0,
.reader = .{
.buffer = &tiny_buffer,
.vtable = &.{ .stream = OneByteReader.stream },
.seek = 0,
.end = 0,
},
};
const res = try one_byte_stream.reader.allocRemaining(std.testing.allocator, .unlimited);
var one_byte_stream: testing.Reader = .init(&tiny_buffer, &.{
.{ .buffer = str },
});
one_byte_stream.artificial_limit = .limited(1);
const res = try one_byte_stream.interface.allocRemaining(std.testing.allocator, .unlimited);
defer std.testing.allocator.free(res);
try std.testing.expectEqualStrings(str, res);
}

View File

@ -1207,12 +1207,14 @@ pub inline fn fuzz(
return @import("root").fuzz(context, testOne, options);
}
/// A `std.io.Reader` that writes a predetermined list of buffers during `stream`.
/// A `std.Io.Reader` that writes a predetermined list of buffers during `stream`.
pub const Reader = struct {
calls: []const Call,
interface: std.io.Reader,
interface: std.Io.Reader,
next_call_index: usize,
next_offset: usize,
/// Further reduces how many bytes are written in each `stream` call.
artificial_limit: std.Io.Limit = .unlimited,
pub const Call = struct {
buffer: []const u8,
@ -1232,11 +1234,11 @@ pub const Reader = struct {
};
}
fn stream(io_r: *std.io.Reader, w: *std.io.Writer, limit: std.io.Limit) std.io.Reader.StreamError!usize {
fn stream(io_r: *std.Io.Reader, w: *std.Io.Writer, limit: std.Io.Limit) std.Io.Reader.StreamError!usize {
const r: *Reader = @alignCast(@fieldParentPtr("interface", io_r));
if (r.calls.len - r.next_call_index == 0) return error.EndOfStream;
const call = r.calls[r.next_call_index];
const buffer = limit.sliceConst(call.buffer[r.next_offset..]);
const buffer = r.artificial_limit.sliceConst(limit.sliceConst(call.buffer[r.next_offset..]));
const n = try w.write(buffer);
r.next_offset += n;
if (call.buffer.len - r.next_offset == 0) {

View File

@ -536,7 +536,8 @@ pub fn readSourceFileToEndAlloc(gpa: Allocator, file_reader: *std.fs.File.Reader
if (file_reader.getSize()) |size| {
const casted_size = std.math.cast(u32, size) orelse return error.StreamTooLong;
try buffer.ensureTotalCapacityPrecise(gpa, casted_size);
// +1 to avoid resizing for the null byte added in toOwnedSliceSentinel below.
try buffer.ensureTotalCapacityPrecise(gpa, casted_size + 1);
} else |_| {}
try file_reader.interface.appendRemaining(gpa, .@"2", &buffer, .limited(max_src_size));