std.Progress: slightly better atomic memcpy

Let's at least do aligned usize loads/stores where possible.
This commit is contained in:
Andrew Kelley 2024-07-09 14:18:23 -07:00
parent b3b923e51f
commit 9363e995fc

View File

@ -90,7 +90,7 @@ pub const Node = struct {
/// 0 means unknown.
/// Little endian.
estimated_total_count: u32,
name: [max_name_len]u8,
name: [max_name_len]u8 align(@alignOf(usize)),
/// Not thread-safe.
fn getIpcFd(s: Storage) ?posix.fd_t {
@ -288,8 +288,9 @@ pub const Node = struct {
@atomicStore(u32, &storage.completed_count, 0, .monotonic);
@atomicStore(u32, &storage.estimated_total_count, std.math.lossyCast(u32, estimated_total_items), .monotonic);
const name_len = @min(max_name_len, name.len);
for (storage.name[0..name_len], name[0..name_len]) |*dest, src| @atomicStore(u8, dest, src, .monotonic);
for (storage.name[name_len..]) |*dest| @atomicStore(u8, dest, 0, .monotonic);
copyAtomicStore(storage.name[0..name_len], name[0..name_len]);
if (name_len < storage.name.len)
@atomicStore(u8, &storage.name[name_len], 0, .monotonic);
const parent_ptr = parentByIndex(free_index);
assert(parent_ptr.* == .unused);
@ -763,7 +764,7 @@ fn serialize(serialized_buffer: *Serialized.Buffer) Serialized {
var begin_parent = @atomicLoad(Node.Parent, parent_ptr, .acquire);
while (begin_parent != .unused) {
const dest_storage = &serialized_buffer.storage[serialized_len];
for (&dest_storage.name, &storage_ptr.name) |*dest, *src| dest.* = @atomicLoad(u8, src, .monotonic);
copyAtomicLoad(&dest_storage.name, &storage_ptr.name);
dest_storage.estimated_total_count = @atomicLoad(u32, &storage_ptr.estimated_total_count, .acquire);
dest_storage.completed_count = @atomicLoad(u32, &storage_ptr.completed_count, .monotonic);
const end_parent = @atomicLoad(Node.Parent, parent_ptr, .acquire);
@ -1384,3 +1385,29 @@ const have_sigwinch = switch (builtin.os.tag) {
/// stderr mutex is held still dumps the stack trace and other debug
/// information.
var stderr_mutex = std.Thread.Mutex.Recursive.init;
fn copyAtomicStore(dest: []align(@alignOf(usize)) u8, src: []const u8) void {
assert(dest.len == src.len);
const chunked_len = dest.len / @sizeOf(usize);
const dest_chunked: []usize = @as([*]usize, @ptrCast(dest))[0..chunked_len];
const src_chunked: []align(1) const usize = @as([*]align(1) const usize, @ptrCast(src))[0..chunked_len];
for (dest_chunked, src_chunked) |*d, s| {
@atomicStore(usize, d, s, .monotonic);
}
const remainder_start = chunked_len * @sizeOf(usize);
for (dest[remainder_start..], src[remainder_start..]) |*d, s| {
@atomicStore(u8, d, s, .monotonic);
}
}
fn copyAtomicLoad(
dest: *align(@alignOf(usize)) [Node.max_name_len]u8,
src: *align(@alignOf(usize)) const [Node.max_name_len]u8,
) void {
const chunked_len = @divExact(dest.len, @sizeOf(usize));
const dest_chunked: *[chunked_len]usize = @ptrCast(dest);
const src_chunked: *const [chunked_len]usize = @ptrCast(src);
for (dest_chunked, src_chunked) |*d, *s| {
d.* = @atomicLoad(usize, s, .monotonic);
}
}