mirror of
https://github.com/ziglang/zig.git
synced 2025-12-06 14:23:09 +00:00
std: add SbrkAllocator and use it for Plan 9
Implements issue #6451. This was needed to support allocation on Plan 9 and now other operating systems like DOS can also use it. It is a modified version of the WasmAllocator since wasm also uses a sbrk-esque allocation system. This commit also adds the necessary system bits for sbrk to work on plan 9.
This commit is contained in:
parent
4d711e8edd
commit
841b54f5e3
@ -21,6 +21,7 @@ pub const WasmAllocator = @import("heap/WasmAllocator.zig");
|
|||||||
pub const WasmPageAllocator = @import("heap/WasmPageAllocator.zig");
|
pub const WasmPageAllocator = @import("heap/WasmPageAllocator.zig");
|
||||||
pub const PageAllocator = @import("heap/PageAllocator.zig");
|
pub const PageAllocator = @import("heap/PageAllocator.zig");
|
||||||
pub const ThreadSafeAllocator = @import("heap/ThreadSafeAllocator.zig");
|
pub const ThreadSafeAllocator = @import("heap/ThreadSafeAllocator.zig");
|
||||||
|
pub const SbrkAllocator = @import("heap/sbrk_allocator.zig").SbrkAllocator;
|
||||||
|
|
||||||
const memory_pool = @import("heap/memory_pool.zig");
|
const memory_pool = @import("heap/memory_pool.zig");
|
||||||
pub const MemoryPool = memory_pool.MemoryPool;
|
pub const MemoryPool = memory_pool.MemoryPool;
|
||||||
@ -228,6 +229,11 @@ pub const page_allocator = if (builtin.target.isWasm())
|
|||||||
.ptr = undefined,
|
.ptr = undefined,
|
||||||
.vtable = &WasmPageAllocator.vtable,
|
.vtable = &WasmPageAllocator.vtable,
|
||||||
}
|
}
|
||||||
|
else if (builtin.target.os.tag == .plan9)
|
||||||
|
Allocator{
|
||||||
|
.ptr = undefined,
|
||||||
|
.vtable = &SbrkAllocator(std.os.plan9.sbrk).vtable,
|
||||||
|
}
|
||||||
else if (builtin.target.os.tag == .freestanding)
|
else if (builtin.target.os.tag == .freestanding)
|
||||||
root.os.heap.page_allocator
|
root.os.heap.page_allocator
|
||||||
else
|
else
|
||||||
|
|||||||
161
lib/std/heap/sbrk_allocator.zig
Normal file
161
lib/std/heap/sbrk_allocator.zig
Normal file
@ -0,0 +1,161 @@
|
|||||||
|
const std = @import("../std.zig");
|
||||||
|
const builtin = @import("builtin");
|
||||||
|
const math = std.math;
|
||||||
|
const Allocator = std.mem.Allocator;
|
||||||
|
const mem = std.mem;
|
||||||
|
const assert = std.debug.assert;
|
||||||
|
|
||||||
|
pub fn SbrkAllocator(comptime sbrk: *const fn (n: usize) usize) type {
|
||||||
|
return struct {
|
||||||
|
pub const vtable = Allocator.VTable{
|
||||||
|
.alloc = alloc,
|
||||||
|
.resize = resize,
|
||||||
|
.free = free,
|
||||||
|
};
|
||||||
|
|
||||||
|
pub const Error = Allocator.Error;
|
||||||
|
|
||||||
|
lock: std.Thread.Mutex = .{},
|
||||||
|
|
||||||
|
const max_usize = math.maxInt(usize);
|
||||||
|
const ushift = math.Log2Int(usize);
|
||||||
|
const bigpage_size = 64 * 1024;
|
||||||
|
const pages_per_bigpage = bigpage_size / mem.page_size;
|
||||||
|
const bigpage_count = max_usize / bigpage_size;
|
||||||
|
|
||||||
|
/// Because of storing free list pointers, the minimum size class is 3.
|
||||||
|
const min_class = math.log2(math.ceilPowerOfTwoAssert(usize, 1 + @sizeOf(usize)));
|
||||||
|
const size_class_count = math.log2(bigpage_size) - min_class;
|
||||||
|
/// 0 - 1 bigpage
|
||||||
|
/// 1 - 2 bigpages
|
||||||
|
/// 2 - 4 bigpages
|
||||||
|
/// etc.
|
||||||
|
const big_size_class_count = math.log2(bigpage_count);
|
||||||
|
|
||||||
|
var next_addrs = [1]usize{0} ** size_class_count;
|
||||||
|
/// For each size class, points to the freed pointer.
|
||||||
|
var frees = [1]usize{0} ** size_class_count;
|
||||||
|
/// For each big size class, points to the freed pointer.
|
||||||
|
var big_frees = [1]usize{0} ** big_size_class_count;
|
||||||
|
|
||||||
|
// TODO don't do the naive locking strategy
|
||||||
|
var lock: std.Thread.Mutex = .{};
|
||||||
|
fn alloc(ctx: *anyopaque, len: usize, log2_align: u8, return_address: usize) ?[*]u8 {
|
||||||
|
_ = ctx;
|
||||||
|
_ = return_address;
|
||||||
|
lock.lock();
|
||||||
|
defer lock.unlock();
|
||||||
|
// Make room for the freelist next pointer.
|
||||||
|
const alignment = @as(usize, 1) << @as(Allocator.Log2Align, @intCast(log2_align));
|
||||||
|
const actual_len = @max(len +| @sizeOf(usize), alignment);
|
||||||
|
const slot_size = math.ceilPowerOfTwo(usize, actual_len) catch return null;
|
||||||
|
const class = math.log2(slot_size) - min_class;
|
||||||
|
if (class < size_class_count) {
|
||||||
|
const addr = a: {
|
||||||
|
const top_free_ptr = frees[class];
|
||||||
|
if (top_free_ptr != 0) {
|
||||||
|
const node = @as(*usize, @ptrFromInt(top_free_ptr + (slot_size - @sizeOf(usize))));
|
||||||
|
frees[class] = node.*;
|
||||||
|
break :a top_free_ptr;
|
||||||
|
}
|
||||||
|
|
||||||
|
const next_addr = next_addrs[class];
|
||||||
|
if (next_addr % mem.page_size == 0) {
|
||||||
|
const addr = allocBigPages(1);
|
||||||
|
if (addr == 0) return null;
|
||||||
|
//std.debug.print("allocated fresh slot_size={d} class={d} addr=0x{x}\n", .{
|
||||||
|
// slot_size, class, addr,
|
||||||
|
//});
|
||||||
|
next_addrs[class] = addr + slot_size;
|
||||||
|
break :a addr;
|
||||||
|
} else {
|
||||||
|
next_addrs[class] = next_addr + slot_size;
|
||||||
|
break :a next_addr;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
return @as([*]u8, @ptrFromInt(addr));
|
||||||
|
}
|
||||||
|
const bigpages_needed = bigPagesNeeded(actual_len);
|
||||||
|
const addr = allocBigPages(bigpages_needed);
|
||||||
|
return @as([*]u8, @ptrFromInt(addr));
|
||||||
|
}
|
||||||
|
|
||||||
|
fn resize(
|
||||||
|
ctx: *anyopaque,
|
||||||
|
buf: []u8,
|
||||||
|
log2_buf_align: u8,
|
||||||
|
new_len: usize,
|
||||||
|
return_address: usize,
|
||||||
|
) bool {
|
||||||
|
_ = ctx;
|
||||||
|
_ = return_address;
|
||||||
|
lock.lock();
|
||||||
|
defer lock.unlock();
|
||||||
|
// We don't want to move anything from one size class to another, but we
|
||||||
|
// can recover bytes in between powers of two.
|
||||||
|
const buf_align = @as(usize, 1) << @as(Allocator.Log2Align, @intCast(log2_buf_align));
|
||||||
|
const old_actual_len = @max(buf.len + @sizeOf(usize), buf_align);
|
||||||
|
const new_actual_len = @max(new_len +| @sizeOf(usize), buf_align);
|
||||||
|
const old_small_slot_size = math.ceilPowerOfTwoAssert(usize, old_actual_len);
|
||||||
|
const old_small_class = math.log2(old_small_slot_size) - min_class;
|
||||||
|
if (old_small_class < size_class_count) {
|
||||||
|
const new_small_slot_size = math.ceilPowerOfTwo(usize, new_actual_len) catch return false;
|
||||||
|
return old_small_slot_size == new_small_slot_size;
|
||||||
|
} else {
|
||||||
|
const old_bigpages_needed = bigPagesNeeded(old_actual_len);
|
||||||
|
const old_big_slot_pages = math.ceilPowerOfTwoAssert(usize, old_bigpages_needed);
|
||||||
|
const new_bigpages_needed = bigPagesNeeded(new_actual_len);
|
||||||
|
const new_big_slot_pages = math.ceilPowerOfTwo(usize, new_bigpages_needed) catch return false;
|
||||||
|
return old_big_slot_pages == new_big_slot_pages;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn free(
|
||||||
|
ctx: *anyopaque,
|
||||||
|
buf: []u8,
|
||||||
|
log2_buf_align: u8,
|
||||||
|
return_address: usize,
|
||||||
|
) void {
|
||||||
|
_ = ctx;
|
||||||
|
_ = return_address;
|
||||||
|
lock.lock();
|
||||||
|
defer lock.unlock();
|
||||||
|
const buf_align = @as(usize, 1) << @as(Allocator.Log2Align, @intCast(log2_buf_align));
|
||||||
|
const actual_len = @max(buf.len + @sizeOf(usize), buf_align);
|
||||||
|
const slot_size = math.ceilPowerOfTwoAssert(usize, actual_len);
|
||||||
|
const class = math.log2(slot_size) - min_class;
|
||||||
|
const addr = @intFromPtr(buf.ptr);
|
||||||
|
if (class < size_class_count) {
|
||||||
|
const node = @as(*usize, @ptrFromInt(addr + (slot_size - @sizeOf(usize))));
|
||||||
|
node.* = frees[class];
|
||||||
|
frees[class] = addr;
|
||||||
|
} else {
|
||||||
|
const bigpages_needed = bigPagesNeeded(actual_len);
|
||||||
|
const pow2_pages = math.ceilPowerOfTwoAssert(usize, bigpages_needed);
|
||||||
|
const big_slot_size_bytes = pow2_pages * bigpage_size;
|
||||||
|
const node = @as(*usize, @ptrFromInt(addr + (big_slot_size_bytes - @sizeOf(usize))));
|
||||||
|
const big_class = math.log2(pow2_pages);
|
||||||
|
node.* = big_frees[big_class];
|
||||||
|
big_frees[big_class] = addr;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
inline fn bigPagesNeeded(byte_count: usize) usize {
|
||||||
|
return (byte_count + (bigpage_size + (@sizeOf(usize) - 1))) / bigpage_size;
|
||||||
|
}
|
||||||
|
|
||||||
|
fn allocBigPages(n: usize) usize {
|
||||||
|
const pow2_pages = math.ceilPowerOfTwoAssert(usize, n);
|
||||||
|
const slot_size_bytes = pow2_pages * bigpage_size;
|
||||||
|
const class = math.log2(pow2_pages);
|
||||||
|
|
||||||
|
const top_free_ptr = big_frees[class];
|
||||||
|
if (top_free_ptr != 0) {
|
||||||
|
const node = @as(*usize, @ptrFromInt(top_free_ptr + (slot_size_bytes - @sizeOf(usize))));
|
||||||
|
big_frees[class] = node.*;
|
||||||
|
return top_free_ptr;
|
||||||
|
}
|
||||||
|
return sbrk(pow2_pages * pages_per_bigpage * mem.page_size);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
}
|
||||||
@ -216,3 +216,37 @@ pub const O = struct {
|
|||||||
pub const RCLOSE = 64; // or'ed in, remove on close
|
pub const RCLOSE = 64; // or'ed in, remove on close
|
||||||
pub const EXCL = 0x1000; // or'ed in, exclusive create
|
pub const EXCL = 0x1000; // or'ed in, exclusive create
|
||||||
};
|
};
|
||||||
|
|
||||||
|
pub const ExecData = struct {
|
||||||
|
pub extern const etext: anyopaque;
|
||||||
|
pub extern const edata: anyopaque;
|
||||||
|
pub extern const end: anyopaque;
|
||||||
|
};
|
||||||
|
|
||||||
|
/// Brk sets the system's idea of the lowest bss location not
|
||||||
|
/// used by the program (called the break) to addr rounded up to
|
||||||
|
/// the next multiple of 8 bytes. Locations not less than addr
|
||||||
|
/// and below the stack pointer may cause a memory violation if
|
||||||
|
/// accessed. -9front brk(2)
|
||||||
|
pub fn brk_(addr: usize) i32 {
|
||||||
|
return @intCast(syscall_bits.syscall1(.BRK_, addr));
|
||||||
|
}
|
||||||
|
var bloc: usize = 0;
|
||||||
|
var bloc_max: usize = 0;
|
||||||
|
|
||||||
|
pub fn sbrk(n: usize) usize {
|
||||||
|
if (bloc == 0) {
|
||||||
|
// we are at the start
|
||||||
|
bloc = @intFromPtr(&ExecData.end);
|
||||||
|
bloc_max = @intFromPtr(&ExecData.end);
|
||||||
|
}
|
||||||
|
var bl = std.mem.alignForward(usize, bloc, std.mem.page_size);
|
||||||
|
const n_aligned = std.mem.alignForward(usize, n, std.mem.page_size);
|
||||||
|
if (bl + n_aligned > bloc_max) {
|
||||||
|
// we need to allocate
|
||||||
|
if (brk_(bl + n_aligned) < 0) return 0;
|
||||||
|
bloc_max = bl + n_aligned;
|
||||||
|
}
|
||||||
|
bloc = bloc + n_aligned;
|
||||||
|
return bl;
|
||||||
|
}
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user