mirror of
https://github.com/ziglang/zig.git
synced 2025-12-06 22:33:08 +00:00
These names aren't matching any formal specification; they're mostly just ripped from LLVM code. Therefore, we should definitely follow Zig naming conventions here.
591 lines
22 KiB
Zig
591 lines
22 KiB
Zig
const std = @import("../std.zig");
|
|
const File = std.fs.File;
|
|
const Allocator = std.mem.Allocator;
|
|
const pdb = std.pdb;
|
|
|
|
const Pdb = @This();
|
|
|
|
in_file: File,
|
|
msf: Msf,
|
|
allocator: Allocator,
|
|
string_table: ?*MsfStream,
|
|
dbi: ?*MsfStream,
|
|
modules: []Module,
|
|
sect_contribs: []pdb.SectionContribEntry,
|
|
guid: [16]u8,
|
|
age: u32,
|
|
|
|
pub const Module = struct {
|
|
mod_info: pdb.ModInfo,
|
|
module_name: []u8,
|
|
obj_file_name: []u8,
|
|
// The fields below are filled on demand.
|
|
populated: bool,
|
|
symbols: []u8,
|
|
subsect_info: []u8,
|
|
checksum_offset: ?usize,
|
|
|
|
pub fn deinit(self: *Module, allocator: Allocator) void {
|
|
allocator.free(self.module_name);
|
|
allocator.free(self.obj_file_name);
|
|
if (self.populated) {
|
|
allocator.free(self.symbols);
|
|
allocator.free(self.subsect_info);
|
|
}
|
|
}
|
|
};
|
|
|
|
pub fn init(allocator: Allocator, path: []const u8) !Pdb {
|
|
const file = try std.fs.cwd().openFile(path, .{});
|
|
errdefer file.close();
|
|
|
|
return .{
|
|
.in_file = file,
|
|
.allocator = allocator,
|
|
.string_table = null,
|
|
.dbi = null,
|
|
.msf = try Msf.init(allocator, file),
|
|
.modules = &[_]Module{},
|
|
.sect_contribs = &[_]pdb.SectionContribEntry{},
|
|
.guid = undefined,
|
|
.age = undefined,
|
|
};
|
|
}
|
|
|
|
pub fn deinit(self: *Pdb) void {
|
|
self.in_file.close();
|
|
self.msf.deinit(self.allocator);
|
|
for (self.modules) |*module| {
|
|
module.deinit(self.allocator);
|
|
}
|
|
self.allocator.free(self.modules);
|
|
self.allocator.free(self.sect_contribs);
|
|
}
|
|
|
|
pub fn parseDbiStream(self: *Pdb) !void {
|
|
var stream = self.getStream(pdb.StreamType.dbi) orelse
|
|
return error.InvalidDebugInfo;
|
|
const reader = stream.reader();
|
|
|
|
const header = try reader.readStruct(std.pdb.DbiStreamHeader);
|
|
if (header.version_header != 19990903) // V70, only value observed by LLVM team
|
|
return error.UnknownPDBVersion;
|
|
// if (header.Age != age)
|
|
// return error.UnmatchingPDB;
|
|
|
|
const mod_info_size = header.mod_info_size;
|
|
const section_contrib_size = header.section_contribution_size;
|
|
|
|
var modules = std.ArrayList(Module).init(self.allocator);
|
|
errdefer modules.deinit();
|
|
|
|
// Module Info Substream
|
|
var mod_info_offset: usize = 0;
|
|
while (mod_info_offset != mod_info_size) {
|
|
const mod_info = try reader.readStruct(pdb.ModInfo);
|
|
var this_record_len: usize = @sizeOf(pdb.ModInfo);
|
|
|
|
const module_name = try reader.readUntilDelimiterAlloc(self.allocator, 0, 1024);
|
|
errdefer self.allocator.free(module_name);
|
|
this_record_len += module_name.len + 1;
|
|
|
|
const obj_file_name = try reader.readUntilDelimiterAlloc(self.allocator, 0, 1024);
|
|
errdefer self.allocator.free(obj_file_name);
|
|
this_record_len += obj_file_name.len + 1;
|
|
|
|
if (this_record_len % 4 != 0) {
|
|
const round_to_next_4 = (this_record_len | 0x3) + 1;
|
|
const march_forward_bytes = round_to_next_4 - this_record_len;
|
|
try stream.seekBy(@as(isize, @intCast(march_forward_bytes)));
|
|
this_record_len += march_forward_bytes;
|
|
}
|
|
|
|
try modules.append(Module{
|
|
.mod_info = mod_info,
|
|
.module_name = module_name,
|
|
.obj_file_name = obj_file_name,
|
|
|
|
.populated = false,
|
|
.symbols = undefined,
|
|
.subsect_info = undefined,
|
|
.checksum_offset = null,
|
|
});
|
|
|
|
mod_info_offset += this_record_len;
|
|
if (mod_info_offset > mod_info_size)
|
|
return error.InvalidDebugInfo;
|
|
}
|
|
|
|
// Section Contribution Substream
|
|
var sect_contribs = std.ArrayList(pdb.SectionContribEntry).init(self.allocator);
|
|
errdefer sect_contribs.deinit();
|
|
|
|
var sect_cont_offset: usize = 0;
|
|
if (section_contrib_size != 0) {
|
|
const version = reader.readEnum(std.pdb.SectionContrSubstreamVersion, .little) catch |err| switch (err) {
|
|
error.InvalidValue => return error.InvalidDebugInfo,
|
|
else => |e| return e,
|
|
};
|
|
_ = version;
|
|
sect_cont_offset += @sizeOf(u32);
|
|
}
|
|
while (sect_cont_offset != section_contrib_size) {
|
|
const entry = try sect_contribs.addOne();
|
|
entry.* = try reader.readStruct(pdb.SectionContribEntry);
|
|
sect_cont_offset += @sizeOf(pdb.SectionContribEntry);
|
|
|
|
if (sect_cont_offset > section_contrib_size)
|
|
return error.InvalidDebugInfo;
|
|
}
|
|
|
|
self.modules = try modules.toOwnedSlice();
|
|
self.sect_contribs = try sect_contribs.toOwnedSlice();
|
|
}
|
|
|
|
pub fn parseInfoStream(self: *Pdb) !void {
|
|
var stream = self.getStream(pdb.StreamType.pdb) orelse
|
|
return error.InvalidDebugInfo;
|
|
const reader = stream.reader();
|
|
|
|
// Parse the InfoStreamHeader.
|
|
const version = try reader.readInt(u32, .little);
|
|
const signature = try reader.readInt(u32, .little);
|
|
_ = signature;
|
|
const age = try reader.readInt(u32, .little);
|
|
const guid = try reader.readBytesNoEof(16);
|
|
|
|
if (version != 20000404) // VC70, only value observed by LLVM team
|
|
return error.UnknownPDBVersion;
|
|
|
|
self.guid = guid;
|
|
self.age = age;
|
|
|
|
// Find the string table.
|
|
const string_table_index = str_tab_index: {
|
|
const name_bytes_len = try reader.readInt(u32, .little);
|
|
const name_bytes = try self.allocator.alloc(u8, name_bytes_len);
|
|
defer self.allocator.free(name_bytes);
|
|
try reader.readNoEof(name_bytes);
|
|
|
|
const HashTableHeader = extern struct {
|
|
size: u32,
|
|
capacity: u32,
|
|
|
|
fn maxLoad(cap: u32) u32 {
|
|
return cap * 2 / 3 + 1;
|
|
}
|
|
};
|
|
const hash_tbl_hdr = try reader.readStruct(HashTableHeader);
|
|
if (hash_tbl_hdr.capacity == 0)
|
|
return error.InvalidDebugInfo;
|
|
|
|
if (hash_tbl_hdr.size > HashTableHeader.maxLoad(hash_tbl_hdr.capacity))
|
|
return error.InvalidDebugInfo;
|
|
|
|
const present = try readSparseBitVector(&reader, self.allocator);
|
|
defer self.allocator.free(present);
|
|
if (present.len != hash_tbl_hdr.size)
|
|
return error.InvalidDebugInfo;
|
|
const deleted = try readSparseBitVector(&reader, self.allocator);
|
|
defer self.allocator.free(deleted);
|
|
|
|
for (present) |_| {
|
|
const name_offset = try reader.readInt(u32, .little);
|
|
const name_index = try reader.readInt(u32, .little);
|
|
if (name_offset > name_bytes.len)
|
|
return error.InvalidDebugInfo;
|
|
const name = std.mem.sliceTo(name_bytes[name_offset..], 0);
|
|
if (std.mem.eql(u8, name, "/names")) {
|
|
break :str_tab_index name_index;
|
|
}
|
|
}
|
|
return error.MissingDebugInfo;
|
|
};
|
|
|
|
self.string_table = self.getStreamById(string_table_index) orelse
|
|
return error.MissingDebugInfo;
|
|
}
|
|
|
|
pub fn getSymbolName(self: *Pdb, module: *Module, address: u64) ?[]const u8 {
|
|
_ = self;
|
|
std.debug.assert(module.populated);
|
|
|
|
var symbol_i: usize = 0;
|
|
while (symbol_i != module.symbols.len) {
|
|
const prefix: *align(1) pdb.RecordPrefix = @ptrCast(&module.symbols[symbol_i]);
|
|
if (prefix.record_len < 2)
|
|
return null;
|
|
switch (prefix.record_kind) {
|
|
.lproc32, .gproc32 => {
|
|
const proc_sym: *align(1) pdb.ProcSym = @ptrCast(&module.symbols[symbol_i + @sizeOf(pdb.RecordPrefix)]);
|
|
if (address >= proc_sym.code_offset and address < proc_sym.code_offset + proc_sym.code_size) {
|
|
return std.mem.sliceTo(@as([*:0]u8, @ptrCast(&proc_sym.name[0])), 0);
|
|
}
|
|
},
|
|
else => {},
|
|
}
|
|
symbol_i += prefix.record_len + @sizeOf(u16);
|
|
}
|
|
|
|
return null;
|
|
}
|
|
|
|
pub fn getLineNumberInfo(self: *Pdb, module: *Module, address: u64) !std.debug.SourceLocation {
|
|
std.debug.assert(module.populated);
|
|
const subsect_info = module.subsect_info;
|
|
|
|
var sect_offset: usize = 0;
|
|
var skip_len: usize = undefined;
|
|
const checksum_offset = module.checksum_offset orelse return error.MissingDebugInfo;
|
|
while (sect_offset != subsect_info.len) : (sect_offset += skip_len) {
|
|
const subsect_hdr: *align(1) pdb.DebugSubsectionHeader = @ptrCast(&subsect_info[sect_offset]);
|
|
skip_len = subsect_hdr.length;
|
|
sect_offset += @sizeOf(pdb.DebugSubsectionHeader);
|
|
|
|
switch (subsect_hdr.kind) {
|
|
.lines => {
|
|
var line_index = sect_offset;
|
|
|
|
const line_hdr: *align(1) pdb.LineFragmentHeader = @ptrCast(&subsect_info[line_index]);
|
|
if (line_hdr.reloc_segment == 0)
|
|
return error.MissingDebugInfo;
|
|
line_index += @sizeOf(pdb.LineFragmentHeader);
|
|
const frag_vaddr_start = line_hdr.reloc_offset;
|
|
const frag_vaddr_end = frag_vaddr_start + line_hdr.code_size;
|
|
|
|
if (address >= frag_vaddr_start and address < frag_vaddr_end) {
|
|
// There is an unknown number of LineBlockFragmentHeaders (and their accompanying line and column records)
|
|
// from now on. We will iterate through them, and eventually find a SourceLocation that we're interested in,
|
|
// breaking out to :subsections. If not, we will make sure to not read anything outside of this subsection.
|
|
const subsection_end_index = sect_offset + subsect_hdr.length;
|
|
|
|
while (line_index < subsection_end_index) {
|
|
const block_hdr: *align(1) pdb.LineBlockFragmentHeader = @ptrCast(&subsect_info[line_index]);
|
|
line_index += @sizeOf(pdb.LineBlockFragmentHeader);
|
|
const start_line_index = line_index;
|
|
|
|
const has_column = line_hdr.flags.have_columns;
|
|
|
|
// All line entries are stored inside their line block by ascending start address.
|
|
// Heuristic: we want to find the last line entry
|
|
// that has a vaddr_start <= address.
|
|
// This is done with a simple linear search.
|
|
var line_i: u32 = 0;
|
|
while (line_i < block_hdr.num_lines) : (line_i += 1) {
|
|
const line_num_entry: *align(1) pdb.LineNumberEntry = @ptrCast(&subsect_info[line_index]);
|
|
line_index += @sizeOf(pdb.LineNumberEntry);
|
|
|
|
const vaddr_start = frag_vaddr_start + line_num_entry.offset;
|
|
if (address < vaddr_start) {
|
|
break;
|
|
}
|
|
}
|
|
|
|
// line_i == 0 would mean that no matching pdb.LineNumberEntry was found.
|
|
if (line_i > 0) {
|
|
const subsect_index = checksum_offset + block_hdr.name_index;
|
|
const chksum_hdr: *align(1) pdb.FileChecksumEntryHeader = @ptrCast(&module.subsect_info[subsect_index]);
|
|
const strtab_offset = @sizeOf(pdb.StringTableHeader) + chksum_hdr.file_name_offset;
|
|
try self.string_table.?.seekTo(strtab_offset);
|
|
const source_file_name = try self.string_table.?.reader().readUntilDelimiterAlloc(self.allocator, 0, 1024);
|
|
|
|
const line_entry_idx = line_i - 1;
|
|
|
|
const column = if (has_column) blk: {
|
|
const start_col_index = start_line_index + @sizeOf(pdb.LineNumberEntry) * block_hdr.num_lines;
|
|
const col_index = start_col_index + @sizeOf(pdb.ColumnNumberEntry) * line_entry_idx;
|
|
const col_num_entry: *align(1) pdb.ColumnNumberEntry = @ptrCast(&subsect_info[col_index]);
|
|
break :blk col_num_entry.start_column;
|
|
} else 0;
|
|
|
|
const found_line_index = start_line_index + line_entry_idx * @sizeOf(pdb.LineNumberEntry);
|
|
const line_num_entry: *align(1) pdb.LineNumberEntry = @ptrCast(&subsect_info[found_line_index]);
|
|
|
|
return .{
|
|
.file_name = source_file_name,
|
|
.line = line_num_entry.flags.start,
|
|
.column = column,
|
|
};
|
|
}
|
|
}
|
|
|
|
// Checking that we are not reading garbage after the (possibly) multiple block fragments.
|
|
if (line_index != subsection_end_index) {
|
|
return error.InvalidDebugInfo;
|
|
}
|
|
}
|
|
},
|
|
else => {},
|
|
}
|
|
|
|
if (sect_offset > subsect_info.len)
|
|
return error.InvalidDebugInfo;
|
|
}
|
|
|
|
return error.MissingDebugInfo;
|
|
}
|
|
|
|
pub fn getModule(self: *Pdb, index: usize) !?*Module {
|
|
if (index >= self.modules.len)
|
|
return null;
|
|
|
|
const mod = &self.modules[index];
|
|
if (mod.populated)
|
|
return mod;
|
|
|
|
// At most one can be non-zero.
|
|
if (mod.mod_info.c11_byte_size != 0 and mod.mod_info.c13_byte_size != 0)
|
|
return error.InvalidDebugInfo;
|
|
if (mod.mod_info.c13_byte_size == 0)
|
|
return error.InvalidDebugInfo;
|
|
|
|
const stream = self.getStreamById(mod.mod_info.module_sym_stream) orelse
|
|
return error.MissingDebugInfo;
|
|
const reader = stream.reader();
|
|
|
|
const signature = try reader.readInt(u32, .little);
|
|
if (signature != 4)
|
|
return error.InvalidDebugInfo;
|
|
|
|
mod.symbols = try self.allocator.alloc(u8, mod.mod_info.sym_byte_size - 4);
|
|
errdefer self.allocator.free(mod.symbols);
|
|
try reader.readNoEof(mod.symbols);
|
|
|
|
mod.subsect_info = try self.allocator.alloc(u8, mod.mod_info.c13_byte_size);
|
|
errdefer self.allocator.free(mod.subsect_info);
|
|
try reader.readNoEof(mod.subsect_info);
|
|
|
|
var sect_offset: usize = 0;
|
|
var skip_len: usize = undefined;
|
|
while (sect_offset != mod.subsect_info.len) : (sect_offset += skip_len) {
|
|
const subsect_hdr: *align(1) pdb.DebugSubsectionHeader = @ptrCast(&mod.subsect_info[sect_offset]);
|
|
skip_len = subsect_hdr.length;
|
|
sect_offset += @sizeOf(pdb.DebugSubsectionHeader);
|
|
|
|
switch (subsect_hdr.kind) {
|
|
.file_checksums => {
|
|
mod.checksum_offset = sect_offset;
|
|
break;
|
|
},
|
|
else => {},
|
|
}
|
|
|
|
if (sect_offset > mod.subsect_info.len)
|
|
return error.InvalidDebugInfo;
|
|
}
|
|
|
|
mod.populated = true;
|
|
return mod;
|
|
}
|
|
|
|
pub fn getStreamById(self: *Pdb, id: u32) ?*MsfStream {
|
|
if (id >= self.msf.streams.len)
|
|
return null;
|
|
return &self.msf.streams[id];
|
|
}
|
|
|
|
pub fn getStream(self: *Pdb, stream: pdb.StreamType) ?*MsfStream {
|
|
const id = @intFromEnum(stream);
|
|
return self.getStreamById(id);
|
|
}
|
|
|
|
/// https://llvm.org/docs/PDB/MsfFile.html
|
|
const Msf = struct {
|
|
directory: MsfStream,
|
|
streams: []MsfStream,
|
|
|
|
fn init(allocator: Allocator, file: File) !Msf {
|
|
const in = file.reader();
|
|
|
|
const superblock = try in.readStruct(pdb.SuperBlock);
|
|
|
|
// Sanity checks
|
|
if (!std.mem.eql(u8, &superblock.file_magic, pdb.SuperBlock.expect_magic))
|
|
return error.InvalidDebugInfo;
|
|
if (superblock.free_block_map_block != 1 and superblock.free_block_map_block != 2)
|
|
return error.InvalidDebugInfo;
|
|
const file_len = try file.getEndPos();
|
|
if (superblock.num_blocks * superblock.block_size != file_len)
|
|
return error.InvalidDebugInfo;
|
|
switch (superblock.block_size) {
|
|
// llvm only supports 4096 but we can handle any of these values
|
|
512, 1024, 2048, 4096 => {},
|
|
else => return error.InvalidDebugInfo,
|
|
}
|
|
|
|
const dir_block_count = blockCountFromSize(superblock.num_directory_bytes, superblock.block_size);
|
|
if (dir_block_count > superblock.block_size / @sizeOf(u32))
|
|
return error.UnhandledBigDirectoryStream; // cf. BlockMapAddr comment.
|
|
|
|
try file.seekTo(superblock.block_size * superblock.block_map_addr);
|
|
const dir_blocks = try allocator.alloc(u32, dir_block_count);
|
|
for (dir_blocks) |*b| {
|
|
b.* = try in.readInt(u32, .little);
|
|
}
|
|
var directory = MsfStream.init(
|
|
superblock.block_size,
|
|
file,
|
|
dir_blocks,
|
|
);
|
|
|
|
const begin = directory.pos;
|
|
const stream_count = try directory.reader().readInt(u32, .little);
|
|
const stream_sizes = try allocator.alloc(u32, stream_count);
|
|
defer allocator.free(stream_sizes);
|
|
|
|
// Microsoft's implementation uses @as(u32, -1) for inexistent streams.
|
|
// These streams are not used, but still participate in the file
|
|
// and must be taken into account when resolving stream indices.
|
|
const Nil = 0xFFFFFFFF;
|
|
for (stream_sizes) |*s| {
|
|
const size = try directory.reader().readInt(u32, .little);
|
|
s.* = if (size == Nil) 0 else blockCountFromSize(size, superblock.block_size);
|
|
}
|
|
|
|
const streams = try allocator.alloc(MsfStream, stream_count);
|
|
for (streams, 0..) |*stream, i| {
|
|
const size = stream_sizes[i];
|
|
if (size == 0) {
|
|
stream.* = MsfStream{
|
|
.blocks = &[_]u32{},
|
|
};
|
|
} else {
|
|
var blocks = try allocator.alloc(u32, size);
|
|
var j: u32 = 0;
|
|
while (j < size) : (j += 1) {
|
|
const block_id = try directory.reader().readInt(u32, .little);
|
|
const n = (block_id % superblock.block_size);
|
|
// 0 is for pdb.SuperBlock, 1 and 2 for FPMs.
|
|
if (block_id == 0 or n == 1 or n == 2 or block_id * superblock.block_size > file_len)
|
|
return error.InvalidBlockIndex;
|
|
blocks[j] = block_id;
|
|
}
|
|
|
|
stream.* = MsfStream.init(
|
|
superblock.block_size,
|
|
file,
|
|
blocks,
|
|
);
|
|
}
|
|
}
|
|
|
|
const end = directory.pos;
|
|
if (end - begin != superblock.num_directory_bytes)
|
|
return error.InvalidStreamDirectory;
|
|
|
|
return Msf{
|
|
.directory = directory,
|
|
.streams = streams,
|
|
};
|
|
}
|
|
|
|
fn deinit(self: *Msf, allocator: Allocator) void {
|
|
allocator.free(self.directory.blocks);
|
|
for (self.streams) |*stream| {
|
|
allocator.free(stream.blocks);
|
|
}
|
|
allocator.free(self.streams);
|
|
}
|
|
};
|
|
|
|
const MsfStream = struct {
|
|
in_file: File = undefined,
|
|
pos: u64 = undefined,
|
|
blocks: []u32 = undefined,
|
|
block_size: u32 = undefined,
|
|
|
|
pub const Error = @typeInfo(@typeInfo(@TypeOf(read)).@"fn".return_type.?).error_union.error_set;
|
|
|
|
fn init(block_size: u32, file: File, blocks: []u32) MsfStream {
|
|
const stream = MsfStream{
|
|
.in_file = file,
|
|
.pos = 0,
|
|
.blocks = blocks,
|
|
.block_size = block_size,
|
|
};
|
|
|
|
return stream;
|
|
}
|
|
|
|
fn read(self: *MsfStream, buffer: []u8) !usize {
|
|
var block_id = @as(usize, @intCast(self.pos / self.block_size));
|
|
if (block_id >= self.blocks.len) return 0; // End of Stream
|
|
var block = self.blocks[block_id];
|
|
var offset = self.pos % self.block_size;
|
|
|
|
try self.in_file.seekTo(block * self.block_size + offset);
|
|
const in = self.in_file.reader();
|
|
|
|
var size: usize = 0;
|
|
var rem_buffer = buffer;
|
|
while (size < buffer.len) {
|
|
const size_to_read = @min(self.block_size - offset, rem_buffer.len);
|
|
size += try in.read(rem_buffer[0..size_to_read]);
|
|
rem_buffer = buffer[size..];
|
|
offset += size_to_read;
|
|
|
|
// If we're at the end of a block, go to the next one.
|
|
if (offset == self.block_size) {
|
|
offset = 0;
|
|
block_id += 1;
|
|
if (block_id >= self.blocks.len) break; // End of Stream
|
|
block = self.blocks[block_id];
|
|
try self.in_file.seekTo(block * self.block_size);
|
|
}
|
|
}
|
|
|
|
self.pos += buffer.len;
|
|
return buffer.len;
|
|
}
|
|
|
|
pub fn seekBy(self: *MsfStream, len: i64) !void {
|
|
self.pos = @as(u64, @intCast(@as(i64, @intCast(self.pos)) + len));
|
|
if (self.pos >= self.blocks.len * self.block_size)
|
|
return error.EOF;
|
|
}
|
|
|
|
pub fn seekTo(self: *MsfStream, len: u64) !void {
|
|
self.pos = len;
|
|
if (self.pos >= self.blocks.len * self.block_size)
|
|
return error.EOF;
|
|
}
|
|
|
|
fn getSize(self: *const MsfStream) u64 {
|
|
return self.blocks.len * self.block_size;
|
|
}
|
|
|
|
fn getFilePos(self: MsfStream) u64 {
|
|
const block_id = self.pos / self.block_size;
|
|
const block = self.blocks[block_id];
|
|
const offset = self.pos % self.block_size;
|
|
|
|
return block * self.block_size + offset;
|
|
}
|
|
|
|
pub fn reader(self: *MsfStream) std.io.Reader(*MsfStream, Error, read) {
|
|
return .{ .context = self };
|
|
}
|
|
};
|
|
|
|
fn readSparseBitVector(stream: anytype, allocator: Allocator) ![]u32 {
|
|
const num_words = try stream.readInt(u32, .little);
|
|
var list = std.ArrayList(u32).init(allocator);
|
|
errdefer list.deinit();
|
|
var word_i: u32 = 0;
|
|
while (word_i != num_words) : (word_i += 1) {
|
|
const word = try stream.readInt(u32, .little);
|
|
var bit_i: u5 = 0;
|
|
while (true) : (bit_i += 1) {
|
|
if (word & (@as(u32, 1) << bit_i) != 0) {
|
|
try list.append(word_i * 32 + bit_i);
|
|
}
|
|
if (bit_i == std.math.maxInt(u5)) break;
|
|
}
|
|
}
|
|
return try list.toOwnedSlice();
|
|
}
|
|
|
|
fn blockCountFromSize(size: u32, block_size: u32) u32 {
|
|
return (size + block_size - 1) / block_size;
|
|
}
|