stage2: simplify Decl src_node field

Also fix "previous definition here" error notes to be correct.
This commit is contained in:
Andrew Kelley 2021-04-08 20:37:19 -07:00
parent 482b995a49
commit 61b868f9a5
6 changed files with 68 additions and 103 deletions

View File

@ -150,9 +150,15 @@ pub const Decl = struct {
/// The direct parent container of the Decl. /// The direct parent container of the Decl.
/// Reference to externally owned memory. /// Reference to externally owned memory.
container: *Scope.Container, container: *Scope.Container,
/// The AST Node decl index or ZIR Inst index that contains this declaration.
/// An integer that can be checked against the corresponding incrementing
/// generation field of Module. This is used to determine whether `complete` status
/// represents pre- or post- re-analysis.
generation: u32,
/// The AST Node index or ZIR Inst index that contains this declaration.
/// Must be recomputed when the corresponding source file is modified. /// Must be recomputed when the corresponding source file is modified.
src_index: usize, src_node: ast.Node.Index,
/// The most recent value of the Decl after a successful semantic analysis. /// The most recent value of the Decl after a successful semantic analysis.
typed_value: union(enum) { typed_value: union(enum) {
never_succeeded: void, never_succeeded: void,
@ -198,11 +204,6 @@ pub const Decl = struct {
/// Whether the corresponding AST decl has a `pub` keyword. /// Whether the corresponding AST decl has a `pub` keyword.
is_pub: bool, is_pub: bool,
/// An integer that can be checked against the corresponding incrementing
/// generation field of Module. This is used to determine whether `complete` status
/// represents pre- or post- re-analysis.
generation: u32,
/// Represents the position of the code in the output file. /// Represents the position of the code in the output file.
/// This is populated regardless of semantic analysis and code generation. /// This is populated regardless of semantic analysis and code generation.
link: link.File.LinkBlock, link: link.File.LinkBlock,
@ -249,11 +250,11 @@ pub const Decl = struct {
} }
pub fn relativeToNodeIndex(decl: Decl, offset: i32) ast.Node.Index { pub fn relativeToNodeIndex(decl: Decl, offset: i32) ast.Node.Index {
return @bitCast(ast.Node.Index, offset + @bitCast(i32, decl.srcNode())); return @bitCast(ast.Node.Index, offset + @bitCast(i32, decl.src_node));
} }
pub fn nodeIndexToRelative(decl: Decl, node_index: ast.Node.Index) i32 { pub fn nodeIndexToRelative(decl: Decl, node_index: ast.Node.Index) i32 {
return @bitCast(i32, node_index) - @bitCast(i32, decl.srcNode()); return @bitCast(i32, node_index) - @bitCast(i32, decl.src_node);
} }
pub fn tokSrcLoc(decl: Decl, token_index: ast.TokenIndex) LazySrcLoc { pub fn tokSrcLoc(decl: Decl, token_index: ast.TokenIndex) LazySrcLoc {
@ -271,14 +272,9 @@ pub const Decl = struct {
}; };
} }
pub fn srcNode(decl: Decl) u32 {
const tree = &decl.container.file_scope.tree;
return tree.rootDecls()[decl.src_index];
}
pub fn srcToken(decl: Decl) u32 { pub fn srcToken(decl: Decl) u32 {
const tree = &decl.container.file_scope.tree; const tree = &decl.container.file_scope.tree;
return tree.firstToken(decl.srcNode()); return tree.firstToken(decl.src_node);
} }
pub fn srcByteOffset(decl: Decl) u32 { pub fn srcByteOffset(decl: Decl) u32 {
@ -2458,7 +2454,7 @@ fn astgenAndSemaDecl(mod: *Module, decl: *Decl) !bool {
const tree = try mod.getAstTree(decl.container.file_scope); const tree = try mod.getAstTree(decl.container.file_scope);
const node_tags = tree.nodes.items(.tag); const node_tags = tree.nodes.items(.tag);
const node_datas = tree.nodes.items(.data); const node_datas = tree.nodes.items(.data);
const decl_node = tree.rootDecls()[decl.src_index]; const decl_node = decl.src_node;
switch (node_tags[decl_node]) { switch (node_tags[decl_node]) {
.fn_decl => { .fn_decl => {
const fn_proto = node_datas[decl_node].lhs; const fn_proto = node_datas[decl_node].lhs;
@ -3292,7 +3288,7 @@ pub fn analyzeContainer(mod: *Module, container_scope: *Scope.Container) !void {
var outdated_decls = std.AutoArrayHashMap(*Decl, void).init(mod.gpa); var outdated_decls = std.AutoArrayHashMap(*Decl, void).init(mod.gpa);
defer outdated_decls.deinit(); defer outdated_decls.deinit();
for (decls) |decl_node, decl_i| switch (node_tags[decl_node]) { for (decls) |decl_node| switch (node_tags[decl_node]) {
.fn_decl => { .fn_decl => {
const fn_proto = node_datas[decl_node].lhs; const fn_proto = node_datas[decl_node].lhs;
const body = node_datas[decl_node].rhs; const body = node_datas[decl_node].rhs;
@ -3304,7 +3300,6 @@ pub fn analyzeContainer(mod: *Module, container_scope: *Scope.Container) !void {
&deleted_decls, &deleted_decls,
&outdated_decls, &outdated_decls,
decl_node, decl_node,
decl_i,
tree.*, tree.*,
body, body,
tree.fnProtoSimple(&params, fn_proto), tree.fnProtoSimple(&params, fn_proto),
@ -3315,7 +3310,6 @@ pub fn analyzeContainer(mod: *Module, container_scope: *Scope.Container) !void {
&deleted_decls, &deleted_decls,
&outdated_decls, &outdated_decls,
decl_node, decl_node,
decl_i,
tree.*, tree.*,
body, body,
tree.fnProtoMulti(fn_proto), tree.fnProtoMulti(fn_proto),
@ -3327,7 +3321,6 @@ pub fn analyzeContainer(mod: *Module, container_scope: *Scope.Container) !void {
&deleted_decls, &deleted_decls,
&outdated_decls, &outdated_decls,
decl_node, decl_node,
decl_i,
tree.*, tree.*,
body, body,
tree.fnProtoOne(&params, fn_proto), tree.fnProtoOne(&params, fn_proto),
@ -3338,7 +3331,6 @@ pub fn analyzeContainer(mod: *Module, container_scope: *Scope.Container) !void {
&deleted_decls, &deleted_decls,
&outdated_decls, &outdated_decls,
decl_node, decl_node,
decl_i,
tree.*, tree.*,
body, body,
tree.fnProto(fn_proto), tree.fnProto(fn_proto),
@ -3353,7 +3345,6 @@ pub fn analyzeContainer(mod: *Module, container_scope: *Scope.Container) !void {
&deleted_decls, &deleted_decls,
&outdated_decls, &outdated_decls,
decl_node, decl_node,
decl_i,
tree.*, tree.*,
0, 0,
tree.fnProtoSimple(&params, decl_node), tree.fnProtoSimple(&params, decl_node),
@ -3364,7 +3355,6 @@ pub fn analyzeContainer(mod: *Module, container_scope: *Scope.Container) !void {
&deleted_decls, &deleted_decls,
&outdated_decls, &outdated_decls,
decl_node, decl_node,
decl_i,
tree.*, tree.*,
0, 0,
tree.fnProtoMulti(decl_node), tree.fnProtoMulti(decl_node),
@ -3376,7 +3366,6 @@ pub fn analyzeContainer(mod: *Module, container_scope: *Scope.Container) !void {
&deleted_decls, &deleted_decls,
&outdated_decls, &outdated_decls,
decl_node, decl_node,
decl_i,
tree.*, tree.*,
0, 0,
tree.fnProtoOne(&params, decl_node), tree.fnProtoOne(&params, decl_node),
@ -3387,7 +3376,6 @@ pub fn analyzeContainer(mod: *Module, container_scope: *Scope.Container) !void {
&deleted_decls, &deleted_decls,
&outdated_decls, &outdated_decls,
decl_node, decl_node,
decl_i,
tree.*, tree.*,
0, 0,
tree.fnProto(decl_node), tree.fnProto(decl_node),
@ -3398,7 +3386,6 @@ pub fn analyzeContainer(mod: *Module, container_scope: *Scope.Container) !void {
&deleted_decls, &deleted_decls,
&outdated_decls, &outdated_decls,
decl_node, decl_node,
decl_i,
tree.*, tree.*,
tree.globalVarDecl(decl_node), tree.globalVarDecl(decl_node),
), ),
@ -3407,7 +3394,6 @@ pub fn analyzeContainer(mod: *Module, container_scope: *Scope.Container) !void {
&deleted_decls, &deleted_decls,
&outdated_decls, &outdated_decls,
decl_node, decl_node,
decl_i,
tree.*, tree.*,
tree.localVarDecl(decl_node), tree.localVarDecl(decl_node),
), ),
@ -3416,7 +3402,6 @@ pub fn analyzeContainer(mod: *Module, container_scope: *Scope.Container) !void {
&deleted_decls, &deleted_decls,
&outdated_decls, &outdated_decls,
decl_node, decl_node,
decl_i,
tree.*, tree.*,
tree.simpleVarDecl(decl_node), tree.simpleVarDecl(decl_node),
), ),
@ -3425,7 +3410,6 @@ pub fn analyzeContainer(mod: *Module, container_scope: *Scope.Container) !void {
&deleted_decls, &deleted_decls,
&outdated_decls, &outdated_decls,
decl_node, decl_node,
decl_i,
tree.*, tree.*,
tree.alignedVarDecl(decl_node), tree.alignedVarDecl(decl_node),
), ),
@ -3438,35 +3422,16 @@ pub fn analyzeContainer(mod: *Module, container_scope: *Scope.Container) !void {
const name_hash = container_scope.fullyQualifiedNameHash(name); const name_hash = container_scope.fullyQualifiedNameHash(name);
const contents_hash = std.zig.hashSrc(tree.getNodeSource(decl_node)); const contents_hash = std.zig.hashSrc(tree.getNodeSource(decl_node));
const new_decl = try mod.createNewDecl(&container_scope.base, name, decl_i, name_hash, contents_hash); const new_decl = try mod.createNewDecl(&container_scope.base, name, decl_node, name_hash, contents_hash);
container_scope.decls.putAssumeCapacity(new_decl, {}); container_scope.decls.putAssumeCapacity(new_decl, {});
mod.comp.work_queue.writeItemAssumeCapacity(.{ .analyze_decl = new_decl }); mod.comp.work_queue.writeItemAssumeCapacity(.{ .analyze_decl = new_decl });
}, },
.container_field_init => try mod.semaContainerField( // Container fields are handled in AstGen.
container_scope, .container_field_init,
&deleted_decls, .container_field_align,
decl_node, .container_field,
decl_i, => continue,
tree.*,
tree.containerFieldInit(decl_node),
),
.container_field_align => try mod.semaContainerField(
container_scope,
&deleted_decls,
decl_node,
decl_i,
tree.*,
tree.containerFieldAlign(decl_node),
),
.container_field => try mod.semaContainerField(
container_scope,
&deleted_decls,
decl_node,
decl_i,
tree.*,
tree.containerField(decl_node),
),
.test_decl => { .test_decl => {
if (mod.comp.bin_file.options.is_test) { if (mod.comp.bin_file.options.is_test) {
@ -3508,7 +3473,6 @@ fn semaContainerFn(
deleted_decls: *std.AutoArrayHashMap(*Decl, void), deleted_decls: *std.AutoArrayHashMap(*Decl, void),
outdated_decls: *std.AutoArrayHashMap(*Decl, void), outdated_decls: *std.AutoArrayHashMap(*Decl, void),
decl_node: ast.Node.Index, decl_node: ast.Node.Index,
decl_i: usize,
tree: ast.Tree, tree: ast.Tree,
body_node: ast.Node.Index, body_node: ast.Node.Index,
fn_proto: ast.full.FnProto, fn_proto: ast.full.FnProto,
@ -3517,25 +3481,30 @@ fn semaContainerFn(
defer tracy.end(); defer tracy.end();
// We will create a Decl for it regardless of analysis status. // We will create a Decl for it regardless of analysis status.
const name_tok = fn_proto.name_token orelse { const name_token = fn_proto.name_token orelse {
// This problem will go away with #1717. // This problem will go away with #1717.
@panic("TODO missing function name"); @panic("TODO missing function name");
}; };
const name = tree.tokenSlice(name_tok); // TODO use identifierTokenString const name = tree.tokenSlice(name_token); // TODO use identifierTokenString
const name_hash = container_scope.fullyQualifiedNameHash(name); const name_hash = container_scope.fullyQualifiedNameHash(name);
const contents_hash = std.zig.hashSrc(tree.getNodeSource(decl_node)); const contents_hash = std.zig.hashSrc(tree.getNodeSource(decl_node));
if (mod.decl_table.get(name_hash)) |decl| { if (mod.decl_table.get(name_hash)) |decl| {
// Update the AST Node index of the decl, even if its contents are unchanged, it may // Update the AST Node index of the decl, even if its contents are unchanged, it may
// have been re-ordered. // have been re-ordered.
decl.src_index = decl_i; const prev_src_node = decl.src_node;
decl.src_node = decl_node;
if (deleted_decls.swapRemove(decl) == null) { if (deleted_decls.swapRemove(decl) == null) {
decl.analysis = .sema_failure; decl.analysis = .sema_failure;
const msg = try ErrorMsg.create(mod.gpa, .{ const msg = try ErrorMsg.create(mod.gpa, .{
.container = .{ .file_scope = container_scope.file_scope }, .container = .{ .file_scope = container_scope.file_scope },
.lazy = .{ .token_abs = name_tok }, .lazy = .{ .token_abs = name_token },
}, "redefinition of '{s}'", .{decl.name}); }, "redefinition of '{s}'", .{decl.name});
errdefer msg.destroy(mod.gpa); errdefer msg.destroy(mod.gpa);
try mod.errNoteNonLazy(decl.srcLoc(), msg, "previous definition here", .{}); const other_src_loc: SrcLoc = .{
.container = .{ .file_scope = decl.container.file_scope },
.lazy = .{ .node_abs = prev_src_node },
};
try mod.errNoteNonLazy(other_src_loc, msg, "previous definition here", .{});
try mod.failed_decls.putNoClobber(mod.gpa, decl, msg); try mod.failed_decls.putNoClobber(mod.gpa, decl, msg);
} else { } else {
if (!srcHashEql(decl.contents_hash, contents_hash)) { if (!srcHashEql(decl.contents_hash, contents_hash)) {
@ -3559,7 +3528,7 @@ fn semaContainerFn(
} }
} }
} else { } else {
const new_decl = try mod.createNewDecl(&container_scope.base, name, decl_i, name_hash, contents_hash); const new_decl = try mod.createNewDecl(&container_scope.base, name, decl_node, name_hash, contents_hash);
container_scope.decls.putAssumeCapacity(new_decl, {}); container_scope.decls.putAssumeCapacity(new_decl, {});
if (fn_proto.extern_export_token) |maybe_export_token| { if (fn_proto.extern_export_token) |maybe_export_token| {
const token_tags = tree.tokens.items(.tag); const token_tags = tree.tokens.items(.tag);
@ -3576,7 +3545,6 @@ fn semaContainerVar(
deleted_decls: *std.AutoArrayHashMap(*Decl, void), deleted_decls: *std.AutoArrayHashMap(*Decl, void),
outdated_decls: *std.AutoArrayHashMap(*Decl, void), outdated_decls: *std.AutoArrayHashMap(*Decl, void),
decl_node: ast.Node.Index, decl_node: ast.Node.Index,
decl_i: usize,
tree: ast.Tree, tree: ast.Tree,
var_decl: ast.full.VarDecl, var_decl: ast.full.VarDecl,
) !void { ) !void {
@ -3590,7 +3558,8 @@ fn semaContainerVar(
if (mod.decl_table.get(name_hash)) |decl| { if (mod.decl_table.get(name_hash)) |decl| {
// Update the AST Node index of the decl, even if its contents are unchanged, it may // Update the AST Node index of the decl, even if its contents are unchanged, it may
// have been re-ordered. // have been re-ordered.
decl.src_index = decl_i; const prev_src_node = decl.src_node;
decl.src_node = decl_node;
if (deleted_decls.swapRemove(decl) == null) { if (deleted_decls.swapRemove(decl) == null) {
decl.analysis = .sema_failure; decl.analysis = .sema_failure;
const msg = try ErrorMsg.create(mod.gpa, .{ const msg = try ErrorMsg.create(mod.gpa, .{
@ -3598,14 +3567,18 @@ fn semaContainerVar(
.lazy = .{ .token_abs = name_token }, .lazy = .{ .token_abs = name_token },
}, "redefinition of '{s}'", .{decl.name}); }, "redefinition of '{s}'", .{decl.name});
errdefer msg.destroy(mod.gpa); errdefer msg.destroy(mod.gpa);
try mod.errNoteNonLazy(decl.srcLoc(), msg, "previous definition here", .{}); const other_src_loc: SrcLoc = .{
.container = .{ .file_scope = decl.container.file_scope },
.lazy = .{ .node_abs = prev_src_node },
};
try mod.errNoteNonLazy(other_src_loc, msg, "previous definition here", .{});
try mod.failed_decls.putNoClobber(mod.gpa, decl, msg); try mod.failed_decls.putNoClobber(mod.gpa, decl, msg);
} else if (!srcHashEql(decl.contents_hash, contents_hash)) { } else if (!srcHashEql(decl.contents_hash, contents_hash)) {
try outdated_decls.put(decl, {}); try outdated_decls.put(decl, {});
decl.contents_hash = contents_hash; decl.contents_hash = contents_hash;
} }
} else { } else {
const new_decl = try mod.createNewDecl(&container_scope.base, name, decl_i, name_hash, contents_hash); const new_decl = try mod.createNewDecl(&container_scope.base, name, decl_node, name_hash, contents_hash);
container_scope.decls.putAssumeCapacity(new_decl, {}); container_scope.decls.putAssumeCapacity(new_decl, {});
if (var_decl.extern_export_token) |maybe_export_token| { if (var_decl.extern_export_token) |maybe_export_token| {
const token_tags = tree.tokens.items(.tag); const token_tags = tree.tokens.items(.tag);
@ -3616,21 +3589,6 @@ fn semaContainerVar(
} }
} }
fn semaContainerField(
mod: *Module,
container_scope: *Scope.Container,
deleted_decls: *std.AutoArrayHashMap(*Decl, void),
decl_node: ast.Node.Index,
decl_i: usize,
tree: ast.Tree,
field: ast.full.ContainerField,
) !void {
const tracy = trace(@src());
defer tracy.end();
log.err("TODO: analyze container field", .{});
}
pub fn deleteDecl( pub fn deleteDecl(
mod: *Module, mod: *Module,
decl: *Decl, decl: *Decl,
@ -3813,7 +3771,7 @@ fn markOutdatedDecl(mod: *Module, decl: *Decl) !void {
fn allocateNewDecl( fn allocateNewDecl(
mod: *Module, mod: *Module,
scope: *Scope, scope: *Scope,
src_index: usize, src_node: ast.Node.Index,
contents_hash: std.zig.SrcHash, contents_hash: std.zig.SrcHash,
) !*Decl { ) !*Decl {
// If we have emit-h then we must allocate a bigger structure to store the emit-h state. // If we have emit-h then we must allocate a bigger structure to store the emit-h state.
@ -3829,7 +3787,7 @@ fn allocateNewDecl(
new_decl.* = .{ new_decl.* = .{
.name = "", .name = "",
.container = scope.namespace(), .container = scope.namespace(),
.src_index = src_index, .src_node = src_node,
.typed_value = .{ .never_succeeded = {} }, .typed_value = .{ .never_succeeded = {} },
.analysis = .unreferenced, .analysis = .unreferenced,
.deletion_flag = false, .deletion_flag = false,
@ -3860,12 +3818,12 @@ fn createNewDecl(
mod: *Module, mod: *Module,
scope: *Scope, scope: *Scope,
decl_name: []const u8, decl_name: []const u8,
src_index: usize, src_node: ast.Node.Index,
name_hash: Scope.NameHash, name_hash: Scope.NameHash,
contents_hash: std.zig.SrcHash, contents_hash: std.zig.SrcHash,
) !*Decl { ) !*Decl {
try mod.decl_table.ensureCapacity(mod.gpa, mod.decl_table.items().len + 1); try mod.decl_table.ensureCapacity(mod.gpa, mod.decl_table.items().len + 1);
const new_decl = try mod.allocateNewDecl(scope, src_index, contents_hash); const new_decl = try mod.allocateNewDecl(scope, src_node, contents_hash);
errdefer mod.gpa.destroy(new_decl); errdefer mod.gpa.destroy(new_decl);
new_decl.name = try mem.dupeZ(mod.gpa, u8, decl_name); new_decl.name = try mem.dupeZ(mod.gpa, u8, decl_name);
mod.decl_table.putAssumeCapacityNoClobber(name_hash, new_decl); mod.decl_table.putAssumeCapacityNoClobber(name_hash, new_decl);
@ -4078,7 +4036,7 @@ pub fn createAnonymousDecl(
defer mod.gpa.free(name); defer mod.gpa.free(name);
const name_hash = scope.namespace().fullyQualifiedNameHash(name); const name_hash = scope.namespace().fullyQualifiedNameHash(name);
const src_hash: std.zig.SrcHash = undefined; const src_hash: std.zig.SrcHash = undefined;
const new_decl = try mod.createNewDecl(scope, name, scope_decl.src_index, name_hash, src_hash); const new_decl = try mod.createNewDecl(scope, name, scope_decl.src_node, name_hash, src_hash);
const decl_arena_state = try decl_arena.allocator.create(std.heap.ArenaAllocator.State); const decl_arena_state = try decl_arena.allocator.create(std.heap.ArenaAllocator.State);
decl_arena_state.* = decl_arena.state; decl_arena_state.* = decl_arena.state;
@ -4114,7 +4072,7 @@ pub fn createContainerDecl(
defer mod.gpa.free(name); defer mod.gpa.free(name);
const name_hash = scope.namespace().fullyQualifiedNameHash(name); const name_hash = scope.namespace().fullyQualifiedNameHash(name);
const src_hash: std.zig.SrcHash = undefined; const src_hash: std.zig.SrcHash = undefined;
const new_decl = try mod.createNewDecl(scope, name, scope_decl.src_index, name_hash, src_hash); const new_decl = try mod.createNewDecl(scope, name, scope_decl.src_node, name_hash, src_hash);
const decl_arena_state = try decl_arena.allocator.create(std.heap.ArenaAllocator.State); const decl_arena_state = try decl_arena.allocator.create(std.heap.ArenaAllocator.State);
decl_arena_state.* = decl_arena.state; decl_arena_state.* = decl_arena.state;

View File

@ -417,7 +417,7 @@ fn Function(comptime arch: std.Target.Cpu.Arch) type {
const node_datas = tree.nodes.items(.data); const node_datas = tree.nodes.items(.data);
const token_starts = tree.tokens.items(.start); const token_starts = tree.tokens.items(.start);
const fn_decl = tree.rootDecls()[module_fn.owner_decl.src_index]; const fn_decl = module_fn.owner_decl.src_node;
assert(node_tags[fn_decl] == .fn_decl); assert(node_tags[fn_decl] == .fn_decl);
const block = node_datas[fn_decl].rhs; const block = node_datas[fn_decl].rhs;
const lbrace_src = token_starts[tree.firstToken(block)]; const lbrace_src = token_starts[tree.firstToken(block)];

View File

@ -2228,10 +2228,9 @@ pub fn updateDecl(self: *Elf, module: *Module, decl: *Module.Decl) !void {
const node_datas = tree.nodes.items(.data); const node_datas = tree.nodes.items(.data);
const token_starts = tree.tokens.items(.start); const token_starts = tree.tokens.items(.start);
const file_ast_decls = tree.rootDecls();
// TODO Look into improving the performance here by adding a token-index-to-line // TODO Look into improving the performance here by adding a token-index-to-line
// lookup table. Currently this involves scanning over the source code for newlines. // lookup table. Currently this involves scanning over the source code for newlines.
const fn_decl = file_ast_decls[decl.src_index]; const fn_decl = decl.src_node;
assert(node_tags[fn_decl] == .fn_decl); assert(node_tags[fn_decl] == .fn_decl);
const block = node_datas[fn_decl].rhs; const block = node_datas[fn_decl].rhs;
const lbrace = tree.firstToken(block); const lbrace = tree.firstToken(block);
@ -2755,10 +2754,9 @@ pub fn updateDeclLineNumber(self: *Elf, module: *Module, decl: *const Module.Dec
const node_datas = tree.nodes.items(.data); const node_datas = tree.nodes.items(.data);
const token_starts = tree.tokens.items(.start); const token_starts = tree.tokens.items(.start);
const file_ast_decls = tree.rootDecls();
// TODO Look into improving the performance here by adding a token-index-to-line // TODO Look into improving the performance here by adding a token-index-to-line
// lookup table. Currently this involves scanning over the source code for newlines. // lookup table. Currently this involves scanning over the source code for newlines.
const fn_decl = file_ast_decls[decl.src_index]; const fn_decl = decl.src_node;
assert(node_tags[fn_decl] == .fn_decl); assert(node_tags[fn_decl] == .fn_decl);
const block = node_datas[fn_decl].rhs; const block = node_datas[fn_decl].rhs;
const lbrace = tree.firstToken(block); const lbrace = tree.firstToken(block);

View File

@ -909,10 +909,9 @@ pub fn updateDeclLineNumber(self: *DebugSymbols, module: *Module, decl: *const M
const node_datas = tree.nodes.items(.data); const node_datas = tree.nodes.items(.data);
const token_starts = tree.tokens.items(.start); const token_starts = tree.tokens.items(.start);
const file_ast_decls = tree.rootDecls();
// TODO Look into improving the performance here by adding a token-index-to-line // TODO Look into improving the performance here by adding a token-index-to-line
// lookup table. Currently this involves scanning over the source code for newlines. // lookup table. Currently this involves scanning over the source code for newlines.
const fn_decl = file_ast_decls[decl.src_index]; const fn_decl = decl.src_node;
assert(node_tags[fn_decl] == .fn_decl); assert(node_tags[fn_decl] == .fn_decl);
const block = node_datas[fn_decl].rhs; const block = node_datas[fn_decl].rhs;
const lbrace = tree.firstToken(block); const lbrace = tree.firstToken(block);
@ -959,10 +958,9 @@ pub fn initDeclDebugBuffers(
const node_datas = tree.nodes.items(.data); const node_datas = tree.nodes.items(.data);
const token_starts = tree.tokens.items(.start); const token_starts = tree.tokens.items(.start);
const file_ast_decls = tree.rootDecls();
// TODO Look into improving the performance here by adding a token-index-to-line // TODO Look into improving the performance here by adding a token-index-to-line
// lookup table. Currently this involves scanning over the source code for newlines. // lookup table. Currently this involves scanning over the source code for newlines.
const fn_decl = file_ast_decls[decl.src_index]; const fn_decl = decl.src_node;
assert(node_tags[fn_decl] == .fn_decl); assert(node_tags[fn_decl] == .fn_decl);
const block = node_datas[fn_decl].rhs; const block = node_datas[fn_decl].rhs;
const lbrace = tree.firstToken(block); const lbrace = tree.firstToken(block);

View File

@ -622,6 +622,7 @@ pub const TestContext = struct {
var root_pkg: Package = .{ var root_pkg: Package = .{
.root_src_directory = .{ .path = tmp_dir_path, .handle = tmp.dir }, .root_src_directory = .{ .path = tmp_dir_path, .handle = tmp.dir },
.root_src_path = tmp_src_path, .root_src_path = tmp_src_path,
.namespace_hash = Package.root_namespace_hash,
}; };
const bin_name = try std.zig.binNameAlloc(arena, .{ const bin_name = try std.zig.binNameAlloc(arena, .{
@ -639,13 +640,10 @@ pub const TestContext = struct {
.directory = emit_directory, .directory = emit_directory,
.basename = bin_name, .basename = bin_name,
}; };
const emit_h: ?Compilation.EmitLoc = if (case.emit_h) const emit_h: ?Compilation.EmitLoc = if (case.emit_h) .{
.{ .directory = emit_directory,
.directory = emit_directory, .basename = "test_case.h",
.basename = "test_case.h", } else null;
}
else
null;
const comp = try Compilation.create(allocator, .{ const comp = try Compilation.create(allocator, .{
.local_cache_directory = zig_cache_directory, .local_cache_directory = zig_cache_directory,
.global_cache_directory = global_cache_directory, .global_cache_directory = global_cache_directory,

View File

@ -1040,9 +1040,22 @@ pub fn addCases(ctx: *TestContext) !void {
} }
ctx.compileError("function redefinition", linux_x64, ctx.compileError("function redefinition", linux_x64,
\\// dummy comment
\\fn entry() void {} \\fn entry() void {}
\\fn entry() void {} \\fn entry() void {}
, &[_][]const u8{":2:4: error: redefinition of 'entry'"}); , &[_][]const u8{
":3:4: error: redefinition of 'entry'",
":2:1: note: previous definition here",
});
ctx.compileError("global variable redefinition", linux_x64,
\\// dummy comment
\\var foo = false;
\\var foo = true;
, &[_][]const u8{
":3:5: error: redefinition of 'foo'",
":2:1: note: previous definition here",
});
ctx.compileError("compileError", linux_x64, ctx.compileError("compileError", linux_x64,
\\export fn _start() noreturn { \\export fn _start() noreturn {