replace most aggregate interns in x86_64/CodeGen

This commit is contained in:
Justus Klausecker 2025-08-07 23:31:38 +02:00
parent 79e5c138c6
commit 76d2782149

View File

@ -170058,12 +170058,9 @@ fn airTrunc(self: *CodeGen, inst: Air.Inst.Index) !void {
}); });
const splat_abi_size: u32 = @intCast(splat_ty.abiSize(zcu)); const splat_abi_size: u32 = @intCast(splat_ty.abiSize(zcu));
const splat_val = try pt.intern(.{ .aggregate = .{ const splat_val = try pt.aggregateSplatValue(splat_ty, mask_val);
.ty = splat_ty.ip_index,
.storage = .{ .repeated_elem = mask_val.ip_index },
} });
const splat_mcv = try self.lowerValue(.fromInterned(splat_val)); const splat_mcv = try self.lowerValue(splat_val);
const splat_addr_mcv: MCValue = switch (splat_mcv) { const splat_addr_mcv: MCValue = switch (splat_mcv) {
.memory, .indirect, .load_frame => splat_mcv.address(), .memory, .indirect, .load_frame => splat_mcv.address(),
else => .{ .register = try self.copyToTmpRegister(.usize, splat_mcv.address()) }, else => .{ .register = try self.copyToTmpRegister(.usize, splat_mcv.address()) },
@ -171693,12 +171690,12 @@ fn airShlShrBinOp(self: *CodeGen, inst: Air.Inst.Index) !void {
defer self.register_manager.unlockReg(shift_lock); defer self.register_manager.unlockReg(shift_lock);
const mask_ty = try pt.vectorType(.{ .len = 16, .child = .u8_type }); const mask_ty = try pt.vectorType(.{ .len = 16, .child = .u8_type });
const mask_mcv = try self.lowerValue(.fromInterned(try pt.intern(.{ .aggregate = .{ const mask_mcv = try self.lowerValue(try pt.aggregateValue(
.ty = mask_ty.toIntern(), mask_ty,
.storage = .{ .elems = &([1]InternPool.Index{ &([1]InternPool.Index{
(try rhs_ty.childType(zcu).maxIntScalar(pt, .u8)).toIntern(), (try rhs_ty.childType(zcu).maxIntScalar(pt, .u8)).toIntern(),
} ++ [1]InternPool.Index{.zero_u8} ** 15) }, } ++ [1]InternPool.Index{.zero_u8} ** 15),
} }))); ));
const mask_addr_reg = try self.copyToTmpRegister(.usize, mask_mcv.address()); const mask_addr_reg = try self.copyToTmpRegister(.usize, mask_mcv.address());
const mask_addr_lock = self.register_manager.lockRegAssumeUnused(mask_addr_reg); const mask_addr_lock = self.register_manager.lockRegAssumeUnused(mask_addr_reg);
defer self.register_manager.unlockReg(mask_addr_lock); defer self.register_manager.unlockReg(mask_addr_lock);
@ -181139,10 +181136,7 @@ fn genSetReg(
.child = .u8_type, .child = .u8_type,
}); });
try self.genSetReg(dst_reg, full_ty, try self.lowerValue( try self.genSetReg(dst_reg, full_ty, try self.lowerValue(
.fromInterned(try pt.intern(.{ .aggregate = .{ try pt.aggregateSplatValue(full_ty, try pt.intValue(.u8, 0xaa)),
.ty = full_ty.toIntern(),
.storage = .{ .repeated_elem = (try pt.intValue(.u8, 0xaa)).toIntern() },
} })),
), opts); ), opts);
}, },
.x87 => try self.genSetReg(dst_reg, .f80, try self.lowerValue( .x87 => try self.genSetReg(dst_reg, .f80, try self.lowerValue(
@ -183565,10 +183559,7 @@ fn airSelect(self: *CodeGen, inst: Air.Inst.Index) !void {
mask_elem_ty, mask_elem_ty,
@as(u8, 1) << @truncate(bit), @as(u8, 1) << @truncate(bit),
)).toIntern(); )).toIntern();
const mask_mcv = try self.lowerValue(.fromInterned(try pt.intern(.{ .aggregate = .{ const mask_mcv = try self.lowerValue(try pt.aggregateValue(mask_ty, mask_elems));
.ty = mask_ty.toIntern(),
.storage = .{ .elems = mask_elems },
} })));
const mask_mem: Memory = .{ const mask_mem: Memory = .{
.base = .{ .reg = try self.copyToTmpRegister(.usize, mask_mcv.address()) }, .base = .{ .reg = try self.copyToTmpRegister(.usize, mask_mcv.address()) },
.mod = .{ .rm = .{ .size = self.memSize(ty) } }, .mod = .{ .rm = .{ .size = self.memSize(ty) } },
@ -184296,10 +184287,9 @@ fn airShuffle(self: *CodeGen, inst: Air.Inst.Index) !void {
else else
try select_mask_elem_ty.minIntScalar(pt, select_mask_elem_ty)).toIntern(); try select_mask_elem_ty.minIntScalar(pt, select_mask_elem_ty)).toIntern();
} }
const select_mask_mcv = try self.lowerValue(.fromInterned(try pt.intern(.{ .aggregate = .{ const select_mask_mcv = try self.lowerValue(
.ty = select_mask_ty.toIntern(), try pt.aggregateValue(select_mask_ty, select_mask_elems[0..mask_elems.len]),
.storage = .{ .elems = select_mask_elems[0..mask_elems.len] }, );
} })));
if (self.hasFeature(.sse4_1)) { if (self.hasFeature(.sse4_1)) {
const mir_tag: Mir.Inst.FixedTag = .{ const mir_tag: Mir.Inst.FixedTag = .{
@ -184441,10 +184431,9 @@ fn airShuffle(self: *CodeGen, inst: Air.Inst.Index) !void {
})).toIntern(); })).toIntern();
} }
const lhs_mask_ty = try pt.vectorType(.{ .len = max_abi_size, .child = .u8_type }); const lhs_mask_ty = try pt.vectorType(.{ .len = max_abi_size, .child = .u8_type });
const lhs_mask_mcv = try self.lowerValue(.fromInterned(try pt.intern(.{ .aggregate = .{ const lhs_mask_mcv = try self.lowerValue(
.ty = lhs_mask_ty.toIntern(), try pt.aggregateValue(lhs_mask_ty, lhs_mask_elems[0..max_abi_size]),
.storage = .{ .elems = lhs_mask_elems[0..max_abi_size] }, );
} })));
const lhs_mask_mem: Memory = .{ const lhs_mask_mem: Memory = .{
.base = .{ .reg = try self.copyToTmpRegister(.usize, lhs_mask_mcv.address()) }, .base = .{ .reg = try self.copyToTmpRegister(.usize, lhs_mask_mcv.address()) },
.mod = .{ .rm = .{ .size = .fromSize(@max(max_abi_size, 16)) } }, .mod = .{ .rm = .{ .size = .fromSize(@max(max_abi_size, 16)) } },
@ -184472,10 +184461,9 @@ fn airShuffle(self: *CodeGen, inst: Air.Inst.Index) !void {
})).toIntern(); })).toIntern();
} }
const rhs_mask_ty = try pt.vectorType(.{ .len = max_abi_size, .child = .u8_type }); const rhs_mask_ty = try pt.vectorType(.{ .len = max_abi_size, .child = .u8_type });
const rhs_mask_mcv = try self.lowerValue(.fromInterned(try pt.intern(.{ .aggregate = .{ const rhs_mask_mcv = try self.lowerValue(
.ty = rhs_mask_ty.toIntern(), try pt.aggregateValue(rhs_mask_ty, rhs_mask_elems[0..max_abi_size]),
.storage = .{ .elems = rhs_mask_elems[0..max_abi_size] }, );
} })));
const rhs_mask_mem: Memory = .{ const rhs_mask_mem: Memory = .{
.base = .{ .reg = try self.copyToTmpRegister(.usize, rhs_mask_mcv.address()) }, .base = .{ .reg = try self.copyToTmpRegister(.usize, rhs_mask_mcv.address()) },
.mod = .{ .rm = .{ .size = .fromSize(@max(max_abi_size, 16)) } }, .mod = .{ .rm = .{ .size = .fromSize(@max(max_abi_size, 16)) } },
@ -192924,36 +192912,30 @@ const Select = struct {
break :res_scalar .{ res_scalar_ty, try pt.intValue_big(res_scalar_ty, res_big_int.toConst()) }; break :res_scalar .{ res_scalar_ty, try pt.intValue_big(res_scalar_ty, res_big_int.toConst()) };
}, },
}; };
const res_val: Value = if (res_vector_len) |len| .fromInterned(try pt.intern(.{ .aggregate = .{ const res_val = if (res_vector_len) |len| try pt.aggregateSplatValue(try pt.vectorType(.{
.ty = (try pt.vectorType(.{ .len = len,
.len = len, .child = res_scalar_ty.toIntern(),
.child = res_scalar_ty.toIntern(), }), res_scalar_val) else res_scalar_val;
})).toIntern(),
.storage = .{ .repeated_elem = res_scalar_val.toIntern() },
} })) else res_scalar_val;
return .{ try cg.tempMemFromValue(res_val), true }; return .{ try cg.tempMemFromValue(res_val), true };
}, },
.f64_0x1p52_0x1p84_mem => .{ try cg.tempMemFromValue(.fromInterned(try pt.intern(.{ .aggregate = .{ .f64_0x1p52_0x1p84_mem => .{ try cg.tempMemFromValue(
.ty = (try pt.vectorType(.{ .len = 2, .child = .f64_type })).toIntern(), try pt.aggregateValue(try pt.vectorType(.{ .len = 2, .child = .f64_type }), &.{
.storage = .{ .elems = &.{
(try pt.floatValue(.f64, @as(f64, 0x1p52))).toIntern(), (try pt.floatValue(.f64, @as(f64, 0x1p52))).toIntern(),
(try pt.floatValue(.f64, @as(f64, 0x1p84))).toIntern(), (try pt.floatValue(.f64, @as(f64, 0x1p84))).toIntern(),
} }, }),
} }))), true }, ), true },
.u32_0x1p52_hi_0x1p84_hi_0_0_mem => .{ try cg.tempMemFromValue(.fromInterned(try pt.intern(.{ .aggregate = .{ .u32_0x1p52_hi_0x1p84_hi_0_0_mem => .{ try cg.tempMemFromValue(
.ty = (try pt.vectorType(.{ .len = 4, .child = .u32_type })).toIntern(), try pt.aggregateValue(try pt.vectorType(.{ .len = 4, .child = .u32_type }), &(.{
.storage = .{ .elems = &(.{
(try pt.intValue(.u32, @as(u64, @bitCast(@as(f64, 0x1p52))) >> 32)).toIntern(), (try pt.intValue(.u32, @as(u64, @bitCast(@as(f64, 0x1p52))) >> 32)).toIntern(),
(try pt.intValue(.u32, @as(u64, @bitCast(@as(f64, 0x1p84))) >> 32)).toIntern(), (try pt.intValue(.u32, @as(u64, @bitCast(@as(f64, 0x1p84))) >> 32)).toIntern(),
} ++ .{(try pt.intValue(.u32, 0)).toIntern()} ** 2) }, } ++ .{(try pt.intValue(.u32, 0)).toIntern()} ** 2)),
} }))), true }, ), true },
.f32_0_0x1p64_mem => .{ try cg.tempMemFromValue(.fromInterned(try pt.intern(.{ .aggregate = .{ .f32_0_0x1p64_mem => .{ try cg.tempMemFromValue(
.ty = (try pt.vectorType(.{ .len = 2, .child = .f32_type })).toIntern(), try pt.aggregateValue(try pt.vectorType(.{ .len = 2, .child = .f32_type }), &.{
.storage = .{ .elems = &.{
(try pt.floatValue(.f32, @as(f32, 0))).toIntern(), (try pt.floatValue(.f32, @as(f32, 0))).toIntern(),
(try pt.floatValue(.f32, @as(f32, 0x1p64))).toIntern(), (try pt.floatValue(.f32, @as(f32, 0x1p64))).toIntern(),
} }, }),
} }))), true }, ), true },
.pshufb_splat_mem => |splat_spec| { .pshufb_splat_mem => |splat_spec| {
const zcu = pt.zcu; const zcu = pt.zcu;
assert(spec.type.isVector(zcu) and spec.type.childType(zcu).toIntern() == .u8_type); assert(spec.type.isVector(zcu) and spec.type.childType(zcu).toIntern() == .u8_type);
@ -193110,13 +193092,10 @@ const Select = struct {
const mem_size = cg.unalignedSize(spec.type); const mem_size = cg.unalignedSize(spec.type);
return .{ try cg.tempMemFromAlignedValue( return .{ try cg.tempMemFromAlignedValue(
if (mem_size < 16) .fromByteUnits(mem_size) else .none, if (mem_size < 16) .fromByteUnits(mem_size) else .none,
.fromInterned(try pt.intern(.{ .aggregate = .{ try pt.aggregateValue(if (mem_size < 16) try pt.arrayType(.{
.ty = if (mem_size < 16) .len = elems.len,
(try pt.arrayType(.{ .len = elems.len, .child = elem_ty.toIntern() })).toIntern() .child = elem_ty.toIntern(),
else }) else spec.type, elems),
spec.type.toIntern(),
.storage = .{ .elems = elems },
} })),
), true }; ), true };
}, },
.splat_float_mem => |splat_spec| { .splat_float_mem => |splat_spec| {
@ -193133,10 +193112,7 @@ const Select = struct {
.zero => 0.0, .zero => 0.0,
}))).toIntern()); }))).toIntern());
@memset(elems[inside_len..], (try pt.floatValue(elem_ty, splat_spec.outside)).toIntern()); @memset(elems[inside_len..], (try pt.floatValue(elem_ty, splat_spec.outside)).toIntern());
return .{ try cg.tempMemFromValue(.fromInterned(try pt.intern(.{ .aggregate = .{ return .{ try cg.tempMemFromValue(try pt.aggregateValue(spec.type, elems)), true };
.ty = spec.type.toIntern(),
.storage = .{ .elems = elems },
} }))), true };
}, },
.frame => |frame_index| .{ try cg.tempInit(spec.type, .{ .load_frame = .{ .frame => |frame_index| .{ try cg.tempInit(spec.type, .{ .load_frame = .{
.index = frame_index, .index = frame_index,