mirror of
https://github.com/ziglang/zig.git
synced 2025-12-06 06:13:07 +00:00
stage2 ARM: implement ptr_add, ptr_sub for all element sizes
Also reduces slice_elem_val to ptr_add, simplifying the implementation
This commit is contained in:
parent
ba17552b4e
commit
3ea603c82a
@ -1434,29 +1434,11 @@ fn airSliceElemVal(self: *Self, inst: Air.Inst.Index) !void {
|
|||||||
break :result dst_mcv;
|
break :result dst_mcv;
|
||||||
},
|
},
|
||||||
else => {
|
else => {
|
||||||
const dst_mcv = try self.allocRegOrMem(inst, true);
|
const dest = try self.allocRegOrMem(inst, true);
|
||||||
|
const addr = try self.binOp(.ptr_add, null, base_mcv, index_mcv, slice_ty, Type.usize);
|
||||||
|
try self.load(dest, addr, slice_ptr_field_type);
|
||||||
|
|
||||||
const offset_mcv = try self.binOp(
|
break :result dest;
|
||||||
.mul,
|
|
||||||
null,
|
|
||||||
index_mcv,
|
|
||||||
.{ .immediate = elem_size },
|
|
||||||
Type.usize,
|
|
||||||
Type.usize,
|
|
||||||
);
|
|
||||||
assert(offset_mcv == .register); // result of multiplication should always be register
|
|
||||||
self.register_manager.freezeRegs(&.{offset_mcv.register});
|
|
||||||
|
|
||||||
const addr_mcv = try self.binOp(.add, null, base_mcv, offset_mcv, Type.usize, Type.usize);
|
|
||||||
|
|
||||||
// At this point in time, neither the base register
|
|
||||||
// nor the offset register contains any valuable data
|
|
||||||
// anymore.
|
|
||||||
self.register_manager.unfreezeRegs(&.{ base_mcv.register, offset_mcv.register });
|
|
||||||
|
|
||||||
try self.load(dst_mcv, addr_mcv, slice_ptr_field_type);
|
|
||||||
|
|
||||||
break :result dst_mcv;
|
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
@ -1710,6 +1692,8 @@ fn store(self: *Self, ptr: MCValue, value: MCValue, ptr_ty: Type, value_ty: Type
|
|||||||
defer self.register_manager.unfreezeRegs(&.{addr_reg});
|
defer self.register_manager.unfreezeRegs(&.{addr_reg});
|
||||||
|
|
||||||
switch (value) {
|
switch (value) {
|
||||||
|
.dead => unreachable,
|
||||||
|
.undef => unreachable,
|
||||||
.register => |value_reg| {
|
.register => |value_reg| {
|
||||||
try self.genStrRegister(value_reg, addr_reg, value_ty);
|
try self.genStrRegister(value_reg, addr_reg, value_ty);
|
||||||
},
|
},
|
||||||
@ -2140,7 +2124,7 @@ fn binOp(
|
|||||||
rhs: MCValue,
|
rhs: MCValue,
|
||||||
lhs_ty: Type,
|
lhs_ty: Type,
|
||||||
rhs_ty: Type,
|
rhs_ty: Type,
|
||||||
) !MCValue {
|
) InnerError!MCValue {
|
||||||
switch (tag) {
|
switch (tag) {
|
||||||
.add,
|
.add,
|
||||||
.sub,
|
.sub,
|
||||||
@ -2281,16 +2265,21 @@ fn binOp(
|
|||||||
switch (lhs_ty.zigTypeTag()) {
|
switch (lhs_ty.zigTypeTag()) {
|
||||||
.Pointer => {
|
.Pointer => {
|
||||||
const ptr_ty = lhs_ty;
|
const ptr_ty = lhs_ty;
|
||||||
const pointee_ty = switch (ptr_ty.ptrSize()) {
|
const elem_ty = switch (ptr_ty.ptrSize()) {
|
||||||
.One => ptr_ty.childType().childType(), // ptr to array, so get array element type
|
.One => ptr_ty.childType().childType(), // ptr to array, so get array element type
|
||||||
else => ptr_ty.childType(),
|
else => ptr_ty.childType(),
|
||||||
};
|
};
|
||||||
|
const elem_size = @intCast(u32, elem_ty.abiSize(self.target.*));
|
||||||
|
|
||||||
if (pointee_ty.abiSize(self.target.*) > 1) {
|
if (elem_size == 1) {
|
||||||
return self.fail("TODO ptr_add, ptr_sub with more element sizes", .{});
|
|
||||||
}
|
|
||||||
|
|
||||||
return try self.binOpRegister(tag, maybe_inst, lhs, rhs, lhs_ty, rhs_ty);
|
return try self.binOpRegister(tag, maybe_inst, lhs, rhs, lhs_ty, rhs_ty);
|
||||||
|
} else {
|
||||||
|
// convert the offset into a byte offset by
|
||||||
|
// multiplying it with elem_size
|
||||||
|
const offset = try self.binOp(.mul, null, rhs, .{ .immediate = elem_size }, Type.usize, Type.usize);
|
||||||
|
const addr = try self.binOp(tag, null, lhs, offset, Type.initTag(.manyptr_u8), Type.usize);
|
||||||
|
return addr;
|
||||||
|
}
|
||||||
},
|
},
|
||||||
else => unreachable,
|
else => unreachable,
|
||||||
}
|
}
|
||||||
@ -3494,6 +3483,12 @@ fn genSetStack(self: *Self, ty: Type, stack_offset: u32, mcv: MCValue) InnerErro
|
|||||||
const reg = try self.copyToTmpRegister(ty, mcv);
|
const reg = try self.copyToTmpRegister(ty, mcv);
|
||||||
return self.genSetStack(ty, stack_offset, MCValue{ .register = reg });
|
return self.genSetStack(ty, stack_offset, MCValue{ .register = reg });
|
||||||
} else {
|
} else {
|
||||||
|
var ptr_ty_payload: Type.Payload.ElemType = .{
|
||||||
|
.base = .{ .tag = .single_mut_pointer },
|
||||||
|
.data = ty,
|
||||||
|
};
|
||||||
|
const ptr_ty = Type.initPayload(&ptr_ty_payload.base);
|
||||||
|
|
||||||
// TODO call extern memcpy
|
// TODO call extern memcpy
|
||||||
const regs = try self.register_manager.allocRegs(5, .{ null, null, null, null, null });
|
const regs = try self.register_manager.allocRegs(5, .{ null, null, null, null, null });
|
||||||
const src_reg = regs[0];
|
const src_reg = regs[0];
|
||||||
@ -3505,20 +3500,9 @@ fn genSetStack(self: *Self, ty: Type, stack_offset: u32, mcv: MCValue) InnerErro
|
|||||||
switch (mcv) {
|
switch (mcv) {
|
||||||
.stack_offset => |off| {
|
.stack_offset => |off| {
|
||||||
// sub src_reg, fp, #off
|
// sub src_reg, fp, #off
|
||||||
const adj_src_offset = off + @intCast(u32, abi_size);
|
try self.genSetReg(ptr_ty, src_reg, .{ .ptr_stack_offset = off });
|
||||||
const src_offset_op: Instruction.Operand = if (Instruction.Operand.fromU32(adj_src_offset)) |x| x else {
|
|
||||||
return self.fail("TODO load: set reg to stack offset with all possible offsets", .{});
|
|
||||||
};
|
|
||||||
_ = try self.addInst(.{
|
|
||||||
.tag = .sub,
|
|
||||||
.data = .{ .rr_op = .{
|
|
||||||
.rd = src_reg,
|
|
||||||
.rn = .fp,
|
|
||||||
.op = src_offset_op,
|
|
||||||
} },
|
|
||||||
});
|
|
||||||
},
|
},
|
||||||
.memory => |addr| try self.genSetReg(Type.usize, src_reg, .{ .immediate = @intCast(u32, addr) }),
|
.memory => |addr| try self.genSetReg(ptr_ty, src_reg, .{ .immediate = @intCast(u32, addr) }),
|
||||||
.embedded_in_code,
|
.embedded_in_code,
|
||||||
.stack_argument_offset,
|
.stack_argument_offset,
|
||||||
=> return self.fail("TODO genSetStack with src={}", .{mcv}),
|
=> return self.fail("TODO genSetStack with src={}", .{mcv}),
|
||||||
@ -3526,18 +3510,7 @@ fn genSetStack(self: *Self, ty: Type, stack_offset: u32, mcv: MCValue) InnerErro
|
|||||||
}
|
}
|
||||||
|
|
||||||
// sub dst_reg, fp, #stack_offset
|
// sub dst_reg, fp, #stack_offset
|
||||||
const adj_dst_offset = stack_offset + abi_size;
|
try self.genSetReg(ptr_ty, dst_reg, .{ .ptr_stack_offset = stack_offset });
|
||||||
const dst_offset_op: Instruction.Operand = if (Instruction.Operand.fromU32(adj_dst_offset)) |x| x else {
|
|
||||||
return self.fail("TODO load: set reg to stack offset with all possible offsets", .{});
|
|
||||||
};
|
|
||||||
_ = try self.addInst(.{
|
|
||||||
.tag = .sub,
|
|
||||||
.data = .{ .rr_op = .{
|
|
||||||
.rd = dst_reg,
|
|
||||||
.rn = .fp,
|
|
||||||
.op = dst_offset_op,
|
|
||||||
} },
|
|
||||||
});
|
|
||||||
|
|
||||||
// mov len, #abi_size
|
// mov len, #abi_size
|
||||||
try self.genSetReg(Type.usize, len_reg, .{ .immediate = abi_size });
|
try self.genSetReg(Type.usize, len_reg, .{ .immediate = abi_size });
|
||||||
@ -3882,6 +3855,12 @@ fn genSetStackArgument(self: *Self, ty: Type, stack_offset: u32, mcv: MCValue) I
|
|||||||
const reg = try self.copyToTmpRegister(ty, mcv);
|
const reg = try self.copyToTmpRegister(ty, mcv);
|
||||||
return self.genSetStackArgument(ty, stack_offset, MCValue{ .register = reg });
|
return self.genSetStackArgument(ty, stack_offset, MCValue{ .register = reg });
|
||||||
} else {
|
} else {
|
||||||
|
var ptr_ty_payload: Type.Payload.ElemType = .{
|
||||||
|
.base = .{ .tag = .single_mut_pointer },
|
||||||
|
.data = ty,
|
||||||
|
};
|
||||||
|
const ptr_ty = Type.initPayload(&ptr_ty_payload.base);
|
||||||
|
|
||||||
// TODO call extern memcpy
|
// TODO call extern memcpy
|
||||||
const regs = try self.register_manager.allocRegs(5, .{ null, null, null, null, null });
|
const regs = try self.register_manager.allocRegs(5, .{ null, null, null, null, null });
|
||||||
const src_reg = regs[0];
|
const src_reg = regs[0];
|
||||||
@ -3893,20 +3872,9 @@ fn genSetStackArgument(self: *Self, ty: Type, stack_offset: u32, mcv: MCValue) I
|
|||||||
switch (mcv) {
|
switch (mcv) {
|
||||||
.stack_offset => |off| {
|
.stack_offset => |off| {
|
||||||
// sub src_reg, fp, #off
|
// sub src_reg, fp, #off
|
||||||
const adj_src_offset = off + abi_size;
|
try self.genSetReg(ptr_ty, src_reg, .{ .ptr_stack_offset = off });
|
||||||
const src_offset_op: Instruction.Operand = if (Instruction.Operand.fromU32(adj_src_offset)) |x| x else {
|
|
||||||
return self.fail("TODO load: set reg to stack offset with all possible offsets", .{});
|
|
||||||
};
|
|
||||||
_ = try self.addInst(.{
|
|
||||||
.tag = .sub,
|
|
||||||
.data = .{ .rr_op = .{
|
|
||||||
.rd = src_reg,
|
|
||||||
.rn = .fp,
|
|
||||||
.op = src_offset_op,
|
|
||||||
} },
|
|
||||||
});
|
|
||||||
},
|
},
|
||||||
.memory => |addr| try self.genSetReg(Type.usize, src_reg, .{ .immediate = @intCast(u32, addr) }),
|
.memory => |addr| try self.genSetReg(ptr_ty, src_reg, .{ .immediate = @intCast(u32, addr) }),
|
||||||
.stack_argument_offset,
|
.stack_argument_offset,
|
||||||
.embedded_in_code,
|
.embedded_in_code,
|
||||||
=> return self.fail("TODO genSetStackArgument src={}", .{mcv}),
|
=> return self.fail("TODO genSetStackArgument src={}", .{mcv}),
|
||||||
|
|||||||
@ -105,7 +105,6 @@ fn fnWithAlignedStack() i32 {
|
|||||||
test "implicitly decreasing slice alignment" {
|
test "implicitly decreasing slice alignment" {
|
||||||
if (builtin.zig_backend == .stage2_c) return error.SkipZigTest;
|
if (builtin.zig_backend == .stage2_c) return error.SkipZigTest;
|
||||||
if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest;
|
if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest;
|
||||||
if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest;
|
|
||||||
|
|
||||||
const a: u32 align(4) = 3;
|
const a: u32 align(4) = 3;
|
||||||
const b: u32 align(8) = 4;
|
const b: u32 align(8) = 4;
|
||||||
|
|||||||
@ -8,7 +8,6 @@ const expectEqual = testing.expectEqual;
|
|||||||
test "array to slice" {
|
test "array to slice" {
|
||||||
if (builtin.zig_backend == .stage2_c) return error.SkipZigTest;
|
if (builtin.zig_backend == .stage2_c) return error.SkipZigTest;
|
||||||
if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest;
|
if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest;
|
||||||
if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest;
|
|
||||||
|
|
||||||
const a: u32 align(4) = 3;
|
const a: u32 align(4) = 3;
|
||||||
const b: u32 align(8) = 4;
|
const b: u32 align(8) = 4;
|
||||||
|
|||||||
@ -984,7 +984,6 @@ test "peer type resolve array pointers, one of them const" {
|
|||||||
test "peer type resolve array pointer and unknown pointer" {
|
test "peer type resolve array pointer and unknown pointer" {
|
||||||
if (builtin.zig_backend == .stage2_wasm) return error.SkipZigTest; // TODO
|
if (builtin.zig_backend == .stage2_wasm) return error.SkipZigTest; // TODO
|
||||||
if (builtin.zig_backend == .stage2_x86_64) return error.SkipZigTest; // TODO
|
if (builtin.zig_backend == .stage2_x86_64) return error.SkipZigTest; // TODO
|
||||||
if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; // TODO
|
|
||||||
if (builtin.zig_backend == .stage2_c) return error.SkipZigTest; // TODO
|
if (builtin.zig_backend == .stage2_c) return error.SkipZigTest; // TODO
|
||||||
|
|
||||||
const const_array: [4]u8 = undefined;
|
const const_array: [4]u8 = undefined;
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user