Merge branch 'feature/exclusive-mem'
This commit is contained in:
commit
9f04f2c892
15 changed files with 323 additions and 15 deletions
|
@ -224,7 +224,7 @@ void A64EmitX64::GenFastmemFallbacks() {
|
||||||
ABI_PopCallerSaveRegistersAndAdjustStack(code);
|
ABI_PopCallerSaveRegistersAndAdjustStack(code);
|
||||||
code.ret();
|
code.ret();
|
||||||
|
|
||||||
if (vaddr_idx == value_idx || value_idx == 4 || value_idx == 15) {
|
if (value_idx == 4 || value_idx == 15) {
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -249,13 +249,18 @@ void A64EmitX64::GenFastmemFallbacks() {
|
||||||
ABI_PushCallerSaveRegistersAndAdjustStack(code);
|
ABI_PushCallerSaveRegistersAndAdjustStack(code);
|
||||||
if (vaddr_idx == code.ABI_PARAM3.getIdx() && value_idx == code.ABI_PARAM2.getIdx()) {
|
if (vaddr_idx == code.ABI_PARAM3.getIdx() && value_idx == code.ABI_PARAM2.getIdx()) {
|
||||||
code.xchg(code.ABI_PARAM2, code.ABI_PARAM3);
|
code.xchg(code.ABI_PARAM2, code.ABI_PARAM3);
|
||||||
} else {
|
} else if (vaddr_idx == code.ABI_PARAM3.getIdx()) {
|
||||||
if (vaddr_idx != code.ABI_PARAM2.getIdx()) {
|
code.mov(code.ABI_PARAM2, Xbyak::Reg64{vaddr_idx});
|
||||||
code.mov(code.ABI_PARAM2, Xbyak::Reg64{vaddr_idx});
|
|
||||||
}
|
|
||||||
if (value_idx != code.ABI_PARAM3.getIdx()) {
|
if (value_idx != code.ABI_PARAM3.getIdx()) {
|
||||||
code.mov(code.ABI_PARAM3, Xbyak::Reg64{value_idx});
|
code.mov(code.ABI_PARAM3, Xbyak::Reg64{value_idx});
|
||||||
}
|
}
|
||||||
|
} else {
|
||||||
|
if (value_idx != code.ABI_PARAM3.getIdx()) {
|
||||||
|
code.mov(code.ABI_PARAM3, Xbyak::Reg64{value_idx});
|
||||||
|
}
|
||||||
|
if (vaddr_idx != code.ABI_PARAM2.getIdx()) {
|
||||||
|
code.mov(code.ABI_PARAM2, Xbyak::Reg64{vaddr_idx});
|
||||||
|
}
|
||||||
}
|
}
|
||||||
callback.EmitCall(code);
|
callback.EmitCall(code);
|
||||||
ABI_PopCallerSaveRegistersAndAdjustStack(code);
|
ABI_PopCallerSaveRegistersAndAdjustStack(code);
|
||||||
|
@ -480,6 +485,19 @@ void A64EmitX64::EmitA64GetTPIDRRO(A64EmitContext& ctx, IR::Inst* inst) {
|
||||||
ctx.reg_alloc.DefineValue(inst, result);
|
ctx.reg_alloc.DefineValue(inst, result);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void A64EmitX64::EmitA64ClearExclusive(A64EmitContext&, IR::Inst*) {
|
||||||
|
code.mov(code.byte[r15 + offsetof(A64JitState, exclusive_state)], u8(0));
|
||||||
|
}
|
||||||
|
|
||||||
|
void A64EmitX64::EmitA64SetExclusive(A64EmitContext& ctx, IR::Inst* inst) {
|
||||||
|
auto args = ctx.reg_alloc.GetArgumentInfo(inst);
|
||||||
|
ASSERT(args[1].IsImmediate());
|
||||||
|
Xbyak::Reg32 address = ctx.reg_alloc.UseGpr(args[0]).cvt32();
|
||||||
|
|
||||||
|
code.mov(code.byte[r15 + offsetof(A64JitState, exclusive_state)], u8(1));
|
||||||
|
code.mov(dword[r15 + offsetof(A64JitState, exclusive_address)], address);
|
||||||
|
}
|
||||||
|
|
||||||
static Xbyak::RegExp EmitVAddrLookup(const A64::UserConfig& conf, BlockOfCode& code, A64EmitContext& ctx, Xbyak::Label& abort, Xbyak::Reg64 vaddr, boost::optional<Xbyak::Reg64> arg_scratch = {}) {
|
static Xbyak::RegExp EmitVAddrLookup(const A64::UserConfig& conf, BlockOfCode& code, A64EmitContext& ctx, Xbyak::Label& abort, Xbyak::Reg64 vaddr, boost::optional<Xbyak::Reg64> arg_scratch = {}) {
|
||||||
constexpr size_t PAGE_BITS = 12;
|
constexpr size_t PAGE_BITS = 12;
|
||||||
constexpr size_t PAGE_SIZE = 1 << PAGE_BITS;
|
constexpr size_t PAGE_SIZE = 1 << PAGE_BITS;
|
||||||
|
@ -722,6 +740,61 @@ void A64EmitX64::EmitA64WriteMemory128(A64EmitContext& ctx, IR::Inst* inst) {
|
||||||
code.CallFunction(memory_write_128);
|
code.CallFunction(memory_write_128);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void A64EmitX64::EmitExclusiveWrite(A64EmitContext& ctx, IR::Inst* inst, size_t bitsize, Xbyak::Reg64 vaddr, size_t value_idx) {
|
||||||
|
Xbyak::Label end;
|
||||||
|
Xbyak::Reg32 passed = ctx.reg_alloc.ScratchGpr().cvt32();
|
||||||
|
Xbyak::Reg32 tmp = ctx.reg_alloc.ScratchGpr().cvt32();
|
||||||
|
|
||||||
|
code.mov(passed, u32(1));
|
||||||
|
code.cmp(code.byte[r15 + offsetof(A64JitState, exclusive_state)], u8(0));
|
||||||
|
code.je(end);
|
||||||
|
code.mov(tmp, vaddr);
|
||||||
|
code.xor_(tmp, dword[r15 + offsetof(A64JitState, exclusive_address)]);
|
||||||
|
code.test(tmp, A64JitState::RESERVATION_GRANULE_MASK);
|
||||||
|
code.jne(end);
|
||||||
|
code.mov(code.byte[r15 + offsetof(A64JitState, exclusive_state)], u8(0));
|
||||||
|
code.call(write_fallbacks[std::make_tuple(bitsize, vaddr.getIdx(), value_idx)]);
|
||||||
|
code.xor_(passed, passed);
|
||||||
|
code.L(end);
|
||||||
|
|
||||||
|
ctx.reg_alloc.DefineValue(inst, passed);
|
||||||
|
}
|
||||||
|
|
||||||
|
void A64EmitX64::EmitA64ExclusiveWriteMemory8(A64EmitContext& ctx, IR::Inst* inst) {
|
||||||
|
auto args = ctx.reg_alloc.GetArgumentInfo(inst);
|
||||||
|
Xbyak::Reg64 vaddr = ctx.reg_alloc.UseGpr(args[0]);
|
||||||
|
Xbyak::Reg64 value = ctx.reg_alloc.UseGpr(args[1]);
|
||||||
|
EmitExclusiveWrite(ctx, inst, 8, vaddr, value.getIdx());
|
||||||
|
}
|
||||||
|
|
||||||
|
void A64EmitX64::EmitA64ExclusiveWriteMemory16(A64EmitContext& ctx, IR::Inst* inst) {
|
||||||
|
auto args = ctx.reg_alloc.GetArgumentInfo(inst);
|
||||||
|
Xbyak::Reg64 vaddr = ctx.reg_alloc.UseGpr(args[0]);
|
||||||
|
Xbyak::Reg64 value = ctx.reg_alloc.UseGpr(args[1]);
|
||||||
|
EmitExclusiveWrite(ctx, inst, 16, vaddr, value.getIdx());
|
||||||
|
}
|
||||||
|
|
||||||
|
void A64EmitX64::EmitA64ExclusiveWriteMemory32(A64EmitContext& ctx, IR::Inst* inst) {
|
||||||
|
auto args = ctx.reg_alloc.GetArgumentInfo(inst);
|
||||||
|
Xbyak::Reg64 vaddr = ctx.reg_alloc.UseGpr(args[0]);
|
||||||
|
Xbyak::Reg64 value = ctx.reg_alloc.UseGpr(args[1]);
|
||||||
|
EmitExclusiveWrite(ctx, inst, 32, vaddr, value.getIdx());
|
||||||
|
}
|
||||||
|
|
||||||
|
void A64EmitX64::EmitA64ExclusiveWriteMemory64(A64EmitContext& ctx, IR::Inst* inst) {
|
||||||
|
auto args = ctx.reg_alloc.GetArgumentInfo(inst);
|
||||||
|
Xbyak::Reg64 vaddr = ctx.reg_alloc.UseGpr(args[0]);
|
||||||
|
Xbyak::Reg64 value = ctx.reg_alloc.UseGpr(args[1]);
|
||||||
|
EmitExclusiveWrite(ctx, inst, 64, vaddr, value.getIdx());
|
||||||
|
}
|
||||||
|
|
||||||
|
void A64EmitX64::EmitA64ExclusiveWriteMemory128(A64EmitContext& ctx, IR::Inst* inst) {
|
||||||
|
auto args = ctx.reg_alloc.GetArgumentInfo(inst);
|
||||||
|
Xbyak::Reg64 vaddr = ctx.reg_alloc.UseGpr(args[0]);
|
||||||
|
Xbyak::Xmm value = ctx.reg_alloc.UseXmm(args[1]);
|
||||||
|
EmitExclusiveWrite(ctx, inst, 128, vaddr, value.getIdx());
|
||||||
|
}
|
||||||
|
|
||||||
void A64EmitX64::EmitTerminalImpl(IR::Term::Interpret terminal, IR::LocationDescriptor) {
|
void A64EmitX64::EmitTerminalImpl(IR::Term::Interpret terminal, IR::LocationDescriptor) {
|
||||||
code.SwitchMxcsrOnExit();
|
code.SwitchMxcsrOnExit();
|
||||||
DEVIRT(conf.callbacks, &A64::UserCallbacks::InterpreterFallback).EmitCall(code, [&](RegList param) {
|
DEVIRT(conf.callbacks, &A64::UserCallbacks::InterpreterFallback).EmitCall(code, [&](RegList param) {
|
||||||
|
|
|
@ -57,6 +57,7 @@ protected:
|
||||||
|
|
||||||
void EmitDirectPageTableMemoryRead(A64EmitContext& ctx, IR::Inst* inst, size_t bitsize);
|
void EmitDirectPageTableMemoryRead(A64EmitContext& ctx, IR::Inst* inst, size_t bitsize);
|
||||||
void EmitDirectPageTableMemoryWrite(A64EmitContext& ctx, IR::Inst* inst, size_t bitsize);
|
void EmitDirectPageTableMemoryWrite(A64EmitContext& ctx, IR::Inst* inst, size_t bitsize);
|
||||||
|
void EmitExclusiveWrite(A64EmitContext& ctx, IR::Inst* inst, size_t bitsize, Xbyak::Reg64 vaddr, size_t value_idx);
|
||||||
|
|
||||||
// Microinstruction emitters
|
// Microinstruction emitters
|
||||||
#define OPCODE(...)
|
#define OPCODE(...)
|
||||||
|
|
|
@ -56,6 +56,11 @@ struct A64JitState {
|
||||||
bool halt_requested = false;
|
bool halt_requested = false;
|
||||||
bool check_bit = false;
|
bool check_bit = false;
|
||||||
|
|
||||||
|
// Exclusive state
|
||||||
|
static constexpr u32 RESERVATION_GRANULE_MASK = 0xFFFFFFF8;
|
||||||
|
u32 exclusive_state = 0;
|
||||||
|
u32 exclusive_address = 0;
|
||||||
|
|
||||||
static constexpr size_t RSBSize = 8; // MUST be a power of 2.
|
static constexpr size_t RSBSize = 8; // MUST be a power of 2.
|
||||||
static constexpr size_t RSBPtrMask = RSBSize - 1;
|
static constexpr size_t RSBPtrMask = RSBSize - 1;
|
||||||
u32 rsb_ptr = 0;
|
u32 rsb_ptr = 0;
|
||||||
|
|
|
@ -28,6 +28,25 @@ void EmitX64::EmitPack2x32To1x64(EmitContext& ctx, IR::Inst* inst) {
|
||||||
ctx.reg_alloc.DefineValue(inst, lo);
|
ctx.reg_alloc.DefineValue(inst, lo);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void EmitX64::EmitPack2x64To1x128(EmitContext& ctx, IR::Inst* inst) {
|
||||||
|
auto args = ctx.reg_alloc.GetArgumentInfo(inst);
|
||||||
|
Xbyak::Reg64 lo = ctx.reg_alloc.UseGpr(args[0]);
|
||||||
|
Xbyak::Reg64 hi = ctx.reg_alloc.UseGpr(args[1]);
|
||||||
|
Xbyak::Xmm result = ctx.reg_alloc.ScratchXmm();
|
||||||
|
|
||||||
|
if (code.DoesCpuSupport(Xbyak::util::Cpu::tSSE41)) {
|
||||||
|
code.movq(result, lo);
|
||||||
|
code.pinsrq(result, hi, 1);
|
||||||
|
} else {
|
||||||
|
Xbyak::Xmm tmp = ctx.reg_alloc.ScratchXmm();
|
||||||
|
code.movq(result, lo);
|
||||||
|
code.movq(tmp, hi);
|
||||||
|
code.punpcklqdq(result, tmp);
|
||||||
|
}
|
||||||
|
|
||||||
|
ctx.reg_alloc.DefineValue(inst, result);
|
||||||
|
}
|
||||||
|
|
||||||
void EmitX64::EmitLeastSignificantWord(EmitContext& ctx, IR::Inst* inst) {
|
void EmitX64::EmitLeastSignificantWord(EmitContext& ctx, IR::Inst* inst) {
|
||||||
auto args = ctx.reg_alloc.GetArgumentInfo(inst);
|
auto args = ctx.reg_alloc.GetArgumentInfo(inst);
|
||||||
ctx.reg_alloc.DefineValue(inst, args[0]);
|
ctx.reg_alloc.DefineValue(inst, args[0]);
|
||||||
|
|
|
@ -61,7 +61,7 @@ INST(SEVL, "SEVL", "11010
|
||||||
//INST(AUTIB_2, "AUTIB, AUTIB1716, AUTIBSP, AUTIBZ, AUTIZB", "1101010100000011001000-111-11111")
|
//INST(AUTIB_2, "AUTIB, AUTIB1716, AUTIBSP, AUTIBZ, AUTIZB", "1101010100000011001000-111-11111")
|
||||||
//INST(ESB, "ESB", "11010101000000110010001000011111")
|
//INST(ESB, "ESB", "11010101000000110010001000011111")
|
||||||
//INST(PSB, "PSB CSYNC", "11010101000000110010001000111111")
|
//INST(PSB, "PSB CSYNC", "11010101000000110010001000111111")
|
||||||
//INST(CLREX, "CLREX", "11010101000000110011MMMM01011111")
|
INST(CLREX, "CLREX", "11010101000000110011MMMM01011111")
|
||||||
INST(DSB, "DSB", "11010101000000110011MMMM10011111")
|
INST(DSB, "DSB", "11010101000000110011MMMM10011111")
|
||||||
INST(DMB, "DMB", "11010101000000110011MMMM10111111")
|
INST(DMB, "DMB", "11010101000000110011MMMM10111111")
|
||||||
//INST(ISB, "ISB", "11010101000000110011MMMM11011111")
|
//INST(ISB, "ISB", "11010101000000110011MMMM11011111")
|
||||||
|
@ -135,14 +135,14 @@ INST(LDx_mult_2, "LDx (multiple structures)", "0Q001
|
||||||
//INST(LD4R_2, "LD4R", "0Q001101111mmmmm1110zznnnnnttttt")
|
//INST(LD4R_2, "LD4R", "0Q001101111mmmmm1110zznnnnnttttt")
|
||||||
|
|
||||||
// Loads and stores - Load/Store Exclusive
|
// Loads and stores - Load/Store Exclusive
|
||||||
//INST(STXR, "STXRB, STXRH, STXR", "zz001000000sssss011111nnnnnttttt")
|
INST(STXR, "STXRB, STXRH, STXR", "zz001000000sssss011111nnnnnttttt")
|
||||||
//INST(STLXR, "STLXRB, STLXRH, STLXR", "zz001000000sssss111111nnnnnttttt")
|
INST(STLXR, "STLXRB, STLXRH, STLXR", "zz001000000sssss111111nnnnnttttt")
|
||||||
//INST(STXP, "STXP", "1z001000001sssss0uuuuunnnnnttttt")
|
INST(STXP, "STXP", "1z001000001sssss0uuuuunnnnnttttt")
|
||||||
//INST(STLXP, "STLXP", "1z001000001sssss1uuuuunnnnnttttt")
|
INST(STLXP, "STLXP", "1z001000001sssss1uuuuunnnnnttttt")
|
||||||
//INST(LDXR, "LDXRB, LDXRH, LDXR", "zz00100001011111011111nnnnnttttt")
|
INST(LDXR, "LDXRB, LDXRH, LDXR", "zz00100001011111011111nnnnnttttt")
|
||||||
//INST(LDAXRB, "LDAXRB", "zz00100001011111111111nnnnnttttt")
|
INST(LDAXR, "LDAXRB, LDAXRH, LDAXR", "zz00100001011111111111nnnnnttttt")
|
||||||
//INST(LDXP, "LDXP", "1z001000011111110uuuuunnnnnttttt")
|
INST(LDXP, "LDXP", "1z001000011111110uuuuunnnnnttttt")
|
||||||
//INST(LDAXP, "LDAXP", "1z001000011111111uuuuunnnnnttttt")
|
INST(LDAXP, "LDAXP", "1z001000011111111uuuuunnnnnttttt")
|
||||||
INST(STLLR, "STLLRB, STLLRH, STLLR", "zz00100010011111011111nnnnnttttt")
|
INST(STLLR, "STLLRB, STLLRH, STLLR", "zz00100010011111011111nnnnnttttt")
|
||||||
INST(STLR, "STLRB, STLRH, STLR", "zz00100010011111111111nnnnnttttt")
|
INST(STLR, "STLRB, STLRH, STLR", "zz00100010011111111111nnnnnttttt")
|
||||||
INST(LDLAR, "LDLARB, LDLARH, LDLAR", "zz00100011011111011111nnnnnttttt")
|
INST(LDLAR, "LDLARB, LDLARH, LDLAR", "zz00100011011111011111nnnnnttttt")
|
||||||
|
|
|
@ -61,6 +61,15 @@ IR::U64 IREmitter::GetTPIDRRO() {
|
||||||
return Inst<IR::U64>(Opcode::A64GetTPIDRRO);
|
return Inst<IR::U64>(Opcode::A64GetTPIDRRO);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void IREmitter::ClearExclusive() {
|
||||||
|
Inst(Opcode::A64ClearExclusive);
|
||||||
|
}
|
||||||
|
|
||||||
|
void IREmitter::SetExclusive(const IR::U64& vaddr, size_t byte_size) {
|
||||||
|
ASSERT(byte_size == 1 || byte_size == 2 || byte_size == 4 || byte_size == 8 || byte_size == 16);
|
||||||
|
Inst(Opcode::A64SetExclusive, vaddr, Imm8(u8(byte_size)));
|
||||||
|
}
|
||||||
|
|
||||||
IR::U8 IREmitter::ReadMemory8(const IR::U64& vaddr) {
|
IR::U8 IREmitter::ReadMemory8(const IR::U64& vaddr) {
|
||||||
return Inst<IR::U8>(Opcode::A64ReadMemory8, vaddr);
|
return Inst<IR::U8>(Opcode::A64ReadMemory8, vaddr);
|
||||||
}
|
}
|
||||||
|
@ -101,6 +110,26 @@ void IREmitter::WriteMemory128(const IR::U64& vaddr, const IR::U128& value) {
|
||||||
Inst(Opcode::A64WriteMemory128, vaddr, value);
|
Inst(Opcode::A64WriteMemory128, vaddr, value);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
IR::U32 IREmitter::ExclusiveWriteMemory8(const IR::U64& vaddr, const IR::U8& value) {
|
||||||
|
return Inst<IR::U32>(Opcode::A64ExclusiveWriteMemory8, vaddr, value);
|
||||||
|
}
|
||||||
|
|
||||||
|
IR::U32 IREmitter::ExclusiveWriteMemory16(const IR::U64& vaddr, const IR::U16& value) {
|
||||||
|
return Inst<IR::U32>(Opcode::A64ExclusiveWriteMemory16, vaddr, value);
|
||||||
|
}
|
||||||
|
|
||||||
|
IR::U32 IREmitter::ExclusiveWriteMemory32(const IR::U64& vaddr, const IR::U32& value) {
|
||||||
|
return Inst<IR::U32>(Opcode::A64ExclusiveWriteMemory32, vaddr, value);
|
||||||
|
}
|
||||||
|
|
||||||
|
IR::U32 IREmitter::ExclusiveWriteMemory64(const IR::U64& vaddr, const IR::U64& value) {
|
||||||
|
return Inst<IR::U32>(Opcode::A64ExclusiveWriteMemory64, vaddr, value);
|
||||||
|
}
|
||||||
|
|
||||||
|
IR::U32 IREmitter::ExclusiveWriteMemory128(const IR::U64& vaddr, const IR::U128& value) {
|
||||||
|
return Inst<IR::U32>(Opcode::A64ExclusiveWriteMemory128, vaddr, value);
|
||||||
|
}
|
||||||
|
|
||||||
IR::U32 IREmitter::GetW(Reg reg) {
|
IR::U32 IREmitter::GetW(Reg reg) {
|
||||||
if (reg == Reg::ZR)
|
if (reg == Reg::ZR)
|
||||||
return Imm32(0);
|
return Imm32(0);
|
||||||
|
|
|
@ -47,6 +47,8 @@ public:
|
||||||
IR::U32 GetDCZID();
|
IR::U32 GetDCZID();
|
||||||
IR::U64 GetTPIDRRO();
|
IR::U64 GetTPIDRRO();
|
||||||
|
|
||||||
|
void ClearExclusive();
|
||||||
|
void SetExclusive(const IR::U64& vaddr, size_t byte_size);
|
||||||
IR::U8 ReadMemory8(const IR::U64& vaddr);
|
IR::U8 ReadMemory8(const IR::U64& vaddr);
|
||||||
IR::U16 ReadMemory16(const IR::U64& vaddr);
|
IR::U16 ReadMemory16(const IR::U64& vaddr);
|
||||||
IR::U32 ReadMemory32(const IR::U64& vaddr);
|
IR::U32 ReadMemory32(const IR::U64& vaddr);
|
||||||
|
@ -57,6 +59,11 @@ public:
|
||||||
void WriteMemory32(const IR::U64& vaddr, const IR::U32& value);
|
void WriteMemory32(const IR::U64& vaddr, const IR::U32& value);
|
||||||
void WriteMemory64(const IR::U64& vaddr, const IR::U64& value);
|
void WriteMemory64(const IR::U64& vaddr, const IR::U64& value);
|
||||||
void WriteMemory128(const IR::U64& vaddr, const IR::U128& value);
|
void WriteMemory128(const IR::U64& vaddr, const IR::U128& value);
|
||||||
|
IR::U32 ExclusiveWriteMemory8(const IR::U64& vaddr, const IR::U8& value);
|
||||||
|
IR::U32 ExclusiveWriteMemory16(const IR::U64& vaddr, const IR::U16& value);
|
||||||
|
IR::U32 ExclusiveWriteMemory32(const IR::U64& vaddr, const IR::U32& value);
|
||||||
|
IR::U32 ExclusiveWriteMemory64(const IR::U64& vaddr, const IR::U64& value);
|
||||||
|
IR::U32 ExclusiveWriteMemory128(const IR::U64& vaddr, const IR::U128& value);
|
||||||
|
|
||||||
IR::U32 GetW(Reg source_reg);
|
IR::U32 GetW(Reg source_reg);
|
||||||
IR::U64 GetX(Reg source_reg);
|
IR::U64 GetX(Reg source_reg);
|
||||||
|
|
|
@ -308,6 +308,24 @@ void TranslatorVisitor::Mem(IR::U64 address, size_t bytesize, AccType /*acctype*
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
IR::U32 TranslatorVisitor::ExclusiveMem(IR::U64 address, size_t bytesize, AccType /*acctype*/, IR::UAnyU128 value) {
|
||||||
|
switch (bytesize) {
|
||||||
|
case 1:
|
||||||
|
return ir.ExclusiveWriteMemory8(address, value);
|
||||||
|
case 2:
|
||||||
|
return ir.ExclusiveWriteMemory16(address, value);
|
||||||
|
case 4:
|
||||||
|
return ir.ExclusiveWriteMemory32(address, value);
|
||||||
|
case 8:
|
||||||
|
return ir.ExclusiveWriteMemory64(address, value);
|
||||||
|
case 16:
|
||||||
|
return ir.ExclusiveWriteMemory128(address, value);
|
||||||
|
default:
|
||||||
|
ASSERT_MSG(false, "Invalid bytesize parameter {}", bytesize);
|
||||||
|
return {};
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
IR::U32U64 TranslatorVisitor::SignExtend(IR::UAny value, size_t to_size) {
|
IR::U32U64 TranslatorVisitor::SignExtend(IR::UAny value, size_t to_size) {
|
||||||
switch (to_size) {
|
switch (to_size) {
|
||||||
case 32:
|
case 32:
|
||||||
|
|
|
@ -63,6 +63,7 @@ struct TranslatorVisitor final {
|
||||||
|
|
||||||
IR::UAnyU128 Mem(IR::U64 address, size_t size, AccType acctype);
|
IR::UAnyU128 Mem(IR::U64 address, size_t size, AccType acctype);
|
||||||
void Mem(IR::U64 address, size_t size, AccType acctype, IR::UAnyU128 value);
|
void Mem(IR::U64 address, size_t size, AccType acctype, IR::UAnyU128 value);
|
||||||
|
IR::U32 ExclusiveMem(IR::U64 address, size_t size, AccType acctype, IR::UAnyU128 value);
|
||||||
|
|
||||||
IR::U32U64 SignExtend(IR::UAny value, size_t to_size);
|
IR::U32U64 SignExtend(IR::UAny value, size_t to_size);
|
||||||
IR::U32U64 ZeroExtend(IR::UAny value, size_t to_size);
|
IR::U32U64 ZeroExtend(IR::UAny value, size_t to_size);
|
||||||
|
@ -211,7 +212,7 @@ struct TranslatorVisitor final {
|
||||||
bool STXP(Imm<1> size, Reg Rs, Reg Rt2, Reg Rn, Reg Rt);
|
bool STXP(Imm<1> size, Reg Rs, Reg Rt2, Reg Rn, Reg Rt);
|
||||||
bool STLXP(Imm<1> size, Reg Rs, Reg Rt2, Reg Rn, Reg Rt);
|
bool STLXP(Imm<1> size, Reg Rs, Reg Rt2, Reg Rn, Reg Rt);
|
||||||
bool LDXR(Imm<2> size, Reg Rn, Reg Rt);
|
bool LDXR(Imm<2> size, Reg Rn, Reg Rt);
|
||||||
bool LDAXRB(Imm<2> size, Reg Rn, Reg Rt);
|
bool LDAXR(Imm<2> size, Reg Rn, Reg Rt);
|
||||||
bool LDXP(Imm<1> size, Reg Rt2, Reg Rn, Reg Rt);
|
bool LDXP(Imm<1> size, Reg Rt2, Reg Rn, Reg Rt);
|
||||||
bool LDAXP(Imm<1> size, Reg Rt2, Reg Rn, Reg Rt);
|
bool LDAXP(Imm<1> size, Reg Rt2, Reg Rn, Reg Rt);
|
||||||
bool STLLR(Imm<2> size, Reg Rn, Reg Rt);
|
bool STLLR(Imm<2> size, Reg Rn, Reg Rt);
|
||||||
|
|
|
@ -4,10 +4,140 @@
|
||||||
* General Public License version 2 or any later version.
|
* General Public License version 2 or any later version.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
|
#include <boost/optional.hpp>
|
||||||
|
|
||||||
#include "frontend/A64/translate/impl/impl.h"
|
#include "frontend/A64/translate/impl/impl.h"
|
||||||
|
|
||||||
namespace Dynarmic::A64 {
|
namespace Dynarmic::A64 {
|
||||||
|
|
||||||
|
static bool ExclusiveSharedDecodeAndOperation(TranslatorVisitor& tv, IREmitter& ir, bool pair, size_t size, bool L, bool o0, boost::optional<Reg> Rs, boost::optional<Reg> Rt2, Reg Rn, Reg Rt) {
|
||||||
|
// Shared Decode
|
||||||
|
|
||||||
|
const AccType acctype = o0 ? AccType::ORDERED : AccType::ATOMIC;
|
||||||
|
const MemOp memop = L ? MemOp::LOAD : MemOp::STORE;
|
||||||
|
const size_t elsize = 8 << size;
|
||||||
|
const size_t regsize = elsize == 64 ? 64 : 32;
|
||||||
|
const size_t datasize = pair ? elsize * 2 : elsize;
|
||||||
|
|
||||||
|
// Operation
|
||||||
|
|
||||||
|
const size_t dbytes = datasize / 8;
|
||||||
|
|
||||||
|
if (memop == MemOp::LOAD && pair && Rt == *Rt2) {
|
||||||
|
return tv.UnpredictableInstruction();
|
||||||
|
} else if (memop == MemOp::STORE && (*Rs == Rt || (pair && *Rs == *Rt2))) {
|
||||||
|
return tv.UnpredictableInstruction();
|
||||||
|
} else if (memop == MemOp::STORE && *Rs == Rn && Rn != Reg::R31) {
|
||||||
|
return tv.UnpredictableInstruction();
|
||||||
|
}
|
||||||
|
|
||||||
|
IR::U64 address;
|
||||||
|
if (Rn == Reg::SP) {
|
||||||
|
// TODO: Check SP Alignment
|
||||||
|
address = tv.SP(64);
|
||||||
|
} else {
|
||||||
|
address = tv.X(64, Rn);
|
||||||
|
}
|
||||||
|
|
||||||
|
switch (memop) {
|
||||||
|
case MemOp::STORE: {
|
||||||
|
IR::UAnyU128 data;
|
||||||
|
if (pair && elsize == 64) {
|
||||||
|
data = ir.Pack2x64To1x128(tv.X(64, Rt), tv.X(64, *Rt2));
|
||||||
|
} else if (pair && elsize == 32) {
|
||||||
|
data = ir.Pack2x32To1x64(tv.X(32, Rt), tv.X(32, *Rt2));
|
||||||
|
} else {
|
||||||
|
data = tv.X(datasize, Rt);
|
||||||
|
}
|
||||||
|
IR::U32 status = tv.ExclusiveMem(address, dbytes, acctype, data);
|
||||||
|
tv.X(32, *Rs, status);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
case MemOp::LOAD: {
|
||||||
|
ir.SetExclusive(address, dbytes);
|
||||||
|
IR::UAnyU128 data = tv.Mem(address, dbytes, acctype);
|
||||||
|
if (pair && elsize == 64) {
|
||||||
|
tv.X(64, Rt, ir.VectorGetElement(64, data, 0));
|
||||||
|
tv.X(64, *Rt2, ir.VectorGetElement(64, data, 1));
|
||||||
|
} else if (pair && elsize == 32) {
|
||||||
|
tv.X(32, Rt, ir.LeastSignificantWord(data));
|
||||||
|
tv.X(32, *Rt2, ir.MostSignificantWord(data).result);
|
||||||
|
} else {
|
||||||
|
tv.X(regsize, Rt, tv.ZeroExtend(data, regsize));
|
||||||
|
}
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
UNREACHABLE();
|
||||||
|
}
|
||||||
|
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
bool TranslatorVisitor::STXR(Imm<2> sz, Reg Rs, Reg Rn, Reg Rt) {
|
||||||
|
const bool pair = false;
|
||||||
|
const size_t size = sz.ZeroExtend<size_t>();
|
||||||
|
const bool L = 0;
|
||||||
|
const bool o0 = 0;
|
||||||
|
return ExclusiveSharedDecodeAndOperation(*this, ir, pair, size, L, o0, Rs, {}, Rn, Rt);
|
||||||
|
}
|
||||||
|
|
||||||
|
bool TranslatorVisitor::STLXR(Imm<2> sz, Reg Rs, Reg Rn, Reg Rt) {
|
||||||
|
const bool pair = false;
|
||||||
|
const size_t size = sz.ZeroExtend<size_t>();
|
||||||
|
const bool L = 0;
|
||||||
|
const bool o0 = 1;
|
||||||
|
return ExclusiveSharedDecodeAndOperation(*this, ir, pair, size, L, o0, Rs, {}, Rn, Rt);
|
||||||
|
}
|
||||||
|
|
||||||
|
bool TranslatorVisitor::STXP(Imm<1> sz, Reg Rs, Reg Rt2, Reg Rn, Reg Rt) {
|
||||||
|
const bool pair = true;
|
||||||
|
const size_t size = concatenate(Imm<1>{1}, sz).ZeroExtend<size_t>();
|
||||||
|
const bool L = 0;
|
||||||
|
const bool o0 = 0;
|
||||||
|
return ExclusiveSharedDecodeAndOperation(*this, ir, pair, size, L, o0, Rs, Rt2, Rn, Rt);
|
||||||
|
}
|
||||||
|
|
||||||
|
bool TranslatorVisitor::STLXP(Imm<1> sz, Reg Rs, Reg Rt2, Reg Rn, Reg Rt) {
|
||||||
|
const bool pair = true;
|
||||||
|
const size_t size = concatenate(Imm<1>{1}, sz).ZeroExtend<size_t>();
|
||||||
|
const bool L = 0;
|
||||||
|
const bool o0 = 1;
|
||||||
|
return ExclusiveSharedDecodeAndOperation(*this, ir, pair, size, L, o0, Rs, Rt2, Rn, Rt);
|
||||||
|
}
|
||||||
|
|
||||||
|
bool TranslatorVisitor::LDXR(Imm<2> sz, Reg Rn, Reg Rt) {
|
||||||
|
const bool pair = false;
|
||||||
|
const size_t size = sz.ZeroExtend<size_t>();
|
||||||
|
const bool L = 1;
|
||||||
|
const bool o0 = 0;
|
||||||
|
return ExclusiveSharedDecodeAndOperation(*this, ir, pair, size, L, o0, {}, {}, Rn, Rt);
|
||||||
|
}
|
||||||
|
|
||||||
|
bool TranslatorVisitor::LDAXR(Imm<2> sz, Reg Rn, Reg Rt) {
|
||||||
|
const bool pair = false;
|
||||||
|
const size_t size = sz.ZeroExtend<size_t>();
|
||||||
|
const bool L = 1;
|
||||||
|
const bool o0 = 1;
|
||||||
|
return ExclusiveSharedDecodeAndOperation(*this, ir, pair, size, L, o0, {}, {}, Rn, Rt);
|
||||||
|
}
|
||||||
|
|
||||||
|
bool TranslatorVisitor::LDXP(Imm<1> sz, Reg Rt2, Reg Rn, Reg Rt) {
|
||||||
|
const bool pair = true;
|
||||||
|
const size_t size = concatenate(Imm<1>{1}, sz).ZeroExtend<size_t>();
|
||||||
|
const bool L = 1;
|
||||||
|
const bool o0 = 0;
|
||||||
|
return ExclusiveSharedDecodeAndOperation(*this, ir, pair, size, L, o0, {}, Rt2, Rn, Rt);
|
||||||
|
}
|
||||||
|
|
||||||
|
bool TranslatorVisitor::LDAXP(Imm<1> sz, Reg Rt2, Reg Rn, Reg Rt) {
|
||||||
|
const bool pair = true;
|
||||||
|
const size_t size = concatenate(Imm<1>{1}, sz).ZeroExtend<size_t>();
|
||||||
|
const bool L = 1;
|
||||||
|
const bool o0 = 1;
|
||||||
|
return ExclusiveSharedDecodeAndOperation(*this, ir, pair, size, L, o0, {}, Rt2, Rn, Rt);
|
||||||
|
}
|
||||||
|
|
||||||
static bool OrderedSharedDecodeAndOperation(TranslatorVisitor& tv, size_t size, bool L, bool o0, Reg Rn, Reg Rt) {
|
static bool OrderedSharedDecodeAndOperation(TranslatorVisitor& tv, size_t size, bool L, bool o0, Reg Rn, Reg Rt) {
|
||||||
// Shared Decode
|
// Shared Decode
|
||||||
|
|
||||||
|
|
|
@ -36,6 +36,11 @@ bool TranslatorVisitor::SEVL() {
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
bool TranslatorVisitor::CLREX(Imm<4> /*CRm*/) {
|
||||||
|
ir.ClearExclusive();
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
bool TranslatorVisitor::DSB(Imm<4> /*CRm*/) {
|
bool TranslatorVisitor::DSB(Imm<4> /*CRm*/) {
|
||||||
ir.DataSynchronizationBarrier();
|
ir.DataSynchronizationBarrier();
|
||||||
return true;
|
return true;
|
||||||
|
|
|
@ -38,6 +38,10 @@ U64 IREmitter::Pack2x32To1x64(const U32& lo, const U32& hi) {
|
||||||
return Inst<U64>(Opcode::Pack2x32To1x64, lo, hi);
|
return Inst<U64>(Opcode::Pack2x32To1x64, lo, hi);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
U128 IREmitter::Pack2x64To1x128(const U64& lo, const U64& hi) {
|
||||||
|
return Inst<U128>(Opcode::Pack2x64To1x128, lo, hi);
|
||||||
|
}
|
||||||
|
|
||||||
U32 IREmitter::LeastSignificantWord(const U64& value) {
|
U32 IREmitter::LeastSignificantWord(const U64& value) {
|
||||||
return Inst<U32>(Opcode::LeastSignificantWord, value);
|
return Inst<U32>(Opcode::LeastSignificantWord, value);
|
||||||
}
|
}
|
||||||
|
|
|
@ -68,6 +68,7 @@ public:
|
||||||
void PushRSB(const LocationDescriptor& return_location);
|
void PushRSB(const LocationDescriptor& return_location);
|
||||||
|
|
||||||
U64 Pack2x32To1x64(const U32& lo, const U32& hi);
|
U64 Pack2x32To1x64(const U32& lo, const U32& hi);
|
||||||
|
U128 Pack2x64To1x128(const U64& lo, const U64& hi);
|
||||||
U32 LeastSignificantWord(const U64& value);
|
U32 LeastSignificantWord(const U64& value);
|
||||||
ResultAndCarry<U32> MostSignificantWord(const U64& value);
|
ResultAndCarry<U32> MostSignificantWord(const U64& value);
|
||||||
U16 LeastSignificantHalf(U32U64 value);
|
U16 LeastSignificantHalf(U32U64 value);
|
||||||
|
|
|
@ -89,6 +89,11 @@ bool Inst::IsExclusiveMemoryWrite() const {
|
||||||
case Opcode::A32ExclusiveWriteMemory16:
|
case Opcode::A32ExclusiveWriteMemory16:
|
||||||
case Opcode::A32ExclusiveWriteMemory32:
|
case Opcode::A32ExclusiveWriteMemory32:
|
||||||
case Opcode::A32ExclusiveWriteMemory64:
|
case Opcode::A32ExclusiveWriteMemory64:
|
||||||
|
case Opcode::A64ExclusiveWriteMemory8:
|
||||||
|
case Opcode::A64ExclusiveWriteMemory16:
|
||||||
|
case Opcode::A64ExclusiveWriteMemory32:
|
||||||
|
case Opcode::A64ExclusiveWriteMemory64:
|
||||||
|
case Opcode::A64ExclusiveWriteMemory128:
|
||||||
return true;
|
return true;
|
||||||
|
|
||||||
default:
|
default:
|
||||||
|
@ -249,6 +254,8 @@ bool Inst::CausesCPUException() const {
|
||||||
bool Inst::AltersExclusiveState() const {
|
bool Inst::AltersExclusiveState() const {
|
||||||
return op == Opcode::A32ClearExclusive ||
|
return op == Opcode::A32ClearExclusive ||
|
||||||
op == Opcode::A32SetExclusive ||
|
op == Opcode::A32SetExclusive ||
|
||||||
|
op == Opcode::A64ClearExclusive ||
|
||||||
|
op == Opcode::A64SetExclusive ||
|
||||||
IsExclusiveMemoryWrite();
|
IsExclusiveMemoryWrite();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -77,6 +77,7 @@ OPCODE(NZCVFromPackedFlags, T::NZCVFlags, T::U32
|
||||||
|
|
||||||
// Calculations
|
// Calculations
|
||||||
OPCODE(Pack2x32To1x64, T::U64, T::U32, T::U32 )
|
OPCODE(Pack2x32To1x64, T::U64, T::U32, T::U32 )
|
||||||
|
OPCODE(Pack2x64To1x128, T::U128, T::U64, T::U64 )
|
||||||
OPCODE(LeastSignificantWord, T::U32, T::U64 )
|
OPCODE(LeastSignificantWord, T::U32, T::U64 )
|
||||||
OPCODE(MostSignificantWord, T::U32, T::U64 )
|
OPCODE(MostSignificantWord, T::U32, T::U64 )
|
||||||
OPCODE(LeastSignificantHalf, T::U16, T::U32 )
|
OPCODE(LeastSignificantHalf, T::U16, T::U32 )
|
||||||
|
@ -325,6 +326,8 @@ A32OPC(ExclusiveWriteMemory32, T::U32, T::U32, T::U32
|
||||||
A32OPC(ExclusiveWriteMemory64, T::U32, T::U32, T::U32, T::U32 )
|
A32OPC(ExclusiveWriteMemory64, T::U32, T::U32, T::U32, T::U32 )
|
||||||
|
|
||||||
// A64 Memory access
|
// A64 Memory access
|
||||||
|
A64OPC(ClearExclusive, T::Void, )
|
||||||
|
A64OPC(SetExclusive, T::Void, T::U64, T::U8 )
|
||||||
A64OPC(ReadMemory8, T::U8, T::U64 )
|
A64OPC(ReadMemory8, T::U8, T::U64 )
|
||||||
A64OPC(ReadMemory16, T::U16, T::U64 )
|
A64OPC(ReadMemory16, T::U16, T::U64 )
|
||||||
A64OPC(ReadMemory32, T::U32, T::U64 )
|
A64OPC(ReadMemory32, T::U32, T::U64 )
|
||||||
|
@ -335,6 +338,11 @@ A64OPC(WriteMemory16, T::Void, T::U64, T::U16
|
||||||
A64OPC(WriteMemory32, T::Void, T::U64, T::U32 )
|
A64OPC(WriteMemory32, T::Void, T::U64, T::U32 )
|
||||||
A64OPC(WriteMemory64, T::Void, T::U64, T::U64 )
|
A64OPC(WriteMemory64, T::Void, T::U64, T::U64 )
|
||||||
A64OPC(WriteMemory128, T::Void, T::U64, T::U128 )
|
A64OPC(WriteMemory128, T::Void, T::U64, T::U128 )
|
||||||
|
A64OPC(ExclusiveWriteMemory8, T::U32, T::U64, T::U8 )
|
||||||
|
A64OPC(ExclusiveWriteMemory16, T::U32, T::U64, T::U16 )
|
||||||
|
A64OPC(ExclusiveWriteMemory32, T::U32, T::U64, T::U32 )
|
||||||
|
A64OPC(ExclusiveWriteMemory64, T::U32, T::U64, T::U64 )
|
||||||
|
A64OPC(ExclusiveWriteMemory128, T::U32, T::U64, T::U128 )
|
||||||
|
|
||||||
// Coprocessor
|
// Coprocessor
|
||||||
A32OPC(CoprocInternalOperation, T::Void, T::CoprocInfo )
|
A32OPC(CoprocInternalOperation, T::Void, T::CoprocInfo )
|
||||||
|
|
Loading…
Reference in a new issue