/* This file is part of the dynarmic project. * Copyright (c) 2016 MerryMage * This software may be used and distributed according to the terms of the GNU * General Public License version 2 or any later version. */ #include #include "backend_x64/block_of_code.h" #include "backend_x64/jitstate.h" #include "common/x64/abi.h" using namespace Gen; namespace Dynarmic { namespace BackendX64 { BlockOfCode::BlockOfCode() : Gen::XCodeBlock() { AllocCodeSpace(128 * 1024 * 1024); ClearCache(false); } void BlockOfCode::ClearCache(bool poison_memory) { if (poison_memory) { ClearCodeSpace(); } else { ResetCodePtr(); } GenConstants(); GenRunCode(); GenReturnFromRunCode(); } size_t BlockOfCode::RunCode(JitState* jit_state, CodePtr basic_block, size_t cycles_to_run) const { constexpr size_t max_cycles_to_run = static_cast(std::numeric_limitscycles_remaining)>::max()); ASSERT(cycles_to_run <= max_cycles_to_run); jit_state->cycles_remaining = cycles_to_run; run_code(jit_state, basic_block); return cycles_to_run - jit_state->cycles_remaining; // Return number of cycles actually run. } void BlockOfCode::ReturnFromRunCode(bool MXCSR_switch) { JMP(MXCSR_switch ? return_from_run_code : return_from_run_code_without_mxcsr_switch, true); } void BlockOfCode::GenConstants() { const_FloatNegativeZero32 = AlignCode16(); Write32(0x80000000u); const_FloatNaN32 = AlignCode16(); Write32(0x7fc00000u); const_FloatNonSignMask32 = AlignCode16(); Write64(0x7fffffffu); const_FloatNegativeZero64 = AlignCode16(); Write64(0x8000000000000000u); const_FloatNaN64 = AlignCode16(); Write64(0x7ff8000000000000u); const_FloatNonSignMask64 = AlignCode16(); Write64(0x7fffffffffffffffu); const_FloatPenultimatePositiveDenormal64 = AlignCode16(); Write64(0x000ffffffffffffeu); AlignCode16(); } void BlockOfCode::GenRunCode() { run_code = reinterpret_cast(const_cast(GetCodePtr())); // This serves two purposes: // 1. It saves all the registers we as a callee need to save. // 2. It aligns the stack so that the code the JIT emits can assume // that the stack is appropriately aligned for CALLs. ABI_PushRegistersAndAdjustStack(ABI_ALL_CALLEE_SAVED, 8); MOV(64, R(R15), R(ABI_PARAM1)); SwitchMxcsrOnEntry(); JMPptr(R(ABI_PARAM2)); } void BlockOfCode::GenReturnFromRunCode() { return_from_run_code = GetCodePtr(); SwitchMxcsrOnExit(); return_from_run_code_without_mxcsr_switch = GetCodePtr(); ABI_PopRegistersAndAdjustStack(ABI_ALL_CALLEE_SAVED, 8); RET(); } void BlockOfCode::SwitchMxcsrOnEntry() { STMXCSR(MDisp(R15, offsetof(JitState, save_host_MXCSR))); LDMXCSR(MDisp(R15, offsetof(JitState, guest_MXCSR))); } void BlockOfCode::SwitchMxcsrOnExit() { STMXCSR(MDisp(R15, offsetof(JitState, guest_MXCSR))); LDMXCSR(MDisp(R15, offsetof(JitState, save_host_MXCSR))); } } // namespace BackendX64 } // namespace Dynarmic