// Copyright 2018 yuzu emulator team // Licensed under GPLv2 or any later version // Refer to the license.txt file included. #include #include #include #include #include "common/logging/log.h" #include "core/arm/dynarmic/arm_dynarmic.h" #include "core/core.h" #include "core/core_timing.h" #include "core/hle/kernel/process.h" #include "core/hle/kernel/svc.h" #include "core/memory.h" namespace Core { using Vector = Dynarmic::A64::Vector; class ARM_Dynarmic_Callbacks : public Dynarmic::A64::UserCallbacks { public: explicit ARM_Dynarmic_Callbacks(ARM_Dynarmic& parent) : parent(parent) {} ~ARM_Dynarmic_Callbacks() = default; u8 MemoryRead8(u64 vaddr) override { return Memory::Read8(vaddr); } u16 MemoryRead16(u64 vaddr) override { return Memory::Read16(vaddr); } u32 MemoryRead32(u64 vaddr) override { return Memory::Read32(vaddr); } u64 MemoryRead64(u64 vaddr) override { return Memory::Read64(vaddr); } Vector MemoryRead128(u64 vaddr) override { return {Memory::Read64(vaddr), Memory::Read64(vaddr + 8)}; } void MemoryWrite8(u64 vaddr, u8 value) override { Memory::Write8(vaddr, value); } void MemoryWrite16(u64 vaddr, u16 value) override { Memory::Write16(vaddr, value); } void MemoryWrite32(u64 vaddr, u32 value) override { Memory::Write32(vaddr, value); } void MemoryWrite64(u64 vaddr, u64 value) override { Memory::Write64(vaddr, value); } void MemoryWrite128(u64 vaddr, Vector value) override { Memory::Write64(vaddr, value[0]); Memory::Write64(vaddr + 8, value[1]); } void InterpreterFallback(u64 pc, size_t num_instructions) override { LOG_INFO(Core_ARM, "Unicorn fallback @ 0x{:X} for {} instructions (instr = {:08X})", pc, num_instructions, MemoryReadCode(pc)); ARM_Interface::ThreadContext ctx; parent.SaveContext(ctx); parent.inner_unicorn.LoadContext(ctx); parent.inner_unicorn.ExecuteInstructions(static_cast(num_instructions)); parent.inner_unicorn.SaveContext(ctx); parent.LoadContext(ctx); num_interpreted_instructions += num_instructions; } void ExceptionRaised(u64 pc, Dynarmic::A64::Exception exception) override { switch (exception) { case Dynarmic::A64::Exception::WaitForInterrupt: case Dynarmic::A64::Exception::WaitForEvent: case Dynarmic::A64::Exception::SendEvent: case Dynarmic::A64::Exception::SendEventLocal: case Dynarmic::A64::Exception::Yield: return; default: ASSERT_MSG(false, "ExceptionRaised(exception = {}, pc = {:X})", static_cast(exception), pc); } } void CallSVC(u32 swi) override { Kernel::CallSVC(swi); } void AddTicks(u64 ticks) override { // Divide the number of ticks by the amount of CPU cores. TODO(Subv): This yields only a // rough approximation of the amount of executed ticks in the system, it may be thrown off // if not all cores are doing a similar amount of work. Instead of doing this, we should // device a way so that timing is consistent across all cores without increasing the ticks 4 // times. u64 amortized_ticks = (ticks - num_interpreted_instructions) / Core::NUM_CPU_CORES; // Always execute at least one tick. amortized_ticks = std::max(amortized_ticks, 1); CoreTiming::AddTicks(amortized_ticks); num_interpreted_instructions = 0; } u64 GetTicksRemaining() override { return std::max(CoreTiming::GetDowncount(), 0); } u64 GetCNTPCT() override { return CoreTiming::GetTicks(); } ARM_Dynarmic& parent; size_t num_interpreted_instructions = 0; u64 tpidrro_el0 = 0; u64 tpidr_el0 = 0; }; std::unique_ptr ARM_Dynarmic::MakeJit() const { auto** const page_table = Core::CurrentProcess()->vm_manager.page_table.pointers.data(); Dynarmic::A64::UserConfig config; // Callbacks config.callbacks = cb.get(); // Memory config.page_table = reinterpret_cast(page_table); config.page_table_address_space_bits = Memory::ADDRESS_SPACE_BITS; config.silently_mirror_page_table = false; // Multi-process state config.processor_id = core_index; config.global_monitor = &exclusive_monitor->monitor; // System registers config.tpidrro_el0 = &cb->tpidrro_el0; config.tpidr_el0 = &cb->tpidr_el0; config.dczid_el0 = 4; config.ctr_el0 = 0x8444c004; // Unpredictable instructions config.define_unpredictable_behaviour = true; return std::make_unique(config); } void ARM_Dynarmic::Run() { ASSERT(Memory::GetCurrentPageTable() == current_page_table); jit->Run(); } void ARM_Dynarmic::Step() { cb->InterpreterFallback(jit->GetPC(), 1); } ARM_Dynarmic::ARM_Dynarmic(std::shared_ptr exclusive_monitor, size_t core_index) : cb(std::make_unique(*this)), core_index{core_index}, exclusive_monitor{std::dynamic_pointer_cast(exclusive_monitor)} { ThreadContext ctx; inner_unicorn.SaveContext(ctx); PageTableChanged(); LoadContext(ctx); } ARM_Dynarmic::~ARM_Dynarmic() = default; void ARM_Dynarmic::MapBackingMemory(u64 address, size_t size, u8* memory, Kernel::VMAPermission perms) { inner_unicorn.MapBackingMemory(address, size, memory, perms); } void ARM_Dynarmic::UnmapMemory(u64 address, size_t size) { inner_unicorn.UnmapMemory(address, size); } void ARM_Dynarmic::SetPC(u64 pc) { jit->SetPC(pc); } u64 ARM_Dynarmic::GetPC() const { return jit->GetPC(); } u64 ARM_Dynarmic::GetReg(int index) const { return jit->GetRegister(index); } void ARM_Dynarmic::SetReg(int index, u64 value) { jit->SetRegister(index, value); } u128 ARM_Dynarmic::GetExtReg(int index) const { return jit->GetVector(index); } void ARM_Dynarmic::SetExtReg(int index, u128 value) { jit->SetVector(index, value); } u32 ARM_Dynarmic::GetVFPReg(int /*index*/) const { UNIMPLEMENTED(); return {}; } void ARM_Dynarmic::SetVFPReg(int /*index*/, u32 /*value*/) { UNIMPLEMENTED(); } u32 ARM_Dynarmic::GetCPSR() const { return jit->GetPstate(); } void ARM_Dynarmic::SetCPSR(u32 cpsr) { jit->SetPstate(cpsr); } u64 ARM_Dynarmic::GetTlsAddress() const { return cb->tpidrro_el0; } void ARM_Dynarmic::SetTlsAddress(VAddr address) { cb->tpidrro_el0 = address; } u64 ARM_Dynarmic::GetTPIDR_EL0() const { return cb->tpidr_el0; } void ARM_Dynarmic::SetTPIDR_EL0(u64 value) { cb->tpidr_el0 = value; } void ARM_Dynarmic::SaveContext(ThreadContext& ctx) { ctx.cpu_registers = jit->GetRegisters(); ctx.sp = jit->GetSP(); ctx.pc = jit->GetPC(); ctx.cpsr = jit->GetPstate(); ctx.fpu_registers = jit->GetVectors(); ctx.fpscr = jit->GetFpcr(); } void ARM_Dynarmic::LoadContext(const ThreadContext& ctx) { jit->SetRegisters(ctx.cpu_registers); jit->SetSP(ctx.sp); jit->SetPC(ctx.pc); jit->SetPstate(static_cast(ctx.cpsr)); jit->SetVectors(ctx.fpu_registers); jit->SetFpcr(static_cast(ctx.fpscr)); } void ARM_Dynarmic::PrepareReschedule() { jit->HaltExecution(); } void ARM_Dynarmic::ClearInstructionCache() { jit->ClearCache(); } void ARM_Dynarmic::ClearExclusiveState() { jit->ClearExclusiveState(); } void ARM_Dynarmic::PageTableChanged() { jit = MakeJit(); current_page_table = Memory::GetCurrentPageTable(); } DynarmicExclusiveMonitor::DynarmicExclusiveMonitor(size_t core_count) : monitor(core_count) {} DynarmicExclusiveMonitor::~DynarmicExclusiveMonitor() = default; void DynarmicExclusiveMonitor::SetExclusive(size_t core_index, VAddr addr) { // Size doesn't actually matter. monitor.Mark(core_index, addr, 16); } void DynarmicExclusiveMonitor::ClearExclusive() { monitor.Clear(); } bool DynarmicExclusiveMonitor::ExclusiveWrite8(size_t core_index, VAddr vaddr, u8 value) { return monitor.DoExclusiveOperation(core_index, vaddr, 1, [&] { Memory::Write8(vaddr, value); }); } bool DynarmicExclusiveMonitor::ExclusiveWrite16(size_t core_index, VAddr vaddr, u16 value) { return monitor.DoExclusiveOperation(core_index, vaddr, 2, [&] { Memory::Write16(vaddr, value); }); } bool DynarmicExclusiveMonitor::ExclusiveWrite32(size_t core_index, VAddr vaddr, u32 value) { return monitor.DoExclusiveOperation(core_index, vaddr, 4, [&] { Memory::Write32(vaddr, value); }); } bool DynarmicExclusiveMonitor::ExclusiveWrite64(size_t core_index, VAddr vaddr, u64 value) { return monitor.DoExclusiveOperation(core_index, vaddr, 8, [&] { Memory::Write64(vaddr, value); }); } bool DynarmicExclusiveMonitor::ExclusiveWrite128(size_t core_index, VAddr vaddr, u128 value) { return monitor.DoExclusiveOperation(core_index, vaddr, 16, [&] { Memory::Write64(vaddr, value[0]); Memory::Write64(vaddr, value[1]); }); } } // namespace Core