1176 lines
48 KiB
C++
1176 lines
48 KiB
C++
// SPDX-License-Identifier: MPL-2.0
|
|
// Copyright © 2020 Skyline Team and Contributors (https://github.com/skyline-emu/)
|
|
|
|
#include <os.h>
|
|
#include <kernel/types/KProcess.h>
|
|
#include <common/trace.h>
|
|
#include <vfs/npdm.h>
|
|
#include "results.h"
|
|
#include "svc.h"
|
|
|
|
namespace skyline::kernel::svc {
|
|
void SetHeapSize(const DeviceState &state) {
|
|
u32 size{state.ctx->gpr.w1};
|
|
|
|
if (!util::IsAligned(size, 0x200000)) {
|
|
state.ctx->gpr.w0 = result::InvalidSize;
|
|
state.ctx->gpr.x1 = 0;
|
|
|
|
state.logger->Warn("'size' not divisible by 2MB: {}", size);
|
|
return;
|
|
}
|
|
|
|
auto &heap{state.process->heap};
|
|
heap->Resize(size);
|
|
|
|
state.ctx->gpr.w0 = Result{};
|
|
state.ctx->gpr.x1 = reinterpret_cast<u64>(heap->ptr);
|
|
|
|
state.logger->Debug("Allocated at 0x{:X} - 0x{:X} (0x{:X} bytes)", heap->ptr, heap->ptr + heap->size, heap->size);
|
|
}
|
|
|
|
void SetMemoryAttribute(const DeviceState &state) {
|
|
auto pointer{reinterpret_cast<u8 *>(state.ctx->gpr.x0)};
|
|
if (!util::PageAligned(pointer)) {
|
|
state.ctx->gpr.w0 = result::InvalidAddress;
|
|
state.logger->Warn("'pointer' not page aligned: 0x{:X}", pointer);
|
|
return;
|
|
}
|
|
|
|
size_t size{state.ctx->gpr.x1};
|
|
if (!util::PageAligned(size)) {
|
|
state.ctx->gpr.w0 = result::InvalidSize;
|
|
state.logger->Warn("'size' {}: 0x{:X}", size ? "not page aligned" : "is zero", size);
|
|
return;
|
|
}
|
|
|
|
memory::MemoryAttribute mask{.value = state.ctx->gpr.w2};
|
|
memory::MemoryAttribute value{.value = state.ctx->gpr.w3};
|
|
|
|
auto maskedValue{mask.value | value.value};
|
|
if (maskedValue != mask.value || !mask.isUncached || mask.isDeviceShared || mask.isBorrowed || mask.isIpcLocked) {
|
|
state.ctx->gpr.w0 = result::InvalidCombination;
|
|
state.logger->Warn("'mask' invalid: 0x{:X}, 0x{:X}", mask.value, value.value);
|
|
return;
|
|
}
|
|
|
|
auto chunk{state.process->memory.Get(pointer)};
|
|
if (!chunk) {
|
|
state.ctx->gpr.w0 = result::InvalidAddress;
|
|
state.logger->Warn("Cannot find memory region: 0x{:X}", pointer);
|
|
return;
|
|
}
|
|
|
|
if (!chunk->state.attributeChangeAllowed) {
|
|
state.ctx->gpr.w0 = result::InvalidState;
|
|
state.logger->Warn("Attribute change not allowed for chunk: 0x{:X}", pointer);
|
|
return;
|
|
}
|
|
|
|
auto newChunk{*chunk};
|
|
newChunk.ptr = pointer;
|
|
newChunk.size = size;
|
|
newChunk.attributes.isUncached = value.isUncached;
|
|
state.process->memory.InsertChunk(newChunk);
|
|
|
|
state.logger->Debug("Set CPU caching to {} at 0x{:X} - 0x{:X} (0x{:X} bytes)", !static_cast<bool>(value.isUncached), pointer, pointer + size, size);
|
|
state.ctx->gpr.w0 = Result{};
|
|
}
|
|
|
|
void MapMemory(const DeviceState &state) {
|
|
auto destination{reinterpret_cast<u8 *>(state.ctx->gpr.x0)};
|
|
auto source{reinterpret_cast<u8 *>(state.ctx->gpr.x1)};
|
|
size_t size{state.ctx->gpr.x2};
|
|
|
|
if (!util::PageAligned(destination) || !util::PageAligned(source)) {
|
|
state.ctx->gpr.w0 = result::InvalidAddress;
|
|
state.logger->Warn("Addresses not page aligned: Source: 0x{:X}, Destination: 0x{:X} (Size: 0x{:X} bytes)", source, destination, size);
|
|
return;
|
|
}
|
|
|
|
if (!util::PageAligned(size)) {
|
|
state.ctx->gpr.w0 = result::InvalidSize;
|
|
state.logger->Warn("'size' {}: 0x{:X}", size ? "not page aligned" : "is zero", size);
|
|
return;
|
|
}
|
|
|
|
auto stack{state.process->memory.stack};
|
|
if (!stack.IsInside(destination)) {
|
|
state.ctx->gpr.w0 = result::InvalidMemoryRegion;
|
|
state.logger->Warn("Destination not within stack region: Source: 0x{:X}, Destination: 0x{:X} (Size: 0x{:X} bytes)", source, destination, size);
|
|
return;
|
|
}
|
|
|
|
auto chunk{state.process->memory.Get(source)};
|
|
if (!chunk) {
|
|
state.ctx->gpr.w0 = result::InvalidAddress;
|
|
state.logger->Warn("Source has no descriptor: Source: 0x{:X}, Destination: 0x{:X} (Size: 0x{:X} bytes)", source, destination, size);
|
|
return;
|
|
}
|
|
if (!chunk->state.mapAllowed) {
|
|
state.ctx->gpr.w0 = result::InvalidState;
|
|
state.logger->Warn("Source doesn't allow usage of svcMapMemory: Source: 0x{:X}, Destination: 0x{:X}, Size: 0x{:X}, MemoryState: 0x{:X}", source, destination, size, chunk->state.value);
|
|
return;
|
|
}
|
|
|
|
state.process->NewHandle<type::KPrivateMemory>(destination, size, chunk->permission, memory::states::Stack);
|
|
std::memcpy(destination, source, size);
|
|
|
|
auto object{state.process->GetMemoryObject(source)};
|
|
if (!object)
|
|
throw exception("svcMapMemory: Cannot find memory object in handle table for address 0x{:X}", source);
|
|
object->item->UpdatePermission(source, size, {false, false, false});
|
|
|
|
state.logger->Debug("Mapped range 0x{:X} - 0x{:X} to 0x{:X} - 0x{:X} (Size: 0x{:X} bytes)", source, source + size, destination, destination + size, size);
|
|
state.ctx->gpr.w0 = Result{};
|
|
}
|
|
|
|
void UnmapMemory(const DeviceState &state) {
|
|
auto source{reinterpret_cast<u8 *>(state.ctx->gpr.x0)};
|
|
auto destination{reinterpret_cast<u8 *>(state.ctx->gpr.x1)};
|
|
size_t size{state.ctx->gpr.x2};
|
|
|
|
if (!util::PageAligned(destination) || !util::PageAligned(source)) {
|
|
state.ctx->gpr.w0 = result::InvalidAddress;
|
|
state.logger->Warn("Addresses not page aligned: Source: 0x{:X}, Destination: 0x{:X} (Size: 0x{:X} bytes)", source, destination, size);
|
|
return;
|
|
}
|
|
|
|
if (!util::PageAligned(size)) {
|
|
state.ctx->gpr.w0 = result::InvalidSize;
|
|
state.logger->Warn("'size' {}: 0x{:X}", size ? "not page aligned" : "is zero", size);
|
|
return;
|
|
}
|
|
|
|
auto stack{state.process->memory.stack};
|
|
if (!stack.IsInside(source)) {
|
|
state.ctx->gpr.w0 = result::InvalidMemoryRegion;
|
|
state.logger->Warn("Source not within stack region: Source: 0x{:X}, Destination: 0x{:X} (Size: 0x{:X} bytes)", source, destination, size);
|
|
return;
|
|
}
|
|
|
|
auto sourceChunk{state.process->memory.Get(source)};
|
|
auto destChunk{state.process->memory.Get(destination)};
|
|
if (!sourceChunk || !destChunk) {
|
|
state.ctx->gpr.w0 = result::InvalidAddress;
|
|
state.logger->Warn("Addresses have no descriptor: Source: 0x{:X}, Destination: 0x{:X} (Size: 0x{:X} bytes)", source, destination, size);
|
|
return;
|
|
}
|
|
|
|
if (!destChunk->state.mapAllowed) {
|
|
state.ctx->gpr.w0 = result::InvalidState;
|
|
state.logger->Warn("Destination doesn't allow usage of svcMapMemory: Source: 0x{:X}, Destination: 0x{:X} (Size: 0x{:X} bytes) 0x{:X}", source, destination, size, destChunk->state.value);
|
|
return;
|
|
}
|
|
|
|
auto destObject{state.process->GetMemoryObject(destination)};
|
|
if (!destObject)
|
|
throw exception("svcUnmapMemory: Cannot find destination memory object in handle table for address 0x{:X}", destination);
|
|
|
|
destObject->item->UpdatePermission(destination, size, sourceChunk->permission);
|
|
|
|
std::memcpy(source, destination, size);
|
|
|
|
auto sourceObject{state.process->GetMemoryObject(source)};
|
|
if (!sourceObject)
|
|
throw exception("svcUnmapMemory: Cannot find source memory object in handle table for address 0x{:X}", source);
|
|
|
|
state.process->CloseHandle(sourceObject->handle);
|
|
|
|
state.logger->Debug("Unmapped range 0x{:X} - 0x{:X} to 0x{:X} - 0x{:X} (Size: 0x{:X} bytes)", source, source + size, destination, destination + size, size);
|
|
state.ctx->gpr.w0 = Result{};
|
|
}
|
|
|
|
void QueryMemory(const DeviceState &state) {
|
|
memory::MemoryInfo memInfo{};
|
|
|
|
auto pointer{reinterpret_cast<u8 *>(state.ctx->gpr.x2)};
|
|
auto chunk{state.process->memory.Get(pointer)};
|
|
|
|
if (chunk) {
|
|
memInfo = {
|
|
.address = reinterpret_cast<u64>(chunk->ptr),
|
|
.size = chunk->size,
|
|
.type = static_cast<u32>(chunk->state.type),
|
|
.attributes = chunk->attributes.value,
|
|
.permissions = static_cast<u32>(chunk->permission.Get()),
|
|
.deviceRefCount = 0,
|
|
.ipcRefCount = 0,
|
|
};
|
|
|
|
state.logger->Debug("Address: 0x{:X}, Region Start: 0x{:X}, Size: 0x{:X}, Type: 0x{:X}, Is Uncached: {}, Permissions: {}{}{}", pointer, memInfo.address, memInfo.size, memInfo.type, static_cast<bool>(chunk->attributes.isUncached), chunk->permission.r ? 'R' : '-', chunk->permission.w ? 'W' : '-', chunk->permission.x ? 'X' : '-');
|
|
} else {
|
|
auto addressSpaceEnd{reinterpret_cast<u64>(state.process->memory.addressSpace.address + state.process->memory.addressSpace.size)};
|
|
|
|
memInfo = {
|
|
.address = addressSpaceEnd,
|
|
.size = ~addressSpaceEnd + 1,
|
|
.type = static_cast<u32>(memory::MemoryType::Reserved),
|
|
};
|
|
|
|
state.logger->Debug("Trying to query memory outside of the application's address space: 0x{:X}", pointer);
|
|
}
|
|
|
|
*reinterpret_cast<memory::MemoryInfo *>(state.ctx->gpr.x0) = memInfo;
|
|
|
|
state.ctx->gpr.w0 = Result{};
|
|
}
|
|
|
|
void ExitProcess(const DeviceState &state) {
|
|
state.logger->Debug("Exiting process");
|
|
if (state.thread->id)
|
|
state.process->Kill(false);
|
|
std::longjmp(state.thread->originalCtx, true);
|
|
}
|
|
|
|
constexpr i32 IdealCoreDontCare{-1};
|
|
constexpr i32 IdealCoreUseProcessValue{-2};
|
|
constexpr i32 IdealCoreNoUpdate{-3};
|
|
|
|
void CreateThread(const DeviceState &state) {
|
|
auto entry{reinterpret_cast<void *>(state.ctx->gpr.x1)};
|
|
auto entryArgument{state.ctx->gpr.x2};
|
|
auto stackTop{reinterpret_cast<u8 *>(state.ctx->gpr.x3)};
|
|
auto priority{static_cast<i8>(static_cast<u32>(state.ctx->gpr.w4))};
|
|
auto idealCore{static_cast<i8>(static_cast<u32>(state.ctx->gpr.w5))};
|
|
|
|
idealCore = (idealCore == IdealCoreUseProcessValue) ? state.process->npdm.meta.idealCore : idealCore;
|
|
if (idealCore < 0 || idealCore >= constant::CoreCount) {
|
|
state.ctx->gpr.w0 = result::InvalidCoreId;
|
|
state.logger->Warn("'idealCore' invalid: {}", idealCore);
|
|
return;
|
|
}
|
|
|
|
if (!state.process->npdm.threadInfo.priority.Valid(priority)) {
|
|
state.ctx->gpr.w0 = result::InvalidPriority;
|
|
state.logger->Warn("'priority' invalid: {}", priority);
|
|
return;
|
|
}
|
|
|
|
auto stack{state.process->GetMemoryObject(stackTop)};
|
|
if (!stack)
|
|
throw exception("svcCreateThread: Cannot find memory object in handle table for thread stack: 0x{:X}", stackTop);
|
|
|
|
auto thread{state.process->CreateThread(entry, entryArgument, stackTop, priority, idealCore)};
|
|
if (thread) {
|
|
state.logger->Debug("Created thread #{} with handle 0x{:X} (Entry Point: 0x{:X}, Argument: 0x{:X}, Stack Pointer: 0x{:X}, Priority: {}, Ideal Core: {})", thread->id, thread->handle, entry, entryArgument, stackTop, priority, idealCore);
|
|
|
|
state.ctx->gpr.w1 = thread->handle;
|
|
state.ctx->gpr.w0 = Result{};
|
|
} else {
|
|
state.logger->Debug("Cannot create thread (Entry Point: 0x{:X}, Argument: 0x{:X}, Stack Pointer: 0x{:X}, Priority: {}, Ideal Core: {})", entry, entryArgument, stackTop, priority, idealCore);
|
|
state.ctx->gpr.w1 = 0;
|
|
state.ctx->gpr.w0 = result::OutOfResource;
|
|
}
|
|
}
|
|
|
|
void StartThread(const DeviceState &state) {
|
|
KHandle handle{state.ctx->gpr.w0};
|
|
try {
|
|
auto thread{state.process->GetHandle<type::KThread>(handle)};
|
|
state.logger->Debug("Starting thread #{}: 0x{:X}", thread->id, handle);
|
|
thread->Start();
|
|
state.ctx->gpr.w0 = Result{};
|
|
} catch (const std::out_of_range &) {
|
|
state.logger->Warn("'handle' invalid: 0x{:X}", handle);
|
|
state.ctx->gpr.w0 = result::InvalidHandle;
|
|
}
|
|
}
|
|
|
|
void ExitThread(const DeviceState &state) {
|
|
state.logger->Debug("Exiting current thread");
|
|
std::longjmp(state.thread->originalCtx, true);
|
|
}
|
|
|
|
void SleepThread(const DeviceState &state) {
|
|
constexpr i64 yieldWithoutCoreMigration{0};
|
|
constexpr i64 yieldWithCoreMigration{-1};
|
|
constexpr i64 yieldToAnyThread{-2};
|
|
|
|
i64 in{static_cast<i64>(state.ctx->gpr.x0)};
|
|
if (in > 0) {
|
|
state.logger->Debug("Sleeping for {}ns", in);
|
|
TRACE_EVENT("kernel", "SleepThread", "duration", in);
|
|
|
|
struct timespec spec{
|
|
.tv_sec = static_cast<time_t>(in / 1000000000),
|
|
.tv_nsec = static_cast<long>(in % 1000000000),
|
|
};
|
|
|
|
SchedulerScopedLock schedulerLock(state);
|
|
nanosleep(&spec, nullptr);
|
|
} else {
|
|
switch (in) {
|
|
case yieldWithCoreMigration: {
|
|
state.logger->Debug("Waking any appropriate parked threads and yielding");
|
|
TRACE_EVENT("kernel", "YieldWithCoreMigration");
|
|
state.scheduler->WakeParkedThread();
|
|
state.scheduler->Rotate();
|
|
state.scheduler->WaitSchedule();
|
|
break;
|
|
}
|
|
|
|
case yieldWithoutCoreMigration: {
|
|
state.logger->Debug("Cooperative yield");
|
|
TRACE_EVENT("kernel", "YieldWithoutCoreMigration");
|
|
state.scheduler->Rotate();
|
|
state.scheduler->WaitSchedule();
|
|
break;
|
|
}
|
|
|
|
case yieldToAnyThread: {
|
|
state.logger->Debug("Parking current thread");
|
|
TRACE_EVENT("kernel", "YieldToAnyThread");
|
|
state.scheduler->ParkThread();
|
|
state.scheduler->WaitSchedule(false);
|
|
break;
|
|
}
|
|
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
void GetThreadPriority(const DeviceState &state) {
|
|
KHandle handle{state.ctx->gpr.w1};
|
|
try {
|
|
auto thread{state.process->GetHandle<type::KThread>(handle)};
|
|
u8 priority{thread->priority};
|
|
state.logger->Debug("Retrieving thread #{}'s priority: {}", thread->id, priority);
|
|
|
|
state.ctx->gpr.w1 = priority;
|
|
state.ctx->gpr.w0 = Result{};
|
|
} catch (const std::out_of_range &) {
|
|
state.logger->Warn("'handle' invalid: 0x{:X}", handle);
|
|
state.ctx->gpr.w0 = result::InvalidHandle;
|
|
}
|
|
}
|
|
|
|
void SetThreadPriority(const DeviceState &state) {
|
|
KHandle handle{state.ctx->gpr.w0};
|
|
u8 priority{static_cast<u8>(state.ctx->gpr.w1)};
|
|
if (!state.process->npdm.threadInfo.priority.Valid(priority)) {
|
|
state.logger->Warn("'priority' invalid: 0x{:X}", priority);
|
|
state.ctx->gpr.w0 = result::InvalidPriority;
|
|
return;
|
|
}
|
|
try {
|
|
auto thread{state.process->GetHandle<type::KThread>(handle)};
|
|
state.logger->Debug("Setting thread #{}'s priority to {}", thread->id, priority);
|
|
if (thread->priority != priority) {
|
|
thread->basePriority = priority;
|
|
u8 newPriority{};
|
|
do {
|
|
// Try to CAS the priority of the thread with its new base priority
|
|
// If the new priority is equivalent to the current priority then we don't need to CAS
|
|
newPriority = thread->priority.load();
|
|
newPriority = std::min(newPriority, priority);
|
|
} while (newPriority != priority && thread->priority.compare_exchange_strong(newPriority, priority));
|
|
state.scheduler->UpdatePriority(thread);
|
|
thread->UpdatePriorityInheritance();
|
|
}
|
|
state.ctx->gpr.w0 = Result{};
|
|
} catch (const std::out_of_range &) {
|
|
state.logger->Warn("'handle' invalid: 0x{:X}", handle);
|
|
state.ctx->gpr.w0 = result::InvalidHandle;
|
|
}
|
|
}
|
|
|
|
void GetThreadCoreMask(const DeviceState &state) {
|
|
KHandle handle{state.ctx->gpr.w2};
|
|
try {
|
|
auto thread{state.process->GetHandle<type::KThread>(handle)};
|
|
auto idealCore{thread->idealCore};
|
|
auto affinityMask{thread->affinityMask};
|
|
state.logger->Debug("Getting thread #{}'s Ideal Core ({}) + Affinity Mask ({})", thread->id, idealCore, affinityMask);
|
|
|
|
state.ctx->gpr.x2 = affinityMask.to_ullong();
|
|
state.ctx->gpr.w1 = idealCore;
|
|
state.ctx->gpr.w0 = Result{};
|
|
} catch (const std::out_of_range &) {
|
|
state.logger->Warn("'handle' invalid: 0x{:X}", handle);
|
|
state.ctx->gpr.w0 = result::InvalidHandle;
|
|
}
|
|
}
|
|
|
|
void SetThreadCoreMask(const DeviceState &state) {
|
|
KHandle handle{state.ctx->gpr.w0};
|
|
i32 idealCore{static_cast<i32>(state.ctx->gpr.w1)};
|
|
CoreMask affinityMask{state.ctx->gpr.x2};
|
|
try {
|
|
auto thread{state.process->GetHandle<type::KThread>(handle)};
|
|
|
|
if (idealCore == IdealCoreUseProcessValue) {
|
|
idealCore = state.process->npdm.meta.idealCore;
|
|
affinityMask.reset().set(idealCore);
|
|
} else if (idealCore == IdealCoreNoUpdate) {
|
|
idealCore = thread->idealCore;
|
|
} else if (idealCore == IdealCoreDontCare) {
|
|
idealCore = std::countr_zero(affinityMask.to_ullong()); // The first enabled core in the affinity mask
|
|
}
|
|
|
|
auto processMask{state.process->npdm.threadInfo.coreMask};
|
|
if ((processMask | affinityMask) != processMask) {
|
|
state.logger->Warn("'affinityMask' invalid: {} (Process Mask: {})", affinityMask, processMask);
|
|
state.ctx->gpr.w0 = result::InvalidCoreId;
|
|
return;
|
|
}
|
|
|
|
if (affinityMask.none() || !affinityMask.test(idealCore)) {
|
|
state.logger->Warn("'affinityMask' invalid: {} (Ideal Core: {})", affinityMask, idealCore);
|
|
state.ctx->gpr.w0 = result::InvalidCombination;
|
|
return;
|
|
}
|
|
|
|
state.logger->Debug("Setting thread #{}'s Ideal Core ({}) + Affinity Mask ({})", thread->id, idealCore, affinityMask);
|
|
|
|
std::lock_guard guard(thread->coreMigrationMutex);
|
|
thread->idealCore = idealCore;
|
|
thread->affinityMask = affinityMask;
|
|
|
|
if (!affinityMask.test(thread->coreId) && thread->coreId != constant::ParkedCoreId) {
|
|
state.logger->Debug("Migrating thread #{} to Ideal Core C{} -> C{}", thread->id, thread->coreId, idealCore);
|
|
|
|
if (thread == state.thread) {
|
|
state.scheduler->RemoveThread();
|
|
thread->coreId = idealCore;
|
|
state.scheduler->InsertThread(state.thread);
|
|
state.scheduler->WaitSchedule();
|
|
} else if (!thread->running) {
|
|
thread->coreId = idealCore;
|
|
} else {
|
|
state.scheduler->UpdateCore(thread);
|
|
}
|
|
}
|
|
|
|
state.ctx->gpr.w0 = Result{};
|
|
} catch (const std::out_of_range &) {
|
|
state.logger->Warn("'handle' invalid: 0x{:X}", handle);
|
|
state.ctx->gpr.w0 = result::InvalidHandle;
|
|
}
|
|
}
|
|
|
|
void GetCurrentProcessorNumber(const DeviceState &state) {
|
|
std::lock_guard guard(state.thread->coreMigrationMutex);
|
|
auto coreId{state.thread->coreId};
|
|
state.logger->Debug("C{}", coreId);
|
|
state.ctx->gpr.w0 = coreId;
|
|
}
|
|
|
|
void ClearEvent(const DeviceState &state) {
|
|
KHandle handle{state.ctx->gpr.w0};
|
|
TRACE_EVENT_FMT("kernel", "ClearEvent 0x{:X}", handle);
|
|
try {
|
|
std::static_pointer_cast<type::KEvent>(state.process->GetHandle(handle))->ResetSignal();
|
|
state.logger->Debug("Clearing 0x{:X}", handle);
|
|
state.ctx->gpr.w0 = Result{};
|
|
} catch (const std::out_of_range &) {
|
|
state.logger->Warn("'handle' invalid: 0x{:X}", handle);
|
|
state.ctx->gpr.w0 = result::InvalidHandle;
|
|
return;
|
|
}
|
|
}
|
|
|
|
void MapSharedMemory(const DeviceState &state) {
|
|
try {
|
|
auto object{state.process->GetHandle<type::KSharedMemory>(state.ctx->gpr.w0)};
|
|
auto pointer{reinterpret_cast<u8 *>(state.ctx->gpr.x1)};
|
|
|
|
if (!util::PageAligned(pointer)) {
|
|
state.ctx->gpr.w0 = result::InvalidAddress;
|
|
state.logger->Warn("'pointer' not page aligned: 0x{:X}", pointer);
|
|
return;
|
|
}
|
|
|
|
size_t size{state.ctx->gpr.x2};
|
|
if (!util::PageAligned(size)) {
|
|
state.ctx->gpr.w0 = result::InvalidSize;
|
|
state.logger->Warn("'size' {}: 0x{:X}", size ? "not page aligned" : "is zero", size);
|
|
return;
|
|
}
|
|
|
|
memory::Permission permission(state.ctx->gpr.w3);
|
|
if ((permission.w && !permission.r) || (permission.x && !permission.r)) {
|
|
state.logger->Warn("'permission' invalid: {}{}{}", permission.r ? 'R' : '-', permission.w ? 'W' : '-', permission.x ? 'X' : '-');
|
|
state.ctx->gpr.w0 = result::InvalidNewMemoryPermission;
|
|
return;
|
|
}
|
|
|
|
state.logger->Debug("Mapping shared memory at 0x{:X} - 0x{:X} (0x{:X} bytes) ({}{}{})", pointer, pointer + size, size, permission.r ? 'R' : '-', permission.w ? 'W' : '-', permission.x ? 'X' : '-');
|
|
|
|
object->Map(pointer, size, permission);
|
|
|
|
state.ctx->gpr.w0 = Result{};
|
|
} catch (const std::out_of_range &) {
|
|
state.logger->Warn("'handle' invalid: 0x{:X}", static_cast<u32>(state.ctx->gpr.w0));
|
|
state.ctx->gpr.w0 = result::InvalidHandle;
|
|
}
|
|
}
|
|
|
|
void CreateTransferMemory(const DeviceState &state) {
|
|
auto pointer{reinterpret_cast<u8 *>(state.ctx->gpr.x1)};
|
|
if (!util::PageAligned(pointer)) {
|
|
state.ctx->gpr.w0 = result::InvalidAddress;
|
|
state.logger->Warn("'pointer' not page aligned: 0x{:X}", pointer);
|
|
return;
|
|
}
|
|
|
|
size_t size{state.ctx->gpr.x2};
|
|
if (!util::PageAligned(size)) {
|
|
state.ctx->gpr.w0 = result::InvalidSize;
|
|
state.logger->Warn("'size' {}: 0x{:X}", size ? "not page aligned" : "is zero", size);
|
|
return;
|
|
}
|
|
|
|
memory::Permission permission(state.ctx->gpr.w3);
|
|
if ((permission.w && !permission.r) || (permission.x && !permission.r)) {
|
|
state.logger->Warn("'permission' invalid: {}{}{}", permission.r ? 'R' : '-', permission.w ? 'W' : '-', permission.x ? 'X' : '-');
|
|
state.ctx->gpr.w0 = result::InvalidNewMemoryPermission;
|
|
return;
|
|
}
|
|
|
|
auto tmem{state.process->NewHandle<type::KTransferMemory>(pointer, size, permission)};
|
|
state.logger->Debug("Creating transfer memory at 0x{:X} - 0x{:X} (0x{:X} bytes) ({}{}{})", pointer, pointer + size, size, permission.r ? 'R' : '-', permission.w ? 'W' : '-', permission.x ? 'X' : '-');
|
|
|
|
state.ctx->gpr.w0 = Result{};
|
|
state.ctx->gpr.w1 = tmem.handle;
|
|
}
|
|
|
|
void CloseHandle(const DeviceState &state) {
|
|
KHandle handle{static_cast<KHandle>(state.ctx->gpr.w0)};
|
|
try {
|
|
state.process->CloseHandle(handle);
|
|
state.logger->Debug("Closing 0x{:X}", handle);
|
|
state.ctx->gpr.w0 = Result{};
|
|
} catch (const std::out_of_range &) {
|
|
state.logger->Warn("'handle' invalid: 0x{:X}", handle);
|
|
state.ctx->gpr.w0 = result::InvalidHandle;
|
|
}
|
|
}
|
|
|
|
void ResetSignal(const DeviceState &state) {
|
|
KHandle handle{state.ctx->gpr.w0};
|
|
TRACE_EVENT_FMT("kernel", "ResetSignal 0x{:X}", handle);
|
|
try {
|
|
auto object{state.process->GetHandle(handle)};
|
|
switch (object->objectType) {
|
|
case type::KType::KEvent:
|
|
case type::KType::KProcess:
|
|
state.ctx->gpr.w0 = std::static_pointer_cast<type::KSyncObject>(object)->ResetSignal() ? Result{} : result::InvalidState;
|
|
break;
|
|
|
|
default: {
|
|
state.logger->Warn("'handle' type invalid: 0x{:X} ({})", handle, object->objectType);
|
|
state.ctx->gpr.w0 = result::InvalidHandle;
|
|
return;
|
|
}
|
|
}
|
|
|
|
state.logger->Debug("Resetting 0x{:X}", handle);
|
|
state.ctx->gpr.w0 = Result{};
|
|
} catch (const std::out_of_range &) {
|
|
state.logger->Warn("'handle' invalid: 0x{:X}", handle);
|
|
state.ctx->gpr.w0 = result::InvalidHandle;
|
|
return;
|
|
}
|
|
}
|
|
|
|
void WaitSynchronization(const DeviceState &state) {
|
|
constexpr u8 maxSyncHandles{0x40}; // The total amount of handles that can be passed to WaitSynchronization
|
|
|
|
u32 numHandles{state.ctx->gpr.w2};
|
|
if (numHandles > maxSyncHandles) {
|
|
state.ctx->gpr.w0 = result::OutOfRange;
|
|
return;
|
|
}
|
|
|
|
span waitHandles(reinterpret_cast<KHandle *>(state.ctx->gpr.x1), numHandles);
|
|
std::vector<std::shared_ptr<type::KSyncObject>> objectTable;
|
|
objectTable.reserve(numHandles);
|
|
|
|
for (const auto &handle : waitHandles) {
|
|
auto object{state.process->GetHandle(handle)};
|
|
switch (object->objectType) {
|
|
case type::KType::KProcess:
|
|
case type::KType::KThread:
|
|
case type::KType::KEvent:
|
|
case type::KType::KSession:
|
|
objectTable.push_back(std::static_pointer_cast<type::KSyncObject>(object));
|
|
break;
|
|
|
|
default: {
|
|
state.logger->Debug("An invalid handle was supplied: 0x{:X}", handle);
|
|
state.ctx->gpr.w0 = result::InvalidHandle;
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
|
|
i64 timeout{static_cast<i64>(state.ctx->gpr.x3)};
|
|
if (waitHandles.size() == 1) {
|
|
state.logger->Debug("Waiting on 0x{:X} for {}ns", waitHandles[0], timeout);
|
|
} else if (Logger::LogLevel::Debug <= state.logger->configLevel) {
|
|
std::string handleString;
|
|
for (const auto &handle : waitHandles)
|
|
handleString += fmt::format("* 0x{:X}\n", handle);
|
|
state.logger->Debug("Waiting on handles:\n{}Timeout: {}ns", handleString, timeout);
|
|
}
|
|
|
|
TRACE_EVENT_FMT("kernel", waitHandles.size() == 1 ? "WaitSynchronization 0x{:X}" : "WaitSynchronizationMultiple 0x{:X}", waitHandles[0]);
|
|
|
|
std::unique_lock lock(type::KSyncObject::syncObjectMutex);
|
|
if (state.thread->cancelSync) {
|
|
state.thread->cancelSync = false;
|
|
state.ctx->gpr.w0 = result::Cancelled;
|
|
return;
|
|
}
|
|
|
|
u32 index{};
|
|
for (const auto &object : objectTable) {
|
|
if (object->signalled) {
|
|
state.logger->Debug("Signalled 0x{:X}", waitHandles[index]);
|
|
state.ctx->gpr.w0 = Result{};
|
|
state.ctx->gpr.w1 = index;
|
|
return;
|
|
}
|
|
index++;
|
|
}
|
|
|
|
if (timeout == 0) {
|
|
state.logger->Debug("No handle is currently signalled");
|
|
state.ctx->gpr.w0 = result::TimedOut;
|
|
return;
|
|
}
|
|
|
|
auto priority{state.thread->priority.load()};
|
|
for (const auto &object : objectTable)
|
|
object->syncObjectWaiters.insert(std::upper_bound(object->syncObjectWaiters.begin(), object->syncObjectWaiters.end(), priority, type::KThread::IsHigherPriority), state.thread);
|
|
|
|
state.thread->isCancellable = true;
|
|
state.thread->wakeObject = nullptr;
|
|
state.scheduler->RemoveThread();
|
|
|
|
lock.unlock();
|
|
if (timeout > 0)
|
|
state.scheduler->TimedWaitSchedule(std::chrono::nanoseconds(timeout));
|
|
else
|
|
state.scheduler->WaitSchedule(false);
|
|
lock.lock();
|
|
|
|
state.thread->isCancellable = false;
|
|
auto wakeObject{state.thread->wakeObject};
|
|
|
|
u32 wakeIndex{};
|
|
index = 0;
|
|
for (const auto &object : objectTable) {
|
|
if (object.get() == wakeObject)
|
|
wakeIndex = index;
|
|
|
|
auto it{std::find(object->syncObjectWaiters.begin(), object->syncObjectWaiters.end(), state.thread)};
|
|
if (it != object->syncObjectWaiters.end())
|
|
object->syncObjectWaiters.erase(it);
|
|
else
|
|
throw exception("svcWaitSynchronization: An object (0x{:X}) has been removed from the syncObjectWaiters queue incorrectly", waitHandles[index]);
|
|
|
|
index++;
|
|
}
|
|
|
|
if (wakeObject) {
|
|
state.logger->Debug("Signalled 0x{:X}", waitHandles[wakeIndex]);
|
|
state.ctx->gpr.w0 = Result{};
|
|
state.ctx->gpr.w1 = wakeIndex;
|
|
} else if (state.thread->cancelSync) {
|
|
state.thread->cancelSync = false;
|
|
state.logger->Debug("Wait has been cancelled");
|
|
state.ctx->gpr.w0 = result::Cancelled;
|
|
} else {
|
|
state.logger->Debug("Wait has timed out");
|
|
state.ctx->gpr.w0 = result::TimedOut;
|
|
lock.unlock();
|
|
state.scheduler->InsertThread(state.thread);
|
|
state.scheduler->WaitSchedule();
|
|
}
|
|
}
|
|
|
|
void CancelSynchronization(const DeviceState &state) {
|
|
try {
|
|
std::unique_lock lock(type::KSyncObject::syncObjectMutex);
|
|
auto thread{state.process->GetHandle<type::KThread>(state.ctx->gpr.w0)};
|
|
thread->cancelSync = true;
|
|
if (thread->isCancellable) {
|
|
thread->isCancellable = false;
|
|
state.scheduler->InsertThread(thread);
|
|
}
|
|
state.ctx->gpr.w0 = Result{};
|
|
} catch (const std::out_of_range &) {
|
|
state.logger->Warn("'handle' invalid: 0x{:X}", static_cast<u32>(state.ctx->gpr.w0));
|
|
state.ctx->gpr.w0 = result::InvalidHandle;
|
|
}
|
|
}
|
|
|
|
void ArbitrateLock(const DeviceState &state) {
|
|
auto mutex{reinterpret_cast<u32 *>(state.ctx->gpr.x1)};
|
|
if (!util::WordAligned(mutex)) {
|
|
state.logger->Warn("'mutex' not word aligned: 0x{:X}", mutex);
|
|
state.ctx->gpr.w0 = result::InvalidAddress;
|
|
return;
|
|
}
|
|
|
|
state.logger->Debug("Locking 0x{:X}", mutex);
|
|
TRACE_EVENT_FMT("kernel", "MutexLock 0x{:X}", mutex);
|
|
|
|
KHandle ownerHandle{state.ctx->gpr.w0};
|
|
KHandle requesterHandle{state.ctx->gpr.w2};
|
|
auto result{state.process->MutexLock(mutex, ownerHandle, requesterHandle)};
|
|
if (result == Result{})
|
|
state.logger->Debug("Locked 0x{:X}", mutex);
|
|
else if (result == result::InvalidCurrentMemory)
|
|
result = Result{}; // If the mutex value isn't expected then it's still successful
|
|
else if (result == result::InvalidHandle)
|
|
state.logger->Warn("'ownerHandle' invalid: 0x{:X} (0x{:X})", ownerHandle, mutex);
|
|
|
|
state.ctx->gpr.w0 = result;
|
|
}
|
|
|
|
void ArbitrateUnlock(const DeviceState &state) {
|
|
auto mutex{reinterpret_cast<u32 *>(state.ctx->gpr.x0)};
|
|
if (!util::WordAligned(mutex)) {
|
|
state.logger->Warn("'mutex' not word aligned: 0x{:X}", mutex);
|
|
state.ctx->gpr.w0 = result::InvalidAddress;
|
|
return;
|
|
}
|
|
|
|
TRACE_EVENT_FMT("kernel", "MutexUnlock 0x{:X}", mutex);
|
|
|
|
state.logger->Debug("Unlocking 0x{:X}", mutex);
|
|
state.process->MutexUnlock(mutex);
|
|
state.logger->Debug("Unlocked 0x{:X}", mutex);
|
|
|
|
state.ctx->gpr.w0 = Result{};
|
|
}
|
|
|
|
void WaitProcessWideKeyAtomic(const DeviceState &state) {
|
|
auto mutex{reinterpret_cast<u32 *>(state.ctx->gpr.x0)};
|
|
if (!util::WordAligned(mutex)) {
|
|
state.logger->Warn("'mutex' not word aligned: 0x{:X}", mutex);
|
|
state.ctx->gpr.w0 = result::InvalidAddress;
|
|
return;
|
|
}
|
|
|
|
auto conditional{reinterpret_cast<u32 *>(state.ctx->gpr.x1)};
|
|
KHandle requesterHandle{state.ctx->gpr.w2};
|
|
|
|
i64 timeout{static_cast<i64>(state.ctx->gpr.x3)};
|
|
state.logger->Debug("Waiting on 0x{:X} with 0x{:X} for {}ns", conditional, mutex, timeout);
|
|
|
|
auto result{state.process->ConditionalVariableWait(conditional, mutex, requesterHandle, timeout)};
|
|
if (result == Result{})
|
|
state.logger->Debug("Waited for 0x{:X} and reacquired 0x{:X}", conditional, mutex);
|
|
else if (result == result::TimedOut)
|
|
state.logger->Debug("Wait on 0x{:X} has timed out after {}ns", conditional, timeout);
|
|
state.ctx->gpr.w0 = result;
|
|
}
|
|
|
|
void SignalProcessWideKey(const DeviceState &state) {
|
|
auto conditional{reinterpret_cast<u32 *>(state.ctx->gpr.x0)};
|
|
i32 count{static_cast<i32>(state.ctx->gpr.w1)};
|
|
|
|
state.logger->Debug("Signalling 0x{:X} for {} waiters", conditional, count);
|
|
state.process->ConditionalVariableSignal(conditional, count);
|
|
state.ctx->gpr.w0 = Result{};
|
|
}
|
|
|
|
void GetSystemTick(const DeviceState &state) {
|
|
u64 tick;
|
|
asm("STR X1, [SP, #-16]!\n\t"
|
|
"MRS %0, CNTVCT_EL0\n\t"
|
|
"MOV X1, #0xF800\n\t"
|
|
"MOVK X1, #0x124, lsl #16\n\t"
|
|
"MUL %0, %0, X1\n\t"
|
|
"MRS X1, CNTFRQ_EL0\n\t"
|
|
"UDIV %0, %0, X1\n\t"
|
|
"LDR X1, [SP], #16" : "=r"(tick));
|
|
state.ctx->gpr.x0 = tick;
|
|
}
|
|
|
|
void ConnectToNamedPort(const DeviceState &state) {
|
|
constexpr u8 portSize = 0x8; //!< The size of a port name string
|
|
std::string_view port(span(reinterpret_cast<char *>(state.ctx->gpr.x1), portSize).as_string(true));
|
|
|
|
KHandle handle{};
|
|
if (port.compare("sm:") >= 0) {
|
|
handle = state.process->NewHandle<type::KSession>(std::static_pointer_cast<service::BaseService>(state.os->serviceManager.smUserInterface)).handle;
|
|
} else {
|
|
state.logger->Warn("Connecting to invalid port: '{}'", port);
|
|
state.ctx->gpr.w0 = result::NotFound;
|
|
return;
|
|
}
|
|
|
|
state.logger->Debug("Connecting to port '{}' at 0x{:X}", port, handle);
|
|
|
|
state.ctx->gpr.w1 = handle;
|
|
state.ctx->gpr.w0 = Result{};
|
|
}
|
|
|
|
void SendSyncRequest(const DeviceState &state) {
|
|
SchedulerScopedLock schedulerLock(state);
|
|
state.os->serviceManager.SyncRequestHandler(static_cast<KHandle>(state.ctx->gpr.x0));
|
|
state.ctx->gpr.w0 = Result{};
|
|
}
|
|
|
|
void GetThreadId(const DeviceState &state) {
|
|
KHandle handle{state.ctx->gpr.w1};
|
|
size_t tid{state.process->GetHandle<type::KThread>(handle)->id};
|
|
|
|
state.logger->Debug("Handle: 0x{:X}, TID: {}", handle, tid);
|
|
|
|
state.ctx->gpr.x1 = tid;
|
|
state.ctx->gpr.w0 = Result{};
|
|
}
|
|
|
|
void Break(const DeviceState &state) {
|
|
auto reason{state.ctx->gpr.x0};
|
|
if (reason & (1ULL << 31)) {
|
|
state.logger->Debug("Debugger is being engaged ({})", reason);
|
|
} else {
|
|
state.logger->Error("Exit Stack Trace ({}){}", reason, state.loader->GetStackTrace());
|
|
if (state.thread->id)
|
|
state.process->Kill(false);
|
|
std::longjmp(state.thread->originalCtx, true);
|
|
}
|
|
}
|
|
|
|
void OutputDebugString(const DeviceState &state) {
|
|
auto string{span(reinterpret_cast<char *>(state.ctx->gpr.x0), state.ctx->gpr.x1).as_string()};
|
|
|
|
if (string.back() == '\n')
|
|
string.remove_suffix(1);
|
|
|
|
state.logger->Info("{}", string);
|
|
state.ctx->gpr.w0 = Result{};
|
|
}
|
|
|
|
void GetInfo(const DeviceState &state) {
|
|
enum class InfoState : u32 {
|
|
// 1.0.0+
|
|
AllowedCpuIdBitmask = 0,
|
|
AllowedThreadPriorityMask = 1,
|
|
AliasRegionBaseAddr = 2,
|
|
AliasRegionSize = 3,
|
|
HeapRegionBaseAddr = 4,
|
|
HeapRegionSize = 5,
|
|
TotalMemoryAvailable = 6,
|
|
TotalMemoryUsage = 7,
|
|
IsCurrentProcessBeingDebugged = 8,
|
|
ResourceLimit = 9,
|
|
IdleTickCount = 10,
|
|
RandomEntropy = 11,
|
|
// 2.0.0+
|
|
AddressSpaceBaseAddr = 12,
|
|
AddressSpaceSize = 13,
|
|
StackRegionBaseAddr = 14,
|
|
StackRegionSize = 15,
|
|
// 3.0.0+
|
|
TotalSystemResourceAvailable = 16,
|
|
TotalSystemResourceUsage = 17,
|
|
ProgramId = 18,
|
|
// 4.0.0+
|
|
PrivilegedProcessId = 19,
|
|
// 5.0.0+
|
|
UserExceptionContextAddr = 20,
|
|
// 6.0.0+
|
|
TotalMemoryAvailableWithoutSystemResource = 21,
|
|
TotalMemoryUsageWithoutSystemResource = 22,
|
|
};
|
|
|
|
InfoState info{static_cast<u32>(state.ctx->gpr.w1)};
|
|
KHandle handle{state.ctx->gpr.w2};
|
|
u64 id1{state.ctx->gpr.x3};
|
|
|
|
constexpr u64 totalPhysicalMemory{0xF8000000}; // ~4 GB of RAM
|
|
|
|
u64 out{};
|
|
switch (info) {
|
|
case InfoState::IsCurrentProcessBeingDebugged:
|
|
case InfoState::PrivilegedProcessId:
|
|
break;
|
|
|
|
case InfoState::AllowedCpuIdBitmask:
|
|
out = state.process->npdm.threadInfo.coreMask.to_ullong();
|
|
break;
|
|
|
|
case InfoState::AllowedThreadPriorityMask:
|
|
out = state.process->npdm.threadInfo.priority.Mask();
|
|
break;
|
|
|
|
case InfoState::AliasRegionBaseAddr:
|
|
out = state.process->memory.alias.address;
|
|
break;
|
|
|
|
case InfoState::AliasRegionSize:
|
|
out = state.process->memory.alias.size;
|
|
break;
|
|
|
|
case InfoState::HeapRegionBaseAddr:
|
|
out = state.process->memory.heap.address;
|
|
break;
|
|
|
|
case InfoState::HeapRegionSize:
|
|
out = state.process->memory.heap.size;
|
|
break;
|
|
|
|
case InfoState::TotalMemoryAvailable:
|
|
out = std::min(totalPhysicalMemory, state.process->memory.heap.size);
|
|
break;
|
|
|
|
case InfoState::TotalMemoryUsage:
|
|
out = state.process->memory.GetUserMemoryUsage() + state.process->memory.GetSystemResourceUsage();
|
|
break;
|
|
|
|
case InfoState::RandomEntropy:
|
|
out = util::GetTimeTicks();
|
|
break;
|
|
|
|
case InfoState::AddressSpaceBaseAddr:
|
|
out = state.process->memory.base.address;
|
|
break;
|
|
|
|
case InfoState::AddressSpaceSize:
|
|
out = state.process->memory.base.size;
|
|
break;
|
|
|
|
case InfoState::StackRegionBaseAddr:
|
|
out = state.process->memory.stack.address;
|
|
break;
|
|
|
|
case InfoState::StackRegionSize:
|
|
out = state.process->memory.stack.size;
|
|
break;
|
|
|
|
case InfoState::TotalSystemResourceAvailable:
|
|
out = state.process->npdm.meta.systemResourceSize;
|
|
break;
|
|
|
|
case InfoState::TotalSystemResourceUsage:
|
|
// A very rough approximation of what this should be on the Switch, the amount of memory allocated for storing the memory blocks (https://switchbrew.org/wiki/Kernel_objects#KMemoryBlockManager)
|
|
out = state.process->memory.GetSystemResourceUsage();
|
|
break;
|
|
|
|
case InfoState::ProgramId:
|
|
out = state.process->npdm.aci0.programId;
|
|
break;
|
|
|
|
case InfoState::TotalMemoryAvailableWithoutSystemResource:
|
|
out = std::min(totalPhysicalMemory, state.process->memory.heap.size) - state.process->npdm.meta.systemResourceSize;
|
|
break;
|
|
|
|
case InfoState::TotalMemoryUsageWithoutSystemResource:
|
|
out = state.process->memory.GetUserMemoryUsage();
|
|
break;
|
|
|
|
case InfoState::UserExceptionContextAddr:
|
|
out = reinterpret_cast<u64>(state.process->tlsExceptionContext);
|
|
break;
|
|
|
|
default:
|
|
state.logger->Warn("Unimplemented case ID0: {}, ID1: {}", static_cast<u32>(info), id1);
|
|
state.ctx->gpr.w0 = result::InvalidEnumValue;
|
|
return;
|
|
}
|
|
|
|
state.logger->Debug("ID0: {}, ID1: {}, Out: 0x{:X}", static_cast<u32>(info), id1, out);
|
|
|
|
state.ctx->gpr.x1 = out;
|
|
state.ctx->gpr.w0 = Result{};
|
|
}
|
|
|
|
void MapPhysicalMemory(const DeviceState &state) {
|
|
auto pointer{reinterpret_cast<u8 *>(state.ctx->gpr.x0)};
|
|
size_t size{state.ctx->gpr.x1};
|
|
|
|
if (!util::PageAligned(pointer)) {
|
|
state.ctx->gpr.w0 = result::InvalidAddress;
|
|
return;
|
|
}
|
|
|
|
if (!size || !util::PageAligned(size)) {
|
|
state.ctx->gpr.w0 = result::InvalidSize;
|
|
return;
|
|
}
|
|
|
|
if (!state.process->memory.alias.IsInside(pointer) || !state.process->memory.alias.IsInside(pointer + size)) {
|
|
state.ctx->gpr.w0 = result::InvalidMemoryRegion;
|
|
return;
|
|
}
|
|
|
|
state.process->NewHandle<type::KPrivateMemory>(pointer, size, memory::Permission{true, true, false}, memory::states::Heap);
|
|
|
|
state.ctx->gpr.w0 = Result{};
|
|
}
|
|
|
|
void UnmapPhysicalMemory(const DeviceState &state) {
|
|
auto pointer{reinterpret_cast<u8 *>(state.ctx->gpr.x0)};
|
|
size_t size{state.ctx->gpr.x1};
|
|
|
|
if (!util::PageAligned(pointer)) {
|
|
state.ctx->gpr.w0 = result::InvalidAddress;
|
|
return;
|
|
}
|
|
|
|
if (!size || !util::PageAligned(size)) {
|
|
state.ctx->gpr.w0 = result::InvalidSize;
|
|
return;
|
|
}
|
|
|
|
if (!state.process->memory.alias.IsInside(pointer) || !state.process->memory.alias.IsInside(pointer + size)) {
|
|
state.ctx->gpr.w0 = result::InvalidMemoryRegion;
|
|
return;
|
|
}
|
|
|
|
auto end{pointer + size};
|
|
while (pointer < end) {
|
|
auto memory{state.process->GetMemoryObject(pointer)};
|
|
if (memory) {
|
|
auto item{static_pointer_cast<type::KPrivateMemory>(memory->item)};
|
|
auto initialSize{item->size};
|
|
if (item->memoryState == memory::states::Heap) {
|
|
if (item->ptr >= pointer) {
|
|
if (item->size <= size) {
|
|
item->Resize(0);
|
|
state.process->CloseHandle(memory->handle);
|
|
} else {
|
|
item->Remap(pointer + size, item->size - (size + (item->ptr - pointer)));
|
|
}
|
|
} else if (item->ptr < pointer) {
|
|
item->Resize(pointer - item->ptr);
|
|
}
|
|
}
|
|
pointer += initialSize;
|
|
size -= initialSize;
|
|
} else {
|
|
auto block{*state.process->memory.Get(pointer)};
|
|
pointer += block.size;
|
|
size -= block.size;
|
|
}
|
|
}
|
|
|
|
state.ctx->gpr.w0 = Result{};
|
|
}
|
|
|
|
void WaitForAddress(const DeviceState &state) {
|
|
auto address{reinterpret_cast<u32 *>(state.ctx->gpr.x0)};
|
|
if (!util::WordAligned(address)) [[unlikely]] {
|
|
state.logger->Warn("'address' not word aligned: 0x{:X}", address);
|
|
state.ctx->gpr.w0 = result::InvalidAddress;
|
|
return;
|
|
}
|
|
|
|
enum class ArbitrationType : u32 {
|
|
WaitIfLessThan = 0,
|
|
DecrementAndWaitIfLessThan = 1,
|
|
WaitIfEqual = 2,
|
|
} arbitrationType{static_cast<ArbitrationType>(static_cast<u32>(state.ctx->gpr.w1))};
|
|
u32 value{state.ctx->gpr.w2};
|
|
i64 timeout{static_cast<i64>(state.ctx->gpr.x3)};
|
|
|
|
Result result;
|
|
switch (arbitrationType) {
|
|
case ArbitrationType::WaitIfLessThan:
|
|
state.logger->Debug("Waiting on 0x{:X} if less than {} for {}ns", address, value, timeout);
|
|
result = state.process->WaitForAddress(address, value, timeout, [](u32 *address, u32 value) {
|
|
return *address < value;
|
|
});
|
|
break;
|
|
|
|
case ArbitrationType::DecrementAndWaitIfLessThan:
|
|
state.logger->Debug("Waiting on and decrementing 0x{:X} if less than {} for {}ns", address, value, timeout);
|
|
result = state.process->WaitForAddress(address, value, timeout, [](u32 *address, u32 value) {
|
|
u32 userValue{__atomic_load_n(address, __ATOMIC_SEQ_CST)};
|
|
do {
|
|
if (value <= userValue) [[unlikely]] // We want to explicitly decrement **after** the check
|
|
return false;
|
|
} while (!__atomic_compare_exchange_n(address, &userValue, userValue - 1, false, __ATOMIC_SEQ_CST, __ATOMIC_SEQ_CST));
|
|
return true;
|
|
});
|
|
break;
|
|
|
|
case ArbitrationType::WaitIfEqual:
|
|
state.logger->Debug("Waiting on 0x{:X} if equal to {} for {}ns", address, value, timeout);
|
|
result = state.process->WaitForAddress(address, value, timeout, [](u32 *address, u32 value) {
|
|
return *address == value;
|
|
});
|
|
break;
|
|
|
|
default:
|
|
[[unlikely]]
|
|
state.logger->Error("'arbitrationType' invalid: {}", arbitrationType);
|
|
state.ctx->gpr.w0 = result::InvalidEnumValue;
|
|
return;
|
|
}
|
|
|
|
if (result == Result{})
|
|
[[likely]]
|
|
state.logger->Debug("Waited on 0x{:X} successfully", address);
|
|
else if (result == result::TimedOut)
|
|
state.logger->Debug("Wait on 0x{:X} has timed out after {}ns", address, timeout);
|
|
else if (result == result::InvalidState)
|
|
state.logger->Debug("The value at 0x{:X} did not satisfy the arbitration condition", address);
|
|
|
|
state.ctx->gpr.w0 = result;
|
|
}
|
|
|
|
void SignalToAddress(const DeviceState &state) {
|
|
auto address{reinterpret_cast<u32 *>(state.ctx->gpr.x0)};
|
|
if (!util::WordAligned(address)) [[unlikely]] {
|
|
state.logger->Warn("'address' not word aligned: 0x{:X}", address);
|
|
state.ctx->gpr.w0 = result::InvalidAddress;
|
|
return;
|
|
}
|
|
|
|
enum class SignalType : u32 {
|
|
Signal = 0,
|
|
SignalAndIncrementIfEqual = 1,
|
|
SignalAndModifyBasedOnWaitingThreadCountIfEqual = 2,
|
|
} signalType{static_cast<SignalType>(static_cast<u32>(state.ctx->gpr.w1))};
|
|
u32 value{state.ctx->gpr.w2};
|
|
i32 count{static_cast<i32>(state.ctx->gpr.w3)};
|
|
|
|
Result result;
|
|
switch (signalType) {
|
|
case SignalType::Signal:
|
|
state.logger->Debug("Signalling 0x{:X} for {} waiters", address, count);
|
|
result = state.process->SignalToAddress(address, value, count);
|
|
break;
|
|
|
|
case SignalType::SignalAndIncrementIfEqual:
|
|
state.logger->Debug("Signalling 0x{:X} and incrementing if equal to {} for {} waiters", address, value, count);
|
|
result = state.process->SignalToAddress(address, value, count, [](u32 *address, u32 value, u32) {
|
|
return __atomic_compare_exchange_n(address, &value, value + 1, false, __ATOMIC_SEQ_CST, __ATOMIC_SEQ_CST);
|
|
});
|
|
break;
|
|
|
|
case SignalType::SignalAndModifyBasedOnWaitingThreadCountIfEqual:
|
|
state.logger->Debug("Signalling 0x{:X} and setting to waiting thread count if equal to {} for {} waiters", address, value, count);
|
|
result = state.process->SignalToAddress(address, value, count, [](u32 *address, u32 value, u32 waiterCount) {
|
|
return __atomic_compare_exchange_n(address, &value, waiterCount, false, __ATOMIC_SEQ_CST, __ATOMIC_SEQ_CST);
|
|
});
|
|
break;
|
|
|
|
default:
|
|
[[unlikely]]
|
|
state.logger->Error("'signalType' invalid: {}", signalType);
|
|
state.ctx->gpr.w0 = result::InvalidEnumValue;
|
|
return;
|
|
}
|
|
|
|
if (result == Result{})
|
|
[[likely]]
|
|
state.logger->Debug("Signalled 0x{:X} for {} successfully", address, count);
|
|
else if (result == result::InvalidState)
|
|
state.logger->Debug("The value at 0x{:X} did not satisfy the mutation condition", address);
|
|
|
|
state.ctx->gpr.w0 = result;
|
|
}
|
|
}
|