|
|
|
@ -5,44 +5,54 @@
|
|
|
|
|
#include "core/hle/kernel/k_light_lock.h"
|
|
|
|
|
#include "core/hle/kernel/k_scheduler.h"
|
|
|
|
|
#include "core/hle/kernel/k_thread.h"
|
|
|
|
|
#include "core/hle/kernel/k_thread_queue.h"
|
|
|
|
|
#include "core/hle/kernel/kernel.h"
|
|
|
|
|
|
|
|
|
|
namespace Kernel {
|
|
|
|
|
|
|
|
|
|
namespace {
|
|
|
|
|
|
|
|
|
|
class ThreadQueueImplForKLightLock final : public KThreadQueue {
|
|
|
|
|
public:
|
|
|
|
|
explicit ThreadQueueImplForKLightLock(KernelCore& kernel_) : KThreadQueue(kernel_) {}
|
|
|
|
|
|
|
|
|
|
virtual void CancelWait([[maybe_unused]] KThread* waiting_thread,
|
|
|
|
|
[[maybe_unused]] ResultCode wait_result,
|
|
|
|
|
[[maybe_unused]] bool cancel_timer_task) override {
|
|
|
|
|
// Do nothing, waiting to acquire a light lock cannot be canceled.
|
|
|
|
|
}
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
} // namespace
|
|
|
|
|
|
|
|
|
|
void KLightLock::Lock() {
|
|
|
|
|
const uintptr_t cur_thread = reinterpret_cast<uintptr_t>(GetCurrentThreadPointer(kernel));
|
|
|
|
|
const uintptr_t cur_thread_tag = (cur_thread | 1);
|
|
|
|
|
|
|
|
|
|
while (true) {
|
|
|
|
|
uintptr_t old_tag = tag.load(std::memory_order_relaxed);
|
|
|
|
|
|
|
|
|
|
while (!tag.compare_exchange_weak(old_tag, (old_tag == 0) ? cur_thread : old_tag | 1,
|
|
|
|
|
while (!tag.compare_exchange_weak(old_tag, (old_tag == 0) ? cur_thread : (old_tag | 1),
|
|
|
|
|
std::memory_order_acquire)) {
|
|
|
|
|
if ((old_tag | 1) == cur_thread_tag) {
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if ((old_tag == 0) || ((old_tag | 1) == cur_thread_tag)) {
|
|
|
|
|
if (old_tag == 0 || this->LockSlowPath(old_tag | 1, cur_thread)) {
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
LockSlowPath(old_tag | 1, cur_thread);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
void KLightLock::Unlock() {
|
|
|
|
|
const uintptr_t cur_thread = reinterpret_cast<uintptr_t>(GetCurrentThreadPointer(kernel));
|
|
|
|
|
|
|
|
|
|
uintptr_t expected = cur_thread;
|
|
|
|
|
do {
|
|
|
|
|
if (expected != cur_thread) {
|
|
|
|
|
return UnlockSlowPath(cur_thread);
|
|
|
|
|
if (!tag.compare_exchange_strong(expected, 0, std::memory_order_release)) {
|
|
|
|
|
this->UnlockSlowPath(cur_thread);
|
|
|
|
|
}
|
|
|
|
|
} while (!tag.compare_exchange_weak(expected, 0, std::memory_order_release));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
void KLightLock::LockSlowPath(uintptr_t _owner, uintptr_t _cur_thread) {
|
|
|
|
|
bool KLightLock::LockSlowPath(uintptr_t _owner, uintptr_t _cur_thread) {
|
|
|
|
|
KThread* cur_thread = reinterpret_cast<KThread*>(_cur_thread);
|
|
|
|
|
ThreadQueueImplForKLightLock wait_queue(kernel);
|
|
|
|
|
|
|
|
|
|
// Pend the current thread waiting on the owner thread.
|
|
|
|
|
{
|
|
|
|
@ -50,30 +60,23 @@ void KLightLock::LockSlowPath(uintptr_t _owner, uintptr_t _cur_thread) {
|
|
|
|
|
|
|
|
|
|
// Ensure we actually have locking to do.
|
|
|
|
|
if (tag.load(std::memory_order_relaxed) != _owner) {
|
|
|
|
|
return;
|
|
|
|
|
return false;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// Add the current thread as a waiter on the owner.
|
|
|
|
|
KThread* owner_thread = reinterpret_cast<KThread*>(_owner & ~1ULL);
|
|
|
|
|
KThread* owner_thread = reinterpret_cast<KThread*>(_owner & ~1ul);
|
|
|
|
|
cur_thread->SetAddressKey(reinterpret_cast<uintptr_t>(std::addressof(tag)));
|
|
|
|
|
owner_thread->AddWaiter(cur_thread);
|
|
|
|
|
|
|
|
|
|
// Set thread states.
|
|
|
|
|
cur_thread->SetState(ThreadState::Waiting);
|
|
|
|
|
// Begin waiting to hold the lock.
|
|
|
|
|
cur_thread->BeginWait(std::addressof(wait_queue));
|
|
|
|
|
|
|
|
|
|
if (owner_thread->IsSuspended()) {
|
|
|
|
|
owner_thread->ContinueIfHasKernelWaiters();
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// We're no longer waiting on the lock owner.
|
|
|
|
|
{
|
|
|
|
|
KScopedSchedulerLock sl{kernel};
|
|
|
|
|
|
|
|
|
|
if (KThread* owner_thread = cur_thread->GetLockOwner(); owner_thread != nullptr) {
|
|
|
|
|
owner_thread->RemoveWaiter(cur_thread);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
return true;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
void KLightLock::UnlockSlowPath(uintptr_t _cur_thread) {
|
|
|
|
@ -81,22 +84,20 @@ void KLightLock::UnlockSlowPath(uintptr_t _cur_thread) {
|
|
|
|
|
|
|
|
|
|
// Unlock.
|
|
|
|
|
{
|
|
|
|
|
KScopedSchedulerLock sl{kernel};
|
|
|
|
|
KScopedSchedulerLock sl(kernel);
|
|
|
|
|
|
|
|
|
|
// Get the next owner.
|
|
|
|
|
s32 num_waiters = 0;
|
|
|
|
|
s32 num_waiters;
|
|
|
|
|
KThread* next_owner = owner_thread->RemoveWaiterByKey(
|
|
|
|
|
std::addressof(num_waiters), reinterpret_cast<uintptr_t>(std::addressof(tag)));
|
|
|
|
|
|
|
|
|
|
// Pass the lock to the next owner.
|
|
|
|
|
uintptr_t next_tag = 0;
|
|
|
|
|
if (next_owner != nullptr) {
|
|
|
|
|
next_tag = reinterpret_cast<uintptr_t>(next_owner);
|
|
|
|
|
if (num_waiters > 1) {
|
|
|
|
|
next_tag |= 0x1;
|
|
|
|
|
}
|
|
|
|
|
next_tag =
|
|
|
|
|
reinterpret_cast<uintptr_t>(next_owner) | static_cast<uintptr_t>(num_waiters > 1);
|
|
|
|
|
|
|
|
|
|
next_owner->SetState(ThreadState::Runnable);
|
|
|
|
|
next_owner->EndWait(ResultSuccess);
|
|
|
|
|
|
|
|
|
|
if (next_owner->IsSuspended()) {
|
|
|
|
|
next_owner->ContinueIfHasKernelWaiters();
|
|
|
|
@ -110,7 +111,7 @@ void KLightLock::UnlockSlowPath(uintptr_t _cur_thread) {
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// Write the new tag value.
|
|
|
|
|
tag.store(next_tag);
|
|
|
|
|
tag.store(next_tag, std::memory_order_release);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|