Merge pull request #7787 from bunnei/scheduler-deadlock-fix

hle: kernel: KScheduler: Fix deadlock with core waiting for a thread lock that has migrated.
master
Morph 2022-01-28 18:30:29 +07:00 committed by GitHub
commit 60b5670577
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
2 changed files with 24 additions and 23 deletions

@ -258,7 +258,7 @@ private:
private: private:
constexpr void ClearAffinityBit(u64& affinity, s32 core) { constexpr void ClearAffinityBit(u64& affinity, s32 core) {
affinity &= ~(u64(1) << core); affinity &= ~(UINT64_C(1) << core);
} }
constexpr s32 GetNextCore(u64& affinity) { constexpr s32 GetNextCore(u64& affinity) {

@ -710,23 +710,19 @@ void KScheduler::Unload(KThread* thread) {
} }
void KScheduler::Reload(KThread* thread) { void KScheduler::Reload(KThread* thread) {
LOG_TRACE(Kernel, "core {}, reload thread {}", core_id, thread ? thread->GetName() : "nullptr"); LOG_TRACE(Kernel, "core {}, reload thread {}", core_id, thread->GetName());
if (thread) { Core::ARM_Interface& cpu_core = system.ArmInterface(core_id);
ASSERT_MSG(thread->GetState() == ThreadState::Runnable, "Thread must be runnable."); cpu_core.LoadContext(thread->GetContext32());
cpu_core.LoadContext(thread->GetContext64());
Core::ARM_Interface& cpu_core = system.ArmInterface(core_id); cpu_core.SetTlsAddress(thread->GetTLSAddress());
cpu_core.LoadContext(thread->GetContext32()); cpu_core.SetTPIDR_EL0(thread->GetTPIDR_EL0());
cpu_core.LoadContext(thread->GetContext64()); cpu_core.ClearExclusiveState();
cpu_core.SetTlsAddress(thread->GetTLSAddress());
cpu_core.SetTPIDR_EL0(thread->GetTPIDR_EL0());
cpu_core.ClearExclusiveState();
}
} }
void KScheduler::SwitchContextStep2() { void KScheduler::SwitchContextStep2() {
// Load context of new thread // Load context of new thread
Reload(current_thread.load()); Reload(GetCurrentThread());
RescheduleCurrentCore(); RescheduleCurrentCore();
} }
@ -735,13 +731,17 @@ void KScheduler::ScheduleImpl() {
KThread* previous_thread = GetCurrentThread(); KThread* previous_thread = GetCurrentThread();
KThread* next_thread = state.highest_priority_thread; KThread* next_thread = state.highest_priority_thread;
state.needs_scheduling = false; state.needs_scheduling.store(false);
// We never want to schedule a null thread, so use the idle thread if we don't have a next. // We never want to schedule a null thread, so use the idle thread if we don't have a next.
if (next_thread == nullptr) { if (next_thread == nullptr) {
next_thread = idle_thread; next_thread = idle_thread;
} }
if (next_thread->GetCurrentCore() != core_id) {
next_thread->SetCurrentCore(core_id);
}
// We never want to schedule a dummy thread, as these are only used by host threads for locking. // We never want to schedule a dummy thread, as these are only used by host threads for locking.
if (next_thread->GetThreadType() == ThreadType::Dummy) { if (next_thread->GetThreadType() == ThreadType::Dummy) {
ASSERT_MSG(false, "Dummy threads should never be scheduled!"); ASSERT_MSG(false, "Dummy threads should never be scheduled!");
@ -755,14 +755,8 @@ void KScheduler::ScheduleImpl() {
return; return;
} }
if (next_thread->GetCurrentCore() != core_id) { // Update the CPU time tracking variables.
next_thread->SetCurrentCore(core_id);
}
current_thread.store(next_thread);
KProcess* const previous_process = system.Kernel().CurrentProcess(); KProcess* const previous_process = system.Kernel().CurrentProcess();
UpdateLastContextSwitchTime(previous_thread, previous_process); UpdateLastContextSwitchTime(previous_thread, previous_process);
// Save context for previous thread // Save context for previous thread
@ -770,6 +764,10 @@ void KScheduler::ScheduleImpl() {
std::shared_ptr<Common::Fiber>* old_context; std::shared_ptr<Common::Fiber>* old_context;
old_context = &previous_thread->GetHostContext(); old_context = &previous_thread->GetHostContext();
// Set the new thread.
current_thread.store(next_thread);
guard.Unlock(); guard.Unlock();
Common::Fiber::YieldTo(*old_context, *switch_fiber); Common::Fiber::YieldTo(*old_context, *switch_fiber);
@ -797,8 +795,8 @@ void KScheduler::SwitchToCurrent() {
do { do {
auto next_thread = current_thread.load(); auto next_thread = current_thread.load();
if (next_thread != nullptr) { if (next_thread != nullptr) {
next_thread->context_guard.Lock(); const auto locked = next_thread->context_guard.TryLock();
if (next_thread->GetRawState() != ThreadState::Runnable) { if (state.needs_scheduling.load()) {
next_thread->context_guard.Unlock(); next_thread->context_guard.Unlock();
break; break;
} }
@ -806,6 +804,9 @@ void KScheduler::SwitchToCurrent() {
next_thread->context_guard.Unlock(); next_thread->context_guard.Unlock();
break; break;
} }
if (!locked) {
continue;
}
} }
auto thread = next_thread ? next_thread : idle_thread; auto thread = next_thread ? next_thread : idle_thread;
Common::Fiber::YieldTo(switch_fiber, *thread->GetHostContext()); Common::Fiber::YieldTo(switch_fiber, *thread->GetHostContext());