mirror of
https://github.com/yuzu-mirror/yuzu.git
synced 2024-11-09 15:49:59 +00:00
kernel: Unconditionally set thread state when appropriate
This commit is contained in:
parent
aa79ca7a7a
commit
ebd38d66db
2 changed files with 12 additions and 23 deletions
|
@ -59,11 +59,7 @@ void KLightLock::LockSlowPath(uintptr_t _owner, uintptr_t _cur_thread) {
|
||||||
owner_thread->AddWaiter(cur_thread);
|
owner_thread->AddWaiter(cur_thread);
|
||||||
|
|
||||||
// Set thread states.
|
// Set thread states.
|
||||||
if (cur_thread->GetState() == ThreadState::Runnable) {
|
cur_thread->SetState(ThreadState::Waiting);
|
||||||
cur_thread->SetState(ThreadState::Waiting);
|
|
||||||
} else {
|
|
||||||
KScheduler::SetSchedulerUpdateNeeded(kernel);
|
|
||||||
}
|
|
||||||
|
|
||||||
if (owner_thread->IsSuspended()) {
|
if (owner_thread->IsSuspended()) {
|
||||||
owner_thread->ContinueIfHasKernelWaiters();
|
owner_thread->ContinueIfHasKernelWaiters();
|
||||||
|
@ -73,10 +69,9 @@ void KLightLock::LockSlowPath(uintptr_t _owner, uintptr_t _cur_thread) {
|
||||||
// We're no longer waiting on the lock owner.
|
// We're no longer waiting on the lock owner.
|
||||||
{
|
{
|
||||||
KScopedSchedulerLock sl{kernel};
|
KScopedSchedulerLock sl{kernel};
|
||||||
KThread* owner_thread = cur_thread->GetLockOwner();
|
|
||||||
if (owner_thread) {
|
if (KThread* owner_thread = cur_thread->GetLockOwner(); owner_thread != nullptr) {
|
||||||
owner_thread->RemoveWaiter(cur_thread);
|
owner_thread->RemoveWaiter(cur_thread);
|
||||||
KScheduler::SetSchedulerUpdateNeeded(kernel);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -95,17 +90,13 @@ void KLightLock::UnlockSlowPath(uintptr_t _cur_thread) {
|
||||||
|
|
||||||
// Pass the lock to the next owner.
|
// Pass the lock to the next owner.
|
||||||
uintptr_t next_tag = 0;
|
uintptr_t next_tag = 0;
|
||||||
if (next_owner) {
|
if (next_owner != nullptr) {
|
||||||
next_tag = reinterpret_cast<uintptr_t>(next_owner);
|
next_tag = reinterpret_cast<uintptr_t>(next_owner);
|
||||||
if (num_waiters > 1) {
|
if (num_waiters > 1) {
|
||||||
next_tag |= 0x1;
|
next_tag |= 0x1;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (next_owner->GetState() == ThreadState::Waiting) {
|
next_owner->SetState(ThreadState::Runnable);
|
||||||
next_owner->SetState(ThreadState::Runnable);
|
|
||||||
} else {
|
|
||||||
KScheduler::SetSchedulerUpdateNeeded(kernel);
|
|
||||||
}
|
|
||||||
|
|
||||||
if (next_owner->IsSuspended()) {
|
if (next_owner->IsSuspended()) {
|
||||||
next_owner->ContinueIfHasKernelWaiters();
|
next_owner->ContinueIfHasKernelWaiters();
|
||||||
|
|
|
@ -201,17 +201,15 @@ bool KProcess::ReleaseUserException(KThread* thread) {
|
||||||
|
|
||||||
// Remove waiter thread.
|
// Remove waiter thread.
|
||||||
s32 num_waiters{};
|
s32 num_waiters{};
|
||||||
KThread* next = thread->RemoveWaiterByKey(
|
if (KThread* next = thread->RemoveWaiterByKey(
|
||||||
std::addressof(num_waiters),
|
std::addressof(num_waiters),
|
||||||
reinterpret_cast<uintptr_t>(std::addressof(exception_thread)));
|
reinterpret_cast<uintptr_t>(std::addressof(exception_thread)));
|
||||||
if (next != nullptr) {
|
next != nullptr) {
|
||||||
if (next->GetState() == ThreadState::Waiting) {
|
next->SetState(ThreadState::Runnable);
|
||||||
next->SetState(ThreadState::Runnable);
|
|
||||||
} else {
|
|
||||||
KScheduler::SetSchedulerUpdateNeeded(kernel);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
KScheduler::SetSchedulerUpdateNeeded(kernel);
|
||||||
|
|
||||||
return true;
|
return true;
|
||||||
} else {
|
} else {
|
||||||
return false;
|
return false;
|
||||||
|
|
Loading…
Reference in a new issue