|
|
@ -22,10 +22,11 @@ std::shared_ptr<EventType> CreateEvent(std::string name, TimedCallback&& callbac
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
struct CoreTiming::Event {
|
|
|
|
struct CoreTiming::Event {
|
|
|
|
u64 time;
|
|
|
|
s64 time;
|
|
|
|
u64 fifo_order;
|
|
|
|
u64 fifo_order;
|
|
|
|
std::uintptr_t user_data;
|
|
|
|
std::uintptr_t user_data;
|
|
|
|
std::weak_ptr<EventType> type;
|
|
|
|
std::weak_ptr<EventType> type;
|
|
|
|
|
|
|
|
s64 reschedule_time;
|
|
|
|
|
|
|
|
|
|
|
|
// Sort by time, unless the times are the same, in which case sort by
|
|
|
|
// Sort by time, unless the times are the same, in which case sort by
|
|
|
|
// the order added to the queue
|
|
|
|
// the order added to the queue
|
|
|
@ -58,7 +59,8 @@ void CoreTiming::Initialize(std::function<void()>&& on_thread_init_) {
|
|
|
|
event_fifo_id = 0;
|
|
|
|
event_fifo_id = 0;
|
|
|
|
shutting_down = false;
|
|
|
|
shutting_down = false;
|
|
|
|
ticks = 0;
|
|
|
|
ticks = 0;
|
|
|
|
const auto empty_timed_callback = [](std::uintptr_t, std::chrono::nanoseconds) {};
|
|
|
|
const auto empty_timed_callback = [](std::uintptr_t, u64, std::chrono::nanoseconds)
|
|
|
|
|
|
|
|
-> std::optional<std::chrono::nanoseconds> { return std::nullopt; };
|
|
|
|
ev_lost = CreateEvent("_lost_event", empty_timed_callback);
|
|
|
|
ev_lost = CreateEvent("_lost_event", empty_timed_callback);
|
|
|
|
if (is_multicore) {
|
|
|
|
if (is_multicore) {
|
|
|
|
worker_threads.emplace_back(ThreadEntry, std::ref(*this), 0);
|
|
|
|
worker_threads.emplace_back(ThreadEntry, std::ref(*this), 0);
|
|
|
@ -76,6 +78,7 @@ void CoreTiming::Shutdown() {
|
|
|
|
thread.join();
|
|
|
|
thread.join();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
worker_threads.clear();
|
|
|
|
worker_threads.clear();
|
|
|
|
|
|
|
|
pause_callbacks.clear();
|
|
|
|
ClearPendingEvents();
|
|
|
|
ClearPendingEvents();
|
|
|
|
has_started = false;
|
|
|
|
has_started = false;
|
|
|
|
}
|
|
|
|
}
|
|
|
@ -93,6 +96,14 @@ void CoreTiming::Pause(bool is_paused_) {
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
paused_state.store(is_paused_, std::memory_order_relaxed);
|
|
|
|
paused_state.store(is_paused_, std::memory_order_relaxed);
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
if (!is_paused_) {
|
|
|
|
|
|
|
|
pause_end_time = GetGlobalTimeNs().count();
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
for (auto& cb : pause_callbacks) {
|
|
|
|
|
|
|
|
cb(is_paused_);
|
|
|
|
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void CoreTiming::SyncPause(bool is_paused_) {
|
|
|
|
void CoreTiming::SyncPause(bool is_paused_) {
|
|
|
@ -116,6 +127,14 @@ void CoreTiming::SyncPause(bool is_paused_) {
|
|
|
|
wait_signal_cv.wait(main_lock, [this] { return pause_count == 0; });
|
|
|
|
wait_signal_cv.wait(main_lock, [this] { return pause_count == 0; });
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
if (!is_paused_) {
|
|
|
|
|
|
|
|
pause_end_time = GetGlobalTimeNs().count();
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
for (auto& cb : pause_callbacks) {
|
|
|
|
|
|
|
|
cb(is_paused_);
|
|
|
|
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
bool CoreTiming::IsRunning() const {
|
|
|
|
bool CoreTiming::IsRunning() const {
|
|
|
@ -129,12 +148,30 @@ bool CoreTiming::HasPendingEvents() const {
|
|
|
|
|
|
|
|
|
|
|
|
void CoreTiming::ScheduleEvent(std::chrono::nanoseconds ns_into_future,
|
|
|
|
void CoreTiming::ScheduleEvent(std::chrono::nanoseconds ns_into_future,
|
|
|
|
const std::shared_ptr<EventType>& event_type,
|
|
|
|
const std::shared_ptr<EventType>& event_type,
|
|
|
|
std::uintptr_t user_data) {
|
|
|
|
std::uintptr_t user_data, bool absolute_time) {
|
|
|
|
|
|
|
|
|
|
|
|
std::unique_lock main_lock(event_mutex);
|
|
|
|
std::unique_lock main_lock(event_mutex);
|
|
|
|
const u64 timeout = static_cast<u64>((GetGlobalTimeNs() + ns_into_future).count());
|
|
|
|
const auto next_time{absolute_time ? ns_into_future : GetGlobalTimeNs() + ns_into_future};
|
|
|
|
|
|
|
|
|
|
|
|
event_queue.emplace_back(Event{timeout, event_fifo_id++, user_data, event_type});
|
|
|
|
event_queue.emplace_back(Event{next_time.count(), event_fifo_id++, user_data, event_type, 0});
|
|
|
|
|
|
|
|
pending_events.fetch_add(1, std::memory_order_relaxed);
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
std::push_heap(event_queue.begin(), event_queue.end(), std::greater<>());
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
if (is_multicore) {
|
|
|
|
|
|
|
|
event_cv.notify_one();
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
void CoreTiming::ScheduleLoopingEvent(std::chrono::nanoseconds start_time,
|
|
|
|
|
|
|
|
std::chrono::nanoseconds resched_time,
|
|
|
|
|
|
|
|
const std::shared_ptr<EventType>& event_type,
|
|
|
|
|
|
|
|
std::uintptr_t user_data, bool absolute_time) {
|
|
|
|
|
|
|
|
std::unique_lock main_lock(event_mutex);
|
|
|
|
|
|
|
|
const auto next_time{absolute_time ? start_time : GetGlobalTimeNs() + start_time};
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
event_queue.emplace_back(
|
|
|
|
|
|
|
|
Event{next_time.count(), event_fifo_id++, user_data, event_type, resched_time.count()});
|
|
|
|
pending_events.fetch_add(1, std::memory_order_relaxed);
|
|
|
|
pending_events.fetch_add(1, std::memory_order_relaxed);
|
|
|
|
|
|
|
|
|
|
|
|
std::push_heap(event_queue.begin(), event_queue.end(), std::greater<>());
|
|
|
|
std::push_heap(event_queue.begin(), event_queue.end(), std::greater<>());
|
|
|
@ -213,6 +250,11 @@ void CoreTiming::RemoveEvent(const std::shared_ptr<EventType>& event_type) {
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
void CoreTiming::RegisterPauseCallback(PauseCallback&& callback) {
|
|
|
|
|
|
|
|
std::unique_lock main_lock(event_mutex);
|
|
|
|
|
|
|
|
pause_callbacks.emplace_back(std::move(callback));
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
std::optional<s64> CoreTiming::Advance() {
|
|
|
|
std::optional<s64> CoreTiming::Advance() {
|
|
|
|
global_timer = GetGlobalTimeNs().count();
|
|
|
|
global_timer = GetGlobalTimeNs().count();
|
|
|
|
|
|
|
|
|
|
|
@ -223,14 +265,31 @@ std::optional<s64> CoreTiming::Advance() {
|
|
|
|
event_queue.pop_back();
|
|
|
|
event_queue.pop_back();
|
|
|
|
|
|
|
|
|
|
|
|
if (const auto event_type{evt.type.lock()}) {
|
|
|
|
if (const auto event_type{evt.type.lock()}) {
|
|
|
|
|
|
|
|
|
|
|
|
event_mutex.unlock();
|
|
|
|
event_mutex.unlock();
|
|
|
|
|
|
|
|
|
|
|
|
const s64 delay = static_cast<s64>(GetGlobalTimeNs().count() - evt.time);
|
|
|
|
const auto new_schedule_time{event_type->callback(
|
|
|
|
event_type->callback(evt.user_data, std::chrono::nanoseconds{delay});
|
|
|
|
evt.user_data, evt.time,
|
|
|
|
|
|
|
|
std::chrono::nanoseconds{GetGlobalTimeNs().count() - evt.time})};
|
|
|
|
|
|
|
|
|
|
|
|
event_mutex.lock();
|
|
|
|
event_mutex.lock();
|
|
|
|
pending_events.fetch_sub(1, std::memory_order_relaxed);
|
|
|
|
pending_events.fetch_sub(1, std::memory_order_relaxed);
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
if (evt.reschedule_time != 0) {
|
|
|
|
|
|
|
|
// If this event was scheduled into a pause, its time now is going to be way behind.
|
|
|
|
|
|
|
|
// Re-set this event to continue from the end of the pause.
|
|
|
|
|
|
|
|
auto next_time{evt.time + evt.reschedule_time};
|
|
|
|
|
|
|
|
if (evt.time < pause_end_time) {
|
|
|
|
|
|
|
|
next_time = pause_end_time + evt.reschedule_time;
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
const auto next_schedule_time{new_schedule_time.has_value()
|
|
|
|
|
|
|
|
? new_schedule_time.value().count()
|
|
|
|
|
|
|
|
: evt.reschedule_time};
|
|
|
|
|
|
|
|
event_queue.emplace_back(
|
|
|
|
|
|
|
|
Event{next_time, event_fifo_id++, evt.user_data, evt.type, next_schedule_time});
|
|
|
|
|
|
|
|
pending_events.fetch_add(1, std::memory_order_relaxed);
|
|
|
|
|
|
|
|
std::push_heap(event_queue.begin(), event_queue.end(), std::greater<>());
|
|
|
|
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
global_timer = GetGlobalTimeNs().count();
|
|
|
|
global_timer = GetGlobalTimeNs().count();
|
|
|
|