2019-01-23 19:17:55 -08:00
|
|
|
// Copyright 2019 yuzu Emulator Project
|
|
|
|
// Licensed under GPLv2 or any later version
|
|
|
|
// Refer to the license.txt file included.
|
|
|
|
|
|
|
|
#include "common/assert.h"
|
|
|
|
#include "common/microprofile.h"
|
2019-03-30 17:08:09 -07:00
|
|
|
#include "core/core.h"
|
|
|
|
#include "core/core_timing.h"
|
|
|
|
#include "core/core_timing_util.h"
|
2019-01-23 19:17:55 -08:00
|
|
|
#include "core/frontend/scope_acquire_window_context.h"
|
|
|
|
#include "video_core/dma_pusher.h"
|
|
|
|
#include "video_core/gpu.h"
|
|
|
|
#include "video_core/gpu_thread.h"
|
|
|
|
#include "video_core/renderer_base.h"
|
|
|
|
|
|
|
|
namespace VideoCommon::GPUThread {
|
|
|
|
|
|
|
|
/// Runs the GPU thread
|
|
|
|
static void RunThread(VideoCore::RendererBase& renderer, Tegra::DmaPusher& dma_pusher,
|
|
|
|
SynchState& state) {
|
|
|
|
MicroProfileOnThreadCreate("GpuThread");
|
|
|
|
|
|
|
|
// Wait for first GPU command before acquiring the window context
|
2019-02-18 17:58:32 -08:00
|
|
|
state.WaitForCommands();
|
2019-01-23 19:17:55 -08:00
|
|
|
|
|
|
|
// If emulation was stopped during disk shader loading, abort before trying to acquire context
|
|
|
|
if (!state.is_running) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
Core::Frontend::ScopeAcquireWindowContext acquire_context{renderer.GetRenderWindow()};
|
|
|
|
|
2019-02-18 17:58:32 -08:00
|
|
|
CommandDataContainer next;
|
2019-01-23 19:17:55 -08:00
|
|
|
while (state.is_running) {
|
2019-02-18 17:58:32 -08:00
|
|
|
state.WaitForCommands();
|
|
|
|
while (!state.queue.Empty()) {
|
|
|
|
state.queue.Pop(next);
|
|
|
|
if (const auto submit_list = std::get_if<SubmitListCommand>(&next.data)) {
|
|
|
|
dma_pusher.Push(std::move(submit_list->entries));
|
|
|
|
dma_pusher.DispatchCalls();
|
|
|
|
} else if (const auto data = std::get_if<SwapBuffersCommand>(&next.data)) {
|
|
|
|
renderer.SwapBuffers(std::move(data->framebuffer));
|
|
|
|
} else if (const auto data = std::get_if<FlushRegionCommand>(&next.data)) {
|
|
|
|
renderer.Rasterizer().FlushRegion(data->addr, data->size);
|
|
|
|
} else if (const auto data = std::get_if<InvalidateRegionCommand>(&next.data)) {
|
|
|
|
renderer.Rasterizer().InvalidateRegion(data->addr, data->size);
|
|
|
|
} else if (const auto data = std::get_if<EndProcessingCommand>(&next.data)) {
|
|
|
|
return;
|
|
|
|
} else {
|
|
|
|
UNREACHABLE();
|
|
|
|
}
|
2019-03-30 17:08:09 -07:00
|
|
|
state.signaled_fence = next.fence;
|
|
|
|
state.TrySynchronize();
|
2019-01-23 19:17:55 -08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-03-30 17:08:09 -07:00
|
|
|
ThreadManager::ThreadManager(Core::System& system, VideoCore::RendererBase& renderer,
|
|
|
|
Tegra::DmaPusher& dma_pusher)
|
|
|
|
: system{system}, thread{RunThread, std::ref(renderer), std::ref(dma_pusher), std::ref(state)} {
|
|
|
|
synchronization_event = system.CoreTiming().RegisterEvent(
|
2019-04-05 15:37:34 -07:00
|
|
|
"GPUThreadSynch", [this](u64 fence, s64) { state.WaitForSynchronization(fence); });
|
2019-03-30 17:08:09 -07:00
|
|
|
}
|
2019-01-23 19:17:55 -08:00
|
|
|
|
|
|
|
ThreadManager::~ThreadManager() {
|
2019-02-18 17:58:32 -08:00
|
|
|
// Notify GPU thread that a shutdown is pending
|
|
|
|
PushCommand(EndProcessingCommand());
|
2019-01-23 19:17:55 -08:00
|
|
|
thread.join();
|
|
|
|
}
|
|
|
|
|
|
|
|
void ThreadManager::SubmitList(Tegra::CommandList&& entries) {
|
2019-03-30 17:08:09 -07:00
|
|
|
const u64 fence{PushCommand(SubmitListCommand(std::move(entries)))};
|
|
|
|
const s64 synchronization_ticks{Core::Timing::usToCycles(9000)};
|
|
|
|
system.CoreTiming().ScheduleEvent(synchronization_ticks, synchronization_event, fence);
|
2019-01-23 19:17:55 -08:00
|
|
|
}
|
|
|
|
|
|
|
|
void ThreadManager::SwapBuffers(
|
|
|
|
std::optional<std::reference_wrapper<const Tegra::FramebufferConfig>> framebuffer) {
|
2019-02-18 17:58:32 -08:00
|
|
|
PushCommand(SwapBuffersCommand(std::move(framebuffer)));
|
2019-01-23 19:17:55 -08:00
|
|
|
}
|
|
|
|
|
2019-02-18 17:58:32 -08:00
|
|
|
void ThreadManager::FlushRegion(CacheAddr addr, u64 size) {
|
|
|
|
PushCommand(FlushRegionCommand(addr, size));
|
2019-01-23 19:17:55 -08:00
|
|
|
}
|
|
|
|
|
2019-02-18 17:58:32 -08:00
|
|
|
void ThreadManager::InvalidateRegion(CacheAddr addr, u64 size) {
|
|
|
|
if (state.queue.Empty()) {
|
|
|
|
// It's quicker to invalidate a single region on the CPU if the queue is already empty
|
2019-03-30 17:08:09 -07:00
|
|
|
system.Renderer().Rasterizer().InvalidateRegion(addr, size);
|
2019-02-18 17:58:32 -08:00
|
|
|
} else {
|
|
|
|
PushCommand(InvalidateRegionCommand(addr, size));
|
|
|
|
}
|
2019-01-23 19:17:55 -08:00
|
|
|
}
|
|
|
|
|
2019-02-18 17:58:32 -08:00
|
|
|
void ThreadManager::FlushAndInvalidateRegion(CacheAddr addr, u64 size) {
|
|
|
|
// Skip flush on asynch mode, as FlushAndInvalidateRegion is not used for anything too important
|
2019-02-08 23:55:45 -08:00
|
|
|
InvalidateRegion(addr, size);
|
2019-01-23 19:17:55 -08:00
|
|
|
}
|
|
|
|
|
2019-03-30 17:08:09 -07:00
|
|
|
u64 ThreadManager::PushCommand(CommandData&& command_data) {
|
|
|
|
const u64 fence{++state.last_fence};
|
|
|
|
state.queue.Push(CommandDataContainer(std::move(command_data), fence));
|
2019-02-18 17:58:32 -08:00
|
|
|
state.SignalCommands();
|
2019-03-30 17:08:09 -07:00
|
|
|
return fence;
|
|
|
|
}
|
|
|
|
|
|
|
|
MICROPROFILE_DEFINE(GPU_wait, "GPU", "Wait for the GPU", MP_RGB(128, 128, 192));
|
|
|
|
void SynchState::WaitForSynchronization(u64 fence) {
|
|
|
|
if (signaled_fence >= fence) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Wait for the GPU to be idle (all commands to be executed)
|
|
|
|
{
|
|
|
|
MICROPROFILE_SCOPE(GPU_wait);
|
|
|
|
std::unique_lock<std::mutex> lock{synchronization_mutex};
|
|
|
|
synchronization_condition.wait(lock, [this, fence] { return signaled_fence >= fence; });
|
|
|
|
}
|
2019-01-23 19:17:55 -08:00
|
|
|
}
|
|
|
|
|
|
|
|
} // namespace VideoCommon::GPUThread
|