rsx/vk: Restructure surface access barriers and implement RCB/RDB

- Implements render target data load (aka Read Color Buffer/Read Depth Buffer)
- Refactors vulkan surface barrier to be much cleaner.
- Removes redundant surface barrier invocations after doing a merged load
  from surface cache.
- Adds explicit access modes when gathering surfaces from cache.
This commit is contained in:
kd-11 2019-08-06 14:51:34 +03:00 committed by kd-11
parent dfe709d464
commit 5e299111cc
6 changed files with 294 additions and 127 deletions

View File

@ -716,7 +716,7 @@ namespace rsx
} }
template <typename commandbuffer_type> template <typename commandbuffer_type>
std::vector<surface_overlap_info> get_merged_texture_memory_region(commandbuffer_type& cmd, u32 texaddr, u32 required_width, u32 required_height, u32 required_pitch, u8 required_bpp) std::vector<surface_overlap_info> get_merged_texture_memory_region(commandbuffer_type& cmd, u32 texaddr, u32 required_width, u32 required_height, u32 required_pitch, u8 required_bpp, rsx::surface_access access)
{ {
std::vector<surface_overlap_info> result; std::vector<surface_overlap_info> result;
std::vector<std::pair<u32, bool>> dirty; std::vector<std::pair<u32, bool>> dirty;
@ -739,12 +739,6 @@ namespace rsx
if ((this_address + texture_size) <= texaddr) if ((this_address + texture_size) <= texaddr)
continue; continue;
if (surface->read_barrier(cmd); !surface->test())
{
dirty.emplace_back(this_address, is_depth);
continue;
}
surface_overlap_info info; surface_overlap_info info;
info.surface = surface; info.surface = surface;
info.base_address = this_address; info.base_address = this_address;
@ -789,6 +783,13 @@ namespace rsx
info.height = std::min<u32>(required_height, normalized_surface_height - info.src_y); info.height = std::min<u32>(required_height, normalized_surface_height - info.src_y);
} }
// Delay this as much as possible to avoid side-effects of spamming barrier
if (surface->memory_barrier(cmd, access); !surface->test())
{
dirty.emplace_back(this_address, is_depth);
continue;
}
info.is_clipped = (info.width < required_width || info.height < required_height); info.is_clipped = (info.width < required_width || info.height < required_height);
if (auto surface_bpp = surface->get_bpp(); UNLIKELY(surface_bpp != required_bpp)) if (auto surface_bpp = surface->get_bpp(); UNLIKELY(surface_bpp != required_bpp))
@ -943,5 +944,26 @@ namespace rsx
rtt = std::make_pair(0, nullptr); rtt = std::make_pair(0, nullptr);
} }
} }
void invalidate_range(const rsx::address_range& range)
{
for (auto &rtt : m_render_targets_storage)
{
if (range.overlaps(rtt.second->get_memory_range()))
{
rtt.second->clear_rw_barrier();
rtt.second->state_flags |= rsx::surface_state_flags::erase_bkgnd;
}
}
for (auto &ds : m_depth_stencil_storage)
{
if (range.overlaps(ds.second->get_memory_range()))
{
ds.second->clear_rw_barrier();
ds.second->state_flags |= rsx::surface_state_flags::erase_bkgnd;
}
}
}
}; };
} }

View File

@ -2084,7 +2084,7 @@ namespace rsx
// NOTE: Compressed formats require a reupload, facilitated by blit synchronization and/or WCB and are not handled here // NOTE: Compressed formats require a reupload, facilitated by blit synchronization and/or WCB and are not handled here
const auto bpp = get_format_block_size_in_bytes(format); const auto bpp = get_format_block_size_in_bytes(format);
const auto overlapping_fbos = m_rtts.get_merged_texture_memory_region(cmd, texaddr, tex_width, required_surface_height, tex_pitch, bpp); const auto overlapping_fbos = m_rtts.get_merged_texture_memory_region(cmd, texaddr, tex_width, required_surface_height, tex_pitch, bpp, rsx::surface_access::read);
if (!overlapping_fbos.empty() || !overlapping_locals.empty()) if (!overlapping_fbos.empty() || !overlapping_locals.empty())
{ {
@ -2342,7 +2342,7 @@ namespace rsx
auto rtt_lookup = [&m_rtts, &cmd, &scale_x, &scale_y, this](u32 address, u32 width, u32 height, u32 pitch, u8 bpp, bool allow_clipped) -> typename surface_store_type::surface_overlap_info auto rtt_lookup = [&m_rtts, &cmd, &scale_x, &scale_y, this](u32 address, u32 width, u32 height, u32 pitch, u8 bpp, bool allow_clipped) -> typename surface_store_type::surface_overlap_info
{ {
const auto list = m_rtts.get_merged_texture_memory_region(cmd, address, width, height, pitch, bpp); const auto list = m_rtts.get_merged_texture_memory_region(cmd, address, width, height, pitch, bpp, rsx::surface_access::transfer);
if (list.empty()) if (list.empty())
{ {
return {}; return {};

View File

@ -1629,7 +1629,7 @@ void GLGSRender::flip(int buffer, bool emu_flip)
else else
{ {
gl::command_context cmd = { gl_state }; gl::command_context cmd = { gl_state };
const auto overlap_info = m_rtts.get_merged_texture_memory_region(cmd, absolute_address, buffer_width, buffer_height, buffer_pitch, render_target_texture->get_bpp()); const auto overlap_info = m_rtts.get_merged_texture_memory_region(cmd, absolute_address, buffer_width, buffer_height, buffer_pitch, render_target_texture->get_bpp(), rsx::surface_access::read);
if (!overlap_info.empty() && overlap_info.back().surface == render_target_texture) if (!overlap_info.empty() && overlap_info.back().surface == render_target_texture)
{ {

View File

@ -223,8 +223,6 @@ struct gl_render_target_traits
prev.target = sink.get(); prev.target = sink.get();
sink->sync_tag();
if (!sink->old_contents.empty()) if (!sink->old_contents.empty())
{ {
// Deal with this, likely only needs to clear // Deal with this, likely only needs to clear

View File

@ -3224,7 +3224,7 @@ void VKGSRender::flip(int buffer, bool emu_flip)
} }
else else
{ {
const auto overlap_info = m_rtts.get_merged_texture_memory_region(*m_current_command_buffer, absolute_address, buffer_width, buffer_height, buffer_pitch, render_target_texture->get_bpp()); const auto overlap_info = m_rtts.get_merged_texture_memory_region(*m_current_command_buffer, absolute_address, buffer_width, buffer_height, buffer_pitch, render_target_texture->get_bpp(), rsx::surface_access::read);
if (!overlap_info.empty() && overlap_info.back().surface == render_target_texture) if (!overlap_info.empty() && overlap_info.back().surface == render_target_texture)
{ {
// Confirmed to be the newest data source in that range // Confirmed to be the newest data source in that range

View File

@ -13,59 +13,42 @@ namespace vk
void resolve_image(vk::command_buffer& cmd, vk::viewable_image* dst, vk::viewable_image* src); void resolve_image(vk::command_buffer& cmd, vk::viewable_image* dst, vk::viewable_image* src);
void unresolve_image(vk::command_buffer& cmd, vk::viewable_image* dst, vk::viewable_image* src); void unresolve_image(vk::command_buffer& cmd, vk::viewable_image* dst, vk::viewable_image* src);
struct render_target : public viewable_image, public rsx::ref_counted, public rsx::render_target_descriptor<vk::viewable_image*> class render_target : public viewable_image, public rsx::ref_counted, public rsx::render_target_descriptor<vk::viewable_image*>
{ {
u64 frame_tag = 0; // frame id when invalidated, 0 if not invalid // Get the linear resolve target bound to this surface. Initialize if none exists
vk::viewable_image* get_resolve_target_safe(vk::command_buffer& cmd)
using viewable_image::viewable_image;
vk::viewable_image* get_surface(rsx::surface_access access_type) override
{ {
if (samples() == 1 || access_type == rsx::surface_access::write) if (!resolve_surface)
{ {
return this; // Create a resolve surface
auto pdev = vk::get_current_renderer();
const auto resolve_w = width() * samples_x;
const auto resolve_h = height() * samples_y;
VkImageUsageFlags usage = VK_IMAGE_USAGE_TRANSFER_SRC_BIT | VK_IMAGE_USAGE_TRANSFER_DST_BIT | VK_IMAGE_USAGE_SAMPLED_BIT;
usage |= (this->info.usage & (VK_IMAGE_USAGE_STORAGE_BIT | VK_IMAGE_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT));
resolve_surface.reset(new vk::viewable_image(
*pdev,
pdev->get_memory_mapping().device_local,
VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT,
VK_IMAGE_TYPE_2D,
format(),
resolve_w, resolve_h, 1, 1, 1,
VK_SAMPLE_COUNT_1_BIT,
VK_IMAGE_LAYOUT_UNDEFINED,
VK_IMAGE_TILING_OPTIMAL,
usage,
0));
resolve_surface->native_component_map = native_component_map;
resolve_surface->change_layout(cmd, VK_IMAGE_LAYOUT_GENERAL);
} }
// A read barrier should have been called before this!
verify("Read access without explicit barrier" HERE), resolve_surface, !(msaa_flags & rsx::surface_state_flags::require_resolve);
return resolve_surface.get(); return resolve_surface.get();
} }
bool is_depth_surface() const override // Resolve the planar MSAA data into a linear block
{
return !!(aspect() & VK_IMAGE_ASPECT_DEPTH_BIT);
}
void release_ref(vk::viewable_image* t) const override
{
static_cast<vk::render_target*>(t)->release();
}
bool matches_dimensions(u16 _width, u16 _height) const
{
//Use forward scaling to account for rounding and clamping errors
return (rsx::apply_resolution_scale(_width, true) == width()) && (rsx::apply_resolution_scale(_height, true) == height());
}
image_view* get_view(u32 remap_encoding, const std::pair<std::array<u8, 4>, std::array<u8, 4>>& remap,
VkImageAspectFlags mask = VK_IMAGE_ASPECT_COLOR_BIT | VK_IMAGE_ASPECT_DEPTH_BIT) override
{
if (remap_encoding != 0xDEADBEEF && resolve_surface)
{
return resolve_surface->get_view(remap_encoding, remap, mask);
}
else
{
if (remap_encoding == 0xDEADBEEF)
{
// Special encoding to skip the resolve target fetch
remap_encoding = 0xAAE4;
}
return vk::viewable_image::get_view(remap_encoding, remap, mask);
}
}
void resolve(vk::command_buffer& cmd) void resolve(vk::command_buffer& cmd)
{ {
VkImageSubresourceRange range = { aspect(), 0, 1, 0, 1 }; VkImageSubresourceRange range = { aspect(), 0, 1, 0, 1 };
@ -140,6 +123,7 @@ namespace vk
msaa_flags &= ~(rsx::surface_state_flags::require_resolve); msaa_flags &= ~(rsx::surface_state_flags::require_resolve);
} }
// Unresolve the linear data into planar MSAA data
void unresolve(vk::command_buffer& cmd) void unresolve(vk::command_buffer& cmd)
{ {
verify(HERE), !(msaa_flags & rsx::surface_state_flags::require_resolve); verify(HERE), !(msaa_flags & rsx::surface_state_flags::require_resolve);
@ -212,94 +196,229 @@ namespace vk
msaa_flags &= ~(rsx::surface_state_flags::require_unresolve); msaa_flags &= ~(rsx::surface_state_flags::require_unresolve);
} }
void memory_barrier(vk::command_buffer& cmd, rsx::surface_access access) // Default-initialize memory without loading
void clear_memory(vk::command_buffer& cmd, vk::image *surface)
{ {
// Helper to optionally clear/initialize memory contents depending on barrier type const auto optimal_layout = (surface->current_layout == VK_IMAGE_LAYOUT_GENERAL) ?
auto clear_surface_impl = [&cmd, this](vk::image* surface) VK_IMAGE_LAYOUT_GENERAL :
VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL;
surface->push_layout(cmd, optimal_layout);
VkImageSubresourceRange range{ surface->aspect(), 0, 1, 0, 1 };
if (surface->aspect() & VK_IMAGE_ASPECT_COLOR_BIT)
{ {
const auto optimal_layout = (surface->current_layout == VK_IMAGE_LAYOUT_GENERAL) ? VkClearColorValue color = { 0.f, 0.f, 0.f, 1.f };
VK_IMAGE_LAYOUT_GENERAL : vkCmdClearColorImage(cmd, surface->value, surface->current_layout, &color, 1, &range);
VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL; }
else
{
VkClearDepthStencilValue clear{ 1.f, 255 };
vkCmdClearDepthStencilImage(cmd, surface->value, surface->current_layout, &clear, 1, &range);
}
surface->push_layout(cmd, optimal_layout); surface->pop_layout(cmd);
VkImageSubresourceRange range{ surface->aspect(), 0, 1, 0, 1 }; if (surface == this)
if (surface->aspect() & VK_IMAGE_ASPECT_COLOR_BIT) {
state_flags &= ~rsx::surface_state_flags::erase_bkgnd;
}
}
// Load memory from cell and use to initialize the surface
void load_memory(vk::command_buffer& cmd)
{
auto& upload_heap = *vk::get_upload_heap();
u32 gcm_format;
if (is_depth_surface())
{
gcm_format = get_compatible_gcm_format(format_info.gcm_depth_format).first;
}
else
{
auto fmt = get_compatible_gcm_format(format_info.gcm_color_format);
if (fmt.second)
{ {
VkClearColorValue color = { 0.f, 0.f, 0.f, 1.f }; switch (fmt.first)
vkCmdClearColorImage(cmd, surface->value, surface->current_layout, &color, 1, &range); {
case CELL_GCM_TEXTURE_A8R8G8B8:
case CELL_GCM_TEXTURE_D8R8G8B8:
//Hack
gcm_format = CELL_GCM_TEXTURE_X32_FLOAT;
break;
default:
gcm_format = fmt.first;
break;
}
} }
else else
{ {
VkClearDepthStencilValue clear{ 1.f, 255 }; gcm_format = fmt.first;
vkCmdClearDepthStencilImage(cmd, surface->value, surface->current_layout, &clear, 1, &range);
} }
}
surface->pop_layout(cmd); rsx_subresource_layout subres{};
subres.width_in_block = surface_width * samples_x;
subres.height_in_block = surface_height * samples_y;
subres.pitch_in_block = rsx_pitch / get_bpp();
subres.depth = 1;
subres.data = { (const gsl::byte*)vm::get_super_ptr(base_addr), s32(rsx_pitch * surface_height * samples_y) };
if (surface == this) if (LIKELY(g_cfg.video.resolution_scale_percent == 100 && samples() == 1))
{
state_flags &= ~rsx::surface_state_flags::erase_bkgnd;
}
};
auto get_resolve_target = [&]()
{ {
if (!resolve_surface) push_layout(cmd, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL);
vk::copy_mipmaped_image_using_buffer(cmd, this, { subres }, gcm_format, false, 1, aspect(), upload_heap, rsx_pitch);
pop_layout(cmd);
}
else
{
vk::image* content = nullptr;
vk::image* final_dst = (samples() > 1) ? get_resolve_target_safe(cmd) : this;
if (LIKELY(g_cfg.video.resolution_scale_percent == 100))
{ {
// Create a resolve surface verify(HERE), samples() > 1;
auto pdev = vk::get_current_renderer(); content = get_resolve_target_safe(cmd);
const auto resolve_w = width() * samples_x; }
const auto resolve_h = height() * samples_y; else
{
VkImageUsageFlags usage = VK_IMAGE_USAGE_TRANSFER_SRC_BIT | VK_IMAGE_USAGE_TRANSFER_DST_BIT | VK_IMAGE_USAGE_SAMPLED_BIT; content = vk::get_typeless_helper(format(), subres.width_in_block, subres.height_in_block);
usage |= (this->info.usage & (VK_IMAGE_USAGE_STORAGE_BIT | VK_IMAGE_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT)); content->change_layout(cmd, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL);
resolve_surface.reset(new vk::viewable_image(
*pdev,
pdev->get_memory_mapping().device_local,
VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT,
VK_IMAGE_TYPE_2D,
format(),
resolve_w, resolve_h, 1, 1, 1,
VK_SAMPLE_COUNT_1_BIT,
VK_IMAGE_LAYOUT_UNDEFINED,
VK_IMAGE_TILING_OPTIMAL,
usage,
0));
resolve_surface->native_component_map = native_component_map;
resolve_surface->change_layout(cmd, VK_IMAGE_LAYOUT_GENERAL);
} }
return resolve_surface.get(); // Load Cell data into temp buffer
}; vk::copy_mipmaped_image_using_buffer(cmd, content, { subres }, gcm_format, false, 1, aspect(), upload_heap, rsx_pitch);
// Write into final image
if (content != final_dst)
{
vk::copy_scaled_image(cmd, content->value, final_dst->value, content->current_layout, final_dst->current_layout,
{ 0, 0, subres.width_in_block, subres.height_in_block }, { 0, 0, (s32)final_dst->width(), (s32)final_dst->height() },
1, aspect(), true, aspect() == VK_IMAGE_ASPECT_COLOR_BIT ? VK_FILTER_LINEAR : VK_FILTER_NEAREST,
format(), format());
}
if (samples() > 1)
{
// Trigger unresolve
msaa_flags = rsx::surface_state_flags::require_unresolve;
}
}
state_flags &= ~rsx::surface_state_flags::erase_bkgnd;
}
void initialize_memory(vk::command_buffer& cmd, bool read_access)
{
const bool memory_load = is_depth_surface() ?
!!g_cfg.video.read_depth_buffer :
!!g_cfg.video.read_color_buffers;
if (!memory_load)
{
clear_memory(cmd, this);
if (read_access && samples() > 1)
{
// Only clear the resolve surface if reading from it, otherwise it's a waste
clear_memory(cmd, get_resolve_target_safe(cmd));
}
msaa_flags = rsx::surface_state_flags::ready;
}
else
{
load_memory(cmd);
}
}
public:
u64 frame_tag = 0; // frame id when invalidated, 0 if not invalid
using viewable_image::viewable_image;
vk::viewable_image* get_surface(rsx::surface_access access_type) override
{
if (samples() == 1 || access_type == rsx::surface_access::write)
{
return this;
}
// A read barrier should have been called before this!
verify("Read access without explicit barrier" HERE), resolve_surface, !(msaa_flags & rsx::surface_state_flags::require_resolve);
return resolve_surface.get();
}
bool is_depth_surface() const override
{
return !!(aspect() & VK_IMAGE_ASPECT_DEPTH_BIT);
}
void release_ref(vk::viewable_image* t) const override
{
static_cast<vk::render_target*>(t)->release();
}
bool matches_dimensions(u16 _width, u16 _height) const
{
//Use forward scaling to account for rounding and clamping errors
return (rsx::apply_resolution_scale(_width, true) == width()) && (rsx::apply_resolution_scale(_height, true) == height());
}
image_view* get_view(u32 remap_encoding, const std::pair<std::array<u8, 4>, std::array<u8, 4>>& remap,
VkImageAspectFlags mask = VK_IMAGE_ASPECT_COLOR_BIT | VK_IMAGE_ASPECT_DEPTH_BIT) override
{
if (remap_encoding != 0xDEADBEEF && resolve_surface)
{
return resolve_surface->get_view(remap_encoding, remap, mask);
}
else
{
if (remap_encoding == 0xDEADBEEF)
{
// Special encoding to skip the resolve target fetch
remap_encoding = 0xAAE4;
}
return vk::viewable_image::get_view(remap_encoding, remap, mask);
}
}
void memory_barrier(vk::command_buffer& cmd, rsx::surface_access access)
{
const bool read_access = (access != rsx::surface_access::write); const bool read_access = (access != rsx::surface_access::write);
if (samples() > 1 && read_access) const bool is_depth = is_depth_surface();
if ((g_cfg.video.read_color_buffers && !is_depth) ||
(g_cfg.video.read_depth_buffer && is_depth))
{ {
get_resolve_target(); // TODO: Decide what to do when memory loads are disabled but the underlying has memory changed
// NOTE: Assume test() is expensive when in a pinch
if (last_use_tag && state_flags == rsx::surface_state_flags::ready && !test())
{
// TODO: Figure out why merely returning and failing the test does not work when reading (TLoU)
// The result should have been the same either way
state_flags |= rsx::surface_state_flags::erase_bkgnd;
}
} }
if (LIKELY(old_contents.empty())) if (LIKELY(old_contents.empty()))
{ {
if (state_flags & rsx::surface_state_flags::erase_bkgnd) if (state_flags & rsx::surface_state_flags::erase_bkgnd)
{ {
clear_surface_impl(this); // NOTE: This step CAN introduce MSAA flags!
initialize_memory(cmd, read_access);
if (resolve_surface && read_access) verify(HERE), state_flags == rsx::surface_state_flags::ready;
{ on_write(rsx::get_shared_tag(), static_cast<rsx::surface_state_flags>(msaa_flags));
// Only clear the resolve surface if reading from it, otherwise it's a waste
clear_surface_impl(resolve_surface.get());
}
on_write(rsx::get_shared_tag(), rsx::surface_state_flags::ready);
} }
else if (msaa_flags & rsx::surface_state_flags::require_resolve)
if (msaa_flags & rsx::surface_state_flags::require_resolve)
{ {
if (read_access) if (read_access)
{ {
// Only do this step when read access is required // Only do this step when read access is required
get_resolve_target_safe(cmd);
resolve(cmd); resolve(cmd);
} }
} }
@ -308,6 +427,7 @@ namespace vk
if (!read_access) if (!read_access)
{ {
// Only do this step when it is needed to start rendering // Only do this step when it is needed to start rendering
verify(HERE), resolve_surface;
unresolve(cmd); unresolve(cmd);
} }
} }
@ -316,11 +436,13 @@ namespace vk
} }
// Memory transfers // Memory transfers
vk::image *target_image = (samples() > 1) ? get_resolve_target() : this; vk::image *target_image = (samples() > 1) ? get_resolve_target_safe(cmd) : this;
vk::blitter hw_blitter; vk::blitter hw_blitter;
bool optimize_copy = true;
const auto dst_bpp = get_bpp(); const auto dst_bpp = get_bpp();
unsigned first = prepare_rw_barrier_for_transfer(this); unsigned first = prepare_rw_barrier_for_transfer(this);
bool optimize_copy = true;
bool any_valid_writes = false;
for (auto i = first; i < old_contents.size(); ++i) for (auto i = first; i < old_contents.size(); ++i)
{ {
@ -328,10 +450,19 @@ namespace vk
auto src_texture = static_cast<vk::render_target*>(section.source); auto src_texture = static_cast<vk::render_target*>(section.source);
src_texture->read_barrier(cmd); src_texture->read_barrier(cmd);
if (LIKELY(src_texture->test()))
{
any_valid_writes = true;
}
else
{
continue;
}
const auto src_bpp = src_texture->get_bpp(); const auto src_bpp = src_texture->get_bpp();
rsx::typeless_xfer typeless_info{}; rsx::typeless_xfer typeless_info{};
if (src_texture->info.format == info.format) if (LIKELY(src_texture->info.format == info.format))
{ {
verify(HERE), src_bpp == dst_bpp; verify(HERE), src_bpp == dst_bpp;
} }
@ -343,7 +474,7 @@ namespace vk
typeless_info.src_is_typeless = true; typeless_info.src_is_typeless = true;
typeless_info.src_context = rsx::texture_upload_context::framebuffer_storage; typeless_info.src_context = rsx::texture_upload_context::framebuffer_storage;
typeless_info.src_native_format_override = (u32)info.format; typeless_info.src_native_format_override = (u32)info.format;
typeless_info.src_is_depth = src_texture->is_depth_surface(); typeless_info.src_is_depth = is_depth;
typeless_info.src_scaling_hint = f32(src_bpp) / dst_bpp; typeless_info.src_scaling_hint = f32(src_bpp) / dst_bpp;
} }
} }
@ -371,12 +502,12 @@ namespace vk
} }
else if (state_flags & rsx::surface_state_flags::erase_bkgnd) else if (state_flags & rsx::surface_state_flags::erase_bkgnd)
{ {
clear_surface_impl(target_image); // Might introduce MSAA flags
initialize_memory(cmd, false);
state_flags &= ~(rsx::surface_state_flags::erase_bkgnd); verify(HERE), state_flags == rsx::surface_state_flags::ready;
msaa_flags = rsx::surface_state_flags::ready;
} }
else if (msaa_flags & rsx::surface_state_flags::require_resolve)
if (msaa_flags & rsx::surface_state_flags::require_resolve)
{ {
// Need to forward resolve this // Need to forward resolve this
resolve(cmd); resolve(cmd);
@ -393,6 +524,24 @@ namespace vk
optimize_copy = optimize_copy && !memory_load; optimize_copy = optimize_copy && !memory_load;
} }
if (UNLIKELY(!any_valid_writes))
{
LOG_WARNING(RSX, "Surface at 0x%x inherited stale references", base_addr);
clear_rw_barrier();
shuffle_tag();
if (!read_access)
{
// This will be modified either way
state_flags |= rsx::surface_state_flags::erase_bkgnd;
memory_barrier(cmd, access);
}
return;
}
// NOTE: Optimize flag relates to stencil resolve/unresolve for NVIDIA.
on_write_copy(0, optimize_copy); on_write_copy(0, optimize_copy);
if (!read_access && samples() > 1) if (!read_access && samples() > 1)
@ -592,8 +741,6 @@ namespace rsx
prev.target = sink.get(); prev.target = sink.get();
sink->rsx_pitch = ref->get_rsx_pitch(); sink->rsx_pitch = ref->get_rsx_pitch();
sink->sync_tag();
if (!sink->old_contents.empty()) if (!sink->old_contents.empty())
{ {
// Deal with this, likely only needs to clear // Deal with this, likely only needs to clear