buffer_cache: Fix copy based uniform bindings tracking
This commit is contained in:
		@@ -680,6 +680,9 @@ void BufferCache<P>::SetUniformBuffersState(const std::array<u32, NUM_STAGES>& m
 | 
				
			|||||||
                                            const UniformBufferSizes* sizes) {
 | 
					                                            const UniformBufferSizes* sizes) {
 | 
				
			||||||
    if constexpr (HAS_PERSISTENT_UNIFORM_BUFFER_BINDINGS) {
 | 
					    if constexpr (HAS_PERSISTENT_UNIFORM_BUFFER_BINDINGS) {
 | 
				
			||||||
        if (enabled_uniform_buffer_masks != mask) {
 | 
					        if (enabled_uniform_buffer_masks != mask) {
 | 
				
			||||||
 | 
					            if constexpr (IS_OPENGL) {
 | 
				
			||||||
 | 
					                fast_bound_uniform_buffers.fill(0);
 | 
				
			||||||
 | 
					            }
 | 
				
			||||||
            dirty_uniform_buffers.fill(~u32{0});
 | 
					            dirty_uniform_buffers.fill(~u32{0});
 | 
				
			||||||
        }
 | 
					        }
 | 
				
			||||||
    }
 | 
					    }
 | 
				
			||||||
@@ -1020,6 +1023,7 @@ void BufferCache<P>::BindHostGraphicsUniformBuffer(size_t stage, u32 index, u32
 | 
				
			|||||||
                // Fast path for Nvidia
 | 
					                // Fast path for Nvidia
 | 
				
			||||||
                if (!HasFastUniformBufferBound(stage, binding_index)) {
 | 
					                if (!HasFastUniformBufferBound(stage, binding_index)) {
 | 
				
			||||||
                    // We only have to bind when the currently bound buffer is not the fast version
 | 
					                    // We only have to bind when the currently bound buffer is not the fast version
 | 
				
			||||||
 | 
					                    fast_bound_uniform_buffers[stage] |= 1U << binding_index;
 | 
				
			||||||
                    runtime.BindFastUniformBuffer(stage, binding_index, size);
 | 
					                    runtime.BindFastUniformBuffer(stage, binding_index, size);
 | 
				
			||||||
                }
 | 
					                }
 | 
				
			||||||
                const auto span = ImmediateBufferWithData(cpu_addr, size);
 | 
					                const auto span = ImmediateBufferWithData(cpu_addr, size);
 | 
				
			||||||
@@ -1027,8 +1031,9 @@ void BufferCache<P>::BindHostGraphicsUniformBuffer(size_t stage, u32 index, u32
 | 
				
			|||||||
                return;
 | 
					                return;
 | 
				
			||||||
            }
 | 
					            }
 | 
				
			||||||
        }
 | 
					        }
 | 
				
			||||||
        fast_bound_uniform_buffers[stage] |= 1U << binding_index;
 | 
					        if constexpr (IS_OPENGL) {
 | 
				
			||||||
 | 
					            fast_bound_uniform_buffers[stage] |= 1U << binding_index;
 | 
				
			||||||
 | 
					        }
 | 
				
			||||||
        // Stream buffer path to avoid stalling on non-Nvidia drivers or Vulkan
 | 
					        // Stream buffer path to avoid stalling on non-Nvidia drivers or Vulkan
 | 
				
			||||||
        const std::span<u8> span = runtime.BindMappedUniformBuffer(stage, binding_index, size);
 | 
					        const std::span<u8> span = runtime.BindMappedUniformBuffer(stage, binding_index, size);
 | 
				
			||||||
        cpu_memory.ReadBlockUnsafe(cpu_addr, span.data(), size);
 | 
					        cpu_memory.ReadBlockUnsafe(cpu_addr, span.data(), size);
 | 
				
			||||||
@@ -1046,9 +1051,15 @@ void BufferCache<P>::BindHostGraphicsUniformBuffer(size_t stage, u32 index, u32
 | 
				
			|||||||
        // This exists to avoid instances where the fast buffer is bound and a GPU write happens
 | 
					        // This exists to avoid instances where the fast buffer is bound and a GPU write happens
 | 
				
			||||||
        return;
 | 
					        return;
 | 
				
			||||||
    }
 | 
					    }
 | 
				
			||||||
    fast_bound_uniform_buffers[stage] &= ~(1U << binding_index);
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
    const u32 offset = buffer.Offset(cpu_addr);
 | 
					    const u32 offset = buffer.Offset(cpu_addr);
 | 
				
			||||||
 | 
					    if constexpr (IS_OPENGL) {
 | 
				
			||||||
 | 
					        // Fast buffer will be unbound
 | 
				
			||||||
 | 
					        fast_bound_uniform_buffers[stage] &= ~(1U << binding_index);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        // Mark the index as dirty if offset doesn't match
 | 
				
			||||||
 | 
					        const bool is_copy_bind = offset != 0 && !runtime.SupportsNonZeroUniformOffset();
 | 
				
			||||||
 | 
					        dirty_uniform_buffers[stage] |= (is_copy_bind ? 1U : 0U) << index;
 | 
				
			||||||
 | 
					    }
 | 
				
			||||||
    if constexpr (NEEDS_BIND_UNIFORM_INDEX) {
 | 
					    if constexpr (NEEDS_BIND_UNIFORM_INDEX) {
 | 
				
			||||||
        runtime.BindUniformBuffer(stage, binding_index, buffer, offset, size);
 | 
					        runtime.BindUniformBuffer(stage, binding_index, buffer, offset, size);
 | 
				
			||||||
    } else {
 | 
					    } else {
 | 
				
			||||||
 
 | 
				
			|||||||
@@ -92,16 +92,14 @@ public:
 | 
				
			|||||||
                         VideoCore::Surface::PixelFormat format);
 | 
					                         VideoCore::Surface::PixelFormat format);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    void BindFastUniformBuffer(size_t stage, u32 binding_index, u32 size) {
 | 
					    void BindFastUniformBuffer(size_t stage, u32 binding_index, u32 size) {
 | 
				
			||||||
 | 
					        const GLuint handle = fast_uniforms[stage][binding_index].handle;
 | 
				
			||||||
 | 
					        const GLsizeiptr gl_size = static_cast<GLsizeiptr>(size);
 | 
				
			||||||
        if (use_assembly_shaders) {
 | 
					        if (use_assembly_shaders) {
 | 
				
			||||||
            const GLuint handle = fast_uniforms[stage][binding_index].handle;
 | 
					 | 
				
			||||||
            const GLsizeiptr gl_size = static_cast<GLsizeiptr>(size);
 | 
					 | 
				
			||||||
            glBindBufferRangeNV(PABO_LUT[stage], binding_index, handle, 0, gl_size);
 | 
					            glBindBufferRangeNV(PABO_LUT[stage], binding_index, handle, 0, gl_size);
 | 
				
			||||||
        } else {
 | 
					        } else {
 | 
				
			||||||
            const GLuint base_binding = graphics_base_uniform_bindings[stage];
 | 
					            const GLuint base_binding = graphics_base_uniform_bindings[stage];
 | 
				
			||||||
            const GLuint binding = base_binding + binding_index;
 | 
					            const GLuint binding = base_binding + binding_index;
 | 
				
			||||||
            glBindBufferRange(GL_UNIFORM_BUFFER, binding,
 | 
					            glBindBufferRange(GL_UNIFORM_BUFFER, binding, handle, 0, gl_size);
 | 
				
			||||||
                              fast_uniforms[stage][binding_index].handle, 0,
 | 
					 | 
				
			||||||
                              static_cast<GLsizeiptr>(size));
 | 
					 | 
				
			||||||
        }
 | 
					        }
 | 
				
			||||||
    }
 | 
					    }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
@@ -134,6 +132,10 @@ public:
 | 
				
			|||||||
        return has_fast_buffer_sub_data;
 | 
					        return has_fast_buffer_sub_data;
 | 
				
			||||||
    }
 | 
					    }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    [[nodiscard]] bool SupportsNonZeroUniformOffset() const noexcept {
 | 
				
			||||||
 | 
					        return !use_assembly_shaders;
 | 
				
			||||||
 | 
					    }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    void SetBaseUniformBindings(const std::array<GLuint, 5>& bindings) {
 | 
					    void SetBaseUniformBindings(const std::array<GLuint, 5>& bindings) {
 | 
				
			||||||
        graphics_base_uniform_bindings = bindings;
 | 
					        graphics_base_uniform_bindings = bindings;
 | 
				
			||||||
    }
 | 
					    }
 | 
				
			||||||
 
 | 
				
			|||||||
		Reference in New Issue
	
	Block a user