Files
mesa/src/compiler/nir/tests/vars_tests.cpp
Jason Ekstrand a700a82bda nir: Distinguish between normal uniforms and UBOs
Previously, NIR had a single nir_var_uniform mode used for atomic
counters, UBOs, samplers, images, and normal uniforms.  This commit
splits this into nir_var_uniform and nir_var_ubo where nir_var_uniform
is still a bit of a catch-all but the nir_var_ubo is specific to UBOs.
While we're at it, we also rename shader_storage to ssbo to follow the
convention.

We need this so that we can distinguish between normal uniforms and UBO
access at the deref level without going all the way back variable and
seeing if it has an interface type.

Reviewed-by: Alejandro Piñeiro <apinheiro@igalia.com>
Reviewed-by: Caio Marcelo de Oliveira Filho <caio.oliveira@intel.com>
2019-01-08 00:38:29 +00:00

727 lines
23 KiB
C++

/*
* Copyright © 2018 Intel Corporation
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice (including the next
* paragraph) shall be included in all copies or substantial portions of the
* Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*/
#include <gtest/gtest.h>
#include "nir.h"
#include "nir_builder.h"
namespace {
class nir_vars_test : public ::testing::Test {
protected:
nir_vars_test();
~nir_vars_test();
nir_variable *create_int(nir_variable_mode mode, const char *name) {
if (mode == nir_var_local)
return nir_local_variable_create(b->impl, glsl_int_type(), name);
return nir_variable_create(b->shader, mode, glsl_int_type(), name);
}
nir_variable *create_ivec2(nir_variable_mode mode, const char *name) {
const glsl_type *var_type = glsl_vector_type(GLSL_TYPE_INT, 2);
if (mode == nir_var_local)
return nir_local_variable_create(b->impl, var_type, name);
return nir_variable_create(b->shader, mode, var_type, name);
}
nir_variable **create_many_int(nir_variable_mode mode, const char *prefix, unsigned count) {
nir_variable **result = (nir_variable **)linear_alloc_child(lin_ctx, sizeof(nir_variable *) * count);
for (unsigned i = 0; i < count; i++)
result[i] = create_int(mode, linear_asprintf(lin_ctx, "%s%u", prefix, i));
return result;
}
nir_variable **create_many_ivec2(nir_variable_mode mode, const char *prefix, unsigned count) {
nir_variable **result = (nir_variable **)linear_alloc_child(lin_ctx, sizeof(nir_variable *) * count);
for (unsigned i = 0; i < count; i++)
result[i] = create_ivec2(mode, linear_asprintf(lin_ctx, "%s%u", prefix, i));
return result;
}
unsigned count_intrinsics(nir_intrinsic_op intrinsic);
nir_intrinsic_instr *find_next_intrinsic(nir_intrinsic_op intrinsic,
nir_intrinsic_instr *after);
void *mem_ctx;
void *lin_ctx;
nir_builder *b;
};
nir_vars_test::nir_vars_test()
{
mem_ctx = ralloc_context(NULL);
lin_ctx = linear_alloc_parent(mem_ctx, 0);
static const nir_shader_compiler_options options = { };
b = rzalloc(mem_ctx, nir_builder);
nir_builder_init_simple_shader(b, mem_ctx, MESA_SHADER_FRAGMENT, &options);
}
nir_vars_test::~nir_vars_test()
{
if (HasFailure()) {
printf("\nShader from the failed test:\n\n");
nir_print_shader(b->shader, stdout);
}
ralloc_free(mem_ctx);
}
unsigned
nir_vars_test::count_intrinsics(nir_intrinsic_op intrinsic)
{
unsigned count = 0;
nir_foreach_block(block, b->impl) {
nir_foreach_instr(instr, block) {
if (instr->type != nir_instr_type_intrinsic)
continue;
nir_intrinsic_instr *intrin = nir_instr_as_intrinsic(instr);
if (intrin->intrinsic == intrinsic)
count++;
}
}
return count;
}
nir_intrinsic_instr *
nir_vars_test::find_next_intrinsic(nir_intrinsic_op intrinsic,
nir_intrinsic_instr *after)
{
bool seen = after == NULL;
nir_foreach_block(block, b->impl) {
/* Skip blocks before the 'after' instruction. */
if (!seen && block != after->instr.block)
continue;
nir_foreach_instr(instr, block) {
if (instr->type != nir_instr_type_intrinsic)
continue;
nir_intrinsic_instr *intrin = nir_instr_as_intrinsic(instr);
if (!seen) {
seen = (after == intrin);
continue;
}
if (intrin->intrinsic == intrinsic)
return intrin;
}
}
return NULL;
}
/* Allow grouping the tests while still sharing the helpers. */
class nir_redundant_load_vars_test : public nir_vars_test {};
class nir_copy_prop_vars_test : public nir_vars_test {};
class nir_dead_write_vars_test : public nir_vars_test {};
} // namespace
TEST_F(nir_redundant_load_vars_test, duplicated_load)
{
/* Load a variable twice in the same block. One should be removed. */
nir_variable *in = create_int(nir_var_shader_in, "in");
nir_variable **out = create_many_int(nir_var_shader_out, "out", 2);
nir_store_var(b, out[0], nir_load_var(b, in), 1);
nir_store_var(b, out[1], nir_load_var(b, in), 1);
nir_validate_shader(b->shader, NULL);
ASSERT_EQ(count_intrinsics(nir_intrinsic_load_deref), 2);
bool progress = nir_opt_copy_prop_vars(b->shader);
EXPECT_TRUE(progress);
nir_validate_shader(b->shader, NULL);
ASSERT_EQ(count_intrinsics(nir_intrinsic_load_deref), 1);
}
TEST_F(nir_redundant_load_vars_test, duplicated_load_in_two_blocks)
{
/* Load a variable twice in different blocks. One should be removed. */
nir_variable *in = create_int(nir_var_shader_in, "in");
nir_variable **out = create_many_int(nir_var_shader_out, "out", 2);
nir_store_var(b, out[0], nir_load_var(b, in), 1);
/* Forces the stores to be in different blocks. */
nir_pop_if(b, nir_push_if(b, nir_imm_int(b, 0)));
nir_store_var(b, out[1], nir_load_var(b, in), 1);
nir_validate_shader(b->shader, NULL);
ASSERT_EQ(count_intrinsics(nir_intrinsic_load_deref), 2);
bool progress = nir_opt_copy_prop_vars(b->shader);
EXPECT_TRUE(progress);
nir_validate_shader(b->shader, NULL);
ASSERT_EQ(count_intrinsics(nir_intrinsic_load_deref), 1);
}
TEST_F(nir_redundant_load_vars_test, invalidate_inside_if_block)
{
/* Load variables, then write to some of then in different branches of the
* if statement. They should be invalidated accordingly.
*/
nir_variable **g = create_many_int(nir_var_global, "g", 3);
nir_variable **out = create_many_int(nir_var_shader_out, "out", 3);
nir_load_var(b, g[0]);
nir_load_var(b, g[1]);
nir_load_var(b, g[2]);
nir_if *if_stmt = nir_push_if(b, nir_imm_int(b, 0));
nir_store_var(b, g[0], nir_imm_int(b, 10), 1);
nir_push_else(b, if_stmt);
nir_store_var(b, g[1], nir_imm_int(b, 20), 1);
nir_pop_if(b, if_stmt);
nir_store_var(b, out[0], nir_load_var(b, g[0]), 1);
nir_store_var(b, out[1], nir_load_var(b, g[1]), 1);
nir_store_var(b, out[2], nir_load_var(b, g[2]), 1);
nir_validate_shader(b->shader, NULL);
bool progress = nir_opt_copy_prop_vars(b->shader);
EXPECT_TRUE(progress);
/* There are 3 initial loads, plus 2 loads for the values invalidated
* inside the if statement.
*/
ASSERT_EQ(count_intrinsics(nir_intrinsic_load_deref), 5);
/* We only load g[2] once. */
unsigned g2_load_count = 0;
nir_intrinsic_instr *load = NULL;
for (int i = 0; i < 5; i++) {
load = find_next_intrinsic(nir_intrinsic_load_deref, load);
if (nir_intrinsic_get_var(load, 0) == g[2])
g2_load_count++;
}
EXPECT_EQ(g2_load_count, 1);
}
TEST_F(nir_redundant_load_vars_test, invalidate_live_load_in_the_end_of_loop)
{
/* Invalidating a load in the end of loop body will apply to the whole loop
* body.
*/
nir_variable *v = create_int(nir_var_ssbo, "v");
nir_load_var(b, v);
nir_loop *loop = nir_push_loop(b);
nir_if *if_stmt = nir_push_if(b, nir_imm_int(b, 0));
nir_jump(b, nir_jump_break);
nir_pop_if(b, if_stmt);
nir_load_var(b, v);
nir_store_var(b, v, nir_imm_int(b, 10), 1);
nir_pop_loop(b, loop);
bool progress = nir_opt_copy_prop_vars(b->shader);
ASSERT_FALSE(progress);
}
TEST_F(nir_copy_prop_vars_test, simple_copies)
{
nir_variable *in = create_int(nir_var_shader_in, "in");
nir_variable *temp = create_int(nir_var_local, "temp");
nir_variable *out = create_int(nir_var_shader_out, "out");
nir_copy_var(b, temp, in);
nir_copy_var(b, out, temp);
nir_validate_shader(b->shader, NULL);
bool progress = nir_opt_copy_prop_vars(b->shader);
EXPECT_TRUE(progress);
nir_validate_shader(b->shader, NULL);
nir_intrinsic_instr *copy = NULL;
copy = find_next_intrinsic(nir_intrinsic_copy_deref, copy);
ASSERT_TRUE(copy->src[1].is_ssa);
nir_ssa_def *first_src = copy->src[1].ssa;
copy = find_next_intrinsic(nir_intrinsic_copy_deref, copy);
ASSERT_TRUE(copy->src[1].is_ssa);
EXPECT_EQ(copy->src[1].ssa, first_src);
}
TEST_F(nir_copy_prop_vars_test, simple_store_load)
{
nir_variable **v = create_many_ivec2(nir_var_local, "v", 2);
unsigned mask = 1 | 2;
nir_ssa_def *stored_value = nir_imm_ivec2(b, 10, 20);
nir_store_var(b, v[0], stored_value, mask);
nir_ssa_def *read_value = nir_load_var(b, v[0]);
nir_store_var(b, v[1], read_value, mask);
nir_validate_shader(b->shader, NULL);
bool progress = nir_opt_copy_prop_vars(b->shader);
EXPECT_TRUE(progress);
nir_validate_shader(b->shader, NULL);
ASSERT_EQ(count_intrinsics(nir_intrinsic_store_deref), 2);
nir_intrinsic_instr *store = NULL;
for (int i = 0; i < 2; i++) {
store = find_next_intrinsic(nir_intrinsic_store_deref, store);
ASSERT_TRUE(store->src[1].is_ssa);
EXPECT_EQ(store->src[1].ssa, stored_value);
}
}
TEST_F(nir_copy_prop_vars_test, store_store_load)
{
nir_variable **v = create_many_ivec2(nir_var_local, "v", 2);
unsigned mask = 1 | 2;
nir_ssa_def *first_value = nir_imm_ivec2(b, 10, 20);
nir_store_var(b, v[0], first_value, mask);
nir_ssa_def *second_value = nir_imm_ivec2(b, 30, 40);
nir_store_var(b, v[0], second_value, mask);
nir_ssa_def *read_value = nir_load_var(b, v[0]);
nir_store_var(b, v[1], read_value, mask);
nir_validate_shader(b->shader, NULL);
bool progress = nir_opt_copy_prop_vars(b->shader);
EXPECT_TRUE(progress);
nir_validate_shader(b->shader, NULL);
/* Store to v[1] should use second_value directly. */
nir_intrinsic_instr *store_to_v1 = NULL;
while ((store_to_v1 = find_next_intrinsic(nir_intrinsic_store_deref, store_to_v1)) != NULL) {
if (nir_intrinsic_get_var(store_to_v1, 0) == v[1]) {
ASSERT_TRUE(store_to_v1->src[1].is_ssa);
EXPECT_EQ(store_to_v1->src[1].ssa, second_value);
break;
}
}
EXPECT_TRUE(store_to_v1);
}
TEST_F(nir_copy_prop_vars_test, store_store_load_different_components)
{
nir_variable **v = create_many_ivec2(nir_var_local, "v", 2);
nir_ssa_def *first_value = nir_imm_ivec2(b, 10, 20);
nir_store_var(b, v[0], first_value, 1 << 1);
nir_ssa_def *second_value = nir_imm_ivec2(b, 30, 40);
nir_store_var(b, v[0], second_value, 1 << 0);
nir_ssa_def *read_value = nir_load_var(b, v[0]);
nir_store_var(b, v[1], read_value, 1 << 1);
nir_validate_shader(b->shader, NULL);
bool progress = nir_opt_copy_prop_vars(b->shader);
EXPECT_TRUE(progress);
nir_validate_shader(b->shader, NULL);
nir_opt_constant_folding(b->shader);
nir_validate_shader(b->shader, NULL);
/* Store to v[1] should use first_value directly. The write of
* second_value did not overwrite the component it uses.
*/
nir_intrinsic_instr *store_to_v1 = NULL;
while ((store_to_v1 = find_next_intrinsic(nir_intrinsic_store_deref, store_to_v1)) != NULL) {
if (nir_intrinsic_get_var(store_to_v1, 0) == v[1]) {
ASSERT_TRUE(store_to_v1->src[1].is_ssa);
ASSERT_TRUE(nir_src_is_const(store_to_v1->src[1]));
ASSERT_EQ(nir_src_comp_as_uint(store_to_v1->src[1], 1), 20);
break;
}
}
EXPECT_TRUE(store_to_v1);
}
TEST_F(nir_copy_prop_vars_test, store_store_load_different_components_in_many_blocks)
{
nir_variable **v = create_many_ivec2(nir_var_local, "v", 2);
nir_ssa_def *first_value = nir_imm_ivec2(b, 10, 20);
nir_store_var(b, v[0], first_value, 1 << 1);
/* Adding an if statement will cause blocks to be created. */
nir_pop_if(b, nir_push_if(b, nir_imm_int(b, 0)));
nir_ssa_def *second_value = nir_imm_ivec2(b, 30, 40);
nir_store_var(b, v[0], second_value, 1 << 0);
/* Adding an if statement will cause blocks to be created. */
nir_pop_if(b, nir_push_if(b, nir_imm_int(b, 0)));
nir_ssa_def *read_value = nir_load_var(b, v[0]);
nir_store_var(b, v[1], read_value, 1 << 1);
nir_validate_shader(b->shader, NULL);
nir_print_shader(b->shader, stdout);
bool progress = nir_opt_copy_prop_vars(b->shader);
EXPECT_TRUE(progress);
nir_print_shader(b->shader, stdout);
nir_validate_shader(b->shader, NULL);
nir_opt_constant_folding(b->shader);
nir_validate_shader(b->shader, NULL);
/* Store to v[1] should use first_value directly. The write of
* second_value did not overwrite the component it uses.
*/
nir_intrinsic_instr *store_to_v1 = NULL;
while ((store_to_v1 = find_next_intrinsic(nir_intrinsic_store_deref, store_to_v1)) != NULL) {
if (nir_intrinsic_get_var(store_to_v1, 0) == v[1]) {
ASSERT_TRUE(store_to_v1->src[1].is_ssa);
ASSERT_TRUE(nir_src_is_const(store_to_v1->src[1]));
ASSERT_EQ(nir_src_comp_as_uint(store_to_v1->src[1], 1), 20);
break;
}
}
EXPECT_TRUE(store_to_v1);
}
TEST_F(nir_copy_prop_vars_test, memory_barrier_in_two_blocks)
{
nir_variable **v = create_many_int(nir_var_ssbo, "v", 4);
nir_store_var(b, v[0], nir_imm_int(b, 1), 1);
nir_store_var(b, v[1], nir_imm_int(b, 2), 1);
/* Split into many blocks. */
nir_pop_if(b, nir_push_if(b, nir_imm_int(b, 0)));
nir_store_var(b, v[2], nir_load_var(b, v[0]), 1);
nir_builder_instr_insert(b, &nir_intrinsic_instr_create(b->shader, nir_intrinsic_memory_barrier)->instr);
nir_store_var(b, v[3], nir_load_var(b, v[1]), 1);
bool progress = nir_opt_copy_prop_vars(b->shader);
ASSERT_TRUE(progress);
/* Only the second load will remain after the optimization. */
ASSERT_EQ(1, count_intrinsics(nir_intrinsic_load_deref));
nir_intrinsic_instr *load = NULL;
load = find_next_intrinsic(nir_intrinsic_load_deref, load);
ASSERT_EQ(nir_intrinsic_get_var(load, 0), v[1]);
}
TEST_F(nir_copy_prop_vars_test, simple_store_load_in_two_blocks)
{
nir_variable **v = create_many_ivec2(nir_var_local, "v", 2);
unsigned mask = 1 | 2;
nir_ssa_def *stored_value = nir_imm_ivec2(b, 10, 20);
nir_store_var(b, v[0], stored_value, mask);
/* Adding an if statement will cause blocks to be created. */
nir_pop_if(b, nir_push_if(b, nir_imm_int(b, 0)));
nir_ssa_def *read_value = nir_load_var(b, v[0]);
nir_store_var(b, v[1], read_value, mask);
nir_validate_shader(b->shader, NULL);
bool progress = nir_opt_copy_prop_vars(b->shader);
EXPECT_TRUE(progress);
nir_validate_shader(b->shader, NULL);
ASSERT_EQ(count_intrinsics(nir_intrinsic_store_deref), 2);
nir_intrinsic_instr *store = NULL;
for (int i = 0; i < 2; i++) {
store = find_next_intrinsic(nir_intrinsic_store_deref, store);
ASSERT_TRUE(store->src[1].is_ssa);
EXPECT_EQ(store->src[1].ssa, stored_value);
}
}
TEST_F(nir_dead_write_vars_test, no_dead_writes_in_block)
{
nir_variable **v = create_many_int(nir_var_ssbo, "v", 2);
nir_store_var(b, v[0], nir_load_var(b, v[1]), 1);
bool progress = nir_opt_dead_write_vars(b->shader);
ASSERT_FALSE(progress);
}
TEST_F(nir_dead_write_vars_test, no_dead_writes_different_components_in_block)
{
nir_variable **v = create_many_ivec2(nir_var_ssbo, "v", 3);
nir_store_var(b, v[0], nir_load_var(b, v[1]), 1 << 0);
nir_store_var(b, v[0], nir_load_var(b, v[2]), 1 << 1);
bool progress = nir_opt_dead_write_vars(b->shader);
ASSERT_FALSE(progress);
}
TEST_F(nir_dead_write_vars_test, no_dead_writes_in_if_statement)
{
nir_variable **v = create_many_int(nir_var_ssbo, "v", 6);
nir_store_var(b, v[2], nir_load_var(b, v[0]), 1);
nir_store_var(b, v[3], nir_load_var(b, v[1]), 1);
/* Each arm of the if statement will overwrite one store. */
nir_if *if_stmt = nir_push_if(b, nir_imm_int(b, 0));
nir_store_var(b, v[2], nir_load_var(b, v[4]), 1);
nir_push_else(b, if_stmt);
nir_store_var(b, v[3], nir_load_var(b, v[5]), 1);
nir_pop_if(b, if_stmt);
bool progress = nir_opt_dead_write_vars(b->shader);
ASSERT_FALSE(progress);
}
TEST_F(nir_dead_write_vars_test, no_dead_writes_in_loop_statement)
{
nir_variable **v = create_many_int(nir_var_ssbo, "v", 3);
nir_store_var(b, v[0], nir_load_var(b, v[1]), 1);
/* Loop will write other value. Since it might not be executed, it doesn't
* kill the first write.
*/
nir_loop *loop = nir_push_loop(b);
nir_if *if_stmt = nir_push_if(b, nir_imm_int(b, 0));
nir_jump(b, nir_jump_break);
nir_pop_if(b, if_stmt);
nir_store_var(b, v[0], nir_load_var(b, v[2]), 1);
nir_pop_loop(b, loop);
bool progress = nir_opt_dead_write_vars(b->shader);
ASSERT_FALSE(progress);
}
TEST_F(nir_dead_write_vars_test, dead_write_in_block)
{
nir_variable **v = create_many_int(nir_var_ssbo, "v", 3);
nir_store_var(b, v[0], nir_load_var(b, v[1]), 1);
nir_ssa_def *load_v2 = nir_load_var(b, v[2]);
nir_store_var(b, v[0], load_v2, 1);
bool progress = nir_opt_dead_write_vars(b->shader);
ASSERT_TRUE(progress);
EXPECT_EQ(1, count_intrinsics(nir_intrinsic_store_deref));
nir_intrinsic_instr *store = find_next_intrinsic(nir_intrinsic_store_deref, NULL);
ASSERT_TRUE(store->src[1].is_ssa);
EXPECT_EQ(store->src[1].ssa, load_v2);
}
TEST_F(nir_dead_write_vars_test, dead_write_components_in_block)
{
nir_variable **v = create_many_ivec2(nir_var_ssbo, "v", 3);
nir_store_var(b, v[0], nir_load_var(b, v[1]), 1 << 0);
nir_ssa_def *load_v2 = nir_load_var(b, v[2]);
nir_store_var(b, v[0], load_v2, 1 << 0);
bool progress = nir_opt_dead_write_vars(b->shader);
ASSERT_TRUE(progress);
EXPECT_EQ(1, count_intrinsics(nir_intrinsic_store_deref));
nir_intrinsic_instr *store = find_next_intrinsic(nir_intrinsic_store_deref, NULL);
ASSERT_TRUE(store->src[1].is_ssa);
EXPECT_EQ(store->src[1].ssa, load_v2);
}
/* TODO: The DISABLED tests below depend on the dead write removal be able to
* identify dead writes between multiple blocks. This is still not
* implemented.
*/
TEST_F(nir_dead_write_vars_test, DISABLED_dead_write_in_two_blocks)
{
nir_variable **v = create_many_int(nir_var_ssbo, "v", 3);
nir_store_var(b, v[0], nir_load_var(b, v[1]), 1);
nir_ssa_def *load_v2 = nir_load_var(b, v[2]);
/* Causes the stores to be in different blocks. */
nir_pop_if(b, nir_push_if(b, nir_imm_int(b, 0)));
nir_store_var(b, v[0], load_v2, 1);
bool progress = nir_opt_dead_write_vars(b->shader);
ASSERT_TRUE(progress);
EXPECT_EQ(1, count_intrinsics(nir_intrinsic_store_deref));
nir_intrinsic_instr *store = find_next_intrinsic(nir_intrinsic_store_deref, NULL);
ASSERT_TRUE(store->src[1].is_ssa);
EXPECT_EQ(store->src[1].ssa, load_v2);
}
TEST_F(nir_dead_write_vars_test, DISABLED_dead_write_components_in_two_blocks)
{
nir_variable **v = create_many_ivec2(nir_var_ssbo, "v", 3);
nir_store_var(b, v[0], nir_load_var(b, v[1]), 1 << 0);
/* Causes the stores to be in different blocks. */
nir_pop_if(b, nir_push_if(b, nir_imm_int(b, 0)));
nir_ssa_def *load_v2 = nir_load_var(b, v[2]);
nir_store_var(b, v[0], load_v2, 1 << 0);
bool progress = nir_opt_dead_write_vars(b->shader);
ASSERT_TRUE(progress);
EXPECT_EQ(1, count_intrinsics(nir_intrinsic_store_deref));
nir_intrinsic_instr *store = find_next_intrinsic(nir_intrinsic_store_deref, NULL);
ASSERT_TRUE(store->src[1].is_ssa);
EXPECT_EQ(store->src[1].ssa, load_v2);
}
TEST_F(nir_dead_write_vars_test, DISABLED_dead_writes_in_if_statement)
{
nir_variable **v = create_many_int(nir_var_ssbo, "v", 4);
/* Both branches will overwrite, making the previous store dead. */
nir_store_var(b, v[0], nir_load_var(b, v[1]), 1);
nir_if *if_stmt = nir_push_if(b, nir_imm_int(b, 0));
nir_ssa_def *load_v2 = nir_load_var(b, v[2]);
nir_store_var(b, v[0], load_v2, 1);
nir_push_else(b, if_stmt);
nir_ssa_def *load_v3 = nir_load_var(b, v[3]);
nir_store_var(b, v[0], load_v3, 1);
nir_pop_if(b, if_stmt);
bool progress = nir_opt_dead_write_vars(b->shader);
ASSERT_TRUE(progress);
EXPECT_EQ(2, count_intrinsics(nir_intrinsic_store_deref));
nir_intrinsic_instr *store = NULL;
store = find_next_intrinsic(nir_intrinsic_store_deref, store);
ASSERT_TRUE(store->src[1].is_ssa);
EXPECT_EQ(store->src[1].ssa, load_v2);
store = find_next_intrinsic(nir_intrinsic_store_deref, store);
ASSERT_TRUE(store->src[1].is_ssa);
EXPECT_EQ(store->src[1].ssa, load_v3);
}
TEST_F(nir_dead_write_vars_test, DISABLED_memory_barrier_in_two_blocks)
{
nir_variable **v = create_many_int(nir_var_ssbo, "v", 2);
nir_store_var(b, v[0], nir_imm_int(b, 1), 1);
nir_store_var(b, v[1], nir_imm_int(b, 2), 1);
/* Split into many blocks. */
nir_pop_if(b, nir_push_if(b, nir_imm_int(b, 0)));
/* Because it is before the barrier, this will kill the previous store to that target. */
nir_store_var(b, v[0], nir_imm_int(b, 3), 1);
nir_builder_instr_insert(b, &nir_intrinsic_instr_create(b->shader, nir_intrinsic_memory_barrier)->instr);
nir_store_var(b, v[1], nir_imm_int(b, 4), 1);
bool progress = nir_opt_dead_write_vars(b->shader);
ASSERT_TRUE(progress);
EXPECT_EQ(3, count_intrinsics(nir_intrinsic_store_deref));
}
TEST_F(nir_dead_write_vars_test, DISABLED_unrelated_barrier_in_two_blocks)
{
nir_variable **v = create_many_int(nir_var_ssbo, "v", 3);
nir_variable *out = create_int(nir_var_shader_out, "out");
nir_store_var(b, out, nir_load_var(b, v[1]), 1);
nir_store_var(b, v[0], nir_load_var(b, v[1]), 1);
/* Split into many blocks. */
nir_pop_if(b, nir_push_if(b, nir_imm_int(b, 0)));
/* Emit vertex will ensure writes to output variables are considered used,
* but should not affect other types of variables. */
nir_builder_instr_insert(b, &nir_intrinsic_instr_create(b->shader, nir_intrinsic_emit_vertex)->instr);
nir_store_var(b, out, nir_load_var(b, v[2]), 1);
nir_store_var(b, v[0], nir_load_var(b, v[2]), 1);
bool progress = nir_opt_dead_write_vars(b->shader);
ASSERT_TRUE(progress);
/* Verify the first write to v[0] was removed. */
EXPECT_EQ(3, count_intrinsics(nir_intrinsic_store_deref));
nir_intrinsic_instr *store = NULL;
store = find_next_intrinsic(nir_intrinsic_store_deref, store);
EXPECT_EQ(nir_intrinsic_get_var(store, 0), out);
store = find_next_intrinsic(nir_intrinsic_store_deref, store);
EXPECT_EQ(nir_intrinsic_get_var(store, 0), out);
store = find_next_intrinsic(nir_intrinsic_store_deref, store);
EXPECT_EQ(nir_intrinsic_get_var(store, 0), v[0]);
}