| /* |
| * Copyright © 2019 Intel Corporation |
| * |
| * Permission is hereby granted, free of charge, to any person obtaining a |
| * copy of this software and associated documentation files (the "Software"), |
| * to deal in the Software without restriction, including without limitation |
| * the rights to use, copy, modify, merge, publish, distribute, sublicense, |
| * and/or sell copies of the Software, and to permit persons to whom the |
| * Software is furnished to do so, subject to the following conditions: |
| * |
| * The above copyright notice and this permission notice (including the next |
| * paragraph) shall be included in all copies or substantial portions of the |
| * Software. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
| * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
| * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL |
| * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
| * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING |
| * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER |
| * DEALINGS IN THE SOFTWARE. |
| */ |
| #include <gtest/gtest.h> |
| #include "nir.h" |
| #include "nir_builder.h" |
| |
| class comparison_pre_test : public ::testing::Test { |
| protected: |
| comparison_pre_test() |
| { |
| glsl_type_singleton_init_or_ref(); |
| |
| static const nir_shader_compiler_options options = { }; |
| nir_builder_init_simple_shader(&bld, NULL, MESA_SHADER_VERTEX, &options); |
| |
| v1 = nir_imm_vec4(&bld, -2.0, -1.0, 1.0, 2.0); |
| v2 = nir_imm_vec4(&bld, 2.0, 1.0, -1.0, -2.0); |
| v3 = nir_imm_vec4(&bld, 3.0, 4.0, 5.0, 6.0); |
| } |
| |
| ~comparison_pre_test() |
| { |
| ralloc_free(bld.shader); |
| glsl_type_singleton_decref(); |
| } |
| |
| struct nir_builder bld; |
| |
| nir_ssa_def *v1; |
| nir_ssa_def *v2; |
| nir_ssa_def *v3; |
| |
| const uint8_t xxxx[4] = { 0, 0, 0, 0 }; |
| const uint8_t wwww[4] = { 3, 3, 3, 3 }; |
| }; |
| |
| TEST_F(comparison_pre_test, a_lt_b_vs_neg_a_plus_b) |
| { |
| /* Before: |
| * |
| * vec4 32 ssa_0 = load_const (-2.0, -1.0, 1.0, 2.0) |
| * vec4 32 ssa_1 = load_const ( 2.0, 1.0, -1.0, -2.0) |
| * vec4 32 ssa_2 = load_const ( 3.0, 4.0, 5.0, 6.0) |
| * vec1 32 ssa_3 = load_const ( 1.0) |
| * vec4 32 ssa_4 = fadd ssa_0, ssa_2 |
| * vec1 32 ssa_5 = mov ssa_4.x |
| * vec1 1 ssa_6 = flt ssa_5, ssa_3 |
| * |
| * if ssa_6 { |
| * vec1 32 ssa_7 = fneg ssa_5 |
| * vec1 32 ssa_8 = fadd ssa_7, ssa_3 |
| * } else { |
| * } |
| * |
| * After: |
| * |
| * vec4 32 ssa_0 = load_const (-2.0, -1.0, 1.0, 2.0) |
| * vec4 32 ssa_1 = load_const ( 2.0, 1.0, -1.0, -2.0) |
| * vec4 32 ssa_2 = load_const ( 3.0, 4.0, 5.0, 6.0) |
| * vec1 32 ssa_3 = load_const ( 1.0) |
| * vec4 32 ssa_4 = fadd ssa_0, ssa_2 |
| * vec1 32 ssa_5 = mov ssa_4.x |
| * vec1 32 ssa_9 = fneg ssa_5 |
| * vec1 32 ssa_10 = fadd ssa_3, ssa_9 |
| * vec1 32 ssa_11 = load_const (0.0) |
| * vec1 1 ssa_12 = flt ssa_11, ssa_10 |
| * vec1 32 ssa_13 = mov ssa_10 |
| * vec1 1 ssa_14 = mov ssa_12 |
| * |
| * if ssa_14 { |
| * vec1 32 ssa_7 = fneg ssa_5 |
| * } else { |
| * } |
| */ |
| nir_ssa_def *one = nir_imm_float(&bld, 1.0f); |
| nir_ssa_def *a = nir_channel(&bld, nir_fadd(&bld, v1, v3), 0); |
| |
| nir_ssa_def *flt = nir_flt(&bld, a, one); |
| |
| nir_if *nif = nir_push_if(&bld, flt); |
| |
| nir_fadd(&bld, nir_fneg(&bld, a), one); |
| |
| nir_pop_if(&bld, nif); |
| |
| EXPECT_TRUE(nir_opt_comparison_pre_impl(bld.impl)); |
| } |
| |
| TEST_F(comparison_pre_test, a_lt_b_vs_a_minus_b) |
| { |
| /* Before: |
| * |
| * vec4 32 ssa_0 = load_const (-2.0, -1.0, 1.0, 2.0) |
| * vec4 32 ssa_1 = load_const ( 2.0, 1.0, -1.0, -2.0) |
| * vec4 32 ssa_2 = load_const ( 3.0, 4.0, 5.0, 6.0) |
| * vec1 32 ssa_3 = load_const ( 1.0) |
| * vec4 32 ssa_4 = fadd ssa_0, ssa_2 |
| * vec1 32 ssa_5 = mov ssa_4.x |
| * vec1 1 ssa_6 = flt ssa_3, ssa_5 |
| * |
| * if ssa_6 { |
| * vec1 32 ssa_7 = fneg ssa_5 |
| * vec1 32 ssa_8 = fadd ssa_3, ssa_7 |
| * } else { |
| * } |
| * |
| * After: |
| * |
| * vec4 32 ssa_0 = load_const (-2.0, -1.0, 1.0, 2.0) |
| * vec4 32 ssa_1 = load_const ( 2.0, 1.0, -1.0, -2.0) |
| * vec4 32 ssa_2 = load_const ( 3.0, 4.0, 5.0, 6.0) |
| * vec1 32 ssa_3 = load_const ( 1.0) |
| * vec4 32 ssa_4 = fadd ssa_0, ssa_2 |
| * vec1 32 ssa_5 = mov ssa_4.x |
| * vec1 32 ssa_9 = fneg ssa_5 |
| * vec1 32 ssa_10 = fadd ssa_3, ssa_9 |
| * vec1 32 ssa_11 = load_const (0.0) |
| * vec1 1 ssa_12 = flt ssa_10, ssa_11 |
| * vec1 32 ssa_13 = mov ssa_10 |
| * vec1 1 ssa_14 = mov ssa_12 |
| * |
| * if ssa_14 { |
| * vec1 32 ssa_7 = fneg ssa_5 |
| * } else { |
| * } |
| */ |
| nir_ssa_def *one = nir_imm_float(&bld, 1.0f); |
| nir_ssa_def *b = nir_channel(&bld, nir_fadd(&bld, v1, v3), 0); |
| |
| nir_ssa_def *flt = nir_flt(&bld, one, b); |
| |
| nir_if *nif = nir_push_if(&bld, flt); |
| |
| nir_fadd(&bld, one, nir_fneg(&bld, b)); |
| |
| nir_pop_if(&bld, nif); |
| |
| EXPECT_TRUE(nir_opt_comparison_pre_impl(bld.impl)); |
| } |
| |
| TEST_F(comparison_pre_test, neg_a_lt_b_vs_a_plus_b) |
| { |
| /* Before: |
| * |
| * vec4 32 ssa_0 = load_const (-2.0, -1.0, 1.0, 2.0) |
| * vec4 32 ssa_1 = load_const ( 2.0, 1.0, -1.0, -2.0) |
| * vec4 32 ssa_2 = load_const ( 3.0, 4.0, 5.0, 6.0) |
| * vec1 32 ssa_3 = load_const ( 1.0) |
| * vec4 32 ssa_4 = fadd ssa_0, ssa_2 |
| * vec1 32 ssa_5 = mov ssa_4.x |
| * vec1 32 ssa_6 = fneg ssa_5 |
| * vec1 1 ssa_7 = flt ssa_6, ssa_3 |
| * |
| * if ssa_7 { |
| * vec1 32 ssa_8 = fadd ssa_5, ssa_3 |
| * } else { |
| * } |
| * |
| * After: |
| * |
| * vec4 32 ssa_0 = load_const (-2.0, -1.0, 1.0, 2.0) |
| * vec4 32 ssa_1 = load_const ( 2.0, 1.0, -1.0, -2.0) |
| * vec4 32 ssa_2 = load_const ( 3.0, 4.0, 5.0, 6.0) |
| * vec1 32 ssa_3 = load_const ( 1.0) |
| * vec4 32 ssa_4 = fadd ssa_0, ssa_2 |
| * vec1 32 ssa_5 = mov ssa_4.x |
| * vec1 32 ssa_9 = fneg ssa_5 |
| * vec1 32 ssa_9 = fneg ssa_6 |
| * vec1 32 ssa_10 = fadd ssa_3, ssa_9 |
| * vec1 32 ssa_11 = load_const ( 0.0) |
| * vec1 1 ssa_12 = flt ssa_11, ssa_10 |
| * vec1 32 ssa_13 = mov ssa_10 |
| * vec1 1 ssa_14 = mov ssa_12 |
| * |
| * if ssa_14 { |
| * } else { |
| * } |
| */ |
| |
| nir_ssa_def *one = nir_imm_float(&bld, 1.0f); |
| nir_ssa_def *a = nir_channel(&bld, nir_fadd(&bld, v1, v3), 0); |
| |
| nir_ssa_def *flt = nir_flt(&bld, nir_fneg(&bld, a), one); |
| |
| nir_if *nif = nir_push_if(&bld, flt); |
| |
| nir_fadd(&bld, a, one); |
| |
| nir_pop_if(&bld, nif); |
| |
| EXPECT_TRUE(nir_opt_comparison_pre_impl(bld.impl)); |
| } |
| |
| TEST_F(comparison_pre_test, a_lt_neg_b_vs_a_plus_b) |
| { |
| /* Before: |
| * |
| * vec4 32 ssa_0 = load_const (-2.0, -1.0, 1.0, 2.0) |
| * vec4 32 ssa_1 = load_const ( 2.0, 1.0, -1.0, -2.0) |
| * vec4 32 ssa_2 = load_const ( 3.0, 4.0, 5.0, 6.0) |
| * vec1 32 ssa_3 = load_const ( 1.0) |
| * vec4 32 ssa_4 = fadd ssa_0, ssa_2 |
| * vec1 32 ssa_5 = mov ssa_4.x |
| * vec1 32 ssa_6 = fneg ssa_5 |
| * vec1 1 ssa_7 = flt ssa_3, ssa_6 |
| * |
| * if ssa_7 { |
| * vec1 32 ssa_8 = fadd ssa_3, ssa_5 |
| * } else { |
| * } |
| * |
| * After: |
| * |
| * vec4 32 ssa_0 = load_const (-2.0, -1.0, 1.0, 2.0) |
| * vec4 32 ssa_1 = load_const ( 2.0, 1.0, -1.0, -2.0) |
| * vec4 32 ssa_2 = load_const ( 3.0, 4.0, 5.0, 6.0) |
| * vec1 32 ssa_3 = load_const ( 1.0) |
| * vec4 32 ssa_4 = fadd ssa_0, ssa_2 |
| * vec1 32 ssa_5 = mov ssa_4.x |
| * vec1 32 ssa_9 = fneg ssa_5 |
| * vec1 32 ssa_9 = fneg ssa_6 |
| * vec1 32 ssa_10 = fadd ssa_3, ssa_9 |
| * vec1 32 ssa_11 = load_const ( 0.0) |
| * vec1 1 ssa_12 = flt ssa_10, ssa_11 |
| * vec1 32 ssa_13 = mov ssa_10 |
| * vec1 1 ssa_14 = mov ssa_12 |
| * |
| * if ssa_14 { |
| * } else { |
| * } |
| */ |
| nir_ssa_def *one = nir_imm_float(&bld, 1.0f); |
| nir_ssa_def *b = nir_channel(&bld, nir_fadd(&bld, v1, v3), 0); |
| |
| nir_ssa_def *flt = nir_flt(&bld, one, nir_fneg(&bld, b)); |
| |
| nir_if *nif = nir_push_if(&bld, flt); |
| |
| nir_fadd(&bld, one, b); |
| |
| nir_pop_if(&bld, nif); |
| |
| EXPECT_TRUE(nir_opt_comparison_pre_impl(bld.impl)); |
| } |
| |
| TEST_F(comparison_pre_test, imm_lt_b_vs_neg_imm_plus_b) |
| { |
| /* Before: |
| * |
| * vec4 32 ssa_0 = load_const (-2.0, -1.0, 1.0, 2.0) |
| * vec4 32 ssa_1 = load_const ( 2.0, 1.0, -1.0, -2.0) |
| * vec4 32 ssa_2 = load_const ( 3.0, 4.0, 5.0, 6.0) |
| * vec1 32 ssa_3 = load_const ( 1.0) |
| * vec1 32 ssa_4 = load_const (-1.0) |
| * vec4 32 ssa_5 = fadd ssa_0, ssa_2 |
| * vec1 32 ssa_6 = mov ssa_5.x |
| * vec1 1 ssa_7 = flt ssa_3, ssa_6 |
| * |
| * if ssa_7 { |
| * vec1 32 ssa_8 = fadd ssa_4, ssa_6 |
| * } else { |
| * } |
| * |
| * After: |
| * |
| * vec4 32 ssa_0 = load_const (-2.0, -1.0, 1.0, 2.0) |
| * vec4 32 ssa_1 = load_const ( 2.0, 1.0, -1.0, -2.0) |
| * vec4 32 ssa_2 = load_const ( 3.0, 4.0, 5.0, 6.0) |
| * vec1 32 ssa_3 = load_const ( 1.0) |
| * vec1 32 ssa_4 = load_const (-1.0) |
| * vec4 32 ssa_5 = fadd ssa_0, ssa_2 |
| * vec1 32 ssa_6 = mov ssa_5.x |
| * vec1 32 ssa_9 = fneg ssa_3 |
| * vec1 32 ssa_10 = fadd ssa_6, ssa_9 |
| * vec1 32 ssa_11 = load_const ( 0.0) |
| * vec1 1 ssa_12 = flt ssa_11, ssa_10 |
| * vec1 32 ssa_13 = mov ssa_10 |
| * vec1 1 ssa_14 = mov ssa_12 |
| * |
| * if ssa_14 { |
| * } else { |
| * } |
| */ |
| nir_ssa_def *one = nir_imm_float(&bld, 1.0f); |
| nir_ssa_def *neg_one = nir_imm_float(&bld, -1.0f); |
| nir_ssa_def *a = nir_channel(&bld, nir_fadd(&bld, v1, v3), 0); |
| |
| nir_ssa_def *flt = nir_flt(&bld, one, a); |
| |
| nir_if *nif = nir_push_if(&bld, flt); |
| |
| nir_fadd(&bld, neg_one, a); |
| |
| nir_pop_if(&bld, nif); |
| |
| EXPECT_TRUE(nir_opt_comparison_pre_impl(bld.impl)); |
| } |
| |
| TEST_F(comparison_pre_test, a_lt_imm_vs_a_minus_imm) |
| { |
| /* Before: |
| * |
| * vec4 32 ssa_0 = load_const (-2.0, -1.0, 1.0, 2.0) |
| * vec4 32 ssa_1 = load_const ( 2.0, 1.0, -1.0, -2.0) |
| * vec4 32 ssa_2 = load_const ( 3.0, 4.0, 5.0, 6.0) |
| * vec1 32 ssa_3 = load_const ( 1.0) |
| * vec1 32 ssa_4 = load_const (-1.0) |
| * vec4 32 ssa_5 = fadd ssa_0, ssa_2 |
| * vec1 32 ssa_6 = mov ssa_5.x |
| * vec1 1 ssa_7 = flt ssa_6, ssa_3 |
| * |
| * if ssa_6 { |
| * vec1 32 ssa_8 = fadd ssa_6, ssa_4 |
| * } else { |
| * } |
| * |
| * After: |
| * |
| * vec4 32 ssa_0 = load_const (-2.0, -1.0, 1.0, 2.0) |
| * vec4 32 ssa_1 = load_const ( 2.0, 1.0, -1.0, -2.0) |
| * vec4 32 ssa_2 = load_const ( 3.0, 4.0, 5.0, 6.0) |
| * vec1 32 ssa_3 = load_const ( 1.0) |
| * vec1 32 ssa_4 = load_const (-1.0) |
| * vec4 32 ssa_5 = fadd ssa_0, ssa_2 |
| * vec1 32 ssa_6 = mov ssa_5.x |
| * vec1 32 ssa_9 = fneg ssa_3 |
| * vec1 32 ssa_10 = fadd ssa_6, ssa_9 |
| * vec1 32 ssa_11 = load_const ( 0.0) |
| * vec1 1 ssa_12 = flt ssa_10, ssa_11 |
| * vec1 32 ssa_13 = mov ssa_10 |
| * vec1 1 ssa_14 = mov ssa_12 |
| * |
| * if ssa_14 { |
| * } else { |
| * } |
| */ |
| nir_ssa_def *one = nir_imm_float(&bld, 1.0f); |
| nir_ssa_def *neg_one = nir_imm_float(&bld, -1.0f); |
| nir_ssa_def *a = nir_channel(&bld, nir_fadd(&bld, v1, v3), 0); |
| |
| nir_ssa_def *flt = nir_flt(&bld, a, one); |
| |
| nir_if *nif = nir_push_if(&bld, flt); |
| |
| nir_fadd(&bld, a, neg_one); |
| |
| nir_pop_if(&bld, nif); |
| |
| EXPECT_TRUE(nir_opt_comparison_pre_impl(bld.impl)); |
| } |
| |
| TEST_F(comparison_pre_test, neg_imm_lt_a_vs_a_plus_imm) |
| { |
| /* Before: |
| * |
| * vec4 32 ssa_0 = load_const (-2.0, -1.0, 1.0, 2.0) |
| * vec4 32 ssa_1 = load_const ( 2.0, 1.0, -1.0, -2.0) |
| * vec4 32 ssa_2 = load_const ( 3.0, 4.0, 5.0, 6.0) |
| * vec1 32 ssa_3 = load_const ( 1.0) |
| * vec1 32 ssa_4 = load_const (-1.0) |
| * vec4 32 ssa_5 = fadd ssa_0, ssa_2 |
| * vec1 32 ssa_6 = mov ssa_5.x |
| * vec1 1 ssa_7 = flt ssa_4, ssa_6 |
| * |
| * if ssa_7 { |
| * vec1 32 ssa_8 = fadd ssa_6, ssa_3 |
| * } else { |
| * } |
| * |
| * After: |
| * |
| * vec4 32 ssa_0 = load_const (-2.0, -1.0, 1.0, 2.0) |
| * vec4 32 ssa_1 = load_const ( 2.0, 1.0, -1.0, -2.0) |
| * vec4 32 ssa_2 = load_const ( 3.0, 4.0, 5.0, 6.0) |
| * vec1 32 ssa_3 = load_const ( 1.0) |
| * vec1 32 ssa_4 = load_const (-1.0) |
| * vec4 32 ssa_5 = fadd ssa_0, ssa_2 |
| * vec1 32 ssa_6 = mov ssa_5.x |
| * vec1 32 ssa_9 = fneg ssa_4 |
| * vec1 32 ssa_10 = fadd ssa_6, ssa_9 |
| * vec1 32 ssa_11 = load_const ( 0.0) |
| * vec1 1 ssa_12 = flt ssa_11, ssa_10 |
| * vec1 32 ssa_13 = mov ssa_10 |
| * vec1 1 ssa_14 = mov ssa_12 |
| * |
| * if ssa_14 { |
| * } else { |
| * } |
| */ |
| |
| nir_ssa_def *one = nir_imm_float(&bld, 1.0f); |
| nir_ssa_def *neg_one = nir_imm_float(&bld, -1.0f); |
| nir_ssa_def *a = nir_channel(&bld, nir_fadd(&bld, v1, v3), 0); |
| |
| nir_ssa_def *flt = nir_flt(&bld, neg_one, a); |
| |
| nir_if *nif = nir_push_if(&bld, flt); |
| |
| nir_fadd(&bld, a, one); |
| |
| nir_pop_if(&bld, nif); |
| |
| EXPECT_TRUE(nir_opt_comparison_pre_impl(bld.impl)); |
| } |
| |
| TEST_F(comparison_pre_test, a_lt_neg_imm_vs_a_plus_imm) |
| { |
| /* Before: |
| * |
| * vec4 32 ssa_0 = load_const (-2.0, -1.0, 1.0, 2.0) |
| * vec4 32 ssa_1 = load_const ( 2.0, 1.0, -1.0, -2.0) |
| * vec4 32 ssa_2 = load_const ( 3.0, 4.0, 5.0, 6.0) |
| * vec1 32 ssa_3 = load_const ( 1.0) |
| * vec1 32 ssa_4 = load_const (-1.0) |
| * vec4 32 ssa_5 = fadd ssa_0, ssa_2 |
| * vec1 32 ssa_6 = mov ssa_5.x |
| * vec1 1 ssa_7 = flt ssa_6, ssa_4 |
| * |
| * if ssa_7 { |
| * vec1 32 ssa_8 = fadd ssa_6, ssa_3 |
| * } else { |
| * } |
| * |
| * After: |
| * |
| * vec4 32 ssa_0 = load_const (-2.0, -1.0, 1.0, 2.0) |
| * vec4 32 ssa_1 = load_const ( 2.0, 1.0, -1.0, -2.0) |
| * vec4 32 ssa_2 = load_const ( 3.0, 4.0, 5.0, 6.0) |
| * vec1 32 ssa_3 = load_const ( 1.0) |
| * vec1 32 ssa_4 = load_const (-1.0) |
| * vec4 32 ssa_5 = fadd ssa_0, ssa_2 |
| * vec1 32 ssa_6 = mov ssa_5.x |
| * vec1 32 ssa_9 = fneg ssa_4 |
| * vec1 32 ssa_10 = fadd ssa_6, ssa_9 |
| * vec1 32 ssa_11 = load_const ( 0.0) |
| * vec1 1 ssa_12 = flt ssa_10, ssa_11 |
| * vec1 32 ssa_13 = mov ssa_10 |
| * vec1 1 ssa_14 = mov ssa_12 |
| * |
| * if ssa_14 { |
| * } else { |
| * } |
| */ |
| nir_ssa_def *one = nir_imm_float(&bld, 1.0f); |
| nir_ssa_def *neg_one = nir_imm_float(&bld, -1.0f); |
| nir_ssa_def *a = nir_channel(&bld, nir_fadd(&bld, v1, v3), 0); |
| |
| nir_ssa_def *flt = nir_flt(&bld, a, neg_one); |
| |
| nir_if *nif = nir_push_if(&bld, flt); |
| |
| nir_fadd(&bld, a, one); |
| |
| nir_pop_if(&bld, nif); |
| |
| EXPECT_TRUE(nir_opt_comparison_pre_impl(bld.impl)); |
| } |
| |
| TEST_F(comparison_pre_test, swizzle_of_same_immediate_vector) |
| { |
| /* Before: |
| * |
| * vec4 32 ssa_0 = load_const (-2.0, -1.0, 1.0, 2.0) |
| * vec4 32 ssa_1 = load_const ( 2.0, 1.0, -1.0, -2.0) |
| * vec4 32 ssa_2 = load_const ( 3.0, 4.0, 5.0, 6.0) |
| * vec4 32 ssa_3 = fadd ssa_0, ssa_2 |
| * vec1 1 ssa_4 = flt ssa_0.x, ssa_3.x |
| * |
| * if ssa_4 { |
| * vec1 32 ssa_5 = fadd ssa_0.w, ssa_3.x |
| * } else { |
| * } |
| */ |
| nir_ssa_def *a = nir_fadd(&bld, v1, v3); |
| |
| nir_alu_instr *flt = nir_alu_instr_create(bld.shader, nir_op_flt); |
| |
| flt->src[0].src = nir_src_for_ssa(v1); |
| flt->src[1].src = nir_src_for_ssa(a); |
| |
| memcpy(&flt->src[0].swizzle, xxxx, sizeof(xxxx)); |
| memcpy(&flt->src[1].swizzle, xxxx, sizeof(xxxx)); |
| |
| nir_builder_alu_instr_finish_and_insert(&bld, flt); |
| |
| flt->dest.dest.ssa.num_components = 1; |
| flt->dest.write_mask = 1; |
| |
| nir_if *nif = nir_push_if(&bld, &flt->dest.dest.ssa); |
| |
| nir_alu_instr *fadd = nir_alu_instr_create(bld.shader, nir_op_fadd); |
| |
| fadd->src[0].src = nir_src_for_ssa(v1); |
| fadd->src[1].src = nir_src_for_ssa(a); |
| |
| memcpy(&fadd->src[0].swizzle, wwww, sizeof(wwww)); |
| memcpy(&fadd->src[1].swizzle, xxxx, sizeof(xxxx)); |
| |
| nir_builder_alu_instr_finish_and_insert(&bld, fadd); |
| |
| fadd->dest.dest.ssa.num_components = 1; |
| fadd->dest.write_mask = 1; |
| |
| nir_pop_if(&bld, nif); |
| |
| EXPECT_TRUE(nir_opt_comparison_pre_impl(bld.impl)); |
| } |
| |
| TEST_F(comparison_pre_test, non_scalar_add_result) |
| { |
| /* The optimization pass should not do anything because the result of the |
| * fadd is not a scalar. |
| * |
| * Before: |
| * |
| * vec4 32 ssa_0 = load_const (-2.0, -1.0, 1.0, 2.0) |
| * vec4 32 ssa_1 = load_const ( 2.0, 1.0, -1.0, -2.0) |
| * vec4 32 ssa_2 = load_const ( 3.0, 4.0, 5.0, 6.0) |
| * vec4 32 ssa_3 = fadd ssa_0, ssa_2 |
| * vec1 1 ssa_4 = flt ssa_0.x, ssa_3.x |
| * |
| * if ssa_4 { |
| * vec2 32 ssa_5 = fadd ssa_1.xx, ssa_3.xx |
| * } else { |
| * } |
| * |
| * After: |
| * |
| * No change. |
| */ |
| nir_ssa_def *a = nir_fadd(&bld, v1, v3); |
| |
| nir_alu_instr *flt = nir_alu_instr_create(bld.shader, nir_op_flt); |
| |
| flt->src[0].src = nir_src_for_ssa(v1); |
| flt->src[1].src = nir_src_for_ssa(a); |
| |
| memcpy(&flt->src[0].swizzle, xxxx, sizeof(xxxx)); |
| memcpy(&flt->src[1].swizzle, xxxx, sizeof(xxxx)); |
| |
| nir_builder_alu_instr_finish_and_insert(&bld, flt); |
| |
| flt->dest.dest.ssa.num_components = 1; |
| flt->dest.write_mask = 1; |
| |
| nir_if *nif = nir_push_if(&bld, &flt->dest.dest.ssa); |
| |
| nir_alu_instr *fadd = nir_alu_instr_create(bld.shader, nir_op_fadd); |
| |
| fadd->src[0].src = nir_src_for_ssa(v2); |
| fadd->src[1].src = nir_src_for_ssa(a); |
| |
| memcpy(&fadd->src[0].swizzle, xxxx, sizeof(xxxx)); |
| memcpy(&fadd->src[1].swizzle, xxxx, sizeof(xxxx)); |
| |
| nir_builder_alu_instr_finish_and_insert(&bld, fadd); |
| |
| fadd->dest.dest.ssa.num_components = 2; |
| fadd->dest.write_mask = 3; |
| |
| nir_pop_if(&bld, nif); |
| |
| EXPECT_FALSE(nir_opt_comparison_pre_impl(bld.impl)); |
| } |