vect: Fix single def-use cycle for ifn reductions [PR108608]
Checks
Commit Message
The patch that added support for fmin/fmax reductions didn't
handle single def-use cycles. In some ways, this seems like
going out of our way to make things slower, but that's a
discussion for another day.
Tested on aarch64-linux-gnu & x86_64-linux-gnu. OK for trunk
and the GCC 12 branch?
Richard
gcc/
PR tree-optimization/108608
* tree-vect-loop.cc (vect_transform_reduction): Handle single
def-use cycles that involve function calls rather than tree codes.
gcc/testsuite/
PR tree-optimization/108608
* gcc.dg/vect/pr108608.c: New test.
* gcc.target/aarch64/sve/pr108608-1.c: Likewise.
---
gcc/testsuite/gcc.dg/vect/pr108608.c | 24 +++++++++++++++++++
.../gcc.target/aarch64/sve/pr108608-1.c | 9 +++++++
gcc/tree-vect-loop.cc | 22 ++++++++++-------
3 files changed, 46 insertions(+), 9 deletions(-)
create mode 100644 gcc/testsuite/gcc.dg/vect/pr108608.c
create mode 100644 gcc/testsuite/gcc.target/aarch64/sve/pr108608-1.c
Comments
On Tue, Jan 31, 2023 at 2:08 PM Richard Sandiford via Gcc-patches
<gcc-patches@gcc.gnu.org> wrote:
>
> The patch that added support for fmin/fmax reductions didn't
> handle single def-use cycles. In some ways, this seems like
> going out of our way to make things slower, but that's a
> discussion for another day.
>
> Tested on aarch64-linux-gnu & x86_64-linux-gnu. OK for trunk
> and the GCC 12 branch?
OK.
Richard.
> Richard
>
>
> gcc/
> PR tree-optimization/108608
> * tree-vect-loop.cc (vect_transform_reduction): Handle single
> def-use cycles that involve function calls rather than tree codes.
>
> gcc/testsuite/
> PR tree-optimization/108608
> * gcc.dg/vect/pr108608.c: New test.
> * gcc.target/aarch64/sve/pr108608-1.c: Likewise.
> ---
> gcc/testsuite/gcc.dg/vect/pr108608.c | 24 +++++++++++++++++++
> .../gcc.target/aarch64/sve/pr108608-1.c | 9 +++++++
> gcc/tree-vect-loop.cc | 22 ++++++++++-------
> 3 files changed, 46 insertions(+), 9 deletions(-)
> create mode 100644 gcc/testsuite/gcc.dg/vect/pr108608.c
> create mode 100644 gcc/testsuite/gcc.target/aarch64/sve/pr108608-1.c
>
> diff --git a/gcc/testsuite/gcc.dg/vect/pr108608.c b/gcc/testsuite/gcc.dg/vect/pr108608.c
> new file mode 100644
> index 00000000000..e968141ba03
> --- /dev/null
> +++ b/gcc/testsuite/gcc.dg/vect/pr108608.c
> @@ -0,0 +1,24 @@
> +#include "tree-vect.h"
> +
> +double __attribute__((noipa))
> +foo (double m, float *ptr)
> +{
> + for (int i = 0; i < 256; i++)
> + m = __builtin_fmax (m, ptr[i]);
> + return m;
> +}
> +
> +int
> +main (void)
> +{
> + check_vect ();
> + float ptr[256];
> + for (int j = 0; j < 16; ++j)
> + {
> + for (int i = 0; i < 256; ++i)
> + ptr[i] = i == 128 + j ? 2 + j : i == 161 ? 1 : 0;
> + if (foo (0, ptr) != 2 + j)
> + __builtin_abort ();
> + }
> + return 0;
> +}
> diff --git a/gcc/testsuite/gcc.target/aarch64/sve/pr108608-1.c b/gcc/testsuite/gcc.target/aarch64/sve/pr108608-1.c
> new file mode 100644
> index 00000000000..0a7d485e047
> --- /dev/null
> +++ b/gcc/testsuite/gcc.target/aarch64/sve/pr108608-1.c
> @@ -0,0 +1,9 @@
> +/* { dg-options "-O3" } */
> +
> +double __attribute__((noipa))
> +foo (double m, float *ptr)
> +{
> + for (int i = 0; i < 256; i++)
> + m = __builtin_fmax (m, ptr[i]);
> + return m;
> +}
> diff --git a/gcc/tree-vect-loop.cc b/gcc/tree-vect-loop.cc
> index f0801c23671..f03af1efd0f 100644
> --- a/gcc/tree-vect-loop.cc
> +++ b/gcc/tree-vect-loop.cc
> @@ -7755,8 +7755,6 @@ vect_transform_reduction (loop_vec_info loop_vinfo,
> gimple_match_op op;
> if (!gimple_extract_op (stmt_info->stmt, &op))
> gcc_unreachable ();
> - gcc_assert (op.code.is_tree_code ());
> - auto code = tree_code (op.code);
>
> /* All uses but the last are expected to be defined in the loop.
> The last use is the reduction variable. In case of nested cycle this
> @@ -7778,7 +7776,8 @@ vect_transform_reduction (loop_vec_info loop_vinfo,
> vec_num = 1;
> }
>
> - internal_fn cond_fn = get_conditional_internal_fn (code);
> + code_helper code = canonicalize_code (op.code, op.type);
> + internal_fn cond_fn = get_conditional_internal_fn (code, op.type);
> vec_loop_masks *masks = &LOOP_VINFO_MASKS (loop_vinfo);
> bool mask_by_cond_expr = use_mask_by_cond_expr_p (code, cond_fn, vectype_in);
>
> @@ -7802,9 +7801,10 @@ vect_transform_reduction (loop_vec_info loop_vinfo,
> if (reduction_type == FOLD_LEFT_REDUCTION)
> {
> internal_fn reduc_fn = STMT_VINFO_REDUC_FN (reduc_info);
> + gcc_assert (code.is_tree_code ());
> return vectorize_fold_left_reduction
> - (loop_vinfo, stmt_info, gsi, vec_stmt, slp_node, reduc_def_phi, code,
> - reduc_fn, op.ops, vectype_in, reduc_index, masks);
> + (loop_vinfo, stmt_info, gsi, vec_stmt, slp_node, reduc_def_phi,
> + tree_code (code), reduc_fn, op.ops, vectype_in, reduc_index, masks);
> }
>
> bool single_defuse_cycle = STMT_VINFO_FORCE_SINGLE_CYCLE (reduc_info);
> @@ -7814,7 +7814,7 @@ vect_transform_reduction (loop_vec_info loop_vinfo,
> || code == SAD_EXPR);
>
> /* Create the destination vector */
> - tree scalar_dest = gimple_assign_lhs (stmt_info->stmt);
> + tree scalar_dest = gimple_get_lhs (stmt_info->stmt);
> tree vec_dest = vect_create_destination_var (scalar_dest, vectype_out);
>
> vect_get_vec_defs (loop_vinfo, stmt_info, slp_node, ncopies,
> @@ -7849,7 +7849,7 @@ vect_transform_reduction (loop_vec_info loop_vinfo,
> /* Make sure that the reduction accumulator is vop[0]. */
> if (reduc_index == 1)
> {
> - gcc_assert (commutative_tree_code (code));
> + gcc_assert (commutative_binary_op_p (code, op.type));
> std::swap (vop[0], vop[1]);
> }
> tree mask = vect_get_loop_mask (gsi, masks, vec_num * ncopies,
> @@ -7877,11 +7877,15 @@ vect_transform_reduction (loop_vec_info loop_vinfo,
> if (emulated_mixed_dot_prod)
> new_stmt = vect_emulate_mixed_dot_prod (loop_vinfo, stmt_info, gsi,
> vec_dest, vop);
> + else if (code.is_internal_fn ())
> + new_stmt = gimple_build_call_internal (internal_fn (code),
> + op.num_ops,
> + vop[0], vop[1], vop[2]);
> else
> - new_stmt = gimple_build_assign (vec_dest, code,
> + new_stmt = gimple_build_assign (vec_dest, tree_code (op.code),
> vop[0], vop[1], vop[2]);
> new_temp = make_ssa_name (vec_dest, new_stmt);
> - gimple_assign_set_lhs (new_stmt, new_temp);
> + gimple_set_lhs (new_stmt, new_temp);
> vect_finish_stmt_generation (loop_vinfo, stmt_info, new_stmt, gsi);
> }
>
> --
> 2.25.1
>
new file mode 100644
@@ -0,0 +1,24 @@
+#include "tree-vect.h"
+
+double __attribute__((noipa))
+foo (double m, float *ptr)
+{
+ for (int i = 0; i < 256; i++)
+ m = __builtin_fmax (m, ptr[i]);
+ return m;
+}
+
+int
+main (void)
+{
+ check_vect ();
+ float ptr[256];
+ for (int j = 0; j < 16; ++j)
+ {
+ for (int i = 0; i < 256; ++i)
+ ptr[i] = i == 128 + j ? 2 + j : i == 161 ? 1 : 0;
+ if (foo (0, ptr) != 2 + j)
+ __builtin_abort ();
+ }
+ return 0;
+}
new file mode 100644
@@ -0,0 +1,9 @@
+/* { dg-options "-O3" } */
+
+double __attribute__((noipa))
+foo (double m, float *ptr)
+{
+ for (int i = 0; i < 256; i++)
+ m = __builtin_fmax (m, ptr[i]);
+ return m;
+}
@@ -7755,8 +7755,6 @@ vect_transform_reduction (loop_vec_info loop_vinfo,
gimple_match_op op;
if (!gimple_extract_op (stmt_info->stmt, &op))
gcc_unreachable ();
- gcc_assert (op.code.is_tree_code ());
- auto code = tree_code (op.code);
/* All uses but the last are expected to be defined in the loop.
The last use is the reduction variable. In case of nested cycle this
@@ -7778,7 +7776,8 @@ vect_transform_reduction (loop_vec_info loop_vinfo,
vec_num = 1;
}
- internal_fn cond_fn = get_conditional_internal_fn (code);
+ code_helper code = canonicalize_code (op.code, op.type);
+ internal_fn cond_fn = get_conditional_internal_fn (code, op.type);
vec_loop_masks *masks = &LOOP_VINFO_MASKS (loop_vinfo);
bool mask_by_cond_expr = use_mask_by_cond_expr_p (code, cond_fn, vectype_in);
@@ -7802,9 +7801,10 @@ vect_transform_reduction (loop_vec_info loop_vinfo,
if (reduction_type == FOLD_LEFT_REDUCTION)
{
internal_fn reduc_fn = STMT_VINFO_REDUC_FN (reduc_info);
+ gcc_assert (code.is_tree_code ());
return vectorize_fold_left_reduction
- (loop_vinfo, stmt_info, gsi, vec_stmt, slp_node, reduc_def_phi, code,
- reduc_fn, op.ops, vectype_in, reduc_index, masks);
+ (loop_vinfo, stmt_info, gsi, vec_stmt, slp_node, reduc_def_phi,
+ tree_code (code), reduc_fn, op.ops, vectype_in, reduc_index, masks);
}
bool single_defuse_cycle = STMT_VINFO_FORCE_SINGLE_CYCLE (reduc_info);
@@ -7814,7 +7814,7 @@ vect_transform_reduction (loop_vec_info loop_vinfo,
|| code == SAD_EXPR);
/* Create the destination vector */
- tree scalar_dest = gimple_assign_lhs (stmt_info->stmt);
+ tree scalar_dest = gimple_get_lhs (stmt_info->stmt);
tree vec_dest = vect_create_destination_var (scalar_dest, vectype_out);
vect_get_vec_defs (loop_vinfo, stmt_info, slp_node, ncopies,
@@ -7849,7 +7849,7 @@ vect_transform_reduction (loop_vec_info loop_vinfo,
/* Make sure that the reduction accumulator is vop[0]. */
if (reduc_index == 1)
{
- gcc_assert (commutative_tree_code (code));
+ gcc_assert (commutative_binary_op_p (code, op.type));
std::swap (vop[0], vop[1]);
}
tree mask = vect_get_loop_mask (gsi, masks, vec_num * ncopies,
@@ -7877,11 +7877,15 @@ vect_transform_reduction (loop_vec_info loop_vinfo,
if (emulated_mixed_dot_prod)
new_stmt = vect_emulate_mixed_dot_prod (loop_vinfo, stmt_info, gsi,
vec_dest, vop);
+ else if (code.is_internal_fn ())
+ new_stmt = gimple_build_call_internal (internal_fn (code),
+ op.num_ops,
+ vop[0], vop[1], vop[2]);
else
- new_stmt = gimple_build_assign (vec_dest, code,
+ new_stmt = gimple_build_assign (vec_dest, tree_code (op.code),
vop[0], vop[1], vop[2]);
new_temp = make_ssa_name (vec_dest, new_stmt);
- gimple_assign_set_lhs (new_stmt, new_temp);
+ gimple_set_lhs (new_stmt, new_temp);
vect_finish_stmt_generation (loop_vinfo, stmt_info, new_stmt, gsi);
}