[3/7] PHIOPT: Move store_elim_worker into pass_cselim::execute
Checks
Context |
Check |
Description |
snail/gcc-patch-check |
success
|
Github commit url
|
Commit Message
This simple patch moves the body of store_elim_worker
direclty into pass_cselim::execute.
Also removes unneeded prototypes too.
OK? Bootstrapped and tested on x86_64-linux-gnu with no regressions.
gcc/ChangeLog:
* tree-ssa-phiopt.cc (cond_store_replacement): Remove
prototype.
(cond_if_else_store_replacement): Likewise.
(get_non_trapping): Likewise.
(store_elim_worker): Move into ...
(pass_cselim::execute): This.
---
gcc/tree-ssa-phiopt.cc | 250 ++++++++++++++++++++---------------------
1 file changed, 119 insertions(+), 131 deletions(-)
Comments
On Mon, Apr 24, 2023 at 11:31 PM Andrew Pinski via Gcc-patches
<gcc-patches@gcc.gnu.org> wrote:
>
> This simple patch moves the body of store_elim_worker
> direclty into pass_cselim::execute.
>
> Also removes unneeded prototypes too.
>
> OK? Bootstrapped and tested on x86_64-linux-gnu with no regressions.
OK.
> gcc/ChangeLog:
>
> * tree-ssa-phiopt.cc (cond_store_replacement): Remove
> prototype.
> (cond_if_else_store_replacement): Likewise.
> (get_non_trapping): Likewise.
> (store_elim_worker): Move into ...
> (pass_cselim::execute): This.
> ---
> gcc/tree-ssa-phiopt.cc | 250 ++++++++++++++++++++---------------------
> 1 file changed, 119 insertions(+), 131 deletions(-)
>
> diff --git a/gcc/tree-ssa-phiopt.cc b/gcc/tree-ssa-phiopt.cc
> index d232fd9b551..fb2d2c9fc1a 100644
> --- a/gcc/tree-ssa-phiopt.cc
> +++ b/gcc/tree-ssa-phiopt.cc
> @@ -55,11 +55,6 @@ along with GCC; see the file COPYING3. If not see
> #include "tree-ssa-propagate.h"
> #include "tree-ssa-dce.h"
>
> -static bool cond_store_replacement (basic_block, basic_block, edge, edge,
> - hash_set<tree> *);
> -static bool cond_if_else_store_replacement (basic_block, basic_block, basic_block);
> -static hash_set<tree> * get_non_trapping ();
> -
> /* Return the singleton PHI in the SEQ of PHIs for edges E0 and E1. */
>
> static gphi *
> @@ -87,130 +82,6 @@ single_non_singleton_phi_for_edges (gimple_seq seq, edge e0, edge e1)
> return phi;
> }
>
> -/* The core routine of conditional store replacement. */
> -static unsigned int
> -store_elim_worker (void)
> -{
> - basic_block bb;
> - basic_block *bb_order;
> - unsigned n, i;
> - bool cfgchanged = false;
> - hash_set<tree> *nontrap = 0;
> -
> - calculate_dominance_info (CDI_DOMINATORS);
> -
> - /* Calculate the set of non-trapping memory accesses. */
> - nontrap = get_non_trapping ();
> -
> - /* Search every basic block for COND_EXPR we may be able to optimize.
> -
> - We walk the blocks in order that guarantees that a block with
> - a single predecessor is processed before the predecessor.
> - This ensures that we collapse inner ifs before visiting the
> - outer ones, and also that we do not try to visit a removed
> - block. */
> - bb_order = single_pred_before_succ_order ();
> - n = n_basic_blocks_for_fn (cfun) - NUM_FIXED_BLOCKS;
> -
> - for (i = 0; i < n; i++)
> - {
> - basic_block bb1, bb2;
> - edge e1, e2;
> - bool diamond_p = false;
> -
> - bb = bb_order[i];
> -
> - /* Check to see if the last statement is a GIMPLE_COND. */
> - gcond *cond_stmt = safe_dyn_cast <gcond *> (*gsi_last_bb (bb));
> - if (!cond_stmt)
> - continue;
> -
> - e1 = EDGE_SUCC (bb, 0);
> - bb1 = e1->dest;
> - e2 = EDGE_SUCC (bb, 1);
> - bb2 = e2->dest;
> -
> - /* We cannot do the optimization on abnormal edges. */
> - if ((e1->flags & EDGE_ABNORMAL) != 0
> - || (e2->flags & EDGE_ABNORMAL) != 0)
> - continue;
> -
> - /* If either bb1's succ or bb2 or bb2's succ is non NULL. */
> - if (EDGE_COUNT (bb1->succs) == 0
> - || EDGE_COUNT (bb2->succs) == 0)
> - continue;
> -
> - /* Find the bb which is the fall through to the other. */
> - if (EDGE_SUCC (bb1, 0)->dest == bb2)
> - ;
> - else if (EDGE_SUCC (bb2, 0)->dest == bb1)
> - {
> - std::swap (bb1, bb2);
> - std::swap (e1, e2);
> - }
> - else if (EDGE_SUCC (bb1, 0)->dest == EDGE_SUCC (bb2, 0)->dest
> - && single_succ_p (bb2))
> - {
> - diamond_p = true;
> - e2 = EDGE_SUCC (bb2, 0);
> - /* Make sure bb2 is just a fall through. */
> - if ((e2->flags & EDGE_FALLTHRU) == 0)
> - continue;
> - }
> - else
> - continue;
> -
> - e1 = EDGE_SUCC (bb1, 0);
> -
> - /* Make sure that bb1 is just a fall through. */
> - if (!single_succ_p (bb1)
> - || (e1->flags & EDGE_FALLTHRU) == 0)
> - continue;
> -
> - if (diamond_p)
> - {
> - basic_block bb3 = e1->dest;
> -
> - /* Only handle sinking of store from 2 bbs only,
> - The middle bbs don't need to come from the
> - if always since we are sinking rather than
> - hoisting. */
> - if (EDGE_COUNT (bb3->preds) != 2)
> - continue;
> - if (cond_if_else_store_replacement (bb1, bb2, bb3))
> - cfgchanged = true;
> - continue;
> - }
> -
> - /* Also make sure that bb1 only have one predecessor and that it
> - is bb. */
> - if (!single_pred_p (bb1)
> - || single_pred (bb1) != bb)
> - continue;
> -
> - /* bb1 is the middle block, bb2 the join block, bb the split block,
> - e1 the fallthrough edge from bb1 to bb2. We can't do the
> - optimization if the join block has more than two predecessors. */
> - if (EDGE_COUNT (bb2->preds) > 2)
> - continue;
> - if (cond_store_replacement (bb1, bb2, e1, e2, nontrap))
> - cfgchanged = true;
> - }
> -
> - free (bb_order);
> -
> - delete nontrap;
> - /* If the CFG has changed, we should cleanup the CFG. */
> - if (cfgchanged)
> - {
> - /* In cond-store replacement we have added some loads on edges
> - and new VOPS (as we moved the store, and created a load). */
> - gsi_commit_edge_inserts ();
> - return TODO_cleanup_cfg | TODO_update_ssa_only_virtuals;
> - }
> - return 0;
> -}
> -
> /* Replace PHI node element whose edge is E in block BB with variable NEW.
> Remove the edge from COND_BLOCK which does not lead to BB (COND_BLOCK
> is known to have two edges, one of which must reach BB). */
> @@ -4403,13 +4274,130 @@ make_pass_cselim (gcc::context *ctxt)
> unsigned int
> pass_cselim::execute (function *)
> {
> - unsigned todo;
> + basic_block bb;
> + basic_block *bb_order;
> + unsigned n, i;
> + bool cfgchanged = false;
> + hash_set<tree> *nontrap = 0;
> + unsigned todo = 0;
> +
> /* ??? We are not interested in loop related info, but the following
> will create it, ICEing as we didn't init loops with pre-headers.
> An interfacing issue of find_data_references_in_bb. */
> loop_optimizer_init (LOOPS_NORMAL);
> scev_initialize ();
> - todo = store_elim_worker ();
> +
> + calculate_dominance_info (CDI_DOMINATORS);
> +
> + /* Calculate the set of non-trapping memory accesses. */
> + nontrap = get_non_trapping ();
> +
> + /* Search every basic block for COND_EXPR we may be able to optimize.
> +
> + We walk the blocks in order that guarantees that a block with
> + a single predecessor is processed before the predecessor.
> + This ensures that we collapse inner ifs before visiting the
> + outer ones, and also that we do not try to visit a removed
> + block. */
> + bb_order = single_pred_before_succ_order ();
> + n = n_basic_blocks_for_fn (cfun) - NUM_FIXED_BLOCKS;
> +
> + for (i = 0; i < n; i++)
> + {
> + basic_block bb1, bb2;
> + edge e1, e2;
> + bool diamond_p = false;
> +
> + bb = bb_order[i];
> +
> + /* Check to see if the last statement is a GIMPLE_COND. */
> + gcond *cond_stmt = safe_dyn_cast <gcond *> (*gsi_last_bb (bb));
> + if (!cond_stmt)
> + continue;
> +
> + e1 = EDGE_SUCC (bb, 0);
> + bb1 = e1->dest;
> + e2 = EDGE_SUCC (bb, 1);
> + bb2 = e2->dest;
> +
> + /* We cannot do the optimization on abnormal edges. */
> + if ((e1->flags & EDGE_ABNORMAL) != 0
> + || (e2->flags & EDGE_ABNORMAL) != 0)
> + continue;
> +
> + /* If either bb1's succ or bb2 or bb2's succ is non NULL. */
> + if (EDGE_COUNT (bb1->succs) == 0
> + || EDGE_COUNT (bb2->succs) == 0)
> + continue;
> +
> + /* Find the bb which is the fall through to the other. */
> + if (EDGE_SUCC (bb1, 0)->dest == bb2)
> + ;
> + else if (EDGE_SUCC (bb2, 0)->dest == bb1)
> + {
> + std::swap (bb1, bb2);
> + std::swap (e1, e2);
> + }
> + else if (EDGE_SUCC (bb1, 0)->dest == EDGE_SUCC (bb2, 0)->dest
> + && single_succ_p (bb2))
> + {
> + diamond_p = true;
> + e2 = EDGE_SUCC (bb2, 0);
> + /* Make sure bb2 is just a fall through. */
> + if ((e2->flags & EDGE_FALLTHRU) == 0)
> + continue;
> + }
> + else
> + continue;
> +
> + e1 = EDGE_SUCC (bb1, 0);
> +
> + /* Make sure that bb1 is just a fall through. */
> + if (!single_succ_p (bb1)
> + || (e1->flags & EDGE_FALLTHRU) == 0)
> + continue;
> +
> + if (diamond_p)
> + {
> + basic_block bb3 = e1->dest;
> +
> + /* Only handle sinking of store from 2 bbs only,
> + The middle bbs don't need to come from the
> + if always since we are sinking rather than
> + hoisting. */
> + if (EDGE_COUNT (bb3->preds) != 2)
> + continue;
> + if (cond_if_else_store_replacement (bb1, bb2, bb3))
> + cfgchanged = true;
> + continue;
> + }
> +
> + /* Also make sure that bb1 only have one predecessor and that it
> + is bb. */
> + if (!single_pred_p (bb1)
> + || single_pred (bb1) != bb)
> + continue;
> +
> + /* bb1 is the middle block, bb2 the join block, bb the split block,
> + e1 the fallthrough edge from bb1 to bb2. We can't do the
> + optimization if the join block has more than two predecessors. */
> + if (EDGE_COUNT (bb2->preds) > 2)
> + continue;
> + if (cond_store_replacement (bb1, bb2, e1, e2, nontrap))
> + cfgchanged = true;
> + }
> +
> + free (bb_order);
> +
> + delete nontrap;
> + /* If the CFG has changed, we should cleanup the CFG. */
> + if (cfgchanged)
> + {
> + /* In cond-store replacement we have added some loads on edges
> + and new VOPS (as we moved the store, and created a load). */
> + gsi_commit_edge_inserts ();
> + todo = TODO_cleanup_cfg | TODO_update_ssa_only_virtuals;
> + }
> scev_finalize ();
> loop_optimizer_finalize ();
> return todo;
> --
> 2.39.1
>
@@ -55,11 +55,6 @@ along with GCC; see the file COPYING3. If not see
#include "tree-ssa-propagate.h"
#include "tree-ssa-dce.h"
-static bool cond_store_replacement (basic_block, basic_block, edge, edge,
- hash_set<tree> *);
-static bool cond_if_else_store_replacement (basic_block, basic_block, basic_block);
-static hash_set<tree> * get_non_trapping ();
-
/* Return the singleton PHI in the SEQ of PHIs for edges E0 and E1. */
static gphi *
@@ -87,130 +82,6 @@ single_non_singleton_phi_for_edges (gimple_seq seq, edge e0, edge e1)
return phi;
}
-/* The core routine of conditional store replacement. */
-static unsigned int
-store_elim_worker (void)
-{
- basic_block bb;
- basic_block *bb_order;
- unsigned n, i;
- bool cfgchanged = false;
- hash_set<tree> *nontrap = 0;
-
- calculate_dominance_info (CDI_DOMINATORS);
-
- /* Calculate the set of non-trapping memory accesses. */
- nontrap = get_non_trapping ();
-
- /* Search every basic block for COND_EXPR we may be able to optimize.
-
- We walk the blocks in order that guarantees that a block with
- a single predecessor is processed before the predecessor.
- This ensures that we collapse inner ifs before visiting the
- outer ones, and also that we do not try to visit a removed
- block. */
- bb_order = single_pred_before_succ_order ();
- n = n_basic_blocks_for_fn (cfun) - NUM_FIXED_BLOCKS;
-
- for (i = 0; i < n; i++)
- {
- basic_block bb1, bb2;
- edge e1, e2;
- bool diamond_p = false;
-
- bb = bb_order[i];
-
- /* Check to see if the last statement is a GIMPLE_COND. */
- gcond *cond_stmt = safe_dyn_cast <gcond *> (*gsi_last_bb (bb));
- if (!cond_stmt)
- continue;
-
- e1 = EDGE_SUCC (bb, 0);
- bb1 = e1->dest;
- e2 = EDGE_SUCC (bb, 1);
- bb2 = e2->dest;
-
- /* We cannot do the optimization on abnormal edges. */
- if ((e1->flags & EDGE_ABNORMAL) != 0
- || (e2->flags & EDGE_ABNORMAL) != 0)
- continue;
-
- /* If either bb1's succ or bb2 or bb2's succ is non NULL. */
- if (EDGE_COUNT (bb1->succs) == 0
- || EDGE_COUNT (bb2->succs) == 0)
- continue;
-
- /* Find the bb which is the fall through to the other. */
- if (EDGE_SUCC (bb1, 0)->dest == bb2)
- ;
- else if (EDGE_SUCC (bb2, 0)->dest == bb1)
- {
- std::swap (bb1, bb2);
- std::swap (e1, e2);
- }
- else if (EDGE_SUCC (bb1, 0)->dest == EDGE_SUCC (bb2, 0)->dest
- && single_succ_p (bb2))
- {
- diamond_p = true;
- e2 = EDGE_SUCC (bb2, 0);
- /* Make sure bb2 is just a fall through. */
- if ((e2->flags & EDGE_FALLTHRU) == 0)
- continue;
- }
- else
- continue;
-
- e1 = EDGE_SUCC (bb1, 0);
-
- /* Make sure that bb1 is just a fall through. */
- if (!single_succ_p (bb1)
- || (e1->flags & EDGE_FALLTHRU) == 0)
- continue;
-
- if (diamond_p)
- {
- basic_block bb3 = e1->dest;
-
- /* Only handle sinking of store from 2 bbs only,
- The middle bbs don't need to come from the
- if always since we are sinking rather than
- hoisting. */
- if (EDGE_COUNT (bb3->preds) != 2)
- continue;
- if (cond_if_else_store_replacement (bb1, bb2, bb3))
- cfgchanged = true;
- continue;
- }
-
- /* Also make sure that bb1 only have one predecessor and that it
- is bb. */
- if (!single_pred_p (bb1)
- || single_pred (bb1) != bb)
- continue;
-
- /* bb1 is the middle block, bb2 the join block, bb the split block,
- e1 the fallthrough edge from bb1 to bb2. We can't do the
- optimization if the join block has more than two predecessors. */
- if (EDGE_COUNT (bb2->preds) > 2)
- continue;
- if (cond_store_replacement (bb1, bb2, e1, e2, nontrap))
- cfgchanged = true;
- }
-
- free (bb_order);
-
- delete nontrap;
- /* If the CFG has changed, we should cleanup the CFG. */
- if (cfgchanged)
- {
- /* In cond-store replacement we have added some loads on edges
- and new VOPS (as we moved the store, and created a load). */
- gsi_commit_edge_inserts ();
- return TODO_cleanup_cfg | TODO_update_ssa_only_virtuals;
- }
- return 0;
-}
-
/* Replace PHI node element whose edge is E in block BB with variable NEW.
Remove the edge from COND_BLOCK which does not lead to BB (COND_BLOCK
is known to have two edges, one of which must reach BB). */
@@ -4403,13 +4274,130 @@ make_pass_cselim (gcc::context *ctxt)
unsigned int
pass_cselim::execute (function *)
{
- unsigned todo;
+ basic_block bb;
+ basic_block *bb_order;
+ unsigned n, i;
+ bool cfgchanged = false;
+ hash_set<tree> *nontrap = 0;
+ unsigned todo = 0;
+
/* ??? We are not interested in loop related info, but the following
will create it, ICEing as we didn't init loops with pre-headers.
An interfacing issue of find_data_references_in_bb. */
loop_optimizer_init (LOOPS_NORMAL);
scev_initialize ();
- todo = store_elim_worker ();
+
+ calculate_dominance_info (CDI_DOMINATORS);
+
+ /* Calculate the set of non-trapping memory accesses. */
+ nontrap = get_non_trapping ();
+
+ /* Search every basic block for COND_EXPR we may be able to optimize.
+
+ We walk the blocks in order that guarantees that a block with
+ a single predecessor is processed before the predecessor.
+ This ensures that we collapse inner ifs before visiting the
+ outer ones, and also that we do not try to visit a removed
+ block. */
+ bb_order = single_pred_before_succ_order ();
+ n = n_basic_blocks_for_fn (cfun) - NUM_FIXED_BLOCKS;
+
+ for (i = 0; i < n; i++)
+ {
+ basic_block bb1, bb2;
+ edge e1, e2;
+ bool diamond_p = false;
+
+ bb = bb_order[i];
+
+ /* Check to see if the last statement is a GIMPLE_COND. */
+ gcond *cond_stmt = safe_dyn_cast <gcond *> (*gsi_last_bb (bb));
+ if (!cond_stmt)
+ continue;
+
+ e1 = EDGE_SUCC (bb, 0);
+ bb1 = e1->dest;
+ e2 = EDGE_SUCC (bb, 1);
+ bb2 = e2->dest;
+
+ /* We cannot do the optimization on abnormal edges. */
+ if ((e1->flags & EDGE_ABNORMAL) != 0
+ || (e2->flags & EDGE_ABNORMAL) != 0)
+ continue;
+
+ /* If either bb1's succ or bb2 or bb2's succ is non NULL. */
+ if (EDGE_COUNT (bb1->succs) == 0
+ || EDGE_COUNT (bb2->succs) == 0)
+ continue;
+
+ /* Find the bb which is the fall through to the other. */
+ if (EDGE_SUCC (bb1, 0)->dest == bb2)
+ ;
+ else if (EDGE_SUCC (bb2, 0)->dest == bb1)
+ {
+ std::swap (bb1, bb2);
+ std::swap (e1, e2);
+ }
+ else if (EDGE_SUCC (bb1, 0)->dest == EDGE_SUCC (bb2, 0)->dest
+ && single_succ_p (bb2))
+ {
+ diamond_p = true;
+ e2 = EDGE_SUCC (bb2, 0);
+ /* Make sure bb2 is just a fall through. */
+ if ((e2->flags & EDGE_FALLTHRU) == 0)
+ continue;
+ }
+ else
+ continue;
+
+ e1 = EDGE_SUCC (bb1, 0);
+
+ /* Make sure that bb1 is just a fall through. */
+ if (!single_succ_p (bb1)
+ || (e1->flags & EDGE_FALLTHRU) == 0)
+ continue;
+
+ if (diamond_p)
+ {
+ basic_block bb3 = e1->dest;
+
+ /* Only handle sinking of store from 2 bbs only,
+ The middle bbs don't need to come from the
+ if always since we are sinking rather than
+ hoisting. */
+ if (EDGE_COUNT (bb3->preds) != 2)
+ continue;
+ if (cond_if_else_store_replacement (bb1, bb2, bb3))
+ cfgchanged = true;
+ continue;
+ }
+
+ /* Also make sure that bb1 only have one predecessor and that it
+ is bb. */
+ if (!single_pred_p (bb1)
+ || single_pred (bb1) != bb)
+ continue;
+
+ /* bb1 is the middle block, bb2 the join block, bb the split block,
+ e1 the fallthrough edge from bb1 to bb2. We can't do the
+ optimization if the join block has more than two predecessors. */
+ if (EDGE_COUNT (bb2->preds) > 2)
+ continue;
+ if (cond_store_replacement (bb1, bb2, e1, e2, nontrap))
+ cfgchanged = true;
+ }
+
+ free (bb_order);
+
+ delete nontrap;
+ /* If the CFG has changed, we should cleanup the CFG. */
+ if (cfgchanged)
+ {
+ /* In cond-store replacement we have added some loads on edges
+ and new VOPS (as we moved the store, and created a load). */
+ gsi_commit_edge_inserts ();
+ todo = TODO_cleanup_cfg | TODO_update_ssa_only_virtuals;
+ }
scev_finalize ();
loop_optimizer_finalize ();
return todo;