aboutsummaryrefslogtreecommitdiff
path: root/gcc
diff options
context:
space:
mode:
authorRichard Biener <rguenther@suse.de>2024-07-17 10:22:47 +0200
committerRichard Biener <rguenth@gcc.gnu.org>2024-07-18 09:43:52 +0200
commit8c67dc40459e3d72e8169b099cc8c5dbdb759da3 (patch)
tree9f31a22e983e39d75d6c3986d220391cd5e0620e /gcc
parent030186cabe8128e752619e101768cf8823a42c38 (diff)
downloadgcc-8c67dc40459e3d72e8169b099cc8c5dbdb759da3.zip
gcc-8c67dc40459e3d72e8169b099cc8c5dbdb759da3.tar.gz
gcc-8c67dc40459e3d72e8169b099cc8c5dbdb759da3.tar.bz2
tree-optimization/104515 - store motion and clobbers
The following addresses an old regression when end-of-object/storage clobbers were introduced. In particular when there's an end-of-object clobber in a loop but no corresponding begin-of-object we can still perform store motion of may-aliased refs when we re-issue the end-of-object/storage on the exits but elide it from the loop. This should be the safest way to deal with this considering stack-slot sharing and it should not cause missed dead store eliminations given DSE can now follow multiple paths in case there are multiple exits. Note when the clobber is re-materialized only on one exit but not on anther we are erroring on the side of removing the clobber on such path. This should be OK (removing clobbers is always OK). Note there's no corresponding code to handle begin-of-object/storage during the hoisting part of loads that are part of a store motion optimization, so this only enables stored-only store motion or cases without such clobber inside the loop. PR tree-optimization/104515 * tree-ssa-loop-im.cc (execute_sm_exit): Add clobbers_to_prune parameter and handle re-materializing of clobbers. (sm_seq_valid_bb): end-of-storage/object clobbers are OK inside an ordered sequence of stores. (sm_seq_push_down): Refuse to push down clobbers. (hoist_memory_references): Prune clobbers from the loop body we re-materialized on an exit. * g++.dg/opt/pr104515.C: New testcase.
Diffstat (limited to 'gcc')
-rw-r--r--gcc/testsuite/g++.dg/opt/pr104515.C18
-rw-r--r--gcc/tree-ssa-loop-im.cc86
2 files changed, 89 insertions, 15 deletions
diff --git a/gcc/testsuite/g++.dg/opt/pr104515.C b/gcc/testsuite/g++.dg/opt/pr104515.C
new file mode 100644
index 0000000..f5455a4
--- /dev/null
+++ b/gcc/testsuite/g++.dg/opt/pr104515.C
@@ -0,0 +1,18 @@
+// { dg-do compile { target c++11 } }
+// { dg-options "-O2 -fdump-tree-lim2-details" }
+
+using T = int;
+struct Vec {
+ T* end;
+};
+void pop_back_many(Vec& v, unsigned n)
+{
+ for (unsigned i = 0; i < n; ++i) {
+ --v.end;
+ // The end-of-object clobber prevented store motion of v
+ v.end->~T();
+ }
+}
+
+// { dg-final { scan-tree-dump "Executing store motion of v" "lim2" } }
+// { dg-final { scan-tree-dump "Re-issueing dependent" "lim2" } }
diff --git a/gcc/tree-ssa-loop-im.cc b/gcc/tree-ssa-loop-im.cc
index 61c6339..c53efbb 100644
--- a/gcc/tree-ssa-loop-im.cc
+++ b/gcc/tree-ssa-loop-im.cc
@@ -2368,7 +2368,8 @@ struct seq_entry
static void
execute_sm_exit (class loop *loop, edge ex, vec<seq_entry> &seq,
hash_map<im_mem_ref *, sm_aux *> &aux_map, sm_kind kind,
- edge &append_cond_position, edge &last_cond_fallthru)
+ edge &append_cond_position, edge &last_cond_fallthru,
+ bitmap clobbers_to_prune)
{
/* Sink the stores to exit from the loop. */
for (unsigned i = seq.length (); i > 0; --i)
@@ -2377,15 +2378,35 @@ execute_sm_exit (class loop *loop, edge ex, vec<seq_entry> &seq,
if (seq[i-1].second == sm_other)
{
gcc_assert (kind == sm_ord && seq[i-1].from != NULL_TREE);
- if (dump_file && (dump_flags & TDF_DETAILS))
+ gassign *store;
+ if (ref->mem.ref == error_mark_node)
{
- fprintf (dump_file, "Re-issueing dependent store of ");
- print_generic_expr (dump_file, ref->mem.ref);
- fprintf (dump_file, " from loop %d on exit %d -> %d\n",
- loop->num, ex->src->index, ex->dest->index);
+ tree lhs = gimple_assign_lhs (ref->accesses_in_loop[0].stmt);
+ if (dump_file && (dump_flags & TDF_DETAILS))
+ {
+ fprintf (dump_file, "Re-issueing dependent ");
+ print_generic_expr (dump_file, unshare_expr (seq[i-1].from));
+ fprintf (dump_file, " of ");
+ print_generic_expr (dump_file, lhs);
+ fprintf (dump_file, " from loop %d on exit %d -> %d\n",
+ loop->num, ex->src->index, ex->dest->index);
+ }
+ store = gimple_build_assign (unshare_expr (lhs),
+ unshare_expr (seq[i-1].from));
+ bitmap_set_bit (clobbers_to_prune, seq[i-1].first);
+ }
+ else
+ {
+ if (dump_file && (dump_flags & TDF_DETAILS))
+ {
+ fprintf (dump_file, "Re-issueing dependent store of ");
+ print_generic_expr (dump_file, ref->mem.ref);
+ fprintf (dump_file, " from loop %d on exit %d -> %d\n",
+ loop->num, ex->src->index, ex->dest->index);
+ }
+ store = gimple_build_assign (unshare_expr (ref->mem.ref),
+ seq[i-1].from);
}
- gassign *store = gimple_build_assign (unshare_expr (ref->mem.ref),
- seq[i-1].from);
gsi_insert_on_edge (ex, store);
}
else
@@ -2426,6 +2447,12 @@ sm_seq_push_down (vec<seq_entry> &seq, unsigned ptr, unsigned *at)
break;
/* We may not ignore self-dependences here. */
if (new_cand.first == against.first
+ /* ??? We could actually handle clobbers here, but not easily
+ with LIMs dependence analysis. */
+ || (memory_accesses.refs_list[new_cand.first]->mem.ref
+ == error_mark_node)
+ || (memory_accesses.refs_list[against.first]->mem.ref
+ == error_mark_node)
|| !refs_independent_p (memory_accesses.refs_list[new_cand.first],
memory_accesses.refs_list[against.first],
false))
@@ -2656,13 +2683,17 @@ sm_seq_valid_bb (class loop *loop, basic_block bb, tree vdef,
return 1;
}
lim_aux_data *data = get_lim_data (def);
- gcc_assert (data);
+ im_mem_ref *ref = memory_accesses.refs_list[data->ref];
if (data->ref == UNANALYZABLE_MEM_ID)
return -1;
/* Stop at memory references which we can't move. */
- else if (memory_accesses.refs_list[data->ref]->mem.ref == error_mark_node
- || TREE_THIS_VOLATILE
- (memory_accesses.refs_list[data->ref]->mem.ref))
+ else if ((ref->mem.ref == error_mark_node
+ /* We can move end-of-storage/object down. */
+ && !gimple_clobber_p (ref->accesses_in_loop[0].stmt,
+ CLOBBER_STORAGE_END)
+ && !gimple_clobber_p (ref->accesses_in_loop[0].stmt,
+ CLOBBER_OBJECT_END))
+ || TREE_THIS_VOLATILE (ref->mem.ref))
{
/* Mark refs_not_in_seq as unsupported. */
bitmap_ior_into (refs_not_supported, refs_not_in_seq);
@@ -2818,6 +2849,7 @@ hoist_memory_references (class loop *loop, bitmap mem_refs,
/* Materialize ordered store sequences on exits. */
edge e;
+ auto_bitmap clobbers_to_prune;
FOR_EACH_VEC_ELT (exits, i, e)
{
edge append_cond_position = NULL;
@@ -2836,10 +2868,21 @@ hoist_memory_references (class loop *loop, bitmap mem_refs,
append_cond_position,
last_cond_fallthru));
execute_sm_exit (loop, insert_e, seq, aux_map, sm_ord,
- append_cond_position, last_cond_fallthru);
+ append_cond_position, last_cond_fallthru,
+ clobbers_to_prune);
gsi_commit_one_edge_insert (insert_e, NULL);
}
+ /* Remove clobbers inside the loop we re-materialized on exits. */
+ EXECUTE_IF_SET_IN_BITMAP (clobbers_to_prune, 0, i, bi)
+ {
+ gimple *stmt = memory_accesses.refs_list[i]->accesses_in_loop[0].stmt;
+ gimple_stmt_iterator gsi = gsi_for_stmt (stmt);
+ unlink_stmt_vdef (stmt);
+ release_defs (stmt);
+ gsi_remove (&gsi, true);
+ }
+
for (hash_map<im_mem_ref *, sm_aux *>::iterator iter = aux_map.begin ();
iter != aux_map.end (); ++iter)
delete (*iter).second;
@@ -2990,6 +3033,7 @@ hoist_memory_references (class loop *loop, bitmap mem_refs,
}
/* Materialize ordered store sequences on exits. */
+ auto_bitmap clobbers_to_prune;
FOR_EACH_VEC_ELT (exits, i, e)
{
edge append_cond_position = NULL;
@@ -2998,17 +3042,29 @@ hoist_memory_references (class loop *loop, bitmap mem_refs,
{
gcc_assert (sms[i].first == e);
execute_sm_exit (loop, e, sms[i].second, aux_map, sm_ord,
- append_cond_position, last_cond_fallthru);
+ append_cond_position, last_cond_fallthru,
+ clobbers_to_prune);
sms[i].second.release ();
}
if (!unord_refs.is_empty ())
execute_sm_exit (loop, e, unord_refs, aux_map, sm_unord,
- append_cond_position, last_cond_fallthru);
+ append_cond_position, last_cond_fallthru,
+ clobbers_to_prune);
/* Commit edge inserts here to preserve the order of stores
when an exit exits multiple loops. */
gsi_commit_one_edge_insert (e, NULL);
}
+ /* Remove clobbers inside the loop we re-materialized on exits. */
+ EXECUTE_IF_SET_IN_BITMAP (clobbers_to_prune, 0, i, bi)
+ {
+ gimple *stmt = memory_accesses.refs_list[i]->accesses_in_loop[0].stmt;
+ gimple_stmt_iterator gsi = gsi_for_stmt (stmt);
+ unlink_stmt_vdef (stmt);
+ release_defs (stmt);
+ gsi_remove (&gsi, true);
+ }
+
for (hash_map<im_mem_ref *, sm_aux *>::iterator iter = aux_map.begin ();
iter != aux_map.end (); ++iter)
delete (*iter).second;