/* Convert a program in SSA form into Normal form. Copyright (C) 2004-2021 Free Software Foundation, Inc. Contributed by Andrew Macleod This file is part of GCC. GCC is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 3, or (at your option) any later version. GCC is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with GCC; see the file COPYING3. If not see . */ #include "config.h" #include "system.h" #include "coretypes.h" #include "backend.h" #include "rtl.h" #include "tree.h" #include "gimple.h" #include "cfghooks.h" #include "ssa.h" #include "tree-ssa.h" #include "memmodel.h" #include "emit-rtl.h" #include "gimple-pretty-print.h" #include "diagnostic-core.h" #include "tree-dfa.h" #include "stor-layout.h" #include "cfgrtl.h" #include "cfganal.h" #include "tree-eh.h" #include "gimple-iterator.h" #include "tree-cfg.h" #include "dumpfile.h" #include "tree-ssa-live.h" #include "tree-ssa-ter.h" #include "tree-ssa-coalesce.h" #include "tree-outof-ssa.h" #include "dojump.h" /* FIXME: A lot of code here deals with expanding to RTL. All that code should be in cfgexpand.c. */ #include "explow.h" #include "expr.h" /* Return TRUE if expression STMT is suitable for replacement. */ bool ssa_is_replaceable_p (gimple *stmt) { use_operand_p use_p; tree def; gimple *use_stmt; /* Only consider modify stmts. */ if (!is_gimple_assign (stmt)) return false; /* If the statement may throw an exception, it cannot be replaced. */ if (stmt_could_throw_p (cfun, stmt)) return false; /* Punt if there is more than 1 def. */ def = SINGLE_SSA_TREE_OPERAND (stmt, SSA_OP_DEF); if (!def) return false; /* Only consider definitions which have a single use. */ if (!single_imm_use (def, &use_p, &use_stmt)) return false; /* Used in this block, but at the TOP of the block, not the end. */ if (gimple_code (use_stmt) == GIMPLE_PHI) return false; /* There must be no VDEFs. */ if (gimple_vdef (stmt)) return false; /* Float expressions must go through memory if float-store is on. */ if (flag_float_store && FLOAT_TYPE_P (gimple_expr_type (stmt))) return false; /* An assignment with a register variable on the RHS is not replaceable. */ if (gimple_assign_rhs_code (stmt) == VAR_DECL && DECL_HARD_REGISTER (gimple_assign_rhs1 (stmt))) return false; /* No function calls can be replaced. */ if (is_gimple_call (stmt)) return false; /* Leave any stmt with volatile operands alone as well. */ if (gimple_has_volatile_ops (stmt)) return false; return true; } /* Used to hold all the components required to do SSA PHI elimination. The node and pred/succ list is a simple linear list of nodes and edges represented as pairs of nodes. The predecessor and successor list: Nodes are entered in pairs, where [0] ->PRED, [1]->SUCC. All the even indexes in the array represent predecessors, all the odd elements are successors. Rationale: When implemented as bitmaps, very large programs SSA->Normal times were being dominated by clearing the interference graph. Typically this list of edges is extremely small since it only includes PHI results and uses from a single edge which have not coalesced with each other. This means that no virtual PHI nodes are included, and empirical evidence suggests that the number of edges rarely exceed 3, and in a bootstrap of GCC, the maximum size encountered was 7. This also limits the number of possible nodes that are involved to rarely more than 6, and in the bootstrap of gcc, the maximum number of nodes encountered was 12. */ class elim_graph { public: elim_graph (var_map map); /* Size of the elimination vectors. */ int size; /* List of nodes in the elimination graph. */ auto_vec nodes; /* The predecessor and successor edge list. */ auto_vec edge_list; /* Source locus on each edge */ auto_vec edge_locus; /* Visited vector. */ auto_sbitmap visited; /* Stack for visited nodes. */ auto_vec stack; /* The variable partition map. */ var_map map; /* Edge being eliminated by this graph. */ edge e; /* List of constant copies to emit. These are pushed on in pairs. */ auto_vec const_dests; auto_vec const_copies; /* Source locations for any constant copies. */ auto_vec copy_locus; }; /* For an edge E find out a good source location to associate with instructions inserted on edge E. If E has an implicit goto set, use its location. Otherwise search instructions in predecessors of E for a location, and use that one. That makes sense because we insert on edges for PHI nodes, and effects of PHIs happen on the end of the predecessor conceptually. An exception is made for EH edges because we don't want to drag the source location of unrelated statements at the beginning of handlers; they would be further reused for various EH constructs, which would damage the coverage information. */ static void set_location_for_edge (edge e) { if (e->goto_locus) set_curr_insn_location (e->goto_locus); else if (e->flags & EDGE_EH) { basic_block bb = e->dest; gimple_stmt_iterator gsi; do { for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi)) { gimple *stmt = gsi_stmt (gsi); if (is_gimple_debug (stmt)) continue; if (gimple_has_location (stmt) || gimple_block (stmt)) { set_curr_insn_location (gimple_location (stmt)); return; } } /* Nothing found in this basic block. Make a half-assed attempt to continue with another block. */ if (single_succ_p (bb)) bb = single_succ (bb); else bb = e->dest; } while (bb != e->dest); } else { basic_block bb = e->src; gimple_stmt_iterator gsi; do { for (gsi = gsi_last_bb (bb); !gsi_end_p (gsi); gsi_prev (&gsi)) { gimple *stmt = gsi_stmt (gsi); if (is_gimple_debug (stmt)) continue; if (gimple_has_location (stmt) || gimple_block (stmt)) { set_curr_insn_location (gimple_location (stmt)); return; } } /* Nothing found in this basic block. Make a half-assed attempt to continue with another block. */ if (single_pred_p (bb)) bb = single_pred (bb); else bb = e->src; } while (bb != e->src); } } /* Emit insns to copy SRC into DEST converting SRC if necessary. As SRC/DEST might be BLKmode memory locations SIZEEXP is a tree from which we deduce the size to copy in that case. */ static inline rtx_insn * emit_partition_copy (rtx dest, rtx src, int unsignedsrcp, tree sizeexp) { start_sequence (); if (GET_MODE (src) != VOIDmode && GET_MODE (src) != GET_MODE (dest)) src = convert_to_mode (GET_MODE (dest), src, unsignedsrcp); if (GET_MODE (src) == BLKmode) { gcc_assert (GET_MODE (dest) == BLKmode); emit_block_move (dest, src, expr_size (sizeexp), BLOCK_OP_NORMAL); } else emit_move_insn (dest, src); do_pending_stack_adjust (); rtx_insn *seq = get_insns (); end_sequence (); return seq; } /* Insert a copy instruction from partition SRC to DEST onto edge E. */ static void insert_partition_copy_on_edge (edge e, int dest, int src, location_t locus) { tree var; if (dump_file && (dump_flags & TDF_DETAILS)) { fprintf (dump_file, "Inserting a partition copy on edge BB%d->BB%d : " "PART.%d = PART.%d", e->src->index, e->dest->index, dest, src); fprintf (dump_file, "\n"); } gcc_assert (SA.partition_to_pseudo[dest]); gcc_assert (SA.partition_to_pseudo[src]); set_location_for_edge (e); /* If a locus is provided, override the default. */ if (locus) set_curr_insn_location (locus); var = partition_to_var (SA.map, src); rtx_insn *seq = emit_partition_copy (copy_rtx (SA.partition_to_pseudo[dest]), copy_rtx (SA.partition_to_pseudo[src]), TYPE_UNSIGNED (TREE_TYPE (var)), var); insert_insn_on_edge (seq, e); } /* Insert a copy instruction from expression SRC to partition DEST onto edge E. */ static void insert_value_copy_on_edge (edge e, int dest, tree src, location_t locus) { rtx dest_rtx, seq, x; machine_mode dest_mode, src_mode; int unsignedp; if (dump_file && (dump_flags & TDF_DETAILS)) { fprintf (dump_file, "Inserting a value copy on edge BB%d->BB%d : PART.%d = ", e->src->index, e->dest->index, dest); print_generic_expr (dump_file, src, TDF_SLIM); fprintf (dump_file, "\n"); } dest_rtx = copy_rtx (SA.partition_to_pseudo[dest]); gcc_assert (dest_rtx); set_location_for_edge (e); /* If a locus is provided, override the default. */ if (locus) set_curr_insn_location (locus); start_sequence (); tree name = partition_to_var (SA.map, dest); src_mode = TYPE_MODE (TREE_TYPE (src)); dest_mode = GET_MODE (dest_rtx); gcc_assert (src_mode == TYPE_MODE (TREE_TYPE (name))); gcc_assert (!REG_P (dest_rtx) || dest_mode == promote_ssa_mode (name, &unsignedp)); if (src_mode != dest_mode) { x = expand_expr (src, NULL, src_mode, EXPAND_NORMAL); x = convert_modes (dest_mode, src_mode, x, unsignedp); } else if (src_mode == BLKmode) { x = dest_rtx; store_expr (src, x, 0, false, false); } else x = expand_expr (src, dest_rtx, dest_mode, EXPAND_NORMAL); if (x != dest_rtx) emit_move_insn (dest_rtx, x); do_pending_stack_adjust (); seq = get_insns (); end_sequence (); insert_insn_on_edge (seq, e); } /* Insert a copy instruction from RTL expression SRC to partition DEST onto edge E. */ static void insert_rtx_to_part_on_edge (edge e, int dest, rtx src, int unsignedsrcp, location_t locus) { if (dump_file && (dump_flags & TDF_DETAILS)) { fprintf (dump_file, "Inserting a temp copy on edge BB%d->BB%d : PART.%d = ", e->src->index, e->dest->index, dest); print_simple_rtl (dump_file, src); fprintf (dump_file, "\n"); } gcc_assert (SA.partition_to_pseudo[dest]); set_location_for_edge (e); /* If a locus is provided, override the default. */ if (locus) set_curr_insn_location (locus); /* We give the destination as sizeexp in case src/dest are BLKmode mems. Usually we give the source. As we result from SSA names the left and right size should be the same (and no WITH_SIZE_EXPR involved), so it doesn't matter. */ rtx_insn *seq = emit_partition_copy (copy_rtx (SA.partition_to_pseudo[dest]), src, unsignedsrcp, partition_to_var (SA.map, dest)); insert_insn_on_edge (seq, e); } /* Insert a copy instruction from partition SRC to RTL lvalue DEST onto edge E. */ static void insert_part_to_rtx_on_edge (edge e, rtx dest, int src, location_t locus) { tree var; if (dump_file && (dump_flags & TDF_DETAILS)) { fprintf (dump_file, "Inserting a temp copy on edge BB%d->BB%d : ", e->src->index, e->dest->index); print_simple_rtl (dump_file, dest); fprintf (dump_file, "= PART.%d\n", src); } gcc_assert (SA.partition_to_pseudo[src]); set_location_for_edge (e); /* If a locus is provided, override the default. */ if (locus) set_curr_insn_location (locus); var = partition_to_var (SA.map, src); rtx_insn *seq = emit_partition_copy (dest, copy_rtx (SA.partition_to_pseudo[src]), TYPE_UNSIGNED (TREE_TYPE (var)), var); insert_insn_on_edge (seq, e); } /* Create an elimination graph for map. */ elim_graph::elim_graph (var_map map) : nodes (30), edge_list (20), edge_locus (10), visited (map->num_partitions), stack (30), map (map), const_dests (20), const_copies (20), copy_locus (10) { } /* Empty elimination graph G. */ static inline void clear_elim_graph (elim_graph *g) { g->nodes.truncate (0); g->edge_list.truncate (0); g->edge_locus.truncate (0); } /* Return the number of nodes in graph G. */ static inline int elim_graph_size (elim_graph *g) { return g->nodes.length (); } /* Add NODE to graph G, if it doesn't exist already. */ static inline void elim_graph_add_node (elim_graph *g, int node) { int x; int t; FOR_EACH_VEC_ELT (g->nodes, x, t) if (t == node) return; g->nodes.safe_push (node); } /* Add the edge PRED->SUCC to graph G. */ static inline void elim_graph_add_edge (elim_graph *g, int pred, int succ, location_t locus) { g->edge_list.safe_push (pred); g->edge_list.safe_push (succ); g->edge_locus.safe_push (locus); } /* Remove an edge from graph G for which NODE is the predecessor, and return the successor node. -1 is returned if there is no such edge. */ static inline int elim_graph_remove_succ_edge (elim_graph *g, int node, location_t *locus) { int y; unsigned x; for (x = 0; x < g->edge_list.length (); x += 2) if (g->edge_list[x] == node) { g->edge_list[x] = -1; y = g->edge_list[x + 1]; g->edge_list[x + 1] = -1; *locus = g->edge_locus[x / 2]; g->edge_locus[x / 2] = UNKNOWN_LOCATION; return y; } *locus = UNKNOWN_LOCATION; return -1; } /* Find all the nodes in GRAPH which are successors to NODE in the edge list. VAR will hold the partition number found. CODE is the code fragment executed for every node found. */ #define FOR_EACH_ELIM_GRAPH_SUCC(GRAPH, NODE, VAR, LOCUS, CODE) \ do { \ unsigned x_; \ int y_; \ for (x_ = 0; x_ < (GRAPH)->edge_list.length (); x_ += 2) \ { \ y_ = (GRAPH)->edge_list[x_]; \ if (y_ != (NODE)) \ continue; \ (void) ((VAR) = (GRAPH)->edge_list[x_ + 1]); \ (void) ((LOCUS) = (GRAPH)->edge_locus[x_ / 2]); \ CODE; \ } \ } while (0) /* Find all the nodes which are predecessors of NODE in the edge list for GRAPH. VAR will hold the partition number found. CODE is the code fragment executed for every node found. */ #define FOR_EACH_ELIM_GRAPH_PRED(GRAPH, NODE, VAR, LOCUS, CODE) \ do { \ unsigned x_; \ int y_; \ for (x_ = 0; x_ < (GRAPH)->edge_list.length (); x_ += 2) \ { \ y_ = (GRAPH)->edge_list[x_ + 1]; \ if (y_ != (NODE)) \ continue; \ (void) ((VAR) = (GRAPH)->edge_list[x_]); \ (void) ((LOCUS) = (GRAPH)->edge_locus[x_ / 2]); \ CODE; \ } \ } while (0) /* Add T to elimination graph G. */ static inline void eliminate_name (elim_graph *g, int T) { elim_graph_add_node (g, T); } /* Return true if this phi argument T should have a copy queued when using var_map MAP. PHI nodes should contain only ssa_names and invariants. A test for ssa_name is definitely simpler, but don't let invalid contents slip through in the meantime. */ static inline bool queue_phi_copy_p (var_map map, tree t) { if (TREE_CODE (t) == SSA_NAME) { if (var_to_partition (map, t) == NO_PARTITION) return true; return false; } gcc_checking_assert (is_gimple_min_invariant (t)); return true; } /* Build elimination graph G for basic block BB on incoming PHI edge G->e. */ static void eliminate_build (elim_graph *g) { tree Ti; int p0, pi; gphi_iterator gsi; clear_elim_graph (g); for (gsi = gsi_start_phis (g->e->dest); !gsi_end_p (gsi); gsi_next (&gsi)) { gphi *phi = gsi.phi (); location_t locus; p0 = var_to_partition (g->map, gimple_phi_result (phi)); /* Ignore results which are not in partitions. */ if (p0 == NO_PARTITION) continue; Ti = PHI_ARG_DEF (phi, g->e->dest_idx); /* See set_location_for_edge for the rationale. */ if (g->e->flags & EDGE_EH) locus = UNKNOWN_LOCATION; else locus = gimple_phi_arg_location_from_edge (phi, g->e); /* If this argument is a constant, or a SSA_NAME which is being left in SSA form, just queue a copy to be emitted on this edge. */ if (queue_phi_copy_p (g->map, Ti)) { /* Save constant copies until all other copies have been emitted on this edge. */ g->const_dests.safe_push (p0); g->const_copies.safe_push (Ti); g->copy_locus.safe_push (locus); } else { pi = var_to_partition (g->map, Ti); if (p0 != pi) { eliminate_name (g, p0); eliminate_name (g, pi); elim_graph_add_edge (g, p0, pi, locus); } } } } /* Push successors of T onto the elimination stack for G. */ static void elim_forward (elim_graph *g, int T) { int S; location_t locus; bitmap_set_bit (g->visited, T); FOR_EACH_ELIM_GRAPH_SUCC (g, T, S, locus, { if (!bitmap_bit_p (g->visited, S)) elim_forward (g, S); }); g->stack.safe_push (T); } /* Return 1 if there unvisited predecessors of T in graph G. */ static int elim_unvisited_predecessor (elim_graph *g, int T) { int P; location_t locus; FOR_EACH_ELIM_GRAPH_PRED (g, T, P, locus, { if (!bitmap_bit_p (g->visited, P)) return 1; }); return 0; } /* Process predecessors first, and insert a copy. */ static void elim_backward (elim_graph *g, int T) { int P; location_t locus; bitmap_set_bit (g->visited, T); FOR_EACH_ELIM_GRAPH_PRED (g, T, P, locus, { if (!bitmap_bit_p (g->visited, P)) { elim_backward (g, P); insert_partition_copy_on_edge (g->e, P, T, locus); } }); } /* Allocate a new pseudo register usable for storing values sitting in NAME (a decl or SSA name), i.e. with matching mode and attributes. */ static rtx get_temp_reg (tree name) { tree type = TREE_TYPE (name); int unsignedp; machine_mode reg_mode = promote_ssa_mode (name, &unsignedp); if (reg_mode == BLKmode) return assign_temp (type, 0, 0); rtx x = gen_reg_rtx (reg_mode); if (POINTER_TYPE_P (type)) mark_reg_pointer (x, TYPE_ALIGN (TREE_TYPE (type))); return x; } /* Insert required copies for T in graph G. Check for a strongly connected region, and create a temporary to break the cycle if one is found. */ static void elim_create (elim_graph *g, int T) { int P, S; location_t locus; if (elim_unvisited_predecessor (g, T)) { tree var = partition_to_var (g->map, T); rtx U = get_temp_reg (var); int unsignedsrcp = TYPE_UNSIGNED (TREE_TYPE (var)); insert_part_to_rtx_on_edge (g->e, U, T, UNKNOWN_LOCATION); FOR_EACH_ELIM_GRAPH_PRED (g, T, P, locus, { if (!bitmap_bit_p (g->visited, P)) { elim_backward (g, P); insert_rtx_to_part_on_edge (g->e, P, U, unsignedsrcp, locus); } }); } else { S = elim_graph_remove_succ_edge (g, T, &locus); if (S != -1) { bitmap_set_bit (g->visited, T); insert_partition_copy_on_edge (g->e, T, S, locus); } } } /* Eliminate all the phi nodes on edge E in graph G. */ static void eliminate_phi (edge e, elim_graph *g) { int x; gcc_assert (g->const_copies.length () == 0); gcc_assert (g->copy_locus.length () == 0); /* Abnormal edges already have everything coalesced. */ if (e->flags & EDGE_ABNORMAL) return; g->e = e; eliminate_build (g); if (elim_graph_size (g) != 0) { int part; bitmap_clear (g->visited); g->stack.truncate (0); FOR_EACH_VEC_ELT (g->nodes, x, part) { if (!bitmap_bit_p (g->visited, part)) elim_forward (g, part); } bitmap_clear (g->visited); while (g->stack.length () > 0) { x = g->stack.pop (); if (!bitmap_bit_p (g->visited, x)) elim_create (g, x); } } /* If there are any pending constant copies, issue them now. */ while (g->const_copies.length () > 0) { int dest; tree src; location_t locus; src = g->const_copies.pop (); dest = g->const_dests.pop (); locus = g->copy_locus.pop (); insert_value_copy_on_edge (e, dest, src, locus); } } /* Remove each argument from PHI. If an arg was the last use of an SSA_NAME, check to see if this allows another PHI node to be removed. */ static void remove_gimple_phi_args (gphi *phi) { use_operand_p arg_p; ssa_op_iter iter; if (dump_file && (dump_flags & TDF_DETAILS)) { fprintf (dump_file, "Removing Dead PHI definition: "); print_gimple_stmt (dump_file, phi, 0, TDF_SLIM); } FOR_EACH_PHI_ARG (arg_p, phi, iter, SSA_OP_USE) { tree arg = USE_FROM_PTR (arg_p); if (TREE_CODE (arg) == SSA_NAME) { /* Remove the reference to the existing argument. */ SET_USE (arg_p, NULL_TREE); if (has_zero_uses (arg)) { gimple *stmt; gimple_stmt_iterator gsi; stmt = SSA_NAME_DEF_STMT (arg); /* Also remove the def if it is a PHI node. */ if (gimple_code (stmt) == GIMPLE_PHI) { remove_gimple_phi_args (as_a (stmt)); gsi = gsi_for_stmt (stmt); remove_phi_node (&gsi, true); } } } } } /* Remove any PHI node which is a virtual PHI, or a PHI with no uses. */ static void eliminate_useless_phis (void) { basic_block bb; gphi_iterator gsi; tree result; FOR_EACH_BB_FN (bb, cfun) { for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); ) { gphi *phi = gsi.phi (); result = gimple_phi_result (phi); if (virtual_operand_p (result)) remove_phi_node (&gsi, true); else { /* Also remove real PHIs with no uses. */ if (has_zero_uses (result)) { remove_gimple_phi_args (phi); remove_phi_node (&gsi, true); } else gsi_next (&gsi); } } } } /* This function will rewrite the current program using the variable mapping found in MAP. If the replacement vector VALUES is provided, any occurrences of partitions with non-null entries in the vector will be replaced with the expression in the vector instead of its mapped variable. */ static void rewrite_trees (var_map map) { if (!flag_checking) return; basic_block bb; /* Search for PHIs where the destination has no partition, but one or more arguments has a partition. This should not happen and can create incorrect code. */ FOR_EACH_BB_FN (bb, cfun) { gphi_iterator gsi; for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); gsi_next (&gsi)) { gphi *phi = gsi.phi (); tree T0 = var_to_partition_to_var (map, gimple_phi_result (phi)); if (T0 == NULL_TREE) { size_t i; for (i = 0; i < gimple_phi_num_args (phi); i++) { tree arg = PHI_ARG_DEF (phi, i); if (TREE_CODE (arg) == SSA_NAME && var_to_partition (map, arg) != NO_PARTITION) { fprintf (stderr, "Argument of PHI is in a partition :("); print_generic_expr (stderr, arg, TDF_SLIM); fprintf (stderr, "), but the result is not :"); print_gimple_stmt (stderr, phi, 0, TDF_SLIM); internal_error ("SSA corruption"); } } } } } } /* Create a default def for VAR. */ static void create_default_def (tree var, void *arg ATTRIBUTE_UNUSED) { if (!is_gimple_reg (var)) return; tree ssa = get_or_create_ssa_default_def (cfun, var); gcc_assert (ssa); } /* Call CALLBACK for all PARM_DECLs and RESULT_DECLs for which assign_parms may ask for a default partition. */ static void for_all_parms (void (*callback)(tree var, void *arg), void *arg) { for (tree var = DECL_ARGUMENTS (current_function_decl); var; var = DECL_CHAIN (var)) callback (var, arg); if (!VOID_TYPE_P (TREE_TYPE (DECL_RESULT (current_function_decl)))) callback (DECL_RESULT (current_function_decl), arg); if (cfun->static_chain_decl) callback (cfun->static_chain_decl, arg); } /* We need to pass two arguments to set_parm_default_def_partition, but for_all_parms only supports one. Use a pair. */ typedef std::pair parm_default_def_partition_arg; /* Set in ARG's PARTS bitmap the bit corresponding to the partition in ARG's MAP containing VAR's default def. */ static void set_parm_default_def_partition (tree var, void *arg_) { parm_default_def_partition_arg *arg = (parm_default_def_partition_arg *)arg_; var_map map = arg->first; bitmap parts = arg->second; if (!is_gimple_reg (var)) return; tree ssa = ssa_default_def (cfun, var); gcc_assert (ssa); int version = var_to_partition (map, ssa); gcc_assert (version != NO_PARTITION); bool changed = bitmap_set_bit (parts, version); gcc_assert (changed); } /* Allocate and return a bitmap that has a bit set for each partition that contains a default def for a parameter. */ static bitmap get_parm_default_def_partitions (var_map map) { bitmap parm_default_def_parts = BITMAP_ALLOC (NULL); parm_default_def_partition_arg arg = std::make_pair (map, parm_default_def_parts); for_all_parms (set_parm_default_def_partition, &arg); return parm_default_def_parts; } /* Allocate and return a bitmap that has a bit set for each partition that contains an undefined value. */ static bitmap get_undefined_value_partitions (var_map map) { bitmap undefined_value_parts = BITMAP_ALLOC (NULL); for (unsigned int i = 1; i < num_ssa_names; i++) { tree var = ssa_name (i); if (var && !virtual_operand_p (var) && !has_zero_uses (var) && ssa_undefined_value_p (var)) { const int p = var_to_partition (map, var); if (p != NO_PARTITION) bitmap_set_bit (undefined_value_parts, p); } } return undefined_value_parts; } /* Given the out-of-ssa info object SA (with prepared partitions) eliminate all phi nodes in all basic blocks. Afterwards no basic block will have phi nodes anymore and there are possibly some RTL instructions inserted on edges. */ void expand_phi_nodes (struct ssaexpand *sa) { basic_block bb; elim_graph g (sa->map); FOR_BB_BETWEEN (bb, ENTRY_BLOCK_PTR_FOR_FN (cfun)->next_bb, EXIT_BLOCK_PTR_FOR_FN (cfun), next_bb) if (!gimple_seq_empty_p (phi_nodes (bb))) { edge e; edge_iterator ei; FOR_EACH_EDGE (e, ei, bb->preds) eliminate_phi (e, &g); set_phi_nodes (bb, NULL); /* We can't redirect EH edges in RTL land, so we need to do this here. Redirection happens only when splitting is necessary, which it is only for critical edges, normally. For EH edges it might also be necessary when the successor has more than one predecessor. In that case the edge is either required to be fallthru (which EH edges aren't), or the predecessor needs to end with a jump (which again, isn't the case with EH edges). Hence, split all EH edges on which we inserted instructions and whose successor has multiple predecessors. */ for (ei = ei_start (bb->preds); (e = ei_safe_edge (ei)); ) { if (e->insns.r && (e->flags & EDGE_EH) && !single_pred_p (e->dest)) { rtx_insn *insns = e->insns.r; basic_block bb; e->insns.r = NULL; bb = split_edge (e); single_pred_edge (bb)->insns.r = insns; } else ei_next (&ei); } } } /* Remove the ssa-names in the current function and translate them into normal compiler variables. PERFORM_TER is true if Temporary Expression Replacement should also be used. */ static void remove_ssa_form (bool perform_ter, struct ssaexpand *sa) { bitmap values = NULL; var_map map; for_all_parms (create_default_def, NULL); map = init_var_map (num_ssa_names); coalesce_ssa_name (map); /* Return to viewing the variable list as just all reference variables after coalescing has been performed. */ partition_view_normal (map); if (dump_file && (dump_flags & TDF_DETAILS)) { fprintf (dump_file, "After Coalescing:\n"); dump_var_map (dump_file, map); } if (perform_ter) { values = find_replaceable_exprs (map); if (values && dump_file && (dump_flags & TDF_DETAILS)) dump_replaceable_exprs (dump_file, values); } rewrite_trees (map); sa->map = map; sa->values = values; sa->partitions_for_parm_default_defs = get_parm_default_def_partitions (map); sa->partitions_for_undefined_values = get_undefined_value_partitions (map); } /* If not already done so for basic block BB, assign increasing uids to each of its instructions. */ static void maybe_renumber_stmts_bb (basic_block bb) { unsigned i = 0; gimple_stmt_iterator gsi; if (!bb->aux) return; bb->aux = NULL; for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi)) { gimple *stmt = gsi_stmt (gsi); gimple_set_uid (stmt, i); i++; } } /* Return true if we can determine that the SSA_NAMEs RESULT (a result of a PHI node) and ARG (one of its arguments) conflict. Return false otherwise, also when we simply aren't sure. */ static bool trivially_conflicts_p (basic_block bb, tree result, tree arg) { use_operand_p use; imm_use_iterator imm_iter; gimple *defa = SSA_NAME_DEF_STMT (arg); /* If ARG isn't defined in the same block it's too complicated for our little mind. */ if (gimple_bb (defa) != bb) return false; FOR_EACH_IMM_USE_FAST (use, imm_iter, result) { gimple *use_stmt = USE_STMT (use); if (is_gimple_debug (use_stmt)) continue; /* Now, if there's a use of RESULT that lies outside this basic block, then there surely is a conflict with ARG. */ if (gimple_bb (use_stmt) != bb) return true; if (gimple_code (use_stmt) == GIMPLE_PHI) continue; /* The use now is in a real stmt of BB, so if ARG was defined in a PHI node (like RESULT) both conflict. */ if (gimple_code (defa) == GIMPLE_PHI) return true; maybe_renumber_stmts_bb (bb); /* If the use of RESULT occurs after the definition of ARG, the two conflict too. */ if (gimple_uid (defa) < gimple_uid (use_stmt)) return true; } return false; } /* Search every PHI node for arguments associated with backedges which we can trivially determine will need a copy (the argument is either not an SSA_NAME or the argument has a different underlying variable than the PHI result). Insert a copy from the PHI argument to a new destination at the end of the block with the backedge to the top of the loop. Update the PHI argument to reference this new destination. */ static void insert_backedge_copies (void) { basic_block bb; gphi_iterator gsi; mark_dfs_back_edges (); FOR_EACH_BB_FN (bb, cfun) { /* Mark block as possibly needing calculation of UIDs. */ bb->aux = &bb->aux; for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); gsi_next (&gsi)) { gphi *phi = gsi.phi (); tree result = gimple_phi_result (phi); size_t i; if (virtual_operand_p (result)) continue; for (i = 0; i < gimple_phi_num_args (phi); i++) { tree arg = gimple_phi_arg_def (phi, i); edge e = gimple_phi_arg_edge (phi, i); /* We are only interested in copies emitted on critical backedges. */ if (!(e->flags & EDGE_DFS_BACK) || !EDGE_CRITICAL_P (e)) continue; /* If the argument is not an SSA_NAME, then we will need a constant initialization. If the argument is an SSA_NAME then a copy statement may be needed. First handle the case where we cannot insert before the argument definition. */ if (TREE_CODE (arg) != SSA_NAME || (gimple_code (SSA_NAME_DEF_STMT (arg)) == GIMPLE_PHI && trivially_conflicts_p (bb, result, arg))) { tree name; gassign *stmt; gimple *last = NULL; gimple_stmt_iterator gsi2; gsi2 = gsi_last_bb (gimple_phi_arg_edge (phi, i)->src); if (!gsi_end_p (gsi2)) last = gsi_stmt (gsi2); /* In theory the only way we ought to get back to the start of a loop should be with a COND_EXPR or GOTO_EXPR. However, better safe than sorry. If the block ends with a control statement or something that might throw, then we have to insert this assignment before the last statement. Else insert it after the last statement. */ if (last && stmt_ends_bb_p (last)) { /* If the last statement in the block is the definition site of the PHI argument, then we can't insert anything after it. */ if (TREE_CODE (arg) == SSA_NAME && SSA_NAME_DEF_STMT (arg) == last) continue; } /* Create a new instance of the underlying variable of the PHI result. */ name = copy_ssa_name (result); stmt = gimple_build_assign (name, gimple_phi_arg_def (phi, i)); /* copy location if present. */ if (gimple_phi_arg_has_location (phi, i)) gimple_set_location (stmt, gimple_phi_arg_location (phi, i)); /* Insert the new statement into the block and update the PHI node. */ if (last && stmt_ends_bb_p (last)) gsi_insert_before (&gsi2, stmt, GSI_NEW_STMT); else gsi_insert_after (&gsi2, stmt, GSI_NEW_STMT); SET_PHI_ARG_DEF (phi, i, name); } /* Insert a copy before the definition of the backedge value and adjust all conflicting uses. */ else if (trivially_conflicts_p (bb, result, arg)) { gimple *def = SSA_NAME_DEF_STMT (arg); if (gimple_nop_p (def) || gimple_code (def) == GIMPLE_PHI) continue; tree name = copy_ssa_name (result); gimple *stmt = gimple_build_assign (name, result); imm_use_iterator imm_iter; gimple *use_stmt; /* The following matches trivially_conflicts_p. */ FOR_EACH_IMM_USE_STMT (use_stmt, imm_iter, result) { if (gimple_bb (use_stmt) != bb || (gimple_code (use_stmt) != GIMPLE_PHI && (maybe_renumber_stmts_bb (bb), true) && gimple_uid (use_stmt) > gimple_uid (def))) { use_operand_p use; FOR_EACH_IMM_USE_ON_STMT (use, imm_iter) SET_USE (use, name); } } gimple_stmt_iterator gsi = gsi_for_stmt (def); gsi_insert_before (&gsi, stmt, GSI_SAME_STMT); } } } /* Unmark this block again. */ bb->aux = NULL; } } /* Free all memory associated with going out of SSA form. SA is the outof-SSA info object. */ void finish_out_of_ssa (struct ssaexpand *sa) { free (sa->partition_to_pseudo); if (sa->values) BITMAP_FREE (sa->values); delete_var_map (sa->map); BITMAP_FREE (sa->partitions_for_parm_default_defs); BITMAP_FREE (sa->partitions_for_undefined_values); memset (sa, 0, sizeof *sa); } /* Take the current function out of SSA form, translating PHIs as described in R. Morgan, ``Building an Optimizing Compiler'', Butterworth-Heinemann, Boston, MA, 1998. pp 176-186. */ unsigned int rewrite_out_of_ssa (struct ssaexpand *sa) { /* If elimination of a PHI requires inserting a copy on a backedge, then we will have to split the backedge which has numerous undesirable performance effects. A significant number of such cases can be handled here by inserting copies into the loop itself. */ insert_backedge_copies (); /* Eliminate PHIs which are of no use, such as virtual or dead phis. */ eliminate_useless_phis (); if (dump_file && (dump_flags & TDF_DETAILS)) gimple_dump_cfg (dump_file, dump_flags & ~TDF_DETAILS); remove_ssa_form (flag_tree_ter, sa); if (dump_file && (dump_flags & TDF_DETAILS)) gimple_dump_cfg (dump_file, dump_flags & ~TDF_DETAILS); return 0; }