diff options
Diffstat (limited to 'gcc/tree-vect-data-refs.c')
-rw-r--r-- | gcc/tree-vect-data-refs.c | 199 |
1 files changed, 103 insertions, 96 deletions
diff --git a/gcc/tree-vect-data-refs.c b/gcc/tree-vect-data-refs.c index 0192aa6..7e9ab3e 100644 --- a/gcc/tree-vect-data-refs.c +++ b/gcc/tree-vect-data-refs.c @@ -695,7 +695,8 @@ vect_slp_analyze_data_ref_dependence (vec_info *vinfo, disambiguating the loads. */ static bool -vect_slp_analyze_node_dependences (slp_instance instance, slp_tree node, +vect_slp_analyze_node_dependences (vec_info *vinfo, + slp_instance instance, slp_tree node, vec<stmt_vec_info> stores, stmt_vec_info last_store_info) { @@ -703,7 +704,6 @@ vect_slp_analyze_node_dependences (slp_instance instance, slp_tree node, in NODE verifying we can sink them up to the last stmt in the group. */ stmt_vec_info last_access_info = vect_find_last_scalar_stmt_in_slp (node); - vec_info *vinfo = last_access_info->vinfo; for (unsigned k = 0; k < SLP_INSTANCE_GROUP_SIZE (instance); ++k) { stmt_vec_info access_info = SLP_TREE_SCALAR_STMTS (node)[k]; @@ -781,7 +781,7 @@ vect_slp_analyze_node_dependences (slp_instance instance, slp_tree node, the maximum vectorization factor the data dependences allow. */ bool -vect_slp_analyze_instance_dependence (slp_instance instance) +vect_slp_analyze_instance_dependence (vec_info *vinfo, slp_instance instance) { DUMP_VECT_SCOPE ("vect_slp_analyze_instance_dependence"); @@ -794,7 +794,8 @@ vect_slp_analyze_instance_dependence (slp_instance instance) stmt_vec_info last_store_info = NULL; if (store) { - if (! vect_slp_analyze_node_dependences (instance, store, vNULL, NULL)) + if (! vect_slp_analyze_node_dependences (vinfo, instance, store, + vNULL, NULL)) return false; /* Mark stores in this instance and remember the last one. */ @@ -810,7 +811,7 @@ vect_slp_analyze_instance_dependence (slp_instance instance) slp_tree load; unsigned int i; FOR_EACH_VEC_ELT (SLP_INSTANCE_LOADS (instance), i, load) - if (! vect_slp_analyze_node_dependences (instance, load, + if (! vect_slp_analyze_node_dependences (vinfo, instance, load, store ? SLP_TREE_SCALAR_STMTS (store) : vNULL, last_store_info)) @@ -831,10 +832,9 @@ vect_slp_analyze_instance_dependence (slp_instance instance) in STMT_INFO. */ static void -vect_record_base_alignment (stmt_vec_info stmt_info, +vect_record_base_alignment (vec_info *vinfo, stmt_vec_info stmt_info, innermost_loop_behavior *drb) { - vec_info *vinfo = stmt_info->vinfo; bool existed; innermost_loop_behavior *&entry = vinfo->base_alignments.get_or_insert (drb->base_address, &existed); @@ -877,13 +877,13 @@ vect_record_base_alignments (vec_info *vinfo) && STMT_VINFO_VECTORIZABLE (stmt_info) && !STMT_VINFO_GATHER_SCATTER_P (stmt_info)) { - vect_record_base_alignment (stmt_info, &DR_INNERMOST (dr)); + vect_record_base_alignment (vinfo, stmt_info, &DR_INNERMOST (dr)); /* If DR is nested in the loop that is being vectorized, we can also record the alignment of the base wrt the outer loop. */ if (loop && nested_in_vect_loop_p (loop, stmt_info)) vect_record_base_alignment - (stmt_info, &STMT_VINFO_DR_WRT_VEC_LOOP (stmt_info)); + (vinfo, stmt_info, &STMT_VINFO_DR_WRT_VEC_LOOP (stmt_info)); } } } @@ -908,11 +908,11 @@ vect_calculate_target_alignment (dr_vec_info *dr_info) only for trivial cases. TODO. */ static void -vect_compute_data_ref_alignment (dr_vec_info *dr_info) +vect_compute_data_ref_alignment (vec_info *vinfo, dr_vec_info *dr_info) { stmt_vec_info stmt_info = dr_info->stmt; - vec_base_alignments *base_alignments = &stmt_info->vinfo->base_alignments; - loop_vec_info loop_vinfo = STMT_VINFO_LOOP_VINFO (stmt_info); + vec_base_alignments *base_alignments = &vinfo->base_alignments; + loop_vec_info loop_vinfo = dyn_cast <loop_vec_info> (vinfo); class loop *loop = NULL; tree ref = DR_REF (dr_info->dr); tree vectype = STMT_VINFO_VECTYPE (stmt_info); @@ -930,7 +930,7 @@ vect_compute_data_ref_alignment (dr_vec_info *dr_info) if (STMT_VINFO_GATHER_SCATTER_P (stmt_info)) return; - innermost_loop_behavior *drb = vect_dr_behavior (dr_info); + innermost_loop_behavior *drb = vect_dr_behavior (vinfo, dr_info); bool step_preserves_misalignment_p; poly_uint64 vector_alignment @@ -1137,10 +1137,10 @@ vect_update_misalignment_for_peel (dr_vec_info *dr_info, Return TRUE if DR_INFO can be handled with respect to alignment. */ static opt_result -verify_data_ref_alignment (dr_vec_info *dr_info) +verify_data_ref_alignment (vec_info *vinfo, dr_vec_info *dr_info) { enum dr_alignment_support supportable_dr_alignment - = vect_supportable_dr_alignment (dr_info, false); + = vect_supportable_dr_alignment (vinfo, dr_info, false); if (!supportable_dr_alignment) return opt_result::failure_at (dr_info->stmt->stmt, @@ -1187,7 +1187,7 @@ vect_verify_datarefs_alignment (loop_vec_info vinfo) && !STMT_VINFO_GROUPED_ACCESS (stmt_info)) continue; - opt_result res = verify_data_ref_alignment (dr_info); + opt_result res = verify_data_ref_alignment (vinfo, dr_info); if (!res) return res; } @@ -1278,14 +1278,14 @@ vector_alignment_reachable_p (dr_vec_info *dr_info) /* Calculate the cost of the memory access represented by DR_INFO. */ static void -vect_get_data_access_cost (dr_vec_info *dr_info, +vect_get_data_access_cost (vec_info *vinfo, dr_vec_info *dr_info, unsigned int *inside_cost, unsigned int *outside_cost, stmt_vector_for_cost *body_cost_vec, stmt_vector_for_cost *prologue_cost_vec) { stmt_vec_info stmt_info = dr_info->stmt; - loop_vec_info loop_vinfo = STMT_VINFO_LOOP_VINFO (stmt_info); + loop_vec_info loop_vinfo = dyn_cast <loop_vec_info> (vinfo); int ncopies; if (PURE_SLP_STMT (stmt_info)) @@ -1294,10 +1294,10 @@ vect_get_data_access_cost (dr_vec_info *dr_info, ncopies = vect_get_num_copies (loop_vinfo, STMT_VINFO_VECTYPE (stmt_info)); if (DR_IS_READ (dr_info->dr)) - vect_get_load_cost (stmt_info, ncopies, true, inside_cost, outside_cost, - prologue_cost_vec, body_cost_vec, false); + vect_get_load_cost (vinfo, stmt_info, ncopies, true, inside_cost, + outside_cost, prologue_cost_vec, body_cost_vec, false); else - vect_get_store_cost (stmt_info, ncopies, inside_cost, body_cost_vec); + vect_get_store_cost (vinfo,stmt_info, ncopies, inside_cost, body_cost_vec); if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, @@ -1315,6 +1315,7 @@ typedef struct _vect_peel_info typedef struct _vect_peel_extended_info { + vec_info *vinfo; struct _vect_peel_info peel_info; unsigned int inside_cost; unsigned int outside_cost; @@ -1352,7 +1353,7 @@ vect_peeling_hash_insert (hash_table<peel_info_hasher> *peeling_htab, struct _vect_peel_info elem, *slot; _vect_peel_info **new_slot; bool supportable_dr_alignment - = vect_supportable_dr_alignment (dr_info, true); + = vect_supportable_dr_alignment (loop_vinfo, dr_info, true); elem.npeel = npeel; slot = peeling_htab->find (&elem); @@ -1440,7 +1441,7 @@ vect_get_peeling_costs_all_drs (loop_vec_info loop_vinfo, SET_DR_MISALIGNMENT (dr_info, 0); else vect_update_misalignment_for_peel (dr_info, dr0_info, npeel); - vect_get_data_access_cost (dr_info, inside_cost, outside_cost, + vect_get_data_access_cost (loop_vinfo, dr_info, inside_cost, outside_cost, body_cost_vec, prologue_cost_vec); SET_DR_MISALIGNMENT (dr_info, save_misalignment); } @@ -1456,8 +1457,7 @@ vect_peeling_hash_get_lowest_cost (_vect_peel_info **slot, vect_peel_info elem = *slot; int dummy; unsigned int inside_cost = 0, outside_cost = 0; - stmt_vec_info stmt_info = elem->dr_info->stmt; - loop_vec_info loop_vinfo = STMT_VINFO_LOOP_VINFO (stmt_info); + loop_vec_info loop_vinfo = dyn_cast <loop_vec_info> (min->vinfo); stmt_vector_for_cost prologue_cost_vec, body_cost_vec, epilogue_cost_vec; @@ -1509,6 +1509,7 @@ vect_peeling_hash_choose_best_peeling (hash_table<peel_info_hasher> *peeling_hta struct _vect_peel_extended_info res; res.peel_info.dr_info = NULL; + res.vinfo = loop_vinfo; if (!unlimited_cost_model (LOOP_VINFO_LOOP (loop_vinfo))) { @@ -1565,7 +1566,7 @@ vect_peeling_supportable (loop_vec_info loop_vinfo, dr_vec_info *dr0_info, save_misalignment = DR_MISALIGNMENT (dr_info); vect_update_misalignment_for_peel (dr_info, dr0_info, npeel); supportable_dr_alignment - = vect_supportable_dr_alignment (dr_info, false); + = vect_supportable_dr_alignment (loop_vinfo, dr_info, false); SET_DR_MISALIGNMENT (dr_info, save_misalignment); if (!supportable_dr_alignment) @@ -1753,7 +1754,8 @@ vect_enhance_data_refs_alignment (loop_vec_info loop_vinfo) && !STMT_VINFO_GROUPED_ACCESS (stmt_info)) continue; - supportable_dr_alignment = vect_supportable_dr_alignment (dr_info, true); + supportable_dr_alignment + = vect_supportable_dr_alignment (loop_vinfo, dr_info, true); do_peeling = vector_alignment_reachable_p (dr_info); if (do_peeling) { @@ -2217,7 +2219,7 @@ vect_enhance_data_refs_alignment (loop_vec_info loop_vinfo) } supportable_dr_alignment - = vect_supportable_dr_alignment (dr_info, false); + = vect_supportable_dr_alignment (loop_vinfo, dr_info, false); if (!supportable_dr_alignment) { @@ -2415,7 +2417,7 @@ vect_analyze_data_refs_alignment (loop_vec_info vinfo) { dr_vec_info *dr_info = vinfo->lookup_dr (dr); if (STMT_VINFO_VECTORIZABLE (dr_info->stmt)) - vect_compute_data_ref_alignment (dr_info); + vect_compute_data_ref_alignment (vinfo, dr_info); } return opt_result::success (); @@ -2425,7 +2427,7 @@ vect_analyze_data_refs_alignment (loop_vec_info vinfo) /* Analyze alignment of DRs of stmts in NODE. */ static bool -vect_slp_analyze_and_verify_node_alignment (slp_tree node) +vect_slp_analyze_and_verify_node_alignment (vec_info *vinfo, slp_tree node) { /* We vectorize from the first scalar stmt in the node unless the node is permuted in which case we start from the first @@ -2436,12 +2438,12 @@ vect_slp_analyze_and_verify_node_alignment (slp_tree node) first_stmt_info = DR_GROUP_FIRST_ELEMENT (first_stmt_info); dr_vec_info *dr_info = STMT_VINFO_DR_INFO (first_stmt_info); - vect_compute_data_ref_alignment (dr_info); + vect_compute_data_ref_alignment (vinfo, dr_info); /* For creating the data-ref pointer we need alignment of the first element anyway. */ if (dr_info != first_dr_info) - vect_compute_data_ref_alignment (first_dr_info); - if (! verify_data_ref_alignment (dr_info)) + vect_compute_data_ref_alignment (vinfo, first_dr_info); + if (! verify_data_ref_alignment (vinfo, dr_info)) { if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, @@ -2459,20 +2461,21 @@ vect_slp_analyze_and_verify_node_alignment (slp_tree node) Return FALSE if a data reference is found that cannot be vectorized. */ bool -vect_slp_analyze_and_verify_instance_alignment (slp_instance instance) +vect_slp_analyze_and_verify_instance_alignment (vec_info *vinfo, + slp_instance instance) { DUMP_VECT_SCOPE ("vect_slp_analyze_and_verify_instance_alignment"); slp_tree node; unsigned i; FOR_EACH_VEC_ELT (SLP_INSTANCE_LOADS (instance), i, node) - if (! vect_slp_analyze_and_verify_node_alignment (node)) + if (! vect_slp_analyze_and_verify_node_alignment (vinfo, node)) return false; node = SLP_INSTANCE_TREE (instance); if (STMT_VINFO_DATA_REF (SLP_TREE_SCALAR_STMTS (node)[0]) && ! vect_slp_analyze_and_verify_node_alignment - (SLP_INSTANCE_TREE (instance))) + (vinfo, SLP_INSTANCE_TREE (instance))) return false; return true; @@ -2486,15 +2489,15 @@ vect_slp_analyze_and_verify_instance_alignment (slp_instance instance) Worker for vect_analyze_group_access. */ static bool -vect_analyze_group_access_1 (dr_vec_info *dr_info) +vect_analyze_group_access_1 (vec_info *vinfo, dr_vec_info *dr_info) { data_reference *dr = dr_info->dr; tree step = DR_STEP (dr); tree scalar_type = TREE_TYPE (DR_REF (dr)); HOST_WIDE_INT type_size = TREE_INT_CST_LOW (TYPE_SIZE_UNIT (scalar_type)); stmt_vec_info stmt_info = dr_info->stmt; - loop_vec_info loop_vinfo = STMT_VINFO_LOOP_VINFO (stmt_info); - bb_vec_info bb_vinfo = STMT_VINFO_BB_VINFO (stmt_info); + loop_vec_info loop_vinfo = dyn_cast <loop_vec_info> (vinfo); + bb_vec_info bb_vinfo = dyn_cast <bb_vec_info> (vinfo); HOST_WIDE_INT dr_step = -1; HOST_WIDE_INT groupsize, last_accessed_element = 1; bool slp_impossible = false; @@ -2696,9 +2699,9 @@ vect_analyze_group_access_1 (dr_vec_info *dr_info) Collect groups of strided stores for further use in SLP analysis. */ static bool -vect_analyze_group_access (dr_vec_info *dr_info) +vect_analyze_group_access (vec_info *vinfo, dr_vec_info *dr_info) { - if (!vect_analyze_group_access_1 (dr_info)) + if (!vect_analyze_group_access_1 (vinfo, dr_info)) { /* Dissolve the group if present. */ stmt_vec_info stmt_info = DR_GROUP_FIRST_ELEMENT (dr_info->stmt); @@ -2719,13 +2722,13 @@ vect_analyze_group_access (dr_vec_info *dr_info) analyze groups of accesses. */ static bool -vect_analyze_data_ref_access (dr_vec_info *dr_info) +vect_analyze_data_ref_access (vec_info *vinfo, dr_vec_info *dr_info) { data_reference *dr = dr_info->dr; tree step = DR_STEP (dr); tree scalar_type = TREE_TYPE (DR_REF (dr)); stmt_vec_info stmt_info = dr_info->stmt; - loop_vec_info loop_vinfo = STMT_VINFO_LOOP_VINFO (stmt_info); + loop_vec_info loop_vinfo = dyn_cast <loop_vec_info> (vinfo); class loop *loop = NULL; if (STMT_VINFO_GATHER_SCATTER_P (stmt_info)) @@ -2804,10 +2807,10 @@ vect_analyze_data_ref_access (dr_vec_info *dr_info) if (TREE_CODE (step) != INTEGER_CST) return (STMT_VINFO_STRIDED_P (stmt_info) && (!STMT_VINFO_GROUPED_ACCESS (stmt_info) - || vect_analyze_group_access (dr_info))); + || vect_analyze_group_access (vinfo, dr_info))); /* Not consecutive access - check if it's a part of interleaving group. */ - return vect_analyze_group_access (dr_info); + return vect_analyze_group_access (vinfo, dr_info); } /* Compare two data-references DRA and DRB to group them into chunks @@ -3153,7 +3156,7 @@ vect_analyze_data_ref_accesses (vec_info *vinfo) { dr_vec_info *dr_info = vinfo->lookup_dr (dr); if (STMT_VINFO_VECTORIZABLE (dr_info->stmt) - && !vect_analyze_data_ref_access (dr_info)) + && !vect_analyze_data_ref_access (vinfo, dr_info)) { if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, @@ -3204,7 +3207,7 @@ vect_vfa_segment_size (dr_vec_info *dr_info, tree length_factor) gives the worst-case number of bytes covered by the segment. */ static unsigned HOST_WIDE_INT -vect_vfa_access_size (dr_vec_info *dr_info) +vect_vfa_access_size (vec_info *vinfo, dr_vec_info *dr_info) { stmt_vec_info stmt_vinfo = dr_info->stmt; tree ref_type = TREE_TYPE (DR_REF (dr_info->dr)); @@ -3216,7 +3219,7 @@ vect_vfa_access_size (dr_vec_info *dr_info) access_size *= DR_GROUP_SIZE (stmt_vinfo) - DR_GROUP_GAP (stmt_vinfo); } if (STMT_VINFO_VEC_STMT (stmt_vinfo) - && (vect_supportable_dr_alignment (dr_info, false) + && (vect_supportable_dr_alignment (vinfo, dr_info, false) == dr_explicit_realign_optimized)) { /* We might access a full vector's worth. */ @@ -3592,8 +3595,8 @@ vect_prune_runtime_alias_test_list (loop_vec_info loop_vinfo) segment_length_a = vect_vfa_segment_size (dr_info_a, length_factor); segment_length_b = vect_vfa_segment_size (dr_info_b, length_factor); } - access_size_a = vect_vfa_access_size (dr_info_a); - access_size_b = vect_vfa_access_size (dr_info_b); + access_size_a = vect_vfa_access_size (loop_vinfo, dr_info_a); + access_size_b = vect_vfa_access_size (loop_vinfo, dr_info_b); align_a = vect_vfa_align (dr_info_a); align_b = vect_vfa_align (dr_info_b); @@ -4580,7 +4583,7 @@ vect_duplicate_ssa_name_ptr_info (tree name, dr_vec_info *dr_info) FORNOW: We are only handling array accesses with step 1. */ tree -vect_create_addr_base_for_vector_ref (stmt_vec_info stmt_info, +vect_create_addr_base_for_vector_ref (vec_info *vinfo, stmt_vec_info stmt_info, gimple_seq *new_stmt_list, tree offset, tree byte_offset) @@ -4593,11 +4596,11 @@ vect_create_addr_base_for_vector_ref (stmt_vec_info stmt_info, gimple_seq seq = NULL; tree vect_ptr_type; tree step = TYPE_SIZE_UNIT (TREE_TYPE (DR_REF (dr))); - loop_vec_info loop_vinfo = STMT_VINFO_LOOP_VINFO (stmt_info); - innermost_loop_behavior *drb = vect_dr_behavior (dr_info); + loop_vec_info loop_vinfo = dyn_cast <loop_vec_info> (vinfo); + innermost_loop_behavior *drb = vect_dr_behavior (vinfo, dr_info); tree data_ref_base = unshare_expr (drb->base_address); - tree base_offset = unshare_expr (get_dr_vinfo_offset (dr_info, true)); + tree base_offset = unshare_expr (get_dr_vinfo_offset (vinfo, dr_info, true)); tree init = unshare_expr (drb->init); if (loop_vinfo) @@ -4714,14 +4717,14 @@ vect_create_addr_base_for_vector_ref (stmt_vec_info stmt_info, 3. Return the pointer. */ tree -vect_create_data_ref_ptr (stmt_vec_info stmt_info, tree aggr_type, - class loop *at_loop, tree offset, +vect_create_data_ref_ptr (vec_info *vinfo, stmt_vec_info stmt_info, + tree aggr_type, class loop *at_loop, tree offset, tree *initial_address, gimple_stmt_iterator *gsi, gimple **ptr_incr, bool only_init, tree byte_offset, tree iv_step) { const char *base_name; - loop_vec_info loop_vinfo = STMT_VINFO_LOOP_VINFO (stmt_info); + loop_vec_info loop_vinfo = dyn_cast <loop_vec_info> (vinfo); class loop *loop = NULL; bool nested_in_vect_loop = false; class loop *containing_loop = NULL; @@ -4739,7 +4742,7 @@ vect_create_data_ref_ptr (stmt_vec_info stmt_info, tree aggr_type, bool insert_after; tree indx_before_incr, indx_after_incr; gimple *incr; - bb_vec_info bb_vinfo = STMT_VINFO_BB_VINFO (stmt_info); + bb_vec_info bb_vinfo = dyn_cast <bb_vec_info> (vinfo); gcc_assert (iv_step != NULL_TREE || TREE_CODE (aggr_type) == ARRAY_TYPE @@ -4848,7 +4851,8 @@ vect_create_data_ref_ptr (stmt_vec_info stmt_info, tree aggr_type, /* Create: (&(base[init_val+offset]+byte_offset) in the loop preheader. */ - new_temp = vect_create_addr_base_for_vector_ref (stmt_info, &new_stmt_list, + new_temp = vect_create_addr_base_for_vector_ref (vinfo, + stmt_info, &new_stmt_list, offset, byte_offset); if (new_stmt_list) { @@ -4875,7 +4879,7 @@ vect_create_data_ref_ptr (stmt_vec_info stmt_info, tree aggr_type, { /* Accesses to invariant addresses should be handled specially by the caller. */ - tree step = vect_dr_behavior (dr_info)->step; + tree step = vect_dr_behavior (vinfo, dr_info)->step; gcc_assert (!integer_zerop (step)); if (iv_step == NULL_TREE) @@ -4977,7 +4981,8 @@ vect_create_data_ref_ptr (stmt_vec_info stmt_info, tree aggr_type, */ tree -bump_vector_ptr (tree dataref_ptr, gimple *ptr_incr, gimple_stmt_iterator *gsi, +bump_vector_ptr (vec_info *vinfo, + tree dataref_ptr, gimple *ptr_incr, gimple_stmt_iterator *gsi, stmt_vec_info stmt_info, tree bump) { struct data_reference *dr = STMT_VINFO_DATA_REF (stmt_info); @@ -4997,7 +5002,7 @@ bump_vector_ptr (tree dataref_ptr, gimple *ptr_incr, gimple_stmt_iterator *gsi, new_dataref_ptr = make_ssa_name (TREE_TYPE (dataref_ptr)); incr_stmt = gimple_build_assign (new_dataref_ptr, POINTER_PLUS_EXPR, dataref_ptr, update); - vect_finish_stmt_generation (stmt_info, incr_stmt, gsi); + vect_finish_stmt_generation (vinfo, stmt_info, incr_stmt, gsi); /* Copy the points-to information if it exists. */ if (DR_PTR_INFO (dr)) @@ -5277,7 +5282,7 @@ vect_store_lanes_supported (tree vectype, unsigned HOST_WIDE_INT count, I4: 6 14 22 30 7 15 23 31. */ void -vect_permute_store_chain (vec<tree> dr_chain, +vect_permute_store_chain (vec_info *vinfo, vec<tree> dr_chain, unsigned int length, stmt_vec_info stmt_info, gimple_stmt_iterator *gsi, @@ -5344,7 +5349,7 @@ vect_permute_store_chain (vec<tree> dr_chain, data_ref = make_temp_ssa_name (vectype, NULL, "vect_shuffle3_low"); perm_stmt = gimple_build_assign (data_ref, VEC_PERM_EXPR, vect1, vect2, perm3_mask_low); - vect_finish_stmt_generation (stmt_info, perm_stmt, gsi); + vect_finish_stmt_generation (vinfo, stmt_info, perm_stmt, gsi); vect1 = data_ref; vect2 = dr_chain[2]; @@ -5355,7 +5360,7 @@ vect_permute_store_chain (vec<tree> dr_chain, data_ref = make_temp_ssa_name (vectype, NULL, "vect_shuffle3_high"); perm_stmt = gimple_build_assign (data_ref, VEC_PERM_EXPR, vect1, vect2, perm3_mask_high); - vect_finish_stmt_generation (stmt_info, perm_stmt, gsi); + vect_finish_stmt_generation (vinfo, stmt_info, perm_stmt, gsi); (*result_chain)[j] = data_ref; } } @@ -5394,7 +5399,7 @@ vect_permute_store_chain (vec<tree> dr_chain, high = make_temp_ssa_name (vectype, NULL, "vect_inter_high"); perm_stmt = gimple_build_assign (high, VEC_PERM_EXPR, vect1, vect2, perm_mask_high); - vect_finish_stmt_generation (stmt_info, perm_stmt, gsi); + vect_finish_stmt_generation (vinfo, stmt_info, perm_stmt, gsi); (*result_chain)[2*j] = high; /* Create interleaving stmt: @@ -5404,7 +5409,7 @@ vect_permute_store_chain (vec<tree> dr_chain, low = make_temp_ssa_name (vectype, NULL, "vect_inter_low"); perm_stmt = gimple_build_assign (low, VEC_PERM_EXPR, vect1, vect2, perm_mask_low); - vect_finish_stmt_generation (stmt_info, perm_stmt, gsi); + vect_finish_stmt_generation (vinfo, stmt_info, perm_stmt, gsi); (*result_chain)[2*j+1] = low; } memcpy (dr_chain.address (), result_chain->address (), @@ -5465,14 +5470,14 @@ vect_permute_store_chain (vec<tree> dr_chain, Return value - the result of the loop-header phi node. */ tree -vect_setup_realignment (stmt_vec_info stmt_info, gimple_stmt_iterator *gsi, - tree *realignment_token, +vect_setup_realignment (vec_info *vinfo, stmt_vec_info stmt_info, + gimple_stmt_iterator *gsi, tree *realignment_token, enum dr_alignment_support alignment_support_scheme, tree init_addr, class loop **at_loop) { tree vectype = STMT_VINFO_VECTYPE (stmt_info); - loop_vec_info loop_vinfo = STMT_VINFO_LOOP_VINFO (stmt_info); + loop_vec_info loop_vinfo = dyn_cast <loop_vec_info> (vinfo); dr_vec_info *dr_info = STMT_VINFO_DR_INFO (stmt_info); struct data_reference *dr = dr_info->dr; class loop *loop = NULL; @@ -5579,7 +5584,7 @@ vect_setup_realignment (stmt_vec_info stmt_info, gimple_stmt_iterator *gsi, gcc_assert (!compute_in_loop); vec_dest = vect_create_destination_var (scalar_dest, vectype); - ptr = vect_create_data_ref_ptr (stmt_info, vectype, + ptr = vect_create_data_ref_ptr (vinfo, stmt_info, vectype, loop_for_initial_load, NULL_TREE, &init_addr, NULL, &inc, true); if (TREE_CODE (ptr) == SSA_NAME) @@ -5626,7 +5631,8 @@ vect_setup_realignment (stmt_vec_info stmt_info, gimple_stmt_iterator *gsi, if (!init_addr) { /* Generate the INIT_ADDR computation outside LOOP. */ - init_addr = vect_create_addr_base_for_vector_ref (stmt_info, &stmts, + init_addr = vect_create_addr_base_for_vector_ref (vinfo, + stmt_info, &stmts, NULL_TREE); if (loop) { @@ -5900,7 +5906,7 @@ vect_load_lanes_supported (tree vectype, unsigned HOST_WIDE_INT count, 4th vec (E4): 3 7 11 15 19 23 27 31. */ static void -vect_permute_load_chain (vec<tree> dr_chain, +vect_permute_load_chain (vec_info *vinfo, vec<tree> dr_chain, unsigned int length, stmt_vec_info stmt_info, gimple_stmt_iterator *gsi, @@ -5953,7 +5959,7 @@ vect_permute_load_chain (vec<tree> dr_chain, data_ref = make_temp_ssa_name (vectype, NULL, "vect_shuffle3_low"); perm_stmt = gimple_build_assign (data_ref, VEC_PERM_EXPR, first_vect, second_vect, perm3_mask_low); - vect_finish_stmt_generation (stmt_info, perm_stmt, gsi); + vect_finish_stmt_generation (vinfo, stmt_info, perm_stmt, gsi); /* Create interleaving stmt (high part of): high = VEC_PERM_EXPR <first_vect, second_vect2, {k, 3 + k, 6 + k, @@ -5963,7 +5969,7 @@ vect_permute_load_chain (vec<tree> dr_chain, data_ref = make_temp_ssa_name (vectype, NULL, "vect_shuffle3_high"); perm_stmt = gimple_build_assign (data_ref, VEC_PERM_EXPR, first_vect, second_vect, perm3_mask_high); - vect_finish_stmt_generation (stmt_info, perm_stmt, gsi); + vect_finish_stmt_generation (vinfo, stmt_info, perm_stmt, gsi); (*result_chain)[k] = data_ref; } } @@ -5998,7 +6004,7 @@ vect_permute_load_chain (vec<tree> dr_chain, perm_stmt = gimple_build_assign (data_ref, VEC_PERM_EXPR, first_vect, second_vect, perm_mask_even); - vect_finish_stmt_generation (stmt_info, perm_stmt, gsi); + vect_finish_stmt_generation (vinfo, stmt_info, perm_stmt, gsi); (*result_chain)[j/2] = data_ref; /* data_ref = permute_odd (first_data_ref, second_data_ref); */ @@ -6006,7 +6012,7 @@ vect_permute_load_chain (vec<tree> dr_chain, perm_stmt = gimple_build_assign (data_ref, VEC_PERM_EXPR, first_vect, second_vect, perm_mask_odd); - vect_finish_stmt_generation (stmt_info, perm_stmt, gsi); + vect_finish_stmt_generation (vinfo, stmt_info, perm_stmt, gsi); (*result_chain)[j/2+length/2] = data_ref; } memcpy (dr_chain.address (), result_chain->address (), @@ -6103,7 +6109,7 @@ vect_permute_load_chain (vec<tree> dr_chain, */ static bool -vect_shift_permute_load_chain (vec<tree> dr_chain, +vect_shift_permute_load_chain (vec_info *vinfo, vec<tree> dr_chain, unsigned int length, stmt_vec_info stmt_info, gimple_stmt_iterator *gsi, @@ -6116,7 +6122,7 @@ vect_shift_permute_load_chain (vec<tree> dr_chain, tree vectype = STMT_VINFO_VECTYPE (stmt_info); unsigned int i; - loop_vec_info loop_vinfo = STMT_VINFO_LOOP_VINFO (stmt_info); + loop_vec_info loop_vinfo = dyn_cast <loop_vec_info> (vinfo); unsigned HOST_WIDE_INT nelt, vf; if (!TYPE_VECTOR_SUBPARTS (vectype).is_constant (&nelt) @@ -6205,26 +6211,26 @@ vect_shift_permute_load_chain (vec<tree> dr_chain, perm_stmt = gimple_build_assign (data_ref, VEC_PERM_EXPR, first_vect, first_vect, perm2_mask1); - vect_finish_stmt_generation (stmt_info, perm_stmt, gsi); + vect_finish_stmt_generation (vinfo, stmt_info, perm_stmt, gsi); vect[0] = data_ref; data_ref = make_temp_ssa_name (vectype, NULL, "vect_shuffle2"); perm_stmt = gimple_build_assign (data_ref, VEC_PERM_EXPR, second_vect, second_vect, perm2_mask2); - vect_finish_stmt_generation (stmt_info, perm_stmt, gsi); + vect_finish_stmt_generation (vinfo, stmt_info, perm_stmt, gsi); vect[1] = data_ref; data_ref = make_temp_ssa_name (vectype, NULL, "vect_shift"); perm_stmt = gimple_build_assign (data_ref, VEC_PERM_EXPR, vect[0], vect[1], shift1_mask); - vect_finish_stmt_generation (stmt_info, perm_stmt, gsi); + vect_finish_stmt_generation (vinfo, stmt_info, perm_stmt, gsi); (*result_chain)[j/2 + length/2] = data_ref; data_ref = make_temp_ssa_name (vectype, NULL, "vect_select"); perm_stmt = gimple_build_assign (data_ref, VEC_PERM_EXPR, vect[0], vect[1], select_mask); - vect_finish_stmt_generation (stmt_info, perm_stmt, gsi); + vect_finish_stmt_generation (vinfo, stmt_info, perm_stmt, gsi); (*result_chain)[j/2] = data_ref; } memcpy (dr_chain.address (), result_chain->address (), @@ -6321,7 +6327,7 @@ vect_shift_permute_load_chain (vec<tree> dr_chain, perm_stmt = gimple_build_assign (data_ref, VEC_PERM_EXPR, dr_chain[k], dr_chain[k], perm3_mask); - vect_finish_stmt_generation (stmt_info, perm_stmt, gsi); + vect_finish_stmt_generation (vinfo, stmt_info, perm_stmt, gsi); vect[k] = data_ref; } @@ -6331,7 +6337,7 @@ vect_shift_permute_load_chain (vec<tree> dr_chain, perm_stmt = gimple_build_assign (data_ref, VEC_PERM_EXPR, vect[k % 3], vect[(k + 1) % 3], shift1_mask); - vect_finish_stmt_generation (stmt_info, perm_stmt, gsi); + vect_finish_stmt_generation (vinfo, stmt_info, perm_stmt, gsi); vect_shift[k] = data_ref; } @@ -6342,7 +6348,7 @@ vect_shift_permute_load_chain (vec<tree> dr_chain, vect_shift[(4 - k) % 3], vect_shift[(3 - k) % 3], shift2_mask); - vect_finish_stmt_generation (stmt_info, perm_stmt, gsi); + vect_finish_stmt_generation (vinfo, stmt_info, perm_stmt, gsi); vect[k] = data_ref; } @@ -6351,13 +6357,13 @@ vect_shift_permute_load_chain (vec<tree> dr_chain, data_ref = make_temp_ssa_name (vectype, NULL, "vect_shift3"); perm_stmt = gimple_build_assign (data_ref, VEC_PERM_EXPR, vect[0], vect[0], shift3_mask); - vect_finish_stmt_generation (stmt_info, perm_stmt, gsi); + vect_finish_stmt_generation (vinfo, stmt_info, perm_stmt, gsi); (*result_chain)[nelt % 3] = data_ref; data_ref = make_temp_ssa_name (vectype, NULL, "vect_shift4"); perm_stmt = gimple_build_assign (data_ref, VEC_PERM_EXPR, vect[1], vect[1], shift4_mask); - vect_finish_stmt_generation (stmt_info, perm_stmt, gsi); + vect_finish_stmt_generation (vinfo, stmt_info, perm_stmt, gsi); (*result_chain)[0] = data_ref; return true; } @@ -6372,7 +6378,8 @@ vect_shift_permute_load_chain (vec<tree> dr_chain, */ void -vect_transform_grouped_load (stmt_vec_info stmt_info, vec<tree> dr_chain, +vect_transform_grouped_load (vec_info *vinfo, stmt_vec_info stmt_info, + vec<tree> dr_chain, int size, gimple_stmt_iterator *gsi) { machine_mode mode; @@ -6389,10 +6396,11 @@ vect_transform_grouped_load (stmt_vec_info stmt_info, vec<tree> dr_chain, mode = TYPE_MODE (STMT_VINFO_VECTYPE (stmt_info)); if (targetm.sched.reassociation_width (VEC_PERM_EXPR, mode) > 1 || pow2p_hwi (size) - || !vect_shift_permute_load_chain (dr_chain, size, stmt_info, + || !vect_shift_permute_load_chain (vinfo, dr_chain, size, stmt_info, gsi, &result_chain)) - vect_permute_load_chain (dr_chain, size, stmt_info, gsi, &result_chain); - vect_record_grouped_load_vectors (stmt_info, result_chain); + vect_permute_load_chain (vinfo, dr_chain, + size, stmt_info, gsi, &result_chain); + vect_record_grouped_load_vectors (vinfo, stmt_info, result_chain); result_chain.release (); } @@ -6401,10 +6409,9 @@ vect_transform_grouped_load (stmt_vec_info stmt_info, vec<tree> dr_chain, for each vector to the associated scalar statement. */ void -vect_record_grouped_load_vectors (stmt_vec_info stmt_info, +vect_record_grouped_load_vectors (vec_info *vinfo, stmt_vec_info stmt_info, vec<tree> result_chain) { - vec_info *vinfo = stmt_info->vinfo; stmt_vec_info first_stmt_info = DR_GROUP_FIRST_ELEMENT (stmt_info); unsigned int i, gap_count; tree tmp_data_ref; @@ -6493,14 +6500,14 @@ vect_can_force_dr_alignment_p (const_tree decl, poly_uint64 alignment) alignment. */ enum dr_alignment_support -vect_supportable_dr_alignment (dr_vec_info *dr_info, +vect_supportable_dr_alignment (vec_info *vinfo, dr_vec_info *dr_info, bool check_aligned_accesses) { data_reference *dr = dr_info->dr; stmt_vec_info stmt_info = dr_info->stmt; tree vectype = STMT_VINFO_VECTYPE (stmt_info); machine_mode mode = TYPE_MODE (vectype); - loop_vec_info loop_vinfo = STMT_VINFO_LOOP_VINFO (stmt_info); + loop_vec_info loop_vinfo = dyn_cast <loop_vec_info> (vinfo); class loop *vect_loop = NULL; bool nested_in_vect_loop = false; |