/* Conversion of SESE regions to Polyhedra.
- Copyright (C) 2009 Free Software Foundation, Inc.
+ Copyright (C) 2009, 2010 Free Software Foundation, Inc.
Contributed by Sebastian Pop <sebastian.pop@amd.com>.
This file is part of GCC.
if (simple_copy_phi_p (phi))
{
- /* FIXME: PRE introduces phi nodes like these, for an example,
+ /* PRE introduces phi nodes like these, for an example,
see id-5.f in the fortran graphite testsuite:
# prephitmp.85_265 = PHI <prephitmp.85_258(33), prephitmp.85_265(18)>
ppl_delete_Coefficient (coef);
}
-/* Saves in NV at index I a new name for variable P. */
-
-static void
-save_var_name (char **nv, int i, tree p)
-{
- const char *name = get_name (SSA_NAME_VAR (p));
-
- if (name)
- {
- int len = strlen (name) + 16;
- nv[i] = XNEWVEC (char, len);
- snprintf (nv[i], len, "%s_%d", name, SSA_NAME_VERSION (p));
- }
- else
- {
- nv[i] = XNEWVEC (char, 16);
- snprintf (nv[i], 2 + 16, "T_%d", SSA_NAME_VERSION (p));
- }
-}
-
/* When parameter NAME is in REGION, returns its index in SESE_PARAMS.
Otherwise returns -1. */
gcc_assert (SESE_ADD_PARAMS (region));
i = VEC_length (tree, SESE_PARAMS (region));
- save_var_name (SESE_PARAMS_NAMES (region), i, name);
- save_clast_name_index (SESE_PARAMS_INDEX (region),
- SESE_PARAMS_NAMES (region)[i], i);
VEC_safe_push (tree, heap, SESE_PARAMS (region), name);
return i;
}
return (gimple_bb_p) bb->aux;
}
+/* Insert in the SCOP context constraints from the estimation of the
+ number of iterations. UB_EXPR is a linear expression describing
+ the number of iterations in a loop. This expression is bounded by
+ the estimation NIT. */
+
+static void
+add_upper_bounds_from_estimated_nit (scop_p scop, double_int nit,
+ ppl_dimension_type dim,
+ ppl_Linear_Expression_t ub_expr)
+{
+ Value val;
+ ppl_Linear_Expression_t nb_iters_le;
+ ppl_Polyhedron_t pol;
+ ppl_Coefficient_t coef;
+ ppl_Constraint_t ub;
+
+ ppl_new_Linear_Expression_with_dimension (&ub_expr, dim);
+ ppl_new_C_Polyhedron_from_space_dimension (&pol, dim, 0);
+ ppl_new_Linear_Expression_from_Linear_Expression (&nb_iters_le,
+ ub_expr);
+
+ /* Construct the negated number of last iteration in VAL. */
+ value_init (val);
+ mpz_set_double_int (val, nit, false);
+ value_sub_int (val, val, 1);
+ value_oppose (val, val);
+
+ /* NB_ITERS_LE holds the number of last iteration in
+ parametrical form. Subtract estimated number of last
+ iteration and assert that result is not positive. */
+ ppl_new_Coefficient_from_mpz_t (&coef, val);
+ ppl_Linear_Expression_add_to_inhomogeneous (nb_iters_le, coef);
+ ppl_delete_Coefficient (coef);
+ ppl_new_Constraint (&ub, nb_iters_le,
+ PPL_CONSTRAINT_TYPE_LESS_OR_EQUAL);
+ ppl_Polyhedron_add_constraint (pol, ub);
+
+ /* Remove all but last GDIM dimensions from POL to obtain
+ only the constraints on the parameters. */
+ {
+ graphite_dim_t gdim = scop_nb_params (scop);
+ ppl_dimension_type *dims = XNEWVEC (ppl_dimension_type, dim - gdim);
+ graphite_dim_t i;
+
+ for (i = 0; i < dim - gdim; i++)
+ dims[i] = i;
+
+ ppl_Polyhedron_remove_space_dimensions (pol, dims, dim - gdim);
+ XDELETEVEC (dims);
+ }
+
+ /* Add the constraints on the parameters to the SCoP context. */
+ {
+ ppl_Pointset_Powerset_C_Polyhedron_t constraints_ps;
+
+ ppl_new_Pointset_Powerset_C_Polyhedron_from_C_Polyhedron
+ (&constraints_ps, pol);
+ ppl_Pointset_Powerset_C_Polyhedron_intersection_assign
+ (SCOP_CONTEXT (scop), constraints_ps);
+ ppl_delete_Pointset_Powerset_C_Polyhedron (constraints_ps);
+ }
+
+ ppl_delete_Polyhedron (pol);
+ ppl_delete_Linear_Expression (nb_iters_le);
+ ppl_delete_Constraint (ub);
+ value_clear (val);
+}
+
/* Builds the constraint polyhedra for LOOP in SCOP. OUTER_PH gives
the constraints for the surrounding loops. */
static void
build_loop_iteration_domains (scop_p scop, struct loop *loop,
- ppl_Polyhedron_t outer_ph, int nb)
+ ppl_Polyhedron_t outer_ph, int nb,
+ ppl_Pointset_Powerset_C_Polyhedron_t *domains)
{
int i;
ppl_Polyhedron_t ph;
scan_tree_for_params (SCOP_REGION (scop), nb_iters, ub_expr, one);
value_clear (one);
- /* N <= estimated_nb_iters
-
- FIXME: This is a workaround that should go away once we will
- have the PIP algorithm. */
if (estimated_loop_iterations (loop, true, &nit))
- {
- Value val;
- ppl_Linear_Expression_t nb_iters_le;
- ppl_Polyhedron_t pol;
- graphite_dim_t n = scop_nb_params (scop);
- ppl_Coefficient_t coef;
-
- ppl_new_C_Polyhedron_from_space_dimension (&pol, dim, 0);
- ppl_new_Linear_Expression_from_Linear_Expression (&nb_iters_le,
- ub_expr);
-
- /* Construct the negated number of last iteration in VAL. */
- value_init (val);
- mpz_set_double_int (val, nit, false);
- value_sub_int (val, val, 1);
- value_oppose (val, val);
-
- /* NB_ITERS_LE holds number of last iteration in parametrical form.
- Subtract estimated number of last iteration and assert that result
- is not positive. */
- ppl_new_Coefficient_from_mpz_t (&coef, val);
- ppl_Linear_Expression_add_to_inhomogeneous (nb_iters_le, coef);
- ppl_delete_Coefficient (coef);
- ppl_new_Constraint (&ub, nb_iters_le,
- PPL_CONSTRAINT_TYPE_LESS_OR_EQUAL);
- ppl_Polyhedron_add_constraint (pol, ub);
-
- /* Remove all but last N dimensions from POL to obtain constraints
- on parameters. */
- {
- ppl_dimension_type *dims = XNEWVEC (ppl_dimension_type, dim - n);
- graphite_dim_t i;
- for (i = 0; i < dim - n; i++)
- dims[i] = i;
- ppl_Polyhedron_remove_space_dimensions (pol, dims, dim - n);
- XDELETEVEC (dims);
- }
-
- /* Add constraints on parameters to SCoP context. */
- {
- ppl_Pointset_Powerset_C_Polyhedron_t constraints_ps;
- ppl_new_Pointset_Powerset_C_Polyhedron_from_C_Polyhedron
- (&constraints_ps, pol);
- ppl_Pointset_Powerset_C_Polyhedron_intersection_assign
- (SCOP_CONTEXT (scop), constraints_ps);
- ppl_delete_Pointset_Powerset_C_Polyhedron (constraints_ps);
- }
-
- ppl_delete_Polyhedron (pol);
- ppl_delete_Linear_Expression (nb_iters_le);
- ppl_delete_Constraint (ub);
- value_clear (val);
- }
+ add_upper_bounds_from_estimated_nit (scop, nit, dim, ub_expr);
/* loop_i <= expr_nb_iters */
ppl_set_coef (ub_expr, nb, -1);
gcc_unreachable ();
if (loop->inner && loop_in_sese_p (loop->inner, region))
- build_loop_iteration_domains (scop, loop->inner, ph, nb + 1);
+ build_loop_iteration_domains (scop, loop->inner, ph, nb + 1, domains);
if (nb != 0
&& loop->next
&& loop_in_sese_p (loop->next, region))
- build_loop_iteration_domains (scop, loop->next, outer_ph, nb);
+ build_loop_iteration_domains (scop, loop->next, outer_ph, nb, domains);
ppl_new_Pointset_Powerset_C_Polyhedron_from_C_Polyhedron
- ((ppl_Pointset_Powerset_C_Polyhedron_t *) &loop->aux, ph);
+ (&domains[loop->num], ph);
ppl_delete_Polyhedron (ph);
}
ppl_Linear_Expression_t le;
tree parameter = VEC_index (tree, SESE_PARAMS (SCOP_REGION (scop)), p);
tree type = TREE_TYPE (parameter);
- tree lb, ub;
+ tree lb = NULL_TREE;
+ tree ub = NULL_TREE;
- /* Disabled until we fix CPU2006. */
- return;
-
- if (!INTEGRAL_TYPE_P (type))
- return;
+ if (POINTER_TYPE_P (type) || !TYPE_MIN_VALUE (type))
+ lb = lower_bound_in_type (type, type);
+ else
+ lb = TYPE_MIN_VALUE (type);
- lb = TYPE_MIN_VALUE (type);
- ub = TYPE_MAX_VALUE (type);
+ if (POINTER_TYPE_P (type) || !TYPE_MAX_VALUE (type))
+ ub = upper_bound_in_type (type, type);
+ else
+ ub = TYPE_MAX_VALUE (type);
if (lb)
{
int i;
ppl_Polyhedron_t ph;
poly_bb_p pbb;
+ int nb_loops = number_of_loops ();
+ ppl_Pointset_Powerset_C_Polyhedron_t *domains
+ = XNEWVEC (ppl_Pointset_Powerset_C_Polyhedron_t, nb_loops);
+
+ for (i = 0; i < nb_loops; i++)
+ domains[i] = NULL;
ppl_new_C_Polyhedron_from_space_dimension (&ph, scop_nb_params (scop), 0);
for (i = 0; VEC_iterate (loop_p, SESE_LOOP_NEST (region), i, loop); i++)
if (!loop_in_sese_p (loop_outer (loop), region))
- build_loop_iteration_domains (scop, loop, ph, 0);
+ build_loop_iteration_domains (scop, loop, ph, 0, domains);
for (i = 0; VEC_iterate (poly_bb_p, SCOP_BBS (scop), i, pbb); i++)
- if (gbb_loop (PBB_BLACK_BOX (pbb))->aux)
+ if (domains[gbb_loop (PBB_BLACK_BOX (pbb))->num])
ppl_new_Pointset_Powerset_C_Polyhedron_from_Pointset_Powerset_C_Polyhedron
(&PBB_DOMAIN (pbb), (ppl_const_Pointset_Powerset_C_Polyhedron_t)
- gbb_loop (PBB_BLACK_BOX (pbb))->aux);
+ domains[gbb_loop (PBB_BLACK_BOX (pbb))->num]);
else
ppl_new_Pointset_Powerset_C_Polyhedron_from_C_Polyhedron
(&PBB_DOMAIN (pbb), ph);
- for (i = 0; VEC_iterate (loop_p, SESE_LOOP_NEST (region), i, loop); i++)
- if (loop->aux)
- {
- ppl_delete_Pointset_Powerset_C_Polyhedron
- ((ppl_Pointset_Powerset_C_Polyhedron_t) loop->aux);
- loop->aux = NULL;
- }
+ for (i = 0; i < nb_loops; i++)
+ if (domains[i])
+ ppl_delete_Pointset_Powerset_C_Polyhedron (domains[i]);
ppl_delete_Polyhedron (ph);
+ free (domains);
}
/* Add a constrain to the ACCESSES polyhedron for the alias set of
/* Creates a zero dimension array of the same type as VAR. */
static tree
-create_zero_dim_array (tree var)
+create_zero_dim_array (tree var, const char *base_name)
{
tree index_type = build_index_type (integer_zero_node);
tree elt_type = TREE_TYPE (var);
tree array_type = build_array_type (elt_type, index_type);
- tree base = create_tmp_var (array_type, "Red");
+ tree base = create_tmp_var (array_type, base_name);
add_referenced_var (base);
gimple phi = gsi_stmt (*psi);
tree res = gimple_phi_result (phi);
tree var = SSA_NAME_VAR (res);
- tree zero_dim_array = create_zero_dim_array (var);
+ tree zero_dim_array = create_zero_dim_array (var, "Close_Phi");
gimple_stmt_iterator gsi = gsi_after_labels (gimple_bb (phi));
gimple stmt = gimple_build_assign (res, zero_dim_array);
tree arg = gimple_phi_arg_def (phi, 0);
- insert_out_of_ssa_copy (zero_dim_array, arg);
+ if (TREE_CODE (arg) == SSA_NAME
+ && !SSA_NAME_IS_DEFAULT_DEF (arg))
+ insert_out_of_ssa_copy (zero_dim_array, arg);
+ else
+ insert_out_of_ssa_copy_on_edge (single_pred_edge (gimple_bb (phi)),
+ zero_dim_array, arg);
remove_phi_node (psi, false);
gsi_insert_before (&gsi, stmt, GSI_NEW_STMT);
basic_block bb = gimple_bb (phi);
tree res = gimple_phi_result (phi);
tree var = SSA_NAME_VAR (res);
- tree zero_dim_array = create_zero_dim_array (var);
+ tree zero_dim_array = create_zero_dim_array (var, "General_Reduction");
gimple_stmt_iterator gsi;
gimple stmt;
gimple_seq stmts;
| end_2
| end_1
- whereas inserting the copy on the incomming edge is correct
+ whereas inserting the copy on the incoming edge is correct
| a = ...
| loop_1
FOR_EACH_IMM_USE_STMT (use_stmt, imm_iter, def)
if (def_bb != gimple_bb (use_stmt)
- && gimple_code (use_stmt) != GIMPLE_PHI)
+ && gimple_code (use_stmt) != GIMPLE_PHI
+ && !is_gimple_debug (use_stmt))
{
if (!zero_dim_array)
{
- zero_dim_array = create_zero_dim_array (SSA_NAME_VAR (def));
+ zero_dim_array = create_zero_dim_array
+ (SSA_NAME_VAR (def), "Cross_BB_scalar_dependence");
insert_out_of_ssa_copy (zero_dim_array, def);
gsi_next (gsi);
}
split_block (bb, stmt);
+ if (gsi_one_before_end_p (gsi_start_nondebug_bb (bb)))
+ return bb;
+
gsi = gsi_last_bb (bb);
gsi_prev (&gsi);
e = split_block (bb, gsi_stmt (gsi));
static inline bool
is_reduction_operation_p (gimple stmt)
{
+ enum tree_code code;
+
+ gcc_assert (is_gimple_assign (stmt));
+ code = gimple_assign_rhs_code (stmt);
+
return flag_associative_math
- && commutative_tree_code (gimple_assign_rhs_code (stmt))
- && associative_tree_code (gimple_assign_rhs_code (stmt));
+ && commutative_tree_code (code)
+ && associative_tree_code (code);
}
/* Returns true when PHI contains an argument ARG. */
stmt = SSA_NAME_DEF_STMT (arg);
+ if (gimple_code (stmt) == GIMPLE_NOP
+ || gimple_code (stmt) == GIMPLE_CALL)
+ return NULL;
+
if (gimple_code (stmt) == GIMPLE_PHI)
{
if (phi_contains_arg (stmt, lhs))
return NULL;
}
+ if (!is_gimple_assign (stmt))
+ return NULL;
+
if (gimple_num_ops (stmt) == 2)
return follow_ssa_with_commutative_ops (gimple_assign_rhs1 (stmt), lhs);
}
/* Detect commutative and associative scalar reductions starting at
- the STMT. */
+ the STMT. Return the phi node of the reduction cycle, or NULL. */
static gimple
detect_commutative_reduction_arg (tree lhs, gimple stmt, tree arg,
{
gimple phi = follow_ssa_with_commutative_ops (arg, lhs);
- if (phi)
- {
- VEC_safe_push (gimple, heap, *in, stmt);
- VEC_safe_push (gimple, heap, *out, stmt);
- return phi;
- }
+ if (!phi)
+ return NULL;
- return NULL;
+ VEC_safe_push (gimple, heap, *in, stmt);
+ VEC_safe_push (gimple, heap, *out, stmt);
+ return phi;
}
/* Detect commutative and associative scalar reductions starting at
- the STMT. */
+ the STMT. Return the phi node of the reduction cycle, or NULL. */
static gimple
detect_commutative_reduction_assign (gimple stmt, VEC (gimple, heap) **in,
}
/* Detect commutative and associative scalar reductions starting at
- the loop closed phi node CLOSE_PHI. */
+ the loop closed phi node CLOSE_PHI. Return the phi node of the
+ reduction cycle, or NULL. */
static gimple
detect_commutative_reduction (gimple stmt, VEC (gimple, heap) **in,
if (scalar_close_phi_node_p (stmt))
{
tree arg = gimple_phi_arg_def (stmt, 0);
- gimple def = SSA_NAME_DEF_STMT (arg);
- gimple loop_phi = detect_commutative_reduction (def, in, out);
+ gimple def, loop_phi;
+
+ if (TREE_CODE (arg) != SSA_NAME)
+ return NULL;
+
+ def = SSA_NAME_DEF_STMT (arg);
+ loop_phi = detect_commutative_reduction (def, in, out);
if (loop_phi)
{
translate_scalar_reduction_to_array_for_stmt (tree red, gimple stmt,
gimple loop_phi)
{
- basic_block bb = gimple_bb (stmt);
- gimple_stmt_iterator insert_gsi = gsi_after_labels (bb);
+ gimple_stmt_iterator insert_gsi = gsi_after_labels (gimple_bb (loop_phi));
tree res = gimple_phi_result (loop_phi);
gimple assign = gimple_build_assign (res, red);
gsi_insert_before (&insert_gsi, assign, GSI_SAME_STMT);
+ insert_gsi = gsi_after_labels (gimple_bb (stmt));
assign = gimple_build_assign (red, gimple_assign_lhs (stmt));
insert_gsi = gsi_for_stmt (stmt);
gsi_insert_after (&insert_gsi, assign, GSI_SAME_STMT);
gsi_insert_seq_on_edge (edge_initial_value_for_loop_phi (loop_phi), stmts);
}
+/* Removes the PHI node and resets all the debug stmts that are using
+ the PHI_RESULT. */
+
+static void
+remove_phi (gimple phi)
+{
+ imm_use_iterator imm_iter;
+ tree def;
+ use_operand_p use_p;
+ gimple_stmt_iterator gsi;
+ VEC (gimple, heap) *update = VEC_alloc (gimple, heap, 3);
+ unsigned int i;
+ gimple stmt;
+
+ def = PHI_RESULT (phi);
+ FOR_EACH_IMM_USE_FAST (use_p, imm_iter, def)
+ {
+ stmt = USE_STMT (use_p);
+
+ if (is_gimple_debug (stmt))
+ {
+ gimple_debug_bind_reset_value (stmt);
+ VEC_safe_push (gimple, heap, update, stmt);
+ }
+ }
+
+ for (i = 0; VEC_iterate (gimple, update, i, stmt); i++)
+ update_stmt (stmt);
+
+ VEC_free (gimple, heap, update);
+
+ gsi = gsi_for_phi_node (phi);
+ remove_phi_node (&gsi, false);
+}
+
/* Rewrite out of SSA the reduction described by the loop phi nodes
IN, and the close phi nodes OUT. IN and OUT are structured by loop
levels like this:
unsigned int i;
gimple loop_phi;
tree red;
- gimple_stmt_iterator gsi;
for (i = 0; VEC_iterate (gimple, in, i, loop_phi); i++)
{
SET_BIT (reductions, bb->index);
gcc_assert (close_phi == loop_phi);
- red = create_zero_dim_array (gimple_assign_lhs (stmt));
+ red = create_zero_dim_array
+ (gimple_assign_lhs (stmt), "Commutative_Associative_Reduction");
translate_scalar_reduction_to_array_for_stmt
(red, stmt, VEC_index (gimple, in, 1));
continue;
insert_copyin (red, loop_phi);
}
- gsi = gsi_for_phi_node (loop_phi);
- remove_phi_node (&gsi, false);
-
- gsi = gsi_for_phi_node (close_phi);
- remove_phi_node (&gsi, false);
+ remove_phi (loop_phi);
+ remove_phi (close_phi);
}
}
#endif
}
+/* A LOOP is in normal form for Graphite when it contains only one
+ scalar phi node that defines the main induction variable of the
+ loop, only one increment of the IV, and only one exit condition. */
+
+static void
+graphite_loop_normal_form (loop_p loop)
+{
+ struct tree_niter_desc niter;
+ tree nit;
+ gimple_seq stmts;
+ edge exit = single_dom_exit (loop);
+
+ bool known_niter = number_of_iterations_exit (loop, exit, &niter, false);
+
+ /* At this point we should know the number of iterations. */
+ gcc_assert (known_niter);
+
+ nit = force_gimple_operand (unshare_expr (niter.niter), &stmts, true,
+ NULL_TREE);
+ if (stmts)
+ gsi_insert_seq_on_edge_immediate (loop_preheader_edge (loop), stmts);
+
+ loop->single_iv = canonicalize_loop_ivs (loop, &nit, false);
+}
+
+/* Rewrite all the loops of SCOP in normal form: one induction
+ variable per loop. */
+
+static void
+scop_canonicalize_loops (scop_p scop)
+{
+ loop_iterator li;
+ loop_p loop;
+
+ FOR_EACH_LOOP (li, loop, 0)
+ if (loop_in_sese_p (loop, SCOP_REGION (scop)))
+ graphite_loop_normal_form (loop);
+}
+
+/* Java does not initialize long_long_integer_type_node. */
+#define my_long_long (long_long_integer_type_node ? long_long_integer_type_node : ssizetype)
+
+/* Can all ivs be represented by a signed integer?
+ As CLooG might generate negative values in its expressions, signed loop ivs
+ are required in the backend. */
+static bool
+scop_ivs_can_be_represented (scop_p scop)
+{
+ loop_iterator li;
+ loop_p loop;
+
+ FOR_EACH_LOOP (li, loop, 0)
+ {
+ tree type;
+ int precision;
+
+ if (!loop_in_sese_p (loop, SCOP_REGION (scop)))
+ continue;
+
+ if (!loop->single_iv)
+ continue;
+
+ type = TREE_TYPE(loop->single_iv);
+ precision = TYPE_PRECISION (type);
+
+ if (TYPE_UNSIGNED (type)
+ && precision >= TYPE_PRECISION (my_long_long))
+ return false;
+ }
+
+ return true;
+}
+
+#undef my_long_long
+
/* Builds the polyhedral representation for a SESE region. */
-bool
+void
build_poly_scop (scop_p scop)
{
sese region = SCOP_REGION (scop);
sbitmap reductions = sbitmap_alloc (last_basic_block * 2);
+ graphite_dim_t max_dim;
sbitmap_zero (reductions);
rewrite_commutative_reductions_out_of_ssa (region, reductions);
sense to optimize a scop containing only PBBs that do not belong
to any loops. */
if (nb_pbbs_in_loops (scop) == 0)
- return false;
+ return;
+
+ scop_canonicalize_loops (scop);
+ if (!scop_ivs_can_be_represented (scop))
+ return;
build_sese_loop_nests (region);
build_sese_conditions (region);
find_scop_parameters (scop);
+ max_dim = PARAM_VALUE (PARAM_GRAPHITE_MAX_NB_SCOP_PARAMS);
+ if (scop_nb_params (scop) > max_dim)
+ return;
+
build_scop_iteration_domain (scop);
build_scop_context (scop);
build_scop_scattering (scop);
build_scop_drs (scop);
- return true;
+ /* This SCoP has been translated to the polyhedral
+ representation. */
+ POLY_SCOP_P (scop) = true;
}
/* Always return false. Exercise the scop_to_clast function. */