X-Git-Url: http://git.sourceforge.jp/view?p=pf3gnuchains%2Fgcc-fork.git;a=blobdiff_plain;f=gcc%2Ftree-ssa-structalias.c;h=05fd6d971f2456e303a866d1d016863622d8dea5;hp=eda053f24302a60eaec774d6a421bae53f1aa832;hb=2f1ce05cf95e354ae01c6801d32fb518f97fa593;hpb=2f8eb9091b18ca1a7b6868980a41678f6d93b9ed diff --git a/gcc/tree-ssa-structalias.c b/gcc/tree-ssa-structalias.c index eda053f2430..05fd6d971f2 100644 --- a/gcc/tree-ssa-structalias.c +++ b/gcc/tree-ssa-structalias.c @@ -1,5 +1,6 @@ /* Tree based points-to analysis - Copyright (C) 2005, 2006, 2007, 2008, 2009 Free Software Foundation, Inc. + Copyright (C) 2005, 2006, 2007, 2008, 2009, 2010 + Free Software Foundation, Inc. Contributed by Daniel Berlin This file is part of GCC. @@ -26,16 +27,11 @@ #include "obstack.h" #include "bitmap.h" #include "flags.h" -#include "rtl.h" -#include "tm_p.h" -#include "hard-reg-set.h" #include "basic-block.h" #include "output.h" #include "tree.h" #include "tree-flow.h" #include "tree-inline.h" -#include "varray.h" -#include "diagnostic.h" #include "toplev.h" #include "gimple.h" #include "hashtab.h" @@ -160,7 +156,48 @@ TODO: We could handle unions, but to be honest, it's probably not worth the pain or slowdown. */ -static GTY ((if_marked ("tree_map_marked_p"), param_is (struct tree_map))) +/* IPA-PTA optimizations possible. + + When the indirect function called is ANYTHING we can add disambiguation + based on the function signatures (or simply the parameter count which + is the varinfo size). We also do not need to consider functions that + do not have their address taken. + + The is_global_var bit which marks escape points is overly conservative + in IPA mode. Split it to is_escape_point and is_global_var - only + externally visible globals are escape points in IPA mode. This is + also needed to fix the pt_solution_includes_global predicate + (and thus ptr_deref_may_alias_global_p). + + The way we introduce DECL_PT_UID to avoid fixing up all points-to + sets in the translation unit when we copy a DECL during inlining + pessimizes precision. The advantage is that the DECL_PT_UID keeps + compile-time and memory usage overhead low - the points-to sets + do not grow or get unshared as they would during a fixup phase. + An alternative solution is to delay IPA PTA until after all + inlining transformations have been applied. + + The way we propagate clobber/use information isn't optimized. + It should use a new complex constraint that properly filters + out local variables of the callee (though that would make + the sets invalid after inlining). OTOH we might as well + admit defeat to WHOPR and simply do all the clobber/use analysis + and propagation after PTA finished but before we threw away + points-to information for memory variables. WHOPR and PTA + do not play along well anyway - the whole constraint solving + would need to be done in WPA phase and it will be very interesting + to apply the results to local SSA names during LTRANS phase. + + We probably should compute a per-function unit-ESCAPE solution + propagating it simply like the clobber / uses solutions. The + solution can go alongside the non-IPA espaced solution and be + used to query which vars escape the unit through a function. + + We never put function decls in points-to sets so we do not + keep the set of called functions for indirect calls. + + And probably more. */ +static GTY ((if_marked ("tree_map_marked_p"), param_is (struct heapvar_map))) htab_t heapvar_for_stmt; static bool use_field_sensitive = true; @@ -211,28 +248,36 @@ struct variable_info /* True if this is a variable created by the constraint analysis, such as heap variables and constraints we had to break up. */ - unsigned int is_artificial_var:1; + unsigned int is_artificial_var : 1; /* True if this is a special variable whose solution set should not be changed. */ - unsigned int is_special_var:1; + unsigned int is_special_var : 1; /* True for variables whose size is not known or variable. */ - unsigned int is_unknown_size_var:1; + unsigned int is_unknown_size_var : 1; /* True for (sub-)fields that represent a whole variable. */ unsigned int is_full_var : 1; /* True if this is a heap variable. */ - unsigned int is_heap_var:1; + unsigned int is_heap_var : 1; - /* True if we may not use TBAA to prune references to this - variable. This is used for C++ placement new. */ - unsigned int no_tbaa_pruning : 1; + /* True if this is a variable tracking a restrict pointer source. */ + unsigned int is_restrict_var : 1; /* True if this field may contain pointers. */ unsigned int may_have_pointers : 1; + /* True if this field has only restrict qualified pointers. */ + unsigned int only_restrict_pointers : 1; + + /* True if this represents a global variable. */ + unsigned int is_global_var : 1; + + /* True if this represents a IPA function info. */ + unsigned int is_fn_info : 1; + /* A link to the variable for the next field in this structure. */ struct variable_info *next; @@ -285,54 +330,42 @@ get_varinfo (unsigned int n) /* Static IDs for the special variables. */ enum { nothing_id = 0, anything_id = 1, readonly_id = 2, - escaped_id = 3, nonlocal_id = 4, callused_id = 5, - storedanything_id = 6, integer_id = 7 }; - -/* Variable that represents the unknown pointer. */ -static varinfo_t var_anything; -static tree anything_tree; - -/* Variable that represents the NULL pointer. */ -static varinfo_t var_nothing; -static tree nothing_tree; - -/* Variable that represents read only memory. */ -static varinfo_t var_readonly; -static tree readonly_tree; - -/* Variable that represents escaped memory. */ -static varinfo_t var_escaped; -static tree escaped_tree; - -/* Variable that represents nonlocal memory. */ -static varinfo_t var_nonlocal; -static tree nonlocal_tree; + escaped_id = 3, nonlocal_id = 4, + storedanything_id = 5, integer_id = 6 }; -/* Variable that represents call-used memory. */ -static varinfo_t var_callused; -static tree callused_tree; +struct GTY(()) heapvar_map { + struct tree_map map; + unsigned HOST_WIDE_INT offset; +}; -/* Variable that represents variables that are stored to anything. */ -static varinfo_t var_storedanything; -static tree storedanything_tree; +static int +heapvar_map_eq (const void *p1, const void *p2) +{ + const struct heapvar_map *h1 = (const struct heapvar_map *)p1; + const struct heapvar_map *h2 = (const struct heapvar_map *)p2; + return (h1->map.base.from == h2->map.base.from + && h1->offset == h2->offset); +} -/* Variable that represents integers. This is used for when people do things - like &0->a.b. */ -static varinfo_t var_integer; -static tree integer_tree; +static unsigned int +heapvar_map_hash (struct heapvar_map *h) +{ + return iterative_hash_host_wide_int (h->offset, + htab_hash_pointer (h->map.base.from)); +} /* Lookup a heap var for FROM, and return it if we find one. */ static tree -heapvar_lookup (tree from) +heapvar_lookup (tree from, unsigned HOST_WIDE_INT offset) { - struct tree_map *h, in; - in.base.from = from; - - h = (struct tree_map *) htab_find_with_hash (heapvar_for_stmt, &in, - htab_hash_pointer (from)); + struct heapvar_map *h, in; + in.map.base.from = from; + in.offset = offset; + h = (struct heapvar_map *) htab_find_with_hash (heapvar_for_stmt, &in, + heapvar_map_hash (&in)); if (h) - return h->to; + return h->map.to; return NULL_TREE; } @@ -340,49 +373,138 @@ heapvar_lookup (tree from) hashtable. */ static void -heapvar_insert (tree from, tree to) +heapvar_insert (tree from, unsigned HOST_WIDE_INT offset, tree to) { - struct tree_map *h; + struct heapvar_map *h; void **loc; - h = GGC_NEW (struct tree_map); - h->hash = htab_hash_pointer (from); - h->base.from = from; - h->to = to; - loc = htab_find_slot_with_hash (heapvar_for_stmt, h, h->hash, INSERT); - *(struct tree_map **) loc = h; + h = ggc_alloc_heapvar_map (); + h->map.base.from = from; + h->offset = offset; + h->map.hash = heapvar_map_hash (h); + h->map.to = to; + loc = htab_find_slot_with_hash (heapvar_for_stmt, h, h->map.hash, INSERT); + gcc_assert (*loc == NULL); + *(struct heapvar_map **) loc = h; } /* Return a new variable info structure consisting for a variable - named NAME, and using constraint graph node NODE. */ + named NAME, and using constraint graph node NODE. Append it + to the vector of variable info structures. */ static varinfo_t -new_var_info (tree t, unsigned int id, const char *name) +new_var_info (tree t, const char *name) { + unsigned index = VEC_length (varinfo_t, varmap); varinfo_t ret = (varinfo_t) pool_alloc (variable_info_pool); - tree var; - ret->id = id; + ret->id = index; ret->name = name; ret->decl = t; - ret->is_artificial_var = false; - ret->is_heap_var = false; + /* Vars without decl are artificial and do not have sub-variables. */ + ret->is_artificial_var = (t == NULL_TREE); ret->is_special_var = false; ret->is_unknown_size_var = false; - ret->is_full_var = false; + ret->is_full_var = (t == NULL_TREE); + ret->is_heap_var = false; + ret->is_restrict_var = false; ret->may_have_pointers = true; - var = t; - if (TREE_CODE (var) == SSA_NAME) - var = SSA_NAME_VAR (var); - ret->no_tbaa_pruning = (DECL_P (var) - && POINTER_TYPE_P (TREE_TYPE (var)) - && DECL_NO_TBAA_P (var)); + ret->only_restrict_pointers = false; + ret->is_global_var = (t == NULL_TREE); + ret->is_fn_info = false; + if (t && DECL_P (t)) + ret->is_global_var = is_global_var (t); ret->solution = BITMAP_ALLOC (&pta_obstack); ret->oldsolution = BITMAP_ALLOC (&oldpta_obstack); ret->next = NULL; + + stats.total_vars++; + + VEC_safe_push (varinfo_t, heap, varmap, ret); + return ret; } + +/* A map mapping call statements to per-stmt variables for uses + and clobbers specific to the call. */ +struct pointer_map_t *call_stmt_vars; + +/* Lookup or create the variable for the call statement CALL. */ + +static varinfo_t +get_call_vi (gimple call) +{ + void **slot_p; + varinfo_t vi, vi2; + + slot_p = pointer_map_insert (call_stmt_vars, call); + if (*slot_p) + return (varinfo_t) *slot_p; + + vi = new_var_info (NULL_TREE, "CALLUSED"); + vi->offset = 0; + vi->size = 1; + vi->fullsize = 2; + vi->is_full_var = true; + + vi->next = vi2 = new_var_info (NULL_TREE, "CALLCLOBBERED"); + vi2->offset = 1; + vi2->size = 1; + vi2->fullsize = 2; + vi2->is_full_var = true; + + *slot_p = (void *) vi; + return vi; +} + +/* Lookup the variable for the call statement CALL representing + the uses. Returns NULL if there is nothing special about this call. */ + +static varinfo_t +lookup_call_use_vi (gimple call) +{ + void **slot_p; + + slot_p = pointer_map_contains (call_stmt_vars, call); + if (slot_p) + return (varinfo_t) *slot_p; + + return NULL; +} + +/* Lookup the variable for the call statement CALL representing + the clobbers. Returns NULL if there is nothing special about this call. */ + +static varinfo_t +lookup_call_clobber_vi (gimple call) +{ + varinfo_t uses = lookup_call_use_vi (call); + if (!uses) + return NULL; + + return uses->next; +} + +/* Lookup or create the variable for the call statement CALL representing + the uses. */ + +static varinfo_t +get_call_use_vi (gimple call) +{ + return get_call_vi (call); +} + +/* Lookup or create the variable for the call statement CALL representing + the clobbers. */ + +static varinfo_t ATTRIBUTE_UNUSED +get_call_clobber_vi (gimple call) +{ + return get_call_vi (call)->next; +} + + typedef enum {SCALAR, DEREF, ADDRESSOF} constraint_expr_type; /* An expression that appears in a constraint. */ @@ -430,10 +552,6 @@ struct constraint static VEC(constraint_t,heap) *constraints; static alloc_pool constraint_pool; - -DEF_VEC_I(int); -DEF_VEC_ALLOC_I(int, heap); - /* The constraint graph is represented as an array of bitmaps containing successor nodes. */ @@ -596,7 +714,7 @@ void debug_sa_points_to_info (void); /* Print out constraint C to stderr. */ -void +DEBUG_FUNCTION void debug_constraint (constraint_t c) { dump_constraint (stderr, c); @@ -605,20 +723,20 @@ debug_constraint (constraint_t c) /* Print out all constraints to FILE */ static void -dump_constraints (FILE *file) +dump_constraints (FILE *file, int from) { int i; constraint_t c; - for (i = 0; VEC_iterate (constraint_t, constraints, i, c); i++) + for (i = from; VEC_iterate (constraint_t, constraints, i, c); i++) dump_constraint (file, c); } /* Print out all constraints to stderr. */ -void +DEBUG_FUNCTION void debug_constraints (void) { - dump_constraints (stderr); + dump_constraints (stderr, 0); } /* Print out to FILE the edge in the constraint graph that is created by @@ -669,7 +787,7 @@ dump_constraint_graph (FILE *file) /* Print the constraints used to produce the constraint graph. The constraints will be printed as comments in the dot file: */ fprintf (file, "\n\n/* Constraints used in the constraint graph:\n"); - dump_constraints (file); + dump_constraints (file, 0); fprintf (file, "*/\n"); /* Prints the header of the dot file: */ @@ -705,7 +823,7 @@ dump_constraint_graph (FILE *file) /* Print out the constraint graph to stderr. */ -void +DEBUG_FUNCTION void debug_constraint_graph (void) { dump_constraint_graph (stderr); @@ -1278,13 +1396,18 @@ build_succ_graph (void) } } - /* Add edges from STOREDANYTHING to all non-direct nodes. */ + /* Add edges from STOREDANYTHING to all non-direct nodes that can + receive pointers. */ t = find (storedanything_id); for (i = integer_id + 1; i < FIRST_REF_NODE; ++i) { - if (!TEST_BIT (graph->direct_nodes, i)) + if (!TEST_BIT (graph->direct_nodes, i) + && get_varinfo (i)->may_have_pointers) add_graph_edge (graph, find (i), t); } + + /* Everything stored to ANYTHING also potentially escapes. */ + add_graph_edge (graph, find (escaped_id), t); } @@ -1292,10 +1415,6 @@ build_succ_graph (void) static unsigned int changed_count; static sbitmap changed; -DEF_VEC_I(unsigned); -DEF_VEC_ALLOC_I(unsigned,heap); - - /* Strongly Connected Component visitation info. */ struct scc_info @@ -1362,7 +1481,6 @@ scc_visit (constraint_graph_t graph, struct scc_info *si, unsigned int n) && si->dfs[VEC_last (unsigned, si->scc_stack)] >= my_dfs) { bitmap scc = BITMAP_ALLOC (NULL); - bool have_ref_node = n >= FIRST_REF_NODE; unsigned int lowest_node; bitmap_iterator bi; @@ -1374,8 +1492,6 @@ scc_visit (constraint_graph_t graph, struct scc_info *si, unsigned int n) unsigned int w = VEC_pop (unsigned, si->scc_stack); bitmap_set_bit (scc, w); - if (w >= FIRST_REF_NODE) - have_ref_node = true; } lowest_node = bitmap_first_set_bit (scc); @@ -1425,9 +1541,6 @@ unify_nodes (constraint_graph_t graph, unsigned int to, unsigned int from, merge_graph_nodes (graph, to, from); merge_node_constraints (graph, to, from); - if (get_varinfo (from)->no_tbaa_pruning) - get_varinfo (to)->no_tbaa_pruning = true; - /* Mark TO as changed if FROM was changed. If TO was already marked as changed, decrease the changed count. */ @@ -1455,10 +1568,10 @@ unify_nodes (constraint_graph_t graph, unsigned int to, unsigned int from, changed_count++; } } - + BITMAP_FREE (get_varinfo (from)->solution); BITMAP_FREE (get_varinfo (from)->oldsolution); - + if (stats.iterations > 0) { BITMAP_FREE (get_varinfo (to)->oldsolution); @@ -1593,7 +1706,8 @@ do_sd_constraint (constraint_graph_t graph, constraint_t c, the set. Use ESCAPED as representative instead. */ else if (v->id == escaped_id) flag |= bitmap_set_bit (sol, escaped_id); - else if (add_graph_edge (graph, lhs, t)) + else if (v->may_have_pointers + && add_graph_edge (graph, lhs, t)) flag |= bitmap_ior_into (sol, get_varinfo (t)->solution); /* If the variable is not exactly at the requested offset @@ -1632,6 +1746,7 @@ do_ds_constraint (constraint_t c, bitmap delta) unsigned int j; bitmap_iterator bi; HOST_WIDE_INT loff = c->lhs.offset; + bool escaped_p = false; /* Our IL does not allow this. */ gcc_assert (c->rhs.offset == 0); @@ -1678,9 +1793,6 @@ do_ds_constraint (constraint_t c, bitmap delta) unsigned int t; HOST_WIDE_INT fieldoffset = v->offset + loff; - if (v->is_special_var) - continue; - if (v->is_full_var) fieldoffset = v->offset; else if (loff != 0) @@ -1693,19 +1805,32 @@ do_ds_constraint (constraint_t c, bitmap delta) { if (v->may_have_pointers) { - t = find (v->id); - if (add_graph_edge (graph, t, rhs)) + /* If v is a global variable then this is an escape point. */ + if (v->is_global_var + && !escaped_p) { - if (bitmap_ior_into (get_varinfo (t)->solution, sol)) + t = find (escaped_id); + if (add_graph_edge (graph, t, rhs) + && bitmap_ior_into (get_varinfo (t)->solution, sol) + && !TEST_BIT (changed, t)) { - if (t == rhs) - sol = get_varinfo (rhs)->solution; - if (!TEST_BIT (changed, t)) - { - SET_BIT (changed, t); - changed_count++; - } + SET_BIT (changed, t); + changed_count++; } + /* Enough to let rhs escape once. */ + escaped_p = true; + } + + if (v->is_special_var) + break; + + t = find (v->id); + if (add_graph_edge (graph, t, rhs) + && bitmap_ior_into (get_varinfo (t)->solution, sol) + && !TEST_BIT (changed, t)) + { + SET_BIT (changed, t); + changed_count++; } } @@ -1878,7 +2003,8 @@ equiv_class_label_eq (const void *p1, const void *p2) { const_equiv_class_label_t const eql1 = (const_equiv_class_label_t) p1; const_equiv_class_label_t const eql2 = (const_equiv_class_label_t) p2; - return bitmap_equal_p (eql1->labels, eql2->labels); + return (eql1->hashcode == eql2->hashcode + && bitmap_equal_p (eql1->labels, eql2->labels)); } /* Lookup a equivalence class in TABLE by the bitmap of LABELS it @@ -2291,10 +2417,10 @@ unite_pointer_equivalences (constraint_graph_t graph) if (label) { int label_rep = graph->pe_rep[label]; - + if (label_rep == -1) continue; - + label_rep = find (label_rep); if (label_rep >= 0 && unite (label_rep, find (i))) unify_nodes (graph, label_rep, i, false); @@ -2371,7 +2497,7 @@ rewrite_constraints (constraint_graph_t graph, { if (dump_file && (dump_flags & TDF_DETAILS)) { - + fprintf (dump_file, "%s is a non-pointer variable," "ignoring constraint:", get_varinfo (lhs.var)->name); @@ -2385,7 +2511,7 @@ rewrite_constraints (constraint_graph_t graph, { if (dump_file && (dump_flags & TDF_DETAILS)) { - + fprintf (dump_file, "%s is a non-pointer variable," "ignoring constraint:", get_varinfo (rhs.var)->name); @@ -2628,10 +2754,14 @@ lookup_vi_for_tree (tree t) static const char * alias_get_name (tree decl) { - const char *res = get_name (decl); + const char *res; char *temp; int num_printed = 0; + if (DECL_ASSEMBLER_NAME_SET_P (decl)) + res = IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (decl)); + else + res= get_name (decl); if (res != NULL) return res; @@ -2670,20 +2800,25 @@ get_vi_for_tree (tree t) return (varinfo_t) *slot; } -/* Get a constraint expression for a new temporary variable. */ +/* Get a scalar constraint expression for a new temporary variable. */ static struct constraint_expr -get_constraint_exp_for_temp (tree t) +new_scalar_tmp_constraint_exp (const char *name) { - struct constraint_expr cexpr; + struct constraint_expr tmp; + varinfo_t vi; - gcc_assert (SSA_VAR_P (t)); + vi = new_var_info (NULL_TREE, name); + vi->offset = 0; + vi->size = -1; + vi->fullsize = -1; + vi->is_full_var = 1; - cexpr.type = SCALAR; - cexpr.var = get_vi_for_tree (t)->id; - cexpr.offset = 0; + tmp.var = vi->id; + tmp.type = SCALAR; + tmp.offset = 0; - return cexpr; + return tmp; } /* Get a constraint expression vector from an SSA_VAR_P node. @@ -2723,7 +2858,8 @@ get_constraint_for_ssa_var (tree t, VEC(ce_s, heap) **results, bool address_p) /* If we are not taking the address of the constraint expr, add all sub-fiels of the variable as well. */ - if (!address_p) + if (!address_p + && !vi->is_full_var) { for (; vi; vi = vi->next) { @@ -2758,27 +2894,30 @@ process_constraint (constraint_t t) /* ADDRESSOF on the lhs is invalid. */ gcc_assert (lhs.type != ADDRESSOF); + /* We shouldn't add constraints from things that cannot have pointers. + It's not completely trivial to avoid in the callers, so do it here. */ + if (rhs.type != ADDRESSOF + && !get_varinfo (rhs.var)->may_have_pointers) + return; + + /* Likewise adding to the solution of a non-pointer var isn't useful. */ + if (!get_varinfo (lhs.var)->may_have_pointers) + return; + /* This can happen in our IR with things like n->a = *p */ if (rhs.type == DEREF && lhs.type == DEREF && rhs.var != anything_id) { /* Split into tmp = *rhs, *lhs = tmp */ - tree rhsdecl = get_varinfo (rhs.var)->decl; - tree pointertype = TREE_TYPE (rhsdecl); - tree pointedtotype = TREE_TYPE (pointertype); - tree tmpvar = create_tmp_var_raw (pointedtotype, "doubledereftmp"); - struct constraint_expr tmplhs = get_constraint_exp_for_temp (tmpvar); - + struct constraint_expr tmplhs; + tmplhs = new_scalar_tmp_constraint_exp ("doubledereftmp"); process_constraint (new_constraint (tmplhs, rhs)); process_constraint (new_constraint (lhs, tmplhs)); } else if (rhs.type == ADDRESSOF && lhs.type == DEREF) { /* Split into tmp = &rhs, *lhs = tmp */ - tree rhsdecl = get_varinfo (rhs.var)->decl; - tree pointertype = TREE_TYPE (rhsdecl); - tree tmpvar = create_tmp_var_raw (pointertype, "derefaddrtmp"); - struct constraint_expr tmplhs = get_constraint_exp_for_temp (tmpvar); - + struct constraint_expr tmplhs; + tmplhs = new_scalar_tmp_constraint_exp ("derefaddrtmp"); process_constraint (new_constraint (tmplhs, rhs)); process_constraint (new_constraint (lhs, tmplhs)); } @@ -2800,6 +2939,12 @@ type_could_have_pointers (tree type) if (TREE_CODE (type) == ARRAY_TYPE) return type_could_have_pointers (TREE_TYPE (type)); + /* A function or method can consume pointers. + ??? We could be more precise here. */ + if (TREE_CODE (type) == FUNCTION_TYPE + || TREE_CODE (type) == METHOD_TYPE) + return true; + return AGGREGATE_TYPE_P (type); } @@ -2809,7 +2954,11 @@ type_could_have_pointers (tree type) static bool could_have_pointers (tree t) { - return type_could_have_pointers (TREE_TYPE (t)); + return (((TREE_CODE (t) == VAR_DECL + || TREE_CODE (t) == PARM_DECL + || TREE_CODE (t) == RESULT_DECL) + && (TREE_PUBLIC (t) || DECL_EXTERNAL (t) || TREE_ADDRESSABLE (t))) + || type_could_have_pointers (TREE_TYPE (t))); } /* Return the position, in bits, of FIELD_DECL from the beginning of its @@ -2835,7 +2984,7 @@ static void get_constraint_for_ptr_offset (tree ptr, tree offset, VEC (ce_s, heap) **results) { - struct constraint_expr *c; + struct constraint_expr c; unsigned int j, n; HOST_WIDE_INT rhsunitoffset, rhsoffset; @@ -2851,7 +3000,8 @@ get_constraint_for_ptr_offset (tree ptr, tree offset, in a HOST_WIDE_INT, we have to fall back to a conservative solution which includes all sub-fields of all pointed-to variables of ptr. */ - if (!host_integerp (offset, 0)) + if (offset == NULL_TREE + || !host_integerp (offset, 0)) rhsoffset = UNKNOWN_OFFSET; else { @@ -2872,14 +3022,14 @@ get_constraint_for_ptr_offset (tree ptr, tree offset, for (j = 0; j < n; j++) { varinfo_t curr; - c = VEC_index (ce_s, *results, j); - curr = get_varinfo (c->var); + c = *VEC_index (ce_s, *results, j); + curr = get_varinfo (c.var); - if (c->type == ADDRESSOF + if (c.type == ADDRESSOF /* If this varinfo represents a full variable just use it. */ && curr->is_full_var) - c->offset = 0; - else if (c->type == ADDRESSOF + c.offset = 0; + else if (c.type == ADDRESSOF /* If we do not know the offset add all subfields. */ && rhsoffset == UNKNOWN_OFFSET) { @@ -2890,12 +3040,13 @@ get_constraint_for_ptr_offset (tree ptr, tree offset, c2.var = temp->id; c2.type = ADDRESSOF; c2.offset = 0; - VEC_safe_push (ce_s, heap, *results, &c2); + if (c2.var != c.var) + VEC_safe_push (ce_s, heap, *results, &c2); temp = temp->next; } while (temp); } - else if (c->type == ADDRESSOF) + else if (c.type == ADDRESSOF) { varinfo_t temp; unsigned HOST_WIDE_INT offset = curr->offset + rhsoffset; @@ -2927,11 +3078,13 @@ get_constraint_for_ptr_offset (tree ptr, tree offset, c2.offset = 0; VEC_safe_push (ce_s, heap, *results, &c2); } - c->var = temp->id; - c->offset = 0; + c.var = temp->id; + c.offset = 0; } else - c->offset = rhsoffset; + c.offset = rhsoffset; + + VEC_replace (ce_s, *results, j, &c); } } @@ -2953,7 +3106,8 @@ get_constraint_for_component_ref (tree t, VEC(ce_s, heap) **results, /* Some people like to do cute things like take the address of &0->a.b */ forzero = t; - while (!SSA_VAR_P (forzero) && !CONSTANT_CLASS_P (forzero)) + while (handled_component_p (forzero) + || INDIRECT_REF_P (forzero)) forzero = TREE_OPERAND (forzero, 0); if (CONSTANT_CLASS_P (forzero) && integer_zerop (forzero)) @@ -3083,8 +3237,8 @@ do_deref (VEC (ce_s, heap) **constraints) c->type = SCALAR; else if (c->type == DEREF) { - tree tmpvar = create_tmp_var_raw (ptr_type_node, "dereftmp"); - struct constraint_expr tmplhs = get_constraint_exp_for_temp (tmpvar); + struct constraint_expr tmplhs; + tmplhs = new_scalar_tmp_constraint_exp ("dereftmp"); process_constraint (new_constraint (tmplhs, *c)); c->var = tmplhs.var; } @@ -3093,6 +3247,28 @@ do_deref (VEC (ce_s, heap) **constraints) } } +static void get_constraint_for_1 (tree, VEC (ce_s, heap) **, bool); + +/* Given a tree T, return the constraint expression for taking the + address of it. */ + +static void +get_constraint_for_address_of (tree t, VEC (ce_s, heap) **results) +{ + struct constraint_expr *c; + unsigned int i; + + get_constraint_for_1 (t, results, true); + + for (i = 0; VEC_iterate (ce_s, *results, i, c); i++) + { + if (c->type == DEREF) + c->type = SCALAR; + else + c->type = ADDRESSOF; + } +} + /* Given a tree T, return the constraint expression for it. */ static void @@ -3113,14 +3289,18 @@ get_constraint_for_1 (tree t, VEC (ce_s, heap) **results, bool address_p) in that case *NULL does not fail, so it _should_ alias *anything. It is not worth adding a new option or renaming the existing one, since this case is relatively obscure. */ - if (flag_delete_null_pointer_checks - && ((TREE_CODE (t) == INTEGER_CST - && integer_zerop (t)) - /* The only valid CONSTRUCTORs in gimple with pointer typed - elements are zero-initializer. */ - || TREE_CODE (t) == CONSTRUCTOR)) - { - temp.var = nothing_id; + if ((TREE_CODE (t) == INTEGER_CST + && integer_zerop (t)) + /* The only valid CONSTRUCTORs in gimple with pointer typed + elements are zero-initializer. But in IPA mode we also + process global initializers, so verify at least. */ + || (TREE_CODE (t) == CONSTRUCTOR + && CONSTRUCTOR_NELTS (t) == 0)) + { + if (flag_delete_null_pointer_checks) + temp.var = nothing_id; + else + temp.var = anything_id; temp.type = ADDRESSOF; temp.offset = 0; VEC_safe_push (ce_s, heap, *results, &temp); @@ -3144,23 +3324,8 @@ get_constraint_for_1 (tree t, VEC (ce_s, heap) **results, bool address_p) switch (TREE_CODE (t)) { case ADDR_EXPR: - { - struct constraint_expr *c; - unsigned int i; - tree exp = TREE_OPERAND (t, 0); - - get_constraint_for_1 (exp, results, true); - - for (i = 0; VEC_iterate (ce_s, *results, i, c); i++) - { - if (c->type == DEREF) - c->type = SCALAR; - else - c->type = ADDRESSOF; - } - return; - } - break; + get_constraint_for_address_of (TREE_OPERAND (t, 0), results); + return; default:; } break; @@ -3197,6 +3362,26 @@ get_constraint_for_1 (tree t, VEC (ce_s, heap) **results, bool address_p) get_constraint_for_ssa_var (t, results, address_p); return; } + case CONSTRUCTOR: + { + unsigned int i; + tree val; + VEC (ce_s, heap) *tmp = NULL; + FOR_EACH_CONSTRUCTOR_VALUE (CONSTRUCTOR_ELTS (t), i, val) + { + struct constraint_expr *rhsp; + unsigned j; + get_constraint_for_1 (val, &tmp, address_p); + for (j = 0; VEC_iterate (ce_s, tmp, j, rhsp); ++j) + VEC_safe_push (ce_s, heap, *results, rhsp); + VEC_truncate (ce_s, tmp, 0); + } + VEC_free (ce_s, heap, tmp); + /* We do not know whether the constructor was complete, + so technically we have to add &NOTHING or &ANYTHING + like we do for an empty constructor as well. */ + return; + } default:; } break; @@ -3226,6 +3411,34 @@ get_constraint_for (tree t, VEC (ce_s, heap) **results) get_constraint_for_1 (t, results, false); } + +/* Efficiently generates constraints from all entries in *RHSC to all + entries in *LHSC. */ + +static void +process_all_all_constraints (VEC (ce_s, heap) *lhsc, VEC (ce_s, heap) *rhsc) +{ + struct constraint_expr *lhsp, *rhsp; + unsigned i, j; + + if (VEC_length (ce_s, lhsc) <= 1 + || VEC_length (ce_s, rhsc) <= 1) + { + for (i = 0; VEC_iterate (ce_s, lhsc, i, lhsp); ++i) + for (j = 0; VEC_iterate (ce_s, rhsc, j, rhsp); ++j) + process_constraint (new_constraint (*lhsp, *rhsp)); + } + else + { + struct constraint_expr tmp; + tmp = new_scalar_tmp_constraint_exp ("allalltmp"); + for (i = 0; VEC_iterate (ce_s, rhsc, i, rhsp); ++i) + process_constraint (new_constraint (tmp, *rhsp)); + for (i = 0; VEC_iterate (ce_s, lhsc, i, lhsp); ++i) + process_constraint (new_constraint (*lhsp, tmp)); + } +} + /* Handle aggregate copies by expanding into copies of the respective fields of the structures. */ @@ -3244,29 +3457,27 @@ do_structure_copy (tree lhsop, tree rhsop) || (lhsp->type == ADDRESSOF && lhsp->var == anything_id) || rhsp->type == DEREF) { - struct constraint_expr tmp; - tree tmpvar = create_tmp_var_raw (ptr_type_node, - "structcopydereftmp"); - tmp.var = get_vi_for_tree (tmpvar)->id; - tmp.type = SCALAR; - tmp.offset = 0; - for (j = 0; VEC_iterate (ce_s, rhsc, j, rhsp); ++j) - process_constraint (new_constraint (tmp, *rhsp)); - for (j = 0; VEC_iterate (ce_s, lhsc, j, lhsp); ++j) - process_constraint (new_constraint (*lhsp, tmp)); + if (lhsp->type == DEREF) + { + gcc_assert (VEC_length (ce_s, lhsc) == 1); + lhsp->offset = UNKNOWN_OFFSET; + } + if (rhsp->type == DEREF) + { + gcc_assert (VEC_length (ce_s, rhsc) == 1); + rhsp->offset = UNKNOWN_OFFSET; + } + process_all_all_constraints (lhsc, rhsc); } else if (lhsp->type == SCALAR && (rhsp->type == SCALAR || rhsp->type == ADDRESSOF)) { - tree lhsbase, rhsbase; HOST_WIDE_INT lhssize, lhsmaxsize, lhsoffset; HOST_WIDE_INT rhssize, rhsmaxsize, rhsoffset; unsigned k = 0; - lhsbase = get_ref_base_and_extent (lhsop, &lhsoffset, - &lhssize, &lhsmaxsize); - rhsbase = get_ref_base_and_extent (rhsop, &rhsoffset, - &rhssize, &rhsmaxsize); + get_ref_base_and_extent (lhsop, &lhsoffset, &lhssize, &lhsmaxsize); + get_ref_base_and_extent (rhsop, &rhsoffset, &rhssize, &rhsmaxsize); for (j = 0; VEC_iterate (ce_s, lhsc, j, lhsp);) { varinfo_t lhsv, rhsv; @@ -3315,6 +3526,40 @@ make_constraint_to (unsigned id, tree op) VEC_free (ce_s, heap, rhsc); } +/* Create a constraint ID = &FROM. */ + +static void +make_constraint_from (varinfo_t vi, int from) +{ + struct constraint_expr lhs, rhs; + + lhs.var = vi->id; + lhs.offset = 0; + lhs.type = SCALAR; + + rhs.var = from; + rhs.offset = 0; + rhs.type = ADDRESSOF; + process_constraint (new_constraint (lhs, rhs)); +} + +/* Create a constraint ID = FROM. */ + +static void +make_copy_constraint (varinfo_t vi, int from) +{ + struct constraint_expr lhs, rhs; + + lhs.var = vi->id; + lhs.offset = 0; + lhs.type = SCALAR; + + rhs.var = from; + rhs.offset = 0; + rhs.type = SCALAR; + process_constraint (new_constraint (lhs, rhs)); +} + /* Make constraints necessary to make OP escape. */ static void @@ -3323,99 +3568,296 @@ make_escape_constraint (tree op) make_constraint_to (escaped_id, op); } -/* For non-IPA mode, generate constraints necessary for a call on the - RHS. */ +/* Add constraints to that the solution of VI is transitively closed. */ static void -handle_rhs_call (gimple stmt, VEC(ce_s, heap) **results) +make_transitive_closure_constraints (varinfo_t vi) { - struct constraint_expr rhsc; - unsigned i; + struct constraint_expr lhs, rhs; - for (i = 0; i < gimple_call_num_args (stmt); ++i) + /* VAR = *VAR; */ + lhs.type = SCALAR; + lhs.var = vi->id; + lhs.offset = 0; + rhs.type = DEREF; + rhs.var = vi->id; + rhs.offset = 0; + process_constraint (new_constraint (lhs, rhs)); + + /* VAR = VAR + UNKNOWN; */ + lhs.type = SCALAR; + lhs.var = vi->id; + lhs.offset = 0; + rhs.type = SCALAR; + rhs.var = vi->id; + rhs.offset = UNKNOWN_OFFSET; + process_constraint (new_constraint (lhs, rhs)); +} + +/* Create a new artificial heap variable with NAME. + Return the created variable. */ + +static varinfo_t +make_heapvar_for (varinfo_t lhs, const char *name) +{ + varinfo_t vi; + tree heapvar = heapvar_lookup (lhs->decl, lhs->offset); + + if (heapvar == NULL_TREE) { - tree arg = gimple_call_arg (stmt, i); + var_ann_t ann; + heapvar = create_tmp_var_raw (ptr_type_node, name); + DECL_EXTERNAL (heapvar) = 1; - /* Find those pointers being passed, and make sure they end up - pointing to anything. */ - if (could_have_pointers (arg)) - make_escape_constraint (arg); + heapvar_insert (lhs->decl, lhs->offset, heapvar); + + ann = get_var_ann (heapvar); + ann->is_heapvar = 1; } - /* The static chain escapes as well. */ - if (gimple_call_chain (stmt)) - make_escape_constraint (gimple_call_chain (stmt)); + /* For global vars we need to add a heapvar to the list of referenced + vars of a different function than it was created for originally. */ + if (cfun && gimple_referenced_vars (cfun)) + add_referenced_var (heapvar); - /* Regular functions return nonlocal memory. */ - rhsc.var = nonlocal_id; - rhsc.offset = 0; - rhsc.type = SCALAR; - VEC_safe_push (ce_s, heap, *results, &rhsc); + vi = new_var_info (heapvar, name); + vi->is_artificial_var = true; + vi->is_heap_var = true; + vi->is_unknown_size_var = true; + vi->offset = 0; + vi->fullsize = ~0; + vi->size = ~0; + vi->is_full_var = true; + insert_vi_for_tree (heapvar, vi); + + return vi; } -/* For non-IPA mode, generate constraints necessary for a call - that returns a pointer and assigns it to LHS. This simply makes - the LHS point to global and escaped variables. */ +/* Create a new artificial heap variable with NAME and make a + constraint from it to LHS. Return the created variable. */ -static void -handle_lhs_call (tree lhs, int flags, VEC(ce_s, heap) *rhsc) +static varinfo_t +make_constraint_from_heapvar (varinfo_t lhs, const char *name) { - VEC(ce_s, heap) *lhsc = NULL; - unsigned int j; - struct constraint_expr *lhsp; + varinfo_t vi = make_heapvar_for (lhs, name); + make_constraint_from (lhs, vi->id); - get_constraint_for (lhs, &lhsc); + return vi; +} - if (flags & ECF_MALLOC) - { - struct constraint_expr rhsc; - tree heapvar = heapvar_lookup (lhs); - varinfo_t vi; +/* Create a new artificial heap variable with NAME and make a + constraint from it to LHS. Set flags according to a tag used + for tracking restrict pointers. */ - if (heapvar == NULL) - { - heapvar = create_tmp_var_raw (ptr_type_node, "HEAP"); - DECL_EXTERNAL (heapvar) = 1; - get_var_ann (heapvar)->is_heapvar = 1; - if (gimple_referenced_vars (cfun)) - add_referenced_var (heapvar); - heapvar_insert (lhs, heapvar); - } +static void +make_constraint_from_restrict (varinfo_t lhs, const char *name) +{ + varinfo_t vi; + vi = make_constraint_from_heapvar (lhs, name); + vi->is_restrict_var = 1; + vi->is_global_var = 0; + vi->is_special_var = 1; + vi->may_have_pointers = 0; +} - rhsc.var = create_variable_info_for (heapvar, - alias_get_name (heapvar)); - vi = get_varinfo (rhsc.var); - vi->is_artificial_var = 1; - vi->is_heap_var = 1; - vi->is_unknown_size_var = true; - vi->fullsize = ~0; - vi->size = ~0; - rhsc.type = ADDRESSOF; - rhsc.offset = 0; - for (j = 0; VEC_iterate (ce_s, lhsc, j, lhsp); j++) - process_constraint (new_constraint (*lhsp, rhsc)); - } - else if (VEC_length (ce_s, rhsc) > 0) - { - struct constraint_expr *lhsp, *rhsp; - unsigned int i, j; - /* If the store is to a global decl make sure to - add proper escape constraints. */ - lhs = get_base_address (lhs); - if (lhs - && DECL_P (lhs) - && is_global_var (lhs)) +/* In IPA mode there are varinfos for different aspects of reach + function designator. One for the points-to set of the return + value, one for the variables that are clobbered by the function, + one for its uses and one for each parameter (including a single + glob for remaining variadic arguments). */ + +enum { fi_clobbers = 1, fi_uses = 2, + fi_static_chain = 3, fi_result = 4, fi_parm_base = 5 }; + +/* Get a constraint for the requested part of a function designator FI + when operating in IPA mode. */ + +static struct constraint_expr +get_function_part_constraint (varinfo_t fi, unsigned part) +{ + struct constraint_expr c; + + gcc_assert (in_ipa_mode); + + if (fi->id == anything_id) + { + /* ??? We probably should have a ANYFN special variable. */ + c.var = anything_id; + c.offset = 0; + c.type = SCALAR; + } + else if (TREE_CODE (fi->decl) == FUNCTION_DECL) + { + varinfo_t ai = first_vi_for_offset (fi, part); + if (ai) + c.var = ai->id; + else + c.var = anything_id; + c.offset = 0; + c.type = SCALAR; + } + else + { + c.var = fi->id; + c.offset = part; + c.type = DEREF; + } + + return c; +} + +/* For non-IPA mode, generate constraints necessary for a call on the + RHS. */ + +static void +handle_rhs_call (gimple stmt, VEC(ce_s, heap) **results) +{ + struct constraint_expr rhsc; + unsigned i; + bool returns_uses = false; + + for (i = 0; i < gimple_call_num_args (stmt); ++i) + { + tree arg = gimple_call_arg (stmt, i); + int flags = gimple_call_arg_flags (stmt, i); + + /* If the argument is not used or it does not contain pointers + we can ignore it. */ + if ((flags & EAF_UNUSED) + || !could_have_pointers (arg)) + continue; + + /* As we compute ESCAPED context-insensitive we do not gain + any precision with just EAF_NOCLOBBER but not EAF_NOESCAPE + set. The argument would still get clobbered through the + escape solution. + ??? We might get away with less (and more precise) constraints + if using a temporary for transitively closing things. */ + if ((flags & EAF_NOCLOBBER) + && (flags & EAF_NOESCAPE)) { - struct constraint_expr tmpc; - tmpc.var = escaped_id; - tmpc.offset = 0; - tmpc.type = SCALAR; - VEC_safe_push (ce_s, heap, lhsc, &tmpc); + varinfo_t uses = get_call_use_vi (stmt); + if (!(flags & EAF_DIRECT)) + make_transitive_closure_constraints (uses); + make_constraint_to (uses->id, arg); + returns_uses = true; } - for (i = 0; VEC_iterate (ce_s, lhsc, i, lhsp); ++i) - for (j = 0; VEC_iterate (ce_s, rhsc, j, rhsp); ++j) - process_constraint (new_constraint (*lhsp, *rhsp)); + else if (flags & EAF_NOESCAPE) + { + varinfo_t uses = get_call_use_vi (stmt); + varinfo_t clobbers = get_call_clobber_vi (stmt); + if (!(flags & EAF_DIRECT)) + { + make_transitive_closure_constraints (uses); + make_transitive_closure_constraints (clobbers); + } + make_constraint_to (uses->id, arg); + make_constraint_to (clobbers->id, arg); + returns_uses = true; + } + else + make_escape_constraint (arg); + } + + /* If we added to the calls uses solution make sure we account for + pointers to it to be returned. */ + if (returns_uses) + { + rhsc.var = get_call_use_vi (stmt)->id; + rhsc.offset = 0; + rhsc.type = SCALAR; + VEC_safe_push (ce_s, heap, *results, &rhsc); + } + + /* The static chain escapes as well. */ + if (gimple_call_chain (stmt)) + make_escape_constraint (gimple_call_chain (stmt)); + + /* And if we applied NRV the address of the return slot escapes as well. */ + if (gimple_call_return_slot_opt_p (stmt) + && gimple_call_lhs (stmt) != NULL_TREE + && TREE_ADDRESSABLE (TREE_TYPE (gimple_call_lhs (stmt)))) + { + VEC(ce_s, heap) *tmpc = NULL; + struct constraint_expr lhsc, *c; + get_constraint_for_address_of (gimple_call_lhs (stmt), &tmpc); + lhsc.var = escaped_id; + lhsc.offset = 0; + lhsc.type = SCALAR; + for (i = 0; VEC_iterate (ce_s, tmpc, i, c); ++i) + process_constraint (new_constraint (lhsc, *c)); + VEC_free(ce_s, heap, tmpc); + } + + /* Regular functions return nonlocal memory. */ + rhsc.var = nonlocal_id; + rhsc.offset = 0; + rhsc.type = SCALAR; + VEC_safe_push (ce_s, heap, *results, &rhsc); +} + +/* For non-IPA mode, generate constraints necessary for a call + that returns a pointer and assigns it to LHS. This simply makes + the LHS point to global and escaped variables. */ + +static void +handle_lhs_call (gimple stmt, tree lhs, int flags, VEC(ce_s, heap) *rhsc, + tree fndecl) +{ + VEC(ce_s, heap) *lhsc = NULL; + + get_constraint_for (lhs, &lhsc); + /* If the store is to a global decl make sure to + add proper escape constraints. */ + lhs = get_base_address (lhs); + if (lhs + && DECL_P (lhs) + && is_global_var (lhs)) + { + struct constraint_expr tmpc; + tmpc.var = escaped_id; + tmpc.offset = 0; + tmpc.type = SCALAR; + VEC_safe_push (ce_s, heap, lhsc, &tmpc); + } + + /* If the call returns an argument unmodified override the rhs + constraints. */ + flags = gimple_call_return_flags (stmt); + if (flags & ERF_RETURNS_ARG + && (flags & ERF_RETURN_ARG_MASK) < gimple_call_num_args (stmt)) + { + tree arg; + rhsc = NULL; + arg = gimple_call_arg (stmt, flags & ERF_RETURN_ARG_MASK); + get_constraint_for (arg, &rhsc); + process_all_all_constraints (lhsc, rhsc); + VEC_free (ce_s, heap, rhsc); } + else if (flags & ERF_NOALIAS) + { + varinfo_t vi; + struct constraint_expr tmpc; + rhsc = NULL; + vi = make_heapvar_for (get_vi_for_tree (lhs), "HEAP"); + /* We delay marking allocated storage global until we know if + it escapes. */ + DECL_EXTERNAL (vi->decl) = 0; + vi->is_global_var = 0; + /* If this is not a real malloc call assume the memory was + initialized and thus may point to global memory. All + builtin functions with the malloc attribute behave in a sane way. */ + if (!fndecl + || DECL_BUILT_IN_CLASS (fndecl) != BUILT_IN_NORMAL) + make_constraint_from (vi, nonlocal_id); + tmpc.var = vi->id; + tmpc.offset = 0; + tmpc.type = ADDRESSOF; + VEC_safe_push (ce_s, heap, rhsc, &tmpc); + } + + process_all_all_constraints (lhsc, rhsc); + VEC_free (ce_s, heap, lhsc); } @@ -3425,16 +3867,17 @@ handle_lhs_call (tree lhs, int flags, VEC(ce_s, heap) *rhsc) static void handle_const_call (gimple stmt, VEC(ce_s, heap) **results) { - struct constraint_expr rhsc, tmpc; - tree tmpvar = NULL_TREE; + struct constraint_expr rhsc; unsigned int k; /* Treat nested const functions the same as pure functions as far as the static chain is concerned. */ if (gimple_call_chain (stmt)) { - make_constraint_to (callused_id, gimple_call_chain (stmt)); - rhsc.var = callused_id; + varinfo_t uses = get_call_use_vi (stmt); + make_transitive_closure_constraints (uses); + make_constraint_to (uses->id, gimple_call_chain (stmt)); + rhsc.var = uses->id; rhsc.offset = 0; rhsc.type = SCALAR; VEC_safe_push (ce_s, heap, *results, &rhsc); @@ -3448,27 +3891,14 @@ handle_const_call (gimple stmt, VEC(ce_s, heap) **results) if (could_have_pointers (arg)) { VEC(ce_s, heap) *argc = NULL; + unsigned i; struct constraint_expr *argp; - int i; - - /* We always use a temporary here, otherwise we end up with - a quadratic amount of constraints for - large_struct = const_call (large_struct); - with field-sensitive PTA. */ - if (tmpvar == NULL_TREE) - { - tmpvar = create_tmp_var_raw (ptr_type_node, "consttmp"); - tmpc = get_constraint_exp_for_temp (tmpvar); - } - get_constraint_for (arg, &argc); - for (i = 0; VEC_iterate (ce_s, argc, i, argp); i++) - process_constraint (new_constraint (tmpc, *argp)); - VEC_free (ce_s, heap, argc); + for (i = 0; VEC_iterate (ce_s, argc, i, argp); ++i) + VEC_safe_push (ce_s, heap, *results, argp); + VEC_free(ce_s, heap, argc); } } - if (tmpvar != NULL_TREE) - VEC_safe_push (ce_s, heap, *results, &tmpc); /* May return addresses of globals. */ rhsc.var = nonlocal_id; @@ -3485,7 +3915,7 @@ handle_pure_call (gimple stmt, VEC(ce_s, heap) **results) { struct constraint_expr rhsc; unsigned i; - bool need_callused = false; + varinfo_t uses = NULL; /* Memory reached from pointer arguments is call-used. */ for (i = 0; i < gimple_call_num_args (stmt); ++i) @@ -3494,22 +3924,30 @@ handle_pure_call (gimple stmt, VEC(ce_s, heap) **results) if (could_have_pointers (arg)) { - make_constraint_to (callused_id, arg); - need_callused = true; + if (!uses) + { + uses = get_call_use_vi (stmt); + make_transitive_closure_constraints (uses); + } + make_constraint_to (uses->id, arg); } } /* The static chain is used as well. */ if (gimple_call_chain (stmt)) { - make_constraint_to (callused_id, gimple_call_chain (stmt)); - need_callused = true; + if (!uses) + { + uses = get_call_use_vi (stmt); + make_transitive_closure_constraints (uses); + } + make_constraint_to (uses->id, gimple_call_chain (stmt)); } - /* Pure functions may return callused and nonlocal memory. */ - if (need_callused) + /* Pure functions may return call-used and nonlocal memory. */ + if (uses) { - rhsc.var = callused_id; + rhsc.var = uses->id; rhsc.offset = 0; rhsc.type = SCALAR; VEC_safe_push (ce_s, heap, *results, &rhsc); @@ -3520,6 +3958,40 @@ handle_pure_call (gimple stmt, VEC(ce_s, heap) **results) VEC_safe_push (ce_s, heap, *results, &rhsc); } + +/* Return the varinfo for the callee of CALL. */ + +static varinfo_t +get_fi_for_callee (gimple call) +{ + tree decl; + + /* If we can directly resolve the function being called, do so. + Otherwise, it must be some sort of indirect expression that + we should still be able to handle. */ + decl = gimple_call_fndecl (call); + if (decl) + return get_vi_for_tree (decl); + + decl = gimple_call_fn (call); + /* The function can be either an SSA name pointer or, + worse, an OBJ_TYPE_REF. In this case we have no + clue and should be getting ANYFN (well, ANYTHING for now). */ + if (TREE_CODE (decl) == SSA_NAME) + { + if (TREE_CODE (decl) == SSA_NAME + && TREE_CODE (SSA_NAME_VAR (decl)) == PARM_DECL + && SSA_NAME_IS_DEFAULT_DEF (decl)) + decl = SSA_NAME_VAR (decl); + return get_vi_for_tree (decl); + } + else if (TREE_CODE (decl) == INTEGER_CST + || TREE_CODE (decl) == OBJ_TYPE_REF) + return get_varinfo (anything_id); + else + gcc_unreachable (); +} + /* Walk statement T setting up aliasing constraints according to the references found in T. This function is the main part of the constraint builder. AI points to auxiliary alias information used @@ -3532,7 +4004,7 @@ find_func_aliases (gimple origt) VEC(ce_s, heap) *lhsc = NULL; VEC(ce_s, heap) *rhsc = NULL; struct constraint_expr *c; - enum escape_type stmt_escape_type; + varinfo_t fi; /* Now build constraints expressions. */ if (gimple_code (t) == GIMPLE_PHI) @@ -3551,11 +4023,9 @@ find_func_aliases (gimple origt) get_constraint_for (gimple_phi_result (t), &lhsc); for (i = 0; i < gimple_phi_num_args (t); i++) { - tree rhstype; tree strippedrhs = PHI_ARG_DEF (t, i); STRIP_NOPS (strippedrhs); - rhstype = TREE_TYPE (strippedrhs); get_constraint_for (gimple_phi_arg_def (t, i), &rhsc); for (j = 0; VEC_iterate (ce_s, lhsc, j, c); j++) @@ -3580,7 +4050,224 @@ find_func_aliases (gimple origt) pointer passed by address. */ else if (is_gimple_call (t)) { - if (!in_ipa_mode) + tree fndecl = gimple_call_fndecl (t); + if (fndecl != NULL_TREE + && DECL_BUILT_IN_CLASS (fndecl) == BUILT_IN_NORMAL) + /* ??? All builtins that are handled here need to be handled + in the alias-oracle query functions explicitly! */ + switch (DECL_FUNCTION_CODE (fndecl)) + { + /* All the following functions return a pointer to the same object + as their first argument points to. The functions do not add + to the ESCAPED solution. The functions make the first argument + pointed to memory point to what the second argument pointed to + memory points to. */ + case BUILT_IN_STRCPY: + case BUILT_IN_STRNCPY: + case BUILT_IN_BCOPY: + case BUILT_IN_MEMCPY: + case BUILT_IN_MEMMOVE: + case BUILT_IN_MEMPCPY: + case BUILT_IN_STPCPY: + case BUILT_IN_STPNCPY: + case BUILT_IN_STRCAT: + case BUILT_IN_STRNCAT: + { + tree res = gimple_call_lhs (t); + tree dest = gimple_call_arg (t, (DECL_FUNCTION_CODE (fndecl) + == BUILT_IN_BCOPY ? 1 : 0)); + tree src = gimple_call_arg (t, (DECL_FUNCTION_CODE (fndecl) + == BUILT_IN_BCOPY ? 0 : 1)); + if (res != NULL_TREE) + { + get_constraint_for (res, &lhsc); + if (DECL_FUNCTION_CODE (fndecl) == BUILT_IN_MEMPCPY + || DECL_FUNCTION_CODE (fndecl) == BUILT_IN_STPCPY + || DECL_FUNCTION_CODE (fndecl) == BUILT_IN_STPNCPY) + get_constraint_for_ptr_offset (dest, NULL_TREE, &rhsc); + else + get_constraint_for (dest, &rhsc); + process_all_all_constraints (lhsc, rhsc); + VEC_free (ce_s, heap, lhsc); + VEC_free (ce_s, heap, rhsc); + } + get_constraint_for_ptr_offset (dest, NULL_TREE, &lhsc); + get_constraint_for_ptr_offset (src, NULL_TREE, &rhsc); + do_deref (&lhsc); + do_deref (&rhsc); + process_all_all_constraints (lhsc, rhsc); + VEC_free (ce_s, heap, lhsc); + VEC_free (ce_s, heap, rhsc); + return; + } + case BUILT_IN_MEMSET: + { + tree res = gimple_call_lhs (t); + tree dest = gimple_call_arg (t, 0); + unsigned i; + ce_s *lhsp; + struct constraint_expr ac; + if (res != NULL_TREE) + { + get_constraint_for (res, &lhsc); + get_constraint_for (dest, &rhsc); + process_all_all_constraints (lhsc, rhsc); + VEC_free (ce_s, heap, lhsc); + VEC_free (ce_s, heap, rhsc); + } + get_constraint_for_ptr_offset (dest, NULL_TREE, &lhsc); + do_deref (&lhsc); + if (flag_delete_null_pointer_checks + && integer_zerop (gimple_call_arg (t, 1))) + { + ac.type = ADDRESSOF; + ac.var = nothing_id; + } + else + { + ac.type = SCALAR; + ac.var = integer_id; + } + ac.offset = 0; + for (i = 0; VEC_iterate (ce_s, lhsc, i, lhsp); ++i) + process_constraint (new_constraint (*lhsp, ac)); + VEC_free (ce_s, heap, lhsc); + return; + } + /* All the following functions do not return pointers, do not + modify the points-to sets of memory reachable from their + arguments and do not add to the ESCAPED solution. */ + case BUILT_IN_SINCOS: + case BUILT_IN_SINCOSF: + case BUILT_IN_SINCOSL: + case BUILT_IN_FREXP: + case BUILT_IN_FREXPF: + case BUILT_IN_FREXPL: + case BUILT_IN_GAMMA_R: + case BUILT_IN_GAMMAF_R: + case BUILT_IN_GAMMAL_R: + case BUILT_IN_LGAMMA_R: + case BUILT_IN_LGAMMAF_R: + case BUILT_IN_LGAMMAL_R: + case BUILT_IN_MODF: + case BUILT_IN_MODFF: + case BUILT_IN_MODFL: + case BUILT_IN_REMQUO: + case BUILT_IN_REMQUOF: + case BUILT_IN_REMQUOL: + case BUILT_IN_FREE: + return; + /* Trampolines are special - they set up passing the static + frame. */ + case BUILT_IN_INIT_TRAMPOLINE: + { + tree tramp = gimple_call_arg (t, 0); + tree nfunc = gimple_call_arg (t, 1); + tree frame = gimple_call_arg (t, 2); + unsigned i; + struct constraint_expr lhs, *rhsp; + if (in_ipa_mode) + { + varinfo_t nfi = NULL; + gcc_assert (TREE_CODE (nfunc) == ADDR_EXPR); + nfi = lookup_vi_for_tree (TREE_OPERAND (nfunc, 0)); + if (nfi) + { + lhs = get_function_part_constraint (nfi, fi_static_chain); + get_constraint_for (frame, &rhsc); + for (i = 0; VEC_iterate (ce_s, rhsc, i, rhsp); ++i) + process_constraint (new_constraint (lhs, *rhsp)); + VEC_free (ce_s, heap, rhsc); + + /* Make the frame point to the function for + the trampoline adjustment call. */ + get_constraint_for (tramp, &lhsc); + do_deref (&lhsc); + get_constraint_for (nfunc, &rhsc); + process_all_all_constraints (lhsc, rhsc); + VEC_free (ce_s, heap, rhsc); + VEC_free (ce_s, heap, lhsc); + + return; + } + } + /* Else fallthru to generic handling which will let + the frame escape. */ + break; + } + case BUILT_IN_ADJUST_TRAMPOLINE: + { + tree tramp = gimple_call_arg (t, 0); + tree res = gimple_call_lhs (t); + if (in_ipa_mode && res) + { + get_constraint_for (res, &lhsc); + get_constraint_for (tramp, &rhsc); + do_deref (&rhsc); + process_all_all_constraints (lhsc, rhsc); + VEC_free (ce_s, heap, rhsc); + VEC_free (ce_s, heap, lhsc); + } + return; + } + /* Variadic argument handling needs to be handled in IPA + mode as well. */ + case BUILT_IN_VA_START: + { + if (in_ipa_mode) + { + tree valist = gimple_call_arg (t, 0); + struct constraint_expr rhs, *lhsp; + unsigned i; + /* The va_list gets access to pointers in variadic + arguments. */ + fi = lookup_vi_for_tree (cfun->decl); + gcc_assert (fi != NULL); + get_constraint_for (valist, &lhsc); + do_deref (&lhsc); + rhs = get_function_part_constraint (fi, ~0); + rhs.type = ADDRESSOF; + for (i = 0; VEC_iterate (ce_s, lhsc, i, lhsp); ++i) + process_constraint (new_constraint (*lhsp, rhs)); + VEC_free (ce_s, heap, lhsc); + /* va_list is clobbered. */ + make_constraint_to (get_call_clobber_vi (t)->id, valist); + return; + } + break; + } + /* va_end doesn't have any effect that matters. */ + case BUILT_IN_VA_END: + return; + /* Alternate return. Simply give up for now. */ + case BUILT_IN_RETURN: + { + fi = NULL; + if (!in_ipa_mode + || !(fi = get_vi_for_tree (cfun->decl))) + make_constraint_from (get_varinfo (escaped_id), anything_id); + else if (in_ipa_mode + && fi != NULL) + { + struct constraint_expr lhs, rhs; + lhs = get_function_part_constraint (fi, fi_result); + rhs.var = anything_id; + rhs.offset = 0; + rhs.type = SCALAR; + process_constraint (new_constraint (lhs, rhs)); + } + return; + } + /* printf-style functions may have hooks to set pointers to + point to somewhere into the generated string. Leave them + for a later excercise... */ + default: + /* Fallthru to general call handling. */; + } + if (!in_ipa_mode + || (fndecl + && (!(fi = lookup_vi_for_tree (fndecl)) + || !fi->is_fn_info))) { VEC(ce_s, heap) *rhsc = NULL; int flags = gimple_call_flags (t); @@ -3602,30 +4289,15 @@ find_func_aliases (gimple origt) handle_rhs_call (t, &rhsc); if (gimple_call_lhs (t) && could_have_pointers (gimple_call_lhs (t))) - handle_lhs_call (gimple_call_lhs (t), flags, rhsc); + handle_lhs_call (t, gimple_call_lhs (t), flags, rhsc, fndecl); VEC_free (ce_s, heap, rhsc); } else { tree lhsop; - varinfo_t fi; - int i = 1; - size_t j; - tree decl; - - lhsop = gimple_call_lhs (t); - decl = gimple_call_fndecl (t); + unsigned j; - /* If we can directly resolve the function being called, do so. - Otherwise, it must be some sort of indirect expression that - we should still be able to handle. */ - if (decl) - fi = get_vi_for_tree (decl); - else - { - decl = gimple_call_fn (t); - fi = get_vi_for_tree (decl); - } + fi = get_fi_for_callee (t); /* Assign all the passed arguments to the appropriate incoming parameters of the function. */ @@ -3635,58 +4307,77 @@ find_func_aliases (gimple origt) struct constraint_expr *rhsp; tree arg = gimple_call_arg (t, j); + if (!could_have_pointers (arg)) + continue; + get_constraint_for (arg, &rhsc); - if (TREE_CODE (decl) != FUNCTION_DECL) - { - lhs.type = DEREF; - lhs.var = fi->id; - lhs.offset = i; - } - else - { - lhs.type = SCALAR; - lhs.var = first_vi_for_offset (fi, i)->id; - lhs.offset = 0; - } + lhs = get_function_part_constraint (fi, fi_parm_base + j); while (VEC_length (ce_s, rhsc) != 0) { rhsp = VEC_last (ce_s, rhsc); process_constraint (new_constraint (lhs, *rhsp)); VEC_pop (ce_s, rhsc); } - i++; } /* If we are returning a value, assign it to the result. */ - if (lhsop) + lhsop = gimple_call_lhs (t); + if (lhsop + && type_could_have_pointers (TREE_TYPE (lhsop))) { struct constraint_expr rhs; struct constraint_expr *lhsp; - unsigned int j = 0; get_constraint_for (lhsop, &lhsc); - if (TREE_CODE (decl) != FUNCTION_DECL) - { - rhs.type = DEREF; - rhs.var = fi->id; - rhs.offset = i; - } - else + rhs = get_function_part_constraint (fi, fi_result); + if (fndecl + && DECL_RESULT (fndecl) + && DECL_BY_REFERENCE (DECL_RESULT (fndecl))) { - rhs.type = SCALAR; - rhs.var = first_vi_for_offset (fi, i)->id; - rhs.offset = 0; + VEC(ce_s, heap) *tem = NULL; + VEC_safe_push (ce_s, heap, tem, &rhs); + do_deref (&tem); + rhs = *VEC_index (ce_s, tem, 0); + VEC_free(ce_s, heap, tem); } for (j = 0; VEC_iterate (ce_s, lhsc, j, lhsp); j++) process_constraint (new_constraint (*lhsp, rhs)); } + + /* If we pass the result decl by reference, honor that. */ + if (lhsop + && fndecl + && DECL_RESULT (fndecl) + && DECL_BY_REFERENCE (DECL_RESULT (fndecl))) + { + struct constraint_expr lhs; + struct constraint_expr *rhsp; + + get_constraint_for_address_of (lhsop, &rhsc); + lhs = get_function_part_constraint (fi, fi_result); + for (j = 0; VEC_iterate (ce_s, rhsc, j, rhsp); j++) + process_constraint (new_constraint (lhs, *rhsp)); + VEC_free (ce_s, heap, rhsc); + } + + /* If we use a static chain, pass it along. */ + if (gimple_call_chain (t)) + { + struct constraint_expr lhs; + struct constraint_expr *rhsp; + + get_constraint_for (gimple_call_chain (t), &rhsc); + lhs = get_function_part_constraint (fi, fi_static_chain); + for (j = 0; VEC_iterate (ce_s, rhsc, j, rhsp); j++) + process_constraint (new_constraint (lhs, *rhsp)); + } } } /* Otherwise, just a regular assignment statement. Only care about operations with pointer result, others are dealt with as escape points if they have pointer operands. */ else if (is_gimple_assign (t) - && could_have_pointers (gimple_assign_lhs (t))) + && type_could_have_pointers (TREE_TYPE (gimple_assign_lhs (t)))) { /* Otherwise, just a regular assignment statement. */ tree lhsop = gimple_assign_lhs (t); @@ -3696,7 +4387,6 @@ find_func_aliases (gimple origt) do_structure_copy (lhsop, rhsop); else { - unsigned int j; struct constraint_expr temp; get_constraint_for (lhsop, &lhsc); @@ -3715,55 +4405,56 @@ find_func_aliases (gimple origt) temp.offset = 0; VEC_safe_push (ce_s, heap, rhsc, &temp); } - for (j = 0; VEC_iterate (ce_s, lhsc, j, c); j++) - { - struct constraint_expr *c2; - unsigned int k; - - for (k = 0; VEC_iterate (ce_s, rhsc, k, c2); k++) - process_constraint (new_constraint (*c, *c2)); - } + process_all_all_constraints (lhsc, rhsc); } + /* If there is a store to a global variable the rhs escapes. */ + if ((lhsop = get_base_address (lhsop)) != NULL_TREE + && DECL_P (lhsop) + && is_global_var (lhsop) + && (!in_ipa_mode + || DECL_EXTERNAL (lhsop) || TREE_PUBLIC (lhsop))) + make_escape_constraint (rhsop); + /* If this is a conversion of a non-restrict pointer to a + restrict pointer track it with a new heapvar. */ + else if (gimple_assign_cast_p (t) + && POINTER_TYPE_P (TREE_TYPE (rhsop)) + && POINTER_TYPE_P (TREE_TYPE (lhsop)) + && !TYPE_RESTRICT (TREE_TYPE (rhsop)) + && TYPE_RESTRICT (TREE_TYPE (lhsop))) + make_constraint_from_restrict (get_vi_for_tree (lhsop), + "CAST_RESTRICT"); } - else if (gimple_code (t) == GIMPLE_CHANGE_DYNAMIC_TYPE) + /* For conversions of pointers to non-pointers the pointer escapes. */ + else if (gimple_assign_cast_p (t) + && POINTER_TYPE_P (TREE_TYPE (gimple_assign_rhs1 (t))) + && !POINTER_TYPE_P (TREE_TYPE (gimple_assign_lhs (t)))) { - unsigned int j; - - get_constraint_for (gimple_cdt_location (t), &lhsc); - for (j = 0; VEC_iterate (ce_s, lhsc, j, c); ++j) - get_varinfo (c->var)->no_tbaa_pruning = true; + make_escape_constraint (gimple_assign_rhs1 (t)); } - - stmt_escape_type = is_escape_site (t); - if (stmt_escape_type == ESCAPE_STORED_IN_GLOBAL) + /* Handle escapes through return. */ + else if (gimple_code (t) == GIMPLE_RETURN + && gimple_return_retval (t) != NULL_TREE + && could_have_pointers (gimple_return_retval (t))) { - gcc_assert (is_gimple_assign (t)); - if (gimple_assign_rhs_code (t) == ADDR_EXPR) - { - tree rhs = gimple_assign_rhs1 (t); - tree base = get_base_address (TREE_OPERAND (rhs, 0)); - if (base - && (!DECL_P (base) - || !is_global_var (base))) - make_escape_constraint (rhs); - } - else if (get_gimple_rhs_class (gimple_assign_rhs_code (t)) - == GIMPLE_SINGLE_RHS) + fi = NULL; + if (!in_ipa_mode + || !(fi = get_vi_for_tree (cfun->decl))) + make_escape_constraint (gimple_return_retval (t)); + else if (in_ipa_mode + && fi != NULL) { - if (could_have_pointers (gimple_assign_rhs1 (t))) - make_escape_constraint (gimple_assign_rhs1 (t)); + struct constraint_expr lhs ; + struct constraint_expr *rhsp; + unsigned i; + + lhs = get_function_part_constraint (fi, fi_result); + get_constraint_for (gimple_return_retval (t), &rhsc); + for (i = 0; VEC_iterate (ce_s, rhsc, i, rhsp); i++) + process_constraint (new_constraint (lhs, *rhsp)); } - else - gcc_unreachable (); } - else if (stmt_escape_type == ESCAPE_BAD_CAST) - { - gcc_assert (is_gimple_assign (t)); - gcc_assert (CONVERT_EXPR_CODE_P (gimple_assign_rhs_code (t)) - || gimple_assign_rhs_code (t) == VIEW_CONVERT_EXPR); - make_escape_constraint (gimple_assign_rhs1 (t)); - } - else if (stmt_escape_type == ESCAPE_TO_ASM) + /* Handle asms conservatively by adding escape constraints to everything. */ + else if (gimple_code (t) == GIMPLE_ASM) { unsigned i, noutputs; const char **oconstraints; @@ -3817,8 +4508,8 @@ find_func_aliases (gimple origt) if (!allows_reg && allows_mem) make_escape_constraint (build_fold_addr_expr (op)); /* Strictly we'd only need the constraint to ESCAPED if - the asm clobbers memory, otherwise using CALLUSED - would be enough. */ + the asm clobbers memory, otherwise using something + along the lines of per-call clobbers/uses would be enough. */ else if (op && could_have_pointers (op)) make_escape_constraint (op); } @@ -3829,106 +4520,342 @@ find_func_aliases (gimple origt) } -/* Find the first varinfo in the same variable as START that overlaps with - OFFSET. Return NULL if we can't find one. */ +/* Create a constraint adding to the clobber set of FI the memory + pointed to by PTR. */ -static varinfo_t -first_vi_for_offset (varinfo_t start, unsigned HOST_WIDE_INT offset) +static void +process_ipa_clobber (varinfo_t fi, tree ptr) { - /* If the offset is outside of the variable, bail out. */ - if (offset >= start->fullsize) - return NULL; + VEC(ce_s, heap) *ptrc = NULL; + struct constraint_expr *c, lhs; + unsigned i; + get_constraint_for (ptr, &ptrc); + lhs = get_function_part_constraint (fi, fi_clobbers); + for (i = 0; VEC_iterate (ce_s, ptrc, i, c); i++) + process_constraint (new_constraint (lhs, *c)); + VEC_free (ce_s, heap, ptrc); +} - /* If we cannot reach offset from start, lookup the first field - and start from there. */ - if (start->offset > offset) - start = lookup_vi_for_tree (start->decl); +/* Walk statement T setting up clobber and use constraints according to the + references found in T. This function is a main part of the + IPA constraint builder. */ - while (start) - { - /* We may not find a variable in the field list with the actual - offset when when we have glommed a structure to a variable. - In that case, however, offset should still be within the size - of the variable. */ - if (offset >= start->offset - && offset < (start->offset + start->size)) - return start; +static void +find_func_clobbers (gimple origt) +{ + gimple t = origt; + VEC(ce_s, heap) *lhsc = NULL; + VEC(ce_s, heap) *rhsc = NULL; + varinfo_t fi; - start= start->next; - } + /* Add constraints for clobbered/used in IPA mode. + We are not interested in what automatic variables are clobbered + or used as we only use the information in the caller to which + they do not escape. */ + gcc_assert (in_ipa_mode); - return NULL; -} + /* If the stmt refers to memory in any way it better had a VUSE. */ + if (gimple_vuse (t) == NULL_TREE) + return; -/* Find the first varinfo in the same variable as START that overlaps with - OFFSET. If there is no such varinfo the varinfo directly preceding - OFFSET is returned. */ + /* We'd better have function information for the current function. */ + fi = lookup_vi_for_tree (cfun->decl); + gcc_assert (fi != NULL); -static varinfo_t -first_or_preceding_vi_for_offset (varinfo_t start, - unsigned HOST_WIDE_INT offset) -{ - /* If we cannot reach offset from start, lookup the first field - and start from there. */ - if (start->offset > offset) - start = lookup_vi_for_tree (start->decl); + /* Account for stores in assignments and calls. */ + if (gimple_vdef (t) != NULL_TREE + && gimple_has_lhs (t)) + { + tree lhs = gimple_get_lhs (t); + tree tem = lhs; + while (handled_component_p (tem)) + tem = TREE_OPERAND (tem, 0); + if ((DECL_P (tem) + && !auto_var_in_fn_p (tem, cfun->decl)) + || INDIRECT_REF_P (tem)) + { + struct constraint_expr lhsc, *rhsp; + unsigned i; + lhsc = get_function_part_constraint (fi, fi_clobbers); + get_constraint_for_address_of (lhs, &rhsc); + for (i = 0; VEC_iterate (ce_s, rhsc, i, rhsp); i++) + process_constraint (new_constraint (lhsc, *rhsp)); + VEC_free (ce_s, heap, rhsc); + } + } - /* We may not find a variable in the field list with the actual - offset when when we have glommed a structure to a variable. - In that case, however, offset should still be within the size - of the variable. - If we got beyond the offset we look for return the field - directly preceding offset which may be the last field. */ - while (start->next - && offset >= start->offset - && !(offset < (start->offset + start->size))) - start = start->next; + /* Account for uses in assigments and returns. */ + if (gimple_assign_single_p (t) + || (gimple_code (t) == GIMPLE_RETURN + && gimple_return_retval (t) != NULL_TREE)) + { + tree rhs = (gimple_assign_single_p (t) + ? gimple_assign_rhs1 (t) : gimple_return_retval (t)); + tree tem = rhs; + while (handled_component_p (tem)) + tem = TREE_OPERAND (tem, 0); + if ((DECL_P (tem) + && !auto_var_in_fn_p (tem, cfun->decl)) + || INDIRECT_REF_P (tem)) + { + struct constraint_expr lhs, *rhsp; + unsigned i; + lhs = get_function_part_constraint (fi, fi_uses); + get_constraint_for_address_of (rhs, &rhsc); + for (i = 0; VEC_iterate (ce_s, rhsc, i, rhsp); i++) + process_constraint (new_constraint (lhs, *rhsp)); + VEC_free (ce_s, heap, rhsc); + } + } - return start; -} + if (is_gimple_call (t)) + { + varinfo_t cfi = NULL; + tree decl = gimple_call_fndecl (t); + struct constraint_expr lhs, rhs; + unsigned i, j; + + /* For builtins we do not have separate function info. For those + we do not generate escapes for we have to generate clobbers/uses. */ + if (decl + && DECL_BUILT_IN_CLASS (decl) == BUILT_IN_NORMAL) + switch (DECL_FUNCTION_CODE (decl)) + { + /* The following functions use and clobber memory pointed to + by their arguments. */ + case BUILT_IN_STRCPY: + case BUILT_IN_STRNCPY: + case BUILT_IN_BCOPY: + case BUILT_IN_MEMCPY: + case BUILT_IN_MEMMOVE: + case BUILT_IN_MEMPCPY: + case BUILT_IN_STPCPY: + case BUILT_IN_STPNCPY: + case BUILT_IN_STRCAT: + case BUILT_IN_STRNCAT: + { + tree dest = gimple_call_arg (t, (DECL_FUNCTION_CODE (decl) + == BUILT_IN_BCOPY ? 1 : 0)); + tree src = gimple_call_arg (t, (DECL_FUNCTION_CODE (decl) + == BUILT_IN_BCOPY ? 0 : 1)); + unsigned i; + struct constraint_expr *rhsp, *lhsp; + get_constraint_for_ptr_offset (dest, NULL_TREE, &lhsc); + lhs = get_function_part_constraint (fi, fi_clobbers); + for (i = 0; VEC_iterate (ce_s, lhsc, i, lhsp); i++) + process_constraint (new_constraint (lhs, *lhsp)); + VEC_free (ce_s, heap, lhsc); + get_constraint_for_ptr_offset (src, NULL_TREE, &rhsc); + lhs = get_function_part_constraint (fi, fi_uses); + for (i = 0; VEC_iterate (ce_s, rhsc, i, rhsp); i++) + process_constraint (new_constraint (lhs, *rhsp)); + VEC_free (ce_s, heap, rhsc); + return; + } + /* The following function clobbers memory pointed to by + its argument. */ + case BUILT_IN_MEMSET: + { + tree dest = gimple_call_arg (t, 0); + unsigned i; + ce_s *lhsp; + get_constraint_for_ptr_offset (dest, NULL_TREE, &lhsc); + lhs = get_function_part_constraint (fi, fi_clobbers); + for (i = 0; VEC_iterate (ce_s, lhsc, i, lhsp); i++) + process_constraint (new_constraint (lhs, *lhsp)); + VEC_free (ce_s, heap, lhsc); + return; + } + /* The following functions clobber their second and third + arguments. */ + case BUILT_IN_SINCOS: + case BUILT_IN_SINCOSF: + case BUILT_IN_SINCOSL: + { + process_ipa_clobber (fi, gimple_call_arg (t, 1)); + process_ipa_clobber (fi, gimple_call_arg (t, 2)); + return; + } + /* The following functions clobber their second argument. */ + case BUILT_IN_FREXP: + case BUILT_IN_FREXPF: + case BUILT_IN_FREXPL: + case BUILT_IN_LGAMMA_R: + case BUILT_IN_LGAMMAF_R: + case BUILT_IN_LGAMMAL_R: + case BUILT_IN_GAMMA_R: + case BUILT_IN_GAMMAF_R: + case BUILT_IN_GAMMAL_R: + case BUILT_IN_MODF: + case BUILT_IN_MODFF: + case BUILT_IN_MODFL: + { + process_ipa_clobber (fi, gimple_call_arg (t, 1)); + return; + } + /* The following functions clobber their third argument. */ + case BUILT_IN_REMQUO: + case BUILT_IN_REMQUOF: + case BUILT_IN_REMQUOL: + { + process_ipa_clobber (fi, gimple_call_arg (t, 2)); + return; + } + /* The following functions neither read nor clobber memory. */ + case BUILT_IN_FREE: + return; + /* Trampolines are of no interest to us. */ + case BUILT_IN_INIT_TRAMPOLINE: + case BUILT_IN_ADJUST_TRAMPOLINE: + return; + case BUILT_IN_VA_START: + case BUILT_IN_VA_END: + return; + /* printf-style functions may have hooks to set pointers to + point to somewhere into the generated string. Leave them + for a later excercise... */ + default: + /* Fallthru to general call handling. */; + } + /* Parameters passed by value are used. */ + lhs = get_function_part_constraint (fi, fi_uses); + for (i = 0; i < gimple_call_num_args (t); i++) + { + struct constraint_expr *rhsp; + tree arg = gimple_call_arg (t, i); -/* Insert the varinfo FIELD into the field list for BASE, at the front - of the list. */ + if (TREE_CODE (arg) == SSA_NAME + || is_gimple_min_invariant (arg)) + continue; -static void -insert_into_field_list (varinfo_t base, varinfo_t field) -{ - varinfo_t prev = base; - varinfo_t curr = base->next; + get_constraint_for_address_of (arg, &rhsc); + for (j = 0; VEC_iterate (ce_s, rhsc, j, rhsp); j++) + process_constraint (new_constraint (lhs, *rhsp)); + VEC_free (ce_s, heap, rhsc); + } - field->next = curr; - prev->next = field; -} + /* Build constraints for propagating clobbers/uses along the + callgraph edges. */ + cfi = get_fi_for_callee (t); + if (cfi->id == anything_id) + { + if (gimple_vdef (t)) + make_constraint_from (first_vi_for_offset (fi, fi_clobbers), + anything_id); + make_constraint_from (first_vi_for_offset (fi, fi_uses), + anything_id); + return; + } -/* Insert the varinfo FIELD into the field list for BASE, ordered by - offset. */ + /* For callees without function info (that's external functions), + ESCAPED is clobbered and used. */ + if (gimple_call_fndecl (t) + && !cfi->is_fn_info) + { + varinfo_t vi; -static void -insert_into_field_list_sorted (varinfo_t base, varinfo_t field) -{ - varinfo_t prev = base; - varinfo_t curr = base->next; + if (gimple_vdef (t)) + make_copy_constraint (first_vi_for_offset (fi, fi_clobbers), + escaped_id); + make_copy_constraint (first_vi_for_offset (fi, fi_uses), escaped_id); + + /* Also honor the call statement use/clobber info. */ + if ((vi = lookup_call_clobber_vi (t)) != NULL) + make_copy_constraint (first_vi_for_offset (fi, fi_clobbers), + vi->id); + if ((vi = lookup_call_use_vi (t)) != NULL) + make_copy_constraint (first_vi_for_offset (fi, fi_uses), + vi->id); + return; + } - if (curr == NULL) + /* Otherwise the caller clobbers and uses what the callee does. + ??? This should use a new complex constraint that filters + local variables of the callee. */ + if (gimple_vdef (t)) + { + lhs = get_function_part_constraint (fi, fi_clobbers); + rhs = get_function_part_constraint (cfi, fi_clobbers); + process_constraint (new_constraint (lhs, rhs)); + } + lhs = get_function_part_constraint (fi, fi_uses); + rhs = get_function_part_constraint (cfi, fi_uses); + process_constraint (new_constraint (lhs, rhs)); + } + else if (gimple_code (t) == GIMPLE_ASM) { - prev->next = field; - field->next = NULL; + /* ??? Ick. We can do better. */ + if (gimple_vdef (t)) + make_constraint_from (first_vi_for_offset (fi, fi_clobbers), + anything_id); + make_constraint_from (first_vi_for_offset (fi, fi_uses), + anything_id); } - else + + VEC_free (ce_s, heap, rhsc); +} + + +/* Find the first varinfo in the same variable as START that overlaps with + OFFSET. Return NULL if we can't find one. */ + +static varinfo_t +first_vi_for_offset (varinfo_t start, unsigned HOST_WIDE_INT offset) +{ + /* If the offset is outside of the variable, bail out. */ + if (offset >= start->fullsize) + return NULL; + + /* If we cannot reach offset from start, lookup the first field + and start from there. */ + if (start->offset > offset) + start = lookup_vi_for_tree (start->decl); + + while (start) { - while (curr) - { - if (field->offset <= curr->offset) - break; - prev = curr; - curr = curr->next; - } - field->next = prev->next; - prev->next = field; + /* We may not find a variable in the field list with the actual + offset when when we have glommed a structure to a variable. + In that case, however, offset should still be within the size + of the variable. */ + if (offset >= start->offset + && (offset - start->offset) < start->size) + return start; + + start= start->next; } + + return NULL; +} + +/* Find the first varinfo in the same variable as START that overlaps with + OFFSET. If there is no such varinfo the varinfo directly preceding + OFFSET is returned. */ + +static varinfo_t +first_or_preceding_vi_for_offset (varinfo_t start, + unsigned HOST_WIDE_INT offset) +{ + /* If we cannot reach offset from start, lookup the first field + and start from there. */ + if (start->offset > offset) + start = lookup_vi_for_tree (start->decl); + + /* We may not find a variable in the field list with the actual + offset when when we have glommed a structure to a variable. + In that case, however, offset should still be within the size + of the variable. + If we got beyond the offset we look for return the field + directly preceding offset which may be the last field. */ + while (start->next + && offset >= start->offset + && !((offset - start->offset) < start->size)) + start = start->next; + + return start; } + /* This structure is used during pushing fields onto the fieldstack to track the offset of the field, since bitpos_of_field gives it relative to its immediate containing type, and we want it relative @@ -3945,6 +4872,8 @@ struct fieldoff unsigned has_unknown_size : 1; unsigned may_have_pointers : 1; + + unsigned only_restrict_pointers : 1; }; typedef struct fieldoff fieldoff_s; @@ -4012,37 +4941,38 @@ var_can_have_subvars (const_tree v) OFFSET is used to keep track of the offset in this entire structure, rather than just the immediately containing structure. - Returns the number of fields pushed. */ + Returns false if the caller is supposed to handle the field we + recursed for. */ -static int +static bool push_fields_onto_fieldstack (tree type, VEC(fieldoff_s,heap) **fieldstack, - HOST_WIDE_INT offset) + HOST_WIDE_INT offset, bool must_have_pointers_p) { tree field; - int count = 0; + bool empty_p = true; if (TREE_CODE (type) != RECORD_TYPE) - return 0; + return false; /* If the vector of fields is growing too big, bail out early. Callers check for VEC_length <= MAX_FIELDS_FOR_FIELD_SENSITIVE, make sure this fails. */ if (VEC_length (fieldoff_s, *fieldstack) > MAX_FIELDS_FOR_FIELD_SENSITIVE) - return 0; + return false; for (field = TYPE_FIELDS (type); field; field = TREE_CHAIN (field)) if (TREE_CODE (field) == FIELD_DECL) { bool push = false; - int pushed = 0; HOST_WIDE_INT foff = bitpos_of_field (field); if (!var_can_have_subvars (field) || TREE_CODE (TREE_TYPE (field)) == QUAL_UNION_TYPE || TREE_CODE (TREE_TYPE (field)) == UNION_TYPE) push = true; - else if (!(pushed = push_fields_onto_fieldstack - (TREE_TYPE (field), fieldstack, offset + foff)) + else if (!push_fields_onto_fieldstack + (TREE_TYPE (field), fieldstack, offset + foff, + must_have_pointers_p) && (DECL_SIZE (field) && !integer_zerop (DECL_SIZE (field)))) /* Empty structures may have actual size, like in C++. So @@ -4065,12 +4995,12 @@ push_fields_onto_fieldstack (tree type, VEC(fieldoff_s,heap) **fieldstack, /* If adjacent fields do not contain pointers merge them. */ if (pair && !pair->may_have_pointers - && !could_have_pointers (field) && !pair->has_unknown_size && !has_unknown_size - && pair->offset + (HOST_WIDE_INT)pair->size == offset + foff) + && pair->offset + (HOST_WIDE_INT)pair->size == offset + foff + && !must_have_pointers_p + && !could_have_pointers (field)) { - pair = VEC_last (fieldoff_s, *fieldstack); pair->size += TREE_INT_CST_LOW (DECL_SIZE (field)); } else @@ -4082,49 +5012,19 @@ push_fields_onto_fieldstack (tree type, VEC(fieldoff_s,heap) **fieldstack, pair->size = TREE_INT_CST_LOW (DECL_SIZE (field)); else pair->size = -1; - pair->may_have_pointers = could_have_pointers (field); - count++; + pair->may_have_pointers + = must_have_pointers_p || could_have_pointers (field); + pair->only_restrict_pointers + = (!has_unknown_size + && POINTER_TYPE_P (TREE_TYPE (field)) + && TYPE_RESTRICT (TREE_TYPE (field))); } } - else - count += pushed; - } - - return count; -} - -/* Create a constraint ID = &FROM. */ - -static void -make_constraint_from (varinfo_t vi, int from) -{ - struct constraint_expr lhs, rhs; - - lhs.var = vi->id; - lhs.offset = 0; - lhs.type = SCALAR; - - rhs.var = from; - rhs.offset = 0; - rhs.type = ADDRESSOF; - process_constraint (new_constraint (lhs, rhs)); -} - -/* Create a constraint ID = FROM. */ - -static void -make_copy_constraint (varinfo_t vi, int from) -{ - struct constraint_expr lhs, rhs; - lhs.var = vi->id; - lhs.offset = 0; - lhs.type = SCALAR; + empty_p = false; + } - rhs.var = from; - rhs.offset = 0; - rhs.type = SCALAR; - process_constraint (new_constraint (lhs, rhs)); + return !empty_p; } /* Count the number of arguments DECL has, and set IS_VARARGS to true @@ -4133,86 +5033,166 @@ make_copy_constraint (varinfo_t vi, int from) static unsigned int count_num_arguments (tree decl, bool *is_varargs) { - unsigned int i = 0; + unsigned int num = 0; tree t; - for (t = TYPE_ARG_TYPES (TREE_TYPE (decl)); - t; - t = TREE_CHAIN (t)) - { - if (TREE_VALUE (t) == void_type_node) - break; - i++; - } + /* Capture named arguments for K&R functions. They do not + have a prototype and thus no TYPE_ARG_TYPES. */ + for (t = DECL_ARGUMENTS (decl); t; t = TREE_CHAIN (t)) + ++num; + /* Check if the function has variadic arguments. */ + for (t = TYPE_ARG_TYPES (TREE_TYPE (decl)); t; t = TREE_CHAIN (t)) + if (TREE_VALUE (t) == void_type_node) + break; if (!t) *is_varargs = true; - return i; + + return num; } /* Creation function node for DECL, using NAME, and return the index of the variable we've created for the function. */ -static unsigned int +static varinfo_t create_function_info_for (tree decl, const char *name) { - unsigned int index = VEC_length (varinfo_t, varmap); - varinfo_t vi; + struct function *fn = DECL_STRUCT_FUNCTION (decl); + varinfo_t vi, prev_vi; tree arg; unsigned int i; bool is_varargs = false; + unsigned int num_args = count_num_arguments (decl, &is_varargs); /* Create the variable info. */ - vi = new_var_info (decl, index, name); - vi->decl = decl; + vi = new_var_info (decl, name); vi->offset = 0; vi->size = 1; - vi->fullsize = count_num_arguments (decl, &is_varargs) + 1; + vi->fullsize = fi_parm_base + num_args; + vi->is_fn_info = 1; + vi->may_have_pointers = false; + if (is_varargs) + vi->fullsize = ~0; insert_vi_for_tree (vi->decl, vi); - VEC_safe_push (varinfo_t, heap, varmap, vi); - stats.total_vars++; + prev_vi = vi; - /* If it's varargs, we don't know how many arguments it has, so we - can't do much. */ - if (is_varargs) + /* Create a variable for things the function clobbers and one for + things the function uses. */ { - vi->fullsize = ~0; - vi->size = ~0; - vi->is_unknown_size_var = true; - return index; + varinfo_t clobbervi, usevi; + const char *newname; + char *tempname; + + asprintf (&tempname, "%s.clobber", name); + newname = ggc_strdup (tempname); + free (tempname); + + clobbervi = new_var_info (NULL, newname); + clobbervi->offset = fi_clobbers; + clobbervi->size = 1; + clobbervi->fullsize = vi->fullsize; + clobbervi->is_full_var = true; + clobbervi->is_global_var = false; + gcc_assert (prev_vi->offset < clobbervi->offset); + prev_vi->next = clobbervi; + prev_vi = clobbervi; + + asprintf (&tempname, "%s.use", name); + newname = ggc_strdup (tempname); + free (tempname); + + usevi = new_var_info (NULL, newname); + usevi->offset = fi_uses; + usevi->size = 1; + usevi->fullsize = vi->fullsize; + usevi->is_full_var = true; + usevi->is_global_var = false; + gcc_assert (prev_vi->offset < usevi->offset); + prev_vi->next = usevi; + prev_vi = usevi; } + /* And one for the static chain. */ + if (fn->static_chain_decl != NULL_TREE) + { + varinfo_t chainvi; + const char *newname; + char *tempname; - arg = DECL_ARGUMENTS (decl); + asprintf (&tempname, "%s.chain", name); + newname = ggc_strdup (tempname); + free (tempname); + + chainvi = new_var_info (fn->static_chain_decl, newname); + chainvi->offset = fi_static_chain; + chainvi->size = 1; + chainvi->fullsize = vi->fullsize; + chainvi->is_full_var = true; + chainvi->is_global_var = false; + gcc_assert (prev_vi->offset < chainvi->offset); + prev_vi->next = chainvi; + prev_vi = chainvi; + insert_vi_for_tree (fn->static_chain_decl, chainvi); + } + + /* Create a variable for the return var. */ + if (DECL_RESULT (decl) != NULL + || !VOID_TYPE_P (TREE_TYPE (TREE_TYPE (decl)))) + { + varinfo_t resultvi; + const char *newname; + char *tempname; + tree resultdecl = decl; + + if (DECL_RESULT (decl)) + resultdecl = DECL_RESULT (decl); + + asprintf (&tempname, "%s.result", name); + newname = ggc_strdup (tempname); + free (tempname); + + resultvi = new_var_info (resultdecl, newname); + resultvi->offset = fi_result; + resultvi->size = 1; + resultvi->fullsize = vi->fullsize; + resultvi->is_full_var = true; + if (DECL_RESULT (decl)) + resultvi->may_have_pointers = could_have_pointers (DECL_RESULT (decl)); + gcc_assert (prev_vi->offset < resultvi->offset); + prev_vi->next = resultvi; + prev_vi = resultvi; + if (DECL_RESULT (decl)) + insert_vi_for_tree (DECL_RESULT (decl), resultvi); + } /* Set up variables for each argument. */ - for (i = 1; i < vi->fullsize; i++) + arg = DECL_ARGUMENTS (decl); + for (i = 0; i < num_args; i++) { varinfo_t argvi; const char *newname; char *tempname; - unsigned int newindex; tree argdecl = decl; if (arg) argdecl = arg; - newindex = VEC_length (varinfo_t, varmap); - asprintf (&tempname, "%s.arg%d", name, i-1); + asprintf (&tempname, "%s.arg%d", name, i); newname = ggc_strdup (tempname); free (tempname); - argvi = new_var_info (argdecl, newindex, newname); - argvi->decl = argdecl; - VEC_safe_push (varinfo_t, heap, varmap, argvi); - argvi->offset = i; + argvi = new_var_info (argdecl, newname); + argvi->offset = fi_parm_base + i; argvi->size = 1; argvi->is_full_var = true; argvi->fullsize = vi->fullsize; - insert_into_field_list_sorted (vi, argvi); - stats.total_vars ++; + if (arg) + argvi->may_have_pointers = could_have_pointers (arg); + gcc_assert (prev_vi->offset < argvi->offset); + prev_vi->next = argvi; + prev_vi = argvi; if (arg) { insert_vi_for_tree (arg, argvi); @@ -4220,39 +5200,34 @@ create_function_info_for (tree decl, const char *name) } } - /* Create a variable for the return var. */ - if (DECL_RESULT (decl) != NULL - || !VOID_TYPE_P (TREE_TYPE (TREE_TYPE (decl)))) + /* Add one representative for all further args. */ + if (is_varargs) { - varinfo_t resultvi; + varinfo_t argvi; const char *newname; char *tempname; - unsigned int newindex; - tree resultdecl = decl; - - vi->fullsize ++; - - if (DECL_RESULT (decl)) - resultdecl = DECL_RESULT (decl); + tree decl; - newindex = VEC_length (varinfo_t, varmap); - asprintf (&tempname, "%s.result", name); + asprintf (&tempname, "%s.varargs", name); newname = ggc_strdup (tempname); free (tempname); - resultvi = new_var_info (resultdecl, newindex, newname); - resultvi->decl = resultdecl; - VEC_safe_push (varinfo_t, heap, varmap, resultvi); - resultvi->offset = i; - resultvi->size = 1; - resultvi->fullsize = vi->fullsize; - resultvi->is_full_var = true; - insert_into_field_list_sorted (vi, resultvi); - stats.total_vars ++; - if (DECL_RESULT (decl)) - insert_vi_for_tree (DECL_RESULT (decl), resultvi); + /* We need sth that can be pointed to for va_start. */ + decl = create_tmp_var_raw (ptr_type_node, name); + get_var_ann (decl); + + argvi = new_var_info (decl, newname); + argvi->offset = fi_parm_base + num_args; + argvi->size = ~0; + argvi->is_full_var = true; + argvi->is_heap_var = true; + argvi->fullsize = vi->fullsize; + gcc_assert (prev_vi->offset < argvi->offset); + prev_vi->next = argvi; + prev_vi = argvi; } - return index; + + return vi; } @@ -4279,79 +5254,54 @@ check_for_overlaps (VEC (fieldoff_s,heap) *fieldstack) This will also create any varinfo structures necessary for fields of DECL. */ -static unsigned int -create_variable_info_for (tree decl, const char *name) +static varinfo_t +create_variable_info_for_1 (tree decl, const char *name) { - unsigned int index = VEC_length (varinfo_t, varmap); - varinfo_t vi; + varinfo_t vi, newvi; tree decl_type = TREE_TYPE (decl); tree declsize = DECL_P (decl) ? DECL_SIZE (decl) : TYPE_SIZE (decl_type); - bool is_global = DECL_P (decl) ? is_global_var (decl) : false; VEC (fieldoff_s,heap) *fieldstack = NULL; + fieldoff_s *fo; + unsigned int i; - if (TREE_CODE (decl) == FUNCTION_DECL && in_ipa_mode) - return create_function_info_for (decl, name); - - if (var_can_have_subvars (decl) && use_field_sensitive - && (!var_ann (decl) - || var_ann (decl)->noalias_state == 0) - && (!var_ann (decl) - || !var_ann (decl)->is_heapvar)) - push_fields_onto_fieldstack (decl_type, &fieldstack, 0); - - /* If the variable doesn't have subvars, we may end up needing to - sort the field list and create fake variables for all the - fields. */ - vi = new_var_info (decl, index, name); - vi->decl = decl; - vi->offset = 0; - vi->may_have_pointers = could_have_pointers (decl); if (!declsize || !host_integerp (declsize, 1)) { - vi->is_unknown_size_var = true; - vi->fullsize = ~0; + vi = new_var_info (decl, name); + vi->offset = 0; vi->size = ~0; - } - else - { - vi->fullsize = TREE_INT_CST_LOW (declsize); - vi->size = vi->fullsize; + vi->fullsize = ~0; + vi->is_unknown_size_var = true; + vi->is_full_var = true; + vi->may_have_pointers = could_have_pointers (decl); + return vi; } - insert_vi_for_tree (vi->decl, vi); - VEC_safe_push (varinfo_t, heap, varmap, vi); - if (is_global && (!flag_whole_program || !in_ipa_mode) - && vi->may_have_pointers) - { - if (var_ann (decl) - && var_ann (decl)->noalias_state == NO_ALIAS_ANYTHING) - make_constraint_from (vi, vi->id); - else - make_copy_constraint (vi, nonlocal_id); - } - - stats.total_vars++; + /* Collect field information. */ if (use_field_sensitive - && !vi->is_unknown_size_var && var_can_have_subvars (decl) - && VEC_length (fieldoff_s, fieldstack) > 1 - && VEC_length (fieldoff_s, fieldstack) <= MAX_FIELDS_FOR_FIELD_SENSITIVE) + /* ??? Force us to not use subfields for global initializers + in IPA mode. Else we'd have to parse arbitrary initializers. */ + && !(in_ipa_mode + && is_global_var (decl) + && DECL_INITIAL (decl))) { - unsigned int newindex = VEC_length (varinfo_t, varmap); fieldoff_s *fo = NULL; bool notokay = false; unsigned int i; + push_fields_onto_fieldstack (decl_type, &fieldstack, 0, + TREE_PUBLIC (decl) + || DECL_EXTERNAL (decl) + || TREE_ADDRESSABLE (decl)); + for (i = 0; !notokay && VEC_iterate (fieldoff_s, fieldstack, i, fo); i++) - { - if (fo->has_unknown_size - || fo->offset < 0) - { - notokay = true; - break; - } - } + if (fo->has_unknown_size + || fo->offset < 0) + { + notokay = true; + break; + } /* We can't sort them if we have a field with a variable sized type, which will make notokay = true. In that case, we are going to return @@ -4367,60 +5317,112 @@ create_variable_info_for (tree decl, const char *name) notokay = check_for_overlaps (fieldstack); } + if (notokay) + VEC_free (fieldoff_s, heap, fieldstack); + } + + /* If we didn't end up collecting sub-variables create a full + variable for the decl. */ + if (VEC_length (fieldoff_s, fieldstack) <= 1 + || VEC_length (fieldoff_s, fieldstack) > MAX_FIELDS_FOR_FIELD_SENSITIVE) + { + vi = new_var_info (decl, name); + vi->offset = 0; + vi->may_have_pointers = could_have_pointers (decl); + vi->fullsize = TREE_INT_CST_LOW (declsize); + vi->size = vi->fullsize; + vi->is_full_var = true; + VEC_free (fieldoff_s, heap, fieldstack); + return vi; + } - if (VEC_length (fieldoff_s, fieldstack) != 0) - fo = VEC_index (fieldoff_s, fieldstack, 0); + vi = new_var_info (decl, name); + vi->fullsize = TREE_INT_CST_LOW (declsize); + for (i = 0, newvi = vi; + VEC_iterate (fieldoff_s, fieldstack, i, fo); + ++i, newvi = newvi->next) + { + const char *newname = "NULL"; + char *tempname; - if (fo == NULL || notokay) + if (dump_file) { - vi->is_unknown_size_var = 1; - vi->fullsize = ~0; - vi->size = ~0; - vi->is_full_var = true; - VEC_free (fieldoff_s, heap, fieldstack); - return index; + asprintf (&tempname, "%s." HOST_WIDE_INT_PRINT_DEC + "+" HOST_WIDE_INT_PRINT_DEC, name, fo->offset, fo->size); + newname = ggc_strdup (tempname); + free (tempname); } + newvi->name = newname; + newvi->offset = fo->offset; + newvi->size = fo->size; + newvi->fullsize = vi->fullsize; + newvi->may_have_pointers = fo->may_have_pointers; + newvi->only_restrict_pointers = fo->only_restrict_pointers; + if (i + 1 < VEC_length (fieldoff_s, fieldstack)) + newvi->next = new_var_info (decl, name); + } - vi->size = fo->size; - vi->offset = fo->offset; - vi->may_have_pointers = fo->may_have_pointers; - for (i = VEC_length (fieldoff_s, fieldstack) - 1; - i >= 1 && VEC_iterate (fieldoff_s, fieldstack, i, fo); - i--) - { - varinfo_t newvi; - const char *newname = "NULL"; - char *tempname; + VEC_free (fieldoff_s, heap, fieldstack); + + return vi; +} + +static unsigned int +create_variable_info_for (tree decl, const char *name) +{ + varinfo_t vi = create_variable_info_for_1 (decl, name); + unsigned int id = vi->id; + + insert_vi_for_tree (decl, vi); - newindex = VEC_length (varinfo_t, varmap); - if (dump_file) + /* Create initial constraints for globals. */ + for (; vi; vi = vi->next) + { + if (!vi->may_have_pointers + || !vi->is_global_var) + continue; + + /* Mark global restrict qualified pointers. */ + if ((POINTER_TYPE_P (TREE_TYPE (decl)) + && TYPE_RESTRICT (TREE_TYPE (decl))) + || vi->only_restrict_pointers) + make_constraint_from_restrict (vi, "GLOBAL_RESTRICT"); + + /* For escaped variables initialize them from nonlocal. */ + if (!in_ipa_mode + || DECL_EXTERNAL (decl) || TREE_PUBLIC (decl)) + make_copy_constraint (vi, nonlocal_id); + + /* If this is a global variable with an initializer and we are in + IPA mode generate constraints for it. In non-IPA mode + the initializer from nonlocal is all we need. */ + if (in_ipa_mode + && DECL_INITIAL (decl)) + { + VEC (ce_s, heap) *rhsc = NULL; + struct constraint_expr lhs, *rhsp; + unsigned i; + get_constraint_for (DECL_INITIAL (decl), &rhsc); + lhs.var = vi->id; + lhs.offset = 0; + lhs.type = SCALAR; + for (i = 0; VEC_iterate (ce_s, rhsc, i, rhsp); ++i) + process_constraint (new_constraint (lhs, *rhsp)); + /* If this is a variable that escapes from the unit + the initializer escapes as well. */ + if (DECL_EXTERNAL (decl) || TREE_PUBLIC (decl)) { - asprintf (&tempname, "%s." HOST_WIDE_INT_PRINT_DEC - "+" HOST_WIDE_INT_PRINT_DEC, - vi->name, fo->offset, fo->size); - newname = ggc_strdup (tempname); - free (tempname); + lhs.var = escaped_id; + lhs.offset = 0; + lhs.type = SCALAR; + for (i = 0; VEC_iterate (ce_s, rhsc, i, rhsp); ++i) + process_constraint (new_constraint (lhs, *rhsp)); } - newvi = new_var_info (decl, newindex, newname); - newvi->offset = fo->offset; - newvi->size = fo->size; - newvi->fullsize = vi->fullsize; - newvi->may_have_pointers = fo->may_have_pointers; - insert_into_field_list (vi, newvi); - VEC_safe_push (varinfo_t, heap, varmap, newvi); - if (is_global && (!flag_whole_program || !in_ipa_mode) - && newvi->may_have_pointers) - make_copy_constraint (newvi, nonlocal_id); - - stats.total_vars++; + VEC_free (ce_s, heap, rhsc); } } - else - vi->is_full_var = true; - VEC_free (fieldoff_s, heap, fieldstack); - - return index; + return id; } /* Print out the points-to solution for VAR to FILE. */ @@ -4432,28 +5434,24 @@ dump_solution_for_var (FILE *file, unsigned int var) unsigned int i; bitmap_iterator bi; - if (find (var) != var) - { - varinfo_t vipt = get_varinfo (find (var)); - fprintf (file, "%s = same as %s\n", vi->name, vipt->name); - } - else - { - fprintf (file, "%s = { ", vi->name); - EXECUTE_IF_SET_IN_BITMAP (vi->solution, 0, i, bi) - { - fprintf (file, "%s ", get_varinfo (i)->name); - } - fprintf (file, "}"); - if (vi->no_tbaa_pruning) - fprintf (file, " no-tbaa-pruning"); - fprintf (file, "\n"); - } + /* Dump the solution for unified vars anyway, this avoids difficulties + in scanning dumps in the testsuite. */ + fprintf (file, "%s = { ", vi->name); + vi = get_varinfo (find (var)); + EXECUTE_IF_SET_IN_BITMAP (vi->solution, 0, i, bi) + fprintf (file, "%s ", get_varinfo (i)->name); + fprintf (file, "}"); + + /* But note when the variable was unified. */ + if (vi->id != var) + fprintf (file, " same as %s", vi->name); + + fprintf (file, "\n"); } /* Print the points-to solution for VAR to stdout. */ -void +DEBUG_FUNCTION void debug_solution_for_var (unsigned int var) { dump_solution_for_var (stdout, var); @@ -4466,10 +5464,10 @@ static void intra_create_variable_infos (void) { tree t; - struct constraint_expr lhs, rhs; /* For each incoming pointer argument arg, create the constraint ARG - = NONLOCAL or a dummy variable if flag_argument_noalias is set. */ + = NONLOCAL or a dummy variable if it is a restrict qualified + passed-by-reference argument. */ for (t = DECL_ARGUMENTS (current_function_decl); t; t = TREE_CHAIN (t)) { varinfo_t p; @@ -4477,65 +5475,48 @@ intra_create_variable_infos (void) if (!could_have_pointers (t)) continue; - /* If flag_argument_noalias is set, then function pointer - arguments are guaranteed not to point to each other. In that - case, create an artificial variable PARM_NOALIAS and the - constraint ARG = &PARM_NOALIAS. */ - if (POINTER_TYPE_P (TREE_TYPE (t)) && flag_argument_noalias > 0) + /* For restrict qualified pointers to objects passed by + reference build a real representative for the pointed-to object. */ + if (DECL_BY_REFERENCE (t) + && POINTER_TYPE_P (TREE_TYPE (t)) + && TYPE_RESTRICT (TREE_TYPE (t))) { + struct constraint_expr lhsc, rhsc; varinfo_t vi; - tree heapvar = heapvar_lookup (t); - - lhs.offset = 0; - lhs.type = SCALAR; - lhs.var = get_vi_for_tree (t)->id; - + tree heapvar = heapvar_lookup (t, 0); if (heapvar == NULL_TREE) { var_ann_t ann; - heapvar = create_tmp_var_raw (ptr_type_node, + heapvar = create_tmp_var_raw (TREE_TYPE (TREE_TYPE (t)), "PARM_NOALIAS"); DECL_EXTERNAL (heapvar) = 1; - if (gimple_referenced_vars (cfun)) - add_referenced_var (heapvar); - - heapvar_insert (t, heapvar); - + heapvar_insert (t, 0, heapvar); ann = get_var_ann (heapvar); ann->is_heapvar = 1; - if (flag_argument_noalias == 1) - ann->noalias_state = NO_ALIAS; - else if (flag_argument_noalias == 2) - ann->noalias_state = NO_ALIAS_GLOBAL; - else if (flag_argument_noalias == 3) - ann->noalias_state = NO_ALIAS_ANYTHING; - else - gcc_unreachable (); - } - - vi = get_vi_for_tree (heapvar); - vi->is_artificial_var = 1; - vi->is_heap_var = 1; - vi->is_unknown_size_var = true; - vi->fullsize = ~0; - vi->size = ~0; - rhs.var = vi->id; - rhs.type = ADDRESSOF; - rhs.offset = 0; - for (p = get_varinfo (lhs.var); p; p = p->next) - { - struct constraint_expr temp = lhs; - temp.var = p->id; - process_constraint (new_constraint (temp, rhs)); } + if (gimple_referenced_vars (cfun)) + add_referenced_var (heapvar); + lhsc.var = get_vi_for_tree (t)->id; + lhsc.type = SCALAR; + lhsc.offset = 0; + rhsc.var = (vi = get_vi_for_tree (heapvar))->id; + rhsc.type = ADDRESSOF; + rhsc.offset = 0; + process_constraint (new_constraint (lhsc, rhsc)); + vi->is_restrict_var = 1; + continue; } - else - { - varinfo_t arg_vi = get_vi_for_tree (t); - for (p = arg_vi; p; p = p->next) + for (p = get_vi_for_tree (t); p; p = p->next) + { + if (p->may_have_pointers) make_constraint_from (p, nonlocal_id); + if (p->only_restrict_pointers) + make_constraint_from_restrict (p, "PARM_RESTRICT"); } + if (POINTER_TYPE_P (TREE_TYPE (t)) + && TYPE_RESTRICT (TREE_TYPE (t))) + make_constraint_from_restrict (get_vi_for_tree (t), "PARM_RESTRICT"); } /* Add a constraint for a result decl that is passed by reference. */ @@ -4628,19 +5609,13 @@ shared_bitmap_add (bitmap pt_vars) } -/* Set bits in INTO corresponding to the variable uids in solution set FROM. - If MEM_ALIAS_SET is not zero, we also use type based alias analysis to - prune the points-to sets with this alias-set. - Returns the number of pruned variables and updates the vars_contains_global - member of *PT . */ +/* Set bits in INTO corresponding to the variable uids in solution set FROM. */ -static unsigned -set_uids_in_ptset (bitmap into, bitmap from, - alias_set_type mem_alias_set, struct pt_solution *pt) +static void +set_uids_in_ptset (bitmap into, bitmap from, struct pt_solution *pt) { unsigned int i; bitmap_iterator bi; - unsigned pruned = 0; EXECUTE_IF_SET_IN_BITMAP (from, 0, i, bi) { @@ -4655,142 +5630,38 @@ set_uids_in_ptset (bitmap into, bitmap from, || TREE_CODE (vi->decl) == PARM_DECL || TREE_CODE (vi->decl) == RESULT_DECL) { - /* Don't type prune artificial vars or points-to sets - for pointers that have not been dereferenced or with - type-based pruning disabled. */ - if (!vi->is_artificial_var - && !vi->no_tbaa_pruning - && mem_alias_set != 0) - { - alias_set_type var_alias_set = get_alias_set (vi->decl); - if (mem_alias_set != var_alias_set - && !alias_set_subset_of (mem_alias_set, var_alias_set)) - { - ++pruned; - continue; - } - } + /* If we are in IPA mode we will not recompute points-to + sets after inlining so make sure they stay valid. */ + if (in_ipa_mode + && !DECL_PT_UID_SET_P (vi->decl)) + SET_DECL_PT_UID (vi->decl, DECL_UID (vi->decl)); /* Add the decl to the points-to set. Note that the points-to set contains global variables. */ - bitmap_set_bit (into, DECL_UID (vi->decl)); - if (is_global_var (vi->decl)) + bitmap_set_bit (into, DECL_PT_UID (vi->decl)); + if (vi->is_global_var) pt->vars_contains_global = true; } } - - return pruned; } -static bool have_alias_info = false; - -/* Emit a note for the pointer initialization point DEF. */ - -static void -emit_pointer_definition (tree ptr, bitmap visited) -{ - gimple def = SSA_NAME_DEF_STMT (ptr); - if (gimple_code (def) == GIMPLE_PHI) - { - use_operand_p argp; - ssa_op_iter oi; - - FOR_EACH_PHI_ARG (argp, def, oi, SSA_OP_USE) - { - tree arg = USE_FROM_PTR (argp); - if (TREE_CODE (arg) == SSA_NAME) - { - if (bitmap_set_bit (visited, SSA_NAME_VERSION (arg))) - emit_pointer_definition (arg, visited); - } - else - inform (0, "initialized from %qE", arg); - } - } - else if (!gimple_nop_p (def)) - inform (gimple_location (def), "initialized from here"); -} - -/* Emit a strict aliasing warning for dereferencing the pointer PTR. */ +/* Compute the points-to solution *PT for the variable VI. */ static void -emit_alias_warning (tree ptr) -{ - gimple use; - imm_use_iterator ui; - bool warned = false; - - FOR_EACH_IMM_USE_STMT (use, ui, ptr) - { - tree deref = NULL_TREE; - - if (gimple_has_lhs (use)) - { - tree lhs = get_base_address (gimple_get_lhs (use)); - if (lhs - && INDIRECT_REF_P (lhs) - && TREE_OPERAND (lhs, 0) == ptr) - deref = lhs; - } - if (gimple_assign_single_p (use)) - { - tree rhs = get_base_address (gimple_assign_rhs1 (use)); - if (rhs - && INDIRECT_REF_P (rhs) - && TREE_OPERAND (rhs, 0) == ptr) - deref = rhs; - } - else if (is_gimple_call (use)) - { - unsigned i; - for (i = 0; i < gimple_call_num_args (use); ++i) - { - tree op = get_base_address (gimple_call_arg (use, i)); - if (op - && INDIRECT_REF_P (op) - && TREE_OPERAND (op, 0) == ptr) - deref = op; - } - } - if (deref - && !TREE_NO_WARNING (deref)) - { - TREE_NO_WARNING (deref) = 1; - warned |= warning_at (gimple_location (use), OPT_Wstrict_aliasing, - "dereferencing pointer %qD does break " - "strict-aliasing rules", SSA_NAME_VAR (ptr)); - } - } - if (warned) - { - bitmap visited = BITMAP_ALLOC (NULL); - emit_pointer_definition (ptr, visited); - BITMAP_FREE (visited); - } -} - -/* Compute the points-to solution *PT for the variable VI. - Prunes the points-to set based on TBAA rules if DO_TBAA_PRUNING - is true. Returns the number of TBAA pruned variables from the - points-to set. */ - -static unsigned int -find_what_var_points_to (varinfo_t vi, struct pt_solution *pt, - bool do_tbaa_pruning) +find_what_var_points_to (varinfo_t orig_vi, struct pt_solution *pt) { - unsigned int i, pruned; + unsigned int i; bitmap_iterator bi; bitmap finished_solution; bitmap result; - tree ptr = vi->decl; - alias_set_type mem_alias_set; + varinfo_t vi; memset (pt, 0, sizeof (struct pt_solution)); /* This variable may have been collapsed, let's get the real variable. */ - vi = get_varinfo (find (vi->id)); + vi = get_varinfo (find (orig_vi->id)); /* Translate artificial variables into SSA_NAME_PTR_INFO attributes. */ @@ -4803,42 +5674,40 @@ find_what_var_points_to (varinfo_t vi, struct pt_solution *pt, if (vi->id == nothing_id) pt->null = 1; else if (vi->id == escaped_id) - pt->escaped = 1; - else if (vi->id == callused_id) - gcc_unreachable (); + { + if (in_ipa_mode) + pt->ipa_escaped = 1; + else + pt->escaped = 1; + } else if (vi->id == nonlocal_id) pt->nonlocal = 1; else if (vi->is_heap_var) /* We represent heapvars in the points-to set properly. */ ; + else if (vi->id == readonly_id) + /* Nobody cares. */ + ; else if (vi->id == anything_id - || vi->id == readonly_id || vi->id == integer_id) pt->anything = 1; } + if (vi->is_restrict_var) + pt->vars_contains_restrict = true; } /* Instead of doing extra work, simply do not create elaborate points-to information for pt_anything pointers. */ - if (pt->anything) - return 0; + if (pt->anything + && (orig_vi->is_artificial_var + || !pt->vars_contains_restrict)) + return; /* Share the final set of variables when possible. */ finished_solution = BITMAP_GGC_ALLOC (); stats.points_to_sets_created++; - if (TREE_CODE (ptr) == SSA_NAME) - ptr = SSA_NAME_VAR (ptr); - - /* If the pointer decl is marked that no TBAA is to be applied, - do not do tbaa pruning. */ - if (!do_tbaa_pruning - || DECL_NO_TBAA_P (ptr)) - mem_alias_set = 0; - else - mem_alias_set = get_deref_alias_set (ptr); - pruned = set_uids_in_ptset (finished_solution, vi->solution, - mem_alias_set, pt); + set_uids_in_ptset (finished_solution, vi->solution, pt); result = shared_bitmap_lookup (finished_solution); if (!result) { @@ -4850,18 +5719,14 @@ find_what_var_points_to (varinfo_t vi, struct pt_solution *pt, pt->vars = result; bitmap_clear (finished_solution); } - - return pruned; } -/* Given a pointer variable P, fill in its points-to set. Apply - type-based pruning if IS_DEREFERENCED is true. */ +/* Given a pointer variable P, fill in its points-to set. */ static void -find_what_p_points_to (tree p, bool is_dereferenced) +find_what_p_points_to (tree p) { struct ptr_info_def *pi; - unsigned int pruned; tree lookup_p = p; varinfo_t vi; @@ -4877,23 +5742,7 @@ find_what_p_points_to (tree p, bool is_dereferenced) return; pi = get_ptr_info (p); - pruned = find_what_var_points_to (vi, &pi->pt, is_dereferenced); - - if (!(pi->pt.anything || pi->pt.nonlocal || pi->pt.escaped) - && bitmap_empty_p (pi->pt.vars) - && pruned > 0 - && is_dereferenced - && warn_strict_aliasing > 0 - && !SSA_NAME_IS_DEFAULT_DEF (p)) - { - if (dump_file && dump_flags & TDF_DETAILS) - { - fprintf (dump_file, "alias warning for "); - print_generic_expr (dump_file, p, 0); - fprintf (dump_file, "\n"); - } - emit_alias_warning (p); - } + find_what_var_points_to (vi, &pi->pt); } @@ -4935,9 +5784,54 @@ pt_solution_reset (struct pt_solution *pt) pt->anything = true; } +/* Set the points-to solution *PT to point only to the variables + in VARS. VARS_CONTAINS_GLOBAL specifies whether that contains + global variables and VARS_CONTAINS_RESTRICT specifies whether + it contains restrict tag variables. */ + +void +pt_solution_set (struct pt_solution *pt, bitmap vars, + bool vars_contains_global, bool vars_contains_restrict) +{ + memset (pt, 0, sizeof (struct pt_solution)); + pt->vars = vars; + pt->vars_contains_global = vars_contains_global; + pt->vars_contains_restrict = vars_contains_restrict; +} + +/* Computes the union of the points-to solutions *DEST and *SRC and + stores the result in *DEST. This changes the points-to bitmap + of *DEST and thus may not be used if that might be shared. + The points-to bitmap of *SRC and *DEST will not be shared after + this function if they were not before. */ + +static void +pt_solution_ior_into (struct pt_solution *dest, struct pt_solution *src) +{ + dest->anything |= src->anything; + if (dest->anything) + { + pt_solution_reset (dest); + return; + } + + dest->nonlocal |= src->nonlocal; + dest->escaped |= src->escaped; + dest->ipa_escaped |= src->ipa_escaped; + dest->null |= src->null; + dest->vars_contains_global |= src->vars_contains_global; + dest->vars_contains_restrict |= src->vars_contains_restrict; + if (!src->vars) + return; + + if (!dest->vars) + dest->vars = BITMAP_GGC_ALLOC (); + bitmap_ior_into (dest->vars, src->vars); +} + /* Return true if the points-to solution *PT is empty. */ -static bool +bool pt_solution_empty_p (struct pt_solution *pt) { if (pt->anything @@ -4953,6 +5847,11 @@ pt_solution_empty_p (struct pt_solution *pt) && !pt_solution_empty_p (&cfun->gimple_df->escaped)) return false; + /* If the solution includes ESCAPED, check if that is empty. */ + if (pt->ipa_escaped + && !pt_solution_empty_p (&ipa_escaped_pt)) + return false; + return true; } @@ -4969,6 +5868,15 @@ pt_solution_includes_global (struct pt_solution *pt) if (pt->escaped) return pt_solution_includes_global (&cfun->gimple_df->escaped); + if (pt->ipa_escaped) + return pt_solution_includes_global (&ipa_escaped_pt); + + /* ??? This predicate is not correct for the IPA-PTA solution + as we do not properly distinguish between unit escape points + and global variables. */ + if (cfun->gimple_df->ipa_pta) + return true; + return false; } @@ -4986,7 +5894,7 @@ pt_solution_includes_1 (struct pt_solution *pt, const_tree decl) return true; if (pt->vars - && bitmap_bit_p (pt->vars, DECL_UID (decl))) + && bitmap_bit_p (pt->vars, DECL_PT_UID (decl))) return true; /* If the solution includes ESCAPED, check it. */ @@ -4994,6 +5902,11 @@ pt_solution_includes_1 (struct pt_solution *pt, const_tree decl) && pt_solution_includes_1 (&cfun->gimple_df->escaped, decl)) return true; + /* If the solution includes ESCAPED, check it. */ + if (pt->ipa_escaped + && pt_solution_includes_1 (&ipa_escaped_pt, decl)) + return true; + return false; } @@ -5044,6 +5957,25 @@ pt_solutions_intersect_1 (struct pt_solution *pt1, struct pt_solution *pt2) return true; } + /* Check the escaped solution if required. + ??? Do we need to check the local against the IPA escaped sets? */ + if ((pt1->ipa_escaped || pt2->ipa_escaped) + && !pt_solution_empty_p (&ipa_escaped_pt)) + { + /* If both point to escaped memory and that solution + is not empty they alias. */ + if (pt1->ipa_escaped && pt2->ipa_escaped) + return true; + + /* If either points to escaped memory see if the escaped solution + intersects with the other. */ + if ((pt1->ipa_escaped + && pt_solutions_intersect_1 (&ipa_escaped_pt, pt2)) + || (pt2->ipa_escaped + && pt_solutions_intersect_1 (&ipa_escaped_pt, pt1))) + return true; + } + /* Now both pointers alias if their points-to solution intersects. */ return (pt1->vars && pt2->vars @@ -5061,6 +5993,27 @@ pt_solutions_intersect (struct pt_solution *pt1, struct pt_solution *pt2) return res; } +/* Return true if both points-to solutions PT1 and PT2 for two restrict + qualified pointers are possibly based on the same pointer. */ + +bool +pt_solutions_same_restrict_base (struct pt_solution *pt1, + struct pt_solution *pt2) +{ + /* If we deal with points-to solutions of two restrict qualified + pointers solely rely on the pointed-to variable bitmap intersection. + For two pointers that are based on each other the bitmaps will + intersect. */ + if (pt1->vars_contains_restrict + && pt2->vars_contains_restrict) + { + gcc_assert (pt1->vars && pt2->vars); + return bitmap_intersect_p (pt1->vars, pt2->vars); + } + + return true; +} + /* Dump points-to information to OUTFILE. */ @@ -5088,13 +6041,18 @@ dump_sa_points_to_info (FILE *outfile) } for (i = 0; i < VEC_length (varinfo_t, varmap); i++) - dump_solution_for_var (outfile, i); + { + varinfo_t vi = get_varinfo (i); + if (!vi->may_have_pointers) + continue; + dump_solution_for_var (outfile, i); + } } /* Debug points-to information to stderr. */ -void +DEBUG_FUNCTION void debug_sa_points_to_info (void) { dump_sa_points_to_info (stderr); @@ -5108,24 +6066,30 @@ static void init_base_vars (void) { struct constraint_expr lhs, rhs; + varinfo_t var_anything; + varinfo_t var_nothing; + varinfo_t var_readonly; + varinfo_t var_escaped; + varinfo_t var_nonlocal; + varinfo_t var_storedanything; + varinfo_t var_integer; /* Create the NULL variable, used to represent that a variable points to NULL. */ - nothing_tree = create_tmp_var_raw (void_type_node, "NULL"); - var_nothing = new_var_info (nothing_tree, nothing_id, "NULL"); - insert_vi_for_tree (nothing_tree, var_nothing); + var_nothing = new_var_info (NULL_TREE, "NULL"); + gcc_assert (var_nothing->id == nothing_id); var_nothing->is_artificial_var = 1; var_nothing->offset = 0; var_nothing->size = ~0; var_nothing->fullsize = ~0; var_nothing->is_special_var = 1; - VEC_safe_push (varinfo_t, heap, varmap, var_nothing); + var_nothing->may_have_pointers = 0; + var_nothing->is_global_var = 0; /* Create the ANYTHING variable, used to represent that a variable points to some unknown piece of memory. */ - anything_tree = create_tmp_var_raw (ptr_type_node, "ANYTHING"); - var_anything = new_var_info (anything_tree, anything_id, "ANYTHING"); - insert_vi_for_tree (anything_tree, var_anything); + var_anything = new_var_info (NULL_TREE, "ANYTHING"); + gcc_assert (var_anything->id == anything_id); var_anything->is_artificial_var = 1; var_anything->size = ~0; var_anything->offset = 0; @@ -5136,7 +6100,6 @@ init_base_vars (void) /* Anything points to anything. This makes deref constraints just work in the presence of linked list and other p = *p type loops, by saying that *ANYTHING = ANYTHING. */ - VEC_safe_push (varinfo_t, heap, varmap, var_anything); lhs.type = SCALAR; lhs.var = anything_id; lhs.offset = 0; @@ -5151,16 +6114,14 @@ init_base_vars (void) /* Create the READONLY variable, used to represent that a variable points to readonly memory. */ - readonly_tree = create_tmp_var_raw (ptr_type_node, "READONLY"); - var_readonly = new_var_info (readonly_tree, readonly_id, "READONLY"); + var_readonly = new_var_info (NULL_TREE, "READONLY"); + gcc_assert (var_readonly->id == readonly_id); var_readonly->is_artificial_var = 1; var_readonly->offset = 0; var_readonly->size = ~0; var_readonly->fullsize = ~0; var_readonly->next = NULL; var_readonly->is_special_var = 1; - insert_vi_for_tree (readonly_tree, var_readonly); - VEC_safe_push (varinfo_t, heap, varmap, var_readonly); /* readonly memory points to anything, in order to make deref easier. In reality, it points to anything the particular @@ -5176,28 +6137,23 @@ init_base_vars (void) /* Create the ESCAPED variable, used to represent the set of escaped memory. */ - escaped_tree = create_tmp_var_raw (ptr_type_node, "ESCAPED"); - var_escaped = new_var_info (escaped_tree, escaped_id, "ESCAPED"); - insert_vi_for_tree (escaped_tree, var_escaped); + var_escaped = new_var_info (NULL_TREE, "ESCAPED"); + gcc_assert (var_escaped->id == escaped_id); var_escaped->is_artificial_var = 1; var_escaped->offset = 0; var_escaped->size = ~0; var_escaped->fullsize = ~0; var_escaped->is_special_var = 0; - VEC_safe_push (varinfo_t, heap, varmap, var_escaped); - gcc_assert (VEC_index (varinfo_t, varmap, 3) == var_escaped); /* Create the NONLOCAL variable, used to represent the set of nonlocal memory. */ - nonlocal_tree = create_tmp_var_raw (ptr_type_node, "NONLOCAL"); - var_nonlocal = new_var_info (nonlocal_tree, nonlocal_id, "NONLOCAL"); - insert_vi_for_tree (nonlocal_tree, var_nonlocal); + var_nonlocal = new_var_info (NULL_TREE, "NONLOCAL"); + gcc_assert (var_nonlocal->id == nonlocal_id); var_nonlocal->is_artificial_var = 1; var_nonlocal->offset = 0; var_nonlocal->size = ~0; var_nonlocal->fullsize = ~0; var_nonlocal->is_special_var = 1; - VEC_safe_push (varinfo_t, heap, varmap, var_nonlocal); /* ESCAPED = *ESCAPED, because escaped is may-deref'd at calls, etc. */ lhs.type = SCALAR; @@ -5243,62 +6199,26 @@ init_base_vars (void) rhs.offset = 0; process_constraint (new_constraint (lhs, rhs)); - /* Create the CALLUSED variable, used to represent the set of call-used - memory. */ - callused_tree = create_tmp_var_raw (ptr_type_node, "CALLUSED"); - var_callused = new_var_info (callused_tree, callused_id, "CALLUSED"); - insert_vi_for_tree (callused_tree, var_callused); - var_callused->is_artificial_var = 1; - var_callused->offset = 0; - var_callused->size = ~0; - var_callused->fullsize = ~0; - var_callused->is_special_var = 0; - VEC_safe_push (varinfo_t, heap, varmap, var_callused); - - /* CALLUSED = *CALLUSED, because call-used is may-deref'd at calls, etc. */ - lhs.type = SCALAR; - lhs.var = callused_id; - lhs.offset = 0; - rhs.type = DEREF; - rhs.var = callused_id; - rhs.offset = 0; - process_constraint (new_constraint (lhs, rhs)); - - /* CALLUSED = CALLUSED + UNKNOWN, because if a sub-field is call-used the - whole variable is call-used. */ - lhs.type = SCALAR; - lhs.var = callused_id; - lhs.offset = 0; - rhs.type = SCALAR; - rhs.var = callused_id; - rhs.offset = UNKNOWN_OFFSET; - process_constraint (new_constraint (lhs, rhs)); - /* Create the STOREDANYTHING variable, used to represent the set of variables stored to *ANYTHING. */ - storedanything_tree = create_tmp_var_raw (ptr_type_node, "STOREDANYTHING"); - var_storedanything = new_var_info (storedanything_tree, storedanything_id, - "STOREDANYTHING"); - insert_vi_for_tree (storedanything_tree, var_storedanything); + var_storedanything = new_var_info (NULL_TREE, "STOREDANYTHING"); + gcc_assert (var_storedanything->id == storedanything_id); var_storedanything->is_artificial_var = 1; var_storedanything->offset = 0; var_storedanything->size = ~0; var_storedanything->fullsize = ~0; var_storedanything->is_special_var = 0; - VEC_safe_push (varinfo_t, heap, varmap, var_storedanything); /* Create the INTEGER variable, used to represent that a variable points to what an INTEGER "points to". */ - integer_tree = create_tmp_var_raw (ptr_type_node, "INTEGER"); - var_integer = new_var_info (integer_tree, integer_id, "INTEGER"); - insert_vi_for_tree (integer_tree, var_integer); + var_integer = new_var_info (NULL_TREE, "INTEGER"); + gcc_assert (var_integer->id == integer_id); var_integer->is_artificial_var = 1; var_integer->size = ~0; var_integer->fullsize = ~0; var_integer->offset = 0; var_integer->next = NULL; var_integer->is_special_var = 1; - VEC_safe_push (varinfo_t, heap, varmap, var_integer); /* INTEGER = ANYTHING, because we don't know where a dereference of a random integer will point to. */ @@ -5329,6 +6249,7 @@ init_alias_vars (void) constraints = VEC_alloc (constraint_t, heap, 8); varmap = VEC_alloc (varinfo_t, heap, 8); vi_for_tree = pointer_map_create (); + call_stmt_vars = pointer_map_create (); memset (&stats, 0, sizeof (stats)); shared_bitmap_table = htab_create (511, shared_bitmap_hash, @@ -5372,157 +6293,82 @@ remove_preds_and_fake_succs (constraint_graph_t graph) bitmap_obstack_release (&predbitmap_obstack); } -/* Compute the set of variables we can't TBAA prune. */ +/* Initialize the heapvar for statement mapping. */ static void -compute_tbaa_pruning (void) +init_alias_heapvars (void) { - unsigned int size = VEC_length (varinfo_t, varmap); - unsigned int i; - bool any; - - changed_count = 0; - changed = sbitmap_alloc (size); - sbitmap_zero (changed); + if (!heapvar_for_stmt) + heapvar_for_stmt = htab_create_ggc (11, tree_map_hash, heapvar_map_eq, + NULL); +} - /* Mark all initial no_tbaa_pruning nodes as changed. */ - any = false; - for (i = 0; i < size; ++i) - { - varinfo_t ivi = get_varinfo (i); +/* Delete the heapvar for statement mapping. */ - if (find (i) == i && ivi->no_tbaa_pruning) - { - any = true; - if ((graph->succs[i] && !bitmap_empty_p (graph->succs[i])) - || VEC_length (constraint_t, graph->complex[i]) > 0) - { - SET_BIT (changed, i); - ++changed_count; - } - } - } +void +delete_alias_heapvars (void) +{ + if (heapvar_for_stmt) + htab_delete (heapvar_for_stmt); + heapvar_for_stmt = NULL; +} - while (changed_count > 0) - { - struct topo_info *ti = init_topo_info (); - ++stats.iterations; +/* Solve the constraint set. */ - compute_topo_order (graph, ti); +static void +solve_constraints (void) +{ + struct scc_info *si; - while (VEC_length (unsigned, ti->topo_order) != 0) - { - bitmap_iterator bi; + if (dump_file) + fprintf (dump_file, + "\nCollapsing static cycles and doing variable " + "substitution\n"); - i = VEC_pop (unsigned, ti->topo_order); + init_graph (VEC_length (varinfo_t, varmap) * 2); - /* If this variable is not a representative, skip it. */ - if (find (i) != i) - continue; + if (dump_file) + fprintf (dump_file, "Building predecessor graph\n"); + build_pred_graph (); - /* If the node has changed, we need to process the complex - constraints and outgoing edges again. */ - if (TEST_BIT (changed, i)) - { - unsigned int j; - constraint_t c; - VEC(constraint_t,heap) *complex = graph->complex[i]; + if (dump_file) + fprintf (dump_file, "Detecting pointer and location " + "equivalences\n"); + si = perform_var_substitution (graph); - RESET_BIT (changed, i); - --changed_count; - - /* Process the complex copy constraints. */ - for (j = 0; VEC_iterate (constraint_t, complex, j, c); ++j) - { - if (c->lhs.type == SCALAR && c->rhs.type == SCALAR) - { - varinfo_t lhsvi = get_varinfo (find (c->lhs.var)); - - if (!lhsvi->no_tbaa_pruning) - { - lhsvi->no_tbaa_pruning = true; - if (!TEST_BIT (changed, lhsvi->id)) - { - SET_BIT (changed, lhsvi->id); - ++changed_count; - } - } - } - } - - /* Propagate to all successors. */ - EXECUTE_IF_IN_NONNULL_BITMAP (graph->succs[i], 0, j, bi) - { - unsigned int to = find (j); - varinfo_t tovi = get_varinfo (to); - - /* Don't propagate to ourselves. */ - if (to == i) - continue; - - if (!tovi->no_tbaa_pruning) - { - tovi->no_tbaa_pruning = true; - if (!TEST_BIT (changed, to)) - { - SET_BIT (changed, to); - ++changed_count; - } - } - } - } - } - - free_topo_info (ti); - } - - sbitmap_free (changed); + if (dump_file) + fprintf (dump_file, "Rewriting constraints and unifying " + "variables\n"); + rewrite_constraints (graph, si); - if (any) - { - for (i = 0; i < size; ++i) - { - varinfo_t ivi = get_varinfo (i); - varinfo_t ivip = get_varinfo (find (i)); + build_succ_graph (); + free_var_substitution_info (si); - if (ivip->no_tbaa_pruning) - { - tree var = ivi->decl; + if (dump_file && (dump_flags & TDF_GRAPH)) + dump_constraint_graph (dump_file); - if (TREE_CODE (var) == SSA_NAME) - var = SSA_NAME_VAR (var); + move_complex_constraints (graph); - if (POINTER_TYPE_P (TREE_TYPE (var))) - { - DECL_NO_TBAA_P (var) = 1; + if (dump_file) + fprintf (dump_file, "Uniting pointer but not location equivalent " + "variables\n"); + unite_pointer_equivalences (graph); - /* Tell the RTL layer that this pointer can alias - anything. */ - DECL_POINTER_ALIAS_SET (var) = 0; - } - } - } - } -} + if (dump_file) + fprintf (dump_file, "Finding indirect cycles\n"); + find_indirect_cycles (graph); -/* Initialize the heapvar for statement mapping. */ + /* Implicit nodes and predecessors are no longer necessary at this + point. */ + remove_preds_and_fake_succs (graph); -static void -init_alias_heapvars (void) -{ - if (!heapvar_for_stmt) - heapvar_for_stmt = htab_create_ggc (11, tree_map_hash, tree_map_eq, - NULL); -} + if (dump_file) + fprintf (dump_file, "Solving graph\n"); -/* Delete the heapvar for statement mapping. */ + solve_graph (graph); -void -delete_alias_heapvars (void) -{ - if (heapvar_for_stmt) - htab_delete (heapvar_for_stmt); - heapvar_for_stmt = NULL; + if (dump_file) + dump_sa_points_to_info (dump_file); } /* Create points-to sets for the current function. See the comments @@ -5531,10 +6377,9 @@ delete_alias_heapvars (void) static void compute_points_to_sets (void) { - struct scc_info *si; basic_block bb; unsigned i; - sbitmap dereferenced_ptrs; + varinfo_t vi; timevar_push (TV_TREE_PTA); @@ -5543,12 +6388,7 @@ compute_points_to_sets (void) intra_create_variable_infos (); - /* A bitmap of SSA_NAME pointers that are dereferenced. This is - used to track which points-to sets may be TBAA pruned. */ - dereferenced_ptrs = sbitmap_alloc (num_ssa_names); - sbitmap_zero (dereferenced_ptrs); - - /* Now walk all statements and derive aliases. */ + /* Now walk all statements and build the constraint set. */ FOR_EACH_BB (bb) { gimple_stmt_iterator gsi; @@ -5564,112 +6404,103 @@ compute_points_to_sets (void) for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi)) { gimple stmt = gsi_stmt (gsi); - use_operand_p use_p; - ssa_op_iter iter; - - /* Mark dereferenced pointers. This is used by TBAA pruning - of the points-to sets and the alias warning machinery. */ - FOR_EACH_SSA_USE_OPERAND (use_p, stmt, iter, SSA_OP_USE) - { - unsigned num_uses, num_loads, num_stores; - tree op = USE_FROM_PTR (use_p); - - if (!POINTER_TYPE_P (TREE_TYPE (op))) - continue; - - /* Determine whether OP is a dereferenced pointer. */ - count_uses_and_derefs (op, stmt, - &num_uses, &num_loads, &num_stores); - if (num_loads + num_stores > 0) - SET_BIT (dereferenced_ptrs, SSA_NAME_VERSION (op)); - } find_func_aliases (stmt); } } - if (dump_file) { fprintf (dump_file, "Points-to analysis\n\nConstraints:\n\n"); - dump_constraints (dump_file); + dump_constraints (dump_file, 0); } - if (dump_file) - fprintf (dump_file, - "\nCollapsing static cycles and doing variable " - "substitution\n"); - - init_graph (VEC_length (varinfo_t, varmap) * 2); - - if (dump_file) - fprintf (dump_file, "Building predecessor graph\n"); - build_pred_graph (); - - if (dump_file) - fprintf (dump_file, "Detecting pointer and location " - "equivalences\n"); - si = perform_var_substitution (graph); - - if (dump_file) - fprintf (dump_file, "Rewriting constraints and unifying " - "variables\n"); - rewrite_constraints (graph, si); - - build_succ_graph (); - free_var_substitution_info (si); - - if (dump_file && (dump_flags & TDF_GRAPH)) - dump_constraint_graph (dump_file); - - move_complex_constraints (graph); - - if (dump_file) - fprintf (dump_file, "Uniting pointer but not location equivalent " - "variables\n"); - unite_pointer_equivalences (graph); - - if (dump_file) - fprintf (dump_file, "Finding indirect cycles\n"); - find_indirect_cycles (graph); - - /* Implicit nodes and predecessors are no longer necessary at this - point. */ - remove_preds_and_fake_succs (graph); + /* From the constraints compute the points-to sets. */ + solve_constraints (); - if (dump_file) - fprintf (dump_file, "Solving graph\n"); - - solve_graph (graph); - - compute_tbaa_pruning (); - - if (dump_file) - dump_sa_points_to_info (dump_file); - - /* Compute the points-to sets for ESCAPED and CALLUSED used for - call-clobber analysis. */ - find_what_var_points_to (var_escaped, &cfun->gimple_df->escaped, false); - find_what_var_points_to (var_callused, &cfun->gimple_df->callused, false); + /* Compute the points-to set for ESCAPED used for call-clobber analysis. */ + find_what_var_points_to (get_varinfo (escaped_id), + &cfun->gimple_df->escaped); /* Make sure the ESCAPED solution (which is used as placeholder in other solutions) does not reference itself. This simplifies points-to solution queries. */ cfun->gimple_df->escaped.escaped = 0; + /* Mark escaped HEAP variables as global. */ + for (i = 0; VEC_iterate (varinfo_t, varmap, i, vi); ++i) + if (vi->is_heap_var + && !vi->is_restrict_var + && !vi->is_global_var) + DECL_EXTERNAL (vi->decl) = vi->is_global_var + = pt_solution_includes (&cfun->gimple_df->escaped, vi->decl); + /* Compute the points-to sets for pointer SSA_NAMEs. */ for (i = 0; i < num_ssa_names; ++i) { tree ptr = ssa_name (i); if (ptr && POINTER_TYPE_P (TREE_TYPE (ptr))) - find_what_p_points_to (ptr, TEST_BIT (dereferenced_ptrs, i)); + find_what_p_points_to (ptr); } - sbitmap_free (dereferenced_ptrs); - timevar_pop (TV_TREE_PTA); + /* Compute the call-used/clobbered sets. */ + FOR_EACH_BB (bb) + { + gimple_stmt_iterator gsi; + + for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi)) + { + gimple stmt = gsi_stmt (gsi); + struct pt_solution *pt; + if (!is_gimple_call (stmt)) + continue; - have_alias_info = true; + pt = gimple_call_use_set (stmt); + if (gimple_call_flags (stmt) & ECF_CONST) + memset (pt, 0, sizeof (struct pt_solution)); + else if ((vi = lookup_call_use_vi (stmt)) != NULL) + { + find_what_var_points_to (vi, pt); + /* Escaped (and thus nonlocal) variables are always + implicitly used by calls. */ + /* ??? ESCAPED can be empty even though NONLOCAL + always escaped. */ + pt->nonlocal = 1; + pt->escaped = 1; + } + else + { + /* If there is nothing special about this call then + we have made everything that is used also escape. */ + *pt = cfun->gimple_df->escaped; + pt->nonlocal = 1; + } + + pt = gimple_call_clobber_set (stmt); + if (gimple_call_flags (stmt) & (ECF_CONST|ECF_PURE|ECF_NOVOPS)) + memset (pt, 0, sizeof (struct pt_solution)); + else if ((vi = lookup_call_clobber_vi (stmt)) != NULL) + { + find_what_var_points_to (vi, pt); + /* Escaped (and thus nonlocal) variables are always + implicitly clobbered by calls. */ + /* ??? ESCAPED can be empty even though NONLOCAL + always escaped. */ + pt->nonlocal = 1; + pt->escaped = 1; + } + else + { + /* If there is nothing special about this call then + we have made everything that is used also escape. */ + *pt = cfun->gimple_df->escaped; + pt->nonlocal = 1; + } + } + } + + timevar_pop (TV_TREE_PTA); } @@ -5686,6 +6517,7 @@ delete_points_to_sets (void) stats.points_to_sets_created); pointer_map_destroy (vi_for_tree); + pointer_map_destroy (call_stmt_vars); bitmap_obstack_release (&pta_obstack); VEC_free (constraint_t, heap, constraints); @@ -5703,7 +6535,6 @@ delete_points_to_sets (void) VEC_free (varinfo_t, heap, varmap); free_alloc_pool (variable_info_pool); free_alloc_pool (constraint_pool); - have_alias_info = false; } @@ -5714,6 +6545,23 @@ delete_points_to_sets (void) unsigned int compute_may_aliases (void) { + if (cfun->gimple_df->ipa_pta) + { + if (dump_file) + { + fprintf (dump_file, "\nNot re-computing points-to information " + "because IPA points-to information is available.\n\n"); + + /* But still dump what we have remaining it. */ + dump_alias_info (dump_file); + + if (dump_flags & TDF_DETAILS) + dump_referenced_vars (dump_file); + } + + return 0; + } + /* For each pointer P_i, determine the sets of variables that P_i may point-to. Compute the reachability set of escaped and call-used variables. */ @@ -5737,6 +6585,11 @@ compute_may_aliases (void) return 0; } +static bool +gate_tree_pta (void) +{ + return flag_tree_pta; +} /* A dummy pass to cause points-to information to be computed via TODO_rebuild_alias. */ @@ -5746,7 +6599,29 @@ struct gimple_opt_pass pass_build_alias = { GIMPLE_PASS, "alias", /* name */ - NULL, /* gate */ + gate_tree_pta, /* gate */ + NULL, /* execute */ + NULL, /* sub */ + NULL, /* next */ + 0, /* static_pass_number */ + TV_NONE, /* tv_id */ + PROP_cfg | PROP_ssa, /* properties_required */ + 0, /* properties_provided */ + 0, /* properties_destroyed */ + 0, /* todo_flags_start */ + TODO_rebuild_alias | TODO_dump_func /* todo_flags_finish */ + } +}; + +/* A dummy pass to cause points-to information to be computed via + TODO_rebuild_alias. */ + +struct gimple_opt_pass pass_build_ealias = +{ + { + GIMPLE_PASS, + "ealias", /* name */ + gate_tree_pta, /* gate */ NULL, /* execute */ NULL, /* sub */ NULL, /* next */ @@ -5765,111 +6640,325 @@ struct gimple_opt_pass pass_build_alias = static bool gate_ipa_pta (void) { - return (flag_ipa_pta + return (optimize + && flag_ipa_pta /* Don't bother doing anything if the program has errors. */ - && !(errorcount || sorrycount)); + && !seen_error ()); } +/* IPA PTA solutions for ESCAPED. */ +struct pt_solution ipa_escaped_pt + = { true, false, false, false, false, false, false, NULL }; + /* Execute the driver for IPA PTA. */ static unsigned int ipa_pta_execute (void) { struct cgraph_node *node; - struct scc_info *si; + struct varpool_node *var; + int from; in_ipa_mode = 1; + init_alias_heapvars (); init_alias_vars (); + /* Build the constraints. */ for (node = cgraph_nodes; node; node = node->next) { - unsigned int varid; + struct cgraph_node *alias; + varinfo_t vi; - varid = create_function_info_for (node->decl, - cgraph_node_name (node)); - if (node->local.externally_visible) - { - varinfo_t fi = get_varinfo (varid); - for (; fi; fi = fi->next) - make_constraint_from (fi, anything_id); - } + /* Nodes without a body are not interesting. Especially do not + visit clones at this point for now - we get duplicate decls + there for inline clones at least. */ + if (!gimple_has_body_p (node->decl) + || node->clone_of) + continue; + + vi = create_function_info_for (node->decl, + alias_get_name (node->decl)); + + /* Associate the varinfo node with all aliases. */ + for (alias = node->same_body; alias; alias = alias->next) + insert_vi_for_tree (alias->decl, vi); } + + /* Create constraints for global variables and their initializers. */ + for (var = varpool_nodes; var; var = var->next) + { + struct varpool_node *alias; + varinfo_t vi; + + vi = get_vi_for_tree (var->decl); + + /* Associate the varinfo node with all aliases. */ + for (alias = var->extra_name; alias; alias = alias->next) + insert_vi_for_tree (alias->decl, vi); + } + + if (dump_file) + { + fprintf (dump_file, + "Generating constraints for global initializers\n\n"); + dump_constraints (dump_file, 0); + fprintf (dump_file, "\n"); + } + from = VEC_length (constraint_t, constraints); + for (node = cgraph_nodes; node; node = node->next) { - if (node->analyzed) + struct function *func; + basic_block bb; + tree old_func_decl; + + /* Nodes without a body are not interesting. */ + if (!gimple_has_body_p (node->decl) + || node->clone_of) + continue; + + if (dump_file) { - struct function *func = DECL_STRUCT_FUNCTION (node->decl); - basic_block bb; - tree old_func_decl = current_function_decl; - if (dump_file) - fprintf (dump_file, - "Generating constraints for %s\n", - cgraph_node_name (node)); - push_cfun (func); - current_function_decl = node->decl; + fprintf (dump_file, + "Generating constraints for %s", cgraph_node_name (node)); + if (DECL_ASSEMBLER_NAME_SET_P (node->decl)) + fprintf (dump_file, " (%s)", + IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (node->decl))); + fprintf (dump_file, "\n"); + } + + func = DECL_STRUCT_FUNCTION (node->decl); + old_func_decl = current_function_decl; + push_cfun (func); + current_function_decl = node->decl; + + /* For externally visible functions use local constraints for + their arguments. For local functions we see all callers + and thus do not need initial constraints for parameters. */ + if (node->local.externally_visible) + intra_create_variable_infos (); + + /* Build constriants for the function body. */ + FOR_EACH_BB_FN (bb, func) + { + gimple_stmt_iterator gsi; - FOR_EACH_BB_FN (bb, func) + for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); + gsi_next (&gsi)) { - gimple_stmt_iterator gsi; + gimple phi = gsi_stmt (gsi); - for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); - gsi_next (&gsi)) - { - gimple phi = gsi_stmt (gsi); + if (is_gimple_reg (gimple_phi_result (phi))) + find_func_aliases (phi); + } - if (is_gimple_reg (gimple_phi_result (phi))) - find_func_aliases (phi); - } + for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi)) + { + gimple stmt = gsi_stmt (gsi); - for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi)) - find_func_aliases (gsi_stmt (gsi)); + find_func_aliases (stmt); + find_func_clobbers (stmt); } - current_function_decl = old_func_decl; - pop_cfun (); } - else + + current_function_decl = old_func_decl; + pop_cfun (); + + if (dump_file) { - /* Make point to anything. */ + fprintf (dump_file, "\n"); + dump_constraints (dump_file, from); + fprintf (dump_file, "\n"); } + from = VEC_length (constraint_t, constraints); } - if (dump_file) + /* From the constraints compute the points-to sets. */ + solve_constraints (); + + /* Compute the global points-to sets for ESCAPED. + ??? Note that the computed escape set is not correct + for the whole unit as we fail to consider graph edges to + externally visible functions. */ + find_what_var_points_to (get_varinfo (escaped_id), &ipa_escaped_pt); + + /* Make sure the ESCAPED solution (which is used as placeholder in + other solutions) does not reference itself. This simplifies + points-to solution queries. */ + ipa_escaped_pt.ipa_escaped = 0; + + /* Assign the points-to sets to the SSA names in the unit. */ + for (node = cgraph_nodes; node; node = node->next) { - fprintf (dump_file, "Points-to analysis\n\nConstraints:\n\n"); - dump_constraints (dump_file); - } + tree ptr; + struct function *fn; + unsigned i; + varinfo_t fi; + basic_block bb; + struct pt_solution uses, clobbers; + struct cgraph_edge *e; + + /* Nodes without a body are not interesting. */ + if (!gimple_has_body_p (node->decl) + || node->clone_of) + continue; - if (dump_file) - fprintf (dump_file, - "\nCollapsing static cycles and doing variable " - "substitution:\n"); + fn = DECL_STRUCT_FUNCTION (node->decl); - init_graph (VEC_length (varinfo_t, varmap) * 2); - build_pred_graph (); - si = perform_var_substitution (graph); - rewrite_constraints (graph, si); + /* Compute the points-to sets for pointer SSA_NAMEs. */ + for (i = 0; VEC_iterate (tree, fn->gimple_df->ssa_names, i, ptr); ++i) + { + if (ptr + && POINTER_TYPE_P (TREE_TYPE (ptr))) + find_what_p_points_to (ptr); + } - build_succ_graph (); - free_var_substitution_info (si); - move_complex_constraints (graph); - unite_pointer_equivalences (graph); - find_indirect_cycles (graph); + /* Compute the call-use and call-clobber sets for all direct calls. */ + fi = lookup_vi_for_tree (node->decl); + gcc_assert (fi->is_fn_info); + find_what_var_points_to (first_vi_for_offset (fi, fi_clobbers), + &clobbers); + find_what_var_points_to (first_vi_for_offset (fi, fi_uses), &uses); + for (e = node->callers; e; e = e->next_caller) + { + if (!e->call_stmt) + continue; - /* Implicit nodes and predecessors are no longer necessary at this - point. */ - remove_preds_and_fake_succs (graph); + *gimple_call_clobber_set (e->call_stmt) = clobbers; + *gimple_call_use_set (e->call_stmt) = uses; + } - if (dump_file) - fprintf (dump_file, "\nSolving graph\n"); + /* Compute the call-use and call-clobber sets for indirect calls + and calls to external functions. */ + FOR_EACH_BB_FN (bb, fn) + { + gimple_stmt_iterator gsi; - solve_graph (graph); + for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi)) + { + gimple stmt = gsi_stmt (gsi); + struct pt_solution *pt; + varinfo_t vi; + tree decl; - if (dump_file) - dump_sa_points_to_info (dump_file); + if (!is_gimple_call (stmt)) + continue; + + /* Handle direct calls to external functions. */ + decl = gimple_call_fndecl (stmt); + if (decl + && (!(fi = lookup_vi_for_tree (decl)) + || !fi->is_fn_info)) + { + pt = gimple_call_use_set (stmt); + if (gimple_call_flags (stmt) & ECF_CONST) + memset (pt, 0, sizeof (struct pt_solution)); + else if ((vi = lookup_call_use_vi (stmt)) != NULL) + { + find_what_var_points_to (vi, pt); + /* Escaped (and thus nonlocal) variables are always + implicitly used by calls. */ + /* ??? ESCAPED can be empty even though NONLOCAL + always escaped. */ + pt->nonlocal = 1; + pt->ipa_escaped = 1; + } + else + { + /* If there is nothing special about this call then + we have made everything that is used also escape. */ + *pt = ipa_escaped_pt; + pt->nonlocal = 1; + } + + pt = gimple_call_clobber_set (stmt); + if (gimple_call_flags (stmt) & (ECF_CONST|ECF_PURE|ECF_NOVOPS)) + memset (pt, 0, sizeof (struct pt_solution)); + else if ((vi = lookup_call_clobber_vi (stmt)) != NULL) + { + find_what_var_points_to (vi, pt); + /* Escaped (and thus nonlocal) variables are always + implicitly clobbered by calls. */ + /* ??? ESCAPED can be empty even though NONLOCAL + always escaped. */ + pt->nonlocal = 1; + pt->ipa_escaped = 1; + } + else + { + /* If there is nothing special about this call then + we have made everything that is used also escape. */ + *pt = ipa_escaped_pt; + pt->nonlocal = 1; + } + } + + /* Handle indirect calls. */ + if (!decl + && (fi = get_fi_for_callee (stmt))) + { + /* We need to accumulate all clobbers/uses of all possible + callees. */ + fi = get_varinfo (find (fi->id)); + /* If we cannot constrain the set of functions we'll end up + calling we end up using/clobbering everything. */ + if (bitmap_bit_p (fi->solution, anything_id) + || bitmap_bit_p (fi->solution, nonlocal_id) + || bitmap_bit_p (fi->solution, escaped_id)) + { + pt_solution_reset (gimple_call_clobber_set (stmt)); + pt_solution_reset (gimple_call_use_set (stmt)); + } + else + { + bitmap_iterator bi; + unsigned i; + struct pt_solution *uses, *clobbers; + + uses = gimple_call_use_set (stmt); + clobbers = gimple_call_clobber_set (stmt); + memset (uses, 0, sizeof (struct pt_solution)); + memset (clobbers, 0, sizeof (struct pt_solution)); + EXECUTE_IF_SET_IN_BITMAP (fi->solution, 0, i, bi) + { + struct pt_solution sol; + + vi = get_varinfo (i); + if (!vi->is_fn_info) + { + /* ??? We could be more precise here? */ + uses->nonlocal = 1; + uses->ipa_escaped = 1; + clobbers->nonlocal = 1; + clobbers->ipa_escaped = 1; + continue; + } + + if (!uses->anything) + { + find_what_var_points_to + (first_vi_for_offset (vi, fi_uses), &sol); + pt_solution_ior_into (uses, &sol); + } + if (!clobbers->anything) + { + find_what_var_points_to + (first_vi_for_offset (vi, fi_clobbers), &sol); + pt_solution_ior_into (clobbers, &sol); + } + } + } + } + } + } + + fn->gimple_df->ipa_pta = true; + } - in_ipa_mode = 0; - delete_alias_heapvars (); delete_points_to_sets (); + + in_ipa_mode = 0; + return 0; }