/* Vector where the parameter infos are actually stored. */
VEC (ipa_edge_args_t, gc) *ipa_edge_args_vector;
-/* Bitmap with all UIDs of call graph edges that have been already processed
- by indirect inlining. */
-static bitmap iinlining_processed_edges;
-
/* Holders of ipa cgraph hooks: */
static struct cgraph_edge_hook_list *edge_removal_hook_holder;
static struct cgraph_node_hook_list *node_removal_hook_holder;
}
}
-/* Count number of arguments callsite CS has and store it in
- ipa_edge_args structure corresponding to this callsite. */
-
-static void
-ipa_count_arguments (struct cgraph_edge *cs)
-{
- gimple stmt;
- int arg_num;
-
- stmt = cs->call_stmt;
- gcc_assert (is_gimple_call (stmt));
- arg_num = gimple_call_num_args (stmt);
- if (VEC_length (ipa_edge_args_t, ipa_edge_args_vector)
- <= (unsigned) cgraph_edge_max_uid)
- VEC_safe_grow_cleared (ipa_edge_args_t, gc,
- ipa_edge_args_vector, cgraph_edge_max_uid + 1);
- ipa_set_cs_argument_count (IPA_EDGE_REF (cs), arg_num);
-}
-
/* Print the jump functions associated with call graph edge CS to file F. */
static void
fprintf (f, "UNKNOWN\n");
else if (type == IPA_JF_KNOWN_TYPE)
{
- tree binfo_type = TREE_TYPE (jump_func->value.base_binfo);
- fprintf (f, "KNOWN TYPE, type in binfo is: ");
- print_generic_expr (f, binfo_type, 0);
- fprintf (f, " (%u)\n", TYPE_UID (binfo_type));
+ fprintf (f, "KNOWN TYPE: base ");
+ print_generic_expr (f, jump_func->value.known_type.base_type, 0);
+ fprintf (f, ", offset "HOST_WIDE_INT_PRINT_DEC", component ",
+ jump_func->value.known_type.offset);
+ print_generic_expr (f, jump_func->value.known_type.component_type, 0);
+ fprintf (f, "\n");
}
else if (type == IPA_JF_CONST)
{
tree_code_name[(int)
jump_func->value.pass_through.operation]);
if (jump_func->value.pass_through.operation != NOP_EXPR)
- print_generic_expr (dump_file,
+ print_generic_expr (f,
jump_func->value.pass_through.operand, 0);
- fprintf (dump_file, "\n");
+ fprintf (f, "\n");
}
else if (type == IPA_JF_ANCESTOR)
{
jump_func->value.ancestor.formal_id,
jump_func->value.ancestor.offset);
print_generic_expr (f, jump_func->value.ancestor.type, 0);
- fprintf (dump_file, "\n");
+ fprintf (f, "\n");
}
}
}
gimple call)
{
HOST_WIDE_INT offset, size, max_size;
- tree base, binfo;
+ tree base;
if (!flag_devirtualize
|| TREE_CODE (op) != ADDR_EXPR
|| is_global_var (base))
return;
- if (detect_type_change (op, base, call, jfunc, offset))
+ if (detect_type_change (op, base, call, jfunc, offset)
+ || !TYPE_BINFO (TREE_TYPE (base)))
return;
- binfo = TYPE_BINFO (TREE_TYPE (base));
- if (!binfo)
- return;
- binfo = get_binfo_at_offset (binfo, offset, TREE_TYPE (op));
- if (binfo)
- {
- jfunc->type = IPA_JF_KNOWN_TYPE;
- jfunc->value.base_binfo = binfo;
- }
+ jfunc->type = IPA_JF_KNOWN_TYPE;
+ jfunc->value.known_type.base_type = TREE_TYPE (base);
+ jfunc->value.known_type.offset = offset;
+ jfunc->value.known_type.component_type = TREE_TYPE (op);
}
static void
compute_scalar_jump_functions (struct ipa_node_params *info,
- struct ipa_jump_func *functions,
+ struct ipa_edge_args *args,
gimple call)
{
tree arg;
for (num = 0; num < gimple_call_num_args (call); num++)
{
+ struct ipa_jump_func *jfunc = ipa_get_ith_jump_func (args, num);
arg = gimple_call_arg (call, num);
if (is_gimple_ip_invariant (arg))
{
- functions[num].type = IPA_JF_CONST;
- functions[num].value.constant = arg;
+ jfunc->type = IPA_JF_CONST;
+ jfunc->value.constant = arg;
}
else if (TREE_CODE (arg) == SSA_NAME)
{
int index = ipa_get_param_decl_index (info, SSA_NAME_VAR (arg));
if (index >= 0
- && !detect_type_change_ssa (arg, call, &functions[num]))
+ && !detect_type_change_ssa (arg, call, jfunc))
{
- functions[num].type = IPA_JF_PASS_THROUGH;
- functions[num].value.pass_through.formal_id = index;
- functions[num].value.pass_through.operation = NOP_EXPR;
+ jfunc->type = IPA_JF_PASS_THROUGH;
+ jfunc->value.pass_through.formal_id = index;
+ jfunc->value.pass_through.operation = NOP_EXPR;
}
}
else
{
gimple stmt = SSA_NAME_DEF_STMT (arg);
if (is_gimple_assign (stmt))
- compute_complex_assign_jump_func (info, &functions[num],
- call, stmt, arg);
+ compute_complex_assign_jump_func (info, jfunc, call, stmt, arg);
else if (gimple_code (stmt) == GIMPLE_PHI)
- compute_complex_ancestor_jump_func (info, &functions[num],
- call, stmt);
+ compute_complex_ancestor_jump_func (info, jfunc, call, stmt);
}
}
else
- compute_known_type_jump_func (arg, &functions[num], call);
+ compute_known_type_jump_func (arg, jfunc, call);
}
}
static bool
compute_pass_through_member_ptrs (struct ipa_node_params *info,
struct param_analysis_info *parms_info,
- struct ipa_jump_func *functions,
+ struct ipa_edge_args *args,
gimple call)
{
bool undecided_members = false;
gcc_assert (index >=0);
if (!is_parm_modified_before_call (&parms_info[index], call, arg))
{
- functions[num].type = IPA_JF_PASS_THROUGH;
- functions[num].value.pass_through.formal_id = index;
- functions[num].value.pass_through.operation = NOP_EXPR;
+ struct ipa_jump_func *jfunc = ipa_get_ith_jump_func (args,
+ num);
+ jfunc->type = IPA_JF_PASS_THROUGH;
+ jfunc->value.pass_through.formal_id = index;
+ jfunc->value.pass_through.operation = NOP_EXPR;
}
else
undecided_members = true;
associated with the call. */
static void
-compute_cst_member_ptr_arguments (struct ipa_jump_func *functions,
+compute_cst_member_ptr_arguments (struct ipa_edge_args *args,
gimple call)
{
unsigned num;
for (num = 0; num < gimple_call_num_args (call); num++)
{
+ struct ipa_jump_func *jfunc = ipa_get_ith_jump_func (args, num);
arg = gimple_call_arg (call, num);
- if (functions[num].type == IPA_JF_UNKNOWN
+ if (jfunc->type == IPA_JF_UNKNOWN
&& type_like_member_ptr_p (TREE_TYPE (arg), &method_field,
&delta_field))
- determine_cst_member_ptr (call, arg, method_field, delta_field,
- &functions[num]);
+ determine_cst_member_ptr (call, arg, method_field, delta_field, jfunc);
}
}
struct cgraph_edge *cs)
{
struct ipa_node_params *info = IPA_NODE_REF (cs->caller);
- struct ipa_edge_args *arguments = IPA_EDGE_REF (cs);
- gimple call;
+ struct ipa_edge_args *args = IPA_EDGE_REF (cs);
+ gimple call = cs->call_stmt;
+ int arg_num = gimple_call_num_args (call);
- if (ipa_get_cs_argument_count (arguments) == 0 || arguments->jump_functions)
+ if (arg_num == 0 || args->jump_functions)
return;
- arguments->jump_functions = ggc_alloc_cleared_vec_ipa_jump_func
- (ipa_get_cs_argument_count (arguments));
-
- call = cs->call_stmt;
- gcc_assert (is_gimple_call (call));
+ VEC_safe_grow_cleared (ipa_jump_func_t, gc, args->jump_functions, arg_num);
/* We will deal with constants and SSA scalars first: */
- compute_scalar_jump_functions (info, arguments->jump_functions, call);
+ compute_scalar_jump_functions (info, args, call);
/* Let's check whether there are any potential member pointers and if so,
whether we can determine their functions as pass_through. */
- if (!compute_pass_through_member_ptrs (info, parms_info,
- arguments->jump_functions, call))
+ if (!compute_pass_through_member_ptrs (info, parms_info, args, call))
return;
/* Finally, let's check whether we actually pass a new constant member
pointer here... */
- compute_cst_member_ptr_arguments (arguments->jump_functions, call);
+ compute_cst_member_ptr_arguments (args, call);
}
/* Compute jump functions for all edges - both direct and indirect - outgoing
for (cs = node->callees; cs; cs = cs->next_callee)
{
- struct cgraph_node *callee = cgraph_function_or_thunk_node (cs->callee, NULL);
+ struct cgraph_node *callee = cgraph_function_or_thunk_node (cs->callee,
+ NULL);
/* We do not need to bother analyzing calls to unknown
functions unless they may become known during lto/whopr. */
- if (!cs->callee->analyzed && !flag_lto)
+ if (!callee->analyzed && !flag_lto)
continue;
- ipa_count_arguments (cs);
- /* If the descriptor of the callee is not initialized yet, we have to do
- it now. */
- if (callee->analyzed)
- ipa_initialize_node_params (callee);
- if (ipa_get_cs_argument_count (IPA_EDGE_REF (cs))
- != ipa_get_param_count (IPA_NODE_REF (callee)))
- ipa_set_called_with_variable_arg (IPA_NODE_REF (callee));
ipa_compute_jump_functions_for_edge (parms_info, cs);
}
for (cs = node->indirect_calls; cs; cs = cs->next_callee)
- {
- ipa_count_arguments (cs);
- ipa_compute_jump_functions_for_edge (parms_info, cs);
- }
+ ipa_compute_jump_functions_for_edge (parms_info, cs);
}
/* If RHS looks like a rhs of a statement loading pfn from a member
combine_known_type_and_ancestor_jfs (struct ipa_jump_func *src,
struct ipa_jump_func *dst)
{
- tree new_binfo;
+ HOST_WIDE_INT combined_offset;
+ tree combined_type;
- new_binfo = get_binfo_at_offset (src->value.base_binfo,
- dst->value.ancestor.offset,
- dst->value.ancestor.type);
- if (new_binfo)
- {
- dst->type = IPA_JF_KNOWN_TYPE;
- dst->value.base_binfo = new_binfo;
- }
- else
- dst->type = IPA_JF_UNKNOWN;
+ combined_offset = src->value.known_type.offset + dst->value.ancestor.offset;
+ combined_type = dst->value.ancestor.type;
+
+ dst->type = IPA_JF_KNOWN_TYPE;
+ dst->value.known_type.base_type = src->value.known_type.base_type;
+ dst->value.known_type.offset = combined_offset;
+ dst->value.known_type.component_type = combined_type;
}
/* Update the jump functions associated with call graph edge E when the call
}
callee = cgraph_function_or_thunk_node (callee, NULL);
- if (ipa_get_cs_argument_count (IPA_EDGE_REF (ie))
- != ipa_get_param_count (IPA_NODE_REF (callee)))
- ipa_set_called_with_variable_arg (IPA_NODE_REF (callee));
-
return ie;
}
try_make_edge_direct_virtual_call (struct cgraph_edge *ie,
struct ipa_jump_func *jfunc)
{
- tree binfo, type, target;
- HOST_WIDE_INT token;
-
- if (jfunc->type == IPA_JF_KNOWN_TYPE)
- binfo = jfunc->value.base_binfo;
- else
- return NULL;
+ tree binfo, target;
- if (!binfo)
+ if (jfunc->type != IPA_JF_KNOWN_TYPE)
return NULL;
- token = ie->indirect_info->otr_token;
- type = ie->indirect_info->otr_type;
- binfo = get_binfo_at_offset (binfo, ie->indirect_info->anc_offset, type);
+ binfo = TYPE_BINFO (jfunc->value.known_type.base_type);
+ gcc_checking_assert (binfo);
+ binfo = get_binfo_at_offset (binfo, jfunc->value.known_type.offset
+ + ie->indirect_info->anc_offset,
+ ie->indirect_info->otr_type);
if (binfo)
- target = gimple_get_virt_method_for_binfo (token, binfo);
+ target = gimple_get_virt_method_for_binfo (ie->indirect_info->otr_token,
+ binfo);
else
return NULL;
struct ipa_jump_func *jfunc;
next_ie = ie->next_callee;
- if (bitmap_bit_p (iinlining_processed_edges, ie->uid))
- continue;
- /* If we ever use indirect edges for anything other than indirect
- inlining, we will need to skip those with negative param_indices. */
if (ici->param_index == -1)
continue;
/* We must check range due to calls with variable number of arguments: */
if (ici->param_index >= ipa_get_cs_argument_count (top))
{
- bitmap_set_bit (iinlining_processed_edges, ie->uid);
+ ici->param_index = -1;
continue;
}
}
else
/* Either we can find a destination for this edge now or never. */
- bitmap_set_bit (iinlining_processed_edges, ie->uid);
+ ici->param_index = -1;
+
+ if (!flag_indirect_inlining)
+ continue;
if (ici->polymorphic)
new_direct_edge = try_make_edge_direct_virtual_call (ie, jfunc);
res |= propagate_info_to_inlined_callees (cs, e->callee, new_edges);
else
update_jump_functions_after_inlining (cs, e);
+ for (e = node->indirect_calls; e; e = e->next_callee)
+ update_jump_functions_after_inlining (cs, e);
return res;
}
ipa_propagate_indirect_call_infos (struct cgraph_edge *cs,
VEC (cgraph_edge_p, heap) **new_edges)
{
+ bool changed;
/* Do nothing if the preparation phase has not been carried out yet
(i.e. during early inlining). */
if (!ipa_node_params_vector)
return false;
gcc_assert (ipa_edge_args_vector);
- return propagate_info_to_inlined_callees (cs, cs->callee, new_edges);
+ changed = propagate_info_to_inlined_callees (cs, cs->callee, new_edges);
+
+ /* We do not keep jump functions of inlined edges up to date. Better to free
+ them so we do not access them accidentally. */
+ ipa_free_edge_args_substructures (IPA_EDGE_REF (cs));
+ return changed;
}
/* Frees all dynamically allocated structures that the argument info points
ipa_free_node_params_substructures (IPA_NODE_REF (node));
}
-static struct ipa_jump_func *
-duplicate_ipa_jump_func_array (const struct ipa_jump_func * src, size_t n)
-{
- struct ipa_jump_func *p;
-
- if (!src)
- return NULL;
-
- p = ggc_alloc_vec_ipa_jump_func (n);
- memcpy (p, src, n * sizeof (struct ipa_jump_func));
- return p;
-}
-
/* Hook that is called by cgraph.c when a node is duplicated. */
static void
__attribute__((unused)) void *data)
{
struct ipa_edge_args *old_args, *new_args;
- int arg_count;
ipa_check_create_edge_args ();
old_args = IPA_EDGE_REF (src);
new_args = IPA_EDGE_REF (dst);
- arg_count = ipa_get_cs_argument_count (old_args);
- ipa_set_cs_argument_count (new_args, arg_count);
- new_args->jump_functions =
- duplicate_ipa_jump_func_array (old_args->jump_functions, arg_count);
-
- if (iinlining_processed_edges
- && bitmap_bit_p (iinlining_processed_edges, src->uid))
- bitmap_set_bit (iinlining_processed_edges, dst->uid);
+ new_args->jump_functions = VEC_copy (ipa_jump_func_t, gc,
+ old_args->jump_functions);
}
/* Hook that is called by cgraph.c when a node is duplicated. */
new_info->lattices = NULL;
new_info->ipcp_orig_node = old_info->ipcp_orig_node;
- new_info->called_with_var_arguments = old_info->called_with_var_arguments;
new_info->uses_analysis_done = old_info->uses_analysis_done;
new_info->node_enqueued = old_info->node_enqueued;
}
function_insertion_hook_holder = NULL;
}
-/* Allocate all necessary data structures necessary for indirect inlining. */
-
-void
-ipa_create_all_structures_for_iinln (void)
-{
- iinlining_processed_edges = BITMAP_ALLOC (NULL);
-}
-
/* Free all ipa_node_params and all ipa_edge_args structures if they are no
longer needed after ipa-cp. */
void
ipa_free_all_structures_after_ipa_cp (void)
{
- if (!flag_indirect_inlining)
+ if (!optimize)
{
ipa_free_all_edge_args ();
ipa_free_all_node_params ();
void
ipa_free_all_structures_after_iinln (void)
{
- BITMAP_FREE (iinlining_processed_edges);
-
ipa_free_all_edge_args ();
ipa_free_all_node_params ();
ipa_unregister_cgraph_hooks ();
case IPA_JF_UNKNOWN:
break;
case IPA_JF_KNOWN_TYPE:
- stream_write_tree (ob, jump_func->value.base_binfo, true);
+ streamer_write_uhwi (ob, jump_func->value.known_type.offset);
+ stream_write_tree (ob, jump_func->value.known_type.base_type, true);
+ stream_write_tree (ob, jump_func->value.known_type.component_type, true);
break;
case IPA_JF_CONST:
stream_write_tree (ob, jump_func->value.constant, true);
case IPA_JF_UNKNOWN:
break;
case IPA_JF_KNOWN_TYPE:
- jump_func->value.base_binfo = stream_read_tree (ib, data_in);
+ jump_func->value.known_type.offset = streamer_read_uhwi (ib);
+ jump_func->value.known_type.base_type = stream_read_tree (ib, data_in);
+ jump_func->value.known_type.component_type = stream_read_tree (ib,
+ data_in);
break;
case IPA_JF_CONST:
jump_func->value.constant = stream_read_tree (ib, data_in);
struct ipa_edge_args *args = IPA_EDGE_REF (e);
int count = streamer_read_uhwi (ib);
- ipa_set_cs_argument_count (args, count);
if (!count)
continue;
+ VEC_safe_grow_cleared (ipa_jump_func_t, gc, args->jump_functions, count);
- args->jump_functions = ggc_alloc_cleared_vec_ipa_jump_func
- (ipa_get_cs_argument_count (args));
for (k = 0; k < ipa_get_cs_argument_count (args); k++)
ipa_read_jump_function (ib, ipa_get_ith_jump_func (args, k), data_in);
}
struct ipa_edge_args *args = IPA_EDGE_REF (e);
int count = streamer_read_uhwi (ib);
- ipa_set_cs_argument_count (args, count);
if (count)
{
- args->jump_functions = ggc_alloc_cleared_vec_ipa_jump_func
- (ipa_get_cs_argument_count (args));
+ VEC_safe_grow_cleared (ipa_jump_func_t, gc, args->jump_functions,
+ count);
for (k = 0; k < ipa_get_cs_argument_count (args); k++)
- ipa_read_jump_function (ib, ipa_get_ith_jump_func (args, k), data_in);
+ ipa_read_jump_function (ib, ipa_get_ith_jump_func (args, k),
+ data_in);
}
ipa_read_indirect_edge_info (ib, data_in, e);
}
ipa_update_after_lto_read (void)
{
struct cgraph_node *node;
- struct cgraph_edge *cs;
ipa_check_create_node_params ();
ipa_check_create_edge_args ();
for (node = cgraph_nodes; node; node = node->next)
if (node->analyzed)
ipa_initialize_node_params (node);
-
- for (node = cgraph_nodes; node; node = node->next)
- if (node->analyzed)
- for (cs = node->callees; cs; cs = cs->next_callee)
- {
- struct cgraph_node *callee;
-
- callee = cgraph_function_or_thunk_node (cs->callee, NULL);
- if (ipa_get_cs_argument_count (IPA_EDGE_REF (cs))
- != ipa_get_param_count (IPA_NODE_REF (callee)))
- ipa_set_called_with_variable_arg (IPA_NODE_REF (callee));
- }
}
-