X-Git-Url: http://git.sourceforge.jp/view?p=pf3gnuchains%2Fgcc-fork.git;a=blobdiff_plain;f=gcc%2Fipa-prop.c;h=6437389876e0f8bc0eda6e4473982b5542b53af4;hp=2842088d8f1b9d075c6f9456170edab6a06b50c7;hb=28918f8e0acd0e75c39839724f8fc13c56ca580a;hpb=5215027d365b04789c45a180119b17153bf9034b diff --git a/gcc/ipa-prop.c b/gcc/ipa-prop.c index 2842088d8f1..6437389876e 100644 --- a/gcc/ipa-prop.c +++ b/gcc/ipa-prop.c @@ -1,5 +1,6 @@ /* Interprocedural analyses. - Copyright (C) 2005, 2007, 2008, 2009 Free Software Foundation, Inc. + Copyright (C) 2005, 2007, 2008, 2009, 2010 + Free Software Foundation, Inc. This file is part of GCC. @@ -29,15 +30,21 @@ along with GCC; see the file COPYING3. If not see #include "tree-flow.h" #include "tree-pass.h" #include "tree-inline.h" +#include "gimple.h" #include "flags.h" #include "timevar.h" #include "flags.h" #include "diagnostic.h" +#include "lto-streamer.h" /* Vector where the parameter infos are actually stored. */ VEC (ipa_node_params_t, heap) *ipa_node_params_vector; /* Vector where the parameter infos are actually stored. */ -VEC (ipa_edge_args_t, heap) *ipa_edge_args_vector; +VEC (ipa_edge_args_t, gc) *ipa_edge_args_vector; + +/* Bitmap with all UIDs of call graph edges that have been already processed + by indirect inlining. */ +static bitmap iinlining_processed_edges; /* Holders of ipa cgraph hooks: */ static struct cgraph_edge_hook_list *edge_removal_hook_holder; @@ -197,11 +204,36 @@ visit_store_addr_for_mod_analysis (gimple stmt ATTRIBUTE_UNUSED, { struct ipa_node_params *info = (struct ipa_node_params *) data; - if (TREE_CODE (op) == PARM_DECL) + op = get_base_address (op); + if (op + && TREE_CODE (op) == PARM_DECL) { int index = ipa_get_param_decl_index (info, op); gcc_assert (index >= 0); info->params[index].modified = true; + info->params[index].used = true; + } + + return false; +} + +/* Callback of walk_stmt_load_store_addr_ops for the visit_load. + If OP is a parameter declaration, mark it as used in the info structure + passed in DATA. */ + +static bool +visit_load_for_mod_analysis (gimple stmt ATTRIBUTE_UNUSED, + tree op, void *data) +{ + struct ipa_node_params *info = (struct ipa_node_params *) data; + + op = get_base_address (op); + if (op + && TREE_CODE (op) == PARM_DECL) + { + int index = ipa_get_param_decl_index (info, op); + gcc_assert (index >= 0); + info->params[index].used = true; } return false; @@ -220,16 +252,35 @@ ipa_detect_param_modifications (struct cgraph_node *node) struct function *func; gimple_stmt_iterator gsi; struct ipa_node_params *info = IPA_NODE_REF (node); + int i; if (ipa_get_param_count (info) == 0 || info->modification_analysis_done) return; + for (i = 0; i < ipa_get_param_count (info); i++) + { + tree parm = ipa_get_param (info, i); + /* For SSA regs see if parameter is used. For non-SSA we compute + the flag during modification analysis. */ + if (is_gimple_reg (parm) + && gimple_default_def (DECL_STRUCT_FUNCTION (node->decl), parm)) + info->params[i].used = true; + } + func = DECL_STRUCT_FUNCTION (decl); FOR_EACH_BB_FN (bb, func) - for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi)) - walk_stmt_load_store_addr_ops (gsi_stmt (gsi), info, NULL, - visit_store_addr_for_mod_analysis, - visit_store_addr_for_mod_analysis); + { + for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi)) + walk_stmt_load_store_addr_ops (gsi_stmt (gsi), info, + visit_load_for_mod_analysis, + visit_store_addr_for_mod_analysis, + visit_store_addr_for_mod_analysis); + for (gsi = gsi_start (phi_nodes (bb)); !gsi_end_p (gsi); gsi_next (&gsi)) + walk_stmt_load_store_addr_ops (gsi_stmt (gsi), info, + visit_load_for_mod_analysis, + visit_store_addr_for_mod_analysis, + visit_store_addr_for_mod_analysis); + } info->modification_analysis_done = 1; } @@ -248,7 +299,7 @@ ipa_count_arguments (struct cgraph_edge *cs) arg_num = gimple_call_num_args (stmt); if (VEC_length (ipa_edge_args_t, ipa_edge_args_vector) <= (unsigned) cgraph_edge_max_uid) - VEC_safe_grow_cleared (ipa_edge_args_t, heap, + VEC_safe_grow_cleared (ipa_edge_args_t, gc, ipa_edge_args_vector, cgraph_edge_max_uid + 1); ipa_set_cs_argument_count (IPA_EDGE_REF (cs), arg_num); } @@ -282,11 +333,25 @@ ipa_print_node_jump_functions (FILE *f, struct cgraph_node *node) fprintf (f, " param %d: ", i); if (type == IPA_JF_UNKNOWN) fprintf (f, "UNKNOWN\n"); + else if (type == IPA_JF_KNOWN_TYPE) + { + tree binfo_type = TREE_TYPE (jump_func->value.base_binfo); + fprintf (f, "KNOWN TYPE, type in binfo is: "); + print_generic_expr (f, binfo_type, 0); + fprintf (f, " (%u)\n", TYPE_UID (binfo_type)); + } else if (type == IPA_JF_CONST) { tree val = jump_func->value.constant; fprintf (f, "CONST: "); print_generic_expr (f, val, 0); + if (TREE_CODE (val) == ADDR_EXPR + && TREE_CODE (TREE_OPERAND (val, 0)) == CONST_DECL) + { + fprintf (f, " -> "); + print_generic_expr (f, DECL_INITIAL (TREE_OPERAND (val, 0)), + 0); + } fprintf (f, "\n"); } else if (type == IPA_JF_CONST_MEMBER_PTR) @@ -312,9 +377,11 @@ ipa_print_node_jump_functions (FILE *f, struct cgraph_node *node) else if (type == IPA_JF_ANCESTOR) { fprintf (f, "ANCESTOR: "); - fprintf (f, "%d, offset "HOST_WIDE_INT_PRINT_DEC"\n", + fprintf (f, "%d, offset "HOST_WIDE_INT_PRINT_DEC", ", jump_func->value.ancestor.formal_id, jump_func->value.ancestor.offset); + print_generic_expr (f, jump_func->value.ancestor.type, 0); + fprintf (dump_file, "\n"); } } } @@ -334,50 +401,65 @@ ipa_print_all_jump_functions (FILE *f) } } -/* Determine whether passing ssa name NAME constitutes a polynomial - pass-through function or getting an address of an acestor and if so, write - such a jump function to JFUNC. INFO describes the caller. */ +/* Given that an actual argument is an SSA_NAME (given in NAME) and is a result + of an assignment statement STMT, try to find out whether NAME can be + described by a (possibly polynomial) pass-through jump-function or an + ancestor jump function and if so, write the appropriate function into + JFUNC */ static void -compute_complex_pass_through (struct ipa_node_params *info, - struct ipa_jump_func *jfunc, - tree name) +compute_complex_assign_jump_func (struct ipa_node_params *info, + struct ipa_jump_func *jfunc, + gimple stmt, tree name) { HOST_WIDE_INT offset, size, max_size; tree op1, op2, type; int index; - gimple stmt = SSA_NAME_DEF_STMT (name); - if (!is_gimple_assign (stmt)) - return; op1 = gimple_assign_rhs1 (stmt); op2 = gimple_assign_rhs2 (stmt); - if (op2) + if (TREE_CODE (op1) == SSA_NAME + && SSA_NAME_IS_DEFAULT_DEF (op1)) { - if (TREE_CODE (op1) != SSA_NAME - || !SSA_NAME_IS_DEFAULT_DEF (op1) - || !is_gimple_ip_invariant (op2)) + index = ipa_get_param_decl_index (info, SSA_NAME_VAR (op1)); + if (index < 0) return; - index = ipa_get_param_decl_index (info, SSA_NAME_VAR (op1)); - if (index >= 0) + if (op2) { + if (!is_gimple_ip_invariant (op2) + || (TREE_CODE_CLASS (gimple_expr_code (stmt)) != tcc_comparison + && !useless_type_conversion_p (TREE_TYPE (name), + TREE_TYPE (op1)))) + return; + jfunc->type = IPA_JF_PASS_THROUGH; jfunc->value.pass_through.formal_id = index; jfunc->value.pass_through.operation = gimple_assign_rhs_code (stmt); jfunc->value.pass_through.operand = op2; } + else if (gimple_assign_unary_nop_p (stmt)) + { + jfunc->type = IPA_JF_PASS_THROUGH; + jfunc->value.pass_through.formal_id = index; + jfunc->value.pass_through.operation = NOP_EXPR; + } return; } if (TREE_CODE (op1) != ADDR_EXPR) return; + op1 = TREE_OPERAND (op1, 0); type = TREE_TYPE (op1); - + if (TREE_CODE (type) != RECORD_TYPE) + return; op1 = get_ref_base_and_extent (op1, &offset, &size, &max_size); - if (TREE_CODE (op1) != INDIRECT_REF) + if (TREE_CODE (op1) != INDIRECT_REF + /* If this is a varying address, punt. */ + || max_size == -1 + || max_size != size) return; op1 = TREE_OPERAND (op1, 0); if (TREE_CODE (op1) != SSA_NAME @@ -395,6 +477,120 @@ compute_complex_pass_through (struct ipa_node_params *info, } +/* Given that an actual argument is an SSA_NAME that is a result of a phi + statement PHI, try to find out whether NAME is in fact a + multiple-inheritance typecast from a descendant into an ancestor of a formal + parameter and thus can be described by an ancestor jump function and if so, + write the appropriate function into JFUNC. + + Essentially we want to match the following pattern: + + if (obj_2(D) != 0B) + goto ; + else + goto ; + + : + iftmp.1_3 = &obj_2(D)->D.1762; + + : + # iftmp.1_1 = PHI + D.1879_6 = middleman_1 (iftmp.1_1, i_5(D)); + return D.1879_6; */ + +static void +compute_complex_ancestor_jump_func (struct ipa_node_params *info, + struct ipa_jump_func *jfunc, + gimple phi) +{ + HOST_WIDE_INT offset, size, max_size; + gimple assign, cond; + basic_block phi_bb, assign_bb, cond_bb; + tree tmp, parm, expr; + int index, i; + + if (gimple_phi_num_args (phi) != 2 + || !integer_zerop (PHI_ARG_DEF (phi, 1))) + return; + + tmp = PHI_ARG_DEF (phi, 0); + if (TREE_CODE (tmp) != SSA_NAME + || SSA_NAME_IS_DEFAULT_DEF (tmp) + || !POINTER_TYPE_P (TREE_TYPE (tmp)) + || TREE_CODE (TREE_TYPE (TREE_TYPE (tmp))) != RECORD_TYPE) + return; + + assign = SSA_NAME_DEF_STMT (tmp); + assign_bb = gimple_bb (assign); + if (!single_pred_p (assign_bb) + || !gimple_assign_single_p (assign)) + return; + expr = gimple_assign_rhs1 (assign); + + if (TREE_CODE (expr) != ADDR_EXPR) + return; + expr = TREE_OPERAND (expr, 0); + expr = get_ref_base_and_extent (expr, &offset, &size, &max_size); + + if (TREE_CODE (expr) != INDIRECT_REF + /* If this is a varying address, punt. */ + || max_size == -1 + || max_size != size) + return; + parm = TREE_OPERAND (expr, 0); + if (TREE_CODE (parm) != SSA_NAME + || !SSA_NAME_IS_DEFAULT_DEF (parm)) + return; + + index = ipa_get_param_decl_index (info, SSA_NAME_VAR (parm)); + if (index < 0) + return; + + cond_bb = single_pred (assign_bb); + cond = last_stmt (cond_bb); + if (gimple_code (cond) != GIMPLE_COND + || gimple_cond_code (cond) != NE_EXPR + || gimple_cond_lhs (cond) != parm + || !integer_zerop (gimple_cond_rhs (cond))) + return; + + + phi_bb = gimple_bb (phi); + for (i = 0; i < 2; i++) + { + basic_block pred = EDGE_PRED (phi_bb, i)->src; + if (pred != assign_bb && pred != cond_bb) + return; + } + + jfunc->type = IPA_JF_ANCESTOR; + jfunc->value.ancestor.formal_id = index; + jfunc->value.ancestor.offset = offset; + jfunc->value.ancestor.type = TREE_TYPE (TREE_TYPE (tmp)); +} + +/* Given OP whch is passed as an actual argument to a called function, + determine if it is possible to construct a KNOWN_TYPE jump function for it + and if so, create one and store it to JFUNC. */ + +static void +compute_known_type_jump_func (tree op, struct ipa_jump_func *jfunc) +{ + tree binfo; + + if (TREE_CODE (op) != ADDR_EXPR) + return; + + op = TREE_OPERAND (op, 0); + binfo = gimple_get_relevant_ref_binfo (op, NULL_TREE); + if (binfo) + { + jfunc->type = IPA_JF_KNOWN_TYPE; + jfunc->value.base_binfo = binfo; + } +} + + /* Determine the jump functions of scalar arguments. Scalar means SSA names and constants of a number of selected types. INFO is the ipa_node_params structure associated with the caller, FUNCTIONS is a pointer to an array of @@ -432,8 +628,18 @@ compute_scalar_jump_functions (struct ipa_node_params *info, } } else - compute_complex_pass_through (info, &functions[num], arg); + { + gimple stmt = SSA_NAME_DEF_STMT (arg); + if (is_gimple_assign (stmt)) + compute_complex_assign_jump_func (info, &functions[num], + stmt, arg); + else if (gimple_code (stmt) == GIMPLE_PHI) + compute_complex_ancestor_jump_func (info, &functions[num], + stmt); + } } + else + compute_known_type_jump_func (arg, &functions[num]); } } @@ -655,8 +861,8 @@ ipa_compute_jump_functions (struct cgraph_edge *cs) if (ipa_get_cs_argument_count (arguments) == 0 || arguments->jump_functions) return; - arguments->jump_functions = XCNEWVEC (struct ipa_jump_func, - ipa_get_cs_argument_count (arguments)); + arguments->jump_functions = GGC_CNEWVEC (struct ipa_jump_func, + ipa_get_cs_argument_count (arguments)); call = cs->call_stmt; gcc_assert (is_gimple_call (call)); @@ -728,39 +934,39 @@ ipa_is_ssa_with_stmt_def (tree t) return false; } -/* Creates a new note describing a call to a parameter number FORMAL_ID and - attaches it to the linked list of INFO. It also sets the called flag of the - parameter. STMT is the corresponding call statement. */ +/* Find the indirect call graph edge corresponding to STMT and add to it all + information necessary to describe a call to a parameter number PARAM_INDEX. + NODE is the caller. POLYMORPHIC should be set to true iff the call is a + virtual one. */ static void -ipa_note_param_call (struct ipa_node_params *info, int formal_id, - gimple stmt) +ipa_note_param_call (struct cgraph_node *node, int param_index, gimple stmt, + bool polymorphic) { - struct ipa_param_call_note *note; - basic_block bb = gimple_bb (stmt); - - info->params[formal_id].called = 1; - - note = XCNEW (struct ipa_param_call_note); - note->formal_id = formal_id; - note->stmt = stmt; - note->count = bb->count; - note->frequency = compute_call_stmt_bb_frequency (current_function_decl, bb); - - note->next = info->param_calls; - info->param_calls = note; + struct cgraph_edge *cs; - return; + cs = cgraph_edge (node, stmt); + cs->indirect_info->param_index = param_index; + cs->indirect_info->anc_offset = 0; + cs->indirect_info->polymorphic = polymorphic; + if (polymorphic) + { + tree otr = gimple_call_fn (stmt); + tree type, token = OBJ_TYPE_REF_TOKEN (otr); + cs->indirect_info->otr_token = tree_low_cst (token, 1); + type = TREE_TYPE (TREE_TYPE (OBJ_TYPE_REF_OBJECT (otr))); + cs->indirect_info->otr_type = type; + } } -/* Analyze the CALL and examine uses of formal parameters of the caller +/* Analyze the CALL and examine uses of formal parameters of the caller NODE (described by INFO). Currently it checks whether the call calls a pointer that is a formal parameter and if so, the parameter is marked with the - called flag and a note describing the call is created. This is very simple - for ordinary pointers represented in SSA but not-so-nice when it comes to - member pointers. The ugly part of this function does nothing more than - tries to match the pattern of such a call. An example of such a pattern is - the gimple dump below, the call is on the last line: + called flag and an indirect call graph edge describing the call is created. + This is very simple for ordinary pointers represented in SSA but not-so-nice + when it comes to member pointers. The ugly part of this function does + nothing more than trying to match the pattern of such a call. An example of + such a pattern is the gimple dump below, the call is on the last line: : f$__delta_5 = f.__delta; @@ -800,11 +1006,11 @@ ipa_note_param_call (struct ipa_node_params *info, int formal_id, */ static void -ipa_analyze_call_uses (struct ipa_node_params *info, gimple call) +ipa_analyze_indirect_call_uses (struct cgraph_node *node, + struct ipa_node_params *info, + gimple call, tree target) { - tree target = gimple_call_fn (call); gimple def; - tree var; tree n1, n2; gimple d1, d2; tree rec, rec2, cond; @@ -812,16 +1018,12 @@ ipa_analyze_call_uses (struct ipa_node_params *info, gimple call) int index; basic_block bb, virt_bb, join; - if (TREE_CODE (target) != SSA_NAME) - return; - - var = SSA_NAME_VAR (target); if (SSA_NAME_IS_DEFAULT_DEF (target)) { - /* assuming TREE_CODE (var) == PARM_DECL */ + tree var = SSA_NAME_VAR (target); index = ipa_get_param_decl_index (info, var); if (index >= 0) - ipa_note_param_call (info, index, call); + ipa_note_param_call (node, index, call, false); return; } @@ -918,20 +1120,73 @@ ipa_analyze_call_uses (struct ipa_node_params *info, gimple call) index = ipa_get_param_decl_index (info, rec); if (index >= 0 && !ipa_is_param_modified (info, index)) - ipa_note_param_call (info, index, call); + ipa_note_param_call (node, index, call, false); return; } -/* Analyze the statement STMT with respect to formal parameters (described in - INFO) and their uses. Currently it only checks whether formal parameters - are called. */ +/* Analyze a CALL to an OBJ_TYPE_REF which is passed in TARGET and if the + object referenced in the expression is a formal parameter of the caller + (described by INFO), create a call note for the statement. */ static void -ipa_analyze_stmt_uses (struct ipa_node_params *info, gimple stmt) +ipa_analyze_virtual_call_uses (struct cgraph_node *node, + struct ipa_node_params *info, gimple call, + tree target) +{ + tree obj = OBJ_TYPE_REF_OBJECT (target); + tree var; + int index; + + if (TREE_CODE (obj) == ADDR_EXPR) + { + do + { + obj = TREE_OPERAND (obj, 0); + } + while (TREE_CODE (obj) == COMPONENT_REF); + if (TREE_CODE (obj) != INDIRECT_REF) + return; + obj = TREE_OPERAND (obj, 0); + } + + if (TREE_CODE (obj) != SSA_NAME + || !SSA_NAME_IS_DEFAULT_DEF (obj)) + return; + + var = SSA_NAME_VAR (obj); + index = ipa_get_param_decl_index (info, var); + + if (index >= 0) + ipa_note_param_call (node, index, call, true); +} + +/* Analyze a call statement CALL whether and how it utilizes formal parameters + of the caller (described by INFO). */ + +static void +ipa_analyze_call_uses (struct cgraph_node *node, + struct ipa_node_params *info, gimple call) +{ + tree target = gimple_call_fn (call); + + if (TREE_CODE (target) == SSA_NAME) + ipa_analyze_indirect_call_uses (node, info, call, target); + else if (TREE_CODE (target) == OBJ_TYPE_REF) + ipa_analyze_virtual_call_uses (node, info, call, target); +} + + +/* Analyze the call statement STMT with respect to formal parameters (described + in INFO) of caller given by NODE. Currently it only checks whether formal + parameters are called. */ + +static void +ipa_analyze_stmt_uses (struct cgraph_node *node, struct ipa_node_params *info, + gimple stmt) { if (is_gimple_call (stmt)) - ipa_analyze_call_uses (info, stmt); + ipa_analyze_call_uses (node, info, stmt); } /* Scan the function body of NODE and inspect the uses of formal parameters. @@ -956,19 +1211,38 @@ ipa_analyze_params_uses (struct cgraph_node *node) for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi)) { gimple stmt = gsi_stmt (gsi); - ipa_analyze_stmt_uses (info, stmt); + ipa_analyze_stmt_uses (node, info, stmt); } } info->uses_analysis_done = 1; } +/* Update the jump function DST when the call graph edge correspondng to SRC is + is being inlined, knowing that DST is of type ancestor and src of known + type. */ + +static void +combine_known_type_and_ancestor_jfs (struct ipa_jump_func *src, + struct ipa_jump_func *dst) +{ + tree new_binfo; + + new_binfo = get_binfo_at_offset (src->value.base_binfo, + dst->value.ancestor.offset, + dst->value.ancestor.type); + if (new_binfo) + { + dst->type = IPA_JF_KNOWN_TYPE; + dst->value.base_binfo = new_binfo; + } + else + dst->type = IPA_JF_UNKNOWN; +} + /* Update the jump functions associated with call graph edge E when the call graph edge CS is being inlined, assuming that E->caller is already (possibly - indirectly) inlined into CS->callee and that E has not been inlined. - - We keep pass through functions only if they do not contain any operation. - This is sufficient for inlining and greately simplifies things. */ + indirectly) inlined into CS->callee and that E has not been inlined. */ static void update_jump_functions_after_inlining (struct cgraph_edge *cs, @@ -981,52 +1255,161 @@ update_jump_functions_after_inlining (struct cgraph_edge *cs, for (i = 0; i < count; i++) { - struct ipa_jump_func *src, *dst = ipa_get_ith_jump_func (args, i); + struct ipa_jump_func *dst = ipa_get_ith_jump_func (args, i); if (dst->type == IPA_JF_ANCESTOR) { - dst->type = IPA_JF_UNKNOWN; - continue; - } + struct ipa_jump_func *src; - if (dst->type != IPA_JF_PASS_THROUGH) - continue; + /* Variable number of arguments can cause havoc if we try to access + one that does not exist in the inlined edge. So make sure we + don't. */ + if (dst->value.ancestor.formal_id >= ipa_get_cs_argument_count (top)) + { + dst->type = IPA_JF_UNKNOWN; + continue; + } + + src = ipa_get_ith_jump_func (top, dst->value.ancestor.formal_id); + if (src->type == IPA_JF_KNOWN_TYPE) + combine_known_type_and_ancestor_jfs (src, dst); + else if (src->type == IPA_JF_CONST) + { + struct ipa_jump_func kt_func; - /* We must check range due to calls with variable number of arguments and - we cannot combine jump functions with operations. */ - if (dst->value.pass_through.operation != NOP_EXPR - || (dst->value.pass_through.formal_id - >= ipa_get_cs_argument_count (top))) + kt_func.type = IPA_JF_UNKNOWN; + compute_known_type_jump_func (src->value.constant, &kt_func); + if (kt_func.type == IPA_JF_KNOWN_TYPE) + combine_known_type_and_ancestor_jfs (&kt_func, dst); + else + dst->type = IPA_JF_UNKNOWN; + } + else if (src->type == IPA_JF_PASS_THROUGH + && src->value.pass_through.operation == NOP_EXPR) + dst->value.ancestor.formal_id = src->value.pass_through.formal_id; + else if (src->type == IPA_JF_ANCESTOR) + { + dst->value.ancestor.formal_id = src->value.ancestor.formal_id; + dst->value.ancestor.offset += src->value.ancestor.offset; + } + else + dst->type = IPA_JF_UNKNOWN; + } + else if (dst->type == IPA_JF_PASS_THROUGH) { - dst->type = IPA_JF_UNKNOWN; - continue; + struct ipa_jump_func *src; + /* We must check range due to calls with variable number of arguments + and we cannot combine jump functions with operations. */ + if (dst->value.pass_through.operation == NOP_EXPR + && (dst->value.pass_through.formal_id + < ipa_get_cs_argument_count (top))) + { + src = ipa_get_ith_jump_func (top, + dst->value.pass_through.formal_id); + *dst = *src; + } + else + dst->type = IPA_JF_UNKNOWN; } + } +} + +/* If TARGET is an addr_expr of a function declaration, make it the destination + of an indirect edge IE and return the edge. Otherwise, return NULL. */ - src = ipa_get_ith_jump_func (top, dst->value.pass_through.formal_id); - *dst = *src; +static struct cgraph_edge * +make_edge_direct_to_target (struct cgraph_edge *ie, tree target) +{ + struct cgraph_node *callee; + + if (TREE_CODE (target) != ADDR_EXPR) + return NULL; + target = TREE_OPERAND (target, 0); + if (TREE_CODE (target) != FUNCTION_DECL) + return NULL; + callee = cgraph_node (target); + if (!callee) + return NULL; + + cgraph_make_edge_direct (ie, callee); + if (dump_file) + { + fprintf (dump_file, "ipa-prop: Discovered %s call to a known target " + "(%s/%i -> %s/%i) for stmt ", + ie->indirect_info->polymorphic ? "a virtual" : "an indirect", + cgraph_node_name (ie->caller), ie->caller->uid, + cgraph_node_name (ie->callee), ie->callee->uid); + + if (ie->call_stmt) + print_gimple_stmt (dump_file, ie->call_stmt, 2, TDF_SLIM); + else + fprintf (dump_file, "with uid %i\n", ie->lto_stmt_uid); } + return ie; } -/* Print out a debug message to file F that we have discovered that an indirect - call described by NT is in fact a call of a known constant function described - by JFUNC. NODE is the node where the call is. */ +/* Try to find a destination for indirect edge IE that corresponds to a simple + call or a call of a member function pointer and where the destination is a + pointer formal parameter described by jump function JFUNC. If it can be + determined, return the newly direct edge, otherwise return NULL. */ -static void -print_edge_addition_message (FILE *f, struct ipa_param_call_note *nt, - struct ipa_jump_func *jfunc, - struct cgraph_node *node) +static struct cgraph_edge * +try_make_edge_direct_simple_call (struct cgraph_edge *ie, + struct ipa_jump_func *jfunc) +{ + tree target; + + if (jfunc->type == IPA_JF_CONST) + target = jfunc->value.constant; + else if (jfunc->type == IPA_JF_CONST_MEMBER_PTR) + target = jfunc->value.member_cst.pfn; + else + return NULL; + + return make_edge_direct_to_target (ie, target); +} + +/* Try to find a destination for indirect edge IE that corresponds to a + virtuall call based on a formal parameter which is described by jump + function JFUNC and if it can be determined, make it direct and return the + direct edge. Otherwise, return NULL. */ + +static struct cgraph_edge * +try_make_edge_direct_virtual_call (struct cgraph_edge *ie, + struct ipa_jump_func *jfunc) { - fprintf (f, "ipa-prop: Discovered an indirect call to a known target ("); - if (jfunc->type == IPA_JF_CONST_MEMBER_PTR) + tree binfo, type, target; + HOST_WIDE_INT token; + + if (jfunc->type == IPA_JF_KNOWN_TYPE) + binfo = jfunc->value.base_binfo; + else if (jfunc->type == IPA_JF_CONST) { - print_node_brief (f, "", jfunc->value.member_cst.pfn, 0); - print_node_brief (f, ", ", jfunc->value.member_cst.delta, 0); + tree cst = jfunc->value.constant; + if (TREE_CODE (cst) == ADDR_EXPR) + binfo = gimple_get_relevant_ref_binfo (TREE_OPERAND (cst, 0), + NULL_TREE); + else + return NULL; } else - print_node_brief(f, "", jfunc->value.constant, 0); + return NULL; + + if (!binfo) + return NULL; - fprintf (f, ") in %s: ", cgraph_node_name (node)); - print_gimple_stmt (f, nt->stmt, 2, TDF_SLIM); + token = ie->indirect_info->otr_token; + type = ie->indirect_info->otr_type; + binfo = get_binfo_at_offset (binfo, ie->indirect_info->anc_offset, type); + if (binfo) + target = gimple_fold_obj_type_ref_known_binfo (token, binfo); + else + return NULL; + + if (target) + return make_edge_direct_to_target (ie, target); + else + return NULL; } /* Update the param called notes associated with NODE when CS is being inlined, @@ -1036,83 +1419,74 @@ print_edge_addition_message (FILE *f, struct ipa_param_call_note *nt, unless NEW_EDGES is NULL. Return true iff a new edge(s) were created. */ static bool -update_call_notes_after_inlining (struct cgraph_edge *cs, - struct cgraph_node *node, - VEC (cgraph_edge_p, heap) **new_edges) +update_indirect_edges_after_inlining (struct cgraph_edge *cs, + struct cgraph_node *node, + VEC (cgraph_edge_p, heap) **new_edges) { - struct ipa_node_params *info = IPA_NODE_REF (node); struct ipa_edge_args *top = IPA_EDGE_REF (cs); - struct ipa_param_call_note *nt; + struct cgraph_edge *ie, *next_ie, *new_direct_edge; bool res = false; - for (nt = info->param_calls; nt; nt = nt->next) + ipa_check_create_edge_args (); + + for (ie = node->indirect_calls; ie; ie = next_ie) { + struct cgraph_indirect_call_info *ici = ie->indirect_info; struct ipa_jump_func *jfunc; - if (nt->processed) + next_ie = ie->next_callee; + if (bitmap_bit_p (iinlining_processed_edges, ie->uid)) + continue; + + /* If we ever use indirect edges for anything other than indirect + inlining, we will need to skip those with negative param_indices. */ + if (ici->param_index == -1) continue; /* We must check range due to calls with variable number of arguments: */ - if (nt->formal_id >= ipa_get_cs_argument_count (top)) + if (ici->param_index >= ipa_get_cs_argument_count (top)) { - nt->processed = true; + bitmap_set_bit (iinlining_processed_edges, ie->uid); continue; } - jfunc = ipa_get_ith_jump_func (top, nt->formal_id); + jfunc = ipa_get_ith_jump_func (top, ici->param_index); if (jfunc->type == IPA_JF_PASS_THROUGH && jfunc->value.pass_through.operation == NOP_EXPR) - nt->formal_id = jfunc->value.pass_through.formal_id; - else if (jfunc->type == IPA_JF_CONST - || jfunc->type == IPA_JF_CONST_MEMBER_PTR) + ici->param_index = jfunc->value.pass_through.formal_id; + else if (jfunc->type == IPA_JF_ANCESTOR) { - struct cgraph_node *callee; - struct cgraph_edge *new_indirect_edge; - tree decl; - - nt->processed = true; - if (jfunc->type == IPA_JF_CONST_MEMBER_PTR) - decl = jfunc->value.member_cst.pfn; - else - decl = jfunc->value.constant; - - if (TREE_CODE (decl) != ADDR_EXPR) - continue; - decl = TREE_OPERAND (decl, 0); - - if (TREE_CODE (decl) != FUNCTION_DECL) - continue; - callee = cgraph_node (decl); - if (!callee || !callee->local.inlinable) - continue; - - res = true; - if (dump_file) - print_edge_addition_message (dump_file, nt, jfunc, node); - - new_indirect_edge = cgraph_create_edge (node, callee, nt->stmt, - nt->count, nt->frequency, - nt->loop_nest); - new_indirect_edge->indirect_call = 1; - ipa_check_create_edge_args (); - if (new_edges) - VEC_safe_push (cgraph_edge_p, heap, *new_edges, new_indirect_edge); - top = IPA_EDGE_REF (cs); + ici->param_index = jfunc->value.ancestor.formal_id; + ici->anc_offset += jfunc->value.ancestor.offset; } else + /* Either we can find a destination for this edge now or never. */ + bitmap_set_bit (iinlining_processed_edges, ie->uid); + + if (ici->polymorphic) + new_direct_edge = try_make_edge_direct_virtual_call (ie, jfunc); + else + new_direct_edge = try_make_edge_direct_simple_call (ie, jfunc); + + if (new_direct_edge) { - /* Ancestor jum functions and pass theoughs with operations should - not be used on parameters that then get called. */ - gcc_assert (jfunc->type == IPA_JF_UNKNOWN); - nt->processed = true; + new_direct_edge->indirect_inlining_edge = 1; + if (new_edges) + { + VEC_safe_push (cgraph_edge_p, heap, *new_edges, + new_direct_edge); + top = IPA_EDGE_REF (cs); + res = true; + } } } + return res; } /* Recursively traverse subtree of NODE (including node) made of inlined cgraph_edges when CS has been inlined and invoke - update_call_notes_after_inlining on all nodes and + update_indirect_edges_after_inlining on all nodes and update_jump_functions_after_inlining on all non-inlined edges that lead out of this subtree. Newly discovered indirect edges will be added to *NEW_EDGES, unless NEW_EDGES is NULL. Return true iff a new edge(s) were @@ -1126,7 +1500,7 @@ propagate_info_to_inlined_callees (struct cgraph_edge *cs, struct cgraph_edge *e; bool res; - res = update_call_notes_after_inlining (cs, node, new_edges); + res = update_indirect_edges_after_inlining (cs, node, new_edges); for (e = node->callees; e; e = e->next_callee) if (!e->inline_failed) @@ -1147,6 +1521,10 @@ bool ipa_propagate_indirect_call_infos (struct cgraph_edge *cs, VEC (cgraph_edge_p, heap) **new_edges) { + /* FIXME lto: We do not stream out indirect call information. */ + if (flag_wpa) + return false; + /* Do nothing if the preparation phase has not been carried out yet (i.e. during early inlining). */ if (!ipa_node_params_vector) @@ -1163,7 +1541,7 @@ void ipa_free_edge_args_substructures (struct ipa_edge_args *args) { if (args->jump_functions) - free (args->jump_functions); + ggc_free (args->jump_functions); memset (args, 0, sizeof (*args)); } @@ -1181,7 +1559,7 @@ ipa_free_all_edge_args (void) i++) ipa_free_edge_args_substructures (args); - VEC_free (ipa_edge_args_t, heap, ipa_edge_args_vector); + VEC_free (ipa_edge_args_t, gc, ipa_edge_args_vector); ipa_edge_args_vector = NULL; } @@ -1194,13 +1572,6 @@ ipa_free_node_params_substructures (struct ipa_node_params *info) if (info->params) free (info->params); - while (info->param_calls) - { - struct ipa_param_call_note *note = info->param_calls; - info->param_calls = note->next; - free (note); - } - memset (info, 0, sizeof (*info)); } @@ -1238,6 +1609,10 @@ ipa_edge_removal_hook (struct cgraph_edge *cs, void *data ATTRIBUTE_UNUSED) static void ipa_node_removal_hook (struct cgraph_node *node, void *data ATTRIBUTE_UNUSED) { + /* During IPA-CP updating we can be called on not-yet analyze clones. */ + if (VEC_length (ipa_node_params_t, ipa_node_params_vector) + <= (unsigned)node->uid) + return; ipa_free_node_params_substructures (IPA_NODE_REF (node)); } @@ -1252,7 +1627,22 @@ duplicate_array (void *src, size_t n) if (!src) return NULL; - p = xcalloc (1, n); + p = xmalloc (n); + memcpy (p, src, n); + return p; +} + +/* Like duplicate_array byt in GGC memory. */ + +static void * +duplicate_ggc_array (void *src, size_t n) +{ + void *p; + + if (!src) + return NULL; + + p = ggc_alloc (n); memcpy (p, src, n); return p; } @@ -1274,8 +1664,12 @@ ipa_edge_duplication_hook (struct cgraph_edge *src, struct cgraph_edge *dst, arg_count = ipa_get_cs_argument_count (old_args); ipa_set_cs_argument_count (new_args, arg_count); new_args->jump_functions = (struct ipa_jump_func *) - duplicate_array (old_args->jump_functions, - sizeof (struct ipa_jump_func) * arg_count); + duplicate_ggc_array (old_args->jump_functions, + sizeof (struct ipa_jump_func) * arg_count); + + if (iinlining_processed_edges + && bitmap_bit_p (iinlining_processed_edges, src->uid)) + bitmap_set_bit (iinlining_processed_edges, dst->uid); } /* Hook that is called by cgraph.c when a node is duplicated. */ @@ -1285,7 +1679,6 @@ ipa_node_duplication_hook (struct cgraph_node *src, struct cgraph_node *dst, __attribute__((unused)) void *data) { struct ipa_node_params *old_info, *new_info; - struct ipa_param_call_note *note; int param_count; ipa_check_create_node_params (); @@ -1299,17 +1692,6 @@ ipa_node_duplication_hook (struct cgraph_node *src, struct cgraph_node *dst, sizeof (struct ipa_param_descriptor) * param_count); new_info->ipcp_orig_node = old_info->ipcp_orig_node; new_info->count_scale = old_info->count_scale; - - for (note = old_info->param_calls; note; note = note->next) - { - struct ipa_param_call_note *nn; - - nn = (struct ipa_param_call_note *) - xcalloc (1, sizeof (struct ipa_param_call_note)); - memcpy (nn, note, sizeof (struct ipa_param_call_note)); - nn->next = new_info->param_calls; - new_info->param_calls = nn; - } } /* Register our cgraph hooks if they are not already there. */ @@ -1346,11 +1728,19 @@ ipa_unregister_cgraph_hooks (void) node_duplication_hook_holder = NULL; } +/* Allocate all necessary data strucutures necessary for indirect inlining. */ + +void +ipa_create_all_structures_for_iinln (void) +{ + iinlining_processed_edges = BITMAP_ALLOC (NULL); +} + /* Free all ipa_node_params and all ipa_edge_args structures if they are no longer needed after ipa-cp. */ void -free_all_ipa_structures_after_ipa_cp (void) +ipa_free_all_structures_after_ipa_cp (void) { if (!flag_indirect_inlining) { @@ -1364,8 +1754,10 @@ free_all_ipa_structures_after_ipa_cp (void) longer needed after indirect inlining. */ void -free_all_ipa_structures_after_iinln (void) +ipa_free_all_structures_after_iinln (void) { + BITMAP_FREE (iinlining_processed_edges); + ipa_free_all_edge_args (); ipa_free_all_node_params (); ipa_unregister_cgraph_hooks (); @@ -1384,18 +1776,21 @@ ipa_print_node_params (FILE * f, struct cgraph_node *node) if (!node->analyzed) return; info = IPA_NODE_REF (node); - fprintf (f, " function %s Trees :: \n", cgraph_node_name (node)); + fprintf (f, " function %s parameter descriptors:\n", + cgraph_node_name (node)); count = ipa_get_param_count (info); for (i = 0; i < count; i++) { temp = ipa_get_param (info, i); if (TREE_CODE (temp) == PARM_DECL) fprintf (f, " param %d : %s", i, - (*lang_hooks.decl_printable_name) (temp, 2)); + (DECL_NAME (temp) + ? (*lang_hooks.decl_printable_name) (temp, 2) + : "(unnamed)")); if (ipa_is_param_modified (info, i)) fprintf (f, " modified"); - if (ipa_is_param_called (info, i)) - fprintf (f, " called"); + if (ipa_is_param_used (info, i)) + fprintf (f, " used"); fprintf (f, "\n"); } } @@ -1663,7 +2058,7 @@ ipa_modify_call_arguments (struct cgraph_edge *cs, gimple stmt, if (!useless_type_conversion_p (ptrtype, TREE_TYPE (expr))) expr = fold_convert (ptrtype, expr); expr = fold_build2 (POINTER_PLUS_EXPR, ptrtype, expr, - build_int_cst (size_type_node, + build_int_cst (sizetype, adj->offset / BITS_PER_UNIT)); if (!adj->by_ref) expr = fold_build1 (INDIRECT_REF, adj->type, expr); @@ -1863,3 +2258,338 @@ ipa_dump_param_adjustments (FILE *file, ipa_parm_adjustment_vec adjustments, VEC_free (tree, heap, parms); } +/* Stream out jump function JUMP_FUNC to OB. */ + +static void +ipa_write_jump_function (struct output_block *ob, + struct ipa_jump_func *jump_func) +{ + lto_output_uleb128_stream (ob->main_stream, + jump_func->type); + + switch (jump_func->type) + { + case IPA_JF_UNKNOWN: + break; + case IPA_JF_KNOWN_TYPE: + lto_output_tree (ob, jump_func->value.base_binfo, true); + break; + case IPA_JF_CONST: + lto_output_tree (ob, jump_func->value.constant, true); + break; + case IPA_JF_PASS_THROUGH: + lto_output_tree (ob, jump_func->value.pass_through.operand, true); + lto_output_uleb128_stream (ob->main_stream, + jump_func->value.pass_through.formal_id); + lto_output_uleb128_stream (ob->main_stream, + jump_func->value.pass_through.operation); + break; + case IPA_JF_ANCESTOR: + lto_output_uleb128_stream (ob->main_stream, + jump_func->value.ancestor.offset); + lto_output_tree (ob, jump_func->value.ancestor.type, true); + lto_output_uleb128_stream (ob->main_stream, + jump_func->value.ancestor.formal_id); + break; + case IPA_JF_CONST_MEMBER_PTR: + lto_output_tree (ob, jump_func->value.member_cst.pfn, true); + lto_output_tree (ob, jump_func->value.member_cst.delta, false); + break; + } +} + +/* Read in jump function JUMP_FUNC from IB. */ + +static void +ipa_read_jump_function (struct lto_input_block *ib, + struct ipa_jump_func *jump_func, + struct data_in *data_in) +{ + jump_func->type = (enum jump_func_type) lto_input_uleb128 (ib); + + switch (jump_func->type) + { + case IPA_JF_UNKNOWN: + break; + case IPA_JF_KNOWN_TYPE: + jump_func->value.base_binfo = lto_input_tree (ib, data_in); + break; + case IPA_JF_CONST: + jump_func->value.constant = lto_input_tree (ib, data_in); + break; + case IPA_JF_PASS_THROUGH: + jump_func->value.pass_through.operand = lto_input_tree (ib, data_in); + jump_func->value.pass_through.formal_id = lto_input_uleb128 (ib); + jump_func->value.pass_through.operation = (enum tree_code) lto_input_uleb128 (ib); + break; + case IPA_JF_ANCESTOR: + jump_func->value.ancestor.offset = lto_input_uleb128 (ib); + jump_func->value.ancestor.type = lto_input_tree (ib, data_in); + jump_func->value.ancestor.formal_id = lto_input_uleb128 (ib); + break; + case IPA_JF_CONST_MEMBER_PTR: + jump_func->value.member_cst.pfn = lto_input_tree (ib, data_in); + jump_func->value.member_cst.delta = lto_input_tree (ib, data_in); + break; + } +} + +/* Stream out parts of cgraph_indirect_call_info corresponding to CS that are + relevant to indirect inlining to OB. */ + +static void +ipa_write_indirect_edge_info (struct output_block *ob, + struct cgraph_edge *cs) +{ + struct cgraph_indirect_call_info *ii = cs->indirect_info; + struct bitpack_d *bp; + + lto_output_sleb128_stream (ob->main_stream, ii->param_index); + lto_output_sleb128_stream (ob->main_stream, ii->anc_offset); + bp = bitpack_create (); + bp_pack_value (bp, ii->polymorphic, 1); + lto_output_bitpack (ob->main_stream, bp); + bitpack_delete (bp); + + if (ii->polymorphic) + { + lto_output_sleb128_stream (ob->main_stream, ii->otr_token); + lto_output_tree (ob, ii->otr_type, true); + } +} + +/* Read in parts of cgraph_indirect_call_info corresponding to CS that are + relevant to indirect inlining from IB. */ + +static void +ipa_read_indirect_edge_info (struct lto_input_block *ib, + struct data_in *data_in ATTRIBUTE_UNUSED, + struct cgraph_edge *cs) +{ + struct cgraph_indirect_call_info *ii = cs->indirect_info; + struct bitpack_d *bp; + + ii->param_index = (int) lto_input_sleb128 (ib); + ii->anc_offset = (HOST_WIDE_INT) lto_input_sleb128 (ib); + bp = lto_input_bitpack (ib); + ii->polymorphic = bp_unpack_value (bp, 1); + bitpack_delete (bp); + if (ii->polymorphic) + { + ii->otr_token = (HOST_WIDE_INT) lto_input_sleb128 (ib); + ii->otr_type = lto_input_tree (ib, data_in); + } +} + +/* Stream out NODE info to OB. */ + +static void +ipa_write_node_info (struct output_block *ob, struct cgraph_node *node) +{ + int node_ref; + lto_cgraph_encoder_t encoder; + struct ipa_node_params *info = IPA_NODE_REF (node); + int j; + struct cgraph_edge *e; + struct bitpack_d *bp; + + encoder = ob->decl_state->cgraph_node_encoder; + node_ref = lto_cgraph_encoder_encode (encoder, node); + lto_output_uleb128_stream (ob->main_stream, node_ref); + + bp = bitpack_create (); + bp_pack_value (bp, info->called_with_var_arguments, 1); + bp_pack_value (bp, info->uses_analysis_done, 1); + gcc_assert (info->modification_analysis_done + || ipa_get_param_count (info) == 0); + gcc_assert (!info->node_enqueued); + gcc_assert (!info->ipcp_orig_node); + for (j = 0; j < ipa_get_param_count (info); j++) + { + bp_pack_value (bp, info->params[j].modified, 1); + bp_pack_value (bp, info->params[j].used, 1); + } + lto_output_bitpack (ob->main_stream, bp); + bitpack_delete (bp); + for (e = node->callees; e; e = e->next_callee) + { + struct ipa_edge_args *args = IPA_EDGE_REF (e); + + lto_output_uleb128_stream (ob->main_stream, + ipa_get_cs_argument_count (args)); + for (j = 0; j < ipa_get_cs_argument_count (args); j++) + ipa_write_jump_function (ob, ipa_get_ith_jump_func (args, j)); + } + for (e = node->indirect_calls; e; e = e->next_callee) + ipa_write_indirect_edge_info (ob, e); +} + +/* Srtream in NODE info from IB. */ + +static void +ipa_read_node_info (struct lto_input_block *ib, struct cgraph_node *node, + struct data_in *data_in) +{ + struct ipa_node_params *info = IPA_NODE_REF (node); + int k; + struct cgraph_edge *e; + struct bitpack_d *bp; + + ipa_initialize_node_params (node); + + bp = lto_input_bitpack (ib); + info->called_with_var_arguments = bp_unpack_value (bp, 1); + info->uses_analysis_done = bp_unpack_value (bp, 1); + if (ipa_get_param_count (info) != 0) + { + info->modification_analysis_done = true; + info->uses_analysis_done = true; + } + info->node_enqueued = false; + for (k = 0; k < ipa_get_param_count (info); k++) + { + info->params[k].modified = bp_unpack_value (bp, 1); + info->params[k].used = bp_unpack_value (bp, 1); + } + bitpack_delete (bp); + for (e = node->callees; e; e = e->next_callee) + { + struct ipa_edge_args *args = IPA_EDGE_REF (e); + int count = lto_input_uleb128 (ib); + + ipa_set_cs_argument_count (args, count); + if (!count) + continue; + + args->jump_functions = GGC_CNEWVEC (struct ipa_jump_func, + ipa_get_cs_argument_count (args)); + for (k = 0; k < ipa_get_cs_argument_count (args); k++) + ipa_read_jump_function (ib, ipa_get_ith_jump_func (args, k), data_in); + } + for (e = node->indirect_calls; e; e = e->next_callee) + ipa_read_indirect_edge_info (ib, data_in, e); +} + +/* Write jump functions for nodes in SET. */ + +void +ipa_prop_write_jump_functions (cgraph_node_set set) +{ + struct cgraph_node *node; + struct output_block *ob = create_output_block (LTO_section_jump_functions); + unsigned int count = 0; + cgraph_node_set_iterator csi; + + ob->cgraph_node = NULL; + + for (csi = csi_start (set); !csi_end_p (csi); csi_next (&csi)) + { + node = csi_node (csi); + if (node->analyzed && IPA_NODE_REF (node) != NULL) + count++; + } + + lto_output_uleb128_stream (ob->main_stream, count); + + /* Process all of the functions. */ + for (csi = csi_start (set); !csi_end_p (csi); csi_next (&csi)) + { + node = csi_node (csi); + if (node->analyzed && IPA_NODE_REF (node) != NULL) + ipa_write_node_info (ob, node); + } + lto_output_1_stream (ob->main_stream, 0); + produce_asm (ob, NULL); + destroy_output_block (ob); +} + +/* Read section in file FILE_DATA of length LEN with data DATA. */ + +static void +ipa_prop_read_section (struct lto_file_decl_data *file_data, const char *data, + size_t len) +{ + const struct lto_function_header *header = + (const struct lto_function_header *) data; + const int32_t cfg_offset = sizeof (struct lto_function_header); + const int32_t main_offset = cfg_offset + header->cfg_size; + const int32_t string_offset = main_offset + header->main_size; + struct data_in *data_in; + struct lto_input_block ib_main; + unsigned int i; + unsigned int count; + + LTO_INIT_INPUT_BLOCK (ib_main, (const char *) data + main_offset, 0, + header->main_size); + + data_in = + lto_data_in_create (file_data, (const char *) data + string_offset, + header->string_size, NULL); + count = lto_input_uleb128 (&ib_main); + + for (i = 0; i < count; i++) + { + unsigned int index; + struct cgraph_node *node; + lto_cgraph_encoder_t encoder; + + index = lto_input_uleb128 (&ib_main); + encoder = file_data->cgraph_node_encoder; + node = lto_cgraph_encoder_deref (encoder, index); + gcc_assert (node->analyzed); + ipa_read_node_info (&ib_main, node, data_in); + } + lto_free_section_data (file_data, LTO_section_jump_functions, NULL, data, + len); + lto_data_in_delete (data_in); +} + +/* Read ipcp jump functions. */ + +void +ipa_prop_read_jump_functions (void) +{ + struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data (); + struct lto_file_decl_data *file_data; + unsigned int j = 0; + + ipa_check_create_node_params (); + ipa_check_create_edge_args (); + ipa_register_cgraph_hooks (); + + while ((file_data = file_data_vec[j++])) + { + size_t len; + const char *data = lto_get_section_data (file_data, LTO_section_jump_functions, NULL, &len); + + if (data) + ipa_prop_read_section (file_data, data, len); + } +} + +/* After merging units, we can get mismatch in argument counts. + Also decl merging might've rendered parameter lists obsolette. + Also compute called_with_variable_arg info. */ + +void +ipa_update_after_lto_read (void) +{ + struct cgraph_node *node; + struct cgraph_edge *cs; + + ipa_check_create_node_params (); + ipa_check_create_edge_args (); + + for (node = cgraph_nodes; node; node = node->next) + if (node->analyzed) + ipa_initialize_node_params (node); + + for (node = cgraph_nodes; node; node = node->next) + if (node->analyzed) + for (cs = node->callees; cs; cs = cs->next_callee) + { + if (ipa_get_cs_argument_count (IPA_EDGE_REF (cs)) + != ipa_get_param_count (IPA_NODE_REF (cs->callee))) + ipa_set_called_with_variable_arg (IPA_NODE_REF (cs->callee)); + } +}