1 /* Integrated Register Allocator (IRA) intercommunication header file.
2 Copyright (C) 2006, 2007, 2008
3 Free Software Foundation, Inc.
4 Contributed by Vladimir Makarov <vmakarov@redhat.com>.
6 This file is part of GCC.
8 GCC is free software; you can redistribute it and/or modify it under
9 the terms of the GNU General Public License as published by the Free
10 Software Foundation; either version 3, or (at your option) any later
13 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
14 WARRANTY; without even the implied warranty of MERCHANTABILITY or
15 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
18 You should have received a copy of the GNU General Public License
19 along with GCC; see the file COPYING3. If not see
20 <http://www.gnu.org/licenses/>. */
24 #include "alloc-pool.h"
26 /* To provide consistency in naming, all IRA external variables,
27 functions, common typedefs start with prefix ira_. */
29 #ifdef ENABLE_CHECKING
30 #define ENABLE_IRA_CHECKING
33 #ifdef ENABLE_IRA_CHECKING
34 #define ira_assert(c) gcc_assert (c)
39 /* Compute register frequency from edge frequency FREQ. It is
40 analogous to REG_FREQ_FROM_BB. When optimizing for size, or
41 profile driven feedback is available and the function is never
42 executed, frequency is always equivalent. Otherwise rescale the
44 #define REG_FREQ_FROM_EDGE_FREQ(freq) \
45 (optimize_size || (flag_branch_probabilities && !ENTRY_BLOCK_PTR->count) \
46 ? REG_FREQ_MAX : (freq * REG_FREQ_MAX / BB_FREQ_MAX) \
47 ? (freq * REG_FREQ_MAX / BB_FREQ_MAX) : 1)
49 /* All natural loops. */
50 extern struct loops ira_loops;
52 /* A modified value of flag `-fira-verbose' used internally. */
53 extern int internal_flag_ira_verbose;
55 /* Dump file of the allocator if it is not NULL. */
56 extern FILE *ira_dump_file;
58 /* Typedefs for pointers to allocno live range, allocno, and copy of
60 typedef struct ira_allocno_live_range *allocno_live_range_t;
61 typedef struct ira_allocno *ira_allocno_t;
62 typedef struct ira_allocno_copy *ira_copy_t;
64 /* Definition of vector of allocnos and copies. */
65 DEF_VEC_P(ira_allocno_t);
66 DEF_VEC_ALLOC_P(ira_allocno_t, heap);
67 DEF_VEC_P(ira_copy_t);
68 DEF_VEC_ALLOC_P(ira_copy_t, heap);
70 /* Typedef for pointer to the subsequent structure. */
71 typedef struct ira_loop_tree_node *ira_loop_tree_node_t;
73 /* In general case, IRA is a regional allocator. The regions are
74 nested and form a tree. Currently regions are natural loops. The
75 following structure describes loop tree node (representing basic
76 block or loop). We need such tree because the loop tree from
77 cfgloop.h is not convenient for the optimization: basic blocks are
78 not a part of the tree from cfgloop.h. We also use the nodes for
79 storing additional information about basic blocks/loops for the
80 register allocation purposes. */
81 struct ira_loop_tree_node
83 /* The node represents basic block if children == NULL. */
84 basic_block bb; /* NULL for loop. */
85 struct loop *loop; /* NULL for BB. */
86 /* NEXT/SUBLOOP_NEXT is the next node/loop-node of the same parent.
87 SUBLOOP_NEXT is always NULL for BBs. */
88 ira_loop_tree_node_t subloop_next, next;
89 /* CHILDREN/SUBLOOPS is the first node/loop-node immediately inside
90 the node. They are NULL for BBs. */
91 ira_loop_tree_node_t subloops, children;
92 /* The node immediately containing given node. */
93 ira_loop_tree_node_t parent;
95 /* Loop level in range [0, ira_loop_tree_height). */
98 /* All the following members are defined only for nodes representing
101 /* Allocnos in the loop corresponding to their regnos. If it is
102 NULL the loop does not form a separate register allocation region
103 (e.g. because it has abnormal enter/exit edges and we can not put
104 code for register shuffling on the edges if a different
105 allocation is used for a pseudo-register on different sides of
106 the edges). Caps are not in the map (remember we can have more
107 one cap with the same regno in a region). */
108 ira_allocno_t *regno_allocno_map;
110 /* True if there is an entry to given loop not from its parent (or
111 grandparent) basic block. For example, it is possible for two
112 adjacent loops inside another loop. */
113 bool entered_from_non_parent_p;
115 /* Maximal register pressure inside loop for given register class
116 (defined only for the cover classes). */
117 int reg_pressure[N_REG_CLASSES];
119 /* Numbers of allocnos referred or living in the loop node (except
120 for its subloops). */
123 /* Numbers of allocnos living at the loop borders. */
124 bitmap border_allocnos;
126 /* Regnos of pseudos modified in the loop node (including its
128 bitmap modified_regnos;
130 /* Numbers of copies referred in the corresponding loop. */
134 /* The root of the loop tree corresponding to the all function. */
135 extern ira_loop_tree_node_t ira_loop_tree_root;
137 /* Height of the loop tree. */
138 extern int ira_loop_tree_height;
140 /* All nodes representing basic blocks are referred through the
141 following array. We can not use basic block member `aux' for this
142 because it is used for insertion of insns on edges. */
143 extern ira_loop_tree_node_t ira_bb_nodes;
145 /* Two access macros to the nodes representing basic blocks. */
146 #if defined ENABLE_IRA_CHECKING && (GCC_VERSION >= 2007)
147 #define IRA_BB_NODE_BY_INDEX(index) __extension__ \
148 (({ ira_loop_tree_node_t _node = (&ira_bb_nodes[index]); \
149 if (_node->children != NULL || _node->loop != NULL || _node->bb == NULL)\
152 "\n%s: %d: error in %s: it is not a block node\n", \
153 __FILE__, __LINE__, __FUNCTION__); \
154 gcc_unreachable (); \
158 #define IRA_BB_NODE_BY_INDEX(index) (&ira_bb_nodes[index])
161 #define IRA_BB_NODE(bb) IRA_BB_NODE_BY_INDEX ((bb)->index)
163 /* All nodes representing loops are referred through the following
165 extern ira_loop_tree_node_t ira_loop_nodes;
167 /* Two access macros to the nodes representing loops. */
168 #if defined ENABLE_IRA_CHECKING && (GCC_VERSION >= 2007)
169 #define IRA_LOOP_NODE_BY_INDEX(index) __extension__ \
170 (({ ira_loop_tree_node_t const _node = (&ira_loop_nodes[index]);\
171 if (_node->children == NULL || _node->bb != NULL || _node->loop == NULL)\
174 "\n%s: %d: error in %s: it is not a loop node\n", \
175 __FILE__, __LINE__, __FUNCTION__); \
176 gcc_unreachable (); \
180 #define IRA_LOOP_NODE_BY_INDEX(index) (&ira_loop_nodes[index])
183 #define IRA_LOOP_NODE(loop) IRA_LOOP_NODE_BY_INDEX ((loop)->num)
187 /* The structure describes program points where a given allocno lives.
188 To save memory we store allocno conflicts only for the same cover
189 class allocnos which is enough to assign hard registers. To find
190 conflicts for other allocnos (e.g. to assign stack memory slot) we
191 use the live ranges. If the live ranges of two allocnos are
192 intersected, the allocnos are in conflict. */
193 struct ira_allocno_live_range
195 /* Allocno whose live range is described by given structure. */
196 ira_allocno_t allocno;
197 /* Program point range. */
199 /* Next structure describing program points where the allocno
201 allocno_live_range_t next;
202 /* Pointer to structures with the same start/finish. */
203 allocno_live_range_t start_next, finish_next;
206 /* Program points are enumerated by numbers from range
207 0..IRA_MAX_POINT-1. There are approximately two times more program
208 points than insns. Program points are places in the program where
209 liveness info can be changed. In most general case (there are more
210 complicated cases too) some program points correspond to places
211 where input operand dies and other ones correspond to places where
212 output operands are born. */
213 extern int ira_max_point;
215 /* Arrays of size IRA_MAX_POINT mapping a program point to the allocno
216 live ranges with given start/finish point. */
217 extern allocno_live_range_t *ira_start_point_ranges, *ira_finish_point_ranges;
219 /* A structure representing an allocno (allocation entity). Allocno
220 represents a pseudo-register in an allocation region. If
221 pseudo-register does not live in a region but it lives in the
222 nested regions, it is represented in the region by special allocno
223 called *cap*. There may be more one cap representing the same
224 pseudo-register in region. It means that the corresponding
225 pseudo-register lives in more one non-intersected subregion. */
228 /* The allocno order number starting with 0. Each allocno has an
229 unique number and the number is never changed for the
232 /* Regno for allocno or cap. */
234 /* Mode of the allocno which is the mode of the corresponding
236 enum machine_mode mode;
237 /* Final rtx representation of the allocno. */
239 /* Hard register assigned to given allocno. Negative value means
240 that memory was allocated to the allocno. During the reload,
241 spilled allocno has value equal to the corresponding stack slot
242 number (0, ...) - 2. Value -1 is used for allocnos spilled by the
243 reload (at this point pseudo-register has only one allocno) which
244 did not get stack slot yet. */
246 /* Allocnos with the same regno are linked by the following member.
247 Allocnos corresponding to inner loops are first in the list (it
248 corresponds to depth-first traverse of the loops). */
249 ira_allocno_t next_regno_allocno;
250 /* There may be different allocnos with the same regno in different
251 regions. Allocnos are bound to the corresponding loop tree node.
252 Pseudo-register may have only one regular allocno with given loop
253 tree node but more than one cap (see comments above). */
254 ira_loop_tree_node_t loop_tree_node;
255 /* Accumulated usage references of the allocno. Here and below,
256 word 'accumulated' means info for given region and all nested
257 subregions. In this case, 'accumulated' means sum of references
258 of the corresponding pseudo-register in this region and in all
259 nested subregions recursively. */
261 /* Accumulated frequency of usage of the allocno. */
263 /* Register class which should be used for allocation for given
264 allocno. NO_REGS means that we should use memory. */
265 enum reg_class cover_class;
266 /* Minimal accumulated and updated costs of usage register of the
267 cover class for the allocno. */
268 int cover_class_cost, updated_cover_class_cost;
269 /* Minimal accumulated, and updated costs of memory for the allocno.
270 At the allocation start, the original and updated costs are
271 equal. The updated cost may be changed after finishing
272 allocation in a region and starting allocation in a subregion.
273 The change reflects the cost of spill/restore code on the
274 subregion border if we assign memory to the pseudo in the
276 int memory_cost, updated_memory_cost;
277 /* Accumulated number of points where the allocno lives and there is
278 excess pressure for its class. Excess pressure for a register
279 class at some point means that there are more allocnos of given
280 register class living at the point than number of hard-registers
281 of the class available for the allocation. */
282 int excess_pressure_points_num;
283 /* Copies to other non-conflicting allocnos. The copies can
284 represent move insn or potential move insn usually because of two
285 operand insn constraints. */
286 ira_copy_t allocno_copies;
287 /* It is a allocno (cap) representing given allocno on upper loop tree
290 /* It is a link to allocno (cap) on lower loop level represented by
291 given cap. Null if given allocno is not a cap. */
292 ira_allocno_t cap_member;
293 /* Coalesced allocnos form a cyclic list. One allocno given by
294 FIRST_COALESCED_ALLOCNO represents all coalesced allocnos. The
295 list is chained by NEXT_COALESCED_ALLOCNO. */
296 ira_allocno_t first_coalesced_allocno;
297 ira_allocno_t next_coalesced_allocno;
298 /* Pointer to structures describing at what program point the
299 allocno lives. We always maintain the list in such way that *the
300 ranges in the list are not intersected and ordered by decreasing
301 their program points*. */
302 allocno_live_range_t live_ranges;
303 /* Before building conflicts the two member values are
304 correspondingly minimal and maximal points of the accumulated
305 allocno live ranges. After building conflicts the values are
306 correspondingly minimal and maximal conflict ids of allocnos with
307 which given allocno can conflict. */
309 /* The unique member value represents given allocno in conflict bit
312 /* Vector of accumulated conflicting allocnos with NULL end marker
313 (if CONFLICT_VEC_P is true) or conflict bit vector otherwise.
314 Only allocnos with the same cover class are in the vector or in
316 void *conflict_allocno_array;
317 /* Allocated size of the previous array. */
318 unsigned int conflict_allocno_array_size;
319 /* Number of accumulated conflicts in the vector of conflicting
321 int conflict_allocnos_num;
322 /* Initial and accumulated hard registers conflicting with this
323 allocno and as a consequences can not be assigned to the allocno.
324 All non-allocatable hard regs and hard regs of cover classes
325 different from given allocno one are included in the sets. */
326 HARD_REG_SET conflict_hard_regs, total_conflict_hard_regs;
327 /* Accumulated frequency of calls which given allocno
330 /* Length of the previous array (number of the intersected calls). */
331 int calls_crossed_num;
332 /* Non NULL if we remove restoring value from given allocno to
333 MEM_OPTIMIZED_DEST at loop exit (see ira-emit.c) because the
334 allocno value is not changed inside the loop. */
335 ira_allocno_t mem_optimized_dest;
336 /* TRUE if the allocno assigned to memory was a destination of
337 removed move (see ira-emit.c) at loop exit because the value of
338 the corresponding pseudo-register is not changed inside the
340 unsigned int mem_optimized_dest_p : 1;
341 /* TRUE if the corresponding pseudo-register has disjoint live
342 ranges and the other allocnos of the pseudo-register except this
344 unsigned int somewhere_renamed_p : 1;
345 /* TRUE if allocno with the same REGNO in a subregion has been
346 renamed, in other words, got a new pseudo-register. */
347 unsigned int child_renamed_p : 1;
348 /* During the reload, value TRUE means that we should not reassign a
349 hard register to the allocno got memory earlier. It is set up
350 when we removed memory-memory move insn before each iteration of
352 unsigned int dont_reassign_p : 1;
354 /* Set to TRUE if allocno can't be assigned to the stack hard
355 register correspondingly in this region and area including the
356 region and all its subregions recursively. */
357 unsigned int no_stack_reg_p : 1, total_no_stack_reg_p : 1;
359 /* TRUE value means that there is no sense to spill the allocno
360 during coloring because the spill will result in additional
361 reloads in reload pass. */
362 unsigned int bad_spill_p : 1;
363 /* TRUE value means that the allocno was not removed yet from the
364 conflicting graph during colouring. */
365 unsigned int in_graph_p : 1;
366 /* TRUE if a hard register or memory has been assigned to the
368 unsigned int assigned_p : 1;
369 /* TRUE if it is put on the stack to make other allocnos
371 unsigned int may_be_spilled_p : 1;
372 /* TRUE if the allocno was removed from the splay tree used to
373 choose allocn for spilling (see ira-color.c::. */
374 unsigned int splay_removed_p : 1;
375 /* TRUE if conflicts for given allocno are represented by vector of
376 pointers to the conflicting allocnos. Otherwise, we use a bit
377 vector where a bit with given index represents allocno with the
379 unsigned int conflict_vec_p : 1;
380 /* Array of usage costs (accumulated and the one updated during
381 coloring) for each hard register of the allocno cover class. The
382 member value can be NULL if all costs are the same and equal to
383 COVER_CLASS_COST. For example, the costs of two different hard
384 registers can be different if one hard register is callee-saved
385 and another one is callee-used and the allocno lives through
386 calls. Another example can be case when for some insn the
387 corresponding pseudo-register value should be put in specific
388 register class (e.g. AREG for x86) which is a strict subset of
389 the allocno cover class (GENERAL_REGS for x86). We have updated
390 costs to reflect the situation when the usage cost of a hard
391 register is decreased because the allocno is connected to another
392 allocno by a copy and the another allocno has been assigned to
393 the hard register. */
394 int *hard_reg_costs, *updated_hard_reg_costs;
395 /* Array of decreasing costs (accumulated and the one updated during
396 coloring) for allocnos conflicting with given allocno for hard
397 regno of the allocno cover class. The member value can be NULL
398 if all costs are the same. These costs are used to reflect
399 preferences of other allocnos not assigned yet during assigning
401 int *conflict_hard_reg_costs, *updated_conflict_hard_reg_costs;
402 /* Number of the same cover class allocnos with TRUE in_graph_p
403 value and conflicting with given allocno during each point of
405 int left_conflicts_num;
406 /* Number of hard registers of the allocno cover class really
407 available for the allocno allocation. */
408 int available_regs_num;
409 /* Allocnos in a bucket (used in coloring) chained by the following
411 ira_allocno_t next_bucket_allocno;
412 ira_allocno_t prev_bucket_allocno;
413 /* Used for temporary purposes. */
417 /* All members of the allocno structures should be accessed only
418 through the following macros. */
419 #define ALLOCNO_NUM(A) ((A)->num)
420 #define ALLOCNO_REGNO(A) ((A)->regno)
421 #define ALLOCNO_REG(A) ((A)->reg)
422 #define ALLOCNO_NEXT_REGNO_ALLOCNO(A) ((A)->next_regno_allocno)
423 #define ALLOCNO_LOOP_TREE_NODE(A) ((A)->loop_tree_node)
424 #define ALLOCNO_CAP(A) ((A)->cap)
425 #define ALLOCNO_CAP_MEMBER(A) ((A)->cap_member)
426 #define ALLOCNO_CONFLICT_ALLOCNO_ARRAY(A) ((A)->conflict_allocno_array)
427 #define ALLOCNO_CONFLICT_ALLOCNO_ARRAY_SIZE(A) \
428 ((A)->conflict_allocno_array_size)
429 #define ALLOCNO_CONFLICT_ALLOCNOS_NUM(A) \
430 ((A)->conflict_allocnos_num)
431 #define ALLOCNO_CONFLICT_HARD_REGS(A) ((A)->conflict_hard_regs)
432 #define ALLOCNO_TOTAL_CONFLICT_HARD_REGS(A) ((A)->total_conflict_hard_regs)
433 #define ALLOCNO_NREFS(A) ((A)->nrefs)
434 #define ALLOCNO_FREQ(A) ((A)->freq)
435 #define ALLOCNO_HARD_REGNO(A) ((A)->hard_regno)
436 #define ALLOCNO_CALL_FREQ(A) ((A)->call_freq)
437 #define ALLOCNO_CALLS_CROSSED_NUM(A) ((A)->calls_crossed_num)
438 #define ALLOCNO_MEM_OPTIMIZED_DEST(A) ((A)->mem_optimized_dest)
439 #define ALLOCNO_MEM_OPTIMIZED_DEST_P(A) ((A)->mem_optimized_dest_p)
440 #define ALLOCNO_SOMEWHERE_RENAMED_P(A) ((A)->somewhere_renamed_p)
441 #define ALLOCNO_CHILD_RENAMED_P(A) ((A)->child_renamed_p)
442 #define ALLOCNO_DONT_REASSIGN_P(A) ((A)->dont_reassign_p)
444 #define ALLOCNO_NO_STACK_REG_P(A) ((A)->no_stack_reg_p)
445 #define ALLOCNO_TOTAL_NO_STACK_REG_P(A) ((A)->total_no_stack_reg_p)
447 #define ALLOCNO_BAD_SPILL_P(A) ((A)->bad_spill_p)
448 #define ALLOCNO_IN_GRAPH_P(A) ((A)->in_graph_p)
449 #define ALLOCNO_ASSIGNED_P(A) ((A)->assigned_p)
450 #define ALLOCNO_MAY_BE_SPILLED_P(A) ((A)->may_be_spilled_p)
451 #define ALLOCNO_SPLAY_REMOVED_P(A) ((A)->splay_removed_p)
452 #define ALLOCNO_CONFLICT_VEC_P(A) ((A)->conflict_vec_p)
453 #define ALLOCNO_MODE(A) ((A)->mode)
454 #define ALLOCNO_COPIES(A) ((A)->allocno_copies)
455 #define ALLOCNO_HARD_REG_COSTS(A) ((A)->hard_reg_costs)
456 #define ALLOCNO_UPDATED_HARD_REG_COSTS(A) ((A)->updated_hard_reg_costs)
457 #define ALLOCNO_CONFLICT_HARD_REG_COSTS(A) \
458 ((A)->conflict_hard_reg_costs)
459 #define ALLOCNO_UPDATED_CONFLICT_HARD_REG_COSTS(A) \
460 ((A)->updated_conflict_hard_reg_costs)
461 #define ALLOCNO_LEFT_CONFLICTS_NUM(A) ((A)->left_conflicts_num)
462 #define ALLOCNO_COVER_CLASS(A) ((A)->cover_class)
463 #define ALLOCNO_COVER_CLASS_COST(A) ((A)->cover_class_cost)
464 #define ALLOCNO_UPDATED_COVER_CLASS_COST(A) ((A)->updated_cover_class_cost)
465 #define ALLOCNO_MEMORY_COST(A) ((A)->memory_cost)
466 #define ALLOCNO_UPDATED_MEMORY_COST(A) ((A)->updated_memory_cost)
467 #define ALLOCNO_EXCESS_PRESSURE_POINTS_NUM(A) ((A)->excess_pressure_points_num)
468 #define ALLOCNO_AVAILABLE_REGS_NUM(A) ((A)->available_regs_num)
469 #define ALLOCNO_NEXT_BUCKET_ALLOCNO(A) ((A)->next_bucket_allocno)
470 #define ALLOCNO_PREV_BUCKET_ALLOCNO(A) ((A)->prev_bucket_allocno)
471 #define ALLOCNO_TEMP(A) ((A)->temp)
472 #define ALLOCNO_FIRST_COALESCED_ALLOCNO(A) ((A)->first_coalesced_allocno)
473 #define ALLOCNO_NEXT_COALESCED_ALLOCNO(A) ((A)->next_coalesced_allocno)
474 #define ALLOCNO_LIVE_RANGES(A) ((A)->live_ranges)
475 #define ALLOCNO_MIN(A) ((A)->min)
476 #define ALLOCNO_MAX(A) ((A)->max)
477 #define ALLOCNO_CONFLICT_ID(A) ((A)->conflict_id)
479 /* Map regno -> allocnos with given regno (see comments for
480 allocno member `next_regno_allocno'). */
481 extern ira_allocno_t *ira_regno_allocno_map;
483 /* Array of references to all allocnos. The order number of the
484 allocno corresponds to the index in the array. Removed allocnos
485 have NULL element value. */
486 extern ira_allocno_t *ira_allocnos;
488 /* Sizes of the previous array. */
489 extern int ira_allocnos_num;
491 /* Map conflict id -> allocno with given conflict id (see comments for
492 allocno member `conflict_id'). */
493 extern ira_allocno_t *ira_conflict_id_allocno_map;
495 /* The following structure represents a copy of two allocnos. The
496 copies represent move insns or potential move insns usually because
497 of two operand insn constraints. To remove register shuffle, we
498 also create copies between allocno which is output of an insn and
499 allocno becoming dead in the insn. */
500 struct ira_allocno_copy
502 /* The unique order number of the copy node starting with 0. */
504 /* Allocnos connected by the copy. The first allocno should have
505 smaller order number than the second one. */
506 ira_allocno_t first, second;
507 /* Execution frequency of the copy. */
510 /* It is a move insn which is an origin of the copy. The member
511 value for the copy representing two operand insn constraints or
512 for the copy created to remove register shuffle is NULL. In last
513 case the copy frequency is smaller than the corresponding insn
514 execution frequency. */
516 /* All copies with the same allocno as FIRST are linked by the two
517 following members. */
518 ira_copy_t prev_first_allocno_copy, next_first_allocno_copy;
519 /* All copies with the same allocno as SECOND are linked by the two
520 following members. */
521 ira_copy_t prev_second_allocno_copy, next_second_allocno_copy;
522 /* Region from which given copy is originated. */
523 ira_loop_tree_node_t loop_tree_node;
526 /* Array of references to all copies. The order number of the copy
527 corresponds to the index in the array. Removed copies have NULL
529 extern ira_copy_t *ira_copies;
531 /* Size of the previous array. */
532 extern int ira_copies_num;
534 /* The following structure describes a stack slot used for spilled
536 struct ira_spilled_reg_stack_slot
538 /* pseudo-registers assigned to the stack slot. */
539 regset_head spilled_regs;
540 /* RTL representation of the stack slot. */
542 /* Size of the stack slot. */
546 /* The number of elements in the following array. */
547 extern int ira_spilled_reg_stack_slots_num;
549 /* The following array contains info about spilled pseudo-registers
550 stack slots used in current function so far. */
551 extern struct ira_spilled_reg_stack_slot *ira_spilled_reg_stack_slots;
553 /* Correspondingly overall cost of the allocation, cost of the
554 allocnos assigned to hard-registers, cost of the allocnos assigned
555 to memory, cost of loads, stores and register move insns generated
556 for pseudo-register live range splitting (see ira-emit.c). */
557 extern int ira_overall_cost;
558 extern int ira_reg_cost, ira_mem_cost;
559 extern int ira_load_cost, ira_store_cost, ira_shuffle_cost;
560 extern int ira_move_loops_num, ira_additional_jumps_num;
562 /* Map: hard register number -> cover class it belongs to. If the
563 corresponding class is NO_REGS, the hard register is not available
565 extern enum reg_class ira_hard_regno_cover_class[FIRST_PSEUDO_REGISTER];
567 /* Map: register class x machine mode -> number of hard registers of
568 given class needed to store value of given mode. If the number for
569 some hard-registers of the register class is different, the size
571 extern int ira_reg_class_nregs[N_REG_CLASSES][MAX_MACHINE_MODE];
573 /* Maximal value of the previous array elements. */
574 extern int ira_max_nregs;
576 /* The number of bits in each element of array used to implement a bit
577 vector of allocnos and what type that element has. We use the
578 largest integer format on the host machine. */
579 #define IRA_INT_BITS HOST_BITS_PER_WIDE_INT
580 #define IRA_INT_TYPE HOST_WIDE_INT
582 /* Set, clear or test bit number I in R, a bit vector of elements with
583 minimal index and maximal index equal correspondingly to MIN and
585 #if defined ENABLE_IRA_CHECKING && (GCC_VERSION >= 2007)
587 #define SET_ALLOCNO_SET_BIT(R, I, MIN, MAX) __extension__ \
588 (({ int _min = (MIN), _max = (MAX), _i = (I); \
589 if (_i < _min || _i > _max) \
592 "\n%s: %d: error in %s: %d not in range [%d,%d]\n", \
593 __FILE__, __LINE__, __FUNCTION__, _i, _min, _max); \
594 gcc_unreachable (); \
596 ((R)[(unsigned) (_i - _min) / IRA_INT_BITS] \
597 |= ((IRA_INT_TYPE) 1 << ((unsigned) (_i - _min) % IRA_INT_BITS))); }))
600 #define CLEAR_ALLOCNO_SET_BIT(R, I, MIN, MAX) __extension__ \
601 (({ int _min = (MIN), _max = (MAX), _i = (I); \
602 if (_i < _min || _i > _max) \
605 "\n%s: %d: error in %s: %d not in range [%d,%d]\n", \
606 __FILE__, __LINE__, __FUNCTION__, _i, _min, _max); \
607 gcc_unreachable (); \
609 ((R)[(unsigned) (_i - _min) / IRA_INT_BITS] \
610 &= ~((IRA_INT_TYPE) 1 << ((unsigned) (_i - _min) % IRA_INT_BITS))); }))
612 #define TEST_ALLOCNO_SET_BIT(R, I, MIN, MAX) __extension__ \
613 (({ int _min = (MIN), _max = (MAX), _i = (I); \
614 if (_i < _min || _i > _max) \
617 "\n%s: %d: error in %s: %d not in range [%d,%d]\n", \
618 __FILE__, __LINE__, __FUNCTION__, _i, _min, _max); \
619 gcc_unreachable (); \
621 ((R)[(unsigned) (_i - _min) / IRA_INT_BITS] \
622 & ((IRA_INT_TYPE) 1 << ((unsigned) (_i - _min) % IRA_INT_BITS))); }))
626 #define SET_ALLOCNO_SET_BIT(R, I, MIN, MAX) \
627 ((R)[(unsigned) ((I) - (MIN)) / IRA_INT_BITS] \
628 |= ((IRA_INT_TYPE) 1 << ((unsigned) ((I) - (MIN)) % IRA_INT_BITS)))
630 #define CLEAR_ALLOCNO_SET_BIT(R, I, MIN, MAX) \
631 ((R)[(unsigned) ((I) - (MIN)) / IRA_INT_BITS] \
632 &= ~((IRA_INT_TYPE) 1 << ((unsigned) ((I) - (MIN)) % IRA_INT_BITS)))
634 #define TEST_ALLOCNO_SET_BIT(R, I, MIN, MAX) \
635 ((R)[(unsigned) ((I) - (MIN)) / IRA_INT_BITS] \
636 & ((IRA_INT_TYPE) 1 << ((unsigned) ((I) - (MIN)) % IRA_INT_BITS)))
640 /* The iterator for allocno set implemented ed as allocno bit
644 /* Array containing the allocno bit vector. */
647 /* The number of the current element in the vector. */
648 unsigned int word_num;
650 /* The number of bits in the bit vector. */
653 /* The current bit index of the bit vector. */
654 unsigned int bit_num;
656 /* Index corresponding to the 1st bit of the bit vector. */
659 /* The word of the bit vector currently visited. */
660 unsigned IRA_INT_TYPE word;
661 } ira_allocno_set_iterator;
663 /* Initialize the iterator I for allocnos bit vector VEC containing
664 minimal and maximal values MIN and MAX. */
666 ira_allocno_set_iter_init (ira_allocno_set_iterator *i,
667 IRA_INT_TYPE *vec, int min, int max)
671 i->nel = max < min ? 0 : max - min + 1;
674 i->word = i->nel == 0 ? 0 : vec[0];
677 /* Return TRUE if we have more allocnos to visit, in which case *N is
678 set to the allocno number to be visited. Otherwise, return
681 ira_allocno_set_iter_cond (ira_allocno_set_iterator *i, int *n)
683 /* Skip words that are zeros. */
684 for (; i->word == 0; i->word = i->vec[i->word_num])
687 i->bit_num = i->word_num * IRA_INT_BITS;
689 /* If we have reached the end, break. */
690 if (i->bit_num >= i->nel)
694 /* Skip bits that are zero. */
695 for (; (i->word & 1) == 0; i->word >>= 1)
698 *n = (int) i->bit_num + i->start_val;
703 /* Advance to the next allocno in the set. */
705 ira_allocno_set_iter_next (ira_allocno_set_iterator *i)
711 /* Loop over all elements of allocno set given by bit vector VEC and
712 their minimal and maximal values MIN and MAX. In each iteration, N
713 is set to the number of next allocno. ITER is an instance of
714 ira_allocno_set_iterator used to iterate the allocnos in the set. */
715 #define FOR_EACH_ALLOCNO_IN_SET(VEC, MIN, MAX, N, ITER) \
716 for (ira_allocno_set_iter_init (&(ITER), (VEC), (MIN), (MAX)); \
717 ira_allocno_set_iter_cond (&(ITER), &(N)); \
718 ira_allocno_set_iter_next (&(ITER)))
722 /* Map: hard regs X modes -> set of hard registers for storing value
723 of given mode starting with given hard register. */
724 extern HARD_REG_SET ira_reg_mode_hard_regset
725 [FIRST_PSEUDO_REGISTER][NUM_MACHINE_MODES];
727 /* Arrays analogous to macros MEMORY_MOVE_COST and
728 REGISTER_MOVE_COST. */
729 extern short ira_memory_move_cost[MAX_MACHINE_MODE][N_REG_CLASSES][2];
730 extern move_table *ira_register_move_cost[MAX_MACHINE_MODE];
732 /* Similar to may_move_in_cost but it is calculated in IRA instead of
733 regclass. Another difference we take only available hard registers
734 into account to figure out that one register class is a subset of
736 extern move_table *ira_may_move_in_cost[MAX_MACHINE_MODE];
738 /* Similar to may_move_out_cost but it is calculated in IRA instead of
739 regclass. Another difference we take only available hard registers
740 into account to figure out that one register class is a subset of
742 extern move_table *ira_may_move_out_cost[MAX_MACHINE_MODE];
744 /* Register class subset relation: TRUE if the first class is a subset
745 of the second one considering only hard registers available for the
747 extern int ira_class_subset_p[N_REG_CLASSES][N_REG_CLASSES];
749 /* Array of number of hard registers of given class which are
750 available for the allocation. The order is defined by the
752 extern short ira_class_hard_regs[N_REG_CLASSES][FIRST_PSEUDO_REGISTER];
754 /* The number of elements of the above array for given register
756 extern int ira_class_hard_regs_num[N_REG_CLASSES];
758 /* Index (in ira_class_hard_regs) for given register class and hard
759 register (in general case a hard register can belong to several
760 register classes). The index is negative for hard registers
761 unavailable for the allocation. */
762 extern short ira_class_hard_reg_index[N_REG_CLASSES][FIRST_PSEUDO_REGISTER];
764 /* Function specific hard registers can not be used for the register
766 extern HARD_REG_SET ira_no_alloc_regs;
768 /* Number of given class hard registers available for the register
769 allocation for given classes. */
770 extern int ira_available_class_regs[N_REG_CLASSES];
772 /* Array whose values are hard regset of hard registers available for
773 the allocation of given register class whose HARD_REGNO_MODE_OK
774 values for given mode are zero. */
775 extern HARD_REG_SET prohibited_class_mode_regs
776 [N_REG_CLASSES][NUM_MACHINE_MODES];
778 /* Array whose values are hard regset of hard registers for which
779 move of the hard register in given mode into itself is
781 extern HARD_REG_SET ira_prohibited_mode_move_regs[NUM_MACHINE_MODES];
783 /* Number of cover classes. Cover classes is non-intersected register
784 classes containing all hard-registers available for the
786 extern int ira_reg_class_cover_size;
788 /* The array containing cover classes (see also comments for macro
789 IRA_COVER_CLASSES). Only first IRA_REG_CLASS_COVER_SIZE elements are
791 extern enum reg_class ira_reg_class_cover[N_REG_CLASSES];
793 /* The value is number of elements in the subsequent array. */
794 extern int ira_important_classes_num;
796 /* The array containing non-empty classes (including non-empty cover
797 classes) which are subclasses of cover classes. Such classes is
798 important for calculation of the hard register usage costs. */
799 extern enum reg_class ira_important_classes[N_REG_CLASSES];
801 /* The array containing indexes of important classes in the previous
802 array. The array elements are defined only for important
804 extern int ira_important_class_nums[N_REG_CLASSES];
806 /* Map of all register classes to corresponding cover class containing
807 the given class. If given class is not a subset of a cover class,
808 we translate it into the cheapest cover class. */
809 extern enum reg_class ira_class_translate[N_REG_CLASSES];
811 /* The biggest important class inside of intersection of the two
812 classes (that is calculated taking only hard registers available
813 for allocation into account). If the both classes contain no hard
814 registers available for allocation, the value is calculated with
815 taking all hard-registers including fixed ones into account. */
816 extern enum reg_class ira_reg_class_intersect[N_REG_CLASSES][N_REG_CLASSES];
818 /* The biggest important class inside of union of the two classes
819 (that is calculated taking only hard registers available for
820 allocation into account). If the both classes contain no hard
821 registers available for allocation, the value is calculated with
822 taking all hard-registers including fixed ones into account. In
823 other words, the value is the corresponding reg_class_subunion
825 extern enum reg_class ira_reg_class_union[N_REG_CLASSES][N_REG_CLASSES];
827 extern void *ira_allocate (size_t);
828 extern void *ira_reallocate (void *, size_t);
829 extern void ira_free (void *addr);
830 extern bitmap ira_allocate_bitmap (void);
831 extern void ira_free_bitmap (bitmap);
832 extern void ira_print_disposition (FILE *);
833 extern void ira_debug_disposition (void);
834 extern void ira_debug_class_cover (void);
835 extern void ira_init_register_move_cost (enum machine_mode);
837 /* The length of the two following arrays. */
838 extern int ira_reg_equiv_len;
840 /* The element value is TRUE if the corresponding regno value is
842 extern bool *ira_reg_equiv_invariant_p;
844 /* The element value is equiv constant of given pseudo-register or
846 extern rtx *ira_reg_equiv_const;
850 /* The current loop tree node and its regno allocno map. */
851 extern ira_loop_tree_node_t ira_curr_loop_tree_node;
852 extern ira_allocno_t *ira_curr_regno_allocno_map;
854 extern void ira_debug_copy (ira_copy_t);
855 extern void ira_debug_copies (void);
856 extern void ira_debug_allocno_copies (ira_allocno_t);
858 extern void ira_traverse_loop_tree (bool, ira_loop_tree_node_t,
859 void (*) (ira_loop_tree_node_t),
860 void (*) (ira_loop_tree_node_t));
861 extern ira_allocno_t ira_create_allocno (int, bool, ira_loop_tree_node_t);
862 extern void ira_set_allocno_cover_class (ira_allocno_t, enum reg_class);
863 extern bool ira_conflict_vector_profitable_p (ira_allocno_t, int);
864 extern void ira_allocate_allocno_conflict_vec (ira_allocno_t, int);
865 extern void ira_allocate_allocno_conflicts (ira_allocno_t, int);
866 extern void ira_add_allocno_conflict (ira_allocno_t, ira_allocno_t);
867 extern void ira_print_expanded_allocno (ira_allocno_t);
868 extern allocno_live_range_t ira_create_allocno_live_range
869 (ira_allocno_t, int, int, allocno_live_range_t);
870 extern allocno_live_range_t ira_copy_allocno_live_range_list
871 (allocno_live_range_t);
872 extern allocno_live_range_t ira_merge_allocno_live_ranges
873 (allocno_live_range_t, allocno_live_range_t);
874 extern bool ira_allocno_live_ranges_intersect_p (allocno_live_range_t,
875 allocno_live_range_t);
876 extern void ira_finish_allocno_live_range (allocno_live_range_t);
877 extern void ira_finish_allocno_live_range_list (allocno_live_range_t);
878 extern void ira_free_allocno_updated_costs (ira_allocno_t);
879 extern ira_copy_t ira_create_copy (ira_allocno_t, ira_allocno_t,
880 int, bool, rtx, ira_loop_tree_node_t);
881 extern void ira_add_allocno_copy_to_list (ira_copy_t);
882 extern void ira_swap_allocno_copy_ends_if_necessary (ira_copy_t);
883 extern void ira_remove_allocno_copy_from_list (ira_copy_t);
884 extern ira_copy_t ira_add_allocno_copy (ira_allocno_t, ira_allocno_t, int,
885 bool, rtx, ira_loop_tree_node_t);
887 extern int *ira_allocate_cost_vector (enum reg_class);
888 extern void ira_free_cost_vector (int *, enum reg_class);
890 extern void ira_flattening (int, int);
891 extern bool ira_build (bool);
892 extern void ira_destroy (void);
895 extern void ira_init_costs_once (void);
896 extern void ira_init_costs (void);
897 extern void ira_finish_costs_once (void);
898 extern void ira_costs (void);
899 extern void ira_tune_allocno_costs_and_cover_classes (void);
903 extern void ira_rebuild_start_finish_chains (void);
904 extern void ira_print_live_range_list (FILE *, allocno_live_range_t);
905 extern void ira_debug_live_range_list (allocno_live_range_t);
906 extern void ira_debug_allocno_live_ranges (ira_allocno_t);
907 extern void ira_debug_live_ranges (void);
908 extern void ira_create_allocno_live_ranges (void);
909 extern void ira_compress_allocno_live_ranges (void);
910 extern void ira_finish_allocno_live_ranges (void);
912 /* ira-conflicts.c */
913 extern void ira_debug_conflicts (bool);
914 extern void ira_build_conflicts (void);
917 extern int ira_loop_edge_freq (ira_loop_tree_node_t, int, bool);
918 extern void ira_reassign_conflict_allocnos (int);
919 extern void ira_initiate_assign (void);
920 extern void ira_finish_assign (void);
921 extern void ira_color (void);
924 extern void ira_emit (bool);
928 /* The iterator for all allocnos. */
930 /* The number of the current element in IRA_ALLOCNOS. */
932 } ira_allocno_iterator;
934 /* Initialize the iterator I. */
936 ira_allocno_iter_init (ira_allocno_iterator *i)
941 /* Return TRUE if we have more allocnos to visit, in which case *A is
942 set to the allocno to be visited. Otherwise, return FALSE. */
944 ira_allocno_iter_cond (ira_allocno_iterator *i, ira_allocno_t *a)
948 for (n = i->n; n < ira_allocnos_num; n++)
949 if (ira_allocnos[n] != NULL)
951 *a = ira_allocnos[n];
958 /* Loop over all allocnos. In each iteration, A is set to the next
959 allocno. ITER is an instance of ira_allocno_iterator used to iterate
961 #define FOR_EACH_ALLOCNO(A, ITER) \
962 for (ira_allocno_iter_init (&(ITER)); \
963 ira_allocno_iter_cond (&(ITER), &(A));)
968 /* The iterator for copies. */
970 /* The number of the current element in IRA_COPIES. */
974 /* Initialize the iterator I. */
976 ira_copy_iter_init (ira_copy_iterator *i)
981 /* Return TRUE if we have more copies to visit, in which case *CP is
982 set to the copy to be visited. Otherwise, return FALSE. */
984 ira_copy_iter_cond (ira_copy_iterator *i, ira_copy_t *cp)
988 for (n = i->n; n < ira_copies_num; n++)
989 if (ira_copies[n] != NULL)
998 /* Loop over all copies. In each iteration, C is set to the next
999 copy. ITER is an instance of ira_copy_iterator used to iterate
1001 #define FOR_EACH_COPY(C, ITER) \
1002 for (ira_copy_iter_init (&(ITER)); \
1003 ira_copy_iter_cond (&(ITER), &(C));)
1008 /* The iterator for allocno conflicts. */
1011 /* TRUE if the conflicts are represented by vector of allocnos. */
1012 bool allocno_conflict_vec_p;
1014 /* The conflict vector or conflict bit vector. */
1017 /* The number of the current element in the vector (of type
1018 ira_allocno_t or IRA_INT_TYPE). */
1019 unsigned int word_num;
1021 /* The bit vector size. It is defined only if
1022 ALLOCNO_CONFLICT_VEC_P is FALSE. */
1025 /* The current bit index of bit vector. It is defined only if
1026 ALLOCNO_CONFLICT_VEC_P is FALSE. */
1027 unsigned int bit_num;
1029 /* Allocno conflict id corresponding to the 1st bit of the bit
1030 vector. It is defined only if ALLOCNO_CONFLICT_VEC_P is
1032 int base_conflict_id;
1034 /* The word of bit vector currently visited. It is defined only if
1035 ALLOCNO_CONFLICT_VEC_P is FALSE. */
1036 unsigned IRA_INT_TYPE word;
1037 } ira_allocno_conflict_iterator;
1039 /* Initialize the iterator I with ALLOCNO conflicts. */
1041 ira_allocno_conflict_iter_init (ira_allocno_conflict_iterator *i,
1042 ira_allocno_t allocno)
1044 i->allocno_conflict_vec_p = ALLOCNO_CONFLICT_VEC_P (allocno);
1045 i->vec = ALLOCNO_CONFLICT_ALLOCNO_ARRAY (allocno);
1047 if (i->allocno_conflict_vec_p)
1048 i->size = i->bit_num = i->base_conflict_id = i->word = 0;
1051 if (ALLOCNO_MIN (allocno) > ALLOCNO_MAX (allocno))
1054 i->size = ((ALLOCNO_MAX (allocno) - ALLOCNO_MIN (allocno)
1056 / IRA_INT_BITS) * sizeof (IRA_INT_TYPE);
1058 i->base_conflict_id = ALLOCNO_MIN (allocno);
1059 i->word = (i->size == 0 ? 0 : ((IRA_INT_TYPE *) i->vec)[0]);
1063 /* Return TRUE if we have more conflicting allocnos to visit, in which
1064 case *A is set to the allocno to be visited. Otherwise, return
1067 ira_allocno_conflict_iter_cond (ira_allocno_conflict_iterator *i,
1070 ira_allocno_t conflict_allocno;
1072 if (i->allocno_conflict_vec_p)
1074 conflict_allocno = ((ira_allocno_t *) i->vec)[i->word_num];
1075 if (conflict_allocno == NULL)
1077 *a = conflict_allocno;
1082 /* Skip words that are zeros. */
1083 for (; i->word == 0; i->word = ((IRA_INT_TYPE *) i->vec)[i->word_num])
1087 /* If we have reached the end, break. */
1088 if (i->word_num * sizeof (IRA_INT_TYPE) >= i->size)
1091 i->bit_num = i->word_num * IRA_INT_BITS;
1094 /* Skip bits that are zero. */
1095 for (; (i->word & 1) == 0; i->word >>= 1)
1098 *a = ira_conflict_id_allocno_map[i->bit_num + i->base_conflict_id];
1104 /* Advance to the next conflicting allocno. */
1106 ira_allocno_conflict_iter_next (ira_allocno_conflict_iterator *i)
1108 if (i->allocno_conflict_vec_p)
1117 /* Loop over all allocnos conflicting with ALLOCNO. In each
1118 iteration, A is set to the next conflicting allocno. ITER is an
1119 instance of ira_allocno_conflict_iterator used to iterate the
1121 #define FOR_EACH_ALLOCNO_CONFLICT(ALLOCNO, A, ITER) \
1122 for (ira_allocno_conflict_iter_init (&(ITER), (ALLOCNO)); \
1123 ira_allocno_conflict_iter_cond (&(ITER), &(A)); \
1124 ira_allocno_conflict_iter_next (&(ITER)))
1128 /* The function returns TRUE if hard registers starting with
1129 HARD_REGNO and containing value of MODE are not in set
1132 ira_hard_reg_not_in_set_p (int hard_regno, enum machine_mode mode,
1133 HARD_REG_SET hard_regset)
1137 ira_assert (hard_regno >= 0);
1138 for (i = hard_regno_nregs[hard_regno][mode] - 1; i >= 0; i--)
1139 if (TEST_HARD_REG_BIT (hard_regset, hard_regno + i))
1146 /* To save memory we use a lazy approach for allocation and
1147 initialization of the cost vectors. We do this only when it is
1148 really necessary. */
1150 /* Allocate cost vector *VEC for hard registers of COVER_CLASS and
1151 initialize the elements by VAL if it is necessary */
1153 ira_allocate_and_set_costs (int **vec, enum reg_class cover_class, int val)
1160 *vec = reg_costs = ira_allocate_cost_vector (cover_class);
1161 len = ira_class_hard_regs_num[cover_class];
1162 for (i = 0; i < len; i++)
1166 /* Allocate cost vector *VEC for hard registers of COVER_CLASS and
1167 copy values of vector SRC into the vector if it is necessary */
1169 ira_allocate_and_copy_costs (int **vec, enum reg_class cover_class, int *src)
1173 if (*vec != NULL || src == NULL)
1175 *vec = ira_allocate_cost_vector (cover_class);
1176 len = ira_class_hard_regs_num[cover_class];
1177 memcpy (*vec, src, sizeof (int) * len);
1180 /* Allocate cost vector *VEC for hard registers of COVER_CLASS and
1181 add values of vector SRC into the vector if it is necessary */
1183 ira_allocate_and_accumulate_costs (int **vec, enum reg_class cover_class,
1190 len = ira_class_hard_regs_num[cover_class];
1193 *vec = ira_allocate_cost_vector (cover_class);
1194 memset (*vec, 0, sizeof (int) * len);
1196 for (i = 0; i < len; i++)
1197 (*vec)[i] += src[i];
1200 /* Allocate cost vector *VEC for hard registers of COVER_CLASS and
1201 copy values of vector SRC into the vector or initialize it by VAL
1202 (if SRC is null). */
1204 ira_allocate_and_set_or_copy_costs (int **vec, enum reg_class cover_class,
1212 *vec = reg_costs = ira_allocate_cost_vector (cover_class);
1213 len = ira_class_hard_regs_num[cover_class];
1215 memcpy (reg_costs, src, sizeof (int) * len);
1218 for (i = 0; i < len; i++)