1 /* The Blackfin code generation auxiliary output file.
2 Copyright (C) 2005, 2006, 2007, 2008 Free Software Foundation, Inc.
3 Contributed by Analog Devices.
5 This file is part of GCC.
7 GCC is free software; you can redistribute it and/or modify it
8 under the terms of the GNU General Public License as published
9 by the Free Software Foundation; either version 3, or (at your
10 option) any later version.
12 GCC is distributed in the hope that it will be useful, but WITHOUT
13 ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
14 or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public
15 License for more details.
17 You should have received a copy of the GNU General Public License
18 along with GCC; see the file COPYING3. If not see
19 <http://www.gnu.org/licenses/>. */
23 #include "coretypes.h"
27 #include "hard-reg-set.h"
29 #include "insn-config.h"
30 #include "insn-codes.h"
31 #include "conditions.h"
32 #include "insn-flags.h"
34 #include "insn-attr.h"
41 #include "target-def.h"
47 #include "integrate.h"
49 #include "langhooks.h"
50 #include "bfin-protos.h"
52 #include "tm-constrs.h"
54 #include "basic-block.h"
55 #include "cfglayout.h"
59 /* A C structure for machine-specific, per-function data.
60 This is added to the cfun structure. */
61 struct machine_function GTY(())
63 int has_hardware_loops;
66 /* Test and compare insns in bfin.md store the information needed to
67 generate branch and scc insns here. */
68 rtx bfin_compare_op0, bfin_compare_op1;
70 /* RTX for condition code flag register and RETS register */
71 extern GTY(()) rtx bfin_cc_rtx;
72 extern GTY(()) rtx bfin_rets_rtx;
73 rtx bfin_cc_rtx, bfin_rets_rtx;
75 int max_arg_registers = 0;
77 /* Arrays used when emitting register names. */
78 const char *short_reg_names[] = SHORT_REGISTER_NAMES;
79 const char *high_reg_names[] = HIGH_REGISTER_NAMES;
80 const char *dregs_pair_names[] = DREGS_PAIR_NAMES;
81 const char *byte_reg_names[] = BYTE_REGISTER_NAMES;
83 static int arg_regs[] = FUNCTION_ARG_REGISTERS;
85 /* Nonzero if -mshared-library-id was given. */
86 static int bfin_lib_id_given;
88 /* Nonzero if -fschedule-insns2 was given. We override it and
89 call the scheduler ourselves during reorg. */
90 static int bfin_flag_schedule_insns2;
92 /* Determines whether we run variable tracking in machine dependent
94 static int bfin_flag_var_tracking;
97 bfin_cpu_t bfin_cpu_type = BFIN_CPU_UNKNOWN;
99 /* -msi-revision support. There are three special values:
100 -1 -msi-revision=none.
101 0xffff -msi-revision=any. */
102 int bfin_si_revision;
104 /* The workarounds enabled */
105 unsigned int bfin_workarounds = 0;
112 unsigned int workarounds;
115 struct bfin_cpu bfin_cpus[] =
117 {"bf522", BFIN_CPU_BF522, 0x0001,
118 WA_SPECULATIVE_LOADS | WA_RETS},
119 {"bf522", BFIN_CPU_BF522, 0x0000,
120 WA_SPECULATIVE_LOADS | WA_RETS},
122 {"bf523", BFIN_CPU_BF523, 0x0001,
123 WA_SPECULATIVE_LOADS | WA_RETS},
124 {"bf523", BFIN_CPU_BF523, 0x0000,
125 WA_SPECULATIVE_LOADS | WA_RETS},
127 {"bf524", BFIN_CPU_BF524, 0x0001,
128 WA_SPECULATIVE_LOADS | WA_RETS},
129 {"bf524", BFIN_CPU_BF524, 0x0000,
130 WA_SPECULATIVE_LOADS | WA_RETS},
132 {"bf525", BFIN_CPU_BF525, 0x0001,
133 WA_SPECULATIVE_LOADS | WA_RETS},
134 {"bf525", BFIN_CPU_BF525, 0x0000,
135 WA_SPECULATIVE_LOADS | WA_RETS},
137 {"bf526", BFIN_CPU_BF526, 0x0001,
138 WA_SPECULATIVE_LOADS | WA_RETS},
139 {"bf526", BFIN_CPU_BF526, 0x0000,
140 WA_SPECULATIVE_LOADS | WA_RETS},
142 {"bf527", BFIN_CPU_BF527, 0x0001,
143 WA_SPECULATIVE_LOADS | WA_RETS},
144 {"bf527", BFIN_CPU_BF527, 0x0000,
145 WA_SPECULATIVE_LOADS | WA_RETS},
147 {"bf531", BFIN_CPU_BF531, 0x0005,
148 WA_SPECULATIVE_LOADS | WA_RETS},
149 {"bf531", BFIN_CPU_BF531, 0x0004,
150 WA_SPECULATIVE_LOADS | WA_SPECULATIVE_SYNCS | WA_RETS},
151 {"bf531", BFIN_CPU_BF531, 0x0003,
152 WA_SPECULATIVE_LOADS | WA_SPECULATIVE_SYNCS | WA_RETS},
154 {"bf532", BFIN_CPU_BF532, 0x0005,
155 WA_SPECULATIVE_LOADS | WA_RETS},
156 {"bf532", BFIN_CPU_BF532, 0x0004,
157 WA_SPECULATIVE_LOADS | WA_SPECULATIVE_SYNCS | WA_RETS},
158 {"bf532", BFIN_CPU_BF532, 0x0003,
159 WA_SPECULATIVE_LOADS | WA_SPECULATIVE_SYNCS | WA_RETS},
161 {"bf533", BFIN_CPU_BF533, 0x0005,
162 WA_SPECULATIVE_LOADS | WA_RETS},
163 {"bf533", BFIN_CPU_BF533, 0x0004,
164 WA_SPECULATIVE_LOADS | WA_SPECULATIVE_SYNCS | WA_RETS},
165 {"bf533", BFIN_CPU_BF533, 0x0003,
166 WA_SPECULATIVE_LOADS | WA_SPECULATIVE_SYNCS | WA_RETS},
168 {"bf534", BFIN_CPU_BF534, 0x0003,
169 WA_SPECULATIVE_LOADS | WA_RETS},
170 {"bf534", BFIN_CPU_BF534, 0x0002,
171 WA_SPECULATIVE_LOADS | WA_SPECULATIVE_SYNCS | WA_RETS},
172 {"bf534", BFIN_CPU_BF534, 0x0001,
173 WA_SPECULATIVE_LOADS | WA_SPECULATIVE_SYNCS | WA_RETS},
175 {"bf536", BFIN_CPU_BF536, 0x0003,
176 WA_SPECULATIVE_LOADS | WA_RETS},
177 {"bf536", BFIN_CPU_BF536, 0x0002,
178 WA_SPECULATIVE_LOADS | WA_SPECULATIVE_SYNCS | WA_RETS},
179 {"bf536", BFIN_CPU_BF536, 0x0001,
180 WA_SPECULATIVE_LOADS | WA_SPECULATIVE_SYNCS | WA_RETS},
182 {"bf537", BFIN_CPU_BF537, 0x0003,
183 WA_SPECULATIVE_LOADS | WA_RETS},
184 {"bf537", BFIN_CPU_BF537, 0x0002,
185 WA_SPECULATIVE_LOADS | WA_SPECULATIVE_SYNCS | WA_RETS},
186 {"bf537", BFIN_CPU_BF537, 0x0001,
187 WA_SPECULATIVE_LOADS | WA_SPECULATIVE_SYNCS | WA_RETS},
189 {"bf538", BFIN_CPU_BF538, 0x0004,
190 WA_SPECULATIVE_LOADS | WA_RETS},
191 {"bf538", BFIN_CPU_BF538, 0x0003,
192 WA_SPECULATIVE_LOADS | WA_RETS},
193 {"bf538", BFIN_CPU_BF538, 0x0002,
194 WA_SPECULATIVE_LOADS | WA_RETS},
196 {"bf539", BFIN_CPU_BF539, 0x0004,
197 WA_SPECULATIVE_LOADS | WA_RETS},
198 {"bf539", BFIN_CPU_BF539, 0x0003,
199 WA_SPECULATIVE_LOADS | WA_RETS},
200 {"bf539", BFIN_CPU_BF539, 0x0002,
201 WA_SPECULATIVE_LOADS | WA_RETS},
203 {"bf542", BFIN_CPU_BF542, 0x0001,
204 WA_SPECULATIVE_LOADS | WA_RETS},
205 {"bf542", BFIN_CPU_BF542, 0x0000,
206 WA_SPECULATIVE_LOADS | WA_RETS},
208 {"bf544", BFIN_CPU_BF544, 0x0001,
209 WA_SPECULATIVE_LOADS | WA_RETS},
210 {"bf544", BFIN_CPU_BF544, 0x0000,
211 WA_SPECULATIVE_LOADS | WA_RETS},
213 {"bf547", BFIN_CPU_BF547, 0x0001,
214 WA_SPECULATIVE_LOADS | WA_RETS},
215 {"bf547", BFIN_CPU_BF547, 0x0000,
216 WA_SPECULATIVE_LOADS | WA_RETS},
218 {"bf548", BFIN_CPU_BF548, 0x0001,
219 WA_SPECULATIVE_LOADS | WA_RETS},
220 {"bf548", BFIN_CPU_BF548, 0x0000,
221 WA_SPECULATIVE_LOADS | WA_RETS},
223 {"bf549", BFIN_CPU_BF549, 0x0001,
224 WA_SPECULATIVE_LOADS | WA_RETS},
225 {"bf549", BFIN_CPU_BF549, 0x0000,
226 WA_SPECULATIVE_LOADS | WA_RETS},
228 {"bf561", BFIN_CPU_BF561, 0x0005, WA_RETS},
229 {"bf561", BFIN_CPU_BF561, 0x0003,
230 WA_SPECULATIVE_LOADS | WA_SPECULATIVE_SYNCS | WA_RETS},
231 {"bf561", BFIN_CPU_BF561, 0x0002,
232 WA_SPECULATIVE_LOADS | WA_SPECULATIVE_SYNCS | WA_RETS},
237 int splitting_for_sched;
240 bfin_globalize_label (FILE *stream, const char *name)
242 fputs (".global ", stream);
243 assemble_name (stream, name);
249 output_file_start (void)
251 FILE *file = asm_out_file;
254 /* Variable tracking should be run after all optimizations which change order
255 of insns. It also needs a valid CFG. This can't be done in
256 override_options, because flag_var_tracking is finalized after
258 bfin_flag_var_tracking = flag_var_tracking;
259 flag_var_tracking = 0;
261 fprintf (file, ".file \"%s\";\n", input_filename);
263 for (i = 0; arg_regs[i] >= 0; i++)
265 max_arg_registers = i; /* how many arg reg used */
268 /* Called early in the compilation to conditionally modify
269 fixed_regs/call_used_regs. */
272 conditional_register_usage (void)
274 /* initialize condition code flag register rtx */
275 bfin_cc_rtx = gen_rtx_REG (BImode, REG_CC);
276 bfin_rets_rtx = gen_rtx_REG (Pmode, REG_RETS);
279 /* Examine machine-dependent attributes of function type FUNTYPE and return its
280 type. See the definition of E_FUNKIND. */
283 funkind (const_tree funtype)
285 tree attrs = TYPE_ATTRIBUTES (funtype);
286 if (lookup_attribute ("interrupt_handler", attrs))
287 return INTERRUPT_HANDLER;
288 else if (lookup_attribute ("exception_handler", attrs))
289 return EXCPT_HANDLER;
290 else if (lookup_attribute ("nmi_handler", attrs))
296 /* Legitimize PIC addresses. If the address is already position-independent,
297 we return ORIG. Newly generated position-independent addresses go into a
298 reg. This is REG if nonzero, otherwise we allocate register(s) as
299 necessary. PICREG is the register holding the pointer to the PIC offset
303 legitimize_pic_address (rtx orig, rtx reg, rtx picreg)
308 if (GET_CODE (addr) == SYMBOL_REF || GET_CODE (addr) == LABEL_REF)
313 if (TARGET_ID_SHARED_LIBRARY)
314 unspec = UNSPEC_MOVE_PIC;
315 else if (GET_CODE (addr) == SYMBOL_REF
316 && SYMBOL_REF_FUNCTION_P (addr))
317 unspec = UNSPEC_FUNCDESC_GOT17M4;
319 unspec = UNSPEC_MOVE_FDPIC;
323 gcc_assert (can_create_pseudo_p ());
324 reg = gen_reg_rtx (Pmode);
327 tmp = gen_rtx_UNSPEC (Pmode, gen_rtvec (1, addr), unspec);
328 new_rtx = gen_const_mem (Pmode, gen_rtx_PLUS (Pmode, picreg, tmp));
330 emit_move_insn (reg, new_rtx);
331 if (picreg == pic_offset_table_rtx)
332 crtl->uses_pic_offset_table = 1;
336 else if (GET_CODE (addr) == CONST || GET_CODE (addr) == PLUS)
340 if (GET_CODE (addr) == CONST)
342 addr = XEXP (addr, 0);
343 gcc_assert (GET_CODE (addr) == PLUS);
346 if (XEXP (addr, 0) == picreg)
351 gcc_assert (can_create_pseudo_p ());
352 reg = gen_reg_rtx (Pmode);
355 base = legitimize_pic_address (XEXP (addr, 0), reg, picreg);
356 addr = legitimize_pic_address (XEXP (addr, 1),
357 base == reg ? NULL_RTX : reg,
360 if (GET_CODE (addr) == CONST_INT)
362 gcc_assert (! reload_in_progress && ! reload_completed);
363 addr = force_reg (Pmode, addr);
366 if (GET_CODE (addr) == PLUS && CONSTANT_P (XEXP (addr, 1)))
368 base = gen_rtx_PLUS (Pmode, base, XEXP (addr, 0));
369 addr = XEXP (addr, 1);
372 return gen_rtx_PLUS (Pmode, base, addr);
378 /* Stack frame layout. */
380 /* For a given REGNO, determine whether it must be saved in the function
381 prologue. IS_INTHANDLER specifies whether we're generating a normal
382 prologue or an interrupt/exception one. */
384 must_save_p (bool is_inthandler, unsigned regno)
386 if (D_REGNO_P (regno))
388 bool is_eh_return_reg = false;
389 if (crtl->calls_eh_return)
394 unsigned test = EH_RETURN_DATA_REGNO (j);
395 if (test == INVALID_REGNUM)
398 is_eh_return_reg = true;
402 return (is_eh_return_reg
403 || (df_regs_ever_live_p (regno)
404 && !fixed_regs[regno]
405 && (is_inthandler || !call_used_regs[regno])));
407 else if (P_REGNO_P (regno))
409 return ((df_regs_ever_live_p (regno)
410 && !fixed_regs[regno]
411 && (is_inthandler || !call_used_regs[regno]))
413 && regno == PIC_OFFSET_TABLE_REGNUM
414 && (crtl->uses_pic_offset_table
415 || (TARGET_ID_SHARED_LIBRARY && !current_function_is_leaf))));
418 return ((is_inthandler || !call_used_regs[regno])
419 && (df_regs_ever_live_p (regno)
420 || (!leaf_function_p () && call_used_regs[regno])));
424 /* Compute the number of DREGS to save with a push_multiple operation.
425 This could include registers that aren't modified in the function,
426 since push_multiple only takes a range of registers.
427 If IS_INTHANDLER, then everything that is live must be saved, even
428 if normally call-clobbered.
429 If CONSECUTIVE, return the number of registers we can save in one
430 instruction with a push/pop multiple instruction. */
433 n_dregs_to_save (bool is_inthandler, bool consecutive)
438 for (i = REG_R7 + 1; i-- != REG_R0;)
440 if (must_save_p (is_inthandler, i))
442 else if (consecutive)
448 /* Like n_dregs_to_save, but compute number of PREGS to save. */
451 n_pregs_to_save (bool is_inthandler, bool consecutive)
456 for (i = REG_P5 + 1; i-- != REG_P0;)
457 if (must_save_p (is_inthandler, i))
459 else if (consecutive)
464 /* Determine if we are going to save the frame pointer in the prologue. */
467 must_save_fp_p (void)
469 return frame_pointer_needed || df_regs_ever_live_p (REG_FP);
473 stack_frame_needed_p (void)
475 /* EH return puts a new return address into the frame using an
476 address relative to the frame pointer. */
477 if (crtl->calls_eh_return)
479 return frame_pointer_needed;
482 /* Emit code to save registers in the prologue. SAVEALL is nonzero if we
483 must save all registers; this is used for interrupt handlers.
484 SPREG contains (reg:SI REG_SP). IS_INTHANDLER is true if we're doing
485 this for an interrupt (or exception) handler. */
488 expand_prologue_reg_save (rtx spreg, int saveall, bool is_inthandler)
490 rtx predec1 = gen_rtx_PRE_DEC (SImode, spreg);
491 rtx predec = gen_rtx_MEM (SImode, predec1);
492 int ndregs = saveall ? 8 : n_dregs_to_save (is_inthandler, false);
493 int npregs = saveall ? 6 : n_pregs_to_save (is_inthandler, false);
494 int ndregs_consec = saveall ? 8 : n_dregs_to_save (is_inthandler, true);
495 int npregs_consec = saveall ? 6 : n_pregs_to_save (is_inthandler, true);
497 int total_consec = ndregs_consec + npregs_consec;
500 if (saveall || is_inthandler)
502 rtx insn = emit_move_insn (predec, gen_rtx_REG (SImode, REG_ASTAT));
503 RTX_FRAME_RELATED_P (insn) = 1;
506 if (total_consec != 0)
509 rtx val = GEN_INT (-total_consec * 4);
510 rtx pat = gen_rtx_PARALLEL (VOIDmode, rtvec_alloc (total_consec + 2));
512 XVECEXP (pat, 0, 0) = gen_rtx_UNSPEC (VOIDmode, gen_rtvec (1, val),
513 UNSPEC_PUSH_MULTIPLE);
514 XVECEXP (pat, 0, total_consec + 1) = gen_rtx_SET (VOIDmode, spreg,
518 RTX_FRAME_RELATED_P (XVECEXP (pat, 0, total_consec + 1)) = 1;
519 d_to_save = ndregs_consec;
520 dregno = REG_R7 + 1 - ndregs_consec;
521 pregno = REG_P5 + 1 - npregs_consec;
522 for (i = 0; i < total_consec; i++)
524 rtx memref = gen_rtx_MEM (word_mode,
525 gen_rtx_PLUS (Pmode, spreg,
526 GEN_INT (- i * 4 - 4)));
530 subpat = gen_rtx_SET (VOIDmode, memref, gen_rtx_REG (word_mode,
536 subpat = gen_rtx_SET (VOIDmode, memref, gen_rtx_REG (word_mode,
539 XVECEXP (pat, 0, i + 1) = subpat;
540 RTX_FRAME_RELATED_P (subpat) = 1;
542 insn = emit_insn (pat);
543 RTX_FRAME_RELATED_P (insn) = 1;
546 for (dregno = REG_R0; ndregs != ndregs_consec; dregno++)
548 if (must_save_p (is_inthandler, dregno))
550 rtx insn = emit_move_insn (predec, gen_rtx_REG (word_mode, dregno));
551 RTX_FRAME_RELATED_P (insn) = 1;
555 for (pregno = REG_P0; npregs != npregs_consec; pregno++)
557 if (must_save_p (is_inthandler, pregno))
559 rtx insn = emit_move_insn (predec, gen_rtx_REG (word_mode, pregno));
560 RTX_FRAME_RELATED_P (insn) = 1;
564 for (i = REG_P7 + 1; i < REG_CC; i++)
567 && (df_regs_ever_live_p (i)
568 || (!leaf_function_p () && call_used_regs[i]))))
571 if (i == REG_A0 || i == REG_A1)
572 insn = emit_move_insn (gen_rtx_MEM (PDImode, predec1),
573 gen_rtx_REG (PDImode, i));
575 insn = emit_move_insn (predec, gen_rtx_REG (SImode, i));
576 RTX_FRAME_RELATED_P (insn) = 1;
580 /* Emit code to restore registers in the epilogue. SAVEALL is nonzero if we
581 must save all registers; this is used for interrupt handlers.
582 SPREG contains (reg:SI REG_SP). IS_INTHANDLER is true if we're doing
583 this for an interrupt (or exception) handler. */
586 expand_epilogue_reg_restore (rtx spreg, bool saveall, bool is_inthandler)
588 rtx postinc1 = gen_rtx_POST_INC (SImode, spreg);
589 rtx postinc = gen_rtx_MEM (SImode, postinc1);
591 int ndregs = saveall ? 8 : n_dregs_to_save (is_inthandler, false);
592 int npregs = saveall ? 6 : n_pregs_to_save (is_inthandler, false);
593 int ndregs_consec = saveall ? 8 : n_dregs_to_save (is_inthandler, true);
594 int npregs_consec = saveall ? 6 : n_pregs_to_save (is_inthandler, true);
595 int total_consec = ndregs_consec + npregs_consec;
599 /* A slightly crude technique to stop flow from trying to delete "dead"
601 MEM_VOLATILE_P (postinc) = 1;
603 for (i = REG_CC - 1; i > REG_P7; i--)
606 && (df_regs_ever_live_p (i)
607 || (!leaf_function_p () && call_used_regs[i]))))
609 if (i == REG_A0 || i == REG_A1)
611 rtx mem = gen_rtx_MEM (PDImode, postinc1);
612 MEM_VOLATILE_P (mem) = 1;
613 emit_move_insn (gen_rtx_REG (PDImode, i), mem);
616 emit_move_insn (gen_rtx_REG (SImode, i), postinc);
619 regno = REG_P5 - npregs_consec;
620 for (; npregs != npregs_consec; regno--)
622 if (must_save_p (is_inthandler, regno))
624 emit_move_insn (gen_rtx_REG (word_mode, regno), postinc);
628 regno = REG_R7 - ndregs_consec;
629 for (; ndregs != ndregs_consec; regno--)
631 if (must_save_p (is_inthandler, regno))
633 emit_move_insn (gen_rtx_REG (word_mode, regno), postinc);
638 if (total_consec != 0)
640 rtx pat = gen_rtx_PARALLEL (VOIDmode, rtvec_alloc (total_consec + 1));
642 = gen_rtx_SET (VOIDmode, spreg,
643 gen_rtx_PLUS (Pmode, spreg,
644 GEN_INT (total_consec * 4)));
646 if (npregs_consec > 0)
651 for (i = 0; i < total_consec; i++)
654 ? gen_rtx_PLUS (Pmode, spreg, GEN_INT (i * 4))
656 rtx memref = gen_rtx_MEM (word_mode, addr);
659 XVECEXP (pat, 0, i + 1)
660 = gen_rtx_SET (VOIDmode, gen_rtx_REG (word_mode, regno), memref);
662 if (npregs_consec > 0)
664 if (--npregs_consec == 0)
669 insn = emit_insn (pat);
670 RTX_FRAME_RELATED_P (insn) = 1;
672 if (saveall || is_inthandler)
673 emit_move_insn (gen_rtx_REG (SImode, REG_ASTAT), postinc);
676 /* Perform any needed actions needed for a function that is receiving a
677 variable number of arguments.
681 MODE and TYPE are the mode and type of the current parameter.
683 PRETEND_SIZE is a variable that should be set to the amount of stack
684 that must be pushed by the prolog to pretend that our caller pushed
687 Normally, this macro will push all remaining incoming registers on the
688 stack and set PRETEND_SIZE to the length of the registers pushed.
691 - VDSP C compiler manual (our ABI) says that a variable args function
692 should save the R0, R1 and R2 registers in the stack.
693 - The caller will always leave space on the stack for the
694 arguments that are passed in registers, so we dont have
695 to leave any extra space.
696 - now, the vastart pointer can access all arguments from the stack. */
699 setup_incoming_varargs (CUMULATIVE_ARGS *cum,
700 enum machine_mode mode ATTRIBUTE_UNUSED,
701 tree type ATTRIBUTE_UNUSED, int *pretend_size,
710 /* The move for named arguments will be generated automatically by the
711 compiler. We need to generate the move rtx for the unnamed arguments
712 if they are in the first 3 words. We assume at least 1 named argument
713 exists, so we never generate [ARGP] = R0 here. */
715 for (i = cum->words + 1; i < max_arg_registers; i++)
717 mem = gen_rtx_MEM (Pmode,
718 plus_constant (arg_pointer_rtx, (i * UNITS_PER_WORD)));
719 emit_move_insn (mem, gen_rtx_REG (Pmode, i));
725 /* Value should be nonzero if functions must have frame pointers.
726 Zero means the frame pointer need not be set up (and parms may
727 be accessed via the stack pointer) in functions that seem suitable. */
730 bfin_frame_pointer_required (void)
732 e_funkind fkind = funkind (TREE_TYPE (current_function_decl));
734 if (fkind != SUBROUTINE)
737 /* We turn on -fomit-frame-pointer if -momit-leaf-frame-pointer is used,
738 so we have to override it for non-leaf functions. */
739 if (TARGET_OMIT_LEAF_FRAME_POINTER && ! current_function_is_leaf)
745 /* Return the number of registers pushed during the prologue. */
748 n_regs_saved_by_prologue (void)
750 e_funkind fkind = funkind (TREE_TYPE (current_function_decl));
751 bool is_inthandler = fkind != SUBROUTINE;
752 tree attrs = TYPE_ATTRIBUTES (TREE_TYPE (current_function_decl));
753 bool all = (lookup_attribute ("saveall", attrs) != NULL_TREE
754 || (is_inthandler && !current_function_is_leaf));
755 int ndregs = all ? 8 : n_dregs_to_save (is_inthandler, false);
756 int npregs = all ? 6 : n_pregs_to_save (is_inthandler, false);
757 int n = ndregs + npregs;
760 if (all || stack_frame_needed_p ())
761 /* We use a LINK instruction in this case. */
765 if (must_save_fp_p ())
767 if (! current_function_is_leaf)
771 if (fkind != SUBROUTINE || all)
772 /* Increment once for ASTAT. */
775 if (fkind != SUBROUTINE)
778 if (lookup_attribute ("nesting", attrs))
782 for (i = REG_P7 + 1; i < REG_CC; i++)
784 || (fkind != SUBROUTINE
785 && (df_regs_ever_live_p (i)
786 || (!leaf_function_p () && call_used_regs[i]))))
787 n += i == REG_A0 || i == REG_A1 ? 2 : 1;
792 /* Return the offset between two registers, one to be eliminated, and the other
793 its replacement, at the start of a routine. */
796 bfin_initial_elimination_offset (int from, int to)
798 HOST_WIDE_INT offset = 0;
800 if (from == ARG_POINTER_REGNUM)
801 offset = n_regs_saved_by_prologue () * 4;
803 if (to == STACK_POINTER_REGNUM)
805 if (crtl->outgoing_args_size >= FIXED_STACK_AREA)
806 offset += crtl->outgoing_args_size;
807 else if (crtl->outgoing_args_size)
808 offset += FIXED_STACK_AREA;
810 offset += get_frame_size ();
816 /* Emit code to load a constant CONSTANT into register REG; setting
817 RTX_FRAME_RELATED_P on all insns we generate if RELATED is true.
818 Make sure that the insns we generate need not be split. */
821 frame_related_constant_load (rtx reg, HOST_WIDE_INT constant, bool related)
824 rtx cst = GEN_INT (constant);
826 if (constant >= -32768 && constant < 65536)
827 insn = emit_move_insn (reg, cst);
830 /* We don't call split_load_immediate here, since dwarf2out.c can get
831 confused about some of the more clever sequences it can generate. */
832 insn = emit_insn (gen_movsi_high (reg, cst));
834 RTX_FRAME_RELATED_P (insn) = 1;
835 insn = emit_insn (gen_movsi_low (reg, reg, cst));
838 RTX_FRAME_RELATED_P (insn) = 1;
841 /* Generate efficient code to add a value to a P register.
842 Set RTX_FRAME_RELATED_P on the generated insns if FRAME is nonzero.
843 EPILOGUE_P is zero if this function is called for prologue,
844 otherwise it's nonzero. And it's less than zero if this is for
848 add_to_reg (rtx reg, HOST_WIDE_INT value, int frame, int epilogue_p)
853 /* Choose whether to use a sequence using a temporary register, or
854 a sequence with multiple adds. We can add a signed 7-bit value
855 in one instruction. */
856 if (value > 120 || value < -120)
864 /* For prologue or normal epilogue, P1 can be safely used
865 as the temporary register. For sibcall epilogue, we try to find
866 a call used P register, which will be restored in epilogue.
867 If we cannot find such a P register, we have to use one I register
871 tmpreg = gen_rtx_REG (SImode, REG_P1);
875 for (i = REG_P0; i <= REG_P5; i++)
876 if ((df_regs_ever_live_p (i) && ! call_used_regs[i])
878 && i == PIC_OFFSET_TABLE_REGNUM
879 && (crtl->uses_pic_offset_table
880 || (TARGET_ID_SHARED_LIBRARY
881 && ! current_function_is_leaf))))
884 tmpreg = gen_rtx_REG (SImode, i);
887 tmpreg = gen_rtx_REG (SImode, REG_P1);
888 tmpreg2 = gen_rtx_REG (SImode, REG_I0);
889 emit_move_insn (tmpreg2, tmpreg);
894 frame_related_constant_load (tmpreg, value, TRUE);
896 insn = emit_move_insn (tmpreg, GEN_INT (value));
898 insn = emit_insn (gen_addsi3 (reg, reg, tmpreg));
900 RTX_FRAME_RELATED_P (insn) = 1;
902 if (tmpreg2 != NULL_RTX)
903 emit_move_insn (tmpreg, tmpreg2);
914 /* We could use -62, but that would leave the stack unaligned, so
918 insn = emit_insn (gen_addsi3 (reg, reg, GEN_INT (size)));
920 RTX_FRAME_RELATED_P (insn) = 1;
926 /* Generate a LINK insn for a frame sized FRAME_SIZE. If this constant
927 is too large, generate a sequence of insns that has the same effect.
928 SPREG contains (reg:SI REG_SP). */
931 emit_link_insn (rtx spreg, HOST_WIDE_INT frame_size)
933 HOST_WIDE_INT link_size = frame_size;
937 if (link_size > 262140)
940 /* Use a LINK insn with as big a constant as possible, then subtract
941 any remaining size from the SP. */
942 insn = emit_insn (gen_link (GEN_INT (-8 - link_size)));
943 RTX_FRAME_RELATED_P (insn) = 1;
945 for (i = 0; i < XVECLEN (PATTERN (insn), 0); i++)
947 rtx set = XVECEXP (PATTERN (insn), 0, i);
948 gcc_assert (GET_CODE (set) == SET);
949 RTX_FRAME_RELATED_P (set) = 1;
952 frame_size -= link_size;
956 /* Must use a call-clobbered PREG that isn't the static chain. */
957 rtx tmpreg = gen_rtx_REG (Pmode, REG_P1);
959 frame_related_constant_load (tmpreg, -frame_size, TRUE);
960 insn = emit_insn (gen_addsi3 (spreg, spreg, tmpreg));
961 RTX_FRAME_RELATED_P (insn) = 1;
965 /* Return the number of bytes we must reserve for outgoing arguments
966 in the current function's stack frame. */
971 if (crtl->outgoing_args_size)
973 if (crtl->outgoing_args_size >= FIXED_STACK_AREA)
974 return crtl->outgoing_args_size;
976 return FIXED_STACK_AREA;
981 /* Save RETS and FP, and allocate a stack frame. ALL is true if the
982 function must save all its registers (true only for certain interrupt
986 do_link (rtx spreg, HOST_WIDE_INT frame_size, bool all)
988 frame_size += arg_area_size ();
990 if (all || stack_frame_needed_p ()
991 || (must_save_fp_p () && ! current_function_is_leaf))
992 emit_link_insn (spreg, frame_size);
995 if (! current_function_is_leaf)
997 rtx pat = gen_movsi (gen_rtx_MEM (Pmode,
998 gen_rtx_PRE_DEC (Pmode, spreg)),
1000 rtx insn = emit_insn (pat);
1001 RTX_FRAME_RELATED_P (insn) = 1;
1003 if (must_save_fp_p ())
1005 rtx pat = gen_movsi (gen_rtx_MEM (Pmode,
1006 gen_rtx_PRE_DEC (Pmode, spreg)),
1007 gen_rtx_REG (Pmode, REG_FP));
1008 rtx insn = emit_insn (pat);
1009 RTX_FRAME_RELATED_P (insn) = 1;
1011 add_to_reg (spreg, -frame_size, 1, 0);
1015 /* Like do_link, but used for epilogues to deallocate the stack frame.
1016 EPILOGUE_P is zero if this function is called for prologue,
1017 otherwise it's nonzero. And it's less than zero if this is for
1018 sibcall epilogue. */
1021 do_unlink (rtx spreg, HOST_WIDE_INT frame_size, bool all, int epilogue_p)
1023 frame_size += arg_area_size ();
1025 if (all || stack_frame_needed_p ())
1026 emit_insn (gen_unlink ());
1029 rtx postinc = gen_rtx_MEM (Pmode, gen_rtx_POST_INC (Pmode, spreg));
1031 add_to_reg (spreg, frame_size, 0, epilogue_p);
1032 if (must_save_fp_p ())
1034 rtx fpreg = gen_rtx_REG (Pmode, REG_FP);
1035 emit_move_insn (fpreg, postinc);
1038 if (! current_function_is_leaf)
1040 emit_move_insn (bfin_rets_rtx, postinc);
1041 emit_use (bfin_rets_rtx);
1046 /* Generate a prologue suitable for a function of kind FKIND. This is
1047 called for interrupt and exception handler prologues.
1048 SPREG contains (reg:SI REG_SP). */
1051 expand_interrupt_handler_prologue (rtx spreg, e_funkind fkind, bool all)
1053 HOST_WIDE_INT frame_size = get_frame_size ();
1054 rtx predec1 = gen_rtx_PRE_DEC (SImode, spreg);
1055 rtx predec = gen_rtx_MEM (SImode, predec1);
1057 tree attrs = TYPE_ATTRIBUTES (TREE_TYPE (current_function_decl));
1058 tree kspisusp = lookup_attribute ("kspisusp", attrs);
1062 insn = emit_move_insn (spreg, gen_rtx_REG (Pmode, REG_USP));
1063 RTX_FRAME_RELATED_P (insn) = 1;
1066 /* We need space on the stack in case we need to save the argument
1068 if (fkind == EXCPT_HANDLER)
1070 insn = emit_insn (gen_addsi3 (spreg, spreg, GEN_INT (-12)));
1071 RTX_FRAME_RELATED_P (insn) = 1;
1074 /* If we're calling other functions, they won't save their call-clobbered
1075 registers, so we must save everything here. */
1076 if (!current_function_is_leaf)
1078 expand_prologue_reg_save (spreg, all, true);
1080 if (lookup_attribute ("nesting", attrs))
1082 rtx srcreg = gen_rtx_REG (Pmode, (fkind == EXCPT_HANDLER ? REG_RETX
1083 : fkind == NMI_HANDLER ? REG_RETN
1085 insn = emit_move_insn (predec, srcreg);
1086 RTX_FRAME_RELATED_P (insn) = 1;
1089 do_link (spreg, frame_size, all);
1091 if (fkind == EXCPT_HANDLER)
1093 rtx r0reg = gen_rtx_REG (SImode, REG_R0);
1094 rtx r1reg = gen_rtx_REG (SImode, REG_R1);
1095 rtx r2reg = gen_rtx_REG (SImode, REG_R2);
1098 insn = emit_move_insn (r0reg, gen_rtx_REG (SImode, REG_SEQSTAT));
1099 insn = emit_insn (gen_ashrsi3 (r0reg, r0reg, GEN_INT (26)));
1100 insn = emit_insn (gen_ashlsi3 (r0reg, r0reg, GEN_INT (26)));
1101 insn = emit_move_insn (r1reg, spreg);
1102 insn = emit_move_insn (r2reg, gen_rtx_REG (Pmode, REG_FP));
1103 insn = emit_insn (gen_addsi3 (r2reg, r2reg, GEN_INT (8)));
1107 /* Generate an epilogue suitable for a function of kind FKIND. This is
1108 called for interrupt and exception handler epilogues.
1109 SPREG contains (reg:SI REG_SP). */
1112 expand_interrupt_handler_epilogue (rtx spreg, e_funkind fkind, bool all)
1114 tree attrs = TYPE_ATTRIBUTES (TREE_TYPE (current_function_decl));
1115 rtx postinc1 = gen_rtx_POST_INC (SImode, spreg);
1116 rtx postinc = gen_rtx_MEM (SImode, postinc1);
1118 /* A slightly crude technique to stop flow from trying to delete "dead"
1120 MEM_VOLATILE_P (postinc) = 1;
1122 do_unlink (spreg, get_frame_size (), all, 1);
1124 if (lookup_attribute ("nesting", attrs))
1126 rtx srcreg = gen_rtx_REG (Pmode, (fkind == EXCPT_HANDLER ? REG_RETX
1127 : fkind == NMI_HANDLER ? REG_RETN
1129 emit_move_insn (srcreg, postinc);
1132 /* If we're calling other functions, they won't save their call-clobbered
1133 registers, so we must save (and restore) everything here. */
1134 if (!current_function_is_leaf)
1137 expand_epilogue_reg_restore (spreg, all, true);
1139 /* Deallocate any space we left on the stack in case we needed to save the
1140 argument registers. */
1141 if (fkind == EXCPT_HANDLER)
1142 emit_insn (gen_addsi3 (spreg, spreg, GEN_INT (12)));
1144 emit_jump_insn (gen_return_internal (GEN_INT (fkind)));
1147 /* Used while emitting the prologue to generate code to load the correct value
1148 into the PIC register, which is passed in DEST. */
1151 bfin_load_pic_reg (rtx dest)
1153 struct cgraph_local_info *i = NULL;
1156 i = cgraph_local_info (current_function_decl);
1158 /* Functions local to the translation unit don't need to reload the
1159 pic reg, since the caller always passes a usable one. */
1161 return pic_offset_table_rtx;
1163 if (bfin_lib_id_given)
1164 addr = plus_constant (pic_offset_table_rtx, -4 - bfin_library_id * 4);
1166 addr = gen_rtx_PLUS (Pmode, pic_offset_table_rtx,
1167 gen_rtx_UNSPEC (Pmode, gen_rtvec (1, const0_rtx),
1168 UNSPEC_LIBRARY_OFFSET));
1169 insn = emit_insn (gen_movsi (dest, gen_rtx_MEM (Pmode, addr)));
1173 /* Generate RTL for the prologue of the current function. */
1176 bfin_expand_prologue (void)
1178 HOST_WIDE_INT frame_size = get_frame_size ();
1179 rtx spreg = gen_rtx_REG (Pmode, REG_SP);
1180 e_funkind fkind = funkind (TREE_TYPE (current_function_decl));
1181 rtx pic_reg_loaded = NULL_RTX;
1182 tree attrs = TYPE_ATTRIBUTES (TREE_TYPE (current_function_decl));
1183 bool all = lookup_attribute ("saveall", attrs) != NULL_TREE;
1185 if (fkind != SUBROUTINE)
1187 expand_interrupt_handler_prologue (spreg, fkind, all);
1191 if (crtl->limit_stack
1192 || (TARGET_STACK_CHECK_L1
1193 && !DECL_NO_LIMIT_STACK (current_function_decl)))
1195 HOST_WIDE_INT offset
1196 = bfin_initial_elimination_offset (ARG_POINTER_REGNUM,
1197 STACK_POINTER_REGNUM);
1198 rtx lim = crtl->limit_stack ? stack_limit_rtx : NULL_RTX;
1199 rtx p2reg = gen_rtx_REG (Pmode, REG_P2);
1203 emit_move_insn (p2reg, gen_int_mode (0xFFB00000, SImode));
1204 emit_move_insn (p2reg, gen_rtx_MEM (Pmode, p2reg));
1207 if (GET_CODE (lim) == SYMBOL_REF)
1209 if (TARGET_ID_SHARED_LIBRARY)
1211 rtx p1reg = gen_rtx_REG (Pmode, REG_P1);
1213 pic_reg_loaded = bfin_load_pic_reg (p2reg);
1214 val = legitimize_pic_address (stack_limit_rtx, p1reg,
1216 emit_move_insn (p1reg, val);
1217 frame_related_constant_load (p2reg, offset, FALSE);
1218 emit_insn (gen_addsi3 (p2reg, p2reg, p1reg));
1223 rtx limit = plus_constant (lim, offset);
1224 emit_move_insn (p2reg, limit);
1231 emit_move_insn (p2reg, lim);
1232 add_to_reg (p2reg, offset, 0, 0);
1235 emit_insn (gen_compare_lt (bfin_cc_rtx, spreg, lim));
1236 emit_insn (gen_trapifcc ());
1238 expand_prologue_reg_save (spreg, all, false);
1240 do_link (spreg, frame_size, false);
1242 if (TARGET_ID_SHARED_LIBRARY
1244 && (crtl->uses_pic_offset_table
1245 || !current_function_is_leaf))
1246 bfin_load_pic_reg (pic_offset_table_rtx);
1249 /* Generate RTL for the epilogue of the current function. NEED_RETURN is zero
1250 if this is for a sibcall. EH_RETURN is nonzero if we're expanding an
1251 eh_return pattern. SIBCALL_P is true if this is a sibcall epilogue,
1255 bfin_expand_epilogue (int need_return, int eh_return, bool sibcall_p)
1257 rtx spreg = gen_rtx_REG (Pmode, REG_SP);
1258 e_funkind fkind = funkind (TREE_TYPE (current_function_decl));
1259 int e = sibcall_p ? -1 : 1;
1260 tree attrs = TYPE_ATTRIBUTES (TREE_TYPE (current_function_decl));
1261 bool all = lookup_attribute ("saveall", attrs) != NULL_TREE;
1263 if (fkind != SUBROUTINE)
1265 expand_interrupt_handler_epilogue (spreg, fkind, all);
1269 do_unlink (spreg, get_frame_size (), false, e);
1271 expand_epilogue_reg_restore (spreg, all, false);
1273 /* Omit the return insn if this is for a sibcall. */
1278 emit_insn (gen_addsi3 (spreg, spreg, gen_rtx_REG (Pmode, REG_P2)));
1280 emit_jump_insn (gen_return_internal (GEN_INT (SUBROUTINE)));
1283 /* Return nonzero if register OLD_REG can be renamed to register NEW_REG. */
1286 bfin_hard_regno_rename_ok (unsigned int old_reg ATTRIBUTE_UNUSED,
1287 unsigned int new_reg)
1289 /* Interrupt functions can only use registers that have already been
1290 saved by the prologue, even if they would normally be
1293 if (funkind (TREE_TYPE (current_function_decl)) != SUBROUTINE
1294 && !df_regs_ever_live_p (new_reg))
1300 /* Return the value of the return address for the frame COUNT steps up
1301 from the current frame, after the prologue.
1302 We punt for everything but the current frame by returning const0_rtx. */
1305 bfin_return_addr_rtx (int count)
1310 return get_hard_reg_initial_val (Pmode, REG_RETS);
1313 /* Try machine-dependent ways of modifying an illegitimate address X
1314 to be legitimate. If we find one, return the new, valid address,
1315 otherwise return NULL_RTX.
1317 OLDX is the address as it was before break_out_memory_refs was called.
1318 In some cases it is useful to look at this to decide what needs to be done.
1320 MODE is the mode of the memory reference. */
1323 legitimize_address (rtx x ATTRIBUTE_UNUSED, rtx oldx ATTRIBUTE_UNUSED,
1324 enum machine_mode mode ATTRIBUTE_UNUSED)
1330 bfin_delegitimize_address (rtx orig_x)
1334 if (GET_CODE (x) != MEM)
1338 if (GET_CODE (x) == PLUS
1339 && GET_CODE (XEXP (x, 1)) == UNSPEC
1340 && XINT (XEXP (x, 1), 1) == UNSPEC_MOVE_PIC
1341 && GET_CODE (XEXP (x, 0)) == REG
1342 && REGNO (XEXP (x, 0)) == PIC_OFFSET_TABLE_REGNUM)
1343 return XVECEXP (XEXP (x, 1), 0, 0);
1348 /* This predicate is used to compute the length of a load/store insn.
1349 OP is a MEM rtx, we return nonzero if its addressing mode requires a
1350 32-bit instruction. */
1353 effective_address_32bit_p (rtx op, enum machine_mode mode)
1355 HOST_WIDE_INT offset;
1357 mode = GET_MODE (op);
1360 if (GET_CODE (op) != PLUS)
1362 gcc_assert (REG_P (op) || GET_CODE (op) == POST_INC
1363 || GET_CODE (op) == PRE_DEC || GET_CODE (op) == POST_DEC);
1367 if (GET_CODE (XEXP (op, 1)) == UNSPEC)
1370 offset = INTVAL (XEXP (op, 1));
1372 /* All byte loads use a 16-bit offset. */
1373 if (GET_MODE_SIZE (mode) == 1)
1376 if (GET_MODE_SIZE (mode) == 4)
1378 /* Frame pointer relative loads can use a negative offset, all others
1379 are restricted to a small positive one. */
1380 if (XEXP (op, 0) == frame_pointer_rtx)
1381 return offset < -128 || offset > 60;
1382 return offset < 0 || offset > 60;
1385 /* Must be HImode now. */
1386 return offset < 0 || offset > 30;
1389 /* Returns true if X is a memory reference using an I register. */
1391 bfin_dsp_memref_p (rtx x)
1396 if (GET_CODE (x) == POST_INC || GET_CODE (x) == PRE_INC
1397 || GET_CODE (x) == POST_DEC || GET_CODE (x) == PRE_DEC)
1402 /* Return cost of the memory address ADDR.
1403 All addressing modes are equally cheap on the Blackfin. */
1406 bfin_address_cost (rtx addr ATTRIBUTE_UNUSED, bool speed ATTRIBUTE_UNUSED)
1411 /* Subroutine of print_operand; used to print a memory reference X to FILE. */
1414 print_address_operand (FILE *file, rtx x)
1416 switch (GET_CODE (x))
1419 output_address (XEXP (x, 0));
1420 fprintf (file, "+");
1421 output_address (XEXP (x, 1));
1425 fprintf (file, "--");
1426 output_address (XEXP (x, 0));
1429 output_address (XEXP (x, 0));
1430 fprintf (file, "++");
1433 output_address (XEXP (x, 0));
1434 fprintf (file, "--");
1438 gcc_assert (GET_CODE (x) != MEM);
1439 print_operand (file, x, 0);
1444 /* Adding intp DImode support by Tony
1450 print_operand (FILE *file, rtx x, char code)
1452 enum machine_mode mode;
1456 if (GET_MODE (current_output_insn) == SImode)
1457 fprintf (file, " ||");
1459 fprintf (file, ";");
1463 mode = GET_MODE (x);
1468 switch (GET_CODE (x))
1471 fprintf (file, "e");
1474 fprintf (file, "ne");
1477 fprintf (file, "g");
1480 fprintf (file, "l");
1483 fprintf (file, "ge");
1486 fprintf (file, "le");
1489 fprintf (file, "g");
1492 fprintf (file, "l");
1495 fprintf (file, "ge");
1498 fprintf (file, "le");
1501 output_operand_lossage ("invalid %%j value");
1505 case 'J': /* reverse logic */
1506 switch (GET_CODE(x))
1509 fprintf (file, "ne");
1512 fprintf (file, "e");
1515 fprintf (file, "le");
1518 fprintf (file, "ge");
1521 fprintf (file, "l");
1524 fprintf (file, "g");
1527 fprintf (file, "le");
1530 fprintf (file, "ge");
1533 fprintf (file, "l");
1536 fprintf (file, "g");
1539 output_operand_lossage ("invalid %%J value");
1544 switch (GET_CODE (x))
1550 fprintf (file, "%s", short_reg_names[REGNO (x)]);
1552 output_operand_lossage ("invalid operand for code '%c'", code);
1554 else if (code == 'd')
1557 fprintf (file, "%s", high_reg_names[REGNO (x)]);
1559 output_operand_lossage ("invalid operand for code '%c'", code);
1561 else if (code == 'w')
1563 if (REGNO (x) == REG_A0 || REGNO (x) == REG_A1)
1564 fprintf (file, "%s.w", reg_names[REGNO (x)]);
1566 output_operand_lossage ("invalid operand for code '%c'", code);
1568 else if (code == 'x')
1570 if (REGNO (x) == REG_A0 || REGNO (x) == REG_A1)
1571 fprintf (file, "%s.x", reg_names[REGNO (x)]);
1573 output_operand_lossage ("invalid operand for code '%c'", code);
1575 else if (code == 'v')
1577 if (REGNO (x) == REG_A0)
1578 fprintf (file, "AV0");
1579 else if (REGNO (x) == REG_A1)
1580 fprintf (file, "AV1");
1582 output_operand_lossage ("invalid operand for code '%c'", code);
1584 else if (code == 'D')
1586 if (D_REGNO_P (REGNO (x)))
1587 fprintf (file, "%s", dregs_pair_names[REGNO (x)]);
1589 output_operand_lossage ("invalid operand for code '%c'", code);
1591 else if (code == 'H')
1593 if ((mode == DImode || mode == DFmode) && REG_P (x))
1594 fprintf (file, "%s", reg_names[REGNO (x) + 1]);
1596 output_operand_lossage ("invalid operand for code '%c'", code);
1598 else if (code == 'T')
1600 if (D_REGNO_P (REGNO (x)))
1601 fprintf (file, "%s", byte_reg_names[REGNO (x)]);
1603 output_operand_lossage ("invalid operand for code '%c'", code);
1606 fprintf (file, "%s", reg_names[REGNO (x)]);
1612 print_address_operand (file, x);
1624 fputs ("(FU)", file);
1627 fputs ("(T)", file);
1630 fputs ("(TFU)", file);
1633 fputs ("(W32)", file);
1636 fputs ("(IS)", file);
1639 fputs ("(IU)", file);
1642 fputs ("(IH)", file);
1645 fputs ("(M)", file);
1648 fputs ("(IS,M)", file);
1651 fputs ("(ISS2)", file);
1654 fputs ("(S2RND)", file);
1661 else if (code == 'b')
1663 if (INTVAL (x) == 0)
1665 else if (INTVAL (x) == 1)
1671 /* Moves to half registers with d or h modifiers always use unsigned
1673 else if (code == 'd')
1674 x = GEN_INT ((INTVAL (x) >> 16) & 0xffff);
1675 else if (code == 'h')
1676 x = GEN_INT (INTVAL (x) & 0xffff);
1677 else if (code == 'N')
1678 x = GEN_INT (-INTVAL (x));
1679 else if (code == 'X')
1680 x = GEN_INT (exact_log2 (0xffffffff & INTVAL (x)));
1681 else if (code == 'Y')
1682 x = GEN_INT (exact_log2 (0xffffffff & ~INTVAL (x)));
1683 else if (code == 'Z')
1684 /* Used for LINK insns. */
1685 x = GEN_INT (-8 - INTVAL (x));
1690 output_addr_const (file, x);
1694 output_operand_lossage ("invalid const_double operand");
1698 switch (XINT (x, 1))
1700 case UNSPEC_MOVE_PIC:
1701 output_addr_const (file, XVECEXP (x, 0, 0));
1702 fprintf (file, "@GOT");
1705 case UNSPEC_MOVE_FDPIC:
1706 output_addr_const (file, XVECEXP (x, 0, 0));
1707 fprintf (file, "@GOT17M4");
1710 case UNSPEC_FUNCDESC_GOT17M4:
1711 output_addr_const (file, XVECEXP (x, 0, 0));
1712 fprintf (file, "@FUNCDESC_GOT17M4");
1715 case UNSPEC_LIBRARY_OFFSET:
1716 fprintf (file, "_current_shared_library_p5_offset_");
1725 output_addr_const (file, x);
1730 /* Argument support functions. */
1732 /* Initialize a variable CUM of type CUMULATIVE_ARGS
1733 for a call to a function whose data type is FNTYPE.
1734 For a library call, FNTYPE is 0.
1735 VDSP C Compiler manual, our ABI says that
1736 first 3 words of arguments will use R0, R1 and R2.
1740 init_cumulative_args (CUMULATIVE_ARGS *cum, tree fntype,
1741 rtx libname ATTRIBUTE_UNUSED)
1743 static CUMULATIVE_ARGS zero_cum;
1747 /* Set up the number of registers to use for passing arguments. */
1749 cum->nregs = max_arg_registers;
1750 cum->arg_regs = arg_regs;
1752 cum->call_cookie = CALL_NORMAL;
1753 /* Check for a longcall attribute. */
1754 if (fntype && lookup_attribute ("shortcall", TYPE_ATTRIBUTES (fntype)))
1755 cum->call_cookie |= CALL_SHORT;
1756 else if (fntype && lookup_attribute ("longcall", TYPE_ATTRIBUTES (fntype)))
1757 cum->call_cookie |= CALL_LONG;
1762 /* Update the data in CUM to advance over an argument
1763 of mode MODE and data type TYPE.
1764 (TYPE is null for libcalls where that information may not be available.) */
1767 function_arg_advance (CUMULATIVE_ARGS *cum, enum machine_mode mode, tree type,
1768 int named ATTRIBUTE_UNUSED)
1770 int count, bytes, words;
1772 bytes = (mode == BLKmode) ? int_size_in_bytes (type) : GET_MODE_SIZE (mode);
1773 words = (bytes + UNITS_PER_WORD - 1) / UNITS_PER_WORD;
1775 cum->words += words;
1776 cum->nregs -= words;
1778 if (cum->nregs <= 0)
1781 cum->arg_regs = NULL;
1785 for (count = 1; count <= words; count++)
1792 /* Define where to put the arguments to a function.
1793 Value is zero to push the argument on the stack,
1794 or a hard register in which to store the argument.
1796 MODE is the argument's machine mode.
1797 TYPE is the data type of the argument (as a tree).
1798 This is null for libcalls where that information may
1800 CUM is a variable of type CUMULATIVE_ARGS which gives info about
1801 the preceding args and about the function being called.
1802 NAMED is nonzero if this argument is a named parameter
1803 (otherwise it is an extra parameter matching an ellipsis). */
1806 function_arg (CUMULATIVE_ARGS *cum, enum machine_mode mode, tree type,
1807 int named ATTRIBUTE_UNUSED)
1810 = (mode == BLKmode) ? int_size_in_bytes (type) : GET_MODE_SIZE (mode);
1812 if (mode == VOIDmode)
1813 /* Compute operand 2 of the call insn. */
1814 return GEN_INT (cum->call_cookie);
1820 return gen_rtx_REG (mode, *(cum->arg_regs));
1825 /* For an arg passed partly in registers and partly in memory,
1826 this is the number of bytes passed in registers.
1827 For args passed entirely in registers or entirely in memory, zero.
1829 Refer VDSP C Compiler manual, our ABI.
1830 First 3 words are in registers. So, if an argument is larger
1831 than the registers available, it will span the register and
1835 bfin_arg_partial_bytes (CUMULATIVE_ARGS *cum, enum machine_mode mode,
1836 tree type ATTRIBUTE_UNUSED,
1837 bool named ATTRIBUTE_UNUSED)
1840 = (mode == BLKmode) ? int_size_in_bytes (type) : GET_MODE_SIZE (mode);
1841 int bytes_left = cum->nregs * UNITS_PER_WORD;
1846 if (bytes_left == 0)
1848 if (bytes > bytes_left)
1853 /* Variable sized types are passed by reference. */
1856 bfin_pass_by_reference (CUMULATIVE_ARGS *cum ATTRIBUTE_UNUSED,
1857 enum machine_mode mode ATTRIBUTE_UNUSED,
1858 const_tree type, bool named ATTRIBUTE_UNUSED)
1860 return type && TREE_CODE (TYPE_SIZE (type)) != INTEGER_CST;
1863 /* Decide whether a type should be returned in memory (true)
1864 or in a register (false). This is called by the macro
1865 TARGET_RETURN_IN_MEMORY. */
1868 bfin_return_in_memory (const_tree type, const_tree fntype ATTRIBUTE_UNUSED)
1870 int size = int_size_in_bytes (type);
1871 return size > 2 * UNITS_PER_WORD || size == -1;
1874 /* Register in which address to store a structure value
1875 is passed to a function. */
1877 bfin_struct_value_rtx (tree fntype ATTRIBUTE_UNUSED,
1878 int incoming ATTRIBUTE_UNUSED)
1880 return gen_rtx_REG (Pmode, REG_P0);
1883 /* Return true when register may be used to pass function parameters. */
1886 function_arg_regno_p (int n)
1889 for (i = 0; arg_regs[i] != -1; i++)
1890 if (n == arg_regs[i])
1895 /* Returns 1 if OP contains a symbol reference */
1898 symbolic_reference_mentioned_p (rtx op)
1900 register const char *fmt;
1903 if (GET_CODE (op) == SYMBOL_REF || GET_CODE (op) == LABEL_REF)
1906 fmt = GET_RTX_FORMAT (GET_CODE (op));
1907 for (i = GET_RTX_LENGTH (GET_CODE (op)) - 1; i >= 0; i--)
1913 for (j = XVECLEN (op, i) - 1; j >= 0; j--)
1914 if (symbolic_reference_mentioned_p (XVECEXP (op, i, j)))
1918 else if (fmt[i] == 'e' && symbolic_reference_mentioned_p (XEXP (op, i)))
1925 /* Decide whether we can make a sibling call to a function. DECL is the
1926 declaration of the function being targeted by the call and EXP is the
1927 CALL_EXPR representing the call. */
1930 bfin_function_ok_for_sibcall (tree decl ATTRIBUTE_UNUSED,
1931 tree exp ATTRIBUTE_UNUSED)
1933 struct cgraph_local_info *this_func, *called_func;
1934 e_funkind fkind = funkind (TREE_TYPE (current_function_decl));
1935 if (fkind != SUBROUTINE)
1937 if (!TARGET_ID_SHARED_LIBRARY || TARGET_SEP_DATA)
1940 /* When compiling for ID shared libraries, can't sibcall a local function
1941 from a non-local function, because the local function thinks it does
1942 not need to reload P5 in the prologue, but the sibcall wil pop P5 in the
1943 sibcall epilogue, and we end up with the wrong value in P5. */
1946 /* Not enough information. */
1949 this_func = cgraph_local_info (current_function_decl);
1950 called_func = cgraph_local_info (decl);
1951 return !called_func->local || this_func->local;
1954 /* Emit RTL insns to initialize the variable parts of a trampoline at
1955 TRAMP. FNADDR is an RTX for the address of the function's pure
1956 code. CXT is an RTX for the static chain value for the function. */
1959 initialize_trampoline (rtx tramp, rtx fnaddr, rtx cxt)
1961 rtx t1 = copy_to_reg (fnaddr);
1962 rtx t2 = copy_to_reg (cxt);
1968 rtx a = memory_address (Pmode, plus_constant (tramp, 8));
1969 addr = memory_address (Pmode, tramp);
1970 emit_move_insn (gen_rtx_MEM (SImode, addr), a);
1974 addr = memory_address (Pmode, plus_constant (tramp, i + 2));
1975 emit_move_insn (gen_rtx_MEM (HImode, addr), gen_lowpart (HImode, t1));
1976 emit_insn (gen_ashrsi3 (t1, t1, GEN_INT (16)));
1977 addr = memory_address (Pmode, plus_constant (tramp, i + 6));
1978 emit_move_insn (gen_rtx_MEM (HImode, addr), gen_lowpart (HImode, t1));
1980 addr = memory_address (Pmode, plus_constant (tramp, i + 10));
1981 emit_move_insn (gen_rtx_MEM (HImode, addr), gen_lowpart (HImode, t2));
1982 emit_insn (gen_ashrsi3 (t2, t2, GEN_INT (16)));
1983 addr = memory_address (Pmode, plus_constant (tramp, i + 14));
1984 emit_move_insn (gen_rtx_MEM (HImode, addr), gen_lowpart (HImode, t2));
1987 /* Emit insns to move operands[1] into operands[0]. */
1990 emit_pic_move (rtx *operands, enum machine_mode mode ATTRIBUTE_UNUSED)
1992 rtx temp = reload_in_progress ? operands[0] : gen_reg_rtx (Pmode);
1994 gcc_assert (!TARGET_FDPIC || !(reload_in_progress || reload_completed));
1995 if (GET_CODE (operands[0]) == MEM && SYMBOLIC_CONST (operands[1]))
1996 operands[1] = force_reg (SImode, operands[1]);
1998 operands[1] = legitimize_pic_address (operands[1], temp,
1999 TARGET_FDPIC ? OUR_FDPIC_REG
2000 : pic_offset_table_rtx);
2003 /* Expand a move operation in mode MODE. The operands are in OPERANDS.
2004 Returns true if no further code must be generated, false if the caller
2005 should generate an insn to move OPERANDS[1] to OPERANDS[0]. */
2008 expand_move (rtx *operands, enum machine_mode mode)
2010 rtx op = operands[1];
2011 if ((TARGET_ID_SHARED_LIBRARY || TARGET_FDPIC)
2012 && SYMBOLIC_CONST (op))
2013 emit_pic_move (operands, mode);
2014 else if (mode == SImode && GET_CODE (op) == CONST
2015 && GET_CODE (XEXP (op, 0)) == PLUS
2016 && GET_CODE (XEXP (XEXP (op, 0), 0)) == SYMBOL_REF
2017 && !bfin_legitimate_constant_p (op))
2019 rtx dest = operands[0];
2021 gcc_assert (!reload_in_progress && !reload_completed);
2023 op0 = force_reg (mode, XEXP (op, 0));
2025 if (!insn_data[CODE_FOR_addsi3].operand[2].predicate (op1, mode))
2026 op1 = force_reg (mode, op1);
2027 if (GET_CODE (dest) == MEM)
2028 dest = gen_reg_rtx (mode);
2029 emit_insn (gen_addsi3 (dest, op0, op1));
2030 if (dest == operands[0])
2034 /* Don't generate memory->memory or constant->memory moves, go through a
2036 else if ((reload_in_progress | reload_completed) == 0
2037 && GET_CODE (operands[0]) == MEM
2038 && GET_CODE (operands[1]) != REG)
2039 operands[1] = force_reg (mode, operands[1]);
2043 /* Split one or more DImode RTL references into pairs of SImode
2044 references. The RTL can be REG, offsettable MEM, integer constant, or
2045 CONST_DOUBLE. "operands" is a pointer to an array of DImode RTL to
2046 split and "num" is its length. lo_half and hi_half are output arrays
2047 that parallel "operands". */
2050 split_di (rtx operands[], int num, rtx lo_half[], rtx hi_half[])
2054 rtx op = operands[num];
2056 /* simplify_subreg refuse to split volatile memory addresses,
2057 but we still have to handle it. */
2058 if (GET_CODE (op) == MEM)
2060 lo_half[num] = adjust_address (op, SImode, 0);
2061 hi_half[num] = adjust_address (op, SImode, 4);
2065 lo_half[num] = simplify_gen_subreg (SImode, op,
2066 GET_MODE (op) == VOIDmode
2067 ? DImode : GET_MODE (op), 0);
2068 hi_half[num] = simplify_gen_subreg (SImode, op,
2069 GET_MODE (op) == VOIDmode
2070 ? DImode : GET_MODE (op), 4);
2076 bfin_longcall_p (rtx op, int call_cookie)
2078 gcc_assert (GET_CODE (op) == SYMBOL_REF);
2079 if (call_cookie & CALL_SHORT)
2081 if (call_cookie & CALL_LONG)
2083 if (TARGET_LONG_CALLS)
2088 /* Expand a call instruction. FNADDR is the call target, RETVAL the return value.
2089 COOKIE is a CONST_INT holding the call_cookie prepared init_cumulative_args.
2090 SIBCALL is nonzero if this is a sibling call. */
2093 bfin_expand_call (rtx retval, rtx fnaddr, rtx callarg1, rtx cookie, int sibcall)
2095 rtx use = NULL, call;
2096 rtx callee = XEXP (fnaddr, 0);
2097 int nelts = 2 + !!sibcall;
2099 rtx picreg = get_hard_reg_initial_val (SImode, FDPIC_REGNO);
2102 /* In an untyped call, we can get NULL for operand 2. */
2103 if (cookie == NULL_RTX)
2104 cookie = const0_rtx;
2106 /* Static functions and indirect calls don't need the pic register. */
2107 if (!TARGET_FDPIC && flag_pic
2108 && GET_CODE (callee) == SYMBOL_REF
2109 && !SYMBOL_REF_LOCAL_P (callee))
2110 use_reg (&use, pic_offset_table_rtx);
2114 int caller_has_l1_text, callee_has_l1_text;
2116 caller_has_l1_text = callee_has_l1_text = 0;
2118 if (lookup_attribute ("l1_text",
2119 DECL_ATTRIBUTES (cfun->decl)) != NULL_TREE)
2120 caller_has_l1_text = 1;
2122 if (GET_CODE (callee) == SYMBOL_REF
2123 && SYMBOL_REF_DECL (callee) && DECL_P (SYMBOL_REF_DECL (callee))
2126 DECL_ATTRIBUTES (SYMBOL_REF_DECL (callee))) != NULL_TREE)
2127 callee_has_l1_text = 1;
2129 if (GET_CODE (callee) != SYMBOL_REF
2130 || bfin_longcall_p (callee, INTVAL (cookie))
2131 || (GET_CODE (callee) == SYMBOL_REF
2132 && !SYMBOL_REF_LOCAL_P (callee)
2133 && TARGET_INLINE_PLT)
2134 || caller_has_l1_text != callee_has_l1_text
2135 || (caller_has_l1_text && callee_has_l1_text
2136 && (GET_CODE (callee) != SYMBOL_REF
2137 || !SYMBOL_REF_LOCAL_P (callee))))
2140 if (! address_operand (addr, Pmode))
2141 addr = force_reg (Pmode, addr);
2143 fnaddr = gen_reg_rtx (SImode);
2144 emit_insn (gen_load_funcdescsi (fnaddr, addr));
2145 fnaddr = gen_rtx_MEM (Pmode, fnaddr);
2147 picreg = gen_reg_rtx (SImode);
2148 emit_insn (gen_load_funcdescsi (picreg,
2149 plus_constant (addr, 4)));
2154 else if ((!register_no_elim_operand (callee, Pmode)
2155 && GET_CODE (callee) != SYMBOL_REF)
2156 || (GET_CODE (callee) == SYMBOL_REF
2157 && ((TARGET_ID_SHARED_LIBRARY && !TARGET_LEAF_ID_SHARED_LIBRARY)
2158 || bfin_longcall_p (callee, INTVAL (cookie)))))
2160 callee = copy_to_mode_reg (Pmode, callee);
2161 fnaddr = gen_rtx_MEM (Pmode, callee);
2163 call = gen_rtx_CALL (VOIDmode, fnaddr, callarg1);
2166 call = gen_rtx_SET (VOIDmode, retval, call);
2168 pat = gen_rtx_PARALLEL (VOIDmode, rtvec_alloc (nelts));
2170 XVECEXP (pat, 0, n++) = call;
2172 XVECEXP (pat, 0, n++) = gen_rtx_USE (VOIDmode, picreg);
2173 XVECEXP (pat, 0, n++) = gen_rtx_USE (VOIDmode, cookie);
2175 XVECEXP (pat, 0, n++) = gen_rtx_RETURN (VOIDmode);
2176 call = emit_call_insn (pat);
2178 CALL_INSN_FUNCTION_USAGE (call) = use;
2181 /* Return 1 if hard register REGNO can hold a value of machine-mode MODE. */
2184 hard_regno_mode_ok (int regno, enum machine_mode mode)
2186 /* Allow only dregs to store value of mode HI or QI */
2187 enum reg_class rclass = REGNO_REG_CLASS (regno);
2192 if (mode == V2HImode)
2193 return D_REGNO_P (regno);
2194 if (rclass == CCREGS)
2195 return mode == BImode;
2196 if (mode == PDImode || mode == V2PDImode)
2197 return regno == REG_A0 || regno == REG_A1;
2199 /* Allow all normal 32-bit regs, except REG_M3, in case regclass ever comes
2200 up with a bad register class (such as ALL_REGS) for DImode. */
2202 return regno < REG_M3;
2205 && TEST_HARD_REG_BIT (reg_class_contents[PROLOGUE_REGS], regno))
2208 return TEST_HARD_REG_BIT (reg_class_contents[MOST_REGS], regno);
2211 /* Implements target hook vector_mode_supported_p. */
2214 bfin_vector_mode_supported_p (enum machine_mode mode)
2216 return mode == V2HImode;
2219 /* Return the cost of moving data from a register in class CLASS1 to
2220 one in class CLASS2. A cost of 2 is the default. */
2223 bfin_register_move_cost (enum machine_mode mode,
2224 enum reg_class class1, enum reg_class class2)
2226 /* These need secondary reloads, so they're more expensive. */
2227 if ((class1 == CCREGS && class2 != DREGS)
2228 || (class1 != DREGS && class2 == CCREGS))
2231 /* If optimizing for size, always prefer reg-reg over reg-memory moves. */
2235 /* There are some stalls involved when moving from a DREG to a different
2236 class reg, and using the value in one of the following instructions.
2237 Attempt to model this by slightly discouraging such moves. */
2238 if (class1 == DREGS && class2 != DREGS)
2241 if (GET_MODE_CLASS (mode) == MODE_INT)
2243 /* Discourage trying to use the accumulators. */
2244 if (TEST_HARD_REG_BIT (reg_class_contents[class1], REG_A0)
2245 || TEST_HARD_REG_BIT (reg_class_contents[class1], REG_A1)
2246 || TEST_HARD_REG_BIT (reg_class_contents[class2], REG_A0)
2247 || TEST_HARD_REG_BIT (reg_class_contents[class2], REG_A1))
2253 /* Return the cost of moving data of mode M between a
2254 register and memory. A value of 2 is the default; this cost is
2255 relative to those in `REGISTER_MOVE_COST'.
2257 ??? In theory L1 memory has single-cycle latency. We should add a switch
2258 that tells the compiler whether we expect to use only L1 memory for the
2259 program; it'll make the costs more accurate. */
2262 bfin_memory_move_cost (enum machine_mode mode ATTRIBUTE_UNUSED,
2263 enum reg_class rclass,
2264 int in ATTRIBUTE_UNUSED)
2266 /* Make memory accesses slightly more expensive than any register-register
2267 move. Also, penalize non-DP registers, since they need secondary
2268 reloads to load and store. */
2269 if (! reg_class_subset_p (rclass, DPREGS))
2275 /* Inform reload about cases where moving X with a mode MODE to a register in
2276 RCLASS requires an extra scratch register. Return the class needed for the
2277 scratch register. */
2279 static enum reg_class
2280 bfin_secondary_reload (bool in_p, rtx x, enum reg_class rclass,
2281 enum machine_mode mode, secondary_reload_info *sri)
2283 /* If we have HImode or QImode, we can only use DREGS as secondary registers;
2284 in most other cases we can also use PREGS. */
2285 enum reg_class default_class = GET_MODE_SIZE (mode) >= 4 ? DPREGS : DREGS;
2286 enum reg_class x_class = NO_REGS;
2287 enum rtx_code code = GET_CODE (x);
2290 x = SUBREG_REG (x), code = GET_CODE (x);
2293 int regno = REGNO (x);
2294 if (regno >= FIRST_PSEUDO_REGISTER)
2295 regno = reg_renumber[regno];
2300 x_class = REGNO_REG_CLASS (regno);
2303 /* We can be asked to reload (plus (FP) (large_constant)) into a DREG.
2304 This happens as a side effect of register elimination, and we need
2305 a scratch register to do it. */
2306 if (fp_plus_const_operand (x, mode))
2308 rtx op2 = XEXP (x, 1);
2309 int large_constant_p = ! satisfies_constraint_Ks7 (op2);
2311 if (rclass == PREGS || rclass == PREGS_CLOBBERED)
2313 /* If destination is a DREG, we can do this without a scratch register
2314 if the constant is valid for an add instruction. */
2315 if ((rclass == DREGS || rclass == DPREGS)
2316 && ! large_constant_p)
2318 /* Reloading to anything other than a DREG? Use a PREG scratch
2320 sri->icode = CODE_FOR_reload_insi;
2324 /* Data can usually be moved freely between registers of most classes.
2325 AREGS are an exception; they can only move to or from another register
2326 in AREGS or one in DREGS. They can also be assigned the constant 0. */
2327 if (x_class == AREGS || x_class == EVEN_AREGS || x_class == ODD_AREGS)
2328 return (rclass == DREGS || rclass == AREGS || rclass == EVEN_AREGS
2329 || rclass == ODD_AREGS
2332 if (rclass == AREGS || rclass == EVEN_AREGS || rclass == ODD_AREGS)
2336 sri->icode = in_p ? CODE_FOR_reload_inpdi : CODE_FOR_reload_outpdi;
2340 if (x != const0_rtx && x_class != DREGS)
2348 /* CCREGS can only be moved from/to DREGS. */
2349 if (rclass == CCREGS && x_class != DREGS)
2351 if (x_class == CCREGS && rclass != DREGS)
2354 /* All registers other than AREGS can load arbitrary constants. The only
2355 case that remains is MEM. */
2357 if (! reg_class_subset_p (rclass, default_class))
2358 return default_class;
2363 /* Implement TARGET_HANDLE_OPTION. */
2366 bfin_handle_option (size_t code, const char *arg, int value)
2370 case OPT_mshared_library_id_:
2371 if (value > MAX_LIBRARY_ID)
2372 error ("-mshared-library-id=%s is not between 0 and %d",
2373 arg, MAX_LIBRARY_ID);
2374 bfin_lib_id_given = 1;
2383 while ((p = bfin_cpus[i].name) != NULL)
2385 if (strncmp (arg, p, strlen (p)) == 0)
2392 error ("-mcpu=%s is not valid", arg);
2396 bfin_cpu_type = bfin_cpus[i].type;
2398 q = arg + strlen (p);
2402 bfin_si_revision = bfin_cpus[i].si_revision;
2403 bfin_workarounds |= bfin_cpus[i].workarounds;
2405 else if (strcmp (q, "-none") == 0)
2406 bfin_si_revision = -1;
2407 else if (strcmp (q, "-any") == 0)
2409 bfin_si_revision = 0xffff;
2410 while (bfin_cpus[i].type == bfin_cpu_type)
2412 bfin_workarounds |= bfin_cpus[i].workarounds;
2418 unsigned int si_major, si_minor;
2421 rev_len = strlen (q);
2423 if (sscanf (q, "-%u.%u%n", &si_major, &si_minor, &n) != 2
2425 || si_major > 0xff || si_minor > 0xff)
2427 invalid_silicon_revision:
2428 error ("-mcpu=%s has invalid silicon revision", arg);
2432 bfin_si_revision = (si_major << 8) | si_minor;
2434 while (bfin_cpus[i].type == bfin_cpu_type
2435 && bfin_cpus[i].si_revision != bfin_si_revision)
2438 if (bfin_cpus[i].type != bfin_cpu_type)
2439 goto invalid_silicon_revision;
2441 bfin_workarounds |= bfin_cpus[i].workarounds;
2452 static struct machine_function *
2453 bfin_init_machine_status (void)
2455 struct machine_function *f;
2457 f = GGC_CNEW (struct machine_function);
2462 /* Implement the macro OVERRIDE_OPTIONS. */
2465 override_options (void)
2467 /* If processor type is not specified, enable all workarounds. */
2468 if (bfin_cpu_type == BFIN_CPU_UNKNOWN)
2472 for (i = 0; bfin_cpus[i].name != NULL; i++)
2473 bfin_workarounds |= bfin_cpus[i].workarounds;
2475 bfin_si_revision = 0xffff;
2478 if (bfin_csync_anomaly == 1)
2479 bfin_workarounds |= WA_SPECULATIVE_SYNCS;
2480 else if (bfin_csync_anomaly == 0)
2481 bfin_workarounds &= ~WA_SPECULATIVE_SYNCS;
2483 if (bfin_specld_anomaly == 1)
2484 bfin_workarounds |= WA_SPECULATIVE_LOADS;
2485 else if (bfin_specld_anomaly == 0)
2486 bfin_workarounds &= ~WA_SPECULATIVE_LOADS;
2488 if (TARGET_OMIT_LEAF_FRAME_POINTER)
2489 flag_omit_frame_pointer = 1;
2491 /* Library identification */
2492 if (bfin_lib_id_given && ! TARGET_ID_SHARED_LIBRARY)
2493 error ("-mshared-library-id= specified without -mid-shared-library");
2495 if (stack_limit_rtx && TARGET_STACK_CHECK_L1)
2496 error ("Can't use multiple stack checking methods together.");
2498 if (TARGET_ID_SHARED_LIBRARY && TARGET_FDPIC)
2499 error ("ID shared libraries and FD-PIC mode can't be used together.");
2501 /* Don't allow the user to specify -mid-shared-library and -msep-data
2502 together, as it makes little sense from a user's point of view... */
2503 if (TARGET_SEP_DATA && TARGET_ID_SHARED_LIBRARY)
2504 error ("cannot specify both -msep-data and -mid-shared-library");
2505 /* ... internally, however, it's nearly the same. */
2506 if (TARGET_SEP_DATA)
2507 target_flags |= MASK_ID_SHARED_LIBRARY | MASK_LEAF_ID_SHARED_LIBRARY;
2509 if (TARGET_ID_SHARED_LIBRARY && flag_pic == 0)
2512 /* There is no single unaligned SI op for PIC code. Sometimes we
2513 need to use ".4byte" and sometimes we need to use ".picptr".
2514 See bfin_assemble_integer for details. */
2516 targetm.asm_out.unaligned_op.si = 0;
2518 /* Silently turn off flag_pic if not doing FDPIC or ID shared libraries,
2519 since we don't support it and it'll just break. */
2520 if (flag_pic && !TARGET_FDPIC && !TARGET_ID_SHARED_LIBRARY)
2523 if (TARGET_MULTICORE && bfin_cpu_type != BFIN_CPU_BF561)
2524 error ("-mmulticore can only be used with BF561");
2526 if (TARGET_COREA && !TARGET_MULTICORE)
2527 error ("-mcorea should be used with -mmulticore");
2529 if (TARGET_COREB && !TARGET_MULTICORE)
2530 error ("-mcoreb should be used with -mmulticore");
2532 if (TARGET_COREA && TARGET_COREB)
2533 error ("-mcorea and -mcoreb can't be used together");
2535 flag_schedule_insns = 0;
2537 /* Passes after sched2 can break the helpful TImode annotations that
2538 haifa-sched puts on every insn. Just do scheduling in reorg. */
2539 bfin_flag_schedule_insns2 = flag_schedule_insns_after_reload;
2540 flag_schedule_insns_after_reload = 0;
2542 init_machine_status = bfin_init_machine_status;
2545 /* Return the destination address of BRANCH.
2546 We need to use this instead of get_attr_length, because the
2547 cbranch_with_nops pattern conservatively sets its length to 6, and
2548 we still prefer to use shorter sequences. */
2551 branch_dest (rtx branch)
2555 rtx pat = PATTERN (branch);
2556 if (GET_CODE (pat) == PARALLEL)
2557 pat = XVECEXP (pat, 0, 0);
2558 dest = SET_SRC (pat);
2559 if (GET_CODE (dest) == IF_THEN_ELSE)
2560 dest = XEXP (dest, 1);
2561 dest = XEXP (dest, 0);
2562 dest_uid = INSN_UID (dest);
2563 return INSN_ADDRESSES (dest_uid);
2566 /* Return nonzero if INSN is annotated with a REG_BR_PROB note that indicates
2567 it's a branch that's predicted taken. */
2570 cbranch_predicted_taken_p (rtx insn)
2572 rtx x = find_reg_note (insn, REG_BR_PROB, 0);
2576 int pred_val = INTVAL (XEXP (x, 0));
2578 return pred_val >= REG_BR_PROB_BASE / 2;
2584 /* Templates for use by asm_conditional_branch. */
2586 static const char *ccbranch_templates[][3] = {
2587 { "if !cc jump %3;", "if cc jump 4 (bp); jump.s %3;", "if cc jump 6 (bp); jump.l %3;" },
2588 { "if cc jump %3;", "if !cc jump 4 (bp); jump.s %3;", "if !cc jump 6 (bp); jump.l %3;" },
2589 { "if !cc jump %3 (bp);", "if cc jump 4; jump.s %3;", "if cc jump 6; jump.l %3;" },
2590 { "if cc jump %3 (bp);", "if !cc jump 4; jump.s %3;", "if !cc jump 6; jump.l %3;" },
2593 /* Output INSN, which is a conditional branch instruction with operands
2596 We deal with the various forms of conditional branches that can be generated
2597 by bfin_reorg to prevent the hardware from doing speculative loads, by
2598 - emitting a sufficient number of nops, if N_NOPS is nonzero, or
2599 - always emitting the branch as predicted taken, if PREDICT_TAKEN is true.
2600 Either of these is only necessary if the branch is short, otherwise the
2601 template we use ends in an unconditional jump which flushes the pipeline
2605 asm_conditional_branch (rtx insn, rtx *operands, int n_nops, int predict_taken)
2607 int offset = branch_dest (insn) - INSN_ADDRESSES (INSN_UID (insn));
2608 /* Note : offset for instructions like if cc jmp; jump.[sl] offset
2609 is to be taken from start of if cc rather than jump.
2610 Range for jump.s is (-4094, 4096) instead of (-4096, 4094)
2612 int len = (offset >= -1024 && offset <= 1022 ? 0
2613 : offset >= -4094 && offset <= 4096 ? 1
2615 int bp = predict_taken && len == 0 ? 1 : cbranch_predicted_taken_p (insn);
2616 int idx = (bp << 1) | (GET_CODE (operands[0]) == EQ ? BRF : BRT);
2617 output_asm_insn (ccbranch_templates[idx][len], operands);
2618 gcc_assert (n_nops == 0 || !bp);
2620 while (n_nops-- > 0)
2621 output_asm_insn ("nop;", NULL);
2624 /* Emit rtl for a comparison operation CMP in mode MODE. Operands have been
2625 stored in bfin_compare_op0 and bfin_compare_op1 already. */
2628 bfin_gen_compare (rtx cmp, enum machine_mode mode ATTRIBUTE_UNUSED)
2630 enum rtx_code code1, code2;
2631 rtx op0 = bfin_compare_op0, op1 = bfin_compare_op1;
2632 rtx tem = bfin_cc_rtx;
2633 enum rtx_code code = GET_CODE (cmp);
2635 /* If we have a BImode input, then we already have a compare result, and
2636 do not need to emit another comparison. */
2637 if (GET_MODE (op0) == BImode)
2639 gcc_assert ((code == NE || code == EQ) && op1 == const0_rtx);
2640 tem = op0, code2 = code;
2645 /* bfin has these conditions */
2655 code1 = reverse_condition (code);
2659 emit_insn (gen_rtx_SET (BImode, tem,
2660 gen_rtx_fmt_ee (code1, BImode, op0, op1)));
2663 return gen_rtx_fmt_ee (code2, BImode, tem, CONST0_RTX (BImode));
2666 /* Return nonzero iff C has exactly one bit set if it is interpreted
2667 as a 32-bit constant. */
2670 log2constp (unsigned HOST_WIDE_INT c)
2673 return c != 0 && (c & (c-1)) == 0;
2676 /* Returns the number of consecutive least significant zeros in the binary
2677 representation of *V.
2678 We modify *V to contain the original value arithmetically shifted right by
2679 the number of zeroes. */
2682 shiftr_zero (HOST_WIDE_INT *v)
2684 unsigned HOST_WIDE_INT tmp = *v;
2685 unsigned HOST_WIDE_INT sgn;
2691 sgn = tmp & ((unsigned HOST_WIDE_INT) 1 << (HOST_BITS_PER_WIDE_INT - 1));
2692 while ((tmp & 0x1) == 0 && n <= 32)
2694 tmp = (tmp >> 1) | sgn;
2701 /* After reload, split the load of an immediate constant. OPERANDS are the
2702 operands of the movsi_insn pattern which we are splitting. We return
2703 nonzero if we emitted a sequence to load the constant, zero if we emitted
2704 nothing because we want to use the splitter's default sequence. */
2707 split_load_immediate (rtx operands[])
2709 HOST_WIDE_INT val = INTVAL (operands[1]);
2711 HOST_WIDE_INT shifted = val;
2712 HOST_WIDE_INT shifted_compl = ~val;
2713 int num_zero = shiftr_zero (&shifted);
2714 int num_compl_zero = shiftr_zero (&shifted_compl);
2715 unsigned int regno = REGNO (operands[0]);
2717 /* This case takes care of single-bit set/clear constants, which we could
2718 also implement with BITSET/BITCLR. */
2720 && shifted >= -32768 && shifted < 65536
2721 && (D_REGNO_P (regno)
2722 || (regno >= REG_P0 && regno <= REG_P7 && num_zero <= 2)))
2724 emit_insn (gen_movsi (operands[0], GEN_INT (shifted)));
2725 emit_insn (gen_ashlsi3 (operands[0], operands[0], GEN_INT (num_zero)));
2730 tmp |= -(tmp & 0x8000);
2732 /* If high word has one bit set or clear, try to use a bit operation. */
2733 if (D_REGNO_P (regno))
2735 if (log2constp (val & 0xFFFF0000))
2737 emit_insn (gen_movsi (operands[0], GEN_INT (val & 0xFFFF)));
2738 emit_insn (gen_iorsi3 (operands[0], operands[0], GEN_INT (val & 0xFFFF0000)));
2741 else if (log2constp (val | 0xFFFF) && (val & 0x8000) != 0)
2743 emit_insn (gen_movsi (operands[0], GEN_INT (tmp)));
2744 emit_insn (gen_andsi3 (operands[0], operands[0], GEN_INT (val | 0xFFFF)));
2748 if (D_REGNO_P (regno))
2750 if (tmp >= -64 && tmp <= 63)
2752 emit_insn (gen_movsi (operands[0], GEN_INT (tmp)));
2753 emit_insn (gen_movstricthi_high (operands[0], GEN_INT (val & -65536)));
2757 if ((val & 0xFFFF0000) == 0)
2759 emit_insn (gen_movsi (operands[0], const0_rtx));
2760 emit_insn (gen_movsi_low (operands[0], operands[0], operands[1]));
2764 if ((val & 0xFFFF0000) == 0xFFFF0000)
2766 emit_insn (gen_movsi (operands[0], constm1_rtx));
2767 emit_insn (gen_movsi_low (operands[0], operands[0], operands[1]));
2772 /* Need DREGs for the remaining case. */
2777 && num_compl_zero && shifted_compl >= -64 && shifted_compl <= 63)
2779 /* If optimizing for size, generate a sequence that has more instructions
2781 emit_insn (gen_movsi (operands[0], GEN_INT (shifted_compl)));
2782 emit_insn (gen_ashlsi3 (operands[0], operands[0],
2783 GEN_INT (num_compl_zero)));
2784 emit_insn (gen_one_cmplsi2 (operands[0], operands[0]));
2790 /* Return true if the legitimate memory address for a memory operand of mode
2791 MODE. Return false if not. */
2794 bfin_valid_add (enum machine_mode mode, HOST_WIDE_INT value)
2796 unsigned HOST_WIDE_INT v = value > 0 ? value : -value;
2797 int sz = GET_MODE_SIZE (mode);
2798 int shift = sz == 1 ? 0 : sz == 2 ? 1 : 2;
2799 /* The usual offsettable_memref machinery doesn't work so well for this
2800 port, so we deal with the problem here. */
2801 if (value > 0 && sz == 8)
2803 return (v & ~(0x7fff << shift)) == 0;
2807 bfin_valid_reg_p (unsigned int regno, int strict, enum machine_mode mode,
2808 enum rtx_code outer_code)
2811 return REGNO_OK_FOR_BASE_STRICT_P (regno, mode, outer_code, SCRATCH);
2813 return REGNO_OK_FOR_BASE_NONSTRICT_P (regno, mode, outer_code, SCRATCH);
2817 bfin_legitimate_address_p (enum machine_mode mode, rtx x, int strict)
2819 switch (GET_CODE (x)) {
2821 if (bfin_valid_reg_p (REGNO (x), strict, mode, MEM))
2825 if (REG_P (XEXP (x, 0))
2826 && bfin_valid_reg_p (REGNO (XEXP (x, 0)), strict, mode, PLUS)
2827 && ((GET_CODE (XEXP (x, 1)) == UNSPEC && mode == SImode)
2828 || (GET_CODE (XEXP (x, 1)) == CONST_INT
2829 && bfin_valid_add (mode, INTVAL (XEXP (x, 1))))))
2834 if (LEGITIMATE_MODE_FOR_AUTOINC_P (mode)
2835 && REG_P (XEXP (x, 0))
2836 && bfin_valid_reg_p (REGNO (XEXP (x, 0)), strict, mode, POST_INC))
2839 if (LEGITIMATE_MODE_FOR_AUTOINC_P (mode)
2840 && XEXP (x, 0) == stack_pointer_rtx
2841 && REG_P (XEXP (x, 0))
2842 && bfin_valid_reg_p (REGNO (XEXP (x, 0)), strict, mode, PRE_DEC))
2851 /* Decide whether we can force certain constants to memory. If we
2852 decide we can't, the caller should be able to cope with it in
2856 bfin_cannot_force_const_mem (rtx x ATTRIBUTE_UNUSED)
2858 /* We have only one class of non-legitimate constants, and our movsi
2859 expander knows how to handle them. Dropping these constants into the
2860 data section would only shift the problem - we'd still get relocs
2861 outside the object, in the data section rather than the text section. */
2865 /* Ensure that for any constant of the form symbol + offset, the offset
2866 remains within the object. Any other constants are ok.
2867 This ensures that flat binaries never have to deal with relocations
2868 crossing section boundaries. */
2871 bfin_legitimate_constant_p (rtx x)
2874 HOST_WIDE_INT offset;
2876 if (GET_CODE (x) != CONST)
2880 gcc_assert (GET_CODE (x) == PLUS);
2884 if (GET_CODE (sym) != SYMBOL_REF
2885 || GET_CODE (x) != CONST_INT)
2887 offset = INTVAL (x);
2889 if (SYMBOL_REF_DECL (sym) == 0)
2892 || offset >= int_size_in_bytes (TREE_TYPE (SYMBOL_REF_DECL (sym))))
2899 bfin_rtx_costs (rtx x, int code, int outer_code, int *total, bool speed)
2901 int cost2 = COSTS_N_INSNS (1);
2907 if (outer_code == SET || outer_code == PLUS)
2908 *total = satisfies_constraint_Ks7 (x) ? 0 : cost2;
2909 else if (outer_code == AND)
2910 *total = log2constp (~INTVAL (x)) ? 0 : cost2;
2911 else if (outer_code == LE || outer_code == LT || outer_code == EQ)
2912 *total = (INTVAL (x) >= -4 && INTVAL (x) <= 3) ? 0 : cost2;
2913 else if (outer_code == LEU || outer_code == LTU)
2914 *total = (INTVAL (x) >= 0 && INTVAL (x) <= 7) ? 0 : cost2;
2915 else if (outer_code == MULT)
2916 *total = (INTVAL (x) == 2 || INTVAL (x) == 4) ? 0 : cost2;
2917 else if (outer_code == ASHIFT && (INTVAL (x) == 1 || INTVAL (x) == 2))
2919 else if (outer_code == ASHIFT || outer_code == ASHIFTRT
2920 || outer_code == LSHIFTRT)
2921 *total = (INTVAL (x) >= 0 && INTVAL (x) <= 31) ? 0 : cost2;
2922 else if (outer_code == IOR || outer_code == XOR)
2923 *total = (INTVAL (x) & (INTVAL (x) - 1)) == 0 ? 0 : cost2;
2932 *total = COSTS_N_INSNS (2);
2938 if (GET_MODE (x) == SImode)
2940 if (GET_CODE (op0) == MULT
2941 && GET_CODE (XEXP (op0, 1)) == CONST_INT)
2943 HOST_WIDE_INT val = INTVAL (XEXP (op0, 1));
2944 if (val == 2 || val == 4)
2947 *total += rtx_cost (XEXP (op0, 0), outer_code, speed);
2948 *total += rtx_cost (op1, outer_code, speed);
2953 if (GET_CODE (op0) != REG
2954 && (GET_CODE (op0) != SUBREG || GET_CODE (SUBREG_REG (op0)) != REG))
2955 *total += rtx_cost (op0, SET, speed);
2956 #if 0 /* We'd like to do this for accuracy, but it biases the loop optimizer
2957 towards creating too many induction variables. */
2958 if (!reg_or_7bit_operand (op1, SImode))
2959 *total += rtx_cost (op1, SET, speed);
2962 else if (GET_MODE (x) == DImode)
2965 if (GET_CODE (op1) != CONST_INT
2966 || !satisfies_constraint_Ks7 (op1))
2967 *total += rtx_cost (op1, PLUS, speed);
2968 if (GET_CODE (op0) != REG
2969 && (GET_CODE (op0) != SUBREG || GET_CODE (SUBREG_REG (op0)) != REG))
2970 *total += rtx_cost (op0, PLUS, speed);
2975 if (GET_MODE (x) == DImode)
2984 if (GET_MODE (x) == DImode)
2991 if (GET_CODE (op0) != REG
2992 && (GET_CODE (op0) != SUBREG || GET_CODE (SUBREG_REG (op0)) != REG))
2993 *total += rtx_cost (op0, code, speed);
3003 /* Handle special cases of IOR: rotates, ALIGN insns, movstricthi_high. */
3006 if ((GET_CODE (op0) == LSHIFTRT && GET_CODE (op1) == ASHIFT)
3007 || (GET_CODE (op0) == ASHIFT && GET_CODE (op1) == ZERO_EXTEND)
3008 || (GET_CODE (op0) == ASHIFT && GET_CODE (op1) == LSHIFTRT)
3009 || (GET_CODE (op0) == AND && GET_CODE (op1) == CONST_INT))
3016 if (GET_CODE (op0) != REG
3017 && (GET_CODE (op0) != SUBREG || GET_CODE (SUBREG_REG (op0)) != REG))
3018 *total += rtx_cost (op0, code, speed);
3020 if (GET_MODE (x) == DImode)
3026 if (GET_MODE (x) != SImode)
3031 if (! rhs_andsi3_operand (XEXP (x, 1), SImode))
3032 *total += rtx_cost (XEXP (x, 1), code, speed);
3036 if (! regorlog2_operand (XEXP (x, 1), SImode))
3037 *total += rtx_cost (XEXP (x, 1), code, speed);
3044 if (outer_code == SET
3045 && XEXP (x, 1) == const1_rtx
3046 && GET_CODE (XEXP (x, 2)) == CONST_INT)
3062 if (GET_CODE (op0) == GET_CODE (op1)
3063 && (GET_CODE (op0) == ZERO_EXTEND
3064 || GET_CODE (op0) == SIGN_EXTEND))
3066 *total = COSTS_N_INSNS (1);
3067 op0 = XEXP (op0, 0);
3068 op1 = XEXP (op1, 0);
3071 *total = COSTS_N_INSNS (1);
3073 *total = COSTS_N_INSNS (3);
3075 if (GET_CODE (op0) != REG
3076 && (GET_CODE (op0) != SUBREG || GET_CODE (SUBREG_REG (op0)) != REG))
3077 *total += rtx_cost (op0, MULT, speed);
3078 if (GET_CODE (op1) != REG
3079 && (GET_CODE (op1) != SUBREG || GET_CODE (SUBREG_REG (op1)) != REG))
3080 *total += rtx_cost (op1, MULT, speed);
3086 *total = COSTS_N_INSNS (32);
3091 if (outer_code == SET)
3100 /* Used for communication between {push,pop}_multiple_operation (which
3101 we use not only as a predicate) and the corresponding output functions. */
3102 static int first_preg_to_save, first_dreg_to_save;
3103 static int n_regs_to_save;
3106 push_multiple_operation (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
3108 int lastdreg = 8, lastpreg = 6;
3111 first_preg_to_save = lastpreg;
3112 first_dreg_to_save = lastdreg;
3113 for (i = 1, group = 0; i < XVECLEN (op, 0) - 1; i++)
3115 rtx t = XVECEXP (op, 0, i);
3119 if (GET_CODE (t) != SET)
3123 dest = SET_DEST (t);
3124 if (GET_CODE (dest) != MEM || ! REG_P (src))
3126 dest = XEXP (dest, 0);
3127 if (GET_CODE (dest) != PLUS
3128 || ! REG_P (XEXP (dest, 0))
3129 || REGNO (XEXP (dest, 0)) != REG_SP
3130 || GET_CODE (XEXP (dest, 1)) != CONST_INT
3131 || INTVAL (XEXP (dest, 1)) != -i * 4)
3134 regno = REGNO (src);
3137 if (D_REGNO_P (regno))
3140 first_dreg_to_save = lastdreg = regno - REG_R0;
3142 else if (regno >= REG_P0 && regno <= REG_P7)
3145 first_preg_to_save = lastpreg = regno - REG_P0;
3155 if (regno >= REG_P0 && regno <= REG_P7)
3158 first_preg_to_save = lastpreg = regno - REG_P0;
3160 else if (regno != REG_R0 + lastdreg + 1)
3165 else if (group == 2)
3167 if (regno != REG_P0 + lastpreg + 1)
3172 n_regs_to_save = 8 - first_dreg_to_save + 6 - first_preg_to_save;
3177 pop_multiple_operation (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
3179 int lastdreg = 8, lastpreg = 6;
3182 for (i = 1, group = 0; i < XVECLEN (op, 0); i++)
3184 rtx t = XVECEXP (op, 0, i);
3188 if (GET_CODE (t) != SET)
3192 dest = SET_DEST (t);
3193 if (GET_CODE (src) != MEM || ! REG_P (dest))
3195 src = XEXP (src, 0);
3199 if (! REG_P (src) || REGNO (src) != REG_SP)
3202 else if (GET_CODE (src) != PLUS
3203 || ! REG_P (XEXP (src, 0))
3204 || REGNO (XEXP (src, 0)) != REG_SP
3205 || GET_CODE (XEXP (src, 1)) != CONST_INT
3206 || INTVAL (XEXP (src, 1)) != (i - 1) * 4)
3209 regno = REGNO (dest);
3212 if (regno == REG_R7)
3217 else if (regno != REG_P0 + lastpreg - 1)
3222 else if (group == 1)
3224 if (regno != REG_R0 + lastdreg - 1)
3230 first_dreg_to_save = lastdreg;
3231 first_preg_to_save = lastpreg;
3232 n_regs_to_save = 8 - first_dreg_to_save + 6 - first_preg_to_save;
3236 /* Emit assembly code for one multi-register push described by INSN, with
3237 operands in OPERANDS. */
3240 output_push_multiple (rtx insn, rtx *operands)
3245 /* Validate the insn again, and compute first_[dp]reg_to_save. */
3246 ok = push_multiple_operation (PATTERN (insn), VOIDmode);
3249 if (first_dreg_to_save == 8)
3250 sprintf (buf, "[--sp] = ( p5:%d );\n", first_preg_to_save);
3251 else if (first_preg_to_save == 6)
3252 sprintf (buf, "[--sp] = ( r7:%d );\n", first_dreg_to_save);
3254 sprintf (buf, "[--sp] = ( r7:%d, p5:%d );\n",
3255 first_dreg_to_save, first_preg_to_save);
3257 output_asm_insn (buf, operands);
3260 /* Emit assembly code for one multi-register pop described by INSN, with
3261 operands in OPERANDS. */
3264 output_pop_multiple (rtx insn, rtx *operands)
3269 /* Validate the insn again, and compute first_[dp]reg_to_save. */
3270 ok = pop_multiple_operation (PATTERN (insn), VOIDmode);
3273 if (first_dreg_to_save == 8)
3274 sprintf (buf, "( p5:%d ) = [sp++];\n", first_preg_to_save);
3275 else if (first_preg_to_save == 6)
3276 sprintf (buf, "( r7:%d ) = [sp++];\n", first_dreg_to_save);
3278 sprintf (buf, "( r7:%d, p5:%d ) = [sp++];\n",
3279 first_dreg_to_save, first_preg_to_save);
3281 output_asm_insn (buf, operands);
3284 /* Adjust DST and SRC by OFFSET bytes, and generate one move in mode MODE. */
3287 single_move_for_movmem (rtx dst, rtx src, enum machine_mode mode, HOST_WIDE_INT offset)
3289 rtx scratch = gen_reg_rtx (mode);
3292 srcmem = adjust_address_nv (src, mode, offset);
3293 dstmem = adjust_address_nv (dst, mode, offset);
3294 emit_move_insn (scratch, srcmem);
3295 emit_move_insn (dstmem, scratch);
3298 /* Expand a string move operation of COUNT_EXP bytes from SRC to DST, with
3299 alignment ALIGN_EXP. Return true if successful, false if we should fall
3300 back on a different method. */
3303 bfin_expand_movmem (rtx dst, rtx src, rtx count_exp, rtx align_exp)
3305 rtx srcreg, destreg, countreg;
3306 HOST_WIDE_INT align = 0;
3307 unsigned HOST_WIDE_INT count = 0;
3309 if (GET_CODE (align_exp) == CONST_INT)
3310 align = INTVAL (align_exp);
3311 if (GET_CODE (count_exp) == CONST_INT)
3313 count = INTVAL (count_exp);
3315 if (!TARGET_INLINE_ALL_STRINGOPS && count > 64)
3320 /* If optimizing for size, only do single copies inline. */
3323 if (count == 2 && align < 2)
3325 if (count == 4 && align < 4)
3327 if (count != 1 && count != 2 && count != 4)
3330 if (align < 2 && count != 1)
3333 destreg = copy_to_mode_reg (Pmode, XEXP (dst, 0));
3334 if (destreg != XEXP (dst, 0))
3335 dst = replace_equiv_address_nv (dst, destreg);
3336 srcreg = copy_to_mode_reg (Pmode, XEXP (src, 0));
3337 if (srcreg != XEXP (src, 0))
3338 src = replace_equiv_address_nv (src, srcreg);
3340 if (count != 0 && align >= 2)
3342 unsigned HOST_WIDE_INT offset = 0;
3346 if ((count & ~3) == 4)
3348 single_move_for_movmem (dst, src, SImode, offset);
3351 else if (count & ~3)
3353 HOST_WIDE_INT new_count = ((count >> 2) & 0x3fffffff) - 1;
3354 countreg = copy_to_mode_reg (Pmode, GEN_INT (new_count));
3356 emit_insn (gen_rep_movsi (destreg, srcreg, countreg, destreg, srcreg));
3360 single_move_for_movmem (dst, src, HImode, offset);
3366 if ((count & ~1) == 2)
3368 single_move_for_movmem (dst, src, HImode, offset);
3371 else if (count & ~1)
3373 HOST_WIDE_INT new_count = ((count >> 1) & 0x7fffffff) - 1;
3374 countreg = copy_to_mode_reg (Pmode, GEN_INT (new_count));
3376 emit_insn (gen_rep_movhi (destreg, srcreg, countreg, destreg, srcreg));
3381 single_move_for_movmem (dst, src, QImode, offset);
3388 /* Compute the alignment for a local variable.
3389 TYPE is the data type, and ALIGN is the alignment that
3390 the object would ordinarily have. The value of this macro is used
3391 instead of that alignment to align the object. */
3394 bfin_local_alignment (tree type, int align)
3396 /* Increasing alignment for (relatively) big types allows the builtin
3397 memcpy can use 32 bit loads/stores. */
3398 if (TYPE_SIZE (type)
3399 && TREE_CODE (TYPE_SIZE (type)) == INTEGER_CST
3400 && (TREE_INT_CST_LOW (TYPE_SIZE (type)) > 8
3401 || TREE_INT_CST_HIGH (TYPE_SIZE (type))) && align < 32)
3406 /* Implement TARGET_SCHED_ISSUE_RATE. */
3409 bfin_issue_rate (void)
3415 bfin_adjust_cost (rtx insn, rtx link, rtx dep_insn, int cost)
3417 enum attr_type insn_type, dep_insn_type;
3418 int dep_insn_code_number;
3420 /* Anti and output dependencies have zero cost. */
3421 if (REG_NOTE_KIND (link) != 0)
3424 dep_insn_code_number = recog_memoized (dep_insn);
3426 /* If we can't recognize the insns, we can't really do anything. */
3427 if (dep_insn_code_number < 0 || recog_memoized (insn) < 0)
3430 insn_type = get_attr_type (insn);
3431 dep_insn_type = get_attr_type (dep_insn);
3433 if (dep_insn_type == TYPE_MOVE || dep_insn_type == TYPE_MCLD)
3435 rtx pat = PATTERN (dep_insn);
3436 if (GET_CODE (pat) == PARALLEL)
3437 pat = XVECEXP (pat, 0, 0);
3438 rtx dest = SET_DEST (pat);
3439 rtx src = SET_SRC (pat);
3440 if (! ADDRESS_REGNO_P (REGNO (dest))
3441 || ! (MEM_P (src) || D_REGNO_P (REGNO (src))))
3443 return cost + (dep_insn_type == TYPE_MOVE ? 4 : 3);
3450 /* Increment the counter for the number of loop instructions in the
3451 current function. */
3454 bfin_hardware_loop (void)
3456 cfun->machine->has_hardware_loops++;
3459 /* Maximum loop nesting depth. */
3460 #define MAX_LOOP_DEPTH 2
3462 /* Maximum size of a loop. */
3463 #define MAX_LOOP_LENGTH 2042
3465 /* Maximum distance of the LSETUP instruction from the loop start. */
3466 #define MAX_LSETUP_DISTANCE 30
3468 /* We need to keep a vector of loops */
3469 typedef struct loop_info *loop_info;
3470 DEF_VEC_P (loop_info);
3471 DEF_VEC_ALLOC_P (loop_info,heap);
3473 /* Information about a loop we have found (or are in the process of
3475 struct loop_info GTY (())
3477 /* loop number, for dumps */
3480 /* All edges that jump into and out of the loop. */
3481 VEC(edge,gc) *incoming;
3483 /* We can handle two cases: all incoming edges have the same destination
3484 block, or all incoming edges have the same source block. These two
3485 members are set to the common source or destination we found, or NULL
3486 if different blocks were found. If both are NULL the loop can't be
3488 basic_block incoming_src;
3489 basic_block incoming_dest;
3491 /* First block in the loop. This is the one branched to by the loop_end
3495 /* Last block in the loop (the one with the loop_end insn). */
3498 /* The successor block of the loop. This is the one the loop_end insn
3500 basic_block successor;
3502 /* The last instruction in the tail. */
3505 /* The loop_end insn. */
3508 /* The iteration register. */
3511 /* The new initialization insn. */
3514 /* The new initialization instruction. */
3517 /* The new label placed at the beginning of the loop. */
3520 /* The new label placed at the end of the loop. */
3523 /* The length of the loop. */
3526 /* The nesting depth of the loop. */
3529 /* Nonzero if we can't optimize this loop. */
3532 /* True if we have visited this loop. */
3535 /* True if this loop body clobbers any of LC0, LT0, or LB0. */
3538 /* True if this loop body clobbers any of LC1, LT1, or LB1. */
3541 /* Next loop in the graph. */
3542 struct loop_info *next;
3544 /* Immediate outer loop of this loop. */
3545 struct loop_info *outer;
3547 /* Vector of blocks only within the loop, including those within
3549 VEC (basic_block,heap) *blocks;
3551 /* Same information in a bitmap. */
3552 bitmap block_bitmap;
3554 /* Vector of inner loops within this loop */
3555 VEC (loop_info,heap) *loops;
3559 bfin_dump_loops (loop_info loops)
3563 for (loop = loops; loop; loop = loop->next)
3569 fprintf (dump_file, ";; loop %d: ", loop->loop_no);
3571 fprintf (dump_file, "(bad) ");
3572 fprintf (dump_file, "{head:%d, depth:%d}", loop->head->index, loop->depth);
3574 fprintf (dump_file, " blocks: [ ");
3575 for (ix = 0; VEC_iterate (basic_block, loop->blocks, ix, b); ix++)
3576 fprintf (dump_file, "%d ", b->index);
3577 fprintf (dump_file, "] ");
3579 fprintf (dump_file, " inner loops: [ ");
3580 for (ix = 0; VEC_iterate (loop_info, loop->loops, ix, i); ix++)
3581 fprintf (dump_file, "%d ", i->loop_no);
3582 fprintf (dump_file, "]\n");
3584 fprintf (dump_file, "\n");
3587 /* Scan the blocks of LOOP (and its inferiors) looking for basic block
3588 BB. Return true, if we find it. */
3591 bfin_bb_in_loop (loop_info loop, basic_block bb)
3593 return bitmap_bit_p (loop->block_bitmap, bb->index);
3596 /* Scan the blocks of LOOP (and its inferiors) looking for uses of
3597 REG. Return true, if we find any. Don't count the loop's loop_end
3598 insn if it matches LOOP_END. */
3601 bfin_scan_loop (loop_info loop, rtx reg, rtx loop_end)
3606 for (ix = 0; VEC_iterate (basic_block, loop->blocks, ix, bb); ix++)
3610 for (insn = BB_HEAD (bb);
3611 insn != NEXT_INSN (BB_END (bb));
3612 insn = NEXT_INSN (insn))
3616 if (insn == loop_end)
3618 if (reg_mentioned_p (reg, PATTERN (insn)))
3625 /* Estimate the length of INSN conservatively. */
3628 length_for_loop (rtx insn)
3631 if (JUMP_P (insn) && any_condjump_p (insn) && !optimize_size)
3633 if (ENABLE_WA_SPECULATIVE_SYNCS)
3635 else if (ENABLE_WA_SPECULATIVE_LOADS)
3638 else if (LABEL_P (insn))
3640 if (ENABLE_WA_SPECULATIVE_SYNCS)
3645 length += get_attr_length (insn);
3650 /* Optimize LOOP. */
3653 bfin_optimize_loop (loop_info loop)
3657 rtx insn, init_insn, last_insn, nop_insn;
3658 rtx loop_init, start_label, end_label;
3659 rtx reg_lc0, reg_lc1, reg_lt0, reg_lt1, reg_lb0, reg_lb1;
3661 rtx lc_reg, lt_reg, lb_reg;
3665 int inner_depth = 0;
3675 fprintf (dump_file, ";; loop %d bad when found\n", loop->loop_no);
3679 /* Every loop contains in its list of inner loops every loop nested inside
3680 it, even if there are intermediate loops. This works because we're doing
3681 a depth-first search here and never visit a loop more than once. */
3682 for (ix = 0; VEC_iterate (loop_info, loop->loops, ix, inner); ix++)
3684 bfin_optimize_loop (inner);
3686 if (!inner->bad && inner_depth < inner->depth)
3688 inner_depth = inner->depth;
3690 loop->clobber_loop0 |= inner->clobber_loop0;
3691 loop->clobber_loop1 |= inner->clobber_loop1;
3695 loop->depth = inner_depth + 1;
3696 if (loop->depth > MAX_LOOP_DEPTH)
3699 fprintf (dump_file, ";; loop %d too deep\n", loop->loop_no);
3703 /* Get the loop iteration register. */
3704 iter_reg = loop->iter_reg;
3706 if (!DPREG_P (iter_reg))
3709 fprintf (dump_file, ";; loop %d iteration count NOT in PREG or DREG\n",
3714 if (loop->incoming_src)
3716 /* Make sure the predecessor is before the loop start label, as required by
3717 the LSETUP instruction. */
3719 for (insn = BB_END (loop->incoming_src);
3720 insn && insn != loop->start_label;
3721 insn = NEXT_INSN (insn))
3722 length += length_for_loop (insn);
3727 fprintf (dump_file, ";; loop %d lsetup not before loop_start\n",
3732 if (length > MAX_LSETUP_DISTANCE)
3735 fprintf (dump_file, ";; loop %d lsetup too far away\n", loop->loop_no);
3740 /* Check if start_label appears before loop_end and calculate the
3741 offset between them. We calculate the length of instructions
3744 for (insn = loop->start_label;
3745 insn && insn != loop->loop_end;
3746 insn = NEXT_INSN (insn))
3747 length += length_for_loop (insn);
3752 fprintf (dump_file, ";; loop %d start_label not before loop_end\n",
3757 loop->length = length;
3758 if (loop->length > MAX_LOOP_LENGTH)
3761 fprintf (dump_file, ";; loop %d too long\n", loop->loop_no);
3765 /* Scan all the blocks to make sure they don't use iter_reg. */
3766 if (bfin_scan_loop (loop, iter_reg, loop->loop_end))
3769 fprintf (dump_file, ";; loop %d uses iterator\n", loop->loop_no);
3773 /* Scan all the insns to see if the loop body clobber
3774 any hardware loop registers. */
3776 reg_lc0 = gen_rtx_REG (SImode, REG_LC0);
3777 reg_lc1 = gen_rtx_REG (SImode, REG_LC1);
3778 reg_lt0 = gen_rtx_REG (SImode, REG_LT0);
3779 reg_lt1 = gen_rtx_REG (SImode, REG_LT1);
3780 reg_lb0 = gen_rtx_REG (SImode, REG_LB0);
3781 reg_lb1 = gen_rtx_REG (SImode, REG_LB1);
3783 for (ix = 0; VEC_iterate (basic_block, loop->blocks, ix, bb); ix++)
3787 for (insn = BB_HEAD (bb);
3788 insn != NEXT_INSN (BB_END (bb));
3789 insn = NEXT_INSN (insn))
3794 if (reg_set_p (reg_lc0, insn)
3795 || reg_set_p (reg_lt0, insn)
3796 || reg_set_p (reg_lb0, insn))
3797 loop->clobber_loop0 = 1;
3799 if (reg_set_p (reg_lc1, insn)
3800 || reg_set_p (reg_lt1, insn)
3801 || reg_set_p (reg_lb1, insn))
3802 loop->clobber_loop1 |= 1;
3806 if ((loop->clobber_loop0 && loop->clobber_loop1)
3807 || (loop->depth == MAX_LOOP_DEPTH && loop->clobber_loop0))
3809 loop->depth = MAX_LOOP_DEPTH + 1;
3811 fprintf (dump_file, ";; loop %d no loop reg available\n",
3816 /* There should be an instruction before the loop_end instruction
3817 in the same basic block. And the instruction must not be
3819 - CONDITIONAL BRANCH
3823 - Returns (RTS, RTN, etc.) */
3826 last_insn = PREV_INSN (loop->loop_end);
3830 for (; last_insn != PREV_INSN (BB_HEAD (bb));
3831 last_insn = PREV_INSN (last_insn))
3832 if (INSN_P (last_insn))
3835 if (last_insn != PREV_INSN (BB_HEAD (bb)))
3838 if (single_pred_p (bb)
3839 && single_pred (bb) != ENTRY_BLOCK_PTR)
3841 bb = single_pred (bb);
3842 last_insn = BB_END (bb);
3847 last_insn = NULL_RTX;
3855 fprintf (dump_file, ";; loop %d has no last instruction\n",
3860 if (JUMP_P (last_insn))
3862 loop_info inner = (loop_info) bb->aux;
3864 && inner->outer == loop
3865 && inner->loop_end == last_insn
3866 && inner->depth == 1)
3867 /* This jump_insn is the exact loop_end of an inner loop
3868 and to be optimized away. So use the inner's last_insn. */
3869 last_insn = inner->last_insn;
3873 fprintf (dump_file, ";; loop %d has bad last instruction\n",
3878 else if (CALL_P (last_insn)
3879 || (GET_CODE (PATTERN (last_insn)) != SEQUENCE
3880 && get_attr_type (last_insn) == TYPE_SYNC)
3881 || recog_memoized (last_insn) == CODE_FOR_return_internal)
3884 fprintf (dump_file, ";; loop %d has bad last instruction\n",
3889 if (GET_CODE (PATTERN (last_insn)) == ASM_INPUT
3890 || asm_noperands (PATTERN (last_insn)) >= 0
3891 || (GET_CODE (PATTERN (last_insn)) != SEQUENCE
3892 && get_attr_seq_insns (last_insn) == SEQ_INSNS_MULTI))
3894 nop_insn = emit_insn_after (gen_nop (), last_insn);
3895 last_insn = nop_insn;
3898 loop->last_insn = last_insn;
3900 /* The loop is good for replacement. */
3901 start_label = loop->start_label;
3902 end_label = gen_label_rtx ();
3903 iter_reg = loop->iter_reg;
3905 if (loop->depth == 1 && !loop->clobber_loop1)
3910 loop->clobber_loop1 = 1;
3917 loop->clobber_loop0 = 1;
3920 /* If iter_reg is a DREG, we need generate an instruction to load
3921 the loop count into LC register. */
3922 if (D_REGNO_P (REGNO (iter_reg)))
3924 init_insn = gen_movsi (lc_reg, iter_reg);
3925 loop_init = gen_lsetup_without_autoinit (lt_reg, start_label,
3929 else if (P_REGNO_P (REGNO (iter_reg)))
3931 init_insn = NULL_RTX;
3932 loop_init = gen_lsetup_with_autoinit (lt_reg, start_label,
3939 loop->init = init_insn;
3940 loop->end_label = end_label;
3941 loop->loop_init = loop_init;
3945 fprintf (dump_file, ";; replacing loop %d initializer with\n",
3947 print_rtl_single (dump_file, loop->loop_init);
3948 fprintf (dump_file, ";; replacing loop %d terminator with\n",
3950 print_rtl_single (dump_file, loop->loop_end);
3955 if (loop->init != NULL_RTX)
3956 emit_insn (loop->init);
3957 seq_end = emit_insn (loop->loop_init);
3962 if (loop->incoming_src)
3964 rtx prev = BB_END (loop->incoming_src);
3965 if (VEC_length (edge, loop->incoming) > 1
3966 || !(VEC_last (edge, loop->incoming)->flags & EDGE_FALLTHRU))
3968 gcc_assert (JUMP_P (prev));
3969 prev = PREV_INSN (prev);
3971 emit_insn_after (seq, prev);
3979 if (loop->head != loop->incoming_dest)
3981 FOR_EACH_EDGE (e, ei, loop->head->preds)
3983 if (e->flags & EDGE_FALLTHRU)
3985 rtx newjump = gen_jump (loop->start_label);
3986 emit_insn_before (newjump, BB_HEAD (loop->head));
3987 new_bb = create_basic_block (newjump, newjump, loop->head->prev_bb);
3988 gcc_assert (new_bb = loop->head->prev_bb);
3994 emit_insn_before (seq, BB_HEAD (loop->head));
3995 seq = emit_label_before (gen_label_rtx (), seq);
3997 new_bb = create_basic_block (seq, seq_end, loop->head->prev_bb);
3998 FOR_EACH_EDGE (e, ei, loop->incoming)
4000 if (!(e->flags & EDGE_FALLTHRU)
4001 || e->dest != loop->head)
4002 redirect_edge_and_branch_force (e, new_bb);
4004 redirect_edge_succ (e, new_bb);
4008 delete_insn (loop->loop_end);
4009 /* Insert the loop end label before the last instruction of the loop. */
4010 emit_label_before (loop->end_label, loop->last_insn);
4017 fprintf (dump_file, ";; loop %d is bad\n", loop->loop_no);
4021 if (DPREG_P (loop->iter_reg))
4023 /* If loop->iter_reg is a DREG or PREG, we can split it here
4024 without scratch register. */
4027 emit_insn_before (gen_addsi3 (loop->iter_reg,
4032 emit_insn_before (gen_cmpsi (loop->iter_reg, const0_rtx),
4035 insn = emit_jump_insn_before (gen_bne (loop->start_label),
4038 JUMP_LABEL (insn) = loop->start_label;
4039 LABEL_NUSES (loop->start_label)++;
4040 delete_insn (loop->loop_end);
4044 /* Called from bfin_reorg_loops when a potential loop end is found. LOOP is
4045 a newly set up structure describing the loop, it is this function's
4046 responsibility to fill most of it. TAIL_BB and TAIL_INSN point to the
4047 loop_end insn and its enclosing basic block. */
4050 bfin_discover_loop (loop_info loop, basic_block tail_bb, rtx tail_insn)
4054 VEC (basic_block,heap) *works = VEC_alloc (basic_block,heap,20);
4056 loop->tail = tail_bb;
4057 loop->head = BRANCH_EDGE (tail_bb)->dest;
4058 loop->successor = FALLTHRU_EDGE (tail_bb)->dest;
4059 loop->loop_end = tail_insn;
4060 loop->last_insn = NULL_RTX;
4061 loop->iter_reg = SET_DEST (XVECEXP (PATTERN (tail_insn), 0, 1));
4062 loop->depth = loop->length = 0;
4064 loop->clobber_loop0 = loop->clobber_loop1 = 0;
4067 loop->incoming = VEC_alloc (edge, gc, 2);
4068 loop->init = loop->loop_init = NULL_RTX;
4069 loop->start_label = XEXP (XEXP (SET_SRC (XVECEXP (PATTERN (tail_insn), 0, 0)), 1), 0);
4070 loop->end_label = NULL_RTX;
4073 VEC_safe_push (basic_block, heap, works, loop->head);
4075 while (VEC_iterate (basic_block, works, dwork++, bb))
4079 if (bb == EXIT_BLOCK_PTR)
4081 /* We've reached the exit block. The loop must be bad. */
4084 ";; Loop is bad - reached exit block while scanning\n");
4089 if (bitmap_bit_p (loop->block_bitmap, bb->index))
4092 /* We've not seen this block before. Add it to the loop's
4093 list and then add each successor to the work list. */
4095 VEC_safe_push (basic_block, heap, loop->blocks, bb);
4096 bitmap_set_bit (loop->block_bitmap, bb->index);
4100 FOR_EACH_EDGE (e, ei, bb->succs)
4102 basic_block succ = EDGE_SUCC (bb, ei.index)->dest;
4103 if (!REGNO_REG_SET_P (df_get_live_in (succ),
4104 REGNO (loop->iter_reg)))
4106 if (!VEC_space (basic_block, works, 1))
4110 VEC_block_remove (basic_block, works, 0, dwork);
4114 VEC_reserve (basic_block, heap, works, 1);
4116 VEC_quick_push (basic_block, works, succ);
4121 /* Find the predecessor, and make sure nothing else jumps into this loop. */
4125 for (dwork = 0; VEC_iterate (basic_block, loop->blocks, dwork, bb); dwork++)
4129 FOR_EACH_EDGE (e, ei, bb->preds)
4131 basic_block pred = e->src;
4133 if (!bfin_bb_in_loop (loop, pred))
4136 fprintf (dump_file, ";; Loop %d: incoming edge %d -> %d\n",