mips.h (ISA_HAS_LDC1_SDC1): New macro.
[official-gcc.git] / gcc / config / mips / mips.c
blob0948832b8ab06b7f4482d3feede00cd57012d09c
1 /* Subroutines used for MIPS code generation.
2 Copyright (C) 1989, 1990, 1991, 1993, 1994, 1995, 1996, 1997, 1998,
3 1999, 2000, 2001, 2002, 2003, 2004, 2005, 2006, 2007
4 Free Software Foundation, Inc.
5 Contributed by A. Lichnewsky, lich@inria.inria.fr.
6 Changes by Michael Meissner, meissner@osf.org.
7 64-bit r4000 support by Ian Lance Taylor, ian@cygnus.com, and
8 Brendan Eich, brendan@microunity.com.
10 This file is part of GCC.
12 GCC is free software; you can redistribute it and/or modify
13 it under the terms of the GNU General Public License as published by
14 the Free Software Foundation; either version 3, or (at your option)
15 any later version.
17 GCC is distributed in the hope that it will be useful,
18 but WITHOUT ANY WARRANTY; without even the implied warranty of
19 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
20 GNU General Public License for more details.
22 You should have received a copy of the GNU General Public License
23 along with GCC; see the file COPYING3. If not see
24 <http://www.gnu.org/licenses/>. */
26 #include "config.h"
27 #include "system.h"
28 #include "coretypes.h"
29 #include "tm.h"
30 #include <signal.h>
31 #include "rtl.h"
32 #include "regs.h"
33 #include "hard-reg-set.h"
34 #include "real.h"
35 #include "insn-config.h"
36 #include "conditions.h"
37 #include "insn-attr.h"
38 #include "recog.h"
39 #include "toplev.h"
40 #include "output.h"
41 #include "tree.h"
42 #include "function.h"
43 #include "expr.h"
44 #include "optabs.h"
45 #include "flags.h"
46 #include "reload.h"
47 #include "tm_p.h"
48 #include "ggc.h"
49 #include "gstab.h"
50 #include "hashtab.h"
51 #include "debug.h"
52 #include "target.h"
53 #include "target-def.h"
54 #include "integrate.h"
55 #include "langhooks.h"
56 #include "cfglayout.h"
57 #include "sched-int.h"
58 #include "tree-gimple.h"
59 #include "bitmap.h"
60 #include "diagnostic.h"
62 /* True if X is an unspec wrapper around a SYMBOL_REF or LABEL_REF. */
63 #define UNSPEC_ADDRESS_P(X) \
64 (GET_CODE (X) == UNSPEC \
65 && XINT (X, 1) >= UNSPEC_ADDRESS_FIRST \
66 && XINT (X, 1) < UNSPEC_ADDRESS_FIRST + NUM_SYMBOL_TYPES)
68 /* Extract the symbol or label from UNSPEC wrapper X. */
69 #define UNSPEC_ADDRESS(X) \
70 XVECEXP (X, 0, 0)
72 /* Extract the symbol type from UNSPEC wrapper X. */
73 #define UNSPEC_ADDRESS_TYPE(X) \
74 ((enum mips_symbol_type) (XINT (X, 1) - UNSPEC_ADDRESS_FIRST))
76 /* The maximum distance between the top of the stack frame and the
77 value $sp has when we save and restore registers.
79 The value for normal-mode code must be a SMALL_OPERAND and must
80 preserve the maximum stack alignment. We therefore use a value
81 of 0x7ff0 in this case.
83 MIPS16e SAVE and RESTORE instructions can adjust the stack pointer by
84 up to 0x7f8 bytes and can usually save or restore all the registers
85 that we need to save or restore. (Note that we can only use these
86 instructions for o32, for which the stack alignment is 8 bytes.)
88 We use a maximum gap of 0x100 or 0x400 for MIPS16 code when SAVE and
89 RESTORE are not available. We can then use unextended instructions
90 to save and restore registers, and to allocate and deallocate the top
91 part of the frame. */
92 #define MIPS_MAX_FIRST_STACK_STEP \
93 (!TARGET_MIPS16 ? 0x7ff0 \
94 : GENERATE_MIPS16E_SAVE_RESTORE ? 0x7f8 \
95 : TARGET_64BIT ? 0x100 : 0x400)
97 /* True if INSN is a mips.md pattern or asm statement. */
98 #define USEFUL_INSN_P(INSN) \
99 (INSN_P (INSN) \
100 && GET_CODE (PATTERN (INSN)) != USE \
101 && GET_CODE (PATTERN (INSN)) != CLOBBER \
102 && GET_CODE (PATTERN (INSN)) != ADDR_VEC \
103 && GET_CODE (PATTERN (INSN)) != ADDR_DIFF_VEC)
105 /* If INSN is a delayed branch sequence, return the first instruction
106 in the sequence, otherwise return INSN itself. */
107 #define SEQ_BEGIN(INSN) \
108 (INSN_P (INSN) && GET_CODE (PATTERN (INSN)) == SEQUENCE \
109 ? XVECEXP (PATTERN (INSN), 0, 0) \
110 : (INSN))
112 /* Likewise for the last instruction in a delayed branch sequence. */
113 #define SEQ_END(INSN) \
114 (INSN_P (INSN) && GET_CODE (PATTERN (INSN)) == SEQUENCE \
115 ? XVECEXP (PATTERN (INSN), 0, XVECLEN (PATTERN (INSN), 0) - 1) \
116 : (INSN))
118 /* Execute the following loop body with SUBINSN set to each instruction
119 between SEQ_BEGIN (INSN) and SEQ_END (INSN) inclusive. */
120 #define FOR_EACH_SUBINSN(SUBINSN, INSN) \
121 for ((SUBINSN) = SEQ_BEGIN (INSN); \
122 (SUBINSN) != NEXT_INSN (SEQ_END (INSN)); \
123 (SUBINSN) = NEXT_INSN (SUBINSN))
125 /* True if bit BIT is set in VALUE. */
126 #define BITSET_P(VALUE, BIT) (((VALUE) & (1 << (BIT))) != 0)
128 /* Classifies an address.
130 ADDRESS_REG
131 A natural register + offset address. The register satisfies
132 mips_valid_base_register_p and the offset is a const_arith_operand.
134 ADDRESS_LO_SUM
135 A LO_SUM rtx. The first operand is a valid base register and
136 the second operand is a symbolic address.
138 ADDRESS_CONST_INT
139 A signed 16-bit constant address.
141 ADDRESS_SYMBOLIC:
142 A constant symbolic address (equivalent to CONSTANT_SYMBOLIC). */
143 enum mips_address_type {
144 ADDRESS_REG,
145 ADDRESS_LO_SUM,
146 ADDRESS_CONST_INT,
147 ADDRESS_SYMBOLIC
150 /* Macros to create an enumeration identifier for a function prototype. */
151 #define MIPS_FTYPE_NAME1(A, B) MIPS_##A##_FTYPE_##B
152 #define MIPS_FTYPE_NAME2(A, B, C) MIPS_##A##_FTYPE_##B##_##C
153 #define MIPS_FTYPE_NAME3(A, B, C, D) MIPS_##A##_FTYPE_##B##_##C##_##D
154 #define MIPS_FTYPE_NAME4(A, B, C, D, E) MIPS_##A##_FTYPE_##B##_##C##_##D##_##E
156 /* Classifies the prototype of a builtin function. */
157 enum mips_function_type
159 #define DEF_MIPS_FTYPE(NARGS, LIST) MIPS_FTYPE_NAME##NARGS LIST,
160 #include "config/mips/mips-ftypes.def"
161 #undef DEF_MIPS_FTYPE
162 MIPS_MAX_FTYPE_MAX
165 /* Specifies how a builtin function should be converted into rtl. */
166 enum mips_builtin_type
168 /* The builtin corresponds directly to an .md pattern. The return
169 value is mapped to operand 0 and the arguments are mapped to
170 operands 1 and above. */
171 MIPS_BUILTIN_DIRECT,
173 /* The builtin corresponds directly to an .md pattern. There is no return
174 value and the arguments are mapped to operands 0 and above. */
175 MIPS_BUILTIN_DIRECT_NO_TARGET,
177 /* The builtin corresponds to a comparison instruction followed by
178 a mips_cond_move_tf_ps pattern. The first two arguments are the
179 values to compare and the second two arguments are the vector
180 operands for the movt.ps or movf.ps instruction (in assembly order). */
181 MIPS_BUILTIN_MOVF,
182 MIPS_BUILTIN_MOVT,
184 /* The builtin corresponds to a V2SF comparison instruction. Operand 0
185 of this instruction is the result of the comparison, which has mode
186 CCV2 or CCV4. The function arguments are mapped to operands 1 and
187 above. The function's return value is an SImode boolean that is
188 true under the following conditions:
190 MIPS_BUILTIN_CMP_ANY: one of the registers is true
191 MIPS_BUILTIN_CMP_ALL: all of the registers are true
192 MIPS_BUILTIN_CMP_LOWER: the first register is true
193 MIPS_BUILTIN_CMP_UPPER: the second register is true. */
194 MIPS_BUILTIN_CMP_ANY,
195 MIPS_BUILTIN_CMP_ALL,
196 MIPS_BUILTIN_CMP_UPPER,
197 MIPS_BUILTIN_CMP_LOWER,
199 /* As above, but the instruction only sets a single $fcc register. */
200 MIPS_BUILTIN_CMP_SINGLE,
202 /* For generating bposge32 branch instructions in MIPS32 DSP ASE. */
203 MIPS_BUILTIN_BPOSGE32
206 /* Invokes MACRO (COND) for each c.cond.fmt condition. */
207 #define MIPS_FP_CONDITIONS(MACRO) \
208 MACRO (f), \
209 MACRO (un), \
210 MACRO (eq), \
211 MACRO (ueq), \
212 MACRO (olt), \
213 MACRO (ult), \
214 MACRO (ole), \
215 MACRO (ule), \
216 MACRO (sf), \
217 MACRO (ngle), \
218 MACRO (seq), \
219 MACRO (ngl), \
220 MACRO (lt), \
221 MACRO (nge), \
222 MACRO (le), \
223 MACRO (ngt)
225 /* Enumerates the codes above as MIPS_FP_COND_<X>. */
226 #define DECLARE_MIPS_COND(X) MIPS_FP_COND_ ## X
227 enum mips_fp_condition {
228 MIPS_FP_CONDITIONS (DECLARE_MIPS_COND)
231 /* Index X provides the string representation of MIPS_FP_COND_<X>. */
232 #define STRINGIFY(X) #X
233 static const char *const mips_fp_conditions[] = {
234 MIPS_FP_CONDITIONS (STRINGIFY)
237 /* Information about a function's frame layout. */
238 struct mips_frame_info GTY(())
240 /* The size of the frame in bytes. */
241 HOST_WIDE_INT total_size;
243 /* The number of bytes allocated to variables. */
244 HOST_WIDE_INT var_size;
246 /* The number of bytes allocated to outgoing function arguments. */
247 HOST_WIDE_INT args_size;
249 /* The number of bytes allocated to the .cprestore slot, or 0 if there
250 is no such slot. */
251 HOST_WIDE_INT cprestore_size;
253 /* Bit X is set if the function saves or restores GPR X. */
254 unsigned int mask;
256 /* Likewise FPR X. */
257 unsigned int fmask;
259 /* The number of GPRs and FPRs saved. */
260 unsigned int num_gp;
261 unsigned int num_fp;
263 /* The offset of the topmost GPR and FPR save slots from the top of
264 the frame, or zero if no such slots are needed. */
265 HOST_WIDE_INT gp_save_offset;
266 HOST_WIDE_INT fp_save_offset;
268 /* Likewise, but giving offsets from the bottom of the frame. */
269 HOST_WIDE_INT gp_sp_offset;
270 HOST_WIDE_INT fp_sp_offset;
272 /* The offset of arg_pointer_rtx from frame_pointer_rtx. */
273 HOST_WIDE_INT arg_pointer_offset;
275 /* The offset of hard_frame_pointer_rtx from frame_pointer_rtx. */
276 HOST_WIDE_INT hard_frame_pointer_offset;
279 struct machine_function GTY(()) {
280 /* Pseudo-reg holding the value of $28 in a mips16 function which
281 refers to GP relative global variables. */
282 rtx mips16_gp_pseudo_rtx;
284 /* The number of extra stack bytes taken up by register varargs.
285 This area is allocated by the callee at the very top of the frame. */
286 int varargs_size;
288 /* Current frame information, calculated by mips_compute_frame_info. */
289 struct mips_frame_info frame;
291 /* The register to use as the global pointer within this function. */
292 unsigned int global_pointer;
294 /* True if mips_adjust_insn_length should ignore an instruction's
295 hazard attribute. */
296 bool ignore_hazard_length_p;
298 /* True if the whole function is suitable for .set noreorder and
299 .set nomacro. */
300 bool all_noreorder_p;
302 /* True if the function is known to have an instruction that needs $gp. */
303 bool has_gp_insn_p;
305 /* True if we have emitted an instruction to initialize
306 mips16_gp_pseudo_rtx. */
307 bool initialized_mips16_gp_pseudo_p;
310 /* Information about a single argument. */
311 struct mips_arg_info
313 /* True if the argument is passed in a floating-point register, or
314 would have been if we hadn't run out of registers. */
315 bool fpr_p;
317 /* The number of words passed in registers, rounded up. */
318 unsigned int reg_words;
320 /* For EABI, the offset of the first register from GP_ARG_FIRST or
321 FP_ARG_FIRST. For other ABIs, the offset of the first register from
322 the start of the ABI's argument structure (see the CUMULATIVE_ARGS
323 comment for details).
325 The value is MAX_ARGS_IN_REGISTERS if the argument is passed entirely
326 on the stack. */
327 unsigned int reg_offset;
329 /* The number of words that must be passed on the stack, rounded up. */
330 unsigned int stack_words;
332 /* The offset from the start of the stack overflow area of the argument's
333 first stack word. Only meaningful when STACK_WORDS is nonzero. */
334 unsigned int stack_offset;
338 /* Information about an address described by mips_address_type.
340 ADDRESS_CONST_INT
341 No fields are used.
343 ADDRESS_REG
344 REG is the base register and OFFSET is the constant offset.
346 ADDRESS_LO_SUM
347 REG is the register that contains the high part of the address,
348 OFFSET is the symbolic address being referenced and SYMBOL_TYPE
349 is the type of OFFSET's symbol.
351 ADDRESS_SYMBOLIC
352 SYMBOL_TYPE is the type of symbol being referenced. */
354 struct mips_address_info
356 enum mips_address_type type;
357 rtx reg;
358 rtx offset;
359 enum mips_symbol_type symbol_type;
363 /* One stage in a constant building sequence. These sequences have
364 the form:
366 A = VALUE[0]
367 A = A CODE[1] VALUE[1]
368 A = A CODE[2] VALUE[2]
371 where A is an accumulator, each CODE[i] is a binary rtl operation
372 and each VALUE[i] is a constant integer. */
373 struct mips_integer_op {
374 enum rtx_code code;
375 unsigned HOST_WIDE_INT value;
379 /* The largest number of operations needed to load an integer constant.
380 The worst accepted case for 64-bit constants is LUI,ORI,SLL,ORI,SLL,ORI.
381 When the lowest bit is clear, we can try, but reject a sequence with
382 an extra SLL at the end. */
383 #define MIPS_MAX_INTEGER_OPS 7
385 /* Information about a MIPS16e SAVE or RESTORE instruction. */
386 struct mips16e_save_restore_info {
387 /* The number of argument registers saved by a SAVE instruction.
388 0 for RESTORE instructions. */
389 unsigned int nargs;
391 /* Bit X is set if the instruction saves or restores GPR X. */
392 unsigned int mask;
394 /* The total number of bytes to allocate. */
395 HOST_WIDE_INT size;
398 /* Global variables for machine-dependent things. */
400 /* Threshold for data being put into the small data/bss area, instead
401 of the normal data area. */
402 int mips_section_threshold = -1;
404 /* Count the number of .file directives, so that .loc is up to date. */
405 int num_source_filenames = 0;
407 /* Name of the file containing the current function. */
408 const char *current_function_file = "";
410 /* Count the number of sdb related labels are generated (to find block
411 start and end boundaries). */
412 int sdb_label_count = 0;
414 /* Next label # for each statement for Silicon Graphics IRIS systems. */
415 int sym_lineno = 0;
417 /* Map GCC register number to debugger register number. */
418 int mips_dbx_regno[FIRST_PSEUDO_REGISTER];
419 int mips_dwarf_regno[FIRST_PSEUDO_REGISTER];
421 /* Number of nested .set noreorder, noat, nomacro, and volatile requests. */
422 int set_noreorder;
423 int set_noat;
424 int set_nomacro;
426 /* The next branch instruction is a branch likely, not branch normal. */
427 int mips_branch_likely;
429 /* The operands passed to the last cmpMM expander. */
430 rtx cmp_operands[2];
432 /* The target cpu for code generation. */
433 enum processor_type mips_arch;
434 const struct mips_cpu_info *mips_arch_info;
436 /* The target cpu for optimization and scheduling. */
437 enum processor_type mips_tune;
438 const struct mips_cpu_info *mips_tune_info;
440 /* Which instruction set architecture to use. */
441 int mips_isa;
443 /* The architecture selected by -mipsN. */
444 static const struct mips_cpu_info *mips_isa_info;
446 /* Which ABI to use. */
447 int mips_abi = MIPS_ABI_DEFAULT;
449 /* Cost information to use. */
450 const struct mips_rtx_cost_data *mips_cost;
452 /* Remember the ambient target flags, excluding mips16. */
453 static int mips_base_target_flags;
454 /* The mips16 command-line target flags only. */
455 static bool mips_base_mips16;
456 /* Similar copies of option settings. */
457 static int mips_flag_delayed_branch; /* flag_delayed_branch */
458 static int mips_base_schedule_insns; /* flag_schedule_insns */
459 static int mips_base_reorder_blocks_and_partition; /* flag_reorder... */
460 static int mips_base_move_loop_invariants; /* flag_move_loop_invariants */
461 static int mips_base_align_loops; /* align_loops */
462 static int mips_base_align_jumps; /* align_jumps */
463 static int mips_base_align_functions; /* align_functions */
465 /* The -mtext-loads setting. */
466 enum mips_code_readable_setting mips_code_readable = CODE_READABLE_YES;
468 /* If TRUE, we split addresses into their high and low parts in the RTL. */
469 int mips_split_addresses;
471 /* Array giving truth value on whether or not a given hard register
472 can support a given mode. */
473 char mips_hard_regno_mode_ok[(int)MAX_MACHINE_MODE][FIRST_PSEUDO_REGISTER];
475 /* List of all MIPS punctuation characters used by print_operand. */
476 char mips_print_operand_punct[256];
478 static GTY (()) int mips_output_filename_first_time = 1;
480 /* mips_split_p[X] is true if symbols of type X can be split by
481 mips_split_symbol(). */
482 bool mips_split_p[NUM_SYMBOL_TYPES];
484 /* mips_lo_relocs[X] is the relocation to use when a symbol of type X
485 appears in a LO_SUM. It can be null if such LO_SUMs aren't valid or
486 if they are matched by a special .md file pattern. */
487 static const char *mips_lo_relocs[NUM_SYMBOL_TYPES];
489 /* Likewise for HIGHs. */
490 static const char *mips_hi_relocs[NUM_SYMBOL_TYPES];
492 /* Map hard register number to register class */
493 const enum reg_class mips_regno_to_class[] =
495 LEA_REGS, LEA_REGS, M16_NA_REGS, V1_REG,
496 M16_REGS, M16_REGS, M16_REGS, M16_REGS,
497 LEA_REGS, LEA_REGS, LEA_REGS, LEA_REGS,
498 LEA_REGS, LEA_REGS, LEA_REGS, LEA_REGS,
499 M16_NA_REGS, M16_NA_REGS, LEA_REGS, LEA_REGS,
500 LEA_REGS, LEA_REGS, LEA_REGS, LEA_REGS,
501 T_REG, PIC_FN_ADDR_REG, LEA_REGS, LEA_REGS,
502 LEA_REGS, LEA_REGS, LEA_REGS, LEA_REGS,
503 FP_REGS, FP_REGS, FP_REGS, FP_REGS,
504 FP_REGS, FP_REGS, FP_REGS, FP_REGS,
505 FP_REGS, FP_REGS, FP_REGS, FP_REGS,
506 FP_REGS, FP_REGS, FP_REGS, FP_REGS,
507 FP_REGS, FP_REGS, FP_REGS, FP_REGS,
508 FP_REGS, FP_REGS, FP_REGS, FP_REGS,
509 FP_REGS, FP_REGS, FP_REGS, FP_REGS,
510 FP_REGS, FP_REGS, FP_REGS, FP_REGS,
511 MD0_REG, MD1_REG, NO_REGS, ST_REGS,
512 ST_REGS, ST_REGS, ST_REGS, ST_REGS,
513 ST_REGS, ST_REGS, ST_REGS, NO_REGS,
514 NO_REGS, ALL_REGS, ALL_REGS, NO_REGS,
515 COP0_REGS, COP0_REGS, COP0_REGS, COP0_REGS,
516 COP0_REGS, COP0_REGS, COP0_REGS, COP0_REGS,
517 COP0_REGS, COP0_REGS, COP0_REGS, COP0_REGS,
518 COP0_REGS, COP0_REGS, COP0_REGS, COP0_REGS,
519 COP0_REGS, COP0_REGS, COP0_REGS, COP0_REGS,
520 COP0_REGS, COP0_REGS, COP0_REGS, COP0_REGS,
521 COP0_REGS, COP0_REGS, COP0_REGS, COP0_REGS,
522 COP0_REGS, COP0_REGS, COP0_REGS, COP0_REGS,
523 COP2_REGS, COP2_REGS, COP2_REGS, COP2_REGS,
524 COP2_REGS, COP2_REGS, COP2_REGS, COP2_REGS,
525 COP2_REGS, COP2_REGS, COP2_REGS, COP2_REGS,
526 COP2_REGS, COP2_REGS, COP2_REGS, COP2_REGS,
527 COP2_REGS, COP2_REGS, COP2_REGS, COP2_REGS,
528 COP2_REGS, COP2_REGS, COP2_REGS, COP2_REGS,
529 COP2_REGS, COP2_REGS, COP2_REGS, COP2_REGS,
530 COP2_REGS, COP2_REGS, COP2_REGS, COP2_REGS,
531 COP3_REGS, COP3_REGS, COP3_REGS, COP3_REGS,
532 COP3_REGS, COP3_REGS, COP3_REGS, COP3_REGS,
533 COP3_REGS, COP3_REGS, COP3_REGS, COP3_REGS,
534 COP3_REGS, COP3_REGS, COP3_REGS, COP3_REGS,
535 COP3_REGS, COP3_REGS, COP3_REGS, COP3_REGS,
536 COP3_REGS, COP3_REGS, COP3_REGS, COP3_REGS,
537 COP3_REGS, COP3_REGS, COP3_REGS, COP3_REGS,
538 COP3_REGS, COP3_REGS, COP3_REGS, COP3_REGS,
539 DSP_ACC_REGS, DSP_ACC_REGS, DSP_ACC_REGS, DSP_ACC_REGS,
540 DSP_ACC_REGS, DSP_ACC_REGS, ALL_REGS, ALL_REGS,
541 ALL_REGS, ALL_REGS, ALL_REGS, ALL_REGS
544 /* Table of machine dependent attributes. */
545 const struct attribute_spec mips_attribute_table[] =
547 { "long_call", 0, 0, false, true, true, NULL },
548 { "far", 0, 0, false, true, true, NULL },
549 { "near", 0, 0, false, true, true, NULL },
550 /* Switch MIPS16 ASE on and off per-function. We would really like
551 to make these type attributes, but GCC doesn't provide the hooks
552 we need to support the right conversion rules. As declaration
553 attributes, they affect code generation but don't carry other
554 semantics. */
555 { "mips16", 0, 0, true, false, false, NULL },
556 { "nomips16", 0, 0, true, false, false, NULL },
557 { NULL, 0, 0, false, false, false, NULL }
560 /* A table describing all the processors gcc knows about. Names are
561 matched in the order listed. The first mention of an ISA level is
562 taken as the canonical name for that ISA.
564 To ease comparison, please keep this table in the same order
565 as gas's mips_cpu_info_table[]. Please also make sure that
566 MIPS_ISA_LEVEL_SPEC and MIPS_ARCH_FLOAT_SPEC handle all -march
567 options correctly. */
568 const struct mips_cpu_info mips_cpu_info_table[] = {
569 /* Entries for generic ISAs */
570 { "mips1", PROCESSOR_R3000, 1, 0 },
571 { "mips2", PROCESSOR_R6000, 2, 0 },
572 { "mips3", PROCESSOR_R4000, 3, 0 },
573 { "mips4", PROCESSOR_R8000, 4, 0 },
574 /* Prefer not to use branch-likely instructions for generic MIPS32rX
575 and MIPS64rX code. The instructions were officially deprecated
576 in revisions 2 and earlier, but revision 3 is likely to downgrade
577 that to a recommendation to avoid the instructions in code that
578 isn't tuned to a specific processor. */
579 { "mips32", PROCESSOR_4KC, 32, PTF_AVOID_BRANCHLIKELY },
580 { "mips32r2", PROCESSOR_M4K, 33, PTF_AVOID_BRANCHLIKELY },
581 { "mips64", PROCESSOR_5KC, 64, PTF_AVOID_BRANCHLIKELY },
583 /* MIPS I */
584 { "r3000", PROCESSOR_R3000, 1, 0 },
585 { "r2000", PROCESSOR_R3000, 1, 0 }, /* = r3000 */
586 { "r3900", PROCESSOR_R3900, 1, 0 },
588 /* MIPS II */
589 { "r6000", PROCESSOR_R6000, 2, 0 },
591 /* MIPS III */
592 { "r4000", PROCESSOR_R4000, 3, 0 },
593 { "vr4100", PROCESSOR_R4100, 3, 0 },
594 { "vr4111", PROCESSOR_R4111, 3, 0 },
595 { "vr4120", PROCESSOR_R4120, 3, 0 },
596 { "vr4130", PROCESSOR_R4130, 3, 0 },
597 { "vr4300", PROCESSOR_R4300, 3, 0 },
598 { "r4400", PROCESSOR_R4000, 3, 0 }, /* = r4000 */
599 { "r4600", PROCESSOR_R4600, 3, 0 },
600 { "orion", PROCESSOR_R4600, 3, 0 }, /* = r4600 */
601 { "r4650", PROCESSOR_R4650, 3, 0 },
603 /* MIPS IV */
604 { "r8000", PROCESSOR_R8000, 4, 0 },
605 { "vr5000", PROCESSOR_R5000, 4, 0 },
606 { "vr5400", PROCESSOR_R5400, 4, 0 },
607 { "vr5500", PROCESSOR_R5500, 4, PTF_AVOID_BRANCHLIKELY },
608 { "rm7000", PROCESSOR_R7000, 4, 0 },
609 { "rm9000", PROCESSOR_R9000, 4, 0 },
611 /* MIPS32 */
612 { "4kc", PROCESSOR_4KC, 32, 0 },
613 { "4km", PROCESSOR_4KC, 32, 0 }, /* = 4kc */
614 { "4kp", PROCESSOR_4KP, 32, 0 },
615 { "4ksc", PROCESSOR_4KC, 32, 0 },
617 /* MIPS32 Release 2 */
618 { "m4k", PROCESSOR_M4K, 33, 0 },
619 { "4kec", PROCESSOR_4KC, 33, 0 },
620 { "4kem", PROCESSOR_4KC, 33, 0 },
621 { "4kep", PROCESSOR_4KP, 33, 0 },
622 { "4ksd", PROCESSOR_4KC, 33, 0 },
624 { "24kc", PROCESSOR_24KC, 33, 0 },
625 { "24kf2_1", PROCESSOR_24KF2_1, 33, 0 },
626 { "24kf", PROCESSOR_24KF2_1, 33, 0 },
627 { "24kf1_1", PROCESSOR_24KF1_1, 33, 0 },
628 { "24kfx", PROCESSOR_24KF1_1, 33, 0 },
629 { "24kx", PROCESSOR_24KF1_1, 33, 0 },
631 { "24kec", PROCESSOR_24KC, 33, 0 }, /* 24K with DSP */
632 { "24kef2_1", PROCESSOR_24KF2_1, 33, 0 },
633 { "24kef", PROCESSOR_24KF2_1, 33, 0 },
634 { "24kef1_1", PROCESSOR_24KF1_1, 33, 0 },
635 { "24kefx", PROCESSOR_24KF1_1, 33, 0 },
636 { "24kex", PROCESSOR_24KF1_1, 33, 0 },
638 { "34kc", PROCESSOR_24KC, 33, 0 }, /* 34K with MT/DSP */
639 { "34kf2_1", PROCESSOR_24KF2_1, 33, 0 },
640 { "34kf", PROCESSOR_24KF2_1, 33, 0 },
641 { "34kf1_1", PROCESSOR_24KF1_1, 33, 0 },
642 { "34kfx", PROCESSOR_24KF1_1, 33, 0 },
643 { "34kx", PROCESSOR_24KF1_1, 33, 0 },
645 { "74kc", PROCESSOR_74KC, 33, 0 }, /* 74K with DSPr2 */
646 { "74kf2_1", PROCESSOR_74KF2_1, 33, 0 },
647 { "74kf", PROCESSOR_74KF2_1, 33, 0 },
648 { "74kf1_1", PROCESSOR_74KF1_1, 33, 0 },
649 { "74kfx", PROCESSOR_74KF1_1, 33, 0 },
650 { "74kx", PROCESSOR_74KF1_1, 33, 0 },
651 { "74kf3_2", PROCESSOR_74KF3_2, 33, 0 },
653 /* MIPS64 */
654 { "5kc", PROCESSOR_5KC, 64, 0 },
655 { "5kf", PROCESSOR_5KF, 64, 0 },
656 { "20kc", PROCESSOR_20KC, 64, PTF_AVOID_BRANCHLIKELY },
657 { "sb1", PROCESSOR_SB1, 64, PTF_AVOID_BRANCHLIKELY },
658 { "sb1a", PROCESSOR_SB1A, 64, PTF_AVOID_BRANCHLIKELY },
659 { "sr71000", PROCESSOR_SR71000, 64, PTF_AVOID_BRANCHLIKELY },
662 /* Default costs. If these are used for a processor we should look
663 up the actual costs. */
664 #define DEFAULT_COSTS COSTS_N_INSNS (6), /* fp_add */ \
665 COSTS_N_INSNS (7), /* fp_mult_sf */ \
666 COSTS_N_INSNS (8), /* fp_mult_df */ \
667 COSTS_N_INSNS (23), /* fp_div_sf */ \
668 COSTS_N_INSNS (36), /* fp_div_df */ \
669 COSTS_N_INSNS (10), /* int_mult_si */ \
670 COSTS_N_INSNS (10), /* int_mult_di */ \
671 COSTS_N_INSNS (69), /* int_div_si */ \
672 COSTS_N_INSNS (69), /* int_div_di */ \
673 2, /* branch_cost */ \
674 4 /* memory_latency */
676 /* Need to replace these with the costs of calling the appropriate
677 libgcc routine. */
678 #define SOFT_FP_COSTS COSTS_N_INSNS (256), /* fp_add */ \
679 COSTS_N_INSNS (256), /* fp_mult_sf */ \
680 COSTS_N_INSNS (256), /* fp_mult_df */ \
681 COSTS_N_INSNS (256), /* fp_div_sf */ \
682 COSTS_N_INSNS (256) /* fp_div_df */
684 static struct mips_rtx_cost_data const mips_rtx_cost_optimize_size =
686 COSTS_N_INSNS (1), /* fp_add */
687 COSTS_N_INSNS (1), /* fp_mult_sf */
688 COSTS_N_INSNS (1), /* fp_mult_df */
689 COSTS_N_INSNS (1), /* fp_div_sf */
690 COSTS_N_INSNS (1), /* fp_div_df */
691 COSTS_N_INSNS (1), /* int_mult_si */
692 COSTS_N_INSNS (1), /* int_mult_di */
693 COSTS_N_INSNS (1), /* int_div_si */
694 COSTS_N_INSNS (1), /* int_div_di */
695 2, /* branch_cost */
696 4 /* memory_latency */
699 static struct mips_rtx_cost_data const mips_rtx_cost_data[PROCESSOR_MAX] =
701 { /* R3000 */
702 COSTS_N_INSNS (2), /* fp_add */
703 COSTS_N_INSNS (4), /* fp_mult_sf */
704 COSTS_N_INSNS (5), /* fp_mult_df */
705 COSTS_N_INSNS (12), /* fp_div_sf */
706 COSTS_N_INSNS (19), /* fp_div_df */
707 COSTS_N_INSNS (12), /* int_mult_si */
708 COSTS_N_INSNS (12), /* int_mult_di */
709 COSTS_N_INSNS (35), /* int_div_si */
710 COSTS_N_INSNS (35), /* int_div_di */
711 1, /* branch_cost */
712 4 /* memory_latency */
715 { /* 4KC */
716 SOFT_FP_COSTS,
717 COSTS_N_INSNS (6), /* int_mult_si */
718 COSTS_N_INSNS (6), /* int_mult_di */
719 COSTS_N_INSNS (36), /* int_div_si */
720 COSTS_N_INSNS (36), /* int_div_di */
721 1, /* branch_cost */
722 4 /* memory_latency */
724 { /* 4KP */
725 SOFT_FP_COSTS,
726 COSTS_N_INSNS (36), /* int_mult_si */
727 COSTS_N_INSNS (36), /* int_mult_di */
728 COSTS_N_INSNS (37), /* int_div_si */
729 COSTS_N_INSNS (37), /* int_div_di */
730 1, /* branch_cost */
731 4 /* memory_latency */
733 { /* 5KC */
734 SOFT_FP_COSTS,
735 COSTS_N_INSNS (4), /* int_mult_si */
736 COSTS_N_INSNS (11), /* int_mult_di */
737 COSTS_N_INSNS (36), /* int_div_si */
738 COSTS_N_INSNS (68), /* int_div_di */
739 1, /* branch_cost */
740 4 /* memory_latency */
742 { /* 5KF */
743 COSTS_N_INSNS (4), /* fp_add */
744 COSTS_N_INSNS (4), /* fp_mult_sf */
745 COSTS_N_INSNS (5), /* fp_mult_df */
746 COSTS_N_INSNS (17), /* fp_div_sf */
747 COSTS_N_INSNS (32), /* fp_div_df */
748 COSTS_N_INSNS (4), /* int_mult_si */
749 COSTS_N_INSNS (11), /* int_mult_di */
750 COSTS_N_INSNS (36), /* int_div_si */
751 COSTS_N_INSNS (68), /* int_div_di */
752 1, /* branch_cost */
753 4 /* memory_latency */
755 { /* 20KC */
756 COSTS_N_INSNS (4), /* fp_add */
757 COSTS_N_INSNS (4), /* fp_mult_sf */
758 COSTS_N_INSNS (5), /* fp_mult_df */
759 COSTS_N_INSNS (17), /* fp_div_sf */
760 COSTS_N_INSNS (32), /* fp_div_df */
761 COSTS_N_INSNS (4), /* int_mult_si */
762 COSTS_N_INSNS (7), /* int_mult_di */
763 COSTS_N_INSNS (42), /* int_div_si */
764 COSTS_N_INSNS (72), /* int_div_di */
765 1, /* branch_cost */
766 4 /* memory_latency */
768 { /* 24KC */
769 SOFT_FP_COSTS,
770 COSTS_N_INSNS (5), /* int_mult_si */
771 COSTS_N_INSNS (5), /* int_mult_di */
772 COSTS_N_INSNS (41), /* int_div_si */
773 COSTS_N_INSNS (41), /* int_div_di */
774 1, /* branch_cost */
775 4 /* memory_latency */
777 { /* 24KF2_1 */
778 COSTS_N_INSNS (8), /* fp_add */
779 COSTS_N_INSNS (8), /* fp_mult_sf */
780 COSTS_N_INSNS (10), /* fp_mult_df */
781 COSTS_N_INSNS (34), /* fp_div_sf */
782 COSTS_N_INSNS (64), /* fp_div_df */
783 COSTS_N_INSNS (5), /* int_mult_si */
784 COSTS_N_INSNS (5), /* int_mult_di */
785 COSTS_N_INSNS (41), /* int_div_si */
786 COSTS_N_INSNS (41), /* int_div_di */
787 1, /* branch_cost */
788 4 /* memory_latency */
790 { /* 24KF1_1 */
791 COSTS_N_INSNS (4), /* fp_add */
792 COSTS_N_INSNS (4), /* fp_mult_sf */
793 COSTS_N_INSNS (5), /* fp_mult_df */
794 COSTS_N_INSNS (17), /* fp_div_sf */
795 COSTS_N_INSNS (32), /* fp_div_df */
796 COSTS_N_INSNS (5), /* int_mult_si */
797 COSTS_N_INSNS (5), /* int_mult_di */
798 COSTS_N_INSNS (41), /* int_div_si */
799 COSTS_N_INSNS (41), /* int_div_di */
800 1, /* branch_cost */
801 4 /* memory_latency */
803 { /* 74KC */
804 SOFT_FP_COSTS,
805 COSTS_N_INSNS (5), /* int_mult_si */
806 COSTS_N_INSNS (5), /* int_mult_di */
807 COSTS_N_INSNS (41), /* int_div_si */
808 COSTS_N_INSNS (41), /* int_div_di */
809 1, /* branch_cost */
810 4 /* memory_latency */
812 { /* 74KF2_1 */
813 COSTS_N_INSNS (8), /* fp_add */
814 COSTS_N_INSNS (8), /* fp_mult_sf */
815 COSTS_N_INSNS (10), /* fp_mult_df */
816 COSTS_N_INSNS (34), /* fp_div_sf */
817 COSTS_N_INSNS (64), /* fp_div_df */
818 COSTS_N_INSNS (5), /* int_mult_si */
819 COSTS_N_INSNS (5), /* int_mult_di */
820 COSTS_N_INSNS (41), /* int_div_si */
821 COSTS_N_INSNS (41), /* int_div_di */
822 1, /* branch_cost */
823 4 /* memory_latency */
825 { /* 74KF1_1 */
826 COSTS_N_INSNS (4), /* fp_add */
827 COSTS_N_INSNS (4), /* fp_mult_sf */
828 COSTS_N_INSNS (5), /* fp_mult_df */
829 COSTS_N_INSNS (17), /* fp_div_sf */
830 COSTS_N_INSNS (32), /* fp_div_df */
831 COSTS_N_INSNS (5), /* int_mult_si */
832 COSTS_N_INSNS (5), /* int_mult_di */
833 COSTS_N_INSNS (41), /* int_div_si */
834 COSTS_N_INSNS (41), /* int_div_di */
835 1, /* branch_cost */
836 4 /* memory_latency */
838 { /* 74KF3_2 */
839 COSTS_N_INSNS (6), /* fp_add */
840 COSTS_N_INSNS (6), /* fp_mult_sf */
841 COSTS_N_INSNS (7), /* fp_mult_df */
842 COSTS_N_INSNS (25), /* fp_div_sf */
843 COSTS_N_INSNS (48), /* fp_div_df */
844 COSTS_N_INSNS (5), /* int_mult_si */
845 COSTS_N_INSNS (5), /* int_mult_di */
846 COSTS_N_INSNS (41), /* int_div_si */
847 COSTS_N_INSNS (41), /* int_div_di */
848 1, /* branch_cost */
849 4 /* memory_latency */
851 { /* M4k */
852 DEFAULT_COSTS
854 { /* R3900 */
855 COSTS_N_INSNS (2), /* fp_add */
856 COSTS_N_INSNS (4), /* fp_mult_sf */
857 COSTS_N_INSNS (5), /* fp_mult_df */
858 COSTS_N_INSNS (12), /* fp_div_sf */
859 COSTS_N_INSNS (19), /* fp_div_df */
860 COSTS_N_INSNS (2), /* int_mult_si */
861 COSTS_N_INSNS (2), /* int_mult_di */
862 COSTS_N_INSNS (35), /* int_div_si */
863 COSTS_N_INSNS (35), /* int_div_di */
864 1, /* branch_cost */
865 4 /* memory_latency */
867 { /* R6000 */
868 COSTS_N_INSNS (3), /* fp_add */
869 COSTS_N_INSNS (5), /* fp_mult_sf */
870 COSTS_N_INSNS (6), /* fp_mult_df */
871 COSTS_N_INSNS (15), /* fp_div_sf */
872 COSTS_N_INSNS (16), /* fp_div_df */
873 COSTS_N_INSNS (17), /* int_mult_si */
874 COSTS_N_INSNS (17), /* int_mult_di */
875 COSTS_N_INSNS (38), /* int_div_si */
876 COSTS_N_INSNS (38), /* int_div_di */
877 2, /* branch_cost */
878 6 /* memory_latency */
880 { /* R4000 */
881 COSTS_N_INSNS (6), /* fp_add */
882 COSTS_N_INSNS (7), /* fp_mult_sf */
883 COSTS_N_INSNS (8), /* fp_mult_df */
884 COSTS_N_INSNS (23), /* fp_div_sf */
885 COSTS_N_INSNS (36), /* fp_div_df */
886 COSTS_N_INSNS (10), /* int_mult_si */
887 COSTS_N_INSNS (10), /* int_mult_di */
888 COSTS_N_INSNS (69), /* int_div_si */
889 COSTS_N_INSNS (69), /* int_div_di */
890 2, /* branch_cost */
891 6 /* memory_latency */
893 { /* R4100 */
894 DEFAULT_COSTS
896 { /* R4111 */
897 DEFAULT_COSTS
899 { /* R4120 */
900 DEFAULT_COSTS
902 { /* R4130 */
903 /* The only costs that appear to be updated here are
904 integer multiplication. */
905 SOFT_FP_COSTS,
906 COSTS_N_INSNS (4), /* int_mult_si */
907 COSTS_N_INSNS (6), /* int_mult_di */
908 COSTS_N_INSNS (69), /* int_div_si */
909 COSTS_N_INSNS (69), /* int_div_di */
910 1, /* branch_cost */
911 4 /* memory_latency */
913 { /* R4300 */
914 DEFAULT_COSTS
916 { /* R4600 */
917 DEFAULT_COSTS
919 { /* R4650 */
920 DEFAULT_COSTS
922 { /* R5000 */
923 COSTS_N_INSNS (6), /* fp_add */
924 COSTS_N_INSNS (4), /* fp_mult_sf */
925 COSTS_N_INSNS (5), /* fp_mult_df */
926 COSTS_N_INSNS (23), /* fp_div_sf */
927 COSTS_N_INSNS (36), /* fp_div_df */
928 COSTS_N_INSNS (5), /* int_mult_si */
929 COSTS_N_INSNS (5), /* int_mult_di */
930 COSTS_N_INSNS (36), /* int_div_si */
931 COSTS_N_INSNS (36), /* int_div_di */
932 1, /* branch_cost */
933 4 /* memory_latency */
935 { /* R5400 */
936 COSTS_N_INSNS (6), /* fp_add */
937 COSTS_N_INSNS (5), /* fp_mult_sf */
938 COSTS_N_INSNS (6), /* fp_mult_df */
939 COSTS_N_INSNS (30), /* fp_div_sf */
940 COSTS_N_INSNS (59), /* fp_div_df */
941 COSTS_N_INSNS (3), /* int_mult_si */
942 COSTS_N_INSNS (4), /* int_mult_di */
943 COSTS_N_INSNS (42), /* int_div_si */
944 COSTS_N_INSNS (74), /* int_div_di */
945 1, /* branch_cost */
946 4 /* memory_latency */
948 { /* R5500 */
949 COSTS_N_INSNS (6), /* fp_add */
950 COSTS_N_INSNS (5), /* fp_mult_sf */
951 COSTS_N_INSNS (6), /* fp_mult_df */
952 COSTS_N_INSNS (30), /* fp_div_sf */
953 COSTS_N_INSNS (59), /* fp_div_df */
954 COSTS_N_INSNS (5), /* int_mult_si */
955 COSTS_N_INSNS (9), /* int_mult_di */
956 COSTS_N_INSNS (42), /* int_div_si */
957 COSTS_N_INSNS (74), /* int_div_di */
958 1, /* branch_cost */
959 4 /* memory_latency */
961 { /* R7000 */
962 /* The only costs that are changed here are
963 integer multiplication. */
964 COSTS_N_INSNS (6), /* fp_add */
965 COSTS_N_INSNS (7), /* fp_mult_sf */
966 COSTS_N_INSNS (8), /* fp_mult_df */
967 COSTS_N_INSNS (23), /* fp_div_sf */
968 COSTS_N_INSNS (36), /* fp_div_df */
969 COSTS_N_INSNS (5), /* int_mult_si */
970 COSTS_N_INSNS (9), /* int_mult_di */
971 COSTS_N_INSNS (69), /* int_div_si */
972 COSTS_N_INSNS (69), /* int_div_di */
973 1, /* branch_cost */
974 4 /* memory_latency */
976 { /* R8000 */
977 DEFAULT_COSTS
979 { /* R9000 */
980 /* The only costs that are changed here are
981 integer multiplication. */
982 COSTS_N_INSNS (6), /* fp_add */
983 COSTS_N_INSNS (7), /* fp_mult_sf */
984 COSTS_N_INSNS (8), /* fp_mult_df */
985 COSTS_N_INSNS (23), /* fp_div_sf */
986 COSTS_N_INSNS (36), /* fp_div_df */
987 COSTS_N_INSNS (3), /* int_mult_si */
988 COSTS_N_INSNS (8), /* int_mult_di */
989 COSTS_N_INSNS (69), /* int_div_si */
990 COSTS_N_INSNS (69), /* int_div_di */
991 1, /* branch_cost */
992 4 /* memory_latency */
994 { /* SB1 */
995 /* These costs are the same as the SB-1A below. */
996 COSTS_N_INSNS (4), /* fp_add */
997 COSTS_N_INSNS (4), /* fp_mult_sf */
998 COSTS_N_INSNS (4), /* fp_mult_df */
999 COSTS_N_INSNS (24), /* fp_div_sf */
1000 COSTS_N_INSNS (32), /* fp_div_df */
1001 COSTS_N_INSNS (3), /* int_mult_si */
1002 COSTS_N_INSNS (4), /* int_mult_di */
1003 COSTS_N_INSNS (36), /* int_div_si */
1004 COSTS_N_INSNS (68), /* int_div_di */
1005 1, /* branch_cost */
1006 4 /* memory_latency */
1008 { /* SB1-A */
1009 /* These costs are the same as the SB-1 above. */
1010 COSTS_N_INSNS (4), /* fp_add */
1011 COSTS_N_INSNS (4), /* fp_mult_sf */
1012 COSTS_N_INSNS (4), /* fp_mult_df */
1013 COSTS_N_INSNS (24), /* fp_div_sf */
1014 COSTS_N_INSNS (32), /* fp_div_df */
1015 COSTS_N_INSNS (3), /* int_mult_si */
1016 COSTS_N_INSNS (4), /* int_mult_di */
1017 COSTS_N_INSNS (36), /* int_div_si */
1018 COSTS_N_INSNS (68), /* int_div_di */
1019 1, /* branch_cost */
1020 4 /* memory_latency */
1022 { /* SR71000 */
1023 DEFAULT_COSTS
1027 /* Use a hash table to keep track of implicit mips16/nomips16 attributes
1028 for -mflip_mips16. It maps decl names onto a boolean mode setting. */
1030 struct mflip_mips16_entry GTY (()) {
1031 const char *name;
1032 bool mips16_p;
1034 static GTY ((param_is (struct mflip_mips16_entry))) htab_t mflip_mips16_htab;
1036 /* Hash table callbacks for mflip_mips16_htab. */
1038 static hashval_t
1039 mflip_mips16_htab_hash (const void *entry)
1041 return htab_hash_string (((const struct mflip_mips16_entry *) entry)->name);
1044 static int
1045 mflip_mips16_htab_eq (const void *entry, const void *name)
1047 return strcmp (((const struct mflip_mips16_entry *) entry)->name,
1048 (const char *) name) == 0;
1051 static GTY(()) int mips16_flipper;
1053 /* DECL is a function that needs a default "mips16" or "nomips16" attribute
1054 for -mflip-mips16. Return true if it should use "mips16" and false if
1055 it should use "nomips16". */
1057 static bool
1058 mflip_mips16_use_mips16_p (tree decl)
1060 struct mflip_mips16_entry *entry;
1061 const char *name;
1062 hashval_t hash;
1063 void **slot;
1065 /* Use the opposite of the command-line setting for anonymous decls. */
1066 if (!DECL_NAME (decl))
1067 return !mips_base_mips16;
1069 if (!mflip_mips16_htab)
1070 mflip_mips16_htab = htab_create_ggc (37, mflip_mips16_htab_hash,
1071 mflip_mips16_htab_eq, NULL);
1073 name = IDENTIFIER_POINTER (DECL_NAME (decl));
1074 hash = htab_hash_string (name);
1075 slot = htab_find_slot_with_hash (mflip_mips16_htab, name, hash, INSERT);
1076 entry = (struct mflip_mips16_entry *) *slot;
1077 if (!entry)
1079 mips16_flipper = !mips16_flipper;
1080 entry = GGC_NEW (struct mflip_mips16_entry);
1081 entry->name = name;
1082 entry->mips16_p = mips16_flipper ? !mips_base_mips16 : mips_base_mips16;
1083 *slot = entry;
1085 return entry->mips16_p;
1088 /* Predicates to test for presence of "near" and "far"/"long_call"
1089 attributes on the given TYPE. */
1091 static bool
1092 mips_near_type_p (const_tree type)
1094 return lookup_attribute ("near", TYPE_ATTRIBUTES (type)) != NULL;
1097 static bool
1098 mips_far_type_p (const_tree type)
1100 return (lookup_attribute ("long_call", TYPE_ATTRIBUTES (type)) != NULL
1101 || lookup_attribute ("far", TYPE_ATTRIBUTES (type)) != NULL);
1104 /* Similar predicates for "mips16"/"nomips16" attributes. */
1106 static bool
1107 mips_mips16_decl_p (const_tree decl)
1109 return lookup_attribute ("mips16", DECL_ATTRIBUTES (decl)) != NULL;
1112 static bool
1113 mips_nomips16_decl_p (const_tree decl)
1115 return lookup_attribute ("nomips16", DECL_ATTRIBUTES (decl)) != NULL;
1118 /* Return true if function DECL is a MIPS16 function. Return the ambient
1119 setting if DECL is null. */
1121 static bool
1122 mips_use_mips16_mode_p (tree decl)
1124 if (decl)
1126 /* Nested functions must use the same frame pointer as their
1127 parent and must therefore use the same ISA mode. */
1128 tree parent = decl_function_context (decl);
1129 if (parent)
1130 decl = parent;
1131 if (mips_mips16_decl_p (decl))
1132 return true;
1133 if (mips_nomips16_decl_p (decl))
1134 return false;
1136 return mips_base_mips16;
1139 /* Return 0 if the attributes for two types are incompatible, 1 if they
1140 are compatible, and 2 if they are nearly compatible (which causes a
1141 warning to be generated). */
1143 static int
1144 mips_comp_type_attributes (const_tree type1, const_tree type2)
1146 /* Check for mismatch of non-default calling convention. */
1147 if (TREE_CODE (type1) != FUNCTION_TYPE)
1148 return 1;
1150 /* Disallow mixed near/far attributes. */
1151 if (mips_far_type_p (type1) && mips_near_type_p (type2))
1152 return 0;
1153 if (mips_near_type_p (type1) && mips_far_type_p (type2))
1154 return 0;
1156 return 1;
1159 /* Implement TARGET_INSERT_ATTRIBUTES. */
1161 static void
1162 mips_insert_attributes (tree decl, tree *attributes)
1164 const char *name;
1165 bool mips16_p, nomips16_p;
1167 /* Check for "mips16" and "nomips16" attributes. */
1168 mips16_p = lookup_attribute ("mips16", *attributes) != NULL;
1169 nomips16_p = lookup_attribute ("nomips16", *attributes) != NULL;
1170 if (TREE_CODE (decl) != FUNCTION_DECL)
1172 if (mips16_p)
1173 error ("%qs attribute only applies to functions", "mips16");
1174 if (nomips16_p)
1175 error ("%qs attribute only applies to functions", "nomips16");
1177 else
1179 mips16_p |= mips_mips16_decl_p (decl);
1180 nomips16_p |= mips_nomips16_decl_p (decl);
1181 if (mips16_p || nomips16_p)
1183 /* DECL cannot be simultaneously mips16 and nomips16. */
1184 if (mips16_p && nomips16_p)
1185 error ("%qs cannot have both %<mips16%> and "
1186 "%<nomips16%> attributes",
1187 IDENTIFIER_POINTER (DECL_NAME (decl)));
1189 else if (TARGET_FLIP_MIPS16 && !DECL_ARTIFICIAL (decl))
1191 /* Implement -mflip-mips16. If DECL has neither a "nomips16" nor a
1192 "mips16" attribute, arbitrarily pick one. We must pick the same
1193 setting for duplicate declarations of a function. */
1194 name = mflip_mips16_use_mips16_p (decl) ? "mips16" : "nomips16";
1195 *attributes = tree_cons (get_identifier (name), NULL, *attributes);
1200 /* Implement TARGET_MERGE_DECL_ATTRIBUTES. */
1202 static tree
1203 mips_merge_decl_attributes (tree olddecl, tree newdecl)
1205 /* The decls' "mips16" and "nomips16" attributes must match exactly. */
1206 if (mips_mips16_decl_p (olddecl) != mips_mips16_decl_p (newdecl))
1207 error ("%qs redeclared with conflicting %qs attributes",
1208 IDENTIFIER_POINTER (DECL_NAME (newdecl)), "mips16");
1209 if (mips_nomips16_decl_p (olddecl) != mips_nomips16_decl_p (newdecl))
1210 error ("%qs redeclared with conflicting %qs attributes",
1211 IDENTIFIER_POINTER (DECL_NAME (newdecl)), "nomips16");
1213 return merge_attributes (DECL_ATTRIBUTES (olddecl),
1214 DECL_ATTRIBUTES (newdecl));
1217 /* If X is a PLUS of a CONST_INT, return the two terms in *BASE_PTR
1218 and *OFFSET_PTR. Return X in *BASE_PTR and 0 in *OFFSET_PTR otherwise. */
1220 static void
1221 mips_split_plus (rtx x, rtx *base_ptr, HOST_WIDE_INT *offset_ptr)
1223 if (GET_CODE (x) == PLUS && GET_CODE (XEXP (x, 1)) == CONST_INT)
1225 *base_ptr = XEXP (x, 0);
1226 *offset_ptr = INTVAL (XEXP (x, 1));
1228 else
1230 *base_ptr = x;
1231 *offset_ptr = 0;
1235 static unsigned int mips_build_integer (struct mips_integer_op *,
1236 unsigned HOST_WIDE_INT);
1238 /* Subroutine of mips_build_integer (with the same interface).
1239 Assume that the final action in the sequence should be a left shift. */
1241 static unsigned int
1242 mips_build_shift (struct mips_integer_op *codes, HOST_WIDE_INT value)
1244 unsigned int i, shift;
1246 /* Shift VALUE right until its lowest bit is set. Shift arithmetically
1247 since signed numbers are easier to load than unsigned ones. */
1248 shift = 0;
1249 while ((value & 1) == 0)
1250 value /= 2, shift++;
1252 i = mips_build_integer (codes, value);
1253 codes[i].code = ASHIFT;
1254 codes[i].value = shift;
1255 return i + 1;
1259 /* As for mips_build_shift, but assume that the final action will be
1260 an IOR or PLUS operation. */
1262 static unsigned int
1263 mips_build_lower (struct mips_integer_op *codes, unsigned HOST_WIDE_INT value)
1265 unsigned HOST_WIDE_INT high;
1266 unsigned int i;
1268 high = value & ~(unsigned HOST_WIDE_INT) 0xffff;
1269 if (!LUI_OPERAND (high) && (value & 0x18000) == 0x18000)
1271 /* The constant is too complex to load with a simple lui/ori pair
1272 so our goal is to clear as many trailing zeros as possible.
1273 In this case, we know bit 16 is set and that the low 16 bits
1274 form a negative number. If we subtract that number from VALUE,
1275 we will clear at least the lowest 17 bits, maybe more. */
1276 i = mips_build_integer (codes, CONST_HIGH_PART (value));
1277 codes[i].code = PLUS;
1278 codes[i].value = CONST_LOW_PART (value);
1280 else
1282 i = mips_build_integer (codes, high);
1283 codes[i].code = IOR;
1284 codes[i].value = value & 0xffff;
1286 return i + 1;
1290 /* Fill CODES with a sequence of rtl operations to load VALUE.
1291 Return the number of operations needed. */
1293 static unsigned int
1294 mips_build_integer (struct mips_integer_op *codes,
1295 unsigned HOST_WIDE_INT value)
1297 if (SMALL_OPERAND (value)
1298 || SMALL_OPERAND_UNSIGNED (value)
1299 || LUI_OPERAND (value))
1301 /* The value can be loaded with a single instruction. */
1302 codes[0].code = UNKNOWN;
1303 codes[0].value = value;
1304 return 1;
1306 else if ((value & 1) != 0 || LUI_OPERAND (CONST_HIGH_PART (value)))
1308 /* Either the constant is a simple LUI/ORI combination or its
1309 lowest bit is set. We don't want to shift in this case. */
1310 return mips_build_lower (codes, value);
1312 else if ((value & 0xffff) == 0)
1314 /* The constant will need at least three actions. The lowest
1315 16 bits are clear, so the final action will be a shift. */
1316 return mips_build_shift (codes, value);
1318 else
1320 /* The final action could be a shift, add or inclusive OR.
1321 Rather than use a complex condition to select the best
1322 approach, try both mips_build_shift and mips_build_lower
1323 and pick the one that gives the shortest sequence.
1324 Note that this case is only used once per constant. */
1325 struct mips_integer_op alt_codes[MIPS_MAX_INTEGER_OPS];
1326 unsigned int cost, alt_cost;
1328 cost = mips_build_shift (codes, value);
1329 alt_cost = mips_build_lower (alt_codes, value);
1330 if (alt_cost < cost)
1332 memcpy (codes, alt_codes, alt_cost * sizeof (codes[0]));
1333 cost = alt_cost;
1335 return cost;
1339 /* Return true if X is a thread-local symbol. */
1341 static bool
1342 mips_tls_operand_p (rtx x)
1344 return GET_CODE (x) == SYMBOL_REF && SYMBOL_REF_TLS_MODEL (x) != 0;
1347 /* Return true if SYMBOL_REF X is associated with a global symbol
1348 (in the STB_GLOBAL sense). */
1350 static bool
1351 mips_global_symbol_p (const_rtx x)
1353 const_tree const decl = SYMBOL_REF_DECL (x);
1355 if (!decl)
1356 return !SYMBOL_REF_LOCAL_P (x);
1358 /* Weakref symbols are not TREE_PUBLIC, but their targets are global
1359 or weak symbols. Relocations in the object file will be against
1360 the target symbol, so it's that symbol's binding that matters here. */
1361 return DECL_P (decl) && (TREE_PUBLIC (decl) || DECL_WEAK (decl));
1364 /* Return true if SYMBOL_REF X binds locally. */
1366 static bool
1367 mips_symbol_binds_local_p (const_rtx x)
1369 return (SYMBOL_REF_DECL (x)
1370 ? targetm.binds_local_p (SYMBOL_REF_DECL (x))
1371 : SYMBOL_REF_LOCAL_P (x));
1374 /* Return true if rtx constants of mode MODE should be put into a small
1375 data section. */
1377 static bool
1378 mips_rtx_constant_in_small_data_p (enum machine_mode mode)
1380 return (!TARGET_EMBEDDED_DATA
1381 && TARGET_LOCAL_SDATA
1382 && GET_MODE_SIZE (mode) <= mips_section_threshold);
1385 /* Return true if X should not be moved directly into register $25.
1386 We need this because many versions of GAS will treat "la $25,foo" as
1387 part of a call sequence and so allow a global "foo" to be lazily bound. */
1389 bool
1390 mips_dangerous_for_la25_p (rtx x)
1392 return (!TARGET_EXPLICIT_RELOCS
1393 && TARGET_USE_GOT
1394 && GET_CODE (x) == SYMBOL_REF
1395 && mips_global_symbol_p (x));
1398 /* Return the method that should be used to access SYMBOL_REF or
1399 LABEL_REF X in context CONTEXT. */
1401 static enum mips_symbol_type
1402 mips_classify_symbol (const_rtx x, enum mips_symbol_context context)
1404 if (TARGET_RTP_PIC)
1405 return SYMBOL_GOT_DISP;
1407 if (GET_CODE (x) == LABEL_REF)
1409 /* LABEL_REFs are used for jump tables as well as text labels.
1410 Only return SYMBOL_PC_RELATIVE if we know the label is in
1411 the text section. */
1412 if (TARGET_MIPS16_SHORT_JUMP_TABLES)
1413 return SYMBOL_PC_RELATIVE;
1414 if (TARGET_ABICALLS && !TARGET_ABSOLUTE_ABICALLS)
1415 return SYMBOL_GOT_PAGE_OFST;
1416 return SYMBOL_ABSOLUTE;
1419 gcc_assert (GET_CODE (x) == SYMBOL_REF);
1421 if (SYMBOL_REF_TLS_MODEL (x))
1422 return SYMBOL_TLS;
1424 if (CONSTANT_POOL_ADDRESS_P (x))
1426 if (TARGET_MIPS16_TEXT_LOADS)
1427 return SYMBOL_PC_RELATIVE;
1429 if (TARGET_MIPS16_PCREL_LOADS && context == SYMBOL_CONTEXT_MEM)
1430 return SYMBOL_PC_RELATIVE;
1432 if (mips_rtx_constant_in_small_data_p (get_pool_mode (x)))
1433 return SYMBOL_GP_RELATIVE;
1436 /* Do not use small-data accesses for weak symbols; they may end up
1437 being zero. */
1438 if (TARGET_GPOPT
1439 && SYMBOL_REF_SMALL_P (x)
1440 && !SYMBOL_REF_WEAK (x))
1441 return SYMBOL_GP_RELATIVE;
1443 /* Don't use GOT accesses for locally-binding symbols when -mno-shared
1444 is in effect. */
1445 if (TARGET_ABICALLS
1446 && !(TARGET_ABSOLUTE_ABICALLS && mips_symbol_binds_local_p (x)))
1448 /* There are three cases to consider:
1450 - o32 PIC (either with or without explicit relocs)
1451 - n32/n64 PIC without explicit relocs
1452 - n32/n64 PIC with explicit relocs
1454 In the first case, both local and global accesses will use an
1455 R_MIPS_GOT16 relocation. We must correctly predict which of
1456 the two semantics (local or global) the assembler and linker
1457 will apply. The choice depends on the symbol's binding rather
1458 than its visibility.
1460 In the second case, the assembler will not use R_MIPS_GOT16
1461 relocations, but it chooses between local and global accesses
1462 in the same way as for o32 PIC.
1464 In the third case we have more freedom since both forms of
1465 access will work for any kind of symbol. However, there seems
1466 little point in doing things differently. */
1467 if (mips_global_symbol_p (x))
1468 return SYMBOL_GOT_DISP;
1470 return SYMBOL_GOT_PAGE_OFST;
1473 if (TARGET_MIPS16_PCREL_LOADS && context != SYMBOL_CONTEXT_CALL)
1474 return SYMBOL_FORCE_TO_MEM;
1475 return SYMBOL_ABSOLUTE;
1478 /* Classify symbolic expression X, given that it appears in context
1479 CONTEXT. */
1481 static enum mips_symbol_type
1482 mips_classify_symbolic_expression (rtx x, enum mips_symbol_context context)
1484 rtx offset;
1486 split_const (x, &x, &offset);
1487 if (UNSPEC_ADDRESS_P (x))
1488 return UNSPEC_ADDRESS_TYPE (x);
1490 return mips_classify_symbol (x, context);
1493 /* Return true if OFFSET is within the range [0, ALIGN), where ALIGN
1494 is the alignment (in bytes) of SYMBOL_REF X. */
1496 static bool
1497 mips_offset_within_alignment_p (rtx x, HOST_WIDE_INT offset)
1499 /* If for some reason we can't get the alignment for the
1500 symbol, initializing this to one means we will only accept
1501 a zero offset. */
1502 HOST_WIDE_INT align = 1;
1503 tree t;
1505 /* Get the alignment of the symbol we're referring to. */
1506 t = SYMBOL_REF_DECL (x);
1507 if (t)
1508 align = DECL_ALIGN_UNIT (t);
1510 return offset >= 0 && offset < align;
1513 /* Return true if X is a symbolic constant that can be used in context
1514 CONTEXT. If it is, store the type of the symbol in *SYMBOL_TYPE. */
1516 bool
1517 mips_symbolic_constant_p (rtx x, enum mips_symbol_context context,
1518 enum mips_symbol_type *symbol_type)
1520 rtx offset;
1522 split_const (x, &x, &offset);
1523 if (UNSPEC_ADDRESS_P (x))
1525 *symbol_type = UNSPEC_ADDRESS_TYPE (x);
1526 x = UNSPEC_ADDRESS (x);
1528 else if (GET_CODE (x) == SYMBOL_REF || GET_CODE (x) == LABEL_REF)
1530 *symbol_type = mips_classify_symbol (x, context);
1531 if (*symbol_type == SYMBOL_TLS)
1532 return false;
1534 else
1535 return false;
1537 if (offset == const0_rtx)
1538 return true;
1540 /* Check whether a nonzero offset is valid for the underlying
1541 relocations. */
1542 switch (*symbol_type)
1544 case SYMBOL_ABSOLUTE:
1545 case SYMBOL_FORCE_TO_MEM:
1546 case SYMBOL_32_HIGH:
1547 case SYMBOL_64_HIGH:
1548 case SYMBOL_64_MID:
1549 case SYMBOL_64_LOW:
1550 /* If the target has 64-bit pointers and the object file only
1551 supports 32-bit symbols, the values of those symbols will be
1552 sign-extended. In this case we can't allow an arbitrary offset
1553 in case the 32-bit value X + OFFSET has a different sign from X. */
1554 if (Pmode == DImode && !ABI_HAS_64BIT_SYMBOLS)
1555 return offset_within_block_p (x, INTVAL (offset));
1557 /* In other cases the relocations can handle any offset. */
1558 return true;
1560 case SYMBOL_PC_RELATIVE:
1561 /* Allow constant pool references to be converted to LABEL+CONSTANT.
1562 In this case, we no longer have access to the underlying constant,
1563 but the original symbol-based access was known to be valid. */
1564 if (GET_CODE (x) == LABEL_REF)
1565 return true;
1567 /* Fall through. */
1569 case SYMBOL_GP_RELATIVE:
1570 /* Make sure that the offset refers to something within the
1571 same object block. This should guarantee that the final
1572 PC- or GP-relative offset is within the 16-bit limit. */
1573 return offset_within_block_p (x, INTVAL (offset));
1575 case SYMBOL_GOT_PAGE_OFST:
1576 case SYMBOL_GOTOFF_PAGE:
1577 /* If the symbol is global, the GOT entry will contain the symbol's
1578 address, and we will apply a 16-bit offset after loading it.
1579 If the symbol is local, the linker should provide enough local
1580 GOT entries for a 16-bit offset, but larger offsets may lead
1581 to GOT overflow. */
1582 return SMALL_INT (offset);
1584 case SYMBOL_TPREL:
1585 case SYMBOL_DTPREL:
1586 /* There is no carry between the HI and LO REL relocations, so the
1587 offset is only valid if we know it won't lead to such a carry. */
1588 return mips_offset_within_alignment_p (x, INTVAL (offset));
1590 case SYMBOL_GOT_DISP:
1591 case SYMBOL_GOTOFF_DISP:
1592 case SYMBOL_GOTOFF_CALL:
1593 case SYMBOL_GOTOFF_LOADGP:
1594 case SYMBOL_TLSGD:
1595 case SYMBOL_TLSLDM:
1596 case SYMBOL_GOTTPREL:
1597 case SYMBOL_TLS:
1598 case SYMBOL_HALF:
1599 return false;
1601 gcc_unreachable ();
1604 /* Like mips_symbol_insns, but treat extended MIPS16 instructions as a
1605 single instruction. We rely on the fact that, in the worst case,
1606 all instructions involved in a MIPS16 address calculation are usually
1607 extended ones. */
1609 static int
1610 mips_symbol_insns_1 (enum mips_symbol_type type, enum machine_mode mode)
1612 switch (type)
1614 case SYMBOL_ABSOLUTE:
1615 /* When using 64-bit symbols, we need 5 preparatory instructions,
1616 such as:
1618 lui $at,%highest(symbol)
1619 daddiu $at,$at,%higher(symbol)
1620 dsll $at,$at,16
1621 daddiu $at,$at,%hi(symbol)
1622 dsll $at,$at,16
1624 The final address is then $at + %lo(symbol). With 32-bit
1625 symbols we just need a preparatory lui for normal mode and
1626 a preparatory "li; sll" for MIPS16. */
1627 return ABI_HAS_64BIT_SYMBOLS ? 6 : TARGET_MIPS16 ? 3 : 2;
1629 case SYMBOL_GP_RELATIVE:
1630 /* Treat GP-relative accesses as taking a single instruction on
1631 MIPS16 too; the copy of $gp can often be shared. */
1632 return 1;
1634 case SYMBOL_PC_RELATIVE:
1635 /* PC-relative constants can be only be used with addiupc,
1636 lwpc and ldpc. */
1637 if (mode == MAX_MACHINE_MODE
1638 || GET_MODE_SIZE (mode) == 4
1639 || GET_MODE_SIZE (mode) == 8)
1640 return 1;
1642 /* The constant must be loaded using addiupc first. */
1643 return 0;
1645 case SYMBOL_FORCE_TO_MEM:
1646 /* LEAs will be converted into constant-pool references by
1647 mips_reorg. */
1648 if (mode == MAX_MACHINE_MODE)
1649 return 1;
1651 /* The constant must be loaded from the constant pool. */
1652 return 0;
1654 case SYMBOL_GOT_DISP:
1655 /* The constant will have to be loaded from the GOT before it
1656 is used in an address. */
1657 if (mode != MAX_MACHINE_MODE)
1658 return 0;
1660 /* Fall through. */
1662 case SYMBOL_GOT_PAGE_OFST:
1663 /* Unless -funit-at-a-time is in effect, we can't be sure whether
1664 the local/global classification is accurate. See override_options
1665 for details.
1667 The worst cases are:
1669 (1) For local symbols when generating o32 or o64 code. The assembler
1670 will use:
1672 lw $at,%got(symbol)
1675 ...and the final address will be $at + %lo(symbol).
1677 (2) For global symbols when -mxgot. The assembler will use:
1679 lui $at,%got_hi(symbol)
1680 (d)addu $at,$at,$gp
1682 ...and the final address will be $at + %got_lo(symbol). */
1683 return 3;
1685 case SYMBOL_GOTOFF_PAGE:
1686 case SYMBOL_GOTOFF_DISP:
1687 case SYMBOL_GOTOFF_CALL:
1688 case SYMBOL_GOTOFF_LOADGP:
1689 case SYMBOL_32_HIGH:
1690 case SYMBOL_64_HIGH:
1691 case SYMBOL_64_MID:
1692 case SYMBOL_64_LOW:
1693 case SYMBOL_TLSGD:
1694 case SYMBOL_TLSLDM:
1695 case SYMBOL_DTPREL:
1696 case SYMBOL_GOTTPREL:
1697 case SYMBOL_TPREL:
1698 case SYMBOL_HALF:
1699 /* A 16-bit constant formed by a single relocation, or a 32-bit
1700 constant formed from a high 16-bit relocation and a low 16-bit
1701 relocation. Use mips_split_p to determine which. */
1702 return !mips_split_p[type] ? 1 : TARGET_MIPS16 ? 3 : 2;
1704 case SYMBOL_TLS:
1705 /* We don't treat a bare TLS symbol as a constant. */
1706 return 0;
1708 gcc_unreachable ();
1711 /* If MODE is MAX_MACHINE_MODE, return the number of instructions needed
1712 to load symbols of type TYPE into a register. Return 0 if the given
1713 type of symbol cannot be used as an immediate operand.
1715 Otherwise, return the number of instructions needed to load or store
1716 values of mode MODE to or from addresses of type TYPE. Return 0 if
1717 the given type of symbol is not valid in addresses.
1719 In both cases, treat extended MIPS16 instructions as two instructions. */
1721 static int
1722 mips_symbol_insns (enum mips_symbol_type type, enum machine_mode mode)
1724 return mips_symbol_insns_1 (type, mode) * (TARGET_MIPS16 ? 2 : 1);
1727 /* Return true if X can not be forced into a constant pool. */
1729 static int
1730 mips_tls_symbol_ref_1 (rtx *x, void *data ATTRIBUTE_UNUSED)
1732 return mips_tls_operand_p (*x);
1735 /* Return true if X can not be forced into a constant pool. */
1737 static bool
1738 mips_cannot_force_const_mem (rtx x)
1740 rtx base, offset;
1742 if (!TARGET_MIPS16)
1744 /* As an optimization, reject constants that mips_legitimize_move
1745 can expand inline.
1747 Suppose we have a multi-instruction sequence that loads constant C
1748 into register R. If R does not get allocated a hard register, and
1749 R is used in an operand that allows both registers and memory
1750 references, reload will consider forcing C into memory and using
1751 one of the instruction's memory alternatives. Returning false
1752 here will force it to use an input reload instead. */
1753 if (GET_CODE (x) == CONST_INT)
1754 return true;
1756 split_const (x, &base, &offset);
1757 if (symbolic_operand (base, VOIDmode) && SMALL_INT (offset))
1758 return true;
1761 if (for_each_rtx (&x, &mips_tls_symbol_ref_1, 0))
1762 return true;
1764 return false;
1767 /* Implement TARGET_USE_BLOCKS_FOR_CONSTANT_P. We can't use blocks for
1768 constants when we're using a per-function constant pool. */
1770 static bool
1771 mips_use_blocks_for_constant_p (enum machine_mode mode ATTRIBUTE_UNUSED,
1772 const_rtx x ATTRIBUTE_UNUSED)
1774 return !TARGET_MIPS16_PCREL_LOADS;
1777 /* This function is used to implement REG_MODE_OK_FOR_BASE_P. */
1780 mips_regno_mode_ok_for_base_p (int regno, enum machine_mode mode, int strict)
1782 if (!HARD_REGISTER_NUM_P (regno))
1784 if (!strict)
1785 return true;
1786 regno = reg_renumber[regno];
1789 /* These fake registers will be eliminated to either the stack or
1790 hard frame pointer, both of which are usually valid base registers.
1791 Reload deals with the cases where the eliminated form isn't valid. */
1792 if (regno == ARG_POINTER_REGNUM || regno == FRAME_POINTER_REGNUM)
1793 return true;
1795 /* In mips16 mode, the stack pointer can only address word and doubleword
1796 values, nothing smaller. There are two problems here:
1798 (a) Instantiating virtual registers can introduce new uses of the
1799 stack pointer. If these virtual registers are valid addresses,
1800 the stack pointer should be too.
1802 (b) Most uses of the stack pointer are not made explicit until
1803 FRAME_POINTER_REGNUM and ARG_POINTER_REGNUM have been eliminated.
1804 We don't know until that stage whether we'll be eliminating to the
1805 stack pointer (which needs the restriction) or the hard frame
1806 pointer (which doesn't).
1808 All in all, it seems more consistent to only enforce this restriction
1809 during and after reload. */
1810 if (TARGET_MIPS16 && regno == STACK_POINTER_REGNUM)
1811 return !strict || GET_MODE_SIZE (mode) == 4 || GET_MODE_SIZE (mode) == 8;
1813 return TARGET_MIPS16 ? M16_REG_P (regno) : GP_REG_P (regno);
1817 /* Return true if X is a valid base register for the given mode.
1818 Allow only hard registers if STRICT. */
1820 static bool
1821 mips_valid_base_register_p (rtx x, enum machine_mode mode, int strict)
1823 if (!strict && GET_CODE (x) == SUBREG)
1824 x = SUBREG_REG (x);
1826 return (REG_P (x)
1827 && mips_regno_mode_ok_for_base_p (REGNO (x), mode, strict));
1831 /* Return true if X is a valid address for machine mode MODE. If it is,
1832 fill in INFO appropriately. STRICT is true if we should only accept
1833 hard base registers. */
1835 static bool
1836 mips_classify_address (struct mips_address_info *info, rtx x,
1837 enum machine_mode mode, int strict)
1839 switch (GET_CODE (x))
1841 case REG:
1842 case SUBREG:
1843 info->type = ADDRESS_REG;
1844 info->reg = x;
1845 info->offset = const0_rtx;
1846 return mips_valid_base_register_p (info->reg, mode, strict);
1848 case PLUS:
1849 info->type = ADDRESS_REG;
1850 info->reg = XEXP (x, 0);
1851 info->offset = XEXP (x, 1);
1852 return (mips_valid_base_register_p (info->reg, mode, strict)
1853 && const_arith_operand (info->offset, VOIDmode));
1855 case LO_SUM:
1856 info->type = ADDRESS_LO_SUM;
1857 info->reg = XEXP (x, 0);
1858 info->offset = XEXP (x, 1);
1859 /* We have to trust the creator of the LO_SUM to do something vaguely
1860 sane. Target-independent code that creates a LO_SUM should also
1861 create and verify the matching HIGH. Target-independent code that
1862 adds an offset to a LO_SUM must prove that the offset will not
1863 induce a carry. Failure to do either of these things would be
1864 a bug, and we are not required to check for it here. The MIPS
1865 backend itself should only create LO_SUMs for valid symbolic
1866 constants, with the high part being either a HIGH or a copy
1867 of _gp. */
1868 info->symbol_type
1869 = mips_classify_symbolic_expression (info->offset, SYMBOL_CONTEXT_MEM);
1870 return (mips_valid_base_register_p (info->reg, mode, strict)
1871 && mips_symbol_insns (info->symbol_type, mode) > 0
1872 && mips_lo_relocs[info->symbol_type] != 0);
1874 case CONST_INT:
1875 /* Small-integer addresses don't occur very often, but they
1876 are legitimate if $0 is a valid base register. */
1877 info->type = ADDRESS_CONST_INT;
1878 return !TARGET_MIPS16 && SMALL_INT (x);
1880 case CONST:
1881 case LABEL_REF:
1882 case SYMBOL_REF:
1883 info->type = ADDRESS_SYMBOLIC;
1884 return (mips_symbolic_constant_p (x, SYMBOL_CONTEXT_MEM,
1885 &info->symbol_type)
1886 && mips_symbol_insns (info->symbol_type, mode) > 0
1887 && !mips_split_p[info->symbol_type]);
1889 default:
1890 return false;
1894 /* This function is used to implement GO_IF_LEGITIMATE_ADDRESS. It
1895 returns a nonzero value if X is a legitimate address for a memory
1896 operand of the indicated MODE. STRICT is nonzero if this function
1897 is called during reload. */
1899 bool
1900 mips_legitimate_address_p (enum machine_mode mode, rtx x, int strict)
1902 struct mips_address_info addr;
1904 return mips_classify_address (&addr, x, mode, strict);
1907 /* Return true if X is a legitimate $sp-based address for mode MDOE. */
1909 bool
1910 mips_stack_address_p (rtx x, enum machine_mode mode)
1912 struct mips_address_info addr;
1914 return (mips_classify_address (&addr, x, mode, false)
1915 && addr.type == ADDRESS_REG
1916 && addr.reg == stack_pointer_rtx);
1919 /* Return true if ADDR matches the pattern for the lwxs load scaled indexed
1920 address instruction. */
1922 static bool
1923 mips_lwxs_address_p (rtx addr)
1925 if (ISA_HAS_LWXS
1926 && GET_CODE (addr) == PLUS
1927 && REG_P (XEXP (addr, 1)))
1929 rtx offset = XEXP (addr, 0);
1930 if (GET_CODE (offset) == MULT
1931 && REG_P (XEXP (offset, 0))
1932 && GET_CODE (XEXP (offset, 1)) == CONST_INT
1933 && INTVAL (XEXP (offset, 1)) == 4)
1934 return true;
1936 return false;
1939 /* Return true if a value at OFFSET bytes from BASE can be accessed
1940 using an unextended mips16 instruction. MODE is the mode of the
1941 value.
1943 Usually the offset in an unextended instruction is a 5-bit field.
1944 The offset is unsigned and shifted left once for HIs, twice
1945 for SIs, and so on. An exception is SImode accesses off the
1946 stack pointer, which have an 8-bit immediate field. */
1948 static bool
1949 mips16_unextended_reference_p (enum machine_mode mode, rtx base, rtx offset)
1951 if (TARGET_MIPS16
1952 && GET_CODE (offset) == CONST_INT
1953 && INTVAL (offset) >= 0
1954 && (INTVAL (offset) & (GET_MODE_SIZE (mode) - 1)) == 0)
1956 if (GET_MODE_SIZE (mode) == 4 && base == stack_pointer_rtx)
1957 return INTVAL (offset) < 256 * GET_MODE_SIZE (mode);
1958 return INTVAL (offset) < 32 * GET_MODE_SIZE (mode);
1960 return false;
1964 /* Return the number of instructions needed to load or store a value
1965 of mode MODE at X. Return 0 if X isn't valid for MODE. Assume that
1966 multiword moves may need to be split into word moves if MIGHT_SPLIT_P,
1967 otherwise assume that a single load or store is enough.
1969 For mips16 code, count extended instructions as two instructions. */
1972 mips_address_insns (rtx x, enum machine_mode mode, bool might_split_p)
1974 struct mips_address_info addr;
1975 int factor;
1977 /* BLKmode is used for single unaligned loads and stores and should
1978 not count as a multiword mode. (GET_MODE_SIZE (BLKmode) is pretty
1979 meaningless, so we have to single it out as a special case one way
1980 or the other.) */
1981 if (mode != BLKmode && might_split_p)
1982 factor = (GET_MODE_SIZE (mode) + UNITS_PER_WORD - 1) / UNITS_PER_WORD;
1983 else
1984 factor = 1;
1986 if (mips_classify_address (&addr, x, mode, false))
1987 switch (addr.type)
1989 case ADDRESS_REG:
1990 if (TARGET_MIPS16
1991 && !mips16_unextended_reference_p (mode, addr.reg, addr.offset))
1992 return factor * 2;
1993 return factor;
1995 case ADDRESS_LO_SUM:
1996 return (TARGET_MIPS16 ? factor * 2 : factor);
1998 case ADDRESS_CONST_INT:
1999 return factor;
2001 case ADDRESS_SYMBOLIC:
2002 return factor * mips_symbol_insns (addr.symbol_type, mode);
2004 return 0;
2008 /* Likewise for constant X. */
2011 mips_const_insns (rtx x)
2013 struct mips_integer_op codes[MIPS_MAX_INTEGER_OPS];
2014 enum mips_symbol_type symbol_type;
2015 rtx offset;
2017 switch (GET_CODE (x))
2019 case HIGH:
2020 if (!mips_symbolic_constant_p (XEXP (x, 0), SYMBOL_CONTEXT_LEA,
2021 &symbol_type)
2022 || !mips_split_p[symbol_type])
2023 return 0;
2025 /* This is simply an lui for normal mode. It is an extended
2026 "li" followed by an extended "sll" for MIPS16. */
2027 return TARGET_MIPS16 ? 4 : 1;
2029 case CONST_INT:
2030 if (TARGET_MIPS16)
2031 /* Unsigned 8-bit constants can be loaded using an unextended
2032 LI instruction. Unsigned 16-bit constants can be loaded
2033 using an extended LI. Negative constants must be loaded
2034 using LI and then negated. */
2035 return (INTVAL (x) >= 0 && INTVAL (x) < 256 ? 1
2036 : SMALL_OPERAND_UNSIGNED (INTVAL (x)) ? 2
2037 : INTVAL (x) > -256 && INTVAL (x) < 0 ? 2
2038 : SMALL_OPERAND_UNSIGNED (-INTVAL (x)) ? 3
2039 : 0);
2041 return mips_build_integer (codes, INTVAL (x));
2043 case CONST_DOUBLE:
2044 case CONST_VECTOR:
2045 return (!TARGET_MIPS16 && x == CONST0_RTX (GET_MODE (x)) ? 1 : 0);
2047 case CONST:
2048 if (CONST_GP_P (x))
2049 return 1;
2051 /* See if we can refer to X directly. */
2052 if (mips_symbolic_constant_p (x, SYMBOL_CONTEXT_LEA, &symbol_type))
2053 return mips_symbol_insns (symbol_type, MAX_MACHINE_MODE);
2055 /* Otherwise try splitting the constant into a base and offset.
2056 16-bit offsets can be added using an extra addiu. Larger offsets
2057 must be calculated separately and then added to the base. */
2058 split_const (x, &x, &offset);
2059 if (offset != 0)
2061 int n = mips_const_insns (x);
2062 if (n != 0)
2064 if (SMALL_INT (offset))
2065 return n + 1;
2066 else
2067 return n + 1 + mips_build_integer (codes, INTVAL (offset));
2070 return 0;
2072 case SYMBOL_REF:
2073 case LABEL_REF:
2074 return mips_symbol_insns (mips_classify_symbol (x, SYMBOL_CONTEXT_LEA),
2075 MAX_MACHINE_MODE);
2077 default:
2078 return 0;
2083 /* Return the number of instructions needed to implement INSN,
2084 given that it loads from or stores to MEM. Count extended
2085 mips16 instructions as two instructions. */
2088 mips_load_store_insns (rtx mem, rtx insn)
2090 enum machine_mode mode;
2091 bool might_split_p;
2092 rtx set;
2094 gcc_assert (MEM_P (mem));
2095 mode = GET_MODE (mem);
2097 /* Try to prove that INSN does not need to be split. */
2098 might_split_p = true;
2099 if (GET_MODE_BITSIZE (mode) == 64)
2101 set = single_set (insn);
2102 if (set && !mips_split_64bit_move_p (SET_DEST (set), SET_SRC (set)))
2103 might_split_p = false;
2106 return mips_address_insns (XEXP (mem, 0), mode, might_split_p);
2110 /* Return the number of instructions needed for an integer division. */
2113 mips_idiv_insns (void)
2115 int count;
2117 count = 1;
2118 if (TARGET_CHECK_ZERO_DIV)
2120 if (GENERATE_DIVIDE_TRAPS)
2121 count++;
2122 else
2123 count += 2;
2126 if (TARGET_FIX_R4000 || TARGET_FIX_R4400)
2127 count++;
2128 return count;
2131 /* Emit a move from SRC to DEST. Assume that the move expanders can
2132 handle all moves if !can_create_pseudo_p (). The distinction is
2133 important because, unlike emit_move_insn, the move expanders know
2134 how to force Pmode objects into the constant pool even when the
2135 constant pool address is not itself legitimate. */
2138 mips_emit_move (rtx dest, rtx src)
2140 return (can_create_pseudo_p ()
2141 ? emit_move_insn (dest, src)
2142 : emit_move_insn_1 (dest, src));
2145 /* Emit an instruction of the form (set TARGET (CODE OP0 OP1)). */
2147 static void
2148 mips_emit_binary (enum rtx_code code, rtx target, rtx op0, rtx op1)
2150 emit_insn (gen_rtx_SET (VOIDmode, target,
2151 gen_rtx_fmt_ee (code, GET_MODE (target), op0, op1)));
2154 /* Copy VALUE to a register and return that register. If new psuedos
2155 are allowed, copy it into a new register, otherwise use DEST. */
2157 static rtx
2158 mips_force_temporary (rtx dest, rtx value)
2160 if (can_create_pseudo_p ())
2161 return force_reg (Pmode, value);
2162 else
2164 mips_emit_move (copy_rtx (dest), value);
2165 return dest;
2169 /* If we can access small data directly (using gp-relative relocation
2170 operators) return the small data pointer, otherwise return null.
2172 For each mips16 function which refers to GP relative symbols, we
2173 use a pseudo register, initialized at the start of the function, to
2174 hold the $gp value. */
2176 static rtx
2177 mips16_gp_pseudo_reg (void)
2179 if (cfun->machine->mips16_gp_pseudo_rtx == NULL_RTX)
2180 cfun->machine->mips16_gp_pseudo_rtx = gen_reg_rtx (Pmode);
2182 /* Don't initialize the pseudo register if we are being called from
2183 the tree optimizers' cost-calculation routines. */
2184 if (!cfun->machine->initialized_mips16_gp_pseudo_p
2185 && (current_ir_type () != IR_GIMPLE || currently_expanding_to_rtl))
2187 rtx insn, scan;
2189 /* We want to initialize this to a value which gcc will believe
2190 is constant. */
2191 insn = gen_load_const_gp (cfun->machine->mips16_gp_pseudo_rtx);
2193 push_topmost_sequence ();
2194 /* We need to emit the initialization after the FUNCTION_BEG
2195 note, so that it will be integrated. */
2196 for (scan = get_insns (); scan != NULL_RTX; scan = NEXT_INSN (scan))
2197 if (NOTE_P (scan)
2198 && NOTE_KIND (scan) == NOTE_INSN_FUNCTION_BEG)
2199 break;
2200 if (scan == NULL_RTX)
2201 scan = get_insns ();
2202 insn = emit_insn_after (insn, scan);
2203 pop_topmost_sequence ();
2205 cfun->machine->initialized_mips16_gp_pseudo_p = true;
2208 return cfun->machine->mips16_gp_pseudo_rtx;
2211 /* If MODE is MAX_MACHINE_MODE, ADDR appears as a move operand, otherwise
2212 it appears in a MEM of that mode. Return true if ADDR is a legitimate
2213 constant in that context and can be split into a high part and a LO_SUM.
2214 If so, and if LO_SUM_OUT is nonnull, emit the high part and return
2215 the LO_SUM in *LO_SUM_OUT. Leave *LO_SUM_OUT unchanged otherwise.
2217 TEMP is as for mips_force_temporary and is used to load the high
2218 part into a register. */
2220 bool
2221 mips_split_symbol (rtx temp, rtx addr, enum machine_mode mode, rtx *lo_sum_out)
2223 enum mips_symbol_context context;
2224 enum mips_symbol_type symbol_type;
2225 rtx high;
2227 context = (mode == MAX_MACHINE_MODE
2228 ? SYMBOL_CONTEXT_LEA
2229 : SYMBOL_CONTEXT_MEM);
2230 if (!mips_symbolic_constant_p (addr, context, &symbol_type)
2231 || mips_symbol_insns (symbol_type, mode) == 0
2232 || !mips_split_p[symbol_type])
2233 return false;
2235 if (lo_sum_out)
2237 if (symbol_type == SYMBOL_GP_RELATIVE)
2239 if (!can_create_pseudo_p ())
2241 emit_insn (gen_load_const_gp (copy_rtx (temp)));
2242 high = temp;
2244 else
2245 high = mips16_gp_pseudo_reg ();
2247 else
2249 high = gen_rtx_HIGH (Pmode, copy_rtx (addr));
2250 high = mips_force_temporary (temp, high);
2252 *lo_sum_out = gen_rtx_LO_SUM (Pmode, high, addr);
2254 return true;
2258 /* Wrap symbol or label BASE in an unspec address of type SYMBOL_TYPE
2259 and add CONST_INT OFFSET to the result. */
2261 static rtx
2262 mips_unspec_address_offset (rtx base, rtx offset,
2263 enum mips_symbol_type symbol_type)
2265 base = gen_rtx_UNSPEC (Pmode, gen_rtvec (1, base),
2266 UNSPEC_ADDRESS_FIRST + symbol_type);
2267 if (offset != const0_rtx)
2268 base = gen_rtx_PLUS (Pmode, base, offset);
2269 return gen_rtx_CONST (Pmode, base);
2272 /* Return an UNSPEC address with underlying address ADDRESS and symbol
2273 type SYMBOL_TYPE. */
2276 mips_unspec_address (rtx address, enum mips_symbol_type symbol_type)
2278 rtx base, offset;
2280 split_const (address, &base, &offset);
2281 return mips_unspec_address_offset (base, offset, symbol_type);
2285 /* If mips_unspec_address (ADDR, SYMBOL_TYPE) is a 32-bit value, add the
2286 high part to BASE and return the result. Just return BASE otherwise.
2287 TEMP is available as a temporary register if needed.
2289 The returned expression can be used as the first operand to a LO_SUM. */
2291 static rtx
2292 mips_unspec_offset_high (rtx temp, rtx base, rtx addr,
2293 enum mips_symbol_type symbol_type)
2295 if (mips_split_p[symbol_type])
2297 addr = gen_rtx_HIGH (Pmode, mips_unspec_address (addr, symbol_type));
2298 addr = mips_force_temporary (temp, addr);
2299 return mips_force_temporary (temp, gen_rtx_PLUS (Pmode, addr, base));
2301 return base;
2305 /* Return a legitimate address for REG + OFFSET. TEMP is as for
2306 mips_force_temporary; it is only needed when OFFSET is not a
2307 SMALL_OPERAND. */
2309 static rtx
2310 mips_add_offset (rtx temp, rtx reg, HOST_WIDE_INT offset)
2312 if (!SMALL_OPERAND (offset))
2314 rtx high;
2315 if (TARGET_MIPS16)
2317 /* Load the full offset into a register so that we can use
2318 an unextended instruction for the address itself. */
2319 high = GEN_INT (offset);
2320 offset = 0;
2322 else
2324 /* Leave OFFSET as a 16-bit offset and put the excess in HIGH. */
2325 high = GEN_INT (CONST_HIGH_PART (offset));
2326 offset = CONST_LOW_PART (offset);
2328 high = mips_force_temporary (temp, high);
2329 reg = mips_force_temporary (temp, gen_rtx_PLUS (Pmode, high, reg));
2331 return plus_constant (reg, offset);
2334 /* Emit a call to __tls_get_addr. SYM is the TLS symbol we are
2335 referencing, and TYPE is the symbol type to use (either global
2336 dynamic or local dynamic). V0 is an RTX for the return value
2337 location. The entire insn sequence is returned. */
2339 static GTY(()) rtx mips_tls_symbol;
2341 static rtx
2342 mips_call_tls_get_addr (rtx sym, enum mips_symbol_type type, rtx v0)
2344 rtx insn, loc, tga, a0;
2346 a0 = gen_rtx_REG (Pmode, GP_ARG_FIRST);
2348 if (!mips_tls_symbol)
2349 mips_tls_symbol = init_one_libfunc ("__tls_get_addr");
2351 loc = mips_unspec_address (sym, type);
2353 start_sequence ();
2355 emit_insn (gen_rtx_SET (Pmode, a0,
2356 gen_rtx_LO_SUM (Pmode, pic_offset_table_rtx, loc)));
2357 tga = gen_const_mem (Pmode, mips_tls_symbol);
2358 insn = emit_call_insn (gen_call_value (v0, tga, const0_rtx, const0_rtx));
2359 CONST_OR_PURE_CALL_P (insn) = 1;
2360 use_reg (&CALL_INSN_FUNCTION_USAGE (insn), a0);
2361 insn = get_insns ();
2363 end_sequence ();
2365 return insn;
2368 /* Generate the code to access LOC, a thread local SYMBOL_REF. The
2369 return value will be a valid address and move_operand (either a REG
2370 or a LO_SUM). */
2372 static rtx
2373 mips_legitimize_tls_address (rtx loc)
2375 rtx dest, insn, v0, v1, tmp1, tmp2, eqv;
2376 enum tls_model model;
2378 if (TARGET_MIPS16)
2380 sorry ("MIPS16 TLS");
2381 return gen_reg_rtx (Pmode);
2384 v0 = gen_rtx_REG (Pmode, GP_RETURN);
2385 v1 = gen_rtx_REG (Pmode, GP_RETURN + 1);
2387 model = SYMBOL_REF_TLS_MODEL (loc);
2388 /* Only TARGET_ABICALLS code can have more than one module; other
2389 code must be be static and should not use a GOT. All TLS models
2390 reduce to local exec in this situation. */
2391 if (!TARGET_ABICALLS)
2392 model = TLS_MODEL_LOCAL_EXEC;
2394 switch (model)
2396 case TLS_MODEL_GLOBAL_DYNAMIC:
2397 insn = mips_call_tls_get_addr (loc, SYMBOL_TLSGD, v0);
2398 dest = gen_reg_rtx (Pmode);
2399 emit_libcall_block (insn, dest, v0, loc);
2400 break;
2402 case TLS_MODEL_LOCAL_DYNAMIC:
2403 insn = mips_call_tls_get_addr (loc, SYMBOL_TLSLDM, v0);
2404 tmp1 = gen_reg_rtx (Pmode);
2406 /* Attach a unique REG_EQUIV, to allow the RTL optimizers to
2407 share the LDM result with other LD model accesses. */
2408 eqv = gen_rtx_UNSPEC (Pmode, gen_rtvec (1, const0_rtx),
2409 UNSPEC_TLS_LDM);
2410 emit_libcall_block (insn, tmp1, v0, eqv);
2412 tmp2 = mips_unspec_offset_high (NULL, tmp1, loc, SYMBOL_DTPREL);
2413 dest = gen_rtx_LO_SUM (Pmode, tmp2,
2414 mips_unspec_address (loc, SYMBOL_DTPREL));
2415 break;
2417 case TLS_MODEL_INITIAL_EXEC:
2418 tmp1 = gen_reg_rtx (Pmode);
2419 tmp2 = mips_unspec_address (loc, SYMBOL_GOTTPREL);
2420 if (Pmode == DImode)
2422 emit_insn (gen_tls_get_tp_di (v1));
2423 emit_insn (gen_load_gotdi (tmp1, pic_offset_table_rtx, tmp2));
2425 else
2427 emit_insn (gen_tls_get_tp_si (v1));
2428 emit_insn (gen_load_gotsi (tmp1, pic_offset_table_rtx, tmp2));
2430 dest = gen_reg_rtx (Pmode);
2431 emit_insn (gen_add3_insn (dest, tmp1, v1));
2432 break;
2434 case TLS_MODEL_LOCAL_EXEC:
2435 if (Pmode == DImode)
2436 emit_insn (gen_tls_get_tp_di (v1));
2437 else
2438 emit_insn (gen_tls_get_tp_si (v1));
2440 tmp1 = mips_unspec_offset_high (NULL, v1, loc, SYMBOL_TPREL);
2441 dest = gen_rtx_LO_SUM (Pmode, tmp1,
2442 mips_unspec_address (loc, SYMBOL_TPREL));
2443 break;
2445 default:
2446 gcc_unreachable ();
2449 return dest;
2452 /* This function is used to implement LEGITIMIZE_ADDRESS. If *XLOC can
2453 be legitimized in a way that the generic machinery might not expect,
2454 put the new address in *XLOC and return true. MODE is the mode of
2455 the memory being accessed. */
2457 bool
2458 mips_legitimize_address (rtx *xloc, enum machine_mode mode)
2460 if (mips_tls_operand_p (*xloc))
2462 *xloc = mips_legitimize_tls_address (*xloc);
2463 return true;
2466 /* See if the address can split into a high part and a LO_SUM. */
2467 if (mips_split_symbol (NULL, *xloc, mode, xloc))
2468 return true;
2470 if (GET_CODE (*xloc) == PLUS && GET_CODE (XEXP (*xloc, 1)) == CONST_INT)
2472 /* Handle REG + CONSTANT using mips_add_offset. */
2473 rtx reg;
2475 reg = XEXP (*xloc, 0);
2476 if (!mips_valid_base_register_p (reg, mode, 0))
2477 reg = copy_to_mode_reg (Pmode, reg);
2478 *xloc = mips_add_offset (0, reg, INTVAL (XEXP (*xloc, 1)));
2479 return true;
2482 return false;
2486 /* Load VALUE into DEST, using TEMP as a temporary register if need be. */
2488 void
2489 mips_move_integer (rtx dest, rtx temp, unsigned HOST_WIDE_INT value)
2491 struct mips_integer_op codes[MIPS_MAX_INTEGER_OPS];
2492 enum machine_mode mode;
2493 unsigned int i, cost;
2494 rtx x;
2496 mode = GET_MODE (dest);
2497 cost = mips_build_integer (codes, value);
2499 /* Apply each binary operation to X. Invariant: X is a legitimate
2500 source operand for a SET pattern. */
2501 x = GEN_INT (codes[0].value);
2502 for (i = 1; i < cost; i++)
2504 if (!can_create_pseudo_p ())
2506 emit_insn (gen_rtx_SET (VOIDmode, temp, x));
2507 x = temp;
2509 else
2510 x = force_reg (mode, x);
2511 x = gen_rtx_fmt_ee (codes[i].code, mode, x, GEN_INT (codes[i].value));
2514 emit_insn (gen_rtx_SET (VOIDmode, dest, x));
2518 /* Subroutine of mips_legitimize_move. Move constant SRC into register
2519 DEST given that SRC satisfies immediate_operand but doesn't satisfy
2520 move_operand. */
2522 static void
2523 mips_legitimize_const_move (enum machine_mode mode, rtx dest, rtx src)
2525 rtx base, offset;
2527 /* Split moves of big integers into smaller pieces. */
2528 if (splittable_const_int_operand (src, mode))
2530 mips_move_integer (dest, dest, INTVAL (src));
2531 return;
2534 /* Split moves of symbolic constants into high/low pairs. */
2535 if (mips_split_symbol (dest, src, MAX_MACHINE_MODE, &src))
2537 emit_insn (gen_rtx_SET (VOIDmode, dest, src));
2538 return;
2541 if (mips_tls_operand_p (src))
2543 mips_emit_move (dest, mips_legitimize_tls_address (src));
2544 return;
2547 /* If we have (const (plus symbol offset)), and that expression cannot
2548 be forced into memory, load the symbol first and add in the offset.
2549 In non-MIPS16 mode, prefer to do this even if the constant _can_ be
2550 forced into memory, as it usually produces better code. */
2551 split_const (src, &base, &offset);
2552 if (offset != const0_rtx
2553 && (targetm.cannot_force_const_mem (src)
2554 || (!TARGET_MIPS16 && can_create_pseudo_p ())))
2556 base = mips_force_temporary (dest, base);
2557 mips_emit_move (dest, mips_add_offset (0, base, INTVAL (offset)));
2558 return;
2561 src = force_const_mem (mode, src);
2563 /* When using explicit relocs, constant pool references are sometimes
2564 not legitimate addresses. */
2565 mips_split_symbol (dest, XEXP (src, 0), mode, &XEXP (src, 0));
2566 mips_emit_move (dest, src);
2570 /* If (set DEST SRC) is not a valid instruction, emit an equivalent
2571 sequence that is valid. */
2573 bool
2574 mips_legitimize_move (enum machine_mode mode, rtx dest, rtx src)
2576 if (!register_operand (dest, mode) && !reg_or_0_operand (src, mode))
2578 mips_emit_move (dest, force_reg (mode, src));
2579 return true;
2582 /* Check for individual, fully-reloaded mflo and mfhi instructions. */
2583 if (GET_MODE_SIZE (mode) <= UNITS_PER_WORD
2584 && REG_P (src) && MD_REG_P (REGNO (src))
2585 && REG_P (dest) && GP_REG_P (REGNO (dest)))
2587 int other_regno = REGNO (src) == HI_REGNUM ? LO_REGNUM : HI_REGNUM;
2588 if (GET_MODE_SIZE (mode) <= 4)
2589 emit_insn (gen_mfhilo_si (gen_rtx_REG (SImode, REGNO (dest)),
2590 gen_rtx_REG (SImode, REGNO (src)),
2591 gen_rtx_REG (SImode, other_regno)));
2592 else
2593 emit_insn (gen_mfhilo_di (gen_rtx_REG (DImode, REGNO (dest)),
2594 gen_rtx_REG (DImode, REGNO (src)),
2595 gen_rtx_REG (DImode, other_regno)));
2596 return true;
2599 /* We need to deal with constants that would be legitimate
2600 immediate_operands but not legitimate move_operands. */
2601 if (CONSTANT_P (src) && !move_operand (src, mode))
2603 mips_legitimize_const_move (mode, dest, src);
2604 set_unique_reg_note (get_last_insn (), REG_EQUAL, copy_rtx (src));
2605 return true;
2607 return false;
2610 /* Return true if X in context CONTEXT is a small data address that can
2611 be rewritten as a LO_SUM. */
2613 static bool
2614 mips_rewrite_small_data_p (rtx x, enum mips_symbol_context context)
2616 enum mips_symbol_type symbol_type;
2618 return (TARGET_EXPLICIT_RELOCS
2619 && mips_symbolic_constant_p (x, context, &symbol_type)
2620 && symbol_type == SYMBOL_GP_RELATIVE);
2624 /* A for_each_rtx callback for mips_small_data_pattern_p. DATA is the
2625 containing MEM, or null if none. */
2627 static int
2628 mips_small_data_pattern_1 (rtx *loc, void *data)
2630 enum mips_symbol_context context;
2632 if (GET_CODE (*loc) == LO_SUM)
2633 return -1;
2635 if (MEM_P (*loc))
2637 if (for_each_rtx (&XEXP (*loc, 0), mips_small_data_pattern_1, *loc))
2638 return 1;
2639 return -1;
2642 context = data ? SYMBOL_CONTEXT_MEM : SYMBOL_CONTEXT_LEA;
2643 return mips_rewrite_small_data_p (*loc, context);
2646 /* Return true if OP refers to small data symbols directly, not through
2647 a LO_SUM. */
2649 bool
2650 mips_small_data_pattern_p (rtx op)
2652 return for_each_rtx (&op, mips_small_data_pattern_1, 0);
2655 /* A for_each_rtx callback, used by mips_rewrite_small_data.
2656 DATA is the containing MEM, or null if none. */
2658 static int
2659 mips_rewrite_small_data_1 (rtx *loc, void *data)
2661 enum mips_symbol_context context;
2663 if (MEM_P (*loc))
2665 for_each_rtx (&XEXP (*loc, 0), mips_rewrite_small_data_1, *loc);
2666 return -1;
2669 context = data ? SYMBOL_CONTEXT_MEM : SYMBOL_CONTEXT_LEA;
2670 if (mips_rewrite_small_data_p (*loc, context))
2671 *loc = gen_rtx_LO_SUM (Pmode, pic_offset_table_rtx, *loc);
2673 if (GET_CODE (*loc) == LO_SUM)
2674 return -1;
2676 return 0;
2679 /* If possible, rewrite OP so that it refers to small data using
2680 explicit relocations. */
2683 mips_rewrite_small_data (rtx op)
2685 op = copy_insn (op);
2686 for_each_rtx (&op, mips_rewrite_small_data_1, 0);
2687 return op;
2690 /* We need a lot of little routines to check constant values on the
2691 mips16. These are used to figure out how long the instruction will
2692 be. It would be much better to do this using constraints, but
2693 there aren't nearly enough letters available. */
2695 static int
2696 m16_check_op (rtx op, int low, int high, int mask)
2698 return (GET_CODE (op) == CONST_INT
2699 && INTVAL (op) >= low
2700 && INTVAL (op) <= high
2701 && (INTVAL (op) & mask) == 0);
2705 m16_uimm3_b (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
2707 return m16_check_op (op, 0x1, 0x8, 0);
2711 m16_simm4_1 (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
2713 return m16_check_op (op, - 0x8, 0x7, 0);
2717 m16_nsimm4_1 (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
2719 return m16_check_op (op, - 0x7, 0x8, 0);
2723 m16_simm5_1 (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
2725 return m16_check_op (op, - 0x10, 0xf, 0);
2729 m16_nsimm5_1 (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
2731 return m16_check_op (op, - 0xf, 0x10, 0);
2735 m16_uimm5_4 (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
2737 return m16_check_op (op, (- 0x10) << 2, 0xf << 2, 3);
2741 m16_nuimm5_4 (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
2743 return m16_check_op (op, (- 0xf) << 2, 0x10 << 2, 3);
2747 m16_simm8_1 (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
2749 return m16_check_op (op, - 0x80, 0x7f, 0);
2753 m16_nsimm8_1 (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
2755 return m16_check_op (op, - 0x7f, 0x80, 0);
2759 m16_uimm8_1 (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
2761 return m16_check_op (op, 0x0, 0xff, 0);
2765 m16_nuimm8_1 (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
2767 return m16_check_op (op, - 0xff, 0x0, 0);
2771 m16_uimm8_m1_1 (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
2773 return m16_check_op (op, - 0x1, 0xfe, 0);
2777 m16_uimm8_4 (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
2779 return m16_check_op (op, 0x0, 0xff << 2, 3);
2783 m16_nuimm8_4 (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
2785 return m16_check_op (op, (- 0xff) << 2, 0x0, 3);
2789 m16_simm8_8 (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
2791 return m16_check_op (op, (- 0x80) << 3, 0x7f << 3, 7);
2795 m16_nsimm8_8 (rtx op, enum machine_mode mode ATTRIBUTE_UNUSED)
2797 return m16_check_op (op, (- 0x7f) << 3, 0x80 << 3, 7);
2800 /* The cost of loading values from the constant pool. It should be
2801 larger than the cost of any constant we want to synthesize inline. */
2803 #define CONSTANT_POOL_COST COSTS_N_INSNS (TARGET_MIPS16 ? 4 : 8)
2805 /* Return the cost of X when used as an operand to the MIPS16 instruction
2806 that implements CODE. Return -1 if there is no such instruction, or if
2807 X is not a valid immediate operand for it. */
2809 static int
2810 mips16_constant_cost (int code, HOST_WIDE_INT x)
2812 switch (code)
2814 case ASHIFT:
2815 case ASHIFTRT:
2816 case LSHIFTRT:
2817 /* Shifts by between 1 and 8 bits (inclusive) are unextended,
2818 other shifts are extended. The shift patterns truncate the shift
2819 count to the right size, so there are no out-of-range values. */
2820 if (IN_RANGE (x, 1, 8))
2821 return 0;
2822 return COSTS_N_INSNS (1);
2824 case PLUS:
2825 if (IN_RANGE (x, -128, 127))
2826 return 0;
2827 if (SMALL_OPERAND (x))
2828 return COSTS_N_INSNS (1);
2829 return -1;
2831 case LEU:
2832 /* Like LE, but reject the always-true case. */
2833 if (x == -1)
2834 return -1;
2835 case LE:
2836 /* We add 1 to the immediate and use SLT. */
2837 x += 1;
2838 case XOR:
2839 /* We can use CMPI for an xor with an unsigned 16-bit X. */
2840 case LT:
2841 case LTU:
2842 if (IN_RANGE (x, 0, 255))
2843 return 0;
2844 if (SMALL_OPERAND_UNSIGNED (x))
2845 return COSTS_N_INSNS (1);
2846 return -1;
2848 case EQ:
2849 case NE:
2850 /* Equality comparisons with 0 are cheap. */
2851 if (x == 0)
2852 return 0;
2853 return -1;
2855 default:
2856 return -1;
2860 /* Return true if there is a non-MIPS16 instruction that implements CODE
2861 and if that instruction accepts X as an immediate operand. */
2863 static int
2864 mips_immediate_operand_p (int code, HOST_WIDE_INT x)
2866 switch (code)
2868 case ASHIFT:
2869 case ASHIFTRT:
2870 case LSHIFTRT:
2871 /* All shift counts are truncated to a valid constant. */
2872 return true;
2874 case ROTATE:
2875 case ROTATERT:
2876 /* Likewise rotates, if the target supports rotates at all. */
2877 return ISA_HAS_ROR;
2879 case AND:
2880 case IOR:
2881 case XOR:
2882 /* These instructions take 16-bit unsigned immediates. */
2883 return SMALL_OPERAND_UNSIGNED (x);
2885 case PLUS:
2886 case LT:
2887 case LTU:
2888 /* These instructions take 16-bit signed immediates. */
2889 return SMALL_OPERAND (x);
2891 case EQ:
2892 case NE:
2893 case GT:
2894 case GTU:
2895 /* The "immediate" forms of these instructions are really
2896 implemented as comparisons with register 0. */
2897 return x == 0;
2899 case GE:
2900 case GEU:
2901 /* Likewise, meaning that the only valid immediate operand is 1. */
2902 return x == 1;
2904 case LE:
2905 /* We add 1 to the immediate and use SLT. */
2906 return SMALL_OPERAND (x + 1);
2908 case LEU:
2909 /* Likewise SLTU, but reject the always-true case. */
2910 return SMALL_OPERAND (x + 1) && x + 1 != 0;
2912 case SIGN_EXTRACT:
2913 case ZERO_EXTRACT:
2914 /* The bit position and size are immediate operands. */
2915 return ISA_HAS_EXT_INS;
2917 default:
2918 /* By default assume that $0 can be used for 0. */
2919 return x == 0;
2923 /* Return the cost of binary operation X, given that the instruction
2924 sequence for a word-sized or smaller operation has cost SINGLE_COST
2925 and that the sequence of a double-word operation has cost DOUBLE_COST. */
2927 static int
2928 mips_binary_cost (rtx x, int single_cost, int double_cost)
2930 int cost;
2932 if (GET_MODE_SIZE (GET_MODE (x)) == UNITS_PER_WORD * 2)
2933 cost = double_cost;
2934 else
2935 cost = single_cost;
2936 return (cost
2937 + rtx_cost (XEXP (x, 0), 0)
2938 + rtx_cost (XEXP (x, 1), GET_CODE (x)));
2941 /* Return the cost of floating-point multiplications of mode MODE. */
2943 static int
2944 mips_fp_mult_cost (enum machine_mode mode)
2946 return mode == DFmode ? mips_cost->fp_mult_df : mips_cost->fp_mult_sf;
2949 /* Return the cost of floating-point divisions of mode MODE. */
2951 static int
2952 mips_fp_div_cost (enum machine_mode mode)
2954 return mode == DFmode ? mips_cost->fp_div_df : mips_cost->fp_div_sf;
2957 /* Return the cost of sign-extending OP to mode MODE, not including the
2958 cost of OP itself. */
2960 static int
2961 mips_sign_extend_cost (enum machine_mode mode, rtx op)
2963 if (MEM_P (op))
2964 /* Extended loads are as cheap as unextended ones. */
2965 return 0;
2967 if (TARGET_64BIT && mode == DImode && GET_MODE (op) == SImode)
2968 /* A sign extension from SImode to DImode in 64-bit mode is free. */
2969 return 0;
2971 if (ISA_HAS_SEB_SEH || GENERATE_MIPS16E)
2972 /* We can use SEB or SEH. */
2973 return COSTS_N_INSNS (1);
2975 /* We need to use a shift left and a shift right. */
2976 return COSTS_N_INSNS (TARGET_MIPS16 ? 4 : 2);
2979 /* Return the cost of zero-extending OP to mode MODE, not including the
2980 cost of OP itself. */
2982 static int
2983 mips_zero_extend_cost (enum machine_mode mode, rtx op)
2985 if (MEM_P (op))
2986 /* Extended loads are as cheap as unextended ones. */
2987 return 0;
2989 if (TARGET_64BIT && mode == DImode && GET_MODE (op) == SImode)
2990 /* We need a shift left by 32 bits and a shift right by 32 bits. */
2991 return COSTS_N_INSNS (TARGET_MIPS16 ? 4 : 2);
2993 if (GENERATE_MIPS16E)
2994 /* We can use ZEB or ZEH. */
2995 return COSTS_N_INSNS (1);
2997 if (TARGET_MIPS16)
2998 /* We need to load 0xff or 0xffff into a register and use AND. */
2999 return COSTS_N_INSNS (GET_MODE (op) == QImode ? 2 : 3);
3001 /* We can use ANDI. */
3002 return COSTS_N_INSNS (1);
3005 /* Implement TARGET_RTX_COSTS. */
3007 static bool
3008 mips_rtx_costs (rtx x, int code, int outer_code, int *total)
3010 enum machine_mode mode = GET_MODE (x);
3011 bool float_mode_p = FLOAT_MODE_P (mode);
3012 int cost;
3013 rtx addr;
3015 /* The cost of a COMPARE is hard to define for MIPS. COMPAREs don't
3016 appear in the instruction stream, and the cost of a comparison is
3017 really the cost of the branch or scc condition. At the time of
3018 writing, gcc only uses an explicit outer COMPARE code when optabs
3019 is testing whether a constant is expensive enough to force into a
3020 register. We want optabs to pass such constants through the MIPS
3021 expanders instead, so make all constants very cheap here. */
3022 if (outer_code == COMPARE)
3024 gcc_assert (CONSTANT_P (x));
3025 *total = 0;
3026 return true;
3029 switch (code)
3031 case CONST_INT:
3032 /* Treat *clear_upper32-style ANDs as having zero cost in the
3033 second operand. The cost is entirely in the first operand.
3035 ??? This is needed because we would otherwise try to CSE
3036 the constant operand. Although that's the right thing for
3037 instructions that continue to be a register operation throughout
3038 compilation, it is disastrous for instructions that could
3039 later be converted into a memory operation. */
3040 if (TARGET_64BIT
3041 && outer_code == AND
3042 && UINTVAL (x) == 0xffffffff)
3044 *total = 0;
3045 return true;
3048 if (TARGET_MIPS16)
3050 cost = mips16_constant_cost (outer_code, INTVAL (x));
3051 if (cost >= 0)
3053 *total = cost;
3054 return true;
3057 else
3059 /* When not optimizing for size, we care more about the cost
3060 of hot code, and hot code is often in a loop. If a constant
3061 operand needs to be forced into a register, we will often be
3062 able to hoist the constant load out of the loop, so the load
3063 should not contribute to the cost. */
3064 if (!optimize_size
3065 || mips_immediate_operand_p (outer_code, INTVAL (x)))
3067 *total = 0;
3068 return true;
3071 /* Fall through. */
3073 case CONST:
3074 case SYMBOL_REF:
3075 case LABEL_REF:
3076 case CONST_DOUBLE:
3077 if (force_to_mem_operand (x, VOIDmode))
3079 *total = COSTS_N_INSNS (1);
3080 return true;
3082 cost = mips_const_insns (x);
3083 if (cost > 0)
3085 /* If the constant is likely to be stored in a GPR, SETs of
3086 single-insn constants are as cheap as register sets; we
3087 never want to CSE them.
3089 Don't reduce the cost of storing a floating-point zero in
3090 FPRs. If we have a zero in an FPR for other reasons, we
3091 can get better cfg-cleanup and delayed-branch results by
3092 using it consistently, rather than using $0 sometimes and
3093 an FPR at other times. Also, moves between floating-point
3094 registers are sometimes cheaper than (D)MTC1 $0. */
3095 if (cost == 1
3096 && outer_code == SET
3097 && !(float_mode_p && TARGET_HARD_FLOAT))
3098 cost = 0;
3099 /* When non-MIPS16 code loads a constant N>1 times, we rarely
3100 want to CSE the constant itself. It is usually better to
3101 have N copies of the last operation in the sequence and one
3102 shared copy of the other operations. (Note that this is
3103 not true for MIPS16 code, where the final operation in the
3104 sequence is often an extended instruction.)
3106 Also, if we have a CONST_INT, we don't know whether it is
3107 for a word or doubleword operation, so we cannot rely on
3108 the result of mips_build_integer. */
3109 else if (!TARGET_MIPS16
3110 && (outer_code == SET || mode == VOIDmode))
3111 cost = 1;
3112 *total = COSTS_N_INSNS (cost);
3113 return true;
3115 /* The value will need to be fetched from the constant pool. */
3116 *total = CONSTANT_POOL_COST;
3117 return true;
3119 case MEM:
3120 /* If the address is legitimate, return the number of
3121 instructions it needs. */
3122 addr = XEXP (x, 0);
3123 cost = mips_address_insns (addr, mode, true);
3124 if (cost > 0)
3126 *total = COSTS_N_INSNS (cost + 1);
3127 return true;
3129 /* Check for a scaled indexed address. */
3130 if (mips_lwxs_address_p (addr))
3132 *total = COSTS_N_INSNS (2);
3133 return true;
3135 /* Otherwise use the default handling. */
3136 return false;
3138 case FFS:
3139 *total = COSTS_N_INSNS (6);
3140 return false;
3142 case NOT:
3143 *total = COSTS_N_INSNS (GET_MODE_SIZE (mode) > UNITS_PER_WORD ? 2 : 1);
3144 return false;
3146 case AND:
3147 /* Check for a *clear_upper32 pattern and treat it like a zero
3148 extension. See the pattern's comment for details. */
3149 if (TARGET_64BIT
3150 && mode == DImode
3151 && CONST_INT_P (XEXP (x, 1))
3152 && UINTVAL (XEXP (x, 1)) == 0xffffffff)
3154 *total = (mips_zero_extend_cost (mode, XEXP (x, 0))
3155 + rtx_cost (XEXP (x, 0), 0));
3156 return true;
3158 /* Fall through. */
3160 case IOR:
3161 case XOR:
3162 /* Double-word operations use two single-word operations. */
3163 *total = mips_binary_cost (x, COSTS_N_INSNS (1), COSTS_N_INSNS (2));
3164 return true;
3166 case ASHIFT:
3167 case ASHIFTRT:
3168 case LSHIFTRT:
3169 case ROTATE:
3170 case ROTATERT:
3171 if (CONSTANT_P (XEXP (x, 1)))
3172 *total = mips_binary_cost (x, COSTS_N_INSNS (1), COSTS_N_INSNS (4));
3173 else
3174 *total = mips_binary_cost (x, COSTS_N_INSNS (1), COSTS_N_INSNS (12));
3175 return true;
3177 case ABS:
3178 if (float_mode_p)
3179 *total = mips_cost->fp_add;
3180 else
3181 *total = COSTS_N_INSNS (4);
3182 return false;
3184 case LO_SUM:
3185 /* Low-part immediates need an extended MIPS16 instruction. */
3186 *total = (COSTS_N_INSNS (TARGET_MIPS16 ? 2 : 1)
3187 + rtx_cost (XEXP (x, 0), 0));
3188 return true;
3190 case LT:
3191 case LTU:
3192 case LE:
3193 case LEU:
3194 case GT:
3195 case GTU:
3196 case GE:
3197 case GEU:
3198 case EQ:
3199 case NE:
3200 case UNORDERED:
3201 case LTGT:
3202 /* Branch comparisons have VOIDmode, so use the first operand's
3203 mode instead. */
3204 mode = GET_MODE (XEXP (x, 0));
3205 if (FLOAT_MODE_P (mode))
3207 *total = mips_cost->fp_add;
3208 return false;
3210 *total = mips_binary_cost (x, COSTS_N_INSNS (1), COSTS_N_INSNS (4));
3211 return true;
3213 case MINUS:
3214 if (float_mode_p
3215 && ISA_HAS_NMADD_NMSUB (mode)
3216 && TARGET_FUSED_MADD
3217 && !HONOR_NANS (mode)
3218 && !HONOR_SIGNED_ZEROS (mode))
3220 /* See if we can use NMADD or NMSUB. See mips.md for the
3221 associated patterns. */
3222 rtx op0 = XEXP (x, 0);
3223 rtx op1 = XEXP (x, 1);
3224 if (GET_CODE (op0) == MULT && GET_CODE (XEXP (op0, 0)) == NEG)
3226 *total = (mips_fp_mult_cost (mode)
3227 + rtx_cost (XEXP (XEXP (op0, 0), 0), 0)
3228 + rtx_cost (XEXP (op0, 1), 0)
3229 + rtx_cost (op1, 0));
3230 return true;
3232 if (GET_CODE (op1) == MULT)
3234 *total = (mips_fp_mult_cost (mode)
3235 + rtx_cost (op0, 0)
3236 + rtx_cost (XEXP (op1, 0), 0)
3237 + rtx_cost (XEXP (op1, 1), 0));
3238 return true;
3241 /* Fall through. */
3243 case PLUS:
3244 if (float_mode_p)
3246 if (ISA_HAS_FP4
3247 && TARGET_FUSED_MADD
3248 && GET_CODE (XEXP (x, 0)) == MULT)
3249 *total = 0;
3250 else
3251 *total = mips_cost->fp_add;
3252 return false;
3255 /* Double-word operations require three single-word operations and
3256 an SLTU. The MIPS16 version then needs to move the result of
3257 the SLTU from $24 to a MIPS16 register. */
3258 *total = mips_binary_cost (x, COSTS_N_INSNS (1),
3259 COSTS_N_INSNS (TARGET_MIPS16 ? 5 : 4));
3260 return true;
3262 case NEG:
3263 if (float_mode_p
3264 && ISA_HAS_NMADD_NMSUB (mode)
3265 && TARGET_FUSED_MADD
3266 && !HONOR_NANS (mode)
3267 && HONOR_SIGNED_ZEROS (mode))
3269 /* See if we can use NMADD or NMSUB. See mips.md for the
3270 associated patterns. */
3271 rtx op = XEXP (x, 0);
3272 if ((GET_CODE (op) == PLUS || GET_CODE (op) == MINUS)
3273 && GET_CODE (XEXP (op, 0)) == MULT)
3275 *total = (mips_fp_mult_cost (mode)
3276 + rtx_cost (XEXP (XEXP (op, 0), 0), 0)
3277 + rtx_cost (XEXP (XEXP (op, 0), 1), 0)
3278 + rtx_cost (XEXP (op, 1), 0));
3279 return true;
3283 if (float_mode_p)
3284 *total = mips_cost->fp_add;
3285 else
3286 *total = COSTS_N_INSNS (GET_MODE_SIZE (mode) > UNITS_PER_WORD ? 4 : 1);
3287 return false;
3289 case MULT:
3290 if (float_mode_p)
3291 *total = mips_fp_mult_cost (mode);
3292 else if (mode == DImode && !TARGET_64BIT)
3293 /* Synthesized from 2 mulsi3s, 1 mulsidi3 and two additions,
3294 where the mulsidi3 always includes an MFHI and an MFLO. */
3295 *total = (optimize_size
3296 ? COSTS_N_INSNS (ISA_HAS_MUL3 ? 7 : 9)
3297 : mips_cost->int_mult_si * 3 + 6);
3298 else if (optimize_size)
3299 *total = (ISA_HAS_MUL3 ? 1 : 2);
3300 else if (mode == DImode)
3301 *total = mips_cost->int_mult_di;
3302 else
3303 *total = mips_cost->int_mult_si;
3304 return false;
3306 case DIV:
3307 /* Check for a reciprocal. */
3308 if (float_mode_p && XEXP (x, 0) == CONST1_RTX (mode))
3310 if (ISA_HAS_FP4
3311 && flag_unsafe_math_optimizations
3312 && (outer_code == SQRT || GET_CODE (XEXP (x, 1)) == SQRT))
3314 /* An rsqrt<mode>a or rsqrt<mode>b pattern. Count the
3315 division as being free. */
3316 *total = rtx_cost (XEXP (x, 1), 0);
3317 return true;
3319 if (!ISA_MIPS1)
3321 *total = mips_fp_div_cost (mode) + rtx_cost (XEXP (x, 1), 0);
3322 return true;
3325 /* Fall through. */
3327 case SQRT:
3328 case MOD:
3329 if (float_mode_p)
3331 *total = mips_fp_div_cost (mode);
3332 return false;
3334 /* Fall through. */
3336 case UDIV:
3337 case UMOD:
3338 if (optimize_size)
3340 /* It is our responsibility to make division by a power of 2
3341 as cheap as 2 register additions if we want the division
3342 expanders to be used for such operations; see the setting
3343 of sdiv_pow2_cheap in optabs.c. Using (D)DIV for MIPS16
3344 should always produce shorter code than using
3345 expand_sdiv2_pow2. */
3346 if (TARGET_MIPS16
3347 && CONST_INT_P (XEXP (x, 1))
3348 && exact_log2 (INTVAL (XEXP (x, 1))) >= 0)
3350 *total = COSTS_N_INSNS (2) + rtx_cost (XEXP (x, 0), 0);
3351 return true;
3353 *total = COSTS_N_INSNS (mips_idiv_insns ());
3355 else if (mode == DImode)
3356 *total = mips_cost->int_div_di;
3357 else
3358 *total = mips_cost->int_div_si;
3359 return false;
3361 case SIGN_EXTEND:
3362 *total = mips_sign_extend_cost (mode, XEXP (x, 0));
3363 return false;
3365 case ZERO_EXTEND:
3366 *total = mips_zero_extend_cost (mode, XEXP (x, 0));
3367 return false;
3369 case FLOAT:
3370 case UNSIGNED_FLOAT:
3371 case FIX:
3372 case FLOAT_EXTEND:
3373 case FLOAT_TRUNCATE:
3374 *total = mips_cost->fp_add;
3375 return false;
3377 default:
3378 return false;
3382 /* Provide the costs of an addressing mode that contains ADDR.
3383 If ADDR is not a valid address, its cost is irrelevant. */
3385 static int
3386 mips_address_cost (rtx addr)
3388 return mips_address_insns (addr, SImode, false);
3391 /* Return one word of double-word value OP, taking into account the fixed
3392 endianness of certain registers. HIGH_P is true to select the high part,
3393 false to select the low part. */
3396 mips_subword (rtx op, int high_p)
3398 unsigned int byte, offset;
3399 enum machine_mode mode;
3401 mode = GET_MODE (op);
3402 if (mode == VOIDmode)
3403 mode = DImode;
3405 if (TARGET_BIG_ENDIAN ? !high_p : high_p)
3406 byte = UNITS_PER_WORD;
3407 else
3408 byte = 0;
3410 if (FP_REG_RTX_P (op))
3412 /* Paired FPRs are always ordered little-endian. */
3413 offset = (UNITS_PER_WORD < UNITS_PER_HWFPVALUE ? high_p : byte != 0);
3414 return gen_rtx_REG (word_mode, REGNO (op) + offset);
3417 if (MEM_P (op))
3418 return mips_rewrite_small_data (adjust_address (op, word_mode, byte));
3420 return simplify_gen_subreg (word_mode, op, mode, byte);
3424 /* Return true if a 64-bit move from SRC to DEST should be split into two. */
3426 bool
3427 mips_split_64bit_move_p (rtx dest, rtx src)
3429 if (TARGET_64BIT)
3430 return false;
3432 /* FP->FP moves can be done in a single instruction. */
3433 if (FP_REG_RTX_P (src) && FP_REG_RTX_P (dest))
3434 return false;
3436 /* Check for floating-point loads and stores. */
3437 if (ISA_HAS_LDC1_SDC1)
3439 if (FP_REG_RTX_P (dest) && MEM_P (src))
3440 return false;
3441 if (FP_REG_RTX_P (src) && MEM_P (dest))
3442 return false;
3444 return true;
3448 /* Split a doubleword move from SRC to DEST. On 32-bit targets,
3449 this function handles 64-bit moves for which mips_split_64bit_move_p
3450 holds. For 64-bit targets, this function handles 128-bit moves. */
3452 void
3453 mips_split_doubleword_move (rtx dest, rtx src)
3455 if (FP_REG_RTX_P (dest) || FP_REG_RTX_P (src))
3457 if (!TARGET_64BIT && GET_MODE (dest) == DImode)
3458 emit_insn (gen_move_doubleword_fprdi (dest, src));
3459 else if (!TARGET_64BIT && GET_MODE (dest) == DFmode)
3460 emit_insn (gen_move_doubleword_fprdf (dest, src));
3461 else if (!TARGET_64BIT && GET_MODE (dest) == V2SFmode)
3462 emit_insn (gen_move_doubleword_fprv2sf (dest, src));
3463 else if (TARGET_64BIT && GET_MODE (dest) == TFmode)
3464 emit_insn (gen_move_doubleword_fprtf (dest, src));
3465 else
3466 gcc_unreachable ();
3468 else
3470 /* The operation can be split into two normal moves. Decide in
3471 which order to do them. */
3472 rtx low_dest;
3474 low_dest = mips_subword (dest, 0);
3475 if (REG_P (low_dest)
3476 && reg_overlap_mentioned_p (low_dest, src))
3478 mips_emit_move (mips_subword (dest, 1), mips_subword (src, 1));
3479 mips_emit_move (low_dest, mips_subword (src, 0));
3481 else
3483 mips_emit_move (low_dest, mips_subword (src, 0));
3484 mips_emit_move (mips_subword (dest, 1), mips_subword (src, 1));
3489 /* Return the appropriate instructions to move SRC into DEST. Assume
3490 that SRC is operand 1 and DEST is operand 0. */
3492 const char *
3493 mips_output_move (rtx dest, rtx src)
3495 enum rtx_code dest_code, src_code;
3496 enum mips_symbol_type symbol_type;
3497 bool dbl_p;
3499 dest_code = GET_CODE (dest);
3500 src_code = GET_CODE (src);
3501 dbl_p = (GET_MODE_SIZE (GET_MODE (dest)) == 8);
3503 if (dbl_p && mips_split_64bit_move_p (dest, src))
3504 return "#";
3506 if ((src_code == REG && GP_REG_P (REGNO (src)))
3507 || (!TARGET_MIPS16 && src == CONST0_RTX (GET_MODE (dest))))
3509 if (dest_code == REG)
3511 if (GP_REG_P (REGNO (dest)))
3512 return "move\t%0,%z1";
3514 if (MD_REG_P (REGNO (dest)))
3515 return "mt%0\t%z1";
3517 if (DSP_ACC_REG_P (REGNO (dest)))
3519 static char retval[] = "mt__\t%z1,%q0";
3520 retval[2] = reg_names[REGNO (dest)][4];
3521 retval[3] = reg_names[REGNO (dest)][5];
3522 return retval;
3525 if (FP_REG_P (REGNO (dest)))
3526 return (dbl_p ? "dmtc1\t%z1,%0" : "mtc1\t%z1,%0");
3528 if (ALL_COP_REG_P (REGNO (dest)))
3530 static char retval[] = "dmtc_\t%z1,%0";
3532 retval[4] = COPNUM_AS_CHAR_FROM_REGNUM (REGNO (dest));
3533 return (dbl_p ? retval : retval + 1);
3536 if (dest_code == MEM)
3537 return (dbl_p ? "sd\t%z1,%0" : "sw\t%z1,%0");
3539 if (dest_code == REG && GP_REG_P (REGNO (dest)))
3541 if (src_code == REG)
3543 if (DSP_ACC_REG_P (REGNO (src)))
3545 static char retval[] = "mf__\t%0,%q1";
3546 retval[2] = reg_names[REGNO (src)][4];
3547 retval[3] = reg_names[REGNO (src)][5];
3548 return retval;
3551 if (ST_REG_P (REGNO (src)) && ISA_HAS_8CC)
3552 return "lui\t%0,0x3f80\n\tmovf\t%0,%.,%1";
3554 if (FP_REG_P (REGNO (src)))
3555 return (dbl_p ? "dmfc1\t%0,%1" : "mfc1\t%0,%1");
3557 if (ALL_COP_REG_P (REGNO (src)))
3559 static char retval[] = "dmfc_\t%0,%1";
3561 retval[4] = COPNUM_AS_CHAR_FROM_REGNUM (REGNO (src));
3562 return (dbl_p ? retval : retval + 1);
3566 if (src_code == MEM)
3567 return (dbl_p ? "ld\t%0,%1" : "lw\t%0,%1");
3569 if (src_code == CONST_INT)
3571 /* Don't use the X format, because that will give out of
3572 range numbers for 64-bit hosts and 32-bit targets. */
3573 if (!TARGET_MIPS16)
3574 return "li\t%0,%1\t\t\t# %X1";
3576 if (INTVAL (src) >= 0 && INTVAL (src) <= 0xffff)
3577 return "li\t%0,%1";
3579 if (INTVAL (src) < 0 && INTVAL (src) >= -0xffff)
3580 return "#";
3583 if (src_code == HIGH)
3584 return TARGET_MIPS16 ? "#" : "lui\t%0,%h1";
3586 if (CONST_GP_P (src))
3587 return "move\t%0,%1";
3589 if (mips_symbolic_constant_p (src, SYMBOL_CONTEXT_LEA, &symbol_type)
3590 && mips_lo_relocs[symbol_type] != 0)
3592 /* A signed 16-bit constant formed by applying a relocation
3593 operator to a symbolic address. */
3594 gcc_assert (!mips_split_p[symbol_type]);
3595 return "li\t%0,%R1";
3598 if (symbolic_operand (src, VOIDmode))
3600 gcc_assert (TARGET_MIPS16
3601 ? TARGET_MIPS16_TEXT_LOADS
3602 : !TARGET_EXPLICIT_RELOCS);
3603 return (dbl_p ? "dla\t%0,%1" : "la\t%0,%1");
3606 if (src_code == REG && FP_REG_P (REGNO (src)))
3608 if (dest_code == REG && FP_REG_P (REGNO (dest)))
3610 if (GET_MODE (dest) == V2SFmode)
3611 return "mov.ps\t%0,%1";
3612 else
3613 return (dbl_p ? "mov.d\t%0,%1" : "mov.s\t%0,%1");
3616 if (dest_code == MEM)
3617 return (dbl_p ? "sdc1\t%1,%0" : "swc1\t%1,%0");
3619 if (dest_code == REG && FP_REG_P (REGNO (dest)))
3621 if (src_code == MEM)
3622 return (dbl_p ? "ldc1\t%0,%1" : "lwc1\t%0,%1");
3624 if (dest_code == REG && ALL_COP_REG_P (REGNO (dest)) && src_code == MEM)
3626 static char retval[] = "l_c_\t%0,%1";
3628 retval[1] = (dbl_p ? 'd' : 'w');
3629 retval[3] = COPNUM_AS_CHAR_FROM_REGNUM (REGNO (dest));
3630 return retval;
3632 if (dest_code == MEM && src_code == REG && ALL_COP_REG_P (REGNO (src)))
3634 static char retval[] = "s_c_\t%1,%0";
3636 retval[1] = (dbl_p ? 'd' : 'w');
3637 retval[3] = COPNUM_AS_CHAR_FROM_REGNUM (REGNO (src));
3638 return retval;
3640 gcc_unreachable ();
3643 /* Return true if CMP1 is a suitable second operand for relational
3644 operator CODE. See also the *sCC patterns in mips.md. */
3646 static bool
3647 mips_relational_operand_ok_p (enum rtx_code code, rtx cmp1)
3649 switch (code)
3651 case GT:
3652 case GTU:
3653 return reg_or_0_operand (cmp1, VOIDmode);
3655 case GE:
3656 case GEU:
3657 return !TARGET_MIPS16 && cmp1 == const1_rtx;
3659 case LT:
3660 case LTU:
3661 return arith_operand (cmp1, VOIDmode);
3663 case LE:
3664 return sle_operand (cmp1, VOIDmode);
3666 case LEU:
3667 return sleu_operand (cmp1, VOIDmode);
3669 default:
3670 gcc_unreachable ();
3674 /* Canonicalize LE or LEU comparisons into LT comparisons when
3675 possible to avoid extra instructions or inverting the
3676 comparison. */
3678 static bool
3679 mips_canonicalize_comparison (enum rtx_code *code, rtx *cmp1,
3680 enum machine_mode mode)
3682 HOST_WIDE_INT plus_one;
3684 if (mips_relational_operand_ok_p (*code, *cmp1))
3685 return true;
3687 if (GET_CODE (*cmp1) == CONST_INT)
3688 switch (*code)
3690 case LE:
3691 plus_one = trunc_int_for_mode (UINTVAL (*cmp1) + 1, mode);
3692 if (INTVAL (*cmp1) < plus_one)
3694 *code = LT;
3695 *cmp1 = force_reg (mode, GEN_INT (plus_one));
3696 return true;
3698 break;
3700 case LEU:
3701 plus_one = trunc_int_for_mode (UINTVAL (*cmp1) + 1, mode);
3702 if (plus_one != 0)
3704 *code = LTU;
3705 *cmp1 = force_reg (mode, GEN_INT (plus_one));
3706 return true;
3708 break;
3710 default:
3711 break;
3713 return false;
3716 /* Compare CMP0 and CMP1 using relational operator CODE and store the
3717 result in TARGET. CMP0 and TARGET are register_operands that have
3718 the same integer mode. If INVERT_PTR is nonnull, it's OK to set
3719 TARGET to the inverse of the result and flip *INVERT_PTR instead. */
3721 static void
3722 mips_emit_int_relational (enum rtx_code code, bool *invert_ptr,
3723 rtx target, rtx cmp0, rtx cmp1)
3725 /* First see if there is a MIPS instruction that can do this operation.
3726 If not, try doing the same for the inverse operation. If that also
3727 fails, force CMP1 into a register and try again. */
3728 if (mips_canonicalize_comparison (&code, &cmp1, GET_MODE (target)))
3729 mips_emit_binary (code, target, cmp0, cmp1);
3730 else
3732 enum rtx_code inv_code = reverse_condition (code);
3733 if (!mips_canonicalize_comparison (&inv_code, &cmp1, GET_MODE (target)))
3735 cmp1 = force_reg (GET_MODE (cmp0), cmp1);
3736 mips_emit_int_relational (code, invert_ptr, target, cmp0, cmp1);
3738 else if (invert_ptr == 0)
3740 rtx inv_target = gen_reg_rtx (GET_MODE (target));
3741 mips_emit_binary (inv_code, inv_target, cmp0, cmp1);
3742 mips_emit_binary (XOR, target, inv_target, const1_rtx);
3744 else
3746 *invert_ptr = !*invert_ptr;
3747 mips_emit_binary (inv_code, target, cmp0, cmp1);
3752 /* Return a register that is zero iff CMP0 and CMP1 are equal.
3753 The register will have the same mode as CMP0. */
3755 static rtx
3756 mips_zero_if_equal (rtx cmp0, rtx cmp1)
3758 if (cmp1 == const0_rtx)
3759 return cmp0;
3761 if (uns_arith_operand (cmp1, VOIDmode))
3762 return expand_binop (GET_MODE (cmp0), xor_optab,
3763 cmp0, cmp1, 0, 0, OPTAB_DIRECT);
3765 return expand_binop (GET_MODE (cmp0), sub_optab,
3766 cmp0, cmp1, 0, 0, OPTAB_DIRECT);
3769 /* Convert *CODE into a code that can be used in a floating-point
3770 scc instruction (c.<cond>.<fmt>). Return true if the values of
3771 the condition code registers will be inverted, with 0 indicating
3772 that the condition holds. */
3774 static bool
3775 mips_reverse_fp_cond_p (enum rtx_code *code)
3777 switch (*code)
3779 case NE:
3780 case LTGT:
3781 case ORDERED:
3782 *code = reverse_condition_maybe_unordered (*code);
3783 return true;
3785 default:
3786 return false;
3790 /* Convert a comparison into something that can be used in a branch or
3791 conditional move. cmp_operands[0] and cmp_operands[1] are the values
3792 being compared and *CODE is the code used to compare them.
3794 Update *CODE, *OP0 and *OP1 so that they describe the final comparison.
3795 If NEED_EQ_NE_P, then only EQ/NE comparisons against zero are possible,
3796 otherwise any standard branch condition can be used. The standard branch
3797 conditions are:
3799 - EQ/NE between two registers.
3800 - any comparison between a register and zero. */
3802 static void
3803 mips_emit_compare (enum rtx_code *code, rtx *op0, rtx *op1, bool need_eq_ne_p)
3805 if (GET_MODE_CLASS (GET_MODE (cmp_operands[0])) == MODE_INT)
3807 if (!need_eq_ne_p && cmp_operands[1] == const0_rtx)
3809 *op0 = cmp_operands[0];
3810 *op1 = cmp_operands[1];
3812 else if (*code == EQ || *code == NE)
3814 if (need_eq_ne_p)
3816 *op0 = mips_zero_if_equal (cmp_operands[0], cmp_operands[1]);
3817 *op1 = const0_rtx;
3819 else
3821 *op0 = cmp_operands[0];
3822 *op1 = force_reg (GET_MODE (*op0), cmp_operands[1]);
3825 else
3827 /* The comparison needs a separate scc instruction. Store the
3828 result of the scc in *OP0 and compare it against zero. */
3829 bool invert = false;
3830 *op0 = gen_reg_rtx (GET_MODE (cmp_operands[0]));
3831 *op1 = const0_rtx;
3832 mips_emit_int_relational (*code, &invert, *op0,
3833 cmp_operands[0], cmp_operands[1]);
3834 *code = (invert ? EQ : NE);
3837 else if (ALL_FIXED_POINT_MODE_P (GET_MODE (cmp_operands[0])))
3839 *op0 = gen_rtx_REG (CCDSPmode, CCDSP_CC_REGNUM);
3840 mips_emit_binary (*code, *op0, cmp_operands[0], cmp_operands[1]);
3841 *code = NE;
3842 *op1 = const0_rtx;
3844 else
3846 enum rtx_code cmp_code;
3848 /* Floating-point tests use a separate c.cond.fmt comparison to
3849 set a condition code register. The branch or conditional move
3850 will then compare that register against zero.
3852 Set CMP_CODE to the code of the comparison instruction and
3853 *CODE to the code that the branch or move should use. */
3854 cmp_code = *code;
3855 *code = mips_reverse_fp_cond_p (&cmp_code) ? EQ : NE;
3856 *op0 = (ISA_HAS_8CC
3857 ? gen_reg_rtx (CCmode)
3858 : gen_rtx_REG (CCmode, FPSW_REGNUM));
3859 *op1 = const0_rtx;
3860 mips_emit_binary (cmp_code, *op0, cmp_operands[0], cmp_operands[1]);
3864 /* Try comparing cmp_operands[0] and cmp_operands[1] using rtl code CODE.
3865 Store the result in TARGET and return true if successful.
3867 On 64-bit targets, TARGET may be wider than cmp_operands[0]. */
3869 bool
3870 mips_emit_scc (enum rtx_code code, rtx target)
3872 if (GET_MODE_CLASS (GET_MODE (cmp_operands[0])) != MODE_INT)
3873 return false;
3875 target = gen_lowpart (GET_MODE (cmp_operands[0]), target);
3876 if (code == EQ || code == NE)
3878 rtx zie = mips_zero_if_equal (cmp_operands[0], cmp_operands[1]);
3879 mips_emit_binary (code, target, zie, const0_rtx);
3881 else
3882 mips_emit_int_relational (code, 0, target,
3883 cmp_operands[0], cmp_operands[1]);
3884 return true;
3887 /* Emit the common code for doing conditional branches.
3888 operand[0] is the label to jump to.
3889 The comparison operands are saved away by cmp{si,di,sf,df}. */
3891 void
3892 gen_conditional_branch (rtx *operands, enum rtx_code code)
3894 rtx op0, op1, condition;
3896 mips_emit_compare (&code, &op0, &op1, TARGET_MIPS16);
3897 condition = gen_rtx_fmt_ee (code, VOIDmode, op0, op1);
3898 emit_jump_insn (gen_condjump (condition, operands[0]));
3901 /* Implement:
3903 (set temp (COND:CCV2 CMP_OP0 CMP_OP1))
3904 (set DEST (unspec [TRUE_SRC FALSE_SRC temp] UNSPEC_MOVE_TF_PS)) */
3906 void
3907 mips_expand_vcondv2sf (rtx dest, rtx true_src, rtx false_src,
3908 enum rtx_code cond, rtx cmp_op0, rtx cmp_op1)
3910 rtx cmp_result;
3911 bool reversed_p;
3913 reversed_p = mips_reverse_fp_cond_p (&cond);
3914 cmp_result = gen_reg_rtx (CCV2mode);
3915 emit_insn (gen_scc_ps (cmp_result,
3916 gen_rtx_fmt_ee (cond, VOIDmode, cmp_op0, cmp_op1)));
3917 if (reversed_p)
3918 emit_insn (gen_mips_cond_move_tf_ps (dest, false_src, true_src,
3919 cmp_result));
3920 else
3921 emit_insn (gen_mips_cond_move_tf_ps (dest, true_src, false_src,
3922 cmp_result));
3925 /* Emit the common code for conditional moves. OPERANDS is the array
3926 of operands passed to the conditional move define_expand. */
3928 void
3929 gen_conditional_move (rtx *operands)
3931 enum rtx_code code;
3932 rtx op0, op1;
3934 code = GET_CODE (operands[1]);
3935 mips_emit_compare (&code, &op0, &op1, true);
3936 emit_insn (gen_rtx_SET (VOIDmode, operands[0],
3937 gen_rtx_IF_THEN_ELSE (GET_MODE (operands[0]),
3938 gen_rtx_fmt_ee (code,
3939 GET_MODE (op0),
3940 op0, op1),
3941 operands[2], operands[3])));
3944 /* Emit a conditional trap. OPERANDS is the array of operands passed to
3945 the conditional_trap expander. */
3947 void
3948 mips_gen_conditional_trap (rtx *operands)
3950 rtx op0, op1;
3951 enum rtx_code cmp_code = GET_CODE (operands[0]);
3952 enum machine_mode mode = GET_MODE (cmp_operands[0]);
3954 /* MIPS conditional trap machine instructions don't have GT or LE
3955 flavors, so we must invert the comparison and convert to LT and
3956 GE, respectively. */
3957 switch (cmp_code)
3959 case GT: cmp_code = LT; break;
3960 case LE: cmp_code = GE; break;
3961 case GTU: cmp_code = LTU; break;
3962 case LEU: cmp_code = GEU; break;
3963 default: break;
3965 if (cmp_code == GET_CODE (operands[0]))
3967 op0 = cmp_operands[0];
3968 op1 = cmp_operands[1];
3970 else
3972 op0 = cmp_operands[1];
3973 op1 = cmp_operands[0];
3975 op0 = force_reg (mode, op0);
3976 if (!arith_operand (op1, mode))
3977 op1 = force_reg (mode, op1);
3979 emit_insn (gen_rtx_TRAP_IF (VOIDmode,
3980 gen_rtx_fmt_ee (cmp_code, mode, op0, op1),
3981 operands[1]));
3984 /* Argument support functions. */
3986 /* Initialize CUMULATIVE_ARGS for a function. */
3988 void
3989 init_cumulative_args (CUMULATIVE_ARGS *cum, tree fntype,
3990 rtx libname ATTRIBUTE_UNUSED)
3992 static CUMULATIVE_ARGS zero_cum;
3993 tree param, next_param;
3995 *cum = zero_cum;
3996 cum->prototype = (fntype && TYPE_ARG_TYPES (fntype));
3998 /* Determine if this function has variable arguments. This is
3999 indicated by the last argument being 'void_type_mode' if there
4000 are no variable arguments. The standard MIPS calling sequence
4001 passes all arguments in the general purpose registers in this case. */
4003 for (param = fntype ? TYPE_ARG_TYPES (fntype) : 0;
4004 param != 0; param = next_param)
4006 next_param = TREE_CHAIN (param);
4007 if (next_param == 0 && TREE_VALUE (param) != void_type_node)
4008 cum->gp_reg_found = 1;
4013 /* Fill INFO with information about a single argument. CUM is the
4014 cumulative state for earlier arguments. MODE is the mode of this
4015 argument and TYPE is its type (if known). NAMED is true if this
4016 is a named (fixed) argument rather than a variable one. */
4018 static void
4019 mips_arg_info (const CUMULATIVE_ARGS *cum, enum machine_mode mode,
4020 tree type, int named, struct mips_arg_info *info)
4022 bool doubleword_aligned_p;
4023 unsigned int num_bytes, num_words, max_regs;
4025 /* Work out the size of the argument. */
4026 num_bytes = type ? int_size_in_bytes (type) : GET_MODE_SIZE (mode);
4027 num_words = (num_bytes + UNITS_PER_WORD - 1) / UNITS_PER_WORD;
4029 /* Decide whether it should go in a floating-point register, assuming
4030 one is free. Later code checks for availability.
4032 The checks against UNITS_PER_FPVALUE handle the soft-float and
4033 single-float cases. */
4034 switch (mips_abi)
4036 case ABI_EABI:
4037 /* The EABI conventions have traditionally been defined in terms
4038 of TYPE_MODE, regardless of the actual type. */
4039 info->fpr_p = ((GET_MODE_CLASS (mode) == MODE_FLOAT
4040 || GET_MODE_CLASS (mode) == MODE_VECTOR_FLOAT)
4041 && GET_MODE_SIZE (mode) <= UNITS_PER_FPVALUE);
4042 break;
4044 case ABI_32:
4045 case ABI_O64:
4046 /* Only leading floating-point scalars are passed in
4047 floating-point registers. We also handle vector floats the same
4048 say, which is OK because they are not covered by the standard ABI. */
4049 info->fpr_p = (!cum->gp_reg_found
4050 && cum->arg_number < 2
4051 && (type == 0 || SCALAR_FLOAT_TYPE_P (type)
4052 || VECTOR_FLOAT_TYPE_P (type))
4053 && (GET_MODE_CLASS (mode) == MODE_FLOAT
4054 || GET_MODE_CLASS (mode) == MODE_VECTOR_FLOAT)
4055 && GET_MODE_SIZE (mode) <= UNITS_PER_FPVALUE);
4056 break;
4058 case ABI_N32:
4059 case ABI_64:
4060 /* Scalar and complex floating-point types are passed in
4061 floating-point registers. */
4062 info->fpr_p = (named
4063 && (type == 0 || FLOAT_TYPE_P (type))
4064 && (GET_MODE_CLASS (mode) == MODE_FLOAT
4065 || GET_MODE_CLASS (mode) == MODE_COMPLEX_FLOAT
4066 || GET_MODE_CLASS (mode) == MODE_VECTOR_FLOAT)
4067 && GET_MODE_UNIT_SIZE (mode) <= UNITS_PER_FPVALUE);
4069 /* ??? According to the ABI documentation, the real and imaginary
4070 parts of complex floats should be passed in individual registers.
4071 The real and imaginary parts of stack arguments are supposed
4072 to be contiguous and there should be an extra word of padding
4073 at the end.
4075 This has two problems. First, it makes it impossible to use a
4076 single "void *" va_list type, since register and stack arguments
4077 are passed differently. (At the time of writing, MIPSpro cannot
4078 handle complex float varargs correctly.) Second, it's unclear
4079 what should happen when there is only one register free.
4081 For now, we assume that named complex floats should go into FPRs
4082 if there are two FPRs free, otherwise they should be passed in the
4083 same way as a struct containing two floats. */
4084 if (info->fpr_p
4085 && GET_MODE_CLASS (mode) == MODE_COMPLEX_FLOAT
4086 && GET_MODE_UNIT_SIZE (mode) < UNITS_PER_FPVALUE)
4088 if (cum->num_gprs >= MAX_ARGS_IN_REGISTERS - 1)
4089 info->fpr_p = false;
4090 else
4091 num_words = 2;
4093 break;
4095 default:
4096 gcc_unreachable ();
4099 /* See whether the argument has doubleword alignment. */
4100 doubleword_aligned_p = FUNCTION_ARG_BOUNDARY (mode, type) > BITS_PER_WORD;
4102 /* Set REG_OFFSET to the register count we're interested in.
4103 The EABI allocates the floating-point registers separately,
4104 but the other ABIs allocate them like integer registers. */
4105 info->reg_offset = (mips_abi == ABI_EABI && info->fpr_p
4106 ? cum->num_fprs
4107 : cum->num_gprs);
4109 /* Advance to an even register if the argument is doubleword-aligned. */
4110 if (doubleword_aligned_p)
4111 info->reg_offset += info->reg_offset & 1;
4113 /* Work out the offset of a stack argument. */
4114 info->stack_offset = cum->stack_words;
4115 if (doubleword_aligned_p)
4116 info->stack_offset += info->stack_offset & 1;
4118 max_regs = MAX_ARGS_IN_REGISTERS - info->reg_offset;
4120 /* Partition the argument between registers and stack. */
4121 info->reg_words = MIN (num_words, max_regs);
4122 info->stack_words = num_words - info->reg_words;
4125 /* INFO describes an argument that is passed in a single-register value.
4126 Return the register it uses, assuming that FPRs are available if
4127 HARD_FLOAT_P. */
4129 static unsigned int
4130 mips_arg_regno (const struct mips_arg_info *info, bool hard_float_p)
4132 if (!info->fpr_p || !hard_float_p)
4133 return GP_ARG_FIRST + info->reg_offset;
4134 else if (mips_abi == ABI_32 && TARGET_DOUBLE_FLOAT && info->reg_offset > 0)
4135 /* In o32, the second argument is always passed in $f14
4136 for TARGET_DOUBLE_FLOAT, regardless of whether the
4137 first argument was a word or doubleword. */
4138 return FP_ARG_FIRST + 2;
4139 else
4140 return FP_ARG_FIRST + info->reg_offset;
4143 static bool
4144 mips_strict_argument_naming (CUMULATIVE_ARGS *ca ATTRIBUTE_UNUSED)
4146 return !TARGET_OLDABI;
4149 /* Implement FUNCTION_ARG. */
4151 struct rtx_def *
4152 function_arg (const CUMULATIVE_ARGS *cum, enum machine_mode mode,
4153 tree type, int named)
4155 struct mips_arg_info info;
4157 /* We will be called with a mode of VOIDmode after the last argument
4158 has been seen. Whatever we return will be passed to the call
4159 insn. If we need a mips16 fp_code, return a REG with the code
4160 stored as the mode. */
4161 if (mode == VOIDmode)
4163 if (TARGET_MIPS16 && cum->fp_code != 0)
4164 return gen_rtx_REG ((enum machine_mode) cum->fp_code, 0);
4166 else
4167 return 0;
4170 mips_arg_info (cum, mode, type, named, &info);
4172 /* Return straight away if the whole argument is passed on the stack. */
4173 if (info.reg_offset == MAX_ARGS_IN_REGISTERS)
4174 return 0;
4176 if (type != 0
4177 && TREE_CODE (type) == RECORD_TYPE
4178 && TARGET_NEWABI
4179 && TYPE_SIZE_UNIT (type)
4180 && host_integerp (TYPE_SIZE_UNIT (type), 1)
4181 && named)
4183 /* The Irix 6 n32/n64 ABIs say that if any 64-bit chunk of the
4184 structure contains a double in its entirety, then that 64-bit
4185 chunk is passed in a floating point register. */
4186 tree field;
4188 /* First check to see if there is any such field. */
4189 for (field = TYPE_FIELDS (type); field; field = TREE_CHAIN (field))
4190 if (TREE_CODE (field) == FIELD_DECL
4191 && TREE_CODE (TREE_TYPE (field)) == REAL_TYPE
4192 && TYPE_PRECISION (TREE_TYPE (field)) == BITS_PER_WORD
4193 && host_integerp (bit_position (field), 0)
4194 && int_bit_position (field) % BITS_PER_WORD == 0)
4195 break;
4197 if (field != 0)
4199 /* Now handle the special case by returning a PARALLEL
4200 indicating where each 64-bit chunk goes. INFO.REG_WORDS
4201 chunks are passed in registers. */
4202 unsigned int i;
4203 HOST_WIDE_INT bitpos;
4204 rtx ret;
4206 /* assign_parms checks the mode of ENTRY_PARM, so we must
4207 use the actual mode here. */
4208 ret = gen_rtx_PARALLEL (mode, rtvec_alloc (info.reg_words));
4210 bitpos = 0;
4211 field = TYPE_FIELDS (type);
4212 for (i = 0; i < info.reg_words; i++)
4214 rtx reg;
4216 for (; field; field = TREE_CHAIN (field))
4217 if (TREE_CODE (field) == FIELD_DECL
4218 && int_bit_position (field) >= bitpos)
4219 break;
4221 if (field
4222 && int_bit_position (field) == bitpos
4223 && TREE_CODE (TREE_TYPE (field)) == REAL_TYPE
4224 && !TARGET_SOFT_FLOAT
4225 && TYPE_PRECISION (TREE_TYPE (field)) == BITS_PER_WORD)
4226 reg = gen_rtx_REG (DFmode, FP_ARG_FIRST + info.reg_offset + i);
4227 else
4228 reg = gen_rtx_REG (DImode, GP_ARG_FIRST + info.reg_offset + i);
4230 XVECEXP (ret, 0, i)
4231 = gen_rtx_EXPR_LIST (VOIDmode, reg,
4232 GEN_INT (bitpos / BITS_PER_UNIT));
4234 bitpos += BITS_PER_WORD;
4236 return ret;
4240 /* Handle the n32/n64 conventions for passing complex floating-point
4241 arguments in FPR pairs. The real part goes in the lower register
4242 and the imaginary part goes in the upper register. */
4243 if (TARGET_NEWABI
4244 && info.fpr_p
4245 && GET_MODE_CLASS (mode) == MODE_COMPLEX_FLOAT)
4247 rtx real, imag;
4248 enum machine_mode inner;
4249 int reg;
4251 inner = GET_MODE_INNER (mode);
4252 reg = FP_ARG_FIRST + info.reg_offset;
4253 if (info.reg_words * UNITS_PER_WORD == GET_MODE_SIZE (inner))
4255 /* Real part in registers, imaginary part on stack. */
4256 gcc_assert (info.stack_words == info.reg_words);
4257 return gen_rtx_REG (inner, reg);
4259 else
4261 gcc_assert (info.stack_words == 0);
4262 real = gen_rtx_EXPR_LIST (VOIDmode,
4263 gen_rtx_REG (inner, reg),
4264 const0_rtx);
4265 imag = gen_rtx_EXPR_LIST (VOIDmode,
4266 gen_rtx_REG (inner,
4267 reg + info.reg_words / 2),
4268 GEN_INT (GET_MODE_SIZE (inner)));
4269 return gen_rtx_PARALLEL (mode, gen_rtvec (2, real, imag));
4273 return gen_rtx_REG (mode, mips_arg_regno (&info, TARGET_HARD_FLOAT));
4276 /* Implement FUNCTION_ARG_ADVANCE. */
4278 void
4279 function_arg_advance (CUMULATIVE_ARGS *cum, enum machine_mode mode,
4280 tree type, int named)
4282 struct mips_arg_info info;
4284 mips_arg_info (cum, mode, type, named, &info);
4286 if (!info.fpr_p)
4287 cum->gp_reg_found = true;
4289 /* See the comment above the cumulative args structure in mips.h
4290 for an explanation of what this code does. It assumes the O32
4291 ABI, which passes at most 2 arguments in float registers. */
4292 if (cum->arg_number < 2 && info.fpr_p)
4293 cum->fp_code += (mode == SFmode ? 1 : 2) << (cum->arg_number * 2);
4295 if (mips_abi != ABI_EABI || !info.fpr_p)
4296 cum->num_gprs = info.reg_offset + info.reg_words;
4297 else if (info.reg_words > 0)
4298 cum->num_fprs += MAX_FPRS_PER_FMT;
4300 if (info.stack_words > 0)
4301 cum->stack_words = info.stack_offset + info.stack_words;
4303 cum->arg_number++;
4306 /* Implement TARGET_ARG_PARTIAL_BYTES. */
4308 static int
4309 mips_arg_partial_bytes (CUMULATIVE_ARGS *cum,
4310 enum machine_mode mode, tree type, bool named)
4312 struct mips_arg_info info;
4314 mips_arg_info (cum, mode, type, named, &info);
4315 return info.stack_words > 0 ? info.reg_words * UNITS_PER_WORD : 0;
4319 /* Implement FUNCTION_ARG_BOUNDARY. Every parameter gets at least
4320 PARM_BOUNDARY bits of alignment, but will be given anything up
4321 to STACK_BOUNDARY bits if the type requires it. */
4324 function_arg_boundary (enum machine_mode mode, tree type)
4326 unsigned int alignment;
4328 alignment = type ? TYPE_ALIGN (type) : GET_MODE_ALIGNMENT (mode);
4329 if (alignment < PARM_BOUNDARY)
4330 alignment = PARM_BOUNDARY;
4331 if (alignment > STACK_BOUNDARY)
4332 alignment = STACK_BOUNDARY;
4333 return alignment;
4336 /* Return true if FUNCTION_ARG_PADDING (MODE, TYPE) should return
4337 upward rather than downward. In other words, return true if the
4338 first byte of the stack slot has useful data, false if the last
4339 byte does. */
4341 bool
4342 mips_pad_arg_upward (enum machine_mode mode, const_tree type)
4344 /* On little-endian targets, the first byte of every stack argument
4345 is passed in the first byte of the stack slot. */
4346 if (!BYTES_BIG_ENDIAN)
4347 return true;
4349 /* Otherwise, integral types are padded downward: the last byte of a
4350 stack argument is passed in the last byte of the stack slot. */
4351 if (type != 0
4352 ? (INTEGRAL_TYPE_P (type)
4353 || POINTER_TYPE_P (type)
4354 || FIXED_POINT_TYPE_P (type))
4355 : (GET_MODE_CLASS (mode) == MODE_INT
4356 || ALL_SCALAR_FIXED_POINT_MODE_P (mode)))
4357 return false;
4359 /* Big-endian o64 pads floating-point arguments downward. */
4360 if (mips_abi == ABI_O64)
4361 if (type != 0 ? FLOAT_TYPE_P (type) : GET_MODE_CLASS (mode) == MODE_FLOAT)
4362 return false;
4364 /* Other types are padded upward for o32, o64, n32 and n64. */
4365 if (mips_abi != ABI_EABI)
4366 return true;
4368 /* Arguments smaller than a stack slot are padded downward. */
4369 if (mode != BLKmode)
4370 return (GET_MODE_BITSIZE (mode) >= PARM_BOUNDARY);
4371 else
4372 return (int_size_in_bytes (type) >= (PARM_BOUNDARY / BITS_PER_UNIT));
4376 /* Likewise BLOCK_REG_PADDING (MODE, TYPE, ...). Return !BYTES_BIG_ENDIAN
4377 if the least significant byte of the register has useful data. Return
4378 the opposite if the most significant byte does. */
4380 bool
4381 mips_pad_reg_upward (enum machine_mode mode, tree type)
4383 /* No shifting is required for floating-point arguments. */
4384 if (type != 0 ? FLOAT_TYPE_P (type) : GET_MODE_CLASS (mode) == MODE_FLOAT)
4385 return !BYTES_BIG_ENDIAN;
4387 /* Otherwise, apply the same padding to register arguments as we do
4388 to stack arguments. */
4389 return mips_pad_arg_upward (mode, type);
4393 /* Return nonzero when an argument must be passed by reference. */
4395 static bool
4396 mips_pass_by_reference (CUMULATIVE_ARGS *cum ATTRIBUTE_UNUSED,
4397 enum machine_mode mode, const_tree type,
4398 bool named ATTRIBUTE_UNUSED)
4400 if (mips_abi == ABI_EABI)
4402 int size;
4404 /* ??? How should SCmode be handled? */
4405 if (mode == DImode || mode == DFmode
4406 || mode == DQmode || mode == UDQmode
4407 || mode == DAmode || mode == UDAmode)
4408 return 0;
4410 size = type ? int_size_in_bytes (type) : GET_MODE_SIZE (mode);
4411 return size == -1 || size > UNITS_PER_WORD;
4413 else
4415 /* If we have a variable-sized parameter, we have no choice. */
4416 return targetm.calls.must_pass_in_stack (mode, type);
4420 static bool
4421 mips_callee_copies (CUMULATIVE_ARGS *cum ATTRIBUTE_UNUSED,
4422 enum machine_mode mode ATTRIBUTE_UNUSED,
4423 const_tree type ATTRIBUTE_UNUSED, bool named)
4425 return mips_abi == ABI_EABI && named;
4428 /* See whether VALTYPE is a record whose fields should be returned in
4429 floating-point registers. If so, return the number of fields and
4430 list them in FIELDS (which should have two elements). Return 0
4431 otherwise.
4433 For n32 & n64, a structure with one or two fields is returned in
4434 floating-point registers as long as every field has a floating-point
4435 type. */
4437 static int
4438 mips_fpr_return_fields (const_tree valtype, tree *fields)
4440 tree field;
4441 int i;
4443 if (!TARGET_NEWABI)
4444 return 0;
4446 if (TREE_CODE (valtype) != RECORD_TYPE)
4447 return 0;
4449 i = 0;
4450 for (field = TYPE_FIELDS (valtype); field != 0; field = TREE_CHAIN (field))
4452 if (TREE_CODE (field) != FIELD_DECL)
4453 continue;
4455 if (TREE_CODE (TREE_TYPE (field)) != REAL_TYPE)
4456 return 0;
4458 if (i == 2)
4459 return 0;
4461 fields[i++] = field;
4463 return i;
4467 /* Implement TARGET_RETURN_IN_MSB. For n32 & n64, we should return
4468 a value in the most significant part of $2/$3 if:
4470 - the target is big-endian;
4472 - the value has a structure or union type (we generalize this to
4473 cover aggregates from other languages too); and
4475 - the structure is not returned in floating-point registers. */
4477 static bool
4478 mips_return_in_msb (const_tree valtype)
4480 tree fields[2];
4482 return (TARGET_NEWABI
4483 && TARGET_BIG_ENDIAN
4484 && AGGREGATE_TYPE_P (valtype)
4485 && mips_fpr_return_fields (valtype, fields) == 0);
4489 /* Return true if the function return value MODE will get returned in a
4490 floating-point register. */
4492 static bool
4493 mips_return_mode_in_fpr_p (enum machine_mode mode)
4495 return ((GET_MODE_CLASS (mode) == MODE_FLOAT
4496 || GET_MODE_CLASS (mode) == MODE_VECTOR_FLOAT
4497 || GET_MODE_CLASS (mode) == MODE_COMPLEX_FLOAT)
4498 && GET_MODE_UNIT_SIZE (mode) <= UNITS_PER_HWFPVALUE);
4501 /* Return a composite value in a pair of floating-point registers.
4502 MODE1 and OFFSET1 are the mode and byte offset for the first value,
4503 likewise MODE2 and OFFSET2 for the second. MODE is the mode of the
4504 complete value.
4506 For n32 & n64, $f0 always holds the first value and $f2 the second.
4507 Otherwise the values are packed together as closely as possible. */
4509 static rtx
4510 mips_return_fpr_pair (enum machine_mode mode,
4511 enum machine_mode mode1, HOST_WIDE_INT offset1,
4512 enum machine_mode mode2, HOST_WIDE_INT offset2)
4514 int inc;
4516 inc = (TARGET_NEWABI ? 2 : MAX_FPRS_PER_FMT);
4517 return gen_rtx_PARALLEL
4518 (mode,
4519 gen_rtvec (2,
4520 gen_rtx_EXPR_LIST (VOIDmode,
4521 gen_rtx_REG (mode1, FP_RETURN),
4522 GEN_INT (offset1)),
4523 gen_rtx_EXPR_LIST (VOIDmode,
4524 gen_rtx_REG (mode2, FP_RETURN + inc),
4525 GEN_INT (offset2))));
4530 /* Implement FUNCTION_VALUE and LIBCALL_VALUE. For normal calls,
4531 VALTYPE is the return type and MODE is VOIDmode. For libcalls,
4532 VALTYPE is null and MODE is the mode of the return value. */
4535 mips_function_value (const_tree valtype, const_tree func ATTRIBUTE_UNUSED,
4536 enum machine_mode mode)
4538 if (valtype)
4540 tree fields[2];
4541 int unsignedp;
4543 mode = TYPE_MODE (valtype);
4544 unsignedp = TYPE_UNSIGNED (valtype);
4546 /* Since we define TARGET_PROMOTE_FUNCTION_RETURN that returns
4547 true, we must promote the mode just as PROMOTE_MODE does. */
4548 mode = promote_mode (valtype, mode, &unsignedp, 1);
4550 /* Handle structures whose fields are returned in $f0/$f2. */
4551 switch (mips_fpr_return_fields (valtype, fields))
4553 case 1:
4554 return gen_rtx_REG (mode, FP_RETURN);
4556 case 2:
4557 return mips_return_fpr_pair (mode,
4558 TYPE_MODE (TREE_TYPE (fields[0])),
4559 int_byte_position (fields[0]),
4560 TYPE_MODE (TREE_TYPE (fields[1])),
4561 int_byte_position (fields[1]));
4564 /* If a value is passed in the most significant part of a register, see
4565 whether we have to round the mode up to a whole number of words. */
4566 if (mips_return_in_msb (valtype))
4568 HOST_WIDE_INT size = int_size_in_bytes (valtype);
4569 if (size % UNITS_PER_WORD != 0)
4571 size += UNITS_PER_WORD - size % UNITS_PER_WORD;
4572 mode = mode_for_size (size * BITS_PER_UNIT, MODE_INT, 0);
4576 /* For EABI, the class of return register depends entirely on MODE.
4577 For example, "struct { some_type x; }" and "union { some_type x; }"
4578 are returned in the same way as a bare "some_type" would be.
4579 Other ABIs only use FPRs for scalar, complex or vector types. */
4580 if (mips_abi != ABI_EABI && !FLOAT_TYPE_P (valtype))
4581 return gen_rtx_REG (mode, GP_RETURN);
4584 if (!TARGET_MIPS16)
4586 /* Handle long doubles for n32 & n64. */
4587 if (mode == TFmode)
4588 return mips_return_fpr_pair (mode,
4589 DImode, 0,
4590 DImode, GET_MODE_SIZE (mode) / 2);
4592 if (mips_return_mode_in_fpr_p (mode))
4594 if (GET_MODE_CLASS (mode) == MODE_COMPLEX_FLOAT)
4595 return mips_return_fpr_pair (mode,
4596 GET_MODE_INNER (mode), 0,
4597 GET_MODE_INNER (mode),
4598 GET_MODE_SIZE (mode) / 2);
4599 else
4600 return gen_rtx_REG (mode, FP_RETURN);
4604 return gen_rtx_REG (mode, GP_RETURN);
4607 /* Implement TARGET_RETURN_IN_MEMORY. Under the old (i.e., 32 and O64 ABIs)
4608 all BLKmode objects are returned in memory. Under the new (N32 and
4609 64-bit MIPS ABIs) small structures are returned in a register.
4610 Objects with varying size must still be returned in memory, of
4611 course. */
4613 static bool
4614 mips_return_in_memory (const_tree type, const_tree fndecl ATTRIBUTE_UNUSED)
4616 if (TARGET_OLDABI)
4617 return (TYPE_MODE (type) == BLKmode);
4618 else
4619 return ((int_size_in_bytes (type) > (2 * UNITS_PER_WORD))
4620 || (int_size_in_bytes (type) == -1));
4623 static void
4624 mips_setup_incoming_varargs (CUMULATIVE_ARGS *cum, enum machine_mode mode,
4625 tree type, int *pretend_size ATTRIBUTE_UNUSED,
4626 int no_rtl)
4628 CUMULATIVE_ARGS local_cum;
4629 int gp_saved, fp_saved;
4631 /* The caller has advanced CUM up to, but not beyond, the last named
4632 argument. Advance a local copy of CUM past the last "real" named
4633 argument, to find out how many registers are left over. */
4635 local_cum = *cum;
4636 FUNCTION_ARG_ADVANCE (local_cum, mode, type, 1);
4638 /* Found out how many registers we need to save. */
4639 gp_saved = MAX_ARGS_IN_REGISTERS - local_cum.num_gprs;
4640 fp_saved = (EABI_FLOAT_VARARGS_P
4641 ? MAX_ARGS_IN_REGISTERS - local_cum.num_fprs
4642 : 0);
4644 if (!no_rtl)
4646 if (gp_saved > 0)
4648 rtx ptr, mem;
4650 ptr = plus_constant (virtual_incoming_args_rtx,
4651 REG_PARM_STACK_SPACE (cfun->decl)
4652 - gp_saved * UNITS_PER_WORD);
4653 mem = gen_frame_mem (BLKmode, ptr);
4654 set_mem_alias_set (mem, get_varargs_alias_set ());
4656 move_block_from_reg (local_cum.num_gprs + GP_ARG_FIRST,
4657 mem, gp_saved);
4659 if (fp_saved > 0)
4661 /* We can't use move_block_from_reg, because it will use
4662 the wrong mode. */
4663 enum machine_mode mode;
4664 int off, i;
4666 /* Set OFF to the offset from virtual_incoming_args_rtx of
4667 the first float register. The FP save area lies below
4668 the integer one, and is aligned to UNITS_PER_FPVALUE bytes. */
4669 off = -gp_saved * UNITS_PER_WORD;
4670 off &= ~(UNITS_PER_FPVALUE - 1);
4671 off -= fp_saved * UNITS_PER_FPREG;
4673 mode = TARGET_SINGLE_FLOAT ? SFmode : DFmode;
4675 for (i = local_cum.num_fprs; i < MAX_ARGS_IN_REGISTERS;
4676 i += MAX_FPRS_PER_FMT)
4678 rtx ptr, mem;
4680 ptr = plus_constant (virtual_incoming_args_rtx, off);
4681 mem = gen_frame_mem (mode, ptr);
4682 set_mem_alias_set (mem, get_varargs_alias_set ());
4683 mips_emit_move (mem, gen_rtx_REG (mode, FP_ARG_FIRST + i));
4684 off += UNITS_PER_HWFPVALUE;
4688 if (REG_PARM_STACK_SPACE (cfun->decl) == 0)
4689 cfun->machine->varargs_size = (gp_saved * UNITS_PER_WORD
4690 + fp_saved * UNITS_PER_FPREG);
4693 /* Create the va_list data type.
4694 We keep 3 pointers, and two offsets.
4695 Two pointers are to the overflow area, which starts at the CFA.
4696 One of these is constant, for addressing into the GPR save area below it.
4697 The other is advanced up the stack through the overflow region.
4698 The third pointer is to the GPR save area. Since the FPR save area
4699 is just below it, we can address FPR slots off this pointer.
4700 We also keep two one-byte offsets, which are to be subtracted from the
4701 constant pointers to yield addresses in the GPR and FPR save areas.
4702 These are downcounted as float or non-float arguments are used,
4703 and when they get to zero, the argument must be obtained from the
4704 overflow region.
4705 If !EABI_FLOAT_VARARGS_P, then no FPR save area exists, and a single
4706 pointer is enough. It's started at the GPR save area, and is
4707 advanced, period.
4708 Note that the GPR save area is not constant size, due to optimization
4709 in the prologue. Hence, we can't use a design with two pointers
4710 and two offsets, although we could have designed this with two pointers
4711 and three offsets. */
4713 static tree
4714 mips_build_builtin_va_list (void)
4716 if (EABI_FLOAT_VARARGS_P)
4718 tree f_ovfl, f_gtop, f_ftop, f_goff, f_foff, f_res, record;
4719 tree array, index;
4721 record = (*lang_hooks.types.make_type) (RECORD_TYPE);
4723 f_ovfl = build_decl (FIELD_DECL, get_identifier ("__overflow_argptr"),
4724 ptr_type_node);
4725 f_gtop = build_decl (FIELD_DECL, get_identifier ("__gpr_top"),
4726 ptr_type_node);
4727 f_ftop = build_decl (FIELD_DECL, get_identifier ("__fpr_top"),
4728 ptr_type_node);
4729 f_goff = build_decl (FIELD_DECL, get_identifier ("__gpr_offset"),
4730 unsigned_char_type_node);
4731 f_foff = build_decl (FIELD_DECL, get_identifier ("__fpr_offset"),
4732 unsigned_char_type_node);
4733 /* Explicitly pad to the size of a pointer, so that -Wpadded won't
4734 warn on every user file. */
4735 index = build_int_cst (NULL_TREE, GET_MODE_SIZE (ptr_mode) - 2 - 1);
4736 array = build_array_type (unsigned_char_type_node,
4737 build_index_type (index));
4738 f_res = build_decl (FIELD_DECL, get_identifier ("__reserved"), array);
4740 DECL_FIELD_CONTEXT (f_ovfl) = record;
4741 DECL_FIELD_CONTEXT (f_gtop) = record;
4742 DECL_FIELD_CONTEXT (f_ftop) = record;
4743 DECL_FIELD_CONTEXT (f_goff) = record;
4744 DECL_FIELD_CONTEXT (f_foff) = record;
4745 DECL_FIELD_CONTEXT (f_res) = record;
4747 TYPE_FIELDS (record) = f_ovfl;
4748 TREE_CHAIN (f_ovfl) = f_gtop;
4749 TREE_CHAIN (f_gtop) = f_ftop;
4750 TREE_CHAIN (f_ftop) = f_goff;
4751 TREE_CHAIN (f_goff) = f_foff;
4752 TREE_CHAIN (f_foff) = f_res;
4754 layout_type (record);
4755 return record;
4757 else if (TARGET_IRIX && TARGET_IRIX6)
4758 /* On IRIX 6, this type is 'char *'. */
4759 return build_pointer_type (char_type_node);
4760 else
4761 /* Otherwise, we use 'void *'. */
4762 return ptr_type_node;
4765 /* Implement va_start. */
4767 void
4768 mips_va_start (tree valist, rtx nextarg)
4770 if (EABI_FLOAT_VARARGS_P)
4772 const CUMULATIVE_ARGS *cum;
4773 tree f_ovfl, f_gtop, f_ftop, f_goff, f_foff;
4774 tree ovfl, gtop, ftop, goff, foff;
4775 tree t;
4776 int gpr_save_area_size;
4777 int fpr_save_area_size;
4778 int fpr_offset;
4780 cum = &current_function_args_info;
4781 gpr_save_area_size
4782 = (MAX_ARGS_IN_REGISTERS - cum->num_gprs) * UNITS_PER_WORD;
4783 fpr_save_area_size
4784 = (MAX_ARGS_IN_REGISTERS - cum->num_fprs) * UNITS_PER_FPREG;
4786 f_ovfl = TYPE_FIELDS (va_list_type_node);
4787 f_gtop = TREE_CHAIN (f_ovfl);
4788 f_ftop = TREE_CHAIN (f_gtop);
4789 f_goff = TREE_CHAIN (f_ftop);
4790 f_foff = TREE_CHAIN (f_goff);
4792 ovfl = build3 (COMPONENT_REF, TREE_TYPE (f_ovfl), valist, f_ovfl,
4793 NULL_TREE);
4794 gtop = build3 (COMPONENT_REF, TREE_TYPE (f_gtop), valist, f_gtop,
4795 NULL_TREE);
4796 ftop = build3 (COMPONENT_REF, TREE_TYPE (f_ftop), valist, f_ftop,
4797 NULL_TREE);
4798 goff = build3 (COMPONENT_REF, TREE_TYPE (f_goff), valist, f_goff,
4799 NULL_TREE);
4800 foff = build3 (COMPONENT_REF, TREE_TYPE (f_foff), valist, f_foff,
4801 NULL_TREE);
4803 /* Emit code to initialize OVFL, which points to the next varargs
4804 stack argument. CUM->STACK_WORDS gives the number of stack
4805 words used by named arguments. */
4806 t = make_tree (TREE_TYPE (ovfl), virtual_incoming_args_rtx);
4807 if (cum->stack_words > 0)
4808 t = build2 (POINTER_PLUS_EXPR, TREE_TYPE (ovfl), t,
4809 size_int (cum->stack_words * UNITS_PER_WORD));
4810 t = build2 (GIMPLE_MODIFY_STMT, TREE_TYPE (ovfl), ovfl, t);
4811 expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
4813 /* Emit code to initialize GTOP, the top of the GPR save area. */
4814 t = make_tree (TREE_TYPE (gtop), virtual_incoming_args_rtx);
4815 t = build2 (GIMPLE_MODIFY_STMT, TREE_TYPE (gtop), gtop, t);
4816 expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
4818 /* Emit code to initialize FTOP, the top of the FPR save area.
4819 This address is gpr_save_area_bytes below GTOP, rounded
4820 down to the next fp-aligned boundary. */
4821 t = make_tree (TREE_TYPE (ftop), virtual_incoming_args_rtx);
4822 fpr_offset = gpr_save_area_size + UNITS_PER_FPVALUE - 1;
4823 fpr_offset &= ~(UNITS_PER_FPVALUE - 1);
4824 if (fpr_offset)
4825 t = build2 (POINTER_PLUS_EXPR, TREE_TYPE (ftop), t,
4826 size_int (-fpr_offset));
4827 t = build2 (GIMPLE_MODIFY_STMT, TREE_TYPE (ftop), ftop, t);
4828 expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
4830 /* Emit code to initialize GOFF, the offset from GTOP of the
4831 next GPR argument. */
4832 t = build2 (GIMPLE_MODIFY_STMT, TREE_TYPE (goff), goff,
4833 build_int_cst (NULL_TREE, gpr_save_area_size));
4834 expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
4836 /* Likewise emit code to initialize FOFF, the offset from FTOP
4837 of the next FPR argument. */
4838 t = build2 (GIMPLE_MODIFY_STMT, TREE_TYPE (foff), foff,
4839 build_int_cst (NULL_TREE, fpr_save_area_size));
4840 expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
4842 else
4844 nextarg = plus_constant (nextarg, -cfun->machine->varargs_size);
4845 std_expand_builtin_va_start (valist, nextarg);
4849 /* Implement va_arg. */
4851 static tree
4852 mips_gimplify_va_arg_expr (tree valist, tree type, tree *pre_p, tree *post_p)
4854 HOST_WIDE_INT size, rsize;
4855 tree addr;
4856 bool indirect;
4858 indirect = pass_by_reference (NULL, TYPE_MODE (type), type, 0);
4860 if (indirect)
4861 type = build_pointer_type (type);
4863 size = int_size_in_bytes (type);
4864 rsize = (size + UNITS_PER_WORD - 1) & -UNITS_PER_WORD;
4866 if (mips_abi != ABI_EABI || !EABI_FLOAT_VARARGS_P)
4867 addr = std_gimplify_va_arg_expr (valist, type, pre_p, post_p);
4868 else
4870 /* Not a simple merged stack. */
4872 tree f_ovfl, f_gtop, f_ftop, f_goff, f_foff;
4873 tree ovfl, top, off, align;
4874 HOST_WIDE_INT osize;
4875 tree t, u;
4877 f_ovfl = TYPE_FIELDS (va_list_type_node);
4878 f_gtop = TREE_CHAIN (f_ovfl);
4879 f_ftop = TREE_CHAIN (f_gtop);
4880 f_goff = TREE_CHAIN (f_ftop);
4881 f_foff = TREE_CHAIN (f_goff);
4883 /* We maintain separate pointers and offsets for floating-point
4884 and integer arguments, but we need similar code in both cases.
4885 Let:
4887 TOP be the top of the register save area;
4888 OFF be the offset from TOP of the next register;
4889 ADDR_RTX be the address of the argument;
4890 RSIZE be the number of bytes used to store the argument
4891 when it's in the register save area;
4892 OSIZE be the number of bytes used to store it when it's
4893 in the stack overflow area; and
4894 PADDING be (BYTES_BIG_ENDIAN ? OSIZE - RSIZE : 0)
4896 The code we want is:
4898 1: off &= -rsize; // round down
4899 2: if (off != 0)
4900 3: {
4901 4: addr_rtx = top - off;
4902 5: off -= rsize;
4903 6: }
4904 7: else
4905 8: {
4906 9: ovfl += ((intptr_t) ovfl + osize - 1) & -osize;
4907 10: addr_rtx = ovfl + PADDING;
4908 11: ovfl += osize;
4909 14: }
4911 [1] and [9] can sometimes be optimized away. */
4913 ovfl = build3 (COMPONENT_REF, TREE_TYPE (f_ovfl), valist, f_ovfl,
4914 NULL_TREE);
4916 if (GET_MODE_CLASS (TYPE_MODE (type)) == MODE_FLOAT
4917 && GET_MODE_SIZE (TYPE_MODE (type)) <= UNITS_PER_FPVALUE)
4919 top = build3 (COMPONENT_REF, TREE_TYPE (f_ftop), valist, f_ftop,
4920 NULL_TREE);
4921 off = build3 (COMPONENT_REF, TREE_TYPE (f_foff), valist, f_foff,
4922 NULL_TREE);
4924 /* When floating-point registers are saved to the stack,
4925 each one will take up UNITS_PER_HWFPVALUE bytes, regardless
4926 of the float's precision. */
4927 rsize = UNITS_PER_HWFPVALUE;
4929 /* Overflow arguments are padded to UNITS_PER_WORD bytes
4930 (= PARM_BOUNDARY bits). This can be different from RSIZE
4931 in two cases:
4933 (1) On 32-bit targets when TYPE is a structure such as:
4935 struct s { float f; };
4937 Such structures are passed in paired FPRs, so RSIZE
4938 will be 8 bytes. However, the structure only takes
4939 up 4 bytes of memory, so OSIZE will only be 4.
4941 (2) In combinations such as -mgp64 -msingle-float
4942 -fshort-double. Doubles passed in registers
4943 will then take up 4 (UNITS_PER_HWFPVALUE) bytes,
4944 but those passed on the stack take up
4945 UNITS_PER_WORD bytes. */
4946 osize = MAX (GET_MODE_SIZE (TYPE_MODE (type)), UNITS_PER_WORD);
4948 else
4950 top = build3 (COMPONENT_REF, TREE_TYPE (f_gtop), valist, f_gtop,
4951 NULL_TREE);
4952 off = build3 (COMPONENT_REF, TREE_TYPE (f_goff), valist, f_goff,
4953 NULL_TREE);
4954 if (rsize > UNITS_PER_WORD)
4956 /* [1] Emit code for: off &= -rsize. */
4957 t = build2 (BIT_AND_EXPR, TREE_TYPE (off), off,
4958 build_int_cst (NULL_TREE, -rsize));
4959 t = build2 (GIMPLE_MODIFY_STMT, TREE_TYPE (off), off, t);
4960 gimplify_and_add (t, pre_p);
4962 osize = rsize;
4965 /* [2] Emit code to branch if off == 0. */
4966 t = build2 (NE_EXPR, boolean_type_node, off,
4967 build_int_cst (TREE_TYPE (off), 0));
4968 addr = build3 (COND_EXPR, ptr_type_node, t, NULL_TREE, NULL_TREE);
4970 /* [5] Emit code for: off -= rsize. We do this as a form of
4971 post-increment not available to C. Also widen for the
4972 coming pointer arithmetic. */
4973 t = fold_convert (TREE_TYPE (off), build_int_cst (NULL_TREE, rsize));
4974 t = build2 (POSTDECREMENT_EXPR, TREE_TYPE (off), off, t);
4975 t = fold_convert (sizetype, t);
4976 t = fold_build1 (NEGATE_EXPR, sizetype, t);
4978 /* [4] Emit code for: addr_rtx = top - off. On big endian machines,
4979 the argument has RSIZE - SIZE bytes of leading padding. */
4980 t = build2 (POINTER_PLUS_EXPR, TREE_TYPE (top), top, t);
4981 if (BYTES_BIG_ENDIAN && rsize > size)
4983 u = size_int (rsize - size);
4984 t = build2 (POINTER_PLUS_EXPR, TREE_TYPE (t), t, u);
4986 COND_EXPR_THEN (addr) = t;
4988 if (osize > UNITS_PER_WORD)
4990 /* [9] Emit: ovfl += ((intptr_t) ovfl + osize - 1) & -osize. */
4991 u = size_int (osize - 1);
4992 t = build2 (POINTER_PLUS_EXPR, TREE_TYPE (ovfl), ovfl, u);
4993 t = fold_convert (sizetype, t);
4994 u = size_int (-osize);
4995 t = build2 (BIT_AND_EXPR, sizetype, t, u);
4996 t = fold_convert (TREE_TYPE (ovfl), t);
4997 align = build2 (GIMPLE_MODIFY_STMT, TREE_TYPE (ovfl), ovfl, t);
4999 else
5000 align = NULL;
5002 /* [10, 11]. Emit code to store ovfl in addr_rtx, then
5003 post-increment ovfl by osize. On big-endian machines,
5004 the argument has OSIZE - SIZE bytes of leading padding. */
5005 u = fold_convert (TREE_TYPE (ovfl),
5006 build_int_cst (NULL_TREE, osize));
5007 t = build2 (POSTINCREMENT_EXPR, TREE_TYPE (ovfl), ovfl, u);
5008 if (BYTES_BIG_ENDIAN && osize > size)
5010 u = size_int (osize - size);
5011 t = build2 (POINTER_PLUS_EXPR, TREE_TYPE (t), t, u);
5014 /* String [9] and [10,11] together. */
5015 if (align)
5016 t = build2 (COMPOUND_EXPR, TREE_TYPE (t), align, t);
5017 COND_EXPR_ELSE (addr) = t;
5019 addr = fold_convert (build_pointer_type (type), addr);
5020 addr = build_va_arg_indirect_ref (addr);
5023 if (indirect)
5024 addr = build_va_arg_indirect_ref (addr);
5026 return addr;
5029 /* We keep a list of functions for which we have already built stubs
5030 in build_mips16_call_stub. */
5032 struct mips16_stub
5034 struct mips16_stub *next;
5035 char *name;
5036 int fpret;
5039 static struct mips16_stub *mips16_stubs;
5041 /* Return a two-character string representing a function floating-point
5042 return mode, used to name MIPS16 function stubs. */
5044 static const char *
5045 mips16_call_stub_mode_suffix (enum machine_mode mode)
5047 if (mode == SFmode)
5048 return "sf";
5049 else if (mode == DFmode)
5050 return "df";
5051 else if (mode == SCmode)
5052 return "sc";
5053 else if (mode == DCmode)
5054 return "dc";
5055 else if (mode == V2SFmode)
5056 return "df";
5057 else
5058 gcc_unreachable ();
5061 /* Write instructions to move a 32-bit value between general register
5062 GPREG and floating-point register FPREG. DIRECTION is 't' to move
5063 from GPREG to FPREG and 'f' to move in the opposite direction. */
5065 static void
5066 mips_output_32bit_xfer (char direction, unsigned int gpreg, unsigned int fpreg)
5068 fprintf (asm_out_file, "\tm%cc1\t%s,%s\n", direction,
5069 reg_names[gpreg], reg_names[fpreg]);
5072 /* Likewise for 64-bit values. */
5074 static void
5075 mips_output_64bit_xfer (char direction, unsigned int gpreg, unsigned int fpreg)
5077 if (TARGET_64BIT)
5078 fprintf (asm_out_file, "\tdm%cc1\t%s,%s\n", direction,
5079 reg_names[gpreg], reg_names[fpreg]);
5080 else if (TARGET_FLOAT64)
5082 fprintf (asm_out_file, "\tm%cc1\t%s,%s\n", direction,
5083 reg_names[gpreg + TARGET_BIG_ENDIAN], reg_names[fpreg]);
5084 fprintf (asm_out_file, "\tm%chc1\t%s,%s\n", direction,
5085 reg_names[gpreg + TARGET_LITTLE_ENDIAN], reg_names[fpreg]);
5087 else
5089 /* Move the least-significant word. */
5090 fprintf (asm_out_file, "\tm%cc1\t%s,%s\n", direction,
5091 reg_names[gpreg + TARGET_BIG_ENDIAN], reg_names[fpreg]);
5092 /* ...then the most significant word. */
5093 fprintf (asm_out_file, "\tm%cc1\t%s,%s\n", direction,
5094 reg_names[gpreg + TARGET_LITTLE_ENDIAN], reg_names[fpreg + 1]);
5098 /* Write out code to move floating-point arguments into or out of
5099 general registers. FP_CODE is the code describing which arguments
5100 are present (see the comment above the definition of CUMULATIVE_ARGS
5101 in mips.h). DIRECTION is as for mips_output_32bit_xfer. */
5103 static void
5104 mips_output_args_xfer (int fp_code, char direction)
5106 unsigned int gparg, fparg, f;
5107 CUMULATIVE_ARGS cum;
5109 /* This code only works for the original 32-bit ABI and the O64 ABI. */
5110 gcc_assert (TARGET_OLDABI);
5112 init_cumulative_args (&cum, NULL, NULL);
5114 for (f = (unsigned int) fp_code; f != 0; f >>= 2)
5116 enum machine_mode mode;
5117 struct mips_arg_info info;
5119 if ((f & 3) == 1)
5120 mode = SFmode;
5121 else if ((f & 3) == 2)
5122 mode = DFmode;
5123 else
5124 gcc_unreachable ();
5126 mips_arg_info (&cum, mode, NULL, true, &info);
5127 gparg = mips_arg_regno (&info, false);
5128 fparg = mips_arg_regno (&info, true);
5130 if (mode == SFmode)
5131 mips_output_32bit_xfer (direction, gparg, fparg);
5132 else
5133 mips_output_64bit_xfer (direction, gparg, fparg);
5135 function_arg_advance (&cum, mode, NULL, true);
5139 /* Build a mips16 function stub. This is used for functions which
5140 take arguments in the floating point registers. It is 32-bit code
5141 that moves the floating point args into the general registers, and
5142 then jumps to the 16-bit code. */
5144 static void
5145 build_mips16_function_stub (void)
5147 const char *fnname;
5148 char *secname, *stubname;
5149 tree stubid, stubdecl;
5150 int need_comma;
5151 unsigned int f;
5153 fnname = XSTR (XEXP (DECL_RTL (current_function_decl), 0), 0);
5154 fnname = targetm.strip_name_encoding (fnname);
5155 secname = (char *) alloca (strlen (fnname) + 20);
5156 sprintf (secname, ".mips16.fn.%s", fnname);
5157 stubname = (char *) alloca (strlen (fnname) + 20);
5158 sprintf (stubname, "__fn_stub_%s", fnname);
5159 stubid = get_identifier (stubname);
5160 stubdecl = build_decl (FUNCTION_DECL, stubid,
5161 build_function_type (void_type_node, NULL_TREE));
5162 DECL_SECTION_NAME (stubdecl) = build_string (strlen (secname), secname);
5163 DECL_RESULT (stubdecl) = build_decl (RESULT_DECL, NULL_TREE, void_type_node);
5165 fprintf (asm_out_file, "\t# Stub function for %s (",
5166 current_function_name ());
5167 need_comma = 0;
5168 for (f = (unsigned int) current_function_args_info.fp_code; f != 0; f >>= 2)
5170 fprintf (asm_out_file, "%s%s",
5171 need_comma ? ", " : "",
5172 (f & 3) == 1 ? "float" : "double");
5173 need_comma = 1;
5175 fprintf (asm_out_file, ")\n");
5177 fprintf (asm_out_file, "\t.set\tnomips16\n");
5178 switch_to_section (function_section (stubdecl));
5179 ASM_OUTPUT_ALIGN (asm_out_file,
5180 floor_log2 (FUNCTION_BOUNDARY / BITS_PER_UNIT));
5182 /* ??? If FUNCTION_NAME_ALREADY_DECLARED is defined, then we are
5183 within a .ent, and we cannot emit another .ent. */
5184 if (!FUNCTION_NAME_ALREADY_DECLARED)
5186 fputs ("\t.ent\t", asm_out_file);
5187 assemble_name (asm_out_file, stubname);
5188 fputs ("\n", asm_out_file);
5191 assemble_name (asm_out_file, stubname);
5192 fputs (":\n", asm_out_file);
5194 /* Load the address of the MIPS16 function into $at. Do this first so
5195 that targets with coprocessor interlocks can use an MFC1 to fill the
5196 delay slot. */
5197 fprintf (asm_out_file, "\t.set\tnoat\n");
5198 fprintf (asm_out_file, "\tla\t%s,", reg_names[GP_REG_FIRST + 1]);
5199 assemble_name (asm_out_file, fnname);
5200 fprintf (asm_out_file, "\n");
5202 mips_output_args_xfer (current_function_args_info.fp_code, 'f');
5204 fprintf (asm_out_file, "\tjr\t%s\n", reg_names[GP_REG_FIRST + 1]);
5205 fprintf (asm_out_file, "\t.set\tat\n");
5207 if (!FUNCTION_NAME_ALREADY_DECLARED)
5209 fputs ("\t.end\t", asm_out_file);
5210 assemble_name (asm_out_file, stubname);
5211 fputs ("\n", asm_out_file);
5214 switch_to_section (function_section (current_function_decl));
5217 /* The current function is a MIPS16 function that returns a value in an FPR.
5218 Copy the return value from its soft-float to its hard-float location.
5219 libgcc2 has special non-MIPS16 helper functions for each case. */
5221 static void
5222 mips16_copy_fpr_return_value (void)
5224 rtx fn, insn, arg, call;
5225 tree id, return_type;
5226 enum machine_mode return_mode;
5228 return_type = DECL_RESULT (current_function_decl);
5229 return_mode = DECL_MODE (return_type);
5231 id = get_identifier (ACONCAT (("__mips16_ret_",
5232 mips16_call_stub_mode_suffix (return_mode),
5233 NULL)));
5234 fn = gen_rtx_SYMBOL_REF (Pmode, IDENTIFIER_POINTER (id));
5235 arg = gen_rtx_REG (return_mode, GP_RETURN);
5236 call = gen_call_value_internal (arg, fn, const0_rtx);
5237 insn = emit_call_insn (call);
5238 use_reg (&CALL_INSN_FUNCTION_USAGE (insn), arg);
5241 /* Build a call stub for a mips16 call. A stub is needed if we are
5242 passing any floating point values which should go into the floating
5243 point registers. If we are, and the call turns out to be to a
5244 32-bit function, the stub will be used to move the values into the
5245 floating point registers before calling the 32-bit function. The
5246 linker will magically adjust the function call to either the 16-bit
5247 function or the 32-bit stub, depending upon where the function call
5248 is actually defined.
5250 Similarly, we need a stub if the return value might come back in a
5251 floating point register.
5253 RETVAL is the location of the return value, or null if this is
5254 a call rather than a call_value. FN is the address of the
5255 function and ARG_SIZE is the size of the arguments. FP_CODE
5256 is the code built by function_arg. This function returns a nonzero
5257 value if it builds the call instruction itself. */
5260 build_mips16_call_stub (rtx retval, rtx fn, rtx arg_size, int fp_code)
5262 int fpret = 0;
5263 const char *fnname;
5264 char *secname, *stubname;
5265 struct mips16_stub *l;
5266 tree stubid, stubdecl;
5267 int need_comma;
5268 unsigned int f;
5269 rtx insn;
5271 /* We don't need to do anything if we aren't in mips16 mode, or if
5272 we were invoked with the -msoft-float option. */
5273 if (!TARGET_MIPS16 || TARGET_SOFT_FLOAT_ABI)
5274 return 0;
5276 /* Figure out whether the value might come back in a floating point
5277 register. */
5278 if (retval)
5279 fpret = mips_return_mode_in_fpr_p (GET_MODE (retval));
5281 /* We don't need to do anything if there were no floating point
5282 arguments and the value will not be returned in a floating point
5283 register. */
5284 if (fp_code == 0 && ! fpret)
5285 return 0;
5287 /* We don't need to do anything if this is a call to a special
5288 mips16 support function. */
5289 if (GET_CODE (fn) == SYMBOL_REF
5290 && strncmp (XSTR (fn, 0), "__mips16_", 9) == 0)
5291 return 0;
5293 /* This code will only work for o32 and o64 abis. The other ABI's
5294 require more sophisticated support. */
5295 gcc_assert (TARGET_OLDABI);
5297 /* If we're calling via a function pointer, then we must always call
5298 via a stub. There are magic stubs provided in libgcc.a for each
5299 of the required cases. Each of them expects the function address
5300 to arrive in register $2. */
5302 if (GET_CODE (fn) != SYMBOL_REF)
5304 char buf[30];
5305 tree id;
5306 rtx stub_fn, insn;
5308 /* ??? If this code is modified to support other ABI's, we need
5309 to handle PARALLEL return values here. */
5311 if (fpret)
5312 sprintf (buf, "__mips16_call_stub_%s_%d",
5313 mips16_call_stub_mode_suffix (GET_MODE (retval)),
5314 fp_code);
5315 else
5316 sprintf (buf, "__mips16_call_stub_%d",
5317 fp_code);
5319 id = get_identifier (buf);
5320 stub_fn = gen_rtx_SYMBOL_REF (Pmode, IDENTIFIER_POINTER (id));
5322 mips_emit_move (gen_rtx_REG (Pmode, 2), fn);
5324 if (retval == NULL_RTX)
5325 insn = gen_call_internal (stub_fn, arg_size);
5326 else
5327 insn = gen_call_value_internal (retval, stub_fn, arg_size);
5328 insn = emit_call_insn (insn);
5330 /* Put the register usage information on the CALL. */
5331 CALL_INSN_FUNCTION_USAGE (insn) =
5332 gen_rtx_EXPR_LIST (VOIDmode,
5333 gen_rtx_USE (VOIDmode, gen_rtx_REG (Pmode, 2)),
5334 CALL_INSN_FUNCTION_USAGE (insn));
5336 /* If we are handling a floating point return value, we need to
5337 save $18 in the function prologue. Putting a note on the
5338 call will mean that df_regs_ever_live_p ($18) will be true if the
5339 call is not eliminated, and we can check that in the prologue
5340 code. */
5341 if (fpret)
5342 CALL_INSN_FUNCTION_USAGE (insn) =
5343 gen_rtx_EXPR_LIST (VOIDmode,
5344 gen_rtx_USE (VOIDmode,
5345 gen_rtx_REG (word_mode, 18)),
5346 CALL_INSN_FUNCTION_USAGE (insn));
5348 /* Return 1 to tell the caller that we've generated the call
5349 insn. */
5350 return 1;
5353 /* We know the function we are going to call. If we have already
5354 built a stub, we don't need to do anything further. */
5356 fnname = targetm.strip_name_encoding (XSTR (fn, 0));
5357 for (l = mips16_stubs; l != NULL; l = l->next)
5358 if (strcmp (l->name, fnname) == 0)
5359 break;
5361 if (l == NULL)
5363 /* Build a special purpose stub. When the linker sees a
5364 function call in mips16 code, it will check where the target
5365 is defined. If the target is a 32-bit call, the linker will
5366 search for the section defined here. It can tell which
5367 symbol this section is associated with by looking at the
5368 relocation information (the name is unreliable, since this
5369 might be a static function). If such a section is found, the
5370 linker will redirect the call to the start of the magic
5371 section.
5373 If the function does not return a floating point value, the
5374 special stub section is named
5375 .mips16.call.FNNAME
5377 If the function does return a floating point value, the stub
5378 section is named
5379 .mips16.call.fp.FNNAME
5382 secname = (char *) alloca (strlen (fnname) + 40);
5383 sprintf (secname, ".mips16.call.%s%s",
5384 fpret ? "fp." : "",
5385 fnname);
5386 stubname = (char *) alloca (strlen (fnname) + 20);
5387 sprintf (stubname, "__call_stub_%s%s",
5388 fpret ? "fp_" : "",
5389 fnname);
5390 stubid = get_identifier (stubname);
5391 stubdecl = build_decl (FUNCTION_DECL, stubid,
5392 build_function_type (void_type_node, NULL_TREE));
5393 DECL_SECTION_NAME (stubdecl) = build_string (strlen (secname), secname);
5394 DECL_RESULT (stubdecl) = build_decl (RESULT_DECL, NULL_TREE, void_type_node);
5396 fprintf (asm_out_file, "\t# Stub function to call %s%s (",
5397 (fpret
5398 ? (GET_MODE (retval) == SFmode ? "float " : "double ")
5399 : ""),
5400 fnname);
5401 need_comma = 0;
5402 for (f = (unsigned int) fp_code; f != 0; f >>= 2)
5404 fprintf (asm_out_file, "%s%s",
5405 need_comma ? ", " : "",
5406 (f & 3) == 1 ? "float" : "double");
5407 need_comma = 1;
5409 fprintf (asm_out_file, ")\n");
5411 fprintf (asm_out_file, "\t.set\tnomips16\n");
5412 assemble_start_function (stubdecl, stubname);
5414 if (!FUNCTION_NAME_ALREADY_DECLARED)
5416 fputs ("\t.ent\t", asm_out_file);
5417 assemble_name (asm_out_file, stubname);
5418 fputs ("\n", asm_out_file);
5420 assemble_name (asm_out_file, stubname);
5421 fputs (":\n", asm_out_file);
5424 /* We build the stub code by hand. That's the only way we can
5425 do it, since we can't generate 32-bit code during a 16-bit
5426 compilation. */
5428 if (! fpret)
5430 /* Load the address of the MIPS16 function into $at. Do this
5431 first so that targets with coprocessor interlocks can use
5432 an MFC1 to fill the delay slot. */
5433 fprintf (asm_out_file, "\t.set\tnoat\n");
5434 fprintf (asm_out_file, "\tla\t%s,%s\n", reg_names[GP_REG_FIRST + 1],
5435 fnname);
5438 mips_output_args_xfer (fp_code, 't');
5440 if (! fpret)
5442 /* Jump to the previously-loaded address. */
5443 fprintf (asm_out_file, "\tjr\t%s\n", reg_names[GP_REG_FIRST + 1]);
5444 fprintf (asm_out_file, "\t.set\tat\n");
5446 else
5448 fprintf (asm_out_file, "\tmove\t%s,%s\n",
5449 reg_names[GP_REG_FIRST + 18], reg_names[GP_REG_FIRST + 31]);
5450 fprintf (asm_out_file, "\tjal\t%s\n", fnname);
5451 switch (GET_MODE (retval))
5453 case SCmode:
5454 mips_output_32bit_xfer ('f', GP_RETURN + 1,
5455 FP_REG_FIRST + MAX_FPRS_PER_FMT);
5456 /* Fall though. */
5457 case SFmode:
5458 mips_output_32bit_xfer ('f', GP_RETURN, FP_REG_FIRST);
5459 if (GET_MODE (retval) == SCmode && TARGET_64BIT)
5461 /* On 64-bit targets, complex floats are returned in
5462 a single GPR, such that "sd" on a suitably-aligned
5463 target would store the value correctly. */
5464 fprintf (asm_out_file, "\tdsll\t%s,%s,32\n",
5465 reg_names[GP_RETURN + TARGET_LITTLE_ENDIAN],
5466 reg_names[GP_RETURN + TARGET_LITTLE_ENDIAN]);
5467 fprintf (asm_out_file, "\tor\t%s,%s,%s\n",
5468 reg_names[GP_RETURN],
5469 reg_names[GP_RETURN],
5470 reg_names[GP_RETURN + 1]);
5472 break;
5474 case DCmode:
5475 mips_output_64bit_xfer ('f', GP_RETURN + (8 / UNITS_PER_WORD),
5476 FP_REG_FIRST + MAX_FPRS_PER_FMT);
5477 /* Fall though. */
5478 case DFmode:
5479 case V2SFmode:
5480 mips_output_64bit_xfer ('f', GP_RETURN, FP_REG_FIRST);
5481 break;
5483 default:
5484 gcc_unreachable ();
5486 fprintf (asm_out_file, "\tj\t%s\n", reg_names[GP_REG_FIRST + 18]);
5489 #ifdef ASM_DECLARE_FUNCTION_SIZE
5490 ASM_DECLARE_FUNCTION_SIZE (asm_out_file, stubname, stubdecl);
5491 #endif
5493 if (!FUNCTION_NAME_ALREADY_DECLARED)
5495 fputs ("\t.end\t", asm_out_file);
5496 assemble_name (asm_out_file, stubname);
5497 fputs ("\n", asm_out_file);
5500 /* Record this stub. */
5501 l = (struct mips16_stub *) xmalloc (sizeof *l);
5502 l->name = xstrdup (fnname);
5503 l->fpret = fpret;
5504 l->next = mips16_stubs;
5505 mips16_stubs = l;
5508 /* If we expect a floating point return value, but we've built a
5509 stub which does not expect one, then we're in trouble. We can't
5510 use the existing stub, because it won't handle the floating point
5511 value. We can't build a new stub, because the linker won't know
5512 which stub to use for the various calls in this object file.
5513 Fortunately, this case is illegal, since it means that a function
5514 was declared in two different ways in a single compilation. */
5515 if (fpret && ! l->fpret)
5516 error ("cannot handle inconsistent calls to %qs", fnname);
5518 if (retval == NULL_RTX)
5519 insn = gen_call_internal_direct (fn, arg_size);
5520 else
5521 insn = gen_call_value_internal_direct (retval, fn, arg_size);
5522 insn = emit_call_insn (insn);
5524 /* If we are calling a stub which handles a floating point return
5525 value, we need to arrange to save $18 in the prologue. We do
5526 this by marking the function call as using the register. The
5527 prologue will later see that it is used, and emit code to save
5528 it. */
5529 if (l->fpret)
5530 CALL_INSN_FUNCTION_USAGE (insn) =
5531 gen_rtx_EXPR_LIST (VOIDmode,
5532 gen_rtx_USE (VOIDmode, gen_rtx_REG (word_mode, 18)),
5533 CALL_INSN_FUNCTION_USAGE (insn));
5535 /* Return 1 to tell the caller that we've generated the call
5536 insn. */
5537 return 1;
5540 /* Return true if calls to X can use R_MIPS_CALL* relocations. */
5542 static bool
5543 mips_ok_for_lazy_binding_p (rtx x)
5545 return (TARGET_USE_GOT
5546 && GET_CODE (x) == SYMBOL_REF
5547 && !mips_symbol_binds_local_p (x));
5550 /* Load function address ADDR into register DEST. SIBCALL_P is true
5551 if the address is needed for a sibling call. Return true if we
5552 used an explicit lazy-binding sequence. */
5554 static bool
5555 mips_load_call_address (rtx dest, rtx addr, int sibcall_p)
5557 /* If we're generating PIC, and this call is to a global function,
5558 try to allow its address to be resolved lazily. This isn't
5559 possible if TARGET_CALL_SAVED_GP since the value of $gp on entry
5560 to the stub would be our caller's gp, not ours. */
5561 if (TARGET_EXPLICIT_RELOCS
5562 && !(sibcall_p && TARGET_CALL_SAVED_GP)
5563 && mips_ok_for_lazy_binding_p (addr))
5565 rtx high, lo_sum_symbol;
5567 high = mips_unspec_offset_high (dest, pic_offset_table_rtx,
5568 addr, SYMBOL_GOTOFF_CALL);
5569 lo_sum_symbol = mips_unspec_address (addr, SYMBOL_GOTOFF_CALL);
5570 if (Pmode == SImode)
5571 emit_insn (gen_load_callsi (dest, high, lo_sum_symbol));
5572 else
5573 emit_insn (gen_load_calldi (dest, high, lo_sum_symbol));
5574 return true;
5576 else
5578 mips_emit_move (dest, addr);
5579 return false;
5584 /* Expand a call or call_value instruction. RESULT is where the
5585 result will go (null for calls), ADDR is the address of the
5586 function, ARGS_SIZE is the size of the arguments and AUX is
5587 the value passed to us by mips_function_arg. SIBCALL_P is true
5588 if we are expanding a sibling call, false if we're expanding
5589 a normal call. */
5591 void
5592 mips_expand_call (rtx result, rtx addr, rtx args_size, rtx aux, int sibcall_p)
5594 rtx orig_addr, pattern, insn;
5595 bool lazy_p;
5597 orig_addr = addr;
5598 lazy_p = false;
5599 if (!call_insn_operand (addr, VOIDmode))
5601 addr = gen_reg_rtx (Pmode);
5602 lazy_p = mips_load_call_address (addr, orig_addr, sibcall_p);
5605 if (TARGET_MIPS16
5606 && TARGET_HARD_FLOAT_ABI
5607 && build_mips16_call_stub (result, addr, args_size,
5608 aux == 0 ? 0 : (int) GET_MODE (aux)))
5609 return;
5611 if (result == 0)
5612 pattern = (sibcall_p
5613 ? gen_sibcall_internal (addr, args_size)
5614 : gen_call_internal (addr, args_size));
5615 else if (GET_CODE (result) == PARALLEL && XVECLEN (result, 0) == 2)
5617 rtx reg1, reg2;
5619 reg1 = XEXP (XVECEXP (result, 0, 0), 0);
5620 reg2 = XEXP (XVECEXP (result, 0, 1), 0);
5621 pattern =
5622 (sibcall_p
5623 ? gen_sibcall_value_multiple_internal (reg1, addr, args_size, reg2)
5624 : gen_call_value_multiple_internal (reg1, addr, args_size, reg2));
5626 else
5627 pattern = (sibcall_p
5628 ? gen_sibcall_value_internal (result, addr, args_size)
5629 : gen_call_value_internal (result, addr, args_size));
5631 insn = emit_call_insn (pattern);
5633 /* Lazy-binding stubs require $gp to be valid on entry. We also pretend
5634 that they use FAKE_CALL_REGNO; see the load_call<mode> patterns for
5635 details. */
5636 if (lazy_p)
5638 use_reg (&CALL_INSN_FUNCTION_USAGE (insn), pic_offset_table_rtx);
5639 use_reg (&CALL_INSN_FUNCTION_USAGE (insn),
5640 gen_rtx_REG (Pmode, FAKE_CALL_REGNO));
5645 /* Implement TARGET_FUNCTION_OK_FOR_SIBCALL. */
5647 static bool
5648 mips_function_ok_for_sibcall (tree decl, tree exp ATTRIBUTE_UNUSED)
5650 if (!TARGET_SIBCALLS)
5651 return false;
5653 /* We can't do a sibcall if the called function is a MIPS16 function
5654 because there is no direct "jx" instruction equivalent to "jalx" to
5655 switch the ISA mode. */
5656 if (mips_use_mips16_mode_p (decl))
5657 return false;
5659 /* ...and when -minterlink-mips16 is in effect, assume that external
5660 functions could be MIPS16 ones unless an attribute explicitly
5661 tells us otherwise. We only care about cases where the sibling
5662 and normal calls would both be direct. */
5663 if (TARGET_INTERLINK_MIPS16
5664 && decl
5665 && DECL_EXTERNAL (decl)
5666 && !mips_nomips16_decl_p (decl)
5667 && const_call_insn_operand (XEXP (DECL_RTL (decl), 0), VOIDmode))
5668 return false;
5670 /* Otherwise OK. */
5671 return true;
5674 /* Emit code to move general operand SRC into condition-code
5675 register DEST. SCRATCH is a scratch TFmode float register.
5676 The sequence is:
5678 FP1 = SRC
5679 FP2 = 0.0f
5680 DEST = FP2 < FP1
5682 where FP1 and FP2 are single-precision float registers
5683 taken from SCRATCH. */
5685 void
5686 mips_emit_fcc_reload (rtx dest, rtx src, rtx scratch)
5688 rtx fp1, fp2;
5690 /* Change the source to SFmode. */
5691 if (MEM_P (src))
5692 src = adjust_address (src, SFmode, 0);
5693 else if (REG_P (src) || GET_CODE (src) == SUBREG)
5694 src = gen_rtx_REG (SFmode, true_regnum (src));
5696 fp1 = gen_rtx_REG (SFmode, REGNO (scratch));
5697 fp2 = gen_rtx_REG (SFmode, REGNO (scratch) + MAX_FPRS_PER_FMT);
5699 mips_emit_move (copy_rtx (fp1), src);
5700 mips_emit_move (copy_rtx (fp2), CONST0_RTX (SFmode));
5701 emit_insn (gen_slt_sf (dest, fp2, fp1));
5704 /* Emit straight-line code to move LENGTH bytes from SRC to DEST.
5705 Assume that the areas do not overlap. */
5707 static void
5708 mips_block_move_straight (rtx dest, rtx src, HOST_WIDE_INT length)
5710 HOST_WIDE_INT offset, delta;
5711 unsigned HOST_WIDE_INT bits;
5712 int i;
5713 enum machine_mode mode;
5714 rtx *regs;
5716 /* Work out how many bits to move at a time. If both operands have
5717 half-word alignment, it is usually better to move in half words.
5718 For instance, lh/lh/sh/sh is usually better than lwl/lwr/swl/swr
5719 and lw/lw/sw/sw is usually better than ldl/ldr/sdl/sdr.
5720 Otherwise move word-sized chunks. */
5721 if (MEM_ALIGN (src) == BITS_PER_WORD / 2
5722 && MEM_ALIGN (dest) == BITS_PER_WORD / 2)
5723 bits = BITS_PER_WORD / 2;
5724 else
5725 bits = BITS_PER_WORD;
5727 mode = mode_for_size (bits, MODE_INT, 0);
5728 delta = bits / BITS_PER_UNIT;
5730 /* Allocate a buffer for the temporary registers. */
5731 regs = alloca (sizeof (rtx) * length / delta);
5733 /* Load as many BITS-sized chunks as possible. Use a normal load if
5734 the source has enough alignment, otherwise use left/right pairs. */
5735 for (offset = 0, i = 0; offset + delta <= length; offset += delta, i++)
5737 regs[i] = gen_reg_rtx (mode);
5738 if (MEM_ALIGN (src) >= bits)
5739 mips_emit_move (regs[i], adjust_address (src, mode, offset));
5740 else
5742 rtx part = adjust_address (src, BLKmode, offset);
5743 if (!mips_expand_unaligned_load (regs[i], part, bits, 0))
5744 gcc_unreachable ();
5748 /* Copy the chunks to the destination. */
5749 for (offset = 0, i = 0; offset + delta <= length; offset += delta, i++)
5750 if (MEM_ALIGN (dest) >= bits)
5751 mips_emit_move (adjust_address (dest, mode, offset), regs[i]);
5752 else
5754 rtx part = adjust_address (dest, BLKmode, offset);
5755 if (!mips_expand_unaligned_store (part, regs[i], bits, 0))
5756 gcc_unreachable ();
5759 /* Mop up any left-over bytes. */
5760 if (offset < length)
5762 src = adjust_address (src, BLKmode, offset);
5763 dest = adjust_address (dest, BLKmode, offset);
5764 move_by_pieces (dest, src, length - offset,
5765 MIN (MEM_ALIGN (src), MEM_ALIGN (dest)), 0);
5769 #define MAX_MOVE_REGS 4
5770 #define MAX_MOVE_BYTES (MAX_MOVE_REGS * UNITS_PER_WORD)
5773 /* Helper function for doing a loop-based block operation on memory
5774 reference MEM. Each iteration of the loop will operate on LENGTH
5775 bytes of MEM.
5777 Create a new base register for use within the loop and point it to
5778 the start of MEM. Create a new memory reference that uses this
5779 register. Store them in *LOOP_REG and *LOOP_MEM respectively. */
5781 static void
5782 mips_adjust_block_mem (rtx mem, HOST_WIDE_INT length,
5783 rtx *loop_reg, rtx *loop_mem)
5785 *loop_reg = copy_addr_to_reg (XEXP (mem, 0));
5787 /* Although the new mem does not refer to a known location,
5788 it does keep up to LENGTH bytes of alignment. */
5789 *loop_mem = change_address (mem, BLKmode, *loop_reg);
5790 set_mem_align (*loop_mem, MIN (MEM_ALIGN (mem), length * BITS_PER_UNIT));
5794 /* Move LENGTH bytes from SRC to DEST using a loop that moves MAX_MOVE_BYTES
5795 per iteration. LENGTH must be at least MAX_MOVE_BYTES. Assume that the
5796 memory regions do not overlap. */
5798 static void
5799 mips_block_move_loop (rtx dest, rtx src, HOST_WIDE_INT length)
5801 rtx label, src_reg, dest_reg, final_src;
5802 HOST_WIDE_INT leftover;
5804 leftover = length % MAX_MOVE_BYTES;
5805 length -= leftover;
5807 /* Create registers and memory references for use within the loop. */
5808 mips_adjust_block_mem (src, MAX_MOVE_BYTES, &src_reg, &src);
5809 mips_adjust_block_mem (dest, MAX_MOVE_BYTES, &dest_reg, &dest);
5811 /* Calculate the value that SRC_REG should have after the last iteration
5812 of the loop. */
5813 final_src = expand_simple_binop (Pmode, PLUS, src_reg, GEN_INT (length),
5814 0, 0, OPTAB_WIDEN);
5816 /* Emit the start of the loop. */
5817 label = gen_label_rtx ();
5818 emit_label (label);
5820 /* Emit the loop body. */
5821 mips_block_move_straight (dest, src, MAX_MOVE_BYTES);
5823 /* Move on to the next block. */
5824 mips_emit_move (src_reg, plus_constant (src_reg, MAX_MOVE_BYTES));
5825 mips_emit_move (dest_reg, plus_constant (dest_reg, MAX_MOVE_BYTES));
5827 /* Emit the loop condition. */
5828 if (Pmode == DImode)
5829 emit_insn (gen_cmpdi (src_reg, final_src));
5830 else
5831 emit_insn (gen_cmpsi (src_reg, final_src));
5832 emit_jump_insn (gen_bne (label));
5834 /* Mop up any left-over bytes. */
5835 if (leftover)
5836 mips_block_move_straight (dest, src, leftover);
5839 /* Expand a movmemsi instruction. */
5841 bool
5842 mips_expand_block_move (rtx dest, rtx src, rtx length)
5844 if (GET_CODE (length) == CONST_INT)
5846 if (INTVAL (length) <= 2 * MAX_MOVE_BYTES)
5848 mips_block_move_straight (dest, src, INTVAL (length));
5849 return true;
5851 else if (optimize)
5853 mips_block_move_loop (dest, src, INTVAL (length));
5854 return true;
5857 return false;
5861 /* Expand a loop of synci insns for the address range [BEGIN, END). */
5863 void
5864 mips_expand_synci_loop (rtx begin, rtx end)
5866 rtx inc, label, cmp, cmp_result;
5868 /* Load INC with the cache line size (rdhwr INC,$1). */
5869 inc = gen_reg_rtx (SImode);
5870 emit_insn (gen_rdhwr (inc, const1_rtx));
5872 /* Loop back to here. */
5873 label = gen_label_rtx ();
5874 emit_label (label);
5876 emit_insn (gen_synci (begin));
5878 cmp = gen_reg_rtx (Pmode);
5879 mips_emit_binary (GTU, cmp, begin, end);
5881 mips_emit_binary (PLUS, begin, begin, inc);
5883 cmp_result = gen_rtx_EQ (VOIDmode, cmp, const0_rtx);
5884 emit_jump_insn (gen_condjump (cmp_result, label));
5887 /* Return true if it is possible to use left/right accesses for a
5888 bitfield of WIDTH bits starting BITPOS bits into *OP. When
5889 returning true, update *OP, *LEFT and *RIGHT as follows:
5891 *OP is a BLKmode reference to the whole field.
5893 *LEFT is a QImode reference to the first byte if big endian or
5894 the last byte if little endian. This address can be used in the
5895 left-side instructions (lwl, swl, ldl, sdl).
5897 *RIGHT is a QImode reference to the opposite end of the field and
5898 can be used in the patterning right-side instruction. */
5900 static bool
5901 mips_get_unaligned_mem (rtx *op, unsigned int width, int bitpos,
5902 rtx *left, rtx *right)
5904 rtx first, last;
5906 /* Check that the operand really is a MEM. Not all the extv and
5907 extzv predicates are checked. */
5908 if (!MEM_P (*op))
5909 return false;
5911 /* Check that the size is valid. */
5912 if (width != 32 && (!TARGET_64BIT || width != 64))
5913 return false;
5915 /* We can only access byte-aligned values. Since we are always passed
5916 a reference to the first byte of the field, it is not necessary to
5917 do anything with BITPOS after this check. */
5918 if (bitpos % BITS_PER_UNIT != 0)
5919 return false;
5921 /* Reject aligned bitfields: we want to use a normal load or store
5922 instead of a left/right pair. */
5923 if (MEM_ALIGN (*op) >= width)
5924 return false;
5926 /* Adjust *OP to refer to the whole field. This also has the effect
5927 of legitimizing *OP's address for BLKmode, possibly simplifying it. */
5928 *op = adjust_address (*op, BLKmode, 0);
5929 set_mem_size (*op, GEN_INT (width / BITS_PER_UNIT));
5931 /* Get references to both ends of the field. We deliberately don't
5932 use the original QImode *OP for FIRST since the new BLKmode one
5933 might have a simpler address. */
5934 first = adjust_address (*op, QImode, 0);
5935 last = adjust_address (*op, QImode, width / BITS_PER_UNIT - 1);
5937 /* Allocate to LEFT and RIGHT according to endianness. LEFT should
5938 be the upper word and RIGHT the lower word. */
5939 if (TARGET_BIG_ENDIAN)
5940 *left = first, *right = last;
5941 else
5942 *left = last, *right = first;
5944 return true;
5948 /* Try to emit the equivalent of (set DEST (zero_extract SRC WIDTH BITPOS)).
5949 Return true on success. We only handle cases where zero_extract is
5950 equivalent to sign_extract. */
5952 bool
5953 mips_expand_unaligned_load (rtx dest, rtx src, unsigned int width, int bitpos)
5955 rtx left, right, temp;
5957 /* If TARGET_64BIT, the destination of a 32-bit load will be a
5958 paradoxical word_mode subreg. This is the only case in which
5959 we allow the destination to be larger than the source. */
5960 if (GET_CODE (dest) == SUBREG
5961 && GET_MODE (dest) == DImode
5962 && SUBREG_BYTE (dest) == 0
5963 && GET_MODE (SUBREG_REG (dest)) == SImode)
5964 dest = SUBREG_REG (dest);
5966 /* After the above adjustment, the destination must be the same
5967 width as the source. */
5968 if (GET_MODE_BITSIZE (GET_MODE (dest)) != width)
5969 return false;
5971 if (!mips_get_unaligned_mem (&src, width, bitpos, &left, &right))
5972 return false;
5974 temp = gen_reg_rtx (GET_MODE (dest));
5975 if (GET_MODE (dest) == DImode)
5977 emit_insn (gen_mov_ldl (temp, src, left));
5978 emit_insn (gen_mov_ldr (dest, copy_rtx (src), right, temp));
5980 else
5982 emit_insn (gen_mov_lwl (temp, src, left));
5983 emit_insn (gen_mov_lwr (dest, copy_rtx (src), right, temp));
5985 return true;
5989 /* Try to expand (set (zero_extract DEST WIDTH BITPOS) SRC). Return
5990 true on success. */
5992 bool
5993 mips_expand_unaligned_store (rtx dest, rtx src, unsigned int width, int bitpos)
5995 rtx left, right;
5996 enum machine_mode mode;
5998 if (!mips_get_unaligned_mem (&dest, width, bitpos, &left, &right))
5999 return false;
6001 mode = mode_for_size (width, MODE_INT, 0);
6002 src = gen_lowpart (mode, src);
6004 if (mode == DImode)
6006 emit_insn (gen_mov_sdl (dest, src, left));
6007 emit_insn (gen_mov_sdr (copy_rtx (dest), copy_rtx (src), right));
6009 else
6011 emit_insn (gen_mov_swl (dest, src, left));
6012 emit_insn (gen_mov_swr (copy_rtx (dest), copy_rtx (src), right));
6014 return true;
6017 /* Return true if X is a MEM with the same size as MODE. */
6019 bool
6020 mips_mem_fits_mode_p (enum machine_mode mode, rtx x)
6022 rtx size;
6024 if (!MEM_P (x))
6025 return false;
6027 size = MEM_SIZE (x);
6028 return size && INTVAL (size) == GET_MODE_SIZE (mode);
6031 /* Return true if (zero_extract OP SIZE POSITION) can be used as the
6032 source of an "ext" instruction or the destination of an "ins"
6033 instruction. OP must be a register operand and the following
6034 conditions must hold:
6036 0 <= POSITION < GET_MODE_BITSIZE (GET_MODE (op))
6037 0 < SIZE <= GET_MODE_BITSIZE (GET_MODE (op))
6038 0 < POSITION + SIZE <= GET_MODE_BITSIZE (GET_MODE (op))
6040 Also reject lengths equal to a word as they are better handled
6041 by the move patterns. */
6043 bool
6044 mips_use_ins_ext_p (rtx op, rtx size, rtx position)
6046 HOST_WIDE_INT len, pos;
6048 if (!ISA_HAS_EXT_INS
6049 || !register_operand (op, VOIDmode)
6050 || GET_MODE_BITSIZE (GET_MODE (op)) > BITS_PER_WORD)
6051 return false;
6053 len = INTVAL (size);
6054 pos = INTVAL (position);
6056 if (len <= 0 || len >= GET_MODE_BITSIZE (GET_MODE (op))
6057 || pos < 0 || pos + len > GET_MODE_BITSIZE (GET_MODE (op)))
6058 return false;
6060 return true;
6063 /* Initialize mips_split_addresses from the associated command-line
6064 settings.
6066 mips_split_addresses is a half-way house between explicit
6067 relocations and the traditional assembler macros. It can
6068 split absolute 32-bit symbolic constants into a high/lo_sum
6069 pair but uses macros for other sorts of access.
6071 Like explicit relocation support for REL targets, it relies
6072 on GNU extensions in the assembler and the linker.
6074 Although this code should work for -O0, it has traditionally
6075 been treated as an optimization. */
6077 static void
6078 mips_init_split_addresses (void)
6080 if (!TARGET_MIPS16 && TARGET_SPLIT_ADDRESSES
6081 && optimize && !flag_pic
6082 && !ABI_HAS_64BIT_SYMBOLS)
6083 mips_split_addresses = 1;
6084 else
6085 mips_split_addresses = 0;
6088 /* (Re-)Initialize information about relocs. */
6090 static void
6091 mips_init_relocs (void)
6093 memset (mips_split_p, '\0', sizeof (mips_split_p));
6094 memset (mips_hi_relocs, '\0', sizeof (mips_hi_relocs));
6095 memset (mips_lo_relocs, '\0', sizeof (mips_lo_relocs));
6097 if (ABI_HAS_64BIT_SYMBOLS)
6099 if (TARGET_EXPLICIT_RELOCS)
6101 mips_split_p[SYMBOL_64_HIGH] = true;
6102 mips_hi_relocs[SYMBOL_64_HIGH] = "%highest(";
6103 mips_lo_relocs[SYMBOL_64_HIGH] = "%higher(";
6105 mips_split_p[SYMBOL_64_MID] = true;
6106 mips_hi_relocs[SYMBOL_64_MID] = "%higher(";
6107 mips_lo_relocs[SYMBOL_64_MID] = "%hi(";
6109 mips_split_p[SYMBOL_64_LOW] = true;
6110 mips_hi_relocs[SYMBOL_64_LOW] = "%hi(";
6111 mips_lo_relocs[SYMBOL_64_LOW] = "%lo(";
6113 mips_split_p[SYMBOL_ABSOLUTE] = true;
6114 mips_lo_relocs[SYMBOL_ABSOLUTE] = "%lo(";
6117 else
6119 if (TARGET_EXPLICIT_RELOCS || mips_split_addresses || TARGET_MIPS16)
6121 mips_split_p[SYMBOL_ABSOLUTE] = true;
6122 mips_hi_relocs[SYMBOL_ABSOLUTE] = "%hi(";
6123 mips_lo_relocs[SYMBOL_ABSOLUTE] = "%lo(";
6125 mips_lo_relocs[SYMBOL_32_HIGH] = "%hi(";
6129 if (TARGET_MIPS16)
6131 /* The high part is provided by a pseudo copy of $gp. */
6132 mips_split_p[SYMBOL_GP_RELATIVE] = true;
6133 mips_lo_relocs[SYMBOL_GP_RELATIVE] = "%gprel(";
6136 if (TARGET_EXPLICIT_RELOCS)
6138 /* Small data constants are kept whole until after reload,
6139 then lowered by mips_rewrite_small_data. */
6140 mips_lo_relocs[SYMBOL_GP_RELATIVE] = "%gp_rel(";
6142 mips_split_p[SYMBOL_GOT_PAGE_OFST] = true;
6143 if (TARGET_NEWABI)
6145 mips_lo_relocs[SYMBOL_GOTOFF_PAGE] = "%got_page(";
6146 mips_lo_relocs[SYMBOL_GOT_PAGE_OFST] = "%got_ofst(";
6148 else
6150 mips_lo_relocs[SYMBOL_GOTOFF_PAGE] = "%got(";
6151 mips_lo_relocs[SYMBOL_GOT_PAGE_OFST] = "%lo(";
6154 if (TARGET_XGOT)
6156 /* The HIGH and LO_SUM are matched by special .md patterns. */
6157 mips_split_p[SYMBOL_GOT_DISP] = true;
6159 mips_split_p[SYMBOL_GOTOFF_DISP] = true;
6160 mips_hi_relocs[SYMBOL_GOTOFF_DISP] = "%got_hi(";
6161 mips_lo_relocs[SYMBOL_GOTOFF_DISP] = "%got_lo(";
6163 mips_split_p[SYMBOL_GOTOFF_CALL] = true;
6164 mips_hi_relocs[SYMBOL_GOTOFF_CALL] = "%call_hi(";
6165 mips_lo_relocs[SYMBOL_GOTOFF_CALL] = "%call_lo(";
6167 else
6169 if (TARGET_NEWABI)
6170 mips_lo_relocs[SYMBOL_GOTOFF_DISP] = "%got_disp(";
6171 else
6172 mips_lo_relocs[SYMBOL_GOTOFF_DISP] = "%got(";
6173 mips_lo_relocs[SYMBOL_GOTOFF_CALL] = "%call16(";
6177 if (TARGET_NEWABI)
6179 mips_split_p[SYMBOL_GOTOFF_LOADGP] = true;
6180 mips_hi_relocs[SYMBOL_GOTOFF_LOADGP] = "%hi(%neg(%gp_rel(";
6181 mips_lo_relocs[SYMBOL_GOTOFF_LOADGP] = "%lo(%neg(%gp_rel(";
6184 /* Thread-local relocation operators. */
6185 mips_lo_relocs[SYMBOL_TLSGD] = "%tlsgd(";
6186 mips_lo_relocs[SYMBOL_TLSLDM] = "%tlsldm(";
6187 mips_split_p[SYMBOL_DTPREL] = 1;
6188 mips_hi_relocs[SYMBOL_DTPREL] = "%dtprel_hi(";
6189 mips_lo_relocs[SYMBOL_DTPREL] = "%dtprel_lo(";
6190 mips_lo_relocs[SYMBOL_GOTTPREL] = "%gottprel(";
6191 mips_split_p[SYMBOL_TPREL] = 1;
6192 mips_hi_relocs[SYMBOL_TPREL] = "%tprel_hi(";
6193 mips_lo_relocs[SYMBOL_TPREL] = "%tprel_lo(";
6195 mips_lo_relocs[SYMBOL_HALF] = "%half(";
6198 /* If OP is an UNSPEC address, return the address to which it refers,
6199 otherwise return OP itself. */
6201 static rtx
6202 mips_strip_unspec_address (rtx op)
6204 rtx base, offset;
6206 split_const (op, &base, &offset);
6207 if (UNSPEC_ADDRESS_P (base))
6208 op = plus_constant (UNSPEC_ADDRESS (base), INTVAL (offset));
6209 return op;
6212 /* Print symbolic operand OP, which is part of a HIGH or LO_SUM
6213 in context CONTEXT. RELOCS is the array of relocations to use. */
6215 static void
6216 print_operand_reloc (FILE *file, rtx op, enum mips_symbol_context context,
6217 const char **relocs)
6219 enum mips_symbol_type symbol_type;
6220 const char *p;
6222 symbol_type = mips_classify_symbolic_expression (op, context);
6223 if (relocs[symbol_type] == 0)
6224 fatal_insn ("PRINT_OPERAND, invalid operand for relocation", op);
6226 fputs (relocs[symbol_type], file);
6227 output_addr_const (file, mips_strip_unspec_address (op));
6228 for (p = relocs[symbol_type]; *p != 0; p++)
6229 if (*p == '(')
6230 fputc (')', file);
6233 /* Print the text for PRINT_OPERAND punctation character CH to FILE.
6234 The punctuation characters are:
6236 '(' Start a nested ".set noreorder" block.
6237 ')' End a nested ".set noreorder" block.
6238 '[' Start a nested ".set noat" block.
6239 ']' End a nested ".set noat" block.
6240 '<' Start a nested ".set nomacro" block.
6241 '>' End a nested ".set nomacro" block.
6242 '*' Behave like %(%< if generating a delayed-branch sequence.
6243 '#' Print a nop if in a ".set noreorder" block.
6244 '/' Like '#', but do nothing within a delayed-branch sequence.
6245 '?' Print "l" if mips_branch_likely is true
6246 '.' Print the name of the register with a hard-wired zero (zero or $0).
6247 '@' Print the name of the assembler temporary register (at or $1).
6248 '^' Print the name of the pic call-through register (t9 or $25).
6249 '+' Print the name of the gp register (usually gp or $28).
6250 '$' Print the name of the stack pointer register (sp or $29).
6251 '|' Print ".set push; .set mips2" if !ISA_HAS_LL_SC.
6252 '-' Print ".set pop" under the same conditions for '|'.
6254 See also mips_init_print_operand_pucnt. */
6256 static void
6257 mips_print_operand_punctuation (FILE *file, int ch)
6259 switch (ch)
6261 case '(':
6262 if (set_noreorder++ == 0)
6263 fputs (".set\tnoreorder\n\t", file);
6264 break;
6266 case ')':
6267 gcc_assert (set_noreorder > 0);
6268 if (--set_noreorder == 0)
6269 fputs ("\n\t.set\treorder", file);
6270 break;
6272 case '[':
6273 if (set_noat++ == 0)
6274 fputs (".set\tnoat\n\t", file);
6275 break;
6277 case ']':
6278 gcc_assert (set_noat > 0);
6279 if (--set_noat == 0)
6280 fputs ("\n\t.set\tat", file);
6281 break;
6283 case '<':
6284 if (set_nomacro++ == 0)
6285 fputs (".set\tnomacro\n\t", file);
6286 break;
6288 case '>':
6289 gcc_assert (set_nomacro > 0);
6290 if (--set_nomacro == 0)
6291 fputs ("\n\t.set\tmacro", file);
6292 break;
6294 case '*':
6295 if (final_sequence != 0)
6297 mips_print_operand_punctuation (file, '(');
6298 mips_print_operand_punctuation (file, '<');
6300 break;
6302 case '#':
6303 if (set_noreorder != 0)
6304 fputs ("\n\tnop", file);
6305 break;
6307 case '/':
6308 /* Print an extra newline so that the delayed insn is separated
6309 from the following ones. This looks neater and is consistent
6310 with non-nop delayed sequences. */
6311 if (set_noreorder != 0 && final_sequence == 0)
6312 fputs ("\n\tnop\n", file);
6313 break;
6315 case '?':
6316 if (mips_branch_likely)
6317 putc ('l', file);
6318 break;
6320 case '.':
6321 fputs (reg_names[GP_REG_FIRST + 0], file);
6322 break;
6324 case '@':
6325 fputs (reg_names[GP_REG_FIRST + 1], file);
6326 break;
6328 case '^':
6329 fputs (reg_names[PIC_FUNCTION_ADDR_REGNUM], file);
6330 break;
6332 case '+':
6333 fputs (reg_names[PIC_OFFSET_TABLE_REGNUM], file);
6334 break;
6336 case '$':
6337 fputs (reg_names[STACK_POINTER_REGNUM], file);
6338 break;
6340 case '|':
6341 if (!ISA_HAS_LL_SC)
6342 fputs (".set\tpush\n\t.set\tmips2\n\t", file);
6343 break;
6345 case '-':
6346 if (!ISA_HAS_LL_SC)
6347 fputs ("\n\t.set\tpop", file);
6348 break;
6350 default:
6351 gcc_unreachable ();
6352 break;
6356 /* Initialize mips_print_operand_punct. */
6358 static void
6359 mips_init_print_operand_punct (void)
6361 const char *p;
6363 for (p = "()[]<>*#/?.@^+$|-"; *p; p++)
6364 mips_print_operand_punct[(unsigned char) *p] = true;
6367 /* PRINT_OPERAND prefix LETTER refers to the integer branch instruction
6368 associated with condition CODE. Print the condition part of the
6369 opcode to FILE. */
6371 static void
6372 mips_print_int_branch_condition (FILE *file, enum rtx_code code, int letter)
6374 switch (code)
6376 case EQ:
6377 case NE:
6378 case GT:
6379 case GE:
6380 case LT:
6381 case LE:
6382 case GTU:
6383 case GEU:
6384 case LTU:
6385 case LEU:
6386 /* Conveniently, the MIPS names for these conditions are the same
6387 as their RTL equivalents. */
6388 fputs (GET_RTX_NAME (code), file);
6389 break;
6391 default:
6392 output_operand_lossage ("'%%%c' is not a valid operand prefix", letter);
6393 break;
6397 /* Likewise floating-point branches. */
6399 static void
6400 mips_print_float_branch_condition (FILE *file, enum rtx_code code, int letter)
6402 switch (code)
6404 case EQ:
6405 fputs ("c1f", file);
6406 break;
6408 case NE:
6409 fputs ("c1t", file);
6410 break;
6412 default:
6413 output_operand_lossage ("'%%%c' is not a valid operand prefix", letter);
6414 break;
6418 /* Implement the PRINT_OPERAND macro. The MIPS-specific operand codes are:
6420 'X' Print CONST_INT OP in hexadecimal format.
6421 'x' Print the low 16 bits of CONST_INT OP in hexadecimal format.
6422 'd' Print CONST_INT OP in decimal.
6423 'h' Print the high-part relocation associated with OP, after stripping
6424 any outermost HIGH.
6425 'R' Print the low-part relocation associated with OP.
6426 'C' Print the integer branch condition for comparison OP.
6427 'N' Print the inverse of the integer branch condition for comparison OP.
6428 'F' Print the FPU branch condition for comparison OP.
6429 'W' Print the inverse of the FPU branch condition for comparison OP.
6430 'T' Print 'f' for (eq:CC ...), 't' for (ne:CC ...),
6431 'z' for (eq:?I ...), 'n' for (ne:?I ...).
6432 't' Like 'T', but with the EQ/NE cases reversed
6433 'Y' Print mips_fp_conditions[INTVAL (OP)]
6434 'Z' Print OP and a comma for ISA_HAS_8CC, otherwise print nothing.
6435 'q' Print a DSP accumulator register.
6436 'D' Print the second part of a double-word register or memory operand.
6437 'L' Print the low-order register in a double-word register operand.
6438 'M' Print high-order register in a double-word register operand.
6439 'z' Print $0 if OP is zero, otherwise print OP normally. */
6441 void
6442 print_operand (FILE *file, rtx op, int letter)
6444 enum rtx_code code;
6446 if (PRINT_OPERAND_PUNCT_VALID_P (letter))
6448 mips_print_operand_punctuation (file, letter);
6449 return;
6452 gcc_assert (op);
6453 code = GET_CODE (op);
6455 switch (letter)
6457 case 'X':
6458 if (GET_CODE (op) == CONST_INT)
6459 fprintf (file, HOST_WIDE_INT_PRINT_HEX, INTVAL (op));
6460 else
6461 output_operand_lossage ("invalid use of '%%%c'", letter);
6462 break;
6464 case 'x':
6465 if (GET_CODE (op) == CONST_INT)
6466 fprintf (file, HOST_WIDE_INT_PRINT_HEX, INTVAL (op) & 0xffff);
6467 else
6468 output_operand_lossage ("invalid use of '%%%c'", letter);
6469 break;
6471 case 'd':
6472 if (GET_CODE (op) == CONST_INT)
6473 fprintf (file, HOST_WIDE_INT_PRINT_DEC, INTVAL (op));
6474 else
6475 output_operand_lossage ("invalid use of '%%%c'", letter);
6476 break;
6478 case 'h':
6479 if (code == HIGH)
6480 op = XEXP (op, 0);
6481 print_operand_reloc (file, op, SYMBOL_CONTEXT_LEA, mips_hi_relocs);
6482 break;
6484 case 'R':
6485 print_operand_reloc (file, op, SYMBOL_CONTEXT_LEA, mips_lo_relocs);
6486 break;
6488 case 'C':
6489 mips_print_int_branch_condition (file, code, letter);
6490 break;
6492 case 'N':
6493 mips_print_int_branch_condition (file, reverse_condition (code), letter);
6494 break;
6496 case 'F':
6497 mips_print_float_branch_condition (file, code, letter);
6498 break;
6500 case 'W':
6501 mips_print_float_branch_condition (file, reverse_condition (code),
6502 letter);
6503 break;
6505 case 'T':
6506 case 't':
6508 int truth = (code == NE) == (letter == 'T');
6509 fputc ("zfnt"[truth * 2 + (GET_MODE (op) == CCmode)], file);
6511 break;
6513 case 'Y':
6514 if (code == CONST_INT && UINTVAL (op) < ARRAY_SIZE (mips_fp_conditions))
6515 fputs (mips_fp_conditions[UINTVAL (op)], file);
6516 else
6517 output_operand_lossage ("'%%%c' is not a valid operand prefix",
6518 letter);
6519 break;
6521 case 'Z':
6522 if (ISA_HAS_8CC)
6524 print_operand (file, op, 0);
6525 fputc (',', file);
6527 break;
6529 case 'q':
6530 if (code == REG && MD_REG_P (REGNO (op)))
6531 fprintf (file, "$ac0");
6532 else if (code == REG && DSP_ACC_REG_P (REGNO (op)))
6533 fprintf (file, "$ac%c", reg_names[REGNO (op)][3]);
6534 else
6535 output_operand_lossage ("invalid use of '%%%c'", letter);
6536 break;
6538 default:
6539 switch (code)
6541 case REG:
6543 unsigned int regno = REGNO (op);
6544 if ((letter == 'M' && TARGET_LITTLE_ENDIAN)
6545 || (letter == 'L' && TARGET_BIG_ENDIAN)
6546 || letter == 'D')
6547 regno++;
6548 fprintf (file, "%s", reg_names[regno]);
6550 break;
6552 case MEM:
6553 if (letter == 'D')
6554 output_address (plus_constant (XEXP (op, 0), 4));
6555 else
6556 output_address (XEXP (op, 0));
6557 break;
6559 default:
6560 if (letter == 'z' && op == CONST0_RTX (GET_MODE (op)))
6561 fputs (reg_names[GP_REG_FIRST], file);
6562 else if (CONST_GP_P (op))
6563 fputs (reg_names[GLOBAL_POINTER_REGNUM], file);
6564 else
6565 output_addr_const (file, mips_strip_unspec_address (op));
6566 break;
6571 /* Output address operand X to FILE. */
6573 void
6574 print_operand_address (FILE *file, rtx x)
6576 struct mips_address_info addr;
6578 if (mips_classify_address (&addr, x, word_mode, true))
6579 switch (addr.type)
6581 case ADDRESS_REG:
6582 print_operand (file, addr.offset, 0);
6583 fprintf (file, "(%s)", reg_names[REGNO (addr.reg)]);
6584 return;
6586 case ADDRESS_LO_SUM:
6587 print_operand_reloc (file, addr.offset, SYMBOL_CONTEXT_MEM,
6588 mips_lo_relocs);
6589 fprintf (file, "(%s)", reg_names[REGNO (addr.reg)]);
6590 return;
6592 case ADDRESS_CONST_INT:
6593 output_addr_const (file, x);
6594 fprintf (file, "(%s)", reg_names[0]);
6595 return;
6597 case ADDRESS_SYMBOLIC:
6598 output_addr_const (file, mips_strip_unspec_address (x));
6599 return;
6601 gcc_unreachable ();
6604 /* Set SYMBOL_REF_FLAGS for the SYMBOL_REF inside RTL, which belongs to DECL.
6605 FIRST is true if this is the first time handling this decl. */
6607 static void
6608 mips_encode_section_info (tree decl, rtx rtl, int first)
6610 default_encode_section_info (decl, rtl, first);
6612 if (TREE_CODE (decl) == FUNCTION_DECL)
6614 rtx symbol = XEXP (rtl, 0);
6615 tree type = TREE_TYPE (decl);
6617 if ((TARGET_LONG_CALLS && !mips_near_type_p (type))
6618 || mips_far_type_p (type))
6619 SYMBOL_REF_FLAGS (symbol) |= SYMBOL_FLAG_LONG_CALL;
6623 /* Implement TARGET_SELECT_RTX_SECTION. */
6625 static section *
6626 mips_select_rtx_section (enum machine_mode mode, rtx x,
6627 unsigned HOST_WIDE_INT align)
6629 /* ??? Consider using mergeable small data sections. */
6630 if (mips_rtx_constant_in_small_data_p (mode))
6631 return get_named_section (NULL, ".sdata", 0);
6633 return default_elf_select_rtx_section (mode, x, align);
6636 /* Implement TARGET_ASM_FUNCTION_RODATA_SECTION.
6638 The complication here is that, with the combination TARGET_ABICALLS
6639 && !TARGET_GPWORD, jump tables will use absolute addresses, and should
6640 therefore not be included in the read-only part of a DSO. Handle such
6641 cases by selecting a normal data section instead of a read-only one.
6642 The logic apes that in default_function_rodata_section. */
6644 static section *
6645 mips_function_rodata_section (tree decl)
6647 if (!TARGET_ABICALLS || TARGET_GPWORD)
6648 return default_function_rodata_section (decl);
6650 if (decl && DECL_SECTION_NAME (decl))
6652 const char *name = TREE_STRING_POINTER (DECL_SECTION_NAME (decl));
6653 if (DECL_ONE_ONLY (decl) && strncmp (name, ".gnu.linkonce.t.", 16) == 0)
6655 char *rname = ASTRDUP (name);
6656 rname[14] = 'd';
6657 return get_section (rname, SECTION_LINKONCE | SECTION_WRITE, decl);
6659 else if (flag_function_sections && flag_data_sections
6660 && strncmp (name, ".text.", 6) == 0)
6662 char *rname = ASTRDUP (name);
6663 memcpy (rname + 1, "data", 4);
6664 return get_section (rname, SECTION_WRITE, decl);
6667 return data_section;
6670 /* Implement TARGET_IN_SMALL_DATA_P. This function controls whether
6671 locally-defined objects go in a small data section. It also controls
6672 the setting of the SYMBOL_REF_SMALL_P flag, which in turn helps
6673 mips_classify_symbol decide when to use %gp_rel(...)($gp) accesses. */
6675 static bool
6676 mips_in_small_data_p (const_tree decl)
6678 HOST_WIDE_INT size;
6680 if (TREE_CODE (decl) == STRING_CST || TREE_CODE (decl) == FUNCTION_DECL)
6681 return false;
6683 /* We don't yet generate small-data references for -mabicalls or
6684 VxWorks RTP code. See the related -G handling in override_options. */
6685 if (TARGET_ABICALLS || TARGET_VXWORKS_RTP)
6686 return false;
6688 if (TREE_CODE (decl) == VAR_DECL && DECL_SECTION_NAME (decl) != 0)
6690 const char *name;
6692 /* Reject anything that isn't in a known small-data section. */
6693 name = TREE_STRING_POINTER (DECL_SECTION_NAME (decl));
6694 if (strcmp (name, ".sdata") != 0 && strcmp (name, ".sbss") != 0)
6695 return false;
6697 /* If a symbol is defined externally, the assembler will use the
6698 usual -G rules when deciding how to implement macros. */
6699 if (mips_lo_relocs[SYMBOL_GP_RELATIVE] || !DECL_EXTERNAL (decl))
6700 return true;
6702 else if (TARGET_EMBEDDED_DATA)
6704 /* Don't put constants into the small data section: we want them
6705 to be in ROM rather than RAM. */
6706 if (TREE_CODE (decl) != VAR_DECL)
6707 return false;
6709 if (TREE_READONLY (decl)
6710 && !TREE_SIDE_EFFECTS (decl)
6711 && (!DECL_INITIAL (decl) || TREE_CONSTANT (DECL_INITIAL (decl))))
6712 return false;
6715 /* Enforce -mlocal-sdata. */
6716 if (!TARGET_LOCAL_SDATA && !TREE_PUBLIC (decl))
6717 return false;
6719 /* Enforce -mextern-sdata. */
6720 if (!TARGET_EXTERN_SDATA && DECL_P (decl))
6722 if (DECL_EXTERNAL (decl))
6723 return false;
6724 if (DECL_COMMON (decl) && DECL_INITIAL (decl) == NULL)
6725 return false;
6728 size = int_size_in_bytes (TREE_TYPE (decl));
6729 return (size > 0 && size <= mips_section_threshold);
6732 /* Implement TARGET_USE_ANCHORS_FOR_SYMBOL_P. We don't want to use
6733 anchors for small data: the GP register acts as an anchor in that
6734 case. We also don't want to use them for PC-relative accesses,
6735 where the PC acts as an anchor. */
6737 static bool
6738 mips_use_anchors_for_symbol_p (const_rtx symbol)
6740 switch (mips_classify_symbol (symbol, SYMBOL_CONTEXT_MEM))
6742 case SYMBOL_PC_RELATIVE:
6743 case SYMBOL_GP_RELATIVE:
6744 return false;
6746 default:
6747 return default_use_anchors_for_symbol_p (symbol);
6751 /* The MIPS debug format wants all automatic variables and arguments
6752 to be in terms of the virtual frame pointer (stack pointer before
6753 any adjustment in the function), while the MIPS 3.0 linker wants
6754 the frame pointer to be the stack pointer after the initial
6755 adjustment. So, we do the adjustment here. The arg pointer (which
6756 is eliminated) points to the virtual frame pointer, while the frame
6757 pointer (which may be eliminated) points to the stack pointer after
6758 the initial adjustments. */
6760 HOST_WIDE_INT
6761 mips_debugger_offset (rtx addr, HOST_WIDE_INT offset)
6763 rtx offset2 = const0_rtx;
6764 rtx reg = eliminate_constant_term (addr, &offset2);
6766 if (offset == 0)
6767 offset = INTVAL (offset2);
6769 if (reg == stack_pointer_rtx || reg == frame_pointer_rtx
6770 || reg == hard_frame_pointer_rtx)
6772 offset -= cfun->machine->frame.total_size;
6773 if (reg == hard_frame_pointer_rtx)
6774 offset += cfun->machine->frame.hard_frame_pointer_offset;
6777 /* sdbout_parms does not want this to crash for unrecognized cases. */
6778 #if 0
6779 else if (reg != arg_pointer_rtx)
6780 fatal_insn ("mips_debugger_offset called with non stack/frame/arg pointer",
6781 addr);
6782 #endif
6784 return offset;
6787 /* When using assembler macros, keep track of all of small-data externs
6788 so that mips_file_end can emit the appropriate declarations for them.
6790 In most cases it would be safe (though pointless) to emit .externs
6791 for other symbols too. One exception is when an object is within
6792 the -G limit but declared by the user to be in a section other
6793 than .sbss or .sdata. */
6795 void
6796 mips_output_external (FILE *file, tree decl, const char *name)
6798 default_elf_asm_output_external (file, decl, name);
6800 /* We output the name if and only if TREE_SYMBOL_REFERENCED is
6801 set in order to avoid putting out names that are never really
6802 used. */
6803 if (TREE_SYMBOL_REFERENCED (DECL_ASSEMBLER_NAME (decl)))
6805 if (!TARGET_EXPLICIT_RELOCS && mips_in_small_data_p (decl))
6807 fputs ("\t.extern\t", file);
6808 assemble_name (file, name);
6809 fprintf (file, ", " HOST_WIDE_INT_PRINT_DEC "\n",
6810 int_size_in_bytes (TREE_TYPE (decl)));
6812 else if (TARGET_IRIX
6813 && mips_abi == ABI_32
6814 && TREE_CODE (decl) == FUNCTION_DECL)
6816 /* In IRIX 5 or IRIX 6 for the O32 ABI, we must output a
6817 `.global name .text' directive for every used but
6818 undefined function. If we don't, the linker may perform
6819 an optimization (skipping over the insns that set $gp)
6820 when it is unsafe. */
6821 fputs ("\t.globl ", file);
6822 assemble_name (file, name);
6823 fputs (" .text\n", file);
6828 /* Emit a new filename to a stream. If we are smuggling stabs, try to
6829 put out a MIPS ECOFF file and a stab. */
6831 void
6832 mips_output_filename (FILE *stream, const char *name)
6835 /* If we are emitting DWARF-2, let dwarf2out handle the ".file"
6836 directives. */
6837 if (write_symbols == DWARF2_DEBUG)
6838 return;
6839 else if (mips_output_filename_first_time)
6841 mips_output_filename_first_time = 0;
6842 num_source_filenames += 1;
6843 current_function_file = name;
6844 fprintf (stream, "\t.file\t%d ", num_source_filenames);
6845 output_quoted_string (stream, name);
6846 putc ('\n', stream);
6849 /* If we are emitting stabs, let dbxout.c handle this (except for
6850 the mips_output_filename_first_time case). */
6851 else if (write_symbols == DBX_DEBUG)
6852 return;
6854 else if (name != current_function_file
6855 && strcmp (name, current_function_file) != 0)
6857 num_source_filenames += 1;
6858 current_function_file = name;
6859 fprintf (stream, "\t.file\t%d ", num_source_filenames);
6860 output_quoted_string (stream, name);
6861 putc ('\n', stream);
6865 /* MIPS implementation of TARGET_ASM_OUTPUT_DWARF_DTPREL. */
6867 static void
6868 mips_output_dwarf_dtprel (FILE *file, int size, rtx x)
6870 switch (size)
6872 case 4:
6873 fputs ("\t.dtprelword\t", file);
6874 break;
6876 case 8:
6877 fputs ("\t.dtpreldword\t", file);
6878 break;
6880 default:
6881 gcc_unreachable ();
6883 output_addr_const (file, x);
6884 fputs ("+0x8000", file);
6887 /* Implement TARGET_DWARF_REGISTER_SPAN. */
6889 static rtx
6890 mips_dwarf_register_span (rtx reg)
6892 rtx high, low;
6893 enum machine_mode mode;
6895 /* By default, GCC maps increasing register numbers to increasing
6896 memory locations, but paired FPRs are always little-endian,
6897 regardless of the prevailing endianness. */
6898 mode = GET_MODE (reg);
6899 if (FP_REG_P (REGNO (reg))
6900 && TARGET_BIG_ENDIAN
6901 && MAX_FPRS_PER_FMT > 1
6902 && GET_MODE_SIZE (mode) > UNITS_PER_FPREG)
6904 gcc_assert (GET_MODE_SIZE (mode) == UNITS_PER_HWFPVALUE);
6905 high = mips_subword (reg, true);
6906 low = mips_subword (reg, false);
6907 return gen_rtx_PARALLEL (VOIDmode, gen_rtvec (2, high, low));
6910 return NULL_RTX;
6913 /* Output an ASCII string, in a space-saving way. PREFIX is the string
6914 that should be written before the opening quote, such as "\t.ascii\t"
6915 for real string data or "\t# " for a comment. */
6917 void
6918 mips_output_ascii (FILE *stream, const char *string_param, size_t len,
6919 const char *prefix)
6921 size_t i;
6922 int cur_pos = 17;
6923 register const unsigned char *string =
6924 (const unsigned char *)string_param;
6926 fprintf (stream, "%s\"", prefix);
6927 for (i = 0; i < len; i++)
6929 register int c = string[i];
6931 if (ISPRINT (c))
6933 if (c == '\\' || c == '\"')
6935 putc ('\\', stream);
6936 cur_pos++;
6938 putc (c, stream);
6939 cur_pos++;
6941 else
6943 fprintf (stream, "\\%03o", c);
6944 cur_pos += 4;
6947 if (cur_pos > 72 && i+1 < len)
6949 cur_pos = 17;
6950 fprintf (stream, "\"\n%s\"", prefix);
6953 fprintf (stream, "\"\n");
6956 #ifdef BSS_SECTION_ASM_OP
6957 /* Implement ASM_OUTPUT_ALIGNED_BSS. This differs from the default only
6958 in the use of sbss. */
6960 void
6961 mips_output_aligned_bss (FILE *stream, tree decl, const char *name,
6962 unsigned HOST_WIDE_INT size, int align)
6964 extern tree last_assemble_variable_decl;
6966 if (mips_in_small_data_p (decl))
6967 switch_to_section (get_named_section (NULL, ".sbss", 0));
6968 else
6969 switch_to_section (bss_section);
6970 ASM_OUTPUT_ALIGN (stream, floor_log2 (align / BITS_PER_UNIT));
6971 last_assemble_variable_decl = decl;
6972 ASM_DECLARE_OBJECT_NAME (stream, name, decl);
6973 ASM_OUTPUT_SKIP (stream, size != 0 ? size : 1);
6975 #endif
6977 /* Emit either a label, .comm, or .lcomm directive. When using assembler
6978 macros, mark the symbol as written so that mips_file_end won't emit an
6979 .extern for it. STREAM is the output file, NAME is the name of the
6980 symbol, INIT_STRING is the string that should be written before the
6981 symbol and FINAL_STRING is the string that should be written after it.
6982 FINAL_STRING is a printf() format that consumes the remaining arguments. */
6984 void
6985 mips_declare_object (FILE *stream, const char *name, const char *init_string,
6986 const char *final_string, ...)
6988 va_list ap;
6990 fputs (init_string, stream);
6991 assemble_name (stream, name);
6992 va_start (ap, final_string);
6993 vfprintf (stream, final_string, ap);
6994 va_end (ap);
6996 if (!TARGET_EXPLICIT_RELOCS)
6998 tree name_tree = get_identifier (name);
6999 TREE_ASM_WRITTEN (name_tree) = 1;
7003 /* Declare a common object of SIZE bytes using asm directive INIT_STRING.
7004 NAME is the name of the object and ALIGN is the required alignment
7005 in bytes. TAKES_ALIGNMENT_P is true if the directive takes a third
7006 alignment argument. */
7008 void
7009 mips_declare_common_object (FILE *stream, const char *name,
7010 const char *init_string,
7011 unsigned HOST_WIDE_INT size,
7012 unsigned int align, bool takes_alignment_p)
7014 if (!takes_alignment_p)
7016 size += (align / BITS_PER_UNIT) - 1;
7017 size -= size % (align / BITS_PER_UNIT);
7018 mips_declare_object (stream, name, init_string,
7019 "," HOST_WIDE_INT_PRINT_UNSIGNED "\n", size);
7021 else
7022 mips_declare_object (stream, name, init_string,
7023 "," HOST_WIDE_INT_PRINT_UNSIGNED ",%u\n",
7024 size, align / BITS_PER_UNIT);
7027 /* Implement ASM_OUTPUT_ALIGNED_DECL_COMMON. This is usually the same as the
7028 elfos.h version, but we also need to handle -muninit-const-in-rodata. */
7030 void
7031 mips_output_aligned_decl_common (FILE *stream, tree decl, const char *name,
7032 unsigned HOST_WIDE_INT size,
7033 unsigned int align)
7035 /* If the target wants uninitialized const declarations in
7036 .rdata then don't put them in .comm. */
7037 if (TARGET_EMBEDDED_DATA && TARGET_UNINIT_CONST_IN_RODATA
7038 && TREE_CODE (decl) == VAR_DECL && TREE_READONLY (decl)
7039 && (DECL_INITIAL (decl) == 0 || DECL_INITIAL (decl) == error_mark_node))
7041 if (TREE_PUBLIC (decl) && DECL_NAME (decl))
7042 targetm.asm_out.globalize_label (stream, name);
7044 switch_to_section (readonly_data_section);
7045 ASM_OUTPUT_ALIGN (stream, floor_log2 (align / BITS_PER_UNIT));
7046 mips_declare_object (stream, name, "",
7047 ":\n\t.space\t" HOST_WIDE_INT_PRINT_UNSIGNED "\n",
7048 size);
7050 else
7051 mips_declare_common_object (stream, name, "\n\t.comm\t",
7052 size, align, true);
7055 #ifdef ASM_OUTPUT_SIZE_DIRECTIVE
7056 extern int size_directive_output;
7058 /* Implement ASM_DECLARE_OBJECT_NAME. This is like most of the standard ELF
7059 definitions except that it uses mips_declare_object() to emit the label. */
7061 void
7062 mips_declare_object_name (FILE *stream, const char *name,
7063 tree decl ATTRIBUTE_UNUSED)
7065 #ifdef ASM_OUTPUT_TYPE_DIRECTIVE
7066 ASM_OUTPUT_TYPE_DIRECTIVE (stream, name, "object");
7067 #endif
7069 size_directive_output = 0;
7070 if (!flag_inhibit_size_directive && DECL_SIZE (decl))
7072 HOST_WIDE_INT size;
7074 size_directive_output = 1;
7075 size = int_size_in_bytes (TREE_TYPE (decl));
7076 ASM_OUTPUT_SIZE_DIRECTIVE (stream, name, size);
7079 mips_declare_object (stream, name, "", ":\n");
7082 /* Implement ASM_FINISH_DECLARE_OBJECT. This is generic ELF stuff. */
7084 void
7085 mips_finish_declare_object (FILE *stream, tree decl, int top_level, int at_end)
7087 const char *name;
7089 name = XSTR (XEXP (DECL_RTL (decl), 0), 0);
7090 if (!flag_inhibit_size_directive
7091 && DECL_SIZE (decl) != 0
7092 && !at_end && top_level
7093 && DECL_INITIAL (decl) == error_mark_node
7094 && !size_directive_output)
7096 HOST_WIDE_INT size;
7098 size_directive_output = 1;
7099 size = int_size_in_bytes (TREE_TYPE (decl));
7100 ASM_OUTPUT_SIZE_DIRECTIVE (stream, name, size);
7103 #endif
7105 /* Return the FOO in the name of the ".mdebug.FOO" section associated
7106 with the current ABI. */
7108 static const char *
7109 mips_mdebug_abi_name (void)
7111 switch (mips_abi)
7113 case ABI_32:
7114 return "abi32";
7115 case ABI_O64:
7116 return "abiO64";
7117 case ABI_N32:
7118 return "abiN32";
7119 case ABI_64:
7120 return "abiN64";
7121 case ABI_EABI:
7122 return TARGET_64BIT ? "eabi64" : "eabi32";
7123 default:
7124 gcc_unreachable ();
7128 /* Implement TARGET_ASM_FILE_START. */
7130 static void
7131 mips_file_start (void)
7133 default_file_start ();
7135 if (!TARGET_IRIX)
7137 /* Generate a special section to describe the ABI switches used to
7138 produce the resultant binary. This used to be done by the assembler
7139 setting bits in the ELF header's flags field, but we have run out of
7140 bits. GDB needs this information in order to be able to correctly
7141 debug these binaries. See the function mips_gdbarch_init() in
7142 gdb/mips-tdep.c. This is unnecessary for the IRIX 5/6 ABIs and
7143 causes unnecessary IRIX 6 ld warnings. */
7144 /* Note - we use fprintf directly rather than calling switch_to_section
7145 because in this way we can avoid creating an allocated section. We
7146 do not want this section to take up any space in the running
7147 executable. */
7148 fprintf (asm_out_file, "\t.section .mdebug.%s\n\t.previous\n",
7149 mips_mdebug_abi_name ());
7151 /* There is no ELF header flag to distinguish long32 forms of the
7152 EABI from long64 forms. Emit a special section to help tools
7153 such as GDB. Do the same for o64, which is sometimes used with
7154 -mlong64. */
7155 if (mips_abi == ABI_EABI || mips_abi == ABI_O64)
7156 fprintf (asm_out_file, "\t.section .gcc_compiled_long%d\n"
7157 "\t.previous\n", TARGET_LONG64 ? 64 : 32);
7159 #ifdef HAVE_AS_GNU_ATTRIBUTE
7160 fprintf (asm_out_file, "\t.gnu_attribute 4, %d\n",
7161 TARGET_HARD_FLOAT_ABI ? (TARGET_DOUBLE_FLOAT ? 1 : 2) : 3);
7162 #endif
7165 /* Generate the pseudo ops that System V.4 wants. */
7166 if (TARGET_ABICALLS)
7167 fprintf (asm_out_file, "\t.abicalls\n");
7169 if (flag_verbose_asm)
7170 fprintf (asm_out_file, "\n%s -G value = %d, Arch = %s, ISA = %d\n",
7171 ASM_COMMENT_START,
7172 mips_section_threshold, mips_arch_info->name, mips_isa);
7176 /* Make the last instruction frame related and note that it performs
7177 the operation described by FRAME_PATTERN. */
7179 static void
7180 mips_set_frame_expr (rtx frame_pattern)
7182 rtx insn;
7184 insn = get_last_insn ();
7185 RTX_FRAME_RELATED_P (insn) = 1;
7186 REG_NOTES (insn) = alloc_EXPR_LIST (REG_FRAME_RELATED_EXPR,
7187 frame_pattern,
7188 REG_NOTES (insn));
7192 /* Return a frame-related rtx that stores REG at MEM.
7193 REG must be a single register. */
7195 static rtx
7196 mips_frame_set (rtx mem, rtx reg)
7198 rtx set;
7200 /* If we're saving the return address register and the dwarf return
7201 address column differs from the hard register number, adjust the
7202 note reg to refer to the former. */
7203 if (REGNO (reg) == GP_REG_FIRST + 31
7204 && DWARF_FRAME_RETURN_COLUMN != GP_REG_FIRST + 31)
7205 reg = gen_rtx_REG (GET_MODE (reg), DWARF_FRAME_RETURN_COLUMN);
7207 set = gen_rtx_SET (VOIDmode, mem, reg);
7208 RTX_FRAME_RELATED_P (set) = 1;
7210 return set;
7213 /* If a MIPS16e SAVE or RESTORE instruction saves or restores register
7214 mips16e_s2_s8_regs[X], it must also save the registers in indexes
7215 X + 1 onwards. Likewise mips16e_a0_a3_regs. */
7216 static const unsigned char mips16e_s2_s8_regs[] = {
7217 30, 23, 22, 21, 20, 19, 18
7219 static const unsigned char mips16e_a0_a3_regs[] = {
7220 4, 5, 6, 7
7223 /* A list of the registers that can be saved by the MIPS16e SAVE instruction,
7224 ordered from the uppermost in memory to the lowest in memory. */
7225 static const unsigned char mips16e_save_restore_regs[] = {
7226 31, 30, 23, 22, 21, 20, 19, 18, 17, 16, 7, 6, 5, 4
7229 /* Return the index of the lowest X in the range [0, SIZE) for which
7230 bit REGS[X] is set in MASK. Return SIZE if there is no such X. */
7232 static unsigned int
7233 mips16e_find_first_register (unsigned int mask, const unsigned char *regs,
7234 unsigned int size)
7236 unsigned int i;
7238 for (i = 0; i < size; i++)
7239 if (BITSET_P (mask, regs[i]))
7240 break;
7242 return i;
7245 /* *MASK_PTR is a mask of general-purpose registers and *NUM_REGS_PTR
7246 is the number of set bits. If *MASK_PTR contains REGS[X] for some X
7247 in [0, SIZE), adjust *MASK_PTR and *NUM_REGS_PTR so that the same
7248 is true for all indexes (X, SIZE). */
7250 static void
7251 mips16e_mask_registers (unsigned int *mask_ptr, const unsigned char *regs,
7252 unsigned int size, unsigned int *num_regs_ptr)
7254 unsigned int i;
7256 i = mips16e_find_first_register (*mask_ptr, regs, size);
7257 for (i++; i < size; i++)
7258 if (!BITSET_P (*mask_ptr, regs[i]))
7260 *num_regs_ptr += 1;
7261 *mask_ptr |= 1 << regs[i];
7265 /* Return a simplified form of X using the register values in REG_VALUES.
7266 REG_VALUES[R] is the last value assigned to hard register R, or null
7267 if R has not been modified.
7269 This function is rather limited, but is good enough for our purposes. */
7271 static rtx
7272 mips16e_collect_propagate_value (rtx x, rtx *reg_values)
7274 rtx x0, x1;
7276 x = avoid_constant_pool_reference (x);
7278 if (UNARY_P (x))
7280 x0 = mips16e_collect_propagate_value (XEXP (x, 0), reg_values);
7281 return simplify_gen_unary (GET_CODE (x), GET_MODE (x),
7282 x0, GET_MODE (XEXP (x, 0)));
7285 if (ARITHMETIC_P (x))
7287 x0 = mips16e_collect_propagate_value (XEXP (x, 0), reg_values);
7288 x1 = mips16e_collect_propagate_value (XEXP (x, 1), reg_values);
7289 return simplify_gen_binary (GET_CODE (x), GET_MODE (x), x0, x1);
7292 if (REG_P (x)
7293 && reg_values[REGNO (x)]
7294 && !rtx_unstable_p (reg_values[REGNO (x)]))
7295 return reg_values[REGNO (x)];
7297 return x;
7300 /* Return true if (set DEST SRC) stores an argument register into its
7301 caller-allocated save slot, storing the number of that argument
7302 register in *REGNO_PTR if so. REG_VALUES is as for
7303 mips16e_collect_propagate_value. */
7305 static bool
7306 mips16e_collect_argument_save_p (rtx dest, rtx src, rtx *reg_values,
7307 unsigned int *regno_ptr)
7309 unsigned int argno, regno;
7310 HOST_WIDE_INT offset, required_offset;
7311 rtx addr, base;
7313 /* Check that this is a word-mode store. */
7314 if (!MEM_P (dest) || !REG_P (src) || GET_MODE (dest) != word_mode)
7315 return false;
7317 /* Check that the register being saved is an unmodified argument
7318 register. */
7319 regno = REGNO (src);
7320 if (regno < GP_ARG_FIRST || regno > GP_ARG_LAST || reg_values[regno])
7321 return false;
7322 argno = regno - GP_ARG_FIRST;
7324 /* Check whether the address is an appropriate stack pointer or
7325 frame pointer access. */
7326 addr = mips16e_collect_propagate_value (XEXP (dest, 0), reg_values);
7327 mips_split_plus (addr, &base, &offset);
7328 required_offset = cfun->machine->frame.total_size + argno * UNITS_PER_WORD;
7329 if (base == hard_frame_pointer_rtx)
7330 required_offset -= cfun->machine->frame.hard_frame_pointer_offset;
7331 else if (base != stack_pointer_rtx)
7332 return false;
7333 if (offset != required_offset)
7334 return false;
7336 *regno_ptr = regno;
7337 return true;
7340 /* A subroutine of mips_expand_prologue, called only when generating
7341 MIPS16e SAVE instructions. Search the start of the function for any
7342 instructions that save argument registers into their caller-allocated
7343 save slots. Delete such instructions and return a value N such that
7344 saving [GP_ARG_FIRST, GP_ARG_FIRST + N) would make all the deleted
7345 instructions redundant. */
7347 static unsigned int
7348 mips16e_collect_argument_saves (void)
7350 rtx reg_values[FIRST_PSEUDO_REGISTER];
7351 rtx insn, next, set, dest, src;
7352 unsigned int nargs, regno;
7354 push_topmost_sequence ();
7355 nargs = 0;
7356 memset (reg_values, 0, sizeof (reg_values));
7357 for (insn = get_insns (); insn; insn = next)
7359 next = NEXT_INSN (insn);
7360 if (NOTE_P (insn))
7361 continue;
7363 if (!INSN_P (insn))
7364 break;
7366 set = PATTERN (insn);
7367 if (GET_CODE (set) != SET)
7368 break;
7370 dest = SET_DEST (set);
7371 src = SET_SRC (set);
7372 if (mips16e_collect_argument_save_p (dest, src, reg_values, &regno))
7374 if (!BITSET_P (cfun->machine->frame.mask, regno))
7376 delete_insn (insn);
7377 nargs = MAX (nargs, (regno - GP_ARG_FIRST) + 1);
7380 else if (REG_P (dest) && GET_MODE (dest) == word_mode)
7381 reg_values[REGNO (dest)]
7382 = mips16e_collect_propagate_value (src, reg_values);
7383 else
7384 break;
7386 pop_topmost_sequence ();
7388 return nargs;
7391 /* Return a move between register REGNO and memory location SP + OFFSET.
7392 Make the move a load if RESTORE_P, otherwise make it a frame-related
7393 store. */
7395 static rtx
7396 mips16e_save_restore_reg (bool restore_p, HOST_WIDE_INT offset,
7397 unsigned int regno)
7399 rtx reg, mem;
7401 mem = gen_frame_mem (SImode, plus_constant (stack_pointer_rtx, offset));
7402 reg = gen_rtx_REG (SImode, regno);
7403 return (restore_p
7404 ? gen_rtx_SET (VOIDmode, reg, mem)
7405 : mips_frame_set (mem, reg));
7408 /* Return RTL for a MIPS16e SAVE or RESTORE instruction; RESTORE_P says which.
7409 The instruction must:
7411 - Allocate or deallocate SIZE bytes in total; SIZE is known
7412 to be nonzero.
7414 - Save or restore as many registers in *MASK_PTR as possible.
7415 The instruction saves the first registers at the top of the
7416 allocated area, with the other registers below it.
7418 - Save NARGS argument registers above the allocated area.
7420 (NARGS is always zero if RESTORE_P.)
7422 The SAVE and RESTORE instructions cannot save and restore all general
7423 registers, so there may be some registers left over for the caller to
7424 handle. Destructively modify *MASK_PTR so that it contains the registers
7425 that still need to be saved or restored. The caller can save these
7426 registers in the memory immediately below *OFFSET_PTR, which is a
7427 byte offset from the bottom of the allocated stack area. */
7429 static rtx
7430 mips16e_build_save_restore (bool restore_p, unsigned int *mask_ptr,
7431 HOST_WIDE_INT *offset_ptr, unsigned int nargs,
7432 HOST_WIDE_INT size)
7434 rtx pattern, set;
7435 HOST_WIDE_INT offset, top_offset;
7436 unsigned int i, regno;
7437 int n;
7439 gcc_assert (cfun->machine->frame.num_fp == 0);
7441 /* Calculate the number of elements in the PARALLEL. We need one element
7442 for the stack adjustment, one for each argument register save, and one
7443 for each additional register move. */
7444 n = 1 + nargs;
7445 for (i = 0; i < ARRAY_SIZE (mips16e_save_restore_regs); i++)
7446 if (BITSET_P (*mask_ptr, mips16e_save_restore_regs[i]))
7447 n++;
7449 /* Create the final PARALLEL. */
7450 pattern = gen_rtx_PARALLEL (VOIDmode, rtvec_alloc (n));
7451 n = 0;
7453 /* Add the stack pointer adjustment. */
7454 set = gen_rtx_SET (VOIDmode, stack_pointer_rtx,
7455 plus_constant (stack_pointer_rtx,
7456 restore_p ? size : -size));
7457 RTX_FRAME_RELATED_P (set) = 1;
7458 XVECEXP (pattern, 0, n++) = set;
7460 /* Stack offsets in the PARALLEL are relative to the old stack pointer. */
7461 top_offset = restore_p ? size : 0;
7463 /* Save the arguments. */
7464 for (i = 0; i < nargs; i++)
7466 offset = top_offset + i * UNITS_PER_WORD;
7467 set = mips16e_save_restore_reg (restore_p, offset, GP_ARG_FIRST + i);
7468 XVECEXP (pattern, 0, n++) = set;
7471 /* Then fill in the other register moves. */
7472 offset = top_offset;
7473 for (i = 0; i < ARRAY_SIZE (mips16e_save_restore_regs); i++)
7475 regno = mips16e_save_restore_regs[i];
7476 if (BITSET_P (*mask_ptr, regno))
7478 offset -= UNITS_PER_WORD;
7479 set = mips16e_save_restore_reg (restore_p, offset, regno);
7480 XVECEXP (pattern, 0, n++) = set;
7481 *mask_ptr &= ~(1 << regno);
7485 /* Tell the caller what offset it should use for the remaining registers. */
7486 *offset_ptr = size + (offset - top_offset);
7488 gcc_assert (n == XVECLEN (pattern, 0));
7490 return pattern;
7493 /* PATTERN is a PARALLEL whose first element adds ADJUST to the stack
7494 pointer. Return true if PATTERN matches the kind of instruction
7495 generated by mips16e_build_save_restore. If INFO is nonnull,
7496 initialize it when returning true. */
7498 bool
7499 mips16e_save_restore_pattern_p (rtx pattern, HOST_WIDE_INT adjust,
7500 struct mips16e_save_restore_info *info)
7502 unsigned int i, nargs, mask, extra;
7503 HOST_WIDE_INT top_offset, save_offset, offset;
7504 rtx set, reg, mem, base;
7505 int n;
7507 if (!GENERATE_MIPS16E_SAVE_RESTORE)
7508 return false;
7510 /* Stack offsets in the PARALLEL are relative to the old stack pointer. */
7511 top_offset = adjust > 0 ? adjust : 0;
7513 /* Interpret all other members of the PARALLEL. */
7514 save_offset = top_offset - UNITS_PER_WORD;
7515 mask = 0;
7516 nargs = 0;
7517 i = 0;
7518 for (n = 1; n < XVECLEN (pattern, 0); n++)
7520 /* Check that we have a SET. */
7521 set = XVECEXP (pattern, 0, n);
7522 if (GET_CODE (set) != SET)
7523 return false;
7525 /* Check that the SET is a load (if restoring) or a store
7526 (if saving). */
7527 mem = adjust > 0 ? SET_SRC (set) : SET_DEST (set);
7528 if (!MEM_P (mem))
7529 return false;
7531 /* Check that the address is the sum of the stack pointer and a
7532 possibly-zero constant offset. */
7533 mips_split_plus (XEXP (mem, 0), &base, &offset);
7534 if (base != stack_pointer_rtx)
7535 return false;
7537 /* Check that SET's other operand is a register. */
7538 reg = adjust > 0 ? SET_DEST (set) : SET_SRC (set);
7539 if (!REG_P (reg))
7540 return false;
7542 /* Check for argument saves. */
7543 if (offset == top_offset + nargs * UNITS_PER_WORD
7544 && REGNO (reg) == GP_ARG_FIRST + nargs)
7545 nargs++;
7546 else if (offset == save_offset)
7548 while (mips16e_save_restore_regs[i++] != REGNO (reg))
7549 if (i == ARRAY_SIZE (mips16e_save_restore_regs))
7550 return false;
7552 mask |= 1 << REGNO (reg);
7553 save_offset -= UNITS_PER_WORD;
7555 else
7556 return false;
7559 /* Check that the restrictions on register ranges are met. */
7560 extra = 0;
7561 mips16e_mask_registers (&mask, mips16e_s2_s8_regs,
7562 ARRAY_SIZE (mips16e_s2_s8_regs), &extra);
7563 mips16e_mask_registers (&mask, mips16e_a0_a3_regs,
7564 ARRAY_SIZE (mips16e_a0_a3_regs), &extra);
7565 if (extra != 0)
7566 return false;
7568 /* Make sure that the topmost argument register is not saved twice.
7569 The checks above ensure that the same is then true for the other
7570 argument registers. */
7571 if (nargs > 0 && BITSET_P (mask, GP_ARG_FIRST + nargs - 1))
7572 return false;
7574 /* Pass back information, if requested. */
7575 if (info)
7577 info->nargs = nargs;
7578 info->mask = mask;
7579 info->size = (adjust > 0 ? adjust : -adjust);
7582 return true;
7585 /* Add a MIPS16e SAVE or RESTORE register-range argument to string S
7586 for the register range [MIN_REG, MAX_REG]. Return a pointer to
7587 the null terminator. */
7589 static char *
7590 mips16e_add_register_range (char *s, unsigned int min_reg,
7591 unsigned int max_reg)
7593 if (min_reg != max_reg)
7594 s += sprintf (s, ",%s-%s", reg_names[min_reg], reg_names[max_reg]);
7595 else
7596 s += sprintf (s, ",%s", reg_names[min_reg]);
7597 return s;
7600 /* Return the assembly instruction for a MIPS16e SAVE or RESTORE instruction.
7601 PATTERN and ADJUST are as for mips16e_save_restore_pattern_p. */
7603 const char *
7604 mips16e_output_save_restore (rtx pattern, HOST_WIDE_INT adjust)
7606 static char buffer[300];
7608 struct mips16e_save_restore_info info;
7609 unsigned int i, end;
7610 char *s;
7612 /* Parse the pattern. */
7613 if (!mips16e_save_restore_pattern_p (pattern, adjust, &info))
7614 gcc_unreachable ();
7616 /* Add the mnemonic. */
7617 s = strcpy (buffer, adjust > 0 ? "restore\t" : "save\t");
7618 s += strlen (s);
7620 /* Save the arguments. */
7621 if (info.nargs > 1)
7622 s += sprintf (s, "%s-%s,", reg_names[GP_ARG_FIRST],
7623 reg_names[GP_ARG_FIRST + info.nargs - 1]);
7624 else if (info.nargs == 1)
7625 s += sprintf (s, "%s,", reg_names[GP_ARG_FIRST]);
7627 /* Emit the amount of stack space to allocate or deallocate. */
7628 s += sprintf (s, "%d", (int) info.size);
7630 /* Save or restore $16. */
7631 if (BITSET_P (info.mask, 16))
7632 s += sprintf (s, ",%s", reg_names[GP_REG_FIRST + 16]);
7634 /* Save or restore $17. */
7635 if (BITSET_P (info.mask, 17))
7636 s += sprintf (s, ",%s", reg_names[GP_REG_FIRST + 17]);
7638 /* Save or restore registers in the range $s2...$s8, which
7639 mips16e_s2_s8_regs lists in decreasing order. Note that this
7640 is a software register range; the hardware registers are not
7641 numbered consecutively. */
7642 end = ARRAY_SIZE (mips16e_s2_s8_regs);
7643 i = mips16e_find_first_register (info.mask, mips16e_s2_s8_regs, end);
7644 if (i < end)
7645 s = mips16e_add_register_range (s, mips16e_s2_s8_regs[end - 1],
7646 mips16e_s2_s8_regs[i]);
7648 /* Save or restore registers in the range $a0...$a3. */
7649 end = ARRAY_SIZE (mips16e_a0_a3_regs);
7650 i = mips16e_find_first_register (info.mask, mips16e_a0_a3_regs, end);
7651 if (i < end)
7652 s = mips16e_add_register_range (s, mips16e_a0_a3_regs[i],
7653 mips16e_a0_a3_regs[end - 1]);
7655 /* Save or restore $31. */
7656 if (BITSET_P (info.mask, 31))
7657 s += sprintf (s, ",%s", reg_names[GP_REG_FIRST + 31]);
7659 return buffer;
7662 /* Return true if the current function has an insn that implicitly
7663 refers to $gp. */
7665 static bool
7666 mips_function_has_gp_insn (void)
7668 /* Don't bother rechecking if we found one last time. */
7669 if (!cfun->machine->has_gp_insn_p)
7671 rtx insn;
7673 push_topmost_sequence ();
7674 for (insn = get_insns (); insn; insn = NEXT_INSN (insn))
7675 if (INSN_P (insn)
7676 && GET_CODE (PATTERN (insn)) != USE
7677 && GET_CODE (PATTERN (insn)) != CLOBBER
7678 && (get_attr_got (insn) != GOT_UNSET
7679 || small_data_pattern (PATTERN (insn), VOIDmode)))
7680 break;
7681 pop_topmost_sequence ();
7683 cfun->machine->has_gp_insn_p = (insn != 0);
7685 return cfun->machine->has_gp_insn_p;
7689 /* Return the register that should be used as the global pointer
7690 within this function. Return 0 if the function doesn't need
7691 a global pointer. */
7693 static unsigned int
7694 mips_global_pointer (void)
7696 unsigned int regno;
7698 /* $gp is always available unless we're using a GOT. */
7699 if (!TARGET_USE_GOT)
7700 return GLOBAL_POINTER_REGNUM;
7702 /* We must always provide $gp when it is used implicitly. */
7703 if (!TARGET_EXPLICIT_RELOCS)
7704 return GLOBAL_POINTER_REGNUM;
7706 /* FUNCTION_PROFILER includes a jal macro, so we need to give it
7707 a valid gp. */
7708 if (current_function_profile)
7709 return GLOBAL_POINTER_REGNUM;
7711 /* If the function has a nonlocal goto, $gp must hold the correct
7712 global pointer for the target function. */
7713 if (current_function_has_nonlocal_goto)
7714 return GLOBAL_POINTER_REGNUM;
7716 /* If the gp is never referenced, there's no need to initialize it.
7717 Note that reload can sometimes introduce constant pool references
7718 into a function that otherwise didn't need them. For example,
7719 suppose we have an instruction like:
7721 (set (reg:DF R1) (float:DF (reg:SI R2)))
7723 If R2 turns out to be constant such as 1, the instruction may have a
7724 REG_EQUAL note saying that R1 == 1.0. Reload then has the option of
7725 using this constant if R2 doesn't get allocated to a register.
7727 In cases like these, reload will have added the constant to the pool
7728 but no instruction will yet refer to it. */
7729 if (!df_regs_ever_live_p (GLOBAL_POINTER_REGNUM)
7730 && !current_function_uses_const_pool
7731 && !mips_function_has_gp_insn ())
7732 return 0;
7734 /* We need a global pointer, but perhaps we can use a call-clobbered
7735 register instead of $gp. */
7736 if (TARGET_CALL_SAVED_GP && current_function_is_leaf)
7737 for (regno = GP_REG_FIRST; regno <= GP_REG_LAST; regno++)
7738 if (!df_regs_ever_live_p (regno)
7739 && call_really_used_regs[regno]
7740 && !fixed_regs[regno]
7741 && regno != PIC_FUNCTION_ADDR_REGNUM)
7742 return regno;
7744 return GLOBAL_POINTER_REGNUM;
7747 /* Return true if the current function returns its value in a floating-point
7748 register in MIPS16 mode. */
7750 static bool
7751 mips16_cfun_returns_in_fpr_p (void)
7753 tree return_type = DECL_RESULT (current_function_decl);
7754 return (TARGET_MIPS16
7755 && TARGET_HARD_FLOAT_ABI
7756 && !aggregate_value_p (return_type, current_function_decl)
7757 && mips_return_mode_in_fpr_p (DECL_MODE (return_type)));
7761 /* Return true if the current function must save REGNO. */
7763 static bool
7764 mips_save_reg_p (unsigned int regno)
7766 /* We only need to save $gp if TARGET_CALL_SAVED_GP and only then
7767 if we have not chosen a call-clobbered substitute. */
7768 if (regno == GLOBAL_POINTER_REGNUM)
7769 return TARGET_CALL_SAVED_GP && cfun->machine->global_pointer == regno;
7771 /* Check call-saved registers. */
7772 if ((current_function_saves_all_registers || df_regs_ever_live_p (regno))
7773 && !call_really_used_regs[regno])
7774 return true;
7776 /* Save both registers in an FPR pair if either one is used. This is
7777 needed for the case when MIN_FPRS_PER_FMT == 1, which allows the odd
7778 register to be used without the even register. */
7779 if (FP_REG_P (regno)
7780 && MAX_FPRS_PER_FMT == 2
7781 && df_regs_ever_live_p (regno + 1)
7782 && !call_really_used_regs[regno + 1])
7783 return true;
7785 /* We need to save the old frame pointer before setting up a new one. */
7786 if (regno == HARD_FRAME_POINTER_REGNUM && frame_pointer_needed)
7787 return true;
7789 /* Check for registers that must be saved for FUNCTION_PROFILER. */
7790 if (current_function_profile && MIPS_SAVE_REG_FOR_PROFILING_P (regno))
7791 return true;
7793 /* We need to save the incoming return address if it is ever clobbered
7794 within the function, if __builtin_eh_return is being used to set a
7795 different return address, or if a stub is being used to return a
7796 value in FPRs. */
7797 if (regno == GP_REG_FIRST + 31
7798 && (df_regs_ever_live_p (regno)
7799 || current_function_calls_eh_return
7800 || mips16_cfun_returns_in_fpr_p ()))
7801 return true;
7803 return false;
7806 /* Populate the current function's mips_frame_info structure.
7808 MIPS stack frames look like:
7810 +-------------------------------+
7812 | incoming stack arguments |
7814 +-------------------------------+
7816 | caller-allocated save area |
7817 A | for register arguments |
7819 +-------------------------------+ <-- incoming stack pointer
7821 | callee-allocated save area |
7822 B | for arguments that are |
7823 | split between registers and |
7824 | the stack |
7826 +-------------------------------+ <-- arg_pointer_rtx
7828 C | callee-allocated save area |
7829 | for register varargs |
7831 +-------------------------------+ <-- frame_pointer_rtx + fp_sp_offset
7832 | | + UNITS_PER_HWFPVALUE
7833 | FPR save area |
7835 +-------------------------------+ <-- frame_pointer_rtx + gp_sp_offset
7836 | | + UNITS_PER_WORD
7837 | GPR save area |
7839 +-------------------------------+
7840 | | \
7841 | local variables | | var_size
7842 | | /
7843 +-------------------------------+
7844 | | \
7845 | $gp save area | | cprestore_size
7846 | | /
7847 P +-------------------------------+ <-- hard_frame_pointer_rtx for
7848 | | MIPS16 code
7849 | outgoing stack arguments |
7851 +-------------------------------+
7853 | caller-allocated save area |
7854 | for register arguments |
7856 +-------------------------------+ <-- stack_pointer_rtx
7857 frame_pointer_rtx
7858 hard_frame_pointer_rtx for
7859 non-MIPS16 code.
7861 At least two of A, B and C will be empty.
7863 Dynamic stack allocations such as alloca insert data at point P.
7864 They decrease stack_pointer_rtx but leave frame_pointer_rtx and
7865 hard_frame_pointer_rtx unchanged. */
7867 static void
7868 mips_compute_frame_info (void)
7870 struct mips_frame_info *frame;
7871 HOST_WIDE_INT offset, size;
7872 unsigned int regno, i;
7874 frame = &cfun->machine->frame;
7875 memset (frame, 0, sizeof (*frame));
7876 size = get_frame_size ();
7878 cfun->machine->global_pointer = mips_global_pointer ();
7880 /* The first STARTING_FRAME_OFFSET bytes contain the outgoing argument
7881 area and the $gp save slot. This area isn't needed in leaf functions,
7882 but if the target-independent frame size is nonzero, we're committed
7883 to allocating it anyway. */
7884 if (size == 0 && current_function_is_leaf)
7886 /* The MIPS 3.0 linker does not like functions that dynamically
7887 allocate the stack and have 0 for STACK_DYNAMIC_OFFSET, since it
7888 looks like we are trying to create a second frame pointer to the
7889 function, so allocate some stack space to make it happy. */
7890 if (current_function_calls_alloca)
7891 frame->args_size = REG_PARM_STACK_SPACE (cfun->decl);
7892 else
7893 frame->args_size = 0;
7894 frame->cprestore_size = 0;
7896 else
7898 frame->args_size = current_function_outgoing_args_size;
7899 frame->cprestore_size = STARTING_FRAME_OFFSET - frame->args_size;
7901 offset = frame->args_size + frame->cprestore_size;
7903 /* Move above the local variables. */
7904 frame->var_size = MIPS_STACK_ALIGN (size);
7905 offset += frame->var_size;
7907 /* Find out which GPRs we need to save. */
7908 for (regno = GP_REG_FIRST; regno <= GP_REG_LAST; regno++)
7909 if (mips_save_reg_p (regno))
7911 frame->num_gp++;
7912 frame->mask |= 1 << (regno - GP_REG_FIRST);
7915 /* If this function calls eh_return, we must also save and restore the
7916 EH data registers. */
7917 if (current_function_calls_eh_return)
7918 for (i = 0; EH_RETURN_DATA_REGNO (i) != INVALID_REGNUM; i++)
7920 frame->num_gp++;
7921 frame->mask |= 1 << (EH_RETURN_DATA_REGNO (i) - GP_REG_FIRST);
7924 /* The MIPS16e SAVE and RESTORE instructions have two ranges of registers:
7925 $a3-$a0 and $s2-$s8. If we save one register in the range, we must
7926 save all later registers too. */
7927 if (GENERATE_MIPS16E_SAVE_RESTORE)
7929 mips16e_mask_registers (&frame->mask, mips16e_s2_s8_regs,
7930 ARRAY_SIZE (mips16e_s2_s8_regs), &frame->num_gp);
7931 mips16e_mask_registers (&frame->mask, mips16e_a0_a3_regs,
7932 ARRAY_SIZE (mips16e_a0_a3_regs), &frame->num_gp);
7935 /* Move above the GPR save area. */
7936 if (frame->num_gp > 0)
7938 offset += MIPS_STACK_ALIGN (frame->num_gp * UNITS_PER_WORD);
7939 frame->gp_sp_offset = offset - UNITS_PER_WORD;
7942 /* Find out which FPRs we need to save. This loop must iterate over
7943 the same space as its companion in mips_for_each_saved_reg. */
7944 if (TARGET_HARD_FLOAT)
7945 for (regno = FP_REG_FIRST; regno <= FP_REG_LAST; regno += MAX_FPRS_PER_FMT)
7946 if (mips_save_reg_p (regno))
7948 frame->num_fp += MAX_FPRS_PER_FMT;
7949 frame->fmask |= ~(~0 << MAX_FPRS_PER_FMT) << (regno - FP_REG_FIRST);
7952 /* Move above the FPR save area. */
7953 if (frame->num_fp > 0)
7955 offset += MIPS_STACK_ALIGN (frame->num_fp * UNITS_PER_FPREG);
7956 frame->fp_sp_offset = offset - UNITS_PER_HWFPVALUE;
7959 /* Move above the callee-allocated varargs save area. */
7960 offset += MIPS_STACK_ALIGN (cfun->machine->varargs_size);
7961 frame->arg_pointer_offset = offset;
7963 /* Move above the callee-allocated area for pretend stack arguments. */
7964 offset += current_function_pretend_args_size;
7965 frame->total_size = offset;
7967 /* Work out the offsets of the save areas from the top of the frame. */
7968 if (frame->gp_sp_offset > 0)
7969 frame->gp_save_offset = frame->gp_sp_offset - offset;
7970 if (frame->fp_sp_offset > 0)
7971 frame->fp_save_offset = frame->fp_sp_offset - offset;
7973 /* MIPS16 code offsets the frame pointer by the size of the outgoing
7974 arguments. This tends to increase the chances of using unextended
7975 instructions for local variables and incoming arguments. */
7976 if (TARGET_MIPS16)
7977 frame->hard_frame_pointer_offset = frame->args_size;
7980 /* Return the style of GP load sequence that is being used for the
7981 current function. */
7983 enum mips_loadgp_style
7984 mips_current_loadgp_style (void)
7986 if (!TARGET_USE_GOT || cfun->machine->global_pointer == 0)
7987 return LOADGP_NONE;
7989 if (TARGET_RTP_PIC)
7990 return LOADGP_RTP;
7992 if (TARGET_ABSOLUTE_ABICALLS)
7993 return LOADGP_ABSOLUTE;
7995 return TARGET_NEWABI ? LOADGP_NEWABI : LOADGP_OLDABI;
7998 /* Implement FRAME_POINTER_REQUIRED. */
8000 bool
8001 mips_frame_pointer_required (void)
8003 /* If the function contains dynamic stack allocations, we need to
8004 use the frame pointer to access the static parts of the frame. */
8005 if (current_function_calls_alloca)
8006 return true;
8008 /* In MIPS16 mode, we need a frame pointer for a large frame; otherwise,
8009 reload may be unable to compute the address of a local variable,
8010 since there is no way to add a large constant to the stack pointer
8011 without using a second temporary register. */
8012 if (TARGET_MIPS16)
8014 mips_compute_frame_info ();
8015 if (!SMALL_OPERAND (cfun->machine->frame.total_size))
8016 return true;
8019 return false;
8022 /* Implement INITIAL_ELIMINATION_OFFSET. FROM is either the frame
8023 pointer or argument pointer. TO is either the stack pointer or
8024 hard frame pointer. */
8026 HOST_WIDE_INT
8027 mips_initial_elimination_offset (int from, int to)
8029 HOST_WIDE_INT offset;
8031 mips_compute_frame_info ();
8033 /* Set OFFSET to the offset from the soft frame pointer, which is also
8034 the offset from the end-of-prologue stack pointer. */
8035 switch (from)
8037 case FRAME_POINTER_REGNUM:
8038 offset = 0;
8039 break;
8041 case ARG_POINTER_REGNUM:
8042 offset = cfun->machine->frame.arg_pointer_offset;
8043 break;
8045 default:
8046 gcc_unreachable ();
8049 if (to == HARD_FRAME_POINTER_REGNUM)
8050 offset -= cfun->machine->frame.hard_frame_pointer_offset;
8052 return offset;
8055 /* Implement TARGET_EXTRA_LIVE_ON_ENTRY. Some code models use the incoming
8056 value of PIC_FUNCTION_ADDR_REGNUM to set up the global pointer. */
8058 static void
8059 mips_extra_live_on_entry (bitmap regs)
8061 if (TARGET_USE_GOT && !TARGET_ABSOLUTE_ABICALLS)
8062 bitmap_set_bit (regs, PIC_FUNCTION_ADDR_REGNUM);
8065 /* Implement RETURN_ADDR_RTX. Note, we do not support moving
8066 back to a previous frame. */
8069 mips_return_addr (int count, rtx frame ATTRIBUTE_UNUSED)
8071 if (count != 0)
8072 return const0_rtx;
8074 return get_hard_reg_initial_val (Pmode, GP_REG_FIRST + 31);
8077 /* Emit code to change the current function's return address to
8078 ADDRESS. SCRATCH is available as a scratch register, if needed.
8079 ADDRESS and SCRATCH are both word-mode GPRs. */
8081 void
8082 mips_set_return_address (rtx address, rtx scratch)
8084 rtx slot_address;
8086 gcc_assert ((cfun->machine->frame.mask >> 31) & 1);
8087 slot_address = mips_add_offset (scratch, stack_pointer_rtx,
8088 cfun->machine->frame.gp_sp_offset);
8090 mips_emit_move (gen_frame_mem (GET_MODE (address), slot_address), address);
8093 /* Restore $gp from its save slot. Valid only when using o32 or
8094 o64 abicalls. */
8096 void
8097 mips_restore_gp (void)
8099 rtx address;
8101 gcc_assert (TARGET_ABICALLS && TARGET_OLDABI);
8103 address = mips_add_offset (pic_offset_table_rtx,
8104 frame_pointer_needed
8105 ? hard_frame_pointer_rtx
8106 : stack_pointer_rtx,
8107 current_function_outgoing_args_size);
8109 mips_emit_move (pic_offset_table_rtx, gen_frame_mem (Pmode, address));
8110 if (!TARGET_EXPLICIT_RELOCS)
8111 emit_insn (gen_blockage ());
8114 /* A function to save or store a register. The first argument is the
8115 register and the second is the stack slot. */
8116 typedef void (*mips_save_restore_fn) (rtx, rtx);
8118 /* Use FN to save or restore register REGNO. MODE is the register's
8119 mode and OFFSET is the offset of its save slot from the current
8120 stack pointer. */
8122 static void
8123 mips_save_restore_reg (enum machine_mode mode, int regno,
8124 HOST_WIDE_INT offset, mips_save_restore_fn fn)
8126 rtx mem;
8128 mem = gen_frame_mem (mode, plus_constant (stack_pointer_rtx, offset));
8130 fn (gen_rtx_REG (mode, regno), mem);
8134 /* Call FN for each register that is saved by the current function.
8135 SP_OFFSET is the offset of the current stack pointer from the start
8136 of the frame. */
8138 static void
8139 mips_for_each_saved_reg (HOST_WIDE_INT sp_offset, mips_save_restore_fn fn)
8141 enum machine_mode fpr_mode;
8142 HOST_WIDE_INT offset;
8143 int regno;
8145 /* Save registers starting from high to low. The debuggers prefer at least
8146 the return register be stored at func+4, and also it allows us not to
8147 need a nop in the epilogue if at least one register is reloaded in
8148 addition to return address. */
8149 offset = cfun->machine->frame.gp_sp_offset - sp_offset;
8150 for (regno = GP_REG_LAST; regno >= GP_REG_FIRST; regno--)
8151 if (BITSET_P (cfun->machine->frame.mask, regno - GP_REG_FIRST))
8153 mips_save_restore_reg (word_mode, regno, offset, fn);
8154 offset -= UNITS_PER_WORD;
8157 /* This loop must iterate over the same space as its companion in
8158 mips_compute_frame_info. */
8159 offset = cfun->machine->frame.fp_sp_offset - sp_offset;
8160 fpr_mode = (TARGET_SINGLE_FLOAT ? SFmode : DFmode);
8161 for (regno = (FP_REG_LAST - MAX_FPRS_PER_FMT + 1);
8162 regno >= FP_REG_FIRST;
8163 regno -= MAX_FPRS_PER_FMT)
8164 if (BITSET_P (cfun->machine->frame.fmask, regno - FP_REG_FIRST))
8166 mips_save_restore_reg (fpr_mode, regno, offset, fn);
8167 offset -= GET_MODE_SIZE (fpr_mode);
8171 /* If we're generating n32 or n64 abicalls, and the current function
8172 does not use $28 as its global pointer, emit a cplocal directive.
8173 Use pic_offset_table_rtx as the argument to the directive. */
8175 static void
8176 mips_output_cplocal (void)
8178 if (!TARGET_EXPLICIT_RELOCS
8179 && cfun->machine->global_pointer > 0
8180 && cfun->machine->global_pointer != GLOBAL_POINTER_REGNUM)
8181 output_asm_insn (".cplocal %+", 0);
8184 /* Set up the stack and frame (if desired) for the function. */
8186 static void
8187 mips_output_function_prologue (FILE *file, HOST_WIDE_INT size ATTRIBUTE_UNUSED)
8189 const char *fnname;
8190 HOST_WIDE_INT tsize = cfun->machine->frame.total_size;
8192 #ifdef SDB_DEBUGGING_INFO
8193 if (debug_info_level != DINFO_LEVEL_TERSE && write_symbols == SDB_DEBUG)
8194 SDB_OUTPUT_SOURCE_LINE (file, DECL_SOURCE_LINE (current_function_decl));
8195 #endif
8197 /* In mips16 mode, we may need to generate a 32 bit to handle
8198 floating point arguments. The linker will arrange for any 32-bit
8199 functions to call this stub, which will then jump to the 16-bit
8200 function proper. */
8201 if (TARGET_MIPS16
8202 && TARGET_HARD_FLOAT_ABI
8203 && current_function_args_info.fp_code != 0)
8204 build_mips16_function_stub ();
8206 /* Select the mips16 mode for this function. */
8207 if (TARGET_MIPS16)
8208 fprintf (file, "\t.set\tmips16\n");
8209 else
8210 fprintf (file, "\t.set\tnomips16\n");
8212 if (!FUNCTION_NAME_ALREADY_DECLARED)
8214 /* Get the function name the same way that toplev.c does before calling
8215 assemble_start_function. This is needed so that the name used here
8216 exactly matches the name used in ASM_DECLARE_FUNCTION_NAME. */
8217 fnname = XSTR (XEXP (DECL_RTL (current_function_decl), 0), 0);
8219 if (!flag_inhibit_size_directive)
8221 fputs ("\t.ent\t", file);
8222 assemble_name (file, fnname);
8223 fputs ("\n", file);
8226 assemble_name (file, fnname);
8227 fputs (":\n", file);
8230 /* Stop mips_file_end from treating this function as external. */
8231 if (TARGET_IRIX && mips_abi == ABI_32)
8232 TREE_ASM_WRITTEN (DECL_NAME (cfun->decl)) = 1;
8234 if (!flag_inhibit_size_directive)
8236 /* .frame FRAMEREG, FRAMESIZE, RETREG */
8237 fprintf (file,
8238 "\t.frame\t%s," HOST_WIDE_INT_PRINT_DEC ",%s\t\t"
8239 "# vars= " HOST_WIDE_INT_PRINT_DEC ", regs= %d/%d"
8240 ", args= " HOST_WIDE_INT_PRINT_DEC
8241 ", gp= " HOST_WIDE_INT_PRINT_DEC "\n",
8242 (reg_names[(frame_pointer_needed)
8243 ? HARD_FRAME_POINTER_REGNUM : STACK_POINTER_REGNUM]),
8244 (frame_pointer_needed
8245 ? tsize - cfun->machine->frame.hard_frame_pointer_offset
8246 : tsize),
8247 reg_names[GP_REG_FIRST + 31],
8248 cfun->machine->frame.var_size,
8249 cfun->machine->frame.num_gp,
8250 cfun->machine->frame.num_fp,
8251 cfun->machine->frame.args_size,
8252 cfun->machine->frame.cprestore_size);
8254 /* .mask MASK, GPOFFSET; .fmask FPOFFSET */
8255 fprintf (file, "\t.mask\t0x%08x," HOST_WIDE_INT_PRINT_DEC "\n",
8256 cfun->machine->frame.mask,
8257 cfun->machine->frame.gp_save_offset);
8258 fprintf (file, "\t.fmask\t0x%08x," HOST_WIDE_INT_PRINT_DEC "\n",
8259 cfun->machine->frame.fmask,
8260 cfun->machine->frame.fp_save_offset);
8262 /* Require:
8263 OLD_SP == *FRAMEREG + FRAMESIZE => can find old_sp from nominated FP reg.
8264 HIGHEST_GP_SAVED == *FRAMEREG + FRAMESIZE + GPOFFSET => can find saved regs. */
8267 if (mips_current_loadgp_style () == LOADGP_OLDABI)
8269 /* Handle the initialization of $gp for SVR4 PIC. */
8270 if (!cfun->machine->all_noreorder_p)
8271 output_asm_insn ("%(.cpload\t%^%)", 0);
8272 else
8273 output_asm_insn ("%(.cpload\t%^\n\t%<", 0);
8275 else if (cfun->machine->all_noreorder_p)
8276 output_asm_insn ("%(%<", 0);
8278 /* Tell the assembler which register we're using as the global
8279 pointer. This is needed for thunks, since they can use either
8280 explicit relocs or assembler macros. */
8281 mips_output_cplocal ();
8284 /* Do any necessary cleanup after a function to restore stack, frame,
8285 and regs. */
8287 #define RA_MASK BITMASK_HIGH /* 1 << 31 */
8289 static void
8290 mips_output_function_epilogue (FILE *file ATTRIBUTE_UNUSED,
8291 HOST_WIDE_INT size ATTRIBUTE_UNUSED)
8293 /* Reinstate the normal $gp. */
8294 SET_REGNO (pic_offset_table_rtx, GLOBAL_POINTER_REGNUM);
8295 mips_output_cplocal ();
8297 if (cfun->machine->all_noreorder_p)
8299 /* Avoid using %>%) since it adds excess whitespace. */
8300 output_asm_insn (".set\tmacro", 0);
8301 output_asm_insn (".set\treorder", 0);
8302 set_noreorder = set_nomacro = 0;
8305 if (!FUNCTION_NAME_ALREADY_DECLARED && !flag_inhibit_size_directive)
8307 const char *fnname;
8309 /* Get the function name the same way that toplev.c does before calling
8310 assemble_start_function. This is needed so that the name used here
8311 exactly matches the name used in ASM_DECLARE_FUNCTION_NAME. */
8312 fnname = XSTR (XEXP (DECL_RTL (current_function_decl), 0), 0);
8313 fputs ("\t.end\t", file);
8314 assemble_name (file, fnname);
8315 fputs ("\n", file);
8319 /* Save register REG to MEM. Make the instruction frame-related. */
8321 static void
8322 mips_save_reg (rtx reg, rtx mem)
8324 if (GET_MODE (reg) == DFmode && !TARGET_FLOAT64)
8326 rtx x1, x2;
8328 if (mips_split_64bit_move_p (mem, reg))
8329 mips_split_doubleword_move (mem, reg);
8330 else
8331 mips_emit_move (mem, reg);
8333 x1 = mips_frame_set (mips_subword (mem, 0), mips_subword (reg, 0));
8334 x2 = mips_frame_set (mips_subword (mem, 1), mips_subword (reg, 1));
8335 mips_set_frame_expr (gen_rtx_PARALLEL (VOIDmode, gen_rtvec (2, x1, x2)));
8337 else
8339 if (TARGET_MIPS16
8340 && REGNO (reg) != GP_REG_FIRST + 31
8341 && !M16_REG_P (REGNO (reg)))
8343 /* Save a non-mips16 register by moving it through a temporary.
8344 We don't need to do this for $31 since there's a special
8345 instruction for it. */
8346 mips_emit_move (MIPS_PROLOGUE_TEMP (GET_MODE (reg)), reg);
8347 mips_emit_move (mem, MIPS_PROLOGUE_TEMP (GET_MODE (reg)));
8349 else
8350 mips_emit_move (mem, reg);
8352 mips_set_frame_expr (mips_frame_set (mem, reg));
8356 /* The __gnu_local_gp symbol. */
8358 static GTY(()) rtx mips_gnu_local_gp;
8360 /* If we're generating n32 or n64 abicalls, emit instructions
8361 to set up the global pointer. */
8363 static void
8364 mips_emit_loadgp (void)
8366 rtx addr, offset, incoming_address, base, index;
8368 switch (mips_current_loadgp_style ())
8370 case LOADGP_ABSOLUTE:
8371 if (mips_gnu_local_gp == NULL)
8373 mips_gnu_local_gp = gen_rtx_SYMBOL_REF (Pmode, "__gnu_local_gp");
8374 SYMBOL_REF_FLAGS (mips_gnu_local_gp) |= SYMBOL_FLAG_LOCAL;
8376 emit_insn (gen_loadgp_absolute (mips_gnu_local_gp));
8377 break;
8379 case LOADGP_NEWABI:
8380 addr = XEXP (DECL_RTL (current_function_decl), 0);
8381 offset = mips_unspec_address (addr, SYMBOL_GOTOFF_LOADGP);
8382 incoming_address = gen_rtx_REG (Pmode, PIC_FUNCTION_ADDR_REGNUM);
8383 emit_insn (gen_loadgp_newabi (offset, incoming_address));
8384 if (!TARGET_EXPLICIT_RELOCS)
8385 emit_insn (gen_loadgp_blockage ());
8386 break;
8388 case LOADGP_RTP:
8389 base = gen_rtx_SYMBOL_REF (Pmode, ggc_strdup (VXWORKS_GOTT_BASE));
8390 index = gen_rtx_SYMBOL_REF (Pmode, ggc_strdup (VXWORKS_GOTT_INDEX));
8391 emit_insn (gen_loadgp_rtp (base, index));
8392 if (!TARGET_EXPLICIT_RELOCS)
8393 emit_insn (gen_loadgp_blockage ());
8394 break;
8396 default:
8397 break;
8401 /* Expand the prologue into a bunch of separate insns. */
8403 void
8404 mips_expand_prologue (void)
8406 HOST_WIDE_INT size;
8407 unsigned int nargs;
8408 rtx insn;
8410 if (cfun->machine->global_pointer > 0)
8411 SET_REGNO (pic_offset_table_rtx, cfun->machine->global_pointer);
8413 size = cfun->machine->frame.total_size;
8415 /* Save the registers. Allocate up to MIPS_MAX_FIRST_STACK_STEP
8416 bytes beforehand; this is enough to cover the register save area
8417 without going out of range. */
8418 if ((cfun->machine->frame.mask | cfun->machine->frame.fmask) != 0)
8420 HOST_WIDE_INT step1;
8422 step1 = MIN (size, MIPS_MAX_FIRST_STACK_STEP);
8424 if (GENERATE_MIPS16E_SAVE_RESTORE)
8426 HOST_WIDE_INT offset;
8427 unsigned int mask, regno;
8429 /* Try to merge argument stores into the save instruction. */
8430 nargs = mips16e_collect_argument_saves ();
8432 /* Build the save instruction. */
8433 mask = cfun->machine->frame.mask;
8434 insn = mips16e_build_save_restore (false, &mask, &offset,
8435 nargs, step1);
8436 RTX_FRAME_RELATED_P (emit_insn (insn)) = 1;
8437 size -= step1;
8439 /* Check if we need to save other registers. */
8440 for (regno = GP_REG_FIRST; regno < GP_REG_LAST; regno++)
8441 if (BITSET_P (mask, regno - GP_REG_FIRST))
8443 offset -= UNITS_PER_WORD;
8444 mips_save_restore_reg (word_mode, regno,
8445 offset, mips_save_reg);
8448 else
8450 insn = gen_add3_insn (stack_pointer_rtx,
8451 stack_pointer_rtx,
8452 GEN_INT (-step1));
8453 RTX_FRAME_RELATED_P (emit_insn (insn)) = 1;
8454 size -= step1;
8455 mips_for_each_saved_reg (size, mips_save_reg);
8459 /* Allocate the rest of the frame. */
8460 if (size > 0)
8462 if (SMALL_OPERAND (-size))
8463 RTX_FRAME_RELATED_P (emit_insn (gen_add3_insn (stack_pointer_rtx,
8464 stack_pointer_rtx,
8465 GEN_INT (-size)))) = 1;
8466 else
8468 mips_emit_move (MIPS_PROLOGUE_TEMP (Pmode), GEN_INT (size));
8469 if (TARGET_MIPS16)
8471 /* There are no instructions to add or subtract registers
8472 from the stack pointer, so use the frame pointer as a
8473 temporary. We should always be using a frame pointer
8474 in this case anyway. */
8475 gcc_assert (frame_pointer_needed);
8476 mips_emit_move (hard_frame_pointer_rtx, stack_pointer_rtx);
8477 emit_insn (gen_sub3_insn (hard_frame_pointer_rtx,
8478 hard_frame_pointer_rtx,
8479 MIPS_PROLOGUE_TEMP (Pmode)));
8480 mips_emit_move (stack_pointer_rtx, hard_frame_pointer_rtx);
8482 else
8483 emit_insn (gen_sub3_insn (stack_pointer_rtx,
8484 stack_pointer_rtx,
8485 MIPS_PROLOGUE_TEMP (Pmode)));
8487 /* Describe the combined effect of the previous instructions. */
8488 mips_set_frame_expr
8489 (gen_rtx_SET (VOIDmode, stack_pointer_rtx,
8490 plus_constant (stack_pointer_rtx, -size)));
8494 /* Set up the frame pointer, if we're using one. */
8495 if (frame_pointer_needed)
8497 HOST_WIDE_INT offset;
8499 offset = cfun->machine->frame.hard_frame_pointer_offset;
8500 if (offset == 0)
8502 insn = mips_emit_move (hard_frame_pointer_rtx, stack_pointer_rtx);
8503 RTX_FRAME_RELATED_P (insn) = 1;
8505 else if (SMALL_OPERAND (offset))
8507 insn = gen_add3_insn (hard_frame_pointer_rtx,
8508 stack_pointer_rtx, GEN_INT (offset));
8509 RTX_FRAME_RELATED_P (emit_insn (insn)) = 1;
8511 else
8513 mips_emit_move (MIPS_PROLOGUE_TEMP (Pmode), GEN_INT (offset));
8514 mips_emit_move (hard_frame_pointer_rtx, stack_pointer_rtx);
8515 emit_insn (gen_add3_insn (hard_frame_pointer_rtx,
8516 hard_frame_pointer_rtx,
8517 MIPS_PROLOGUE_TEMP (Pmode)));
8518 mips_set_frame_expr
8519 (gen_rtx_SET (VOIDmode, hard_frame_pointer_rtx,
8520 plus_constant (stack_pointer_rtx, offset)));
8524 mips_emit_loadgp ();
8526 /* If generating o32/o64 abicalls, save $gp on the stack. */
8527 if (TARGET_ABICALLS && TARGET_OLDABI && !current_function_is_leaf)
8528 emit_insn (gen_cprestore (GEN_INT (current_function_outgoing_args_size)));
8530 /* If we are profiling, make sure no instructions are scheduled before
8531 the call to mcount. */
8533 if (current_function_profile)
8534 emit_insn (gen_blockage ());
8537 /* Emit instructions to restore register REG from slot MEM. */
8539 static void
8540 mips_restore_reg (rtx reg, rtx mem)
8542 /* There's no mips16 instruction to load $31 directly. Load into
8543 $7 instead and adjust the return insn appropriately. */
8544 if (TARGET_MIPS16 && REGNO (reg) == GP_REG_FIRST + 31)
8545 reg = gen_rtx_REG (GET_MODE (reg), 7);
8547 if (TARGET_MIPS16 && !M16_REG_P (REGNO (reg)))
8549 /* Can't restore directly; move through a temporary. */
8550 mips_emit_move (MIPS_EPILOGUE_TEMP (GET_MODE (reg)), mem);
8551 mips_emit_move (reg, MIPS_EPILOGUE_TEMP (GET_MODE (reg)));
8553 else
8554 mips_emit_move (reg, mem);
8558 /* Expand the epilogue into a bunch of separate insns. SIBCALL_P is true
8559 if this epilogue precedes a sibling call, false if it is for a normal
8560 "epilogue" pattern. */
8562 void
8563 mips_expand_epilogue (int sibcall_p)
8565 HOST_WIDE_INT step1, step2;
8566 rtx base, target;
8568 if (!sibcall_p && mips_can_use_return_insn ())
8570 emit_jump_insn (gen_return ());
8571 return;
8574 /* In mips16 mode, if the return value should go into a floating-point
8575 register, we need to call a helper routine to copy it over. */
8576 if (mips16_cfun_returns_in_fpr_p ())
8577 mips16_copy_fpr_return_value ();
8579 /* Split the frame into two. STEP1 is the amount of stack we should
8580 deallocate before restoring the registers. STEP2 is the amount we
8581 should deallocate afterwards.
8583 Start off by assuming that no registers need to be restored. */
8584 step1 = cfun->machine->frame.total_size;
8585 step2 = 0;
8587 /* Work out which register holds the frame address. */
8588 if (!frame_pointer_needed)
8589 base = stack_pointer_rtx;
8590 else
8592 base = hard_frame_pointer_rtx;
8593 step1 -= cfun->machine->frame.hard_frame_pointer_offset;
8596 /* If we need to restore registers, deallocate as much stack as
8597 possible in the second step without going out of range. */
8598 if ((cfun->machine->frame.mask | cfun->machine->frame.fmask) != 0)
8600 step2 = MIN (step1, MIPS_MAX_FIRST_STACK_STEP);
8601 step1 -= step2;
8604 /* Set TARGET to BASE + STEP1. */
8605 target = base;
8606 if (step1 > 0)
8608 rtx adjust;
8610 /* Get an rtx for STEP1 that we can add to BASE. */
8611 adjust = GEN_INT (step1);
8612 if (!SMALL_OPERAND (step1))
8614 mips_emit_move (MIPS_EPILOGUE_TEMP (Pmode), adjust);
8615 adjust = MIPS_EPILOGUE_TEMP (Pmode);
8618 /* Normal mode code can copy the result straight into $sp. */
8619 if (!TARGET_MIPS16)
8620 target = stack_pointer_rtx;
8622 emit_insn (gen_add3_insn (target, base, adjust));
8625 /* Copy TARGET into the stack pointer. */
8626 if (target != stack_pointer_rtx)
8627 mips_emit_move (stack_pointer_rtx, target);
8629 /* If we're using addressing macros, $gp is implicitly used by all
8630 SYMBOL_REFs. We must emit a blockage insn before restoring $gp
8631 from the stack. */
8632 if (TARGET_CALL_SAVED_GP && !TARGET_EXPLICIT_RELOCS)
8633 emit_insn (gen_blockage ());
8635 if (GENERATE_MIPS16E_SAVE_RESTORE && cfun->machine->frame.mask != 0)
8637 unsigned int regno, mask;
8638 HOST_WIDE_INT offset;
8639 rtx restore;
8641 /* Generate the restore instruction. */
8642 mask = cfun->machine->frame.mask;
8643 restore = mips16e_build_save_restore (true, &mask, &offset, 0, step2);
8645 /* Restore any other registers manually. */
8646 for (regno = GP_REG_FIRST; regno < GP_REG_LAST; regno++)
8647 if (BITSET_P (mask, regno - GP_REG_FIRST))
8649 offset -= UNITS_PER_WORD;
8650 mips_save_restore_reg (word_mode, regno, offset, mips_restore_reg);
8653 /* Restore the remaining registers and deallocate the final bit
8654 of the frame. */
8655 emit_insn (restore);
8657 else
8659 /* Restore the registers. */
8660 mips_for_each_saved_reg (cfun->machine->frame.total_size - step2,
8661 mips_restore_reg);
8663 /* Deallocate the final bit of the frame. */
8664 if (step2 > 0)
8665 emit_insn (gen_add3_insn (stack_pointer_rtx,
8666 stack_pointer_rtx,
8667 GEN_INT (step2)));
8670 /* Add in the __builtin_eh_return stack adjustment. We need to
8671 use a temporary in mips16 code. */
8672 if (current_function_calls_eh_return)
8674 if (TARGET_MIPS16)
8676 mips_emit_move (MIPS_EPILOGUE_TEMP (Pmode), stack_pointer_rtx);
8677 emit_insn (gen_add3_insn (MIPS_EPILOGUE_TEMP (Pmode),
8678 MIPS_EPILOGUE_TEMP (Pmode),
8679 EH_RETURN_STACKADJ_RTX));
8680 mips_emit_move (stack_pointer_rtx, MIPS_EPILOGUE_TEMP (Pmode));
8682 else
8683 emit_insn (gen_add3_insn (stack_pointer_rtx,
8684 stack_pointer_rtx,
8685 EH_RETURN_STACKADJ_RTX));
8688 if (!sibcall_p)
8690 /* When generating MIPS16 code, the normal mips_for_each_saved_reg
8691 path will restore the return address into $7 rather than $31. */
8692 if (TARGET_MIPS16
8693 && !GENERATE_MIPS16E_SAVE_RESTORE
8694 && (cfun->machine->frame.mask & RA_MASK) != 0)
8695 emit_jump_insn (gen_return_internal (gen_rtx_REG (Pmode,
8696 GP_REG_FIRST + 7)));
8697 else
8698 emit_jump_insn (gen_return_internal (gen_rtx_REG (Pmode,
8699 GP_REG_FIRST + 31)));
8703 /* Return nonzero if this function is known to have a null epilogue.
8704 This allows the optimizer to omit jumps to jumps if no stack
8705 was created. */
8708 mips_can_use_return_insn (void)
8710 if (! reload_completed)
8711 return 0;
8713 if (current_function_profile)
8714 return 0;
8716 /* In mips16 mode, a function that returns a floating point value
8717 needs to arrange to copy the return value into the floating point
8718 registers. */
8719 if (mips16_cfun_returns_in_fpr_p ())
8720 return 0;
8722 return cfun->machine->frame.total_size == 0;
8725 /* Return true if register REGNO can store a value of mode MODE.
8726 The result of this function is cached in mips_hard_regno_mode_ok. */
8728 static bool
8729 mips_hard_regno_mode_ok_p (unsigned int regno, enum machine_mode mode)
8731 unsigned int size;
8732 enum mode_class class;
8734 if (mode == CCV2mode)
8735 return (ISA_HAS_8CC
8736 && ST_REG_P (regno)
8737 && (regno - ST_REG_FIRST) % 2 == 0);
8739 if (mode == CCV4mode)
8740 return (ISA_HAS_8CC
8741 && ST_REG_P (regno)
8742 && (regno - ST_REG_FIRST) % 4 == 0);
8744 if (mode == CCmode)
8746 if (!ISA_HAS_8CC)
8747 return regno == FPSW_REGNUM;
8749 return (ST_REG_P (regno)
8750 || GP_REG_P (regno)
8751 || FP_REG_P (regno));
8754 size = GET_MODE_SIZE (mode);
8755 class = GET_MODE_CLASS (mode);
8757 if (GP_REG_P (regno))
8758 return ((regno - GP_REG_FIRST) & 1) == 0 || size <= UNITS_PER_WORD;
8760 if (FP_REG_P (regno)
8761 && (((regno - FP_REG_FIRST) % MAX_FPRS_PER_FMT) == 0
8762 || (MIN_FPRS_PER_FMT == 1 && size <= UNITS_PER_FPREG)))
8764 /* Allow TFmode for CCmode reloads. */
8765 if (mode == TFmode && ISA_HAS_8CC)
8766 return true;
8768 if (class == MODE_FLOAT
8769 || class == MODE_COMPLEX_FLOAT
8770 || class == MODE_VECTOR_FLOAT)
8771 return size <= UNITS_PER_FPVALUE;
8773 /* Allow integer modes that fit into a single register. We need
8774 to put integers into FPRs when using instructions like CVT
8775 and TRUNC. There's no point allowing sizes smaller than a word,
8776 because the FPU has no appropriate load/store instructions. */
8777 if (class == MODE_INT)
8778 return size >= MIN_UNITS_PER_WORD && size <= UNITS_PER_FPREG;
8781 if (ACC_REG_P (regno)
8782 && (INTEGRAL_MODE_P (mode) || ALL_FIXED_POINT_MODE_P (mode)))
8784 if (size <= UNITS_PER_WORD)
8785 return true;
8787 if (size <= UNITS_PER_WORD * 2)
8788 return (DSP_ACC_REG_P (regno)
8789 ? ((regno - DSP_ACC_REG_FIRST) & 1) == 0
8790 : regno == MD_REG_FIRST);
8793 if (ALL_COP_REG_P (regno))
8794 return class == MODE_INT && size <= UNITS_PER_WORD;
8796 return false;
8799 /* Implement HARD_REGNO_NREGS. */
8801 unsigned int
8802 mips_hard_regno_nregs (int regno, enum machine_mode mode)
8804 if (ST_REG_P (regno))
8805 /* The size of FP status registers is always 4, because they only hold
8806 CCmode values, and CCmode is always considered to be 4 bytes wide. */
8807 return (GET_MODE_SIZE (mode) + 3) / 4;
8809 if (FP_REG_P (regno))
8810 return (GET_MODE_SIZE (mode) + UNITS_PER_FPREG - 1) / UNITS_PER_FPREG;
8812 /* All other registers are word-sized. */
8813 return (GET_MODE_SIZE (mode) + UNITS_PER_WORD - 1) / UNITS_PER_WORD;
8816 /* Implement CLASS_MAX_NREGS, taking the maximum of the cases
8817 in mips_hard_regno_nregs. */
8820 mips_class_max_nregs (enum reg_class class, enum machine_mode mode)
8822 int size;
8823 HARD_REG_SET left;
8825 size = 0x8000;
8826 COPY_HARD_REG_SET (left, reg_class_contents[(int) class]);
8827 if (hard_reg_set_intersect_p (left, reg_class_contents[(int) ST_REGS]))
8829 size = MIN (size, 4);
8830 AND_COMPL_HARD_REG_SET (left, reg_class_contents[(int) ST_REGS]);
8832 if (hard_reg_set_intersect_p (left, reg_class_contents[(int) FP_REGS]))
8834 size = MIN (size, UNITS_PER_FPREG);
8835 AND_COMPL_HARD_REG_SET (left, reg_class_contents[(int) FP_REGS]);
8837 if (!hard_reg_set_empty_p (left))
8838 size = MIN (size, UNITS_PER_WORD);
8839 return (GET_MODE_SIZE (mode) + size - 1) / size;
8842 /* Return true if registers of class CLASS cannot change from mode FROM
8843 to mode TO. */
8845 bool
8846 mips_cannot_change_mode_class (enum machine_mode from ATTRIBUTE_UNUSED,
8847 enum machine_mode to ATTRIBUTE_UNUSED,
8848 enum reg_class class)
8850 /* There are several problems with changing the modes of values
8851 in floating-point registers:
8853 - When a multi-word value is stored in paired floating-point
8854 registers, the first register always holds the low word.
8855 We therefore can't allow FPRs to change between single-word
8856 and multi-word modes on big-endian targets.
8858 - GCC assumes that each word of a multiword register can be accessed
8859 individually using SUBREGs. This is not true for floating-point
8860 registers if they are bigger than a word.
8862 - Loading a 32-bit value into a 64-bit floating-point register
8863 will not sign-extend the value, despite what LOAD_EXTEND_OP says.
8864 We can't allow FPRs to change from SImode to to a wider mode on
8865 64-bit targets.
8867 - If the FPU has already interpreted a value in one format, we must
8868 not ask it to treat the value as having a different format.
8870 We therefore only allow changes between 4-byte and smaller integer
8871 values, all of which have the "W" format as far as the FPU is
8872 concerned. */
8873 return (reg_classes_intersect_p (FP_REGS, class)
8874 && (GET_MODE_CLASS (from) != MODE_INT
8875 || GET_MODE_CLASS (to) != MODE_INT
8876 || GET_MODE_SIZE (from) > 4
8877 || GET_MODE_SIZE (to) > 4));
8880 /* Return true if moves in mode MODE can use the FPU's mov.fmt instruction. */
8882 static bool
8883 mips_mode_ok_for_mov_fmt_p (enum machine_mode mode)
8885 switch (mode)
8887 case SFmode:
8888 return TARGET_HARD_FLOAT;
8890 case DFmode:
8891 return TARGET_HARD_FLOAT && TARGET_DOUBLE_FLOAT;
8893 case V2SFmode:
8894 return TARGET_HARD_FLOAT && TARGET_PAIRED_SINGLE_FLOAT;
8896 default:
8897 return false;
8901 /* Implement MODES_TIEABLE_P. */
8903 bool
8904 mips_modes_tieable_p (enum machine_mode mode1, enum machine_mode mode2)
8906 /* FPRs allow no mode punning, so it's not worth tying modes if we'd
8907 prefer to put one of them in FPRs. */
8908 return (mode1 == mode2
8909 || (!mips_mode_ok_for_mov_fmt_p (mode1)
8910 && !mips_mode_ok_for_mov_fmt_p (mode2)));
8913 /* Implement PREFERRED_RELOAD_CLASS. */
8915 enum reg_class
8916 mips_preferred_reload_class (rtx x, enum reg_class class)
8918 if (mips_dangerous_for_la25_p (x) && reg_class_subset_p (LEA_REGS, class))
8919 return LEA_REGS;
8921 if (reg_class_subset_p (FP_REGS, class)
8922 && mips_mode_ok_for_mov_fmt_p (GET_MODE (x)))
8923 return FP_REGS;
8925 if (reg_class_subset_p (GR_REGS, class))
8926 class = GR_REGS;
8928 if (TARGET_MIPS16 && reg_class_subset_p (M16_REGS, class))
8929 class = M16_REGS;
8931 return class;
8934 /* Return a number assessing the cost of moving a register in class
8935 FROM to class TO. The classes are expressed using the enumeration
8936 values such as `GENERAL_REGS'. A value of 2 is the default; other
8937 values are interpreted relative to that.
8939 It is not required that the cost always equal 2 when FROM is the
8940 same as TO; on some machines it is expensive to move between
8941 registers if they are not general registers.
8943 If reload sees an insn consisting of a single `set' between two
8944 hard registers, and if `REGISTER_MOVE_COST' applied to their
8945 classes returns a value of 2, reload does not check to ensure that
8946 the constraints of the insn are met. Setting a cost of other than
8947 2 will allow reload to verify that the constraints are met. You
8948 should do this if the `movM' pattern's constraints do not allow
8949 such copying.
8951 ??? We make the cost of moving from HI/LO into general
8952 registers the same as for one of moving general registers to
8953 HI/LO for TARGET_MIPS16 in order to prevent allocating a
8954 pseudo to HI/LO. This might hurt optimizations though, it
8955 isn't clear if it is wise. And it might not work in all cases. We
8956 could solve the DImode LO reg problem by using a multiply, just
8957 like reload_{in,out}si. We could solve the SImode/HImode HI reg
8958 problem by using divide instructions. divu puts the remainder in
8959 the HI reg, so doing a divide by -1 will move the value in the HI
8960 reg for all values except -1. We could handle that case by using a
8961 signed divide, e.g. -1 / 2 (or maybe 1 / -2?). We'd have to emit
8962 a compare/branch to test the input value to see which instruction
8963 we need to use. This gets pretty messy, but it is feasible. */
8966 mips_register_move_cost (enum machine_mode mode,
8967 enum reg_class to, enum reg_class from)
8969 if (TARGET_MIPS16)
8971 if (reg_class_subset_p (from, GENERAL_REGS)
8972 && reg_class_subset_p (to, GENERAL_REGS))
8974 if (reg_class_subset_p (from, M16_REGS)
8975 || reg_class_subset_p (to, M16_REGS))
8976 return 2;
8977 /* Two MOVEs. */
8978 return 4;
8981 else if (reg_class_subset_p (from, GENERAL_REGS))
8983 if (reg_class_subset_p (to, GENERAL_REGS))
8984 return 2;
8985 if (reg_class_subset_p (to, FP_REGS))
8986 return 4;
8987 if (reg_class_subset_p (to, ALL_COP_AND_GR_REGS))
8988 return 5;
8989 if (reg_class_subset_p (to, ACC_REGS))
8990 return 6;
8992 else if (reg_class_subset_p (to, GENERAL_REGS))
8994 if (reg_class_subset_p (from, FP_REGS))
8995 return 4;
8996 if (reg_class_subset_p (from, ST_REGS))
8997 /* LUI followed by MOVF. */
8998 return 4;
8999 if (reg_class_subset_p (from, ALL_COP_AND_GR_REGS))
9000 return 5;
9001 if (reg_class_subset_p (from, ACC_REGS))
9002 return 6;
9004 else if (reg_class_subset_p (from, FP_REGS))
9006 if (reg_class_subset_p (to, FP_REGS)
9007 && mips_mode_ok_for_mov_fmt_p (mode))
9008 return 4;
9009 if (reg_class_subset_p (to, ST_REGS))
9010 /* An expensive sequence. */
9011 return 8;
9014 return 12;
9017 /* This function returns the register class required for a secondary
9018 register when copying between one of the registers in CLASS, and X,
9019 using MODE. If IN_P is nonzero, the copy is going from X to the
9020 register, otherwise the register is the source. A return value of
9021 NO_REGS means that no secondary register is required. */
9023 enum reg_class
9024 mips_secondary_reload_class (enum reg_class class,
9025 enum machine_mode mode, rtx x, int in_p)
9027 int regno;
9029 /* If X is a constant that cannot be loaded into $25, it must be loaded
9030 into some other GPR. No other register class allows a direct move. */
9031 if (mips_dangerous_for_la25_p (x))
9032 return reg_class_subset_p (class, LEA_REGS) ? NO_REGS : LEA_REGS;
9034 regno = true_regnum (x);
9035 if (TARGET_MIPS16)
9037 /* In MIPS16 mode, every move must involve a member of M16_REGS. */
9038 if (!reg_class_subset_p (class, M16_REGS) && !M16_REG_P (regno))
9039 return M16_REGS;
9041 /* We can't really copy to HI or LO at all in MIPS16 mode. */
9042 if (in_p ? reg_classes_intersect_p (class, ACC_REGS) : ACC_REG_P (regno))
9043 return M16_REGS;
9045 return NO_REGS;
9048 /* Copying from accumulator registers to anywhere other than a general
9049 register requires a temporary general register. */
9050 if (reg_class_subset_p (class, ACC_REGS))
9051 return GP_REG_P (regno) ? NO_REGS : GR_REGS;
9052 if (ACC_REG_P (regno))
9053 return reg_class_subset_p (class, GR_REGS) ? NO_REGS : GR_REGS;
9055 /* We can only copy a value to a condition code register from a
9056 floating point register, and even then we require a scratch
9057 floating point register. We can only copy a value out of a
9058 condition code register into a general register. */
9059 if (reg_class_subset_p (class, ST_REGS))
9061 if (in_p)
9062 return FP_REGS;
9063 return GP_REG_P (regno) ? NO_REGS : GR_REGS;
9065 if (ST_REG_P (regno))
9067 if (!in_p)
9068 return FP_REGS;
9069 return reg_class_subset_p (class, GR_REGS) ? NO_REGS : GR_REGS;
9072 if (reg_class_subset_p (class, FP_REGS))
9074 if (MEM_P (x)
9075 && (GET_MODE_SIZE (mode) == 4 || GET_MODE_SIZE (mode) == 8))
9076 /* In this case we can use lwc1, swc1, ldc1 or sdc1. We'll use
9077 pairs of lwc1s and swc1s if ldc1 and sdc1 are not supported. */
9078 return NO_REGS;
9080 if (GP_REG_P (regno) || x == CONST0_RTX (mode))
9081 /* In this case we can use mtc1, mfc1, dmtc1 or dmfc1. */
9082 return NO_REGS;
9084 if (CONSTANT_P (x) && !targetm.cannot_force_const_mem (x))
9085 /* We can force the constant to memory and use lwc1
9086 and ldc1. As above, we will use pairs of lwc1s if
9087 ldc1 is not supported. */
9088 return NO_REGS;
9090 if (FP_REG_P (regno) && mips_mode_ok_for_mov_fmt_p (mode))
9091 /* In this case we can use mov.fmt. */
9092 return NO_REGS;
9094 /* Otherwise, we need to reload through an integer register. */
9095 return GR_REGS;
9097 if (FP_REG_P (regno))
9098 return reg_class_subset_p (class, GR_REGS) ? NO_REGS : GR_REGS;
9100 return NO_REGS;
9103 /* SImode values are represented as sign-extended to DImode. */
9105 static int
9106 mips_mode_rep_extended (enum machine_mode mode, enum machine_mode mode_rep)
9108 if (TARGET_64BIT && mode == SImode && mode_rep == DImode)
9109 return SIGN_EXTEND;
9111 return UNKNOWN;
9114 static bool
9115 mips_valid_pointer_mode (enum machine_mode mode)
9117 return (mode == SImode || (TARGET_64BIT && mode == DImode));
9120 /* Target hook for vector_mode_supported_p. */
9122 static bool
9123 mips_vector_mode_supported_p (enum machine_mode mode)
9125 switch (mode)
9127 case V2SFmode:
9128 return TARGET_PAIRED_SINGLE_FLOAT;
9130 case V2HImode:
9131 case V4QImode:
9132 case V2HQmode:
9133 case V2UHQmode:
9134 case V2HAmode:
9135 case V2UHAmode:
9136 case V4QQmode:
9137 case V4UQQmode:
9138 return TARGET_DSP;
9140 default:
9141 return false;
9145 /* Implement TARGET_SCALAR_MODE_SUPPORTED_P. */
9147 static bool
9148 mips_scalar_mode_supported_p (enum machine_mode mode)
9150 if (ALL_FIXED_POINT_MODE_P (mode)
9151 && GET_MODE_PRECISION (mode) <= 2 * BITS_PER_WORD)
9152 return true;
9154 return default_scalar_mode_supported_p (mode);
9156 /* This function does three things:
9158 - Register the special divsi3 and modsi3 functions if -mfix-vr4120.
9159 - Register the mips16 hardware floating point stubs.
9160 - Register the gofast functions if selected using --enable-gofast. */
9162 #include "config/gofast.h"
9164 static void
9165 mips_init_libfuncs (void)
9167 if (TARGET_FIX_VR4120)
9169 set_optab_libfunc (sdiv_optab, SImode, "__vr4120_divsi3");
9170 set_optab_libfunc (smod_optab, SImode, "__vr4120_modsi3");
9173 if (TARGET_MIPS16 && TARGET_HARD_FLOAT_ABI)
9175 set_optab_libfunc (add_optab, SFmode, "__mips16_addsf3");
9176 set_optab_libfunc (sub_optab, SFmode, "__mips16_subsf3");
9177 set_optab_libfunc (smul_optab, SFmode, "__mips16_mulsf3");
9178 set_optab_libfunc (sdiv_optab, SFmode, "__mips16_divsf3");
9180 set_optab_libfunc (eq_optab, SFmode, "__mips16_eqsf2");
9181 set_optab_libfunc (ne_optab, SFmode, "__mips16_nesf2");
9182 set_optab_libfunc (gt_optab, SFmode, "__mips16_gtsf2");
9183 set_optab_libfunc (ge_optab, SFmode, "__mips16_gesf2");
9184 set_optab_libfunc (lt_optab, SFmode, "__mips16_ltsf2");
9185 set_optab_libfunc (le_optab, SFmode, "__mips16_lesf2");
9186 set_optab_libfunc (unord_optab, SFmode, "__mips16_unordsf2");
9188 set_conv_libfunc (sfix_optab, SImode, SFmode, "__mips16_fix_truncsfsi");
9189 set_conv_libfunc (sfloat_optab, SFmode, SImode, "__mips16_floatsisf");
9190 set_conv_libfunc (ufloat_optab, SFmode, SImode, "__mips16_floatunsisf");
9192 if (TARGET_DOUBLE_FLOAT)
9194 set_optab_libfunc (add_optab, DFmode, "__mips16_adddf3");
9195 set_optab_libfunc (sub_optab, DFmode, "__mips16_subdf3");
9196 set_optab_libfunc (smul_optab, DFmode, "__mips16_muldf3");
9197 set_optab_libfunc (sdiv_optab, DFmode, "__mips16_divdf3");
9199 set_optab_libfunc (eq_optab, DFmode, "__mips16_eqdf2");
9200 set_optab_libfunc (ne_optab, DFmode, "__mips16_nedf2");
9201 set_optab_libfunc (gt_optab, DFmode, "__mips16_gtdf2");
9202 set_optab_libfunc (ge_optab, DFmode, "__mips16_gedf2");
9203 set_optab_libfunc (lt_optab, DFmode, "__mips16_ltdf2");
9204 set_optab_libfunc (le_optab, DFmode, "__mips16_ledf2");
9205 set_optab_libfunc (unord_optab, DFmode, "__mips16_unorddf2");
9207 set_conv_libfunc (sext_optab, DFmode, SFmode, "__mips16_extendsfdf2");
9208 set_conv_libfunc (trunc_optab, SFmode, DFmode, "__mips16_truncdfsf2");
9210 set_conv_libfunc (sfix_optab, SImode, DFmode, "__mips16_fix_truncdfsi");
9211 set_conv_libfunc (sfloat_optab, DFmode, SImode, "__mips16_floatsidf");
9212 set_conv_libfunc (ufloat_optab, DFmode, SImode, "__mips16_floatunsidf");
9215 else
9216 gofast_maybe_init_libfuncs ();
9219 /* Return the length of INSN. LENGTH is the initial length computed by
9220 attributes in the machine-description file. */
9223 mips_adjust_insn_length (rtx insn, int length)
9225 /* A unconditional jump has an unfilled delay slot if it is not part
9226 of a sequence. A conditional jump normally has a delay slot, but
9227 does not on MIPS16. */
9228 if (CALL_P (insn) || (TARGET_MIPS16 ? simplejump_p (insn) : JUMP_P (insn)))
9229 length += 4;
9231 /* See how many nops might be needed to avoid hardware hazards. */
9232 if (!cfun->machine->ignore_hazard_length_p && INSN_CODE (insn) >= 0)
9233 switch (get_attr_hazard (insn))
9235 case HAZARD_NONE:
9236 break;
9238 case HAZARD_DELAY:
9239 length += 4;
9240 break;
9242 case HAZARD_HILO:
9243 length += 8;
9244 break;
9247 /* All MIPS16 instructions are a measly two bytes. */
9248 if (TARGET_MIPS16)
9249 length /= 2;
9251 return length;
9255 /* Return an asm sequence to start a noat block and load the address
9256 of a label into $1. */
9258 const char *
9259 mips_output_load_label (void)
9261 if (TARGET_EXPLICIT_RELOCS)
9262 switch (mips_abi)
9264 case ABI_N32:
9265 return "%[lw\t%@,%%got_page(%0)(%+)\n\taddiu\t%@,%@,%%got_ofst(%0)";
9267 case ABI_64:
9268 return "%[ld\t%@,%%got_page(%0)(%+)\n\tdaddiu\t%@,%@,%%got_ofst(%0)";
9270 default:
9271 if (ISA_HAS_LOAD_DELAY)
9272 return "%[lw\t%@,%%got(%0)(%+)%#\n\taddiu\t%@,%@,%%lo(%0)";
9273 return "%[lw\t%@,%%got(%0)(%+)\n\taddiu\t%@,%@,%%lo(%0)";
9275 else
9277 if (Pmode == DImode)
9278 return "%[dla\t%@,%0";
9279 else
9280 return "%[la\t%@,%0";
9284 /* Return the assembly code for INSN, which has the operands given by
9285 OPERANDS, and which branches to OPERANDS[1] if some condition is true.
9286 BRANCH_IF_TRUE is the asm template that should be used if OPERANDS[1]
9287 is in range of a direct branch. BRANCH_IF_FALSE is an inverted
9288 version of BRANCH_IF_TRUE. */
9290 const char *
9291 mips_output_conditional_branch (rtx insn, rtx *operands,
9292 const char *branch_if_true,
9293 const char *branch_if_false)
9295 unsigned int length;
9296 rtx taken, not_taken;
9298 length = get_attr_length (insn);
9299 if (length <= 8)
9301 /* Just a simple conditional branch. */
9302 mips_branch_likely = (final_sequence && INSN_ANNULLED_BRANCH_P (insn));
9303 return branch_if_true;
9306 /* Generate a reversed branch around a direct jump. This fallback does
9307 not use branch-likely instructions. */
9308 mips_branch_likely = false;
9309 not_taken = gen_label_rtx ();
9310 taken = operands[1];
9312 /* Generate the reversed branch to NOT_TAKEN. */
9313 operands[1] = not_taken;
9314 output_asm_insn (branch_if_false, operands);
9316 /* If INSN has a delay slot, we must provide delay slots for both the
9317 branch to NOT_TAKEN and the conditional jump. We must also ensure
9318 that INSN's delay slot is executed in the appropriate cases. */
9319 if (final_sequence)
9321 /* This first delay slot will always be executed, so use INSN's
9322 delay slot if is not annulled. */
9323 if (!INSN_ANNULLED_BRANCH_P (insn))
9325 final_scan_insn (XVECEXP (final_sequence, 0, 1),
9326 asm_out_file, optimize, 1, NULL);
9327 INSN_DELETED_P (XVECEXP (final_sequence, 0, 1)) = 1;
9329 else
9330 output_asm_insn ("nop", 0);
9331 fprintf (asm_out_file, "\n");
9334 /* Output the unconditional branch to TAKEN. */
9335 if (length <= 16)
9336 output_asm_insn ("j\t%0%/", &taken);
9337 else
9339 output_asm_insn (mips_output_load_label (), &taken);
9340 output_asm_insn ("jr\t%@%]%/", 0);
9343 /* Now deal with its delay slot; see above. */
9344 if (final_sequence)
9346 /* This delay slot will only be executed if the branch is taken.
9347 Use INSN's delay slot if is annulled. */
9348 if (INSN_ANNULLED_BRANCH_P (insn))
9350 final_scan_insn (XVECEXP (final_sequence, 0, 1),
9351 asm_out_file, optimize, 1, NULL);
9352 INSN_DELETED_P (XVECEXP (final_sequence, 0, 1)) = 1;
9354 else
9355 output_asm_insn ("nop", 0);
9356 fprintf (asm_out_file, "\n");
9359 /* Output NOT_TAKEN. */
9360 (*targetm.asm_out.internal_label) (asm_out_file, "L",
9361 CODE_LABEL_NUMBER (not_taken));
9362 return "";
9365 /* Return the assembly code for INSN, which branches to OPERANDS[1]
9366 if some ordered condition is true. The condition is given by
9367 OPERANDS[0] if !INVERTED_P, otherwise it is the inverse of
9368 OPERANDS[0]. OPERANDS[2] is the comparison's first operand;
9369 its second is always zero. */
9371 const char *
9372 mips_output_order_conditional_branch (rtx insn, rtx *operands, bool inverted_p)
9374 const char *branch[2];
9376 /* Make BRANCH[1] branch to OPERANDS[1] when the condition is true.
9377 Make BRANCH[0] branch on the inverse condition. */
9378 switch (GET_CODE (operands[0]))
9380 /* These cases are equivalent to comparisons against zero. */
9381 case LEU:
9382 inverted_p = !inverted_p;
9383 /* Fall through. */
9384 case GTU:
9385 branch[!inverted_p] = MIPS_BRANCH ("bne", "%2,%.,%1");
9386 branch[inverted_p] = MIPS_BRANCH ("beq", "%2,%.,%1");
9387 break;
9389 /* These cases are always true or always false. */
9390 case LTU:
9391 inverted_p = !inverted_p;
9392 /* Fall through. */
9393 case GEU:
9394 branch[!inverted_p] = MIPS_BRANCH ("beq", "%.,%.,%1");
9395 branch[inverted_p] = MIPS_BRANCH ("bne", "%.,%.,%1");
9396 break;
9398 default:
9399 branch[!inverted_p] = MIPS_BRANCH ("b%C0z", "%2,%1");
9400 branch[inverted_p] = MIPS_BRANCH ("b%N0z", "%2,%1");
9401 break;
9403 return mips_output_conditional_branch (insn, operands, branch[1], branch[0]);
9406 /* Used to output div or ddiv instruction DIVISION, which has the operands
9407 given by OPERANDS. Add in a divide-by-zero check if needed.
9409 When working around R4000 and R4400 errata, we need to make sure that
9410 the division is not immediately followed by a shift[1][2]. We also
9411 need to stop the division from being put into a branch delay slot[3].
9412 The easiest way to avoid both problems is to add a nop after the
9413 division. When a divide-by-zero check is needed, this nop can be
9414 used to fill the branch delay slot.
9416 [1] If a double-word or a variable shift executes immediately
9417 after starting an integer division, the shift may give an
9418 incorrect result. See quotations of errata #16 and #28 from
9419 "MIPS R4000PC/SC Errata, Processor Revision 2.2 and 3.0"
9420 in mips.md for details.
9422 [2] A similar bug to [1] exists for all revisions of the
9423 R4000 and the R4400 when run in an MC configuration.
9424 From "MIPS R4000MC Errata, Processor Revision 2.2 and 3.0":
9426 "19. In this following sequence:
9428 ddiv (or ddivu or div or divu)
9429 dsll32 (or dsrl32, dsra32)
9431 if an MPT stall occurs, while the divide is slipping the cpu
9432 pipeline, then the following double shift would end up with an
9433 incorrect result.
9435 Workaround: The compiler needs to avoid generating any
9436 sequence with divide followed by extended double shift."
9438 This erratum is also present in "MIPS R4400MC Errata, Processor
9439 Revision 1.0" and "MIPS R4400MC Errata, Processor Revision 2.0
9440 & 3.0" as errata #10 and #4, respectively.
9442 [3] From "MIPS R4000PC/SC Errata, Processor Revision 2.2 and 3.0"
9443 (also valid for MIPS R4000MC processors):
9445 "52. R4000SC: This bug does not apply for the R4000PC.
9447 There are two flavors of this bug:
9449 1) If the instruction just after divide takes an RF exception
9450 (tlb-refill, tlb-invalid) and gets an instruction cache
9451 miss (both primary and secondary) and the line which is
9452 currently in secondary cache at this index had the first
9453 data word, where the bits 5..2 are set, then R4000 would
9454 get a wrong result for the div.
9458 div r8, r9
9459 ------------------- # end-of page. -tlb-refill
9463 div r8, r9
9464 ------------------- # end-of page. -tlb-invalid
9467 2) If the divide is in the taken branch delay slot, where the
9468 target takes RF exception and gets an I-cache miss for the
9469 exception vector or where I-cache miss occurs for the
9470 target address, under the above mentioned scenarios, the
9471 div would get wrong results.
9474 j r2 # to next page mapped or unmapped
9475 div r8,r9 # this bug would be there as long
9476 # as there is an ICache miss and
9477 nop # the "data pattern" is present
9480 beq r0, r0, NextPage # to Next page
9481 div r8,r9
9484 This bug is present for div, divu, ddiv, and ddivu
9485 instructions.
9487 Workaround: For item 1), OS could make sure that the next page
9488 after the divide instruction is also mapped. For item 2), the
9489 compiler could make sure that the divide instruction is not in
9490 the branch delay slot."
9492 These processors have PRId values of 0x00004220 and 0x00004300 for
9493 the R4000 and 0x00004400, 0x00004500 and 0x00004600 for the R4400. */
9495 const char *
9496 mips_output_division (const char *division, rtx *operands)
9498 const char *s;
9500 s = division;
9501 if (TARGET_FIX_R4000 || TARGET_FIX_R4400)
9503 output_asm_insn (s, operands);
9504 s = "nop";
9506 if (TARGET_CHECK_ZERO_DIV)
9508 if (TARGET_MIPS16)
9510 output_asm_insn (s, operands);
9511 s = "bnez\t%2,1f\n\tbreak\t7\n1:";
9513 else if (GENERATE_DIVIDE_TRAPS)
9515 output_asm_insn (s, operands);
9516 s = "teq\t%2,%.,7";
9518 else
9520 output_asm_insn ("%(bne\t%2,%.,1f", operands);
9521 output_asm_insn (s, operands);
9522 s = "break\t7%)\n1:";
9525 return s;
9528 /* Return true if INSN is a multiply-add or multiply-subtract
9529 instruction and PREV assigns to the accumulator operand. */
9531 bool
9532 mips_linked_madd_p (rtx prev, rtx insn)
9534 rtx x;
9536 x = single_set (insn);
9537 if (x == 0)
9538 return false;
9540 x = SET_SRC (x);
9542 if (GET_CODE (x) == PLUS
9543 && GET_CODE (XEXP (x, 0)) == MULT
9544 && reg_set_p (XEXP (x, 1), prev))
9545 return true;
9547 if (GET_CODE (x) == MINUS
9548 && GET_CODE (XEXP (x, 1)) == MULT
9549 && reg_set_p (XEXP (x, 0), prev))
9550 return true;
9552 return false;
9555 /* Implements a store data bypass check. We need this because the cprestore
9556 pattern is type store, but defined using an UNSPEC. This UNSPEC causes the
9557 default routine to abort. We just return false for that case. */
9558 /* ??? Should try to give a better result here than assuming false. */
9561 mips_store_data_bypass_p (rtx out_insn, rtx in_insn)
9563 if (GET_CODE (PATTERN (in_insn)) == UNSPEC_VOLATILE)
9564 return false;
9566 return ! store_data_bypass_p (out_insn, in_insn);
9569 /* Implement TARGET_SCHED_ADJUST_COST. We assume that anti and output
9570 dependencies have no cost, except on the 20Kc where output-dependence
9571 is treated like input-dependence. */
9573 static int
9574 mips_adjust_cost (rtx insn ATTRIBUTE_UNUSED, rtx link,
9575 rtx dep ATTRIBUTE_UNUSED, int cost)
9577 if (REG_NOTE_KIND (link) == REG_DEP_OUTPUT
9578 && TUNE_20KC)
9579 return cost;
9580 if (REG_NOTE_KIND (link) != 0)
9581 return 0;
9582 return cost;
9585 /* Return the number of instructions that can be issued per cycle. */
9587 static int
9588 mips_issue_rate (void)
9590 switch (mips_tune)
9592 case PROCESSOR_74KC:
9593 case PROCESSOR_74KF2_1:
9594 case PROCESSOR_74KF1_1:
9595 case PROCESSOR_74KF3_2:
9596 /* The 74k is not strictly quad-issue cpu, but can be seen as one
9597 by the scheduler. It can issue 1 ALU, 1 AGEN and 2 FPU insns,
9598 but in reality only a maximum of 3 insns can be issued as the
9599 floating point load/stores also require a slot in the AGEN pipe. */
9600 return 4;
9602 case PROCESSOR_20KC:
9603 case PROCESSOR_R4130:
9604 case PROCESSOR_R5400:
9605 case PROCESSOR_R5500:
9606 case PROCESSOR_R7000:
9607 case PROCESSOR_R9000:
9608 return 2;
9610 case PROCESSOR_SB1:
9611 case PROCESSOR_SB1A:
9612 /* This is actually 4, but we get better performance if we claim 3.
9613 This is partly because of unwanted speculative code motion with the
9614 larger number, and partly because in most common cases we can't
9615 reach the theoretical max of 4. */
9616 return 3;
9618 default:
9619 return 1;
9623 /* Implements TARGET_SCHED_FIRST_CYCLE_MULTIPASS_DFA_LOOKAHEAD. This should
9624 be as wide as the scheduling freedom in the DFA. */
9626 static int
9627 mips_multipass_dfa_lookahead (void)
9629 /* Can schedule up to 4 of the 6 function units in any one cycle. */
9630 if (TUNE_SB1)
9631 return 4;
9633 return 0;
9636 /* Remove the instruction at index LOWER from ready queue READY and
9637 reinsert it in front of the instruction at index HIGHER. LOWER must
9638 be <= HIGHER. */
9640 static void
9641 mips_promote_ready (rtx *ready, int lower, int higher)
9643 rtx new_head;
9644 int i;
9646 new_head = ready[lower];
9647 for (i = lower; i < higher; i++)
9648 ready[i] = ready[i + 1];
9649 ready[i] = new_head;
9652 /* If the priority of the instruction at POS2 in the ready queue READY
9653 is within LIMIT units of that of the instruction at POS1, swap the
9654 instructions if POS2 is not already less than POS1. */
9656 static void
9657 mips_maybe_swap_ready (rtx *ready, int pos1, int pos2, int limit)
9659 if (pos1 < pos2
9660 && INSN_PRIORITY (ready[pos1]) + limit >= INSN_PRIORITY (ready[pos2]))
9662 rtx temp;
9663 temp = ready[pos1];
9664 ready[pos1] = ready[pos2];
9665 ready[pos2] = temp;
9669 /* Used by TUNE_MACC_CHAINS to record the last scheduled instruction
9670 that may clobber hi or lo. */
9672 static rtx mips_macc_chains_last_hilo;
9674 /* A TUNE_MACC_CHAINS helper function. Record that instruction INSN has
9675 been scheduled, updating mips_macc_chains_last_hilo appropriately. */
9677 static void
9678 mips_macc_chains_record (rtx insn)
9680 if (get_attr_may_clobber_hilo (insn))
9681 mips_macc_chains_last_hilo = insn;
9684 /* A TUNE_MACC_CHAINS helper function. Search ready queue READY, which
9685 has NREADY elements, looking for a multiply-add or multiply-subtract
9686 instruction that is cumulative with mips_macc_chains_last_hilo.
9687 If there is one, promote it ahead of anything else that might
9688 clobber hi or lo. */
9690 static void
9691 mips_macc_chains_reorder (rtx *ready, int nready)
9693 int i, j;
9695 if (mips_macc_chains_last_hilo != 0)
9696 for (i = nready - 1; i >= 0; i--)
9697 if (mips_linked_madd_p (mips_macc_chains_last_hilo, ready[i]))
9699 for (j = nready - 1; j > i; j--)
9700 if (recog_memoized (ready[j]) >= 0
9701 && get_attr_may_clobber_hilo (ready[j]))
9703 mips_promote_ready (ready, i, j);
9704 break;
9706 break;
9710 /* The last instruction to be scheduled. */
9712 static rtx vr4130_last_insn;
9714 /* A note_stores callback used by vr4130_true_reg_dependence_p. DATA
9715 points to an rtx that is initially an instruction. Nullify the rtx
9716 if the instruction uses the value of register X. */
9718 static void
9719 vr4130_true_reg_dependence_p_1 (rtx x, const_rtx pat ATTRIBUTE_UNUSED, void *data)
9721 rtx *insn_ptr = data;
9722 if (REG_P (x)
9723 && *insn_ptr != 0
9724 && reg_referenced_p (x, PATTERN (*insn_ptr)))
9725 *insn_ptr = 0;
9728 /* Return true if there is true register dependence between vr4130_last_insn
9729 and INSN. */
9731 static bool
9732 vr4130_true_reg_dependence_p (rtx insn)
9734 note_stores (PATTERN (vr4130_last_insn),
9735 vr4130_true_reg_dependence_p_1, &insn);
9736 return insn == 0;
9739 /* A TUNE_MIPS4130 helper function. Given that INSN1 is at the head of
9740 the ready queue and that INSN2 is the instruction after it, return
9741 true if it is worth promoting INSN2 ahead of INSN1. Look for cases
9742 in which INSN1 and INSN2 can probably issue in parallel, but for
9743 which (INSN2, INSN1) should be less sensitive to instruction
9744 alignment than (INSN1, INSN2). See 4130.md for more details. */
9746 static bool
9747 vr4130_swap_insns_p (rtx insn1, rtx insn2)
9749 sd_iterator_def sd_it;
9750 dep_t dep;
9752 /* Check for the following case:
9754 1) there is some other instruction X with an anti dependence on INSN1;
9755 2) X has a higher priority than INSN2; and
9756 3) X is an arithmetic instruction (and thus has no unit restrictions).
9758 If INSN1 is the last instruction blocking X, it would better to
9759 choose (INSN1, X) over (INSN2, INSN1). */
9760 FOR_EACH_DEP (insn1, SD_LIST_FORW, sd_it, dep)
9761 if (DEP_TYPE (dep) == REG_DEP_ANTI
9762 && INSN_PRIORITY (DEP_CON (dep)) > INSN_PRIORITY (insn2)
9763 && recog_memoized (DEP_CON (dep)) >= 0
9764 && get_attr_vr4130_class (DEP_CON (dep)) == VR4130_CLASS_ALU)
9765 return false;
9767 if (vr4130_last_insn != 0
9768 && recog_memoized (insn1) >= 0
9769 && recog_memoized (insn2) >= 0)
9771 /* See whether INSN1 and INSN2 use different execution units,
9772 or if they are both ALU-type instructions. If so, they can
9773 probably execute in parallel. */
9774 enum attr_vr4130_class class1 = get_attr_vr4130_class (insn1);
9775 enum attr_vr4130_class class2 = get_attr_vr4130_class (insn2);
9776 if (class1 != class2 || class1 == VR4130_CLASS_ALU)
9778 /* If only one of the instructions has a dependence on
9779 vr4130_last_insn, prefer to schedule the other one first. */
9780 bool dep1 = vr4130_true_reg_dependence_p (insn1);
9781 bool dep2 = vr4130_true_reg_dependence_p (insn2);
9782 if (dep1 != dep2)
9783 return dep1;
9785 /* Prefer to schedule INSN2 ahead of INSN1 if vr4130_last_insn
9786 is not an ALU-type instruction and if INSN1 uses the same
9787 execution unit. (Note that if this condition holds, we already
9788 know that INSN2 uses a different execution unit.) */
9789 if (class1 != VR4130_CLASS_ALU
9790 && recog_memoized (vr4130_last_insn) >= 0
9791 && class1 == get_attr_vr4130_class (vr4130_last_insn))
9792 return true;
9795 return false;
9798 /* A TUNE_MIPS4130 helper function. (READY, NREADY) describes a ready
9799 queue with at least two instructions. Swap the first two if
9800 vr4130_swap_insns_p says that it could be worthwhile. */
9802 static void
9803 vr4130_reorder (rtx *ready, int nready)
9805 if (vr4130_swap_insns_p (ready[nready - 1], ready[nready - 2]))
9806 mips_promote_ready (ready, nready - 2, nready - 1);
9809 /* Record whether last 74k AGEN instruction was a load or store. */
9811 static enum attr_type mips_last_74k_agen_insn = TYPE_UNKNOWN;
9813 /* Initialize mips_last_74k_agen_insn from INSN. A null argument
9814 resets to TYPE_UNKNOWN state. */
9816 static void
9817 mips_74k_agen_init (rtx insn)
9819 if (!insn || !NONJUMP_INSN_P (insn))
9820 mips_last_74k_agen_insn = TYPE_UNKNOWN;
9821 else if (USEFUL_INSN_P (insn))
9823 enum attr_type type = get_attr_type (insn);
9824 if (type == TYPE_LOAD || type == TYPE_STORE)
9825 mips_last_74k_agen_insn = type;
9829 /* A TUNE_74K helper function. The 74K AGEN pipeline likes multiple
9830 loads to be grouped together, and multiple stores to be grouped
9831 together. Swap things around in the ready queue to make this happen. */
9833 static void
9834 mips_74k_agen_reorder (rtx *ready, int nready)
9836 int i;
9837 int store_pos, load_pos;
9839 store_pos = -1;
9840 load_pos = -1;
9842 for (i = nready - 1; i >= 0; i--)
9844 rtx insn = ready[i];
9845 if (USEFUL_INSN_P (insn))
9846 switch (get_attr_type (insn))
9848 case TYPE_STORE:
9849 if (store_pos == -1)
9850 store_pos = i;
9851 break;
9853 case TYPE_LOAD:
9854 if (load_pos == -1)
9855 load_pos = i;
9856 break;
9858 default:
9859 break;
9863 if (load_pos == -1 || store_pos == -1)
9864 return;
9866 switch (mips_last_74k_agen_insn)
9868 case TYPE_UNKNOWN:
9869 /* Prefer to schedule loads since they have a higher latency. */
9870 case TYPE_LOAD:
9871 /* Swap loads to the front of the queue. */
9872 mips_maybe_swap_ready (ready, load_pos, store_pos, 4);
9873 break;
9874 case TYPE_STORE:
9875 /* Swap stores to the front of the queue. */
9876 mips_maybe_swap_ready (ready, store_pos, load_pos, 4);
9877 break;
9878 default:
9879 break;
9883 /* Implement TARGET_SCHED_INIT. */
9885 static void
9886 mips_sched_init (FILE *file ATTRIBUTE_UNUSED, int verbose ATTRIBUTE_UNUSED,
9887 int max_ready ATTRIBUTE_UNUSED)
9889 mips_macc_chains_last_hilo = 0;
9890 vr4130_last_insn = 0;
9891 mips_74k_agen_init (NULL_RTX);
9894 /* Implement TARGET_SCHED_REORDER and TARG_SCHED_REORDER2. */
9896 static int
9897 mips_sched_reorder (FILE *file ATTRIBUTE_UNUSED, int verbose ATTRIBUTE_UNUSED,
9898 rtx *ready, int *nreadyp, int cycle ATTRIBUTE_UNUSED)
9900 if (!reload_completed
9901 && TUNE_MACC_CHAINS
9902 && *nreadyp > 0)
9903 mips_macc_chains_reorder (ready, *nreadyp);
9904 if (reload_completed
9905 && TUNE_MIPS4130
9906 && !TARGET_VR4130_ALIGN
9907 && *nreadyp > 1)
9908 vr4130_reorder (ready, *nreadyp);
9909 if (TUNE_74K)
9910 mips_74k_agen_reorder (ready, *nreadyp);
9911 return mips_issue_rate ();
9914 /* Implement TARGET_SCHED_VARIABLE_ISSUE. */
9916 static int
9917 mips_variable_issue (FILE *file ATTRIBUTE_UNUSED, int verbose ATTRIBUTE_UNUSED,
9918 rtx insn, int more)
9920 if (TUNE_74K)
9921 mips_74k_agen_init (insn);
9922 switch (GET_CODE (PATTERN (insn)))
9924 case USE:
9925 case CLOBBER:
9926 /* Don't count USEs and CLOBBERs against the issue rate. */
9927 break;
9929 default:
9930 more--;
9931 if (!reload_completed && TUNE_MACC_CHAINS)
9932 mips_macc_chains_record (insn);
9933 vr4130_last_insn = insn;
9934 break;
9936 return more;
9939 /* Given that we have an rtx of the form (prefetch ... WRITE LOCALITY),
9940 return the first operand of the associated "pref" or "prefx" insn. */
9943 mips_prefetch_cookie (rtx write, rtx locality)
9945 /* store_streamed / load_streamed. */
9946 if (INTVAL (locality) <= 0)
9947 return GEN_INT (INTVAL (write) + 4);
9949 /* store / load. */
9950 if (INTVAL (locality) <= 2)
9951 return write;
9953 /* store_retained / load_retained. */
9954 return GEN_INT (INTVAL (write) + 6);
9957 /* MIPS builtin function support. */
9959 struct builtin_description
9961 /* The code of the main .md file instruction. See mips_builtin_type
9962 for more information. */
9963 enum insn_code icode;
9965 /* The floating-point comparison code to use with ICODE, if any. */
9966 enum mips_fp_condition cond;
9968 /* The name of the builtin function. */
9969 const char *name;
9971 /* Specifies how the function should be expanded. */
9972 enum mips_builtin_type builtin_type;
9974 /* The function's prototype. */
9975 enum mips_function_type function_type;
9977 /* The target flags required for this function. */
9978 int target_flags;
9981 /* Define a MIPS_BUILTIN_DIRECT function for instruction CODE_FOR_mips_<INSN>.
9982 FUNCTION_TYPE and TARGET_FLAGS are builtin_description fields. */
9983 #define DIRECT_BUILTIN(INSN, FUNCTION_TYPE, TARGET_FLAGS) \
9984 { CODE_FOR_mips_ ## INSN, 0, "__builtin_mips_" #INSN, \
9985 MIPS_BUILTIN_DIRECT, FUNCTION_TYPE, TARGET_FLAGS }
9987 /* Define __builtin_mips_<INSN>_<COND>_{s,d}, both of which require
9988 TARGET_FLAGS. */
9989 #define CMP_SCALAR_BUILTINS(INSN, COND, TARGET_FLAGS) \
9990 { CODE_FOR_mips_ ## INSN ## _cond_s, MIPS_FP_COND_ ## COND, \
9991 "__builtin_mips_" #INSN "_" #COND "_s", \
9992 MIPS_BUILTIN_CMP_SINGLE, MIPS_INT_FTYPE_SF_SF, TARGET_FLAGS }, \
9993 { CODE_FOR_mips_ ## INSN ## _cond_d, MIPS_FP_COND_ ## COND, \
9994 "__builtin_mips_" #INSN "_" #COND "_d", \
9995 MIPS_BUILTIN_CMP_SINGLE, MIPS_INT_FTYPE_DF_DF, TARGET_FLAGS }
9997 /* Define __builtin_mips_{any,all,upper,lower}_<INSN>_<COND>_ps.
9998 The lower and upper forms require TARGET_FLAGS while the any and all
9999 forms require MASK_MIPS3D. */
10000 #define CMP_PS_BUILTINS(INSN, COND, TARGET_FLAGS) \
10001 { CODE_FOR_mips_ ## INSN ## _cond_ps, MIPS_FP_COND_ ## COND, \
10002 "__builtin_mips_any_" #INSN "_" #COND "_ps", \
10003 MIPS_BUILTIN_CMP_ANY, MIPS_INT_FTYPE_V2SF_V2SF, MASK_MIPS3D }, \
10004 { CODE_FOR_mips_ ## INSN ## _cond_ps, MIPS_FP_COND_ ## COND, \
10005 "__builtin_mips_all_" #INSN "_" #COND "_ps", \
10006 MIPS_BUILTIN_CMP_ALL, MIPS_INT_FTYPE_V2SF_V2SF, MASK_MIPS3D }, \
10007 { CODE_FOR_mips_ ## INSN ## _cond_ps, MIPS_FP_COND_ ## COND, \
10008 "__builtin_mips_lower_" #INSN "_" #COND "_ps", \
10009 MIPS_BUILTIN_CMP_LOWER, MIPS_INT_FTYPE_V2SF_V2SF, TARGET_FLAGS }, \
10010 { CODE_FOR_mips_ ## INSN ## _cond_ps, MIPS_FP_COND_ ## COND, \
10011 "__builtin_mips_upper_" #INSN "_" #COND "_ps", \
10012 MIPS_BUILTIN_CMP_UPPER, MIPS_INT_FTYPE_V2SF_V2SF, TARGET_FLAGS }
10014 /* Define __builtin_mips_{any,all}_<INSN>_<COND>_4s. The functions
10015 require MASK_MIPS3D. */
10016 #define CMP_4S_BUILTINS(INSN, COND) \
10017 { CODE_FOR_mips_ ## INSN ## _cond_4s, MIPS_FP_COND_ ## COND, \
10018 "__builtin_mips_any_" #INSN "_" #COND "_4s", \
10019 MIPS_BUILTIN_CMP_ANY, MIPS_INT_FTYPE_V2SF_V2SF_V2SF_V2SF, \
10020 MASK_MIPS3D }, \
10021 { CODE_FOR_mips_ ## INSN ## _cond_4s, MIPS_FP_COND_ ## COND, \
10022 "__builtin_mips_all_" #INSN "_" #COND "_4s", \
10023 MIPS_BUILTIN_CMP_ALL, MIPS_INT_FTYPE_V2SF_V2SF_V2SF_V2SF, \
10024 MASK_MIPS3D }
10026 /* Define __builtin_mips_mov{t,f}_<INSN>_<COND>_ps. The comparison
10027 instruction requires TARGET_FLAGS. */
10028 #define MOVTF_BUILTINS(INSN, COND, TARGET_FLAGS) \
10029 { CODE_FOR_mips_ ## INSN ## _cond_ps, MIPS_FP_COND_ ## COND, \
10030 "__builtin_mips_movt_" #INSN "_" #COND "_ps", \
10031 MIPS_BUILTIN_MOVT, MIPS_V2SF_FTYPE_V2SF_V2SF_V2SF_V2SF, \
10032 TARGET_FLAGS }, \
10033 { CODE_FOR_mips_ ## INSN ## _cond_ps, MIPS_FP_COND_ ## COND, \
10034 "__builtin_mips_movf_" #INSN "_" #COND "_ps", \
10035 MIPS_BUILTIN_MOVF, MIPS_V2SF_FTYPE_V2SF_V2SF_V2SF_V2SF, \
10036 TARGET_FLAGS }
10038 /* Define all the builtins related to c.cond.fmt condition COND. */
10039 #define CMP_BUILTINS(COND) \
10040 MOVTF_BUILTINS (c, COND, MASK_PAIRED_SINGLE_FLOAT), \
10041 MOVTF_BUILTINS (cabs, COND, MASK_MIPS3D), \
10042 CMP_SCALAR_BUILTINS (cabs, COND, MASK_MIPS3D), \
10043 CMP_PS_BUILTINS (c, COND, MASK_PAIRED_SINGLE_FLOAT), \
10044 CMP_PS_BUILTINS (cabs, COND, MASK_MIPS3D), \
10045 CMP_4S_BUILTINS (c, COND), \
10046 CMP_4S_BUILTINS (cabs, COND)
10048 static const struct builtin_description mips_bdesc[] =
10050 DIRECT_BUILTIN (pll_ps, MIPS_V2SF_FTYPE_V2SF_V2SF, MASK_PAIRED_SINGLE_FLOAT),
10051 DIRECT_BUILTIN (pul_ps, MIPS_V2SF_FTYPE_V2SF_V2SF, MASK_PAIRED_SINGLE_FLOAT),
10052 DIRECT_BUILTIN (plu_ps, MIPS_V2SF_FTYPE_V2SF_V2SF, MASK_PAIRED_SINGLE_FLOAT),
10053 DIRECT_BUILTIN (puu_ps, MIPS_V2SF_FTYPE_V2SF_V2SF, MASK_PAIRED_SINGLE_FLOAT),
10054 DIRECT_BUILTIN (cvt_ps_s, MIPS_V2SF_FTYPE_SF_SF, MASK_PAIRED_SINGLE_FLOAT),
10055 DIRECT_BUILTIN (cvt_s_pl, MIPS_SF_FTYPE_V2SF, MASK_PAIRED_SINGLE_FLOAT),
10056 DIRECT_BUILTIN (cvt_s_pu, MIPS_SF_FTYPE_V2SF, MASK_PAIRED_SINGLE_FLOAT),
10057 DIRECT_BUILTIN (abs_ps, MIPS_V2SF_FTYPE_V2SF, MASK_PAIRED_SINGLE_FLOAT),
10059 DIRECT_BUILTIN (alnv_ps, MIPS_V2SF_FTYPE_V2SF_V2SF_INT,
10060 MASK_PAIRED_SINGLE_FLOAT),
10061 DIRECT_BUILTIN (addr_ps, MIPS_V2SF_FTYPE_V2SF_V2SF, MASK_MIPS3D),
10062 DIRECT_BUILTIN (mulr_ps, MIPS_V2SF_FTYPE_V2SF_V2SF, MASK_MIPS3D),
10063 DIRECT_BUILTIN (cvt_pw_ps, MIPS_V2SF_FTYPE_V2SF, MASK_MIPS3D),
10064 DIRECT_BUILTIN (cvt_ps_pw, MIPS_V2SF_FTYPE_V2SF, MASK_MIPS3D),
10066 DIRECT_BUILTIN (recip1_s, MIPS_SF_FTYPE_SF, MASK_MIPS3D),
10067 DIRECT_BUILTIN (recip1_d, MIPS_DF_FTYPE_DF, MASK_MIPS3D),
10068 DIRECT_BUILTIN (recip1_ps, MIPS_V2SF_FTYPE_V2SF, MASK_MIPS3D),
10069 DIRECT_BUILTIN (recip2_s, MIPS_SF_FTYPE_SF_SF, MASK_MIPS3D),
10070 DIRECT_BUILTIN (recip2_d, MIPS_DF_FTYPE_DF_DF, MASK_MIPS3D),
10071 DIRECT_BUILTIN (recip2_ps, MIPS_V2SF_FTYPE_V2SF_V2SF, MASK_MIPS3D),
10073 DIRECT_BUILTIN (rsqrt1_s, MIPS_SF_FTYPE_SF, MASK_MIPS3D),
10074 DIRECT_BUILTIN (rsqrt1_d, MIPS_DF_FTYPE_DF, MASK_MIPS3D),
10075 DIRECT_BUILTIN (rsqrt1_ps, MIPS_V2SF_FTYPE_V2SF, MASK_MIPS3D),
10076 DIRECT_BUILTIN (rsqrt2_s, MIPS_SF_FTYPE_SF_SF, MASK_MIPS3D),
10077 DIRECT_BUILTIN (rsqrt2_d, MIPS_DF_FTYPE_DF_DF, MASK_MIPS3D),
10078 DIRECT_BUILTIN (rsqrt2_ps, MIPS_V2SF_FTYPE_V2SF_V2SF, MASK_MIPS3D),
10080 MIPS_FP_CONDITIONS (CMP_BUILTINS)
10083 /* Builtin functions for the SB-1 processor. */
10085 #define CODE_FOR_mips_sqrt_ps CODE_FOR_sqrtv2sf2
10087 static const struct builtin_description sb1_bdesc[] =
10089 DIRECT_BUILTIN (sqrt_ps, MIPS_V2SF_FTYPE_V2SF, MASK_PAIRED_SINGLE_FLOAT)
10092 /* Builtin functions for DSP ASE. */
10094 #define CODE_FOR_mips_addq_ph CODE_FOR_addv2hi3
10095 #define CODE_FOR_mips_addu_qb CODE_FOR_addv4qi3
10096 #define CODE_FOR_mips_subq_ph CODE_FOR_subv2hi3
10097 #define CODE_FOR_mips_subu_qb CODE_FOR_subv4qi3
10098 #define CODE_FOR_mips_mul_ph CODE_FOR_mulv2hi3
10100 /* Define a MIPS_BUILTIN_DIRECT_NO_TARGET function for instruction
10101 CODE_FOR_mips_<INSN>. FUNCTION_TYPE and TARGET_FLAGS are
10102 builtin_description fields. */
10103 #define DIRECT_NO_TARGET_BUILTIN(INSN, FUNCTION_TYPE, TARGET_FLAGS) \
10104 { CODE_FOR_mips_ ## INSN, 0, "__builtin_mips_" #INSN, \
10105 MIPS_BUILTIN_DIRECT_NO_TARGET, FUNCTION_TYPE, TARGET_FLAGS }
10107 /* Define __builtin_mips_bposge<VALUE>. <VALUE> is 32 for the MIPS32 DSP
10108 branch instruction. TARGET_FLAGS is a builtin_description field. */
10109 #define BPOSGE_BUILTIN(VALUE, TARGET_FLAGS) \
10110 { CODE_FOR_mips_bposge, 0, "__builtin_mips_bposge" #VALUE, \
10111 MIPS_BUILTIN_BPOSGE ## VALUE, MIPS_SI_FTYPE_VOID, TARGET_FLAGS }
10113 static const struct builtin_description dsp_bdesc[] =
10115 DIRECT_BUILTIN (addq_ph, MIPS_V2HI_FTYPE_V2HI_V2HI, MASK_DSP),
10116 DIRECT_BUILTIN (addq_s_ph, MIPS_V2HI_FTYPE_V2HI_V2HI, MASK_DSP),
10117 DIRECT_BUILTIN (addq_s_w, MIPS_SI_FTYPE_SI_SI, MASK_DSP),
10118 DIRECT_BUILTIN (addu_qb, MIPS_V4QI_FTYPE_V4QI_V4QI, MASK_DSP),
10119 DIRECT_BUILTIN (addu_s_qb, MIPS_V4QI_FTYPE_V4QI_V4QI, MASK_DSP),
10120 DIRECT_BUILTIN (subq_ph, MIPS_V2HI_FTYPE_V2HI_V2HI, MASK_DSP),
10121 DIRECT_BUILTIN (subq_s_ph, MIPS_V2HI_FTYPE_V2HI_V2HI, MASK_DSP),
10122 DIRECT_BUILTIN (subq_s_w, MIPS_SI_FTYPE_SI_SI, MASK_DSP),
10123 DIRECT_BUILTIN (subu_qb, MIPS_V4QI_FTYPE_V4QI_V4QI, MASK_DSP),
10124 DIRECT_BUILTIN (subu_s_qb, MIPS_V4QI_FTYPE_V4QI_V4QI, MASK_DSP),
10125 DIRECT_BUILTIN (addsc, MIPS_SI_FTYPE_SI_SI, MASK_DSP),
10126 DIRECT_BUILTIN (addwc, MIPS_SI_FTYPE_SI_SI, MASK_DSP),
10127 DIRECT_BUILTIN (modsub, MIPS_SI_FTYPE_SI_SI, MASK_DSP),
10128 DIRECT_BUILTIN (raddu_w_qb, MIPS_SI_FTYPE_V4QI, MASK_DSP),
10129 DIRECT_BUILTIN (absq_s_ph, MIPS_V2HI_FTYPE_V2HI, MASK_DSP),
10130 DIRECT_BUILTIN (absq_s_w, MIPS_SI_FTYPE_SI, MASK_DSP),
10131 DIRECT_BUILTIN (precrq_qb_ph, MIPS_V4QI_FTYPE_V2HI_V2HI, MASK_DSP),
10132 DIRECT_BUILTIN (precrq_ph_w, MIPS_V2HI_FTYPE_SI_SI, MASK_DSP),
10133 DIRECT_BUILTIN (precrq_rs_ph_w, MIPS_V2HI_FTYPE_SI_SI, MASK_DSP),
10134 DIRECT_BUILTIN (precrqu_s_qb_ph, MIPS_V4QI_FTYPE_V2HI_V2HI, MASK_DSP),
10135 DIRECT_BUILTIN (preceq_w_phl, MIPS_SI_FTYPE_V2HI, MASK_DSP),
10136 DIRECT_BUILTIN (preceq_w_phr, MIPS_SI_FTYPE_V2HI, MASK_DSP),
10137 DIRECT_BUILTIN (precequ_ph_qbl, MIPS_V2HI_FTYPE_V4QI, MASK_DSP),
10138 DIRECT_BUILTIN (precequ_ph_qbr, MIPS_V2HI_FTYPE_V4QI, MASK_DSP),
10139 DIRECT_BUILTIN (precequ_ph_qbla, MIPS_V2HI_FTYPE_V4QI, MASK_DSP),
10140 DIRECT_BUILTIN (precequ_ph_qbra, MIPS_V2HI_FTYPE_V4QI, MASK_DSP),
10141 DIRECT_BUILTIN (preceu_ph_qbl, MIPS_V2HI_FTYPE_V4QI, MASK_DSP),
10142 DIRECT_BUILTIN (preceu_ph_qbr, MIPS_V2HI_FTYPE_V4QI, MASK_DSP),
10143 DIRECT_BUILTIN (preceu_ph_qbla, MIPS_V2HI_FTYPE_V4QI, MASK_DSP),
10144 DIRECT_BUILTIN (preceu_ph_qbra, MIPS_V2HI_FTYPE_V4QI, MASK_DSP),
10145 DIRECT_BUILTIN (shll_qb, MIPS_V4QI_FTYPE_V4QI_SI, MASK_DSP),
10146 DIRECT_BUILTIN (shll_ph, MIPS_V2HI_FTYPE_V2HI_SI, MASK_DSP),
10147 DIRECT_BUILTIN (shll_s_ph, MIPS_V2HI_FTYPE_V2HI_SI, MASK_DSP),
10148 DIRECT_BUILTIN (shll_s_w, MIPS_SI_FTYPE_SI_SI, MASK_DSP),
10149 DIRECT_BUILTIN (shrl_qb, MIPS_V4QI_FTYPE_V4QI_SI, MASK_DSP),
10150 DIRECT_BUILTIN (shra_ph, MIPS_V2HI_FTYPE_V2HI_SI, MASK_DSP),
10151 DIRECT_BUILTIN (shra_r_ph, MIPS_V2HI_FTYPE_V2HI_SI, MASK_DSP),
10152 DIRECT_BUILTIN (shra_r_w, MIPS_SI_FTYPE_SI_SI, MASK_DSP),
10153 DIRECT_BUILTIN (muleu_s_ph_qbl, MIPS_V2HI_FTYPE_V4QI_V2HI, MASK_DSP),
10154 DIRECT_BUILTIN (muleu_s_ph_qbr, MIPS_V2HI_FTYPE_V4QI_V2HI, MASK_DSP),
10155 DIRECT_BUILTIN (mulq_rs_ph, MIPS_V2HI_FTYPE_V2HI_V2HI, MASK_DSP),
10156 DIRECT_BUILTIN (muleq_s_w_phl, MIPS_SI_FTYPE_V2HI_V2HI, MASK_DSP),
10157 DIRECT_BUILTIN (muleq_s_w_phr, MIPS_SI_FTYPE_V2HI_V2HI, MASK_DSP),
10158 DIRECT_BUILTIN (bitrev, MIPS_SI_FTYPE_SI, MASK_DSP),
10159 DIRECT_BUILTIN (insv, MIPS_SI_FTYPE_SI_SI, MASK_DSP),
10160 DIRECT_BUILTIN (repl_qb, MIPS_V4QI_FTYPE_SI, MASK_DSP),
10161 DIRECT_BUILTIN (repl_ph, MIPS_V2HI_FTYPE_SI, MASK_DSP),
10162 DIRECT_NO_TARGET_BUILTIN (cmpu_eq_qb, MIPS_VOID_FTYPE_V4QI_V4QI, MASK_DSP),
10163 DIRECT_NO_TARGET_BUILTIN (cmpu_lt_qb, MIPS_VOID_FTYPE_V4QI_V4QI, MASK_DSP),
10164 DIRECT_NO_TARGET_BUILTIN (cmpu_le_qb, MIPS_VOID_FTYPE_V4QI_V4QI, MASK_DSP),
10165 DIRECT_BUILTIN (cmpgu_eq_qb, MIPS_SI_FTYPE_V4QI_V4QI, MASK_DSP),
10166 DIRECT_BUILTIN (cmpgu_lt_qb, MIPS_SI_FTYPE_V4QI_V4QI, MASK_DSP),
10167 DIRECT_BUILTIN (cmpgu_le_qb, MIPS_SI_FTYPE_V4QI_V4QI, MASK_DSP),
10168 DIRECT_NO_TARGET_BUILTIN (cmp_eq_ph, MIPS_VOID_FTYPE_V2HI_V2HI, MASK_DSP),
10169 DIRECT_NO_TARGET_BUILTIN (cmp_lt_ph, MIPS_VOID_FTYPE_V2HI_V2HI, MASK_DSP),
10170 DIRECT_NO_TARGET_BUILTIN (cmp_le_ph, MIPS_VOID_FTYPE_V2HI_V2HI, MASK_DSP),
10171 DIRECT_BUILTIN (pick_qb, MIPS_V4QI_FTYPE_V4QI_V4QI, MASK_DSP),
10172 DIRECT_BUILTIN (pick_ph, MIPS_V2HI_FTYPE_V2HI_V2HI, MASK_DSP),
10173 DIRECT_BUILTIN (packrl_ph, MIPS_V2HI_FTYPE_V2HI_V2HI, MASK_DSP),
10174 DIRECT_NO_TARGET_BUILTIN (wrdsp, MIPS_VOID_FTYPE_SI_SI, MASK_DSP),
10175 DIRECT_BUILTIN (rddsp, MIPS_SI_FTYPE_SI, MASK_DSP),
10176 DIRECT_BUILTIN (lbux, MIPS_SI_FTYPE_POINTER_SI, MASK_DSP),
10177 DIRECT_BUILTIN (lhx, MIPS_SI_FTYPE_POINTER_SI, MASK_DSP),
10178 DIRECT_BUILTIN (lwx, MIPS_SI_FTYPE_POINTER_SI, MASK_DSP),
10179 BPOSGE_BUILTIN (32, MASK_DSP),
10181 /* The following are for the MIPS DSP ASE REV 2. */
10182 DIRECT_BUILTIN (absq_s_qb, MIPS_V4QI_FTYPE_V4QI, MASK_DSPR2),
10183 DIRECT_BUILTIN (addu_ph, MIPS_V2HI_FTYPE_V2HI_V2HI, MASK_DSPR2),
10184 DIRECT_BUILTIN (addu_s_ph, MIPS_V2HI_FTYPE_V2HI_V2HI, MASK_DSPR2),
10185 DIRECT_BUILTIN (adduh_qb, MIPS_V4QI_FTYPE_V4QI_V4QI, MASK_DSPR2),
10186 DIRECT_BUILTIN (adduh_r_qb, MIPS_V4QI_FTYPE_V4QI_V4QI, MASK_DSPR2),
10187 DIRECT_BUILTIN (append, MIPS_SI_FTYPE_SI_SI_SI, MASK_DSPR2),
10188 DIRECT_BUILTIN (balign, MIPS_SI_FTYPE_SI_SI_SI, MASK_DSPR2),
10189 DIRECT_BUILTIN (cmpgdu_eq_qb, MIPS_SI_FTYPE_V4QI_V4QI, MASK_DSPR2),
10190 DIRECT_BUILTIN (cmpgdu_lt_qb, MIPS_SI_FTYPE_V4QI_V4QI, MASK_DSPR2),
10191 DIRECT_BUILTIN (cmpgdu_le_qb, MIPS_SI_FTYPE_V4QI_V4QI, MASK_DSPR2),
10192 DIRECT_BUILTIN (mul_ph, MIPS_V2HI_FTYPE_V2HI_V2HI, MASK_DSPR2),
10193 DIRECT_BUILTIN (mul_s_ph, MIPS_V2HI_FTYPE_V2HI_V2HI, MASK_DSPR2),
10194 DIRECT_BUILTIN (mulq_rs_w, MIPS_SI_FTYPE_SI_SI, MASK_DSPR2),
10195 DIRECT_BUILTIN (mulq_s_ph, MIPS_V2HI_FTYPE_V2HI_V2HI, MASK_DSPR2),
10196 DIRECT_BUILTIN (mulq_s_w, MIPS_SI_FTYPE_SI_SI, MASK_DSPR2),
10197 DIRECT_BUILTIN (precr_qb_ph, MIPS_V4QI_FTYPE_V2HI_V2HI, MASK_DSPR2),
10198 DIRECT_BUILTIN (precr_sra_ph_w, MIPS_V2HI_FTYPE_SI_SI_SI, MASK_DSPR2),
10199 DIRECT_BUILTIN (precr_sra_r_ph_w, MIPS_V2HI_FTYPE_SI_SI_SI, MASK_DSPR2),
10200 DIRECT_BUILTIN (prepend, MIPS_SI_FTYPE_SI_SI_SI, MASK_DSPR2),
10201 DIRECT_BUILTIN (shra_qb, MIPS_V4QI_FTYPE_V4QI_SI, MASK_DSPR2),
10202 DIRECT_BUILTIN (shra_r_qb, MIPS_V4QI_FTYPE_V4QI_SI, MASK_DSPR2),
10203 DIRECT_BUILTIN (shrl_ph, MIPS_V2HI_FTYPE_V2HI_SI, MASK_DSPR2),
10204 DIRECT_BUILTIN (subu_ph, MIPS_V2HI_FTYPE_V2HI_V2HI, MASK_DSPR2),
10205 DIRECT_BUILTIN (subu_s_ph, MIPS_V2HI_FTYPE_V2HI_V2HI, MASK_DSPR2),
10206 DIRECT_BUILTIN (subuh_qb, MIPS_V4QI_FTYPE_V4QI_V4QI, MASK_DSPR2),
10207 DIRECT_BUILTIN (subuh_r_qb, MIPS_V4QI_FTYPE_V4QI_V4QI, MASK_DSPR2),
10208 DIRECT_BUILTIN (addqh_ph, MIPS_V2HI_FTYPE_V2HI_V2HI, MASK_DSPR2),
10209 DIRECT_BUILTIN (addqh_r_ph, MIPS_V2HI_FTYPE_V2HI_V2HI, MASK_DSPR2),
10210 DIRECT_BUILTIN (addqh_w, MIPS_SI_FTYPE_SI_SI, MASK_DSPR2),
10211 DIRECT_BUILTIN (addqh_r_w, MIPS_SI_FTYPE_SI_SI, MASK_DSPR2),
10212 DIRECT_BUILTIN (subqh_ph, MIPS_V2HI_FTYPE_V2HI_V2HI, MASK_DSPR2),
10213 DIRECT_BUILTIN (subqh_r_ph, MIPS_V2HI_FTYPE_V2HI_V2HI, MASK_DSPR2),
10214 DIRECT_BUILTIN (subqh_w, MIPS_SI_FTYPE_SI_SI, MASK_DSPR2),
10215 DIRECT_BUILTIN (subqh_r_w, MIPS_SI_FTYPE_SI_SI, MASK_DSPR2)
10218 static const struct builtin_description dsp_32only_bdesc[] =
10220 DIRECT_BUILTIN (dpau_h_qbl, MIPS_DI_FTYPE_DI_V4QI_V4QI, MASK_DSP),
10221 DIRECT_BUILTIN (dpau_h_qbr, MIPS_DI_FTYPE_DI_V4QI_V4QI, MASK_DSP),
10222 DIRECT_BUILTIN (dpsu_h_qbl, MIPS_DI_FTYPE_DI_V4QI_V4QI, MASK_DSP),
10223 DIRECT_BUILTIN (dpsu_h_qbr, MIPS_DI_FTYPE_DI_V4QI_V4QI, MASK_DSP),
10224 DIRECT_BUILTIN (dpaq_s_w_ph, MIPS_DI_FTYPE_DI_V2HI_V2HI, MASK_DSP),
10225 DIRECT_BUILTIN (dpsq_s_w_ph, MIPS_DI_FTYPE_DI_V2HI_V2HI, MASK_DSP),
10226 DIRECT_BUILTIN (mulsaq_s_w_ph, MIPS_DI_FTYPE_DI_V2HI_V2HI, MASK_DSP),
10227 DIRECT_BUILTIN (dpaq_sa_l_w, MIPS_DI_FTYPE_DI_SI_SI, MASK_DSP),
10228 DIRECT_BUILTIN (dpsq_sa_l_w, MIPS_DI_FTYPE_DI_SI_SI, MASK_DSP),
10229 DIRECT_BUILTIN (maq_s_w_phl, MIPS_DI_FTYPE_DI_V2HI_V2HI, MASK_DSP),
10230 DIRECT_BUILTIN (maq_s_w_phr, MIPS_DI_FTYPE_DI_V2HI_V2HI, MASK_DSP),
10231 DIRECT_BUILTIN (maq_sa_w_phl, MIPS_DI_FTYPE_DI_V2HI_V2HI, MASK_DSP),
10232 DIRECT_BUILTIN (maq_sa_w_phr, MIPS_DI_FTYPE_DI_V2HI_V2HI, MASK_DSP),
10233 DIRECT_BUILTIN (extr_w, MIPS_SI_FTYPE_DI_SI, MASK_DSP),
10234 DIRECT_BUILTIN (extr_r_w, MIPS_SI_FTYPE_DI_SI, MASK_DSP),
10235 DIRECT_BUILTIN (extr_rs_w, MIPS_SI_FTYPE_DI_SI, MASK_DSP),
10236 DIRECT_BUILTIN (extr_s_h, MIPS_SI_FTYPE_DI_SI, MASK_DSP),
10237 DIRECT_BUILTIN (extp, MIPS_SI_FTYPE_DI_SI, MASK_DSP),
10238 DIRECT_BUILTIN (extpdp, MIPS_SI_FTYPE_DI_SI, MASK_DSP),
10239 DIRECT_BUILTIN (shilo, MIPS_DI_FTYPE_DI_SI, MASK_DSP),
10240 DIRECT_BUILTIN (mthlip, MIPS_DI_FTYPE_DI_SI, MASK_DSP),
10242 /* The following are for the MIPS DSP ASE REV 2. */
10243 DIRECT_BUILTIN (dpa_w_ph, MIPS_DI_FTYPE_DI_V2HI_V2HI, MASK_DSPR2),
10244 DIRECT_BUILTIN (dps_w_ph, MIPS_DI_FTYPE_DI_V2HI_V2HI, MASK_DSPR2),
10245 DIRECT_BUILTIN (madd, MIPS_DI_FTYPE_DI_SI_SI, MASK_DSPR2),
10246 DIRECT_BUILTIN (maddu, MIPS_DI_FTYPE_DI_USI_USI, MASK_DSPR2),
10247 DIRECT_BUILTIN (msub, MIPS_DI_FTYPE_DI_SI_SI, MASK_DSPR2),
10248 DIRECT_BUILTIN (msubu, MIPS_DI_FTYPE_DI_USI_USI, MASK_DSPR2),
10249 DIRECT_BUILTIN (mulsa_w_ph, MIPS_DI_FTYPE_DI_V2HI_V2HI, MASK_DSPR2),
10250 DIRECT_BUILTIN (mult, MIPS_DI_FTYPE_SI_SI, MASK_DSPR2),
10251 DIRECT_BUILTIN (multu, MIPS_DI_FTYPE_USI_USI, MASK_DSPR2),
10252 DIRECT_BUILTIN (dpax_w_ph, MIPS_DI_FTYPE_DI_V2HI_V2HI, MASK_DSPR2),
10253 DIRECT_BUILTIN (dpsx_w_ph, MIPS_DI_FTYPE_DI_V2HI_V2HI, MASK_DSPR2),
10254 DIRECT_BUILTIN (dpaqx_s_w_ph, MIPS_DI_FTYPE_DI_V2HI_V2HI, MASK_DSPR2),
10255 DIRECT_BUILTIN (dpaqx_sa_w_ph, MIPS_DI_FTYPE_DI_V2HI_V2HI, MASK_DSPR2),
10256 DIRECT_BUILTIN (dpsqx_s_w_ph, MIPS_DI_FTYPE_DI_V2HI_V2HI, MASK_DSPR2),
10257 DIRECT_BUILTIN (dpsqx_sa_w_ph, MIPS_DI_FTYPE_DI_V2HI_V2HI, MASK_DSPR2)
10260 /* This helps provide a mapping from builtin function codes to bdesc
10261 arrays. */
10263 struct bdesc_map
10265 /* The builtin function table that this entry describes. */
10266 const struct builtin_description *bdesc;
10268 /* The number of entries in the builtin function table. */
10269 unsigned int size;
10271 /* The target processor that supports these builtin functions.
10272 PROCESSOR_MAX means we enable them for all processors. */
10273 enum processor_type proc;
10275 /* If the target has these flags, this builtin function table
10276 will not be supported. */
10277 int unsupported_target_flags;
10280 static const struct bdesc_map bdesc_arrays[] =
10282 { mips_bdesc, ARRAY_SIZE (mips_bdesc), PROCESSOR_MAX, 0 },
10283 { sb1_bdesc, ARRAY_SIZE (sb1_bdesc), PROCESSOR_SB1, 0 },
10284 { dsp_bdesc, ARRAY_SIZE (dsp_bdesc), PROCESSOR_MAX, 0 },
10285 { dsp_32only_bdesc, ARRAY_SIZE (dsp_32only_bdesc), PROCESSOR_MAX,
10286 MASK_64BIT }
10289 /* MODE is a vector mode whose elements have type TYPE. Return the type
10290 of the vector itself. */
10292 static tree
10293 mips_builtin_vector_type (tree type, enum machine_mode mode)
10295 static tree types[(int) MAX_MACHINE_MODE];
10297 if (types[(int) mode] == NULL_TREE)
10298 types[(int) mode] = build_vector_type_for_mode (type, mode);
10299 return types[(int) mode];
10302 /* Source-level argument types. */
10303 #define MIPS_ATYPE_VOID void_type_node
10304 #define MIPS_ATYPE_INT integer_type_node
10305 #define MIPS_ATYPE_POINTER ptr_type_node
10307 /* Standard mode-based argument types. */
10308 #define MIPS_ATYPE_SI intSI_type_node
10309 #define MIPS_ATYPE_USI unsigned_intSI_type_node
10310 #define MIPS_ATYPE_DI intDI_type_node
10311 #define MIPS_ATYPE_SF float_type_node
10312 #define MIPS_ATYPE_DF double_type_node
10314 /* Vector argument types. */
10315 #define MIPS_ATYPE_V2SF mips_builtin_vector_type (float_type_node, V2SFmode)
10316 #define MIPS_ATYPE_V2HI mips_builtin_vector_type (intHI_type_node, V2HImode)
10317 #define MIPS_ATYPE_V4QI mips_builtin_vector_type (intQI_type_node, V4QImode)
10319 /* MIPS_FTYPE_ATYPESN takes N MIPS_FTYPES-like type codes and lists
10320 their associated MIPS_ATYPEs. */
10321 #define MIPS_FTYPE_ATYPES1(A, B) \
10322 MIPS_ATYPE_##A, MIPS_ATYPE_##B
10324 #define MIPS_FTYPE_ATYPES2(A, B, C) \
10325 MIPS_ATYPE_##A, MIPS_ATYPE_##B, MIPS_ATYPE_##C
10327 #define MIPS_FTYPE_ATYPES3(A, B, C, D) \
10328 MIPS_ATYPE_##A, MIPS_ATYPE_##B, MIPS_ATYPE_##C, MIPS_ATYPE_##D
10330 #define MIPS_FTYPE_ATYPES4(A, B, C, D, E) \
10331 MIPS_ATYPE_##A, MIPS_ATYPE_##B, MIPS_ATYPE_##C, MIPS_ATYPE_##D, \
10332 MIPS_ATYPE_##E
10334 /* Return the function type associated with function prototype TYPE. */
10336 static tree
10337 mips_build_function_type (enum mips_function_type type)
10339 static tree types[(int) MIPS_MAX_FTYPE_MAX];
10341 if (types[(int) type] == NULL_TREE)
10342 switch (type)
10344 #define DEF_MIPS_FTYPE(NUM, ARGS) \
10345 case MIPS_FTYPE_NAME##NUM ARGS: \
10346 types[(int) type] \
10347 = build_function_type_list (MIPS_FTYPE_ATYPES##NUM ARGS, \
10348 NULL_TREE); \
10349 break;
10350 #include "config/mips/mips-ftypes.def"
10351 #undef DEF_MIPS_FTYPE
10352 default:
10353 gcc_unreachable ();
10356 return types[(int) type];
10359 /* Init builtin functions. This is called from TARGET_INIT_BUILTIN. */
10361 static void
10362 mips_init_builtins (void)
10364 const struct builtin_description *d;
10365 const struct bdesc_map *m;
10366 unsigned int offset;
10368 /* Iterate through all of the bdesc arrays, initializing all of the
10369 builtin functions. */
10371 offset = 0;
10372 for (m = bdesc_arrays; m < &bdesc_arrays[ARRAY_SIZE (bdesc_arrays)]; m++)
10374 if ((m->proc == PROCESSOR_MAX || (m->proc == mips_arch))
10375 && (m->unsupported_target_flags & target_flags) == 0)
10376 for (d = m->bdesc; d < &m->bdesc[m->size]; d++)
10377 if ((d->target_flags & target_flags) == d->target_flags)
10378 add_builtin_function (d->name,
10379 mips_build_function_type (d->function_type),
10380 d - m->bdesc + offset,
10381 BUILT_IN_MD, NULL, NULL);
10382 offset += m->size;
10386 /* Take the argument ARGNUM of the arglist of EXP and convert it into a form
10387 suitable for input operand OP of instruction ICODE. Return the value. */
10389 static rtx
10390 mips_prepare_builtin_arg (enum insn_code icode,
10391 unsigned int op, tree exp, unsigned int argnum)
10393 rtx value;
10394 enum machine_mode mode;
10396 value = expand_normal (CALL_EXPR_ARG (exp, argnum));
10397 mode = insn_data[icode].operand[op].mode;
10398 if (!insn_data[icode].operand[op].predicate (value, mode))
10400 value = copy_to_mode_reg (mode, value);
10401 /* Check the predicate again. */
10402 if (!insn_data[icode].operand[op].predicate (value, mode))
10404 error ("invalid argument to builtin function");
10405 return const0_rtx;
10409 return value;
10412 /* Return an rtx suitable for output operand OP of instruction ICODE.
10413 If TARGET is non-null, try to use it where possible. */
10415 static rtx
10416 mips_prepare_builtin_target (enum insn_code icode, unsigned int op, rtx target)
10418 enum machine_mode mode;
10420 mode = insn_data[icode].operand[op].mode;
10421 if (target == 0 || !insn_data[icode].operand[op].predicate (target, mode))
10422 target = gen_reg_rtx (mode);
10424 return target;
10427 /* Expand a MIPS_BUILTIN_DIRECT function. ICODE is the code of the
10428 .md pattern and CALL is the function expr with arguments. TARGET,
10429 if nonnull, suggests a good place to put the result.
10430 HAS_TARGET indicates the function must return something. */
10432 static rtx
10433 mips_expand_builtin_direct (enum insn_code icode, rtx target, tree exp,
10434 bool has_target)
10436 rtx ops[MAX_RECOG_OPERANDS];
10437 int i = 0;
10438 int j = 0;
10440 if (has_target)
10442 /* We save target to ops[0]. */
10443 ops[0] = mips_prepare_builtin_target (icode, 0, target);
10444 i = 1;
10447 /* We need to test if the arglist is not zero. Some instructions have extra
10448 clobber registers. */
10449 for (; i < insn_data[icode].n_operands && i <= call_expr_nargs (exp); i++, j++)
10450 ops[i] = mips_prepare_builtin_arg (icode, i, exp, j);
10452 switch (i)
10454 case 2:
10455 emit_insn (GEN_FCN (icode) (ops[0], ops[1]));
10456 break;
10458 case 3:
10459 emit_insn (GEN_FCN (icode) (ops[0], ops[1], ops[2]));
10460 break;
10462 case 4:
10463 emit_insn (GEN_FCN (icode) (ops[0], ops[1], ops[2], ops[3]));
10464 break;
10466 default:
10467 gcc_unreachable ();
10469 return target;
10472 /* Expand a __builtin_mips_movt_*_ps() or __builtin_mips_movf_*_ps()
10473 function (TYPE says which). EXP is the tree for the function
10474 function, ICODE is the instruction that should be used to compare
10475 the first two arguments, and COND is the condition it should test.
10476 TARGET, if nonnull, suggests a good place to put the result. */
10478 static rtx
10479 mips_expand_builtin_movtf (enum mips_builtin_type type,
10480 enum insn_code icode, enum mips_fp_condition cond,
10481 rtx target, tree exp)
10483 rtx cmp_result, op0, op1;
10485 cmp_result = mips_prepare_builtin_target (icode, 0, 0);
10486 op0 = mips_prepare_builtin_arg (icode, 1, exp, 0);
10487 op1 = mips_prepare_builtin_arg (icode, 2, exp, 1);
10488 emit_insn (GEN_FCN (icode) (cmp_result, op0, op1, GEN_INT (cond)));
10490 icode = CODE_FOR_mips_cond_move_tf_ps;
10491 target = mips_prepare_builtin_target (icode, 0, target);
10492 if (type == MIPS_BUILTIN_MOVT)
10494 op1 = mips_prepare_builtin_arg (icode, 2, exp, 2);
10495 op0 = mips_prepare_builtin_arg (icode, 1, exp, 3);
10497 else
10499 op0 = mips_prepare_builtin_arg (icode, 1, exp, 2);
10500 op1 = mips_prepare_builtin_arg (icode, 2, exp, 3);
10502 emit_insn (gen_mips_cond_move_tf_ps (target, op0, op1, cmp_result));
10503 return target;
10506 /* Move VALUE_IF_TRUE into TARGET if CONDITION is true; move VALUE_IF_FALSE
10507 into TARGET otherwise. Return TARGET. */
10509 static rtx
10510 mips_builtin_branch_and_move (rtx condition, rtx target,
10511 rtx value_if_true, rtx value_if_false)
10513 rtx true_label, done_label;
10515 true_label = gen_label_rtx ();
10516 done_label = gen_label_rtx ();
10518 /* First assume that CONDITION is false. */
10519 mips_emit_move (target, value_if_false);
10521 /* Branch to TRUE_LABEL if CONDITION is true and DONE_LABEL otherwise. */
10522 emit_jump_insn (gen_condjump (condition, true_label));
10523 emit_jump_insn (gen_jump (done_label));
10524 emit_barrier ();
10526 /* Fix TARGET if CONDITION is true. */
10527 emit_label (true_label);
10528 mips_emit_move (target, value_if_true);
10530 emit_label (done_label);
10531 return target;
10534 /* Expand a comparison builtin of type BUILTIN_TYPE. ICODE is the code
10535 of the comparison instruction and COND is the condition it should test.
10536 EXP is the function call and arguments and TARGET, if nonnull,
10537 suggests a good place to put the boolean result. */
10539 static rtx
10540 mips_expand_builtin_compare (enum mips_builtin_type builtin_type,
10541 enum insn_code icode, enum mips_fp_condition cond,
10542 rtx target, tree exp)
10544 rtx offset, condition, cmp_result, ops[MAX_RECOG_OPERANDS];
10545 int i;
10546 int j = 0;
10548 if (target == 0 || GET_MODE (target) != SImode)
10549 target = gen_reg_rtx (SImode);
10551 /* Prepare the operands to the comparison. */
10552 cmp_result = mips_prepare_builtin_target (icode, 0, 0);
10553 for (i = 1; i < insn_data[icode].n_operands - 1; i++, j++)
10554 ops[i] = mips_prepare_builtin_arg (icode, i, exp, j);
10556 switch (insn_data[icode].n_operands)
10558 case 4:
10559 emit_insn (GEN_FCN (icode) (cmp_result, ops[1], ops[2], GEN_INT (cond)));
10560 break;
10562 case 6:
10563 emit_insn (GEN_FCN (icode) (cmp_result, ops[1], ops[2],
10564 ops[3], ops[4], GEN_INT (cond)));
10565 break;
10567 default:
10568 gcc_unreachable ();
10571 /* If the comparison sets more than one register, we define the result
10572 to be 0 if all registers are false and -1 if all registers are true.
10573 The value of the complete result is indeterminate otherwise. */
10574 switch (builtin_type)
10576 case MIPS_BUILTIN_CMP_ALL:
10577 condition = gen_rtx_NE (VOIDmode, cmp_result, constm1_rtx);
10578 return mips_builtin_branch_and_move (condition, target,
10579 const0_rtx, const1_rtx);
10581 case MIPS_BUILTIN_CMP_UPPER:
10582 case MIPS_BUILTIN_CMP_LOWER:
10583 offset = GEN_INT (builtin_type == MIPS_BUILTIN_CMP_UPPER);
10584 condition = gen_single_cc (cmp_result, offset);
10585 return mips_builtin_branch_and_move (condition, target,
10586 const1_rtx, const0_rtx);
10588 default:
10589 condition = gen_rtx_NE (VOIDmode, cmp_result, const0_rtx);
10590 return mips_builtin_branch_and_move (condition, target,
10591 const1_rtx, const0_rtx);
10595 /* Expand a bposge builtin of type BUILTIN_TYPE. TARGET, if nonnull,
10596 suggests a good place to put the boolean result. */
10598 static rtx
10599 mips_expand_builtin_bposge (enum mips_builtin_type builtin_type, rtx target)
10601 rtx condition, cmp_result;
10602 int cmp_value;
10604 if (target == 0 || GET_MODE (target) != SImode)
10605 target = gen_reg_rtx (SImode);
10607 cmp_result = gen_rtx_REG (CCDSPmode, CCDSP_PO_REGNUM);
10609 if (builtin_type == MIPS_BUILTIN_BPOSGE32)
10610 cmp_value = 32;
10611 else
10612 gcc_assert (0);
10614 condition = gen_rtx_GE (VOIDmode, cmp_result, GEN_INT (cmp_value));
10615 return mips_builtin_branch_and_move (condition, target,
10616 const1_rtx, const0_rtx);
10619 /* EXP is a CALL_EXPR that calls the function described by BDESC.
10620 Expand the call and return an rtx for its return value.
10621 TARGET, if nonnull, suggests a good place to put this value. */
10623 static rtx
10624 mips_expand_builtin_1 (const struct builtin_description *bdesc,
10625 tree exp, rtx target)
10627 switch (bdesc->builtin_type)
10629 case MIPS_BUILTIN_DIRECT:
10630 return mips_expand_builtin_direct (bdesc->icode, target, exp, true);
10632 case MIPS_BUILTIN_DIRECT_NO_TARGET:
10633 return mips_expand_builtin_direct (bdesc->icode, target, exp, false);
10635 case MIPS_BUILTIN_MOVT:
10636 case MIPS_BUILTIN_MOVF:
10637 return mips_expand_builtin_movtf (bdesc->builtin_type, bdesc->icode,
10638 bdesc->cond, target, exp);
10640 case MIPS_BUILTIN_CMP_ANY:
10641 case MIPS_BUILTIN_CMP_ALL:
10642 case MIPS_BUILTIN_CMP_UPPER:
10643 case MIPS_BUILTIN_CMP_LOWER:
10644 case MIPS_BUILTIN_CMP_SINGLE:
10645 return mips_expand_builtin_compare (bdesc->builtin_type, bdesc->icode,
10646 bdesc->cond, target, exp);
10648 case MIPS_BUILTIN_BPOSGE32:
10649 return mips_expand_builtin_bposge (bdesc->builtin_type, target);
10651 gcc_unreachable ();
10654 /* Expand builtin functions. This is called from TARGET_EXPAND_BUILTIN. */
10656 static rtx
10657 mips_expand_builtin (tree exp, rtx target, rtx subtarget ATTRIBUTE_UNUSED,
10658 enum machine_mode mode ATTRIBUTE_UNUSED,
10659 int ignore ATTRIBUTE_UNUSED)
10661 tree fndecl;
10662 unsigned int fcode;
10663 const struct bdesc_map *m;
10665 fndecl = TREE_OPERAND (CALL_EXPR_FN (exp), 0);
10666 fcode = DECL_FUNCTION_CODE (fndecl);
10668 if (TARGET_MIPS16)
10670 error ("built-in function %qs not supported for MIPS16",
10671 IDENTIFIER_POINTER (DECL_NAME (fndecl)));
10672 return const0_rtx;
10675 for (m = bdesc_arrays; m < &bdesc_arrays[ARRAY_SIZE (bdesc_arrays)]; m++)
10677 if (fcode < m->size)
10678 return mips_expand_builtin_1 (m->bdesc + fcode, exp, target);
10679 fcode -= m->size;
10681 gcc_unreachable ();
10684 /* An entry in the mips16 constant pool. VALUE is the pool constant,
10685 MODE is its mode, and LABEL is the CODE_LABEL associated with it. */
10687 struct mips16_constant {
10688 struct mips16_constant *next;
10689 rtx value;
10690 rtx label;
10691 enum machine_mode mode;
10694 /* Information about an incomplete mips16 constant pool. FIRST is the
10695 first constant, HIGHEST_ADDRESS is the highest address that the first
10696 byte of the pool can have, and INSN_ADDRESS is the current instruction
10697 address. */
10699 struct mips16_constant_pool {
10700 struct mips16_constant *first;
10701 int highest_address;
10702 int insn_address;
10705 /* Add constant VALUE to POOL and return its label. MODE is the
10706 value's mode (used for CONST_INTs, etc.). */
10708 static rtx
10709 add_constant (struct mips16_constant_pool *pool,
10710 rtx value, enum machine_mode mode)
10712 struct mips16_constant **p, *c;
10713 bool first_of_size_p;
10715 /* See whether the constant is already in the pool. If so, return the
10716 existing label, otherwise leave P pointing to the place where the
10717 constant should be added.
10719 Keep the pool sorted in increasing order of mode size so that we can
10720 reduce the number of alignments needed. */
10721 first_of_size_p = true;
10722 for (p = &pool->first; *p != 0; p = &(*p)->next)
10724 if (mode == (*p)->mode && rtx_equal_p (value, (*p)->value))
10725 return (*p)->label;
10726 if (GET_MODE_SIZE (mode) < GET_MODE_SIZE ((*p)->mode))
10727 break;
10728 if (GET_MODE_SIZE (mode) == GET_MODE_SIZE ((*p)->mode))
10729 first_of_size_p = false;
10732 /* In the worst case, the constant needed by the earliest instruction
10733 will end up at the end of the pool. The entire pool must then be
10734 accessible from that instruction.
10736 When adding the first constant, set the pool's highest address to
10737 the address of the first out-of-range byte. Adjust this address
10738 downwards each time a new constant is added. */
10739 if (pool->first == 0)
10740 /* For pc-relative lw, addiu and daddiu instructions, the base PC value
10741 is the address of the instruction with the lowest two bits clear.
10742 The base PC value for ld has the lowest three bits clear. Assume
10743 the worst case here. */
10744 pool->highest_address = pool->insn_address - (UNITS_PER_WORD - 2) + 0x8000;
10745 pool->highest_address -= GET_MODE_SIZE (mode);
10746 if (first_of_size_p)
10747 /* Take into account the worst possible padding due to alignment. */
10748 pool->highest_address -= GET_MODE_SIZE (mode) - 1;
10750 /* Create a new entry. */
10751 c = (struct mips16_constant *) xmalloc (sizeof *c);
10752 c->value = value;
10753 c->mode = mode;
10754 c->label = gen_label_rtx ();
10755 c->next = *p;
10756 *p = c;
10758 return c->label;
10761 /* Output constant VALUE after instruction INSN and return the last
10762 instruction emitted. MODE is the mode of the constant. */
10764 static rtx
10765 dump_constants_1 (enum machine_mode mode, rtx value, rtx insn)
10767 if (SCALAR_INT_MODE_P (mode)
10768 || ALL_SCALAR_FRACT_MODE_P (mode)
10769 || ALL_SCALAR_ACCUM_MODE_P (mode))
10771 rtx size = GEN_INT (GET_MODE_SIZE (mode));
10772 return emit_insn_after (gen_consttable_int (value, size), insn);
10775 if (SCALAR_FLOAT_MODE_P (mode))
10776 return emit_insn_after (gen_consttable_float (value), insn);
10778 if (VECTOR_MODE_P (mode))
10780 int i;
10782 for (i = 0; i < CONST_VECTOR_NUNITS (value); i++)
10783 insn = dump_constants_1 (GET_MODE_INNER (mode),
10784 CONST_VECTOR_ELT (value, i), insn);
10785 return insn;
10788 gcc_unreachable ();
10792 /* Dump out the constants in CONSTANTS after INSN. */
10794 static void
10795 dump_constants (struct mips16_constant *constants, rtx insn)
10797 struct mips16_constant *c, *next;
10798 int align;
10800 align = 0;
10801 for (c = constants; c != NULL; c = next)
10803 /* If necessary, increase the alignment of PC. */
10804 if (align < GET_MODE_SIZE (c->mode))
10806 int align_log = floor_log2 (GET_MODE_SIZE (c->mode));
10807 insn = emit_insn_after (gen_align (GEN_INT (align_log)), insn);
10809 align = GET_MODE_SIZE (c->mode);
10811 insn = emit_label_after (c->label, insn);
10812 insn = dump_constants_1 (c->mode, c->value, insn);
10814 next = c->next;
10815 free (c);
10818 emit_barrier_after (insn);
10821 /* Return the length of instruction INSN. */
10823 static int
10824 mips16_insn_length (rtx insn)
10826 if (JUMP_P (insn))
10828 rtx body = PATTERN (insn);
10829 if (GET_CODE (body) == ADDR_VEC)
10830 return GET_MODE_SIZE (GET_MODE (body)) * XVECLEN (body, 0);
10831 if (GET_CODE (body) == ADDR_DIFF_VEC)
10832 return GET_MODE_SIZE (GET_MODE (body)) * XVECLEN (body, 1);
10834 return get_attr_length (insn);
10837 /* If *X is a symbolic constant that refers to the constant pool, add
10838 the constant to POOL and rewrite *X to use the constant's label. */
10840 static void
10841 mips16_rewrite_pool_constant (struct mips16_constant_pool *pool, rtx *x)
10843 rtx base, offset, label;
10845 split_const (*x, &base, &offset);
10846 if (GET_CODE (base) == SYMBOL_REF && CONSTANT_POOL_ADDRESS_P (base))
10848 label = add_constant (pool, get_pool_constant (base),
10849 get_pool_mode (base));
10850 base = gen_rtx_LABEL_REF (Pmode, label);
10851 *x = mips_unspec_address_offset (base, offset, SYMBOL_PC_RELATIVE);
10855 /* This structure is used to communicate with mips16_rewrite_pool_refs.
10856 INSN is the instruction we're rewriting and POOL points to the current
10857 constant pool. */
10858 struct mips16_rewrite_pool_refs_info {
10859 rtx insn;
10860 struct mips16_constant_pool *pool;
10863 /* Rewrite *X so that constant pool references refer to the constant's
10864 label instead. DATA points to a mips16_rewrite_pool_refs_info
10865 structure. */
10867 static int
10868 mips16_rewrite_pool_refs (rtx *x, void *data)
10870 struct mips16_rewrite_pool_refs_info *info = data;
10872 if (force_to_mem_operand (*x, Pmode))
10874 rtx mem = force_const_mem (GET_MODE (*x), *x);
10875 validate_change (info->insn, x, mem, false);
10878 if (MEM_P (*x))
10880 mips16_rewrite_pool_constant (info->pool, &XEXP (*x, 0));
10881 return -1;
10884 if (TARGET_MIPS16_TEXT_LOADS)
10885 mips16_rewrite_pool_constant (info->pool, x);
10887 return GET_CODE (*x) == CONST ? -1 : 0;
10890 /* Build MIPS16 constant pools. */
10892 static void
10893 mips16_lay_out_constants (void)
10895 struct mips16_constant_pool pool;
10896 struct mips16_rewrite_pool_refs_info info;
10897 rtx insn, barrier;
10899 if (!TARGET_MIPS16_PCREL_LOADS)
10900 return;
10902 barrier = 0;
10903 memset (&pool, 0, sizeof (pool));
10904 for (insn = get_insns (); insn; insn = NEXT_INSN (insn))
10906 /* Rewrite constant pool references in INSN. */
10907 if (INSN_P (insn))
10909 info.insn = insn;
10910 info.pool = &pool;
10911 for_each_rtx (&PATTERN (insn), mips16_rewrite_pool_refs, &info);
10914 pool.insn_address += mips16_insn_length (insn);
10916 if (pool.first != NULL)
10918 /* If there are no natural barriers between the first user of
10919 the pool and the highest acceptable address, we'll need to
10920 create a new instruction to jump around the constant pool.
10921 In the worst case, this instruction will be 4 bytes long.
10923 If it's too late to do this transformation after INSN,
10924 do it immediately before INSN. */
10925 if (barrier == 0 && pool.insn_address + 4 > pool.highest_address)
10927 rtx label, jump;
10929 label = gen_label_rtx ();
10931 jump = emit_jump_insn_before (gen_jump (label), insn);
10932 JUMP_LABEL (jump) = label;
10933 LABEL_NUSES (label) = 1;
10934 barrier = emit_barrier_after (jump);
10936 emit_label_after (label, barrier);
10937 pool.insn_address += 4;
10940 /* See whether the constant pool is now out of range of the first
10941 user. If so, output the constants after the previous barrier.
10942 Note that any instructions between BARRIER and INSN (inclusive)
10943 will use negative offsets to refer to the pool. */
10944 if (pool.insn_address > pool.highest_address)
10946 dump_constants (pool.first, barrier);
10947 pool.first = NULL;
10948 barrier = 0;
10950 else if (BARRIER_P (insn))
10951 barrier = insn;
10954 dump_constants (pool.first, get_last_insn ());
10957 /* A temporary variable used by for_each_rtx callbacks, etc. */
10958 static rtx mips_sim_insn;
10960 /* A structure representing the state of the processor pipeline.
10961 Used by the mips_sim_* family of functions. */
10962 struct mips_sim {
10963 /* The maximum number of instructions that can be issued in a cycle.
10964 (Caches mips_issue_rate.) */
10965 unsigned int issue_rate;
10967 /* The current simulation time. */
10968 unsigned int time;
10970 /* How many more instructions can be issued in the current cycle. */
10971 unsigned int insns_left;
10973 /* LAST_SET[X].INSN is the last instruction to set register X.
10974 LAST_SET[X].TIME is the time at which that instruction was issued.
10975 INSN is null if no instruction has yet set register X. */
10976 struct {
10977 rtx insn;
10978 unsigned int time;
10979 } last_set[FIRST_PSEUDO_REGISTER];
10981 /* The pipeline's current DFA state. */
10982 state_t dfa_state;
10985 /* Reset STATE to the initial simulation state. */
10987 static void
10988 mips_sim_reset (struct mips_sim *state)
10990 state->time = 0;
10991 state->insns_left = state->issue_rate;
10992 memset (&state->last_set, 0, sizeof (state->last_set));
10993 state_reset (state->dfa_state);
10996 /* Initialize STATE before its first use. DFA_STATE points to an
10997 allocated but uninitialized DFA state. */
10999 static void
11000 mips_sim_init (struct mips_sim *state, state_t dfa_state)
11002 state->issue_rate = mips_issue_rate ();
11003 state->dfa_state = dfa_state;
11004 mips_sim_reset (state);
11007 /* Advance STATE by one clock cycle. */
11009 static void
11010 mips_sim_next_cycle (struct mips_sim *state)
11012 state->time++;
11013 state->insns_left = state->issue_rate;
11014 state_transition (state->dfa_state, 0);
11017 /* Advance simulation state STATE until instruction INSN can read
11018 register REG. */
11020 static void
11021 mips_sim_wait_reg (struct mips_sim *state, rtx insn, rtx reg)
11023 unsigned int i;
11025 for (i = 0; i < HARD_REGNO_NREGS (REGNO (reg), GET_MODE (reg)); i++)
11026 if (state->last_set[REGNO (reg) + i].insn != 0)
11028 unsigned int t;
11030 t = state->last_set[REGNO (reg) + i].time;
11031 t += insn_latency (state->last_set[REGNO (reg) + i].insn, insn);
11032 while (state->time < t)
11033 mips_sim_next_cycle (state);
11037 /* A for_each_rtx callback. If *X is a register, advance simulation state
11038 DATA until mips_sim_insn can read the register's value. */
11040 static int
11041 mips_sim_wait_regs_2 (rtx *x, void *data)
11043 if (REG_P (*x))
11044 mips_sim_wait_reg (data, mips_sim_insn, *x);
11045 return 0;
11048 /* Call mips_sim_wait_regs_2 (R, DATA) for each register R mentioned in *X. */
11050 static void
11051 mips_sim_wait_regs_1 (rtx *x, void *data)
11053 for_each_rtx (x, mips_sim_wait_regs_2, data);
11056 /* Advance simulation state STATE until all of INSN's register
11057 dependencies are satisfied. */
11059 static void
11060 mips_sim_wait_regs (struct mips_sim *state, rtx insn)
11062 mips_sim_insn = insn;
11063 note_uses (&PATTERN (insn), mips_sim_wait_regs_1, state);
11066 /* Advance simulation state STATE until the units required by
11067 instruction INSN are available. */
11069 static void
11070 mips_sim_wait_units (struct mips_sim *state, rtx insn)
11072 state_t tmp_state;
11074 tmp_state = alloca (state_size ());
11075 while (state->insns_left == 0
11076 || (memcpy (tmp_state, state->dfa_state, state_size ()),
11077 state_transition (tmp_state, insn) >= 0))
11078 mips_sim_next_cycle (state);
11081 /* Advance simulation state STATE until INSN is ready to issue. */
11083 static void
11084 mips_sim_wait_insn (struct mips_sim *state, rtx insn)
11086 mips_sim_wait_regs (state, insn);
11087 mips_sim_wait_units (state, insn);
11090 /* mips_sim_insn has just set X. Update the LAST_SET array
11091 in simulation state DATA. */
11093 static void
11094 mips_sim_record_set (rtx x, const_rtx pat ATTRIBUTE_UNUSED, void *data)
11096 struct mips_sim *state;
11097 unsigned int i;
11099 state = data;
11100 if (REG_P (x))
11101 for (i = 0; i < HARD_REGNO_NREGS (REGNO (x), GET_MODE (x)); i++)
11103 state->last_set[REGNO (x) + i].insn = mips_sim_insn;
11104 state->last_set[REGNO (x) + i].time = state->time;
11108 /* Issue instruction INSN in scheduler state STATE. Assume that INSN
11109 can issue immediately (i.e., that mips_sim_wait_insn has already
11110 been called). */
11112 static void
11113 mips_sim_issue_insn (struct mips_sim *state, rtx insn)
11115 state_transition (state->dfa_state, insn);
11116 state->insns_left--;
11118 mips_sim_insn = insn;
11119 note_stores (PATTERN (insn), mips_sim_record_set, state);
11122 /* Simulate issuing a NOP in state STATE. */
11124 static void
11125 mips_sim_issue_nop (struct mips_sim *state)
11127 if (state->insns_left == 0)
11128 mips_sim_next_cycle (state);
11129 state->insns_left--;
11132 /* Update simulation state STATE so that it's ready to accept the instruction
11133 after INSN. INSN should be part of the main rtl chain, not a member of a
11134 SEQUENCE. */
11136 static void
11137 mips_sim_finish_insn (struct mips_sim *state, rtx insn)
11139 /* If INSN is a jump with an implicit delay slot, simulate a nop. */
11140 if (JUMP_P (insn))
11141 mips_sim_issue_nop (state);
11143 switch (GET_CODE (SEQ_BEGIN (insn)))
11145 case CODE_LABEL:
11146 case CALL_INSN:
11147 /* We can't predict the processor state after a call or label. */
11148 mips_sim_reset (state);
11149 break;
11151 case JUMP_INSN:
11152 /* The delay slots of branch likely instructions are only executed
11153 when the branch is taken. Therefore, if the caller has simulated
11154 the delay slot instruction, STATE does not really reflect the state
11155 of the pipeline for the instruction after the delay slot. Also,
11156 branch likely instructions tend to incur a penalty when not taken,
11157 so there will probably be an extra delay between the branch and
11158 the instruction after the delay slot. */
11159 if (INSN_ANNULLED_BRANCH_P (SEQ_BEGIN (insn)))
11160 mips_sim_reset (state);
11161 break;
11163 default:
11164 break;
11168 /* The VR4130 pipeline issues aligned pairs of instructions together,
11169 but it stalls the second instruction if it depends on the first.
11170 In order to cut down the amount of logic required, this dependence
11171 check is not based on a full instruction decode. Instead, any non-SPECIAL
11172 instruction is assumed to modify the register specified by bits 20-16
11173 (which is usually the "rt" field).
11175 In beq, beql, bne and bnel instructions, the rt field is actually an
11176 input, so we can end up with a false dependence between the branch
11177 and its delay slot. If this situation occurs in instruction INSN,
11178 try to avoid it by swapping rs and rt. */
11180 static void
11181 vr4130_avoid_branch_rt_conflict (rtx insn)
11183 rtx first, second;
11185 first = SEQ_BEGIN (insn);
11186 second = SEQ_END (insn);
11187 if (JUMP_P (first)
11188 && NONJUMP_INSN_P (second)
11189 && GET_CODE (PATTERN (first)) == SET
11190 && GET_CODE (SET_DEST (PATTERN (first))) == PC
11191 && GET_CODE (SET_SRC (PATTERN (first))) == IF_THEN_ELSE)
11193 /* Check for the right kind of condition. */
11194 rtx cond = XEXP (SET_SRC (PATTERN (first)), 0);
11195 if ((GET_CODE (cond) == EQ || GET_CODE (cond) == NE)
11196 && REG_P (XEXP (cond, 0))
11197 && REG_P (XEXP (cond, 1))
11198 && reg_referenced_p (XEXP (cond, 1), PATTERN (second))
11199 && !reg_referenced_p (XEXP (cond, 0), PATTERN (second)))
11201 /* SECOND mentions the rt register but not the rs register. */
11202 rtx tmp = XEXP (cond, 0);
11203 XEXP (cond, 0) = XEXP (cond, 1);
11204 XEXP (cond, 1) = tmp;
11209 /* Implement -mvr4130-align. Go through each basic block and simulate the
11210 processor pipeline. If we find that a pair of instructions could execute
11211 in parallel, and the first of those instruction is not 8-byte aligned,
11212 insert a nop to make it aligned. */
11214 static void
11215 vr4130_align_insns (void)
11217 struct mips_sim state;
11218 rtx insn, subinsn, last, last2, next;
11219 bool aligned_p;
11221 dfa_start ();
11223 /* LAST is the last instruction before INSN to have a nonzero length.
11224 LAST2 is the last such instruction before LAST. */
11225 last = 0;
11226 last2 = 0;
11228 /* ALIGNED_P is true if INSN is known to be at an aligned address. */
11229 aligned_p = true;
11231 mips_sim_init (&state, alloca (state_size ()));
11232 for (insn = get_insns (); insn != 0; insn = next)
11234 unsigned int length;
11236 next = NEXT_INSN (insn);
11238 /* See the comment above vr4130_avoid_branch_rt_conflict for details.
11239 This isn't really related to the alignment pass, but we do it on
11240 the fly to avoid a separate instruction walk. */
11241 vr4130_avoid_branch_rt_conflict (insn);
11243 if (USEFUL_INSN_P (insn))
11244 FOR_EACH_SUBINSN (subinsn, insn)
11246 mips_sim_wait_insn (&state, subinsn);
11248 /* If we want this instruction to issue in parallel with the
11249 previous one, make sure that the previous instruction is
11250 aligned. There are several reasons why this isn't worthwhile
11251 when the second instruction is a call:
11253 - Calls are less likely to be performance critical,
11254 - There's a good chance that the delay slot can execute
11255 in parallel with the call.
11256 - The return address would then be unaligned.
11258 In general, if we're going to insert a nop between instructions
11259 X and Y, it's better to insert it immediately after X. That
11260 way, if the nop makes Y aligned, it will also align any labels
11261 between X and Y. */
11262 if (state.insns_left != state.issue_rate
11263 && !CALL_P (subinsn))
11265 if (subinsn == SEQ_BEGIN (insn) && aligned_p)
11267 /* SUBINSN is the first instruction in INSN and INSN is
11268 aligned. We want to align the previous instruction
11269 instead, so insert a nop between LAST2 and LAST.
11271 Note that LAST could be either a single instruction
11272 or a branch with a delay slot. In the latter case,
11273 LAST, like INSN, is already aligned, but the delay
11274 slot must have some extra delay that stops it from
11275 issuing at the same time as the branch. We therefore
11276 insert a nop before the branch in order to align its
11277 delay slot. */
11278 emit_insn_after (gen_nop (), last2);
11279 aligned_p = false;
11281 else if (subinsn != SEQ_BEGIN (insn) && !aligned_p)
11283 /* SUBINSN is the delay slot of INSN, but INSN is
11284 currently unaligned. Insert a nop between
11285 LAST and INSN to align it. */
11286 emit_insn_after (gen_nop (), last);
11287 aligned_p = true;
11290 mips_sim_issue_insn (&state, subinsn);
11292 mips_sim_finish_insn (&state, insn);
11294 /* Update LAST, LAST2 and ALIGNED_P for the next instruction. */
11295 length = get_attr_length (insn);
11296 if (length > 0)
11298 /* If the instruction is an asm statement or multi-instruction
11299 mips.md patern, the length is only an estimate. Insert an
11300 8 byte alignment after it so that the following instructions
11301 can be handled correctly. */
11302 if (NONJUMP_INSN_P (SEQ_BEGIN (insn))
11303 && (recog_memoized (insn) < 0 || length >= 8))
11305 next = emit_insn_after (gen_align (GEN_INT (3)), insn);
11306 next = NEXT_INSN (next);
11307 mips_sim_next_cycle (&state);
11308 aligned_p = true;
11310 else if (length & 4)
11311 aligned_p = !aligned_p;
11312 last2 = last;
11313 last = insn;
11316 /* See whether INSN is an aligned label. */
11317 if (LABEL_P (insn) && label_to_alignment (insn) >= 3)
11318 aligned_p = true;
11320 dfa_finish ();
11323 /* Subroutine of mips_reorg. If there is a hazard between INSN
11324 and a previous instruction, avoid it by inserting nops after
11325 instruction AFTER.
11327 *DELAYED_REG and *HILO_DELAY describe the hazards that apply at
11328 this point. If *DELAYED_REG is non-null, INSN must wait a cycle
11329 before using the value of that register. *HILO_DELAY counts the
11330 number of instructions since the last hilo hazard (that is,
11331 the number of instructions since the last mflo or mfhi).
11333 After inserting nops for INSN, update *DELAYED_REG and *HILO_DELAY
11334 for the next instruction.
11336 LO_REG is an rtx for the LO register, used in dependence checking. */
11338 static void
11339 mips_avoid_hazard (rtx after, rtx insn, int *hilo_delay,
11340 rtx *delayed_reg, rtx lo_reg)
11342 rtx pattern, set;
11343 int nops, ninsns, hazard_set;
11345 if (!INSN_P (insn))
11346 return;
11348 pattern = PATTERN (insn);
11350 /* Do not put the whole function in .set noreorder if it contains
11351 an asm statement. We don't know whether there will be hazards
11352 between the asm statement and the gcc-generated code. */
11353 if (GET_CODE (pattern) == ASM_INPUT || asm_noperands (pattern) >= 0)
11354 cfun->machine->all_noreorder_p = false;
11356 /* Ignore zero-length instructions (barriers and the like). */
11357 ninsns = get_attr_length (insn) / 4;
11358 if (ninsns == 0)
11359 return;
11361 /* Work out how many nops are needed. Note that we only care about
11362 registers that are explicitly mentioned in the instruction's pattern.
11363 It doesn't matter that calls use the argument registers or that they
11364 clobber hi and lo. */
11365 if (*hilo_delay < 2 && reg_set_p (lo_reg, pattern))
11366 nops = 2 - *hilo_delay;
11367 else if (*delayed_reg != 0 && reg_referenced_p (*delayed_reg, pattern))
11368 nops = 1;
11369 else
11370 nops = 0;
11372 /* Insert the nops between this instruction and the previous one.
11373 Each new nop takes us further from the last hilo hazard. */
11374 *hilo_delay += nops;
11375 while (nops-- > 0)
11376 emit_insn_after (gen_hazard_nop (), after);
11378 /* Set up the state for the next instruction. */
11379 *hilo_delay += ninsns;
11380 *delayed_reg = 0;
11381 if (INSN_CODE (insn) >= 0)
11382 switch (get_attr_hazard (insn))
11384 case HAZARD_NONE:
11385 break;
11387 case HAZARD_HILO:
11388 *hilo_delay = 0;
11389 break;
11391 case HAZARD_DELAY:
11392 hazard_set = (int) get_attr_hazard_set (insn);
11393 if (hazard_set == 0)
11394 set = single_set (insn);
11395 else
11397 gcc_assert (GET_CODE (PATTERN (insn)) == PARALLEL);
11398 set = XVECEXP (PATTERN (insn), 0, hazard_set - 1);
11400 gcc_assert (set && GET_CODE (set) == SET);
11401 *delayed_reg = SET_DEST (set);
11402 break;
11407 /* Go through the instruction stream and insert nops where necessary.
11408 See if the whole function can then be put into .set noreorder &
11409 .set nomacro. */
11411 static void
11412 mips_avoid_hazards (void)
11414 rtx insn, last_insn, lo_reg, delayed_reg;
11415 int hilo_delay, i;
11417 /* Force all instructions to be split into their final form. */
11418 split_all_insns_noflow ();
11420 /* Recalculate instruction lengths without taking nops into account. */
11421 cfun->machine->ignore_hazard_length_p = true;
11422 shorten_branches (get_insns ());
11424 cfun->machine->all_noreorder_p = true;
11426 /* Profiled functions can't be all noreorder because the profiler
11427 support uses assembler macros. */
11428 if (current_function_profile)
11429 cfun->machine->all_noreorder_p = false;
11431 /* Code compiled with -mfix-vr4120 can't be all noreorder because
11432 we rely on the assembler to work around some errata. */
11433 if (TARGET_FIX_VR4120)
11434 cfun->machine->all_noreorder_p = false;
11436 /* The same is true for -mfix-vr4130 if we might generate mflo or
11437 mfhi instructions. Note that we avoid using mflo and mfhi if
11438 the VR4130 macc and dmacc instructions are available instead;
11439 see the *mfhilo_{si,di}_macc patterns. */
11440 if (TARGET_FIX_VR4130 && !ISA_HAS_MACCHI)
11441 cfun->machine->all_noreorder_p = false;
11443 last_insn = 0;
11444 hilo_delay = 2;
11445 delayed_reg = 0;
11446 lo_reg = gen_rtx_REG (SImode, LO_REGNUM);
11448 for (insn = get_insns (); insn != 0; insn = NEXT_INSN (insn))
11449 if (INSN_P (insn))
11451 if (GET_CODE (PATTERN (insn)) == SEQUENCE)
11452 for (i = 0; i < XVECLEN (PATTERN (insn), 0); i++)
11453 mips_avoid_hazard (last_insn, XVECEXP (PATTERN (insn), 0, i),
11454 &hilo_delay, &delayed_reg, lo_reg);
11455 else
11456 mips_avoid_hazard (last_insn, insn, &hilo_delay,
11457 &delayed_reg, lo_reg);
11459 last_insn = insn;
11464 /* Implement TARGET_MACHINE_DEPENDENT_REORG. */
11466 static void
11467 mips_reorg (void)
11469 mips16_lay_out_constants ();
11470 if (TARGET_EXPLICIT_RELOCS)
11472 if (mips_flag_delayed_branch)
11473 dbr_schedule (get_insns ());
11474 mips_avoid_hazards ();
11475 if (TUNE_MIPS4130 && TARGET_VR4130_ALIGN)
11476 vr4130_align_insns ();
11480 /* Implement TARGET_ASM_OUTPUT_MI_THUNK. Generate rtl rather than asm text
11481 in order to avoid duplicating too much logic from elsewhere. */
11483 static void
11484 mips_output_mi_thunk (FILE *file, tree thunk_fndecl ATTRIBUTE_UNUSED,
11485 HOST_WIDE_INT delta, HOST_WIDE_INT vcall_offset,
11486 tree function)
11488 rtx this, temp1, temp2, insn, fnaddr;
11489 bool use_sibcall_p;
11491 /* Pretend to be a post-reload pass while generating rtl. */
11492 reload_completed = 1;
11494 /* Mark the end of the (empty) prologue. */
11495 emit_note (NOTE_INSN_PROLOGUE_END);
11497 /* Determine if we can use a sibcall to call FUNCTION directly. */
11498 fnaddr = XEXP (DECL_RTL (function), 0);
11499 use_sibcall_p = (mips_function_ok_for_sibcall (function, NULL)
11500 && const_call_insn_operand (fnaddr, Pmode));
11502 /* Determine if we need to load FNADDR from the GOT. */
11503 if (!use_sibcall_p)
11504 switch (mips_classify_symbol (fnaddr, SYMBOL_CONTEXT_LEA))
11506 case SYMBOL_GOT_PAGE_OFST:
11507 case SYMBOL_GOT_DISP:
11508 /* Pick a global pointer. Use a call-clobbered register if
11509 TARGET_CALL_SAVED_GP. */
11510 cfun->machine->global_pointer =
11511 TARGET_CALL_SAVED_GP ? 15 : GLOBAL_POINTER_REGNUM;
11512 SET_REGNO (pic_offset_table_rtx, cfun->machine->global_pointer);
11514 /* Set up the global pointer for n32 or n64 abicalls. */
11515 mips_emit_loadgp ();
11516 break;
11518 default:
11519 break;
11522 /* We need two temporary registers in some cases. */
11523 temp1 = gen_rtx_REG (Pmode, 2);
11524 temp2 = gen_rtx_REG (Pmode, 3);
11526 /* Find out which register contains the "this" pointer. */
11527 if (aggregate_value_p (TREE_TYPE (TREE_TYPE (function)), function))
11528 this = gen_rtx_REG (Pmode, GP_ARG_FIRST + 1);
11529 else
11530 this = gen_rtx_REG (Pmode, GP_ARG_FIRST);
11532 /* Add DELTA to THIS. */
11533 if (delta != 0)
11535 rtx offset = GEN_INT (delta);
11536 if (!SMALL_OPERAND (delta))
11538 mips_emit_move (temp1, offset);
11539 offset = temp1;
11541 emit_insn (gen_add3_insn (this, this, offset));
11544 /* If needed, add *(*THIS + VCALL_OFFSET) to THIS. */
11545 if (vcall_offset != 0)
11547 rtx addr;
11549 /* Set TEMP1 to *THIS. */
11550 mips_emit_move (temp1, gen_rtx_MEM (Pmode, this));
11552 /* Set ADDR to a legitimate address for *THIS + VCALL_OFFSET. */
11553 addr = mips_add_offset (temp2, temp1, vcall_offset);
11555 /* Load the offset and add it to THIS. */
11556 mips_emit_move (temp1, gen_rtx_MEM (Pmode, addr));
11557 emit_insn (gen_add3_insn (this, this, temp1));
11560 /* Jump to the target function. Use a sibcall if direct jumps are
11561 allowed, otherwise load the address into a register first. */
11562 if (use_sibcall_p)
11564 insn = emit_call_insn (gen_sibcall_internal (fnaddr, const0_rtx));
11565 SIBLING_CALL_P (insn) = 1;
11567 else
11569 /* This is messy. gas treats "la $25,foo" as part of a call
11570 sequence and may allow a global "foo" to be lazily bound.
11571 The general move patterns therefore reject this combination.
11573 In this context, lazy binding would actually be OK
11574 for TARGET_CALL_CLOBBERED_GP, but it's still wrong for
11575 TARGET_CALL_SAVED_GP; see mips_load_call_address.
11576 We must therefore load the address via a temporary
11577 register if mips_dangerous_for_la25_p.
11579 If we jump to the temporary register rather than $25, the assembler
11580 can use the move insn to fill the jump's delay slot. */
11581 if (TARGET_USE_PIC_FN_ADDR_REG
11582 && !mips_dangerous_for_la25_p (fnaddr))
11583 temp1 = gen_rtx_REG (Pmode, PIC_FUNCTION_ADDR_REGNUM);
11584 mips_load_call_address (temp1, fnaddr, true);
11586 if (TARGET_USE_PIC_FN_ADDR_REG
11587 && REGNO (temp1) != PIC_FUNCTION_ADDR_REGNUM)
11588 mips_emit_move (gen_rtx_REG (Pmode, PIC_FUNCTION_ADDR_REGNUM), temp1);
11589 emit_jump_insn (gen_indirect_jump (temp1));
11592 /* Run just enough of rest_of_compilation. This sequence was
11593 "borrowed" from alpha.c. */
11594 insn = get_insns ();
11595 insn_locators_alloc ();
11596 split_all_insns_noflow ();
11597 mips16_lay_out_constants ();
11598 shorten_branches (insn);
11599 final_start_function (insn, file, 1);
11600 final (insn, file, 1);
11601 final_end_function ();
11603 /* Clean up the vars set above. Note that final_end_function resets
11604 the global pointer for us. */
11605 reload_completed = 0;
11608 static GTY(()) int was_mips16_p = -1;
11610 /* Set up the target-dependent global state so that it matches the
11611 current function's ISA mode. */
11613 static void
11614 mips_set_mips16_mode (int mips16_p)
11616 if (mips16_p == was_mips16_p)
11617 return;
11619 /* Restore base settings of various flags. */
11620 target_flags = mips_base_target_flags;
11621 flag_delayed_branch = mips_flag_delayed_branch;
11622 flag_schedule_insns = mips_base_schedule_insns;
11623 flag_reorder_blocks_and_partition = mips_base_reorder_blocks_and_partition;
11624 flag_move_loop_invariants = mips_base_move_loop_invariants;
11625 align_loops = mips_base_align_loops;
11626 align_jumps = mips_base_align_jumps;
11627 align_functions = mips_base_align_functions;
11629 if (mips16_p)
11631 /* Select mips16 instruction set. */
11632 target_flags |= MASK_MIPS16;
11634 /* Don't run the scheduler before reload, since it tends to
11635 increase register pressure. */
11636 flag_schedule_insns = 0;
11638 /* Don't do hot/cold partitioning. The constant layout code expects
11639 the whole function to be in a single section. */
11640 flag_reorder_blocks_and_partition = 0;
11642 /* Don't move loop invariants, because it tends to increase
11643 register pressure. It also introduces an extra move in cases
11644 where the constant is the first operand in a two-operand binary
11645 instruction, or when it forms a register argument to a functon
11646 call. */
11647 flag_move_loop_invariants = 0;
11649 /* Silently disable -mexplicit-relocs since it doesn't apply
11650 to mips16 code. Even so, it would overly pedantic to warn
11651 about "-mips16 -mexplicit-relocs", especially given that
11652 we use a %gprel() operator. */
11653 target_flags &= ~MASK_EXPLICIT_RELOCS;
11655 /* Experiments suggest we get the best overall results from using
11656 the range of an unextended lw or sw. Code that makes heavy use
11657 of byte or short accesses can do better with ranges of 0...31
11658 and 0...63 respectively, but most code is sensitive to the range
11659 of lw and sw instead. */
11660 targetm.min_anchor_offset = 0;
11661 targetm.max_anchor_offset = 127;
11663 if (flag_pic || TARGET_ABICALLS)
11664 sorry ("MIPS16 PIC");
11666 else
11668 /* Reset to select base non-mips16 ISA. */
11669 target_flags &= ~MASK_MIPS16;
11671 /* When using explicit relocs, we call dbr_schedule from within
11672 mips_reorg. */
11673 if (TARGET_EXPLICIT_RELOCS)
11674 flag_delayed_branch = 0;
11676 /* Provide default values for align_* for 64-bit targets. */
11677 if (TARGET_64BIT)
11679 if (align_loops == 0)
11680 align_loops = 8;
11681 if (align_jumps == 0)
11682 align_jumps = 8;
11683 if (align_functions == 0)
11684 align_functions = 8;
11687 targetm.min_anchor_offset = -32768;
11688 targetm.max_anchor_offset = 32767;
11691 /* (Re)initialize mips target internals for new ISA. */
11692 mips_init_split_addresses ();
11693 mips_init_relocs ();
11695 if (was_mips16_p >= 0)
11696 /* Reinitialize target-dependent state. */
11697 target_reinit ();
11699 was_mips16_p = TARGET_MIPS16;
11702 /* Implement TARGET_SET_CURRENT_FUNCTION. Decide whether the current
11703 function should use the MIPS16 ISA and switch modes accordingly. */
11705 static void
11706 mips_set_current_function (tree fndecl)
11708 mips_set_mips16_mode (mips_use_mips16_mode_p (fndecl));
11711 /* Allocate a chunk of memory for per-function machine-dependent data. */
11712 static struct machine_function *
11713 mips_init_machine_status (void)
11715 return ((struct machine_function *)
11716 ggc_alloc_cleared (sizeof (struct machine_function)));
11719 /* Return the processor associated with the given ISA level, or null
11720 if the ISA isn't valid. */
11722 static const struct mips_cpu_info *
11723 mips_cpu_info_from_isa (int isa)
11725 unsigned int i;
11727 for (i = 0; i < ARRAY_SIZE (mips_cpu_info_table); i++)
11728 if (mips_cpu_info_table[i].isa == isa)
11729 return mips_cpu_info_table + i;
11731 return 0;
11734 /* Return true if GIVEN is the same as CANONICAL, or if it is CANONICAL
11735 with a final "000" replaced by "k". Ignore case.
11737 Note: this function is shared between GCC and GAS. */
11739 static bool
11740 mips_strict_matching_cpu_name_p (const char *canonical, const char *given)
11742 while (*given != 0 && TOLOWER (*given) == TOLOWER (*canonical))
11743 given++, canonical++;
11745 return ((*given == 0 && *canonical == 0)
11746 || (strcmp (canonical, "000") == 0 && strcasecmp (given, "k") == 0));
11750 /* Return true if GIVEN matches CANONICAL, where GIVEN is a user-supplied
11751 CPU name. We've traditionally allowed a lot of variation here.
11753 Note: this function is shared between GCC and GAS. */
11755 static bool
11756 mips_matching_cpu_name_p (const char *canonical, const char *given)
11758 /* First see if the name matches exactly, or with a final "000"
11759 turned into "k". */
11760 if (mips_strict_matching_cpu_name_p (canonical, given))
11761 return true;
11763 /* If not, try comparing based on numerical designation alone.
11764 See if GIVEN is an unadorned number, or 'r' followed by a number. */
11765 if (TOLOWER (*given) == 'r')
11766 given++;
11767 if (!ISDIGIT (*given))
11768 return false;
11770 /* Skip over some well-known prefixes in the canonical name,
11771 hoping to find a number there too. */
11772 if (TOLOWER (canonical[0]) == 'v' && TOLOWER (canonical[1]) == 'r')
11773 canonical += 2;
11774 else if (TOLOWER (canonical[0]) == 'r' && TOLOWER (canonical[1]) == 'm')
11775 canonical += 2;
11776 else if (TOLOWER (canonical[0]) == 'r')
11777 canonical += 1;
11779 return mips_strict_matching_cpu_name_p (canonical, given);
11783 /* Return the mips_cpu_info entry for the processor or ISA given
11784 by CPU_STRING. Return null if the string isn't recognized.
11786 A similar function exists in GAS. */
11788 static const struct mips_cpu_info *
11789 mips_parse_cpu (const char *cpu_string)
11791 unsigned int i;
11792 const char *s;
11794 /* In the past, we allowed upper-case CPU names, but it doesn't
11795 work well with the multilib machinery. */
11796 for (s = cpu_string; *s != 0; s++)
11797 if (ISUPPER (*s))
11799 warning (0, "the cpu name must be lower case");
11800 break;
11803 /* 'from-abi' selects the most compatible architecture for the given
11804 ABI: MIPS I for 32-bit ABIs and MIPS III for 64-bit ABIs. For the
11805 EABIs, we have to decide whether we're using the 32-bit or 64-bit
11806 version. Look first at the -mgp options, if given, otherwise base
11807 the choice on MASK_64BIT in TARGET_DEFAULT. */
11808 if (strcasecmp (cpu_string, "from-abi") == 0)
11809 return mips_cpu_info_from_isa (ABI_NEEDS_32BIT_REGS ? 1
11810 : ABI_NEEDS_64BIT_REGS ? 3
11811 : (TARGET_64BIT ? 3 : 1));
11813 /* 'default' has traditionally been a no-op. Probably not very useful. */
11814 if (strcasecmp (cpu_string, "default") == 0)
11815 return 0;
11817 for (i = 0; i < ARRAY_SIZE (mips_cpu_info_table); i++)
11818 if (mips_matching_cpu_name_p (mips_cpu_info_table[i].name, cpu_string))
11819 return mips_cpu_info_table + i;
11821 return 0;
11825 /* Set up globals to generate code for the ISA or processor
11826 described by INFO. */
11828 static void
11829 mips_set_architecture (const struct mips_cpu_info *info)
11831 if (info != 0)
11833 mips_arch_info = info;
11834 mips_arch = info->cpu;
11835 mips_isa = info->isa;
11840 /* Likewise for tuning. */
11842 static void
11843 mips_set_tune (const struct mips_cpu_info *info)
11845 if (info != 0)
11847 mips_tune_info = info;
11848 mips_tune = info->cpu;
11852 /* Implement TARGET_HANDLE_OPTION. */
11854 static bool
11855 mips_handle_option (size_t code, const char *arg, int value ATTRIBUTE_UNUSED)
11857 switch (code)
11859 case OPT_mabi_:
11860 if (strcmp (arg, "32") == 0)
11861 mips_abi = ABI_32;
11862 else if (strcmp (arg, "o64") == 0)
11863 mips_abi = ABI_O64;
11864 else if (strcmp (arg, "n32") == 0)
11865 mips_abi = ABI_N32;
11866 else if (strcmp (arg, "64") == 0)
11867 mips_abi = ABI_64;
11868 else if (strcmp (arg, "eabi") == 0)
11869 mips_abi = ABI_EABI;
11870 else
11871 return false;
11872 return true;
11874 case OPT_march_:
11875 case OPT_mtune_:
11876 return mips_parse_cpu (arg) != 0;
11878 case OPT_mips:
11879 mips_isa_info = mips_parse_cpu (ACONCAT (("mips", arg, NULL)));
11880 return mips_isa_info != 0;
11882 case OPT_mno_flush_func:
11883 mips_cache_flush_func = NULL;
11884 return true;
11886 case OPT_mcode_readable_:
11887 if (strcmp (arg, "yes") == 0)
11888 mips_code_readable = CODE_READABLE_YES;
11889 else if (strcmp (arg, "pcrel") == 0)
11890 mips_code_readable = CODE_READABLE_PCREL;
11891 else if (strcmp (arg, "no") == 0)
11892 mips_code_readable = CODE_READABLE_NO;
11893 else
11894 return false;
11895 return true;
11897 default:
11898 return true;
11902 /* Set up the threshold for data to go into the small data area, instead
11903 of the normal data area, and detect any conflicts in the switches. */
11905 void
11906 override_options (void)
11908 int i, start, regno;
11909 enum machine_mode mode;
11911 #ifdef SUBTARGET_OVERRIDE_OPTIONS
11912 SUBTARGET_OVERRIDE_OPTIONS;
11913 #endif
11915 mips_section_threshold = g_switch_set ? g_switch_value : MIPS_DEFAULT_GVALUE;
11917 /* The following code determines the architecture and register size.
11918 Similar code was added to GAS 2.14 (see tc-mips.c:md_after_parse_args()).
11919 The GAS and GCC code should be kept in sync as much as possible. */
11921 if (mips_arch_string != 0)
11922 mips_set_architecture (mips_parse_cpu (mips_arch_string));
11924 if (mips_isa_info != 0)
11926 if (mips_arch_info == 0)
11927 mips_set_architecture (mips_isa_info);
11928 else if (mips_arch_info->isa != mips_isa_info->isa)
11929 error ("-%s conflicts with the other architecture options, "
11930 "which specify a %s processor",
11931 mips_isa_info->name,
11932 mips_cpu_info_from_isa (mips_arch_info->isa)->name);
11935 if (mips_arch_info == 0)
11937 #ifdef MIPS_CPU_STRING_DEFAULT
11938 mips_set_architecture (mips_parse_cpu (MIPS_CPU_STRING_DEFAULT));
11939 #else
11940 mips_set_architecture (mips_cpu_info_from_isa (MIPS_ISA_DEFAULT));
11941 #endif
11944 if (ABI_NEEDS_64BIT_REGS && !ISA_HAS_64BIT_REGS)
11945 error ("-march=%s is not compatible with the selected ABI",
11946 mips_arch_info->name);
11948 /* Optimize for mips_arch, unless -mtune selects a different processor. */
11949 if (mips_tune_string != 0)
11950 mips_set_tune (mips_parse_cpu (mips_tune_string));
11952 if (mips_tune_info == 0)
11953 mips_set_tune (mips_arch_info);
11955 /* Set cost structure for the processor. */
11956 if (optimize_size)
11957 mips_cost = &mips_rtx_cost_optimize_size;
11958 else
11959 mips_cost = &mips_rtx_cost_data[mips_tune];
11961 /* If the user hasn't specified a branch cost, use the processor's
11962 default. */
11963 if (mips_branch_cost == 0)
11964 mips_branch_cost = mips_cost->branch_cost;
11966 if ((target_flags_explicit & MASK_64BIT) != 0)
11968 /* The user specified the size of the integer registers. Make sure
11969 it agrees with the ABI and ISA. */
11970 if (TARGET_64BIT && !ISA_HAS_64BIT_REGS)
11971 error ("-mgp64 used with a 32-bit processor");
11972 else if (!TARGET_64BIT && ABI_NEEDS_64BIT_REGS)
11973 error ("-mgp32 used with a 64-bit ABI");
11974 else if (TARGET_64BIT && ABI_NEEDS_32BIT_REGS)
11975 error ("-mgp64 used with a 32-bit ABI");
11977 else
11979 /* Infer the integer register size from the ABI and processor.
11980 Restrict ourselves to 32-bit registers if that's all the
11981 processor has, or if the ABI cannot handle 64-bit registers. */
11982 if (ABI_NEEDS_32BIT_REGS || !ISA_HAS_64BIT_REGS)
11983 target_flags &= ~MASK_64BIT;
11984 else
11985 target_flags |= MASK_64BIT;
11988 if ((target_flags_explicit & MASK_FLOAT64) != 0)
11990 /* Really, -mfp32 and -mfp64 are ornamental options. There's
11991 only one right answer here. */
11992 if (TARGET_64BIT && TARGET_DOUBLE_FLOAT && !TARGET_FLOAT64)
11993 error ("unsupported combination: %s", "-mgp64 -mfp32 -mdouble-float");
11994 else if (!TARGET_64BIT && TARGET_FLOAT64
11995 && !(ISA_HAS_MXHC1 && mips_abi == ABI_32))
11996 error ("-mgp32 and -mfp64 can only be combined if the target"
11997 " supports the mfhc1 and mthc1 instructions");
11998 else if (TARGET_SINGLE_FLOAT && TARGET_FLOAT64)
11999 error ("unsupported combination: %s", "-mfp64 -msingle-float");
12001 else
12003 /* -msingle-float selects 32-bit float registers. Otherwise the
12004 float registers should be the same size as the integer ones. */
12005 if (TARGET_64BIT && TARGET_DOUBLE_FLOAT)
12006 target_flags |= MASK_FLOAT64;
12007 else
12008 target_flags &= ~MASK_FLOAT64;
12011 /* End of code shared with GAS. */
12013 if ((target_flags_explicit & MASK_LONG64) == 0)
12015 if ((mips_abi == ABI_EABI && TARGET_64BIT) || mips_abi == ABI_64)
12016 target_flags |= MASK_LONG64;
12017 else
12018 target_flags &= ~MASK_LONG64;
12021 if (!TARGET_OLDABI)
12022 flag_pcc_struct_return = 0;
12024 if ((target_flags_explicit & MASK_BRANCHLIKELY) == 0)
12026 /* If neither -mbranch-likely nor -mno-branch-likely was given
12027 on the command line, set MASK_BRANCHLIKELY based on the target
12028 architecture and tuning flags. Annulled delay slots are a
12029 size win, so we only consider the processor-specific tuning
12030 for !optimize_size. */
12031 if (ISA_HAS_BRANCHLIKELY
12032 && (optimize_size
12033 || (mips_tune_info->tune_flags & PTF_AVOID_BRANCHLIKELY) == 0))
12034 target_flags |= MASK_BRANCHLIKELY;
12035 else
12036 target_flags &= ~MASK_BRANCHLIKELY;
12038 else if (TARGET_BRANCHLIKELY && !ISA_HAS_BRANCHLIKELY)
12039 warning (0, "the %qs architecture does not support branch-likely"
12040 " instructions", mips_arch_info->name);
12042 /* The effect of -mabicalls isn't defined for the EABI. */
12043 if (mips_abi == ABI_EABI && TARGET_ABICALLS)
12045 error ("unsupported combination: %s", "-mabicalls -mabi=eabi");
12046 target_flags &= ~MASK_ABICALLS;
12049 /* MIPS16 cannot generate PIC yet. */
12050 if (TARGET_MIPS16 && (flag_pic || TARGET_ABICALLS))
12052 sorry ("MIPS16 PIC");
12053 target_flags &= ~MASK_ABICALLS;
12054 flag_pic = flag_pie = flag_shlib = 0;
12057 if (TARGET_ABICALLS)
12058 /* We need to set flag_pic for executables as well as DSOs
12059 because we may reference symbols that are not defined in
12060 the final executable. (MIPS does not use things like
12061 copy relocs, for example.)
12063 Also, there is a body of code that uses __PIC__ to distinguish
12064 between -mabicalls and -mno-abicalls code. */
12065 flag_pic = 1;
12067 /* -mvr4130-align is a "speed over size" optimization: it usually produces
12068 faster code, but at the expense of more nops. Enable it at -O3 and
12069 above. */
12070 if (optimize > 2 && (target_flags_explicit & MASK_VR4130_ALIGN) == 0)
12071 target_flags |= MASK_VR4130_ALIGN;
12073 /* Prefer a call to memcpy over inline code when optimizing for size,
12074 though see MOVE_RATIO in mips.h. */
12075 if (optimize_size && (target_flags_explicit & MASK_MEMCPY) == 0)
12076 target_flags |= MASK_MEMCPY;
12078 /* If we have a nonzero small-data limit, check that the -mgpopt
12079 setting is consistent with the other target flags. */
12080 if (mips_section_threshold > 0)
12082 if (!TARGET_GPOPT)
12084 if (!TARGET_MIPS16 && !TARGET_EXPLICIT_RELOCS)
12085 error ("%<-mno-gpopt%> needs %<-mexplicit-relocs%>");
12087 TARGET_LOCAL_SDATA = false;
12088 TARGET_EXTERN_SDATA = false;
12090 else
12092 if (TARGET_VXWORKS_RTP)
12093 warning (0, "cannot use small-data accesses for %qs", "-mrtp");
12095 if (TARGET_ABICALLS)
12096 warning (0, "cannot use small-data accesses for %qs",
12097 "-mabicalls");
12101 #ifdef MIPS_TFMODE_FORMAT
12102 REAL_MODE_FORMAT (TFmode) = &MIPS_TFMODE_FORMAT;
12103 #endif
12105 /* Make sure that the user didn't turn off paired single support when
12106 MIPS-3D support is requested. */
12107 if (TARGET_MIPS3D && (target_flags_explicit & MASK_PAIRED_SINGLE_FLOAT)
12108 && !TARGET_PAIRED_SINGLE_FLOAT)
12109 error ("-mips3d requires -mpaired-single");
12111 /* If TARGET_MIPS3D, enable MASK_PAIRED_SINGLE_FLOAT. */
12112 if (TARGET_MIPS3D)
12113 target_flags |= MASK_PAIRED_SINGLE_FLOAT;
12115 /* Make sure that when TARGET_PAIRED_SINGLE_FLOAT is true, TARGET_FLOAT64
12116 and TARGET_HARD_FLOAT_ABI are both true. */
12117 if (TARGET_PAIRED_SINGLE_FLOAT && !(TARGET_FLOAT64 && TARGET_HARD_FLOAT_ABI))
12118 error ("-mips3d/-mpaired-single must be used with -mfp64 -mhard-float");
12120 /* Make sure that the ISA supports TARGET_PAIRED_SINGLE_FLOAT when it is
12121 enabled. */
12122 if (TARGET_PAIRED_SINGLE_FLOAT && !ISA_HAS_PAIRED_SINGLE)
12123 warning (0, "the %qs architecture does not support paired-single"
12124 " instructions", mips_arch_info->name);
12126 /* If TARGET_DSPR2, enable MASK_DSP. */
12127 if (TARGET_DSPR2)
12128 target_flags |= MASK_DSP;
12130 mips_init_print_operand_punct ();
12132 /* Set up array to map GCC register number to debug register number.
12133 Ignore the special purpose register numbers. */
12135 for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
12137 mips_dbx_regno[i] = INVALID_REGNUM;
12138 if (GP_REG_P (i) || FP_REG_P (i) || ALL_COP_REG_P (i))
12139 mips_dwarf_regno[i] = i;
12140 else
12141 mips_dwarf_regno[i] = INVALID_REGNUM;
12144 start = GP_DBX_FIRST - GP_REG_FIRST;
12145 for (i = GP_REG_FIRST; i <= GP_REG_LAST; i++)
12146 mips_dbx_regno[i] = i + start;
12148 start = FP_DBX_FIRST - FP_REG_FIRST;
12149 for (i = FP_REG_FIRST; i <= FP_REG_LAST; i++)
12150 mips_dbx_regno[i] = i + start;
12152 /* HI and LO debug registers use big-endian ordering. */
12153 mips_dbx_regno[HI_REGNUM] = MD_DBX_FIRST + 0;
12154 mips_dbx_regno[LO_REGNUM] = MD_DBX_FIRST + 1;
12155 mips_dwarf_regno[HI_REGNUM] = MD_REG_FIRST + 0;
12156 mips_dwarf_regno[LO_REGNUM] = MD_REG_FIRST + 1;
12157 for (i = DSP_ACC_REG_FIRST; i <= DSP_ACC_REG_LAST; i += 2)
12159 mips_dwarf_regno[i + TARGET_LITTLE_ENDIAN] = i;
12160 mips_dwarf_regno[i + TARGET_BIG_ENDIAN] = i + 1;
12163 /* Set up mips_hard_regno_mode_ok. */
12164 for (mode = 0; mode < MAX_MACHINE_MODE; mode++)
12165 for (regno = 0; regno < FIRST_PSEUDO_REGISTER; regno++)
12166 mips_hard_regno_mode_ok[(int)mode][regno]
12167 = mips_hard_regno_mode_ok_p (regno, mode);
12169 /* Function to allocate machine-dependent function status. */
12170 init_machine_status = &mips_init_machine_status;
12172 /* Default to working around R4000 errata only if the processor
12173 was selected explicitly. */
12174 if ((target_flags_explicit & MASK_FIX_R4000) == 0
12175 && mips_matching_cpu_name_p (mips_arch_info->name, "r4000"))
12176 target_flags |= MASK_FIX_R4000;
12178 /* Default to working around R4400 errata only if the processor
12179 was selected explicitly. */
12180 if ((target_flags_explicit & MASK_FIX_R4400) == 0
12181 && mips_matching_cpu_name_p (mips_arch_info->name, "r4400"))
12182 target_flags |= MASK_FIX_R4400;
12184 /* Save base state of options. */
12185 mips_base_mips16 = TARGET_MIPS16;
12186 mips_base_target_flags = target_flags;
12187 mips_flag_delayed_branch = flag_delayed_branch;
12188 mips_base_schedule_insns = flag_schedule_insns;
12189 mips_base_reorder_blocks_and_partition = flag_reorder_blocks_and_partition;
12190 mips_base_move_loop_invariants = flag_move_loop_invariants;
12191 mips_base_align_loops = align_loops;
12192 mips_base_align_jumps = align_jumps;
12193 mips_base_align_functions = align_functions;
12195 /* Now select the mips16 or 32-bit instruction set, as requested. */
12196 mips_set_mips16_mode (mips_base_mips16);
12199 /* Swap the register information for registers I and I + 1, which
12200 currently have the wrong endianness. Note that the registers'
12201 fixedness and call-clobberedness might have been set on the
12202 command line. */
12204 static void
12205 mips_swap_registers (unsigned int i)
12207 int tmpi;
12208 const char *tmps;
12210 #define SWAP_INT(X, Y) (tmpi = (X), (X) = (Y), (Y) = tmpi)
12211 #define SWAP_STRING(X, Y) (tmps = (X), (X) = (Y), (Y) = tmps)
12213 SWAP_INT (fixed_regs[i], fixed_regs[i + 1]);
12214 SWAP_INT (call_used_regs[i], call_used_regs[i + 1]);
12215 SWAP_INT (call_really_used_regs[i], call_really_used_regs[i + 1]);
12216 SWAP_STRING (reg_names[i], reg_names[i + 1]);
12218 #undef SWAP_STRING
12219 #undef SWAP_INT
12222 /* Implement CONDITIONAL_REGISTER_USAGE. */
12224 void
12225 mips_conditional_register_usage (void)
12227 if (!ISA_HAS_DSP)
12229 int regno;
12231 for (regno = DSP_ACC_REG_FIRST; regno <= DSP_ACC_REG_LAST; regno++)
12232 fixed_regs[regno] = call_used_regs[regno] = 1;
12234 if (!TARGET_HARD_FLOAT)
12236 int regno;
12238 for (regno = FP_REG_FIRST; regno <= FP_REG_LAST; regno++)
12239 fixed_regs[regno] = call_used_regs[regno] = 1;
12240 for (regno = ST_REG_FIRST; regno <= ST_REG_LAST; regno++)
12241 fixed_regs[regno] = call_used_regs[regno] = 1;
12243 else if (! ISA_HAS_8CC)
12245 int regno;
12247 /* We only have a single condition code register. We
12248 implement this by hiding all the condition code registers,
12249 and generating RTL that refers directly to ST_REG_FIRST. */
12250 for (regno = ST_REG_FIRST; regno <= ST_REG_LAST; regno++)
12251 fixed_regs[regno] = call_used_regs[regno] = 1;
12253 /* In mips16 mode, we permit the $t temporary registers to be used
12254 for reload. We prohibit the unused $s registers, since they
12255 are caller saved, and saving them via a mips16 register would
12256 probably waste more time than just reloading the value. */
12257 if (TARGET_MIPS16)
12259 fixed_regs[18] = call_used_regs[18] = 1;
12260 fixed_regs[19] = call_used_regs[19] = 1;
12261 fixed_regs[20] = call_used_regs[20] = 1;
12262 fixed_regs[21] = call_used_regs[21] = 1;
12263 fixed_regs[22] = call_used_regs[22] = 1;
12264 fixed_regs[23] = call_used_regs[23] = 1;
12265 fixed_regs[26] = call_used_regs[26] = 1;
12266 fixed_regs[27] = call_used_regs[27] = 1;
12267 fixed_regs[30] = call_used_regs[30] = 1;
12269 /* fp20-23 are now caller saved. */
12270 if (mips_abi == ABI_64)
12272 int regno;
12273 for (regno = FP_REG_FIRST + 20; regno < FP_REG_FIRST + 24; regno++)
12274 call_really_used_regs[regno] = call_used_regs[regno] = 1;
12276 /* Odd registers from fp21 to fp31 are now caller saved. */
12277 if (mips_abi == ABI_N32)
12279 int regno;
12280 for (regno = FP_REG_FIRST + 21; regno <= FP_REG_FIRST + 31; regno+=2)
12281 call_really_used_regs[regno] = call_used_regs[regno] = 1;
12283 /* Make sure that double-register accumulator values are correctly
12284 ordered for the current endianness. */
12285 if (TARGET_LITTLE_ENDIAN)
12287 int regno;
12288 mips_swap_registers (MD_REG_FIRST);
12289 for (regno = DSP_ACC_REG_FIRST; regno <= DSP_ACC_REG_LAST; regno += 2)
12290 mips_swap_registers (regno);
12294 /* On the mips16, we want to allocate $24 (T_REG) before other
12295 registers for instructions for which it is possible. This helps
12296 avoid shuffling registers around in order to set up for an xor,
12297 encouraging the compiler to use a cmp instead. */
12299 void
12300 mips_order_regs_for_local_alloc (void)
12302 register int i;
12304 for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
12305 reg_alloc_order[i] = i;
12307 if (TARGET_MIPS16)
12309 /* It really doesn't matter where we put register 0, since it is
12310 a fixed register anyhow. */
12311 reg_alloc_order[0] = 24;
12312 reg_alloc_order[24] = 0;
12316 /* Initialize the GCC target structure. */
12317 #undef TARGET_ASM_ALIGNED_HI_OP
12318 #define TARGET_ASM_ALIGNED_HI_OP "\t.half\t"
12319 #undef TARGET_ASM_ALIGNED_SI_OP
12320 #define TARGET_ASM_ALIGNED_SI_OP "\t.word\t"
12321 #undef TARGET_ASM_ALIGNED_DI_OP
12322 #define TARGET_ASM_ALIGNED_DI_OP "\t.dword\t"
12324 #undef TARGET_ASM_FUNCTION_PROLOGUE
12325 #define TARGET_ASM_FUNCTION_PROLOGUE mips_output_function_prologue
12326 #undef TARGET_ASM_FUNCTION_EPILOGUE
12327 #define TARGET_ASM_FUNCTION_EPILOGUE mips_output_function_epilogue
12328 #undef TARGET_ASM_SELECT_RTX_SECTION
12329 #define TARGET_ASM_SELECT_RTX_SECTION mips_select_rtx_section
12330 #undef TARGET_ASM_FUNCTION_RODATA_SECTION
12331 #define TARGET_ASM_FUNCTION_RODATA_SECTION mips_function_rodata_section
12333 #undef TARGET_SCHED_INIT
12334 #define TARGET_SCHED_INIT mips_sched_init
12335 #undef TARGET_SCHED_REORDER
12336 #define TARGET_SCHED_REORDER mips_sched_reorder
12337 #undef TARGET_SCHED_REORDER2
12338 #define TARGET_SCHED_REORDER2 mips_sched_reorder
12339 #undef TARGET_SCHED_VARIABLE_ISSUE
12340 #define TARGET_SCHED_VARIABLE_ISSUE mips_variable_issue
12341 #undef TARGET_SCHED_ADJUST_COST
12342 #define TARGET_SCHED_ADJUST_COST mips_adjust_cost
12343 #undef TARGET_SCHED_ISSUE_RATE
12344 #define TARGET_SCHED_ISSUE_RATE mips_issue_rate
12345 #undef TARGET_SCHED_FIRST_CYCLE_MULTIPASS_DFA_LOOKAHEAD
12346 #define TARGET_SCHED_FIRST_CYCLE_MULTIPASS_DFA_LOOKAHEAD \
12347 mips_multipass_dfa_lookahead
12349 #undef TARGET_DEFAULT_TARGET_FLAGS
12350 #define TARGET_DEFAULT_TARGET_FLAGS \
12351 (TARGET_DEFAULT \
12352 | TARGET_CPU_DEFAULT \
12353 | TARGET_ENDIAN_DEFAULT \
12354 | TARGET_FP_EXCEPTIONS_DEFAULT \
12355 | MASK_CHECK_ZERO_DIV \
12356 | MASK_FUSED_MADD)
12357 #undef TARGET_HANDLE_OPTION
12358 #define TARGET_HANDLE_OPTION mips_handle_option
12360 #undef TARGET_FUNCTION_OK_FOR_SIBCALL
12361 #define TARGET_FUNCTION_OK_FOR_SIBCALL mips_function_ok_for_sibcall
12363 #undef TARGET_INSERT_ATTRIBUTES
12364 #define TARGET_INSERT_ATTRIBUTES mips_insert_attributes
12365 #undef TARGET_MERGE_DECL_ATTRIBUTES
12366 #define TARGET_MERGE_DECL_ATTRIBUTES mips_merge_decl_attributes
12367 #undef TARGET_SET_CURRENT_FUNCTION
12368 #define TARGET_SET_CURRENT_FUNCTION mips_set_current_function
12370 #undef TARGET_VALID_POINTER_MODE
12371 #define TARGET_VALID_POINTER_MODE mips_valid_pointer_mode
12372 #undef TARGET_RTX_COSTS
12373 #define TARGET_RTX_COSTS mips_rtx_costs
12374 #undef TARGET_ADDRESS_COST
12375 #define TARGET_ADDRESS_COST mips_address_cost
12377 #undef TARGET_IN_SMALL_DATA_P
12378 #define TARGET_IN_SMALL_DATA_P mips_in_small_data_p
12380 #undef TARGET_MACHINE_DEPENDENT_REORG
12381 #define TARGET_MACHINE_DEPENDENT_REORG mips_reorg
12383 #undef TARGET_ASM_FILE_START
12384 #define TARGET_ASM_FILE_START mips_file_start
12385 #undef TARGET_ASM_FILE_START_FILE_DIRECTIVE
12386 #define TARGET_ASM_FILE_START_FILE_DIRECTIVE true
12388 #undef TARGET_INIT_LIBFUNCS
12389 #define TARGET_INIT_LIBFUNCS mips_init_libfuncs
12391 #undef TARGET_BUILD_BUILTIN_VA_LIST
12392 #define TARGET_BUILD_BUILTIN_VA_LIST mips_build_builtin_va_list
12393 #undef TARGET_GIMPLIFY_VA_ARG_EXPR
12394 #define TARGET_GIMPLIFY_VA_ARG_EXPR mips_gimplify_va_arg_expr
12396 #undef TARGET_PROMOTE_FUNCTION_ARGS
12397 #define TARGET_PROMOTE_FUNCTION_ARGS hook_bool_const_tree_true
12398 #undef TARGET_PROMOTE_FUNCTION_RETURN
12399 #define TARGET_PROMOTE_FUNCTION_RETURN hook_bool_const_tree_true
12400 #undef TARGET_PROMOTE_PROTOTYPES
12401 #define TARGET_PROMOTE_PROTOTYPES hook_bool_const_tree_true
12403 #undef TARGET_RETURN_IN_MEMORY
12404 #define TARGET_RETURN_IN_MEMORY mips_return_in_memory
12405 #undef TARGET_RETURN_IN_MSB
12406 #define TARGET_RETURN_IN_MSB mips_return_in_msb
12408 #undef TARGET_ASM_OUTPUT_MI_THUNK
12409 #define TARGET_ASM_OUTPUT_MI_THUNK mips_output_mi_thunk
12410 #undef TARGET_ASM_CAN_OUTPUT_MI_THUNK
12411 #define TARGET_ASM_CAN_OUTPUT_MI_THUNK hook_bool_const_tree_hwi_hwi_const_tree_true
12413 #undef TARGET_SETUP_INCOMING_VARARGS
12414 #define TARGET_SETUP_INCOMING_VARARGS mips_setup_incoming_varargs
12415 #undef TARGET_STRICT_ARGUMENT_NAMING
12416 #define TARGET_STRICT_ARGUMENT_NAMING mips_strict_argument_naming
12417 #undef TARGET_MUST_PASS_IN_STACK
12418 #define TARGET_MUST_PASS_IN_STACK must_pass_in_stack_var_size
12419 #undef TARGET_PASS_BY_REFERENCE
12420 #define TARGET_PASS_BY_REFERENCE mips_pass_by_reference
12421 #undef TARGET_CALLEE_COPIES
12422 #define TARGET_CALLEE_COPIES mips_callee_copies
12423 #undef TARGET_ARG_PARTIAL_BYTES
12424 #define TARGET_ARG_PARTIAL_BYTES mips_arg_partial_bytes
12426 #undef TARGET_MODE_REP_EXTENDED
12427 #define TARGET_MODE_REP_EXTENDED mips_mode_rep_extended
12429 #undef TARGET_VECTOR_MODE_SUPPORTED_P
12430 #define TARGET_VECTOR_MODE_SUPPORTED_P mips_vector_mode_supported_p
12432 #undef TARGET_SCALAR_MODE_SUPPORTED_P
12433 #define TARGET_SCALAR_MODE_SUPPORTED_P mips_scalar_mode_supported_p
12435 #undef TARGET_INIT_BUILTINS
12436 #define TARGET_INIT_BUILTINS mips_init_builtins
12437 #undef TARGET_EXPAND_BUILTIN
12438 #define TARGET_EXPAND_BUILTIN mips_expand_builtin
12440 #undef TARGET_HAVE_TLS
12441 #define TARGET_HAVE_TLS HAVE_AS_TLS
12443 #undef TARGET_CANNOT_FORCE_CONST_MEM
12444 #define TARGET_CANNOT_FORCE_CONST_MEM mips_cannot_force_const_mem
12446 #undef TARGET_ENCODE_SECTION_INFO
12447 #define TARGET_ENCODE_SECTION_INFO mips_encode_section_info
12449 #undef TARGET_ATTRIBUTE_TABLE
12450 #define TARGET_ATTRIBUTE_TABLE mips_attribute_table
12451 /* All our function attributes are related to how out-of-line copies should
12452 be compiled or called. They don't in themselves prevent inlining. */
12453 #undef TARGET_FUNCTION_ATTRIBUTE_INLINABLE_P
12454 #define TARGET_FUNCTION_ATTRIBUTE_INLINABLE_P hook_bool_const_tree_true
12456 #undef TARGET_EXTRA_LIVE_ON_ENTRY
12457 #define TARGET_EXTRA_LIVE_ON_ENTRY mips_extra_live_on_entry
12459 #undef TARGET_USE_BLOCKS_FOR_CONSTANT_P
12460 #define TARGET_USE_BLOCKS_FOR_CONSTANT_P mips_use_blocks_for_constant_p
12461 #undef TARGET_USE_ANCHORS_FOR_SYMBOL_P
12462 #define TARGET_USE_ANCHORS_FOR_SYMBOL_P mips_use_anchors_for_symbol_p
12464 #undef TARGET_COMP_TYPE_ATTRIBUTES
12465 #define TARGET_COMP_TYPE_ATTRIBUTES mips_comp_type_attributes
12467 #ifdef HAVE_AS_DTPRELWORD
12468 #undef TARGET_ASM_OUTPUT_DWARF_DTPREL
12469 #define TARGET_ASM_OUTPUT_DWARF_DTPREL mips_output_dwarf_dtprel
12470 #endif
12471 #undef TARGET_DWARF_REGISTER_SPAN
12472 #define TARGET_DWARF_REGISTER_SPAN mips_dwarf_register_span
12474 struct gcc_target targetm = TARGET_INITIALIZER;
12476 #include "gt-mips.h"