1 /* Subroutines used for code generation on the DEC Alpha.
2 Copyright (C) 1992, 1993, 1994, 1995, 1996, 1997, 1998, 1999, 2000, 2001,
3 2002, 2003, 2004, 2005, 2006, 2007, 2008, 2009, 2010
4 Free Software Foundation, Inc.
5 Contributed by Richard Kenner (kenner@vlsi1.ultra.nyu.edu)
7 This file is part of GCC.
9 GCC is free software; you can redistribute it and/or modify
10 it under the terms of the GNU General Public License as published by
11 the Free Software Foundation; either version 3, or (at your option)
14 GCC is distributed in the hope that it will be useful,
15 but WITHOUT ANY WARRANTY; without even the implied warranty of
16 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 GNU General Public License for more details.
19 You should have received a copy of the GNU General Public License
20 along with GCC; see the file COPYING3. If not see
21 <http://www.gnu.org/licenses/>. */
26 #include "coretypes.h"
31 #include "hard-reg-set.h"
33 #include "insn-config.h"
34 #include "conditions.h"
36 #include "insn-attr.h"
47 #include "integrate.h"
50 #include "target-def.h"
52 #include "langhooks.h"
53 #include <splay-tree.h>
54 #include "cfglayout.h"
56 #include "tree-flow.h"
57 #include "tree-stdarg.h"
58 #include "tm-constrs.h"
62 /* Specify which cpu to schedule for. */
63 enum processor_type alpha_tune
;
65 /* Which cpu we're generating code for. */
66 enum processor_type alpha_cpu
;
68 static const char * const alpha_cpu_name
[] =
73 /* Specify how accurate floating-point traps need to be. */
75 enum alpha_trap_precision alpha_tp
;
77 /* Specify the floating-point rounding mode. */
79 enum alpha_fp_rounding_mode alpha_fprm
;
81 /* Specify which things cause traps. */
83 enum alpha_fp_trap_mode alpha_fptm
;
85 /* Nonzero if inside of a function, because the Alpha asm can't
86 handle .files inside of functions. */
88 static int inside_function
= FALSE
;
90 /* The number of cycles of latency we should assume on memory reads. */
92 int alpha_memory_latency
= 3;
94 /* Whether the function needs the GP. */
96 static int alpha_function_needs_gp
;
98 /* The alias set for prologue/epilogue register save/restore. */
100 static GTY(()) alias_set_type alpha_sr_alias_set
;
102 /* The assembler name of the current function. */
104 static const char *alpha_fnname
;
106 /* The next explicit relocation sequence number. */
107 extern GTY(()) int alpha_next_sequence_number
;
108 int alpha_next_sequence_number
= 1;
110 /* The literal and gpdisp sequence numbers for this insn, as printed
111 by %# and %* respectively. */
112 extern GTY(()) int alpha_this_literal_sequence_number
;
113 extern GTY(()) int alpha_this_gpdisp_sequence_number
;
114 int alpha_this_literal_sequence_number
;
115 int alpha_this_gpdisp_sequence_number
;
117 /* Costs of various operations on the different architectures. */
119 struct alpha_rtx_cost_data
121 unsigned char fp_add
;
122 unsigned char fp_mult
;
123 unsigned char fp_div_sf
;
124 unsigned char fp_div_df
;
125 unsigned char int_mult_si
;
126 unsigned char int_mult_di
;
127 unsigned char int_shift
;
128 unsigned char int_cmov
;
129 unsigned short int_div
;
132 static struct alpha_rtx_cost_data
const alpha_rtx_cost_data
[PROCESSOR_MAX
] =
135 COSTS_N_INSNS (6), /* fp_add */
136 COSTS_N_INSNS (6), /* fp_mult */
137 COSTS_N_INSNS (34), /* fp_div_sf */
138 COSTS_N_INSNS (63), /* fp_div_df */
139 COSTS_N_INSNS (23), /* int_mult_si */
140 COSTS_N_INSNS (23), /* int_mult_di */
141 COSTS_N_INSNS (2), /* int_shift */
142 COSTS_N_INSNS (2), /* int_cmov */
143 COSTS_N_INSNS (97), /* int_div */
146 COSTS_N_INSNS (4), /* fp_add */
147 COSTS_N_INSNS (4), /* fp_mult */
148 COSTS_N_INSNS (15), /* fp_div_sf */
149 COSTS_N_INSNS (22), /* fp_div_df */
150 COSTS_N_INSNS (8), /* int_mult_si */
151 COSTS_N_INSNS (12), /* int_mult_di */
152 COSTS_N_INSNS (1) + 1, /* int_shift */
153 COSTS_N_INSNS (1), /* int_cmov */
154 COSTS_N_INSNS (83), /* int_div */
157 COSTS_N_INSNS (4), /* fp_add */
158 COSTS_N_INSNS (4), /* fp_mult */
159 COSTS_N_INSNS (12), /* fp_div_sf */
160 COSTS_N_INSNS (15), /* fp_div_df */
161 COSTS_N_INSNS (7), /* int_mult_si */
162 COSTS_N_INSNS (7), /* int_mult_di */
163 COSTS_N_INSNS (1), /* int_shift */
164 COSTS_N_INSNS (2), /* int_cmov */
165 COSTS_N_INSNS (86), /* int_div */
169 /* Similar but tuned for code size instead of execution latency. The
170 extra +N is fractional cost tuning based on latency. It's used to
171 encourage use of cheaper insns like shift, but only if there's just
174 static struct alpha_rtx_cost_data
const alpha_rtx_cost_size
=
176 COSTS_N_INSNS (1), /* fp_add */
177 COSTS_N_INSNS (1), /* fp_mult */
178 COSTS_N_INSNS (1), /* fp_div_sf */
179 COSTS_N_INSNS (1) + 1, /* fp_div_df */
180 COSTS_N_INSNS (1) + 1, /* int_mult_si */
181 COSTS_N_INSNS (1) + 2, /* int_mult_di */
182 COSTS_N_INSNS (1), /* int_shift */
183 COSTS_N_INSNS (1), /* int_cmov */
184 COSTS_N_INSNS (6), /* int_div */
187 /* Get the number of args of a function in one of two ways. */
188 #if TARGET_ABI_OPEN_VMS || TARGET_ABI_UNICOSMK
189 #define NUM_ARGS crtl->args.info.num_args
191 #define NUM_ARGS crtl->args.info
197 /* Declarations of static functions. */
198 static struct machine_function
*alpha_init_machine_status (void);
199 static rtx
alpha_emit_xfloating_compare (enum rtx_code
*, rtx
, rtx
);
201 #if TARGET_ABI_OPEN_VMS
202 static void alpha_write_linkage (FILE *, const char *, tree
);
203 static bool vms_valid_pointer_mode (enum machine_mode
);
206 static void unicosmk_output_deferred_case_vectors (FILE *);
207 static void unicosmk_gen_dsib (unsigned long *);
208 static void unicosmk_output_ssib (FILE *, const char *);
209 static int unicosmk_need_dex (rtx
);
211 /* Implement TARGET_HANDLE_OPTION. */
214 alpha_handle_option (size_t code
, const char *arg
, int value
)
220 target_flags
|= MASK_SOFT_FP
;
224 case OPT_mieee_with_inexact
:
225 target_flags
|= MASK_IEEE_CONFORMANT
;
229 if (value
!= 16 && value
!= 32 && value
!= 64)
230 error ("bad value %qs for -mtls-size switch", arg
);
237 #ifdef TARGET_ALTERNATE_LONG_DOUBLE_MANGLING
238 /* Implement TARGET_MANGLE_TYPE. */
241 alpha_mangle_type (const_tree type
)
243 if (TYPE_MAIN_VARIANT (type
) == long_double_type_node
244 && TARGET_LONG_DOUBLE_128
)
247 /* For all other types, use normal C++ mangling. */
252 /* Parse target option strings. */
255 override_options (void)
257 static const struct cpu_table
{
258 const char *const name
;
259 const enum processor_type processor
;
262 { "ev4", PROCESSOR_EV4
, 0 },
263 { "ev45", PROCESSOR_EV4
, 0 },
264 { "21064", PROCESSOR_EV4
, 0 },
265 { "ev5", PROCESSOR_EV5
, 0 },
266 { "21164", PROCESSOR_EV5
, 0 },
267 { "ev56", PROCESSOR_EV5
, MASK_BWX
},
268 { "21164a", PROCESSOR_EV5
, MASK_BWX
},
269 { "pca56", PROCESSOR_EV5
, MASK_BWX
|MASK_MAX
},
270 { "21164PC",PROCESSOR_EV5
, MASK_BWX
|MASK_MAX
},
271 { "21164pc",PROCESSOR_EV5
, MASK_BWX
|MASK_MAX
},
272 { "ev6", PROCESSOR_EV6
, MASK_BWX
|MASK_MAX
|MASK_FIX
},
273 { "21264", PROCESSOR_EV6
, MASK_BWX
|MASK_MAX
|MASK_FIX
},
274 { "ev67", PROCESSOR_EV6
, MASK_BWX
|MASK_MAX
|MASK_FIX
|MASK_CIX
},
275 { "21264a", PROCESSOR_EV6
, MASK_BWX
|MASK_MAX
|MASK_FIX
|MASK_CIX
}
278 int const ct_size
= ARRAY_SIZE (cpu_table
);
281 /* Unicos/Mk doesn't have shared libraries. */
282 if (TARGET_ABI_UNICOSMK
&& flag_pic
)
284 warning (0, "-f%s ignored for Unicos/Mk (not supported)",
285 (flag_pic
> 1) ? "PIC" : "pic");
289 /* On Unicos/Mk, the native compiler consistently generates /d suffices for
290 floating-point instructions. Make that the default for this target. */
291 if (TARGET_ABI_UNICOSMK
)
292 alpha_fprm
= ALPHA_FPRM_DYN
;
294 alpha_fprm
= ALPHA_FPRM_NORM
;
296 alpha_tp
= ALPHA_TP_PROG
;
297 alpha_fptm
= ALPHA_FPTM_N
;
299 /* We cannot use su and sui qualifiers for conversion instructions on
300 Unicos/Mk. I'm not sure if this is due to assembler or hardware
301 limitations. Right now, we issue a warning if -mieee is specified
302 and then ignore it; eventually, we should either get it right or
303 disable the option altogether. */
307 if (TARGET_ABI_UNICOSMK
)
308 warning (0, "-mieee not supported on Unicos/Mk");
311 alpha_tp
= ALPHA_TP_INSN
;
312 alpha_fptm
= ALPHA_FPTM_SU
;
316 if (TARGET_IEEE_WITH_INEXACT
)
318 if (TARGET_ABI_UNICOSMK
)
319 warning (0, "-mieee-with-inexact not supported on Unicos/Mk");
322 alpha_tp
= ALPHA_TP_INSN
;
323 alpha_fptm
= ALPHA_FPTM_SUI
;
329 if (! strcmp (alpha_tp_string
, "p"))
330 alpha_tp
= ALPHA_TP_PROG
;
331 else if (! strcmp (alpha_tp_string
, "f"))
332 alpha_tp
= ALPHA_TP_FUNC
;
333 else if (! strcmp (alpha_tp_string
, "i"))
334 alpha_tp
= ALPHA_TP_INSN
;
336 error ("bad value %qs for -mtrap-precision switch", alpha_tp_string
);
339 if (alpha_fprm_string
)
341 if (! strcmp (alpha_fprm_string
, "n"))
342 alpha_fprm
= ALPHA_FPRM_NORM
;
343 else if (! strcmp (alpha_fprm_string
, "m"))
344 alpha_fprm
= ALPHA_FPRM_MINF
;
345 else if (! strcmp (alpha_fprm_string
, "c"))
346 alpha_fprm
= ALPHA_FPRM_CHOP
;
347 else if (! strcmp (alpha_fprm_string
,"d"))
348 alpha_fprm
= ALPHA_FPRM_DYN
;
350 error ("bad value %qs for -mfp-rounding-mode switch",
354 if (alpha_fptm_string
)
356 if (strcmp (alpha_fptm_string
, "n") == 0)
357 alpha_fptm
= ALPHA_FPTM_N
;
358 else if (strcmp (alpha_fptm_string
, "u") == 0)
359 alpha_fptm
= ALPHA_FPTM_U
;
360 else if (strcmp (alpha_fptm_string
, "su") == 0)
361 alpha_fptm
= ALPHA_FPTM_SU
;
362 else if (strcmp (alpha_fptm_string
, "sui") == 0)
363 alpha_fptm
= ALPHA_FPTM_SUI
;
365 error ("bad value %qs for -mfp-trap-mode switch", alpha_fptm_string
);
368 if (alpha_cpu_string
)
370 for (i
= 0; i
< ct_size
; i
++)
371 if (! strcmp (alpha_cpu_string
, cpu_table
[i
].name
))
373 alpha_tune
= alpha_cpu
= cpu_table
[i
].processor
;
374 target_flags
&= ~ (MASK_BWX
| MASK_MAX
| MASK_FIX
| MASK_CIX
);
375 target_flags
|= cpu_table
[i
].flags
;
379 error ("bad value %qs for -mcpu switch", alpha_cpu_string
);
382 if (alpha_tune_string
)
384 for (i
= 0; i
< ct_size
; i
++)
385 if (! strcmp (alpha_tune_string
, cpu_table
[i
].name
))
387 alpha_tune
= cpu_table
[i
].processor
;
391 error ("bad value %qs for -mtune switch", alpha_tune_string
);
394 /* Do some sanity checks on the above options. */
396 if (TARGET_ABI_UNICOSMK
&& alpha_fptm
!= ALPHA_FPTM_N
)
398 warning (0, "trap mode not supported on Unicos/Mk");
399 alpha_fptm
= ALPHA_FPTM_N
;
402 if ((alpha_fptm
== ALPHA_FPTM_SU
|| alpha_fptm
== ALPHA_FPTM_SUI
)
403 && alpha_tp
!= ALPHA_TP_INSN
&& alpha_cpu
!= PROCESSOR_EV6
)
405 warning (0, "fp software completion requires -mtrap-precision=i");
406 alpha_tp
= ALPHA_TP_INSN
;
409 if (alpha_cpu
== PROCESSOR_EV6
)
411 /* Except for EV6 pass 1 (not released), we always have precise
412 arithmetic traps. Which means we can do software completion
413 without minding trap shadows. */
414 alpha_tp
= ALPHA_TP_PROG
;
417 if (TARGET_FLOAT_VAX
)
419 if (alpha_fprm
== ALPHA_FPRM_MINF
|| alpha_fprm
== ALPHA_FPRM_DYN
)
421 warning (0, "rounding mode not supported for VAX floats");
422 alpha_fprm
= ALPHA_FPRM_NORM
;
424 if (alpha_fptm
== ALPHA_FPTM_SUI
)
426 warning (0, "trap mode not supported for VAX floats");
427 alpha_fptm
= ALPHA_FPTM_SU
;
429 if (target_flags_explicit
& MASK_LONG_DOUBLE_128
)
430 warning (0, "128-bit long double not supported for VAX floats");
431 target_flags
&= ~MASK_LONG_DOUBLE_128
;
438 if (!alpha_mlat_string
)
439 alpha_mlat_string
= "L1";
441 if (ISDIGIT ((unsigned char)alpha_mlat_string
[0])
442 && (lat
= strtol (alpha_mlat_string
, &end
, 10), *end
== '\0'))
444 else if ((alpha_mlat_string
[0] == 'L' || alpha_mlat_string
[0] == 'l')
445 && ISDIGIT ((unsigned char)alpha_mlat_string
[1])
446 && alpha_mlat_string
[2] == '\0')
448 static int const cache_latency
[][4] =
450 { 3, 30, -1 }, /* ev4 -- Bcache is a guess */
451 { 2, 12, 38 }, /* ev5 -- Bcache from PC164 LMbench numbers */
452 { 3, 12, 30 }, /* ev6 -- Bcache from DS20 LMbench. */
455 lat
= alpha_mlat_string
[1] - '0';
456 if (lat
<= 0 || lat
> 3 || cache_latency
[alpha_tune
][lat
-1] == -1)
458 warning (0, "L%d cache latency unknown for %s",
459 lat
, alpha_cpu_name
[alpha_tune
]);
463 lat
= cache_latency
[alpha_tune
][lat
-1];
465 else if (! strcmp (alpha_mlat_string
, "main"))
467 /* Most current memories have about 370ns latency. This is
468 a reasonable guess for a fast cpu. */
473 warning (0, "bad value %qs for -mmemory-latency", alpha_mlat_string
);
477 alpha_memory_latency
= lat
;
480 /* Default the definition of "small data" to 8 bytes. */
484 /* Infer TARGET_SMALL_DATA from -fpic/-fPIC. */
486 target_flags
|= MASK_SMALL_DATA
;
487 else if (flag_pic
== 2)
488 target_flags
&= ~MASK_SMALL_DATA
;
490 /* Align labels and loops for optimal branching. */
491 /* ??? Kludge these by not doing anything if we don't optimize and also if
492 we are writing ECOFF symbols to work around a bug in DEC's assembler. */
493 if (optimize
> 0 && write_symbols
!= SDB_DEBUG
)
495 if (align_loops
<= 0)
497 if (align_jumps
<= 0)
500 if (align_functions
<= 0)
501 align_functions
= 16;
503 /* Acquire a unique set number for our register saves and restores. */
504 alpha_sr_alias_set
= new_alias_set ();
506 /* Register variables and functions with the garbage collector. */
508 /* Set up function hooks. */
509 init_machine_status
= alpha_init_machine_status
;
511 /* Tell the compiler when we're using VAX floating point. */
512 if (TARGET_FLOAT_VAX
)
514 REAL_MODE_FORMAT (SFmode
) = &vax_f_format
;
515 REAL_MODE_FORMAT (DFmode
) = &vax_g_format
;
516 REAL_MODE_FORMAT (TFmode
) = NULL
;
519 #ifdef TARGET_DEFAULT_LONG_DOUBLE_128
520 if (!(target_flags_explicit
& MASK_LONG_DOUBLE_128
))
521 target_flags
|= MASK_LONG_DOUBLE_128
;
524 /* If using typedef char *va_list, signal that __builtin_va_start (&ap, 0)
525 can be optimized to ap = __builtin_next_arg (0). */
526 if (TARGET_ABI_UNICOSMK
)
527 targetm
.expand_builtin_va_start
= NULL
;
530 /* Returns 1 if VALUE is a mask that contains full bytes of zero or ones. */
533 zap_mask (HOST_WIDE_INT value
)
537 for (i
= 0; i
< HOST_BITS_PER_WIDE_INT
/ HOST_BITS_PER_CHAR
;
539 if ((value
& 0xff) != 0 && (value
& 0xff) != 0xff)
545 /* Return true if OP is valid for a particular TLS relocation.
546 We are already guaranteed that OP is a CONST. */
549 tls_symbolic_operand_1 (rtx op
, int size
, int unspec
)
553 if (GET_CODE (op
) != UNSPEC
|| XINT (op
, 1) != unspec
)
555 op
= XVECEXP (op
, 0, 0);
557 if (GET_CODE (op
) != SYMBOL_REF
)
560 switch (SYMBOL_REF_TLS_MODEL (op
))
562 case TLS_MODEL_LOCAL_DYNAMIC
:
563 return unspec
== UNSPEC_DTPREL
&& size
== alpha_tls_size
;
564 case TLS_MODEL_INITIAL_EXEC
:
565 return unspec
== UNSPEC_TPREL
&& size
== 64;
566 case TLS_MODEL_LOCAL_EXEC
:
567 return unspec
== UNSPEC_TPREL
&& size
== alpha_tls_size
;
573 /* Used by aligned_memory_operand and unaligned_memory_operand to
574 resolve what reload is going to do with OP if it's a register. */
577 resolve_reload_operand (rtx op
)
579 if (reload_in_progress
)
582 if (GET_CODE (tmp
) == SUBREG
)
583 tmp
= SUBREG_REG (tmp
);
585 && REGNO (tmp
) >= FIRST_PSEUDO_REGISTER
)
587 op
= reg_equiv_memory_loc
[REGNO (tmp
)];
595 /* The scalar modes supported differs from the default check-what-c-supports
596 version in that sometimes TFmode is available even when long double
597 indicates only DFmode. On unicosmk, we have the situation that HImode
598 doesn't map to any C type, but of course we still support that. */
601 alpha_scalar_mode_supported_p (enum machine_mode mode
)
609 case TImode
: /* via optabs.c */
617 return TARGET_HAS_XFLOATING_LIBS
;
624 /* Alpha implements a couple of integer vector mode operations when
625 TARGET_MAX is enabled. We do not check TARGET_MAX here, however,
626 which allows the vectorizer to operate on e.g. move instructions,
627 or when expand_vector_operations can do something useful. */
630 alpha_vector_mode_supported_p (enum machine_mode mode
)
632 return mode
== V8QImode
|| mode
== V4HImode
|| mode
== V2SImode
;
635 /* Return 1 if this function can directly return via $26. */
640 return (! TARGET_ABI_OPEN_VMS
&& ! TARGET_ABI_UNICOSMK
642 && alpha_sa_size () == 0
643 && get_frame_size () == 0
644 && crtl
->outgoing_args_size
== 0
645 && crtl
->args
.pretend_args_size
== 0);
648 /* Return the ADDR_VEC associated with a tablejump insn. */
651 alpha_tablejump_addr_vec (rtx insn
)
655 tmp
= JUMP_LABEL (insn
);
658 tmp
= NEXT_INSN (tmp
);
662 && GET_CODE (PATTERN (tmp
)) == ADDR_DIFF_VEC
)
663 return PATTERN (tmp
);
667 /* Return the label of the predicted edge, or CONST0_RTX if we don't know. */
670 alpha_tablejump_best_label (rtx insn
)
672 rtx jump_table
= alpha_tablejump_addr_vec (insn
);
673 rtx best_label
= NULL_RTX
;
675 /* ??? Once the CFG doesn't keep getting completely rebuilt, look
676 there for edge frequency counts from profile data. */
680 int n_labels
= XVECLEN (jump_table
, 1);
684 for (i
= 0; i
< n_labels
; i
++)
688 for (j
= i
+ 1; j
< n_labels
; j
++)
689 if (XEXP (XVECEXP (jump_table
, 1, i
), 0)
690 == XEXP (XVECEXP (jump_table
, 1, j
), 0))
693 if (count
> best_count
)
694 best_count
= count
, best_label
= XVECEXP (jump_table
, 1, i
);
698 return best_label
? best_label
: const0_rtx
;
701 /* Return the TLS model to use for SYMBOL. */
703 static enum tls_model
704 tls_symbolic_operand_type (rtx symbol
)
706 enum tls_model model
;
708 if (GET_CODE (symbol
) != SYMBOL_REF
)
709 return TLS_MODEL_NONE
;
710 model
= SYMBOL_REF_TLS_MODEL (symbol
);
712 /* Local-exec with a 64-bit size is the same code as initial-exec. */
713 if (model
== TLS_MODEL_LOCAL_EXEC
&& alpha_tls_size
== 64)
714 model
= TLS_MODEL_INITIAL_EXEC
;
719 /* Return true if the function DECL will share the same GP as any
720 function in the current unit of translation. */
723 decl_has_samegp (const_tree decl
)
725 /* Functions that are not local can be overridden, and thus may
726 not share the same gp. */
727 if (!(*targetm
.binds_local_p
) (decl
))
730 /* If -msmall-data is in effect, assume that there is only one GP
731 for the module, and so any local symbol has this property. We
732 need explicit relocations to be able to enforce this for symbols
733 not defined in this unit of translation, however. */
734 if (TARGET_EXPLICIT_RELOCS
&& TARGET_SMALL_DATA
)
737 /* Functions that are not external are defined in this UoT. */
738 /* ??? Irritatingly, static functions not yet emitted are still
739 marked "external". Apply this to non-static functions only. */
740 return !TREE_PUBLIC (decl
) || !DECL_EXTERNAL (decl
);
743 /* Return true if EXP should be placed in the small data section. */
746 alpha_in_small_data_p (const_tree exp
)
748 /* We want to merge strings, so we never consider them small data. */
749 if (TREE_CODE (exp
) == STRING_CST
)
752 /* Functions are never in the small data area. Duh. */
753 if (TREE_CODE (exp
) == FUNCTION_DECL
)
756 if (TREE_CODE (exp
) == VAR_DECL
&& DECL_SECTION_NAME (exp
))
758 const char *section
= TREE_STRING_POINTER (DECL_SECTION_NAME (exp
));
759 if (strcmp (section
, ".sdata") == 0
760 || strcmp (section
, ".sbss") == 0)
765 HOST_WIDE_INT size
= int_size_in_bytes (TREE_TYPE (exp
));
767 /* If this is an incomplete type with size 0, then we can't put it
768 in sdata because it might be too big when completed. */
769 if (size
> 0 && (unsigned HOST_WIDE_INT
) size
<= g_switch_value
)
776 #if TARGET_ABI_OPEN_VMS
778 vms_valid_pointer_mode (enum machine_mode mode
)
780 return (mode
== SImode
|| mode
== DImode
);
784 alpha_linkage_symbol_p (const char *symname
)
786 int symlen
= strlen (symname
);
789 return strcmp (&symname
[symlen
- 4], "..lk") == 0;
794 #define LINKAGE_SYMBOL_REF_P(X) \
795 ((GET_CODE (X) == SYMBOL_REF \
796 && alpha_linkage_symbol_p (XSTR (X, 0))) \
797 || (GET_CODE (X) == CONST \
798 && GET_CODE (XEXP (X, 0)) == PLUS \
799 && GET_CODE (XEXP (XEXP (X, 0), 0)) == SYMBOL_REF \
800 && alpha_linkage_symbol_p (XSTR (XEXP (XEXP (X, 0), 0), 0))))
803 /* legitimate_address_p recognizes an RTL expression that is a valid
804 memory address for an instruction. The MODE argument is the
805 machine mode for the MEM expression that wants to use this address.
807 For Alpha, we have either a constant address or the sum of a
808 register and a constant address, or just a register. For DImode,
809 any of those forms can be surrounded with an AND that clear the
810 low-order three bits; this is an "unaligned" access. */
813 alpha_legitimate_address_p (enum machine_mode mode
, rtx x
, bool strict
)
815 /* If this is an ldq_u type address, discard the outer AND. */
817 && GET_CODE (x
) == AND
818 && CONST_INT_P (XEXP (x
, 1))
819 && INTVAL (XEXP (x
, 1)) == -8)
822 /* Discard non-paradoxical subregs. */
823 if (GET_CODE (x
) == SUBREG
824 && (GET_MODE_SIZE (GET_MODE (x
))
825 < GET_MODE_SIZE (GET_MODE (SUBREG_REG (x
)))))
828 /* Unadorned general registers are valid. */
831 ? STRICT_REG_OK_FOR_BASE_P (x
)
832 : NONSTRICT_REG_OK_FOR_BASE_P (x
)))
835 /* Constant addresses (i.e. +/- 32k) are valid. */
836 if (CONSTANT_ADDRESS_P (x
))
839 #if TARGET_ABI_OPEN_VMS
840 if (LINKAGE_SYMBOL_REF_P (x
))
844 /* Register plus a small constant offset is valid. */
845 if (GET_CODE (x
) == PLUS
)
847 rtx ofs
= XEXP (x
, 1);
850 /* Discard non-paradoxical subregs. */
851 if (GET_CODE (x
) == SUBREG
852 && (GET_MODE_SIZE (GET_MODE (x
))
853 < GET_MODE_SIZE (GET_MODE (SUBREG_REG (x
)))))
859 && NONSTRICT_REG_OK_FP_BASE_P (x
)
860 && CONST_INT_P (ofs
))
863 ? STRICT_REG_OK_FOR_BASE_P (x
)
864 : NONSTRICT_REG_OK_FOR_BASE_P (x
))
865 && CONSTANT_ADDRESS_P (ofs
))
870 /* If we're managing explicit relocations, LO_SUM is valid, as are small
871 data symbols. Avoid explicit relocations of modes larger than word
872 mode since i.e. $LC0+8($1) can fold around +/- 32k offset. */
873 else if (TARGET_EXPLICIT_RELOCS
874 && GET_MODE_SIZE (mode
) <= UNITS_PER_WORD
)
876 if (small_symbolic_operand (x
, Pmode
))
879 if (GET_CODE (x
) == LO_SUM
)
881 rtx ofs
= XEXP (x
, 1);
884 /* Discard non-paradoxical subregs. */
885 if (GET_CODE (x
) == SUBREG
886 && (GET_MODE_SIZE (GET_MODE (x
))
887 < GET_MODE_SIZE (GET_MODE (SUBREG_REG (x
)))))
890 /* Must have a valid base register. */
893 ? STRICT_REG_OK_FOR_BASE_P (x
)
894 : NONSTRICT_REG_OK_FOR_BASE_P (x
))))
897 /* The symbol must be local. */
898 if (local_symbolic_operand (ofs
, Pmode
)
899 || dtp32_symbolic_operand (ofs
, Pmode
)
900 || tp32_symbolic_operand (ofs
, Pmode
))
908 /* Build the SYMBOL_REF for __tls_get_addr. */
910 static GTY(()) rtx tls_get_addr_libfunc
;
913 get_tls_get_addr (void)
915 if (!tls_get_addr_libfunc
)
916 tls_get_addr_libfunc
= init_one_libfunc ("__tls_get_addr");
917 return tls_get_addr_libfunc
;
920 /* Try machine-dependent ways of modifying an illegitimate address
921 to be legitimate. If we find one, return the new, valid address. */
924 alpha_legitimize_address_1 (rtx x
, rtx scratch
, enum machine_mode mode
)
926 HOST_WIDE_INT addend
;
928 /* If the address is (plus reg const_int) and the CONST_INT is not a
929 valid offset, compute the high part of the constant and add it to
930 the register. Then our address is (plus temp low-part-const). */
931 if (GET_CODE (x
) == PLUS
932 && REG_P (XEXP (x
, 0))
933 && CONST_INT_P (XEXP (x
, 1))
934 && ! CONSTANT_ADDRESS_P (XEXP (x
, 1)))
936 addend
= INTVAL (XEXP (x
, 1));
941 /* If the address is (const (plus FOO const_int)), find the low-order
942 part of the CONST_INT. Then load FOO plus any high-order part of the
943 CONST_INT into a register. Our address is (plus reg low-part-const).
944 This is done to reduce the number of GOT entries. */
945 if (can_create_pseudo_p ()
946 && GET_CODE (x
) == CONST
947 && GET_CODE (XEXP (x
, 0)) == PLUS
948 && CONST_INT_P (XEXP (XEXP (x
, 0), 1)))
950 addend
= INTVAL (XEXP (XEXP (x
, 0), 1));
951 x
= force_reg (Pmode
, XEXP (XEXP (x
, 0), 0));
955 /* If we have a (plus reg const), emit the load as in (2), then add
956 the two registers, and finally generate (plus reg low-part-const) as
958 if (can_create_pseudo_p ()
959 && GET_CODE (x
) == PLUS
960 && REG_P (XEXP (x
, 0))
961 && GET_CODE (XEXP (x
, 1)) == CONST
962 && GET_CODE (XEXP (XEXP (x
, 1), 0)) == PLUS
963 && CONST_INT_P (XEXP (XEXP (XEXP (x
, 1), 0), 1)))
965 addend
= INTVAL (XEXP (XEXP (XEXP (x
, 1), 0), 1));
966 x
= expand_simple_binop (Pmode
, PLUS
, XEXP (x
, 0),
967 XEXP (XEXP (XEXP (x
, 1), 0), 0),
968 NULL_RTX
, 1, OPTAB_LIB_WIDEN
);
972 /* If this is a local symbol, split the address into HIGH/LO_SUM parts.
973 Avoid modes larger than word mode since i.e. $LC0+8($1) can fold
974 around +/- 32k offset. */
975 if (TARGET_EXPLICIT_RELOCS
976 && GET_MODE_SIZE (mode
) <= UNITS_PER_WORD
977 && symbolic_operand (x
, Pmode
))
979 rtx r0
, r16
, eqv
, tga
, tp
, insn
, dest
, seq
;
981 switch (tls_symbolic_operand_type (x
))
986 case TLS_MODEL_GLOBAL_DYNAMIC
:
989 r0
= gen_rtx_REG (Pmode
, 0);
990 r16
= gen_rtx_REG (Pmode
, 16);
991 tga
= get_tls_get_addr ();
992 dest
= gen_reg_rtx (Pmode
);
993 seq
= GEN_INT (alpha_next_sequence_number
++);
995 emit_insn (gen_movdi_er_tlsgd (r16
, pic_offset_table_rtx
, x
, seq
));
996 insn
= gen_call_value_osf_tlsgd (r0
, tga
, seq
);
997 insn
= emit_call_insn (insn
);
998 RTL_CONST_CALL_P (insn
) = 1;
999 use_reg (&CALL_INSN_FUNCTION_USAGE (insn
), r16
);
1001 insn
= get_insns ();
1004 emit_libcall_block (insn
, dest
, r0
, x
);
1007 case TLS_MODEL_LOCAL_DYNAMIC
:
1010 r0
= gen_rtx_REG (Pmode
, 0);
1011 r16
= gen_rtx_REG (Pmode
, 16);
1012 tga
= get_tls_get_addr ();
1013 scratch
= gen_reg_rtx (Pmode
);
1014 seq
= GEN_INT (alpha_next_sequence_number
++);
1016 emit_insn (gen_movdi_er_tlsldm (r16
, pic_offset_table_rtx
, seq
));
1017 insn
= gen_call_value_osf_tlsldm (r0
, tga
, seq
);
1018 insn
= emit_call_insn (insn
);
1019 RTL_CONST_CALL_P (insn
) = 1;
1020 use_reg (&CALL_INSN_FUNCTION_USAGE (insn
), r16
);
1022 insn
= get_insns ();
1025 eqv
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, const0_rtx
),
1026 UNSPEC_TLSLDM_CALL
);
1027 emit_libcall_block (insn
, scratch
, r0
, eqv
);
1029 eqv
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, x
), UNSPEC_DTPREL
);
1030 eqv
= gen_rtx_CONST (Pmode
, eqv
);
1032 if (alpha_tls_size
== 64)
1034 dest
= gen_reg_rtx (Pmode
);
1035 emit_insn (gen_rtx_SET (VOIDmode
, dest
, eqv
));
1036 emit_insn (gen_adddi3 (dest
, dest
, scratch
));
1039 if (alpha_tls_size
== 32)
1041 insn
= gen_rtx_HIGH (Pmode
, eqv
);
1042 insn
= gen_rtx_PLUS (Pmode
, scratch
, insn
);
1043 scratch
= gen_reg_rtx (Pmode
);
1044 emit_insn (gen_rtx_SET (VOIDmode
, scratch
, insn
));
1046 return gen_rtx_LO_SUM (Pmode
, scratch
, eqv
);
1048 case TLS_MODEL_INITIAL_EXEC
:
1049 eqv
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, x
), UNSPEC_TPREL
);
1050 eqv
= gen_rtx_CONST (Pmode
, eqv
);
1051 tp
= gen_reg_rtx (Pmode
);
1052 scratch
= gen_reg_rtx (Pmode
);
1053 dest
= gen_reg_rtx (Pmode
);
1055 emit_insn (gen_load_tp (tp
));
1056 emit_insn (gen_rtx_SET (VOIDmode
, scratch
, eqv
));
1057 emit_insn (gen_adddi3 (dest
, tp
, scratch
));
1060 case TLS_MODEL_LOCAL_EXEC
:
1061 eqv
= gen_rtx_UNSPEC (Pmode
, gen_rtvec (1, x
), UNSPEC_TPREL
);
1062 eqv
= gen_rtx_CONST (Pmode
, eqv
);
1063 tp
= gen_reg_rtx (Pmode
);
1065 emit_insn (gen_load_tp (tp
));
1066 if (alpha_tls_size
== 32)
1068 insn
= gen_rtx_HIGH (Pmode
, eqv
);
1069 insn
= gen_rtx_PLUS (Pmode
, tp
, insn
);
1070 tp
= gen_reg_rtx (Pmode
);
1071 emit_insn (gen_rtx_SET (VOIDmode
, tp
, insn
));
1073 return gen_rtx_LO_SUM (Pmode
, tp
, eqv
);
1079 if (local_symbolic_operand (x
, Pmode
))
1081 if (small_symbolic_operand (x
, Pmode
))
1085 if (can_create_pseudo_p ())
1086 scratch
= gen_reg_rtx (Pmode
);
1087 emit_insn (gen_rtx_SET (VOIDmode
, scratch
,
1088 gen_rtx_HIGH (Pmode
, x
)));
1089 return gen_rtx_LO_SUM (Pmode
, scratch
, x
);
1098 HOST_WIDE_INT low
, high
;
1100 low
= ((addend
& 0xffff) ^ 0x8000) - 0x8000;
1102 high
= ((addend
& 0xffffffff) ^ 0x80000000) - 0x80000000;
1106 x
= expand_simple_binop (Pmode
, PLUS
, x
, GEN_INT (addend
),
1107 (!can_create_pseudo_p () ? scratch
: NULL_RTX
),
1108 1, OPTAB_LIB_WIDEN
);
1110 x
= expand_simple_binop (Pmode
, PLUS
, x
, GEN_INT (high
),
1111 (!can_create_pseudo_p () ? scratch
: NULL_RTX
),
1112 1, OPTAB_LIB_WIDEN
);
1114 return plus_constant (x
, low
);
1119 /* Try machine-dependent ways of modifying an illegitimate address
1120 to be legitimate. Return X or the new, valid address. */
1123 alpha_legitimize_address (rtx x
, rtx oldx ATTRIBUTE_UNUSED
,
1124 enum machine_mode mode
)
1126 rtx new_x
= alpha_legitimize_address_1 (x
, NULL_RTX
, mode
);
1127 return new_x
? new_x
: x
;
1130 /* Primarily this is required for TLS symbols, but given that our move
1131 patterns *ought* to be able to handle any symbol at any time, we
1132 should never be spilling symbolic operands to the constant pool, ever. */
1135 alpha_cannot_force_const_mem (rtx x
)
1137 enum rtx_code code
= GET_CODE (x
);
1138 return code
== SYMBOL_REF
|| code
== LABEL_REF
|| code
== CONST
;
1141 /* We do not allow indirect calls to be optimized into sibling calls, nor
1142 can we allow a call to a function with a different GP to be optimized
1146 alpha_function_ok_for_sibcall (tree decl
, tree exp ATTRIBUTE_UNUSED
)
1148 /* Can't do indirect tail calls, since we don't know if the target
1149 uses the same GP. */
1153 /* Otherwise, we can make a tail call if the target function shares
1155 return decl_has_samegp (decl
);
1159 some_small_symbolic_operand_int (rtx
*px
, void *data ATTRIBUTE_UNUSED
)
1163 /* Don't re-split. */
1164 if (GET_CODE (x
) == LO_SUM
)
1167 return small_symbolic_operand (x
, Pmode
) != 0;
1171 split_small_symbolic_operand_1 (rtx
*px
, void *data ATTRIBUTE_UNUSED
)
1175 /* Don't re-split. */
1176 if (GET_CODE (x
) == LO_SUM
)
1179 if (small_symbolic_operand (x
, Pmode
))
1181 x
= gen_rtx_LO_SUM (Pmode
, pic_offset_table_rtx
, x
);
1190 split_small_symbolic_operand (rtx x
)
1193 for_each_rtx (&x
, split_small_symbolic_operand_1
, NULL
);
1197 /* Indicate that INSN cannot be duplicated. This is true for any insn
1198 that we've marked with gpdisp relocs, since those have to stay in
1199 1-1 correspondence with one another.
1201 Technically we could copy them if we could set up a mapping from one
1202 sequence number to another, across the set of insns to be duplicated.
1203 This seems overly complicated and error-prone since interblock motion
1204 from sched-ebb could move one of the pair of insns to a different block.
1206 Also cannot allow jsr insns to be duplicated. If they throw exceptions,
1207 then they'll be in a different block from their ldgp. Which could lead
1208 the bb reorder code to think that it would be ok to copy just the block
1209 containing the call and branch to the block containing the ldgp. */
1212 alpha_cannot_copy_insn_p (rtx insn
)
1214 if (!reload_completed
|| !TARGET_EXPLICIT_RELOCS
)
1216 if (recog_memoized (insn
) >= 0)
1217 return get_attr_cannot_copy (insn
);
1223 /* Try a machine-dependent way of reloading an illegitimate address
1224 operand. If we find one, push the reload and return the new rtx. */
1227 alpha_legitimize_reload_address (rtx x
,
1228 enum machine_mode mode ATTRIBUTE_UNUSED
,
1229 int opnum
, int type
,
1230 int ind_levels ATTRIBUTE_UNUSED
)
1232 /* We must recognize output that we have already generated ourselves. */
1233 if (GET_CODE (x
) == PLUS
1234 && GET_CODE (XEXP (x
, 0)) == PLUS
1235 && REG_P (XEXP (XEXP (x
, 0), 0))
1236 && CONST_INT_P (XEXP (XEXP (x
, 0), 1))
1237 && CONST_INT_P (XEXP (x
, 1)))
1239 push_reload (XEXP (x
, 0), NULL_RTX
, &XEXP (x
, 0), NULL
,
1240 BASE_REG_CLASS
, GET_MODE (x
), VOIDmode
, 0, 0,
1241 opnum
, (enum reload_type
) type
);
1245 /* We wish to handle large displacements off a base register by
1246 splitting the addend across an ldah and the mem insn. This
1247 cuts number of extra insns needed from 3 to 1. */
1248 if (GET_CODE (x
) == PLUS
1249 && REG_P (XEXP (x
, 0))
1250 && REGNO (XEXP (x
, 0)) < FIRST_PSEUDO_REGISTER
1251 && REGNO_OK_FOR_BASE_P (REGNO (XEXP (x
, 0)))
1252 && GET_CODE (XEXP (x
, 1)) == CONST_INT
)
1254 HOST_WIDE_INT val
= INTVAL (XEXP (x
, 1));
1255 HOST_WIDE_INT low
= ((val
& 0xffff) ^ 0x8000) - 0x8000;
1257 = (((val
- low
) & 0xffffffff) ^ 0x80000000) - 0x80000000;
1259 /* Check for 32-bit overflow. */
1260 if (high
+ low
!= val
)
1263 /* Reload the high part into a base reg; leave the low part
1264 in the mem directly. */
1265 x
= gen_rtx_PLUS (GET_MODE (x
),
1266 gen_rtx_PLUS (GET_MODE (x
), XEXP (x
, 0),
1270 push_reload (XEXP (x
, 0), NULL_RTX
, &XEXP (x
, 0), NULL
,
1271 BASE_REG_CLASS
, GET_MODE (x
), VOIDmode
, 0, 0,
1272 opnum
, (enum reload_type
) type
);
1279 /* Compute a (partial) cost for rtx X. Return true if the complete
1280 cost has been computed, and false if subexpressions should be
1281 scanned. In either case, *TOTAL contains the cost result. */
1284 alpha_rtx_costs (rtx x
, int code
, int outer_code
, int *total
,
1287 enum machine_mode mode
= GET_MODE (x
);
1288 bool float_mode_p
= FLOAT_MODE_P (mode
);
1289 const struct alpha_rtx_cost_data
*cost_data
;
1292 cost_data
= &alpha_rtx_cost_size
;
1294 cost_data
= &alpha_rtx_cost_data
[alpha_tune
];
1299 /* If this is an 8-bit constant, return zero since it can be used
1300 nearly anywhere with no cost. If it is a valid operand for an
1301 ADD or AND, likewise return 0 if we know it will be used in that
1302 context. Otherwise, return 2 since it might be used there later.
1303 All other constants take at least two insns. */
1304 if (INTVAL (x
) >= 0 && INTVAL (x
) < 256)
1312 if (x
== CONST0_RTX (mode
))
1314 else if ((outer_code
== PLUS
&& add_operand (x
, VOIDmode
))
1315 || (outer_code
== AND
&& and_operand (x
, VOIDmode
)))
1317 else if (add_operand (x
, VOIDmode
) || and_operand (x
, VOIDmode
))
1320 *total
= COSTS_N_INSNS (2);
1326 if (TARGET_EXPLICIT_RELOCS
&& small_symbolic_operand (x
, VOIDmode
))
1327 *total
= COSTS_N_INSNS (outer_code
!= MEM
);
1328 else if (TARGET_EXPLICIT_RELOCS
&& local_symbolic_operand (x
, VOIDmode
))
1329 *total
= COSTS_N_INSNS (1 + (outer_code
!= MEM
));
1330 else if (tls_symbolic_operand_type (x
))
1331 /* Estimate of cost for call_pal rduniq. */
1332 /* ??? How many insns do we emit here? More than one... */
1333 *total
= COSTS_N_INSNS (15);
1335 /* Otherwise we do a load from the GOT. */
1336 *total
= COSTS_N_INSNS (!speed
? 1 : alpha_memory_latency
);
1340 /* This is effectively an add_operand. */
1347 *total
= cost_data
->fp_add
;
1348 else if (GET_CODE (XEXP (x
, 0)) == MULT
1349 && const48_operand (XEXP (XEXP (x
, 0), 1), VOIDmode
))
1351 *total
= (rtx_cost (XEXP (XEXP (x
, 0), 0),
1352 (enum rtx_code
) outer_code
, speed
)
1353 + rtx_cost (XEXP (x
, 1),
1354 (enum rtx_code
) outer_code
, speed
)
1355 + COSTS_N_INSNS (1));
1362 *total
= cost_data
->fp_mult
;
1363 else if (mode
== DImode
)
1364 *total
= cost_data
->int_mult_di
;
1366 *total
= cost_data
->int_mult_si
;
1370 if (CONST_INT_P (XEXP (x
, 1))
1371 && INTVAL (XEXP (x
, 1)) <= 3)
1373 *total
= COSTS_N_INSNS (1);
1380 *total
= cost_data
->int_shift
;
1385 *total
= cost_data
->fp_add
;
1387 *total
= cost_data
->int_cmov
;
1395 *total
= cost_data
->int_div
;
1396 else if (mode
== SFmode
)
1397 *total
= cost_data
->fp_div_sf
;
1399 *total
= cost_data
->fp_div_df
;
1403 *total
= COSTS_N_INSNS (!speed
? 1 : alpha_memory_latency
);
1409 *total
= COSTS_N_INSNS (1);
1417 *total
= COSTS_N_INSNS (1) + cost_data
->int_cmov
;
1423 case UNSIGNED_FLOAT
:
1426 case FLOAT_TRUNCATE
:
1427 *total
= cost_data
->fp_add
;
1431 if (MEM_P (XEXP (x
, 0)))
1434 *total
= cost_data
->fp_add
;
1442 /* REF is an alignable memory location. Place an aligned SImode
1443 reference into *PALIGNED_MEM and the number of bits to shift into
1444 *PBITNUM. SCRATCH is a free register for use in reloading out
1445 of range stack slots. */
1448 get_aligned_mem (rtx ref
, rtx
*paligned_mem
, rtx
*pbitnum
)
1451 HOST_WIDE_INT disp
, offset
;
1453 gcc_assert (MEM_P (ref
));
1455 if (reload_in_progress
1456 && ! memory_address_p (GET_MODE (ref
), XEXP (ref
, 0)))
1458 base
= find_replacement (&XEXP (ref
, 0));
1459 gcc_assert (memory_address_p (GET_MODE (ref
), base
));
1462 base
= XEXP (ref
, 0);
1464 if (GET_CODE (base
) == PLUS
)
1465 disp
= INTVAL (XEXP (base
, 1)), base
= XEXP (base
, 0);
1469 /* Find the byte offset within an aligned word. If the memory itself is
1470 claimed to be aligned, believe it. Otherwise, aligned_memory_operand
1471 will have examined the base register and determined it is aligned, and
1472 thus displacements from it are naturally alignable. */
1473 if (MEM_ALIGN (ref
) >= 32)
1478 /* The location should not cross aligned word boundary. */
1479 gcc_assert (offset
+ GET_MODE_SIZE (GET_MODE (ref
))
1480 <= GET_MODE_SIZE (SImode
));
1482 /* Access the entire aligned word. */
1483 *paligned_mem
= widen_memory_access (ref
, SImode
, -offset
);
1485 /* Convert the byte offset within the word to a bit offset. */
1486 if (WORDS_BIG_ENDIAN
)
1487 offset
= 32 - (GET_MODE_BITSIZE (GET_MODE (ref
)) + offset
* 8);
1490 *pbitnum
= GEN_INT (offset
);
1493 /* Similar, but just get the address. Handle the two reload cases.
1494 Add EXTRA_OFFSET to the address we return. */
1497 get_unaligned_address (rtx ref
)
1500 HOST_WIDE_INT offset
= 0;
1502 gcc_assert (MEM_P (ref
));
1504 if (reload_in_progress
1505 && ! memory_address_p (GET_MODE (ref
), XEXP (ref
, 0)))
1507 base
= find_replacement (&XEXP (ref
, 0));
1509 gcc_assert (memory_address_p (GET_MODE (ref
), base
));
1512 base
= XEXP (ref
, 0);
1514 if (GET_CODE (base
) == PLUS
)
1515 offset
+= INTVAL (XEXP (base
, 1)), base
= XEXP (base
, 0);
1517 return plus_constant (base
, offset
);
1520 /* Compute a value X, such that X & 7 == (ADDR + OFS) & 7.
1521 X is always returned in a register. */
1524 get_unaligned_offset (rtx addr
, HOST_WIDE_INT ofs
)
1526 if (GET_CODE (addr
) == PLUS
)
1528 ofs
+= INTVAL (XEXP (addr
, 1));
1529 addr
= XEXP (addr
, 0);
1532 return expand_simple_binop (Pmode
, PLUS
, addr
, GEN_INT (ofs
& 7),
1533 NULL_RTX
, 1, OPTAB_LIB_WIDEN
);
1536 /* On the Alpha, all (non-symbolic) constants except zero go into
1537 a floating-point register via memory. Note that we cannot
1538 return anything that is not a subset of RCLASS, and that some
1539 symbolic constants cannot be dropped to memory. */
1542 alpha_preferred_reload_class(rtx x
, enum reg_class rclass
)
1544 /* Zero is present in any register class. */
1545 if (x
== CONST0_RTX (GET_MODE (x
)))
1548 /* These sorts of constants we can easily drop to memory. */
1550 || GET_CODE (x
) == CONST_DOUBLE
1551 || GET_CODE (x
) == CONST_VECTOR
)
1553 if (rclass
== FLOAT_REGS
)
1555 if (rclass
== ALL_REGS
)
1556 return GENERAL_REGS
;
1560 /* All other kinds of constants should not (and in the case of HIGH
1561 cannot) be dropped to memory -- instead we use a GENERAL_REGS
1562 secondary reload. */
1564 return (rclass
== ALL_REGS
? GENERAL_REGS
: rclass
);
1569 /* Inform reload about cases where moving X with a mode MODE to a register in
1570 RCLASS requires an extra scratch or immediate register. Return the class
1571 needed for the immediate register. */
1573 static enum reg_class
1574 alpha_secondary_reload (bool in_p
, rtx x
, enum reg_class rclass
,
1575 enum machine_mode mode
, secondary_reload_info
*sri
)
1577 /* Loading and storing HImode or QImode values to and from memory
1578 usually requires a scratch register. */
1579 if (!TARGET_BWX
&& (mode
== QImode
|| mode
== HImode
|| mode
== CQImode
))
1581 if (any_memory_operand (x
, mode
))
1585 if (!aligned_memory_operand (x
, mode
))
1586 sri
->icode
= reload_in_optab
[mode
];
1589 sri
->icode
= reload_out_optab
[mode
];
1594 /* We also cannot do integral arithmetic into FP regs, as might result
1595 from register elimination into a DImode fp register. */
1596 if (rclass
== FLOAT_REGS
)
1598 if (MEM_P (x
) && GET_CODE (XEXP (x
, 0)) == AND
)
1599 return GENERAL_REGS
;
1600 if (in_p
&& INTEGRAL_MODE_P (mode
)
1601 && !MEM_P (x
) && !REG_P (x
) && !CONST_INT_P (x
))
1602 return GENERAL_REGS
;
1608 /* Subfunction of the following function. Update the flags of any MEM
1609 found in part of X. */
1612 alpha_set_memflags_1 (rtx
*xp
, void *data
)
1614 rtx x
= *xp
, orig
= (rtx
) data
;
1619 MEM_VOLATILE_P (x
) = MEM_VOLATILE_P (orig
);
1620 MEM_IN_STRUCT_P (x
) = MEM_IN_STRUCT_P (orig
);
1621 MEM_SCALAR_P (x
) = MEM_SCALAR_P (orig
);
1622 MEM_NOTRAP_P (x
) = MEM_NOTRAP_P (orig
);
1623 MEM_READONLY_P (x
) = MEM_READONLY_P (orig
);
1625 /* Sadly, we cannot use alias sets because the extra aliasing
1626 produced by the AND interferes. Given that two-byte quantities
1627 are the only thing we would be able to differentiate anyway,
1628 there does not seem to be any point in convoluting the early
1629 out of the alias check. */
1634 /* Given SEQ, which is an INSN list, look for any MEMs in either
1635 a SET_DEST or a SET_SRC and copy the in-struct, unchanging, and
1636 volatile flags from REF into each of the MEMs found. If REF is not
1637 a MEM, don't do anything. */
1640 alpha_set_memflags (rtx seq
, rtx ref
)
1647 /* This is only called from alpha.md, after having had something
1648 generated from one of the insn patterns. So if everything is
1649 zero, the pattern is already up-to-date. */
1650 if (!MEM_VOLATILE_P (ref
)
1651 && !MEM_IN_STRUCT_P (ref
)
1652 && !MEM_SCALAR_P (ref
)
1653 && !MEM_NOTRAP_P (ref
)
1654 && !MEM_READONLY_P (ref
))
1657 for (insn
= seq
; insn
; insn
= NEXT_INSN (insn
))
1659 for_each_rtx (&PATTERN (insn
), alpha_set_memflags_1
, (void *) ref
);
1664 static rtx
alpha_emit_set_const (rtx
, enum machine_mode
, HOST_WIDE_INT
,
1667 /* Internal routine for alpha_emit_set_const to check for N or below insns.
1668 If NO_OUTPUT is true, then we only check to see if N insns are possible,
1669 and return pc_rtx if successful. */
1672 alpha_emit_set_const_1 (rtx target
, enum machine_mode mode
,
1673 HOST_WIDE_INT c
, int n
, bool no_output
)
1675 HOST_WIDE_INT new_const
;
1677 /* Use a pseudo if highly optimizing and still generating RTL. */
1679 = (flag_expensive_optimizations
&& can_create_pseudo_p () ? 0 : target
);
1682 /* If this is a sign-extended 32-bit constant, we can do this in at most
1683 three insns, so do it if we have enough insns left. We always have
1684 a sign-extended 32-bit constant when compiling on a narrow machine. */
1686 if (HOST_BITS_PER_WIDE_INT
!= 64
1687 || c
>> 31 == -1 || c
>> 31 == 0)
1689 HOST_WIDE_INT low
= ((c
& 0xffff) ^ 0x8000) - 0x8000;
1690 HOST_WIDE_INT tmp1
= c
- low
;
1691 HOST_WIDE_INT high
= (((tmp1
>> 16) & 0xffff) ^ 0x8000) - 0x8000;
1692 HOST_WIDE_INT extra
= 0;
1694 /* If HIGH will be interpreted as negative but the constant is
1695 positive, we must adjust it to do two ldha insns. */
1697 if ((high
& 0x8000) != 0 && c
>= 0)
1701 high
= ((tmp1
>> 16) & 0xffff) - 2 * ((tmp1
>> 16) & 0x8000);
1704 if (c
== low
|| (low
== 0 && extra
== 0))
1706 /* We used to use copy_to_suggested_reg (GEN_INT (c), target, mode)
1707 but that meant that we can't handle INT_MIN on 32-bit machines
1708 (like NT/Alpha), because we recurse indefinitely through
1709 emit_move_insn to gen_movdi. So instead, since we know exactly
1710 what we want, create it explicitly. */
1715 target
= gen_reg_rtx (mode
);
1716 emit_insn (gen_rtx_SET (VOIDmode
, target
, GEN_INT (c
)));
1719 else if (n
>= 2 + (extra
!= 0))
1723 if (!can_create_pseudo_p ())
1725 emit_insn (gen_rtx_SET (VOIDmode
, target
, GEN_INT (high
<< 16)));
1729 temp
= copy_to_suggested_reg (GEN_INT (high
<< 16),
1732 /* As of 2002-02-23, addsi3 is only available when not optimizing.
1733 This means that if we go through expand_binop, we'll try to
1734 generate extensions, etc, which will require new pseudos, which
1735 will fail during some split phases. The SImode add patterns
1736 still exist, but are not named. So build the insns by hand. */
1741 subtarget
= gen_reg_rtx (mode
);
1742 insn
= gen_rtx_PLUS (mode
, temp
, GEN_INT (extra
<< 16));
1743 insn
= gen_rtx_SET (VOIDmode
, subtarget
, insn
);
1749 target
= gen_reg_rtx (mode
);
1750 insn
= gen_rtx_PLUS (mode
, temp
, GEN_INT (low
));
1751 insn
= gen_rtx_SET (VOIDmode
, target
, insn
);
1757 /* If we couldn't do it that way, try some other methods. But if we have
1758 no instructions left, don't bother. Likewise, if this is SImode and
1759 we can't make pseudos, we can't do anything since the expand_binop
1760 and expand_unop calls will widen and try to make pseudos. */
1762 if (n
== 1 || (mode
== SImode
&& !can_create_pseudo_p ()))
1765 /* Next, see if we can load a related constant and then shift and possibly
1766 negate it to get the constant we want. Try this once each increasing
1767 numbers of insns. */
1769 for (i
= 1; i
< n
; i
++)
1771 /* First, see if minus some low bits, we've an easy load of
1774 new_const
= ((c
& 0xffff) ^ 0x8000) - 0x8000;
1777 temp
= alpha_emit_set_const (subtarget
, mode
, c
- new_const
, i
, no_output
);
1782 return expand_binop (mode
, add_optab
, temp
, GEN_INT (new_const
),
1783 target
, 0, OPTAB_WIDEN
);
1787 /* Next try complementing. */
1788 temp
= alpha_emit_set_const (subtarget
, mode
, ~c
, i
, no_output
);
1793 return expand_unop (mode
, one_cmpl_optab
, temp
, target
, 0);
1796 /* Next try to form a constant and do a left shift. We can do this
1797 if some low-order bits are zero; the exact_log2 call below tells
1798 us that information. The bits we are shifting out could be any
1799 value, but here we'll just try the 0- and sign-extended forms of
1800 the constant. To try to increase the chance of having the same
1801 constant in more than one insn, start at the highest number of
1802 bits to shift, but try all possibilities in case a ZAPNOT will
1805 bits
= exact_log2 (c
& -c
);
1807 for (; bits
> 0; bits
--)
1809 new_const
= c
>> bits
;
1810 temp
= alpha_emit_set_const (subtarget
, mode
, new_const
, i
, no_output
);
1813 new_const
= (unsigned HOST_WIDE_INT
)c
>> bits
;
1814 temp
= alpha_emit_set_const (subtarget
, mode
, new_const
,
1821 return expand_binop (mode
, ashl_optab
, temp
, GEN_INT (bits
),
1822 target
, 0, OPTAB_WIDEN
);
1826 /* Now try high-order zero bits. Here we try the shifted-in bits as
1827 all zero and all ones. Be careful to avoid shifting outside the
1828 mode and to avoid shifting outside the host wide int size. */
1829 /* On narrow hosts, don't shift a 1 into the high bit, since we'll
1830 confuse the recursive call and set all of the high 32 bits. */
1832 bits
= (MIN (HOST_BITS_PER_WIDE_INT
, GET_MODE_SIZE (mode
) * 8)
1833 - floor_log2 (c
) - 1 - (HOST_BITS_PER_WIDE_INT
< 64));
1835 for (; bits
> 0; bits
--)
1837 new_const
= c
<< bits
;
1838 temp
= alpha_emit_set_const (subtarget
, mode
, new_const
, i
, no_output
);
1841 new_const
= (c
<< bits
) | (((HOST_WIDE_INT
) 1 << bits
) - 1);
1842 temp
= alpha_emit_set_const (subtarget
, mode
, new_const
,
1849 return expand_binop (mode
, lshr_optab
, temp
, GEN_INT (bits
),
1850 target
, 1, OPTAB_WIDEN
);
1854 /* Now try high-order 1 bits. We get that with a sign-extension.
1855 But one bit isn't enough here. Be careful to avoid shifting outside
1856 the mode and to avoid shifting outside the host wide int size. */
1858 bits
= (MIN (HOST_BITS_PER_WIDE_INT
, GET_MODE_SIZE (mode
) * 8)
1859 - floor_log2 (~ c
) - 2);
1861 for (; bits
> 0; bits
--)
1863 new_const
= c
<< bits
;
1864 temp
= alpha_emit_set_const (subtarget
, mode
, new_const
, i
, no_output
);
1867 new_const
= (c
<< bits
) | (((HOST_WIDE_INT
) 1 << bits
) - 1);
1868 temp
= alpha_emit_set_const (subtarget
, mode
, new_const
,
1875 return expand_binop (mode
, ashr_optab
, temp
, GEN_INT (bits
),
1876 target
, 0, OPTAB_WIDEN
);
1881 #if HOST_BITS_PER_WIDE_INT == 64
1882 /* Finally, see if can load a value into the target that is the same as the
1883 constant except that all bytes that are 0 are changed to be 0xff. If we
1884 can, then we can do a ZAPNOT to obtain the desired constant. */
1887 for (i
= 0; i
< 64; i
+= 8)
1888 if ((new_const
& ((HOST_WIDE_INT
) 0xff << i
)) == 0)
1889 new_const
|= (HOST_WIDE_INT
) 0xff << i
;
1891 /* We are only called for SImode and DImode. If this is SImode, ensure that
1892 we are sign extended to a full word. */
1895 new_const
= ((new_const
& 0xffffffff) ^ 0x80000000) - 0x80000000;
1899 temp
= alpha_emit_set_const (subtarget
, mode
, new_const
, n
- 1, no_output
);
1904 return expand_binop (mode
, and_optab
, temp
, GEN_INT (c
| ~ new_const
),
1905 target
, 0, OPTAB_WIDEN
);
1913 /* Try to output insns to set TARGET equal to the constant C if it can be
1914 done in less than N insns. Do all computations in MODE. Returns the place
1915 where the output has been placed if it can be done and the insns have been
1916 emitted. If it would take more than N insns, zero is returned and no
1917 insns and emitted. */
1920 alpha_emit_set_const (rtx target
, enum machine_mode mode
,
1921 HOST_WIDE_INT c
, int n
, bool no_output
)
1923 enum machine_mode orig_mode
= mode
;
1924 rtx orig_target
= target
;
1928 /* If we can't make any pseudos, TARGET is an SImode hard register, we
1929 can't load this constant in one insn, do this in DImode. */
1930 if (!can_create_pseudo_p () && mode
== SImode
1931 && REG_P (target
) && REGNO (target
) < FIRST_PSEUDO_REGISTER
)
1933 result
= alpha_emit_set_const_1 (target
, mode
, c
, 1, no_output
);
1937 target
= no_output
? NULL
: gen_lowpart (DImode
, target
);
1940 else if (mode
== V8QImode
|| mode
== V4HImode
|| mode
== V2SImode
)
1942 target
= no_output
? NULL
: gen_lowpart (DImode
, target
);
1946 /* Try 1 insn, then 2, then up to N. */
1947 for (i
= 1; i
<= n
; i
++)
1949 result
= alpha_emit_set_const_1 (target
, mode
, c
, i
, no_output
);
1957 insn
= get_last_insn ();
1958 set
= single_set (insn
);
1959 if (! CONSTANT_P (SET_SRC (set
)))
1960 set_unique_reg_note (get_last_insn (), REG_EQUAL
, GEN_INT (c
));
1965 /* Allow for the case where we changed the mode of TARGET. */
1968 if (result
== target
)
1969 result
= orig_target
;
1970 else if (mode
!= orig_mode
)
1971 result
= gen_lowpart (orig_mode
, result
);
1977 /* Having failed to find a 3 insn sequence in alpha_emit_set_const,
1978 fall back to a straight forward decomposition. We do this to avoid
1979 exponential run times encountered when looking for longer sequences
1980 with alpha_emit_set_const. */
1983 alpha_emit_set_long_const (rtx target
, HOST_WIDE_INT c1
, HOST_WIDE_INT c2
)
1985 HOST_WIDE_INT d1
, d2
, d3
, d4
;
1987 /* Decompose the entire word */
1988 #if HOST_BITS_PER_WIDE_INT >= 64
1989 gcc_assert (c2
== -(c1
< 0));
1990 d1
= ((c1
& 0xffff) ^ 0x8000) - 0x8000;
1992 d2
= ((c1
& 0xffffffff) ^ 0x80000000) - 0x80000000;
1993 c1
= (c1
- d2
) >> 32;
1994 d3
= ((c1
& 0xffff) ^ 0x8000) - 0x8000;
1996 d4
= ((c1
& 0xffffffff) ^ 0x80000000) - 0x80000000;
1997 gcc_assert (c1
== d4
);
1999 d1
= ((c1
& 0xffff) ^ 0x8000) - 0x8000;
2001 d2
= ((c1
& 0xffffffff) ^ 0x80000000) - 0x80000000;
2002 gcc_assert (c1
== d2
);
2004 d3
= ((c2
& 0xffff) ^ 0x8000) - 0x8000;
2006 d4
= ((c2
& 0xffffffff) ^ 0x80000000) - 0x80000000;
2007 gcc_assert (c2
== d4
);
2010 /* Construct the high word */
2013 emit_move_insn (target
, GEN_INT (d4
));
2015 emit_move_insn (target
, gen_rtx_PLUS (DImode
, target
, GEN_INT (d3
)));
2018 emit_move_insn (target
, GEN_INT (d3
));
2020 /* Shift it into place */
2021 emit_move_insn (target
, gen_rtx_ASHIFT (DImode
, target
, GEN_INT (32)));
2023 /* Add in the low bits. */
2025 emit_move_insn (target
, gen_rtx_PLUS (DImode
, target
, GEN_INT (d2
)));
2027 emit_move_insn (target
, gen_rtx_PLUS (DImode
, target
, GEN_INT (d1
)));
2032 /* Given an integral CONST_INT, CONST_DOUBLE, or CONST_VECTOR, return
2036 alpha_extract_integer (rtx x
, HOST_WIDE_INT
*p0
, HOST_WIDE_INT
*p1
)
2038 HOST_WIDE_INT i0
, i1
;
2040 if (GET_CODE (x
) == CONST_VECTOR
)
2041 x
= simplify_subreg (DImode
, x
, GET_MODE (x
), 0);
2044 if (CONST_INT_P (x
))
2049 else if (HOST_BITS_PER_WIDE_INT
>= 64)
2051 i0
= CONST_DOUBLE_LOW (x
);
2056 i0
= CONST_DOUBLE_LOW (x
);
2057 i1
= CONST_DOUBLE_HIGH (x
);
2064 /* Implement LEGITIMATE_CONSTANT_P. This is all constants for which we
2065 are willing to load the value into a register via a move pattern.
2066 Normally this is all symbolic constants, integral constants that
2067 take three or fewer instructions, and floating-point zero. */
2070 alpha_legitimate_constant_p (rtx x
)
2072 enum machine_mode mode
= GET_MODE (x
);
2073 HOST_WIDE_INT i0
, i1
;
2075 switch (GET_CODE (x
))
2082 if (GET_CODE (XEXP (x
, 0)) == PLUS
2083 && GET_CODE (XEXP (XEXP (x
, 0), 1)) == CONST_INT
)
2084 x
= XEXP (XEXP (x
, 0), 0);
2088 if (GET_CODE (x
) != SYMBOL_REF
)
2094 /* TLS symbols are never valid. */
2095 return SYMBOL_REF_TLS_MODEL (x
) == 0;
2098 if (x
== CONST0_RTX (mode
))
2100 if (FLOAT_MODE_P (mode
))
2105 if (x
== CONST0_RTX (mode
))
2107 if (GET_MODE_CLASS (mode
) != MODE_VECTOR_INT
)
2109 if (GET_MODE_SIZE (mode
) != 8)
2115 if (TARGET_BUILD_CONSTANTS
)
2117 alpha_extract_integer (x
, &i0
, &i1
);
2118 if (HOST_BITS_PER_WIDE_INT
>= 64 || i1
== (-i0
< 0))
2119 return alpha_emit_set_const_1 (x
, mode
, i0
, 3, true) != NULL
;
2127 /* Operand 1 is known to be a constant, and should require more than one
2128 instruction to load. Emit that multi-part load. */
2131 alpha_split_const_mov (enum machine_mode mode
, rtx
*operands
)
2133 HOST_WIDE_INT i0
, i1
;
2134 rtx temp
= NULL_RTX
;
2136 alpha_extract_integer (operands
[1], &i0
, &i1
);
2138 if (HOST_BITS_PER_WIDE_INT
>= 64 || i1
== -(i0
< 0))
2139 temp
= alpha_emit_set_const (operands
[0], mode
, i0
, 3, false);
2141 if (!temp
&& TARGET_BUILD_CONSTANTS
)
2142 temp
= alpha_emit_set_long_const (operands
[0], i0
, i1
);
2146 if (!rtx_equal_p (operands
[0], temp
))
2147 emit_move_insn (operands
[0], temp
);
2154 /* Expand a move instruction; return true if all work is done.
2155 We don't handle non-bwx subword loads here. */
2158 alpha_expand_mov (enum machine_mode mode
, rtx
*operands
)
2162 /* If the output is not a register, the input must be. */
2163 if (MEM_P (operands
[0])
2164 && ! reg_or_0_operand (operands
[1], mode
))
2165 operands
[1] = force_reg (mode
, operands
[1]);
2167 /* Allow legitimize_address to perform some simplifications. */
2168 if (mode
== Pmode
&& symbolic_operand (operands
[1], mode
))
2170 tmp
= alpha_legitimize_address_1 (operands
[1], operands
[0], mode
);
2173 if (tmp
== operands
[0])
2180 /* Early out for non-constants and valid constants. */
2181 if (! CONSTANT_P (operands
[1]) || input_operand (operands
[1], mode
))
2184 /* Split large integers. */
2185 if (CONST_INT_P (operands
[1])
2186 || GET_CODE (operands
[1]) == CONST_DOUBLE
2187 || GET_CODE (operands
[1]) == CONST_VECTOR
)
2189 if (alpha_split_const_mov (mode
, operands
))
2193 /* Otherwise we've nothing left but to drop the thing to memory. */
2194 tmp
= force_const_mem (mode
, operands
[1]);
2196 if (tmp
== NULL_RTX
)
2199 if (reload_in_progress
)
2201 emit_move_insn (operands
[0], XEXP (tmp
, 0));
2202 operands
[1] = replace_equiv_address (tmp
, operands
[0]);
2205 operands
[1] = validize_mem (tmp
);
2209 /* Expand a non-bwx QImode or HImode move instruction;
2210 return true if all work is done. */
2213 alpha_expand_mov_nobwx (enum machine_mode mode
, rtx
*operands
)
2217 /* If the output is not a register, the input must be. */
2218 if (MEM_P (operands
[0]))
2219 operands
[1] = force_reg (mode
, operands
[1]);
2221 /* Handle four memory cases, unaligned and aligned for either the input
2222 or the output. The only case where we can be called during reload is
2223 for aligned loads; all other cases require temporaries. */
2225 if (any_memory_operand (operands
[1], mode
))
2227 if (aligned_memory_operand (operands
[1], mode
))
2229 if (reload_in_progress
)
2232 seq
= gen_reload_inqi_aligned (operands
[0], operands
[1]);
2234 seq
= gen_reload_inhi_aligned (operands
[0], operands
[1]);
2239 rtx aligned_mem
, bitnum
;
2240 rtx scratch
= gen_reg_rtx (SImode
);
2244 get_aligned_mem (operands
[1], &aligned_mem
, &bitnum
);
2246 subtarget
= operands
[0];
2247 if (REG_P (subtarget
))
2248 subtarget
= gen_lowpart (DImode
, subtarget
), copyout
= false;
2250 subtarget
= gen_reg_rtx (DImode
), copyout
= true;
2253 seq
= gen_aligned_loadqi (subtarget
, aligned_mem
,
2256 seq
= gen_aligned_loadhi (subtarget
, aligned_mem
,
2261 emit_move_insn (operands
[0], gen_lowpart (mode
, subtarget
));
2266 /* Don't pass these as parameters since that makes the generated
2267 code depend on parameter evaluation order which will cause
2268 bootstrap failures. */
2270 rtx temp1
, temp2
, subtarget
, ua
;
2273 temp1
= gen_reg_rtx (DImode
);
2274 temp2
= gen_reg_rtx (DImode
);
2276 subtarget
= operands
[0];
2277 if (REG_P (subtarget
))
2278 subtarget
= gen_lowpart (DImode
, subtarget
), copyout
= false;
2280 subtarget
= gen_reg_rtx (DImode
), copyout
= true;
2282 ua
= get_unaligned_address (operands
[1]);
2284 seq
= gen_unaligned_loadqi (subtarget
, ua
, temp1
, temp2
);
2286 seq
= gen_unaligned_loadhi (subtarget
, ua
, temp1
, temp2
);
2288 alpha_set_memflags (seq
, operands
[1]);
2292 emit_move_insn (operands
[0], gen_lowpart (mode
, subtarget
));
2297 if (any_memory_operand (operands
[0], mode
))
2299 if (aligned_memory_operand (operands
[0], mode
))
2301 rtx aligned_mem
, bitnum
;
2302 rtx temp1
= gen_reg_rtx (SImode
);
2303 rtx temp2
= gen_reg_rtx (SImode
);
2305 get_aligned_mem (operands
[0], &aligned_mem
, &bitnum
);
2307 emit_insn (gen_aligned_store (aligned_mem
, operands
[1], bitnum
,
2312 rtx temp1
= gen_reg_rtx (DImode
);
2313 rtx temp2
= gen_reg_rtx (DImode
);
2314 rtx temp3
= gen_reg_rtx (DImode
);
2315 rtx ua
= get_unaligned_address (operands
[0]);
2318 seq
= gen_unaligned_storeqi (ua
, operands
[1], temp1
, temp2
, temp3
);
2320 seq
= gen_unaligned_storehi (ua
, operands
[1], temp1
, temp2
, temp3
);
2322 alpha_set_memflags (seq
, operands
[0]);
2331 /* Implement the movmisalign patterns. One of the operands is a memory
2332 that is not naturally aligned. Emit instructions to load it. */
2335 alpha_expand_movmisalign (enum machine_mode mode
, rtx
*operands
)
2337 /* Honor misaligned loads, for those we promised to do so. */
2338 if (MEM_P (operands
[1]))
2342 if (register_operand (operands
[0], mode
))
2345 tmp
= gen_reg_rtx (mode
);
2347 alpha_expand_unaligned_load (tmp
, operands
[1], 8, 0, 0);
2348 if (tmp
!= operands
[0])
2349 emit_move_insn (operands
[0], tmp
);
2351 else if (MEM_P (operands
[0]))
2353 if (!reg_or_0_operand (operands
[1], mode
))
2354 operands
[1] = force_reg (mode
, operands
[1]);
2355 alpha_expand_unaligned_store (operands
[0], operands
[1], 8, 0);
2361 /* Generate an unsigned DImode to FP conversion. This is the same code
2362 optabs would emit if we didn't have TFmode patterns.
2364 For SFmode, this is the only construction I've found that can pass
2365 gcc.c-torture/execute/ieee/rbug.c. No scenario that uses DFmode
2366 intermediates will work, because you'll get intermediate rounding
2367 that ruins the end result. Some of this could be fixed by turning
2368 on round-to-positive-infinity, but that requires diddling the fpsr,
2369 which kills performance. I tried turning this around and converting
2370 to a negative number, so that I could turn on /m, but either I did
2371 it wrong or there's something else cause I wound up with the exact
2372 same single-bit error. There is a branch-less form of this same code:
2383 fcmoveq $f10,$f11,$f0
2385 I'm not using it because it's the same number of instructions as
2386 this branch-full form, and it has more serialized long latency
2387 instructions on the critical path.
2389 For DFmode, we can avoid rounding errors by breaking up the word
2390 into two pieces, converting them separately, and adding them back:
2392 LC0: .long 0,0x5f800000
2397 cpyse $f11,$f31,$f10
2398 cpyse $f31,$f11,$f11
2406 This doesn't seem to be a clear-cut win over the optabs form.
2407 It probably all depends on the distribution of numbers being
2408 converted -- in the optabs form, all but high-bit-set has a
2409 much lower minimum execution time. */
2412 alpha_emit_floatuns (rtx operands
[2])
2414 rtx neglab
, donelab
, i0
, i1
, f0
, in
, out
;
2415 enum machine_mode mode
;
2418 in
= force_reg (DImode
, operands
[1]);
2419 mode
= GET_MODE (out
);
2420 neglab
= gen_label_rtx ();
2421 donelab
= gen_label_rtx ();
2422 i0
= gen_reg_rtx (DImode
);
2423 i1
= gen_reg_rtx (DImode
);
2424 f0
= gen_reg_rtx (mode
);
2426 emit_cmp_and_jump_insns (in
, const0_rtx
, LT
, const0_rtx
, DImode
, 0, neglab
);
2428 emit_insn (gen_rtx_SET (VOIDmode
, out
, gen_rtx_FLOAT (mode
, in
)));
2429 emit_jump_insn (gen_jump (donelab
));
2432 emit_label (neglab
);
2434 emit_insn (gen_lshrdi3 (i0
, in
, const1_rtx
));
2435 emit_insn (gen_anddi3 (i1
, in
, const1_rtx
));
2436 emit_insn (gen_iordi3 (i0
, i0
, i1
));
2437 emit_insn (gen_rtx_SET (VOIDmode
, f0
, gen_rtx_FLOAT (mode
, i0
)));
2438 emit_insn (gen_rtx_SET (VOIDmode
, out
, gen_rtx_PLUS (mode
, f0
, f0
)));
2440 emit_label (donelab
);
2443 /* Generate the comparison for a conditional branch. */
2446 alpha_emit_conditional_branch (rtx operands
[], enum machine_mode cmp_mode
)
2448 enum rtx_code cmp_code
, branch_code
;
2449 enum machine_mode branch_mode
= VOIDmode
;
2450 enum rtx_code code
= GET_CODE (operands
[0]);
2451 rtx op0
= operands
[1], op1
= operands
[2];
2454 if (cmp_mode
== TFmode
)
2456 op0
= alpha_emit_xfloating_compare (&code
, op0
, op1
);
2461 /* The general case: fold the comparison code to the types of compares
2462 that we have, choosing the branch as necessary. */
2465 case EQ
: case LE
: case LT
: case LEU
: case LTU
:
2467 /* We have these compares: */
2468 cmp_code
= code
, branch_code
= NE
;
2473 /* These must be reversed. */
2474 cmp_code
= reverse_condition (code
), branch_code
= EQ
;
2477 case GE
: case GT
: case GEU
: case GTU
:
2478 /* For FP, we swap them, for INT, we reverse them. */
2479 if (cmp_mode
== DFmode
)
2481 cmp_code
= swap_condition (code
);
2483 tem
= op0
, op0
= op1
, op1
= tem
;
2487 cmp_code
= reverse_condition (code
);
2496 if (cmp_mode
== DFmode
)
2498 if (flag_unsafe_math_optimizations
&& cmp_code
!= UNORDERED
)
2500 /* When we are not as concerned about non-finite values, and we
2501 are comparing against zero, we can branch directly. */
2502 if (op1
== CONST0_RTX (DFmode
))
2503 cmp_code
= UNKNOWN
, branch_code
= code
;
2504 else if (op0
== CONST0_RTX (DFmode
))
2506 /* Undo the swap we probably did just above. */
2507 tem
= op0
, op0
= op1
, op1
= tem
;
2508 branch_code
= swap_condition (cmp_code
);
2514 /* ??? We mark the branch mode to be CCmode to prevent the
2515 compare and branch from being combined, since the compare
2516 insn follows IEEE rules that the branch does not. */
2517 branch_mode
= CCmode
;
2522 /* The following optimizations are only for signed compares. */
2523 if (code
!= LEU
&& code
!= LTU
&& code
!= GEU
&& code
!= GTU
)
2525 /* Whee. Compare and branch against 0 directly. */
2526 if (op1
== const0_rtx
)
2527 cmp_code
= UNKNOWN
, branch_code
= code
;
2529 /* If the constants doesn't fit into an immediate, but can
2530 be generated by lda/ldah, we adjust the argument and
2531 compare against zero, so we can use beq/bne directly. */
2532 /* ??? Don't do this when comparing against symbols, otherwise
2533 we'll reduce (&x == 0x1234) to (&x-0x1234 == 0), which will
2534 be declared false out of hand (at least for non-weak). */
2535 else if (CONST_INT_P (op1
)
2536 && (code
== EQ
|| code
== NE
)
2537 && !(symbolic_operand (op0
, VOIDmode
)
2538 || (REG_P (op0
) && REG_POINTER (op0
))))
2540 rtx n_op1
= GEN_INT (-INTVAL (op1
));
2542 if (! satisfies_constraint_I (op1
)
2543 && (satisfies_constraint_K (n_op1
)
2544 || satisfies_constraint_L (n_op1
)))
2545 cmp_code
= PLUS
, branch_code
= code
, op1
= n_op1
;
2549 if (!reg_or_0_operand (op0
, DImode
))
2550 op0
= force_reg (DImode
, op0
);
2551 if (cmp_code
!= PLUS
&& !reg_or_8bit_operand (op1
, DImode
))
2552 op1
= force_reg (DImode
, op1
);
2555 /* Emit an initial compare instruction, if necessary. */
2557 if (cmp_code
!= UNKNOWN
)
2559 tem
= gen_reg_rtx (cmp_mode
);
2560 emit_move_insn (tem
, gen_rtx_fmt_ee (cmp_code
, cmp_mode
, op0
, op1
));
2563 /* Emit the branch instruction. */
2564 tem
= gen_rtx_SET (VOIDmode
, pc_rtx
,
2565 gen_rtx_IF_THEN_ELSE (VOIDmode
,
2566 gen_rtx_fmt_ee (branch_code
,
2568 CONST0_RTX (cmp_mode
)),
2569 gen_rtx_LABEL_REF (VOIDmode
,
2572 emit_jump_insn (tem
);
2575 /* Certain simplifications can be done to make invalid setcc operations
2576 valid. Return the final comparison, or NULL if we can't work. */
2579 alpha_emit_setcc (rtx operands
[], enum machine_mode cmp_mode
)
2581 enum rtx_code cmp_code
;
2582 enum rtx_code code
= GET_CODE (operands
[1]);
2583 rtx op0
= operands
[2], op1
= operands
[3];
2586 if (cmp_mode
== TFmode
)
2588 op0
= alpha_emit_xfloating_compare (&code
, op0
, op1
);
2593 if (cmp_mode
== DFmode
&& !TARGET_FIX
)
2596 /* The general case: fold the comparison code to the types of compares
2597 that we have, choosing the branch as necessary. */
2602 case EQ
: case LE
: case LT
: case LEU
: case LTU
:
2604 /* We have these compares. */
2605 if (cmp_mode
== DFmode
)
2606 cmp_code
= code
, code
= NE
;
2610 if (cmp_mode
== DImode
&& op1
== const0_rtx
)
2615 cmp_code
= reverse_condition (code
);
2619 case GE
: case GT
: case GEU
: case GTU
:
2620 /* These normally need swapping, but for integer zero we have
2621 special patterns that recognize swapped operands. */
2622 if (cmp_mode
== DImode
&& op1
== const0_rtx
)
2624 code
= swap_condition (code
);
2625 if (cmp_mode
== DFmode
)
2626 cmp_code
= code
, code
= NE
;
2627 tmp
= op0
, op0
= op1
, op1
= tmp
;
2634 if (cmp_mode
== DImode
)
2636 if (!register_operand (op0
, DImode
))
2637 op0
= force_reg (DImode
, op0
);
2638 if (!reg_or_8bit_operand (op1
, DImode
))
2639 op1
= force_reg (DImode
, op1
);
2642 /* Emit an initial compare instruction, if necessary. */
2643 if (cmp_code
!= UNKNOWN
)
2645 tmp
= gen_reg_rtx (cmp_mode
);
2646 emit_insn (gen_rtx_SET (VOIDmode
, tmp
,
2647 gen_rtx_fmt_ee (cmp_code
, cmp_mode
, op0
, op1
)));
2649 op0
= cmp_mode
!= DImode
? gen_lowpart (DImode
, tmp
) : tmp
;
2653 /* Emit the setcc instruction. */
2654 emit_insn (gen_rtx_SET (VOIDmode
, operands
[0],
2655 gen_rtx_fmt_ee (code
, DImode
, op0
, op1
)));
2660 /* Rewrite a comparison against zero CMP of the form
2661 (CODE (cc0) (const_int 0)) so it can be written validly in
2662 a conditional move (if_then_else CMP ...).
2663 If both of the operands that set cc0 are nonzero we must emit
2664 an insn to perform the compare (it can't be done within
2665 the conditional move). */
2668 alpha_emit_conditional_move (rtx cmp
, enum machine_mode mode
)
2670 enum rtx_code code
= GET_CODE (cmp
);
2671 enum rtx_code cmov_code
= NE
;
2672 rtx op0
= XEXP (cmp
, 0);
2673 rtx op1
= XEXP (cmp
, 1);
2674 enum machine_mode cmp_mode
2675 = (GET_MODE (op0
) == VOIDmode
? DImode
: GET_MODE (op0
));
2676 enum machine_mode cmov_mode
= VOIDmode
;
2677 int local_fast_math
= flag_unsafe_math_optimizations
;
2680 if (cmp_mode
== TFmode
)
2682 op0
= alpha_emit_xfloating_compare (&code
, op0
, op1
);
2687 gcc_assert (cmp_mode
== DFmode
|| cmp_mode
== DImode
);
2689 if (FLOAT_MODE_P (cmp_mode
) != FLOAT_MODE_P (mode
))
2691 enum rtx_code cmp_code
;
2696 /* If we have fp<->int register move instructions, do a cmov by
2697 performing the comparison in fp registers, and move the
2698 zero/nonzero value to integer registers, where we can then
2699 use a normal cmov, or vice-versa. */
2703 case EQ
: case LE
: case LT
: case LEU
: case LTU
:
2704 /* We have these compares. */
2705 cmp_code
= code
, code
= NE
;
2709 /* This must be reversed. */
2710 cmp_code
= EQ
, code
= EQ
;
2713 case GE
: case GT
: case GEU
: case GTU
:
2714 /* These normally need swapping, but for integer zero we have
2715 special patterns that recognize swapped operands. */
2716 if (cmp_mode
== DImode
&& op1
== const0_rtx
)
2717 cmp_code
= code
, code
= NE
;
2720 cmp_code
= swap_condition (code
);
2722 tem
= op0
, op0
= op1
, op1
= tem
;
2730 tem
= gen_reg_rtx (cmp_mode
);
2731 emit_insn (gen_rtx_SET (VOIDmode
, tem
,
2732 gen_rtx_fmt_ee (cmp_code
, cmp_mode
,
2735 cmp_mode
= cmp_mode
== DImode
? DFmode
: DImode
;
2736 op0
= gen_lowpart (cmp_mode
, tem
);
2737 op1
= CONST0_RTX (cmp_mode
);
2738 local_fast_math
= 1;
2741 /* We may be able to use a conditional move directly.
2742 This avoids emitting spurious compares. */
2743 if (signed_comparison_operator (cmp
, VOIDmode
)
2744 && (cmp_mode
== DImode
|| local_fast_math
)
2745 && (op0
== CONST0_RTX (cmp_mode
) || op1
== CONST0_RTX (cmp_mode
)))
2746 return gen_rtx_fmt_ee (code
, VOIDmode
, op0
, op1
);
2748 /* We can't put the comparison inside the conditional move;
2749 emit a compare instruction and put that inside the
2750 conditional move. Make sure we emit only comparisons we have;
2751 swap or reverse as necessary. */
2753 if (!can_create_pseudo_p ())
2758 case EQ
: case LE
: case LT
: case LEU
: case LTU
:
2759 /* We have these compares: */
2763 /* This must be reversed. */
2764 code
= reverse_condition (code
);
2768 case GE
: case GT
: case GEU
: case GTU
:
2769 /* These must be swapped. */
2770 if (op1
!= CONST0_RTX (cmp_mode
))
2772 code
= swap_condition (code
);
2773 tem
= op0
, op0
= op1
, op1
= tem
;
2781 if (cmp_mode
== DImode
)
2783 if (!reg_or_0_operand (op0
, DImode
))
2784 op0
= force_reg (DImode
, op0
);
2785 if (!reg_or_8bit_operand (op1
, DImode
))
2786 op1
= force_reg (DImode
, op1
);
2789 /* ??? We mark the branch mode to be CCmode to prevent the compare
2790 and cmov from being combined, since the compare insn follows IEEE
2791 rules that the cmov does not. */
2792 if (cmp_mode
== DFmode
&& !local_fast_math
)
2795 tem
= gen_reg_rtx (cmp_mode
);
2796 emit_move_insn (tem
, gen_rtx_fmt_ee (code
, cmp_mode
, op0
, op1
));
2797 return gen_rtx_fmt_ee (cmov_code
, cmov_mode
, tem
, CONST0_RTX (cmp_mode
));
2800 /* Simplify a conditional move of two constants into a setcc with
2801 arithmetic. This is done with a splitter since combine would
2802 just undo the work if done during code generation. It also catches
2803 cases we wouldn't have before cse. */
2806 alpha_split_conditional_move (enum rtx_code code
, rtx dest
, rtx cond
,
2807 rtx t_rtx
, rtx f_rtx
)
2809 HOST_WIDE_INT t
, f
, diff
;
2810 enum machine_mode mode
;
2811 rtx target
, subtarget
, tmp
;
2813 mode
= GET_MODE (dest
);
2818 if (((code
== NE
|| code
== EQ
) && diff
< 0)
2819 || (code
== GE
|| code
== GT
))
2821 code
= reverse_condition (code
);
2822 diff
= t
, t
= f
, f
= diff
;
2826 subtarget
= target
= dest
;
2829 target
= gen_lowpart (DImode
, dest
);
2830 if (can_create_pseudo_p ())
2831 subtarget
= gen_reg_rtx (DImode
);
2835 /* Below, we must be careful to use copy_rtx on target and subtarget
2836 in intermediate insns, as they may be a subreg rtx, which may not
2839 if (f
== 0 && exact_log2 (diff
) > 0
2840 /* On EV6, we've got enough shifters to make non-arithmetic shifts
2841 viable over a longer latency cmove. On EV5, the E0 slot is a
2842 scarce resource, and on EV4 shift has the same latency as a cmove. */
2843 && (diff
<= 8 || alpha_tune
== PROCESSOR_EV6
))
2845 tmp
= gen_rtx_fmt_ee (code
, DImode
, cond
, const0_rtx
);
2846 emit_insn (gen_rtx_SET (VOIDmode
, copy_rtx (subtarget
), tmp
));
2848 tmp
= gen_rtx_ASHIFT (DImode
, copy_rtx (subtarget
),
2849 GEN_INT (exact_log2 (t
)));
2850 emit_insn (gen_rtx_SET (VOIDmode
, target
, tmp
));
2852 else if (f
== 0 && t
== -1)
2854 tmp
= gen_rtx_fmt_ee (code
, DImode
, cond
, const0_rtx
);
2855 emit_insn (gen_rtx_SET (VOIDmode
, copy_rtx (subtarget
), tmp
));
2857 emit_insn (gen_negdi2 (target
, copy_rtx (subtarget
)));
2859 else if (diff
== 1 || diff
== 4 || diff
== 8)
2863 tmp
= gen_rtx_fmt_ee (code
, DImode
, cond
, const0_rtx
);
2864 emit_insn (gen_rtx_SET (VOIDmode
, copy_rtx (subtarget
), tmp
));
2867 emit_insn (gen_adddi3 (target
, copy_rtx (subtarget
), GEN_INT (f
)));
2870 add_op
= GEN_INT (f
);
2871 if (sext_add_operand (add_op
, mode
))
2873 tmp
= gen_rtx_MULT (DImode
, copy_rtx (subtarget
),
2875 tmp
= gen_rtx_PLUS (DImode
, tmp
, add_op
);
2876 emit_insn (gen_rtx_SET (VOIDmode
, target
, tmp
));
2888 /* Look up the function X_floating library function name for the
2891 struct GTY(()) xfloating_op
2893 const enum rtx_code code
;
2894 const char *const GTY((skip
)) osf_func
;
2895 const char *const GTY((skip
)) vms_func
;
2899 static GTY(()) struct xfloating_op xfloating_ops
[] =
2901 { PLUS
, "_OtsAddX", "OTS$ADD_X", 0 },
2902 { MINUS
, "_OtsSubX", "OTS$SUB_X", 0 },
2903 { MULT
, "_OtsMulX", "OTS$MUL_X", 0 },
2904 { DIV
, "_OtsDivX", "OTS$DIV_X", 0 },
2905 { EQ
, "_OtsEqlX", "OTS$EQL_X", 0 },
2906 { NE
, "_OtsNeqX", "OTS$NEQ_X", 0 },
2907 { LT
, "_OtsLssX", "OTS$LSS_X", 0 },
2908 { LE
, "_OtsLeqX", "OTS$LEQ_X", 0 },
2909 { GT
, "_OtsGtrX", "OTS$GTR_X", 0 },
2910 { GE
, "_OtsGeqX", "OTS$GEQ_X", 0 },
2911 { FIX
, "_OtsCvtXQ", "OTS$CVTXQ", 0 },
2912 { FLOAT
, "_OtsCvtQX", "OTS$CVTQX", 0 },
2913 { UNSIGNED_FLOAT
, "_OtsCvtQUX", "OTS$CVTQUX", 0 },
2914 { FLOAT_EXTEND
, "_OtsConvertFloatTX", "OTS$CVT_FLOAT_T_X", 0 },
2915 { FLOAT_TRUNCATE
, "_OtsConvertFloatXT", "OTS$CVT_FLOAT_X_T", 0 }
2918 static GTY(()) struct xfloating_op vax_cvt_ops
[] =
2920 { FLOAT_EXTEND
, "_OtsConvertFloatGX", "OTS$CVT_FLOAT_G_X", 0 },
2921 { FLOAT_TRUNCATE
, "_OtsConvertFloatXG", "OTS$CVT_FLOAT_X_G", 0 }
2925 alpha_lookup_xfloating_lib_func (enum rtx_code code
)
2927 struct xfloating_op
*ops
= xfloating_ops
;
2928 long n
= ARRAY_SIZE (xfloating_ops
);
2931 gcc_assert (TARGET_HAS_XFLOATING_LIBS
);
2933 /* How irritating. Nothing to key off for the main table. */
2934 if (TARGET_FLOAT_VAX
&& (code
== FLOAT_EXTEND
|| code
== FLOAT_TRUNCATE
))
2937 n
= ARRAY_SIZE (vax_cvt_ops
);
2940 for (i
= 0; i
< n
; ++i
, ++ops
)
2941 if (ops
->code
== code
)
2943 rtx func
= ops
->libcall
;
2946 func
= init_one_libfunc (TARGET_ABI_OPEN_VMS
2947 ? ops
->vms_func
: ops
->osf_func
);
2948 ops
->libcall
= func
;
2956 /* Most X_floating operations take the rounding mode as an argument.
2957 Compute that here. */
2960 alpha_compute_xfloating_mode_arg (enum rtx_code code
,
2961 enum alpha_fp_rounding_mode round
)
2967 case ALPHA_FPRM_NORM
:
2970 case ALPHA_FPRM_MINF
:
2973 case ALPHA_FPRM_CHOP
:
2976 case ALPHA_FPRM_DYN
:
2982 /* XXX For reference, round to +inf is mode = 3. */
2985 if (code
== FLOAT_TRUNCATE
&& alpha_fptm
== ALPHA_FPTM_N
)
2991 /* Emit an X_floating library function call.
2993 Note that these functions do not follow normal calling conventions:
2994 TFmode arguments are passed in two integer registers (as opposed to
2995 indirect); TFmode return values appear in R16+R17.
2997 FUNC is the function to call.
2998 TARGET is where the output belongs.
2999 OPERANDS are the inputs.
3000 NOPERANDS is the count of inputs.
3001 EQUIV is the expression equivalent for the function.
3005 alpha_emit_xfloating_libcall (rtx func
, rtx target
, rtx operands
[],
3006 int noperands
, rtx equiv
)
3008 rtx usage
= NULL_RTX
, tmp
, reg
;
3013 for (i
= 0; i
< noperands
; ++i
)
3015 switch (GET_MODE (operands
[i
]))
3018 reg
= gen_rtx_REG (TFmode
, regno
);
3023 reg
= gen_rtx_REG (DFmode
, regno
+ 32);
3028 gcc_assert (CONST_INT_P (operands
[i
]));
3031 reg
= gen_rtx_REG (DImode
, regno
);
3039 emit_move_insn (reg
, operands
[i
]);
3040 usage
= alloc_EXPR_LIST (0, gen_rtx_USE (VOIDmode
, reg
), usage
);
3043 switch (GET_MODE (target
))
3046 reg
= gen_rtx_REG (TFmode
, 16);
3049 reg
= gen_rtx_REG (DFmode
, 32);
3052 reg
= gen_rtx_REG (DImode
, 0);
3058 tmp
= gen_rtx_MEM (QImode
, func
);
3059 tmp
= emit_call_insn (GEN_CALL_VALUE (reg
, tmp
, const0_rtx
,
3060 const0_rtx
, const0_rtx
));
3061 CALL_INSN_FUNCTION_USAGE (tmp
) = usage
;
3062 RTL_CONST_CALL_P (tmp
) = 1;
3067 emit_libcall_block (tmp
, target
, reg
, equiv
);
3070 /* Emit an X_floating library function call for arithmetic (+,-,*,/). */
3073 alpha_emit_xfloating_arith (enum rtx_code code
, rtx operands
[])
3077 rtx out_operands
[3];
3079 func
= alpha_lookup_xfloating_lib_func (code
);
3080 mode
= alpha_compute_xfloating_mode_arg (code
, alpha_fprm
);
3082 out_operands
[0] = operands
[1];
3083 out_operands
[1] = operands
[2];
3084 out_operands
[2] = GEN_INT (mode
);
3085 alpha_emit_xfloating_libcall (func
, operands
[0], out_operands
, 3,
3086 gen_rtx_fmt_ee (code
, TFmode
, operands
[1],
3090 /* Emit an X_floating library function call for a comparison. */
3093 alpha_emit_xfloating_compare (enum rtx_code
*pcode
, rtx op0
, rtx op1
)
3095 enum rtx_code cmp_code
, res_code
;
3096 rtx func
, out
, operands
[2], note
;
3098 /* X_floating library comparison functions return
3102 Convert the compare against the raw return value. */
3130 func
= alpha_lookup_xfloating_lib_func (cmp_code
);
3134 out
= gen_reg_rtx (DImode
);
3136 /* What's actually returned is -1,0,1, not a proper boolean value,
3137 so use an EXPR_LIST as with a generic libcall instead of a
3138 comparison type expression. */
3139 note
= gen_rtx_EXPR_LIST (VOIDmode
, op1
, NULL_RTX
);
3140 note
= gen_rtx_EXPR_LIST (VOIDmode
, op0
, note
);
3141 note
= gen_rtx_EXPR_LIST (VOIDmode
, func
, note
);
3142 alpha_emit_xfloating_libcall (func
, out
, operands
, 2, note
);
3147 /* Emit an X_floating library function call for a conversion. */
3150 alpha_emit_xfloating_cvt (enum rtx_code orig_code
, rtx operands
[])
3152 int noperands
= 1, mode
;
3153 rtx out_operands
[2];
3155 enum rtx_code code
= orig_code
;
3157 if (code
== UNSIGNED_FIX
)
3160 func
= alpha_lookup_xfloating_lib_func (code
);
3162 out_operands
[0] = operands
[1];
3167 mode
= alpha_compute_xfloating_mode_arg (code
, ALPHA_FPRM_CHOP
);
3168 out_operands
[1] = GEN_INT (mode
);
3171 case FLOAT_TRUNCATE
:
3172 mode
= alpha_compute_xfloating_mode_arg (code
, alpha_fprm
);
3173 out_operands
[1] = GEN_INT (mode
);
3180 alpha_emit_xfloating_libcall (func
, operands
[0], out_operands
, noperands
,
3181 gen_rtx_fmt_e (orig_code
,
3182 GET_MODE (operands
[0]),
3186 /* Split a TImode or TFmode move from OP[1] to OP[0] into a pair of
3187 DImode moves from OP[2,3] to OP[0,1]. If FIXUP_OVERLAP is true,
3188 guarantee that the sequence
3191 is valid. Naturally, output operand ordering is little-endian.
3192 This is used by *movtf_internal and *movti_internal. */
3195 alpha_split_tmode_pair (rtx operands
[4], enum machine_mode mode
,
3198 switch (GET_CODE (operands
[1]))
3201 operands
[3] = gen_rtx_REG (DImode
, REGNO (operands
[1]) + 1);
3202 operands
[2] = gen_rtx_REG (DImode
, REGNO (operands
[1]));
3206 operands
[3] = adjust_address (operands
[1], DImode
, 8);
3207 operands
[2] = adjust_address (operands
[1], DImode
, 0);
3212 gcc_assert (operands
[1] == CONST0_RTX (mode
));
3213 operands
[2] = operands
[3] = const0_rtx
;
3220 switch (GET_CODE (operands
[0]))
3223 operands
[1] = gen_rtx_REG (DImode
, REGNO (operands
[0]) + 1);
3224 operands
[0] = gen_rtx_REG (DImode
, REGNO (operands
[0]));
3228 operands
[1] = adjust_address (operands
[0], DImode
, 8);
3229 operands
[0] = adjust_address (operands
[0], DImode
, 0);
3236 if (fixup_overlap
&& reg_overlap_mentioned_p (operands
[0], operands
[3]))
3239 tmp
= operands
[0], operands
[0] = operands
[1], operands
[1] = tmp
;
3240 tmp
= operands
[2], operands
[2] = operands
[3], operands
[3] = tmp
;
3244 /* Implement negtf2 or abstf2. Op0 is destination, op1 is source,
3245 op2 is a register containing the sign bit, operation is the
3246 logical operation to be performed. */
3249 alpha_split_tfmode_frobsign (rtx operands
[3], rtx (*operation
) (rtx
, rtx
, rtx
))
3251 rtx high_bit
= operands
[2];
3255 alpha_split_tmode_pair (operands
, TFmode
, false);
3257 /* Detect three flavors of operand overlap. */
3259 if (rtx_equal_p (operands
[0], operands
[2]))
3261 else if (rtx_equal_p (operands
[1], operands
[2]))
3263 if (rtx_equal_p (operands
[0], high_bit
))
3270 emit_move_insn (operands
[0], operands
[2]);
3272 /* ??? If the destination overlaps both source tf and high_bit, then
3273 assume source tf is dead in its entirety and use the other half
3274 for a scratch register. Otherwise "scratch" is just the proper
3275 destination register. */
3276 scratch
= operands
[move
< 2 ? 1 : 3];
3278 emit_insn ((*operation
) (scratch
, high_bit
, operands
[3]));
3282 emit_move_insn (operands
[0], operands
[2]);
3284 emit_move_insn (operands
[1], scratch
);
3288 /* Use ext[wlq][lh] as the Architecture Handbook describes for extracting
3292 word: ldq_u r1,X(r11) ldq_u r1,X(r11)
3293 ldq_u r2,X+1(r11) ldq_u r2,X+1(r11)
3294 lda r3,X(r11) lda r3,X+2(r11)
3295 extwl r1,r3,r1 extql r1,r3,r1
3296 extwh r2,r3,r2 extqh r2,r3,r2
3297 or r1.r2.r1 or r1,r2,r1
3300 long: ldq_u r1,X(r11) ldq_u r1,X(r11)
3301 ldq_u r2,X+3(r11) ldq_u r2,X+3(r11)
3302 lda r3,X(r11) lda r3,X(r11)
3303 extll r1,r3,r1 extll r1,r3,r1
3304 extlh r2,r3,r2 extlh r2,r3,r2
3305 or r1.r2.r1 addl r1,r2,r1
3307 quad: ldq_u r1,X(r11)
3316 alpha_expand_unaligned_load (rtx tgt
, rtx mem
, HOST_WIDE_INT size
,
3317 HOST_WIDE_INT ofs
, int sign
)
3319 rtx meml
, memh
, addr
, extl
, exth
, tmp
, mema
;
3320 enum machine_mode mode
;
3322 if (TARGET_BWX
&& size
== 2)
3324 meml
= adjust_address (mem
, QImode
, ofs
);
3325 memh
= adjust_address (mem
, QImode
, ofs
+1);
3326 if (BYTES_BIG_ENDIAN
)
3327 tmp
= meml
, meml
= memh
, memh
= tmp
;
3328 extl
= gen_reg_rtx (DImode
);
3329 exth
= gen_reg_rtx (DImode
);
3330 emit_insn (gen_zero_extendqidi2 (extl
, meml
));
3331 emit_insn (gen_zero_extendqidi2 (exth
, memh
));
3332 exth
= expand_simple_binop (DImode
, ASHIFT
, exth
, GEN_INT (8),
3333 NULL
, 1, OPTAB_LIB_WIDEN
);
3334 addr
= expand_simple_binop (DImode
, IOR
, extl
, exth
,
3335 NULL
, 1, OPTAB_LIB_WIDEN
);
3337 if (sign
&& GET_MODE (tgt
) != HImode
)
3339 addr
= gen_lowpart (HImode
, addr
);
3340 emit_insn (gen_extend_insn (tgt
, addr
, GET_MODE (tgt
), HImode
, 0));
3344 if (GET_MODE (tgt
) != DImode
)
3345 addr
= gen_lowpart (GET_MODE (tgt
), addr
);
3346 emit_move_insn (tgt
, addr
);
3351 meml
= gen_reg_rtx (DImode
);
3352 memh
= gen_reg_rtx (DImode
);
3353 addr
= gen_reg_rtx (DImode
);
3354 extl
= gen_reg_rtx (DImode
);
3355 exth
= gen_reg_rtx (DImode
);
3357 mema
= XEXP (mem
, 0);
3358 if (GET_CODE (mema
) == LO_SUM
)
3359 mema
= force_reg (Pmode
, mema
);
3361 /* AND addresses cannot be in any alias set, since they may implicitly
3362 alias surrounding code. Ideally we'd have some alias set that
3363 covered all types except those with alignment 8 or higher. */
3365 tmp
= change_address (mem
, DImode
,
3366 gen_rtx_AND (DImode
,
3367 plus_constant (mema
, ofs
),
3369 set_mem_alias_set (tmp
, 0);
3370 emit_move_insn (meml
, tmp
);
3372 tmp
= change_address (mem
, DImode
,
3373 gen_rtx_AND (DImode
,
3374 plus_constant (mema
, ofs
+ size
- 1),
3376 set_mem_alias_set (tmp
, 0);
3377 emit_move_insn (memh
, tmp
);
3379 if (WORDS_BIG_ENDIAN
&& sign
&& (size
== 2 || size
== 4))
3381 emit_move_insn (addr
, plus_constant (mema
, -1));
3383 emit_insn (gen_extqh_be (extl
, meml
, addr
));
3384 emit_insn (gen_extxl_be (exth
, memh
, GEN_INT (64), addr
));
3386 addr
= expand_binop (DImode
, ior_optab
, extl
, exth
, tgt
, 1, OPTAB_WIDEN
);
3387 addr
= expand_binop (DImode
, ashr_optab
, addr
, GEN_INT (64 - size
*8),
3388 addr
, 1, OPTAB_WIDEN
);
3390 else if (sign
&& size
== 2)
3392 emit_move_insn (addr
, plus_constant (mema
, ofs
+2));
3394 emit_insn (gen_extxl_le (extl
, meml
, GEN_INT (64), addr
));
3395 emit_insn (gen_extqh_le (exth
, memh
, addr
));
3397 /* We must use tgt here for the target. Alpha-vms port fails if we use
3398 addr for the target, because addr is marked as a pointer and combine
3399 knows that pointers are always sign-extended 32-bit values. */
3400 addr
= expand_binop (DImode
, ior_optab
, extl
, exth
, tgt
, 1, OPTAB_WIDEN
);
3401 addr
= expand_binop (DImode
, ashr_optab
, addr
, GEN_INT (48),
3402 addr
, 1, OPTAB_WIDEN
);
3406 if (WORDS_BIG_ENDIAN
)
3408 emit_move_insn (addr
, plus_constant (mema
, ofs
+size
-1));
3412 emit_insn (gen_extwh_be (extl
, meml
, addr
));
3417 emit_insn (gen_extlh_be (extl
, meml
, addr
));
3422 emit_insn (gen_extqh_be (extl
, meml
, addr
));
3429 emit_insn (gen_extxl_be (exth
, memh
, GEN_INT (size
*8), addr
));
3433 emit_move_insn (addr
, plus_constant (mema
, ofs
));
3434 emit_insn (gen_extxl_le (extl
, meml
, GEN_INT (size
*8), addr
));
3438 emit_insn (gen_extwh_le (exth
, memh
, addr
));
3443 emit_insn (gen_extlh_le (exth
, memh
, addr
));
3448 emit_insn (gen_extqh_le (exth
, memh
, addr
));
3457 addr
= expand_binop (mode
, ior_optab
, gen_lowpart (mode
, extl
),
3458 gen_lowpart (mode
, exth
), gen_lowpart (mode
, tgt
),
3463 emit_move_insn (tgt
, gen_lowpart (GET_MODE (tgt
), addr
));
3466 /* Similarly, use ins and msk instructions to perform unaligned stores. */
3469 alpha_expand_unaligned_store (rtx dst
, rtx src
,
3470 HOST_WIDE_INT size
, HOST_WIDE_INT ofs
)
3472 rtx dstl
, dsth
, addr
, insl
, insh
, meml
, memh
, dsta
;
3474 if (TARGET_BWX
&& size
== 2)
3476 if (src
!= const0_rtx
)
3478 dstl
= gen_lowpart (QImode
, src
);
3479 dsth
= expand_simple_binop (DImode
, LSHIFTRT
, src
, GEN_INT (8),
3480 NULL
, 1, OPTAB_LIB_WIDEN
);
3481 dsth
= gen_lowpart (QImode
, dsth
);
3484 dstl
= dsth
= const0_rtx
;
3486 meml
= adjust_address (dst
, QImode
, ofs
);
3487 memh
= adjust_address (dst
, QImode
, ofs
+1);
3488 if (BYTES_BIG_ENDIAN
)
3489 addr
= meml
, meml
= memh
, memh
= addr
;
3491 emit_move_insn (meml
, dstl
);
3492 emit_move_insn (memh
, dsth
);
3496 dstl
= gen_reg_rtx (DImode
);
3497 dsth
= gen_reg_rtx (DImode
);
3498 insl
= gen_reg_rtx (DImode
);
3499 insh
= gen_reg_rtx (DImode
);
3501 dsta
= XEXP (dst
, 0);
3502 if (GET_CODE (dsta
) == LO_SUM
)
3503 dsta
= force_reg (Pmode
, dsta
);
3505 /* AND addresses cannot be in any alias set, since they may implicitly
3506 alias surrounding code. Ideally we'd have some alias set that
3507 covered all types except those with alignment 8 or higher. */
3509 meml
= change_address (dst
, DImode
,
3510 gen_rtx_AND (DImode
,
3511 plus_constant (dsta
, ofs
),
3513 set_mem_alias_set (meml
, 0);
3515 memh
= change_address (dst
, DImode
,
3516 gen_rtx_AND (DImode
,
3517 plus_constant (dsta
, ofs
+ size
- 1),
3519 set_mem_alias_set (memh
, 0);
3521 emit_move_insn (dsth
, memh
);
3522 emit_move_insn (dstl
, meml
);
3523 if (WORDS_BIG_ENDIAN
)
3525 addr
= copy_addr_to_reg (plus_constant (dsta
, ofs
+size
-1));
3527 if (src
!= const0_rtx
)
3532 emit_insn (gen_inswl_be (insh
, gen_lowpart (HImode
,src
), addr
));
3535 emit_insn (gen_insll_be (insh
, gen_lowpart (SImode
,src
), addr
));
3538 emit_insn (gen_insql_be (insh
, gen_lowpart (DImode
,src
), addr
));
3541 emit_insn (gen_insxh (insl
, gen_lowpart (DImode
, src
),
3542 GEN_INT (size
*8), addr
));
3548 emit_insn (gen_mskxl_be (dsth
, dsth
, GEN_INT (0xffff), addr
));
3552 rtx msk
= immed_double_const (0xffffffff, 0, DImode
);
3553 emit_insn (gen_mskxl_be (dsth
, dsth
, msk
, addr
));
3557 emit_insn (gen_mskxl_be (dsth
, dsth
, constm1_rtx
, addr
));
3561 emit_insn (gen_mskxh (dstl
, dstl
, GEN_INT (size
*8), addr
));
3565 addr
= copy_addr_to_reg (plus_constant (dsta
, ofs
));
3567 if (src
!= CONST0_RTX (GET_MODE (src
)))
3569 emit_insn (gen_insxh (insh
, gen_lowpart (DImode
, src
),
3570 GEN_INT (size
*8), addr
));
3575 emit_insn (gen_inswl_le (insl
, gen_lowpart (HImode
, src
), addr
));
3578 emit_insn (gen_insll_le (insl
, gen_lowpart (SImode
, src
), addr
));
3581 emit_insn (gen_insql_le (insl
, gen_lowpart (DImode
, src
), addr
));
3586 emit_insn (gen_mskxh (dsth
, dsth
, GEN_INT (size
*8), addr
));
3591 emit_insn (gen_mskxl_le (dstl
, dstl
, GEN_INT (0xffff), addr
));
3595 rtx msk
= immed_double_const (0xffffffff, 0, DImode
);
3596 emit_insn (gen_mskxl_le (dstl
, dstl
, msk
, addr
));
3600 emit_insn (gen_mskxl_le (dstl
, dstl
, constm1_rtx
, addr
));
3605 if (src
!= CONST0_RTX (GET_MODE (src
)))
3607 dsth
= expand_binop (DImode
, ior_optab
, insh
, dsth
, dsth
, 0, OPTAB_WIDEN
);
3608 dstl
= expand_binop (DImode
, ior_optab
, insl
, dstl
, dstl
, 0, OPTAB_WIDEN
);
3611 if (WORDS_BIG_ENDIAN
)
3613 emit_move_insn (meml
, dstl
);
3614 emit_move_insn (memh
, dsth
);
3618 /* Must store high before low for degenerate case of aligned. */
3619 emit_move_insn (memh
, dsth
);
3620 emit_move_insn (meml
, dstl
);
3624 /* The block move code tries to maximize speed by separating loads and
3625 stores at the expense of register pressure: we load all of the data
3626 before we store it back out. There are two secondary effects worth
3627 mentioning, that this speeds copying to/from aligned and unaligned
3628 buffers, and that it makes the code significantly easier to write. */
3630 #define MAX_MOVE_WORDS 8
3632 /* Load an integral number of consecutive unaligned quadwords. */
3635 alpha_expand_unaligned_load_words (rtx
*out_regs
, rtx smem
,
3636 HOST_WIDE_INT words
, HOST_WIDE_INT ofs
)
3638 rtx
const im8
= GEN_INT (-8);
3639 rtx
const i64
= GEN_INT (64);
3640 rtx ext_tmps
[MAX_MOVE_WORDS
], data_regs
[MAX_MOVE_WORDS
+1];
3641 rtx sreg
, areg
, tmp
, smema
;
3644 smema
= XEXP (smem
, 0);
3645 if (GET_CODE (smema
) == LO_SUM
)
3646 smema
= force_reg (Pmode
, smema
);
3648 /* Generate all the tmp registers we need. */
3649 for (i
= 0; i
< words
; ++i
)
3651 data_regs
[i
] = out_regs
[i
];
3652 ext_tmps
[i
] = gen_reg_rtx (DImode
);
3654 data_regs
[words
] = gen_reg_rtx (DImode
);
3657 smem
= adjust_address (smem
, GET_MODE (smem
), ofs
);
3659 /* Load up all of the source data. */
3660 for (i
= 0; i
< words
; ++i
)
3662 tmp
= change_address (smem
, DImode
,
3663 gen_rtx_AND (DImode
,
3664 plus_constant (smema
, 8*i
),
3666 set_mem_alias_set (tmp
, 0);
3667 emit_move_insn (data_regs
[i
], tmp
);
3670 tmp
= change_address (smem
, DImode
,
3671 gen_rtx_AND (DImode
,
3672 plus_constant (smema
, 8*words
- 1),
3674 set_mem_alias_set (tmp
, 0);
3675 emit_move_insn (data_regs
[words
], tmp
);
3677 /* Extract the half-word fragments. Unfortunately DEC decided to make
3678 extxh with offset zero a noop instead of zeroing the register, so
3679 we must take care of that edge condition ourselves with cmov. */
3681 sreg
= copy_addr_to_reg (smema
);
3682 areg
= expand_binop (DImode
, and_optab
, sreg
, GEN_INT (7), NULL
,
3684 if (WORDS_BIG_ENDIAN
)
3685 emit_move_insn (sreg
, plus_constant (sreg
, 7));
3686 for (i
= 0; i
< words
; ++i
)
3688 if (WORDS_BIG_ENDIAN
)
3690 emit_insn (gen_extqh_be (data_regs
[i
], data_regs
[i
], sreg
));
3691 emit_insn (gen_extxl_be (ext_tmps
[i
], data_regs
[i
+1], i64
, sreg
));
3695 emit_insn (gen_extxl_le (data_regs
[i
], data_regs
[i
], i64
, sreg
));
3696 emit_insn (gen_extqh_le (ext_tmps
[i
], data_regs
[i
+1], sreg
));
3698 emit_insn (gen_rtx_SET (VOIDmode
, ext_tmps
[i
],
3699 gen_rtx_IF_THEN_ELSE (DImode
,
3700 gen_rtx_EQ (DImode
, areg
,
3702 const0_rtx
, ext_tmps
[i
])));
3705 /* Merge the half-words into whole words. */
3706 for (i
= 0; i
< words
; ++i
)
3708 out_regs
[i
] = expand_binop (DImode
, ior_optab
, data_regs
[i
],
3709 ext_tmps
[i
], data_regs
[i
], 1, OPTAB_WIDEN
);
3713 /* Store an integral number of consecutive unaligned quadwords. DATA_REGS
3714 may be NULL to store zeros. */
3717 alpha_expand_unaligned_store_words (rtx
*data_regs
, rtx dmem
,
3718 HOST_WIDE_INT words
, HOST_WIDE_INT ofs
)
3720 rtx
const im8
= GEN_INT (-8);
3721 rtx
const i64
= GEN_INT (64);
3722 rtx ins_tmps
[MAX_MOVE_WORDS
];
3723 rtx st_tmp_1
, st_tmp_2
, dreg
;
3724 rtx st_addr_1
, st_addr_2
, dmema
;
3727 dmema
= XEXP (dmem
, 0);
3728 if (GET_CODE (dmema
) == LO_SUM
)
3729 dmema
= force_reg (Pmode
, dmema
);
3731 /* Generate all the tmp registers we need. */
3732 if (data_regs
!= NULL
)
3733 for (i
= 0; i
< words
; ++i
)
3734 ins_tmps
[i
] = gen_reg_rtx(DImode
);
3735 st_tmp_1
= gen_reg_rtx(DImode
);
3736 st_tmp_2
= gen_reg_rtx(DImode
);
3739 dmem
= adjust_address (dmem
, GET_MODE (dmem
), ofs
);
3741 st_addr_2
= change_address (dmem
, DImode
,
3742 gen_rtx_AND (DImode
,
3743 plus_constant (dmema
, words
*8 - 1),
3745 set_mem_alias_set (st_addr_2
, 0);
3747 st_addr_1
= change_address (dmem
, DImode
,
3748 gen_rtx_AND (DImode
, dmema
, im8
));
3749 set_mem_alias_set (st_addr_1
, 0);
3751 /* Load up the destination end bits. */
3752 emit_move_insn (st_tmp_2
, st_addr_2
);
3753 emit_move_insn (st_tmp_1
, st_addr_1
);
3755 /* Shift the input data into place. */
3756 dreg
= copy_addr_to_reg (dmema
);
3757 if (WORDS_BIG_ENDIAN
)
3758 emit_move_insn (dreg
, plus_constant (dreg
, 7));
3759 if (data_regs
!= NULL
)
3761 for (i
= words
-1; i
>= 0; --i
)
3763 if (WORDS_BIG_ENDIAN
)
3765 emit_insn (gen_insql_be (ins_tmps
[i
], data_regs
[i
], dreg
));
3766 emit_insn (gen_insxh (data_regs
[i
], data_regs
[i
], i64
, dreg
));
3770 emit_insn (gen_insxh (ins_tmps
[i
], data_regs
[i
], i64
, dreg
));
3771 emit_insn (gen_insql_le (data_regs
[i
], data_regs
[i
], dreg
));
3774 for (i
= words
-1; i
> 0; --i
)
3776 ins_tmps
[i
-1] = expand_binop (DImode
, ior_optab
, data_regs
[i
],
3777 ins_tmps
[i
-1], ins_tmps
[i
-1], 1,
3782 /* Split and merge the ends with the destination data. */
3783 if (WORDS_BIG_ENDIAN
)
3785 emit_insn (gen_mskxl_be (st_tmp_2
, st_tmp_2
, constm1_rtx
, dreg
));
3786 emit_insn (gen_mskxh (st_tmp_1
, st_tmp_1
, i64
, dreg
));
3790 emit_insn (gen_mskxh (st_tmp_2
, st_tmp_2
, i64
, dreg
));
3791 emit_insn (gen_mskxl_le (st_tmp_1
, st_tmp_1
, constm1_rtx
, dreg
));
3794 if (data_regs
!= NULL
)
3796 st_tmp_2
= expand_binop (DImode
, ior_optab
, st_tmp_2
, ins_tmps
[words
-1],
3797 st_tmp_2
, 1, OPTAB_WIDEN
);
3798 st_tmp_1
= expand_binop (DImode
, ior_optab
, st_tmp_1
, data_regs
[0],
3799 st_tmp_1
, 1, OPTAB_WIDEN
);
3803 if (WORDS_BIG_ENDIAN
)
3804 emit_move_insn (st_addr_1
, st_tmp_1
);
3806 emit_move_insn (st_addr_2
, st_tmp_2
);
3807 for (i
= words
-1; i
> 0; --i
)
3809 rtx tmp
= change_address (dmem
, DImode
,
3810 gen_rtx_AND (DImode
,
3811 plus_constant(dmema
,
3812 WORDS_BIG_ENDIAN
? i
*8-1 : i
*8),
3814 set_mem_alias_set (tmp
, 0);
3815 emit_move_insn (tmp
, data_regs
? ins_tmps
[i
-1] : const0_rtx
);
3817 if (WORDS_BIG_ENDIAN
)
3818 emit_move_insn (st_addr_2
, st_tmp_2
);
3820 emit_move_insn (st_addr_1
, st_tmp_1
);
3824 /* Expand string/block move operations.
3826 operands[0] is the pointer to the destination.
3827 operands[1] is the pointer to the source.
3828 operands[2] is the number of bytes to move.
3829 operands[3] is the alignment. */
3832 alpha_expand_block_move (rtx operands
[])
3834 rtx bytes_rtx
= operands
[2];
3835 rtx align_rtx
= operands
[3];
3836 HOST_WIDE_INT orig_bytes
= INTVAL (bytes_rtx
);
3837 HOST_WIDE_INT bytes
= orig_bytes
;
3838 HOST_WIDE_INT src_align
= INTVAL (align_rtx
) * BITS_PER_UNIT
;
3839 HOST_WIDE_INT dst_align
= src_align
;
3840 rtx orig_src
= operands
[1];
3841 rtx orig_dst
= operands
[0];
3842 rtx data_regs
[2 * MAX_MOVE_WORDS
+ 16];
3844 unsigned int i
, words
, ofs
, nregs
= 0;
3846 if (orig_bytes
<= 0)
3848 else if (orig_bytes
> MAX_MOVE_WORDS
* UNITS_PER_WORD
)
3851 /* Look for additional alignment information from recorded register info. */
3853 tmp
= XEXP (orig_src
, 0);
3855 src_align
= MAX (src_align
, REGNO_POINTER_ALIGN (REGNO (tmp
)));
3856 else if (GET_CODE (tmp
) == PLUS
3857 && REG_P (XEXP (tmp
, 0))
3858 && CONST_INT_P (XEXP (tmp
, 1)))
3860 unsigned HOST_WIDE_INT c
= INTVAL (XEXP (tmp
, 1));
3861 unsigned int a
= REGNO_POINTER_ALIGN (REGNO (XEXP (tmp
, 0)));
3865 if (a
>= 64 && c
% 8 == 0)
3867 else if (a
>= 32 && c
% 4 == 0)
3869 else if (a
>= 16 && c
% 2 == 0)
3874 tmp
= XEXP (orig_dst
, 0);
3876 dst_align
= MAX (dst_align
, REGNO_POINTER_ALIGN (REGNO (tmp
)));
3877 else if (GET_CODE (tmp
) == PLUS
3878 && REG_P (XEXP (tmp
, 0))
3879 && CONST_INT_P (XEXP (tmp
, 1)))
3881 unsigned HOST_WIDE_INT c
= INTVAL (XEXP (tmp
, 1));
3882 unsigned int a
= REGNO_POINTER_ALIGN (REGNO (XEXP (tmp
, 0)));
3886 if (a
>= 64 && c
% 8 == 0)
3888 else if (a
>= 32 && c
% 4 == 0)
3890 else if (a
>= 16 && c
% 2 == 0)
3896 if (src_align
>= 64 && bytes
>= 8)
3900 for (i
= 0; i
< words
; ++i
)
3901 data_regs
[nregs
+ i
] = gen_reg_rtx (DImode
);
3903 for (i
= 0; i
< words
; ++i
)
3904 emit_move_insn (data_regs
[nregs
+ i
],
3905 adjust_address (orig_src
, DImode
, ofs
+ i
* 8));
3912 if (src_align
>= 32 && bytes
>= 4)
3916 for (i
= 0; i
< words
; ++i
)
3917 data_regs
[nregs
+ i
] = gen_reg_rtx (SImode
);
3919 for (i
= 0; i
< words
; ++i
)
3920 emit_move_insn (data_regs
[nregs
+ i
],
3921 adjust_address (orig_src
, SImode
, ofs
+ i
* 4));
3932 for (i
= 0; i
< words
+1; ++i
)
3933 data_regs
[nregs
+ i
] = gen_reg_rtx (DImode
);
3935 alpha_expand_unaligned_load_words (data_regs
+ nregs
, orig_src
,
3943 if (! TARGET_BWX
&& bytes
>= 4)
3945 data_regs
[nregs
++] = tmp
= gen_reg_rtx (SImode
);
3946 alpha_expand_unaligned_load (tmp
, orig_src
, 4, ofs
, 0);
3953 if (src_align
>= 16)
3956 data_regs
[nregs
++] = tmp
= gen_reg_rtx (HImode
);
3957 emit_move_insn (tmp
, adjust_address (orig_src
, HImode
, ofs
));
3960 } while (bytes
>= 2);
3962 else if (! TARGET_BWX
)
3964 data_regs
[nregs
++] = tmp
= gen_reg_rtx (HImode
);
3965 alpha_expand_unaligned_load (tmp
, orig_src
, 2, ofs
, 0);
3973 data_regs
[nregs
++] = tmp
= gen_reg_rtx (QImode
);
3974 emit_move_insn (tmp
, adjust_address (orig_src
, QImode
, ofs
));
3979 gcc_assert (nregs
<= ARRAY_SIZE (data_regs
));
3981 /* Now save it back out again. */
3985 /* Write out the data in whatever chunks reading the source allowed. */
3986 if (dst_align
>= 64)
3988 while (i
< nregs
&& GET_MODE (data_regs
[i
]) == DImode
)
3990 emit_move_insn (adjust_address (orig_dst
, DImode
, ofs
),
3997 if (dst_align
>= 32)
3999 /* If the source has remaining DImode regs, write them out in
4001 while (i
< nregs
&& GET_MODE (data_regs
[i
]) == DImode
)
4003 tmp
= expand_binop (DImode
, lshr_optab
, data_regs
[i
], GEN_INT (32),
4004 NULL_RTX
, 1, OPTAB_WIDEN
);
4006 emit_move_insn (adjust_address (orig_dst
, SImode
, ofs
),
4007 gen_lowpart (SImode
, data_regs
[i
]));
4008 emit_move_insn (adjust_address (orig_dst
, SImode
, ofs
+ 4),
4009 gen_lowpart (SImode
, tmp
));
4014 while (i
< nregs
&& GET_MODE (data_regs
[i
]) == SImode
)
4016 emit_move_insn (adjust_address (orig_dst
, SImode
, ofs
),
4023 if (i
< nregs
&& GET_MODE (data_regs
[i
]) == DImode
)
4025 /* Write out a remaining block of words using unaligned methods. */
4027 for (words
= 1; i
+ words
< nregs
; words
++)
4028 if (GET_MODE (data_regs
[i
+ words
]) != DImode
)
4032 alpha_expand_unaligned_store (orig_dst
, data_regs
[i
], 8, ofs
);
4034 alpha_expand_unaligned_store_words (data_regs
+ i
, orig_dst
,
4041 /* Due to the above, this won't be aligned. */
4042 /* ??? If we have more than one of these, consider constructing full
4043 words in registers and using alpha_expand_unaligned_store_words. */
4044 while (i
< nregs
&& GET_MODE (data_regs
[i
]) == SImode
)
4046 alpha_expand_unaligned_store (orig_dst
, data_regs
[i
], 4, ofs
);
4051 if (dst_align
>= 16)
4052 while (i
< nregs
&& GET_MODE (data_regs
[i
]) == HImode
)
4054 emit_move_insn (adjust_address (orig_dst
, HImode
, ofs
), data_regs
[i
]);
4059 while (i
< nregs
&& GET_MODE (data_regs
[i
]) == HImode
)
4061 alpha_expand_unaligned_store (orig_dst
, data_regs
[i
], 2, ofs
);
4066 /* The remainder must be byte copies. */
4069 gcc_assert (GET_MODE (data_regs
[i
]) == QImode
);
4070 emit_move_insn (adjust_address (orig_dst
, QImode
, ofs
), data_regs
[i
]);
4079 alpha_expand_block_clear (rtx operands
[])
4081 rtx bytes_rtx
= operands
[1];
4082 rtx align_rtx
= operands
[3];
4083 HOST_WIDE_INT orig_bytes
= INTVAL (bytes_rtx
);
4084 HOST_WIDE_INT bytes
= orig_bytes
;
4085 HOST_WIDE_INT align
= INTVAL (align_rtx
) * BITS_PER_UNIT
;
4086 HOST_WIDE_INT alignofs
= 0;
4087 rtx orig_dst
= operands
[0];
4089 int i
, words
, ofs
= 0;
4091 if (orig_bytes
<= 0)
4093 if (orig_bytes
> MAX_MOVE_WORDS
* UNITS_PER_WORD
)
4096 /* Look for stricter alignment. */
4097 tmp
= XEXP (orig_dst
, 0);
4099 align
= MAX (align
, REGNO_POINTER_ALIGN (REGNO (tmp
)));
4100 else if (GET_CODE (tmp
) == PLUS
4101 && REG_P (XEXP (tmp
, 0))
4102 && CONST_INT_P (XEXP (tmp
, 1)))
4104 HOST_WIDE_INT c
= INTVAL (XEXP (tmp
, 1));
4105 int a
= REGNO_POINTER_ALIGN (REGNO (XEXP (tmp
, 0)));
4110 align
= a
, alignofs
= 8 - c
% 8;
4112 align
= a
, alignofs
= 4 - c
% 4;
4114 align
= a
, alignofs
= 2 - c
% 2;
4118 /* Handle an unaligned prefix first. */
4122 #if HOST_BITS_PER_WIDE_INT >= 64
4123 /* Given that alignofs is bounded by align, the only time BWX could
4124 generate three stores is for a 7 byte fill. Prefer two individual
4125 stores over a load/mask/store sequence. */
4126 if ((!TARGET_BWX
|| alignofs
== 7)
4128 && !(alignofs
== 4 && bytes
>= 4))
4130 enum machine_mode mode
= (align
>= 64 ? DImode
: SImode
);
4131 int inv_alignofs
= (align
>= 64 ? 8 : 4) - alignofs
;
4135 mem
= adjust_address (orig_dst
, mode
, ofs
- inv_alignofs
);
4136 set_mem_alias_set (mem
, 0);
4138 mask
= ~(~(HOST_WIDE_INT
)0 << (inv_alignofs
* 8));
4139 if (bytes
< alignofs
)
4141 mask
|= ~(HOST_WIDE_INT
)0 << ((inv_alignofs
+ bytes
) * 8);
4152 tmp
= expand_binop (mode
, and_optab
, mem
, GEN_INT (mask
),
4153 NULL_RTX
, 1, OPTAB_WIDEN
);
4155 emit_move_insn (mem
, tmp
);
4159 if (TARGET_BWX
&& (alignofs
& 1) && bytes
>= 1)
4161 emit_move_insn (adjust_address (orig_dst
, QImode
, ofs
), const0_rtx
);
4166 if (TARGET_BWX
&& align
>= 16 && (alignofs
& 3) == 2 && bytes
>= 2)
4168 emit_move_insn (adjust_address (orig_dst
, HImode
, ofs
), const0_rtx
);
4173 if (alignofs
== 4 && bytes
>= 4)
4175 emit_move_insn (adjust_address (orig_dst
, SImode
, ofs
), const0_rtx
);
4181 /* If we've not used the extra lead alignment information by now,
4182 we won't be able to. Downgrade align to match what's left over. */
4185 alignofs
= alignofs
& -alignofs
;
4186 align
= MIN (align
, alignofs
* BITS_PER_UNIT
);
4190 /* Handle a block of contiguous long-words. */
4192 if (align
>= 64 && bytes
>= 8)
4196 for (i
= 0; i
< words
; ++i
)
4197 emit_move_insn (adjust_address (orig_dst
, DImode
, ofs
+ i
* 8),
4204 /* If the block is large and appropriately aligned, emit a single
4205 store followed by a sequence of stq_u insns. */
4207 if (align
>= 32 && bytes
> 16)
4211 emit_move_insn (adjust_address (orig_dst
, SImode
, ofs
), const0_rtx
);
4215 orig_dsta
= XEXP (orig_dst
, 0);
4216 if (GET_CODE (orig_dsta
) == LO_SUM
)
4217 orig_dsta
= force_reg (Pmode
, orig_dsta
);
4220 for (i
= 0; i
< words
; ++i
)
4223 = change_address (orig_dst
, DImode
,
4224 gen_rtx_AND (DImode
,
4225 plus_constant (orig_dsta
, ofs
+ i
*8),
4227 set_mem_alias_set (mem
, 0);
4228 emit_move_insn (mem
, const0_rtx
);
4231 /* Depending on the alignment, the first stq_u may have overlapped
4232 with the initial stl, which means that the last stq_u didn't
4233 write as much as it would appear. Leave those questionable bytes
4235 bytes
-= words
* 8 - 4;
4236 ofs
+= words
* 8 - 4;
4239 /* Handle a smaller block of aligned words. */
4241 if ((align
>= 64 && bytes
== 4)
4242 || (align
== 32 && bytes
>= 4))
4246 for (i
= 0; i
< words
; ++i
)
4247 emit_move_insn (adjust_address (orig_dst
, SImode
, ofs
+ i
* 4),
4254 /* An unaligned block uses stq_u stores for as many as possible. */
4260 alpha_expand_unaligned_store_words (NULL
, orig_dst
, words
, ofs
);
4266 /* Next clean up any trailing pieces. */
4268 #if HOST_BITS_PER_WIDE_INT >= 64
4269 /* Count the number of bits in BYTES for which aligned stores could
4272 for (i
= (TARGET_BWX
? 1 : 4); i
* BITS_PER_UNIT
<= align
; i
<<= 1)
4276 /* If we have appropriate alignment (and it wouldn't take too many
4277 instructions otherwise), mask out the bytes we need. */
4278 if (TARGET_BWX
? words
> 2 : bytes
> 0)
4285 mem
= adjust_address (orig_dst
, DImode
, ofs
);
4286 set_mem_alias_set (mem
, 0);
4288 mask
= ~(HOST_WIDE_INT
)0 << (bytes
* 8);
4290 tmp
= expand_binop (DImode
, and_optab
, mem
, GEN_INT (mask
),
4291 NULL_RTX
, 1, OPTAB_WIDEN
);
4293 emit_move_insn (mem
, tmp
);
4296 else if (align
>= 32 && bytes
< 4)
4301 mem
= adjust_address (orig_dst
, SImode
, ofs
);
4302 set_mem_alias_set (mem
, 0);
4304 mask
= ~(HOST_WIDE_INT
)0 << (bytes
* 8);
4306 tmp
= expand_binop (SImode
, and_optab
, mem
, GEN_INT (mask
),
4307 NULL_RTX
, 1, OPTAB_WIDEN
);
4309 emit_move_insn (mem
, tmp
);
4315 if (!TARGET_BWX
&& bytes
>= 4)
4317 alpha_expand_unaligned_store (orig_dst
, const0_rtx
, 4, ofs
);
4327 emit_move_insn (adjust_address (orig_dst
, HImode
, ofs
),
4331 } while (bytes
>= 2);
4333 else if (! TARGET_BWX
)
4335 alpha_expand_unaligned_store (orig_dst
, const0_rtx
, 2, ofs
);
4343 emit_move_insn (adjust_address (orig_dst
, QImode
, ofs
), const0_rtx
);
4351 /* Returns a mask so that zap(x, value) == x & mask. */
4354 alpha_expand_zap_mask (HOST_WIDE_INT value
)
4359 if (HOST_BITS_PER_WIDE_INT
>= 64)
4361 HOST_WIDE_INT mask
= 0;
4363 for (i
= 7; i
>= 0; --i
)
4366 if (!((value
>> i
) & 1))
4370 result
= gen_int_mode (mask
, DImode
);
4374 HOST_WIDE_INT mask_lo
= 0, mask_hi
= 0;
4376 gcc_assert (HOST_BITS_PER_WIDE_INT
== 32);
4378 for (i
= 7; i
>= 4; --i
)
4381 if (!((value
>> i
) & 1))
4385 for (i
= 3; i
>= 0; --i
)
4388 if (!((value
>> i
) & 1))
4392 result
= immed_double_const (mask_lo
, mask_hi
, DImode
);
4399 alpha_expand_builtin_vector_binop (rtx (*gen
) (rtx
, rtx
, rtx
),
4400 enum machine_mode mode
,
4401 rtx op0
, rtx op1
, rtx op2
)
4403 op0
= gen_lowpart (mode
, op0
);
4405 if (op1
== const0_rtx
)
4406 op1
= CONST0_RTX (mode
);
4408 op1
= gen_lowpart (mode
, op1
);
4410 if (op2
== const0_rtx
)
4411 op2
= CONST0_RTX (mode
);
4413 op2
= gen_lowpart (mode
, op2
);
4415 emit_insn ((*gen
) (op0
, op1
, op2
));
4418 /* A subroutine of the atomic operation splitters. Jump to LABEL if
4419 COND is true. Mark the jump as unlikely to be taken. */
4422 emit_unlikely_jump (rtx cond
, rtx label
)
4424 rtx very_unlikely
= GEN_INT (REG_BR_PROB_BASE
/ 100 - 1);
4427 x
= gen_rtx_IF_THEN_ELSE (VOIDmode
, cond
, label
, pc_rtx
);
4428 x
= emit_jump_insn (gen_rtx_SET (VOIDmode
, pc_rtx
, x
));
4429 add_reg_note (x
, REG_BR_PROB
, very_unlikely
);
4432 /* A subroutine of the atomic operation splitters. Emit a load-locked
4433 instruction in MODE. */
4436 emit_load_locked (enum machine_mode mode
, rtx reg
, rtx mem
)
4438 rtx (*fn
) (rtx
, rtx
) = NULL
;
4440 fn
= gen_load_locked_si
;
4441 else if (mode
== DImode
)
4442 fn
= gen_load_locked_di
;
4443 emit_insn (fn (reg
, mem
));
4446 /* A subroutine of the atomic operation splitters. Emit a store-conditional
4447 instruction in MODE. */
4450 emit_store_conditional (enum machine_mode mode
, rtx res
, rtx mem
, rtx val
)
4452 rtx (*fn
) (rtx
, rtx
, rtx
) = NULL
;
4454 fn
= gen_store_conditional_si
;
4455 else if (mode
== DImode
)
4456 fn
= gen_store_conditional_di
;
4457 emit_insn (fn (res
, mem
, val
));
4460 /* A subroutine of the atomic operation splitters. Emit an insxl
4461 instruction in MODE. */
4464 emit_insxl (enum machine_mode mode
, rtx op1
, rtx op2
)
4466 rtx ret
= gen_reg_rtx (DImode
);
4467 rtx (*fn
) (rtx
, rtx
, rtx
);
4469 if (WORDS_BIG_ENDIAN
)
4483 /* The insbl and inswl patterns require a register operand. */
4484 op1
= force_reg (mode
, op1
);
4485 emit_insn (fn (ret
, op1
, op2
));
4490 /* Expand an atomic fetch-and-operate pattern. CODE is the binary operation
4491 to perform. MEM is the memory on which to operate. VAL is the second
4492 operand of the binary operator. BEFORE and AFTER are optional locations to
4493 return the value of MEM either before of after the operation. SCRATCH is
4494 a scratch register. */
4497 alpha_split_atomic_op (enum rtx_code code
, rtx mem
, rtx val
,
4498 rtx before
, rtx after
, rtx scratch
)
4500 enum machine_mode mode
= GET_MODE (mem
);
4501 rtx label
, x
, cond
= gen_rtx_REG (DImode
, REGNO (scratch
));
4503 emit_insn (gen_memory_barrier ());
4505 label
= gen_label_rtx ();
4507 label
= gen_rtx_LABEL_REF (DImode
, label
);
4511 emit_load_locked (mode
, before
, mem
);
4515 x
= gen_rtx_AND (mode
, before
, val
);
4516 emit_insn (gen_rtx_SET (VOIDmode
, val
, x
));
4518 x
= gen_rtx_NOT (mode
, val
);
4521 x
= gen_rtx_fmt_ee (code
, mode
, before
, val
);
4523 emit_insn (gen_rtx_SET (VOIDmode
, after
, copy_rtx (x
)));
4524 emit_insn (gen_rtx_SET (VOIDmode
, scratch
, x
));
4526 emit_store_conditional (mode
, cond
, mem
, scratch
);
4528 x
= gen_rtx_EQ (DImode
, cond
, const0_rtx
);
4529 emit_unlikely_jump (x
, label
);
4531 emit_insn (gen_memory_barrier ());
4534 /* Expand a compare and swap operation. */
4537 alpha_split_compare_and_swap (rtx retval
, rtx mem
, rtx oldval
, rtx newval
,
4540 enum machine_mode mode
= GET_MODE (mem
);
4541 rtx label1
, label2
, x
, cond
= gen_lowpart (DImode
, scratch
);
4543 emit_insn (gen_memory_barrier ());
4545 label1
= gen_rtx_LABEL_REF (DImode
, gen_label_rtx ());
4546 label2
= gen_rtx_LABEL_REF (DImode
, gen_label_rtx ());
4547 emit_label (XEXP (label1
, 0));
4549 emit_load_locked (mode
, retval
, mem
);
4551 x
= gen_lowpart (DImode
, retval
);
4552 if (oldval
== const0_rtx
)
4553 x
= gen_rtx_NE (DImode
, x
, const0_rtx
);
4556 x
= gen_rtx_EQ (DImode
, x
, oldval
);
4557 emit_insn (gen_rtx_SET (VOIDmode
, cond
, x
));
4558 x
= gen_rtx_EQ (DImode
, cond
, const0_rtx
);
4560 emit_unlikely_jump (x
, label2
);
4562 emit_move_insn (scratch
, newval
);
4563 emit_store_conditional (mode
, cond
, mem
, scratch
);
4565 x
= gen_rtx_EQ (DImode
, cond
, const0_rtx
);
4566 emit_unlikely_jump (x
, label1
);
4568 emit_insn (gen_memory_barrier ());
4569 emit_label (XEXP (label2
, 0));
4573 alpha_expand_compare_and_swap_12 (rtx dst
, rtx mem
, rtx oldval
, rtx newval
)
4575 enum machine_mode mode
= GET_MODE (mem
);
4576 rtx addr
, align
, wdst
;
4577 rtx (*fn5
) (rtx
, rtx
, rtx
, rtx
, rtx
);
4579 addr
= force_reg (DImode
, XEXP (mem
, 0));
4580 align
= expand_simple_binop (Pmode
, AND
, addr
, GEN_INT (-8),
4581 NULL_RTX
, 1, OPTAB_DIRECT
);
4583 oldval
= convert_modes (DImode
, mode
, oldval
, 1);
4584 newval
= emit_insxl (mode
, newval
, addr
);
4586 wdst
= gen_reg_rtx (DImode
);
4588 fn5
= gen_sync_compare_and_swapqi_1
;
4590 fn5
= gen_sync_compare_and_swaphi_1
;
4591 emit_insn (fn5 (wdst
, addr
, oldval
, newval
, align
));
4593 emit_move_insn (dst
, gen_lowpart (mode
, wdst
));
4597 alpha_split_compare_and_swap_12 (enum machine_mode mode
, rtx dest
, rtx addr
,
4598 rtx oldval
, rtx newval
, rtx align
,
4599 rtx scratch
, rtx cond
)
4601 rtx label1
, label2
, mem
, width
, mask
, x
;
4603 mem
= gen_rtx_MEM (DImode
, align
);
4604 MEM_VOLATILE_P (mem
) = 1;
4606 emit_insn (gen_memory_barrier ());
4607 label1
= gen_rtx_LABEL_REF (DImode
, gen_label_rtx ());
4608 label2
= gen_rtx_LABEL_REF (DImode
, gen_label_rtx ());
4609 emit_label (XEXP (label1
, 0));
4611 emit_load_locked (DImode
, scratch
, mem
);
4613 width
= GEN_INT (GET_MODE_BITSIZE (mode
));
4614 mask
= GEN_INT (mode
== QImode
? 0xff : 0xffff);
4615 if (WORDS_BIG_ENDIAN
)
4616 emit_insn (gen_extxl_be (dest
, scratch
, width
, addr
));
4618 emit_insn (gen_extxl_le (dest
, scratch
, width
, addr
));
4620 if (oldval
== const0_rtx
)
4621 x
= gen_rtx_NE (DImode
, dest
, const0_rtx
);
4624 x
= gen_rtx_EQ (DImode
, dest
, oldval
);
4625 emit_insn (gen_rtx_SET (VOIDmode
, cond
, x
));
4626 x
= gen_rtx_EQ (DImode
, cond
, const0_rtx
);
4628 emit_unlikely_jump (x
, label2
);
4630 if (WORDS_BIG_ENDIAN
)
4631 emit_insn (gen_mskxl_be (scratch
, scratch
, mask
, addr
));
4633 emit_insn (gen_mskxl_le (scratch
, scratch
, mask
, addr
));
4634 emit_insn (gen_iordi3 (scratch
, scratch
, newval
));
4636 emit_store_conditional (DImode
, scratch
, mem
, scratch
);
4638 x
= gen_rtx_EQ (DImode
, scratch
, const0_rtx
);
4639 emit_unlikely_jump (x
, label1
);
4641 emit_insn (gen_memory_barrier ());
4642 emit_label (XEXP (label2
, 0));
4645 /* Expand an atomic exchange operation. */
4648 alpha_split_lock_test_and_set (rtx retval
, rtx mem
, rtx val
, rtx scratch
)
4650 enum machine_mode mode
= GET_MODE (mem
);
4651 rtx label
, x
, cond
= gen_lowpart (DImode
, scratch
);
4653 label
= gen_rtx_LABEL_REF (DImode
, gen_label_rtx ());
4654 emit_label (XEXP (label
, 0));
4656 emit_load_locked (mode
, retval
, mem
);
4657 emit_move_insn (scratch
, val
);
4658 emit_store_conditional (mode
, cond
, mem
, scratch
);
4660 x
= gen_rtx_EQ (DImode
, cond
, const0_rtx
);
4661 emit_unlikely_jump (x
, label
);
4663 emit_insn (gen_memory_barrier ());
4667 alpha_expand_lock_test_and_set_12 (rtx dst
, rtx mem
, rtx val
)
4669 enum machine_mode mode
= GET_MODE (mem
);
4670 rtx addr
, align
, wdst
;
4671 rtx (*fn4
) (rtx
, rtx
, rtx
, rtx
);
4673 /* Force the address into a register. */
4674 addr
= force_reg (DImode
, XEXP (mem
, 0));
4676 /* Align it to a multiple of 8. */
4677 align
= expand_simple_binop (Pmode
, AND
, addr
, GEN_INT (-8),
4678 NULL_RTX
, 1, OPTAB_DIRECT
);
4680 /* Insert val into the correct byte location within the word. */
4681 val
= emit_insxl (mode
, val
, addr
);
4683 wdst
= gen_reg_rtx (DImode
);
4685 fn4
= gen_sync_lock_test_and_setqi_1
;
4687 fn4
= gen_sync_lock_test_and_sethi_1
;
4688 emit_insn (fn4 (wdst
, addr
, val
, align
));
4690 emit_move_insn (dst
, gen_lowpart (mode
, wdst
));
4694 alpha_split_lock_test_and_set_12 (enum machine_mode mode
, rtx dest
, rtx addr
,
4695 rtx val
, rtx align
, rtx scratch
)
4697 rtx label
, mem
, width
, mask
, x
;
4699 mem
= gen_rtx_MEM (DImode
, align
);
4700 MEM_VOLATILE_P (mem
) = 1;
4702 label
= gen_rtx_LABEL_REF (DImode
, gen_label_rtx ());
4703 emit_label (XEXP (label
, 0));
4705 emit_load_locked (DImode
, scratch
, mem
);
4707 width
= GEN_INT (GET_MODE_BITSIZE (mode
));
4708 mask
= GEN_INT (mode
== QImode
? 0xff : 0xffff);
4709 if (WORDS_BIG_ENDIAN
)
4711 emit_insn (gen_extxl_be (dest
, scratch
, width
, addr
));
4712 emit_insn (gen_mskxl_be (scratch
, scratch
, mask
, addr
));
4716 emit_insn (gen_extxl_le (dest
, scratch
, width
, addr
));
4717 emit_insn (gen_mskxl_le (scratch
, scratch
, mask
, addr
));
4719 emit_insn (gen_iordi3 (scratch
, scratch
, val
));
4721 emit_store_conditional (DImode
, scratch
, mem
, scratch
);
4723 x
= gen_rtx_EQ (DImode
, scratch
, const0_rtx
);
4724 emit_unlikely_jump (x
, label
);
4726 emit_insn (gen_memory_barrier ());
4729 /* Adjust the cost of a scheduling dependency. Return the new cost of
4730 a dependency LINK or INSN on DEP_INSN. COST is the current cost. */
4733 alpha_adjust_cost (rtx insn
, rtx link
, rtx dep_insn
, int cost
)
4735 enum attr_type dep_insn_type
;
4737 /* If the dependence is an anti-dependence, there is no cost. For an
4738 output dependence, there is sometimes a cost, but it doesn't seem
4739 worth handling those few cases. */
4740 if (REG_NOTE_KIND (link
) != 0)
4743 /* If we can't recognize the insns, we can't really do anything. */
4744 if (recog_memoized (insn
) < 0 || recog_memoized (dep_insn
) < 0)
4747 dep_insn_type
= get_attr_type (dep_insn
);
4749 /* Bring in the user-defined memory latency. */
4750 if (dep_insn_type
== TYPE_ILD
4751 || dep_insn_type
== TYPE_FLD
4752 || dep_insn_type
== TYPE_LDSYM
)
4753 cost
+= alpha_memory_latency
-1;
4755 /* Everything else handled in DFA bypasses now. */
4760 /* The number of instructions that can be issued per cycle. */
4763 alpha_issue_rate (void)
4765 return (alpha_tune
== PROCESSOR_EV4
? 2 : 4);
4768 /* How many alternative schedules to try. This should be as wide as the
4769 scheduling freedom in the DFA, but no wider. Making this value too
4770 large results extra work for the scheduler.
4772 For EV4, loads can be issued to either IB0 or IB1, thus we have 2
4773 alternative schedules. For EV5, we can choose between E0/E1 and
4774 FA/FM. For EV6, an arithmetic insn can be issued to U0/U1/L0/L1. */
4777 alpha_multipass_dfa_lookahead (void)
4779 return (alpha_tune
== PROCESSOR_EV6
? 4 : 2);
4782 /* Machine-specific function data. */
4784 struct GTY(()) machine_function
4787 /* List of call information words for calls from this function. */
4788 struct rtx_def
*first_ciw
;
4789 struct rtx_def
*last_ciw
;
4792 /* List of deferred case vectors. */
4793 struct rtx_def
*addr_list
;
4796 const char *some_ld_name
;
4798 /* For TARGET_LD_BUGGY_LDGP. */
4799 struct rtx_def
*gp_save_rtx
;
4801 /* For VMS condition handlers. */
4802 bool uses_condition_handler
;
4805 /* How to allocate a 'struct machine_function'. */
4807 static struct machine_function
*
4808 alpha_init_machine_status (void)
4810 return ((struct machine_function
*)
4811 ggc_alloc_cleared (sizeof (struct machine_function
)));
4814 /* Support for frame based VMS condition handlers. */
4816 /* A VMS condition handler may be established for a function with a call to
4817 __builtin_establish_vms_condition_handler, and cancelled with a call to
4818 __builtin_revert_vms_condition_handler.
4820 The VMS Condition Handling Facility knows about the existence of a handler
4821 from the procedure descriptor .handler field. As the VMS native compilers,
4822 we store the user specified handler's address at a fixed location in the
4823 stack frame and point the procedure descriptor at a common wrapper which
4824 fetches the real handler's address and issues an indirect call.
4826 The indirection wrapper is "__gcc_shell_handler", provided by libgcc.
4828 We force the procedure kind to PT_STACK, and the fixed frame location is
4829 fp+8, just before the register save area. We use the handler_data field in
4830 the procedure descriptor to state the fp offset at which the installed
4831 handler address can be found. */
4833 #define VMS_COND_HANDLER_FP_OFFSET 8
4835 /* Expand code to store the currently installed user VMS condition handler
4836 into TARGET and install HANDLER as the new condition handler. */
4839 alpha_expand_builtin_establish_vms_condition_handler (rtx target
, rtx handler
)
4841 rtx handler_slot_address
4842 = plus_constant (hard_frame_pointer_rtx
, VMS_COND_HANDLER_FP_OFFSET
);
4845 = gen_rtx_MEM (DImode
, handler_slot_address
);
4847 emit_move_insn (target
, handler_slot
);
4848 emit_move_insn (handler_slot
, handler
);
4850 /* Notify the start/prologue/epilogue emitters that the condition handler
4851 slot is needed. In addition to reserving the slot space, this will force
4852 the procedure kind to PT_STACK so ensure that the hard_frame_pointer_rtx
4853 use above is correct. */
4854 cfun
->machine
->uses_condition_handler
= true;
4857 /* Expand code to store the current VMS condition handler into TARGET and
4861 alpha_expand_builtin_revert_vms_condition_handler (rtx target
)
4863 /* We implement this by establishing a null condition handler, with the tiny
4864 side effect of setting uses_condition_handler. This is a little bit
4865 pessimistic if no actual builtin_establish call is ever issued, which is
4866 not a real problem and expected never to happen anyway. */
4868 alpha_expand_builtin_establish_vms_condition_handler (target
, const0_rtx
);
4871 /* Functions to save and restore alpha_return_addr_rtx. */
4873 /* Start the ball rolling with RETURN_ADDR_RTX. */
4876 alpha_return_addr (int count
, rtx frame ATTRIBUTE_UNUSED
)
4881 return get_hard_reg_initial_val (Pmode
, REG_RA
);
4884 /* Return or create a memory slot containing the gp value for the current
4885 function. Needed only if TARGET_LD_BUGGY_LDGP. */
4888 alpha_gp_save_rtx (void)
4890 rtx seq
, m
= cfun
->machine
->gp_save_rtx
;
4896 m
= assign_stack_local (DImode
, UNITS_PER_WORD
, BITS_PER_WORD
);
4897 m
= validize_mem (m
);
4898 emit_move_insn (m
, pic_offset_table_rtx
);
4903 /* We used to simply emit the sequence after entry_of_function.
4904 However this breaks the CFG if the first instruction in the
4905 first block is not the NOTE_INSN_BASIC_BLOCK, for example a
4906 label. Emit the sequence properly on the edge. We are only
4907 invoked from dw2_build_landing_pads and finish_eh_generation
4908 will call commit_edge_insertions thanks to a kludge. */
4909 insert_insn_on_edge (seq
, single_succ_edge (ENTRY_BLOCK_PTR
));
4911 cfun
->machine
->gp_save_rtx
= m
;
4918 alpha_ra_ever_killed (void)
4922 if (!has_hard_reg_initial_val (Pmode
, REG_RA
))
4923 return (int)df_regs_ever_live_p (REG_RA
);
4925 push_topmost_sequence ();
4927 pop_topmost_sequence ();
4929 return reg_set_between_p (gen_rtx_REG (Pmode
, REG_RA
), top
, NULL_RTX
);
4933 /* Return the trap mode suffix applicable to the current
4934 instruction, or NULL. */
4937 get_trap_mode_suffix (void)
4939 enum attr_trap_suffix s
= get_attr_trap_suffix (current_output_insn
);
4943 case TRAP_SUFFIX_NONE
:
4946 case TRAP_SUFFIX_SU
:
4947 if (alpha_fptm
>= ALPHA_FPTM_SU
)
4951 case TRAP_SUFFIX_SUI
:
4952 if (alpha_fptm
>= ALPHA_FPTM_SUI
)
4956 case TRAP_SUFFIX_V_SV
:
4964 case ALPHA_FPTM_SUI
:
4970 case TRAP_SUFFIX_V_SV_SVI
:
4979 case ALPHA_FPTM_SUI
:
4986 case TRAP_SUFFIX_U_SU_SUI
:
4995 case ALPHA_FPTM_SUI
:
5008 /* Return the rounding mode suffix applicable to the current
5009 instruction, or NULL. */
5012 get_round_mode_suffix (void)
5014 enum attr_round_suffix s
= get_attr_round_suffix (current_output_insn
);
5018 case ROUND_SUFFIX_NONE
:
5020 case ROUND_SUFFIX_NORMAL
:
5023 case ALPHA_FPRM_NORM
:
5025 case ALPHA_FPRM_MINF
:
5027 case ALPHA_FPRM_CHOP
:
5029 case ALPHA_FPRM_DYN
:
5036 case ROUND_SUFFIX_C
:
5045 /* Locate some local-dynamic symbol still in use by this function
5046 so that we can print its name in some movdi_er_tlsldm pattern. */
5049 get_some_local_dynamic_name_1 (rtx
*px
, void *data ATTRIBUTE_UNUSED
)
5053 if (GET_CODE (x
) == SYMBOL_REF
5054 && SYMBOL_REF_TLS_MODEL (x
) == TLS_MODEL_LOCAL_DYNAMIC
)
5056 cfun
->machine
->some_ld_name
= XSTR (x
, 0);
5064 get_some_local_dynamic_name (void)
5068 if (cfun
->machine
->some_ld_name
)
5069 return cfun
->machine
->some_ld_name
;
5071 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
5073 && for_each_rtx (&PATTERN (insn
), get_some_local_dynamic_name_1
, 0))
5074 return cfun
->machine
->some_ld_name
;
5079 /* Print an operand. Recognize special options, documented below. */
5082 print_operand (FILE *file
, rtx x
, int code
)
5089 /* Print the assembler name of the current function. */
5090 assemble_name (file
, alpha_fnname
);
5094 assemble_name (file
, get_some_local_dynamic_name ());
5099 const char *trap
= get_trap_mode_suffix ();
5100 const char *round
= get_round_mode_suffix ();
5103 fprintf (file
, (TARGET_AS_SLASH_BEFORE_SUFFIX
? "/%s%s" : "%s%s"),
5104 (trap
? trap
: ""), (round
? round
: ""));
5109 /* Generates single precision instruction suffix. */
5110 fputc ((TARGET_FLOAT_VAX
? 'f' : 's'), file
);
5114 /* Generates double precision instruction suffix. */
5115 fputc ((TARGET_FLOAT_VAX
? 'g' : 't'), file
);
5119 if (alpha_this_literal_sequence_number
== 0)
5120 alpha_this_literal_sequence_number
= alpha_next_sequence_number
++;
5121 fprintf (file
, "%d", alpha_this_literal_sequence_number
);
5125 if (alpha_this_gpdisp_sequence_number
== 0)
5126 alpha_this_gpdisp_sequence_number
= alpha_next_sequence_number
++;
5127 fprintf (file
, "%d", alpha_this_gpdisp_sequence_number
);
5131 if (GET_CODE (x
) == HIGH
)
5132 output_addr_const (file
, XEXP (x
, 0));
5134 output_operand_lossage ("invalid %%H value");
5141 if (GET_CODE (x
) == UNSPEC
&& XINT (x
, 1) == UNSPEC_TLSGD_CALL
)
5143 x
= XVECEXP (x
, 0, 0);
5144 lituse
= "lituse_tlsgd";
5146 else if (GET_CODE (x
) == UNSPEC
&& XINT (x
, 1) == UNSPEC_TLSLDM_CALL
)
5148 x
= XVECEXP (x
, 0, 0);
5149 lituse
= "lituse_tlsldm";
5151 else if (CONST_INT_P (x
))
5152 lituse
= "lituse_jsr";
5155 output_operand_lossage ("invalid %%J value");
5159 if (x
!= const0_rtx
)
5160 fprintf (file
, "\t\t!%s!%d", lituse
, (int) INTVAL (x
));
5168 #ifdef HAVE_AS_JSRDIRECT_RELOCS
5169 lituse
= "lituse_jsrdirect";
5171 lituse
= "lituse_jsr";
5174 gcc_assert (INTVAL (x
) != 0);
5175 fprintf (file
, "\t\t!%s!%d", lituse
, (int) INTVAL (x
));
5179 /* If this operand is the constant zero, write it as "$31". */
5181 fprintf (file
, "%s", reg_names
[REGNO (x
)]);
5182 else if (x
== CONST0_RTX (GET_MODE (x
)))
5183 fprintf (file
, "$31");
5185 output_operand_lossage ("invalid %%r value");
5189 /* Similar, but for floating-point. */
5191 fprintf (file
, "%s", reg_names
[REGNO (x
)]);
5192 else if (x
== CONST0_RTX (GET_MODE (x
)))
5193 fprintf (file
, "$f31");
5195 output_operand_lossage ("invalid %%R value");
5199 /* Write the 1's complement of a constant. */
5200 if (!CONST_INT_P (x
))
5201 output_operand_lossage ("invalid %%N value");
5203 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, ~ INTVAL (x
));
5207 /* Write 1 << C, for a constant C. */
5208 if (!CONST_INT_P (x
))
5209 output_operand_lossage ("invalid %%P value");
5211 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, (HOST_WIDE_INT
) 1 << INTVAL (x
));
5215 /* Write the high-order 16 bits of a constant, sign-extended. */
5216 if (!CONST_INT_P (x
))
5217 output_operand_lossage ("invalid %%h value");
5219 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, INTVAL (x
) >> 16);
5223 /* Write the low-order 16 bits of a constant, sign-extended. */
5224 if (!CONST_INT_P (x
))
5225 output_operand_lossage ("invalid %%L value");
5227 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
,
5228 (INTVAL (x
) & 0xffff) - 2 * (INTVAL (x
) & 0x8000));
5232 /* Write mask for ZAP insn. */
5233 if (GET_CODE (x
) == CONST_DOUBLE
)
5235 HOST_WIDE_INT mask
= 0;
5236 HOST_WIDE_INT value
;
5238 value
= CONST_DOUBLE_LOW (x
);
5239 for (i
= 0; i
< HOST_BITS_PER_WIDE_INT
/ HOST_BITS_PER_CHAR
;
5244 value
= CONST_DOUBLE_HIGH (x
);
5245 for (i
= 0; i
< HOST_BITS_PER_WIDE_INT
/ HOST_BITS_PER_CHAR
;
5248 mask
|= (1 << (i
+ sizeof (int)));
5250 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, mask
& 0xff);
5253 else if (CONST_INT_P (x
))
5255 HOST_WIDE_INT mask
= 0, value
= INTVAL (x
);
5257 for (i
= 0; i
< 8; i
++, value
>>= 8)
5261 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, mask
);
5264 output_operand_lossage ("invalid %%m value");
5268 /* 'b', 'w', 'l', or 'q' as the value of the constant. */
5269 if (!CONST_INT_P (x
)
5270 || (INTVAL (x
) != 8 && INTVAL (x
) != 16
5271 && INTVAL (x
) != 32 && INTVAL (x
) != 64))
5272 output_operand_lossage ("invalid %%M value");
5274 fprintf (file
, "%s",
5275 (INTVAL (x
) == 8 ? "b"
5276 : INTVAL (x
) == 16 ? "w"
5277 : INTVAL (x
) == 32 ? "l"
5282 /* Similar, except do it from the mask. */
5283 if (CONST_INT_P (x
))
5285 HOST_WIDE_INT value
= INTVAL (x
);
5292 if (value
== 0xffff)
5297 if (value
== 0xffffffff)
5308 else if (HOST_BITS_PER_WIDE_INT
== 32
5309 && GET_CODE (x
) == CONST_DOUBLE
5310 && CONST_DOUBLE_LOW (x
) == 0xffffffff
5311 && CONST_DOUBLE_HIGH (x
) == 0)
5316 output_operand_lossage ("invalid %%U value");
5320 /* Write the constant value divided by 8 for little-endian mode or
5321 (56 - value) / 8 for big-endian mode. */
5323 if (!CONST_INT_P (x
)
5324 || (unsigned HOST_WIDE_INT
) INTVAL (x
) >= (WORDS_BIG_ENDIAN
5327 || (INTVAL (x
) & 7) != 0)
5328 output_operand_lossage ("invalid %%s value");
5330 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
,
5332 ? (56 - INTVAL (x
)) / 8
5337 /* Same, except compute (64 - c) / 8 */
5339 if (!CONST_INT_P (x
)
5340 && (unsigned HOST_WIDE_INT
) INTVAL (x
) >= 64
5341 && (INTVAL (x
) & 7) != 8)
5342 output_operand_lossage ("invalid %%s value");
5344 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
, (64 - INTVAL (x
)) / 8);
5349 /* On Unicos/Mk systems: use a DEX expression if the symbol
5350 clashes with a register name. */
5351 int dex
= unicosmk_need_dex (x
);
5353 fprintf (file
, "DEX(%d)", dex
);
5355 output_addr_const (file
, x
);
5359 case 'C': case 'D': case 'c': case 'd':
5360 /* Write out comparison name. */
5362 enum rtx_code c
= GET_CODE (x
);
5364 if (!COMPARISON_P (x
))
5365 output_operand_lossage ("invalid %%C value");
5367 else if (code
== 'D')
5368 c
= reverse_condition (c
);
5369 else if (code
== 'c')
5370 c
= swap_condition (c
);
5371 else if (code
== 'd')
5372 c
= swap_condition (reverse_condition (c
));
5375 fprintf (file
, "ule");
5377 fprintf (file
, "ult");
5378 else if (c
== UNORDERED
)
5379 fprintf (file
, "un");
5381 fprintf (file
, "%s", GET_RTX_NAME (c
));
5386 /* Write the divide or modulus operator. */
5387 switch (GET_CODE (x
))
5390 fprintf (file
, "div%s", GET_MODE (x
) == SImode
? "l" : "q");
5393 fprintf (file
, "div%su", GET_MODE (x
) == SImode
? "l" : "q");
5396 fprintf (file
, "rem%s", GET_MODE (x
) == SImode
? "l" : "q");
5399 fprintf (file
, "rem%su", GET_MODE (x
) == SImode
? "l" : "q");
5402 output_operand_lossage ("invalid %%E value");
5408 /* Write "_u" for unaligned access. */
5409 if (MEM_P (x
) && GET_CODE (XEXP (x
, 0)) == AND
)
5410 fprintf (file
, "_u");
5415 fprintf (file
, "%s", reg_names
[REGNO (x
)]);
5417 output_address (XEXP (x
, 0));
5418 else if (GET_CODE (x
) == CONST
&& GET_CODE (XEXP (x
, 0)) == UNSPEC
)
5420 switch (XINT (XEXP (x
, 0), 1))
5424 output_addr_const (file
, XVECEXP (XEXP (x
, 0), 0, 0));
5427 output_operand_lossage ("unknown relocation unspec");
5432 output_addr_const (file
, x
);
5436 output_operand_lossage ("invalid %%xn code");
5441 print_operand_address (FILE *file
, rtx addr
)
5444 HOST_WIDE_INT offset
= 0;
5446 if (GET_CODE (addr
) == AND
)
5447 addr
= XEXP (addr
, 0);
5449 if (GET_CODE (addr
) == PLUS
5450 && CONST_INT_P (XEXP (addr
, 1)))
5452 offset
= INTVAL (XEXP (addr
, 1));
5453 addr
= XEXP (addr
, 0);
5456 if (GET_CODE (addr
) == LO_SUM
)
5458 const char *reloc16
, *reloclo
;
5459 rtx op1
= XEXP (addr
, 1);
5461 if (GET_CODE (op1
) == CONST
&& GET_CODE (XEXP (op1
, 0)) == UNSPEC
)
5463 op1
= XEXP (op1
, 0);
5464 switch (XINT (op1
, 1))
5468 reloclo
= (alpha_tls_size
== 16 ? "dtprel" : "dtprello");
5472 reloclo
= (alpha_tls_size
== 16 ? "tprel" : "tprello");
5475 output_operand_lossage ("unknown relocation unspec");
5479 output_addr_const (file
, XVECEXP (op1
, 0, 0));
5484 reloclo
= "gprellow";
5485 output_addr_const (file
, op1
);
5489 fprintf (file
, "+" HOST_WIDE_INT_PRINT_DEC
, offset
);
5491 addr
= XEXP (addr
, 0);
5492 switch (GET_CODE (addr
))
5495 basereg
= REGNO (addr
);
5499 basereg
= subreg_regno (addr
);
5506 fprintf (file
, "($%d)\t\t!%s", basereg
,
5507 (basereg
== 29 ? reloc16
: reloclo
));
5511 switch (GET_CODE (addr
))
5514 basereg
= REGNO (addr
);
5518 basereg
= subreg_regno (addr
);
5522 offset
= INTVAL (addr
);
5525 #if TARGET_ABI_OPEN_VMS
5527 fprintf (file
, "%s", XSTR (addr
, 0));
5531 gcc_assert (GET_CODE (XEXP (addr
, 0)) == PLUS
5532 && GET_CODE (XEXP (XEXP (addr
, 0), 0)) == SYMBOL_REF
);
5533 fprintf (file
, "%s+" HOST_WIDE_INT_PRINT_DEC
,
5534 XSTR (XEXP (XEXP (addr
, 0), 0), 0),
5535 INTVAL (XEXP (XEXP (addr
, 0), 1)));
5543 fprintf (file
, HOST_WIDE_INT_PRINT_DEC
"($%d)", offset
, basereg
);
5546 /* Emit RTL insns to initialize the variable parts of a trampoline at
5547 M_TRAMP. FNDECL is target function's decl. CHAIN_VALUE is an rtx
5548 for the static chain value for the function. */
5551 alpha_trampoline_init (rtx m_tramp
, tree fndecl
, rtx chain_value
)
5553 rtx fnaddr
, mem
, word1
, word2
;
5555 fnaddr
= XEXP (DECL_RTL (fndecl
), 0);
5557 #ifdef POINTERS_EXTEND_UNSIGNED
5558 fnaddr
= convert_memory_address (Pmode
, fnaddr
);
5559 chain_value
= convert_memory_address (Pmode
, chain_value
);
5562 if (TARGET_ABI_OPEN_VMS
)
5567 /* Construct the name of the trampoline entry point. */
5568 fnname
= XSTR (fnaddr
, 0);
5569 trname
= (char *) alloca (strlen (fnname
) + 5);
5570 strcpy (trname
, fnname
);
5571 strcat (trname
, "..tr");
5572 fnname
= ggc_alloc_string (trname
, strlen (trname
) + 1);
5573 word2
= gen_rtx_SYMBOL_REF (Pmode
, fnname
);
5575 /* Trampoline (or "bounded") procedure descriptor is constructed from
5576 the function's procedure descriptor with certain fields zeroed IAW
5577 the VMS calling standard. This is stored in the first quadword. */
5578 word1
= force_reg (DImode
, gen_const_mem (DImode
, fnaddr
));
5579 word1
= expand_and (DImode
, word1
, GEN_INT (0xffff0fff0000fff0), NULL
);
5583 /* These 4 instructions are:
5588 We don't bother setting the HINT field of the jump; the nop
5589 is merely there for padding. */
5590 word1
= GEN_INT (0xa77b0010a43b0018);
5591 word2
= GEN_INT (0x47ff041f6bfb0000);
5594 /* Store the first two words, as computed above. */
5595 mem
= adjust_address (m_tramp
, DImode
, 0);
5596 emit_move_insn (mem
, word1
);
5597 mem
= adjust_address (m_tramp
, DImode
, 8);
5598 emit_move_insn (mem
, word2
);
5600 /* Store function address and static chain value. */
5601 mem
= adjust_address (m_tramp
, Pmode
, 16);
5602 emit_move_insn (mem
, fnaddr
);
5603 mem
= adjust_address (m_tramp
, Pmode
, 24);
5604 emit_move_insn (mem
, chain_value
);
5606 if (!TARGET_ABI_OPEN_VMS
)
5608 emit_insn (gen_imb ());
5609 #ifdef ENABLE_EXECUTE_STACK
5610 emit_library_call (init_one_libfunc ("__enable_execute_stack"),
5611 LCT_NORMAL
, VOIDmode
, 1, XEXP (m_tramp
, 0), Pmode
);
5616 /* Determine where to put an argument to a function.
5617 Value is zero to push the argument on the stack,
5618 or a hard register in which to store the argument.
5620 MODE is the argument's machine mode.
5621 TYPE is the data type of the argument (as a tree).
5622 This is null for libcalls where that information may
5624 CUM is a variable of type CUMULATIVE_ARGS which gives info about
5625 the preceding args and about the function being called.
5626 NAMED is nonzero if this argument is a named parameter
5627 (otherwise it is an extra parameter matching an ellipsis).
5629 On Alpha the first 6 words of args are normally in registers
5630 and the rest are pushed. */
5633 function_arg (CUMULATIVE_ARGS cum
, enum machine_mode mode
, tree type
,
5634 int named ATTRIBUTE_UNUSED
)
5639 /* Don't get confused and pass small structures in FP registers. */
5640 if (type
&& AGGREGATE_TYPE_P (type
))
5644 #ifdef ENABLE_CHECKING
5645 /* With alpha_split_complex_arg, we shouldn't see any raw complex
5647 gcc_assert (!COMPLEX_MODE_P (mode
));
5650 /* Set up defaults for FP operands passed in FP registers, and
5651 integral operands passed in integer registers. */
5652 if (TARGET_FPREGS
&& GET_MODE_CLASS (mode
) == MODE_FLOAT
)
5658 /* ??? Irritatingly, the definition of CUMULATIVE_ARGS is different for
5659 the three platforms, so we can't avoid conditional compilation. */
5660 #if TARGET_ABI_OPEN_VMS
5662 if (mode
== VOIDmode
)
5663 return alpha_arg_info_reg_val (cum
);
5665 num_args
= cum
.num_args
;
5667 || targetm
.calls
.must_pass_in_stack (mode
, type
))
5670 #elif TARGET_ABI_UNICOSMK
5674 /* If this is the last argument, generate the call info word (CIW). */
5675 /* ??? We don't include the caller's line number in the CIW because
5676 I don't know how to determine it if debug infos are turned off. */
5677 if (mode
== VOIDmode
)
5686 for (i
= 0; i
< cum
.num_reg_words
&& i
< 5; i
++)
5687 if (cum
.reg_args_type
[i
])
5688 lo
|= (1 << (7 - i
));
5690 if (cum
.num_reg_words
== 6 && cum
.reg_args_type
[5])
5693 lo
|= cum
.num_reg_words
;
5695 #if HOST_BITS_PER_WIDE_INT == 32
5696 hi
= (cum
.num_args
<< 20) | cum
.num_arg_words
;
5698 lo
= lo
| ((HOST_WIDE_INT
) cum
.num_args
<< 52)
5699 | ((HOST_WIDE_INT
) cum
.num_arg_words
<< 32);
5702 ciw
= immed_double_const (lo
, hi
, DImode
);
5704 return gen_rtx_UNSPEC (DImode
, gen_rtvec (1, ciw
),
5705 UNSPEC_UMK_LOAD_CIW
);
5708 size
= ALPHA_ARG_SIZE (mode
, type
, named
);
5709 num_args
= cum
.num_reg_words
;
5711 || cum
.num_reg_words
+ size
> 6
5712 || targetm
.calls
.must_pass_in_stack (mode
, type
))
5714 else if (type
&& TYPE_MODE (type
) == BLKmode
)
5718 reg1
= gen_rtx_REG (DImode
, num_args
+ 16);
5719 reg1
= gen_rtx_EXPR_LIST (DImode
, reg1
, const0_rtx
);
5721 /* The argument fits in two registers. Note that we still need to
5722 reserve a register for empty structures. */
5726 return gen_rtx_PARALLEL (mode
, gen_rtvec (1, reg1
));
5729 reg2
= gen_rtx_REG (DImode
, num_args
+ 17);
5730 reg2
= gen_rtx_EXPR_LIST (DImode
, reg2
, GEN_INT (8));
5731 return gen_rtx_PARALLEL (mode
, gen_rtvec (2, reg1
, reg2
));
5735 #elif TARGET_ABI_OSF
5741 /* VOID is passed as a special flag for "last argument". */
5742 if (type
== void_type_node
)
5744 else if (targetm
.calls
.must_pass_in_stack (mode
, type
))
5748 #error Unhandled ABI
5751 return gen_rtx_REG (mode
, num_args
+ basereg
);
5755 alpha_arg_partial_bytes (CUMULATIVE_ARGS
*cum ATTRIBUTE_UNUSED
,
5756 enum machine_mode mode ATTRIBUTE_UNUSED
,
5757 tree type ATTRIBUTE_UNUSED
,
5758 bool named ATTRIBUTE_UNUSED
)
5762 #if TARGET_ABI_OPEN_VMS
5763 if (cum
->num_args
< 6
5764 && 6 < cum
->num_args
+ ALPHA_ARG_SIZE (mode
, type
, named
))
5765 words
= 6 - cum
->num_args
;
5766 #elif TARGET_ABI_UNICOSMK
5767 /* Never any split arguments. */
5768 #elif TARGET_ABI_OSF
5769 if (*cum
< 6 && 6 < *cum
+ ALPHA_ARG_SIZE (mode
, type
, named
))
5772 #error Unhandled ABI
5775 return words
* UNITS_PER_WORD
;
5779 /* Return true if TYPE must be returned in memory, instead of in registers. */
5782 alpha_return_in_memory (const_tree type
, const_tree fndecl ATTRIBUTE_UNUSED
)
5784 enum machine_mode mode
= VOIDmode
;
5789 mode
= TYPE_MODE (type
);
5791 /* All aggregates are returned in memory, except on OpenVMS where
5792 records that fit 64 bits should be returned by immediate value
5793 as required by section 3.8.7.1 of the OpenVMS Calling Standard. */
5794 if (TARGET_ABI_OPEN_VMS
5795 && TREE_CODE (type
) != ARRAY_TYPE
5796 && (unsigned HOST_WIDE_INT
) int_size_in_bytes(type
) <= 8)
5799 if (AGGREGATE_TYPE_P (type
))
5803 size
= GET_MODE_SIZE (mode
);
5804 switch (GET_MODE_CLASS (mode
))
5806 case MODE_VECTOR_FLOAT
:
5807 /* Pass all float vectors in memory, like an aggregate. */
5810 case MODE_COMPLEX_FLOAT
:
5811 /* We judge complex floats on the size of their element,
5812 not the size of the whole type. */
5813 size
= GET_MODE_UNIT_SIZE (mode
);
5818 case MODE_COMPLEX_INT
:
5819 case MODE_VECTOR_INT
:
5823 /* ??? We get called on all sorts of random stuff from
5824 aggregate_value_p. We must return something, but it's not
5825 clear what's safe to return. Pretend it's a struct I
5830 /* Otherwise types must fit in one register. */
5831 return size
> UNITS_PER_WORD
;
5834 /* Return true if TYPE should be passed by invisible reference. */
5837 alpha_pass_by_reference (CUMULATIVE_ARGS
*ca ATTRIBUTE_UNUSED
,
5838 enum machine_mode mode
,
5839 const_tree type ATTRIBUTE_UNUSED
,
5840 bool named ATTRIBUTE_UNUSED
)
5842 return mode
== TFmode
|| mode
== TCmode
;
5845 /* Define how to find the value returned by a function. VALTYPE is the
5846 data type of the value (as a tree). If the precise function being
5847 called is known, FUNC is its FUNCTION_DECL; otherwise, FUNC is 0.
5848 MODE is set instead of VALTYPE for libcalls.
5850 On Alpha the value is found in $0 for integer functions and
5851 $f0 for floating-point functions. */
5854 function_value (const_tree valtype
, const_tree func ATTRIBUTE_UNUSED
,
5855 enum machine_mode mode
)
5857 unsigned int regnum
, dummy ATTRIBUTE_UNUSED
;
5858 enum mode_class mclass
;
5860 gcc_assert (!valtype
|| !alpha_return_in_memory (valtype
, func
));
5863 mode
= TYPE_MODE (valtype
);
5865 mclass
= GET_MODE_CLASS (mode
);
5869 /* Do the same thing as PROMOTE_MODE except for libcalls on VMS,
5870 where we have them returning both SImode and DImode. */
5871 if (!(TARGET_ABI_OPEN_VMS
&& valtype
&& AGGREGATE_TYPE_P (valtype
)))
5872 PROMOTE_MODE (mode
, dummy
, valtype
);
5875 case MODE_COMPLEX_INT
:
5876 case MODE_VECTOR_INT
:
5884 case MODE_COMPLEX_FLOAT
:
5886 enum machine_mode cmode
= GET_MODE_INNER (mode
);
5888 return gen_rtx_PARALLEL
5891 gen_rtx_EXPR_LIST (VOIDmode
, gen_rtx_REG (cmode
, 32),
5893 gen_rtx_EXPR_LIST (VOIDmode
, gen_rtx_REG (cmode
, 33),
5894 GEN_INT (GET_MODE_SIZE (cmode
)))));
5898 /* We should only reach here for BLKmode on VMS. */
5899 gcc_assert (TARGET_ABI_OPEN_VMS
&& mode
== BLKmode
);
5907 return gen_rtx_REG (mode
, regnum
);
5910 /* TCmode complex values are passed by invisible reference. We
5911 should not split these values. */
5914 alpha_split_complex_arg (const_tree type
)
5916 return TYPE_MODE (type
) != TCmode
;
5920 alpha_build_builtin_va_list (void)
5922 tree base
, ofs
, space
, record
, type_decl
;
5924 if (TARGET_ABI_OPEN_VMS
|| TARGET_ABI_UNICOSMK
)
5925 return ptr_type_node
;
5927 record
= (*lang_hooks
.types
.make_type
) (RECORD_TYPE
);
5928 type_decl
= build_decl (BUILTINS_LOCATION
,
5929 TYPE_DECL
, get_identifier ("__va_list_tag"), record
);
5930 TREE_CHAIN (record
) = type_decl
;
5931 TYPE_NAME (record
) = type_decl
;
5933 /* C++? SET_IS_AGGR_TYPE (record, 1); */
5935 /* Dummy field to prevent alignment warnings. */
5936 space
= build_decl (BUILTINS_LOCATION
,
5937 FIELD_DECL
, NULL_TREE
, integer_type_node
);
5938 DECL_FIELD_CONTEXT (space
) = record
;
5939 DECL_ARTIFICIAL (space
) = 1;
5940 DECL_IGNORED_P (space
) = 1;
5942 ofs
= build_decl (BUILTINS_LOCATION
,
5943 FIELD_DECL
, get_identifier ("__offset"),
5945 DECL_FIELD_CONTEXT (ofs
) = record
;
5946 TREE_CHAIN (ofs
) = space
;
5948 base
= build_decl (BUILTINS_LOCATION
,
5949 FIELD_DECL
, get_identifier ("__base"),
5951 DECL_FIELD_CONTEXT (base
) = record
;
5952 TREE_CHAIN (base
) = ofs
;
5954 TYPE_FIELDS (record
) = base
;
5955 layout_type (record
);
5957 va_list_gpr_counter_field
= ofs
;
5962 /* Helper function for alpha_stdarg_optimize_hook. Skip over casts
5963 and constant additions. */
5966 va_list_skip_additions (tree lhs
)
5972 enum tree_code code
;
5974 stmt
= SSA_NAME_DEF_STMT (lhs
);
5976 if (gimple_code (stmt
) == GIMPLE_PHI
)
5979 if (!is_gimple_assign (stmt
)
5980 || gimple_assign_lhs (stmt
) != lhs
)
5983 if (TREE_CODE (gimple_assign_rhs1 (stmt
)) != SSA_NAME
)
5985 code
= gimple_assign_rhs_code (stmt
);
5986 if (!CONVERT_EXPR_CODE_P (code
)
5987 && ((code
!= PLUS_EXPR
&& code
!= POINTER_PLUS_EXPR
)
5988 || TREE_CODE (gimple_assign_rhs2 (stmt
)) != INTEGER_CST
5989 || !host_integerp (gimple_assign_rhs2 (stmt
), 1)))
5992 lhs
= gimple_assign_rhs1 (stmt
);
5996 /* Check if LHS = RHS statement is
5997 LHS = *(ap.__base + ap.__offset + cst)
6000 + ((ap.__offset + cst <= 47)
6001 ? ap.__offset + cst - 48 : ap.__offset + cst) + cst2).
6002 If the former, indicate that GPR registers are needed,
6003 if the latter, indicate that FPR registers are needed.
6005 Also look for LHS = (*ptr).field, where ptr is one of the forms
6008 On alpha, cfun->va_list_gpr_size is used as size of the needed
6009 regs and cfun->va_list_fpr_size is a bitmask, bit 0 set if GPR
6010 registers are needed and bit 1 set if FPR registers are needed.
6011 Return true if va_list references should not be scanned for the
6012 current statement. */
6015 alpha_stdarg_optimize_hook (struct stdarg_info
*si
, const_gimple stmt
)
6017 tree base
, offset
, rhs
;
6021 if (get_gimple_rhs_class (gimple_assign_rhs_code (stmt
))
6022 != GIMPLE_SINGLE_RHS
)
6025 rhs
= gimple_assign_rhs1 (stmt
);
6026 while (handled_component_p (rhs
))
6027 rhs
= TREE_OPERAND (rhs
, 0);
6028 if (TREE_CODE (rhs
) != INDIRECT_REF
6029 || TREE_CODE (TREE_OPERAND (rhs
, 0)) != SSA_NAME
)
6032 stmt
= va_list_skip_additions (TREE_OPERAND (rhs
, 0));
6034 || !is_gimple_assign (stmt
)
6035 || gimple_assign_rhs_code (stmt
) != POINTER_PLUS_EXPR
)
6038 base
= gimple_assign_rhs1 (stmt
);
6039 if (TREE_CODE (base
) == SSA_NAME
)
6041 base_stmt
= va_list_skip_additions (base
);
6043 && is_gimple_assign (base_stmt
)
6044 && gimple_assign_rhs_code (base_stmt
) == COMPONENT_REF
)
6045 base
= gimple_assign_rhs1 (base_stmt
);
6048 if (TREE_CODE (base
) != COMPONENT_REF
6049 || TREE_OPERAND (base
, 1) != TYPE_FIELDS (va_list_type_node
))
6051 base
= gimple_assign_rhs2 (stmt
);
6052 if (TREE_CODE (base
) == SSA_NAME
)
6054 base_stmt
= va_list_skip_additions (base
);
6056 && is_gimple_assign (base_stmt
)
6057 && gimple_assign_rhs_code (base_stmt
) == COMPONENT_REF
)
6058 base
= gimple_assign_rhs1 (base_stmt
);
6061 if (TREE_CODE (base
) != COMPONENT_REF
6062 || TREE_OPERAND (base
, 1) != TYPE_FIELDS (va_list_type_node
))
6068 base
= get_base_address (base
);
6069 if (TREE_CODE (base
) != VAR_DECL
6070 || !bitmap_bit_p (si
->va_list_vars
, DECL_UID (base
)))
6073 offset
= gimple_op (stmt
, 1 + offset_arg
);
6074 if (TREE_CODE (offset
) == SSA_NAME
)
6076 gimple offset_stmt
= va_list_skip_additions (offset
);
6079 && gimple_code (offset_stmt
) == GIMPLE_PHI
)
6082 gimple arg1_stmt
, arg2_stmt
;
6084 enum tree_code code1
, code2
;
6086 if (gimple_phi_num_args (offset_stmt
) != 2)
6090 = va_list_skip_additions (gimple_phi_arg_def (offset_stmt
, 0));
6092 = va_list_skip_additions (gimple_phi_arg_def (offset_stmt
, 1));
6093 if (arg1_stmt
== NULL
6094 || !is_gimple_assign (arg1_stmt
)
6095 || arg2_stmt
== NULL
6096 || !is_gimple_assign (arg2_stmt
))
6099 code1
= gimple_assign_rhs_code (arg1_stmt
);
6100 code2
= gimple_assign_rhs_code (arg2_stmt
);
6101 if (code1
== COMPONENT_REF
6102 && (code2
== MINUS_EXPR
|| code2
== PLUS_EXPR
))
6104 else if (code2
== COMPONENT_REF
6105 && (code1
== MINUS_EXPR
|| code1
== PLUS_EXPR
))
6107 gimple tem
= arg1_stmt
;
6109 arg1_stmt
= arg2_stmt
;
6115 if (!host_integerp (gimple_assign_rhs2 (arg2_stmt
), 0))
6118 sub
= tree_low_cst (gimple_assign_rhs2 (arg2_stmt
), 0);
6119 if (code2
== MINUS_EXPR
)
6121 if (sub
< -48 || sub
> -32)
6124 arg1
= gimple_assign_rhs1 (arg1_stmt
);
6125 arg2
= gimple_assign_rhs1 (arg2_stmt
);
6126 if (TREE_CODE (arg2
) == SSA_NAME
)
6128 arg2_stmt
= va_list_skip_additions (arg2
);
6129 if (arg2_stmt
== NULL
6130 || !is_gimple_assign (arg2_stmt
)
6131 || gimple_assign_rhs_code (arg2_stmt
) != COMPONENT_REF
)
6133 arg2
= gimple_assign_rhs1 (arg2_stmt
);
6138 if (TREE_CODE (arg1
) != COMPONENT_REF
6139 || TREE_OPERAND (arg1
, 1) != va_list_gpr_counter_field
6140 || get_base_address (arg1
) != base
)
6143 /* Need floating point regs. */
6144 cfun
->va_list_fpr_size
|= 2;
6148 && is_gimple_assign (offset_stmt
)
6149 && gimple_assign_rhs_code (offset_stmt
) == COMPONENT_REF
)
6150 offset
= gimple_assign_rhs1 (offset_stmt
);
6152 if (TREE_CODE (offset
) != COMPONENT_REF
6153 || TREE_OPERAND (offset
, 1) != va_list_gpr_counter_field
6154 || get_base_address (offset
) != base
)
6157 /* Need general regs. */
6158 cfun
->va_list_fpr_size
|= 1;
6162 si
->va_list_escapes
= true;
6167 /* Perform any needed actions needed for a function that is receiving a
6168 variable number of arguments. */
6171 alpha_setup_incoming_varargs (CUMULATIVE_ARGS
*pcum
, enum machine_mode mode
,
6172 tree type
, int *pretend_size
, int no_rtl
)
6174 CUMULATIVE_ARGS cum
= *pcum
;
6176 /* Skip the current argument. */
6177 FUNCTION_ARG_ADVANCE (cum
, mode
, type
, 1);
6179 #if TARGET_ABI_UNICOSMK
6180 /* On Unicos/Mk, the standard subroutine __T3E_MISMATCH stores all register
6181 arguments on the stack. Unfortunately, it doesn't always store the first
6182 one (i.e. the one that arrives in $16 or $f16). This is not a problem
6183 with stdargs as we always have at least one named argument there. */
6184 if (cum
.num_reg_words
< 6)
6188 emit_insn (gen_umk_mismatch_args (GEN_INT (cum
.num_reg_words
)));
6189 emit_insn (gen_arg_home_umk ());
6193 #elif TARGET_ABI_OPEN_VMS
6194 /* For VMS, we allocate space for all 6 arg registers plus a count.
6196 However, if NO registers need to be saved, don't allocate any space.
6197 This is not only because we won't need the space, but because AP
6198 includes the current_pretend_args_size and we don't want to mess up
6199 any ap-relative addresses already made. */
6200 if (cum
.num_args
< 6)
6204 emit_move_insn (gen_rtx_REG (DImode
, 1), virtual_incoming_args_rtx
);
6205 emit_insn (gen_arg_home ());
6207 *pretend_size
= 7 * UNITS_PER_WORD
;
6210 /* On OSF/1 and friends, we allocate space for all 12 arg registers, but
6211 only push those that are remaining. However, if NO registers need to
6212 be saved, don't allocate any space. This is not only because we won't
6213 need the space, but because AP includes the current_pretend_args_size
6214 and we don't want to mess up any ap-relative addresses already made.
6216 If we are not to use the floating-point registers, save the integer
6217 registers where we would put the floating-point registers. This is
6218 not the most efficient way to implement varargs with just one register
6219 class, but it isn't worth doing anything more efficient in this rare
6227 alias_set_type set
= get_varargs_alias_set ();
6230 count
= cfun
->va_list_gpr_size
/ UNITS_PER_WORD
;
6231 if (count
> 6 - cum
)
6234 /* Detect whether integer registers or floating-point registers
6235 are needed by the detected va_arg statements. See above for
6236 how these values are computed. Note that the "escape" value
6237 is VA_LIST_MAX_FPR_SIZE, which is 255, which has both of
6239 gcc_assert ((VA_LIST_MAX_FPR_SIZE
& 3) == 3);
6241 if (cfun
->va_list_fpr_size
& 1)
6243 tmp
= gen_rtx_MEM (BLKmode
,
6244 plus_constant (virtual_incoming_args_rtx
,
6245 (cum
+ 6) * UNITS_PER_WORD
));
6246 MEM_NOTRAP_P (tmp
) = 1;
6247 set_mem_alias_set (tmp
, set
);
6248 move_block_from_reg (16 + cum
, tmp
, count
);
6251 if (cfun
->va_list_fpr_size
& 2)
6253 tmp
= gen_rtx_MEM (BLKmode
,
6254 plus_constant (virtual_incoming_args_rtx
,
6255 cum
* UNITS_PER_WORD
));
6256 MEM_NOTRAP_P (tmp
) = 1;
6257 set_mem_alias_set (tmp
, set
);
6258 move_block_from_reg (16 + cum
+ TARGET_FPREGS
*32, tmp
, count
);
6261 *pretend_size
= 12 * UNITS_PER_WORD
;
6266 alpha_va_start (tree valist
, rtx nextarg ATTRIBUTE_UNUSED
)
6268 HOST_WIDE_INT offset
;
6269 tree t
, offset_field
, base_field
;
6271 if (TREE_CODE (TREE_TYPE (valist
)) == ERROR_MARK
)
6274 if (TARGET_ABI_UNICOSMK
)
6275 std_expand_builtin_va_start (valist
, nextarg
);
6277 /* For Unix, TARGET_SETUP_INCOMING_VARARGS moves the starting address base
6278 up by 48, storing fp arg registers in the first 48 bytes, and the
6279 integer arg registers in the next 48 bytes. This is only done,
6280 however, if any integer registers need to be stored.
6282 If no integer registers need be stored, then we must subtract 48
6283 in order to account for the integer arg registers which are counted
6284 in argsize above, but which are not actually stored on the stack.
6285 Must further be careful here about structures straddling the last
6286 integer argument register; that futzes with pretend_args_size,
6287 which changes the meaning of AP. */
6290 offset
= TARGET_ABI_OPEN_VMS
? UNITS_PER_WORD
: 6 * UNITS_PER_WORD
;
6292 offset
= -6 * UNITS_PER_WORD
+ crtl
->args
.pretend_args_size
;
6294 if (TARGET_ABI_OPEN_VMS
)
6296 t
= make_tree (ptr_type_node
, virtual_incoming_args_rtx
);
6297 t
= build2 (POINTER_PLUS_EXPR
, ptr_type_node
, t
,
6298 size_int (offset
+ NUM_ARGS
* UNITS_PER_WORD
));
6299 t
= build2 (MODIFY_EXPR
, TREE_TYPE (valist
), valist
, t
);
6300 TREE_SIDE_EFFECTS (t
) = 1;
6301 expand_expr (t
, const0_rtx
, VOIDmode
, EXPAND_NORMAL
);
6305 base_field
= TYPE_FIELDS (TREE_TYPE (valist
));
6306 offset_field
= TREE_CHAIN (base_field
);
6308 base_field
= build3 (COMPONENT_REF
, TREE_TYPE (base_field
),
6309 valist
, base_field
, NULL_TREE
);
6310 offset_field
= build3 (COMPONENT_REF
, TREE_TYPE (offset_field
),
6311 valist
, offset_field
, NULL_TREE
);
6313 t
= make_tree (ptr_type_node
, virtual_incoming_args_rtx
);
6314 t
= build2 (POINTER_PLUS_EXPR
, ptr_type_node
, t
,
6316 t
= build2 (MODIFY_EXPR
, TREE_TYPE (base_field
), base_field
, t
);
6317 TREE_SIDE_EFFECTS (t
) = 1;
6318 expand_expr (t
, const0_rtx
, VOIDmode
, EXPAND_NORMAL
);
6320 t
= build_int_cst (NULL_TREE
, NUM_ARGS
* UNITS_PER_WORD
);
6321 t
= build2 (MODIFY_EXPR
, TREE_TYPE (offset_field
), offset_field
, t
);
6322 TREE_SIDE_EFFECTS (t
) = 1;
6323 expand_expr (t
, const0_rtx
, VOIDmode
, EXPAND_NORMAL
);
6328 alpha_gimplify_va_arg_1 (tree type
, tree base
, tree offset
,
6331 tree type_size
, ptr_type
, addend
, t
, addr
;
6332 gimple_seq internal_post
;
6334 /* If the type could not be passed in registers, skip the block
6335 reserved for the registers. */
6336 if (targetm
.calls
.must_pass_in_stack (TYPE_MODE (type
), type
))
6338 t
= build_int_cst (TREE_TYPE (offset
), 6*8);
6339 gimplify_assign (offset
,
6340 build2 (MAX_EXPR
, TREE_TYPE (offset
), offset
, t
),
6345 ptr_type
= build_pointer_type_for_mode (type
, ptr_mode
, true);
6347 if (TREE_CODE (type
) == COMPLEX_TYPE
)
6349 tree real_part
, imag_part
, real_temp
;
6351 real_part
= alpha_gimplify_va_arg_1 (TREE_TYPE (type
), base
,
6354 /* Copy the value into a new temporary, lest the formal temporary
6355 be reused out from under us. */
6356 real_temp
= get_initialized_tmp_var (real_part
, pre_p
, NULL
);
6358 imag_part
= alpha_gimplify_va_arg_1 (TREE_TYPE (type
), base
,
6361 return build2 (COMPLEX_EXPR
, type
, real_temp
, imag_part
);
6363 else if (TREE_CODE (type
) == REAL_TYPE
)
6365 tree fpaddend
, cond
, fourtyeight
;
6367 fourtyeight
= build_int_cst (TREE_TYPE (addend
), 6*8);
6368 fpaddend
= fold_build2 (MINUS_EXPR
, TREE_TYPE (addend
),
6369 addend
, fourtyeight
);
6370 cond
= fold_build2 (LT_EXPR
, boolean_type_node
, addend
, fourtyeight
);
6371 addend
= fold_build3 (COND_EXPR
, TREE_TYPE (addend
), cond
,
6375 /* Build the final address and force that value into a temporary. */
6376 addr
= build2 (POINTER_PLUS_EXPR
, ptr_type
, fold_convert (ptr_type
, base
),
6377 fold_convert (sizetype
, addend
));
6378 internal_post
= NULL
;
6379 gimplify_expr (&addr
, pre_p
, &internal_post
, is_gimple_val
, fb_rvalue
);
6380 gimple_seq_add_seq (pre_p
, internal_post
);
6382 /* Update the offset field. */
6383 type_size
= TYPE_SIZE_UNIT (TYPE_MAIN_VARIANT (type
));
6384 if (type_size
== NULL
|| TREE_OVERFLOW (type_size
))
6388 t
= size_binop (PLUS_EXPR
, type_size
, size_int (7));
6389 t
= size_binop (TRUNC_DIV_EXPR
, t
, size_int (8));
6390 t
= size_binop (MULT_EXPR
, t
, size_int (8));
6392 t
= fold_convert (TREE_TYPE (offset
), t
);
6393 gimplify_assign (offset
, build2 (PLUS_EXPR
, TREE_TYPE (offset
), offset
, t
),
6396 return build_va_arg_indirect_ref (addr
);
6400 alpha_gimplify_va_arg (tree valist
, tree type
, gimple_seq
*pre_p
,
6403 tree offset_field
, base_field
, offset
, base
, t
, r
;
6406 if (TARGET_ABI_OPEN_VMS
|| TARGET_ABI_UNICOSMK
)
6407 return std_gimplify_va_arg_expr (valist
, type
, pre_p
, post_p
);
6409 base_field
= TYPE_FIELDS (va_list_type_node
);
6410 offset_field
= TREE_CHAIN (base_field
);
6411 base_field
= build3 (COMPONENT_REF
, TREE_TYPE (base_field
),
6412 valist
, base_field
, NULL_TREE
);
6413 offset_field
= build3 (COMPONENT_REF
, TREE_TYPE (offset_field
),
6414 valist
, offset_field
, NULL_TREE
);
6416 /* Pull the fields of the structure out into temporaries. Since we never
6417 modify the base field, we can use a formal temporary. Sign-extend the
6418 offset field so that it's the proper width for pointer arithmetic. */
6419 base
= get_formal_tmp_var (base_field
, pre_p
);
6421 t
= fold_convert (lang_hooks
.types
.type_for_size (64, 0), offset_field
);
6422 offset
= get_initialized_tmp_var (t
, pre_p
, NULL
);
6424 indirect
= pass_by_reference (NULL
, TYPE_MODE (type
), type
, false);
6426 type
= build_pointer_type_for_mode (type
, ptr_mode
, true);
6428 /* Find the value. Note that this will be a stable indirection, or
6429 a composite of stable indirections in the case of complex. */
6430 r
= alpha_gimplify_va_arg_1 (type
, base
, offset
, pre_p
);
6432 /* Stuff the offset temporary back into its field. */
6433 gimplify_assign (unshare_expr (offset_field
),
6434 fold_convert (TREE_TYPE (offset_field
), offset
), pre_p
);
6437 r
= build_va_arg_indirect_ref (r
);
6446 ALPHA_BUILTIN_CMPBGE
,
6447 ALPHA_BUILTIN_EXTBL
,
6448 ALPHA_BUILTIN_EXTWL
,
6449 ALPHA_BUILTIN_EXTLL
,
6450 ALPHA_BUILTIN_EXTQL
,
6451 ALPHA_BUILTIN_EXTWH
,
6452 ALPHA_BUILTIN_EXTLH
,
6453 ALPHA_BUILTIN_EXTQH
,
6454 ALPHA_BUILTIN_INSBL
,
6455 ALPHA_BUILTIN_INSWL
,
6456 ALPHA_BUILTIN_INSLL
,
6457 ALPHA_BUILTIN_INSQL
,
6458 ALPHA_BUILTIN_INSWH
,
6459 ALPHA_BUILTIN_INSLH
,
6460 ALPHA_BUILTIN_INSQH
,
6461 ALPHA_BUILTIN_MSKBL
,
6462 ALPHA_BUILTIN_MSKWL
,
6463 ALPHA_BUILTIN_MSKLL
,
6464 ALPHA_BUILTIN_MSKQL
,
6465 ALPHA_BUILTIN_MSKWH
,
6466 ALPHA_BUILTIN_MSKLH
,
6467 ALPHA_BUILTIN_MSKQH
,
6468 ALPHA_BUILTIN_UMULH
,
6470 ALPHA_BUILTIN_ZAPNOT
,
6471 ALPHA_BUILTIN_AMASK
,
6472 ALPHA_BUILTIN_IMPLVER
,
6474 ALPHA_BUILTIN_THREAD_POINTER
,
6475 ALPHA_BUILTIN_SET_THREAD_POINTER
,
6476 ALPHA_BUILTIN_ESTABLISH_VMS_CONDITION_HANDLER
,
6477 ALPHA_BUILTIN_REVERT_VMS_CONDITION_HANDLER
,
6480 ALPHA_BUILTIN_MINUB8
,
6481 ALPHA_BUILTIN_MINSB8
,
6482 ALPHA_BUILTIN_MINUW4
,
6483 ALPHA_BUILTIN_MINSW4
,
6484 ALPHA_BUILTIN_MAXUB8
,
6485 ALPHA_BUILTIN_MAXSB8
,
6486 ALPHA_BUILTIN_MAXUW4
,
6487 ALPHA_BUILTIN_MAXSW4
,
6491 ALPHA_BUILTIN_UNPKBL
,
6492 ALPHA_BUILTIN_UNPKBW
,
6497 ALPHA_BUILTIN_CTPOP
,
6502 static enum insn_code
const code_for_builtin
[ALPHA_BUILTIN_max
] = {
6503 CODE_FOR_builtin_cmpbge
,
6504 CODE_FOR_builtin_extbl
,
6505 CODE_FOR_builtin_extwl
,
6506 CODE_FOR_builtin_extll
,
6507 CODE_FOR_builtin_extql
,
6508 CODE_FOR_builtin_extwh
,
6509 CODE_FOR_builtin_extlh
,
6510 CODE_FOR_builtin_extqh
,
6511 CODE_FOR_builtin_insbl
,
6512 CODE_FOR_builtin_inswl
,
6513 CODE_FOR_builtin_insll
,
6514 CODE_FOR_builtin_insql
,
6515 CODE_FOR_builtin_inswh
,
6516 CODE_FOR_builtin_inslh
,
6517 CODE_FOR_builtin_insqh
,
6518 CODE_FOR_builtin_mskbl
,
6519 CODE_FOR_builtin_mskwl
,
6520 CODE_FOR_builtin_mskll
,
6521 CODE_FOR_builtin_mskql
,
6522 CODE_FOR_builtin_mskwh
,
6523 CODE_FOR_builtin_msklh
,
6524 CODE_FOR_builtin_mskqh
,
6525 CODE_FOR_umuldi3_highpart
,
6526 CODE_FOR_builtin_zap
,
6527 CODE_FOR_builtin_zapnot
,
6528 CODE_FOR_builtin_amask
,
6529 CODE_FOR_builtin_implver
,
6530 CODE_FOR_builtin_rpcc
,
6533 CODE_FOR_builtin_establish_vms_condition_handler
,
6534 CODE_FOR_builtin_revert_vms_condition_handler
,
6537 CODE_FOR_builtin_minub8
,
6538 CODE_FOR_builtin_minsb8
,
6539 CODE_FOR_builtin_minuw4
,
6540 CODE_FOR_builtin_minsw4
,
6541 CODE_FOR_builtin_maxub8
,
6542 CODE_FOR_builtin_maxsb8
,
6543 CODE_FOR_builtin_maxuw4
,
6544 CODE_FOR_builtin_maxsw4
,
6545 CODE_FOR_builtin_perr
,
6546 CODE_FOR_builtin_pklb
,
6547 CODE_FOR_builtin_pkwb
,
6548 CODE_FOR_builtin_unpkbl
,
6549 CODE_FOR_builtin_unpkbw
,
6554 CODE_FOR_popcountdi2
6557 struct alpha_builtin_def
6560 enum alpha_builtin code
;
6561 unsigned int target_mask
;
6565 static struct alpha_builtin_def
const zero_arg_builtins
[] = {
6566 { "__builtin_alpha_implver", ALPHA_BUILTIN_IMPLVER
, 0, true },
6567 { "__builtin_alpha_rpcc", ALPHA_BUILTIN_RPCC
, 0, false }
6570 static struct alpha_builtin_def
const one_arg_builtins
[] = {
6571 { "__builtin_alpha_amask", ALPHA_BUILTIN_AMASK
, 0, true },
6572 { "__builtin_alpha_pklb", ALPHA_BUILTIN_PKLB
, MASK_MAX
, true },
6573 { "__builtin_alpha_pkwb", ALPHA_BUILTIN_PKWB
, MASK_MAX
, true },
6574 { "__builtin_alpha_unpkbl", ALPHA_BUILTIN_UNPKBL
, MASK_MAX
, true },
6575 { "__builtin_alpha_unpkbw", ALPHA_BUILTIN_UNPKBW
, MASK_MAX
, true },
6576 { "__builtin_alpha_cttz", ALPHA_BUILTIN_CTTZ
, MASK_CIX
, true },
6577 { "__builtin_alpha_ctlz", ALPHA_BUILTIN_CTLZ
, MASK_CIX
, true },
6578 { "__builtin_alpha_ctpop", ALPHA_BUILTIN_CTPOP
, MASK_CIX
, true }
6581 static struct alpha_builtin_def
const two_arg_builtins
[] = {
6582 { "__builtin_alpha_cmpbge", ALPHA_BUILTIN_CMPBGE
, 0, true },
6583 { "__builtin_alpha_extbl", ALPHA_BUILTIN_EXTBL
, 0, true },
6584 { "__builtin_alpha_extwl", ALPHA_BUILTIN_EXTWL
, 0, true },
6585 { "__builtin_alpha_extll", ALPHA_BUILTIN_EXTLL
, 0, true },
6586 { "__builtin_alpha_extql", ALPHA_BUILTIN_EXTQL
, 0, true },
6587 { "__builtin_alpha_extwh", ALPHA_BUILTIN_EXTWH
, 0, true },
6588 { "__builtin_alpha_extlh", ALPHA_BUILTIN_EXTLH
, 0, true },
6589 { "__builtin_alpha_extqh", ALPHA_BUILTIN_EXTQH
, 0, true },
6590 { "__builtin_alpha_insbl", ALPHA_BUILTIN_INSBL
, 0, true },
6591 { "__builtin_alpha_inswl", ALPHA_BUILTIN_INSWL
, 0, true },
6592 { "__builtin_alpha_insll", ALPHA_BUILTIN_INSLL
, 0, true },
6593 { "__builtin_alpha_insql", ALPHA_BUILTIN_INSQL
, 0, true },
6594 { "__builtin_alpha_inswh", ALPHA_BUILTIN_INSWH
, 0, true },
6595 { "__builtin_alpha_inslh", ALPHA_BUILTIN_INSLH
, 0, true },
6596 { "__builtin_alpha_insqh", ALPHA_BUILTIN_INSQH
, 0, true },
6597 { "__builtin_alpha_mskbl", ALPHA_BUILTIN_MSKBL
, 0, true },
6598 { "__builtin_alpha_mskwl", ALPHA_BUILTIN_MSKWL
, 0, true },
6599 { "__builtin_alpha_mskll", ALPHA_BUILTIN_MSKLL
, 0, true },
6600 { "__builtin_alpha_mskql", ALPHA_BUILTIN_MSKQL
, 0, true },
6601 { "__builtin_alpha_mskwh", ALPHA_BUILTIN_MSKWH
, 0, true },
6602 { "__builtin_alpha_msklh", ALPHA_BUILTIN_MSKLH
, 0, true },
6603 { "__builtin_alpha_mskqh", ALPHA_BUILTIN_MSKQH
, 0, true },
6604 { "__builtin_alpha_umulh", ALPHA_BUILTIN_UMULH
, 0, true },
6605 { "__builtin_alpha_zap", ALPHA_BUILTIN_ZAP
, 0, true },
6606 { "__builtin_alpha_zapnot", ALPHA_BUILTIN_ZAPNOT
, 0, true },
6607 { "__builtin_alpha_minub8", ALPHA_BUILTIN_MINUB8
, MASK_MAX
, true },
6608 { "__builtin_alpha_minsb8", ALPHA_BUILTIN_MINSB8
, MASK_MAX
, true },
6609 { "__builtin_alpha_minuw4", ALPHA_BUILTIN_MINUW4
, MASK_MAX
, true },
6610 { "__builtin_alpha_minsw4", ALPHA_BUILTIN_MINSW4
, MASK_MAX
, true },
6611 { "__builtin_alpha_maxub8", ALPHA_BUILTIN_MAXUB8
, MASK_MAX
, true },
6612 { "__builtin_alpha_maxsb8", ALPHA_BUILTIN_MAXSB8
, MASK_MAX
, true },
6613 { "__builtin_alpha_maxuw4", ALPHA_BUILTIN_MAXUW4
, MASK_MAX
, true },
6614 { "__builtin_alpha_maxsw4", ALPHA_BUILTIN_MAXSW4
, MASK_MAX
, true },
6615 { "__builtin_alpha_perr", ALPHA_BUILTIN_PERR
, MASK_MAX
, true }
6618 static GTY(()) tree alpha_v8qi_u
;
6619 static GTY(()) tree alpha_v8qi_s
;
6620 static GTY(()) tree alpha_v4hi_u
;
6621 static GTY(()) tree alpha_v4hi_s
;
6623 /* Helper function of alpha_init_builtins. Add the COUNT built-in
6624 functions pointed to by P, with function type FTYPE. */
6627 alpha_add_builtins (const struct alpha_builtin_def
*p
, size_t count
,
6633 for (i
= 0; i
< count
; ++i
, ++p
)
6634 if ((target_flags
& p
->target_mask
) == p
->target_mask
)
6636 decl
= add_builtin_function (p
->name
, ftype
, p
->code
, BUILT_IN_MD
,
6639 TREE_READONLY (decl
) = 1;
6640 TREE_NOTHROW (decl
) = 1;
6646 alpha_init_builtins (void)
6648 tree dimode_integer_type_node
;
6651 dimode_integer_type_node
= lang_hooks
.types
.type_for_mode (DImode
, 0);
6653 /* Fwrite on VMS is non-standard. */
6654 #if TARGET_ABI_OPEN_VMS
6655 implicit_built_in_decls
[(int) BUILT_IN_FWRITE
] = NULL_TREE
;
6656 implicit_built_in_decls
[(int) BUILT_IN_FWRITE_UNLOCKED
] = NULL_TREE
;
6659 ftype
= build_function_type (dimode_integer_type_node
, void_list_node
);
6660 alpha_add_builtins (zero_arg_builtins
, ARRAY_SIZE (zero_arg_builtins
),
6663 ftype
= build_function_type_list (dimode_integer_type_node
,
6664 dimode_integer_type_node
, NULL_TREE
);
6665 alpha_add_builtins (one_arg_builtins
, ARRAY_SIZE (one_arg_builtins
),
6668 ftype
= build_function_type_list (dimode_integer_type_node
,
6669 dimode_integer_type_node
,
6670 dimode_integer_type_node
, NULL_TREE
);
6671 alpha_add_builtins (two_arg_builtins
, ARRAY_SIZE (two_arg_builtins
),
6674 ftype
= build_function_type (ptr_type_node
, void_list_node
);
6675 decl
= add_builtin_function ("__builtin_thread_pointer", ftype
,
6676 ALPHA_BUILTIN_THREAD_POINTER
, BUILT_IN_MD
,
6678 TREE_NOTHROW (decl
) = 1;
6680 ftype
= build_function_type_list (void_type_node
, ptr_type_node
, NULL_TREE
);
6681 decl
= add_builtin_function ("__builtin_set_thread_pointer", ftype
,
6682 ALPHA_BUILTIN_SET_THREAD_POINTER
, BUILT_IN_MD
,
6684 TREE_NOTHROW (decl
) = 1;
6686 if (TARGET_ABI_OPEN_VMS
)
6688 ftype
= build_function_type_list (ptr_type_node
, ptr_type_node
,
6690 add_builtin_function ("__builtin_establish_vms_condition_handler", ftype
,
6691 ALPHA_BUILTIN_ESTABLISH_VMS_CONDITION_HANDLER
,
6692 BUILT_IN_MD
, NULL
, NULL_TREE
);
6694 ftype
= build_function_type_list (ptr_type_node
, void_type_node
,
6696 add_builtin_function ("__builtin_revert_vms_condition_handler", ftype
,
6697 ALPHA_BUILTIN_REVERT_VMS_CONDITION_HANDLER
,
6698 BUILT_IN_MD
, NULL
, NULL_TREE
);
6701 alpha_v8qi_u
= build_vector_type (unsigned_intQI_type_node
, 8);
6702 alpha_v8qi_s
= build_vector_type (intQI_type_node
, 8);
6703 alpha_v4hi_u
= build_vector_type (unsigned_intHI_type_node
, 4);
6704 alpha_v4hi_s
= build_vector_type (intHI_type_node
, 4);
6707 /* Expand an expression EXP that calls a built-in function,
6708 with result going to TARGET if that's convenient
6709 (and in mode MODE if that's convenient).
6710 SUBTARGET may be used as the target for computing one of EXP's operands.
6711 IGNORE is nonzero if the value is to be ignored. */
6714 alpha_expand_builtin (tree exp
, rtx target
,
6715 rtx subtarget ATTRIBUTE_UNUSED
,
6716 enum machine_mode mode ATTRIBUTE_UNUSED
,
6717 int ignore ATTRIBUTE_UNUSED
)
6721 tree fndecl
= TREE_OPERAND (CALL_EXPR_FN (exp
), 0);
6722 unsigned int fcode
= DECL_FUNCTION_CODE (fndecl
);
6724 call_expr_arg_iterator iter
;
6725 enum insn_code icode
;
6726 rtx op
[MAX_ARGS
], pat
;
6730 if (fcode
>= ALPHA_BUILTIN_max
)
6731 internal_error ("bad builtin fcode");
6732 icode
= code_for_builtin
[fcode
];
6734 internal_error ("bad builtin fcode");
6736 nonvoid
= TREE_TYPE (TREE_TYPE (fndecl
)) != void_type_node
;
6739 FOR_EACH_CALL_EXPR_ARG (arg
, iter
, exp
)
6741 const struct insn_operand_data
*insn_op
;
6743 if (arg
== error_mark_node
)
6745 if (arity
> MAX_ARGS
)
6748 insn_op
= &insn_data
[icode
].operand
[arity
+ nonvoid
];
6750 op
[arity
] = expand_expr (arg
, NULL_RTX
, insn_op
->mode
, EXPAND_NORMAL
);
6752 if (!(*insn_op
->predicate
) (op
[arity
], insn_op
->mode
))
6753 op
[arity
] = copy_to_mode_reg (insn_op
->mode
, op
[arity
]);
6759 enum machine_mode tmode
= insn_data
[icode
].operand
[0].mode
;
6761 || GET_MODE (target
) != tmode
6762 || !(*insn_data
[icode
].operand
[0].predicate
) (target
, tmode
))
6763 target
= gen_reg_rtx (tmode
);
6769 pat
= GEN_FCN (icode
) (target
);
6773 pat
= GEN_FCN (icode
) (target
, op
[0]);
6775 pat
= GEN_FCN (icode
) (op
[0]);
6778 pat
= GEN_FCN (icode
) (target
, op
[0], op
[1]);
6794 /* Several bits below assume HWI >= 64 bits. This should be enforced
6796 #if HOST_BITS_PER_WIDE_INT < 64
6797 # error "HOST_WIDE_INT too small"
6800 /* Fold the builtin for the CMPBGE instruction. This is a vector comparison
6801 with an 8-bit output vector. OPINT contains the integer operands; bit N
6802 of OP_CONST is set if OPINT[N] is valid. */
6805 alpha_fold_builtin_cmpbge (unsigned HOST_WIDE_INT opint
[], long op_const
)
6810 for (i
= 0, val
= 0; i
< 8; ++i
)
6812 unsigned HOST_WIDE_INT c0
= (opint
[0] >> (i
* 8)) & 0xff;
6813 unsigned HOST_WIDE_INT c1
= (opint
[1] >> (i
* 8)) & 0xff;
6817 return build_int_cst (long_integer_type_node
, val
);
6819 else if (op_const
== 2 && opint
[1] == 0)
6820 return build_int_cst (long_integer_type_node
, 0xff);
6824 /* Fold the builtin for the ZAPNOT instruction. This is essentially a
6825 specialized form of an AND operation. Other byte manipulation instructions
6826 are defined in terms of this instruction, so this is also used as a
6827 subroutine for other builtins.
6829 OP contains the tree operands; OPINT contains the extracted integer values.
6830 Bit N of OP_CONST it set if OPINT[N] is valid. OP may be null if only
6831 OPINT may be considered. */
6834 alpha_fold_builtin_zapnot (tree
*op
, unsigned HOST_WIDE_INT opint
[],
6839 unsigned HOST_WIDE_INT mask
= 0;
6842 for (i
= 0; i
< 8; ++i
)
6843 if ((opint
[1] >> i
) & 1)
6844 mask
|= (unsigned HOST_WIDE_INT
)0xff << (i
* 8);
6847 return build_int_cst (long_integer_type_node
, opint
[0] & mask
);
6850 return fold_build2 (BIT_AND_EXPR
, long_integer_type_node
, op
[0],
6851 build_int_cst (long_integer_type_node
, mask
));
6853 else if ((op_const
& 1) && opint
[0] == 0)
6854 return build_int_cst (long_integer_type_node
, 0);
6858 /* Fold the builtins for the EXT family of instructions. */
6861 alpha_fold_builtin_extxx (tree op
[], unsigned HOST_WIDE_INT opint
[],
6862 long op_const
, unsigned HOST_WIDE_INT bytemask
,
6866 tree
*zap_op
= NULL
;
6870 unsigned HOST_WIDE_INT loc
;
6873 if (BYTES_BIG_ENDIAN
)
6881 unsigned HOST_WIDE_INT temp
= opint
[0];
6894 opint
[1] = bytemask
;
6895 return alpha_fold_builtin_zapnot (zap_op
, opint
, zap_const
);
6898 /* Fold the builtins for the INS family of instructions. */
6901 alpha_fold_builtin_insxx (tree op
[], unsigned HOST_WIDE_INT opint
[],
6902 long op_const
, unsigned HOST_WIDE_INT bytemask
,
6905 if ((op_const
& 1) && opint
[0] == 0)
6906 return build_int_cst (long_integer_type_node
, 0);
6910 unsigned HOST_WIDE_INT temp
, loc
, byteloc
;
6911 tree
*zap_op
= NULL
;
6914 if (BYTES_BIG_ENDIAN
)
6921 byteloc
= (64 - (loc
* 8)) & 0x3f;
6938 opint
[1] = bytemask
;
6939 return alpha_fold_builtin_zapnot (zap_op
, opint
, op_const
);
6946 alpha_fold_builtin_mskxx (tree op
[], unsigned HOST_WIDE_INT opint
[],
6947 long op_const
, unsigned HOST_WIDE_INT bytemask
,
6952 unsigned HOST_WIDE_INT loc
;
6955 if (BYTES_BIG_ENDIAN
)
6962 opint
[1] = bytemask
^ 0xff;
6965 return alpha_fold_builtin_zapnot (op
, opint
, op_const
);
6969 alpha_fold_builtin_umulh (unsigned HOST_WIDE_INT opint
[], long op_const
)
6975 unsigned HOST_WIDE_INT l
;
6978 mul_double (opint
[0], 0, opint
[1], 0, &l
, &h
);
6980 #if HOST_BITS_PER_WIDE_INT > 64
6984 return build_int_cst (long_integer_type_node
, h
);
6988 opint
[1] = opint
[0];
6991 /* Note that (X*1) >> 64 == 0. */
6992 if (opint
[1] == 0 || opint
[1] == 1)
6993 return build_int_cst (long_integer_type_node
, 0);
7000 alpha_fold_vector_minmax (enum tree_code code
, tree op
[], tree vtype
)
7002 tree op0
= fold_convert (vtype
, op
[0]);
7003 tree op1
= fold_convert (vtype
, op
[1]);
7004 tree val
= fold_build2 (code
, vtype
, op0
, op1
);
7005 return fold_build1 (VIEW_CONVERT_EXPR
, long_integer_type_node
, val
);
7009 alpha_fold_builtin_perr (unsigned HOST_WIDE_INT opint
[], long op_const
)
7011 unsigned HOST_WIDE_INT temp
= 0;
7017 for (i
= 0; i
< 8; ++i
)
7019 unsigned HOST_WIDE_INT a
= (opint
[0] >> (i
* 8)) & 0xff;
7020 unsigned HOST_WIDE_INT b
= (opint
[1] >> (i
* 8)) & 0xff;
7027 return build_int_cst (long_integer_type_node
, temp
);
7031 alpha_fold_builtin_pklb (unsigned HOST_WIDE_INT opint
[], long op_const
)
7033 unsigned HOST_WIDE_INT temp
;
7038 temp
= opint
[0] & 0xff;
7039 temp
|= (opint
[0] >> 24) & 0xff00;
7041 return build_int_cst (long_integer_type_node
, temp
);
7045 alpha_fold_builtin_pkwb (unsigned HOST_WIDE_INT opint
[], long op_const
)
7047 unsigned HOST_WIDE_INT temp
;
7052 temp
= opint
[0] & 0xff;
7053 temp
|= (opint
[0] >> 8) & 0xff00;
7054 temp
|= (opint
[0] >> 16) & 0xff0000;
7055 temp
|= (opint
[0] >> 24) & 0xff000000;
7057 return build_int_cst (long_integer_type_node
, temp
);
7061 alpha_fold_builtin_unpkbl (unsigned HOST_WIDE_INT opint
[], long op_const
)
7063 unsigned HOST_WIDE_INT temp
;
7068 temp
= opint
[0] & 0xff;
7069 temp
|= (opint
[0] & 0xff00) << 24;
7071 return build_int_cst (long_integer_type_node
, temp
);
7075 alpha_fold_builtin_unpkbw (unsigned HOST_WIDE_INT opint
[], long op_const
)
7077 unsigned HOST_WIDE_INT temp
;
7082 temp
= opint
[0] & 0xff;
7083 temp
|= (opint
[0] & 0x0000ff00) << 8;
7084 temp
|= (opint
[0] & 0x00ff0000) << 16;
7085 temp
|= (opint
[0] & 0xff000000) << 24;
7087 return build_int_cst (long_integer_type_node
, temp
);
7091 alpha_fold_builtin_cttz (unsigned HOST_WIDE_INT opint
[], long op_const
)
7093 unsigned HOST_WIDE_INT temp
;
7101 temp
= exact_log2 (opint
[0] & -opint
[0]);
7103 return build_int_cst (long_integer_type_node
, temp
);
7107 alpha_fold_builtin_ctlz (unsigned HOST_WIDE_INT opint
[], long op_const
)
7109 unsigned HOST_WIDE_INT temp
;
7117 temp
= 64 - floor_log2 (opint
[0]) - 1;
7119 return build_int_cst (long_integer_type_node
, temp
);
7123 alpha_fold_builtin_ctpop (unsigned HOST_WIDE_INT opint
[], long op_const
)
7125 unsigned HOST_WIDE_INT temp
, op
;
7133 temp
++, op
&= op
- 1;
7135 return build_int_cst (long_integer_type_node
, temp
);
7138 /* Fold one of our builtin functions. */
7141 alpha_fold_builtin (tree fndecl
, tree call
, bool ignore ATTRIBUTE_UNUSED
)
7143 tree
*op
= CALL_EXPR_ARGP (call
);
7144 unsigned HOST_WIDE_INT opint
[MAX_ARGS
];
7148 for (i
= 0; i
< call_expr_nargs (call
); i
++)
7150 tree arg
= CALL_EXPR_ARG (call
, i
);
7151 if (arg
== error_mark_node
)
7153 if (arity
>= MAX_ARGS
)
7157 if (TREE_CODE (arg
) == INTEGER_CST
)
7159 op_const
|= 1L << arity
;
7160 opint
[arity
] = int_cst_value (arg
);
7164 switch (DECL_FUNCTION_CODE (fndecl
))
7166 case ALPHA_BUILTIN_CMPBGE
:
7167 return alpha_fold_builtin_cmpbge (opint
, op_const
);
7169 case ALPHA_BUILTIN_EXTBL
:
7170 return alpha_fold_builtin_extxx (op
, opint
, op_const
, 0x01, false);
7171 case ALPHA_BUILTIN_EXTWL
:
7172 return alpha_fold_builtin_extxx (op
, opint
, op_const
, 0x03, false);
7173 case ALPHA_BUILTIN_EXTLL
:
7174 return alpha_fold_builtin_extxx (op
, opint
, op_const
, 0x0f, false);
7175 case ALPHA_BUILTIN_EXTQL
:
7176 return alpha_fold_builtin_extxx (op
, opint
, op_const
, 0xff, false);
7177 case ALPHA_BUILTIN_EXTWH
:
7178 return alpha_fold_builtin_extxx (op
, opint
, op_const
, 0x03, true);
7179 case ALPHA_BUILTIN_EXTLH
:
7180 return alpha_fold_builtin_extxx (op
, opint
, op_const
, 0x0f, true);
7181 case ALPHA_BUILTIN_EXTQH
:
7182 return alpha_fold_builtin_extxx (op
, opint
, op_const
, 0xff, true);
7184 case ALPHA_BUILTIN_INSBL
:
7185 return alpha_fold_builtin_insxx (op
, opint
, op_const
, 0x01, false);
7186 case ALPHA_BUILTIN_INSWL
:
7187 return alpha_fold_builtin_insxx (op
, opint
, op_const
, 0x03, false);
7188 case ALPHA_BUILTIN_INSLL
:
7189 return alpha_fold_builtin_insxx (op
, opint
, op_const
, 0x0f, false);
7190 case ALPHA_BUILTIN_INSQL
:
7191 return alpha_fold_builtin_insxx (op
, opint
, op_const
, 0xff, false);
7192 case ALPHA_BUILTIN_INSWH
:
7193 return alpha_fold_builtin_insxx (op
, opint
, op_const
, 0x03, true);
7194 case ALPHA_BUILTIN_INSLH
:
7195 return alpha_fold_builtin_insxx (op
, opint
, op_const
, 0x0f, true);
7196 case ALPHA_BUILTIN_INSQH
:
7197 return alpha_fold_builtin_insxx (op
, opint
, op_const
, 0xff, true);
7199 case ALPHA_BUILTIN_MSKBL
:
7200 return alpha_fold_builtin_mskxx (op
, opint
, op_const
, 0x01, false);
7201 case ALPHA_BUILTIN_MSKWL
:
7202 return alpha_fold_builtin_mskxx (op
, opint
, op_const
, 0x03, false);
7203 case ALPHA_BUILTIN_MSKLL
:
7204 return alpha_fold_builtin_mskxx (op
, opint
, op_const
, 0x0f, false);
7205 case ALPHA_BUILTIN_MSKQL
:
7206 return alpha_fold_builtin_mskxx (op
, opint
, op_const
, 0xff, false);
7207 case ALPHA_BUILTIN_MSKWH
:
7208 return alpha_fold_builtin_mskxx (op
, opint
, op_const
, 0x03, true);
7209 case ALPHA_BUILTIN_MSKLH
:
7210 return alpha_fold_builtin_mskxx (op
, opint
, op_const
, 0x0f, true);
7211 case ALPHA_BUILTIN_MSKQH
:
7212 return alpha_fold_builtin_mskxx (op
, opint
, op_const
, 0xff, true);
7214 case ALPHA_BUILTIN_UMULH
:
7215 return alpha_fold_builtin_umulh (opint
, op_const
);
7217 case ALPHA_BUILTIN_ZAP
:
7220 case ALPHA_BUILTIN_ZAPNOT
:
7221 return alpha_fold_builtin_zapnot (op
, opint
, op_const
);
7223 case ALPHA_BUILTIN_MINUB8
:
7224 return alpha_fold_vector_minmax (MIN_EXPR
, op
, alpha_v8qi_u
);
7225 case ALPHA_BUILTIN_MINSB8
:
7226 return alpha_fold_vector_minmax (MIN_EXPR
, op
, alpha_v8qi_s
);
7227 case ALPHA_BUILTIN_MINUW4
:
7228 return alpha_fold_vector_minmax (MIN_EXPR
, op
, alpha_v4hi_u
);
7229 case ALPHA_BUILTIN_MINSW4
:
7230 return alpha_fold_vector_minmax (MIN_EXPR
, op
, alpha_v4hi_s
);
7231 case ALPHA_BUILTIN_MAXUB8
:
7232 return alpha_fold_vector_minmax (MAX_EXPR
, op
, alpha_v8qi_u
);
7233 case ALPHA_BUILTIN_MAXSB8
:
7234 return alpha_fold_vector_minmax (MAX_EXPR
, op
, alpha_v8qi_s
);
7235 case ALPHA_BUILTIN_MAXUW4
:
7236 return alpha_fold_vector_minmax (MAX_EXPR
, op
, alpha_v4hi_u
);
7237 case ALPHA_BUILTIN_MAXSW4
:
7238 return alpha_fold_vector_minmax (MAX_EXPR
, op
, alpha_v4hi_s
);
7240 case ALPHA_BUILTIN_PERR
:
7241 return alpha_fold_builtin_perr (opint
, op_const
);
7242 case ALPHA_BUILTIN_PKLB
:
7243 return alpha_fold_builtin_pklb (opint
, op_const
);
7244 case ALPHA_BUILTIN_PKWB
:
7245 return alpha_fold_builtin_pkwb (opint
, op_const
);
7246 case ALPHA_BUILTIN_UNPKBL
:
7247 return alpha_fold_builtin_unpkbl (opint
, op_const
);
7248 case ALPHA_BUILTIN_UNPKBW
:
7249 return alpha_fold_builtin_unpkbw (opint
, op_const
);
7251 case ALPHA_BUILTIN_CTTZ
:
7252 return alpha_fold_builtin_cttz (opint
, op_const
);
7253 case ALPHA_BUILTIN_CTLZ
:
7254 return alpha_fold_builtin_ctlz (opint
, op_const
);
7255 case ALPHA_BUILTIN_CTPOP
:
7256 return alpha_fold_builtin_ctpop (opint
, op_const
);
7258 case ALPHA_BUILTIN_AMASK
:
7259 case ALPHA_BUILTIN_IMPLVER
:
7260 case ALPHA_BUILTIN_RPCC
:
7261 case ALPHA_BUILTIN_THREAD_POINTER
:
7262 case ALPHA_BUILTIN_SET_THREAD_POINTER
:
7263 /* None of these are foldable at compile-time. */
7269 /* This page contains routines that are used to determine what the function
7270 prologue and epilogue code will do and write them out. */
7272 /* Compute the size of the save area in the stack. */
7274 /* These variables are used for communication between the following functions.
7275 They indicate various things about the current function being compiled
7276 that are used to tell what kind of prologue, epilogue and procedure
7277 descriptor to generate. */
7279 /* Nonzero if we need a stack procedure. */
7280 enum alpha_procedure_types
{PT_NULL
= 0, PT_REGISTER
= 1, PT_STACK
= 2};
7281 static enum alpha_procedure_types alpha_procedure_type
;
7283 /* Register number (either FP or SP) that is used to unwind the frame. */
7284 static int vms_unwind_regno
;
7286 /* Register number used to save FP. We need not have one for RA since
7287 we don't modify it for register procedures. This is only defined
7288 for register frame procedures. */
7289 static int vms_save_fp_regno
;
7291 /* Register number used to reference objects off our PV. */
7292 static int vms_base_regno
;
7294 /* Compute register masks for saved registers. */
7297 alpha_sa_mask (unsigned long *imaskP
, unsigned long *fmaskP
)
7299 unsigned long imask
= 0;
7300 unsigned long fmask
= 0;
7303 /* When outputting a thunk, we don't have valid register life info,
7304 but assemble_start_function wants to output .frame and .mask
7313 if (TARGET_ABI_OPEN_VMS
&& alpha_procedure_type
== PT_STACK
)
7314 imask
|= (1UL << HARD_FRAME_POINTER_REGNUM
);
7316 /* One for every register we have to save. */
7317 for (i
= 0; i
< FIRST_PSEUDO_REGISTER
; i
++)
7318 if (! fixed_regs
[i
] && ! call_used_regs
[i
]
7319 && df_regs_ever_live_p (i
) && i
!= REG_RA
7320 && (!TARGET_ABI_UNICOSMK
|| i
!= HARD_FRAME_POINTER_REGNUM
))
7323 imask
|= (1UL << i
);
7325 fmask
|= (1UL << (i
- 32));
7328 /* We need to restore these for the handler. */
7329 if (crtl
->calls_eh_return
)
7333 unsigned regno
= EH_RETURN_DATA_REGNO (i
);
7334 if (regno
== INVALID_REGNUM
)
7336 imask
|= 1UL << regno
;
7340 /* If any register spilled, then spill the return address also. */
7341 /* ??? This is required by the Digital stack unwind specification
7342 and isn't needed if we're doing Dwarf2 unwinding. */
7343 if (imask
|| fmask
|| alpha_ra_ever_killed ())
7344 imask
|= (1UL << REG_RA
);
7351 alpha_sa_size (void)
7353 unsigned long mask
[2];
7357 alpha_sa_mask (&mask
[0], &mask
[1]);
7359 if (TARGET_ABI_UNICOSMK
)
7361 if (mask
[0] || mask
[1])
7366 for (j
= 0; j
< 2; ++j
)
7367 for (i
= 0; i
< 32; ++i
)
7368 if ((mask
[j
] >> i
) & 1)
7372 if (TARGET_ABI_UNICOSMK
)
7374 /* We might not need to generate a frame if we don't make any calls
7375 (including calls to __T3E_MISMATCH if this is a vararg function),
7376 don't have any local variables which require stack slots, don't
7377 use alloca and have not determined that we need a frame for other
7380 alpha_procedure_type
7381 = (sa_size
|| get_frame_size() != 0
7382 || crtl
->outgoing_args_size
7383 || cfun
->stdarg
|| cfun
->calls_alloca
7384 || frame_pointer_needed
)
7385 ? PT_STACK
: PT_REGISTER
;
7387 /* Always reserve space for saving callee-saved registers if we
7388 need a frame as required by the calling convention. */
7389 if (alpha_procedure_type
== PT_STACK
)
7392 else if (TARGET_ABI_OPEN_VMS
)
7394 /* Start with a stack procedure if we make any calls (REG_RA used), or
7395 need a frame pointer, with a register procedure if we otherwise need
7396 at least a slot, and with a null procedure in other cases. */
7397 if ((mask
[0] >> REG_RA
) & 1 || frame_pointer_needed
)
7398 alpha_procedure_type
= PT_STACK
;
7399 else if (get_frame_size() != 0)
7400 alpha_procedure_type
= PT_REGISTER
;
7402 alpha_procedure_type
= PT_NULL
;
7404 /* Don't reserve space for saving FP & RA yet. Do that later after we've
7405 made the final decision on stack procedure vs register procedure. */
7406 if (alpha_procedure_type
== PT_STACK
)
7409 /* Decide whether to refer to objects off our PV via FP or PV.
7410 If we need FP for something else or if we receive a nonlocal
7411 goto (which expects PV to contain the value), we must use PV.
7412 Otherwise, start by assuming we can use FP. */
7415 = (frame_pointer_needed
7416 || cfun
->has_nonlocal_label
7417 || alpha_procedure_type
== PT_STACK
7418 || crtl
->outgoing_args_size
)
7419 ? REG_PV
: HARD_FRAME_POINTER_REGNUM
;
7421 /* If we want to copy PV into FP, we need to find some register
7422 in which to save FP. */
7424 vms_save_fp_regno
= -1;
7425 if (vms_base_regno
== HARD_FRAME_POINTER_REGNUM
)
7426 for (i
= 0; i
< 32; i
++)
7427 if (! fixed_regs
[i
] && call_used_regs
[i
] && ! df_regs_ever_live_p (i
))
7428 vms_save_fp_regno
= i
;
7430 /* A VMS condition handler requires a stack procedure in our
7431 implementation. (not required by the calling standard). */
7432 if ((vms_save_fp_regno
== -1 && alpha_procedure_type
== PT_REGISTER
)
7433 || cfun
->machine
->uses_condition_handler
)
7434 vms_base_regno
= REG_PV
, alpha_procedure_type
= PT_STACK
;
7435 else if (alpha_procedure_type
== PT_NULL
)
7436 vms_base_regno
= REG_PV
;
7438 /* Stack unwinding should be done via FP unless we use it for PV. */
7439 vms_unwind_regno
= (vms_base_regno
== REG_PV
7440 ? HARD_FRAME_POINTER_REGNUM
: STACK_POINTER_REGNUM
);
7442 /* If this is a stack procedure, allow space for saving FP, RA and
7443 a condition handler slot if needed. */
7444 if (alpha_procedure_type
== PT_STACK
)
7445 sa_size
+= 2 + cfun
->machine
->uses_condition_handler
;
7449 /* Our size must be even (multiple of 16 bytes). */
7457 /* Define the offset between two registers, one to be eliminated,
7458 and the other its replacement, at the start of a routine. */
7461 alpha_initial_elimination_offset (unsigned int from
,
7462 unsigned int to ATTRIBUTE_UNUSED
)
7466 ret
= alpha_sa_size ();
7467 ret
+= ALPHA_ROUND (crtl
->outgoing_args_size
);
7471 case FRAME_POINTER_REGNUM
:
7474 case ARG_POINTER_REGNUM
:
7475 ret
+= (ALPHA_ROUND (get_frame_size ()
7476 + crtl
->args
.pretend_args_size
)
7477 - crtl
->args
.pretend_args_size
);
7487 #if TARGET_ABI_OPEN_VMS
7489 /* Worker function for TARGET_CAN_ELIMINATE. */
7492 alpha_vms_can_eliminate (const int from ATTRIBUTE_UNUSED
, const int to
)
7494 /* We need the alpha_procedure_type to decide. Evaluate it now. */
7497 switch (alpha_procedure_type
)
7500 /* NULL procedures have no frame of their own and we only
7501 know how to resolve from the current stack pointer. */
7502 return to
== STACK_POINTER_REGNUM
;
7506 /* We always eliminate except to the stack pointer if there is no
7507 usable frame pointer at hand. */
7508 return (to
!= STACK_POINTER_REGNUM
7509 || vms_unwind_regno
!= HARD_FRAME_POINTER_REGNUM
);
7515 /* FROM is to be eliminated for TO. Return the offset so that TO+offset
7516 designates the same location as FROM. */
7519 alpha_vms_initial_elimination_offset (unsigned int from
, unsigned int to
)
7521 /* The only possible attempts we ever expect are ARG or FRAME_PTR to
7522 HARD_FRAME or STACK_PTR. We need the alpha_procedure_type to decide
7523 on the proper computations and will need the register save area size
7526 HOST_WIDE_INT sa_size
= alpha_sa_size ();
7528 /* PT_NULL procedures have no frame of their own and we only allow
7529 elimination to the stack pointer. This is the argument pointer and we
7530 resolve the soft frame pointer to that as well. */
7532 if (alpha_procedure_type
== PT_NULL
)
7535 /* For a PT_STACK procedure the frame layout looks as follows
7537 -----> decreasing addresses
7539 < size rounded up to 16 | likewise >
7540 --------------#------------------------------+++--------------+++-------#
7541 incoming args # pretended args | "frame" | regs sa | PV | outgoing args #
7542 --------------#---------------------------------------------------------#
7544 ARG_PTR FRAME_PTR HARD_FRAME_PTR STACK_PTR
7547 PT_REGISTER procedures are similar in that they may have a frame of their
7548 own. They have no regs-sa/pv/outgoing-args area.
7550 We first compute offset to HARD_FRAME_PTR, then add what we need to get
7551 to STACK_PTR if need be. */
7554 HOST_WIDE_INT offset
;
7555 HOST_WIDE_INT pv_save_size
= alpha_procedure_type
== PT_STACK
? 8 : 0;
7559 case FRAME_POINTER_REGNUM
:
7560 offset
= ALPHA_ROUND (sa_size
+ pv_save_size
);
7562 case ARG_POINTER_REGNUM
:
7563 offset
= (ALPHA_ROUND (sa_size
+ pv_save_size
7565 + crtl
->args
.pretend_args_size
)
7566 - crtl
->args
.pretend_args_size
);
7572 if (to
== STACK_POINTER_REGNUM
)
7573 offset
+= ALPHA_ROUND (crtl
->outgoing_args_size
);
7579 #define COMMON_OBJECT "common_object"
7582 common_object_handler (tree
*node
, tree name ATTRIBUTE_UNUSED
,
7583 tree args ATTRIBUTE_UNUSED
, int flags ATTRIBUTE_UNUSED
,
7584 bool *no_add_attrs ATTRIBUTE_UNUSED
)
7587 gcc_assert (DECL_P (decl
));
7589 DECL_COMMON (decl
) = 1;
7593 static const struct attribute_spec vms_attribute_table
[] =
7595 /* { name, min_len, max_len, decl_req, type_req, fn_type_req, handler } */
7596 { COMMON_OBJECT
, 0, 1, true, false, false, common_object_handler
},
7597 { NULL
, 0, 0, false, false, false, NULL
}
7601 vms_output_aligned_decl_common(FILE *file
, tree decl
, const char *name
,
7602 unsigned HOST_WIDE_INT size
,
7605 tree attr
= DECL_ATTRIBUTES (decl
);
7606 fprintf (file
, "%s", COMMON_ASM_OP
);
7607 assemble_name (file
, name
);
7608 fprintf (file
, "," HOST_WIDE_INT_PRINT_UNSIGNED
, size
);
7609 /* ??? Unlike on OSF/1, the alignment factor is not in log units. */
7610 fprintf (file
, ",%u", align
/ BITS_PER_UNIT
);
7613 attr
= lookup_attribute (COMMON_OBJECT
, attr
);
7615 fprintf (file
, ",%s",
7616 IDENTIFIER_POINTER (TREE_VALUE (TREE_VALUE (attr
))));
7621 #undef COMMON_OBJECT
7626 find_lo_sum_using_gp (rtx
*px
, void *data ATTRIBUTE_UNUSED
)
7628 return GET_CODE (*px
) == LO_SUM
&& XEXP (*px
, 0) == pic_offset_table_rtx
;
7632 alpha_find_lo_sum_using_gp (rtx insn
)
7634 return for_each_rtx (&PATTERN (insn
), find_lo_sum_using_gp
, NULL
) > 0;
7638 alpha_does_function_need_gp (void)
7642 /* The GP being variable is an OSF abi thing. */
7643 if (! TARGET_ABI_OSF
)
7646 /* We need the gp to load the address of __mcount. */
7647 if (TARGET_PROFILING_NEEDS_GP
&& crtl
->profile
)
7650 /* The code emitted by alpha_output_mi_thunk_osf uses the gp. */
7654 /* The nonlocal receiver pattern assumes that the gp is valid for
7655 the nested function. Reasonable because it's almost always set
7656 correctly already. For the cases where that's wrong, make sure
7657 the nested function loads its gp on entry. */
7658 if (crtl
->has_nonlocal_goto
)
7661 /* If we need a GP (we have a LDSYM insn or a CALL_INSN), load it first.
7662 Even if we are a static function, we still need to do this in case
7663 our address is taken and passed to something like qsort. */
7665 push_topmost_sequence ();
7666 insn
= get_insns ();
7667 pop_topmost_sequence ();
7669 for (; insn
; insn
= NEXT_INSN (insn
))
7670 if (NONDEBUG_INSN_P (insn
)
7671 && ! JUMP_TABLE_DATA_P (insn
)
7672 && GET_CODE (PATTERN (insn
)) != USE
7673 && GET_CODE (PATTERN (insn
)) != CLOBBER
7674 && get_attr_usegp (insn
))
7681 /* Helper function to set RTX_FRAME_RELATED_P on instructions, including
7685 set_frame_related_p (void)
7687 rtx seq
= get_insns ();
7698 while (insn
!= NULL_RTX
)
7700 RTX_FRAME_RELATED_P (insn
) = 1;
7701 insn
= NEXT_INSN (insn
);
7703 seq
= emit_insn (seq
);
7707 seq
= emit_insn (seq
);
7708 RTX_FRAME_RELATED_P (seq
) = 1;
7713 #define FRP(exp) (start_sequence (), exp, set_frame_related_p ())
7715 /* Generates a store with the proper unwind info attached. VALUE is
7716 stored at BASE_REG+BASE_OFS. If FRAME_BIAS is nonzero, then BASE_REG
7717 contains SP+FRAME_BIAS, and that is the unwind info that should be
7718 generated. If FRAME_REG != VALUE, then VALUE is being stored on
7719 behalf of FRAME_REG, and FRAME_REG should be present in the unwind. */
7722 emit_frame_store_1 (rtx value
, rtx base_reg
, HOST_WIDE_INT frame_bias
,
7723 HOST_WIDE_INT base_ofs
, rtx frame_reg
)
7725 rtx addr
, mem
, insn
;
7727 addr
= plus_constant (base_reg
, base_ofs
);
7728 mem
= gen_rtx_MEM (DImode
, addr
);
7729 set_mem_alias_set (mem
, alpha_sr_alias_set
);
7731 insn
= emit_move_insn (mem
, value
);
7732 RTX_FRAME_RELATED_P (insn
) = 1;
7734 if (frame_bias
|| value
!= frame_reg
)
7738 addr
= plus_constant (stack_pointer_rtx
, frame_bias
+ base_ofs
);
7739 mem
= gen_rtx_MEM (DImode
, addr
);
7742 add_reg_note (insn
, REG_FRAME_RELATED_EXPR
,
7743 gen_rtx_SET (VOIDmode
, mem
, frame_reg
));
7748 emit_frame_store (unsigned int regno
, rtx base_reg
,
7749 HOST_WIDE_INT frame_bias
, HOST_WIDE_INT base_ofs
)
7751 rtx reg
= gen_rtx_REG (DImode
, regno
);
7752 emit_frame_store_1 (reg
, base_reg
, frame_bias
, base_ofs
, reg
);
7755 /* Write function prologue. */
7757 /* On vms we have two kinds of functions:
7759 - stack frame (PROC_STACK)
7760 these are 'normal' functions with local vars and which are
7761 calling other functions
7762 - register frame (PROC_REGISTER)
7763 keeps all data in registers, needs no stack
7765 We must pass this to the assembler so it can generate the
7766 proper pdsc (procedure descriptor)
7767 This is done with the '.pdesc' command.
7769 On not-vms, we don't really differentiate between the two, as we can
7770 simply allocate stack without saving registers. */
7773 alpha_expand_prologue (void)
7775 /* Registers to save. */
7776 unsigned long imask
= 0;
7777 unsigned long fmask
= 0;
7778 /* Stack space needed for pushing registers clobbered by us. */
7779 HOST_WIDE_INT sa_size
;
7780 /* Complete stack size needed. */
7781 HOST_WIDE_INT frame_size
;
7782 /* Offset from base reg to register save area. */
7783 HOST_WIDE_INT reg_offset
;
7787 sa_size
= alpha_sa_size ();
7789 frame_size
= get_frame_size ();
7790 if (TARGET_ABI_OPEN_VMS
)
7791 frame_size
= ALPHA_ROUND (sa_size
7792 + (alpha_procedure_type
== PT_STACK
? 8 : 0)
7794 + crtl
->args
.pretend_args_size
);
7795 else if (TARGET_ABI_UNICOSMK
)
7796 /* We have to allocate space for the DSIB if we generate a frame. */
7797 frame_size
= ALPHA_ROUND (sa_size
7798 + (alpha_procedure_type
== PT_STACK
? 48 : 0))
7799 + ALPHA_ROUND (frame_size
7800 + crtl
->outgoing_args_size
);
7802 frame_size
= (ALPHA_ROUND (crtl
->outgoing_args_size
)
7804 + ALPHA_ROUND (frame_size
7805 + crtl
->args
.pretend_args_size
));
7807 if (TARGET_ABI_OPEN_VMS
)
7808 reg_offset
= 8 + 8 * cfun
->machine
->uses_condition_handler
;
7810 reg_offset
= ALPHA_ROUND (crtl
->outgoing_args_size
);
7812 alpha_sa_mask (&imask
, &fmask
);
7814 /* Emit an insn to reload GP, if needed. */
7817 alpha_function_needs_gp
= alpha_does_function_need_gp ();
7818 if (alpha_function_needs_gp
)
7819 emit_insn (gen_prologue_ldgp ());
7822 /* TARGET_PROFILING_NEEDS_GP actually implies that we need to insert
7823 the call to mcount ourselves, rather than having the linker do it
7824 magically in response to -pg. Since _mcount has special linkage,
7825 don't represent the call as a call. */
7826 if (TARGET_PROFILING_NEEDS_GP
&& crtl
->profile
)
7827 emit_insn (gen_prologue_mcount ());
7829 if (TARGET_ABI_UNICOSMK
)
7830 unicosmk_gen_dsib (&imask
);
7832 /* Adjust the stack by the frame size. If the frame size is > 4096
7833 bytes, we need to be sure we probe somewhere in the first and last
7834 4096 bytes (we can probably get away without the latter test) and
7835 every 8192 bytes in between. If the frame size is > 32768, we
7836 do this in a loop. Otherwise, we generate the explicit probe
7839 Note that we are only allowed to adjust sp once in the prologue. */
7841 if (frame_size
<= 32768)
7843 if (frame_size
> 4096)
7847 for (probed
= 4096; probed
< frame_size
; probed
+= 8192)
7848 emit_insn (gen_probe_stack (GEN_INT (TARGET_ABI_UNICOSMK
7852 /* We only have to do this probe if we aren't saving registers. */
7853 if (sa_size
== 0 && frame_size
> probed
- 4096)
7854 emit_insn (gen_probe_stack (GEN_INT (-frame_size
)));
7857 if (frame_size
!= 0)
7858 FRP (emit_insn (gen_adddi3 (stack_pointer_rtx
, stack_pointer_rtx
,
7859 GEN_INT (TARGET_ABI_UNICOSMK
7865 /* Here we generate code to set R22 to SP + 4096 and set R23 to the
7866 number of 8192 byte blocks to probe. We then probe each block
7867 in the loop and then set SP to the proper location. If the
7868 amount remaining is > 4096, we have to do one more probe if we
7869 are not saving any registers. */
7871 HOST_WIDE_INT blocks
= (frame_size
+ 4096) / 8192;
7872 HOST_WIDE_INT leftover
= frame_size
+ 4096 - blocks
* 8192;
7873 rtx ptr
= gen_rtx_REG (DImode
, 22);
7874 rtx count
= gen_rtx_REG (DImode
, 23);
7877 emit_move_insn (count
, GEN_INT (blocks
));
7878 emit_insn (gen_adddi3 (ptr
, stack_pointer_rtx
,
7879 GEN_INT (TARGET_ABI_UNICOSMK
? 4096 - 64 : 4096)));
7881 /* Because of the difficulty in emitting a new basic block this
7882 late in the compilation, generate the loop as a single insn. */
7883 emit_insn (gen_prologue_stack_probe_loop (count
, ptr
));
7885 if (leftover
> 4096 && sa_size
== 0)
7887 rtx last
= gen_rtx_MEM (DImode
, plus_constant (ptr
, -leftover
));
7888 MEM_VOLATILE_P (last
) = 1;
7889 emit_move_insn (last
, const0_rtx
);
7892 if (TARGET_ABI_WINDOWS_NT
)
7894 /* For NT stack unwind (done by 'reverse execution'), it's
7895 not OK to take the result of a loop, even though the value
7896 is already in ptr, so we reload it via a single operation
7897 and subtract it to sp.
7899 Yes, that's correct -- we have to reload the whole constant
7900 into a temporary via ldah+lda then subtract from sp. */
7902 HOST_WIDE_INT lo
, hi
;
7903 lo
= ((frame_size
& 0xffff) ^ 0x8000) - 0x8000;
7904 hi
= frame_size
- lo
;
7906 emit_move_insn (ptr
, GEN_INT (hi
));
7907 emit_insn (gen_adddi3 (ptr
, ptr
, GEN_INT (lo
)));
7908 seq
= emit_insn (gen_subdi3 (stack_pointer_rtx
, stack_pointer_rtx
,
7913 seq
= emit_insn (gen_adddi3 (stack_pointer_rtx
, ptr
,
7914 GEN_INT (-leftover
)));
7917 /* This alternative is special, because the DWARF code cannot
7918 possibly intuit through the loop above. So we invent this
7919 note it looks at instead. */
7920 RTX_FRAME_RELATED_P (seq
) = 1;
7921 add_reg_note (seq
, REG_FRAME_RELATED_EXPR
,
7922 gen_rtx_SET (VOIDmode
, stack_pointer_rtx
,
7923 gen_rtx_PLUS (Pmode
, stack_pointer_rtx
,
7924 GEN_INT (TARGET_ABI_UNICOSMK
7929 if (!TARGET_ABI_UNICOSMK
)
7931 HOST_WIDE_INT sa_bias
= 0;
7933 /* Cope with very large offsets to the register save area. */
7934 sa_reg
= stack_pointer_rtx
;
7935 if (reg_offset
+ sa_size
> 0x8000)
7937 int low
= ((reg_offset
& 0xffff) ^ 0x8000) - 0x8000;
7940 if (low
+ sa_size
<= 0x8000)
7941 sa_bias
= reg_offset
- low
, reg_offset
= low
;
7943 sa_bias
= reg_offset
, reg_offset
= 0;
7945 sa_reg
= gen_rtx_REG (DImode
, 24);
7946 sa_bias_rtx
= GEN_INT (sa_bias
);
7948 if (add_operand (sa_bias_rtx
, DImode
))
7949 emit_insn (gen_adddi3 (sa_reg
, stack_pointer_rtx
, sa_bias_rtx
));
7952 emit_move_insn (sa_reg
, sa_bias_rtx
);
7953 emit_insn (gen_adddi3 (sa_reg
, stack_pointer_rtx
, sa_reg
));
7957 /* Save regs in stack order. Beginning with VMS PV. */
7958 if (TARGET_ABI_OPEN_VMS
&& alpha_procedure_type
== PT_STACK
)
7959 emit_frame_store (REG_PV
, stack_pointer_rtx
, 0, 0);
7961 /* Save register RA next. */
7962 if (imask
& (1UL << REG_RA
))
7964 emit_frame_store (REG_RA
, sa_reg
, sa_bias
, reg_offset
);
7965 imask
&= ~(1UL << REG_RA
);
7969 /* Now save any other registers required to be saved. */
7970 for (i
= 0; i
< 31; i
++)
7971 if (imask
& (1UL << i
))
7973 emit_frame_store (i
, sa_reg
, sa_bias
, reg_offset
);
7977 for (i
= 0; i
< 31; i
++)
7978 if (fmask
& (1UL << i
))
7980 emit_frame_store (i
+32, sa_reg
, sa_bias
, reg_offset
);
7984 else if (TARGET_ABI_UNICOSMK
&& alpha_procedure_type
== PT_STACK
)
7986 /* The standard frame on the T3E includes space for saving registers.
7987 We just have to use it. We don't have to save the return address and
7988 the old frame pointer here - they are saved in the DSIB. */
7991 for (i
= 9; i
< 15; i
++)
7992 if (imask
& (1UL << i
))
7994 emit_frame_store (i
, hard_frame_pointer_rtx
, 0, reg_offset
);
7997 for (i
= 2; i
< 10; i
++)
7998 if (fmask
& (1UL << i
))
8000 emit_frame_store (i
+32, hard_frame_pointer_rtx
, 0, reg_offset
);
8005 if (TARGET_ABI_OPEN_VMS
)
8007 /* Register frame procedures save the fp. */
8008 if (alpha_procedure_type
== PT_REGISTER
)
8010 rtx insn
= emit_move_insn (gen_rtx_REG (DImode
, vms_save_fp_regno
),
8011 hard_frame_pointer_rtx
);
8012 add_reg_note (insn
, REG_CFA_REGISTER
, NULL
);
8013 RTX_FRAME_RELATED_P (insn
) = 1;
8016 if (alpha_procedure_type
!= PT_NULL
&& vms_base_regno
!= REG_PV
)
8017 emit_insn (gen_force_movdi (gen_rtx_REG (DImode
, vms_base_regno
),
8018 gen_rtx_REG (DImode
, REG_PV
)));
8020 if (alpha_procedure_type
!= PT_NULL
8021 && vms_unwind_regno
== HARD_FRAME_POINTER_REGNUM
)
8022 FRP (emit_move_insn (hard_frame_pointer_rtx
, stack_pointer_rtx
));
8024 /* If we have to allocate space for outgoing args, do it now. */
8025 if (crtl
->outgoing_args_size
!= 0)
8028 = emit_move_insn (stack_pointer_rtx
,
8030 (hard_frame_pointer_rtx
,
8032 (crtl
->outgoing_args_size
))));
8034 /* Only set FRAME_RELATED_P on the stack adjustment we just emitted
8035 if ! frame_pointer_needed. Setting the bit will change the CFA
8036 computation rule to use sp again, which would be wrong if we had
8037 frame_pointer_needed, as this means sp might move unpredictably
8041 frame_pointer_needed
8042 => vms_unwind_regno == HARD_FRAME_POINTER_REGNUM
8044 crtl->outgoing_args_size != 0
8045 => alpha_procedure_type != PT_NULL,
8047 so when we are not setting the bit here, we are guaranteed to
8048 have emitted an FRP frame pointer update just before. */
8049 RTX_FRAME_RELATED_P (seq
) = ! frame_pointer_needed
;
8052 else if (!TARGET_ABI_UNICOSMK
)
8054 /* If we need a frame pointer, set it from the stack pointer. */
8055 if (frame_pointer_needed
)
8057 if (TARGET_CAN_FAULT_IN_PROLOGUE
)
8058 FRP (emit_move_insn (hard_frame_pointer_rtx
, stack_pointer_rtx
));
8060 /* This must always be the last instruction in the
8061 prologue, thus we emit a special move + clobber. */
8062 FRP (emit_insn (gen_init_fp (hard_frame_pointer_rtx
,
8063 stack_pointer_rtx
, sa_reg
)));
8067 /* The ABIs for VMS and OSF/1 say that while we can schedule insns into
8068 the prologue, for exception handling reasons, we cannot do this for
8069 any insn that might fault. We could prevent this for mems with a
8070 (clobber:BLK (scratch)), but this doesn't work for fp insns. So we
8071 have to prevent all such scheduling with a blockage.
8073 Linux, on the other hand, never bothered to implement OSF/1's
8074 exception handling, and so doesn't care about such things. Anyone
8075 planning to use dwarf2 frame-unwind info can also omit the blockage. */
8077 if (! TARGET_CAN_FAULT_IN_PROLOGUE
)
8078 emit_insn (gen_blockage ());
8081 /* Count the number of .file directives, so that .loc is up to date. */
8082 int num_source_filenames
= 0;
8084 /* Output the textual info surrounding the prologue. */
8087 alpha_start_function (FILE *file
, const char *fnname
,
8088 tree decl ATTRIBUTE_UNUSED
)
8090 unsigned long imask
= 0;
8091 unsigned long fmask
= 0;
8092 /* Stack space needed for pushing registers clobbered by us. */
8093 HOST_WIDE_INT sa_size
;
8094 /* Complete stack size needed. */
8095 unsigned HOST_WIDE_INT frame_size
;
8096 /* The maximum debuggable frame size (512 Kbytes using Tru64 as). */
8097 unsigned HOST_WIDE_INT max_frame_size
= TARGET_ABI_OSF
&& !TARGET_GAS
8100 /* Offset from base reg to register save area. */
8101 HOST_WIDE_INT reg_offset
;
8102 char *entry_label
= (char *) alloca (strlen (fnname
) + 6);
8103 char *tramp_label
= (char *) alloca (strlen (fnname
) + 6);
8106 /* Don't emit an extern directive for functions defined in the same file. */
8107 if (TARGET_ABI_UNICOSMK
)
8110 name_tree
= get_identifier (fnname
);
8111 TREE_ASM_WRITTEN (name_tree
) = 1;
8114 #if TARGET_ABI_OPEN_VMS
8116 && strncmp (vms_debug_main
, fnname
, strlen (vms_debug_main
)) == 0)
8118 targetm
.asm_out
.globalize_label (asm_out_file
, VMS_DEBUG_MAIN_POINTER
);
8119 ASM_OUTPUT_DEF (asm_out_file
, VMS_DEBUG_MAIN_POINTER
, fnname
);
8120 switch_to_section (text_section
);
8121 vms_debug_main
= NULL
;
8125 alpha_fnname
= fnname
;
8126 sa_size
= alpha_sa_size ();
8128 frame_size
= get_frame_size ();
8129 if (TARGET_ABI_OPEN_VMS
)
8130 frame_size
= ALPHA_ROUND (sa_size
8131 + (alpha_procedure_type
== PT_STACK
? 8 : 0)
8133 + crtl
->args
.pretend_args_size
);
8134 else if (TARGET_ABI_UNICOSMK
)
8135 frame_size
= ALPHA_ROUND (sa_size
8136 + (alpha_procedure_type
== PT_STACK
? 48 : 0))
8137 + ALPHA_ROUND (frame_size
8138 + crtl
->outgoing_args_size
);
8140 frame_size
= (ALPHA_ROUND (crtl
->outgoing_args_size
)
8142 + ALPHA_ROUND (frame_size
8143 + crtl
->args
.pretend_args_size
));
8145 if (TARGET_ABI_OPEN_VMS
)
8146 reg_offset
= 8 + 8 * cfun
->machine
->uses_condition_handler
;
8148 reg_offset
= ALPHA_ROUND (crtl
->outgoing_args_size
);
8150 alpha_sa_mask (&imask
, &fmask
);
8152 /* Ecoff can handle multiple .file directives, so put out file and lineno.
8153 We have to do that before the .ent directive as we cannot switch
8154 files within procedures with native ecoff because line numbers are
8155 linked to procedure descriptors.
8156 Outputting the lineno helps debugging of one line functions as they
8157 would otherwise get no line number at all. Please note that we would
8158 like to put out last_linenum from final.c, but it is not accessible. */
8160 if (write_symbols
== SDB_DEBUG
)
8162 #ifdef ASM_OUTPUT_SOURCE_FILENAME
8163 ASM_OUTPUT_SOURCE_FILENAME (file
,
8164 DECL_SOURCE_FILE (current_function_decl
));
8166 #ifdef SDB_OUTPUT_SOURCE_LINE
8167 if (debug_info_level
!= DINFO_LEVEL_TERSE
)
8168 SDB_OUTPUT_SOURCE_LINE (file
,
8169 DECL_SOURCE_LINE (current_function_decl
));
8173 /* Issue function start and label. */
8174 if (TARGET_ABI_OPEN_VMS
8175 || (!TARGET_ABI_UNICOSMK
&& !flag_inhibit_size_directive
))
8177 fputs ("\t.ent ", file
);
8178 assemble_name (file
, fnname
);
8181 /* If the function needs GP, we'll write the "..ng" label there.
8182 Otherwise, do it here. */
8184 && ! alpha_function_needs_gp
8185 && ! cfun
->is_thunk
)
8188 assemble_name (file
, fnname
);
8189 fputs ("..ng:\n", file
);
8192 /* Nested functions on VMS that are potentially called via trampoline
8193 get a special transfer entry point that loads the called functions
8194 procedure descriptor and static chain. */
8195 if (TARGET_ABI_OPEN_VMS
8196 && !TREE_PUBLIC (decl
)
8197 && DECL_CONTEXT (decl
)
8198 && !TYPE_P (DECL_CONTEXT (decl
)))
8200 strcpy (tramp_label
, fnname
);
8201 strcat (tramp_label
, "..tr");
8202 ASM_OUTPUT_LABEL (file
, tramp_label
);
8203 fprintf (file
, "\tldq $1,24($27)\n");
8204 fprintf (file
, "\tldq $27,16($27)\n");
8207 strcpy (entry_label
, fnname
);
8208 if (TARGET_ABI_OPEN_VMS
)
8209 strcat (entry_label
, "..en");
8211 /* For public functions, the label must be globalized by appending an
8212 additional colon. */
8213 if (TARGET_ABI_UNICOSMK
&& TREE_PUBLIC (decl
))
8214 strcat (entry_label
, ":");
8216 ASM_OUTPUT_LABEL (file
, entry_label
);
8217 inside_function
= TRUE
;
8219 if (TARGET_ABI_OPEN_VMS
)
8220 fprintf (file
, "\t.base $%d\n", vms_base_regno
);
8222 if (!TARGET_ABI_OPEN_VMS
&& !TARGET_ABI_UNICOSMK
&& TARGET_IEEE_CONFORMANT
8223 && !flag_inhibit_size_directive
)
8225 /* Set flags in procedure descriptor to request IEEE-conformant
8226 math-library routines. The value we set it to is PDSC_EXC_IEEE
8227 (/usr/include/pdsc.h). */
8228 fputs ("\t.eflag 48\n", file
);
8231 /* Set up offsets to alpha virtual arg/local debugging pointer. */
8232 alpha_auto_offset
= -frame_size
+ crtl
->args
.pretend_args_size
;
8233 alpha_arg_offset
= -frame_size
+ 48;
8235 /* Describe our frame. If the frame size is larger than an integer,
8236 print it as zero to avoid an assembler error. We won't be
8237 properly describing such a frame, but that's the best we can do. */
8238 if (TARGET_ABI_UNICOSMK
)
8240 else if (TARGET_ABI_OPEN_VMS
)
8241 fprintf (file
, "\t.frame $%d," HOST_WIDE_INT_PRINT_DEC
",$26,"
8242 HOST_WIDE_INT_PRINT_DEC
"\n",
8244 frame_size
>= (1UL << 31) ? 0 : frame_size
,
8246 else if (!flag_inhibit_size_directive
)
8247 fprintf (file
, "\t.frame $%d," HOST_WIDE_INT_PRINT_DEC
",$26,%d\n",
8248 (frame_pointer_needed
8249 ? HARD_FRAME_POINTER_REGNUM
: STACK_POINTER_REGNUM
),
8250 frame_size
>= max_frame_size
? 0 : frame_size
,
8251 crtl
->args
.pretend_args_size
);
8253 /* Describe which registers were spilled. */
8254 if (TARGET_ABI_UNICOSMK
)
8256 else if (TARGET_ABI_OPEN_VMS
)
8259 /* ??? Does VMS care if mask contains ra? The old code didn't
8260 set it, so I don't here. */
8261 fprintf (file
, "\t.mask 0x%lx,0\n", imask
& ~(1UL << REG_RA
));
8263 fprintf (file
, "\t.fmask 0x%lx,0\n", fmask
);
8264 if (alpha_procedure_type
== PT_REGISTER
)
8265 fprintf (file
, "\t.fp_save $%d\n", vms_save_fp_regno
);
8267 else if (!flag_inhibit_size_directive
)
8271 fprintf (file
, "\t.mask 0x%lx," HOST_WIDE_INT_PRINT_DEC
"\n", imask
,
8272 frame_size
>= max_frame_size
? 0 : reg_offset
- frame_size
);
8274 for (i
= 0; i
< 32; ++i
)
8275 if (imask
& (1UL << i
))
8280 fprintf (file
, "\t.fmask 0x%lx," HOST_WIDE_INT_PRINT_DEC
"\n", fmask
,
8281 frame_size
>= max_frame_size
? 0 : reg_offset
- frame_size
);
8284 #if TARGET_ABI_OPEN_VMS
8285 /* If a user condition handler has been installed at some point, emit
8286 the procedure descriptor bits to point the Condition Handling Facility
8287 at the indirection wrapper, and state the fp offset at which the user
8288 handler may be found. */
8289 if (cfun
->machine
->uses_condition_handler
)
8291 fprintf (file
, "\t.handler __gcc_shell_handler\n");
8292 fprintf (file
, "\t.handler_data %d\n", VMS_COND_HANDLER_FP_OFFSET
);
8295 /* Ifdef'ed cause link_section are only available then. */
8296 switch_to_section (readonly_data_section
);
8297 fprintf (file
, "\t.align 3\n");
8298 assemble_name (file
, fnname
); fputs ("..na:\n", file
);
8299 fputs ("\t.ascii \"", file
);
8300 assemble_name (file
, fnname
);
8301 fputs ("\\0\"\n", file
);
8302 alpha_need_linkage (fnname
, 1);
8303 switch_to_section (text_section
);
8307 /* Emit the .prologue note at the scheduled end of the prologue. */
8310 alpha_output_function_end_prologue (FILE *file
)
8312 if (TARGET_ABI_UNICOSMK
)
8314 else if (TARGET_ABI_OPEN_VMS
)
8315 fputs ("\t.prologue\n", file
);
8316 else if (TARGET_ABI_WINDOWS_NT
)
8317 fputs ("\t.prologue 0\n", file
);
8318 else if (!flag_inhibit_size_directive
)
8319 fprintf (file
, "\t.prologue %d\n",
8320 alpha_function_needs_gp
|| cfun
->is_thunk
);
8323 /* Write function epilogue. */
8326 alpha_expand_epilogue (void)
8328 /* Registers to save. */
8329 unsigned long imask
= 0;
8330 unsigned long fmask
= 0;
8331 /* Stack space needed for pushing registers clobbered by us. */
8332 HOST_WIDE_INT sa_size
;
8333 /* Complete stack size needed. */
8334 HOST_WIDE_INT frame_size
;
8335 /* Offset from base reg to register save area. */
8336 HOST_WIDE_INT reg_offset
;
8337 int fp_is_frame_pointer
, fp_offset
;
8338 rtx sa_reg
, sa_reg_exp
= NULL
;
8339 rtx sp_adj1
, sp_adj2
, mem
, reg
, insn
;
8341 rtx cfa_restores
= NULL_RTX
;
8344 sa_size
= alpha_sa_size ();
8346 frame_size
= get_frame_size ();
8347 if (TARGET_ABI_OPEN_VMS
)
8348 frame_size
= ALPHA_ROUND (sa_size
8349 + (alpha_procedure_type
== PT_STACK
? 8 : 0)
8351 + crtl
->args
.pretend_args_size
);
8352 else if (TARGET_ABI_UNICOSMK
)
8353 frame_size
= ALPHA_ROUND (sa_size
8354 + (alpha_procedure_type
== PT_STACK
? 48 : 0))
8355 + ALPHA_ROUND (frame_size
8356 + crtl
->outgoing_args_size
);
8358 frame_size
= (ALPHA_ROUND (crtl
->outgoing_args_size
)
8360 + ALPHA_ROUND (frame_size
8361 + crtl
->args
.pretend_args_size
));
8363 if (TARGET_ABI_OPEN_VMS
)
8365 if (alpha_procedure_type
== PT_STACK
)
8366 reg_offset
= 8 + 8 * cfun
->machine
->uses_condition_handler
;
8371 reg_offset
= ALPHA_ROUND (crtl
->outgoing_args_size
);
8373 alpha_sa_mask (&imask
, &fmask
);
8376 = ((TARGET_ABI_OPEN_VMS
&& alpha_procedure_type
== PT_STACK
)
8377 || (!TARGET_ABI_OPEN_VMS
&& frame_pointer_needed
));
8379 sa_reg
= stack_pointer_rtx
;
8381 if (crtl
->calls_eh_return
)
8382 eh_ofs
= EH_RETURN_STACKADJ_RTX
;
8386 if (!TARGET_ABI_UNICOSMK
&& sa_size
)
8388 /* If we have a frame pointer, restore SP from it. */
8389 if ((TARGET_ABI_OPEN_VMS
8390 && vms_unwind_regno
== HARD_FRAME_POINTER_REGNUM
)
8391 || (!TARGET_ABI_OPEN_VMS
&& frame_pointer_needed
))
8392 emit_move_insn (stack_pointer_rtx
, hard_frame_pointer_rtx
);
8394 /* Cope with very large offsets to the register save area. */
8395 if (reg_offset
+ sa_size
> 0x8000)
8397 int low
= ((reg_offset
& 0xffff) ^ 0x8000) - 0x8000;
8400 if (low
+ sa_size
<= 0x8000)
8401 bias
= reg_offset
- low
, reg_offset
= low
;
8403 bias
= reg_offset
, reg_offset
= 0;
8405 sa_reg
= gen_rtx_REG (DImode
, 22);
8406 sa_reg_exp
= plus_constant (stack_pointer_rtx
, bias
);
8408 emit_move_insn (sa_reg
, sa_reg_exp
);
8411 /* Restore registers in order, excepting a true frame pointer. */
8413 mem
= gen_rtx_MEM (DImode
, plus_constant (sa_reg
, reg_offset
));
8415 set_mem_alias_set (mem
, alpha_sr_alias_set
);
8416 reg
= gen_rtx_REG (DImode
, REG_RA
);
8417 emit_move_insn (reg
, mem
);
8418 cfa_restores
= alloc_reg_note (REG_CFA_RESTORE
, reg
, cfa_restores
);
8421 imask
&= ~(1UL << REG_RA
);
8423 for (i
= 0; i
< 31; ++i
)
8424 if (imask
& (1UL << i
))
8426 if (i
== HARD_FRAME_POINTER_REGNUM
&& fp_is_frame_pointer
)
8427 fp_offset
= reg_offset
;
8430 mem
= gen_rtx_MEM (DImode
, plus_constant(sa_reg
, reg_offset
));
8431 set_mem_alias_set (mem
, alpha_sr_alias_set
);
8432 reg
= gen_rtx_REG (DImode
, i
);
8433 emit_move_insn (reg
, mem
);
8434 cfa_restores
= alloc_reg_note (REG_CFA_RESTORE
, reg
,
8440 for (i
= 0; i
< 31; ++i
)
8441 if (fmask
& (1UL << i
))
8443 mem
= gen_rtx_MEM (DFmode
, plus_constant(sa_reg
, reg_offset
));
8444 set_mem_alias_set (mem
, alpha_sr_alias_set
);
8445 reg
= gen_rtx_REG (DFmode
, i
+32);
8446 emit_move_insn (reg
, mem
);
8447 cfa_restores
= alloc_reg_note (REG_CFA_RESTORE
, reg
, cfa_restores
);
8451 else if (TARGET_ABI_UNICOSMK
&& alpha_procedure_type
== PT_STACK
)
8453 /* Restore callee-saved general-purpose registers. */
8457 for (i
= 9; i
< 15; i
++)
8458 if (imask
& (1UL << i
))
8460 mem
= gen_rtx_MEM (DImode
, plus_constant(hard_frame_pointer_rtx
,
8462 set_mem_alias_set (mem
, alpha_sr_alias_set
);
8463 reg
= gen_rtx_REG (DImode
, i
);
8464 emit_move_insn (reg
, mem
);
8465 cfa_restores
= alloc_reg_note (REG_CFA_RESTORE
, reg
, cfa_restores
);
8469 for (i
= 2; i
< 10; i
++)
8470 if (fmask
& (1UL << i
))
8472 mem
= gen_rtx_MEM (DFmode
, plus_constant(hard_frame_pointer_rtx
,
8474 set_mem_alias_set (mem
, alpha_sr_alias_set
);
8475 reg
= gen_rtx_REG (DFmode
, i
+32);
8476 emit_move_insn (reg
, mem
);
8477 cfa_restores
= alloc_reg_note (REG_CFA_RESTORE
, reg
, cfa_restores
);
8481 /* Restore the return address from the DSIB. */
8482 mem
= gen_rtx_MEM (DImode
, plus_constant (hard_frame_pointer_rtx
, -8));
8483 set_mem_alias_set (mem
, alpha_sr_alias_set
);
8484 reg
= gen_rtx_REG (DImode
, REG_RA
);
8485 emit_move_insn (reg
, mem
);
8486 cfa_restores
= alloc_reg_note (REG_CFA_RESTORE
, reg
, cfa_restores
);
8489 if (frame_size
|| eh_ofs
)
8491 sp_adj1
= stack_pointer_rtx
;
8495 sp_adj1
= gen_rtx_REG (DImode
, 23);
8496 emit_move_insn (sp_adj1
,
8497 gen_rtx_PLUS (Pmode
, stack_pointer_rtx
, eh_ofs
));
8500 /* If the stack size is large, begin computation into a temporary
8501 register so as not to interfere with a potential fp restore,
8502 which must be consecutive with an SP restore. */
8503 if (frame_size
< 32768
8504 && ! (TARGET_ABI_UNICOSMK
&& cfun
->calls_alloca
))
8505 sp_adj2
= GEN_INT (frame_size
);
8506 else if (TARGET_ABI_UNICOSMK
)
8508 sp_adj1
= gen_rtx_REG (DImode
, 23);
8509 emit_move_insn (sp_adj1
, hard_frame_pointer_rtx
);
8510 sp_adj2
= const0_rtx
;
8512 else if (frame_size
< 0x40007fffL
)
8514 int low
= ((frame_size
& 0xffff) ^ 0x8000) - 0x8000;
8516 sp_adj2
= plus_constant (sp_adj1
, frame_size
- low
);
8517 if (sa_reg_exp
&& rtx_equal_p (sa_reg_exp
, sp_adj2
))
8521 sp_adj1
= gen_rtx_REG (DImode
, 23);
8522 emit_move_insn (sp_adj1
, sp_adj2
);
8524 sp_adj2
= GEN_INT (low
);
8528 rtx tmp
= gen_rtx_REG (DImode
, 23);
8529 sp_adj2
= alpha_emit_set_const (tmp
, DImode
, frame_size
, 3, false);
8532 /* We can't drop new things to memory this late, afaik,
8533 so build it up by pieces. */
8534 sp_adj2
= alpha_emit_set_long_const (tmp
, frame_size
,
8536 gcc_assert (sp_adj2
);
8540 /* From now on, things must be in order. So emit blockages. */
8542 /* Restore the frame pointer. */
8543 if (TARGET_ABI_UNICOSMK
)
8545 emit_insn (gen_blockage ());
8546 mem
= gen_rtx_MEM (DImode
,
8547 plus_constant (hard_frame_pointer_rtx
, -16));
8548 set_mem_alias_set (mem
, alpha_sr_alias_set
);
8549 emit_move_insn (hard_frame_pointer_rtx
, mem
);
8550 cfa_restores
= alloc_reg_note (REG_CFA_RESTORE
,
8551 hard_frame_pointer_rtx
, cfa_restores
);
8553 else if (fp_is_frame_pointer
)
8555 emit_insn (gen_blockage ());
8556 mem
= gen_rtx_MEM (DImode
, plus_constant (sa_reg
, fp_offset
));
8557 set_mem_alias_set (mem
, alpha_sr_alias_set
);
8558 emit_move_insn (hard_frame_pointer_rtx
, mem
);
8559 cfa_restores
= alloc_reg_note (REG_CFA_RESTORE
,
8560 hard_frame_pointer_rtx
, cfa_restores
);
8562 else if (TARGET_ABI_OPEN_VMS
)
8564 emit_insn (gen_blockage ());
8565 emit_move_insn (hard_frame_pointer_rtx
,
8566 gen_rtx_REG (DImode
, vms_save_fp_regno
));
8567 cfa_restores
= alloc_reg_note (REG_CFA_RESTORE
,
8568 hard_frame_pointer_rtx
, cfa_restores
);
8571 /* Restore the stack pointer. */
8572 emit_insn (gen_blockage ());
8573 if (sp_adj2
== const0_rtx
)
8574 insn
= emit_move_insn (stack_pointer_rtx
, sp_adj1
);
8576 insn
= emit_move_insn (stack_pointer_rtx
,
8577 gen_rtx_PLUS (DImode
, sp_adj1
, sp_adj2
));
8578 REG_NOTES (insn
) = cfa_restores
;
8579 add_reg_note (insn
, REG_CFA_DEF_CFA
, stack_pointer_rtx
);
8580 RTX_FRAME_RELATED_P (insn
) = 1;
8584 gcc_assert (cfa_restores
== NULL
);
8586 if (TARGET_ABI_OPEN_VMS
&& alpha_procedure_type
== PT_REGISTER
)
8588 emit_insn (gen_blockage ());
8589 insn
= emit_move_insn (hard_frame_pointer_rtx
,
8590 gen_rtx_REG (DImode
, vms_save_fp_regno
));
8591 add_reg_note (insn
, REG_CFA_RESTORE
, hard_frame_pointer_rtx
);
8592 RTX_FRAME_RELATED_P (insn
) = 1;
8594 else if (TARGET_ABI_UNICOSMK
&& alpha_procedure_type
!= PT_STACK
)
8596 /* Decrement the frame pointer if the function does not have a
8598 emit_insn (gen_blockage ());
8599 emit_insn (gen_adddi3 (hard_frame_pointer_rtx
,
8600 hard_frame_pointer_rtx
, constm1_rtx
));
8605 /* Output the rest of the textual info surrounding the epilogue. */
8608 alpha_end_function (FILE *file
, const char *fnname
, tree decl ATTRIBUTE_UNUSED
)
8612 /* We output a nop after noreturn calls at the very end of the function to
8613 ensure that the return address always remains in the caller's code range,
8614 as not doing so might confuse unwinding engines. */
8615 insn
= get_last_insn ();
8617 insn
= prev_active_insn (insn
);
8618 if (insn
&& CALL_P (insn
))
8619 output_asm_insn (get_insn_template (CODE_FOR_nop
, NULL
), NULL
);
8621 #if TARGET_ABI_OPEN_VMS
8622 alpha_write_linkage (file
, fnname
, decl
);
8625 /* End the function. */
8626 if (!TARGET_ABI_UNICOSMK
&& !flag_inhibit_size_directive
)
8628 fputs ("\t.end ", file
);
8629 assemble_name (file
, fnname
);
8632 inside_function
= FALSE
;
8634 /* Output jump tables and the static subroutine information block. */
8635 if (TARGET_ABI_UNICOSMK
)
8637 unicosmk_output_ssib (file
, fnname
);
8638 unicosmk_output_deferred_case_vectors (file
);
8642 #if TARGET_ABI_OPEN_VMS
8643 void avms_asm_output_external (FILE *file
, tree decl ATTRIBUTE_UNUSED
, const char *name
)
8645 #ifdef DO_CRTL_NAMES
8652 /* Emit a tail call to FUNCTION after adjusting THIS by DELTA.
8654 In order to avoid the hordes of differences between generated code
8655 with and without TARGET_EXPLICIT_RELOCS, and to avoid duplicating
8656 lots of code loading up large constants, generate rtl and emit it
8657 instead of going straight to text.
8659 Not sure why this idea hasn't been explored before... */
8662 alpha_output_mi_thunk_osf (FILE *file
, tree thunk_fndecl ATTRIBUTE_UNUSED
,
8663 HOST_WIDE_INT delta
, HOST_WIDE_INT vcall_offset
,
8666 HOST_WIDE_INT hi
, lo
;
8667 rtx this_rtx
, insn
, funexp
;
8669 /* We always require a valid GP. */
8670 emit_insn (gen_prologue_ldgp ());
8671 emit_note (NOTE_INSN_PROLOGUE_END
);
8673 /* Find the "this" pointer. If the function returns a structure,
8674 the structure return pointer is in $16. */
8675 if (aggregate_value_p (TREE_TYPE (TREE_TYPE (function
)), function
))
8676 this_rtx
= gen_rtx_REG (Pmode
, 17);
8678 this_rtx
= gen_rtx_REG (Pmode
, 16);
8680 /* Add DELTA. When possible we use ldah+lda. Otherwise load the
8681 entire constant for the add. */
8682 lo
= ((delta
& 0xffff) ^ 0x8000) - 0x8000;
8683 hi
= (((delta
- lo
) & 0xffffffff) ^ 0x80000000) - 0x80000000;
8684 if (hi
+ lo
== delta
)
8687 emit_insn (gen_adddi3 (this_rtx
, this_rtx
, GEN_INT (hi
)));
8689 emit_insn (gen_adddi3 (this_rtx
, this_rtx
, GEN_INT (lo
)));
8693 rtx tmp
= alpha_emit_set_long_const (gen_rtx_REG (Pmode
, 0),
8694 delta
, -(delta
< 0));
8695 emit_insn (gen_adddi3 (this_rtx
, this_rtx
, tmp
));
8698 /* Add a delta stored in the vtable at VCALL_OFFSET. */
8703 tmp
= gen_rtx_REG (Pmode
, 0);
8704 emit_move_insn (tmp
, gen_rtx_MEM (Pmode
, this_rtx
));
8706 lo
= ((vcall_offset
& 0xffff) ^ 0x8000) - 0x8000;
8707 hi
= (((vcall_offset
- lo
) & 0xffffffff) ^ 0x80000000) - 0x80000000;
8708 if (hi
+ lo
== vcall_offset
)
8711 emit_insn (gen_adddi3 (tmp
, tmp
, GEN_INT (hi
)));
8715 tmp2
= alpha_emit_set_long_const (gen_rtx_REG (Pmode
, 1),
8716 vcall_offset
, -(vcall_offset
< 0));
8717 emit_insn (gen_adddi3 (tmp
, tmp
, tmp2
));
8721 tmp2
= gen_rtx_PLUS (Pmode
, tmp
, GEN_INT (lo
));
8724 emit_move_insn (tmp
, gen_rtx_MEM (Pmode
, tmp2
));
8726 emit_insn (gen_adddi3 (this_rtx
, this_rtx
, tmp
));
8729 /* Generate a tail call to the target function. */
8730 if (! TREE_USED (function
))
8732 assemble_external (function
);
8733 TREE_USED (function
) = 1;
8735 funexp
= XEXP (DECL_RTL (function
), 0);
8736 funexp
= gen_rtx_MEM (FUNCTION_MODE
, funexp
);
8737 insn
= emit_call_insn (gen_sibcall (funexp
, const0_rtx
));
8738 SIBLING_CALL_P (insn
) = 1;
8740 /* Run just enough of rest_of_compilation to get the insns emitted.
8741 There's not really enough bulk here to make other passes such as
8742 instruction scheduling worth while. Note that use_thunk calls
8743 assemble_start_function and assemble_end_function. */
8744 insn
= get_insns ();
8745 insn_locators_alloc ();
8746 shorten_branches (insn
);
8747 final_start_function (insn
, file
, 1);
8748 final (insn
, file
, 1);
8749 final_end_function ();
8751 #endif /* TARGET_ABI_OSF */
8753 /* Debugging support. */
8757 /* Count the number of sdb related labels are generated (to find block
8758 start and end boundaries). */
8760 int sdb_label_count
= 0;
8762 /* Name of the file containing the current function. */
8764 static const char *current_function_file
= "";
8766 /* Offsets to alpha virtual arg/local debugging pointers. */
8768 long alpha_arg_offset
;
8769 long alpha_auto_offset
;
8771 /* Emit a new filename to a stream. */
8774 alpha_output_filename (FILE *stream
, const char *name
)
8776 static int first_time
= TRUE
;
8781 ++num_source_filenames
;
8782 current_function_file
= name
;
8783 fprintf (stream
, "\t.file\t%d ", num_source_filenames
);
8784 output_quoted_string (stream
, name
);
8785 fprintf (stream
, "\n");
8786 if (!TARGET_GAS
&& write_symbols
== DBX_DEBUG
)
8787 fprintf (stream
, "\t#@stabs\n");
8790 else if (write_symbols
== DBX_DEBUG
)
8791 /* dbxout.c will emit an appropriate .stabs directive. */
8794 else if (name
!= current_function_file
8795 && strcmp (name
, current_function_file
) != 0)
8797 if (inside_function
&& ! TARGET_GAS
)
8798 fprintf (stream
, "\t#.file\t%d ", num_source_filenames
);
8801 ++num_source_filenames
;
8802 current_function_file
= name
;
8803 fprintf (stream
, "\t.file\t%d ", num_source_filenames
);
8806 output_quoted_string (stream
, name
);
8807 fprintf (stream
, "\n");
8811 /* Structure to show the current status of registers and memory. */
8813 struct shadow_summary
8816 unsigned int i
: 31; /* Mask of int regs */
8817 unsigned int fp
: 31; /* Mask of fp regs */
8818 unsigned int mem
: 1; /* mem == imem | fpmem */
8822 /* Summary the effects of expression X on the machine. Update SUM, a pointer
8823 to the summary structure. SET is nonzero if the insn is setting the
8824 object, otherwise zero. */
8827 summarize_insn (rtx x
, struct shadow_summary
*sum
, int set
)
8829 const char *format_ptr
;
8835 switch (GET_CODE (x
))
8837 /* ??? Note that this case would be incorrect if the Alpha had a
8838 ZERO_EXTRACT in SET_DEST. */
8840 summarize_insn (SET_SRC (x
), sum
, 0);
8841 summarize_insn (SET_DEST (x
), sum
, 1);
8845 summarize_insn (XEXP (x
, 0), sum
, 1);
8849 summarize_insn (XEXP (x
, 0), sum
, 0);
8853 for (i
= ASM_OPERANDS_INPUT_LENGTH (x
) - 1; i
>= 0; i
--)
8854 summarize_insn (ASM_OPERANDS_INPUT (x
, i
), sum
, 0);
8858 for (i
= XVECLEN (x
, 0) - 1; i
>= 0; i
--)
8859 summarize_insn (XVECEXP (x
, 0, i
), sum
, 0);
8863 summarize_insn (SUBREG_REG (x
), sum
, 0);
8868 int regno
= REGNO (x
);
8869 unsigned long mask
= ((unsigned long) 1) << (regno
% 32);
8871 if (regno
== 31 || regno
== 63)
8877 sum
->defd
.i
|= mask
;
8879 sum
->defd
.fp
|= mask
;
8884 sum
->used
.i
|= mask
;
8886 sum
->used
.fp
|= mask
;
8897 /* Find the regs used in memory address computation: */
8898 summarize_insn (XEXP (x
, 0), sum
, 0);
8901 case CONST_INT
: case CONST_DOUBLE
:
8902 case SYMBOL_REF
: case LABEL_REF
: case CONST
:
8903 case SCRATCH
: case ASM_INPUT
:
8906 /* Handle common unary and binary ops for efficiency. */
8907 case COMPARE
: case PLUS
: case MINUS
: case MULT
: case DIV
:
8908 case MOD
: case UDIV
: case UMOD
: case AND
: case IOR
:
8909 case XOR
: case ASHIFT
: case ROTATE
: case ASHIFTRT
: case LSHIFTRT
:
8910 case ROTATERT
: case SMIN
: case SMAX
: case UMIN
: case UMAX
:
8911 case NE
: case EQ
: case GE
: case GT
: case LE
:
8912 case LT
: case GEU
: case GTU
: case LEU
: case LTU
:
8913 summarize_insn (XEXP (x
, 0), sum
, 0);
8914 summarize_insn (XEXP (x
, 1), sum
, 0);
8917 case NEG
: case NOT
: case SIGN_EXTEND
: case ZERO_EXTEND
:
8918 case TRUNCATE
: case FLOAT_EXTEND
: case FLOAT_TRUNCATE
: case FLOAT
:
8919 case FIX
: case UNSIGNED_FLOAT
: case UNSIGNED_FIX
: case ABS
:
8920 case SQRT
: case FFS
:
8921 summarize_insn (XEXP (x
, 0), sum
, 0);
8925 format_ptr
= GET_RTX_FORMAT (GET_CODE (x
));
8926 for (i
= GET_RTX_LENGTH (GET_CODE (x
)) - 1; i
>= 0; i
--)
8927 switch (format_ptr
[i
])
8930 summarize_insn (XEXP (x
, i
), sum
, 0);
8934 for (j
= XVECLEN (x
, i
) - 1; j
>= 0; j
--)
8935 summarize_insn (XVECEXP (x
, i
, j
), sum
, 0);
8947 /* Ensure a sufficient number of `trapb' insns are in the code when
8948 the user requests code with a trap precision of functions or
8951 In naive mode, when the user requests a trap-precision of
8952 "instruction", a trapb is needed after every instruction that may
8953 generate a trap. This ensures that the code is resumption safe but
8956 When optimizations are turned on, we delay issuing a trapb as long
8957 as possible. In this context, a trap shadow is the sequence of
8958 instructions that starts with a (potentially) trap generating
8959 instruction and extends to the next trapb or call_pal instruction
8960 (but GCC never generates call_pal by itself). We can delay (and
8961 therefore sometimes omit) a trapb subject to the following
8964 (a) On entry to the trap shadow, if any Alpha register or memory
8965 location contains a value that is used as an operand value by some
8966 instruction in the trap shadow (live on entry), then no instruction
8967 in the trap shadow may modify the register or memory location.
8969 (b) Within the trap shadow, the computation of the base register
8970 for a memory load or store instruction may not involve using the
8971 result of an instruction that might generate an UNPREDICTABLE
8974 (c) Within the trap shadow, no register may be used more than once
8975 as a destination register. (This is to make life easier for the
8978 (d) The trap shadow may not include any branch instructions. */
8981 alpha_handle_trap_shadows (void)
8983 struct shadow_summary shadow
;
8984 int trap_pending
, exception_nesting
;
8988 exception_nesting
= 0;
8991 shadow
.used
.mem
= 0;
8992 shadow
.defd
= shadow
.used
;
8994 for (i
= get_insns (); i
; i
= NEXT_INSN (i
))
8998 switch (NOTE_KIND (i
))
9000 case NOTE_INSN_EH_REGION_BEG
:
9001 exception_nesting
++;
9006 case NOTE_INSN_EH_REGION_END
:
9007 exception_nesting
--;
9012 case NOTE_INSN_EPILOGUE_BEG
:
9013 if (trap_pending
&& alpha_tp
>= ALPHA_TP_FUNC
)
9018 else if (trap_pending
)
9020 if (alpha_tp
== ALPHA_TP_FUNC
)
9023 && GET_CODE (PATTERN (i
)) == RETURN
)
9026 else if (alpha_tp
== ALPHA_TP_INSN
)
9030 struct shadow_summary sum
;
9035 sum
.defd
= sum
.used
;
9037 switch (GET_CODE (i
))
9040 /* Annoyingly, get_attr_trap will die on these. */
9041 if (GET_CODE (PATTERN (i
)) == USE
9042 || GET_CODE (PATTERN (i
)) == CLOBBER
)
9045 summarize_insn (PATTERN (i
), &sum
, 0);
9047 if ((sum
.defd
.i
& shadow
.defd
.i
)
9048 || (sum
.defd
.fp
& shadow
.defd
.fp
))
9050 /* (c) would be violated */
9054 /* Combine shadow with summary of current insn: */
9055 shadow
.used
.i
|= sum
.used
.i
;
9056 shadow
.used
.fp
|= sum
.used
.fp
;
9057 shadow
.used
.mem
|= sum
.used
.mem
;
9058 shadow
.defd
.i
|= sum
.defd
.i
;
9059 shadow
.defd
.fp
|= sum
.defd
.fp
;
9060 shadow
.defd
.mem
|= sum
.defd
.mem
;
9062 if ((sum
.defd
.i
& shadow
.used
.i
)
9063 || (sum
.defd
.fp
& shadow
.used
.fp
)
9064 || (sum
.defd
.mem
& shadow
.used
.mem
))
9066 /* (a) would be violated (also takes care of (b)) */
9067 gcc_assert (get_attr_trap (i
) != TRAP_YES
9068 || (!(sum
.defd
.i
& sum
.used
.i
)
9069 && !(sum
.defd
.fp
& sum
.used
.fp
)));
9087 n
= emit_insn_before (gen_trapb (), i
);
9088 PUT_MODE (n
, TImode
);
9089 PUT_MODE (i
, TImode
);
9093 shadow
.used
.mem
= 0;
9094 shadow
.defd
= shadow
.used
;
9099 if ((exception_nesting
> 0 || alpha_tp
>= ALPHA_TP_FUNC
)
9100 && NONJUMP_INSN_P (i
)
9101 && GET_CODE (PATTERN (i
)) != USE
9102 && GET_CODE (PATTERN (i
)) != CLOBBER
9103 && get_attr_trap (i
) == TRAP_YES
)
9105 if (optimize
&& !trap_pending
)
9106 summarize_insn (PATTERN (i
), &shadow
, 0);
9112 /* Alpha can only issue instruction groups simultaneously if they are
9113 suitably aligned. This is very processor-specific. */
9114 /* There are a number of entries in alphaev4_insn_pipe and alphaev5_insn_pipe
9115 that are marked "fake". These instructions do not exist on that target,
9116 but it is possible to see these insns with deranged combinations of
9117 command-line options, such as "-mtune=ev4 -mmax". Instead of aborting,
9118 choose a result at random. */
9120 enum alphaev4_pipe
{
9127 enum alphaev5_pipe
{
9138 static enum alphaev4_pipe
9139 alphaev4_insn_pipe (rtx insn
)
9141 if (recog_memoized (insn
) < 0)
9143 if (get_attr_length (insn
) != 4)
9146 switch (get_attr_type (insn
))
9162 case TYPE_MVI
: /* fake */
9177 case TYPE_FSQRT
: /* fake */
9178 case TYPE_FTOI
: /* fake */
9179 case TYPE_ITOF
: /* fake */
9187 static enum alphaev5_pipe
9188 alphaev5_insn_pipe (rtx insn
)
9190 if (recog_memoized (insn
) < 0)
9192 if (get_attr_length (insn
) != 4)
9195 switch (get_attr_type (insn
))
9215 case TYPE_FTOI
: /* fake */
9216 case TYPE_ITOF
: /* fake */
9231 case TYPE_FSQRT
: /* fake */
9242 /* IN_USE is a mask of the slots currently filled within the insn group.
9243 The mask bits come from alphaev4_pipe above. If EV4_IBX is set, then
9244 the insn in EV4_IB0 can be swapped by the hardware into EV4_IB1.
9246 LEN is, of course, the length of the group in bytes. */
9249 alphaev4_next_group (rtx insn
, int *pin_use
, int *plen
)
9256 || GET_CODE (PATTERN (insn
)) == CLOBBER
9257 || GET_CODE (PATTERN (insn
)) == USE
)
9262 enum alphaev4_pipe pipe
;
9264 pipe
= alphaev4_insn_pipe (insn
);
9268 /* Force complex instructions to start new groups. */
9272 /* If this is a completely unrecognized insn, it's an asm.
9273 We don't know how long it is, so record length as -1 to
9274 signal a needed realignment. */
9275 if (recog_memoized (insn
) < 0)
9278 len
= get_attr_length (insn
);
9282 if (in_use
& EV4_IB0
)
9284 if (in_use
& EV4_IB1
)
9289 in_use
|= EV4_IB0
| EV4_IBX
;
9293 if (in_use
& EV4_IB0
)
9295 if (!(in_use
& EV4_IBX
) || (in_use
& EV4_IB1
))
9303 if (in_use
& EV4_IB1
)
9313 /* Haifa doesn't do well scheduling branches. */
9318 insn
= next_nonnote_insn (insn
);
9320 if (!insn
|| ! INSN_P (insn
))
9323 /* Let Haifa tell us where it thinks insn group boundaries are. */
9324 if (GET_MODE (insn
) == TImode
)
9327 if (GET_CODE (insn
) == CLOBBER
|| GET_CODE (insn
) == USE
)
9332 insn
= next_nonnote_insn (insn
);
9340 /* IN_USE is a mask of the slots currently filled within the insn group.
9341 The mask bits come from alphaev5_pipe above. If EV5_E01 is set, then
9342 the insn in EV5_E0 can be swapped by the hardware into EV5_E1.
9344 LEN is, of course, the length of the group in bytes. */
9347 alphaev5_next_group (rtx insn
, int *pin_use
, int *plen
)
9354 || GET_CODE (PATTERN (insn
)) == CLOBBER
9355 || GET_CODE (PATTERN (insn
)) == USE
)
9360 enum alphaev5_pipe pipe
;
9362 pipe
= alphaev5_insn_pipe (insn
);
9366 /* Force complex instructions to start new groups. */
9370 /* If this is a completely unrecognized insn, it's an asm.
9371 We don't know how long it is, so record length as -1 to
9372 signal a needed realignment. */
9373 if (recog_memoized (insn
) < 0)
9376 len
= get_attr_length (insn
);
9379 /* ??? Most of the places below, we would like to assert never
9380 happen, as it would indicate an error either in Haifa, or
9381 in the scheduling description. Unfortunately, Haifa never
9382 schedules the last instruction of the BB, so we don't have
9383 an accurate TI bit to go off. */
9385 if (in_use
& EV5_E0
)
9387 if (in_use
& EV5_E1
)
9392 in_use
|= EV5_E0
| EV5_E01
;
9396 if (in_use
& EV5_E0
)
9398 if (!(in_use
& EV5_E01
) || (in_use
& EV5_E1
))
9406 if (in_use
& EV5_E1
)
9412 if (in_use
& EV5_FA
)
9414 if (in_use
& EV5_FM
)
9419 in_use
|= EV5_FA
| EV5_FAM
;
9423 if (in_use
& EV5_FA
)
9429 if (in_use
& EV5_FM
)
9442 /* Haifa doesn't do well scheduling branches. */
9443 /* ??? If this is predicted not-taken, slotting continues, except
9444 that no more IBR, FBR, or JSR insns may be slotted. */
9449 insn
= next_nonnote_insn (insn
);
9451 if (!insn
|| ! INSN_P (insn
))
9454 /* Let Haifa tell us where it thinks insn group boundaries are. */
9455 if (GET_MODE (insn
) == TImode
)
9458 if (GET_CODE (insn
) == CLOBBER
|| GET_CODE (insn
) == USE
)
9463 insn
= next_nonnote_insn (insn
);
9472 alphaev4_next_nop (int *pin_use
)
9474 int in_use
= *pin_use
;
9477 if (!(in_use
& EV4_IB0
))
9482 else if ((in_use
& (EV4_IBX
|EV4_IB1
)) == EV4_IBX
)
9487 else if (TARGET_FP
&& !(in_use
& EV4_IB1
))
9500 alphaev5_next_nop (int *pin_use
)
9502 int in_use
= *pin_use
;
9505 if (!(in_use
& EV5_E1
))
9510 else if (TARGET_FP
&& !(in_use
& EV5_FA
))
9515 else if (TARGET_FP
&& !(in_use
& EV5_FM
))
9527 /* The instruction group alignment main loop. */
9530 alpha_align_insns (unsigned int max_align
,
9531 rtx (*next_group
) (rtx
, int *, int *),
9532 rtx (*next_nop
) (int *))
9534 /* ALIGN is the known alignment for the insn group. */
9536 /* OFS is the offset of the current insn in the insn group. */
9538 int prev_in_use
, in_use
, len
, ldgp
;
9541 /* Let shorten branches care for assigning alignments to code labels. */
9542 shorten_branches (get_insns ());
9544 if (align_functions
< 4)
9546 else if ((unsigned int) align_functions
< max_align
)
9547 align
= align_functions
;
9551 ofs
= prev_in_use
= 0;
9554 i
= next_nonnote_insn (i
);
9556 ldgp
= alpha_function_needs_gp
? 8 : 0;
9560 next
= (*next_group
) (i
, &in_use
, &len
);
9562 /* When we see a label, resync alignment etc. */
9565 unsigned int new_align
= 1 << label_to_alignment (i
);
9567 if (new_align
>= align
)
9569 align
= new_align
< max_align
? new_align
: max_align
;
9573 else if (ofs
& (new_align
-1))
9574 ofs
= (ofs
| (new_align
-1)) + 1;
9578 /* Handle complex instructions special. */
9579 else if (in_use
== 0)
9581 /* Asms will have length < 0. This is a signal that we have
9582 lost alignment knowledge. Assume, however, that the asm
9583 will not mis-align instructions. */
9592 /* If the known alignment is smaller than the recognized insn group,
9593 realign the output. */
9594 else if ((int) align
< len
)
9596 unsigned int new_log_align
= len
> 8 ? 4 : 3;
9599 where
= prev
= prev_nonnote_insn (i
);
9600 if (!where
|| !LABEL_P (where
))
9603 /* Can't realign between a call and its gp reload. */
9604 if (! (TARGET_EXPLICIT_RELOCS
9605 && prev
&& CALL_P (prev
)))
9607 emit_insn_before (gen_realign (GEN_INT (new_log_align
)), where
);
9608 align
= 1 << new_log_align
;
9613 /* We may not insert padding inside the initial ldgp sequence. */
9617 /* If the group won't fit in the same INT16 as the previous,
9618 we need to add padding to keep the group together. Rather
9619 than simply leaving the insn filling to the assembler, we
9620 can make use of the knowledge of what sorts of instructions
9621 were issued in the previous group to make sure that all of
9622 the added nops are really free. */
9623 else if (ofs
+ len
> (int) align
)
9625 int nop_count
= (align
- ofs
) / 4;
9628 /* Insert nops before labels, branches, and calls to truly merge
9629 the execution of the nops with the previous instruction group. */
9630 where
= prev_nonnote_insn (i
);
9633 if (LABEL_P (where
))
9635 rtx where2
= prev_nonnote_insn (where
);
9636 if (where2
&& JUMP_P (where2
))
9639 else if (NONJUMP_INSN_P (where
))
9646 emit_insn_before ((*next_nop
)(&prev_in_use
), where
);
9647 while (--nop_count
);
9651 ofs
= (ofs
+ len
) & (align
- 1);
9652 prev_in_use
= in_use
;
9657 /* Insert an unop between a noreturn function call and GP load. */
9660 alpha_pad_noreturn (void)
9664 for (insn
= get_insns (); insn
; insn
= NEXT_INSN (insn
))
9667 || !find_reg_note (insn
, REG_NORETURN
, NULL_RTX
))
9670 next
= next_active_insn (insn
);
9674 rtx pat
= PATTERN (next
);
9676 if (GET_CODE (pat
) == SET
9677 && GET_CODE (SET_SRC (pat
)) == UNSPEC_VOLATILE
9678 && XINT (SET_SRC (pat
), 1) == UNSPECV_LDGP1
)
9679 emit_insn_after (gen_unop (), insn
);
9684 /* Machine dependent reorg pass. */
9689 /* Workaround for a linker error that triggers when an
9690 exception handler immediatelly follows a noreturn function.
9692 The instruction stream from an object file:
9694 54: 00 40 5b 6b jsr ra,(t12),58 <__func+0x58>
9695 58: 00 00 ba 27 ldah gp,0(ra)
9696 5c: 00 00 bd 23 lda gp,0(gp)
9697 60: 00 00 7d a7 ldq t12,0(gp)
9698 64: 00 40 5b 6b jsr ra,(t12),68 <__func+0x68>
9700 was converted in the final link pass to:
9702 fdb24: a0 03 40 d3 bsr ra,fe9a8 <_called_func+0x8>
9703 fdb28: 00 00 fe 2f unop
9704 fdb2c: 00 00 fe 2f unop
9705 fdb30: 30 82 7d a7 ldq t12,-32208(gp)
9706 fdb34: 00 40 5b 6b jsr ra,(t12),fdb38 <__func+0x68>
9708 GP load instructions were wrongly cleared by the linker relaxation
9709 pass. This workaround prevents removal of GP loads by inserting
9710 an unop instruction between a noreturn function call and
9711 exception handler prologue. */
9713 if (current_function_has_exception_handlers ())
9714 alpha_pad_noreturn ();
9716 if (alpha_tp
!= ALPHA_TP_PROG
|| flag_exceptions
)
9717 alpha_handle_trap_shadows ();
9719 /* Due to the number of extra trapb insns, don't bother fixing up
9720 alignment when trap precision is instruction. Moreover, we can
9721 only do our job when sched2 is run. */
9722 if (optimize
&& !optimize_size
9723 && alpha_tp
!= ALPHA_TP_INSN
9724 && flag_schedule_insns_after_reload
)
9726 if (alpha_tune
== PROCESSOR_EV4
)
9727 alpha_align_insns (8, alphaev4_next_group
, alphaev4_next_nop
);
9728 else if (alpha_tune
== PROCESSOR_EV5
)
9729 alpha_align_insns (16, alphaev5_next_group
, alphaev5_next_nop
);
9733 #if !TARGET_ABI_UNICOSMK
9740 alpha_file_start (void)
9742 #ifdef OBJECT_FORMAT_ELF
9743 /* If emitting dwarf2 debug information, we cannot generate a .file
9744 directive to start the file, as it will conflict with dwarf2out
9745 file numbers. So it's only useful when emitting mdebug output. */
9746 targetm
.file_start_file_directive
= (write_symbols
== DBX_DEBUG
);
9749 default_file_start ();
9751 fprintf (asm_out_file
, "\t.verstamp %d %d\n", MS_STAMP
, LS_STAMP
);
9754 fputs ("\t.set noreorder\n", asm_out_file
);
9755 fputs ("\t.set volatile\n", asm_out_file
);
9756 if (!TARGET_ABI_OPEN_VMS
)
9757 fputs ("\t.set noat\n", asm_out_file
);
9758 if (TARGET_EXPLICIT_RELOCS
)
9759 fputs ("\t.set nomacro\n", asm_out_file
);
9760 if (TARGET_SUPPORT_ARCH
| TARGET_BWX
| TARGET_MAX
| TARGET_FIX
| TARGET_CIX
)
9764 if (alpha_cpu
== PROCESSOR_EV6
|| TARGET_FIX
|| TARGET_CIX
)
9766 else if (TARGET_MAX
)
9768 else if (TARGET_BWX
)
9770 else if (alpha_cpu
== PROCESSOR_EV5
)
9775 fprintf (asm_out_file
, "\t.arch %s\n", arch
);
9780 #ifdef OBJECT_FORMAT_ELF
9781 /* Since we don't have a .dynbss section, we should not allow global
9782 relocations in the .rodata section. */
9785 alpha_elf_reloc_rw_mask (void)
9787 return flag_pic
? 3 : 2;
9790 /* Return a section for X. The only special thing we do here is to
9791 honor small data. */
9794 alpha_elf_select_rtx_section (enum machine_mode mode
, rtx x
,
9795 unsigned HOST_WIDE_INT align
)
9797 if (TARGET_SMALL_DATA
&& GET_MODE_SIZE (mode
) <= g_switch_value
)
9798 /* ??? Consider using mergeable sdata sections. */
9799 return sdata_section
;
9801 return default_elf_select_rtx_section (mode
, x
, align
);
9805 alpha_elf_section_type_flags (tree decl
, const char *name
, int reloc
)
9807 unsigned int flags
= 0;
9809 if (strcmp (name
, ".sdata") == 0
9810 || strncmp (name
, ".sdata.", 7) == 0
9811 || strncmp (name
, ".gnu.linkonce.s.", 16) == 0
9812 || strcmp (name
, ".sbss") == 0
9813 || strncmp (name
, ".sbss.", 6) == 0
9814 || strncmp (name
, ".gnu.linkonce.sb.", 17) == 0)
9815 flags
= SECTION_SMALL
;
9817 flags
|= default_section_type_flags (decl
, name
, reloc
);
9820 #endif /* OBJECT_FORMAT_ELF */
9822 /* Structure to collect function names for final output in link section. */
9823 /* Note that items marked with GTY can't be ifdef'ed out. */
9825 enum links_kind
{KIND_UNUSED
, KIND_LOCAL
, KIND_EXTERN
};
9826 enum reloc_kind
{KIND_LINKAGE
, KIND_CODEADDR
};
9828 struct GTY(()) alpha_links
9833 enum links_kind lkind
;
9834 enum reloc_kind rkind
;
9837 struct GTY(()) alpha_funcs
9840 splay_tree
GTY ((param1_is (char *), param2_is (struct alpha_links
*)))
9844 static GTY ((param1_is (char *), param2_is (struct alpha_links
*)))
9845 splay_tree alpha_links_tree
;
9846 static GTY ((param1_is (tree
), param2_is (struct alpha_funcs
*)))
9847 splay_tree alpha_funcs_tree
;
9849 static GTY(()) int alpha_funcs_num
;
9851 #if TARGET_ABI_OPEN_VMS
9853 /* Return the VMS argument type corresponding to MODE. */
9856 alpha_arg_type (enum machine_mode mode
)
9861 return TARGET_FLOAT_VAX
? FF
: FS
;
9863 return TARGET_FLOAT_VAX
? FD
: FT
;
9869 /* Return an rtx for an integer representing the VMS Argument Information
9873 alpha_arg_info_reg_val (CUMULATIVE_ARGS cum
)
9875 unsigned HOST_WIDE_INT regval
= cum
.num_args
;
9878 for (i
= 0; i
< 6; i
++)
9879 regval
|= ((int) cum
.atypes
[i
]) << (i
* 3 + 8);
9881 return GEN_INT (regval
);
9884 /* Register the need for a (fake) .linkage entry for calls to function NAME.
9885 IS_LOCAL is 1 if this is for a definition, 0 if this is for a real call.
9886 Return a SYMBOL_REF suited to the call instruction. */
9889 alpha_need_linkage (const char *name
, int is_local
)
9891 splay_tree_node node
;
9892 struct alpha_links
*al
;
9901 struct alpha_funcs
*cfaf
;
9903 if (!alpha_funcs_tree
)
9904 alpha_funcs_tree
= splay_tree_new_ggc ((splay_tree_compare_fn
)
9905 splay_tree_compare_pointers
);
9907 cfaf
= (struct alpha_funcs
*) ggc_alloc (sizeof (struct alpha_funcs
));
9910 cfaf
->num
= ++alpha_funcs_num
;
9912 splay_tree_insert (alpha_funcs_tree
,
9913 (splay_tree_key
) current_function_decl
,
9914 (splay_tree_value
) cfaf
);
9917 if (alpha_links_tree
)
9919 /* Is this name already defined? */
9921 node
= splay_tree_lookup (alpha_links_tree
, (splay_tree_key
) name
);
9924 al
= (struct alpha_links
*) node
->value
;
9927 /* Defined here but external assumed. */
9928 if (al
->lkind
== KIND_EXTERN
)
9929 al
->lkind
= KIND_LOCAL
;
9933 /* Used here but unused assumed. */
9934 if (al
->lkind
== KIND_UNUSED
)
9935 al
->lkind
= KIND_LOCAL
;
9941 alpha_links_tree
= splay_tree_new_ggc ((splay_tree_compare_fn
) strcmp
);
9943 al
= (struct alpha_links
*) ggc_alloc (sizeof (struct alpha_links
));
9944 name
= ggc_strdup (name
);
9946 /* Assume external if no definition. */
9947 al
->lkind
= (is_local
? KIND_UNUSED
: KIND_EXTERN
);
9949 /* Ensure we have an IDENTIFIER so assemble_name can mark it used
9950 and find the ultimate alias target like assemble_name. */
9951 id
= get_identifier (name
);
9953 while (IDENTIFIER_TRANSPARENT_ALIAS (id
))
9955 id
= TREE_CHAIN (id
);
9956 target
= IDENTIFIER_POINTER (id
);
9959 al
->target
= target
? target
: name
;
9960 al
->linkage
= gen_rtx_SYMBOL_REF (Pmode
, name
);
9962 splay_tree_insert (alpha_links_tree
, (splay_tree_key
) name
,
9963 (splay_tree_value
) al
);
9968 /* Return a SYMBOL_REF representing the reference to the .linkage entry
9969 of function FUNC built for calls made from CFUNDECL. LFLAG is 1 if
9970 this is the reference to the linkage pointer value, 0 if this is the
9971 reference to the function entry value. RFLAG is 1 if this a reduced
9972 reference (code address only), 0 if this is a full reference. */
9975 alpha_use_linkage (rtx func
, tree cfundecl
, int lflag
, int rflag
)
9977 splay_tree_node cfunnode
;
9978 struct alpha_funcs
*cfaf
;
9979 struct alpha_links
*al
;
9980 const char *name
= XSTR (func
, 0);
9982 cfaf
= (struct alpha_funcs
*) 0;
9983 al
= (struct alpha_links
*) 0;
9985 cfunnode
= splay_tree_lookup (alpha_funcs_tree
, (splay_tree_key
) cfundecl
);
9986 cfaf
= (struct alpha_funcs
*) cfunnode
->value
;
9990 splay_tree_node lnode
;
9992 /* Is this name already defined? */
9994 lnode
= splay_tree_lookup (cfaf
->links
, (splay_tree_key
) name
);
9996 al
= (struct alpha_links
*) lnode
->value
;
9999 cfaf
->links
= splay_tree_new_ggc ((splay_tree_compare_fn
) strcmp
);
10006 splay_tree_node node
= 0;
10007 struct alpha_links
*anl
;
10009 if (name
[0] == '*')
10012 name_len
= strlen (name
);
10013 linksym
= (char *) alloca (name_len
+ 50);
10015 al
= (struct alpha_links
*) ggc_alloc (sizeof (struct alpha_links
));
10016 al
->num
= cfaf
->num
;
10018 node
= splay_tree_lookup (alpha_links_tree
, (splay_tree_key
) name
);
10021 anl
= (struct alpha_links
*) node
->value
;
10022 al
->lkind
= anl
->lkind
;
10023 name
= anl
->target
;
10026 sprintf (linksym
, "$%d..%s..lk", cfaf
->num
, name
);
10027 buflen
= strlen (linksym
);
10029 al
->linkage
= gen_rtx_SYMBOL_REF
10030 (Pmode
, ggc_alloc_string (linksym
, buflen
+ 1));
10032 splay_tree_insert (cfaf
->links
, (splay_tree_key
) name
,
10033 (splay_tree_value
) al
);
10037 al
->rkind
= KIND_CODEADDR
;
10039 al
->rkind
= KIND_LINKAGE
;
10042 return gen_rtx_MEM (Pmode
, plus_constant (al
->linkage
, 8));
10044 return al
->linkage
;
10048 alpha_write_one_linkage (splay_tree_node node
, void *data
)
10050 const char *const name
= (const char *) node
->key
;
10051 struct alpha_links
*link
= (struct alpha_links
*) node
->value
;
10052 FILE *stream
= (FILE *) data
;
10054 fprintf (stream
, "$%d..%s..lk:\n", link
->num
, name
);
10055 if (link
->rkind
== KIND_CODEADDR
)
10057 if (link
->lkind
== KIND_LOCAL
)
10059 /* Local and used */
10060 fprintf (stream
, "\t.quad %s..en\n", name
);
10064 /* External and used, request code address. */
10065 fprintf (stream
, "\t.code_address %s\n", name
);
10070 if (link
->lkind
== KIND_LOCAL
)
10072 /* Local and used, build linkage pair. */
10073 fprintf (stream
, "\t.quad %s..en\n", name
);
10074 fprintf (stream
, "\t.quad %s\n", name
);
10078 /* External and used, request linkage pair. */
10079 fprintf (stream
, "\t.linkage %s\n", name
);
10087 alpha_write_linkage (FILE *stream
, const char *funname
, tree fundecl
)
10089 splay_tree_node node
;
10090 struct alpha_funcs
*func
;
10092 fprintf (stream
, "\t.link\n");
10093 fprintf (stream
, "\t.align 3\n");
10096 node
= splay_tree_lookup (alpha_funcs_tree
, (splay_tree_key
) fundecl
);
10097 func
= (struct alpha_funcs
*) node
->value
;
10099 fputs ("\t.name ", stream
);
10100 assemble_name (stream
, funname
);
10101 fputs ("..na\n", stream
);
10102 ASM_OUTPUT_LABEL (stream
, funname
);
10103 fprintf (stream
, "\t.pdesc ");
10104 assemble_name (stream
, funname
);
10105 fprintf (stream
, "..en,%s\n",
10106 alpha_procedure_type
== PT_STACK
? "stack"
10107 : alpha_procedure_type
== PT_REGISTER
? "reg" : "null");
10111 splay_tree_foreach (func
->links
, alpha_write_one_linkage
, stream
);
10112 /* splay_tree_delete (func->links); */
10116 /* Switch to an arbitrary section NAME with attributes as specified
10117 by FLAGS. ALIGN specifies any known alignment requirements for
10118 the section; 0 if the default should be used. */
10121 vms_asm_named_section (const char *name
, unsigned int flags
,
10122 tree decl ATTRIBUTE_UNUSED
)
10124 fputc ('\n', asm_out_file
);
10125 fprintf (asm_out_file
, ".section\t%s", name
);
10127 if (flags
& SECTION_DEBUG
)
10128 fprintf (asm_out_file
, ",NOWRT");
10130 fputc ('\n', asm_out_file
);
10133 /* Record an element in the table of global constructors. SYMBOL is
10134 a SYMBOL_REF of the function to be called; PRIORITY is a number
10135 between 0 and MAX_INIT_PRIORITY.
10137 Differs from default_ctors_section_asm_out_constructor in that the
10138 width of the .ctors entry is always 64 bits, rather than the 32 bits
10139 used by a normal pointer. */
10142 vms_asm_out_constructor (rtx symbol
, int priority ATTRIBUTE_UNUSED
)
10144 switch_to_section (ctors_section
);
10145 assemble_align (BITS_PER_WORD
);
10146 assemble_integer (symbol
, UNITS_PER_WORD
, BITS_PER_WORD
, 1);
10150 vms_asm_out_destructor (rtx symbol
, int priority ATTRIBUTE_UNUSED
)
10152 switch_to_section (dtors_section
);
10153 assemble_align (BITS_PER_WORD
);
10154 assemble_integer (symbol
, UNITS_PER_WORD
, BITS_PER_WORD
, 1);
10159 alpha_need_linkage (const char *name ATTRIBUTE_UNUSED
,
10160 int is_local ATTRIBUTE_UNUSED
)
10166 alpha_use_linkage (rtx func ATTRIBUTE_UNUSED
,
10167 tree cfundecl ATTRIBUTE_UNUSED
,
10168 int lflag ATTRIBUTE_UNUSED
,
10169 int rflag ATTRIBUTE_UNUSED
)
10174 #endif /* TARGET_ABI_OPEN_VMS */
10176 #if TARGET_ABI_UNICOSMK
10178 /* This evaluates to true if we do not know how to pass TYPE solely in
10179 registers. This is the case for all arguments that do not fit in two
10183 unicosmk_must_pass_in_stack (enum machine_mode mode
, const_tree type
)
10188 if (TREE_CODE (TYPE_SIZE (type
)) != INTEGER_CST
)
10190 if (TREE_ADDRESSABLE (type
))
10193 return ALPHA_ARG_SIZE (mode
, type
, 0) > 2;
10196 /* Define the offset between two registers, one to be eliminated, and the
10197 other its replacement, at the start of a routine. */
10200 unicosmk_initial_elimination_offset (int from
, int to
)
10204 fixed_size
= alpha_sa_size();
10205 if (fixed_size
!= 0)
10208 if (from
== FRAME_POINTER_REGNUM
&& to
== HARD_FRAME_POINTER_REGNUM
)
10209 return -fixed_size
;
10210 else if (from
== ARG_POINTER_REGNUM
&& to
== HARD_FRAME_POINTER_REGNUM
)
10212 else if (from
== FRAME_POINTER_REGNUM
&& to
== STACK_POINTER_REGNUM
)
10213 return (ALPHA_ROUND (crtl
->outgoing_args_size
)
10214 + ALPHA_ROUND (get_frame_size()));
10215 else if (from
== ARG_POINTER_REGNUM
&& to
== STACK_POINTER_REGNUM
)
10216 return (ALPHA_ROUND (fixed_size
)
10217 + ALPHA_ROUND (get_frame_size()
10218 + crtl
->outgoing_args_size
));
10220 gcc_unreachable ();
10223 /* Output the module name for .ident and .end directives. We have to strip
10224 directories and add make sure that the module name starts with a letter
10228 unicosmk_output_module_name (FILE *file
)
10230 const char *name
= lbasename (main_input_filename
);
10231 unsigned len
= strlen (name
);
10232 char *clean_name
= alloca (len
+ 2);
10233 char *ptr
= clean_name
;
10235 /* CAM only accepts module names that start with a letter or '$'. We
10236 prefix the module name with a '$' if necessary. */
10238 if (!ISALPHA (*name
))
10240 memcpy (ptr
, name
, len
+ 1);
10241 clean_symbol_name (clean_name
);
10242 fputs (clean_name
, file
);
10245 /* Output the definition of a common variable. */
10248 unicosmk_output_common (FILE *file
, const char *name
, int size
, int align
)
10251 printf ("T3E__: common %s\n", name
);
10254 fputs("\t.endp\n\n\t.psect ", file
);
10255 assemble_name(file
, name
);
10256 fprintf(file
, ",%d,common\n", floor_log2 (align
/ BITS_PER_UNIT
));
10257 fprintf(file
, "\t.byte\t0:%d\n", size
);
10259 /* Mark the symbol as defined in this module. */
10260 name_tree
= get_identifier (name
);
10261 TREE_ASM_WRITTEN (name_tree
) = 1;
10264 #define SECTION_PUBLIC SECTION_MACH_DEP
10265 #define SECTION_MAIN (SECTION_PUBLIC << 1)
10266 static int current_section_align
;
10268 /* A get_unnamed_section callback for switching to the text section. */
10271 unicosmk_output_text_section_asm_op (const void *data ATTRIBUTE_UNUSED
)
10273 static int count
= 0;
10274 fprintf (asm_out_file
, "\t.endp\n\n\t.psect\tgcc@text___%d,code\n", count
++);
10277 /* A get_unnamed_section callback for switching to the data section. */
10280 unicosmk_output_data_section_asm_op (const void *data ATTRIBUTE_UNUSED
)
10282 static int count
= 1;
10283 fprintf (asm_out_file
, "\t.endp\n\n\t.psect\tgcc@data___%d,data\n", count
++);
10286 /* Implement TARGET_ASM_INIT_SECTIONS.
10288 The Cray assembler is really weird with respect to sections. It has only
10289 named sections and you can't reopen a section once it has been closed.
10290 This means that we have to generate unique names whenever we want to
10291 reenter the text or the data section. */
10294 unicosmk_init_sections (void)
10296 text_section
= get_unnamed_section (SECTION_CODE
,
10297 unicosmk_output_text_section_asm_op
,
10299 data_section
= get_unnamed_section (SECTION_WRITE
,
10300 unicosmk_output_data_section_asm_op
,
10302 readonly_data_section
= data_section
;
10305 static unsigned int
10306 unicosmk_section_type_flags (tree decl
, const char *name
,
10307 int reloc ATTRIBUTE_UNUSED
)
10309 unsigned int flags
= default_section_type_flags (decl
, name
, reloc
);
10314 if (TREE_CODE (decl
) == FUNCTION_DECL
)
10316 current_section_align
= floor_log2 (FUNCTION_BOUNDARY
/ BITS_PER_UNIT
);
10317 if (align_functions_log
> current_section_align
)
10318 current_section_align
= align_functions_log
;
10320 if (! strcmp (IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (decl
)), "main"))
10321 flags
|= SECTION_MAIN
;
10324 current_section_align
= floor_log2 (DECL_ALIGN (decl
) / BITS_PER_UNIT
);
10326 if (TREE_PUBLIC (decl
))
10327 flags
|= SECTION_PUBLIC
;
10332 /* Generate a section name for decl and associate it with the
10336 unicosmk_unique_section (tree decl
, int reloc ATTRIBUTE_UNUSED
)
10343 name
= IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (decl
));
10344 name
= default_strip_name_encoding (name
);
10345 len
= strlen (name
);
10347 if (TREE_CODE (decl
) == FUNCTION_DECL
)
10351 /* It is essential that we prefix the section name here because
10352 otherwise the section names generated for constructors and
10353 destructors confuse collect2. */
10355 string
= alloca (len
+ 6);
10356 sprintf (string
, "code@%s", name
);
10357 DECL_SECTION_NAME (decl
) = build_string (len
+ 5, string
);
10359 else if (TREE_PUBLIC (decl
))
10360 DECL_SECTION_NAME (decl
) = build_string (len
, name
);
10365 string
= alloca (len
+ 6);
10366 sprintf (string
, "data@%s", name
);
10367 DECL_SECTION_NAME (decl
) = build_string (len
+ 5, string
);
10371 /* Switch to an arbitrary section NAME with attributes as specified
10372 by FLAGS. ALIGN specifies any known alignment requirements for
10373 the section; 0 if the default should be used. */
10376 unicosmk_asm_named_section (const char *name
, unsigned int flags
,
10377 tree decl ATTRIBUTE_UNUSED
)
10381 /* Close the previous section. */
10383 fputs ("\t.endp\n\n", asm_out_file
);
10385 /* Find out what kind of section we are opening. */
10387 if (flags
& SECTION_MAIN
)
10388 fputs ("\t.start\tmain\n", asm_out_file
);
10390 if (flags
& SECTION_CODE
)
10392 else if (flags
& SECTION_PUBLIC
)
10397 if (current_section_align
!= 0)
10398 fprintf (asm_out_file
, "\t.psect\t%s,%d,%s\n", name
,
10399 current_section_align
, kind
);
10401 fprintf (asm_out_file
, "\t.psect\t%s,%s\n", name
, kind
);
10405 unicosmk_insert_attributes (tree decl
, tree
*attr_ptr ATTRIBUTE_UNUSED
)
10408 && (TREE_PUBLIC (decl
) || TREE_CODE (decl
) == FUNCTION_DECL
))
10409 unicosmk_unique_section (decl
, 0);
10412 /* Output an alignment directive. We have to use the macro 'gcc@code@align'
10413 in code sections because .align fill unused space with zeroes. */
10416 unicosmk_output_align (FILE *file
, int align
)
10418 if (inside_function
)
10419 fprintf (file
, "\tgcc@code@align\t%d\n", align
);
10421 fprintf (file
, "\t.align\t%d\n", align
);
10424 /* Add a case vector to the current function's list of deferred case
10425 vectors. Case vectors have to be put into a separate section because CAM
10426 does not allow data definitions in code sections. */
10429 unicosmk_defer_case_vector (rtx lab
, rtx vec
)
10431 struct machine_function
*machine
= cfun
->machine
;
10433 vec
= gen_rtx_EXPR_LIST (VOIDmode
, lab
, vec
);
10434 machine
->addr_list
= gen_rtx_EXPR_LIST (VOIDmode
, vec
,
10435 machine
->addr_list
);
10438 /* Output a case vector. */
10441 unicosmk_output_addr_vec (FILE *file
, rtx vec
)
10443 rtx lab
= XEXP (vec
, 0);
10444 rtx body
= XEXP (vec
, 1);
10445 int vlen
= XVECLEN (body
, 0);
10448 (*targetm
.asm_out
.internal_label
) (file
, "L", CODE_LABEL_NUMBER (lab
));
10450 for (idx
= 0; idx
< vlen
; idx
++)
10452 ASM_OUTPUT_ADDR_VEC_ELT
10453 (file
, CODE_LABEL_NUMBER (XEXP (XVECEXP (body
, 0, idx
), 0)));
10457 /* Output current function's deferred case vectors. */
10460 unicosmk_output_deferred_case_vectors (FILE *file
)
10462 struct machine_function
*machine
= cfun
->machine
;
10465 if (machine
->addr_list
== NULL_RTX
)
10468 switch_to_section (data_section
);
10469 for (t
= machine
->addr_list
; t
; t
= XEXP (t
, 1))
10470 unicosmk_output_addr_vec (file
, XEXP (t
, 0));
10473 /* Generate the name of the SSIB section for the current function. */
10475 #define SSIB_PREFIX "__SSIB_"
10476 #define SSIB_PREFIX_LEN 7
10478 static const char *
10479 unicosmk_ssib_name (void)
10481 /* This is ok since CAM won't be able to deal with names longer than that
10484 static char name
[256];
10487 const char *fnname
;
10490 x
= DECL_RTL (cfun
->decl
);
10491 gcc_assert (MEM_P (x
));
10493 gcc_assert (GET_CODE (x
) == SYMBOL_REF
);
10494 fnname
= XSTR (x
, 0);
10496 len
= strlen (fnname
);
10497 if (len
+ SSIB_PREFIX_LEN
> 255)
10498 len
= 255 - SSIB_PREFIX_LEN
;
10500 strcpy (name
, SSIB_PREFIX
);
10501 strncpy (name
+ SSIB_PREFIX_LEN
, fnname
, len
);
10502 name
[len
+ SSIB_PREFIX_LEN
] = 0;
10507 /* Set up the dynamic subprogram information block (DSIB) and update the
10508 frame pointer register ($15) for subroutines which have a frame. If the
10509 subroutine doesn't have a frame, simply increment $15. */
10512 unicosmk_gen_dsib (unsigned long *imaskP
)
10514 if (alpha_procedure_type
== PT_STACK
)
10516 const char *ssib_name
;
10519 /* Allocate 64 bytes for the DSIB. */
10521 FRP (emit_insn (gen_adddi3 (stack_pointer_rtx
, stack_pointer_rtx
,
10523 emit_insn (gen_blockage ());
10525 /* Save the return address. */
10527 mem
= gen_rtx_MEM (DImode
, plus_constant (stack_pointer_rtx
, 56));
10528 set_mem_alias_set (mem
, alpha_sr_alias_set
);
10529 FRP (emit_move_insn (mem
, gen_rtx_REG (DImode
, REG_RA
)));
10530 (*imaskP
) &= ~(1UL << REG_RA
);
10532 /* Save the old frame pointer. */
10534 mem
= gen_rtx_MEM (DImode
, plus_constant (stack_pointer_rtx
, 48));
10535 set_mem_alias_set (mem
, alpha_sr_alias_set
);
10536 FRP (emit_move_insn (mem
, hard_frame_pointer_rtx
));
10537 (*imaskP
) &= ~(1UL << HARD_FRAME_POINTER_REGNUM
);
10539 emit_insn (gen_blockage ());
10541 /* Store the SSIB pointer. */
10543 ssib_name
= ggc_strdup (unicosmk_ssib_name ());
10544 mem
= gen_rtx_MEM (DImode
, plus_constant (stack_pointer_rtx
, 32));
10545 set_mem_alias_set (mem
, alpha_sr_alias_set
);
10547 FRP (emit_move_insn (gen_rtx_REG (DImode
, 5),
10548 gen_rtx_SYMBOL_REF (Pmode
, ssib_name
)));
10549 FRP (emit_move_insn (mem
, gen_rtx_REG (DImode
, 5)));
10551 /* Save the CIW index. */
10553 mem
= gen_rtx_MEM (DImode
, plus_constant (stack_pointer_rtx
, 24));
10554 set_mem_alias_set (mem
, alpha_sr_alias_set
);
10555 FRP (emit_move_insn (mem
, gen_rtx_REG (DImode
, 25)));
10557 emit_insn (gen_blockage ());
10559 /* Set the new frame pointer. */
10560 FRP (emit_insn (gen_adddi3 (hard_frame_pointer_rtx
,
10561 stack_pointer_rtx
, GEN_INT (64))));
10565 /* Increment the frame pointer register to indicate that we do not
10567 emit_insn (gen_adddi3 (hard_frame_pointer_rtx
,
10568 hard_frame_pointer_rtx
, const1_rtx
));
10572 /* Output the static subroutine information block for the current
10576 unicosmk_output_ssib (FILE *file
, const char *fnname
)
10582 struct machine_function
*machine
= cfun
->machine
;
10585 fprintf (file
, "\t.endp\n\n\t.psect\t%s%s,data\n", user_label_prefix
,
10586 unicosmk_ssib_name ());
10588 /* Some required stuff and the function name length. */
10590 len
= strlen (fnname
);
10591 fprintf (file
, "\t.quad\t^X20008%2.2X28\n", len
);
10594 ??? We don't do that yet. */
10596 fputs ("\t.quad\t0\n", file
);
10598 /* Function address. */
10600 fputs ("\t.quad\t", file
);
10601 assemble_name (file
, fnname
);
10604 fputs ("\t.quad\t0\n", file
);
10605 fputs ("\t.quad\t0\n", file
);
10608 ??? We do it the same way Cray CC does it but this could be
10611 for( i
= 0; i
< len
; i
++ )
10612 fprintf (file
, "\t.byte\t%d\n", (int)(fnname
[i
]));
10613 if( (len
% 8) == 0 )
10614 fputs ("\t.quad\t0\n", file
);
10616 fprintf (file
, "\t.bits\t%d : 0\n", (8 - (len
% 8))*8);
10618 /* All call information words used in the function. */
10620 for (x
= machine
->first_ciw
; x
; x
= XEXP (x
, 1))
10623 #if HOST_BITS_PER_WIDE_INT == 32
10624 fprintf (file
, "\t.quad\t" HOST_WIDE_INT_PRINT_DOUBLE_HEX
"\n",
10625 CONST_DOUBLE_HIGH (ciw
), CONST_DOUBLE_LOW (ciw
));
10627 fprintf (file
, "\t.quad\t" HOST_WIDE_INT_PRINT_HEX
"\n", INTVAL (ciw
));
10632 /* Add a call information word (CIW) to the list of the current function's
10633 CIWs and return its index.
10635 X is a CONST_INT or CONST_DOUBLE representing the CIW. */
10638 unicosmk_add_call_info_word (rtx x
)
10641 struct machine_function
*machine
= cfun
->machine
;
10643 node
= gen_rtx_EXPR_LIST (VOIDmode
, x
, NULL_RTX
);
10644 if (machine
->first_ciw
== NULL_RTX
)
10645 machine
->first_ciw
= node
;
10647 XEXP (machine
->last_ciw
, 1) = node
;
10649 machine
->last_ciw
= node
;
10650 ++machine
->ciw_count
;
10652 return GEN_INT (machine
->ciw_count
10653 + strlen (current_function_name ())/8 + 5);
10656 /* The Cray assembler doesn't accept extern declarations for symbols which
10657 are defined in the same file. We have to keep track of all global
10658 symbols which are referenced and/or defined in a source file and output
10659 extern declarations for those which are referenced but not defined at
10660 the end of file. */
10662 /* List of identifiers for which an extern declaration might have to be
10664 /* FIXME: needs to use GC, so it can be saved and restored for PCH. */
10666 struct unicosmk_extern_list
10668 struct unicosmk_extern_list
*next
;
10672 static struct unicosmk_extern_list
*unicosmk_extern_head
= 0;
10674 /* Output extern declarations which are required for every asm file. */
10677 unicosmk_output_default_externs (FILE *file
)
10679 static const char *const externs
[] =
10680 { "__T3E_MISMATCH" };
10685 n
= ARRAY_SIZE (externs
);
10687 for (i
= 0; i
< n
; i
++)
10688 fprintf (file
, "\t.extern\t%s\n", externs
[i
]);
10691 /* Output extern declarations for global symbols which are have been
10692 referenced but not defined. */
10695 unicosmk_output_externs (FILE *file
)
10697 struct unicosmk_extern_list
*p
;
10698 const char *real_name
;
10702 len
= strlen (user_label_prefix
);
10703 for (p
= unicosmk_extern_head
; p
!= 0; p
= p
->next
)
10705 /* We have to strip the encoding and possibly remove user_label_prefix
10706 from the identifier in order to handle -fleading-underscore and
10707 explicit asm names correctly (cf. gcc.dg/asm-names-1.c). */
10708 real_name
= default_strip_name_encoding (p
->name
);
10709 if (len
&& p
->name
[0] == '*'
10710 && !memcmp (real_name
, user_label_prefix
, len
))
10713 name_tree
= get_identifier (real_name
);
10714 if (! TREE_ASM_WRITTEN (name_tree
))
10716 TREE_ASM_WRITTEN (name_tree
) = 1;
10717 fputs ("\t.extern\t", file
);
10718 assemble_name (file
, p
->name
);
10724 /* Record an extern. */
10727 unicosmk_add_extern (const char *name
)
10729 struct unicosmk_extern_list
*p
;
10731 p
= (struct unicosmk_extern_list
*)
10732 xmalloc (sizeof (struct unicosmk_extern_list
));
10733 p
->next
= unicosmk_extern_head
;
10735 unicosmk_extern_head
= p
;
10738 /* The Cray assembler generates incorrect code if identifiers which
10739 conflict with register names are used as instruction operands. We have
10740 to replace such identifiers with DEX expressions. */
10742 /* Structure to collect identifiers which have been replaced by DEX
10744 /* FIXME: needs to use GC, so it can be saved and restored for PCH. */
10746 struct unicosmk_dex
{
10747 struct unicosmk_dex
*next
;
10751 /* List of identifiers which have been replaced by DEX expressions. The DEX
10752 number is determined by the position in the list. */
10754 static struct unicosmk_dex
*unicosmk_dex_list
= NULL
;
10756 /* The number of elements in the DEX list. */
10758 static int unicosmk_dex_count
= 0;
10760 /* Check if NAME must be replaced by a DEX expression. */
10763 unicosmk_special_name (const char *name
)
10765 if (name
[0] == '*')
10768 if (name
[0] == '$')
10771 if (name
[0] != 'r' && name
[0] != 'f' && name
[0] != 'R' && name
[0] != 'F')
10776 case '1': case '2':
10777 return (name
[2] == '\0' || (ISDIGIT (name
[2]) && name
[3] == '\0'));
10780 return (name
[2] == '\0'
10781 || ((name
[2] == '0' || name
[2] == '1') && name
[3] == '\0'));
10784 return (ISDIGIT (name
[1]) && name
[2] == '\0');
10788 /* Return the DEX number if X must be replaced by a DEX expression and 0
10792 unicosmk_need_dex (rtx x
)
10794 struct unicosmk_dex
*dex
;
10798 if (GET_CODE (x
) != SYMBOL_REF
)
10802 if (! unicosmk_special_name (name
))
10805 i
= unicosmk_dex_count
;
10806 for (dex
= unicosmk_dex_list
; dex
; dex
= dex
->next
)
10808 if (! strcmp (name
, dex
->name
))
10813 dex
= (struct unicosmk_dex
*) xmalloc (sizeof (struct unicosmk_dex
));
10815 dex
->next
= unicosmk_dex_list
;
10816 unicosmk_dex_list
= dex
;
10818 ++unicosmk_dex_count
;
10819 return unicosmk_dex_count
;
10822 /* Output the DEX definitions for this file. */
10825 unicosmk_output_dex (FILE *file
)
10827 struct unicosmk_dex
*dex
;
10830 if (unicosmk_dex_list
== NULL
)
10833 fprintf (file
, "\t.dexstart\n");
10835 i
= unicosmk_dex_count
;
10836 for (dex
= unicosmk_dex_list
; dex
; dex
= dex
->next
)
10838 fprintf (file
, "\tDEX (%d) = ", i
);
10839 assemble_name (file
, dex
->name
);
10844 fprintf (file
, "\t.dexend\n");
10847 /* Output text that to appear at the beginning of an assembler file. */
10850 unicosmk_file_start (void)
10854 fputs ("\t.ident\t", asm_out_file
);
10855 unicosmk_output_module_name (asm_out_file
);
10856 fputs ("\n\n", asm_out_file
);
10858 /* The Unicos/Mk assembler uses different register names. Instead of trying
10859 to support them, we simply use micro definitions. */
10861 /* CAM has different register names: rN for the integer register N and fN
10862 for the floating-point register N. Instead of trying to use these in
10863 alpha.md, we define the symbols $N and $fN to refer to the appropriate
10866 for (i
= 0; i
< 32; ++i
)
10867 fprintf (asm_out_file
, "$%d <- r%d\n", i
, i
);
10869 for (i
= 0; i
< 32; ++i
)
10870 fprintf (asm_out_file
, "$f%d <- f%d\n", i
, i
);
10872 putc ('\n', asm_out_file
);
10874 /* The .align directive fill unused space with zeroes which does not work
10875 in code sections. We define the macro 'gcc@code@align' which uses nops
10876 instead. Note that it assumes that code sections always have the
10877 biggest possible alignment since . refers to the current offset from
10878 the beginning of the section. */
10880 fputs ("\t.macro gcc@code@align n\n", asm_out_file
);
10881 fputs ("gcc@n@bytes = 1 << n\n", asm_out_file
);
10882 fputs ("gcc@here = . % gcc@n@bytes\n", asm_out_file
);
10883 fputs ("\t.if ne, gcc@here, 0\n", asm_out_file
);
10884 fputs ("\t.repeat (gcc@n@bytes - gcc@here) / 4\n", asm_out_file
);
10885 fputs ("\tbis r31,r31,r31\n", asm_out_file
);
10886 fputs ("\t.endr\n", asm_out_file
);
10887 fputs ("\t.endif\n", asm_out_file
);
10888 fputs ("\t.endm gcc@code@align\n\n", asm_out_file
);
10890 /* Output extern declarations which should always be visible. */
10891 unicosmk_output_default_externs (asm_out_file
);
10893 /* Open a dummy section. We always need to be inside a section for the
10894 section-switching code to work correctly.
10895 ??? This should be a module id or something like that. I still have to
10896 figure out what the rules for those are. */
10897 fputs ("\n\t.psect\t$SG00000,data\n", asm_out_file
);
10900 /* Output text to appear at the end of an assembler file. This includes all
10901 pending extern declarations and DEX expressions. */
10904 unicosmk_file_end (void)
10906 fputs ("\t.endp\n\n", asm_out_file
);
10908 /* Output all pending externs. */
10910 unicosmk_output_externs (asm_out_file
);
10912 /* Output dex definitions used for functions whose names conflict with
10915 unicosmk_output_dex (asm_out_file
);
10917 fputs ("\t.end\t", asm_out_file
);
10918 unicosmk_output_module_name (asm_out_file
);
10919 putc ('\n', asm_out_file
);
10925 unicosmk_output_deferred_case_vectors (FILE *file ATTRIBUTE_UNUSED
)
10929 unicosmk_gen_dsib (unsigned long *imaskP ATTRIBUTE_UNUSED
)
10933 unicosmk_output_ssib (FILE * file ATTRIBUTE_UNUSED
,
10934 const char * fnname ATTRIBUTE_UNUSED
)
10938 unicosmk_add_call_info_word (rtx x ATTRIBUTE_UNUSED
)
10944 unicosmk_need_dex (rtx x ATTRIBUTE_UNUSED
)
10949 #endif /* TARGET_ABI_UNICOSMK */
10952 alpha_init_libfuncs (void)
10954 if (TARGET_ABI_UNICOSMK
)
10956 /* Prevent gcc from generating calls to __divsi3. */
10957 set_optab_libfunc (sdiv_optab
, SImode
, 0);
10958 set_optab_libfunc (udiv_optab
, SImode
, 0);
10960 /* Use the functions provided by the system library
10961 for DImode integer division. */
10962 set_optab_libfunc (sdiv_optab
, DImode
, "$sldiv");
10963 set_optab_libfunc (udiv_optab
, DImode
, "$uldiv");
10965 else if (TARGET_ABI_OPEN_VMS
)
10967 /* Use the VMS runtime library functions for division and
10969 set_optab_libfunc (sdiv_optab
, SImode
, "OTS$DIV_I");
10970 set_optab_libfunc (sdiv_optab
, DImode
, "OTS$DIV_L");
10971 set_optab_libfunc (udiv_optab
, SImode
, "OTS$DIV_UI");
10972 set_optab_libfunc (udiv_optab
, DImode
, "OTS$DIV_UL");
10973 set_optab_libfunc (smod_optab
, SImode
, "OTS$REM_I");
10974 set_optab_libfunc (smod_optab
, DImode
, "OTS$REM_L");
10975 set_optab_libfunc (umod_optab
, SImode
, "OTS$REM_UI");
10976 set_optab_libfunc (umod_optab
, DImode
, "OTS$REM_UL");
10977 abort_libfunc
= init_one_libfunc ("decc$abort");
10978 memcmp_libfunc
= init_one_libfunc ("decc$memcmp");
10979 #ifdef MEM_LIBFUNCS_INIT
10986 /* Initialize the GCC target structure. */
10987 #if TARGET_ABI_OPEN_VMS
10988 # undef TARGET_ATTRIBUTE_TABLE
10989 # define TARGET_ATTRIBUTE_TABLE vms_attribute_table
10990 # undef TARGET_CAN_ELIMINATE
10991 # define TARGET_CAN_ELIMINATE alpha_vms_can_eliminate
10994 #undef TARGET_IN_SMALL_DATA_P
10995 #define TARGET_IN_SMALL_DATA_P alpha_in_small_data_p
10997 #if TARGET_ABI_UNICOSMK
10998 # undef TARGET_INSERT_ATTRIBUTES
10999 # define TARGET_INSERT_ATTRIBUTES unicosmk_insert_attributes
11000 # undef TARGET_SECTION_TYPE_FLAGS
11001 # define TARGET_SECTION_TYPE_FLAGS unicosmk_section_type_flags
11002 # undef TARGET_ASM_UNIQUE_SECTION
11003 # define TARGET_ASM_UNIQUE_SECTION unicosmk_unique_section
11004 #undef TARGET_ASM_FUNCTION_RODATA_SECTION
11005 #define TARGET_ASM_FUNCTION_RODATA_SECTION default_no_function_rodata_section
11006 # undef TARGET_ASM_GLOBALIZE_LABEL
11007 # define TARGET_ASM_GLOBALIZE_LABEL hook_void_FILEptr_constcharptr
11008 # undef TARGET_MUST_PASS_IN_STACK
11009 # define TARGET_MUST_PASS_IN_STACK unicosmk_must_pass_in_stack
11012 #undef TARGET_ASM_ALIGNED_HI_OP
11013 #define TARGET_ASM_ALIGNED_HI_OP "\t.word\t"
11014 #undef TARGET_ASM_ALIGNED_DI_OP
11015 #define TARGET_ASM_ALIGNED_DI_OP "\t.quad\t"
11017 /* Default unaligned ops are provided for ELF systems. To get unaligned
11018 data for non-ELF systems, we have to turn off auto alignment. */
11019 #if !defined (OBJECT_FORMAT_ELF) || TARGET_ABI_OPEN_VMS
11020 #undef TARGET_ASM_UNALIGNED_HI_OP
11021 #define TARGET_ASM_UNALIGNED_HI_OP "\t.align 0\n\t.word\t"
11022 #undef TARGET_ASM_UNALIGNED_SI_OP
11023 #define TARGET_ASM_UNALIGNED_SI_OP "\t.align 0\n\t.long\t"
11024 #undef TARGET_ASM_UNALIGNED_DI_OP
11025 #define TARGET_ASM_UNALIGNED_DI_OP "\t.align 0\n\t.quad\t"
11028 #ifdef OBJECT_FORMAT_ELF
11029 #undef TARGET_ASM_RELOC_RW_MASK
11030 #define TARGET_ASM_RELOC_RW_MASK alpha_elf_reloc_rw_mask
11031 #undef TARGET_ASM_SELECT_RTX_SECTION
11032 #define TARGET_ASM_SELECT_RTX_SECTION alpha_elf_select_rtx_section
11033 #undef TARGET_SECTION_TYPE_FLAGS
11034 #define TARGET_SECTION_TYPE_FLAGS alpha_elf_section_type_flags
11037 #undef TARGET_ASM_FUNCTION_END_PROLOGUE
11038 #define TARGET_ASM_FUNCTION_END_PROLOGUE alpha_output_function_end_prologue
11040 #undef TARGET_INIT_LIBFUNCS
11041 #define TARGET_INIT_LIBFUNCS alpha_init_libfuncs
11043 #undef TARGET_LEGITIMIZE_ADDRESS
11044 #define TARGET_LEGITIMIZE_ADDRESS alpha_legitimize_address
11046 #if TARGET_ABI_UNICOSMK
11047 #undef TARGET_ASM_FILE_START
11048 #define TARGET_ASM_FILE_START unicosmk_file_start
11049 #undef TARGET_ASM_FILE_END
11050 #define TARGET_ASM_FILE_END unicosmk_file_end
11052 #undef TARGET_ASM_FILE_START
11053 #define TARGET_ASM_FILE_START alpha_file_start
11054 #undef TARGET_ASM_FILE_START_FILE_DIRECTIVE
11055 #define TARGET_ASM_FILE_START_FILE_DIRECTIVE true
11058 #undef TARGET_SCHED_ADJUST_COST
11059 #define TARGET_SCHED_ADJUST_COST alpha_adjust_cost
11060 #undef TARGET_SCHED_ISSUE_RATE
11061 #define TARGET_SCHED_ISSUE_RATE alpha_issue_rate
11062 #undef TARGET_SCHED_FIRST_CYCLE_MULTIPASS_DFA_LOOKAHEAD
11063 #define TARGET_SCHED_FIRST_CYCLE_MULTIPASS_DFA_LOOKAHEAD \
11064 alpha_multipass_dfa_lookahead
11066 #undef TARGET_HAVE_TLS
11067 #define TARGET_HAVE_TLS HAVE_AS_TLS
11069 #undef TARGET_INIT_BUILTINS
11070 #define TARGET_INIT_BUILTINS alpha_init_builtins
11071 #undef TARGET_EXPAND_BUILTIN
11072 #define TARGET_EXPAND_BUILTIN alpha_expand_builtin
11073 #undef TARGET_FOLD_BUILTIN
11074 #define TARGET_FOLD_BUILTIN alpha_fold_builtin
11076 #undef TARGET_FUNCTION_OK_FOR_SIBCALL
11077 #define TARGET_FUNCTION_OK_FOR_SIBCALL alpha_function_ok_for_sibcall
11078 #undef TARGET_CANNOT_COPY_INSN_P
11079 #define TARGET_CANNOT_COPY_INSN_P alpha_cannot_copy_insn_p
11080 #undef TARGET_CANNOT_FORCE_CONST_MEM
11081 #define TARGET_CANNOT_FORCE_CONST_MEM alpha_cannot_force_const_mem
11084 #undef TARGET_ASM_OUTPUT_MI_THUNK
11085 #define TARGET_ASM_OUTPUT_MI_THUNK alpha_output_mi_thunk_osf
11086 #undef TARGET_ASM_CAN_OUTPUT_MI_THUNK
11087 #define TARGET_ASM_CAN_OUTPUT_MI_THUNK hook_bool_const_tree_hwi_hwi_const_tree_true
11088 #undef TARGET_STDARG_OPTIMIZE_HOOK
11089 #define TARGET_STDARG_OPTIMIZE_HOOK alpha_stdarg_optimize_hook
11092 #undef TARGET_RTX_COSTS
11093 #define TARGET_RTX_COSTS alpha_rtx_costs
11094 #undef TARGET_ADDRESS_COST
11095 #define TARGET_ADDRESS_COST hook_int_rtx_bool_0
11097 #undef TARGET_MACHINE_DEPENDENT_REORG
11098 #define TARGET_MACHINE_DEPENDENT_REORG alpha_reorg
11100 #undef TARGET_PROMOTE_FUNCTION_MODE
11101 #define TARGET_PROMOTE_FUNCTION_MODE default_promote_function_mode_always_promote
11102 #undef TARGET_PROMOTE_PROTOTYPES
11103 #define TARGET_PROMOTE_PROTOTYPES hook_bool_const_tree_false
11104 #undef TARGET_RETURN_IN_MEMORY
11105 #define TARGET_RETURN_IN_MEMORY alpha_return_in_memory
11106 #undef TARGET_PASS_BY_REFERENCE
11107 #define TARGET_PASS_BY_REFERENCE alpha_pass_by_reference
11108 #undef TARGET_SETUP_INCOMING_VARARGS
11109 #define TARGET_SETUP_INCOMING_VARARGS alpha_setup_incoming_varargs
11110 #undef TARGET_STRICT_ARGUMENT_NAMING
11111 #define TARGET_STRICT_ARGUMENT_NAMING hook_bool_CUMULATIVE_ARGS_true
11112 #undef TARGET_PRETEND_OUTGOING_VARARGS_NAMED
11113 #define TARGET_PRETEND_OUTGOING_VARARGS_NAMED hook_bool_CUMULATIVE_ARGS_true
11114 #undef TARGET_SPLIT_COMPLEX_ARG
11115 #define TARGET_SPLIT_COMPLEX_ARG alpha_split_complex_arg
11116 #undef TARGET_GIMPLIFY_VA_ARG_EXPR
11117 #define TARGET_GIMPLIFY_VA_ARG_EXPR alpha_gimplify_va_arg
11118 #undef TARGET_ARG_PARTIAL_BYTES
11119 #define TARGET_ARG_PARTIAL_BYTES alpha_arg_partial_bytes
11120 #undef TARGET_TRAMPOLINE_INIT
11121 #define TARGET_TRAMPOLINE_INIT alpha_trampoline_init
11123 #undef TARGET_SECONDARY_RELOAD
11124 #define TARGET_SECONDARY_RELOAD alpha_secondary_reload
11126 #undef TARGET_SCALAR_MODE_SUPPORTED_P
11127 #define TARGET_SCALAR_MODE_SUPPORTED_P alpha_scalar_mode_supported_p
11128 #undef TARGET_VECTOR_MODE_SUPPORTED_P
11129 #define TARGET_VECTOR_MODE_SUPPORTED_P alpha_vector_mode_supported_p
11131 #undef TARGET_BUILD_BUILTIN_VA_LIST
11132 #define TARGET_BUILD_BUILTIN_VA_LIST alpha_build_builtin_va_list
11134 #undef TARGET_EXPAND_BUILTIN_VA_START
11135 #define TARGET_EXPAND_BUILTIN_VA_START alpha_va_start
11137 /* The Alpha architecture does not require sequential consistency. See
11138 http://www.cs.umd.edu/~pugh/java/memoryModel/AlphaReordering.html
11139 for an example of how it can be violated in practice. */
11140 #undef TARGET_RELAXED_ORDERING
11141 #define TARGET_RELAXED_ORDERING true
11143 #undef TARGET_DEFAULT_TARGET_FLAGS
11144 #define TARGET_DEFAULT_TARGET_FLAGS \
11145 (TARGET_DEFAULT | TARGET_CPU_DEFAULT | TARGET_DEFAULT_EXPLICIT_RELOCS)
11146 #undef TARGET_HANDLE_OPTION
11147 #define TARGET_HANDLE_OPTION alpha_handle_option
11149 #ifdef TARGET_ALTERNATE_LONG_DOUBLE_MANGLING
11150 #undef TARGET_MANGLE_TYPE
11151 #define TARGET_MANGLE_TYPE alpha_mangle_type
11154 #undef TARGET_LEGITIMATE_ADDRESS_P
11155 #define TARGET_LEGITIMATE_ADDRESS_P alpha_legitimate_address_p
11157 struct gcc_target targetm
= TARGET_INITIALIZER
;
11160 #include "gt-alpha.h"