Rebase.
[official-gcc.git] / gcc / config / alpha / alpha.c
blobe79621ff058f595b5946ba7037861cdfcb4b0857
1 /* Subroutines used for code generation on the DEC Alpha.
2 Copyright (C) 1992-2014 Free Software Foundation, Inc.
3 Contributed by Richard Kenner (kenner@vlsi1.ultra.nyu.edu)
5 This file is part of GCC.
7 GCC is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License as published by
9 the Free Software Foundation; either version 3, or (at your option)
10 any later version.
12 GCC is distributed in the hope that it will be useful,
13 but WITHOUT ANY WARRANTY; without even the implied warranty of
14 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 GNU General Public License for more details.
17 You should have received a copy of the GNU General Public License
18 along with GCC; see the file COPYING3. If not see
19 <http://www.gnu.org/licenses/>. */
22 #include "config.h"
23 #include "system.h"
24 #include "coretypes.h"
25 #include "tm.h"
26 #include "rtl.h"
27 #include "tree.h"
28 #include "stor-layout.h"
29 #include "calls.h"
30 #include "varasm.h"
31 #include "regs.h"
32 #include "hard-reg-set.h"
33 #include "insn-config.h"
34 #include "conditions.h"
35 #include "output.h"
36 #include "insn-attr.h"
37 #include "flags.h"
38 #include "recog.h"
39 #include "expr.h"
40 #include "optabs.h"
41 #include "reload.h"
42 #include "obstack.h"
43 #include "except.h"
44 #include "function.h"
45 #include "diagnostic-core.h"
46 #include "ggc.h"
47 #include "tm_p.h"
48 #include "target.h"
49 #include "target-def.h"
50 #include "common/common-target.h"
51 #include "debug.h"
52 #include "langhooks.h"
53 #include "splay-tree.h"
54 #include "pointer-set.h"
55 #include "hash-table.h"
56 #include "vec.h"
57 #include "basic-block.h"
58 #include "tree-ssa-alias.h"
59 #include "internal-fn.h"
60 #include "gimple-fold.h"
61 #include "tree-eh.h"
62 #include "gimple-expr.h"
63 #include "is-a.h"
64 #include "gimple.h"
65 #include "tree-pass.h"
66 #include "context.h"
67 #include "pass_manager.h"
68 #include "gimple-iterator.h"
69 #include "gimplify.h"
70 #include "gimple-ssa.h"
71 #include "stringpool.h"
72 #include "tree-ssanames.h"
73 #include "tree-stdarg.h"
74 #include "tm-constrs.h"
75 #include "df.h"
76 #include "libfuncs.h"
77 #include "opts.h"
78 #include "params.h"
79 #include "builtins.h"
81 /* Specify which cpu to schedule for. */
82 enum processor_type alpha_tune;
84 /* Which cpu we're generating code for. */
85 enum processor_type alpha_cpu;
87 static const char * const alpha_cpu_name[] =
89 "ev4", "ev5", "ev6"
92 /* Specify how accurate floating-point traps need to be. */
94 enum alpha_trap_precision alpha_tp;
96 /* Specify the floating-point rounding mode. */
98 enum alpha_fp_rounding_mode alpha_fprm;
100 /* Specify which things cause traps. */
102 enum alpha_fp_trap_mode alpha_fptm;
104 /* Nonzero if inside of a function, because the Alpha asm can't
105 handle .files inside of functions. */
107 static int inside_function = FALSE;
109 /* The number of cycles of latency we should assume on memory reads. */
111 int alpha_memory_latency = 3;
113 /* Whether the function needs the GP. */
115 static int alpha_function_needs_gp;
117 /* The assembler name of the current function. */
119 static const char *alpha_fnname;
121 /* The next explicit relocation sequence number. */
122 extern GTY(()) int alpha_next_sequence_number;
123 int alpha_next_sequence_number = 1;
125 /* The literal and gpdisp sequence numbers for this insn, as printed
126 by %# and %* respectively. */
127 extern GTY(()) int alpha_this_literal_sequence_number;
128 extern GTY(()) int alpha_this_gpdisp_sequence_number;
129 int alpha_this_literal_sequence_number;
130 int alpha_this_gpdisp_sequence_number;
132 /* Costs of various operations on the different architectures. */
134 struct alpha_rtx_cost_data
136 unsigned char fp_add;
137 unsigned char fp_mult;
138 unsigned char fp_div_sf;
139 unsigned char fp_div_df;
140 unsigned char int_mult_si;
141 unsigned char int_mult_di;
142 unsigned char int_shift;
143 unsigned char int_cmov;
144 unsigned short int_div;
147 static struct alpha_rtx_cost_data const alpha_rtx_cost_data[PROCESSOR_MAX] =
149 { /* EV4 */
150 COSTS_N_INSNS (6), /* fp_add */
151 COSTS_N_INSNS (6), /* fp_mult */
152 COSTS_N_INSNS (34), /* fp_div_sf */
153 COSTS_N_INSNS (63), /* fp_div_df */
154 COSTS_N_INSNS (23), /* int_mult_si */
155 COSTS_N_INSNS (23), /* int_mult_di */
156 COSTS_N_INSNS (2), /* int_shift */
157 COSTS_N_INSNS (2), /* int_cmov */
158 COSTS_N_INSNS (97), /* int_div */
160 { /* EV5 */
161 COSTS_N_INSNS (4), /* fp_add */
162 COSTS_N_INSNS (4), /* fp_mult */
163 COSTS_N_INSNS (15), /* fp_div_sf */
164 COSTS_N_INSNS (22), /* fp_div_df */
165 COSTS_N_INSNS (8), /* int_mult_si */
166 COSTS_N_INSNS (12), /* int_mult_di */
167 COSTS_N_INSNS (1) + 1, /* int_shift */
168 COSTS_N_INSNS (1), /* int_cmov */
169 COSTS_N_INSNS (83), /* int_div */
171 { /* EV6 */
172 COSTS_N_INSNS (4), /* fp_add */
173 COSTS_N_INSNS (4), /* fp_mult */
174 COSTS_N_INSNS (12), /* fp_div_sf */
175 COSTS_N_INSNS (15), /* fp_div_df */
176 COSTS_N_INSNS (7), /* int_mult_si */
177 COSTS_N_INSNS (7), /* int_mult_di */
178 COSTS_N_INSNS (1), /* int_shift */
179 COSTS_N_INSNS (2), /* int_cmov */
180 COSTS_N_INSNS (86), /* int_div */
184 /* Similar but tuned for code size instead of execution latency. The
185 extra +N is fractional cost tuning based on latency. It's used to
186 encourage use of cheaper insns like shift, but only if there's just
187 one of them. */
189 static struct alpha_rtx_cost_data const alpha_rtx_cost_size =
191 COSTS_N_INSNS (1), /* fp_add */
192 COSTS_N_INSNS (1), /* fp_mult */
193 COSTS_N_INSNS (1), /* fp_div_sf */
194 COSTS_N_INSNS (1) + 1, /* fp_div_df */
195 COSTS_N_INSNS (1) + 1, /* int_mult_si */
196 COSTS_N_INSNS (1) + 2, /* int_mult_di */
197 COSTS_N_INSNS (1), /* int_shift */
198 COSTS_N_INSNS (1), /* int_cmov */
199 COSTS_N_INSNS (6), /* int_div */
202 /* Get the number of args of a function in one of two ways. */
203 #if TARGET_ABI_OPEN_VMS
204 #define NUM_ARGS crtl->args.info.num_args
205 #else
206 #define NUM_ARGS crtl->args.info
207 #endif
209 #define REG_PV 27
210 #define REG_RA 26
212 /* Declarations of static functions. */
213 static struct machine_function *alpha_init_machine_status (void);
214 static rtx alpha_emit_xfloating_compare (enum rtx_code *, rtx, rtx);
215 static void alpha_handle_trap_shadows (void);
216 static void alpha_align_insns (void);
218 #if TARGET_ABI_OPEN_VMS
219 static void alpha_write_linkage (FILE *, const char *);
220 static bool vms_valid_pointer_mode (enum machine_mode);
221 #else
222 #define vms_patch_builtins() gcc_unreachable()
223 #endif
225 static unsigned int
226 rest_of_handle_trap_shadows (void)
228 alpha_handle_trap_shadows ();
229 return 0;
232 namespace {
234 const pass_data pass_data_handle_trap_shadows =
236 RTL_PASS,
237 "trap_shadows", /* name */
238 OPTGROUP_NONE, /* optinfo_flags */
239 TV_NONE, /* tv_id */
240 0, /* properties_required */
241 0, /* properties_provided */
242 0, /* properties_destroyed */
243 0, /* todo_flags_start */
244 TODO_df_finish, /* todo_flags_finish */
247 class pass_handle_trap_shadows : public rtl_opt_pass
249 public:
250 pass_handle_trap_shadows(gcc::context *ctxt)
251 : rtl_opt_pass(pass_data_handle_trap_shadows, ctxt)
254 /* opt_pass methods: */
255 virtual bool gate (function *)
257 return alpha_tp != ALPHA_TP_PROG || flag_exceptions;
260 virtual unsigned int execute (function *)
262 return rest_of_handle_trap_shadows ();
265 }; // class pass_handle_trap_shadows
267 } // anon namespace
269 rtl_opt_pass *
270 make_pass_handle_trap_shadows (gcc::context *ctxt)
272 return new pass_handle_trap_shadows (ctxt);
275 static unsigned int
276 rest_of_align_insns (void)
278 alpha_align_insns ();
279 return 0;
282 namespace {
284 const pass_data pass_data_align_insns =
286 RTL_PASS,
287 "align_insns", /* name */
288 OPTGROUP_NONE, /* optinfo_flags */
289 TV_NONE, /* tv_id */
290 0, /* properties_required */
291 0, /* properties_provided */
292 0, /* properties_destroyed */
293 0, /* todo_flags_start */
294 TODO_df_finish, /* todo_flags_finish */
297 class pass_align_insns : public rtl_opt_pass
299 public:
300 pass_align_insns(gcc::context *ctxt)
301 : rtl_opt_pass(pass_data_align_insns, ctxt)
304 /* opt_pass methods: */
305 virtual bool gate (function *)
307 /* Due to the number of extra trapb insns, don't bother fixing up
308 alignment when trap precision is instruction. Moreover, we can
309 only do our job when sched2 is run. */
310 return ((alpha_tune == PROCESSOR_EV4
311 || alpha_tune == PROCESSOR_EV5)
312 && optimize && !optimize_size
313 && alpha_tp != ALPHA_TP_INSN
314 && flag_schedule_insns_after_reload);
317 virtual unsigned int execute (function *)
319 return rest_of_align_insns ();
322 }; // class pass_align_insns
324 } // anon namespace
326 rtl_opt_pass *
327 make_pass_align_insns (gcc::context *ctxt)
329 return new pass_align_insns (ctxt);
332 #ifdef TARGET_ALTERNATE_LONG_DOUBLE_MANGLING
333 /* Implement TARGET_MANGLE_TYPE. */
335 static const char *
336 alpha_mangle_type (const_tree type)
338 if (TYPE_MAIN_VARIANT (type) == long_double_type_node
339 && TARGET_LONG_DOUBLE_128)
340 return "g";
342 /* For all other types, use normal C++ mangling. */
343 return NULL;
345 #endif
347 /* Parse target option strings. */
349 static void
350 alpha_option_override (void)
352 static const struct cpu_table {
353 const char *const name;
354 const enum processor_type processor;
355 const int flags;
356 const unsigned short line_size; /* in bytes */
357 const unsigned short l1_size; /* in kb. */
358 const unsigned short l2_size; /* in kb. */
359 } cpu_table[] = {
360 /* EV4/LCA45 had 8k L1 caches; EV45 had 16k L1 caches.
361 EV4/EV45 had 128k to 16M 32-byte direct Bcache. LCA45
362 had 64k to 8M 8-byte direct Bcache. */
363 { "ev4", PROCESSOR_EV4, 0, 32, 8, 8*1024 },
364 { "21064", PROCESSOR_EV4, 0, 32, 8, 8*1024 },
365 { "ev45", PROCESSOR_EV4, 0, 32, 16, 16*1024 },
367 /* EV5 or EV56 had 8k 32 byte L1, 96k 32 or 64 byte L2,
368 and 1M to 16M 64 byte L3 (not modeled).
369 PCA56 had 16k 64-byte cache; PCA57 had 32k Icache.
370 PCA56 had 8k 64-byte cache; PCA57 had 16k Dcache. */
371 { "ev5", PROCESSOR_EV5, 0, 32, 8, 96 },
372 { "21164", PROCESSOR_EV5, 0, 32, 8, 96 },
373 { "ev56", PROCESSOR_EV5, MASK_BWX, 32, 8, 96 },
374 { "21164a", PROCESSOR_EV5, MASK_BWX, 32, 8, 96 },
375 { "pca56", PROCESSOR_EV5, MASK_BWX|MASK_MAX, 64, 16, 4*1024 },
376 { "21164PC",PROCESSOR_EV5, MASK_BWX|MASK_MAX, 64, 16, 4*1024 },
377 { "21164pc",PROCESSOR_EV5, MASK_BWX|MASK_MAX, 64, 16, 4*1024 },
379 /* EV6 had 64k 64 byte L1, 1M to 16M Bcache. */
380 { "ev6", PROCESSOR_EV6, MASK_BWX|MASK_MAX|MASK_FIX, 64, 64, 16*1024 },
381 { "21264", PROCESSOR_EV6, MASK_BWX|MASK_MAX|MASK_FIX, 64, 64, 16*1024 },
382 { "ev67", PROCESSOR_EV6, MASK_BWX|MASK_MAX|MASK_FIX|MASK_CIX,
383 64, 64, 16*1024 },
384 { "21264a", PROCESSOR_EV6, MASK_BWX|MASK_MAX|MASK_FIX|MASK_CIX,
385 64, 64, 16*1024 }
388 opt_pass *pass_handle_trap_shadows = make_pass_handle_trap_shadows (g);
389 static struct register_pass_info handle_trap_shadows_info
390 = { pass_handle_trap_shadows, "eh_ranges",
391 1, PASS_POS_INSERT_AFTER
394 opt_pass *pass_align_insns = make_pass_align_insns (g);
395 static struct register_pass_info align_insns_info
396 = { pass_align_insns, "shorten",
397 1, PASS_POS_INSERT_BEFORE
400 int const ct_size = ARRAY_SIZE (cpu_table);
401 int line_size = 0, l1_size = 0, l2_size = 0;
402 int i;
404 #ifdef SUBTARGET_OVERRIDE_OPTIONS
405 SUBTARGET_OVERRIDE_OPTIONS;
406 #endif
408 /* Default to full IEEE compliance mode for Go language. */
409 if (strcmp (lang_hooks.name, "GNU Go") == 0
410 && !(target_flags_explicit & MASK_IEEE))
411 target_flags |= MASK_IEEE;
413 alpha_fprm = ALPHA_FPRM_NORM;
414 alpha_tp = ALPHA_TP_PROG;
415 alpha_fptm = ALPHA_FPTM_N;
417 if (TARGET_IEEE)
419 alpha_tp = ALPHA_TP_INSN;
420 alpha_fptm = ALPHA_FPTM_SU;
422 if (TARGET_IEEE_WITH_INEXACT)
424 alpha_tp = ALPHA_TP_INSN;
425 alpha_fptm = ALPHA_FPTM_SUI;
428 if (alpha_tp_string)
430 if (! strcmp (alpha_tp_string, "p"))
431 alpha_tp = ALPHA_TP_PROG;
432 else if (! strcmp (alpha_tp_string, "f"))
433 alpha_tp = ALPHA_TP_FUNC;
434 else if (! strcmp (alpha_tp_string, "i"))
435 alpha_tp = ALPHA_TP_INSN;
436 else
437 error ("bad value %qs for -mtrap-precision switch", alpha_tp_string);
440 if (alpha_fprm_string)
442 if (! strcmp (alpha_fprm_string, "n"))
443 alpha_fprm = ALPHA_FPRM_NORM;
444 else if (! strcmp (alpha_fprm_string, "m"))
445 alpha_fprm = ALPHA_FPRM_MINF;
446 else if (! strcmp (alpha_fprm_string, "c"))
447 alpha_fprm = ALPHA_FPRM_CHOP;
448 else if (! strcmp (alpha_fprm_string,"d"))
449 alpha_fprm = ALPHA_FPRM_DYN;
450 else
451 error ("bad value %qs for -mfp-rounding-mode switch",
452 alpha_fprm_string);
455 if (alpha_fptm_string)
457 if (strcmp (alpha_fptm_string, "n") == 0)
458 alpha_fptm = ALPHA_FPTM_N;
459 else if (strcmp (alpha_fptm_string, "u") == 0)
460 alpha_fptm = ALPHA_FPTM_U;
461 else if (strcmp (alpha_fptm_string, "su") == 0)
462 alpha_fptm = ALPHA_FPTM_SU;
463 else if (strcmp (alpha_fptm_string, "sui") == 0)
464 alpha_fptm = ALPHA_FPTM_SUI;
465 else
466 error ("bad value %qs for -mfp-trap-mode switch", alpha_fptm_string);
469 if (alpha_cpu_string)
471 for (i = 0; i < ct_size; i++)
472 if (! strcmp (alpha_cpu_string, cpu_table [i].name))
474 alpha_tune = alpha_cpu = cpu_table[i].processor;
475 line_size = cpu_table[i].line_size;
476 l1_size = cpu_table[i].l1_size;
477 l2_size = cpu_table[i].l2_size;
478 target_flags &= ~ (MASK_BWX | MASK_MAX | MASK_FIX | MASK_CIX);
479 target_flags |= cpu_table[i].flags;
480 break;
482 if (i == ct_size)
483 error ("bad value %qs for -mcpu switch", alpha_cpu_string);
486 if (alpha_tune_string)
488 for (i = 0; i < ct_size; i++)
489 if (! strcmp (alpha_tune_string, cpu_table [i].name))
491 alpha_tune = cpu_table[i].processor;
492 line_size = cpu_table[i].line_size;
493 l1_size = cpu_table[i].l1_size;
494 l2_size = cpu_table[i].l2_size;
495 break;
497 if (i == ct_size)
498 error ("bad value %qs for -mtune switch", alpha_tune_string);
501 if (line_size)
502 maybe_set_param_value (PARAM_L1_CACHE_LINE_SIZE, line_size,
503 global_options.x_param_values,
504 global_options_set.x_param_values);
505 if (l1_size)
506 maybe_set_param_value (PARAM_L1_CACHE_SIZE, l1_size,
507 global_options.x_param_values,
508 global_options_set.x_param_values);
509 if (l2_size)
510 maybe_set_param_value (PARAM_L2_CACHE_SIZE, l2_size,
511 global_options.x_param_values,
512 global_options_set.x_param_values);
514 /* Do some sanity checks on the above options. */
516 if ((alpha_fptm == ALPHA_FPTM_SU || alpha_fptm == ALPHA_FPTM_SUI)
517 && alpha_tp != ALPHA_TP_INSN && alpha_cpu != PROCESSOR_EV6)
519 warning (0, "fp software completion requires -mtrap-precision=i");
520 alpha_tp = ALPHA_TP_INSN;
523 if (alpha_cpu == PROCESSOR_EV6)
525 /* Except for EV6 pass 1 (not released), we always have precise
526 arithmetic traps. Which means we can do software completion
527 without minding trap shadows. */
528 alpha_tp = ALPHA_TP_PROG;
531 if (TARGET_FLOAT_VAX)
533 if (alpha_fprm == ALPHA_FPRM_MINF || alpha_fprm == ALPHA_FPRM_DYN)
535 warning (0, "rounding mode not supported for VAX floats");
536 alpha_fprm = ALPHA_FPRM_NORM;
538 if (alpha_fptm == ALPHA_FPTM_SUI)
540 warning (0, "trap mode not supported for VAX floats");
541 alpha_fptm = ALPHA_FPTM_SU;
543 if (target_flags_explicit & MASK_LONG_DOUBLE_128)
544 warning (0, "128-bit long double not supported for VAX floats");
545 target_flags &= ~MASK_LONG_DOUBLE_128;
549 char *end;
550 int lat;
552 if (!alpha_mlat_string)
553 alpha_mlat_string = "L1";
555 if (ISDIGIT ((unsigned char)alpha_mlat_string[0])
556 && (lat = strtol (alpha_mlat_string, &end, 10), *end == '\0'))
558 else if ((alpha_mlat_string[0] == 'L' || alpha_mlat_string[0] == 'l')
559 && ISDIGIT ((unsigned char)alpha_mlat_string[1])
560 && alpha_mlat_string[2] == '\0')
562 static int const cache_latency[][4] =
564 { 3, 30, -1 }, /* ev4 -- Bcache is a guess */
565 { 2, 12, 38 }, /* ev5 -- Bcache from PC164 LMbench numbers */
566 { 3, 12, 30 }, /* ev6 -- Bcache from DS20 LMbench. */
569 lat = alpha_mlat_string[1] - '0';
570 if (lat <= 0 || lat > 3 || cache_latency[alpha_tune][lat-1] == -1)
572 warning (0, "L%d cache latency unknown for %s",
573 lat, alpha_cpu_name[alpha_tune]);
574 lat = 3;
576 else
577 lat = cache_latency[alpha_tune][lat-1];
579 else if (! strcmp (alpha_mlat_string, "main"))
581 /* Most current memories have about 370ns latency. This is
582 a reasonable guess for a fast cpu. */
583 lat = 150;
585 else
587 warning (0, "bad value %qs for -mmemory-latency", alpha_mlat_string);
588 lat = 3;
591 alpha_memory_latency = lat;
594 /* Default the definition of "small data" to 8 bytes. */
595 if (!global_options_set.x_g_switch_value)
596 g_switch_value = 8;
598 /* Infer TARGET_SMALL_DATA from -fpic/-fPIC. */
599 if (flag_pic == 1)
600 target_flags |= MASK_SMALL_DATA;
601 else if (flag_pic == 2)
602 target_flags &= ~MASK_SMALL_DATA;
604 /* Align labels and loops for optimal branching. */
605 /* ??? Kludge these by not doing anything if we don't optimize. */
606 if (optimize > 0)
608 if (align_loops <= 0)
609 align_loops = 16;
610 if (align_jumps <= 0)
611 align_jumps = 16;
613 if (align_functions <= 0)
614 align_functions = 16;
616 /* Register variables and functions with the garbage collector. */
618 /* Set up function hooks. */
619 init_machine_status = alpha_init_machine_status;
621 /* Tell the compiler when we're using VAX floating point. */
622 if (TARGET_FLOAT_VAX)
624 REAL_MODE_FORMAT (SFmode) = &vax_f_format;
625 REAL_MODE_FORMAT (DFmode) = &vax_g_format;
626 REAL_MODE_FORMAT (TFmode) = NULL;
629 #ifdef TARGET_DEFAULT_LONG_DOUBLE_128
630 if (!(target_flags_explicit & MASK_LONG_DOUBLE_128))
631 target_flags |= MASK_LONG_DOUBLE_128;
632 #endif
634 /* This needs to be done at start up. It's convenient to do it here. */
635 register_pass (&handle_trap_shadows_info);
636 register_pass (&align_insns_info);
639 /* Returns 1 if VALUE is a mask that contains full bytes of zero or ones. */
642 zap_mask (HOST_WIDE_INT value)
644 int i;
646 for (i = 0; i < HOST_BITS_PER_WIDE_INT / HOST_BITS_PER_CHAR;
647 i++, value >>= 8)
648 if ((value & 0xff) != 0 && (value & 0xff) != 0xff)
649 return 0;
651 return 1;
654 /* Return true if OP is valid for a particular TLS relocation.
655 We are already guaranteed that OP is a CONST. */
658 tls_symbolic_operand_1 (rtx op, int size, int unspec)
660 op = XEXP (op, 0);
662 if (GET_CODE (op) != UNSPEC || XINT (op, 1) != unspec)
663 return 0;
664 op = XVECEXP (op, 0, 0);
666 if (GET_CODE (op) != SYMBOL_REF)
667 return 0;
669 switch (SYMBOL_REF_TLS_MODEL (op))
671 case TLS_MODEL_LOCAL_DYNAMIC:
672 return unspec == UNSPEC_DTPREL && size == alpha_tls_size;
673 case TLS_MODEL_INITIAL_EXEC:
674 return unspec == UNSPEC_TPREL && size == 64;
675 case TLS_MODEL_LOCAL_EXEC:
676 return unspec == UNSPEC_TPREL && size == alpha_tls_size;
677 default:
678 gcc_unreachable ();
682 /* Used by aligned_memory_operand and unaligned_memory_operand to
683 resolve what reload is going to do with OP if it's a register. */
686 resolve_reload_operand (rtx op)
688 if (reload_in_progress)
690 rtx tmp = op;
691 if (GET_CODE (tmp) == SUBREG)
692 tmp = SUBREG_REG (tmp);
693 if (REG_P (tmp)
694 && REGNO (tmp) >= FIRST_PSEUDO_REGISTER)
696 op = reg_equiv_memory_loc (REGNO (tmp));
697 if (op == 0)
698 return 0;
701 return op;
704 /* The scalar modes supported differs from the default check-what-c-supports
705 version in that sometimes TFmode is available even when long double
706 indicates only DFmode. */
708 static bool
709 alpha_scalar_mode_supported_p (enum machine_mode mode)
711 switch (mode)
713 case QImode:
714 case HImode:
715 case SImode:
716 case DImode:
717 case TImode: /* via optabs.c */
718 return true;
720 case SFmode:
721 case DFmode:
722 return true;
724 case TFmode:
725 return TARGET_HAS_XFLOATING_LIBS;
727 default:
728 return false;
732 /* Alpha implements a couple of integer vector mode operations when
733 TARGET_MAX is enabled. We do not check TARGET_MAX here, however,
734 which allows the vectorizer to operate on e.g. move instructions,
735 or when expand_vector_operations can do something useful. */
737 static bool
738 alpha_vector_mode_supported_p (enum machine_mode mode)
740 return mode == V8QImode || mode == V4HImode || mode == V2SImode;
743 /* Return 1 if this function can directly return via $26. */
746 direct_return (void)
748 return (TARGET_ABI_OSF
749 && reload_completed
750 && alpha_sa_size () == 0
751 && get_frame_size () == 0
752 && crtl->outgoing_args_size == 0
753 && crtl->args.pretend_args_size == 0);
756 /* Return the TLS model to use for SYMBOL. */
758 static enum tls_model
759 tls_symbolic_operand_type (rtx symbol)
761 enum tls_model model;
763 if (GET_CODE (symbol) != SYMBOL_REF)
764 return TLS_MODEL_NONE;
765 model = SYMBOL_REF_TLS_MODEL (symbol);
767 /* Local-exec with a 64-bit size is the same code as initial-exec. */
768 if (model == TLS_MODEL_LOCAL_EXEC && alpha_tls_size == 64)
769 model = TLS_MODEL_INITIAL_EXEC;
771 return model;
774 /* Return true if the function DECL will share the same GP as any
775 function in the current unit of translation. */
777 static bool
778 decl_has_samegp (const_tree decl)
780 /* Functions that are not local can be overridden, and thus may
781 not share the same gp. */
782 if (!(*targetm.binds_local_p) (decl))
783 return false;
785 /* If -msmall-data is in effect, assume that there is only one GP
786 for the module, and so any local symbol has this property. We
787 need explicit relocations to be able to enforce this for symbols
788 not defined in this unit of translation, however. */
789 if (TARGET_EXPLICIT_RELOCS && TARGET_SMALL_DATA)
790 return true;
792 /* Functions that are not external are defined in this UoT. */
793 /* ??? Irritatingly, static functions not yet emitted are still
794 marked "external". Apply this to non-static functions only. */
795 return !TREE_PUBLIC (decl) || !DECL_EXTERNAL (decl);
798 /* Return true if EXP should be placed in the small data section. */
800 static bool
801 alpha_in_small_data_p (const_tree exp)
803 /* We want to merge strings, so we never consider them small data. */
804 if (TREE_CODE (exp) == STRING_CST)
805 return false;
807 /* Functions are never in the small data area. Duh. */
808 if (TREE_CODE (exp) == FUNCTION_DECL)
809 return false;
811 if (TREE_CODE (exp) == VAR_DECL && DECL_SECTION_NAME (exp))
813 const char *section = DECL_SECTION_NAME (exp);
814 if (strcmp (section, ".sdata") == 0
815 || strcmp (section, ".sbss") == 0)
816 return true;
818 else
820 HOST_WIDE_INT size = int_size_in_bytes (TREE_TYPE (exp));
822 /* If this is an incomplete type with size 0, then we can't put it
823 in sdata because it might be too big when completed. */
824 if (size > 0 && size <= g_switch_value)
825 return true;
828 return false;
831 #if TARGET_ABI_OPEN_VMS
832 static bool
833 vms_valid_pointer_mode (enum machine_mode mode)
835 return (mode == SImode || mode == DImode);
838 static bool
839 alpha_linkage_symbol_p (const char *symname)
841 int symlen = strlen (symname);
843 if (symlen > 4)
844 return strcmp (&symname [symlen - 4], "..lk") == 0;
846 return false;
849 #define LINKAGE_SYMBOL_REF_P(X) \
850 ((GET_CODE (X) == SYMBOL_REF \
851 && alpha_linkage_symbol_p (XSTR (X, 0))) \
852 || (GET_CODE (X) == CONST \
853 && GET_CODE (XEXP (X, 0)) == PLUS \
854 && GET_CODE (XEXP (XEXP (X, 0), 0)) == SYMBOL_REF \
855 && alpha_linkage_symbol_p (XSTR (XEXP (XEXP (X, 0), 0), 0))))
856 #endif
858 /* legitimate_address_p recognizes an RTL expression that is a valid
859 memory address for an instruction. The MODE argument is the
860 machine mode for the MEM expression that wants to use this address.
862 For Alpha, we have either a constant address or the sum of a
863 register and a constant address, or just a register. For DImode,
864 any of those forms can be surrounded with an AND that clear the
865 low-order three bits; this is an "unaligned" access. */
867 static bool
868 alpha_legitimate_address_p (enum machine_mode mode, rtx x, bool strict)
870 /* If this is an ldq_u type address, discard the outer AND. */
871 if (mode == DImode
872 && GET_CODE (x) == AND
873 && CONST_INT_P (XEXP (x, 1))
874 && INTVAL (XEXP (x, 1)) == -8)
875 x = XEXP (x, 0);
877 /* Discard non-paradoxical subregs. */
878 if (GET_CODE (x) == SUBREG
879 && (GET_MODE_SIZE (GET_MODE (x))
880 < GET_MODE_SIZE (GET_MODE (SUBREG_REG (x)))))
881 x = SUBREG_REG (x);
883 /* Unadorned general registers are valid. */
884 if (REG_P (x)
885 && (strict
886 ? STRICT_REG_OK_FOR_BASE_P (x)
887 : NONSTRICT_REG_OK_FOR_BASE_P (x)))
888 return true;
890 /* Constant addresses (i.e. +/- 32k) are valid. */
891 if (CONSTANT_ADDRESS_P (x))
892 return true;
894 #if TARGET_ABI_OPEN_VMS
895 if (LINKAGE_SYMBOL_REF_P (x))
896 return true;
897 #endif
899 /* Register plus a small constant offset is valid. */
900 if (GET_CODE (x) == PLUS)
902 rtx ofs = XEXP (x, 1);
903 x = XEXP (x, 0);
905 /* Discard non-paradoxical subregs. */
906 if (GET_CODE (x) == SUBREG
907 && (GET_MODE_SIZE (GET_MODE (x))
908 < GET_MODE_SIZE (GET_MODE (SUBREG_REG (x)))))
909 x = SUBREG_REG (x);
911 if (REG_P (x))
913 if (! strict
914 && NONSTRICT_REG_OK_FP_BASE_P (x)
915 && CONST_INT_P (ofs))
916 return true;
917 if ((strict
918 ? STRICT_REG_OK_FOR_BASE_P (x)
919 : NONSTRICT_REG_OK_FOR_BASE_P (x))
920 && CONSTANT_ADDRESS_P (ofs))
921 return true;
925 /* If we're managing explicit relocations, LO_SUM is valid, as are small
926 data symbols. Avoid explicit relocations of modes larger than word
927 mode since i.e. $LC0+8($1) can fold around +/- 32k offset. */
928 else if (TARGET_EXPLICIT_RELOCS
929 && GET_MODE_SIZE (mode) <= UNITS_PER_WORD)
931 if (small_symbolic_operand (x, Pmode))
932 return true;
934 if (GET_CODE (x) == LO_SUM)
936 rtx ofs = XEXP (x, 1);
937 x = XEXP (x, 0);
939 /* Discard non-paradoxical subregs. */
940 if (GET_CODE (x) == SUBREG
941 && (GET_MODE_SIZE (GET_MODE (x))
942 < GET_MODE_SIZE (GET_MODE (SUBREG_REG (x)))))
943 x = SUBREG_REG (x);
945 /* Must have a valid base register. */
946 if (! (REG_P (x)
947 && (strict
948 ? STRICT_REG_OK_FOR_BASE_P (x)
949 : NONSTRICT_REG_OK_FOR_BASE_P (x))))
950 return false;
952 /* The symbol must be local. */
953 if (local_symbolic_operand (ofs, Pmode)
954 || dtp32_symbolic_operand (ofs, Pmode)
955 || tp32_symbolic_operand (ofs, Pmode))
956 return true;
960 return false;
963 /* Build the SYMBOL_REF for __tls_get_addr. */
965 static GTY(()) rtx tls_get_addr_libfunc;
967 static rtx
968 get_tls_get_addr (void)
970 if (!tls_get_addr_libfunc)
971 tls_get_addr_libfunc = init_one_libfunc ("__tls_get_addr");
972 return tls_get_addr_libfunc;
975 /* Try machine-dependent ways of modifying an illegitimate address
976 to be legitimate. If we find one, return the new, valid address. */
978 static rtx
979 alpha_legitimize_address_1 (rtx x, rtx scratch, enum machine_mode mode)
981 HOST_WIDE_INT addend;
983 /* If the address is (plus reg const_int) and the CONST_INT is not a
984 valid offset, compute the high part of the constant and add it to
985 the register. Then our address is (plus temp low-part-const). */
986 if (GET_CODE (x) == PLUS
987 && REG_P (XEXP (x, 0))
988 && CONST_INT_P (XEXP (x, 1))
989 && ! CONSTANT_ADDRESS_P (XEXP (x, 1)))
991 addend = INTVAL (XEXP (x, 1));
992 x = XEXP (x, 0);
993 goto split_addend;
996 /* If the address is (const (plus FOO const_int)), find the low-order
997 part of the CONST_INT. Then load FOO plus any high-order part of the
998 CONST_INT into a register. Our address is (plus reg low-part-const).
999 This is done to reduce the number of GOT entries. */
1000 if (can_create_pseudo_p ()
1001 && GET_CODE (x) == CONST
1002 && GET_CODE (XEXP (x, 0)) == PLUS
1003 && CONST_INT_P (XEXP (XEXP (x, 0), 1)))
1005 addend = INTVAL (XEXP (XEXP (x, 0), 1));
1006 x = force_reg (Pmode, XEXP (XEXP (x, 0), 0));
1007 goto split_addend;
1010 /* If we have a (plus reg const), emit the load as in (2), then add
1011 the two registers, and finally generate (plus reg low-part-const) as
1012 our address. */
1013 if (can_create_pseudo_p ()
1014 && GET_CODE (x) == PLUS
1015 && REG_P (XEXP (x, 0))
1016 && GET_CODE (XEXP (x, 1)) == CONST
1017 && GET_CODE (XEXP (XEXP (x, 1), 0)) == PLUS
1018 && CONST_INT_P (XEXP (XEXP (XEXP (x, 1), 0), 1)))
1020 addend = INTVAL (XEXP (XEXP (XEXP (x, 1), 0), 1));
1021 x = expand_simple_binop (Pmode, PLUS, XEXP (x, 0),
1022 XEXP (XEXP (XEXP (x, 1), 0), 0),
1023 NULL_RTX, 1, OPTAB_LIB_WIDEN);
1024 goto split_addend;
1027 /* If this is a local symbol, split the address into HIGH/LO_SUM parts.
1028 Avoid modes larger than word mode since i.e. $LC0+8($1) can fold
1029 around +/- 32k offset. */
1030 if (TARGET_EXPLICIT_RELOCS
1031 && GET_MODE_SIZE (mode) <= UNITS_PER_WORD
1032 && symbolic_operand (x, Pmode))
1034 rtx r0, r16, eqv, tga, tp, insn, dest, seq;
1036 switch (tls_symbolic_operand_type (x))
1038 case TLS_MODEL_NONE:
1039 break;
1041 case TLS_MODEL_GLOBAL_DYNAMIC:
1042 start_sequence ();
1044 r0 = gen_rtx_REG (Pmode, 0);
1045 r16 = gen_rtx_REG (Pmode, 16);
1046 tga = get_tls_get_addr ();
1047 dest = gen_reg_rtx (Pmode);
1048 seq = GEN_INT (alpha_next_sequence_number++);
1050 emit_insn (gen_movdi_er_tlsgd (r16, pic_offset_table_rtx, x, seq));
1051 insn = gen_call_value_osf_tlsgd (r0, tga, seq);
1052 insn = emit_call_insn (insn);
1053 RTL_CONST_CALL_P (insn) = 1;
1054 use_reg (&CALL_INSN_FUNCTION_USAGE (insn), r16);
1056 insn = get_insns ();
1057 end_sequence ();
1059 emit_libcall_block (insn, dest, r0, x);
1060 return dest;
1062 case TLS_MODEL_LOCAL_DYNAMIC:
1063 start_sequence ();
1065 r0 = gen_rtx_REG (Pmode, 0);
1066 r16 = gen_rtx_REG (Pmode, 16);
1067 tga = get_tls_get_addr ();
1068 scratch = gen_reg_rtx (Pmode);
1069 seq = GEN_INT (alpha_next_sequence_number++);
1071 emit_insn (gen_movdi_er_tlsldm (r16, pic_offset_table_rtx, seq));
1072 insn = gen_call_value_osf_tlsldm (r0, tga, seq);
1073 insn = emit_call_insn (insn);
1074 RTL_CONST_CALL_P (insn) = 1;
1075 use_reg (&CALL_INSN_FUNCTION_USAGE (insn), r16);
1077 insn = get_insns ();
1078 end_sequence ();
1080 eqv = gen_rtx_UNSPEC (Pmode, gen_rtvec (1, const0_rtx),
1081 UNSPEC_TLSLDM_CALL);
1082 emit_libcall_block (insn, scratch, r0, eqv);
1084 eqv = gen_rtx_UNSPEC (Pmode, gen_rtvec (1, x), UNSPEC_DTPREL);
1085 eqv = gen_rtx_CONST (Pmode, eqv);
1087 if (alpha_tls_size == 64)
1089 dest = gen_reg_rtx (Pmode);
1090 emit_insn (gen_rtx_SET (VOIDmode, dest, eqv));
1091 emit_insn (gen_adddi3 (dest, dest, scratch));
1092 return dest;
1094 if (alpha_tls_size == 32)
1096 insn = gen_rtx_HIGH (Pmode, eqv);
1097 insn = gen_rtx_PLUS (Pmode, scratch, insn);
1098 scratch = gen_reg_rtx (Pmode);
1099 emit_insn (gen_rtx_SET (VOIDmode, scratch, insn));
1101 return gen_rtx_LO_SUM (Pmode, scratch, eqv);
1103 case TLS_MODEL_INITIAL_EXEC:
1104 eqv = gen_rtx_UNSPEC (Pmode, gen_rtvec (1, x), UNSPEC_TPREL);
1105 eqv = gen_rtx_CONST (Pmode, eqv);
1106 tp = gen_reg_rtx (Pmode);
1107 scratch = gen_reg_rtx (Pmode);
1108 dest = gen_reg_rtx (Pmode);
1110 emit_insn (gen_get_thread_pointerdi (tp));
1111 emit_insn (gen_rtx_SET (VOIDmode, scratch, eqv));
1112 emit_insn (gen_adddi3 (dest, tp, scratch));
1113 return dest;
1115 case TLS_MODEL_LOCAL_EXEC:
1116 eqv = gen_rtx_UNSPEC (Pmode, gen_rtvec (1, x), UNSPEC_TPREL);
1117 eqv = gen_rtx_CONST (Pmode, eqv);
1118 tp = gen_reg_rtx (Pmode);
1120 emit_insn (gen_get_thread_pointerdi (tp));
1121 if (alpha_tls_size == 32)
1123 insn = gen_rtx_HIGH (Pmode, eqv);
1124 insn = gen_rtx_PLUS (Pmode, tp, insn);
1125 tp = gen_reg_rtx (Pmode);
1126 emit_insn (gen_rtx_SET (VOIDmode, tp, insn));
1128 return gen_rtx_LO_SUM (Pmode, tp, eqv);
1130 default:
1131 gcc_unreachable ();
1134 if (local_symbolic_operand (x, Pmode))
1136 if (small_symbolic_operand (x, Pmode))
1137 return x;
1138 else
1140 if (can_create_pseudo_p ())
1141 scratch = gen_reg_rtx (Pmode);
1142 emit_insn (gen_rtx_SET (VOIDmode, scratch,
1143 gen_rtx_HIGH (Pmode, x)));
1144 return gen_rtx_LO_SUM (Pmode, scratch, x);
1149 return NULL;
1151 split_addend:
1153 HOST_WIDE_INT low, high;
1155 low = ((addend & 0xffff) ^ 0x8000) - 0x8000;
1156 addend -= low;
1157 high = ((addend & 0xffffffff) ^ 0x80000000) - 0x80000000;
1158 addend -= high;
1160 if (addend)
1161 x = expand_simple_binop (Pmode, PLUS, x, GEN_INT (addend),
1162 (!can_create_pseudo_p () ? scratch : NULL_RTX),
1163 1, OPTAB_LIB_WIDEN);
1164 if (high)
1165 x = expand_simple_binop (Pmode, PLUS, x, GEN_INT (high),
1166 (!can_create_pseudo_p () ? scratch : NULL_RTX),
1167 1, OPTAB_LIB_WIDEN);
1169 return plus_constant (Pmode, x, low);
1174 /* Try machine-dependent ways of modifying an illegitimate address
1175 to be legitimate. Return X or the new, valid address. */
1177 static rtx
1178 alpha_legitimize_address (rtx x, rtx oldx ATTRIBUTE_UNUSED,
1179 enum machine_mode mode)
1181 rtx new_x = alpha_legitimize_address_1 (x, NULL_RTX, mode);
1182 return new_x ? new_x : x;
1185 /* Return true if ADDR has an effect that depends on the machine mode it
1186 is used for. On the Alpha this is true only for the unaligned modes.
1187 We can simplify the test since we know that the address must be valid. */
1189 static bool
1190 alpha_mode_dependent_address_p (const_rtx addr,
1191 addr_space_t as ATTRIBUTE_UNUSED)
1193 return GET_CODE (addr) == AND;
1196 /* Primarily this is required for TLS symbols, but given that our move
1197 patterns *ought* to be able to handle any symbol at any time, we
1198 should never be spilling symbolic operands to the constant pool, ever. */
1200 static bool
1201 alpha_cannot_force_const_mem (enum machine_mode mode ATTRIBUTE_UNUSED, rtx x)
1203 enum rtx_code code = GET_CODE (x);
1204 return code == SYMBOL_REF || code == LABEL_REF || code == CONST;
1207 /* We do not allow indirect calls to be optimized into sibling calls, nor
1208 can we allow a call to a function with a different GP to be optimized
1209 into a sibcall. */
1211 static bool
1212 alpha_function_ok_for_sibcall (tree decl, tree exp ATTRIBUTE_UNUSED)
1214 /* Can't do indirect tail calls, since we don't know if the target
1215 uses the same GP. */
1216 if (!decl)
1217 return false;
1219 /* Otherwise, we can make a tail call if the target function shares
1220 the same GP. */
1221 return decl_has_samegp (decl);
1225 some_small_symbolic_operand_int (rtx *px, void *data ATTRIBUTE_UNUSED)
1227 rtx x = *px;
1229 /* Don't re-split. */
1230 if (GET_CODE (x) == LO_SUM)
1231 return -1;
1233 return small_symbolic_operand (x, Pmode) != 0;
1236 static int
1237 split_small_symbolic_operand_1 (rtx *px, void *data ATTRIBUTE_UNUSED)
1239 rtx x = *px;
1241 /* Don't re-split. */
1242 if (GET_CODE (x) == LO_SUM)
1243 return -1;
1245 if (small_symbolic_operand (x, Pmode))
1247 x = gen_rtx_LO_SUM (Pmode, pic_offset_table_rtx, x);
1248 *px = x;
1249 return -1;
1252 return 0;
1256 split_small_symbolic_operand (rtx x)
1258 x = copy_insn (x);
1259 for_each_rtx (&x, split_small_symbolic_operand_1, NULL);
1260 return x;
1263 /* Indicate that INSN cannot be duplicated. This is true for any insn
1264 that we've marked with gpdisp relocs, since those have to stay in
1265 1-1 correspondence with one another.
1267 Technically we could copy them if we could set up a mapping from one
1268 sequence number to another, across the set of insns to be duplicated.
1269 This seems overly complicated and error-prone since interblock motion
1270 from sched-ebb could move one of the pair of insns to a different block.
1272 Also cannot allow jsr insns to be duplicated. If they throw exceptions,
1273 then they'll be in a different block from their ldgp. Which could lead
1274 the bb reorder code to think that it would be ok to copy just the block
1275 containing the call and branch to the block containing the ldgp. */
1277 static bool
1278 alpha_cannot_copy_insn_p (rtx insn)
1280 if (!reload_completed || !TARGET_EXPLICIT_RELOCS)
1281 return false;
1282 if (recog_memoized (insn) >= 0)
1283 return get_attr_cannot_copy (insn);
1284 else
1285 return false;
1289 /* Try a machine-dependent way of reloading an illegitimate address
1290 operand. If we find one, push the reload and return the new rtx. */
1293 alpha_legitimize_reload_address (rtx x,
1294 enum machine_mode mode ATTRIBUTE_UNUSED,
1295 int opnum, int type,
1296 int ind_levels ATTRIBUTE_UNUSED)
1298 /* We must recognize output that we have already generated ourselves. */
1299 if (GET_CODE (x) == PLUS
1300 && GET_CODE (XEXP (x, 0)) == PLUS
1301 && REG_P (XEXP (XEXP (x, 0), 0))
1302 && CONST_INT_P (XEXP (XEXP (x, 0), 1))
1303 && CONST_INT_P (XEXP (x, 1)))
1305 push_reload (XEXP (x, 0), NULL_RTX, &XEXP (x, 0), NULL,
1306 BASE_REG_CLASS, GET_MODE (x), VOIDmode, 0, 0,
1307 opnum, (enum reload_type) type);
1308 return x;
1311 /* We wish to handle large displacements off a base register by
1312 splitting the addend across an ldah and the mem insn. This
1313 cuts number of extra insns needed from 3 to 1. */
1314 if (GET_CODE (x) == PLUS
1315 && REG_P (XEXP (x, 0))
1316 && REGNO (XEXP (x, 0)) < FIRST_PSEUDO_REGISTER
1317 && REGNO_OK_FOR_BASE_P (REGNO (XEXP (x, 0)))
1318 && GET_CODE (XEXP (x, 1)) == CONST_INT)
1320 HOST_WIDE_INT val = INTVAL (XEXP (x, 1));
1321 HOST_WIDE_INT low = ((val & 0xffff) ^ 0x8000) - 0x8000;
1322 HOST_WIDE_INT high
1323 = (((val - low) & 0xffffffff) ^ 0x80000000) - 0x80000000;
1325 /* Check for 32-bit overflow. */
1326 if (high + low != val)
1327 return NULL_RTX;
1329 /* Reload the high part into a base reg; leave the low part
1330 in the mem directly. */
1331 x = gen_rtx_PLUS (GET_MODE (x),
1332 gen_rtx_PLUS (GET_MODE (x), XEXP (x, 0),
1333 GEN_INT (high)),
1334 GEN_INT (low));
1336 push_reload (XEXP (x, 0), NULL_RTX, &XEXP (x, 0), NULL,
1337 BASE_REG_CLASS, GET_MODE (x), VOIDmode, 0, 0,
1338 opnum, (enum reload_type) type);
1339 return x;
1342 return NULL_RTX;
1345 /* Compute a (partial) cost for rtx X. Return true if the complete
1346 cost has been computed, and false if subexpressions should be
1347 scanned. In either case, *TOTAL contains the cost result. */
1349 static bool
1350 alpha_rtx_costs (rtx x, int code, int outer_code, int opno, int *total,
1351 bool speed)
1353 enum machine_mode mode = GET_MODE (x);
1354 bool float_mode_p = FLOAT_MODE_P (mode);
1355 const struct alpha_rtx_cost_data *cost_data;
1357 if (!speed)
1358 cost_data = &alpha_rtx_cost_size;
1359 else
1360 cost_data = &alpha_rtx_cost_data[alpha_tune];
1362 switch (code)
1364 case CONST_INT:
1365 /* If this is an 8-bit constant, return zero since it can be used
1366 nearly anywhere with no cost. If it is a valid operand for an
1367 ADD or AND, likewise return 0 if we know it will be used in that
1368 context. Otherwise, return 2 since it might be used there later.
1369 All other constants take at least two insns. */
1370 if (INTVAL (x) >= 0 && INTVAL (x) < 256)
1372 *total = 0;
1373 return true;
1375 /* FALLTHRU */
1377 case CONST_DOUBLE:
1378 if (x == CONST0_RTX (mode))
1379 *total = 0;
1380 else if ((outer_code == PLUS && add_operand (x, VOIDmode))
1381 || (outer_code == AND && and_operand (x, VOIDmode)))
1382 *total = 0;
1383 else if (add_operand (x, VOIDmode) || and_operand (x, VOIDmode))
1384 *total = 2;
1385 else
1386 *total = COSTS_N_INSNS (2);
1387 return true;
1389 case CONST:
1390 case SYMBOL_REF:
1391 case LABEL_REF:
1392 if (TARGET_EXPLICIT_RELOCS && small_symbolic_operand (x, VOIDmode))
1393 *total = COSTS_N_INSNS (outer_code != MEM);
1394 else if (TARGET_EXPLICIT_RELOCS && local_symbolic_operand (x, VOIDmode))
1395 *total = COSTS_N_INSNS (1 + (outer_code != MEM));
1396 else if (tls_symbolic_operand_type (x))
1397 /* Estimate of cost for call_pal rduniq. */
1398 /* ??? How many insns do we emit here? More than one... */
1399 *total = COSTS_N_INSNS (15);
1400 else
1401 /* Otherwise we do a load from the GOT. */
1402 *total = COSTS_N_INSNS (!speed ? 1 : alpha_memory_latency);
1403 return true;
1405 case HIGH:
1406 /* This is effectively an add_operand. */
1407 *total = 2;
1408 return true;
1410 case PLUS:
1411 case MINUS:
1412 if (float_mode_p)
1413 *total = cost_data->fp_add;
1414 else if (GET_CODE (XEXP (x, 0)) == MULT
1415 && const48_operand (XEXP (XEXP (x, 0), 1), VOIDmode))
1417 *total = (rtx_cost (XEXP (XEXP (x, 0), 0),
1418 (enum rtx_code) outer_code, opno, speed)
1419 + rtx_cost (XEXP (x, 1),
1420 (enum rtx_code) outer_code, opno, speed)
1421 + COSTS_N_INSNS (1));
1422 return true;
1424 return false;
1426 case MULT:
1427 if (float_mode_p)
1428 *total = cost_data->fp_mult;
1429 else if (mode == DImode)
1430 *total = cost_data->int_mult_di;
1431 else
1432 *total = cost_data->int_mult_si;
1433 return false;
1435 case ASHIFT:
1436 if (CONST_INT_P (XEXP (x, 1))
1437 && INTVAL (XEXP (x, 1)) <= 3)
1439 *total = COSTS_N_INSNS (1);
1440 return false;
1442 /* FALLTHRU */
1444 case ASHIFTRT:
1445 case LSHIFTRT:
1446 *total = cost_data->int_shift;
1447 return false;
1449 case IF_THEN_ELSE:
1450 if (float_mode_p)
1451 *total = cost_data->fp_add;
1452 else
1453 *total = cost_data->int_cmov;
1454 return false;
1456 case DIV:
1457 case UDIV:
1458 case MOD:
1459 case UMOD:
1460 if (!float_mode_p)
1461 *total = cost_data->int_div;
1462 else if (mode == SFmode)
1463 *total = cost_data->fp_div_sf;
1464 else
1465 *total = cost_data->fp_div_df;
1466 return false;
1468 case MEM:
1469 *total = COSTS_N_INSNS (!speed ? 1 : alpha_memory_latency);
1470 return true;
1472 case NEG:
1473 if (! float_mode_p)
1475 *total = COSTS_N_INSNS (1);
1476 return false;
1478 /* FALLTHRU */
1480 case ABS:
1481 if (! float_mode_p)
1483 *total = COSTS_N_INSNS (1) + cost_data->int_cmov;
1484 return false;
1486 /* FALLTHRU */
1488 case FLOAT:
1489 case UNSIGNED_FLOAT:
1490 case FIX:
1491 case UNSIGNED_FIX:
1492 case FLOAT_TRUNCATE:
1493 *total = cost_data->fp_add;
1494 return false;
1496 case FLOAT_EXTEND:
1497 if (MEM_P (XEXP (x, 0)))
1498 *total = 0;
1499 else
1500 *total = cost_data->fp_add;
1501 return false;
1503 default:
1504 return false;
1508 /* REF is an alignable memory location. Place an aligned SImode
1509 reference into *PALIGNED_MEM and the number of bits to shift into
1510 *PBITNUM. SCRATCH is a free register for use in reloading out
1511 of range stack slots. */
1513 void
1514 get_aligned_mem (rtx ref, rtx *paligned_mem, rtx *pbitnum)
1516 rtx base;
1517 HOST_WIDE_INT disp, offset;
1519 gcc_assert (MEM_P (ref));
1521 if (reload_in_progress
1522 && ! memory_address_p (GET_MODE (ref), XEXP (ref, 0)))
1524 base = find_replacement (&XEXP (ref, 0));
1525 gcc_assert (memory_address_p (GET_MODE (ref), base));
1527 else
1528 base = XEXP (ref, 0);
1530 if (GET_CODE (base) == PLUS)
1531 disp = INTVAL (XEXP (base, 1)), base = XEXP (base, 0);
1532 else
1533 disp = 0;
1535 /* Find the byte offset within an aligned word. If the memory itself is
1536 claimed to be aligned, believe it. Otherwise, aligned_memory_operand
1537 will have examined the base register and determined it is aligned, and
1538 thus displacements from it are naturally alignable. */
1539 if (MEM_ALIGN (ref) >= 32)
1540 offset = 0;
1541 else
1542 offset = disp & 3;
1544 /* The location should not cross aligned word boundary. */
1545 gcc_assert (offset + GET_MODE_SIZE (GET_MODE (ref))
1546 <= GET_MODE_SIZE (SImode));
1548 /* Access the entire aligned word. */
1549 *paligned_mem = widen_memory_access (ref, SImode, -offset);
1551 /* Convert the byte offset within the word to a bit offset. */
1552 offset *= BITS_PER_UNIT;
1553 *pbitnum = GEN_INT (offset);
1556 /* Similar, but just get the address. Handle the two reload cases.
1557 Add EXTRA_OFFSET to the address we return. */
1560 get_unaligned_address (rtx ref)
1562 rtx base;
1563 HOST_WIDE_INT offset = 0;
1565 gcc_assert (MEM_P (ref));
1567 if (reload_in_progress
1568 && ! memory_address_p (GET_MODE (ref), XEXP (ref, 0)))
1570 base = find_replacement (&XEXP (ref, 0));
1572 gcc_assert (memory_address_p (GET_MODE (ref), base));
1574 else
1575 base = XEXP (ref, 0);
1577 if (GET_CODE (base) == PLUS)
1578 offset += INTVAL (XEXP (base, 1)), base = XEXP (base, 0);
1580 return plus_constant (Pmode, base, offset);
1583 /* Compute a value X, such that X & 7 == (ADDR + OFS) & 7.
1584 X is always returned in a register. */
1587 get_unaligned_offset (rtx addr, HOST_WIDE_INT ofs)
1589 if (GET_CODE (addr) == PLUS)
1591 ofs += INTVAL (XEXP (addr, 1));
1592 addr = XEXP (addr, 0);
1595 return expand_simple_binop (Pmode, PLUS, addr, GEN_INT (ofs & 7),
1596 NULL_RTX, 1, OPTAB_LIB_WIDEN);
1599 /* On the Alpha, all (non-symbolic) constants except zero go into
1600 a floating-point register via memory. Note that we cannot
1601 return anything that is not a subset of RCLASS, and that some
1602 symbolic constants cannot be dropped to memory. */
1604 enum reg_class
1605 alpha_preferred_reload_class(rtx x, enum reg_class rclass)
1607 /* Zero is present in any register class. */
1608 if (x == CONST0_RTX (GET_MODE (x)))
1609 return rclass;
1611 /* These sorts of constants we can easily drop to memory. */
1612 if (CONST_INT_P (x)
1613 || GET_CODE (x) == CONST_DOUBLE
1614 || GET_CODE (x) == CONST_VECTOR)
1616 if (rclass == FLOAT_REGS)
1617 return NO_REGS;
1618 if (rclass == ALL_REGS)
1619 return GENERAL_REGS;
1620 return rclass;
1623 /* All other kinds of constants should not (and in the case of HIGH
1624 cannot) be dropped to memory -- instead we use a GENERAL_REGS
1625 secondary reload. */
1626 if (CONSTANT_P (x))
1627 return (rclass == ALL_REGS ? GENERAL_REGS : rclass);
1629 return rclass;
1632 /* Inform reload about cases where moving X with a mode MODE to a register in
1633 RCLASS requires an extra scratch or immediate register. Return the class
1634 needed for the immediate register. */
1636 static reg_class_t
1637 alpha_secondary_reload (bool in_p, rtx x, reg_class_t rclass_i,
1638 enum machine_mode mode, secondary_reload_info *sri)
1640 enum reg_class rclass = (enum reg_class) rclass_i;
1642 /* Loading and storing HImode or QImode values to and from memory
1643 usually requires a scratch register. */
1644 if (!TARGET_BWX && (mode == QImode || mode == HImode || mode == CQImode))
1646 if (any_memory_operand (x, mode))
1648 if (in_p)
1650 if (!aligned_memory_operand (x, mode))
1651 sri->icode = direct_optab_handler (reload_in_optab, mode);
1653 else
1654 sri->icode = direct_optab_handler (reload_out_optab, mode);
1655 return NO_REGS;
1659 /* We also cannot do integral arithmetic into FP regs, as might result
1660 from register elimination into a DImode fp register. */
1661 if (rclass == FLOAT_REGS)
1663 if (MEM_P (x) && GET_CODE (XEXP (x, 0)) == AND)
1664 return GENERAL_REGS;
1665 if (in_p && INTEGRAL_MODE_P (mode)
1666 && !MEM_P (x) && !REG_P (x) && !CONST_INT_P (x))
1667 return GENERAL_REGS;
1670 return NO_REGS;
1673 /* Subfunction of the following function. Update the flags of any MEM
1674 found in part of X. */
1676 static int
1677 alpha_set_memflags_1 (rtx *xp, void *data)
1679 rtx x = *xp, orig = (rtx) data;
1681 if (!MEM_P (x))
1682 return 0;
1684 MEM_VOLATILE_P (x) = MEM_VOLATILE_P (orig);
1685 MEM_NOTRAP_P (x) = MEM_NOTRAP_P (orig);
1686 MEM_READONLY_P (x) = MEM_READONLY_P (orig);
1688 /* Sadly, we cannot use alias sets because the extra aliasing
1689 produced by the AND interferes. Given that two-byte quantities
1690 are the only thing we would be able to differentiate anyway,
1691 there does not seem to be any point in convoluting the early
1692 out of the alias check. */
1694 return -1;
1697 /* Given SEQ, which is an INSN list, look for any MEMs in either
1698 a SET_DEST or a SET_SRC and copy the in-struct, unchanging, and
1699 volatile flags from REF into each of the MEMs found. If REF is not
1700 a MEM, don't do anything. */
1702 void
1703 alpha_set_memflags (rtx seq, rtx ref)
1705 rtx insn;
1707 if (!MEM_P (ref))
1708 return;
1710 /* This is only called from alpha.md, after having had something
1711 generated from one of the insn patterns. So if everything is
1712 zero, the pattern is already up-to-date. */
1713 if (!MEM_VOLATILE_P (ref)
1714 && !MEM_NOTRAP_P (ref)
1715 && !MEM_READONLY_P (ref))
1716 return;
1718 for (insn = seq; insn; insn = NEXT_INSN (insn))
1719 if (INSN_P (insn))
1720 for_each_rtx (&PATTERN (insn), alpha_set_memflags_1, (void *) ref);
1721 else
1722 gcc_unreachable ();
1725 static rtx alpha_emit_set_const (rtx, enum machine_mode, HOST_WIDE_INT,
1726 int, bool);
1728 /* Internal routine for alpha_emit_set_const to check for N or below insns.
1729 If NO_OUTPUT is true, then we only check to see if N insns are possible,
1730 and return pc_rtx if successful. */
1732 static rtx
1733 alpha_emit_set_const_1 (rtx target, enum machine_mode mode,
1734 HOST_WIDE_INT c, int n, bool no_output)
1736 HOST_WIDE_INT new_const;
1737 int i, bits;
1738 /* Use a pseudo if highly optimizing and still generating RTL. */
1739 rtx subtarget
1740 = (flag_expensive_optimizations && can_create_pseudo_p () ? 0 : target);
1741 rtx temp, insn;
1743 /* If this is a sign-extended 32-bit constant, we can do this in at most
1744 three insns, so do it if we have enough insns left. We always have
1745 a sign-extended 32-bit constant when compiling on a narrow machine. */
1747 if (HOST_BITS_PER_WIDE_INT != 64
1748 || c >> 31 == -1 || c >> 31 == 0)
1750 HOST_WIDE_INT low = ((c & 0xffff) ^ 0x8000) - 0x8000;
1751 HOST_WIDE_INT tmp1 = c - low;
1752 HOST_WIDE_INT high = (((tmp1 >> 16) & 0xffff) ^ 0x8000) - 0x8000;
1753 HOST_WIDE_INT extra = 0;
1755 /* If HIGH will be interpreted as negative but the constant is
1756 positive, we must adjust it to do two ldha insns. */
1758 if ((high & 0x8000) != 0 && c >= 0)
1760 extra = 0x4000;
1761 tmp1 -= 0x40000000;
1762 high = ((tmp1 >> 16) & 0xffff) - 2 * ((tmp1 >> 16) & 0x8000);
1765 if (c == low || (low == 0 && extra == 0))
1767 /* We used to use copy_to_suggested_reg (GEN_INT (c), target, mode)
1768 but that meant that we can't handle INT_MIN on 32-bit machines
1769 (like NT/Alpha), because we recurse indefinitely through
1770 emit_move_insn to gen_movdi. So instead, since we know exactly
1771 what we want, create it explicitly. */
1773 if (no_output)
1774 return pc_rtx;
1775 if (target == NULL)
1776 target = gen_reg_rtx (mode);
1777 emit_insn (gen_rtx_SET (VOIDmode, target, GEN_INT (c)));
1778 return target;
1780 else if (n >= 2 + (extra != 0))
1782 if (no_output)
1783 return pc_rtx;
1784 if (!can_create_pseudo_p ())
1786 emit_insn (gen_rtx_SET (VOIDmode, target, GEN_INT (high << 16)));
1787 temp = target;
1789 else
1790 temp = copy_to_suggested_reg (GEN_INT (high << 16),
1791 subtarget, mode);
1793 /* As of 2002-02-23, addsi3 is only available when not optimizing.
1794 This means that if we go through expand_binop, we'll try to
1795 generate extensions, etc, which will require new pseudos, which
1796 will fail during some split phases. The SImode add patterns
1797 still exist, but are not named. So build the insns by hand. */
1799 if (extra != 0)
1801 if (! subtarget)
1802 subtarget = gen_reg_rtx (mode);
1803 insn = gen_rtx_PLUS (mode, temp, GEN_INT (extra << 16));
1804 insn = gen_rtx_SET (VOIDmode, subtarget, insn);
1805 emit_insn (insn);
1806 temp = subtarget;
1809 if (target == NULL)
1810 target = gen_reg_rtx (mode);
1811 insn = gen_rtx_PLUS (mode, temp, GEN_INT (low));
1812 insn = gen_rtx_SET (VOIDmode, target, insn);
1813 emit_insn (insn);
1814 return target;
1818 /* If we couldn't do it that way, try some other methods. But if we have
1819 no instructions left, don't bother. Likewise, if this is SImode and
1820 we can't make pseudos, we can't do anything since the expand_binop
1821 and expand_unop calls will widen and try to make pseudos. */
1823 if (n == 1 || (mode == SImode && !can_create_pseudo_p ()))
1824 return 0;
1826 /* Next, see if we can load a related constant and then shift and possibly
1827 negate it to get the constant we want. Try this once each increasing
1828 numbers of insns. */
1830 for (i = 1; i < n; i++)
1832 /* First, see if minus some low bits, we've an easy load of
1833 high bits. */
1835 new_const = ((c & 0xffff) ^ 0x8000) - 0x8000;
1836 if (new_const != 0)
1838 temp = alpha_emit_set_const (subtarget, mode, c - new_const, i, no_output);
1839 if (temp)
1841 if (no_output)
1842 return temp;
1843 return expand_binop (mode, add_optab, temp, GEN_INT (new_const),
1844 target, 0, OPTAB_WIDEN);
1848 /* Next try complementing. */
1849 temp = alpha_emit_set_const (subtarget, mode, ~c, i, no_output);
1850 if (temp)
1852 if (no_output)
1853 return temp;
1854 return expand_unop (mode, one_cmpl_optab, temp, target, 0);
1857 /* Next try to form a constant and do a left shift. We can do this
1858 if some low-order bits are zero; the exact_log2 call below tells
1859 us that information. The bits we are shifting out could be any
1860 value, but here we'll just try the 0- and sign-extended forms of
1861 the constant. To try to increase the chance of having the same
1862 constant in more than one insn, start at the highest number of
1863 bits to shift, but try all possibilities in case a ZAPNOT will
1864 be useful. */
1866 bits = exact_log2 (c & -c);
1867 if (bits > 0)
1868 for (; bits > 0; bits--)
1870 new_const = c >> bits;
1871 temp = alpha_emit_set_const (subtarget, mode, new_const, i, no_output);
1872 if (!temp && c < 0)
1874 new_const = (unsigned HOST_WIDE_INT)c >> bits;
1875 temp = alpha_emit_set_const (subtarget, mode, new_const,
1876 i, no_output);
1878 if (temp)
1880 if (no_output)
1881 return temp;
1882 return expand_binop (mode, ashl_optab, temp, GEN_INT (bits),
1883 target, 0, OPTAB_WIDEN);
1887 /* Now try high-order zero bits. Here we try the shifted-in bits as
1888 all zero and all ones. Be careful to avoid shifting outside the
1889 mode and to avoid shifting outside the host wide int size. */
1890 /* On narrow hosts, don't shift a 1 into the high bit, since we'll
1891 confuse the recursive call and set all of the high 32 bits. */
1893 bits = (MIN (HOST_BITS_PER_WIDE_INT, GET_MODE_SIZE (mode) * 8)
1894 - floor_log2 (c) - 1 - (HOST_BITS_PER_WIDE_INT < 64));
1895 if (bits > 0)
1896 for (; bits > 0; bits--)
1898 new_const = c << bits;
1899 temp = alpha_emit_set_const (subtarget, mode, new_const, i, no_output);
1900 if (!temp)
1902 new_const = (c << bits) | (((HOST_WIDE_INT) 1 << bits) - 1);
1903 temp = alpha_emit_set_const (subtarget, mode, new_const,
1904 i, no_output);
1906 if (temp)
1908 if (no_output)
1909 return temp;
1910 return expand_binop (mode, lshr_optab, temp, GEN_INT (bits),
1911 target, 1, OPTAB_WIDEN);
1915 /* Now try high-order 1 bits. We get that with a sign-extension.
1916 But one bit isn't enough here. Be careful to avoid shifting outside
1917 the mode and to avoid shifting outside the host wide int size. */
1919 bits = (MIN (HOST_BITS_PER_WIDE_INT, GET_MODE_SIZE (mode) * 8)
1920 - floor_log2 (~ c) - 2);
1921 if (bits > 0)
1922 for (; bits > 0; bits--)
1924 new_const = c << bits;
1925 temp = alpha_emit_set_const (subtarget, mode, new_const, i, no_output);
1926 if (!temp)
1928 new_const = (c << bits) | (((HOST_WIDE_INT) 1 << bits) - 1);
1929 temp = alpha_emit_set_const (subtarget, mode, new_const,
1930 i, no_output);
1932 if (temp)
1934 if (no_output)
1935 return temp;
1936 return expand_binop (mode, ashr_optab, temp, GEN_INT (bits),
1937 target, 0, OPTAB_WIDEN);
1942 #if HOST_BITS_PER_WIDE_INT == 64
1943 /* Finally, see if can load a value into the target that is the same as the
1944 constant except that all bytes that are 0 are changed to be 0xff. If we
1945 can, then we can do a ZAPNOT to obtain the desired constant. */
1947 new_const = c;
1948 for (i = 0; i < 64; i += 8)
1949 if ((new_const & ((HOST_WIDE_INT) 0xff << i)) == 0)
1950 new_const |= (HOST_WIDE_INT) 0xff << i;
1952 /* We are only called for SImode and DImode. If this is SImode, ensure that
1953 we are sign extended to a full word. */
1955 if (mode == SImode)
1956 new_const = ((new_const & 0xffffffff) ^ 0x80000000) - 0x80000000;
1958 if (new_const != c)
1960 temp = alpha_emit_set_const (subtarget, mode, new_const, n - 1, no_output);
1961 if (temp)
1963 if (no_output)
1964 return temp;
1965 return expand_binop (mode, and_optab, temp, GEN_INT (c | ~ new_const),
1966 target, 0, OPTAB_WIDEN);
1969 #endif
1971 return 0;
1974 /* Try to output insns to set TARGET equal to the constant C if it can be
1975 done in less than N insns. Do all computations in MODE. Returns the place
1976 where the output has been placed if it can be done and the insns have been
1977 emitted. If it would take more than N insns, zero is returned and no
1978 insns and emitted. */
1980 static rtx
1981 alpha_emit_set_const (rtx target, enum machine_mode mode,
1982 HOST_WIDE_INT c, int n, bool no_output)
1984 enum machine_mode orig_mode = mode;
1985 rtx orig_target = target;
1986 rtx result = 0;
1987 int i;
1989 /* If we can't make any pseudos, TARGET is an SImode hard register, we
1990 can't load this constant in one insn, do this in DImode. */
1991 if (!can_create_pseudo_p () && mode == SImode
1992 && REG_P (target) && REGNO (target) < FIRST_PSEUDO_REGISTER)
1994 result = alpha_emit_set_const_1 (target, mode, c, 1, no_output);
1995 if (result)
1996 return result;
1998 target = no_output ? NULL : gen_lowpart (DImode, target);
1999 mode = DImode;
2001 else if (mode == V8QImode || mode == V4HImode || mode == V2SImode)
2003 target = no_output ? NULL : gen_lowpart (DImode, target);
2004 mode = DImode;
2007 /* Try 1 insn, then 2, then up to N. */
2008 for (i = 1; i <= n; i++)
2010 result = alpha_emit_set_const_1 (target, mode, c, i, no_output);
2011 if (result)
2013 rtx insn, set;
2015 if (no_output)
2016 return result;
2018 insn = get_last_insn ();
2019 set = single_set (insn);
2020 if (! CONSTANT_P (SET_SRC (set)))
2021 set_unique_reg_note (get_last_insn (), REG_EQUAL, GEN_INT (c));
2022 break;
2026 /* Allow for the case where we changed the mode of TARGET. */
2027 if (result)
2029 if (result == target)
2030 result = orig_target;
2031 else if (mode != orig_mode)
2032 result = gen_lowpart (orig_mode, result);
2035 return result;
2038 /* Having failed to find a 3 insn sequence in alpha_emit_set_const,
2039 fall back to a straight forward decomposition. We do this to avoid
2040 exponential run times encountered when looking for longer sequences
2041 with alpha_emit_set_const. */
2043 static rtx
2044 alpha_emit_set_long_const (rtx target, HOST_WIDE_INT c1, HOST_WIDE_INT c2)
2046 HOST_WIDE_INT d1, d2, d3, d4;
2048 /* Decompose the entire word */
2049 #if HOST_BITS_PER_WIDE_INT >= 64
2050 gcc_assert (c2 == -(c1 < 0));
2051 d1 = ((c1 & 0xffff) ^ 0x8000) - 0x8000;
2052 c1 -= d1;
2053 d2 = ((c1 & 0xffffffff) ^ 0x80000000) - 0x80000000;
2054 c1 = (c1 - d2) >> 32;
2055 d3 = ((c1 & 0xffff) ^ 0x8000) - 0x8000;
2056 c1 -= d3;
2057 d4 = ((c1 & 0xffffffff) ^ 0x80000000) - 0x80000000;
2058 gcc_assert (c1 == d4);
2059 #else
2060 d1 = ((c1 & 0xffff) ^ 0x8000) - 0x8000;
2061 c1 -= d1;
2062 d2 = ((c1 & 0xffffffff) ^ 0x80000000) - 0x80000000;
2063 gcc_assert (c1 == d2);
2064 c2 += (d2 < 0);
2065 d3 = ((c2 & 0xffff) ^ 0x8000) - 0x8000;
2066 c2 -= d3;
2067 d4 = ((c2 & 0xffffffff) ^ 0x80000000) - 0x80000000;
2068 gcc_assert (c2 == d4);
2069 #endif
2071 /* Construct the high word */
2072 if (d4)
2074 emit_move_insn (target, GEN_INT (d4));
2075 if (d3)
2076 emit_move_insn (target, gen_rtx_PLUS (DImode, target, GEN_INT (d3)));
2078 else
2079 emit_move_insn (target, GEN_INT (d3));
2081 /* Shift it into place */
2082 emit_move_insn (target, gen_rtx_ASHIFT (DImode, target, GEN_INT (32)));
2084 /* Add in the low bits. */
2085 if (d2)
2086 emit_move_insn (target, gen_rtx_PLUS (DImode, target, GEN_INT (d2)));
2087 if (d1)
2088 emit_move_insn (target, gen_rtx_PLUS (DImode, target, GEN_INT (d1)));
2090 return target;
2093 /* Given an integral CONST_INT, CONST_DOUBLE, or CONST_VECTOR, return
2094 the low 64 bits. */
2096 static void
2097 alpha_extract_integer (rtx x, HOST_WIDE_INT *p0, HOST_WIDE_INT *p1)
2099 HOST_WIDE_INT i0, i1;
2101 if (GET_CODE (x) == CONST_VECTOR)
2102 x = simplify_subreg (DImode, x, GET_MODE (x), 0);
2105 if (CONST_INT_P (x))
2107 i0 = INTVAL (x);
2108 i1 = -(i0 < 0);
2110 else if (HOST_BITS_PER_WIDE_INT >= 64)
2112 i0 = CONST_DOUBLE_LOW (x);
2113 i1 = -(i0 < 0);
2115 else
2117 i0 = CONST_DOUBLE_LOW (x);
2118 i1 = CONST_DOUBLE_HIGH (x);
2121 *p0 = i0;
2122 *p1 = i1;
2125 /* Implement TARGET_LEGITIMATE_CONSTANT_P. This is all constants for which
2126 we are willing to load the value into a register via a move pattern.
2127 Normally this is all symbolic constants, integral constants that
2128 take three or fewer instructions, and floating-point zero. */
2130 bool
2131 alpha_legitimate_constant_p (enum machine_mode mode, rtx x)
2133 HOST_WIDE_INT i0, i1;
2135 switch (GET_CODE (x))
2137 case LABEL_REF:
2138 case HIGH:
2139 return true;
2141 case CONST:
2142 if (GET_CODE (XEXP (x, 0)) == PLUS
2143 && GET_CODE (XEXP (XEXP (x, 0), 1)) == CONST_INT)
2144 x = XEXP (XEXP (x, 0), 0);
2145 else
2146 return true;
2148 if (GET_CODE (x) != SYMBOL_REF)
2149 return true;
2151 /* FALLTHRU */
2153 case SYMBOL_REF:
2154 /* TLS symbols are never valid. */
2155 return SYMBOL_REF_TLS_MODEL (x) == 0;
2157 case CONST_DOUBLE:
2158 if (x == CONST0_RTX (mode))
2159 return true;
2160 if (FLOAT_MODE_P (mode))
2161 return false;
2162 goto do_integer;
2164 case CONST_VECTOR:
2165 if (x == CONST0_RTX (mode))
2166 return true;
2167 if (GET_MODE_CLASS (mode) != MODE_VECTOR_INT)
2168 return false;
2169 if (GET_MODE_SIZE (mode) != 8)
2170 return false;
2171 goto do_integer;
2173 case CONST_INT:
2174 do_integer:
2175 if (TARGET_BUILD_CONSTANTS)
2176 return true;
2177 alpha_extract_integer (x, &i0, &i1);
2178 if (HOST_BITS_PER_WIDE_INT >= 64 || i1 == (-i0 < 0))
2179 return alpha_emit_set_const_1 (x, mode, i0, 3, true) != NULL;
2180 return false;
2182 default:
2183 return false;
2187 /* Operand 1 is known to be a constant, and should require more than one
2188 instruction to load. Emit that multi-part load. */
2190 bool
2191 alpha_split_const_mov (enum machine_mode mode, rtx *operands)
2193 HOST_WIDE_INT i0, i1;
2194 rtx temp = NULL_RTX;
2196 alpha_extract_integer (operands[1], &i0, &i1);
2198 if (HOST_BITS_PER_WIDE_INT >= 64 || i1 == -(i0 < 0))
2199 temp = alpha_emit_set_const (operands[0], mode, i0, 3, false);
2201 if (!temp && TARGET_BUILD_CONSTANTS)
2202 temp = alpha_emit_set_long_const (operands[0], i0, i1);
2204 if (temp)
2206 if (!rtx_equal_p (operands[0], temp))
2207 emit_move_insn (operands[0], temp);
2208 return true;
2211 return false;
2214 /* Expand a move instruction; return true if all work is done.
2215 We don't handle non-bwx subword loads here. */
2217 bool
2218 alpha_expand_mov (enum machine_mode mode, rtx *operands)
2220 rtx tmp;
2222 /* If the output is not a register, the input must be. */
2223 if (MEM_P (operands[0])
2224 && ! reg_or_0_operand (operands[1], mode))
2225 operands[1] = force_reg (mode, operands[1]);
2227 /* Allow legitimize_address to perform some simplifications. */
2228 if (mode == Pmode && symbolic_operand (operands[1], mode))
2230 tmp = alpha_legitimize_address_1 (operands[1], operands[0], mode);
2231 if (tmp)
2233 if (tmp == operands[0])
2234 return true;
2235 operands[1] = tmp;
2236 return false;
2240 /* Early out for non-constants and valid constants. */
2241 if (! CONSTANT_P (operands[1]) || input_operand (operands[1], mode))
2242 return false;
2244 /* Split large integers. */
2245 if (CONST_INT_P (operands[1])
2246 || GET_CODE (operands[1]) == CONST_DOUBLE
2247 || GET_CODE (operands[1]) == CONST_VECTOR)
2249 if (alpha_split_const_mov (mode, operands))
2250 return true;
2253 /* Otherwise we've nothing left but to drop the thing to memory. */
2254 tmp = force_const_mem (mode, operands[1]);
2256 if (tmp == NULL_RTX)
2257 return false;
2259 if (reload_in_progress)
2261 emit_move_insn (operands[0], XEXP (tmp, 0));
2262 operands[1] = replace_equiv_address (tmp, operands[0]);
2264 else
2265 operands[1] = validize_mem (tmp);
2266 return false;
2269 /* Expand a non-bwx QImode or HImode move instruction;
2270 return true if all work is done. */
2272 bool
2273 alpha_expand_mov_nobwx (enum machine_mode mode, rtx *operands)
2275 rtx seq;
2277 /* If the output is not a register, the input must be. */
2278 if (MEM_P (operands[0]))
2279 operands[1] = force_reg (mode, operands[1]);
2281 /* Handle four memory cases, unaligned and aligned for either the input
2282 or the output. The only case where we can be called during reload is
2283 for aligned loads; all other cases require temporaries. */
2285 if (any_memory_operand (operands[1], mode))
2287 if (aligned_memory_operand (operands[1], mode))
2289 if (reload_in_progress)
2291 if (mode == QImode)
2292 seq = gen_reload_inqi_aligned (operands[0], operands[1]);
2293 else
2294 seq = gen_reload_inhi_aligned (operands[0], operands[1]);
2295 emit_insn (seq);
2297 else
2299 rtx aligned_mem, bitnum;
2300 rtx scratch = gen_reg_rtx (SImode);
2301 rtx subtarget;
2302 bool copyout;
2304 get_aligned_mem (operands[1], &aligned_mem, &bitnum);
2306 subtarget = operands[0];
2307 if (REG_P (subtarget))
2308 subtarget = gen_lowpart (DImode, subtarget), copyout = false;
2309 else
2310 subtarget = gen_reg_rtx (DImode), copyout = true;
2312 if (mode == QImode)
2313 seq = gen_aligned_loadqi (subtarget, aligned_mem,
2314 bitnum, scratch);
2315 else
2316 seq = gen_aligned_loadhi (subtarget, aligned_mem,
2317 bitnum, scratch);
2318 emit_insn (seq);
2320 if (copyout)
2321 emit_move_insn (operands[0], gen_lowpart (mode, subtarget));
2324 else
2326 /* Don't pass these as parameters since that makes the generated
2327 code depend on parameter evaluation order which will cause
2328 bootstrap failures. */
2330 rtx temp1, temp2, subtarget, ua;
2331 bool copyout;
2333 temp1 = gen_reg_rtx (DImode);
2334 temp2 = gen_reg_rtx (DImode);
2336 subtarget = operands[0];
2337 if (REG_P (subtarget))
2338 subtarget = gen_lowpart (DImode, subtarget), copyout = false;
2339 else
2340 subtarget = gen_reg_rtx (DImode), copyout = true;
2342 ua = get_unaligned_address (operands[1]);
2343 if (mode == QImode)
2344 seq = gen_unaligned_loadqi (subtarget, ua, temp1, temp2);
2345 else
2346 seq = gen_unaligned_loadhi (subtarget, ua, temp1, temp2);
2348 alpha_set_memflags (seq, operands[1]);
2349 emit_insn (seq);
2351 if (copyout)
2352 emit_move_insn (operands[0], gen_lowpart (mode, subtarget));
2354 return true;
2357 if (any_memory_operand (operands[0], mode))
2359 if (aligned_memory_operand (operands[0], mode))
2361 rtx aligned_mem, bitnum;
2362 rtx temp1 = gen_reg_rtx (SImode);
2363 rtx temp2 = gen_reg_rtx (SImode);
2365 get_aligned_mem (operands[0], &aligned_mem, &bitnum);
2367 emit_insn (gen_aligned_store (aligned_mem, operands[1], bitnum,
2368 temp1, temp2));
2370 else
2372 rtx temp1 = gen_reg_rtx (DImode);
2373 rtx temp2 = gen_reg_rtx (DImode);
2374 rtx temp3 = gen_reg_rtx (DImode);
2375 rtx ua = get_unaligned_address (operands[0]);
2377 if (mode == QImode)
2378 seq = gen_unaligned_storeqi (ua, operands[1], temp1, temp2, temp3);
2379 else
2380 seq = gen_unaligned_storehi (ua, operands[1], temp1, temp2, temp3);
2382 alpha_set_memflags (seq, operands[0]);
2383 emit_insn (seq);
2385 return true;
2388 return false;
2391 /* Implement the movmisalign patterns. One of the operands is a memory
2392 that is not naturally aligned. Emit instructions to load it. */
2394 void
2395 alpha_expand_movmisalign (enum machine_mode mode, rtx *operands)
2397 /* Honor misaligned loads, for those we promised to do so. */
2398 if (MEM_P (operands[1]))
2400 rtx tmp;
2402 if (register_operand (operands[0], mode))
2403 tmp = operands[0];
2404 else
2405 tmp = gen_reg_rtx (mode);
2407 alpha_expand_unaligned_load (tmp, operands[1], 8, 0, 0);
2408 if (tmp != operands[0])
2409 emit_move_insn (operands[0], tmp);
2411 else if (MEM_P (operands[0]))
2413 if (!reg_or_0_operand (operands[1], mode))
2414 operands[1] = force_reg (mode, operands[1]);
2415 alpha_expand_unaligned_store (operands[0], operands[1], 8, 0);
2417 else
2418 gcc_unreachable ();
2421 /* Generate an unsigned DImode to FP conversion. This is the same code
2422 optabs would emit if we didn't have TFmode patterns.
2424 For SFmode, this is the only construction I've found that can pass
2425 gcc.c-torture/execute/ieee/rbug.c. No scenario that uses DFmode
2426 intermediates will work, because you'll get intermediate rounding
2427 that ruins the end result. Some of this could be fixed by turning
2428 on round-to-positive-infinity, but that requires diddling the fpsr,
2429 which kills performance. I tried turning this around and converting
2430 to a negative number, so that I could turn on /m, but either I did
2431 it wrong or there's something else cause I wound up with the exact
2432 same single-bit error. There is a branch-less form of this same code:
2434 srl $16,1,$1
2435 and $16,1,$2
2436 cmplt $16,0,$3
2437 or $1,$2,$2
2438 cmovge $16,$16,$2
2439 itoft $3,$f10
2440 itoft $2,$f11
2441 cvtqs $f11,$f11
2442 adds $f11,$f11,$f0
2443 fcmoveq $f10,$f11,$f0
2445 I'm not using it because it's the same number of instructions as
2446 this branch-full form, and it has more serialized long latency
2447 instructions on the critical path.
2449 For DFmode, we can avoid rounding errors by breaking up the word
2450 into two pieces, converting them separately, and adding them back:
2452 LC0: .long 0,0x5f800000
2454 itoft $16,$f11
2455 lda $2,LC0
2456 cmplt $16,0,$1
2457 cpyse $f11,$f31,$f10
2458 cpyse $f31,$f11,$f11
2459 s4addq $1,$2,$1
2460 lds $f12,0($1)
2461 cvtqt $f10,$f10
2462 cvtqt $f11,$f11
2463 addt $f12,$f10,$f0
2464 addt $f0,$f11,$f0
2466 This doesn't seem to be a clear-cut win over the optabs form.
2467 It probably all depends on the distribution of numbers being
2468 converted -- in the optabs form, all but high-bit-set has a
2469 much lower minimum execution time. */
2471 void
2472 alpha_emit_floatuns (rtx operands[2])
2474 rtx neglab, donelab, i0, i1, f0, in, out;
2475 enum machine_mode mode;
2477 out = operands[0];
2478 in = force_reg (DImode, operands[1]);
2479 mode = GET_MODE (out);
2480 neglab = gen_label_rtx ();
2481 donelab = gen_label_rtx ();
2482 i0 = gen_reg_rtx (DImode);
2483 i1 = gen_reg_rtx (DImode);
2484 f0 = gen_reg_rtx (mode);
2486 emit_cmp_and_jump_insns (in, const0_rtx, LT, const0_rtx, DImode, 0, neglab);
2488 emit_insn (gen_rtx_SET (VOIDmode, out, gen_rtx_FLOAT (mode, in)));
2489 emit_jump_insn (gen_jump (donelab));
2490 emit_barrier ();
2492 emit_label (neglab);
2494 emit_insn (gen_lshrdi3 (i0, in, const1_rtx));
2495 emit_insn (gen_anddi3 (i1, in, const1_rtx));
2496 emit_insn (gen_iordi3 (i0, i0, i1));
2497 emit_insn (gen_rtx_SET (VOIDmode, f0, gen_rtx_FLOAT (mode, i0)));
2498 emit_insn (gen_rtx_SET (VOIDmode, out, gen_rtx_PLUS (mode, f0, f0)));
2500 emit_label (donelab);
2503 /* Generate the comparison for a conditional branch. */
2505 void
2506 alpha_emit_conditional_branch (rtx operands[], enum machine_mode cmp_mode)
2508 enum rtx_code cmp_code, branch_code;
2509 enum machine_mode branch_mode = VOIDmode;
2510 enum rtx_code code = GET_CODE (operands[0]);
2511 rtx op0 = operands[1], op1 = operands[2];
2512 rtx tem;
2514 if (cmp_mode == TFmode)
2516 op0 = alpha_emit_xfloating_compare (&code, op0, op1);
2517 op1 = const0_rtx;
2518 cmp_mode = DImode;
2521 /* The general case: fold the comparison code to the types of compares
2522 that we have, choosing the branch as necessary. */
2523 switch (code)
2525 case EQ: case LE: case LT: case LEU: case LTU:
2526 case UNORDERED:
2527 /* We have these compares. */
2528 cmp_code = code, branch_code = NE;
2529 break;
2531 case NE:
2532 case ORDERED:
2533 /* These must be reversed. */
2534 cmp_code = reverse_condition (code), branch_code = EQ;
2535 break;
2537 case GE: case GT: case GEU: case GTU:
2538 /* For FP, we swap them, for INT, we reverse them. */
2539 if (cmp_mode == DFmode)
2541 cmp_code = swap_condition (code);
2542 branch_code = NE;
2543 tem = op0, op0 = op1, op1 = tem;
2545 else
2547 cmp_code = reverse_condition (code);
2548 branch_code = EQ;
2550 break;
2552 default:
2553 gcc_unreachable ();
2556 if (cmp_mode == DFmode)
2558 if (flag_unsafe_math_optimizations && cmp_code != UNORDERED)
2560 /* When we are not as concerned about non-finite values, and we
2561 are comparing against zero, we can branch directly. */
2562 if (op1 == CONST0_RTX (DFmode))
2563 cmp_code = UNKNOWN, branch_code = code;
2564 else if (op0 == CONST0_RTX (DFmode))
2566 /* Undo the swap we probably did just above. */
2567 tem = op0, op0 = op1, op1 = tem;
2568 branch_code = swap_condition (cmp_code);
2569 cmp_code = UNKNOWN;
2572 else
2574 /* ??? We mark the branch mode to be CCmode to prevent the
2575 compare and branch from being combined, since the compare
2576 insn follows IEEE rules that the branch does not. */
2577 branch_mode = CCmode;
2580 else
2582 /* The following optimizations are only for signed compares. */
2583 if (code != LEU && code != LTU && code != GEU && code != GTU)
2585 /* Whee. Compare and branch against 0 directly. */
2586 if (op1 == const0_rtx)
2587 cmp_code = UNKNOWN, branch_code = code;
2589 /* If the constants doesn't fit into an immediate, but can
2590 be generated by lda/ldah, we adjust the argument and
2591 compare against zero, so we can use beq/bne directly. */
2592 /* ??? Don't do this when comparing against symbols, otherwise
2593 we'll reduce (&x == 0x1234) to (&x-0x1234 == 0), which will
2594 be declared false out of hand (at least for non-weak). */
2595 else if (CONST_INT_P (op1)
2596 && (code == EQ || code == NE)
2597 && !(symbolic_operand (op0, VOIDmode)
2598 || (REG_P (op0) && REG_POINTER (op0))))
2600 rtx n_op1 = GEN_INT (-INTVAL (op1));
2602 if (! satisfies_constraint_I (op1)
2603 && (satisfies_constraint_K (n_op1)
2604 || satisfies_constraint_L (n_op1)))
2605 cmp_code = PLUS, branch_code = code, op1 = n_op1;
2609 if (!reg_or_0_operand (op0, DImode))
2610 op0 = force_reg (DImode, op0);
2611 if (cmp_code != PLUS && !reg_or_8bit_operand (op1, DImode))
2612 op1 = force_reg (DImode, op1);
2615 /* Emit an initial compare instruction, if necessary. */
2616 tem = op0;
2617 if (cmp_code != UNKNOWN)
2619 tem = gen_reg_rtx (cmp_mode);
2620 emit_move_insn (tem, gen_rtx_fmt_ee (cmp_code, cmp_mode, op0, op1));
2623 /* Emit the branch instruction. */
2624 tem = gen_rtx_SET (VOIDmode, pc_rtx,
2625 gen_rtx_IF_THEN_ELSE (VOIDmode,
2626 gen_rtx_fmt_ee (branch_code,
2627 branch_mode, tem,
2628 CONST0_RTX (cmp_mode)),
2629 gen_rtx_LABEL_REF (VOIDmode,
2630 operands[3]),
2631 pc_rtx));
2632 emit_jump_insn (tem);
2635 /* Certain simplifications can be done to make invalid setcc operations
2636 valid. Return the final comparison, or NULL if we can't work. */
2638 bool
2639 alpha_emit_setcc (rtx operands[], enum machine_mode cmp_mode)
2641 enum rtx_code cmp_code;
2642 enum rtx_code code = GET_CODE (operands[1]);
2643 rtx op0 = operands[2], op1 = operands[3];
2644 rtx tmp;
2646 if (cmp_mode == TFmode)
2648 op0 = alpha_emit_xfloating_compare (&code, op0, op1);
2649 op1 = const0_rtx;
2650 cmp_mode = DImode;
2653 if (cmp_mode == DFmode && !TARGET_FIX)
2654 return 0;
2656 /* The general case: fold the comparison code to the types of compares
2657 that we have, choosing the branch as necessary. */
2659 cmp_code = UNKNOWN;
2660 switch (code)
2662 case EQ: case LE: case LT: case LEU: case LTU:
2663 case UNORDERED:
2664 /* We have these compares. */
2665 if (cmp_mode == DFmode)
2666 cmp_code = code, code = NE;
2667 break;
2669 case NE:
2670 if (cmp_mode == DImode && op1 == const0_rtx)
2671 break;
2672 /* FALLTHRU */
2674 case ORDERED:
2675 cmp_code = reverse_condition (code);
2676 code = EQ;
2677 break;
2679 case GE: case GT: case GEU: case GTU:
2680 /* These normally need swapping, but for integer zero we have
2681 special patterns that recognize swapped operands. */
2682 if (cmp_mode == DImode && op1 == const0_rtx)
2683 break;
2684 code = swap_condition (code);
2685 if (cmp_mode == DFmode)
2686 cmp_code = code, code = NE;
2687 tmp = op0, op0 = op1, op1 = tmp;
2688 break;
2690 default:
2691 gcc_unreachable ();
2694 if (cmp_mode == DImode)
2696 if (!register_operand (op0, DImode))
2697 op0 = force_reg (DImode, op0);
2698 if (!reg_or_8bit_operand (op1, DImode))
2699 op1 = force_reg (DImode, op1);
2702 /* Emit an initial compare instruction, if necessary. */
2703 if (cmp_code != UNKNOWN)
2705 tmp = gen_reg_rtx (cmp_mode);
2706 emit_insn (gen_rtx_SET (VOIDmode, tmp,
2707 gen_rtx_fmt_ee (cmp_code, cmp_mode, op0, op1)));
2709 op0 = cmp_mode != DImode ? gen_lowpart (DImode, tmp) : tmp;
2710 op1 = const0_rtx;
2713 /* Emit the setcc instruction. */
2714 emit_insn (gen_rtx_SET (VOIDmode, operands[0],
2715 gen_rtx_fmt_ee (code, DImode, op0, op1)));
2716 return true;
2720 /* Rewrite a comparison against zero CMP of the form
2721 (CODE (cc0) (const_int 0)) so it can be written validly in
2722 a conditional move (if_then_else CMP ...).
2723 If both of the operands that set cc0 are nonzero we must emit
2724 an insn to perform the compare (it can't be done within
2725 the conditional move). */
2728 alpha_emit_conditional_move (rtx cmp, enum machine_mode mode)
2730 enum rtx_code code = GET_CODE (cmp);
2731 enum rtx_code cmov_code = NE;
2732 rtx op0 = XEXP (cmp, 0);
2733 rtx op1 = XEXP (cmp, 1);
2734 enum machine_mode cmp_mode
2735 = (GET_MODE (op0) == VOIDmode ? DImode : GET_MODE (op0));
2736 enum machine_mode cmov_mode = VOIDmode;
2737 int local_fast_math = flag_unsafe_math_optimizations;
2738 rtx tem;
2740 if (cmp_mode == TFmode)
2742 op0 = alpha_emit_xfloating_compare (&code, op0, op1);
2743 op1 = const0_rtx;
2744 cmp_mode = DImode;
2747 gcc_assert (cmp_mode == DFmode || cmp_mode == DImode);
2749 if (FLOAT_MODE_P (cmp_mode) != FLOAT_MODE_P (mode))
2751 enum rtx_code cmp_code;
2753 if (! TARGET_FIX)
2754 return 0;
2756 /* If we have fp<->int register move instructions, do a cmov by
2757 performing the comparison in fp registers, and move the
2758 zero/nonzero value to integer registers, where we can then
2759 use a normal cmov, or vice-versa. */
2761 switch (code)
2763 case EQ: case LE: case LT: case LEU: case LTU:
2764 case UNORDERED:
2765 /* We have these compares. */
2766 cmp_code = code, code = NE;
2767 break;
2769 case NE:
2770 case ORDERED:
2771 /* These must be reversed. */
2772 cmp_code = reverse_condition (code), code = EQ;
2773 break;
2775 case GE: case GT: case GEU: case GTU:
2776 /* These normally need swapping, but for integer zero we have
2777 special patterns that recognize swapped operands. */
2778 if (cmp_mode == DImode && op1 == const0_rtx)
2779 cmp_code = code, code = NE;
2780 else
2782 cmp_code = swap_condition (code);
2783 code = NE;
2784 tem = op0, op0 = op1, op1 = tem;
2786 break;
2788 default:
2789 gcc_unreachable ();
2792 if (cmp_mode == DImode)
2794 if (!reg_or_0_operand (op0, DImode))
2795 op0 = force_reg (DImode, op0);
2796 if (!reg_or_8bit_operand (op1, DImode))
2797 op1 = force_reg (DImode, op1);
2800 tem = gen_reg_rtx (cmp_mode);
2801 emit_insn (gen_rtx_SET (VOIDmode, tem,
2802 gen_rtx_fmt_ee (cmp_code, cmp_mode,
2803 op0, op1)));
2805 cmp_mode = cmp_mode == DImode ? DFmode : DImode;
2806 op0 = gen_lowpart (cmp_mode, tem);
2807 op1 = CONST0_RTX (cmp_mode);
2808 cmp = gen_rtx_fmt_ee (code, VOIDmode, op0, op1);
2809 local_fast_math = 1;
2812 if (cmp_mode == DImode)
2814 if (!reg_or_0_operand (op0, DImode))
2815 op0 = force_reg (DImode, op0);
2816 if (!reg_or_8bit_operand (op1, DImode))
2817 op1 = force_reg (DImode, op1);
2820 /* We may be able to use a conditional move directly.
2821 This avoids emitting spurious compares. */
2822 if (signed_comparison_operator (cmp, VOIDmode)
2823 && (cmp_mode == DImode || local_fast_math)
2824 && (op0 == CONST0_RTX (cmp_mode) || op1 == CONST0_RTX (cmp_mode)))
2825 return gen_rtx_fmt_ee (code, VOIDmode, op0, op1);
2827 /* We can't put the comparison inside the conditional move;
2828 emit a compare instruction and put that inside the
2829 conditional move. Make sure we emit only comparisons we have;
2830 swap or reverse as necessary. */
2832 if (!can_create_pseudo_p ())
2833 return NULL_RTX;
2835 switch (code)
2837 case EQ: case LE: case LT: case LEU: case LTU:
2838 case UNORDERED:
2839 /* We have these compares: */
2840 break;
2842 case NE:
2843 case ORDERED:
2844 /* These must be reversed. */
2845 code = reverse_condition (code);
2846 cmov_code = EQ;
2847 break;
2849 case GE: case GT: case GEU: case GTU:
2850 /* These normally need swapping, but for integer zero we have
2851 special patterns that recognize swapped operands. */
2852 if (cmp_mode == DImode && op1 == const0_rtx)
2853 break;
2854 code = swap_condition (code);
2855 tem = op0, op0 = op1, op1 = tem;
2856 break;
2858 default:
2859 gcc_unreachable ();
2862 if (cmp_mode == DImode)
2864 if (!reg_or_0_operand (op0, DImode))
2865 op0 = force_reg (DImode, op0);
2866 if (!reg_or_8bit_operand (op1, DImode))
2867 op1 = force_reg (DImode, op1);
2870 /* ??? We mark the branch mode to be CCmode to prevent the compare
2871 and cmov from being combined, since the compare insn follows IEEE
2872 rules that the cmov does not. */
2873 if (cmp_mode == DFmode && !local_fast_math)
2874 cmov_mode = CCmode;
2876 tem = gen_reg_rtx (cmp_mode);
2877 emit_move_insn (tem, gen_rtx_fmt_ee (code, cmp_mode, op0, op1));
2878 return gen_rtx_fmt_ee (cmov_code, cmov_mode, tem, CONST0_RTX (cmp_mode));
2881 /* Simplify a conditional move of two constants into a setcc with
2882 arithmetic. This is done with a splitter since combine would
2883 just undo the work if done during code generation. It also catches
2884 cases we wouldn't have before cse. */
2887 alpha_split_conditional_move (enum rtx_code code, rtx dest, rtx cond,
2888 rtx t_rtx, rtx f_rtx)
2890 HOST_WIDE_INT t, f, diff;
2891 enum machine_mode mode;
2892 rtx target, subtarget, tmp;
2894 mode = GET_MODE (dest);
2895 t = INTVAL (t_rtx);
2896 f = INTVAL (f_rtx);
2897 diff = t - f;
2899 if (((code == NE || code == EQ) && diff < 0)
2900 || (code == GE || code == GT))
2902 code = reverse_condition (code);
2903 diff = t, t = f, f = diff;
2904 diff = t - f;
2907 subtarget = target = dest;
2908 if (mode != DImode)
2910 target = gen_lowpart (DImode, dest);
2911 if (can_create_pseudo_p ())
2912 subtarget = gen_reg_rtx (DImode);
2913 else
2914 subtarget = target;
2916 /* Below, we must be careful to use copy_rtx on target and subtarget
2917 in intermediate insns, as they may be a subreg rtx, which may not
2918 be shared. */
2920 if (f == 0 && exact_log2 (diff) > 0
2921 /* On EV6, we've got enough shifters to make non-arithmetic shifts
2922 viable over a longer latency cmove. On EV5, the E0 slot is a
2923 scarce resource, and on EV4 shift has the same latency as a cmove. */
2924 && (diff <= 8 || alpha_tune == PROCESSOR_EV6))
2926 tmp = gen_rtx_fmt_ee (code, DImode, cond, const0_rtx);
2927 emit_insn (gen_rtx_SET (VOIDmode, copy_rtx (subtarget), tmp));
2929 tmp = gen_rtx_ASHIFT (DImode, copy_rtx (subtarget),
2930 GEN_INT (exact_log2 (t)));
2931 emit_insn (gen_rtx_SET (VOIDmode, target, tmp));
2933 else if (f == 0 && t == -1)
2935 tmp = gen_rtx_fmt_ee (code, DImode, cond, const0_rtx);
2936 emit_insn (gen_rtx_SET (VOIDmode, copy_rtx (subtarget), tmp));
2938 emit_insn (gen_negdi2 (target, copy_rtx (subtarget)));
2940 else if (diff == 1 || diff == 4 || diff == 8)
2942 rtx add_op;
2944 tmp = gen_rtx_fmt_ee (code, DImode, cond, const0_rtx);
2945 emit_insn (gen_rtx_SET (VOIDmode, copy_rtx (subtarget), tmp));
2947 if (diff == 1)
2948 emit_insn (gen_adddi3 (target, copy_rtx (subtarget), GEN_INT (f)));
2949 else
2951 add_op = GEN_INT (f);
2952 if (sext_add_operand (add_op, mode))
2954 tmp = gen_rtx_MULT (DImode, copy_rtx (subtarget),
2955 GEN_INT (diff));
2956 tmp = gen_rtx_PLUS (DImode, tmp, add_op);
2957 emit_insn (gen_rtx_SET (VOIDmode, target, tmp));
2959 else
2960 return 0;
2963 else
2964 return 0;
2966 return 1;
2969 /* Look up the function X_floating library function name for the
2970 given operation. */
2972 struct GTY(()) xfloating_op
2974 const enum rtx_code code;
2975 const char *const GTY((skip)) osf_func;
2976 const char *const GTY((skip)) vms_func;
2977 rtx libcall;
2980 static GTY(()) struct xfloating_op xfloating_ops[] =
2982 { PLUS, "_OtsAddX", "OTS$ADD_X", 0 },
2983 { MINUS, "_OtsSubX", "OTS$SUB_X", 0 },
2984 { MULT, "_OtsMulX", "OTS$MUL_X", 0 },
2985 { DIV, "_OtsDivX", "OTS$DIV_X", 0 },
2986 { EQ, "_OtsEqlX", "OTS$EQL_X", 0 },
2987 { NE, "_OtsNeqX", "OTS$NEQ_X", 0 },
2988 { LT, "_OtsLssX", "OTS$LSS_X", 0 },
2989 { LE, "_OtsLeqX", "OTS$LEQ_X", 0 },
2990 { GT, "_OtsGtrX", "OTS$GTR_X", 0 },
2991 { GE, "_OtsGeqX", "OTS$GEQ_X", 0 },
2992 { FIX, "_OtsCvtXQ", "OTS$CVTXQ", 0 },
2993 { FLOAT, "_OtsCvtQX", "OTS$CVTQX", 0 },
2994 { UNSIGNED_FLOAT, "_OtsCvtQUX", "OTS$CVTQUX", 0 },
2995 { FLOAT_EXTEND, "_OtsConvertFloatTX", "OTS$CVT_FLOAT_T_X", 0 },
2996 { FLOAT_TRUNCATE, "_OtsConvertFloatXT", "OTS$CVT_FLOAT_X_T", 0 }
2999 static GTY(()) struct xfloating_op vax_cvt_ops[] =
3001 { FLOAT_EXTEND, "_OtsConvertFloatGX", "OTS$CVT_FLOAT_G_X", 0 },
3002 { FLOAT_TRUNCATE, "_OtsConvertFloatXG", "OTS$CVT_FLOAT_X_G", 0 }
3005 static rtx
3006 alpha_lookup_xfloating_lib_func (enum rtx_code code)
3008 struct xfloating_op *ops = xfloating_ops;
3009 long n = ARRAY_SIZE (xfloating_ops);
3010 long i;
3012 gcc_assert (TARGET_HAS_XFLOATING_LIBS);
3014 /* How irritating. Nothing to key off for the main table. */
3015 if (TARGET_FLOAT_VAX && (code == FLOAT_EXTEND || code == FLOAT_TRUNCATE))
3017 ops = vax_cvt_ops;
3018 n = ARRAY_SIZE (vax_cvt_ops);
3021 for (i = 0; i < n; ++i, ++ops)
3022 if (ops->code == code)
3024 rtx func = ops->libcall;
3025 if (!func)
3027 func = init_one_libfunc (TARGET_ABI_OPEN_VMS
3028 ? ops->vms_func : ops->osf_func);
3029 ops->libcall = func;
3031 return func;
3034 gcc_unreachable ();
3037 /* Most X_floating operations take the rounding mode as an argument.
3038 Compute that here. */
3040 static int
3041 alpha_compute_xfloating_mode_arg (enum rtx_code code,
3042 enum alpha_fp_rounding_mode round)
3044 int mode;
3046 switch (round)
3048 case ALPHA_FPRM_NORM:
3049 mode = 2;
3050 break;
3051 case ALPHA_FPRM_MINF:
3052 mode = 1;
3053 break;
3054 case ALPHA_FPRM_CHOP:
3055 mode = 0;
3056 break;
3057 case ALPHA_FPRM_DYN:
3058 mode = 4;
3059 break;
3060 default:
3061 gcc_unreachable ();
3063 /* XXX For reference, round to +inf is mode = 3. */
3066 if (code == FLOAT_TRUNCATE && alpha_fptm == ALPHA_FPTM_N)
3067 mode |= 0x10000;
3069 return mode;
3072 /* Emit an X_floating library function call.
3074 Note that these functions do not follow normal calling conventions:
3075 TFmode arguments are passed in two integer registers (as opposed to
3076 indirect); TFmode return values appear in R16+R17.
3078 FUNC is the function to call.
3079 TARGET is where the output belongs.
3080 OPERANDS are the inputs.
3081 NOPERANDS is the count of inputs.
3082 EQUIV is the expression equivalent for the function.
3085 static void
3086 alpha_emit_xfloating_libcall (rtx func, rtx target, rtx operands[],
3087 int noperands, rtx equiv)
3089 rtx usage = NULL_RTX, tmp, reg;
3090 int regno = 16, i;
3092 start_sequence ();
3094 for (i = 0; i < noperands; ++i)
3096 switch (GET_MODE (operands[i]))
3098 case TFmode:
3099 reg = gen_rtx_REG (TFmode, regno);
3100 regno += 2;
3101 break;
3103 case DFmode:
3104 reg = gen_rtx_REG (DFmode, regno + 32);
3105 regno += 1;
3106 break;
3108 case VOIDmode:
3109 gcc_assert (CONST_INT_P (operands[i]));
3110 /* FALLTHRU */
3111 case DImode:
3112 reg = gen_rtx_REG (DImode, regno);
3113 regno += 1;
3114 break;
3116 default:
3117 gcc_unreachable ();
3120 emit_move_insn (reg, operands[i]);
3121 use_reg (&usage, reg);
3124 switch (GET_MODE (target))
3126 case TFmode:
3127 reg = gen_rtx_REG (TFmode, 16);
3128 break;
3129 case DFmode:
3130 reg = gen_rtx_REG (DFmode, 32);
3131 break;
3132 case DImode:
3133 reg = gen_rtx_REG (DImode, 0);
3134 break;
3135 default:
3136 gcc_unreachable ();
3139 tmp = gen_rtx_MEM (QImode, func);
3140 tmp = emit_call_insn (GEN_CALL_VALUE (reg, tmp, const0_rtx,
3141 const0_rtx, const0_rtx));
3142 CALL_INSN_FUNCTION_USAGE (tmp) = usage;
3143 RTL_CONST_CALL_P (tmp) = 1;
3145 tmp = get_insns ();
3146 end_sequence ();
3148 emit_libcall_block (tmp, target, reg, equiv);
3151 /* Emit an X_floating library function call for arithmetic (+,-,*,/). */
3153 void
3154 alpha_emit_xfloating_arith (enum rtx_code code, rtx operands[])
3156 rtx func;
3157 int mode;
3158 rtx out_operands[3];
3160 func = alpha_lookup_xfloating_lib_func (code);
3161 mode = alpha_compute_xfloating_mode_arg (code, alpha_fprm);
3163 out_operands[0] = operands[1];
3164 out_operands[1] = operands[2];
3165 out_operands[2] = GEN_INT (mode);
3166 alpha_emit_xfloating_libcall (func, operands[0], out_operands, 3,
3167 gen_rtx_fmt_ee (code, TFmode, operands[1],
3168 operands[2]));
3171 /* Emit an X_floating library function call for a comparison. */
3173 static rtx
3174 alpha_emit_xfloating_compare (enum rtx_code *pcode, rtx op0, rtx op1)
3176 enum rtx_code cmp_code, res_code;
3177 rtx func, out, operands[2], note;
3179 /* X_floating library comparison functions return
3180 -1 unordered
3181 0 false
3182 1 true
3183 Convert the compare against the raw return value. */
3185 cmp_code = *pcode;
3186 switch (cmp_code)
3188 case UNORDERED:
3189 cmp_code = EQ;
3190 res_code = LT;
3191 break;
3192 case ORDERED:
3193 cmp_code = EQ;
3194 res_code = GE;
3195 break;
3196 case NE:
3197 res_code = NE;
3198 break;
3199 case EQ:
3200 case LT:
3201 case GT:
3202 case LE:
3203 case GE:
3204 res_code = GT;
3205 break;
3206 default:
3207 gcc_unreachable ();
3209 *pcode = res_code;
3211 func = alpha_lookup_xfloating_lib_func (cmp_code);
3213 operands[0] = op0;
3214 operands[1] = op1;
3215 out = gen_reg_rtx (DImode);
3217 /* What's actually returned is -1,0,1, not a proper boolean value. */
3218 note = gen_rtx_fmt_ee (cmp_code, VOIDmode, op0, op1);
3219 note = gen_rtx_UNSPEC (DImode, gen_rtvec (1, note), UNSPEC_XFLT_COMPARE);
3220 alpha_emit_xfloating_libcall (func, out, operands, 2, note);
3222 return out;
3225 /* Emit an X_floating library function call for a conversion. */
3227 void
3228 alpha_emit_xfloating_cvt (enum rtx_code orig_code, rtx operands[])
3230 int noperands = 1, mode;
3231 rtx out_operands[2];
3232 rtx func;
3233 enum rtx_code code = orig_code;
3235 if (code == UNSIGNED_FIX)
3236 code = FIX;
3238 func = alpha_lookup_xfloating_lib_func (code);
3240 out_operands[0] = operands[1];
3242 switch (code)
3244 case FIX:
3245 mode = alpha_compute_xfloating_mode_arg (code, ALPHA_FPRM_CHOP);
3246 out_operands[1] = GEN_INT (mode);
3247 noperands = 2;
3248 break;
3249 case FLOAT_TRUNCATE:
3250 mode = alpha_compute_xfloating_mode_arg (code, alpha_fprm);
3251 out_operands[1] = GEN_INT (mode);
3252 noperands = 2;
3253 break;
3254 default:
3255 break;
3258 alpha_emit_xfloating_libcall (func, operands[0], out_operands, noperands,
3259 gen_rtx_fmt_e (orig_code,
3260 GET_MODE (operands[0]),
3261 operands[1]));
3264 /* Split a TImode or TFmode move from OP[1] to OP[0] into a pair of
3265 DImode moves from OP[2,3] to OP[0,1]. If FIXUP_OVERLAP is true,
3266 guarantee that the sequence
3267 set (OP[0] OP[2])
3268 set (OP[1] OP[3])
3269 is valid. Naturally, output operand ordering is little-endian.
3270 This is used by *movtf_internal and *movti_internal. */
3272 void
3273 alpha_split_tmode_pair (rtx operands[4], enum machine_mode mode,
3274 bool fixup_overlap)
3276 switch (GET_CODE (operands[1]))
3278 case REG:
3279 operands[3] = gen_rtx_REG (DImode, REGNO (operands[1]) + 1);
3280 operands[2] = gen_rtx_REG (DImode, REGNO (operands[1]));
3281 break;
3283 case MEM:
3284 operands[3] = adjust_address (operands[1], DImode, 8);
3285 operands[2] = adjust_address (operands[1], DImode, 0);
3286 break;
3288 case CONST_INT:
3289 case CONST_DOUBLE:
3290 gcc_assert (operands[1] == CONST0_RTX (mode));
3291 operands[2] = operands[3] = const0_rtx;
3292 break;
3294 default:
3295 gcc_unreachable ();
3298 switch (GET_CODE (operands[0]))
3300 case REG:
3301 operands[1] = gen_rtx_REG (DImode, REGNO (operands[0]) + 1);
3302 operands[0] = gen_rtx_REG (DImode, REGNO (operands[0]));
3303 break;
3305 case MEM:
3306 operands[1] = adjust_address (operands[0], DImode, 8);
3307 operands[0] = adjust_address (operands[0], DImode, 0);
3308 break;
3310 default:
3311 gcc_unreachable ();
3314 if (fixup_overlap && reg_overlap_mentioned_p (operands[0], operands[3]))
3316 rtx tmp;
3317 tmp = operands[0], operands[0] = operands[1], operands[1] = tmp;
3318 tmp = operands[2], operands[2] = operands[3], operands[3] = tmp;
3322 /* Implement negtf2 or abstf2. Op0 is destination, op1 is source,
3323 op2 is a register containing the sign bit, operation is the
3324 logical operation to be performed. */
3326 void
3327 alpha_split_tfmode_frobsign (rtx operands[3], rtx (*operation) (rtx, rtx, rtx))
3329 rtx high_bit = operands[2];
3330 rtx scratch;
3331 int move;
3333 alpha_split_tmode_pair (operands, TFmode, false);
3335 /* Detect three flavors of operand overlap. */
3336 move = 1;
3337 if (rtx_equal_p (operands[0], operands[2]))
3338 move = 0;
3339 else if (rtx_equal_p (operands[1], operands[2]))
3341 if (rtx_equal_p (operands[0], high_bit))
3342 move = 2;
3343 else
3344 move = -1;
3347 if (move < 0)
3348 emit_move_insn (operands[0], operands[2]);
3350 /* ??? If the destination overlaps both source tf and high_bit, then
3351 assume source tf is dead in its entirety and use the other half
3352 for a scratch register. Otherwise "scratch" is just the proper
3353 destination register. */
3354 scratch = operands[move < 2 ? 1 : 3];
3356 emit_insn ((*operation) (scratch, high_bit, operands[3]));
3358 if (move > 0)
3360 emit_move_insn (operands[0], operands[2]);
3361 if (move > 1)
3362 emit_move_insn (operands[1], scratch);
3366 /* Use ext[wlq][lh] as the Architecture Handbook describes for extracting
3367 unaligned data:
3369 unsigned: signed:
3370 word: ldq_u r1,X(r11) ldq_u r1,X(r11)
3371 ldq_u r2,X+1(r11) ldq_u r2,X+1(r11)
3372 lda r3,X(r11) lda r3,X+2(r11)
3373 extwl r1,r3,r1 extql r1,r3,r1
3374 extwh r2,r3,r2 extqh r2,r3,r2
3375 or r1.r2.r1 or r1,r2,r1
3376 sra r1,48,r1
3378 long: ldq_u r1,X(r11) ldq_u r1,X(r11)
3379 ldq_u r2,X+3(r11) ldq_u r2,X+3(r11)
3380 lda r3,X(r11) lda r3,X(r11)
3381 extll r1,r3,r1 extll r1,r3,r1
3382 extlh r2,r3,r2 extlh r2,r3,r2
3383 or r1.r2.r1 addl r1,r2,r1
3385 quad: ldq_u r1,X(r11)
3386 ldq_u r2,X+7(r11)
3387 lda r3,X(r11)
3388 extql r1,r3,r1
3389 extqh r2,r3,r2
3390 or r1.r2.r1
3393 void
3394 alpha_expand_unaligned_load (rtx tgt, rtx mem, HOST_WIDE_INT size,
3395 HOST_WIDE_INT ofs, int sign)
3397 rtx meml, memh, addr, extl, exth, tmp, mema;
3398 enum machine_mode mode;
3400 if (TARGET_BWX && size == 2)
3402 meml = adjust_address (mem, QImode, ofs);
3403 memh = adjust_address (mem, QImode, ofs+1);
3404 extl = gen_reg_rtx (DImode);
3405 exth = gen_reg_rtx (DImode);
3406 emit_insn (gen_zero_extendqidi2 (extl, meml));
3407 emit_insn (gen_zero_extendqidi2 (exth, memh));
3408 exth = expand_simple_binop (DImode, ASHIFT, exth, GEN_INT (8),
3409 NULL, 1, OPTAB_LIB_WIDEN);
3410 addr = expand_simple_binop (DImode, IOR, extl, exth,
3411 NULL, 1, OPTAB_LIB_WIDEN);
3413 if (sign && GET_MODE (tgt) != HImode)
3415 addr = gen_lowpart (HImode, addr);
3416 emit_insn (gen_extend_insn (tgt, addr, GET_MODE (tgt), HImode, 0));
3418 else
3420 if (GET_MODE (tgt) != DImode)
3421 addr = gen_lowpart (GET_MODE (tgt), addr);
3422 emit_move_insn (tgt, addr);
3424 return;
3427 meml = gen_reg_rtx (DImode);
3428 memh = gen_reg_rtx (DImode);
3429 addr = gen_reg_rtx (DImode);
3430 extl = gen_reg_rtx (DImode);
3431 exth = gen_reg_rtx (DImode);
3433 mema = XEXP (mem, 0);
3434 if (GET_CODE (mema) == LO_SUM)
3435 mema = force_reg (Pmode, mema);
3437 /* AND addresses cannot be in any alias set, since they may implicitly
3438 alias surrounding code. Ideally we'd have some alias set that
3439 covered all types except those with alignment 8 or higher. */
3441 tmp = change_address (mem, DImode,
3442 gen_rtx_AND (DImode,
3443 plus_constant (DImode, mema, ofs),
3444 GEN_INT (-8)));
3445 set_mem_alias_set (tmp, 0);
3446 emit_move_insn (meml, tmp);
3448 tmp = change_address (mem, DImode,
3449 gen_rtx_AND (DImode,
3450 plus_constant (DImode, mema,
3451 ofs + size - 1),
3452 GEN_INT (-8)));
3453 set_mem_alias_set (tmp, 0);
3454 emit_move_insn (memh, tmp);
3456 if (sign && size == 2)
3458 emit_move_insn (addr, plus_constant (Pmode, mema, ofs+2));
3460 emit_insn (gen_extql (extl, meml, addr));
3461 emit_insn (gen_extqh (exth, memh, addr));
3463 /* We must use tgt here for the target. Alpha-vms port fails if we use
3464 addr for the target, because addr is marked as a pointer and combine
3465 knows that pointers are always sign-extended 32-bit values. */
3466 addr = expand_binop (DImode, ior_optab, extl, exth, tgt, 1, OPTAB_WIDEN);
3467 addr = expand_binop (DImode, ashr_optab, addr, GEN_INT (48),
3468 addr, 1, OPTAB_WIDEN);
3470 else
3472 emit_move_insn (addr, plus_constant (Pmode, mema, ofs));
3473 emit_insn (gen_extxl (extl, meml, GEN_INT (size*8), addr));
3474 switch ((int) size)
3476 case 2:
3477 emit_insn (gen_extwh (exth, memh, addr));
3478 mode = HImode;
3479 break;
3480 case 4:
3481 emit_insn (gen_extlh (exth, memh, addr));
3482 mode = SImode;
3483 break;
3484 case 8:
3485 emit_insn (gen_extqh (exth, memh, addr));
3486 mode = DImode;
3487 break;
3488 default:
3489 gcc_unreachable ();
3492 addr = expand_binop (mode, ior_optab, gen_lowpart (mode, extl),
3493 gen_lowpart (mode, exth), gen_lowpart (mode, tgt),
3494 sign, OPTAB_WIDEN);
3497 if (addr != tgt)
3498 emit_move_insn (tgt, gen_lowpart (GET_MODE (tgt), addr));
3501 /* Similarly, use ins and msk instructions to perform unaligned stores. */
3503 void
3504 alpha_expand_unaligned_store (rtx dst, rtx src,
3505 HOST_WIDE_INT size, HOST_WIDE_INT ofs)
3507 rtx dstl, dsth, addr, insl, insh, meml, memh, dsta;
3509 if (TARGET_BWX && size == 2)
3511 if (src != const0_rtx)
3513 dstl = gen_lowpart (QImode, src);
3514 dsth = expand_simple_binop (DImode, LSHIFTRT, src, GEN_INT (8),
3515 NULL, 1, OPTAB_LIB_WIDEN);
3516 dsth = gen_lowpart (QImode, dsth);
3518 else
3519 dstl = dsth = const0_rtx;
3521 meml = adjust_address (dst, QImode, ofs);
3522 memh = adjust_address (dst, QImode, ofs+1);
3524 emit_move_insn (meml, dstl);
3525 emit_move_insn (memh, dsth);
3526 return;
3529 dstl = gen_reg_rtx (DImode);
3530 dsth = gen_reg_rtx (DImode);
3531 insl = gen_reg_rtx (DImode);
3532 insh = gen_reg_rtx (DImode);
3534 dsta = XEXP (dst, 0);
3535 if (GET_CODE (dsta) == LO_SUM)
3536 dsta = force_reg (Pmode, dsta);
3538 /* AND addresses cannot be in any alias set, since they may implicitly
3539 alias surrounding code. Ideally we'd have some alias set that
3540 covered all types except those with alignment 8 or higher. */
3542 meml = change_address (dst, DImode,
3543 gen_rtx_AND (DImode,
3544 plus_constant (DImode, dsta, ofs),
3545 GEN_INT (-8)));
3546 set_mem_alias_set (meml, 0);
3548 memh = change_address (dst, DImode,
3549 gen_rtx_AND (DImode,
3550 plus_constant (DImode, dsta,
3551 ofs + size - 1),
3552 GEN_INT (-8)));
3553 set_mem_alias_set (memh, 0);
3555 emit_move_insn (dsth, memh);
3556 emit_move_insn (dstl, meml);
3558 addr = copy_addr_to_reg (plus_constant (Pmode, dsta, ofs));
3560 if (src != CONST0_RTX (GET_MODE (src)))
3562 emit_insn (gen_insxh (insh, gen_lowpart (DImode, src),
3563 GEN_INT (size*8), addr));
3565 switch ((int) size)
3567 case 2:
3568 emit_insn (gen_inswl (insl, gen_lowpart (HImode, src), addr));
3569 break;
3570 case 4:
3571 emit_insn (gen_insll (insl, gen_lowpart (SImode, src), addr));
3572 break;
3573 case 8:
3574 emit_insn (gen_insql (insl, gen_lowpart (DImode, src), addr));
3575 break;
3576 default:
3577 gcc_unreachable ();
3581 emit_insn (gen_mskxh (dsth, dsth, GEN_INT (size*8), addr));
3583 switch ((int) size)
3585 case 2:
3586 emit_insn (gen_mskwl (dstl, dstl, addr));
3587 break;
3588 case 4:
3589 emit_insn (gen_mskll (dstl, dstl, addr));
3590 break;
3591 case 8:
3592 emit_insn (gen_mskql (dstl, dstl, addr));
3593 break;
3594 default:
3595 gcc_unreachable ();
3598 if (src != CONST0_RTX (GET_MODE (src)))
3600 dsth = expand_binop (DImode, ior_optab, insh, dsth, dsth, 0, OPTAB_WIDEN);
3601 dstl = expand_binop (DImode, ior_optab, insl, dstl, dstl, 0, OPTAB_WIDEN);
3604 /* Must store high before low for degenerate case of aligned. */
3605 emit_move_insn (memh, dsth);
3606 emit_move_insn (meml, dstl);
3609 /* The block move code tries to maximize speed by separating loads and
3610 stores at the expense of register pressure: we load all of the data
3611 before we store it back out. There are two secondary effects worth
3612 mentioning, that this speeds copying to/from aligned and unaligned
3613 buffers, and that it makes the code significantly easier to write. */
3615 #define MAX_MOVE_WORDS 8
3617 /* Load an integral number of consecutive unaligned quadwords. */
3619 static void
3620 alpha_expand_unaligned_load_words (rtx *out_regs, rtx smem,
3621 HOST_WIDE_INT words, HOST_WIDE_INT ofs)
3623 rtx const im8 = GEN_INT (-8);
3624 rtx ext_tmps[MAX_MOVE_WORDS], data_regs[MAX_MOVE_WORDS+1];
3625 rtx sreg, areg, tmp, smema;
3626 HOST_WIDE_INT i;
3628 smema = XEXP (smem, 0);
3629 if (GET_CODE (smema) == LO_SUM)
3630 smema = force_reg (Pmode, smema);
3632 /* Generate all the tmp registers we need. */
3633 for (i = 0; i < words; ++i)
3635 data_regs[i] = out_regs[i];
3636 ext_tmps[i] = gen_reg_rtx (DImode);
3638 data_regs[words] = gen_reg_rtx (DImode);
3640 if (ofs != 0)
3641 smem = adjust_address (smem, GET_MODE (smem), ofs);
3643 /* Load up all of the source data. */
3644 for (i = 0; i < words; ++i)
3646 tmp = change_address (smem, DImode,
3647 gen_rtx_AND (DImode,
3648 plus_constant (DImode, smema, 8*i),
3649 im8));
3650 set_mem_alias_set (tmp, 0);
3651 emit_move_insn (data_regs[i], tmp);
3654 tmp = change_address (smem, DImode,
3655 gen_rtx_AND (DImode,
3656 plus_constant (DImode, smema,
3657 8*words - 1),
3658 im8));
3659 set_mem_alias_set (tmp, 0);
3660 emit_move_insn (data_regs[words], tmp);
3662 /* Extract the half-word fragments. Unfortunately DEC decided to make
3663 extxh with offset zero a noop instead of zeroing the register, so
3664 we must take care of that edge condition ourselves with cmov. */
3666 sreg = copy_addr_to_reg (smema);
3667 areg = expand_binop (DImode, and_optab, sreg, GEN_INT (7), NULL,
3668 1, OPTAB_WIDEN);
3669 for (i = 0; i < words; ++i)
3671 emit_insn (gen_extql (data_regs[i], data_regs[i], sreg));
3672 emit_insn (gen_extqh (ext_tmps[i], data_regs[i+1], sreg));
3673 emit_insn (gen_rtx_SET (VOIDmode, ext_tmps[i],
3674 gen_rtx_IF_THEN_ELSE (DImode,
3675 gen_rtx_EQ (DImode, areg,
3676 const0_rtx),
3677 const0_rtx, ext_tmps[i])));
3680 /* Merge the half-words into whole words. */
3681 for (i = 0; i < words; ++i)
3683 out_regs[i] = expand_binop (DImode, ior_optab, data_regs[i],
3684 ext_tmps[i], data_regs[i], 1, OPTAB_WIDEN);
3688 /* Store an integral number of consecutive unaligned quadwords. DATA_REGS
3689 may be NULL to store zeros. */
3691 static void
3692 alpha_expand_unaligned_store_words (rtx *data_regs, rtx dmem,
3693 HOST_WIDE_INT words, HOST_WIDE_INT ofs)
3695 rtx const im8 = GEN_INT (-8);
3696 rtx ins_tmps[MAX_MOVE_WORDS];
3697 rtx st_tmp_1, st_tmp_2, dreg;
3698 rtx st_addr_1, st_addr_2, dmema;
3699 HOST_WIDE_INT i;
3701 dmema = XEXP (dmem, 0);
3702 if (GET_CODE (dmema) == LO_SUM)
3703 dmema = force_reg (Pmode, dmema);
3705 /* Generate all the tmp registers we need. */
3706 if (data_regs != NULL)
3707 for (i = 0; i < words; ++i)
3708 ins_tmps[i] = gen_reg_rtx(DImode);
3709 st_tmp_1 = gen_reg_rtx(DImode);
3710 st_tmp_2 = gen_reg_rtx(DImode);
3712 if (ofs != 0)
3713 dmem = adjust_address (dmem, GET_MODE (dmem), ofs);
3715 st_addr_2 = change_address (dmem, DImode,
3716 gen_rtx_AND (DImode,
3717 plus_constant (DImode, dmema,
3718 words*8 - 1),
3719 im8));
3720 set_mem_alias_set (st_addr_2, 0);
3722 st_addr_1 = change_address (dmem, DImode,
3723 gen_rtx_AND (DImode, dmema, im8));
3724 set_mem_alias_set (st_addr_1, 0);
3726 /* Load up the destination end bits. */
3727 emit_move_insn (st_tmp_2, st_addr_2);
3728 emit_move_insn (st_tmp_1, st_addr_1);
3730 /* Shift the input data into place. */
3731 dreg = copy_addr_to_reg (dmema);
3732 if (data_regs != NULL)
3734 for (i = words-1; i >= 0; --i)
3736 emit_insn (gen_insqh (ins_tmps[i], data_regs[i], dreg));
3737 emit_insn (gen_insql (data_regs[i], data_regs[i], dreg));
3739 for (i = words-1; i > 0; --i)
3741 ins_tmps[i-1] = expand_binop (DImode, ior_optab, data_regs[i],
3742 ins_tmps[i-1], ins_tmps[i-1], 1,
3743 OPTAB_WIDEN);
3747 /* Split and merge the ends with the destination data. */
3748 emit_insn (gen_mskqh (st_tmp_2, st_tmp_2, dreg));
3749 emit_insn (gen_mskql (st_tmp_1, st_tmp_1, dreg));
3751 if (data_regs != NULL)
3753 st_tmp_2 = expand_binop (DImode, ior_optab, st_tmp_2, ins_tmps[words-1],
3754 st_tmp_2, 1, OPTAB_WIDEN);
3755 st_tmp_1 = expand_binop (DImode, ior_optab, st_tmp_1, data_regs[0],
3756 st_tmp_1, 1, OPTAB_WIDEN);
3759 /* Store it all. */
3760 emit_move_insn (st_addr_2, st_tmp_2);
3761 for (i = words-1; i > 0; --i)
3763 rtx tmp = change_address (dmem, DImode,
3764 gen_rtx_AND (DImode,
3765 plus_constant (DImode,
3766 dmema, i*8),
3767 im8));
3768 set_mem_alias_set (tmp, 0);
3769 emit_move_insn (tmp, data_regs ? ins_tmps[i-1] : const0_rtx);
3771 emit_move_insn (st_addr_1, st_tmp_1);
3775 /* Expand string/block move operations.
3777 operands[0] is the pointer to the destination.
3778 operands[1] is the pointer to the source.
3779 operands[2] is the number of bytes to move.
3780 operands[3] is the alignment. */
3783 alpha_expand_block_move (rtx operands[])
3785 rtx bytes_rtx = operands[2];
3786 rtx align_rtx = operands[3];
3787 HOST_WIDE_INT orig_bytes = INTVAL (bytes_rtx);
3788 HOST_WIDE_INT bytes = orig_bytes;
3789 HOST_WIDE_INT src_align = INTVAL (align_rtx) * BITS_PER_UNIT;
3790 HOST_WIDE_INT dst_align = src_align;
3791 rtx orig_src = operands[1];
3792 rtx orig_dst = operands[0];
3793 rtx data_regs[2 * MAX_MOVE_WORDS + 16];
3794 rtx tmp;
3795 unsigned int i, words, ofs, nregs = 0;
3797 if (orig_bytes <= 0)
3798 return 1;
3799 else if (orig_bytes > MAX_MOVE_WORDS * UNITS_PER_WORD)
3800 return 0;
3802 /* Look for additional alignment information from recorded register info. */
3804 tmp = XEXP (orig_src, 0);
3805 if (REG_P (tmp))
3806 src_align = MAX (src_align, REGNO_POINTER_ALIGN (REGNO (tmp)));
3807 else if (GET_CODE (tmp) == PLUS
3808 && REG_P (XEXP (tmp, 0))
3809 && CONST_INT_P (XEXP (tmp, 1)))
3811 unsigned HOST_WIDE_INT c = INTVAL (XEXP (tmp, 1));
3812 unsigned int a = REGNO_POINTER_ALIGN (REGNO (XEXP (tmp, 0)));
3814 if (a > src_align)
3816 if (a >= 64 && c % 8 == 0)
3817 src_align = 64;
3818 else if (a >= 32 && c % 4 == 0)
3819 src_align = 32;
3820 else if (a >= 16 && c % 2 == 0)
3821 src_align = 16;
3825 tmp = XEXP (orig_dst, 0);
3826 if (REG_P (tmp))
3827 dst_align = MAX (dst_align, REGNO_POINTER_ALIGN (REGNO (tmp)));
3828 else if (GET_CODE (tmp) == PLUS
3829 && REG_P (XEXP (tmp, 0))
3830 && CONST_INT_P (XEXP (tmp, 1)))
3832 unsigned HOST_WIDE_INT c = INTVAL (XEXP (tmp, 1));
3833 unsigned int a = REGNO_POINTER_ALIGN (REGNO (XEXP (tmp, 0)));
3835 if (a > dst_align)
3837 if (a >= 64 && c % 8 == 0)
3838 dst_align = 64;
3839 else if (a >= 32 && c % 4 == 0)
3840 dst_align = 32;
3841 else if (a >= 16 && c % 2 == 0)
3842 dst_align = 16;
3846 ofs = 0;
3847 if (src_align >= 64 && bytes >= 8)
3849 words = bytes / 8;
3851 for (i = 0; i < words; ++i)
3852 data_regs[nregs + i] = gen_reg_rtx (DImode);
3854 for (i = 0; i < words; ++i)
3855 emit_move_insn (data_regs[nregs + i],
3856 adjust_address (orig_src, DImode, ofs + i * 8));
3858 nregs += words;
3859 bytes -= words * 8;
3860 ofs += words * 8;
3863 if (src_align >= 32 && bytes >= 4)
3865 words = bytes / 4;
3867 for (i = 0; i < words; ++i)
3868 data_regs[nregs + i] = gen_reg_rtx (SImode);
3870 for (i = 0; i < words; ++i)
3871 emit_move_insn (data_regs[nregs + i],
3872 adjust_address (orig_src, SImode, ofs + i * 4));
3874 nregs += words;
3875 bytes -= words * 4;
3876 ofs += words * 4;
3879 if (bytes >= 8)
3881 words = bytes / 8;
3883 for (i = 0; i < words+1; ++i)
3884 data_regs[nregs + i] = gen_reg_rtx (DImode);
3886 alpha_expand_unaligned_load_words (data_regs + nregs, orig_src,
3887 words, ofs);
3889 nregs += words;
3890 bytes -= words * 8;
3891 ofs += words * 8;
3894 if (! TARGET_BWX && bytes >= 4)
3896 data_regs[nregs++] = tmp = gen_reg_rtx (SImode);
3897 alpha_expand_unaligned_load (tmp, orig_src, 4, ofs, 0);
3898 bytes -= 4;
3899 ofs += 4;
3902 if (bytes >= 2)
3904 if (src_align >= 16)
3906 do {
3907 data_regs[nregs++] = tmp = gen_reg_rtx (HImode);
3908 emit_move_insn (tmp, adjust_address (orig_src, HImode, ofs));
3909 bytes -= 2;
3910 ofs += 2;
3911 } while (bytes >= 2);
3913 else if (! TARGET_BWX)
3915 data_regs[nregs++] = tmp = gen_reg_rtx (HImode);
3916 alpha_expand_unaligned_load (tmp, orig_src, 2, ofs, 0);
3917 bytes -= 2;
3918 ofs += 2;
3922 while (bytes > 0)
3924 data_regs[nregs++] = tmp = gen_reg_rtx (QImode);
3925 emit_move_insn (tmp, adjust_address (orig_src, QImode, ofs));
3926 bytes -= 1;
3927 ofs += 1;
3930 gcc_assert (nregs <= ARRAY_SIZE (data_regs));
3932 /* Now save it back out again. */
3934 i = 0, ofs = 0;
3936 /* Write out the data in whatever chunks reading the source allowed. */
3937 if (dst_align >= 64)
3939 while (i < nregs && GET_MODE (data_regs[i]) == DImode)
3941 emit_move_insn (adjust_address (orig_dst, DImode, ofs),
3942 data_regs[i]);
3943 ofs += 8;
3944 i++;
3948 if (dst_align >= 32)
3950 /* If the source has remaining DImode regs, write them out in
3951 two pieces. */
3952 while (i < nregs && GET_MODE (data_regs[i]) == DImode)
3954 tmp = expand_binop (DImode, lshr_optab, data_regs[i], GEN_INT (32),
3955 NULL_RTX, 1, OPTAB_WIDEN);
3957 emit_move_insn (adjust_address (orig_dst, SImode, ofs),
3958 gen_lowpart (SImode, data_regs[i]));
3959 emit_move_insn (adjust_address (orig_dst, SImode, ofs + 4),
3960 gen_lowpart (SImode, tmp));
3961 ofs += 8;
3962 i++;
3965 while (i < nregs && GET_MODE (data_regs[i]) == SImode)
3967 emit_move_insn (adjust_address (orig_dst, SImode, ofs),
3968 data_regs[i]);
3969 ofs += 4;
3970 i++;
3974 if (i < nregs && GET_MODE (data_regs[i]) == DImode)
3976 /* Write out a remaining block of words using unaligned methods. */
3978 for (words = 1; i + words < nregs; words++)
3979 if (GET_MODE (data_regs[i + words]) != DImode)
3980 break;
3982 if (words == 1)
3983 alpha_expand_unaligned_store (orig_dst, data_regs[i], 8, ofs);
3984 else
3985 alpha_expand_unaligned_store_words (data_regs + i, orig_dst,
3986 words, ofs);
3988 i += words;
3989 ofs += words * 8;
3992 /* Due to the above, this won't be aligned. */
3993 /* ??? If we have more than one of these, consider constructing full
3994 words in registers and using alpha_expand_unaligned_store_words. */
3995 while (i < nregs && GET_MODE (data_regs[i]) == SImode)
3997 alpha_expand_unaligned_store (orig_dst, data_regs[i], 4, ofs);
3998 ofs += 4;
3999 i++;
4002 if (dst_align >= 16)
4003 while (i < nregs && GET_MODE (data_regs[i]) == HImode)
4005 emit_move_insn (adjust_address (orig_dst, HImode, ofs), data_regs[i]);
4006 i++;
4007 ofs += 2;
4009 else
4010 while (i < nregs && GET_MODE (data_regs[i]) == HImode)
4012 alpha_expand_unaligned_store (orig_dst, data_regs[i], 2, ofs);
4013 i++;
4014 ofs += 2;
4017 /* The remainder must be byte copies. */
4018 while (i < nregs)
4020 gcc_assert (GET_MODE (data_regs[i]) == QImode);
4021 emit_move_insn (adjust_address (orig_dst, QImode, ofs), data_regs[i]);
4022 i++;
4023 ofs += 1;
4026 return 1;
4030 alpha_expand_block_clear (rtx operands[])
4032 rtx bytes_rtx = operands[1];
4033 rtx align_rtx = operands[3];
4034 HOST_WIDE_INT orig_bytes = INTVAL (bytes_rtx);
4035 HOST_WIDE_INT bytes = orig_bytes;
4036 HOST_WIDE_INT align = INTVAL (align_rtx) * BITS_PER_UNIT;
4037 HOST_WIDE_INT alignofs = 0;
4038 rtx orig_dst = operands[0];
4039 rtx tmp;
4040 int i, words, ofs = 0;
4042 if (orig_bytes <= 0)
4043 return 1;
4044 if (orig_bytes > MAX_MOVE_WORDS * UNITS_PER_WORD)
4045 return 0;
4047 /* Look for stricter alignment. */
4048 tmp = XEXP (orig_dst, 0);
4049 if (REG_P (tmp))
4050 align = MAX (align, REGNO_POINTER_ALIGN (REGNO (tmp)));
4051 else if (GET_CODE (tmp) == PLUS
4052 && REG_P (XEXP (tmp, 0))
4053 && CONST_INT_P (XEXP (tmp, 1)))
4055 HOST_WIDE_INT c = INTVAL (XEXP (tmp, 1));
4056 int a = REGNO_POINTER_ALIGN (REGNO (XEXP (tmp, 0)));
4058 if (a > align)
4060 if (a >= 64)
4061 align = a, alignofs = 8 - c % 8;
4062 else if (a >= 32)
4063 align = a, alignofs = 4 - c % 4;
4064 else if (a >= 16)
4065 align = a, alignofs = 2 - c % 2;
4069 /* Handle an unaligned prefix first. */
4071 if (alignofs > 0)
4073 #if HOST_BITS_PER_WIDE_INT >= 64
4074 /* Given that alignofs is bounded by align, the only time BWX could
4075 generate three stores is for a 7 byte fill. Prefer two individual
4076 stores over a load/mask/store sequence. */
4077 if ((!TARGET_BWX || alignofs == 7)
4078 && align >= 32
4079 && !(alignofs == 4 && bytes >= 4))
4081 enum machine_mode mode = (align >= 64 ? DImode : SImode);
4082 int inv_alignofs = (align >= 64 ? 8 : 4) - alignofs;
4083 rtx mem, tmp;
4084 HOST_WIDE_INT mask;
4086 mem = adjust_address (orig_dst, mode, ofs - inv_alignofs);
4087 set_mem_alias_set (mem, 0);
4089 mask = ~(~(HOST_WIDE_INT)0 << (inv_alignofs * 8));
4090 if (bytes < alignofs)
4092 mask |= ~(HOST_WIDE_INT)0 << ((inv_alignofs + bytes) * 8);
4093 ofs += bytes;
4094 bytes = 0;
4096 else
4098 bytes -= alignofs;
4099 ofs += alignofs;
4101 alignofs = 0;
4103 tmp = expand_binop (mode, and_optab, mem, GEN_INT (mask),
4104 NULL_RTX, 1, OPTAB_WIDEN);
4106 emit_move_insn (mem, tmp);
4108 #endif
4110 if (TARGET_BWX && (alignofs & 1) && bytes >= 1)
4112 emit_move_insn (adjust_address (orig_dst, QImode, ofs), const0_rtx);
4113 bytes -= 1;
4114 ofs += 1;
4115 alignofs -= 1;
4117 if (TARGET_BWX && align >= 16 && (alignofs & 3) == 2 && bytes >= 2)
4119 emit_move_insn (adjust_address (orig_dst, HImode, ofs), const0_rtx);
4120 bytes -= 2;
4121 ofs += 2;
4122 alignofs -= 2;
4124 if (alignofs == 4 && bytes >= 4)
4126 emit_move_insn (adjust_address (orig_dst, SImode, ofs), const0_rtx);
4127 bytes -= 4;
4128 ofs += 4;
4129 alignofs = 0;
4132 /* If we've not used the extra lead alignment information by now,
4133 we won't be able to. Downgrade align to match what's left over. */
4134 if (alignofs > 0)
4136 alignofs = alignofs & -alignofs;
4137 align = MIN (align, alignofs * BITS_PER_UNIT);
4141 /* Handle a block of contiguous long-words. */
4143 if (align >= 64 && bytes >= 8)
4145 words = bytes / 8;
4147 for (i = 0; i < words; ++i)
4148 emit_move_insn (adjust_address (orig_dst, DImode, ofs + i * 8),
4149 const0_rtx);
4151 bytes -= words * 8;
4152 ofs += words * 8;
4155 /* If the block is large and appropriately aligned, emit a single
4156 store followed by a sequence of stq_u insns. */
4158 if (align >= 32 && bytes > 16)
4160 rtx orig_dsta;
4162 emit_move_insn (adjust_address (orig_dst, SImode, ofs), const0_rtx);
4163 bytes -= 4;
4164 ofs += 4;
4166 orig_dsta = XEXP (orig_dst, 0);
4167 if (GET_CODE (orig_dsta) == LO_SUM)
4168 orig_dsta = force_reg (Pmode, orig_dsta);
4170 words = bytes / 8;
4171 for (i = 0; i < words; ++i)
4173 rtx mem
4174 = change_address (orig_dst, DImode,
4175 gen_rtx_AND (DImode,
4176 plus_constant (DImode, orig_dsta,
4177 ofs + i*8),
4178 GEN_INT (-8)));
4179 set_mem_alias_set (mem, 0);
4180 emit_move_insn (mem, const0_rtx);
4183 /* Depending on the alignment, the first stq_u may have overlapped
4184 with the initial stl, which means that the last stq_u didn't
4185 write as much as it would appear. Leave those questionable bytes
4186 unaccounted for. */
4187 bytes -= words * 8 - 4;
4188 ofs += words * 8 - 4;
4191 /* Handle a smaller block of aligned words. */
4193 if ((align >= 64 && bytes == 4)
4194 || (align == 32 && bytes >= 4))
4196 words = bytes / 4;
4198 for (i = 0; i < words; ++i)
4199 emit_move_insn (adjust_address (orig_dst, SImode, ofs + i * 4),
4200 const0_rtx);
4202 bytes -= words * 4;
4203 ofs += words * 4;
4206 /* An unaligned block uses stq_u stores for as many as possible. */
4208 if (bytes >= 8)
4210 words = bytes / 8;
4212 alpha_expand_unaligned_store_words (NULL, orig_dst, words, ofs);
4214 bytes -= words * 8;
4215 ofs += words * 8;
4218 /* Next clean up any trailing pieces. */
4220 #if HOST_BITS_PER_WIDE_INT >= 64
4221 /* Count the number of bits in BYTES for which aligned stores could
4222 be emitted. */
4223 words = 0;
4224 for (i = (TARGET_BWX ? 1 : 4); i * BITS_PER_UNIT <= align ; i <<= 1)
4225 if (bytes & i)
4226 words += 1;
4228 /* If we have appropriate alignment (and it wouldn't take too many
4229 instructions otherwise), mask out the bytes we need. */
4230 if (TARGET_BWX ? words > 2 : bytes > 0)
4232 if (align >= 64)
4234 rtx mem, tmp;
4235 HOST_WIDE_INT mask;
4237 mem = adjust_address (orig_dst, DImode, ofs);
4238 set_mem_alias_set (mem, 0);
4240 mask = ~(HOST_WIDE_INT)0 << (bytes * 8);
4242 tmp = expand_binop (DImode, and_optab, mem, GEN_INT (mask),
4243 NULL_RTX, 1, OPTAB_WIDEN);
4245 emit_move_insn (mem, tmp);
4246 return 1;
4248 else if (align >= 32 && bytes < 4)
4250 rtx mem, tmp;
4251 HOST_WIDE_INT mask;
4253 mem = adjust_address (orig_dst, SImode, ofs);
4254 set_mem_alias_set (mem, 0);
4256 mask = ~(HOST_WIDE_INT)0 << (bytes * 8);
4258 tmp = expand_binop (SImode, and_optab, mem, GEN_INT (mask),
4259 NULL_RTX, 1, OPTAB_WIDEN);
4261 emit_move_insn (mem, tmp);
4262 return 1;
4265 #endif
4267 if (!TARGET_BWX && bytes >= 4)
4269 alpha_expand_unaligned_store (orig_dst, const0_rtx, 4, ofs);
4270 bytes -= 4;
4271 ofs += 4;
4274 if (bytes >= 2)
4276 if (align >= 16)
4278 do {
4279 emit_move_insn (adjust_address (orig_dst, HImode, ofs),
4280 const0_rtx);
4281 bytes -= 2;
4282 ofs += 2;
4283 } while (bytes >= 2);
4285 else if (! TARGET_BWX)
4287 alpha_expand_unaligned_store (orig_dst, const0_rtx, 2, ofs);
4288 bytes -= 2;
4289 ofs += 2;
4293 while (bytes > 0)
4295 emit_move_insn (adjust_address (orig_dst, QImode, ofs), const0_rtx);
4296 bytes -= 1;
4297 ofs += 1;
4300 return 1;
4303 /* Returns a mask so that zap(x, value) == x & mask. */
4306 alpha_expand_zap_mask (HOST_WIDE_INT value)
4308 rtx result;
4309 int i;
4311 if (HOST_BITS_PER_WIDE_INT >= 64)
4313 HOST_WIDE_INT mask = 0;
4315 for (i = 7; i >= 0; --i)
4317 mask <<= 8;
4318 if (!((value >> i) & 1))
4319 mask |= 0xff;
4322 result = gen_int_mode (mask, DImode);
4324 else
4326 HOST_WIDE_INT mask_lo = 0, mask_hi = 0;
4328 gcc_assert (HOST_BITS_PER_WIDE_INT == 32);
4330 for (i = 7; i >= 4; --i)
4332 mask_hi <<= 8;
4333 if (!((value >> i) & 1))
4334 mask_hi |= 0xff;
4337 for (i = 3; i >= 0; --i)
4339 mask_lo <<= 8;
4340 if (!((value >> i) & 1))
4341 mask_lo |= 0xff;
4344 result = immed_double_const (mask_lo, mask_hi, DImode);
4347 return result;
4350 void
4351 alpha_expand_builtin_vector_binop (rtx (*gen) (rtx, rtx, rtx),
4352 enum machine_mode mode,
4353 rtx op0, rtx op1, rtx op2)
4355 op0 = gen_lowpart (mode, op0);
4357 if (op1 == const0_rtx)
4358 op1 = CONST0_RTX (mode);
4359 else
4360 op1 = gen_lowpart (mode, op1);
4362 if (op2 == const0_rtx)
4363 op2 = CONST0_RTX (mode);
4364 else
4365 op2 = gen_lowpart (mode, op2);
4367 emit_insn ((*gen) (op0, op1, op2));
4370 /* A subroutine of the atomic operation splitters. Jump to LABEL if
4371 COND is true. Mark the jump as unlikely to be taken. */
4373 static void
4374 emit_unlikely_jump (rtx cond, rtx label)
4376 int very_unlikely = REG_BR_PROB_BASE / 100 - 1;
4377 rtx x;
4379 x = gen_rtx_IF_THEN_ELSE (VOIDmode, cond, label, pc_rtx);
4380 x = emit_jump_insn (gen_rtx_SET (VOIDmode, pc_rtx, x));
4381 add_int_reg_note (x, REG_BR_PROB, very_unlikely);
4384 /* A subroutine of the atomic operation splitters. Emit a load-locked
4385 instruction in MODE. */
4387 static void
4388 emit_load_locked (enum machine_mode mode, rtx reg, rtx mem)
4390 rtx (*fn) (rtx, rtx) = NULL;
4391 if (mode == SImode)
4392 fn = gen_load_locked_si;
4393 else if (mode == DImode)
4394 fn = gen_load_locked_di;
4395 emit_insn (fn (reg, mem));
4398 /* A subroutine of the atomic operation splitters. Emit a store-conditional
4399 instruction in MODE. */
4401 static void
4402 emit_store_conditional (enum machine_mode mode, rtx res, rtx mem, rtx val)
4404 rtx (*fn) (rtx, rtx, rtx) = NULL;
4405 if (mode == SImode)
4406 fn = gen_store_conditional_si;
4407 else if (mode == DImode)
4408 fn = gen_store_conditional_di;
4409 emit_insn (fn (res, mem, val));
4412 /* Subroutines of the atomic operation splitters. Emit barriers
4413 as needed for the memory MODEL. */
4415 static void
4416 alpha_pre_atomic_barrier (enum memmodel model)
4418 if (need_atomic_barrier_p (model, true))
4419 emit_insn (gen_memory_barrier ());
4422 static void
4423 alpha_post_atomic_barrier (enum memmodel model)
4425 if (need_atomic_barrier_p (model, false))
4426 emit_insn (gen_memory_barrier ());
4429 /* A subroutine of the atomic operation splitters. Emit an insxl
4430 instruction in MODE. */
4432 static rtx
4433 emit_insxl (enum machine_mode mode, rtx op1, rtx op2)
4435 rtx ret = gen_reg_rtx (DImode);
4436 rtx (*fn) (rtx, rtx, rtx);
4438 switch (mode)
4440 case QImode:
4441 fn = gen_insbl;
4442 break;
4443 case HImode:
4444 fn = gen_inswl;
4445 break;
4446 case SImode:
4447 fn = gen_insll;
4448 break;
4449 case DImode:
4450 fn = gen_insql;
4451 break;
4452 default:
4453 gcc_unreachable ();
4456 op1 = force_reg (mode, op1);
4457 emit_insn (fn (ret, op1, op2));
4459 return ret;
4462 /* Expand an atomic fetch-and-operate pattern. CODE is the binary operation
4463 to perform. MEM is the memory on which to operate. VAL is the second
4464 operand of the binary operator. BEFORE and AFTER are optional locations to
4465 return the value of MEM either before of after the operation. SCRATCH is
4466 a scratch register. */
4468 void
4469 alpha_split_atomic_op (enum rtx_code code, rtx mem, rtx val, rtx before,
4470 rtx after, rtx scratch, enum memmodel model)
4472 enum machine_mode mode = GET_MODE (mem);
4473 rtx label, x, cond = gen_rtx_REG (DImode, REGNO (scratch));
4475 alpha_pre_atomic_barrier (model);
4477 label = gen_label_rtx ();
4478 emit_label (label);
4479 label = gen_rtx_LABEL_REF (DImode, label);
4481 if (before == NULL)
4482 before = scratch;
4483 emit_load_locked (mode, before, mem);
4485 if (code == NOT)
4487 x = gen_rtx_AND (mode, before, val);
4488 emit_insn (gen_rtx_SET (VOIDmode, val, x));
4490 x = gen_rtx_NOT (mode, val);
4492 else
4493 x = gen_rtx_fmt_ee (code, mode, before, val);
4494 if (after)
4495 emit_insn (gen_rtx_SET (VOIDmode, after, copy_rtx (x)));
4496 emit_insn (gen_rtx_SET (VOIDmode, scratch, x));
4498 emit_store_conditional (mode, cond, mem, scratch);
4500 x = gen_rtx_EQ (DImode, cond, const0_rtx);
4501 emit_unlikely_jump (x, label);
4503 alpha_post_atomic_barrier (model);
4506 /* Expand a compare and swap operation. */
4508 void
4509 alpha_split_compare_and_swap (rtx operands[])
4511 rtx cond, retval, mem, oldval, newval;
4512 bool is_weak;
4513 enum memmodel mod_s, mod_f;
4514 enum machine_mode mode;
4515 rtx label1, label2, x;
4517 cond = operands[0];
4518 retval = operands[1];
4519 mem = operands[2];
4520 oldval = operands[3];
4521 newval = operands[4];
4522 is_weak = (operands[5] != const0_rtx);
4523 mod_s = (enum memmodel) INTVAL (operands[6]);
4524 mod_f = (enum memmodel) INTVAL (operands[7]);
4525 mode = GET_MODE (mem);
4527 alpha_pre_atomic_barrier (mod_s);
4529 label1 = NULL_RTX;
4530 if (!is_weak)
4532 label1 = gen_rtx_LABEL_REF (DImode, gen_label_rtx ());
4533 emit_label (XEXP (label1, 0));
4535 label2 = gen_rtx_LABEL_REF (DImode, gen_label_rtx ());
4537 emit_load_locked (mode, retval, mem);
4539 x = gen_lowpart (DImode, retval);
4540 if (oldval == const0_rtx)
4542 emit_move_insn (cond, const0_rtx);
4543 x = gen_rtx_NE (DImode, x, const0_rtx);
4545 else
4547 x = gen_rtx_EQ (DImode, x, oldval);
4548 emit_insn (gen_rtx_SET (VOIDmode, cond, x));
4549 x = gen_rtx_EQ (DImode, cond, const0_rtx);
4551 emit_unlikely_jump (x, label2);
4553 emit_move_insn (cond, newval);
4554 emit_store_conditional (mode, cond, mem, gen_lowpart (mode, cond));
4556 if (!is_weak)
4558 x = gen_rtx_EQ (DImode, cond, const0_rtx);
4559 emit_unlikely_jump (x, label1);
4562 if (mod_f != MEMMODEL_RELAXED)
4563 emit_label (XEXP (label2, 0));
4565 alpha_post_atomic_barrier (mod_s);
4567 if (mod_f == MEMMODEL_RELAXED)
4568 emit_label (XEXP (label2, 0));
4571 void
4572 alpha_expand_compare_and_swap_12 (rtx operands[])
4574 rtx cond, dst, mem, oldval, newval, is_weak, mod_s, mod_f;
4575 enum machine_mode mode;
4576 rtx addr, align, wdst;
4577 rtx (*gen) (rtx, rtx, rtx, rtx, rtx, rtx, rtx, rtx, rtx);
4579 cond = operands[0];
4580 dst = operands[1];
4581 mem = operands[2];
4582 oldval = operands[3];
4583 newval = operands[4];
4584 is_weak = operands[5];
4585 mod_s = operands[6];
4586 mod_f = operands[7];
4587 mode = GET_MODE (mem);
4589 /* We forced the address into a register via mem_noofs_operand. */
4590 addr = XEXP (mem, 0);
4591 gcc_assert (register_operand (addr, DImode));
4593 align = expand_simple_binop (Pmode, AND, addr, GEN_INT (-8),
4594 NULL_RTX, 1, OPTAB_DIRECT);
4596 oldval = convert_modes (DImode, mode, oldval, 1);
4598 if (newval != const0_rtx)
4599 newval = emit_insxl (mode, newval, addr);
4601 wdst = gen_reg_rtx (DImode);
4602 if (mode == QImode)
4603 gen = gen_atomic_compare_and_swapqi_1;
4604 else
4605 gen = gen_atomic_compare_and_swaphi_1;
4606 emit_insn (gen (cond, wdst, mem, oldval, newval, align,
4607 is_weak, mod_s, mod_f));
4609 emit_move_insn (dst, gen_lowpart (mode, wdst));
4612 void
4613 alpha_split_compare_and_swap_12 (rtx operands[])
4615 rtx cond, dest, orig_mem, oldval, newval, align, scratch;
4616 enum machine_mode mode;
4617 bool is_weak;
4618 enum memmodel mod_s, mod_f;
4619 rtx label1, label2, mem, addr, width, mask, x;
4621 cond = operands[0];
4622 dest = operands[1];
4623 orig_mem = operands[2];
4624 oldval = operands[3];
4625 newval = operands[4];
4626 align = operands[5];
4627 is_weak = (operands[6] != const0_rtx);
4628 mod_s = (enum memmodel) INTVAL (operands[7]);
4629 mod_f = (enum memmodel) INTVAL (operands[8]);
4630 scratch = operands[9];
4631 mode = GET_MODE (orig_mem);
4632 addr = XEXP (orig_mem, 0);
4634 mem = gen_rtx_MEM (DImode, align);
4635 MEM_VOLATILE_P (mem) = MEM_VOLATILE_P (orig_mem);
4636 if (MEM_ALIAS_SET (orig_mem) == ALIAS_SET_MEMORY_BARRIER)
4637 set_mem_alias_set (mem, ALIAS_SET_MEMORY_BARRIER);
4639 alpha_pre_atomic_barrier (mod_s);
4641 label1 = NULL_RTX;
4642 if (!is_weak)
4644 label1 = gen_rtx_LABEL_REF (DImode, gen_label_rtx ());
4645 emit_label (XEXP (label1, 0));
4647 label2 = gen_rtx_LABEL_REF (DImode, gen_label_rtx ());
4649 emit_load_locked (DImode, scratch, mem);
4651 width = GEN_INT (GET_MODE_BITSIZE (mode));
4652 mask = GEN_INT (mode == QImode ? 0xff : 0xffff);
4653 emit_insn (gen_extxl (dest, scratch, width, addr));
4655 if (oldval == const0_rtx)
4657 emit_move_insn (cond, const0_rtx);
4658 x = gen_rtx_NE (DImode, dest, const0_rtx);
4660 else
4662 x = gen_rtx_EQ (DImode, dest, oldval);
4663 emit_insn (gen_rtx_SET (VOIDmode, cond, x));
4664 x = gen_rtx_EQ (DImode, cond, const0_rtx);
4666 emit_unlikely_jump (x, label2);
4668 emit_insn (gen_mskxl (cond, scratch, mask, addr));
4670 if (newval != const0_rtx)
4671 emit_insn (gen_iordi3 (cond, cond, newval));
4673 emit_store_conditional (DImode, cond, mem, cond);
4675 if (!is_weak)
4677 x = gen_rtx_EQ (DImode, cond, const0_rtx);
4678 emit_unlikely_jump (x, label1);
4681 if (mod_f != MEMMODEL_RELAXED)
4682 emit_label (XEXP (label2, 0));
4684 alpha_post_atomic_barrier (mod_s);
4686 if (mod_f == MEMMODEL_RELAXED)
4687 emit_label (XEXP (label2, 0));
4690 /* Expand an atomic exchange operation. */
4692 void
4693 alpha_split_atomic_exchange (rtx operands[])
4695 rtx retval, mem, val, scratch;
4696 enum memmodel model;
4697 enum machine_mode mode;
4698 rtx label, x, cond;
4700 retval = operands[0];
4701 mem = operands[1];
4702 val = operands[2];
4703 model = (enum memmodel) INTVAL (operands[3]);
4704 scratch = operands[4];
4705 mode = GET_MODE (mem);
4706 cond = gen_lowpart (DImode, scratch);
4708 alpha_pre_atomic_barrier (model);
4710 label = gen_rtx_LABEL_REF (DImode, gen_label_rtx ());
4711 emit_label (XEXP (label, 0));
4713 emit_load_locked (mode, retval, mem);
4714 emit_move_insn (scratch, val);
4715 emit_store_conditional (mode, cond, mem, scratch);
4717 x = gen_rtx_EQ (DImode, cond, const0_rtx);
4718 emit_unlikely_jump (x, label);
4720 alpha_post_atomic_barrier (model);
4723 void
4724 alpha_expand_atomic_exchange_12 (rtx operands[])
4726 rtx dst, mem, val, model;
4727 enum machine_mode mode;
4728 rtx addr, align, wdst;
4729 rtx (*gen) (rtx, rtx, rtx, rtx, rtx);
4731 dst = operands[0];
4732 mem = operands[1];
4733 val = operands[2];
4734 model = operands[3];
4735 mode = GET_MODE (mem);
4737 /* We forced the address into a register via mem_noofs_operand. */
4738 addr = XEXP (mem, 0);
4739 gcc_assert (register_operand (addr, DImode));
4741 align = expand_simple_binop (Pmode, AND, addr, GEN_INT (-8),
4742 NULL_RTX, 1, OPTAB_DIRECT);
4744 /* Insert val into the correct byte location within the word. */
4745 if (val != const0_rtx)
4746 val = emit_insxl (mode, val, addr);
4748 wdst = gen_reg_rtx (DImode);
4749 if (mode == QImode)
4750 gen = gen_atomic_exchangeqi_1;
4751 else
4752 gen = gen_atomic_exchangehi_1;
4753 emit_insn (gen (wdst, mem, val, align, model));
4755 emit_move_insn (dst, gen_lowpart (mode, wdst));
4758 void
4759 alpha_split_atomic_exchange_12 (rtx operands[])
4761 rtx dest, orig_mem, addr, val, align, scratch;
4762 rtx label, mem, width, mask, x;
4763 enum machine_mode mode;
4764 enum memmodel model;
4766 dest = operands[0];
4767 orig_mem = operands[1];
4768 val = operands[2];
4769 align = operands[3];
4770 model = (enum memmodel) INTVAL (operands[4]);
4771 scratch = operands[5];
4772 mode = GET_MODE (orig_mem);
4773 addr = XEXP (orig_mem, 0);
4775 mem = gen_rtx_MEM (DImode, align);
4776 MEM_VOLATILE_P (mem) = MEM_VOLATILE_P (orig_mem);
4777 if (MEM_ALIAS_SET (orig_mem) == ALIAS_SET_MEMORY_BARRIER)
4778 set_mem_alias_set (mem, ALIAS_SET_MEMORY_BARRIER);
4780 alpha_pre_atomic_barrier (model);
4782 label = gen_rtx_LABEL_REF (DImode, gen_label_rtx ());
4783 emit_label (XEXP (label, 0));
4785 emit_load_locked (DImode, scratch, mem);
4787 width = GEN_INT (GET_MODE_BITSIZE (mode));
4788 mask = GEN_INT (mode == QImode ? 0xff : 0xffff);
4789 emit_insn (gen_extxl (dest, scratch, width, addr));
4790 emit_insn (gen_mskxl (scratch, scratch, mask, addr));
4791 if (val != const0_rtx)
4792 emit_insn (gen_iordi3 (scratch, scratch, val));
4794 emit_store_conditional (DImode, scratch, mem, scratch);
4796 x = gen_rtx_EQ (DImode, scratch, const0_rtx);
4797 emit_unlikely_jump (x, label);
4799 alpha_post_atomic_barrier (model);
4802 /* Adjust the cost of a scheduling dependency. Return the new cost of
4803 a dependency LINK or INSN on DEP_INSN. COST is the current cost. */
4805 static int
4806 alpha_adjust_cost (rtx insn, rtx link, rtx dep_insn, int cost)
4808 enum attr_type dep_insn_type;
4810 /* If the dependence is an anti-dependence, there is no cost. For an
4811 output dependence, there is sometimes a cost, but it doesn't seem
4812 worth handling those few cases. */
4813 if (REG_NOTE_KIND (link) != 0)
4814 return cost;
4816 /* If we can't recognize the insns, we can't really do anything. */
4817 if (recog_memoized (insn) < 0 || recog_memoized (dep_insn) < 0)
4818 return cost;
4820 dep_insn_type = get_attr_type (dep_insn);
4822 /* Bring in the user-defined memory latency. */
4823 if (dep_insn_type == TYPE_ILD
4824 || dep_insn_type == TYPE_FLD
4825 || dep_insn_type == TYPE_LDSYM)
4826 cost += alpha_memory_latency-1;
4828 /* Everything else handled in DFA bypasses now. */
4830 return cost;
4833 /* The number of instructions that can be issued per cycle. */
4835 static int
4836 alpha_issue_rate (void)
4838 return (alpha_tune == PROCESSOR_EV4 ? 2 : 4);
4841 /* How many alternative schedules to try. This should be as wide as the
4842 scheduling freedom in the DFA, but no wider. Making this value too
4843 large results extra work for the scheduler.
4845 For EV4, loads can be issued to either IB0 or IB1, thus we have 2
4846 alternative schedules. For EV5, we can choose between E0/E1 and
4847 FA/FM. For EV6, an arithmetic insn can be issued to U0/U1/L0/L1. */
4849 static int
4850 alpha_multipass_dfa_lookahead (void)
4852 return (alpha_tune == PROCESSOR_EV6 ? 4 : 2);
4855 /* Machine-specific function data. */
4857 struct GTY(()) alpha_links;
4859 struct GTY(()) machine_function
4861 /* For OSF. */
4862 const char *some_ld_name;
4864 /* For flag_reorder_blocks_and_partition. */
4865 rtx gp_save_rtx;
4867 /* For VMS condition handlers. */
4868 bool uses_condition_handler;
4870 /* Linkage entries. */
4871 splay_tree GTY ((param1_is (char *), param2_is (struct alpha_links *)))
4872 links;
4875 /* How to allocate a 'struct machine_function'. */
4877 static struct machine_function *
4878 alpha_init_machine_status (void)
4880 return ggc_cleared_alloc<machine_function> ();
4883 /* Support for frame based VMS condition handlers. */
4885 /* A VMS condition handler may be established for a function with a call to
4886 __builtin_establish_vms_condition_handler, and cancelled with a call to
4887 __builtin_revert_vms_condition_handler.
4889 The VMS Condition Handling Facility knows about the existence of a handler
4890 from the procedure descriptor .handler field. As the VMS native compilers,
4891 we store the user specified handler's address at a fixed location in the
4892 stack frame and point the procedure descriptor at a common wrapper which
4893 fetches the real handler's address and issues an indirect call.
4895 The indirection wrapper is "__gcc_shell_handler", provided by libgcc.
4897 We force the procedure kind to PT_STACK, and the fixed frame location is
4898 fp+8, just before the register save area. We use the handler_data field in
4899 the procedure descriptor to state the fp offset at which the installed
4900 handler address can be found. */
4902 #define VMS_COND_HANDLER_FP_OFFSET 8
4904 /* Expand code to store the currently installed user VMS condition handler
4905 into TARGET and install HANDLER as the new condition handler. */
4907 void
4908 alpha_expand_builtin_establish_vms_condition_handler (rtx target, rtx handler)
4910 rtx handler_slot_address = plus_constant (Pmode, hard_frame_pointer_rtx,
4911 VMS_COND_HANDLER_FP_OFFSET);
4913 rtx handler_slot
4914 = gen_rtx_MEM (DImode, handler_slot_address);
4916 emit_move_insn (target, handler_slot);
4917 emit_move_insn (handler_slot, handler);
4919 /* Notify the start/prologue/epilogue emitters that the condition handler
4920 slot is needed. In addition to reserving the slot space, this will force
4921 the procedure kind to PT_STACK so ensure that the hard_frame_pointer_rtx
4922 use above is correct. */
4923 cfun->machine->uses_condition_handler = true;
4926 /* Expand code to store the current VMS condition handler into TARGET and
4927 nullify it. */
4929 void
4930 alpha_expand_builtin_revert_vms_condition_handler (rtx target)
4932 /* We implement this by establishing a null condition handler, with the tiny
4933 side effect of setting uses_condition_handler. This is a little bit
4934 pessimistic if no actual builtin_establish call is ever issued, which is
4935 not a real problem and expected never to happen anyway. */
4937 alpha_expand_builtin_establish_vms_condition_handler (target, const0_rtx);
4940 /* Functions to save and restore alpha_return_addr_rtx. */
4942 /* Start the ball rolling with RETURN_ADDR_RTX. */
4945 alpha_return_addr (int count, rtx frame ATTRIBUTE_UNUSED)
4947 if (count != 0)
4948 return const0_rtx;
4950 return get_hard_reg_initial_val (Pmode, REG_RA);
4953 /* Return or create a memory slot containing the gp value for the current
4954 function. Needed only if TARGET_LD_BUGGY_LDGP. */
4957 alpha_gp_save_rtx (void)
4959 rtx seq, m = cfun->machine->gp_save_rtx;
4961 if (m == NULL)
4963 start_sequence ();
4965 m = assign_stack_local (DImode, UNITS_PER_WORD, BITS_PER_WORD);
4966 m = validize_mem (m);
4967 emit_move_insn (m, pic_offset_table_rtx);
4969 seq = get_insns ();
4970 end_sequence ();
4972 /* We used to simply emit the sequence after entry_of_function.
4973 However this breaks the CFG if the first instruction in the
4974 first block is not the NOTE_INSN_BASIC_BLOCK, for example a
4975 label. Emit the sequence properly on the edge. We are only
4976 invoked from dw2_build_landing_pads and finish_eh_generation
4977 will call commit_edge_insertions thanks to a kludge. */
4978 insert_insn_on_edge (seq,
4979 single_succ_edge (ENTRY_BLOCK_PTR_FOR_FN (cfun)));
4981 cfun->machine->gp_save_rtx = m;
4984 return m;
4987 static void
4988 alpha_instantiate_decls (void)
4990 if (cfun->machine->gp_save_rtx != NULL_RTX)
4991 instantiate_decl_rtl (cfun->machine->gp_save_rtx);
4994 static int
4995 alpha_ra_ever_killed (void)
4997 rtx top;
4999 if (!has_hard_reg_initial_val (Pmode, REG_RA))
5000 return (int)df_regs_ever_live_p (REG_RA);
5002 push_topmost_sequence ();
5003 top = get_insns ();
5004 pop_topmost_sequence ();
5006 return reg_set_between_p (gen_rtx_REG (Pmode, REG_RA), top, NULL_RTX);
5010 /* Return the trap mode suffix applicable to the current
5011 instruction, or NULL. */
5013 static const char *
5014 get_trap_mode_suffix (void)
5016 enum attr_trap_suffix s = get_attr_trap_suffix (current_output_insn);
5018 switch (s)
5020 case TRAP_SUFFIX_NONE:
5021 return NULL;
5023 case TRAP_SUFFIX_SU:
5024 if (alpha_fptm >= ALPHA_FPTM_SU)
5025 return "su";
5026 return NULL;
5028 case TRAP_SUFFIX_SUI:
5029 if (alpha_fptm >= ALPHA_FPTM_SUI)
5030 return "sui";
5031 return NULL;
5033 case TRAP_SUFFIX_V_SV:
5034 switch (alpha_fptm)
5036 case ALPHA_FPTM_N:
5037 return NULL;
5038 case ALPHA_FPTM_U:
5039 return "v";
5040 case ALPHA_FPTM_SU:
5041 case ALPHA_FPTM_SUI:
5042 return "sv";
5043 default:
5044 gcc_unreachable ();
5047 case TRAP_SUFFIX_V_SV_SVI:
5048 switch (alpha_fptm)
5050 case ALPHA_FPTM_N:
5051 return NULL;
5052 case ALPHA_FPTM_U:
5053 return "v";
5054 case ALPHA_FPTM_SU:
5055 return "sv";
5056 case ALPHA_FPTM_SUI:
5057 return "svi";
5058 default:
5059 gcc_unreachable ();
5061 break;
5063 case TRAP_SUFFIX_U_SU_SUI:
5064 switch (alpha_fptm)
5066 case ALPHA_FPTM_N:
5067 return NULL;
5068 case ALPHA_FPTM_U:
5069 return "u";
5070 case ALPHA_FPTM_SU:
5071 return "su";
5072 case ALPHA_FPTM_SUI:
5073 return "sui";
5074 default:
5075 gcc_unreachable ();
5077 break;
5079 default:
5080 gcc_unreachable ();
5082 gcc_unreachable ();
5085 /* Return the rounding mode suffix applicable to the current
5086 instruction, or NULL. */
5088 static const char *
5089 get_round_mode_suffix (void)
5091 enum attr_round_suffix s = get_attr_round_suffix (current_output_insn);
5093 switch (s)
5095 case ROUND_SUFFIX_NONE:
5096 return NULL;
5097 case ROUND_SUFFIX_NORMAL:
5098 switch (alpha_fprm)
5100 case ALPHA_FPRM_NORM:
5101 return NULL;
5102 case ALPHA_FPRM_MINF:
5103 return "m";
5104 case ALPHA_FPRM_CHOP:
5105 return "c";
5106 case ALPHA_FPRM_DYN:
5107 return "d";
5108 default:
5109 gcc_unreachable ();
5111 break;
5113 case ROUND_SUFFIX_C:
5114 return "c";
5116 default:
5117 gcc_unreachable ();
5119 gcc_unreachable ();
5122 /* Locate some local-dynamic symbol still in use by this function
5123 so that we can print its name in some movdi_er_tlsldm pattern. */
5125 static int
5126 get_some_local_dynamic_name_1 (rtx *px, void *data ATTRIBUTE_UNUSED)
5128 rtx x = *px;
5130 if (GET_CODE (x) == SYMBOL_REF
5131 && SYMBOL_REF_TLS_MODEL (x) == TLS_MODEL_LOCAL_DYNAMIC)
5133 cfun->machine->some_ld_name = XSTR (x, 0);
5134 return 1;
5137 return 0;
5140 static const char *
5141 get_some_local_dynamic_name (void)
5143 rtx insn;
5145 if (cfun->machine->some_ld_name)
5146 return cfun->machine->some_ld_name;
5148 for (insn = get_insns (); insn ; insn = NEXT_INSN (insn))
5149 if (INSN_P (insn)
5150 && for_each_rtx (&PATTERN (insn), get_some_local_dynamic_name_1, 0))
5151 return cfun->machine->some_ld_name;
5153 gcc_unreachable ();
5156 /* Print an operand. Recognize special options, documented below. */
5158 void
5159 print_operand (FILE *file, rtx x, int code)
5161 int i;
5163 switch (code)
5165 case '~':
5166 /* Print the assembler name of the current function. */
5167 assemble_name (file, alpha_fnname);
5168 break;
5170 case '&':
5171 assemble_name (file, get_some_local_dynamic_name ());
5172 break;
5174 case '/':
5176 const char *trap = get_trap_mode_suffix ();
5177 const char *round = get_round_mode_suffix ();
5179 if (trap || round)
5180 fprintf (file, "/%s%s", (trap ? trap : ""), (round ? round : ""));
5181 break;
5184 case ',':
5185 /* Generates single precision instruction suffix. */
5186 fputc ((TARGET_FLOAT_VAX ? 'f' : 's'), file);
5187 break;
5189 case '-':
5190 /* Generates double precision instruction suffix. */
5191 fputc ((TARGET_FLOAT_VAX ? 'g' : 't'), file);
5192 break;
5194 case '#':
5195 if (alpha_this_literal_sequence_number == 0)
5196 alpha_this_literal_sequence_number = alpha_next_sequence_number++;
5197 fprintf (file, "%d", alpha_this_literal_sequence_number);
5198 break;
5200 case '*':
5201 if (alpha_this_gpdisp_sequence_number == 0)
5202 alpha_this_gpdisp_sequence_number = alpha_next_sequence_number++;
5203 fprintf (file, "%d", alpha_this_gpdisp_sequence_number);
5204 break;
5206 case 'H':
5207 if (GET_CODE (x) == HIGH)
5208 output_addr_const (file, XEXP (x, 0));
5209 else
5210 output_operand_lossage ("invalid %%H value");
5211 break;
5213 case 'J':
5215 const char *lituse;
5217 if (GET_CODE (x) == UNSPEC && XINT (x, 1) == UNSPEC_TLSGD_CALL)
5219 x = XVECEXP (x, 0, 0);
5220 lituse = "lituse_tlsgd";
5222 else if (GET_CODE (x) == UNSPEC && XINT (x, 1) == UNSPEC_TLSLDM_CALL)
5224 x = XVECEXP (x, 0, 0);
5225 lituse = "lituse_tlsldm";
5227 else if (CONST_INT_P (x))
5228 lituse = "lituse_jsr";
5229 else
5231 output_operand_lossage ("invalid %%J value");
5232 break;
5235 if (x != const0_rtx)
5236 fprintf (file, "\t\t!%s!%d", lituse, (int) INTVAL (x));
5238 break;
5240 case 'j':
5242 const char *lituse;
5244 #ifdef HAVE_AS_JSRDIRECT_RELOCS
5245 lituse = "lituse_jsrdirect";
5246 #else
5247 lituse = "lituse_jsr";
5248 #endif
5250 gcc_assert (INTVAL (x) != 0);
5251 fprintf (file, "\t\t!%s!%d", lituse, (int) INTVAL (x));
5253 break;
5254 case 'r':
5255 /* If this operand is the constant zero, write it as "$31". */
5256 if (REG_P (x))
5257 fprintf (file, "%s", reg_names[REGNO (x)]);
5258 else if (x == CONST0_RTX (GET_MODE (x)))
5259 fprintf (file, "$31");
5260 else
5261 output_operand_lossage ("invalid %%r value");
5262 break;
5264 case 'R':
5265 /* Similar, but for floating-point. */
5266 if (REG_P (x))
5267 fprintf (file, "%s", reg_names[REGNO (x)]);
5268 else if (x == CONST0_RTX (GET_MODE (x)))
5269 fprintf (file, "$f31");
5270 else
5271 output_operand_lossage ("invalid %%R value");
5272 break;
5274 case 'N':
5275 /* Write the 1's complement of a constant. */
5276 if (!CONST_INT_P (x))
5277 output_operand_lossage ("invalid %%N value");
5279 fprintf (file, HOST_WIDE_INT_PRINT_DEC, ~ INTVAL (x));
5280 break;
5282 case 'P':
5283 /* Write 1 << C, for a constant C. */
5284 if (!CONST_INT_P (x))
5285 output_operand_lossage ("invalid %%P value");
5287 fprintf (file, HOST_WIDE_INT_PRINT_DEC, (HOST_WIDE_INT) 1 << INTVAL (x));
5288 break;
5290 case 'h':
5291 /* Write the high-order 16 bits of a constant, sign-extended. */
5292 if (!CONST_INT_P (x))
5293 output_operand_lossage ("invalid %%h value");
5295 fprintf (file, HOST_WIDE_INT_PRINT_DEC, INTVAL (x) >> 16);
5296 break;
5298 case 'L':
5299 /* Write the low-order 16 bits of a constant, sign-extended. */
5300 if (!CONST_INT_P (x))
5301 output_operand_lossage ("invalid %%L value");
5303 fprintf (file, HOST_WIDE_INT_PRINT_DEC,
5304 (INTVAL (x) & 0xffff) - 2 * (INTVAL (x) & 0x8000));
5305 break;
5307 case 'm':
5308 /* Write mask for ZAP insn. */
5309 if (GET_CODE (x) == CONST_DOUBLE)
5311 HOST_WIDE_INT mask = 0;
5312 HOST_WIDE_INT value;
5314 value = CONST_DOUBLE_LOW (x);
5315 for (i = 0; i < HOST_BITS_PER_WIDE_INT / HOST_BITS_PER_CHAR;
5316 i++, value >>= 8)
5317 if (value & 0xff)
5318 mask |= (1 << i);
5320 value = CONST_DOUBLE_HIGH (x);
5321 for (i = 0; i < HOST_BITS_PER_WIDE_INT / HOST_BITS_PER_CHAR;
5322 i++, value >>= 8)
5323 if (value & 0xff)
5324 mask |= (1 << (i + sizeof (int)));
5326 fprintf (file, HOST_WIDE_INT_PRINT_DEC, mask & 0xff);
5329 else if (CONST_INT_P (x))
5331 HOST_WIDE_INT mask = 0, value = INTVAL (x);
5333 for (i = 0; i < 8; i++, value >>= 8)
5334 if (value & 0xff)
5335 mask |= (1 << i);
5337 fprintf (file, HOST_WIDE_INT_PRINT_DEC, mask);
5339 else
5340 output_operand_lossage ("invalid %%m value");
5341 break;
5343 case 'M':
5344 /* 'b', 'w', 'l', or 'q' as the value of the constant. */
5345 if (!CONST_INT_P (x)
5346 || (INTVAL (x) != 8 && INTVAL (x) != 16
5347 && INTVAL (x) != 32 && INTVAL (x) != 64))
5348 output_operand_lossage ("invalid %%M value");
5350 fprintf (file, "%s",
5351 (INTVAL (x) == 8 ? "b"
5352 : INTVAL (x) == 16 ? "w"
5353 : INTVAL (x) == 32 ? "l"
5354 : "q"));
5355 break;
5357 case 'U':
5358 /* Similar, except do it from the mask. */
5359 if (CONST_INT_P (x))
5361 HOST_WIDE_INT value = INTVAL (x);
5363 if (value == 0xff)
5365 fputc ('b', file);
5366 break;
5368 if (value == 0xffff)
5370 fputc ('w', file);
5371 break;
5373 if (value == 0xffffffff)
5375 fputc ('l', file);
5376 break;
5378 if (value == -1)
5380 fputc ('q', file);
5381 break;
5384 else if (HOST_BITS_PER_WIDE_INT == 32
5385 && GET_CODE (x) == CONST_DOUBLE
5386 && CONST_DOUBLE_LOW (x) == 0xffffffff
5387 && CONST_DOUBLE_HIGH (x) == 0)
5389 fputc ('l', file);
5390 break;
5392 output_operand_lossage ("invalid %%U value");
5393 break;
5395 case 's':
5396 /* Write the constant value divided by 8. */
5397 if (!CONST_INT_P (x)
5398 || (unsigned HOST_WIDE_INT) INTVAL (x) >= 64
5399 || (INTVAL (x) & 7) != 0)
5400 output_operand_lossage ("invalid %%s value");
5402 fprintf (file, HOST_WIDE_INT_PRINT_DEC, INTVAL (x) / 8);
5403 break;
5405 case 'S':
5406 /* Same, except compute (64 - c) / 8 */
5408 if (!CONST_INT_P (x)
5409 && (unsigned HOST_WIDE_INT) INTVAL (x) >= 64
5410 && (INTVAL (x) & 7) != 8)
5411 output_operand_lossage ("invalid %%s value");
5413 fprintf (file, HOST_WIDE_INT_PRINT_DEC, (64 - INTVAL (x)) / 8);
5414 break;
5416 case 'C': case 'D': case 'c': case 'd':
5417 /* Write out comparison name. */
5419 enum rtx_code c = GET_CODE (x);
5421 if (!COMPARISON_P (x))
5422 output_operand_lossage ("invalid %%C value");
5424 else if (code == 'D')
5425 c = reverse_condition (c);
5426 else if (code == 'c')
5427 c = swap_condition (c);
5428 else if (code == 'd')
5429 c = swap_condition (reverse_condition (c));
5431 if (c == LEU)
5432 fprintf (file, "ule");
5433 else if (c == LTU)
5434 fprintf (file, "ult");
5435 else if (c == UNORDERED)
5436 fprintf (file, "un");
5437 else
5438 fprintf (file, "%s", GET_RTX_NAME (c));
5440 break;
5442 case 'E':
5443 /* Write the divide or modulus operator. */
5444 switch (GET_CODE (x))
5446 case DIV:
5447 fprintf (file, "div%s", GET_MODE (x) == SImode ? "l" : "q");
5448 break;
5449 case UDIV:
5450 fprintf (file, "div%su", GET_MODE (x) == SImode ? "l" : "q");
5451 break;
5452 case MOD:
5453 fprintf (file, "rem%s", GET_MODE (x) == SImode ? "l" : "q");
5454 break;
5455 case UMOD:
5456 fprintf (file, "rem%su", GET_MODE (x) == SImode ? "l" : "q");
5457 break;
5458 default:
5459 output_operand_lossage ("invalid %%E value");
5460 break;
5462 break;
5464 case 'A':
5465 /* Write "_u" for unaligned access. */
5466 if (MEM_P (x) && GET_CODE (XEXP (x, 0)) == AND)
5467 fprintf (file, "_u");
5468 break;
5470 case 0:
5471 if (REG_P (x))
5472 fprintf (file, "%s", reg_names[REGNO (x)]);
5473 else if (MEM_P (x))
5474 output_address (XEXP (x, 0));
5475 else if (GET_CODE (x) == CONST && GET_CODE (XEXP (x, 0)) == UNSPEC)
5477 switch (XINT (XEXP (x, 0), 1))
5479 case UNSPEC_DTPREL:
5480 case UNSPEC_TPREL:
5481 output_addr_const (file, XVECEXP (XEXP (x, 0), 0, 0));
5482 break;
5483 default:
5484 output_operand_lossage ("unknown relocation unspec");
5485 break;
5488 else
5489 output_addr_const (file, x);
5490 break;
5492 default:
5493 output_operand_lossage ("invalid %%xn code");
5497 void
5498 print_operand_address (FILE *file, rtx addr)
5500 int basereg = 31;
5501 HOST_WIDE_INT offset = 0;
5503 if (GET_CODE (addr) == AND)
5504 addr = XEXP (addr, 0);
5506 if (GET_CODE (addr) == PLUS
5507 && CONST_INT_P (XEXP (addr, 1)))
5509 offset = INTVAL (XEXP (addr, 1));
5510 addr = XEXP (addr, 0);
5513 if (GET_CODE (addr) == LO_SUM)
5515 const char *reloc16, *reloclo;
5516 rtx op1 = XEXP (addr, 1);
5518 if (GET_CODE (op1) == CONST && GET_CODE (XEXP (op1, 0)) == UNSPEC)
5520 op1 = XEXP (op1, 0);
5521 switch (XINT (op1, 1))
5523 case UNSPEC_DTPREL:
5524 reloc16 = NULL;
5525 reloclo = (alpha_tls_size == 16 ? "dtprel" : "dtprello");
5526 break;
5527 case UNSPEC_TPREL:
5528 reloc16 = NULL;
5529 reloclo = (alpha_tls_size == 16 ? "tprel" : "tprello");
5530 break;
5531 default:
5532 output_operand_lossage ("unknown relocation unspec");
5533 return;
5536 output_addr_const (file, XVECEXP (op1, 0, 0));
5538 else
5540 reloc16 = "gprel";
5541 reloclo = "gprellow";
5542 output_addr_const (file, op1);
5545 if (offset)
5546 fprintf (file, "+" HOST_WIDE_INT_PRINT_DEC, offset);
5548 addr = XEXP (addr, 0);
5549 switch (GET_CODE (addr))
5551 case REG:
5552 basereg = REGNO (addr);
5553 break;
5555 case SUBREG:
5556 basereg = subreg_regno (addr);
5557 break;
5559 default:
5560 gcc_unreachable ();
5563 fprintf (file, "($%d)\t\t!%s", basereg,
5564 (basereg == 29 ? reloc16 : reloclo));
5565 return;
5568 switch (GET_CODE (addr))
5570 case REG:
5571 basereg = REGNO (addr);
5572 break;
5574 case SUBREG:
5575 basereg = subreg_regno (addr);
5576 break;
5578 case CONST_INT:
5579 offset = INTVAL (addr);
5580 break;
5582 case SYMBOL_REF:
5583 gcc_assert(TARGET_ABI_OPEN_VMS || this_is_asm_operands);
5584 fprintf (file, "%s", XSTR (addr, 0));
5585 return;
5587 case CONST:
5588 gcc_assert(TARGET_ABI_OPEN_VMS || this_is_asm_operands);
5589 gcc_assert (GET_CODE (XEXP (addr, 0)) == PLUS
5590 && GET_CODE (XEXP (XEXP (addr, 0), 0)) == SYMBOL_REF);
5591 fprintf (file, "%s+" HOST_WIDE_INT_PRINT_DEC,
5592 XSTR (XEXP (XEXP (addr, 0), 0), 0),
5593 INTVAL (XEXP (XEXP (addr, 0), 1)));
5594 return;
5596 default:
5597 output_operand_lossage ("invalid operand address");
5598 return;
5601 fprintf (file, HOST_WIDE_INT_PRINT_DEC "($%d)", offset, basereg);
5604 /* Emit RTL insns to initialize the variable parts of a trampoline at
5605 M_TRAMP. FNDECL is target function's decl. CHAIN_VALUE is an rtx
5606 for the static chain value for the function. */
5608 static void
5609 alpha_trampoline_init (rtx m_tramp, tree fndecl, rtx chain_value)
5611 rtx fnaddr, mem, word1, word2;
5613 fnaddr = XEXP (DECL_RTL (fndecl), 0);
5615 #ifdef POINTERS_EXTEND_UNSIGNED
5616 fnaddr = convert_memory_address (Pmode, fnaddr);
5617 chain_value = convert_memory_address (Pmode, chain_value);
5618 #endif
5620 if (TARGET_ABI_OPEN_VMS)
5622 const char *fnname;
5623 char *trname;
5625 /* Construct the name of the trampoline entry point. */
5626 fnname = XSTR (fnaddr, 0);
5627 trname = (char *) alloca (strlen (fnname) + 5);
5628 strcpy (trname, fnname);
5629 strcat (trname, "..tr");
5630 fnname = ggc_alloc_string (trname, strlen (trname) + 1);
5631 word2 = gen_rtx_SYMBOL_REF (Pmode, fnname);
5633 /* Trampoline (or "bounded") procedure descriptor is constructed from
5634 the function's procedure descriptor with certain fields zeroed IAW
5635 the VMS calling standard. This is stored in the first quadword. */
5636 word1 = force_reg (DImode, gen_const_mem (DImode, fnaddr));
5637 word1 = expand_and (DImode, word1,
5638 GEN_INT (HOST_WIDE_INT_C (0xffff0fff0000fff0)),
5639 NULL);
5641 else
5643 /* These 4 instructions are:
5644 ldq $1,24($27)
5645 ldq $27,16($27)
5646 jmp $31,($27),0
5648 We don't bother setting the HINT field of the jump; the nop
5649 is merely there for padding. */
5650 word1 = GEN_INT (HOST_WIDE_INT_C (0xa77b0010a43b0018));
5651 word2 = GEN_INT (HOST_WIDE_INT_C (0x47ff041f6bfb0000));
5654 /* Store the first two words, as computed above. */
5655 mem = adjust_address (m_tramp, DImode, 0);
5656 emit_move_insn (mem, word1);
5657 mem = adjust_address (m_tramp, DImode, 8);
5658 emit_move_insn (mem, word2);
5660 /* Store function address and static chain value. */
5661 mem = adjust_address (m_tramp, Pmode, 16);
5662 emit_move_insn (mem, fnaddr);
5663 mem = adjust_address (m_tramp, Pmode, 24);
5664 emit_move_insn (mem, chain_value);
5666 if (TARGET_ABI_OSF)
5668 emit_insn (gen_imb ());
5669 #ifdef HAVE_ENABLE_EXECUTE_STACK
5670 emit_library_call (init_one_libfunc ("__enable_execute_stack"),
5671 LCT_NORMAL, VOIDmode, 1, XEXP (m_tramp, 0), Pmode);
5672 #endif
5676 /* Determine where to put an argument to a function.
5677 Value is zero to push the argument on the stack,
5678 or a hard register in which to store the argument.
5680 MODE is the argument's machine mode.
5681 TYPE is the data type of the argument (as a tree).
5682 This is null for libcalls where that information may
5683 not be available.
5684 CUM is a variable of type CUMULATIVE_ARGS which gives info about
5685 the preceding args and about the function being called.
5686 NAMED is nonzero if this argument is a named parameter
5687 (otherwise it is an extra parameter matching an ellipsis).
5689 On Alpha the first 6 words of args are normally in registers
5690 and the rest are pushed. */
5692 static rtx
5693 alpha_function_arg (cumulative_args_t cum_v, enum machine_mode mode,
5694 const_tree type, bool named ATTRIBUTE_UNUSED)
5696 CUMULATIVE_ARGS *cum = get_cumulative_args (cum_v);
5697 int basereg;
5698 int num_args;
5700 /* Don't get confused and pass small structures in FP registers. */
5701 if (type && AGGREGATE_TYPE_P (type))
5702 basereg = 16;
5703 else
5705 #ifdef ENABLE_CHECKING
5706 /* With alpha_split_complex_arg, we shouldn't see any raw complex
5707 values here. */
5708 gcc_assert (!COMPLEX_MODE_P (mode));
5709 #endif
5711 /* Set up defaults for FP operands passed in FP registers, and
5712 integral operands passed in integer registers. */
5713 if (TARGET_FPREGS && GET_MODE_CLASS (mode) == MODE_FLOAT)
5714 basereg = 32 + 16;
5715 else
5716 basereg = 16;
5719 /* ??? Irritatingly, the definition of CUMULATIVE_ARGS is different for
5720 the two platforms, so we can't avoid conditional compilation. */
5721 #if TARGET_ABI_OPEN_VMS
5723 if (mode == VOIDmode)
5724 return alpha_arg_info_reg_val (*cum);
5726 num_args = cum->num_args;
5727 if (num_args >= 6
5728 || targetm.calls.must_pass_in_stack (mode, type))
5729 return NULL_RTX;
5731 #elif TARGET_ABI_OSF
5733 if (*cum >= 6)
5734 return NULL_RTX;
5735 num_args = *cum;
5737 /* VOID is passed as a special flag for "last argument". */
5738 if (type == void_type_node)
5739 basereg = 16;
5740 else if (targetm.calls.must_pass_in_stack (mode, type))
5741 return NULL_RTX;
5743 #else
5744 #error Unhandled ABI
5745 #endif
5747 return gen_rtx_REG (mode, num_args + basereg);
5750 /* Update the data in CUM to advance over an argument
5751 of mode MODE and data type TYPE.
5752 (TYPE is null for libcalls where that information may not be available.) */
5754 static void
5755 alpha_function_arg_advance (cumulative_args_t cum_v, enum machine_mode mode,
5756 const_tree type, bool named ATTRIBUTE_UNUSED)
5758 CUMULATIVE_ARGS *cum = get_cumulative_args (cum_v);
5759 bool onstack = targetm.calls.must_pass_in_stack (mode, type);
5760 int increment = onstack ? 6 : ALPHA_ARG_SIZE (mode, type, named);
5762 #if TARGET_ABI_OSF
5763 *cum += increment;
5764 #else
5765 if (!onstack && cum->num_args < 6)
5766 cum->atypes[cum->num_args] = alpha_arg_type (mode);
5767 cum->num_args += increment;
5768 #endif
5771 static int
5772 alpha_arg_partial_bytes (cumulative_args_t cum_v,
5773 enum machine_mode mode ATTRIBUTE_UNUSED,
5774 tree type ATTRIBUTE_UNUSED,
5775 bool named ATTRIBUTE_UNUSED)
5777 int words = 0;
5778 CUMULATIVE_ARGS *cum ATTRIBUTE_UNUSED = get_cumulative_args (cum_v);
5780 #if TARGET_ABI_OPEN_VMS
5781 if (cum->num_args < 6
5782 && 6 < cum->num_args + ALPHA_ARG_SIZE (mode, type, named))
5783 words = 6 - cum->num_args;
5784 #elif TARGET_ABI_OSF
5785 if (*cum < 6 && 6 < *cum + ALPHA_ARG_SIZE (mode, type, named))
5786 words = 6 - *cum;
5787 #else
5788 #error Unhandled ABI
5789 #endif
5791 return words * UNITS_PER_WORD;
5795 /* Return true if TYPE must be returned in memory, instead of in registers. */
5797 static bool
5798 alpha_return_in_memory (const_tree type, const_tree fndecl ATTRIBUTE_UNUSED)
5800 enum machine_mode mode = VOIDmode;
5801 int size;
5803 if (type)
5805 mode = TYPE_MODE (type);
5807 /* All aggregates are returned in memory, except on OpenVMS where
5808 records that fit 64 bits should be returned by immediate value
5809 as required by section 3.8.7.1 of the OpenVMS Calling Standard. */
5810 if (TARGET_ABI_OPEN_VMS
5811 && TREE_CODE (type) != ARRAY_TYPE
5812 && (unsigned HOST_WIDE_INT) int_size_in_bytes(type) <= 8)
5813 return false;
5815 if (AGGREGATE_TYPE_P (type))
5816 return true;
5819 size = GET_MODE_SIZE (mode);
5820 switch (GET_MODE_CLASS (mode))
5822 case MODE_VECTOR_FLOAT:
5823 /* Pass all float vectors in memory, like an aggregate. */
5824 return true;
5826 case MODE_COMPLEX_FLOAT:
5827 /* We judge complex floats on the size of their element,
5828 not the size of the whole type. */
5829 size = GET_MODE_UNIT_SIZE (mode);
5830 break;
5832 case MODE_INT:
5833 case MODE_FLOAT:
5834 case MODE_COMPLEX_INT:
5835 case MODE_VECTOR_INT:
5836 break;
5838 default:
5839 /* ??? We get called on all sorts of random stuff from
5840 aggregate_value_p. We must return something, but it's not
5841 clear what's safe to return. Pretend it's a struct I
5842 guess. */
5843 return true;
5846 /* Otherwise types must fit in one register. */
5847 return size > UNITS_PER_WORD;
5850 /* Return true if TYPE should be passed by invisible reference. */
5852 static bool
5853 alpha_pass_by_reference (cumulative_args_t ca ATTRIBUTE_UNUSED,
5854 enum machine_mode mode,
5855 const_tree type ATTRIBUTE_UNUSED,
5856 bool named ATTRIBUTE_UNUSED)
5858 return mode == TFmode || mode == TCmode;
5861 /* Define how to find the value returned by a function. VALTYPE is the
5862 data type of the value (as a tree). If the precise function being
5863 called is known, FUNC is its FUNCTION_DECL; otherwise, FUNC is 0.
5864 MODE is set instead of VALTYPE for libcalls.
5866 On Alpha the value is found in $0 for integer functions and
5867 $f0 for floating-point functions. */
5870 function_value (const_tree valtype, const_tree func ATTRIBUTE_UNUSED,
5871 enum machine_mode mode)
5873 unsigned int regnum, dummy ATTRIBUTE_UNUSED;
5874 enum mode_class mclass;
5876 gcc_assert (!valtype || !alpha_return_in_memory (valtype, func));
5878 if (valtype)
5879 mode = TYPE_MODE (valtype);
5881 mclass = GET_MODE_CLASS (mode);
5882 switch (mclass)
5884 case MODE_INT:
5885 /* Do the same thing as PROMOTE_MODE except for libcalls on VMS,
5886 where we have them returning both SImode and DImode. */
5887 if (!(TARGET_ABI_OPEN_VMS && valtype && AGGREGATE_TYPE_P (valtype)))
5888 PROMOTE_MODE (mode, dummy, valtype);
5889 /* FALLTHRU */
5891 case MODE_COMPLEX_INT:
5892 case MODE_VECTOR_INT:
5893 regnum = 0;
5894 break;
5896 case MODE_FLOAT:
5897 regnum = 32;
5898 break;
5900 case MODE_COMPLEX_FLOAT:
5902 enum machine_mode cmode = GET_MODE_INNER (mode);
5904 return gen_rtx_PARALLEL
5905 (VOIDmode,
5906 gen_rtvec (2,
5907 gen_rtx_EXPR_LIST (VOIDmode, gen_rtx_REG (cmode, 32),
5908 const0_rtx),
5909 gen_rtx_EXPR_LIST (VOIDmode, gen_rtx_REG (cmode, 33),
5910 GEN_INT (GET_MODE_SIZE (cmode)))));
5913 case MODE_RANDOM:
5914 /* We should only reach here for BLKmode on VMS. */
5915 gcc_assert (TARGET_ABI_OPEN_VMS && mode == BLKmode);
5916 regnum = 0;
5917 break;
5919 default:
5920 gcc_unreachable ();
5923 return gen_rtx_REG (mode, regnum);
5926 /* TCmode complex values are passed by invisible reference. We
5927 should not split these values. */
5929 static bool
5930 alpha_split_complex_arg (const_tree type)
5932 return TYPE_MODE (type) != TCmode;
5935 static tree
5936 alpha_build_builtin_va_list (void)
5938 tree base, ofs, space, record, type_decl;
5940 if (TARGET_ABI_OPEN_VMS)
5941 return ptr_type_node;
5943 record = (*lang_hooks.types.make_type) (RECORD_TYPE);
5944 type_decl = build_decl (BUILTINS_LOCATION,
5945 TYPE_DECL, get_identifier ("__va_list_tag"), record);
5946 TYPE_STUB_DECL (record) = type_decl;
5947 TYPE_NAME (record) = type_decl;
5949 /* C++? SET_IS_AGGR_TYPE (record, 1); */
5951 /* Dummy field to prevent alignment warnings. */
5952 space = build_decl (BUILTINS_LOCATION,
5953 FIELD_DECL, NULL_TREE, integer_type_node);
5954 DECL_FIELD_CONTEXT (space) = record;
5955 DECL_ARTIFICIAL (space) = 1;
5956 DECL_IGNORED_P (space) = 1;
5958 ofs = build_decl (BUILTINS_LOCATION,
5959 FIELD_DECL, get_identifier ("__offset"),
5960 integer_type_node);
5961 DECL_FIELD_CONTEXT (ofs) = record;
5962 DECL_CHAIN (ofs) = space;
5963 /* ??? This is a hack, __offset is marked volatile to prevent
5964 DCE that confuses stdarg optimization and results in
5965 gcc.c-torture/execute/stdarg-1.c failure. See PR 41089. */
5966 TREE_THIS_VOLATILE (ofs) = 1;
5968 base = build_decl (BUILTINS_LOCATION,
5969 FIELD_DECL, get_identifier ("__base"),
5970 ptr_type_node);
5971 DECL_FIELD_CONTEXT (base) = record;
5972 DECL_CHAIN (base) = ofs;
5974 TYPE_FIELDS (record) = base;
5975 layout_type (record);
5977 va_list_gpr_counter_field = ofs;
5978 return record;
5981 #if TARGET_ABI_OSF
5982 /* Helper function for alpha_stdarg_optimize_hook. Skip over casts
5983 and constant additions. */
5985 static gimple
5986 va_list_skip_additions (tree lhs)
5988 gimple stmt;
5990 for (;;)
5992 enum tree_code code;
5994 stmt = SSA_NAME_DEF_STMT (lhs);
5996 if (gimple_code (stmt) == GIMPLE_PHI)
5997 return stmt;
5999 if (!is_gimple_assign (stmt)
6000 || gimple_assign_lhs (stmt) != lhs)
6001 return NULL;
6003 if (TREE_CODE (gimple_assign_rhs1 (stmt)) != SSA_NAME)
6004 return stmt;
6005 code = gimple_assign_rhs_code (stmt);
6006 if (!CONVERT_EXPR_CODE_P (code)
6007 && ((code != PLUS_EXPR && code != POINTER_PLUS_EXPR)
6008 || TREE_CODE (gimple_assign_rhs2 (stmt)) != INTEGER_CST
6009 || !tree_fits_uhwi_p (gimple_assign_rhs2 (stmt))))
6010 return stmt;
6012 lhs = gimple_assign_rhs1 (stmt);
6016 /* Check if LHS = RHS statement is
6017 LHS = *(ap.__base + ap.__offset + cst)
6019 LHS = *(ap.__base
6020 + ((ap.__offset + cst <= 47)
6021 ? ap.__offset + cst - 48 : ap.__offset + cst) + cst2).
6022 If the former, indicate that GPR registers are needed,
6023 if the latter, indicate that FPR registers are needed.
6025 Also look for LHS = (*ptr).field, where ptr is one of the forms
6026 listed above.
6028 On alpha, cfun->va_list_gpr_size is used as size of the needed
6029 regs and cfun->va_list_fpr_size is a bitmask, bit 0 set if GPR
6030 registers are needed and bit 1 set if FPR registers are needed.
6031 Return true if va_list references should not be scanned for the
6032 current statement. */
6034 static bool
6035 alpha_stdarg_optimize_hook (struct stdarg_info *si, const_gimple stmt)
6037 tree base, offset, rhs;
6038 int offset_arg = 1;
6039 gimple base_stmt;
6041 if (get_gimple_rhs_class (gimple_assign_rhs_code (stmt))
6042 != GIMPLE_SINGLE_RHS)
6043 return false;
6045 rhs = gimple_assign_rhs1 (stmt);
6046 while (handled_component_p (rhs))
6047 rhs = TREE_OPERAND (rhs, 0);
6048 if (TREE_CODE (rhs) != MEM_REF
6049 || TREE_CODE (TREE_OPERAND (rhs, 0)) != SSA_NAME)
6050 return false;
6052 stmt = va_list_skip_additions (TREE_OPERAND (rhs, 0));
6053 if (stmt == NULL
6054 || !is_gimple_assign (stmt)
6055 || gimple_assign_rhs_code (stmt) != POINTER_PLUS_EXPR)
6056 return false;
6058 base = gimple_assign_rhs1 (stmt);
6059 if (TREE_CODE (base) == SSA_NAME)
6061 base_stmt = va_list_skip_additions (base);
6062 if (base_stmt
6063 && is_gimple_assign (base_stmt)
6064 && gimple_assign_rhs_code (base_stmt) == COMPONENT_REF)
6065 base = gimple_assign_rhs1 (base_stmt);
6068 if (TREE_CODE (base) != COMPONENT_REF
6069 || TREE_OPERAND (base, 1) != TYPE_FIELDS (va_list_type_node))
6071 base = gimple_assign_rhs2 (stmt);
6072 if (TREE_CODE (base) == SSA_NAME)
6074 base_stmt = va_list_skip_additions (base);
6075 if (base_stmt
6076 && is_gimple_assign (base_stmt)
6077 && gimple_assign_rhs_code (base_stmt) == COMPONENT_REF)
6078 base = gimple_assign_rhs1 (base_stmt);
6081 if (TREE_CODE (base) != COMPONENT_REF
6082 || TREE_OPERAND (base, 1) != TYPE_FIELDS (va_list_type_node))
6083 return false;
6085 offset_arg = 0;
6088 base = get_base_address (base);
6089 if (TREE_CODE (base) != VAR_DECL
6090 || !bitmap_bit_p (si->va_list_vars, DECL_UID (base) + num_ssa_names))
6091 return false;
6093 offset = gimple_op (stmt, 1 + offset_arg);
6094 if (TREE_CODE (offset) == SSA_NAME)
6096 gimple offset_stmt = va_list_skip_additions (offset);
6098 if (offset_stmt
6099 && gimple_code (offset_stmt) == GIMPLE_PHI)
6101 HOST_WIDE_INT sub;
6102 gimple arg1_stmt, arg2_stmt;
6103 tree arg1, arg2;
6104 enum tree_code code1, code2;
6106 if (gimple_phi_num_args (offset_stmt) != 2)
6107 goto escapes;
6109 arg1_stmt
6110 = va_list_skip_additions (gimple_phi_arg_def (offset_stmt, 0));
6111 arg2_stmt
6112 = va_list_skip_additions (gimple_phi_arg_def (offset_stmt, 1));
6113 if (arg1_stmt == NULL
6114 || !is_gimple_assign (arg1_stmt)
6115 || arg2_stmt == NULL
6116 || !is_gimple_assign (arg2_stmt))
6117 goto escapes;
6119 code1 = gimple_assign_rhs_code (arg1_stmt);
6120 code2 = gimple_assign_rhs_code (arg2_stmt);
6121 if (code1 == COMPONENT_REF
6122 && (code2 == MINUS_EXPR || code2 == PLUS_EXPR))
6123 /* Do nothing. */;
6124 else if (code2 == COMPONENT_REF
6125 && (code1 == MINUS_EXPR || code1 == PLUS_EXPR))
6127 gimple tem = arg1_stmt;
6128 code2 = code1;
6129 arg1_stmt = arg2_stmt;
6130 arg2_stmt = tem;
6132 else
6133 goto escapes;
6135 if (!tree_fits_shwi_p (gimple_assign_rhs2 (arg2_stmt)))
6136 goto escapes;
6138 sub = tree_to_shwi (gimple_assign_rhs2 (arg2_stmt));
6139 if (code2 == MINUS_EXPR)
6140 sub = -sub;
6141 if (sub < -48 || sub > -32)
6142 goto escapes;
6144 arg1 = gimple_assign_rhs1 (arg1_stmt);
6145 arg2 = gimple_assign_rhs1 (arg2_stmt);
6146 if (TREE_CODE (arg2) == SSA_NAME)
6148 arg2_stmt = va_list_skip_additions (arg2);
6149 if (arg2_stmt == NULL
6150 || !is_gimple_assign (arg2_stmt)
6151 || gimple_assign_rhs_code (arg2_stmt) != COMPONENT_REF)
6152 goto escapes;
6153 arg2 = gimple_assign_rhs1 (arg2_stmt);
6155 if (arg1 != arg2)
6156 goto escapes;
6158 if (TREE_CODE (arg1) != COMPONENT_REF
6159 || TREE_OPERAND (arg1, 1) != va_list_gpr_counter_field
6160 || get_base_address (arg1) != base)
6161 goto escapes;
6163 /* Need floating point regs. */
6164 cfun->va_list_fpr_size |= 2;
6165 return false;
6167 if (offset_stmt
6168 && is_gimple_assign (offset_stmt)
6169 && gimple_assign_rhs_code (offset_stmt) == COMPONENT_REF)
6170 offset = gimple_assign_rhs1 (offset_stmt);
6172 if (TREE_CODE (offset) != COMPONENT_REF
6173 || TREE_OPERAND (offset, 1) != va_list_gpr_counter_field
6174 || get_base_address (offset) != base)
6175 goto escapes;
6176 else
6177 /* Need general regs. */
6178 cfun->va_list_fpr_size |= 1;
6179 return false;
6181 escapes:
6182 si->va_list_escapes = true;
6183 return false;
6185 #endif
6187 /* Perform any needed actions needed for a function that is receiving a
6188 variable number of arguments. */
6190 static void
6191 alpha_setup_incoming_varargs (cumulative_args_t pcum, enum machine_mode mode,
6192 tree type, int *pretend_size, int no_rtl)
6194 CUMULATIVE_ARGS cum = *get_cumulative_args (pcum);
6196 /* Skip the current argument. */
6197 targetm.calls.function_arg_advance (pack_cumulative_args (&cum), mode, type,
6198 true);
6200 #if TARGET_ABI_OPEN_VMS
6201 /* For VMS, we allocate space for all 6 arg registers plus a count.
6203 However, if NO registers need to be saved, don't allocate any space.
6204 This is not only because we won't need the space, but because AP
6205 includes the current_pretend_args_size and we don't want to mess up
6206 any ap-relative addresses already made. */
6207 if (cum.num_args < 6)
6209 if (!no_rtl)
6211 emit_move_insn (gen_rtx_REG (DImode, 1), virtual_incoming_args_rtx);
6212 emit_insn (gen_arg_home ());
6214 *pretend_size = 7 * UNITS_PER_WORD;
6216 #else
6217 /* On OSF/1 and friends, we allocate space for all 12 arg registers, but
6218 only push those that are remaining. However, if NO registers need to
6219 be saved, don't allocate any space. This is not only because we won't
6220 need the space, but because AP includes the current_pretend_args_size
6221 and we don't want to mess up any ap-relative addresses already made.
6223 If we are not to use the floating-point registers, save the integer
6224 registers where we would put the floating-point registers. This is
6225 not the most efficient way to implement varargs with just one register
6226 class, but it isn't worth doing anything more efficient in this rare
6227 case. */
6228 if (cum >= 6)
6229 return;
6231 if (!no_rtl)
6233 int count;
6234 alias_set_type set = get_varargs_alias_set ();
6235 rtx tmp;
6237 count = cfun->va_list_gpr_size / UNITS_PER_WORD;
6238 if (count > 6 - cum)
6239 count = 6 - cum;
6241 /* Detect whether integer registers or floating-point registers
6242 are needed by the detected va_arg statements. See above for
6243 how these values are computed. Note that the "escape" value
6244 is VA_LIST_MAX_FPR_SIZE, which is 255, which has both of
6245 these bits set. */
6246 gcc_assert ((VA_LIST_MAX_FPR_SIZE & 3) == 3);
6248 if (cfun->va_list_fpr_size & 1)
6250 tmp = gen_rtx_MEM (BLKmode,
6251 plus_constant (Pmode, virtual_incoming_args_rtx,
6252 (cum + 6) * UNITS_PER_WORD));
6253 MEM_NOTRAP_P (tmp) = 1;
6254 set_mem_alias_set (tmp, set);
6255 move_block_from_reg (16 + cum, tmp, count);
6258 if (cfun->va_list_fpr_size & 2)
6260 tmp = gen_rtx_MEM (BLKmode,
6261 plus_constant (Pmode, virtual_incoming_args_rtx,
6262 cum * UNITS_PER_WORD));
6263 MEM_NOTRAP_P (tmp) = 1;
6264 set_mem_alias_set (tmp, set);
6265 move_block_from_reg (16 + cum + TARGET_FPREGS*32, tmp, count);
6268 *pretend_size = 12 * UNITS_PER_WORD;
6269 #endif
6272 static void
6273 alpha_va_start (tree valist, rtx nextarg ATTRIBUTE_UNUSED)
6275 HOST_WIDE_INT offset;
6276 tree t, offset_field, base_field;
6278 if (TREE_CODE (TREE_TYPE (valist)) == ERROR_MARK)
6279 return;
6281 /* For Unix, TARGET_SETUP_INCOMING_VARARGS moves the starting address base
6282 up by 48, storing fp arg registers in the first 48 bytes, and the
6283 integer arg registers in the next 48 bytes. This is only done,
6284 however, if any integer registers need to be stored.
6286 If no integer registers need be stored, then we must subtract 48
6287 in order to account for the integer arg registers which are counted
6288 in argsize above, but which are not actually stored on the stack.
6289 Must further be careful here about structures straddling the last
6290 integer argument register; that futzes with pretend_args_size,
6291 which changes the meaning of AP. */
6293 if (NUM_ARGS < 6)
6294 offset = TARGET_ABI_OPEN_VMS ? UNITS_PER_WORD : 6 * UNITS_PER_WORD;
6295 else
6296 offset = -6 * UNITS_PER_WORD + crtl->args.pretend_args_size;
6298 if (TARGET_ABI_OPEN_VMS)
6300 t = make_tree (ptr_type_node, virtual_incoming_args_rtx);
6301 t = fold_build_pointer_plus_hwi (t, offset + NUM_ARGS * UNITS_PER_WORD);
6302 t = build2 (MODIFY_EXPR, TREE_TYPE (valist), valist, t);
6303 TREE_SIDE_EFFECTS (t) = 1;
6304 expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
6306 else
6308 base_field = TYPE_FIELDS (TREE_TYPE (valist));
6309 offset_field = DECL_CHAIN (base_field);
6311 base_field = build3 (COMPONENT_REF, TREE_TYPE (base_field),
6312 valist, base_field, NULL_TREE);
6313 offset_field = build3 (COMPONENT_REF, TREE_TYPE (offset_field),
6314 valist, offset_field, NULL_TREE);
6316 t = make_tree (ptr_type_node, virtual_incoming_args_rtx);
6317 t = fold_build_pointer_plus_hwi (t, offset);
6318 t = build2 (MODIFY_EXPR, TREE_TYPE (base_field), base_field, t);
6319 TREE_SIDE_EFFECTS (t) = 1;
6320 expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
6322 t = build_int_cst (NULL_TREE, NUM_ARGS * UNITS_PER_WORD);
6323 t = build2 (MODIFY_EXPR, TREE_TYPE (offset_field), offset_field, t);
6324 TREE_SIDE_EFFECTS (t) = 1;
6325 expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL);
6329 static tree
6330 alpha_gimplify_va_arg_1 (tree type, tree base, tree offset,
6331 gimple_seq *pre_p)
6333 tree type_size, ptr_type, addend, t, addr;
6334 gimple_seq internal_post;
6336 /* If the type could not be passed in registers, skip the block
6337 reserved for the registers. */
6338 if (targetm.calls.must_pass_in_stack (TYPE_MODE (type), type))
6340 t = build_int_cst (TREE_TYPE (offset), 6*8);
6341 gimplify_assign (offset,
6342 build2 (MAX_EXPR, TREE_TYPE (offset), offset, t),
6343 pre_p);
6346 addend = offset;
6347 ptr_type = build_pointer_type_for_mode (type, ptr_mode, true);
6349 if (TREE_CODE (type) == COMPLEX_TYPE)
6351 tree real_part, imag_part, real_temp;
6353 real_part = alpha_gimplify_va_arg_1 (TREE_TYPE (type), base,
6354 offset, pre_p);
6356 /* Copy the value into a new temporary, lest the formal temporary
6357 be reused out from under us. */
6358 real_temp = get_initialized_tmp_var (real_part, pre_p, NULL);
6360 imag_part = alpha_gimplify_va_arg_1 (TREE_TYPE (type), base,
6361 offset, pre_p);
6363 return build2 (COMPLEX_EXPR, type, real_temp, imag_part);
6365 else if (TREE_CODE (type) == REAL_TYPE)
6367 tree fpaddend, cond, fourtyeight;
6369 fourtyeight = build_int_cst (TREE_TYPE (addend), 6*8);
6370 fpaddend = fold_build2 (MINUS_EXPR, TREE_TYPE (addend),
6371 addend, fourtyeight);
6372 cond = fold_build2 (LT_EXPR, boolean_type_node, addend, fourtyeight);
6373 addend = fold_build3 (COND_EXPR, TREE_TYPE (addend), cond,
6374 fpaddend, addend);
6377 /* Build the final address and force that value into a temporary. */
6378 addr = fold_build_pointer_plus (fold_convert (ptr_type, base), addend);
6379 internal_post = NULL;
6380 gimplify_expr (&addr, pre_p, &internal_post, is_gimple_val, fb_rvalue);
6381 gimple_seq_add_seq (pre_p, internal_post);
6383 /* Update the offset field. */
6384 type_size = TYPE_SIZE_UNIT (TYPE_MAIN_VARIANT (type));
6385 if (type_size == NULL || TREE_OVERFLOW (type_size))
6386 t = size_zero_node;
6387 else
6389 t = size_binop (PLUS_EXPR, type_size, size_int (7));
6390 t = size_binop (TRUNC_DIV_EXPR, t, size_int (8));
6391 t = size_binop (MULT_EXPR, t, size_int (8));
6393 t = fold_convert (TREE_TYPE (offset), t);
6394 gimplify_assign (offset, build2 (PLUS_EXPR, TREE_TYPE (offset), offset, t),
6395 pre_p);
6397 return build_va_arg_indirect_ref (addr);
6400 static tree
6401 alpha_gimplify_va_arg (tree valist, tree type, gimple_seq *pre_p,
6402 gimple_seq *post_p)
6404 tree offset_field, base_field, offset, base, t, r;
6405 bool indirect;
6407 if (TARGET_ABI_OPEN_VMS)
6408 return std_gimplify_va_arg_expr (valist, type, pre_p, post_p);
6410 base_field = TYPE_FIELDS (va_list_type_node);
6411 offset_field = DECL_CHAIN (base_field);
6412 base_field = build3 (COMPONENT_REF, TREE_TYPE (base_field),
6413 valist, base_field, NULL_TREE);
6414 offset_field = build3 (COMPONENT_REF, TREE_TYPE (offset_field),
6415 valist, offset_field, NULL_TREE);
6417 /* Pull the fields of the structure out into temporaries. Since we never
6418 modify the base field, we can use a formal temporary. Sign-extend the
6419 offset field so that it's the proper width for pointer arithmetic. */
6420 base = get_formal_tmp_var (base_field, pre_p);
6422 t = fold_convert (build_nonstandard_integer_type (64, 0), offset_field);
6423 offset = get_initialized_tmp_var (t, pre_p, NULL);
6425 indirect = pass_by_reference (NULL, TYPE_MODE (type), type, false);
6426 if (indirect)
6427 type = build_pointer_type_for_mode (type, ptr_mode, true);
6429 /* Find the value. Note that this will be a stable indirection, or
6430 a composite of stable indirections in the case of complex. */
6431 r = alpha_gimplify_va_arg_1 (type, base, offset, pre_p);
6433 /* Stuff the offset temporary back into its field. */
6434 gimplify_assign (unshare_expr (offset_field),
6435 fold_convert (TREE_TYPE (offset_field), offset), pre_p);
6437 if (indirect)
6438 r = build_va_arg_indirect_ref (r);
6440 return r;
6443 /* Builtins. */
6445 enum alpha_builtin
6447 ALPHA_BUILTIN_CMPBGE,
6448 ALPHA_BUILTIN_EXTBL,
6449 ALPHA_BUILTIN_EXTWL,
6450 ALPHA_BUILTIN_EXTLL,
6451 ALPHA_BUILTIN_EXTQL,
6452 ALPHA_BUILTIN_EXTWH,
6453 ALPHA_BUILTIN_EXTLH,
6454 ALPHA_BUILTIN_EXTQH,
6455 ALPHA_BUILTIN_INSBL,
6456 ALPHA_BUILTIN_INSWL,
6457 ALPHA_BUILTIN_INSLL,
6458 ALPHA_BUILTIN_INSQL,
6459 ALPHA_BUILTIN_INSWH,
6460 ALPHA_BUILTIN_INSLH,
6461 ALPHA_BUILTIN_INSQH,
6462 ALPHA_BUILTIN_MSKBL,
6463 ALPHA_BUILTIN_MSKWL,
6464 ALPHA_BUILTIN_MSKLL,
6465 ALPHA_BUILTIN_MSKQL,
6466 ALPHA_BUILTIN_MSKWH,
6467 ALPHA_BUILTIN_MSKLH,
6468 ALPHA_BUILTIN_MSKQH,
6469 ALPHA_BUILTIN_UMULH,
6470 ALPHA_BUILTIN_ZAP,
6471 ALPHA_BUILTIN_ZAPNOT,
6472 ALPHA_BUILTIN_AMASK,
6473 ALPHA_BUILTIN_IMPLVER,
6474 ALPHA_BUILTIN_RPCC,
6475 ALPHA_BUILTIN_ESTABLISH_VMS_CONDITION_HANDLER,
6476 ALPHA_BUILTIN_REVERT_VMS_CONDITION_HANDLER,
6478 /* TARGET_MAX */
6479 ALPHA_BUILTIN_MINUB8,
6480 ALPHA_BUILTIN_MINSB8,
6481 ALPHA_BUILTIN_MINUW4,
6482 ALPHA_BUILTIN_MINSW4,
6483 ALPHA_BUILTIN_MAXUB8,
6484 ALPHA_BUILTIN_MAXSB8,
6485 ALPHA_BUILTIN_MAXUW4,
6486 ALPHA_BUILTIN_MAXSW4,
6487 ALPHA_BUILTIN_PERR,
6488 ALPHA_BUILTIN_PKLB,
6489 ALPHA_BUILTIN_PKWB,
6490 ALPHA_BUILTIN_UNPKBL,
6491 ALPHA_BUILTIN_UNPKBW,
6493 /* TARGET_CIX */
6494 ALPHA_BUILTIN_CTTZ,
6495 ALPHA_BUILTIN_CTLZ,
6496 ALPHA_BUILTIN_CTPOP,
6498 ALPHA_BUILTIN_max
6501 static enum insn_code const code_for_builtin[ALPHA_BUILTIN_max] = {
6502 CODE_FOR_builtin_cmpbge,
6503 CODE_FOR_extbl,
6504 CODE_FOR_extwl,
6505 CODE_FOR_extll,
6506 CODE_FOR_extql,
6507 CODE_FOR_extwh,
6508 CODE_FOR_extlh,
6509 CODE_FOR_extqh,
6510 CODE_FOR_builtin_insbl,
6511 CODE_FOR_builtin_inswl,
6512 CODE_FOR_builtin_insll,
6513 CODE_FOR_insql,
6514 CODE_FOR_inswh,
6515 CODE_FOR_inslh,
6516 CODE_FOR_insqh,
6517 CODE_FOR_mskbl,
6518 CODE_FOR_mskwl,
6519 CODE_FOR_mskll,
6520 CODE_FOR_mskql,
6521 CODE_FOR_mskwh,
6522 CODE_FOR_msklh,
6523 CODE_FOR_mskqh,
6524 CODE_FOR_umuldi3_highpart,
6525 CODE_FOR_builtin_zap,
6526 CODE_FOR_builtin_zapnot,
6527 CODE_FOR_builtin_amask,
6528 CODE_FOR_builtin_implver,
6529 CODE_FOR_builtin_rpcc,
6530 CODE_FOR_builtin_establish_vms_condition_handler,
6531 CODE_FOR_builtin_revert_vms_condition_handler,
6533 /* TARGET_MAX */
6534 CODE_FOR_builtin_minub8,
6535 CODE_FOR_builtin_minsb8,
6536 CODE_FOR_builtin_minuw4,
6537 CODE_FOR_builtin_minsw4,
6538 CODE_FOR_builtin_maxub8,
6539 CODE_FOR_builtin_maxsb8,
6540 CODE_FOR_builtin_maxuw4,
6541 CODE_FOR_builtin_maxsw4,
6542 CODE_FOR_builtin_perr,
6543 CODE_FOR_builtin_pklb,
6544 CODE_FOR_builtin_pkwb,
6545 CODE_FOR_builtin_unpkbl,
6546 CODE_FOR_builtin_unpkbw,
6548 /* TARGET_CIX */
6549 CODE_FOR_ctzdi2,
6550 CODE_FOR_clzdi2,
6551 CODE_FOR_popcountdi2
6554 struct alpha_builtin_def
6556 const char *name;
6557 enum alpha_builtin code;
6558 unsigned int target_mask;
6559 bool is_const;
6562 static struct alpha_builtin_def const zero_arg_builtins[] = {
6563 { "__builtin_alpha_implver", ALPHA_BUILTIN_IMPLVER, 0, true },
6564 { "__builtin_alpha_rpcc", ALPHA_BUILTIN_RPCC, 0, false }
6567 static struct alpha_builtin_def const one_arg_builtins[] = {
6568 { "__builtin_alpha_amask", ALPHA_BUILTIN_AMASK, 0, true },
6569 { "__builtin_alpha_pklb", ALPHA_BUILTIN_PKLB, MASK_MAX, true },
6570 { "__builtin_alpha_pkwb", ALPHA_BUILTIN_PKWB, MASK_MAX, true },
6571 { "__builtin_alpha_unpkbl", ALPHA_BUILTIN_UNPKBL, MASK_MAX, true },
6572 { "__builtin_alpha_unpkbw", ALPHA_BUILTIN_UNPKBW, MASK_MAX, true },
6573 { "__builtin_alpha_cttz", ALPHA_BUILTIN_CTTZ, MASK_CIX, true },
6574 { "__builtin_alpha_ctlz", ALPHA_BUILTIN_CTLZ, MASK_CIX, true },
6575 { "__builtin_alpha_ctpop", ALPHA_BUILTIN_CTPOP, MASK_CIX, true }
6578 static struct alpha_builtin_def const two_arg_builtins[] = {
6579 { "__builtin_alpha_cmpbge", ALPHA_BUILTIN_CMPBGE, 0, true },
6580 { "__builtin_alpha_extbl", ALPHA_BUILTIN_EXTBL, 0, true },
6581 { "__builtin_alpha_extwl", ALPHA_BUILTIN_EXTWL, 0, true },
6582 { "__builtin_alpha_extll", ALPHA_BUILTIN_EXTLL, 0, true },
6583 { "__builtin_alpha_extql", ALPHA_BUILTIN_EXTQL, 0, true },
6584 { "__builtin_alpha_extwh", ALPHA_BUILTIN_EXTWH, 0, true },
6585 { "__builtin_alpha_extlh", ALPHA_BUILTIN_EXTLH, 0, true },
6586 { "__builtin_alpha_extqh", ALPHA_BUILTIN_EXTQH, 0, true },
6587 { "__builtin_alpha_insbl", ALPHA_BUILTIN_INSBL, 0, true },
6588 { "__builtin_alpha_inswl", ALPHA_BUILTIN_INSWL, 0, true },
6589 { "__builtin_alpha_insll", ALPHA_BUILTIN_INSLL, 0, true },
6590 { "__builtin_alpha_insql", ALPHA_BUILTIN_INSQL, 0, true },
6591 { "__builtin_alpha_inswh", ALPHA_BUILTIN_INSWH, 0, true },
6592 { "__builtin_alpha_inslh", ALPHA_BUILTIN_INSLH, 0, true },
6593 { "__builtin_alpha_insqh", ALPHA_BUILTIN_INSQH, 0, true },
6594 { "__builtin_alpha_mskbl", ALPHA_BUILTIN_MSKBL, 0, true },
6595 { "__builtin_alpha_mskwl", ALPHA_BUILTIN_MSKWL, 0, true },
6596 { "__builtin_alpha_mskll", ALPHA_BUILTIN_MSKLL, 0, true },
6597 { "__builtin_alpha_mskql", ALPHA_BUILTIN_MSKQL, 0, true },
6598 { "__builtin_alpha_mskwh", ALPHA_BUILTIN_MSKWH, 0, true },
6599 { "__builtin_alpha_msklh", ALPHA_BUILTIN_MSKLH, 0, true },
6600 { "__builtin_alpha_mskqh", ALPHA_BUILTIN_MSKQH, 0, true },
6601 { "__builtin_alpha_umulh", ALPHA_BUILTIN_UMULH, 0, true },
6602 { "__builtin_alpha_zap", ALPHA_BUILTIN_ZAP, 0, true },
6603 { "__builtin_alpha_zapnot", ALPHA_BUILTIN_ZAPNOT, 0, true },
6604 { "__builtin_alpha_minub8", ALPHA_BUILTIN_MINUB8, MASK_MAX, true },
6605 { "__builtin_alpha_minsb8", ALPHA_BUILTIN_MINSB8, MASK_MAX, true },
6606 { "__builtin_alpha_minuw4", ALPHA_BUILTIN_MINUW4, MASK_MAX, true },
6607 { "__builtin_alpha_minsw4", ALPHA_BUILTIN_MINSW4, MASK_MAX, true },
6608 { "__builtin_alpha_maxub8", ALPHA_BUILTIN_MAXUB8, MASK_MAX, true },
6609 { "__builtin_alpha_maxsb8", ALPHA_BUILTIN_MAXSB8, MASK_MAX, true },
6610 { "__builtin_alpha_maxuw4", ALPHA_BUILTIN_MAXUW4, MASK_MAX, true },
6611 { "__builtin_alpha_maxsw4", ALPHA_BUILTIN_MAXSW4, MASK_MAX, true },
6612 { "__builtin_alpha_perr", ALPHA_BUILTIN_PERR, MASK_MAX, true }
6615 static GTY(()) tree alpha_dimode_u;
6616 static GTY(()) tree alpha_v8qi_u;
6617 static GTY(()) tree alpha_v8qi_s;
6618 static GTY(()) tree alpha_v4hi_u;
6619 static GTY(()) tree alpha_v4hi_s;
6621 static GTY(()) tree alpha_builtins[(int) ALPHA_BUILTIN_max];
6623 /* Return the alpha builtin for CODE. */
6625 static tree
6626 alpha_builtin_decl (unsigned code, bool initialize_p ATTRIBUTE_UNUSED)
6628 if (code >= ALPHA_BUILTIN_max)
6629 return error_mark_node;
6630 return alpha_builtins[code];
6633 /* Helper function of alpha_init_builtins. Add the built-in specified
6634 by NAME, TYPE, CODE, and ECF. */
6636 static void
6637 alpha_builtin_function (const char *name, tree ftype,
6638 enum alpha_builtin code, unsigned ecf)
6640 tree decl = add_builtin_function (name, ftype, (int) code,
6641 BUILT_IN_MD, NULL, NULL_TREE);
6643 if (ecf & ECF_CONST)
6644 TREE_READONLY (decl) = 1;
6645 if (ecf & ECF_NOTHROW)
6646 TREE_NOTHROW (decl) = 1;
6648 alpha_builtins [(int) code] = decl;
6651 /* Helper function of alpha_init_builtins. Add the COUNT built-in
6652 functions pointed to by P, with function type FTYPE. */
6654 static void
6655 alpha_add_builtins (const struct alpha_builtin_def *p, size_t count,
6656 tree ftype)
6658 size_t i;
6660 for (i = 0; i < count; ++i, ++p)
6661 if ((target_flags & p->target_mask) == p->target_mask)
6662 alpha_builtin_function (p->name, ftype, p->code,
6663 (p->is_const ? ECF_CONST : 0) | ECF_NOTHROW);
6666 static void
6667 alpha_init_builtins (void)
6669 tree ftype;
6671 alpha_dimode_u = lang_hooks.types.type_for_mode (DImode, 1);
6672 alpha_v8qi_u = build_vector_type (unsigned_intQI_type_node, 8);
6673 alpha_v8qi_s = build_vector_type (intQI_type_node, 8);
6674 alpha_v4hi_u = build_vector_type (unsigned_intHI_type_node, 4);
6675 alpha_v4hi_s = build_vector_type (intHI_type_node, 4);
6677 ftype = build_function_type_list (alpha_dimode_u, NULL_TREE);
6678 alpha_add_builtins (zero_arg_builtins, ARRAY_SIZE (zero_arg_builtins), ftype);
6680 ftype = build_function_type_list (alpha_dimode_u, alpha_dimode_u, NULL_TREE);
6681 alpha_add_builtins (one_arg_builtins, ARRAY_SIZE (one_arg_builtins), ftype);
6683 ftype = build_function_type_list (alpha_dimode_u, alpha_dimode_u,
6684 alpha_dimode_u, NULL_TREE);
6685 alpha_add_builtins (two_arg_builtins, ARRAY_SIZE (two_arg_builtins), ftype);
6687 if (TARGET_ABI_OPEN_VMS)
6689 ftype = build_function_type_list (ptr_type_node, ptr_type_node,
6690 NULL_TREE);
6691 alpha_builtin_function ("__builtin_establish_vms_condition_handler",
6692 ftype,
6693 ALPHA_BUILTIN_ESTABLISH_VMS_CONDITION_HANDLER,
6696 ftype = build_function_type_list (ptr_type_node, void_type_node,
6697 NULL_TREE);
6698 alpha_builtin_function ("__builtin_revert_vms_condition_handler", ftype,
6699 ALPHA_BUILTIN_REVERT_VMS_CONDITION_HANDLER, 0);
6701 vms_patch_builtins ();
6705 /* Expand an expression EXP that calls a built-in function,
6706 with result going to TARGET if that's convenient
6707 (and in mode MODE if that's convenient).
6708 SUBTARGET may be used as the target for computing one of EXP's operands.
6709 IGNORE is nonzero if the value is to be ignored. */
6711 static rtx
6712 alpha_expand_builtin (tree exp, rtx target,
6713 rtx subtarget ATTRIBUTE_UNUSED,
6714 enum machine_mode mode ATTRIBUTE_UNUSED,
6715 int ignore ATTRIBUTE_UNUSED)
6717 #define MAX_ARGS 2
6719 tree fndecl = TREE_OPERAND (CALL_EXPR_FN (exp), 0);
6720 unsigned int fcode = DECL_FUNCTION_CODE (fndecl);
6721 tree arg;
6722 call_expr_arg_iterator iter;
6723 enum insn_code icode;
6724 rtx op[MAX_ARGS], pat;
6725 int arity;
6726 bool nonvoid;
6728 if (fcode >= ALPHA_BUILTIN_max)
6729 internal_error ("bad builtin fcode");
6730 icode = code_for_builtin[fcode];
6731 if (icode == 0)
6732 internal_error ("bad builtin fcode");
6734 nonvoid = TREE_TYPE (TREE_TYPE (fndecl)) != void_type_node;
6736 arity = 0;
6737 FOR_EACH_CALL_EXPR_ARG (arg, iter, exp)
6739 const struct insn_operand_data *insn_op;
6741 if (arg == error_mark_node)
6742 return NULL_RTX;
6743 if (arity > MAX_ARGS)
6744 return NULL_RTX;
6746 insn_op = &insn_data[icode].operand[arity + nonvoid];
6748 op[arity] = expand_expr (arg, NULL_RTX, insn_op->mode, EXPAND_NORMAL);
6750 if (!(*insn_op->predicate) (op[arity], insn_op->mode))
6751 op[arity] = copy_to_mode_reg (insn_op->mode, op[arity]);
6752 arity++;
6755 if (nonvoid)
6757 enum machine_mode tmode = insn_data[icode].operand[0].mode;
6758 if (!target
6759 || GET_MODE (target) != tmode
6760 || !(*insn_data[icode].operand[0].predicate) (target, tmode))
6761 target = gen_reg_rtx (tmode);
6764 switch (arity)
6766 case 0:
6767 pat = GEN_FCN (icode) (target);
6768 break;
6769 case 1:
6770 if (nonvoid)
6771 pat = GEN_FCN (icode) (target, op[0]);
6772 else
6773 pat = GEN_FCN (icode) (op[0]);
6774 break;
6775 case 2:
6776 pat = GEN_FCN (icode) (target, op[0], op[1]);
6777 break;
6778 default:
6779 gcc_unreachable ();
6781 if (!pat)
6782 return NULL_RTX;
6783 emit_insn (pat);
6785 if (nonvoid)
6786 return target;
6787 else
6788 return const0_rtx;
6792 /* Several bits below assume HWI >= 64 bits. This should be enforced
6793 by config.gcc. */
6794 #if HOST_BITS_PER_WIDE_INT < 64
6795 # error "HOST_WIDE_INT too small"
6796 #endif
6798 /* Fold the builtin for the CMPBGE instruction. This is a vector comparison
6799 with an 8-bit output vector. OPINT contains the integer operands; bit N
6800 of OP_CONST is set if OPINT[N] is valid. */
6802 static tree
6803 alpha_fold_builtin_cmpbge (unsigned HOST_WIDE_INT opint[], long op_const)
6805 if (op_const == 3)
6807 int i, val;
6808 for (i = 0, val = 0; i < 8; ++i)
6810 unsigned HOST_WIDE_INT c0 = (opint[0] >> (i * 8)) & 0xff;
6811 unsigned HOST_WIDE_INT c1 = (opint[1] >> (i * 8)) & 0xff;
6812 if (c0 >= c1)
6813 val |= 1 << i;
6815 return build_int_cst (alpha_dimode_u, val);
6817 else if (op_const == 2 && opint[1] == 0)
6818 return build_int_cst (alpha_dimode_u, 0xff);
6819 return NULL;
6822 /* Fold the builtin for the ZAPNOT instruction. This is essentially a
6823 specialized form of an AND operation. Other byte manipulation instructions
6824 are defined in terms of this instruction, so this is also used as a
6825 subroutine for other builtins.
6827 OP contains the tree operands; OPINT contains the extracted integer values.
6828 Bit N of OP_CONST it set if OPINT[N] is valid. OP may be null if only
6829 OPINT may be considered. */
6831 static tree
6832 alpha_fold_builtin_zapnot (tree *op, unsigned HOST_WIDE_INT opint[],
6833 long op_const)
6835 if (op_const & 2)
6837 unsigned HOST_WIDE_INT mask = 0;
6838 int i;
6840 for (i = 0; i < 8; ++i)
6841 if ((opint[1] >> i) & 1)
6842 mask |= (unsigned HOST_WIDE_INT)0xff << (i * 8);
6844 if (op_const & 1)
6845 return build_int_cst (alpha_dimode_u, opint[0] & mask);
6847 if (op)
6848 return fold_build2 (BIT_AND_EXPR, alpha_dimode_u, op[0],
6849 build_int_cst (alpha_dimode_u, mask));
6851 else if ((op_const & 1) && opint[0] == 0)
6852 return build_int_cst (alpha_dimode_u, 0);
6853 return NULL;
6856 /* Fold the builtins for the EXT family of instructions. */
6858 static tree
6859 alpha_fold_builtin_extxx (tree op[], unsigned HOST_WIDE_INT opint[],
6860 long op_const, unsigned HOST_WIDE_INT bytemask,
6861 bool is_high)
6863 long zap_const = 2;
6864 tree *zap_op = NULL;
6866 if (op_const & 2)
6868 unsigned HOST_WIDE_INT loc;
6870 loc = opint[1] & 7;
6871 loc *= BITS_PER_UNIT;
6873 if (loc != 0)
6875 if (op_const & 1)
6877 unsigned HOST_WIDE_INT temp = opint[0];
6878 if (is_high)
6879 temp <<= loc;
6880 else
6881 temp >>= loc;
6882 opint[0] = temp;
6883 zap_const = 3;
6886 else
6887 zap_op = op;
6890 opint[1] = bytemask;
6891 return alpha_fold_builtin_zapnot (zap_op, opint, zap_const);
6894 /* Fold the builtins for the INS family of instructions. */
6896 static tree
6897 alpha_fold_builtin_insxx (tree op[], unsigned HOST_WIDE_INT opint[],
6898 long op_const, unsigned HOST_WIDE_INT bytemask,
6899 bool is_high)
6901 if ((op_const & 1) && opint[0] == 0)
6902 return build_int_cst (alpha_dimode_u, 0);
6904 if (op_const & 2)
6906 unsigned HOST_WIDE_INT temp, loc, byteloc;
6907 tree *zap_op = NULL;
6909 loc = opint[1] & 7;
6910 bytemask <<= loc;
6912 temp = opint[0];
6913 if (is_high)
6915 byteloc = (64 - (loc * 8)) & 0x3f;
6916 if (byteloc == 0)
6917 zap_op = op;
6918 else
6919 temp >>= byteloc;
6920 bytemask >>= 8;
6922 else
6924 byteloc = loc * 8;
6925 if (byteloc == 0)
6926 zap_op = op;
6927 else
6928 temp <<= byteloc;
6931 opint[0] = temp;
6932 opint[1] = bytemask;
6933 return alpha_fold_builtin_zapnot (zap_op, opint, op_const);
6936 return NULL;
6939 static tree
6940 alpha_fold_builtin_mskxx (tree op[], unsigned HOST_WIDE_INT opint[],
6941 long op_const, unsigned HOST_WIDE_INT bytemask,
6942 bool is_high)
6944 if (op_const & 2)
6946 unsigned HOST_WIDE_INT loc;
6948 loc = opint[1] & 7;
6949 bytemask <<= loc;
6951 if (is_high)
6952 bytemask >>= 8;
6954 opint[1] = bytemask ^ 0xff;
6957 return alpha_fold_builtin_zapnot (op, opint, op_const);
6960 static tree
6961 alpha_fold_vector_minmax (enum tree_code code, tree op[], tree vtype)
6963 tree op0 = fold_convert (vtype, op[0]);
6964 tree op1 = fold_convert (vtype, op[1]);
6965 tree val = fold_build2 (code, vtype, op0, op1);
6966 return fold_build1 (VIEW_CONVERT_EXPR, alpha_dimode_u, val);
6969 static tree
6970 alpha_fold_builtin_perr (unsigned HOST_WIDE_INT opint[], long op_const)
6972 unsigned HOST_WIDE_INT temp = 0;
6973 int i;
6975 if (op_const != 3)
6976 return NULL;
6978 for (i = 0; i < 8; ++i)
6980 unsigned HOST_WIDE_INT a = (opint[0] >> (i * 8)) & 0xff;
6981 unsigned HOST_WIDE_INT b = (opint[1] >> (i * 8)) & 0xff;
6982 if (a >= b)
6983 temp += a - b;
6984 else
6985 temp += b - a;
6988 return build_int_cst (alpha_dimode_u, temp);
6991 static tree
6992 alpha_fold_builtin_pklb (unsigned HOST_WIDE_INT opint[], long op_const)
6994 unsigned HOST_WIDE_INT temp;
6996 if (op_const == 0)
6997 return NULL;
6999 temp = opint[0] & 0xff;
7000 temp |= (opint[0] >> 24) & 0xff00;
7002 return build_int_cst (alpha_dimode_u, temp);
7005 static tree
7006 alpha_fold_builtin_pkwb (unsigned HOST_WIDE_INT opint[], long op_const)
7008 unsigned HOST_WIDE_INT temp;
7010 if (op_const == 0)
7011 return NULL;
7013 temp = opint[0] & 0xff;
7014 temp |= (opint[0] >> 8) & 0xff00;
7015 temp |= (opint[0] >> 16) & 0xff0000;
7016 temp |= (opint[0] >> 24) & 0xff000000;
7018 return build_int_cst (alpha_dimode_u, temp);
7021 static tree
7022 alpha_fold_builtin_unpkbl (unsigned HOST_WIDE_INT opint[], long op_const)
7024 unsigned HOST_WIDE_INT temp;
7026 if (op_const == 0)
7027 return NULL;
7029 temp = opint[0] & 0xff;
7030 temp |= (opint[0] & 0xff00) << 24;
7032 return build_int_cst (alpha_dimode_u, temp);
7035 static tree
7036 alpha_fold_builtin_unpkbw (unsigned HOST_WIDE_INT opint[], long op_const)
7038 unsigned HOST_WIDE_INT temp;
7040 if (op_const == 0)
7041 return NULL;
7043 temp = opint[0] & 0xff;
7044 temp |= (opint[0] & 0x0000ff00) << 8;
7045 temp |= (opint[0] & 0x00ff0000) << 16;
7046 temp |= (opint[0] & 0xff000000) << 24;
7048 return build_int_cst (alpha_dimode_u, temp);
7051 static tree
7052 alpha_fold_builtin_cttz (unsigned HOST_WIDE_INT opint[], long op_const)
7054 unsigned HOST_WIDE_INT temp;
7056 if (op_const == 0)
7057 return NULL;
7059 if (opint[0] == 0)
7060 temp = 64;
7061 else
7062 temp = exact_log2 (opint[0] & -opint[0]);
7064 return build_int_cst (alpha_dimode_u, temp);
7067 static tree
7068 alpha_fold_builtin_ctlz (unsigned HOST_WIDE_INT opint[], long op_const)
7070 unsigned HOST_WIDE_INT temp;
7072 if (op_const == 0)
7073 return NULL;
7075 if (opint[0] == 0)
7076 temp = 64;
7077 else
7078 temp = 64 - floor_log2 (opint[0]) - 1;
7080 return build_int_cst (alpha_dimode_u, temp);
7083 static tree
7084 alpha_fold_builtin_ctpop (unsigned HOST_WIDE_INT opint[], long op_const)
7086 unsigned HOST_WIDE_INT temp, op;
7088 if (op_const == 0)
7089 return NULL;
7091 op = opint[0];
7092 temp = 0;
7093 while (op)
7094 temp++, op &= op - 1;
7096 return build_int_cst (alpha_dimode_u, temp);
7099 /* Fold one of our builtin functions. */
7101 static tree
7102 alpha_fold_builtin (tree fndecl, int n_args, tree *op,
7103 bool ignore ATTRIBUTE_UNUSED)
7105 unsigned HOST_WIDE_INT opint[MAX_ARGS];
7106 long op_const = 0;
7107 int i;
7109 if (n_args > MAX_ARGS)
7110 return NULL;
7112 for (i = 0; i < n_args; i++)
7114 tree arg = op[i];
7115 if (arg == error_mark_node)
7116 return NULL;
7118 opint[i] = 0;
7119 if (TREE_CODE (arg) == INTEGER_CST)
7121 op_const |= 1L << i;
7122 opint[i] = int_cst_value (arg);
7126 switch (DECL_FUNCTION_CODE (fndecl))
7128 case ALPHA_BUILTIN_CMPBGE:
7129 return alpha_fold_builtin_cmpbge (opint, op_const);
7131 case ALPHA_BUILTIN_EXTBL:
7132 return alpha_fold_builtin_extxx (op, opint, op_const, 0x01, false);
7133 case ALPHA_BUILTIN_EXTWL:
7134 return alpha_fold_builtin_extxx (op, opint, op_const, 0x03, false);
7135 case ALPHA_BUILTIN_EXTLL:
7136 return alpha_fold_builtin_extxx (op, opint, op_const, 0x0f, false);
7137 case ALPHA_BUILTIN_EXTQL:
7138 return alpha_fold_builtin_extxx (op, opint, op_const, 0xff, false);
7139 case ALPHA_BUILTIN_EXTWH:
7140 return alpha_fold_builtin_extxx (op, opint, op_const, 0x03, true);
7141 case ALPHA_BUILTIN_EXTLH:
7142 return alpha_fold_builtin_extxx (op, opint, op_const, 0x0f, true);
7143 case ALPHA_BUILTIN_EXTQH:
7144 return alpha_fold_builtin_extxx (op, opint, op_const, 0xff, true);
7146 case ALPHA_BUILTIN_INSBL:
7147 return alpha_fold_builtin_insxx (op, opint, op_const, 0x01, false);
7148 case ALPHA_BUILTIN_INSWL:
7149 return alpha_fold_builtin_insxx (op, opint, op_const, 0x03, false);
7150 case ALPHA_BUILTIN_INSLL:
7151 return alpha_fold_builtin_insxx (op, opint, op_const, 0x0f, false);
7152 case ALPHA_BUILTIN_INSQL:
7153 return alpha_fold_builtin_insxx (op, opint, op_const, 0xff, false);
7154 case ALPHA_BUILTIN_INSWH:
7155 return alpha_fold_builtin_insxx (op, opint, op_const, 0x03, true);
7156 case ALPHA_BUILTIN_INSLH:
7157 return alpha_fold_builtin_insxx (op, opint, op_const, 0x0f, true);
7158 case ALPHA_BUILTIN_INSQH:
7159 return alpha_fold_builtin_insxx (op, opint, op_const, 0xff, true);
7161 case ALPHA_BUILTIN_MSKBL:
7162 return alpha_fold_builtin_mskxx (op, opint, op_const, 0x01, false);
7163 case ALPHA_BUILTIN_MSKWL:
7164 return alpha_fold_builtin_mskxx (op, opint, op_const, 0x03, false);
7165 case ALPHA_BUILTIN_MSKLL:
7166 return alpha_fold_builtin_mskxx (op, opint, op_const, 0x0f, false);
7167 case ALPHA_BUILTIN_MSKQL:
7168 return alpha_fold_builtin_mskxx (op, opint, op_const, 0xff, false);
7169 case ALPHA_BUILTIN_MSKWH:
7170 return alpha_fold_builtin_mskxx (op, opint, op_const, 0x03, true);
7171 case ALPHA_BUILTIN_MSKLH:
7172 return alpha_fold_builtin_mskxx (op, opint, op_const, 0x0f, true);
7173 case ALPHA_BUILTIN_MSKQH:
7174 return alpha_fold_builtin_mskxx (op, opint, op_const, 0xff, true);
7176 case ALPHA_BUILTIN_ZAP:
7177 opint[1] ^= 0xff;
7178 /* FALLTHRU */
7179 case ALPHA_BUILTIN_ZAPNOT:
7180 return alpha_fold_builtin_zapnot (op, opint, op_const);
7182 case ALPHA_BUILTIN_MINUB8:
7183 return alpha_fold_vector_minmax (MIN_EXPR, op, alpha_v8qi_u);
7184 case ALPHA_BUILTIN_MINSB8:
7185 return alpha_fold_vector_minmax (MIN_EXPR, op, alpha_v8qi_s);
7186 case ALPHA_BUILTIN_MINUW4:
7187 return alpha_fold_vector_minmax (MIN_EXPR, op, alpha_v4hi_u);
7188 case ALPHA_BUILTIN_MINSW4:
7189 return alpha_fold_vector_minmax (MIN_EXPR, op, alpha_v4hi_s);
7190 case ALPHA_BUILTIN_MAXUB8:
7191 return alpha_fold_vector_minmax (MAX_EXPR, op, alpha_v8qi_u);
7192 case ALPHA_BUILTIN_MAXSB8:
7193 return alpha_fold_vector_minmax (MAX_EXPR, op, alpha_v8qi_s);
7194 case ALPHA_BUILTIN_MAXUW4:
7195 return alpha_fold_vector_minmax (MAX_EXPR, op, alpha_v4hi_u);
7196 case ALPHA_BUILTIN_MAXSW4:
7197 return alpha_fold_vector_minmax (MAX_EXPR, op, alpha_v4hi_s);
7199 case ALPHA_BUILTIN_PERR:
7200 return alpha_fold_builtin_perr (opint, op_const);
7201 case ALPHA_BUILTIN_PKLB:
7202 return alpha_fold_builtin_pklb (opint, op_const);
7203 case ALPHA_BUILTIN_PKWB:
7204 return alpha_fold_builtin_pkwb (opint, op_const);
7205 case ALPHA_BUILTIN_UNPKBL:
7206 return alpha_fold_builtin_unpkbl (opint, op_const);
7207 case ALPHA_BUILTIN_UNPKBW:
7208 return alpha_fold_builtin_unpkbw (opint, op_const);
7210 case ALPHA_BUILTIN_CTTZ:
7211 return alpha_fold_builtin_cttz (opint, op_const);
7212 case ALPHA_BUILTIN_CTLZ:
7213 return alpha_fold_builtin_ctlz (opint, op_const);
7214 case ALPHA_BUILTIN_CTPOP:
7215 return alpha_fold_builtin_ctpop (opint, op_const);
7217 case ALPHA_BUILTIN_AMASK:
7218 case ALPHA_BUILTIN_IMPLVER:
7219 case ALPHA_BUILTIN_RPCC:
7220 /* None of these are foldable at compile-time. */
7221 default:
7222 return NULL;
7226 bool
7227 alpha_gimple_fold_builtin (gimple_stmt_iterator *gsi)
7229 bool changed = false;
7230 gimple stmt = gsi_stmt (*gsi);
7231 tree call = gimple_call_fn (stmt);
7232 gimple new_stmt = NULL;
7234 if (call)
7236 tree fndecl = gimple_call_fndecl (stmt);
7238 if (fndecl)
7240 tree arg0, arg1;
7242 switch (DECL_FUNCTION_CODE (fndecl))
7244 case ALPHA_BUILTIN_UMULH:
7245 arg0 = gimple_call_arg (stmt, 0);
7246 arg1 = gimple_call_arg (stmt, 1);
7248 new_stmt
7249 = gimple_build_assign_with_ops (MULT_HIGHPART_EXPR,
7250 gimple_call_lhs (stmt),
7251 arg0,
7252 arg1);
7253 break;
7254 default:
7255 break;
7260 if (new_stmt)
7262 gsi_replace (gsi, new_stmt, true);
7263 changed = true;
7266 return changed;
7269 /* This page contains routines that are used to determine what the function
7270 prologue and epilogue code will do and write them out. */
7272 /* Compute the size of the save area in the stack. */
7274 /* These variables are used for communication between the following functions.
7275 They indicate various things about the current function being compiled
7276 that are used to tell what kind of prologue, epilogue and procedure
7277 descriptor to generate. */
7279 /* Nonzero if we need a stack procedure. */
7280 enum alpha_procedure_types {PT_NULL = 0, PT_REGISTER = 1, PT_STACK = 2};
7281 static enum alpha_procedure_types alpha_procedure_type;
7283 /* Register number (either FP or SP) that is used to unwind the frame. */
7284 static int vms_unwind_regno;
7286 /* Register number used to save FP. We need not have one for RA since
7287 we don't modify it for register procedures. This is only defined
7288 for register frame procedures. */
7289 static int vms_save_fp_regno;
7291 /* Register number used to reference objects off our PV. */
7292 static int vms_base_regno;
7294 /* Compute register masks for saved registers. */
7296 static void
7297 alpha_sa_mask (unsigned long *imaskP, unsigned long *fmaskP)
7299 unsigned long imask = 0;
7300 unsigned long fmask = 0;
7301 unsigned int i;
7303 /* When outputting a thunk, we don't have valid register life info,
7304 but assemble_start_function wants to output .frame and .mask
7305 directives. */
7306 if (cfun->is_thunk)
7308 *imaskP = 0;
7309 *fmaskP = 0;
7310 return;
7313 if (TARGET_ABI_OPEN_VMS && alpha_procedure_type == PT_STACK)
7314 imask |= (1UL << HARD_FRAME_POINTER_REGNUM);
7316 /* One for every register we have to save. */
7317 for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
7318 if (! fixed_regs[i] && ! call_used_regs[i]
7319 && df_regs_ever_live_p (i) && i != REG_RA)
7321 if (i < 32)
7322 imask |= (1UL << i);
7323 else
7324 fmask |= (1UL << (i - 32));
7327 /* We need to restore these for the handler. */
7328 if (crtl->calls_eh_return)
7330 for (i = 0; ; ++i)
7332 unsigned regno = EH_RETURN_DATA_REGNO (i);
7333 if (regno == INVALID_REGNUM)
7334 break;
7335 imask |= 1UL << regno;
7339 /* If any register spilled, then spill the return address also. */
7340 /* ??? This is required by the Digital stack unwind specification
7341 and isn't needed if we're doing Dwarf2 unwinding. */
7342 if (imask || fmask || alpha_ra_ever_killed ())
7343 imask |= (1UL << REG_RA);
7345 *imaskP = imask;
7346 *fmaskP = fmask;
7350 alpha_sa_size (void)
7352 unsigned long mask[2];
7353 int sa_size = 0;
7354 int i, j;
7356 alpha_sa_mask (&mask[0], &mask[1]);
7358 for (j = 0; j < 2; ++j)
7359 for (i = 0; i < 32; ++i)
7360 if ((mask[j] >> i) & 1)
7361 sa_size++;
7363 if (TARGET_ABI_OPEN_VMS)
7365 /* Start with a stack procedure if we make any calls (REG_RA used), or
7366 need a frame pointer, with a register procedure if we otherwise need
7367 at least a slot, and with a null procedure in other cases. */
7368 if ((mask[0] >> REG_RA) & 1 || frame_pointer_needed)
7369 alpha_procedure_type = PT_STACK;
7370 else if (get_frame_size() != 0)
7371 alpha_procedure_type = PT_REGISTER;
7372 else
7373 alpha_procedure_type = PT_NULL;
7375 /* Don't reserve space for saving FP & RA yet. Do that later after we've
7376 made the final decision on stack procedure vs register procedure. */
7377 if (alpha_procedure_type == PT_STACK)
7378 sa_size -= 2;
7380 /* Decide whether to refer to objects off our PV via FP or PV.
7381 If we need FP for something else or if we receive a nonlocal
7382 goto (which expects PV to contain the value), we must use PV.
7383 Otherwise, start by assuming we can use FP. */
7385 vms_base_regno
7386 = (frame_pointer_needed
7387 || cfun->has_nonlocal_label
7388 || alpha_procedure_type == PT_STACK
7389 || crtl->outgoing_args_size)
7390 ? REG_PV : HARD_FRAME_POINTER_REGNUM;
7392 /* If we want to copy PV into FP, we need to find some register
7393 in which to save FP. */
7395 vms_save_fp_regno = -1;
7396 if (vms_base_regno == HARD_FRAME_POINTER_REGNUM)
7397 for (i = 0; i < 32; i++)
7398 if (! fixed_regs[i] && call_used_regs[i] && ! df_regs_ever_live_p (i))
7399 vms_save_fp_regno = i;
7401 /* A VMS condition handler requires a stack procedure in our
7402 implementation. (not required by the calling standard). */
7403 if ((vms_save_fp_regno == -1 && alpha_procedure_type == PT_REGISTER)
7404 || cfun->machine->uses_condition_handler)
7405 vms_base_regno = REG_PV, alpha_procedure_type = PT_STACK;
7406 else if (alpha_procedure_type == PT_NULL)
7407 vms_base_regno = REG_PV;
7409 /* Stack unwinding should be done via FP unless we use it for PV. */
7410 vms_unwind_regno = (vms_base_regno == REG_PV
7411 ? HARD_FRAME_POINTER_REGNUM : STACK_POINTER_REGNUM);
7413 /* If this is a stack procedure, allow space for saving FP, RA and
7414 a condition handler slot if needed. */
7415 if (alpha_procedure_type == PT_STACK)
7416 sa_size += 2 + cfun->machine->uses_condition_handler;
7418 else
7420 /* Our size must be even (multiple of 16 bytes). */
7421 if (sa_size & 1)
7422 sa_size++;
7425 return sa_size * 8;
7428 /* Define the offset between two registers, one to be eliminated,
7429 and the other its replacement, at the start of a routine. */
7431 HOST_WIDE_INT
7432 alpha_initial_elimination_offset (unsigned int from,
7433 unsigned int to ATTRIBUTE_UNUSED)
7435 HOST_WIDE_INT ret;
7437 ret = alpha_sa_size ();
7438 ret += ALPHA_ROUND (crtl->outgoing_args_size);
7440 switch (from)
7442 case FRAME_POINTER_REGNUM:
7443 break;
7445 case ARG_POINTER_REGNUM:
7446 ret += (ALPHA_ROUND (get_frame_size ()
7447 + crtl->args.pretend_args_size)
7448 - crtl->args.pretend_args_size);
7449 break;
7451 default:
7452 gcc_unreachable ();
7455 return ret;
7458 #if TARGET_ABI_OPEN_VMS
7460 /* Worker function for TARGET_CAN_ELIMINATE. */
7462 static bool
7463 alpha_vms_can_eliminate (const int from ATTRIBUTE_UNUSED, const int to)
7465 /* We need the alpha_procedure_type to decide. Evaluate it now. */
7466 alpha_sa_size ();
7468 switch (alpha_procedure_type)
7470 case PT_NULL:
7471 /* NULL procedures have no frame of their own and we only
7472 know how to resolve from the current stack pointer. */
7473 return to == STACK_POINTER_REGNUM;
7475 case PT_REGISTER:
7476 case PT_STACK:
7477 /* We always eliminate except to the stack pointer if there is no
7478 usable frame pointer at hand. */
7479 return (to != STACK_POINTER_REGNUM
7480 || vms_unwind_regno != HARD_FRAME_POINTER_REGNUM);
7483 gcc_unreachable ();
7486 /* FROM is to be eliminated for TO. Return the offset so that TO+offset
7487 designates the same location as FROM. */
7489 HOST_WIDE_INT
7490 alpha_vms_initial_elimination_offset (unsigned int from, unsigned int to)
7492 /* The only possible attempts we ever expect are ARG or FRAME_PTR to
7493 HARD_FRAME or STACK_PTR. We need the alpha_procedure_type to decide
7494 on the proper computations and will need the register save area size
7495 in most cases. */
7497 HOST_WIDE_INT sa_size = alpha_sa_size ();
7499 /* PT_NULL procedures have no frame of their own and we only allow
7500 elimination to the stack pointer. This is the argument pointer and we
7501 resolve the soft frame pointer to that as well. */
7503 if (alpha_procedure_type == PT_NULL)
7504 return 0;
7506 /* For a PT_STACK procedure the frame layout looks as follows
7508 -----> decreasing addresses
7510 < size rounded up to 16 | likewise >
7511 --------------#------------------------------+++--------------+++-------#
7512 incoming args # pretended args | "frame" | regs sa | PV | outgoing args #
7513 --------------#---------------------------------------------------------#
7514 ^ ^ ^ ^
7515 ARG_PTR FRAME_PTR HARD_FRAME_PTR STACK_PTR
7518 PT_REGISTER procedures are similar in that they may have a frame of their
7519 own. They have no regs-sa/pv/outgoing-args area.
7521 We first compute offset to HARD_FRAME_PTR, then add what we need to get
7522 to STACK_PTR if need be. */
7525 HOST_WIDE_INT offset;
7526 HOST_WIDE_INT pv_save_size = alpha_procedure_type == PT_STACK ? 8 : 0;
7528 switch (from)
7530 case FRAME_POINTER_REGNUM:
7531 offset = ALPHA_ROUND (sa_size + pv_save_size);
7532 break;
7533 case ARG_POINTER_REGNUM:
7534 offset = (ALPHA_ROUND (sa_size + pv_save_size
7535 + get_frame_size ()
7536 + crtl->args.pretend_args_size)
7537 - crtl->args.pretend_args_size);
7538 break;
7539 default:
7540 gcc_unreachable ();
7543 if (to == STACK_POINTER_REGNUM)
7544 offset += ALPHA_ROUND (crtl->outgoing_args_size);
7546 return offset;
7550 #define COMMON_OBJECT "common_object"
7552 static tree
7553 common_object_handler (tree *node, tree name ATTRIBUTE_UNUSED,
7554 tree args ATTRIBUTE_UNUSED, int flags ATTRIBUTE_UNUSED,
7555 bool *no_add_attrs ATTRIBUTE_UNUSED)
7557 tree decl = *node;
7558 gcc_assert (DECL_P (decl));
7560 DECL_COMMON (decl) = 1;
7561 return NULL_TREE;
7564 static const struct attribute_spec vms_attribute_table[] =
7566 /* { name, min_len, max_len, decl_req, type_req, fn_type_req, handler,
7567 affects_type_identity } */
7568 { COMMON_OBJECT, 0, 1, true, false, false, common_object_handler, false },
7569 { NULL, 0, 0, false, false, false, NULL, false }
7572 void
7573 vms_output_aligned_decl_common(FILE *file, tree decl, const char *name,
7574 unsigned HOST_WIDE_INT size,
7575 unsigned int align)
7577 tree attr = DECL_ATTRIBUTES (decl);
7578 fprintf (file, "%s", COMMON_ASM_OP);
7579 assemble_name (file, name);
7580 fprintf (file, "," HOST_WIDE_INT_PRINT_UNSIGNED, size);
7581 /* ??? Unlike on OSF/1, the alignment factor is not in log units. */
7582 fprintf (file, ",%u", align / BITS_PER_UNIT);
7583 if (attr)
7585 attr = lookup_attribute (COMMON_OBJECT, attr);
7586 if (attr)
7587 fprintf (file, ",%s",
7588 IDENTIFIER_POINTER (TREE_VALUE (TREE_VALUE (attr))));
7590 fputc ('\n', file);
7593 #undef COMMON_OBJECT
7595 #endif
7597 static int
7598 find_lo_sum_using_gp (rtx *px, void *data ATTRIBUTE_UNUSED)
7600 return GET_CODE (*px) == LO_SUM && XEXP (*px, 0) == pic_offset_table_rtx;
7604 alpha_find_lo_sum_using_gp (rtx insn)
7606 return for_each_rtx (&PATTERN (insn), find_lo_sum_using_gp, NULL) > 0;
7609 static int
7610 alpha_does_function_need_gp (void)
7612 rtx insn;
7614 /* The GP being variable is an OSF abi thing. */
7615 if (! TARGET_ABI_OSF)
7616 return 0;
7618 /* We need the gp to load the address of __mcount. */
7619 if (TARGET_PROFILING_NEEDS_GP && crtl->profile)
7620 return 1;
7622 /* The code emitted by alpha_output_mi_thunk_osf uses the gp. */
7623 if (cfun->is_thunk)
7624 return 1;
7626 /* The nonlocal receiver pattern assumes that the gp is valid for
7627 the nested function. Reasonable because it's almost always set
7628 correctly already. For the cases where that's wrong, make sure
7629 the nested function loads its gp on entry. */
7630 if (crtl->has_nonlocal_goto)
7631 return 1;
7633 /* If we need a GP (we have a LDSYM insn or a CALL_INSN), load it first.
7634 Even if we are a static function, we still need to do this in case
7635 our address is taken and passed to something like qsort. */
7637 push_topmost_sequence ();
7638 insn = get_insns ();
7639 pop_topmost_sequence ();
7641 for (; insn; insn = NEXT_INSN (insn))
7642 if (NONDEBUG_INSN_P (insn)
7643 && GET_CODE (PATTERN (insn)) != USE
7644 && GET_CODE (PATTERN (insn)) != CLOBBER
7645 && get_attr_usegp (insn))
7646 return 1;
7648 return 0;
7652 /* Helper function to set RTX_FRAME_RELATED_P on instructions, including
7653 sequences. */
7655 static rtx
7656 set_frame_related_p (void)
7658 rtx seq = get_insns ();
7659 rtx insn;
7661 end_sequence ();
7663 if (!seq)
7664 return NULL_RTX;
7666 if (INSN_P (seq))
7668 insn = seq;
7669 while (insn != NULL_RTX)
7671 RTX_FRAME_RELATED_P (insn) = 1;
7672 insn = NEXT_INSN (insn);
7674 seq = emit_insn (seq);
7676 else
7678 seq = emit_insn (seq);
7679 RTX_FRAME_RELATED_P (seq) = 1;
7681 return seq;
7684 #define FRP(exp) (start_sequence (), exp, set_frame_related_p ())
7686 /* Generates a store with the proper unwind info attached. VALUE is
7687 stored at BASE_REG+BASE_OFS. If FRAME_BIAS is nonzero, then BASE_REG
7688 contains SP+FRAME_BIAS, and that is the unwind info that should be
7689 generated. If FRAME_REG != VALUE, then VALUE is being stored on
7690 behalf of FRAME_REG, and FRAME_REG should be present in the unwind. */
7692 static void
7693 emit_frame_store_1 (rtx value, rtx base_reg, HOST_WIDE_INT frame_bias,
7694 HOST_WIDE_INT base_ofs, rtx frame_reg)
7696 rtx addr, mem, insn;
7698 addr = plus_constant (Pmode, base_reg, base_ofs);
7699 mem = gen_frame_mem (DImode, addr);
7701 insn = emit_move_insn (mem, value);
7702 RTX_FRAME_RELATED_P (insn) = 1;
7704 if (frame_bias || value != frame_reg)
7706 if (frame_bias)
7708 addr = plus_constant (Pmode, stack_pointer_rtx,
7709 frame_bias + base_ofs);
7710 mem = gen_rtx_MEM (DImode, addr);
7713 add_reg_note (insn, REG_FRAME_RELATED_EXPR,
7714 gen_rtx_SET (VOIDmode, mem, frame_reg));
7718 static void
7719 emit_frame_store (unsigned int regno, rtx base_reg,
7720 HOST_WIDE_INT frame_bias, HOST_WIDE_INT base_ofs)
7722 rtx reg = gen_rtx_REG (DImode, regno);
7723 emit_frame_store_1 (reg, base_reg, frame_bias, base_ofs, reg);
7726 /* Compute the frame size. SIZE is the size of the "naked" frame
7727 and SA_SIZE is the size of the register save area. */
7729 static HOST_WIDE_INT
7730 compute_frame_size (HOST_WIDE_INT size, HOST_WIDE_INT sa_size)
7732 if (TARGET_ABI_OPEN_VMS)
7733 return ALPHA_ROUND (sa_size
7734 + (alpha_procedure_type == PT_STACK ? 8 : 0)
7735 + size
7736 + crtl->args.pretend_args_size);
7737 else
7738 return ALPHA_ROUND (crtl->outgoing_args_size)
7739 + sa_size
7740 + ALPHA_ROUND (size
7741 + crtl->args.pretend_args_size);
7744 /* Write function prologue. */
7746 /* On vms we have two kinds of functions:
7748 - stack frame (PROC_STACK)
7749 these are 'normal' functions with local vars and which are
7750 calling other functions
7751 - register frame (PROC_REGISTER)
7752 keeps all data in registers, needs no stack
7754 We must pass this to the assembler so it can generate the
7755 proper pdsc (procedure descriptor)
7756 This is done with the '.pdesc' command.
7758 On not-vms, we don't really differentiate between the two, as we can
7759 simply allocate stack without saving registers. */
7761 void
7762 alpha_expand_prologue (void)
7764 /* Registers to save. */
7765 unsigned long imask = 0;
7766 unsigned long fmask = 0;
7767 /* Stack space needed for pushing registers clobbered by us. */
7768 HOST_WIDE_INT sa_size, sa_bias;
7769 /* Complete stack size needed. */
7770 HOST_WIDE_INT frame_size;
7771 /* Probed stack size; it additionally includes the size of
7772 the "reserve region" if any. */
7773 HOST_WIDE_INT probed_size;
7774 /* Offset from base reg to register save area. */
7775 HOST_WIDE_INT reg_offset;
7776 rtx sa_reg;
7777 int i;
7779 sa_size = alpha_sa_size ();
7780 frame_size = compute_frame_size (get_frame_size (), sa_size);
7782 if (flag_stack_usage_info)
7783 current_function_static_stack_size = frame_size;
7785 if (TARGET_ABI_OPEN_VMS)
7786 reg_offset = 8 + 8 * cfun->machine->uses_condition_handler;
7787 else
7788 reg_offset = ALPHA_ROUND (crtl->outgoing_args_size);
7790 alpha_sa_mask (&imask, &fmask);
7792 /* Emit an insn to reload GP, if needed. */
7793 if (TARGET_ABI_OSF)
7795 alpha_function_needs_gp = alpha_does_function_need_gp ();
7796 if (alpha_function_needs_gp)
7797 emit_insn (gen_prologue_ldgp ());
7800 /* TARGET_PROFILING_NEEDS_GP actually implies that we need to insert
7801 the call to mcount ourselves, rather than having the linker do it
7802 magically in response to -pg. Since _mcount has special linkage,
7803 don't represent the call as a call. */
7804 if (TARGET_PROFILING_NEEDS_GP && crtl->profile)
7805 emit_insn (gen_prologue_mcount ());
7807 /* Adjust the stack by the frame size. If the frame size is > 4096
7808 bytes, we need to be sure we probe somewhere in the first and last
7809 4096 bytes (we can probably get away without the latter test) and
7810 every 8192 bytes in between. If the frame size is > 32768, we
7811 do this in a loop. Otherwise, we generate the explicit probe
7812 instructions.
7814 Note that we are only allowed to adjust sp once in the prologue. */
7816 probed_size = frame_size;
7817 if (flag_stack_check)
7818 probed_size += STACK_CHECK_PROTECT;
7820 if (probed_size <= 32768)
7822 if (probed_size > 4096)
7824 int probed;
7826 for (probed = 4096; probed < probed_size; probed += 8192)
7827 emit_insn (gen_probe_stack (GEN_INT (-probed)));
7829 /* We only have to do this probe if we aren't saving registers or
7830 if we are probing beyond the frame because of -fstack-check. */
7831 if ((sa_size == 0 && probed_size > probed - 4096)
7832 || flag_stack_check)
7833 emit_insn (gen_probe_stack (GEN_INT (-probed_size)));
7836 if (frame_size != 0)
7837 FRP (emit_insn (gen_adddi3 (stack_pointer_rtx, stack_pointer_rtx,
7838 GEN_INT (-frame_size))));
7840 else
7842 /* Here we generate code to set R22 to SP + 4096 and set R23 to the
7843 number of 8192 byte blocks to probe. We then probe each block
7844 in the loop and then set SP to the proper location. If the
7845 amount remaining is > 4096, we have to do one more probe if we
7846 are not saving any registers or if we are probing beyond the
7847 frame because of -fstack-check. */
7849 HOST_WIDE_INT blocks = (probed_size + 4096) / 8192;
7850 HOST_WIDE_INT leftover = probed_size + 4096 - blocks * 8192;
7851 rtx ptr = gen_rtx_REG (DImode, 22);
7852 rtx count = gen_rtx_REG (DImode, 23);
7853 rtx seq;
7855 emit_move_insn (count, GEN_INT (blocks));
7856 emit_insn (gen_adddi3 (ptr, stack_pointer_rtx, GEN_INT (4096)));
7858 /* Because of the difficulty in emitting a new basic block this
7859 late in the compilation, generate the loop as a single insn. */
7860 emit_insn (gen_prologue_stack_probe_loop (count, ptr));
7862 if ((leftover > 4096 && sa_size == 0) || flag_stack_check)
7864 rtx last = gen_rtx_MEM (DImode,
7865 plus_constant (Pmode, ptr, -leftover));
7866 MEM_VOLATILE_P (last) = 1;
7867 emit_move_insn (last, const0_rtx);
7870 if (flag_stack_check)
7872 /* If -fstack-check is specified we have to load the entire
7873 constant into a register and subtract from the sp in one go,
7874 because the probed stack size is not equal to the frame size. */
7875 HOST_WIDE_INT lo, hi;
7876 lo = ((frame_size & 0xffff) ^ 0x8000) - 0x8000;
7877 hi = frame_size - lo;
7879 emit_move_insn (ptr, GEN_INT (hi));
7880 emit_insn (gen_adddi3 (ptr, ptr, GEN_INT (lo)));
7881 seq = emit_insn (gen_subdi3 (stack_pointer_rtx, stack_pointer_rtx,
7882 ptr));
7884 else
7886 seq = emit_insn (gen_adddi3 (stack_pointer_rtx, ptr,
7887 GEN_INT (-leftover)));
7890 /* This alternative is special, because the DWARF code cannot
7891 possibly intuit through the loop above. So we invent this
7892 note it looks at instead. */
7893 RTX_FRAME_RELATED_P (seq) = 1;
7894 add_reg_note (seq, REG_FRAME_RELATED_EXPR,
7895 gen_rtx_SET (VOIDmode, stack_pointer_rtx,
7896 plus_constant (Pmode, stack_pointer_rtx,
7897 -frame_size)));
7900 /* Cope with very large offsets to the register save area. */
7901 sa_bias = 0;
7902 sa_reg = stack_pointer_rtx;
7903 if (reg_offset + sa_size > 0x8000)
7905 int low = ((reg_offset & 0xffff) ^ 0x8000) - 0x8000;
7906 rtx sa_bias_rtx;
7908 if (low + sa_size <= 0x8000)
7909 sa_bias = reg_offset - low, reg_offset = low;
7910 else
7911 sa_bias = reg_offset, reg_offset = 0;
7913 sa_reg = gen_rtx_REG (DImode, 24);
7914 sa_bias_rtx = GEN_INT (sa_bias);
7916 if (add_operand (sa_bias_rtx, DImode))
7917 emit_insn (gen_adddi3 (sa_reg, stack_pointer_rtx, sa_bias_rtx));
7918 else
7920 emit_move_insn (sa_reg, sa_bias_rtx);
7921 emit_insn (gen_adddi3 (sa_reg, stack_pointer_rtx, sa_reg));
7925 /* Save regs in stack order. Beginning with VMS PV. */
7926 if (TARGET_ABI_OPEN_VMS && alpha_procedure_type == PT_STACK)
7927 emit_frame_store (REG_PV, stack_pointer_rtx, 0, 0);
7929 /* Save register RA next. */
7930 if (imask & (1UL << REG_RA))
7932 emit_frame_store (REG_RA, sa_reg, sa_bias, reg_offset);
7933 imask &= ~(1UL << REG_RA);
7934 reg_offset += 8;
7937 /* Now save any other registers required to be saved. */
7938 for (i = 0; i < 31; i++)
7939 if (imask & (1UL << i))
7941 emit_frame_store (i, sa_reg, sa_bias, reg_offset);
7942 reg_offset += 8;
7945 for (i = 0; i < 31; i++)
7946 if (fmask & (1UL << i))
7948 emit_frame_store (i+32, sa_reg, sa_bias, reg_offset);
7949 reg_offset += 8;
7952 if (TARGET_ABI_OPEN_VMS)
7954 /* Register frame procedures save the fp. */
7955 if (alpha_procedure_type == PT_REGISTER)
7957 rtx insn = emit_move_insn (gen_rtx_REG (DImode, vms_save_fp_regno),
7958 hard_frame_pointer_rtx);
7959 add_reg_note (insn, REG_CFA_REGISTER, NULL);
7960 RTX_FRAME_RELATED_P (insn) = 1;
7963 if (alpha_procedure_type != PT_NULL && vms_base_regno != REG_PV)
7964 emit_insn (gen_force_movdi (gen_rtx_REG (DImode, vms_base_regno),
7965 gen_rtx_REG (DImode, REG_PV)));
7967 if (alpha_procedure_type != PT_NULL
7968 && vms_unwind_regno == HARD_FRAME_POINTER_REGNUM)
7969 FRP (emit_move_insn (hard_frame_pointer_rtx, stack_pointer_rtx));
7971 /* If we have to allocate space for outgoing args, do it now. */
7972 if (crtl->outgoing_args_size != 0)
7974 rtx seq
7975 = emit_move_insn (stack_pointer_rtx,
7976 plus_constant
7977 (Pmode, hard_frame_pointer_rtx,
7978 - (ALPHA_ROUND
7979 (crtl->outgoing_args_size))));
7981 /* Only set FRAME_RELATED_P on the stack adjustment we just emitted
7982 if ! frame_pointer_needed. Setting the bit will change the CFA
7983 computation rule to use sp again, which would be wrong if we had
7984 frame_pointer_needed, as this means sp might move unpredictably
7985 later on.
7987 Also, note that
7988 frame_pointer_needed
7989 => vms_unwind_regno == HARD_FRAME_POINTER_REGNUM
7991 crtl->outgoing_args_size != 0
7992 => alpha_procedure_type != PT_NULL,
7994 so when we are not setting the bit here, we are guaranteed to
7995 have emitted an FRP frame pointer update just before. */
7996 RTX_FRAME_RELATED_P (seq) = ! frame_pointer_needed;
7999 else
8001 /* If we need a frame pointer, set it from the stack pointer. */
8002 if (frame_pointer_needed)
8004 if (TARGET_CAN_FAULT_IN_PROLOGUE)
8005 FRP (emit_move_insn (hard_frame_pointer_rtx, stack_pointer_rtx));
8006 else
8007 /* This must always be the last instruction in the
8008 prologue, thus we emit a special move + clobber. */
8009 FRP (emit_insn (gen_init_fp (hard_frame_pointer_rtx,
8010 stack_pointer_rtx, sa_reg)));
8014 /* The ABIs for VMS and OSF/1 say that while we can schedule insns into
8015 the prologue, for exception handling reasons, we cannot do this for
8016 any insn that might fault. We could prevent this for mems with a
8017 (clobber:BLK (scratch)), but this doesn't work for fp insns. So we
8018 have to prevent all such scheduling with a blockage.
8020 Linux, on the other hand, never bothered to implement OSF/1's
8021 exception handling, and so doesn't care about such things. Anyone
8022 planning to use dwarf2 frame-unwind info can also omit the blockage. */
8024 if (! TARGET_CAN_FAULT_IN_PROLOGUE)
8025 emit_insn (gen_blockage ());
8028 /* Count the number of .file directives, so that .loc is up to date. */
8029 int num_source_filenames = 0;
8031 /* Output the textual info surrounding the prologue. */
8033 void
8034 alpha_start_function (FILE *file, const char *fnname,
8035 tree decl ATTRIBUTE_UNUSED)
8037 unsigned long imask = 0;
8038 unsigned long fmask = 0;
8039 /* Stack space needed for pushing registers clobbered by us. */
8040 HOST_WIDE_INT sa_size;
8041 /* Complete stack size needed. */
8042 unsigned HOST_WIDE_INT frame_size;
8043 /* The maximum debuggable frame size. */
8044 unsigned HOST_WIDE_INT max_frame_size = 1UL << 31;
8045 /* Offset from base reg to register save area. */
8046 HOST_WIDE_INT reg_offset;
8047 char *entry_label = (char *) alloca (strlen (fnname) + 6);
8048 char *tramp_label = (char *) alloca (strlen (fnname) + 6);
8049 int i;
8051 #if TARGET_ABI_OPEN_VMS
8052 vms_start_function (fnname);
8053 #endif
8055 alpha_fnname = fnname;
8056 sa_size = alpha_sa_size ();
8057 frame_size = compute_frame_size (get_frame_size (), sa_size);
8059 if (TARGET_ABI_OPEN_VMS)
8060 reg_offset = 8 + 8 * cfun->machine->uses_condition_handler;
8061 else
8062 reg_offset = ALPHA_ROUND (crtl->outgoing_args_size);
8064 alpha_sa_mask (&imask, &fmask);
8066 /* Issue function start and label. */
8067 if (TARGET_ABI_OPEN_VMS || !flag_inhibit_size_directive)
8069 fputs ("\t.ent ", file);
8070 assemble_name (file, fnname);
8071 putc ('\n', file);
8073 /* If the function needs GP, we'll write the "..ng" label there.
8074 Otherwise, do it here. */
8075 if (TARGET_ABI_OSF
8076 && ! alpha_function_needs_gp
8077 && ! cfun->is_thunk)
8079 putc ('$', file);
8080 assemble_name (file, fnname);
8081 fputs ("..ng:\n", file);
8084 /* Nested functions on VMS that are potentially called via trampoline
8085 get a special transfer entry point that loads the called functions
8086 procedure descriptor and static chain. */
8087 if (TARGET_ABI_OPEN_VMS
8088 && !TREE_PUBLIC (decl)
8089 && DECL_CONTEXT (decl)
8090 && !TYPE_P (DECL_CONTEXT (decl))
8091 && TREE_CODE (DECL_CONTEXT (decl)) != TRANSLATION_UNIT_DECL)
8093 strcpy (tramp_label, fnname);
8094 strcat (tramp_label, "..tr");
8095 ASM_OUTPUT_LABEL (file, tramp_label);
8096 fprintf (file, "\tldq $1,24($27)\n");
8097 fprintf (file, "\tldq $27,16($27)\n");
8100 strcpy (entry_label, fnname);
8101 if (TARGET_ABI_OPEN_VMS)
8102 strcat (entry_label, "..en");
8104 ASM_OUTPUT_LABEL (file, entry_label);
8105 inside_function = TRUE;
8107 if (TARGET_ABI_OPEN_VMS)
8108 fprintf (file, "\t.base $%d\n", vms_base_regno);
8110 if (TARGET_ABI_OSF
8111 && TARGET_IEEE_CONFORMANT
8112 && !flag_inhibit_size_directive)
8114 /* Set flags in procedure descriptor to request IEEE-conformant
8115 math-library routines. The value we set it to is PDSC_EXC_IEEE
8116 (/usr/include/pdsc.h). */
8117 fputs ("\t.eflag 48\n", file);
8120 /* Set up offsets to alpha virtual arg/local debugging pointer. */
8121 alpha_auto_offset = -frame_size + crtl->args.pretend_args_size;
8122 alpha_arg_offset = -frame_size + 48;
8124 /* Describe our frame. If the frame size is larger than an integer,
8125 print it as zero to avoid an assembler error. We won't be
8126 properly describing such a frame, but that's the best we can do. */
8127 if (TARGET_ABI_OPEN_VMS)
8128 fprintf (file, "\t.frame $%d," HOST_WIDE_INT_PRINT_DEC ",$26,"
8129 HOST_WIDE_INT_PRINT_DEC "\n",
8130 vms_unwind_regno,
8131 frame_size >= (1UL << 31) ? 0 : frame_size,
8132 reg_offset);
8133 else if (!flag_inhibit_size_directive)
8134 fprintf (file, "\t.frame $%d," HOST_WIDE_INT_PRINT_DEC ",$26,%d\n",
8135 (frame_pointer_needed
8136 ? HARD_FRAME_POINTER_REGNUM : STACK_POINTER_REGNUM),
8137 frame_size >= max_frame_size ? 0 : frame_size,
8138 crtl->args.pretend_args_size);
8140 /* Describe which registers were spilled. */
8141 if (TARGET_ABI_OPEN_VMS)
8143 if (imask)
8144 /* ??? Does VMS care if mask contains ra? The old code didn't
8145 set it, so I don't here. */
8146 fprintf (file, "\t.mask 0x%lx,0\n", imask & ~(1UL << REG_RA));
8147 if (fmask)
8148 fprintf (file, "\t.fmask 0x%lx,0\n", fmask);
8149 if (alpha_procedure_type == PT_REGISTER)
8150 fprintf (file, "\t.fp_save $%d\n", vms_save_fp_regno);
8152 else if (!flag_inhibit_size_directive)
8154 if (imask)
8156 fprintf (file, "\t.mask 0x%lx," HOST_WIDE_INT_PRINT_DEC "\n", imask,
8157 frame_size >= max_frame_size ? 0 : reg_offset - frame_size);
8159 for (i = 0; i < 32; ++i)
8160 if (imask & (1UL << i))
8161 reg_offset += 8;
8164 if (fmask)
8165 fprintf (file, "\t.fmask 0x%lx," HOST_WIDE_INT_PRINT_DEC "\n", fmask,
8166 frame_size >= max_frame_size ? 0 : reg_offset - frame_size);
8169 #if TARGET_ABI_OPEN_VMS
8170 /* If a user condition handler has been installed at some point, emit
8171 the procedure descriptor bits to point the Condition Handling Facility
8172 at the indirection wrapper, and state the fp offset at which the user
8173 handler may be found. */
8174 if (cfun->machine->uses_condition_handler)
8176 fprintf (file, "\t.handler __gcc_shell_handler\n");
8177 fprintf (file, "\t.handler_data %d\n", VMS_COND_HANDLER_FP_OFFSET);
8180 #ifdef TARGET_VMS_CRASH_DEBUG
8181 /* Support of minimal traceback info. */
8182 switch_to_section (readonly_data_section);
8183 fprintf (file, "\t.align 3\n");
8184 assemble_name (file, fnname); fputs ("..na:\n", file);
8185 fputs ("\t.ascii \"", file);
8186 assemble_name (file, fnname);
8187 fputs ("\\0\"\n", file);
8188 switch_to_section (text_section);
8189 #endif
8190 #endif /* TARGET_ABI_OPEN_VMS */
8193 /* Emit the .prologue note at the scheduled end of the prologue. */
8195 static void
8196 alpha_output_function_end_prologue (FILE *file)
8198 if (TARGET_ABI_OPEN_VMS)
8199 fputs ("\t.prologue\n", file);
8200 else if (!flag_inhibit_size_directive)
8201 fprintf (file, "\t.prologue %d\n",
8202 alpha_function_needs_gp || cfun->is_thunk);
8205 /* Write function epilogue. */
8207 void
8208 alpha_expand_epilogue (void)
8210 /* Registers to save. */
8211 unsigned long imask = 0;
8212 unsigned long fmask = 0;
8213 /* Stack space needed for pushing registers clobbered by us. */
8214 HOST_WIDE_INT sa_size;
8215 /* Complete stack size needed. */
8216 HOST_WIDE_INT frame_size;
8217 /* Offset from base reg to register save area. */
8218 HOST_WIDE_INT reg_offset;
8219 int fp_is_frame_pointer, fp_offset;
8220 rtx sa_reg, sa_reg_exp = NULL;
8221 rtx sp_adj1, sp_adj2, mem, reg, insn;
8222 rtx eh_ofs;
8223 rtx cfa_restores = NULL_RTX;
8224 int i;
8226 sa_size = alpha_sa_size ();
8227 frame_size = compute_frame_size (get_frame_size (), sa_size);
8229 if (TARGET_ABI_OPEN_VMS)
8231 if (alpha_procedure_type == PT_STACK)
8232 reg_offset = 8 + 8 * cfun->machine->uses_condition_handler;
8233 else
8234 reg_offset = 0;
8236 else
8237 reg_offset = ALPHA_ROUND (crtl->outgoing_args_size);
8239 alpha_sa_mask (&imask, &fmask);
8241 fp_is_frame_pointer
8242 = (TARGET_ABI_OPEN_VMS
8243 ? alpha_procedure_type == PT_STACK
8244 : frame_pointer_needed);
8245 fp_offset = 0;
8246 sa_reg = stack_pointer_rtx;
8248 if (crtl->calls_eh_return)
8249 eh_ofs = EH_RETURN_STACKADJ_RTX;
8250 else
8251 eh_ofs = NULL_RTX;
8253 if (sa_size)
8255 /* If we have a frame pointer, restore SP from it. */
8256 if (TARGET_ABI_OPEN_VMS
8257 ? vms_unwind_regno == HARD_FRAME_POINTER_REGNUM
8258 : frame_pointer_needed)
8259 emit_move_insn (stack_pointer_rtx, hard_frame_pointer_rtx);
8261 /* Cope with very large offsets to the register save area. */
8262 if (reg_offset + sa_size > 0x8000)
8264 int low = ((reg_offset & 0xffff) ^ 0x8000) - 0x8000;
8265 HOST_WIDE_INT bias;
8267 if (low + sa_size <= 0x8000)
8268 bias = reg_offset - low, reg_offset = low;
8269 else
8270 bias = reg_offset, reg_offset = 0;
8272 sa_reg = gen_rtx_REG (DImode, 22);
8273 sa_reg_exp = plus_constant (Pmode, stack_pointer_rtx, bias);
8275 emit_move_insn (sa_reg, sa_reg_exp);
8278 /* Restore registers in order, excepting a true frame pointer. */
8280 mem = gen_frame_mem (DImode, plus_constant (Pmode, sa_reg, reg_offset));
8281 reg = gen_rtx_REG (DImode, REG_RA);
8282 emit_move_insn (reg, mem);
8283 cfa_restores = alloc_reg_note (REG_CFA_RESTORE, reg, cfa_restores);
8285 reg_offset += 8;
8286 imask &= ~(1UL << REG_RA);
8288 for (i = 0; i < 31; ++i)
8289 if (imask & (1UL << i))
8291 if (i == HARD_FRAME_POINTER_REGNUM && fp_is_frame_pointer)
8292 fp_offset = reg_offset;
8293 else
8295 mem = gen_frame_mem (DImode,
8296 plus_constant (Pmode, sa_reg,
8297 reg_offset));
8298 reg = gen_rtx_REG (DImode, i);
8299 emit_move_insn (reg, mem);
8300 cfa_restores = alloc_reg_note (REG_CFA_RESTORE, reg,
8301 cfa_restores);
8303 reg_offset += 8;
8306 for (i = 0; i < 31; ++i)
8307 if (fmask & (1UL << i))
8309 mem = gen_frame_mem (DFmode, plus_constant (Pmode, sa_reg,
8310 reg_offset));
8311 reg = gen_rtx_REG (DFmode, i+32);
8312 emit_move_insn (reg, mem);
8313 cfa_restores = alloc_reg_note (REG_CFA_RESTORE, reg, cfa_restores);
8314 reg_offset += 8;
8318 if (frame_size || eh_ofs)
8320 sp_adj1 = stack_pointer_rtx;
8322 if (eh_ofs)
8324 sp_adj1 = gen_rtx_REG (DImode, 23);
8325 emit_move_insn (sp_adj1,
8326 gen_rtx_PLUS (Pmode, stack_pointer_rtx, eh_ofs));
8329 /* If the stack size is large, begin computation into a temporary
8330 register so as not to interfere with a potential fp restore,
8331 which must be consecutive with an SP restore. */
8332 if (frame_size < 32768 && !cfun->calls_alloca)
8333 sp_adj2 = GEN_INT (frame_size);
8334 else if (frame_size < 0x40007fffL)
8336 int low = ((frame_size & 0xffff) ^ 0x8000) - 0x8000;
8338 sp_adj2 = plus_constant (Pmode, sp_adj1, frame_size - low);
8339 if (sa_reg_exp && rtx_equal_p (sa_reg_exp, sp_adj2))
8340 sp_adj1 = sa_reg;
8341 else
8343 sp_adj1 = gen_rtx_REG (DImode, 23);
8344 emit_move_insn (sp_adj1, sp_adj2);
8346 sp_adj2 = GEN_INT (low);
8348 else
8350 rtx tmp = gen_rtx_REG (DImode, 23);
8351 sp_adj2 = alpha_emit_set_const (tmp, DImode, frame_size, 3, false);
8352 if (!sp_adj2)
8354 /* We can't drop new things to memory this late, afaik,
8355 so build it up by pieces. */
8356 sp_adj2 = alpha_emit_set_long_const (tmp, frame_size,
8357 -(frame_size < 0));
8358 gcc_assert (sp_adj2);
8362 /* From now on, things must be in order. So emit blockages. */
8364 /* Restore the frame pointer. */
8365 if (fp_is_frame_pointer)
8367 emit_insn (gen_blockage ());
8368 mem = gen_frame_mem (DImode, plus_constant (Pmode, sa_reg,
8369 fp_offset));
8370 emit_move_insn (hard_frame_pointer_rtx, mem);
8371 cfa_restores = alloc_reg_note (REG_CFA_RESTORE,
8372 hard_frame_pointer_rtx, cfa_restores);
8374 else if (TARGET_ABI_OPEN_VMS)
8376 emit_insn (gen_blockage ());
8377 emit_move_insn (hard_frame_pointer_rtx,
8378 gen_rtx_REG (DImode, vms_save_fp_regno));
8379 cfa_restores = alloc_reg_note (REG_CFA_RESTORE,
8380 hard_frame_pointer_rtx, cfa_restores);
8383 /* Restore the stack pointer. */
8384 emit_insn (gen_blockage ());
8385 if (sp_adj2 == const0_rtx)
8386 insn = emit_move_insn (stack_pointer_rtx, sp_adj1);
8387 else
8388 insn = emit_move_insn (stack_pointer_rtx,
8389 gen_rtx_PLUS (DImode, sp_adj1, sp_adj2));
8390 REG_NOTES (insn) = cfa_restores;
8391 add_reg_note (insn, REG_CFA_DEF_CFA, stack_pointer_rtx);
8392 RTX_FRAME_RELATED_P (insn) = 1;
8394 else
8396 gcc_assert (cfa_restores == NULL);
8398 if (TARGET_ABI_OPEN_VMS && alpha_procedure_type == PT_REGISTER)
8400 emit_insn (gen_blockage ());
8401 insn = emit_move_insn (hard_frame_pointer_rtx,
8402 gen_rtx_REG (DImode, vms_save_fp_regno));
8403 add_reg_note (insn, REG_CFA_RESTORE, hard_frame_pointer_rtx);
8404 RTX_FRAME_RELATED_P (insn) = 1;
8409 /* Output the rest of the textual info surrounding the epilogue. */
8411 void
8412 alpha_end_function (FILE *file, const char *fnname, tree decl ATTRIBUTE_UNUSED)
8414 rtx insn;
8416 /* We output a nop after noreturn calls at the very end of the function to
8417 ensure that the return address always remains in the caller's code range,
8418 as not doing so might confuse unwinding engines. */
8419 insn = get_last_insn ();
8420 if (!INSN_P (insn))
8421 insn = prev_active_insn (insn);
8422 if (insn && CALL_P (insn))
8423 output_asm_insn (get_insn_template (CODE_FOR_nop, NULL), NULL);
8425 #if TARGET_ABI_OPEN_VMS
8426 /* Write the linkage entries. */
8427 alpha_write_linkage (file, fnname);
8428 #endif
8430 /* End the function. */
8431 if (TARGET_ABI_OPEN_VMS
8432 || !flag_inhibit_size_directive)
8434 fputs ("\t.end ", file);
8435 assemble_name (file, fnname);
8436 putc ('\n', file);
8438 inside_function = FALSE;
8441 #if TARGET_ABI_OSF
8442 /* Emit a tail call to FUNCTION after adjusting THIS by DELTA.
8444 In order to avoid the hordes of differences between generated code
8445 with and without TARGET_EXPLICIT_RELOCS, and to avoid duplicating
8446 lots of code loading up large constants, generate rtl and emit it
8447 instead of going straight to text.
8449 Not sure why this idea hasn't been explored before... */
8451 static void
8452 alpha_output_mi_thunk_osf (FILE *file, tree thunk_fndecl ATTRIBUTE_UNUSED,
8453 HOST_WIDE_INT delta, HOST_WIDE_INT vcall_offset,
8454 tree function)
8456 HOST_WIDE_INT hi, lo;
8457 rtx this_rtx, insn, funexp;
8459 /* We always require a valid GP. */
8460 emit_insn (gen_prologue_ldgp ());
8461 emit_note (NOTE_INSN_PROLOGUE_END);
8463 /* Find the "this" pointer. If the function returns a structure,
8464 the structure return pointer is in $16. */
8465 if (aggregate_value_p (TREE_TYPE (TREE_TYPE (function)), function))
8466 this_rtx = gen_rtx_REG (Pmode, 17);
8467 else
8468 this_rtx = gen_rtx_REG (Pmode, 16);
8470 /* Add DELTA. When possible we use ldah+lda. Otherwise load the
8471 entire constant for the add. */
8472 lo = ((delta & 0xffff) ^ 0x8000) - 0x8000;
8473 hi = (((delta - lo) & 0xffffffff) ^ 0x80000000) - 0x80000000;
8474 if (hi + lo == delta)
8476 if (hi)
8477 emit_insn (gen_adddi3 (this_rtx, this_rtx, GEN_INT (hi)));
8478 if (lo)
8479 emit_insn (gen_adddi3 (this_rtx, this_rtx, GEN_INT (lo)));
8481 else
8483 rtx tmp = alpha_emit_set_long_const (gen_rtx_REG (Pmode, 0),
8484 delta, -(delta < 0));
8485 emit_insn (gen_adddi3 (this_rtx, this_rtx, tmp));
8488 /* Add a delta stored in the vtable at VCALL_OFFSET. */
8489 if (vcall_offset)
8491 rtx tmp, tmp2;
8493 tmp = gen_rtx_REG (Pmode, 0);
8494 emit_move_insn (tmp, gen_rtx_MEM (Pmode, this_rtx));
8496 lo = ((vcall_offset & 0xffff) ^ 0x8000) - 0x8000;
8497 hi = (((vcall_offset - lo) & 0xffffffff) ^ 0x80000000) - 0x80000000;
8498 if (hi + lo == vcall_offset)
8500 if (hi)
8501 emit_insn (gen_adddi3 (tmp, tmp, GEN_INT (hi)));
8503 else
8505 tmp2 = alpha_emit_set_long_const (gen_rtx_REG (Pmode, 1),
8506 vcall_offset, -(vcall_offset < 0));
8507 emit_insn (gen_adddi3 (tmp, tmp, tmp2));
8508 lo = 0;
8510 if (lo)
8511 tmp2 = gen_rtx_PLUS (Pmode, tmp, GEN_INT (lo));
8512 else
8513 tmp2 = tmp;
8514 emit_move_insn (tmp, gen_rtx_MEM (Pmode, tmp2));
8516 emit_insn (gen_adddi3 (this_rtx, this_rtx, tmp));
8519 /* Generate a tail call to the target function. */
8520 if (! TREE_USED (function))
8522 assemble_external (function);
8523 TREE_USED (function) = 1;
8525 funexp = XEXP (DECL_RTL (function), 0);
8526 funexp = gen_rtx_MEM (FUNCTION_MODE, funexp);
8527 insn = emit_call_insn (gen_sibcall (funexp, const0_rtx));
8528 SIBLING_CALL_P (insn) = 1;
8530 /* Run just enough of rest_of_compilation to get the insns emitted.
8531 There's not really enough bulk here to make other passes such as
8532 instruction scheduling worth while. Note that use_thunk calls
8533 assemble_start_function and assemble_end_function. */
8534 insn = get_insns ();
8535 shorten_branches (insn);
8536 final_start_function (insn, file, 1);
8537 final (insn, file, 1);
8538 final_end_function ();
8540 #endif /* TARGET_ABI_OSF */
8542 /* Debugging support. */
8544 #include "gstab.h"
8546 /* Name of the file containing the current function. */
8548 static const char *current_function_file = "";
8550 /* Offsets to alpha virtual arg/local debugging pointers. */
8552 long alpha_arg_offset;
8553 long alpha_auto_offset;
8555 /* Emit a new filename to a stream. */
8557 void
8558 alpha_output_filename (FILE *stream, const char *name)
8560 static int first_time = TRUE;
8562 if (first_time)
8564 first_time = FALSE;
8565 ++num_source_filenames;
8566 current_function_file = name;
8567 fprintf (stream, "\t.file\t%d ", num_source_filenames);
8568 output_quoted_string (stream, name);
8569 fprintf (stream, "\n");
8572 else if (name != current_function_file
8573 && strcmp (name, current_function_file) != 0)
8575 ++num_source_filenames;
8576 current_function_file = name;
8577 fprintf (stream, "\t.file\t%d ", num_source_filenames);
8579 output_quoted_string (stream, name);
8580 fprintf (stream, "\n");
8584 /* Structure to show the current status of registers and memory. */
8586 struct shadow_summary
8588 struct {
8589 unsigned int i : 31; /* Mask of int regs */
8590 unsigned int fp : 31; /* Mask of fp regs */
8591 unsigned int mem : 1; /* mem == imem | fpmem */
8592 } used, defd;
8595 /* Summary the effects of expression X on the machine. Update SUM, a pointer
8596 to the summary structure. SET is nonzero if the insn is setting the
8597 object, otherwise zero. */
8599 static void
8600 summarize_insn (rtx x, struct shadow_summary *sum, int set)
8602 const char *format_ptr;
8603 int i, j;
8605 if (x == 0)
8606 return;
8608 switch (GET_CODE (x))
8610 /* ??? Note that this case would be incorrect if the Alpha had a
8611 ZERO_EXTRACT in SET_DEST. */
8612 case SET:
8613 summarize_insn (SET_SRC (x), sum, 0);
8614 summarize_insn (SET_DEST (x), sum, 1);
8615 break;
8617 case CLOBBER:
8618 summarize_insn (XEXP (x, 0), sum, 1);
8619 break;
8621 case USE:
8622 summarize_insn (XEXP (x, 0), sum, 0);
8623 break;
8625 case ASM_OPERANDS:
8626 for (i = ASM_OPERANDS_INPUT_LENGTH (x) - 1; i >= 0; i--)
8627 summarize_insn (ASM_OPERANDS_INPUT (x, i), sum, 0);
8628 break;
8630 case PARALLEL:
8631 for (i = XVECLEN (x, 0) - 1; i >= 0; i--)
8632 summarize_insn (XVECEXP (x, 0, i), sum, 0);
8633 break;
8635 case SUBREG:
8636 summarize_insn (SUBREG_REG (x), sum, 0);
8637 break;
8639 case REG:
8641 int regno = REGNO (x);
8642 unsigned long mask = ((unsigned long) 1) << (regno % 32);
8644 if (regno == 31 || regno == 63)
8645 break;
8647 if (set)
8649 if (regno < 32)
8650 sum->defd.i |= mask;
8651 else
8652 sum->defd.fp |= mask;
8654 else
8656 if (regno < 32)
8657 sum->used.i |= mask;
8658 else
8659 sum->used.fp |= mask;
8662 break;
8664 case MEM:
8665 if (set)
8666 sum->defd.mem = 1;
8667 else
8668 sum->used.mem = 1;
8670 /* Find the regs used in memory address computation: */
8671 summarize_insn (XEXP (x, 0), sum, 0);
8672 break;
8674 case CONST_INT: case CONST_DOUBLE:
8675 case SYMBOL_REF: case LABEL_REF: case CONST:
8676 case SCRATCH: case ASM_INPUT:
8677 break;
8679 /* Handle common unary and binary ops for efficiency. */
8680 case COMPARE: case PLUS: case MINUS: case MULT: case DIV:
8681 case MOD: case UDIV: case UMOD: case AND: case IOR:
8682 case XOR: case ASHIFT: case ROTATE: case ASHIFTRT: case LSHIFTRT:
8683 case ROTATERT: case SMIN: case SMAX: case UMIN: case UMAX:
8684 case NE: case EQ: case GE: case GT: case LE:
8685 case LT: case GEU: case GTU: case LEU: case LTU:
8686 summarize_insn (XEXP (x, 0), sum, 0);
8687 summarize_insn (XEXP (x, 1), sum, 0);
8688 break;
8690 case NEG: case NOT: case SIGN_EXTEND: case ZERO_EXTEND:
8691 case TRUNCATE: case FLOAT_EXTEND: case FLOAT_TRUNCATE: case FLOAT:
8692 case FIX: case UNSIGNED_FLOAT: case UNSIGNED_FIX: case ABS:
8693 case SQRT: case FFS:
8694 summarize_insn (XEXP (x, 0), sum, 0);
8695 break;
8697 default:
8698 format_ptr = GET_RTX_FORMAT (GET_CODE (x));
8699 for (i = GET_RTX_LENGTH (GET_CODE (x)) - 1; i >= 0; i--)
8700 switch (format_ptr[i])
8702 case 'e':
8703 summarize_insn (XEXP (x, i), sum, 0);
8704 break;
8706 case 'E':
8707 for (j = XVECLEN (x, i) - 1; j >= 0; j--)
8708 summarize_insn (XVECEXP (x, i, j), sum, 0);
8709 break;
8711 case 'i':
8712 break;
8714 default:
8715 gcc_unreachable ();
8720 /* Ensure a sufficient number of `trapb' insns are in the code when
8721 the user requests code with a trap precision of functions or
8722 instructions.
8724 In naive mode, when the user requests a trap-precision of
8725 "instruction", a trapb is needed after every instruction that may
8726 generate a trap. This ensures that the code is resumption safe but
8727 it is also slow.
8729 When optimizations are turned on, we delay issuing a trapb as long
8730 as possible. In this context, a trap shadow is the sequence of
8731 instructions that starts with a (potentially) trap generating
8732 instruction and extends to the next trapb or call_pal instruction
8733 (but GCC never generates call_pal by itself). We can delay (and
8734 therefore sometimes omit) a trapb subject to the following
8735 conditions:
8737 (a) On entry to the trap shadow, if any Alpha register or memory
8738 location contains a value that is used as an operand value by some
8739 instruction in the trap shadow (live on entry), then no instruction
8740 in the trap shadow may modify the register or memory location.
8742 (b) Within the trap shadow, the computation of the base register
8743 for a memory load or store instruction may not involve using the
8744 result of an instruction that might generate an UNPREDICTABLE
8745 result.
8747 (c) Within the trap shadow, no register may be used more than once
8748 as a destination register. (This is to make life easier for the
8749 trap-handler.)
8751 (d) The trap shadow may not include any branch instructions. */
8753 static void
8754 alpha_handle_trap_shadows (void)
8756 struct shadow_summary shadow;
8757 int trap_pending, exception_nesting;
8758 rtx i, n;
8760 trap_pending = 0;
8761 exception_nesting = 0;
8762 shadow.used.i = 0;
8763 shadow.used.fp = 0;
8764 shadow.used.mem = 0;
8765 shadow.defd = shadow.used;
8767 for (i = get_insns (); i ; i = NEXT_INSN (i))
8769 if (NOTE_P (i))
8771 switch (NOTE_KIND (i))
8773 case NOTE_INSN_EH_REGION_BEG:
8774 exception_nesting++;
8775 if (trap_pending)
8776 goto close_shadow;
8777 break;
8779 case NOTE_INSN_EH_REGION_END:
8780 exception_nesting--;
8781 if (trap_pending)
8782 goto close_shadow;
8783 break;
8785 case NOTE_INSN_EPILOGUE_BEG:
8786 if (trap_pending && alpha_tp >= ALPHA_TP_FUNC)
8787 goto close_shadow;
8788 break;
8791 else if (trap_pending)
8793 if (alpha_tp == ALPHA_TP_FUNC)
8795 if (JUMP_P (i)
8796 && GET_CODE (PATTERN (i)) == RETURN)
8797 goto close_shadow;
8799 else if (alpha_tp == ALPHA_TP_INSN)
8801 if (optimize > 0)
8803 struct shadow_summary sum;
8805 sum.used.i = 0;
8806 sum.used.fp = 0;
8807 sum.used.mem = 0;
8808 sum.defd = sum.used;
8810 switch (GET_CODE (i))
8812 case INSN:
8813 /* Annoyingly, get_attr_trap will die on these. */
8814 if (GET_CODE (PATTERN (i)) == USE
8815 || GET_CODE (PATTERN (i)) == CLOBBER)
8816 break;
8818 summarize_insn (PATTERN (i), &sum, 0);
8820 if ((sum.defd.i & shadow.defd.i)
8821 || (sum.defd.fp & shadow.defd.fp))
8823 /* (c) would be violated */
8824 goto close_shadow;
8827 /* Combine shadow with summary of current insn: */
8828 shadow.used.i |= sum.used.i;
8829 shadow.used.fp |= sum.used.fp;
8830 shadow.used.mem |= sum.used.mem;
8831 shadow.defd.i |= sum.defd.i;
8832 shadow.defd.fp |= sum.defd.fp;
8833 shadow.defd.mem |= sum.defd.mem;
8835 if ((sum.defd.i & shadow.used.i)
8836 || (sum.defd.fp & shadow.used.fp)
8837 || (sum.defd.mem & shadow.used.mem))
8839 /* (a) would be violated (also takes care of (b)) */
8840 gcc_assert (get_attr_trap (i) != TRAP_YES
8841 || (!(sum.defd.i & sum.used.i)
8842 && !(sum.defd.fp & sum.used.fp)));
8844 goto close_shadow;
8846 break;
8848 case BARRIER:
8849 /* __builtin_unreachable can expand to no code at all,
8850 leaving (barrier) RTXes in the instruction stream. */
8851 goto close_shadow_notrapb;
8853 case JUMP_INSN:
8854 case CALL_INSN:
8855 case CODE_LABEL:
8856 goto close_shadow;
8858 default:
8859 gcc_unreachable ();
8862 else
8864 close_shadow:
8865 n = emit_insn_before (gen_trapb (), i);
8866 PUT_MODE (n, TImode);
8867 PUT_MODE (i, TImode);
8868 close_shadow_notrapb:
8869 trap_pending = 0;
8870 shadow.used.i = 0;
8871 shadow.used.fp = 0;
8872 shadow.used.mem = 0;
8873 shadow.defd = shadow.used;
8878 if ((exception_nesting > 0 || alpha_tp >= ALPHA_TP_FUNC)
8879 && NONJUMP_INSN_P (i)
8880 && GET_CODE (PATTERN (i)) != USE
8881 && GET_CODE (PATTERN (i)) != CLOBBER
8882 && get_attr_trap (i) == TRAP_YES)
8884 if (optimize && !trap_pending)
8885 summarize_insn (PATTERN (i), &shadow, 0);
8886 trap_pending = 1;
8891 /* Alpha can only issue instruction groups simultaneously if they are
8892 suitably aligned. This is very processor-specific. */
8893 /* There are a number of entries in alphaev4_insn_pipe and alphaev5_insn_pipe
8894 that are marked "fake". These instructions do not exist on that target,
8895 but it is possible to see these insns with deranged combinations of
8896 command-line options, such as "-mtune=ev4 -mmax". Instead of aborting,
8897 choose a result at random. */
8899 enum alphaev4_pipe {
8900 EV4_STOP = 0,
8901 EV4_IB0 = 1,
8902 EV4_IB1 = 2,
8903 EV4_IBX = 4
8906 enum alphaev5_pipe {
8907 EV5_STOP = 0,
8908 EV5_NONE = 1,
8909 EV5_E01 = 2,
8910 EV5_E0 = 4,
8911 EV5_E1 = 8,
8912 EV5_FAM = 16,
8913 EV5_FA = 32,
8914 EV5_FM = 64
8917 static enum alphaev4_pipe
8918 alphaev4_insn_pipe (rtx insn)
8920 if (recog_memoized (insn) < 0)
8921 return EV4_STOP;
8922 if (get_attr_length (insn) != 4)
8923 return EV4_STOP;
8925 switch (get_attr_type (insn))
8927 case TYPE_ILD:
8928 case TYPE_LDSYM:
8929 case TYPE_FLD:
8930 case TYPE_LD_L:
8931 return EV4_IBX;
8933 case TYPE_IADD:
8934 case TYPE_ILOG:
8935 case TYPE_ICMOV:
8936 case TYPE_ICMP:
8937 case TYPE_FST:
8938 case TYPE_SHIFT:
8939 case TYPE_IMUL:
8940 case TYPE_FBR:
8941 case TYPE_MVI: /* fake */
8942 return EV4_IB0;
8944 case TYPE_IST:
8945 case TYPE_MISC:
8946 case TYPE_IBR:
8947 case TYPE_JSR:
8948 case TYPE_CALLPAL:
8949 case TYPE_FCPYS:
8950 case TYPE_FCMOV:
8951 case TYPE_FADD:
8952 case TYPE_FDIV:
8953 case TYPE_FMUL:
8954 case TYPE_ST_C:
8955 case TYPE_MB:
8956 case TYPE_FSQRT: /* fake */
8957 case TYPE_FTOI: /* fake */
8958 case TYPE_ITOF: /* fake */
8959 return EV4_IB1;
8961 default:
8962 gcc_unreachable ();
8966 static enum alphaev5_pipe
8967 alphaev5_insn_pipe (rtx insn)
8969 if (recog_memoized (insn) < 0)
8970 return EV5_STOP;
8971 if (get_attr_length (insn) != 4)
8972 return EV5_STOP;
8974 switch (get_attr_type (insn))
8976 case TYPE_ILD:
8977 case TYPE_FLD:
8978 case TYPE_LDSYM:
8979 case TYPE_IADD:
8980 case TYPE_ILOG:
8981 case TYPE_ICMOV:
8982 case TYPE_ICMP:
8983 return EV5_E01;
8985 case TYPE_IST:
8986 case TYPE_FST:
8987 case TYPE_SHIFT:
8988 case TYPE_IMUL:
8989 case TYPE_MISC:
8990 case TYPE_MVI:
8991 case TYPE_LD_L:
8992 case TYPE_ST_C:
8993 case TYPE_MB:
8994 case TYPE_FTOI: /* fake */
8995 case TYPE_ITOF: /* fake */
8996 return EV5_E0;
8998 case TYPE_IBR:
8999 case TYPE_JSR:
9000 case TYPE_CALLPAL:
9001 return EV5_E1;
9003 case TYPE_FCPYS:
9004 return EV5_FAM;
9006 case TYPE_FBR:
9007 case TYPE_FCMOV:
9008 case TYPE_FADD:
9009 case TYPE_FDIV:
9010 case TYPE_FSQRT: /* fake */
9011 return EV5_FA;
9013 case TYPE_FMUL:
9014 return EV5_FM;
9016 default:
9017 gcc_unreachable ();
9021 /* IN_USE is a mask of the slots currently filled within the insn group.
9022 The mask bits come from alphaev4_pipe above. If EV4_IBX is set, then
9023 the insn in EV4_IB0 can be swapped by the hardware into EV4_IB1.
9025 LEN is, of course, the length of the group in bytes. */
9027 static rtx
9028 alphaev4_next_group (rtx insn, int *pin_use, int *plen)
9030 int len, in_use;
9032 len = in_use = 0;
9034 if (! INSN_P (insn)
9035 || GET_CODE (PATTERN (insn)) == CLOBBER
9036 || GET_CODE (PATTERN (insn)) == USE)
9037 goto next_and_done;
9039 while (1)
9041 enum alphaev4_pipe pipe;
9043 pipe = alphaev4_insn_pipe (insn);
9044 switch (pipe)
9046 case EV4_STOP:
9047 /* Force complex instructions to start new groups. */
9048 if (in_use)
9049 goto done;
9051 /* If this is a completely unrecognized insn, it's an asm.
9052 We don't know how long it is, so record length as -1 to
9053 signal a needed realignment. */
9054 if (recog_memoized (insn) < 0)
9055 len = -1;
9056 else
9057 len = get_attr_length (insn);
9058 goto next_and_done;
9060 case EV4_IBX:
9061 if (in_use & EV4_IB0)
9063 if (in_use & EV4_IB1)
9064 goto done;
9065 in_use |= EV4_IB1;
9067 else
9068 in_use |= EV4_IB0 | EV4_IBX;
9069 break;
9071 case EV4_IB0:
9072 if (in_use & EV4_IB0)
9074 if (!(in_use & EV4_IBX) || (in_use & EV4_IB1))
9075 goto done;
9076 in_use |= EV4_IB1;
9078 in_use |= EV4_IB0;
9079 break;
9081 case EV4_IB1:
9082 if (in_use & EV4_IB1)
9083 goto done;
9084 in_use |= EV4_IB1;
9085 break;
9087 default:
9088 gcc_unreachable ();
9090 len += 4;
9092 /* Haifa doesn't do well scheduling branches. */
9093 if (JUMP_P (insn))
9094 goto next_and_done;
9096 next:
9097 insn = next_nonnote_insn (insn);
9099 if (!insn || ! INSN_P (insn))
9100 goto done;
9102 /* Let Haifa tell us where it thinks insn group boundaries are. */
9103 if (GET_MODE (insn) == TImode)
9104 goto done;
9106 if (GET_CODE (insn) == CLOBBER || GET_CODE (insn) == USE)
9107 goto next;
9110 next_and_done:
9111 insn = next_nonnote_insn (insn);
9113 done:
9114 *plen = len;
9115 *pin_use = in_use;
9116 return insn;
9119 /* IN_USE is a mask of the slots currently filled within the insn group.
9120 The mask bits come from alphaev5_pipe above. If EV5_E01 is set, then
9121 the insn in EV5_E0 can be swapped by the hardware into EV5_E1.
9123 LEN is, of course, the length of the group in bytes. */
9125 static rtx
9126 alphaev5_next_group (rtx insn, int *pin_use, int *plen)
9128 int len, in_use;
9130 len = in_use = 0;
9132 if (! INSN_P (insn)
9133 || GET_CODE (PATTERN (insn)) == CLOBBER
9134 || GET_CODE (PATTERN (insn)) == USE)
9135 goto next_and_done;
9137 while (1)
9139 enum alphaev5_pipe pipe;
9141 pipe = alphaev5_insn_pipe (insn);
9142 switch (pipe)
9144 case EV5_STOP:
9145 /* Force complex instructions to start new groups. */
9146 if (in_use)
9147 goto done;
9149 /* If this is a completely unrecognized insn, it's an asm.
9150 We don't know how long it is, so record length as -1 to
9151 signal a needed realignment. */
9152 if (recog_memoized (insn) < 0)
9153 len = -1;
9154 else
9155 len = get_attr_length (insn);
9156 goto next_and_done;
9158 /* ??? Most of the places below, we would like to assert never
9159 happen, as it would indicate an error either in Haifa, or
9160 in the scheduling description. Unfortunately, Haifa never
9161 schedules the last instruction of the BB, so we don't have
9162 an accurate TI bit to go off. */
9163 case EV5_E01:
9164 if (in_use & EV5_E0)
9166 if (in_use & EV5_E1)
9167 goto done;
9168 in_use |= EV5_E1;
9170 else
9171 in_use |= EV5_E0 | EV5_E01;
9172 break;
9174 case EV5_E0:
9175 if (in_use & EV5_E0)
9177 if (!(in_use & EV5_E01) || (in_use & EV5_E1))
9178 goto done;
9179 in_use |= EV5_E1;
9181 in_use |= EV5_E0;
9182 break;
9184 case EV5_E1:
9185 if (in_use & EV5_E1)
9186 goto done;
9187 in_use |= EV5_E1;
9188 break;
9190 case EV5_FAM:
9191 if (in_use & EV5_FA)
9193 if (in_use & EV5_FM)
9194 goto done;
9195 in_use |= EV5_FM;
9197 else
9198 in_use |= EV5_FA | EV5_FAM;
9199 break;
9201 case EV5_FA:
9202 if (in_use & EV5_FA)
9203 goto done;
9204 in_use |= EV5_FA;
9205 break;
9207 case EV5_FM:
9208 if (in_use & EV5_FM)
9209 goto done;
9210 in_use |= EV5_FM;
9211 break;
9213 case EV5_NONE:
9214 break;
9216 default:
9217 gcc_unreachable ();
9219 len += 4;
9221 /* Haifa doesn't do well scheduling branches. */
9222 /* ??? If this is predicted not-taken, slotting continues, except
9223 that no more IBR, FBR, or JSR insns may be slotted. */
9224 if (JUMP_P (insn))
9225 goto next_and_done;
9227 next:
9228 insn = next_nonnote_insn (insn);
9230 if (!insn || ! INSN_P (insn))
9231 goto done;
9233 /* Let Haifa tell us where it thinks insn group boundaries are. */
9234 if (GET_MODE (insn) == TImode)
9235 goto done;
9237 if (GET_CODE (insn) == CLOBBER || GET_CODE (insn) == USE)
9238 goto next;
9241 next_and_done:
9242 insn = next_nonnote_insn (insn);
9244 done:
9245 *plen = len;
9246 *pin_use = in_use;
9247 return insn;
9250 static rtx
9251 alphaev4_next_nop (int *pin_use)
9253 int in_use = *pin_use;
9254 rtx nop;
9256 if (!(in_use & EV4_IB0))
9258 in_use |= EV4_IB0;
9259 nop = gen_nop ();
9261 else if ((in_use & (EV4_IBX|EV4_IB1)) == EV4_IBX)
9263 in_use |= EV4_IB1;
9264 nop = gen_nop ();
9266 else if (TARGET_FP && !(in_use & EV4_IB1))
9268 in_use |= EV4_IB1;
9269 nop = gen_fnop ();
9271 else
9272 nop = gen_unop ();
9274 *pin_use = in_use;
9275 return nop;
9278 static rtx
9279 alphaev5_next_nop (int *pin_use)
9281 int in_use = *pin_use;
9282 rtx nop;
9284 if (!(in_use & EV5_E1))
9286 in_use |= EV5_E1;
9287 nop = gen_nop ();
9289 else if (TARGET_FP && !(in_use & EV5_FA))
9291 in_use |= EV5_FA;
9292 nop = gen_fnop ();
9294 else if (TARGET_FP && !(in_use & EV5_FM))
9296 in_use |= EV5_FM;
9297 nop = gen_fnop ();
9299 else
9300 nop = gen_unop ();
9302 *pin_use = in_use;
9303 return nop;
9306 /* The instruction group alignment main loop. */
9308 static void
9309 alpha_align_insns_1 (unsigned int max_align,
9310 rtx (*next_group) (rtx, int *, int *),
9311 rtx (*next_nop) (int *))
9313 /* ALIGN is the known alignment for the insn group. */
9314 unsigned int align;
9315 /* OFS is the offset of the current insn in the insn group. */
9316 int ofs;
9317 int prev_in_use, in_use, len, ldgp;
9318 rtx i, next;
9320 /* Let shorten branches care for assigning alignments to code labels. */
9321 shorten_branches (get_insns ());
9323 if (align_functions < 4)
9324 align = 4;
9325 else if ((unsigned int) align_functions < max_align)
9326 align = align_functions;
9327 else
9328 align = max_align;
9330 ofs = prev_in_use = 0;
9331 i = get_insns ();
9332 if (NOTE_P (i))
9333 i = next_nonnote_insn (i);
9335 ldgp = alpha_function_needs_gp ? 8 : 0;
9337 while (i)
9339 next = (*next_group) (i, &in_use, &len);
9341 /* When we see a label, resync alignment etc. */
9342 if (LABEL_P (i))
9344 unsigned int new_align = 1 << label_to_alignment (i);
9346 if (new_align >= align)
9348 align = new_align < max_align ? new_align : max_align;
9349 ofs = 0;
9352 else if (ofs & (new_align-1))
9353 ofs = (ofs | (new_align-1)) + 1;
9354 gcc_assert (!len);
9357 /* Handle complex instructions special. */
9358 else if (in_use == 0)
9360 /* Asms will have length < 0. This is a signal that we have
9361 lost alignment knowledge. Assume, however, that the asm
9362 will not mis-align instructions. */
9363 if (len < 0)
9365 ofs = 0;
9366 align = 4;
9367 len = 0;
9371 /* If the known alignment is smaller than the recognized insn group,
9372 realign the output. */
9373 else if ((int) align < len)
9375 unsigned int new_log_align = len > 8 ? 4 : 3;
9376 rtx prev, where;
9378 where = prev = prev_nonnote_insn (i);
9379 if (!where || !LABEL_P (where))
9380 where = i;
9382 /* Can't realign between a call and its gp reload. */
9383 if (! (TARGET_EXPLICIT_RELOCS
9384 && prev && CALL_P (prev)))
9386 emit_insn_before (gen_realign (GEN_INT (new_log_align)), where);
9387 align = 1 << new_log_align;
9388 ofs = 0;
9392 /* We may not insert padding inside the initial ldgp sequence. */
9393 else if (ldgp > 0)
9394 ldgp -= len;
9396 /* If the group won't fit in the same INT16 as the previous,
9397 we need to add padding to keep the group together. Rather
9398 than simply leaving the insn filling to the assembler, we
9399 can make use of the knowledge of what sorts of instructions
9400 were issued in the previous group to make sure that all of
9401 the added nops are really free. */
9402 else if (ofs + len > (int) align)
9404 int nop_count = (align - ofs) / 4;
9405 rtx where;
9407 /* Insert nops before labels, branches, and calls to truly merge
9408 the execution of the nops with the previous instruction group. */
9409 where = prev_nonnote_insn (i);
9410 if (where)
9412 if (LABEL_P (where))
9414 rtx where2 = prev_nonnote_insn (where);
9415 if (where2 && JUMP_P (where2))
9416 where = where2;
9418 else if (NONJUMP_INSN_P (where))
9419 where = i;
9421 else
9422 where = i;
9425 emit_insn_before ((*next_nop)(&prev_in_use), where);
9426 while (--nop_count);
9427 ofs = 0;
9430 ofs = (ofs + len) & (align - 1);
9431 prev_in_use = in_use;
9432 i = next;
9436 static void
9437 alpha_align_insns (void)
9439 if (alpha_tune == PROCESSOR_EV4)
9440 alpha_align_insns_1 (8, alphaev4_next_group, alphaev4_next_nop);
9441 else if (alpha_tune == PROCESSOR_EV5)
9442 alpha_align_insns_1 (16, alphaev5_next_group, alphaev5_next_nop);
9443 else
9444 gcc_unreachable ();
9447 /* Insert an unop between sibcall or noreturn function call and GP load. */
9449 static void
9450 alpha_pad_function_end (void)
9452 rtx insn, next;
9454 for (insn = get_insns (); insn; insn = NEXT_INSN (insn))
9456 if (!CALL_P (insn)
9457 || !(SIBLING_CALL_P (insn)
9458 || find_reg_note (insn, REG_NORETURN, NULL_RTX)))
9459 continue;
9461 /* Make sure we do not split a call and its corresponding
9462 CALL_ARG_LOCATION note. */
9463 next = NEXT_INSN (insn);
9464 if (next == NULL)
9465 continue;
9466 if (NOTE_P (next) && NOTE_KIND (next) == NOTE_INSN_CALL_ARG_LOCATION)
9467 insn = next;
9469 next = next_active_insn (insn);
9470 if (next)
9472 rtx pat = PATTERN (next);
9474 if (GET_CODE (pat) == SET
9475 && GET_CODE (SET_SRC (pat)) == UNSPEC_VOLATILE
9476 && XINT (SET_SRC (pat), 1) == UNSPECV_LDGP1)
9477 emit_insn_after (gen_unop (), insn);
9482 /* Machine dependent reorg pass. */
9484 static void
9485 alpha_reorg (void)
9487 /* Workaround for a linker error that triggers when an exception
9488 handler immediatelly follows a sibcall or a noreturn function.
9490 In the sibcall case:
9492 The instruction stream from an object file:
9494 1d8: 00 00 fb 6b jmp (t12)
9495 1dc: 00 00 ba 27 ldah gp,0(ra)
9496 1e0: 00 00 bd 23 lda gp,0(gp)
9497 1e4: 00 00 7d a7 ldq t12,0(gp)
9498 1e8: 00 40 5b 6b jsr ra,(t12),1ec <__funcZ+0x1ec>
9500 was converted in the final link pass to:
9502 12003aa88: 67 fa ff c3 br 120039428 <...>
9503 12003aa8c: 00 00 fe 2f unop
9504 12003aa90: 00 00 fe 2f unop
9505 12003aa94: 48 83 7d a7 ldq t12,-31928(gp)
9506 12003aa98: 00 40 5b 6b jsr ra,(t12),12003aa9c <__func+0x1ec>
9508 And in the noreturn case:
9510 The instruction stream from an object file:
9512 54: 00 40 5b 6b jsr ra,(t12),58 <__func+0x58>
9513 58: 00 00 ba 27 ldah gp,0(ra)
9514 5c: 00 00 bd 23 lda gp,0(gp)
9515 60: 00 00 7d a7 ldq t12,0(gp)
9516 64: 00 40 5b 6b jsr ra,(t12),68 <__func+0x68>
9518 was converted in the final link pass to:
9520 fdb24: a0 03 40 d3 bsr ra,fe9a8 <_called_func+0x8>
9521 fdb28: 00 00 fe 2f unop
9522 fdb2c: 00 00 fe 2f unop
9523 fdb30: 30 82 7d a7 ldq t12,-32208(gp)
9524 fdb34: 00 40 5b 6b jsr ra,(t12),fdb38 <__func+0x68>
9526 GP load instructions were wrongly cleared by the linker relaxation
9527 pass. This workaround prevents removal of GP loads by inserting
9528 an unop instruction between a sibcall or noreturn function call and
9529 exception handler prologue. */
9531 if (current_function_has_exception_handlers ())
9532 alpha_pad_function_end ();
9535 static void
9536 alpha_file_start (void)
9538 default_file_start ();
9540 fputs ("\t.set noreorder\n", asm_out_file);
9541 fputs ("\t.set volatile\n", asm_out_file);
9542 if (TARGET_ABI_OSF)
9543 fputs ("\t.set noat\n", asm_out_file);
9544 if (TARGET_EXPLICIT_RELOCS)
9545 fputs ("\t.set nomacro\n", asm_out_file);
9546 if (TARGET_SUPPORT_ARCH | TARGET_BWX | TARGET_MAX | TARGET_FIX | TARGET_CIX)
9548 const char *arch;
9550 if (alpha_cpu == PROCESSOR_EV6 || TARGET_FIX || TARGET_CIX)
9551 arch = "ev6";
9552 else if (TARGET_MAX)
9553 arch = "pca56";
9554 else if (TARGET_BWX)
9555 arch = "ev56";
9556 else if (alpha_cpu == PROCESSOR_EV5)
9557 arch = "ev5";
9558 else
9559 arch = "ev4";
9561 fprintf (asm_out_file, "\t.arch %s\n", arch);
9565 /* Since we don't have a .dynbss section, we should not allow global
9566 relocations in the .rodata section. */
9568 static int
9569 alpha_elf_reloc_rw_mask (void)
9571 return flag_pic ? 3 : 2;
9574 /* Return a section for X. The only special thing we do here is to
9575 honor small data. */
9577 static section *
9578 alpha_elf_select_rtx_section (enum machine_mode mode, rtx x,
9579 unsigned HOST_WIDE_INT align)
9581 if (TARGET_SMALL_DATA && GET_MODE_SIZE (mode) <= g_switch_value)
9582 /* ??? Consider using mergeable sdata sections. */
9583 return sdata_section;
9584 else
9585 return default_elf_select_rtx_section (mode, x, align);
9588 static unsigned int
9589 alpha_elf_section_type_flags (tree decl, const char *name, int reloc)
9591 unsigned int flags = 0;
9593 if (strcmp (name, ".sdata") == 0
9594 || strncmp (name, ".sdata.", 7) == 0
9595 || strncmp (name, ".gnu.linkonce.s.", 16) == 0
9596 || strcmp (name, ".sbss") == 0
9597 || strncmp (name, ".sbss.", 6) == 0
9598 || strncmp (name, ".gnu.linkonce.sb.", 17) == 0)
9599 flags = SECTION_SMALL;
9601 flags |= default_section_type_flags (decl, name, reloc);
9602 return flags;
9605 /* Structure to collect function names for final output in link section. */
9606 /* Note that items marked with GTY can't be ifdef'ed out. */
9608 enum reloc_kind
9610 KIND_LINKAGE,
9611 KIND_CODEADDR
9614 struct GTY(()) alpha_links
9616 rtx func;
9617 rtx linkage;
9618 enum reloc_kind rkind;
9621 #if TARGET_ABI_OPEN_VMS
9623 /* Return the VMS argument type corresponding to MODE. */
9625 enum avms_arg_type
9626 alpha_arg_type (enum machine_mode mode)
9628 switch (mode)
9630 case SFmode:
9631 return TARGET_FLOAT_VAX ? FF : FS;
9632 case DFmode:
9633 return TARGET_FLOAT_VAX ? FD : FT;
9634 default:
9635 return I64;
9639 /* Return an rtx for an integer representing the VMS Argument Information
9640 register value. */
9643 alpha_arg_info_reg_val (CUMULATIVE_ARGS cum)
9645 unsigned HOST_WIDE_INT regval = cum.num_args;
9646 int i;
9648 for (i = 0; i < 6; i++)
9649 regval |= ((int) cum.atypes[i]) << (i * 3 + 8);
9651 return GEN_INT (regval);
9655 /* Return a SYMBOL_REF representing the reference to the .linkage entry
9656 of function FUNC built for calls made from CFUNDECL. LFLAG is 1 if
9657 this is the reference to the linkage pointer value, 0 if this is the
9658 reference to the function entry value. RFLAG is 1 if this a reduced
9659 reference (code address only), 0 if this is a full reference. */
9662 alpha_use_linkage (rtx func, bool lflag, bool rflag)
9664 struct alpha_links *al = NULL;
9665 const char *name = XSTR (func, 0);
9667 if (cfun->machine->links)
9669 splay_tree_node lnode;
9671 /* Is this name already defined? */
9672 lnode = splay_tree_lookup (cfun->machine->links, (splay_tree_key) name);
9673 if (lnode)
9674 al = (struct alpha_links *) lnode->value;
9676 else
9677 cfun->machine->links = splay_tree_new_ggc
9678 ((splay_tree_compare_fn) strcmp,
9679 ggc_alloc_splay_tree_str_alpha_links_splay_tree_s,
9680 ggc_alloc_splay_tree_str_alpha_links_splay_tree_node_s);
9682 if (al == NULL)
9684 size_t buf_len;
9685 char *linksym;
9686 tree id;
9688 if (name[0] == '*')
9689 name++;
9691 /* Follow transparent alias, as this is used for CRTL translations. */
9692 id = maybe_get_identifier (name);
9693 if (id)
9695 while (IDENTIFIER_TRANSPARENT_ALIAS (id))
9696 id = TREE_CHAIN (id);
9697 name = IDENTIFIER_POINTER (id);
9700 buf_len = strlen (name) + 8 + 9;
9701 linksym = (char *) alloca (buf_len);
9702 snprintf (linksym, buf_len, "$%d..%s..lk", cfun->funcdef_no, name);
9704 al = ggc_alloc<alpha_links> ();
9705 al->func = func;
9706 al->linkage = gen_rtx_SYMBOL_REF (Pmode, ggc_strdup (linksym));
9708 splay_tree_insert (cfun->machine->links,
9709 (splay_tree_key) ggc_strdup (name),
9710 (splay_tree_value) al);
9713 al->rkind = rflag ? KIND_CODEADDR : KIND_LINKAGE;
9715 if (lflag)
9716 return gen_rtx_MEM (Pmode, plus_constant (Pmode, al->linkage, 8));
9717 else
9718 return al->linkage;
9721 static int
9722 alpha_write_one_linkage (splay_tree_node node, void *data)
9724 const char *const name = (const char *) node->key;
9725 struct alpha_links *link = (struct alpha_links *) node->value;
9726 FILE *stream = (FILE *) data;
9728 ASM_OUTPUT_INTERNAL_LABEL (stream, XSTR (link->linkage, 0));
9729 if (link->rkind == KIND_CODEADDR)
9731 /* External and used, request code address. */
9732 fprintf (stream, "\t.code_address ");
9734 else
9736 if (!SYMBOL_REF_EXTERNAL_P (link->func)
9737 && SYMBOL_REF_LOCAL_P (link->func))
9739 /* Locally defined, build linkage pair. */
9740 fprintf (stream, "\t.quad %s..en\n", name);
9741 fprintf (stream, "\t.quad ");
9743 else
9745 /* External, request linkage pair. */
9746 fprintf (stream, "\t.linkage ");
9749 assemble_name (stream, name);
9750 fputs ("\n", stream);
9752 return 0;
9755 static void
9756 alpha_write_linkage (FILE *stream, const char *funname)
9758 fprintf (stream, "\t.link\n");
9759 fprintf (stream, "\t.align 3\n");
9760 in_section = NULL;
9762 #ifdef TARGET_VMS_CRASH_DEBUG
9763 fputs ("\t.name ", stream);
9764 assemble_name (stream, funname);
9765 fputs ("..na\n", stream);
9766 #endif
9768 ASM_OUTPUT_LABEL (stream, funname);
9769 fprintf (stream, "\t.pdesc ");
9770 assemble_name (stream, funname);
9771 fprintf (stream, "..en,%s\n",
9772 alpha_procedure_type == PT_STACK ? "stack"
9773 : alpha_procedure_type == PT_REGISTER ? "reg" : "null");
9775 if (cfun->machine->links)
9777 splay_tree_foreach (cfun->machine->links, alpha_write_one_linkage, stream);
9778 /* splay_tree_delete (func->links); */
9782 /* Switch to an arbitrary section NAME with attributes as specified
9783 by FLAGS. ALIGN specifies any known alignment requirements for
9784 the section; 0 if the default should be used. */
9786 static void
9787 vms_asm_named_section (const char *name, unsigned int flags,
9788 tree decl ATTRIBUTE_UNUSED)
9790 fputc ('\n', asm_out_file);
9791 fprintf (asm_out_file, ".section\t%s", name);
9793 if (flags & SECTION_DEBUG)
9794 fprintf (asm_out_file, ",NOWRT");
9796 fputc ('\n', asm_out_file);
9799 /* Record an element in the table of global constructors. SYMBOL is
9800 a SYMBOL_REF of the function to be called; PRIORITY is a number
9801 between 0 and MAX_INIT_PRIORITY.
9803 Differs from default_ctors_section_asm_out_constructor in that the
9804 width of the .ctors entry is always 64 bits, rather than the 32 bits
9805 used by a normal pointer. */
9807 static void
9808 vms_asm_out_constructor (rtx symbol, int priority ATTRIBUTE_UNUSED)
9810 switch_to_section (ctors_section);
9811 assemble_align (BITS_PER_WORD);
9812 assemble_integer (symbol, UNITS_PER_WORD, BITS_PER_WORD, 1);
9815 static void
9816 vms_asm_out_destructor (rtx symbol, int priority ATTRIBUTE_UNUSED)
9818 switch_to_section (dtors_section);
9819 assemble_align (BITS_PER_WORD);
9820 assemble_integer (symbol, UNITS_PER_WORD, BITS_PER_WORD, 1);
9822 #else
9824 alpha_use_linkage (rtx func ATTRIBUTE_UNUSED,
9825 bool lflag ATTRIBUTE_UNUSED,
9826 bool rflag ATTRIBUTE_UNUSED)
9828 return NULL_RTX;
9831 #endif /* TARGET_ABI_OPEN_VMS */
9833 static void
9834 alpha_init_libfuncs (void)
9836 if (TARGET_ABI_OPEN_VMS)
9838 /* Use the VMS runtime library functions for division and
9839 remainder. */
9840 set_optab_libfunc (sdiv_optab, SImode, "OTS$DIV_I");
9841 set_optab_libfunc (sdiv_optab, DImode, "OTS$DIV_L");
9842 set_optab_libfunc (udiv_optab, SImode, "OTS$DIV_UI");
9843 set_optab_libfunc (udiv_optab, DImode, "OTS$DIV_UL");
9844 set_optab_libfunc (smod_optab, SImode, "OTS$REM_I");
9845 set_optab_libfunc (smod_optab, DImode, "OTS$REM_L");
9846 set_optab_libfunc (umod_optab, SImode, "OTS$REM_UI");
9847 set_optab_libfunc (umod_optab, DImode, "OTS$REM_UL");
9848 abort_libfunc = init_one_libfunc ("decc$abort");
9849 memcmp_libfunc = init_one_libfunc ("decc$memcmp");
9850 #ifdef MEM_LIBFUNCS_INIT
9851 MEM_LIBFUNCS_INIT;
9852 #endif
9856 /* On the Alpha, we use this to disable the floating-point registers
9857 when they don't exist. */
9859 static void
9860 alpha_conditional_register_usage (void)
9862 int i;
9863 if (! TARGET_FPREGS)
9864 for (i = 32; i < 63; i++)
9865 fixed_regs[i] = call_used_regs[i] = 1;
9868 /* Canonicalize a comparison from one we don't have to one we do have. */
9870 static void
9871 alpha_canonicalize_comparison (int *code, rtx *op0, rtx *op1,
9872 bool op0_preserve_value)
9874 if (!op0_preserve_value
9875 && (*code == GE || *code == GT || *code == GEU || *code == GTU)
9876 && (REG_P (*op1) || *op1 == const0_rtx))
9878 rtx tem = *op0;
9879 *op0 = *op1;
9880 *op1 = tem;
9881 *code = (int)swap_condition ((enum rtx_code)*code);
9884 if ((*code == LT || *code == LTU)
9885 && CONST_INT_P (*op1) && INTVAL (*op1) == 256)
9887 *code = *code == LT ? LE : LEU;
9888 *op1 = GEN_INT (255);
9892 /* Implement TARGET_ATOMIC_ASSIGN_EXPAND_FENV. */
9894 static void
9895 alpha_atomic_assign_expand_fenv (tree *hold, tree *clear, tree *update)
9897 const unsigned HOST_WIDE_INT SWCR_STATUS_MASK = (0x3fUL << 17);
9899 tree fenv_var, get_fpscr, set_fpscr, mask, ld_fenv, masked_fenv;
9900 tree new_fenv_var, reload_fenv, restore_fnenv;
9901 tree update_call, atomic_feraiseexcept, hold_fnclex;
9903 /* Assume OSF/1 compatible interfaces. */
9904 if (!TARGET_ABI_OSF)
9905 return;
9907 /* Generate the equivalent of :
9908 unsigned long fenv_var;
9909 fenv_var = __ieee_get_fp_control ();
9911 unsigned long masked_fenv;
9912 masked_fenv = fenv_var & mask;
9914 __ieee_set_fp_control (masked_fenv); */
9916 fenv_var = create_tmp_var (long_unsigned_type_node, NULL);
9917 get_fpscr
9918 = build_fn_decl ("__ieee_get_fp_control",
9919 build_function_type_list (long_unsigned_type_node, NULL));
9920 set_fpscr
9921 = build_fn_decl ("__ieee_set_fp_control",
9922 build_function_type_list (void_type_node, NULL));
9923 mask = build_int_cst (long_unsigned_type_node, ~SWCR_STATUS_MASK);
9924 ld_fenv = build2 (MODIFY_EXPR, long_unsigned_type_node,
9925 fenv_var, build_call_expr (get_fpscr, 0));
9926 masked_fenv = build2 (BIT_AND_EXPR, long_unsigned_type_node, fenv_var, mask);
9927 hold_fnclex = build_call_expr (set_fpscr, 1, masked_fenv);
9928 *hold = build2 (COMPOUND_EXPR, void_type_node,
9929 build2 (COMPOUND_EXPR, void_type_node, masked_fenv, ld_fenv),
9930 hold_fnclex);
9932 /* Store the value of masked_fenv to clear the exceptions:
9933 __ieee_set_fp_control (masked_fenv); */
9935 *clear = build_call_expr (set_fpscr, 1, masked_fenv);
9937 /* Generate the equivalent of :
9938 unsigned long new_fenv_var;
9939 new_fenv_var = __ieee_get_fp_control ();
9941 __ieee_set_fp_control (fenv_var);
9943 __atomic_feraiseexcept (new_fenv_var); */
9945 new_fenv_var = create_tmp_var (long_unsigned_type_node, NULL);
9946 reload_fenv = build2 (MODIFY_EXPR, long_unsigned_type_node, new_fenv_var,
9947 build_call_expr (get_fpscr, 0));
9948 restore_fnenv = build_call_expr (set_fpscr, 1, fenv_var);
9949 atomic_feraiseexcept = builtin_decl_implicit (BUILT_IN_ATOMIC_FERAISEEXCEPT);
9950 update_call
9951 = build_call_expr (atomic_feraiseexcept, 1,
9952 fold_convert (integer_type_node, new_fenv_var));
9953 *update = build2 (COMPOUND_EXPR, void_type_node,
9954 build2 (COMPOUND_EXPR, void_type_node,
9955 reload_fenv, restore_fnenv), update_call);
9958 /* Initialize the GCC target structure. */
9959 #if TARGET_ABI_OPEN_VMS
9960 # undef TARGET_ATTRIBUTE_TABLE
9961 # define TARGET_ATTRIBUTE_TABLE vms_attribute_table
9962 # undef TARGET_CAN_ELIMINATE
9963 # define TARGET_CAN_ELIMINATE alpha_vms_can_eliminate
9964 #endif
9966 #undef TARGET_IN_SMALL_DATA_P
9967 #define TARGET_IN_SMALL_DATA_P alpha_in_small_data_p
9969 #undef TARGET_ASM_ALIGNED_HI_OP
9970 #define TARGET_ASM_ALIGNED_HI_OP "\t.word\t"
9971 #undef TARGET_ASM_ALIGNED_DI_OP
9972 #define TARGET_ASM_ALIGNED_DI_OP "\t.quad\t"
9974 /* Default unaligned ops are provided for ELF systems. To get unaligned
9975 data for non-ELF systems, we have to turn off auto alignment. */
9976 #if TARGET_ABI_OPEN_VMS
9977 #undef TARGET_ASM_UNALIGNED_HI_OP
9978 #define TARGET_ASM_UNALIGNED_HI_OP "\t.align 0\n\t.word\t"
9979 #undef TARGET_ASM_UNALIGNED_SI_OP
9980 #define TARGET_ASM_UNALIGNED_SI_OP "\t.align 0\n\t.long\t"
9981 #undef TARGET_ASM_UNALIGNED_DI_OP
9982 #define TARGET_ASM_UNALIGNED_DI_OP "\t.align 0\n\t.quad\t"
9983 #endif
9985 #undef TARGET_ASM_RELOC_RW_MASK
9986 #define TARGET_ASM_RELOC_RW_MASK alpha_elf_reloc_rw_mask
9987 #undef TARGET_ASM_SELECT_RTX_SECTION
9988 #define TARGET_ASM_SELECT_RTX_SECTION alpha_elf_select_rtx_section
9989 #undef TARGET_SECTION_TYPE_FLAGS
9990 #define TARGET_SECTION_TYPE_FLAGS alpha_elf_section_type_flags
9992 #undef TARGET_ASM_FUNCTION_END_PROLOGUE
9993 #define TARGET_ASM_FUNCTION_END_PROLOGUE alpha_output_function_end_prologue
9995 #undef TARGET_INIT_LIBFUNCS
9996 #define TARGET_INIT_LIBFUNCS alpha_init_libfuncs
9998 #undef TARGET_LEGITIMIZE_ADDRESS
9999 #define TARGET_LEGITIMIZE_ADDRESS alpha_legitimize_address
10000 #undef TARGET_MODE_DEPENDENT_ADDRESS_P
10001 #define TARGET_MODE_DEPENDENT_ADDRESS_P alpha_mode_dependent_address_p
10003 #undef TARGET_ASM_FILE_START
10004 #define TARGET_ASM_FILE_START alpha_file_start
10006 #undef TARGET_SCHED_ADJUST_COST
10007 #define TARGET_SCHED_ADJUST_COST alpha_adjust_cost
10008 #undef TARGET_SCHED_ISSUE_RATE
10009 #define TARGET_SCHED_ISSUE_RATE alpha_issue_rate
10010 #undef TARGET_SCHED_FIRST_CYCLE_MULTIPASS_DFA_LOOKAHEAD
10011 #define TARGET_SCHED_FIRST_CYCLE_MULTIPASS_DFA_LOOKAHEAD \
10012 alpha_multipass_dfa_lookahead
10014 #undef TARGET_HAVE_TLS
10015 #define TARGET_HAVE_TLS HAVE_AS_TLS
10017 #undef TARGET_BUILTIN_DECL
10018 #define TARGET_BUILTIN_DECL alpha_builtin_decl
10019 #undef TARGET_INIT_BUILTINS
10020 #define TARGET_INIT_BUILTINS alpha_init_builtins
10021 #undef TARGET_EXPAND_BUILTIN
10022 #define TARGET_EXPAND_BUILTIN alpha_expand_builtin
10023 #undef TARGET_FOLD_BUILTIN
10024 #define TARGET_FOLD_BUILTIN alpha_fold_builtin
10025 #undef TARGET_GIMPLE_FOLD_BUILTIN
10026 #define TARGET_GIMPLE_FOLD_BUILTIN alpha_gimple_fold_builtin
10028 #undef TARGET_FUNCTION_OK_FOR_SIBCALL
10029 #define TARGET_FUNCTION_OK_FOR_SIBCALL alpha_function_ok_for_sibcall
10030 #undef TARGET_CANNOT_COPY_INSN_P
10031 #define TARGET_CANNOT_COPY_INSN_P alpha_cannot_copy_insn_p
10032 #undef TARGET_LEGITIMATE_CONSTANT_P
10033 #define TARGET_LEGITIMATE_CONSTANT_P alpha_legitimate_constant_p
10034 #undef TARGET_CANNOT_FORCE_CONST_MEM
10035 #define TARGET_CANNOT_FORCE_CONST_MEM alpha_cannot_force_const_mem
10037 #if TARGET_ABI_OSF
10038 #undef TARGET_ASM_OUTPUT_MI_THUNK
10039 #define TARGET_ASM_OUTPUT_MI_THUNK alpha_output_mi_thunk_osf
10040 #undef TARGET_ASM_CAN_OUTPUT_MI_THUNK
10041 #define TARGET_ASM_CAN_OUTPUT_MI_THUNK hook_bool_const_tree_hwi_hwi_const_tree_true
10042 #undef TARGET_STDARG_OPTIMIZE_HOOK
10043 #define TARGET_STDARG_OPTIMIZE_HOOK alpha_stdarg_optimize_hook
10044 #endif
10046 /* Use 16-bits anchor. */
10047 #undef TARGET_MIN_ANCHOR_OFFSET
10048 #define TARGET_MIN_ANCHOR_OFFSET -0x7fff - 1
10049 #undef TARGET_MAX_ANCHOR_OFFSET
10050 #define TARGET_MAX_ANCHOR_OFFSET 0x7fff
10051 #undef TARGET_USE_BLOCKS_FOR_CONSTANT_P
10052 #define TARGET_USE_BLOCKS_FOR_CONSTANT_P hook_bool_mode_const_rtx_true
10054 #undef TARGET_RTX_COSTS
10055 #define TARGET_RTX_COSTS alpha_rtx_costs
10056 #undef TARGET_ADDRESS_COST
10057 #define TARGET_ADDRESS_COST hook_int_rtx_mode_as_bool_0
10059 #undef TARGET_MACHINE_DEPENDENT_REORG
10060 #define TARGET_MACHINE_DEPENDENT_REORG alpha_reorg
10062 #undef TARGET_PROMOTE_FUNCTION_MODE
10063 #define TARGET_PROMOTE_FUNCTION_MODE default_promote_function_mode_always_promote
10064 #undef TARGET_PROMOTE_PROTOTYPES
10065 #define TARGET_PROMOTE_PROTOTYPES hook_bool_const_tree_false
10066 #undef TARGET_RETURN_IN_MEMORY
10067 #define TARGET_RETURN_IN_MEMORY alpha_return_in_memory
10068 #undef TARGET_PASS_BY_REFERENCE
10069 #define TARGET_PASS_BY_REFERENCE alpha_pass_by_reference
10070 #undef TARGET_SETUP_INCOMING_VARARGS
10071 #define TARGET_SETUP_INCOMING_VARARGS alpha_setup_incoming_varargs
10072 #undef TARGET_STRICT_ARGUMENT_NAMING
10073 #define TARGET_STRICT_ARGUMENT_NAMING hook_bool_CUMULATIVE_ARGS_true
10074 #undef TARGET_PRETEND_OUTGOING_VARARGS_NAMED
10075 #define TARGET_PRETEND_OUTGOING_VARARGS_NAMED hook_bool_CUMULATIVE_ARGS_true
10076 #undef TARGET_SPLIT_COMPLEX_ARG
10077 #define TARGET_SPLIT_COMPLEX_ARG alpha_split_complex_arg
10078 #undef TARGET_GIMPLIFY_VA_ARG_EXPR
10079 #define TARGET_GIMPLIFY_VA_ARG_EXPR alpha_gimplify_va_arg
10080 #undef TARGET_ARG_PARTIAL_BYTES
10081 #define TARGET_ARG_PARTIAL_BYTES alpha_arg_partial_bytes
10082 #undef TARGET_FUNCTION_ARG
10083 #define TARGET_FUNCTION_ARG alpha_function_arg
10084 #undef TARGET_FUNCTION_ARG_ADVANCE
10085 #define TARGET_FUNCTION_ARG_ADVANCE alpha_function_arg_advance
10086 #undef TARGET_TRAMPOLINE_INIT
10087 #define TARGET_TRAMPOLINE_INIT alpha_trampoline_init
10089 #undef TARGET_INSTANTIATE_DECLS
10090 #define TARGET_INSTANTIATE_DECLS alpha_instantiate_decls
10092 #undef TARGET_SECONDARY_RELOAD
10093 #define TARGET_SECONDARY_RELOAD alpha_secondary_reload
10095 #undef TARGET_SCALAR_MODE_SUPPORTED_P
10096 #define TARGET_SCALAR_MODE_SUPPORTED_P alpha_scalar_mode_supported_p
10097 #undef TARGET_VECTOR_MODE_SUPPORTED_P
10098 #define TARGET_VECTOR_MODE_SUPPORTED_P alpha_vector_mode_supported_p
10100 #undef TARGET_BUILD_BUILTIN_VA_LIST
10101 #define TARGET_BUILD_BUILTIN_VA_LIST alpha_build_builtin_va_list
10103 #undef TARGET_EXPAND_BUILTIN_VA_START
10104 #define TARGET_EXPAND_BUILTIN_VA_START alpha_va_start
10106 /* The Alpha architecture does not require sequential consistency. See
10107 http://www.cs.umd.edu/~pugh/java/memoryModel/AlphaReordering.html
10108 for an example of how it can be violated in practice. */
10109 #undef TARGET_RELAXED_ORDERING
10110 #define TARGET_RELAXED_ORDERING true
10112 #undef TARGET_OPTION_OVERRIDE
10113 #define TARGET_OPTION_OVERRIDE alpha_option_override
10115 #ifdef TARGET_ALTERNATE_LONG_DOUBLE_MANGLING
10116 #undef TARGET_MANGLE_TYPE
10117 #define TARGET_MANGLE_TYPE alpha_mangle_type
10118 #endif
10120 #undef TARGET_LEGITIMATE_ADDRESS_P
10121 #define TARGET_LEGITIMATE_ADDRESS_P alpha_legitimate_address_p
10123 #undef TARGET_CONDITIONAL_REGISTER_USAGE
10124 #define TARGET_CONDITIONAL_REGISTER_USAGE alpha_conditional_register_usage
10126 #undef TARGET_CANONICALIZE_COMPARISON
10127 #define TARGET_CANONICALIZE_COMPARISON alpha_canonicalize_comparison
10129 #undef TARGET_ATOMIC_ASSIGN_EXPAND_FENV
10130 #define TARGET_ATOMIC_ASSIGN_EXPAND_FENV alpha_atomic_assign_expand_fenv
10132 struct gcc_target targetm = TARGET_INITIALIZER;
10135 #include "gt-alpha.h"