Add configuration for semi-hosted ARM.
[official-gcc.git] / gcc / integrate.c
blob9ab7adf46354bca3b7d11ab87272e21335f785b3
1 /* Procedure integration for GNU CC.
2 Copyright (C) 1988, 1991, 1993, 1994, 1995 Free Software Foundation, Inc.
3 Contributed by Michael Tiemann (tiemann@cygnus.com)
5 This file is part of GNU CC.
7 GNU CC is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License as published by
9 the Free Software Foundation; either version 2, or (at your option)
10 any later version.
12 GNU CC is distributed in the hope that it will be useful,
13 but WITHOUT ANY WARRANTY; without even the implied warranty of
14 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 GNU General Public License for more details.
17 You should have received a copy of the GNU General Public License
18 along with GNU CC; see the file COPYING. If not, write to
19 the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA. */
22 #include <stdio.h>
24 #include "config.h"
25 #include "rtl.h"
26 #include "tree.h"
27 #include "flags.h"
28 #include "insn-config.h"
29 #include "insn-flags.h"
30 #include "expr.h"
31 #include "output.h"
32 #include "integrate.h"
33 #include "real.h"
34 #include "function.h"
35 #include "bytecode.h"
37 #include "obstack.h"
38 #define obstack_chunk_alloc xmalloc
39 #define obstack_chunk_free free
41 extern struct obstack *function_maybepermanent_obstack;
43 extern tree pushdecl ();
44 extern tree poplevel ();
46 /* Similar, but round to the next highest integer that meets the
47 alignment. */
48 #define CEIL_ROUND(VALUE,ALIGN) (((VALUE) + (ALIGN) - 1) & ~((ALIGN)- 1))
50 /* Default max number of insns a function can have and still be inline.
51 This is overridden on RISC machines. */
52 #ifndef INTEGRATE_THRESHOLD
53 #define INTEGRATE_THRESHOLD(DECL) \
54 (8 * (8 + list_length (DECL_ARGUMENTS (DECL))))
55 #endif
57 static rtx initialize_for_inline PROTO((tree, int, int, int, int));
58 static void finish_inline PROTO((tree, rtx));
59 static void adjust_copied_decl_tree PROTO((tree));
60 static tree copy_decl_list PROTO((tree));
61 static tree copy_decl_tree PROTO((tree));
62 static void copy_decl_rtls PROTO((tree));
63 static void save_constants PROTO((rtx *));
64 static void note_modified_parmregs PROTO((rtx, rtx));
65 static rtx copy_for_inline PROTO((rtx));
66 static void integrate_parm_decls PROTO((tree, struct inline_remap *, rtvec));
67 static void integrate_decl_tree PROTO((tree, int, struct inline_remap *));
68 static void subst_constants PROTO((rtx *, rtx, struct inline_remap *));
69 static void restore_constants PROTO((rtx *));
70 static void set_block_origin_self PROTO((tree));
71 static void set_decl_origin_self PROTO((tree));
72 static void set_block_abstract_flags PROTO((tree, int));
74 void set_decl_abstract_flags PROTO((tree, int));
76 /* Zero if the current function (whose FUNCTION_DECL is FNDECL)
77 is safe and reasonable to integrate into other functions.
78 Nonzero means value is a warning message with a single %s
79 for the function's name. */
81 char *
82 function_cannot_inline_p (fndecl)
83 register tree fndecl;
85 register rtx insn;
86 tree last = tree_last (TYPE_ARG_TYPES (TREE_TYPE (fndecl)));
87 int max_insns = INTEGRATE_THRESHOLD (fndecl);
88 register int ninsns = 0;
89 register tree parms;
91 /* No inlines with varargs. `grokdeclarator' gives a warning
92 message about that if `inline' is specified. This code
93 it put in to catch the volunteers. */
94 if ((last && TREE_VALUE (last) != void_type_node)
95 || current_function_varargs)
96 return "varargs function cannot be inline";
98 if (current_function_calls_alloca)
99 return "function using alloca cannot be inline";
101 if (current_function_contains_functions)
102 return "function with nested functions cannot be inline";
104 /* If its not even close, don't even look. */
105 if (!DECL_INLINE (fndecl) && get_max_uid () > 3 * max_insns)
106 return "function too large to be inline";
108 #if 0
109 /* Large stacks are OK now that inlined functions can share them. */
110 /* Don't inline functions with large stack usage,
111 since they can make other recursive functions burn up stack. */
112 if (!DECL_INLINE (fndecl) && get_frame_size () > 100)
113 return "function stack frame for inlining";
114 #endif
116 #if 0
117 /* Don't inline functions which do not specify a function prototype and
118 have BLKmode argument or take the address of a parameter. */
119 for (parms = DECL_ARGUMENTS (fndecl); parms; parms = TREE_CHAIN (parms))
121 if (TYPE_MODE (TREE_TYPE (parms)) == BLKmode)
122 TREE_ADDRESSABLE (parms) = 1;
123 if (last == NULL_TREE && TREE_ADDRESSABLE (parms))
124 return "no prototype, and parameter address used; cannot be inline";
126 #endif
128 /* We can't inline functions that return structures
129 the old-fashioned PCC way, copying into a static block. */
130 if (current_function_returns_pcc_struct)
131 return "inline functions not supported for this return value type";
133 /* We can't inline functions that return BLKmode structures in registers. */
134 if (TYPE_MODE (TREE_TYPE (TREE_TYPE (fndecl))) == BLKmode
135 && ! aggregate_value_p (TREE_TYPE (TREE_TYPE (fndecl))))
136 return "inline functions not supported for this return value type";
138 /* We can't inline functions that return structures of varying size. */
139 if (int_size_in_bytes (TREE_TYPE (TREE_TYPE (fndecl))) < 0)
140 return "function with varying-size return value cannot be inline";
142 /* Cannot inline a function with a varying size argument or one that
143 receives a transparent union. */
144 for (parms = DECL_ARGUMENTS (fndecl); parms; parms = TREE_CHAIN (parms))
146 if (int_size_in_bytes (TREE_TYPE (parms)) < 0)
147 return "function with varying-size parameter cannot be inline";
148 else if (TYPE_TRANSPARENT_UNION (TREE_TYPE (parms)))
149 return "function with transparent unit parameter cannot be inline";
152 if (!DECL_INLINE (fndecl) && get_max_uid () > max_insns)
154 for (ninsns = 0, insn = get_first_nonparm_insn (); insn && ninsns < max_insns;
155 insn = NEXT_INSN (insn))
157 if (GET_RTX_CLASS (GET_CODE (insn)) == 'i')
158 ninsns++;
161 if (ninsns >= max_insns)
162 return "function too large to be inline";
165 /* We cannot inline this function if forced_labels is non-zero. This
166 implies that a label in this function was used as an initializer.
167 Because labels can not be duplicated, all labels in the function
168 will be renamed when it is inlined. However, there is no way to find
169 and fix all variables initialized with addresses of labels in this
170 function, hence inlining is impossible. */
172 if (forced_labels)
173 return "function with label addresses used in initializers cannot inline";
175 /* We cannot inline a nested function that jumps to a nonlocal label. */
176 if (current_function_has_nonlocal_goto)
177 return "function with nonlocal goto cannot be inline";
179 return 0;
182 /* Variables used within save_for_inline. */
184 /* Mapping from old pseudo-register to new pseudo-registers.
185 The first element of this map is reg_map[FIRST_PSEUDO_REGISTER].
186 It is allocated in `save_for_inline' and `expand_inline_function',
187 and deallocated on exit from each of those routines. */
188 static rtx *reg_map;
190 /* Mapping from old code-labels to new code-labels.
191 The first element of this map is label_map[min_labelno].
192 It is allocated in `save_for_inline' and `expand_inline_function',
193 and deallocated on exit from each of those routines. */
194 static rtx *label_map;
196 /* Mapping from old insn uid's to copied insns.
197 It is allocated in `save_for_inline' and `expand_inline_function',
198 and deallocated on exit from each of those routines. */
199 static rtx *insn_map;
201 /* Map pseudo reg number into the PARM_DECL for the parm living in the reg.
202 Zero for a reg that isn't a parm's home.
203 Only reg numbers less than max_parm_reg are mapped here. */
204 static tree *parmdecl_map;
206 /* Keep track of first pseudo-register beyond those that are parms. */
207 static int max_parm_reg;
209 /* When an insn is being copied by copy_for_inline,
210 this is nonzero if we have copied an ASM_OPERANDS.
211 In that case, it is the original input-operand vector. */
212 static rtvec orig_asm_operands_vector;
214 /* When an insn is being copied by copy_for_inline,
215 this is nonzero if we have copied an ASM_OPERANDS.
216 In that case, it is the copied input-operand vector. */
217 static rtvec copy_asm_operands_vector;
219 /* Likewise, this is the copied constraints vector. */
220 static rtvec copy_asm_constraints_vector;
222 /* In save_for_inline, nonzero if past the parm-initialization insns. */
223 static int in_nonparm_insns;
225 /* Subroutine for `save_for_inline{copying,nocopy}'. Performs initialization
226 needed to save FNDECL's insns and info for future inline expansion. */
228 static rtx
229 initialize_for_inline (fndecl, min_labelno, max_labelno, max_reg, copy)
230 tree fndecl;
231 int min_labelno;
232 int max_labelno;
233 int max_reg;
234 int copy;
236 int function_flags, i;
237 rtvec arg_vector;
238 tree parms;
240 /* Compute the values of any flags we must restore when inlining this. */
242 function_flags
243 = (current_function_calls_alloca * FUNCTION_FLAGS_CALLS_ALLOCA
244 + current_function_calls_setjmp * FUNCTION_FLAGS_CALLS_SETJMP
245 + current_function_calls_longjmp * FUNCTION_FLAGS_CALLS_LONGJMP
246 + current_function_returns_struct * FUNCTION_FLAGS_RETURNS_STRUCT
247 + current_function_returns_pcc_struct * FUNCTION_FLAGS_RETURNS_PCC_STRUCT
248 + current_function_needs_context * FUNCTION_FLAGS_NEEDS_CONTEXT
249 + current_function_has_nonlocal_label * FUNCTION_FLAGS_HAS_NONLOCAL_LABEL
250 + current_function_returns_pointer * FUNCTION_FLAGS_RETURNS_POINTER
251 + current_function_uses_const_pool * FUNCTION_FLAGS_USES_CONST_POOL
252 + current_function_uses_pic_offset_table * FUNCTION_FLAGS_USES_PIC_OFFSET_TABLE);
254 /* Clear out PARMDECL_MAP. It was allocated in the caller's frame. */
255 bzero ((char *) parmdecl_map, max_parm_reg * sizeof (tree));
256 arg_vector = rtvec_alloc (list_length (DECL_ARGUMENTS (fndecl)));
258 for (parms = DECL_ARGUMENTS (fndecl), i = 0;
259 parms;
260 parms = TREE_CHAIN (parms), i++)
262 rtx p = DECL_RTL (parms);
264 if (GET_CODE (p) == MEM && copy)
266 /* Copy the rtl so that modifications of the addresses
267 later in compilation won't affect this arg_vector.
268 Virtual register instantiation can screw the address
269 of the rtl. */
270 rtx new = copy_rtx (p);
272 /* Don't leave the old copy anywhere in this decl. */
273 if (DECL_RTL (parms) == DECL_INCOMING_RTL (parms)
274 || (GET_CODE (DECL_RTL (parms)) == MEM
275 && GET_CODE (DECL_INCOMING_RTL (parms)) == MEM
276 && (XEXP (DECL_RTL (parms), 0)
277 == XEXP (DECL_INCOMING_RTL (parms), 0))))
278 DECL_INCOMING_RTL (parms) = new;
279 DECL_RTL (parms) = new;
282 RTVEC_ELT (arg_vector, i) = p;
284 if (GET_CODE (p) == REG)
285 parmdecl_map[REGNO (p)] = parms;
286 else if (GET_CODE (p) == CONCAT)
288 rtx preal = gen_realpart (GET_MODE (XEXP (p, 0)), p);
289 rtx pimag = gen_imagpart (GET_MODE (preal), p);
291 if (GET_CODE (preal) == REG)
292 parmdecl_map[REGNO (preal)] = parms;
293 if (GET_CODE (pimag) == REG)
294 parmdecl_map[REGNO (pimag)] = parms;
297 /* This flag is cleared later
298 if the function ever modifies the value of the parm. */
299 TREE_READONLY (parms) = 1;
302 /* Assume we start out in the insns that set up the parameters. */
303 in_nonparm_insns = 0;
305 /* The list of DECL_SAVED_INSNS, starts off with a header which
306 contains the following information:
308 the first insn of the function (not including the insns that copy
309 parameters into registers).
310 the first parameter insn of the function,
311 the first label used by that function,
312 the last label used by that function,
313 the highest register number used for parameters,
314 the total number of registers used,
315 the size of the incoming stack area for parameters,
316 the number of bytes popped on return,
317 the stack slot list,
318 some flags that are used to restore compiler globals,
319 the value of current_function_outgoing_args_size,
320 the original argument vector,
321 and the original DECL_INITIAL. */
323 return gen_inline_header_rtx (NULL_RTX, NULL_RTX, min_labelno, max_labelno,
324 max_parm_reg, max_reg,
325 current_function_args_size,
326 current_function_pops_args,
327 stack_slot_list, forced_labels, function_flags,
328 current_function_outgoing_args_size,
329 arg_vector, (rtx) DECL_INITIAL (fndecl));
332 /* Subroutine for `save_for_inline{copying,nocopy}'. Finishes up the
333 things that must be done to make FNDECL expandable as an inline function.
334 HEAD contains the chain of insns to which FNDECL will expand. */
336 static void
337 finish_inline (fndecl, head)
338 tree fndecl;
339 rtx head;
341 NEXT_INSN (head) = get_first_nonparm_insn ();
342 FIRST_PARM_INSN (head) = get_insns ();
343 DECL_SAVED_INSNS (fndecl) = head;
344 DECL_FRAME_SIZE (fndecl) = get_frame_size ();
347 /* Adjust the BLOCK_END_NOTE pointers in a given copied DECL tree so that
348 they all point to the new (copied) rtxs. */
350 static void
351 adjust_copied_decl_tree (block)
352 register tree block;
354 register tree subblock;
355 register rtx original_end;
357 original_end = BLOCK_END_NOTE (block);
358 if (original_end)
360 BLOCK_END_NOTE (block) = (rtx) NOTE_SOURCE_FILE (original_end);
361 NOTE_SOURCE_FILE (original_end) = 0;
364 /* Process all subblocks. */
365 for (subblock = BLOCK_SUBBLOCKS (block);
366 subblock;
367 subblock = TREE_CHAIN (subblock))
368 adjust_copied_decl_tree (subblock);
371 /* Make the insns and PARM_DECLs of the current function permanent
372 and record other information in DECL_SAVED_INSNS to allow inlining
373 of this function in subsequent calls.
375 This function is called when we are going to immediately compile
376 the insns for FNDECL. The insns in maybepermanent_obstack cannot be
377 modified by the compilation process, so we copy all of them to
378 new storage and consider the new insns to be the insn chain to be
379 compiled. Our caller (rest_of_compilation) saves the original
380 DECL_INITIAL and DECL_ARGUMENTS; here we copy them. */
382 /* ??? The nonlocal_label list should be adjusted also. However, since
383 a function that contains a nested function never gets inlined currently,
384 the nonlocal_label list will always be empty, so we don't worry about
385 it for now. */
387 void
388 save_for_inline_copying (fndecl)
389 tree fndecl;
391 rtx first_insn, last_insn, insn;
392 rtx head, copy;
393 int max_labelno, min_labelno, i, len;
394 int max_reg;
395 int max_uid;
396 rtx first_nonparm_insn;
398 /* Make and emit a return-label if we have not already done so.
399 Do this before recording the bounds on label numbers. */
401 if (return_label == 0)
403 return_label = gen_label_rtx ();
404 emit_label (return_label);
407 /* Get some bounds on the labels and registers used. */
409 max_labelno = max_label_num ();
410 min_labelno = get_first_label_num ();
411 max_reg = max_reg_num ();
413 /* Set up PARMDECL_MAP which maps pseudo-reg number to its PARM_DECL.
414 Later we set TREE_READONLY to 0 if the parm is modified inside the fn.
415 Also set up ARG_VECTOR, which holds the unmodified DECL_RTX values
416 for the parms, prior to elimination of virtual registers.
417 These values are needed for substituting parms properly. */
419 max_parm_reg = max_parm_reg_num ();
420 parmdecl_map = (tree *) alloca (max_parm_reg * sizeof (tree));
422 head = initialize_for_inline (fndecl, min_labelno, max_labelno, max_reg, 1);
424 if (current_function_uses_const_pool)
426 /* Replace any constant pool references with the actual constant. We
427 will put the constants back in the copy made below. */
428 for (insn = get_insns (); insn; insn = NEXT_INSN (insn))
429 if (GET_RTX_CLASS (GET_CODE (insn)) == 'i')
431 save_constants (&PATTERN (insn));
432 if (REG_NOTES (insn))
433 save_constants (&REG_NOTES (insn));
436 /* Clear out the constant pool so that we can recreate it with the
437 copied constants below. */
438 init_const_rtx_hash_table ();
439 clear_const_double_mem ();
442 max_uid = INSN_UID (head);
444 /* We have now allocated all that needs to be allocated permanently
445 on the rtx obstack. Set our high-water mark, so that we
446 can free the rest of this when the time comes. */
448 preserve_data ();
450 /* Copy the chain insns of this function.
451 Install the copied chain as the insns of this function,
452 for continued compilation;
453 the original chain is recorded as the DECL_SAVED_INSNS
454 for inlining future calls. */
456 /* If there are insns that copy parms from the stack into pseudo registers,
457 those insns are not copied. `expand_inline_function' must
458 emit the correct code to handle such things. */
460 insn = get_insns ();
461 if (GET_CODE (insn) != NOTE)
462 abort ();
463 first_insn = rtx_alloc (NOTE);
464 NOTE_SOURCE_FILE (first_insn) = NOTE_SOURCE_FILE (insn);
465 NOTE_LINE_NUMBER (first_insn) = NOTE_LINE_NUMBER (insn);
466 INSN_UID (first_insn) = INSN_UID (insn);
467 PREV_INSN (first_insn) = NULL;
468 NEXT_INSN (first_insn) = NULL;
469 last_insn = first_insn;
471 /* Each pseudo-reg in the old insn chain must have a unique rtx in the copy.
472 Make these new rtx's now, and install them in regno_reg_rtx, so they
473 will be the official pseudo-reg rtx's for the rest of compilation. */
475 reg_map = (rtx *) alloca ((max_reg + 1) * sizeof (rtx));
477 len = sizeof (struct rtx_def) + (GET_RTX_LENGTH (REG) - 1) * sizeof (rtunion);
478 for (i = max_reg - 1; i > LAST_VIRTUAL_REGISTER; i--)
479 reg_map[i] = (rtx)obstack_copy (function_maybepermanent_obstack,
480 regno_reg_rtx[i], len);
482 bcopy ((char *) (reg_map + LAST_VIRTUAL_REGISTER + 1),
483 (char *) (regno_reg_rtx + LAST_VIRTUAL_REGISTER + 1),
484 (max_reg - (LAST_VIRTUAL_REGISTER + 1)) * sizeof (rtx));
486 /* Likewise each label rtx must have a unique rtx as its copy. */
488 label_map = (rtx *)alloca ((max_labelno - min_labelno) * sizeof (rtx));
489 label_map -= min_labelno;
491 for (i = min_labelno; i < max_labelno; i++)
492 label_map[i] = gen_label_rtx ();
494 /* Record the mapping of old insns to copied insns. */
496 insn_map = (rtx *) alloca (max_uid * sizeof (rtx));
497 bzero ((char *) insn_map, max_uid * sizeof (rtx));
499 /* Get the insn which signals the end of parameter setup code. */
500 first_nonparm_insn = get_first_nonparm_insn ();
502 /* Copy any entries in regno_reg_rtx or DECL_RTLs that reference MEM
503 (the former occurs when a variable has its address taken)
504 since these may be shared and can be changed by virtual
505 register instantiation. DECL_RTL values for our arguments
506 have already been copied by initialize_for_inline. */
507 for (i = LAST_VIRTUAL_REGISTER + 1; i < max_reg; i++)
508 if (GET_CODE (regno_reg_rtx[i]) == MEM)
509 XEXP (regno_reg_rtx[i], 0)
510 = copy_for_inline (XEXP (regno_reg_rtx[i], 0));
512 /* Copy the tree of subblocks of the function, and the decls in them.
513 We will use the copy for compiling this function, then restore the original
514 subblocks and decls for use when inlining this function.
516 Several parts of the compiler modify BLOCK trees. In particular,
517 instantiate_virtual_regs will instantiate any virtual regs
518 mentioned in the DECL_RTLs of the decls, and loop
519 unrolling will replicate any BLOCK trees inside an unrolled loop.
521 The modified subblocks or DECL_RTLs would be incorrect for the original rtl
522 which we will use for inlining. The rtl might even contain pseudoregs
523 whose space has been freed. */
525 DECL_INITIAL (fndecl) = copy_decl_tree (DECL_INITIAL (fndecl));
526 DECL_ARGUMENTS (fndecl) = copy_decl_list (DECL_ARGUMENTS (fndecl));
528 /* Now copy each DECL_RTL which is a MEM,
529 so it is safe to modify their addresses. */
530 copy_decl_rtls (DECL_INITIAL (fndecl));
532 /* The fndecl node acts as its own progenitor, so mark it as such. */
533 DECL_ABSTRACT_ORIGIN (fndecl) = fndecl;
535 /* Now copy the chain of insns. Do this twice. The first copy the insn
536 itself and its body. The second time copy of REG_NOTES. This is because
537 a REG_NOTE may have a forward pointer to another insn. */
539 for (insn = NEXT_INSN (insn); insn; insn = NEXT_INSN (insn))
541 orig_asm_operands_vector = 0;
543 if (insn == first_nonparm_insn)
544 in_nonparm_insns = 1;
546 switch (GET_CODE (insn))
548 case NOTE:
549 /* No need to keep these. */
550 if (NOTE_LINE_NUMBER (insn) == NOTE_INSN_DELETED)
551 continue;
553 copy = rtx_alloc (NOTE);
554 NOTE_LINE_NUMBER (copy) = NOTE_LINE_NUMBER (insn);
555 if (NOTE_LINE_NUMBER (insn) != NOTE_INSN_BLOCK_END)
556 NOTE_SOURCE_FILE (copy) = NOTE_SOURCE_FILE (insn);
557 else
559 NOTE_SOURCE_FILE (insn) = (char *) copy;
560 NOTE_SOURCE_FILE (copy) = 0;
562 break;
564 case INSN:
565 case JUMP_INSN:
566 case CALL_INSN:
567 copy = rtx_alloc (GET_CODE (insn));
569 if (GET_CODE (insn) == CALL_INSN)
570 CALL_INSN_FUNCTION_USAGE (copy) =
571 copy_for_inline (CALL_INSN_FUNCTION_USAGE (insn));
573 PATTERN (copy) = copy_for_inline (PATTERN (insn));
574 INSN_CODE (copy) = -1;
575 LOG_LINKS (copy) = NULL_RTX;
576 RTX_INTEGRATED_P (copy) = RTX_INTEGRATED_P (insn);
577 break;
579 case CODE_LABEL:
580 copy = label_map[CODE_LABEL_NUMBER (insn)];
581 LABEL_NAME (copy) = LABEL_NAME (insn);
582 break;
584 case BARRIER:
585 copy = rtx_alloc (BARRIER);
586 break;
588 default:
589 abort ();
591 INSN_UID (copy) = INSN_UID (insn);
592 insn_map[INSN_UID (insn)] = copy;
593 NEXT_INSN (last_insn) = copy;
594 PREV_INSN (copy) = last_insn;
595 last_insn = copy;
598 adjust_copied_decl_tree (DECL_INITIAL (fndecl));
600 /* Now copy the REG_NOTES. */
601 for (insn = NEXT_INSN (get_insns ()); insn; insn = NEXT_INSN (insn))
602 if (GET_RTX_CLASS (GET_CODE (insn)) == 'i'
603 && insn_map[INSN_UID(insn)])
604 REG_NOTES (insn_map[INSN_UID (insn)])
605 = copy_for_inline (REG_NOTES (insn));
607 NEXT_INSN (last_insn) = NULL;
609 finish_inline (fndecl, head);
611 set_new_first_and_last_insn (first_insn, last_insn);
614 /* Return a copy of a chain of nodes, chained through the TREE_CHAIN field.
615 For example, this can copy a list made of TREE_LIST nodes. While copying,
616 for each node copied which doesn't already have is DECL_ABSTRACT_ORIGIN
617 set to some non-zero value, set the DECL_ABSTRACT_ORIGIN of the copy to
618 point to the corresponding (abstract) original node. */
620 static tree
621 copy_decl_list (list)
622 tree list;
624 tree head;
625 register tree prev, next;
627 if (list == 0)
628 return 0;
630 head = prev = copy_node (list);
631 if (DECL_ABSTRACT_ORIGIN (head) == NULL_TREE)
632 DECL_ABSTRACT_ORIGIN (head) = list;
633 next = TREE_CHAIN (list);
634 while (next)
636 register tree copy;
638 copy = copy_node (next);
639 if (DECL_ABSTRACT_ORIGIN (copy) == NULL_TREE)
640 DECL_ABSTRACT_ORIGIN (copy) = next;
641 TREE_CHAIN (prev) = copy;
642 prev = copy;
643 next = TREE_CHAIN (next);
645 return head;
648 /* Make a copy of the entire tree of blocks BLOCK, and return it. */
650 static tree
651 copy_decl_tree (block)
652 tree block;
654 tree t, vars, subblocks;
656 vars = copy_decl_list (BLOCK_VARS (block));
657 subblocks = 0;
659 /* Process all subblocks. */
660 for (t = BLOCK_SUBBLOCKS (block); t; t = TREE_CHAIN (t))
662 tree copy = copy_decl_tree (t);
663 TREE_CHAIN (copy) = subblocks;
664 subblocks = copy;
667 t = copy_node (block);
668 BLOCK_VARS (t) = vars;
669 BLOCK_SUBBLOCKS (t) = nreverse (subblocks);
670 /* If the BLOCK being cloned is already marked as having been instantiated
671 from something else, then leave that `origin' marking alone. Elsewise,
672 mark the clone as having originated from the BLOCK we are cloning. */
673 if (BLOCK_ABSTRACT_ORIGIN (t) == NULL_TREE)
674 BLOCK_ABSTRACT_ORIGIN (t) = block;
675 return t;
678 /* Copy DECL_RTLs in all decls in the given BLOCK node. */
680 static void
681 copy_decl_rtls (block)
682 tree block;
684 tree t;
686 for (t = BLOCK_VARS (block); t; t = TREE_CHAIN (t))
687 if (DECL_RTL (t) && GET_CODE (DECL_RTL (t)) == MEM)
688 DECL_RTL (t) = copy_for_inline (DECL_RTL (t));
690 /* Process all subblocks. */
691 for (t = BLOCK_SUBBLOCKS (block); t; t = TREE_CHAIN (t))
692 copy_decl_rtls (t);
695 /* Make the insns and PARM_DECLs of the current function permanent
696 and record other information in DECL_SAVED_INSNS to allow inlining
697 of this function in subsequent calls.
699 This routine need not copy any insns because we are not going
700 to immediately compile the insns in the insn chain. There
701 are two cases when we would compile the insns for FNDECL:
702 (1) when FNDECL is expanded inline, and (2) when FNDECL needs to
703 be output at the end of other compilation, because somebody took
704 its address. In the first case, the insns of FNDECL are copied
705 as it is expanded inline, so FNDECL's saved insns are not
706 modified. In the second case, FNDECL is used for the last time,
707 so modifying the rtl is not a problem.
709 ??? Actually, we do not verify that FNDECL is not inline expanded
710 by other functions which must also be written down at the end
711 of compilation. We could set flag_no_inline to nonzero when
712 the time comes to write down such functions. */
714 void
715 save_for_inline_nocopy (fndecl)
716 tree fndecl;
718 rtx insn;
719 rtx head;
720 rtx first_nonparm_insn;
722 /* Set up PARMDECL_MAP which maps pseudo-reg number to its PARM_DECL.
723 Later we set TREE_READONLY to 0 if the parm is modified inside the fn.
724 Also set up ARG_VECTOR, which holds the unmodified DECL_RTX values
725 for the parms, prior to elimination of virtual registers.
726 These values are needed for substituting parms properly. */
728 max_parm_reg = max_parm_reg_num ();
729 parmdecl_map = (tree *) alloca (max_parm_reg * sizeof (tree));
731 /* Make and emit a return-label if we have not already done so. */
733 if (return_label == 0)
735 return_label = gen_label_rtx ();
736 emit_label (return_label);
739 head = initialize_for_inline (fndecl, get_first_label_num (),
740 max_label_num (), max_reg_num (), 0);
742 /* If there are insns that copy parms from the stack into pseudo registers,
743 those insns are not copied. `expand_inline_function' must
744 emit the correct code to handle such things. */
746 insn = get_insns ();
747 if (GET_CODE (insn) != NOTE)
748 abort ();
750 /* Get the insn which signals the end of parameter setup code. */
751 first_nonparm_insn = get_first_nonparm_insn ();
753 /* Now just scan the chain of insns to see what happens to our
754 PARM_DECLs. If a PARM_DECL is used but never modified, we
755 can substitute its rtl directly when expanding inline (and
756 perform constant folding when its incoming value is constant).
757 Otherwise, we have to copy its value into a new register and track
758 the new register's life. */
760 for (insn = NEXT_INSN (insn); insn; insn = NEXT_INSN (insn))
762 if (insn == first_nonparm_insn)
763 in_nonparm_insns = 1;
765 if (GET_RTX_CLASS (GET_CODE (insn)) == 'i')
767 if (current_function_uses_const_pool)
769 /* Replace any constant pool references with the actual constant.
770 We will put the constant back if we need to write the
771 function out after all. */
772 save_constants (&PATTERN (insn));
773 if (REG_NOTES (insn))
774 save_constants (&REG_NOTES (insn));
777 /* Record what interesting things happen to our parameters. */
778 note_stores (PATTERN (insn), note_modified_parmregs);
782 /* We have now allocated all that needs to be allocated permanently
783 on the rtx obstack. Set our high-water mark, so that we
784 can free the rest of this when the time comes. */
786 preserve_data ();
788 finish_inline (fndecl, head);
791 /* Given PX, a pointer into an insn, search for references to the constant
792 pool. Replace each with a CONST that has the mode of the original
793 constant, contains the constant, and has RTX_INTEGRATED_P set.
794 Similarly, constant pool addresses not enclosed in a MEM are replaced
795 with an ADDRESS rtx which also gives the constant, mode, and has
796 RTX_INTEGRATED_P set. */
798 static void
799 save_constants (px)
800 rtx *px;
802 rtx x;
803 int i, j;
805 again:
806 x = *px;
808 /* If this is a CONST_DOUBLE, don't try to fix things up in
809 CONST_DOUBLE_MEM, because this is an infinite recursion. */
810 if (GET_CODE (x) == CONST_DOUBLE)
811 return;
812 else if (GET_CODE (x) == MEM && GET_CODE (XEXP (x, 0)) == SYMBOL_REF
813 && CONSTANT_POOL_ADDRESS_P (XEXP (x,0)))
815 enum machine_mode const_mode = get_pool_mode (XEXP (x, 0));
816 rtx new = gen_rtx (CONST, const_mode, get_pool_constant (XEXP (x, 0)));
817 RTX_INTEGRATED_P (new) = 1;
819 /* If the MEM was in a different mode than the constant (perhaps we
820 were only looking at the low-order part), surround it with a
821 SUBREG so we can save both modes. */
823 if (GET_MODE (x) != const_mode)
825 new = gen_rtx (SUBREG, GET_MODE (x), new, 0);
826 RTX_INTEGRATED_P (new) = 1;
829 *px = new;
830 save_constants (&XEXP (*px, 0));
832 else if (GET_CODE (x) == SYMBOL_REF
833 && CONSTANT_POOL_ADDRESS_P (x))
835 *px = gen_rtx (ADDRESS, get_pool_mode (x), get_pool_constant (x));
836 save_constants (&XEXP (*px, 0));
837 RTX_INTEGRATED_P (*px) = 1;
840 else
842 char *fmt = GET_RTX_FORMAT (GET_CODE (x));
843 int len = GET_RTX_LENGTH (GET_CODE (x));
845 for (i = len-1; i >= 0; i--)
847 switch (fmt[i])
849 case 'E':
850 for (j = 0; j < XVECLEN (x, i); j++)
851 save_constants (&XVECEXP (x, i, j));
852 break;
854 case 'e':
855 if (XEXP (x, i) == 0)
856 continue;
857 if (i == 0)
859 /* Hack tail-recursion here. */
860 px = &XEXP (x, 0);
861 goto again;
863 save_constants (&XEXP (x, i));
864 break;
870 /* Note whether a parameter is modified or not. */
872 static void
873 note_modified_parmregs (reg, x)
874 rtx reg;
875 rtx x;
877 if (GET_CODE (reg) == REG && in_nonparm_insns
878 && REGNO (reg) < max_parm_reg
879 && REGNO (reg) >= FIRST_PSEUDO_REGISTER
880 && parmdecl_map[REGNO (reg)] != 0)
881 TREE_READONLY (parmdecl_map[REGNO (reg)]) = 0;
884 /* Copy the rtx ORIG recursively, replacing pseudo-regs and labels
885 according to `reg_map' and `label_map'. The original rtl insns
886 will be saved for inlining; this is used to make a copy
887 which is used to finish compiling the inline function itself.
889 If we find a "saved" constant pool entry, one which was replaced with
890 the value of the constant, convert it back to a constant pool entry.
891 Since the pool wasn't touched, this should simply restore the old
892 address.
894 All other kinds of rtx are copied except those that can never be
895 changed during compilation. */
897 static rtx
898 copy_for_inline (orig)
899 rtx orig;
901 register rtx x = orig;
902 register int i;
903 register enum rtx_code code;
904 register char *format_ptr;
906 if (x == 0)
907 return x;
909 code = GET_CODE (x);
911 /* These types may be freely shared. */
913 switch (code)
915 case QUEUED:
916 case CONST_INT:
917 case SYMBOL_REF:
918 case PC:
919 case CC0:
920 return x;
922 case CONST_DOUBLE:
923 /* We have to make a new CONST_DOUBLE to ensure that we account for
924 it correctly. Using the old CONST_DOUBLE_MEM data is wrong. */
925 if (GET_MODE_CLASS (GET_MODE (x)) == MODE_FLOAT)
927 REAL_VALUE_TYPE d;
929 REAL_VALUE_FROM_CONST_DOUBLE (d, x);
930 return CONST_DOUBLE_FROM_REAL_VALUE (d, GET_MODE (x));
932 else
933 return immed_double_const (CONST_DOUBLE_LOW (x), CONST_DOUBLE_HIGH (x),
934 VOIDmode);
936 case CONST:
937 /* Get constant pool entry for constant in the pool. */
938 if (RTX_INTEGRATED_P (x))
939 return validize_mem (force_const_mem (GET_MODE (x),
940 copy_for_inline (XEXP (x, 0))));
941 break;
943 case SUBREG:
944 /* Get constant pool entry, but access in different mode. */
945 if (RTX_INTEGRATED_P (x))
947 rtx new
948 = force_const_mem (GET_MODE (SUBREG_REG (x)),
949 copy_for_inline (XEXP (SUBREG_REG (x), 0)));
951 PUT_MODE (new, GET_MODE (x));
952 return validize_mem (new);
954 break;
956 case ADDRESS:
957 /* If not special for constant pool error. Else get constant pool
958 address. */
959 if (! RTX_INTEGRATED_P (x))
960 abort ();
962 return XEXP (force_const_mem (GET_MODE (x),
963 copy_for_inline (XEXP (x, 0))), 0);
965 case ASM_OPERANDS:
966 /* If a single asm insn contains multiple output operands
967 then it contains multiple ASM_OPERANDS rtx's that share operand 3.
968 We must make sure that the copied insn continues to share it. */
969 if (orig_asm_operands_vector == XVEC (orig, 3))
971 x = rtx_alloc (ASM_OPERANDS);
972 x->volatil = orig->volatil;
973 XSTR (x, 0) = XSTR (orig, 0);
974 XSTR (x, 1) = XSTR (orig, 1);
975 XINT (x, 2) = XINT (orig, 2);
976 XVEC (x, 3) = copy_asm_operands_vector;
977 XVEC (x, 4) = copy_asm_constraints_vector;
978 XSTR (x, 5) = XSTR (orig, 5);
979 XINT (x, 6) = XINT (orig, 6);
980 return x;
982 break;
984 case MEM:
985 /* A MEM is usually allowed to be shared if its address is constant
986 or is a constant plus one of the special registers.
988 We do not allow sharing of addresses that are either a special
989 register or the sum of a constant and a special register because
990 it is possible for unshare_all_rtl to copy the address, into memory
991 that won't be saved. Although the MEM can safely be shared, and
992 won't be copied there, the address itself cannot be shared, and may
993 need to be copied.
995 There are also two exceptions with constants: The first is if the
996 constant is a LABEL_REF or the sum of the LABEL_REF
997 and an integer. This case can happen if we have an inline
998 function that supplies a constant operand to the call of another
999 inline function that uses it in a switch statement. In this case,
1000 we will be replacing the LABEL_REF, so we have to replace this MEM
1001 as well.
1003 The second case is if we have a (const (plus (address ..) ...)).
1004 In that case we need to put back the address of the constant pool
1005 entry. */
1007 if (CONSTANT_ADDRESS_P (XEXP (x, 0))
1008 && GET_CODE (XEXP (x, 0)) != LABEL_REF
1009 && ! (GET_CODE (XEXP (x, 0)) == CONST
1010 && (GET_CODE (XEXP (XEXP (x, 0), 0)) == PLUS
1011 && ((GET_CODE (XEXP (XEXP (XEXP (x, 0), 0), 0))
1012 == LABEL_REF)
1013 || (GET_CODE (XEXP (XEXP (XEXP (x, 0), 0), 0))
1014 == ADDRESS)))))
1015 return x;
1016 break;
1018 case LABEL_REF:
1019 /* If this is a non-local label, just make a new LABEL_REF.
1020 Otherwise, use the new label as well. */
1021 x = gen_rtx (LABEL_REF, GET_MODE (orig),
1022 LABEL_REF_NONLOCAL_P (orig) ? XEXP (orig, 0)
1023 : label_map[CODE_LABEL_NUMBER (XEXP (orig, 0))]);
1024 LABEL_REF_NONLOCAL_P (x) = LABEL_REF_NONLOCAL_P (orig);
1025 LABEL_OUTSIDE_LOOP_P (x) = LABEL_OUTSIDE_LOOP_P (orig);
1026 return x;
1028 case REG:
1029 if (REGNO (x) > LAST_VIRTUAL_REGISTER)
1030 return reg_map [REGNO (x)];
1031 else
1032 return x;
1034 case SET:
1035 /* If a parm that gets modified lives in a pseudo-reg,
1036 clear its TREE_READONLY to prevent certain optimizations. */
1038 rtx dest = SET_DEST (x);
1040 while (GET_CODE (dest) == STRICT_LOW_PART
1041 || GET_CODE (dest) == ZERO_EXTRACT
1042 || GET_CODE (dest) == SUBREG)
1043 dest = XEXP (dest, 0);
1045 if (GET_CODE (dest) == REG
1046 && REGNO (dest) < max_parm_reg
1047 && REGNO (dest) >= FIRST_PSEUDO_REGISTER
1048 && parmdecl_map[REGNO (dest)] != 0
1049 /* The insn to load an arg pseudo from a stack slot
1050 does not count as modifying it. */
1051 && in_nonparm_insns)
1052 TREE_READONLY (parmdecl_map[REGNO (dest)]) = 0;
1054 break;
1056 #if 0 /* This is a good idea, but here is the wrong place for it. */
1057 /* Arrange that CONST_INTs always appear as the second operand
1058 if they appear, and that `frame_pointer_rtx' or `arg_pointer_rtx'
1059 always appear as the first. */
1060 case PLUS:
1061 if (GET_CODE (XEXP (x, 0)) == CONST_INT
1062 || (XEXP (x, 1) == frame_pointer_rtx
1063 || (ARG_POINTER_REGNUM != FRAME_POINTER_REGNUM
1064 && XEXP (x, 1) == arg_pointer_rtx)))
1066 rtx t = XEXP (x, 0);
1067 XEXP (x, 0) = XEXP (x, 1);
1068 XEXP (x, 1) = t;
1070 break;
1071 #endif
1074 /* Replace this rtx with a copy of itself. */
1076 x = rtx_alloc (code);
1077 bcopy ((char *) orig, (char *) x,
1078 (sizeof (*x) - sizeof (x->fld)
1079 + sizeof (x->fld[0]) * GET_RTX_LENGTH (code)));
1081 /* Now scan the subexpressions recursively.
1082 We can store any replaced subexpressions directly into X
1083 since we know X is not shared! Any vectors in X
1084 must be copied if X was copied. */
1086 format_ptr = GET_RTX_FORMAT (code);
1088 for (i = 0; i < GET_RTX_LENGTH (code); i++)
1090 switch (*format_ptr++)
1092 case 'e':
1093 XEXP (x, i) = copy_for_inline (XEXP (x, i));
1094 break;
1096 case 'u':
1097 /* Change any references to old-insns to point to the
1098 corresponding copied insns. */
1099 XEXP (x, i) = insn_map[INSN_UID (XEXP (x, i))];
1100 break;
1102 case 'E':
1103 if (XVEC (x, i) != NULL && XVECLEN (x, i) != 0)
1105 register int j;
1107 XVEC (x, i) = gen_rtvec_v (XVECLEN (x, i), &XVECEXP (x, i, 0));
1108 for (j = 0; j < XVECLEN (x, i); j++)
1109 XVECEXP (x, i, j)
1110 = copy_for_inline (XVECEXP (x, i, j));
1112 break;
1116 if (code == ASM_OPERANDS && orig_asm_operands_vector == 0)
1118 orig_asm_operands_vector = XVEC (orig, 3);
1119 copy_asm_operands_vector = XVEC (x, 3);
1120 copy_asm_constraints_vector = XVEC (x, 4);
1123 return x;
1126 /* Unfortunately, we need a global copy of const_equiv map for communication
1127 with a function called from note_stores. Be *very* careful that this
1128 is used properly in the presence of recursion. */
1130 rtx *global_const_equiv_map;
1131 int global_const_equiv_map_size;
1133 #define FIXED_BASE_PLUS_P(X) \
1134 (GET_CODE (X) == PLUS && GET_CODE (XEXP (X, 1)) == CONST_INT \
1135 && GET_CODE (XEXP (X, 0)) == REG \
1136 && REGNO (XEXP (X, 0)) >= FIRST_VIRTUAL_REGISTER \
1137 && REGNO (XEXP (X, 0)) <= LAST_VIRTUAL_REGISTER)
1139 /* Integrate the procedure defined by FNDECL. Note that this function
1140 may wind up calling itself. Since the static variables are not
1141 reentrant, we do not assign them until after the possibility
1142 of recursion is eliminated.
1144 If IGNORE is nonzero, do not produce a value.
1145 Otherwise store the value in TARGET if it is nonzero and that is convenient.
1147 Value is:
1148 (rtx)-1 if we could not substitute the function
1149 0 if we substituted it and it does not produce a value
1150 else an rtx for where the value is stored. */
1153 expand_inline_function (fndecl, parms, target, ignore, type, structure_value_addr)
1154 tree fndecl, parms;
1155 rtx target;
1156 int ignore;
1157 tree type;
1158 rtx structure_value_addr;
1160 tree formal, actual, block;
1161 rtx header = DECL_SAVED_INSNS (fndecl);
1162 rtx insns = FIRST_FUNCTION_INSN (header);
1163 rtx parm_insns = FIRST_PARM_INSN (header);
1164 tree *arg_trees;
1165 rtx *arg_vals;
1166 rtx insn;
1167 int max_regno;
1168 register int i;
1169 int min_labelno = FIRST_LABELNO (header);
1170 int max_labelno = LAST_LABELNO (header);
1171 int nargs;
1172 rtx local_return_label = 0;
1173 rtx loc;
1174 rtx stack_save = 0;
1175 rtx temp;
1176 struct inline_remap *map;
1177 rtx cc0_insn = 0;
1178 rtvec arg_vector = ORIGINAL_ARG_VECTOR (header);
1179 rtx static_chain_value = 0;
1181 /* Allow for equivalences of the pseudos we make for virtual fp and ap. */
1182 max_regno = MAX_REGNUM (header) + 3;
1183 if (max_regno < FIRST_PSEUDO_REGISTER)
1184 abort ();
1186 nargs = list_length (DECL_ARGUMENTS (fndecl));
1188 /* Check that the parms type match and that sufficient arguments were
1189 passed. Since the appropriate conversions or default promotions have
1190 already been applied, the machine modes should match exactly. */
1192 for (formal = DECL_ARGUMENTS (fndecl),
1193 actual = parms;
1194 formal;
1195 formal = TREE_CHAIN (formal),
1196 actual = TREE_CHAIN (actual))
1198 tree arg;
1199 enum machine_mode mode;
1201 if (actual == 0)
1202 return (rtx) (HOST_WIDE_INT) -1;
1204 arg = TREE_VALUE (actual);
1205 mode= TYPE_MODE (DECL_ARG_TYPE (formal));
1207 if (mode != TYPE_MODE (TREE_TYPE (arg))
1208 /* If they are block mode, the types should match exactly.
1209 They don't match exactly if TREE_TYPE (FORMAL) == ERROR_MARK_NODE,
1210 which could happen if the parameter has incomplete type. */
1211 || (mode == BLKmode && TREE_TYPE (arg) != TREE_TYPE (formal)))
1212 return (rtx) (HOST_WIDE_INT) -1;
1215 /* Extra arguments are valid, but will be ignored below, so we must
1216 evaluate them here for side-effects. */
1217 for (; actual; actual = TREE_CHAIN (actual))
1218 expand_expr (TREE_VALUE (actual), const0_rtx,
1219 TYPE_MODE (TREE_TYPE (TREE_VALUE (actual))), 0);
1221 /* Make a binding contour to keep inline cleanups called at
1222 outer function-scope level from looking like they are shadowing
1223 parameter declarations. */
1224 pushlevel (0);
1226 /* Make a fresh binding contour that we can easily remove. */
1227 pushlevel (0);
1228 expand_start_bindings (0);
1230 /* Expand the function arguments. Do this first so that any
1231 new registers get created before we allocate the maps. */
1233 arg_vals = (rtx *) alloca (nargs * sizeof (rtx));
1234 arg_trees = (tree *) alloca (nargs * sizeof (tree));
1236 for (formal = DECL_ARGUMENTS (fndecl), actual = parms, i = 0;
1237 formal;
1238 formal = TREE_CHAIN (formal), actual = TREE_CHAIN (actual), i++)
1240 /* Actual parameter, converted to the type of the argument within the
1241 function. */
1242 tree arg = convert (TREE_TYPE (formal), TREE_VALUE (actual));
1243 /* Mode of the variable used within the function. */
1244 enum machine_mode mode = TYPE_MODE (TREE_TYPE (formal));
1245 int invisiref = 0;
1247 arg_trees[i] = arg;
1248 loc = RTVEC_ELT (arg_vector, i);
1250 /* If this is an object passed by invisible reference, we copy the
1251 object into a stack slot and save its address. If this will go
1252 into memory, we do nothing now. Otherwise, we just expand the
1253 argument. */
1254 if (GET_CODE (loc) == MEM && GET_CODE (XEXP (loc, 0)) == REG
1255 && REGNO (XEXP (loc, 0)) > LAST_VIRTUAL_REGISTER)
1257 rtx stack_slot
1258 = assign_stack_temp (TYPE_MODE (TREE_TYPE (arg)),
1259 int_size_in_bytes (TREE_TYPE (arg)), 1);
1260 MEM_IN_STRUCT_P (stack_slot) = AGGREGATE_TYPE_P (TREE_TYPE (arg));
1262 store_expr (arg, stack_slot, 0);
1264 arg_vals[i] = XEXP (stack_slot, 0);
1265 invisiref = 1;
1267 else if (GET_CODE (loc) != MEM)
1269 if (GET_MODE (loc) != TYPE_MODE (TREE_TYPE (arg)))
1270 /* The mode if LOC and ARG can differ if LOC was a variable
1271 that had its mode promoted via PROMOTED_MODE. */
1272 arg_vals[i] = convert_modes (GET_MODE (loc),
1273 TYPE_MODE (TREE_TYPE (arg)),
1274 expand_expr (arg, NULL_RTX, mode,
1275 EXPAND_SUM),
1276 TREE_UNSIGNED (TREE_TYPE (formal)));
1277 else
1278 arg_vals[i] = expand_expr (arg, NULL_RTX, mode, EXPAND_SUM);
1280 else
1281 arg_vals[i] = 0;
1283 if (arg_vals[i] != 0
1284 && (! TREE_READONLY (formal)
1285 /* If the parameter is not read-only, copy our argument through
1286 a register. Also, we cannot use ARG_VALS[I] if it overlaps
1287 TARGET in any way. In the inline function, they will likely
1288 be two different pseudos, and `safe_from_p' will make all
1289 sorts of smart assumptions about their not conflicting.
1290 But if ARG_VALS[I] overlaps TARGET, these assumptions are
1291 wrong, so put ARG_VALS[I] into a fresh register.
1292 Don't worry about invisible references, since their stack
1293 temps will never overlap the target. */
1294 || (target != 0
1295 && ! invisiref
1296 && (GET_CODE (arg_vals[i]) == REG
1297 || GET_CODE (arg_vals[i]) == SUBREG
1298 || GET_CODE (arg_vals[i]) == MEM)
1299 && reg_overlap_mentioned_p (arg_vals[i], target))
1300 /* ??? We must always copy a SUBREG into a REG, because it might
1301 get substituted into an address, and not all ports correctly
1302 handle SUBREGs in addresses. */
1303 || (GET_CODE (arg_vals[i]) == SUBREG)))
1304 arg_vals[i] = copy_to_mode_reg (GET_MODE (loc), arg_vals[i]);
1307 /* Allocate the structures we use to remap things. */
1309 map = (struct inline_remap *) alloca (sizeof (struct inline_remap));
1310 map->fndecl = fndecl;
1312 map->reg_map = (rtx *) alloca (max_regno * sizeof (rtx));
1313 bzero ((char *) map->reg_map, max_regno * sizeof (rtx));
1315 map->label_map = (rtx *)alloca ((max_labelno - min_labelno) * sizeof (rtx));
1316 map->label_map -= min_labelno;
1318 map->insn_map = (rtx *) alloca (INSN_UID (header) * sizeof (rtx));
1319 bzero ((char *) map->insn_map, INSN_UID (header) * sizeof (rtx));
1320 map->min_insnno = 0;
1321 map->max_insnno = INSN_UID (header);
1323 map->integrating = 1;
1325 /* const_equiv_map maps pseudos in our routine to constants, so it needs to
1326 be large enough for all our pseudos. This is the number we are currently
1327 using plus the number in the called routine, plus 15 for each arg,
1328 five to compute the virtual frame pointer, and five for the return value.
1329 This should be enough for most cases. We do not reference entries
1330 outside the range of the map.
1332 ??? These numbers are quite arbitrary and were obtained by
1333 experimentation. At some point, we should try to allocate the
1334 table after all the parameters are set up so we an more accurately
1335 estimate the number of pseudos we will need. */
1337 map->const_equiv_map_size
1338 = max_reg_num () + (max_regno - FIRST_PSEUDO_REGISTER) + 15 * nargs + 10;
1340 map->const_equiv_map
1341 = (rtx *)alloca (map->const_equiv_map_size * sizeof (rtx));
1342 bzero ((char *) map->const_equiv_map,
1343 map->const_equiv_map_size * sizeof (rtx));
1345 map->const_age_map
1346 = (unsigned *)alloca (map->const_equiv_map_size * sizeof (unsigned));
1347 bzero ((char *) map->const_age_map,
1348 map->const_equiv_map_size * sizeof (unsigned));
1349 map->const_age = 0;
1351 /* Record the current insn in case we have to set up pointers to frame
1352 and argument memory blocks. */
1353 map->insns_at_start = get_last_insn ();
1355 /* Update the outgoing argument size to allow for those in the inlined
1356 function. */
1357 if (OUTGOING_ARGS_SIZE (header) > current_function_outgoing_args_size)
1358 current_function_outgoing_args_size = OUTGOING_ARGS_SIZE (header);
1360 /* If the inline function needs to make PIC references, that means
1361 that this function's PIC offset table must be used. */
1362 if (FUNCTION_FLAGS (header) & FUNCTION_FLAGS_USES_PIC_OFFSET_TABLE)
1363 current_function_uses_pic_offset_table = 1;
1365 /* If this function needs a context, set it up. */
1366 if (FUNCTION_FLAGS (header) & FUNCTION_FLAGS_NEEDS_CONTEXT)
1367 static_chain_value = lookup_static_chain (fndecl);
1369 if (GET_CODE (parm_insns) == NOTE
1370 && NOTE_LINE_NUMBER (parm_insns) > 0)
1372 rtx note = emit_note (NOTE_SOURCE_FILE (parm_insns),
1373 NOTE_LINE_NUMBER (parm_insns));
1374 if (note)
1375 RTX_INTEGRATED_P (note) = 1;
1378 /* Process each argument. For each, set up things so that the function's
1379 reference to the argument will refer to the argument being passed.
1380 We only replace REG with REG here. Any simplifications are done
1381 via const_equiv_map.
1383 We make two passes: In the first, we deal with parameters that will
1384 be placed into registers, since we need to ensure that the allocated
1385 register number fits in const_equiv_map. Then we store all non-register
1386 parameters into their memory location. */
1388 /* Don't try to free temp stack slots here, because we may put one of the
1389 parameters into a temp stack slot. */
1391 for (i = 0; i < nargs; i++)
1393 rtx copy = arg_vals[i];
1395 loc = RTVEC_ELT (arg_vector, i);
1397 /* There are three cases, each handled separately. */
1398 if (GET_CODE (loc) == MEM && GET_CODE (XEXP (loc, 0)) == REG
1399 && REGNO (XEXP (loc, 0)) > LAST_VIRTUAL_REGISTER)
1401 /* This must be an object passed by invisible reference (it could
1402 also be a variable-sized object, but we forbid inlining functions
1403 with variable-sized arguments). COPY is the address of the
1404 actual value (this computation will cause it to be copied). We
1405 map that address for the register, noting the actual address as
1406 an equivalent in case it can be substituted into the insns. */
1408 if (GET_CODE (copy) != REG)
1410 temp = copy_addr_to_reg (copy);
1411 if ((CONSTANT_P (copy) || FIXED_BASE_PLUS_P (copy))
1412 && REGNO (temp) < map->const_equiv_map_size)
1414 map->const_equiv_map[REGNO (temp)] = copy;
1415 map->const_age_map[REGNO (temp)] = CONST_AGE_PARM;
1417 copy = temp;
1419 map->reg_map[REGNO (XEXP (loc, 0))] = copy;
1421 else if (GET_CODE (loc) == MEM)
1423 /* This is the case of a parameter that lives in memory.
1424 It will live in the block we allocate in the called routine's
1425 frame that simulates the incoming argument area. Do nothing
1426 now; we will call store_expr later. */
1429 else if (GET_CODE (loc) == REG)
1431 /* This is the good case where the parameter is in a register.
1432 If it is read-only and our argument is a constant, set up the
1433 constant equivalence.
1435 If LOC is REG_USERVAR_P, the usual case, COPY must also have
1436 that flag set if it is a register.
1438 Also, don't allow hard registers here; they might not be valid
1439 when substituted into insns. */
1441 if ((GET_CODE (copy) != REG && GET_CODE (copy) != SUBREG)
1442 || (GET_CODE (copy) == REG && REG_USERVAR_P (loc)
1443 && ! REG_USERVAR_P (copy))
1444 || (GET_CODE (copy) == REG
1445 && REGNO (copy) < FIRST_PSEUDO_REGISTER))
1447 temp = copy_to_mode_reg (GET_MODE (loc), copy);
1448 REG_USERVAR_P (temp) = REG_USERVAR_P (loc);
1449 if ((CONSTANT_P (copy) || FIXED_BASE_PLUS_P (copy))
1450 && REGNO (temp) < map->const_equiv_map_size)
1452 map->const_equiv_map[REGNO (temp)] = copy;
1453 map->const_age_map[REGNO (temp)] = CONST_AGE_PARM;
1455 copy = temp;
1457 map->reg_map[REGNO (loc)] = copy;
1459 else if (GET_CODE (loc) == CONCAT)
1461 /* This is the good case where the parameter is in a
1462 pair of separate pseudos.
1463 If it is read-only and our argument is a constant, set up the
1464 constant equivalence.
1466 If LOC is REG_USERVAR_P, the usual case, COPY must also have
1467 that flag set if it is a register.
1469 Also, don't allow hard registers here; they might not be valid
1470 when substituted into insns. */
1471 rtx locreal = gen_realpart (GET_MODE (XEXP (loc, 0)), loc);
1472 rtx locimag = gen_imagpart (GET_MODE (XEXP (loc, 0)), loc);
1473 rtx copyreal = gen_realpart (GET_MODE (locreal), copy);
1474 rtx copyimag = gen_imagpart (GET_MODE (locimag), copy);
1476 if ((GET_CODE (copyreal) != REG && GET_CODE (copyreal) != SUBREG)
1477 || (GET_CODE (copyreal) == REG && REG_USERVAR_P (locreal)
1478 && ! REG_USERVAR_P (copyreal))
1479 || (GET_CODE (copyreal) == REG
1480 && REGNO (copyreal) < FIRST_PSEUDO_REGISTER))
1482 temp = copy_to_mode_reg (GET_MODE (locreal), copyreal);
1483 REG_USERVAR_P (temp) = REG_USERVAR_P (locreal);
1484 if ((CONSTANT_P (copyreal) || FIXED_BASE_PLUS_P (copyreal))
1485 && REGNO (temp) < map->const_equiv_map_size)
1487 map->const_equiv_map[REGNO (temp)] = copyreal;
1488 map->const_age_map[REGNO (temp)] = CONST_AGE_PARM;
1490 copyreal = temp;
1492 map->reg_map[REGNO (locreal)] = copyreal;
1494 if ((GET_CODE (copyimag) != REG && GET_CODE (copyimag) != SUBREG)
1495 || (GET_CODE (copyimag) == REG && REG_USERVAR_P (locimag)
1496 && ! REG_USERVAR_P (copyimag))
1497 || (GET_CODE (copyimag) == REG
1498 && REGNO (copyimag) < FIRST_PSEUDO_REGISTER))
1500 temp = copy_to_mode_reg (GET_MODE (locimag), copyimag);
1501 REG_USERVAR_P (temp) = REG_USERVAR_P (locimag);
1502 if ((CONSTANT_P (copyimag) || FIXED_BASE_PLUS_P (copyimag))
1503 && REGNO (temp) < map->const_equiv_map_size)
1505 map->const_equiv_map[REGNO (temp)] = copyimag;
1506 map->const_age_map[REGNO (temp)] = CONST_AGE_PARM;
1508 copyimag = temp;
1510 map->reg_map[REGNO (locimag)] = copyimag;
1512 else
1513 abort ();
1516 /* Now do the parameters that will be placed in memory. */
1518 for (formal = DECL_ARGUMENTS (fndecl), i = 0;
1519 formal; formal = TREE_CHAIN (formal), i++)
1521 loc = RTVEC_ELT (arg_vector, i);
1523 if (GET_CODE (loc) == MEM
1524 /* Exclude case handled above. */
1525 && ! (GET_CODE (XEXP (loc, 0)) == REG
1526 && REGNO (XEXP (loc, 0)) > LAST_VIRTUAL_REGISTER))
1528 rtx note = emit_note (DECL_SOURCE_FILE (formal),
1529 DECL_SOURCE_LINE (formal));
1530 if (note)
1531 RTX_INTEGRATED_P (note) = 1;
1533 /* Compute the address in the area we reserved and store the
1534 value there. */
1535 temp = copy_rtx_and_substitute (loc, map);
1536 subst_constants (&temp, NULL_RTX, map);
1537 apply_change_group ();
1538 if (! memory_address_p (GET_MODE (temp), XEXP (temp, 0)))
1539 temp = change_address (temp, VOIDmode, XEXP (temp, 0));
1540 store_expr (arg_trees[i], temp, 0);
1544 /* Deal with the places that the function puts its result.
1545 We are driven by what is placed into DECL_RESULT.
1547 Initially, we assume that we don't have anything special handling for
1548 REG_FUNCTION_RETURN_VALUE_P. */
1550 map->inline_target = 0;
1551 loc = DECL_RTL (DECL_RESULT (fndecl));
1552 if (TYPE_MODE (type) == VOIDmode)
1553 /* There is no return value to worry about. */
1555 else if (GET_CODE (loc) == MEM)
1557 if (! structure_value_addr || ! aggregate_value_p (DECL_RESULT (fndecl)))
1558 abort ();
1560 /* Pass the function the address in which to return a structure value.
1561 Note that a constructor can cause someone to call us with
1562 STRUCTURE_VALUE_ADDR, but the initialization takes place
1563 via the first parameter, rather than the struct return address.
1565 We have two cases: If the address is a simple register indirect,
1566 use the mapping mechanism to point that register to our structure
1567 return address. Otherwise, store the structure return value into
1568 the place that it will be referenced from. */
1570 if (GET_CODE (XEXP (loc, 0)) == REG)
1572 temp = force_reg (Pmode, structure_value_addr);
1573 map->reg_map[REGNO (XEXP (loc, 0))] = temp;
1574 if ((CONSTANT_P (structure_value_addr)
1575 || (GET_CODE (structure_value_addr) == PLUS
1576 && XEXP (structure_value_addr, 0) == virtual_stack_vars_rtx
1577 && GET_CODE (XEXP (structure_value_addr, 1)) == CONST_INT))
1578 && REGNO (temp) < map->const_equiv_map_size)
1580 map->const_equiv_map[REGNO (temp)] = structure_value_addr;
1581 map->const_age_map[REGNO (temp)] = CONST_AGE_PARM;
1584 else
1586 temp = copy_rtx_and_substitute (loc, map);
1587 subst_constants (&temp, NULL_RTX, map);
1588 apply_change_group ();
1589 emit_move_insn (temp, structure_value_addr);
1592 else if (ignore)
1593 /* We will ignore the result value, so don't look at its structure.
1594 Note that preparations for an aggregate return value
1595 do need to be made (above) even if it will be ignored. */
1597 else if (GET_CODE (loc) == REG)
1599 /* The function returns an object in a register and we use the return
1600 value. Set up our target for remapping. */
1602 /* Machine mode function was declared to return. */
1603 enum machine_mode departing_mode = TYPE_MODE (type);
1604 /* (Possibly wider) machine mode it actually computes
1605 (for the sake of callers that fail to declare it right). */
1606 enum machine_mode arriving_mode
1607 = TYPE_MODE (TREE_TYPE (DECL_RESULT (fndecl)));
1608 rtx reg_to_map;
1610 /* Don't use MEMs as direct targets because on some machines
1611 substituting a MEM for a REG makes invalid insns.
1612 Let the combiner substitute the MEM if that is valid. */
1613 if (target == 0 || GET_CODE (target) != REG
1614 || GET_MODE (target) != departing_mode)
1615 target = gen_reg_rtx (departing_mode);
1617 /* If function's value was promoted before return,
1618 avoid machine mode mismatch when we substitute INLINE_TARGET.
1619 But TARGET is what we will return to the caller. */
1620 if (arriving_mode != departing_mode)
1621 reg_to_map = gen_rtx (SUBREG, arriving_mode, target, 0);
1622 else
1623 reg_to_map = target;
1625 /* Usually, the result value is the machine's return register.
1626 Sometimes it may be a pseudo. Handle both cases. */
1627 if (REG_FUNCTION_VALUE_P (loc))
1628 map->inline_target = reg_to_map;
1629 else
1630 map->reg_map[REGNO (loc)] = reg_to_map;
1633 /* Make new label equivalences for the labels in the called function. */
1634 for (i = min_labelno; i < max_labelno; i++)
1635 map->label_map[i] = gen_label_rtx ();
1637 /* Perform postincrements before actually calling the function. */
1638 emit_queue ();
1640 /* Clean up stack so that variables might have smaller offsets. */
1641 do_pending_stack_adjust ();
1643 /* Save a copy of the location of const_equiv_map for mark_stores, called
1644 via note_stores. */
1645 global_const_equiv_map = map->const_equiv_map;
1646 global_const_equiv_map_size = map->const_equiv_map_size;
1648 /* If the called function does an alloca, save and restore the
1649 stack pointer around the call. This saves stack space, but
1650 also is required if this inline is being done between two
1651 pushes. */
1652 if (FUNCTION_FLAGS (header) & FUNCTION_FLAGS_CALLS_ALLOCA)
1653 emit_stack_save (SAVE_BLOCK, &stack_save, NULL_RTX);
1655 /* Now copy the insns one by one. Do this in two passes, first the insns and
1656 then their REG_NOTES, just like save_for_inline. */
1658 /* This loop is very similar to the loop in copy_loop_body in unroll.c. */
1660 for (insn = insns; insn; insn = NEXT_INSN (insn))
1662 rtx copy, pattern, set;
1664 map->orig_asm_operands_vector = 0;
1666 switch (GET_CODE (insn))
1668 case INSN:
1669 pattern = PATTERN (insn);
1670 set = single_set (insn);
1671 copy = 0;
1672 if (GET_CODE (pattern) == USE
1673 && GET_CODE (XEXP (pattern, 0)) == REG
1674 && REG_FUNCTION_VALUE_P (XEXP (pattern, 0)))
1675 /* The (USE (REG n)) at return from the function should
1676 be ignored since we are changing (REG n) into
1677 inline_target. */
1678 break;
1680 /* Ignore setting a function value that we don't want to use. */
1681 if (map->inline_target == 0
1682 && set != 0
1683 && GET_CODE (SET_DEST (set)) == REG
1684 && REG_FUNCTION_VALUE_P (SET_DEST (set)))
1686 if (volatile_refs_p (SET_SRC (set)))
1688 rtx new_set;
1690 /* If we must not delete the source,
1691 load it into a new temporary. */
1692 copy = emit_insn (copy_rtx_and_substitute (pattern, map));
1694 new_set = single_set (copy);
1695 if (new_set == 0)
1696 abort ();
1698 SET_DEST (new_set)
1699 = gen_reg_rtx (GET_MODE (SET_DEST (new_set)));
1701 else
1702 break;
1705 /* If this is setting the static chain rtx, omit it. */
1706 else if (static_chain_value != 0
1707 && set != 0
1708 && GET_CODE (SET_DEST (set)) == REG
1709 && rtx_equal_p (SET_DEST (set),
1710 static_chain_incoming_rtx))
1711 break;
1713 /* If this is setting the static chain pseudo, set it from
1714 the value we want to give it instead. */
1715 else if (static_chain_value != 0
1716 && set != 0
1717 && rtx_equal_p (SET_SRC (set),
1718 static_chain_incoming_rtx))
1720 rtx newdest = copy_rtx_and_substitute (SET_DEST (set), map);
1722 copy = emit_move_insn (newdest, static_chain_value);
1723 static_chain_value = 0;
1725 else
1726 copy = emit_insn (copy_rtx_and_substitute (pattern, map));
1727 /* REG_NOTES will be copied later. */
1729 #ifdef HAVE_cc0
1730 /* If this insn is setting CC0, it may need to look at
1731 the insn that uses CC0 to see what type of insn it is.
1732 In that case, the call to recog via validate_change will
1733 fail. So don't substitute constants here. Instead,
1734 do it when we emit the following insn.
1736 For example, see the pyr.md file. That machine has signed and
1737 unsigned compares. The compare patterns must check the
1738 following branch insn to see which what kind of compare to
1739 emit.
1741 If the previous insn set CC0, substitute constants on it as
1742 well. */
1743 if (sets_cc0_p (PATTERN (copy)) != 0)
1744 cc0_insn = copy;
1745 else
1747 if (cc0_insn)
1748 try_constants (cc0_insn, map);
1749 cc0_insn = 0;
1750 try_constants (copy, map);
1752 #else
1753 try_constants (copy, map);
1754 #endif
1755 break;
1757 case JUMP_INSN:
1758 if (GET_CODE (PATTERN (insn)) == RETURN)
1760 if (local_return_label == 0)
1761 local_return_label = gen_label_rtx ();
1762 pattern = gen_jump (local_return_label);
1764 else
1765 pattern = copy_rtx_and_substitute (PATTERN (insn), map);
1767 copy = emit_jump_insn (pattern);
1769 #ifdef HAVE_cc0
1770 if (cc0_insn)
1771 try_constants (cc0_insn, map);
1772 cc0_insn = 0;
1773 #endif
1774 try_constants (copy, map);
1776 /* If this used to be a conditional jump insn but whose branch
1777 direction is now know, we must do something special. */
1778 if (condjump_p (insn) && ! simplejump_p (insn) && map->last_pc_value)
1780 #ifdef HAVE_cc0
1781 /* The previous insn set cc0 for us. So delete it. */
1782 delete_insn (PREV_INSN (copy));
1783 #endif
1785 /* If this is now a no-op, delete it. */
1786 if (map->last_pc_value == pc_rtx)
1788 delete_insn (copy);
1789 copy = 0;
1791 else
1792 /* Otherwise, this is unconditional jump so we must put a
1793 BARRIER after it. We could do some dead code elimination
1794 here, but jump.c will do it just as well. */
1795 emit_barrier ();
1797 break;
1799 case CALL_INSN:
1800 pattern = copy_rtx_and_substitute (PATTERN (insn), map);
1801 copy = emit_call_insn (pattern);
1803 /* Because the USAGE information potentially contains objects other
1804 than hard registers, we need to copy it. */
1805 CALL_INSN_FUNCTION_USAGE (copy) =
1806 copy_rtx_and_substitute (CALL_INSN_FUNCTION_USAGE (insn), map);
1808 #ifdef HAVE_cc0
1809 if (cc0_insn)
1810 try_constants (cc0_insn, map);
1811 cc0_insn = 0;
1812 #endif
1813 try_constants (copy, map);
1815 /* Be lazy and assume CALL_INSNs clobber all hard registers. */
1816 for (i = 0; i < FIRST_PSEUDO_REGISTER; i++)
1817 map->const_equiv_map[i] = 0;
1818 break;
1820 case CODE_LABEL:
1821 copy = emit_label (map->label_map[CODE_LABEL_NUMBER (insn)]);
1822 LABEL_NAME (copy) = LABEL_NAME (insn);
1823 map->const_age++;
1824 break;
1826 case BARRIER:
1827 copy = emit_barrier ();
1828 break;
1830 case NOTE:
1831 /* It is important to discard function-end and function-beg notes,
1832 so we have only one of each in the current function.
1833 Also, NOTE_INSN_DELETED notes aren't useful (save_for_inline
1834 deleted these in the copy used for continuing compilation,
1835 not the copy used for inlining). */
1836 if (NOTE_LINE_NUMBER (insn) != NOTE_INSN_FUNCTION_END
1837 && NOTE_LINE_NUMBER (insn) != NOTE_INSN_FUNCTION_BEG
1838 && NOTE_LINE_NUMBER (insn) != NOTE_INSN_DELETED)
1839 copy = emit_note (NOTE_SOURCE_FILE (insn), NOTE_LINE_NUMBER (insn));
1840 else
1841 copy = 0;
1842 break;
1844 default:
1845 abort ();
1846 break;
1849 if (copy)
1850 RTX_INTEGRATED_P (copy) = 1;
1852 map->insn_map[INSN_UID (insn)] = copy;
1855 /* Now copy the REG_NOTES. Increment const_age, so that only constants
1856 from parameters can be substituted in. These are the only ones that
1857 are valid across the entire function. */
1858 map->const_age++;
1859 for (insn = insns; insn; insn = NEXT_INSN (insn))
1860 if (GET_RTX_CLASS (GET_CODE (insn)) == 'i'
1861 && map->insn_map[INSN_UID (insn)]
1862 && REG_NOTES (insn))
1864 rtx tem = copy_rtx_and_substitute (REG_NOTES (insn), map);
1865 /* We must also do subst_constants, in case one of our parameters
1866 has const type and constant value. */
1867 subst_constants (&tem, NULL_RTX, map);
1868 apply_change_group ();
1869 REG_NOTES (map->insn_map[INSN_UID (insn)]) = tem;
1872 if (local_return_label)
1873 emit_label (local_return_label);
1875 /* Restore the stack pointer if we saved it above. */
1876 if (FUNCTION_FLAGS (header) & FUNCTION_FLAGS_CALLS_ALLOCA)
1877 emit_stack_restore (SAVE_BLOCK, stack_save, NULL_RTX);
1879 /* Make copies of the decls of the symbols in the inline function, so that
1880 the copies of the variables get declared in the current function. Set
1881 up things so that lookup_static_chain knows that to interpret registers
1882 in SAVE_EXPRs for TYPE_SIZEs as local. */
1884 inline_function_decl = fndecl;
1885 integrate_parm_decls (DECL_ARGUMENTS (fndecl), map, arg_vector);
1886 integrate_decl_tree ((tree) ORIGINAL_DECL_INITIAL (header), 0, map);
1887 inline_function_decl = 0;
1889 /* End the scope containing the copied formal parameter variables
1890 and copied LABEL_DECLs. */
1892 expand_end_bindings (getdecls (), 1, 1);
1893 block = poplevel (1, 1, 0);
1894 BLOCK_ABSTRACT_ORIGIN (block) = (DECL_ABSTRACT_ORIGIN (fndecl) == NULL
1895 ? fndecl : DECL_ABSTRACT_ORIGIN (fndecl));
1896 poplevel (0, 0, 0);
1897 emit_line_note (input_filename, lineno);
1899 if (structure_value_addr)
1901 target = gen_rtx (MEM, TYPE_MODE (type),
1902 memory_address (TYPE_MODE (type), structure_value_addr));
1903 MEM_IN_STRUCT_P (target) = 1;
1905 return target;
1908 /* Given a chain of PARM_DECLs, ARGS, copy each decl into a VAR_DECL,
1909 push all of those decls and give each one the corresponding home. */
1911 static void
1912 integrate_parm_decls (args, map, arg_vector)
1913 tree args;
1914 struct inline_remap *map;
1915 rtvec arg_vector;
1917 register tree tail;
1918 register int i;
1920 for (tail = args, i = 0; tail; tail = TREE_CHAIN (tail), i++)
1922 register tree decl = build_decl (VAR_DECL, DECL_NAME (tail),
1923 TREE_TYPE (tail));
1924 rtx new_decl_rtl
1925 = copy_rtx_and_substitute (RTVEC_ELT (arg_vector, i), map);
1927 DECL_ARG_TYPE (decl) = DECL_ARG_TYPE (tail);
1928 /* We really should be setting DECL_INCOMING_RTL to something reasonable
1929 here, but that's going to require some more work. */
1930 /* DECL_INCOMING_RTL (decl) = ?; */
1931 /* These args would always appear unused, if not for this. */
1932 TREE_USED (decl) = 1;
1933 /* Prevent warning for shadowing with these. */
1934 DECL_ABSTRACT_ORIGIN (decl) = tail;
1935 pushdecl (decl);
1936 /* Fully instantiate the address with the equivalent form so that the
1937 debugging information contains the actual register, instead of the
1938 virtual register. Do this by not passing an insn to
1939 subst_constants. */
1940 subst_constants (&new_decl_rtl, NULL_RTX, map);
1941 apply_change_group ();
1942 DECL_RTL (decl) = new_decl_rtl;
1946 /* Given a BLOCK node LET, push decls and levels so as to construct in the
1947 current function a tree of contexts isomorphic to the one that is given.
1949 LEVEL indicates how far down into the BLOCK tree is the node we are
1950 currently traversing. It is always zero except for recursive calls.
1952 MAP, if nonzero, is a pointer to an inline_remap map which indicates how
1953 registers used in the DECL_RTL field should be remapped. If it is zero,
1954 no mapping is necessary. */
1956 static void
1957 integrate_decl_tree (let, level, map)
1958 tree let;
1959 int level;
1960 struct inline_remap *map;
1962 tree t, node;
1964 if (level > 0)
1965 pushlevel (0);
1967 for (t = BLOCK_VARS (let); t; t = TREE_CHAIN (t))
1969 tree d;
1971 push_obstacks_nochange ();
1972 saveable_allocation ();
1973 d = copy_node (t);
1974 pop_obstacks ();
1976 if (DECL_RTL (t) != 0)
1978 DECL_RTL (d) = copy_rtx_and_substitute (DECL_RTL (t), map);
1979 /* Fully instantiate the address with the equivalent form so that the
1980 debugging information contains the actual register, instead of the
1981 virtual register. Do this by not passing an insn to
1982 subst_constants. */
1983 subst_constants (&DECL_RTL (d), NULL_RTX, map);
1984 apply_change_group ();
1986 /* These args would always appear unused, if not for this. */
1987 TREE_USED (d) = 1;
1988 /* Prevent warning for shadowing with these. */
1989 DECL_ABSTRACT_ORIGIN (d) = t;
1991 if (DECL_LANG_SPECIFIC (d))
1992 copy_lang_decl (d);
1994 pushdecl (d);
1997 for (t = BLOCK_SUBBLOCKS (let); t; t = TREE_CHAIN (t))
1998 integrate_decl_tree (t, level + 1, map);
2000 if (level > 0)
2002 node = poplevel (1, 0, 0);
2003 if (node)
2005 TREE_USED (node) = TREE_USED (let);
2006 BLOCK_ABSTRACT_ORIGIN (node) = let;
2011 /* Create a new copy of an rtx.
2012 Recursively copies the operands of the rtx,
2013 except for those few rtx codes that are sharable.
2015 We always return an rtx that is similar to that incoming rtx, with the
2016 exception of possibly changing a REG to a SUBREG or vice versa. No
2017 rtl is ever emitted.
2019 Handle constants that need to be placed in the constant pool by
2020 calling `force_const_mem'. */
2023 copy_rtx_and_substitute (orig, map)
2024 register rtx orig;
2025 struct inline_remap *map;
2027 register rtx copy, temp;
2028 register int i, j;
2029 register RTX_CODE code;
2030 register enum machine_mode mode;
2031 register char *format_ptr;
2032 int regno;
2034 if (orig == 0)
2035 return 0;
2037 code = GET_CODE (orig);
2038 mode = GET_MODE (orig);
2040 switch (code)
2042 case REG:
2043 /* If the stack pointer register shows up, it must be part of
2044 stack-adjustments (*not* because we eliminated the frame pointer!).
2045 Small hard registers are returned as-is. Pseudo-registers
2046 go through their `reg_map'. */
2047 regno = REGNO (orig);
2048 if (regno <= LAST_VIRTUAL_REGISTER)
2050 /* Some hard registers are also mapped,
2051 but others are not translated. */
2052 if (map->reg_map[regno] != 0)
2053 return map->reg_map[regno];
2055 /* If this is the virtual frame pointer, make space in current
2056 function's stack frame for the stack frame of the inline function.
2058 Copy the address of this area into a pseudo. Map
2059 virtual_stack_vars_rtx to this pseudo and set up a constant
2060 equivalence for it to be the address. This will substitute the
2061 address into insns where it can be substituted and use the new
2062 pseudo where it can't. */
2063 if (regno == VIRTUAL_STACK_VARS_REGNUM)
2065 rtx loc, seq;
2066 int size = DECL_FRAME_SIZE (map->fndecl);
2067 int rounded;
2069 start_sequence ();
2070 loc = assign_stack_temp (BLKmode, size, 1);
2071 loc = XEXP (loc, 0);
2072 #ifdef FRAME_GROWS_DOWNWARD
2073 /* In this case, virtual_stack_vars_rtx points to one byte
2074 higher than the top of the frame area. So compute the offset
2075 to one byte higher than our substitute frame.
2076 Keep the fake frame pointer aligned like a real one. */
2077 rounded = CEIL_ROUND (size, BIGGEST_ALIGNMENT / BITS_PER_UNIT);
2078 loc = plus_constant (loc, rounded);
2079 #endif
2080 map->reg_map[regno] = temp
2081 = force_reg (Pmode, force_operand (loc, NULL_RTX));
2083 if (REGNO (temp) < map->const_equiv_map_size)
2085 map->const_equiv_map[REGNO (temp)] = loc;
2086 map->const_age_map[REGNO (temp)] = CONST_AGE_PARM;
2089 seq = gen_sequence ();
2090 end_sequence ();
2091 emit_insn_after (seq, map->insns_at_start);
2092 return temp;
2094 else if (regno == VIRTUAL_INCOMING_ARGS_REGNUM)
2096 /* Do the same for a block to contain any arguments referenced
2097 in memory. */
2098 rtx loc, seq;
2099 int size = FUNCTION_ARGS_SIZE (DECL_SAVED_INSNS (map->fndecl));
2101 start_sequence ();
2102 loc = assign_stack_temp (BLKmode, size, 1);
2103 loc = XEXP (loc, 0);
2104 /* When arguments grow downward, the virtual incoming
2105 args pointer points to the top of the argument block,
2106 so the remapped location better do the same. */
2107 #ifdef ARGS_GROW_DOWNWARD
2108 loc = plus_constant (loc, size);
2109 #endif
2110 map->reg_map[regno] = temp
2111 = force_reg (Pmode, force_operand (loc, NULL_RTX));
2113 if (REGNO (temp) < map->const_equiv_map_size)
2115 map->const_equiv_map[REGNO (temp)] = loc;
2116 map->const_age_map[REGNO (temp)] = CONST_AGE_PARM;
2119 seq = gen_sequence ();
2120 end_sequence ();
2121 emit_insn_after (seq, map->insns_at_start);
2122 return temp;
2124 else if (REG_FUNCTION_VALUE_P (orig))
2126 /* This is a reference to the function return value. If
2127 the function doesn't have a return value, error. If the
2128 mode doesn't agree, make a SUBREG. */
2129 if (map->inline_target == 0)
2130 /* Must be unrolling loops or replicating code if we
2131 reach here, so return the register unchanged. */
2132 return orig;
2133 else if (mode != GET_MODE (map->inline_target))
2134 return gen_lowpart (mode, map->inline_target);
2135 else
2136 return map->inline_target;
2138 return orig;
2140 if (map->reg_map[regno] == NULL)
2142 map->reg_map[regno] = gen_reg_rtx (mode);
2143 REG_USERVAR_P (map->reg_map[regno]) = REG_USERVAR_P (orig);
2144 REG_LOOP_TEST_P (map->reg_map[regno]) = REG_LOOP_TEST_P (orig);
2145 RTX_UNCHANGING_P (map->reg_map[regno]) = RTX_UNCHANGING_P (orig);
2146 /* A reg with REG_FUNCTION_VALUE_P true will never reach here. */
2148 return map->reg_map[regno];
2150 case SUBREG:
2151 copy = copy_rtx_and_substitute (SUBREG_REG (orig), map);
2152 /* SUBREG is ordinary, but don't make nested SUBREGs. */
2153 if (GET_CODE (copy) == SUBREG)
2154 return gen_rtx (SUBREG, GET_MODE (orig), SUBREG_REG (copy),
2155 SUBREG_WORD (orig) + SUBREG_WORD (copy));
2156 else if (GET_CODE (copy) == CONCAT)
2157 return (subreg_realpart_p (orig) ? XEXP (copy, 0) : XEXP (copy, 1));
2158 else
2159 return gen_rtx (SUBREG, GET_MODE (orig), copy,
2160 SUBREG_WORD (orig));
2162 case USE:
2163 case CLOBBER:
2164 /* USE and CLOBBER are ordinary, but we convert (use (subreg foo))
2165 to (use foo) if the original insn didn't have a subreg.
2166 Removing the subreg distorts the VAX movstrhi pattern
2167 by changing the mode of an operand. */
2168 copy = copy_rtx_and_substitute (XEXP (orig, 0), map);
2169 if (GET_CODE (copy) == SUBREG && GET_CODE (XEXP (orig, 0)) != SUBREG)
2170 copy = SUBREG_REG (copy);
2171 return gen_rtx (code, VOIDmode, copy);
2173 case CODE_LABEL:
2174 LABEL_PRESERVE_P (map->label_map[CODE_LABEL_NUMBER (orig)])
2175 = LABEL_PRESERVE_P (orig);
2176 return map->label_map[CODE_LABEL_NUMBER (orig)];
2178 case LABEL_REF:
2179 copy = gen_rtx (LABEL_REF, mode,
2180 LABEL_REF_NONLOCAL_P (orig) ? XEXP (orig, 0)
2181 : map->label_map[CODE_LABEL_NUMBER (XEXP (orig, 0))]);
2182 LABEL_OUTSIDE_LOOP_P (copy) = LABEL_OUTSIDE_LOOP_P (orig);
2184 /* The fact that this label was previously nonlocal does not mean
2185 it still is, so we must check if it is within the range of
2186 this function's labels. */
2187 LABEL_REF_NONLOCAL_P (copy)
2188 = (LABEL_REF_NONLOCAL_P (orig)
2189 && ! (CODE_LABEL_NUMBER (XEXP (copy, 0)) >= get_first_label_num ()
2190 && CODE_LABEL_NUMBER (XEXP (copy, 0)) < max_label_num ()));
2192 /* If we have made a nonlocal label local, it means that this
2193 inlined call will be refering to our nonlocal goto handler.
2194 So make sure we create one for this block; we normally would
2195 not since this is not otherwise considered a "call". */
2196 if (LABEL_REF_NONLOCAL_P (orig) && ! LABEL_REF_NONLOCAL_P (copy))
2197 function_call_count++;
2199 return copy;
2201 case PC:
2202 case CC0:
2203 case CONST_INT:
2204 return orig;
2206 case SYMBOL_REF:
2207 /* Symbols which represent the address of a label stored in the constant
2208 pool must be modified to point to a constant pool entry for the
2209 remapped label. Otherwise, symbols are returned unchanged. */
2210 if (CONSTANT_POOL_ADDRESS_P (orig))
2212 rtx constant = get_pool_constant (orig);
2213 if (GET_CODE (constant) == LABEL_REF)
2214 return XEXP (force_const_mem (Pmode,
2215 copy_rtx_and_substitute (constant,
2216 map)),
2220 return orig;
2222 case CONST_DOUBLE:
2223 /* We have to make a new copy of this CONST_DOUBLE because don't want
2224 to use the old value of CONST_DOUBLE_MEM. Also, this may be a
2225 duplicate of a CONST_DOUBLE we have already seen. */
2226 if (GET_MODE_CLASS (GET_MODE (orig)) == MODE_FLOAT)
2228 REAL_VALUE_TYPE d;
2230 REAL_VALUE_FROM_CONST_DOUBLE (d, orig);
2231 return CONST_DOUBLE_FROM_REAL_VALUE (d, GET_MODE (orig));
2233 else
2234 return immed_double_const (CONST_DOUBLE_LOW (orig),
2235 CONST_DOUBLE_HIGH (orig), VOIDmode);
2237 case CONST:
2238 /* Make new constant pool entry for a constant
2239 that was in the pool of the inline function. */
2240 if (RTX_INTEGRATED_P (orig))
2242 /* If this was an address of a constant pool entry that itself
2243 had to be placed in the constant pool, it might not be a
2244 valid address. So the recursive call below might turn it
2245 into a register. In that case, it isn't a constant any
2246 more, so return it. This has the potential of changing a
2247 MEM into a REG, but we'll assume that it safe. */
2248 temp = copy_rtx_and_substitute (XEXP (orig, 0), map);
2249 if (! CONSTANT_P (temp))
2250 return temp;
2251 return validize_mem (force_const_mem (GET_MODE (orig), temp));
2253 break;
2255 case ADDRESS:
2256 /* If from constant pool address, make new constant pool entry and
2257 return its address. */
2258 if (! RTX_INTEGRATED_P (orig))
2259 abort ();
2261 temp = force_const_mem (GET_MODE (orig),
2262 copy_rtx_and_substitute (XEXP (orig, 0), map));
2264 #if 0
2265 /* Legitimizing the address here is incorrect.
2267 The only ADDRESS rtx's that can reach here are ones created by
2268 save_constants. Hence the operand of the ADDRESS is always valid
2269 in this position of the instruction, since the original rtx without
2270 the ADDRESS was valid.
2272 The reason we don't legitimize the address here is that on the
2273 Sparc, the caller may have a (high ...) surrounding this ADDRESS.
2274 This code forces the operand of the address to a register, which
2275 fails because we can not take the HIGH part of a register.
2277 Also, change_address may create new registers. These registers
2278 will not have valid reg_map entries. This can cause try_constants()
2279 to fail because assumes that all registers in the rtx have valid
2280 reg_map entries, and it may end up replacing one of these new
2281 registers with junk. */
2283 if (! memory_address_p (GET_MODE (temp), XEXP (temp, 0)))
2284 temp = change_address (temp, GET_MODE (temp), XEXP (temp, 0));
2285 #endif
2287 return XEXP (temp, 0);
2289 case ASM_OPERANDS:
2290 /* If a single asm insn contains multiple output operands
2291 then it contains multiple ASM_OPERANDS rtx's that share operand 3.
2292 We must make sure that the copied insn continues to share it. */
2293 if (map->orig_asm_operands_vector == XVEC (orig, 3))
2295 copy = rtx_alloc (ASM_OPERANDS);
2296 copy->volatil = orig->volatil;
2297 XSTR (copy, 0) = XSTR (orig, 0);
2298 XSTR (copy, 1) = XSTR (orig, 1);
2299 XINT (copy, 2) = XINT (orig, 2);
2300 XVEC (copy, 3) = map->copy_asm_operands_vector;
2301 XVEC (copy, 4) = map->copy_asm_constraints_vector;
2302 XSTR (copy, 5) = XSTR (orig, 5);
2303 XINT (copy, 6) = XINT (orig, 6);
2304 return copy;
2306 break;
2308 case CALL:
2309 /* This is given special treatment because the first
2310 operand of a CALL is a (MEM ...) which may get
2311 forced into a register for cse. This is undesirable
2312 if function-address cse isn't wanted or if we won't do cse. */
2313 #ifndef NO_FUNCTION_CSE
2314 if (! (optimize && ! flag_no_function_cse))
2315 #endif
2316 return gen_rtx (CALL, GET_MODE (orig),
2317 gen_rtx (MEM, GET_MODE (XEXP (orig, 0)),
2318 copy_rtx_and_substitute (XEXP (XEXP (orig, 0), 0), map)),
2319 copy_rtx_and_substitute (XEXP (orig, 1), map));
2320 break;
2322 #if 0
2323 /* Must be ifdefed out for loop unrolling to work. */
2324 case RETURN:
2325 abort ();
2326 #endif
2328 case SET:
2329 /* If this is setting fp or ap, it means that we have a nonlocal goto.
2330 Don't alter that.
2331 If the nonlocal goto is into the current function,
2332 this will result in unnecessarily bad code, but should work. */
2333 if (SET_DEST (orig) == virtual_stack_vars_rtx
2334 || SET_DEST (orig) == virtual_incoming_args_rtx)
2335 return gen_rtx (SET, VOIDmode, SET_DEST (orig),
2336 copy_rtx_and_substitute (SET_SRC (orig), map));
2337 break;
2339 case MEM:
2340 copy = rtx_alloc (MEM);
2341 PUT_MODE (copy, mode);
2342 XEXP (copy, 0) = copy_rtx_and_substitute (XEXP (orig, 0), map);
2343 MEM_IN_STRUCT_P (copy) = MEM_IN_STRUCT_P (orig);
2344 MEM_VOLATILE_P (copy) = MEM_VOLATILE_P (orig);
2346 /* If doing function inlining, this MEM might not be const in the
2347 function that it is being inlined into, and thus may not be
2348 unchanging after function inlining. Constant pool references are
2349 handled elsewhere, so this doesn't lose RTX_UNCHANGING_P bits
2350 for them. */
2351 if (! map->integrating)
2352 RTX_UNCHANGING_P (copy) = RTX_UNCHANGING_P (orig);
2354 return copy;
2357 copy = rtx_alloc (code);
2358 PUT_MODE (copy, mode);
2359 copy->in_struct = orig->in_struct;
2360 copy->volatil = orig->volatil;
2361 copy->unchanging = orig->unchanging;
2363 format_ptr = GET_RTX_FORMAT (GET_CODE (copy));
2365 for (i = 0; i < GET_RTX_LENGTH (GET_CODE (copy)); i++)
2367 switch (*format_ptr++)
2369 case '0':
2370 break;
2372 case 'e':
2373 XEXP (copy, i) = copy_rtx_and_substitute (XEXP (orig, i), map);
2374 break;
2376 case 'u':
2377 /* Change any references to old-insns to point to the
2378 corresponding copied insns. */
2379 XEXP (copy, i) = map->insn_map[INSN_UID (XEXP (orig, i))];
2380 break;
2382 case 'E':
2383 XVEC (copy, i) = XVEC (orig, i);
2384 if (XVEC (orig, i) != NULL && XVECLEN (orig, i) != 0)
2386 XVEC (copy, i) = rtvec_alloc (XVECLEN (orig, i));
2387 for (j = 0; j < XVECLEN (copy, i); j++)
2388 XVECEXP (copy, i, j)
2389 = copy_rtx_and_substitute (XVECEXP (orig, i, j), map);
2391 break;
2393 case 'w':
2394 XWINT (copy, i) = XWINT (orig, i);
2395 break;
2397 case 'i':
2398 XINT (copy, i) = XINT (orig, i);
2399 break;
2401 case 's':
2402 XSTR (copy, i) = XSTR (orig, i);
2403 break;
2405 default:
2406 abort ();
2410 if (code == ASM_OPERANDS && map->orig_asm_operands_vector == 0)
2412 map->orig_asm_operands_vector = XVEC (orig, 3);
2413 map->copy_asm_operands_vector = XVEC (copy, 3);
2414 map->copy_asm_constraints_vector = XVEC (copy, 4);
2417 return copy;
2420 /* Substitute known constant values into INSN, if that is valid. */
2422 void
2423 try_constants (insn, map)
2424 rtx insn;
2425 struct inline_remap *map;
2427 int i;
2429 map->num_sets = 0;
2430 subst_constants (&PATTERN (insn), insn, map);
2432 /* Apply the changes if they are valid; otherwise discard them. */
2433 apply_change_group ();
2435 /* Show we don't know the value of anything stored or clobbered. */
2436 note_stores (PATTERN (insn), mark_stores);
2437 map->last_pc_value = 0;
2438 #ifdef HAVE_cc0
2439 map->last_cc0_value = 0;
2440 #endif
2442 /* Set up any constant equivalences made in this insn. */
2443 for (i = 0; i < map->num_sets; i++)
2445 if (GET_CODE (map->equiv_sets[i].dest) == REG)
2447 int regno = REGNO (map->equiv_sets[i].dest);
2449 if (regno < map->const_equiv_map_size
2450 && (map->const_equiv_map[regno] == 0
2451 /* Following clause is a hack to make case work where GNU C++
2452 reassigns a variable to make cse work right. */
2453 || ! rtx_equal_p (map->const_equiv_map[regno],
2454 map->equiv_sets[i].equiv)))
2456 map->const_equiv_map[regno] = map->equiv_sets[i].equiv;
2457 map->const_age_map[regno] = map->const_age;
2460 else if (map->equiv_sets[i].dest == pc_rtx)
2461 map->last_pc_value = map->equiv_sets[i].equiv;
2462 #ifdef HAVE_cc0
2463 else if (map->equiv_sets[i].dest == cc0_rtx)
2464 map->last_cc0_value = map->equiv_sets[i].equiv;
2465 #endif
2469 /* Substitute known constants for pseudo regs in the contents of LOC,
2470 which are part of INSN.
2471 If INSN is zero, the substitution should always be done (this is used to
2472 update DECL_RTL).
2473 These changes are taken out by try_constants if the result is not valid.
2475 Note that we are more concerned with determining when the result of a SET
2476 is a constant, for further propagation, than actually inserting constants
2477 into insns; cse will do the latter task better.
2479 This function is also used to adjust address of items previously addressed
2480 via the virtual stack variable or virtual incoming arguments registers. */
2482 static void
2483 subst_constants (loc, insn, map)
2484 rtx *loc;
2485 rtx insn;
2486 struct inline_remap *map;
2488 rtx x = *loc;
2489 register int i;
2490 register enum rtx_code code;
2491 register char *format_ptr;
2492 int num_changes = num_validated_changes ();
2493 rtx new = 0;
2494 enum machine_mode op0_mode;
2496 code = GET_CODE (x);
2498 switch (code)
2500 case PC:
2501 case CONST_INT:
2502 case CONST_DOUBLE:
2503 case SYMBOL_REF:
2504 case CONST:
2505 case LABEL_REF:
2506 case ADDRESS:
2507 return;
2509 #ifdef HAVE_cc0
2510 case CC0:
2511 validate_change (insn, loc, map->last_cc0_value, 1);
2512 return;
2513 #endif
2515 case USE:
2516 case CLOBBER:
2517 /* The only thing we can do with a USE or CLOBBER is possibly do
2518 some substitutions in a MEM within it. */
2519 if (GET_CODE (XEXP (x, 0)) == MEM)
2520 subst_constants (&XEXP (XEXP (x, 0), 0), insn, map);
2521 return;
2523 case REG:
2524 /* Substitute for parms and known constants. Don't replace
2525 hard regs used as user variables with constants. */
2527 int regno = REGNO (x);
2529 if (! (regno < FIRST_PSEUDO_REGISTER && REG_USERVAR_P (x))
2530 && regno < map->const_equiv_map_size
2531 && map->const_equiv_map[regno] != 0
2532 && map->const_age_map[regno] >= map->const_age)
2533 validate_change (insn, loc, map->const_equiv_map[regno], 1);
2534 return;
2537 case SUBREG:
2538 /* SUBREG applied to something other than a reg
2539 should be treated as ordinary, since that must
2540 be a special hack and we don't know how to treat it specially.
2541 Consider for example mulsidi3 in m68k.md.
2542 Ordinary SUBREG of a REG needs this special treatment. */
2543 if (GET_CODE (SUBREG_REG (x)) == REG)
2545 rtx inner = SUBREG_REG (x);
2546 rtx new = 0;
2548 /* We can't call subst_constants on &SUBREG_REG (x) because any
2549 constant or SUBREG wouldn't be valid inside our SUBEG. Instead,
2550 see what is inside, try to form the new SUBREG and see if that is
2551 valid. We handle two cases: extracting a full word in an
2552 integral mode and extracting the low part. */
2553 subst_constants (&inner, NULL_RTX, map);
2555 if (GET_MODE_CLASS (GET_MODE (x)) == MODE_INT
2556 && GET_MODE_SIZE (GET_MODE (x)) == UNITS_PER_WORD
2557 && GET_MODE (SUBREG_REG (x)) != VOIDmode)
2558 new = operand_subword (inner, SUBREG_WORD (x), 0,
2559 GET_MODE (SUBREG_REG (x)));
2561 if (new == 0 && subreg_lowpart_p (x))
2562 new = gen_lowpart_common (GET_MODE (x), inner);
2564 if (new)
2565 validate_change (insn, loc, new, 1);
2567 return;
2569 break;
2571 case MEM:
2572 subst_constants (&XEXP (x, 0), insn, map);
2574 /* If a memory address got spoiled, change it back. */
2575 if (insn != 0 && num_validated_changes () != num_changes
2576 && !memory_address_p (GET_MODE (x), XEXP (x, 0)))
2577 cancel_changes (num_changes);
2578 return;
2580 case SET:
2582 /* Substitute constants in our source, and in any arguments to a
2583 complex (e..g, ZERO_EXTRACT) destination, but not in the destination
2584 itself. */
2585 rtx *dest_loc = &SET_DEST (x);
2586 rtx dest = *dest_loc;
2587 rtx src, tem;
2589 subst_constants (&SET_SRC (x), insn, map);
2590 src = SET_SRC (x);
2592 while (GET_CODE (*dest_loc) == ZERO_EXTRACT
2593 /* By convention, we always use ZERO_EXTRACT in the dest. */
2594 /* || GET_CODE (*dest_loc) == SIGN_EXTRACT */
2595 || GET_CODE (*dest_loc) == SUBREG
2596 || GET_CODE (*dest_loc) == STRICT_LOW_PART)
2598 if (GET_CODE (*dest_loc) == ZERO_EXTRACT)
2600 subst_constants (&XEXP (*dest_loc, 1), insn, map);
2601 subst_constants (&XEXP (*dest_loc, 2), insn, map);
2603 dest_loc = &XEXP (*dest_loc, 0);
2606 /* Do substitute in the address of a destination in memory. */
2607 if (GET_CODE (*dest_loc) == MEM)
2608 subst_constants (&XEXP (*dest_loc, 0), insn, map);
2610 /* Check for the case of DEST a SUBREG, both it and the underlying
2611 register are less than one word, and the SUBREG has the wider mode.
2612 In the case, we are really setting the underlying register to the
2613 source converted to the mode of DEST. So indicate that. */
2614 if (GET_CODE (dest) == SUBREG
2615 && GET_MODE_SIZE (GET_MODE (dest)) <= UNITS_PER_WORD
2616 && GET_MODE_SIZE (GET_MODE (SUBREG_REG (dest))) <= UNITS_PER_WORD
2617 && (GET_MODE_SIZE (GET_MODE (SUBREG_REG (dest)))
2618 <= GET_MODE_SIZE (GET_MODE (dest)))
2619 && (tem = gen_lowpart_if_possible (GET_MODE (SUBREG_REG (dest)),
2620 src)))
2621 src = tem, dest = SUBREG_REG (dest);
2623 /* If storing a recognizable value save it for later recording. */
2624 if ((map->num_sets < MAX_RECOG_OPERANDS)
2625 && (CONSTANT_P (src)
2626 || (GET_CODE (src) == REG
2627 && (REGNO (src) == VIRTUAL_INCOMING_ARGS_REGNUM
2628 || REGNO (src) == VIRTUAL_STACK_VARS_REGNUM))
2629 || (GET_CODE (src) == PLUS
2630 && GET_CODE (XEXP (src, 0)) == REG
2631 && (REGNO (XEXP (src, 0)) == VIRTUAL_INCOMING_ARGS_REGNUM
2632 || REGNO (XEXP (src, 0)) == VIRTUAL_STACK_VARS_REGNUM)
2633 && CONSTANT_P (XEXP (src, 1)))
2634 || GET_CODE (src) == COMPARE
2635 #ifdef HAVE_cc0
2636 || dest == cc0_rtx
2637 #endif
2638 || (dest == pc_rtx
2639 && (src == pc_rtx || GET_CODE (src) == RETURN
2640 || GET_CODE (src) == LABEL_REF))))
2642 /* Normally, this copy won't do anything. But, if SRC is a COMPARE
2643 it will cause us to save the COMPARE with any constants
2644 substituted, which is what we want for later. */
2645 map->equiv_sets[map->num_sets].equiv = copy_rtx (src);
2646 map->equiv_sets[map->num_sets++].dest = dest;
2649 return;
2653 format_ptr = GET_RTX_FORMAT (code);
2655 /* If the first operand is an expression, save its mode for later. */
2656 if (*format_ptr == 'e')
2657 op0_mode = GET_MODE (XEXP (x, 0));
2659 for (i = 0; i < GET_RTX_LENGTH (code); i++)
2661 switch (*format_ptr++)
2663 case '0':
2664 break;
2666 case 'e':
2667 if (XEXP (x, i))
2668 subst_constants (&XEXP (x, i), insn, map);
2669 break;
2671 case 'u':
2672 case 'i':
2673 case 's':
2674 case 'w':
2675 break;
2677 case 'E':
2678 if (XVEC (x, i) != NULL && XVECLEN (x, i) != 0)
2680 int j;
2681 for (j = 0; j < XVECLEN (x, i); j++)
2682 subst_constants (&XVECEXP (x, i, j), insn, map);
2684 break;
2686 default:
2687 abort ();
2691 /* If this is a commutative operation, move a constant to the second
2692 operand unless the second operand is already a CONST_INT. */
2693 if ((GET_RTX_CLASS (code) == 'c' || code == NE || code == EQ)
2694 && CONSTANT_P (XEXP (x, 0)) && GET_CODE (XEXP (x, 1)) != CONST_INT)
2696 rtx tem = XEXP (x, 0);
2697 validate_change (insn, &XEXP (x, 0), XEXP (x, 1), 1);
2698 validate_change (insn, &XEXP (x, 1), tem, 1);
2701 /* Simplify the expression in case we put in some constants. */
2702 switch (GET_RTX_CLASS (code))
2704 case '1':
2705 new = simplify_unary_operation (code, GET_MODE (x),
2706 XEXP (x, 0), op0_mode);
2707 break;
2709 case '<':
2711 enum machine_mode op_mode = GET_MODE (XEXP (x, 0));
2712 if (op_mode == VOIDmode)
2713 op_mode = GET_MODE (XEXP (x, 1));
2714 new = simplify_relational_operation (code, op_mode,
2715 XEXP (x, 0), XEXP (x, 1));
2716 #ifdef FLOAT_STORE_FLAG_VALUE
2717 if (new != 0 && GET_MODE_CLASS (GET_MODE (x)) == MODE_FLOAT)
2718 new = ((new == const0_rtx) ? CONST0_RTX (GET_MODE (x))
2719 : CONST_DOUBLE_FROM_REAL_VALUE (FLOAT_STORE_FLAG_VALUE,
2720 GET_MODE (x)));
2721 #endif
2722 break;
2725 case '2':
2726 case 'c':
2727 new = simplify_binary_operation (code, GET_MODE (x),
2728 XEXP (x, 0), XEXP (x, 1));
2729 break;
2731 case 'b':
2732 case '3':
2733 new = simplify_ternary_operation (code, GET_MODE (x), op0_mode,
2734 XEXP (x, 0), XEXP (x, 1), XEXP (x, 2));
2735 break;
2738 if (new)
2739 validate_change (insn, loc, new, 1);
2742 /* Show that register modified no longer contain known constants. We are
2743 called from note_stores with parts of the new insn. */
2745 void
2746 mark_stores (dest, x)
2747 rtx dest;
2748 rtx x;
2750 int regno = -1;
2751 enum machine_mode mode;
2753 /* DEST is always the innermost thing set, except in the case of
2754 SUBREGs of hard registers. */
2756 if (GET_CODE (dest) == REG)
2757 regno = REGNO (dest), mode = GET_MODE (dest);
2758 else if (GET_CODE (dest) == SUBREG && GET_CODE (SUBREG_REG (dest)) == REG)
2760 regno = REGNO (SUBREG_REG (dest)) + SUBREG_WORD (dest);
2761 mode = GET_MODE (SUBREG_REG (dest));
2764 if (regno >= 0)
2766 int last_reg = (regno >= FIRST_PSEUDO_REGISTER ? regno
2767 : regno + HARD_REGNO_NREGS (regno, mode) - 1);
2768 int i;
2770 for (i = regno; i <= last_reg; i++)
2771 if (i < global_const_equiv_map_size)
2772 global_const_equiv_map[i] = 0;
2776 /* If any CONST expressions with RTX_INTEGRATED_P are present in the rtx
2777 pointed to by PX, they represent constants in the constant pool.
2778 Replace these with a new memory reference obtained from force_const_mem.
2779 Similarly, ADDRESS expressions with RTX_INTEGRATED_P represent the
2780 address of a constant pool entry. Replace them with the address of
2781 a new constant pool entry obtained from force_const_mem. */
2783 static void
2784 restore_constants (px)
2785 rtx *px;
2787 rtx x = *px;
2788 int i, j;
2789 char *fmt;
2791 if (x == 0)
2792 return;
2794 if (GET_CODE (x) == CONST_DOUBLE)
2796 /* We have to make a new CONST_DOUBLE to ensure that we account for
2797 it correctly. Using the old CONST_DOUBLE_MEM data is wrong. */
2798 if (GET_MODE_CLASS (GET_MODE (x)) == MODE_FLOAT)
2800 REAL_VALUE_TYPE d;
2802 REAL_VALUE_FROM_CONST_DOUBLE (d, x);
2803 *px = CONST_DOUBLE_FROM_REAL_VALUE (d, GET_MODE (x));
2805 else
2806 *px = immed_double_const (CONST_DOUBLE_LOW (x), CONST_DOUBLE_HIGH (x),
2807 VOIDmode);
2810 else if (RTX_INTEGRATED_P (x) && GET_CODE (x) == CONST)
2812 restore_constants (&XEXP (x, 0));
2813 *px = validize_mem (force_const_mem (GET_MODE (x), XEXP (x, 0)));
2815 else if (RTX_INTEGRATED_P (x) && GET_CODE (x) == SUBREG)
2817 /* This must be (subreg/i:M1 (const/i:M2 ...) 0). */
2818 rtx new = XEXP (SUBREG_REG (x), 0);
2820 restore_constants (&new);
2821 new = force_const_mem (GET_MODE (SUBREG_REG (x)), new);
2822 PUT_MODE (new, GET_MODE (x));
2823 *px = validize_mem (new);
2825 else if (RTX_INTEGRATED_P (x) && GET_CODE (x) == ADDRESS)
2827 restore_constants (&XEXP (x, 0));
2828 *px = XEXP (force_const_mem (GET_MODE (x), XEXP (x, 0)), 0);
2830 else
2832 fmt = GET_RTX_FORMAT (GET_CODE (x));
2833 for (i = 0; i < GET_RTX_LENGTH (GET_CODE (x)); i++)
2835 switch (*fmt++)
2837 case 'E':
2838 for (j = 0; j < XVECLEN (x, i); j++)
2839 restore_constants (&XVECEXP (x, i, j));
2840 break;
2842 case 'e':
2843 restore_constants (&XEXP (x, i));
2844 break;
2850 /* Given a pointer to some BLOCK node, if the BLOCK_ABSTRACT_ORIGIN for the
2851 given BLOCK node is NULL, set the BLOCK_ABSTRACT_ORIGIN for the node so
2852 that it points to the node itself, thus indicating that the node is its
2853 own (abstract) origin. Additionally, if the BLOCK_ABSTRACT_ORIGIN for
2854 the given node is NULL, recursively descend the decl/block tree which
2855 it is the root of, and for each other ..._DECL or BLOCK node contained
2856 therein whose DECL_ABSTRACT_ORIGINs or BLOCK_ABSTRACT_ORIGINs are also
2857 still NULL, set *their* DECL_ABSTRACT_ORIGIN or BLOCK_ABSTRACT_ORIGIN
2858 values to point to themselves. */
2860 static void
2861 set_block_origin_self (stmt)
2862 register tree stmt;
2864 if (BLOCK_ABSTRACT_ORIGIN (stmt) == NULL_TREE)
2866 BLOCK_ABSTRACT_ORIGIN (stmt) = stmt;
2869 register tree local_decl;
2871 for (local_decl = BLOCK_VARS (stmt);
2872 local_decl != NULL_TREE;
2873 local_decl = TREE_CHAIN (local_decl))
2874 set_decl_origin_self (local_decl); /* Potential recursion. */
2878 register tree subblock;
2880 for (subblock = BLOCK_SUBBLOCKS (stmt);
2881 subblock != NULL_TREE;
2882 subblock = BLOCK_CHAIN (subblock))
2883 set_block_origin_self (subblock); /* Recurse. */
2888 /* Given a pointer to some ..._DECL node, if the DECL_ABSTRACT_ORIGIN for
2889 the given ..._DECL node is NULL, set the DECL_ABSTRACT_ORIGIN for the
2890 node to so that it points to the node itself, thus indicating that the
2891 node represents its own (abstract) origin. Additionally, if the
2892 DECL_ABSTRACT_ORIGIN for the given node is NULL, recursively descend
2893 the decl/block tree of which the given node is the root of, and for
2894 each other ..._DECL or BLOCK node contained therein whose
2895 DECL_ABSTRACT_ORIGINs or BLOCK_ABSTRACT_ORIGINs are also still NULL,
2896 set *their* DECL_ABSTRACT_ORIGIN or BLOCK_ABSTRACT_ORIGIN values to
2897 point to themselves. */
2899 static void
2900 set_decl_origin_self (decl)
2901 register tree decl;
2903 if (DECL_ABSTRACT_ORIGIN (decl) == NULL_TREE)
2905 DECL_ABSTRACT_ORIGIN (decl) = decl;
2906 if (TREE_CODE (decl) == FUNCTION_DECL)
2908 register tree arg;
2910 for (arg = DECL_ARGUMENTS (decl); arg; arg = TREE_CHAIN (arg))
2911 DECL_ABSTRACT_ORIGIN (arg) = arg;
2912 if (DECL_INITIAL (decl) != NULL_TREE)
2913 set_block_origin_self (DECL_INITIAL (decl));
2918 /* Given a pointer to some BLOCK node, and a boolean value to set the
2919 "abstract" flags to, set that value into the BLOCK_ABSTRACT flag for
2920 the given block, and for all local decls and all local sub-blocks
2921 (recursively) which are contained therein. */
2923 static void
2924 set_block_abstract_flags (stmt, setting)
2925 register tree stmt;
2926 register int setting;
2928 BLOCK_ABSTRACT (stmt) = setting;
2931 register tree local_decl;
2933 for (local_decl = BLOCK_VARS (stmt);
2934 local_decl != NULL_TREE;
2935 local_decl = TREE_CHAIN (local_decl))
2936 set_decl_abstract_flags (local_decl, setting);
2940 register tree subblock;
2942 for (subblock = BLOCK_SUBBLOCKS (stmt);
2943 subblock != NULL_TREE;
2944 subblock = BLOCK_CHAIN (subblock))
2945 set_block_abstract_flags (subblock, setting);
2949 /* Given a pointer to some ..._DECL node, and a boolean value to set the
2950 "abstract" flags to, set that value into the DECL_ABSTRACT flag for the
2951 given decl, and (in the case where the decl is a FUNCTION_DECL) also
2952 set the abstract flags for all of the parameters, local vars, local
2953 blocks and sub-blocks (recursively) to the same setting. */
2955 void
2956 set_decl_abstract_flags (decl, setting)
2957 register tree decl;
2958 register int setting;
2960 DECL_ABSTRACT (decl) = setting;
2961 if (TREE_CODE (decl) == FUNCTION_DECL)
2963 register tree arg;
2965 for (arg = DECL_ARGUMENTS (decl); arg; arg = TREE_CHAIN (arg))
2966 DECL_ABSTRACT (arg) = setting;
2967 if (DECL_INITIAL (decl) != NULL_TREE)
2968 set_block_abstract_flags (DECL_INITIAL (decl), setting);
2972 /* Output the assembly language code for the function FNDECL
2973 from its DECL_SAVED_INSNS. Used for inline functions that are output
2974 at end of compilation instead of where they came in the source. */
2976 void
2977 output_inline_function (fndecl)
2978 tree fndecl;
2980 rtx head;
2981 rtx last;
2983 if (output_bytecode)
2985 warning ("`inline' ignored for bytecode output");
2986 return;
2989 head = DECL_SAVED_INSNS (fndecl);
2990 current_function_decl = fndecl;
2992 /* This call is only used to initialize global variables. */
2993 init_function_start (fndecl, "lossage", 1);
2995 /* Redo parameter determinations in case the FUNCTION_...
2996 macros took machine-specific actions that need to be redone. */
2997 assign_parms (fndecl, 1);
2999 /* Set stack frame size. */
3000 assign_stack_local (BLKmode, DECL_FRAME_SIZE (fndecl), 0);
3002 restore_reg_data (FIRST_PARM_INSN (head));
3004 stack_slot_list = STACK_SLOT_LIST (head);
3005 forced_labels = FORCED_LABELS (head);
3007 if (FUNCTION_FLAGS (head) & FUNCTION_FLAGS_CALLS_ALLOCA)
3008 current_function_calls_alloca = 1;
3010 if (FUNCTION_FLAGS (head) & FUNCTION_FLAGS_CALLS_SETJMP)
3011 current_function_calls_setjmp = 1;
3013 if (FUNCTION_FLAGS (head) & FUNCTION_FLAGS_CALLS_LONGJMP)
3014 current_function_calls_longjmp = 1;
3016 if (FUNCTION_FLAGS (head) & FUNCTION_FLAGS_RETURNS_STRUCT)
3017 current_function_returns_struct = 1;
3019 if (FUNCTION_FLAGS (head) & FUNCTION_FLAGS_RETURNS_PCC_STRUCT)
3020 current_function_returns_pcc_struct = 1;
3022 if (FUNCTION_FLAGS (head) & FUNCTION_FLAGS_NEEDS_CONTEXT)
3023 current_function_needs_context = 1;
3025 if (FUNCTION_FLAGS (head) & FUNCTION_FLAGS_HAS_NONLOCAL_LABEL)
3026 current_function_has_nonlocal_label = 1;
3028 if (FUNCTION_FLAGS (head) & FUNCTION_FLAGS_RETURNS_POINTER)
3029 current_function_returns_pointer = 1;
3031 if (FUNCTION_FLAGS (head) & FUNCTION_FLAGS_USES_CONST_POOL)
3032 current_function_uses_const_pool = 1;
3034 if (FUNCTION_FLAGS (head) & FUNCTION_FLAGS_USES_PIC_OFFSET_TABLE)
3035 current_function_uses_pic_offset_table = 1;
3037 current_function_outgoing_args_size = OUTGOING_ARGS_SIZE (head);
3038 current_function_pops_args = POPS_ARGS (head);
3040 /* There is no need to output a return label again. */
3041 return_label = 0;
3043 expand_function_end (DECL_SOURCE_FILE (fndecl), DECL_SOURCE_LINE (fndecl), 0);
3045 /* Find last insn and rebuild the constant pool. */
3046 for (last = FIRST_PARM_INSN (head);
3047 NEXT_INSN (last); last = NEXT_INSN (last))
3049 if (GET_RTX_CLASS (GET_CODE (last)) == 'i')
3051 restore_constants (&PATTERN (last));
3052 restore_constants (&REG_NOTES (last));
3056 set_new_first_and_last_insn (FIRST_PARM_INSN (head), last);
3057 set_new_first_and_last_label_num (FIRST_LABELNO (head), LAST_LABELNO (head));
3059 /* We must have already output DWARF debugging information for the
3060 original (abstract) inline function declaration/definition, so
3061 we want to make sure that the debugging information we generate
3062 for this special instance of the inline function refers back to
3063 the information we already generated. To make sure that happens,
3064 we simply have to set the DECL_ABSTRACT_ORIGIN for the function
3065 node (and for all of the local ..._DECL nodes which are its children)
3066 so that they all point to themselves. */
3068 set_decl_origin_self (fndecl);
3070 /* We're not deferring this any longer. */
3071 DECL_DEFER_OUTPUT (fndecl) = 0;
3073 /* Compile this function all the way down to assembly code. */
3074 rest_of_compilation (fndecl);
3076 current_function_decl = 0;