1 /* Store motion via Lazy Code Motion on the reverse CFG.
2 Copyright (C) 1997, 1998, 1999, 2000, 2001, 2002, 2003, 2004, 2005,
3 2006, 2007, 2008, 2009, 2010 Free Software Foundation, Inc.
5 This file is part of GCC.
7 GCC is free software; you can redistribute it and/or modify it under
8 the terms of the GNU General Public License as published by the Free
9 Software Foundation; either version 3, or (at your option) any later
12 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
13 WARRANTY; without even the implied warranty of MERCHANTABILITY or
14 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
17 You should have received a copy of the GNU General Public License
18 along with GCC; see the file COPYING3. If not see
19 <http://www.gnu.org/licenses/>. */
23 #include "coretypes.h"
25 #include "diagnostic-core.h"
32 #include "hard-reg-set.h"
34 #include "insn-config.h"
36 #include "basic-block.h"
42 #include "tree-pass.h"
47 /* This pass implements downward store motion.
48 As of May 1, 2009, the pass is not enabled by default on any target,
49 but bootstrap completes on ia64 and x86_64 with the pass enabled. */
52 - remove_reachable_equiv_notes is an incomprehensible pile of goo and
53 a compile time hog that needs a rewrite (maybe cache st_exprs to
54 invalidate REG_EQUAL/REG_EQUIV notes for?).
55 - pattern_regs in st_expr should be a regset (on its own obstack).
56 - antic_stores and avail_stores should be VECs instead of lists.
57 - store_motion_mems should be a VEC instead of a list.
58 - there should be an alloc pool for struct st_expr objects.
59 - investigate whether it is helpful to make the address of an st_expr
61 - when GIMPLE alias information is exported, the effectiveness of this
62 pass should be re-evaluated.
65 /* This is a list of store expressions (MEMs). The structure is used
66 as an expression table to track stores which look interesting, and
67 might be moveable towards the exit block. */
71 /* Pattern of this mem. */
73 /* List of registers mentioned by the mem. */
75 /* INSN list of stores that are locally anticipatable. */
77 /* INSN list of stores that are locally available. */
79 /* Next in the list. */
80 struct st_expr
* next
;
81 /* Store ID in the dataflow bitmaps. */
83 /* Hash value for the hash table. */
84 unsigned int hash_index
;
85 /* Register holding the stored expression when a store is moved.
86 This field is also used as a cache in find_moveable_store, see
87 LAST_AVAIL_CHECK_FAILURE below. */
91 /* Head of the list of load/store memory refs. */
92 static struct st_expr
* store_motion_mems
= NULL
;
94 /* Hashtable for the load/store memory refs. */
95 static htab_t store_motion_mems_table
= NULL
;
97 /* These bitmaps will hold the local dataflow properties per basic block. */
98 static sbitmap
*st_kill
, *st_avloc
, *st_antloc
, *st_transp
;
100 /* Nonzero for expressions which should be inserted on a specific edge. */
101 static sbitmap
*st_insert_map
;
103 /* Nonzero for expressions which should be deleted in a specific block. */
104 static sbitmap
*st_delete_map
;
106 /* Global holding the number of store expressions we are dealing with. */
107 static int num_stores
;
109 /* Contains the edge_list returned by pre_edge_lcm. */
110 static struct edge_list
*edge_list
;
113 pre_st_expr_hash (const void *p
)
115 int do_not_record_p
= 0;
116 const struct st_expr
*const x
= (const struct st_expr
*) p
;
117 return hash_rtx (x
->pattern
, GET_MODE (x
->pattern
), &do_not_record_p
, NULL
, false);
121 pre_st_expr_eq (const void *p1
, const void *p2
)
123 const struct st_expr
*const ptr1
= (const struct st_expr
*) p1
,
124 *const ptr2
= (const struct st_expr
*) p2
;
125 return exp_equiv_p (ptr1
->pattern
, ptr2
->pattern
, 0, true);
128 /* This will search the st_expr list for a matching expression. If it
129 doesn't find one, we create one and initialize it. */
131 static struct st_expr
*
132 st_expr_entry (rtx x
)
134 int do_not_record_p
= 0;
135 struct st_expr
* ptr
;
140 hash
= hash_rtx (x
, GET_MODE (x
), &do_not_record_p
,
141 NULL
, /*have_reg_qty=*/false);
144 slot
= htab_find_slot_with_hash (store_motion_mems_table
, &e
, hash
, INSERT
);
146 return (struct st_expr
*)*slot
;
148 ptr
= XNEW (struct st_expr
);
150 ptr
->next
= store_motion_mems
;
152 ptr
->pattern_regs
= NULL_RTX
;
153 ptr
->antic_stores
= NULL_RTX
;
154 ptr
->avail_stores
= NULL_RTX
;
155 ptr
->reaching_reg
= NULL_RTX
;
157 ptr
->hash_index
= hash
;
158 store_motion_mems
= ptr
;
164 /* Free up an individual st_expr entry. */
167 free_st_expr_entry (struct st_expr
* ptr
)
169 free_INSN_LIST_list (& ptr
->antic_stores
);
170 free_INSN_LIST_list (& ptr
->avail_stores
);
175 /* Free up all memory associated with the st_expr list. */
178 free_store_motion_mems (void)
180 if (store_motion_mems_table
)
181 htab_delete (store_motion_mems_table
);
182 store_motion_mems_table
= NULL
;
184 while (store_motion_mems
)
186 struct st_expr
* tmp
= store_motion_mems
;
187 store_motion_mems
= store_motion_mems
->next
;
188 free_st_expr_entry (tmp
);
190 store_motion_mems
= NULL
;
193 /* Assign each element of the list of mems a monotonically increasing value. */
196 enumerate_store_motion_mems (void)
198 struct st_expr
* ptr
;
201 for (ptr
= store_motion_mems
; ptr
!= NULL
; ptr
= ptr
->next
)
207 /* Return first item in the list. */
209 static inline struct st_expr
*
212 return store_motion_mems
;
215 /* Return the next item in the list after the specified one. */
217 static inline struct st_expr
*
218 next_st_expr (struct st_expr
* ptr
)
223 /* Dump debugging info about the store_motion_mems list. */
226 print_store_motion_mems (FILE * file
)
228 struct st_expr
* ptr
;
230 fprintf (dump_file
, "STORE_MOTION list of MEM exprs considered:\n");
232 for (ptr
= first_st_expr (); ptr
!= NULL
; ptr
= next_st_expr (ptr
))
234 fprintf (file
, " Pattern (%3d): ", ptr
->index
);
236 print_rtl (file
, ptr
->pattern
);
238 fprintf (file
, "\n ANTIC stores : ");
240 if (ptr
->antic_stores
)
241 print_rtl (file
, ptr
->antic_stores
);
243 fprintf (file
, "(nil)");
245 fprintf (file
, "\n AVAIL stores : ");
247 if (ptr
->avail_stores
)
248 print_rtl (file
, ptr
->avail_stores
);
250 fprintf (file
, "(nil)");
252 fprintf (file
, "\n\n");
255 fprintf (file
, "\n");
258 /* Return zero if some of the registers in list X are killed
259 due to set of registers in bitmap REGS_SET. */
262 store_ops_ok (const_rtx x
, int *regs_set
)
266 for (; x
; x
= XEXP (x
, 1))
269 if (regs_set
[REGNO(reg
)])
276 /* Helper for extract_mentioned_regs. */
279 extract_mentioned_regs_1 (rtx
*loc
, void *data
)
281 rtx
*mentioned_regs_p
= (rtx
*) data
;
284 *mentioned_regs_p
= alloc_EXPR_LIST (0, *loc
, *mentioned_regs_p
);
289 /* Returns a list of registers mentioned in X.
290 FIXME: A regset would be prettier and less expensive. */
293 extract_mentioned_regs (rtx x
)
295 rtx mentioned_regs
= NULL
;
296 for_each_rtx (&x
, extract_mentioned_regs_1
, &mentioned_regs
);
297 return mentioned_regs
;
300 /* Check to see if the load X is aliased with STORE_PATTERN.
301 AFTER is true if we are checking the case when STORE_PATTERN occurs
305 load_kills_store (const_rtx x
, const_rtx store_pattern
, int after
)
308 return anti_dependence (x
, store_pattern
);
310 return true_dependence (store_pattern
, GET_MODE (store_pattern
), x
);
313 /* Go through the entire rtx X, looking for any loads which might alias
314 STORE_PATTERN. Return true if found.
315 AFTER is true if we are checking the case when STORE_PATTERN occurs
319 find_loads (const_rtx x
, const_rtx store_pattern
, int after
)
328 if (GET_CODE (x
) == SET
)
333 if (load_kills_store (x
, store_pattern
, after
))
337 /* Recursively process the insn. */
338 fmt
= GET_RTX_FORMAT (GET_CODE (x
));
340 for (i
= GET_RTX_LENGTH (GET_CODE (x
)) - 1; i
>= 0 && !ret
; i
--)
343 ret
|= find_loads (XEXP (x
, i
), store_pattern
, after
);
344 else if (fmt
[i
] == 'E')
345 for (j
= XVECLEN (x
, i
) - 1; j
>= 0; j
--)
346 ret
|= find_loads (XVECEXP (x
, i
, j
), store_pattern
, after
);
351 /* Go through pattern PAT looking for any loads which might kill the
352 store in X. Return true if found.
353 AFTER is true if we are checking the case when loads kill X occurs
354 after the insn for PAT. */
357 store_killed_in_pat (const_rtx x
, const_rtx pat
, int after
)
359 if (GET_CODE (pat
) == SET
)
361 rtx dest
= SET_DEST (pat
);
363 if (GET_CODE (dest
) == ZERO_EXTRACT
)
364 dest
= XEXP (dest
, 0);
366 /* Check for memory stores to aliased objects. */
368 && !exp_equiv_p (dest
, x
, 0, true))
372 if (output_dependence (dest
, x
))
377 if (output_dependence (x
, dest
))
383 if (find_loads (pat
, x
, after
))
389 /* Check if INSN kills the store pattern X (is aliased with it).
390 AFTER is true if we are checking the case when store X occurs
391 after the insn. Return true if it does. */
394 store_killed_in_insn (const_rtx x
, const_rtx x_regs
, const_rtx insn
, int after
)
396 const_rtx reg
, note
, pat
;
398 if (! NONDEBUG_INSN_P (insn
))
403 /* A normal or pure call might read from pattern,
404 but a const call will not. */
405 if (!RTL_CONST_CALL_P (insn
))
408 /* But even a const call reads its parameters. Check whether the
409 base of some of registers used in mem is stack pointer. */
410 for (reg
= x_regs
; reg
; reg
= XEXP (reg
, 1))
411 if (may_be_sp_based_p (XEXP (reg
, 0)))
417 pat
= PATTERN (insn
);
418 if (GET_CODE (pat
) == SET
)
420 if (store_killed_in_pat (x
, pat
, after
))
423 else if (GET_CODE (pat
) == PARALLEL
)
427 for (i
= 0; i
< XVECLEN (pat
, 0); i
++)
428 if (store_killed_in_pat (x
, XVECEXP (pat
, 0, i
), after
))
431 else if (find_loads (PATTERN (insn
), x
, after
))
434 /* If this insn has a REG_EQUAL or REG_EQUIV note referencing a memory
435 location aliased with X, then this insn kills X. */
436 note
= find_reg_equal_equiv_note (insn
);
439 note
= XEXP (note
, 0);
441 /* However, if the note represents a must alias rather than a may
442 alias relationship, then it does not kill X. */
443 if (exp_equiv_p (note
, x
, 0, true))
446 /* See if there are any aliased loads in the note. */
447 return find_loads (note
, x
, after
);
450 /* Returns true if the expression X is loaded or clobbered on or after INSN
451 within basic block BB. REGS_SET_AFTER is bitmap of registers set in
452 or after the insn. X_REGS is list of registers mentioned in X. If the store
453 is killed, return the last insn in that it occurs in FAIL_INSN. */
456 store_killed_after (const_rtx x
, const_rtx x_regs
, const_rtx insn
, const_basic_block bb
,
457 int *regs_set_after
, rtx
*fail_insn
)
459 rtx last
= BB_END (bb
), act
;
461 if (!store_ops_ok (x_regs
, regs_set_after
))
463 /* We do not know where it will happen. */
465 *fail_insn
= NULL_RTX
;
469 /* Scan from the end, so that fail_insn is determined correctly. */
470 for (act
= last
; act
!= PREV_INSN (insn
); act
= PREV_INSN (act
))
471 if (store_killed_in_insn (x
, x_regs
, act
, false))
481 /* Returns true if the expression X is loaded or clobbered on or before INSN
482 within basic block BB. X_REGS is list of registers mentioned in X.
483 REGS_SET_BEFORE is bitmap of registers set before or in this insn. */
485 store_killed_before (const_rtx x
, const_rtx x_regs
, const_rtx insn
, const_basic_block bb
,
486 int *regs_set_before
)
488 rtx first
= BB_HEAD (bb
);
490 if (!store_ops_ok (x_regs
, regs_set_before
))
493 for ( ; insn
!= PREV_INSN (first
); insn
= PREV_INSN (insn
))
494 if (store_killed_in_insn (x
, x_regs
, insn
, true))
500 /* The last insn in the basic block that compute_store_table is processing,
501 where store_killed_after is true for X.
502 Since we go through the basic block from BB_END to BB_HEAD, this is
503 also the available store at the end of the basic block. Therefore
504 this is in effect a cache, to avoid calling store_killed_after for
505 equivalent aliasing store expressions.
506 This value is only meaningful during the computation of the store
507 table. We hi-jack the REACHING_REG field of struct st_expr to save
509 #define LAST_AVAIL_CHECK_FAILURE(x) ((x)->reaching_reg)
511 /* Determine whether INSN is MEM store pattern that we will consider moving.
512 REGS_SET_BEFORE is bitmap of registers set before (and including) the
513 current insn, REGS_SET_AFTER is bitmap of registers set after (and
514 including) the insn in this basic block. We must be passing through BB from
515 head to end, as we are using this fact to speed things up.
517 The results are stored this way:
519 -- the first anticipatable expression is added into ANTIC_STORES
520 -- if the processed expression is not anticipatable, NULL_RTX is added
521 there instead, so that we can use it as indicator that no further
522 expression of this type may be anticipatable
523 -- if the expression is available, it is added as head of AVAIL_STORES;
524 consequently, all of them but this head are dead and may be deleted.
525 -- if the expression is not available, the insn due to that it fails to be
526 available is stored in REACHING_REG (via LAST_AVAIL_CHECK_FAILURE).
528 The things are complicated a bit by fact that there already may be stores
529 to the same MEM from other blocks; also caller must take care of the
530 necessary cleanup of the temporary markers after end of the basic block.
534 find_moveable_store (rtx insn
, int *regs_set_before
, int *regs_set_after
)
536 struct st_expr
* ptr
;
538 int check_anticipatable
, check_available
;
539 basic_block bb
= BLOCK_FOR_INSN (insn
);
541 set
= single_set (insn
);
545 dest
= SET_DEST (set
);
547 if (! MEM_P (dest
) || MEM_VOLATILE_P (dest
)
548 || GET_MODE (dest
) == BLKmode
)
551 if (side_effects_p (dest
))
554 /* If we are handling exceptions, we must be careful with memory references
555 that may trap. If we are not, the behavior is undefined, so we may just
557 if (cfun
->can_throw_non_call_exceptions
&& may_trap_p (dest
))
560 /* Even if the destination cannot trap, the source may. In this case we'd
561 need to handle updating the REG_EH_REGION note. */
562 if (find_reg_note (insn
, REG_EH_REGION
, NULL_RTX
))
565 /* Make sure that the SET_SRC of this store insns can be assigned to
566 a register, or we will fail later on in replace_store_insn, which
567 assumes that we can do this. But sometimes the target machine has
568 oddities like MEM read-modify-write instruction. See for example
570 if (!can_assign_to_reg_without_clobbers_p (SET_SRC (set
)))
573 ptr
= st_expr_entry (dest
);
574 if (!ptr
->pattern_regs
)
575 ptr
->pattern_regs
= extract_mentioned_regs (dest
);
577 /* Do not check for anticipatability if we either found one anticipatable
578 store already, or tested for one and found out that it was killed. */
579 check_anticipatable
= 0;
580 if (!ptr
->antic_stores
)
581 check_anticipatable
= 1;
584 tmp
= XEXP (ptr
->antic_stores
, 0);
586 && BLOCK_FOR_INSN (tmp
) != bb
)
587 check_anticipatable
= 1;
589 if (check_anticipatable
)
591 if (store_killed_before (dest
, ptr
->pattern_regs
, insn
, bb
, regs_set_before
))
595 ptr
->antic_stores
= alloc_INSN_LIST (tmp
, ptr
->antic_stores
);
598 /* It is not necessary to check whether store is available if we did
599 it successfully before; if we failed before, do not bother to check
600 until we reach the insn that caused us to fail. */
602 if (!ptr
->avail_stores
)
606 tmp
= XEXP (ptr
->avail_stores
, 0);
607 if (BLOCK_FOR_INSN (tmp
) != bb
)
612 /* Check that we have already reached the insn at that the check
614 if (LAST_AVAIL_CHECK_FAILURE (ptr
))
616 for (tmp
= BB_END (bb
);
617 tmp
!= insn
&& tmp
!= LAST_AVAIL_CHECK_FAILURE (ptr
);
618 tmp
= PREV_INSN (tmp
))
624 check_available
= store_killed_after (dest
, ptr
->pattern_regs
, insn
,
626 &LAST_AVAIL_CHECK_FAILURE (ptr
));
628 if (!check_available
)
629 ptr
->avail_stores
= alloc_INSN_LIST (insn
, ptr
->avail_stores
);
632 /* Find available and anticipatable stores. */
635 compute_store_table (void)
639 #ifdef ENABLE_CHECKING
644 int *last_set_in
, *already_set
;
645 struct st_expr
* ptr
, **prev_next_ptr_ptr
;
646 unsigned int max_gcse_regno
= max_reg_num ();
648 store_motion_mems
= NULL
;
649 store_motion_mems_table
= htab_create (13, pre_st_expr_hash
,
650 pre_st_expr_eq
, NULL
);
651 last_set_in
= XCNEWVEC (int, max_gcse_regno
);
652 already_set
= XNEWVEC (int, max_gcse_regno
);
654 /* Find all the stores we care about. */
657 /* First compute the registers set in this block. */
658 FOR_BB_INSNS (bb
, insn
)
661 if (! NONDEBUG_INSN_P (insn
))
664 for (def_rec
= DF_INSN_DEFS (insn
); *def_rec
; def_rec
++)
665 last_set_in
[DF_REF_REGNO (*def_rec
)] = INSN_UID (insn
);
668 /* Now find the stores. */
669 memset (already_set
, 0, sizeof (int) * max_gcse_regno
);
670 FOR_BB_INSNS (bb
, insn
)
672 if (! NONDEBUG_INSN_P (insn
))
675 for (def_rec
= DF_INSN_DEFS (insn
); *def_rec
; def_rec
++)
676 already_set
[DF_REF_REGNO (*def_rec
)] = INSN_UID (insn
);
678 /* Now that we've marked regs, look for stores. */
679 find_moveable_store (insn
, already_set
, last_set_in
);
681 /* Unmark regs that are no longer set. */
682 for (def_rec
= DF_INSN_DEFS (insn
); *def_rec
; def_rec
++)
683 if (last_set_in
[DF_REF_REGNO (*def_rec
)] == INSN_UID (insn
))
684 last_set_in
[DF_REF_REGNO (*def_rec
)] = 0;
687 #ifdef ENABLE_CHECKING
688 /* last_set_in should now be all-zero. */
689 for (regno
= 0; regno
< max_gcse_regno
; regno
++)
690 gcc_assert (!last_set_in
[regno
]);
693 /* Clear temporary marks. */
694 for (ptr
= first_st_expr (); ptr
!= NULL
; ptr
= next_st_expr (ptr
))
696 LAST_AVAIL_CHECK_FAILURE (ptr
) = NULL_RTX
;
697 if (ptr
->antic_stores
698 && (tmp
= XEXP (ptr
->antic_stores
, 0)) == NULL_RTX
)
699 ptr
->antic_stores
= XEXP (ptr
->antic_stores
, 1);
703 /* Remove the stores that are not available anywhere, as there will
704 be no opportunity to optimize them. */
705 for (ptr
= store_motion_mems
, prev_next_ptr_ptr
= &store_motion_mems
;
707 ptr
= *prev_next_ptr_ptr
)
709 if (! ptr
->avail_stores
)
711 *prev_next_ptr_ptr
= ptr
->next
;
712 htab_remove_elt_with_hash (store_motion_mems_table
,
713 ptr
, ptr
->hash_index
);
714 free_st_expr_entry (ptr
);
717 prev_next_ptr_ptr
= &ptr
->next
;
720 ret
= enumerate_store_motion_mems ();
723 print_store_motion_mems (dump_file
);
730 /* In all code following after this, REACHING_REG has its original
731 meaning again. Avoid confusion, and undef the accessor macro for
732 the temporary marks usage in compute_store_table. */
733 #undef LAST_AVAIL_CHECK_FAILURE
735 /* Insert an instruction at the beginning of a basic block, and update
736 the BB_HEAD if needed. */
739 insert_insn_start_basic_block (rtx insn
, basic_block bb
)
741 /* Insert at start of successor block. */
742 rtx prev
= PREV_INSN (BB_HEAD (bb
));
743 rtx before
= BB_HEAD (bb
);
746 if (! LABEL_P (before
)
747 && !NOTE_INSN_BASIC_BLOCK_P (before
))
750 if (prev
== BB_END (bb
))
752 before
= NEXT_INSN (before
);
755 insn
= emit_insn_after_noloc (insn
, prev
, bb
);
759 fprintf (dump_file
, "STORE_MOTION insert store at start of BB %d:\n",
761 print_inline_rtx (dump_file
, insn
, 6);
762 fprintf (dump_file
, "\n");
766 /* This routine will insert a store on an edge. EXPR is the st_expr entry for
767 the memory reference, and E is the edge to insert it on. Returns nonzero
768 if an edge insertion was performed. */
771 insert_store (struct st_expr
* expr
, edge e
)
778 /* We did all the deleted before this insert, so if we didn't delete a
779 store, then we haven't set the reaching reg yet either. */
780 if (expr
->reaching_reg
== NULL_RTX
)
783 if (e
->flags
& EDGE_FAKE
)
786 reg
= expr
->reaching_reg
;
787 insn
= gen_move_insn (copy_rtx (expr
->pattern
), reg
);
789 /* If we are inserting this expression on ALL predecessor edges of a BB,
790 insert it at the start of the BB, and reset the insert bits on the other
791 edges so we don't try to insert it on the other edges. */
793 FOR_EACH_EDGE (tmp
, ei
, e
->dest
->preds
)
794 if (!(tmp
->flags
& EDGE_FAKE
))
796 int index
= EDGE_INDEX (edge_list
, tmp
->src
, tmp
->dest
);
798 gcc_assert (index
!= EDGE_INDEX_NO_EDGE
);
799 if (! TEST_BIT (st_insert_map
[index
], expr
->index
))
803 /* If tmp is NULL, we found an insertion on every edge, blank the
804 insertion vector for these edges, and insert at the start of the BB. */
805 if (!tmp
&& bb
!= EXIT_BLOCK_PTR
)
807 FOR_EACH_EDGE (tmp
, ei
, e
->dest
->preds
)
809 int index
= EDGE_INDEX (edge_list
, tmp
->src
, tmp
->dest
);
810 RESET_BIT (st_insert_map
[index
], expr
->index
);
812 insert_insn_start_basic_block (insn
, bb
);
816 /* We can't put stores in the front of blocks pointed to by abnormal
817 edges since that may put a store where one didn't used to be. */
818 gcc_assert (!(e
->flags
& EDGE_ABNORMAL
));
820 insert_insn_on_edge (insn
, e
);
824 fprintf (dump_file
, "STORE_MOTION insert insn on edge (%d, %d):\n",
825 e
->src
->index
, e
->dest
->index
);
826 print_inline_rtx (dump_file
, insn
, 6);
827 fprintf (dump_file
, "\n");
833 /* Remove any REG_EQUAL or REG_EQUIV notes containing a reference to the
834 memory location in SMEXPR set in basic block BB.
836 This could be rather expensive. */
839 remove_reachable_equiv_notes (basic_block bb
, struct st_expr
*smexpr
)
841 edge_iterator
*stack
, ei
;
844 sbitmap visited
= sbitmap_alloc (last_basic_block
);
845 rtx last
, insn
, note
;
846 rtx mem
= smexpr
->pattern
;
848 stack
= XNEWVEC (edge_iterator
, n_basic_blocks
);
850 ei
= ei_start (bb
->succs
);
852 sbitmap_zero (visited
);
854 act
= (EDGE_COUNT (ei_container (ei
)) > 0 ? EDGE_I (ei_container (ei
), 0) : NULL
);
862 sbitmap_free (visited
);
865 act
= ei_edge (stack
[--sp
]);
869 if (bb
== EXIT_BLOCK_PTR
870 || TEST_BIT (visited
, bb
->index
))
874 act
= (! ei_end_p (ei
)) ? ei_edge (ei
) : NULL
;
877 SET_BIT (visited
, bb
->index
);
879 if (TEST_BIT (st_antloc
[bb
->index
], smexpr
->index
))
881 for (last
= smexpr
->antic_stores
;
882 BLOCK_FOR_INSN (XEXP (last
, 0)) != bb
;
883 last
= XEXP (last
, 1))
885 last
= XEXP (last
, 0);
888 last
= NEXT_INSN (BB_END (bb
));
890 for (insn
= BB_HEAD (bb
); insn
!= last
; insn
= NEXT_INSN (insn
))
891 if (NONDEBUG_INSN_P (insn
))
893 note
= find_reg_equal_equiv_note (insn
);
894 if (!note
|| !exp_equiv_p (XEXP (note
, 0), mem
, 0, true))
898 fprintf (dump_file
, "STORE_MOTION drop REG_EQUAL note at insn %d:\n",
900 remove_note (insn
, note
);
905 act
= (! ei_end_p (ei
)) ? ei_edge (ei
) : NULL
;
907 if (EDGE_COUNT (bb
->succs
) > 0)
911 ei
= ei_start (bb
->succs
);
912 act
= (EDGE_COUNT (ei_container (ei
)) > 0 ? EDGE_I (ei_container (ei
), 0) : NULL
);
917 /* This routine will replace a store with a SET to a specified register. */
920 replace_store_insn (rtx reg
, rtx del
, basic_block bb
, struct st_expr
*smexpr
)
922 rtx insn
, mem
, note
, set
, ptr
;
924 mem
= smexpr
->pattern
;
925 insn
= gen_move_insn (reg
, SET_SRC (single_set (del
)));
927 for (ptr
= smexpr
->antic_stores
; ptr
; ptr
= XEXP (ptr
, 1))
928 if (XEXP (ptr
, 0) == del
)
930 XEXP (ptr
, 0) = insn
;
934 /* Move the notes from the deleted insn to its replacement. */
935 REG_NOTES (insn
) = REG_NOTES (del
);
937 /* Emit the insn AFTER all the notes are transferred.
938 This is cheaper since we avoid df rescanning for the note change. */
939 insn
= emit_insn_after (insn
, del
);
944 "STORE_MOTION delete insn in BB %d:\n ", bb
->index
);
945 print_inline_rtx (dump_file
, del
, 6);
946 fprintf (dump_file
, "\nSTORE_MOTION replaced with insn:\n ");
947 print_inline_rtx (dump_file
, insn
, 6);
948 fprintf (dump_file
, "\n");
953 /* Now we must handle REG_EQUAL notes whose contents is equal to the mem;
954 they are no longer accurate provided that they are reached by this
955 definition, so drop them. */
956 for (; insn
!= NEXT_INSN (BB_END (bb
)); insn
= NEXT_INSN (insn
))
957 if (NONDEBUG_INSN_P (insn
))
959 set
= single_set (insn
);
962 if (exp_equiv_p (SET_DEST (set
), mem
, 0, true))
964 note
= find_reg_equal_equiv_note (insn
);
965 if (!note
|| !exp_equiv_p (XEXP (note
, 0), mem
, 0, true))
969 fprintf (dump_file
, "STORE_MOTION drop REG_EQUAL note at insn %d:\n",
971 remove_note (insn
, note
);
973 remove_reachable_equiv_notes (bb
, smexpr
);
977 /* Delete a store, but copy the value that would have been stored into
978 the reaching_reg for later storing. */
981 delete_store (struct st_expr
* expr
, basic_block bb
)
985 if (expr
->reaching_reg
== NULL_RTX
)
986 expr
->reaching_reg
= gen_reg_rtx_and_attrs (expr
->pattern
);
988 reg
= expr
->reaching_reg
;
990 for (i
= expr
->avail_stores
; i
; i
= XEXP (i
, 1))
993 if (BLOCK_FOR_INSN (del
) == bb
)
995 /* We know there is only one since we deleted redundant
996 ones during the available computation. */
997 replace_store_insn (reg
, del
, bb
, expr
);
1003 /* Fill in available, anticipatable, transparent and kill vectors in
1004 STORE_DATA, based on lists of available and anticipatable stores. */
1006 build_store_vectors (void)
1009 int *regs_set_in_block
;
1011 struct st_expr
* ptr
;
1012 unsigned int max_gcse_regno
= max_reg_num ();
1014 /* Build the gen_vector. This is any store in the table which is not killed
1015 by aliasing later in its block. */
1016 st_avloc
= sbitmap_vector_alloc (last_basic_block
, num_stores
);
1017 sbitmap_vector_zero (st_avloc
, last_basic_block
);
1019 st_antloc
= sbitmap_vector_alloc (last_basic_block
, num_stores
);
1020 sbitmap_vector_zero (st_antloc
, last_basic_block
);
1022 for (ptr
= first_st_expr (); ptr
!= NULL
; ptr
= next_st_expr (ptr
))
1024 for (st
= ptr
->avail_stores
; st
!= NULL
; st
= XEXP (st
, 1))
1026 insn
= XEXP (st
, 0);
1027 bb
= BLOCK_FOR_INSN (insn
);
1029 /* If we've already seen an available expression in this block,
1030 we can delete this one (It occurs earlier in the block). We'll
1031 copy the SRC expression to an unused register in case there
1032 are any side effects. */
1033 if (TEST_BIT (st_avloc
[bb
->index
], ptr
->index
))
1035 rtx r
= gen_reg_rtx_and_attrs (ptr
->pattern
);
1037 fprintf (dump_file
, "Removing redundant store:\n");
1038 replace_store_insn (r
, XEXP (st
, 0), bb
, ptr
);
1041 SET_BIT (st_avloc
[bb
->index
], ptr
->index
);
1044 for (st
= ptr
->antic_stores
; st
!= NULL
; st
= XEXP (st
, 1))
1046 insn
= XEXP (st
, 0);
1047 bb
= BLOCK_FOR_INSN (insn
);
1048 SET_BIT (st_antloc
[bb
->index
], ptr
->index
);
1052 st_kill
= sbitmap_vector_alloc (last_basic_block
, num_stores
);
1053 sbitmap_vector_zero (st_kill
, last_basic_block
);
1055 st_transp
= sbitmap_vector_alloc (last_basic_block
, num_stores
);
1056 sbitmap_vector_zero (st_transp
, last_basic_block
);
1057 regs_set_in_block
= XNEWVEC (int, max_gcse_regno
);
1061 memset (regs_set_in_block
, 0, sizeof (int) * max_gcse_regno
);
1063 FOR_BB_INSNS (bb
, insn
)
1064 if (NONDEBUG_INSN_P (insn
))
1067 for (def_rec
= DF_INSN_DEFS (insn
); *def_rec
; def_rec
++)
1069 unsigned int ref_regno
= DF_REF_REGNO (*def_rec
);
1070 if (ref_regno
< max_gcse_regno
)
1071 regs_set_in_block
[DF_REF_REGNO (*def_rec
)] = 1;
1075 for (ptr
= first_st_expr (); ptr
!= NULL
; ptr
= next_st_expr (ptr
))
1077 if (store_killed_after (ptr
->pattern
, ptr
->pattern_regs
, BB_HEAD (bb
),
1078 bb
, regs_set_in_block
, NULL
))
1080 /* It should not be necessary to consider the expression
1081 killed if it is both anticipatable and available. */
1082 if (!TEST_BIT (st_antloc
[bb
->index
], ptr
->index
)
1083 || !TEST_BIT (st_avloc
[bb
->index
], ptr
->index
))
1084 SET_BIT (st_kill
[bb
->index
], ptr
->index
);
1087 SET_BIT (st_transp
[bb
->index
], ptr
->index
);
1091 free (regs_set_in_block
);
1095 dump_sbitmap_vector (dump_file
, "st_antloc", "", st_antloc
, last_basic_block
);
1096 dump_sbitmap_vector (dump_file
, "st_kill", "", st_kill
, last_basic_block
);
1097 dump_sbitmap_vector (dump_file
, "st_transp", "", st_transp
, last_basic_block
);
1098 dump_sbitmap_vector (dump_file
, "st_avloc", "", st_avloc
, last_basic_block
);
1102 /* Free memory used by store motion. */
1105 free_store_memory (void)
1107 free_store_motion_mems ();
1110 sbitmap_vector_free (st_avloc
);
1112 sbitmap_vector_free (st_kill
);
1114 sbitmap_vector_free (st_transp
);
1116 sbitmap_vector_free (st_antloc
);
1118 sbitmap_vector_free (st_insert_map
);
1120 sbitmap_vector_free (st_delete_map
);
1122 st_avloc
= st_kill
= st_transp
= st_antloc
= NULL
;
1123 st_insert_map
= st_delete_map
= NULL
;
1126 /* Perform store motion. Much like gcse, except we move expressions the
1127 other way by looking at the flowgraph in reverse.
1128 Return non-zero if transformations are performed by the pass. */
1131 one_store_motion_pass (void)
1135 struct st_expr
* ptr
;
1136 int did_edge_inserts
= 0;
1137 int n_stores_deleted
= 0;
1138 int n_stores_created
= 0;
1140 init_alias_analysis ();
1142 /* Find all the available and anticipatable stores. */
1143 num_stores
= compute_store_table ();
1144 if (num_stores
== 0)
1146 htab_delete (store_motion_mems_table
);
1147 store_motion_mems_table
= NULL
;
1148 end_alias_analysis ();
1152 /* Now compute kill & transp vectors. */
1153 build_store_vectors ();
1154 add_noreturn_fake_exit_edges ();
1155 connect_infinite_loops_to_exit ();
1157 edge_list
= pre_edge_rev_lcm (num_stores
, st_transp
, st_avloc
,
1158 st_antloc
, st_kill
, &st_insert_map
,
1161 /* Now we want to insert the new stores which are going to be needed. */
1162 for (ptr
= first_st_expr (); ptr
!= NULL
; ptr
= next_st_expr (ptr
))
1164 /* If any of the edges we have above are abnormal, we can't move this
1166 for (x
= NUM_EDGES (edge_list
) - 1; x
>= 0; x
--)
1167 if (TEST_BIT (st_insert_map
[x
], ptr
->index
)
1168 && (INDEX_EDGE (edge_list
, x
)->flags
& EDGE_ABNORMAL
))
1173 if (dump_file
!= NULL
)
1175 "Can't replace store %d: abnormal edge from %d to %d\n",
1176 ptr
->index
, INDEX_EDGE (edge_list
, x
)->src
->index
,
1177 INDEX_EDGE (edge_list
, x
)->dest
->index
);
1181 /* Now we want to insert the new stores which are going to be needed. */
1184 if (TEST_BIT (st_delete_map
[bb
->index
], ptr
->index
))
1186 delete_store (ptr
, bb
);
1190 for (x
= 0; x
< NUM_EDGES (edge_list
); x
++)
1191 if (TEST_BIT (st_insert_map
[x
], ptr
->index
))
1193 did_edge_inserts
|= insert_store (ptr
, INDEX_EDGE (edge_list
, x
));
1198 if (did_edge_inserts
)
1199 commit_edge_insertions ();
1201 free_store_memory ();
1202 free_edge_list (edge_list
);
1203 remove_fake_exit_edges ();
1204 end_alias_analysis ();
1208 fprintf (dump_file
, "STORE_MOTION of %s, %d basic blocks, ",
1209 current_function_name (), n_basic_blocks
);
1210 fprintf (dump_file
, "%d insns deleted, %d insns created\n",
1211 n_stores_deleted
, n_stores_created
);
1214 return (n_stores_deleted
> 0 || n_stores_created
> 0);
1219 gate_rtl_store_motion (void)
1221 return optimize
> 0 && flag_gcse_sm
1222 && !cfun
->calls_setjmp
1223 && optimize_function_for_speed_p (cfun
)
1224 && dbg_cnt (store_motion
);
1228 execute_rtl_store_motion (void)
1230 delete_unreachable_blocks ();
1232 flag_rerun_cse_after_global_opts
|= one_store_motion_pass ();
1236 struct rtl_opt_pass pass_rtl_store_motion
=
1240 "store_motion", /* name */
1241 gate_rtl_store_motion
, /* gate */
1242 execute_rtl_store_motion
, /* execute */
1245 0, /* static_pass_number */
1247 PROP_cfglayout
, /* properties_required */
1248 0, /* properties_provided */
1249 0, /* properties_destroyed */
1250 0, /* todo_flags_start */
1251 TODO_df_finish
| TODO_verify_rtl_sharing
|
1252 TODO_verify_flow
| TODO_ggc_collect
/* todo_flags_finish */