1 /* Gimple walk support.
3 Copyright (C) 2007-2014 Free Software Foundation, Inc.
4 Contributed by Aldy Hernandez <aldyh@redhat.com>
6 This file is part of GCC.
8 GCC is free software; you can redistribute it and/or modify it under
9 the terms of the GNU General Public License as published by the Free
10 Software Foundation; either version 3, or (at your option) any later
13 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
14 WARRANTY; without even the implied warranty of MERCHANTABILITY or
15 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
18 You should have received a copy of the GNU General Public License
19 along with GCC; see the file COPYING3. If not see
20 <http://www.gnu.org/licenses/>. */
24 #include "coretypes.h"
28 #include "basic-block.h"
29 #include "tree-ssa-alias.h"
30 #include "internal-fn.h"
31 #include "gimple-expr.h"
34 #include "gimple-iterator.h"
35 #include "gimple-walk.h"
36 #include "gimple-walk.h"
39 /* Walk all the statements in the sequence *PSEQ calling walk_gimple_stmt
40 on each one. WI is as in walk_gimple_stmt.
42 If walk_gimple_stmt returns non-NULL, the walk is stopped, and the
43 value is stored in WI->CALLBACK_RESULT. Also, the statement that
44 produced the value is returned if this statement has not been
45 removed by a callback (wi->removed_stmt). If the statement has
46 been removed, NULL is returned.
48 Otherwise, all the statements are walked and NULL returned. */
51 walk_gimple_seq_mod (gimple_seq
*pseq
, walk_stmt_fn callback_stmt
,
52 walk_tree_fn callback_op
, struct walk_stmt_info
*wi
)
54 gimple_stmt_iterator gsi
;
56 for (gsi
= gsi_start (*pseq
); !gsi_end_p (gsi
); )
58 tree ret
= walk_gimple_stmt (&gsi
, callback_stmt
, callback_op
, wi
);
61 /* If CALLBACK_STMT or CALLBACK_OP return a value, WI must exist
64 wi
->callback_result
= ret
;
66 return wi
->removed_stmt
? NULL
: gsi_stmt (gsi
);
69 if (!wi
->removed_stmt
)
74 wi
->callback_result
= NULL_TREE
;
80 /* Like walk_gimple_seq_mod, but ensure that the head of SEQ isn't
81 changed by the callbacks. */
84 walk_gimple_seq (gimple_seq seq
, walk_stmt_fn callback_stmt
,
85 walk_tree_fn callback_op
, struct walk_stmt_info
*wi
)
87 gimple_seq seq2
= seq
;
88 gimple ret
= walk_gimple_seq_mod (&seq2
, callback_stmt
, callback_op
, wi
);
89 gcc_assert (seq2
== seq
);
94 /* Helper function for walk_gimple_stmt. Walk operands of a GIMPLE_ASM. */
97 walk_gimple_asm (gimple_asm stmt
, walk_tree_fn callback_op
,
98 struct walk_stmt_info
*wi
)
102 const char **oconstraints
;
104 const char *constraint
;
105 bool allows_mem
, allows_reg
, is_inout
;
107 noutputs
= gimple_asm_noutputs (stmt
);
108 oconstraints
= (const char **) alloca ((noutputs
) * sizeof (const char *));
113 for (i
= 0; i
< noutputs
; i
++)
115 op
= gimple_asm_output_op (stmt
, i
);
116 constraint
= TREE_STRING_POINTER (TREE_VALUE (TREE_PURPOSE (op
)));
117 oconstraints
[i
] = constraint
;
118 parse_output_constraint (&constraint
, i
, 0, 0, &allows_mem
, &allows_reg
,
121 wi
->val_only
= (allows_reg
|| !allows_mem
);
122 ret
= walk_tree (&TREE_VALUE (op
), callback_op
, wi
, NULL
);
127 n
= gimple_asm_ninputs (stmt
);
128 for (i
= 0; i
< n
; i
++)
130 op
= gimple_asm_input_op (stmt
, i
);
131 constraint
= TREE_STRING_POINTER (TREE_VALUE (TREE_PURPOSE (op
)));
132 parse_input_constraint (&constraint
, 0, 0, noutputs
, 0,
133 oconstraints
, &allows_mem
, &allows_reg
);
136 wi
->val_only
= (allows_reg
|| !allows_mem
);
137 /* Although input "m" is not really a LHS, we need a lvalue. */
138 wi
->is_lhs
= !wi
->val_only
;
140 ret
= walk_tree (&TREE_VALUE (op
), callback_op
, wi
, NULL
);
151 n
= gimple_asm_nlabels (stmt
);
152 for (i
= 0; i
< n
; i
++)
154 op
= gimple_asm_label_op (stmt
, i
);
155 ret
= walk_tree (&TREE_VALUE (op
), callback_op
, wi
, NULL
);
164 /* Helper function of WALK_GIMPLE_STMT. Walk every tree operand in
165 STMT. CALLBACK_OP and WI are as in WALK_GIMPLE_STMT.
167 CALLBACK_OP is called on each operand of STMT via walk_tree.
168 Additional parameters to walk_tree must be stored in WI. For each operand
169 OP, walk_tree is called as:
171 walk_tree (&OP, CALLBACK_OP, WI, WI->PSET)
173 If CALLBACK_OP returns non-NULL for an operand, the remaining
174 operands are not scanned.
176 The return value is that returned by the last call to walk_tree, or
177 NULL_TREE if no CALLBACK_OP is specified. */
180 walk_gimple_op (gimple stmt
, walk_tree_fn callback_op
,
181 struct walk_stmt_info
*wi
)
183 hash_set
<tree
> *pset
= (wi
) ? wi
->pset
: NULL
;
185 tree ret
= NULL_TREE
;
187 switch (gimple_code (stmt
))
190 /* Walk the RHS operands. If the LHS is of a non-renamable type or
191 is a register variable, we may use a COMPONENT_REF on the RHS. */
194 tree lhs
= gimple_assign_lhs (stmt
);
196 = (is_gimple_reg_type (TREE_TYPE (lhs
)) && !is_gimple_reg (lhs
))
197 || gimple_assign_rhs_class (stmt
) != GIMPLE_SINGLE_RHS
;
200 for (i
= 1; i
< gimple_num_ops (stmt
); i
++)
202 ret
= walk_tree (gimple_op_ptr (stmt
, i
), callback_op
, wi
,
208 /* Walk the LHS. If the RHS is appropriate for a memory, we
209 may use a COMPONENT_REF on the LHS. */
212 /* If the RHS is of a non-renamable type or is a register variable,
213 we may use a COMPONENT_REF on the LHS. */
214 tree rhs1
= gimple_assign_rhs1 (stmt
);
216 = (is_gimple_reg_type (TREE_TYPE (rhs1
)) && !is_gimple_reg (rhs1
))
217 || gimple_assign_rhs_class (stmt
) != GIMPLE_SINGLE_RHS
;
221 ret
= walk_tree (gimple_op_ptr (stmt
, 0), callback_op
, wi
, pset
);
239 ret
= walk_tree (gimple_call_chain_ptr (as_a
<gimple_call
> (stmt
)),
240 callback_op
, wi
, pset
);
244 ret
= walk_tree (gimple_call_fn_ptr (stmt
), callback_op
, wi
, pset
);
248 for (i
= 0; i
< gimple_call_num_args (stmt
); i
++)
252 = is_gimple_reg_type (TREE_TYPE (gimple_call_arg (stmt
, i
)));
253 ret
= walk_tree (gimple_call_arg_ptr (stmt
, i
), callback_op
, wi
,
259 if (gimple_call_lhs (stmt
))
265 = is_gimple_reg_type (TREE_TYPE (gimple_call_lhs (stmt
)));
268 ret
= walk_tree (gimple_call_lhs_ptr (stmt
), callback_op
, wi
, pset
);
281 ret
= walk_tree (gimple_catch_types_ptr (as_a
<gimple_catch
> (stmt
)),
282 callback_op
, wi
, pset
);
287 case GIMPLE_EH_FILTER
:
288 ret
= walk_tree (gimple_eh_filter_types_ptr (stmt
), callback_op
, wi
,
295 ret
= walk_gimple_asm (as_a
<gimple_asm
> (stmt
), callback_op
, wi
);
300 case GIMPLE_OMP_CONTINUE
:
302 gimple_omp_continue cont_stmt
= as_a
<gimple_omp_continue
> (stmt
);
303 ret
= walk_tree (gimple_omp_continue_control_def_ptr (cont_stmt
),
304 callback_op
, wi
, pset
);
308 ret
= walk_tree (gimple_omp_continue_control_use_ptr (cont_stmt
),
309 callback_op
, wi
, pset
);
315 case GIMPLE_OMP_CRITICAL
:
316 ret
= walk_tree (gimple_omp_critical_name_ptr (
317 as_a
<gimple_omp_critical
> (stmt
)),
318 callback_op
, wi
, pset
);
324 ret
= walk_tree (gimple_omp_for_clauses_ptr (stmt
), callback_op
, wi
,
328 for (i
= 0; i
< gimple_omp_for_collapse (stmt
); i
++)
330 ret
= walk_tree (gimple_omp_for_index_ptr (stmt
, i
), callback_op
,
334 ret
= walk_tree (gimple_omp_for_initial_ptr (stmt
, i
), callback_op
,
338 ret
= walk_tree (gimple_omp_for_final_ptr (stmt
, i
), callback_op
,
342 ret
= walk_tree (gimple_omp_for_incr_ptr (stmt
, i
), callback_op
,
349 case GIMPLE_OMP_PARALLEL
:
351 gimple_omp_parallel omp_par_stmt
= as_a
<gimple_omp_parallel
> (stmt
);
352 ret
= walk_tree (gimple_omp_parallel_clauses_ptr (omp_par_stmt
),
353 callback_op
, wi
, pset
);
356 ret
= walk_tree (gimple_omp_parallel_child_fn_ptr (omp_par_stmt
),
357 callback_op
, wi
, pset
);
360 ret
= walk_tree (gimple_omp_parallel_data_arg_ptr (omp_par_stmt
),
361 callback_op
, wi
, pset
);
367 case GIMPLE_OMP_TASK
:
368 ret
= walk_tree (gimple_omp_task_clauses_ptr (stmt
), callback_op
,
372 ret
= walk_tree (gimple_omp_task_child_fn_ptr (stmt
), callback_op
,
376 ret
= walk_tree (gimple_omp_task_data_arg_ptr (stmt
), callback_op
,
380 ret
= walk_tree (gimple_omp_task_copy_fn_ptr (stmt
), callback_op
,
384 ret
= walk_tree (gimple_omp_task_arg_size_ptr (stmt
), callback_op
,
388 ret
= walk_tree (gimple_omp_task_arg_align_ptr (stmt
), callback_op
,
394 case GIMPLE_OMP_SECTIONS
:
395 ret
= walk_tree (gimple_omp_sections_clauses_ptr (stmt
), callback_op
,
400 ret
= walk_tree (gimple_omp_sections_control_ptr (stmt
), callback_op
,
407 case GIMPLE_OMP_SINGLE
:
408 ret
= walk_tree (gimple_omp_single_clauses_ptr (stmt
), callback_op
, wi
,
414 case GIMPLE_OMP_TARGET
:
415 ret
= walk_tree (gimple_omp_target_clauses_ptr (stmt
), callback_op
, wi
,
421 case GIMPLE_OMP_TEAMS
:
422 ret
= walk_tree (gimple_omp_teams_clauses_ptr (stmt
), callback_op
, wi
,
428 case GIMPLE_OMP_ATOMIC_LOAD
:
430 gimple_omp_atomic_load omp_stmt
= as_a
<gimple_omp_atomic_load
> (stmt
);
431 ret
= walk_tree (gimple_omp_atomic_load_lhs_ptr (omp_stmt
),
432 callback_op
, wi
, pset
);
436 ret
= walk_tree (gimple_omp_atomic_load_rhs_ptr (omp_stmt
),
437 callback_op
, wi
, pset
);
443 case GIMPLE_OMP_ATOMIC_STORE
:
444 ret
= walk_tree (gimple_omp_atomic_store_val_ptr (
445 as_a
<gimple_omp_atomic_store
> (stmt
)),
446 callback_op
, wi
, pset
);
451 case GIMPLE_TRANSACTION
:
452 ret
= walk_tree (gimple_transaction_label_ptr (
453 as_a
<gimple_transaction
> (stmt
)),
454 callback_op
, wi
, pset
);
459 case GIMPLE_OMP_RETURN
:
460 ret
= walk_tree (gimple_omp_return_lhs_ptr (stmt
), callback_op
, wi
,
466 /* Tuples that do not have operands. */
474 enum gimple_statement_structure_enum gss
;
475 gss
= gimple_statement_structure (stmt
);
476 if (gss
== GSS_WITH_OPS
|| gss
== GSS_WITH_MEM_OPS
)
477 for (i
= 0; i
< gimple_num_ops (stmt
); i
++)
479 ret
= walk_tree (gimple_op_ptr (stmt
, i
), callback_op
, wi
, pset
);
491 /* Walk the current statement in GSI (optionally using traversal state
492 stored in WI). If WI is NULL, no state is kept during traversal.
493 The callback CALLBACK_STMT is called. If CALLBACK_STMT indicates
494 that it has handled all the operands of the statement, its return
495 value is returned. Otherwise, the return value from CALLBACK_STMT
496 is discarded and its operands are scanned.
498 If CALLBACK_STMT is NULL or it didn't handle the operands,
499 CALLBACK_OP is called on each operand of the statement via
500 walk_gimple_op. If walk_gimple_op returns non-NULL for any
501 operand, the remaining operands are not scanned. In this case, the
502 return value from CALLBACK_OP is returned.
504 In any other case, NULL_TREE is returned. */
507 walk_gimple_stmt (gimple_stmt_iterator
*gsi
, walk_stmt_fn callback_stmt
,
508 walk_tree_fn callback_op
, struct walk_stmt_info
*wi
)
512 gimple stmt
= gsi_stmt (*gsi
);
517 wi
->removed_stmt
= false;
519 if (wi
->want_locations
&& gimple_has_location (stmt
))
520 input_location
= gimple_location (stmt
);
525 /* Invoke the statement callback. Return if the callback handled
526 all of STMT operands by itself. */
529 bool handled_ops
= false;
530 tree_ret
= callback_stmt (gsi
, &handled_ops
, wi
);
534 /* If CALLBACK_STMT did not handle operands, it should not have
535 a value to return. */
536 gcc_assert (tree_ret
== NULL
);
538 if (wi
&& wi
->removed_stmt
)
541 /* Re-read stmt in case the callback changed it. */
542 stmt
= gsi_stmt (*gsi
);
545 /* If CALLBACK_OP is defined, invoke it on every operand of STMT. */
548 tree_ret
= walk_gimple_op (stmt
, callback_op
, wi
);
553 /* If STMT can have statements inside (e.g. GIMPLE_BIND), walk them. */
554 switch (gimple_code (stmt
))
558 walk_gimple_seq_mod (gimple_bind_body_ptr (as_a
<gimple_bind
> (stmt
)),
559 callback_stmt
, callback_op
, wi
);
561 return wi
->callback_result
;
565 ret
= walk_gimple_seq_mod (gimple_catch_handler_ptr (
566 as_a
<gimple_catch
> (stmt
)),
567 callback_stmt
, callback_op
, wi
);
569 return wi
->callback_result
;
572 case GIMPLE_EH_FILTER
:
573 ret
= walk_gimple_seq_mod (gimple_eh_filter_failure_ptr (stmt
), callback_stmt
,
576 return wi
->callback_result
;
581 gimple_eh_else eh_else_stmt
= as_a
<gimple_eh_else
> (stmt
);
582 ret
= walk_gimple_seq_mod (gimple_eh_else_n_body_ptr (eh_else_stmt
),
583 callback_stmt
, callback_op
, wi
);
585 return wi
->callback_result
;
586 ret
= walk_gimple_seq_mod (gimple_eh_else_e_body_ptr (eh_else_stmt
),
587 callback_stmt
, callback_op
, wi
);
589 return wi
->callback_result
;
594 ret
= walk_gimple_seq_mod (gimple_try_eval_ptr (stmt
), callback_stmt
, callback_op
,
597 return wi
->callback_result
;
599 ret
= walk_gimple_seq_mod (gimple_try_cleanup_ptr (stmt
), callback_stmt
,
602 return wi
->callback_result
;
606 ret
= walk_gimple_seq_mod (gimple_omp_for_pre_body_ptr (stmt
), callback_stmt
,
609 return wi
->callback_result
;
612 case GIMPLE_OMP_CRITICAL
:
613 case GIMPLE_OMP_MASTER
:
614 case GIMPLE_OMP_TASKGROUP
:
615 case GIMPLE_OMP_ORDERED
:
616 case GIMPLE_OMP_SECTION
:
617 case GIMPLE_OMP_PARALLEL
:
618 case GIMPLE_OMP_TASK
:
619 case GIMPLE_OMP_SECTIONS
:
620 case GIMPLE_OMP_SINGLE
:
621 case GIMPLE_OMP_TARGET
:
622 case GIMPLE_OMP_TEAMS
:
623 ret
= walk_gimple_seq_mod (gimple_omp_body_ptr (stmt
), callback_stmt
,
626 return wi
->callback_result
;
629 case GIMPLE_WITH_CLEANUP_EXPR
:
630 ret
= walk_gimple_seq_mod (gimple_wce_cleanup_ptr (stmt
), callback_stmt
,
633 return wi
->callback_result
;
636 case GIMPLE_TRANSACTION
:
637 ret
= walk_gimple_seq_mod (gimple_transaction_body_ptr (
638 as_a
<gimple_transaction
> (stmt
)),
639 callback_stmt
, callback_op
, wi
);
641 return wi
->callback_result
;
645 gcc_assert (!gimple_has_substatements (stmt
));
652 /* From a tree operand OP return the base of a load or store operation
653 or NULL_TREE if OP is not a load or a store. */
656 get_base_loadstore (tree op
)
658 while (handled_component_p (op
))
659 op
= TREE_OPERAND (op
, 0);
661 || INDIRECT_REF_P (op
)
662 || TREE_CODE (op
) == MEM_REF
663 || TREE_CODE (op
) == TARGET_MEM_REF
)
669 /* For the statement STMT call the callbacks VISIT_LOAD, VISIT_STORE and
670 VISIT_ADDR if non-NULL on loads, store and address-taken operands
671 passing the STMT, the base of the operand, the operand itself containing
672 the base and DATA to it. The base will be either a decl, an indirect
673 reference (including TARGET_MEM_REF) or the argument of an address
675 Returns the results of these callbacks or'ed. */
678 walk_stmt_load_store_addr_ops (gimple stmt
, void *data
,
679 walk_stmt_load_store_addr_fn visit_load
,
680 walk_stmt_load_store_addr_fn visit_store
,
681 walk_stmt_load_store_addr_fn visit_addr
)
685 if (gimple_assign_single_p (stmt
))
690 arg
= gimple_assign_lhs (stmt
);
691 lhs
= get_base_loadstore (arg
);
693 ret
|= visit_store (stmt
, lhs
, arg
, data
);
695 arg
= gimple_assign_rhs1 (stmt
);
697 while (handled_component_p (rhs
))
698 rhs
= TREE_OPERAND (rhs
, 0);
701 if (TREE_CODE (rhs
) == ADDR_EXPR
)
702 ret
|= visit_addr (stmt
, TREE_OPERAND (rhs
, 0), arg
, data
);
703 else if (TREE_CODE (rhs
) == TARGET_MEM_REF
704 && TREE_CODE (TMR_BASE (rhs
)) == ADDR_EXPR
)
705 ret
|= visit_addr (stmt
, TREE_OPERAND (TMR_BASE (rhs
), 0), arg
,
707 else if (TREE_CODE (rhs
) == OBJ_TYPE_REF
708 && TREE_CODE (OBJ_TYPE_REF_OBJECT (rhs
)) == ADDR_EXPR
)
709 ret
|= visit_addr (stmt
, TREE_OPERAND (OBJ_TYPE_REF_OBJECT (rhs
),
711 else if (TREE_CODE (rhs
) == CONSTRUCTOR
)
716 FOR_EACH_CONSTRUCTOR_VALUE (CONSTRUCTOR_ELTS (rhs
), ix
, val
)
717 if (TREE_CODE (val
) == ADDR_EXPR
)
718 ret
|= visit_addr (stmt
, TREE_OPERAND (val
, 0), arg
, data
);
719 else if (TREE_CODE (val
) == OBJ_TYPE_REF
720 && TREE_CODE (OBJ_TYPE_REF_OBJECT (val
)) == ADDR_EXPR
)
721 ret
|= visit_addr (stmt
,
722 TREE_OPERAND (OBJ_TYPE_REF_OBJECT (val
),
725 lhs
= gimple_assign_lhs (stmt
);
726 if (TREE_CODE (lhs
) == TARGET_MEM_REF
727 && TREE_CODE (TMR_BASE (lhs
)) == ADDR_EXPR
)
728 ret
|= visit_addr (stmt
, TREE_OPERAND (TMR_BASE (lhs
), 0), lhs
, data
);
732 rhs
= get_base_loadstore (rhs
);
734 ret
|= visit_load (stmt
, rhs
, arg
, data
);
738 && (is_gimple_assign (stmt
)
739 || gimple_code (stmt
) == GIMPLE_COND
))
741 for (i
= 0; i
< gimple_num_ops (stmt
); ++i
)
743 tree op
= gimple_op (stmt
, i
);
746 else if (TREE_CODE (op
) == ADDR_EXPR
)
747 ret
|= visit_addr (stmt
, TREE_OPERAND (op
, 0), op
, data
);
748 /* COND_EXPR and VCOND_EXPR rhs1 argument is a comparison
749 tree with two operands. */
750 else if (i
== 1 && COMPARISON_CLASS_P (op
))
752 if (TREE_CODE (TREE_OPERAND (op
, 0)) == ADDR_EXPR
)
753 ret
|= visit_addr (stmt
, TREE_OPERAND (TREE_OPERAND (op
, 0),
755 if (TREE_CODE (TREE_OPERAND (op
, 1)) == ADDR_EXPR
)
756 ret
|= visit_addr (stmt
, TREE_OPERAND (TREE_OPERAND (op
, 1),
761 else if (gimple_call call_stmt
= dyn_cast
<gimple_call
> (stmt
))
765 tree arg
= gimple_call_lhs (call_stmt
);
768 tree lhs
= get_base_loadstore (arg
);
770 ret
|= visit_store (stmt
, lhs
, arg
, data
);
773 if (visit_load
|| visit_addr
)
774 for (i
= 0; i
< gimple_call_num_args (call_stmt
); ++i
)
776 tree arg
= gimple_call_arg (call_stmt
, i
);
778 && TREE_CODE (arg
) == ADDR_EXPR
)
779 ret
|= visit_addr (stmt
, TREE_OPERAND (arg
, 0), arg
, data
);
782 tree rhs
= get_base_loadstore (arg
);
784 ret
|= visit_load (stmt
, rhs
, arg
, data
);
788 && gimple_call_chain (call_stmt
)
789 && TREE_CODE (gimple_call_chain (call_stmt
)) == ADDR_EXPR
)
790 ret
|= visit_addr (stmt
, TREE_OPERAND (gimple_call_chain (call_stmt
), 0),
791 gimple_call_chain (call_stmt
), data
);
793 && gimple_call_return_slot_opt_p (call_stmt
)
794 && gimple_call_lhs (call_stmt
) != NULL_TREE
795 && TREE_ADDRESSABLE (TREE_TYPE (gimple_call_lhs (call_stmt
))))
796 ret
|= visit_addr (stmt
, gimple_call_lhs (call_stmt
),
797 gimple_call_lhs (call_stmt
), data
);
799 else if (gimple_asm asm_stmt
= dyn_cast
<gimple_asm
> (stmt
))
802 const char *constraint
;
803 const char **oconstraints
;
804 bool allows_mem
, allows_reg
, is_inout
;
805 noutputs
= gimple_asm_noutputs (asm_stmt
);
806 oconstraints
= XALLOCAVEC (const char *, noutputs
);
807 if (visit_store
|| visit_addr
)
808 for (i
= 0; i
< gimple_asm_noutputs (asm_stmt
); ++i
)
810 tree link
= gimple_asm_output_op (asm_stmt
, i
);
811 tree op
= get_base_loadstore (TREE_VALUE (link
));
812 if (op
&& visit_store
)
813 ret
|= visit_store (stmt
, op
, TREE_VALUE (link
), data
);
816 constraint
= TREE_STRING_POINTER
817 (TREE_VALUE (TREE_PURPOSE (link
)));
818 oconstraints
[i
] = constraint
;
819 parse_output_constraint (&constraint
, i
, 0, 0, &allows_mem
,
820 &allows_reg
, &is_inout
);
821 if (op
&& !allows_reg
&& allows_mem
)
822 ret
|= visit_addr (stmt
, op
, TREE_VALUE (link
), data
);
825 if (visit_load
|| visit_addr
)
826 for (i
= 0; i
< gimple_asm_ninputs (asm_stmt
); ++i
)
828 tree link
= gimple_asm_input_op (asm_stmt
, i
);
829 tree op
= TREE_VALUE (link
);
831 && TREE_CODE (op
) == ADDR_EXPR
)
832 ret
|= visit_addr (stmt
, TREE_OPERAND (op
, 0), op
, data
);
833 else if (visit_load
|| visit_addr
)
835 op
= get_base_loadstore (op
);
839 ret
|= visit_load (stmt
, op
, TREE_VALUE (link
), data
);
842 constraint
= TREE_STRING_POINTER
843 (TREE_VALUE (TREE_PURPOSE (link
)));
844 parse_input_constraint (&constraint
, 0, 0, noutputs
,
846 &allows_mem
, &allows_reg
);
847 if (!allows_reg
&& allows_mem
)
848 ret
|= visit_addr (stmt
, op
, TREE_VALUE (link
),
855 else if (gimple_return return_stmt
= dyn_cast
<gimple_return
> (stmt
))
857 tree op
= gimple_return_retval (return_stmt
);
861 && TREE_CODE (op
) == ADDR_EXPR
)
862 ret
|= visit_addr (stmt
, TREE_OPERAND (op
, 0), op
, data
);
865 tree base
= get_base_loadstore (op
);
867 ret
|= visit_load (stmt
, base
, op
, data
);
872 && gimple_code (stmt
) == GIMPLE_PHI
)
874 for (i
= 0; i
< gimple_phi_num_args (stmt
); ++i
)
876 tree op
= gimple_phi_arg_def (stmt
, i
);
877 if (TREE_CODE (op
) == ADDR_EXPR
)
878 ret
|= visit_addr (stmt
, TREE_OPERAND (op
, 0), op
, data
);
882 && gimple_code (stmt
) == GIMPLE_GOTO
)
884 tree op
= gimple_goto_dest (stmt
);
885 if (TREE_CODE (op
) == ADDR_EXPR
)
886 ret
|= visit_addr (stmt
, TREE_OPERAND (op
, 0), op
, data
);
892 /* Like walk_stmt_load_store_addr_ops but with NULL visit_addr. IPA-CP
893 should make a faster clone for this case. */
896 walk_stmt_load_store_ops (gimple stmt
, void *data
,
897 walk_stmt_load_store_addr_fn visit_load
,
898 walk_stmt_load_store_addr_fn visit_store
)
900 return walk_stmt_load_store_addr_ops (stmt
, data
,
901 visit_load
, visit_store
, NULL
);