Merged r157653 through r157895 into branch.
[official-gcc.git] / gcc / gimplify.c
blob8f6223c40c6768b1e8cedd98bc50686964b361e2
1 /* Tree lowering pass. This pass converts the GENERIC functions-as-trees
2 tree representation into the GIMPLE form.
3 Copyright (C) 2002, 2003, 2004, 2005, 2006, 2007, 2008, 2009, 2010
4 Free Software Foundation, Inc.
5 Major work done by Sebastian Pop <s.pop@laposte.net>,
6 Diego Novillo <dnovillo@redhat.com> and Jason Merrill <jason@redhat.com>.
8 This file is part of GCC.
10 GCC is free software; you can redistribute it and/or modify it under
11 the terms of the GNU General Public License as published by the Free
12 Software Foundation; either version 3, or (at your option) any later
13 version.
15 GCC is distributed in the hope that it will be useful, but WITHOUT ANY
16 WARRANTY; without even the implied warranty of MERCHANTABILITY or
17 FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
18 for more details.
20 You should have received a copy of the GNU General Public License
21 along with GCC; see the file COPYING3. If not see
22 <http://www.gnu.org/licenses/>. */
24 #include "config.h"
25 #include "system.h"
26 #include "coretypes.h"
27 #include "tm.h"
28 #include "tree.h"
29 #include "rtl.h"
30 #include "varray.h"
31 #include "gimple.h"
32 #include "tree-iterator.h"
33 #include "tree-inline.h"
34 #include "diagnostic.h"
35 #include "langhooks.h"
36 #include "langhooks-def.h"
37 #include "tree-flow.h"
38 #include "cgraph.h"
39 #include "timevar.h"
40 #include "except.h"
41 #include "hashtab.h"
42 #include "flags.h"
43 #include "real.h"
44 #include "function.h"
45 #include "output.h"
46 #include "expr.h"
47 #include "ggc.h"
48 #include "toplev.h"
49 #include "target.h"
50 #include "optabs.h"
51 #include "pointer-set.h"
52 #include "splay-tree.h"
53 #include "vec.h"
54 #include "gimple.h"
55 #include "tree-pass.h"
58 enum gimplify_omp_var_data
60 GOVD_SEEN = 1,
61 GOVD_EXPLICIT = 2,
62 GOVD_SHARED = 4,
63 GOVD_PRIVATE = 8,
64 GOVD_FIRSTPRIVATE = 16,
65 GOVD_LASTPRIVATE = 32,
66 GOVD_REDUCTION = 64,
67 GOVD_LOCAL = 128,
68 GOVD_DEBUG_PRIVATE = 256,
69 GOVD_PRIVATE_OUTER_REF = 512,
70 GOVD_DATA_SHARE_CLASS = (GOVD_SHARED | GOVD_PRIVATE | GOVD_FIRSTPRIVATE
71 | GOVD_LASTPRIVATE | GOVD_REDUCTION | GOVD_LOCAL)
75 enum omp_region_type
77 ORT_WORKSHARE = 0,
78 ORT_TASK = 1,
79 ORT_PARALLEL = 2,
80 ORT_COMBINED_PARALLEL = 3
83 struct gimplify_omp_ctx
85 struct gimplify_omp_ctx *outer_context;
86 splay_tree variables;
87 struct pointer_set_t *privatized_types;
88 location_t location;
89 enum omp_clause_default_kind default_kind;
90 enum omp_region_type region_type;
93 static struct gimplify_ctx *gimplify_ctxp;
94 static struct gimplify_omp_ctx *gimplify_omp_ctxp;
97 /* Formal (expression) temporary table handling: Multiple occurrences of
98 the same scalar expression are evaluated into the same temporary. */
100 typedef struct gimple_temp_hash_elt
102 tree val; /* Key */
103 tree temp; /* Value */
104 } elt_t;
106 /* Forward declarations. */
107 static enum gimplify_status gimplify_compound_expr (tree *, gimple_seq *, bool);
109 /* Mark X addressable. Unlike the langhook we expect X to be in gimple
110 form and we don't do any syntax checking. */
111 void
112 mark_addressable (tree x)
114 while (handled_component_p (x))
115 x = TREE_OPERAND (x, 0);
116 if (TREE_CODE (x) != VAR_DECL
117 && TREE_CODE (x) != PARM_DECL
118 && TREE_CODE (x) != RESULT_DECL)
119 return ;
120 TREE_ADDRESSABLE (x) = 1;
123 /* Return a hash value for a formal temporary table entry. */
125 static hashval_t
126 gimple_tree_hash (const void *p)
128 tree t = ((const elt_t *) p)->val;
129 return iterative_hash_expr (t, 0);
132 /* Compare two formal temporary table entries. */
134 static int
135 gimple_tree_eq (const void *p1, const void *p2)
137 tree t1 = ((const elt_t *) p1)->val;
138 tree t2 = ((const elt_t *) p2)->val;
139 enum tree_code code = TREE_CODE (t1);
141 if (TREE_CODE (t2) != code
142 || TREE_TYPE (t1) != TREE_TYPE (t2))
143 return 0;
145 if (!operand_equal_p (t1, t2, 0))
146 return 0;
148 /* Only allow them to compare equal if they also hash equal; otherwise
149 results are nondeterminate, and we fail bootstrap comparison. */
150 gcc_assert (gimple_tree_hash (p1) == gimple_tree_hash (p2));
152 return 1;
155 /* Link gimple statement GS to the end of the sequence *SEQ_P. If
156 *SEQ_P is NULL, a new sequence is allocated. This function is
157 similar to gimple_seq_add_stmt, but does not scan the operands.
158 During gimplification, we need to manipulate statement sequences
159 before the def/use vectors have been constructed. */
161 static void
162 gimplify_seq_add_stmt (gimple_seq *seq_p, gimple gs)
164 gimple_stmt_iterator si;
166 if (gs == NULL)
167 return;
169 if (*seq_p == NULL)
170 *seq_p = gimple_seq_alloc ();
172 si = gsi_last (*seq_p);
174 gsi_insert_after_without_update (&si, gs, GSI_NEW_STMT);
177 /* Append sequence SRC to the end of sequence *DST_P. If *DST_P is
178 NULL, a new sequence is allocated. This function is
179 similar to gimple_seq_add_seq, but does not scan the operands.
180 During gimplification, we need to manipulate statement sequences
181 before the def/use vectors have been constructed. */
183 static void
184 gimplify_seq_add_seq (gimple_seq *dst_p, gimple_seq src)
186 gimple_stmt_iterator si;
188 if (src == NULL)
189 return;
191 if (*dst_p == NULL)
192 *dst_p = gimple_seq_alloc ();
194 si = gsi_last (*dst_p);
195 gsi_insert_seq_after_without_update (&si, src, GSI_NEW_STMT);
198 /* Set up a context for the gimplifier. */
200 void
201 push_gimplify_context (struct gimplify_ctx *c)
203 memset (c, '\0', sizeof (*c));
204 c->prev_context = gimplify_ctxp;
205 gimplify_ctxp = c;
208 /* Tear down a context for the gimplifier. If BODY is non-null, then
209 put the temporaries into the outer BIND_EXPR. Otherwise, put them
210 in the local_decls.
212 BODY is not a sequence, but the first tuple in a sequence. */
214 void
215 pop_gimplify_context (gimple body)
217 struct gimplify_ctx *c = gimplify_ctxp;
219 gcc_assert (c && (c->bind_expr_stack == NULL
220 || VEC_empty (gimple, c->bind_expr_stack)));
221 VEC_free (gimple, heap, c->bind_expr_stack);
222 gimplify_ctxp = c->prev_context;
224 if (body)
225 declare_vars (c->temps, body, false);
226 else
227 record_vars (c->temps);
229 if (c->temp_htab)
230 htab_delete (c->temp_htab);
233 static void
234 gimple_push_bind_expr (gimple gimple_bind)
236 if (gimplify_ctxp->bind_expr_stack == NULL)
237 gimplify_ctxp->bind_expr_stack = VEC_alloc (gimple, heap, 8);
238 VEC_safe_push (gimple, heap, gimplify_ctxp->bind_expr_stack, gimple_bind);
241 static void
242 gimple_pop_bind_expr (void)
244 VEC_pop (gimple, gimplify_ctxp->bind_expr_stack);
247 gimple
248 gimple_current_bind_expr (void)
250 return VEC_last (gimple, gimplify_ctxp->bind_expr_stack);
253 /* Return the stack GIMPLE_BINDs created during gimplification. */
255 VEC(gimple, heap) *
256 gimple_bind_expr_stack (void)
258 return gimplify_ctxp->bind_expr_stack;
261 /* Returns true iff there is a COND_EXPR between us and the innermost
262 CLEANUP_POINT_EXPR. This info is used by gimple_push_cleanup. */
264 static bool
265 gimple_conditional_context (void)
267 return gimplify_ctxp->conditions > 0;
270 /* Note that we've entered a COND_EXPR. */
272 static void
273 gimple_push_condition (void)
275 #ifdef ENABLE_GIMPLE_CHECKING
276 if (gimplify_ctxp->conditions == 0)
277 gcc_assert (gimple_seq_empty_p (gimplify_ctxp->conditional_cleanups));
278 #endif
279 ++(gimplify_ctxp->conditions);
282 /* Note that we've left a COND_EXPR. If we're back at unconditional scope
283 now, add any conditional cleanups we've seen to the prequeue. */
285 static void
286 gimple_pop_condition (gimple_seq *pre_p)
288 int conds = --(gimplify_ctxp->conditions);
290 gcc_assert (conds >= 0);
291 if (conds == 0)
293 gimplify_seq_add_seq (pre_p, gimplify_ctxp->conditional_cleanups);
294 gimplify_ctxp->conditional_cleanups = NULL;
298 /* A stable comparison routine for use with splay trees and DECLs. */
300 static int
301 splay_tree_compare_decl_uid (splay_tree_key xa, splay_tree_key xb)
303 tree a = (tree) xa;
304 tree b = (tree) xb;
306 return DECL_UID (a) - DECL_UID (b);
309 /* Create a new omp construct that deals with variable remapping. */
311 static struct gimplify_omp_ctx *
312 new_omp_context (enum omp_region_type region_type)
314 struct gimplify_omp_ctx *c;
316 c = XCNEW (struct gimplify_omp_ctx);
317 c->outer_context = gimplify_omp_ctxp;
318 c->variables = splay_tree_new (splay_tree_compare_decl_uid, 0, 0);
319 c->privatized_types = pointer_set_create ();
320 c->location = input_location;
321 c->region_type = region_type;
322 if (region_type != ORT_TASK)
323 c->default_kind = OMP_CLAUSE_DEFAULT_SHARED;
324 else
325 c->default_kind = OMP_CLAUSE_DEFAULT_UNSPECIFIED;
327 return c;
330 /* Destroy an omp construct that deals with variable remapping. */
332 static void
333 delete_omp_context (struct gimplify_omp_ctx *c)
335 splay_tree_delete (c->variables);
336 pointer_set_destroy (c->privatized_types);
337 XDELETE (c);
340 static void omp_add_variable (struct gimplify_omp_ctx *, tree, unsigned int);
341 static bool omp_notice_variable (struct gimplify_omp_ctx *, tree, bool);
343 /* A subroutine of append_to_statement_list{,_force}. T is not NULL. */
345 static void
346 append_to_statement_list_1 (tree t, tree *list_p)
348 tree list = *list_p;
349 tree_stmt_iterator i;
351 if (!list)
353 if (t && TREE_CODE (t) == STATEMENT_LIST)
355 *list_p = t;
356 return;
358 *list_p = list = alloc_stmt_list ();
361 i = tsi_last (list);
362 tsi_link_after (&i, t, TSI_CONTINUE_LINKING);
365 /* Add T to the end of the list container pointed to by LIST_P.
366 If T is an expression with no effects, it is ignored. */
368 void
369 append_to_statement_list (tree t, tree *list_p)
371 if (t && TREE_SIDE_EFFECTS (t))
372 append_to_statement_list_1 (t, list_p);
375 /* Similar, but the statement is always added, regardless of side effects. */
377 void
378 append_to_statement_list_force (tree t, tree *list_p)
380 if (t != NULL_TREE)
381 append_to_statement_list_1 (t, list_p);
384 /* Both gimplify the statement T and append it to *SEQ_P. This function
385 behaves exactly as gimplify_stmt, but you don't have to pass T as a
386 reference. */
388 void
389 gimplify_and_add (tree t, gimple_seq *seq_p)
391 gimplify_stmt (&t, seq_p);
394 /* Gimplify statement T into sequence *SEQ_P, and return the first
395 tuple in the sequence of generated tuples for this statement.
396 Return NULL if gimplifying T produced no tuples. */
398 static gimple
399 gimplify_and_return_first (tree t, gimple_seq *seq_p)
401 gimple_stmt_iterator last = gsi_last (*seq_p);
403 gimplify_and_add (t, seq_p);
405 if (!gsi_end_p (last))
407 gsi_next (&last);
408 return gsi_stmt (last);
410 else
411 return gimple_seq_first_stmt (*seq_p);
414 /* Strip off a legitimate source ending from the input string NAME of
415 length LEN. Rather than having to know the names used by all of
416 our front ends, we strip off an ending of a period followed by
417 up to five characters. (Java uses ".class".) */
419 static inline void
420 remove_suffix (char *name, int len)
422 int i;
424 for (i = 2; i < 8 && len > i; i++)
426 if (name[len - i] == '.')
428 name[len - i] = '\0';
429 break;
434 /* Create a new temporary name with PREFIX. Returns an identifier. */
436 static GTY(()) unsigned int tmp_var_id_num;
438 tree
439 create_tmp_var_name (const char *prefix)
441 char *tmp_name;
443 if (prefix)
445 char *preftmp = ASTRDUP (prefix);
447 remove_suffix (preftmp, strlen (preftmp));
448 prefix = preftmp;
451 ASM_FORMAT_PRIVATE_NAME (tmp_name, prefix ? prefix : "T", tmp_var_id_num++);
452 return get_identifier (tmp_name);
456 /* Create a new temporary variable declaration of type TYPE.
457 Does NOT push it into the current binding. */
459 tree
460 create_tmp_var_raw (tree type, const char *prefix)
462 tree tmp_var;
463 tree new_type;
465 /* Make the type of the variable writable. */
466 new_type = build_type_variant (type, 0, 0);
467 TYPE_ATTRIBUTES (new_type) = TYPE_ATTRIBUTES (type);
469 tmp_var = build_decl (input_location,
470 VAR_DECL, prefix ? create_tmp_var_name (prefix) : NULL,
471 type);
473 /* The variable was declared by the compiler. */
474 DECL_ARTIFICIAL (tmp_var) = 1;
475 /* And we don't want debug info for it. */
476 DECL_IGNORED_P (tmp_var) = 1;
478 /* Make the variable writable. */
479 TREE_READONLY (tmp_var) = 0;
481 DECL_EXTERNAL (tmp_var) = 0;
482 TREE_STATIC (tmp_var) = 0;
483 TREE_USED (tmp_var) = 1;
485 return tmp_var;
488 /* Create a new temporary variable declaration of type TYPE. DOES push the
489 variable into the current binding. Further, assume that this is called
490 only from gimplification or optimization, at which point the creation of
491 certain types are bugs. */
493 tree
494 create_tmp_var (tree type, const char *prefix)
496 tree tmp_var;
498 /* We don't allow types that are addressable (meaning we can't make copies),
499 or incomplete. We also used to reject every variable size objects here,
500 but now support those for which a constant upper bound can be obtained.
501 The processing for variable sizes is performed in gimple_add_tmp_var,
502 point at which it really matters and possibly reached via paths not going
503 through this function, e.g. after direct calls to create_tmp_var_raw. */
504 gcc_assert (!TREE_ADDRESSABLE (type) && COMPLETE_TYPE_P (type));
506 tmp_var = create_tmp_var_raw (type, prefix);
507 gimple_add_tmp_var (tmp_var);
508 return tmp_var;
511 /* Create a temporary with a name derived from VAL. Subroutine of
512 lookup_tmp_var; nobody else should call this function. */
514 static inline tree
515 create_tmp_from_val (tree val)
517 return create_tmp_var (TREE_TYPE (val), get_name (val));
520 /* Create a temporary to hold the value of VAL. If IS_FORMAL, try to reuse
521 an existing expression temporary. */
523 static tree
524 lookup_tmp_var (tree val, bool is_formal)
526 tree ret;
528 /* If not optimizing, never really reuse a temporary. local-alloc
529 won't allocate any variable that is used in more than one basic
530 block, which means it will go into memory, causing much extra
531 work in reload and final and poorer code generation, outweighing
532 the extra memory allocation here. */
533 if (!optimize || !is_formal || TREE_SIDE_EFFECTS (val))
534 ret = create_tmp_from_val (val);
535 else
537 elt_t elt, *elt_p;
538 void **slot;
540 elt.val = val;
541 if (gimplify_ctxp->temp_htab == NULL)
542 gimplify_ctxp->temp_htab
543 = htab_create (1000, gimple_tree_hash, gimple_tree_eq, free);
544 slot = htab_find_slot (gimplify_ctxp->temp_htab, (void *)&elt, INSERT);
545 if (*slot == NULL)
547 elt_p = XNEW (elt_t);
548 elt_p->val = val;
549 elt_p->temp = ret = create_tmp_from_val (val);
550 *slot = (void *) elt_p;
552 else
554 elt_p = (elt_t *) *slot;
555 ret = elt_p->temp;
559 return ret;
563 /* Return true if T is a CALL_EXPR or an expression that can be
564 assignmed to a temporary. Note that this predicate should only be
565 used during gimplification. See the rationale for this in
566 gimplify_modify_expr. */
568 static bool
569 is_gimple_reg_rhs_or_call (tree t)
571 return (get_gimple_rhs_class (TREE_CODE (t)) != GIMPLE_INVALID_RHS
572 || TREE_CODE (t) == CALL_EXPR);
575 /* Return true if T is a valid memory RHS or a CALL_EXPR. Note that
576 this predicate should only be used during gimplification. See the
577 rationale for this in gimplify_modify_expr. */
579 static bool
580 is_gimple_mem_rhs_or_call (tree t)
582 /* If we're dealing with a renamable type, either source or dest must be
583 a renamed variable. */
584 if (is_gimple_reg_type (TREE_TYPE (t)))
585 return is_gimple_val (t);
586 else
587 return (is_gimple_val (t) || is_gimple_lvalue (t)
588 || TREE_CODE (t) == CALL_EXPR);
591 /* Helper for get_formal_tmp_var and get_initialized_tmp_var. */
593 static tree
594 internal_get_tmp_var (tree val, gimple_seq *pre_p, gimple_seq *post_p,
595 bool is_formal)
597 tree t, mod;
599 /* Notice that we explicitly allow VAL to be a CALL_EXPR so that we
600 can create an INIT_EXPR and convert it into a GIMPLE_CALL below. */
601 gimplify_expr (&val, pre_p, post_p, is_gimple_reg_rhs_or_call,
602 fb_rvalue);
604 t = lookup_tmp_var (val, is_formal);
606 if (is_formal
607 && (TREE_CODE (TREE_TYPE (t)) == COMPLEX_TYPE
608 || TREE_CODE (TREE_TYPE (t)) == VECTOR_TYPE))
609 DECL_GIMPLE_REG_P (t) = 1;
611 mod = build2 (INIT_EXPR, TREE_TYPE (t), t, unshare_expr (val));
613 if (EXPR_HAS_LOCATION (val))
614 SET_EXPR_LOCATION (mod, EXPR_LOCATION (val));
615 else
616 SET_EXPR_LOCATION (mod, input_location);
618 /* gimplify_modify_expr might want to reduce this further. */
619 gimplify_and_add (mod, pre_p);
620 ggc_free (mod);
622 /* If we're gimplifying into ssa, gimplify_modify_expr will have
623 given our temporary an SSA name. Find and return it. */
624 if (gimplify_ctxp->into_ssa)
626 gimple last = gimple_seq_last_stmt (*pre_p);
627 t = gimple_get_lhs (last);
630 return t;
633 /* Returns a formal temporary variable initialized with VAL. PRE_P is as
634 in gimplify_expr. Only use this function if:
636 1) The value of the unfactored expression represented by VAL will not
637 change between the initialization and use of the temporary, and
638 2) The temporary will not be otherwise modified.
640 For instance, #1 means that this is inappropriate for SAVE_EXPR temps,
641 and #2 means it is inappropriate for && temps.
643 For other cases, use get_initialized_tmp_var instead. */
645 tree
646 get_formal_tmp_var (tree val, gimple_seq *pre_p)
648 return internal_get_tmp_var (val, pre_p, NULL, true);
651 /* Returns a temporary variable initialized with VAL. PRE_P and POST_P
652 are as in gimplify_expr. */
654 tree
655 get_initialized_tmp_var (tree val, gimple_seq *pre_p, gimple_seq *post_p)
657 return internal_get_tmp_var (val, pre_p, post_p, false);
660 /* Declares all the variables in VARS in SCOPE. If DEBUG_INFO is
661 true, generate debug info for them; otherwise don't. */
663 void
664 declare_vars (tree vars, gimple scope, bool debug_info)
666 tree last = vars;
667 if (last)
669 tree temps, block;
671 gcc_assert (gimple_code (scope) == GIMPLE_BIND);
673 temps = nreverse (last);
675 block = gimple_bind_block (scope);
676 gcc_assert (!block || TREE_CODE (block) == BLOCK);
677 if (!block || !debug_info)
679 TREE_CHAIN (last) = gimple_bind_vars (scope);
680 gimple_bind_set_vars (scope, temps);
682 else
684 /* We need to attach the nodes both to the BIND_EXPR and to its
685 associated BLOCK for debugging purposes. The key point here
686 is that the BLOCK_VARS of the BIND_EXPR_BLOCK of a BIND_EXPR
687 is a subchain of the BIND_EXPR_VARS of the BIND_EXPR. */
688 if (BLOCK_VARS (block))
689 BLOCK_VARS (block) = chainon (BLOCK_VARS (block), temps);
690 else
692 gimple_bind_set_vars (scope,
693 chainon (gimple_bind_vars (scope), temps));
694 BLOCK_VARS (block) = temps;
700 /* For VAR a VAR_DECL of variable size, try to find a constant upper bound
701 for the size and adjust DECL_SIZE/DECL_SIZE_UNIT accordingly. Abort if
702 no such upper bound can be obtained. */
704 static void
705 force_constant_size (tree var)
707 /* The only attempt we make is by querying the maximum size of objects
708 of the variable's type. */
710 HOST_WIDE_INT max_size;
712 gcc_assert (TREE_CODE (var) == VAR_DECL);
714 max_size = max_int_size_in_bytes (TREE_TYPE (var));
716 gcc_assert (max_size >= 0);
718 DECL_SIZE_UNIT (var)
719 = build_int_cst (TREE_TYPE (DECL_SIZE_UNIT (var)), max_size);
720 DECL_SIZE (var)
721 = build_int_cst (TREE_TYPE (DECL_SIZE (var)), max_size * BITS_PER_UNIT);
724 void
725 gimple_add_tmp_var (tree tmp)
727 gcc_assert (!TREE_CHAIN (tmp) && !DECL_SEEN_IN_BIND_EXPR_P (tmp));
729 /* Later processing assumes that the object size is constant, which might
730 not be true at this point. Force the use of a constant upper bound in
731 this case. */
732 if (!host_integerp (DECL_SIZE_UNIT (tmp), 1))
733 force_constant_size (tmp);
735 DECL_CONTEXT (tmp) = current_function_decl;
736 DECL_SEEN_IN_BIND_EXPR_P (tmp) = 1;
738 if (gimplify_ctxp)
740 TREE_CHAIN (tmp) = gimplify_ctxp->temps;
741 gimplify_ctxp->temps = tmp;
743 /* Mark temporaries local within the nearest enclosing parallel. */
744 if (gimplify_omp_ctxp)
746 struct gimplify_omp_ctx *ctx = gimplify_omp_ctxp;
747 while (ctx && ctx->region_type == ORT_WORKSHARE)
748 ctx = ctx->outer_context;
749 if (ctx)
750 omp_add_variable (ctx, tmp, GOVD_LOCAL | GOVD_SEEN);
753 else if (cfun)
754 record_vars (tmp);
755 else
757 gimple_seq body_seq;
759 /* This case is for nested functions. We need to expose the locals
760 they create. */
761 body_seq = gimple_body (current_function_decl);
762 declare_vars (tmp, gimple_seq_first_stmt (body_seq), false);
766 /* Determines whether to assign a location to the statement GS. */
768 static bool
769 should_carry_location_p (gimple gs)
771 /* Don't emit a line note for a label. We particularly don't want to
772 emit one for the break label, since it doesn't actually correspond
773 to the beginning of the loop/switch. */
774 if (gimple_code (gs) == GIMPLE_LABEL)
775 return false;
777 return true;
781 /* Return true if a location should not be emitted for this statement
782 by annotate_one_with_location. */
784 static inline bool
785 gimple_do_not_emit_location_p (gimple g)
787 return gimple_plf (g, GF_PLF_1);
790 /* Mark statement G so a location will not be emitted by
791 annotate_one_with_location. */
793 static inline void
794 gimple_set_do_not_emit_location (gimple g)
796 /* The PLF flags are initialized to 0 when a new tuple is created,
797 so no need to initialize it anywhere. */
798 gimple_set_plf (g, GF_PLF_1, true);
801 /* Set the location for gimple statement GS to LOCATION. */
803 static void
804 annotate_one_with_location (gimple gs, location_t location)
806 if (!gimple_has_location (gs)
807 && !gimple_do_not_emit_location_p (gs)
808 && should_carry_location_p (gs))
809 gimple_set_location (gs, location);
813 /* Set LOCATION for all the statements after iterator GSI in sequence
814 SEQ. If GSI is pointing to the end of the sequence, start with the
815 first statement in SEQ. */
817 static void
818 annotate_all_with_location_after (gimple_seq seq, gimple_stmt_iterator gsi,
819 location_t location)
821 if (gsi_end_p (gsi))
822 gsi = gsi_start (seq);
823 else
824 gsi_next (&gsi);
826 for (; !gsi_end_p (gsi); gsi_next (&gsi))
827 annotate_one_with_location (gsi_stmt (gsi), location);
831 /* Set the location for all the statements in a sequence STMT_P to LOCATION. */
833 void
834 annotate_all_with_location (gimple_seq stmt_p, location_t location)
836 gimple_stmt_iterator i;
838 if (gimple_seq_empty_p (stmt_p))
839 return;
841 for (i = gsi_start (stmt_p); !gsi_end_p (i); gsi_next (&i))
843 gimple gs = gsi_stmt (i);
844 annotate_one_with_location (gs, location);
849 /* Similar to copy_tree_r() but do not copy SAVE_EXPR or TARGET_EXPR nodes.
850 These nodes model computations that should only be done once. If we
851 were to unshare something like SAVE_EXPR(i++), the gimplification
852 process would create wrong code. */
854 static tree
855 mostly_copy_tree_r (tree *tp, int *walk_subtrees, void *data)
857 enum tree_code code = TREE_CODE (*tp);
858 /* Don't unshare types, decls, constants and SAVE_EXPR nodes. */
859 if (TREE_CODE_CLASS (code) == tcc_type
860 || TREE_CODE_CLASS (code) == tcc_declaration
861 || TREE_CODE_CLASS (code) == tcc_constant
862 || code == SAVE_EXPR || code == TARGET_EXPR
863 /* We can't do anything sensible with a BLOCK used as an expression,
864 but we also can't just die when we see it because of non-expression
865 uses. So just avert our eyes and cross our fingers. Silly Java. */
866 || code == BLOCK)
867 *walk_subtrees = 0;
868 else
870 gcc_assert (code != BIND_EXPR);
871 copy_tree_r (tp, walk_subtrees, data);
874 return NULL_TREE;
877 /* Callback for walk_tree to unshare most of the shared trees rooted at
878 *TP. If *TP has been visited already (i.e., TREE_VISITED (*TP) == 1),
879 then *TP is deep copied by calling copy_tree_r.
881 This unshares the same trees as copy_tree_r with the exception of
882 SAVE_EXPR nodes. These nodes model computations that should only be
883 done once. If we were to unshare something like SAVE_EXPR(i++), the
884 gimplification process would create wrong code. */
886 static tree
887 copy_if_shared_r (tree *tp, int *walk_subtrees ATTRIBUTE_UNUSED,
888 void *data ATTRIBUTE_UNUSED)
890 tree t = *tp;
891 enum tree_code code = TREE_CODE (t);
893 /* Skip types, decls, and constants. But we do want to look at their
894 types and the bounds of types. Mark them as visited so we properly
895 unmark their subtrees on the unmark pass. If we've already seen them,
896 don't look down further. */
897 if (TREE_CODE_CLASS (code) == tcc_type
898 || TREE_CODE_CLASS (code) == tcc_declaration
899 || TREE_CODE_CLASS (code) == tcc_constant)
901 if (TREE_VISITED (t))
902 *walk_subtrees = 0;
903 else
904 TREE_VISITED (t) = 1;
907 /* If this node has been visited already, unshare it and don't look
908 any deeper. */
909 else if (TREE_VISITED (t))
911 walk_tree (tp, mostly_copy_tree_r, NULL, NULL);
912 *walk_subtrees = 0;
915 /* Otherwise, mark the tree as visited and keep looking. */
916 else
917 TREE_VISITED (t) = 1;
919 return NULL_TREE;
922 static tree
923 unmark_visited_r (tree *tp, int *walk_subtrees ATTRIBUTE_UNUSED,
924 void *data ATTRIBUTE_UNUSED)
926 if (TREE_VISITED (*tp))
927 TREE_VISITED (*tp) = 0;
928 else
929 *walk_subtrees = 0;
931 return NULL_TREE;
934 /* Unshare all the trees in BODY_P, a pointer into the body of FNDECL, and the
935 bodies of any nested functions if we are unsharing the entire body of
936 FNDECL. */
938 static void
939 unshare_body (tree *body_p, tree fndecl)
941 struct cgraph_node *cgn = cgraph_node (fndecl);
943 walk_tree (body_p, copy_if_shared_r, NULL, NULL);
944 if (body_p == &DECL_SAVED_TREE (fndecl))
945 for (cgn = cgn->nested; cgn; cgn = cgn->next_nested)
946 unshare_body (&DECL_SAVED_TREE (cgn->decl), cgn->decl);
949 /* Likewise, but mark all trees as not visited. */
951 static void
952 unvisit_body (tree *body_p, tree fndecl)
954 struct cgraph_node *cgn = cgraph_node (fndecl);
956 walk_tree (body_p, unmark_visited_r, NULL, NULL);
957 if (body_p == &DECL_SAVED_TREE (fndecl))
958 for (cgn = cgn->nested; cgn; cgn = cgn->next_nested)
959 unvisit_body (&DECL_SAVED_TREE (cgn->decl), cgn->decl);
962 /* Unconditionally make an unshared copy of EXPR. This is used when using
963 stored expressions which span multiple functions, such as BINFO_VTABLE,
964 as the normal unsharing process can't tell that they're shared. */
966 tree
967 unshare_expr (tree expr)
969 walk_tree (&expr, mostly_copy_tree_r, NULL, NULL);
970 return expr;
973 /* WRAPPER is a code such as BIND_EXPR or CLEANUP_POINT_EXPR which can both
974 contain statements and have a value. Assign its value to a temporary
975 and give it void_type_node. Returns the temporary, or NULL_TREE if
976 WRAPPER was already void. */
978 tree
979 voidify_wrapper_expr (tree wrapper, tree temp)
981 tree type = TREE_TYPE (wrapper);
982 if (type && !VOID_TYPE_P (type))
984 tree *p;
986 /* Set p to point to the body of the wrapper. Loop until we find
987 something that isn't a wrapper. */
988 for (p = &wrapper; p && *p; )
990 switch (TREE_CODE (*p))
992 case BIND_EXPR:
993 TREE_SIDE_EFFECTS (*p) = 1;
994 TREE_TYPE (*p) = void_type_node;
995 /* For a BIND_EXPR, the body is operand 1. */
996 p = &BIND_EXPR_BODY (*p);
997 break;
999 case CLEANUP_POINT_EXPR:
1000 case TRY_FINALLY_EXPR:
1001 case TRY_CATCH_EXPR:
1002 TREE_SIDE_EFFECTS (*p) = 1;
1003 TREE_TYPE (*p) = void_type_node;
1004 p = &TREE_OPERAND (*p, 0);
1005 break;
1007 case STATEMENT_LIST:
1009 tree_stmt_iterator i = tsi_last (*p);
1010 TREE_SIDE_EFFECTS (*p) = 1;
1011 TREE_TYPE (*p) = void_type_node;
1012 p = tsi_end_p (i) ? NULL : tsi_stmt_ptr (i);
1014 break;
1016 case COMPOUND_EXPR:
1017 /* Advance to the last statement. Set all container types to void. */
1018 for (; TREE_CODE (*p) == COMPOUND_EXPR; p = &TREE_OPERAND (*p, 1))
1020 TREE_SIDE_EFFECTS (*p) = 1;
1021 TREE_TYPE (*p) = void_type_node;
1023 break;
1025 default:
1026 goto out;
1030 out:
1031 if (p == NULL || IS_EMPTY_STMT (*p))
1032 temp = NULL_TREE;
1033 else if (temp)
1035 /* The wrapper is on the RHS of an assignment that we're pushing
1036 down. */
1037 gcc_assert (TREE_CODE (temp) == INIT_EXPR
1038 || TREE_CODE (temp) == MODIFY_EXPR);
1039 TREE_OPERAND (temp, 1) = *p;
1040 *p = temp;
1042 else
1044 temp = create_tmp_var (type, "retval");
1045 *p = build2 (INIT_EXPR, type, temp, *p);
1048 return temp;
1051 return NULL_TREE;
1054 /* Prepare calls to builtins to SAVE and RESTORE the stack as well as
1055 a temporary through which they communicate. */
1057 static void
1058 build_stack_save_restore (gimple *save, gimple *restore)
1060 tree tmp_var;
1062 *save = gimple_build_call (implicit_built_in_decls[BUILT_IN_STACK_SAVE], 0);
1063 tmp_var = create_tmp_var (ptr_type_node, "saved_stack");
1064 gimple_call_set_lhs (*save, tmp_var);
1066 *restore = gimple_build_call (implicit_built_in_decls[BUILT_IN_STACK_RESTORE],
1067 1, tmp_var);
1070 /* Gimplify a BIND_EXPR. Just voidify and recurse. */
1072 static enum gimplify_status
1073 gimplify_bind_expr (tree *expr_p, gimple_seq *pre_p)
1075 tree bind_expr = *expr_p;
1076 bool old_save_stack = gimplify_ctxp->save_stack;
1077 tree t;
1078 gimple gimple_bind;
1079 gimple_seq body;
1081 tree temp = voidify_wrapper_expr (bind_expr, NULL);
1083 /* Mark variables seen in this bind expr. */
1084 for (t = BIND_EXPR_VARS (bind_expr); t ; t = TREE_CHAIN (t))
1086 if (TREE_CODE (t) == VAR_DECL)
1088 struct gimplify_omp_ctx *ctx = gimplify_omp_ctxp;
1090 /* Mark variable as local. */
1091 if (ctx && !is_global_var (t)
1092 && (! DECL_SEEN_IN_BIND_EXPR_P (t)
1093 || splay_tree_lookup (ctx->variables,
1094 (splay_tree_key) t) == NULL))
1095 omp_add_variable (gimplify_omp_ctxp, t, GOVD_LOCAL | GOVD_SEEN);
1097 DECL_SEEN_IN_BIND_EXPR_P (t) = 1;
1099 if (DECL_HARD_REGISTER (t) && !is_global_var (t) && cfun)
1100 cfun->has_local_explicit_reg_vars = true;
1103 /* Preliminarily mark non-addressed complex variables as eligible
1104 for promotion to gimple registers. We'll transform their uses
1105 as we find them.
1106 We exclude complex types if not optimizing because they can be
1107 subject to partial stores in GNU C by means of the __real__ and
1108 __imag__ operators and we cannot promote them to total stores
1109 (see gimplify_modify_expr_complex_part). */
1110 if (optimize
1111 && (TREE_CODE (TREE_TYPE (t)) == COMPLEX_TYPE
1112 || TREE_CODE (TREE_TYPE (t)) == VECTOR_TYPE)
1113 && !TREE_THIS_VOLATILE (t)
1114 && (TREE_CODE (t) == VAR_DECL && !DECL_HARD_REGISTER (t))
1115 && !needs_to_live_in_memory (t))
1116 DECL_GIMPLE_REG_P (t) = 1;
1119 gimple_bind = gimple_build_bind (BIND_EXPR_VARS (bind_expr), NULL,
1120 BIND_EXPR_BLOCK (bind_expr));
1121 gimple_push_bind_expr (gimple_bind);
1123 gimplify_ctxp->save_stack = false;
1125 /* Gimplify the body into the GIMPLE_BIND tuple's body. */
1126 body = NULL;
1127 gimplify_stmt (&BIND_EXPR_BODY (bind_expr), &body);
1128 gimple_bind_set_body (gimple_bind, body);
1130 if (gimplify_ctxp->save_stack)
1132 gimple stack_save, stack_restore, gs;
1133 gimple_seq cleanup, new_body;
1135 /* Save stack on entry and restore it on exit. Add a try_finally
1136 block to achieve this. Note that mudflap depends on the
1137 format of the emitted code: see mx_register_decls(). */
1138 build_stack_save_restore (&stack_save, &stack_restore);
1140 cleanup = new_body = NULL;
1141 gimplify_seq_add_stmt (&cleanup, stack_restore);
1142 gs = gimple_build_try (gimple_bind_body (gimple_bind), cleanup,
1143 GIMPLE_TRY_FINALLY);
1145 gimplify_seq_add_stmt (&new_body, stack_save);
1146 gimplify_seq_add_stmt (&new_body, gs);
1147 gimple_bind_set_body (gimple_bind, new_body);
1150 gimplify_ctxp->save_stack = old_save_stack;
1151 gimple_pop_bind_expr ();
1153 gimplify_seq_add_stmt (pre_p, gimple_bind);
1155 if (temp)
1157 *expr_p = temp;
1158 return GS_OK;
1161 *expr_p = NULL_TREE;
1162 return GS_ALL_DONE;
1165 /* Gimplify a RETURN_EXPR. If the expression to be returned is not a
1166 GIMPLE value, it is assigned to a new temporary and the statement is
1167 re-written to return the temporary.
1169 PRE_P points to the sequence where side effects that must happen before
1170 STMT should be stored. */
1172 static enum gimplify_status
1173 gimplify_return_expr (tree stmt, gimple_seq *pre_p)
1175 gimple ret;
1176 tree ret_expr = TREE_OPERAND (stmt, 0);
1177 tree result_decl, result;
1179 if (ret_expr == error_mark_node)
1180 return GS_ERROR;
1182 if (!ret_expr
1183 || TREE_CODE (ret_expr) == RESULT_DECL
1184 || ret_expr == error_mark_node)
1186 gimple ret = gimple_build_return (ret_expr);
1187 gimple_set_no_warning (ret, TREE_NO_WARNING (stmt));
1188 gimplify_seq_add_stmt (pre_p, ret);
1189 return GS_ALL_DONE;
1192 if (VOID_TYPE_P (function_return_type (current_function_decl)))
1193 result_decl = NULL_TREE;
1194 else
1196 result_decl = TREE_OPERAND (ret_expr, 0);
1198 /* See through a return by reference. */
1199 if (TREE_CODE (result_decl) == INDIRECT_REF)
1200 result_decl = TREE_OPERAND (result_decl, 0);
1202 gcc_assert ((TREE_CODE (ret_expr) == MODIFY_EXPR
1203 || TREE_CODE (ret_expr) == INIT_EXPR)
1204 && TREE_CODE (result_decl) == RESULT_DECL);
1207 /* If aggregate_value_p is true, then we can return the bare RESULT_DECL.
1208 Recall that aggregate_value_p is FALSE for any aggregate type that is
1209 returned in registers. If we're returning values in registers, then
1210 we don't want to extend the lifetime of the RESULT_DECL, particularly
1211 across another call. In addition, for those aggregates for which
1212 hard_function_value generates a PARALLEL, we'll die during normal
1213 expansion of structure assignments; there's special code in expand_return
1214 to handle this case that does not exist in expand_expr. */
1215 if (!result_decl
1216 || aggregate_value_p (result_decl, TREE_TYPE (current_function_decl)))
1217 result = result_decl;
1218 else if (gimplify_ctxp->return_temp)
1219 result = gimplify_ctxp->return_temp;
1220 else
1222 result = create_tmp_var (TREE_TYPE (result_decl), NULL);
1223 if (TREE_CODE (TREE_TYPE (result)) == COMPLEX_TYPE
1224 || TREE_CODE (TREE_TYPE (result)) == VECTOR_TYPE)
1225 DECL_GIMPLE_REG_P (result) = 1;
1227 /* ??? With complex control flow (usually involving abnormal edges),
1228 we can wind up warning about an uninitialized value for this. Due
1229 to how this variable is constructed and initialized, this is never
1230 true. Give up and never warn. */
1231 TREE_NO_WARNING (result) = 1;
1233 gimplify_ctxp->return_temp = result;
1236 /* Smash the lhs of the MODIFY_EXPR to the temporary we plan to use.
1237 Then gimplify the whole thing. */
1238 if (result != result_decl)
1239 TREE_OPERAND (ret_expr, 0) = result;
1241 gimplify_and_add (TREE_OPERAND (stmt, 0), pre_p);
1243 ret = gimple_build_return (result);
1244 gimple_set_no_warning (ret, TREE_NO_WARNING (stmt));
1245 gimplify_seq_add_stmt (pre_p, ret);
1247 return GS_ALL_DONE;
1250 static void
1251 gimplify_vla_decl (tree decl, gimple_seq *seq_p)
1253 /* This is a variable-sized decl. Simplify its size and mark it
1254 for deferred expansion. Note that mudflap depends on the format
1255 of the emitted code: see mx_register_decls(). */
1256 tree t, addr, ptr_type;
1258 gimplify_one_sizepos (&DECL_SIZE (decl), seq_p);
1259 gimplify_one_sizepos (&DECL_SIZE_UNIT (decl), seq_p);
1261 /* All occurrences of this decl in final gimplified code will be
1262 replaced by indirection. Setting DECL_VALUE_EXPR does two
1263 things: First, it lets the rest of the gimplifier know what
1264 replacement to use. Second, it lets the debug info know
1265 where to find the value. */
1266 ptr_type = build_pointer_type (TREE_TYPE (decl));
1267 addr = create_tmp_var (ptr_type, get_name (decl));
1268 DECL_IGNORED_P (addr) = 0;
1269 t = build_fold_indirect_ref (addr);
1270 SET_DECL_VALUE_EXPR (decl, t);
1271 DECL_HAS_VALUE_EXPR_P (decl) = 1;
1273 t = built_in_decls[BUILT_IN_ALLOCA];
1274 t = build_call_expr (t, 1, DECL_SIZE_UNIT (decl));
1275 t = fold_convert (ptr_type, t);
1276 t = build2 (MODIFY_EXPR, TREE_TYPE (addr), addr, t);
1278 gimplify_and_add (t, seq_p);
1280 /* Indicate that we need to restore the stack level when the
1281 enclosing BIND_EXPR is exited. */
1282 gimplify_ctxp->save_stack = true;
1286 /* Gimplifies a DECL_EXPR node *STMT_P by making any necessary allocation
1287 and initialization explicit. */
1289 static enum gimplify_status
1290 gimplify_decl_expr (tree *stmt_p, gimple_seq *seq_p)
1292 tree stmt = *stmt_p;
1293 tree decl = DECL_EXPR_DECL (stmt);
1295 *stmt_p = NULL_TREE;
1297 if (TREE_TYPE (decl) == error_mark_node)
1298 return GS_ERROR;
1300 if ((TREE_CODE (decl) == TYPE_DECL
1301 || TREE_CODE (decl) == VAR_DECL)
1302 && !TYPE_SIZES_GIMPLIFIED (TREE_TYPE (decl)))
1303 gimplify_type_sizes (TREE_TYPE (decl), seq_p);
1305 if (TREE_CODE (decl) == VAR_DECL && !DECL_EXTERNAL (decl))
1307 tree init = DECL_INITIAL (decl);
1309 if (TREE_CODE (DECL_SIZE_UNIT (decl)) != INTEGER_CST
1310 || (!TREE_STATIC (decl)
1311 && flag_stack_check == GENERIC_STACK_CHECK
1312 && compare_tree_int (DECL_SIZE_UNIT (decl),
1313 STACK_CHECK_MAX_VAR_SIZE) > 0))
1314 gimplify_vla_decl (decl, seq_p);
1316 if (init && init != error_mark_node)
1318 if (!TREE_STATIC (decl))
1320 DECL_INITIAL (decl) = NULL_TREE;
1321 init = build2 (INIT_EXPR, void_type_node, decl, init);
1322 gimplify_and_add (init, seq_p);
1323 ggc_free (init);
1325 else
1326 /* We must still examine initializers for static variables
1327 as they may contain a label address. */
1328 walk_tree (&init, force_labels_r, NULL, NULL);
1331 /* Some front ends do not explicitly declare all anonymous
1332 artificial variables. We compensate here by declaring the
1333 variables, though it would be better if the front ends would
1334 explicitly declare them. */
1335 if (!DECL_SEEN_IN_BIND_EXPR_P (decl)
1336 && DECL_ARTIFICIAL (decl) && DECL_NAME (decl) == NULL_TREE)
1337 gimple_add_tmp_var (decl);
1340 return GS_ALL_DONE;
1343 /* Gimplify a LOOP_EXPR. Normally this just involves gimplifying the body
1344 and replacing the LOOP_EXPR with goto, but if the loop contains an
1345 EXIT_EXPR, we need to append a label for it to jump to. */
1347 static enum gimplify_status
1348 gimplify_loop_expr (tree *expr_p, gimple_seq *pre_p)
1350 tree saved_label = gimplify_ctxp->exit_label;
1351 tree start_label = create_artificial_label (UNKNOWN_LOCATION);
1353 gimplify_seq_add_stmt (pre_p, gimple_build_label (start_label));
1355 gimplify_ctxp->exit_label = NULL_TREE;
1357 gimplify_and_add (LOOP_EXPR_BODY (*expr_p), pre_p);
1359 gimplify_seq_add_stmt (pre_p, gimple_build_goto (start_label));
1361 if (gimplify_ctxp->exit_label)
1362 gimplify_seq_add_stmt (pre_p, gimple_build_label (gimplify_ctxp->exit_label));
1364 gimplify_ctxp->exit_label = saved_label;
1366 *expr_p = NULL;
1367 return GS_ALL_DONE;
1370 /* Gimplifies a statement list onto a sequence. These may be created either
1371 by an enlightened front-end, or by shortcut_cond_expr. */
1373 static enum gimplify_status
1374 gimplify_statement_list (tree *expr_p, gimple_seq *pre_p)
1376 tree temp = voidify_wrapper_expr (*expr_p, NULL);
1378 tree_stmt_iterator i = tsi_start (*expr_p);
1380 while (!tsi_end_p (i))
1382 gimplify_stmt (tsi_stmt_ptr (i), pre_p);
1383 tsi_delink (&i);
1386 if (temp)
1388 *expr_p = temp;
1389 return GS_OK;
1392 return GS_ALL_DONE;
1395 /* Compare two case labels. Because the front end should already have
1396 made sure that case ranges do not overlap, it is enough to only compare
1397 the CASE_LOW values of each case label. */
1399 static int
1400 compare_case_labels (const void *p1, const void *p2)
1402 const_tree const case1 = *(const_tree const*)p1;
1403 const_tree const case2 = *(const_tree const*)p2;
1405 /* The 'default' case label always goes first. */
1406 if (!CASE_LOW (case1))
1407 return -1;
1408 else if (!CASE_LOW (case2))
1409 return 1;
1410 else
1411 return tree_int_cst_compare (CASE_LOW (case1), CASE_LOW (case2));
1415 /* Sort the case labels in LABEL_VEC in place in ascending order. */
1417 void
1418 sort_case_labels (VEC(tree,heap)* label_vec)
1420 size_t len = VEC_length (tree, label_vec);
1421 qsort (VEC_address (tree, label_vec), len, sizeof (tree),
1422 compare_case_labels);
1426 /* Gimplify a SWITCH_EXPR, and collect a TREE_VEC of the labels it can
1427 branch to. */
1429 static enum gimplify_status
1430 gimplify_switch_expr (tree *expr_p, gimple_seq *pre_p)
1432 tree switch_expr = *expr_p;
1433 gimple_seq switch_body_seq = NULL;
1434 enum gimplify_status ret;
1436 ret = gimplify_expr (&SWITCH_COND (switch_expr), pre_p, NULL, is_gimple_val,
1437 fb_rvalue);
1438 if (ret == GS_ERROR || ret == GS_UNHANDLED)
1439 return ret;
1441 if (SWITCH_BODY (switch_expr))
1443 VEC (tree,heap) *labels;
1444 VEC (tree,heap) *saved_labels;
1445 tree default_case = NULL_TREE;
1446 size_t i, len;
1447 gimple gimple_switch;
1449 /* If someone can be bothered to fill in the labels, they can
1450 be bothered to null out the body too. */
1451 gcc_assert (!SWITCH_LABELS (switch_expr));
1453 /* save old labels, get new ones from body, then restore the old
1454 labels. Save all the things from the switch body to append after. */
1455 saved_labels = gimplify_ctxp->case_labels;
1456 gimplify_ctxp->case_labels = VEC_alloc (tree, heap, 8);
1458 gimplify_stmt (&SWITCH_BODY (switch_expr), &switch_body_seq);
1459 labels = gimplify_ctxp->case_labels;
1460 gimplify_ctxp->case_labels = saved_labels;
1462 i = 0;
1463 while (i < VEC_length (tree, labels))
1465 tree elt = VEC_index (tree, labels, i);
1466 tree low = CASE_LOW (elt);
1467 bool remove_element = FALSE;
1469 if (low)
1471 /* Discard empty ranges. */
1472 tree high = CASE_HIGH (elt);
1473 if (high && tree_int_cst_lt (high, low))
1474 remove_element = TRUE;
1476 else
1478 /* The default case must be the last label in the list. */
1479 gcc_assert (!default_case);
1480 default_case = elt;
1481 remove_element = TRUE;
1484 if (remove_element)
1485 VEC_ordered_remove (tree, labels, i);
1486 else
1487 i++;
1489 len = i;
1491 if (!VEC_empty (tree, labels))
1492 sort_case_labels (labels);
1494 if (!default_case)
1496 tree type = TREE_TYPE (switch_expr);
1498 /* If the switch has no default label, add one, so that we jump
1499 around the switch body. If the labels already cover the whole
1500 range of type, add the default label pointing to one of the
1501 existing labels. */
1502 if (type == void_type_node)
1503 type = TREE_TYPE (SWITCH_COND (switch_expr));
1504 if (len
1505 && INTEGRAL_TYPE_P (type)
1506 && TYPE_MIN_VALUE (type)
1507 && TYPE_MAX_VALUE (type)
1508 && tree_int_cst_equal (CASE_LOW (VEC_index (tree, labels, 0)),
1509 TYPE_MIN_VALUE (type)))
1511 tree low, high = CASE_HIGH (VEC_index (tree, labels, len - 1));
1512 if (!high)
1513 high = CASE_LOW (VEC_index (tree, labels, len - 1));
1514 if (tree_int_cst_equal (high, TYPE_MAX_VALUE (type)))
1516 for (i = 1; i < len; i++)
1518 high = CASE_LOW (VEC_index (tree, labels, i));
1519 low = CASE_HIGH (VEC_index (tree, labels, i - 1));
1520 if (!low)
1521 low = CASE_LOW (VEC_index (tree, labels, i - 1));
1522 if ((TREE_INT_CST_LOW (low) + 1
1523 != TREE_INT_CST_LOW (high))
1524 || (TREE_INT_CST_HIGH (low)
1525 + (TREE_INT_CST_LOW (high) == 0)
1526 != TREE_INT_CST_HIGH (high)))
1527 break;
1529 if (i == len)
1530 default_case = build3 (CASE_LABEL_EXPR, void_type_node,
1531 NULL_TREE, NULL_TREE,
1532 CASE_LABEL (VEC_index (tree,
1533 labels, 0)));
1537 if (!default_case)
1539 gimple new_default;
1541 default_case
1542 = build3 (CASE_LABEL_EXPR, void_type_node,
1543 NULL_TREE, NULL_TREE,
1544 create_artificial_label (UNKNOWN_LOCATION));
1545 new_default = gimple_build_label (CASE_LABEL (default_case));
1546 gimplify_seq_add_stmt (&switch_body_seq, new_default);
1550 gimple_switch = gimple_build_switch_vec (SWITCH_COND (switch_expr),
1551 default_case, labels);
1552 gimplify_seq_add_stmt (pre_p, gimple_switch);
1553 gimplify_seq_add_seq (pre_p, switch_body_seq);
1554 VEC_free(tree, heap, labels);
1556 else
1557 gcc_assert (SWITCH_LABELS (switch_expr));
1559 return GS_ALL_DONE;
1563 static enum gimplify_status
1564 gimplify_case_label_expr (tree *expr_p, gimple_seq *pre_p)
1566 struct gimplify_ctx *ctxp;
1567 gimple gimple_label;
1569 /* Invalid OpenMP programs can play Duff's Device type games with
1570 #pragma omp parallel. At least in the C front end, we don't
1571 detect such invalid branches until after gimplification. */
1572 for (ctxp = gimplify_ctxp; ; ctxp = ctxp->prev_context)
1573 if (ctxp->case_labels)
1574 break;
1576 gimple_label = gimple_build_label (CASE_LABEL (*expr_p));
1577 VEC_safe_push (tree, heap, ctxp->case_labels, *expr_p);
1578 gimplify_seq_add_stmt (pre_p, gimple_label);
1580 return GS_ALL_DONE;
1583 /* Build a GOTO to the LABEL_DECL pointed to by LABEL_P, building it first
1584 if necessary. */
1586 tree
1587 build_and_jump (tree *label_p)
1589 if (label_p == NULL)
1590 /* If there's nowhere to jump, just fall through. */
1591 return NULL_TREE;
1593 if (*label_p == NULL_TREE)
1595 tree label = create_artificial_label (UNKNOWN_LOCATION);
1596 *label_p = label;
1599 return build1 (GOTO_EXPR, void_type_node, *label_p);
1602 /* Gimplify an EXIT_EXPR by converting to a GOTO_EXPR inside a COND_EXPR.
1603 This also involves building a label to jump to and communicating it to
1604 gimplify_loop_expr through gimplify_ctxp->exit_label. */
1606 static enum gimplify_status
1607 gimplify_exit_expr (tree *expr_p)
1609 tree cond = TREE_OPERAND (*expr_p, 0);
1610 tree expr;
1612 expr = build_and_jump (&gimplify_ctxp->exit_label);
1613 expr = build3 (COND_EXPR, void_type_node, cond, expr, NULL_TREE);
1614 *expr_p = expr;
1616 return GS_OK;
1619 /* A helper function to be called via walk_tree. Mark all labels under *TP
1620 as being forced. To be called for DECL_INITIAL of static variables. */
1622 tree
1623 force_labels_r (tree *tp, int *walk_subtrees, void *data ATTRIBUTE_UNUSED)
1625 if (TYPE_P (*tp))
1626 *walk_subtrees = 0;
1627 if (TREE_CODE (*tp) == LABEL_DECL)
1628 FORCED_LABEL (*tp) = 1;
1630 return NULL_TREE;
1633 /* *EXPR_P is a COMPONENT_REF being used as an rvalue. If its type is
1634 different from its canonical type, wrap the whole thing inside a
1635 NOP_EXPR and force the type of the COMPONENT_REF to be the canonical
1636 type.
1638 The canonical type of a COMPONENT_REF is the type of the field being
1639 referenced--unless the field is a bit-field which can be read directly
1640 in a smaller mode, in which case the canonical type is the
1641 sign-appropriate type corresponding to that mode. */
1643 static void
1644 canonicalize_component_ref (tree *expr_p)
1646 tree expr = *expr_p;
1647 tree type;
1649 gcc_assert (TREE_CODE (expr) == COMPONENT_REF);
1651 if (INTEGRAL_TYPE_P (TREE_TYPE (expr)))
1652 type = TREE_TYPE (get_unwidened (expr, NULL_TREE));
1653 else
1654 type = TREE_TYPE (TREE_OPERAND (expr, 1));
1656 /* One could argue that all the stuff below is not necessary for
1657 the non-bitfield case and declare it a FE error if type
1658 adjustment would be needed. */
1659 if (TREE_TYPE (expr) != type)
1661 #ifdef ENABLE_TYPES_CHECKING
1662 tree old_type = TREE_TYPE (expr);
1663 #endif
1664 int type_quals;
1666 /* We need to preserve qualifiers and propagate them from
1667 operand 0. */
1668 type_quals = TYPE_QUALS (type)
1669 | TYPE_QUALS (TREE_TYPE (TREE_OPERAND (expr, 0)));
1670 if (TYPE_QUALS (type) != type_quals)
1671 type = build_qualified_type (TYPE_MAIN_VARIANT (type), type_quals);
1673 /* Set the type of the COMPONENT_REF to the underlying type. */
1674 TREE_TYPE (expr) = type;
1676 #ifdef ENABLE_TYPES_CHECKING
1677 /* It is now a FE error, if the conversion from the canonical
1678 type to the original expression type is not useless. */
1679 gcc_assert (useless_type_conversion_p (old_type, type));
1680 #endif
1684 /* If a NOP conversion is changing a pointer to array of foo to a pointer
1685 to foo, embed that change in the ADDR_EXPR by converting
1686 T array[U];
1687 (T *)&array
1689 &array[L]
1690 where L is the lower bound. For simplicity, only do this for constant
1691 lower bound.
1692 The constraint is that the type of &array[L] is trivially convertible
1693 to T *. */
1695 static void
1696 canonicalize_addr_expr (tree *expr_p)
1698 tree expr = *expr_p;
1699 tree addr_expr = TREE_OPERAND (expr, 0);
1700 tree datype, ddatype, pddatype;
1702 /* We simplify only conversions from an ADDR_EXPR to a pointer type. */
1703 if (!POINTER_TYPE_P (TREE_TYPE (expr))
1704 || TREE_CODE (addr_expr) != ADDR_EXPR)
1705 return;
1707 /* The addr_expr type should be a pointer to an array. */
1708 datype = TREE_TYPE (TREE_TYPE (addr_expr));
1709 if (TREE_CODE (datype) != ARRAY_TYPE)
1710 return;
1712 /* The pointer to element type shall be trivially convertible to
1713 the expression pointer type. */
1714 ddatype = TREE_TYPE (datype);
1715 pddatype = build_pointer_type (ddatype);
1716 if (!useless_type_conversion_p (TYPE_MAIN_VARIANT (TREE_TYPE (expr)),
1717 pddatype))
1718 return;
1720 /* The lower bound and element sizes must be constant. */
1721 if (!TYPE_SIZE_UNIT (ddatype)
1722 || TREE_CODE (TYPE_SIZE_UNIT (ddatype)) != INTEGER_CST
1723 || !TYPE_DOMAIN (datype) || !TYPE_MIN_VALUE (TYPE_DOMAIN (datype))
1724 || TREE_CODE (TYPE_MIN_VALUE (TYPE_DOMAIN (datype))) != INTEGER_CST)
1725 return;
1727 /* All checks succeeded. Build a new node to merge the cast. */
1728 *expr_p = build4 (ARRAY_REF, ddatype, TREE_OPERAND (addr_expr, 0),
1729 TYPE_MIN_VALUE (TYPE_DOMAIN (datype)),
1730 NULL_TREE, NULL_TREE);
1731 *expr_p = build1 (ADDR_EXPR, pddatype, *expr_p);
1733 /* We can have stripped a required restrict qualifier above. */
1734 if (!useless_type_conversion_p (TREE_TYPE (expr), TREE_TYPE (*expr_p)))
1735 *expr_p = fold_convert (TREE_TYPE (expr), *expr_p);
1738 /* *EXPR_P is a NOP_EXPR or CONVERT_EXPR. Remove it and/or other conversions
1739 underneath as appropriate. */
1741 static enum gimplify_status
1742 gimplify_conversion (tree *expr_p)
1744 tree tem;
1745 location_t loc = EXPR_LOCATION (*expr_p);
1746 gcc_assert (CONVERT_EXPR_P (*expr_p));
1748 /* Then strip away all but the outermost conversion. */
1749 STRIP_SIGN_NOPS (TREE_OPERAND (*expr_p, 0));
1751 /* And remove the outermost conversion if it's useless. */
1752 if (tree_ssa_useless_type_conversion (*expr_p))
1753 *expr_p = TREE_OPERAND (*expr_p, 0);
1755 /* Attempt to avoid NOP_EXPR by producing reference to a subtype.
1756 For example this fold (subclass *)&A into &A->subclass avoiding
1757 a need for statement. */
1758 if (CONVERT_EXPR_P (*expr_p)
1759 && POINTER_TYPE_P (TREE_TYPE (*expr_p))
1760 && POINTER_TYPE_P (TREE_TYPE (TREE_OPERAND (*expr_p, 0)))
1761 && (tem = maybe_fold_offset_to_address
1762 (EXPR_LOCATION (*expr_p), TREE_OPERAND (*expr_p, 0),
1763 integer_zero_node, TREE_TYPE (*expr_p))) != NULL_TREE)
1764 *expr_p = tem;
1766 /* If we still have a conversion at the toplevel,
1767 then canonicalize some constructs. */
1768 if (CONVERT_EXPR_P (*expr_p))
1770 tree sub = TREE_OPERAND (*expr_p, 0);
1772 /* If a NOP conversion is changing the type of a COMPONENT_REF
1773 expression, then canonicalize its type now in order to expose more
1774 redundant conversions. */
1775 if (TREE_CODE (sub) == COMPONENT_REF)
1776 canonicalize_component_ref (&TREE_OPERAND (*expr_p, 0));
1778 /* If a NOP conversion is changing a pointer to array of foo
1779 to a pointer to foo, embed that change in the ADDR_EXPR. */
1780 else if (TREE_CODE (sub) == ADDR_EXPR)
1781 canonicalize_addr_expr (expr_p);
1784 /* If we have a conversion to a non-register type force the
1785 use of a VIEW_CONVERT_EXPR instead. */
1786 if (CONVERT_EXPR_P (*expr_p) && !is_gimple_reg_type (TREE_TYPE (*expr_p)))
1787 *expr_p = fold_build1_loc (loc, VIEW_CONVERT_EXPR, TREE_TYPE (*expr_p),
1788 TREE_OPERAND (*expr_p, 0));
1790 return GS_OK;
1793 /* Nonlocal VLAs seen in the current function. */
1794 static struct pointer_set_t *nonlocal_vlas;
1796 /* Gimplify a VAR_DECL or PARM_DECL. Returns GS_OK if we expanded a
1797 DECL_VALUE_EXPR, and it's worth re-examining things. */
1799 static enum gimplify_status
1800 gimplify_var_or_parm_decl (tree *expr_p)
1802 tree decl = *expr_p;
1804 /* ??? If this is a local variable, and it has not been seen in any
1805 outer BIND_EXPR, then it's probably the result of a duplicate
1806 declaration, for which we've already issued an error. It would
1807 be really nice if the front end wouldn't leak these at all.
1808 Currently the only known culprit is C++ destructors, as seen
1809 in g++.old-deja/g++.jason/binding.C. */
1810 if (TREE_CODE (decl) == VAR_DECL
1811 && !DECL_SEEN_IN_BIND_EXPR_P (decl)
1812 && !TREE_STATIC (decl) && !DECL_EXTERNAL (decl)
1813 && decl_function_context (decl) == current_function_decl)
1815 gcc_assert (errorcount || sorrycount);
1816 return GS_ERROR;
1819 /* When within an OpenMP context, notice uses of variables. */
1820 if (gimplify_omp_ctxp && omp_notice_variable (gimplify_omp_ctxp, decl, true))
1821 return GS_ALL_DONE;
1823 /* If the decl is an alias for another expression, substitute it now. */
1824 if (DECL_HAS_VALUE_EXPR_P (decl))
1826 tree value_expr = DECL_VALUE_EXPR (decl);
1828 /* For referenced nonlocal VLAs add a decl for debugging purposes
1829 to the current function. */
1830 if (TREE_CODE (decl) == VAR_DECL
1831 && TREE_CODE (DECL_SIZE_UNIT (decl)) != INTEGER_CST
1832 && nonlocal_vlas != NULL
1833 && TREE_CODE (value_expr) == INDIRECT_REF
1834 && TREE_CODE (TREE_OPERAND (value_expr, 0)) == VAR_DECL
1835 && decl_function_context (decl) != current_function_decl)
1837 struct gimplify_omp_ctx *ctx = gimplify_omp_ctxp;
1838 while (ctx && ctx->region_type == ORT_WORKSHARE)
1839 ctx = ctx->outer_context;
1840 if (!ctx && !pointer_set_insert (nonlocal_vlas, decl))
1842 tree copy = copy_node (decl), block;
1844 lang_hooks.dup_lang_specific_decl (copy);
1845 SET_DECL_RTL (copy, NULL_RTX);
1846 TREE_USED (copy) = 1;
1847 block = DECL_INITIAL (current_function_decl);
1848 TREE_CHAIN (copy) = BLOCK_VARS (block);
1849 BLOCK_VARS (block) = copy;
1850 SET_DECL_VALUE_EXPR (copy, unshare_expr (value_expr));
1851 DECL_HAS_VALUE_EXPR_P (copy) = 1;
1855 *expr_p = unshare_expr (value_expr);
1856 return GS_OK;
1859 return GS_ALL_DONE;
1863 /* Gimplify the COMPONENT_REF, ARRAY_REF, REALPART_EXPR or IMAGPART_EXPR
1864 node *EXPR_P.
1866 compound_lval
1867 : min_lval '[' val ']'
1868 | min_lval '.' ID
1869 | compound_lval '[' val ']'
1870 | compound_lval '.' ID
1872 This is not part of the original SIMPLE definition, which separates
1873 array and member references, but it seems reasonable to handle them
1874 together. Also, this way we don't run into problems with union
1875 aliasing; gcc requires that for accesses through a union to alias, the
1876 union reference must be explicit, which was not always the case when we
1877 were splitting up array and member refs.
1879 PRE_P points to the sequence where side effects that must happen before
1880 *EXPR_P should be stored.
1882 POST_P points to the sequence where side effects that must happen after
1883 *EXPR_P should be stored. */
1885 static enum gimplify_status
1886 gimplify_compound_lval (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p,
1887 fallback_t fallback)
1889 tree *p;
1890 VEC(tree,heap) *stack;
1891 enum gimplify_status ret = GS_OK, tret;
1892 int i;
1893 location_t loc = EXPR_LOCATION (*expr_p);
1895 /* Create a stack of the subexpressions so later we can walk them in
1896 order from inner to outer. */
1897 stack = VEC_alloc (tree, heap, 10);
1899 /* We can handle anything that get_inner_reference can deal with. */
1900 for (p = expr_p; ; p = &TREE_OPERAND (*p, 0))
1902 restart:
1903 /* Fold INDIRECT_REFs now to turn them into ARRAY_REFs. */
1904 if (TREE_CODE (*p) == INDIRECT_REF)
1905 *p = fold_indirect_ref_loc (loc, *p);
1907 if (handled_component_p (*p))
1909 /* Expand DECL_VALUE_EXPR now. In some cases that may expose
1910 additional COMPONENT_REFs. */
1911 else if ((TREE_CODE (*p) == VAR_DECL || TREE_CODE (*p) == PARM_DECL)
1912 && gimplify_var_or_parm_decl (p) == GS_OK)
1913 goto restart;
1914 else
1915 break;
1917 VEC_safe_push (tree, heap, stack, *p);
1920 gcc_assert (VEC_length (tree, stack));
1922 /* Now STACK is a stack of pointers to all the refs we've walked through
1923 and P points to the innermost expression.
1925 Java requires that we elaborated nodes in source order. That
1926 means we must gimplify the inner expression followed by each of
1927 the indices, in order. But we can't gimplify the inner
1928 expression until we deal with any variable bounds, sizes, or
1929 positions in order to deal with PLACEHOLDER_EXPRs.
1931 So we do this in three steps. First we deal with the annotations
1932 for any variables in the components, then we gimplify the base,
1933 then we gimplify any indices, from left to right. */
1934 for (i = VEC_length (tree, stack) - 1; i >= 0; i--)
1936 tree t = VEC_index (tree, stack, i);
1938 if (TREE_CODE (t) == ARRAY_REF || TREE_CODE (t) == ARRAY_RANGE_REF)
1940 /* Gimplify the low bound and element type size and put them into
1941 the ARRAY_REF. If these values are set, they have already been
1942 gimplified. */
1943 if (TREE_OPERAND (t, 2) == NULL_TREE)
1945 tree low = unshare_expr (array_ref_low_bound (t));
1946 if (!is_gimple_min_invariant (low))
1948 TREE_OPERAND (t, 2) = low;
1949 tret = gimplify_expr (&TREE_OPERAND (t, 2), pre_p,
1950 post_p, is_gimple_reg,
1951 fb_rvalue);
1952 ret = MIN (ret, tret);
1956 if (!TREE_OPERAND (t, 3))
1958 tree elmt_type = TREE_TYPE (TREE_TYPE (TREE_OPERAND (t, 0)));
1959 tree elmt_size = unshare_expr (array_ref_element_size (t));
1960 tree factor = size_int (TYPE_ALIGN_UNIT (elmt_type));
1962 /* Divide the element size by the alignment of the element
1963 type (above). */
1964 elmt_size = size_binop_loc (loc, EXACT_DIV_EXPR, elmt_size, factor);
1966 if (!is_gimple_min_invariant (elmt_size))
1968 TREE_OPERAND (t, 3) = elmt_size;
1969 tret = gimplify_expr (&TREE_OPERAND (t, 3), pre_p,
1970 post_p, is_gimple_reg,
1971 fb_rvalue);
1972 ret = MIN (ret, tret);
1976 else if (TREE_CODE (t) == COMPONENT_REF)
1978 /* Set the field offset into T and gimplify it. */
1979 if (!TREE_OPERAND (t, 2))
1981 tree offset = unshare_expr (component_ref_field_offset (t));
1982 tree field = TREE_OPERAND (t, 1);
1983 tree factor
1984 = size_int (DECL_OFFSET_ALIGN (field) / BITS_PER_UNIT);
1986 /* Divide the offset by its alignment. */
1987 offset = size_binop_loc (loc, EXACT_DIV_EXPR, offset, factor);
1989 if (!is_gimple_min_invariant (offset))
1991 TREE_OPERAND (t, 2) = offset;
1992 tret = gimplify_expr (&TREE_OPERAND (t, 2), pre_p,
1993 post_p, is_gimple_reg,
1994 fb_rvalue);
1995 ret = MIN (ret, tret);
2001 /* Step 2 is to gimplify the base expression. Make sure lvalue is set
2002 so as to match the min_lval predicate. Failure to do so may result
2003 in the creation of large aggregate temporaries. */
2004 tret = gimplify_expr (p, pre_p, post_p, is_gimple_min_lval,
2005 fallback | fb_lvalue);
2006 ret = MIN (ret, tret);
2008 /* And finally, the indices and operands to BIT_FIELD_REF. During this
2009 loop we also remove any useless conversions. */
2010 for (; VEC_length (tree, stack) > 0; )
2012 tree t = VEC_pop (tree, stack);
2014 if (TREE_CODE (t) == ARRAY_REF || TREE_CODE (t) == ARRAY_RANGE_REF)
2016 /* Gimplify the dimension. */
2017 if (!is_gimple_min_invariant (TREE_OPERAND (t, 1)))
2019 tret = gimplify_expr (&TREE_OPERAND (t, 1), pre_p, post_p,
2020 is_gimple_val, fb_rvalue);
2021 ret = MIN (ret, tret);
2024 else if (TREE_CODE (t) == BIT_FIELD_REF)
2026 tret = gimplify_expr (&TREE_OPERAND (t, 1), pre_p, post_p,
2027 is_gimple_val, fb_rvalue);
2028 ret = MIN (ret, tret);
2029 tret = gimplify_expr (&TREE_OPERAND (t, 2), pre_p, post_p,
2030 is_gimple_val, fb_rvalue);
2031 ret = MIN (ret, tret);
2034 STRIP_USELESS_TYPE_CONVERSION (TREE_OPERAND (t, 0));
2036 /* The innermost expression P may have originally had
2037 TREE_SIDE_EFFECTS set which would have caused all the outer
2038 expressions in *EXPR_P leading to P to also have had
2039 TREE_SIDE_EFFECTS set. */
2040 recalculate_side_effects (t);
2043 /* If the outermost expression is a COMPONENT_REF, canonicalize its type. */
2044 if ((fallback & fb_rvalue) && TREE_CODE (*expr_p) == COMPONENT_REF)
2046 canonicalize_component_ref (expr_p);
2047 ret = MIN (ret, GS_OK);
2050 VEC_free (tree, heap, stack);
2052 return ret;
2055 /* Gimplify the self modifying expression pointed to by EXPR_P
2056 (++, --, +=, -=).
2058 PRE_P points to the list where side effects that must happen before
2059 *EXPR_P should be stored.
2061 POST_P points to the list where side effects that must happen after
2062 *EXPR_P should be stored.
2064 WANT_VALUE is nonzero iff we want to use the value of this expression
2065 in another expression. */
2067 static enum gimplify_status
2068 gimplify_self_mod_expr (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p,
2069 bool want_value)
2071 enum tree_code code;
2072 tree lhs, lvalue, rhs, t1;
2073 gimple_seq post = NULL, *orig_post_p = post_p;
2074 bool postfix;
2075 enum tree_code arith_code;
2076 enum gimplify_status ret;
2077 location_t loc = EXPR_LOCATION (*expr_p);
2079 code = TREE_CODE (*expr_p);
2081 gcc_assert (code == POSTINCREMENT_EXPR || code == POSTDECREMENT_EXPR
2082 || code == PREINCREMENT_EXPR || code == PREDECREMENT_EXPR);
2084 /* Prefix or postfix? */
2085 if (code == POSTINCREMENT_EXPR || code == POSTDECREMENT_EXPR)
2086 /* Faster to treat as prefix if result is not used. */
2087 postfix = want_value;
2088 else
2089 postfix = false;
2091 /* For postfix, make sure the inner expression's post side effects
2092 are executed after side effects from this expression. */
2093 if (postfix)
2094 post_p = &post;
2096 /* Add or subtract? */
2097 if (code == PREINCREMENT_EXPR || code == POSTINCREMENT_EXPR)
2098 arith_code = PLUS_EXPR;
2099 else
2100 arith_code = MINUS_EXPR;
2102 /* Gimplify the LHS into a GIMPLE lvalue. */
2103 lvalue = TREE_OPERAND (*expr_p, 0);
2104 ret = gimplify_expr (&lvalue, pre_p, post_p, is_gimple_lvalue, fb_lvalue);
2105 if (ret == GS_ERROR)
2106 return ret;
2108 /* Extract the operands to the arithmetic operation. */
2109 lhs = lvalue;
2110 rhs = TREE_OPERAND (*expr_p, 1);
2112 /* For postfix operator, we evaluate the LHS to an rvalue and then use
2113 that as the result value and in the postqueue operation. We also
2114 make sure to make lvalue a minimal lval, see
2115 gcc.c-torture/execute/20040313-1.c for an example where this matters. */
2116 if (postfix)
2118 if (!is_gimple_min_lval (lvalue))
2120 mark_addressable (lvalue);
2121 lvalue = build_fold_addr_expr_loc (input_location, lvalue);
2122 gimplify_expr (&lvalue, pre_p, post_p, is_gimple_val, fb_rvalue);
2123 lvalue = build_fold_indirect_ref_loc (input_location, lvalue);
2125 ret = gimplify_expr (&lhs, pre_p, post_p, is_gimple_val, fb_rvalue);
2126 if (ret == GS_ERROR)
2127 return ret;
2130 /* For POINTERs increment, use POINTER_PLUS_EXPR. */
2131 if (POINTER_TYPE_P (TREE_TYPE (lhs)))
2133 rhs = fold_convert_loc (loc, sizetype, rhs);
2134 if (arith_code == MINUS_EXPR)
2135 rhs = fold_build1_loc (loc, NEGATE_EXPR, TREE_TYPE (rhs), rhs);
2136 arith_code = POINTER_PLUS_EXPR;
2139 t1 = build2 (arith_code, TREE_TYPE (*expr_p), lhs, rhs);
2141 if (postfix)
2143 gimplify_assign (lvalue, t1, orig_post_p);
2144 gimplify_seq_add_seq (orig_post_p, post);
2145 *expr_p = lhs;
2146 return GS_ALL_DONE;
2148 else
2150 *expr_p = build2 (MODIFY_EXPR, TREE_TYPE (lvalue), lvalue, t1);
2151 return GS_OK;
2156 /* If *EXPR_P has a variable sized type, wrap it in a WITH_SIZE_EXPR. */
2158 static void
2159 maybe_with_size_expr (tree *expr_p)
2161 tree expr = *expr_p;
2162 tree type = TREE_TYPE (expr);
2163 tree size;
2165 /* If we've already wrapped this or the type is error_mark_node, we can't do
2166 anything. */
2167 if (TREE_CODE (expr) == WITH_SIZE_EXPR
2168 || type == error_mark_node)
2169 return;
2171 /* If the size isn't known or is a constant, we have nothing to do. */
2172 size = TYPE_SIZE_UNIT (type);
2173 if (!size || TREE_CODE (size) == INTEGER_CST)
2174 return;
2176 /* Otherwise, make a WITH_SIZE_EXPR. */
2177 size = unshare_expr (size);
2178 size = SUBSTITUTE_PLACEHOLDER_IN_EXPR (size, expr);
2179 *expr_p = build2 (WITH_SIZE_EXPR, type, expr, size);
2183 /* Helper for gimplify_call_expr. Gimplify a single argument *ARG_P
2184 Store any side-effects in PRE_P. CALL_LOCATION is the location of
2185 the CALL_EXPR. */
2187 static enum gimplify_status
2188 gimplify_arg (tree *arg_p, gimple_seq *pre_p, location_t call_location)
2190 bool (*test) (tree);
2191 fallback_t fb;
2193 /* In general, we allow lvalues for function arguments to avoid
2194 extra overhead of copying large aggregates out of even larger
2195 aggregates into temporaries only to copy the temporaries to
2196 the argument list. Make optimizers happy by pulling out to
2197 temporaries those types that fit in registers. */
2198 if (is_gimple_reg_type (TREE_TYPE (*arg_p)))
2199 test = is_gimple_val, fb = fb_rvalue;
2200 else
2201 test = is_gimple_lvalue, fb = fb_either;
2203 /* If this is a variable sized type, we must remember the size. */
2204 maybe_with_size_expr (arg_p);
2206 /* FIXME diagnostics: This will mess up gcc.dg/Warray-bounds.c. */
2207 /* Make sure arguments have the same location as the function call
2208 itself. */
2209 protected_set_expr_location (*arg_p, call_location);
2211 /* There is a sequence point before a function call. Side effects in
2212 the argument list must occur before the actual call. So, when
2213 gimplifying arguments, force gimplify_expr to use an internal
2214 post queue which is then appended to the end of PRE_P. */
2215 return gimplify_expr (arg_p, pre_p, NULL, test, fb);
2219 /* Gimplify the CALL_EXPR node *EXPR_P into the GIMPLE sequence PRE_P.
2220 WANT_VALUE is true if the result of the call is desired. */
2222 static enum gimplify_status
2223 gimplify_call_expr (tree *expr_p, gimple_seq *pre_p, bool want_value)
2225 tree fndecl, parms, p;
2226 enum gimplify_status ret;
2227 int i, nargs;
2228 gimple call;
2229 bool builtin_va_start_p = FALSE;
2230 location_t loc = EXPR_LOCATION (*expr_p);
2232 gcc_assert (TREE_CODE (*expr_p) == CALL_EXPR);
2234 /* For reliable diagnostics during inlining, it is necessary that
2235 every call_expr be annotated with file and line. */
2236 if (! EXPR_HAS_LOCATION (*expr_p))
2237 SET_EXPR_LOCATION (*expr_p, input_location);
2239 /* This may be a call to a builtin function.
2241 Builtin function calls may be transformed into different
2242 (and more efficient) builtin function calls under certain
2243 circumstances. Unfortunately, gimplification can muck things
2244 up enough that the builtin expanders are not aware that certain
2245 transformations are still valid.
2247 So we attempt transformation/gimplification of the call before
2248 we gimplify the CALL_EXPR. At this time we do not manage to
2249 transform all calls in the same manner as the expanders do, but
2250 we do transform most of them. */
2251 fndecl = get_callee_fndecl (*expr_p);
2252 if (fndecl && DECL_BUILT_IN (fndecl))
2254 tree new_tree = fold_call_expr (input_location, *expr_p, !want_value);
2256 if (new_tree && new_tree != *expr_p)
2258 /* There was a transformation of this call which computes the
2259 same value, but in a more efficient way. Return and try
2260 again. */
2261 *expr_p = new_tree;
2262 return GS_OK;
2265 if (DECL_BUILT_IN_CLASS (fndecl) == BUILT_IN_NORMAL
2266 && DECL_FUNCTION_CODE (fndecl) == BUILT_IN_VA_START)
2268 builtin_va_start_p = TRUE;
2269 if (call_expr_nargs (*expr_p) < 2)
2271 error ("too few arguments to function %<va_start%>");
2272 *expr_p = build_empty_stmt (EXPR_LOCATION (*expr_p));
2273 return GS_OK;
2276 if (fold_builtin_next_arg (*expr_p, true))
2278 *expr_p = build_empty_stmt (EXPR_LOCATION (*expr_p));
2279 return GS_OK;
2284 /* There is a sequence point before the call, so any side effects in
2285 the calling expression must occur before the actual call. Force
2286 gimplify_expr to use an internal post queue. */
2287 ret = gimplify_expr (&CALL_EXPR_FN (*expr_p), pre_p, NULL,
2288 is_gimple_call_addr, fb_rvalue);
2290 nargs = call_expr_nargs (*expr_p);
2292 /* Get argument types for verification. */
2293 fndecl = get_callee_fndecl (*expr_p);
2294 parms = NULL_TREE;
2295 if (fndecl)
2296 parms = TYPE_ARG_TYPES (TREE_TYPE (fndecl));
2297 else if (POINTER_TYPE_P (TREE_TYPE (CALL_EXPR_FN (*expr_p))))
2298 parms = TYPE_ARG_TYPES (TREE_TYPE (TREE_TYPE (CALL_EXPR_FN (*expr_p))));
2300 if (fndecl && DECL_ARGUMENTS (fndecl))
2301 p = DECL_ARGUMENTS (fndecl);
2302 else if (parms)
2303 p = parms;
2304 else
2305 p = NULL_TREE;
2306 for (i = 0; i < nargs && p; i++, p = TREE_CHAIN (p))
2309 /* If the last argument is __builtin_va_arg_pack () and it is not
2310 passed as a named argument, decrease the number of CALL_EXPR
2311 arguments and set instead the CALL_EXPR_VA_ARG_PACK flag. */
2312 if (!p
2313 && i < nargs
2314 && TREE_CODE (CALL_EXPR_ARG (*expr_p, nargs - 1)) == CALL_EXPR)
2316 tree last_arg = CALL_EXPR_ARG (*expr_p, nargs - 1);
2317 tree last_arg_fndecl = get_callee_fndecl (last_arg);
2319 if (last_arg_fndecl
2320 && TREE_CODE (last_arg_fndecl) == FUNCTION_DECL
2321 && DECL_BUILT_IN_CLASS (last_arg_fndecl) == BUILT_IN_NORMAL
2322 && DECL_FUNCTION_CODE (last_arg_fndecl) == BUILT_IN_VA_ARG_PACK)
2324 tree call = *expr_p;
2326 --nargs;
2327 *expr_p = build_call_array_loc (loc, TREE_TYPE (call),
2328 CALL_EXPR_FN (call),
2329 nargs, CALL_EXPR_ARGP (call));
2331 /* Copy all CALL_EXPR flags, location and block, except
2332 CALL_EXPR_VA_ARG_PACK flag. */
2333 CALL_EXPR_STATIC_CHAIN (*expr_p) = CALL_EXPR_STATIC_CHAIN (call);
2334 CALL_EXPR_TAILCALL (*expr_p) = CALL_EXPR_TAILCALL (call);
2335 CALL_EXPR_RETURN_SLOT_OPT (*expr_p)
2336 = CALL_EXPR_RETURN_SLOT_OPT (call);
2337 CALL_FROM_THUNK_P (*expr_p) = CALL_FROM_THUNK_P (call);
2338 CALL_CANNOT_INLINE_P (*expr_p) = CALL_CANNOT_INLINE_P (call);
2339 SET_EXPR_LOCATION (*expr_p, EXPR_LOCATION (call));
2340 TREE_BLOCK (*expr_p) = TREE_BLOCK (call);
2342 /* Set CALL_EXPR_VA_ARG_PACK. */
2343 CALL_EXPR_VA_ARG_PACK (*expr_p) = 1;
2347 /* Finally, gimplify the function arguments. */
2348 if (nargs > 0)
2350 for (i = (PUSH_ARGS_REVERSED ? nargs - 1 : 0);
2351 PUSH_ARGS_REVERSED ? i >= 0 : i < nargs;
2352 PUSH_ARGS_REVERSED ? i-- : i++)
2354 enum gimplify_status t;
2356 /* Avoid gimplifying the second argument to va_start, which needs to
2357 be the plain PARM_DECL. */
2358 if ((i != 1) || !builtin_va_start_p)
2360 t = gimplify_arg (&CALL_EXPR_ARG (*expr_p, i), pre_p,
2361 EXPR_LOCATION (*expr_p));
2363 if (t == GS_ERROR)
2364 ret = GS_ERROR;
2369 /* Verify the function result. */
2370 if (want_value && fndecl
2371 && VOID_TYPE_P (TREE_TYPE (TREE_TYPE (fndecl))))
2373 error_at (loc, "using result of function returning %<void%>");
2374 ret = GS_ERROR;
2377 /* Try this again in case gimplification exposed something. */
2378 if (ret != GS_ERROR)
2380 tree new_tree = fold_call_expr (input_location, *expr_p, !want_value);
2382 if (new_tree && new_tree != *expr_p)
2384 /* There was a transformation of this call which computes the
2385 same value, but in a more efficient way. Return and try
2386 again. */
2387 *expr_p = new_tree;
2388 return GS_OK;
2391 else
2393 *expr_p = error_mark_node;
2394 return GS_ERROR;
2397 /* If the function is "const" or "pure", then clear TREE_SIDE_EFFECTS on its
2398 decl. This allows us to eliminate redundant or useless
2399 calls to "const" functions. */
2400 if (TREE_CODE (*expr_p) == CALL_EXPR)
2402 int flags = call_expr_flags (*expr_p);
2403 if (flags & (ECF_CONST | ECF_PURE)
2404 /* An infinite loop is considered a side effect. */
2405 && !(flags & (ECF_LOOPING_CONST_OR_PURE)))
2406 TREE_SIDE_EFFECTS (*expr_p) = 0;
2409 /* If the value is not needed by the caller, emit a new GIMPLE_CALL
2410 and clear *EXPR_P. Otherwise, leave *EXPR_P in its gimplified
2411 form and delegate the creation of a GIMPLE_CALL to
2412 gimplify_modify_expr. This is always possible because when
2413 WANT_VALUE is true, the caller wants the result of this call into
2414 a temporary, which means that we will emit an INIT_EXPR in
2415 internal_get_tmp_var which will then be handled by
2416 gimplify_modify_expr. */
2417 if (!want_value)
2419 /* The CALL_EXPR in *EXPR_P is already in GIMPLE form, so all we
2420 have to do is replicate it as a GIMPLE_CALL tuple. */
2421 call = gimple_build_call_from_tree (*expr_p);
2422 gimplify_seq_add_stmt (pre_p, call);
2423 *expr_p = NULL_TREE;
2426 return ret;
2429 /* Handle shortcut semantics in the predicate operand of a COND_EXPR by
2430 rewriting it into multiple COND_EXPRs, and possibly GOTO_EXPRs.
2432 TRUE_LABEL_P and FALSE_LABEL_P point to the labels to jump to if the
2433 condition is true or false, respectively. If null, we should generate
2434 our own to skip over the evaluation of this specific expression.
2436 LOCUS is the source location of the COND_EXPR.
2438 This function is the tree equivalent of do_jump.
2440 shortcut_cond_r should only be called by shortcut_cond_expr. */
2442 static tree
2443 shortcut_cond_r (tree pred, tree *true_label_p, tree *false_label_p,
2444 location_t locus)
2446 tree local_label = NULL_TREE;
2447 tree t, expr = NULL;
2449 /* OK, it's not a simple case; we need to pull apart the COND_EXPR to
2450 retain the shortcut semantics. Just insert the gotos here;
2451 shortcut_cond_expr will append the real blocks later. */
2452 if (TREE_CODE (pred) == TRUTH_ANDIF_EXPR)
2454 location_t new_locus;
2456 /* Turn if (a && b) into
2458 if (a); else goto no;
2459 if (b) goto yes; else goto no;
2460 (no:) */
2462 if (false_label_p == NULL)
2463 false_label_p = &local_label;
2465 /* Keep the original source location on the first 'if'. */
2466 t = shortcut_cond_r (TREE_OPERAND (pred, 0), NULL, false_label_p, locus);
2467 append_to_statement_list (t, &expr);
2469 /* Set the source location of the && on the second 'if'. */
2470 new_locus = EXPR_HAS_LOCATION (pred) ? EXPR_LOCATION (pred) : locus;
2471 t = shortcut_cond_r (TREE_OPERAND (pred, 1), true_label_p, false_label_p,
2472 new_locus);
2473 append_to_statement_list (t, &expr);
2475 else if (TREE_CODE (pred) == TRUTH_ORIF_EXPR)
2477 location_t new_locus;
2479 /* Turn if (a || b) into
2481 if (a) goto yes;
2482 if (b) goto yes; else goto no;
2483 (yes:) */
2485 if (true_label_p == NULL)
2486 true_label_p = &local_label;
2488 /* Keep the original source location on the first 'if'. */
2489 t = shortcut_cond_r (TREE_OPERAND (pred, 0), true_label_p, NULL, locus);
2490 append_to_statement_list (t, &expr);
2492 /* Set the source location of the || on the second 'if'. */
2493 new_locus = EXPR_HAS_LOCATION (pred) ? EXPR_LOCATION (pred) : locus;
2494 t = shortcut_cond_r (TREE_OPERAND (pred, 1), true_label_p, false_label_p,
2495 new_locus);
2496 append_to_statement_list (t, &expr);
2498 else if (TREE_CODE (pred) == COND_EXPR)
2500 location_t new_locus;
2502 /* As long as we're messing with gotos, turn if (a ? b : c) into
2503 if (a)
2504 if (b) goto yes; else goto no;
2505 else
2506 if (c) goto yes; else goto no; */
2508 /* Keep the original source location on the first 'if'. Set the source
2509 location of the ? on the second 'if'. */
2510 new_locus = EXPR_HAS_LOCATION (pred) ? EXPR_LOCATION (pred) : locus;
2511 expr = build3 (COND_EXPR, void_type_node, TREE_OPERAND (pred, 0),
2512 shortcut_cond_r (TREE_OPERAND (pred, 1), true_label_p,
2513 false_label_p, locus),
2514 shortcut_cond_r (TREE_OPERAND (pred, 2), true_label_p,
2515 false_label_p, new_locus));
2517 else
2519 expr = build3 (COND_EXPR, void_type_node, pred,
2520 build_and_jump (true_label_p),
2521 build_and_jump (false_label_p));
2522 SET_EXPR_LOCATION (expr, locus);
2525 if (local_label)
2527 t = build1 (LABEL_EXPR, void_type_node, local_label);
2528 append_to_statement_list (t, &expr);
2531 return expr;
2534 /* Given a conditional expression EXPR with short-circuit boolean
2535 predicates using TRUTH_ANDIF_EXPR or TRUTH_ORIF_EXPR, break the
2536 predicate appart into the equivalent sequence of conditionals. */
2538 static tree
2539 shortcut_cond_expr (tree expr)
2541 tree pred = TREE_OPERAND (expr, 0);
2542 tree then_ = TREE_OPERAND (expr, 1);
2543 tree else_ = TREE_OPERAND (expr, 2);
2544 tree true_label, false_label, end_label, t;
2545 tree *true_label_p;
2546 tree *false_label_p;
2547 bool emit_end, emit_false, jump_over_else;
2548 bool then_se = then_ && TREE_SIDE_EFFECTS (then_);
2549 bool else_se = else_ && TREE_SIDE_EFFECTS (else_);
2551 /* First do simple transformations. */
2552 if (!else_se)
2554 /* If there is no 'else', turn
2555 if (a && b) then c
2556 into
2557 if (a) if (b) then c. */
2558 while (TREE_CODE (pred) == TRUTH_ANDIF_EXPR)
2560 /* Keep the original source location on the first 'if'. */
2561 location_t locus = EXPR_HAS_LOCATION (expr)
2562 ? EXPR_LOCATION (expr) : input_location;
2563 TREE_OPERAND (expr, 0) = TREE_OPERAND (pred, 1);
2564 /* Set the source location of the && on the second 'if'. */
2565 if (EXPR_HAS_LOCATION (pred))
2566 SET_EXPR_LOCATION (expr, EXPR_LOCATION (pred));
2567 then_ = shortcut_cond_expr (expr);
2568 then_se = then_ && TREE_SIDE_EFFECTS (then_);
2569 pred = TREE_OPERAND (pred, 0);
2570 expr = build3 (COND_EXPR, void_type_node, pred, then_, NULL_TREE);
2571 SET_EXPR_LOCATION (expr, locus);
2575 if (!then_se)
2577 /* If there is no 'then', turn
2578 if (a || b); else d
2579 into
2580 if (a); else if (b); else d. */
2581 while (TREE_CODE (pred) == TRUTH_ORIF_EXPR)
2583 /* Keep the original source location on the first 'if'. */
2584 location_t locus = EXPR_HAS_LOCATION (expr)
2585 ? EXPR_LOCATION (expr) : input_location;
2586 TREE_OPERAND (expr, 0) = TREE_OPERAND (pred, 1);
2587 /* Set the source location of the || on the second 'if'. */
2588 if (EXPR_HAS_LOCATION (pred))
2589 SET_EXPR_LOCATION (expr, EXPR_LOCATION (pred));
2590 else_ = shortcut_cond_expr (expr);
2591 else_se = else_ && TREE_SIDE_EFFECTS (else_);
2592 pred = TREE_OPERAND (pred, 0);
2593 expr = build3 (COND_EXPR, void_type_node, pred, NULL_TREE, else_);
2594 SET_EXPR_LOCATION (expr, locus);
2598 /* If we're done, great. */
2599 if (TREE_CODE (pred) != TRUTH_ANDIF_EXPR
2600 && TREE_CODE (pred) != TRUTH_ORIF_EXPR)
2601 return expr;
2603 /* Otherwise we need to mess with gotos. Change
2604 if (a) c; else d;
2606 if (a); else goto no;
2607 c; goto end;
2608 no: d; end:
2609 and recursively gimplify the condition. */
2611 true_label = false_label = end_label = NULL_TREE;
2613 /* If our arms just jump somewhere, hijack those labels so we don't
2614 generate jumps to jumps. */
2616 if (then_
2617 && TREE_CODE (then_) == GOTO_EXPR
2618 && TREE_CODE (GOTO_DESTINATION (then_)) == LABEL_DECL)
2620 true_label = GOTO_DESTINATION (then_);
2621 then_ = NULL;
2622 then_se = false;
2625 if (else_
2626 && TREE_CODE (else_) == GOTO_EXPR
2627 && TREE_CODE (GOTO_DESTINATION (else_)) == LABEL_DECL)
2629 false_label = GOTO_DESTINATION (else_);
2630 else_ = NULL;
2631 else_se = false;
2634 /* If we aren't hijacking a label for the 'then' branch, it falls through. */
2635 if (true_label)
2636 true_label_p = &true_label;
2637 else
2638 true_label_p = NULL;
2640 /* The 'else' branch also needs a label if it contains interesting code. */
2641 if (false_label || else_se)
2642 false_label_p = &false_label;
2643 else
2644 false_label_p = NULL;
2646 /* If there was nothing else in our arms, just forward the label(s). */
2647 if (!then_se && !else_se)
2648 return shortcut_cond_r (pred, true_label_p, false_label_p,
2649 EXPR_HAS_LOCATION (expr)
2650 ? EXPR_LOCATION (expr) : input_location);
2652 /* If our last subexpression already has a terminal label, reuse it. */
2653 if (else_se)
2654 t = expr_last (else_);
2655 else if (then_se)
2656 t = expr_last (then_);
2657 else
2658 t = NULL;
2659 if (t && TREE_CODE (t) == LABEL_EXPR)
2660 end_label = LABEL_EXPR_LABEL (t);
2662 /* If we don't care about jumping to the 'else' branch, jump to the end
2663 if the condition is false. */
2664 if (!false_label_p)
2665 false_label_p = &end_label;
2667 /* We only want to emit these labels if we aren't hijacking them. */
2668 emit_end = (end_label == NULL_TREE);
2669 emit_false = (false_label == NULL_TREE);
2671 /* We only emit the jump over the else clause if we have to--if the
2672 then clause may fall through. Otherwise we can wind up with a
2673 useless jump and a useless label at the end of gimplified code,
2674 which will cause us to think that this conditional as a whole
2675 falls through even if it doesn't. If we then inline a function
2676 which ends with such a condition, that can cause us to issue an
2677 inappropriate warning about control reaching the end of a
2678 non-void function. */
2679 jump_over_else = block_may_fallthru (then_);
2681 pred = shortcut_cond_r (pred, true_label_p, false_label_p,
2682 EXPR_HAS_LOCATION (expr)
2683 ? EXPR_LOCATION (expr) : input_location);
2685 expr = NULL;
2686 append_to_statement_list (pred, &expr);
2688 append_to_statement_list (then_, &expr);
2689 if (else_se)
2691 if (jump_over_else)
2693 tree last = expr_last (expr);
2694 t = build_and_jump (&end_label);
2695 if (EXPR_HAS_LOCATION (last))
2696 SET_EXPR_LOCATION (t, EXPR_LOCATION (last));
2697 append_to_statement_list (t, &expr);
2699 if (emit_false)
2701 t = build1 (LABEL_EXPR, void_type_node, false_label);
2702 append_to_statement_list (t, &expr);
2704 append_to_statement_list (else_, &expr);
2706 if (emit_end && end_label)
2708 t = build1 (LABEL_EXPR, void_type_node, end_label);
2709 append_to_statement_list (t, &expr);
2712 return expr;
2715 /* EXPR is used in a boolean context; make sure it has BOOLEAN_TYPE. */
2717 tree
2718 gimple_boolify (tree expr)
2720 tree type = TREE_TYPE (expr);
2721 location_t loc = EXPR_LOCATION (expr);
2723 if (TREE_CODE (expr) == NE_EXPR
2724 && TREE_CODE (TREE_OPERAND (expr, 0)) == CALL_EXPR
2725 && integer_zerop (TREE_OPERAND (expr, 1)))
2727 tree call = TREE_OPERAND (expr, 0);
2728 tree fn = get_callee_fndecl (call);
2730 /* For __builtin_expect ((long) (x), y) recurse into x as well
2731 if x is truth_value_p. */
2732 if (fn
2733 && DECL_BUILT_IN_CLASS (fn) == BUILT_IN_NORMAL
2734 && DECL_FUNCTION_CODE (fn) == BUILT_IN_EXPECT
2735 && call_expr_nargs (call) == 2)
2737 tree arg = CALL_EXPR_ARG (call, 0);
2738 if (arg)
2740 if (TREE_CODE (arg) == NOP_EXPR
2741 && TREE_TYPE (arg) == TREE_TYPE (call))
2742 arg = TREE_OPERAND (arg, 0);
2743 if (truth_value_p (TREE_CODE (arg)))
2745 arg = gimple_boolify (arg);
2746 CALL_EXPR_ARG (call, 0)
2747 = fold_convert_loc (loc, TREE_TYPE (call), arg);
2753 if (TREE_CODE (type) == BOOLEAN_TYPE)
2754 return expr;
2756 switch (TREE_CODE (expr))
2758 case TRUTH_AND_EXPR:
2759 case TRUTH_OR_EXPR:
2760 case TRUTH_XOR_EXPR:
2761 case TRUTH_ANDIF_EXPR:
2762 case TRUTH_ORIF_EXPR:
2763 /* Also boolify the arguments of truth exprs. */
2764 TREE_OPERAND (expr, 1) = gimple_boolify (TREE_OPERAND (expr, 1));
2765 /* FALLTHRU */
2767 case TRUTH_NOT_EXPR:
2768 TREE_OPERAND (expr, 0) = gimple_boolify (TREE_OPERAND (expr, 0));
2769 /* FALLTHRU */
2771 case EQ_EXPR: case NE_EXPR:
2772 case LE_EXPR: case GE_EXPR: case LT_EXPR: case GT_EXPR:
2773 /* These expressions always produce boolean results. */
2774 TREE_TYPE (expr) = boolean_type_node;
2775 return expr;
2777 default:
2778 /* Other expressions that get here must have boolean values, but
2779 might need to be converted to the appropriate mode. */
2780 return fold_convert_loc (loc, boolean_type_node, expr);
2784 /* Given a conditional expression *EXPR_P without side effects, gimplify
2785 its operands. New statements are inserted to PRE_P. */
2787 static enum gimplify_status
2788 gimplify_pure_cond_expr (tree *expr_p, gimple_seq *pre_p)
2790 tree expr = *expr_p, cond;
2791 enum gimplify_status ret, tret;
2792 enum tree_code code;
2794 cond = gimple_boolify (COND_EXPR_COND (expr));
2796 /* We need to handle && and || specially, as their gimplification
2797 creates pure cond_expr, thus leading to an infinite cycle otherwise. */
2798 code = TREE_CODE (cond);
2799 if (code == TRUTH_ANDIF_EXPR)
2800 TREE_SET_CODE (cond, TRUTH_AND_EXPR);
2801 else if (code == TRUTH_ORIF_EXPR)
2802 TREE_SET_CODE (cond, TRUTH_OR_EXPR);
2803 ret = gimplify_expr (&cond, pre_p, NULL, is_gimple_condexpr, fb_rvalue);
2804 COND_EXPR_COND (*expr_p) = cond;
2806 tret = gimplify_expr (&COND_EXPR_THEN (expr), pre_p, NULL,
2807 is_gimple_val, fb_rvalue);
2808 ret = MIN (ret, tret);
2809 tret = gimplify_expr (&COND_EXPR_ELSE (expr), pre_p, NULL,
2810 is_gimple_val, fb_rvalue);
2812 return MIN (ret, tret);
2815 /* Returns true if evaluating EXPR could trap.
2816 EXPR is GENERIC, while tree_could_trap_p can be called
2817 only on GIMPLE. */
2819 static bool
2820 generic_expr_could_trap_p (tree expr)
2822 unsigned i, n;
2824 if (!expr || is_gimple_val (expr))
2825 return false;
2827 if (!EXPR_P (expr) || tree_could_trap_p (expr))
2828 return true;
2830 n = TREE_OPERAND_LENGTH (expr);
2831 for (i = 0; i < n; i++)
2832 if (generic_expr_could_trap_p (TREE_OPERAND (expr, i)))
2833 return true;
2835 return false;
2838 /* Convert the conditional expression pointed to by EXPR_P '(p) ? a : b;'
2839 into
2841 if (p) if (p)
2842 t1 = a; a;
2843 else or else
2844 t1 = b; b;
2847 The second form is used when *EXPR_P is of type void.
2849 PRE_P points to the list where side effects that must happen before
2850 *EXPR_P should be stored. */
2852 static enum gimplify_status
2853 gimplify_cond_expr (tree *expr_p, gimple_seq *pre_p, fallback_t fallback)
2855 tree expr = *expr_p;
2856 tree tmp, type, arm1, arm2;
2857 enum gimplify_status ret;
2858 tree label_true, label_false, label_cont;
2859 bool have_then_clause_p, have_else_clause_p;
2860 gimple gimple_cond;
2861 enum tree_code pred_code;
2862 gimple_seq seq = NULL;
2863 location_t loc = EXPR_LOCATION (*expr_p);
2865 type = TREE_TYPE (expr);
2867 /* If this COND_EXPR has a value, copy the values into a temporary within
2868 the arms. */
2869 if (! VOID_TYPE_P (type))
2871 tree result;
2873 /* If an rvalue is ok or we do not require an lvalue, avoid creating
2874 an addressable temporary. */
2875 if (((fallback & fb_rvalue)
2876 || !(fallback & fb_lvalue))
2877 && !TREE_ADDRESSABLE (type))
2879 if (gimplify_ctxp->allow_rhs_cond_expr
2880 /* If either branch has side effects or could trap, it can't be
2881 evaluated unconditionally. */
2882 && !TREE_SIDE_EFFECTS (TREE_OPERAND (*expr_p, 1))
2883 && !generic_expr_could_trap_p (TREE_OPERAND (*expr_p, 1))
2884 && !TREE_SIDE_EFFECTS (TREE_OPERAND (*expr_p, 2))
2885 && !generic_expr_could_trap_p (TREE_OPERAND (*expr_p, 2)))
2886 return gimplify_pure_cond_expr (expr_p, pre_p);
2888 result = tmp = create_tmp_var (TREE_TYPE (expr), "iftmp");
2889 ret = GS_ALL_DONE;
2891 else
2893 tree type = build_pointer_type (TREE_TYPE (expr));
2895 if (TREE_TYPE (TREE_OPERAND (expr, 1)) != void_type_node)
2896 TREE_OPERAND (expr, 1) =
2897 build_fold_addr_expr_loc (loc, TREE_OPERAND (expr, 1));
2899 if (TREE_TYPE (TREE_OPERAND (expr, 2)) != void_type_node)
2900 TREE_OPERAND (expr, 2) =
2901 build_fold_addr_expr_loc (loc, TREE_OPERAND (expr, 2));
2903 tmp = create_tmp_var (type, "iftmp");
2905 expr = build3 (COND_EXPR, void_type_node, TREE_OPERAND (expr, 0),
2906 TREE_OPERAND (expr, 1), TREE_OPERAND (expr, 2));
2908 result = build_fold_indirect_ref_loc (loc, tmp);
2911 /* Build the then clause, 't1 = a;'. But don't build an assignment
2912 if this branch is void; in C++ it can be, if it's a throw. */
2913 if (TREE_TYPE (TREE_OPERAND (expr, 1)) != void_type_node)
2914 TREE_OPERAND (expr, 1)
2915 = build2 (MODIFY_EXPR, TREE_TYPE (tmp), tmp, TREE_OPERAND (expr, 1));
2917 /* Build the else clause, 't1 = b;'. */
2918 if (TREE_TYPE (TREE_OPERAND (expr, 2)) != void_type_node)
2919 TREE_OPERAND (expr, 2)
2920 = build2 (MODIFY_EXPR, TREE_TYPE (tmp), tmp, TREE_OPERAND (expr, 2));
2922 TREE_TYPE (expr) = void_type_node;
2923 recalculate_side_effects (expr);
2925 /* Move the COND_EXPR to the prequeue. */
2926 gimplify_stmt (&expr, pre_p);
2928 *expr_p = result;
2929 return GS_ALL_DONE;
2932 /* Make sure the condition has BOOLEAN_TYPE. */
2933 TREE_OPERAND (expr, 0) = gimple_boolify (TREE_OPERAND (expr, 0));
2935 /* Break apart && and || conditions. */
2936 if (TREE_CODE (TREE_OPERAND (expr, 0)) == TRUTH_ANDIF_EXPR
2937 || TREE_CODE (TREE_OPERAND (expr, 0)) == TRUTH_ORIF_EXPR)
2939 expr = shortcut_cond_expr (expr);
2941 if (expr != *expr_p)
2943 *expr_p = expr;
2945 /* We can't rely on gimplify_expr to re-gimplify the expanded
2946 form properly, as cleanups might cause the target labels to be
2947 wrapped in a TRY_FINALLY_EXPR. To prevent that, we need to
2948 set up a conditional context. */
2949 gimple_push_condition ();
2950 gimplify_stmt (expr_p, &seq);
2951 gimple_pop_condition (pre_p);
2952 gimple_seq_add_seq (pre_p, seq);
2954 return GS_ALL_DONE;
2958 /* Now do the normal gimplification. */
2960 /* Gimplify condition. */
2961 ret = gimplify_expr (&TREE_OPERAND (expr, 0), pre_p, NULL, is_gimple_condexpr,
2962 fb_rvalue);
2963 if (ret == GS_ERROR)
2964 return GS_ERROR;
2965 gcc_assert (TREE_OPERAND (expr, 0) != NULL_TREE);
2967 gimple_push_condition ();
2969 have_then_clause_p = have_else_clause_p = false;
2970 if (TREE_OPERAND (expr, 1) != NULL
2971 && TREE_CODE (TREE_OPERAND (expr, 1)) == GOTO_EXPR
2972 && TREE_CODE (GOTO_DESTINATION (TREE_OPERAND (expr, 1))) == LABEL_DECL
2973 && (DECL_CONTEXT (GOTO_DESTINATION (TREE_OPERAND (expr, 1)))
2974 == current_function_decl)
2975 /* For -O0 avoid this optimization if the COND_EXPR and GOTO_EXPR
2976 have different locations, otherwise we end up with incorrect
2977 location information on the branches. */
2978 && (optimize
2979 || !EXPR_HAS_LOCATION (expr)
2980 || !EXPR_HAS_LOCATION (TREE_OPERAND (expr, 1))
2981 || EXPR_LOCATION (expr) == EXPR_LOCATION (TREE_OPERAND (expr, 1))))
2983 label_true = GOTO_DESTINATION (TREE_OPERAND (expr, 1));
2984 have_then_clause_p = true;
2986 else
2987 label_true = create_artificial_label (UNKNOWN_LOCATION);
2988 if (TREE_OPERAND (expr, 2) != NULL
2989 && TREE_CODE (TREE_OPERAND (expr, 2)) == GOTO_EXPR
2990 && TREE_CODE (GOTO_DESTINATION (TREE_OPERAND (expr, 2))) == LABEL_DECL
2991 && (DECL_CONTEXT (GOTO_DESTINATION (TREE_OPERAND (expr, 2)))
2992 == current_function_decl)
2993 /* For -O0 avoid this optimization if the COND_EXPR and GOTO_EXPR
2994 have different locations, otherwise we end up with incorrect
2995 location information on the branches. */
2996 && (optimize
2997 || !EXPR_HAS_LOCATION (expr)
2998 || !EXPR_HAS_LOCATION (TREE_OPERAND (expr, 2))
2999 || EXPR_LOCATION (expr) == EXPR_LOCATION (TREE_OPERAND (expr, 2))))
3001 label_false = GOTO_DESTINATION (TREE_OPERAND (expr, 2));
3002 have_else_clause_p = true;
3004 else
3005 label_false = create_artificial_label (UNKNOWN_LOCATION);
3007 gimple_cond_get_ops_from_tree (COND_EXPR_COND (expr), &pred_code, &arm1,
3008 &arm2);
3010 gimple_cond = gimple_build_cond (pred_code, arm1, arm2, label_true,
3011 label_false);
3013 gimplify_seq_add_stmt (&seq, gimple_cond);
3014 label_cont = NULL_TREE;
3015 if (!have_then_clause_p)
3017 /* For if (...) {} else { code; } put label_true after
3018 the else block. */
3019 if (TREE_OPERAND (expr, 1) == NULL_TREE
3020 && !have_else_clause_p
3021 && TREE_OPERAND (expr, 2) != NULL_TREE)
3022 label_cont = label_true;
3023 else
3025 gimplify_seq_add_stmt (&seq, gimple_build_label (label_true));
3026 have_then_clause_p = gimplify_stmt (&TREE_OPERAND (expr, 1), &seq);
3027 /* For if (...) { code; } else {} or
3028 if (...) { code; } else goto label; or
3029 if (...) { code; return; } else { ... }
3030 label_cont isn't needed. */
3031 if (!have_else_clause_p
3032 && TREE_OPERAND (expr, 2) != NULL_TREE
3033 && gimple_seq_may_fallthru (seq))
3035 gimple g;
3036 label_cont = create_artificial_label (UNKNOWN_LOCATION);
3038 g = gimple_build_goto (label_cont);
3040 /* GIMPLE_COND's are very low level; they have embedded
3041 gotos. This particular embedded goto should not be marked
3042 with the location of the original COND_EXPR, as it would
3043 correspond to the COND_EXPR's condition, not the ELSE or the
3044 THEN arms. To avoid marking it with the wrong location, flag
3045 it as "no location". */
3046 gimple_set_do_not_emit_location (g);
3048 gimplify_seq_add_stmt (&seq, g);
3052 if (!have_else_clause_p)
3054 gimplify_seq_add_stmt (&seq, gimple_build_label (label_false));
3055 have_else_clause_p = gimplify_stmt (&TREE_OPERAND (expr, 2), &seq);
3057 if (label_cont)
3058 gimplify_seq_add_stmt (&seq, gimple_build_label (label_cont));
3060 gimple_pop_condition (pre_p);
3061 gimple_seq_add_seq (pre_p, seq);
3063 if (ret == GS_ERROR)
3064 ; /* Do nothing. */
3065 else if (have_then_clause_p || have_else_clause_p)
3066 ret = GS_ALL_DONE;
3067 else
3069 /* Both arms are empty; replace the COND_EXPR with its predicate. */
3070 expr = TREE_OPERAND (expr, 0);
3071 gimplify_stmt (&expr, pre_p);
3074 *expr_p = NULL;
3075 return ret;
3078 /* Prepare the node pointed to by EXPR_P, an is_gimple_addressable expression,
3079 to be marked addressable.
3081 We cannot rely on such an expression being directly markable if a temporary
3082 has been created by the gimplification. In this case, we create another
3083 temporary and initialize it with a copy, which will become a store after we
3084 mark it addressable. This can happen if the front-end passed us something
3085 that it could not mark addressable yet, like a Fortran pass-by-reference
3086 parameter (int) floatvar. */
3088 static void
3089 prepare_gimple_addressable (tree *expr_p, gimple_seq *seq_p)
3091 while (handled_component_p (*expr_p))
3092 expr_p = &TREE_OPERAND (*expr_p, 0);
3093 if (is_gimple_reg (*expr_p))
3094 *expr_p = get_initialized_tmp_var (*expr_p, seq_p, NULL);
3097 /* A subroutine of gimplify_modify_expr. Replace a MODIFY_EXPR with
3098 a call to __builtin_memcpy. */
3100 static enum gimplify_status
3101 gimplify_modify_expr_to_memcpy (tree *expr_p, tree size, bool want_value,
3102 gimple_seq *seq_p)
3104 tree t, to, to_ptr, from, from_ptr;
3105 gimple gs;
3106 location_t loc = EXPR_LOCATION (*expr_p);
3108 to = TREE_OPERAND (*expr_p, 0);
3109 from = TREE_OPERAND (*expr_p, 1);
3111 /* Mark the RHS addressable. Beware that it may not be possible to do so
3112 directly if a temporary has been created by the gimplification. */
3113 prepare_gimple_addressable (&from, seq_p);
3115 mark_addressable (from);
3116 from_ptr = build_fold_addr_expr_loc (loc, from);
3117 gimplify_arg (&from_ptr, seq_p, loc);
3119 mark_addressable (to);
3120 to_ptr = build_fold_addr_expr_loc (loc, to);
3121 gimplify_arg (&to_ptr, seq_p, loc);
3123 t = implicit_built_in_decls[BUILT_IN_MEMCPY];
3125 gs = gimple_build_call (t, 3, to_ptr, from_ptr, size);
3127 if (want_value)
3129 /* tmp = memcpy() */
3130 t = create_tmp_var (TREE_TYPE (to_ptr), NULL);
3131 gimple_call_set_lhs (gs, t);
3132 gimplify_seq_add_stmt (seq_p, gs);
3134 *expr_p = build1 (INDIRECT_REF, TREE_TYPE (to), t);
3135 return GS_ALL_DONE;
3138 gimplify_seq_add_stmt (seq_p, gs);
3139 *expr_p = NULL;
3140 return GS_ALL_DONE;
3143 /* A subroutine of gimplify_modify_expr. Replace a MODIFY_EXPR with
3144 a call to __builtin_memset. In this case we know that the RHS is
3145 a CONSTRUCTOR with an empty element list. */
3147 static enum gimplify_status
3148 gimplify_modify_expr_to_memset (tree *expr_p, tree size, bool want_value,
3149 gimple_seq *seq_p)
3151 tree t, from, to, to_ptr;
3152 gimple gs;
3153 location_t loc = EXPR_LOCATION (*expr_p);
3155 /* Assert our assumptions, to abort instead of producing wrong code
3156 silently if they are not met. Beware that the RHS CONSTRUCTOR might
3157 not be immediately exposed. */
3158 from = TREE_OPERAND (*expr_p, 1);
3159 if (TREE_CODE (from) == WITH_SIZE_EXPR)
3160 from = TREE_OPERAND (from, 0);
3162 gcc_assert (TREE_CODE (from) == CONSTRUCTOR
3163 && VEC_empty (constructor_elt, CONSTRUCTOR_ELTS (from)));
3165 /* Now proceed. */
3166 to = TREE_OPERAND (*expr_p, 0);
3168 to_ptr = build_fold_addr_expr_loc (loc, to);
3169 gimplify_arg (&to_ptr, seq_p, loc);
3170 t = implicit_built_in_decls[BUILT_IN_MEMSET];
3172 gs = gimple_build_call (t, 3, to_ptr, integer_zero_node, size);
3174 if (want_value)
3176 /* tmp = memset() */
3177 t = create_tmp_var (TREE_TYPE (to_ptr), NULL);
3178 gimple_call_set_lhs (gs, t);
3179 gimplify_seq_add_stmt (seq_p, gs);
3181 *expr_p = build1 (INDIRECT_REF, TREE_TYPE (to), t);
3182 return GS_ALL_DONE;
3185 gimplify_seq_add_stmt (seq_p, gs);
3186 *expr_p = NULL;
3187 return GS_ALL_DONE;
3190 /* A subroutine of gimplify_init_ctor_preeval. Called via walk_tree,
3191 determine, cautiously, if a CONSTRUCTOR overlaps the lhs of an
3192 assignment. Returns non-null if we detect a potential overlap. */
3194 struct gimplify_init_ctor_preeval_data
3196 /* The base decl of the lhs object. May be NULL, in which case we
3197 have to assume the lhs is indirect. */
3198 tree lhs_base_decl;
3200 /* The alias set of the lhs object. */
3201 alias_set_type lhs_alias_set;
3204 static tree
3205 gimplify_init_ctor_preeval_1 (tree *tp, int *walk_subtrees, void *xdata)
3207 struct gimplify_init_ctor_preeval_data *data
3208 = (struct gimplify_init_ctor_preeval_data *) xdata;
3209 tree t = *tp;
3211 /* If we find the base object, obviously we have overlap. */
3212 if (data->lhs_base_decl == t)
3213 return t;
3215 /* If the constructor component is indirect, determine if we have a
3216 potential overlap with the lhs. The only bits of information we
3217 have to go on at this point are addressability and alias sets. */
3218 if (TREE_CODE (t) == INDIRECT_REF
3219 && (!data->lhs_base_decl || TREE_ADDRESSABLE (data->lhs_base_decl))
3220 && alias_sets_conflict_p (data->lhs_alias_set, get_alias_set (t)))
3221 return t;
3223 /* If the constructor component is a call, determine if it can hide a
3224 potential overlap with the lhs through an INDIRECT_REF like above. */
3225 if (TREE_CODE (t) == CALL_EXPR)
3227 tree type, fntype = TREE_TYPE (TREE_TYPE (CALL_EXPR_FN (t)));
3229 for (type = TYPE_ARG_TYPES (fntype); type; type = TREE_CHAIN (type))
3230 if (POINTER_TYPE_P (TREE_VALUE (type))
3231 && (!data->lhs_base_decl || TREE_ADDRESSABLE (data->lhs_base_decl))
3232 && alias_sets_conflict_p (data->lhs_alias_set,
3233 get_alias_set
3234 (TREE_TYPE (TREE_VALUE (type)))))
3235 return t;
3238 if (IS_TYPE_OR_DECL_P (t))
3239 *walk_subtrees = 0;
3240 return NULL;
3243 /* A subroutine of gimplify_init_constructor. Pre-evaluate EXPR,
3244 force values that overlap with the lhs (as described by *DATA)
3245 into temporaries. */
3247 static void
3248 gimplify_init_ctor_preeval (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p,
3249 struct gimplify_init_ctor_preeval_data *data)
3251 enum gimplify_status one;
3253 /* If the value is constant, then there's nothing to pre-evaluate. */
3254 if (TREE_CONSTANT (*expr_p))
3256 /* Ensure it does not have side effects, it might contain a reference to
3257 the object we're initializing. */
3258 gcc_assert (!TREE_SIDE_EFFECTS (*expr_p));
3259 return;
3262 /* If the type has non-trivial constructors, we can't pre-evaluate. */
3263 if (TREE_ADDRESSABLE (TREE_TYPE (*expr_p)))
3264 return;
3266 /* Recurse for nested constructors. */
3267 if (TREE_CODE (*expr_p) == CONSTRUCTOR)
3269 unsigned HOST_WIDE_INT ix;
3270 constructor_elt *ce;
3271 VEC(constructor_elt,gc) *v = CONSTRUCTOR_ELTS (*expr_p);
3273 for (ix = 0; VEC_iterate (constructor_elt, v, ix, ce); ix++)
3274 gimplify_init_ctor_preeval (&ce->value, pre_p, post_p, data);
3276 return;
3279 /* If this is a variable sized type, we must remember the size. */
3280 maybe_with_size_expr (expr_p);
3282 /* Gimplify the constructor element to something appropriate for the rhs
3283 of a MODIFY_EXPR. Given that we know the LHS is an aggregate, we know
3284 the gimplifier will consider this a store to memory. Doing this
3285 gimplification now means that we won't have to deal with complicated
3286 language-specific trees, nor trees like SAVE_EXPR that can induce
3287 exponential search behavior. */
3288 one = gimplify_expr (expr_p, pre_p, post_p, is_gimple_mem_rhs, fb_rvalue);
3289 if (one == GS_ERROR)
3291 *expr_p = NULL;
3292 return;
3295 /* If we gimplified to a bare decl, we can be sure that it doesn't overlap
3296 with the lhs, since "a = { .x=a }" doesn't make sense. This will
3297 always be true for all scalars, since is_gimple_mem_rhs insists on a
3298 temporary variable for them. */
3299 if (DECL_P (*expr_p))
3300 return;
3302 /* If this is of variable size, we have no choice but to assume it doesn't
3303 overlap since we can't make a temporary for it. */
3304 if (TREE_CODE (TYPE_SIZE (TREE_TYPE (*expr_p))) != INTEGER_CST)
3305 return;
3307 /* Otherwise, we must search for overlap ... */
3308 if (!walk_tree (expr_p, gimplify_init_ctor_preeval_1, data, NULL))
3309 return;
3311 /* ... and if found, force the value into a temporary. */
3312 *expr_p = get_formal_tmp_var (*expr_p, pre_p);
3315 /* A subroutine of gimplify_init_ctor_eval. Create a loop for
3316 a RANGE_EXPR in a CONSTRUCTOR for an array.
3318 var = lower;
3319 loop_entry:
3320 object[var] = value;
3321 if (var == upper)
3322 goto loop_exit;
3323 var = var + 1;
3324 goto loop_entry;
3325 loop_exit:
3327 We increment var _after_ the loop exit check because we might otherwise
3328 fail if upper == TYPE_MAX_VALUE (type for upper).
3330 Note that we never have to deal with SAVE_EXPRs here, because this has
3331 already been taken care of for us, in gimplify_init_ctor_preeval(). */
3333 static void gimplify_init_ctor_eval (tree, VEC(constructor_elt,gc) *,
3334 gimple_seq *, bool);
3336 static void
3337 gimplify_init_ctor_eval_range (tree object, tree lower, tree upper,
3338 tree value, tree array_elt_type,
3339 gimple_seq *pre_p, bool cleared)
3341 tree loop_entry_label, loop_exit_label, fall_thru_label;
3342 tree var, var_type, cref, tmp;
3344 loop_entry_label = create_artificial_label (UNKNOWN_LOCATION);
3345 loop_exit_label = create_artificial_label (UNKNOWN_LOCATION);
3346 fall_thru_label = create_artificial_label (UNKNOWN_LOCATION);
3348 /* Create and initialize the index variable. */
3349 var_type = TREE_TYPE (upper);
3350 var = create_tmp_var (var_type, NULL);
3351 gimplify_seq_add_stmt (pre_p, gimple_build_assign (var, lower));
3353 /* Add the loop entry label. */
3354 gimplify_seq_add_stmt (pre_p, gimple_build_label (loop_entry_label));
3356 /* Build the reference. */
3357 cref = build4 (ARRAY_REF, array_elt_type, unshare_expr (object),
3358 var, NULL_TREE, NULL_TREE);
3360 /* If we are a constructor, just call gimplify_init_ctor_eval to do
3361 the store. Otherwise just assign value to the reference. */
3363 if (TREE_CODE (value) == CONSTRUCTOR)
3364 /* NB we might have to call ourself recursively through
3365 gimplify_init_ctor_eval if the value is a constructor. */
3366 gimplify_init_ctor_eval (cref, CONSTRUCTOR_ELTS (value),
3367 pre_p, cleared);
3368 else
3369 gimplify_seq_add_stmt (pre_p, gimple_build_assign (cref, value));
3371 /* We exit the loop when the index var is equal to the upper bound. */
3372 gimplify_seq_add_stmt (pre_p,
3373 gimple_build_cond (EQ_EXPR, var, upper,
3374 loop_exit_label, fall_thru_label));
3376 gimplify_seq_add_stmt (pre_p, gimple_build_label (fall_thru_label));
3378 /* Otherwise, increment the index var... */
3379 tmp = build2 (PLUS_EXPR, var_type, var,
3380 fold_convert (var_type, integer_one_node));
3381 gimplify_seq_add_stmt (pre_p, gimple_build_assign (var, tmp));
3383 /* ...and jump back to the loop entry. */
3384 gimplify_seq_add_stmt (pre_p, gimple_build_goto (loop_entry_label));
3386 /* Add the loop exit label. */
3387 gimplify_seq_add_stmt (pre_p, gimple_build_label (loop_exit_label));
3390 /* Return true if FDECL is accessing a field that is zero sized. */
3392 static bool
3393 zero_sized_field_decl (const_tree fdecl)
3395 if (TREE_CODE (fdecl) == FIELD_DECL && DECL_SIZE (fdecl)
3396 && integer_zerop (DECL_SIZE (fdecl)))
3397 return true;
3398 return false;
3401 /* Return true if TYPE is zero sized. */
3403 static bool
3404 zero_sized_type (const_tree type)
3406 if (AGGREGATE_TYPE_P (type) && TYPE_SIZE (type)
3407 && integer_zerop (TYPE_SIZE (type)))
3408 return true;
3409 return false;
3412 /* A subroutine of gimplify_init_constructor. Generate individual
3413 MODIFY_EXPRs for a CONSTRUCTOR. OBJECT is the LHS against which the
3414 assignments should happen. ELTS is the CONSTRUCTOR_ELTS of the
3415 CONSTRUCTOR. CLEARED is true if the entire LHS object has been
3416 zeroed first. */
3418 static void
3419 gimplify_init_ctor_eval (tree object, VEC(constructor_elt,gc) *elts,
3420 gimple_seq *pre_p, bool cleared)
3422 tree array_elt_type = NULL;
3423 unsigned HOST_WIDE_INT ix;
3424 tree purpose, value;
3426 if (TREE_CODE (TREE_TYPE (object)) == ARRAY_TYPE)
3427 array_elt_type = TYPE_MAIN_VARIANT (TREE_TYPE (TREE_TYPE (object)));
3429 FOR_EACH_CONSTRUCTOR_ELT (elts, ix, purpose, value)
3431 tree cref;
3433 /* NULL values are created above for gimplification errors. */
3434 if (value == NULL)
3435 continue;
3437 if (cleared && initializer_zerop (value))
3438 continue;
3440 /* ??? Here's to hoping the front end fills in all of the indices,
3441 so we don't have to figure out what's missing ourselves. */
3442 gcc_assert (purpose);
3444 /* Skip zero-sized fields, unless value has side-effects. This can
3445 happen with calls to functions returning a zero-sized type, which
3446 we shouldn't discard. As a number of downstream passes don't
3447 expect sets of zero-sized fields, we rely on the gimplification of
3448 the MODIFY_EXPR we make below to drop the assignment statement. */
3449 if (! TREE_SIDE_EFFECTS (value) && zero_sized_field_decl (purpose))
3450 continue;
3452 /* If we have a RANGE_EXPR, we have to build a loop to assign the
3453 whole range. */
3454 if (TREE_CODE (purpose) == RANGE_EXPR)
3456 tree lower = TREE_OPERAND (purpose, 0);
3457 tree upper = TREE_OPERAND (purpose, 1);
3459 /* If the lower bound is equal to upper, just treat it as if
3460 upper was the index. */
3461 if (simple_cst_equal (lower, upper))
3462 purpose = upper;
3463 else
3465 gimplify_init_ctor_eval_range (object, lower, upper, value,
3466 array_elt_type, pre_p, cleared);
3467 continue;
3471 if (array_elt_type)
3473 /* Do not use bitsizetype for ARRAY_REF indices. */
3474 if (TYPE_DOMAIN (TREE_TYPE (object)))
3475 purpose = fold_convert (TREE_TYPE (TYPE_DOMAIN (TREE_TYPE (object))),
3476 purpose);
3477 cref = build4 (ARRAY_REF, array_elt_type, unshare_expr (object),
3478 purpose, NULL_TREE, NULL_TREE);
3480 else
3482 gcc_assert (TREE_CODE (purpose) == FIELD_DECL);
3483 cref = build3 (COMPONENT_REF, TREE_TYPE (purpose),
3484 unshare_expr (object), purpose, NULL_TREE);
3487 if (TREE_CODE (value) == CONSTRUCTOR
3488 && TREE_CODE (TREE_TYPE (value)) != VECTOR_TYPE)
3489 gimplify_init_ctor_eval (cref, CONSTRUCTOR_ELTS (value),
3490 pre_p, cleared);
3491 else
3493 tree init = build2 (INIT_EXPR, TREE_TYPE (cref), cref, value);
3494 gimplify_and_add (init, pre_p);
3495 ggc_free (init);
3501 /* Returns the appropriate RHS predicate for this LHS. */
3503 gimple_predicate
3504 rhs_predicate_for (tree lhs)
3506 if (is_gimple_reg (lhs))
3507 return is_gimple_reg_rhs_or_call;
3508 else
3509 return is_gimple_mem_rhs_or_call;
3512 /* Gimplify a C99 compound literal expression. This just means adding
3513 the DECL_EXPR before the current statement and using its anonymous
3514 decl instead. */
3516 static enum gimplify_status
3517 gimplify_compound_literal_expr (tree *expr_p, gimple_seq *pre_p)
3519 tree decl_s = COMPOUND_LITERAL_EXPR_DECL_EXPR (*expr_p);
3520 tree decl = DECL_EXPR_DECL (decl_s);
3521 /* Mark the decl as addressable if the compound literal
3522 expression is addressable now, otherwise it is marked too late
3523 after we gimplify the initialization expression. */
3524 if (TREE_ADDRESSABLE (*expr_p))
3525 TREE_ADDRESSABLE (decl) = 1;
3527 /* Preliminarily mark non-addressed complex variables as eligible
3528 for promotion to gimple registers. We'll transform their uses
3529 as we find them. */
3530 if ((TREE_CODE (TREE_TYPE (decl)) == COMPLEX_TYPE
3531 || TREE_CODE (TREE_TYPE (decl)) == VECTOR_TYPE)
3532 && !TREE_THIS_VOLATILE (decl)
3533 && !needs_to_live_in_memory (decl))
3534 DECL_GIMPLE_REG_P (decl) = 1;
3536 /* This decl isn't mentioned in the enclosing block, so add it to the
3537 list of temps. FIXME it seems a bit of a kludge to say that
3538 anonymous artificial vars aren't pushed, but everything else is. */
3539 if (DECL_NAME (decl) == NULL_TREE && !DECL_SEEN_IN_BIND_EXPR_P (decl))
3540 gimple_add_tmp_var (decl);
3542 gimplify_and_add (decl_s, pre_p);
3543 *expr_p = decl;
3544 return GS_OK;
3547 /* Optimize embedded COMPOUND_LITERAL_EXPRs within a CONSTRUCTOR,
3548 return a new CONSTRUCTOR if something changed. */
3550 static tree
3551 optimize_compound_literals_in_ctor (tree orig_ctor)
3553 tree ctor = orig_ctor;
3554 VEC(constructor_elt,gc) *elts = CONSTRUCTOR_ELTS (ctor);
3555 unsigned int idx, num = VEC_length (constructor_elt, elts);
3557 for (idx = 0; idx < num; idx++)
3559 tree value = VEC_index (constructor_elt, elts, idx)->value;
3560 tree newval = value;
3561 if (TREE_CODE (value) == CONSTRUCTOR)
3562 newval = optimize_compound_literals_in_ctor (value);
3563 else if (TREE_CODE (value) == COMPOUND_LITERAL_EXPR)
3565 tree decl_s = COMPOUND_LITERAL_EXPR_DECL_EXPR (value);
3566 tree decl = DECL_EXPR_DECL (decl_s);
3567 tree init = DECL_INITIAL (decl);
3569 if (!TREE_ADDRESSABLE (value)
3570 && !TREE_ADDRESSABLE (decl)
3571 && init)
3572 newval = optimize_compound_literals_in_ctor (init);
3574 if (newval == value)
3575 continue;
3577 if (ctor == orig_ctor)
3579 ctor = copy_node (orig_ctor);
3580 CONSTRUCTOR_ELTS (ctor) = VEC_copy (constructor_elt, gc, elts);
3581 elts = CONSTRUCTOR_ELTS (ctor);
3583 VEC_index (constructor_elt, elts, idx)->value = newval;
3585 return ctor;
3590 /* A subroutine of gimplify_modify_expr. Break out elements of a
3591 CONSTRUCTOR used as an initializer into separate MODIFY_EXPRs.
3593 Note that we still need to clear any elements that don't have explicit
3594 initializers, so if not all elements are initialized we keep the
3595 original MODIFY_EXPR, we just remove all of the constructor elements.
3597 If NOTIFY_TEMP_CREATION is true, do not gimplify, just return
3598 GS_ERROR if we would have to create a temporary when gimplifying
3599 this constructor. Otherwise, return GS_OK.
3601 If NOTIFY_TEMP_CREATION is false, just do the gimplification. */
3603 static enum gimplify_status
3604 gimplify_init_constructor (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p,
3605 bool want_value, bool notify_temp_creation)
3607 tree object, ctor, type;
3608 enum gimplify_status ret;
3609 VEC(constructor_elt,gc) *elts;
3611 gcc_assert (TREE_CODE (TREE_OPERAND (*expr_p, 1)) == CONSTRUCTOR);
3613 if (!notify_temp_creation)
3615 ret = gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p, post_p,
3616 is_gimple_lvalue, fb_lvalue);
3617 if (ret == GS_ERROR)
3618 return ret;
3621 object = TREE_OPERAND (*expr_p, 0);
3622 ctor = TREE_OPERAND (*expr_p, 1) =
3623 optimize_compound_literals_in_ctor (TREE_OPERAND (*expr_p, 1));
3624 type = TREE_TYPE (ctor);
3625 elts = CONSTRUCTOR_ELTS (ctor);
3626 ret = GS_ALL_DONE;
3628 switch (TREE_CODE (type))
3630 case RECORD_TYPE:
3631 case UNION_TYPE:
3632 case QUAL_UNION_TYPE:
3633 case ARRAY_TYPE:
3635 struct gimplify_init_ctor_preeval_data preeval_data;
3636 HOST_WIDE_INT num_type_elements, num_ctor_elements;
3637 HOST_WIDE_INT num_nonzero_elements;
3638 bool cleared, valid_const_initializer;
3640 /* Aggregate types must lower constructors to initialization of
3641 individual elements. The exception is that a CONSTRUCTOR node
3642 with no elements indicates zero-initialization of the whole. */
3643 if (VEC_empty (constructor_elt, elts))
3645 if (notify_temp_creation)
3646 return GS_OK;
3647 break;
3650 /* Fetch information about the constructor to direct later processing.
3651 We might want to make static versions of it in various cases, and
3652 can only do so if it known to be a valid constant initializer. */
3653 valid_const_initializer
3654 = categorize_ctor_elements (ctor, &num_nonzero_elements,
3655 &num_ctor_elements, &cleared);
3657 /* If a const aggregate variable is being initialized, then it
3658 should never be a lose to promote the variable to be static. */
3659 if (valid_const_initializer
3660 && num_nonzero_elements > 1
3661 && TREE_READONLY (object)
3662 && TREE_CODE (object) == VAR_DECL
3663 && (flag_merge_constants >= 2 || !TREE_ADDRESSABLE (object)))
3665 if (notify_temp_creation)
3666 return GS_ERROR;
3667 DECL_INITIAL (object) = ctor;
3668 TREE_STATIC (object) = 1;
3669 if (!DECL_NAME (object))
3670 DECL_NAME (object) = create_tmp_var_name ("C");
3671 walk_tree (&DECL_INITIAL (object), force_labels_r, NULL, NULL);
3673 /* ??? C++ doesn't automatically append a .<number> to the
3674 assembler name, and even when it does, it looks a FE private
3675 data structures to figure out what that number should be,
3676 which are not set for this variable. I suppose this is
3677 important for local statics for inline functions, which aren't
3678 "local" in the object file sense. So in order to get a unique
3679 TU-local symbol, we must invoke the lhd version now. */
3680 lhd_set_decl_assembler_name (object);
3682 *expr_p = NULL_TREE;
3683 break;
3686 /* If there are "lots" of initialized elements, even discounting
3687 those that are not address constants (and thus *must* be
3688 computed at runtime), then partition the constructor into
3689 constant and non-constant parts. Block copy the constant
3690 parts in, then generate code for the non-constant parts. */
3691 /* TODO. There's code in cp/typeck.c to do this. */
3693 num_type_elements = count_type_elements (type, true);
3695 /* If count_type_elements could not determine number of type elements
3696 for a constant-sized object, assume clearing is needed.
3697 Don't do this for variable-sized objects, as store_constructor
3698 will ignore the clearing of variable-sized objects. */
3699 if (num_type_elements < 0 && int_size_in_bytes (type) >= 0)
3700 cleared = true;
3701 /* If there are "lots" of zeros, then block clear the object first. */
3702 else if (num_type_elements - num_nonzero_elements
3703 > CLEAR_RATIO (optimize_function_for_speed_p (cfun))
3704 && num_nonzero_elements < num_type_elements/4)
3705 cleared = true;
3706 /* ??? This bit ought not be needed. For any element not present
3707 in the initializer, we should simply set them to zero. Except
3708 we'd need to *find* the elements that are not present, and that
3709 requires trickery to avoid quadratic compile-time behavior in
3710 large cases or excessive memory use in small cases. */
3711 else if (num_ctor_elements < num_type_elements)
3712 cleared = true;
3714 /* If there are "lots" of initialized elements, and all of them
3715 are valid address constants, then the entire initializer can
3716 be dropped to memory, and then memcpy'd out. Don't do this
3717 for sparse arrays, though, as it's more efficient to follow
3718 the standard CONSTRUCTOR behavior of memset followed by
3719 individual element initialization. Also don't do this for small
3720 all-zero initializers (which aren't big enough to merit
3721 clearing), and don't try to make bitwise copies of
3722 TREE_ADDRESSABLE types. */
3723 if (valid_const_initializer
3724 && !(cleared || num_nonzero_elements == 0)
3725 && !TREE_ADDRESSABLE (type))
3727 HOST_WIDE_INT size = int_size_in_bytes (type);
3728 unsigned int align;
3730 /* ??? We can still get unbounded array types, at least
3731 from the C++ front end. This seems wrong, but attempt
3732 to work around it for now. */
3733 if (size < 0)
3735 size = int_size_in_bytes (TREE_TYPE (object));
3736 if (size >= 0)
3737 TREE_TYPE (ctor) = type = TREE_TYPE (object);
3740 /* Find the maximum alignment we can assume for the object. */
3741 /* ??? Make use of DECL_OFFSET_ALIGN. */
3742 if (DECL_P (object))
3743 align = DECL_ALIGN (object);
3744 else
3745 align = TYPE_ALIGN (type);
3747 if (size > 0
3748 && num_nonzero_elements > 1
3749 && !can_move_by_pieces (size, align))
3751 tree new_tree;
3753 if (notify_temp_creation)
3754 return GS_ERROR;
3756 new_tree = create_tmp_var_raw (type, "C");
3758 gimple_add_tmp_var (new_tree);
3759 TREE_STATIC (new_tree) = 1;
3760 TREE_READONLY (new_tree) = 1;
3761 DECL_INITIAL (new_tree) = ctor;
3762 if (align > DECL_ALIGN (new_tree))
3764 DECL_ALIGN (new_tree) = align;
3765 DECL_USER_ALIGN (new_tree) = 1;
3767 walk_tree (&DECL_INITIAL (new_tree), force_labels_r, NULL, NULL);
3769 TREE_OPERAND (*expr_p, 1) = new_tree;
3771 /* This is no longer an assignment of a CONSTRUCTOR, but
3772 we still may have processing to do on the LHS. So
3773 pretend we didn't do anything here to let that happen. */
3774 return GS_UNHANDLED;
3778 /* If the target is volatile and we have non-zero elements
3779 initialize the target from a temporary. */
3780 if (TREE_THIS_VOLATILE (object)
3781 && !TREE_ADDRESSABLE (type)
3782 && num_nonzero_elements > 0)
3784 tree temp = create_tmp_var (TYPE_MAIN_VARIANT (type), NULL);
3785 TREE_OPERAND (*expr_p, 0) = temp;
3786 *expr_p = build2 (COMPOUND_EXPR, TREE_TYPE (*expr_p),
3787 *expr_p,
3788 build2 (MODIFY_EXPR, void_type_node,
3789 object, temp));
3790 return GS_OK;
3793 if (notify_temp_creation)
3794 return GS_OK;
3796 /* If there are nonzero elements, pre-evaluate to capture elements
3797 overlapping with the lhs into temporaries. We must do this before
3798 clearing to fetch the values before they are zeroed-out. */
3799 if (num_nonzero_elements > 0)
3801 preeval_data.lhs_base_decl = get_base_address (object);
3802 if (!DECL_P (preeval_data.lhs_base_decl))
3803 preeval_data.lhs_base_decl = NULL;
3804 preeval_data.lhs_alias_set = get_alias_set (object);
3806 gimplify_init_ctor_preeval (&TREE_OPERAND (*expr_p, 1),
3807 pre_p, post_p, &preeval_data);
3810 if (cleared)
3812 /* Zap the CONSTRUCTOR element list, which simplifies this case.
3813 Note that we still have to gimplify, in order to handle the
3814 case of variable sized types. Avoid shared tree structures. */
3815 CONSTRUCTOR_ELTS (ctor) = NULL;
3816 TREE_SIDE_EFFECTS (ctor) = 0;
3817 object = unshare_expr (object);
3818 gimplify_stmt (expr_p, pre_p);
3821 /* If we have not block cleared the object, or if there are nonzero
3822 elements in the constructor, add assignments to the individual
3823 scalar fields of the object. */
3824 if (!cleared || num_nonzero_elements > 0)
3825 gimplify_init_ctor_eval (object, elts, pre_p, cleared);
3827 *expr_p = NULL_TREE;
3829 break;
3831 case COMPLEX_TYPE:
3833 tree r, i;
3835 if (notify_temp_creation)
3836 return GS_OK;
3838 /* Extract the real and imaginary parts out of the ctor. */
3839 gcc_assert (VEC_length (constructor_elt, elts) == 2);
3840 r = VEC_index (constructor_elt, elts, 0)->value;
3841 i = VEC_index (constructor_elt, elts, 1)->value;
3842 if (r == NULL || i == NULL)
3844 tree zero = fold_convert (TREE_TYPE (type), integer_zero_node);
3845 if (r == NULL)
3846 r = zero;
3847 if (i == NULL)
3848 i = zero;
3851 /* Complex types have either COMPLEX_CST or COMPLEX_EXPR to
3852 represent creation of a complex value. */
3853 if (TREE_CONSTANT (r) && TREE_CONSTANT (i))
3855 ctor = build_complex (type, r, i);
3856 TREE_OPERAND (*expr_p, 1) = ctor;
3858 else
3860 ctor = build2 (COMPLEX_EXPR, type, r, i);
3861 TREE_OPERAND (*expr_p, 1) = ctor;
3862 ret = gimplify_expr (&TREE_OPERAND (*expr_p, 1),
3863 pre_p,
3864 post_p,
3865 rhs_predicate_for (TREE_OPERAND (*expr_p, 0)),
3866 fb_rvalue);
3869 break;
3871 case VECTOR_TYPE:
3873 unsigned HOST_WIDE_INT ix;
3874 constructor_elt *ce;
3876 if (notify_temp_creation)
3877 return GS_OK;
3879 /* Go ahead and simplify constant constructors to VECTOR_CST. */
3880 if (TREE_CONSTANT (ctor))
3882 bool constant_p = true;
3883 tree value;
3885 /* Even when ctor is constant, it might contain non-*_CST
3886 elements, such as addresses or trapping values like
3887 1.0/0.0 - 1.0/0.0. Such expressions don't belong
3888 in VECTOR_CST nodes. */
3889 FOR_EACH_CONSTRUCTOR_VALUE (elts, ix, value)
3890 if (!CONSTANT_CLASS_P (value))
3892 constant_p = false;
3893 break;
3896 if (constant_p)
3898 TREE_OPERAND (*expr_p, 1) = build_vector_from_ctor (type, elts);
3899 break;
3902 /* Don't reduce an initializer constant even if we can't
3903 make a VECTOR_CST. It won't do anything for us, and it'll
3904 prevent us from representing it as a single constant. */
3905 if (initializer_constant_valid_p (ctor, type))
3906 break;
3908 TREE_CONSTANT (ctor) = 0;
3911 /* Vector types use CONSTRUCTOR all the way through gimple
3912 compilation as a general initializer. */
3913 for (ix = 0; VEC_iterate (constructor_elt, elts, ix, ce); ix++)
3915 enum gimplify_status tret;
3916 tret = gimplify_expr (&ce->value, pre_p, post_p, is_gimple_val,
3917 fb_rvalue);
3918 if (tret == GS_ERROR)
3919 ret = GS_ERROR;
3921 if (!is_gimple_reg (TREE_OPERAND (*expr_p, 0)))
3922 TREE_OPERAND (*expr_p, 1) = get_formal_tmp_var (ctor, pre_p);
3924 break;
3926 default:
3927 /* So how did we get a CONSTRUCTOR for a scalar type? */
3928 gcc_unreachable ();
3931 if (ret == GS_ERROR)
3932 return GS_ERROR;
3933 else if (want_value)
3935 *expr_p = object;
3936 return GS_OK;
3938 else
3940 /* If we have gimplified both sides of the initializer but have
3941 not emitted an assignment, do so now. */
3942 if (*expr_p)
3944 tree lhs = TREE_OPERAND (*expr_p, 0);
3945 tree rhs = TREE_OPERAND (*expr_p, 1);
3946 gimple init = gimple_build_assign (lhs, rhs);
3947 gimplify_seq_add_stmt (pre_p, init);
3948 *expr_p = NULL;
3951 return GS_ALL_DONE;
3955 /* Given a pointer value OP0, return a simplified version of an
3956 indirection through OP0, or NULL_TREE if no simplification is
3957 possible. Note that the resulting type may be different from
3958 the type pointed to in the sense that it is still compatible
3959 from the langhooks point of view. */
3961 tree
3962 gimple_fold_indirect_ref (tree t)
3964 tree type = TREE_TYPE (TREE_TYPE (t));
3965 tree sub = t;
3966 tree subtype;
3968 STRIP_NOPS (sub);
3969 subtype = TREE_TYPE (sub);
3970 if (!POINTER_TYPE_P (subtype))
3971 return NULL_TREE;
3973 if (TREE_CODE (sub) == ADDR_EXPR)
3975 tree op = TREE_OPERAND (sub, 0);
3976 tree optype = TREE_TYPE (op);
3977 /* *&p => p */
3978 if (useless_type_conversion_p (type, optype))
3979 return op;
3981 /* *(foo *)&fooarray => fooarray[0] */
3982 if (TREE_CODE (optype) == ARRAY_TYPE
3983 && TREE_CODE (TYPE_SIZE (TREE_TYPE (optype))) == INTEGER_CST
3984 && useless_type_conversion_p (type, TREE_TYPE (optype)))
3986 tree type_domain = TYPE_DOMAIN (optype);
3987 tree min_val = size_zero_node;
3988 if (type_domain && TYPE_MIN_VALUE (type_domain))
3989 min_val = TYPE_MIN_VALUE (type_domain);
3990 if (TREE_CODE (min_val) == INTEGER_CST)
3991 return build4 (ARRAY_REF, type, op, min_val, NULL_TREE, NULL_TREE);
3993 /* *(foo *)&complexfoo => __real__ complexfoo */
3994 else if (TREE_CODE (optype) == COMPLEX_TYPE
3995 && useless_type_conversion_p (type, TREE_TYPE (optype)))
3996 return fold_build1 (REALPART_EXPR, type, op);
3997 /* *(foo *)&vectorfoo => BIT_FIELD_REF<vectorfoo,...> */
3998 else if (TREE_CODE (optype) == VECTOR_TYPE
3999 && useless_type_conversion_p (type, TREE_TYPE (optype)))
4001 tree part_width = TYPE_SIZE (type);
4002 tree index = bitsize_int (0);
4003 return fold_build3 (BIT_FIELD_REF, type, op, part_width, index);
4007 /* ((foo*)&vectorfoo)[1] => BIT_FIELD_REF<vectorfoo,...> */
4008 if (TREE_CODE (sub) == POINTER_PLUS_EXPR
4009 && TREE_CODE (TREE_OPERAND (sub, 1)) == INTEGER_CST)
4011 tree op00 = TREE_OPERAND (sub, 0);
4012 tree op01 = TREE_OPERAND (sub, 1);
4013 tree op00type;
4015 STRIP_NOPS (op00);
4016 op00type = TREE_TYPE (op00);
4017 if (TREE_CODE (op00) == ADDR_EXPR
4018 && TREE_CODE (TREE_TYPE (op00type)) == VECTOR_TYPE
4019 && useless_type_conversion_p (type, TREE_TYPE (TREE_TYPE (op00type))))
4021 HOST_WIDE_INT offset = tree_low_cst (op01, 0);
4022 tree part_width = TYPE_SIZE (type);
4023 unsigned HOST_WIDE_INT part_widthi
4024 = tree_low_cst (part_width, 0) / BITS_PER_UNIT;
4025 unsigned HOST_WIDE_INT indexi = offset * BITS_PER_UNIT;
4026 tree index = bitsize_int (indexi);
4027 if (offset / part_widthi
4028 <= TYPE_VECTOR_SUBPARTS (TREE_TYPE (op00type)))
4029 return fold_build3 (BIT_FIELD_REF, type, TREE_OPERAND (op00, 0),
4030 part_width, index);
4034 /* ((foo*)&complexfoo)[1] => __imag__ complexfoo */
4035 if (TREE_CODE (sub) == POINTER_PLUS_EXPR
4036 && TREE_CODE (TREE_OPERAND (sub, 1)) == INTEGER_CST)
4038 tree op00 = TREE_OPERAND (sub, 0);
4039 tree op01 = TREE_OPERAND (sub, 1);
4040 tree op00type;
4042 STRIP_NOPS (op00);
4043 op00type = TREE_TYPE (op00);
4044 if (TREE_CODE (op00) == ADDR_EXPR
4045 && TREE_CODE (TREE_TYPE (op00type)) == COMPLEX_TYPE
4046 && useless_type_conversion_p (type, TREE_TYPE (TREE_TYPE (op00type))))
4048 tree size = TYPE_SIZE_UNIT (type);
4049 if (tree_int_cst_equal (size, op01))
4050 return fold_build1 (IMAGPART_EXPR, type, TREE_OPERAND (op00, 0));
4054 /* *(foo *)fooarrptr => (*fooarrptr)[0] */
4055 if (TREE_CODE (TREE_TYPE (subtype)) == ARRAY_TYPE
4056 && TREE_CODE (TYPE_SIZE (TREE_TYPE (TREE_TYPE (subtype)))) == INTEGER_CST
4057 && useless_type_conversion_p (type, TREE_TYPE (TREE_TYPE (subtype))))
4059 tree type_domain;
4060 tree min_val = size_zero_node;
4061 tree osub = sub;
4062 sub = gimple_fold_indirect_ref (sub);
4063 if (! sub)
4064 sub = build1 (INDIRECT_REF, TREE_TYPE (subtype), osub);
4065 type_domain = TYPE_DOMAIN (TREE_TYPE (sub));
4066 if (type_domain && TYPE_MIN_VALUE (type_domain))
4067 min_val = TYPE_MIN_VALUE (type_domain);
4068 if (TREE_CODE (min_val) == INTEGER_CST)
4069 return build4 (ARRAY_REF, type, sub, min_val, NULL_TREE, NULL_TREE);
4072 return NULL_TREE;
4075 /* Given a pointer value OP0, return a simplified version of an
4076 indirection through OP0, or NULL_TREE if no simplification is
4077 possible. This may only be applied to a rhs of an expression.
4078 Note that the resulting type may be different from the type pointed
4079 to in the sense that it is still compatible from the langhooks
4080 point of view. */
4082 static tree
4083 gimple_fold_indirect_ref_rhs (tree t)
4085 return gimple_fold_indirect_ref (t);
4088 /* Subroutine of gimplify_modify_expr to do simplifications of
4089 MODIFY_EXPRs based on the code of the RHS. We loop for as long as
4090 something changes. */
4092 static enum gimplify_status
4093 gimplify_modify_expr_rhs (tree *expr_p, tree *from_p, tree *to_p,
4094 gimple_seq *pre_p, gimple_seq *post_p,
4095 bool want_value)
4097 enum gimplify_status ret = GS_OK;
4099 while (ret != GS_UNHANDLED)
4100 switch (TREE_CODE (*from_p))
4102 case VAR_DECL:
4103 /* If we're assigning from a read-only variable initialized with
4104 a constructor, do the direct assignment from the constructor,
4105 but only if neither source nor target are volatile since this
4106 latter assignment might end up being done on a per-field basis. */
4107 if (DECL_INITIAL (*from_p)
4108 && TREE_READONLY (*from_p)
4109 && !TREE_THIS_VOLATILE (*from_p)
4110 && !TREE_THIS_VOLATILE (*to_p)
4111 && TREE_CODE (DECL_INITIAL (*from_p)) == CONSTRUCTOR)
4113 tree old_from = *from_p;
4115 /* Move the constructor into the RHS. */
4116 *from_p = unshare_expr (DECL_INITIAL (*from_p));
4118 /* Let's see if gimplify_init_constructor will need to put
4119 it in memory. If so, revert the change. */
4120 ret = gimplify_init_constructor (expr_p, NULL, NULL, false, true);
4121 if (ret == GS_ERROR)
4123 *from_p = old_from;
4124 /* Fall through. */
4126 else
4128 ret = GS_OK;
4129 break;
4132 ret = GS_UNHANDLED;
4133 break;
4134 case INDIRECT_REF:
4136 /* If we have code like
4138 *(const A*)(A*)&x
4140 where the type of "x" is a (possibly cv-qualified variant
4141 of "A"), treat the entire expression as identical to "x".
4142 This kind of code arises in C++ when an object is bound
4143 to a const reference, and if "x" is a TARGET_EXPR we want
4144 to take advantage of the optimization below. */
4145 tree t = gimple_fold_indirect_ref_rhs (TREE_OPERAND (*from_p, 0));
4146 if (t)
4148 *from_p = t;
4149 ret = GS_OK;
4151 else
4152 ret = GS_UNHANDLED;
4153 break;
4156 case TARGET_EXPR:
4158 /* If we are initializing something from a TARGET_EXPR, strip the
4159 TARGET_EXPR and initialize it directly, if possible. This can't
4160 be done if the initializer is void, since that implies that the
4161 temporary is set in some non-trivial way.
4163 ??? What about code that pulls out the temp and uses it
4164 elsewhere? I think that such code never uses the TARGET_EXPR as
4165 an initializer. If I'm wrong, we'll die because the temp won't
4166 have any RTL. In that case, I guess we'll need to replace
4167 references somehow. */
4168 tree init = TARGET_EXPR_INITIAL (*from_p);
4170 if (init
4171 && !VOID_TYPE_P (TREE_TYPE (init)))
4173 *from_p = init;
4174 ret = GS_OK;
4176 else
4177 ret = GS_UNHANDLED;
4179 break;
4181 case COMPOUND_EXPR:
4182 /* Remove any COMPOUND_EXPR in the RHS so the following cases will be
4183 caught. */
4184 gimplify_compound_expr (from_p, pre_p, true);
4185 ret = GS_OK;
4186 break;
4188 case CONSTRUCTOR:
4189 /* If we're initializing from a CONSTRUCTOR, break this into
4190 individual MODIFY_EXPRs. */
4191 return gimplify_init_constructor (expr_p, pre_p, post_p, want_value,
4192 false);
4194 case COND_EXPR:
4195 /* If we're assigning to a non-register type, push the assignment
4196 down into the branches. This is mandatory for ADDRESSABLE types,
4197 since we cannot generate temporaries for such, but it saves a
4198 copy in other cases as well. */
4199 if (!is_gimple_reg_type (TREE_TYPE (*from_p)))
4201 /* This code should mirror the code in gimplify_cond_expr. */
4202 enum tree_code code = TREE_CODE (*expr_p);
4203 tree cond = *from_p;
4204 tree result = *to_p;
4206 ret = gimplify_expr (&result, pre_p, post_p,
4207 is_gimple_lvalue, fb_lvalue);
4208 if (ret != GS_ERROR)
4209 ret = GS_OK;
4211 if (TREE_TYPE (TREE_OPERAND (cond, 1)) != void_type_node)
4212 TREE_OPERAND (cond, 1)
4213 = build2 (code, void_type_node, result,
4214 TREE_OPERAND (cond, 1));
4215 if (TREE_TYPE (TREE_OPERAND (cond, 2)) != void_type_node)
4216 TREE_OPERAND (cond, 2)
4217 = build2 (code, void_type_node, unshare_expr (result),
4218 TREE_OPERAND (cond, 2));
4220 TREE_TYPE (cond) = void_type_node;
4221 recalculate_side_effects (cond);
4223 if (want_value)
4225 gimplify_and_add (cond, pre_p);
4226 *expr_p = unshare_expr (result);
4228 else
4229 *expr_p = cond;
4230 return ret;
4232 else
4233 ret = GS_UNHANDLED;
4234 break;
4236 case CALL_EXPR:
4237 /* For calls that return in memory, give *to_p as the CALL_EXPR's
4238 return slot so that we don't generate a temporary. */
4239 if (!CALL_EXPR_RETURN_SLOT_OPT (*from_p)
4240 && aggregate_value_p (*from_p, *from_p))
4242 bool use_target;
4244 if (!(rhs_predicate_for (*to_p))(*from_p))
4245 /* If we need a temporary, *to_p isn't accurate. */
4246 use_target = false;
4247 else if (TREE_CODE (*to_p) == RESULT_DECL
4248 && DECL_NAME (*to_p) == NULL_TREE
4249 && needs_to_live_in_memory (*to_p))
4250 /* It's OK to use the return slot directly unless it's an NRV. */
4251 use_target = true;
4252 else if (is_gimple_reg_type (TREE_TYPE (*to_p))
4253 || (DECL_P (*to_p) && DECL_REGISTER (*to_p)))
4254 /* Don't force regs into memory. */
4255 use_target = false;
4256 else if (TREE_CODE (*expr_p) == INIT_EXPR)
4257 /* It's OK to use the target directly if it's being
4258 initialized. */
4259 use_target = true;
4260 else if (!is_gimple_non_addressable (*to_p))
4261 /* Don't use the original target if it's already addressable;
4262 if its address escapes, and the called function uses the
4263 NRV optimization, a conforming program could see *to_p
4264 change before the called function returns; see c++/19317.
4265 When optimizing, the return_slot pass marks more functions
4266 as safe after we have escape info. */
4267 use_target = false;
4268 else
4269 use_target = true;
4271 if (use_target)
4273 CALL_EXPR_RETURN_SLOT_OPT (*from_p) = 1;
4274 mark_addressable (*to_p);
4278 ret = GS_UNHANDLED;
4279 break;
4281 /* If we're initializing from a container, push the initialization
4282 inside it. */
4283 case CLEANUP_POINT_EXPR:
4284 case BIND_EXPR:
4285 case STATEMENT_LIST:
4287 tree wrap = *from_p;
4288 tree t;
4290 ret = gimplify_expr (to_p, pre_p, post_p, is_gimple_min_lval,
4291 fb_lvalue);
4292 if (ret != GS_ERROR)
4293 ret = GS_OK;
4295 t = voidify_wrapper_expr (wrap, *expr_p);
4296 gcc_assert (t == *expr_p);
4298 if (want_value)
4300 gimplify_and_add (wrap, pre_p);
4301 *expr_p = unshare_expr (*to_p);
4303 else
4304 *expr_p = wrap;
4305 return GS_OK;
4308 case COMPOUND_LITERAL_EXPR:
4310 tree complit = TREE_OPERAND (*expr_p, 1);
4311 tree decl_s = COMPOUND_LITERAL_EXPR_DECL_EXPR (complit);
4312 tree decl = DECL_EXPR_DECL (decl_s);
4313 tree init = DECL_INITIAL (decl);
4315 /* struct T x = (struct T) { 0, 1, 2 } can be optimized
4316 into struct T x = { 0, 1, 2 } if the address of the
4317 compound literal has never been taken. */
4318 if (!TREE_ADDRESSABLE (complit)
4319 && !TREE_ADDRESSABLE (decl)
4320 && init)
4322 *expr_p = copy_node (*expr_p);
4323 TREE_OPERAND (*expr_p, 1) = init;
4324 return GS_OK;
4328 default:
4329 ret = GS_UNHANDLED;
4330 break;
4333 return ret;
4337 /* Promote partial stores to COMPLEX variables to total stores. *EXPR_P is
4338 a MODIFY_EXPR with a lhs of a REAL/IMAGPART_EXPR of a variable with
4339 DECL_GIMPLE_REG_P set.
4341 IMPORTANT NOTE: This promotion is performed by introducing a load of the
4342 other, unmodified part of the complex object just before the total store.
4343 As a consequence, if the object is still uninitialized, an undefined value
4344 will be loaded into a register, which may result in a spurious exception
4345 if the register is floating-point and the value happens to be a signaling
4346 NaN for example. Then the fully-fledged complex operations lowering pass
4347 followed by a DCE pass are necessary in order to fix things up. */
4349 static enum gimplify_status
4350 gimplify_modify_expr_complex_part (tree *expr_p, gimple_seq *pre_p,
4351 bool want_value)
4353 enum tree_code code, ocode;
4354 tree lhs, rhs, new_rhs, other, realpart, imagpart;
4356 lhs = TREE_OPERAND (*expr_p, 0);
4357 rhs = TREE_OPERAND (*expr_p, 1);
4358 code = TREE_CODE (lhs);
4359 lhs = TREE_OPERAND (lhs, 0);
4361 ocode = code == REALPART_EXPR ? IMAGPART_EXPR : REALPART_EXPR;
4362 other = build1 (ocode, TREE_TYPE (rhs), lhs);
4363 other = get_formal_tmp_var (other, pre_p);
4365 realpart = code == REALPART_EXPR ? rhs : other;
4366 imagpart = code == REALPART_EXPR ? other : rhs;
4368 if (TREE_CONSTANT (realpart) && TREE_CONSTANT (imagpart))
4369 new_rhs = build_complex (TREE_TYPE (lhs), realpart, imagpart);
4370 else
4371 new_rhs = build2 (COMPLEX_EXPR, TREE_TYPE (lhs), realpart, imagpart);
4373 gimplify_seq_add_stmt (pre_p, gimple_build_assign (lhs, new_rhs));
4374 *expr_p = (want_value) ? rhs : NULL_TREE;
4376 return GS_ALL_DONE;
4380 /* Gimplify the MODIFY_EXPR node pointed to by EXPR_P.
4382 modify_expr
4383 : varname '=' rhs
4384 | '*' ID '=' rhs
4386 PRE_P points to the list where side effects that must happen before
4387 *EXPR_P should be stored.
4389 POST_P points to the list where side effects that must happen after
4390 *EXPR_P should be stored.
4392 WANT_VALUE is nonzero iff we want to use the value of this expression
4393 in another expression. */
4395 static enum gimplify_status
4396 gimplify_modify_expr (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p,
4397 bool want_value)
4399 tree *from_p = &TREE_OPERAND (*expr_p, 1);
4400 tree *to_p = &TREE_OPERAND (*expr_p, 0);
4401 enum gimplify_status ret = GS_UNHANDLED;
4402 gimple assign;
4403 location_t loc = EXPR_LOCATION (*expr_p);
4405 gcc_assert (TREE_CODE (*expr_p) == MODIFY_EXPR
4406 || TREE_CODE (*expr_p) == INIT_EXPR);
4408 /* Insert pointer conversions required by the middle-end that are not
4409 required by the frontend. This fixes middle-end type checking for
4410 for example gcc.dg/redecl-6.c. */
4411 if (POINTER_TYPE_P (TREE_TYPE (*to_p)))
4413 STRIP_USELESS_TYPE_CONVERSION (*from_p);
4414 if (!useless_type_conversion_p (TREE_TYPE (*to_p), TREE_TYPE (*from_p)))
4415 *from_p = fold_convert_loc (loc, TREE_TYPE (*to_p), *from_p);
4418 /* See if any simplifications can be done based on what the RHS is. */
4419 ret = gimplify_modify_expr_rhs (expr_p, from_p, to_p, pre_p, post_p,
4420 want_value);
4421 if (ret != GS_UNHANDLED)
4422 return ret;
4424 /* For zero sized types only gimplify the left hand side and right hand
4425 side as statements and throw away the assignment. Do this after
4426 gimplify_modify_expr_rhs so we handle TARGET_EXPRs of addressable
4427 types properly. */
4428 if (zero_sized_type (TREE_TYPE (*from_p)) && !want_value)
4430 gimplify_stmt (from_p, pre_p);
4431 gimplify_stmt (to_p, pre_p);
4432 *expr_p = NULL_TREE;
4433 return GS_ALL_DONE;
4436 /* If the value being copied is of variable width, compute the length
4437 of the copy into a WITH_SIZE_EXPR. Note that we need to do this
4438 before gimplifying any of the operands so that we can resolve any
4439 PLACEHOLDER_EXPRs in the size. Also note that the RTL expander uses
4440 the size of the expression to be copied, not of the destination, so
4441 that is what we must do here. */
4442 maybe_with_size_expr (from_p);
4444 ret = gimplify_expr (to_p, pre_p, post_p, is_gimple_lvalue, fb_lvalue);
4445 if (ret == GS_ERROR)
4446 return ret;
4448 /* As a special case, we have to temporarily allow for assignments
4449 with a CALL_EXPR on the RHS. Since in GIMPLE a function call is
4450 a toplevel statement, when gimplifying the GENERIC expression
4451 MODIFY_EXPR <a, CALL_EXPR <foo>>, we cannot create the tuple
4452 GIMPLE_ASSIGN <a, GIMPLE_CALL <foo>>.
4454 Instead, we need to create the tuple GIMPLE_CALL <a, foo>. To
4455 prevent gimplify_expr from trying to create a new temporary for
4456 foo's LHS, we tell it that it should only gimplify until it
4457 reaches the CALL_EXPR. On return from gimplify_expr, the newly
4458 created GIMPLE_CALL <foo> will be the last statement in *PRE_P
4459 and all we need to do here is set 'a' to be its LHS. */
4460 ret = gimplify_expr (from_p, pre_p, post_p, rhs_predicate_for (*to_p),
4461 fb_rvalue);
4462 if (ret == GS_ERROR)
4463 return ret;
4465 /* Now see if the above changed *from_p to something we handle specially. */
4466 ret = gimplify_modify_expr_rhs (expr_p, from_p, to_p, pre_p, post_p,
4467 want_value);
4468 if (ret != GS_UNHANDLED)
4469 return ret;
4471 /* If we've got a variable sized assignment between two lvalues (i.e. does
4472 not involve a call), then we can make things a bit more straightforward
4473 by converting the assignment to memcpy or memset. */
4474 if (TREE_CODE (*from_p) == WITH_SIZE_EXPR)
4476 tree from = TREE_OPERAND (*from_p, 0);
4477 tree size = TREE_OPERAND (*from_p, 1);
4479 if (TREE_CODE (from) == CONSTRUCTOR)
4480 return gimplify_modify_expr_to_memset (expr_p, size, want_value, pre_p);
4482 if (is_gimple_addressable (from))
4484 *from_p = from;
4485 return gimplify_modify_expr_to_memcpy (expr_p, size, want_value,
4486 pre_p);
4490 /* Transform partial stores to non-addressable complex variables into
4491 total stores. This allows us to use real instead of virtual operands
4492 for these variables, which improves optimization. */
4493 if ((TREE_CODE (*to_p) == REALPART_EXPR
4494 || TREE_CODE (*to_p) == IMAGPART_EXPR)
4495 && is_gimple_reg (TREE_OPERAND (*to_p, 0)))
4496 return gimplify_modify_expr_complex_part (expr_p, pre_p, want_value);
4498 /* Try to alleviate the effects of the gimplification creating artificial
4499 temporaries (see for example is_gimple_reg_rhs) on the debug info. */
4500 if (!gimplify_ctxp->into_ssa
4501 && DECL_P (*from_p)
4502 && DECL_IGNORED_P (*from_p)
4503 && DECL_P (*to_p)
4504 && !DECL_IGNORED_P (*to_p))
4506 if (!DECL_NAME (*from_p) && DECL_NAME (*to_p))
4507 DECL_NAME (*from_p)
4508 = create_tmp_var_name (IDENTIFIER_POINTER (DECL_NAME (*to_p)));
4509 DECL_DEBUG_EXPR_IS_FROM (*from_p) = 1;
4510 SET_DECL_DEBUG_EXPR (*from_p, *to_p);
4513 if (TREE_CODE (*from_p) == CALL_EXPR)
4515 /* Since the RHS is a CALL_EXPR, we need to create a GIMPLE_CALL
4516 instead of a GIMPLE_ASSIGN. */
4517 assign = gimple_build_call_from_tree (*from_p);
4518 if (!gimple_call_noreturn_p (assign))
4519 gimple_call_set_lhs (assign, *to_p);
4521 else
4523 assign = gimple_build_assign (*to_p, *from_p);
4524 gimple_set_location (assign, EXPR_LOCATION (*expr_p));
4527 gimplify_seq_add_stmt (pre_p, assign);
4529 if (gimplify_ctxp->into_ssa && is_gimple_reg (*to_p))
4531 /* If we've somehow already got an SSA_NAME on the LHS, then
4532 we've probably modified it twice. Not good. */
4533 gcc_assert (TREE_CODE (*to_p) != SSA_NAME);
4534 *to_p = make_ssa_name (*to_p, assign);
4535 gimple_set_lhs (assign, *to_p);
4538 if (want_value)
4540 *expr_p = unshare_expr (*to_p);
4541 return GS_OK;
4543 else
4544 *expr_p = NULL;
4546 return GS_ALL_DONE;
4549 /* Gimplify a comparison between two variable-sized objects. Do this
4550 with a call to BUILT_IN_MEMCMP. */
4552 static enum gimplify_status
4553 gimplify_variable_sized_compare (tree *expr_p)
4555 tree op0 = TREE_OPERAND (*expr_p, 0);
4556 tree op1 = TREE_OPERAND (*expr_p, 1);
4557 tree t, arg, dest, src;
4558 location_t loc = EXPR_LOCATION (*expr_p);
4560 arg = TYPE_SIZE_UNIT (TREE_TYPE (op0));
4561 arg = unshare_expr (arg);
4562 arg = SUBSTITUTE_PLACEHOLDER_IN_EXPR (arg, op0);
4563 src = build_fold_addr_expr_loc (loc, op1);
4564 dest = build_fold_addr_expr_loc (loc, op0);
4565 t = implicit_built_in_decls[BUILT_IN_MEMCMP];
4566 t = build_call_expr_loc (loc, t, 3, dest, src, arg);
4567 *expr_p
4568 = build2 (TREE_CODE (*expr_p), TREE_TYPE (*expr_p), t, integer_zero_node);
4570 return GS_OK;
4573 /* Gimplify a comparison between two aggregate objects of integral scalar
4574 mode as a comparison between the bitwise equivalent scalar values. */
4576 static enum gimplify_status
4577 gimplify_scalar_mode_aggregate_compare (tree *expr_p)
4579 location_t loc = EXPR_LOCATION (*expr_p);
4580 tree op0 = TREE_OPERAND (*expr_p, 0);
4581 tree op1 = TREE_OPERAND (*expr_p, 1);
4583 tree type = TREE_TYPE (op0);
4584 tree scalar_type = lang_hooks.types.type_for_mode (TYPE_MODE (type), 1);
4586 op0 = fold_build1_loc (loc, VIEW_CONVERT_EXPR, scalar_type, op0);
4587 op1 = fold_build1_loc (loc, VIEW_CONVERT_EXPR, scalar_type, op1);
4589 *expr_p
4590 = fold_build2_loc (loc, TREE_CODE (*expr_p), TREE_TYPE (*expr_p), op0, op1);
4592 return GS_OK;
4595 /* Gimplify TRUTH_ANDIF_EXPR and TRUTH_ORIF_EXPR expressions. EXPR_P
4596 points to the expression to gimplify.
4598 Expressions of the form 'a && b' are gimplified to:
4600 a && b ? true : false
4602 LOCUS is the source location to be put on the generated COND_EXPR.
4603 gimplify_cond_expr will do the rest. */
4605 static enum gimplify_status
4606 gimplify_boolean_expr (tree *expr_p, location_t locus)
4608 /* Preserve the original type of the expression. */
4609 tree type = TREE_TYPE (*expr_p);
4611 *expr_p = build3 (COND_EXPR, type, *expr_p,
4612 fold_convert_loc (locus, type, boolean_true_node),
4613 fold_convert_loc (locus, type, boolean_false_node));
4615 SET_EXPR_LOCATION (*expr_p, locus);
4617 return GS_OK;
4620 /* Gimplifies an expression sequence. This function gimplifies each
4621 expression and re-writes the original expression with the last
4622 expression of the sequence in GIMPLE form.
4624 PRE_P points to the list where the side effects for all the
4625 expressions in the sequence will be emitted.
4627 WANT_VALUE is true when the result of the last COMPOUND_EXPR is used. */
4629 static enum gimplify_status
4630 gimplify_compound_expr (tree *expr_p, gimple_seq *pre_p, bool want_value)
4632 tree t = *expr_p;
4636 tree *sub_p = &TREE_OPERAND (t, 0);
4638 if (TREE_CODE (*sub_p) == COMPOUND_EXPR)
4639 gimplify_compound_expr (sub_p, pre_p, false);
4640 else
4641 gimplify_stmt (sub_p, pre_p);
4643 t = TREE_OPERAND (t, 1);
4645 while (TREE_CODE (t) == COMPOUND_EXPR);
4647 *expr_p = t;
4648 if (want_value)
4649 return GS_OK;
4650 else
4652 gimplify_stmt (expr_p, pre_p);
4653 return GS_ALL_DONE;
4658 /* Gimplify a SAVE_EXPR node. EXPR_P points to the expression to
4659 gimplify. After gimplification, EXPR_P will point to a new temporary
4660 that holds the original value of the SAVE_EXPR node.
4662 PRE_P points to the list where side effects that must happen before
4663 *EXPR_P should be stored. */
4665 static enum gimplify_status
4666 gimplify_save_expr (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p)
4668 enum gimplify_status ret = GS_ALL_DONE;
4669 tree val;
4671 gcc_assert (TREE_CODE (*expr_p) == SAVE_EXPR);
4672 val = TREE_OPERAND (*expr_p, 0);
4674 /* If the SAVE_EXPR has not been resolved, then evaluate it once. */
4675 if (!SAVE_EXPR_RESOLVED_P (*expr_p))
4677 /* The operand may be a void-valued expression such as SAVE_EXPRs
4678 generated by the Java frontend for class initialization. It is
4679 being executed only for its side-effects. */
4680 if (TREE_TYPE (val) == void_type_node)
4682 ret = gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p, post_p,
4683 is_gimple_stmt, fb_none);
4684 val = NULL;
4686 else
4687 val = get_initialized_tmp_var (val, pre_p, post_p);
4689 TREE_OPERAND (*expr_p, 0) = val;
4690 SAVE_EXPR_RESOLVED_P (*expr_p) = 1;
4693 *expr_p = val;
4695 return ret;
4698 /* Re-write the ADDR_EXPR node pointed to by EXPR_P
4700 unary_expr
4701 : ...
4702 | '&' varname
4705 PRE_P points to the list where side effects that must happen before
4706 *EXPR_P should be stored.
4708 POST_P points to the list where side effects that must happen after
4709 *EXPR_P should be stored. */
4711 static enum gimplify_status
4712 gimplify_addr_expr (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p)
4714 tree expr = *expr_p;
4715 tree op0 = TREE_OPERAND (expr, 0);
4716 enum gimplify_status ret;
4717 location_t loc = EXPR_LOCATION (*expr_p);
4719 switch (TREE_CODE (op0))
4721 case INDIRECT_REF:
4722 case MISALIGNED_INDIRECT_REF:
4723 do_indirect_ref:
4724 /* Check if we are dealing with an expression of the form '&*ptr'.
4725 While the front end folds away '&*ptr' into 'ptr', these
4726 expressions may be generated internally by the compiler (e.g.,
4727 builtins like __builtin_va_end). */
4728 /* Caution: the silent array decomposition semantics we allow for
4729 ADDR_EXPR means we can't always discard the pair. */
4730 /* Gimplification of the ADDR_EXPR operand may drop
4731 cv-qualification conversions, so make sure we add them if
4732 needed. */
4734 tree op00 = TREE_OPERAND (op0, 0);
4735 tree t_expr = TREE_TYPE (expr);
4736 tree t_op00 = TREE_TYPE (op00);
4738 if (!useless_type_conversion_p (t_expr, t_op00))
4739 op00 = fold_convert_loc (loc, TREE_TYPE (expr), op00);
4740 *expr_p = op00;
4741 ret = GS_OK;
4743 break;
4745 case VIEW_CONVERT_EXPR:
4746 /* Take the address of our operand and then convert it to the type of
4747 this ADDR_EXPR.
4749 ??? The interactions of VIEW_CONVERT_EXPR and aliasing is not at
4750 all clear. The impact of this transformation is even less clear. */
4752 /* If the operand is a useless conversion, look through it. Doing so
4753 guarantees that the ADDR_EXPR and its operand will remain of the
4754 same type. */
4755 if (tree_ssa_useless_type_conversion (TREE_OPERAND (op0, 0)))
4756 op0 = TREE_OPERAND (op0, 0);
4758 *expr_p = fold_convert_loc (loc, TREE_TYPE (expr),
4759 build_fold_addr_expr_loc (loc,
4760 TREE_OPERAND (op0, 0)));
4761 ret = GS_OK;
4762 break;
4764 default:
4765 /* We use fb_either here because the C frontend sometimes takes
4766 the address of a call that returns a struct; see
4767 gcc.dg/c99-array-lval-1.c. The gimplifier will correctly make
4768 the implied temporary explicit. */
4770 /* Make the operand addressable. */
4771 ret = gimplify_expr (&TREE_OPERAND (expr, 0), pre_p, post_p,
4772 is_gimple_addressable, fb_either);
4773 if (ret == GS_ERROR)
4774 break;
4776 /* Then mark it. Beware that it may not be possible to do so directly
4777 if a temporary has been created by the gimplification. */
4778 prepare_gimple_addressable (&TREE_OPERAND (expr, 0), pre_p);
4780 op0 = TREE_OPERAND (expr, 0);
4782 /* For various reasons, the gimplification of the expression
4783 may have made a new INDIRECT_REF. */
4784 if (TREE_CODE (op0) == INDIRECT_REF)
4785 goto do_indirect_ref;
4787 mark_addressable (TREE_OPERAND (expr, 0));
4789 /* The FEs may end up building ADDR_EXPRs early on a decl with
4790 an incomplete type. Re-build ADDR_EXPRs in canonical form
4791 here. */
4792 if (!types_compatible_p (TREE_TYPE (op0), TREE_TYPE (TREE_TYPE (expr))))
4793 *expr_p = build_fold_addr_expr (op0);
4795 /* Make sure TREE_CONSTANT and TREE_SIDE_EFFECTS are set properly. */
4796 recompute_tree_invariant_for_addr_expr (*expr_p);
4798 /* If we re-built the ADDR_EXPR add a conversion to the original type
4799 if required. */
4800 if (!useless_type_conversion_p (TREE_TYPE (expr), TREE_TYPE (*expr_p)))
4801 *expr_p = fold_convert (TREE_TYPE (expr), *expr_p);
4803 break;
4806 return ret;
4809 /* Gimplify the operands of an ASM_EXPR. Input operands should be a gimple
4810 value; output operands should be a gimple lvalue. */
4812 static enum gimplify_status
4813 gimplify_asm_expr (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p)
4815 tree expr;
4816 int noutputs;
4817 const char **oconstraints;
4818 int i;
4819 tree link;
4820 const char *constraint;
4821 bool allows_mem, allows_reg, is_inout;
4822 enum gimplify_status ret, tret;
4823 gimple stmt;
4824 VEC(tree, gc) *inputs;
4825 VEC(tree, gc) *outputs;
4826 VEC(tree, gc) *clobbers;
4827 VEC(tree, gc) *labels;
4828 tree link_next;
4830 expr = *expr_p;
4831 noutputs = list_length (ASM_OUTPUTS (expr));
4832 oconstraints = (const char **) alloca ((noutputs) * sizeof (const char *));
4834 inputs = outputs = clobbers = labels = NULL;
4836 ret = GS_ALL_DONE;
4837 link_next = NULL_TREE;
4838 for (i = 0, link = ASM_OUTPUTS (expr); link; ++i, link = link_next)
4840 bool ok;
4841 size_t constraint_len;
4843 link_next = TREE_CHAIN (link);
4845 oconstraints[i]
4846 = constraint
4847 = TREE_STRING_POINTER (TREE_VALUE (TREE_PURPOSE (link)));
4848 constraint_len = strlen (constraint);
4849 if (constraint_len == 0)
4850 continue;
4852 ok = parse_output_constraint (&constraint, i, 0, 0,
4853 &allows_mem, &allows_reg, &is_inout);
4854 if (!ok)
4856 ret = GS_ERROR;
4857 is_inout = false;
4860 if (!allows_reg && allows_mem)
4861 mark_addressable (TREE_VALUE (link));
4863 tret = gimplify_expr (&TREE_VALUE (link), pre_p, post_p,
4864 is_inout ? is_gimple_min_lval : is_gimple_lvalue,
4865 fb_lvalue | fb_mayfail);
4866 if (tret == GS_ERROR)
4868 error ("invalid lvalue in asm output %d", i);
4869 ret = tret;
4872 VEC_safe_push (tree, gc, outputs, link);
4873 TREE_CHAIN (link) = NULL_TREE;
4875 if (is_inout)
4877 /* An input/output operand. To give the optimizers more
4878 flexibility, split it into separate input and output
4879 operands. */
4880 tree input;
4881 char buf[10];
4883 /* Turn the in/out constraint into an output constraint. */
4884 char *p = xstrdup (constraint);
4885 p[0] = '=';
4886 TREE_VALUE (TREE_PURPOSE (link)) = build_string (constraint_len, p);
4888 /* And add a matching input constraint. */
4889 if (allows_reg)
4891 sprintf (buf, "%d", i);
4893 /* If there are multiple alternatives in the constraint,
4894 handle each of them individually. Those that allow register
4895 will be replaced with operand number, the others will stay
4896 unchanged. */
4897 if (strchr (p, ',') != NULL)
4899 size_t len = 0, buflen = strlen (buf);
4900 char *beg, *end, *str, *dst;
4902 for (beg = p + 1;;)
4904 end = strchr (beg, ',');
4905 if (end == NULL)
4906 end = strchr (beg, '\0');
4907 if ((size_t) (end - beg) < buflen)
4908 len += buflen + 1;
4909 else
4910 len += end - beg + 1;
4911 if (*end)
4912 beg = end + 1;
4913 else
4914 break;
4917 str = (char *) alloca (len);
4918 for (beg = p + 1, dst = str;;)
4920 const char *tem;
4921 bool mem_p, reg_p, inout_p;
4923 end = strchr (beg, ',');
4924 if (end)
4925 *end = '\0';
4926 beg[-1] = '=';
4927 tem = beg - 1;
4928 parse_output_constraint (&tem, i, 0, 0,
4929 &mem_p, &reg_p, &inout_p);
4930 if (dst != str)
4931 *dst++ = ',';
4932 if (reg_p)
4934 memcpy (dst, buf, buflen);
4935 dst += buflen;
4937 else
4939 if (end)
4940 len = end - beg;
4941 else
4942 len = strlen (beg);
4943 memcpy (dst, beg, len);
4944 dst += len;
4946 if (end)
4947 beg = end + 1;
4948 else
4949 break;
4951 *dst = '\0';
4952 input = build_string (dst - str, str);
4954 else
4955 input = build_string (strlen (buf), buf);
4957 else
4958 input = build_string (constraint_len - 1, constraint + 1);
4960 free (p);
4962 input = build_tree_list (build_tree_list (NULL_TREE, input),
4963 unshare_expr (TREE_VALUE (link)));
4964 ASM_INPUTS (expr) = chainon (ASM_INPUTS (expr), input);
4968 link_next = NULL_TREE;
4969 for (link = ASM_INPUTS (expr); link; ++i, link = link_next)
4971 link_next = TREE_CHAIN (link);
4972 constraint = TREE_STRING_POINTER (TREE_VALUE (TREE_PURPOSE (link)));
4973 parse_input_constraint (&constraint, 0, 0, noutputs, 0,
4974 oconstraints, &allows_mem, &allows_reg);
4976 /* If we can't make copies, we can only accept memory. */
4977 if (TREE_ADDRESSABLE (TREE_TYPE (TREE_VALUE (link))))
4979 if (allows_mem)
4980 allows_reg = 0;
4981 else
4983 error ("impossible constraint in %<asm%>");
4984 error ("non-memory input %d must stay in memory", i);
4985 return GS_ERROR;
4989 /* If the operand is a memory input, it should be an lvalue. */
4990 if (!allows_reg && allows_mem)
4992 tret = gimplify_expr (&TREE_VALUE (link), pre_p, post_p,
4993 is_gimple_lvalue, fb_lvalue | fb_mayfail);
4994 mark_addressable (TREE_VALUE (link));
4995 if (tret == GS_ERROR)
4997 if (EXPR_HAS_LOCATION (TREE_VALUE (link)))
4998 input_location = EXPR_LOCATION (TREE_VALUE (link));
4999 error ("memory input %d is not directly addressable", i);
5000 ret = tret;
5003 else
5005 tret = gimplify_expr (&TREE_VALUE (link), pre_p, post_p,
5006 is_gimple_asm_val, fb_rvalue);
5007 if (tret == GS_ERROR)
5008 ret = tret;
5011 TREE_CHAIN (link) = NULL_TREE;
5012 VEC_safe_push (tree, gc, inputs, link);
5015 for (link = ASM_CLOBBERS (expr); link; ++i, link = TREE_CHAIN (link))
5016 VEC_safe_push (tree, gc, clobbers, link);
5018 for (link = ASM_LABELS (expr); link; ++i, link = TREE_CHAIN (link))
5019 VEC_safe_push (tree, gc, labels, link);
5021 /* Do not add ASMs with errors to the gimple IL stream. */
5022 if (ret != GS_ERROR)
5024 stmt = gimple_build_asm_vec (TREE_STRING_POINTER (ASM_STRING (expr)),
5025 inputs, outputs, clobbers, labels);
5027 gimple_asm_set_volatile (stmt, ASM_VOLATILE_P (expr));
5028 gimple_asm_set_input (stmt, ASM_INPUT_P (expr));
5030 gimplify_seq_add_stmt (pre_p, stmt);
5033 return ret;
5036 /* Gimplify a CLEANUP_POINT_EXPR. Currently this works by adding
5037 GIMPLE_WITH_CLEANUP_EXPRs to the prequeue as we encounter cleanups while
5038 gimplifying the body, and converting them to TRY_FINALLY_EXPRs when we
5039 return to this function.
5041 FIXME should we complexify the prequeue handling instead? Or use flags
5042 for all the cleanups and let the optimizer tighten them up? The current
5043 code seems pretty fragile; it will break on a cleanup within any
5044 non-conditional nesting. But any such nesting would be broken, anyway;
5045 we can't write a TRY_FINALLY_EXPR that starts inside a nesting construct
5046 and continues out of it. We can do that at the RTL level, though, so
5047 having an optimizer to tighten up try/finally regions would be a Good
5048 Thing. */
5050 static enum gimplify_status
5051 gimplify_cleanup_point_expr (tree *expr_p, gimple_seq *pre_p)
5053 gimple_stmt_iterator iter;
5054 gimple_seq body_sequence = NULL;
5056 tree temp = voidify_wrapper_expr (*expr_p, NULL);
5058 /* We only care about the number of conditions between the innermost
5059 CLEANUP_POINT_EXPR and the cleanup. So save and reset the count and
5060 any cleanups collected outside the CLEANUP_POINT_EXPR. */
5061 int old_conds = gimplify_ctxp->conditions;
5062 gimple_seq old_cleanups = gimplify_ctxp->conditional_cleanups;
5063 gimplify_ctxp->conditions = 0;
5064 gimplify_ctxp->conditional_cleanups = NULL;
5066 gimplify_stmt (&TREE_OPERAND (*expr_p, 0), &body_sequence);
5068 gimplify_ctxp->conditions = old_conds;
5069 gimplify_ctxp->conditional_cleanups = old_cleanups;
5071 for (iter = gsi_start (body_sequence); !gsi_end_p (iter); )
5073 gimple wce = gsi_stmt (iter);
5075 if (gimple_code (wce) == GIMPLE_WITH_CLEANUP_EXPR)
5077 if (gsi_one_before_end_p (iter))
5079 /* Note that gsi_insert_seq_before and gsi_remove do not
5080 scan operands, unlike some other sequence mutators. */
5081 gsi_insert_seq_before_without_update (&iter,
5082 gimple_wce_cleanup (wce),
5083 GSI_SAME_STMT);
5084 gsi_remove (&iter, true);
5085 break;
5087 else
5089 gimple gtry;
5090 gimple_seq seq;
5091 enum gimple_try_flags kind;
5093 if (gimple_wce_cleanup_eh_only (wce))
5094 kind = GIMPLE_TRY_CATCH;
5095 else
5096 kind = GIMPLE_TRY_FINALLY;
5097 seq = gsi_split_seq_after (iter);
5099 gtry = gimple_build_try (seq, gimple_wce_cleanup (wce), kind);
5100 /* Do not use gsi_replace here, as it may scan operands.
5101 We want to do a simple structural modification only. */
5102 *gsi_stmt_ptr (&iter) = gtry;
5103 iter = gsi_start (seq);
5106 else
5107 gsi_next (&iter);
5110 gimplify_seq_add_seq (pre_p, body_sequence);
5111 if (temp)
5113 *expr_p = temp;
5114 return GS_OK;
5116 else
5118 *expr_p = NULL;
5119 return GS_ALL_DONE;
5123 /* Insert a cleanup marker for gimplify_cleanup_point_expr. CLEANUP
5124 is the cleanup action required. EH_ONLY is true if the cleanup should
5125 only be executed if an exception is thrown, not on normal exit. */
5127 static void
5128 gimple_push_cleanup (tree var, tree cleanup, bool eh_only, gimple_seq *pre_p)
5130 gimple wce;
5131 gimple_seq cleanup_stmts = NULL;
5133 /* Errors can result in improperly nested cleanups. Which results in
5134 confusion when trying to resolve the GIMPLE_WITH_CLEANUP_EXPR. */
5135 if (errorcount || sorrycount)
5136 return;
5138 if (gimple_conditional_context ())
5140 /* If we're in a conditional context, this is more complex. We only
5141 want to run the cleanup if we actually ran the initialization that
5142 necessitates it, but we want to run it after the end of the
5143 conditional context. So we wrap the try/finally around the
5144 condition and use a flag to determine whether or not to actually
5145 run the destructor. Thus
5147 test ? f(A()) : 0
5149 becomes (approximately)
5151 flag = 0;
5152 try {
5153 if (test) { A::A(temp); flag = 1; val = f(temp); }
5154 else { val = 0; }
5155 } finally {
5156 if (flag) A::~A(temp);
5160 tree flag = create_tmp_var (boolean_type_node, "cleanup");
5161 gimple ffalse = gimple_build_assign (flag, boolean_false_node);
5162 gimple ftrue = gimple_build_assign (flag, boolean_true_node);
5164 cleanup = build3 (COND_EXPR, void_type_node, flag, cleanup, NULL);
5165 gimplify_stmt (&cleanup, &cleanup_stmts);
5166 wce = gimple_build_wce (cleanup_stmts);
5168 gimplify_seq_add_stmt (&gimplify_ctxp->conditional_cleanups, ffalse);
5169 gimplify_seq_add_stmt (&gimplify_ctxp->conditional_cleanups, wce);
5170 gimplify_seq_add_stmt (pre_p, ftrue);
5172 /* Because of this manipulation, and the EH edges that jump
5173 threading cannot redirect, the temporary (VAR) will appear
5174 to be used uninitialized. Don't warn. */
5175 TREE_NO_WARNING (var) = 1;
5177 else
5179 gimplify_stmt (&cleanup, &cleanup_stmts);
5180 wce = gimple_build_wce (cleanup_stmts);
5181 gimple_wce_set_cleanup_eh_only (wce, eh_only);
5182 gimplify_seq_add_stmt (pre_p, wce);
5186 /* Gimplify a TARGET_EXPR which doesn't appear on the rhs of an INIT_EXPR. */
5188 static enum gimplify_status
5189 gimplify_target_expr (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p)
5191 tree targ = *expr_p;
5192 tree temp = TARGET_EXPR_SLOT (targ);
5193 tree init = TARGET_EXPR_INITIAL (targ);
5194 enum gimplify_status ret;
5196 if (init)
5198 /* TARGET_EXPR temps aren't part of the enclosing block, so add it
5199 to the temps list. Handle also variable length TARGET_EXPRs. */
5200 if (TREE_CODE (DECL_SIZE (temp)) != INTEGER_CST)
5202 if (!TYPE_SIZES_GIMPLIFIED (TREE_TYPE (temp)))
5203 gimplify_type_sizes (TREE_TYPE (temp), pre_p);
5204 gimplify_vla_decl (temp, pre_p);
5206 else
5207 gimple_add_tmp_var (temp);
5209 /* If TARGET_EXPR_INITIAL is void, then the mere evaluation of the
5210 expression is supposed to initialize the slot. */
5211 if (VOID_TYPE_P (TREE_TYPE (init)))
5212 ret = gimplify_expr (&init, pre_p, post_p, is_gimple_stmt, fb_none);
5213 else
5215 tree init_expr = build2 (INIT_EXPR, void_type_node, temp, init);
5216 init = init_expr;
5217 ret = gimplify_expr (&init, pre_p, post_p, is_gimple_stmt, fb_none);
5218 init = NULL;
5219 ggc_free (init_expr);
5221 if (ret == GS_ERROR)
5223 /* PR c++/28266 Make sure this is expanded only once. */
5224 TARGET_EXPR_INITIAL (targ) = NULL_TREE;
5225 return GS_ERROR;
5227 if (init)
5228 gimplify_and_add (init, pre_p);
5230 /* If needed, push the cleanup for the temp. */
5231 if (TARGET_EXPR_CLEANUP (targ))
5232 gimple_push_cleanup (temp, TARGET_EXPR_CLEANUP (targ),
5233 CLEANUP_EH_ONLY (targ), pre_p);
5235 /* Only expand this once. */
5236 TREE_OPERAND (targ, 3) = init;
5237 TARGET_EXPR_INITIAL (targ) = NULL_TREE;
5239 else
5240 /* We should have expanded this before. */
5241 gcc_assert (DECL_SEEN_IN_BIND_EXPR_P (temp));
5243 *expr_p = temp;
5244 return GS_OK;
5247 /* Gimplification of expression trees. */
5249 /* Gimplify an expression which appears at statement context. The
5250 corresponding GIMPLE statements are added to *SEQ_P. If *SEQ_P is
5251 NULL, a new sequence is allocated.
5253 Return true if we actually added a statement to the queue. */
5255 bool
5256 gimplify_stmt (tree *stmt_p, gimple_seq *seq_p)
5258 gimple_seq_node last;
5260 if (!*seq_p)
5261 *seq_p = gimple_seq_alloc ();
5263 last = gimple_seq_last (*seq_p);
5264 gimplify_expr (stmt_p, seq_p, NULL, is_gimple_stmt, fb_none);
5265 return last != gimple_seq_last (*seq_p);
5269 /* Add FIRSTPRIVATE entries for DECL in the OpenMP the surrounding parallels
5270 to CTX. If entries already exist, force them to be some flavor of private.
5271 If there is no enclosing parallel, do nothing. */
5273 void
5274 omp_firstprivatize_variable (struct gimplify_omp_ctx *ctx, tree decl)
5276 splay_tree_node n;
5278 if (decl == NULL || !DECL_P (decl))
5279 return;
5283 n = splay_tree_lookup (ctx->variables, (splay_tree_key)decl);
5284 if (n != NULL)
5286 if (n->value & GOVD_SHARED)
5287 n->value = GOVD_FIRSTPRIVATE | (n->value & GOVD_SEEN);
5288 else
5289 return;
5291 else if (ctx->region_type != ORT_WORKSHARE)
5292 omp_add_variable (ctx, decl, GOVD_FIRSTPRIVATE);
5294 ctx = ctx->outer_context;
5296 while (ctx);
5299 /* Similarly for each of the type sizes of TYPE. */
5301 static void
5302 omp_firstprivatize_type_sizes (struct gimplify_omp_ctx *ctx, tree type)
5304 if (type == NULL || type == error_mark_node)
5305 return;
5306 type = TYPE_MAIN_VARIANT (type);
5308 if (pointer_set_insert (ctx->privatized_types, type))
5309 return;
5311 switch (TREE_CODE (type))
5313 case INTEGER_TYPE:
5314 case ENUMERAL_TYPE:
5315 case BOOLEAN_TYPE:
5316 case REAL_TYPE:
5317 case FIXED_POINT_TYPE:
5318 omp_firstprivatize_variable (ctx, TYPE_MIN_VALUE (type));
5319 omp_firstprivatize_variable (ctx, TYPE_MAX_VALUE (type));
5320 break;
5322 case ARRAY_TYPE:
5323 omp_firstprivatize_type_sizes (ctx, TREE_TYPE (type));
5324 omp_firstprivatize_type_sizes (ctx, TYPE_DOMAIN (type));
5325 break;
5327 case RECORD_TYPE:
5328 case UNION_TYPE:
5329 case QUAL_UNION_TYPE:
5331 tree field;
5332 for (field = TYPE_FIELDS (type); field; field = TREE_CHAIN (field))
5333 if (TREE_CODE (field) == FIELD_DECL)
5335 omp_firstprivatize_variable (ctx, DECL_FIELD_OFFSET (field));
5336 omp_firstprivatize_type_sizes (ctx, TREE_TYPE (field));
5339 break;
5341 case POINTER_TYPE:
5342 case REFERENCE_TYPE:
5343 omp_firstprivatize_type_sizes (ctx, TREE_TYPE (type));
5344 break;
5346 default:
5347 break;
5350 omp_firstprivatize_variable (ctx, TYPE_SIZE (type));
5351 omp_firstprivatize_variable (ctx, TYPE_SIZE_UNIT (type));
5352 lang_hooks.types.omp_firstprivatize_type_sizes (ctx, type);
5355 /* Add an entry for DECL in the OpenMP context CTX with FLAGS. */
5357 static void
5358 omp_add_variable (struct gimplify_omp_ctx *ctx, tree decl, unsigned int flags)
5360 splay_tree_node n;
5361 unsigned int nflags;
5362 tree t;
5364 if (decl == error_mark_node || TREE_TYPE (decl) == error_mark_node)
5365 return;
5367 /* Never elide decls whose type has TREE_ADDRESSABLE set. This means
5368 there are constructors involved somewhere. */
5369 if (TREE_ADDRESSABLE (TREE_TYPE (decl))
5370 || TYPE_NEEDS_CONSTRUCTING (TREE_TYPE (decl)))
5371 flags |= GOVD_SEEN;
5373 n = splay_tree_lookup (ctx->variables, (splay_tree_key)decl);
5374 if (n != NULL)
5376 /* We shouldn't be re-adding the decl with the same data
5377 sharing class. */
5378 gcc_assert ((n->value & GOVD_DATA_SHARE_CLASS & flags) == 0);
5379 /* The only combination of data sharing classes we should see is
5380 FIRSTPRIVATE and LASTPRIVATE. */
5381 nflags = n->value | flags;
5382 gcc_assert ((nflags & GOVD_DATA_SHARE_CLASS)
5383 == (GOVD_FIRSTPRIVATE | GOVD_LASTPRIVATE));
5384 n->value = nflags;
5385 return;
5388 /* When adding a variable-sized variable, we have to handle all sorts
5389 of additional bits of data: the pointer replacement variable, and
5390 the parameters of the type. */
5391 if (DECL_SIZE (decl) && TREE_CODE (DECL_SIZE (decl)) != INTEGER_CST)
5393 /* Add the pointer replacement variable as PRIVATE if the variable
5394 replacement is private, else FIRSTPRIVATE since we'll need the
5395 address of the original variable either for SHARED, or for the
5396 copy into or out of the context. */
5397 if (!(flags & GOVD_LOCAL))
5399 nflags = flags & GOVD_PRIVATE ? GOVD_PRIVATE : GOVD_FIRSTPRIVATE;
5400 nflags |= flags & GOVD_SEEN;
5401 t = DECL_VALUE_EXPR (decl);
5402 gcc_assert (TREE_CODE (t) == INDIRECT_REF);
5403 t = TREE_OPERAND (t, 0);
5404 gcc_assert (DECL_P (t));
5405 omp_add_variable (ctx, t, nflags);
5408 /* Add all of the variable and type parameters (which should have
5409 been gimplified to a formal temporary) as FIRSTPRIVATE. */
5410 omp_firstprivatize_variable (ctx, DECL_SIZE_UNIT (decl));
5411 omp_firstprivatize_variable (ctx, DECL_SIZE (decl));
5412 omp_firstprivatize_type_sizes (ctx, TREE_TYPE (decl));
5414 /* The variable-sized variable itself is never SHARED, only some form
5415 of PRIVATE. The sharing would take place via the pointer variable
5416 which we remapped above. */
5417 if (flags & GOVD_SHARED)
5418 flags = GOVD_PRIVATE | GOVD_DEBUG_PRIVATE
5419 | (flags & (GOVD_SEEN | GOVD_EXPLICIT));
5421 /* We're going to make use of the TYPE_SIZE_UNIT at least in the
5422 alloca statement we generate for the variable, so make sure it
5423 is available. This isn't automatically needed for the SHARED
5424 case, since we won't be allocating local storage then.
5425 For local variables TYPE_SIZE_UNIT might not be gimplified yet,
5426 in this case omp_notice_variable will be called later
5427 on when it is gimplified. */
5428 else if (! (flags & GOVD_LOCAL))
5429 omp_notice_variable (ctx, TYPE_SIZE_UNIT (TREE_TYPE (decl)), true);
5431 else if (lang_hooks.decls.omp_privatize_by_reference (decl))
5433 gcc_assert ((flags & GOVD_LOCAL) == 0);
5434 omp_firstprivatize_type_sizes (ctx, TREE_TYPE (decl));
5436 /* Similar to the direct variable sized case above, we'll need the
5437 size of references being privatized. */
5438 if ((flags & GOVD_SHARED) == 0)
5440 t = TYPE_SIZE_UNIT (TREE_TYPE (TREE_TYPE (decl)));
5441 if (TREE_CODE (t) != INTEGER_CST)
5442 omp_notice_variable (ctx, t, true);
5446 splay_tree_insert (ctx->variables, (splay_tree_key)decl, flags);
5449 /* Record the fact that DECL was used within the OpenMP context CTX.
5450 IN_CODE is true when real code uses DECL, and false when we should
5451 merely emit default(none) errors. Return true if DECL is going to
5452 be remapped and thus DECL shouldn't be gimplified into its
5453 DECL_VALUE_EXPR (if any). */
5455 static bool
5456 omp_notice_variable (struct gimplify_omp_ctx *ctx, tree decl, bool in_code)
5458 splay_tree_node n;
5459 unsigned flags = in_code ? GOVD_SEEN : 0;
5460 bool ret = false, shared;
5462 if (decl == error_mark_node || TREE_TYPE (decl) == error_mark_node)
5463 return false;
5465 /* Threadprivate variables are predetermined. */
5466 if (is_global_var (decl))
5468 if (DECL_THREAD_LOCAL_P (decl))
5469 return false;
5471 if (DECL_HAS_VALUE_EXPR_P (decl))
5473 tree value = get_base_address (DECL_VALUE_EXPR (decl));
5475 if (value && DECL_P (value) && DECL_THREAD_LOCAL_P (value))
5476 return false;
5480 n = splay_tree_lookup (ctx->variables, (splay_tree_key)decl);
5481 if (n == NULL)
5483 enum omp_clause_default_kind default_kind, kind;
5484 struct gimplify_omp_ctx *octx;
5486 if (ctx->region_type == ORT_WORKSHARE)
5487 goto do_outer;
5489 /* ??? Some compiler-generated variables (like SAVE_EXPRs) could be
5490 remapped firstprivate instead of shared. To some extent this is
5491 addressed in omp_firstprivatize_type_sizes, but not effectively. */
5492 default_kind = ctx->default_kind;
5493 kind = lang_hooks.decls.omp_predetermined_sharing (decl);
5494 if (kind != OMP_CLAUSE_DEFAULT_UNSPECIFIED)
5495 default_kind = kind;
5497 switch (default_kind)
5499 case OMP_CLAUSE_DEFAULT_NONE:
5500 error ("%qE not specified in enclosing parallel",
5501 DECL_NAME (decl));
5502 error_at (ctx->location, "enclosing parallel");
5503 /* FALLTHRU */
5504 case OMP_CLAUSE_DEFAULT_SHARED:
5505 flags |= GOVD_SHARED;
5506 break;
5507 case OMP_CLAUSE_DEFAULT_PRIVATE:
5508 flags |= GOVD_PRIVATE;
5509 break;
5510 case OMP_CLAUSE_DEFAULT_FIRSTPRIVATE:
5511 flags |= GOVD_FIRSTPRIVATE;
5512 break;
5513 case OMP_CLAUSE_DEFAULT_UNSPECIFIED:
5514 /* decl will be either GOVD_FIRSTPRIVATE or GOVD_SHARED. */
5515 gcc_assert (ctx->region_type == ORT_TASK);
5516 if (ctx->outer_context)
5517 omp_notice_variable (ctx->outer_context, decl, in_code);
5518 for (octx = ctx->outer_context; octx; octx = octx->outer_context)
5520 splay_tree_node n2;
5522 n2 = splay_tree_lookup (octx->variables, (splay_tree_key) decl);
5523 if (n2 && (n2->value & GOVD_DATA_SHARE_CLASS) != GOVD_SHARED)
5525 flags |= GOVD_FIRSTPRIVATE;
5526 break;
5528 if ((octx->region_type & ORT_PARALLEL) != 0)
5529 break;
5531 if (flags & GOVD_FIRSTPRIVATE)
5532 break;
5533 if (octx == NULL
5534 && (TREE_CODE (decl) == PARM_DECL
5535 || (!is_global_var (decl)
5536 && DECL_CONTEXT (decl) == current_function_decl)))
5538 flags |= GOVD_FIRSTPRIVATE;
5539 break;
5541 flags |= GOVD_SHARED;
5542 break;
5543 default:
5544 gcc_unreachable ();
5547 if ((flags & GOVD_PRIVATE)
5548 && lang_hooks.decls.omp_private_outer_ref (decl))
5549 flags |= GOVD_PRIVATE_OUTER_REF;
5551 omp_add_variable (ctx, decl, flags);
5553 shared = (flags & GOVD_SHARED) != 0;
5554 ret = lang_hooks.decls.omp_disregard_value_expr (decl, shared);
5555 goto do_outer;
5558 if ((n->value & (GOVD_SEEN | GOVD_LOCAL)) == 0
5559 && (flags & (GOVD_SEEN | GOVD_LOCAL)) == GOVD_SEEN
5560 && DECL_SIZE (decl)
5561 && TREE_CODE (DECL_SIZE (decl)) != INTEGER_CST)
5563 splay_tree_node n2;
5564 tree t = DECL_VALUE_EXPR (decl);
5565 gcc_assert (TREE_CODE (t) == INDIRECT_REF);
5566 t = TREE_OPERAND (t, 0);
5567 gcc_assert (DECL_P (t));
5568 n2 = splay_tree_lookup (ctx->variables, (splay_tree_key) t);
5569 n2->value |= GOVD_SEEN;
5572 shared = ((flags | n->value) & GOVD_SHARED) != 0;
5573 ret = lang_hooks.decls.omp_disregard_value_expr (decl, shared);
5575 /* If nothing changed, there's nothing left to do. */
5576 if ((n->value & flags) == flags)
5577 return ret;
5578 flags |= n->value;
5579 n->value = flags;
5581 do_outer:
5582 /* If the variable is private in the current context, then we don't
5583 need to propagate anything to an outer context. */
5584 if ((flags & GOVD_PRIVATE) && !(flags & GOVD_PRIVATE_OUTER_REF))
5585 return ret;
5586 if (ctx->outer_context
5587 && omp_notice_variable (ctx->outer_context, decl, in_code))
5588 return true;
5589 return ret;
5592 /* Verify that DECL is private within CTX. If there's specific information
5593 to the contrary in the innermost scope, generate an error. */
5595 static bool
5596 omp_is_private (struct gimplify_omp_ctx *ctx, tree decl)
5598 splay_tree_node n;
5600 n = splay_tree_lookup (ctx->variables, (splay_tree_key)decl);
5601 if (n != NULL)
5603 if (n->value & GOVD_SHARED)
5605 if (ctx == gimplify_omp_ctxp)
5607 error ("iteration variable %qE should be private",
5608 DECL_NAME (decl));
5609 n->value = GOVD_PRIVATE;
5610 return true;
5612 else
5613 return false;
5615 else if ((n->value & GOVD_EXPLICIT) != 0
5616 && (ctx == gimplify_omp_ctxp
5617 || (ctx->region_type == ORT_COMBINED_PARALLEL
5618 && gimplify_omp_ctxp->outer_context == ctx)))
5620 if ((n->value & GOVD_FIRSTPRIVATE) != 0)
5621 error ("iteration variable %qE should not be firstprivate",
5622 DECL_NAME (decl));
5623 else if ((n->value & GOVD_REDUCTION) != 0)
5624 error ("iteration variable %qE should not be reduction",
5625 DECL_NAME (decl));
5627 return (ctx == gimplify_omp_ctxp
5628 || (ctx->region_type == ORT_COMBINED_PARALLEL
5629 && gimplify_omp_ctxp->outer_context == ctx));
5632 if (ctx->region_type != ORT_WORKSHARE)
5633 return false;
5634 else if (ctx->outer_context)
5635 return omp_is_private (ctx->outer_context, decl);
5636 return false;
5639 /* Return true if DECL is private within a parallel region
5640 that binds to the current construct's context or in parallel
5641 region's REDUCTION clause. */
5643 static bool
5644 omp_check_private (struct gimplify_omp_ctx *ctx, tree decl)
5646 splay_tree_node n;
5650 ctx = ctx->outer_context;
5651 if (ctx == NULL)
5652 return !(is_global_var (decl)
5653 /* References might be private, but might be shared too. */
5654 || lang_hooks.decls.omp_privatize_by_reference (decl));
5656 n = splay_tree_lookup (ctx->variables, (splay_tree_key) decl);
5657 if (n != NULL)
5658 return (n->value & GOVD_SHARED) == 0;
5660 while (ctx->region_type == ORT_WORKSHARE);
5661 return false;
5664 /* Scan the OpenMP clauses in *LIST_P, installing mappings into a new
5665 and previous omp contexts. */
5667 static void
5668 gimplify_scan_omp_clauses (tree *list_p, gimple_seq *pre_p,
5669 enum omp_region_type region_type)
5671 struct gimplify_omp_ctx *ctx, *outer_ctx;
5672 struct gimplify_ctx gctx;
5673 tree c;
5675 ctx = new_omp_context (region_type);
5676 outer_ctx = ctx->outer_context;
5678 while ((c = *list_p) != NULL)
5680 bool remove = false;
5681 bool notice_outer = true;
5682 const char *check_non_private = NULL;
5683 unsigned int flags;
5684 tree decl;
5686 switch (OMP_CLAUSE_CODE (c))
5688 case OMP_CLAUSE_PRIVATE:
5689 flags = GOVD_PRIVATE | GOVD_EXPLICIT;
5690 if (lang_hooks.decls.omp_private_outer_ref (OMP_CLAUSE_DECL (c)))
5692 flags |= GOVD_PRIVATE_OUTER_REF;
5693 OMP_CLAUSE_PRIVATE_OUTER_REF (c) = 1;
5695 else
5696 notice_outer = false;
5697 goto do_add;
5698 case OMP_CLAUSE_SHARED:
5699 flags = GOVD_SHARED | GOVD_EXPLICIT;
5700 goto do_add;
5701 case OMP_CLAUSE_FIRSTPRIVATE:
5702 flags = GOVD_FIRSTPRIVATE | GOVD_EXPLICIT;
5703 check_non_private = "firstprivate";
5704 goto do_add;
5705 case OMP_CLAUSE_LASTPRIVATE:
5706 flags = GOVD_LASTPRIVATE | GOVD_SEEN | GOVD_EXPLICIT;
5707 check_non_private = "lastprivate";
5708 goto do_add;
5709 case OMP_CLAUSE_REDUCTION:
5710 flags = GOVD_REDUCTION | GOVD_SEEN | GOVD_EXPLICIT;
5711 check_non_private = "reduction";
5712 goto do_add;
5714 do_add:
5715 decl = OMP_CLAUSE_DECL (c);
5716 if (decl == error_mark_node || TREE_TYPE (decl) == error_mark_node)
5718 remove = true;
5719 break;
5721 omp_add_variable (ctx, decl, flags);
5722 if (OMP_CLAUSE_CODE (c) == OMP_CLAUSE_REDUCTION
5723 && OMP_CLAUSE_REDUCTION_PLACEHOLDER (c))
5725 omp_add_variable (ctx, OMP_CLAUSE_REDUCTION_PLACEHOLDER (c),
5726 GOVD_LOCAL | GOVD_SEEN);
5727 gimplify_omp_ctxp = ctx;
5728 push_gimplify_context (&gctx);
5730 OMP_CLAUSE_REDUCTION_GIMPLE_INIT (c) = gimple_seq_alloc ();
5731 OMP_CLAUSE_REDUCTION_GIMPLE_MERGE (c) = gimple_seq_alloc ();
5733 gimplify_and_add (OMP_CLAUSE_REDUCTION_INIT (c),
5734 &OMP_CLAUSE_REDUCTION_GIMPLE_INIT (c));
5735 pop_gimplify_context
5736 (gimple_seq_first_stmt (OMP_CLAUSE_REDUCTION_GIMPLE_INIT (c)));
5737 push_gimplify_context (&gctx);
5738 gimplify_and_add (OMP_CLAUSE_REDUCTION_MERGE (c),
5739 &OMP_CLAUSE_REDUCTION_GIMPLE_MERGE (c));
5740 pop_gimplify_context
5741 (gimple_seq_first_stmt (OMP_CLAUSE_REDUCTION_GIMPLE_MERGE (c)));
5742 OMP_CLAUSE_REDUCTION_INIT (c) = NULL_TREE;
5743 OMP_CLAUSE_REDUCTION_MERGE (c) = NULL_TREE;
5745 gimplify_omp_ctxp = outer_ctx;
5747 else if (OMP_CLAUSE_CODE (c) == OMP_CLAUSE_LASTPRIVATE
5748 && OMP_CLAUSE_LASTPRIVATE_STMT (c))
5750 gimplify_omp_ctxp = ctx;
5751 push_gimplify_context (&gctx);
5752 if (TREE_CODE (OMP_CLAUSE_LASTPRIVATE_STMT (c)) != BIND_EXPR)
5754 tree bind = build3 (BIND_EXPR, void_type_node, NULL,
5755 NULL, NULL);
5756 TREE_SIDE_EFFECTS (bind) = 1;
5757 BIND_EXPR_BODY (bind) = OMP_CLAUSE_LASTPRIVATE_STMT (c);
5758 OMP_CLAUSE_LASTPRIVATE_STMT (c) = bind;
5760 gimplify_and_add (OMP_CLAUSE_LASTPRIVATE_STMT (c),
5761 &OMP_CLAUSE_LASTPRIVATE_GIMPLE_SEQ (c));
5762 pop_gimplify_context
5763 (gimple_seq_first_stmt (OMP_CLAUSE_LASTPRIVATE_GIMPLE_SEQ (c)));
5764 OMP_CLAUSE_LASTPRIVATE_STMT (c) = NULL_TREE;
5766 gimplify_omp_ctxp = outer_ctx;
5768 if (notice_outer)
5769 goto do_notice;
5770 break;
5772 case OMP_CLAUSE_COPYIN:
5773 case OMP_CLAUSE_COPYPRIVATE:
5774 decl = OMP_CLAUSE_DECL (c);
5775 if (decl == error_mark_node || TREE_TYPE (decl) == error_mark_node)
5777 remove = true;
5778 break;
5780 do_notice:
5781 if (outer_ctx)
5782 omp_notice_variable (outer_ctx, decl, true);
5783 if (check_non_private
5784 && region_type == ORT_WORKSHARE
5785 && omp_check_private (ctx, decl))
5787 error ("%s variable %qE is private in outer context",
5788 check_non_private, DECL_NAME (decl));
5789 remove = true;
5791 break;
5793 case OMP_CLAUSE_IF:
5794 OMP_CLAUSE_OPERAND (c, 0)
5795 = gimple_boolify (OMP_CLAUSE_OPERAND (c, 0));
5796 /* Fall through. */
5798 case OMP_CLAUSE_SCHEDULE:
5799 case OMP_CLAUSE_NUM_THREADS:
5800 if (gimplify_expr (&OMP_CLAUSE_OPERAND (c, 0), pre_p, NULL,
5801 is_gimple_val, fb_rvalue) == GS_ERROR)
5802 remove = true;
5803 break;
5805 case OMP_CLAUSE_NOWAIT:
5806 case OMP_CLAUSE_ORDERED:
5807 case OMP_CLAUSE_UNTIED:
5808 case OMP_CLAUSE_COLLAPSE:
5809 break;
5811 case OMP_CLAUSE_DEFAULT:
5812 ctx->default_kind = OMP_CLAUSE_DEFAULT_KIND (c);
5813 break;
5815 default:
5816 gcc_unreachable ();
5819 if (remove)
5820 *list_p = OMP_CLAUSE_CHAIN (c);
5821 else
5822 list_p = &OMP_CLAUSE_CHAIN (c);
5825 gimplify_omp_ctxp = ctx;
5828 /* For all variables that were not actually used within the context,
5829 remove PRIVATE, SHARED, and FIRSTPRIVATE clauses. */
5831 static int
5832 gimplify_adjust_omp_clauses_1 (splay_tree_node n, void *data)
5834 tree *list_p = (tree *) data;
5835 tree decl = (tree) n->key;
5836 unsigned flags = n->value;
5837 enum omp_clause_code code;
5838 tree clause;
5839 bool private_debug;
5841 if (flags & (GOVD_EXPLICIT | GOVD_LOCAL))
5842 return 0;
5843 if ((flags & GOVD_SEEN) == 0)
5844 return 0;
5845 if (flags & GOVD_DEBUG_PRIVATE)
5847 gcc_assert ((flags & GOVD_DATA_SHARE_CLASS) == GOVD_PRIVATE);
5848 private_debug = true;
5850 else
5851 private_debug
5852 = lang_hooks.decls.omp_private_debug_clause (decl,
5853 !!(flags & GOVD_SHARED));
5854 if (private_debug)
5855 code = OMP_CLAUSE_PRIVATE;
5856 else if (flags & GOVD_SHARED)
5858 if (is_global_var (decl))
5860 struct gimplify_omp_ctx *ctx = gimplify_omp_ctxp->outer_context;
5861 while (ctx != NULL)
5863 splay_tree_node on
5864 = splay_tree_lookup (ctx->variables, (splay_tree_key) decl);
5865 if (on && (on->value & (GOVD_FIRSTPRIVATE | GOVD_LASTPRIVATE
5866 | GOVD_PRIVATE | GOVD_REDUCTION)) != 0)
5867 break;
5868 ctx = ctx->outer_context;
5870 if (ctx == NULL)
5871 return 0;
5873 code = OMP_CLAUSE_SHARED;
5875 else if (flags & GOVD_PRIVATE)
5876 code = OMP_CLAUSE_PRIVATE;
5877 else if (flags & GOVD_FIRSTPRIVATE)
5878 code = OMP_CLAUSE_FIRSTPRIVATE;
5879 else
5880 gcc_unreachable ();
5882 clause = build_omp_clause (input_location, code);
5883 OMP_CLAUSE_DECL (clause) = decl;
5884 OMP_CLAUSE_CHAIN (clause) = *list_p;
5885 if (private_debug)
5886 OMP_CLAUSE_PRIVATE_DEBUG (clause) = 1;
5887 else if (code == OMP_CLAUSE_PRIVATE && (flags & GOVD_PRIVATE_OUTER_REF))
5888 OMP_CLAUSE_PRIVATE_OUTER_REF (clause) = 1;
5889 *list_p = clause;
5890 lang_hooks.decls.omp_finish_clause (clause);
5892 return 0;
5895 static void
5896 gimplify_adjust_omp_clauses (tree *list_p)
5898 struct gimplify_omp_ctx *ctx = gimplify_omp_ctxp;
5899 tree c, decl;
5901 while ((c = *list_p) != NULL)
5903 splay_tree_node n;
5904 bool remove = false;
5906 switch (OMP_CLAUSE_CODE (c))
5908 case OMP_CLAUSE_PRIVATE:
5909 case OMP_CLAUSE_SHARED:
5910 case OMP_CLAUSE_FIRSTPRIVATE:
5911 decl = OMP_CLAUSE_DECL (c);
5912 n = splay_tree_lookup (ctx->variables, (splay_tree_key) decl);
5913 remove = !(n->value & GOVD_SEEN);
5914 if (! remove)
5916 bool shared = OMP_CLAUSE_CODE (c) == OMP_CLAUSE_SHARED;
5917 if ((n->value & GOVD_DEBUG_PRIVATE)
5918 || lang_hooks.decls.omp_private_debug_clause (decl, shared))
5920 gcc_assert ((n->value & GOVD_DEBUG_PRIVATE) == 0
5921 || ((n->value & GOVD_DATA_SHARE_CLASS)
5922 == GOVD_PRIVATE));
5923 OMP_CLAUSE_SET_CODE (c, OMP_CLAUSE_PRIVATE);
5924 OMP_CLAUSE_PRIVATE_DEBUG (c) = 1;
5927 break;
5929 case OMP_CLAUSE_LASTPRIVATE:
5930 /* Make sure OMP_CLAUSE_LASTPRIVATE_FIRSTPRIVATE is set to
5931 accurately reflect the presence of a FIRSTPRIVATE clause. */
5932 decl = OMP_CLAUSE_DECL (c);
5933 n = splay_tree_lookup (ctx->variables, (splay_tree_key) decl);
5934 OMP_CLAUSE_LASTPRIVATE_FIRSTPRIVATE (c)
5935 = (n->value & GOVD_FIRSTPRIVATE) != 0;
5936 break;
5938 case OMP_CLAUSE_REDUCTION:
5939 case OMP_CLAUSE_COPYIN:
5940 case OMP_CLAUSE_COPYPRIVATE:
5941 case OMP_CLAUSE_IF:
5942 case OMP_CLAUSE_NUM_THREADS:
5943 case OMP_CLAUSE_SCHEDULE:
5944 case OMP_CLAUSE_NOWAIT:
5945 case OMP_CLAUSE_ORDERED:
5946 case OMP_CLAUSE_DEFAULT:
5947 case OMP_CLAUSE_UNTIED:
5948 case OMP_CLAUSE_COLLAPSE:
5949 break;
5951 default:
5952 gcc_unreachable ();
5955 if (remove)
5956 *list_p = OMP_CLAUSE_CHAIN (c);
5957 else
5958 list_p = &OMP_CLAUSE_CHAIN (c);
5961 /* Add in any implicit data sharing. */
5962 splay_tree_foreach (ctx->variables, gimplify_adjust_omp_clauses_1, list_p);
5964 gimplify_omp_ctxp = ctx->outer_context;
5965 delete_omp_context (ctx);
5968 /* Gimplify the contents of an OMP_PARALLEL statement. This involves
5969 gimplification of the body, as well as scanning the body for used
5970 variables. We need to do this scan now, because variable-sized
5971 decls will be decomposed during gimplification. */
5973 static void
5974 gimplify_omp_parallel (tree *expr_p, gimple_seq *pre_p)
5976 tree expr = *expr_p;
5977 gimple g;
5978 gimple_seq body = NULL;
5979 struct gimplify_ctx gctx;
5981 gimplify_scan_omp_clauses (&OMP_PARALLEL_CLAUSES (expr), pre_p,
5982 OMP_PARALLEL_COMBINED (expr)
5983 ? ORT_COMBINED_PARALLEL
5984 : ORT_PARALLEL);
5986 push_gimplify_context (&gctx);
5988 g = gimplify_and_return_first (OMP_PARALLEL_BODY (expr), &body);
5989 if (gimple_code (g) == GIMPLE_BIND)
5990 pop_gimplify_context (g);
5991 else
5992 pop_gimplify_context (NULL);
5994 gimplify_adjust_omp_clauses (&OMP_PARALLEL_CLAUSES (expr));
5996 g = gimple_build_omp_parallel (body,
5997 OMP_PARALLEL_CLAUSES (expr),
5998 NULL_TREE, NULL_TREE);
5999 if (OMP_PARALLEL_COMBINED (expr))
6000 gimple_omp_set_subcode (g, GF_OMP_PARALLEL_COMBINED);
6001 gimplify_seq_add_stmt (pre_p, g);
6002 *expr_p = NULL_TREE;
6005 /* Gimplify the contents of an OMP_TASK statement. This involves
6006 gimplification of the body, as well as scanning the body for used
6007 variables. We need to do this scan now, because variable-sized
6008 decls will be decomposed during gimplification. */
6010 static void
6011 gimplify_omp_task (tree *expr_p, gimple_seq *pre_p)
6013 tree expr = *expr_p;
6014 gimple g;
6015 gimple_seq body = NULL;
6016 struct gimplify_ctx gctx;
6018 gimplify_scan_omp_clauses (&OMP_TASK_CLAUSES (expr), pre_p, ORT_TASK);
6020 push_gimplify_context (&gctx);
6022 g = gimplify_and_return_first (OMP_TASK_BODY (expr), &body);
6023 if (gimple_code (g) == GIMPLE_BIND)
6024 pop_gimplify_context (g);
6025 else
6026 pop_gimplify_context (NULL);
6028 gimplify_adjust_omp_clauses (&OMP_TASK_CLAUSES (expr));
6030 g = gimple_build_omp_task (body,
6031 OMP_TASK_CLAUSES (expr),
6032 NULL_TREE, NULL_TREE,
6033 NULL_TREE, NULL_TREE, NULL_TREE);
6034 gimplify_seq_add_stmt (pre_p, g);
6035 *expr_p = NULL_TREE;
6038 /* Gimplify the gross structure of an OMP_FOR statement. */
6040 static enum gimplify_status
6041 gimplify_omp_for (tree *expr_p, gimple_seq *pre_p)
6043 tree for_stmt, decl, var, t;
6044 enum gimplify_status ret = GS_ALL_DONE;
6045 enum gimplify_status tret;
6046 gimple gfor;
6047 gimple_seq for_body, for_pre_body;
6048 int i;
6050 for_stmt = *expr_p;
6052 gimplify_scan_omp_clauses (&OMP_FOR_CLAUSES (for_stmt), pre_p,
6053 ORT_WORKSHARE);
6055 /* Handle OMP_FOR_INIT. */
6056 for_pre_body = NULL;
6057 gimplify_and_add (OMP_FOR_PRE_BODY (for_stmt), &for_pre_body);
6058 OMP_FOR_PRE_BODY (for_stmt) = NULL_TREE;
6060 for_body = gimple_seq_alloc ();
6061 gcc_assert (TREE_VEC_LENGTH (OMP_FOR_INIT (for_stmt))
6062 == TREE_VEC_LENGTH (OMP_FOR_COND (for_stmt)));
6063 gcc_assert (TREE_VEC_LENGTH (OMP_FOR_INIT (for_stmt))
6064 == TREE_VEC_LENGTH (OMP_FOR_INCR (for_stmt)));
6065 for (i = 0; i < TREE_VEC_LENGTH (OMP_FOR_INIT (for_stmt)); i++)
6067 t = TREE_VEC_ELT (OMP_FOR_INIT (for_stmt), i);
6068 gcc_assert (TREE_CODE (t) == MODIFY_EXPR);
6069 decl = TREE_OPERAND (t, 0);
6070 gcc_assert (DECL_P (decl));
6071 gcc_assert (INTEGRAL_TYPE_P (TREE_TYPE (decl))
6072 || POINTER_TYPE_P (TREE_TYPE (decl)));
6074 /* Make sure the iteration variable is private. */
6075 if (omp_is_private (gimplify_omp_ctxp, decl))
6076 omp_notice_variable (gimplify_omp_ctxp, decl, true);
6077 else
6078 omp_add_variable (gimplify_omp_ctxp, decl, GOVD_PRIVATE | GOVD_SEEN);
6080 /* If DECL is not a gimple register, create a temporary variable to act
6081 as an iteration counter. This is valid, since DECL cannot be
6082 modified in the body of the loop. */
6083 if (!is_gimple_reg (decl))
6085 var = create_tmp_var (TREE_TYPE (decl), get_name (decl));
6086 TREE_OPERAND (t, 0) = var;
6088 gimplify_seq_add_stmt (&for_body, gimple_build_assign (decl, var));
6090 omp_add_variable (gimplify_omp_ctxp, var, GOVD_PRIVATE | GOVD_SEEN);
6092 else
6093 var = decl;
6095 tret = gimplify_expr (&TREE_OPERAND (t, 1), &for_pre_body, NULL,
6096 is_gimple_val, fb_rvalue);
6097 ret = MIN (ret, tret);
6098 if (ret == GS_ERROR)
6099 return ret;
6101 /* Handle OMP_FOR_COND. */
6102 t = TREE_VEC_ELT (OMP_FOR_COND (for_stmt), i);
6103 gcc_assert (COMPARISON_CLASS_P (t));
6104 gcc_assert (TREE_OPERAND (t, 0) == decl);
6106 tret = gimplify_expr (&TREE_OPERAND (t, 1), &for_pre_body, NULL,
6107 is_gimple_val, fb_rvalue);
6108 ret = MIN (ret, tret);
6110 /* Handle OMP_FOR_INCR. */
6111 t = TREE_VEC_ELT (OMP_FOR_INCR (for_stmt), i);
6112 switch (TREE_CODE (t))
6114 case PREINCREMENT_EXPR:
6115 case POSTINCREMENT_EXPR:
6116 t = build_int_cst (TREE_TYPE (decl), 1);
6117 t = build2 (PLUS_EXPR, TREE_TYPE (decl), var, t);
6118 t = build2 (MODIFY_EXPR, TREE_TYPE (var), var, t);
6119 TREE_VEC_ELT (OMP_FOR_INCR (for_stmt), i) = t;
6120 break;
6122 case PREDECREMENT_EXPR:
6123 case POSTDECREMENT_EXPR:
6124 t = build_int_cst (TREE_TYPE (decl), -1);
6125 t = build2 (PLUS_EXPR, TREE_TYPE (decl), var, t);
6126 t = build2 (MODIFY_EXPR, TREE_TYPE (var), var, t);
6127 TREE_VEC_ELT (OMP_FOR_INCR (for_stmt), i) = t;
6128 break;
6130 case MODIFY_EXPR:
6131 gcc_assert (TREE_OPERAND (t, 0) == decl);
6132 TREE_OPERAND (t, 0) = var;
6134 t = TREE_OPERAND (t, 1);
6135 switch (TREE_CODE (t))
6137 case PLUS_EXPR:
6138 if (TREE_OPERAND (t, 1) == decl)
6140 TREE_OPERAND (t, 1) = TREE_OPERAND (t, 0);
6141 TREE_OPERAND (t, 0) = var;
6142 break;
6145 /* Fallthru. */
6146 case MINUS_EXPR:
6147 case POINTER_PLUS_EXPR:
6148 gcc_assert (TREE_OPERAND (t, 0) == decl);
6149 TREE_OPERAND (t, 0) = var;
6150 break;
6151 default:
6152 gcc_unreachable ();
6155 tret = gimplify_expr (&TREE_OPERAND (t, 1), &for_pre_body, NULL,
6156 is_gimple_val, fb_rvalue);
6157 ret = MIN (ret, tret);
6158 break;
6160 default:
6161 gcc_unreachable ();
6164 if (var != decl || TREE_VEC_LENGTH (OMP_FOR_INIT (for_stmt)) > 1)
6166 tree c;
6167 for (c = OMP_FOR_CLAUSES (for_stmt); c ; c = OMP_CLAUSE_CHAIN (c))
6168 if (OMP_CLAUSE_CODE (c) == OMP_CLAUSE_LASTPRIVATE
6169 && OMP_CLAUSE_DECL (c) == decl
6170 && OMP_CLAUSE_LASTPRIVATE_GIMPLE_SEQ (c) == NULL)
6172 t = TREE_VEC_ELT (OMP_FOR_INCR (for_stmt), i);
6173 gcc_assert (TREE_CODE (t) == MODIFY_EXPR);
6174 gcc_assert (TREE_OPERAND (t, 0) == var);
6175 t = TREE_OPERAND (t, 1);
6176 gcc_assert (TREE_CODE (t) == PLUS_EXPR
6177 || TREE_CODE (t) == MINUS_EXPR
6178 || TREE_CODE (t) == POINTER_PLUS_EXPR);
6179 gcc_assert (TREE_OPERAND (t, 0) == var);
6180 t = build2 (TREE_CODE (t), TREE_TYPE (decl), decl,
6181 TREE_OPERAND (t, 1));
6182 gimplify_assign (decl, t,
6183 &OMP_CLAUSE_LASTPRIVATE_GIMPLE_SEQ (c));
6188 gimplify_and_add (OMP_FOR_BODY (for_stmt), &for_body);
6190 gimplify_adjust_omp_clauses (&OMP_FOR_CLAUSES (for_stmt));
6192 gfor = gimple_build_omp_for (for_body, OMP_FOR_CLAUSES (for_stmt),
6193 TREE_VEC_LENGTH (OMP_FOR_INIT (for_stmt)),
6194 for_pre_body);
6196 for (i = 0; i < TREE_VEC_LENGTH (OMP_FOR_INIT (for_stmt)); i++)
6198 t = TREE_VEC_ELT (OMP_FOR_INIT (for_stmt), i);
6199 gimple_omp_for_set_index (gfor, i, TREE_OPERAND (t, 0));
6200 gimple_omp_for_set_initial (gfor, i, TREE_OPERAND (t, 1));
6201 t = TREE_VEC_ELT (OMP_FOR_COND (for_stmt), i);
6202 gimple_omp_for_set_cond (gfor, i, TREE_CODE (t));
6203 gimple_omp_for_set_final (gfor, i, TREE_OPERAND (t, 1));
6204 t = TREE_VEC_ELT (OMP_FOR_INCR (for_stmt), i);
6205 gimple_omp_for_set_incr (gfor, i, TREE_OPERAND (t, 1));
6208 gimplify_seq_add_stmt (pre_p, gfor);
6209 return ret == GS_ALL_DONE ? GS_ALL_DONE : GS_ERROR;
6212 /* Gimplify the gross structure of other OpenMP worksharing constructs.
6213 In particular, OMP_SECTIONS and OMP_SINGLE. */
6215 static void
6216 gimplify_omp_workshare (tree *expr_p, gimple_seq *pre_p)
6218 tree expr = *expr_p;
6219 gimple stmt;
6220 gimple_seq body = NULL;
6222 gimplify_scan_omp_clauses (&OMP_CLAUSES (expr), pre_p, ORT_WORKSHARE);
6223 gimplify_and_add (OMP_BODY (expr), &body);
6224 gimplify_adjust_omp_clauses (&OMP_CLAUSES (expr));
6226 if (TREE_CODE (expr) == OMP_SECTIONS)
6227 stmt = gimple_build_omp_sections (body, OMP_CLAUSES (expr));
6228 else if (TREE_CODE (expr) == OMP_SINGLE)
6229 stmt = gimple_build_omp_single (body, OMP_CLAUSES (expr));
6230 else
6231 gcc_unreachable ();
6233 gimplify_seq_add_stmt (pre_p, stmt);
6236 /* A subroutine of gimplify_omp_atomic. The front end is supposed to have
6237 stabilized the lhs of the atomic operation as *ADDR. Return true if
6238 EXPR is this stabilized form. */
6240 static bool
6241 goa_lhs_expr_p (tree expr, tree addr)
6243 /* Also include casts to other type variants. The C front end is fond
6244 of adding these for e.g. volatile variables. This is like
6245 STRIP_TYPE_NOPS but includes the main variant lookup. */
6246 STRIP_USELESS_TYPE_CONVERSION (expr);
6248 if (TREE_CODE (expr) == INDIRECT_REF)
6250 expr = TREE_OPERAND (expr, 0);
6251 while (expr != addr
6252 && (CONVERT_EXPR_P (expr)
6253 || TREE_CODE (expr) == NON_LVALUE_EXPR)
6254 && TREE_CODE (expr) == TREE_CODE (addr)
6255 && types_compatible_p (TREE_TYPE (expr), TREE_TYPE (addr)))
6257 expr = TREE_OPERAND (expr, 0);
6258 addr = TREE_OPERAND (addr, 0);
6260 if (expr == addr)
6261 return true;
6262 return (TREE_CODE (addr) == ADDR_EXPR
6263 && TREE_CODE (expr) == ADDR_EXPR
6264 && TREE_OPERAND (addr, 0) == TREE_OPERAND (expr, 0));
6266 if (TREE_CODE (addr) == ADDR_EXPR && expr == TREE_OPERAND (addr, 0))
6267 return true;
6268 return false;
6271 /* Walk *EXPR_P and replace
6272 appearances of *LHS_ADDR with LHS_VAR. If an expression does not involve
6273 the lhs, evaluate it into a temporary. Return 1 if the lhs appeared as
6274 a subexpression, 0 if it did not, or -1 if an error was encountered. */
6276 static int
6277 goa_stabilize_expr (tree *expr_p, gimple_seq *pre_p, tree lhs_addr,
6278 tree lhs_var)
6280 tree expr = *expr_p;
6281 int saw_lhs;
6283 if (goa_lhs_expr_p (expr, lhs_addr))
6285 *expr_p = lhs_var;
6286 return 1;
6288 if (is_gimple_val (expr))
6289 return 0;
6291 saw_lhs = 0;
6292 switch (TREE_CODE_CLASS (TREE_CODE (expr)))
6294 case tcc_binary:
6295 case tcc_comparison:
6296 saw_lhs |= goa_stabilize_expr (&TREE_OPERAND (expr, 1), pre_p, lhs_addr,
6297 lhs_var);
6298 case tcc_unary:
6299 saw_lhs |= goa_stabilize_expr (&TREE_OPERAND (expr, 0), pre_p, lhs_addr,
6300 lhs_var);
6301 break;
6302 case tcc_expression:
6303 switch (TREE_CODE (expr))
6305 case TRUTH_ANDIF_EXPR:
6306 case TRUTH_ORIF_EXPR:
6307 saw_lhs |= goa_stabilize_expr (&TREE_OPERAND (expr, 1), pre_p,
6308 lhs_addr, lhs_var);
6309 saw_lhs |= goa_stabilize_expr (&TREE_OPERAND (expr, 0), pre_p,
6310 lhs_addr, lhs_var);
6311 break;
6312 default:
6313 break;
6315 break;
6316 default:
6317 break;
6320 if (saw_lhs == 0)
6322 enum gimplify_status gs;
6323 gs = gimplify_expr (expr_p, pre_p, NULL, is_gimple_val, fb_rvalue);
6324 if (gs != GS_ALL_DONE)
6325 saw_lhs = -1;
6328 return saw_lhs;
6332 /* Gimplify an OMP_ATOMIC statement. */
6334 static enum gimplify_status
6335 gimplify_omp_atomic (tree *expr_p, gimple_seq *pre_p)
6337 tree addr = TREE_OPERAND (*expr_p, 0);
6338 tree rhs = TREE_OPERAND (*expr_p, 1);
6339 tree type = TYPE_MAIN_VARIANT (TREE_TYPE (TREE_TYPE (addr)));
6340 tree tmp_load;
6342 tmp_load = create_tmp_var (type, NULL);
6343 if (TREE_CODE (type) == COMPLEX_TYPE || TREE_CODE (type) == VECTOR_TYPE)
6344 DECL_GIMPLE_REG_P (tmp_load) = 1;
6345 if (goa_stabilize_expr (&rhs, pre_p, addr, tmp_load) < 0)
6346 return GS_ERROR;
6348 if (gimplify_expr (&addr, pre_p, NULL, is_gimple_val, fb_rvalue)
6349 != GS_ALL_DONE)
6350 return GS_ERROR;
6352 gimplify_seq_add_stmt (pre_p, gimple_build_omp_atomic_load (tmp_load, addr));
6353 if (gimplify_expr (&rhs, pre_p, NULL, is_gimple_val, fb_rvalue)
6354 != GS_ALL_DONE)
6355 return GS_ERROR;
6356 gimplify_seq_add_stmt (pre_p, gimple_build_omp_atomic_store (rhs));
6357 *expr_p = NULL;
6359 return GS_ALL_DONE;
6363 /* Converts the GENERIC expression tree *EXPR_P to GIMPLE. If the
6364 expression produces a value to be used as an operand inside a GIMPLE
6365 statement, the value will be stored back in *EXPR_P. This value will
6366 be a tree of class tcc_declaration, tcc_constant, tcc_reference or
6367 an SSA_NAME. The corresponding sequence of GIMPLE statements is
6368 emitted in PRE_P and POST_P.
6370 Additionally, this process may overwrite parts of the input
6371 expression during gimplification. Ideally, it should be
6372 possible to do non-destructive gimplification.
6374 EXPR_P points to the GENERIC expression to convert to GIMPLE. If
6375 the expression needs to evaluate to a value to be used as
6376 an operand in a GIMPLE statement, this value will be stored in
6377 *EXPR_P on exit. This happens when the caller specifies one
6378 of fb_lvalue or fb_rvalue fallback flags.
6380 PRE_P will contain the sequence of GIMPLE statements corresponding
6381 to the evaluation of EXPR and all the side-effects that must
6382 be executed before the main expression. On exit, the last
6383 statement of PRE_P is the core statement being gimplified. For
6384 instance, when gimplifying 'if (++a)' the last statement in
6385 PRE_P will be 'if (t.1)' where t.1 is the result of
6386 pre-incrementing 'a'.
6388 POST_P will contain the sequence of GIMPLE statements corresponding
6389 to the evaluation of all the side-effects that must be executed
6390 after the main expression. If this is NULL, the post
6391 side-effects are stored at the end of PRE_P.
6393 The reason why the output is split in two is to handle post
6394 side-effects explicitly. In some cases, an expression may have
6395 inner and outer post side-effects which need to be emitted in
6396 an order different from the one given by the recursive
6397 traversal. For instance, for the expression (*p--)++ the post
6398 side-effects of '--' must actually occur *after* the post
6399 side-effects of '++'. However, gimplification will first visit
6400 the inner expression, so if a separate POST sequence was not
6401 used, the resulting sequence would be:
6403 1 t.1 = *p
6404 2 p = p - 1
6405 3 t.2 = t.1 + 1
6406 4 *p = t.2
6408 However, the post-decrement operation in line #2 must not be
6409 evaluated until after the store to *p at line #4, so the
6410 correct sequence should be:
6412 1 t.1 = *p
6413 2 t.2 = t.1 + 1
6414 3 *p = t.2
6415 4 p = p - 1
6417 So, by specifying a separate post queue, it is possible
6418 to emit the post side-effects in the correct order.
6419 If POST_P is NULL, an internal queue will be used. Before
6420 returning to the caller, the sequence POST_P is appended to
6421 the main output sequence PRE_P.
6423 GIMPLE_TEST_F points to a function that takes a tree T and
6424 returns nonzero if T is in the GIMPLE form requested by the
6425 caller. The GIMPLE predicates are in tree-gimple.c.
6427 FALLBACK tells the function what sort of a temporary we want if
6428 gimplification cannot produce an expression that complies with
6429 GIMPLE_TEST_F.
6431 fb_none means that no temporary should be generated
6432 fb_rvalue means that an rvalue is OK to generate
6433 fb_lvalue means that an lvalue is OK to generate
6434 fb_either means that either is OK, but an lvalue is preferable.
6435 fb_mayfail means that gimplification may fail (in which case
6436 GS_ERROR will be returned)
6438 The return value is either GS_ERROR or GS_ALL_DONE, since this
6439 function iterates until EXPR is completely gimplified or an error
6440 occurs. */
6442 enum gimplify_status
6443 gimplify_expr (tree *expr_p, gimple_seq *pre_p, gimple_seq *post_p,
6444 bool (*gimple_test_f) (tree), fallback_t fallback)
6446 tree tmp;
6447 gimple_seq internal_pre = NULL;
6448 gimple_seq internal_post = NULL;
6449 tree save_expr;
6450 bool is_statement;
6451 location_t saved_location;
6452 enum gimplify_status ret;
6453 gimple_stmt_iterator pre_last_gsi, post_last_gsi;
6455 save_expr = *expr_p;
6456 if (save_expr == NULL_TREE)
6457 return GS_ALL_DONE;
6459 /* If we are gimplifying a top-level statement, PRE_P must be valid. */
6460 is_statement = gimple_test_f == is_gimple_stmt;
6461 if (is_statement)
6462 gcc_assert (pre_p);
6464 /* Consistency checks. */
6465 if (gimple_test_f == is_gimple_reg)
6466 gcc_assert (fallback & (fb_rvalue | fb_lvalue));
6467 else if (gimple_test_f == is_gimple_val
6468 || gimple_test_f == is_gimple_call_addr
6469 || gimple_test_f == is_gimple_condexpr
6470 || gimple_test_f == is_gimple_mem_rhs
6471 || gimple_test_f == is_gimple_mem_rhs_or_call
6472 || gimple_test_f == is_gimple_reg_rhs
6473 || gimple_test_f == is_gimple_reg_rhs_or_call
6474 || gimple_test_f == is_gimple_asm_val)
6475 gcc_assert (fallback & fb_rvalue);
6476 else if (gimple_test_f == is_gimple_min_lval
6477 || gimple_test_f == is_gimple_lvalue)
6478 gcc_assert (fallback & fb_lvalue);
6479 else if (gimple_test_f == is_gimple_addressable)
6480 gcc_assert (fallback & fb_either);
6481 else if (gimple_test_f == is_gimple_stmt)
6482 gcc_assert (fallback == fb_none);
6483 else
6485 /* We should have recognized the GIMPLE_TEST_F predicate to
6486 know what kind of fallback to use in case a temporary is
6487 needed to hold the value or address of *EXPR_P. */
6488 gcc_unreachable ();
6491 /* We used to check the predicate here and return immediately if it
6492 succeeds. This is wrong; the design is for gimplification to be
6493 idempotent, and for the predicates to only test for valid forms, not
6494 whether they are fully simplified. */
6495 if (pre_p == NULL)
6496 pre_p = &internal_pre;
6498 if (post_p == NULL)
6499 post_p = &internal_post;
6501 /* Remember the last statements added to PRE_P and POST_P. Every
6502 new statement added by the gimplification helpers needs to be
6503 annotated with location information. To centralize the
6504 responsibility, we remember the last statement that had been
6505 added to both queues before gimplifying *EXPR_P. If
6506 gimplification produces new statements in PRE_P and POST_P, those
6507 statements will be annotated with the same location information
6508 as *EXPR_P. */
6509 pre_last_gsi = gsi_last (*pre_p);
6510 post_last_gsi = gsi_last (*post_p);
6512 saved_location = input_location;
6513 if (save_expr != error_mark_node
6514 && EXPR_HAS_LOCATION (*expr_p))
6515 input_location = EXPR_LOCATION (*expr_p);
6517 /* Loop over the specific gimplifiers until the toplevel node
6518 remains the same. */
6521 /* Strip away as many useless type conversions as possible
6522 at the toplevel. */
6523 STRIP_USELESS_TYPE_CONVERSION (*expr_p);
6525 /* Remember the expr. */
6526 save_expr = *expr_p;
6528 /* Die, die, die, my darling. */
6529 if (save_expr == error_mark_node
6530 || (TREE_TYPE (save_expr)
6531 && TREE_TYPE (save_expr) == error_mark_node))
6533 ret = GS_ERROR;
6534 break;
6537 /* Do any language-specific gimplification. */
6538 ret = ((enum gimplify_status)
6539 lang_hooks.gimplify_expr (expr_p, pre_p, post_p));
6540 if (ret == GS_OK)
6542 if (*expr_p == NULL_TREE)
6543 break;
6544 if (*expr_p != save_expr)
6545 continue;
6547 else if (ret != GS_UNHANDLED)
6548 break;
6550 ret = GS_OK;
6551 switch (TREE_CODE (*expr_p))
6553 /* First deal with the special cases. */
6555 case POSTINCREMENT_EXPR:
6556 case POSTDECREMENT_EXPR:
6557 case PREINCREMENT_EXPR:
6558 case PREDECREMENT_EXPR:
6559 ret = gimplify_self_mod_expr (expr_p, pre_p, post_p,
6560 fallback != fb_none);
6561 break;
6563 case ARRAY_REF:
6564 case ARRAY_RANGE_REF:
6565 case REALPART_EXPR:
6566 case IMAGPART_EXPR:
6567 case COMPONENT_REF:
6568 case VIEW_CONVERT_EXPR:
6569 ret = gimplify_compound_lval (expr_p, pre_p, post_p,
6570 fallback ? fallback : fb_rvalue);
6571 break;
6573 case COND_EXPR:
6574 ret = gimplify_cond_expr (expr_p, pre_p, fallback);
6576 /* C99 code may assign to an array in a structure value of a
6577 conditional expression, and this has undefined behavior
6578 only on execution, so create a temporary if an lvalue is
6579 required. */
6580 if (fallback == fb_lvalue)
6582 *expr_p = get_initialized_tmp_var (*expr_p, pre_p, post_p);
6583 mark_addressable (*expr_p);
6585 break;
6587 case CALL_EXPR:
6588 ret = gimplify_call_expr (expr_p, pre_p, fallback != fb_none);
6590 /* C99 code may assign to an array in a structure returned
6591 from a function, and this has undefined behavior only on
6592 execution, so create a temporary if an lvalue is
6593 required. */
6594 if (fallback == fb_lvalue)
6596 *expr_p = get_initialized_tmp_var (*expr_p, pre_p, post_p);
6597 mark_addressable (*expr_p);
6599 break;
6601 case TREE_LIST:
6602 gcc_unreachable ();
6604 case COMPOUND_EXPR:
6605 ret = gimplify_compound_expr (expr_p, pre_p, fallback != fb_none);
6606 break;
6608 case COMPOUND_LITERAL_EXPR:
6609 ret = gimplify_compound_literal_expr (expr_p, pre_p);
6610 break;
6612 case MODIFY_EXPR:
6613 case INIT_EXPR:
6614 ret = gimplify_modify_expr (expr_p, pre_p, post_p,
6615 fallback != fb_none);
6616 break;
6618 case TRUTH_ANDIF_EXPR:
6619 case TRUTH_ORIF_EXPR:
6620 /* Pass the source location of the outer expression. */
6621 ret = gimplify_boolean_expr (expr_p, saved_location);
6622 break;
6624 case TRUTH_NOT_EXPR:
6625 if (TREE_CODE (TREE_TYPE (*expr_p)) != BOOLEAN_TYPE)
6627 tree type = TREE_TYPE (*expr_p);
6628 *expr_p = fold_convert (type, gimple_boolify (*expr_p));
6629 ret = GS_OK;
6630 break;
6633 ret = gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p, post_p,
6634 is_gimple_val, fb_rvalue);
6635 recalculate_side_effects (*expr_p);
6636 break;
6638 case ADDR_EXPR:
6639 ret = gimplify_addr_expr (expr_p, pre_p, post_p);
6640 break;
6642 case VA_ARG_EXPR:
6643 ret = gimplify_va_arg_expr (expr_p, pre_p, post_p);
6644 break;
6646 CASE_CONVERT:
6647 if (IS_EMPTY_STMT (*expr_p))
6649 ret = GS_ALL_DONE;
6650 break;
6653 if (VOID_TYPE_P (TREE_TYPE (*expr_p))
6654 || fallback == fb_none)
6656 /* Just strip a conversion to void (or in void context) and
6657 try again. */
6658 *expr_p = TREE_OPERAND (*expr_p, 0);
6659 break;
6662 ret = gimplify_conversion (expr_p);
6663 if (ret == GS_ERROR)
6664 break;
6665 if (*expr_p != save_expr)
6666 break;
6667 /* FALLTHRU */
6669 case FIX_TRUNC_EXPR:
6670 /* unary_expr: ... | '(' cast ')' val | ... */
6671 ret = gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p, post_p,
6672 is_gimple_val, fb_rvalue);
6673 recalculate_side_effects (*expr_p);
6674 break;
6676 case INDIRECT_REF:
6677 *expr_p = fold_indirect_ref_loc (input_location, *expr_p);
6678 if (*expr_p != save_expr)
6679 break;
6680 /* else fall through. */
6681 case ALIGN_INDIRECT_REF:
6682 case MISALIGNED_INDIRECT_REF:
6683 ret = gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p, post_p,
6684 is_gimple_reg, fb_rvalue);
6685 recalculate_side_effects (*expr_p);
6686 break;
6688 /* Constants need not be gimplified. */
6689 case INTEGER_CST:
6690 case REAL_CST:
6691 case FIXED_CST:
6692 case STRING_CST:
6693 case COMPLEX_CST:
6694 case VECTOR_CST:
6695 ret = GS_ALL_DONE;
6696 break;
6698 case CONST_DECL:
6699 /* If we require an lvalue, such as for ADDR_EXPR, retain the
6700 CONST_DECL node. Otherwise the decl is replaceable by its
6701 value. */
6702 /* ??? Should be == fb_lvalue, but ADDR_EXPR passes fb_either. */
6703 if (fallback & fb_lvalue)
6704 ret = GS_ALL_DONE;
6705 else
6706 *expr_p = DECL_INITIAL (*expr_p);
6707 break;
6709 case DECL_EXPR:
6710 ret = gimplify_decl_expr (expr_p, pre_p);
6711 break;
6713 case BIND_EXPR:
6714 ret = gimplify_bind_expr (expr_p, pre_p);
6715 break;
6717 case LOOP_EXPR:
6718 ret = gimplify_loop_expr (expr_p, pre_p);
6719 break;
6721 case SWITCH_EXPR:
6722 ret = gimplify_switch_expr (expr_p, pre_p);
6723 break;
6725 case EXIT_EXPR:
6726 ret = gimplify_exit_expr (expr_p);
6727 break;
6729 case GOTO_EXPR:
6730 /* If the target is not LABEL, then it is a computed jump
6731 and the target needs to be gimplified. */
6732 if (TREE_CODE (GOTO_DESTINATION (*expr_p)) != LABEL_DECL)
6734 ret = gimplify_expr (&GOTO_DESTINATION (*expr_p), pre_p,
6735 NULL, is_gimple_val, fb_rvalue);
6736 if (ret == GS_ERROR)
6737 break;
6739 gimplify_seq_add_stmt (pre_p,
6740 gimple_build_goto (GOTO_DESTINATION (*expr_p)));
6741 break;
6743 case PREDICT_EXPR:
6744 gimplify_seq_add_stmt (pre_p,
6745 gimple_build_predict (PREDICT_EXPR_PREDICTOR (*expr_p),
6746 PREDICT_EXPR_OUTCOME (*expr_p)));
6747 ret = GS_ALL_DONE;
6748 break;
6750 case LABEL_EXPR:
6751 ret = GS_ALL_DONE;
6752 gcc_assert (decl_function_context (LABEL_EXPR_LABEL (*expr_p))
6753 == current_function_decl);
6754 gimplify_seq_add_stmt (pre_p,
6755 gimple_build_label (LABEL_EXPR_LABEL (*expr_p)));
6756 break;
6758 case CASE_LABEL_EXPR:
6759 ret = gimplify_case_label_expr (expr_p, pre_p);
6760 break;
6762 case RETURN_EXPR:
6763 ret = gimplify_return_expr (*expr_p, pre_p);
6764 break;
6766 case CONSTRUCTOR:
6767 /* Don't reduce this in place; let gimplify_init_constructor work its
6768 magic. Buf if we're just elaborating this for side effects, just
6769 gimplify any element that has side-effects. */
6770 if (fallback == fb_none)
6772 unsigned HOST_WIDE_INT ix;
6773 constructor_elt *ce;
6774 tree temp = NULL_TREE;
6775 for (ix = 0;
6776 VEC_iterate (constructor_elt, CONSTRUCTOR_ELTS (*expr_p),
6777 ix, ce);
6778 ix++)
6779 if (TREE_SIDE_EFFECTS (ce->value))
6780 append_to_statement_list (ce->value, &temp);
6782 *expr_p = temp;
6783 ret = GS_OK;
6785 /* C99 code may assign to an array in a constructed
6786 structure or union, and this has undefined behavior only
6787 on execution, so create a temporary if an lvalue is
6788 required. */
6789 else if (fallback == fb_lvalue)
6791 *expr_p = get_initialized_tmp_var (*expr_p, pre_p, post_p);
6792 mark_addressable (*expr_p);
6794 else
6795 ret = GS_ALL_DONE;
6796 break;
6798 /* The following are special cases that are not handled by the
6799 original GIMPLE grammar. */
6801 /* SAVE_EXPR nodes are converted into a GIMPLE identifier and
6802 eliminated. */
6803 case SAVE_EXPR:
6804 ret = gimplify_save_expr (expr_p, pre_p, post_p);
6805 break;
6807 case BIT_FIELD_REF:
6809 enum gimplify_status r0, r1, r2;
6811 r0 = gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p,
6812 post_p, is_gimple_lvalue, fb_either);
6813 r1 = gimplify_expr (&TREE_OPERAND (*expr_p, 1), pre_p,
6814 post_p, is_gimple_val, fb_rvalue);
6815 r2 = gimplify_expr (&TREE_OPERAND (*expr_p, 2), pre_p,
6816 post_p, is_gimple_val, fb_rvalue);
6817 recalculate_side_effects (*expr_p);
6819 ret = MIN (r0, MIN (r1, r2));
6821 break;
6823 case TARGET_MEM_REF:
6825 enum gimplify_status r0 = GS_ALL_DONE, r1 = GS_ALL_DONE;
6827 if (TMR_SYMBOL (*expr_p))
6828 r0 = gimplify_expr (&TMR_SYMBOL (*expr_p), pre_p,
6829 post_p, is_gimple_lvalue, fb_either);
6830 else if (TMR_BASE (*expr_p))
6831 r0 = gimplify_expr (&TMR_BASE (*expr_p), pre_p,
6832 post_p, is_gimple_val, fb_either);
6833 if (TMR_INDEX (*expr_p))
6834 r1 = gimplify_expr (&TMR_INDEX (*expr_p), pre_p,
6835 post_p, is_gimple_val, fb_rvalue);
6836 /* TMR_STEP and TMR_OFFSET are always integer constants. */
6837 ret = MIN (r0, r1);
6839 break;
6841 case NON_LVALUE_EXPR:
6842 /* This should have been stripped above. */
6843 gcc_unreachable ();
6845 case ASM_EXPR:
6846 ret = gimplify_asm_expr (expr_p, pre_p, post_p);
6847 break;
6849 case TRY_FINALLY_EXPR:
6850 case TRY_CATCH_EXPR:
6852 gimple_seq eval, cleanup;
6853 gimple try_;
6855 eval = cleanup = NULL;
6856 gimplify_and_add (TREE_OPERAND (*expr_p, 0), &eval);
6857 gimplify_and_add (TREE_OPERAND (*expr_p, 1), &cleanup);
6858 /* Don't create bogus GIMPLE_TRY with empty cleanup. */
6859 if (gimple_seq_empty_p (cleanup))
6861 gimple_seq_add_seq (pre_p, eval);
6862 ret = GS_ALL_DONE;
6863 break;
6865 try_ = gimple_build_try (eval, cleanup,
6866 TREE_CODE (*expr_p) == TRY_FINALLY_EXPR
6867 ? GIMPLE_TRY_FINALLY
6868 : GIMPLE_TRY_CATCH);
6869 if (TREE_CODE (*expr_p) == TRY_CATCH_EXPR)
6870 gimple_try_set_catch_is_cleanup (try_,
6871 TRY_CATCH_IS_CLEANUP (*expr_p));
6872 gimplify_seq_add_stmt (pre_p, try_);
6873 ret = GS_ALL_DONE;
6874 break;
6877 case CLEANUP_POINT_EXPR:
6878 ret = gimplify_cleanup_point_expr (expr_p, pre_p);
6879 break;
6881 case TARGET_EXPR:
6882 ret = gimplify_target_expr (expr_p, pre_p, post_p);
6883 break;
6885 case CATCH_EXPR:
6887 gimple c;
6888 gimple_seq handler = NULL;
6889 gimplify_and_add (CATCH_BODY (*expr_p), &handler);
6890 c = gimple_build_catch (CATCH_TYPES (*expr_p), handler);
6891 gimplify_seq_add_stmt (pre_p, c);
6892 ret = GS_ALL_DONE;
6893 break;
6896 case EH_FILTER_EXPR:
6898 gimple ehf;
6899 gimple_seq failure = NULL;
6901 gimplify_and_add (EH_FILTER_FAILURE (*expr_p), &failure);
6902 ehf = gimple_build_eh_filter (EH_FILTER_TYPES (*expr_p), failure);
6903 gimple_set_no_warning (ehf, TREE_NO_WARNING (*expr_p));
6904 gimplify_seq_add_stmt (pre_p, ehf);
6905 ret = GS_ALL_DONE;
6906 break;
6909 case OBJ_TYPE_REF:
6911 enum gimplify_status r0, r1;
6912 r0 = gimplify_expr (&OBJ_TYPE_REF_OBJECT (*expr_p), pre_p,
6913 post_p, is_gimple_val, fb_rvalue);
6914 r1 = gimplify_expr (&OBJ_TYPE_REF_EXPR (*expr_p), pre_p,
6915 post_p, is_gimple_val, fb_rvalue);
6916 TREE_SIDE_EFFECTS (*expr_p) = 0;
6917 ret = MIN (r0, r1);
6919 break;
6921 case LABEL_DECL:
6922 /* We get here when taking the address of a label. We mark
6923 the label as "forced"; meaning it can never be removed and
6924 it is a potential target for any computed goto. */
6925 FORCED_LABEL (*expr_p) = 1;
6926 ret = GS_ALL_DONE;
6927 break;
6929 case STATEMENT_LIST:
6930 ret = gimplify_statement_list (expr_p, pre_p);
6931 break;
6933 case WITH_SIZE_EXPR:
6935 gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p,
6936 post_p == &internal_post ? NULL : post_p,
6937 gimple_test_f, fallback);
6938 gimplify_expr (&TREE_OPERAND (*expr_p, 1), pre_p, post_p,
6939 is_gimple_val, fb_rvalue);
6941 break;
6943 case VAR_DECL:
6944 case PARM_DECL:
6945 ret = gimplify_var_or_parm_decl (expr_p);
6946 break;
6948 case RESULT_DECL:
6949 /* When within an OpenMP context, notice uses of variables. */
6950 if (gimplify_omp_ctxp)
6951 omp_notice_variable (gimplify_omp_ctxp, *expr_p, true);
6952 ret = GS_ALL_DONE;
6953 break;
6955 case SSA_NAME:
6956 /* Allow callbacks into the gimplifier during optimization. */
6957 ret = GS_ALL_DONE;
6958 break;
6960 case OMP_PARALLEL:
6961 gimplify_omp_parallel (expr_p, pre_p);
6962 ret = GS_ALL_DONE;
6963 break;
6965 case OMP_TASK:
6966 gimplify_omp_task (expr_p, pre_p);
6967 ret = GS_ALL_DONE;
6968 break;
6970 case OMP_FOR:
6971 ret = gimplify_omp_for (expr_p, pre_p);
6972 break;
6974 case OMP_SECTIONS:
6975 case OMP_SINGLE:
6976 gimplify_omp_workshare (expr_p, pre_p);
6977 ret = GS_ALL_DONE;
6978 break;
6980 case OMP_SECTION:
6981 case OMP_MASTER:
6982 case OMP_ORDERED:
6983 case OMP_CRITICAL:
6985 gimple_seq body = NULL;
6986 gimple g;
6988 gimplify_and_add (OMP_BODY (*expr_p), &body);
6989 switch (TREE_CODE (*expr_p))
6991 case OMP_SECTION:
6992 g = gimple_build_omp_section (body);
6993 break;
6994 case OMP_MASTER:
6995 g = gimple_build_omp_master (body);
6996 break;
6997 case OMP_ORDERED:
6998 g = gimple_build_omp_ordered (body);
6999 break;
7000 case OMP_CRITICAL:
7001 g = gimple_build_omp_critical (body,
7002 OMP_CRITICAL_NAME (*expr_p));
7003 break;
7004 default:
7005 gcc_unreachable ();
7007 gimplify_seq_add_stmt (pre_p, g);
7008 ret = GS_ALL_DONE;
7009 break;
7012 case OMP_ATOMIC:
7013 ret = gimplify_omp_atomic (expr_p, pre_p);
7014 break;
7016 case POINTER_PLUS_EXPR:
7017 /* Convert ((type *)A)+offset into &A->field_of_type_and_offset.
7018 The second is gimple immediate saving a need for extra statement.
7020 if (TREE_CODE (TREE_OPERAND (*expr_p, 1)) == INTEGER_CST
7021 && (tmp = maybe_fold_offset_to_address
7022 (EXPR_LOCATION (*expr_p),
7023 TREE_OPERAND (*expr_p, 0), TREE_OPERAND (*expr_p, 1),
7024 TREE_TYPE (*expr_p))))
7026 *expr_p = tmp;
7027 break;
7029 /* Convert (void *)&a + 4 into (void *)&a[1]. */
7030 if (TREE_CODE (TREE_OPERAND (*expr_p, 0)) == NOP_EXPR
7031 && TREE_CODE (TREE_OPERAND (*expr_p, 1)) == INTEGER_CST
7032 && POINTER_TYPE_P (TREE_TYPE (TREE_OPERAND (TREE_OPERAND (*expr_p,
7033 0),0)))
7034 && (tmp = maybe_fold_offset_to_address
7035 (EXPR_LOCATION (*expr_p),
7036 TREE_OPERAND (TREE_OPERAND (*expr_p, 0), 0),
7037 TREE_OPERAND (*expr_p, 1),
7038 TREE_TYPE (TREE_OPERAND (TREE_OPERAND (*expr_p, 0),
7039 0)))))
7041 *expr_p = fold_convert (TREE_TYPE (*expr_p), tmp);
7042 break;
7044 /* FALLTHRU */
7046 default:
7047 switch (TREE_CODE_CLASS (TREE_CODE (*expr_p)))
7049 case tcc_comparison:
7050 /* Handle comparison of objects of non scalar mode aggregates
7051 with a call to memcmp. It would be nice to only have to do
7052 this for variable-sized objects, but then we'd have to allow
7053 the same nest of reference nodes we allow for MODIFY_EXPR and
7054 that's too complex.
7056 Compare scalar mode aggregates as scalar mode values. Using
7057 memcmp for them would be very inefficient at best, and is
7058 plain wrong if bitfields are involved. */
7060 tree type = TREE_TYPE (TREE_OPERAND (*expr_p, 1));
7062 if (!AGGREGATE_TYPE_P (type))
7063 goto expr_2;
7064 else if (TYPE_MODE (type) != BLKmode)
7065 ret = gimplify_scalar_mode_aggregate_compare (expr_p);
7066 else
7067 ret = gimplify_variable_sized_compare (expr_p);
7069 break;
7072 /* If *EXPR_P does not need to be special-cased, handle it
7073 according to its class. */
7074 case tcc_unary:
7075 ret = gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p,
7076 post_p, is_gimple_val, fb_rvalue);
7077 break;
7079 case tcc_binary:
7080 expr_2:
7082 enum gimplify_status r0, r1;
7084 r0 = gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p,
7085 post_p, is_gimple_val, fb_rvalue);
7086 r1 = gimplify_expr (&TREE_OPERAND (*expr_p, 1), pre_p,
7087 post_p, is_gimple_val, fb_rvalue);
7089 ret = MIN (r0, r1);
7090 break;
7093 case tcc_declaration:
7094 case tcc_constant:
7095 ret = GS_ALL_DONE;
7096 goto dont_recalculate;
7098 default:
7099 gcc_assert (TREE_CODE (*expr_p) == TRUTH_AND_EXPR
7100 || TREE_CODE (*expr_p) == TRUTH_OR_EXPR
7101 || TREE_CODE (*expr_p) == TRUTH_XOR_EXPR);
7102 goto expr_2;
7105 recalculate_side_effects (*expr_p);
7107 dont_recalculate:
7108 break;
7111 /* If we replaced *expr_p, gimplify again. */
7112 if (ret == GS_OK && (*expr_p == NULL || *expr_p == save_expr))
7113 ret = GS_ALL_DONE;
7115 while (ret == GS_OK);
7117 /* If we encountered an error_mark somewhere nested inside, either
7118 stub out the statement or propagate the error back out. */
7119 if (ret == GS_ERROR)
7121 if (is_statement)
7122 *expr_p = NULL;
7123 goto out;
7126 /* This was only valid as a return value from the langhook, which
7127 we handled. Make sure it doesn't escape from any other context. */
7128 gcc_assert (ret != GS_UNHANDLED);
7130 if (fallback == fb_none && *expr_p && !is_gimple_stmt (*expr_p))
7132 /* We aren't looking for a value, and we don't have a valid
7133 statement. If it doesn't have side-effects, throw it away. */
7134 if (!TREE_SIDE_EFFECTS (*expr_p))
7135 *expr_p = NULL;
7136 else if (!TREE_THIS_VOLATILE (*expr_p))
7138 /* This is probably a _REF that contains something nested that
7139 has side effects. Recurse through the operands to find it. */
7140 enum tree_code code = TREE_CODE (*expr_p);
7142 switch (code)
7144 case COMPONENT_REF:
7145 case REALPART_EXPR:
7146 case IMAGPART_EXPR:
7147 case VIEW_CONVERT_EXPR:
7148 gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p, post_p,
7149 gimple_test_f, fallback);
7150 break;
7152 case ARRAY_REF:
7153 case ARRAY_RANGE_REF:
7154 gimplify_expr (&TREE_OPERAND (*expr_p, 0), pre_p, post_p,
7155 gimple_test_f, fallback);
7156 gimplify_expr (&TREE_OPERAND (*expr_p, 1), pre_p, post_p,
7157 gimple_test_f, fallback);
7158 break;
7160 default:
7161 /* Anything else with side-effects must be converted to
7162 a valid statement before we get here. */
7163 gcc_unreachable ();
7166 *expr_p = NULL;
7168 else if (COMPLETE_TYPE_P (TREE_TYPE (*expr_p))
7169 && TYPE_MODE (TREE_TYPE (*expr_p)) != BLKmode)
7171 /* Historically, the compiler has treated a bare reference
7172 to a non-BLKmode volatile lvalue as forcing a load. */
7173 tree type = TYPE_MAIN_VARIANT (TREE_TYPE (*expr_p));
7175 /* Normally, we do not want to create a temporary for a
7176 TREE_ADDRESSABLE type because such a type should not be
7177 copied by bitwise-assignment. However, we make an
7178 exception here, as all we are doing here is ensuring that
7179 we read the bytes that make up the type. We use
7180 create_tmp_var_raw because create_tmp_var will abort when
7181 given a TREE_ADDRESSABLE type. */
7182 tree tmp = create_tmp_var_raw (type, "vol");
7183 gimple_add_tmp_var (tmp);
7184 gimplify_assign (tmp, *expr_p, pre_p);
7185 *expr_p = NULL;
7187 else
7188 /* We can't do anything useful with a volatile reference to
7189 an incomplete type, so just throw it away. Likewise for
7190 a BLKmode type, since any implicit inner load should
7191 already have been turned into an explicit one by the
7192 gimplification process. */
7193 *expr_p = NULL;
7196 /* If we are gimplifying at the statement level, we're done. Tack
7197 everything together and return. */
7198 if (fallback == fb_none || is_statement)
7200 /* Since *EXPR_P has been converted into a GIMPLE tuple, clear
7201 it out for GC to reclaim it. */
7202 *expr_p = NULL_TREE;
7204 if (!gimple_seq_empty_p (internal_pre)
7205 || !gimple_seq_empty_p (internal_post))
7207 gimplify_seq_add_seq (&internal_pre, internal_post);
7208 gimplify_seq_add_seq (pre_p, internal_pre);
7211 /* The result of gimplifying *EXPR_P is going to be the last few
7212 statements in *PRE_P and *POST_P. Add location information
7213 to all the statements that were added by the gimplification
7214 helpers. */
7215 if (!gimple_seq_empty_p (*pre_p))
7216 annotate_all_with_location_after (*pre_p, pre_last_gsi, input_location);
7218 if (!gimple_seq_empty_p (*post_p))
7219 annotate_all_with_location_after (*post_p, post_last_gsi,
7220 input_location);
7222 goto out;
7225 #ifdef ENABLE_GIMPLE_CHECKING
7226 if (*expr_p)
7228 enum tree_code code = TREE_CODE (*expr_p);
7229 /* These expressions should already be in gimple IR form. */
7230 gcc_assert (code != MODIFY_EXPR
7231 && code != ASM_EXPR
7232 && code != BIND_EXPR
7233 && code != CATCH_EXPR
7234 && (code != COND_EXPR || gimplify_ctxp->allow_rhs_cond_expr)
7235 && code != EH_FILTER_EXPR
7236 && code != GOTO_EXPR
7237 && code != LABEL_EXPR
7238 && code != LOOP_EXPR
7239 && code != SWITCH_EXPR
7240 && code != TRY_FINALLY_EXPR
7241 && code != OMP_CRITICAL
7242 && code != OMP_FOR
7243 && code != OMP_MASTER
7244 && code != OMP_ORDERED
7245 && code != OMP_PARALLEL
7246 && code != OMP_SECTIONS
7247 && code != OMP_SECTION
7248 && code != OMP_SINGLE);
7250 #endif
7252 /* Otherwise we're gimplifying a subexpression, so the resulting
7253 value is interesting. If it's a valid operand that matches
7254 GIMPLE_TEST_F, we're done. Unless we are handling some
7255 post-effects internally; if that's the case, we need to copy into
7256 a temporary before adding the post-effects to POST_P. */
7257 if (gimple_seq_empty_p (internal_post) && (*gimple_test_f) (*expr_p))
7258 goto out;
7260 /* Otherwise, we need to create a new temporary for the gimplified
7261 expression. */
7263 /* We can't return an lvalue if we have an internal postqueue. The
7264 object the lvalue refers to would (probably) be modified by the
7265 postqueue; we need to copy the value out first, which means an
7266 rvalue. */
7267 if ((fallback & fb_lvalue)
7268 && gimple_seq_empty_p (internal_post)
7269 && is_gimple_addressable (*expr_p))
7271 /* An lvalue will do. Take the address of the expression, store it
7272 in a temporary, and replace the expression with an INDIRECT_REF of
7273 that temporary. */
7274 tmp = build_fold_addr_expr_loc (input_location, *expr_p);
7275 gimplify_expr (&tmp, pre_p, post_p, is_gimple_reg, fb_rvalue);
7276 *expr_p = build1 (INDIRECT_REF, TREE_TYPE (TREE_TYPE (tmp)), tmp);
7278 else if ((fallback & fb_rvalue) && is_gimple_reg_rhs_or_call (*expr_p))
7280 /* An rvalue will do. Assign the gimplified expression into a
7281 new temporary TMP and replace the original expression with
7282 TMP. First, make sure that the expression has a type so that
7283 it can be assigned into a temporary. */
7284 gcc_assert (!VOID_TYPE_P (TREE_TYPE (*expr_p)));
7286 if (!gimple_seq_empty_p (internal_post) || (fallback & fb_lvalue))
7287 /* The postqueue might change the value of the expression between
7288 the initialization and use of the temporary, so we can't use a
7289 formal temp. FIXME do we care? */
7291 *expr_p = get_initialized_tmp_var (*expr_p, pre_p, post_p);
7292 if (TREE_CODE (TREE_TYPE (*expr_p)) == COMPLEX_TYPE
7293 || TREE_CODE (TREE_TYPE (*expr_p)) == VECTOR_TYPE)
7294 DECL_GIMPLE_REG_P (*expr_p) = 1;
7296 else
7297 *expr_p = get_formal_tmp_var (*expr_p, pre_p);
7299 else
7301 #ifdef ENABLE_GIMPLE_CHECKING
7302 if (!(fallback & fb_mayfail))
7304 fprintf (stderr, "gimplification failed:\n");
7305 print_generic_expr (stderr, *expr_p, 0);
7306 debug_tree (*expr_p);
7307 internal_error ("gimplification failed");
7309 #endif
7310 gcc_assert (fallback & fb_mayfail);
7312 /* If this is an asm statement, and the user asked for the
7313 impossible, don't die. Fail and let gimplify_asm_expr
7314 issue an error. */
7315 ret = GS_ERROR;
7316 goto out;
7319 /* Make sure the temporary matches our predicate. */
7320 gcc_assert ((*gimple_test_f) (*expr_p));
7322 if (!gimple_seq_empty_p (internal_post))
7324 annotate_all_with_location (internal_post, input_location);
7325 gimplify_seq_add_seq (pre_p, internal_post);
7328 out:
7329 input_location = saved_location;
7330 return ret;
7333 /* Look through TYPE for variable-sized objects and gimplify each such
7334 size that we find. Add to LIST_P any statements generated. */
7336 void
7337 gimplify_type_sizes (tree type, gimple_seq *list_p)
7339 tree field, t;
7341 if (type == NULL || type == error_mark_node)
7342 return;
7344 /* We first do the main variant, then copy into any other variants. */
7345 type = TYPE_MAIN_VARIANT (type);
7347 /* Avoid infinite recursion. */
7348 if (TYPE_SIZES_GIMPLIFIED (type))
7349 return;
7351 TYPE_SIZES_GIMPLIFIED (type) = 1;
7353 switch (TREE_CODE (type))
7355 case INTEGER_TYPE:
7356 case ENUMERAL_TYPE:
7357 case BOOLEAN_TYPE:
7358 case REAL_TYPE:
7359 case FIXED_POINT_TYPE:
7360 gimplify_one_sizepos (&TYPE_MIN_VALUE (type), list_p);
7361 gimplify_one_sizepos (&TYPE_MAX_VALUE (type), list_p);
7363 for (t = TYPE_NEXT_VARIANT (type); t; t = TYPE_NEXT_VARIANT (t))
7365 TYPE_MIN_VALUE (t) = TYPE_MIN_VALUE (type);
7366 TYPE_MAX_VALUE (t) = TYPE_MAX_VALUE (type);
7368 break;
7370 case ARRAY_TYPE:
7371 /* These types may not have declarations, so handle them here. */
7372 gimplify_type_sizes (TREE_TYPE (type), list_p);
7373 gimplify_type_sizes (TYPE_DOMAIN (type), list_p);
7374 /* Ensure VLA bounds aren't removed, for -O0 they should be variables
7375 with assigned stack slots, for -O1+ -g they should be tracked
7376 by VTA. */
7377 if (TYPE_DOMAIN (type)
7378 && INTEGRAL_TYPE_P (TYPE_DOMAIN (type)))
7380 t = TYPE_MIN_VALUE (TYPE_DOMAIN (type));
7381 if (t && TREE_CODE (t) == VAR_DECL && DECL_ARTIFICIAL (t))
7382 DECL_IGNORED_P (t) = 0;
7383 t = TYPE_MAX_VALUE (TYPE_DOMAIN (type));
7384 if (t && TREE_CODE (t) == VAR_DECL && DECL_ARTIFICIAL (t))
7385 DECL_IGNORED_P (t) = 0;
7387 break;
7389 case RECORD_TYPE:
7390 case UNION_TYPE:
7391 case QUAL_UNION_TYPE:
7392 for (field = TYPE_FIELDS (type); field; field = TREE_CHAIN (field))
7393 if (TREE_CODE (field) == FIELD_DECL)
7395 gimplify_one_sizepos (&DECL_FIELD_OFFSET (field), list_p);
7396 gimplify_one_sizepos (&DECL_SIZE (field), list_p);
7397 gimplify_one_sizepos (&DECL_SIZE_UNIT (field), list_p);
7398 gimplify_type_sizes (TREE_TYPE (field), list_p);
7400 break;
7402 case POINTER_TYPE:
7403 case REFERENCE_TYPE:
7404 /* We used to recurse on the pointed-to type here, which turned out to
7405 be incorrect because its definition might refer to variables not
7406 yet initialized at this point if a forward declaration is involved.
7408 It was actually useful for anonymous pointed-to types to ensure
7409 that the sizes evaluation dominates every possible later use of the
7410 values. Restricting to such types here would be safe since there
7411 is no possible forward declaration around, but would introduce an
7412 undesirable middle-end semantic to anonymity. We then defer to
7413 front-ends the responsibility of ensuring that the sizes are
7414 evaluated both early and late enough, e.g. by attaching artificial
7415 type declarations to the tree. */
7416 break;
7418 default:
7419 break;
7422 gimplify_one_sizepos (&TYPE_SIZE (type), list_p);
7423 gimplify_one_sizepos (&TYPE_SIZE_UNIT (type), list_p);
7425 for (t = TYPE_NEXT_VARIANT (type); t; t = TYPE_NEXT_VARIANT (t))
7427 TYPE_SIZE (t) = TYPE_SIZE (type);
7428 TYPE_SIZE_UNIT (t) = TYPE_SIZE_UNIT (type);
7429 TYPE_SIZES_GIMPLIFIED (t) = 1;
7433 /* A subroutine of gimplify_type_sizes to make sure that *EXPR_P,
7434 a size or position, has had all of its SAVE_EXPRs evaluated.
7435 We add any required statements to *STMT_P. */
7437 void
7438 gimplify_one_sizepos (tree *expr_p, gimple_seq *stmt_p)
7440 tree type, expr = *expr_p;
7442 /* We don't do anything if the value isn't there, is constant, or contains
7443 A PLACEHOLDER_EXPR. We also don't want to do anything if it's already
7444 a VAR_DECL. If it's a VAR_DECL from another function, the gimplifier
7445 will want to replace it with a new variable, but that will cause problems
7446 if this type is from outside the function. It's OK to have that here. */
7447 if (expr == NULL_TREE || TREE_CONSTANT (expr)
7448 || TREE_CODE (expr) == VAR_DECL
7449 || CONTAINS_PLACEHOLDER_P (expr))
7450 return;
7452 type = TREE_TYPE (expr);
7453 *expr_p = unshare_expr (expr);
7455 gimplify_expr (expr_p, stmt_p, NULL, is_gimple_val, fb_rvalue);
7456 expr = *expr_p;
7458 /* Verify that we've an exact type match with the original expression.
7459 In particular, we do not wish to drop a "sizetype" in favour of a
7460 type of similar dimensions. We don't want to pollute the generic
7461 type-stripping code with this knowledge because it doesn't matter
7462 for the bulk of GENERIC/GIMPLE. It only matters that TYPE_SIZE_UNIT
7463 and friends retain their "sizetype-ness". */
7464 if (TREE_TYPE (expr) != type
7465 && TREE_CODE (type) == INTEGER_TYPE
7466 && TYPE_IS_SIZETYPE (type))
7468 tree tmp;
7469 gimple stmt;
7471 *expr_p = create_tmp_var (type, NULL);
7472 tmp = build1 (NOP_EXPR, type, expr);
7473 stmt = gimplify_assign (*expr_p, tmp, stmt_p);
7474 if (EXPR_HAS_LOCATION (expr))
7475 gimple_set_location (stmt, EXPR_LOCATION (expr));
7476 else
7477 gimple_set_location (stmt, input_location);
7482 /* Gimplify the body of statements pointed to by BODY_P and return a
7483 GIMPLE_BIND containing the sequence of GIMPLE statements
7484 corresponding to BODY_P. FNDECL is the function decl containing
7485 *BODY_P. */
7487 gimple
7488 gimplify_body (tree *body_p, tree fndecl, bool do_parms)
7490 location_t saved_location = input_location;
7491 gimple_seq parm_stmts, seq;
7492 gimple outer_bind;
7493 struct gimplify_ctx gctx;
7495 timevar_push (TV_TREE_GIMPLIFY);
7497 /* Initialize for optimize_insn_for_s{ize,peed}_p possibly called during
7498 gimplification. */
7499 default_rtl_profile ();
7501 gcc_assert (gimplify_ctxp == NULL);
7502 push_gimplify_context (&gctx);
7504 /* Unshare most shared trees in the body and in that of any nested functions.
7505 It would seem we don't have to do this for nested functions because
7506 they are supposed to be output and then the outer function gimplified
7507 first, but the g++ front end doesn't always do it that way. */
7508 unshare_body (body_p, fndecl);
7509 unvisit_body (body_p, fndecl);
7511 if (cgraph_node (fndecl)->origin)
7512 nonlocal_vlas = pointer_set_create ();
7514 /* Make sure input_location isn't set to something weird. */
7515 input_location = DECL_SOURCE_LOCATION (fndecl);
7517 /* Resolve callee-copies. This has to be done before processing
7518 the body so that DECL_VALUE_EXPR gets processed correctly. */
7519 parm_stmts = (do_parms) ? gimplify_parameters () : NULL;
7521 /* Gimplify the function's body. */
7522 seq = NULL;
7523 gimplify_stmt (body_p, &seq);
7524 outer_bind = gimple_seq_first_stmt (seq);
7525 if (!outer_bind)
7527 outer_bind = gimple_build_nop ();
7528 gimplify_seq_add_stmt (&seq, outer_bind);
7531 /* The body must contain exactly one statement, a GIMPLE_BIND. If this is
7532 not the case, wrap everything in a GIMPLE_BIND to make it so. */
7533 if (gimple_code (outer_bind) == GIMPLE_BIND
7534 && gimple_seq_first (seq) == gimple_seq_last (seq))
7536 else
7537 outer_bind = gimple_build_bind (NULL_TREE, seq, NULL);
7539 *body_p = NULL_TREE;
7541 /* If we had callee-copies statements, insert them at the beginning
7542 of the function and clear DECL_VALUE_EXPR_P on the parameters. */
7543 if (!gimple_seq_empty_p (parm_stmts))
7545 tree parm;
7547 gimplify_seq_add_seq (&parm_stmts, gimple_bind_body (outer_bind));
7548 gimple_bind_set_body (outer_bind, parm_stmts);
7550 for (parm = DECL_ARGUMENTS (current_function_decl);
7551 parm; parm = TREE_CHAIN (parm))
7552 if (DECL_HAS_VALUE_EXPR_P (parm))
7554 DECL_HAS_VALUE_EXPR_P (parm) = 0;
7555 DECL_IGNORED_P (parm) = 0;
7559 if (nonlocal_vlas)
7561 pointer_set_destroy (nonlocal_vlas);
7562 nonlocal_vlas = NULL;
7565 pop_gimplify_context (outer_bind);
7566 gcc_assert (gimplify_ctxp == NULL);
7568 #ifdef ENABLE_TYPES_CHECKING
7569 if (!errorcount && !sorrycount)
7570 verify_types_in_gimple_seq (gimple_bind_body (outer_bind));
7571 #endif
7573 timevar_pop (TV_TREE_GIMPLIFY);
7574 input_location = saved_location;
7576 return outer_bind;
7579 /* Entry point to the gimplification pass. FNDECL is the FUNCTION_DECL
7580 node for the function we want to gimplify.
7582 Returns the sequence of GIMPLE statements corresponding to the body
7583 of FNDECL. */
7585 void
7586 gimplify_function_tree (tree fndecl)
7588 tree oldfn, parm, ret;
7589 gimple_seq seq;
7590 gimple bind;
7592 gcc_assert (!gimple_body (fndecl));
7594 oldfn = current_function_decl;
7595 current_function_decl = fndecl;
7596 if (DECL_STRUCT_FUNCTION (fndecl))
7597 push_cfun (DECL_STRUCT_FUNCTION (fndecl));
7598 else
7599 push_struct_function (fndecl);
7601 for (parm = DECL_ARGUMENTS (fndecl); parm ; parm = TREE_CHAIN (parm))
7603 /* Preliminarily mark non-addressed complex variables as eligible
7604 for promotion to gimple registers. We'll transform their uses
7605 as we find them. */
7606 if ((TREE_CODE (TREE_TYPE (parm)) == COMPLEX_TYPE
7607 || TREE_CODE (TREE_TYPE (parm)) == VECTOR_TYPE)
7608 && !TREE_THIS_VOLATILE (parm)
7609 && !needs_to_live_in_memory (parm))
7610 DECL_GIMPLE_REG_P (parm) = 1;
7613 ret = DECL_RESULT (fndecl);
7614 if ((TREE_CODE (TREE_TYPE (ret)) == COMPLEX_TYPE
7615 || TREE_CODE (TREE_TYPE (ret)) == VECTOR_TYPE)
7616 && !needs_to_live_in_memory (ret))
7617 DECL_GIMPLE_REG_P (ret) = 1;
7619 bind = gimplify_body (&DECL_SAVED_TREE (fndecl), fndecl, true);
7621 /* The tree body of the function is no longer needed, replace it
7622 with the new GIMPLE body. */
7623 seq = gimple_seq_alloc ();
7624 gimple_seq_add_stmt (&seq, bind);
7625 gimple_set_body (fndecl, seq);
7627 /* If we're instrumenting function entry/exit, then prepend the call to
7628 the entry hook and wrap the whole function in a TRY_FINALLY_EXPR to
7629 catch the exit hook. */
7630 /* ??? Add some way to ignore exceptions for this TFE. */
7631 if (flag_instrument_function_entry_exit
7632 && !DECL_NO_INSTRUMENT_FUNCTION_ENTRY_EXIT (fndecl)
7633 && !flag_instrument_functions_exclude_p (fndecl))
7635 tree x;
7636 gimple new_bind;
7637 gimple tf;
7638 gimple_seq cleanup = NULL, body = NULL;
7640 x = implicit_built_in_decls[BUILT_IN_PROFILE_FUNC_EXIT];
7641 gimplify_seq_add_stmt (&cleanup, gimple_build_call (x, 0));
7642 tf = gimple_build_try (seq, cleanup, GIMPLE_TRY_FINALLY);
7644 x = implicit_built_in_decls[BUILT_IN_PROFILE_FUNC_ENTER];
7645 gimplify_seq_add_stmt (&body, gimple_build_call (x, 0));
7646 gimplify_seq_add_stmt (&body, tf);
7647 new_bind = gimple_build_bind (NULL, body, gimple_bind_block (bind));
7648 /* Clear the block for BIND, since it is no longer directly inside
7649 the function, but within a try block. */
7650 gimple_bind_set_block (bind, NULL);
7652 /* Replace the current function body with the body
7653 wrapped in the try/finally TF. */
7654 seq = gimple_seq_alloc ();
7655 gimple_seq_add_stmt (&seq, new_bind);
7656 gimple_set_body (fndecl, seq);
7659 DECL_SAVED_TREE (fndecl) = NULL_TREE;
7660 cfun->curr_properties = PROP_gimple_any;
7662 current_function_decl = oldfn;
7663 pop_cfun ();
7667 /* Some transformations like inlining may invalidate the GIMPLE form
7668 for operands. This function traverses all the operands in STMT and
7669 gimplifies anything that is not a valid gimple operand. Any new
7670 GIMPLE statements are inserted before *GSI_P. */
7672 void
7673 gimple_regimplify_operands (gimple stmt, gimple_stmt_iterator *gsi_p)
7675 size_t i, num_ops;
7676 tree orig_lhs = NULL_TREE, lhs, t;
7677 gimple_seq pre = NULL;
7678 gimple post_stmt = NULL;
7679 struct gimplify_ctx gctx;
7681 push_gimplify_context (&gctx);
7682 gimplify_ctxp->into_ssa = gimple_in_ssa_p (cfun);
7684 switch (gimple_code (stmt))
7686 case GIMPLE_COND:
7687 gimplify_expr (gimple_cond_lhs_ptr (stmt), &pre, NULL,
7688 is_gimple_val, fb_rvalue);
7689 gimplify_expr (gimple_cond_rhs_ptr (stmt), &pre, NULL,
7690 is_gimple_val, fb_rvalue);
7691 break;
7692 case GIMPLE_SWITCH:
7693 gimplify_expr (gimple_switch_index_ptr (stmt), &pre, NULL,
7694 is_gimple_val, fb_rvalue);
7695 break;
7696 case GIMPLE_OMP_ATOMIC_LOAD:
7697 gimplify_expr (gimple_omp_atomic_load_rhs_ptr (stmt), &pre, NULL,
7698 is_gimple_val, fb_rvalue);
7699 break;
7700 case GIMPLE_ASM:
7702 size_t i, noutputs = gimple_asm_noutputs (stmt);
7703 const char *constraint, **oconstraints;
7704 bool allows_mem, allows_reg, is_inout;
7706 oconstraints
7707 = (const char **) alloca ((noutputs) * sizeof (const char *));
7708 for (i = 0; i < noutputs; i++)
7710 tree op = gimple_asm_output_op (stmt, i);
7711 constraint = TREE_STRING_POINTER (TREE_VALUE (TREE_PURPOSE (op)));
7712 oconstraints[i] = constraint;
7713 parse_output_constraint (&constraint, i, 0, 0, &allows_mem,
7714 &allows_reg, &is_inout);
7715 gimplify_expr (&TREE_VALUE (op), &pre, NULL,
7716 is_inout ? is_gimple_min_lval : is_gimple_lvalue,
7717 fb_lvalue | fb_mayfail);
7719 for (i = 0; i < gimple_asm_ninputs (stmt); i++)
7721 tree op = gimple_asm_input_op (stmt, i);
7722 constraint = TREE_STRING_POINTER (TREE_VALUE (TREE_PURPOSE (op)));
7723 parse_input_constraint (&constraint, 0, 0, noutputs, 0,
7724 oconstraints, &allows_mem, &allows_reg);
7725 if (TREE_ADDRESSABLE (TREE_TYPE (TREE_VALUE (op))) && allows_mem)
7726 allows_reg = 0;
7727 if (!allows_reg && allows_mem)
7728 gimplify_expr (&TREE_VALUE (op), &pre, NULL,
7729 is_gimple_lvalue, fb_lvalue | fb_mayfail);
7730 else
7731 gimplify_expr (&TREE_VALUE (op), &pre, NULL,
7732 is_gimple_asm_val, fb_rvalue);
7735 break;
7736 default:
7737 /* NOTE: We start gimplifying operands from last to first to
7738 make sure that side-effects on the RHS of calls, assignments
7739 and ASMs are executed before the LHS. The ordering is not
7740 important for other statements. */
7741 num_ops = gimple_num_ops (stmt);
7742 orig_lhs = gimple_get_lhs (stmt);
7743 for (i = num_ops; i > 0; i--)
7745 tree op = gimple_op (stmt, i - 1);
7746 if (op == NULL_TREE)
7747 continue;
7748 if (i == 1 && (is_gimple_call (stmt) || is_gimple_assign (stmt)))
7749 gimplify_expr (&op, &pre, NULL, is_gimple_lvalue, fb_lvalue);
7750 else if (i == 2
7751 && is_gimple_assign (stmt)
7752 && num_ops == 2
7753 && get_gimple_rhs_class (gimple_expr_code (stmt))
7754 == GIMPLE_SINGLE_RHS)
7755 gimplify_expr (&op, &pre, NULL,
7756 rhs_predicate_for (gimple_assign_lhs (stmt)),
7757 fb_rvalue);
7758 else if (i == 2 && is_gimple_call (stmt))
7760 if (TREE_CODE (op) == FUNCTION_DECL)
7761 continue;
7762 gimplify_expr (&op, &pre, NULL, is_gimple_call_addr, fb_rvalue);
7764 else
7765 gimplify_expr (&op, &pre, NULL, is_gimple_val, fb_rvalue);
7766 gimple_set_op (stmt, i - 1, op);
7769 lhs = gimple_get_lhs (stmt);
7770 /* If the LHS changed it in a way that requires a simple RHS,
7771 create temporary. */
7772 if (lhs && !is_gimple_reg (lhs))
7774 bool need_temp = false;
7776 if (is_gimple_assign (stmt)
7777 && num_ops == 2
7778 && get_gimple_rhs_class (gimple_expr_code (stmt))
7779 == GIMPLE_SINGLE_RHS)
7780 gimplify_expr (gimple_assign_rhs1_ptr (stmt), &pre, NULL,
7781 rhs_predicate_for (gimple_assign_lhs (stmt)),
7782 fb_rvalue);
7783 else if (is_gimple_reg (lhs))
7785 if (is_gimple_reg_type (TREE_TYPE (lhs)))
7787 if (is_gimple_call (stmt))
7789 i = gimple_call_flags (stmt);
7790 if ((i & ECF_LOOPING_CONST_OR_PURE)
7791 || !(i & (ECF_CONST | ECF_PURE)))
7792 need_temp = true;
7794 if (stmt_can_throw_internal (stmt))
7795 need_temp = true;
7798 else
7800 if (is_gimple_reg_type (TREE_TYPE (lhs)))
7801 need_temp = true;
7802 else if (TYPE_MODE (TREE_TYPE (lhs)) != BLKmode)
7804 if (is_gimple_call (stmt))
7806 tree fndecl = gimple_call_fndecl (stmt);
7808 if (!aggregate_value_p (TREE_TYPE (lhs), fndecl)
7809 && !(fndecl && DECL_RESULT (fndecl)
7810 && DECL_BY_REFERENCE (DECL_RESULT (fndecl))))
7811 need_temp = true;
7813 else
7814 need_temp = true;
7817 if (need_temp)
7819 tree temp = create_tmp_var (TREE_TYPE (lhs), NULL);
7821 if (TREE_CODE (TREE_TYPE (lhs)) == COMPLEX_TYPE
7822 || TREE_CODE (TREE_TYPE (lhs)) == VECTOR_TYPE)
7823 DECL_GIMPLE_REG_P (temp) = 1;
7824 if (TREE_CODE (orig_lhs) == SSA_NAME)
7825 orig_lhs = SSA_NAME_VAR (orig_lhs);
7827 if (gimple_in_ssa_p (cfun))
7828 temp = make_ssa_name (temp, NULL);
7829 gimple_set_lhs (stmt, temp);
7830 post_stmt = gimple_build_assign (lhs, temp);
7831 if (TREE_CODE (lhs) == SSA_NAME)
7832 SSA_NAME_DEF_STMT (lhs) = post_stmt;
7835 break;
7838 if (gimple_referenced_vars (cfun))
7839 for (t = gimplify_ctxp->temps; t ; t = TREE_CHAIN (t))
7840 add_referenced_var (t);
7842 if (!gimple_seq_empty_p (pre))
7844 if (gimple_in_ssa_p (cfun))
7846 gimple_stmt_iterator i;
7848 for (i = gsi_start (pre); !gsi_end_p (i); gsi_next (&i))
7849 mark_symbols_for_renaming (gsi_stmt (i));
7851 gsi_insert_seq_before (gsi_p, pre, GSI_SAME_STMT);
7853 if (post_stmt)
7854 gsi_insert_after (gsi_p, post_stmt, GSI_NEW_STMT);
7856 pop_gimplify_context (NULL);
7860 /* Expands EXPR to list of gimple statements STMTS. If SIMPLE is true,
7861 force the result to be either ssa_name or an invariant, otherwise
7862 just force it to be a rhs expression. If VAR is not NULL, make the
7863 base variable of the final destination be VAR if suitable. */
7865 tree
7866 force_gimple_operand (tree expr, gimple_seq *stmts, bool simple, tree var)
7868 tree t;
7869 enum gimplify_status ret;
7870 gimple_predicate gimple_test_f;
7871 struct gimplify_ctx gctx;
7873 *stmts = NULL;
7875 if (is_gimple_val (expr))
7876 return expr;
7878 gimple_test_f = simple ? is_gimple_val : is_gimple_reg_rhs;
7880 push_gimplify_context (&gctx);
7881 gimplify_ctxp->into_ssa = gimple_in_ssa_p (cfun);
7882 gimplify_ctxp->allow_rhs_cond_expr = true;
7884 if (var)
7885 expr = build2 (MODIFY_EXPR, TREE_TYPE (var), var, expr);
7887 if (TREE_CODE (expr) != MODIFY_EXPR
7888 && TREE_TYPE (expr) == void_type_node)
7890 gimplify_and_add (expr, stmts);
7891 expr = NULL_TREE;
7893 else
7895 ret = gimplify_expr (&expr, stmts, NULL, gimple_test_f, fb_rvalue);
7896 gcc_assert (ret != GS_ERROR);
7899 if (gimple_referenced_vars (cfun))
7900 for (t = gimplify_ctxp->temps; t ; t = TREE_CHAIN (t))
7901 add_referenced_var (t);
7903 pop_gimplify_context (NULL);
7905 return expr;
7908 /* Invokes force_gimple_operand for EXPR with parameters SIMPLE_P and VAR. If
7909 some statements are produced, emits them at GSI. If BEFORE is true.
7910 the statements are appended before GSI, otherwise they are appended after
7911 it. M specifies the way GSI moves after insertion (GSI_SAME_STMT or
7912 GSI_CONTINUE_LINKING are the usual values). */
7914 tree
7915 force_gimple_operand_gsi (gimple_stmt_iterator *gsi, tree expr,
7916 bool simple_p, tree var, bool before,
7917 enum gsi_iterator_update m)
7919 gimple_seq stmts;
7921 expr = force_gimple_operand (expr, &stmts, simple_p, var);
7923 if (!gimple_seq_empty_p (stmts))
7925 if (gimple_in_ssa_p (cfun))
7927 gimple_stmt_iterator i;
7929 for (i = gsi_start (stmts); !gsi_end_p (i); gsi_next (&i))
7930 mark_symbols_for_renaming (gsi_stmt (i));
7933 if (before)
7934 gsi_insert_seq_before (gsi, stmts, m);
7935 else
7936 gsi_insert_seq_after (gsi, stmts, m);
7939 return expr;
7942 #include "gt-gimplify.h"