Add C++11 header <cuchar>.
[official-gcc.git] / gcc / cp / lambda.c
blobea9dba05227556fd958adcbdd187d05e13920303
1 /* Perform the semantic phase of lambda parsing, i.e., the process of
2 building tree structure, checking semantic consistency, and
3 building RTL. These routines are used both during actual parsing
4 and during the instantiation of template functions.
6 Copyright (C) 1998-2015 Free Software Foundation, Inc.
8 This file is part of GCC.
10 GCC is free software; you can redistribute it and/or modify it
11 under the terms of the GNU General Public License as published by
12 the Free Software Foundation; either version 3, or (at your option)
13 any later version.
15 GCC is distributed in the hope that it will be useful, but
16 WITHOUT ANY WARRANTY; without even the implied warranty of
17 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
18 General Public License for more details.
20 You should have received a copy of the GNU General Public License
21 along with GCC; see the file COPYING3. If not see
22 <http://www.gnu.org/licenses/>. */
24 #include "config.h"
25 #include "system.h"
26 #include "coretypes.h"
27 #include "alias.h"
28 #include "tree.h"
29 #include "options.h"
30 #include "stringpool.h"
31 #include "tm.h"
32 #include "hard-reg-set.h"
33 #include "function.h"
34 #include "cgraph.h"
35 #include "tree-iterator.h"
36 #include "cp-tree.h"
37 #include "toplev.h"
39 /* Constructor for a lambda expression. */
41 tree
42 build_lambda_expr (void)
44 tree lambda = make_node (LAMBDA_EXPR);
45 LAMBDA_EXPR_DEFAULT_CAPTURE_MODE (lambda) = CPLD_NONE;
46 LAMBDA_EXPR_CAPTURE_LIST (lambda) = NULL_TREE;
47 LAMBDA_EXPR_THIS_CAPTURE (lambda) = NULL_TREE;
48 LAMBDA_EXPR_PENDING_PROXIES (lambda) = NULL;
49 LAMBDA_EXPR_RETURN_TYPE (lambda) = NULL_TREE;
50 LAMBDA_EXPR_MUTABLE_P (lambda) = false;
51 return lambda;
54 /* Create the closure object for a LAMBDA_EXPR. */
56 tree
57 build_lambda_object (tree lambda_expr)
59 /* Build aggregate constructor call.
60 - cp_parser_braced_list
61 - cp_parser_functional_cast */
62 vec<constructor_elt, va_gc> *elts = NULL;
63 tree node, expr, type;
64 location_t saved_loc;
66 if (processing_template_decl)
67 return lambda_expr;
69 /* Make sure any error messages refer to the lambda-introducer. */
70 saved_loc = input_location;
71 input_location = LAMBDA_EXPR_LOCATION (lambda_expr);
73 for (node = LAMBDA_EXPR_CAPTURE_LIST (lambda_expr);
74 node;
75 node = TREE_CHAIN (node))
77 tree field = TREE_PURPOSE (node);
78 tree val = TREE_VALUE (node);
80 if (field == error_mark_node)
82 expr = error_mark_node;
83 goto out;
86 if (DECL_P (val))
87 mark_used (val);
89 /* Mere mortals can't copy arrays with aggregate initialization, so
90 do some magic to make it work here. */
91 if (TREE_CODE (TREE_TYPE (field)) == ARRAY_TYPE)
92 val = build_array_copy (val);
93 else if (DECL_NORMAL_CAPTURE_P (field)
94 && !DECL_VLA_CAPTURE_P (field)
95 && TREE_CODE (TREE_TYPE (field)) != REFERENCE_TYPE)
97 /* "the entities that are captured by copy are used to
98 direct-initialize each corresponding non-static data
99 member of the resulting closure object."
101 There's normally no way to express direct-initialization
102 from an element of a CONSTRUCTOR, so we build up a special
103 TARGET_EXPR to bypass the usual copy-initialization. */
104 val = force_rvalue (val, tf_warning_or_error);
105 if (TREE_CODE (val) == TARGET_EXPR)
106 TARGET_EXPR_DIRECT_INIT_P (val) = true;
109 CONSTRUCTOR_APPEND_ELT (elts, DECL_NAME (field), val);
112 expr = build_constructor (init_list_type_node, elts);
113 CONSTRUCTOR_IS_DIRECT_INIT (expr) = 1;
115 /* N2927: "[The closure] class type is not an aggregate."
116 But we briefly treat it as an aggregate to make this simpler. */
117 type = LAMBDA_EXPR_CLOSURE (lambda_expr);
118 CLASSTYPE_NON_AGGREGATE (type) = 0;
119 expr = finish_compound_literal (type, expr, tf_warning_or_error);
120 CLASSTYPE_NON_AGGREGATE (type) = 1;
122 out:
123 input_location = saved_loc;
124 return expr;
127 /* Return an initialized RECORD_TYPE for LAMBDA.
128 LAMBDA must have its explicit captures already. */
130 tree
131 begin_lambda_type (tree lambda)
133 tree type;
136 /* Unique name. This is just like an unnamed class, but we cannot use
137 make_anon_name because of certain checks against TYPE_ANONYMOUS_P. */
138 tree name;
139 name = make_lambda_name ();
141 /* Create the new RECORD_TYPE for this lambda. */
142 type = xref_tag (/*tag_code=*/record_type,
143 name,
144 /*scope=*/ts_lambda,
145 /*template_header_p=*/false);
146 if (type == error_mark_node)
147 return error_mark_node;
150 /* Designate it as a struct so that we can use aggregate initialization. */
151 CLASSTYPE_DECLARED_CLASS (type) = false;
153 /* Cross-reference the expression and the type. */
154 LAMBDA_EXPR_CLOSURE (lambda) = type;
155 CLASSTYPE_LAMBDA_EXPR (type) = lambda;
157 /* Clear base types. */
158 xref_basetypes (type, /*bases=*/NULL_TREE);
160 /* Start the class. */
161 type = begin_class_definition (type);
163 return type;
166 /* Returns the type to use for the return type of the operator() of a
167 closure class. */
169 tree
170 lambda_return_type (tree expr)
172 if (expr == NULL_TREE)
173 return void_type_node;
174 if (type_unknown_p (expr)
175 || BRACE_ENCLOSED_INITIALIZER_P (expr))
177 cxx_incomplete_type_error (expr, TREE_TYPE (expr));
178 return void_type_node;
180 gcc_checking_assert (!type_dependent_expression_p (expr));
181 return cv_unqualified (type_decays_to (unlowered_expr_type (expr)));
184 /* Given a LAMBDA_EXPR or closure type LAMBDA, return the op() of the
185 closure type. */
187 tree
188 lambda_function (tree lambda)
190 tree type;
191 if (TREE_CODE (lambda) == LAMBDA_EXPR)
192 type = LAMBDA_EXPR_CLOSURE (lambda);
193 else
194 type = lambda;
195 gcc_assert (LAMBDA_TYPE_P (type));
196 /* Don't let debug_tree cause instantiation. */
197 if (CLASSTYPE_TEMPLATE_INSTANTIATION (type)
198 && !COMPLETE_OR_OPEN_TYPE_P (type))
199 return NULL_TREE;
200 lambda = lookup_member (type, ansi_opname (CALL_EXPR),
201 /*protect=*/0, /*want_type=*/false,
202 tf_warning_or_error);
203 if (lambda)
204 lambda = STRIP_TEMPLATE (get_first_fn (lambda));
205 return lambda;
208 /* Returns the type to use for the FIELD_DECL corresponding to the
209 capture of EXPR.
210 The caller should add REFERENCE_TYPE for capture by reference. */
212 tree
213 lambda_capture_field_type (tree expr, bool explicit_init_p)
215 tree type;
216 if (explicit_init_p)
218 type = make_auto ();
219 type = do_auto_deduction (type, expr, type);
221 else
222 type = non_reference (unlowered_expr_type (expr));
223 if (type_dependent_expression_p (expr)
224 && !is_this_parameter (tree_strip_nop_conversions (expr)))
226 type = cxx_make_type (DECLTYPE_TYPE);
227 DECLTYPE_TYPE_EXPR (type) = expr;
228 DECLTYPE_FOR_LAMBDA_CAPTURE (type) = true;
229 DECLTYPE_FOR_INIT_CAPTURE (type) = explicit_init_p;
230 SET_TYPE_STRUCTURAL_EQUALITY (type);
232 return type;
235 /* Returns true iff DECL is a lambda capture proxy variable created by
236 build_capture_proxy. */
238 bool
239 is_capture_proxy (tree decl)
241 return (VAR_P (decl)
242 && DECL_HAS_VALUE_EXPR_P (decl)
243 && !DECL_ANON_UNION_VAR_P (decl)
244 && LAMBDA_FUNCTION_P (DECL_CONTEXT (decl)));
247 /* Returns true iff DECL is a capture proxy for a normal capture
248 (i.e. without explicit initializer). */
250 bool
251 is_normal_capture_proxy (tree decl)
253 if (!is_capture_proxy (decl))
254 /* It's not a capture proxy. */
255 return false;
257 if (variably_modified_type_p (TREE_TYPE (decl), NULL_TREE))
258 /* VLA capture. */
259 return true;
261 /* It is a capture proxy, is it a normal capture? */
262 tree val = DECL_VALUE_EXPR (decl);
263 if (val == error_mark_node)
264 return true;
266 gcc_assert (TREE_CODE (val) == COMPONENT_REF);
267 val = TREE_OPERAND (val, 1);
268 return DECL_NORMAL_CAPTURE_P (val);
271 /* VAR is a capture proxy created by build_capture_proxy; add it to the
272 current function, which is the operator() for the appropriate lambda. */
274 void
275 insert_capture_proxy (tree var)
277 cp_binding_level *b;
278 tree stmt_list;
280 /* Put the capture proxy in the extra body block so that it won't clash
281 with a later local variable. */
282 b = current_binding_level;
283 for (;;)
285 cp_binding_level *n = b->level_chain;
286 if (n->kind == sk_function_parms)
287 break;
288 b = n;
290 pushdecl_with_scope (var, b, false);
292 /* And put a DECL_EXPR in the STATEMENT_LIST for the same block. */
293 var = build_stmt (DECL_SOURCE_LOCATION (var), DECL_EXPR, var);
294 stmt_list = (*stmt_list_stack)[1];
295 gcc_assert (stmt_list);
296 append_to_statement_list_force (var, &stmt_list);
299 /* We've just finished processing a lambda; if the containing scope is also
300 a lambda, insert any capture proxies that were created while processing
301 the nested lambda. */
303 void
304 insert_pending_capture_proxies (void)
306 tree lam;
307 vec<tree, va_gc> *proxies;
308 unsigned i;
310 if (!current_function_decl || !LAMBDA_FUNCTION_P (current_function_decl))
311 return;
313 lam = CLASSTYPE_LAMBDA_EXPR (DECL_CONTEXT (current_function_decl));
314 proxies = LAMBDA_EXPR_PENDING_PROXIES (lam);
315 for (i = 0; i < vec_safe_length (proxies); ++i)
317 tree var = (*proxies)[i];
318 insert_capture_proxy (var);
320 release_tree_vector (LAMBDA_EXPR_PENDING_PROXIES (lam));
321 LAMBDA_EXPR_PENDING_PROXIES (lam) = NULL;
324 /* Given REF, a COMPONENT_REF designating a field in the lambda closure,
325 return the type we want the proxy to have: the type of the field itself,
326 with added const-qualification if the lambda isn't mutable and the
327 capture is by value. */
329 tree
330 lambda_proxy_type (tree ref)
332 tree type;
333 if (ref == error_mark_node)
334 return error_mark_node;
335 if (REFERENCE_REF_P (ref))
336 ref = TREE_OPERAND (ref, 0);
337 gcc_assert (TREE_CODE (ref) == COMPONENT_REF);
338 type = TREE_TYPE (ref);
339 if (!type || WILDCARD_TYPE_P (non_reference (type)))
341 type = cxx_make_type (DECLTYPE_TYPE);
342 DECLTYPE_TYPE_EXPR (type) = ref;
343 DECLTYPE_FOR_LAMBDA_PROXY (type) = true;
344 SET_TYPE_STRUCTURAL_EQUALITY (type);
346 if (DECL_PACK_P (TREE_OPERAND (ref, 1)))
347 type = make_pack_expansion (type);
348 return type;
351 /* MEMBER is a capture field in a lambda closure class. Now that we're
352 inside the operator(), build a placeholder var for future lookups and
353 debugging. */
355 tree
356 build_capture_proxy (tree member)
358 tree var, object, fn, closure, name, lam, type;
360 if (PACK_EXPANSION_P (member))
361 member = PACK_EXPANSION_PATTERN (member);
363 closure = DECL_CONTEXT (member);
364 fn = lambda_function (closure);
365 lam = CLASSTYPE_LAMBDA_EXPR (closure);
367 /* The proxy variable forwards to the capture field. */
368 object = build_fold_indirect_ref (DECL_ARGUMENTS (fn));
369 object = finish_non_static_data_member (member, object, NULL_TREE);
370 if (REFERENCE_REF_P (object))
371 object = TREE_OPERAND (object, 0);
373 /* Remove the __ inserted by add_capture. */
374 name = get_identifier (IDENTIFIER_POINTER (DECL_NAME (member)) + 2);
376 type = lambda_proxy_type (object);
378 if (DECL_VLA_CAPTURE_P (member))
380 /* Rebuild the VLA type from the pointer and maxindex. */
381 tree field = next_initializable_field (TYPE_FIELDS (type));
382 tree ptr = build_simple_component_ref (object, field);
383 field = next_initializable_field (DECL_CHAIN (field));
384 tree max = build_simple_component_ref (object, field);
385 type = build_cplus_array_type (TREE_TYPE (TREE_TYPE (ptr)),
386 build_index_type (max));
387 type = build_reference_type (type);
388 REFERENCE_VLA_OK (type) = true;
389 object = convert (type, ptr);
392 var = build_decl (input_location, VAR_DECL, name, type);
393 SET_DECL_VALUE_EXPR (var, object);
394 DECL_HAS_VALUE_EXPR_P (var) = 1;
395 DECL_ARTIFICIAL (var) = 1;
396 TREE_USED (var) = 1;
397 DECL_CONTEXT (var) = fn;
399 if (name == this_identifier)
401 gcc_assert (LAMBDA_EXPR_THIS_CAPTURE (lam) == member);
402 LAMBDA_EXPR_THIS_CAPTURE (lam) = var;
405 if (fn == current_function_decl)
406 insert_capture_proxy (var);
407 else
408 vec_safe_push (LAMBDA_EXPR_PENDING_PROXIES (lam), var);
410 return var;
413 /* Return a struct containing a pointer and a length for lambda capture of
414 an array of runtime length. */
416 static tree
417 vla_capture_type (tree array_type)
419 static tree ptr_id, max_id;
420 tree type = xref_tag (record_type, make_anon_name (), ts_current, false);
421 xref_basetypes (type, NULL_TREE);
422 type = begin_class_definition (type);
423 if (!ptr_id)
425 ptr_id = get_identifier ("ptr");
426 max_id = get_identifier ("max");
428 tree ptrtype = build_pointer_type (TREE_TYPE (array_type));
429 tree field = build_decl (input_location, FIELD_DECL, ptr_id, ptrtype);
430 finish_member_declaration (field);
431 field = build_decl (input_location, FIELD_DECL, max_id, sizetype);
432 finish_member_declaration (field);
433 return finish_struct (type, NULL_TREE);
436 /* From an ID and INITIALIZER, create a capture (by reference if
437 BY_REFERENCE_P is true), add it to the capture-list for LAMBDA,
438 and return it. */
440 tree
441 add_capture (tree lambda, tree id, tree orig_init, bool by_reference_p,
442 bool explicit_init_p)
444 char *buf;
445 tree type, member, name;
446 bool vla = false;
447 bool variadic = false;
448 tree initializer = orig_init;
450 if (PACK_EXPANSION_P (initializer))
452 initializer = PACK_EXPANSION_PATTERN (initializer);
453 variadic = true;
456 if (TREE_CODE (initializer) == TREE_LIST)
457 initializer = build_x_compound_expr_from_list (initializer, ELK_INIT,
458 tf_warning_or_error);
459 type = TREE_TYPE (initializer);
460 if (type == error_mark_node)
461 return error_mark_node;
463 if (array_of_runtime_bound_p (type))
465 vla = true;
466 if (!by_reference_p)
467 error ("array of runtime bound cannot be captured by copy, "
468 "only by reference");
470 /* For a VLA, we capture the address of the first element and the
471 maximum index, and then reconstruct the VLA for the proxy. */
472 tree elt = cp_build_array_ref (input_location, initializer,
473 integer_zero_node, tf_warning_or_error);
474 initializer = build_constructor_va (init_list_type_node, 2,
475 NULL_TREE, build_address (elt),
476 NULL_TREE, array_type_nelts (type));
477 type = vla_capture_type (type);
479 else if (!dependent_type_p (type)
480 && variably_modified_type_p (type, NULL_TREE))
482 error ("capture of variable-size type %qT that is not an N3639 array "
483 "of runtime bound", type);
484 if (TREE_CODE (type) == ARRAY_TYPE
485 && variably_modified_type_p (TREE_TYPE (type), NULL_TREE))
486 inform (input_location, "because the array element type %qT has "
487 "variable size", TREE_TYPE (type));
488 type = error_mark_node;
490 else
492 type = lambda_capture_field_type (initializer, explicit_init_p);
493 if (by_reference_p)
495 type = build_reference_type (type);
496 if (!dependent_type_p (type) && !real_lvalue_p (initializer))
497 error ("cannot capture %qE by reference", initializer);
499 else
501 /* Capture by copy requires a complete type. */
502 type = complete_type (type);
503 if (!dependent_type_p (type) && !COMPLETE_TYPE_P (type))
505 error ("capture by copy of incomplete type %qT", type);
506 cxx_incomplete_type_inform (type);
507 return error_mark_node;
512 /* Add __ to the beginning of the field name so that user code
513 won't find the field with name lookup. We can't just leave the name
514 unset because template instantiation uses the name to find
515 instantiated fields. */
516 buf = (char *) alloca (IDENTIFIER_LENGTH (id) + 3);
517 buf[1] = buf[0] = '_';
518 memcpy (buf + 2, IDENTIFIER_POINTER (id),
519 IDENTIFIER_LENGTH (id) + 1);
520 name = get_identifier (buf);
522 /* If TREE_TYPE isn't set, we're still in the introducer, so check
523 for duplicates. */
524 if (!LAMBDA_EXPR_CLOSURE (lambda))
526 if (IDENTIFIER_MARKED (name))
528 pedwarn (input_location, 0,
529 "already captured %qD in lambda expression", id);
530 return NULL_TREE;
532 IDENTIFIER_MARKED (name) = true;
535 if (variadic)
536 type = make_pack_expansion (type);
538 /* Make member variable. */
539 member = build_decl (input_location, FIELD_DECL, name, type);
540 DECL_VLA_CAPTURE_P (member) = vla;
542 if (!explicit_init_p)
543 /* Normal captures are invisible to name lookup but uses are replaced
544 with references to the capture field; we implement this by only
545 really making them invisible in unevaluated context; see
546 qualify_lookup. For now, let's make explicitly initialized captures
547 always visible. */
548 DECL_NORMAL_CAPTURE_P (member) = true;
550 if (id == this_identifier)
551 LAMBDA_EXPR_THIS_CAPTURE (lambda) = member;
553 /* Add it to the appropriate closure class if we've started it. */
554 if (current_class_type
555 && current_class_type == LAMBDA_EXPR_CLOSURE (lambda))
556 finish_member_declaration (member);
558 tree listmem = member;
559 if (variadic)
561 listmem = make_pack_expansion (member);
562 initializer = orig_init;
564 LAMBDA_EXPR_CAPTURE_LIST (lambda)
565 = tree_cons (listmem, initializer, LAMBDA_EXPR_CAPTURE_LIST (lambda));
567 if (LAMBDA_EXPR_CLOSURE (lambda))
568 return build_capture_proxy (member);
569 /* For explicit captures we haven't started the function yet, so we wait
570 and build the proxy from cp_parser_lambda_body. */
571 return NULL_TREE;
574 /* Register all the capture members on the list CAPTURES, which is the
575 LAMBDA_EXPR_CAPTURE_LIST for the lambda after the introducer. */
577 void
578 register_capture_members (tree captures)
580 if (captures == NULL_TREE)
581 return;
583 register_capture_members (TREE_CHAIN (captures));
585 tree field = TREE_PURPOSE (captures);
586 if (PACK_EXPANSION_P (field))
587 field = PACK_EXPANSION_PATTERN (field);
589 /* We set this in add_capture to avoid duplicates. */
590 IDENTIFIER_MARKED (DECL_NAME (field)) = false;
591 finish_member_declaration (field);
594 /* Similar to add_capture, except this works on a stack of nested lambdas.
595 BY_REFERENCE_P in this case is derived from the default capture mode.
596 Returns the capture for the lambda at the bottom of the stack. */
598 tree
599 add_default_capture (tree lambda_stack, tree id, tree initializer)
601 bool this_capture_p = (id == this_identifier);
603 tree var = NULL_TREE;
605 tree saved_class_type = current_class_type;
607 tree node;
609 for (node = lambda_stack;
610 node;
611 node = TREE_CHAIN (node))
613 tree lambda = TREE_VALUE (node);
615 current_class_type = LAMBDA_EXPR_CLOSURE (lambda);
616 if (DECL_PACK_P (initializer))
617 initializer = make_pack_expansion (initializer);
618 var = add_capture (lambda,
620 initializer,
621 /*by_reference_p=*/
622 (!this_capture_p
623 && (LAMBDA_EXPR_DEFAULT_CAPTURE_MODE (lambda)
624 == CPLD_REFERENCE)),
625 /*explicit_init_p=*/false);
626 initializer = convert_from_reference (var);
629 current_class_type = saved_class_type;
631 return var;
634 /* Return the capture pertaining to a use of 'this' in LAMBDA, in the
635 form of an INDIRECT_REF, possibly adding it through default
636 capturing, if ADD_CAPTURE_P is true. */
638 tree
639 lambda_expr_this_capture (tree lambda, bool add_capture_p)
641 tree result;
643 tree this_capture = LAMBDA_EXPR_THIS_CAPTURE (lambda);
645 /* In unevaluated context this isn't an odr-use, so don't capture. */
646 if (cp_unevaluated_operand)
647 add_capture_p = false;
649 /* Try to default capture 'this' if we can. */
650 if (!this_capture
651 && (!add_capture_p
652 || LAMBDA_EXPR_DEFAULT_CAPTURE_MODE (lambda) != CPLD_NONE))
654 tree lambda_stack = NULL_TREE;
655 tree init = NULL_TREE;
657 /* If we are in a lambda function, we can move out until we hit:
658 1. a non-lambda function or NSDMI,
659 2. a lambda function capturing 'this', or
660 3. a non-default capturing lambda function. */
661 for (tree tlambda = lambda; ;)
663 lambda_stack = tree_cons (NULL_TREE,
664 tlambda,
665 lambda_stack);
667 if (LAMBDA_EXPR_EXTRA_SCOPE (tlambda)
668 && TREE_CODE (LAMBDA_EXPR_EXTRA_SCOPE (tlambda)) == FIELD_DECL)
670 /* In an NSDMI, we don't have a function to look up the decl in,
671 but the fake 'this' pointer that we're using for parsing is
672 in scope_chain. */
673 init = scope_chain->x_current_class_ptr;
674 gcc_checking_assert
675 (init && (TREE_TYPE (TREE_TYPE (init))
676 == current_nonlambda_class_type ()));
677 break;
680 tree closure_decl = TYPE_NAME (LAMBDA_EXPR_CLOSURE (tlambda));
681 tree containing_function = decl_function_context (closure_decl);
683 if (containing_function == NULL_TREE)
684 /* We ran out of scopes; there's no 'this' to capture. */
685 break;
687 if (!LAMBDA_FUNCTION_P (containing_function))
689 /* We found a non-lambda function. */
690 if (DECL_NONSTATIC_MEMBER_FUNCTION_P (containing_function))
691 /* First parameter is 'this'. */
692 init = DECL_ARGUMENTS (containing_function);
693 break;
696 tlambda
697 = CLASSTYPE_LAMBDA_EXPR (DECL_CONTEXT (containing_function));
699 if (LAMBDA_EXPR_THIS_CAPTURE (tlambda))
701 /* An outer lambda has already captured 'this'. */
702 init = LAMBDA_EXPR_THIS_CAPTURE (tlambda);
703 break;
706 if (LAMBDA_EXPR_DEFAULT_CAPTURE_MODE (tlambda) == CPLD_NONE)
707 /* An outer lambda won't let us capture 'this'. */
708 break;
711 if (init)
713 if (add_capture_p)
714 this_capture = add_default_capture (lambda_stack,
715 /*id=*/this_identifier,
716 init);
717 else
718 this_capture = init;
722 if (cp_unevaluated_operand)
723 result = this_capture;
724 else if (!this_capture)
726 if (add_capture_p)
728 error ("%<this%> was not captured for this lambda function");
729 result = error_mark_node;
731 else
732 result = NULL_TREE;
734 else
736 /* To make sure that current_class_ref is for the lambda. */
737 gcc_assert (TYPE_MAIN_VARIANT (TREE_TYPE (current_class_ref))
738 == LAMBDA_EXPR_CLOSURE (lambda));
740 result = this_capture;
742 /* If 'this' is captured, each use of 'this' is transformed into an
743 access to the corresponding unnamed data member of the closure
744 type cast (_expr.cast_ 5.4) to the type of 'this'. [ The cast
745 ensures that the transformed expression is an rvalue. ] */
746 result = rvalue (result);
749 return result;
752 /* We don't want to capture 'this' until we know we need it, i.e. after
753 overload resolution has chosen a non-static member function. At that
754 point we call this function to turn a dummy object into a use of the
755 'this' capture. */
757 tree
758 maybe_resolve_dummy (tree object, bool add_capture_p)
760 if (!is_dummy_object (object))
761 return object;
763 tree type = TYPE_MAIN_VARIANT (TREE_TYPE (object));
764 gcc_assert (!TYPE_PTR_P (type));
766 if (type != current_class_type
767 && current_class_type
768 && LAMBDA_TYPE_P (current_class_type)
769 && lambda_function (current_class_type)
770 && DERIVED_FROM_P (type, current_nonlambda_class_type ()))
772 /* In a lambda, need to go through 'this' capture. */
773 tree lam = CLASSTYPE_LAMBDA_EXPR (current_class_type);
774 tree cap = lambda_expr_this_capture (lam, add_capture_p);
775 if (cap && cap != error_mark_node)
776 object = build_x_indirect_ref (EXPR_LOCATION (object), cap,
777 RO_NULL, tf_warning_or_error);
780 return object;
783 /* Returns the innermost non-lambda function. */
785 tree
786 current_nonlambda_function (void)
788 tree fn = current_function_decl;
789 while (fn && LAMBDA_FUNCTION_P (fn))
790 fn = decl_function_context (fn);
791 return fn;
794 /* Returns the method basetype of the innermost non-lambda function, or
795 NULL_TREE if none. */
797 tree
798 nonlambda_method_basetype (void)
800 tree fn, type;
801 if (!current_class_ref)
802 return NULL_TREE;
804 type = current_class_type;
805 if (!LAMBDA_TYPE_P (type))
806 return type;
808 /* Find the nearest enclosing non-lambda function. */
809 fn = TYPE_NAME (type);
811 fn = decl_function_context (fn);
812 while (fn && LAMBDA_FUNCTION_P (fn));
814 if (!fn || !DECL_NONSTATIC_MEMBER_FUNCTION_P (fn))
815 return NULL_TREE;
817 return TYPE_METHOD_BASETYPE (TREE_TYPE (fn));
820 /* Like current_scope, but looking through lambdas. */
822 tree
823 current_nonlambda_scope (void)
825 tree scope = current_scope ();
826 for (;;)
828 if (TREE_CODE (scope) == FUNCTION_DECL
829 && LAMBDA_FUNCTION_P (scope))
831 scope = CP_TYPE_CONTEXT (DECL_CONTEXT (scope));
832 continue;
834 else if (LAMBDA_TYPE_P (scope))
836 scope = CP_TYPE_CONTEXT (scope);
837 continue;
839 break;
841 return scope;
844 /* Helper function for maybe_add_lambda_conv_op; build a CALL_EXPR with
845 indicated FN and NARGS, but do not initialize the return type or any of the
846 argument slots. */
848 static tree
849 prepare_op_call (tree fn, int nargs)
851 tree t;
853 t = build_vl_exp (CALL_EXPR, nargs + 3);
854 CALL_EXPR_FN (t) = fn;
855 CALL_EXPR_STATIC_CHAIN (t) = NULL;
857 return t;
860 /* If the closure TYPE has a static op(), also add a conversion to function
861 pointer. */
863 void
864 maybe_add_lambda_conv_op (tree type)
866 bool nested = (cfun != NULL);
867 bool nested_def = decl_function_context (TYPE_MAIN_DECL (type));
868 tree callop = lambda_function (type);
870 if (LAMBDA_EXPR_CAPTURE_LIST (CLASSTYPE_LAMBDA_EXPR (type)) != NULL_TREE)
871 return;
873 if (processing_template_decl)
874 return;
876 bool const generic_lambda_p
877 = (DECL_TEMPLATE_INFO (callop)
878 && DECL_TEMPLATE_RESULT (DECL_TI_TEMPLATE (callop)) == callop);
880 if (!generic_lambda_p && DECL_INITIAL (callop) == NULL_TREE)
882 /* If the op() wasn't instantiated due to errors, give up. */
883 gcc_assert (errorcount || sorrycount);
884 return;
887 /* Non-template conversion operators are defined directly with build_call_a
888 and using DIRECT_ARGVEC for arguments (including 'this'). Templates are
889 deferred and the CALL is built in-place. In the case of a deduced return
890 call op, the decltype expression, DECLTYPE_CALL, used as a substitute for
891 the return type is also built in-place. The arguments of DECLTYPE_CALL in
892 the return expression may differ in flags from those in the body CALL. In
893 particular, parameter pack expansions are marked PACK_EXPANSION_LOCAL_P in
894 the body CALL, but not in DECLTYPE_CALL. */
896 vec<tree, va_gc> *direct_argvec = 0;
897 tree decltype_call = 0, call = 0;
898 tree fn_result = TREE_TYPE (TREE_TYPE (callop));
900 if (generic_lambda_p)
902 /* Prepare the dependent member call for the static member function
903 '_FUN' and, potentially, prepare another call to be used in a decltype
904 return expression for a deduced return call op to allow for simple
905 implementation of the conversion operator. */
907 tree instance = build_nop (type, null_pointer_node);
908 tree objfn = build_min (COMPONENT_REF, NULL_TREE,
909 instance, DECL_NAME (callop), NULL_TREE);
910 int nargs = list_length (DECL_ARGUMENTS (callop)) - 1;
912 call = prepare_op_call (objfn, nargs);
913 if (type_uses_auto (fn_result))
914 decltype_call = prepare_op_call (objfn, nargs);
916 else
918 direct_argvec = make_tree_vector ();
919 direct_argvec->quick_push (build1 (NOP_EXPR,
920 TREE_TYPE (DECL_ARGUMENTS (callop)),
921 null_pointer_node));
924 /* Copy CALLOP's argument list (as per 'copy_list') as FN_ARGS in order to
925 declare the static member function "_FUN" below. For each arg append to
926 DIRECT_ARGVEC (for the non-template case) or populate the pre-allocated
927 call args (for the template case). If a parameter pack is found, expand
928 it, flagging it as PACK_EXPANSION_LOCAL_P for the body call. */
930 tree fn_args = NULL_TREE;
932 int ix = 0;
933 tree src = DECL_CHAIN (DECL_ARGUMENTS (callop));
934 tree tgt;
936 while (src)
938 tree new_node = copy_node (src);
940 if (!fn_args)
941 fn_args = tgt = new_node;
942 else
944 TREE_CHAIN (tgt) = new_node;
945 tgt = new_node;
948 mark_exp_read (tgt);
950 if (generic_lambda_p)
952 if (DECL_PACK_P (tgt))
954 tree a = make_pack_expansion (tgt);
955 if (decltype_call)
956 CALL_EXPR_ARG (decltype_call, ix) = copy_node (a);
957 PACK_EXPANSION_LOCAL_P (a) = true;
958 CALL_EXPR_ARG (call, ix) = a;
960 else
962 tree a = convert_from_reference (tgt);
963 CALL_EXPR_ARG (call, ix) = a;
964 if (decltype_call)
965 CALL_EXPR_ARG (decltype_call, ix) = copy_node (a);
967 ++ix;
969 else
970 vec_safe_push (direct_argvec, tgt);
972 src = TREE_CHAIN (src);
977 if (generic_lambda_p)
979 if (decltype_call)
981 ++processing_template_decl;
982 fn_result = finish_decltype_type
983 (decltype_call, /*id_expression_or_member_access_p=*/false,
984 tf_warning_or_error);
985 --processing_template_decl;
988 else
989 call = build_call_a (callop,
990 direct_argvec->length (),
991 direct_argvec->address ());
993 CALL_FROM_THUNK_P (call) = 1;
995 tree stattype = build_function_type (fn_result, FUNCTION_ARG_CHAIN (callop));
997 /* First build up the conversion op. */
999 tree rettype = build_pointer_type (stattype);
1000 tree name = mangle_conv_op_name_for_type (rettype);
1001 tree thistype = cp_build_qualified_type (type, TYPE_QUAL_CONST);
1002 tree fntype = build_method_type_directly (thistype, rettype, void_list_node);
1003 tree convfn = build_lang_decl (FUNCTION_DECL, name, fntype);
1004 tree fn = convfn;
1005 DECL_SOURCE_LOCATION (fn) = DECL_SOURCE_LOCATION (callop);
1007 if (TARGET_PTRMEMFUNC_VBIT_LOCATION == ptrmemfunc_vbit_in_pfn
1008 && DECL_ALIGN (fn) < 2 * BITS_PER_UNIT)
1009 DECL_ALIGN (fn) = 2 * BITS_PER_UNIT;
1011 SET_OVERLOADED_OPERATOR_CODE (fn, TYPE_EXPR);
1012 grokclassfn (type, fn, NO_SPECIAL);
1013 set_linkage_according_to_type (type, fn);
1014 rest_of_decl_compilation (fn, toplevel_bindings_p (), at_eof);
1015 DECL_IN_AGGR_P (fn) = 1;
1016 DECL_ARTIFICIAL (fn) = 1;
1017 DECL_NOT_REALLY_EXTERN (fn) = 1;
1018 DECL_DECLARED_INLINE_P (fn) = 1;
1019 DECL_ARGUMENTS (fn) = build_this_parm (fntype, TYPE_QUAL_CONST);
1020 if (nested_def)
1021 DECL_INTERFACE_KNOWN (fn) = 1;
1023 if (generic_lambda_p)
1024 fn = add_inherited_template_parms (fn, DECL_TI_TEMPLATE (callop));
1026 add_method (type, fn, NULL_TREE);
1028 /* Generic thunk code fails for varargs; we'll complain in mark_used if
1029 the conversion op is used. */
1030 if (varargs_function_p (callop))
1032 DECL_DELETED_FN (fn) = 1;
1033 return;
1036 /* Now build up the thunk to be returned. */
1038 name = get_identifier ("_FUN");
1039 tree statfn = build_lang_decl (FUNCTION_DECL, name, stattype);
1040 fn = statfn;
1041 DECL_SOURCE_LOCATION (fn) = DECL_SOURCE_LOCATION (callop);
1042 if (TARGET_PTRMEMFUNC_VBIT_LOCATION == ptrmemfunc_vbit_in_pfn
1043 && DECL_ALIGN (fn) < 2 * BITS_PER_UNIT)
1044 DECL_ALIGN (fn) = 2 * BITS_PER_UNIT;
1045 grokclassfn (type, fn, NO_SPECIAL);
1046 set_linkage_according_to_type (type, fn);
1047 rest_of_decl_compilation (fn, toplevel_bindings_p (), at_eof);
1048 DECL_IN_AGGR_P (fn) = 1;
1049 DECL_ARTIFICIAL (fn) = 1;
1050 DECL_NOT_REALLY_EXTERN (fn) = 1;
1051 DECL_DECLARED_INLINE_P (fn) = 1;
1052 DECL_STATIC_FUNCTION_P (fn) = 1;
1053 DECL_ARGUMENTS (fn) = fn_args;
1054 for (tree arg = fn_args; arg; arg = DECL_CHAIN (arg))
1056 /* Avoid duplicate -Wshadow warnings. */
1057 DECL_NAME (arg) = NULL_TREE;
1058 DECL_CONTEXT (arg) = fn;
1060 if (nested_def)
1061 DECL_INTERFACE_KNOWN (fn) = 1;
1063 if (generic_lambda_p)
1064 fn = add_inherited_template_parms (fn, DECL_TI_TEMPLATE (callop));
1066 add_method (type, fn, NULL_TREE);
1068 if (nested)
1069 push_function_context ();
1070 else
1071 /* Still increment function_depth so that we don't GC in the
1072 middle of an expression. */
1073 ++function_depth;
1075 /* Generate the body of the thunk. */
1077 start_preparsed_function (statfn, NULL_TREE,
1078 SF_PRE_PARSED | SF_INCLASS_INLINE);
1079 if (DECL_ONE_ONLY (statfn))
1081 /* Put the thunk in the same comdat group as the call op. */
1082 cgraph_node::get_create (statfn)->add_to_same_comdat_group
1083 (cgraph_node::get_create (callop));
1085 tree body = begin_function_body ();
1086 tree compound_stmt = begin_compound_stmt (0);
1087 if (!generic_lambda_p)
1089 set_flags_from_callee (call);
1090 if (MAYBE_CLASS_TYPE_P (TREE_TYPE (call)))
1091 call = build_cplus_new (TREE_TYPE (call), call, tf_warning_or_error);
1093 call = convert_from_reference (call);
1094 finish_return_stmt (call);
1096 finish_compound_stmt (compound_stmt);
1097 finish_function_body (body);
1099 fn = finish_function (/*inline*/2);
1100 if (!generic_lambda_p)
1101 expand_or_defer_fn (fn);
1103 /* Generate the body of the conversion op. */
1105 start_preparsed_function (convfn, NULL_TREE,
1106 SF_PRE_PARSED | SF_INCLASS_INLINE);
1107 body = begin_function_body ();
1108 compound_stmt = begin_compound_stmt (0);
1110 /* decl_needed_p needs to see that it's used. */
1111 TREE_USED (statfn) = 1;
1112 finish_return_stmt (decay_conversion (statfn, tf_warning_or_error));
1114 finish_compound_stmt (compound_stmt);
1115 finish_function_body (body);
1117 fn = finish_function (/*inline*/2);
1118 if (!generic_lambda_p)
1119 expand_or_defer_fn (fn);
1121 if (nested)
1122 pop_function_context ();
1123 else
1124 --function_depth;
1127 /* Returns true iff VAL is a lambda-related declaration which should
1128 be ignored by unqualified lookup. */
1130 bool
1131 is_lambda_ignored_entity (tree val)
1133 /* In unevaluated context, look past normal capture proxies. */
1134 if (cp_unevaluated_operand && is_normal_capture_proxy (val))
1135 return true;
1137 /* Always ignore lambda fields, their names are only for debugging. */
1138 if (TREE_CODE (val) == FIELD_DECL
1139 && CLASSTYPE_LAMBDA_EXPR (DECL_CONTEXT (val)))
1140 return true;
1142 /* None of the lookups that use qualify_lookup want the op() from the
1143 lambda; they want the one from the enclosing class. */
1144 if (TREE_CODE (val) == FUNCTION_DECL && LAMBDA_FUNCTION_P (val))
1145 return true;
1147 return false;