aboutsummaryrefslogtreecommitdiff
path: root/gcc/ada/gcc-interface/trans.cc
diff options
context:
space:
mode:
Diffstat (limited to 'gcc/ada/gcc-interface/trans.cc')
-rw-r--r--gcc/ada/gcc-interface/trans.cc11009
1 files changed, 11009 insertions, 0 deletions
diff --git a/gcc/ada/gcc-interface/trans.cc b/gcc/ada/gcc-interface/trans.cc
new file mode 100644
index 0000000..39059cb
--- /dev/null
+++ b/gcc/ada/gcc-interface/trans.cc
@@ -0,0 +1,11009 @@
+/****************************************************************************
+ * *
+ * GNAT COMPILER COMPONENTS *
+ * *
+ * T R A N S *
+ * *
+ * C Implementation File *
+ * *
+ * Copyright (C) 1992-2021, Free Software Foundation, Inc. *
+ * *
+ * GNAT is free software; you can redistribute it and/or modify it under *
+ * terms of the GNU General Public License as published by the Free Soft- *
+ * ware Foundation; either version 3, or (at your option) any later ver- *
+ * sion. GNAT is distributed in the hope that it will be useful, but WITH- *
+ * OUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY *
+ * or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License *
+ * for more details. You should have received a copy of the GNU General *
+ * Public License distributed with GNAT; see file COPYING3. If not see *
+ * <http://www.gnu.org/licenses/>. *
+ * *
+ * GNAT was originally developed by the GNAT team at New York University. *
+ * Extensive contributions were provided by Ada Core Technologies Inc. *
+ * *
+ ****************************************************************************/
+
+#include "config.h"
+#include "system.h"
+#include "coretypes.h"
+#include "target.h"
+#include "function.h"
+#include "bitmap.h"
+#include "tree.h"
+#include "gimple-expr.h"
+#include "stringpool.h"
+#include "cgraph.h"
+#include "predict.h"
+#include "diagnostic.h"
+#include "alias.h"
+#include "fold-const.h"
+#include "stor-layout.h"
+#include "stmt.h"
+#include "varasm.h"
+#include "output.h"
+#include "debug.h"
+#include "libfuncs.h" /* For set_stack_check_libfunc. */
+#include "tree-iterator.h"
+#include "gimplify.h"
+#include "opts.h"
+#include "common/common-target.h"
+#include "gomp-constants.h"
+#include "stringpool.h"
+#include "attribs.h"
+#include "tree-nested.h"
+
+#include "ada.h"
+#include "adadecode.h"
+#include "types.h"
+#include "atree.h"
+#include "namet.h"
+#include "nlists.h"
+#include "snames.h"
+#include "stringt.h"
+#include "uintp.h"
+#include "urealp.h"
+#include "fe.h"
+#include "sinfo.h"
+#include "einfo.h"
+#include "gadaint.h"
+#include "ada-tree.h"
+#include "gigi.h"
+
+/* We should avoid allocating more than ALLOCA_THRESHOLD bytes via alloca,
+ for fear of running out of stack space. If we need more, we use xmalloc
+ instead. */
+#define ALLOCA_THRESHOLD 1000
+
+/* Pointers to front-end tables accessed through macros. */
+Node_Header *Node_Offsets_Ptr;
+any_slot *Slots_Ptr;
+Node_Id *Next_Node_Ptr;
+Node_Id *Prev_Node_Ptr;
+struct Elist_Header *Elists_Ptr;
+struct Elmt_Item *Elmts_Ptr;
+struct String_Entry *Strings_Ptr;
+Char_Code *String_Chars_Ptr;
+struct List_Header *List_Headers_Ptr;
+
+/* Highest number in the front-end node table. */
+int max_gnat_nodes;
+
+/* True when gigi is being called on an analyzed but unexpanded
+ tree, and the only purpose of the call is to properly annotate
+ types with representation information. */
+bool type_annotate_only;
+
+/* List of N_Validate_Unchecked_Conversion nodes in the unit. */
+static vec<Node_Id> gnat_validate_uc_list;
+
+/* List of expressions of pragma Compile_Time_{Error|Warning} in the unit. */
+static vec<Node_Id> gnat_compile_time_expr_list;
+
+/* When not optimizing, we cache the 'First, 'Last and 'Length attributes
+ of unconstrained array IN parameters to avoid emitting a great deal of
+ redundant instructions to recompute them each time. */
+struct GTY (()) parm_attr_d {
+ int id; /* GTY doesn't like Entity_Id. */
+ int dim;
+ tree first;
+ tree last;
+ tree length;
+};
+
+typedef struct parm_attr_d *parm_attr;
+
+/* Structure used to record information for a function. */
+struct GTY(()) language_function {
+ vec<parm_attr, va_gc> *parm_attr_cache;
+ bitmap named_ret_val;
+ vec<tree, va_gc> *other_ret_val;
+ int gnat_ret;
+};
+
+#define f_parm_attr_cache \
+ DECL_STRUCT_FUNCTION (current_function_decl)->language->parm_attr_cache
+
+#define f_named_ret_val \
+ DECL_STRUCT_FUNCTION (current_function_decl)->language->named_ret_val
+
+#define f_other_ret_val \
+ DECL_STRUCT_FUNCTION (current_function_decl)->language->other_ret_val
+
+#define f_gnat_ret \
+ DECL_STRUCT_FUNCTION (current_function_decl)->language->gnat_ret
+
+/* A structure used to gather together information about a statement group.
+ We use this to gather related statements, for example the "then" part
+ of a IF. In the case where it represents a lexical scope, we may also
+ have a BLOCK node corresponding to it and/or cleanups. */
+
+struct GTY((chain_next ("%h.previous"))) stmt_group {
+ struct stmt_group *previous; /* Previous code group. */
+ tree stmt_list; /* List of statements for this code group. */
+ tree block; /* BLOCK for this code group, if any. */
+ tree cleanups; /* Cleanups for this code group, if any. */
+};
+
+static GTY(()) struct stmt_group *current_stmt_group;
+
+/* List of unused struct stmt_group nodes. */
+static GTY((deletable)) struct stmt_group *stmt_group_free_list;
+
+/* A structure used to record information on elaboration procedures
+ we've made and need to process.
+
+ ??? gnat_node should be Node_Id, but gengtype gets confused. */
+
+struct GTY((chain_next ("%h.next"))) elab_info {
+ struct elab_info *next; /* Pointer to next in chain. */
+ tree elab_proc; /* Elaboration procedure. */
+ int gnat_node; /* The N_Compilation_Unit. */
+};
+
+static GTY(()) struct elab_info *elab_info_list;
+
+/* Stack of exception pointer variables. Each entry is the VAR_DECL
+ that stores the address of the raised exception. Nonzero means we
+ are in an exception handler. Not used in the zero-cost case. */
+static GTY(()) vec<tree, va_gc> *gnu_except_ptr_stack;
+
+/* In ZCX case, current exception pointer. Used to re-raise it. */
+static GTY(()) tree gnu_incoming_exc_ptr;
+
+/* Stack for storing the current elaboration procedure decl. */
+static GTY(()) vec<tree, va_gc> *gnu_elab_proc_stack;
+
+/* Stack of labels to be used as a goto target instead of a return in
+ some functions. See processing for N_Subprogram_Body. */
+static GTY(()) vec<tree, va_gc> *gnu_return_label_stack;
+
+/* Stack of variable for the return value of a function with copy-in/copy-out
+ parameters. See processing for N_Subprogram_Body. */
+static GTY(()) vec<tree, va_gc> *gnu_return_var_stack;
+
+/* Structure used to record information for a range check. */
+struct GTY(()) range_check_info_d {
+ tree low_bound;
+ tree high_bound;
+ tree disp;
+ bool neg_p;
+ tree type;
+ tree invariant_cond;
+ tree inserted_cond;
+};
+
+typedef struct range_check_info_d *range_check_info;
+
+/* Structure used to record information for a loop. */
+struct GTY(()) loop_info_d {
+ tree fndecl;
+ tree stmt;
+ tree loop_var;
+ tree low_bound;
+ tree high_bound;
+ tree omp_loop_clauses;
+ tree omp_construct_clauses;
+ enum tree_code omp_code;
+ vec<range_check_info, va_gc> *checks;
+ vec<tree, va_gc> *invariants;
+};
+
+typedef struct loop_info_d *loop_info;
+
+/* Stack of loop_info structures associated with LOOP_STMT nodes. */
+static GTY(()) vec<loop_info, va_gc> *gnu_loop_stack;
+
+/* The stacks for N_{Push,Pop}_*_Label. */
+static vec<Entity_Id> gnu_constraint_error_label_stack;
+static vec<Entity_Id> gnu_storage_error_label_stack;
+static vec<Entity_Id> gnu_program_error_label_stack;
+
+/* Map GNAT tree codes to GCC tree codes for simple expressions. */
+static enum tree_code gnu_codes[Number_Node_Kinds];
+
+static void init_code_table (void);
+static tree get_elaboration_procedure (void);
+static void Compilation_Unit_to_gnu (Node_Id);
+static bool empty_stmt_list_p (tree);
+static void record_code_position (Node_Id);
+static void insert_code_for (Node_Id);
+static void add_cleanup (tree, Node_Id);
+static void add_stmt_list (List_Id);
+static tree build_stmt_group (List_Id, bool);
+static inline bool stmt_group_may_fallthru (void);
+static enum gimplify_status gnat_gimplify_stmt (tree *);
+static void elaborate_all_entities (Node_Id);
+static void process_freeze_entity (Node_Id);
+static void process_decls (List_Id, List_Id, Node_Id, bool, bool);
+static tree emit_check (tree, tree, int, Node_Id);
+static tree build_unary_op_trapv (enum tree_code, tree, tree, Node_Id);
+static tree build_binary_op_trapv (enum tree_code, tree, tree, tree, Node_Id);
+static tree convert_with_check (Entity_Id, tree, bool, bool, Node_Id);
+static bool addressable_p (tree, tree);
+static tree assoc_to_constructor (Entity_Id, Node_Id, tree);
+static tree pos_to_constructor (Node_Id, tree);
+static void validate_unchecked_conversion (Node_Id);
+static void set_expr_location_from_node (tree, Node_Id, bool = false);
+static void set_gnu_expr_location_from_node (tree, Node_Id);
+static bool set_end_locus_from_node (tree, Node_Id);
+static int lvalue_required_p (Node_Id, tree, bool, bool);
+static tree build_raise_check (int, enum exception_info_kind);
+static tree create_init_temporary (const char *, tree, tree *, Node_Id);
+static bool maybe_make_gnu_thunk (Entity_Id gnat_thunk, tree gnu_thunk);
+
+/* This makes gigi's file_info_ptr visible in this translation unit,
+ so that Sloc_to_locus can look it up when deciding whether to map
+ decls to instances. */
+
+static struct File_Info_Type *file_map;
+
+/* Return the string of the identifier allocated for the file name Id. */
+
+static const char*
+File_Name_to_gnu (Name_Id Id)
+{
+ /* __gnat_to_canonical_file_spec translates file names from pragmas
+ Source_Reference that contain host style syntax not understood by GDB. */
+ const char *name = __gnat_to_canonical_file_spec (Get_Name_String (Id));
+
+ /* Use the identifier table to make a permanent copy of the file name as
+ the name table gets reallocated after Gigi returns but before all the
+ debugging information is output. */
+ return IDENTIFIER_POINTER (get_identifier (name));
+}
+
+/* This is the main program of the back-end. It sets up all the table
+ structures and then generates code. */
+
+void
+gigi (Node_Id gnat_root,
+ int max_gnat_node,
+ int number_name ATTRIBUTE_UNUSED,
+ Node_Header *node_offsets_ptr,
+ any_slot *slots_ptr,
+ Node_Id *next_node_ptr,
+ Node_Id *prev_node_ptr,
+ struct Elist_Header *elists_ptr,
+ struct Elmt_Item *elmts_ptr,
+ struct String_Entry *strings_ptr,
+ Char_Code *string_chars_ptr,
+ struct List_Header *list_headers_ptr,
+ Nat number_file,
+ struct File_Info_Type *file_info_ptr,
+ Entity_Id standard_boolean,
+ Entity_Id standard_integer,
+ Entity_Id standard_character,
+ Entity_Id standard_long_long_float,
+ Entity_Id standard_exception_type,
+ Int gigi_operating_mode)
+{
+ Node_Id gnat_iter;
+ Entity_Id gnat_literal;
+ tree t, ftype, int64_type;
+ struct elab_info *info;
+ int i;
+
+ max_gnat_nodes = max_gnat_node;
+
+ Node_Offsets_Ptr = node_offsets_ptr;
+ Slots_Ptr = slots_ptr;
+ Next_Node_Ptr = next_node_ptr;
+ Prev_Node_Ptr = prev_node_ptr;
+ Elists_Ptr = elists_ptr;
+ Elmts_Ptr = elmts_ptr;
+ Strings_Ptr = strings_ptr;
+ String_Chars_Ptr = string_chars_ptr;
+ List_Headers_Ptr = list_headers_ptr;
+
+ type_annotate_only = (gigi_operating_mode == 1);
+
+ if (Generate_SCO_Instance_Table != 0)
+ {
+ file_map = file_info_ptr;
+ maybe_create_decl_to_instance_map (number_file);
+ }
+
+ for (i = 0; i < number_file; i++)
+ {
+ /* We rely on the order isomorphism between files and line maps. */
+ if ((int) LINEMAPS_ORDINARY_USED (line_table) != i)
+ {
+ gcc_assert (i > 0);
+ error ("%s contains too many lines",
+ File_Name_to_gnu (file_info_ptr[i - 1].File_Name));
+ }
+
+ /* We create the line map for a source file at once, with a fixed number
+ of columns chosen to avoid jumping over the next power of 2. */
+ linemap_add (line_table, LC_ENTER, 0,
+ File_Name_to_gnu (file_info_ptr[i].File_Name), 1);
+ linemap_line_start (line_table, file_info_ptr[i].Num_Source_Lines, 252);
+ linemap_position_for_column (line_table, 252 - 1);
+ linemap_add (line_table, LC_LEAVE, 0, NULL, 0);
+ }
+
+ gcc_assert (Nkind (gnat_root) == N_Compilation_Unit);
+
+ /* Declare the name of the compilation unit as the first global
+ name in order to make the middle-end fully deterministic. */
+ t = create_concat_name (Defining_Entity (Unit (gnat_root)), NULL);
+ first_global_object_name = ggc_strdup (IDENTIFIER_POINTER (t));
+
+ /* Initialize ourselves. */
+ init_code_table ();
+ init_gnat_decl ();
+ init_gnat_utils ();
+
+ /* If we are just annotating types, give VOID_TYPE zero sizes to avoid
+ errors. */
+ if (type_annotate_only)
+ {
+ TYPE_SIZE (void_type_node) = bitsize_zero_node;
+ TYPE_SIZE_UNIT (void_type_node) = size_zero_node;
+ }
+
+ /* Enable GNAT stack checking method if needed */
+ if (!Stack_Check_Probes_On_Target)
+ set_stack_check_libfunc ("_gnat_stack_check");
+
+ /* Retrieve alignment settings. */
+ double_float_alignment = get_target_double_float_alignment ();
+ double_scalar_alignment = get_target_double_scalar_alignment ();
+
+ /* Record the builtin types. Define `integer' and `character' first so that
+ dbx will output them first. */
+ record_builtin_type ("integer", integer_type_node, false);
+ record_builtin_type ("character", char_type_node, false);
+ record_builtin_type ("boolean", boolean_type_node, false);
+ record_builtin_type ("void", void_type_node, false);
+
+ /* Save the type we made for integer as the type for Standard.Integer. */
+ save_gnu_tree (Base_Type (standard_integer),
+ TYPE_NAME (integer_type_node),
+ false);
+
+ /* Likewise for character as the type for Standard.Character. */
+ finish_character_type (char_type_node);
+ save_gnu_tree (Base_Type (standard_character),
+ TYPE_NAME (char_type_node),
+ false);
+
+ /* Likewise for boolean as the type for Standard.Boolean. */
+ save_gnu_tree (Base_Type (standard_boolean),
+ TYPE_NAME (boolean_type_node),
+ false);
+ gnat_literal = First_Literal (Base_Type (standard_boolean));
+ t = UI_To_gnu (Enumeration_Rep (gnat_literal), boolean_type_node);
+ gcc_assert (t == boolean_false_node);
+ t = create_var_decl (get_entity_name (gnat_literal), NULL_TREE,
+ boolean_type_node, t, true, false, false, false, false,
+ true, false, NULL, gnat_literal);
+ save_gnu_tree (gnat_literal, t, false);
+ gnat_literal = Next_Literal (gnat_literal);
+ t = UI_To_gnu (Enumeration_Rep (gnat_literal), boolean_type_node);
+ gcc_assert (t == boolean_true_node);
+ t = create_var_decl (get_entity_name (gnat_literal), NULL_TREE,
+ boolean_type_node, t, true, false, false, false, false,
+ true, false, NULL, gnat_literal);
+ save_gnu_tree (gnat_literal, t, false);
+
+ /* Declare the building blocks of function nodes. */
+ void_list_node = build_tree_list (NULL_TREE, void_type_node);
+ void_ftype = build_function_type_list (void_type_node, NULL_TREE);
+ ptr_void_ftype = build_pointer_type (void_ftype);
+
+ /* Now declare run-time functions. */
+ malloc_decl
+ = create_subprog_decl (get_identifier ("__gnat_malloc"), NULL_TREE,
+ build_function_type_list (ptr_type_node, sizetype,
+ NULL_TREE),
+ NULL_TREE, is_default, true, true, true, false,
+ false, NULL, Empty);
+ DECL_IS_MALLOC (malloc_decl) = 1;
+
+ free_decl
+ = create_subprog_decl (get_identifier ("__gnat_free"), NULL_TREE,
+ build_function_type_list (void_type_node,
+ ptr_type_node, NULL_TREE),
+ NULL_TREE, is_default, true, true, true, false,
+ false, NULL, Empty);
+
+ realloc_decl
+ = create_subprog_decl (get_identifier ("__gnat_realloc"), NULL_TREE,
+ build_function_type_list (ptr_type_node,
+ ptr_type_node, sizetype,
+ NULL_TREE),
+ NULL_TREE, is_default, true, true, true, false,
+ false, NULL, Empty);
+
+ /* This is used for 64-bit multiplication with overflow checking. */
+ int64_type = gnat_type_for_size (64, 0);
+ mulv64_decl
+ = create_subprog_decl (get_identifier ("__gnat_mulv64"), NULL_TREE,
+ build_function_type_list (int64_type, int64_type,
+ int64_type, NULL_TREE),
+ NULL_TREE, is_default, true, true, true, false,
+ false, NULL, Empty);
+
+ if (Enable_128bit_Types)
+ {
+ tree int128_type = gnat_type_for_size (128, 0);
+ mulv128_decl
+ = create_subprog_decl (get_identifier ("__gnat_mulv128"), NULL_TREE,
+ build_function_type_list (int128_type,
+ int128_type,
+ int128_type,
+ NULL_TREE),
+ NULL_TREE, is_default, true, true, true, false,
+ false, NULL, Empty);
+ }
+
+ /* Name of the _Parent field in tagged record types. */
+ parent_name_id = get_identifier (Get_Name_String (Name_uParent));
+
+ /* Name of the Not_Handled_By_Others field in exception record types. */
+ not_handled_by_others_name_id = get_identifier ("not_handled_by_others");
+
+ /* Make the types and functions used for exception processing. */
+ except_type_node = gnat_to_gnu_type (Base_Type (standard_exception_type));
+
+ for (t = TYPE_FIELDS (except_type_node); t; t = DECL_CHAIN (t))
+ if (DECL_NAME (t) == not_handled_by_others_name_id)
+ {
+ not_handled_by_others_decl = t;
+ break;
+ }
+ gcc_assert (DECL_P (not_handled_by_others_decl));
+
+ jmpbuf_type
+ = build_array_type (gnat_type_for_mode (Pmode, 0),
+ build_index_type (size_int (5)));
+ record_builtin_type ("JMPBUF_T", jmpbuf_type, true);
+ jmpbuf_ptr_type = build_pointer_type (jmpbuf_type);
+
+ /* Functions to get and set the jumpbuf pointer for the current thread. */
+ get_jmpbuf_decl
+ = create_subprog_decl
+ (get_identifier ("system__soft_links__get_jmpbuf_address_soft"),
+ NULL_TREE, build_function_type_list (jmpbuf_ptr_type, NULL_TREE),
+ NULL_TREE, is_default, true, true, true, false, false, NULL, Empty);
+
+ set_jmpbuf_decl
+ = create_subprog_decl
+ (get_identifier ("system__soft_links__set_jmpbuf_address_soft"),
+ NULL_TREE, build_function_type_list (void_type_node, jmpbuf_ptr_type,
+ NULL_TREE),
+ NULL_TREE, is_default, true, true, true, false, false, NULL, Empty);
+
+ get_excptr_decl
+ = create_subprog_decl
+ (get_identifier ("system__soft_links__get_gnat_exception"), NULL_TREE,
+ build_function_type_list (build_pointer_type (except_type_node),
+ NULL_TREE),
+ NULL_TREE, is_default, true, true, true, false, false, NULL, Empty);
+
+ /* setjmp returns an integer and has one operand, which is a pointer to
+ a jmpbuf. */
+ setjmp_decl
+ = create_subprog_decl
+ (get_identifier ("__builtin_setjmp"), NULL_TREE,
+ build_function_type_list (integer_type_node, jmpbuf_ptr_type,
+ NULL_TREE),
+ NULL_TREE, is_default, true, true, true, false, false, NULL, Empty);
+ set_decl_built_in_function (setjmp_decl, BUILT_IN_NORMAL, BUILT_IN_SETJMP);
+
+ /* update_setjmp_buf updates a setjmp buffer from the current stack pointer
+ address. */
+ update_setjmp_buf_decl
+ = create_subprog_decl
+ (get_identifier ("__builtin_update_setjmp_buf"), NULL_TREE,
+ build_function_type_list (void_type_node, jmpbuf_ptr_type, NULL_TREE),
+ NULL_TREE, is_default, true, true, true, false, false, NULL, Empty);
+ set_decl_built_in_function (update_setjmp_buf_decl, BUILT_IN_NORMAL,
+ BUILT_IN_UPDATE_SETJMP_BUF);
+
+ /* Indicate that it never returns. */
+ ftype = build_function_type_list (void_type_node,
+ build_pointer_type (except_type_node),
+ NULL_TREE);
+ ftype = build_qualified_type (ftype, TYPE_QUAL_VOLATILE);
+ raise_nodefer_decl
+ = create_subprog_decl
+ (get_identifier ("__gnat_raise_nodefer_with_msg"), NULL_TREE, ftype,
+ NULL_TREE, is_default, true, true, true, false, false, NULL, Empty);
+
+ set_exception_parameter_decl
+ = create_subprog_decl
+ (get_identifier ("__gnat_set_exception_parameter"), NULL_TREE,
+ build_function_type_list (void_type_node, ptr_type_node, ptr_type_node,
+ NULL_TREE),
+ NULL_TREE, is_default, true, true, true, false, false, NULL, Empty);
+
+ /* Hooks to call when entering/leaving an exception handler. */
+ ftype = build_function_type_list (ptr_type_node,
+ ptr_type_node, NULL_TREE);
+ begin_handler_decl
+ = create_subprog_decl (get_identifier ("__gnat_begin_handler_v1"),
+ NULL_TREE, ftype, NULL_TREE,
+ is_default, true, true, true, false, false, NULL,
+ Empty);
+ /* __gnat_begin_handler_v1 is not a dummy procedure, but we arrange
+ for it not to throw. */
+ TREE_NOTHROW (begin_handler_decl) = 1;
+
+ ftype = build_function_type_list (ptr_type_node,
+ ptr_type_node, ptr_type_node,
+ ptr_type_node, NULL_TREE);
+ end_handler_decl
+ = create_subprog_decl (get_identifier ("__gnat_end_handler_v1"), NULL_TREE,
+ ftype, NULL_TREE,
+ is_default, true, true, true, false, false, NULL,
+ Empty);
+
+ ftype = build_function_type_list (void_type_node, ptr_type_node, NULL_TREE);
+ unhandled_except_decl
+ = create_subprog_decl (get_identifier ("__gnat_unhandled_except_handler"),
+ NULL_TREE, ftype, NULL_TREE,
+ is_default, true, true, true, false, false, NULL,
+ Empty);
+
+ /* Indicate that it never returns. */
+ ftype = build_qualified_type (ftype, TYPE_QUAL_VOLATILE);
+ reraise_zcx_decl
+ = create_subprog_decl (get_identifier ("__gnat_reraise_zcx"), NULL_TREE,
+ ftype, NULL_TREE,
+ is_default, true, true, true, false, false, NULL,
+ Empty);
+
+ /* Dummy objects to materialize "others" and "all others" in the exception
+ tables. These are exported by a-exexpr-gcc.adb, so see this unit for
+ the types to use. */
+ others_decl
+ = create_var_decl (get_identifier ("OTHERS"),
+ get_identifier ("__gnat_others_value"),
+ char_type_node, NULL_TREE,
+ true, false, true, false, false, true, false,
+ NULL, Empty);
+
+ all_others_decl
+ = create_var_decl (get_identifier ("ALL_OTHERS"),
+ get_identifier ("__gnat_all_others_value"),
+ char_type_node, NULL_TREE,
+ true, false, true, false, false, true, false,
+ NULL, Empty);
+
+ unhandled_others_decl
+ = create_var_decl (get_identifier ("UNHANDLED_OTHERS"),
+ get_identifier ("__gnat_unhandled_others_value"),
+ char_type_node, NULL_TREE,
+ true, false, true, false, false, true, false,
+ NULL, Empty);
+
+ /* If in no exception handlers mode, all raise statements are redirected to
+ __gnat_last_chance_handler. No need to redefine raise_nodefer_decl since
+ this procedure will never be called in this mode. */
+ if (No_Exception_Handlers_Set ())
+ {
+ /* Indicate that it never returns. */
+ ftype = build_function_type_list (void_type_node,
+ build_pointer_type (char_type_node),
+ integer_type_node, NULL_TREE);
+ ftype = build_qualified_type (ftype, TYPE_QUAL_VOLATILE);
+ tree decl
+ = create_subprog_decl
+ (get_identifier ("__gnat_last_chance_handler"), NULL_TREE, ftype,
+ NULL_TREE, is_default, true, true, true, false, false, NULL,
+ Empty);
+ for (i = 0; i < (int) ARRAY_SIZE (gnat_raise_decls); i++)
+ gnat_raise_decls[i] = decl;
+ }
+ else
+ {
+ /* Otherwise, make one decl for each exception reason. */
+ for (i = 0; i < (int) ARRAY_SIZE (gnat_raise_decls); i++)
+ gnat_raise_decls[i] = build_raise_check (i, exception_simple);
+ for (i = 0; i < (int) ARRAY_SIZE (gnat_raise_decls_ext); i++)
+ gnat_raise_decls_ext[i]
+ = build_raise_check (i,
+ i == CE_Index_Check_Failed
+ || i == CE_Range_Check_Failed
+ || i == CE_Invalid_Data
+ ? exception_range : exception_column);
+ }
+
+ /* Build the special descriptor type and its null node if needed. */
+ if (TARGET_VTABLE_USES_DESCRIPTORS)
+ {
+ tree null_node = fold_convert (ptr_void_ftype, null_pointer_node);
+ tree field_list = NULL_TREE;
+ int j;
+ vec<constructor_elt, va_gc> *null_vec = NULL;
+ constructor_elt *elt;
+
+ fdesc_type_node = make_node (RECORD_TYPE);
+ vec_safe_grow (null_vec, TARGET_VTABLE_USES_DESCRIPTORS, true);
+ elt = (null_vec->address () + TARGET_VTABLE_USES_DESCRIPTORS - 1);
+
+ for (j = 0; j < TARGET_VTABLE_USES_DESCRIPTORS; j++)
+ {
+ tree field
+ = create_field_decl (NULL_TREE, ptr_void_ftype, fdesc_type_node,
+ NULL_TREE, NULL_TREE, 0, 1);
+ DECL_CHAIN (field) = field_list;
+ field_list = field;
+ elt->index = field;
+ elt->value = null_node;
+ elt--;
+ }
+
+ finish_record_type (fdesc_type_node, nreverse (field_list), 0, false);
+ record_builtin_type ("descriptor", fdesc_type_node, true);
+ null_fdesc_node = gnat_build_constructor (fdesc_type_node, null_vec);
+ }
+
+ longest_float_type_node
+ = get_unpadded_type (Base_Type (standard_long_long_float));
+
+ main_identifier_node = get_identifier ("main");
+
+ /* If we are using the GCC exception mechanism, let GCC know. */
+ if (Back_End_Exceptions ())
+ gnat_init_gcc_eh ();
+
+ /* Initialize the GCC support for FP operations. */
+ gnat_init_gcc_fp ();
+
+ /* Install the builtins we might need, either internally or as user-available
+ facilities for Intrinsic imports. Note that this must be done after the
+ GCC exception mechanism is initialized. */
+ gnat_install_builtins ();
+
+ vec_safe_push (gnu_except_ptr_stack, NULL_TREE);
+
+ gnu_constraint_error_label_stack.safe_push (Empty);
+ gnu_storage_error_label_stack.safe_push (Empty);
+ gnu_program_error_label_stack.safe_push (Empty);
+
+ /* Process any Pragma Ident for the main unit. */
+ if (Present (Ident_String (Main_Unit)))
+ targetm.asm_out.output_ident
+ (TREE_STRING_POINTER (gnat_to_gnu (Ident_String (Main_Unit))));
+
+ /* Force -fno-strict-aliasing if the configuration pragma was seen. */
+ if (No_Strict_Aliasing_CP)
+ flag_strict_aliasing = 0;
+
+ /* Save the current optimization options again after the above possible
+ global_options changes. */
+ optimization_default_node
+ = build_optimization_node (&global_options, &global_options_set);
+ optimization_current_node = optimization_default_node;
+
+ /* Now translate the compilation unit proper. */
+ Compilation_Unit_to_gnu (gnat_root);
+
+ /* Then process the N_Validate_Unchecked_Conversion nodes. We do this at
+ the very end to avoid having to second-guess the front-end when we run
+ into dummy nodes during the regular processing. */
+ for (i = 0; gnat_validate_uc_list.iterate (i, &gnat_iter); i++)
+ validate_unchecked_conversion (gnat_iter);
+ gnat_validate_uc_list.release ();
+
+ /* Finally see if we have any elaboration procedures to deal with. */
+ for (info = elab_info_list; info; info = info->next)
+ {
+ tree gnu_body = DECL_SAVED_TREE (info->elab_proc);
+
+ /* We should have a BIND_EXPR but it may not have any statements in it.
+ If it doesn't have any, we have nothing to do except for setting the
+ flag on the GNAT node. Otherwise, process the function as others. */
+ tree gnu_stmts = gnu_body;
+ if (TREE_CODE (gnu_stmts) == BIND_EXPR)
+ gnu_stmts = BIND_EXPR_BODY (gnu_stmts);
+ if (!gnu_stmts || empty_stmt_list_p (gnu_stmts))
+ Set_Has_No_Elaboration_Code (info->gnat_node, 1);
+ else
+ {
+ begin_subprog_body (info->elab_proc);
+ end_subprog_body (gnu_body);
+ rest_of_subprog_body_compilation (info->elab_proc);
+ }
+ }
+
+ /* Destroy ourselves. */
+ file_map = NULL;
+ destroy_gnat_decl ();
+ destroy_gnat_utils ();
+
+ /* We cannot track the location of errors past this point. */
+ Current_Error_Node = Empty;
+}
+
+/* Return a subprogram decl corresponding to __gnat_rcheck_xx for the given
+ CHECK if KIND is EXCEPTION_SIMPLE, or else to __gnat_rcheck_xx_ext. */
+
+static tree
+build_raise_check (int check, enum exception_info_kind kind)
+{
+ tree result, ftype;
+ const char pfx[] = "__gnat_rcheck_";
+
+ strcpy (Name_Buffer, pfx);
+ Name_Len = sizeof (pfx) - 1;
+ Get_RT_Exception_Name ((enum RT_Exception_Code) check);
+
+ if (kind == exception_simple)
+ {
+ Name_Buffer[Name_Len] = 0;
+ ftype
+ = build_function_type_list (void_type_node,
+ build_pointer_type (char_type_node),
+ integer_type_node, NULL_TREE);
+ }
+ else
+ {
+ tree t = (kind == exception_column ? NULL_TREE : integer_type_node);
+
+ strcpy (Name_Buffer + Name_Len, "_ext");
+ Name_Buffer[Name_Len + 4] = 0;
+ ftype
+ = build_function_type_list (void_type_node,
+ build_pointer_type (char_type_node),
+ integer_type_node, integer_type_node,
+ t, t, NULL_TREE);
+ }
+
+ /* Indicate that it never returns. */
+ ftype = build_qualified_type (ftype, TYPE_QUAL_VOLATILE);
+ result
+ = create_subprog_decl (get_identifier (Name_Buffer), NULL_TREE, ftype,
+ NULL_TREE, is_default, true, true, true, false,
+ false, NULL, Empty);
+
+ return result;
+}
+
+/* Return a positive value if an lvalue is required for GNAT_NODE, which is
+ an N_Attribute_Reference. */
+
+static int
+lvalue_required_for_attribute_p (Node_Id gnat_node)
+{
+ switch (Get_Attribute_Id (Attribute_Name (gnat_node)))
+ {
+ case Attr_Pred:
+ case Attr_Succ:
+ case Attr_First:
+ case Attr_Last:
+ case Attr_Range_Length:
+ case Attr_Length:
+ case Attr_Object_Size:
+ case Attr_Size:
+ case Attr_Value_Size:
+ case Attr_Component_Size:
+ case Attr_Descriptor_Size:
+ case Attr_Max_Size_In_Storage_Elements:
+ case Attr_Min:
+ case Attr_Max:
+ case Attr_Null_Parameter:
+ case Attr_Passed_By_Reference:
+ case Attr_Mechanism_Code:
+ case Attr_Machine:
+ case Attr_Model:
+ return 0;
+
+ case Attr_Address:
+ case Attr_Access:
+ case Attr_Unchecked_Access:
+ case Attr_Unrestricted_Access:
+ case Attr_Code_Address:
+ case Attr_Pool_Address:
+ case Attr_Alignment:
+ case Attr_Bit_Position:
+ case Attr_Position:
+ case Attr_First_Bit:
+ case Attr_Last_Bit:
+ case Attr_Bit:
+ case Attr_Asm_Input:
+ case Attr_Asm_Output:
+ default:
+ return 1;
+ }
+}
+
+/* Return a positive value if an lvalue is required for GNAT_NODE. GNU_TYPE
+ is the type that will be used for GNAT_NODE in the translated GNU tree.
+ CONSTANT indicates whether the underlying object represented by GNAT_NODE
+ is constant in the Ada sense. If it is, ADDRESS_OF_CONSTANT indicates
+ whether its value is the address of another constant. If it isn't, then
+ ADDRESS_OF_CONSTANT is ignored.
+
+ The function climbs up the GNAT tree starting from the node and returns 1
+ upon encountering a node that effectively requires an lvalue downstream.
+ It returns int instead of bool to facilitate usage in non-purely binary
+ logic contexts. */
+
+static int
+lvalue_required_p (Node_Id gnat_node, tree gnu_type, bool constant,
+ bool address_of_constant)
+{
+ Node_Id gnat_parent = Parent (gnat_node), gnat_temp;
+
+ switch (Nkind (gnat_parent))
+ {
+ case N_Reference:
+ return 1;
+
+ case N_Attribute_Reference:
+ return lvalue_required_for_attribute_p (gnat_parent);
+
+ case N_Parameter_Association:
+ case N_Function_Call:
+ case N_Procedure_Call_Statement:
+ /* If the parameter is by reference, an lvalue is required. */
+ return (!constant
+ || must_pass_by_ref (gnu_type)
+ || default_pass_by_ref (gnu_type));
+
+ case N_Pragma_Argument_Association:
+ return lvalue_required_p (gnat_parent, gnu_type, constant,
+ address_of_constant);
+
+ case N_Pragma:
+ if (Is_Pragma_Name (Chars (Pragma_Identifier (gnat_parent))))
+ {
+ const unsigned char id
+ = Get_Pragma_Id (Chars (Pragma_Identifier (gnat_parent)));
+ return id == Pragma_Inspection_Point;
+ }
+ else
+ return 0;
+
+ case N_Indexed_Component:
+ /* Only the array expression can require an lvalue. */
+ if (Prefix (gnat_parent) != gnat_node)
+ return 0;
+
+ /* ??? Consider that referencing an indexed component with a variable
+ index forces the whole aggregate to memory. Note that testing only
+ for literals is conservative, any static expression in the RM sense
+ could probably be accepted with some additional work. */
+ for (gnat_temp = First (Expressions (gnat_parent));
+ Present (gnat_temp);
+ gnat_temp = Next (gnat_temp))
+ if (Nkind (gnat_temp) != N_Character_Literal
+ && Nkind (gnat_temp) != N_Integer_Literal
+ && !(Is_Entity_Name (gnat_temp)
+ && Ekind (Entity (gnat_temp)) == E_Enumeration_Literal))
+ return 1;
+
+ /* ... fall through ... */
+
+ case N_Selected_Component:
+ case N_Slice:
+ /* Only the prefix expression can require an lvalue. */
+ if (Prefix (gnat_parent) != gnat_node)
+ return 0;
+
+ return lvalue_required_p (gnat_parent,
+ get_unpadded_type (Etype (gnat_parent)),
+ constant, address_of_constant);
+
+ case N_Object_Renaming_Declaration:
+ /* We need to preserve addresses through a renaming. */
+ return 1;
+
+ case N_Object_Declaration:
+ /* We cannot use a constructor if this is an atomic object because
+ the actual assignment might end up being done component-wise. */
+ return (!constant
+ ||(Is_Composite_Type (Underlying_Type (Etype (gnat_node)))
+ && Is_Full_Access (Defining_Entity (gnat_parent)))
+ /* We don't use a constructor if this is a class-wide object
+ because the effective type of the object is the equivalent
+ type of the class-wide subtype and it smashes most of the
+ data into an array of bytes to which we cannot convert. */
+ || Ekind ((Etype (Defining_Entity (gnat_parent))))
+ == E_Class_Wide_Subtype);
+
+ case N_Assignment_Statement:
+ /* We cannot use a constructor if the LHS is an atomic object because
+ the actual assignment might end up being done component-wise. */
+ return (!constant
+ || Name (gnat_parent) == gnat_node
+ || (Is_Composite_Type (Underlying_Type (Etype (gnat_node)))
+ && Is_Entity_Name (Name (gnat_parent))
+ && Is_Full_Access (Entity (Name (gnat_parent)))));
+
+ case N_Unchecked_Type_Conversion:
+ if (!constant)
+ return 1;
+
+ /* ... fall through ... */
+
+ case N_Type_Conversion:
+ case N_Qualified_Expression:
+ /* We must look through all conversions because we may need to bypass
+ an intermediate conversion that is meant to be purely formal. */
+ return lvalue_required_p (gnat_parent,
+ get_unpadded_type (Etype (gnat_parent)),
+ constant, address_of_constant);
+
+ case N_Explicit_Dereference:
+ /* We look through dereferences for address of constant because we need
+ to handle the special cases listed above. */
+ if (constant && address_of_constant)
+ return lvalue_required_p (gnat_parent,
+ get_unpadded_type (Etype (gnat_parent)),
+ true, false);
+
+ /* ... fall through ... */
+
+ default:
+ return 0;
+ }
+
+ gcc_unreachable ();
+}
+
+/* Return true if an lvalue should be used for GNAT_NODE. GNU_TYPE is the type
+ that will be used for GNAT_NODE in the translated GNU tree and is assumed to
+ be an aggregate type.
+
+ The function climbs up the GNAT tree starting from the node and returns true
+ upon encountering a node that makes it doable to decide. lvalue_required_p
+ should have been previously invoked on the arguments and returned false. */
+
+static bool
+lvalue_for_aggregate_p (Node_Id gnat_node, tree gnu_type)
+{
+ Node_Id gnat_parent = Parent (gnat_node);
+
+ switch (Nkind (gnat_parent))
+ {
+ case N_Parameter_Association:
+ case N_Function_Call:
+ case N_Procedure_Call_Statement:
+ /* Even if the parameter is by copy, prefer an lvalue. */
+ return true;
+
+ case N_Simple_Return_Statement:
+ /* Likewise for a return value. */
+ return true;
+
+ case N_Indexed_Component:
+ case N_Selected_Component:
+ /* If an elementary component is used, take it from the constant. */
+ if (!Is_Composite_Type (Underlying_Type (Etype (gnat_parent))))
+ return false;
+
+ /* ... fall through ... */
+
+ case N_Slice:
+ return lvalue_for_aggregate_p (gnat_parent,
+ get_unpadded_type (Etype (gnat_parent)));
+
+ case N_Object_Declaration:
+ /* For an aggregate object declaration, return false consistently. */
+ return false;
+
+ case N_Assignment_Statement:
+ /* For an aggregate assignment, decide based on the size. */
+ {
+ const HOST_WIDE_INT size = int_size_in_bytes (gnu_type);
+ return size < 0 || size >= param_large_stack_frame / 4;
+ }
+
+ case N_Unchecked_Type_Conversion:
+ case N_Type_Conversion:
+ case N_Qualified_Expression:
+ return lvalue_for_aggregate_p (gnat_parent,
+ get_unpadded_type (Etype (gnat_parent)));
+
+ case N_Allocator:
+ /* We should only reach here through the N_Qualified_Expression case.
+ Force an lvalue for aggregate types since a block-copy to the newly
+ allocated area of memory is made. */
+ return true;
+
+ default:
+ return false;
+ }
+
+ gcc_unreachable ();
+}
+
+
+/* Return true if T is a constant DECL node that can be safely replaced
+ by its initializer. */
+
+static bool
+constant_decl_with_initializer_p (tree t)
+{
+ if (!TREE_CONSTANT (t) || !DECL_P (t) || !DECL_INITIAL (t))
+ return false;
+
+ /* Return false for aggregate types that contain a placeholder since
+ their initializers cannot be manipulated easily. */
+ if (AGGREGATE_TYPE_P (TREE_TYPE (t))
+ && !TYPE_IS_FAT_POINTER_P (TREE_TYPE (t))
+ && type_contains_placeholder_p (TREE_TYPE (t)))
+ return false;
+
+ return true;
+}
+
+/* Return an expression equivalent to EXP but where constant DECL nodes
+ have been replaced by their initializer. */
+
+static tree
+fold_constant_decl_in_expr (tree exp)
+{
+ enum tree_code code = TREE_CODE (exp);
+ tree op0;
+
+ switch (code)
+ {
+ case CONST_DECL:
+ case VAR_DECL:
+ if (!constant_decl_with_initializer_p (exp))
+ return exp;
+
+ return DECL_INITIAL (exp);
+
+ case COMPONENT_REF:
+ op0 = fold_constant_decl_in_expr (TREE_OPERAND (exp, 0));
+ if (op0 == TREE_OPERAND (exp, 0))
+ return exp;
+
+ return fold_build3 (COMPONENT_REF, TREE_TYPE (exp), op0,
+ TREE_OPERAND (exp, 1), NULL_TREE);
+
+ case BIT_FIELD_REF:
+ op0 = fold_constant_decl_in_expr (TREE_OPERAND (exp, 0));
+ if (op0 == TREE_OPERAND (exp, 0))
+ return exp;
+
+ return fold_build3 (BIT_FIELD_REF, TREE_TYPE (exp), op0,
+ TREE_OPERAND (exp, 1), TREE_OPERAND (exp, 2));
+
+ case ARRAY_REF:
+ case ARRAY_RANGE_REF:
+ /* If the index is not itself constant, then nothing can be folded. */
+ if (!TREE_CONSTANT (TREE_OPERAND (exp, 1)))
+ return exp;
+ op0 = fold_constant_decl_in_expr (TREE_OPERAND (exp, 0));
+ if (op0 == TREE_OPERAND (exp, 0))
+ return exp;
+
+ return fold (build4 (code, TREE_TYPE (exp), op0, TREE_OPERAND (exp, 1),
+ TREE_OPERAND (exp, 2), NULL_TREE));
+
+ case REALPART_EXPR:
+ case IMAGPART_EXPR:
+ case VIEW_CONVERT_EXPR:
+ op0 = fold_constant_decl_in_expr (TREE_OPERAND (exp, 0));
+ if (op0 == TREE_OPERAND (exp, 0))
+ return exp;
+
+ return fold_build1 (code, TREE_TYPE (exp), op0);
+
+ default:
+ return exp;
+ }
+
+ gcc_unreachable ();
+}
+
+/* Return true if TYPE and DEF_TYPE are compatible GNAT types for Gigi. */
+
+static bool
+Gigi_Types_Compatible (Entity_Id type, Entity_Id def_type)
+{
+ /* The trivial case. */
+ if (type == def_type)
+ return true;
+
+ /* A class-wide type is equivalent to a subtype of itself. */
+ if (Is_Class_Wide_Type (type))
+ return true;
+
+ /* A packed array type is compatible with its implementation type. */
+ if (Is_Packed (def_type) && type == Packed_Array_Impl_Type (def_type))
+ return true;
+
+ /* If both types are Itypes, one may be a copy of the other. */
+ if (Is_Itype (def_type) && Is_Itype (type))
+ return true;
+
+ /* If the type is incomplete and comes from a limited context, then also
+ consider its non-limited view. */
+ if (Is_Incomplete_Type (def_type)
+ && From_Limited_With (def_type)
+ && Present (Non_Limited_View (def_type)))
+ return Gigi_Types_Compatible (type, Non_Limited_View (def_type));
+
+ /* If the type is incomplete/private, then also consider its full view. */
+ if (Is_Incomplete_Or_Private_Type (def_type)
+ && Present (Full_View (def_type)))
+ return Gigi_Types_Compatible (type, Full_View (def_type));
+
+ return false;
+}
+
+/* Subroutine of gnat_to_gnu to translate gnat_node, an N_Identifier,
+ to a GCC tree, which is returned. GNU_RESULT_TYPE_P is a pointer
+ to where we should place the result type. */
+
+static tree
+Identifier_to_gnu (Node_Id gnat_node, tree *gnu_result_type_p)
+{
+ /* The entity of GNAT_NODE and its type. */
+ Node_Id gnat_entity = (Nkind (gnat_node) == N_Defining_Identifier
+ || Nkind (gnat_node) == N_Defining_Operator_Symbol)
+ ? gnat_node : Entity (gnat_node);
+ Node_Id gnat_entity_type = Etype (gnat_entity);
+ /* If GNAT_NODE is a constant, whether we should use the initialization
+ value instead of the constant entity, typically for scalars with an
+ address clause when the parent doesn't require an lvalue. */
+ bool use_constant_initializer = false;
+ /* Whether we should require an lvalue for GNAT_NODE. Needed in
+ specific circumstances only, so evaluated lazily. < 0 means
+ unknown, > 0 means known true, 0 means known false. */
+ int require_lvalue = -1;
+ Entity_Id gnat_result_type;
+ tree gnu_result, gnu_result_type;
+
+ /* If the Etype of this node is not the same as that of the Entity, then
+ something went wrong, probably in generic instantiation. However, this
+ does not apply to types. Since we sometime have strange Ekind's, just
+ do this test for objects, except for discriminants because their type
+ may have been changed to a subtype by Exp_Ch3.Adjust_Discriminants. */
+ gcc_assert (!Is_Object (gnat_entity)
+ || Ekind (gnat_entity) == E_Discriminant
+ || Etype (gnat_node) == gnat_entity_type
+ || Gigi_Types_Compatible (Etype (gnat_node), gnat_entity_type));
+
+ /* If this is a reference to a deferred constant whose partial view is an
+ unconstrained private type, the proper type is on the full view of the
+ constant, not on the full view of the type, which may be unconstrained.
+
+ This may be a reference to a type, for example in the prefix of the
+ attribute Position, generated for dispatching code (see Make_DT in
+ exp_disp,adb). In that case we need the type itself, not is parent,
+ in particular if it is a derived type */
+ if (Ekind (gnat_entity) == E_Constant
+ && Is_Private_Type (gnat_entity_type)
+ && (Has_Unknown_Discriminants (gnat_entity_type)
+ || (Present (Full_View (gnat_entity_type))
+ && Has_Discriminants (Full_View (gnat_entity_type))))
+ && Present (Full_View (gnat_entity)))
+ {
+ gnat_entity = Full_View (gnat_entity);
+ gnat_result_type = Etype (gnat_entity);
+ }
+ else
+ {
+ /* We use the Actual_Subtype only if it has already been elaborated,
+ as we may be invoked precisely during its elaboration, otherwise
+ the Etype. Avoid using it for packed arrays to simplify things,
+ except in a return statement because we need the actual size and
+ the front-end does not make it explicit in this case. */
+ if ((Ekind (gnat_entity) == E_Constant
+ || Ekind (gnat_entity) == E_Variable
+ || Is_Formal (gnat_entity))
+ && !(Is_Array_Type (Etype (gnat_entity))
+ && Present (Packed_Array_Impl_Type (Etype (gnat_entity)))
+ && Nkind (Parent (gnat_node)) != N_Simple_Return_Statement)
+ && Present (Actual_Subtype (gnat_entity))
+ && present_gnu_tree (Actual_Subtype (gnat_entity)))
+ gnat_result_type = Actual_Subtype (gnat_entity);
+ else
+ gnat_result_type = Etype (gnat_node);
+ }
+
+ /* Expand the type of this identifier first, in case it is an enumeral
+ literal, which only get made when the type is expanded. There is no
+ order-of-elaboration issue here. */
+ gnu_result_type = get_unpadded_type (gnat_result_type);
+
+ /* If this is a non-imported elementary constant with an address clause,
+ retrieve the value instead of a pointer to be dereferenced unless
+ an lvalue is required. This is generally more efficient and actually
+ required if this is a static expression because it might be used
+ in a context where a dereference is inappropriate, such as a case
+ statement alternative or a record discriminant. There is no possible
+ volatile-ness short-circuit here since Volatile constants must be
+ imported per C.6. */
+ if (Ekind (gnat_entity) == E_Constant
+ && Is_Elementary_Type (gnat_result_type)
+ && !Is_Imported (gnat_entity)
+ && Present (Address_Clause (gnat_entity)))
+ {
+ require_lvalue
+ = lvalue_required_p (gnat_node, gnu_result_type, true, false);
+ use_constant_initializer = !require_lvalue;
+ }
+
+ if (use_constant_initializer)
+ {
+ /* If this is a deferred constant, the initializer is attached to
+ the full view. */
+ if (Present (Full_View (gnat_entity)))
+ gnat_entity = Full_View (gnat_entity);
+
+ gnu_result = gnat_to_gnu (Expression (Declaration_Node (gnat_entity)));
+ }
+ else
+ gnu_result = gnat_to_gnu_entity (gnat_entity, NULL_TREE, false);
+
+ /* Some objects (such as parameters passed by reference, globals of
+ variable size, and renamed objects) actually represent the address
+ of the object. In that case, we must do the dereference. Likewise,
+ deal with parameters to foreign convention subprograms. */
+ if (DECL_P (gnu_result)
+ && (DECL_BY_REF_P (gnu_result)
+ || (TREE_CODE (gnu_result) == PARM_DECL
+ && DECL_BY_COMPONENT_PTR_P (gnu_result))))
+ {
+ const bool read_only = DECL_POINTS_TO_READONLY_P (gnu_result);
+
+ /* If it's a PARM_DECL to foreign convention subprogram, convert it. */
+ if (TREE_CODE (gnu_result) == PARM_DECL
+ && DECL_BY_COMPONENT_PTR_P (gnu_result))
+ gnu_result
+ = convert (build_pointer_type (gnu_result_type), gnu_result);
+
+ /* If it's a CONST_DECL, return the underlying constant like below. */
+ else if (TREE_CODE (gnu_result) == CONST_DECL
+ && !(DECL_CONST_ADDRESS_P (gnu_result)
+ && lvalue_required_p (gnat_node, gnu_result_type, true,
+ true)))
+ gnu_result = DECL_INITIAL (gnu_result);
+
+ /* Do the final dereference. */
+ gnu_result = build_unary_op (INDIRECT_REF, NULL_TREE, gnu_result);
+
+ if ((TREE_CODE (gnu_result) == INDIRECT_REF
+ || TREE_CODE (gnu_result) == UNCONSTRAINED_ARRAY_REF)
+ && No (Address_Clause (gnat_entity)))
+ TREE_THIS_NOTRAP (gnu_result) = 1;
+
+ if (read_only)
+ TREE_READONLY (gnu_result) = 1;
+ }
+
+ /* If we have a constant declaration and its initializer, try to return the
+ latter to avoid the need to call fold in lots of places and the need for
+ elaboration code if this identifier is used as an initializer itself. */
+ if (constant_decl_with_initializer_p (gnu_result))
+ {
+ bool constant_only = (TREE_CODE (gnu_result) == CONST_DECL
+ && !DECL_CONST_CORRESPONDING_VAR (gnu_result));
+ bool address_of_constant = (TREE_CODE (gnu_result) == CONST_DECL
+ && DECL_CONST_ADDRESS_P (gnu_result));
+
+ /* If there is a (corresponding) variable or this is the address of a
+ constant, we only want to return the initializer if an lvalue isn't
+ required. Evaluate this now if we have not already done so. */
+ if ((!constant_only || address_of_constant) && require_lvalue < 0)
+ require_lvalue
+ = lvalue_required_p (gnat_node, gnu_result_type, true,
+ address_of_constant)
+ || (AGGREGATE_TYPE_P (gnu_result_type)
+ && lvalue_for_aggregate_p (gnat_node, gnu_result_type));
+
+ /* Finally retrieve the initializer if this is deemed valid. */
+ if ((constant_only && !address_of_constant) || !require_lvalue)
+ gnu_result = DECL_INITIAL (gnu_result);
+ }
+
+ /* But for a constant renaming we couldn't do that incrementally for its
+ definition because of the need to return an lvalue so, if the present
+ context doesn't itself require an lvalue, we try again here. */
+ else if (Ekind (gnat_entity) == E_Constant
+ && Is_Elementary_Type (gnat_result_type)
+ && Present (Renamed_Object (gnat_entity)))
+ {
+ if (require_lvalue < 0)
+ require_lvalue
+ = lvalue_required_p (gnat_node, gnu_result_type, true, false);
+ if (!require_lvalue)
+ gnu_result = fold_constant_decl_in_expr (gnu_result);
+ }
+
+ /* The GNAT tree has the type of a function set to its result type, so we
+ adjust here. Also use the type of the result if the Etype is a subtype
+ that is nominally unconstrained. Likewise if this is a deferred constant
+ of a discriminated type whose full view can be elaborated statically, to
+ avoid problematic conversions to the nominal subtype. But remove any
+ padding from the resulting type. */
+ if (FUNC_OR_METHOD_TYPE_P (TREE_TYPE (gnu_result))
+ || Is_Constr_Subt_For_UN_Aliased (gnat_result_type)
+ || (Ekind (gnat_entity) == E_Constant
+ && Present (Full_View (gnat_entity))
+ && Has_Discriminants (gnat_result_type)
+ && TREE_CODE (gnu_result) == CONSTRUCTOR))
+ {
+ gnu_result_type = TREE_TYPE (gnu_result);
+ if (TYPE_IS_PADDING_P (gnu_result_type))
+ gnu_result_type = TREE_TYPE (TYPE_FIELDS (gnu_result_type));
+ }
+
+ *gnu_result_type_p = gnu_result_type;
+
+ return gnu_result;
+}
+
+/* Subroutine of gnat_to_gnu to process gnat_node, an N_Pragma. Return
+ any statements we generate. */
+
+static tree
+Pragma_to_gnu (Node_Id gnat_node)
+{
+ tree gnu_result = alloc_stmt_list ();
+ Node_Id gnat_temp;
+
+ /* Check for (and ignore) unrecognized pragmas. */
+ if (!Is_Pragma_Name (Chars (Pragma_Identifier (gnat_node))))
+ return gnu_result;
+
+ const unsigned char id
+ = Get_Pragma_Id (Chars (Pragma_Identifier (gnat_node)));
+
+ /* Save the expression of pragma Compile_Time_{Error|Warning} for later. */
+ if (id == Pragma_Compile_Time_Error || id == Pragma_Compile_Time_Warning)
+ {
+ gnat_temp = First (Pragma_Argument_Associations (gnat_node));
+ gnat_compile_time_expr_list.safe_push (Expression (gnat_temp));
+ return gnu_result;
+ }
+
+ /* Stop there if we are just annotating types. */
+ if (type_annotate_only)
+ return gnu_result;
+
+ switch (id)
+ {
+ case Pragma_Inspection_Point:
+ /* Do nothing at top level: all such variables are already viewable. */
+ if (global_bindings_p ())
+ break;
+
+ for (gnat_temp = First (Pragma_Argument_Associations (gnat_node));
+ Present (gnat_temp);
+ gnat_temp = Next (gnat_temp))
+ {
+ Node_Id gnat_expr = Expression (gnat_temp);
+ tree gnu_expr = gnat_to_gnu (gnat_expr);
+ tree asm_constraint = NULL_TREE;
+#ifdef ASM_COMMENT_START
+ char *comment;
+#endif
+ gnu_expr = maybe_unconstrained_array (gnu_expr);
+ if (TREE_CODE (gnu_expr) == CONST_DECL
+ && DECL_CONST_CORRESPONDING_VAR (gnu_expr))
+ gnu_expr = DECL_CONST_CORRESPONDING_VAR (gnu_expr);
+ gnat_mark_addressable (gnu_expr);
+
+#ifdef ASM_COMMENT_START
+ comment = concat (ASM_COMMENT_START,
+ " inspection point: ",
+ Get_Name_String (Chars (gnat_expr)),
+ " is at %0",
+ NULL);
+ asm_constraint = build_string (strlen (comment), comment);
+ free (comment);
+#endif
+ gnu_expr = build5 (ASM_EXPR, void_type_node,
+ asm_constraint,
+ NULL_TREE,
+ tree_cons
+ (build_tree_list (NULL_TREE,
+ build_string (1, "m")),
+ gnu_expr, NULL_TREE),
+ NULL_TREE, NULL_TREE);
+ ASM_VOLATILE_P (gnu_expr) = 1;
+ set_expr_location_from_node (gnu_expr, gnat_node);
+ append_to_statement_list (gnu_expr, &gnu_result);
+ }
+ break;
+
+ case Pragma_Loop_Optimize:
+ for (gnat_temp = First (Pragma_Argument_Associations (gnat_node));
+ Present (gnat_temp);
+ gnat_temp = Next (gnat_temp))
+ {
+ tree gnu_loop_stmt = gnu_loop_stack->last ()->stmt;
+
+ switch (Chars (Expression (gnat_temp)))
+ {
+ case Name_Ivdep:
+ LOOP_STMT_IVDEP (gnu_loop_stmt) = 1;
+ break;
+
+ case Name_No_Unroll:
+ LOOP_STMT_NO_UNROLL (gnu_loop_stmt) = 1;
+ break;
+
+ case Name_Unroll:
+ LOOP_STMT_UNROLL (gnu_loop_stmt) = 1;
+ break;
+
+ case Name_No_Vector:
+ LOOP_STMT_NO_VECTOR (gnu_loop_stmt) = 1;
+ break;
+
+ case Name_Vector:
+ LOOP_STMT_VECTOR (gnu_loop_stmt) = 1;
+ break;
+
+ default:
+ gcc_unreachable ();
+ }
+ }
+ break;
+
+ case Pragma_Optimize:
+ switch (Chars (Expression
+ (First (Pragma_Argument_Associations (gnat_node)))))
+ {
+ case Name_Off:
+ if (optimize)
+ post_error ("must specify -O0??", gnat_node);
+ break;
+
+ case Name_Space:
+ if (!optimize_size)
+ post_error ("must specify -Os??", gnat_node);
+ break;
+
+ case Name_Time:
+ if (!optimize)
+ post_error ("insufficient -O value??", gnat_node);
+ break;
+
+ default:
+ gcc_unreachable ();
+ }
+ break;
+
+ case Pragma_Reviewable:
+ if (write_symbols == NO_DEBUG)
+ post_error ("must specify -g??", gnat_node);
+ break;
+
+ case Pragma_Warning_As_Error:
+ case Pragma_Warnings:
+ {
+ Node_Id gnat_expr;
+ /* Preserve the location of the pragma. */
+ const location_t location = input_location;
+ struct cl_option_handlers handlers;
+ unsigned int option_index;
+ diagnostic_t kind;
+ bool imply;
+
+ gnat_temp = First (Pragma_Argument_Associations (gnat_node));
+
+ /* This is the String form: pragma Warning{s|_As_Error}(String). */
+ if (Nkind (Expression (gnat_temp)) == N_String_Literal)
+ {
+ switch (id)
+ {
+ case Pragma_Warning_As_Error:
+ kind = DK_ERROR;
+ imply = false;
+ break;
+
+ case Pragma_Warnings:
+ kind = DK_WARNING;
+ imply = true;
+ break;
+
+ default:
+ gcc_unreachable ();
+ }
+
+ gnat_expr = Expression (gnat_temp);
+ }
+
+ /* This is the On/Off form: pragma Warnings (On | Off [,String]). */
+ else if (Nkind (Expression (gnat_temp)) == N_Identifier)
+ {
+ switch (Chars (Expression (gnat_temp)))
+ {
+ case Name_Off:
+ kind = DK_IGNORED;
+ break;
+
+ case Name_On:
+ kind = DK_WARNING;
+ break;
+
+ default:
+ gcc_unreachable ();
+ }
+
+ /* Deal with optional pattern (but ignore Reason => "..."). */
+ if (Present (Next (gnat_temp))
+ && Chars (Next (gnat_temp)) != Name_Reason)
+ {
+ /* pragma Warnings (On | Off, Name) is handled differently. */
+ if (Nkind (Expression (Next (gnat_temp))) != N_String_Literal)
+ break;
+
+ gnat_expr = Expression (Next (gnat_temp));
+ }
+ else
+ {
+ gnat_expr = Empty;
+
+ /* For pragma Warnings (Off), we save the current state... */
+ if (kind == DK_IGNORED)
+ diagnostic_push_diagnostics (global_dc, location);
+
+ /* ...so that, for pragma Warnings (On), we do not enable all
+ the warnings but just restore the previous state. */
+ else
+ {
+ diagnostic_pop_diagnostics (global_dc, location);
+ break;
+ }
+ }
+
+ imply = false;
+ }
+
+ else
+ gcc_unreachable ();
+
+ /* This is the same implementation as in the C family of compilers. */
+ const unsigned int lang_mask = CL_Ada | CL_COMMON;
+ const char *arg = NULL;
+ if (Present (gnat_expr))
+ {
+ tree gnu_expr = gnat_to_gnu (gnat_expr);
+ const char *option_string = TREE_STRING_POINTER (gnu_expr);
+ const int len = TREE_STRING_LENGTH (gnu_expr);
+ if (len < 3 || option_string[0] != '-' || option_string[1] != 'W')
+ break;
+ option_index = find_opt (option_string + 1, lang_mask);
+ if (option_index == OPT_SPECIAL_unknown)
+ {
+ post_error ("unknown -W switch??", gnat_node);
+ break;
+ }
+ else if (!(cl_options[option_index].flags & CL_WARNING))
+ {
+ post_error ("-W switch does not control warning??", gnat_node);
+ break;
+ }
+ else if (!(cl_options[option_index].flags & lang_mask))
+ {
+ post_error ("-W switch not valid for Ada??", gnat_node);
+ break;
+ }
+ if (cl_options[option_index].flags & CL_JOINED)
+ arg = option_string + 1 + cl_options[option_index].opt_len;
+ }
+ else
+ option_index = 0;
+
+ set_default_handlers (&handlers, NULL);
+ control_warning_option (option_index, (int) kind, arg, imply, location,
+ lang_mask, &handlers, &global_options,
+ &global_options_set, global_dc);
+ }
+ break;
+
+ default:
+ break;
+ }
+
+ return gnu_result;
+}
+
+/* Check the inline status of nested function FNDECL wrt its parent function.
+
+ If a non-inline nested function is referenced from an inline external
+ function, we cannot honor both requests at the same time without cloning
+ the nested function in the current unit since it is private to its unit.
+ We could inline it as well but it's probably better to err on the side
+ of too little inlining.
+
+ This must be done only on nested functions present in the source code
+ and not on nested functions generated by the compiler, e.g. finalizers,
+ because they may be not marked inline and we don't want them to block
+ the inlining of the parent function. */
+
+static void
+check_inlining_for_nested_subprog (tree fndecl)
+{
+ if (DECL_IGNORED_P (current_function_decl) || DECL_IGNORED_P (fndecl))
+ return;
+
+ if (DECL_DECLARED_INLINE_P (fndecl))
+ return;
+
+ tree parent_decl = decl_function_context (fndecl);
+ if (DECL_EXTERNAL (parent_decl) && DECL_DECLARED_INLINE_P (parent_decl))
+ {
+ const location_t loc1 = DECL_SOURCE_LOCATION (fndecl);
+ const location_t loc2 = DECL_SOURCE_LOCATION (parent_decl);
+
+ if (lookup_attribute ("always_inline", DECL_ATTRIBUTES (parent_decl)))
+ {
+ error_at (loc1, "subprogram %q+F not marked %<Inline_Always%>",
+ fndecl);
+ error_at (loc2, "parent subprogram cannot be inlined");
+ }
+ else
+ {
+ warning_at (loc1, OPT_Winline, "subprogram %q+F not marked %<Inline%>",
+ fndecl);
+ warning_at (loc2, OPT_Winline, "parent subprogram cannot be inlined");
+ }
+
+ DECL_DECLARED_INLINE_P (parent_decl) = 0;
+ DECL_UNINLINABLE (parent_decl) = 1;
+ }
+}
+
+/* Return an expression for the length of TYPE, an integral type, computed in
+ RESULT_TYPE, another integral type.
+
+ We used to compute the length as MAX (hb - lb + 1, 0) which could overflow
+ when lb == TYPE'First. We now compute it as (hb >= lb) ? hb - lb + 1 : 0
+ which would only overflow in much rarer cases, for extremely large arrays
+ we expect never to encounter in practice. Besides, the former computation
+ required the use of potentially constraining signed arithmetics while the
+ latter does not. Note that the comparison must be done in the original
+ base index type in order to avoid any overflow during the conversion. */
+
+static tree
+get_type_length (tree type, tree result_type)
+{
+ tree comp_type = get_base_type (result_type);
+ tree base_type = maybe_character_type (get_base_type (type));
+ tree lb = convert (base_type, TYPE_MIN_VALUE (type));
+ tree hb = convert (base_type, TYPE_MAX_VALUE (type));
+ tree length
+ = build_binary_op (PLUS_EXPR, comp_type,
+ build_binary_op (MINUS_EXPR, comp_type,
+ convert (comp_type, hb),
+ convert (comp_type, lb)),
+ build_int_cst (comp_type, 1));
+ length
+ = build_cond_expr (result_type,
+ build_binary_op (GE_EXPR, boolean_type_node, hb, lb),
+ convert (result_type, length),
+ build_int_cst (result_type, 0));
+ return length;
+}
+
+/* Subroutine of gnat_to_gnu to translate GNAT_NODE, an N_Attribute node,
+ to a GCC tree, which is returned. GNU_RESULT_TYPE_P is a pointer to
+ where we should place the result type. ATTRIBUTE is the attribute ID. */
+
+static tree
+Attribute_to_gnu (Node_Id gnat_node, tree *gnu_result_type_p, int attribute)
+{
+ const Node_Id gnat_prefix = Prefix (gnat_node);
+ tree gnu_prefix = gnat_to_gnu (gnat_prefix);
+ tree gnu_type = TREE_TYPE (gnu_prefix);
+ tree gnu_expr, gnu_result_type, gnu_result = error_mark_node;
+ bool prefix_unused = false;
+
+ /* If the input is a NULL_EXPR, make a new one. */
+ if (TREE_CODE (gnu_prefix) == NULL_EXPR)
+ {
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ *gnu_result_type_p = gnu_result_type;
+ return build1 (NULL_EXPR, gnu_result_type, TREE_OPERAND (gnu_prefix, 0));
+ }
+
+ switch (attribute)
+ {
+ case Attr_Pred:
+ case Attr_Succ:
+ /* These just add or subtract the constant 1 since representation
+ clauses for enumeration types are handled in the front-end. */
+ gnu_expr = gnat_to_gnu (First (Expressions (gnat_node)));
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ gnu_type = maybe_character_type (gnu_result_type);
+ if (TREE_TYPE (gnu_expr) != gnu_type)
+ gnu_expr = convert (gnu_type, gnu_expr);
+ gnu_result
+ = build_binary_op (attribute == Attr_Pred ? MINUS_EXPR : PLUS_EXPR,
+ gnu_type, gnu_expr, build_int_cst (gnu_type, 1));
+ break;
+
+ case Attr_Address:
+ case Attr_Unrestricted_Access:
+ /* Conversions don't change the address of references but can cause
+ build_unary_op to miss the references below, so strip them off.
+ On the contrary, if the address-of operation causes a temporary
+ to be created, then it must be created with the proper type. */
+ gnu_expr = remove_conversions (gnu_prefix,
+ !Must_Be_Byte_Aligned (gnat_node));
+ if (REFERENCE_CLASS_P (gnu_expr))
+ gnu_prefix = gnu_expr;
+
+ /* If we are taking 'Address of an unconstrained object, this is the
+ pointer to the underlying array. */
+ if (attribute == Attr_Address)
+ gnu_prefix = maybe_unconstrained_array (gnu_prefix);
+
+ /* If we are building a static dispatch table, we have to honor
+ TARGET_VTABLE_USES_DESCRIPTORS if we want to be compatible
+ with the C++ ABI. We do it in the non-static case as well,
+ see gnat_to_gnu_entity, case E_Access_Subprogram_Type. */
+ else if (TARGET_VTABLE_USES_DESCRIPTORS
+ && Is_Dispatch_Table_Entity (Etype (gnat_node)))
+ {
+ tree gnu_field, t;
+ /* Descriptors can only be built here for top-level functions. */
+ bool build_descriptor = (global_bindings_p () != 0);
+ int i;
+ vec<constructor_elt, va_gc> *gnu_vec = NULL;
+ constructor_elt *elt;
+
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+
+ /* If we're not going to build the descriptor, we have to retrieve
+ the one which will be built by the linker (or by the compiler
+ later if a static chain is requested). */
+ if (!build_descriptor)
+ {
+ gnu_result = build_unary_op (ADDR_EXPR, NULL_TREE, gnu_prefix);
+ gnu_result = fold_convert (build_pointer_type (gnu_result_type),
+ gnu_result);
+ gnu_result = build1 (INDIRECT_REF, gnu_result_type, gnu_result);
+ }
+
+ vec_safe_grow (gnu_vec, TARGET_VTABLE_USES_DESCRIPTORS, true);
+ elt = (gnu_vec->address () + TARGET_VTABLE_USES_DESCRIPTORS - 1);
+ for (gnu_field = TYPE_FIELDS (gnu_result_type), i = 0;
+ i < TARGET_VTABLE_USES_DESCRIPTORS;
+ gnu_field = DECL_CHAIN (gnu_field), i++)
+ {
+ if (build_descriptor)
+ {
+ t = build2 (FDESC_EXPR, TREE_TYPE (gnu_field), gnu_prefix,
+ build_int_cst (NULL_TREE, i));
+ TREE_CONSTANT (t) = 1;
+ }
+ else
+ t = build3 (COMPONENT_REF, ptr_void_ftype, gnu_result,
+ gnu_field, NULL_TREE);
+
+ elt->index = gnu_field;
+ elt->value = t;
+ elt--;
+ }
+
+ gnu_result = gnat_build_constructor (gnu_result_type, gnu_vec);
+ break;
+ }
+
+ /* ... fall through ... */
+
+ case Attr_Access:
+ case Attr_Unchecked_Access:
+ case Attr_Code_Address:
+ /* Taking the address of a type does not make sense. */
+ gcc_assert (TREE_CODE (gnu_prefix) != TYPE_DECL);
+
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ gnu_result
+ = build_unary_op (((attribute == Attr_Address
+ || attribute == Attr_Unrestricted_Access)
+ && !Must_Be_Byte_Aligned (gnat_node))
+ ? ATTR_ADDR_EXPR : ADDR_EXPR,
+ gnu_result_type, gnu_prefix);
+
+ /* For 'Code_Address, find an inner ADDR_EXPR and mark it so that we
+ don't try to build a trampoline. */
+ if (attribute == Attr_Code_Address)
+ {
+ gnu_expr = remove_conversions (gnu_result, false);
+
+ if (TREE_CODE (gnu_expr) == ADDR_EXPR)
+ TREE_NO_TRAMPOLINE (gnu_expr) = TREE_CONSTANT (gnu_expr) = 1;
+
+ /* On targets for which function symbols denote a descriptor, the
+ code address is stored within the first slot of the descriptor
+ so we do an additional dereference:
+ result = *((result_type *) result)
+ where we expect result to be of some pointer type already. */
+ if (targetm.calls.custom_function_descriptors == 0)
+ gnu_result
+ = build_unary_op (INDIRECT_REF, NULL_TREE,
+ convert (build_pointer_type (gnu_result_type),
+ gnu_result));
+ }
+
+ /* For 'Access, issue an error message if the prefix is a C++ method
+ since it can use a special calling convention on some platforms,
+ which cannot be propagated to the access type. */
+ else if (attribute == Attr_Access
+ && TREE_CODE (TREE_TYPE (gnu_prefix)) == METHOD_TYPE)
+ post_error ("access to C++ constructor or member function not allowed",
+ gnat_node);
+
+ /* For other address attributes applied to a nested function,
+ find an inner ADDR_EXPR and annotate it so that we can issue
+ a useful warning with -Wtrampolines. */
+ else if (FUNC_OR_METHOD_TYPE_P (TREE_TYPE (gnu_prefix))
+ && (gnu_expr = remove_conversions (gnu_result, false))
+ && TREE_CODE (gnu_expr) == ADDR_EXPR
+ && decl_function_context (TREE_OPERAND (gnu_expr, 0)))
+ {
+ set_expr_location_from_node (gnu_expr, gnat_node);
+
+ /* Also check the inlining status. */
+ check_inlining_for_nested_subprog (TREE_OPERAND (gnu_expr, 0));
+
+ /* Moreover, for 'Access or 'Unrestricted_Access with non-
+ foreign-compatible representation, mark the ADDR_EXPR so
+ that we can build a descriptor instead of a trampoline. */
+ if ((attribute == Attr_Access
+ || attribute == Attr_Unrestricted_Access)
+ && targetm.calls.custom_function_descriptors > 0
+ && Can_Use_Internal_Rep (Underlying_Type (Etype (gnat_node))))
+ FUNC_ADDR_BY_DESCRIPTOR (gnu_expr) = 1;
+
+ /* Otherwise, we need to check that we are not violating the
+ No_Implicit_Dynamic_Code restriction. */
+ else if (targetm.calls.custom_function_descriptors != 0)
+ Check_Implicit_Dynamic_Code_Allowed (gnat_node);
+ }
+ break;
+
+ case Attr_Pool_Address:
+ {
+ tree gnu_ptr = gnu_prefix;
+ tree gnu_obj_type;
+
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+
+ /* If this is fat pointer, the object must have been allocated with the
+ template in front of the array. So compute the template address; do
+ it by converting to a thin pointer. */
+ if (TYPE_IS_FAT_POINTER_P (TREE_TYPE (gnu_ptr)))
+ gnu_ptr
+ = convert (build_pointer_type
+ (TYPE_OBJECT_RECORD_TYPE
+ (TYPE_UNCONSTRAINED_ARRAY (TREE_TYPE (gnu_ptr)))),
+ gnu_ptr);
+
+ gnu_obj_type = TREE_TYPE (TREE_TYPE (gnu_ptr));
+
+ /* If this is a thin pointer, the object must have been allocated with
+ the template in front of the array. So compute the template address
+ and return it. */
+ if (TYPE_IS_THIN_POINTER_P (TREE_TYPE (gnu_ptr)))
+ gnu_ptr
+ = build_binary_op (POINTER_PLUS_EXPR, TREE_TYPE (gnu_ptr),
+ gnu_ptr,
+ fold_build1 (NEGATE_EXPR, sizetype,
+ byte_position
+ (DECL_CHAIN
+ TYPE_FIELDS ((gnu_obj_type)))));
+
+ gnu_result = convert (gnu_result_type, gnu_ptr);
+ }
+ break;
+
+ case Attr_Size:
+ case Attr_Object_Size:
+ case Attr_Value_Size:
+ case Attr_Max_Size_In_Storage_Elements:
+ /* Strip NOPs, conversions between original and packable versions, and
+ unpadding from GNU_PREFIX. Note that we cannot simply strip every
+ VIEW_CONVERT_EXPR because some of them may give the actual size, e.g.
+ for nominally unconstrained packed array. We use GNU_EXPR to see
+ if a COMPONENT_REF was involved. */
+ while (CONVERT_EXPR_P (gnu_prefix)
+ || TREE_CODE (gnu_prefix) == NON_LVALUE_EXPR
+ || (TREE_CODE (gnu_prefix) == VIEW_CONVERT_EXPR
+ && TREE_CODE (TREE_TYPE (gnu_prefix)) == RECORD_TYPE
+ && TREE_CODE (TREE_TYPE (TREE_OPERAND (gnu_prefix, 0)))
+ == RECORD_TYPE
+ && TYPE_NAME (TREE_TYPE (gnu_prefix))
+ == TYPE_NAME (TREE_TYPE (TREE_OPERAND (gnu_prefix, 0)))))
+ gnu_prefix = TREE_OPERAND (gnu_prefix, 0);
+ gnu_expr = gnu_prefix;
+ if (TREE_CODE (gnu_prefix) == COMPONENT_REF
+ && TYPE_IS_PADDING_P (TREE_TYPE (TREE_OPERAND (gnu_prefix, 0))))
+ gnu_prefix = TREE_OPERAND (gnu_prefix, 0);
+ prefix_unused = true;
+ gnu_type = TREE_TYPE (gnu_prefix);
+
+ /* Replace an unconstrained array type with the type of the underlying
+ array, except for 'Max_Size_In_Storage_Elements because we need to
+ return the (maximum) size requested for an allocator. */
+ if (TREE_CODE (gnu_type) == UNCONSTRAINED_ARRAY_TYPE)
+ {
+ gnu_type = TYPE_OBJECT_RECORD_TYPE (gnu_type);
+ if (attribute != Attr_Max_Size_In_Storage_Elements)
+ gnu_type = TREE_TYPE (DECL_CHAIN (TYPE_FIELDS (gnu_type)));
+ }
+
+ /* The type must be frozen at this point. */
+ gcc_assert (COMPLETE_TYPE_P (gnu_type));
+
+ /* If we're looking for the size of a field, return the field size. */
+ if (TREE_CODE (gnu_prefix) == COMPONENT_REF)
+ gnu_result = DECL_SIZE (TREE_OPERAND (gnu_prefix, 1));
+
+ /* Otherwise, if the prefix is an object, or if we are looking for
+ 'Object_Size or 'Max_Size_In_Storage_Elements, the result is the
+ GCC size of the type. We make an exception for padded objects,
+ as we do not take into account alignment promotions for the size.
+ This is in keeping with the object case of gnat_to_gnu_entity. */
+ else if ((TREE_CODE (gnu_prefix) != TYPE_DECL
+ && !(TYPE_IS_PADDING_P (gnu_type)
+ && TREE_CODE (gnu_expr) == COMPONENT_REF
+ && pad_type_has_rm_size (gnu_type)))
+ || attribute == Attr_Object_Size
+ || attribute == Attr_Max_Size_In_Storage_Elements)
+ {
+ /* If this is a dereference and we have a special dynamic constrained
+ subtype on the prefix, use it to compute the size; otherwise, use
+ the designated subtype. */
+ if (Nkind (gnat_prefix) == N_Explicit_Dereference)
+ {
+ Node_Id gnat_actual_subtype
+ = Actual_Designated_Subtype (gnat_prefix);
+ tree gnu_ptr_type
+ = TREE_TYPE (gnat_to_gnu (Prefix (gnat_prefix)));
+
+ if (TYPE_IS_FAT_OR_THIN_POINTER_P (gnu_ptr_type)
+ && Present (gnat_actual_subtype))
+ {
+ tree gnu_actual_obj_type
+ = gnat_to_gnu_type (gnat_actual_subtype);
+ gnu_type
+ = build_unc_object_type_from_ptr (gnu_ptr_type,
+ gnu_actual_obj_type,
+ get_identifier ("SIZE"),
+ false);
+ }
+ }
+
+ gnu_result = TYPE_SIZE (gnu_type);
+ }
+
+ /* Otherwise, the result is the RM size of the type. */
+ else
+ gnu_result = rm_size (gnu_type);
+
+ /* Deal with a self-referential size by qualifying the size with the
+ object or returning the maximum size for a type. */
+ if (TREE_CODE (gnu_prefix) != TYPE_DECL)
+ gnu_result = SUBSTITUTE_PLACEHOLDER_IN_EXPR (gnu_result, gnu_prefix);
+ else if (CONTAINS_PLACEHOLDER_P (gnu_result))
+ gnu_result = max_size (gnu_result, true);
+
+ /* If the type contains a template, subtract the padded size of the
+ template, except for 'Max_Size_In_Storage_Elements because we need
+ to return the (maximum) size requested for an allocator. */
+ if (TREE_CODE (gnu_type) == RECORD_TYPE
+ && TYPE_CONTAINS_TEMPLATE_P (gnu_type)
+ && attribute != Attr_Max_Size_In_Storage_Elements)
+ gnu_result
+ = size_binop (MINUS_EXPR, gnu_result,
+ bit_position (DECL_CHAIN (TYPE_FIELDS (gnu_type))));
+
+ /* For 'Max_Size_In_Storage_Elements, adjust the unit. */
+ if (attribute == Attr_Max_Size_In_Storage_Elements)
+ gnu_result = size_binop (CEIL_DIV_EXPR, gnu_result, bitsize_unit_node);
+
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ break;
+
+ case Attr_Alignment:
+ {
+ unsigned int align;
+
+ if (TREE_CODE (gnu_prefix) == COMPONENT_REF
+ && TYPE_IS_PADDING_P (TREE_TYPE (TREE_OPERAND (gnu_prefix, 0))))
+ gnu_prefix = TREE_OPERAND (gnu_prefix, 0);
+
+ gnu_type = TREE_TYPE (gnu_prefix);
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ prefix_unused = true;
+
+ if (TREE_CODE (gnu_prefix) == COMPONENT_REF)
+ align = DECL_ALIGN (TREE_OPERAND (gnu_prefix, 1)) / BITS_PER_UNIT;
+ else
+ {
+ Entity_Id gnat_type = Etype (gnat_prefix);
+ unsigned int double_align;
+ bool is_capped_double, align_clause;
+
+ /* If the default alignment of "double" or larger scalar types is
+ specifically capped and there is an alignment clause neither
+ on the type nor on the prefix itself, return the cap. */
+ if ((double_align = double_float_alignment) > 0)
+ is_capped_double
+ = is_double_float_or_array (gnat_type, &align_clause);
+ else if ((double_align = double_scalar_alignment) > 0)
+ is_capped_double
+ = is_double_scalar_or_array (gnat_type, &align_clause);
+ else
+ is_capped_double = align_clause = false;
+
+ if (is_capped_double
+ && Nkind (gnat_prefix) == N_Identifier
+ && Present (Alignment_Clause (Entity (gnat_prefix))))
+ align_clause = true;
+
+ if (is_capped_double && !align_clause)
+ align = double_align;
+ else
+ align = TYPE_ALIGN (gnu_type) / BITS_PER_UNIT;
+ }
+
+ gnu_result = size_int (align);
+ }
+ break;
+
+ case Attr_First:
+ case Attr_Last:
+ case Attr_Range_Length:
+ prefix_unused = true;
+
+ if (INTEGRAL_TYPE_P (gnu_type) || SCALAR_FLOAT_TYPE_P (gnu_type))
+ {
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+
+ if (attribute == Attr_First)
+ gnu_result = TYPE_MIN_VALUE (gnu_type);
+ else if (attribute == Attr_Last)
+ gnu_result = TYPE_MAX_VALUE (gnu_type);
+ else
+ gnu_result = get_type_length (gnu_type, gnu_result_type);
+ break;
+ }
+
+ /* ... fall through ... */
+
+ case Attr_Length:
+ {
+ int Dimension = (Present (Expressions (gnat_node))
+ ? UI_To_Int (Intval (First (Expressions (gnat_node))))
+ : 1), i;
+ struct parm_attr_d *pa = NULL;
+ Entity_Id gnat_param = Empty;
+ bool unconstrained_ptr_deref = false;
+
+ gnu_prefix = maybe_padded_object (gnu_prefix);
+ gnu_prefix = maybe_unconstrained_array (gnu_prefix);
+
+ /* We treat unconstrained array In parameters specially. We also note
+ whether we are dereferencing a pointer to unconstrained array. */
+ if (!Is_Constrained (Etype (gnat_prefix)))
+ switch (Nkind (gnat_prefix))
+ {
+ case N_Identifier:
+ /* This is the direct case. */
+ if (Ekind (Entity (gnat_prefix)) == E_In_Parameter)
+ gnat_param = Entity (gnat_prefix);
+ break;
+
+ case N_Explicit_Dereference:
+ /* This is the indirect case. Note that we need to be sure that
+ the access value cannot be null as we'll hoist the load. */
+ if (Nkind (Prefix (gnat_prefix)) == N_Identifier
+ && Ekind (Entity (Prefix (gnat_prefix))) == E_In_Parameter)
+ {
+ if (Can_Never_Be_Null (Entity (Prefix (gnat_prefix))))
+ gnat_param = Entity (Prefix (gnat_prefix));
+ }
+ else
+ unconstrained_ptr_deref = true;
+ break;
+
+ default:
+ break;
+ }
+
+ /* If the prefix is the view conversion of a constrained array to an
+ unconstrained form, we retrieve the constrained array because we
+ might not be able to substitute the PLACEHOLDER_EXPR coming from
+ the conversion. This can occur with the 'Old attribute applied
+ to a parameter with an unconstrained type, which gets rewritten
+ into a constrained local variable very late in the game. */
+ if (TREE_CODE (gnu_prefix) == VIEW_CONVERT_EXPR
+ && CONTAINS_PLACEHOLDER_P (TYPE_SIZE (TREE_TYPE (gnu_prefix)))
+ && !CONTAINS_PLACEHOLDER_P
+ (TYPE_SIZE (TREE_TYPE (TREE_OPERAND (gnu_prefix, 0)))))
+ gnu_type = TREE_TYPE (TREE_OPERAND (gnu_prefix, 0));
+ else
+ gnu_type = TREE_TYPE (gnu_prefix);
+
+ prefix_unused = true;
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+
+ if (TYPE_CONVENTION_FORTRAN_P (gnu_type))
+ {
+ int ndim;
+ tree gnu_type_temp;
+
+ for (ndim = 1, gnu_type_temp = gnu_type;
+ TREE_CODE (TREE_TYPE (gnu_type_temp)) == ARRAY_TYPE
+ && TYPE_MULTI_ARRAY_P (TREE_TYPE (gnu_type_temp));
+ ndim++, gnu_type_temp = TREE_TYPE (gnu_type_temp))
+ ;
+
+ Dimension = ndim + 1 - Dimension;
+ }
+
+ for (i = 1; i < Dimension; i++)
+ gnu_type = TREE_TYPE (gnu_type);
+
+ gcc_assert (TREE_CODE (gnu_type) == ARRAY_TYPE);
+
+ /* When not optimizing, look up the slot associated with the parameter
+ and the dimension in the cache and create a new one on failure.
+ Don't do this when the actual subtype needs debug info (this happens
+ with -gnatD): in elaborate_expression_1, we create variables that
+ hold the bounds, so caching attributes isn't very interesting and
+ causes dependency issues between these variables and cached
+ expressions. */
+ if (!optimize
+ && Present (gnat_param)
+ && !(Present (Actual_Subtype (gnat_param))
+ && Needs_Debug_Info (Actual_Subtype (gnat_param))))
+ {
+ FOR_EACH_VEC_SAFE_ELT (f_parm_attr_cache, i, pa)
+ if (pa->id == gnat_param && pa->dim == Dimension)
+ break;
+
+ if (!pa)
+ {
+ pa = ggc_cleared_alloc<parm_attr_d> ();
+ pa->id = gnat_param;
+ pa->dim = Dimension;
+ vec_safe_push (f_parm_attr_cache, pa);
+ }
+ }
+
+ /* Return the cached expression or build a new one. */
+ if (attribute == Attr_First)
+ {
+ if (pa && pa->first)
+ {
+ gnu_result = pa->first;
+ break;
+ }
+
+ gnu_result
+ = TYPE_MIN_VALUE (TYPE_INDEX_TYPE (TYPE_DOMAIN (gnu_type)));
+ }
+
+ else if (attribute == Attr_Last)
+ {
+ if (pa && pa->last)
+ {
+ gnu_result = pa->last;
+ break;
+ }
+
+ gnu_result
+ = TYPE_MAX_VALUE (TYPE_INDEX_TYPE (TYPE_DOMAIN (gnu_type)));
+ }
+
+ else /* attribute == Attr_Range_Length || attribute == Attr_Length */
+ {
+ if (pa && pa->length)
+ {
+ gnu_result = pa->length;
+ break;
+ }
+
+ gnu_result
+ = get_type_length (TYPE_INDEX_TYPE (TYPE_DOMAIN (gnu_type)),
+ gnu_result_type);
+ }
+
+ /* If this has a PLACEHOLDER_EXPR, qualify it by the object we are
+ handling. Note that these attributes could not have been used on
+ an unconstrained array type. */
+ gnu_result = SUBSTITUTE_PLACEHOLDER_IN_EXPR (gnu_result, gnu_prefix);
+
+ /* Cache the expression we have just computed. Since we want to do it
+ at run time, we force the use of a SAVE_EXPR and let the gimplifier
+ create the temporary in the outermost binding level. We will make
+ sure in Subprogram_Body_to_gnu that it is evaluated on all possible
+ paths by forcing its evaluation on entry of the function. */
+ if (pa)
+ {
+ gnu_result
+ = build1 (SAVE_EXPR, TREE_TYPE (gnu_result), gnu_result);
+ switch (attribute)
+ {
+ case Attr_First:
+ pa->first = gnu_result;
+ break;
+
+ case Attr_Last:
+ pa->last = gnu_result;
+ break;
+
+ case Attr_Length:
+ case Attr_Range_Length:
+ pa->length = gnu_result;
+ break;
+
+ default:
+ gcc_unreachable ();
+ }
+ }
+
+ /* Otherwise, evaluate it each time it is referenced. */
+ else
+ switch (attribute)
+ {
+ case Attr_First:
+ case Attr_Last:
+ /* If we are dereferencing a pointer to unconstrained array, we
+ need to capture the value because the pointed-to bounds may
+ subsequently be released. */
+ if (unconstrained_ptr_deref)
+ gnu_result
+ = build1 (SAVE_EXPR, TREE_TYPE (gnu_result), gnu_result);
+ break;
+
+ case Attr_Length:
+ case Attr_Range_Length:
+ /* Set the source location onto the predicate of the condition
+ but not if the expression is cached to avoid messing up the
+ debug info. */
+ if (TREE_CODE (gnu_result) == COND_EXPR
+ && EXPR_P (TREE_OPERAND (gnu_result, 0)))
+ set_expr_location_from_node (TREE_OPERAND (gnu_result, 0),
+ gnat_node);
+ break;
+
+ default:
+ gcc_unreachable ();
+ }
+
+ break;
+ }
+
+ case Attr_Bit_Position:
+ case Attr_Position:
+ case Attr_First_Bit:
+ case Attr_Last_Bit:
+ case Attr_Bit:
+ {
+ poly_int64 bitsize;
+ poly_int64 bitpos;
+ tree gnu_offset;
+ tree gnu_field_bitpos;
+ tree gnu_field_offset;
+ tree gnu_inner;
+ machine_mode mode;
+ int unsignedp, reversep, volatilep;
+
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ gnu_prefix = remove_conversions (gnu_prefix, true);
+ prefix_unused = true;
+
+ /* We can have 'Bit on any object, but if it isn't a COMPONENT_REF,
+ the result is 0. Don't allow 'Bit on a bare component, though. */
+ if (attribute == Attr_Bit
+ && TREE_CODE (gnu_prefix) != COMPONENT_REF
+ && TREE_CODE (gnu_prefix) != FIELD_DECL)
+ {
+ gnu_result = integer_zero_node;
+ break;
+ }
+
+ else
+ gcc_assert (TREE_CODE (gnu_prefix) == COMPONENT_REF
+ || (attribute == Attr_Bit_Position
+ && TREE_CODE (gnu_prefix) == FIELD_DECL));
+
+ get_inner_reference (gnu_prefix, &bitsize, &bitpos, &gnu_offset,
+ &mode, &unsignedp, &reversep, &volatilep);
+
+ if (TREE_CODE (gnu_prefix) == COMPONENT_REF)
+ {
+ gnu_field_bitpos = bit_position (TREE_OPERAND (gnu_prefix, 1));
+ gnu_field_offset = byte_position (TREE_OPERAND (gnu_prefix, 1));
+
+ for (gnu_inner = TREE_OPERAND (gnu_prefix, 0);
+ TREE_CODE (gnu_inner) == COMPONENT_REF
+ && DECL_INTERNAL_P (TREE_OPERAND (gnu_inner, 1));
+ gnu_inner = TREE_OPERAND (gnu_inner, 0))
+ {
+ gnu_field_bitpos
+ = size_binop (PLUS_EXPR, gnu_field_bitpos,
+ bit_position (TREE_OPERAND (gnu_inner, 1)));
+ gnu_field_offset
+ = size_binop (PLUS_EXPR, gnu_field_offset,
+ byte_position (TREE_OPERAND (gnu_inner, 1)));
+ }
+ }
+ else if (TREE_CODE (gnu_prefix) == FIELD_DECL)
+ {
+ gnu_field_bitpos = bit_position (gnu_prefix);
+ gnu_field_offset = byte_position (gnu_prefix);
+ }
+ else
+ {
+ gnu_field_bitpos = bitsize_zero_node;
+ gnu_field_offset = size_zero_node;
+ }
+
+ switch (attribute)
+ {
+ case Attr_Position:
+ gnu_result = gnu_field_offset;
+ break;
+
+ case Attr_First_Bit:
+ case Attr_Bit:
+ gnu_result = size_int (num_trailing_bits (bitpos));
+ break;
+
+ case Attr_Last_Bit:
+ gnu_result = bitsize_int (num_trailing_bits (bitpos));
+ gnu_result = size_binop (PLUS_EXPR, gnu_result,
+ TYPE_SIZE (TREE_TYPE (gnu_prefix)));
+ /* ??? Avoid a large unsigned result that will overflow when
+ converted to the signed universal_integer. */
+ if (integer_zerop (gnu_result))
+ gnu_result = integer_minus_one_node;
+ else
+ gnu_result
+ = size_binop (MINUS_EXPR, gnu_result, bitsize_one_node);
+ break;
+
+ case Attr_Bit_Position:
+ gnu_result = gnu_field_bitpos;
+ break;
+ }
+
+ /* If this has a PLACEHOLDER_EXPR, qualify it by the object we are
+ handling. */
+ gnu_result = SUBSTITUTE_PLACEHOLDER_IN_EXPR (gnu_result, gnu_prefix);
+ break;
+ }
+
+ case Attr_Min:
+ case Attr_Max:
+ {
+ tree gnu_lhs = gnat_to_gnu (First (Expressions (gnat_node)));
+ tree gnu_rhs = gnat_to_gnu (Next (First (Expressions (gnat_node))));
+
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+
+ /* The result of {MIN,MAX}_EXPR is unspecified if either operand is
+ a NaN so we implement the semantics of C99 f{min,max} to make it
+ predictable in this case: if either operand is a NaN, the other
+ is returned; if both operands are NaN's, a NaN is returned. */
+ if (SCALAR_FLOAT_TYPE_P (gnu_result_type)
+ && !Machine_Overflows_On_Target)
+ {
+ const bool lhs_side_effects_p = TREE_SIDE_EFFECTS (gnu_lhs);
+ const bool rhs_side_effects_p = TREE_SIDE_EFFECTS (gnu_rhs);
+ tree t = builtin_decl_explicit (BUILT_IN_ISNAN);
+ tree lhs_is_nan, rhs_is_nan;
+
+ /* If the operands have side-effects, they need to be evaluated
+ only once in spite of the multiple references in the result. */
+ if (lhs_side_effects_p)
+ gnu_lhs = gnat_protect_expr (gnu_lhs);
+ if (rhs_side_effects_p)
+ gnu_rhs = gnat_protect_expr (gnu_rhs);
+
+ lhs_is_nan = fold_build2 (NE_EXPR, boolean_type_node,
+ build_call_expr (t, 1, gnu_lhs),
+ integer_zero_node);
+
+ rhs_is_nan = fold_build2 (NE_EXPR, boolean_type_node,
+ build_call_expr (t, 1, gnu_rhs),
+ integer_zero_node);
+
+ gnu_result = build_binary_op (attribute == Attr_Min
+ ? MIN_EXPR : MAX_EXPR,
+ gnu_result_type, gnu_lhs, gnu_rhs);
+ gnu_result = fold_build3 (COND_EXPR, gnu_result_type,
+ rhs_is_nan, gnu_lhs, gnu_result);
+ gnu_result = fold_build3 (COND_EXPR, gnu_result_type,
+ lhs_is_nan, gnu_rhs, gnu_result);
+
+ /* If the operands have side-effects, they need to be evaluated
+ before doing the tests above since the place they otherwise
+ would end up being evaluated at run time could be wrong. */
+ if (lhs_side_effects_p)
+ gnu_result
+ = build2 (COMPOUND_EXPR, gnu_result_type, gnu_lhs, gnu_result);
+
+ if (rhs_side_effects_p)
+ gnu_result
+ = build2 (COMPOUND_EXPR, gnu_result_type, gnu_rhs, gnu_result);
+ }
+ else
+ gnu_result = build_binary_op (attribute == Attr_Min
+ ? MIN_EXPR : MAX_EXPR,
+ gnu_result_type, gnu_lhs, gnu_rhs);
+ }
+ break;
+
+ case Attr_Passed_By_Reference:
+ gnu_result = size_int (default_pass_by_ref (gnu_type)
+ || must_pass_by_ref (gnu_type));
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ break;
+
+ case Attr_Component_Size:
+ gnu_prefix = maybe_padded_object (gnu_prefix);
+ gnu_type = TREE_TYPE (gnu_prefix);
+
+ if (TREE_CODE (gnu_type) == UNCONSTRAINED_ARRAY_TYPE)
+ gnu_type = TREE_TYPE (TREE_TYPE (TYPE_FIELDS (TREE_TYPE (gnu_type))));
+
+ while (TREE_CODE (TREE_TYPE (gnu_type)) == ARRAY_TYPE
+ && TYPE_MULTI_ARRAY_P (TREE_TYPE (gnu_type)))
+ gnu_type = TREE_TYPE (gnu_type);
+
+ gcc_assert (TREE_CODE (gnu_type) == ARRAY_TYPE);
+
+ /* Note this size cannot be self-referential. */
+ gnu_result = TYPE_SIZE (TREE_TYPE (gnu_type));
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ prefix_unused = true;
+ break;
+
+ case Attr_Descriptor_Size:
+ gnu_type = TREE_TYPE (gnu_prefix);
+ gcc_assert (TREE_CODE (gnu_type) == UNCONSTRAINED_ARRAY_TYPE);
+
+ /* Return the padded size of the template in the object record type. */
+ gnu_type = TYPE_OBJECT_RECORD_TYPE (gnu_type);
+ gnu_result = bit_position (DECL_CHAIN (TYPE_FIELDS (gnu_type)));
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ prefix_unused = true;
+ break;
+
+ case Attr_Null_Parameter:
+ /* This is just a zero cast to the pointer type for our prefix and
+ dereferenced. */
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ gnu_result
+ = build_unary_op (INDIRECT_REF, NULL_TREE,
+ convert (build_pointer_type (gnu_result_type),
+ integer_zero_node));
+ break;
+
+ case Attr_Mechanism_Code:
+ {
+ Entity_Id gnat_obj = Entity (gnat_prefix);
+ int code;
+
+ prefix_unused = true;
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ if (Present (Expressions (gnat_node)))
+ {
+ int i = UI_To_Int (Intval (First (Expressions (gnat_node))));
+
+ for (gnat_obj = First_Formal (gnat_obj); i > 1;
+ i--, gnat_obj = Next_Formal (gnat_obj))
+ ;
+ }
+
+ code = Mechanism (gnat_obj);
+ if (code == Default)
+ code = ((present_gnu_tree (gnat_obj)
+ && (DECL_BY_REF_P (get_gnu_tree (gnat_obj))
+ || ((TREE_CODE (get_gnu_tree (gnat_obj))
+ == PARM_DECL)
+ && (DECL_BY_COMPONENT_PTR_P
+ (get_gnu_tree (gnat_obj))))))
+ ? By_Reference : By_Copy);
+ gnu_result = convert (gnu_result_type, size_int (- code));
+ }
+ break;
+
+ case Attr_Model:
+ /* We treat Model as identical to Machine. This is true for at least
+ IEEE and some other nice floating-point systems. */
+
+ /* ... fall through ... */
+
+ case Attr_Machine:
+ /* The trick is to force the compiler to store the result in memory so
+ that we do not have extra precision used. But do this only when this
+ is necessary, i.e. if FP_ARITH_MAY_WIDEN is true and the precision of
+ the type is lower than that of the longest floating-point type. */
+ prefix_unused = true;
+ gnu_expr = gnat_to_gnu (First (Expressions (gnat_node)));
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ gnu_result = convert (gnu_result_type, gnu_expr);
+
+ if (TREE_CODE (gnu_result) != REAL_CST
+ && fp_arith_may_widen
+ && TYPE_PRECISION (gnu_result_type)
+ < TYPE_PRECISION (longest_float_type_node))
+ {
+ tree rec_type = make_node (RECORD_TYPE);
+ tree field
+ = create_field_decl (get_identifier ("OBJ"), gnu_result_type,
+ rec_type, NULL_TREE, NULL_TREE, 0, 0);
+ tree rec_val, asm_expr;
+
+ finish_record_type (rec_type, field, 0, false);
+
+ rec_val = build_constructor_single (rec_type, field, gnu_result);
+ rec_val = build1 (SAVE_EXPR, rec_type, rec_val);
+
+ asm_expr
+ = build5 (ASM_EXPR, void_type_node,
+ build_string (0, ""),
+ tree_cons (build_tree_list (NULL_TREE,
+ build_string (2, "=m")),
+ rec_val, NULL_TREE),
+ tree_cons (build_tree_list (NULL_TREE,
+ build_string (1, "m")),
+ rec_val, NULL_TREE),
+ NULL_TREE, NULL_TREE);
+ ASM_VOLATILE_P (asm_expr) = 1;
+
+ gnu_result
+ = build_compound_expr (gnu_result_type, asm_expr,
+ build_component_ref (rec_val, field,
+ false));
+ }
+ break;
+
+ case Attr_Deref:
+ prefix_unused = true;
+ gnu_expr = gnat_to_gnu (First (Expressions (gnat_node)));
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ /* This can be a random address so build an alias-all pointer type. */
+ gnu_expr
+ = convert (build_pointer_type_for_mode (gnu_result_type, ptr_mode,
+ true),
+ gnu_expr);
+ gnu_result = build_unary_op (INDIRECT_REF, NULL_TREE, gnu_expr);
+ break;
+
+ default:
+ /* This abort means that we have an unimplemented attribute. */
+ gcc_unreachable ();
+ }
+
+ /* If this is an attribute where the prefix was unused, force a use of it if
+ it has a side-effect. But don't do it if the prefix is just an entity
+ name. However, if an access check is needed, we must do it. See second
+ example in AARM 11.6(5.e). */
+ if (prefix_unused
+ && TREE_SIDE_EFFECTS (gnu_prefix)
+ && !Is_Entity_Name (gnat_prefix))
+ gnu_result
+ = build_compound_expr (TREE_TYPE (gnu_result), gnu_prefix, gnu_result);
+
+ *gnu_result_type_p = gnu_result_type;
+ return gnu_result;
+}
+
+/* Subroutine of gnat_to_gnu to translate gnat_node, an N_Case_Statement,
+ to a GCC tree, which is returned. */
+
+static tree
+Case_Statement_to_gnu (Node_Id gnat_node)
+{
+ tree gnu_result, gnu_expr, gnu_type, gnu_label;
+ Node_Id gnat_when;
+ location_t end_locus;
+ bool may_fallthru = false;
+
+ gnu_expr = gnat_to_gnu (Expression (gnat_node));
+ gnu_expr = convert (get_base_type (TREE_TYPE (gnu_expr)), gnu_expr);
+ gnu_expr = maybe_character_value (gnu_expr);
+ gnu_type = TREE_TYPE (gnu_expr);
+
+ /* We build a SWITCH_EXPR that contains the code with interspersed
+ CASE_LABEL_EXPRs for each label. */
+ if (!Sloc_to_locus (End_Location (gnat_node), &end_locus))
+ end_locus = input_location;
+ gnu_label = create_artificial_label (end_locus);
+ start_stmt_group ();
+
+ for (gnat_when = First_Non_Pragma (Alternatives (gnat_node));
+ Present (gnat_when);
+ gnat_when = Next_Non_Pragma (gnat_when))
+ {
+ bool choices_added_p = false;
+ Node_Id gnat_choice;
+
+ /* First compile all the different case choices for the current WHEN
+ alternative. */
+ for (gnat_choice = First (Discrete_Choices (gnat_when));
+ Present (gnat_choice);
+ gnat_choice = Next (gnat_choice))
+ {
+ tree gnu_low = NULL_TREE, gnu_high = NULL_TREE;
+ tree label = create_artificial_label (input_location);
+
+ switch (Nkind (gnat_choice))
+ {
+ case N_Range:
+ gnu_low = gnat_to_gnu (Low_Bound (gnat_choice));
+ gnu_high = gnat_to_gnu (High_Bound (gnat_choice));
+ break;
+
+ case N_Subtype_Indication:
+ gnu_low = gnat_to_gnu (Low_Bound (Range_Expression
+ (Constraint (gnat_choice))));
+ gnu_high = gnat_to_gnu (High_Bound (Range_Expression
+ (Constraint (gnat_choice))));
+ break;
+
+ case N_Identifier:
+ case N_Expanded_Name:
+ /* This represents either a subtype range or a static value of
+ some kind; Ekind says which. */
+ if (Is_Type (Entity (gnat_choice)))
+ {
+ tree gnu_type = get_unpadded_type (Entity (gnat_choice));
+
+ gnu_low = TYPE_MIN_VALUE (gnu_type);
+ gnu_high = TYPE_MAX_VALUE (gnu_type);
+ break;
+ }
+
+ /* ... fall through ... */
+
+ case N_Character_Literal:
+ case N_Integer_Literal:
+ gnu_low = gnat_to_gnu (gnat_choice);
+ break;
+
+ case N_Others_Choice:
+ break;
+
+ default:
+ gcc_unreachable ();
+ }
+
+ /* Everything should be folded into constants at this point. */
+ gcc_assert (!gnu_low || TREE_CODE (gnu_low) == INTEGER_CST);
+ gcc_assert (!gnu_high || TREE_CODE (gnu_high) == INTEGER_CST);
+
+ if (gnu_low && TREE_TYPE (gnu_low) != gnu_type)
+ gnu_low = convert (gnu_type, gnu_low);
+ if (gnu_high && TREE_TYPE (gnu_high) != gnu_type)
+ gnu_high = convert (gnu_type, gnu_high);
+
+ add_stmt_with_node (build_case_label (gnu_low, gnu_high, label),
+ gnat_choice);
+ choices_added_p = true;
+ }
+
+ /* This construct doesn't define a scope so we shouldn't push a binding
+ level around the statement list. Except that we have always done so
+ historically and this makes it possible to reduce stack usage. As a
+ compromise, we keep doing it for case statements, for which this has
+ never been problematic, but not for case expressions in Ada 2012. */
+ if (choices_added_p)
+ {
+ const bool is_case_expression
+ = (Nkind (Parent (gnat_node)) == N_Expression_With_Actions);
+ tree group
+ = build_stmt_group (Statements (gnat_when), !is_case_expression);
+ bool group_may_fallthru = block_may_fallthru (group);
+ add_stmt (group);
+ if (group_may_fallthru)
+ {
+ tree stmt = build1 (GOTO_EXPR, void_type_node, gnu_label);
+ SET_EXPR_LOCATION (stmt, end_locus);
+ add_stmt (stmt);
+ may_fallthru = true;
+ }
+ }
+ }
+
+ /* Now emit a definition of the label the cases branch to, if any. */
+ if (may_fallthru)
+ add_stmt (build1 (LABEL_EXPR, void_type_node, gnu_label));
+ gnu_result = build2 (SWITCH_EXPR, gnu_type, gnu_expr, end_stmt_group ());
+
+ return gnu_result;
+}
+
+/* Return true if we are in the body of a loop. */
+
+static inline bool
+inside_loop_p (void)
+{
+ return !vec_safe_is_empty (gnu_loop_stack);
+}
+
+/* Find out whether EXPR is a simple additive expression based on the iteration
+ variable of some enclosing loop in the current function. If so, return the
+ loop and set *DISP to the displacement and *NEG_P to true if this is for a
+ subtraction; otherwise, return NULL. */
+
+static struct loop_info_d *
+find_loop_for (tree expr, tree *disp, bool *neg_p)
+{
+ tree var, add, cst;
+ bool minus_p;
+ struct loop_info_d *iter = NULL;
+ unsigned int i;
+
+ if (is_simple_additive_expression (expr, &add, &cst, &minus_p))
+ {
+ var = add;
+ if (disp)
+ *disp = cst;
+ if (neg_p)
+ *neg_p = minus_p;
+ }
+ else
+ {
+ var = expr;
+ if (disp)
+ *disp = NULL_TREE;
+ if (neg_p)
+ *neg_p = false;
+ }
+
+ var = remove_conversions (var, false);
+
+ if (TREE_CODE (var) != VAR_DECL)
+ return NULL;
+
+ gcc_checking_assert (vec_safe_length (gnu_loop_stack) > 0);
+
+ FOR_EACH_VEC_ELT_REVERSE (*gnu_loop_stack, i, iter)
+ if (iter->loop_var == var && iter->fndecl == current_function_decl)
+ break;
+
+ return iter;
+}
+
+/* Return the innermost enclosing loop in the current function. */
+
+static struct loop_info_d *
+find_loop (void)
+{
+ struct loop_info_d *iter = NULL;
+ unsigned int i;
+
+ gcc_checking_assert (vec_safe_length (gnu_loop_stack) > 0);
+
+ FOR_EACH_VEC_ELT_REVERSE (*gnu_loop_stack, i, iter)
+ if (iter->fndecl == current_function_decl)
+ break;
+
+ return iter;
+}
+
+/* Return true if VAL (of type TYPE) can equal the minimum value if MAX is
+ false, or the maximum value if MAX is true, of TYPE. */
+
+static bool
+can_equal_min_or_max_val_p (tree val, tree type, bool max)
+{
+ tree min_or_max_val = (max ? TYPE_MAX_VALUE (type) : TYPE_MIN_VALUE (type));
+
+ if (TREE_CODE (min_or_max_val) != INTEGER_CST)
+ return true;
+
+ if (TREE_CODE (val) == NOP_EXPR)
+ val = (max
+ ? TYPE_MAX_VALUE (TREE_TYPE (TREE_OPERAND (val, 0)))
+ : TYPE_MIN_VALUE (TREE_TYPE (TREE_OPERAND (val, 0))));
+
+ if (TREE_CODE (val) != INTEGER_CST)
+ return true;
+
+ if (max)
+ return tree_int_cst_lt (val, min_or_max_val) == 0;
+ else
+ return tree_int_cst_lt (min_or_max_val, val) == 0;
+}
+
+/* Return true if VAL (of type TYPE) can equal the minimum value of TYPE.
+ If REVERSE is true, minimum value is taken as maximum value. */
+
+static inline bool
+can_equal_min_val_p (tree val, tree type, bool reverse)
+{
+ return can_equal_min_or_max_val_p (val, type, reverse);
+}
+
+/* Return true if VAL (of type TYPE) can equal the maximum value of TYPE.
+ If REVERSE is true, maximum value is taken as minimum value. */
+
+static inline bool
+can_equal_max_val_p (tree val, tree type, bool reverse)
+{
+ return can_equal_min_or_max_val_p (val, type, !reverse);
+}
+
+/* Replace EXPR1 and EXPR2 by invariant expressions if possible. Return
+ true if both expressions have been replaced and false otherwise. */
+
+static bool
+make_invariant (tree *expr1, tree *expr2)
+{
+ tree inv_expr1 = gnat_invariant_expr (*expr1);
+ tree inv_expr2 = gnat_invariant_expr (*expr2);
+
+ if (inv_expr1)
+ *expr1 = inv_expr1;
+
+ if (inv_expr2)
+ *expr2 = inv_expr2;
+
+ return inv_expr1 && inv_expr2;
+}
+
+/* Helper function for walk_tree, used by independent_iterations_p below. */
+
+static tree
+scan_rhs_r (tree *tp, int *walk_subtrees, void *data)
+{
+ bitmap *params = (bitmap *)data;
+ tree t = *tp;
+
+ /* No need to walk into types or decls. */
+ if (IS_TYPE_OR_DECL_P (t))
+ *walk_subtrees = 0;
+
+ if (TREE_CODE (t) == PARM_DECL && bitmap_bit_p (*params, DECL_UID (t)))
+ return t;
+
+ return NULL_TREE;
+}
+
+/* Return true if STMT_LIST generates independent iterations in a loop. */
+
+static bool
+independent_iterations_p (tree stmt_list)
+{
+ tree_stmt_iterator tsi;
+ bitmap params = BITMAP_GGC_ALLOC();
+ auto_vec<tree, 16> rhs;
+ tree iter;
+ int i;
+
+ if (TREE_CODE (stmt_list) == BIND_EXPR)
+ stmt_list = BIND_EXPR_BODY (stmt_list);
+
+ /* Scan the list and return false on anything that is not either a check
+ or an assignment to a parameter with restricted aliasing. */
+ for (tsi = tsi_start (stmt_list); !tsi_end_p (tsi); tsi_next (&tsi))
+ {
+ tree stmt = tsi_stmt (tsi);
+
+ switch (TREE_CODE (stmt))
+ {
+ case COND_EXPR:
+ {
+ if (COND_EXPR_ELSE (stmt))
+ return false;
+ if (TREE_CODE (COND_EXPR_THEN (stmt)) != CALL_EXPR)
+ return false;
+ tree func = get_callee_fndecl (COND_EXPR_THEN (stmt));
+ if (!(func && TREE_THIS_VOLATILE (func)))
+ return false;
+ break;
+ }
+
+ case MODIFY_EXPR:
+ {
+ tree lhs = TREE_OPERAND (stmt, 0);
+ while (handled_component_p (lhs))
+ lhs = TREE_OPERAND (lhs, 0);
+ if (TREE_CODE (lhs) != INDIRECT_REF)
+ return false;
+ lhs = TREE_OPERAND (lhs, 0);
+ if (!(TREE_CODE (lhs) == PARM_DECL
+ && DECL_RESTRICTED_ALIASING_P (lhs)))
+ return false;
+ bitmap_set_bit (params, DECL_UID (lhs));
+ rhs.safe_push (TREE_OPERAND (stmt, 1));
+ break;
+ }
+
+ default:
+ return false;
+ }
+ }
+
+ /* At this point we know that the list contains only statements that will
+ modify parameters with restricted aliasing. Check that the statements
+ don't at the time read from these parameters. */
+ FOR_EACH_VEC_ELT (rhs, i, iter)
+ if (walk_tree_without_duplicates (&iter, scan_rhs_r, &params))
+ return false;
+
+ return true;
+}
+
+/* Subroutine of gnat_to_gnu to translate gnat_node, an N_Loop_Statement,
+ to a GCC tree, which is returned. */
+
+static tree
+Loop_Statement_to_gnu (Node_Id gnat_node)
+{
+ const Node_Id gnat_iter_scheme = Iteration_Scheme (gnat_node);
+ struct loop_info_d *gnu_loop_info = ggc_cleared_alloc<loop_info_d> ();
+ tree gnu_loop_stmt = build4 (LOOP_STMT, void_type_node, NULL_TREE,
+ NULL_TREE, NULL_TREE, NULL_TREE);
+ tree gnu_loop_label = create_artificial_label (input_location);
+ tree gnu_cond_expr = NULL_TREE, gnu_low = NULL_TREE, gnu_high = NULL_TREE;
+ tree gnu_result;
+
+ /* Push the loop_info structure associated with the LOOP_STMT. */
+ gnu_loop_info->fndecl = current_function_decl;
+ gnu_loop_info->stmt = gnu_loop_stmt;
+ vec_safe_push (gnu_loop_stack, gnu_loop_info);
+
+ /* Set location information for statement and end label. */
+ set_expr_location_from_node (gnu_loop_stmt, gnat_node);
+ Sloc_to_locus (Sloc (End_Label (gnat_node)),
+ &DECL_SOURCE_LOCATION (gnu_loop_label));
+ LOOP_STMT_LABEL (gnu_loop_stmt) = gnu_loop_label;
+
+ /* Set the condition under which the loop must keep going. If we have an
+ explicit condition, use it to set the location information throughout
+ the translation of the loop statement to avoid having multiple SLOCs.
+
+ For the case "LOOP .... END LOOP;" the condition is always true. */
+ if (No (gnat_iter_scheme))
+ ;
+
+ /* For the case "WHILE condition LOOP ..... END LOOP;" it's immediate. */
+ else if (Present (Condition (gnat_iter_scheme)))
+ {
+ LOOP_STMT_COND (gnu_loop_stmt)
+ = gnat_to_gnu (Condition (gnat_iter_scheme));
+
+ set_expr_location_from_node (gnu_loop_stmt, gnat_iter_scheme);
+ }
+
+ /* Otherwise we have an iteration scheme and the condition is given by the
+ bounds of the subtype of the iteration variable. */
+ else
+ {
+ Node_Id gnat_loop_spec = Loop_Parameter_Specification (gnat_iter_scheme);
+ Entity_Id gnat_loop_var = Defining_Entity (gnat_loop_spec);
+ Entity_Id gnat_type = Etype (gnat_loop_var);
+ tree gnu_type = get_unpadded_type (gnat_type);
+ tree gnu_base_type = maybe_character_type (get_base_type (gnu_type));
+ tree gnu_one_node = build_int_cst (gnu_base_type, 1);
+ tree gnu_loop_var, gnu_loop_iv, gnu_first, gnu_last, gnu_stmt;
+ enum tree_code update_code, test_code, shift_code;
+ bool reverse = Reverse_Present (gnat_loop_spec), use_iv = false;
+
+ gnu_low = convert (gnu_base_type, TYPE_MIN_VALUE (gnu_type));
+ gnu_high = convert (gnu_base_type, TYPE_MAX_VALUE (gnu_type));
+
+ /* We must disable modulo reduction for the iteration variable, if any,
+ in order for the loop comparison to be effective. */
+ if (reverse)
+ {
+ gnu_first = gnu_high;
+ gnu_last = gnu_low;
+ update_code = MINUS_NOMOD_EXPR;
+ test_code = GE_EXPR;
+ shift_code = PLUS_NOMOD_EXPR;
+ }
+ else
+ {
+ gnu_first = gnu_low;
+ gnu_last = gnu_high;
+ update_code = PLUS_NOMOD_EXPR;
+ test_code = LE_EXPR;
+ shift_code = MINUS_NOMOD_EXPR;
+ }
+
+ /* We use two different strategies to translate the loop, depending on
+ whether optimization is enabled.
+
+ If it is, we generate the canonical loop form expected by the loop
+ optimizer and the loop vectorizer, which is the do-while form:
+
+ ENTRY_COND
+ loop:
+ TOP_UPDATE
+ BODY
+ BOTTOM_COND
+ GOTO loop
+
+ This avoids an implicit dependency on loop header copying and makes
+ it possible to turn BOTTOM_COND into an inequality test.
+
+ If optimization is disabled, loop header copying doesn't come into
+ play and we try to generate the loop form with the fewer conditional
+ branches. First, the default form, which is:
+
+ loop:
+ TOP_COND
+ BODY
+ BOTTOM_UPDATE
+ GOTO loop
+
+ It should catch most loops with constant ending point. Then, if we
+ cannot, we try to generate the shifted form:
+
+ loop:
+ TOP_COND
+ TOP_UPDATE
+ BODY
+ GOTO loop
+
+ which should catch loops with constant starting point. Otherwise, if
+ we cannot, we generate the fallback form:
+
+ ENTRY_COND
+ loop:
+ BODY
+ BOTTOM_COND
+ BOTTOM_UPDATE
+ GOTO loop
+
+ which works in all cases. */
+
+ if (optimize && !optimize_debug)
+ {
+ /* We can use the do-while form directly if GNU_FIRST-1 doesn't
+ overflow. */
+ if (!can_equal_min_val_p (gnu_first, gnu_base_type, reverse))
+ ;
+
+ /* Otherwise, use the do-while form with the help of a special
+ induction variable in the unsigned version of the base type
+ or the unsigned version of the size type, whichever is the
+ largest, in order to have wrap-around arithmetics for it. */
+ else
+ {
+ if (TYPE_PRECISION (gnu_base_type)
+ > TYPE_PRECISION (size_type_node))
+ gnu_base_type
+ = gnat_type_for_size (TYPE_PRECISION (gnu_base_type), 1);
+ else
+ gnu_base_type = size_type_node;
+
+ gnu_first = convert (gnu_base_type, gnu_first);
+ gnu_last = convert (gnu_base_type, gnu_last);
+ gnu_one_node = build_int_cst (gnu_base_type, 1);
+ use_iv = true;
+ }
+
+ gnu_first
+ = build_binary_op (shift_code, gnu_base_type, gnu_first,
+ gnu_one_node);
+ LOOP_STMT_TOP_UPDATE_P (gnu_loop_stmt) = 1;
+ LOOP_STMT_BOTTOM_COND_P (gnu_loop_stmt) = 1;
+ }
+ else
+ {
+ /* We can use the default form if GNU_LAST+1 doesn't overflow. */
+ if (!can_equal_max_val_p (gnu_last, gnu_base_type, reverse))
+ ;
+
+ /* Otherwise, we can use the shifted form if neither GNU_FIRST-1 nor
+ GNU_LAST-1 does. */
+ else if (!can_equal_min_val_p (gnu_first, gnu_base_type, reverse)
+ && !can_equal_min_val_p (gnu_last, gnu_base_type, reverse))
+ {
+ gnu_first
+ = build_binary_op (shift_code, gnu_base_type, gnu_first,
+ gnu_one_node);
+ gnu_last
+ = build_binary_op (shift_code, gnu_base_type, gnu_last,
+ gnu_one_node);
+ LOOP_STMT_TOP_UPDATE_P (gnu_loop_stmt) = 1;
+ }
+
+ /* Otherwise, use the fallback form. */
+ else
+ LOOP_STMT_BOTTOM_COND_P (gnu_loop_stmt) = 1;
+ }
+
+ /* If we use the BOTTOM_COND, we can turn the test into an inequality
+ test but we have to add ENTRY_COND to protect the empty loop. */
+ if (LOOP_STMT_BOTTOM_COND_P (gnu_loop_stmt))
+ {
+ test_code = NE_EXPR;
+ gnu_cond_expr
+ = build3 (COND_EXPR, void_type_node,
+ build_binary_op (LE_EXPR, boolean_type_node,
+ gnu_low, gnu_high),
+ NULL_TREE, alloc_stmt_list ());
+ set_expr_location_from_node (gnu_cond_expr, gnat_iter_scheme);
+ }
+
+ /* Open a new nesting level that will surround the loop to declare the
+ iteration variable. */
+ start_stmt_group ();
+ gnat_pushlevel ();
+
+ /* If we use the special induction variable, create it and set it to
+ its initial value. Morever, the regular iteration variable cannot
+ itself be initialized, lest the initial value wrapped around. */
+ if (use_iv)
+ {
+ gnu_loop_iv
+ = create_init_temporary ("I", gnu_first, &gnu_stmt, gnat_loop_var);
+ add_stmt (gnu_stmt);
+ gnu_first = NULL_TREE;
+ }
+ else
+ gnu_loop_iv = NULL_TREE;
+
+ /* Declare the iteration variable and set it to its initial value. */
+ gnu_loop_var = gnat_to_gnu_entity (gnat_loop_var, gnu_first, true);
+ if (DECL_BY_REF_P (gnu_loop_var))
+ gnu_loop_var = build_unary_op (INDIRECT_REF, NULL_TREE, gnu_loop_var);
+ else if (use_iv)
+ {
+ gcc_assert (DECL_LOOP_PARM_P (gnu_loop_var));
+ SET_DECL_INDUCTION_VAR (gnu_loop_var, gnu_loop_iv);
+ }
+ gnu_loop_info->loop_var = gnu_loop_var;
+ gnu_loop_info->low_bound = gnu_low;
+ gnu_loop_info->high_bound = gnu_high;
+
+ /* Do all the arithmetics in the base type. */
+ gnu_loop_var = convert (gnu_base_type, gnu_loop_var);
+
+ /* Set either the top or bottom exit condition. */
+ if (use_iv)
+ LOOP_STMT_COND (gnu_loop_stmt)
+ = build_binary_op (test_code, boolean_type_node, gnu_loop_iv,
+ gnu_last);
+ else
+ LOOP_STMT_COND (gnu_loop_stmt)
+ = build_binary_op (test_code, boolean_type_node, gnu_loop_var,
+ gnu_last);
+
+ /* Set either the top or bottom update statement and give it the source
+ location of the iteration for better coverage info. */
+ if (use_iv)
+ {
+ gnu_stmt
+ = build_binary_op (MODIFY_EXPR, NULL_TREE, gnu_loop_iv,
+ build_binary_op (update_code, gnu_base_type,
+ gnu_loop_iv, gnu_one_node));
+ set_expr_location_from_node (gnu_stmt, gnat_iter_scheme);
+ append_to_statement_list (gnu_stmt,
+ &LOOP_STMT_UPDATE (gnu_loop_stmt));
+ gnu_stmt
+ = build_binary_op (MODIFY_EXPR, NULL_TREE, gnu_loop_var,
+ gnu_loop_iv);
+ set_expr_location_from_node (gnu_stmt, gnat_iter_scheme);
+ append_to_statement_list (gnu_stmt,
+ &LOOP_STMT_UPDATE (gnu_loop_stmt));
+ }
+ else
+ {
+ gnu_stmt
+ = build_binary_op (MODIFY_EXPR, NULL_TREE, gnu_loop_var,
+ build_binary_op (update_code, gnu_base_type,
+ gnu_loop_var, gnu_one_node));
+ set_expr_location_from_node (gnu_stmt, gnat_iter_scheme);
+ LOOP_STMT_UPDATE (gnu_loop_stmt) = gnu_stmt;
+ }
+
+ set_expr_location_from_node (gnu_loop_stmt, gnat_iter_scheme);
+ }
+
+ /* If the loop was named, have the name point to this loop. In this case,
+ the association is not a DECL node, but the end label of the loop. */
+ if (Present (Identifier (gnat_node)))
+ save_gnu_tree (Entity (Identifier (gnat_node)), gnu_loop_label, true);
+
+ /* Make the loop body into its own block, so any allocated storage will be
+ released every iteration. This is needed for stack allocation. */
+ LOOP_STMT_BODY (gnu_loop_stmt)
+ = build_stmt_group (Statements (gnat_node), true);
+ TREE_SIDE_EFFECTS (gnu_loop_stmt) = 1;
+
+ /* If we have an iteration scheme, then we are in a statement group. Add
+ the LOOP_STMT to it, finish it and make it the "loop". */
+ if (Present (gnat_iter_scheme) && No (Condition (gnat_iter_scheme)))
+ {
+ /* First, if we have computed invariant conditions for range (or index)
+ checks applied to the iteration variable, find out whether they can
+ be evaluated to false at compile time; otherwise, if there are not
+ too many of them, combine them with the original checks. If loop
+ unswitching is enabled, do not require the loop bounds to be also
+ invariant, as their evaluation will still be ahead of the loop. */
+ if (vec_safe_length (gnu_loop_info->checks) > 0
+ && (make_invariant (&gnu_low, &gnu_high) || optimize >= 3))
+ {
+ struct range_check_info_d *rci;
+ unsigned int i, n_remaining_checks = 0;
+
+ FOR_EACH_VEC_ELT (*gnu_loop_info->checks, i, rci)
+ {
+ tree low_ok, high_ok;
+
+ if (rci->low_bound)
+ {
+ tree gnu_adjusted_low = convert (rci->type, gnu_low);
+ if (rci->disp)
+ gnu_adjusted_low
+ = fold_build2 (rci->neg_p ? MINUS_EXPR : PLUS_EXPR,
+ rci->type, gnu_adjusted_low, rci->disp);
+ low_ok
+ = build_binary_op (GE_EXPR, boolean_type_node,
+ gnu_adjusted_low, rci->low_bound);
+ }
+ else
+ low_ok = boolean_true_node;
+
+ if (rci->high_bound)
+ {
+ tree gnu_adjusted_high = convert (rci->type, gnu_high);
+ if (rci->disp)
+ gnu_adjusted_high
+ = fold_build2 (rci->neg_p ? MINUS_EXPR : PLUS_EXPR,
+ rci->type, gnu_adjusted_high, rci->disp);
+ high_ok
+ = build_binary_op (LE_EXPR, boolean_type_node,
+ gnu_adjusted_high, rci->high_bound);
+ }
+ else
+ high_ok = boolean_true_node;
+
+ tree range_ok
+ = build_binary_op (TRUTH_ANDIF_EXPR, boolean_type_node,
+ low_ok, high_ok);
+
+ rci->invariant_cond
+ = build_unary_op (TRUTH_NOT_EXPR, boolean_type_node, range_ok);
+
+ if (rci->invariant_cond == boolean_false_node)
+ TREE_OPERAND (rci->inserted_cond, 0) = rci->invariant_cond;
+ else
+ n_remaining_checks++;
+ }
+
+ /* Note that loop unswitching can only be applied a small number of
+ times to a given loop (PARAM_MAX_UNSWITCH_LEVEL default to 3). */
+ if (IN_RANGE (n_remaining_checks, 1, 3)
+ && optimize >= 2
+ && !optimize_size)
+ FOR_EACH_VEC_ELT (*gnu_loop_info->checks, i, rci)
+ if (rci->invariant_cond != boolean_false_node)
+ {
+ TREE_OPERAND (rci->inserted_cond, 0) = rci->invariant_cond;
+
+ if (optimize >= 3)
+ add_stmt_with_node_force (rci->inserted_cond, gnat_node);
+ }
+ }
+
+ /* Second, if we have recorded invariants to be hoisted, emit them. */
+ if (vec_safe_length (gnu_loop_info->invariants) > 0)
+ {
+ tree *iter;
+ unsigned int i;
+ FOR_EACH_VEC_ELT (*gnu_loop_info->invariants, i, iter)
+ add_stmt_with_node_force (*iter, gnat_node);
+ }
+
+ /* Third, if loop vectorization is enabled and the iterations of the
+ loop can easily be proved as independent, mark the loop. */
+ if (optimize >= 3
+ && independent_iterations_p (LOOP_STMT_BODY (gnu_loop_stmt)))
+ LOOP_STMT_IVDEP (gnu_loop_stmt) = 1;
+
+ add_stmt (gnu_loop_stmt);
+ gnat_poplevel ();
+ gnu_loop_stmt = end_stmt_group ();
+ }
+
+ /* If we have an outer COND_EXPR, that's our result and this loop is its
+ "true" statement. Otherwise, the result is the LOOP_STMT. */
+ if (gnu_cond_expr)
+ {
+ COND_EXPR_THEN (gnu_cond_expr) = gnu_loop_stmt;
+ TREE_SIDE_EFFECTS (gnu_cond_expr) = 1;
+ gnu_result = gnu_cond_expr;
+ }
+ else
+ gnu_result = gnu_loop_stmt;
+
+ gnu_loop_stack->pop ();
+
+ return gnu_result;
+}
+
+/* This page implements a form of Named Return Value optimization modeled
+ on the C++ optimization of the same name. The main difference is that
+ we disregard any semantical considerations when applying it here, the
+ counterpart being that we don't try to apply it to semantically loaded
+ return types, i.e. types with the TYPE_BY_REFERENCE_P flag set.
+
+ We consider a function body of the following GENERIC form:
+
+ return_type R1;
+ [...]
+ RETURN_EXPR [<retval> = ...]
+ [...]
+ RETURN_EXPR [<retval> = R1]
+ [...]
+ return_type Ri;
+ [...]
+ RETURN_EXPR [<retval> = ...]
+ [...]
+ RETURN_EXPR [<retval> = Ri]
+ [...]
+
+ where the Ri are not addressable and we try to fulfill a simple criterion
+ that would make it possible to replace one or several Ri variables by the
+ single RESULT_DECL of the function.
+
+ The first observation is that RETURN_EXPRs that don't directly reference
+ any of the Ri variables on the RHS of their assignment are transparent wrt
+ the optimization. This is because the Ri variables aren't addressable so
+ any transformation applied to them doesn't affect the RHS; moreover, the
+ assignment writes the full <retval> object so existing values are entirely
+ discarded.
+
+ This property can be extended to some forms of RETURN_EXPRs that reference
+ the Ri variables, for example CONSTRUCTORs, but isn't true in the general
+ case, in particular when function calls are involved.
+
+ Therefore the algorithm is as follows:
+
+ 1. Collect the list of candidates for a Named Return Value (Ri variables
+ on the RHS of assignments of RETURN_EXPRs) as well as the list of the
+ other expressions on the RHS of such assignments.
+
+ 2. Prune the members of the first list (candidates) that are referenced
+ by a member of the second list (expressions).
+
+ 3. Extract a set of candidates with non-overlapping live ranges from the
+ first list. These are the Named Return Values.
+
+ 4. Adjust the relevant RETURN_EXPRs and replace the occurrences of the
+ Named Return Values in the function with the RESULT_DECL.
+
+ If the function returns an unconstrained type, things are a bit different
+ because the anonymous return object is allocated on the secondary stack
+ and RESULT_DECL is only a pointer to it. Each return object can be of a
+ different size and is allocated separately so we need not care about the
+ addressability and the aforementioned overlapping issues. Therefore, we
+ don't collect the other expressions and skip step #2 in the algorithm. */
+
+struct nrv_data
+{
+ bitmap nrv;
+ tree result;
+ Node_Id gnat_ret;
+ hash_set<tree> *visited;
+};
+
+/* Return true if T is a Named Return Value. */
+
+static inline bool
+is_nrv_p (bitmap nrv, tree t)
+{
+ return TREE_CODE (t) == VAR_DECL && bitmap_bit_p (nrv, DECL_UID (t));
+}
+
+/* Helper function for walk_tree, used by finalize_nrv below. */
+
+static tree
+prune_nrv_r (tree *tp, int *walk_subtrees, void *data)
+{
+ struct nrv_data *dp = (struct nrv_data *)data;
+ tree t = *tp;
+
+ /* No need to walk into types or decls. */
+ if (IS_TYPE_OR_DECL_P (t))
+ *walk_subtrees = 0;
+
+ if (is_nrv_p (dp->nrv, t))
+ bitmap_clear_bit (dp->nrv, DECL_UID (t));
+
+ return NULL_TREE;
+}
+
+/* Prune Named Return Values in BLOCK and return true if there is still a
+ Named Return Value in BLOCK or one of its sub-blocks. */
+
+static bool
+prune_nrv_in_block (bitmap nrv, tree block)
+{
+ bool has_nrv = false;
+ tree t;
+
+ /* First recurse on the sub-blocks. */
+ for (t = BLOCK_SUBBLOCKS (block); t; t = BLOCK_CHAIN (t))
+ has_nrv |= prune_nrv_in_block (nrv, t);
+
+ /* Then make sure to keep at most one NRV per block. */
+ for (t = BLOCK_VARS (block); t; t = DECL_CHAIN (t))
+ if (is_nrv_p (nrv, t))
+ {
+ if (has_nrv)
+ bitmap_clear_bit (nrv, DECL_UID (t));
+ else
+ has_nrv = true;
+ }
+
+ return has_nrv;
+}
+
+/* Helper function for walk_tree, used by finalize_nrv below. */
+
+static tree
+finalize_nrv_r (tree *tp, int *walk_subtrees, void *data)
+{
+ struct nrv_data *dp = (struct nrv_data *)data;
+ tree t = *tp;
+
+ /* No need to walk into types. */
+ if (TYPE_P (t))
+ *walk_subtrees = 0;
+
+ /* Change RETURN_EXPRs of NRVs to just refer to the RESULT_DECL; this is a
+ nop, but differs from using NULL_TREE in that it indicates that we care
+ about the value of the RESULT_DECL. */
+ else if (TREE_CODE (t) == RETURN_EXPR
+ && TREE_CODE (TREE_OPERAND (t, 0)) == INIT_EXPR)
+ {
+ tree ret_val = TREE_OPERAND (TREE_OPERAND (t, 0), 1);
+
+ /* Strip useless conversions around the return value. */
+ if (gnat_useless_type_conversion (ret_val))
+ ret_val = TREE_OPERAND (ret_val, 0);
+
+ if (is_nrv_p (dp->nrv, ret_val))
+ TREE_OPERAND (t, 0) = dp->result;
+ }
+
+ /* Replace the DECL_EXPR of NRVs with an initialization of the RESULT_DECL,
+ if needed. */
+ else if (TREE_CODE (t) == DECL_EXPR
+ && is_nrv_p (dp->nrv, DECL_EXPR_DECL (t)))
+ {
+ tree var = DECL_EXPR_DECL (t), init;
+
+ if (DECL_INITIAL (var))
+ {
+ init = build_binary_op (INIT_EXPR, NULL_TREE, dp->result,
+ DECL_INITIAL (var));
+ SET_EXPR_LOCATION (init, EXPR_LOCATION (t));
+ DECL_INITIAL (var) = NULL_TREE;
+ }
+ else
+ init = build_empty_stmt (EXPR_LOCATION (t));
+ *tp = init;
+
+ /* Identify the NRV to the RESULT_DECL for debugging purposes. */
+ SET_DECL_VALUE_EXPR (var, dp->result);
+ DECL_HAS_VALUE_EXPR_P (var) = 1;
+ /* ??? Kludge to avoid an assertion failure during inlining. */
+ DECL_SIZE (var) = bitsize_unit_node;
+ DECL_SIZE_UNIT (var) = size_one_node;
+ }
+
+ /* And replace all uses of NRVs with the RESULT_DECL. */
+ else if (is_nrv_p (dp->nrv, t))
+ *tp = convert (TREE_TYPE (t), dp->result);
+
+ /* Avoid walking into the same tree more than once. Unfortunately, we
+ can't just use walk_tree_without_duplicates because it would only
+ call us for the first occurrence of NRVs in the function body. */
+ if (dp->visited->add (*tp))
+ *walk_subtrees = 0;
+
+ return NULL_TREE;
+}
+
+/* Likewise, but used when the function returns an unconstrained type. */
+
+static tree
+finalize_nrv_unc_r (tree *tp, int *walk_subtrees, void *data)
+{
+ struct nrv_data *dp = (struct nrv_data *)data;
+ tree t = *tp;
+
+ /* No need to walk into types. */
+ if (TYPE_P (t))
+ *walk_subtrees = 0;
+
+ /* We need to see the DECL_EXPR of NRVs before any other references so we
+ walk the body of BIND_EXPR before walking its variables. */
+ else if (TREE_CODE (t) == BIND_EXPR)
+ walk_tree (&BIND_EXPR_BODY (t), finalize_nrv_unc_r, data, NULL);
+
+ /* Change RETURN_EXPRs of NRVs to assign to the RESULT_DECL only the final
+ return value built by the allocator instead of the whole construct. */
+ else if (TREE_CODE (t) == RETURN_EXPR
+ && TREE_CODE (TREE_OPERAND (t, 0)) == INIT_EXPR)
+ {
+ tree ret_val = TREE_OPERAND (TREE_OPERAND (t, 0), 1);
+
+ /* This is the construct returned by the allocator. */
+ if (TREE_CODE (ret_val) == COMPOUND_EXPR
+ && TREE_CODE (TREE_OPERAND (ret_val, 0)) == INIT_EXPR)
+ {
+ tree rhs = TREE_OPERAND (TREE_OPERAND (ret_val, 0), 1);
+
+ if (TYPE_IS_FAT_POINTER_P (TREE_TYPE (ret_val)))
+ ret_val = CONSTRUCTOR_ELT (rhs, 1)->value;
+ else
+ ret_val = rhs;
+ }
+
+ /* Strip useless conversions around the return value. */
+ if (gnat_useless_type_conversion (ret_val)
+ || TREE_CODE (ret_val) == VIEW_CONVERT_EXPR)
+ ret_val = TREE_OPERAND (ret_val, 0);
+
+ /* Strip unpadding around the return value. */
+ if (TREE_CODE (ret_val) == COMPONENT_REF
+ && TYPE_IS_PADDING_P (TREE_TYPE (TREE_OPERAND (ret_val, 0))))
+ ret_val = TREE_OPERAND (ret_val, 0);
+
+ /* Assign the new return value to the RESULT_DECL. */
+ if (is_nrv_p (dp->nrv, ret_val))
+ TREE_OPERAND (TREE_OPERAND (t, 0), 1)
+ = TREE_OPERAND (DECL_INITIAL (ret_val), 0);
+ }
+
+ /* Adjust the DECL_EXPR of NRVs to call the allocator and save the result
+ into a new variable. */
+ else if (TREE_CODE (t) == DECL_EXPR
+ && is_nrv_p (dp->nrv, DECL_EXPR_DECL (t)))
+ {
+ tree saved_current_function_decl = current_function_decl;
+ tree var = DECL_EXPR_DECL (t);
+ tree alloc, p_array, new_var, new_ret;
+ vec<constructor_elt, va_gc> *v;
+ vec_alloc (v, 2);
+
+ /* Create an artificial context to build the allocation. */
+ current_function_decl = decl_function_context (var);
+ start_stmt_group ();
+ gnat_pushlevel ();
+
+ /* This will return a COMPOUND_EXPR with the allocation in the first
+ arm and the final return value in the second arm. */
+ alloc = build_allocator (TREE_TYPE (var), DECL_INITIAL (var),
+ TREE_TYPE (dp->result),
+ Procedure_To_Call (dp->gnat_ret),
+ Storage_Pool (dp->gnat_ret),
+ Empty, false);
+
+ /* The new variable is built as a reference to the allocated space. */
+ new_var
+ = build_decl (DECL_SOURCE_LOCATION (var), VAR_DECL, DECL_NAME (var),
+ build_reference_type (TREE_TYPE (var)));
+ DECL_BY_REFERENCE (new_var) = 1;
+
+ if (TYPE_IS_FAT_POINTER_P (TREE_TYPE (alloc)))
+ {
+ tree cst = TREE_OPERAND (alloc, 1);
+
+ /* The new initial value is a COMPOUND_EXPR with the allocation in
+ the first arm and the value of P_ARRAY in the second arm. */
+ DECL_INITIAL (new_var)
+ = build2 (COMPOUND_EXPR, TREE_TYPE (new_var),
+ TREE_OPERAND (alloc, 0),
+ CONSTRUCTOR_ELT (cst, 0)->value);
+
+ /* Build a modified CONSTRUCTOR that references NEW_VAR. */
+ p_array = TYPE_FIELDS (TREE_TYPE (alloc));
+ CONSTRUCTOR_APPEND_ELT (v, p_array,
+ fold_convert (TREE_TYPE (p_array), new_var));
+ CONSTRUCTOR_APPEND_ELT (v, DECL_CHAIN (p_array),
+ CONSTRUCTOR_ELT (cst, 1)->value);
+ new_ret = build_constructor (TREE_TYPE (alloc), v);
+ }
+ else
+ {
+ /* The new initial value is just the allocation. */
+ DECL_INITIAL (new_var) = alloc;
+ new_ret = fold_convert (TREE_TYPE (alloc), new_var);
+ }
+
+ gnat_pushdecl (new_var, Empty);
+
+ /* Destroy the artificial context and insert the new statements. */
+ gnat_zaplevel ();
+ *tp = end_stmt_group ();
+ current_function_decl = saved_current_function_decl;
+
+ /* Chain NEW_VAR immediately after VAR and ignore the latter. */
+ DECL_CHAIN (new_var) = DECL_CHAIN (var);
+ DECL_CHAIN (var) = new_var;
+ DECL_IGNORED_P (var) = 1;
+
+ /* Save the new return value and the dereference of NEW_VAR. */
+ DECL_INITIAL (var)
+ = build2 (COMPOUND_EXPR, TREE_TYPE (var), new_ret,
+ build1 (INDIRECT_REF, TREE_TYPE (var), new_var));
+ /* ??? Kludge to avoid messing up during inlining. */
+ DECL_CONTEXT (var) = NULL_TREE;
+ }
+
+ /* And replace all uses of NRVs with the dereference of NEW_VAR. */
+ else if (is_nrv_p (dp->nrv, t))
+ *tp = TREE_OPERAND (DECL_INITIAL (t), 1);
+
+ /* Avoid walking into the same tree more than once. Unfortunately, we
+ can't just use walk_tree_without_duplicates because it would only
+ call us for the first occurrence of NRVs in the function body. */
+ if (dp->visited->add (*tp))
+ *walk_subtrees = 0;
+
+ return NULL_TREE;
+}
+
+/* Apply FUNC to all the sub-trees of nested functions in NODE. FUNC is called
+ with the DATA and the address of each sub-tree. If FUNC returns a non-NULL
+ value, the traversal is stopped. */
+
+static void
+walk_nesting_tree (struct cgraph_node *node, walk_tree_fn func, void *data)
+{
+ for (node = first_nested_function (node);
+ node; node = next_nested_function (node))
+ {
+ walk_tree_without_duplicates (&DECL_SAVED_TREE (node->decl), func, data);
+ walk_nesting_tree (node, func, data);
+ }
+}
+
+/* Finalize the Named Return Value optimization for FNDECL. The NRV bitmap
+ contains the candidates for Named Return Value and OTHER is a list of
+ the other return values. GNAT_RET is a representative return node. */
+
+static void
+finalize_nrv (tree fndecl, bitmap nrv, vec<tree, va_gc> *other, Node_Id gnat_ret)
+{
+ struct nrv_data data;
+ walk_tree_fn func;
+ unsigned int i;
+ tree iter;
+
+ /* We shouldn't be applying the optimization to return types that we aren't
+ allowed to manipulate freely. */
+ gcc_assert (!TYPE_IS_BY_REFERENCE_P (TREE_TYPE (TREE_TYPE (fndecl))));
+
+ /* Prune the candidates that are referenced by other return values. */
+ data.nrv = nrv;
+ data.result = NULL_TREE;
+ data.gnat_ret = Empty;
+ data.visited = NULL;
+ FOR_EACH_VEC_SAFE_ELT (other, i, iter)
+ walk_tree_without_duplicates (&iter, prune_nrv_r, &data);
+ if (bitmap_empty_p (nrv))
+ return;
+
+ /* Prune also the candidates that are referenced by nested functions. */
+ walk_nesting_tree (cgraph_node::get_create (fndecl), prune_nrv_r, &data);
+ if (bitmap_empty_p (nrv))
+ return;
+
+ /* Extract a set of NRVs with non-overlapping live ranges. */
+ if (!prune_nrv_in_block (nrv, DECL_INITIAL (fndecl)))
+ return;
+
+ /* Adjust the relevant RETURN_EXPRs and replace the occurrences of NRVs. */
+ data.nrv = nrv;
+ data.result = DECL_RESULT (fndecl);
+ data.gnat_ret = gnat_ret;
+ data.visited = new hash_set<tree>;
+ if (TYPE_RETURN_UNCONSTRAINED_P (TREE_TYPE (fndecl)))
+ func = finalize_nrv_unc_r;
+ else
+ func = finalize_nrv_r;
+ walk_tree (&DECL_SAVED_TREE (fndecl), func, &data, NULL);
+ delete data.visited;
+}
+
+/* Return true if RET_VAL can be used as a Named Return Value for the
+ anonymous return object RET_OBJ. */
+
+static bool
+return_value_ok_for_nrv_p (tree ret_obj, tree ret_val)
+{
+ if (TREE_CODE (ret_val) != VAR_DECL)
+ return false;
+
+ if (TREE_THIS_VOLATILE (ret_val))
+ return false;
+
+ if (DECL_CONTEXT (ret_val) != current_function_decl)
+ return false;
+
+ if (TREE_STATIC (ret_val))
+ return false;
+
+ /* For the constrained case, test for addressability. */
+ if (ret_obj && TREE_ADDRESSABLE (ret_val))
+ return false;
+
+ /* For the constrained case, test for overalignment. */
+ if (ret_obj && DECL_ALIGN (ret_val) > DECL_ALIGN (ret_obj))
+ return false;
+
+ /* For the unconstrained case, test for bogus initialization. */
+ if (!ret_obj
+ && DECL_INITIAL (ret_val)
+ && TREE_CODE (DECL_INITIAL (ret_val)) == NULL_EXPR)
+ return false;
+
+ return true;
+}
+
+/* Build a RETURN_EXPR. If RET_VAL is non-null, build a RETURN_EXPR around
+ the assignment of RET_VAL to RET_OBJ. Otherwise build a bare RETURN_EXPR
+ around RESULT_OBJ, which may be null in this case. */
+
+static tree
+build_return_expr (tree ret_obj, tree ret_val)
+{
+ tree result_expr;
+
+ if (ret_val)
+ {
+ /* The gimplifier explicitly enforces the following invariant:
+
+ RETURN_EXPR
+ |
+ INIT_EXPR
+ / \
+ / \
+ RET_OBJ ...
+
+ As a consequence, type consistency dictates that we use the type
+ of the RET_OBJ as the operation type. */
+ tree operation_type = TREE_TYPE (ret_obj);
+
+ /* Convert the right operand to the operation type. Note that this is
+ the transformation applied in the INIT_EXPR case of build_binary_op,
+ with the assumption that the type cannot involve a placeholder. */
+ if (operation_type != TREE_TYPE (ret_val))
+ ret_val = convert (operation_type, ret_val);
+
+ /* We always can use an INIT_EXPR for the return object. */
+ result_expr = build2 (INIT_EXPR, void_type_node, ret_obj, ret_val);
+
+ /* If the function returns an aggregate type, find out whether this is
+ a candidate for Named Return Value. If so, record it. Otherwise,
+ if this is an expression of some kind, record it elsewhere. */
+ if (optimize
+ && !optimize_debug
+ && AGGREGATE_TYPE_P (operation_type)
+ && !TYPE_IS_FAT_POINTER_P (operation_type)
+ && TYPE_MODE (operation_type) == BLKmode
+ && aggregate_value_p (operation_type, current_function_decl))
+ {
+ /* Strip useless conversions around the return value. */
+ if (gnat_useless_type_conversion (ret_val))
+ ret_val = TREE_OPERAND (ret_val, 0);
+
+ /* Now apply the test to the return value. */
+ if (return_value_ok_for_nrv_p (ret_obj, ret_val))
+ {
+ if (!f_named_ret_val)
+ f_named_ret_val = BITMAP_GGC_ALLOC ();
+ bitmap_set_bit (f_named_ret_val, DECL_UID (ret_val));
+ }
+
+ /* Note that we need not care about CONSTRUCTORs here, as they are
+ totally transparent given the read-compose-write semantics of
+ assignments from CONSTRUCTORs. */
+ else if (EXPR_P (ret_val))
+ vec_safe_push (f_other_ret_val, ret_val);
+ }
+ }
+ else
+ result_expr = ret_obj;
+
+ return build1 (RETURN_EXPR, void_type_node, result_expr);
+}
+
+/* Subroutine of gnat_to_gnu to process gnat_node, an N_Subprogram_Body. We
+ don't return anything. */
+
+static void
+Subprogram_Body_to_gnu (Node_Id gnat_node)
+{
+ /* Defining identifier of a parameter to the subprogram. */
+ Entity_Id gnat_param;
+ /* The defining identifier for the subprogram body. Note that if a
+ specification has appeared before for this body, then the identifier
+ occurring in that specification will also be a defining identifier and all
+ the calls to this subprogram will point to that specification. */
+ Entity_Id gnat_subprog_id
+ = (Present (Corresponding_Spec (gnat_node))
+ ? Corresponding_Spec (gnat_node) : Defining_Entity (gnat_node));
+ /* The FUNCTION_DECL node corresponding to the subprogram spec. */
+ tree gnu_subprog_decl;
+ /* Its RESULT_DECL node. */
+ tree gnu_result_decl;
+ /* Its FUNCTION_TYPE node. */
+ tree gnu_subprog_type;
+ /* The TYPE_CI_CO_LIST of its FUNCTION_TYPE node, if any. */
+ tree gnu_cico_list;
+ /* The entry in the CI_CO_LIST that represents a function return, if any. */
+ tree gnu_return_var_elmt = NULL_TREE;
+ tree gnu_result;
+ location_t locus;
+ struct language_function *gnu_subprog_language;
+ vec<parm_attr, va_gc> *cache;
+
+ /* If this is a generic object or if it has been eliminated,
+ ignore it. */
+ if (Ekind (gnat_subprog_id) == E_Generic_Procedure
+ || Ekind (gnat_subprog_id) == E_Generic_Function
+ || Is_Eliminated (gnat_subprog_id))
+ return;
+
+ /* If this subprogram acts as its own spec, define it. Otherwise, just get
+ the already-elaborated tree node. However, if this subprogram had its
+ elaboration deferred, we will already have made a tree node for it. So
+ treat it as not being defined in that case. Such a subprogram cannot
+ have an address clause or a freeze node, so this test is safe, though it
+ does disable some otherwise-useful error checking. */
+ gnu_subprog_decl
+ = gnat_to_gnu_entity (gnat_subprog_id, NULL_TREE,
+ Acts_As_Spec (gnat_node)
+ && !present_gnu_tree (gnat_subprog_id));
+ DECL_FUNCTION_IS_DEF (gnu_subprog_decl) = true;
+ gnu_result_decl = DECL_RESULT (gnu_subprog_decl);
+ gnu_subprog_type = TREE_TYPE (gnu_subprog_decl);
+ gnu_cico_list = TYPE_CI_CO_LIST (gnu_subprog_type);
+ if (gnu_cico_list && TREE_VALUE (gnu_cico_list) == void_type_node)
+ gnu_return_var_elmt = gnu_cico_list;
+
+ /* If the function returns by invisible reference, make it explicit in the
+ function body. See gnat_to_gnu_entity, E_Subprogram_Type case. */
+ if (TREE_ADDRESSABLE (gnu_subprog_type))
+ {
+ TREE_TYPE (gnu_result_decl)
+ = build_reference_type (TREE_TYPE (gnu_result_decl));
+ relayout_decl (gnu_result_decl);
+ }
+
+ /* Set the line number in the decl to correspond to that of the body. */
+ if (DECL_IGNORED_P (gnu_subprog_decl))
+ locus = UNKNOWN_LOCATION;
+ else if (!Sloc_to_locus (Sloc (gnat_node), &locus, false, gnu_subprog_decl))
+ locus = input_location;
+ DECL_SOURCE_LOCATION (gnu_subprog_decl) = locus;
+
+ /* If the body comes from an expression function, arrange it to be inlined
+ in almost all cases. */
+ if (Was_Expression_Function (gnat_node) && !Debug_Flag_Dot_8)
+ DECL_DISREGARD_INLINE_LIMITS (gnu_subprog_decl) = 1;
+
+ /* Try to create a bona-fide thunk and hand it over to the middle-end. */
+ if (Is_Thunk (gnat_subprog_id)
+ && maybe_make_gnu_thunk (gnat_subprog_id, gnu_subprog_decl))
+ return;
+
+ /* Initialize the information structure for the function. */
+ allocate_struct_function (gnu_subprog_decl, false);
+ gnu_subprog_language = ggc_cleared_alloc<language_function> ();
+ DECL_STRUCT_FUNCTION (gnu_subprog_decl)->language = gnu_subprog_language;
+ DECL_STRUCT_FUNCTION (gnu_subprog_decl)->function_start_locus = locus;
+ set_cfun (NULL);
+
+ begin_subprog_body (gnu_subprog_decl);
+
+ /* If there are copy-in/copy-out parameters, we need to ensure that they are
+ properly copied out by the return statement. We do this by making a new
+ block and converting any return into a goto to a label at the end of the
+ block. */
+ if (gnu_cico_list)
+ {
+ tree gnu_return_var = NULL_TREE;
+
+ vec_safe_push (gnu_return_label_stack,
+ create_artificial_label (input_location));
+
+ start_stmt_group ();
+ gnat_pushlevel ();
+
+ /* If this is a function with copy-in/copy-out parameters and which does
+ not return by invisible reference, we also need a variable for the
+ return value to be placed. */
+ if (gnu_return_var_elmt && !TREE_ADDRESSABLE (gnu_subprog_type))
+ {
+ tree gnu_return_type
+ = TREE_TYPE (TREE_PURPOSE (gnu_return_var_elmt));
+
+ gnu_return_var
+ = create_var_decl (get_identifier ("RETVAL"), NULL_TREE,
+ gnu_return_type, NULL_TREE,
+ false, false, false, false, false,
+ true, false, NULL, gnat_subprog_id);
+ TREE_VALUE (gnu_return_var_elmt) = gnu_return_var;
+ }
+
+ vec_safe_push (gnu_return_var_stack, gnu_return_var);
+
+ /* See whether there are parameters for which we don't have a GCC tree
+ yet. These must be Out parameters. Make a VAR_DECL for them and
+ put it into TYPE_CI_CO_LIST, which must contain an empty entry too.
+ We can match up the entries because TYPE_CI_CO_LIST is in the order
+ of the parameters. */
+ for (gnat_param = First_Formal_With_Extras (gnat_subprog_id);
+ Present (gnat_param);
+ gnat_param = Next_Formal_With_Extras (gnat_param))
+ if (!present_gnu_tree (gnat_param))
+ {
+ tree gnu_cico_entry = gnu_cico_list;
+ tree gnu_decl;
+
+ /* Skip any entries that have been already filled in; they must
+ correspond to In Out parameters. */
+ while (gnu_cico_entry && TREE_VALUE (gnu_cico_entry))
+ gnu_cico_entry = TREE_CHAIN (gnu_cico_entry);
+
+ /* Do any needed dereferences for by-ref objects. */
+ gnu_decl = gnat_to_gnu_entity (gnat_param, NULL_TREE, true);
+ gcc_assert (DECL_P (gnu_decl));
+ if (DECL_BY_REF_P (gnu_decl))
+ gnu_decl = build_unary_op (INDIRECT_REF, NULL_TREE, gnu_decl);
+
+ /* Do any needed references for padded types. */
+ TREE_VALUE (gnu_cico_entry)
+ = convert (TREE_TYPE (TREE_PURPOSE (gnu_cico_entry)), gnu_decl);
+ }
+ }
+ else
+ vec_safe_push (gnu_return_label_stack, NULL_TREE);
+
+ /* Get a tree corresponding to the code for the subprogram. */
+ start_stmt_group ();
+ gnat_pushlevel ();
+
+ process_decls (Declarations (gnat_node), Empty, Empty, true, true);
+
+ /* Generate the code of the subprogram itself. A return statement will be
+ present and any Out parameters will be handled there. */
+ add_stmt (gnat_to_gnu (Handled_Statement_Sequence (gnat_node)));
+ gnat_poplevel ();
+ gnu_result = end_stmt_group ();
+
+ /* Attempt setting the end_locus of our GCC body tree, typically a BIND_EXPR,
+ then the end_locus of our GCC subprogram declaration tree. */
+ set_end_locus_from_node (gnu_result, gnat_node);
+ set_end_locus_from_node (gnu_subprog_decl, gnat_node);
+
+ /* If we populated the parameter attributes cache, we need to make sure that
+ the cached expressions are evaluated on all the possible paths leading to
+ their uses. So we force their evaluation on entry of the function. */
+ cache = gnu_subprog_language->parm_attr_cache;
+ if (cache)
+ {
+ struct parm_attr_d *pa;
+ int i;
+
+ start_stmt_group ();
+
+ FOR_EACH_VEC_ELT (*cache, i, pa)
+ {
+ if (pa->first)
+ add_stmt_with_node_force (pa->first, gnat_node);
+ if (pa->last)
+ add_stmt_with_node_force (pa->last, gnat_node);
+ if (pa->length)
+ add_stmt_with_node_force (pa->length, gnat_node);
+ }
+
+ add_stmt (gnu_result);
+ gnu_result = end_stmt_group ();
+
+ gnu_subprog_language->parm_attr_cache = NULL;
+ }
+
+ /* If we are dealing with a return from an Ada procedure with parameters
+ passed by copy-in/copy-out, we need to return a record containing the
+ final values of these parameters. If the list contains only one entry,
+ return just that entry though.
+
+ For a full description of the copy-in/copy-out parameter mechanism, see
+ the part of the gnat_to_gnu_entity routine dealing with the translation
+ of subprograms.
+
+ We need to make a block that contains the definition of that label and
+ the copying of the return value. It first contains the function, then
+ the label and copy statement. */
+ if (gnu_cico_list)
+ {
+ const Node_Id gnat_end_label
+ = End_Label (Handled_Statement_Sequence (gnat_node));
+
+ gnu_return_var_stack->pop ();
+
+ add_stmt (gnu_result);
+ add_stmt (build1 (LABEL_EXPR, void_type_node,
+ gnu_return_label_stack->last ()));
+
+ /* If this is a function which returns by invisible reference, the
+ return value has already been dealt with at the return statements,
+ so we only need to indirectly copy out the parameters. */
+ if (TREE_ADDRESSABLE (gnu_subprog_type))
+ {
+ tree gnu_ret_deref
+ = build_unary_op (INDIRECT_REF, NULL_TREE, gnu_result_decl);
+ tree t;
+
+ gcc_assert (TREE_VALUE (gnu_cico_list) == void_type_node);
+
+ for (t = TREE_CHAIN (gnu_cico_list); t; t = TREE_CHAIN (t))
+ {
+ tree gnu_field_deref
+ = build_component_ref (gnu_ret_deref, TREE_PURPOSE (t), true);
+ gnu_result = build2 (MODIFY_EXPR, void_type_node,
+ gnu_field_deref, TREE_VALUE (t));
+ add_stmt_with_node (gnu_result, gnat_end_label);
+ }
+ }
+
+ /* Otherwise, if this is a procedure or a function which does not return
+ by invisible reference, we can do a direct block-copy out. */
+ else
+ {
+ tree gnu_retval;
+
+ if (list_length (gnu_cico_list) == 1)
+ gnu_retval = TREE_VALUE (gnu_cico_list);
+ else
+ gnu_retval
+ = build_constructor_from_list (TREE_TYPE (gnu_subprog_type),
+ gnu_cico_list);
+
+ gnu_result = build_return_expr (gnu_result_decl, gnu_retval);
+ add_stmt_with_node (gnu_result, gnat_end_label);
+ }
+
+ gnat_poplevel ();
+ gnu_result = end_stmt_group ();
+ }
+
+ gnu_return_label_stack->pop ();
+
+ /* On SEH targets, install an exception handler around the main entry
+ point to catch unhandled exceptions. */
+ if (DECL_NAME (gnu_subprog_decl) == main_identifier_node
+ && targetm_common.except_unwind_info (&global_options) == UI_SEH)
+ {
+ tree t;
+ tree etype;
+
+ t = build_call_expr (builtin_decl_explicit (BUILT_IN_EH_POINTER),
+ 1, integer_zero_node);
+ t = build_call_n_expr (unhandled_except_decl, 1, t);
+
+ etype = build_unary_op (ADDR_EXPR, NULL_TREE, unhandled_others_decl);
+ etype = tree_cons (NULL_TREE, etype, NULL_TREE);
+
+ t = build2 (CATCH_EXPR, void_type_node, etype, t);
+ gnu_result = build2 (TRY_CATCH_EXPR, TREE_TYPE (gnu_result),
+ gnu_result, t);
+ }
+
+ end_subprog_body (gnu_result);
+
+ /* Finally annotate the parameters and disconnect the trees for parameters
+ that we have turned into variables since they are now unusable. */
+ for (gnat_param = First_Formal_With_Extras (gnat_subprog_id);
+ Present (gnat_param);
+ gnat_param = Next_Formal_With_Extras (gnat_param))
+ {
+ tree gnu_param = get_gnu_tree (gnat_param);
+ bool is_var_decl = (TREE_CODE (gnu_param) == VAR_DECL);
+
+ annotate_object (gnat_param, TREE_TYPE (gnu_param), NULL_TREE,
+ DECL_BY_REF_P (gnu_param));
+
+ if (is_var_decl)
+ save_gnu_tree (gnat_param, NULL_TREE, false);
+ }
+
+ /* Disconnect the variable created for the return value. */
+ if (gnu_return_var_elmt)
+ TREE_VALUE (gnu_return_var_elmt) = void_type_node;
+
+ /* If the function returns an aggregate type and we have candidates for
+ a Named Return Value, finalize the optimization. */
+ if (optimize && !optimize_debug && gnu_subprog_language->named_ret_val)
+ {
+ finalize_nrv (gnu_subprog_decl,
+ gnu_subprog_language->named_ret_val,
+ gnu_subprog_language->other_ret_val,
+ gnu_subprog_language->gnat_ret);
+ gnu_subprog_language->named_ret_val = NULL;
+ gnu_subprog_language->other_ret_val = NULL;
+ }
+
+ /* If this is an inlined external function that has been marked uninlinable,
+ drop the body and stop there. Otherwise compile the body. */
+ if (DECL_EXTERNAL (gnu_subprog_decl) && DECL_UNINLINABLE (gnu_subprog_decl))
+ DECL_SAVED_TREE (gnu_subprog_decl) = NULL_TREE;
+ else
+ rest_of_subprog_body_compilation (gnu_subprog_decl);
+}
+
+/* The type of an atomic access. */
+
+typedef enum { NOT_ATOMIC, SIMPLE_ATOMIC, OUTER_ATOMIC } atomic_acces_t;
+
+/* Return true if GNAT_NODE references an Atomic entity. This is modeled on
+ the Is_Atomic_Object predicate of the front-end, but additionally handles
+ explicit dereferences. */
+
+static bool
+node_is_atomic (Node_Id gnat_node)
+{
+ Entity_Id gnat_entity;
+
+ switch (Nkind (gnat_node))
+ {
+ case N_Identifier:
+ case N_Expanded_Name:
+ gnat_entity = Entity (gnat_node);
+ if (Ekind (gnat_entity) != E_Variable)
+ break;
+ return Is_Atomic (gnat_entity) || Is_Atomic (Etype (gnat_entity));
+
+ case N_Selected_Component:
+ return Is_Atomic (Etype (gnat_node))
+ || Is_Atomic (Entity (Selector_Name (gnat_node)));
+
+ case N_Indexed_Component:
+ return Is_Atomic (Etype (gnat_node))
+ || Has_Atomic_Components (Etype (Prefix (gnat_node)))
+ || (Is_Entity_Name (Prefix (gnat_node))
+ && Has_Atomic_Components (Entity (Prefix (gnat_node))));
+
+ case N_Explicit_Dereference:
+ return Is_Atomic (Etype (gnat_node));
+
+ default:
+ break;
+ }
+
+ return false;
+}
+
+/* Return true if GNAT_NODE references a Volatile_Full_Access entity. This is
+ modeled on the Is_Volatile_Full_Access_Object predicate of the front-end,
+ but additionally handles explicit dereferences. */
+
+static bool
+node_is_volatile_full_access (Node_Id gnat_node)
+{
+ Entity_Id gnat_entity;
+
+ switch (Nkind (gnat_node))
+ {
+ case N_Identifier:
+ case N_Expanded_Name:
+ gnat_entity = Entity (gnat_node);
+ if (!Is_Object (gnat_entity))
+ break;
+ return Is_Volatile_Full_Access (gnat_entity)
+ || Is_Volatile_Full_Access (Etype (gnat_entity));
+
+ case N_Selected_Component:
+ return Is_Volatile_Full_Access (Etype (gnat_node))
+ || Is_Volatile_Full_Access (Entity (Selector_Name (gnat_node)));
+
+ case N_Indexed_Component:
+ case N_Explicit_Dereference:
+ return Is_Volatile_Full_Access (Etype (gnat_node));
+
+ default:
+ break;
+ }
+
+ return false;
+}
+
+/* Return true if GNAT_NODE references a component of a larger object. */
+
+static inline bool
+node_is_component (Node_Id gnat_node)
+{
+ const Node_Kind k = Nkind (gnat_node);
+ return
+ (k == N_Indexed_Component || k == N_Selected_Component || k == N_Slice);
+}
+
+/* Compute whether GNAT_NODE requires atomic access and set TYPE to the type
+ of access and SYNC according to the associated synchronization setting.
+
+ We implement 3 different semantics of atomicity in this function:
+
+ 1. the Ada 95/2005/2012 semantics of the Atomic aspect/pragma,
+ 2. the Ada 2022 semantics of the Atomic aspect/pragma,
+ 3. the semantics of the Volatile_Full_Access GNAT aspect/pragma.
+
+ They are mutually exclusive and the FE should have rejected conflicts. */
+
+static void
+get_atomic_access (Node_Id gnat_node, atomic_acces_t *type, bool *sync)
+{
+ Node_Id gnat_parent, gnat_temp;
+ unsigned char attr_id;
+
+ /* First, scan the parent to filter out irrelevant cases. */
+ gnat_parent = Parent (gnat_node);
+ switch (Nkind (gnat_parent))
+ {
+ case N_Attribute_Reference:
+ attr_id = Get_Attribute_Id (Attribute_Name (gnat_parent));
+ /* Do not mess up machine code insertions. */
+ if (attr_id == Attr_Asm_Input || attr_id == Attr_Asm_Output)
+ goto not_atomic;
+
+ /* Nothing to do if we are the prefix of an attribute, since we do not
+ want an atomic access for things like 'Size. */
+
+ /* ... fall through ... */
+
+ case N_Reference:
+ /* The N_Reference node is like an attribute. */
+ if (Prefix (gnat_parent) == gnat_node)
+ goto not_atomic;
+ break;
+
+ case N_Object_Renaming_Declaration:
+ /* Nothing to do for the identifier in an object renaming declaration,
+ the renaming itself does not need atomic access. */
+ goto not_atomic;
+
+ default:
+ break;
+ }
+
+ /* Now strip any type conversion from GNAT_NODE. */
+ if (Nkind (gnat_node) == N_Type_Conversion
+ || Nkind (gnat_node) == N_Unchecked_Type_Conversion)
+ gnat_node = Expression (gnat_node);
+
+ /* Up to Ada 2012, for Atomic itself, only reads and updates of the object as
+ a whole require atomic access (RM C.6(15)). But, starting with Ada 2022,
+ reads of or writes to a nonatomic subcomponent of the object also require
+ atomic access (RM C.6(19)). */
+ if (node_is_atomic (gnat_node))
+ {
+ bool as_a_whole = true;
+
+ /* If we are the prefix of the parent, then the access is partial. */
+ for (gnat_temp = gnat_node, gnat_parent = Parent (gnat_temp);
+ node_is_component (gnat_parent) && Prefix (gnat_parent) == gnat_temp;
+ gnat_temp = gnat_parent, gnat_parent = Parent (gnat_temp))
+ if (Ada_Version < Ada_2022 || node_is_atomic (gnat_parent))
+ goto not_atomic;
+ else
+ as_a_whole = false;
+
+ /* We consider that partial accesses are not sequential actions and,
+ therefore, do not require synchronization. */
+ *type = SIMPLE_ATOMIC;
+ *sync = as_a_whole ? Atomic_Sync_Required (gnat_node) : false;
+ return;
+ }
+
+ /* Look for an outer atomic access of a nonatomic subcomponent. Note that,
+ for VFA, we do this before looking at the node itself because we need to
+ access the outermost VFA object atomically, unlike for Atomic where it is
+ the innermost atomic object (RM C.6(19)). */
+ for (gnat_temp = gnat_node;
+ node_is_component (gnat_temp);
+ gnat_temp = Prefix (gnat_temp))
+ if ((Ada_Version >= Ada_2022 && node_is_atomic (Prefix (gnat_temp)))
+ || node_is_volatile_full_access (Prefix (gnat_temp)))
+ {
+ *type = OUTER_ATOMIC;
+ *sync = false;
+ return;
+ }
+
+ /* Unlike Atomic, accessing a VFA object always requires atomic access. */
+ if (node_is_volatile_full_access (gnat_node))
+ {
+ *type = SIMPLE_ATOMIC;
+ *sync = false;
+ return;
+ }
+
+not_atomic:
+ *type = NOT_ATOMIC;
+ *sync = false;
+}
+
+/* Return true if GNAT_NODE requires simple atomic access and, if so, set SYNC
+ according to the associated synchronization setting. */
+
+static inline bool
+simple_atomic_access_required_p (Node_Id gnat_node, bool *sync)
+{
+ atomic_acces_t type;
+ get_atomic_access (gnat_node, &type, sync);
+ return type == SIMPLE_ATOMIC;
+}
+
+/* Create a temporary variable with PREFIX and TYPE, and return it. */
+
+static tree
+create_temporary (const char *prefix, tree type)
+{
+ tree gnu_temp
+ = create_var_decl (create_tmp_var_name (prefix), NULL_TREE,
+ type, NULL_TREE,
+ false, false, false, false, false,
+ true, false, NULL, Empty);
+ return gnu_temp;
+}
+
+/* Create a temporary variable with PREFIX and initialize it with GNU_INIT.
+ Put the initialization statement into GNU_INIT_STMT and annotate it with
+ the SLOC of GNAT_NODE. Return the temporary variable. */
+
+static tree
+create_init_temporary (const char *prefix, tree gnu_init, tree *gnu_init_stmt,
+ Node_Id gnat_node)
+{
+ tree gnu_temp = create_temporary (prefix, TREE_TYPE (gnu_init));
+
+ *gnu_init_stmt = build_binary_op (INIT_EXPR, NULL_TREE, gnu_temp, gnu_init);
+ set_expr_location_from_node (*gnu_init_stmt, gnat_node);
+
+ return gnu_temp;
+}
+
+/* Return true if TYPE is an array of scalar type. */
+
+static bool
+is_array_of_scalar_type (tree type)
+{
+ if (TREE_CODE (type) != ARRAY_TYPE)
+ return false;
+
+ type = TREE_TYPE (type);
+
+ return !AGGREGATE_TYPE_P (type) && !POINTER_TYPE_P (type);
+}
+
+/* Helper function for walk_tree, used by return_slot_opt_for_pure_call_p. */
+
+static tree
+find_decls_r (tree *tp, int *walk_subtrees, void *data)
+{
+ bitmap decls = (bitmap) data;
+
+ if (TYPE_P (*tp))
+ *walk_subtrees = 0;
+
+ else if (DECL_P (*tp))
+ bitmap_set_bit (decls, DECL_UID (*tp));
+
+ return NULL_TREE;
+}
+
+/* Return whether the assignment TARGET = CALL can be subject to the return
+ slot optimization, under the assumption that the called function be pure
+ in the Ada sense and return an array of scalar type. */
+
+static bool
+return_slot_opt_for_pure_call_p (tree target, tree call)
+{
+ /* Check that the target is a DECL. */
+ if (!DECL_P (target))
+ return false;
+
+ const bitmap decls = BITMAP_GGC_ALLOC ();
+ call_expr_arg_iterator iter;
+ tree arg;
+
+ /* Check that all the arguments have either a scalar type (we assume that
+ this means by-copy passing mechanism) or array of scalar type. */
+ FOR_EACH_CALL_EXPR_ARG (arg, iter, call)
+ {
+ tree arg_type = TREE_TYPE (arg);
+ if (TREE_CODE (arg_type) == REFERENCE_TYPE)
+ arg_type = TREE_TYPE (arg_type);
+
+ if (is_array_of_scalar_type (arg_type))
+ walk_tree_without_duplicates (&arg, find_decls_r, decls);
+
+ else if (AGGREGATE_TYPE_P (arg_type) || POINTER_TYPE_P (arg_type))
+ return false;
+ }
+
+ /* Check that the target is not referenced by the non-scalar arguments. */
+ return !bitmap_bit_p (decls, DECL_UID (target));
+}
+
+/* Elaborate types referenced in the profile (FIRST_FORMAL, RESULT_TYPE). */
+
+static void
+elaborate_profile (Entity_Id first_formal, Entity_Id result_type)
+{
+ Entity_Id formal;
+
+ for (formal = first_formal;
+ Present (formal);
+ formal = Next_Formal_With_Extras (formal))
+ (void) gnat_to_gnu_type (Etype (formal));
+
+ if (Present (result_type) && Ekind (result_type) != E_Void)
+ (void) gnat_to_gnu_type (result_type);
+}
+
+/* Subroutine of gnat_to_gnu to translate gnat_node, either an N_Function_Call
+ or an N_Procedure_Call_Statement, to a GCC tree, which is returned.
+ GNU_RESULT_TYPE_P is a pointer to where we should place the result type.
+ If GNU_TARGET is non-null, this must be a function call on the RHS of a
+ N_Assignment_Statement and the result is to be placed into that object.
+ ATOMIC_ACCESS is the type of atomic access to be used for the assignment
+ to GNU_TARGET. If, in addition, ATOMIC_SYNC is true, then the assignment
+ to GNU_TARGET requires atomic synchronization. */
+
+static tree
+Call_to_gnu (Node_Id gnat_node, tree *gnu_result_type_p, tree gnu_target,
+ atomic_acces_t atomic_access, bool atomic_sync)
+{
+ const bool function_call = (Nkind (gnat_node) == N_Function_Call);
+ const bool returning_value = (function_call && !gnu_target);
+ /* The GCC node corresponding to the GNAT subprogram name. This can either
+ be a FUNCTION_DECL node if we are dealing with a standard subprogram call,
+ or an indirect reference expression (an INDIRECT_REF node) pointing to a
+ subprogram. */
+ const Node_Id gnat_subprog = Name (gnat_node);
+ tree gnu_subprog = gnat_to_gnu (gnat_subprog);
+ /* The FUNCTION_TYPE node giving the GCC type of the subprogram. */
+ tree gnu_subprog_type = TREE_TYPE (gnu_subprog);
+ /* The return type of the FUNCTION_TYPE. */
+ tree gnu_result_type;;
+ const bool frontend_builtin
+ = (TREE_CODE (gnu_subprog) == FUNCTION_DECL
+ && DECL_BUILT_IN_CLASS (gnu_subprog) == BUILT_IN_FRONTEND);
+ auto_vec<tree, 16> gnu_actual_vec;
+ tree gnu_name_list = NULL_TREE;
+ tree gnu_stmt_list = NULL_TREE;
+ tree gnu_after_list = NULL_TREE;
+ tree gnu_retval = NULL_TREE;
+ tree gnu_call, gnu_result;
+ bool went_into_elab_proc;
+ bool pushed_binding_level;
+ bool variadic;
+ bool by_descriptor;
+ Entity_Id gnat_formal;
+ Entity_Id gnat_result_type;
+ Node_Id gnat_actual;
+ atomic_acces_t aa_type;
+ bool aa_sync;
+
+ /* The only way we can make a call via an access type is if GNAT_NAME is an
+ explicit dereference. In that case, get the list of formal args from the
+ type the access type is pointing to. Otherwise, get the formals from the
+ entity being called. */
+ if (Nkind (gnat_subprog) == N_Explicit_Dereference)
+ {
+ const Entity_Id gnat_prefix_type
+ = Underlying_Type (Etype (Prefix (gnat_subprog)));
+
+ gnat_formal = First_Formal_With_Extras (Etype (gnat_subprog));
+ gnat_result_type = Etype (Etype (gnat_subprog));
+ variadic = IN (Convention (gnat_prefix_type), Convention_C_Variadic);
+
+ /* If the access type doesn't require foreign-compatible representation,
+ be prepared for descriptors. */
+ by_descriptor
+ = targetm.calls.custom_function_descriptors > 0
+ && Can_Use_Internal_Rep (gnat_prefix_type);
+ }
+
+ else if (Nkind (gnat_subprog) == N_Attribute_Reference)
+ {
+ /* Assume here that this must be 'Elab_Body or 'Elab_Spec. */
+ gnat_formal = Empty;
+ gnat_result_type = Empty;
+ variadic = false;
+ by_descriptor = false;
+ }
+
+ else
+ {
+ gcc_checking_assert (Is_Entity_Name (gnat_subprog));
+
+ gnat_formal = First_Formal_With_Extras (Entity (gnat_subprog));
+ gnat_result_type = Etype (Entity_Id (gnat_subprog));
+ variadic = IN (Convention (Entity (gnat_subprog)), Convention_C_Variadic);
+ by_descriptor = false;
+
+ /* If we are calling a stubbed function, then raise Program_Error, but
+ elaborate all our args first. */
+ if (Convention (Entity (gnat_subprog)) == Convention_Stubbed)
+ {
+ tree call_expr = build_call_raise (PE_Stubbed_Subprogram_Called,
+ gnat_node, N_Raise_Program_Error);
+
+ for (gnat_actual = First_Actual (gnat_node);
+ Present (gnat_actual);
+ gnat_actual = Next_Actual (gnat_actual))
+ add_stmt (gnat_to_gnu (gnat_actual));
+
+ if (returning_value)
+ {
+ gnu_result_type = TREE_TYPE (gnu_subprog_type);
+ *gnu_result_type_p = gnu_result_type;
+ return build1 (NULL_EXPR, gnu_result_type, call_expr);
+ }
+
+ return call_expr;
+ }
+ }
+
+ /* We must elaborate the entire profile now because, if it references types
+ that were initially incomplete,, their elaboration changes the contents
+ of GNU_SUBPROG_TYPE and, in particular, may change the result type. */
+ elaborate_profile (gnat_formal, gnat_result_type);
+
+ gcc_assert (FUNC_OR_METHOD_TYPE_P (gnu_subprog_type));
+ gnu_result_type = TREE_TYPE (gnu_subprog_type);
+
+ if (TREE_CODE (gnu_subprog) == FUNCTION_DECL)
+ {
+ /* For a call to a nested function, check the inlining status. */
+ if (decl_function_context (gnu_subprog))
+ check_inlining_for_nested_subprog (gnu_subprog);
+
+ /* For a recursive call, avoid explosion due to recursive inlining. */
+ if (gnu_subprog == current_function_decl)
+ DECL_DISREGARD_INLINE_LIMITS (gnu_subprog) = 0;
+ }
+
+ /* The lifetime of the temporaries created for the call ends right after the
+ return value is copied, so we can give them the scope of the elaboration
+ routine at top level. */
+ if (!current_function_decl)
+ {
+ current_function_decl = get_elaboration_procedure ();
+ went_into_elab_proc = true;
+ }
+ else
+ went_into_elab_proc = false;
+
+ /* First, create the temporary for the return value when:
+
+ 1. There is no target and the function has copy-in/copy-out parameters,
+ because we need to preserve the return value before copying back the
+ parameters.
+
+ 2. There is no target and the call is made for neither the declaration
+ of an object (regular or renaming), nor a return statement, nor an
+ allocator, nor an aggregate, and the return type has variable size
+ because in this case the gimplifier cannot create the temporary, or
+ more generally is an aggregate type, because the gimplifier would
+ create the temporary in the outermost scope instead of locally here.
+ But there is an exception for an allocator of unconstrained record
+ type with default discriminant because we allocate the actual size
+ in this case, unlike in the other cases, so we need a temporary to
+ fetch the discriminant and we create it here.
+
+ 3. There is a target and it is a slice or an array with fixed size,
+ and the return type has variable size, because the gimplifier
+ doesn't handle these cases.
+
+ 4. There is a target which is a bit-field and the function returns an
+ unconstrained record type with default discriminant, because the
+ return may copy more data than the bit-field can contain.
+
+ 5. There is no target and we have misaligned In Out or Out parameters
+ passed by reference, because we need to preserve the return value
+ before copying back the parameters. However, in this case, we'll
+ defer creating the temporary, see below.
+
+ This must be done before we push a binding level around the call, since
+ we will pop it before copying the return value. */
+ if (function_call
+ && ((!gnu_target && TYPE_CI_CO_LIST (gnu_subprog_type))
+ || (!gnu_target
+ && Nkind (Parent (gnat_node)) != N_Object_Declaration
+ && Nkind (Parent (gnat_node)) != N_Object_Renaming_Declaration
+ && Nkind (Parent (gnat_node)) != N_Simple_Return_Statement
+ && (!(Nkind (Parent (gnat_node)) == N_Qualified_Expression
+ && Nkind (Parent (Parent (gnat_node))) == N_Allocator)
+ || type_is_padding_self_referential (gnu_result_type))
+ && Nkind (Parent (gnat_node)) != N_Aggregate
+ && AGGREGATE_TYPE_P (gnu_result_type)
+ && !TYPE_IS_FAT_POINTER_P (gnu_result_type))
+ || (gnu_target
+ && (TREE_CODE (gnu_target) == ARRAY_RANGE_REF
+ || (TREE_CODE (TREE_TYPE (gnu_target)) == ARRAY_TYPE
+ && TREE_CODE (TYPE_SIZE (TREE_TYPE (gnu_target)))
+ == INTEGER_CST))
+ && TREE_CODE (TYPE_SIZE (gnu_result_type)) != INTEGER_CST)
+ || (gnu_target
+ && TREE_CODE (gnu_target) == COMPONENT_REF
+ && DECL_BIT_FIELD (TREE_OPERAND (gnu_target, 1))
+ && DECL_SIZE (TREE_OPERAND (gnu_target, 1))
+ != TYPE_SIZE (TREE_TYPE (gnu_target))
+ && type_is_padding_self_referential (gnu_result_type))))
+ {
+ gnu_retval = create_temporary ("R", gnu_result_type);
+ DECL_RETURN_VALUE_P (gnu_retval) = 1;
+ }
+
+ /* If we don't need a value or have already created it, push a binding level
+ around the call. This will narrow the lifetime of the temporaries we may
+ need to make when translating the parameters as much as possible. */
+ if (!returning_value || gnu_retval)
+ {
+ start_stmt_group ();
+ gnat_pushlevel ();
+ pushed_binding_level = true;
+ }
+ else
+ pushed_binding_level = false;
+
+ /* Create the list of the actual parameters as GCC expects it, namely a
+ chain of TREE_LIST nodes in which the TREE_VALUE field of each node
+ is an expression and the TREE_PURPOSE field is null. But skip Out
+ parameters not passed by reference and that need not be copied in. */
+ for (gnat_actual = First_Actual (gnat_node);
+ Present (gnat_actual);
+ gnat_formal = Next_Formal_With_Extras (gnat_formal),
+ gnat_actual = Next_Actual (gnat_actual))
+ {
+ Entity_Id gnat_formal_type = Etype (gnat_formal);
+ tree gnu_formal_type = gnat_to_gnu_type (gnat_formal_type);
+ tree gnu_formal = present_gnu_tree (gnat_formal)
+ ? get_gnu_tree (gnat_formal) : NULL_TREE;
+ const bool in_param = (Ekind (gnat_formal) == E_In_Parameter);
+ const bool is_true_formal_parm
+ = gnu_formal && TREE_CODE (gnu_formal) == PARM_DECL;
+ const bool is_by_ref_formal_parm
+ = is_true_formal_parm
+ && (DECL_BY_REF_P (gnu_formal)
+ || DECL_BY_COMPONENT_PTR_P (gnu_formal));
+ /* In the In Out or Out case, we must suppress conversions that yield
+ an lvalue but can nevertheless cause the creation of a temporary,
+ because we need the real object in this case, either to pass its
+ address if it's passed by reference or as target of the back copy
+ done after the call if it uses the copy-in/copy-out mechanism.
+ We do it in the In case too, except for an unchecked conversion
+ to an elementary type or a constrained composite type because it
+ alone can cause the actual to be misaligned and the addressability
+ test is applied to the real object. */
+ const bool suppress_type_conversion
+ = ((Nkind (gnat_actual) == N_Unchecked_Type_Conversion
+ && (!in_param
+ || !is_by_ref_formal_parm
+ || (Is_Composite_Type (Underlying_Type (gnat_formal_type))
+ && !Is_Constrained (Underlying_Type (gnat_formal_type)))))
+ || (Nkind (gnat_actual) == N_Type_Conversion
+ && Is_Composite_Type (Underlying_Type (gnat_formal_type))));
+ Node_Id gnat_name = suppress_type_conversion
+ ? Expression (gnat_actual) : gnat_actual;
+ tree gnu_name = gnat_to_gnu (gnat_name), gnu_name_type;
+
+ /* If it's possible we may need to use this expression twice, make sure
+ that any side-effects are handled via SAVE_EXPRs; likewise if we need
+ to force side-effects before the call. */
+ if (!in_param && !is_by_ref_formal_parm)
+ {
+ tree init = NULL_TREE;
+ gnu_name = gnat_stabilize_reference (gnu_name, true, &init);
+ if (init)
+ gnu_name
+ = build_compound_expr (TREE_TYPE (gnu_name), init, gnu_name);
+ }
+
+ /* If we are passing a non-addressable parameter by reference, pass the
+ address of a copy. In the In Out or Out case, set up to copy back
+ out after the call. */
+ if (is_by_ref_formal_parm
+ && (gnu_name_type = gnat_to_gnu_type (Etype (gnat_name)))
+ && !addressable_p (gnu_name, gnu_name_type))
+ {
+ tree gnu_orig = gnu_name, gnu_temp, gnu_stmt;
+
+ /* Do not issue warnings for CONSTRUCTORs since this is not a copy
+ but sort of an instantiation for them. */
+ if (TREE_CODE (remove_conversions (gnu_name, true)) == CONSTRUCTOR)
+ ;
+
+ /* If the formal is passed by reference, a copy is not allowed. */
+ else if (TYPE_IS_BY_REFERENCE_P (gnu_formal_type)
+ || Is_Aliased (gnat_formal))
+ post_error ("misaligned actual cannot be passed by reference",
+ gnat_actual);
+
+ /* If the mechanism was forced to by-ref, a copy is not allowed but
+ we issue only a warning because this case is not strict Ada. */
+ else if (DECL_FORCED_BY_REF_P (gnu_formal))
+ post_error ("misaligned actual cannot be passed by reference??",
+ gnat_actual);
+
+ /* If the actual type of the object is already the nominal type,
+ we have nothing to do, except if the size is self-referential
+ in which case we'll remove the unpadding below. */
+ if (TREE_TYPE (gnu_name) == gnu_name_type
+ && !CONTAINS_PLACEHOLDER_P (TYPE_SIZE (gnu_name_type)))
+ ;
+
+ /* Otherwise remove the unpadding from all the objects. */
+ else if (TREE_CODE (gnu_name) == COMPONENT_REF
+ && TYPE_IS_PADDING_P
+ (TREE_TYPE (TREE_OPERAND (gnu_name, 0))))
+ gnu_orig = gnu_name = TREE_OPERAND (gnu_name, 0);
+
+ /* Otherwise convert to the nominal type of the object if needed.
+ There are several cases in which we need to make the temporary
+ using this type instead of the actual type of the object when
+ they are distinct, because the expectations of the callee would
+ otherwise not be met:
+ - if it's a justified modular type,
+ - if the actual type is a smaller form of it,
+ - if it's a smaller form of the actual type. */
+ else if ((TREE_CODE (gnu_name_type) == RECORD_TYPE
+ && (TYPE_JUSTIFIED_MODULAR_P (gnu_name_type)
+ || smaller_form_type_p (TREE_TYPE (gnu_name),
+ gnu_name_type)))
+ || (INTEGRAL_TYPE_P (gnu_name_type)
+ && smaller_form_type_p (gnu_name_type,
+ TREE_TYPE (gnu_name))))
+ gnu_name = convert (gnu_name_type, gnu_name);
+
+ /* If this is an In Out or Out parameter and we're returning a value,
+ we need to create a temporary for the return value because we must
+ preserve it before copying back at the very end. */
+ if (!in_param && returning_value && !gnu_retval)
+ {
+ gnu_retval = create_temporary ("R", gnu_result_type);
+ DECL_RETURN_VALUE_P (gnu_retval) = 1;
+ }
+
+ /* If we haven't pushed a binding level, push it now. This will
+ narrow the lifetime of the temporary we are about to make as
+ much as possible. */
+ if (!pushed_binding_level && (!returning_value || gnu_retval))
+ {
+ start_stmt_group ();
+ gnat_pushlevel ();
+ pushed_binding_level = true;
+ }
+
+ /* Create an explicit temporary holding the copy. */
+ /* Do not initialize it for the _Init parameter of an initialization
+ procedure since no data is meant to be passed in. */
+ if (Ekind (gnat_formal) == E_Out_Parameter
+ && Is_Entity_Name (gnat_subprog)
+ && Is_Init_Proc (Entity (gnat_subprog)))
+ gnu_name = gnu_temp = create_temporary ("A", TREE_TYPE (gnu_name));
+
+ /* Initialize it on the fly like for an implicit temporary in the
+ other cases, as we don't necessarily have a statement list. */
+ else
+ {
+ gnu_temp = create_init_temporary ("A", gnu_name, &gnu_stmt,
+ gnat_actual);
+ gnu_name = build_compound_expr (TREE_TYPE (gnu_name), gnu_stmt,
+ gnu_temp);
+ }
+
+ /* Set up to move the copy back to the original if needed. */
+ if (!in_param)
+ {
+ /* If the original is a COND_EXPR whose first arm isn't meant to
+ be further used, just deal with the second arm. This is very
+ likely the conditional expression built for a check. */
+ if (TREE_CODE (gnu_orig) == COND_EXPR
+ && TREE_CODE (TREE_OPERAND (gnu_orig, 1)) == COMPOUND_EXPR
+ && integer_zerop
+ (TREE_OPERAND (TREE_OPERAND (gnu_orig, 1), 1)))
+ gnu_orig = TREE_OPERAND (gnu_orig, 2);
+
+ gnu_stmt
+ = build_binary_op (MODIFY_EXPR, NULL_TREE, gnu_orig, gnu_temp);
+ set_expr_location_from_node (gnu_stmt, gnat_node);
+
+ append_to_statement_list (gnu_stmt, &gnu_after_list);
+ }
+ }
+
+ /* Start from the real object and build the actual. */
+ tree gnu_actual = gnu_name;
+
+ /* If atomic access is required for an In or In Out actual parameter,
+ build the atomic load. */
+ if (is_true_formal_parm
+ && !is_by_ref_formal_parm
+ && Ekind (gnat_formal) != E_Out_Parameter
+ && simple_atomic_access_required_p (gnat_actual, &aa_sync))
+ gnu_actual = build_atomic_load (gnu_actual, aa_sync);
+
+ /* If this was a procedure call, we may not have removed any padding.
+ So do it here for the part we will use as an input, if any. */
+ if (Ekind (gnat_formal) != E_Out_Parameter
+ && TYPE_IS_PADDING_P (TREE_TYPE (gnu_actual)))
+ gnu_actual
+ = convert (get_unpadded_type (Etype (gnat_actual)), gnu_actual);
+
+ /* Put back the conversion we suppressed above in the computation of the
+ real object. And even if we didn't suppress any conversion there, we
+ may have suppressed a conversion to the Etype of the actual earlier,
+ since the parent is a procedure call, so put it back here. Note that
+ we might have a dummy type here if the actual is the dereference of a
+ pointer to it, but that's OK when the formal is passed by reference.
+ We also do not put back a conversion between an actual and a formal
+ that are unconstrained array types to avoid creating local bounds. */
+ tree gnu_actual_type = get_unpadded_type (Etype (gnat_actual));
+ if (TYPE_IS_DUMMY_P (gnu_actual_type))
+ gcc_assert (is_true_formal_parm && DECL_BY_REF_P (gnu_formal));
+ else if (suppress_type_conversion
+ && Nkind (gnat_actual) == N_Unchecked_Type_Conversion)
+ gnu_actual = unchecked_convert (gnu_actual_type, gnu_actual,
+ No_Truncation (gnat_actual));
+ else if ((TREE_CODE (TREE_TYPE (gnu_actual)) == UNCONSTRAINED_ARRAY_TYPE
+ || (TREE_CODE (TREE_TYPE (gnu_actual)) == RECORD_TYPE
+ && TYPE_CONTAINS_TEMPLATE_P (TREE_TYPE (gnu_actual))))
+ && TREE_CODE (gnu_formal_type) == UNCONSTRAINED_ARRAY_TYPE)
+ ;
+ else
+ gnu_actual = convert (gnu_actual_type, gnu_actual);
+
+ gigi_checking_assert (!Do_Range_Check (gnat_actual));
+
+ /* First see if the parameter is passed by reference. */
+ if (is_true_formal_parm && DECL_BY_REF_P (gnu_formal))
+ {
+ if (!in_param)
+ {
+ /* In Out or Out parameters passed by reference don't use the
+ copy-in/copy-out mechanism so the address of the real object
+ must be passed to the function. */
+ gnu_actual = gnu_name;
+
+ /* If we have a padded type, be sure we've removed padding. */
+ if (TYPE_IS_PADDING_P (TREE_TYPE (gnu_actual)))
+ gnu_actual = convert (get_unpadded_type (Etype (gnat_actual)),
+ gnu_actual);
+
+ /* If we have the constructed subtype of an aliased object
+ with an unconstrained nominal subtype, the type of the
+ actual includes the template, although it is formally
+ constrained. So we need to convert it back to the real
+ constructed subtype to retrieve the constrained part
+ and takes its address. */
+ if (TREE_CODE (TREE_TYPE (gnu_actual)) == RECORD_TYPE
+ && TYPE_CONTAINS_TEMPLATE_P (TREE_TYPE (gnu_actual))
+ && Is_Constr_Subt_For_UN_Aliased (Etype (gnat_actual))
+ && Is_Array_Type (Underlying_Type (Etype (gnat_actual))))
+ gnu_actual = convert (gnu_actual_type, gnu_actual);
+ }
+
+ /* There is no need to convert the actual to the formal's type before
+ taking its address. The only exception is for unconstrained array
+ types because of the way we build fat pointers. */
+ if (TREE_CODE (gnu_formal_type) == UNCONSTRAINED_ARRAY_TYPE)
+ {
+ /* Put back the conversion we suppressed above for In Out or Out
+ parameters, since it may set the bounds of the actual. */
+ if (!in_param && suppress_type_conversion)
+ gnu_actual = convert (gnu_actual_type, gnu_actual);
+ gnu_actual = convert (gnu_formal_type, gnu_actual);
+ }
+
+ /* Take the address of the object and convert to the proper pointer
+ type. */
+ gnu_formal_type = TREE_TYPE (gnu_formal);
+ gnu_actual = build_unary_op (ADDR_EXPR, gnu_formal_type, gnu_actual);
+ }
+
+ /* Then see if the parameter is an array passed to a foreign convention
+ subprogram. */
+ else if (is_true_formal_parm && DECL_BY_COMPONENT_PTR_P (gnu_formal))
+ {
+ gnu_actual = maybe_padded_object (gnu_actual);
+ gnu_actual = maybe_unconstrained_array (gnu_actual);
+
+ /* Take the address of the object and convert to the proper pointer
+ type. We'd like to actually compute the address of the beginning
+ of the array using an ADDR_EXPR of an ARRAY_REF, but there's a
+ possibility that the ARRAY_REF might return a constant and we'd be
+ getting the wrong address. Neither approach is exactly correct,
+ but this is the most likely to work in all cases. */
+ gnu_formal_type = TREE_TYPE (gnu_formal);
+ gnu_actual = build_unary_op (ADDR_EXPR, gnu_formal_type, gnu_actual);
+ }
+
+ /* Then see if the parameter is passed by copy. */
+ else if (is_true_formal_parm)
+ {
+ if (!in_param)
+ gnu_name_list = tree_cons (NULL_TREE, gnu_name, gnu_name_list);
+
+ gnu_actual = convert (gnu_formal_type, gnu_actual);
+
+ /* If this is a front-end built-in function, there is no need to
+ convert to the type used to pass the argument. */
+ if (!frontend_builtin)
+ gnu_actual = convert (DECL_ARG_TYPE (gnu_formal), gnu_actual);
+ }
+
+ /* Then see if this is an unnamed parameter in a variadic C function. */
+ else if (variadic)
+ {
+ /* This is based on the processing done in gnat_to_gnu_param, but
+ we expect the mechanism to be set in (almost) all cases. */
+ const Mechanism_Type mech = Mechanism (gnat_formal);
+
+ /* Strip off possible padding type. */
+ if (TYPE_IS_PADDING_P (gnu_formal_type))
+ gnu_formal_type = TREE_TYPE (TYPE_FIELDS (gnu_formal_type));
+
+ /* Arrays are passed as pointers to element type. First check for
+ unconstrained array and get the underlying array. */
+ if (TREE_CODE (gnu_formal_type) == UNCONSTRAINED_ARRAY_TYPE)
+ gnu_formal_type
+ = TREE_TYPE
+ (TREE_TYPE (TYPE_FIELDS (TREE_TYPE (gnu_formal_type))));
+
+ /* Arrays are passed as pointers to element type. */
+ if (mech != By_Copy && TREE_CODE (gnu_formal_type) == ARRAY_TYPE)
+ {
+ gnu_actual = maybe_padded_object (gnu_actual);
+ gnu_actual = maybe_unconstrained_array (gnu_actual);
+
+ /* Strip off any multi-dimensional entries, then strip
+ off the last array to get the component type. */
+ while (TREE_CODE (TREE_TYPE (gnu_formal_type)) == ARRAY_TYPE
+ && TYPE_MULTI_ARRAY_P (TREE_TYPE (gnu_formal_type)))
+ gnu_formal_type = TREE_TYPE (gnu_formal_type);
+
+ gnu_formal_type = TREE_TYPE (gnu_formal_type);
+ gnu_formal_type = build_pointer_type (gnu_formal_type);
+ gnu_actual
+ = build_unary_op (ADDR_EXPR, gnu_formal_type, gnu_actual);
+ }
+
+ /* Fat pointers are passed as thin pointers. */
+ else if (TYPE_IS_FAT_POINTER_P (gnu_formal_type))
+ gnu_formal_type
+ = make_type_from_size (gnu_formal_type,
+ size_int (POINTER_SIZE), 0);
+
+ /* If we were requested or muss pass by reference, do so.
+ If we were requested to pass by copy, do so.
+ Otherwise, pass In Out or Out parameters or aggregates by
+ reference. */
+ else if (mech == By_Reference
+ || must_pass_by_ref (gnu_formal_type)
+ || (mech != By_Copy
+ && (!in_param || AGGREGATE_TYPE_P (gnu_formal_type))))
+ {
+ gnu_formal_type = build_reference_type (gnu_formal_type);
+ gnu_actual
+ = build_unary_op (ADDR_EXPR, gnu_formal_type, gnu_actual);
+ }
+
+ /* Otherwise pass by copy after applying default C promotions. */
+ else
+ {
+ if (INTEGRAL_TYPE_P (gnu_formal_type)
+ && TYPE_PRECISION (gnu_formal_type)
+ < TYPE_PRECISION (integer_type_node))
+ gnu_formal_type = integer_type_node;
+
+ else if (SCALAR_FLOAT_TYPE_P (gnu_formal_type)
+ && TYPE_PRECISION (gnu_formal_type)
+ < TYPE_PRECISION (double_type_node))
+ gnu_formal_type = double_type_node;
+ }
+
+ gnu_actual = convert (gnu_formal_type, gnu_actual);
+ }
+
+ /* If we didn't create a PARM_DECL for the formal, this means that
+ it is an Out parameter not passed by reference and that need not
+ be copied in. In this case, the value of the actual need not be
+ read. However, we still need to make sure that its side-effects
+ are evaluated before the call, so we evaluate its address. */
+ else
+ {
+ if (!in_param)
+ gnu_name_list = tree_cons (NULL_TREE, gnu_name, gnu_name_list);
+
+ if (TREE_SIDE_EFFECTS (gnu_name))
+ {
+ tree addr = build_unary_op (ADDR_EXPR, NULL_TREE, gnu_name);
+ append_to_statement_list (addr, &gnu_stmt_list);
+ }
+
+ continue;
+ }
+
+ gnu_actual_vec.safe_push (gnu_actual);
+ }
+
+ if (frontend_builtin)
+ {
+ tree pred_cst = build_int_cst (integer_type_node, PRED_BUILTIN_EXPECT);
+ enum internal_fn icode = IFN_BUILTIN_EXPECT;
+
+ switch (DECL_FE_FUNCTION_CODE (gnu_subprog))
+ {
+ case BUILT_IN_EXPECT:
+ break;
+ case BUILT_IN_LIKELY:
+ gnu_actual_vec.safe_push (boolean_true_node);
+ break;
+ case BUILT_IN_UNLIKELY:
+ gnu_actual_vec.safe_push (boolean_false_node);
+ break;
+ default:
+ gcc_unreachable ();
+ }
+
+ gnu_actual_vec.safe_push (pred_cst);
+
+ gnu_call
+ = build_call_expr_internal_loc_array (UNKNOWN_LOCATION,
+ icode,
+ gnu_result_type,
+ gnu_actual_vec.length (),
+ gnu_actual_vec.begin ());
+ }
+ else
+ {
+ gnu_call
+ = build_call_array_loc (UNKNOWN_LOCATION,
+ gnu_result_type,
+ build_unary_op (ADDR_EXPR, NULL_TREE,
+ gnu_subprog),
+ gnu_actual_vec.length (),
+ gnu_actual_vec.begin ());
+ CALL_EXPR_BY_DESCRIPTOR (gnu_call) = by_descriptor;
+ }
+
+ set_expr_location_from_node (gnu_call, gnat_node);
+
+ /* If we have created a temporary for the return value, initialize it. */
+ if (gnu_retval)
+ {
+ tree gnu_stmt
+ = build_binary_op (INIT_EXPR, NULL_TREE, gnu_retval, gnu_call);
+ set_expr_location_from_node (gnu_stmt, gnat_node);
+ append_to_statement_list (gnu_stmt, &gnu_stmt_list);
+ gnu_call = gnu_retval;
+ }
+
+ /* If this is a subprogram with copy-in/copy-out parameters, we need to
+ unpack the valued returned from the function into the In Out or Out
+ parameters. We deal with the function return (if this is an Ada
+ function) below. */
+ if (TYPE_CI_CO_LIST (gnu_subprog_type))
+ {
+ /* List of FIELD_DECLs associated with the PARM_DECLs of the copy-in/
+ copy-out parameters. */
+ tree gnu_cico_list = TYPE_CI_CO_LIST (gnu_subprog_type);
+ const int length = list_length (gnu_cico_list);
+
+ /* The call sequence must contain one and only one call, even though the
+ function is pure. Save the result into a temporary if needed. */
+ if (length > 1)
+ {
+ if (!gnu_retval)
+ {
+ tree gnu_stmt;
+ gnu_call
+ = create_init_temporary ("P", gnu_call, &gnu_stmt, gnat_node);
+ append_to_statement_list (gnu_stmt, &gnu_stmt_list);
+ }
+
+ gnu_name_list = nreverse (gnu_name_list);
+ }
+
+ /* The first entry is for the actual return value if this is a
+ function, so skip it. */
+ if (function_call)
+ gnu_cico_list = TREE_CHAIN (gnu_cico_list);
+
+ if (Nkind (gnat_subprog) == N_Explicit_Dereference)
+ gnat_formal = First_Formal_With_Extras (Etype (gnat_subprog));
+ else
+ gnat_formal = First_Formal_With_Extras (Entity (gnat_subprog));
+
+ for (gnat_actual = First_Actual (gnat_node);
+ Present (gnat_actual);
+ gnat_formal = Next_Formal_With_Extras (gnat_formal),
+ gnat_actual = Next_Actual (gnat_actual))
+ /* If we are dealing with a copy-in/copy-out parameter, we must
+ retrieve its value from the record returned in the call. */
+ if (!(present_gnu_tree (gnat_formal)
+ && TREE_CODE (get_gnu_tree (gnat_formal)) == PARM_DECL
+ && (DECL_BY_REF_P (get_gnu_tree (gnat_formal))
+ || DECL_BY_COMPONENT_PTR_P (get_gnu_tree (gnat_formal))))
+ && Ekind (gnat_formal) != E_In_Parameter)
+ {
+ /* Get the value to assign to this In Out or Out parameter. It is
+ either the result of the function if there is only a single such
+ parameter or the appropriate field from the record returned. */
+ tree gnu_result
+ = length == 1
+ ? gnu_call
+ : build_component_ref (gnu_call, TREE_PURPOSE (gnu_cico_list),
+ false);
+
+ /* If the actual is a conversion, get the inner expression, which
+ will be the real destination, and convert the result to the
+ type of the actual parameter. */
+ tree gnu_actual
+ = maybe_unconstrained_array (TREE_VALUE (gnu_name_list));
+
+ /* If the result is padded, remove the padding. */
+ gnu_result = maybe_padded_object (gnu_result);
+
+ /* If the actual is a type conversion, the real target object is
+ denoted by the inner Expression and we need to convert the
+ result to the associated type.
+ We also need to convert our gnu assignment target to this type
+ if the corresponding GNU_NAME was constructed from the GNAT
+ conversion node and not from the inner Expression. */
+ if (Nkind (gnat_actual) == N_Type_Conversion)
+ {
+ const Node_Id gnat_expr = Expression (gnat_actual);
+
+ gigi_checking_assert (!Do_Range_Check (gnat_expr));
+
+ gnu_result
+ = convert_with_check (Etype (gnat_expr), gnu_result,
+ Do_Overflow_Check (gnat_actual),
+ Float_Truncate (gnat_actual),
+ gnat_actual);
+
+ if (!Is_Composite_Type (Underlying_Type (Etype (gnat_formal))))
+ gnu_actual = convert (TREE_TYPE (gnu_result), gnu_actual);
+ }
+
+ /* Unchecked conversions as actuals for Out parameters are not
+ allowed in user code because they are not variables, but do
+ occur in front-end expansions. The associated GNU_NAME is
+ always obtained from the inner expression in such cases. */
+ else if (Nkind (gnat_actual) == N_Unchecked_Type_Conversion)
+ gnu_result = unchecked_convert (TREE_TYPE (gnu_actual),
+ gnu_result,
+ No_Truncation (gnat_actual));
+ else
+ {
+ gigi_checking_assert (!Do_Range_Check (gnat_actual));
+
+ if (!(!TREE_CONSTANT (TYPE_SIZE (TREE_TYPE (gnu_actual)))
+ && TREE_CONSTANT (TYPE_SIZE (TREE_TYPE (gnu_result)))))
+ gnu_result = convert (TREE_TYPE (gnu_actual), gnu_result);
+ }
+
+ get_atomic_access (gnat_actual, &aa_type, &aa_sync);
+
+ /* If an outer atomic access is required for an actual parameter,
+ build the load-modify-store sequence. */
+ if (aa_type == OUTER_ATOMIC)
+ gnu_result
+ = build_load_modify_store (gnu_actual, gnu_result, gnat_node);
+
+ /* Or else, if a simple atomic access is required, build the atomic
+ store. */
+ else if (aa_type == SIMPLE_ATOMIC)
+ gnu_result
+ = build_atomic_store (gnu_actual, gnu_result, aa_sync);
+
+ /* Otherwise build a regular assignment. */
+ else
+ gnu_result = build_binary_op (MODIFY_EXPR, NULL_TREE,
+ gnu_actual, gnu_result);
+
+ if (EXPR_P (gnu_result))
+ set_expr_location_from_node (gnu_result, gnat_node);
+ append_to_statement_list (gnu_result, &gnu_stmt_list);
+ gnu_cico_list = TREE_CHAIN (gnu_cico_list);
+ gnu_name_list = TREE_CHAIN (gnu_name_list);
+ }
+ }
+
+ /* If this is a function call, the result is the call expression unless a
+ target is specified, in which case we copy the result into the target
+ and return the assignment statement. */
+ if (function_call)
+ {
+ /* If this is a function with copy-in/copy-out parameters, extract the
+ return value from it and update the return type. */
+ if (TYPE_CI_CO_LIST (gnu_subprog_type))
+ {
+ tree gnu_elmt = TYPE_CI_CO_LIST (gnu_subprog_type);
+ gnu_call
+ = build_component_ref (gnu_call, TREE_PURPOSE (gnu_elmt), false);
+ gnu_result_type = TREE_TYPE (gnu_call);
+ }
+
+ /* If the function returns an unconstrained array or by direct reference,
+ we have to dereference the pointer. */
+ if (TYPE_RETURN_UNCONSTRAINED_P (gnu_subprog_type)
+ || TYPE_RETURN_BY_DIRECT_REF_P (gnu_subprog_type))
+ gnu_call = build_unary_op (INDIRECT_REF, NULL_TREE, gnu_call);
+
+ if (gnu_target)
+ {
+ Node_Id gnat_parent = Parent (gnat_node);
+ enum tree_code op_code;
+
+ gigi_checking_assert (!Do_Range_Check (gnat_node));
+
+ /* ??? If the return type has variable size, then force the return
+ slot optimization as we would not be able to create a temporary.
+ That's what has been done historically. */
+ if (return_type_with_variable_size_p (gnu_result_type))
+ op_code = INIT_EXPR;
+
+ /* If this is a call to a pure function returning an array of scalar
+ type, try to apply the return slot optimization. */
+ else if ((TYPE_READONLY (gnu_subprog_type)
+ || TYPE_RESTRICT (gnu_subprog_type))
+ && is_array_of_scalar_type (gnu_result_type)
+ && TYPE_MODE (gnu_result_type) == BLKmode
+ && aggregate_value_p (gnu_result_type, gnu_subprog_type)
+ && return_slot_opt_for_pure_call_p (gnu_target, gnu_call))
+ op_code = INIT_EXPR;
+
+ else
+ op_code = MODIFY_EXPR;
+
+ /* Use the required method to move the result to the target. */
+ if (atomic_access == OUTER_ATOMIC)
+ gnu_call
+ = build_load_modify_store (gnu_target, gnu_call, gnat_node);
+ else if (atomic_access == SIMPLE_ATOMIC)
+ gnu_call = build_atomic_store (gnu_target, gnu_call, atomic_sync);
+ else
+ gnu_call
+ = build_binary_op (op_code, NULL_TREE, gnu_target, gnu_call);
+
+ if (EXPR_P (gnu_call))
+ set_expr_location_from_node (gnu_call, gnat_parent);
+ append_to_statement_list (gnu_call, &gnu_stmt_list);
+ }
+ else
+ *gnu_result_type_p = get_unpadded_type (Etype (gnat_node));
+ }
+
+ /* Otherwise, if this is a procedure call statement without copy-in/copy-out
+ parameters, the result is just the call statement. */
+ else if (!TYPE_CI_CO_LIST (gnu_subprog_type))
+ append_to_statement_list (gnu_call, &gnu_stmt_list);
+
+ /* Finally, add the copy back statements, if any. */
+ append_to_statement_list (gnu_after_list, &gnu_stmt_list);
+
+ if (went_into_elab_proc)
+ current_function_decl = NULL_TREE;
+
+ /* If we have pushed a binding level, pop it and finish up the enclosing
+ statement group. */
+ if (pushed_binding_level)
+ {
+ add_stmt (gnu_stmt_list);
+ gnat_poplevel ();
+ gnu_result = end_stmt_group ();
+ }
+
+ /* Otherwise, retrieve the statement list, if any. */
+ else if (gnu_stmt_list)
+ gnu_result = gnu_stmt_list;
+
+ /* Otherwise, just return the call expression. */
+ else
+ return gnu_call;
+
+ /* If we nevertheless need a value, make a COMPOUND_EXPR to return it.
+ But first simplify if we have only one statement in the list. */
+ if (returning_value)
+ {
+ tree first = expr_first (gnu_result), last = expr_last (gnu_result);
+ if (first == last)
+ gnu_result = first;
+ gnu_result
+ = build_compound_expr (TREE_TYPE (gnu_call), gnu_result, gnu_call);
+ }
+
+ return gnu_result;
+}
+
+/* Subroutine of gnat_to_gnu to translate gnat_node, an
+ N_Handled_Sequence_Of_Statements, to a GCC tree, which is returned. */
+
+static tree
+Handled_Sequence_Of_Statements_to_gnu (Node_Id gnat_node)
+{
+ /* If just annotating, ignore all EH and cleanups. */
+ const bool gcc_eh
+ = (!type_annotate_only
+ && Present (Exception_Handlers (gnat_node))
+ && Back_End_Exceptions ());
+ const bool fe_sjlj_eh
+ = (!type_annotate_only
+ && Present (Exception_Handlers (gnat_node))
+ && Exception_Mechanism == Front_End_SJLJ);
+ const bool at_end = !type_annotate_only && Present (At_End_Proc (gnat_node));
+ const bool binding_for_block = (at_end || gcc_eh || fe_sjlj_eh);
+ tree gnu_jmpsave_decl = NULL_TREE;
+ tree gnu_jmpbuf_decl = NULL_TREE;
+ tree gnu_inner_block; /* The statement(s) for the block itself. */
+ tree gnu_result;
+ tree gnu_expr;
+ Node_Id gnat_temp;
+
+ /* The GCC exception handling mechanism can handle both ZCX and SJLJ schemes
+ and the front-end has its own SJLJ mechanism. To call the GCC mechanism,
+ we call add_cleanup, and when we leave the binding, end_stmt_group will
+ create the TRY_FINALLY_EXPR construct.
+
+ ??? The region level calls down there have been specifically put in place
+ for a ZCX context and currently the order in which things are emitted
+ (region/handlers) is different from the SJLJ case. Instead of putting
+ other calls with different conditions at other places for the SJLJ case,
+ it seems cleaner to reorder things for the SJLJ case and generalize the
+ condition to make it not ZCX specific.
+
+ If there are any exceptions or cleanup processing involved, we need an
+ outer statement group (for front-end SJLJ) and binding level. */
+ if (binding_for_block)
+ {
+ start_stmt_group ();
+ gnat_pushlevel ();
+ }
+
+ /* If using fe_sjlj_eh, make the variables for the setjmp buffer and save
+ area for address of previous buffer. Do this first since we need to have
+ the setjmp buf known for any decls in this block. */
+ if (fe_sjlj_eh)
+ {
+ gnu_jmpsave_decl
+ = create_var_decl (get_identifier ("JMPBUF_SAVE"), NULL_TREE,
+ jmpbuf_ptr_type,
+ build_call_n_expr (get_jmpbuf_decl, 0),
+ false, false, false, false, false, true, false,
+ NULL, gnat_node);
+
+ /* The __builtin_setjmp receivers will immediately reinstall it. Now
+ because of the unstructured form of EH used by fe_sjlj_eh, there
+ might be forward edges going to __builtin_setjmp receivers on which
+ it is uninitialized, although they will never be actually taken. */
+ suppress_warning (gnu_jmpsave_decl, OPT_Wuninitialized);
+ gnu_jmpbuf_decl
+ = create_var_decl (get_identifier ("JMP_BUF"), NULL_TREE,
+ jmpbuf_type,
+ NULL_TREE,
+ false, false, false, false, false, true, false,
+ NULL, gnat_node);
+
+ set_block_jmpbuf_decl (gnu_jmpbuf_decl);
+
+ /* When we exit this block, restore the saved value. */
+ add_cleanup (build_call_n_expr (set_jmpbuf_decl, 1, gnu_jmpsave_decl),
+ Present (End_Label (gnat_node))
+ ? End_Label (gnat_node) : gnat_node);
+ }
+
+ /* If we are to call a function when exiting this block, add a cleanup
+ to the binding level we made above. Note that add_cleanup is FIFO
+ so we must register this cleanup after the EH cleanup just above. */
+ if (at_end)
+ {
+ tree proc_decl = gnat_to_gnu (At_End_Proc (gnat_node));
+
+ /* When not optimizing, disable inlining of finalizers as this can
+ create a more complex CFG in the parent function. */
+ if (!optimize || optimize_debug)
+ DECL_DECLARED_INLINE_P (proc_decl) = 0;
+
+ /* If there is no end label attached, we use the location of the At_End
+ procedure because Expand_Cleanup_Actions might reset the location of
+ the enclosing construct to that of an inner statement. */
+ add_cleanup (build_call_n_expr (proc_decl, 0),
+ Present (End_Label (gnat_node))
+ ? End_Label (gnat_node) : At_End_Proc (gnat_node));
+ }
+
+ /* Now build the tree for the declarations and statements inside this block.
+ If this is SJLJ, set our jmp_buf as the current buffer. */
+ start_stmt_group ();
+
+ if (fe_sjlj_eh)
+ {
+ gnu_expr = build_call_n_expr (set_jmpbuf_decl, 1,
+ build_unary_op (ADDR_EXPR, NULL_TREE,
+ gnu_jmpbuf_decl));
+ set_expr_location_from_node (gnu_expr, gnat_node);
+ add_stmt (gnu_expr);
+ }
+
+ if (Present (First_Real_Statement (gnat_node)))
+ process_decls (Statements (gnat_node), Empty,
+ First_Real_Statement (gnat_node), true, true);
+
+ /* Generate code for each statement in the block. */
+ for (gnat_temp = (Present (First_Real_Statement (gnat_node))
+ ? First_Real_Statement (gnat_node)
+ : First (Statements (gnat_node)));
+ Present (gnat_temp); gnat_temp = Next (gnat_temp))
+ add_stmt (gnat_to_gnu (gnat_temp));
+
+ gnu_inner_block = end_stmt_group ();
+
+ /* Now generate code for the two exception models, if either is relevant for
+ this block. */
+ if (fe_sjlj_eh)
+ {
+ tree *gnu_else_ptr = 0;
+ tree gnu_handler;
+
+ /* Make a binding level for the exception handling declarations and code
+ and set up gnu_except_ptr_stack for the handlers to use. */
+ start_stmt_group ();
+ gnat_pushlevel ();
+
+ vec_safe_push (gnu_except_ptr_stack,
+ create_var_decl (get_identifier ("EXCEPT_PTR"), NULL_TREE,
+ build_pointer_type (except_type_node),
+ build_call_n_expr (get_excptr_decl, 0),
+ false, false, false, false, false,
+ true, false, NULL, gnat_node));
+
+ /* Generate code for each handler. The N_Exception_Handler case does the
+ real work and returns a COND_EXPR for each handler, which we chain
+ together here. */
+ for (gnat_temp = First_Non_Pragma (Exception_Handlers (gnat_node));
+ Present (gnat_temp); gnat_temp = Next_Non_Pragma (gnat_temp))
+ {
+ gnu_expr = gnat_to_gnu (gnat_temp);
+
+ /* If this is the first one, set it as the outer one. Otherwise,
+ point the "else" part of the previous handler to us. Then point
+ to our "else" part. */
+ if (!gnu_else_ptr)
+ add_stmt (gnu_expr);
+ else
+ *gnu_else_ptr = gnu_expr;
+
+ gnu_else_ptr = &COND_EXPR_ELSE (gnu_expr);
+ }
+
+ /* If none of the exception handlers did anything, re-raise but do not
+ defer abortion. */
+ gnu_expr = build_call_n_expr (raise_nodefer_decl, 1,
+ gnu_except_ptr_stack->last ());
+ set_expr_location_from_node
+ (gnu_expr,
+ Present (End_Label (gnat_node)) ? End_Label (gnat_node) : gnat_node);
+
+ if (gnu_else_ptr)
+ *gnu_else_ptr = gnu_expr;
+ else
+ add_stmt (gnu_expr);
+
+ /* End the binding level dedicated to the exception handlers and get the
+ whole statement group. */
+ gnu_except_ptr_stack->pop ();
+ gnat_poplevel ();
+ gnu_handler = end_stmt_group ();
+
+ /* If the setjmp returns 1, we restore our incoming longjmp value and
+ then check the handlers. */
+ start_stmt_group ();
+ add_stmt_with_node (build_call_n_expr (set_jmpbuf_decl, 1,
+ gnu_jmpsave_decl),
+ gnat_node);
+ add_stmt (gnu_handler);
+ gnu_handler = end_stmt_group ();
+
+ /* This block is now "if (setjmp) ... <handlers> else <block>". */
+ gnu_result = build3 (COND_EXPR, void_type_node,
+ (build_call_n_expr
+ (setjmp_decl, 1,
+ build_unary_op (ADDR_EXPR, NULL_TREE,
+ gnu_jmpbuf_decl))),
+ gnu_handler, gnu_inner_block);
+ }
+ else if (gcc_eh)
+ {
+ tree gnu_handlers;
+ location_t locus;
+
+ /* First make a block containing the handlers. */
+ start_stmt_group ();
+ for (gnat_temp = First_Non_Pragma (Exception_Handlers (gnat_node));
+ Present (gnat_temp);
+ gnat_temp = Next_Non_Pragma (gnat_temp))
+ add_stmt (gnat_to_gnu (gnat_temp));
+ gnu_handlers = end_stmt_group ();
+
+ /* Now make the TRY_CATCH_EXPR for the block. */
+ gnu_result = build2 (TRY_CATCH_EXPR, void_type_node,
+ gnu_inner_block, gnu_handlers);
+ /* Set a location. We need to find a unique location for the dispatching
+ code, otherwise we can get coverage or debugging issues. Try with
+ the location of the end label. */
+ if (Present (End_Label (gnat_node))
+ && Sloc_to_locus (Sloc (End_Label (gnat_node)), &locus))
+ SET_EXPR_LOCATION (gnu_result, locus);
+ else
+ /* Clear column information so that the exception handler of an
+ implicit transient block does not incorrectly inherit the slocs
+ of a decision, which would otherwise confuse control flow based
+ coverage analysis tools. */
+ set_expr_location_from_node (gnu_result, gnat_node, true);
+ }
+ else
+ gnu_result = gnu_inner_block;
+
+ /* Now close our outer block, if we had to make one. */
+ if (binding_for_block)
+ {
+ add_stmt (gnu_result);
+ gnat_poplevel ();
+ gnu_result = end_stmt_group ();
+ }
+
+ return gnu_result;
+}
+
+/* Subroutine of gnat_to_gnu to translate gnat_node, an N_Exception_Handler,
+ to a GCC tree, which is returned. This is the variant for front-end sjlj
+ exception handling. */
+
+static tree
+Exception_Handler_to_gnu_fe_sjlj (Node_Id gnat_node)
+{
+ /* Unless this is "Others" or the special "Non-Ada" exception for Ada, make
+ an "if" statement to select the proper exceptions. For "Others", exclude
+ exceptions where Handled_By_Others is nonzero unless the All_Others flag
+ is set. For "Non-ada", accept an exception if "Lang" is 'V'. */
+ tree gnu_choice = boolean_false_node;
+ tree gnu_body = build_stmt_group (Statements (gnat_node), false);
+ Node_Id gnat_temp;
+
+ for (gnat_temp = First (Exception_Choices (gnat_node));
+ gnat_temp; gnat_temp = Next (gnat_temp))
+ {
+ tree this_choice;
+
+ if (Nkind (gnat_temp) == N_Others_Choice)
+ {
+ if (All_Others (gnat_temp))
+ this_choice = boolean_true_node;
+ else
+ this_choice
+ = build_binary_op
+ (EQ_EXPR, boolean_type_node,
+ convert
+ (integer_type_node,
+ build_component_ref
+ (build_unary_op
+ (INDIRECT_REF, NULL_TREE,
+ gnu_except_ptr_stack->last ()),
+ not_handled_by_others_decl,
+ false)),
+ integer_zero_node);
+ }
+
+ else if (Nkind (gnat_temp) == N_Identifier
+ || Nkind (gnat_temp) == N_Expanded_Name)
+ {
+ Entity_Id gnat_ex_id = Entity (gnat_temp);
+ tree gnu_expr;
+
+ /* Exception may be a renaming. Recover original exception which is
+ the one elaborated and registered. */
+ if (Present (Renamed_Object (gnat_ex_id)))
+ gnat_ex_id = Renamed_Object (gnat_ex_id);
+
+ gnu_expr = gnat_to_gnu_entity (gnat_ex_id, NULL_TREE, false);
+
+ this_choice
+ = build_binary_op
+ (EQ_EXPR, boolean_type_node,
+ gnu_except_ptr_stack->last (),
+ convert (TREE_TYPE (gnu_except_ptr_stack->last ()),
+ build_unary_op (ADDR_EXPR, NULL_TREE, gnu_expr)));
+ }
+ else
+ gcc_unreachable ();
+
+ gnu_choice = build_binary_op (TRUTH_ORIF_EXPR, boolean_type_node,
+ gnu_choice, this_choice);
+ }
+
+ return build3 (COND_EXPR, void_type_node, gnu_choice, gnu_body, NULL_TREE);
+}
+
+/* Return true if no statement in GNAT_LIST can alter the control flow. */
+
+static bool
+stmt_list_cannot_alter_control_flow_p (List_Id gnat_list)
+{
+ if (No (gnat_list))
+ return true;
+
+ /* This is very conservative, we reject everything except for simple
+ assignments between identifiers or literals. */
+ for (Node_Id gnat_node = First (gnat_list);
+ Present (gnat_node);
+ gnat_node = Next (gnat_node))
+ {
+ if (Nkind (gnat_node) != N_Assignment_Statement)
+ return false;
+
+ if (Nkind (Name (gnat_node)) != N_Identifier)
+ return false;
+
+ Node_Kind nkind = Nkind (Expression (gnat_node));
+ if (nkind != N_Identifier
+ && nkind != N_Integer_Literal
+ && nkind != N_Real_Literal)
+ return false;
+ }
+
+ return true;
+}
+
+/* Subroutine of gnat_to_gnu to translate gnat_node, an N_Exception_Handler,
+ to a GCC tree, which is returned. This is the variant for GCC exception
+ schemes. */
+
+static tree
+Exception_Handler_to_gnu_gcc (Node_Id gnat_node)
+{
+ tree gnu_etypes_list = NULL_TREE;
+
+ /* We build a TREE_LIST of nodes representing what exception types this
+ handler can catch, with special cases for others and all others cases.
+
+ Each exception type is actually identified by a pointer to the exception
+ id, or to a dummy object for "others" and "all others". */
+ for (Node_Id gnat_temp = First (Exception_Choices (gnat_node));
+ gnat_temp;
+ gnat_temp = Next (gnat_temp))
+ {
+ tree gnu_expr, gnu_etype;
+
+ if (Nkind (gnat_temp) == N_Others_Choice)
+ {
+ gnu_expr = All_Others (gnat_temp) ? all_others_decl : others_decl;
+ gnu_etype = build_unary_op (ADDR_EXPR, NULL_TREE, gnu_expr);
+ }
+ else if (Nkind (gnat_temp) == N_Identifier
+ || Nkind (gnat_temp) == N_Expanded_Name)
+ {
+ Entity_Id gnat_ex_id = Entity (gnat_temp);
+
+ /* Exception may be a renaming. Recover original exception which is
+ the one elaborated and registered. */
+ if (Present (Renamed_Object (gnat_ex_id)))
+ gnat_ex_id = Renamed_Object (gnat_ex_id);
+
+ gnu_expr = gnat_to_gnu_entity (gnat_ex_id, NULL_TREE, false);
+ gnu_etype = build_unary_op (ADDR_EXPR, NULL_TREE, gnu_expr);
+ }
+ else
+ gcc_unreachable ();
+
+ /* The GCC interface expects NULL to be passed for catch all handlers, so
+ it would be quite tempting to set gnu_etypes_list to NULL if gnu_etype
+ is integer_zero_node. It would not work, however, because GCC's
+ notion of "catch all" is stronger than our notion of "others". Until
+ we correctly use the cleanup interface as well, doing that would
+ prevent the "all others" handlers from being seen, because nothing
+ can be caught beyond a catch all from GCC's point of view. */
+ gnu_etypes_list = tree_cons (NULL_TREE, gnu_etype, gnu_etypes_list);
+ }
+
+ start_stmt_group ();
+ gnat_pushlevel ();
+
+ /* Expand a call to the begin_handler hook at the beginning of the
+ handler, and arrange for a call to the end_handler hook to occur
+ on every possible exit path. GDB sets a breakpoint in the
+ begin_handler for catchpoints.
+
+ A v1 begin handler saves the cleanup from the exception object,
+ and marks the exception as in use, so that it will not be
+ released by other handlers. A v1 end handler restores the
+ cleanup and releases the exception object, unless it is still
+ claimed, or the exception is being propagated (reraised).
+
+ __builtin_eh_pointer references the exception occurrence being
+ handled or propagated. Within the handler region, it is the
+ former, but within the else branch of the EH_ELSE_EXPR, i.e. the
+ exceptional cleanup path, it is the latter, so we must save the
+ occurrence being handled early on, so that, should an exception
+ be (re)raised, we can release the current exception, or figure
+ out we're not to release it because we're propagating a reraise
+ thereof.
+
+ We use local variables to retrieve the incoming value at handler
+ entry time (EXPTR), the saved cleanup (EXCLN) and the token
+ (EXVTK), and reuse them to feed the end_handler hook's argument
+ at exit. */
+
+ /* CODE: void *EXPTR = __builtin_eh_pointer (0); */
+ tree gnu_current_exc_ptr
+ = build_call_expr (builtin_decl_explicit (BUILT_IN_EH_POINTER),
+ 1, integer_zero_node);
+ tree exc_ptr
+ = create_var_decl (get_identifier ("EXPTR"), NULL_TREE,
+ ptr_type_node, gnu_current_exc_ptr,
+ true, false, false, false, false, true, true,
+ NULL, gnat_node);
+
+ tree prev_gnu_incoming_exc_ptr = gnu_incoming_exc_ptr;
+ gnu_incoming_exc_ptr = exc_ptr;
+
+ /* begin_handler_decl must not throw, so we can use it as an
+ initializer for a variable used in cleanups.
+
+ CODE: void *EXCLN = __gnat_begin_handler_v1 (EXPTR); */
+ tree exc_cleanup
+ = create_var_decl (get_identifier ("EXCLN"), NULL_TREE,
+ ptr_type_node,
+ build_call_n_expr (begin_handler_decl, 1,
+ exc_ptr),
+ true, false, false, false, false,
+ true, true, NULL, gnat_node);
+
+ /* Declare and initialize the choice parameter, if present. */
+ if (Present (Choice_Parameter (gnat_node)))
+ {
+ tree gnu_param
+ = gnat_to_gnu_entity (Choice_Parameter (gnat_node), NULL_TREE, true);
+
+ /* CODE: __gnat_set_exception_parameter (&choice_param, EXPTR); */
+ add_stmt (build_call_n_expr
+ (set_exception_parameter_decl, 2,
+ build_unary_op (ADDR_EXPR, NULL_TREE, gnu_param),
+ gnu_incoming_exc_ptr));
+ }
+
+ /* CODE: <handler proper> */
+ add_stmt_list (Statements (gnat_node));
+
+ tree call = build_call_n_expr (end_handler_decl, 3,
+ exc_ptr,
+ exc_cleanup,
+ null_pointer_node);
+ /* If the handler can only end by falling off the end, don't bother
+ with cleanups. */
+ if (stmt_list_cannot_alter_control_flow_p (Statements (gnat_node)))
+ /* CODE: __gnat_end_handler_v1 (EXPTR, EXCLN, NULL); */
+ add_stmt_with_node (call, gnat_node);
+ /* Otherwise, all of the above is after
+ CODE: try {
+
+ The call above will appear after
+ CODE: } finally {
+
+ And the code below will appear after
+ CODE: } else {
+
+ The else block to a finally block is taken instead of the finally
+ block when an exception propagates out of the try block. */
+ else
+ {
+ start_stmt_group ();
+ gnat_pushlevel ();
+ /* CODE: void *EXPRP = __builtin_eh_handler (0); */
+ tree prop_ptr
+ = create_var_decl (get_identifier ("EXPRP"), NULL_TREE,
+ ptr_type_node,
+ build_call_expr (builtin_decl_explicit
+ (BUILT_IN_EH_POINTER),
+ 1, integer_zero_node),
+ true, false, false, false, false,
+ true, true, NULL, gnat_node);
+
+ /* CODE: __gnat_end_handler_v1 (EXPTR, EXCLN, EXPRP); */
+ tree ecall = build_call_n_expr (end_handler_decl, 3,
+ exc_ptr,
+ exc_cleanup,
+ prop_ptr);
+
+ add_stmt_with_node (ecall, gnat_node);
+
+ /* CODE: } */
+ gnat_poplevel ();
+ tree eblk = end_stmt_group ();
+ tree ehls = build2 (EH_ELSE_EXPR, void_type_node, call, eblk);
+ add_cleanup (ehls, gnat_node);
+ }
+
+ gnat_poplevel ();
+
+ gnu_incoming_exc_ptr = prev_gnu_incoming_exc_ptr;
+
+ return
+ build2 (CATCH_EXPR, void_type_node, gnu_etypes_list, end_stmt_group ());
+}
+
+/* Subroutine of gnat_to_gnu to generate code for an N_Compilation unit. */
+
+static void
+Compilation_Unit_to_gnu (Node_Id gnat_node)
+{
+ const Node_Id gnat_unit = Unit (gnat_node);
+ const bool body_p = (Nkind (gnat_unit) == N_Package_Body
+ || Nkind (gnat_unit) == N_Subprogram_Body);
+ const Entity_Id gnat_unit_entity = Defining_Entity (gnat_unit);
+ Entity_Id gnat_entity;
+ Node_Id gnat_pragma, gnat_iter;
+ /* Make the decl for the elaboration procedure. Emit debug info for it, so
+ that users can break into their elaboration code in debuggers. Kludge:
+ don't consider it as a definition so that we have a line map for its
+ body, but no subprogram description in debug info. In addition, don't
+ qualify it as artificial, even though it is not a user subprogram per se,
+ in particular for specs. Unlike, say, clones created internally by the
+ compiler, this subprogram materializes specific user code and flagging it
+ artificial would take elab code away from gcov's analysis. */
+ tree gnu_elab_proc_decl
+ = create_subprog_decl
+ (create_concat_name (gnat_unit_entity, body_p ? "elabb" : "elabs"),
+ NULL_TREE, void_ftype, NULL_TREE,
+ is_default, true, false, false, true, false, NULL, gnat_unit);
+ struct elab_info *info;
+
+ vec_safe_push (gnu_elab_proc_stack, gnu_elab_proc_decl);
+ DECL_ELABORATION_PROC_P (gnu_elab_proc_decl) = 1;
+
+ /* Initialize the information structure for the function. */
+ allocate_struct_function (gnu_elab_proc_decl, false);
+ set_cfun (NULL);
+
+ current_function_decl = NULL_TREE;
+
+ start_stmt_group ();
+ gnat_pushlevel ();
+
+ /* For a body, first process the spec if there is one. */
+ if (Nkind (gnat_unit) == N_Package_Body
+ || (Nkind (gnat_unit) == N_Subprogram_Body && !Acts_As_Spec (gnat_node)))
+ add_stmt (gnat_to_gnu (Library_Unit (gnat_node)));
+
+ if (type_annotate_only && gnat_node == Cunit (Main_Unit))
+ {
+ elaborate_all_entities (gnat_node);
+
+ if (Nkind (gnat_unit) == N_Subprogram_Declaration
+ || Nkind (gnat_unit) == N_Generic_Package_Declaration
+ || Nkind (gnat_unit) == N_Generic_Subprogram_Declaration)
+ return;
+ }
+
+ /* Then process any pragmas and declarations preceding the unit. */
+ for (gnat_pragma = First (Context_Items (gnat_node));
+ Present (gnat_pragma);
+ gnat_pragma = Next (gnat_pragma))
+ if (Nkind (gnat_pragma) == N_Pragma)
+ add_stmt (gnat_to_gnu (gnat_pragma));
+ process_decls (Declarations (Aux_Decls_Node (gnat_node)), Empty, Empty,
+ true, true);
+
+ /* Process the unit itself. */
+ add_stmt (gnat_to_gnu (gnat_unit));
+
+ /* Generate code for all the inlined subprograms. */
+ for (gnat_entity = First_Inlined_Subprogram (gnat_node);
+ Present (gnat_entity);
+ gnat_entity = Next_Inlined_Subprogram (gnat_entity))
+ {
+ Node_Id gnat_body;
+
+ /* Without optimization, process only the required subprograms. */
+ if (!optimize && !Has_Pragma_Inline_Always (gnat_entity))
+ continue;
+
+ /* The set of inlined subprograms is computed from data recorded early
+ during expansion and it can be a strict superset of the final set
+ computed after semantic analysis, for example if a call to such a
+ subprogram occurs in a pragma Assert and assertions are disabled.
+ In that case, semantic analysis resets Is_Public to false but the
+ entry for the subprogram in the inlining tables is stalled. */
+ if (!Is_Public (gnat_entity))
+ continue;
+
+ gnat_body = Parent (Declaration_Node (gnat_entity));
+ if (Nkind (gnat_body) != N_Subprogram_Body)
+ {
+ /* ??? This happens when only the spec of a package is provided. */
+ if (No (Corresponding_Body (gnat_body)))
+ continue;
+
+ gnat_body
+ = Parent (Declaration_Node (Corresponding_Body (gnat_body)));
+ }
+
+ /* Define the entity first so we set DECL_EXTERNAL. */
+ gnat_to_gnu_entity (gnat_entity, NULL_TREE, false);
+ add_stmt (gnat_to_gnu (gnat_body));
+ }
+
+ /* Process any pragmas and actions following the unit. */
+ add_stmt_list (Pragmas_After (Aux_Decls_Node (gnat_node)));
+ add_stmt_list (Actions (Aux_Decls_Node (gnat_node)));
+ finalize_from_limited_with ();
+
+ /* Then process the expressions of pragma Compile_Time_{Error|Warning} to
+ annotate types referenced therein if they have not been annotated. */
+ for (int i = 0; gnat_compile_time_expr_list.iterate (i, &gnat_iter); i++)
+ (void) gnat_to_gnu_external (gnat_iter);
+ gnat_compile_time_expr_list.release ();
+
+ /* Save away what we've made so far and finish it up. */
+ set_current_block_context (gnu_elab_proc_decl);
+ gnat_poplevel ();
+ DECL_SAVED_TREE (gnu_elab_proc_decl) = end_stmt_group ();
+ set_end_locus_from_node (gnu_elab_proc_decl, gnat_unit);
+ gnu_elab_proc_stack->pop ();
+
+ /* Record this potential elaboration procedure for later processing. */
+ info = ggc_alloc<elab_info> ();
+ info->next = elab_info_list;
+ info->elab_proc = gnu_elab_proc_decl;
+ info->gnat_node = gnat_node;
+ elab_info_list = info;
+
+ /* Force the processing for all nodes that remain in the queue. */
+ process_deferred_decl_context (true);
+}
+
+/* Mark COND, a boolean expression, as predicating a call to a noreturn
+ function, i.e. predict that it is very likely false, and return it.
+
+ The compiler will automatically predict the last edge leading to a call
+ to a noreturn function as very unlikely taken. This function makes it
+ possible to extend the prediction to predecessors in case the condition
+ is made up of several short-circuit operators. */
+
+static tree
+build_noreturn_cond (tree cond)
+{
+ tree pred_cst = build_int_cst (integer_type_node, PRED_NORETURN);
+ return
+ build_call_expr_internal_loc (UNKNOWN_LOCATION, IFN_BUILTIN_EXPECT,
+ boolean_type_node, 3, cond,
+ boolean_false_node, pred_cst);
+}
+
+/* Subroutine of gnat_to_gnu to translate GNAT_RANGE, a node representing a
+ range of values, into GNU_LOW and GNU_HIGH bounds. */
+
+static void
+Range_to_gnu (Node_Id gnat_range, tree *gnu_low, tree *gnu_high)
+{
+ /* GNAT_RANGE is either an N_Range or an identifier denoting a subtype. */
+ switch (Nkind (gnat_range))
+ {
+ case N_Range:
+ *gnu_low = gnat_to_gnu (Low_Bound (gnat_range));
+ *gnu_high = gnat_to_gnu (High_Bound (gnat_range));
+ break;
+
+ case N_Expanded_Name:
+ case N_Identifier:
+ {
+ tree gnu_range_type = get_unpadded_type (Entity (gnat_range));
+ tree gnu_range_base_type = get_base_type (gnu_range_type);
+
+ *gnu_low
+ = convert (gnu_range_base_type, TYPE_MIN_VALUE (gnu_range_type));
+ *gnu_high
+ = convert (gnu_range_base_type, TYPE_MAX_VALUE (gnu_range_type));
+ }
+ break;
+
+ default:
+ gcc_unreachable ();
+ }
+}
+
+/* Subroutine of gnat_to_gnu to translate GNAT_NODE, an N_Raise_xxx_Error,
+ to a GCC tree and return it. GNU_RESULT_TYPE_P is a pointer to where
+ we should place the result type. */
+
+static tree
+Raise_Error_to_gnu (Node_Id gnat_node, tree *gnu_result_type_p)
+{
+ const Node_Kind kind = Nkind (gnat_node);
+ const Node_Id gnat_cond = Condition (gnat_node);
+ const int reason = UI_To_Int (Reason (gnat_node));
+ const bool with_extra_info
+ = Exception_Extra_Info
+ && !No_Exception_Handlers_Set ()
+ && No (get_exception_label (kind));
+ tree gnu_result = NULL_TREE, gnu_cond = NULL_TREE;
+ Node_Id gnat_rcond;
+
+ /* The following processing is not required for correctness. Its purpose is
+ to give more precise error messages and to record some information. */
+ switch (reason)
+ {
+ case CE_Access_Check_Failed:
+ if (with_extra_info)
+ gnu_result = build_call_raise_column (reason, gnat_node, kind);
+ break;
+
+ case CE_Index_Check_Failed:
+ case CE_Range_Check_Failed:
+ case CE_Invalid_Data:
+ if (No (gnat_cond) || Nkind (gnat_cond) != N_Op_Not)
+ break;
+ gnat_rcond = Right_Opnd (gnat_cond);
+ if (Nkind (gnat_rcond) == N_In
+ || Nkind (gnat_rcond) == N_Op_Ge
+ || Nkind (gnat_rcond) == N_Op_Le)
+ {
+ const Node_Id gnat_index = Left_Opnd (gnat_rcond);
+ const Node_Id gnat_type = Etype (gnat_index);
+ tree gnu_index = gnat_to_gnu (gnat_index);
+ tree gnu_type = get_unpadded_type (gnat_type);
+ tree gnu_low_bound, gnu_high_bound, disp;
+ struct loop_info_d *loop;
+ bool neg_p;
+
+ switch (Nkind (gnat_rcond))
+ {
+ case N_In:
+ Range_to_gnu (Right_Opnd (gnat_rcond),
+ &gnu_low_bound, &gnu_high_bound);
+ break;
+
+ case N_Op_Ge:
+ gnu_low_bound = gnat_to_gnu (Right_Opnd (gnat_rcond));
+ gnu_high_bound = TYPE_MAX_VALUE (gnu_type);
+ break;
+
+ case N_Op_Le:
+ gnu_low_bound = TYPE_MIN_VALUE (gnu_type);
+ gnu_high_bound = gnat_to_gnu (Right_Opnd (gnat_rcond));
+ break;
+
+ default:
+ gcc_unreachable ();
+ }
+
+ gnu_type = maybe_character_type (gnu_type);
+ if (TREE_TYPE (gnu_index) != gnu_type)
+ {
+ gnu_low_bound = convert (gnu_type, gnu_low_bound);
+ gnu_high_bound = convert (gnu_type, gnu_high_bound);
+ gnu_index = convert (gnu_type, gnu_index);
+ }
+
+ if (with_extra_info
+ && Known_Esize (gnat_type)
+ && UI_To_Int (Esize (gnat_type)) <= 32)
+ gnu_result
+ = build_call_raise_range (reason, gnat_node, kind, gnu_index,
+ gnu_low_bound, gnu_high_bound);
+
+ /* If optimization is enabled and we are inside a loop, we try to
+ compute invariant conditions for checks applied to the iteration
+ variable, i.e. conditions that are independent of the variable
+ and necessary in order for the checks to fail in the course of
+ some iteration. If we succeed, we consider an alternative:
+
+ 1. If loop unswitching is enabled, we prepend these conditions
+ to the original conditions of the checks. This will make it
+ possible for the loop unswitching pass to replace the loop
+ with two loops, one of which has the checks eliminated and
+ the other has the original checks reinstated, and a prologue
+ implementing a run-time selection. The former loop will be
+ for example suitable for vectorization.
+
+ 2. Otherwise, we instead append the conditions to the original
+ conditions of the checks. At worse, if the conditions cannot
+ be evaluated at compile time, they will be evaluated as true
+ at run time only when the checks have already failed, thus
+ contributing negatively only to the size of the executable.
+ But the hope is that these invariant conditions be evaluated
+ at compile time to false, thus taking away the entire checks
+ with them. */
+ if (optimize
+ && inside_loop_p ()
+ && (!gnu_low_bound
+ || (gnu_low_bound = gnat_invariant_expr (gnu_low_bound)))
+ && (!gnu_high_bound
+ || (gnu_high_bound = gnat_invariant_expr (gnu_high_bound)))
+ && (loop = find_loop_for (gnu_index, &disp, &neg_p)))
+ {
+ struct range_check_info_d *rci = ggc_alloc<range_check_info_d> ();
+ rci->low_bound = gnu_low_bound;
+ rci->high_bound = gnu_high_bound;
+ rci->disp = disp;
+ rci->neg_p = neg_p;
+ rci->type = gnu_type;
+ rci->inserted_cond
+ = build1 (SAVE_EXPR, boolean_type_node, boolean_true_node);
+ vec_safe_push (loop->checks, rci);
+ gnu_cond = build_noreturn_cond (gnat_to_gnu (gnat_cond));
+ if (optimize >= 3)
+ gnu_cond = build_binary_op (TRUTH_ANDIF_EXPR,
+ boolean_type_node,
+ rci->inserted_cond,
+ gnu_cond);
+ else
+ gnu_cond = build_binary_op (TRUTH_ANDIF_EXPR,
+ boolean_type_node,
+ gnu_cond,
+ rci->inserted_cond);
+ }
+ }
+ break;
+
+ default:
+ break;
+ }
+
+ /* The following processing does the real work, but we must nevertheless make
+ sure not to override the result of the previous processing. */
+ if (!gnu_result)
+ gnu_result = build_call_raise (reason, gnat_node, kind);
+ set_expr_location_from_node (gnu_result, gnat_node);
+
+ *gnu_result_type_p = get_unpadded_type (Etype (gnat_node));
+
+ /* If the type is VOID, this is a statement, so we need to generate the code
+ for the call. Handle a condition, if there is one. */
+ if (VOID_TYPE_P (*gnu_result_type_p))
+ {
+ if (Present (gnat_cond))
+ {
+ if (!gnu_cond)
+ gnu_cond = gnat_to_gnu (gnat_cond);
+ if (integer_zerop (gnu_cond))
+ return alloc_stmt_list ();
+ gnu_result = build3 (COND_EXPR, void_type_node, gnu_cond, gnu_result,
+ alloc_stmt_list ());
+ }
+ }
+ else
+ {
+ /* The condition field must not be present when the node is used as an
+ expression form. */
+ gigi_checking_assert (No (gnat_cond));
+ gnu_result = build1 (NULL_EXPR, *gnu_result_type_p, gnu_result);
+ }
+
+ return gnu_result;
+}
+
+/* Return true if GNAT_NODE is on the LHS of an assignment or an actual
+ parameter of a call. */
+
+static bool
+lhs_or_actual_p (Node_Id gnat_node)
+{
+ const Node_Id gnat_parent = Parent (gnat_node);
+ const Node_Kind kind = Nkind (gnat_parent);
+
+ if (kind == N_Assignment_Statement && Name (gnat_parent) == gnat_node)
+ return true;
+
+ if ((kind == N_Procedure_Call_Statement || kind == N_Function_Call)
+ && Name (gnat_parent) != gnat_node)
+ return true;
+
+ if (kind == N_Parameter_Association)
+ return true;
+
+ return false;
+}
+
+/* Return true if either GNAT_NODE or a view of GNAT_NODE is on the LHS
+ of an assignment or an actual parameter of a call. */
+
+static bool
+present_in_lhs_or_actual_p (Node_Id gnat_node)
+{
+ if (lhs_or_actual_p (gnat_node))
+ return true;
+
+ const Node_Kind kind = Nkind (Parent (gnat_node));
+
+ if ((kind == N_Type_Conversion || kind == N_Unchecked_Type_Conversion)
+ && lhs_or_actual_p (Parent (gnat_node)))
+ return true;
+
+ return false;
+}
+
+/* Return true if GNAT_NODE, an unchecked type conversion, is a no-op as far
+ as gigi is concerned. This is used to avoid conversions on the LHS. */
+
+static bool
+unchecked_conversion_nop (Node_Id gnat_node)
+{
+ Entity_Id from_type, to_type;
+
+ /* The conversion must be on the LHS of an assignment or an actual parameter
+ of a call. Otherwise, even if the conversion was essentially a no-op, it
+ could de facto ensure type consistency and this should be preserved. */
+ if (!lhs_or_actual_p (gnat_node))
+ return false;
+
+ from_type = Etype (Expression (gnat_node));
+
+ /* We're interested in artificial conversions generated by the front-end
+ to make private types explicit, e.g. in Expand_Assign_Array. */
+ if (!Is_Private_Type (from_type))
+ return false;
+
+ from_type = Underlying_Type (from_type);
+ to_type = Etype (gnat_node);
+
+ /* The direct conversion to the underlying type is a no-op. */
+ if (to_type == from_type)
+ return true;
+
+ /* For an array subtype, the conversion to the PAIT is a no-op. */
+ if (Ekind (from_type) == E_Array_Subtype
+ && to_type == Packed_Array_Impl_Type (from_type))
+ return true;
+
+ /* For a record subtype, the conversion to the type is a no-op. */
+ if (Ekind (from_type) == E_Record_Subtype
+ && to_type == Etype (from_type))
+ return true;
+
+ return false;
+}
+
+/* Return true if GNAT_NODE represents a statement. */
+
+static bool
+statement_node_p (Node_Id gnat_node)
+{
+ const Node_Kind kind = Nkind (gnat_node);
+
+ if (kind == N_Label)
+ return true;
+
+ if (IN (kind, N_Statement_Other_Than_Procedure_Call))
+ return true;
+
+ if (kind == N_Procedure_Call_Statement)
+ return true;
+
+ if (IN (kind, N_Raise_xxx_Error) && Ekind (Etype (gnat_node)) == E_Void)
+ return true;
+
+ return false;
+}
+
+/* This function is the driver of the GNAT to GCC tree transformation process.
+ It is the entry point of the tree transformer. GNAT_NODE is the root of
+ some GNAT tree. Return the root of the corresponding GCC tree. If this
+ is an expression, return the GCC equivalent of the expression. If this
+ is a statement, return the statement or add it to the current statement
+ group, in which case anything returned is to be interpreted as occurring
+ after anything added. */
+
+tree
+gnat_to_gnu (Node_Id gnat_node)
+{
+ const Node_Kind kind = Nkind (gnat_node);
+ tree gnu_result = error_mark_node; /* Default to no value. */
+ tree gnu_result_type = void_type_node;
+ tree gnu_expr, gnu_lhs, gnu_rhs;
+ Node_Id gnat_temp;
+ atomic_acces_t aa_type;
+ bool went_into_elab_proc;
+ bool aa_sync;
+
+ /* Save node number for error message and set location information. */
+ Current_Error_Node = gnat_node;
+ Sloc_to_locus (Sloc (gnat_node), &input_location);
+
+ /* If we are only annotating types and this node is a statement, return
+ an empty statement list. */
+ if (type_annotate_only && statement_node_p (gnat_node))
+ return alloc_stmt_list ();
+
+ /* If we are only annotating types and this node is a subexpression, return
+ a NULL_EXPR, but filter out nodes appearing in the expressions attached
+ to packed array implementation types. */
+ if (type_annotate_only
+ && IN (kind, N_Subexpr)
+ && !(((IN (kind, N_Op) && kind != N_Op_Expon)
+ || kind == N_Type_Conversion)
+ && Is_Integer_Type (Etype (gnat_node)))
+ && !(kind == N_Attribute_Reference
+ && (Get_Attribute_Id (Attribute_Name (gnat_node)) == Attr_Length
+ || Get_Attribute_Id (Attribute_Name (gnat_node)) == Attr_Size)
+ && Is_Constrained (Etype (Prefix (gnat_node)))
+ && !Is_Constr_Subt_For_U_Nominal (Etype (Prefix (gnat_node))))
+ && kind != N_Expanded_Name
+ && kind != N_Identifier
+ && !Compile_Time_Known_Value (gnat_node))
+ return build1 (NULL_EXPR, get_unpadded_type (Etype (gnat_node)),
+ build_call_raise (CE_Range_Check_Failed, gnat_node,
+ N_Raise_Constraint_Error));
+
+ /* If this is a statement and we are at top level, it must be part of the
+ elaboration procedure, so mark us as being in that procedure. */
+ if ((statement_node_p (gnat_node)
+ || kind == N_Handled_Sequence_Of_Statements
+ || kind == N_Implicit_Label_Declaration)
+ && !current_function_decl)
+ {
+ current_function_decl = get_elaboration_procedure ();
+ went_into_elab_proc = true;
+ }
+ else
+ went_into_elab_proc = false;
+
+ switch (kind)
+ {
+ /********************************/
+ /* Chapter 2: Lexical Elements */
+ /********************************/
+
+ case N_Identifier:
+ case N_Expanded_Name:
+ case N_Operator_Symbol:
+ case N_Defining_Identifier:
+ case N_Defining_Operator_Symbol:
+ gnu_result = Identifier_to_gnu (gnat_node, &gnu_result_type);
+
+ /* If atomic access is required on the RHS, build the atomic load. */
+ if (simple_atomic_access_required_p (gnat_node, &aa_sync)
+ && !present_in_lhs_or_actual_p (gnat_node))
+ gnu_result = build_atomic_load (gnu_result, aa_sync);
+ break;
+
+ case N_Integer_Literal:
+ {
+ tree gnu_type;
+
+ /* Get the type of the result, looking inside any padding and
+ justified modular types. Then get the value in that type. */
+ gnu_type = gnu_result_type = get_unpadded_type (Etype (gnat_node));
+
+ if (TREE_CODE (gnu_type) == RECORD_TYPE
+ && TYPE_JUSTIFIED_MODULAR_P (gnu_type))
+ gnu_type = TREE_TYPE (TYPE_FIELDS (gnu_type));
+
+ gnu_result = UI_To_gnu (Intval (gnat_node), gnu_type);
+
+ /* If the result overflows (meaning it doesn't fit in its base type),
+ abort, unless this is for a named number because that's not fatal.
+ We would like to check that the value is within the range of the
+ subtype, but that causes problems with subtypes whose usage will
+ raise Constraint_Error and also with biased representation. */
+ if (TREE_OVERFLOW (gnu_result))
+ {
+ if (Nkind (Parent (gnat_node)) == N_Number_Declaration)
+ gnu_result = error_mark_node;
+ else
+ gcc_unreachable ();
+ }
+ }
+ break;
+
+ case N_Character_Literal:
+ /* If a Entity is present, it means that this was one of the
+ literals in a user-defined character type. In that case,
+ just return the value in the CONST_DECL. Otherwise, use the
+ character code. In that case, the base type should be an
+ INTEGER_TYPE, but we won't bother checking for that. */
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ if (Present (Entity (gnat_node)))
+ gnu_result = DECL_INITIAL (get_gnu_tree (Entity (gnat_node)));
+ else
+ gnu_result
+ = build_int_cst (gnu_result_type,
+ UI_To_CC (Char_Literal_Value (gnat_node)));
+ break;
+
+ case N_Real_Literal:
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+
+ /* If this is of a fixed-point type, the value we want is the value of
+ the corresponding integer. */
+ if (Is_Fixed_Point_Type (Underlying_Type (Etype (gnat_node))))
+ {
+ gnu_result = UI_To_gnu (Corresponding_Integer_Value (gnat_node),
+ gnu_result_type);
+ gcc_assert (!TREE_OVERFLOW (gnu_result));
+ }
+
+ else
+ {
+ Ureal ur_realval = Realval (gnat_node);
+
+ /* First convert the value to a machine number if it isn't already.
+ That will force the base to 2 for non-zero values and simplify
+ the rest of the logic. */
+ if (!Is_Machine_Number (gnat_node))
+ ur_realval
+ = Machine (Base_Type (Underlying_Type (Etype (gnat_node))),
+ ur_realval, Round_Even, gnat_node);
+
+ if (UR_Is_Zero (ur_realval))
+ gnu_result = build_real (gnu_result_type, dconst0);
+ else
+ {
+ REAL_VALUE_TYPE tmp;
+
+ gnu_result = UI_To_gnu (Numerator (ur_realval), gnu_result_type);
+
+ /* The base must be 2 as Machine guarantees this, so we scale
+ the value, which we know can fit in the mantissa of the type
+ (hence the use of that type above). */
+ gcc_assert (Rbase (ur_realval) == 2);
+ real_ldexp (&tmp, &TREE_REAL_CST (gnu_result),
+ - UI_To_Int (Denominator (ur_realval)));
+ gnu_result = build_real (gnu_result_type, tmp);
+ }
+
+ /* Now see if we need to negate the result. Do it this way to
+ properly handle -0. */
+ if (UR_Is_Negative (Realval (gnat_node)))
+ gnu_result
+ = build_unary_op (NEGATE_EXPR, get_base_type (gnu_result_type),
+ gnu_result);
+ }
+
+ break;
+
+ case N_String_Literal:
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ if (TYPE_PRECISION (TREE_TYPE (gnu_result_type)) == HOST_BITS_PER_CHAR)
+ {
+ String_Id gnat_string = Strval (gnat_node);
+ int length = String_Length (gnat_string);
+ int i;
+ char *string;
+ if (length >= ALLOCA_THRESHOLD)
+ string = XNEWVEC (char, length);
+ else
+ string = (char *) alloca (length);
+
+ /* Build the string with the characters in the literal. Note
+ that Ada strings are 1-origin. */
+ for (i = 0; i < length; i++)
+ string[i] = Get_String_Char (gnat_string, i + 1);
+
+ gnu_result = build_string (length, string);
+
+ /* Strings in GCC don't normally have types, but we want
+ this to not be converted to the array type. */
+ TREE_TYPE (gnu_result) = gnu_result_type;
+
+ if (length >= ALLOCA_THRESHOLD)
+ free (string);
+ }
+ else
+ {
+ /* Build a list consisting of each character, then make
+ the aggregate. */
+ String_Id gnat_string = Strval (gnat_node);
+ int length = String_Length (gnat_string);
+ int i;
+ tree gnu_idx = TYPE_MIN_VALUE (TYPE_DOMAIN (gnu_result_type));
+ tree gnu_one_node = convert (TREE_TYPE (gnu_idx), integer_one_node);
+ vec<constructor_elt, va_gc> *gnu_vec;
+ vec_alloc (gnu_vec, length);
+
+ for (i = 0; i < length; i++)
+ {
+ tree t = build_int_cst (TREE_TYPE (gnu_result_type),
+ Get_String_Char (gnat_string, i + 1));
+
+ CONSTRUCTOR_APPEND_ELT (gnu_vec, gnu_idx, t);
+ gnu_idx = int_const_binop (PLUS_EXPR, gnu_idx, gnu_one_node);
+ }
+
+ gnu_result = gnat_build_constructor (gnu_result_type, gnu_vec);
+ }
+ break;
+
+ case N_Pragma:
+ gnu_result = Pragma_to_gnu (gnat_node);
+ break;
+
+ /**************************************/
+ /* Chapter 3: Declarations and Types */
+ /**************************************/
+
+ case N_Subtype_Declaration:
+ case N_Full_Type_Declaration:
+ case N_Incomplete_Type_Declaration:
+ case N_Private_Type_Declaration:
+ case N_Private_Extension_Declaration:
+ case N_Task_Type_Declaration:
+ process_type (Defining_Entity (gnat_node));
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ case N_Object_Declaration:
+ case N_Number_Declaration:
+ case N_Exception_Declaration:
+ gnat_temp = Defining_Entity (gnat_node);
+ gnu_result = alloc_stmt_list ();
+
+ /* If we are just annotating types and this object has an unconstrained
+ or task type, don't elaborate it. */
+ if (type_annotate_only
+ && (((Is_Array_Type (Etype (gnat_temp))
+ || Is_Record_Type (Etype (gnat_temp)))
+ && !Is_Constrained (Etype (gnat_temp)))
+ || Is_Concurrent_Type (Etype (gnat_temp))))
+ break;
+
+ if (Present (Expression (gnat_node))
+ && !(kind == N_Object_Declaration && No_Initialization (gnat_node))
+ && (!type_annotate_only
+ || Compile_Time_Known_Value (Expression (gnat_node))))
+ {
+ gigi_checking_assert (!Do_Range_Check (Expression (gnat_node)));
+
+ gnu_expr = gnat_to_gnu (Expression (gnat_node));
+
+ /* First deal with erroneous expressions. */
+ if (TREE_CODE (gnu_expr) == ERROR_MARK)
+ {
+ /* If this is a named number for which we cannot manipulate
+ the value, just skip the declaration altogether. */
+ if (kind == N_Number_Declaration)
+ break;
+ else if (type_annotate_only)
+ gnu_expr = NULL_TREE;
+ }
+
+ /* Then a special case: we do not want the SLOC of the expression
+ of the tag to pop up every time it is referenced somewhere. */
+ else if (EXPR_P (gnu_expr) && Is_Tag (gnat_temp))
+ SET_EXPR_LOCATION (gnu_expr, UNKNOWN_LOCATION);
+ }
+ else
+ gnu_expr = NULL_TREE;
+
+ /* If this is a deferred constant with an address clause, we ignore the
+ full view since the clause is on the partial view and we cannot have
+ 2 different GCC trees for the object. The only bits of the full view
+ we will use is the initializer, but it will be directly fetched. */
+ if (Ekind (gnat_temp) == E_Constant
+ && Present (Address_Clause (gnat_temp))
+ && Present (Full_View (gnat_temp)))
+ save_gnu_tree (Full_View (gnat_temp), error_mark_node, true);
+
+ /* If this object has its elaboration delayed, we must force evaluation
+ of GNU_EXPR now and save it for the freeze point. Note that we need
+ not do anything special at the global level since the lifetime of the
+ temporary is fully contained within the elaboration routine. */
+ if (Present (Freeze_Node (gnat_temp)))
+ {
+ if (gnu_expr)
+ {
+ gnu_result = gnat_save_expr (gnu_expr);
+ save_gnu_tree (gnat_node, gnu_result, true);
+ }
+ }
+ else
+ gnat_to_gnu_entity (gnat_temp, gnu_expr, true);
+ break;
+
+ case N_Object_Renaming_Declaration:
+ gnat_temp = Defining_Entity (gnat_node);
+ gnu_result = alloc_stmt_list ();
+
+ /* Don't do anything if this renaming is handled by the front end and it
+ does not need debug info. Note that we consider renamings don't need
+ debug info when optimizing: our way to describe them has a
+ memory/elaboration footprint.
+
+ Don't do anything neither if we are just annotating types and this
+ object has a composite or task type, don't elaborate it. */
+ if ((!Is_Renaming_Of_Object (gnat_temp)
+ || (Needs_Debug_Info (gnat_temp)
+ && !optimize
+ && can_materialize_object_renaming_p
+ (Renamed_Object (gnat_temp))))
+ && ! (type_annotate_only
+ && (Is_Array_Type (Etype (gnat_temp))
+ || Is_Record_Type (Etype (gnat_temp))
+ || Is_Concurrent_Type (Etype (gnat_temp)))))
+ gnat_to_gnu_entity (gnat_temp,
+ gnat_to_gnu (Renamed_Object (gnat_temp)),
+ true);
+ break;
+
+ case N_Exception_Renaming_Declaration:
+ gnat_temp = Defining_Entity (gnat_node);
+ gnu_result = alloc_stmt_list ();
+
+ if (Present (Renamed_Entity (gnat_temp)))
+ gnat_to_gnu_entity (gnat_temp,
+ gnat_to_gnu (Renamed_Entity (gnat_temp)),
+ true);
+ break;
+
+ case N_Subprogram_Renaming_Declaration:
+ {
+ const Node_Id gnat_renaming = Defining_Entity (gnat_node);
+ const Node_Id gnat_renamed = Renamed_Entity (gnat_renaming);
+
+ gnu_result = alloc_stmt_list ();
+
+ /* Materializing renamed subprograms will only benefit the debugging
+ information as they aren't referenced in the generated code. So
+ skip them when they aren't needed. Avoid doing this if:
+
+ - there is a freeze node: in this case the renamed entity is not
+ elaborated yet,
+ - the renamed subprogram is intrinsic: it will not be available in
+ the debugging information (note that both or only one of the
+ renaming and the renamed subprograms can be intrinsic). */
+ if (!type_annotate_only
+ && Needs_Debug_Info (gnat_renaming)
+ && No (Freeze_Node (gnat_renaming))
+ && Present (gnat_renamed)
+ && (Ekind (gnat_renamed) == E_Function
+ || Ekind (gnat_renamed) == E_Procedure)
+ && !Is_Intrinsic_Subprogram (gnat_renaming)
+ && !Is_Intrinsic_Subprogram (gnat_renamed))
+ gnat_to_gnu_entity (gnat_renaming, gnat_to_gnu (gnat_renamed), true);
+ break;
+ }
+
+ case N_Implicit_Label_Declaration:
+ gnat_to_gnu_entity (Defining_Entity (gnat_node), NULL_TREE, true);
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ case N_Package_Renaming_Declaration:
+ /* These are fully handled in the front end. */
+ /* ??? For package renamings, find a way to use GENERIC namespaces so
+ that we get proper debug information for them. */
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ /*************************************/
+ /* Chapter 4: Names and Expressions */
+ /*************************************/
+
+ case N_Explicit_Dereference:
+ /* Make sure the designated type is complete before dereferencing. */
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ gnu_result = gnat_to_gnu (Prefix (gnat_node));
+ gnu_result = build_unary_op (INDIRECT_REF, NULL_TREE, gnu_result);
+
+ /* If atomic access is required on the RHS, build the atomic load. */
+ if (simple_atomic_access_required_p (gnat_node, &aa_sync)
+ && !present_in_lhs_or_actual_p (gnat_node))
+ gnu_result = build_atomic_load (gnu_result, aa_sync);
+ break;
+
+ case N_Indexed_Component:
+ {
+ tree gnu_array_object = gnat_to_gnu ((Prefix (gnat_node)));
+ tree gnu_type;
+ int ndim, i;
+ Node_Id *gnat_expr_array;
+
+ gnu_array_object = maybe_padded_object (gnu_array_object);
+ gnu_array_object = maybe_unconstrained_array (gnu_array_object);
+
+ /* Convert vector inputs to their representative array type, to fit
+ what the code below expects. */
+ if (VECTOR_TYPE_P (TREE_TYPE (gnu_array_object)))
+ {
+ if (present_in_lhs_or_actual_p (gnat_node))
+ gnat_mark_addressable (gnu_array_object);
+ gnu_array_object = maybe_vector_array (gnu_array_object);
+ }
+
+ /* The failure of this assertion will very likely come from a missing
+ expansion for a packed array access. */
+ gcc_assert (TREE_CODE (TREE_TYPE (gnu_array_object)) == ARRAY_TYPE);
+
+ /* First compute the number of dimensions of the array, then
+ fill the expression array, the order depending on whether
+ this is a Convention_Fortran array or not. */
+ for (ndim = 1, gnu_type = TREE_TYPE (gnu_array_object);
+ TREE_CODE (TREE_TYPE (gnu_type)) == ARRAY_TYPE
+ && TYPE_MULTI_ARRAY_P (TREE_TYPE (gnu_type));
+ ndim++, gnu_type = TREE_TYPE (gnu_type))
+ ;
+
+ gnat_expr_array = XALLOCAVEC (Node_Id, ndim);
+
+ if (TYPE_CONVENTION_FORTRAN_P (TREE_TYPE (gnu_array_object)))
+ for (i = ndim - 1, gnat_temp = First (Expressions (gnat_node));
+ i >= 0;
+ i--, gnat_temp = Next (gnat_temp))
+ gnat_expr_array[i] = gnat_temp;
+ else
+ for (i = 0, gnat_temp = First (Expressions (gnat_node));
+ i < ndim;
+ i++, gnat_temp = Next (gnat_temp))
+ gnat_expr_array[i] = gnat_temp;
+
+ /* Start with the prefix and build the successive references. */
+ gnu_result = gnu_array_object;
+
+ for (i = 0, gnu_type = TREE_TYPE (gnu_array_object);
+ i < ndim;
+ i++, gnu_type = TREE_TYPE (gnu_type))
+ {
+ gcc_assert (TREE_CODE (gnu_type) == ARRAY_TYPE);
+ gnat_temp = gnat_expr_array[i];
+ gnu_expr = maybe_character_value (gnat_to_gnu (gnat_temp));
+
+ gnu_result
+ = build_binary_op (ARRAY_REF, NULL_TREE, gnu_result, gnu_expr);
+ }
+
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+
+ /* If atomic access is required on the RHS, build the atomic load. */
+ if (simple_atomic_access_required_p (gnat_node, &aa_sync)
+ && !present_in_lhs_or_actual_p (gnat_node))
+ gnu_result = build_atomic_load (gnu_result, aa_sync);
+ }
+ break;
+
+ case N_Slice:
+ {
+ tree gnu_array_object = gnat_to_gnu (Prefix (gnat_node));
+
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+
+ gnu_array_object = maybe_padded_object (gnu_array_object);
+ gnu_array_object = maybe_unconstrained_array (gnu_array_object);
+
+ gnu_expr = TYPE_MIN_VALUE (TYPE_DOMAIN (gnu_result_type));
+ gnu_expr = maybe_character_value (gnu_expr);
+
+ /* If this is a slice with non-constant size of an array with constant
+ size, set the maximum size for the allocation of temporaries. */
+ if (!TREE_CONSTANT (TYPE_SIZE_UNIT (gnu_result_type))
+ && TREE_CONSTANT (TYPE_SIZE_UNIT (TREE_TYPE (gnu_array_object))))
+ TYPE_ARRAY_MAX_SIZE (gnu_result_type)
+ = TYPE_SIZE_UNIT (TREE_TYPE (gnu_array_object));
+
+ gnu_result = build_binary_op (ARRAY_RANGE_REF, gnu_result_type,
+ gnu_array_object, gnu_expr);
+ }
+ break;
+
+ case N_Selected_Component:
+ {
+ const Entity_Id gnat_prefix = Prefix (gnat_node);
+ Entity_Id gnat_field = Entity (Selector_Name (gnat_node));
+ tree gnu_prefix = gnat_to_gnu (gnat_prefix);
+
+ gnu_prefix = maybe_padded_object (gnu_prefix);
+
+ /* gnat_to_gnu_entity does not save the GNU tree made for renamed
+ discriminants so avoid making recursive calls on each reference
+ to them by following the appropriate link directly here. */
+ if (Ekind (gnat_field) == E_Discriminant)
+ {
+ /* For discriminant references in tagged types always substitute
+ the corresponding discriminant as the actual component. */
+ if (Is_Tagged_Type (Underlying_Type (Etype (gnat_prefix))))
+ while (Present (Corresponding_Discriminant (gnat_field)))
+ gnat_field = Corresponding_Discriminant (gnat_field);
+
+ /* For discriminant references in untagged types always substitute
+ the corresponding stored discriminant. */
+ else if (Present (Corresponding_Discriminant (gnat_field)))
+ gnat_field = Original_Record_Component (gnat_field);
+ }
+
+ /* Handle extracting the real or imaginary part of a complex.
+ The real part is the first field and the imaginary the last. */
+ if (TREE_CODE (TREE_TYPE (gnu_prefix)) == COMPLEX_TYPE)
+ gnu_result = build_unary_op (Present (Next_Entity (gnat_field))
+ ? REALPART_EXPR : IMAGPART_EXPR,
+ NULL_TREE, gnu_prefix);
+ else
+ {
+ tree gnu_field = gnat_to_gnu_field_decl (gnat_field);
+ tree gnu_offset;
+ struct loop_info_d *loop;
+
+ gnu_result
+ = build_component_ref (gnu_prefix, gnu_field,
+ (Nkind (Parent (gnat_node))
+ == N_Attribute_Reference)
+ && lvalue_required_for_attribute_p
+ (Parent (gnat_node)));
+
+ /* If optimization is enabled and we are inside a loop, we try to
+ hoist nonconstant but invariant offset computations outside of
+ the loop, since they very likely contain loads that could turn
+ out to be hard to move if they end up in active EH regions. */
+ if (optimize
+ && inside_loop_p ()
+ && TREE_CODE (gnu_result) == COMPONENT_REF
+ && (gnu_offset = component_ref_field_offset (gnu_result))
+ && !TREE_CONSTANT (gnu_offset)
+ && (gnu_offset = gnat_invariant_expr (gnu_offset))
+ && (loop = find_loop ()))
+ {
+ tree invariant
+ = build1 (SAVE_EXPR, TREE_TYPE (gnu_offset), gnu_offset);
+ vec_safe_push (loop->invariants, invariant);
+ tree field = TREE_OPERAND (gnu_result, 1);
+ tree factor
+ = size_int (DECL_OFFSET_ALIGN (field) / BITS_PER_UNIT);
+ /* Divide the offset by its alignment. */
+ TREE_OPERAND (gnu_result, 2)
+ = size_binop (EXACT_DIV_EXPR, invariant, factor);
+ }
+ }
+
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+
+ /* If atomic access is required on the RHS, build the atomic load. */
+ if (simple_atomic_access_required_p (gnat_node, &aa_sync)
+ && !present_in_lhs_or_actual_p (gnat_node))
+ gnu_result = build_atomic_load (gnu_result, aa_sync);
+ }
+ break;
+
+ case N_Attribute_Reference:
+ {
+ /* The attribute designator. */
+ const int attr = Get_Attribute_Id (Attribute_Name (gnat_node));
+
+ /* The Elab_Spec and Elab_Body attributes are special in that Prefix
+ is a unit, not an object with a GCC equivalent. */
+ if (attr == Attr_Elab_Spec || attr == Attr_Elab_Body)
+ return
+ create_subprog_decl (create_concat_name
+ (Entity (Prefix (gnat_node)),
+ attr == Attr_Elab_Body ? "elabb" : "elabs"),
+ NULL_TREE, void_ftype, NULL_TREE, is_default,
+ true, true, true, true, false, NULL,
+ gnat_node);
+
+ gnu_result = Attribute_to_gnu (gnat_node, &gnu_result_type, attr);
+ }
+ break;
+
+ case N_Reference:
+ /* Like 'Access as far as we are concerned. */
+ gnu_result = gnat_to_gnu (Prefix (gnat_node));
+ gnu_result = build_unary_op (ADDR_EXPR, NULL_TREE, gnu_result);
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ break;
+
+ case N_Aggregate:
+ case N_Extension_Aggregate:
+ {
+ tree gnu_aggr_type;
+
+ /* Check that this aggregate has not slipped through the cracks. */
+ gcc_assert (!Expansion_Delayed (gnat_node));
+
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+
+ if (TREE_CODE (gnu_result_type) == RECORD_TYPE
+ && TYPE_CONTAINS_TEMPLATE_P (gnu_result_type))
+ gnu_aggr_type
+ = TREE_TYPE (DECL_CHAIN (TYPE_FIELDS (gnu_result_type)));
+ else if (TREE_CODE (gnu_result_type) == VECTOR_TYPE)
+ gnu_aggr_type = TYPE_REPRESENTATIVE_ARRAY (gnu_result_type);
+ else
+ gnu_aggr_type = gnu_result_type;
+
+ if (Null_Record_Present (gnat_node))
+ gnu_result = gnat_build_constructor (gnu_aggr_type, NULL);
+
+ else if (TREE_CODE (gnu_aggr_type) == RECORD_TYPE
+ || TREE_CODE (gnu_aggr_type) == UNION_TYPE)
+ gnu_result
+ = assoc_to_constructor (Etype (gnat_node),
+ First (Component_Associations (gnat_node)),
+ gnu_aggr_type);
+ else if (TREE_CODE (gnu_aggr_type) == ARRAY_TYPE)
+ gnu_result = pos_to_constructor (First (Expressions (gnat_node)),
+ gnu_aggr_type);
+ else if (TREE_CODE (gnu_aggr_type) == COMPLEX_TYPE)
+ gnu_result
+ = build_binary_op
+ (COMPLEX_EXPR, gnu_aggr_type,
+ gnat_to_gnu (Expression (First
+ (Component_Associations (gnat_node)))),
+ gnat_to_gnu (Expression
+ (Next
+ (First (Component_Associations (gnat_node))))));
+ else
+ gcc_unreachable ();
+
+ gnu_result = convert (gnu_result_type, gnu_result);
+ }
+ break;
+
+ case N_Null:
+ if (TARGET_VTABLE_USES_DESCRIPTORS
+ && Ekind (Etype (gnat_node)) == E_Access_Subprogram_Type
+ && Is_Dispatch_Table_Entity (Etype (gnat_node)))
+ gnu_result = null_fdesc_node;
+ else
+ gnu_result = null_pointer_node;
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ break;
+
+ case N_Type_Conversion:
+ case N_Qualified_Expression:
+ gnu_expr = maybe_character_value (gnat_to_gnu (Expression (gnat_node)));
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+
+ /* If this is a qualified expression for a tagged type, we mark the type
+ as used. Because of polymorphism, this might be the only reference to
+ the tagged type in the program while objects have it as dynamic type.
+ The debugger needs to see it to display these objects properly. */
+ if (kind == N_Qualified_Expression && Is_Tagged_Type (Etype (gnat_node)))
+ used_types_insert (gnu_result_type);
+
+ gigi_checking_assert (!Do_Range_Check (Expression (gnat_node)));
+
+ gnu_result
+ = convert_with_check (Etype (gnat_node), gnu_expr,
+ Do_Overflow_Check (gnat_node),
+ kind == N_Type_Conversion
+ && Float_Truncate (gnat_node), gnat_node);
+ break;
+
+ case N_Unchecked_Type_Conversion:
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ gnu_expr = maybe_character_value (gnat_to_gnu (Expression (gnat_node)));
+
+ /* Skip further processing if the conversion is deemed a no-op. */
+ if (unchecked_conversion_nop (gnat_node))
+ {
+ gnu_result = gnu_expr;
+ gnu_result_type = TREE_TYPE (gnu_result);
+ break;
+ }
+
+ /* If the result is a pointer type, see if we are improperly
+ converting to a stricter alignment. */
+ if (STRICT_ALIGNMENT && POINTER_TYPE_P (gnu_result_type)
+ && Is_Access_Type (Etype (gnat_node)))
+ {
+ unsigned int align = known_alignment (gnu_expr);
+ tree gnu_obj_type = TREE_TYPE (gnu_result_type);
+ unsigned int oalign = TYPE_ALIGN (gnu_obj_type);
+
+ if (align != 0 && align < oalign && !TYPE_ALIGN_OK (gnu_obj_type))
+ post_error_ne_tree_2
+ ("??source alignment (^) '< alignment of & (^)",
+ gnat_node, Designated_Type (Etype (gnat_node)),
+ size_int (align / BITS_PER_UNIT), oalign / BITS_PER_UNIT);
+ }
+
+ /* If we are converting a descriptor to a function pointer, first
+ build the pointer. */
+ if (TARGET_VTABLE_USES_DESCRIPTORS
+ && TREE_TYPE (gnu_expr) == fdesc_type_node
+ && POINTER_TYPE_P (gnu_result_type))
+ gnu_expr = build_unary_op (ADDR_EXPR, NULL_TREE, gnu_expr);
+
+ gnu_result = unchecked_convert (gnu_result_type, gnu_expr,
+ No_Truncation (gnat_node));
+ break;
+
+ case N_In:
+ case N_Not_In:
+ {
+ tree gnu_obj = gnat_to_gnu (Left_Opnd (gnat_node));
+ tree gnu_low, gnu_high;
+
+ Range_to_gnu (Right_Opnd (gnat_node), &gnu_low, &gnu_high);
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+
+ tree gnu_op_type = maybe_character_type (TREE_TYPE (gnu_obj));
+ if (TREE_TYPE (gnu_obj) != gnu_op_type)
+ {
+ gnu_obj = convert (gnu_op_type, gnu_obj);
+ gnu_low = convert (gnu_op_type, gnu_low);
+ gnu_high = convert (gnu_op_type, gnu_high);
+ }
+
+ /* If LOW and HIGH are identical, perform an equality test. Otherwise,
+ ensure that GNU_OBJ is evaluated only once and perform a full range
+ test. */
+ if (operand_equal_p (gnu_low, gnu_high, 0))
+ gnu_result
+ = build_binary_op (EQ_EXPR, gnu_result_type, gnu_obj, gnu_low);
+ else
+ {
+ tree t1, t2;
+ gnu_obj = gnat_protect_expr (gnu_obj);
+ t1 = build_binary_op (GE_EXPR, gnu_result_type, gnu_obj, gnu_low);
+ if (EXPR_P (t1))
+ set_expr_location_from_node (t1, gnat_node);
+ t2 = build_binary_op (LE_EXPR, gnu_result_type, gnu_obj, gnu_high);
+ if (EXPR_P (t2))
+ set_expr_location_from_node (t2, gnat_node);
+ gnu_result
+ = build_binary_op (TRUTH_ANDIF_EXPR, gnu_result_type, t1, t2);
+ }
+
+ if (kind == N_Not_In)
+ gnu_result
+ = invert_truthvalue_loc (EXPR_LOCATION (gnu_result), gnu_result);
+ }
+ break;
+
+ case N_Op_Divide:
+ gnu_lhs = gnat_to_gnu (Left_Opnd (gnat_node));
+ gnu_rhs = gnat_to_gnu (Right_Opnd (gnat_node));
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ gnu_result = build_binary_op (FLOAT_TYPE_P (gnu_result_type)
+ ? RDIV_EXPR
+ : (Rounded_Result (gnat_node)
+ ? ROUND_DIV_EXPR : TRUNC_DIV_EXPR),
+ gnu_result_type, gnu_lhs, gnu_rhs);
+ break;
+
+ case N_Op_Eq:
+ case N_Op_Ne:
+ case N_Op_Lt:
+ case N_Op_Le:
+ case N_Op_Gt:
+ case N_Op_Ge:
+ case N_Op_Add:
+ case N_Op_Subtract:
+ case N_Op_Multiply:
+ case N_Op_Mod:
+ case N_Op_Rem:
+ case N_Op_Rotate_Left:
+ case N_Op_Rotate_Right:
+ case N_Op_Shift_Left:
+ case N_Op_Shift_Right:
+ case N_Op_Shift_Right_Arithmetic:
+ case N_Op_And:
+ case N_Op_Or:
+ case N_Op_Xor:
+ case N_And_Then:
+ case N_Or_Else:
+ {
+ enum tree_code code = gnu_codes[kind];
+ bool ignore_lhs_overflow = false;
+ location_t saved_location = input_location;
+ tree gnu_type, gnu_max_shift = NULL_TREE;
+
+ /* Fix operations set up for boolean types in GNU_CODES above. */
+ if (Is_Modular_Integer_Type (Underlying_Type (Etype (gnat_node))))
+ switch (kind)
+ {
+ case N_Op_And:
+ code = BIT_AND_EXPR;
+ break;
+ case N_Op_Or:
+ code = BIT_IOR_EXPR;
+ break;
+ case N_Op_Xor:
+ code = BIT_XOR_EXPR;
+ break;
+ default:
+ break;
+ }
+
+ gnu_lhs = gnat_to_gnu (Left_Opnd (gnat_node));
+ gnu_rhs = gnat_to_gnu (Right_Opnd (gnat_node));
+ gnu_type = gnu_result_type = get_unpadded_type (Etype (gnat_node));
+
+ /* If this is a shift, take the count as unsigned since that is what
+ most machines do and will generate simpler adjustments below. */
+ if (IN (kind, N_Op_Shift))
+ {
+ tree gnu_count_type
+ = gnat_unsigned_type_for (get_base_type (TREE_TYPE (gnu_rhs)));
+ gnu_rhs = convert (gnu_count_type, gnu_rhs);
+ gnu_max_shift
+ = convert (TREE_TYPE (gnu_rhs), TYPE_SIZE (gnu_type));
+ }
+
+ /* Pending generic support for efficient vector logical operations in
+ GCC, convert vectors to their representative array type view and
+ fallthrough. */
+ gnu_lhs = maybe_vector_array (gnu_lhs);
+ gnu_rhs = maybe_vector_array (gnu_rhs);
+
+ /* If this is a comparison operator, convert any references to an
+ unconstrained array value into a reference to the actual array. */
+ if (TREE_CODE_CLASS (code) == tcc_comparison)
+ {
+ gnu_lhs = maybe_unconstrained_array (gnu_lhs);
+ gnu_rhs = maybe_unconstrained_array (gnu_rhs);
+
+ tree gnu_op_type = maybe_character_type (TREE_TYPE (gnu_lhs));
+ if (TREE_TYPE (gnu_lhs) != gnu_op_type)
+ {
+ gnu_lhs = convert (gnu_op_type, gnu_lhs);
+ gnu_rhs = convert (gnu_op_type, gnu_rhs);
+ }
+ }
+
+ /* If this is a shift whose count is not guaranteed to be correct,
+ we need to adjust the shift count. */
+ if ((kind == N_Op_Rotate_Left || kind == N_Op_Rotate_Right)
+ && !Shift_Count_OK (gnat_node))
+ gnu_rhs = build_binary_op (TRUNC_MOD_EXPR, TREE_TYPE (gnu_rhs),
+ gnu_rhs, gnu_max_shift);
+ else if (kind == N_Op_Shift_Right_Arithmetic
+ && !Shift_Count_OK (gnat_node))
+ gnu_rhs
+ = build_binary_op (MIN_EXPR, TREE_TYPE (gnu_rhs),
+ build_binary_op (MINUS_EXPR,
+ TREE_TYPE (gnu_rhs),
+ gnu_max_shift,
+ build_int_cst
+ (TREE_TYPE (gnu_rhs), 1)),
+ gnu_rhs);
+
+ /* For right shifts, the type says what kind of shift to do,
+ so we may need to choose a different type. In this case,
+ we have to ignore integer overflow lest it propagates all
+ the way down and causes a CE to be explicitly raised. */
+ if (kind == N_Op_Shift_Right && !TYPE_UNSIGNED (gnu_type))
+ {
+ gnu_type = gnat_unsigned_type_for (gnu_type);
+ ignore_lhs_overflow = true;
+ }
+ else if (kind == N_Op_Shift_Right_Arithmetic
+ && TYPE_UNSIGNED (gnu_type))
+ {
+ gnu_type = gnat_signed_type_for (gnu_type);
+ ignore_lhs_overflow = true;
+ }
+
+ if (gnu_type != gnu_result_type)
+ {
+ tree gnu_old_lhs = gnu_lhs;
+ gnu_lhs = convert (gnu_type, gnu_lhs);
+ if (TREE_CODE (gnu_lhs) == INTEGER_CST && ignore_lhs_overflow)
+ TREE_OVERFLOW (gnu_lhs) = TREE_OVERFLOW (gnu_old_lhs);
+ gnu_rhs = convert (gnu_type, gnu_rhs);
+ if (gnu_max_shift)
+ gnu_max_shift = convert (gnu_type, gnu_max_shift);
+ }
+
+ /* For signed integer addition, subtraction and multiplication, do an
+ overflow check if required. */
+ if (Do_Overflow_Check (gnat_node)
+ && (code == PLUS_EXPR || code == MINUS_EXPR || code == MULT_EXPR)
+ && !TYPE_UNSIGNED (gnu_type)
+ && !FLOAT_TYPE_P (gnu_type))
+ gnu_result
+ = build_binary_op_trapv (code, gnu_type, gnu_lhs, gnu_rhs,
+ gnat_node);
+ else
+ {
+ /* Some operations, e.g. comparisons of arrays, generate complex
+ trees that need to be annotated while they are being built. */
+ input_location = saved_location;
+ gnu_result = build_binary_op (code, gnu_type, gnu_lhs, gnu_rhs);
+ }
+
+ /* If this is a logical shift with the shift count not verified,
+ we must return zero if it is too large. We cannot compensate
+ beforehand in this case. */
+ if ((kind == N_Op_Shift_Left || kind == N_Op_Shift_Right)
+ && !Shift_Count_OK (gnat_node))
+ gnu_result
+ = build_cond_expr (gnu_type,
+ build_binary_op (GE_EXPR, boolean_type_node,
+ gnu_rhs, gnu_max_shift),
+ build_int_cst (gnu_type, 0),
+ gnu_result);
+ }
+ break;
+
+ case N_If_Expression:
+ {
+ tree gnu_cond = gnat_to_gnu (First (Expressions (gnat_node)));
+ tree gnu_true = gnat_to_gnu (Next (First (Expressions (gnat_node))));
+ tree gnu_false
+ = gnat_to_gnu (Next (Next (First (Expressions (gnat_node)))));
+
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ gnu_result
+ = build_cond_expr (gnu_result_type, gnu_cond, gnu_true, gnu_false);
+ }
+ break;
+
+ case N_Op_Plus:
+ gnu_result = gnat_to_gnu (Right_Opnd (gnat_node));
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ break;
+
+ case N_Op_Not:
+ /* This case can apply to a boolean or a modular type.
+ Fall through for a boolean operand since GNU_CODES is set
+ up to handle this. */
+ if (Is_Modular_Integer_Type (Underlying_Type (Etype (gnat_node))))
+ {
+ gnu_expr = gnat_to_gnu (Right_Opnd (gnat_node));
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ gnu_result = build_unary_op (BIT_NOT_EXPR, gnu_result_type,
+ gnu_expr);
+ break;
+ }
+
+ /* ... fall through ... */
+
+ case N_Op_Minus:
+ case N_Op_Abs:
+ gnu_expr = gnat_to_gnu (Right_Opnd (gnat_node));
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+
+ /* For signed integer negation and absolute value, do an overflow check
+ if required. */
+ if (Do_Overflow_Check (gnat_node)
+ && !TYPE_UNSIGNED (gnu_result_type)
+ && !FLOAT_TYPE_P (gnu_result_type))
+ gnu_result
+ = build_unary_op_trapv (gnu_codes[kind], gnu_result_type, gnu_expr,
+ gnat_node);
+ else
+ gnu_result
+ = build_unary_op (gnu_codes[kind], gnu_result_type, gnu_expr);
+ break;
+
+ case N_Allocator:
+ {
+ tree gnu_type, gnu_init;
+ bool ignore_init_type;
+
+ gnat_temp = Expression (gnat_node);
+
+ /* The expression can be either an N_Identifier or an Expanded_Name,
+ which must represent a type, or a N_Qualified_Expression, which
+ contains both the type and an initial value for the object. */
+ if (Nkind (gnat_temp) == N_Identifier
+ || Nkind (gnat_temp) == N_Expanded_Name)
+ {
+ ignore_init_type = false;
+ gnu_init = NULL_TREE;
+ gnu_type = gnat_to_gnu_type (Entity (gnat_temp));
+ }
+
+ else if (Nkind (gnat_temp) == N_Qualified_Expression)
+ {
+ const Entity_Id gnat_desig_type
+ = Designated_Type (Underlying_Type (Etype (gnat_node)));
+
+ ignore_init_type = Has_Constrained_Partial_View (gnat_desig_type);
+
+ gnu_init = gnat_to_gnu (Expression (gnat_temp));
+ gnu_init = maybe_unconstrained_array (gnu_init);
+
+ gigi_checking_assert (!Do_Range_Check (Expression (gnat_temp)));
+
+ if (Is_Elementary_Type (gnat_desig_type)
+ || Is_Constrained (gnat_desig_type))
+ gnu_type = gnat_to_gnu_type (gnat_desig_type);
+ else
+ {
+ gnu_type = gnat_to_gnu_type (Etype (Expression (gnat_temp)));
+ if (TREE_CODE (gnu_type) == UNCONSTRAINED_ARRAY_TYPE)
+ gnu_type = TREE_TYPE (gnu_init);
+ }
+
+ /* See the N_Qualified_Expression case for the rationale. */
+ if (Is_Tagged_Type (gnat_desig_type))
+ used_types_insert (gnu_type);
+
+ gnu_init = convert (gnu_type, gnu_init);
+ }
+ else
+ gcc_unreachable ();
+
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ return build_allocator (gnu_type, gnu_init, gnu_result_type,
+ Procedure_To_Call (gnat_node),
+ Storage_Pool (gnat_node), gnat_node,
+ ignore_init_type);
+ }
+ break;
+
+ /**************************/
+ /* Chapter 5: Statements */
+ /**************************/
+
+ case N_Label:
+ gnu_result = build1 (LABEL_EXPR, void_type_node,
+ gnat_to_gnu (Identifier (gnat_node)));
+ break;
+
+ case N_Null_Statement:
+ /* When not optimizing, turn null statements from source into gotos to
+ the next statement that the middle-end knows how to preserve. */
+ if (!optimize && Comes_From_Source (gnat_node))
+ {
+ tree stmt, label = create_label_decl (NULL_TREE, gnat_node);
+ DECL_IGNORED_P (label) = 1;
+ start_stmt_group ();
+ stmt = build1 (GOTO_EXPR, void_type_node, label);
+ set_expr_location_from_node (stmt, gnat_node);
+ add_stmt (stmt);
+ stmt = build1 (LABEL_EXPR, void_type_node, label);
+ set_expr_location_from_node (stmt, gnat_node);
+ add_stmt (stmt);
+ gnu_result = end_stmt_group ();
+ }
+ else
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ case N_Assignment_Statement:
+ /* Get the LHS and RHS of the statement and convert any reference to an
+ unconstrained array into a reference to the underlying array. */
+ gnu_lhs = maybe_unconstrained_array (gnat_to_gnu (Name (gnat_node)));
+
+ /* If the type has a size that overflows, convert this into raise of
+ Storage_Error: execution shouldn't have gotten here anyway. */
+ if (TREE_CODE (TYPE_SIZE_UNIT (TREE_TYPE (gnu_lhs))) == INTEGER_CST
+ && !valid_constant_size_p (TYPE_SIZE_UNIT (TREE_TYPE (gnu_lhs))))
+ gnu_result = build_call_raise (SE_Object_Too_Large, gnat_node,
+ N_Raise_Storage_Error);
+ else if (Nkind (Expression (gnat_node)) == N_Function_Call)
+ {
+ get_atomic_access (Name (gnat_node), &aa_type, &aa_sync);
+ gnu_result
+ = Call_to_gnu (Expression (gnat_node), &gnu_result_type, gnu_lhs,
+ aa_type, aa_sync);
+ }
+ else
+ {
+ const Node_Id gnat_expr = Expression (gnat_node);
+ const Node_Id gnat_inner
+ = Nkind (gnat_expr) == N_Qualified_Expression
+ ? Expression (gnat_expr)
+ : gnat_expr;
+ const Entity_Id gnat_type
+ = Underlying_Type (Etype (Name (gnat_node)));
+ const bool use_memset_p
+ = Is_Array_Type (gnat_type)
+ && Nkind (gnat_inner) == N_Aggregate
+ && Is_Single_Aggregate (gnat_inner);
+
+ /* If we use memset, we need to find the innermost expression. */
+ if (use_memset_p)
+ {
+ gnat_temp = gnat_inner;
+ do {
+ gnat_temp
+ = Expression (First (Component_Associations (gnat_temp)));
+ } while (Nkind (gnat_temp) == N_Aggregate
+ && Is_Single_Aggregate (gnat_temp));
+ gnu_rhs = gnat_to_gnu (gnat_temp);
+ }
+ else
+ gnu_rhs = maybe_unconstrained_array (gnat_to_gnu (gnat_expr));
+
+ gigi_checking_assert (!Do_Range_Check (gnat_expr));
+
+ get_atomic_access (Name (gnat_node), &aa_type, &aa_sync);
+
+ /* If an outer atomic access is required on the LHS, build the load-
+ modify-store sequence. */
+ if (aa_type == OUTER_ATOMIC)
+ gnu_result = build_load_modify_store (gnu_lhs, gnu_rhs, gnat_node);
+
+ /* Or else, if a simple atomic access is required, build the atomic
+ store. */
+ else if (aa_type == SIMPLE_ATOMIC)
+ gnu_result = build_atomic_store (gnu_lhs, gnu_rhs, aa_sync);
+
+ /* Or else, use memset when the conditions are met. This has already
+ been validated by Aggr_Assignment_OK_For_Backend in the front-end
+ and the RHS is thus guaranteed to be of the appropriate form. */
+ else if (use_memset_p)
+ {
+ tree value
+ = real_zerop (gnu_rhs)
+ ? integer_zero_node
+ : fold_convert (integer_type_node, gnu_rhs);
+ tree dest = build_fold_addr_expr (gnu_lhs);
+ tree t = builtin_decl_explicit (BUILT_IN_MEMSET);
+ /* Be extra careful not to write too much data. */
+ tree size;
+ if (TREE_CODE (gnu_lhs) == COMPONENT_REF)
+ size = DECL_SIZE_UNIT (TREE_OPERAND (gnu_lhs, 1));
+ else if (DECL_P (gnu_lhs))
+ size = DECL_SIZE_UNIT (gnu_lhs);
+ else
+ size = TYPE_SIZE_UNIT (TREE_TYPE (gnu_lhs));
+ size = SUBSTITUTE_PLACEHOLDER_IN_EXPR (size, gnu_lhs);
+ if (TREE_CODE (value) == INTEGER_CST && !integer_zerop (value))
+ {
+ tree mask
+ = build_int_cst (integer_type_node,
+ ((HOST_WIDE_INT) 1 << BITS_PER_UNIT) - 1);
+ value = int_const_binop (BIT_AND_EXPR, value, mask);
+ }
+ gnu_result = build_call_expr (t, 3, dest, value, size);
+ }
+
+ /* Otherwise build a regular assignment. */
+ else
+ gnu_result
+ = build_binary_op (MODIFY_EXPR, NULL_TREE, gnu_lhs, gnu_rhs);
+
+ /* If the assignment type is a regular array and the two sides are
+ not completely disjoint, play safe and use memmove. But don't do
+ it for a bit-packed array as it might not be byte-aligned. */
+ if (TREE_CODE (gnu_result) == MODIFY_EXPR
+ && Is_Array_Type (gnat_type)
+ && !Is_Bit_Packed_Array (gnat_type)
+ && !(Forwards_OK (gnat_node) && Backwards_OK (gnat_node)))
+ {
+ tree to = TREE_OPERAND (gnu_result, 0);
+ tree from = TREE_OPERAND (gnu_result, 1);
+ tree type = TREE_TYPE (from);
+ tree size
+ = SUBSTITUTE_PLACEHOLDER_IN_EXPR (TYPE_SIZE_UNIT (type), from);
+ tree to_ptr = build_fold_addr_expr (to);
+ tree from_ptr = build_fold_addr_expr (from);
+ tree t = builtin_decl_explicit (BUILT_IN_MEMMOVE);
+ gnu_result = build_call_expr (t, 3, to_ptr, from_ptr, size);
+ }
+ }
+ break;
+
+ case N_If_Statement:
+ {
+ tree *gnu_else_ptr; /* Point to put next "else if" or "else". */
+
+ /* Make the outer COND_EXPR. Avoid non-determinism. */
+ gnu_result = build3 (COND_EXPR, void_type_node,
+ gnat_to_gnu (Condition (gnat_node)),
+ NULL_TREE, NULL_TREE);
+ COND_EXPR_THEN (gnu_result)
+ = build_stmt_group (Then_Statements (gnat_node), false);
+ TREE_SIDE_EFFECTS (gnu_result) = 1;
+ gnu_else_ptr = &COND_EXPR_ELSE (gnu_result);
+
+ /* Now make a COND_EXPR for each of the "else if" parts. Put each
+ into the previous "else" part and point to where to put any
+ outer "else". Also avoid non-determinism. */
+ if (Present (Elsif_Parts (gnat_node)))
+ for (gnat_temp = First (Elsif_Parts (gnat_node));
+ Present (gnat_temp); gnat_temp = Next (gnat_temp))
+ {
+ gnu_expr = build3 (COND_EXPR, void_type_node,
+ gnat_to_gnu (Condition (gnat_temp)),
+ NULL_TREE, NULL_TREE);
+ COND_EXPR_THEN (gnu_expr)
+ = build_stmt_group (Then_Statements (gnat_temp), false);
+ TREE_SIDE_EFFECTS (gnu_expr) = 1;
+ set_expr_location_from_node (gnu_expr, gnat_temp);
+ *gnu_else_ptr = gnu_expr;
+ gnu_else_ptr = &COND_EXPR_ELSE (gnu_expr);
+ }
+
+ *gnu_else_ptr = build_stmt_group (Else_Statements (gnat_node), false);
+ }
+ break;
+
+ case N_Case_Statement:
+ gnu_result = Case_Statement_to_gnu (gnat_node);
+ break;
+
+ case N_Loop_Statement:
+ gnu_result = Loop_Statement_to_gnu (gnat_node);
+ break;
+
+ case N_Block_Statement:
+ /* The only way to enter the block is to fall through to it. */
+ if (stmt_group_may_fallthru ())
+ {
+ start_stmt_group ();
+ gnat_pushlevel ();
+ process_decls (Declarations (gnat_node), Empty, Empty, true, true);
+ add_stmt (gnat_to_gnu (Handled_Statement_Sequence (gnat_node)));
+ gnat_poplevel ();
+ gnu_result = end_stmt_group ();
+ }
+ else
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ case N_Exit_Statement:
+ gnu_result
+ = build2 (EXIT_STMT, void_type_node,
+ (Present (Condition (gnat_node))
+ ? gnat_to_gnu (Condition (gnat_node)) : NULL_TREE),
+ (Present (Name (gnat_node))
+ ? get_gnu_tree (Entity (Name (gnat_node)))
+ : LOOP_STMT_LABEL (gnu_loop_stack->last ()->stmt)));
+ break;
+
+ case N_Simple_Return_Statement:
+ {
+ tree gnu_ret_obj, gnu_ret_val;
+
+ /* If the subprogram is a function, we must return the expression. */
+ if (Present (Expression (gnat_node)))
+ {
+ tree gnu_subprog_type = TREE_TYPE (current_function_decl);
+
+ /* If this function has copy-in/copy-out parameters parameters and
+ doesn't return by invisible reference, get the real object for
+ the return. See Subprogram_Body_to_gnu. */
+ if (TYPE_CI_CO_LIST (gnu_subprog_type)
+ && !TREE_ADDRESSABLE (gnu_subprog_type))
+ gnu_ret_obj = gnu_return_var_stack->last ();
+ else
+ gnu_ret_obj = DECL_RESULT (current_function_decl);
+
+ /* Get the GCC tree for the expression to be returned. */
+ gnu_ret_val = gnat_to_gnu (Expression (gnat_node));
+
+ /* Do not remove the padding from GNU_RET_VAL if the inner type is
+ self-referential since we want to allocate the fixed size. */
+ if (TREE_CODE (gnu_ret_val) == COMPONENT_REF
+ && type_is_padding_self_referential
+ (TREE_TYPE (TREE_OPERAND (gnu_ret_val, 0))))
+ gnu_ret_val = TREE_OPERAND (gnu_ret_val, 0);
+
+ /* If the function returns by direct reference, return a pointer
+ to the return value. */
+ if (TYPE_RETURN_BY_DIRECT_REF_P (gnu_subprog_type)
+ || By_Ref (gnat_node))
+ gnu_ret_val = build_unary_op (ADDR_EXPR, NULL_TREE, gnu_ret_val);
+
+ /* Otherwise, if it returns an unconstrained array, we have to
+ allocate a new version of the result and return it. */
+ else if (TYPE_RETURN_UNCONSTRAINED_P (gnu_subprog_type))
+ {
+ gnu_ret_val = maybe_unconstrained_array (gnu_ret_val);
+
+ /* And find out whether this is a candidate for Named Return
+ Value. If so, record it. */
+ if (optimize
+ && !optimize_debug
+ && !TYPE_CI_CO_LIST (gnu_subprog_type))
+ {
+ tree ret_val = gnu_ret_val;
+
+ /* Strip useless conversions around the return value. */
+ if (gnat_useless_type_conversion (ret_val))
+ ret_val = TREE_OPERAND (ret_val, 0);
+
+ /* Strip unpadding around the return value. */
+ if (TREE_CODE (ret_val) == COMPONENT_REF
+ && TYPE_IS_PADDING_P
+ (TREE_TYPE (TREE_OPERAND (ret_val, 0))))
+ ret_val = TREE_OPERAND (ret_val, 0);
+
+ /* Now apply the test to the return value. */
+ if (return_value_ok_for_nrv_p (NULL_TREE, ret_val))
+ {
+ if (!f_named_ret_val)
+ f_named_ret_val = BITMAP_GGC_ALLOC ();
+ bitmap_set_bit (f_named_ret_val, DECL_UID (ret_val));
+ if (!f_gnat_ret)
+ f_gnat_ret = gnat_node;
+ }
+ }
+
+ gnu_ret_val = build_allocator (TREE_TYPE (gnu_ret_val),
+ gnu_ret_val,
+ TREE_TYPE (gnu_ret_obj),
+ Procedure_To_Call (gnat_node),
+ Storage_Pool (gnat_node),
+ gnat_node, false);
+ }
+
+ /* Otherwise, if it returns by invisible reference, dereference
+ the pointer it is passed using the type of the return value
+ and build the copy operation manually. This ensures that we
+ don't copy too much data, for example if the return type is
+ unconstrained with a maximum size. */
+ else if (TREE_ADDRESSABLE (gnu_subprog_type))
+ {
+ tree gnu_ret_deref
+ = build_unary_op (INDIRECT_REF, TREE_TYPE (gnu_ret_val),
+ gnu_ret_obj);
+ gnu_result = build2 (INIT_EXPR, void_type_node,
+ gnu_ret_deref, gnu_ret_val);
+ add_stmt_with_node (gnu_result, gnat_node);
+ gnu_ret_val = NULL_TREE;
+ }
+ }
+
+ else
+ gnu_ret_obj = gnu_ret_val = NULL_TREE;
+
+ /* If we have a return label defined, convert this into a branch to
+ that label. The return proper will be handled elsewhere. */
+ if (gnu_return_label_stack->last ())
+ {
+ if (gnu_ret_val)
+ add_stmt_with_node (build_binary_op (MODIFY_EXPR,
+ NULL_TREE, gnu_ret_obj,
+ gnu_ret_val),
+ gnat_node);
+
+ gnu_result = build1 (GOTO_EXPR, void_type_node,
+ gnu_return_label_stack->last ());
+
+ /* When not optimizing, make sure the return is preserved. */
+ if (!optimize && Comes_From_Source (gnat_node))
+ DECL_ARTIFICIAL (gnu_return_label_stack->last ()) = 0;
+ }
+
+ /* Otherwise, build a regular return. */
+ else
+ gnu_result = build_return_expr (gnu_ret_obj, gnu_ret_val);
+ }
+ break;
+
+ case N_Goto_Statement:
+ gnu_expr = gnat_to_gnu (Name (gnat_node));
+ gnu_result = build1 (GOTO_EXPR, void_type_node, gnu_expr);
+ TREE_USED (gnu_expr) = 1;
+ break;
+
+ /***************************/
+ /* Chapter 6: Subprograms */
+ /***************************/
+
+ case N_Subprogram_Declaration:
+ /* Unless there is a freeze node, declare the entity. We consider
+ this a definition even though we're not generating code for the
+ subprogram because we will be making the corresponding GCC node.
+ When there is a freeze node, it is considered the definition of
+ the subprogram and we do nothing until after it is encountered.
+ That's an efficiency issue: the types involved in the profile
+ are far more likely to be frozen between the declaration and
+ the freeze node than before the declaration, so we save some
+ updates of the GCC node by waiting until the freeze node.
+ The counterpart is that we assume that there is no reference
+ to the subprogram between the declaration and the freeze node
+ in the expanded code; otherwise, it will be interpreted as an
+ external reference and very likely give rise to a link failure. */
+ if (No (Freeze_Node (Defining_Entity (Specification (gnat_node)))))
+ gnat_to_gnu_entity (Defining_Entity (Specification (gnat_node)),
+ NULL_TREE, true);
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ case N_Abstract_Subprogram_Declaration:
+ /* This subprogram doesn't exist for code generation purposes, but we
+ have to elaborate the types of any parameters and result, unless
+ they are imported types (nothing to generate in this case).
+
+ The parameter list may contain types with freeze nodes, e.g. not null
+ subtypes, so the subprogram itself may carry a freeze node, in which
+ case its elaboration must be deferred. */
+
+ /* Process the parameter types first. */
+ if (No (Freeze_Node (Defining_Entity (Specification (gnat_node)))))
+ for (gnat_temp
+ = First_Formal_With_Extras
+ (Defining_Entity (Specification (gnat_node)));
+ Present (gnat_temp);
+ gnat_temp = Next_Formal_With_Extras (gnat_temp))
+ if (Is_Itype (Etype (gnat_temp))
+ && !From_Limited_With (Etype (gnat_temp)))
+ gnat_to_gnu_entity (Etype (gnat_temp), NULL_TREE, false);
+
+ /* Then the result type, set to Standard_Void_Type for procedures. */
+ {
+ Entity_Id gnat_temp_type
+ = Etype (Defining_Entity (Specification (gnat_node)));
+
+ if (Is_Itype (gnat_temp_type) && !From_Limited_With (gnat_temp_type))
+ gnat_to_gnu_entity (Etype (gnat_temp_type), NULL_TREE, false);
+ }
+
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ case N_Defining_Program_Unit_Name:
+ /* For a child unit identifier go up a level to get the specification.
+ We get this when we try to find the spec of a child unit package
+ that is the compilation unit being compiled. */
+ gnu_result = gnat_to_gnu (Parent (gnat_node));
+ break;
+
+ case N_Subprogram_Body:
+ Subprogram_Body_to_gnu (gnat_node);
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ case N_Function_Call:
+ case N_Procedure_Call_Statement:
+ gnu_result = Call_to_gnu (gnat_node, &gnu_result_type, NULL_TREE,
+ NOT_ATOMIC, false);
+ break;
+
+ /************************/
+ /* Chapter 7: Packages */
+ /************************/
+
+ case N_Package_Declaration:
+ gnu_result = gnat_to_gnu (Specification (gnat_node));
+ break;
+
+ case N_Package_Specification:
+
+ start_stmt_group ();
+ process_decls (Visible_Declarations (gnat_node),
+ Private_Declarations (gnat_node), Empty, true, true);
+ gnu_result = end_stmt_group ();
+ break;
+
+ case N_Package_Body:
+
+ /* If this is the body of a generic package - do nothing. */
+ if (Ekind (Corresponding_Spec (gnat_node)) == E_Generic_Package)
+ {
+ gnu_result = alloc_stmt_list ();
+ break;
+ }
+
+ start_stmt_group ();
+ process_decls (Declarations (gnat_node), Empty, Empty, true, true);
+
+ if (Present (Handled_Statement_Sequence (gnat_node)))
+ add_stmt (gnat_to_gnu (Handled_Statement_Sequence (gnat_node)));
+
+ gnu_result = end_stmt_group ();
+ break;
+
+ /********************************/
+ /* Chapter 8: Visibility Rules */
+ /********************************/
+
+ case N_Use_Package_Clause:
+ case N_Use_Type_Clause:
+ /* Nothing to do here - but these may appear in list of declarations. */
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ /*********************/
+ /* Chapter 9: Tasks */
+ /*********************/
+
+ case N_Protected_Type_Declaration:
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ case N_Single_Task_Declaration:
+ gnat_to_gnu_entity (Defining_Entity (gnat_node), NULL_TREE, true);
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ /*********************************************************/
+ /* Chapter 10: Program Structure and Compilation Issues */
+ /*********************************************************/
+
+ case N_Compilation_Unit:
+ /* This is not called for the main unit on which gigi is invoked. */
+ Compilation_Unit_to_gnu (gnat_node);
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ case N_Subunit:
+ gnu_result = gnat_to_gnu (Proper_Body (gnat_node));
+ break;
+
+ case N_Entry_Body:
+ case N_Protected_Body:
+ case N_Task_Body:
+ /* These nodes should only be present when annotating types. */
+ gcc_assert (type_annotate_only);
+ process_decls (Declarations (gnat_node), Empty, Empty, true, true);
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ case N_Subprogram_Body_Stub:
+ case N_Package_Body_Stub:
+ case N_Protected_Body_Stub:
+ case N_Task_Body_Stub:
+ /* Simply process whatever unit is being inserted. */
+ if (Present (Library_Unit (gnat_node)))
+ gnu_result = gnat_to_gnu (Unit (Library_Unit (gnat_node)));
+ else
+ {
+ gcc_assert (type_annotate_only);
+ gnu_result = alloc_stmt_list ();
+ }
+ break;
+
+ /***************************/
+ /* Chapter 11: Exceptions */
+ /***************************/
+
+ case N_Handled_Sequence_Of_Statements:
+ /* If there is an At_End procedure attached to this node, and the EH
+ mechanism is front-end, we must have at least a corresponding At_End
+ handler, unless the No_Exception_Handlers restriction is set. */
+ gcc_assert (type_annotate_only
+ || !Front_End_Exceptions ()
+ || No (At_End_Proc (gnat_node))
+ || Present (Exception_Handlers (gnat_node))
+ || No_Exception_Handlers_Set ());
+
+ gnu_result = Handled_Sequence_Of_Statements_to_gnu (gnat_node);
+ break;
+
+ case N_Exception_Handler:
+ if (Back_End_Exceptions ())
+ gnu_result = Exception_Handler_to_gnu_gcc (gnat_node);
+ else if (Exception_Mechanism == Front_End_SJLJ)
+ gnu_result = Exception_Handler_to_gnu_fe_sjlj (gnat_node);
+ else
+ gcc_unreachable ();
+ break;
+
+ case N_Raise_Statement:
+ /* Only for reraise in back-end exceptions mode. */
+ gcc_assert (No (Name (gnat_node)) && Back_End_Exceptions ());
+
+ start_stmt_group ();
+
+ add_stmt_with_node (build_call_n_expr (reraise_zcx_decl, 1,
+ gnu_incoming_exc_ptr),
+ gnat_node);
+
+ gnu_result = end_stmt_group ();
+ break;
+
+ case N_Push_Constraint_Error_Label:
+ gnu_constraint_error_label_stack.safe_push (Exception_Label (gnat_node));
+ break;
+
+ case N_Push_Storage_Error_Label:
+ gnu_storage_error_label_stack.safe_push (Exception_Label (gnat_node));
+ break;
+
+ case N_Push_Program_Error_Label:
+ gnu_program_error_label_stack.safe_push (Exception_Label (gnat_node));
+ break;
+
+ case N_Pop_Constraint_Error_Label:
+ gnat_temp = gnu_constraint_error_label_stack.pop ();
+ if (Present (gnat_temp)
+ && !TREE_USED (gnat_to_gnu_entity (gnat_temp, NULL_TREE, false))
+ && No_Exception_Propagation_Active ())
+ Warn_If_No_Local_Raise (gnat_temp);
+ break;
+
+ case N_Pop_Storage_Error_Label:
+ gnat_temp = gnu_storage_error_label_stack.pop ();
+ if (Present (gnat_temp)
+ && !TREE_USED (gnat_to_gnu_entity (gnat_temp, NULL_TREE, false))
+ && No_Exception_Propagation_Active ())
+ Warn_If_No_Local_Raise (gnat_temp);
+ break;
+
+ case N_Pop_Program_Error_Label:
+ gnat_temp = gnu_program_error_label_stack.pop ();
+ if (Present (gnat_temp)
+ && !TREE_USED (gnat_to_gnu_entity (gnat_temp, NULL_TREE, false))
+ && No_Exception_Propagation_Active ())
+ Warn_If_No_Local_Raise (gnat_temp);
+ break;
+
+ /******************************/
+ /* Chapter 12: Generic Units */
+ /******************************/
+
+ case N_Generic_Function_Renaming_Declaration:
+ case N_Generic_Package_Renaming_Declaration:
+ case N_Generic_Procedure_Renaming_Declaration:
+ case N_Generic_Package_Declaration:
+ case N_Generic_Subprogram_Declaration:
+ case N_Package_Instantiation:
+ case N_Procedure_Instantiation:
+ case N_Function_Instantiation:
+ /* These nodes can appear on a declaration list but there is nothing to
+ to be done with them. */
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ /**************************************************/
+ /* Chapter 13: Representation Clauses and */
+ /* Implementation-Dependent Features */
+ /**************************************************/
+
+ case N_Attribute_Definition_Clause:
+ gnu_result = alloc_stmt_list ();
+
+ /* The only one we need to deal with is 'Address since, for the others,
+ the front-end puts the information elsewhere. */
+ if (Get_Attribute_Id (Chars (gnat_node)) != Attr_Address)
+ break;
+
+ /* And we only deal with 'Address if the object has a Freeze node. */
+ gnat_temp = Entity (Name (gnat_node));
+ if (Freeze_Node (gnat_temp))
+ {
+ tree gnu_address = gnat_to_gnu (Expression (gnat_node)), gnu_temp;
+
+ /* Get the value to use as the address and save it as the equivalent
+ for the object; when it is frozen, gnat_to_gnu_entity will do the
+ right thing. For a subprogram, put the naked address but build a
+ meaningfull expression for an object in case its address is taken
+ before the Freeze node is encountered; this can happen if the type
+ of the object is limited and it is initialized with the result of
+ a function call. */
+ if (Is_Subprogram (gnat_temp))
+ gnu_temp = gnu_address;
+ else
+ {
+ tree gnu_type = gnat_to_gnu_type (Etype (gnat_temp));
+ /* Drop atomic and volatile qualifiers for the expression. */
+ gnu_type = TYPE_MAIN_VARIANT (gnu_type);
+ gnu_type
+ = build_reference_type_for_mode (gnu_type, ptr_mode, true);
+ gnu_address = convert (gnu_type, gnu_address);
+ gnu_temp
+ = build_unary_op (INDIRECT_REF, NULL_TREE, gnu_address);
+ }
+
+ save_gnu_tree (gnat_temp, gnu_temp, true);
+ }
+ break;
+
+ case N_Enumeration_Representation_Clause:
+ case N_Record_Representation_Clause:
+ case N_At_Clause:
+ /* We do nothing with these. SEM puts the information elsewhere. */
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ case N_Code_Statement:
+ if (!type_annotate_only)
+ {
+ tree gnu_template = gnat_to_gnu (Asm_Template (gnat_node));
+ tree gnu_inputs = NULL_TREE, gnu_outputs = NULL_TREE;
+ tree gnu_clobbers = NULL_TREE, tail;
+ bool allows_mem, allows_reg, fake;
+ int ninputs, noutputs, i;
+ const char **oconstraints;
+ const char *constraint;
+ char *clobber;
+
+ /* First retrieve the 3 operand lists built by the front-end. */
+ Setup_Asm_Outputs (gnat_node);
+ while (Present (gnat_temp = Asm_Output_Variable ()))
+ {
+ tree gnu_value = gnat_to_gnu (gnat_temp);
+ tree gnu_constr = build_tree_list (NULL_TREE, gnat_to_gnu
+ (Asm_Output_Constraint ()));
+
+ gnu_outputs = tree_cons (gnu_constr, gnu_value, gnu_outputs);
+ Next_Asm_Output ();
+ }
+
+ Setup_Asm_Inputs (gnat_node);
+ while (Present (gnat_temp = Asm_Input_Value ()))
+ {
+ tree gnu_value = gnat_to_gnu (gnat_temp);
+ tree gnu_constr = build_tree_list (NULL_TREE, gnat_to_gnu
+ (Asm_Input_Constraint ()));
+
+ gnu_inputs = tree_cons (gnu_constr, gnu_value, gnu_inputs);
+ Next_Asm_Input ();
+ }
+
+ Clobber_Setup (gnat_node);
+ while ((clobber = (char *) Clobber_Get_Next ()))
+ gnu_clobbers
+ = tree_cons (NULL_TREE,
+ build_string (strlen (clobber) + 1, clobber),
+ gnu_clobbers);
+
+ /* Then perform some standard checking and processing on the
+ operands. In particular, mark them addressable if needed. */
+ gnu_outputs = nreverse (gnu_outputs);
+ noutputs = list_length (gnu_outputs);
+ gnu_inputs = nreverse (gnu_inputs);
+ ninputs = list_length (gnu_inputs);
+ oconstraints = XALLOCAVEC (const char *, noutputs);
+
+ for (i = 0, tail = gnu_outputs; tail; ++i, tail = TREE_CHAIN (tail))
+ {
+ tree output = TREE_VALUE (tail);
+ constraint
+ = TREE_STRING_POINTER (TREE_VALUE (TREE_PURPOSE (tail)));
+ oconstraints[i] = constraint;
+
+ if (parse_output_constraint (&constraint, i, ninputs, noutputs,
+ &allows_mem, &allows_reg, &fake))
+ {
+ /* If the operand is going to end up in memory,
+ mark it addressable. Note that we don't test
+ allows_mem like in the input case below; this
+ is modeled on the C front-end. */
+ if (!allows_reg)
+ {
+ output = remove_conversions (output, false);
+ if (TREE_CODE (output) == CONST_DECL
+ && DECL_CONST_CORRESPONDING_VAR (output))
+ output = DECL_CONST_CORRESPONDING_VAR (output);
+ if (!gnat_mark_addressable (output))
+ output = error_mark_node;
+ }
+ }
+ else
+ output = error_mark_node;
+
+ TREE_VALUE (tail) = output;
+ }
+
+ for (i = 0, tail = gnu_inputs; tail; ++i, tail = TREE_CHAIN (tail))
+ {
+ tree input = TREE_VALUE (tail);
+ constraint
+ = TREE_STRING_POINTER (TREE_VALUE (TREE_PURPOSE (tail)));
+
+ if (parse_input_constraint (&constraint, i, ninputs, noutputs,
+ 0, oconstraints,
+ &allows_mem, &allows_reg))
+ {
+ /* If the operand is going to end up in memory,
+ mark it addressable. */
+ if (!allows_reg && allows_mem)
+ {
+ input = remove_conversions (input, false);
+ if (TREE_CODE (input) == CONST_DECL
+ && DECL_CONST_CORRESPONDING_VAR (input))
+ input = DECL_CONST_CORRESPONDING_VAR (input);
+ if (!gnat_mark_addressable (input))
+ input = error_mark_node;
+ }
+ }
+ else
+ input = error_mark_node;
+
+ TREE_VALUE (tail) = input;
+ }
+
+ gnu_result = build5 (ASM_EXPR, void_type_node,
+ gnu_template, gnu_outputs,
+ gnu_inputs, gnu_clobbers, NULL_TREE);
+ ASM_VOLATILE_P (gnu_result) = Is_Asm_Volatile (gnat_node);
+ }
+ else
+ gnu_result = alloc_stmt_list ();
+
+ break;
+
+ /****************/
+ /* Added Nodes */
+ /****************/
+
+ /* Markers are created by the ABE mechanism to capture information which
+ is either unavailable of expensive to recompute. Markers do not have
+ and runtime semantics, and should be ignored. */
+
+ case N_Call_Marker:
+ case N_Variable_Reference_Marker:
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ case N_Expression_With_Actions:
+ /* This construct doesn't define a scope so we don't push a binding
+ level around the statement list, but we wrap it in a SAVE_EXPR to
+ protect it from unsharing. Elaborate the expression as part of the
+ same statement group as the actions so that the type declaration
+ gets inserted there as well. This ensures that the type elaboration
+ code is issued past the actions computing values on which it might
+ depend. */
+ start_stmt_group ();
+ add_stmt_list (Actions (gnat_node));
+ gnu_expr = gnat_to_gnu (Expression (gnat_node));
+ gnu_result = end_stmt_group ();
+
+ gnu_result = build1 (SAVE_EXPR, void_type_node, gnu_result);
+ TREE_SIDE_EFFECTS (gnu_result) = 1;
+
+ gnu_result
+ = build_compound_expr (TREE_TYPE (gnu_expr), gnu_result, gnu_expr);
+ gnu_result_type = get_unpadded_type (Etype (gnat_node));
+ break;
+
+ case N_Freeze_Entity:
+ start_stmt_group ();
+ process_freeze_entity (gnat_node);
+ process_decls (Actions (gnat_node), Empty, Empty, true, true);
+ gnu_result = end_stmt_group ();
+ break;
+
+ case N_Freeze_Generic_Entity:
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ case N_Itype_Reference:
+ if (!present_gnu_tree (Itype (gnat_node)))
+ process_type (Itype (gnat_node));
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ case N_Free_Statement:
+ gnat_temp = Expression (gnat_node);
+
+ if (!type_annotate_only)
+ {
+ tree gnu_ptr, gnu_ptr_type, gnu_obj_type, gnu_actual_obj_type;
+
+ const Entity_Id gnat_desig_type
+ = Designated_Type (Underlying_Type (Etype (gnat_temp)));
+
+ /* Make sure the designated type is complete before dereferencing,
+ in case it is a Taft Amendment type. */
+ (void) gnat_to_gnu_entity (gnat_desig_type, NULL_TREE, false);
+
+ gnu_ptr = gnat_to_gnu (gnat_temp);
+ gnu_ptr_type = TREE_TYPE (gnu_ptr);
+
+ /* If this is a thin pointer, we must first dereference it to create
+ a fat pointer, then go back below to a thin pointer. The reason
+ for this is that we need to have a fat pointer someplace in order
+ to properly compute the size. */
+ if (TYPE_IS_THIN_POINTER_P (TREE_TYPE (gnu_ptr)))
+ gnu_ptr = build_unary_op (ADDR_EXPR, NULL_TREE,
+ build_unary_op (INDIRECT_REF, NULL_TREE,
+ gnu_ptr));
+
+ /* If this is a fat pointer, the object must have been allocated with
+ the template in front of the array. So pass the template address,
+ and get the total size; do it by converting to a thin pointer. */
+ if (TYPE_IS_FAT_POINTER_P (TREE_TYPE (gnu_ptr)))
+ gnu_ptr
+ = convert (build_pointer_type
+ (TYPE_OBJECT_RECORD_TYPE
+ (TYPE_UNCONSTRAINED_ARRAY (TREE_TYPE (gnu_ptr)))),
+ gnu_ptr);
+
+ gnu_obj_type = TREE_TYPE (TREE_TYPE (gnu_ptr));
+
+ /* If this is a thin pointer, the object must have been allocated with
+ the template in front of the array. So pass the template address,
+ and get the total size. */
+ if (TYPE_IS_THIN_POINTER_P (TREE_TYPE (gnu_ptr)))
+ gnu_ptr
+ = build_binary_op (POINTER_PLUS_EXPR, TREE_TYPE (gnu_ptr),
+ gnu_ptr,
+ fold_build1 (NEGATE_EXPR, sizetype,
+ byte_position
+ (DECL_CHAIN
+ TYPE_FIELDS ((gnu_obj_type)))));
+
+ /* If we have a special dynamic constrained subtype on the node, use
+ it to compute the size; otherwise, use the designated subtype. */
+ if (Present (Actual_Designated_Subtype (gnat_node)))
+ {
+ gnu_actual_obj_type
+ = gnat_to_gnu_type (Actual_Designated_Subtype (gnat_node));
+
+ if (TYPE_IS_FAT_OR_THIN_POINTER_P (gnu_ptr_type))
+ gnu_actual_obj_type
+ = build_unc_object_type_from_ptr (gnu_ptr_type,
+ gnu_actual_obj_type,
+ get_identifier ("DEALLOC"),
+ false);
+ }
+ else
+ gnu_actual_obj_type = gnu_obj_type;
+
+ tree gnu_size = TYPE_SIZE_UNIT (gnu_actual_obj_type);
+ gnu_size = SUBSTITUTE_PLACEHOLDER_IN_EXPR (gnu_size, gnu_ptr);
+
+ gnu_result
+ = build_call_alloc_dealloc (gnu_ptr, gnu_size, gnu_obj_type,
+ Procedure_To_Call (gnat_node),
+ Storage_Pool (gnat_node),
+ gnat_node);
+ }
+ break;
+
+ case N_Raise_Constraint_Error:
+ case N_Raise_Program_Error:
+ case N_Raise_Storage_Error:
+ if (type_annotate_only)
+ gnu_result = alloc_stmt_list ();
+ else
+ gnu_result = Raise_Error_to_gnu (gnat_node, &gnu_result_type);
+ break;
+
+ case N_Validate_Unchecked_Conversion:
+ /* The only validation we currently do on an unchecked conversion is
+ that of aliasing assumptions. */
+ if (flag_strict_aliasing)
+ gnat_validate_uc_list.safe_push (gnat_node);
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ case N_Function_Specification:
+ case N_Procedure_Specification:
+ case N_Op_Concat:
+ case N_Component_Association:
+ /* These nodes should only be present when annotating types. */
+ gcc_assert (type_annotate_only);
+ gnu_result = alloc_stmt_list ();
+ break;
+
+ default:
+ /* Other nodes are not supposed to reach here. */
+ gcc_unreachable ();
+ }
+
+ /* If we are in the elaboration procedure, check if we are violating the
+ No_Elaboration_Code restriction by having a non-empty statement. */
+ if (statement_node_p (gnat_node)
+ && !(TREE_CODE (gnu_result) == STATEMENT_LIST
+ && empty_stmt_list_p (gnu_result))
+ && current_function_decl == get_elaboration_procedure ())
+ Check_Elaboration_Code_Allowed (gnat_node);
+
+ /* If we pushed the processing of the elaboration routine, pop it back. */
+ if (went_into_elab_proc)
+ current_function_decl = NULL_TREE;
+
+ /* When not optimizing, turn boolean rvalues B into B != false tests
+ so that we can put the location information of the reference to B on
+ the inequality operator for better debug info. */
+ if (!optimize
+ && TREE_CODE (gnu_result) != INTEGER_CST
+ && TREE_CODE (gnu_result) != TYPE_DECL
+ && (kind == N_Identifier
+ || kind == N_Expanded_Name
+ || kind == N_Explicit_Dereference
+ || kind == N_Indexed_Component
+ || kind == N_Selected_Component)
+ && TREE_CODE (get_base_type (gnu_result_type)) == BOOLEAN_TYPE
+ && Nkind (Parent (gnat_node)) != N_Attribute_Reference
+ && Nkind (Parent (gnat_node)) != N_Pragma_Argument_Association
+ && Nkind (Parent (gnat_node)) != N_Variant_Part
+ && !lvalue_required_p (gnat_node, gnu_result_type, false, false))
+ {
+ gnu_result
+ = build_binary_op (NE_EXPR, gnu_result_type,
+ convert (gnu_result_type, gnu_result),
+ convert (gnu_result_type, boolean_false_node));
+ if (TREE_CODE (gnu_result) != INTEGER_CST)
+ set_gnu_expr_location_from_node (gnu_result, gnat_node);
+ }
+
+ /* Set the location information on the result if it's not a simple name
+ or something that contains a simple name, for example a tag, because
+ we don"t want all the references to get the location of the first use.
+ Note that we may have no result if we tried to build a CALL_EXPR node
+ to a procedure with no side-effects and optimization is enabled. */
+ else if (kind != N_Identifier
+ && !(kind == N_Selected_Component
+ && Chars (Selector_Name (gnat_node)) == Name_uTag)
+ && gnu_result
+ && EXPR_P (gnu_result))
+ set_gnu_expr_location_from_node (gnu_result, gnat_node);
+
+ /* If we're supposed to return something of void_type, it means we have
+ something we're elaborating for effect, so just return. */
+ if (TREE_CODE (gnu_result_type) == VOID_TYPE)
+ return gnu_result;
+
+ /* If the result is a constant that overflowed, raise Constraint_Error. */
+ if (TREE_CODE (gnu_result) == INTEGER_CST && TREE_OVERFLOW (gnu_result))
+ {
+ post_error ("??Constraint_Error will be raised at run time", gnat_node);
+ gnu_result
+ = build1 (NULL_EXPR, gnu_result_type,
+ build_call_raise (CE_Overflow_Check_Failed, gnat_node,
+ N_Raise_Constraint_Error));
+ }
+
+ /* If the result has side-effects and is of an unconstrained type, protect
+ the expression in case it will be referenced multiple times, i.e. for
+ its value and to compute the size of an object. But do it neither for
+ an object nor a renaming declaration, nor a return statement of a call
+ to a function that returns an unconstrained record type with default
+ discriminant, because there is no size to be computed in these cases
+ and this will create a useless temporary. We must do this before any
+ conversions. */
+ if (TREE_SIDE_EFFECTS (gnu_result)
+ && (TREE_CODE (gnu_result_type) == UNCONSTRAINED_ARRAY_TYPE
+ || CONTAINS_PLACEHOLDER_P (TYPE_SIZE (gnu_result_type)))
+ && !(TREE_CODE (gnu_result) == CALL_EXPR
+ && type_is_padding_self_referential (TREE_TYPE (gnu_result))
+ && (Nkind (Parent (gnat_node)) == N_Object_Declaration
+ || Nkind (Parent (gnat_node)) == N_Object_Renaming_Declaration
+ || Nkind (Parent (gnat_node)) == N_Simple_Return_Statement)))
+ gnu_result = gnat_protect_expr (gnu_result);
+
+ /* Now convert the result to the result type, unless we are in one of the
+ following cases:
+
+ 1. If this is the LHS of an assignment or an actual parameter of a
+ call, return the result almost unmodified since the RHS will have
+ to be converted to our type in that case, unless the result type
+ has a simpler size or for array types because this size might be
+ changed in-between. Likewise if there is just a no-op unchecked
+ conversion in-between. Similarly, don't convert integral types
+ that are the operands of an unchecked conversion since we need
+ to ignore those conversions (for 'Valid).
+
+ 2. If we have a label (which doesn't have any well-defined type), a
+ field or an error, return the result almost unmodified. Similarly,
+ if the two types are record types with the same name, don't convert.
+ This will be the case when we are converting from a packable version
+ of a type to its original type and we need those conversions to be
+ NOPs in order for assignments into these types to work properly.
+
+ 3. If the type is void or if we have no result, return error_mark_node
+ to show we have no result.
+
+ 4. If this is a call to a function that returns with variable size and
+ the call is used as the expression in either an object or a renaming
+ declaration, return the result unmodified because we want to use the
+ return slot optimization in this case.
+
+ 5. If this is a reference to an unconstrained array which is used as the
+ prefix of an attribute reference that requires an lvalue, return the
+ result unmodified because we want to return the original bounds.
+
+ 6. Finally, if the type of the result is already correct. */
+
+ if (Present (Parent (gnat_node))
+ && (lhs_or_actual_p (gnat_node)
+ || (Nkind (Parent (gnat_node)) == N_Unchecked_Type_Conversion
+ && unchecked_conversion_nop (Parent (gnat_node)))
+ || (Nkind (Parent (gnat_node)) == N_Unchecked_Type_Conversion
+ && !AGGREGATE_TYPE_P (gnu_result_type)
+ && !AGGREGATE_TYPE_P (TREE_TYPE (gnu_result))))
+ && !(TYPE_SIZE (gnu_result_type)
+ && TYPE_SIZE (TREE_TYPE (gnu_result))
+ && AGGREGATE_TYPE_P (gnu_result_type)
+ == AGGREGATE_TYPE_P (TREE_TYPE (gnu_result))
+ && ((TREE_CODE (TYPE_SIZE (gnu_result_type)) == INTEGER_CST
+ && (TREE_CODE (TYPE_SIZE (TREE_TYPE (gnu_result)))
+ != INTEGER_CST))
+ || (TREE_CODE (TYPE_SIZE (gnu_result_type)) != INTEGER_CST
+ && !CONTAINS_PLACEHOLDER_P (TYPE_SIZE (gnu_result_type))
+ && (CONTAINS_PLACEHOLDER_P
+ (TYPE_SIZE (TREE_TYPE (gnu_result)))))
+ || (TREE_CODE (gnu_result_type) == ARRAY_TYPE
+ && TREE_CODE (TREE_TYPE (gnu_result)) == ARRAY_TYPE))
+ && !(TREE_CODE (gnu_result_type) == RECORD_TYPE
+ && TYPE_JUSTIFIED_MODULAR_P (gnu_result_type))))
+ {
+ /* Remove padding only if the inner object is of self-referential
+ size: in that case it must be an object of unconstrained type
+ with a default discriminant and we want to avoid copying too
+ much data. But do not remove it if it is already too small. */
+ if (type_is_padding_self_referential (TREE_TYPE (gnu_result))
+ && !(TREE_CODE (gnu_result) == COMPONENT_REF
+ && DECL_BIT_FIELD (TREE_OPERAND (gnu_result, 1))
+ && DECL_SIZE (TREE_OPERAND (gnu_result, 1))
+ != TYPE_SIZE (TREE_TYPE (gnu_result))))
+ gnu_result = convert (TREE_TYPE (TYPE_FIELDS (TREE_TYPE (gnu_result))),
+ gnu_result);
+ }
+
+ else if (TREE_CODE (gnu_result) == LABEL_DECL
+ || TREE_CODE (gnu_result) == FIELD_DECL
+ || TREE_CODE (gnu_result) == ERROR_MARK
+ || (TYPE_NAME (gnu_result_type)
+ == TYPE_NAME (TREE_TYPE (gnu_result))
+ && TREE_CODE (gnu_result_type) == RECORD_TYPE
+ && TREE_CODE (TREE_TYPE (gnu_result)) == RECORD_TYPE))
+ {
+ /* Remove any padding. */
+ gnu_result = maybe_padded_object (gnu_result);
+ }
+
+ else if (gnu_result == error_mark_node || gnu_result_type == void_type_node)
+ gnu_result = error_mark_node;
+
+ else if (TREE_CODE (gnu_result) == CALL_EXPR
+ && Present (Parent (gnat_node))
+ && (Nkind (Parent (gnat_node)) == N_Object_Declaration
+ || Nkind (Parent (gnat_node)) == N_Object_Renaming_Declaration)
+ && return_type_with_variable_size_p (TREE_TYPE (gnu_result)))
+ ;
+
+ else if (TREE_CODE (gnu_result) == UNCONSTRAINED_ARRAY_REF
+ && Present (Parent (gnat_node))
+ && Nkind (Parent (gnat_node)) == N_Attribute_Reference
+ && lvalue_required_for_attribute_p (Parent (gnat_node)))
+ ;
+
+ else if (TREE_TYPE (gnu_result) != gnu_result_type)
+ gnu_result = convert (gnu_result_type, gnu_result);
+
+ /* We don't need any NOP_EXPR or NON_LVALUE_EXPR on the result. */
+ while ((TREE_CODE (gnu_result) == NOP_EXPR
+ || TREE_CODE (gnu_result) == NON_LVALUE_EXPR)
+ && TREE_TYPE (TREE_OPERAND (gnu_result, 0)) == TREE_TYPE (gnu_result))
+ gnu_result = TREE_OPERAND (gnu_result, 0);
+
+ return gnu_result;
+}
+
+/* Similar to gnat_to_gnu, but discard any object that might be created in
+ the course of the translation of GNAT_NODE, which must be an "external"
+ expression in the sense that it will be elaborated elsewhere. */
+
+tree
+gnat_to_gnu_external (Node_Id gnat_node)
+{
+ const int save_force_global = force_global;
+ bool went_into_elab_proc;
+
+ /* Force the local context and create a fake scope that we zap
+ at the end so declarations will not be stuck either in the
+ global varpool or in the current scope. */
+ if (!current_function_decl)
+ {
+ current_function_decl = get_elaboration_procedure ();
+ went_into_elab_proc = true;
+ }
+ else
+ went_into_elab_proc = false;
+ force_global = 0;
+ gnat_pushlevel ();
+
+ tree gnu_result = gnat_to_gnu (gnat_node);
+
+ gnat_zaplevel ();
+ force_global = save_force_global;
+ if (went_into_elab_proc)
+ current_function_decl = NULL_TREE;
+
+ /* Do not import locations from external units. */
+ if (gnu_result && EXPR_P (gnu_result))
+ SET_EXPR_LOCATION (gnu_result, UNKNOWN_LOCATION);
+
+ return gnu_result;
+}
+
+/* Return true if the statement list STMT_LIST is empty. */
+
+static bool
+empty_stmt_list_p (tree stmt_list)
+{
+ tree_stmt_iterator tsi;
+
+ for (tsi = tsi_start (stmt_list); !tsi_end_p (tsi); tsi_next (&tsi))
+ {
+ tree stmt = tsi_stmt (tsi);
+
+ /* Anything else than an empty STMT_STMT counts as something. */
+ if (TREE_CODE (stmt) != STMT_STMT || STMT_STMT_STMT (stmt))
+ return false;
+ }
+
+ return true;
+}
+
+/* Record the current code position in GNAT_NODE. */
+
+static void
+record_code_position (Node_Id gnat_node)
+{
+ tree stmt_stmt = build1 (STMT_STMT, void_type_node, NULL_TREE);
+
+ add_stmt_with_node (stmt_stmt, gnat_node);
+ save_gnu_tree (gnat_node, stmt_stmt, true);
+}
+
+/* Insert the code for GNAT_NODE at the position saved for that node. */
+
+static void
+insert_code_for (Node_Id gnat_node)
+{
+ tree code = gnat_to_gnu (gnat_node);
+
+ /* It's too late to remove the STMT_STMT itself at this point. */
+ if (!empty_stmt_list_p (code))
+ STMT_STMT_STMT (get_gnu_tree (gnat_node)) = code;
+
+ save_gnu_tree (gnat_node, NULL_TREE, true);
+}
+
+/* Start a new statement group chained to the previous group. */
+
+void
+start_stmt_group (void)
+{
+ struct stmt_group *group = stmt_group_free_list;
+
+ /* First see if we can get one from the free list. */
+ if (group)
+ stmt_group_free_list = group->previous;
+ else
+ group = ggc_alloc<stmt_group> ();
+
+ group->previous = current_stmt_group;
+ group->stmt_list = group->block = group->cleanups = NULL_TREE;
+ current_stmt_group = group;
+}
+
+/* Add GNU_STMT to the current statement group. If it is an expression with
+ no effects, it is ignored. */
+
+void
+add_stmt (tree gnu_stmt)
+{
+ append_to_statement_list (gnu_stmt, &current_stmt_group->stmt_list);
+}
+
+/* Similar, but the statement is always added, regardless of side-effects. */
+
+void
+add_stmt_force (tree gnu_stmt)
+{
+ append_to_statement_list_force (gnu_stmt, &current_stmt_group->stmt_list);
+}
+
+/* Like add_stmt, but set the location of GNU_STMT to that of GNAT_NODE. */
+
+void
+add_stmt_with_node (tree gnu_stmt, Node_Id gnat_node)
+{
+ if (Present (gnat_node))
+ set_expr_location_from_node (gnu_stmt, gnat_node);
+ add_stmt (gnu_stmt);
+}
+
+/* Similar, but the statement is always added, regardless of side-effects. */
+
+void
+add_stmt_with_node_force (tree gnu_stmt, Node_Id gnat_node)
+{
+ if (Present (gnat_node))
+ set_expr_location_from_node (gnu_stmt, gnat_node);
+ add_stmt_force (gnu_stmt);
+}
+
+/* Add a declaration statement for GNU_DECL to the current statement group.
+ Get the SLOC to be put onto the statement from GNAT_NODE. */
+
+void
+add_decl_expr (tree gnu_decl, Node_Id gnat_node)
+{
+ tree type = TREE_TYPE (gnu_decl);
+ tree gnu_stmt, gnu_init;
+
+ /* If this is a variable that Gigi is to ignore, we may have been given
+ an ERROR_MARK. So test for it. We also might have been given a
+ reference for a renaming. So only do something for a decl. Also
+ ignore a TYPE_DECL for an UNCONSTRAINED_ARRAY_TYPE. */
+ if (!DECL_P (gnu_decl)
+ || (TREE_CODE (gnu_decl) == TYPE_DECL
+ && TREE_CODE (type) == UNCONSTRAINED_ARRAY_TYPE))
+ return;
+
+ gnu_stmt = build1 (DECL_EXPR, void_type_node, gnu_decl);
+
+ /* If we are external or global, we don't want to output the DECL_EXPR for
+ this DECL node since we already have evaluated the expressions in the
+ sizes and positions as globals and doing it again would be wrong. */
+ if (DECL_EXTERNAL (gnu_decl) || global_bindings_p ())
+ {
+ /* Mark everything as used to prevent node sharing with subprograms.
+ Note that walk_tree knows how to deal with TYPE_DECL, but neither
+ VAR_DECL nor CONST_DECL. This appears to be somewhat arbitrary. */
+ MARK_VISITED (gnu_stmt);
+ if (TREE_CODE (gnu_decl) == VAR_DECL
+ || TREE_CODE (gnu_decl) == CONST_DECL)
+ {
+ MARK_VISITED (DECL_SIZE (gnu_decl));
+ MARK_VISITED (DECL_SIZE_UNIT (gnu_decl));
+ MARK_VISITED (DECL_INITIAL (gnu_decl));
+ }
+ }
+ else
+ add_stmt_with_node (gnu_stmt, gnat_node);
+
+ /* Mark our TYPE_ADA_SIZE field now since it will not be gimplified. */
+ if (TREE_CODE (gnu_decl) == TYPE_DECL
+ && RECORD_OR_UNION_TYPE_P (type)
+ && !TYPE_FAT_POINTER_P (type))
+ MARK_VISITED (TYPE_ADA_SIZE (type));
+
+ /* If this is a variable and an initializer is attached to it, it must be
+ valid for the context. Similar to init_const in create_var_decl. */
+ if (TREE_CODE (gnu_decl) == VAR_DECL
+ && (gnu_init = DECL_INITIAL (gnu_decl))
+ && (!gnat_types_compatible_p (type, TREE_TYPE (gnu_init))
+ || (TREE_STATIC (gnu_decl)
+ && !initializer_constant_valid_p (gnu_init,
+ TREE_TYPE (gnu_init)))))
+ {
+ DECL_INITIAL (gnu_decl) = NULL_TREE;
+ if (TREE_READONLY (gnu_decl))
+ {
+ TREE_READONLY (gnu_decl) = 0;
+ DECL_READONLY_ONCE_ELAB (gnu_decl) = 1;
+ }
+
+ /* Remove any padding so the assignment is done properly. */
+ gnu_decl = maybe_padded_object (gnu_decl);
+
+ gnu_stmt = build_binary_op (INIT_EXPR, NULL_TREE, gnu_decl, gnu_init);
+ add_stmt_with_node (gnu_stmt, gnat_node);
+ }
+}
+
+/* Callback for walk_tree to mark the visited trees rooted at *TP. */
+
+static tree
+mark_visited_r (tree *tp, int *walk_subtrees, void *data ATTRIBUTE_UNUSED)
+{
+ tree t = *tp;
+
+ if (TREE_VISITED (t))
+ *walk_subtrees = 0;
+
+ /* Don't mark a dummy type as visited because we want to mark its sizes
+ and fields once it's filled in. */
+ else if (!TYPE_IS_DUMMY_P (t))
+ TREE_VISITED (t) = 1;
+
+ /* The test in gimplify_type_sizes is on the main variant. */
+ if (TYPE_P (t))
+ TYPE_SIZES_GIMPLIFIED (TYPE_MAIN_VARIANT (t)) = 1;
+
+ return NULL_TREE;
+}
+
+/* Mark nodes rooted at T with TREE_VISITED and types as having their
+ sized gimplified. We use this to indicate all variable sizes and
+ positions in global types may not be shared by any subprogram. */
+
+void
+mark_visited (tree t)
+{
+ walk_tree (&t, mark_visited_r, NULL, NULL);
+}
+
+/* Add GNU_CLEANUP, a cleanup action, to the current code group and
+ set its location to that of GNAT_NODE if present, but with column info
+ cleared so that conditional branches generated as part of the cleanup
+ code do not interfere with coverage analysis tools. */
+
+static void
+add_cleanup (tree gnu_cleanup, Node_Id gnat_node)
+{
+ if (Present (gnat_node))
+ set_expr_location_from_node (gnu_cleanup, gnat_node, true);
+
+ /* An EH_ELSE_EXPR must be by itself, and that's all we need when we
+ use it. The assert below makes sure that is so. Should we ever
+ need more than that, we could combine EH_ELSE_EXPRs, and copy
+ non-EH_ELSE_EXPR stmts into both cleanup paths of an
+ EH_ELSE_EXPR. */
+ if (TREE_CODE (gnu_cleanup) == EH_ELSE_EXPR)
+ {
+ gcc_assert (!current_stmt_group->cleanups);
+ current_stmt_group->cleanups = gnu_cleanup;
+ }
+ else
+ {
+ gcc_assert (!current_stmt_group->cleanups
+ || (TREE_CODE (current_stmt_group->cleanups)
+ != EH_ELSE_EXPR));
+ append_to_statement_list (gnu_cleanup, &current_stmt_group->cleanups);
+ }
+}
+
+/* Set the BLOCK node corresponding to the current code group to GNU_BLOCK. */
+
+void
+set_block_for_group (tree gnu_block)
+{
+ gcc_assert (!current_stmt_group->block);
+ current_stmt_group->block = gnu_block;
+}
+
+/* Return code corresponding to the current code group. It is normally
+ a STATEMENT_LIST, but may also be a BIND_EXPR or TRY_FINALLY_EXPR if
+ BLOCK or cleanups were set. */
+
+tree
+end_stmt_group (void)
+{
+ struct stmt_group *group = current_stmt_group;
+ tree gnu_retval = group->stmt_list;
+
+ /* If this is a null list, allocate a new STATEMENT_LIST. Then, if there
+ are cleanups, make a TRY_FINALLY_EXPR. Last, if there is a BLOCK,
+ make a BIND_EXPR. Note that we nest in that because the cleanup may
+ reference variables in the block. */
+ if (!gnu_retval)
+ gnu_retval = alloc_stmt_list ();
+
+ if (group->cleanups)
+ gnu_retval = build2 (TRY_FINALLY_EXPR, void_type_node, gnu_retval,
+ group->cleanups);
+
+ if (current_stmt_group->block)
+ gnu_retval = build3 (BIND_EXPR, void_type_node, BLOCK_VARS (group->block),
+ gnu_retval, group->block);
+
+ /* Remove this group from the stack and add it to the free list. */
+ current_stmt_group = group->previous;
+ group->previous = stmt_group_free_list;
+ stmt_group_free_list = group;
+
+ return gnu_retval;
+}
+
+/* Return whether the current statement group may fall through. */
+
+static inline bool
+stmt_group_may_fallthru (void)
+{
+ if (current_stmt_group->stmt_list)
+ return block_may_fallthru (current_stmt_group->stmt_list);
+ else
+ return true;
+}
+
+/* Add a list of statements from GNAT_LIST, a possibly-empty list of
+ statements.*/
+
+static void
+add_stmt_list (List_Id gnat_list)
+{
+ Node_Id gnat_node;
+
+ if (Present (gnat_list))
+ for (gnat_node = First (gnat_list); Present (gnat_node);
+ gnat_node = Next (gnat_node))
+ add_stmt (gnat_to_gnu (gnat_node));
+}
+
+/* Build a tree from GNAT_LIST, a possibly-empty list of statements.
+ If BINDING_P is true, push and pop a binding level around the list. */
+
+static tree
+build_stmt_group (List_Id gnat_list, bool binding_p)
+{
+ start_stmt_group ();
+
+ if (binding_p)
+ gnat_pushlevel ();
+
+ add_stmt_list (gnat_list);
+
+ if (binding_p)
+ gnat_poplevel ();
+
+ return end_stmt_group ();
+}
+
+/* Generate GIMPLE in place for the expression at *EXPR_P. */
+
+int
+gnat_gimplify_expr (tree *expr_p, gimple_seq *pre_p,
+ gimple_seq *post_p ATTRIBUTE_UNUSED)
+{
+ tree expr = *expr_p;
+ tree type = TREE_TYPE (expr);
+ tree op;
+
+ if (IS_ADA_STMT (expr))
+ return gnat_gimplify_stmt (expr_p);
+
+ switch (TREE_CODE (expr))
+ {
+ case NULL_EXPR:
+ /* If this is an aggregate type, build a null pointer of the appropriate
+ type and dereference it. */
+ if (AGGREGATE_TYPE_P (type)
+ || TREE_CODE (type) == UNCONSTRAINED_ARRAY_TYPE)
+ *expr_p = build_unary_op (INDIRECT_REF, NULL_TREE,
+ convert (build_pointer_type (type),
+ integer_zero_node));
+ /* Otherwise, just make a VAR_DECL. */
+ else
+ {
+ *expr_p = create_tmp_var (type, NULL);
+ suppress_warning (*expr_p);
+ }
+
+ gimplify_and_add (TREE_OPERAND (expr, 0), pre_p);
+ return GS_OK;
+
+ case UNCONSTRAINED_ARRAY_REF:
+ /* We should only do this if we are just elaborating for side-effects,
+ but we can't know that yet. */
+ *expr_p = TREE_OPERAND (*expr_p, 0);
+ return GS_OK;
+
+ case ADDR_EXPR:
+ op = TREE_OPERAND (expr, 0);
+
+ /* If we are taking the address of a constant CONSTRUCTOR, make sure it
+ is put into static memory. We know that it's going to be read-only
+ given the semantics we have and it must be in static memory when the
+ reference is in an elaboration procedure. */
+ if (TREE_CODE (op) == CONSTRUCTOR && TREE_CONSTANT (op))
+ {
+ tree addr = build_fold_addr_expr (tree_output_constant_def (op));
+ *expr_p = fold_convert (type, addr);
+ return GS_ALL_DONE;
+ }
+
+ /* Replace atomic loads with their first argument. That's necessary
+ because the gimplifier would create a temporary otherwise. */
+ if (TREE_SIDE_EFFECTS (op))
+ while (handled_component_p (op) || CONVERT_EXPR_P (op))
+ {
+ tree inner = TREE_OPERAND (op, 0);
+ if (TREE_CODE (inner) == CALL_EXPR && call_is_atomic_load (inner))
+ {
+ tree t = CALL_EXPR_ARG (inner, 0);
+ if (TREE_CODE (t) == NOP_EXPR)
+ t = TREE_OPERAND (t, 0);
+ if (TREE_CODE (t) == ADDR_EXPR)
+ TREE_OPERAND (op, 0) = TREE_OPERAND (t, 0);
+ else
+ TREE_OPERAND (op, 0) = build_fold_indirect_ref (t);
+ }
+ else
+ op = inner;
+ }
+
+ return GS_UNHANDLED;
+
+ case CALL_EXPR:
+ /* If we are passing a constant fat pointer CONSTRUCTOR, make sure it is
+ put into static memory; this performs a restricted version of constant
+ propagation on fat pointers in calls. But do not do it for strings to
+ avoid blocking concatenation in the caller when it is inlined. */
+ for (int i = 0; i < call_expr_nargs (expr); i++)
+ {
+ tree arg = *(CALL_EXPR_ARGP (expr) + i);
+
+ if (TREE_CODE (arg) == CONSTRUCTOR
+ && TREE_CONSTANT (arg)
+ && TYPE_IS_FAT_POINTER_P (TREE_TYPE (arg)))
+ {
+ tree t = CONSTRUCTOR_ELT (arg, 0)->value;
+ if (TREE_CODE (t) == NOP_EXPR)
+ t = TREE_OPERAND (t, 0);
+ if (TREE_CODE (t) == ADDR_EXPR)
+ t = TREE_OPERAND (t, 0);
+ if (TREE_CODE (t) != STRING_CST)
+ *(CALL_EXPR_ARGP (expr) + i) = tree_output_constant_def (arg);
+ }
+ }
+
+ return GS_UNHANDLED;
+
+ case VIEW_CONVERT_EXPR:
+ op = TREE_OPERAND (expr, 0);
+
+ /* If we are view-converting a CONSTRUCTOR or a call from an aggregate
+ type to a scalar one, explicitly create the local temporary. That's
+ required if the type is passed by reference. */
+ if ((TREE_CODE (op) == CONSTRUCTOR || TREE_CODE (op) == CALL_EXPR)
+ && AGGREGATE_TYPE_P (TREE_TYPE (op))
+ && !AGGREGATE_TYPE_P (type))
+ {
+ tree mod, new_var = create_tmp_var_raw (TREE_TYPE (op), "C");
+ gimple_add_tmp_var (new_var);
+
+ mod = build2 (INIT_EXPR, TREE_TYPE (new_var), new_var, op);
+ gimplify_and_add (mod, pre_p);
+
+ TREE_OPERAND (expr, 0) = new_var;
+ return GS_OK;
+ }
+
+ return GS_UNHANDLED;
+
+ case DECL_EXPR:
+ op = DECL_EXPR_DECL (expr);
+
+ /* The expressions for the RM bounds must be gimplified to ensure that
+ they are properly elaborated. See gimplify_decl_expr. */
+ if ((TREE_CODE (op) == TYPE_DECL || TREE_CODE (op) == VAR_DECL)
+ && !TYPE_SIZES_GIMPLIFIED (TREE_TYPE (op)))
+ switch (TREE_CODE (TREE_TYPE (op)))
+ {
+ case INTEGER_TYPE:
+ case ENUMERAL_TYPE:
+ case BOOLEAN_TYPE:
+ case REAL_TYPE:
+ {
+ tree type = TYPE_MAIN_VARIANT (TREE_TYPE (op)), t, val;
+
+ val = TYPE_RM_MIN_VALUE (type);
+ if (val)
+ {
+ gimplify_one_sizepos (&val, pre_p);
+ for (t = type; t; t = TYPE_NEXT_VARIANT (t))
+ SET_TYPE_RM_MIN_VALUE (t, val);
+ }
+
+ val = TYPE_RM_MAX_VALUE (type);
+ if (val)
+ {
+ gimplify_one_sizepos (&val, pre_p);
+ for (t = type; t; t = TYPE_NEXT_VARIANT (t))
+ SET_TYPE_RM_MAX_VALUE (t, val);
+ }
+
+ }
+ break;
+
+ default:
+ break;
+ }
+
+ /* ... fall through ... */
+
+ default:
+ return GS_UNHANDLED;
+ }
+}
+
+/* Generate GIMPLE in place for the statement at *STMT_P. */
+
+static enum gimplify_status
+gnat_gimplify_stmt (tree *stmt_p)
+{
+ tree stmt = *stmt_p;
+
+ switch (TREE_CODE (stmt))
+ {
+ case STMT_STMT:
+ *stmt_p = STMT_STMT_STMT (stmt);
+ return GS_OK;
+
+ case LOOP_STMT:
+ {
+ tree gnu_start_label = create_artificial_label (input_location);
+ tree gnu_cond = LOOP_STMT_COND (stmt);
+ tree gnu_update = LOOP_STMT_UPDATE (stmt);
+ tree gnu_end_label = LOOP_STMT_LABEL (stmt);
+
+ /* Build the condition expression from the test, if any. */
+ if (gnu_cond)
+ {
+ /* Deal with the optimization hints. */
+ if (LOOP_STMT_IVDEP (stmt))
+ gnu_cond = build3 (ANNOTATE_EXPR, TREE_TYPE (gnu_cond), gnu_cond,
+ build_int_cst (integer_type_node,
+ annot_expr_ivdep_kind),
+ integer_zero_node);
+ if (LOOP_STMT_NO_UNROLL (stmt))
+ gnu_cond = build3 (ANNOTATE_EXPR, TREE_TYPE (gnu_cond), gnu_cond,
+ build_int_cst (integer_type_node,
+ annot_expr_unroll_kind),
+ integer_one_node);
+ if (LOOP_STMT_UNROLL (stmt))
+ gnu_cond = build3 (ANNOTATE_EXPR, TREE_TYPE (gnu_cond), gnu_cond,
+ build_int_cst (integer_type_node,
+ annot_expr_unroll_kind),
+ build_int_cst (NULL_TREE, USHRT_MAX));
+ if (LOOP_STMT_NO_VECTOR (stmt))
+ gnu_cond = build3 (ANNOTATE_EXPR, TREE_TYPE (gnu_cond), gnu_cond,
+ build_int_cst (integer_type_node,
+ annot_expr_no_vector_kind),
+ integer_zero_node);
+ if (LOOP_STMT_VECTOR (stmt))
+ gnu_cond = build3 (ANNOTATE_EXPR, TREE_TYPE (gnu_cond), gnu_cond,
+ build_int_cst (integer_type_node,
+ annot_expr_vector_kind),
+ integer_zero_node);
+
+ gnu_cond
+ = build3 (COND_EXPR, void_type_node, gnu_cond, NULL_TREE,
+ build1 (GOTO_EXPR, void_type_node, gnu_end_label));
+ }
+
+ /* Set to emit the statements of the loop. */
+ *stmt_p = NULL_TREE;
+
+ /* We first emit the start label and then a conditional jump to the
+ end label if there's a top condition, then the update if it's at
+ the top, then the body of the loop, then a conditional jump to
+ the end label if there's a bottom condition, then the update if
+ it's at the bottom, and finally a jump to the start label and the
+ definition of the end label. */
+ append_to_statement_list (build1 (LABEL_EXPR, void_type_node,
+ gnu_start_label),
+ stmt_p);
+
+ if (gnu_cond && !LOOP_STMT_BOTTOM_COND_P (stmt))
+ append_to_statement_list (gnu_cond, stmt_p);
+
+ if (gnu_update && LOOP_STMT_TOP_UPDATE_P (stmt))
+ append_to_statement_list (gnu_update, stmt_p);
+
+ append_to_statement_list (LOOP_STMT_BODY (stmt), stmt_p);
+
+ if (gnu_cond && LOOP_STMT_BOTTOM_COND_P (stmt))
+ append_to_statement_list (gnu_cond, stmt_p);
+
+ if (gnu_update && !LOOP_STMT_TOP_UPDATE_P (stmt))
+ append_to_statement_list (gnu_update, stmt_p);
+
+ tree t = build1 (GOTO_EXPR, void_type_node, gnu_start_label);
+ SET_EXPR_LOCATION (t, DECL_SOURCE_LOCATION (gnu_end_label));
+ append_to_statement_list (t, stmt_p);
+
+ append_to_statement_list (build1 (LABEL_EXPR, void_type_node,
+ gnu_end_label),
+ stmt_p);
+ return GS_OK;
+ }
+
+ case EXIT_STMT:
+ /* Build a statement to jump to the corresponding end label, then
+ see if it needs to be conditional. */
+ *stmt_p = build1 (GOTO_EXPR, void_type_node, EXIT_STMT_LABEL (stmt));
+ if (EXIT_STMT_COND (stmt))
+ *stmt_p = build3 (COND_EXPR, void_type_node,
+ EXIT_STMT_COND (stmt), *stmt_p, alloc_stmt_list ());
+ return GS_OK;
+
+ default:
+ gcc_unreachable ();
+ }
+}
+
+/* Force a reference to each of the entities in GNAT_PACKAGE recursively.
+
+ This routine is exclusively called in type_annotate mode, to compute DDA
+ information for types in withed units, for ASIS use. */
+
+static void
+elaborate_all_entities_for_package (Entity_Id gnat_package)
+{
+ Entity_Id gnat_entity;
+
+ for (gnat_entity = First_Entity (gnat_package);
+ Present (gnat_entity);
+ gnat_entity = Next_Entity (gnat_entity))
+ {
+ const Entity_Kind kind = Ekind (gnat_entity);
+
+ /* We are interested only in entities visible from the main unit. */
+ if (!Is_Public (gnat_entity))
+ continue;
+
+ /* Skip stuff internal to the compiler. */
+ if (Is_Intrinsic_Subprogram (gnat_entity))
+ continue;
+ if (kind == E_Operator)
+ continue;
+ if (IN (kind, Subprogram_Kind)
+ && (Present (Alias (gnat_entity))
+ || Is_Intrinsic_Subprogram (gnat_entity)))
+ continue;
+ if (Is_Itype (gnat_entity))
+ continue;
+
+ /* Skip named numbers. */
+ if (IN (kind, Named_Kind))
+ continue;
+
+ /* Skip generic declarations. */
+ if (IN (kind, Generic_Unit_Kind))
+ continue;
+
+ /* Skip formal objects. */
+ if (IN (kind, Formal_Object_Kind))
+ continue;
+
+ /* Skip package bodies. */
+ if (kind == E_Package_Body)
+ continue;
+
+ /* Skip limited views that point back to the main unit. */
+ if (IN (kind, Incomplete_Kind)
+ && From_Limited_With (gnat_entity)
+ && In_Extended_Main_Code_Unit (Non_Limited_View (gnat_entity)))
+ continue;
+
+ /* Skip types that aren't frozen. */
+ if (IN (kind, Type_Kind) && !Is_Frozen (gnat_entity))
+ continue;
+
+ /* Recurse on real packages that aren't in the main unit. */
+ if (kind == E_Package)
+ {
+ if (No (Renamed_Entity (gnat_entity))
+ && !In_Extended_Main_Code_Unit (gnat_entity))
+ elaborate_all_entities_for_package (gnat_entity);
+ }
+ else
+ gnat_to_gnu_entity (gnat_entity, NULL_TREE, false);
+ }
+}
+
+/* Force a reference to each of the entities in packages withed by GNAT_NODE.
+ Operate recursively but check that we aren't elaborating something more
+ than once.
+
+ This routine is exclusively called in type_annotate mode, to compute DDA
+ information for types in withed units, for ASIS use. */
+
+static void
+elaborate_all_entities (Node_Id gnat_node)
+{
+ Entity_Id gnat_with_clause;
+
+ /* Process each unit only once. As we trace the context of all relevant
+ units transitively, including generic bodies, we may encounter the
+ same generic unit repeatedly. */
+ if (!present_gnu_tree (gnat_node))
+ save_gnu_tree (gnat_node, integer_zero_node, true);
+
+ /* Save entities in all context units. A body may have an implicit_with
+ on its own spec, if the context includes a child unit, so don't save
+ the spec twice. */
+ for (gnat_with_clause = First (Context_Items (gnat_node));
+ Present (gnat_with_clause);
+ gnat_with_clause = Next (gnat_with_clause))
+ if (Nkind (gnat_with_clause) == N_With_Clause
+ && !present_gnu_tree (Library_Unit (gnat_with_clause))
+ && Library_Unit (gnat_with_clause) != Library_Unit (Cunit (Main_Unit)))
+ {
+ Node_Id gnat_unit = Library_Unit (gnat_with_clause);
+ Entity_Id gnat_entity = Entity (Name (gnat_with_clause));
+
+ elaborate_all_entities (gnat_unit);
+
+ if (Ekind (gnat_entity) == E_Package
+ && No (Renamed_Entity (gnat_entity)))
+ elaborate_all_entities_for_package (gnat_entity);
+
+ else if (Ekind (gnat_entity) == E_Generic_Package)
+ {
+ Node_Id gnat_body = Corresponding_Body (Unit (gnat_unit));
+
+ /* Retrieve compilation unit node of generic body. */
+ while (Present (gnat_body)
+ && Nkind (gnat_body) != N_Compilation_Unit)
+ gnat_body = Parent (gnat_body);
+
+ /* If body is available, elaborate its context. */
+ if (Present (gnat_body))
+ elaborate_all_entities (gnat_body);
+ }
+ }
+
+ if (Nkind (Unit (gnat_node)) == N_Package_Body)
+ elaborate_all_entities (Library_Unit (gnat_node));
+}
+
+/* Do the processing of GNAT_NODE, an N_Freeze_Entity. */
+
+static void
+process_freeze_entity (Node_Id gnat_node)
+{
+ const Entity_Id gnat_entity = Entity (gnat_node);
+ const Entity_Kind kind = Ekind (gnat_entity);
+ tree gnu_old, gnu_new;
+
+ /* If this is a package, generate code for the package body, if any. */
+ if (kind == E_Package)
+ {
+ const Node_Id gnat_decl = Parent (Declaration_Node (gnat_entity));
+ if (Present (Corresponding_Body (gnat_decl)))
+ insert_code_for (Parent (Corresponding_Body (gnat_decl)));
+ return;
+ }
+
+ /* Don't do anything for class-wide types as they are always transformed
+ into their root type. */
+ if (kind == E_Class_Wide_Type)
+ return;
+
+ /* Check for an old definition if this isn't an object with address clause,
+ since the saved GCC tree is the address expression in that case. */
+ gnu_old
+ = present_gnu_tree (gnat_entity) && No (Address_Clause (gnat_entity))
+ ? get_gnu_tree (gnat_entity) : NULL_TREE;
+
+ /* Don't do anything for subprograms that may have been elaborated before
+ their freeze nodes. This can happen, for example, because of an inner
+ call in an instance body or because of previous compilation of a spec
+ for inlining purposes. */
+ if (gnu_old
+ && ((TREE_CODE (gnu_old) == FUNCTION_DECL
+ && (kind == E_Function || kind == E_Procedure))
+ || (FUNC_OR_METHOD_TYPE_P (TREE_TYPE (gnu_old))
+ && kind == E_Subprogram_Type)))
+ return;
+
+ /* If we have a non-dummy type old tree, we have nothing to do, except for
+ aborting, since this node was never delayed as it should have been. We
+ let this happen for concurrent types and their Corresponding_Record_Type,
+ however, because each might legitimately be elaborated before its own
+ freeze node, e.g. while processing the other. */
+ if (gnu_old
+ && !(TREE_CODE (gnu_old) == TYPE_DECL
+ && TYPE_IS_DUMMY_P (TREE_TYPE (gnu_old))))
+ {
+ gcc_assert (Is_Concurrent_Type (gnat_entity)
+ || (Is_Record_Type (gnat_entity)
+ && Is_Concurrent_Record_Type (gnat_entity)));
+ return;
+ }
+
+ /* Reset the saved tree, if any, and elaborate the object or type for real.
+ If there is a full view, elaborate it and use the result. And, if this
+ is the root type of a class-wide type, reuse it for the latter. */
+ if (gnu_old)
+ {
+ save_gnu_tree (gnat_entity, NULL_TREE, false);
+
+ if (Is_Incomplete_Or_Private_Type (gnat_entity)
+ && Present (Full_View (gnat_entity)))
+ {
+ Entity_Id full_view = Full_View (gnat_entity);
+
+ save_gnu_tree (full_view, NULL_TREE, false);
+
+ if (Is_Private_Type (full_view)
+ && Present (Underlying_Full_View (full_view)))
+ {
+ full_view = Underlying_Full_View (full_view);
+ save_gnu_tree (full_view, NULL_TREE, false);
+ }
+ }
+
+ if (Is_Type (gnat_entity)
+ && Present (Class_Wide_Type (gnat_entity))
+ && Root_Type (Class_Wide_Type (gnat_entity)) == gnat_entity)
+ save_gnu_tree (Class_Wide_Type (gnat_entity), NULL_TREE, false);
+ }
+
+ if (Is_Incomplete_Or_Private_Type (gnat_entity)
+ && Present (Full_View (gnat_entity)))
+ {
+ Entity_Id full_view = Full_View (gnat_entity);
+
+ if (Is_Private_Type (full_view)
+ && Present (Underlying_Full_View (full_view)))
+ full_view = Underlying_Full_View (full_view);
+
+ gnu_new = gnat_to_gnu_entity (full_view, NULL_TREE, true);
+
+ /* Propagate back-annotations from full view to partial view. */
+ if (!Known_Alignment (gnat_entity))
+ Copy_Alignment (gnat_entity, full_view);
+
+ if (!Known_Esize (gnat_entity))
+ Copy_Esize (gnat_entity, full_view);
+
+ if (!Known_RM_Size (gnat_entity))
+ Copy_RM_Size (gnat_entity, full_view);
+
+ /* The above call may have defined this entity (the simplest example
+ of this is when we have a private enumeral type since the bounds
+ will have the public view). */
+ if (!present_gnu_tree (gnat_entity))
+ save_gnu_tree (gnat_entity, gnu_new, false);
+ }
+ else
+ {
+ tree gnu_init
+ = (Nkind (Declaration_Node (gnat_entity)) == N_Object_Declaration
+ && present_gnu_tree (Declaration_Node (gnat_entity)))
+ ? get_gnu_tree (Declaration_Node (gnat_entity)) : NULL_TREE;
+
+ gnu_new = gnat_to_gnu_entity (gnat_entity, gnu_init, true);
+ }
+
+ if (Is_Type (gnat_entity)
+ && Present (Class_Wide_Type (gnat_entity))
+ && Root_Type (Class_Wide_Type (gnat_entity)) == gnat_entity)
+ save_gnu_tree (Class_Wide_Type (gnat_entity), gnu_new, false);
+
+ /* If we have an old type and we've made pointers to this type, update those
+ pointers. If this is a Taft amendment type in the main unit, we need to
+ mark the type as used since other units referencing it don't see the full
+ declaration and, therefore, cannot mark it as used themselves. */
+ if (gnu_old)
+ {
+ update_pointer_to (TYPE_MAIN_VARIANT (TREE_TYPE (gnu_old)),
+ TREE_TYPE (gnu_new));
+ if (TYPE_DUMMY_IN_PROFILE_P (TREE_TYPE (gnu_old)))
+ update_profiles_with (TREE_TYPE (gnu_old));
+ if (DECL_TAFT_TYPE_P (gnu_old))
+ used_types_insert (TREE_TYPE (gnu_new));
+ }
+}
+
+/* Elaborate decls in the lists GNAT_DECLS and GNAT_DECLS2, if present.
+ We make two passes, one to elaborate anything other than bodies (but
+ we declare a function if there was no spec). The second pass
+ elaborates the bodies.
+
+ GNAT_END_LIST gives the element in the list past the end. Normally,
+ this is Empty, but can be First_Real_Statement for a
+ Handled_Sequence_Of_Statements.
+
+ We make a complete pass through both lists if PASS1P is true, then make
+ the second pass over both lists if PASS2P is true. The lists usually
+ correspond to the public and private parts of a package. */
+
+static void
+process_decls (List_Id gnat_decls, List_Id gnat_decls2,
+ Node_Id gnat_end_list, bool pass1p, bool pass2p)
+{
+ List_Id gnat_decl_array[2];
+ Node_Id gnat_decl;
+ int i;
+
+ gnat_decl_array[0] = gnat_decls, gnat_decl_array[1] = gnat_decls2;
+
+ if (pass1p)
+ for (i = 0; i <= 1; i++)
+ if (Present (gnat_decl_array[i]))
+ for (gnat_decl = First (gnat_decl_array[i]);
+ gnat_decl != gnat_end_list; gnat_decl = Next (gnat_decl))
+ {
+ /* For package specs, we recurse inside the declarations,
+ thus taking the two pass approach inside the boundary. */
+ if (Nkind (gnat_decl) == N_Package_Declaration
+ && (Nkind (Specification (gnat_decl)
+ == N_Package_Specification)))
+ process_decls (Visible_Declarations (Specification (gnat_decl)),
+ Private_Declarations (Specification (gnat_decl)),
+ Empty, true, false);
+
+ /* Similarly for any declarations in the actions of a
+ freeze node. */
+ else if (Nkind (gnat_decl) == N_Freeze_Entity)
+ {
+ process_freeze_entity (gnat_decl);
+ process_decls (Actions (gnat_decl), Empty, Empty, true, false);
+ }
+
+ /* Package bodies with freeze nodes get their elaboration deferred
+ until the freeze node, but the code must be placed in the right
+ place, so record the code position now. */
+ else if (Nkind (gnat_decl) == N_Package_Body
+ && Present (Freeze_Node (Corresponding_Spec (gnat_decl))))
+ record_code_position (gnat_decl);
+
+ else if (Nkind (gnat_decl) == N_Package_Body_Stub
+ && Present (Library_Unit (gnat_decl))
+ && Present (Freeze_Node
+ (Corresponding_Spec
+ (Proper_Body (Unit
+ (Library_Unit (gnat_decl)))))))
+ record_code_position
+ (Proper_Body (Unit (Library_Unit (gnat_decl))));
+
+ /* We defer most subprogram bodies to the second pass. */
+ else if (Nkind (gnat_decl) == N_Subprogram_Body)
+ {
+ if (Acts_As_Spec (gnat_decl))
+ {
+ Node_Id gnat_subprog_id = Defining_Entity (gnat_decl);
+
+ if (Ekind (gnat_subprog_id) != E_Generic_Procedure
+ && Ekind (gnat_subprog_id) != E_Generic_Function)
+ gnat_to_gnu_entity (gnat_subprog_id, NULL_TREE, true);
+ }
+ }
+
+ /* For bodies and stubs that act as their own specs, the entity
+ itself must be elaborated in the first pass, because it may
+ be used in other declarations. */
+ else if (Nkind (gnat_decl) == N_Subprogram_Body_Stub)
+ {
+ Node_Id gnat_subprog_id
+ = Defining_Entity (Specification (gnat_decl));
+
+ if (Ekind (gnat_subprog_id) != E_Subprogram_Body
+ && Ekind (gnat_subprog_id) != E_Generic_Procedure
+ && Ekind (gnat_subprog_id) != E_Generic_Function)
+ gnat_to_gnu_entity (gnat_subprog_id, NULL_TREE, true);
+ }
+
+ /* Concurrent stubs stand for the corresponding subprogram bodies,
+ which are deferred like other bodies. */
+ else if (Nkind (gnat_decl) == N_Task_Body_Stub
+ || Nkind (gnat_decl) == N_Protected_Body_Stub)
+ ;
+
+ /* Renamed subprograms may not be elaborated yet at this point
+ since renamings do not trigger freezing. Wait for the second
+ pass to take care of them. */
+ else if (Nkind (gnat_decl) == N_Subprogram_Renaming_Declaration)
+ ;
+
+ else
+ add_stmt (gnat_to_gnu (gnat_decl));
+ }
+
+ /* Here we elaborate everything we deferred above except for package bodies,
+ which are elaborated at their freeze nodes. Note that we must also
+ go inside things (package specs and freeze nodes) the first pass did. */
+ if (pass2p)
+ for (i = 0; i <= 1; i++)
+ if (Present (gnat_decl_array[i]))
+ for (gnat_decl = First (gnat_decl_array[i]);
+ gnat_decl != gnat_end_list; gnat_decl = Next (gnat_decl))
+ {
+ if (Nkind (gnat_decl) == N_Subprogram_Body
+ || Nkind (gnat_decl) == N_Subprogram_Body_Stub
+ || Nkind (gnat_decl) == N_Task_Body_Stub
+ || Nkind (gnat_decl) == N_Protected_Body_Stub)
+ add_stmt (gnat_to_gnu (gnat_decl));
+
+ else if (Nkind (gnat_decl) == N_Package_Declaration
+ && (Nkind (Specification (gnat_decl)
+ == N_Package_Specification)))
+ process_decls (Visible_Declarations (Specification (gnat_decl)),
+ Private_Declarations (Specification (gnat_decl)),
+ Empty, false, true);
+
+ else if (Nkind (gnat_decl) == N_Freeze_Entity)
+ process_decls (Actions (gnat_decl), Empty, Empty, false, true);
+
+ else if (Nkind (gnat_decl) == N_Subprogram_Renaming_Declaration)
+ add_stmt (gnat_to_gnu (gnat_decl));
+ }
+}
+
+/* Make a unary operation of kind CODE using build_unary_op, but guard
+ the operation by an overflow check. CODE can be one of NEGATE_EXPR
+ or ABS_EXPR. GNU_TYPE is the type desired for the result. Usually
+ the operation is to be performed in that type. GNAT_NODE is the gnat
+ node conveying the source location for which the error should be
+ signaled. */
+
+static tree
+build_unary_op_trapv (enum tree_code code, tree gnu_type, tree operand,
+ Node_Id gnat_node)
+{
+ gcc_assert (code == NEGATE_EXPR || code == ABS_EXPR);
+
+ operand = gnat_protect_expr (operand);
+
+ return emit_check (build_binary_op (EQ_EXPR, boolean_type_node,
+ operand, TYPE_MIN_VALUE (gnu_type)),
+ build_unary_op (code, gnu_type, operand),
+ CE_Overflow_Check_Failed, gnat_node);
+}
+
+/* Make a binary operation of kind CODE using build_binary_op, but guard
+ the operation by an overflow check. CODE can be one of PLUS_EXPR,
+ MINUS_EXPR or MULT_EXPR. GNU_TYPE is the type desired for the result.
+ Usually the operation is to be performed in that type. GNAT_NODE is
+ the GNAT node conveying the source location for which the error should
+ be signaled. */
+
+static tree
+build_binary_op_trapv (enum tree_code code, tree gnu_type, tree left,
+ tree right, Node_Id gnat_node)
+{
+ const unsigned int precision = TYPE_PRECISION (gnu_type);
+ tree lhs = gnat_protect_expr (left);
+ tree rhs = gnat_protect_expr (right);
+ tree type_max = TYPE_MAX_VALUE (gnu_type);
+ tree type_min = TYPE_MIN_VALUE (gnu_type);
+ tree gnu_expr, check;
+ int sgn;
+
+ /* Assert that the precision is a power of 2. */
+ gcc_assert ((precision & (precision - 1)) == 0);
+
+ /* Prefer a constant on the RHS to simplify checks. */
+ if (TREE_CODE (rhs) != INTEGER_CST
+ && TREE_CODE (lhs) == INTEGER_CST
+ && (code == PLUS_EXPR || code == MULT_EXPR))
+ {
+ tree tmp = lhs;
+ lhs = rhs;
+ rhs = tmp;
+ }
+
+ gnu_expr = build_binary_op (code, gnu_type, lhs, rhs);
+
+ /* If we can fold the expression to a constant, just return it.
+ The caller will deal with overflow, no need to generate a check. */
+ if (TREE_CODE (gnu_expr) == INTEGER_CST)
+ return gnu_expr;
+
+ /* If no operand is a constant, we use the generic implementation. */
+ if (TREE_CODE (lhs) != INTEGER_CST && TREE_CODE (rhs) != INTEGER_CST)
+ {
+ /* First convert the operands to the result type like build_binary_op.
+ This is where the bias is made explicit for biased types. */
+ lhs = convert (gnu_type, lhs);
+ rhs = convert (gnu_type, rhs);
+
+ /* Never inline a 64-bit mult for a 32-bit target, it's way too long. */
+ if (code == MULT_EXPR && precision == 64 && BITS_PER_WORD < 64)
+ {
+ tree int64 = gnat_type_for_size (64, 0);
+ return convert (gnu_type, build_call_n_expr (mulv64_decl, 2,
+ convert (int64, lhs),
+ convert (int64, rhs)));
+ }
+
+ /* Likewise for a 128-bit mult and a 64-bit target. */
+ else if (code == MULT_EXPR && precision == 128 && BITS_PER_WORD < 128)
+ {
+ tree int128 = gnat_type_for_size (128, 0);
+ return convert (gnu_type, build_call_n_expr (mulv128_decl, 2,
+ convert (int128, lhs),
+ convert (int128, rhs)));
+ }
+
+ enum internal_fn icode;
+
+ switch (code)
+ {
+ case PLUS_EXPR:
+ icode = IFN_ADD_OVERFLOW;
+ break;
+ case MINUS_EXPR:
+ icode = IFN_SUB_OVERFLOW;
+ break;
+ case MULT_EXPR:
+ icode = IFN_MUL_OVERFLOW;
+ break;
+ default:
+ gcc_unreachable ();
+ }
+
+ tree gnu_ctype = build_complex_type (gnu_type);
+ tree call
+ = build_call_expr_internal_loc (UNKNOWN_LOCATION, icode, gnu_ctype, 2,
+ lhs, rhs);
+ tree tgt = save_expr (call);
+ gnu_expr = build1 (REALPART_EXPR, gnu_type, tgt);
+ check = fold_build2 (NE_EXPR, boolean_type_node,
+ build1 (IMAGPART_EXPR, gnu_type, tgt),
+ build_int_cst (gnu_type, 0));
+ return
+ emit_check (check, gnu_expr, CE_Overflow_Check_Failed, gnat_node);
+ }
+
+ /* If one operand is a constant, we expose the overflow condition to enable
+ a subsequent simplication or even elimination. */
+ switch (code)
+ {
+ case PLUS_EXPR:
+ sgn = tree_int_cst_sgn (rhs);
+ if (sgn > 0)
+ /* When rhs > 0, overflow when lhs > type_max - rhs. */
+ check = build_binary_op (GT_EXPR, boolean_type_node, lhs,
+ build_binary_op (MINUS_EXPR, gnu_type,
+ type_max, rhs));
+ else if (sgn < 0)
+ /* When rhs < 0, overflow when lhs < type_min - rhs. */
+ check = build_binary_op (LT_EXPR, boolean_type_node, lhs,
+ build_binary_op (MINUS_EXPR, gnu_type,
+ type_min, rhs));
+ else
+ return gnu_expr;
+ break;
+
+ case MINUS_EXPR:
+ if (TREE_CODE (lhs) == INTEGER_CST)
+ {
+ sgn = tree_int_cst_sgn (lhs);
+ if (sgn > 0)
+ /* When lhs > 0, overflow when rhs < lhs - type_max. */
+ check = build_binary_op (LT_EXPR, boolean_type_node, rhs,
+ build_binary_op (MINUS_EXPR, gnu_type,
+ lhs, type_max));
+ else if (sgn < 0)
+ /* When lhs < 0, overflow when rhs > lhs - type_min. */
+ check = build_binary_op (GT_EXPR, boolean_type_node, rhs,
+ build_binary_op (MINUS_EXPR, gnu_type,
+ lhs, type_min));
+ else
+ return gnu_expr;
+ }
+ else
+ {
+ sgn = tree_int_cst_sgn (rhs);
+ if (sgn > 0)
+ /* When rhs > 0, overflow when lhs < type_min + rhs. */
+ check = build_binary_op (LT_EXPR, boolean_type_node, lhs,
+ build_binary_op (PLUS_EXPR, gnu_type,
+ type_min, rhs));
+ else if (sgn < 0)
+ /* When rhs < 0, overflow when lhs > type_max + rhs. */
+ check = build_binary_op (GT_EXPR, boolean_type_node, lhs,
+ build_binary_op (PLUS_EXPR, gnu_type,
+ type_max, rhs));
+ else
+ return gnu_expr;
+ }
+ break;
+
+ case MULT_EXPR:
+ sgn = tree_int_cst_sgn (rhs);
+ if (sgn > 0)
+ {
+ if (integer_onep (rhs))
+ return gnu_expr;
+
+ tree lb = build_binary_op (TRUNC_DIV_EXPR, gnu_type, type_min, rhs);
+ tree ub = build_binary_op (TRUNC_DIV_EXPR, gnu_type, type_max, rhs);
+
+ /* When rhs > 1, overflow outside [type_min/rhs; type_max/rhs]. */
+ check
+ = build_binary_op (TRUTH_ORIF_EXPR, boolean_type_node,
+ build_binary_op (LT_EXPR, boolean_type_node,
+ lhs, lb),
+ build_binary_op (GT_EXPR, boolean_type_node,
+ lhs, ub));
+ }
+ else if (sgn < 0)
+ {
+ tree lb = build_binary_op (TRUNC_DIV_EXPR, gnu_type, type_max, rhs);
+ tree ub = build_binary_op (TRUNC_DIV_EXPR, gnu_type, type_min, rhs);
+
+ if (integer_minus_onep (rhs))
+ /* When rhs == -1, overflow if lhs == type_min. */
+ check
+ = build_binary_op (EQ_EXPR, boolean_type_node, lhs, type_min);
+ else
+ /* When rhs < -1, overflow outside [type_max/rhs; type_min/rhs]. */
+ check
+ = build_binary_op (TRUTH_ORIF_EXPR, boolean_type_node,
+ build_binary_op (LT_EXPR, boolean_type_node,
+ lhs, lb),
+ build_binary_op (GT_EXPR, boolean_type_node,
+ lhs, ub));
+ }
+ else
+ return gnu_expr;
+ break;
+
+ default:
+ gcc_unreachable ();
+ }
+
+ return emit_check (check, gnu_expr, CE_Overflow_Check_Failed, gnat_node);
+}
+
+/* GNU_COND contains the condition corresponding to an index, overflow or
+ range check of value GNU_EXPR. Build a COND_EXPR that returns GNU_EXPR
+ if GNU_COND is false and raises a CONSTRAINT_ERROR if GNU_COND is true.
+ REASON is the code that says why the exception is raised. GNAT_NODE is
+ the node conveying the source location for which the error should be
+ signaled.
+
+ We used to propagate TREE_SIDE_EFFECTS from GNU_EXPR to the COND_EXPR,
+ overwriting the setting inherited from the call statement, on the ground
+ that the expression need not be evaluated just for the check. However
+ that's incorrect because, in the GCC type system, its value is presumed
+ to be valid so its comparison against the type bounds always yields true
+ and, therefore, could be done without evaluating it; given that it can
+ be a computation that overflows the bounds, the language may require the
+ check to fail and thus the expression to be evaluated in this case. */
+
+static tree
+emit_check (tree gnu_cond, tree gnu_expr, int reason, Node_Id gnat_node)
+{
+ tree gnu_call
+ = build_call_raise (reason, gnat_node, N_Raise_Constraint_Error);
+ return
+ fold_build3 (COND_EXPR, TREE_TYPE (gnu_expr), gnu_cond,
+ build2 (COMPOUND_EXPR, TREE_TYPE (gnu_expr), gnu_call,
+ SCALAR_FLOAT_TYPE_P (TREE_TYPE (gnu_expr))
+ ? build_real (TREE_TYPE (gnu_expr), dconst0)
+ : build_int_cst (TREE_TYPE (gnu_expr), 0)),
+ gnu_expr);
+}
+
+/* Return an expression that converts GNU_EXPR to GNAT_TYPE, doing overflow
+ checks if OVERFLOW_P is true. If TRUNCATE_P is true, do a fp-to-integer
+ conversion with truncation, otherwise round. GNAT_NODE is the GNAT node
+ conveying the source location for which the error should be signaled. */
+
+static tree
+convert_with_check (Entity_Id gnat_type, tree gnu_expr, bool overflow_p,
+ bool truncate_p, Node_Id gnat_node)
+{
+ tree gnu_type = get_unpadded_type (gnat_type);
+ tree gnu_base_type = get_base_type (gnu_type);
+ tree gnu_in_type = TREE_TYPE (gnu_expr);
+ tree gnu_in_base_type = get_base_type (gnu_in_type);
+ tree gnu_result = gnu_expr;
+
+ /* If we are not doing any checks, the output is an integral type and the
+ input is not a floating-point type, just do the conversion. This is
+ required for packed array types and is simpler in all cases anyway. */
+ if (!overflow_p
+ && INTEGRAL_TYPE_P (gnu_base_type)
+ && !FLOAT_TYPE_P (gnu_in_base_type))
+ return convert (gnu_type, gnu_expr);
+
+ /* If the mode of the input base type is larger, then converting to it below
+ may pessimize the final conversion step, for example generate a libcall
+ instead of a simple instruction, so use a narrower type in this case. */
+ if (TYPE_MODE (gnu_in_base_type) != TYPE_MODE (gnu_in_type)
+ && !(TREE_CODE (gnu_in_type) == INTEGER_TYPE
+ && TYPE_BIASED_REPRESENTATION_P (gnu_in_type)))
+ gnu_in_base_type = gnat_type_for_mode (TYPE_MODE (gnu_in_type),
+ TYPE_UNSIGNED (gnu_in_type));
+
+ /* First convert the expression to the base type. This will never generate
+ code, but makes the tests below simpler. But don't do this if converting
+ from an integer type to an unconstrained array type since then we need to
+ get the bounds from the original (unpacked) type. */
+ if (TREE_CODE (gnu_type) != UNCONSTRAINED_ARRAY_TYPE)
+ gnu_result = convert (gnu_in_base_type, gnu_result);
+
+ /* If overflow checks are requested, we need to be sure the result will fit
+ in the output base type. But don't do this if the input is integer and
+ the output floating-point. */
+ if (overflow_p
+ && !(FLOAT_TYPE_P (gnu_base_type) && INTEGRAL_TYPE_P (gnu_in_base_type)))
+ {
+ /* Ensure GNU_EXPR only gets evaluated once. */
+ tree gnu_input = gnat_protect_expr (gnu_result);
+ tree gnu_cond = boolean_false_node;
+ tree gnu_in_lb = TYPE_MIN_VALUE (gnu_in_base_type);
+ tree gnu_in_ub = TYPE_MAX_VALUE (gnu_in_base_type);
+ tree gnu_out_lb = TYPE_MIN_VALUE (gnu_base_type);
+ tree gnu_out_ub
+ = (TREE_CODE (gnu_base_type) == INTEGER_TYPE
+ && TYPE_MODULAR_P (gnu_base_type))
+ ? fold_build2 (MINUS_EXPR, gnu_base_type,
+ TYPE_MODULUS (gnu_base_type),
+ build_int_cst (gnu_base_type, 1))
+ : TYPE_MAX_VALUE (gnu_base_type);
+
+ /* Convert the lower bounds to signed types, so we're sure we're
+ comparing them properly. Likewise, convert the upper bounds
+ to unsigned types. */
+ if (INTEGRAL_TYPE_P (gnu_in_base_type)
+ && TYPE_UNSIGNED (gnu_in_base_type))
+ gnu_in_lb
+ = convert (gnat_signed_type_for (gnu_in_base_type), gnu_in_lb);
+
+ if (INTEGRAL_TYPE_P (gnu_in_base_type)
+ && !TYPE_UNSIGNED (gnu_in_base_type))
+ gnu_in_ub
+ = convert (gnat_unsigned_type_for (gnu_in_base_type), gnu_in_ub);
+
+ if (INTEGRAL_TYPE_P (gnu_base_type) && TYPE_UNSIGNED (gnu_base_type))
+ gnu_out_lb
+ = convert (gnat_signed_type_for (gnu_base_type), gnu_out_lb);
+
+ if (INTEGRAL_TYPE_P (gnu_base_type) && !TYPE_UNSIGNED (gnu_base_type))
+ gnu_out_ub
+ = convert (gnat_unsigned_type_for (gnu_base_type), gnu_out_ub);
+
+ /* Check each bound separately and only if the result bound
+ is tighter than the bound on the input type. Note that all the
+ types are base types, so the bounds must be constant. Also,
+ the comparison is done in the base type of the input, which
+ always has the proper signedness. First check for input
+ integer (which means output integer), output float (which means
+ both float), or mixed, in which case we always compare.
+ Note that we have to do the comparison which would *fail* in the
+ case of an error since if it's an FP comparison and one of the
+ values is a NaN or Inf, the comparison will fail. */
+ if (INTEGRAL_TYPE_P (gnu_in_base_type)
+ ? tree_int_cst_lt (gnu_in_lb, gnu_out_lb)
+ : (FLOAT_TYPE_P (gnu_base_type)
+ ? real_less (&TREE_REAL_CST (gnu_in_lb),
+ &TREE_REAL_CST (gnu_out_lb))
+ : 1))
+ gnu_cond
+ = invert_truthvalue
+ (build_binary_op (GE_EXPR, boolean_type_node,
+ gnu_input, convert (gnu_in_base_type,
+ gnu_out_lb)));
+
+ if (INTEGRAL_TYPE_P (gnu_in_base_type)
+ ? tree_int_cst_lt (gnu_out_ub, gnu_in_ub)
+ : (FLOAT_TYPE_P (gnu_base_type)
+ ? real_less (&TREE_REAL_CST (gnu_out_ub),
+ &TREE_REAL_CST (gnu_in_ub))
+ : 1))
+ gnu_cond
+ = build_binary_op (TRUTH_ORIF_EXPR, boolean_type_node, gnu_cond,
+ invert_truthvalue
+ (build_binary_op (LE_EXPR, boolean_type_node,
+ gnu_input,
+ convert (gnu_in_base_type,
+ gnu_out_ub))));
+
+ if (!integer_zerop (gnu_cond))
+ gnu_result = emit_check (gnu_cond, gnu_input,
+ CE_Overflow_Check_Failed, gnat_node);
+ }
+
+ /* Now convert to the result base type. If this is a non-truncating
+ float-to-integer conversion, round. */
+ if (INTEGRAL_TYPE_P (gnu_base_type)
+ && FLOAT_TYPE_P (gnu_in_base_type)
+ && !truncate_p)
+ {
+ REAL_VALUE_TYPE half_minus_pred_half, pred_half;
+ tree gnu_conv, gnu_zero, gnu_comp, calc_type;
+ tree gnu_pred_half, gnu_add_pred_half, gnu_subtract_pred_half;
+ const struct real_format *fmt;
+
+ /* The following calculations depend on proper rounding to even
+ of each arithmetic operation. In order to prevent excess
+ precision from spoiling this property, use the widest hardware
+ floating-point type if FP_ARITH_MAY_WIDEN is true. */
+ calc_type
+ = fp_arith_may_widen ? longest_float_type_node : gnu_in_base_type;
+
+ /* Compute the exact value calc_type'Pred (0.5) at compile time. */
+ fmt = REAL_MODE_FORMAT (TYPE_MODE (calc_type));
+ real_2expN (&half_minus_pred_half, -(fmt->p) - 1, TYPE_MODE (calc_type));
+ real_arithmetic (&pred_half, MINUS_EXPR, &dconsthalf,
+ &half_minus_pred_half);
+ gnu_pred_half = build_real (calc_type, pred_half);
+
+ /* If the input is strictly negative, subtract this value
+ and otherwise add it from the input. For 0.5, the result
+ is exactly between 1.0 and the machine number preceding 1.0
+ (for calc_type). Since the last bit of 1.0 is even, this 0.5
+ will round to 1.0, while all other number with an absolute
+ value less than 0.5 round to 0.0. For larger numbers exactly
+ halfway between integers, rounding will always be correct as
+ the true mathematical result will be closer to the higher
+ integer compared to the lower one. So, this constant works
+ for all floating-point numbers.
+
+ The reason to use the same constant with subtract/add instead
+ of a positive and negative constant is to allow the comparison
+ to be scheduled in parallel with retrieval of the constant and
+ conversion of the input to the calc_type (if necessary). */
+
+ gnu_zero = build_real (gnu_in_base_type, dconst0);
+ gnu_result = gnat_protect_expr (gnu_result);
+ gnu_conv = convert (calc_type, gnu_result);
+ gnu_comp
+ = fold_build2 (GE_EXPR, boolean_type_node, gnu_result, gnu_zero);
+ gnu_add_pred_half
+ = fold_build2 (PLUS_EXPR, calc_type, gnu_conv, gnu_pred_half);
+ gnu_subtract_pred_half
+ = fold_build2 (MINUS_EXPR, calc_type, gnu_conv, gnu_pred_half);
+ gnu_result = fold_build3 (COND_EXPR, calc_type, gnu_comp,
+ gnu_add_pred_half, gnu_subtract_pred_half);
+ }
+
+ if (TREE_CODE (gnu_base_type) == INTEGER_TYPE
+ && TYPE_HAS_ACTUAL_BOUNDS_P (gnu_base_type)
+ && TREE_CODE (gnu_result) == UNCONSTRAINED_ARRAY_REF)
+ gnu_result = unchecked_convert (gnu_base_type, gnu_result, false);
+ else
+ gnu_result = convert (gnu_base_type, gnu_result);
+
+ return convert (gnu_type, gnu_result);
+}
+
+/* Return true if GNU_EXPR can be directly addressed. This is the case
+ unless it is an expression involving computation or if it involves a
+ reference to a bitfield or to an object not sufficiently aligned for
+ its type. If GNU_TYPE is non-null, return true only if GNU_EXPR can
+ be directly addressed as an object of this type.
+
+ *** Notes on addressability issues in the Ada compiler ***
+
+ This predicate is necessary in order to bridge the gap between Gigi
+ and the middle-end about addressability of GENERIC trees. A tree
+ is said to be addressable if it can be directly addressed, i.e. if
+ its address can be taken, is a multiple of the type's alignment on
+ strict-alignment architectures and returns the first storage unit
+ assigned to the object represented by the tree.
+
+ In the C family of languages, everything is in practice addressable
+ at the language level, except for bit-fields. This means that these
+ compilers will take the address of any tree that doesn't represent
+ a bit-field reference and expect the result to be the first storage
+ unit assigned to the object. Even in cases where this will result
+ in unaligned accesses at run time, nothing is supposed to be done
+ and the program is considered as erroneous instead (see PR c/18287).
+
+ The implicit assumptions made in the middle-end are in keeping with
+ the C viewpoint described above:
+ - the address of a bit-field reference is supposed to be never
+ taken; the compiler (generally) will stop on such a construct,
+ - any other tree is addressable if it is formally addressable,
+ i.e. if it is formally allowed to be the operand of ADDR_EXPR.
+
+ In Ada, the viewpoint is the opposite one: nothing is addressable
+ at the language level unless explicitly declared so. This means
+ that the compiler will both make sure that the trees representing
+ references to addressable ("aliased" in Ada parlance) objects are
+ addressable and make no real attempts at ensuring that the trees
+ representing references to non-addressable objects are addressable.
+
+ In the first case, Ada is effectively equivalent to C and handing
+ down the direct result of applying ADDR_EXPR to these trees to the
+ middle-end works flawlessly. In the second case, Ada cannot afford
+ to consider the program as erroneous if the address of trees that
+ are not addressable is requested for technical reasons, unlike C;
+ as a consequence, the Ada compiler must arrange for either making
+ sure that this address is not requested in the middle-end or for
+ compensating by inserting temporaries if it is requested in Gigi.
+
+ The first goal can be achieved because the middle-end should not
+ request the address of non-addressable trees on its own; the only
+ exception is for the invocation of low-level block operations like
+ memcpy, for which the addressability requirements are lower since
+ the type's alignment can be disregarded. In practice, this means
+ that Gigi must make sure that such operations cannot be applied to
+ non-BLKmode bit-fields.
+
+ The second goal is achieved by means of the addressable_p predicate,
+ which computes whether a temporary must be inserted by Gigi when the
+ address of a tree is requested; if so, the address of the temporary
+ will be used in lieu of that of the original tree and some glue code
+ generated to connect everything together. */
+
+static bool
+addressable_p (tree gnu_expr, tree gnu_type)
+{
+ /* For an integral type, the size of the actual type of the object may not
+ be greater than that of the expected type, otherwise an indirect access
+ in the latter type wouldn't correctly set all the bits of the object. */
+ if (gnu_type
+ && INTEGRAL_TYPE_P (gnu_type)
+ && smaller_form_type_p (gnu_type, TREE_TYPE (gnu_expr)))
+ return false;
+
+ /* The size of the actual type of the object may not be smaller than that
+ of the expected type, otherwise an indirect access in the latter type
+ would be larger than the object. But only record types need to be
+ considered in practice for this case. */
+ if (gnu_type
+ && TREE_CODE (gnu_type) == RECORD_TYPE
+ && smaller_form_type_p (TREE_TYPE (gnu_expr), gnu_type))
+ return false;
+
+ switch (TREE_CODE (gnu_expr))
+ {
+ case VAR_DECL:
+ case PARM_DECL:
+ case FUNCTION_DECL:
+ case RESULT_DECL:
+ /* All DECLs are addressable: if they are in a register, we can force
+ them to memory. */
+ return true;
+
+ case UNCONSTRAINED_ARRAY_REF:
+ case INDIRECT_REF:
+ /* Taking the address of a dereference yields the original pointer. */
+ return true;
+
+ case STRING_CST:
+ case INTEGER_CST:
+ case REAL_CST:
+ /* Taking the address yields a pointer to the constant pool. */
+ return true;
+
+ case CONSTRUCTOR:
+ /* Taking the address of a static constructor yields a pointer to the
+ tree constant pool. */
+ return TREE_STATIC (gnu_expr) ? true : false;
+
+ case NULL_EXPR:
+ case ADDR_EXPR:
+ case SAVE_EXPR:
+ case CALL_EXPR:
+ case PLUS_EXPR:
+ case MINUS_EXPR:
+ case BIT_IOR_EXPR:
+ case BIT_XOR_EXPR:
+ case BIT_AND_EXPR:
+ case BIT_NOT_EXPR:
+ /* All rvalues are deemed addressable since taking their address will
+ force a temporary to be created by the middle-end. */
+ return true;
+
+ case COMPOUND_EXPR:
+ /* The address of a compound expression is that of its 2nd operand. */
+ return addressable_p (TREE_OPERAND (gnu_expr, 1), gnu_type);
+
+ case COND_EXPR:
+ /* We accept &COND_EXPR as soon as both operands are addressable and
+ expect the outcome to be the address of the selected operand. */
+ return (addressable_p (TREE_OPERAND (gnu_expr, 1), NULL_TREE)
+ && addressable_p (TREE_OPERAND (gnu_expr, 2), NULL_TREE));
+
+ case COMPONENT_REF:
+ return (((!DECL_BIT_FIELD (TREE_OPERAND (gnu_expr, 1))
+ /* Even with DECL_BIT_FIELD cleared, we have to ensure that
+ the field is sufficiently aligned, in case it is subject
+ to a pragma Component_Alignment. But we don't need to
+ check the alignment of the containing record, as it is
+ guaranteed to be not smaller than that of its most
+ aligned field that is not a bit-field. */
+ && (!STRICT_ALIGNMENT
+ || DECL_ALIGN (TREE_OPERAND (gnu_expr, 1))
+ >= TYPE_ALIGN (TREE_TYPE (gnu_expr))))
+ /* The field of a padding record is always addressable. */
+ || TYPE_IS_PADDING_P (TREE_TYPE (TREE_OPERAND (gnu_expr, 0))))
+ && addressable_p (TREE_OPERAND (gnu_expr, 0), NULL_TREE));
+
+ case ARRAY_REF: case ARRAY_RANGE_REF:
+ case REALPART_EXPR: case IMAGPART_EXPR:
+ case NOP_EXPR:
+ return addressable_p (TREE_OPERAND (gnu_expr, 0), NULL_TREE);
+
+ case CONVERT_EXPR:
+ return (AGGREGATE_TYPE_P (TREE_TYPE (gnu_expr))
+ && addressable_p (TREE_OPERAND (gnu_expr, 0), NULL_TREE));
+
+ case VIEW_CONVERT_EXPR:
+ {
+ /* This is addressable if we can avoid a copy. */
+ tree type = TREE_TYPE (gnu_expr);
+ tree inner_type = TREE_TYPE (TREE_OPERAND (gnu_expr, 0));
+ return (((TYPE_MODE (type) == TYPE_MODE (inner_type)
+ && (!STRICT_ALIGNMENT
+ || TYPE_ALIGN (type) <= TYPE_ALIGN (inner_type)
+ || TYPE_ALIGN (inner_type) >= BIGGEST_ALIGNMENT))
+ || ((TYPE_MODE (type) == BLKmode
+ || TYPE_MODE (inner_type) == BLKmode)
+ && (!STRICT_ALIGNMENT
+ || TYPE_ALIGN (type) <= TYPE_ALIGN (inner_type)
+ || TYPE_ALIGN (inner_type) >= BIGGEST_ALIGNMENT
+ || TYPE_ALIGN_OK (type)
+ || TYPE_ALIGN_OK (inner_type))))
+ && addressable_p (TREE_OPERAND (gnu_expr, 0), NULL_TREE));
+ }
+
+ default:
+ return false;
+ }
+}
+
+/* Do the processing for the declaration of a GNAT_ENTITY, a type or subtype.
+ If a Freeze node exists for the entity, delay the bulk of the processing.
+ Otherwise make a GCC type for GNAT_ENTITY and set up the correspondence. */
+
+void
+process_type (Entity_Id gnat_entity)
+{
+ tree gnu_old
+ = present_gnu_tree (gnat_entity) ? get_gnu_tree (gnat_entity) : NULL_TREE;
+
+ /* If we are to delay elaboration of this type, just do any elaboration
+ needed for expressions within the declaration and make a dummy node
+ for it and its Full_View (if any), in case something points to it.
+ Do not do this if it has already been done (the only way that can
+ happen is if the private completion is also delayed). */
+ if (Present (Freeze_Node (gnat_entity)))
+ {
+ elaborate_entity (gnat_entity);
+
+ if (!gnu_old)
+ {
+ tree gnu_decl = TYPE_STUB_DECL (make_dummy_type (gnat_entity));
+ save_gnu_tree (gnat_entity, gnu_decl, false);
+ if (Is_Incomplete_Or_Private_Type (gnat_entity)
+ && Present (Full_View (gnat_entity)))
+ {
+ if (Has_Completion_In_Body (gnat_entity))
+ DECL_TAFT_TYPE_P (gnu_decl) = 1;
+ save_gnu_tree (Full_View (gnat_entity), gnu_decl, false);
+ }
+ }
+
+ return;
+ }
+
+ /* If we saved away a dummy type for this node, it means that this made the
+ type that corresponds to the full type of an incomplete type. Clear that
+ type for now and then update the type in the pointers below. But, if the
+ saved type is not dummy, it very likely means that we have a use before
+ declaration for the type in the tree, what we really cannot handle. */
+ if (gnu_old)
+ {
+ gcc_assert (TREE_CODE (gnu_old) == TYPE_DECL
+ && TYPE_IS_DUMMY_P (TREE_TYPE (gnu_old)));
+
+ save_gnu_tree (gnat_entity, NULL_TREE, false);
+ }
+
+ /* Now fully elaborate the type. */
+ tree gnu_new = gnat_to_gnu_entity (gnat_entity, NULL_TREE, true);
+ gcc_assert (TREE_CODE (gnu_new) == TYPE_DECL);
+
+ /* If we have an old type and we've made pointers to this type, update those
+ pointers. If this is a Taft amendment type in the main unit, we need to
+ mark the type as used since other units referencing it don't see the full
+ declaration and, therefore, cannot mark it as used themselves. */
+ if (gnu_old)
+ {
+ update_pointer_to (TYPE_MAIN_VARIANT (TREE_TYPE (gnu_old)),
+ TREE_TYPE (gnu_new));
+ if (DECL_TAFT_TYPE_P (gnu_old))
+ used_types_insert (TREE_TYPE (gnu_new));
+ }
+
+ /* If this is a record type corresponding to a task or protected type
+ that is a completion of an incomplete type, perform a similar update
+ on the type. ??? Including protected types here is a guess. */
+ if (Is_Record_Type (gnat_entity)
+ && Is_Concurrent_Record_Type (gnat_entity)
+ && present_gnu_tree (Corresponding_Concurrent_Type (gnat_entity)))
+ {
+ tree gnu_task_old
+ = get_gnu_tree (Corresponding_Concurrent_Type (gnat_entity));
+
+ save_gnu_tree (Corresponding_Concurrent_Type (gnat_entity),
+ NULL_TREE, false);
+ save_gnu_tree (Corresponding_Concurrent_Type (gnat_entity),
+ gnu_new, false);
+
+ update_pointer_to (TYPE_MAIN_VARIANT (TREE_TYPE (gnu_task_old)),
+ TREE_TYPE (gnu_new));
+ }
+}
+
+/* Subroutine of assoc_to_constructor: VALUES is a list of field associations,
+ some of which are from RECORD_TYPE. Return a CONSTRUCTOR consisting of the
+ associations that are from RECORD_TYPE. If we see an internal record, make
+ a recursive call to fill it in as well. */
+
+static tree
+extract_values (tree values, tree record_type)
+{
+ vec<constructor_elt, va_gc> *v = NULL;
+ tree field;
+
+ for (field = TYPE_FIELDS (record_type); field; field = DECL_CHAIN (field))
+ {
+ tree tem, value = NULL_TREE;
+
+ /* _Parent is an internal field, but may have values in the aggregate,
+ so check for values first. */
+ if ((tem = purpose_member (field, values)))
+ {
+ value = TREE_VALUE (tem);
+ TREE_ADDRESSABLE (tem) = 1;
+ }
+
+ else if (DECL_INTERNAL_P (field))
+ {
+ value = extract_values (values, TREE_TYPE (field));
+ if (TREE_CODE (value) == CONSTRUCTOR
+ && vec_safe_is_empty (CONSTRUCTOR_ELTS (value)))
+ value = NULL_TREE;
+ }
+ else
+ /* If we have a record subtype, the names will match, but not the
+ actual FIELD_DECLs. */
+ for (tem = values; tem; tem = TREE_CHAIN (tem))
+ if (DECL_NAME (TREE_PURPOSE (tem)) == DECL_NAME (field))
+ {
+ value = convert (TREE_TYPE (field), TREE_VALUE (tem));
+ TREE_ADDRESSABLE (tem) = 1;
+ }
+
+ if (!value)
+ continue;
+
+ CONSTRUCTOR_APPEND_ELT (v, field, value);
+ }
+
+ return gnat_build_constructor (record_type, v);
+}
+
+/* GNAT_ENTITY is the type of the resulting constructor, GNAT_ASSOC is the
+ front of the Component_Associations of an N_Aggregate and GNU_TYPE is the
+ GCC type of the corresponding record type. Return the CONSTRUCTOR. */
+
+static tree
+assoc_to_constructor (Entity_Id gnat_entity, Node_Id gnat_assoc, tree gnu_type)
+{
+ tree gnu_list = NULL_TREE, gnu_result;
+
+ /* We test for GNU_FIELD being empty in the case where a variant
+ was the last thing since we don't take things off GNAT_ASSOC in
+ that case. We check GNAT_ASSOC in case we have a variant, but it
+ has no fields. */
+
+ for (; Present (gnat_assoc); gnat_assoc = Next (gnat_assoc))
+ {
+ const Node_Id gnat_field = First (Choices (gnat_assoc));
+ const Node_Id gnat_expr = Expression (gnat_assoc);
+ tree gnu_field = gnat_to_gnu_field_decl (Entity (gnat_field));
+ tree gnu_expr = gnat_to_gnu (Expression (gnat_assoc));
+
+ /* The expander is supposed to put a single component selector name
+ in every record component association. */
+ gcc_assert (No (Next (gnat_field)));
+
+ /* Ignore discriminants that have Corresponding_Discriminants in tagged
+ types since we'll be setting those fields in the parent subtype. */
+ if (Ekind (Entity (gnat_field)) == E_Discriminant
+ && Present (Corresponding_Discriminant (Entity (gnat_field)))
+ && Is_Tagged_Type (Scope (Entity (gnat_field))))
+ continue;
+
+ /* Also ignore discriminants of Unchecked_Unions. */
+ if (Ekind (Entity (gnat_field)) == E_Discriminant
+ && Is_Unchecked_Union (gnat_entity))
+ continue;
+
+ gigi_checking_assert (!Do_Range_Check (gnat_expr));
+
+ /* Convert to the type of the field. */
+ gnu_expr = convert (TREE_TYPE (gnu_field), gnu_expr);
+
+ /* Add the field and expression to the list. */
+ gnu_list = tree_cons (gnu_field, gnu_expr, gnu_list);
+ }
+
+ gnu_result = extract_values (gnu_list, gnu_type);
+
+ if (flag_checking)
+ {
+ /* Verify that every entry in GNU_LIST was used. */
+ for (; gnu_list; gnu_list = TREE_CHAIN (gnu_list))
+ gcc_assert (TREE_ADDRESSABLE (gnu_list));
+ }
+
+ return gnu_result;
+}
+
+/* Build a possibly nested constructor for array aggregates. GNAT_EXPR is
+ the first element of an array aggregate. It may itself be an aggregate.
+ GNU_ARRAY_TYPE is the GCC type corresponding to the array aggregate. */
+
+static tree
+pos_to_constructor (Node_Id gnat_expr, tree gnu_array_type)
+{
+ tree gnu_index = TYPE_MIN_VALUE (TYPE_DOMAIN (gnu_array_type));
+ vec<constructor_elt, va_gc> *gnu_expr_vec = NULL;
+
+ for (; Present (gnat_expr); gnat_expr = Next (gnat_expr))
+ {
+ tree gnu_expr;
+
+ /* If the expression is itself an array aggregate then first build the
+ innermost constructor if it is part of our array (multi-dimensional
+ case). */
+ if (Nkind (gnat_expr) == N_Aggregate
+ && TREE_CODE (TREE_TYPE (gnu_array_type)) == ARRAY_TYPE
+ && TYPE_MULTI_ARRAY_P (TREE_TYPE (gnu_array_type)))
+ gnu_expr = pos_to_constructor (First (Expressions (gnat_expr)),
+ TREE_TYPE (gnu_array_type));
+ else
+ {
+ /* If the expression is a conversion to an unconstrained array type,
+ skip it to avoid spilling to memory. */
+ if (Nkind (gnat_expr) == N_Type_Conversion
+ && Is_Array_Type (Etype (gnat_expr))
+ && !Is_Constrained (Etype (gnat_expr)))
+ gnu_expr = gnat_to_gnu (Expression (gnat_expr));
+ else
+ gnu_expr = gnat_to_gnu (gnat_expr);
+
+ gigi_checking_assert (!Do_Range_Check (gnat_expr));
+ }
+
+ CONSTRUCTOR_APPEND_ELT (gnu_expr_vec, gnu_index,
+ convert (TREE_TYPE (gnu_array_type), gnu_expr));
+
+ gnu_index = int_const_binop (PLUS_EXPR, gnu_index,
+ convert (TREE_TYPE (gnu_index),
+ integer_one_node));
+ }
+
+ return gnat_build_constructor (gnu_array_type, gnu_expr_vec);
+}
+
+/* Process a N_Validate_Unchecked_Conversion node. */
+
+static void
+validate_unchecked_conversion (Node_Id gnat_node)
+{
+ tree gnu_source_type = gnat_to_gnu_type (Source_Type (gnat_node));
+ tree gnu_target_type = gnat_to_gnu_type (Target_Type (gnat_node));
+
+ /* If the target is a pointer type, see if we are either converting from a
+ non-pointer or from a pointer to a type with a different alias set and
+ warn if so, unless the pointer has been marked to alias everything. */
+ if (POINTER_TYPE_P (gnu_target_type)
+ && !TYPE_REF_CAN_ALIAS_ALL (gnu_target_type))
+ {
+ tree gnu_source_desig_type = POINTER_TYPE_P (gnu_source_type)
+ ? TREE_TYPE (gnu_source_type)
+ : NULL_TREE;
+ tree gnu_target_desig_type = TREE_TYPE (gnu_target_type);
+ alias_set_type target_alias_set = get_alias_set (gnu_target_desig_type);
+
+ if (target_alias_set != 0
+ && (!POINTER_TYPE_P (gnu_source_type)
+ || !alias_sets_conflict_p (get_alias_set (gnu_source_desig_type),
+ target_alias_set)))
+ {
+ post_error_ne ("??possible aliasing problem for type&",
+ gnat_node, Target_Type (gnat_node));
+ post_error ("\\?use -fno-strict-aliasing switch for references",
+ gnat_node);
+ post_error_ne ("\\?or use `pragma No_Strict_Aliasing (&);`",
+ gnat_node, Target_Type (gnat_node));
+ }
+ }
+
+ /* Likewise if the target is a fat pointer type, but we have no mechanism to
+ mitigate the problem in this case, so we unconditionally warn. */
+ else if (TYPE_IS_FAT_POINTER_P (gnu_target_type))
+ {
+ tree gnu_source_desig_type
+ = TYPE_IS_FAT_POINTER_P (gnu_source_type)
+ ? TREE_TYPE (TREE_TYPE (TYPE_FIELDS (gnu_source_type)))
+ : NULL_TREE;
+ tree gnu_target_desig_type
+ = TREE_TYPE (TREE_TYPE (TYPE_FIELDS (gnu_target_type)));
+ alias_set_type target_alias_set = get_alias_set (gnu_target_desig_type);
+
+ if (target_alias_set != 0
+ && (!TYPE_IS_FAT_POINTER_P (gnu_source_type)
+ || !alias_sets_conflict_p (get_alias_set (gnu_source_desig_type),
+ target_alias_set)))
+ {
+ post_error_ne ("??possible aliasing problem for type&",
+ gnat_node, Target_Type (gnat_node));
+ post_error ("\\?use -fno-strict-aliasing switch for references",
+ gnat_node);
+ }
+ }
+}
+
+/* Convert SLOC into LOCUS. Return true if SLOC corresponds to a
+ source code location and false if it doesn't. If CLEAR_COLUMN is
+ true, set the column information to 0. If DECL is given and SLOC
+ refers to a File with an instance, map DECL to that instance. */
+
+bool
+Sloc_to_locus (Source_Ptr Sloc, location_t *locus, bool clear_column,
+ const_tree decl)
+{
+ if (Sloc == No_Location)
+ return false;
+
+ if (Sloc <= Standard_Location)
+ {
+ *locus = BUILTINS_LOCATION;
+ return false;
+ }
+
+ Source_File_Index file = Get_Source_File_Index (Sloc);
+ Line_Number_Type line = Get_Logical_Line_Number (Sloc);
+ Column_Number_Type column = (clear_column ? 0 : Get_Column_Number (Sloc));
+ line_map_ordinary *map = LINEMAPS_ORDINARY_MAP_AT (line_table, file - 1);
+
+ /* We can have zero if pragma Source_Reference is in effect. */
+ if (line < 1)
+ line = 1;
+
+ /* Translate the location. */
+ *locus
+ = linemap_position_for_line_and_column (line_table, map, line, column);
+
+ if (file_map && file_map[file - 1].Instance)
+ decl_to_instance_map->put (decl, file_map[file - 1].Instance);
+
+ return true;
+}
+
+/* Return whether GNAT_NODE is a defining identifier for a renaming that comes
+ from the parameter association for the instantiation of a generic. We do
+ not want to emit source location for them: the code generated for their
+ initialization is likely to disturb debugging. */
+
+bool
+renaming_from_instantiation_p (Node_Id gnat_node)
+{
+ if (Nkind (gnat_node) != N_Defining_Identifier
+ || !Is_Object (gnat_node)
+ || Comes_From_Source (gnat_node)
+ || !Present (Renamed_Object (gnat_node)))
+ return false;
+
+ /* Get the object declaration of the renamed object, if any and if the
+ renamed object is a mere identifier. */
+ gnat_node = Renamed_Object (gnat_node);
+ if (Nkind (gnat_node) != N_Identifier)
+ return false;
+
+ gnat_node = Parent (Entity (gnat_node));
+ return (Present (gnat_node)
+ && Nkind (gnat_node) == N_Object_Declaration
+ && Present (Corresponding_Generic_Association (gnat_node)));
+}
+
+/* Similar to set_expr_location, but start with the Sloc of GNAT_NODE and
+ don't do anything if it doesn't correspond to a source location. And,
+ if CLEAR_COLUMN is true, set the column information to 0. */
+
+static void
+set_expr_location_from_node (tree node, Node_Id gnat_node, bool clear_column)
+{
+ location_t locus;
+
+ /* Do not set a location for constructs likely to disturb debugging. */
+ if (Nkind (gnat_node) == N_Defining_Identifier)
+ {
+ if (Is_Type (gnat_node) && Is_Actual_Subtype (gnat_node))
+ return;
+
+ if (renaming_from_instantiation_p (gnat_node))
+ return;
+ }
+
+ if (!Sloc_to_locus (Sloc (gnat_node), &locus, clear_column))
+ return;
+
+ SET_EXPR_LOCATION (node, locus);
+}
+
+/* More elaborate version of set_expr_location_from_node to be used in more
+ general contexts, for example the result of the translation of a generic
+ GNAT node. */
+
+static void
+set_gnu_expr_location_from_node (tree node, Node_Id gnat_node)
+{
+ /* Set the location information on the node if it is a real expression.
+ References can be reused for multiple GNAT nodes and they would get
+ the location information of their last use. Also make sure not to
+ overwrite an existing location as it is probably more precise. */
+
+ switch (TREE_CODE (node))
+ {
+ CASE_CONVERT:
+ case NON_LVALUE_EXPR:
+ case SAVE_EXPR:
+ break;
+
+ case COMPOUND_EXPR:
+ if (EXPR_P (TREE_OPERAND (node, 1)))
+ set_gnu_expr_location_from_node (TREE_OPERAND (node, 1), gnat_node);
+
+ /* ... fall through ... */
+
+ default:
+ if (!REFERENCE_CLASS_P (node) && !EXPR_HAS_LOCATION (node))
+ {
+ set_expr_location_from_node (node, gnat_node);
+ set_end_locus_from_node (node, gnat_node);
+ }
+ break;
+ }
+}
+
+/* Set the end_locus information for GNU_NODE, if any, from an explicit end
+ location associated with GNAT_NODE or GNAT_NODE itself, whichever makes
+ most sense. Return true if a sensible assignment was performed. */
+
+static bool
+set_end_locus_from_node (tree gnu_node, Node_Id gnat_node)
+{
+ Node_Id gnat_end_label;
+ location_t end_locus;
+
+ /* Pick the GNAT node of which we'll take the sloc to assign to the GCC node
+ end_locus when there is one. We consider only GNAT nodes with a possible
+ End_Label attached. If the End_Label actually was unassigned, fallback
+ on the original node. We'd better assign an explicit sloc associated with
+ the outer construct in any case. */
+
+ switch (Nkind (gnat_node))
+ {
+ case N_Package_Body:
+ case N_Subprogram_Body:
+ case N_Block_Statement:
+ if (Present (Handled_Statement_Sequence (gnat_node)))
+ gnat_end_label = End_Label (Handled_Statement_Sequence (gnat_node));
+ else
+ gnat_end_label = Empty;
+
+ break;
+
+ case N_Package_Declaration:
+ gcc_checking_assert (Present (Specification (gnat_node)));
+ gnat_end_label = End_Label (Specification (gnat_node));
+ break;
+
+ default:
+ return false;
+ }
+
+ if (Present (gnat_end_label))
+ gnat_node = gnat_end_label;
+
+ /* Some expanded subprograms have neither an End_Label nor a Sloc
+ attached. Notify that to callers. For a block statement with no
+ End_Label, clear column information, so that the tree for a
+ transient block does not receive the sloc of a source condition. */
+ if (!Sloc_to_locus (Sloc (gnat_node), &end_locus,
+ No (gnat_end_label)
+ && (Nkind (gnat_node) == N_Block_Statement)))
+ return false;
+
+ switch (TREE_CODE (gnu_node))
+ {
+ case BIND_EXPR:
+ BLOCK_SOURCE_END_LOCATION (BIND_EXPR_BLOCK (gnu_node)) = end_locus;
+ return true;
+
+ case FUNCTION_DECL:
+ DECL_STRUCT_FUNCTION (gnu_node)->function_end_locus = end_locus;
+ return true;
+
+ default:
+ return false;
+ }
+}
+
+/* Post an error message. MSG is the error message, properly annotated.
+ NODE is the node at which to post the error and the node to use for the
+ '&' substitution. */
+
+void
+post_error (const char *msg, Node_Id node)
+{
+ String_Template temp;
+ String_Pointer sp;
+
+ if (No (node))
+ return;
+
+ temp.Low_Bound = 1;
+ temp.High_Bound = strlen (msg);
+ sp.Bounds = &temp;
+ sp.Array = msg;
+ Error_Msg_N (sp, node);
+}
+
+/* Similar to post_error, but NODE is the node at which to post the error and
+ ENT is the node to use for the '&' substitution. */
+
+void
+post_error_ne (const char *msg, Node_Id node, Entity_Id ent)
+{
+ String_Template temp;
+ String_Pointer sp;
+
+ if (No (node))
+ return;
+
+ temp.Low_Bound = 1;
+ temp.High_Bound = strlen (msg);
+ sp.Bounds = &temp;
+ sp.Array = msg;
+ Error_Msg_NE (sp, node, ent);
+}
+
+/* Similar to post_error_ne, but NUM is the number to use for the '^'. */
+
+void
+post_error_ne_num (const char *msg, Node_Id node, Entity_Id ent, int num)
+{
+ Error_Msg_Uint_1 = UI_From_Int (num);
+ post_error_ne (msg, node, ent);
+}
+
+/* Similar to post_error_ne, but T is a GCC tree representing the number to
+ write. If T represents a constant, the text inside curly brackets in
+ MSG will be output (presumably including a '^'). Otherwise it will not
+ be output and the text inside square brackets will be output instead. */
+
+void
+post_error_ne_tree (const char *msg, Node_Id node, Entity_Id ent, tree t)
+{
+ char *new_msg = XALLOCAVEC (char, strlen (msg) + 1);
+ char start_yes, end_yes, start_no, end_no;
+ const char *p;
+ char *q;
+
+ if (TREE_CODE (t) == INTEGER_CST)
+ {
+ Error_Msg_Uint_1 = UI_From_gnu (t);
+ start_yes = '{', end_yes = '}', start_no = '[', end_no = ']';
+ }
+ else
+ start_yes = '[', end_yes = ']', start_no = '{', end_no = '}';
+
+ for (p = msg, q = new_msg; *p; p++)
+ {
+ if (*p == start_yes)
+ for (p++; *p != end_yes; p++)
+ *q++ = *p;
+ else if (*p == start_no)
+ for (p++; *p != end_no; p++)
+ ;
+ else
+ *q++ = *p;
+ }
+
+ *q = 0;
+
+ post_error_ne (new_msg, node, ent);
+}
+
+/* Similar to post_error_ne_tree, but NUM is a second integer to write. */
+
+void
+post_error_ne_tree_2 (const char *msg, Node_Id node, Entity_Id ent, tree t,
+ int num)
+{
+ Error_Msg_Uint_2 = UI_From_Int (num);
+ post_error_ne_tree (msg, node, ent, t);
+}
+
+/* Return a label to branch to for the exception type in KIND or Empty
+ if none. */
+
+Entity_Id
+get_exception_label (char kind)
+{
+ switch (kind)
+ {
+ case N_Raise_Constraint_Error:
+ return gnu_constraint_error_label_stack.last ();
+
+ case N_Raise_Storage_Error:
+ return gnu_storage_error_label_stack.last ();
+
+ case N_Raise_Program_Error:
+ return gnu_program_error_label_stack.last ();
+
+ default:
+ return Empty;
+ }
+
+ gcc_unreachable ();
+}
+
+/* Return the decl for the current elaboration procedure. */
+
+static tree
+get_elaboration_procedure (void)
+{
+ return gnu_elab_proc_stack->last ();
+}
+
+/* Return the controlling type of a dispatching subprogram. */
+
+static Entity_Id
+get_controlling_type (Entity_Id subprog)
+{
+ /* This is modeled on Expand_Interface_Thunk. */
+ Entity_Id controlling_type = Etype (First_Formal (subprog));
+ if (Is_Access_Type (controlling_type))
+ controlling_type = Directly_Designated_Type (controlling_type);
+ controlling_type = Underlying_Type (controlling_type);
+ if (Is_Concurrent_Type (controlling_type))
+ controlling_type = Corresponding_Record_Type (controlling_type);
+ controlling_type = Base_Type (controlling_type);
+ return controlling_type;
+}
+
+/* Return whether we should use an alias for the TARGET of a thunk
+ in order to make the call generated in the thunk local. */
+
+static bool
+use_alias_for_thunk_p (tree target)
+{
+ /* We cannot generate a local call in this case. */
+ if (DECL_EXTERNAL (target))
+ return false;
+
+ /* The call is already local in this case. */
+ if (TREE_CODE (DECL_CONTEXT (target)) == FUNCTION_DECL)
+ return false;
+
+ return TARGET_USE_LOCAL_THUNK_ALIAS_P (target);
+}
+
+static GTY(()) unsigned long thunk_labelno = 0;
+
+/* Create an alias for TARGET to be used as the target of a thunk. */
+
+static tree
+make_alias_for_thunk (tree target)
+{
+ char buf[64];
+ targetm.asm_out.generate_internal_label (buf, "LTHUNK", thunk_labelno++);
+
+ tree alias = build_decl (DECL_SOURCE_LOCATION (target), TREE_CODE (target),
+ get_identifier (buf), TREE_TYPE (target));
+
+ DECL_LANG_SPECIFIC (alias) = DECL_LANG_SPECIFIC (target);
+ DECL_CONTEXT (alias) = DECL_CONTEXT (target);
+ TREE_READONLY (alias) = TREE_READONLY (target);
+ TREE_THIS_VOLATILE (alias) = TREE_THIS_VOLATILE (target);
+ DECL_ARTIFICIAL (alias) = 1;
+ DECL_INITIAL (alias) = error_mark_node;
+ DECL_ARGUMENTS (alias) = copy_list (DECL_ARGUMENTS (target));
+ TREE_ADDRESSABLE (alias) = 1;
+ SET_DECL_ASSEMBLER_NAME (alias, DECL_NAME (alias));
+
+ cgraph_node *n = cgraph_node::create_same_body_alias (alias, target);
+ gcc_assert (n);
+
+ return alias;
+}
+
+/* Create the local covariant part of {GNAT,GNU}_THUNK. */
+
+static tree
+make_covariant_thunk (Entity_Id gnat_thunk, tree gnu_thunk)
+{
+ tree gnu_name = create_concat_name (gnat_thunk, "CV");
+ tree gnu_cv_thunk
+ = build_decl (DECL_SOURCE_LOCATION (gnu_thunk), TREE_CODE (gnu_thunk),
+ gnu_name, TREE_TYPE (gnu_thunk));
+
+ DECL_ARGUMENTS (gnu_cv_thunk) = copy_list (DECL_ARGUMENTS (gnu_thunk));
+ for (tree param_decl = DECL_ARGUMENTS (gnu_cv_thunk);
+ param_decl;
+ param_decl = DECL_CHAIN (param_decl))
+ DECL_CONTEXT (param_decl) = gnu_cv_thunk;
+
+ DECL_RESULT (gnu_cv_thunk) = copy_node (DECL_RESULT (gnu_thunk));
+ DECL_CONTEXT (DECL_RESULT (gnu_cv_thunk)) = gnu_cv_thunk;
+
+ DECL_LANG_SPECIFIC (gnu_cv_thunk) = DECL_LANG_SPECIFIC (gnu_thunk);
+ DECL_CONTEXT (gnu_cv_thunk) = DECL_CONTEXT (gnu_thunk);
+ TREE_READONLY (gnu_cv_thunk) = TREE_READONLY (gnu_thunk);
+ TREE_THIS_VOLATILE (gnu_cv_thunk) = TREE_THIS_VOLATILE (gnu_thunk);
+ DECL_ARTIFICIAL (gnu_cv_thunk) = 1;
+
+ return gnu_cv_thunk;
+}
+
+/* Try to create a GNU thunk for {GNAT,GNU}_THUNK and return true on success.
+
+ GNU thunks are more efficient than GNAT thunks because they don't call into
+ the runtime to retrieve the offset used in the displacement operation, but
+ they are tailored to C++ and thus too limited to support the full range of
+ thunks generated in Ada. Here's the complete list of limitations:
+
+ 1. Multi-controlling thunks, i.e thunks with more than one controlling
+ parameter, are simply not supported.
+
+ 2. Covariant thunks, i.e. thunks for which the result is also controlling,
+ are split into a pair of (this, covariant-only) thunks.
+
+ 3. Variable-offset thunks, i.e. thunks for which the offset depends on the
+ object and not only on its type, are supported as 2nd class citizens.
+
+ 4. External thunks, i.e. thunks for which the target is not declared in
+ the same unit as the thunk, are supported as 2nd class citizens.
+
+ 5. Local thunks, i.e. thunks generated for a local type, are supported as
+ 2nd class citizens. */
+
+static bool
+maybe_make_gnu_thunk (Entity_Id gnat_thunk, tree gnu_thunk)
+{
+ const Entity_Id gnat_target = Thunk_Entity (gnat_thunk);
+
+ /* Check that the first formal of the target is the only controlling one. */
+ Entity_Id gnat_formal = First_Formal (gnat_target);
+ if (!Is_Controlling_Formal (gnat_formal))
+ return false;
+ for (gnat_formal = Next_Formal (gnat_formal);
+ Present (gnat_formal);
+ gnat_formal = Next_Formal (gnat_formal))
+ if (Is_Controlling_Formal (gnat_formal))
+ return false;
+
+ /* Look for the types that control the target and the thunk. */
+ const Entity_Id gnat_controlling_type = get_controlling_type (gnat_target);
+ const Entity_Id gnat_interface_type = get_controlling_type (gnat_thunk);
+
+ /* We must have an interface type at this point. */
+ gcc_assert (Is_Interface (gnat_interface_type));
+
+ /* Now compute whether the former covers the latter. */
+ const Entity_Id gnat_interface_tag
+ = Find_Interface_Tag (gnat_controlling_type, gnat_interface_type);
+ tree gnu_interface_tag
+ = Present (gnat_interface_tag)
+ ? gnat_to_gnu_field_decl (gnat_interface_tag)
+ : NULL_TREE;
+ tree gnu_interface_offset
+ = gnu_interface_tag ? byte_position (gnu_interface_tag) : NULL_TREE;
+
+ /* There are three ways to retrieve the offset between the interface view
+ and the base object. Either the controlling type covers the interface
+ type and the offset of the corresponding tag is fixed, in which case it
+ can be statically encoded in the thunk (see FIXED_OFFSET below). Or the
+ controlling type doesn't cover the interface type but is of fixed size,
+ in which case the offset is stored in the dispatch table, two pointers
+ above the dispatch table address (see VIRTUAL_VALUE below). Otherwise,
+ the offset is variable and is stored right after the tag in every object
+ (see INDIRECT_OFFSET below). See also a-tags.ads for more details. */
+ HOST_WIDE_INT fixed_offset, virtual_value, indirect_offset;
+ tree virtual_offset;
+
+ if (gnu_interface_offset && TREE_CODE (gnu_interface_offset) == INTEGER_CST)
+ {
+ fixed_offset = - tree_to_shwi (gnu_interface_offset);
+ virtual_value = 0;
+ virtual_offset = NULL_TREE;
+ indirect_offset = 0;
+ }
+ else if (!gnu_interface_offset
+ && !Is_Variable_Size_Record (gnat_controlling_type))
+ {
+ fixed_offset = 0;
+ virtual_value = - 2 * (HOST_WIDE_INT) (POINTER_SIZE / BITS_PER_UNIT);
+ virtual_offset = build_int_cst (integer_type_node, virtual_value);
+ indirect_offset = 0;
+ }
+ else
+ {
+ /* Covariant thunks with variable offset are not supported. */
+ if (Has_Controlling_Result (gnat_target))
+ return false;
+
+ fixed_offset = 0;
+ virtual_value = 0;
+ virtual_offset = NULL_TREE;
+ indirect_offset = (HOST_WIDE_INT) (POINTER_SIZE / BITS_PER_UNIT);
+ }
+
+ tree gnu_target = gnat_to_gnu_entity (gnat_target, NULL_TREE, false);
+
+ /* If the target is local, then thunk and target must have the same context
+ because cgraph_node::expand_thunk can only forward the static chain. */
+ if (DECL_STATIC_CHAIN (gnu_target)
+ && DECL_CONTEXT (gnu_thunk) != DECL_CONTEXT (gnu_target))
+ return false;
+
+ /* If the target returns by invisible reference and is external, apply the
+ same transformation as Subprogram_Body_to_gnu here. */
+ if (TREE_ADDRESSABLE (TREE_TYPE (gnu_target))
+ && DECL_EXTERNAL (gnu_target)
+ && !POINTER_TYPE_P (TREE_TYPE (DECL_RESULT (gnu_target))))
+ {
+ TREE_TYPE (DECL_RESULT (gnu_target))
+ = build_reference_type (TREE_TYPE (DECL_RESULT (gnu_target)));
+ relayout_decl (DECL_RESULT (gnu_target));
+ }
+
+ /* The thunk expander requires the return types of thunk and target to be
+ compatible, which is not fully the case with the CICO mechanism. */
+ if (TYPE_CI_CO_LIST (TREE_TYPE (gnu_thunk)))
+ {
+ tree gnu_target_type = TREE_TYPE (gnu_target);
+ gcc_assert (TYPE_CI_CO_LIST (gnu_target_type));
+ TYPE_CANONICAL (TREE_TYPE (TREE_TYPE (gnu_thunk)))
+ = TYPE_CANONICAL (TREE_TYPE (gnu_target_type));
+ }
+
+ cgraph_node *target_node = cgraph_node::get_create (gnu_target);
+
+ /* We may also need to create an alias for the target in order to make
+ the call local, depending on the linkage of the target. */
+ tree gnu_alias = use_alias_for_thunk_p (gnu_target)
+ ? make_alias_for_thunk (gnu_target)
+ : gnu_target;
+
+ /* If the return type of the target is a controlling type, then we need
+ both an usual this thunk and a covariant thunk in this order:
+
+ this thunk --> covariant thunk --> target
+
+ For covariant thunks, we can only handle a fixed offset. */
+ if (Has_Controlling_Result (gnat_target))
+ {
+ gcc_assert (fixed_offset < 0);
+ tree gnu_cv_thunk = make_covariant_thunk (gnat_thunk, gnu_thunk);
+ target_node->create_thunk (gnu_cv_thunk, gnu_target, false,
+ - fixed_offset, 0, 0,
+ NULL_TREE, gnu_alias);
+
+ gnu_alias = gnu_target = gnu_cv_thunk;
+ }
+
+ target_node->create_thunk (gnu_thunk, gnu_target, true,
+ fixed_offset, virtual_value, indirect_offset,
+ virtual_offset, gnu_alias);
+
+ return true;
+}
+
+/* Initialize the table that maps GNAT codes to GCC codes for simple
+ binary and unary operations. */
+
+static void
+init_code_table (void)
+{
+ gnu_codes[N_Op_And] = TRUTH_AND_EXPR;
+ gnu_codes[N_Op_Or] = TRUTH_OR_EXPR;
+ gnu_codes[N_Op_Xor] = TRUTH_XOR_EXPR;
+ gnu_codes[N_Op_Eq] = EQ_EXPR;
+ gnu_codes[N_Op_Ne] = NE_EXPR;
+ gnu_codes[N_Op_Lt] = LT_EXPR;
+ gnu_codes[N_Op_Le] = LE_EXPR;
+ gnu_codes[N_Op_Gt] = GT_EXPR;
+ gnu_codes[N_Op_Ge] = GE_EXPR;
+ gnu_codes[N_Op_Add] = PLUS_EXPR;
+ gnu_codes[N_Op_Subtract] = MINUS_EXPR;
+ gnu_codes[N_Op_Multiply] = MULT_EXPR;
+ gnu_codes[N_Op_Mod] = FLOOR_MOD_EXPR;
+ gnu_codes[N_Op_Rem] = TRUNC_MOD_EXPR;
+ gnu_codes[N_Op_Minus] = NEGATE_EXPR;
+ gnu_codes[N_Op_Abs] = ABS_EXPR;
+ gnu_codes[N_Op_Not] = TRUTH_NOT_EXPR;
+ gnu_codes[N_Op_Rotate_Left] = LROTATE_EXPR;
+ gnu_codes[N_Op_Rotate_Right] = RROTATE_EXPR;
+ gnu_codes[N_Op_Shift_Left] = LSHIFT_EXPR;
+ gnu_codes[N_Op_Shift_Right] = RSHIFT_EXPR;
+ gnu_codes[N_Op_Shift_Right_Arithmetic] = RSHIFT_EXPR;
+ gnu_codes[N_And_Then] = TRUTH_ANDIF_EXPR;
+ gnu_codes[N_Or_Else] = TRUTH_ORIF_EXPR;
+}
+
+#include "gt-ada-trans.h"