/* Read the GIMPLE representation from a file stream. Copyright (C) 2009-2019 Free Software Foundation, Inc. Contributed by Kenneth Zadeck Re-implemented by Diego Novillo This file is part of GCC. GCC is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 3, or (at your option) any later version. GCC is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with GCC; see the file COPYING3. If not see . */ #include "config.h" #include "system.h" #include "coretypes.h" #include "backend.h" #include "target.h" #include "rtl.h" #include "tree.h" #include "gimple.h" #include "cfghooks.h" #include "tree-pass.h" #include "ssa.h" #include "gimple-streamer.h" #include "toplev.h" #include "gimple-iterator.h" #include "tree-cfg.h" #include "tree-into-ssa.h" #include "tree-dfa.h" #include "tree-ssa.h" #include "except.h" #include "cgraph.h" #include "cfgloop.h" #include "debug.h" #include "alloc-pool.h" /* Allocator used to hold string slot entries for line map streaming. */ static struct object_allocator *string_slot_allocator; /* The table to hold the file names. */ static hash_table *file_name_hash_table; /* This obstack holds file names used in locators. Line map datastructures points here and thus it needs to be kept allocated as long as linemaps exists. */ static struct obstack file_name_obstack; /* Check that tag ACTUAL has one of the given values. NUM_TAGS is the number of valid tag values to check. */ void lto_tag_check_set (enum LTO_tags actual, int ntags, ...) { va_list ap; int i; va_start (ap, ntags); for (i = 0; i < ntags; i++) if ((unsigned) actual == va_arg (ap, unsigned)) { va_end (ap); return; } va_end (ap); internal_error ("bytecode stream: unexpected tag %s", lto_tag_name (actual)); } /* Read LENGTH bytes from STREAM to ADDR. */ void lto_input_data_block (class lto_input_block *ib, void *addr, size_t length) { size_t i; unsigned char *const buffer = (unsigned char *) addr; for (i = 0; i < length; i++) buffer[i] = streamer_read_uchar (ib); } /* Lookup STRING in file_name_hash_table. If found, return the existing string, otherwise insert STRING as the canonical version. */ static const char * canon_file_name (const char *string) { string_slot **slot; struct string_slot s_slot; size_t len = strlen (string); s_slot.s = string; s_slot.len = len; slot = file_name_hash_table->find_slot (&s_slot, INSERT); if (*slot == NULL) { char *saved_string; struct string_slot *new_slot; saved_string = XOBNEWVEC (&file_name_obstack, char, len + 1); new_slot = string_slot_allocator->allocate (); memcpy (saved_string, string, len + 1); new_slot->s = saved_string; new_slot->len = len; *slot = new_slot; return saved_string; } else { struct string_slot *old_slot = *slot; return old_slot->s; } } /* Pointer to currently alive instance of lto_location_cache. */ lto_location_cache *lto_location_cache::current_cache; /* Sort locations in source order. Start with file from last application. */ int lto_location_cache::cmp_loc (const void *pa, const void *pb) { const cached_location *a = ((const cached_location *)pa); const cached_location *b = ((const cached_location *)pb); const char *current_file = current_cache->current_file; int current_line = current_cache->current_line; if (a->file == current_file && b->file != current_file) return -1; if (a->file != current_file && b->file == current_file) return 1; if (a->file == current_file && b->file == current_file) { if (a->line == current_line && b->line != current_line) return -1; if (a->line != current_line && b->line == current_line) return 1; } if (a->file != b->file) return strcmp (a->file, b->file); if (a->sysp != b->sysp) return a->sysp ? 1 : -1; if (a->line != b->line) return a->line - b->line; return a->col - b->col; } /* Apply all changes in location cache. Add locations into linemap and patch trees. */ bool lto_location_cache::apply_location_cache () { static const char *prev_file; if (!loc_cache.length ()) return false; if (loc_cache.length () > 1) loc_cache.qsort (cmp_loc); for (unsigned int i = 0; i < loc_cache.length (); i++) { struct cached_location loc = loc_cache[i]; if (current_file != loc.file) linemap_add (line_table, prev_file ? LC_RENAME : LC_ENTER, loc.sysp, loc.file, loc.line); else if (current_line != loc.line) { int max = loc.col; for (unsigned int j = i + 1; j < loc_cache.length (); j++) if (loc.file != loc_cache[j].file || loc.line != loc_cache[j].line) break; else if (max < loc_cache[j].col) max = loc_cache[j].col; linemap_line_start (line_table, loc.line, max + 1); } gcc_assert (*loc.loc == BUILTINS_LOCATION + 1); if (current_file == loc.file && current_line == loc.line && current_col == loc.col) *loc.loc = current_loc; else current_loc = *loc.loc = linemap_position_for_column (line_table, loc.col); current_line = loc.line; prev_file = current_file = loc.file; current_col = loc.col; } loc_cache.truncate (0); accepted_length = 0; return true; } /* Tree merging did not suceed; mark all changes in the cache as accepted. */ void lto_location_cache::accept_location_cache () { gcc_assert (current_cache == this); accepted_length = loc_cache.length (); } /* Tree merging did suceed; throw away recent changes. */ void lto_location_cache::revert_location_cache () { loc_cache.truncate (accepted_length); } /* Read a location bitpack from input block IB and either update *LOC directly or add it to the location cache. It is neccesary to call apply_location_cache to get *LOC updated. */ void lto_location_cache::input_location (location_t *loc, struct bitpack_d *bp, class data_in *data_in) { static const char *stream_file; static int stream_line; static int stream_col; static bool stream_sysp; bool file_change, line_change, column_change; gcc_assert (current_cache == this); *loc = bp_unpack_int_in_range (bp, "location", 0, RESERVED_LOCATION_COUNT); if (*loc < RESERVED_LOCATION_COUNT) return; /* Keep value RESERVED_LOCATION_COUNT in *loc as linemap lookups will ICE on it. */ file_change = bp_unpack_value (bp, 1); line_change = bp_unpack_value (bp, 1); column_change = bp_unpack_value (bp, 1); if (file_change) { stream_file = canon_file_name (bp_unpack_string (data_in, bp)); stream_sysp = bp_unpack_value (bp, 1); } if (line_change) stream_line = bp_unpack_var_len_unsigned (bp); if (column_change) stream_col = bp_unpack_var_len_unsigned (bp); /* This optimization saves location cache operations druing gimple streaming. */ if (current_file == stream_file && current_line == stream_line && current_col == stream_col && current_sysp == stream_sysp) { *loc = current_loc; return; } struct cached_location entry = {stream_file, loc, stream_line, stream_col, stream_sysp}; loc_cache.safe_push (entry); } /* Read a location bitpack from input block IB and either update *LOC directly or add it to the location cache. It is neccesary to call apply_location_cache to get *LOC updated. */ void lto_input_location (location_t *loc, struct bitpack_d *bp, class data_in *data_in) { data_in->location_cache.input_location (loc, bp, data_in); } /* Read location and return it instead of going through location caching. This should be used only when the resulting location is not going to be discarded. */ location_t stream_input_location_now (struct bitpack_d *bp, class data_in *data_in) { location_t loc; stream_input_location (&loc, bp, data_in); data_in->location_cache.apply_location_cache (); return loc; } /* Read a reference to a tree node from DATA_IN using input block IB. TAG is the expected node that should be found in IB, if TAG belongs to one of the indexable trees, expect to read a reference index to be looked up in one of the symbol tables, otherwise read the pysical representation of the tree using stream_read_tree. FN is the function scope for the read tree. */ tree lto_input_tree_ref (class lto_input_block *ib, class data_in *data_in, struct function *fn, enum LTO_tags tag) { unsigned HOST_WIDE_INT ix_u; tree result = NULL_TREE; lto_tag_check_range (tag, LTO_field_decl_ref, LTO_namelist_decl_ref); switch (tag) { case LTO_type_ref: ix_u = streamer_read_uhwi (ib); result = lto_file_decl_data_get_type (data_in->file_data, ix_u); break; case LTO_ssa_name_ref: ix_u = streamer_read_uhwi (ib); result = (*SSANAMES (fn))[ix_u]; break; case LTO_field_decl_ref: ix_u = streamer_read_uhwi (ib); result = lto_file_decl_data_get_field_decl (data_in->file_data, ix_u); break; case LTO_function_decl_ref: ix_u = streamer_read_uhwi (ib); result = lto_file_decl_data_get_fn_decl (data_in->file_data, ix_u); break; case LTO_type_decl_ref: ix_u = streamer_read_uhwi (ib); result = lto_file_decl_data_get_type_decl (data_in->file_data, ix_u); break; case LTO_namespace_decl_ref: ix_u = streamer_read_uhwi (ib); result = lto_file_decl_data_get_namespace_decl (data_in->file_data, ix_u); break; case LTO_global_decl_ref: case LTO_result_decl_ref: case LTO_const_decl_ref: case LTO_imported_decl_ref: case LTO_label_decl_ref: case LTO_translation_unit_decl_ref: case LTO_namelist_decl_ref: ix_u = streamer_read_uhwi (ib); result = lto_file_decl_data_get_var_decl (data_in->file_data, ix_u); break; default: gcc_unreachable (); } gcc_assert (result); return result; } /* Read and return a double-linked list of catch handlers from input block IB, using descriptors in DATA_IN. */ static struct eh_catch_d * lto_input_eh_catch_list (class lto_input_block *ib, class data_in *data_in, eh_catch *last_p) { eh_catch first; enum LTO_tags tag; *last_p = first = NULL; tag = streamer_read_record_start (ib); while (tag) { tree list; eh_catch n; lto_tag_check_range (tag, LTO_eh_catch, LTO_eh_catch); /* Read the catch node. */ n = ggc_cleared_alloc (); n->type_list = stream_read_tree (ib, data_in); n->filter_list = stream_read_tree (ib, data_in); n->label = stream_read_tree (ib, data_in); /* Register all the types in N->FILTER_LIST. */ for (list = n->filter_list; list; list = TREE_CHAIN (list)) add_type_for_runtime (TREE_VALUE (list)); /* Chain N to the end of the list. */ if (*last_p) (*last_p)->next_catch = n; n->prev_catch = *last_p; *last_p = n; /* Set the head of the list the first time through the loop. */ if (first == NULL) first = n; tag = streamer_read_record_start (ib); } return first; } /* Read and return EH region IX from input block IB, using descriptors in DATA_IN. */ static eh_region input_eh_region (class lto_input_block *ib, class data_in *data_in, int ix) { enum LTO_tags tag; eh_region r; /* Read the region header. */ tag = streamer_read_record_start (ib); if (tag == LTO_null) return NULL; r = ggc_cleared_alloc (); r->index = streamer_read_hwi (ib); gcc_assert (r->index == ix); /* Read all the region pointers as region numbers. We'll fix up the pointers once the whole array has been read. */ r->outer = (eh_region) (intptr_t) streamer_read_hwi (ib); r->inner = (eh_region) (intptr_t) streamer_read_hwi (ib); r->next_peer = (eh_region) (intptr_t) streamer_read_hwi (ib); switch (tag) { case LTO_ert_cleanup: r->type = ERT_CLEANUP; break; case LTO_ert_try: { struct eh_catch_d *last_catch; r->type = ERT_TRY; r->u.eh_try.first_catch = lto_input_eh_catch_list (ib, data_in, &last_catch); r->u.eh_try.last_catch = last_catch; break; } case LTO_ert_allowed_exceptions: { tree l; r->type = ERT_ALLOWED_EXCEPTIONS; r->u.allowed.type_list = stream_read_tree (ib, data_in); r->u.allowed.label = stream_read_tree (ib, data_in); r->u.allowed.filter = streamer_read_uhwi (ib); for (l = r->u.allowed.type_list; l ; l = TREE_CHAIN (l)) add_type_for_runtime (TREE_VALUE (l)); } break; case LTO_ert_must_not_throw: { r->type = ERT_MUST_NOT_THROW; r->u.must_not_throw.failure_decl = stream_read_tree (ib, data_in); bitpack_d bp = streamer_read_bitpack (ib); r->u.must_not_throw.failure_loc = stream_input_location_now (&bp, data_in); } break; default: gcc_unreachable (); } r->landing_pads = (eh_landing_pad) (intptr_t) streamer_read_hwi (ib); return r; } /* Read and return EH landing pad IX from input block IB, using descriptors in DATA_IN. */ static eh_landing_pad input_eh_lp (class lto_input_block *ib, class data_in *data_in, int ix) { enum LTO_tags tag; eh_landing_pad lp; /* Read the landing pad header. */ tag = streamer_read_record_start (ib); if (tag == LTO_null) return NULL; lto_tag_check_range (tag, LTO_eh_landing_pad, LTO_eh_landing_pad); lp = ggc_cleared_alloc (); lp->index = streamer_read_hwi (ib); gcc_assert (lp->index == ix); lp->next_lp = (eh_landing_pad) (intptr_t) streamer_read_hwi (ib); lp->region = (eh_region) (intptr_t) streamer_read_hwi (ib); lp->post_landing_pad = stream_read_tree (ib, data_in); return lp; } /* After reading the EH regions, pointers to peer and children regions are region numbers. This converts all these region numbers into real pointers into the rematerialized regions for FN. ROOT_REGION is the region number for the root EH region in FN. */ static void fixup_eh_region_pointers (struct function *fn, HOST_WIDE_INT root_region) { unsigned i; vec *eh_array = fn->eh->region_array; vec *lp_array = fn->eh->lp_array; eh_region r; eh_landing_pad lp; gcc_assert (eh_array && lp_array); gcc_assert (root_region >= 0); fn->eh->region_tree = (*eh_array)[root_region]; #define FIXUP_EH_REGION(r) (r) = (*eh_array)[(HOST_WIDE_INT) (intptr_t) (r)] #define FIXUP_EH_LP(p) (p) = (*lp_array)[(HOST_WIDE_INT) (intptr_t) (p)] /* Convert all the index numbers stored in pointer fields into pointers to the corresponding slots in the EH region array. */ FOR_EACH_VEC_ELT (*eh_array, i, r) { /* The array may contain NULL regions. */ if (r == NULL) continue; gcc_assert (i == (unsigned) r->index); FIXUP_EH_REGION (r->outer); FIXUP_EH_REGION (r->inner); FIXUP_EH_REGION (r->next_peer); FIXUP_EH_LP (r->landing_pads); } /* Convert all the index numbers stored in pointer fields into pointers to the corresponding slots in the EH landing pad array. */ FOR_EACH_VEC_ELT (*lp_array, i, lp) { /* The array may contain NULL landing pads. */ if (lp == NULL) continue; gcc_assert (i == (unsigned) lp->index); FIXUP_EH_LP (lp->next_lp); FIXUP_EH_REGION (lp->region); } #undef FIXUP_EH_REGION #undef FIXUP_EH_LP } /* Initialize EH support. */ void lto_init_eh (void) { static bool eh_initialized_p = false; if (eh_initialized_p) return; /* Contrary to most other FEs, we only initialize EH support when at least one of the files in the set contains exception regions in it. Since this happens much later than the call to init_eh in lang_dependent_init, we have to set flag_exceptions and call init_eh again to initialize the EH tables. */ flag_exceptions = 1; init_eh (); eh_initialized_p = true; } /* Read the exception table for FN from IB using the data descriptors in DATA_IN. */ static void input_eh_regions (class lto_input_block *ib, class data_in *data_in, struct function *fn) { HOST_WIDE_INT i, root_region, len; enum LTO_tags tag; tag = streamer_read_record_start (ib); if (tag == LTO_null) return; lto_tag_check_range (tag, LTO_eh_table, LTO_eh_table); gcc_assert (fn->eh); root_region = streamer_read_hwi (ib); gcc_assert (root_region == (int) root_region); /* Read the EH region array. */ len = streamer_read_hwi (ib); gcc_assert (len == (int) len); if (len > 0) { vec_safe_grow_cleared (fn->eh->region_array, len); for (i = 0; i < len; i++) { eh_region r = input_eh_region (ib, data_in, i); (*fn->eh->region_array)[i] = r; } } /* Read the landing pads. */ len = streamer_read_hwi (ib); gcc_assert (len == (int) len); if (len > 0) { vec_safe_grow_cleared (fn->eh->lp_array, len); for (i = 0; i < len; i++) { eh_landing_pad lp = input_eh_lp (ib, data_in, i); (*fn->eh->lp_array)[i] = lp; } } /* Read the runtime type data. */ len = streamer_read_hwi (ib); gcc_assert (len == (int) len); if (len > 0) { vec_safe_grow_cleared (fn->eh->ttype_data, len); for (i = 0; i < len; i++) { tree ttype = stream_read_tree (ib, data_in); (*fn->eh->ttype_data)[i] = ttype; } } /* Read the table of action chains. */ len = streamer_read_hwi (ib); gcc_assert (len == (int) len); if (len > 0) { if (targetm.arm_eabi_unwinder) { vec_safe_grow_cleared (fn->eh->ehspec_data.arm_eabi, len); for (i = 0; i < len; i++) { tree t = stream_read_tree (ib, data_in); (*fn->eh->ehspec_data.arm_eabi)[i] = t; } } else { vec_safe_grow_cleared (fn->eh->ehspec_data.other, len); for (i = 0; i < len; i++) { uchar c = streamer_read_uchar (ib); (*fn->eh->ehspec_data.other)[i] = c; } } } /* Reconstruct the EH region tree by fixing up the peer/children pointers. */ fixup_eh_region_pointers (fn, root_region); tag = streamer_read_record_start (ib); lto_tag_check_range (tag, LTO_null, LTO_null); } /* Make a new basic block with index INDEX in function FN. */ static basic_block make_new_block (struct function *fn, unsigned int index) { basic_block bb = alloc_block (); bb->index = index; SET_BASIC_BLOCK_FOR_FN (fn, index, bb); n_basic_blocks_for_fn (fn)++; return bb; } /* Read the CFG for function FN from input block IB. */ static void input_cfg (class lto_input_block *ib, class data_in *data_in, struct function *fn) { unsigned int bb_count; basic_block p_bb; unsigned int i; int index; init_empty_tree_cfg_for_function (fn); init_ssa_operands (fn); profile_status_for_fn (fn) = streamer_read_enum (ib, profile_status_d, PROFILE_LAST); bb_count = streamer_read_uhwi (ib); last_basic_block_for_fn (fn) = bb_count; if (bb_count > basic_block_info_for_fn (fn)->length ()) vec_safe_grow_cleared (basic_block_info_for_fn (fn), bb_count); if (bb_count > label_to_block_map_for_fn (fn)->length ()) vec_safe_grow_cleared (label_to_block_map_for_fn (fn), bb_count); index = streamer_read_hwi (ib); while (index != -1) { basic_block bb = BASIC_BLOCK_FOR_FN (fn, index); unsigned int edge_count; if (bb == NULL) bb = make_new_block (fn, index); edge_count = streamer_read_uhwi (ib); /* Connect up the CFG. */ for (i = 0; i < edge_count; i++) { unsigned int dest_index; unsigned int edge_flags; basic_block dest; profile_probability probability; edge e; dest_index = streamer_read_uhwi (ib); probability = profile_probability::stream_in (ib); edge_flags = streamer_read_uhwi (ib); dest = BASIC_BLOCK_FOR_FN (fn, dest_index); if (dest == NULL) dest = make_new_block (fn, dest_index); e = make_edge (bb, dest, edge_flags); e->probability = probability; } index = streamer_read_hwi (ib); } p_bb = ENTRY_BLOCK_PTR_FOR_FN (fn); index = streamer_read_hwi (ib); while (index != -1) { basic_block bb = BASIC_BLOCK_FOR_FN (fn, index); bb->prev_bb = p_bb; p_bb->next_bb = bb; p_bb = bb; index = streamer_read_hwi (ib); } /* ??? The cfgloop interface is tied to cfun. */ gcc_assert (cfun == fn); /* Input the loop tree. */ unsigned n_loops = streamer_read_uhwi (ib); if (n_loops == 0) return; struct loops *loops = ggc_cleared_alloc (); init_loops_structure (fn, loops, n_loops); set_loops_for_fn (fn, loops); /* Input each loop and associate it with its loop header so flow_loops_find can rebuild the loop tree. */ for (unsigned i = 1; i < n_loops; ++i) { int header_index = streamer_read_hwi (ib); if (header_index == -1) { loops->larray->quick_push (NULL); continue; } class loop *loop = alloc_loop (); loop->header = BASIC_BLOCK_FOR_FN (fn, header_index); loop->header->loop_father = loop; /* Read everything copy_loop_info copies. */ loop->estimate_state = streamer_read_enum (ib, loop_estimation, EST_LAST); loop->any_upper_bound = streamer_read_hwi (ib); if (loop->any_upper_bound) loop->nb_iterations_upper_bound = streamer_read_widest_int (ib); loop->any_likely_upper_bound = streamer_read_hwi (ib); if (loop->any_likely_upper_bound) loop->nb_iterations_likely_upper_bound = streamer_read_widest_int (ib); loop->any_estimate = streamer_read_hwi (ib); if (loop->any_estimate) loop->nb_iterations_estimate = streamer_read_widest_int (ib); /* Read OMP SIMD related info. */ loop->safelen = streamer_read_hwi (ib); loop->unroll = streamer_read_hwi (ib); loop->owned_clique = streamer_read_hwi (ib); loop->dont_vectorize = streamer_read_hwi (ib); loop->force_vectorize = streamer_read_hwi (ib); loop->simduid = stream_read_tree (ib, data_in); place_new_loop (fn, loop); /* flow_loops_find doesn't like loops not in the tree, hook them all as siblings of the tree root temporarily. */ flow_loop_tree_node_add (loops->tree_root, loop); } /* Rebuild the loop tree. */ flow_loops_find (loops); } /* Read the SSA names array for function FN from DATA_IN using input block IB. */ static void input_ssa_names (class lto_input_block *ib, class data_in *data_in, struct function *fn) { unsigned int i, size; size = streamer_read_uhwi (ib); init_ssanames (fn, size); i = streamer_read_uhwi (ib); while (i) { tree ssa_name, name; bool is_default_def; /* Skip over the elements that had been freed. */ while (SSANAMES (fn)->length () < i) SSANAMES (fn)->quick_push (NULL_TREE); is_default_def = (streamer_read_uchar (ib) != 0); name = stream_read_tree (ib, data_in); ssa_name = make_ssa_name_fn (fn, name, NULL); if (is_default_def) { set_ssa_default_def (cfun, SSA_NAME_VAR (ssa_name), ssa_name); SSA_NAME_DEF_STMT (ssa_name) = gimple_build_nop (); } i = streamer_read_uhwi (ib); } } /* Go through all NODE edges and fixup call_stmt pointers so they point to STMTS. */ static void fixup_call_stmt_edges_1 (struct cgraph_node *node, gimple **stmts, struct function *fn) { #define STMT_UID_NOT_IN_RANGE(uid) \ (gimple_stmt_max_uid (fn) < uid || uid == 0) struct cgraph_edge *cedge; struct ipa_ref *ref = NULL; unsigned int i; for (cedge = node->callees; cedge; cedge = cedge->next_callee) { if (STMT_UID_NOT_IN_RANGE (cedge->lto_stmt_uid)) fatal_error (input_location, "Cgraph edge statement index out of range"); cedge->call_stmt = as_a (stmts[cedge->lto_stmt_uid - 1]); if (!cedge->call_stmt) fatal_error (input_location, "Cgraph edge statement index not found"); } for (cedge = node->indirect_calls; cedge; cedge = cedge->next_callee) { if (STMT_UID_NOT_IN_RANGE (cedge->lto_stmt_uid)) fatal_error (input_location, "Cgraph edge statement index out of range"); cedge->call_stmt = as_a (stmts[cedge->lto_stmt_uid - 1]); if (!cedge->call_stmt) fatal_error (input_location, "Cgraph edge statement index not found"); } for (i = 0; node->iterate_reference (i, ref); i++) if (ref->lto_stmt_uid) { if (STMT_UID_NOT_IN_RANGE (ref->lto_stmt_uid)) fatal_error (input_location, "Reference statement index out of range"); ref->stmt = stmts[ref->lto_stmt_uid - 1]; if (!ref->stmt) fatal_error (input_location, "Reference statement index not found"); } } /* Fixup call_stmt pointers in NODE and all clones. */ static void fixup_call_stmt_edges (struct cgraph_node *orig, gimple **stmts) { struct cgraph_node *node; struct function *fn; while (orig->clone_of) orig = orig->clone_of; fn = DECL_STRUCT_FUNCTION (orig->decl); if (!orig->thunk.thunk_p) fixup_call_stmt_edges_1 (orig, stmts, fn); if (orig->clones) for (node = orig->clones; node != orig;) { if (!node->thunk.thunk_p) fixup_call_stmt_edges_1 (node, stmts, fn); if (node->clones) node = node->clones; else if (node->next_sibling_clone) node = node->next_sibling_clone; else { while (node != orig && !node->next_sibling_clone) node = node->clone_of; if (node != orig) node = node->next_sibling_clone; } } } /* Input the base body of struct function FN from DATA_IN using input block IB. */ static void input_struct_function_base (struct function *fn, class data_in *data_in, class lto_input_block *ib) { struct bitpack_d bp; int len; /* Read the static chain and non-local goto save area. */ fn->static_chain_decl = stream_read_tree (ib, data_in); fn->nonlocal_goto_save_area = stream_read_tree (ib, data_in); /* Read all the local symbols. */ len = streamer_read_hwi (ib); if (len > 0) { int i; vec_safe_grow_cleared (fn->local_decls, len); for (i = 0; i < len; i++) { tree t = stream_read_tree (ib, data_in); (*fn->local_decls)[i] = t; } } /* Input the current IL state of the function. */ fn->curr_properties = streamer_read_uhwi (ib); /* Read all the attributes for FN. */ bp = streamer_read_bitpack (ib); fn->is_thunk = bp_unpack_value (&bp, 1); fn->has_local_explicit_reg_vars = bp_unpack_value (&bp, 1); fn->returns_pcc_struct = bp_unpack_value (&bp, 1); fn->returns_struct = bp_unpack_value (&bp, 1); fn->can_throw_non_call_exceptions = bp_unpack_value (&bp, 1); fn->can_delete_dead_exceptions = bp_unpack_value (&bp, 1); fn->always_inline_functions_inlined = bp_unpack_value (&bp, 1); fn->after_inlining = bp_unpack_value (&bp, 1); fn->stdarg = bp_unpack_value (&bp, 1); fn->has_nonlocal_label = bp_unpack_value (&bp, 1); fn->has_forced_label_in_static = bp_unpack_value (&bp, 1); fn->calls_alloca = bp_unpack_value (&bp, 1); fn->calls_setjmp = bp_unpack_value (&bp, 1); fn->calls_eh_return = bp_unpack_value (&bp, 1); fn->has_force_vectorize_loops = bp_unpack_value (&bp, 1); fn->has_simduid_loops = bp_unpack_value (&bp, 1); fn->va_list_fpr_size = bp_unpack_value (&bp, 8); fn->va_list_gpr_size = bp_unpack_value (&bp, 8); fn->last_clique = bp_unpack_value (&bp, sizeof (short) * 8); /* Input the function start and end loci. */ fn->function_start_locus = stream_input_location_now (&bp, data_in); fn->function_end_locus = stream_input_location_now (&bp, data_in); /* Restore the instance discriminators if present. */ int instance_number = bp_unpack_value (&bp, 1); if (instance_number) { instance_number = bp_unpack_value (&bp, sizeof (int) * CHAR_BIT); maybe_create_decl_to_instance_map ()->put (fn->decl, instance_number); } } /* Read the body of function FN_DECL from DATA_IN using input block IB. */ static void input_function (tree fn_decl, class data_in *data_in, class lto_input_block *ib, class lto_input_block *ib_cfg) { struct function *fn; enum LTO_tags tag; gimple **stmts; basic_block bb; struct cgraph_node *node; tag = streamer_read_record_start (ib); lto_tag_check (tag, LTO_function); /* Read decls for parameters and args. */ DECL_RESULT (fn_decl) = stream_read_tree (ib, data_in); DECL_ARGUMENTS (fn_decl) = streamer_read_chain (ib, data_in); /* Read debug args if available. */ unsigned n_debugargs = streamer_read_uhwi (ib); if (n_debugargs) { vec **debugargs = decl_debug_args_insert (fn_decl); vec_safe_grow (*debugargs, n_debugargs); for (unsigned i = 0; i < n_debugargs; ++i) (**debugargs)[i] = stream_read_tree (ib, data_in); } /* Read the tree of lexical scopes for the function. */ DECL_INITIAL (fn_decl) = stream_read_tree (ib, data_in); unsigned block_leaf_count = streamer_read_uhwi (ib); while (block_leaf_count--) stream_read_tree (ib, data_in); if (!streamer_read_uhwi (ib)) return; push_struct_function (fn_decl); fn = DECL_STRUCT_FUNCTION (fn_decl); init_tree_ssa (fn); /* We input IL in SSA form. */ cfun->gimple_df->in_ssa_p = true; gimple_register_cfg_hooks (); node = cgraph_node::get (fn_decl); if (!node) node = cgraph_node::create (fn_decl); input_struct_function_base (fn, data_in, ib); input_cfg (ib_cfg, data_in, fn); /* Read all the SSA names. */ input_ssa_names (ib, data_in, fn); /* Read the exception handling regions in the function. */ input_eh_regions (ib, data_in, fn); gcc_assert (DECL_INITIAL (fn_decl)); DECL_SAVED_TREE (fn_decl) = NULL_TREE; /* Read all the basic blocks. */ tag = streamer_read_record_start (ib); while (tag) { input_bb (ib, tag, data_in, fn, node->count_materialization_scale); tag = streamer_read_record_start (ib); } /* Fix up the call statements that are mentioned in the callgraph edges. */ set_gimple_stmt_max_uid (cfun, 0); FOR_ALL_BB_FN (bb, cfun) { gimple_stmt_iterator gsi; for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); gsi_next (&gsi)) { gimple *stmt = gsi_stmt (gsi); gimple_set_uid (stmt, inc_gimple_stmt_max_uid (cfun)); } for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi)) { gimple *stmt = gsi_stmt (gsi); gimple_set_uid (stmt, inc_gimple_stmt_max_uid (cfun)); } } stmts = (gimple **) xcalloc (gimple_stmt_max_uid (fn), sizeof (gimple *)); FOR_ALL_BB_FN (bb, cfun) { gimple_stmt_iterator bsi = gsi_start_phis (bb); while (!gsi_end_p (bsi)) { gimple *stmt = gsi_stmt (bsi); gsi_next (&bsi); stmts[gimple_uid (stmt)] = stmt; } bsi = gsi_start_bb (bb); while (!gsi_end_p (bsi)) { gimple *stmt = gsi_stmt (bsi); bool remove = false; /* If we're recompiling LTO objects with debug stmts but we're not supposed to have debug stmts, remove them now. We can't remove them earlier because this would cause uid mismatches in fixups, but we can do it at this point, as long as debug stmts don't require fixups. Similarly remove all IFN_*SAN_* internal calls */ if (!flag_wpa) { if (is_gimple_debug (stmt) && (gimple_debug_nonbind_marker_p (stmt) ? !MAY_HAVE_DEBUG_MARKER_STMTS : !MAY_HAVE_DEBUG_BIND_STMTS)) remove = true; /* In case the linemap overflows locations can be dropped to zero. Thus do not keep nonsensical inline entry markers we'd later ICE on. */ tree block; if (gimple_debug_inline_entry_p (stmt) && (block = gimple_block (stmt)) && !inlined_function_outer_scope_p (block)) remove = true; if (is_gimple_call (stmt) && gimple_call_internal_p (stmt)) { bool replace = false; switch (gimple_call_internal_fn (stmt)) { case IFN_UBSAN_NULL: if ((flag_sanitize & (SANITIZE_NULL | SANITIZE_ALIGNMENT)) == 0) replace = true; break; case IFN_UBSAN_BOUNDS: if ((flag_sanitize & SANITIZE_BOUNDS) == 0) replace = true; break; case IFN_UBSAN_VPTR: if ((flag_sanitize & SANITIZE_VPTR) == 0) replace = true; break; case IFN_UBSAN_OBJECT_SIZE: if ((flag_sanitize & SANITIZE_OBJECT_SIZE) == 0) replace = true; break; case IFN_UBSAN_PTR: if ((flag_sanitize & SANITIZE_POINTER_OVERFLOW) == 0) replace = true; break; case IFN_ASAN_MARK: if ((flag_sanitize & SANITIZE_ADDRESS) == 0) replace = true; break; case IFN_TSAN_FUNC_EXIT: if ((flag_sanitize & SANITIZE_THREAD) == 0) replace = true; break; default: break; } if (replace) { gimple_call_set_internal_fn (as_a (stmt), IFN_NOP); update_stmt (stmt); } } } if (remove) { gimple_stmt_iterator gsi = bsi; gsi_next (&bsi); unlink_stmt_vdef (stmt); release_defs (stmt); gsi_remove (&gsi, true); } else { gsi_next (&bsi); stmts[gimple_uid (stmt)] = stmt; /* Remember that the input function has begin stmt markers, so that we know to expect them when emitting debug info. */ if (!cfun->debug_nonbind_markers && gimple_debug_nonbind_marker_p (stmt)) cfun->debug_nonbind_markers = true; } } } /* Set the gimple body to the statement sequence in the entry basic block. FIXME lto, this is fairly hacky. The existence of a gimple body is used by the cgraph routines, but we should really use the presence of the CFG. */ { edge_iterator ei = ei_start (ENTRY_BLOCK_PTR_FOR_FN (cfun)->succs); gimple_set_body (fn_decl, bb_seq (ei_edge (ei)->dest)); } update_max_bb_count (); fixup_call_stmt_edges (node, stmts); execute_all_ipa_stmt_fixups (node, stmts); update_ssa (TODO_update_ssa_only_virtuals); free_dominance_info (CDI_DOMINATORS); free_dominance_info (CDI_POST_DOMINATORS); free (stmts); pop_cfun (); } /* Read the body of function FN_DECL from DATA_IN using input block IB. */ static void input_constructor (tree var, class data_in *data_in, class lto_input_block *ib) { DECL_INITIAL (var) = stream_read_tree (ib, data_in); } /* Read the body from DATA for function NODE and fill it in. FILE_DATA are the global decls and types. SECTION_TYPE is either LTO_section_function_body or LTO_section_static_initializer. If section type is LTO_section_function_body, FN must be the decl for that function. */ static void lto_read_body_or_constructor (struct lto_file_decl_data *file_data, struct symtab_node *node, const char *data, enum lto_section_type section_type) { const struct lto_function_header *header; class data_in *data_in; int cfg_offset; int main_offset; int string_offset; tree fn_decl = node->decl; header = (const struct lto_function_header *) data; if (TREE_CODE (node->decl) == FUNCTION_DECL) { cfg_offset = sizeof (struct lto_function_header); main_offset = cfg_offset + header->cfg_size; string_offset = main_offset + header->main_size; } else { main_offset = sizeof (struct lto_function_header); string_offset = main_offset + header->main_size; } data_in = lto_data_in_create (file_data, data + string_offset, header->string_size, vNULL); if (section_type == LTO_section_function_body) { struct lto_in_decl_state *decl_state; unsigned from; gcc_checking_assert (node); /* Use the function's decl state. */ decl_state = lto_get_function_in_decl_state (file_data, fn_decl); gcc_assert (decl_state); file_data->current_decl_state = decl_state; /* Set up the struct function. */ from = data_in->reader_cache->nodes.length (); lto_input_block ib_main (data + main_offset, header->main_size, file_data->mode_table); if (TREE_CODE (node->decl) == FUNCTION_DECL) { lto_input_block ib_cfg (data + cfg_offset, header->cfg_size, file_data->mode_table); input_function (fn_decl, data_in, &ib_main, &ib_cfg); } else input_constructor (fn_decl, data_in, &ib_main); data_in->location_cache.apply_location_cache (); /* And fixup types we streamed locally. */ { struct streamer_tree_cache_d *cache = data_in->reader_cache; unsigned len = cache->nodes.length (); unsigned i; for (i = len; i-- > from;) { tree t = streamer_tree_cache_get_tree (cache, i); if (t == NULL_TREE) continue; if (TYPE_P (t)) { gcc_assert (TYPE_CANONICAL (t) == NULL_TREE); if (type_with_alias_set_p (t) && canonical_type_used_p (t)) TYPE_CANONICAL (t) = TYPE_MAIN_VARIANT (t); if (TYPE_MAIN_VARIANT (t) != t) { gcc_assert (TYPE_NEXT_VARIANT (t) == NULL_TREE); TYPE_NEXT_VARIANT (t) = TYPE_NEXT_VARIANT (TYPE_MAIN_VARIANT (t)); TYPE_NEXT_VARIANT (TYPE_MAIN_VARIANT (t)) = t; } } } } /* Restore decl state */ file_data->current_decl_state = file_data->global_decl_state; } lto_data_in_delete (data_in); } /* Read the body of NODE using DATA. FILE_DATA holds the global decls and types. */ void lto_input_function_body (struct lto_file_decl_data *file_data, struct cgraph_node *node, const char *data) { lto_read_body_or_constructor (file_data, node, data, LTO_section_function_body); } /* Read the body of NODE using DATA. FILE_DATA holds the global decls and types. */ void lto_input_variable_constructor (struct lto_file_decl_data *file_data, struct varpool_node *node, const char *data) { lto_read_body_or_constructor (file_data, node, data, LTO_section_function_body); } /* Queue of acummulated decl -> DIE mappings. Similar to locations those are only applied to prevailing tree nodes during tree merging. */ vec dref_queue; /* Read the physical representation of a tree node EXPR from input block IB using the per-file context in DATA_IN. */ static void lto_read_tree_1 (class lto_input_block *ib, class data_in *data_in, tree expr) { /* Read all the bitfield values in EXPR. Note that for LTO, we only write language-independent bitfields, so no more unpacking is needed. */ streamer_read_tree_bitfields (ib, data_in, expr); /* Read all the pointer fields in EXPR. */ streamer_read_tree_body (ib, data_in, expr); /* Read any LTO-specific data not read by the tree streamer. */ if (DECL_P (expr) && TREE_CODE (expr) != FUNCTION_DECL && TREE_CODE (expr) != TRANSLATION_UNIT_DECL) DECL_INITIAL (expr) = stream_read_tree (ib, data_in); /* Stream references to early generated DIEs. Keep in sync with the trees handled in dwarf2out_register_external_die. */ if ((DECL_P (expr) && TREE_CODE (expr) != FIELD_DECL && TREE_CODE (expr) != DEBUG_EXPR_DECL && TREE_CODE (expr) != TYPE_DECL) || TREE_CODE (expr) == BLOCK) { const char *str = streamer_read_string (data_in, ib); if (str) { unsigned HOST_WIDE_INT off = streamer_read_uhwi (ib); dref_entry e = { expr, str, off }; dref_queue.safe_push (e); } } } /* Read the physical representation of a tree node with tag TAG from input block IB using the per-file context in DATA_IN. */ static tree lto_read_tree (class lto_input_block *ib, class data_in *data_in, enum LTO_tags tag, hashval_t hash) { /* Instantiate a new tree node. */ tree result = streamer_alloc_tree (ib, data_in, tag); /* Enter RESULT in the reader cache. This will make RESULT available so that circular references in the rest of the tree structure can be resolved in subsequent calls to stream_read_tree. */ streamer_tree_cache_append (data_in->reader_cache, result, hash); lto_read_tree_1 (ib, data_in, result); /* end_marker = */ streamer_read_uchar (ib); return result; } /* Populate the reader cache with trees materialized from the SCC following in the IB, DATA_IN stream. */ hashval_t lto_input_scc (class lto_input_block *ib, class data_in *data_in, unsigned *len, unsigned *entry_len) { /* A blob of unnamed tree nodes, fill the cache from it and recurse. */ unsigned size = streamer_read_uhwi (ib); hashval_t scc_hash = streamer_read_uhwi (ib); unsigned scc_entry_len = 1; if (size == 1) { enum LTO_tags tag = streamer_read_record_start (ib); lto_input_tree_1 (ib, data_in, tag, scc_hash); } else { unsigned int first = data_in->reader_cache->nodes.length (); tree result; scc_entry_len = streamer_read_uhwi (ib); /* Materialize size trees by reading their headers. */ for (unsigned i = 0; i < size; ++i) { enum LTO_tags tag = streamer_read_record_start (ib); if (tag == LTO_null || (tag >= LTO_field_decl_ref && tag <= LTO_global_decl_ref) || tag == LTO_tree_pickle_reference || tag == LTO_integer_cst || tag == LTO_tree_scc) gcc_unreachable (); result = streamer_alloc_tree (ib, data_in, tag); streamer_tree_cache_append (data_in->reader_cache, result, 0); } /* Read the tree bitpacks and references. */ for (unsigned i = 0; i < size; ++i) { result = streamer_tree_cache_get_tree (data_in->reader_cache, first + i); lto_read_tree_1 (ib, data_in, result); /* end_marker = */ streamer_read_uchar (ib); } } *len = size; *entry_len = scc_entry_len; return scc_hash; } /* Read a tree from input block IB using the per-file context in DATA_IN. This context is used, for example, to resolve references to previously read nodes. */ tree lto_input_tree_1 (class lto_input_block *ib, class data_in *data_in, enum LTO_tags tag, hashval_t hash) { tree result; gcc_assert ((unsigned) tag < (unsigned) LTO_NUM_TAGS); if (tag == LTO_null) result = NULL_TREE; else if (tag >= LTO_field_decl_ref && tag <= LTO_namelist_decl_ref) { /* If TAG is a reference to an indexable tree, the next value in IB is the index into the table where we expect to find that tree. */ result = lto_input_tree_ref (ib, data_in, cfun, tag); } else if (tag == LTO_tree_pickle_reference) { /* If TAG is a reference to a previously read tree, look it up in the reader cache. */ result = streamer_get_pickled_tree (ib, data_in); } else if (tag == LTO_integer_cst) { /* For shared integer constants in singletons we can use the existing tree integer constant merging code. */ tree type = stream_read_tree (ib, data_in); unsigned HOST_WIDE_INT len = streamer_read_uhwi (ib); unsigned HOST_WIDE_INT i; HOST_WIDE_INT a[WIDE_INT_MAX_ELTS]; for (i = 0; i < len; i++) a[i] = streamer_read_hwi (ib); gcc_assert (TYPE_PRECISION (type) <= MAX_BITSIZE_MODE_ANY_INT); result = wide_int_to_tree (type, wide_int::from_array (a, len, TYPE_PRECISION (type))); streamer_tree_cache_append (data_in->reader_cache, result, hash); } else if (tag == LTO_tree_scc) gcc_unreachable (); else { /* Otherwise, materialize a new node from IB. */ result = lto_read_tree (ib, data_in, tag, hash); } return result; } tree lto_input_tree (class lto_input_block *ib, class data_in *data_in) { enum LTO_tags tag; /* Input and skip SCCs. */ while ((tag = streamer_read_record_start (ib)) == LTO_tree_scc) { unsigned len, entry_len; lto_input_scc (ib, data_in, &len, &entry_len); /* Register DECLs with the debuginfo machinery. */ while (!dref_queue.is_empty ()) { dref_entry e = dref_queue.pop (); debug_hooks->register_external_die (e.decl, e.sym, e.off); } } return lto_input_tree_1 (ib, data_in, tag, 0); } /* Input toplevel asms. */ void lto_input_toplevel_asms (struct lto_file_decl_data *file_data, int order_base) { size_t len; const char *data = lto_get_summary_section_data (file_data, LTO_section_asm, &len); const struct lto_simple_header_with_strings *header = (const struct lto_simple_header_with_strings *) data; int string_offset; class data_in *data_in; tree str; if (! data) return; string_offset = sizeof (*header) + header->main_size; lto_input_block ib (data + sizeof (*header), header->main_size, file_data->mode_table); data_in = lto_data_in_create (file_data, data + string_offset, header->string_size, vNULL); while ((str = streamer_read_string_cst (data_in, &ib))) { asm_node *node = symtab->finalize_toplevel_asm (str); node->order = streamer_read_hwi (&ib) + order_base; if (node->order >= symtab->order) symtab->order = node->order + 1; } lto_data_in_delete (data_in); lto_free_section_data (file_data, LTO_section_asm, NULL, data, len); } /* Input mode table. */ void lto_input_mode_table (struct lto_file_decl_data *file_data) { size_t len; const char *data = lto_get_summary_section_data (file_data, LTO_section_mode_table, &len); if (! data) { internal_error ("cannot read LTO mode table from %s", file_data->file_name); return; } unsigned char *table = ggc_cleared_vec_alloc (1 << 8); file_data->mode_table = table; const struct lto_simple_header_with_strings *header = (const struct lto_simple_header_with_strings *) data; int string_offset; class data_in *data_in; string_offset = sizeof (*header) + header->main_size; lto_input_block ib (data + sizeof (*header), header->main_size, NULL); data_in = lto_data_in_create (file_data, data + string_offset, header->string_size, vNULL); bitpack_d bp = streamer_read_bitpack (&ib); table[VOIDmode] = VOIDmode; table[BLKmode] = BLKmode; unsigned int m; while ((m = bp_unpack_value (&bp, 8)) != VOIDmode) { enum mode_class mclass = bp_unpack_enum (&bp, mode_class, MAX_MODE_CLASS); poly_uint16 size = bp_unpack_poly_value (&bp, 16); poly_uint16 prec = bp_unpack_poly_value (&bp, 16); machine_mode inner = (machine_mode) bp_unpack_value (&bp, 8); poly_uint16 nunits = bp_unpack_poly_value (&bp, 16); unsigned int ibit = 0, fbit = 0; unsigned int real_fmt_len = 0; const char *real_fmt_name = NULL; switch (mclass) { case MODE_FRACT: case MODE_UFRACT: case MODE_ACCUM: case MODE_UACCUM: ibit = bp_unpack_value (&bp, 8); fbit = bp_unpack_value (&bp, 8); break; case MODE_FLOAT: case MODE_DECIMAL_FLOAT: real_fmt_name = bp_unpack_indexed_string (data_in, &bp, &real_fmt_len); break; default: break; } /* First search just the GET_CLASS_NARROWEST_MODE to wider modes, if not found, fallback to all modes. */ int pass; for (pass = 0; pass < 2; pass++) for (machine_mode mr = pass ? VOIDmode : GET_CLASS_NARROWEST_MODE (mclass); pass ? mr < MAX_MACHINE_MODE : mr != VOIDmode; pass ? mr = (machine_mode) (mr + 1) : mr = GET_MODE_WIDER_MODE (mr).else_void ()) if (GET_MODE_CLASS (mr) != mclass || maybe_ne (GET_MODE_SIZE (mr), size) || maybe_ne (GET_MODE_PRECISION (mr), prec) || (inner == m ? GET_MODE_INNER (mr) != mr : GET_MODE_INNER (mr) != table[(int) inner]) || GET_MODE_IBIT (mr) != ibit || GET_MODE_FBIT (mr) != fbit || maybe_ne (GET_MODE_NUNITS (mr), nunits)) continue; else if ((mclass == MODE_FLOAT || mclass == MODE_DECIMAL_FLOAT) && strcmp (REAL_MODE_FORMAT (mr)->name, real_fmt_name) != 0) continue; else { table[m] = mr; pass = 2; break; } unsigned int mname_len; const char *mname = bp_unpack_indexed_string (data_in, &bp, &mname_len); if (pass == 2) { switch (mclass) { case MODE_VECTOR_BOOL: case MODE_VECTOR_INT: case MODE_VECTOR_FLOAT: case MODE_VECTOR_FRACT: case MODE_VECTOR_UFRACT: case MODE_VECTOR_ACCUM: case MODE_VECTOR_UACCUM: /* For unsupported vector modes just use BLKmode, if the scalar mode is supported. */ if (table[(int) inner] != VOIDmode) { table[m] = BLKmode; break; } /* FALLTHRU */ default: fatal_error (UNKNOWN_LOCATION, "unsupported mode %qs", mname); break; } } } lto_data_in_delete (data_in); lto_free_section_data (file_data, LTO_section_mode_table, NULL, data, len); } /* Initialization for the LTO reader. */ void lto_reader_init (void) { lto_streamer_init (); file_name_hash_table = new hash_table (37); string_slot_allocator = new object_allocator ("line map file name hash"); gcc_obstack_init (&file_name_obstack); } /* Free hash table used to stream in location file names. */ void lto_free_file_name_hash (void) { delete file_name_hash_table; file_name_hash_table = NULL; delete string_slot_allocator; string_slot_allocator = NULL; /* file_name_obstack must stay allocated since it is referred to by line map table. */ } /* Create a new data_in object for FILE_DATA. STRINGS is the string table to use with LEN strings. RESOLUTIONS is the vector of linker resolutions (NULL if not using a linker plugin). */ class data_in * lto_data_in_create (struct lto_file_decl_data *file_data, const char *strings, unsigned len, vec resolutions) { class data_in *data_in = new (class data_in); data_in->file_data = file_data; data_in->strings = strings; data_in->strings_len = len; data_in->globals_resolution = resolutions; data_in->reader_cache = streamer_tree_cache_create (false, false, true); return data_in; } /* Remove DATA_IN. */ void lto_data_in_delete (class data_in *data_in) { data_in->globals_resolution.release (); streamer_tree_cache_delete (data_in->reader_cache); delete data_in; }