aboutsummaryrefslogtreecommitdiff
path: root/gcc/cgraph.cc
diff options
context:
space:
mode:
Diffstat (limited to 'gcc/cgraph.cc')
-rw-r--r--gcc/cgraph.cc4273
1 files changed, 4273 insertions, 0 deletions
diff --git a/gcc/cgraph.cc b/gcc/cgraph.cc
new file mode 100644
index 0000000..1bbca9d
--- /dev/null
+++ b/gcc/cgraph.cc
@@ -0,0 +1,4273 @@
+/* Callgraph handling code.
+ Copyright (C) 2003-2022 Free Software Foundation, Inc.
+ Contributed by Jan Hubicka
+
+This file is part of GCC.
+
+GCC is free software; you can redistribute it and/or modify it under
+the terms of the GNU General Public License as published by the Free
+Software Foundation; either version 3, or (at your option) any later
+version.
+
+GCC is distributed in the hope that it will be useful, but WITHOUT ANY
+WARRANTY; without even the implied warranty of MERCHANTABILITY or
+FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
+for more details.
+
+You should have received a copy of the GNU General Public License
+along with GCC; see the file COPYING3. If not see
+<http://www.gnu.org/licenses/>. */
+
+/* This file contains basic routines manipulating call graph
+
+ The call-graph is a data structure designed for inter-procedural
+ optimization. It represents a multi-graph where nodes are functions
+ (symbols within symbol table) and edges are call sites. */
+
+#include "config.h"
+#include "system.h"
+#include "coretypes.h"
+#include "backend.h"
+#include "target.h"
+#include "rtl.h"
+#include "tree.h"
+#include "gimple.h"
+#include "predict.h"
+#include "alloc-pool.h"
+#include "gimple-ssa.h"
+#include "cgraph.h"
+#include "lto-streamer.h"
+#include "fold-const.h"
+#include "varasm.h"
+#include "calls.h"
+#include "print-tree.h"
+#include "langhooks.h"
+#include "intl.h"
+#include "tree-eh.h"
+#include "gimple-iterator.h"
+#include "tree-cfg.h"
+#include "tree-ssa.h"
+#include "value-prof.h"
+#include "ipa-utils.h"
+#include "symbol-summary.h"
+#include "tree-vrp.h"
+#include "ipa-prop.h"
+#include "ipa-fnsummary.h"
+#include "cfgloop.h"
+#include "gimple-pretty-print.h"
+#include "tree-dfa.h"
+#include "profile.h"
+#include "context.h"
+#include "gimplify.h"
+#include "stringpool.h"
+#include "attribs.h"
+#include "selftest.h"
+#include "tree-into-ssa.h"
+#include "ipa-inline.h"
+#include "tree-nested.h"
+#include "symtab-thunks.h"
+#include "symtab-clones.h"
+
+/* FIXME: Only for PROP_loops, but cgraph shouldn't have to know about this. */
+#include "tree-pass.h"
+
+/* Queue of cgraph nodes scheduled to be lowered. */
+symtab_node *x_cgraph_nodes_queue;
+#define cgraph_nodes_queue ((cgraph_node *)x_cgraph_nodes_queue)
+
+/* Symbol table global context. */
+symbol_table *symtab;
+
+/* List of hooks triggered on cgraph_edge events. */
+struct cgraph_edge_hook_list {
+ cgraph_edge_hook hook;
+ void *data;
+ struct cgraph_edge_hook_list *next;
+};
+
+/* List of hooks triggered on cgraph_node events. */
+struct cgraph_node_hook_list {
+ cgraph_node_hook hook;
+ void *data;
+ struct cgraph_node_hook_list *next;
+};
+
+/* List of hooks triggered on events involving two cgraph_edges. */
+struct cgraph_2edge_hook_list {
+ cgraph_2edge_hook hook;
+ void *data;
+ struct cgraph_2edge_hook_list *next;
+};
+
+/* List of hooks triggered on events involving two cgraph_nodes. */
+struct cgraph_2node_hook_list {
+ cgraph_2node_hook hook;
+ void *data;
+ struct cgraph_2node_hook_list *next;
+};
+
+/* Hash descriptor for cgraph_function_version_info. */
+
+struct function_version_hasher : ggc_ptr_hash<cgraph_function_version_info>
+{
+ static hashval_t hash (cgraph_function_version_info *);
+ static bool equal (cgraph_function_version_info *,
+ cgraph_function_version_info *);
+};
+
+/* Map a cgraph_node to cgraph_function_version_info using this htab.
+ The cgraph_function_version_info has a THIS_NODE field that is the
+ corresponding cgraph_node.. */
+
+static GTY(()) hash_table<function_version_hasher> *cgraph_fnver_htab = NULL;
+
+/* Hash function for cgraph_fnver_htab. */
+hashval_t
+function_version_hasher::hash (cgraph_function_version_info *ptr)
+{
+ int uid = ptr->this_node->get_uid ();
+ return (hashval_t)(uid);
+}
+
+/* eq function for cgraph_fnver_htab. */
+bool
+function_version_hasher::equal (cgraph_function_version_info *n1,
+ cgraph_function_version_info *n2)
+{
+ return n1->this_node->get_uid () == n2->this_node->get_uid ();
+}
+
+/* Mark as GC root all allocated nodes. */
+static GTY(()) struct cgraph_function_version_info *
+ version_info_node = NULL;
+
+/* Return true if NODE's address can be compared. */
+
+bool
+symtab_node::address_can_be_compared_p ()
+{
+ /* Address of virtual tables and functions is never compared. */
+ if (DECL_VIRTUAL_P (decl))
+ return false;
+ /* Address of C++ cdtors is never compared. */
+ if (is_a <cgraph_node *> (this)
+ && (DECL_CXX_CONSTRUCTOR_P (decl)
+ || DECL_CXX_DESTRUCTOR_P (decl)))
+ return false;
+ /* Constant pool symbols addresses are never compared.
+ flag_merge_constants permits us to assume the same on readonly vars. */
+ if (is_a <varpool_node *> (this)
+ && (DECL_IN_CONSTANT_POOL (decl)
+ || (flag_merge_constants >= 2
+ && TREE_READONLY (decl) && !TREE_THIS_VOLATILE (decl))))
+ return false;
+ return true;
+}
+
+/* Get the cgraph_function_version_info node corresponding to node. */
+cgraph_function_version_info *
+cgraph_node::function_version (void)
+{
+ cgraph_function_version_info key;
+ key.this_node = this;
+
+ if (cgraph_fnver_htab == NULL)
+ return NULL;
+
+ return cgraph_fnver_htab->find (&key);
+}
+
+/* Insert a new cgraph_function_version_info node into cgraph_fnver_htab
+ corresponding to cgraph_node NODE. */
+cgraph_function_version_info *
+cgraph_node::insert_new_function_version (void)
+{
+ version_info_node = NULL;
+ version_info_node = ggc_cleared_alloc<cgraph_function_version_info> ();
+ version_info_node->this_node = this;
+
+ if (cgraph_fnver_htab == NULL)
+ cgraph_fnver_htab = hash_table<function_version_hasher>::create_ggc (2);
+
+ *cgraph_fnver_htab->find_slot (version_info_node, INSERT)
+ = version_info_node;
+ return version_info_node;
+}
+
+/* Remove the cgraph_function_version_info node given by DECL_V. */
+static void
+delete_function_version (cgraph_function_version_info *decl_v)
+{
+ if (decl_v == NULL)
+ return;
+
+ if (version_info_node == decl_v)
+ version_info_node = NULL;
+
+ if (decl_v->prev != NULL)
+ decl_v->prev->next = decl_v->next;
+
+ if (decl_v->next != NULL)
+ decl_v->next->prev = decl_v->prev;
+
+ if (cgraph_fnver_htab != NULL)
+ cgraph_fnver_htab->remove_elt (decl_v);
+}
+
+/* Remove the cgraph_function_version_info and cgraph_node for DECL. This
+ DECL is a duplicate declaration. */
+void
+cgraph_node::delete_function_version_by_decl (tree decl)
+{
+ cgraph_node *decl_node = cgraph_node::get (decl);
+
+ if (decl_node == NULL)
+ return;
+
+ delete_function_version (decl_node->function_version ());
+
+ decl_node->remove ();
+}
+
+/* Record that DECL1 and DECL2 are semantically identical function
+ versions. */
+void
+cgraph_node::record_function_versions (tree decl1, tree decl2)
+{
+ cgraph_node *decl1_node = cgraph_node::get_create (decl1);
+ cgraph_node *decl2_node = cgraph_node::get_create (decl2);
+ cgraph_function_version_info *decl1_v = NULL;
+ cgraph_function_version_info *decl2_v = NULL;
+ cgraph_function_version_info *before;
+ cgraph_function_version_info *after;
+
+ gcc_assert (decl1_node != NULL && decl2_node != NULL);
+ decl1_v = decl1_node->function_version ();
+ decl2_v = decl2_node->function_version ();
+
+ if (decl1_v != NULL && decl2_v != NULL)
+ return;
+
+ if (decl1_v == NULL)
+ decl1_v = decl1_node->insert_new_function_version ();
+
+ if (decl2_v == NULL)
+ decl2_v = decl2_node->insert_new_function_version ();
+
+ /* Chain decl2_v and decl1_v. All semantically identical versions
+ will be chained together. */
+
+ before = decl1_v;
+ after = decl2_v;
+
+ while (before->next != NULL)
+ before = before->next;
+
+ while (after->prev != NULL)
+ after= after->prev;
+
+ before->next = after;
+ after->prev = before;
+}
+
+/* Initialize callgraph dump file. */
+
+void
+symbol_table::initialize (void)
+{
+ if (!dump_file)
+ dump_file = dump_begin (TDI_cgraph, NULL);
+
+ if (!ipa_clones_dump_file)
+ ipa_clones_dump_file = dump_begin (TDI_clones, NULL);
+}
+
+/* Allocate new callgraph node and insert it into basic data structures. */
+
+cgraph_node *
+symbol_table::create_empty (void)
+{
+ cgraph_count++;
+ return new (ggc_alloc<cgraph_node> ()) cgraph_node (cgraph_max_uid++);
+}
+
+/* Register HOOK to be called with DATA on each removed edge. */
+cgraph_edge_hook_list *
+symbol_table::add_edge_removal_hook (cgraph_edge_hook hook, void *data)
+{
+ cgraph_edge_hook_list *entry;
+ cgraph_edge_hook_list **ptr = &m_first_edge_removal_hook;
+
+ entry = (cgraph_edge_hook_list *) xmalloc (sizeof (*entry));
+ entry->hook = hook;
+ entry->data = data;
+ entry->next = NULL;
+ while (*ptr)
+ ptr = &(*ptr)->next;
+ *ptr = entry;
+ return entry;
+}
+
+/* Remove ENTRY from the list of hooks called on removing edges. */
+void
+symbol_table::remove_edge_removal_hook (cgraph_edge_hook_list *entry)
+{
+ cgraph_edge_hook_list **ptr = &m_first_edge_removal_hook;
+
+ while (*ptr != entry)
+ ptr = &(*ptr)->next;
+ *ptr = entry->next;
+ free (entry);
+}
+
+/* Call all edge removal hooks. */
+void
+symbol_table::call_edge_removal_hooks (cgraph_edge *e)
+{
+ cgraph_edge_hook_list *entry = m_first_edge_removal_hook;
+ while (entry)
+ {
+ entry->hook (e, entry->data);
+ entry = entry->next;
+ }
+}
+
+/* Register HOOK to be called with DATA on each removed node. */
+cgraph_node_hook_list *
+symbol_table::add_cgraph_removal_hook (cgraph_node_hook hook, void *data)
+{
+ cgraph_node_hook_list *entry;
+ cgraph_node_hook_list **ptr = &m_first_cgraph_removal_hook;
+
+ entry = (cgraph_node_hook_list *) xmalloc (sizeof (*entry));
+ entry->hook = hook;
+ entry->data = data;
+ entry->next = NULL;
+ while (*ptr)
+ ptr = &(*ptr)->next;
+ *ptr = entry;
+ return entry;
+}
+
+/* Remove ENTRY from the list of hooks called on removing nodes. */
+void
+symbol_table::remove_cgraph_removal_hook (cgraph_node_hook_list *entry)
+{
+ cgraph_node_hook_list **ptr = &m_first_cgraph_removal_hook;
+
+ while (*ptr != entry)
+ ptr = &(*ptr)->next;
+ *ptr = entry->next;
+ free (entry);
+}
+
+/* Call all node removal hooks. */
+void
+symbol_table::call_cgraph_removal_hooks (cgraph_node *node)
+{
+ cgraph_node_hook_list *entry = m_first_cgraph_removal_hook;
+ while (entry)
+ {
+ entry->hook (node, entry->data);
+ entry = entry->next;
+ }
+}
+
+/* Call all node removal hooks. */
+void
+symbol_table::call_cgraph_insertion_hooks (cgraph_node *node)
+{
+ cgraph_node_hook_list *entry = m_first_cgraph_insertion_hook;
+ while (entry)
+ {
+ entry->hook (node, entry->data);
+ entry = entry->next;
+ }
+}
+
+
+/* Register HOOK to be called with DATA on each inserted node. */
+cgraph_node_hook_list *
+symbol_table::add_cgraph_insertion_hook (cgraph_node_hook hook, void *data)
+{
+ cgraph_node_hook_list *entry;
+ cgraph_node_hook_list **ptr = &m_first_cgraph_insertion_hook;
+
+ entry = (cgraph_node_hook_list *) xmalloc (sizeof (*entry));
+ entry->hook = hook;
+ entry->data = data;
+ entry->next = NULL;
+ while (*ptr)
+ ptr = &(*ptr)->next;
+ *ptr = entry;
+ return entry;
+}
+
+/* Remove ENTRY from the list of hooks called on inserted nodes. */
+void
+symbol_table::remove_cgraph_insertion_hook (cgraph_node_hook_list *entry)
+{
+ cgraph_node_hook_list **ptr = &m_first_cgraph_insertion_hook;
+
+ while (*ptr != entry)
+ ptr = &(*ptr)->next;
+ *ptr = entry->next;
+ free (entry);
+}
+
+/* Register HOOK to be called with DATA on each duplicated edge. */
+cgraph_2edge_hook_list *
+symbol_table::add_edge_duplication_hook (cgraph_2edge_hook hook, void *data)
+{
+ cgraph_2edge_hook_list *entry;
+ cgraph_2edge_hook_list **ptr = &m_first_edge_duplicated_hook;
+
+ entry = (cgraph_2edge_hook_list *) xmalloc (sizeof (*entry));
+ entry->hook = hook;
+ entry->data = data;
+ entry->next = NULL;
+ while (*ptr)
+ ptr = &(*ptr)->next;
+ *ptr = entry;
+ return entry;
+}
+
+/* Remove ENTRY from the list of hooks called on duplicating edges. */
+void
+symbol_table::remove_edge_duplication_hook (cgraph_2edge_hook_list *entry)
+{
+ cgraph_2edge_hook_list **ptr = &m_first_edge_duplicated_hook;
+
+ while (*ptr != entry)
+ ptr = &(*ptr)->next;
+ *ptr = entry->next;
+ free (entry);
+}
+
+/* Call all edge duplication hooks. */
+void
+symbol_table::call_edge_duplication_hooks (cgraph_edge *cs1, cgraph_edge *cs2)
+{
+ cgraph_2edge_hook_list *entry = m_first_edge_duplicated_hook;
+ while (entry)
+ {
+ entry->hook (cs1, cs2, entry->data);
+ entry = entry->next;
+ }
+}
+
+/* Register HOOK to be called with DATA on each duplicated node. */
+cgraph_2node_hook_list *
+symbol_table::add_cgraph_duplication_hook (cgraph_2node_hook hook, void *data)
+{
+ cgraph_2node_hook_list *entry;
+ cgraph_2node_hook_list **ptr = &m_first_cgraph_duplicated_hook;
+
+ entry = (cgraph_2node_hook_list *) xmalloc (sizeof (*entry));
+ entry->hook = hook;
+ entry->data = data;
+ entry->next = NULL;
+ while (*ptr)
+ ptr = &(*ptr)->next;
+ *ptr = entry;
+ return entry;
+}
+
+/* Remove ENTRY from the list of hooks called on duplicating nodes. */
+void
+symbol_table::remove_cgraph_duplication_hook (cgraph_2node_hook_list *entry)
+{
+ cgraph_2node_hook_list **ptr = &m_first_cgraph_duplicated_hook;
+
+ while (*ptr != entry)
+ ptr = &(*ptr)->next;
+ *ptr = entry->next;
+ free (entry);
+}
+
+/* Call all node duplication hooks. */
+void
+symbol_table::call_cgraph_duplication_hooks (cgraph_node *node,
+ cgraph_node *node2)
+{
+ cgraph_2node_hook_list *entry = m_first_cgraph_duplicated_hook;
+ while (entry)
+ {
+ entry->hook (node, node2, entry->data);
+ entry = entry->next;
+ }
+}
+
+/* Return cgraph node assigned to DECL. Create new one when needed. */
+
+cgraph_node *
+cgraph_node::create (tree decl)
+{
+ cgraph_node *node = symtab->create_empty ();
+ gcc_assert (TREE_CODE (decl) == FUNCTION_DECL);
+
+ node->decl = decl;
+
+ if ((flag_openacc || flag_openmp)
+ && lookup_attribute ("omp declare target", DECL_ATTRIBUTES (decl)))
+ {
+ node->offloadable = 1;
+ if (ENABLE_OFFLOADING)
+ g->have_offload = true;
+ }
+
+ if (lookup_attribute ("ifunc", DECL_ATTRIBUTES (decl)))
+ node->ifunc_resolver = true;
+
+ node->register_symbol ();
+ maybe_record_nested_function (node);
+
+ return node;
+}
+
+/* Try to find a call graph node for declaration DECL and if it does not exist
+ or if it corresponds to an inline clone, create a new one. */
+
+cgraph_node *
+cgraph_node::get_create (tree decl)
+{
+ cgraph_node *first_clone = cgraph_node::get (decl);
+
+ if (first_clone && !first_clone->inlined_to)
+ return first_clone;
+
+ cgraph_node *node = cgraph_node::create (decl);
+ if (first_clone)
+ {
+ first_clone->clone_of = node;
+ node->clones = first_clone;
+ node->order = first_clone->order;
+ symtab->symtab_prevail_in_asm_name_hash (node);
+ node->decl->decl_with_vis.symtab_node = node;
+ if (dump_file)
+ fprintf (dump_file, "Introduced new external node "
+ "(%s) and turned into root of the clone tree.\n",
+ node->dump_name ());
+ }
+ else if (dump_file)
+ fprintf (dump_file, "Introduced new external node "
+ "(%s).\n", node->dump_name ());
+ return node;
+}
+
+/* Mark ALIAS as an alias to DECL. DECL_NODE is cgraph node representing
+ the function body is associated with
+ (not necessarily cgraph_node (DECL)). */
+
+cgraph_node *
+cgraph_node::create_alias (tree alias, tree target)
+{
+ cgraph_node *alias_node;
+
+ gcc_assert (TREE_CODE (target) == FUNCTION_DECL
+ || TREE_CODE (target) == IDENTIFIER_NODE);
+ gcc_assert (TREE_CODE (alias) == FUNCTION_DECL);
+ alias_node = cgraph_node::get_create (alias);
+ gcc_assert (!alias_node->definition);
+ alias_node->alias_target = target;
+ alias_node->definition = true;
+ alias_node->alias = true;
+ if (lookup_attribute ("weakref", DECL_ATTRIBUTES (alias)) != NULL)
+ alias_node->transparent_alias = alias_node->weakref = true;
+ if (lookup_attribute ("ifunc", DECL_ATTRIBUTES (alias)))
+ alias_node->ifunc_resolver = true;
+ return alias_node;
+}
+
+/* Attempt to mark ALIAS as an alias to DECL. Return alias node if successful
+ and NULL otherwise.
+ Same body aliases are output whenever the body of DECL is output,
+ and cgraph_node::get (ALIAS) transparently returns
+ cgraph_node::get (DECL). */
+
+cgraph_node *
+cgraph_node::create_same_body_alias (tree alias, tree decl)
+{
+ cgraph_node *n;
+
+ /* If aliases aren't supported by the assembler, fail. */
+ if (!TARGET_SUPPORTS_ALIASES)
+ return NULL;
+
+ /* Langhooks can create same body aliases of symbols not defined.
+ Those are useless. Drop them on the floor. */
+ if (symtab->global_info_ready)
+ return NULL;
+
+ n = cgraph_node::create_alias (alias, decl);
+ n->cpp_implicit_alias = true;
+ if (symtab->cpp_implicit_aliases_done)
+ n->resolve_alias (cgraph_node::get (decl));
+ return n;
+}
+
+/* Add thunk alias into callgraph. The alias declaration is ALIAS and it
+ aliases DECL with an adjustments made into the first parameter.
+ See comments in struct cgraph_thunk_info for detail on the parameters. */
+
+cgraph_node *
+cgraph_node::create_thunk (tree alias, tree, bool this_adjusting,
+ HOST_WIDE_INT fixed_offset,
+ HOST_WIDE_INT virtual_value,
+ HOST_WIDE_INT indirect_offset,
+ tree virtual_offset,
+ tree real_alias)
+{
+ cgraph_node *node;
+
+ node = cgraph_node::get (alias);
+ if (node)
+ node->reset ();
+ else
+ node = cgraph_node::create (alias);
+
+ /* Make sure that if VIRTUAL_OFFSET is in sync with VIRTUAL_VALUE. */
+ gcc_checking_assert (virtual_offset
+ ? virtual_value == wi::to_wide (virtual_offset)
+ : virtual_value == 0);
+
+ node->thunk = true;
+ node->definition = true;
+
+ thunk_info *i;
+ thunk_info local_info;
+ if (symtab->state < CONSTRUCTION)
+ i = &local_info;
+ else
+ i = thunk_info::get_create (node);
+ i->fixed_offset = fixed_offset;
+ i->virtual_value = virtual_value;
+ i->indirect_offset = indirect_offset;
+ i->alias = real_alias;
+ i->this_adjusting = this_adjusting;
+ i->virtual_offset_p = virtual_offset != NULL;
+ if (symtab->state < CONSTRUCTION)
+ i->register_early (node);
+
+ return node;
+}
+
+/* Return the cgraph node that has ASMNAME for its DECL_ASSEMBLER_NAME.
+ Return NULL if there's no such node. */
+
+cgraph_node *
+cgraph_node::get_for_asmname (tree asmname)
+{
+ /* We do not want to look at inline clones. */
+ for (symtab_node *node = symtab_node::get_for_asmname (asmname);
+ node;
+ node = node->next_sharing_asm_name)
+ {
+ cgraph_node *cn = dyn_cast <cgraph_node *> (node);
+ if (cn && !cn->inlined_to)
+ return cn;
+ }
+ return NULL;
+}
+
+/* Returns a hash value for X (which really is a cgraph_edge). */
+
+hashval_t
+cgraph_edge_hasher::hash (cgraph_edge *e)
+{
+ /* This is a really poor hash function, but it is what htab_hash_pointer
+ uses. */
+ return (hashval_t) ((intptr_t)e->call_stmt >> 3);
+}
+
+/* Returns a hash value for X (which really is a cgraph_edge). */
+
+hashval_t
+cgraph_edge_hasher::hash (gimple *call_stmt)
+{
+ /* This is a really poor hash function, but it is what htab_hash_pointer
+ uses. */
+ return (hashval_t) ((intptr_t)call_stmt >> 3);
+}
+
+/* Return nonzero if the call_stmt of cgraph_edge X is stmt *Y. */
+
+inline bool
+cgraph_edge_hasher::equal (cgraph_edge *x, gimple *y)
+{
+ return x->call_stmt == y;
+}
+
+/* Add call graph edge E to call site hash of its caller. */
+
+static inline void
+cgraph_update_edge_in_call_site_hash (cgraph_edge *e)
+{
+ gimple *call = e->call_stmt;
+ *e->caller->call_site_hash->find_slot_with_hash
+ (call, cgraph_edge_hasher::hash (call), INSERT) = e;
+}
+
+/* Add call graph edge E to call site hash of its caller. */
+
+static inline void
+cgraph_add_edge_to_call_site_hash (cgraph_edge *e)
+{
+ /* There are two speculative edges for every statement (one direct,
+ one indirect); always hash the direct one. */
+ if (e->speculative && e->indirect_unknown_callee)
+ return;
+ cgraph_edge **slot = e->caller->call_site_hash->find_slot_with_hash
+ (e->call_stmt, cgraph_edge_hasher::hash (e->call_stmt), INSERT);
+ if (*slot)
+ {
+ gcc_assert (((cgraph_edge *)*slot)->speculative);
+ if (e->callee && (!e->prev_callee
+ || !e->prev_callee->speculative
+ || e->prev_callee->call_stmt != e->call_stmt))
+ *slot = e;
+ return;
+ }
+ gcc_assert (!*slot || e->speculative);
+ *slot = e;
+}
+
+/* Return the callgraph edge representing the GIMPLE_CALL statement
+ CALL_STMT. */
+
+cgraph_edge *
+cgraph_node::get_edge (gimple *call_stmt)
+{
+ cgraph_edge *e, *e2;
+ int n = 0;
+
+ if (call_site_hash)
+ return call_site_hash->find_with_hash
+ (call_stmt, cgraph_edge_hasher::hash (call_stmt));
+
+ /* This loop may turn out to be performance problem. In such case adding
+ hashtables into call nodes with very many edges is probably best
+ solution. It is not good idea to add pointer into CALL_EXPR itself
+ because we want to make possible having multiple cgraph nodes representing
+ different clones of the same body before the body is actually cloned. */
+ for (e = callees; e; e = e->next_callee)
+ {
+ if (e->call_stmt == call_stmt)
+ break;
+ n++;
+ }
+
+ if (!e)
+ for (e = indirect_calls; e; e = e->next_callee)
+ {
+ if (e->call_stmt == call_stmt)
+ break;
+ n++;
+ }
+
+ if (n > 100)
+ {
+ call_site_hash = hash_table<cgraph_edge_hasher>::create_ggc (120);
+ for (e2 = callees; e2; e2 = e2->next_callee)
+ cgraph_add_edge_to_call_site_hash (e2);
+ for (e2 = indirect_calls; e2; e2 = e2->next_callee)
+ cgraph_add_edge_to_call_site_hash (e2);
+ }
+
+ return e;
+}
+
+
+/* Change field call_stmt of edge E to NEW_STMT. If UPDATE_SPECULATIVE and E
+ is any component of speculative edge, then update all components.
+ Speculations can be resolved in the process and EDGE can be removed and
+ deallocated. Return the edge that now represents the call. */
+
+cgraph_edge *
+cgraph_edge::set_call_stmt (cgraph_edge *e, gcall *new_stmt,
+ bool update_speculative)
+{
+ tree decl;
+
+ cgraph_node *new_direct_callee = NULL;
+ if ((e->indirect_unknown_callee || e->speculative)
+ && (decl = gimple_call_fndecl (new_stmt)))
+ {
+ /* Constant propagation and especially inlining can turn an indirect call
+ into a direct one. */
+ new_direct_callee = cgraph_node::get (decl);
+ gcc_checking_assert (new_direct_callee);
+ }
+
+ /* Speculative edges has three component, update all of them
+ when asked to. */
+ if (update_speculative && e->speculative
+ /* If we are about to resolve the speculation by calling make_direct
+ below, do not bother going over all the speculative edges now. */
+ && !new_direct_callee)
+ {
+ cgraph_edge *direct, *indirect, *next;
+ ipa_ref *ref;
+ bool e_indirect = e->indirect_unknown_callee;
+ int n = 0;
+
+ direct = e->first_speculative_call_target ();
+ indirect = e->speculative_call_indirect_edge ();
+
+ gcall *old_stmt = direct->call_stmt;
+ for (cgraph_edge *d = direct; d; d = next)
+ {
+ next = d->next_speculative_call_target ();
+ cgraph_edge *d2 = set_call_stmt (d, new_stmt, false);
+ gcc_assert (d2 == d);
+ n++;
+ }
+ gcc_checking_assert (indirect->num_speculative_call_targets_p () == n);
+ for (unsigned int i = 0; e->caller->iterate_reference (i, ref); i++)
+ if (ref->speculative && ref->stmt == old_stmt)
+ {
+ ref->stmt = new_stmt;
+ n--;
+ }
+
+ indirect = set_call_stmt (indirect, new_stmt, false);
+ return e_indirect ? indirect : direct;
+ }
+
+ if (new_direct_callee)
+ e = make_direct (e, new_direct_callee);
+
+ /* Only direct speculative edges go to call_site_hash. */
+ if (e->caller->call_site_hash
+ && (!e->speculative || !e->indirect_unknown_callee)
+ /* It is possible that edge was previously speculative. In this case
+ we have different value in call stmt hash which needs preserving. */
+ && e->caller->get_edge (e->call_stmt) == e)
+ e->caller->call_site_hash->remove_elt_with_hash
+ (e->call_stmt, cgraph_edge_hasher::hash (e->call_stmt));
+
+ e->call_stmt = new_stmt;
+
+ function *fun = DECL_STRUCT_FUNCTION (e->caller->decl);
+ e->can_throw_external = stmt_can_throw_external (fun, new_stmt);
+ /* Update call stite hash. For speculative calls we only record the first
+ direct edge. */
+ if (e->caller->call_site_hash
+ && (!e->speculative
+ || (e->callee
+ && (!e->prev_callee || !e->prev_callee->speculative
+ || e->prev_callee->call_stmt != e->call_stmt))
+ || (e->speculative && !e->callee)))
+ cgraph_add_edge_to_call_site_hash (e);
+ return e;
+}
+
+/* Allocate a cgraph_edge structure and fill it with data according to the
+ parameters of which only CALLEE can be NULL (when creating an indirect call
+ edge). CLONING_P should be set if properties that are copied from an
+ original edge should not be calculated. */
+
+cgraph_edge *
+symbol_table::create_edge (cgraph_node *caller, cgraph_node *callee,
+ gcall *call_stmt, profile_count count,
+ bool indir_unknown_callee, bool cloning_p)
+{
+ cgraph_edge *edge;
+
+ /* LTO does not actually have access to the call_stmt since these
+ have not been loaded yet. */
+ if (call_stmt)
+ {
+ /* This is a rather expensive check possibly triggering
+ construction of call stmt hashtable. */
+ cgraph_edge *e;
+ gcc_checking_assert (!(e = caller->get_edge (call_stmt))
+ || e->speculative);
+
+ gcc_assert (is_gimple_call (call_stmt));
+ }
+
+ edge = ggc_alloc<cgraph_edge> ();
+ edge->m_summary_id = -1;
+ edges_count++;
+
+ gcc_assert (++edges_max_uid != 0);
+ edge->m_uid = edges_max_uid;
+ edge->aux = NULL;
+ edge->caller = caller;
+ edge->callee = callee;
+ edge->prev_caller = NULL;
+ edge->next_caller = NULL;
+ edge->prev_callee = NULL;
+ edge->next_callee = NULL;
+ edge->lto_stmt_uid = 0;
+ edge->speculative_id = 0;
+
+ edge->count = count;
+ edge->call_stmt = call_stmt;
+ edge->indirect_info = NULL;
+ edge->indirect_inlining_edge = 0;
+ edge->speculative = false;
+ edge->indirect_unknown_callee = indir_unknown_callee;
+ if (call_stmt && caller->call_site_hash)
+ cgraph_add_edge_to_call_site_hash (edge);
+
+ if (cloning_p)
+ return edge;
+
+ edge->can_throw_external
+ = call_stmt ? stmt_can_throw_external (DECL_STRUCT_FUNCTION (caller->decl),
+ call_stmt) : false;
+ edge->inline_failed = CIF_FUNCTION_NOT_CONSIDERED;
+ edge->call_stmt_cannot_inline_p = false;
+
+ if (opt_for_fn (edge->caller->decl, flag_devirtualize)
+ && call_stmt && DECL_STRUCT_FUNCTION (caller->decl))
+ edge->in_polymorphic_cdtor
+ = decl_maybe_in_construction_p (NULL, NULL, call_stmt,
+ caller->decl);
+ else
+ edge->in_polymorphic_cdtor = caller->thunk;
+ if (callee)
+ caller->calls_declare_variant_alt |= callee->declare_variant_alt;
+
+ if (callee && symtab->state != LTO_STREAMING
+ && edge->callee->comdat_local_p ())
+ edge->caller->calls_comdat_local = true;
+
+ return edge;
+}
+
+/* Create edge from a given function to CALLEE in the cgraph. CLONING_P should
+ be set if properties that are copied from an original edge should not be
+ calculated. */
+
+cgraph_edge *
+cgraph_node::create_edge (cgraph_node *callee,
+ gcall *call_stmt, profile_count count, bool cloning_p)
+{
+ cgraph_edge *edge = symtab->create_edge (this, callee, call_stmt, count,
+ false, cloning_p);
+
+ if (!cloning_p)
+ initialize_inline_failed (edge);
+
+ edge->next_caller = callee->callers;
+ if (callee->callers)
+ callee->callers->prev_caller = edge;
+ edge->next_callee = callees;
+ if (callees)
+ callees->prev_callee = edge;
+ callees = edge;
+ callee->callers = edge;
+
+ return edge;
+}
+
+/* Allocate cgraph_indirect_call_info and set its fields to default values. */
+
+cgraph_indirect_call_info *
+cgraph_allocate_init_indirect_info (void)
+{
+ cgraph_indirect_call_info *ii;
+
+ ii = ggc_cleared_alloc<cgraph_indirect_call_info> ();
+ ii->param_index = -1;
+ return ii;
+}
+
+/* Create an indirect edge with a yet-undetermined callee where the call
+ statement destination is a formal parameter of the caller with index
+ PARAM_INDEX. CLONING_P should be set if properties that are copied from an
+ original edge should not be calculated and indirect_info structure should
+ not be calculated. */
+
+cgraph_edge *
+cgraph_node::create_indirect_edge (gcall *call_stmt, int ecf_flags,
+ profile_count count,
+ bool cloning_p)
+{
+ cgraph_edge *edge = symtab->create_edge (this, NULL, call_stmt, count, true,
+ cloning_p);
+ tree target;
+
+ if (!cloning_p)
+ initialize_inline_failed (edge);
+
+ edge->indirect_info = cgraph_allocate_init_indirect_info ();
+ edge->indirect_info->ecf_flags = ecf_flags;
+ edge->indirect_info->vptr_changed = true;
+
+ /* Record polymorphic call info. */
+ if (!cloning_p
+ && call_stmt
+ && (target = gimple_call_fn (call_stmt))
+ && virtual_method_call_p (target))
+ {
+ ipa_polymorphic_call_context context (decl, target, call_stmt);
+
+ /* Only record types can have virtual calls. */
+ edge->indirect_info->polymorphic = true;
+ edge->indirect_info->param_index = -1;
+ edge->indirect_info->otr_token
+ = tree_to_uhwi (OBJ_TYPE_REF_TOKEN (target));
+ edge->indirect_info->otr_type = obj_type_ref_class (target);
+ gcc_assert (TREE_CODE (edge->indirect_info->otr_type) == RECORD_TYPE);
+ edge->indirect_info->context = context;
+ }
+
+ edge->next_callee = indirect_calls;
+ if (indirect_calls)
+ indirect_calls->prev_callee = edge;
+ indirect_calls = edge;
+
+ return edge;
+}
+
+/* Remove the edge from the list of the callees of the caller. */
+
+void
+cgraph_edge::remove_caller (void)
+{
+ if (prev_callee)
+ prev_callee->next_callee = next_callee;
+ if (next_callee)
+ next_callee->prev_callee = prev_callee;
+ if (!prev_callee)
+ {
+ if (indirect_unknown_callee)
+ caller->indirect_calls = next_callee;
+ else
+ caller->callees = next_callee;
+ }
+ if (caller->call_site_hash
+ && this == caller->get_edge (call_stmt))
+ caller->call_site_hash->remove_elt_with_hash
+ (call_stmt, cgraph_edge_hasher::hash (call_stmt));
+}
+
+/* Put the edge onto the free list. */
+
+void
+symbol_table::free_edge (cgraph_edge *e)
+{
+ edges_count--;
+ if (e->m_summary_id != -1)
+ edge_released_summary_ids.safe_push (e->m_summary_id);
+
+ if (e->indirect_info)
+ ggc_free (e->indirect_info);
+ ggc_free (e);
+}
+
+/* Remove the edge in the cgraph. */
+
+void
+cgraph_edge::remove (cgraph_edge *edge)
+{
+ /* Call all edge removal hooks. */
+ symtab->call_edge_removal_hooks (edge);
+
+ if (!edge->indirect_unknown_callee)
+ /* Remove from callers list of the callee. */
+ edge->remove_callee ();
+
+ /* Remove from callees list of the callers. */
+ edge->remove_caller ();
+
+ /* Put the edge onto the free list. */
+ symtab->free_edge (edge);
+}
+
+/* Turn edge into speculative call calling N2. Update
+ the profile so the direct call is taken COUNT times
+ with FREQUENCY.
+
+ At clone materialization time, the indirect call E will
+ be expanded as:
+
+ if (call_dest == N2)
+ n2 ();
+ else
+ call call_dest
+
+ At this time the function just creates the direct call,
+ the reference representing the if conditional and attaches
+ them all to the original indirect call statement.
+
+ speculative_id is used to link direct calls with their corresponding
+ IPA_REF_ADDR references when representing speculative calls.
+
+ Return direct edge created. */
+
+cgraph_edge *
+cgraph_edge::make_speculative (cgraph_node *n2, profile_count direct_count,
+ unsigned int speculative_id)
+{
+ cgraph_node *n = caller;
+ ipa_ref *ref = NULL;
+ cgraph_edge *e2;
+
+ if (dump_file)
+ fprintf (dump_file, "Indirect call -> speculative call %s => %s\n",
+ n->dump_name (), n2->dump_name ());
+ speculative = true;
+ e2 = n->create_edge (n2, call_stmt, direct_count);
+ initialize_inline_failed (e2);
+ e2->speculative = true;
+ if (TREE_NOTHROW (n2->decl))
+ e2->can_throw_external = false;
+ else
+ e2->can_throw_external = can_throw_external;
+ e2->lto_stmt_uid = lto_stmt_uid;
+ e2->speculative_id = speculative_id;
+ e2->in_polymorphic_cdtor = in_polymorphic_cdtor;
+ indirect_info->num_speculative_call_targets++;
+ count -= e2->count;
+ symtab->call_edge_duplication_hooks (this, e2);
+ ref = n->create_reference (n2, IPA_REF_ADDR, call_stmt);
+ ref->lto_stmt_uid = lto_stmt_uid;
+ ref->speculative_id = speculative_id;
+ ref->speculative = speculative;
+ n2->mark_address_taken ();
+ return e2;
+}
+
+/* Speculative call consists of an indirect edge and one or more
+ direct edge+ref pairs.
+
+ Given an edge which is part of speculative call, return the first
+ direct call edge in the speculative call sequence. */
+
+cgraph_edge *
+cgraph_edge::first_speculative_call_target ()
+{
+ cgraph_edge *e = this;
+
+ gcc_checking_assert (e->speculative);
+ if (e->callee)
+ {
+ while (e->prev_callee && e->prev_callee->speculative
+ && e->prev_callee->call_stmt == e->call_stmt
+ && e->prev_callee->lto_stmt_uid == e->lto_stmt_uid)
+ e = e->prev_callee;
+ return e;
+ }
+ /* Call stmt site hash always points to the first target of the
+ speculative call sequence. */
+ if (e->call_stmt)
+ return e->caller->get_edge (e->call_stmt);
+ for (cgraph_edge *e2 = e->caller->callees; true; e2 = e2->next_callee)
+ if (e2->speculative
+ && e->call_stmt == e2->call_stmt
+ && e->lto_stmt_uid == e2->lto_stmt_uid)
+ return e2;
+}
+
+/* We always maintain first direct edge in the call site hash, if one
+ exists. E is going to be removed. See if it is first one and update
+ hash accordingly. INDIRECT is the indirect edge of speculative call.
+ We assume that INDIRECT->num_speculative_call_targets_p () is already
+ updated for removal of E. */
+static void
+update_call_stmt_hash_for_removing_direct_edge (cgraph_edge *e,
+ cgraph_edge *indirect)
+{
+ if (e->caller->call_site_hash)
+ {
+ if (e->caller->get_edge (e->call_stmt) != e)
+ ;
+ else if (!indirect->num_speculative_call_targets_p ())
+ cgraph_update_edge_in_call_site_hash (indirect);
+ else
+ {
+ gcc_checking_assert (e->next_callee && e->next_callee->speculative
+ && e->next_callee->call_stmt == e->call_stmt);
+ cgraph_update_edge_in_call_site_hash (e->next_callee);
+ }
+ }
+}
+
+/* Speculative call EDGE turned out to be direct call to CALLEE_DECL. Remove
+ the speculative call sequence and return edge representing the call, the
+ original EDGE can be removed and deallocated. Return the edge that now
+ represents the call.
+
+ For "speculative" indirect call that contains multiple "speculative"
+ targets (i.e. edge->indirect_info->num_speculative_call_targets > 1),
+ decrease the count and only remove current direct edge.
+
+ If no speculative direct call left to the speculative indirect call, remove
+ the speculative of both the indirect call and corresponding direct edge.
+
+ It is up to caller to iteratively resolve each "speculative" direct call and
+ redirect the call as appropriate. */
+
+cgraph_edge *
+cgraph_edge::resolve_speculation (cgraph_edge *edge, tree callee_decl)
+{
+ cgraph_edge *e2;
+ ipa_ref *ref;
+
+ gcc_assert (edge->speculative && (!callee_decl || edge->callee));
+ if (!edge->callee)
+ e2 = edge->first_speculative_call_target ();
+ else
+ e2 = edge;
+ ref = e2->speculative_call_target_ref ();
+ edge = edge->speculative_call_indirect_edge ();
+ if (!callee_decl
+ || !ref->referred->semantically_equivalent_p
+ (symtab_node::get (callee_decl)))
+ {
+ if (dump_file)
+ {
+ if (callee_decl)
+ {
+ fprintf (dump_file, "Speculative indirect call %s => %s has "
+ "turned out to have contradicting known target ",
+ edge->caller->dump_name (),
+ e2->callee->dump_name ());
+ print_generic_expr (dump_file, callee_decl);
+ fprintf (dump_file, "\n");
+ }
+ else
+ {
+ fprintf (dump_file, "Removing speculative call %s => %s\n",
+ edge->caller->dump_name (),
+ e2->callee->dump_name ());
+ }
+ }
+ }
+ else
+ {
+ cgraph_edge *tmp = edge;
+ if (dump_file)
+ fprintf (dump_file, "Speculative call turned into direct call.\n");
+ edge = e2;
+ e2 = tmp;
+ /* FIXME: If EDGE is inlined, we should scale up the frequencies
+ and counts in the functions inlined through it. */
+ }
+ edge->count += e2->count;
+ if (edge->num_speculative_call_targets_p ())
+ {
+ /* The indirect edge has multiple speculative targets, don't remove
+ speculative until all related direct edges are resolved. */
+ edge->indirect_info->num_speculative_call_targets--;
+ if (!edge->indirect_info->num_speculative_call_targets)
+ edge->speculative = false;
+ }
+ else
+ edge->speculative = false;
+ e2->speculative = false;
+ update_call_stmt_hash_for_removing_direct_edge (e2, edge);
+ ref->remove_reference ();
+ if (e2->indirect_unknown_callee || e2->inline_failed)
+ remove (e2);
+ else
+ e2->callee->remove_symbol_and_inline_clones ();
+ return edge;
+}
+
+/* Return edge corresponding to speculative call to a given target.
+ NULL if speculative call does not have one. */
+
+cgraph_edge *
+cgraph_edge::speculative_call_for_target (cgraph_node *target)
+{
+ for (cgraph_edge *direct = first_speculative_call_target ();
+ direct;
+ direct = direct->next_speculative_call_target ())
+ if (direct->speculative_call_target_ref ()
+ ->referred->semantically_equivalent_p (target))
+ return direct;
+ return NULL;
+}
+
+/* Make an indirect or speculative EDGE with an unknown callee an ordinary edge
+ leading to CALLEE. Speculations can be resolved in the process and EDGE can
+ be removed and deallocated. Return the edge that now represents the
+ call. */
+
+cgraph_edge *
+cgraph_edge::make_direct (cgraph_edge *edge, cgraph_node *callee)
+{
+ gcc_assert (edge->indirect_unknown_callee || edge->speculative);
+
+ /* If we are redirecting speculative call, make it non-speculative. */
+ if (edge->speculative)
+ {
+ cgraph_edge *found = NULL;
+ cgraph_edge *direct, *next;
+
+ edge = edge->speculative_call_indirect_edge ();
+
+ /* Look all speculative targets and remove all but one corresponding
+ to callee (if it exists). */
+ for (direct = edge->first_speculative_call_target ();
+ direct;
+ direct = next)
+ {
+ next = direct->next_speculative_call_target ();
+
+ /* Compare ref not direct->callee. Direct edge is possibly
+ inlined or redirected. */
+ if (!direct->speculative_call_target_ref ()
+ ->referred->semantically_equivalent_p (callee))
+ edge = direct->resolve_speculation (direct, NULL);
+ else
+ {
+ gcc_checking_assert (!found);
+ found = direct;
+ }
+ }
+
+ /* On successful speculation just remove the indirect edge and
+ return the pre existing direct edge.
+ It is important to not remove it and redirect because the direct
+ edge may be inlined or redirected. */
+ if (found)
+ {
+ cgraph_edge *e2 = resolve_speculation (found, callee->decl);
+ gcc_checking_assert (!found->speculative && e2 == found);
+ return found;
+ }
+ gcc_checking_assert (!edge->speculative);
+ }
+
+ edge->indirect_unknown_callee = 0;
+ ggc_free (edge->indirect_info);
+ edge->indirect_info = NULL;
+
+ /* Get the edge out of the indirect edge list. */
+ if (edge->prev_callee)
+ edge->prev_callee->next_callee = edge->next_callee;
+ if (edge->next_callee)
+ edge->next_callee->prev_callee = edge->prev_callee;
+ if (!edge->prev_callee)
+ edge->caller->indirect_calls = edge->next_callee;
+
+ /* Put it into the normal callee list */
+ edge->prev_callee = NULL;
+ edge->next_callee = edge->caller->callees;
+ if (edge->caller->callees)
+ edge->caller->callees->prev_callee = edge;
+ edge->caller->callees = edge;
+
+ /* Insert to callers list of the new callee. */
+ edge->set_callee (callee);
+
+ /* We need to re-determine the inlining status of the edge. */
+ initialize_inline_failed (edge);
+ return edge;
+}
+
+/* Redirect callee of the edge to N. The function does not update underlying
+ call expression. */
+
+void
+cgraph_edge::redirect_callee (cgraph_node *n)
+{
+ bool loc = callee->comdat_local_p ();
+ /* Remove from callers list of the current callee. */
+ remove_callee ();
+
+ /* Insert to callers list of the new callee. */
+ set_callee (n);
+
+ if (!inline_failed)
+ return;
+ if (!loc && n->comdat_local_p ())
+ {
+ cgraph_node *to = caller->inlined_to ? caller->inlined_to : caller;
+ to->calls_comdat_local = true;
+ }
+ else if (loc && !n->comdat_local_p ())
+ {
+ cgraph_node *to = caller->inlined_to ? caller->inlined_to : caller;
+ gcc_checking_assert (to->calls_comdat_local);
+ to->calls_comdat_local = to->check_calls_comdat_local_p ();
+ }
+}
+
+/* If necessary, change the function declaration in the call statement
+ associated with E so that it corresponds to the edge callee. Speculations
+ can be resolved in the process and EDGE can be removed and deallocated.
+
+ The edge could be one of speculative direct call generated from speculative
+ indirect call. In this circumstance, decrease the speculative targets
+ count (i.e. num_speculative_call_targets) and redirect call stmt to the
+ corresponding i-th target. If no speculative direct call left to the
+ speculative indirect call, remove "speculative" of the indirect call and
+ also redirect stmt to it's final direct target.
+
+ It is up to caller to iteratively transform each "speculative"
+ direct call as appropriate. */
+
+gimple *
+cgraph_edge::redirect_call_stmt_to_callee (cgraph_edge *e)
+{
+ tree decl = gimple_call_fndecl (e->call_stmt);
+ gcall *new_stmt;
+ gimple_stmt_iterator gsi;
+
+ if (e->speculative)
+ {
+ /* If there already is an direct call (i.e. as a result of inliner's
+ substitution), forget about speculating. */
+ if (decl)
+ e = make_direct (e->speculative_call_indirect_edge (),
+ cgraph_node::get (decl));
+ else
+ {
+ /* Be sure we redirect all speculative targets before poking
+ abou tindirect edge. */
+ gcc_checking_assert (e->callee);
+ cgraph_edge *indirect = e->speculative_call_indirect_edge ();
+ gcall *new_stmt;
+ ipa_ref *ref;
+
+ /* Expand speculation into GIMPLE code. */
+ if (dump_file)
+ {
+ fprintf (dump_file,
+ "Expanding speculative call of %s -> %s count: ",
+ e->caller->dump_name (),
+ e->callee->dump_name ());
+ e->count.dump (dump_file);
+ fprintf (dump_file, "\n");
+ }
+ push_cfun (DECL_STRUCT_FUNCTION (e->caller->decl));
+
+ profile_count all = indirect->count;
+ for (cgraph_edge *e2 = e->first_speculative_call_target ();
+ e2;
+ e2 = e2->next_speculative_call_target ())
+ all = all + e2->count;
+ profile_probability prob = e->count.probability_in (all);
+ if (!prob.initialized_p ())
+ prob = profile_probability::even ();
+ ref = e->speculative_call_target_ref ();
+ new_stmt = gimple_ic (e->call_stmt,
+ dyn_cast<cgraph_node *> (ref->referred),
+ prob);
+ e->speculative = false;
+ if (indirect->num_speculative_call_targets_p ())
+ {
+ /* The indirect edge has multiple speculative targets, don't
+ remove speculative until all related direct edges are
+ redirected. */
+ indirect->indirect_info->num_speculative_call_targets--;
+ if (!indirect->indirect_info->num_speculative_call_targets)
+ indirect->speculative = false;
+ }
+ else
+ indirect->speculative = false;
+ /* Indirect edges are not both in the call site hash.
+ get it updated. */
+ update_call_stmt_hash_for_removing_direct_edge (e, indirect);
+ cgraph_edge::set_call_stmt (e, new_stmt, false);
+ e->count = gimple_bb (e->call_stmt)->count;
+
+ /* Once we are done with expanding the sequence, update also indirect
+ call probability. Until then the basic block accounts for the
+ sum of indirect edge and all non-expanded speculations. */
+ if (!indirect->speculative)
+ indirect->count = gimple_bb (indirect->call_stmt)->count;
+ ref->speculative = false;
+ ref->stmt = NULL;
+ pop_cfun ();
+ /* Continue redirecting E to proper target. */
+ }
+ }
+
+
+ if (e->indirect_unknown_callee
+ || decl == e->callee->decl)
+ return e->call_stmt;
+
+ if (decl && ipa_saved_clone_sources)
+ {
+ tree *p = ipa_saved_clone_sources->get (e->callee);
+ if (p && decl == *p)
+ {
+ gimple_call_set_fndecl (e->call_stmt, e->callee->decl);
+ return e->call_stmt;
+ }
+ }
+ if (flag_checking && decl)
+ {
+ if (cgraph_node *node = cgraph_node::get (decl))
+ {
+ clone_info *info = clone_info::get (node);
+ gcc_assert (!info || !info->param_adjustments);
+ }
+ }
+
+ clone_info *callee_info = clone_info::get (e->callee);
+ if (symtab->dump_file)
+ {
+ fprintf (symtab->dump_file, "updating call of %s -> %s: ",
+ e->caller->dump_name (), e->callee->dump_name ());
+ print_gimple_stmt (symtab->dump_file, e->call_stmt, 0, dump_flags);
+ if (callee_info && callee_info->param_adjustments)
+ callee_info->param_adjustments->dump (symtab->dump_file);
+ }
+
+ if (ipa_param_adjustments *padjs
+ = callee_info ? callee_info->param_adjustments : NULL)
+ {
+ /* We need to defer cleaning EH info on the new statement to
+ fixup-cfg. We may not have dominator information at this point
+ and thus would end up with unreachable blocks and have no way
+ to communicate that we need to run CFG cleanup then. */
+ int lp_nr = lookup_stmt_eh_lp (e->call_stmt);
+ if (lp_nr != 0)
+ remove_stmt_from_eh_lp (e->call_stmt);
+
+ tree old_fntype = gimple_call_fntype (e->call_stmt);
+ new_stmt = padjs->modify_call (e, false);
+ cgraph_node *origin = e->callee;
+ while (origin->clone_of)
+ origin = origin->clone_of;
+
+ if ((origin->former_clone_of
+ && old_fntype == TREE_TYPE (origin->former_clone_of))
+ || old_fntype == TREE_TYPE (origin->decl))
+ gimple_call_set_fntype (new_stmt, TREE_TYPE (e->callee->decl));
+ else
+ {
+ tree new_fntype = padjs->build_new_function_type (old_fntype, true);
+ gimple_call_set_fntype (new_stmt, new_fntype);
+ }
+
+ if (lp_nr != 0)
+ add_stmt_to_eh_lp (new_stmt, lp_nr);
+ }
+ else
+ {
+ if (flag_checking
+ && !fndecl_built_in_p (e->callee->decl, BUILT_IN_UNREACHABLE))
+ ipa_verify_edge_has_no_modifications (e);
+ new_stmt = e->call_stmt;
+ gimple_call_set_fndecl (new_stmt, e->callee->decl);
+ update_stmt_fn (DECL_STRUCT_FUNCTION (e->caller->decl), new_stmt);
+ }
+
+ /* If changing the call to __cxa_pure_virtual or similar noreturn function,
+ adjust gimple_call_fntype too. */
+ if (gimple_call_noreturn_p (new_stmt)
+ && VOID_TYPE_P (TREE_TYPE (TREE_TYPE (e->callee->decl)))
+ && TYPE_ARG_TYPES (TREE_TYPE (e->callee->decl))
+ && (TREE_VALUE (TYPE_ARG_TYPES (TREE_TYPE (e->callee->decl)))
+ == void_type_node))
+ gimple_call_set_fntype (new_stmt, TREE_TYPE (e->callee->decl));
+
+ /* If the call becomes noreturn, remove the LHS if possible. */
+ tree lhs = gimple_call_lhs (new_stmt);
+ if (lhs
+ && gimple_call_noreturn_p (new_stmt)
+ && (VOID_TYPE_P (TREE_TYPE (gimple_call_fntype (new_stmt)))
+ || should_remove_lhs_p (lhs)))
+ {
+ if (TREE_CODE (lhs) == SSA_NAME)
+ {
+ tree var = create_tmp_reg_fn (DECL_STRUCT_FUNCTION (e->caller->decl),
+ TREE_TYPE (lhs), NULL);
+ var = get_or_create_ssa_default_def
+ (DECL_STRUCT_FUNCTION (e->caller->decl), var);
+ gimple *set_stmt = gimple_build_assign (lhs, var);
+ gsi = gsi_for_stmt (new_stmt);
+ gsi_insert_before_without_update (&gsi, set_stmt, GSI_SAME_STMT);
+ update_stmt_fn (DECL_STRUCT_FUNCTION (e->caller->decl), set_stmt);
+ }
+ gimple_call_set_lhs (new_stmt, NULL_TREE);
+ update_stmt_fn (DECL_STRUCT_FUNCTION (e->caller->decl), new_stmt);
+ }
+
+ /* If new callee has no static chain, remove it. */
+ if (gimple_call_chain (new_stmt) && !DECL_STATIC_CHAIN (e->callee->decl))
+ {
+ gimple_call_set_chain (new_stmt, NULL);
+ update_stmt_fn (DECL_STRUCT_FUNCTION (e->caller->decl), new_stmt);
+ }
+
+ maybe_remove_unused_call_args (DECL_STRUCT_FUNCTION (e->caller->decl),
+ new_stmt);
+
+ e->caller->set_call_stmt_including_clones (e->call_stmt, new_stmt, false);
+
+ if (symtab->dump_file)
+ {
+ fprintf (symtab->dump_file, " updated to:");
+ print_gimple_stmt (symtab->dump_file, e->call_stmt, 0, dump_flags);
+ }
+ return new_stmt;
+}
+
+/* Update or remove the corresponding cgraph edge if a GIMPLE_CALL
+ OLD_STMT changed into NEW_STMT. OLD_CALL is gimple_call_fndecl
+ of OLD_STMT if it was previously call statement.
+ If NEW_STMT is NULL, the call has been dropped without any
+ replacement. */
+
+static void
+cgraph_update_edges_for_call_stmt_node (cgraph_node *node,
+ gimple *old_stmt, tree old_call,
+ gimple *new_stmt)
+{
+ tree new_call = (new_stmt && is_gimple_call (new_stmt))
+ ? gimple_call_fndecl (new_stmt) : 0;
+
+ /* We are seeing indirect calls, then there is nothing to update. */
+ if (!new_call && !old_call)
+ return;
+ /* See if we turned indirect call into direct call or folded call to one builtin
+ into different builtin. */
+ if (old_call != new_call)
+ {
+ cgraph_edge *e = node->get_edge (old_stmt);
+ cgraph_edge *ne = NULL;
+ profile_count count;
+
+ if (e)
+ {
+ /* Keep calls marked as dead dead. */
+ if (new_stmt && is_gimple_call (new_stmt) && e->callee
+ && fndecl_built_in_p (e->callee->decl, BUILT_IN_UNREACHABLE))
+ {
+ cgraph_edge::set_call_stmt (node->get_edge (old_stmt),
+ as_a <gcall *> (new_stmt));
+ return;
+ }
+ /* See if the edge is already there and has the correct callee. It
+ might be so because of indirect inlining has already updated
+ it. We also might've cloned and redirected the edge. */
+ if (new_call && e->callee)
+ {
+ cgraph_node *callee = e->callee;
+ while (callee)
+ {
+ if (callee->decl == new_call
+ || callee->former_clone_of == new_call)
+ {
+ cgraph_edge::set_call_stmt (e, as_a <gcall *> (new_stmt));
+ return;
+ }
+ callee = callee->clone_of;
+ }
+ }
+
+ /* Otherwise remove edge and create new one; we can't simply redirect
+ since function has changed, so inline plan and other information
+ attached to edge is invalid. */
+ count = e->count;
+ if (e->indirect_unknown_callee || e->inline_failed)
+ cgraph_edge::remove (e);
+ else
+ e->callee->remove_symbol_and_inline_clones ();
+ }
+ else if (new_call)
+ {
+ /* We are seeing new direct call; compute profile info based on BB. */
+ basic_block bb = gimple_bb (new_stmt);
+ count = bb->count;
+ }
+
+ if (new_call)
+ {
+ ne = node->create_edge (cgraph_node::get_create (new_call),
+ as_a <gcall *> (new_stmt), count);
+ gcc_assert (ne->inline_failed);
+ }
+ }
+ /* We only updated the call stmt; update pointer in cgraph edge.. */
+ else if (old_stmt != new_stmt)
+ cgraph_edge::set_call_stmt (node->get_edge (old_stmt),
+ as_a <gcall *> (new_stmt));
+}
+
+/* Update or remove the corresponding cgraph edge if a GIMPLE_CALL
+ OLD_STMT changed into NEW_STMT. OLD_DECL is gimple_call_fndecl
+ of OLD_STMT before it was updated (updating can happen inplace). */
+
+void
+cgraph_update_edges_for_call_stmt (gimple *old_stmt, tree old_decl,
+ gimple *new_stmt)
+{
+ cgraph_node *orig = cgraph_node::get (cfun->decl);
+ cgraph_node *node;
+
+ gcc_checking_assert (orig);
+ cgraph_update_edges_for_call_stmt_node (orig, old_stmt, old_decl, new_stmt);
+ if (orig->clones)
+ for (node = orig->clones; node != orig;)
+ {
+ cgraph_update_edges_for_call_stmt_node (node, old_stmt, old_decl,
+ new_stmt);
+ if (node->clones)
+ node = node->clones;
+ else if (node->next_sibling_clone)
+ node = node->next_sibling_clone;
+ else
+ {
+ while (node != orig && !node->next_sibling_clone)
+ node = node->clone_of;
+ if (node != orig)
+ node = node->next_sibling_clone;
+ }
+ }
+}
+
+
+/* Remove all callees from the node. */
+
+void
+cgraph_node::remove_callees (void)
+{
+ cgraph_edge *e, *f;
+
+ calls_comdat_local = false;
+
+ /* It is sufficient to remove the edges from the lists of callers of
+ the callees. The callee list of the node can be zapped with one
+ assignment. */
+ for (e = callees; e; e = f)
+ {
+ f = e->next_callee;
+ symtab->call_edge_removal_hooks (e);
+ if (!e->indirect_unknown_callee)
+ e->remove_callee ();
+ symtab->free_edge (e);
+ }
+ for (e = indirect_calls; e; e = f)
+ {
+ f = e->next_callee;
+ symtab->call_edge_removal_hooks (e);
+ if (!e->indirect_unknown_callee)
+ e->remove_callee ();
+ symtab->free_edge (e);
+ }
+ indirect_calls = NULL;
+ callees = NULL;
+ if (call_site_hash)
+ {
+ call_site_hash->empty ();
+ call_site_hash = NULL;
+ }
+}
+
+/* Remove all callers from the node. */
+
+void
+cgraph_node::remove_callers (void)
+{
+ cgraph_edge *e, *f;
+
+ /* It is sufficient to remove the edges from the lists of callees of
+ the callers. The caller list of the node can be zapped with one
+ assignment. */
+ for (e = callers; e; e = f)
+ {
+ f = e->next_caller;
+ symtab->call_edge_removal_hooks (e);
+ e->remove_caller ();
+ symtab->free_edge (e);
+ }
+ callers = NULL;
+}
+
+/* Helper function for cgraph_release_function_body and free_lang_data.
+ It releases body from function DECL without having to inspect its
+ possibly non-existent symtab node. */
+
+void
+release_function_body (tree decl)
+{
+ function *fn = DECL_STRUCT_FUNCTION (decl);
+ if (fn)
+ {
+ if (fn->cfg
+ && loops_for_fn (fn))
+ {
+ fn->curr_properties &= ~PROP_loops;
+ loop_optimizer_finalize (fn);
+ }
+ if (fn->gimple_df)
+ {
+ delete_tree_ssa (fn);
+ fn->eh = NULL;
+ }
+ if (fn->cfg)
+ {
+ gcc_assert (!dom_info_available_p (fn, CDI_DOMINATORS));
+ gcc_assert (!dom_info_available_p (fn, CDI_POST_DOMINATORS));
+ delete_tree_cfg_annotations (fn);
+ free_cfg (fn);
+ fn->cfg = NULL;
+ }
+ if (fn->value_histograms)
+ free_histograms (fn);
+ gimple_set_body (decl, NULL);
+ /* Struct function hangs a lot of data that would leak if we didn't
+ removed all pointers to it. */
+ ggc_free (fn);
+ DECL_STRUCT_FUNCTION (decl) = NULL;
+ }
+ DECL_SAVED_TREE (decl) = NULL;
+}
+
+/* Release memory used to represent body of function.
+ Use this only for functions that are released before being translated to
+ target code (i.e. RTL). Functions that are compiled to RTL and beyond
+ are free'd in final.c via free_after_compilation().
+ KEEP_ARGUMENTS are useful only if you want to rebuild body as thunk. */
+
+void
+cgraph_node::release_body (bool keep_arguments)
+{
+ ipa_transforms_to_apply.release ();
+ if (!used_as_abstract_origin && symtab->state != PARSING)
+ {
+ DECL_RESULT (decl) = NULL;
+
+ if (!keep_arguments)
+ DECL_ARGUMENTS (decl) = NULL;
+ }
+ /* If the node is abstract and needed, then do not clear
+ DECL_INITIAL of its associated function declaration because it's
+ needed to emit debug info later. */
+ if (!used_as_abstract_origin && DECL_INITIAL (decl))
+ DECL_INITIAL (decl) = error_mark_node;
+ release_function_body (decl);
+ if (lto_file_data)
+ {
+ lto_free_function_in_decl_state_for_node (this);
+ lto_file_data = NULL;
+ }
+ if (flag_checking && clones)
+ {
+ /* It is invalid to release body before materializing clones except
+ for thunks that don't really need a body. Verify also that we do
+ not leak pointers to the call statements. */
+ for (cgraph_node *node = clones; node;
+ node = node->next_sibling_clone)
+ gcc_assert (node->thunk && !node->callees->call_stmt);
+ }
+ remove_callees ();
+ remove_all_references ();
+}
+
+/* Remove function from symbol table. */
+
+void
+cgraph_node::remove (void)
+{
+ bool clone_info_set = false;
+ clone_info *info, saved_info;
+ if (symtab->ipa_clones_dump_file && symtab->cloned_nodes.contains (this))
+ fprintf (symtab->ipa_clones_dump_file,
+ "Callgraph removal;%s;%d;%s;%d;%d\n", asm_name (), order,
+ DECL_SOURCE_FILE (decl), DECL_SOURCE_LINE (decl),
+ DECL_SOURCE_COLUMN (decl));
+
+ if ((info = clone_info::get (this)) != NULL)
+ {
+ saved_info = *info;
+ clone_info_set = true;
+ }
+ symtab->call_cgraph_removal_hooks (this);
+ remove_callers ();
+ remove_callees ();
+ ipa_transforms_to_apply.release ();
+ delete_function_version (function_version ());
+
+ /* Incremental inlining access removed nodes stored in the postorder list.
+ */
+ force_output = false;
+ forced_by_abi = false;
+
+ unregister (clone_info_set ? &saved_info : NULL);
+ if (prev_sibling_clone)
+ prev_sibling_clone->next_sibling_clone = next_sibling_clone;
+ else if (clone_of)
+ clone_of->clones = next_sibling_clone;
+ if (next_sibling_clone)
+ next_sibling_clone->prev_sibling_clone = prev_sibling_clone;
+ if (clones)
+ {
+ cgraph_node *n, *next;
+
+ if (clone_of)
+ {
+ for (n = clones; n->next_sibling_clone; n = n->next_sibling_clone)
+ n->clone_of = clone_of;
+ n->clone_of = clone_of;
+ n->next_sibling_clone = clone_of->clones;
+ if (clone_of->clones)
+ clone_of->clones->prev_sibling_clone = n;
+ clone_of->clones = clones;
+ }
+ else
+ {
+ /* We are removing node with clones. This makes clones inconsistent,
+ but assume they will be removed subsequently and just keep clone
+ tree intact. This can happen in unreachable function removal since
+ we remove unreachable functions in random order, not by bottom-up
+ walk of clone trees. */
+ for (n = clones; n; n = next)
+ {
+ next = n->next_sibling_clone;
+ n->next_sibling_clone = NULL;
+ n->prev_sibling_clone = NULL;
+ n->clone_of = NULL;
+ }
+ }
+ }
+
+ /* While all the clones are removed after being proceeded, the function
+ itself is kept in the cgraph even after it is compiled. Check whether
+ we are done with this body and reclaim it proactively if this is the case.
+ */
+ if (symtab->state != LTO_STREAMING)
+ {
+ cgraph_node *n = cgraph_node::get (decl);
+ if (!n
+ || (!n->clones && !n->clone_of && !n->inlined_to
+ && ((symtab->global_info_ready || in_lto_p)
+ && (TREE_ASM_WRITTEN (n->decl)
+ || DECL_EXTERNAL (n->decl)
+ || !n->analyzed
+ || (!flag_wpa && n->in_other_partition)))))
+ release_body ();
+ }
+ else
+ {
+ lto_free_function_in_decl_state_for_node (this);
+ lto_file_data = NULL;
+ }
+
+ decl = NULL;
+ if (call_site_hash)
+ {
+ call_site_hash->empty ();
+ call_site_hash = NULL;
+ }
+
+ symtab->release_symbol (this);
+}
+
+/* Likewise indicate that a node is having address taken. */
+
+void
+cgraph_node::mark_address_taken (void)
+{
+ /* Indirect inlining can figure out that all uses of the address are
+ inlined. */
+ if (inlined_to)
+ {
+ gcc_assert (cfun->after_inlining);
+ gcc_assert (callers->indirect_inlining_edge);
+ return;
+ }
+ /* FIXME: address_taken flag is used both as a shortcut for testing whether
+ IPA_REF_ADDR reference exists (and thus it should be set on node
+ representing alias we take address of) and as a test whether address
+ of the object was taken (and thus it should be set on node alias is
+ referring to). We should remove the first use and the remove the
+ following set. */
+ address_taken = 1;
+ cgraph_node *node = ultimate_alias_target ();
+ node->address_taken = 1;
+}
+
+/* Return local info node for the compiled function. */
+
+cgraph_node *
+cgraph_node::local_info_node (tree decl)
+{
+ gcc_assert (TREE_CODE (decl) == FUNCTION_DECL);
+ cgraph_node *node = get (decl);
+ if (!node)
+ return NULL;
+ return node->ultimate_alias_target ();
+}
+
+/* Return RTL info for the compiled function. */
+
+cgraph_rtl_info *
+cgraph_node::rtl_info (const_tree decl)
+{
+ gcc_assert (TREE_CODE (decl) == FUNCTION_DECL);
+ cgraph_node *node = get (decl);
+ if (!node)
+ return NULL;
+ enum availability avail;
+ node = node->ultimate_alias_target (&avail);
+ if (decl != current_function_decl
+ && (avail < AVAIL_AVAILABLE
+ || (node->decl != current_function_decl
+ && !TREE_ASM_WRITTEN (node->decl))))
+ return NULL;
+ /* Allocate if it doesn't exist. */
+ if (node->rtl == NULL)
+ {
+ node->rtl = ggc_cleared_alloc<cgraph_rtl_info> ();
+ SET_HARD_REG_SET (node->rtl->function_used_regs);
+ }
+ return node->rtl;
+}
+
+/* Return a string describing the failure REASON. */
+
+const char*
+cgraph_inline_failed_string (cgraph_inline_failed_t reason)
+{
+#undef DEFCIFCODE
+#define DEFCIFCODE(code, type, string) string,
+
+ static const char *cif_string_table[CIF_N_REASONS] = {
+#include "cif-code.def"
+ };
+
+ /* Signedness of an enum type is implementation defined, so cast it
+ to unsigned before testing. */
+ gcc_assert ((unsigned) reason < CIF_N_REASONS);
+ return cif_string_table[reason];
+}
+
+/* Return a type describing the failure REASON. */
+
+cgraph_inline_failed_type_t
+cgraph_inline_failed_type (cgraph_inline_failed_t reason)
+{
+#undef DEFCIFCODE
+#define DEFCIFCODE(code, type, string) type,
+
+ static cgraph_inline_failed_type_t cif_type_table[CIF_N_REASONS] = {
+#include "cif-code.def"
+ };
+
+ /* Signedness of an enum type is implementation defined, so cast it
+ to unsigned before testing. */
+ gcc_assert ((unsigned) reason < CIF_N_REASONS);
+ return cif_type_table[reason];
+}
+
+/* Names used to print out the availability enum. */
+const char * const cgraph_availability_names[] =
+ {"unset", "not_available", "overwritable", "available", "local"};
+
+/* Output flags of edge to a file F. */
+
+void
+cgraph_edge::dump_edge_flags (FILE *f)
+{
+ if (speculative)
+ fprintf (f, "(speculative) ");
+ if (!inline_failed)
+ fprintf (f, "(inlined) ");
+ if (call_stmt_cannot_inline_p)
+ fprintf (f, "(call_stmt_cannot_inline_p) ");
+ if (indirect_inlining_edge)
+ fprintf (f, "(indirect_inlining) ");
+ if (count.initialized_p ())
+ {
+ fprintf (f, "(");
+ count.dump (f);
+ fprintf (f, ",");
+ fprintf (f, "%.2f per call) ", sreal_frequency ().to_double ());
+ }
+ if (can_throw_external)
+ fprintf (f, "(can throw external) ");
+}
+
+/* Dump edge to stderr. */
+
+void
+cgraph_edge::debug (void)
+{
+ fprintf (stderr, "%s -> %s ", caller->dump_asm_name (),
+ callee == NULL ? "(null)" : callee->dump_asm_name ());
+ dump_edge_flags (stderr);
+ fprintf (stderr, "\n\n");
+ caller->debug ();
+ if (callee != NULL)
+ callee->debug ();
+}
+
+/* Dump call graph node to file F. */
+
+void
+cgraph_node::dump (FILE *f)
+{
+ cgraph_edge *edge;
+
+ dump_base (f);
+
+ if (inlined_to)
+ fprintf (f, " Function %s is inline copy in %s\n",
+ dump_name (),
+ inlined_to->dump_name ());
+ if (clone_of)
+ fprintf (f, " Clone of %s\n", clone_of->dump_asm_name ());
+ if (symtab->function_flags_ready)
+ fprintf (f, " Availability: %s\n",
+ cgraph_availability_names [get_availability ()]);
+
+ if (profile_id)
+ fprintf (f, " Profile id: %i\n",
+ profile_id);
+ if (unit_id)
+ fprintf (f, " Unit id: %i\n",
+ unit_id);
+ cgraph_function_version_info *vi = function_version ();
+ if (vi != NULL)
+ {
+ fprintf (f, " Version info: ");
+ if (vi->prev != NULL)
+ {
+ fprintf (f, "prev: ");
+ fprintf (f, "%s ", vi->prev->this_node->dump_asm_name ());
+ }
+ if (vi->next != NULL)
+ {
+ fprintf (f, "next: ");
+ fprintf (f, "%s ", vi->next->this_node->dump_asm_name ());
+ }
+ if (vi->dispatcher_resolver != NULL_TREE)
+ fprintf (f, "dispatcher: %s",
+ lang_hooks.decl_printable_name (vi->dispatcher_resolver, 2));
+
+ fprintf (f, "\n");
+ }
+ fprintf (f, " Function flags:");
+ if (count.initialized_p ())
+ {
+ fprintf (f, " count:");
+ count.dump (f);
+ }
+ if (tp_first_run > 0)
+ fprintf (f, " first_run:%" PRId64, (int64_t) tp_first_run);
+ if (cgraph_node *origin = nested_function_origin (this))
+ fprintf (f, " nested in:%s", origin->dump_asm_name ());
+ if (gimple_has_body_p (decl))
+ fprintf (f, " body");
+ if (process)
+ fprintf (f, " process");
+ if (local)
+ fprintf (f, " local");
+ if (redefined_extern_inline)
+ fprintf (f, " redefined_extern_inline");
+ if (only_called_at_startup)
+ fprintf (f, " only_called_at_startup");
+ if (only_called_at_exit)
+ fprintf (f, " only_called_at_exit");
+ if (tm_clone)
+ fprintf (f, " tm_clone");
+ if (calls_comdat_local)
+ fprintf (f, " calls_comdat_local");
+ if (icf_merged)
+ fprintf (f, " icf_merged");
+ if (merged_comdat)
+ fprintf (f, " merged_comdat");
+ if (merged_extern_inline)
+ fprintf (f, " merged_extern_inline");
+ if (split_part)
+ fprintf (f, " split_part");
+ if (indirect_call_target)
+ fprintf (f, " indirect_call_target");
+ if (nonfreeing_fn)
+ fprintf (f, " nonfreeing_fn");
+ if (DECL_STATIC_CONSTRUCTOR (decl))
+ fprintf (f," static_constructor (priority:%i)", get_init_priority ());
+ if (DECL_STATIC_DESTRUCTOR (decl))
+ fprintf (f," static_destructor (priority:%i)", get_fini_priority ());
+ if (frequency == NODE_FREQUENCY_HOT)
+ fprintf (f, " hot");
+ if (frequency == NODE_FREQUENCY_UNLIKELY_EXECUTED)
+ fprintf (f, " unlikely_executed");
+ if (frequency == NODE_FREQUENCY_EXECUTED_ONCE)
+ fprintf (f, " executed_once");
+ if (opt_for_fn (decl, optimize_size))
+ fprintf (f, " optimize_size");
+ if (parallelized_function)
+ fprintf (f, " parallelized_function");
+ if (DECL_IS_MALLOC (decl))
+ fprintf (f, " decl_is_malloc");
+ if (DECL_IS_OPERATOR_NEW_P (decl))
+ fprintf (f, " %soperator_new",
+ DECL_IS_REPLACEABLE_OPERATOR (decl) ? "replaceable_" : "");
+ if (DECL_IS_OPERATOR_DELETE_P (decl))
+ fprintf (f, " %soperator_delete",
+ DECL_IS_REPLACEABLE_OPERATOR (decl) ? "replaceable_" : "");
+
+ if (DECL_STATIC_CHAIN (decl))
+ fprintf (f, " static_chain");
+
+ fprintf (f, "\n");
+
+ if (thunk)
+ {
+ fprintf (f, " Thunk");
+ thunk_info::get (this)->dump (f);
+ }
+ else if (former_thunk_p ())
+ {
+ fprintf (f, " Former thunk ");
+ thunk_info::get (this)->dump (f);
+ }
+ else gcc_checking_assert (!thunk_info::get (this));
+
+ fprintf (f, " Called by: ");
+
+ profile_count sum = profile_count::zero ();
+ for (edge = callers; edge; edge = edge->next_caller)
+ {
+ fprintf (f, "%s ", edge->caller->dump_asm_name ());
+ edge->dump_edge_flags (f);
+ if (edge->count.initialized_p ())
+ sum += edge->count.ipa ();
+ }
+
+ fprintf (f, "\n Calls: ");
+ for (edge = callees; edge; edge = edge->next_callee)
+ {
+ fprintf (f, "%s ", edge->callee->dump_asm_name ());
+ edge->dump_edge_flags (f);
+ }
+ fprintf (f, "\n");
+
+ if (!body_removed && count.ipa ().initialized_p ())
+ {
+ bool ok = true;
+ bool min = false;
+ ipa_ref *ref;
+
+ FOR_EACH_ALIAS (this, ref)
+ if (dyn_cast <cgraph_node *> (ref->referring)->count.initialized_p ())
+ sum += dyn_cast <cgraph_node *> (ref->referring)->count.ipa ();
+
+ if (inlined_to
+ || (symtab->state < EXPANSION
+ && ultimate_alias_target () == this && only_called_directly_p ()))
+ ok = !count.ipa ().differs_from_p (sum);
+ else if (count.ipa () > profile_count::from_gcov_type (100)
+ && count.ipa () < sum.apply_scale (99, 100))
+ ok = false, min = true;
+ if (!ok)
+ {
+ fprintf (f, " Invalid sum of caller counts ");
+ sum.dump (f);
+ if (min)
+ fprintf (f, ", should be at most ");
+ else
+ fprintf (f, ", should be ");
+ count.ipa ().dump (f);
+ fprintf (f, "\n");
+ }
+ }
+
+ for (edge = indirect_calls; edge; edge = edge->next_callee)
+ {
+ if (edge->indirect_info->polymorphic)
+ {
+ fprintf (f, " Polymorphic indirect call of type ");
+ print_generic_expr (f, edge->indirect_info->otr_type, TDF_SLIM);
+ fprintf (f, " token:%i", (int) edge->indirect_info->otr_token);
+ }
+ else
+ fprintf (f, " Indirect call");
+ edge->dump_edge_flags (f);
+ if (edge->indirect_info->param_index != -1)
+ {
+ fprintf (f, "of param:%i ", edge->indirect_info->param_index);
+ if (edge->indirect_info->agg_contents)
+ fprintf (f, "loaded from %s %s at offset %i ",
+ edge->indirect_info->member_ptr ? "member ptr" : "aggregate",
+ edge->indirect_info->by_ref ? "passed by reference":"",
+ (int)edge->indirect_info->offset);
+ if (edge->indirect_info->vptr_changed)
+ fprintf (f, "(vptr maybe changed) ");
+ }
+ fprintf (f, "num speculative call targets: %i\n",
+ edge->indirect_info->num_speculative_call_targets);
+ if (edge->indirect_info->polymorphic)
+ edge->indirect_info->context.dump (f);
+ }
+}
+
+/* Dump call graph node to file F in graphviz format. */
+
+void
+cgraph_node::dump_graphviz (FILE *f)
+{
+ cgraph_edge *edge;
+
+ for (edge = callees; edge; edge = edge->next_callee)
+ {
+ cgraph_node *callee = edge->callee;
+
+ fprintf (f, "\t\"%s\" -> \"%s\"\n", dump_name (), callee->dump_name ());
+ }
+}
+
+
+/* Dump call graph node NODE to stderr. */
+
+DEBUG_FUNCTION void
+cgraph_node::debug (void)
+{
+ dump (stderr);
+}
+
+/* Dump the callgraph to file F. */
+
+void
+cgraph_node::dump_cgraph (FILE *f)
+{
+ cgraph_node *node;
+
+ fprintf (f, "callgraph:\n\n");
+ FOR_EACH_FUNCTION (node)
+ node->dump (f);
+}
+
+/* Return true when the DECL can possibly be inlined. */
+
+bool
+cgraph_function_possibly_inlined_p (tree decl)
+{
+ if (!symtab->global_info_ready)
+ return !DECL_UNINLINABLE (decl);
+ return DECL_POSSIBLY_INLINED (decl);
+}
+
+/* Return function availability. See cgraph.h for description of individual
+ return values. */
+enum availability
+cgraph_node::get_availability (symtab_node *ref)
+{
+ if (ref)
+ {
+ cgraph_node *cref = dyn_cast <cgraph_node *> (ref);
+ if (cref)
+ ref = cref->inlined_to;
+ }
+ enum availability avail;
+ if (!analyzed && !in_other_partition)
+ avail = AVAIL_NOT_AVAILABLE;
+ else if (local)
+ avail = AVAIL_LOCAL;
+ else if (inlined_to)
+ avail = AVAIL_AVAILABLE;
+ else if (transparent_alias)
+ ultimate_alias_target (&avail, ref);
+ else if (ifunc_resolver
+ || lookup_attribute ("noipa", DECL_ATTRIBUTES (decl)))
+ avail = AVAIL_INTERPOSABLE;
+ else if (!externally_visible)
+ avail = AVAIL_AVAILABLE;
+ /* If this is a reference from symbol itself and there are no aliases, we
+ may be sure that the symbol was not interposed by something else because
+ the symbol itself would be unreachable otherwise.
+
+ Also comdat groups are always resolved in groups. */
+ else if ((this == ref && !has_aliases_p ())
+ || (ref && get_comdat_group ()
+ && get_comdat_group () == ref->get_comdat_group ()))
+ avail = AVAIL_AVAILABLE;
+ /* Inline functions are safe to be analyzed even if their symbol can
+ be overwritten at runtime. It is not meaningful to enforce any sane
+ behavior on replacing inline function by different body. */
+ else if (DECL_DECLARED_INLINE_P (decl))
+ avail = AVAIL_AVAILABLE;
+
+ /* If the function can be overwritten, return OVERWRITABLE. Take
+ care at least of two notable extensions - the COMDAT functions
+ used to share template instantiations in C++ (this is symmetric
+ to code cp_cannot_inline_tree_fn and probably shall be shared and
+ the inlinability hooks completely eliminated). */
+
+ else if (decl_replaceable_p (decl, semantic_interposition)
+ && !DECL_EXTERNAL (decl))
+ avail = AVAIL_INTERPOSABLE;
+ else avail = AVAIL_AVAILABLE;
+
+ return avail;
+}
+
+/* Worker for cgraph_node_can_be_local_p. */
+static bool
+cgraph_node_cannot_be_local_p_1 (cgraph_node *node, void *)
+{
+ return !(!node->force_output
+ && !node->ifunc_resolver
+ /* Limitation of gas requires us to output targets of symver aliases
+ as global symbols. This is binutils PR 25295. */
+ && !node->symver
+ && ((DECL_COMDAT (node->decl)
+ && !node->forced_by_abi
+ && !node->used_from_object_file_p ()
+ && !node->same_comdat_group)
+ || !node->externally_visible));
+}
+
+/* Return true if cgraph_node can be made local for API change.
+ Extern inline functions and C++ COMDAT functions can be made local
+ at the expense of possible code size growth if function is used in multiple
+ compilation units. */
+bool
+cgraph_node::can_be_local_p (void)
+{
+ return (!address_taken
+ && !call_for_symbol_thunks_and_aliases (cgraph_node_cannot_be_local_p_1,
+ NULL, true));
+}
+
+/* Call callback on cgraph_node, thunks and aliases associated to cgraph_node.
+ When INCLUDE_OVERWRITABLE is false, overwritable symbols are
+ skipped. When EXCLUDE_VIRTUAL_THUNKS is true, virtual thunks are
+ skipped. */
+bool
+cgraph_node::call_for_symbol_thunks_and_aliases (bool (*callback)
+ (cgraph_node *, void *),
+ void *data,
+ bool include_overwritable,
+ bool exclude_virtual_thunks)
+{
+ cgraph_edge *e;
+ ipa_ref *ref;
+ enum availability avail = AVAIL_AVAILABLE;
+
+ if (include_overwritable
+ || (avail = get_availability ()) > AVAIL_INTERPOSABLE)
+ {
+ if (callback (this, data))
+ return true;
+ }
+ FOR_EACH_ALIAS (this, ref)
+ {
+ cgraph_node *alias = dyn_cast <cgraph_node *> (ref->referring);
+ if (include_overwritable
+ || alias->get_availability () > AVAIL_INTERPOSABLE)
+ if (alias->call_for_symbol_thunks_and_aliases (callback, data,
+ include_overwritable,
+ exclude_virtual_thunks))
+ return true;
+ }
+ if (avail <= AVAIL_INTERPOSABLE)
+ return false;
+ for (e = callers; e; e = e->next_caller)
+ if (e->caller->thunk
+ && (include_overwritable
+ || e->caller->get_availability () > AVAIL_INTERPOSABLE)
+ && !(exclude_virtual_thunks
+ && thunk_info::get (e->caller)->virtual_offset_p))
+ if (e->caller->call_for_symbol_thunks_and_aliases (callback, data,
+ include_overwritable,
+ exclude_virtual_thunks))
+ return true;
+
+ return false;
+}
+
+/* Worker to bring NODE local. */
+
+bool
+cgraph_node::make_local (cgraph_node *node, void *)
+{
+ gcc_checking_assert (node->can_be_local_p ());
+ if (DECL_COMDAT (node->decl) || DECL_EXTERNAL (node->decl))
+ {
+ node->make_decl_local ();
+ node->set_section (NULL);
+ node->set_comdat_group (NULL);
+ node->externally_visible = false;
+ node->forced_by_abi = false;
+ node->local = true;
+ node->set_section (NULL);
+ node->unique_name = ((node->resolution == LDPR_PREVAILING_DEF_IRONLY
+ || node->resolution == LDPR_PREVAILING_DEF_IRONLY_EXP)
+ && !flag_incremental_link);
+ node->resolution = LDPR_PREVAILING_DEF_IRONLY;
+ gcc_assert (node->get_availability () == AVAIL_LOCAL);
+ }
+ return false;
+}
+
+/* Bring cgraph node local. */
+
+void
+cgraph_node::make_local (void)
+{
+ call_for_symbol_thunks_and_aliases (cgraph_node::make_local, NULL, true);
+}
+
+/* Worker to set nothrow flag. */
+
+static void
+set_nothrow_flag_1 (cgraph_node *node, bool nothrow, bool non_call,
+ bool *changed)
+{
+ cgraph_edge *e;
+
+ if (nothrow && !TREE_NOTHROW (node->decl))
+ {
+ /* With non-call exceptions we can't say for sure if other function body
+ was not possibly optimized to still throw. */
+ if (!non_call || node->binds_to_current_def_p ())
+ {
+ TREE_NOTHROW (node->decl) = true;
+ *changed = true;
+ for (e = node->callers; e; e = e->next_caller)
+ e->can_throw_external = false;
+ }
+ }
+ else if (!nothrow && TREE_NOTHROW (node->decl))
+ {
+ TREE_NOTHROW (node->decl) = false;
+ *changed = true;
+ }
+ ipa_ref *ref;
+ FOR_EACH_ALIAS (node, ref)
+ {
+ cgraph_node *alias = dyn_cast <cgraph_node *> (ref->referring);
+ if (!nothrow || alias->get_availability () > AVAIL_INTERPOSABLE)
+ set_nothrow_flag_1 (alias, nothrow, non_call, changed);
+ }
+ for (cgraph_edge *e = node->callers; e; e = e->next_caller)
+ if (e->caller->thunk
+ && (!nothrow || e->caller->get_availability () > AVAIL_INTERPOSABLE))
+ set_nothrow_flag_1 (e->caller, nothrow, non_call, changed);
+}
+
+/* Set TREE_NOTHROW on NODE's decl and on aliases of NODE
+ if any to NOTHROW. */
+
+bool
+cgraph_node::set_nothrow_flag (bool nothrow)
+{
+ bool changed = false;
+ bool non_call = opt_for_fn (decl, flag_non_call_exceptions);
+
+ if (!nothrow || get_availability () > AVAIL_INTERPOSABLE)
+ set_nothrow_flag_1 (this, nothrow, non_call, &changed);
+ else
+ {
+ ipa_ref *ref;
+
+ FOR_EACH_ALIAS (this, ref)
+ {
+ cgraph_node *alias = dyn_cast <cgraph_node *> (ref->referring);
+ if (!nothrow || alias->get_availability () > AVAIL_INTERPOSABLE)
+ set_nothrow_flag_1 (alias, nothrow, non_call, &changed);
+ }
+ }
+ return changed;
+}
+
+/* Worker to set malloc flag. */
+static void
+set_malloc_flag_1 (cgraph_node *node, bool malloc_p, bool *changed)
+{
+ if (malloc_p && !DECL_IS_MALLOC (node->decl))
+ {
+ DECL_IS_MALLOC (node->decl) = true;
+ *changed = true;
+ }
+
+ ipa_ref *ref;
+ FOR_EACH_ALIAS (node, ref)
+ {
+ cgraph_node *alias = dyn_cast<cgraph_node *> (ref->referring);
+ if (!malloc_p || alias->get_availability () > AVAIL_INTERPOSABLE)
+ set_malloc_flag_1 (alias, malloc_p, changed);
+ }
+
+ for (cgraph_edge *e = node->callers; e; e = e->next_caller)
+ if (e->caller->thunk
+ && (!malloc_p || e->caller->get_availability () > AVAIL_INTERPOSABLE))
+ set_malloc_flag_1 (e->caller, malloc_p, changed);
+}
+
+/* Set DECL_IS_MALLOC on NODE's decl and on NODE's aliases if any. */
+
+bool
+cgraph_node::set_malloc_flag (bool malloc_p)
+{
+ bool changed = false;
+
+ if (!malloc_p || get_availability () > AVAIL_INTERPOSABLE)
+ set_malloc_flag_1 (this, malloc_p, &changed);
+ else
+ {
+ ipa_ref *ref;
+
+ FOR_EACH_ALIAS (this, ref)
+ {
+ cgraph_node *alias = dyn_cast<cgraph_node *> (ref->referring);
+ if (!malloc_p || alias->get_availability () > AVAIL_INTERPOSABLE)
+ set_malloc_flag_1 (alias, malloc_p, &changed);
+ }
+ }
+ return changed;
+}
+
+/* Worker to set noreturng flag. */
+static void
+set_noreturn_flag_1 (cgraph_node *node, bool noreturn_p, bool *changed)
+{
+ if (noreturn_p && !TREE_THIS_VOLATILE (node->decl))
+ {
+ TREE_THIS_VOLATILE (node->decl) = true;
+ *changed = true;
+ }
+
+ ipa_ref *ref;
+ FOR_EACH_ALIAS (node, ref)
+ {
+ cgraph_node *alias = dyn_cast<cgraph_node *> (ref->referring);
+ if (!noreturn_p || alias->get_availability () > AVAIL_INTERPOSABLE)
+ set_noreturn_flag_1 (alias, noreturn_p, changed);
+ }
+
+ for (cgraph_edge *e = node->callers; e; e = e->next_caller)
+ if (e->caller->thunk
+ && (!noreturn_p || e->caller->get_availability () > AVAIL_INTERPOSABLE))
+ set_noreturn_flag_1 (e->caller, noreturn_p, changed);
+}
+
+/* Set TREE_THIS_VOLATILE on NODE's decl and on NODE's aliases if any. */
+
+bool
+cgraph_node::set_noreturn_flag (bool noreturn_p)
+{
+ bool changed = false;
+
+ if (!noreturn_p || get_availability () > AVAIL_INTERPOSABLE)
+ set_noreturn_flag_1 (this, noreturn_p, &changed);
+ else
+ {
+ ipa_ref *ref;
+
+ FOR_EACH_ALIAS (this, ref)
+ {
+ cgraph_node *alias = dyn_cast<cgraph_node *> (ref->referring);
+ if (!noreturn_p || alias->get_availability () > AVAIL_INTERPOSABLE)
+ set_noreturn_flag_1 (alias, noreturn_p, &changed);
+ }
+ }
+ return changed;
+}
+
+/* Worker to set_const_flag. */
+
+static void
+set_const_flag_1 (cgraph_node *node, bool set_const, bool looping,
+ bool *changed)
+{
+ /* Static constructors and destructors without a side effect can be
+ optimized out. */
+ if (set_const && !looping)
+ {
+ if (DECL_STATIC_CONSTRUCTOR (node->decl))
+ {
+ DECL_STATIC_CONSTRUCTOR (node->decl) = 0;
+ *changed = true;
+ }
+ if (DECL_STATIC_DESTRUCTOR (node->decl))
+ {
+ DECL_STATIC_DESTRUCTOR (node->decl) = 0;
+ *changed = true;
+ }
+ }
+ if (!set_const)
+ {
+ if (TREE_READONLY (node->decl))
+ {
+ TREE_READONLY (node->decl) = 0;
+ DECL_LOOPING_CONST_OR_PURE_P (node->decl) = false;
+ *changed = true;
+ }
+ }
+ else
+ {
+ /* Consider function:
+
+ bool a(int *p)
+ {
+ return *p==*p;
+ }
+
+ During early optimization we will turn this into:
+
+ bool a(int *p)
+ {
+ return true;
+ }
+
+ Now if this function will be detected as CONST however when interposed
+ it may end up being just pure. We always must assume the worst
+ scenario here. */
+ if (TREE_READONLY (node->decl))
+ {
+ if (!looping && DECL_LOOPING_CONST_OR_PURE_P (node->decl))
+ {
+ DECL_LOOPING_CONST_OR_PURE_P (node->decl) = false;
+ *changed = true;
+ }
+ }
+ else if (node->binds_to_current_def_p ())
+ {
+ TREE_READONLY (node->decl) = true;
+ DECL_LOOPING_CONST_OR_PURE_P (node->decl) = looping;
+ DECL_PURE_P (node->decl) = false;
+ *changed = true;
+ }
+ else
+ {
+ if (dump_file && (dump_flags & TDF_DETAILS))
+ fprintf (dump_file, "Dropping state to PURE because function does "
+ "not bind to current def.\n");
+ if (!DECL_PURE_P (node->decl))
+ {
+ DECL_PURE_P (node->decl) = true;
+ DECL_LOOPING_CONST_OR_PURE_P (node->decl) = looping;
+ *changed = true;
+ }
+ else if (!looping && DECL_LOOPING_CONST_OR_PURE_P (node->decl))
+ {
+ DECL_LOOPING_CONST_OR_PURE_P (node->decl) = false;
+ *changed = true;
+ }
+ }
+ }
+
+ ipa_ref *ref;
+ FOR_EACH_ALIAS (node, ref)
+ {
+ cgraph_node *alias = dyn_cast <cgraph_node *> (ref->referring);
+ if (!set_const || alias->get_availability () > AVAIL_INTERPOSABLE)
+ set_const_flag_1 (alias, set_const, looping, changed);
+ }
+ for (cgraph_edge *e = node->callers; e; e = e->next_caller)
+ if (e->caller->thunk
+ && (!set_const || e->caller->get_availability () > AVAIL_INTERPOSABLE))
+ {
+ /* Virtual thunks access virtual offset in the vtable, so they can
+ only be pure, never const. */
+ if (set_const
+ && (thunk_info::get (e->caller)->virtual_offset_p
+ || !node->binds_to_current_def_p (e->caller)))
+ *changed |= e->caller->set_pure_flag (true, looping);
+ else
+ set_const_flag_1 (e->caller, set_const, looping, changed);
+ }
+}
+
+/* If SET_CONST is true, mark function, aliases and thunks to be ECF_CONST.
+ If SET_CONST if false, clear the flag.
+
+ When setting the flag be careful about possible interposition and
+ do not set the flag for functions that can be interposed and set pure
+ flag for functions that can bind to other definition.
+
+ Return true if any change was done. */
+
+bool
+cgraph_node::set_const_flag (bool set_const, bool looping)
+{
+ bool changed = false;
+ if (!set_const || get_availability () > AVAIL_INTERPOSABLE)
+ set_const_flag_1 (this, set_const, looping, &changed);
+ else
+ {
+ ipa_ref *ref;
+
+ FOR_EACH_ALIAS (this, ref)
+ {
+ cgraph_node *alias = dyn_cast <cgraph_node *> (ref->referring);
+ if (!set_const || alias->get_availability () > AVAIL_INTERPOSABLE)
+ set_const_flag_1 (alias, set_const, looping, &changed);
+ }
+ }
+ return changed;
+}
+
+/* Info used by set_pure_flag_1. */
+
+struct set_pure_flag_info
+{
+ bool pure;
+ bool looping;
+ bool changed;
+};
+
+/* Worker to set_pure_flag. */
+
+static bool
+set_pure_flag_1 (cgraph_node *node, void *data)
+{
+ struct set_pure_flag_info *info = (struct set_pure_flag_info *)data;
+ /* Static constructors and destructors without a side effect can be
+ optimized out. */
+ if (info->pure && !info->looping)
+ {
+ if (DECL_STATIC_CONSTRUCTOR (node->decl))
+ {
+ DECL_STATIC_CONSTRUCTOR (node->decl) = 0;
+ info->changed = true;
+ }
+ if (DECL_STATIC_DESTRUCTOR (node->decl))
+ {
+ DECL_STATIC_DESTRUCTOR (node->decl) = 0;
+ info->changed = true;
+ }
+ }
+ if (info->pure)
+ {
+ if (!DECL_PURE_P (node->decl) && !TREE_READONLY (node->decl))
+ {
+ DECL_PURE_P (node->decl) = true;
+ DECL_LOOPING_CONST_OR_PURE_P (node->decl) = info->looping;
+ info->changed = true;
+ }
+ else if (DECL_LOOPING_CONST_OR_PURE_P (node->decl)
+ && !info->looping)
+ {
+ DECL_LOOPING_CONST_OR_PURE_P (node->decl) = false;
+ info->changed = true;
+ }
+ }
+ else
+ {
+ if (DECL_PURE_P (node->decl))
+ {
+ DECL_PURE_P (node->decl) = false;
+ DECL_LOOPING_CONST_OR_PURE_P (node->decl) = false;
+ info->changed = true;
+ }
+ }
+ return false;
+}
+
+/* Set DECL_PURE_P on cgraph_node's decl and on aliases of the node
+ if any to PURE.
+
+ When setting the flag, be careful about possible interposition.
+ Return true if any change was done. */
+
+bool
+cgraph_node::set_pure_flag (bool pure, bool looping)
+{
+ struct set_pure_flag_info info = {pure, looping, false};
+ call_for_symbol_thunks_and_aliases (set_pure_flag_1, &info, !pure, true);
+ return info.changed;
+}
+
+/* Return true when cgraph_node cannot return or throw and thus
+ it is safe to ignore its side effects for IPA analysis. */
+
+bool
+cgraph_node::cannot_return_p (void)
+{
+ int flags = flags_from_decl_or_type (decl);
+ if (!opt_for_fn (decl, flag_exceptions))
+ return (flags & ECF_NORETURN) != 0;
+ else
+ return ((flags & (ECF_NORETURN | ECF_NOTHROW))
+ == (ECF_NORETURN | ECF_NOTHROW));
+}
+
+/* Return true when call of edge cannot lead to return from caller
+ and thus it is safe to ignore its side effects for IPA analysis
+ when computing side effects of the caller.
+ FIXME: We could actually mark all edges that have no reaching
+ patch to the exit block or throw to get better results. */
+bool
+cgraph_edge::cannot_lead_to_return_p (void)
+{
+ if (caller->cannot_return_p ())
+ return true;
+ if (indirect_unknown_callee)
+ {
+ int flags = indirect_info->ecf_flags;
+ if (!opt_for_fn (caller->decl, flag_exceptions))
+ return (flags & ECF_NORETURN) != 0;
+ else
+ return ((flags & (ECF_NORETURN | ECF_NOTHROW))
+ == (ECF_NORETURN | ECF_NOTHROW));
+ }
+ else
+ return callee->cannot_return_p ();
+}
+
+/* Return true if the edge may be considered hot. */
+
+bool
+cgraph_edge::maybe_hot_p (void)
+{
+ if (!maybe_hot_count_p (NULL, count.ipa ()))
+ return false;
+ if (caller->frequency == NODE_FREQUENCY_UNLIKELY_EXECUTED
+ || (callee
+ && callee->frequency == NODE_FREQUENCY_UNLIKELY_EXECUTED))
+ return false;
+ if (caller->frequency > NODE_FREQUENCY_UNLIKELY_EXECUTED
+ && (callee
+ && callee->frequency <= NODE_FREQUENCY_EXECUTED_ONCE))
+ return false;
+ if (opt_for_fn (caller->decl, optimize_size))
+ return false;
+ if (caller->frequency == NODE_FREQUENCY_HOT)
+ return true;
+ if (!count.initialized_p ())
+ return true;
+ cgraph_node *where = caller->inlined_to ? caller->inlined_to : caller;
+ if (!where->count.initialized_p ())
+ return false;
+ if (caller->frequency == NODE_FREQUENCY_EXECUTED_ONCE)
+ {
+ if (count.apply_scale (2, 1) < where->count.apply_scale (3, 1))
+ return false;
+ }
+ else if (count.apply_scale (param_hot_bb_frequency_fraction , 1)
+ < where->count)
+ return false;
+ return true;
+}
+
+/* Worker for cgraph_can_remove_if_no_direct_calls_p. */
+
+static bool
+nonremovable_p (cgraph_node *node, void *)
+{
+ return !node->can_remove_if_no_direct_calls_and_refs_p ();
+}
+
+/* Return true if whole comdat group can be removed if there are no direct
+ calls to THIS. */
+
+bool
+cgraph_node::can_remove_if_no_direct_calls_p (bool will_inline)
+{
+ struct ipa_ref *ref;
+
+ /* For local symbols or non-comdat group it is the same as
+ can_remove_if_no_direct_calls_p. */
+ if (!externally_visible || !same_comdat_group)
+ {
+ if (DECL_EXTERNAL (decl))
+ return true;
+ if (address_taken)
+ return false;
+ return !call_for_symbol_and_aliases (nonremovable_p, NULL, true);
+ }
+
+ if (will_inline && address_taken)
+ return false;
+
+ /* Otherwise check if we can remove the symbol itself and then verify
+ that only uses of the comdat groups are direct call to THIS
+ or its aliases. */
+ if (!can_remove_if_no_direct_calls_and_refs_p ())
+ return false;
+
+ /* Check that all refs come from within the comdat group. */
+ for (int i = 0; iterate_referring (i, ref); i++)
+ if (ref->referring->get_comdat_group () != get_comdat_group ())
+ return false;
+
+ struct cgraph_node *target = ultimate_alias_target ();
+ for (cgraph_node *next = dyn_cast<cgraph_node *> (same_comdat_group);
+ next != this; next = dyn_cast<cgraph_node *> (next->same_comdat_group))
+ {
+ if (!externally_visible)
+ continue;
+ if (!next->alias
+ && !next->can_remove_if_no_direct_calls_and_refs_p ())
+ return false;
+
+ /* If we see different symbol than THIS, be sure to check calls. */
+ if (next->ultimate_alias_target () != target)
+ for (cgraph_edge *e = next->callers; e; e = e->next_caller)
+ if (e->caller->get_comdat_group () != get_comdat_group ()
+ || will_inline)
+ return false;
+
+ /* If function is not being inlined, we care only about
+ references outside of the comdat group. */
+ if (!will_inline)
+ for (int i = 0; next->iterate_referring (i, ref); i++)
+ if (ref->referring->get_comdat_group () != get_comdat_group ())
+ return false;
+ }
+ return true;
+}
+
+/* Return true when function cgraph_node can be expected to be removed
+ from program when direct calls in this compilation unit are removed.
+
+ As a special case COMDAT functions are
+ cgraph_can_remove_if_no_direct_calls_p while the are not
+ cgraph_only_called_directly_p (it is possible they are called from other
+ unit)
+
+ This function behaves as cgraph_only_called_directly_p because eliminating
+ all uses of COMDAT function does not make it necessarily disappear from
+ the program unless we are compiling whole program or we do LTO. In this
+ case we know we win since dynamic linking will not really discard the
+ linkonce section. */
+
+bool
+cgraph_node::will_be_removed_from_program_if_no_direct_calls_p
+ (bool will_inline)
+{
+ gcc_assert (!inlined_to);
+ if (DECL_EXTERNAL (decl))
+ return true;
+
+ if (!in_lto_p && !flag_whole_program)
+ {
+ /* If the symbol is in comdat group, we need to verify that whole comdat
+ group becomes unreachable. Technically we could skip references from
+ within the group, too. */
+ if (!only_called_directly_p ())
+ return false;
+ if (same_comdat_group && externally_visible)
+ {
+ struct cgraph_node *target = ultimate_alias_target ();
+
+ if (will_inline && address_taken)
+ return true;
+ for (cgraph_node *next = dyn_cast<cgraph_node *> (same_comdat_group);
+ next != this;
+ next = dyn_cast<cgraph_node *> (next->same_comdat_group))
+ {
+ if (!externally_visible)
+ continue;
+ if (!next->alias
+ && !next->only_called_directly_p ())
+ return false;
+
+ /* If we see different symbol than THIS,
+ be sure to check calls. */
+ if (next->ultimate_alias_target () != target)
+ for (cgraph_edge *e = next->callers; e; e = e->next_caller)
+ if (e->caller->get_comdat_group () != get_comdat_group ()
+ || will_inline)
+ return false;
+ }
+ }
+ return true;
+ }
+ else
+ return can_remove_if_no_direct_calls_p (will_inline);
+}
+
+
+/* Worker for cgraph_only_called_directly_p. */
+
+static bool
+cgraph_not_only_called_directly_p_1 (cgraph_node *node, void *)
+{
+ return !node->only_called_directly_or_aliased_p ();
+}
+
+/* Return true when function cgraph_node and all its aliases are only called
+ directly.
+ i.e. it is not externally visible, address was not taken and
+ it is not used in any other non-standard way. */
+
+bool
+cgraph_node::only_called_directly_p (void)
+{
+ gcc_assert (ultimate_alias_target () == this);
+ return !call_for_symbol_and_aliases (cgraph_not_only_called_directly_p_1,
+ NULL, true);
+}
+
+
+/* Collect all callers of NODE. Worker for collect_callers_of_node. */
+
+static bool
+collect_callers_of_node_1 (cgraph_node *node, void *data)
+{
+ vec<cgraph_edge *> *redirect_callers = (vec<cgraph_edge *> *)data;
+ cgraph_edge *cs;
+ enum availability avail;
+ node->ultimate_alias_target (&avail);
+
+ if (avail > AVAIL_INTERPOSABLE)
+ for (cs = node->callers; cs != NULL; cs = cs->next_caller)
+ if (!cs->indirect_inlining_edge
+ && !cs->caller->thunk)
+ redirect_callers->safe_push (cs);
+ return false;
+}
+
+/* Collect all callers of cgraph_node and its aliases that are known to lead to
+ cgraph_node (i.e. are not overwritable). */
+
+auto_vec<cgraph_edge *>
+cgraph_node::collect_callers (void)
+{
+ auto_vec<cgraph_edge *> redirect_callers;
+ call_for_symbol_thunks_and_aliases (collect_callers_of_node_1,
+ &redirect_callers, false);
+ return redirect_callers;
+}
+
+
+/* Return TRUE if NODE2 a clone of NODE or is equivalent to it. Return
+ optimistically true if this cannot be determined. */
+
+static bool
+clone_of_p (cgraph_node *node, cgraph_node *node2)
+{
+ node = node->ultimate_alias_target ();
+ node2 = node2->ultimate_alias_target ();
+
+ if (node2->clone_of == node
+ || node2->former_clone_of == node->decl)
+ return true;
+
+ if (!node->thunk && !node->former_thunk_p ())
+ {
+ while (node2
+ && node->decl != node2->decl
+ && node->decl != node2->former_clone_of)
+ node2 = node2->clone_of;
+ return node2 != NULL;
+ }
+
+ /* There are no virtual clones of thunks so check former_clone_of or if we
+ might have skipped thunks because this adjustments are no longer
+ necessary. */
+ while (node->thunk || node->former_thunk_p ())
+ {
+ if (!thunk_info::get (node)->this_adjusting)
+ return false;
+ /* In case of instrumented expanded thunks, which can have multiple calls
+ in them, we do not know how to continue and just have to be
+ optimistic. The same applies if all calls have already been inlined
+ into the thunk. */
+ if (!node->callees || node->callees->next_callee)
+ return true;
+ node = node->callees->callee->ultimate_alias_target ();
+
+ clone_info *info = clone_info::get (node2);
+ if (!info || !info->param_adjustments
+ || info->param_adjustments->first_param_intact_p ())
+ return false;
+ if (node2->former_clone_of == node->decl
+ || node2->former_clone_of == node->former_clone_of)
+ return true;
+
+ cgraph_node *n2 = node2;
+ while (n2 && node->decl != n2->decl)
+ n2 = n2->clone_of;
+ if (n2)
+ return true;
+ }
+
+ return false;
+}
+
+/* Verify edge count and frequency. */
+
+bool
+cgraph_edge::verify_count ()
+{
+ bool error_found = false;
+ if (!count.verify ())
+ {
+ error ("caller edge count invalid");
+ error_found = true;
+ }
+ return error_found;
+}
+
+/* Switch to THIS_CFUN if needed and print STMT to stderr. */
+static void
+cgraph_debug_gimple_stmt (function *this_cfun, gimple *stmt)
+{
+ bool fndecl_was_null = false;
+ /* debug_gimple_stmt needs correct cfun */
+ if (cfun != this_cfun)
+ set_cfun (this_cfun);
+ /* ...and an actual current_function_decl */
+ if (!current_function_decl)
+ {
+ current_function_decl = this_cfun->decl;
+ fndecl_was_null = true;
+ }
+ debug_gimple_stmt (stmt);
+ if (fndecl_was_null)
+ current_function_decl = NULL;
+}
+
+/* Verify that call graph edge corresponds to DECL from the associated
+ statement. Return true if the verification should fail. */
+
+bool
+cgraph_edge::verify_corresponds_to_fndecl (tree decl)
+{
+ cgraph_node *node;
+
+ if (!decl || callee->inlined_to)
+ return false;
+ if (symtab->state == LTO_STREAMING)
+ return false;
+ node = cgraph_node::get (decl);
+
+ /* We do not know if a node from a different partition is an alias or what it
+ aliases and therefore cannot do the former_clone_of check reliably. When
+ body_removed is set, we have lost all information about what was alias or
+ thunk of and also cannot proceed. */
+ if (!node
+ || node->body_removed
+ || node->in_other_partition
+ || callee->icf_merged
+ || callee->in_other_partition)
+ return false;
+
+ node = node->ultimate_alias_target ();
+
+ /* Optimizers can redirect unreachable calls or calls triggering undefined
+ behavior to builtin_unreachable. */
+
+ if (fndecl_built_in_p (callee->decl, BUILT_IN_UNREACHABLE))
+ return false;
+
+ if (callee->former_clone_of != node->decl
+ && (node != callee->ultimate_alias_target ())
+ && !clone_of_p (node, callee))
+ return true;
+ else
+ return false;
+}
+
+/* Disable warnings about missing quoting in GCC diagnostics for
+ the verification errors. Their format strings don't follow GCC
+ diagnostic conventions and the calls are ultimately followed by
+ one to internal_error. */
+#if __GNUC__ >= 10
+# pragma GCC diagnostic push
+# pragma GCC diagnostic ignored "-Wformat-diag"
+#endif
+
+/* Verify consistency of speculative call in NODE corresponding to STMT
+ and LTO_STMT_UID. If INDIRECT is set, assume that it is the indirect
+ edge of call sequence. Return true if error is found.
+
+ This function is called to every component of indirect call (direct edges,
+ indirect edge and refs). To save duplicated work, do full testing only
+ in that case. */
+static bool
+verify_speculative_call (struct cgraph_node *node, gimple *stmt,
+ unsigned int lto_stmt_uid,
+ struct cgraph_edge *indirect)
+{
+ if (indirect == NULL)
+ {
+ for (indirect = node->indirect_calls; indirect;
+ indirect = indirect->next_callee)
+ if (indirect->call_stmt == stmt
+ && indirect->lto_stmt_uid == lto_stmt_uid)
+ break;
+ if (!indirect)
+ {
+ error ("missing indirect call in speculative call sequence");
+ return true;
+ }
+ if (!indirect->speculative)
+ {
+ error ("indirect call in speculative call sequence has no "
+ "speculative flag");
+ return true;
+ }
+ return false;
+ }
+
+ /* Maximal number of targets. We probably will never want to have more than
+ this. */
+ const unsigned int num = 256;
+ cgraph_edge *direct_calls[num];
+ ipa_ref *refs[num];
+
+ for (unsigned int i = 0; i < num; i++)
+ {
+ direct_calls[i] = NULL;
+ refs[i] = NULL;
+ }
+
+ cgraph_edge *first_call = NULL;
+ cgraph_edge *prev_call = NULL;
+
+ for (cgraph_edge *direct = node->callees; direct;
+ direct = direct->next_callee)
+ if (direct->call_stmt == stmt && direct->lto_stmt_uid == lto_stmt_uid)
+ {
+ if (!first_call)
+ first_call = direct;
+ if (prev_call && direct != prev_call->next_callee)
+ {
+ error ("speculative edges are not adjacent");
+ return true;
+ }
+ prev_call = direct;
+ if (!direct->speculative)
+ {
+ error ("direct call to %s in speculative call sequence has no "
+ "speculative flag", direct->callee->dump_name ());
+ return true;
+ }
+ if (direct->speculative_id >= num)
+ {
+ error ("direct call to %s in speculative call sequence has "
+ "speculative_id %i out of range",
+ direct->callee->dump_name (), direct->speculative_id);
+ return true;
+ }
+ if (direct_calls[direct->speculative_id])
+ {
+ error ("duplicate direct call to %s in speculative call sequence "
+ "with speculative_id %i",
+ direct->callee->dump_name (), direct->speculative_id);
+ return true;
+ }
+ direct_calls[direct->speculative_id] = direct;
+ }
+
+ if (first_call->call_stmt
+ && first_call != node->get_edge (first_call->call_stmt))
+ {
+ error ("call stmt hash does not point to first direct edge of "
+ "speculative call sequence");
+ return true;
+ }
+
+ ipa_ref *ref;
+ for (int i = 0; node->iterate_reference (i, ref); i++)
+ if (ref->speculative
+ && ref->stmt == stmt && ref->lto_stmt_uid == lto_stmt_uid)
+ {
+ if (ref->speculative_id >= num)
+ {
+ error ("direct call to %s in speculative call sequence has "
+ "speculative_id %i out of range",
+ ref->referred->dump_name (), ref->speculative_id);
+ return true;
+ }
+ if (refs[ref->speculative_id])
+ {
+ error ("duplicate reference %s in speculative call sequence "
+ "with speculative_id %i",
+ ref->referred->dump_name (), ref->speculative_id);
+ return true;
+ }
+ refs[ref->speculative_id] = ref;
+ }
+
+ int num_targets = 0;
+ for (unsigned int i = 0 ; i < num ; i++)
+ {
+ if (refs[i] && !direct_calls[i])
+ {
+ error ("missing direct call for speculation %i", i);
+ return true;
+ }
+ if (!refs[i] && direct_calls[i])
+ {
+ error ("missing ref for speculation %i", i);
+ return true;
+ }
+ if (refs[i] != NULL)
+ num_targets++;
+ }
+
+ if (num_targets != indirect->num_speculative_call_targets_p ())
+ {
+ error ("number of speculative targets %i mismatched with "
+ "num_speculative_call_targets %i",
+ num_targets,
+ indirect->num_speculative_call_targets_p ());
+ return true;
+ }
+ return false;
+}
+
+/* Verify cgraph nodes of given cgraph node. */
+DEBUG_FUNCTION void
+cgraph_node::verify_node (void)
+{
+ cgraph_edge *e;
+ function *this_cfun = DECL_STRUCT_FUNCTION (decl);
+ basic_block this_block;
+ gimple_stmt_iterator gsi;
+ bool error_found = false;
+ int i;
+ ipa_ref *ref = NULL;
+
+ if (seen_error ())
+ return;
+
+ timevar_push (TV_CGRAPH_VERIFY);
+ error_found |= verify_base ();
+ for (e = callees; e; e = e->next_callee)
+ if (e->aux)
+ {
+ error ("aux field set for edge %s->%s",
+ identifier_to_locale (e->caller->name ()),
+ identifier_to_locale (e->callee->name ()));
+ error_found = true;
+ }
+ if (!count.verify ())
+ {
+ error ("cgraph count invalid");
+ error_found = true;
+ }
+ if (inlined_to && same_comdat_group)
+ {
+ error ("inline clone in same comdat group list");
+ error_found = true;
+ }
+ if (inlined_to && !count.compatible_p (inlined_to->count))
+ {
+ error ("inline clone count is not compatible");
+ count.debug ();
+ inlined_to->count.debug ();
+ error_found = true;
+ }
+ if (tp_first_run < 0)
+ {
+ error ("tp_first_run must be non-negative");
+ error_found = true;
+ }
+ if (!definition && !in_other_partition && local)
+ {
+ error ("local symbols must be defined");
+ error_found = true;
+ }
+ if (inlined_to && externally_visible)
+ {
+ error ("externally visible inline clone");
+ error_found = true;
+ }
+ if (inlined_to && address_taken)
+ {
+ error ("inline clone with address taken");
+ error_found = true;
+ }
+ if (inlined_to && force_output)
+ {
+ error ("inline clone is forced to output");
+ error_found = true;
+ }
+ if (symtab->state != LTO_STREAMING)
+ {
+ if (calls_comdat_local && !same_comdat_group)
+ {
+ error ("calls_comdat_local is set outside of a comdat group");
+ error_found = true;
+ }
+ if (!inlined_to && calls_comdat_local != check_calls_comdat_local_p ())
+ {
+ error ("invalid calls_comdat_local flag");
+ error_found = true;
+ }
+ }
+ if (DECL_IS_MALLOC (decl)
+ && !POINTER_TYPE_P (TREE_TYPE (TREE_TYPE (decl))))
+ {
+ error ("malloc attribute should be used for a function that "
+ "returns a pointer");
+ error_found = true;
+ }
+ if (definition && externally_visible
+ && semantic_interposition
+ != opt_for_fn (decl, flag_semantic_interposition))
+ {
+ error ("semantic interposition mismatch");
+ error_found = true;
+ }
+ for (e = indirect_calls; e; e = e->next_callee)
+ {
+ if (e->aux)
+ {
+ error ("aux field set for indirect edge from %s",
+ identifier_to_locale (e->caller->name ()));
+ error_found = true;
+ }
+ if (!e->count.compatible_p (count))
+ {
+ error ("edge count is not compatible with function count");
+ e->count.debug ();
+ count.debug ();
+ error_found = true;
+ }
+ if (!e->indirect_unknown_callee
+ || !e->indirect_info)
+ {
+ error ("An indirect edge from %s is not marked as indirect or has "
+ "associated indirect_info, the corresponding statement is: ",
+ identifier_to_locale (e->caller->name ()));
+ cgraph_debug_gimple_stmt (this_cfun, e->call_stmt);
+ error_found = true;
+ }
+ if (e->call_stmt && e->lto_stmt_uid)
+ {
+ error ("edge has both call_stmt and lto_stmt_uid set");
+ error_found = true;
+ }
+ }
+ bool check_comdat = comdat_local_p ();
+ for (e = callers; e; e = e->next_caller)
+ {
+ if (e->verify_count ())
+ error_found = true;
+ if (check_comdat
+ && !in_same_comdat_group_p (e->caller))
+ {
+ error ("comdat-local function called by %s outside its comdat",
+ identifier_to_locale (e->caller->name ()));
+ error_found = true;
+ }
+ if (!e->inline_failed)
+ {
+ if (inlined_to
+ != (e->caller->inlined_to
+ ? e->caller->inlined_to : e->caller))
+ {
+ error ("inlined_to pointer is wrong");
+ error_found = true;
+ }
+ if (callers->next_caller)
+ {
+ error ("multiple inline callers");
+ error_found = true;
+ }
+ }
+ else
+ if (inlined_to)
+ {
+ error ("inlined_to pointer set for noninline callers");
+ error_found = true;
+ }
+ }
+ for (e = callees; e; e = e->next_callee)
+ {
+ if (e->verify_count ())
+ error_found = true;
+ if (!e->count.compatible_p (count))
+ {
+ error ("edge count is not compatible with function count");
+ e->count.debug ();
+ count.debug ();
+ error_found = true;
+ }
+ if (gimple_has_body_p (e->caller->decl)
+ && !e->caller->inlined_to
+ && !e->speculative
+ /* Optimized out calls are redirected to __builtin_unreachable. */
+ && (e->count.nonzero_p ()
+ || ! e->callee->decl
+ || !fndecl_built_in_p (e->callee->decl, BUILT_IN_UNREACHABLE))
+ && count
+ == ENTRY_BLOCK_PTR_FOR_FN (DECL_STRUCT_FUNCTION (decl))->count
+ && (!e->count.ipa_p ()
+ && e->count.differs_from_p (gimple_bb (e->call_stmt)->count)))
+ {
+ error ("caller edge count does not match BB count");
+ fprintf (stderr, "edge count: ");
+ e->count.dump (stderr);
+ fprintf (stderr, "\n bb count: ");
+ gimple_bb (e->call_stmt)->count.dump (stderr);
+ fprintf (stderr, "\n");
+ error_found = true;
+ }
+ if (e->call_stmt && e->lto_stmt_uid)
+ {
+ error ("edge has both call_stmt and lto_stmt_uid set");
+ error_found = true;
+ }
+ if (e->speculative
+ && verify_speculative_call (e->caller, e->call_stmt, e->lto_stmt_uid,
+ NULL))
+ error_found = true;
+ }
+ for (e = indirect_calls; e; e = e->next_callee)
+ {
+ if (e->verify_count ())
+ error_found = true;
+ if (gimple_has_body_p (e->caller->decl)
+ && !e->caller->inlined_to
+ && !e->speculative
+ && e->count.ipa_p ()
+ && count
+ == ENTRY_BLOCK_PTR_FOR_FN (DECL_STRUCT_FUNCTION (decl))->count
+ && (!e->count.ipa_p ()
+ && e->count.differs_from_p (gimple_bb (e->call_stmt)->count)))
+ {
+ error ("indirect call count does not match BB count");
+ fprintf (stderr, "edge count: ");
+ e->count.dump (stderr);
+ fprintf (stderr, "\n bb count: ");
+ gimple_bb (e->call_stmt)->count.dump (stderr);
+ fprintf (stderr, "\n");
+ error_found = true;
+ }
+ if (e->speculative
+ && verify_speculative_call (e->caller, e->call_stmt, e->lto_stmt_uid,
+ e))
+ error_found = true;
+ }
+ for (i = 0; iterate_reference (i, ref); i++)
+ {
+ if (ref->stmt && ref->lto_stmt_uid)
+ {
+ error ("reference has both stmt and lto_stmt_uid set");
+ error_found = true;
+ }
+ if (ref->speculative
+ && verify_speculative_call (this, ref->stmt,
+ ref->lto_stmt_uid, NULL))
+ error_found = true;
+ }
+
+ if (!callers && inlined_to)
+ {
+ error ("inlined_to pointer is set but no predecessors found");
+ error_found = true;
+ }
+ if (inlined_to == this)
+ {
+ error ("inlined_to pointer refers to itself");
+ error_found = true;
+ }
+
+ if (clone_of)
+ {
+ cgraph_node *first_clone = clone_of->clones;
+ if (first_clone != this)
+ {
+ if (prev_sibling_clone->clone_of != clone_of)
+ {
+ error ("cgraph_node has wrong clone_of");
+ error_found = true;
+ }
+ }
+ }
+ if (clones)
+ {
+ cgraph_node *n;
+ for (n = clones; n; n = n->next_sibling_clone)
+ if (n->clone_of != this)
+ break;
+ if (n)
+ {
+ error ("cgraph_node has wrong clone list");
+ error_found = true;
+ }
+ }
+ if ((prev_sibling_clone || next_sibling_clone) && !clone_of)
+ {
+ error ("cgraph_node is in clone list but it is not clone");
+ error_found = true;
+ }
+ if (!prev_sibling_clone && clone_of && clone_of->clones != this)
+ {
+ error ("cgraph_node has wrong prev_clone pointer");
+ error_found = true;
+ }
+ if (prev_sibling_clone && prev_sibling_clone->next_sibling_clone != this)
+ {
+ error ("double linked list of clones corrupted");
+ error_found = true;
+ }
+
+ if (analyzed && alias)
+ {
+ bool ref_found = false;
+ int i;
+ ipa_ref *ref = NULL;
+
+ if (callees)
+ {
+ error ("Alias has call edges");
+ error_found = true;
+ }
+ for (i = 0; iterate_reference (i, ref); i++)
+ if (ref->use != IPA_REF_ALIAS)
+ {
+ error ("Alias has non-alias reference");
+ error_found = true;
+ }
+ else if (ref_found)
+ {
+ error ("Alias has more than one alias reference");
+ error_found = true;
+ }
+ else
+ ref_found = true;
+ if (!ref_found)
+ {
+ error ("Analyzed alias has no reference");
+ error_found = true;
+ }
+ }
+
+ if (analyzed && thunk)
+ {
+ if (!callees)
+ {
+ error ("No edge out of thunk node");
+ error_found = true;
+ }
+ else if (callees->next_callee)
+ {
+ error ("More than one edge out of thunk node");
+ error_found = true;
+ }
+ if (gimple_has_body_p (decl) && !inlined_to)
+ {
+ error ("Thunk is not supposed to have body");
+ error_found = true;
+ }
+ }
+ else if (analyzed && gimple_has_body_p (decl)
+ && !TREE_ASM_WRITTEN (decl)
+ && (!DECL_EXTERNAL (decl) || inlined_to)
+ && !flag_wpa)
+ {
+ if (this_cfun->cfg)
+ {
+ hash_set<gimple *> stmts;
+
+ /* Reach the trees by walking over the CFG, and note the
+ enclosing basic-blocks in the call edges. */
+ FOR_EACH_BB_FN (this_block, this_cfun)
+ {
+ for (gsi = gsi_start_phis (this_block);
+ !gsi_end_p (gsi); gsi_next (&gsi))
+ stmts.add (gsi_stmt (gsi));
+ for (gsi = gsi_start_bb (this_block);
+ !gsi_end_p (gsi);
+ gsi_next (&gsi))
+ {
+ gimple *stmt = gsi_stmt (gsi);
+ stmts.add (stmt);
+ if (is_gimple_call (stmt))
+ {
+ cgraph_edge *e = get_edge (stmt);
+ tree decl = gimple_call_fndecl (stmt);
+ if (e)
+ {
+ if (e->aux)
+ {
+ error ("shared call_stmt:");
+ cgraph_debug_gimple_stmt (this_cfun, stmt);
+ error_found = true;
+ }
+ if (!e->indirect_unknown_callee)
+ {
+ if (e->verify_corresponds_to_fndecl (decl))
+ {
+ error ("edge points to wrong declaration:");
+ debug_tree (e->callee->decl);
+ fprintf (stderr," Instead of:");
+ debug_tree (decl);
+ error_found = true;
+ }
+ }
+ else if (decl)
+ {
+ error ("an indirect edge with unknown callee "
+ "corresponding to a call_stmt with "
+ "a known declaration:");
+ error_found = true;
+ cgraph_debug_gimple_stmt (this_cfun, e->call_stmt);
+ }
+ e->aux = (void *)1;
+ }
+ else if (decl)
+ {
+ error ("missing callgraph edge for call stmt:");
+ cgraph_debug_gimple_stmt (this_cfun, stmt);
+ error_found = true;
+ }
+ }
+ }
+ }
+ for (i = 0; iterate_reference (i, ref); i++)
+ if (ref->stmt && !stmts.contains (ref->stmt))
+ {
+ error ("reference to dead statement");
+ cgraph_debug_gimple_stmt (this_cfun, ref->stmt);
+ error_found = true;
+ }
+ }
+ else
+ /* No CFG available?! */
+ gcc_unreachable ();
+
+ for (e = callees; e; e = e->next_callee)
+ {
+ if (!e->aux && !e->speculative)
+ {
+ error ("edge %s->%s has no corresponding call_stmt",
+ identifier_to_locale (e->caller->name ()),
+ identifier_to_locale (e->callee->name ()));
+ cgraph_debug_gimple_stmt (this_cfun, e->call_stmt);
+ error_found = true;
+ }
+ e->aux = 0;
+ }
+ for (e = indirect_calls; e; e = e->next_callee)
+ {
+ if (!e->aux && !e->speculative)
+ {
+ error ("an indirect edge from %s has no corresponding call_stmt",
+ identifier_to_locale (e->caller->name ()));
+ cgraph_debug_gimple_stmt (this_cfun, e->call_stmt);
+ error_found = true;
+ }
+ e->aux = 0;
+ }
+ }
+
+ if (nested_function_info *info = nested_function_info::get (this))
+ {
+ if (info->nested != NULL)
+ {
+ for (cgraph_node *n = info->nested; n != NULL;
+ n = next_nested_function (n))
+ {
+ nested_function_info *ninfo = nested_function_info::get (n);
+ if (ninfo->origin == NULL)
+ {
+ error ("missing origin for a node in a nested list");
+ error_found = true;
+ }
+ else if (ninfo->origin != this)
+ {
+ error ("origin points to a different parent");
+ error_found = true;
+ break;
+ }
+ }
+ }
+ if (info->next_nested != NULL && info->origin == NULL)
+ {
+ error ("missing origin for a node in a nested list");
+ error_found = true;
+ }
+ }
+
+ if (error_found)
+ {
+ dump (stderr);
+ internal_error ("verify_cgraph_node failed");
+ }
+ timevar_pop (TV_CGRAPH_VERIFY);
+}
+
+/* Verify whole cgraph structure. */
+DEBUG_FUNCTION void
+cgraph_node::verify_cgraph_nodes (void)
+{
+ cgraph_node *node;
+
+ if (seen_error ())
+ return;
+
+ FOR_EACH_FUNCTION (node)
+ node->verify ();
+}
+
+#if __GNUC__ >= 10
+# pragma GCC diagnostic pop
+#endif
+
+/* Walk the alias chain to return the function cgraph_node is alias of.
+ Walk through thunks, too.
+ When AVAILABILITY is non-NULL, get minimal availability in the chain.
+ When REF is non-NULL, assume that reference happens in symbol REF
+ when determining the availability. */
+
+cgraph_node *
+cgraph_node::function_symbol (enum availability *availability,
+ struct symtab_node *ref)
+{
+ cgraph_node *node = ultimate_alias_target (availability, ref);
+
+ while (node->thunk)
+ {
+ enum availability a;
+
+ ref = node;
+ node = node->callees->callee;
+ node = node->ultimate_alias_target (availability ? &a : NULL, ref);
+ if (availability && a < *availability)
+ *availability = a;
+ }
+ return node;
+}
+
+/* Walk the alias chain to return the function cgraph_node is alias of.
+ Walk through non virtual thunks, too. Thus we return either a function
+ or a virtual thunk node.
+ When AVAILABILITY is non-NULL, get minimal availability in the chain.
+ When REF is non-NULL, assume that reference happens in symbol REF
+ when determining the availability. */
+
+cgraph_node *
+cgraph_node::function_or_virtual_thunk_symbol
+ (enum availability *availability,
+ struct symtab_node *ref)
+{
+ cgraph_node *node = ultimate_alias_target (availability, ref);
+
+ while (node->thunk && !thunk_info::get (node)->virtual_offset_p)
+ {
+ enum availability a;
+
+ ref = node;
+ node = node->callees->callee;
+ node = node->ultimate_alias_target (availability ? &a : NULL, ref);
+ if (availability && a < *availability)
+ *availability = a;
+ }
+ return node;
+}
+
+/* When doing LTO, read cgraph_node's body from disk if it is not already
+ present. Also perform any necessary clone materializations. */
+
+bool
+cgraph_node::get_untransformed_body ()
+{
+ lto_file_decl_data *file_data;
+ const char *data, *name;
+ size_t len;
+ tree decl = this->decl;
+
+ /* See if there is clone to be materialized.
+ (inline clones does not need materialization, but we can be seeing
+ an inline clone of real clone). */
+ cgraph_node *p = this;
+ for (cgraph_node *c = clone_of; c; c = c->clone_of)
+ {
+ if (c->decl != decl)
+ p->materialize_clone ();
+ p = c;
+ }
+
+ /* Check if body is already there. Either we have gimple body or
+ the function is thunk and in that case we set DECL_ARGUMENTS. */
+ if (DECL_ARGUMENTS (decl) || gimple_has_body_p (decl))
+ return false;
+
+ gcc_assert (in_lto_p && !DECL_RESULT (decl));
+
+ timevar_push (TV_IPA_LTO_GIMPLE_IN);
+
+ file_data = lto_file_data;
+ name = IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (decl));
+
+ /* We may have renamed the declaration, e.g., a static function. */
+ name = lto_get_decl_name_mapping (file_data, name);
+ struct lto_in_decl_state *decl_state
+ = lto_get_function_in_decl_state (file_data, decl);
+
+ cgraph_node *origin = this;
+ while (origin->clone_of)
+ origin = origin->clone_of;
+
+ int stream_order = origin->order - file_data->order_base;
+ data = lto_get_section_data (file_data, LTO_section_function_body,
+ name, stream_order, &len,
+ decl_state->compressed);
+ if (!data)
+ fatal_error (input_location, "%s: section %s.%d is missing",
+ file_data->file_name, name, stream_order);
+
+ gcc_assert (DECL_STRUCT_FUNCTION (decl) == NULL);
+
+ if (!quiet_flag)
+ fprintf (stderr, " in:%s", IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (decl)));
+ lto_input_function_body (file_data, this, data);
+ lto_stats.num_function_bodies++;
+ lto_free_section_data (file_data, LTO_section_function_body, name,
+ data, len, decl_state->compressed);
+ lto_free_function_in_decl_state_for_node (this);
+ /* Keep lto file data so ipa-inline-analysis knows about cross module
+ inlining. */
+
+ timevar_pop (TV_IPA_LTO_GIMPLE_IN);
+
+ return true;
+}
+
+/* Prepare function body. When doing LTO, read cgraph_node's body from disk
+ if it is not already present. When some IPA transformations are scheduled,
+ apply them. */
+
+bool
+cgraph_node::get_body (void)
+{
+ bool updated;
+
+ updated = get_untransformed_body ();
+
+ /* Getting transformed body makes no sense for inline clones;
+ we should never use this on real clones because they are materialized
+ early.
+ TODO: Materializing clones here will likely lead to smaller LTRANS
+ footprint. */
+ gcc_assert (!inlined_to && !clone_of);
+ if (ipa_transforms_to_apply.exists ())
+ {
+ opt_pass *saved_current_pass = current_pass;
+ FILE *saved_dump_file = dump_file;
+ const char *saved_dump_file_name = dump_file_name;
+ dump_flags_t saved_dump_flags = dump_flags;
+ dump_file_name = NULL;
+ set_dump_file (NULL);
+
+ push_cfun (DECL_STRUCT_FUNCTION (decl));
+
+ update_ssa (TODO_update_ssa_only_virtuals);
+ execute_all_ipa_transforms (true);
+ cgraph_edge::rebuild_edges ();
+ free_dominance_info (CDI_DOMINATORS);
+ free_dominance_info (CDI_POST_DOMINATORS);
+ pop_cfun ();
+ updated = true;
+
+ current_pass = saved_current_pass;
+ set_dump_file (saved_dump_file);
+ dump_file_name = saved_dump_file_name;
+ dump_flags = saved_dump_flags;
+ }
+ return updated;
+}
+
+/* Return the DECL_STRUCT_FUNCTION of the function. */
+
+struct function *
+cgraph_node::get_fun () const
+{
+ const cgraph_node *node = this;
+ struct function *fun = DECL_STRUCT_FUNCTION (node->decl);
+
+ while (!fun && node->clone_of)
+ {
+ node = node->clone_of;
+ fun = DECL_STRUCT_FUNCTION (node->decl);
+ }
+
+ return fun;
+}
+
+/* Reset all state within cgraph.c so that we can rerun the compiler
+ within the same process. For use by toplev::finalize. */
+
+void
+cgraph_c_finalize (void)
+{
+ nested_function_info::release ();
+ thunk_info::release ();
+ clone_info::release ();
+ symtab = NULL;
+
+ x_cgraph_nodes_queue = NULL;
+
+ cgraph_fnver_htab = NULL;
+ version_info_node = NULL;
+}
+
+/* A worker for call_for_symbol_and_aliases. */
+
+bool
+cgraph_node::call_for_symbol_and_aliases_1 (bool (*callback) (cgraph_node *,
+ void *),
+ void *data,
+ bool include_overwritable)
+{
+ ipa_ref *ref;
+ FOR_EACH_ALIAS (this, ref)
+ {
+ cgraph_node *alias = dyn_cast <cgraph_node *> (ref->referring);
+ if (include_overwritable
+ || alias->get_availability () > AVAIL_INTERPOSABLE)
+ if (alias->call_for_symbol_and_aliases (callback, data,
+ include_overwritable))
+ return true;
+ }
+ return false;
+}
+
+/* Return true if NODE has thunk. */
+
+bool
+cgraph_node::has_thunk_p (cgraph_node *node, void *)
+{
+ for (cgraph_edge *e = node->callers; e; e = e->next_caller)
+ if (e->caller->thunk)
+ return true;
+ return false;
+}
+
+/* Expected frequency of executions within the function. */
+
+sreal
+cgraph_edge::sreal_frequency ()
+{
+ return count.to_sreal_scale (caller->inlined_to
+ ? caller->inlined_to->count
+ : caller->count);
+}
+
+
+/* During LTO stream in this can be used to check whether call can possibly
+ be internal to the current translation unit. */
+
+bool
+cgraph_edge::possibly_call_in_translation_unit_p (void)
+{
+ gcc_checking_assert (in_lto_p && caller->prevailing_p ());
+
+ /* While incremental linking we may end up getting function body later. */
+ if (flag_incremental_link == INCREMENTAL_LINK_LTO)
+ return true;
+
+ /* We may be smarter here and avoid streaming in indirect calls we can't
+ track, but that would require arranging streaming the indirect call
+ summary first. */
+ if (!callee)
+ return true;
+
+ /* If callee is local to the original translation unit, it will be
+ defined. */
+ if (!TREE_PUBLIC (callee->decl) && !DECL_EXTERNAL (callee->decl))
+ return true;
+
+ /* Otherwise we need to lookup prevailing symbol (symbol table is not merged,
+ yet) and see if it is a definition. In fact we may also resolve aliases,
+ but that is probably not too important. */
+ symtab_node *node = callee;
+ for (int n = 10; node->previous_sharing_asm_name && n ; n--)
+ node = node->previous_sharing_asm_name;
+ if (node->previous_sharing_asm_name)
+ node = symtab_node::get_for_asmname (DECL_ASSEMBLER_NAME (callee->decl));
+ gcc_assert (TREE_PUBLIC (node->decl));
+ return node->get_availability () >= AVAIL_INTERPOSABLE;
+}
+
+/* Return num_speculative_targets of this edge. */
+
+int
+cgraph_edge::num_speculative_call_targets_p (void)
+{
+ return indirect_info ? indirect_info->num_speculative_call_targets : 0;
+}
+
+/* Check if function calls comdat local. This is used to recompute
+ calls_comdat_local flag after function transformations. */
+bool
+cgraph_node::check_calls_comdat_local_p ()
+{
+ for (cgraph_edge *e = callees; e; e = e->next_callee)
+ if (e->inline_failed
+ ? e->callee->comdat_local_p ()
+ : e->callee->check_calls_comdat_local_p ())
+ return true;
+ return false;
+}
+
+/* Return true if this node represents a former, i.e. an expanded, thunk. */
+
+bool
+cgraph_node::former_thunk_p (void)
+{
+ if (thunk)
+ return false;
+ thunk_info *i = thunk_info::get (this);
+ if (!i)
+ return false;
+ gcc_checking_assert (i->fixed_offset || i->virtual_offset_p
+ || i->indirect_offset);
+ return true;
+}
+
+/* A stashed copy of "symtab" for use by selftest::symbol_table_test.
+ This needs to be a global so that it can be a GC root, and thus
+ prevent the stashed copy from being garbage-collected if the GC runs
+ during a symbol_table_test. */
+
+symbol_table *saved_symtab;
+
+#if CHECKING_P
+
+namespace selftest {
+
+/* class selftest::symbol_table_test. */
+
+/* Constructor. Store the old value of symtab, and create a new one. */
+
+symbol_table_test::symbol_table_test ()
+{
+ gcc_assert (saved_symtab == NULL);
+ saved_symtab = symtab;
+ symtab = new (ggc_alloc<symbol_table> ()) symbol_table ();
+}
+
+/* Destructor. Restore the old value of symtab. */
+
+symbol_table_test::~symbol_table_test ()
+{
+ gcc_assert (saved_symtab != NULL);
+ symtab = saved_symtab;
+ saved_symtab = NULL;
+}
+
+/* Verify that symbol_table_test works. */
+
+static void
+test_symbol_table_test ()
+{
+ /* Simulate running two selftests involving symbol tables. */
+ for (int i = 0; i < 2; i++)
+ {
+ symbol_table_test stt;
+ tree test_decl = build_decl (UNKNOWN_LOCATION, FUNCTION_DECL,
+ get_identifier ("test_decl"),
+ build_function_type_list (void_type_node,
+ NULL_TREE));
+ cgraph_node *node = cgraph_node::get_create (test_decl);
+ gcc_assert (node);
+
+ /* Verify that the node has order 0 on both iterations,
+ and thus that nodes have predictable dump names in selftests. */
+ ASSERT_EQ (node->order, 0);
+ ASSERT_STREQ (node->dump_name (), "test_decl/0");
+ }
+}
+
+/* Run all of the selftests within this file. */
+
+void
+cgraph_c_tests ()
+{
+ test_symbol_table_test ();
+}
+
+} // namespace selftest
+
+#endif /* CHECKING_P */
+
+#include "gt-cgraph.h"