mirror of
https://github.com/Stichting-MINIX-Research-Foundation/netbsd.git
synced 2025-09-11 16:15:07 -04:00
3420 lines
98 KiB
C
3420 lines
98 KiB
C
/* Top-level LTO routines.
|
|
Copyright (C) 2009-2013 Free Software Foundation, Inc.
|
|
Contributed by CodeSourcery, Inc.
|
|
|
|
This file is part of GCC.
|
|
|
|
GCC is free software; you can redistribute it and/or modify it under
|
|
the terms of the GNU General Public License as published by the Free
|
|
Software Foundation; either version 3, or (at your option) any later
|
|
version.
|
|
|
|
GCC is distributed in the hope that it will be useful, but WITHOUT ANY
|
|
WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with GCC; see the file COPYING3. If not see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
#include "config.h"
|
|
#include "system.h"
|
|
#include "coretypes.h"
|
|
#include "opts.h"
|
|
#include "toplev.h"
|
|
#include "tree.h"
|
|
#include "tree-flow.h"
|
|
#include "diagnostic-core.h"
|
|
#include "tm.h"
|
|
#include "cgraph.h"
|
|
#include "ggc.h"
|
|
#include "tree-ssa-operands.h"
|
|
#include "tree-pass.h"
|
|
#include "langhooks.h"
|
|
#include "vec.h"
|
|
#include "bitmap.h"
|
|
#include "pointer-set.h"
|
|
#include "ipa-prop.h"
|
|
#include "common.h"
|
|
#include "debug.h"
|
|
#include "gimple.h"
|
|
#include "lto.h"
|
|
#include "lto-tree.h"
|
|
#include "lto-streamer.h"
|
|
#include "tree-streamer.h"
|
|
#include "splay-tree.h"
|
|
#include "lto-partition.h"
|
|
|
|
static GTY(()) tree first_personality_decl;
|
|
|
|
/* Returns a hash code for P. */
|
|
|
|
static hashval_t
|
|
hash_name (const void *p)
|
|
{
|
|
const struct lto_section_slot *ds = (const struct lto_section_slot *) p;
|
|
return (hashval_t) htab_hash_string (ds->name);
|
|
}
|
|
|
|
|
|
/* Returns nonzero if P1 and P2 are equal. */
|
|
|
|
static int
|
|
eq_name (const void *p1, const void *p2)
|
|
{
|
|
const struct lto_section_slot *s1 =
|
|
(const struct lto_section_slot *) p1;
|
|
const struct lto_section_slot *s2 =
|
|
(const struct lto_section_slot *) p2;
|
|
|
|
return strcmp (s1->name, s2->name) == 0;
|
|
}
|
|
|
|
/* Free lto_section_slot */
|
|
|
|
static void
|
|
free_with_string (void *arg)
|
|
{
|
|
struct lto_section_slot *s = (struct lto_section_slot *)arg;
|
|
|
|
free (CONST_CAST (char *, s->name));
|
|
free (arg);
|
|
}
|
|
|
|
/* Create section hash table */
|
|
|
|
htab_t
|
|
lto_obj_create_section_hash_table (void)
|
|
{
|
|
return htab_create (37, hash_name, eq_name, free_with_string);
|
|
}
|
|
|
|
/* Delete an allocated integer KEY in the splay tree. */
|
|
|
|
static void
|
|
lto_splay_tree_delete_id (splay_tree_key key)
|
|
{
|
|
free ((void *) key);
|
|
}
|
|
|
|
/* Compare splay tree node ids A and B. */
|
|
|
|
static int
|
|
lto_splay_tree_compare_ids (splay_tree_key a, splay_tree_key b)
|
|
{
|
|
unsigned HOST_WIDE_INT ai;
|
|
unsigned HOST_WIDE_INT bi;
|
|
|
|
ai = *(unsigned HOST_WIDE_INT *) a;
|
|
bi = *(unsigned HOST_WIDE_INT *) b;
|
|
|
|
if (ai < bi)
|
|
return -1;
|
|
else if (ai > bi)
|
|
return 1;
|
|
return 0;
|
|
}
|
|
|
|
/* Look up splay tree node by ID in splay tree T. */
|
|
|
|
static splay_tree_node
|
|
lto_splay_tree_lookup (splay_tree t, unsigned HOST_WIDE_INT id)
|
|
{
|
|
return splay_tree_lookup (t, (splay_tree_key) &id);
|
|
}
|
|
|
|
/* Check if KEY has ID. */
|
|
|
|
static bool
|
|
lto_splay_tree_id_equal_p (splay_tree_key key, unsigned HOST_WIDE_INT id)
|
|
{
|
|
return *(unsigned HOST_WIDE_INT *) key == id;
|
|
}
|
|
|
|
/* Insert a splay tree node into tree T with ID as key and FILE_DATA as value.
|
|
The ID is allocated separately because we need HOST_WIDE_INTs which may
|
|
be wider than a splay_tree_key. */
|
|
|
|
static void
|
|
lto_splay_tree_insert (splay_tree t, unsigned HOST_WIDE_INT id,
|
|
struct lto_file_decl_data *file_data)
|
|
{
|
|
unsigned HOST_WIDE_INT *idp = XCNEW (unsigned HOST_WIDE_INT);
|
|
*idp = id;
|
|
splay_tree_insert (t, (splay_tree_key) idp, (splay_tree_value) file_data);
|
|
}
|
|
|
|
/* Create a splay tree. */
|
|
|
|
static splay_tree
|
|
lto_splay_tree_new (void)
|
|
{
|
|
return splay_tree_new (lto_splay_tree_compare_ids,
|
|
lto_splay_tree_delete_id,
|
|
NULL);
|
|
}
|
|
|
|
/* Return true when NODE has a clone that is analyzed (i.e. we need
|
|
to load its body even if the node itself is not needed). */
|
|
|
|
static bool
|
|
has_analyzed_clone_p (struct cgraph_node *node)
|
|
{
|
|
struct cgraph_node *orig = node;
|
|
node = node->clones;
|
|
if (node)
|
|
while (node != orig)
|
|
{
|
|
if (node->analyzed)
|
|
return true;
|
|
if (node->clones)
|
|
node = node->clones;
|
|
else if (node->next_sibling_clone)
|
|
node = node->next_sibling_clone;
|
|
else
|
|
{
|
|
while (node != orig && !node->next_sibling_clone)
|
|
node = node->clone_of;
|
|
if (node != orig)
|
|
node = node->next_sibling_clone;
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
|
|
/* Read the function body for the function associated with NODE. */
|
|
|
|
static void
|
|
lto_materialize_function (struct cgraph_node *node)
|
|
{
|
|
tree decl;
|
|
struct lto_file_decl_data *file_data;
|
|
const char *data, *name;
|
|
size_t len;
|
|
|
|
decl = node->symbol.decl;
|
|
/* Read in functions with body (analyzed nodes)
|
|
and also functions that are needed to produce virtual clones. */
|
|
if (cgraph_function_with_gimple_body_p (node) || has_analyzed_clone_p (node))
|
|
{
|
|
/* Clones don't need to be read. */
|
|
if (node->clone_of)
|
|
return;
|
|
|
|
/* Load the function body only if not operating in WPA mode. In
|
|
WPA mode, the body of the function is not needed. */
|
|
if (!flag_wpa)
|
|
{
|
|
file_data = node->symbol.lto_file_data;
|
|
name = IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (decl));
|
|
|
|
/* We may have renamed the declaration, e.g., a static function. */
|
|
name = lto_get_decl_name_mapping (file_data, name);
|
|
|
|
data = lto_get_section_data (file_data, LTO_section_function_body,
|
|
name, &len);
|
|
if (!data)
|
|
fatal_error ("%s: section %s is missing",
|
|
file_data->file_name,
|
|
name);
|
|
|
|
gcc_assert (DECL_STRUCT_FUNCTION (decl) == NULL);
|
|
|
|
push_struct_function (decl);
|
|
announce_function (decl);
|
|
lto_input_function_body (file_data, decl, data);
|
|
if (DECL_FUNCTION_PERSONALITY (decl) && !first_personality_decl)
|
|
first_personality_decl = DECL_FUNCTION_PERSONALITY (decl);
|
|
lto_stats.num_function_bodies++;
|
|
lto_free_section_data (file_data, LTO_section_function_body, name,
|
|
data, len);
|
|
pop_cfun ();
|
|
ggc_collect ();
|
|
}
|
|
}
|
|
|
|
/* Let the middle end know about the function. */
|
|
rest_of_decl_compilation (decl, 1, 0);
|
|
}
|
|
|
|
|
|
/* Decode the content of memory pointed to by DATA in the in decl
|
|
state object STATE. DATA_IN points to a data_in structure for
|
|
decoding. Return the address after the decoded object in the
|
|
input. */
|
|
|
|
static const uint32_t *
|
|
lto_read_in_decl_state (struct data_in *data_in, const uint32_t *data,
|
|
struct lto_in_decl_state *state)
|
|
{
|
|
uint32_t ix;
|
|
tree decl;
|
|
uint32_t i, j;
|
|
|
|
ix = *data++;
|
|
decl = streamer_tree_cache_get (data_in->reader_cache, ix);
|
|
if (TREE_CODE (decl) != FUNCTION_DECL)
|
|
{
|
|
gcc_assert (decl == void_type_node);
|
|
decl = NULL_TREE;
|
|
}
|
|
state->fn_decl = decl;
|
|
|
|
for (i = 0; i < LTO_N_DECL_STREAMS; i++)
|
|
{
|
|
uint32_t size = *data++;
|
|
tree *decls = ggc_alloc_vec_tree (size);
|
|
|
|
for (j = 0; j < size; j++)
|
|
decls[j] = streamer_tree_cache_get (data_in->reader_cache, data[j]);
|
|
|
|
state->streams[i].size = size;
|
|
state->streams[i].trees = decls;
|
|
data += size;
|
|
}
|
|
|
|
return data;
|
|
}
|
|
|
|
|
|
|
|
/* Global type table. FIXME, it should be possible to re-use some
|
|
of the type hashing routines in tree.c (type_hash_canon, type_hash_lookup,
|
|
etc), but those assume that types were built with the various
|
|
build_*_type routines which is not the case with the streamer. */
|
|
static GTY((if_marked ("ggc_marked_p"), param_is (union tree_node)))
|
|
htab_t gimple_types;
|
|
static GTY((if_marked ("tree_int_map_marked_p"), param_is (struct tree_int_map)))
|
|
htab_t type_hash_cache;
|
|
|
|
static hashval_t gimple_type_hash (const void *);
|
|
|
|
/* Structure used to maintain a cache of some type pairs compared by
|
|
gimple_types_compatible_p when comparing aggregate types. There are
|
|
three possible values for SAME_P:
|
|
|
|
-2: The pair (T1, T2) has just been inserted in the table.
|
|
0: T1 and T2 are different types.
|
|
1: T1 and T2 are the same type. */
|
|
|
|
struct type_pair_d
|
|
{
|
|
unsigned int uid1;
|
|
unsigned int uid2;
|
|
signed char same_p;
|
|
};
|
|
typedef struct type_pair_d *type_pair_t;
|
|
|
|
#define GIMPLE_TYPE_PAIR_SIZE 16381
|
|
struct type_pair_d *type_pair_cache;
|
|
|
|
|
|
/* Lookup the pair of types T1 and T2 in *VISITED_P. Insert a new
|
|
entry if none existed. */
|
|
|
|
static inline type_pair_t
|
|
lookup_type_pair (tree t1, tree t2)
|
|
{
|
|
unsigned int index;
|
|
unsigned int uid1, uid2;
|
|
|
|
if (TYPE_UID (t1) < TYPE_UID (t2))
|
|
{
|
|
uid1 = TYPE_UID (t1);
|
|
uid2 = TYPE_UID (t2);
|
|
}
|
|
else
|
|
{
|
|
uid1 = TYPE_UID (t2);
|
|
uid2 = TYPE_UID (t1);
|
|
}
|
|
gcc_checking_assert (uid1 != uid2);
|
|
|
|
/* iterative_hash_hashval_t imply an function calls.
|
|
We know that UIDS are in limited range. */
|
|
index = ((((unsigned HOST_WIDE_INT)uid1 << HOST_BITS_PER_WIDE_INT / 2) + uid2)
|
|
% GIMPLE_TYPE_PAIR_SIZE);
|
|
if (type_pair_cache [index].uid1 == uid1
|
|
&& type_pair_cache [index].uid2 == uid2)
|
|
return &type_pair_cache[index];
|
|
|
|
type_pair_cache [index].uid1 = uid1;
|
|
type_pair_cache [index].uid2 = uid2;
|
|
type_pair_cache [index].same_p = -2;
|
|
|
|
return &type_pair_cache[index];
|
|
}
|
|
|
|
/* Per pointer state for the SCC finding. The on_sccstack flag
|
|
is not strictly required, it is true when there is no hash value
|
|
recorded for the type and false otherwise. But querying that
|
|
is slower. */
|
|
|
|
struct sccs
|
|
{
|
|
unsigned int dfsnum;
|
|
unsigned int low;
|
|
bool on_sccstack;
|
|
union {
|
|
hashval_t hash;
|
|
signed char same_p;
|
|
} u;
|
|
};
|
|
|
|
static unsigned int next_dfs_num;
|
|
static unsigned int gtc_next_dfs_num;
|
|
|
|
/* GIMPLE type merging cache. A direct-mapped cache based on TYPE_UID. */
|
|
|
|
typedef struct GTY(()) gimple_type_leader_entry_s {
|
|
tree type;
|
|
tree leader;
|
|
} gimple_type_leader_entry;
|
|
|
|
#define GIMPLE_TYPE_LEADER_SIZE 16381
|
|
static GTY((length("GIMPLE_TYPE_LEADER_SIZE")))
|
|
gimple_type_leader_entry *gimple_type_leader;
|
|
|
|
/* Lookup an existing leader for T and return it or NULL_TREE, if
|
|
there is none in the cache. */
|
|
|
|
static inline tree
|
|
gimple_lookup_type_leader (tree t)
|
|
{
|
|
gimple_type_leader_entry *leader;
|
|
|
|
leader = &gimple_type_leader[TYPE_UID (t) % GIMPLE_TYPE_LEADER_SIZE];
|
|
if (leader->type != t)
|
|
return NULL_TREE;
|
|
|
|
return leader->leader;
|
|
}
|
|
|
|
|
|
/* Return true if T1 and T2 have the same name. If FOR_COMPLETION_P is
|
|
true then if any type has no name return false, otherwise return
|
|
true if both types have no names. */
|
|
|
|
static bool
|
|
compare_type_names_p (tree t1, tree t2)
|
|
{
|
|
tree name1 = TYPE_NAME (t1);
|
|
tree name2 = TYPE_NAME (t2);
|
|
|
|
if ((name1 != NULL_TREE) != (name2 != NULL_TREE))
|
|
return false;
|
|
|
|
if (name1 == NULL_TREE)
|
|
return true;
|
|
|
|
/* Either both should be a TYPE_DECL or both an IDENTIFIER_NODE. */
|
|
if (TREE_CODE (name1) != TREE_CODE (name2))
|
|
return false;
|
|
|
|
if (TREE_CODE (name1) == TYPE_DECL)
|
|
name1 = DECL_NAME (name1);
|
|
gcc_checking_assert (!name1 || TREE_CODE (name1) == IDENTIFIER_NODE);
|
|
|
|
if (TREE_CODE (name2) == TYPE_DECL)
|
|
name2 = DECL_NAME (name2);
|
|
gcc_checking_assert (!name2 || TREE_CODE (name2) == IDENTIFIER_NODE);
|
|
|
|
/* Identifiers can be compared with pointer equality rather
|
|
than a string comparison. */
|
|
if (name1 == name2)
|
|
return true;
|
|
|
|
return false;
|
|
}
|
|
|
|
static bool
|
|
gimple_types_compatible_p_1 (tree, tree, type_pair_t,
|
|
vec<type_pair_t> *,
|
|
struct pointer_map_t *, struct obstack *);
|
|
|
|
/* DFS visit the edge from the callers type pair with state *STATE to
|
|
the pair T1, T2 while operating in FOR_MERGING_P mode.
|
|
Update the merging status if it is not part of the SCC containing the
|
|
callers pair and return it.
|
|
SCCSTACK, SCCSTATE and SCCSTATE_OBSTACK are state for the DFS walk done. */
|
|
|
|
static bool
|
|
gtc_visit (tree t1, tree t2,
|
|
struct sccs *state,
|
|
vec<type_pair_t> *sccstack,
|
|
struct pointer_map_t *sccstate,
|
|
struct obstack *sccstate_obstack)
|
|
{
|
|
struct sccs *cstate = NULL;
|
|
type_pair_t p;
|
|
void **slot;
|
|
tree leader1, leader2;
|
|
|
|
/* Check first for the obvious case of pointer identity. */
|
|
if (t1 == t2)
|
|
return true;
|
|
|
|
/* Check that we have two types to compare. */
|
|
if (t1 == NULL_TREE || t2 == NULL_TREE)
|
|
return false;
|
|
|
|
/* Can't be the same type if the types don't have the same code. */
|
|
if (TREE_CODE (t1) != TREE_CODE (t2))
|
|
return false;
|
|
|
|
/* Can't be the same type if they have different CV qualifiers. */
|
|
if (TYPE_QUALS (t1) != TYPE_QUALS (t2))
|
|
return false;
|
|
|
|
if (TREE_ADDRESSABLE (t1) != TREE_ADDRESSABLE (t2))
|
|
return false;
|
|
|
|
/* Void types and nullptr types are always the same. */
|
|
if (TREE_CODE (t1) == VOID_TYPE
|
|
|| TREE_CODE (t1) == NULLPTR_TYPE)
|
|
return true;
|
|
|
|
/* Can't be the same type if they have different alignment or mode. */
|
|
if (TYPE_ALIGN (t1) != TYPE_ALIGN (t2)
|
|
|| TYPE_MODE (t1) != TYPE_MODE (t2))
|
|
return false;
|
|
|
|
/* Do some simple checks before doing three hashtable queries. */
|
|
if (INTEGRAL_TYPE_P (t1)
|
|
|| SCALAR_FLOAT_TYPE_P (t1)
|
|
|| FIXED_POINT_TYPE_P (t1)
|
|
|| TREE_CODE (t1) == VECTOR_TYPE
|
|
|| TREE_CODE (t1) == COMPLEX_TYPE
|
|
|| TREE_CODE (t1) == OFFSET_TYPE
|
|
|| POINTER_TYPE_P (t1))
|
|
{
|
|
/* Can't be the same type if they have different sign or precision. */
|
|
if (TYPE_PRECISION (t1) != TYPE_PRECISION (t2)
|
|
|| TYPE_UNSIGNED (t1) != TYPE_UNSIGNED (t2))
|
|
return false;
|
|
|
|
if (TREE_CODE (t1) == INTEGER_TYPE
|
|
&& TYPE_STRING_FLAG (t1) != TYPE_STRING_FLAG (t2))
|
|
return false;
|
|
|
|
/* That's all we need to check for float and fixed-point types. */
|
|
if (SCALAR_FLOAT_TYPE_P (t1)
|
|
|| FIXED_POINT_TYPE_P (t1))
|
|
return true;
|
|
|
|
/* For other types fall through to more complex checks. */
|
|
}
|
|
|
|
/* If the types have been previously registered and found equal
|
|
they still are. */
|
|
leader1 = gimple_lookup_type_leader (t1);
|
|
leader2 = gimple_lookup_type_leader (t2);
|
|
if (leader1 == t2
|
|
|| t1 == leader2
|
|
|| (leader1 && leader1 == leader2))
|
|
return true;
|
|
|
|
/* If the hash values of t1 and t2 are different the types can't
|
|
possibly be the same. This helps keeping the type-pair hashtable
|
|
small, only tracking comparisons for hash collisions. */
|
|
if (gimple_type_hash (t1) != gimple_type_hash (t2))
|
|
return false;
|
|
|
|
/* Allocate a new cache entry for this comparison. */
|
|
p = lookup_type_pair (t1, t2);
|
|
if (p->same_p == 0 || p->same_p == 1)
|
|
{
|
|
/* We have already decided whether T1 and T2 are the
|
|
same, return the cached result. */
|
|
return p->same_p == 1;
|
|
}
|
|
|
|
if ((slot = pointer_map_contains (sccstate, p)) != NULL)
|
|
cstate = (struct sccs *)*slot;
|
|
/* Not yet visited. DFS recurse. */
|
|
if (!cstate)
|
|
{
|
|
gimple_types_compatible_p_1 (t1, t2, p,
|
|
sccstack, sccstate, sccstate_obstack);
|
|
cstate = (struct sccs *)* pointer_map_contains (sccstate, p);
|
|
state->low = MIN (state->low, cstate->low);
|
|
}
|
|
/* If the type is still on the SCC stack adjust the parents low. */
|
|
if (cstate->dfsnum < state->dfsnum
|
|
&& cstate->on_sccstack)
|
|
state->low = MIN (cstate->dfsnum, state->low);
|
|
|
|
/* Return the current lattice value. We start with an equality
|
|
assumption so types part of a SCC will be optimistically
|
|
treated equal unless proven otherwise. */
|
|
return cstate->u.same_p;
|
|
}
|
|
|
|
/* Worker for gimple_types_compatible.
|
|
SCCSTACK, SCCSTATE and SCCSTATE_OBSTACK are state for the DFS walk done. */
|
|
|
|
static bool
|
|
gimple_types_compatible_p_1 (tree t1, tree t2, type_pair_t p,
|
|
vec<type_pair_t> *sccstack,
|
|
struct pointer_map_t *sccstate,
|
|
struct obstack *sccstate_obstack)
|
|
{
|
|
struct sccs *state;
|
|
|
|
gcc_assert (p->same_p == -2);
|
|
|
|
state = XOBNEW (sccstate_obstack, struct sccs);
|
|
*pointer_map_insert (sccstate, p) = state;
|
|
|
|
sccstack->safe_push (p);
|
|
state->dfsnum = gtc_next_dfs_num++;
|
|
state->low = state->dfsnum;
|
|
state->on_sccstack = true;
|
|
/* Start with an equality assumption. As we DFS recurse into child
|
|
SCCs this assumption may get revisited. */
|
|
state->u.same_p = 1;
|
|
|
|
/* The struct tags shall compare equal. */
|
|
if (!compare_type_names_p (t1, t2))
|
|
goto different_types;
|
|
|
|
/* The main variant of both types should compare equal. */
|
|
if (TYPE_MAIN_VARIANT (t1) != t1
|
|
|| TYPE_MAIN_VARIANT (t2) != t2)
|
|
{
|
|
if (!gtc_visit (TYPE_MAIN_VARIANT (t1), TYPE_MAIN_VARIANT (t2),
|
|
state, sccstack, sccstate, sccstate_obstack))
|
|
goto different_types;
|
|
}
|
|
|
|
/* We may not merge typedef types to the same type in different
|
|
contexts. */
|
|
if (TYPE_NAME (t1)
|
|
&& TREE_CODE (TYPE_NAME (t1)) == TYPE_DECL
|
|
&& DECL_CONTEXT (TYPE_NAME (t1))
|
|
&& TYPE_P (DECL_CONTEXT (TYPE_NAME (t1))))
|
|
{
|
|
if (!gtc_visit (DECL_CONTEXT (TYPE_NAME (t1)),
|
|
DECL_CONTEXT (TYPE_NAME (t2)),
|
|
state, sccstack, sccstate, sccstate_obstack))
|
|
goto different_types;
|
|
}
|
|
|
|
/* If their attributes are not the same they can't be the same type. */
|
|
if (!attribute_list_equal (TYPE_ATTRIBUTES (t1), TYPE_ATTRIBUTES (t2)))
|
|
goto different_types;
|
|
|
|
/* Do type-specific comparisons. */
|
|
switch (TREE_CODE (t1))
|
|
{
|
|
case VECTOR_TYPE:
|
|
case COMPLEX_TYPE:
|
|
if (!gtc_visit (TREE_TYPE (t1), TREE_TYPE (t2),
|
|
state, sccstack, sccstate, sccstate_obstack))
|
|
goto different_types;
|
|
goto same_types;
|
|
|
|
case ARRAY_TYPE:
|
|
/* Array types are the same if the element types are the same and
|
|
the number of elements are the same. */
|
|
if (!gtc_visit (TREE_TYPE (t1), TREE_TYPE (t2),
|
|
state, sccstack, sccstate, sccstate_obstack)
|
|
|| TYPE_STRING_FLAG (t1) != TYPE_STRING_FLAG (t2)
|
|
|| TYPE_NONALIASED_COMPONENT (t1) != TYPE_NONALIASED_COMPONENT (t2))
|
|
goto different_types;
|
|
else
|
|
{
|
|
tree i1 = TYPE_DOMAIN (t1);
|
|
tree i2 = TYPE_DOMAIN (t2);
|
|
|
|
/* For an incomplete external array, the type domain can be
|
|
NULL_TREE. Check this condition also. */
|
|
if (i1 == NULL_TREE && i2 == NULL_TREE)
|
|
goto same_types;
|
|
else if (i1 == NULL_TREE || i2 == NULL_TREE)
|
|
goto different_types;
|
|
else
|
|
{
|
|
tree min1 = TYPE_MIN_VALUE (i1);
|
|
tree min2 = TYPE_MIN_VALUE (i2);
|
|
tree max1 = TYPE_MAX_VALUE (i1);
|
|
tree max2 = TYPE_MAX_VALUE (i2);
|
|
|
|
/* The minimum/maximum values have to be the same. */
|
|
if ((min1 == min2
|
|
|| (min1 && min2
|
|
&& ((TREE_CODE (min1) == PLACEHOLDER_EXPR
|
|
&& TREE_CODE (min2) == PLACEHOLDER_EXPR)
|
|
|| operand_equal_p (min1, min2, 0))))
|
|
&& (max1 == max2
|
|
|| (max1 && max2
|
|
&& ((TREE_CODE (max1) == PLACEHOLDER_EXPR
|
|
&& TREE_CODE (max2) == PLACEHOLDER_EXPR)
|
|
|| operand_equal_p (max1, max2, 0)))))
|
|
goto same_types;
|
|
else
|
|
goto different_types;
|
|
}
|
|
}
|
|
|
|
case METHOD_TYPE:
|
|
/* Method types should belong to the same class. */
|
|
if (!gtc_visit (TYPE_METHOD_BASETYPE (t1), TYPE_METHOD_BASETYPE (t2),
|
|
state, sccstack, sccstate, sccstate_obstack))
|
|
goto different_types;
|
|
|
|
/* Fallthru */
|
|
|
|
case FUNCTION_TYPE:
|
|
/* Function types are the same if the return type and arguments types
|
|
are the same. */
|
|
if (!gtc_visit (TREE_TYPE (t1), TREE_TYPE (t2),
|
|
state, sccstack, sccstate, sccstate_obstack))
|
|
goto different_types;
|
|
|
|
if (!comp_type_attributes (t1, t2))
|
|
goto different_types;
|
|
|
|
if (TYPE_ARG_TYPES (t1) == TYPE_ARG_TYPES (t2))
|
|
goto same_types;
|
|
else
|
|
{
|
|
tree parms1, parms2;
|
|
|
|
for (parms1 = TYPE_ARG_TYPES (t1), parms2 = TYPE_ARG_TYPES (t2);
|
|
parms1 && parms2;
|
|
parms1 = TREE_CHAIN (parms1), parms2 = TREE_CHAIN (parms2))
|
|
{
|
|
if (!gtc_visit (TREE_VALUE (parms1), TREE_VALUE (parms2),
|
|
state, sccstack, sccstate, sccstate_obstack))
|
|
goto different_types;
|
|
}
|
|
|
|
if (parms1 || parms2)
|
|
goto different_types;
|
|
|
|
goto same_types;
|
|
}
|
|
|
|
case OFFSET_TYPE:
|
|
{
|
|
if (!gtc_visit (TREE_TYPE (t1), TREE_TYPE (t2),
|
|
state, sccstack, sccstate, sccstate_obstack)
|
|
|| !gtc_visit (TYPE_OFFSET_BASETYPE (t1),
|
|
TYPE_OFFSET_BASETYPE (t2),
|
|
state, sccstack, sccstate, sccstate_obstack))
|
|
goto different_types;
|
|
|
|
goto same_types;
|
|
}
|
|
|
|
case POINTER_TYPE:
|
|
case REFERENCE_TYPE:
|
|
{
|
|
/* If the two pointers have different ref-all attributes,
|
|
they can't be the same type. */
|
|
if (TYPE_REF_CAN_ALIAS_ALL (t1) != TYPE_REF_CAN_ALIAS_ALL (t2))
|
|
goto different_types;
|
|
|
|
/* Otherwise, pointer and reference types are the same if the
|
|
pointed-to types are the same. */
|
|
if (gtc_visit (TREE_TYPE (t1), TREE_TYPE (t2),
|
|
state, sccstack, sccstate, sccstate_obstack))
|
|
goto same_types;
|
|
|
|
goto different_types;
|
|
}
|
|
|
|
case INTEGER_TYPE:
|
|
case BOOLEAN_TYPE:
|
|
{
|
|
tree min1 = TYPE_MIN_VALUE (t1);
|
|
tree max1 = TYPE_MAX_VALUE (t1);
|
|
tree min2 = TYPE_MIN_VALUE (t2);
|
|
tree max2 = TYPE_MAX_VALUE (t2);
|
|
bool min_equal_p = false;
|
|
bool max_equal_p = false;
|
|
|
|
/* If either type has a minimum value, the other type must
|
|
have the same. */
|
|
if (min1 == NULL_TREE && min2 == NULL_TREE)
|
|
min_equal_p = true;
|
|
else if (min1 && min2 && operand_equal_p (min1, min2, 0))
|
|
min_equal_p = true;
|
|
|
|
/* Likewise, if either type has a maximum value, the other
|
|
type must have the same. */
|
|
if (max1 == NULL_TREE && max2 == NULL_TREE)
|
|
max_equal_p = true;
|
|
else if (max1 && max2 && operand_equal_p (max1, max2, 0))
|
|
max_equal_p = true;
|
|
|
|
if (!min_equal_p || !max_equal_p)
|
|
goto different_types;
|
|
|
|
goto same_types;
|
|
}
|
|
|
|
case ENUMERAL_TYPE:
|
|
{
|
|
/* FIXME lto, we cannot check bounds on enumeral types because
|
|
different front ends will produce different values.
|
|
In C, enumeral types are integers, while in C++ each element
|
|
will have its own symbolic value. We should decide how enums
|
|
are to be represented in GIMPLE and have each front end lower
|
|
to that. */
|
|
tree v1, v2;
|
|
|
|
/* For enumeral types, all the values must be the same. */
|
|
if (TYPE_VALUES (t1) == TYPE_VALUES (t2))
|
|
goto same_types;
|
|
|
|
for (v1 = TYPE_VALUES (t1), v2 = TYPE_VALUES (t2);
|
|
v1 && v2;
|
|
v1 = TREE_CHAIN (v1), v2 = TREE_CHAIN (v2))
|
|
{
|
|
tree c1 = TREE_VALUE (v1);
|
|
tree c2 = TREE_VALUE (v2);
|
|
|
|
if (TREE_CODE (c1) == CONST_DECL)
|
|
c1 = DECL_INITIAL (c1);
|
|
|
|
if (TREE_CODE (c2) == CONST_DECL)
|
|
c2 = DECL_INITIAL (c2);
|
|
|
|
if (tree_int_cst_equal (c1, c2) != 1)
|
|
goto different_types;
|
|
|
|
if (TREE_PURPOSE (v1) != TREE_PURPOSE (v2))
|
|
goto different_types;
|
|
}
|
|
|
|
/* If one enumeration has more values than the other, they
|
|
are not the same. */
|
|
if (v1 || v2)
|
|
goto different_types;
|
|
|
|
goto same_types;
|
|
}
|
|
|
|
case RECORD_TYPE:
|
|
case UNION_TYPE:
|
|
case QUAL_UNION_TYPE:
|
|
{
|
|
tree f1, f2;
|
|
|
|
/* For aggregate types, all the fields must be the same. */
|
|
for (f1 = TYPE_FIELDS (t1), f2 = TYPE_FIELDS (t2);
|
|
f1 && f2;
|
|
f1 = TREE_CHAIN (f1), f2 = TREE_CHAIN (f2))
|
|
{
|
|
/* Different field kinds are not compatible. */
|
|
if (TREE_CODE (f1) != TREE_CODE (f2))
|
|
goto different_types;
|
|
/* Field decls must have the same name and offset. */
|
|
if (TREE_CODE (f1) == FIELD_DECL
|
|
&& (DECL_NONADDRESSABLE_P (f1) != DECL_NONADDRESSABLE_P (f2)
|
|
|| !gimple_compare_field_offset (f1, f2)))
|
|
goto different_types;
|
|
/* All entities should have the same name and type. */
|
|
if (DECL_NAME (f1) != DECL_NAME (f2)
|
|
|| !gtc_visit (TREE_TYPE (f1), TREE_TYPE (f2),
|
|
state, sccstack, sccstate, sccstate_obstack))
|
|
goto different_types;
|
|
}
|
|
|
|
/* If one aggregate has more fields than the other, they
|
|
are not the same. */
|
|
if (f1 || f2)
|
|
goto different_types;
|
|
|
|
goto same_types;
|
|
}
|
|
|
|
default:
|
|
gcc_unreachable ();
|
|
}
|
|
|
|
/* Common exit path for types that are not compatible. */
|
|
different_types:
|
|
state->u.same_p = 0;
|
|
goto pop;
|
|
|
|
/* Common exit path for types that are compatible. */
|
|
same_types:
|
|
gcc_assert (state->u.same_p == 1);
|
|
|
|
pop:
|
|
if (state->low == state->dfsnum)
|
|
{
|
|
type_pair_t x;
|
|
|
|
/* Pop off the SCC and set its cache values to the final
|
|
comparison result. */
|
|
do
|
|
{
|
|
struct sccs *cstate;
|
|
x = sccstack->pop ();
|
|
cstate = (struct sccs *)*pointer_map_contains (sccstate, x);
|
|
cstate->on_sccstack = false;
|
|
x->same_p = state->u.same_p;
|
|
}
|
|
while (x != p);
|
|
}
|
|
|
|
return state->u.same_p;
|
|
}
|
|
|
|
/* Return true iff T1 and T2 are structurally identical. When
|
|
FOR_MERGING_P is true the an incomplete type and a complete type
|
|
are considered different, otherwise they are considered compatible. */
|
|
|
|
static bool
|
|
gimple_types_compatible_p (tree t1, tree t2)
|
|
{
|
|
vec<type_pair_t> sccstack = vNULL;
|
|
struct pointer_map_t *sccstate;
|
|
struct obstack sccstate_obstack;
|
|
type_pair_t p = NULL;
|
|
bool res;
|
|
tree leader1, leader2;
|
|
|
|
/* Before starting to set up the SCC machinery handle simple cases. */
|
|
|
|
/* Check first for the obvious case of pointer identity. */
|
|
if (t1 == t2)
|
|
return true;
|
|
|
|
/* Check that we have two types to compare. */
|
|
if (t1 == NULL_TREE || t2 == NULL_TREE)
|
|
return false;
|
|
|
|
/* Can't be the same type if the types don't have the same code. */
|
|
if (TREE_CODE (t1) != TREE_CODE (t2))
|
|
return false;
|
|
|
|
/* Can't be the same type if they have different CV qualifiers. */
|
|
if (TYPE_QUALS (t1) != TYPE_QUALS (t2))
|
|
return false;
|
|
|
|
if (TREE_ADDRESSABLE (t1) != TREE_ADDRESSABLE (t2))
|
|
return false;
|
|
|
|
/* Void types and nullptr types are always the same. */
|
|
if (TREE_CODE (t1) == VOID_TYPE
|
|
|| TREE_CODE (t1) == NULLPTR_TYPE)
|
|
return true;
|
|
|
|
/* Can't be the same type if they have different alignment or mode. */
|
|
if (TYPE_ALIGN (t1) != TYPE_ALIGN (t2)
|
|
|| TYPE_MODE (t1) != TYPE_MODE (t2))
|
|
return false;
|
|
|
|
/* Do some simple checks before doing three hashtable queries. */
|
|
if (INTEGRAL_TYPE_P (t1)
|
|
|| SCALAR_FLOAT_TYPE_P (t1)
|
|
|| FIXED_POINT_TYPE_P (t1)
|
|
|| TREE_CODE (t1) == VECTOR_TYPE
|
|
|| TREE_CODE (t1) == COMPLEX_TYPE
|
|
|| TREE_CODE (t1) == OFFSET_TYPE
|
|
|| POINTER_TYPE_P (t1))
|
|
{
|
|
/* Can't be the same type if they have different sign or precision. */
|
|
if (TYPE_PRECISION (t1) != TYPE_PRECISION (t2)
|
|
|| TYPE_UNSIGNED (t1) != TYPE_UNSIGNED (t2))
|
|
return false;
|
|
|
|
if (TREE_CODE (t1) == INTEGER_TYPE
|
|
&& TYPE_STRING_FLAG (t1) != TYPE_STRING_FLAG (t2))
|
|
return false;
|
|
|
|
/* That's all we need to check for float and fixed-point types. */
|
|
if (SCALAR_FLOAT_TYPE_P (t1)
|
|
|| FIXED_POINT_TYPE_P (t1))
|
|
return true;
|
|
|
|
/* For other types fall through to more complex checks. */
|
|
}
|
|
|
|
/* If the types have been previously registered and found equal
|
|
they still are. */
|
|
leader1 = gimple_lookup_type_leader (t1);
|
|
leader2 = gimple_lookup_type_leader (t2);
|
|
if (leader1 == t2
|
|
|| t1 == leader2
|
|
|| (leader1 && leader1 == leader2))
|
|
return true;
|
|
|
|
/* If the hash values of t1 and t2 are different the types can't
|
|
possibly be the same. This helps keeping the type-pair hashtable
|
|
small, only tracking comparisons for hash collisions. */
|
|
if (gimple_type_hash (t1) != gimple_type_hash (t2))
|
|
return false;
|
|
|
|
/* If we've visited this type pair before (in the case of aggregates
|
|
with self-referential types), and we made a decision, return it. */
|
|
p = lookup_type_pair (t1, t2);
|
|
if (p->same_p == 0 || p->same_p == 1)
|
|
{
|
|
/* We have already decided whether T1 and T2 are the
|
|
same, return the cached result. */
|
|
return p->same_p == 1;
|
|
}
|
|
|
|
/* Now set up the SCC machinery for the comparison. */
|
|
gtc_next_dfs_num = 1;
|
|
sccstate = pointer_map_create ();
|
|
gcc_obstack_init (&sccstate_obstack);
|
|
res = gimple_types_compatible_p_1 (t1, t2, p,
|
|
&sccstack, sccstate, &sccstate_obstack);
|
|
sccstack.release ();
|
|
pointer_map_destroy (sccstate);
|
|
obstack_free (&sccstate_obstack, NULL);
|
|
|
|
return res;
|
|
}
|
|
|
|
static hashval_t
|
|
iterative_hash_gimple_type (tree, hashval_t, vec<tree> *,
|
|
struct pointer_map_t *, struct obstack *);
|
|
|
|
/* DFS visit the edge from the callers type with state *STATE to T.
|
|
Update the callers type hash V with the hash for T if it is not part
|
|
of the SCC containing the callers type and return it.
|
|
SCCSTACK, SCCSTATE and SCCSTATE_OBSTACK are state for the DFS walk done. */
|
|
|
|
static hashval_t
|
|
visit (tree t, struct sccs *state, hashval_t v,
|
|
vec<tree> *sccstack,
|
|
struct pointer_map_t *sccstate,
|
|
struct obstack *sccstate_obstack)
|
|
{
|
|
struct sccs *cstate = NULL;
|
|
struct tree_int_map m;
|
|
void **slot;
|
|
|
|
/* If there is a hash value recorded for this type then it can't
|
|
possibly be part of our parent SCC. Simply mix in its hash. */
|
|
m.base.from = t;
|
|
if ((slot = htab_find_slot (type_hash_cache, &m, NO_INSERT))
|
|
&& *slot)
|
|
return iterative_hash_hashval_t (((struct tree_int_map *) *slot)->to, v);
|
|
|
|
if ((slot = pointer_map_contains (sccstate, t)) != NULL)
|
|
cstate = (struct sccs *)*slot;
|
|
if (!cstate)
|
|
{
|
|
hashval_t tem;
|
|
/* Not yet visited. DFS recurse. */
|
|
tem = iterative_hash_gimple_type (t, v,
|
|
sccstack, sccstate, sccstate_obstack);
|
|
if (!cstate)
|
|
cstate = (struct sccs *)* pointer_map_contains (sccstate, t);
|
|
state->low = MIN (state->low, cstate->low);
|
|
/* If the type is no longer on the SCC stack and thus is not part
|
|
of the parents SCC mix in its hash value. Otherwise we will
|
|
ignore the type for hashing purposes and return the unaltered
|
|
hash value. */
|
|
if (!cstate->on_sccstack)
|
|
return tem;
|
|
}
|
|
if (cstate->dfsnum < state->dfsnum
|
|
&& cstate->on_sccstack)
|
|
state->low = MIN (cstate->dfsnum, state->low);
|
|
|
|
/* We are part of our parents SCC, skip this type during hashing
|
|
and return the unaltered hash value. */
|
|
return v;
|
|
}
|
|
|
|
/* Hash NAME with the previous hash value V and return it. */
|
|
|
|
static hashval_t
|
|
iterative_hash_name (tree name, hashval_t v)
|
|
{
|
|
if (!name)
|
|
return v;
|
|
v = iterative_hash_hashval_t (TREE_CODE (name), v);
|
|
if (TREE_CODE (name) == TYPE_DECL)
|
|
name = DECL_NAME (name);
|
|
if (!name)
|
|
return v;
|
|
gcc_assert (TREE_CODE (name) == IDENTIFIER_NODE);
|
|
return iterative_hash_object (IDENTIFIER_HASH_VALUE (name), v);
|
|
}
|
|
|
|
/* A type, hashvalue pair for sorting SCC members. */
|
|
|
|
struct type_hash_pair {
|
|
tree type;
|
|
hashval_t hash;
|
|
};
|
|
|
|
/* Compare two type, hashvalue pairs. */
|
|
|
|
static int
|
|
type_hash_pair_compare (const void *p1_, const void *p2_)
|
|
{
|
|
const struct type_hash_pair *p1 = (const struct type_hash_pair *) p1_;
|
|
const struct type_hash_pair *p2 = (const struct type_hash_pair *) p2_;
|
|
if (p1->hash < p2->hash)
|
|
return -1;
|
|
else if (p1->hash > p2->hash)
|
|
return 1;
|
|
return 0;
|
|
}
|
|
|
|
/* Returning a hash value for gimple type TYPE combined with VAL.
|
|
SCCSTACK, SCCSTATE and SCCSTATE_OBSTACK are state for the DFS walk done.
|
|
|
|
To hash a type we end up hashing in types that are reachable.
|
|
Through pointers we can end up with cycles which messes up the
|
|
required property that we need to compute the same hash value
|
|
for structurally equivalent types. To avoid this we have to
|
|
hash all types in a cycle (the SCC) in a commutative way. The
|
|
easiest way is to not mix in the hashes of the SCC members at
|
|
all. To make this work we have to delay setting the hash
|
|
values of the SCC until it is complete. */
|
|
|
|
static hashval_t
|
|
iterative_hash_gimple_type (tree type, hashval_t val,
|
|
vec<tree> *sccstack,
|
|
struct pointer_map_t *sccstate,
|
|
struct obstack *sccstate_obstack)
|
|
{
|
|
hashval_t v;
|
|
void **slot;
|
|
struct sccs *state;
|
|
|
|
/* Not visited during this DFS walk. */
|
|
gcc_checking_assert (!pointer_map_contains (sccstate, type));
|
|
state = XOBNEW (sccstate_obstack, struct sccs);
|
|
*pointer_map_insert (sccstate, type) = state;
|
|
|
|
sccstack->safe_push (type);
|
|
state->dfsnum = next_dfs_num++;
|
|
state->low = state->dfsnum;
|
|
state->on_sccstack = true;
|
|
|
|
/* Combine a few common features of types so that types are grouped into
|
|
smaller sets; when searching for existing matching types to merge,
|
|
only existing types having the same features as the new type will be
|
|
checked. */
|
|
v = iterative_hash_name (TYPE_NAME (type), 0);
|
|
if (TYPE_NAME (type)
|
|
&& TREE_CODE (TYPE_NAME (type)) == TYPE_DECL
|
|
&& DECL_CONTEXT (TYPE_NAME (type))
|
|
&& TYPE_P (DECL_CONTEXT (TYPE_NAME (type))))
|
|
v = visit (DECL_CONTEXT (TYPE_NAME (type)), state, v,
|
|
sccstack, sccstate, sccstate_obstack);
|
|
|
|
/* Factor in the variant structure. */
|
|
if (TYPE_MAIN_VARIANT (type) != type)
|
|
v = visit (TYPE_MAIN_VARIANT (type), state, v,
|
|
sccstack, sccstate, sccstate_obstack);
|
|
|
|
v = iterative_hash_hashval_t (TREE_CODE (type), v);
|
|
v = iterative_hash_hashval_t (TYPE_QUALS (type), v);
|
|
v = iterative_hash_hashval_t (TREE_ADDRESSABLE (type), v);
|
|
|
|
/* Do not hash the types size as this will cause differences in
|
|
hash values for the complete vs. the incomplete type variant. */
|
|
|
|
/* Incorporate common features of numerical types. */
|
|
if (INTEGRAL_TYPE_P (type)
|
|
|| SCALAR_FLOAT_TYPE_P (type)
|
|
|| FIXED_POINT_TYPE_P (type))
|
|
{
|
|
v = iterative_hash_hashval_t (TYPE_PRECISION (type), v);
|
|
v = iterative_hash_hashval_t (TYPE_MODE (type), v);
|
|
v = iterative_hash_hashval_t (TYPE_UNSIGNED (type), v);
|
|
}
|
|
|
|
/* For pointer and reference types, fold in information about the type
|
|
pointed to. */
|
|
if (POINTER_TYPE_P (type))
|
|
v = visit (TREE_TYPE (type), state, v,
|
|
sccstack, sccstate, sccstate_obstack);
|
|
|
|
/* For integer types hash the types min/max values and the string flag. */
|
|
if (TREE_CODE (type) == INTEGER_TYPE)
|
|
{
|
|
/* OMP lowering can introduce error_mark_node in place of
|
|
random local decls in types. */
|
|
if (TYPE_MIN_VALUE (type) != error_mark_node)
|
|
v = iterative_hash_expr (TYPE_MIN_VALUE (type), v);
|
|
if (TYPE_MAX_VALUE (type) != error_mark_node)
|
|
v = iterative_hash_expr (TYPE_MAX_VALUE (type), v);
|
|
v = iterative_hash_hashval_t (TYPE_STRING_FLAG (type), v);
|
|
}
|
|
|
|
/* For array types hash the domain and the string flag. */
|
|
if (TREE_CODE (type) == ARRAY_TYPE && TYPE_DOMAIN (type))
|
|
{
|
|
v = iterative_hash_hashval_t (TYPE_STRING_FLAG (type), v);
|
|
v = visit (TYPE_DOMAIN (type), state, v,
|
|
sccstack, sccstate, sccstate_obstack);
|
|
}
|
|
|
|
/* Recurse for aggregates with a single element type. */
|
|
if (TREE_CODE (type) == ARRAY_TYPE
|
|
|| TREE_CODE (type) == COMPLEX_TYPE
|
|
|| TREE_CODE (type) == VECTOR_TYPE)
|
|
v = visit (TREE_TYPE (type), state, v,
|
|
sccstack, sccstate, sccstate_obstack);
|
|
|
|
/* Incorporate function return and argument types. */
|
|
if (TREE_CODE (type) == FUNCTION_TYPE || TREE_CODE (type) == METHOD_TYPE)
|
|
{
|
|
unsigned na;
|
|
tree p;
|
|
|
|
/* For method types also incorporate their parent class. */
|
|
if (TREE_CODE (type) == METHOD_TYPE)
|
|
v = visit (TYPE_METHOD_BASETYPE (type), state, v,
|
|
sccstack, sccstate, sccstate_obstack);
|
|
|
|
/* Check result and argument types. */
|
|
v = visit (TREE_TYPE (type), state, v,
|
|
sccstack, sccstate, sccstate_obstack);
|
|
for (p = TYPE_ARG_TYPES (type), na = 0; p; p = TREE_CHAIN (p))
|
|
{
|
|
v = visit (TREE_VALUE (p), state, v,
|
|
sccstack, sccstate, sccstate_obstack);
|
|
na++;
|
|
}
|
|
|
|
v = iterative_hash_hashval_t (na, v);
|
|
}
|
|
|
|
if (RECORD_OR_UNION_TYPE_P (type))
|
|
{
|
|
unsigned nf;
|
|
tree f;
|
|
|
|
for (f = TYPE_FIELDS (type), nf = 0; f; f = TREE_CHAIN (f))
|
|
{
|
|
v = iterative_hash_name (DECL_NAME (f), v);
|
|
v = visit (TREE_TYPE (f), state, v,
|
|
sccstack, sccstate, sccstate_obstack);
|
|
nf++;
|
|
}
|
|
|
|
v = iterative_hash_hashval_t (nf, v);
|
|
}
|
|
|
|
/* Record hash for us. */
|
|
state->u.hash = v;
|
|
|
|
/* See if we found an SCC. */
|
|
if (state->low == state->dfsnum)
|
|
{
|
|
tree x;
|
|
struct tree_int_map *m;
|
|
|
|
/* Pop off the SCC and set its hash values. */
|
|
x = sccstack->pop ();
|
|
/* Optimize SCC size one. */
|
|
if (x == type)
|
|
{
|
|
state->on_sccstack = false;
|
|
m = ggc_alloc_cleared_tree_int_map ();
|
|
m->base.from = x;
|
|
m->to = v;
|
|
slot = htab_find_slot (type_hash_cache, m, INSERT);
|
|
gcc_assert (!*slot);
|
|
*slot = (void *) m;
|
|
}
|
|
else
|
|
{
|
|
struct sccs *cstate;
|
|
unsigned first, i, size, j;
|
|
struct type_hash_pair *pairs;
|
|
/* Pop off the SCC and build an array of type, hash pairs. */
|
|
first = sccstack->length () - 1;
|
|
while ((*sccstack)[first] != type)
|
|
--first;
|
|
size = sccstack->length () - first + 1;
|
|
pairs = XALLOCAVEC (struct type_hash_pair, size);
|
|
i = 0;
|
|
cstate = (struct sccs *)*pointer_map_contains (sccstate, x);
|
|
cstate->on_sccstack = false;
|
|
pairs[i].type = x;
|
|
pairs[i].hash = cstate->u.hash;
|
|
do
|
|
{
|
|
x = sccstack->pop ();
|
|
cstate = (struct sccs *)*pointer_map_contains (sccstate, x);
|
|
cstate->on_sccstack = false;
|
|
++i;
|
|
pairs[i].type = x;
|
|
pairs[i].hash = cstate->u.hash;
|
|
}
|
|
while (x != type);
|
|
gcc_assert (i + 1 == size);
|
|
/* Sort the arrays of type, hash pairs so that when we mix in
|
|
all members of the SCC the hash value becomes independent on
|
|
the order we visited the SCC. Disregard hashes equal to
|
|
the hash of the type we mix into because we cannot guarantee
|
|
a stable sort for those across different TUs. */
|
|
qsort (pairs, size, sizeof (struct type_hash_pair),
|
|
type_hash_pair_compare);
|
|
for (i = 0; i < size; ++i)
|
|
{
|
|
hashval_t hash;
|
|
m = ggc_alloc_cleared_tree_int_map ();
|
|
m->base.from = pairs[i].type;
|
|
hash = pairs[i].hash;
|
|
/* Skip same hashes. */
|
|
for (j = i + 1; j < size && pairs[j].hash == pairs[i].hash; ++j)
|
|
;
|
|
for (; j < size; ++j)
|
|
hash = iterative_hash_hashval_t (pairs[j].hash, hash);
|
|
for (j = 0; pairs[j].hash != pairs[i].hash; ++j)
|
|
hash = iterative_hash_hashval_t (pairs[j].hash, hash);
|
|
m->to = hash;
|
|
if (pairs[i].type == type)
|
|
v = hash;
|
|
slot = htab_find_slot (type_hash_cache, m, INSERT);
|
|
gcc_assert (!*slot);
|
|
*slot = (void *) m;
|
|
}
|
|
}
|
|
}
|
|
|
|
return iterative_hash_hashval_t (v, val);
|
|
}
|
|
|
|
/* Returns a hash value for P (assumed to be a type). The hash value
|
|
is computed using some distinguishing features of the type. Note
|
|
that we cannot use pointer hashing here as we may be dealing with
|
|
two distinct instances of the same type.
|
|
|
|
This function should produce the same hash value for two compatible
|
|
types according to gimple_types_compatible_p. */
|
|
|
|
static hashval_t
|
|
gimple_type_hash (const void *p)
|
|
{
|
|
const_tree t = (const_tree) p;
|
|
vec<tree> sccstack = vNULL;
|
|
struct pointer_map_t *sccstate;
|
|
struct obstack sccstate_obstack;
|
|
hashval_t val;
|
|
void **slot;
|
|
struct tree_int_map m;
|
|
|
|
m.base.from = CONST_CAST_TREE (t);
|
|
if ((slot = htab_find_slot (type_hash_cache, &m, NO_INSERT))
|
|
&& *slot)
|
|
return iterative_hash_hashval_t (((struct tree_int_map *) *slot)->to, 0);
|
|
|
|
/* Perform a DFS walk and pre-hash all reachable types. */
|
|
next_dfs_num = 1;
|
|
sccstate = pointer_map_create ();
|
|
gcc_obstack_init (&sccstate_obstack);
|
|
val = iterative_hash_gimple_type (CONST_CAST_TREE (t), 0,
|
|
&sccstack, sccstate, &sccstate_obstack);
|
|
sccstack.release ();
|
|
pointer_map_destroy (sccstate);
|
|
obstack_free (&sccstate_obstack, NULL);
|
|
|
|
return val;
|
|
}
|
|
|
|
/* Returns nonzero if P1 and P2 are equal. */
|
|
|
|
static int
|
|
gimple_type_eq (const void *p1, const void *p2)
|
|
{
|
|
const_tree t1 = (const_tree) p1;
|
|
const_tree t2 = (const_tree) p2;
|
|
return gimple_types_compatible_p (CONST_CAST_TREE (t1),
|
|
CONST_CAST_TREE (t2));
|
|
}
|
|
|
|
|
|
/* Worker for gimple_register_type.
|
|
Register type T in the global type table gimple_types.
|
|
When REGISTERING_MV is false first recurse for the main variant of T. */
|
|
|
|
static tree
|
|
gimple_register_type_1 (tree t, bool registering_mv)
|
|
{
|
|
void **slot;
|
|
gimple_type_leader_entry *leader;
|
|
|
|
/* If we registered this type before return the cached result. */
|
|
leader = &gimple_type_leader[TYPE_UID (t) % GIMPLE_TYPE_LEADER_SIZE];
|
|
if (leader->type == t)
|
|
return leader->leader;
|
|
|
|
/* Always register the main variant first. This is important so we
|
|
pick up the non-typedef variants as canonical, otherwise we'll end
|
|
up taking typedef ids for structure tags during comparison.
|
|
It also makes sure that main variants will be merged to main variants.
|
|
As we are operating on a possibly partially fixed up type graph
|
|
do not bother to recurse more than once, otherwise we may end up
|
|
walking in circles.
|
|
If we are registering a main variant it will either remain its
|
|
own main variant or it will be merged to something else in which
|
|
case we do not care for the main variant leader. */
|
|
if (!registering_mv
|
|
&& TYPE_MAIN_VARIANT (t) != t)
|
|
gimple_register_type_1 (TYPE_MAIN_VARIANT (t), true);
|
|
|
|
/* See if we already have an equivalent type registered. */
|
|
slot = htab_find_slot (gimple_types, t, INSERT);
|
|
if (*slot
|
|
&& *(tree *)slot != t)
|
|
{
|
|
tree new_type = (tree) *((tree *) slot);
|
|
leader->type = t;
|
|
leader->leader = new_type;
|
|
return new_type;
|
|
}
|
|
|
|
/* If not, insert it to the cache and the hash. */
|
|
leader->type = t;
|
|
leader->leader = t;
|
|
*slot = (void *) t;
|
|
return t;
|
|
}
|
|
|
|
/* Register type T in the global type table gimple_types.
|
|
If another type T', compatible with T, already existed in
|
|
gimple_types then return T', otherwise return T. This is used by
|
|
LTO to merge identical types read from different TUs. */
|
|
|
|
static tree
|
|
gimple_register_type (tree t)
|
|
{
|
|
gcc_assert (TYPE_P (t));
|
|
return gimple_register_type_1 (t, false);
|
|
}
|
|
|
|
#define GIMPLE_REGISTER_TYPE(tt) \
|
|
(TREE_VISITED (tt) ? gimple_register_type (tt) : tt)
|
|
|
|
|
|
|
|
/* A hashtable of trees that potentially refer to variables or functions
|
|
that must be replaced with their prevailing variant. */
|
|
static GTY((if_marked ("ggc_marked_p"), param_is (union tree_node))) htab_t
|
|
tree_with_vars;
|
|
|
|
/* Remember that T is a tree that (potentially) refers to a variable
|
|
or function decl that may be replaced with its prevailing variant. */
|
|
static void
|
|
remember_with_vars (tree t)
|
|
{
|
|
*(tree *) htab_find_slot (tree_with_vars, t, INSERT) = t;
|
|
}
|
|
|
|
#define LTO_FIXUP_TREE(tt) \
|
|
do \
|
|
{ \
|
|
if (tt) \
|
|
{ \
|
|
if (TYPE_P (tt)) \
|
|
(tt) = GIMPLE_REGISTER_TYPE (tt); \
|
|
if (VAR_OR_FUNCTION_DECL_P (tt) && TREE_PUBLIC (tt)) \
|
|
remember_with_vars (t); \
|
|
if (TREE_CODE (tt) == INTEGER_CST) \
|
|
(tt) = fixup_integer_cst (tt); \
|
|
} \
|
|
} while (0)
|
|
|
|
static void lto_fixup_types (tree);
|
|
|
|
/* Return integer_cst T with updated type. */
|
|
|
|
static tree
|
|
fixup_integer_cst (tree t)
|
|
{
|
|
tree type = GIMPLE_REGISTER_TYPE (TREE_TYPE (t));
|
|
|
|
if (type == TREE_TYPE (t))
|
|
return t;
|
|
|
|
/* If overflow was set, streamer_read_integer_cst
|
|
produced local copy of T. */
|
|
if (TREE_OVERFLOW (t))
|
|
{
|
|
TREE_TYPE (t) = type;
|
|
return t;
|
|
}
|
|
else
|
|
/* Otherwise produce new shared node for the new type. */
|
|
return build_int_cst_wide (type, TREE_INT_CST_LOW (t),
|
|
TREE_INT_CST_HIGH (t));
|
|
}
|
|
|
|
/* Fix up fields of a tree_typed T. */
|
|
|
|
static void
|
|
lto_ft_typed (tree t)
|
|
{
|
|
LTO_FIXUP_TREE (TREE_TYPE (t));
|
|
}
|
|
|
|
/* Fix up fields of a tree_common T. */
|
|
|
|
static void
|
|
lto_ft_common (tree t)
|
|
{
|
|
lto_ft_typed (t);
|
|
LTO_FIXUP_TREE (TREE_CHAIN (t));
|
|
}
|
|
|
|
/* Fix up fields of a decl_minimal T. */
|
|
|
|
static void
|
|
lto_ft_decl_minimal (tree t)
|
|
{
|
|
lto_ft_common (t);
|
|
LTO_FIXUP_TREE (DECL_NAME (t));
|
|
LTO_FIXUP_TREE (DECL_CONTEXT (t));
|
|
}
|
|
|
|
/* Fix up fields of a decl_common T. */
|
|
|
|
static void
|
|
lto_ft_decl_common (tree t)
|
|
{
|
|
lto_ft_decl_minimal (t);
|
|
LTO_FIXUP_TREE (DECL_SIZE (t));
|
|
LTO_FIXUP_TREE (DECL_SIZE_UNIT (t));
|
|
LTO_FIXUP_TREE (DECL_INITIAL (t));
|
|
LTO_FIXUP_TREE (DECL_ATTRIBUTES (t));
|
|
LTO_FIXUP_TREE (DECL_ABSTRACT_ORIGIN (t));
|
|
}
|
|
|
|
/* Fix up fields of a decl_with_vis T. */
|
|
|
|
static void
|
|
lto_ft_decl_with_vis (tree t)
|
|
{
|
|
lto_ft_decl_common (t);
|
|
|
|
/* Accessor macro has side-effects, use field-name here. */
|
|
LTO_FIXUP_TREE (t->decl_with_vis.assembler_name);
|
|
LTO_FIXUP_TREE (DECL_SECTION_NAME (t));
|
|
}
|
|
|
|
/* Fix up fields of a decl_non_common T. */
|
|
|
|
static void
|
|
lto_ft_decl_non_common (tree t)
|
|
{
|
|
lto_ft_decl_with_vis (t);
|
|
LTO_FIXUP_TREE (DECL_ARGUMENT_FLD (t));
|
|
LTO_FIXUP_TREE (DECL_RESULT_FLD (t));
|
|
LTO_FIXUP_TREE (DECL_VINDEX (t));
|
|
/* The C frontends may create exact duplicates for DECL_ORIGINAL_TYPE
|
|
like for 'typedef enum foo foo'. We have no way of avoiding to
|
|
merge them and dwarf2out.c cannot deal with this,
|
|
so fix this up by clearing DECL_ORIGINAL_TYPE in this case. */
|
|
if (TREE_CODE (t) == TYPE_DECL
|
|
&& DECL_ORIGINAL_TYPE (t) == TREE_TYPE (t))
|
|
DECL_ORIGINAL_TYPE (t) = NULL_TREE;
|
|
}
|
|
|
|
/* Fix up fields of a decl_non_common T. */
|
|
|
|
static void
|
|
lto_ft_function (tree t)
|
|
{
|
|
lto_ft_decl_non_common (t);
|
|
LTO_FIXUP_TREE (DECL_FUNCTION_PERSONALITY (t));
|
|
}
|
|
|
|
/* Fix up fields of a field_decl T. */
|
|
|
|
static void
|
|
lto_ft_field_decl (tree t)
|
|
{
|
|
lto_ft_decl_common (t);
|
|
LTO_FIXUP_TREE (DECL_FIELD_OFFSET (t));
|
|
LTO_FIXUP_TREE (DECL_BIT_FIELD_TYPE (t));
|
|
LTO_FIXUP_TREE (DECL_QUALIFIER (t));
|
|
LTO_FIXUP_TREE (DECL_FIELD_BIT_OFFSET (t));
|
|
LTO_FIXUP_TREE (DECL_FCONTEXT (t));
|
|
}
|
|
|
|
/* Fix up fields of a type T. */
|
|
|
|
static void
|
|
lto_ft_type (tree t)
|
|
{
|
|
lto_ft_common (t);
|
|
LTO_FIXUP_TREE (TYPE_CACHED_VALUES (t));
|
|
LTO_FIXUP_TREE (TYPE_SIZE (t));
|
|
LTO_FIXUP_TREE (TYPE_SIZE_UNIT (t));
|
|
LTO_FIXUP_TREE (TYPE_ATTRIBUTES (t));
|
|
LTO_FIXUP_TREE (TYPE_NAME (t));
|
|
|
|
/* Accessors are for derived node types only. */
|
|
if (!POINTER_TYPE_P (t))
|
|
LTO_FIXUP_TREE (TYPE_MINVAL (t));
|
|
LTO_FIXUP_TREE (TYPE_MAXVAL (t));
|
|
|
|
/* Accessor is for derived node types only. */
|
|
LTO_FIXUP_TREE (t->type_non_common.binfo);
|
|
|
|
LTO_FIXUP_TREE (TYPE_CONTEXT (t));
|
|
}
|
|
|
|
/* Fix up fields of a BINFO T. */
|
|
|
|
static void
|
|
lto_ft_binfo (tree t)
|
|
{
|
|
unsigned HOST_WIDE_INT i, n;
|
|
tree base, saved_base;
|
|
|
|
lto_ft_common (t);
|
|
LTO_FIXUP_TREE (BINFO_VTABLE (t));
|
|
LTO_FIXUP_TREE (BINFO_OFFSET (t));
|
|
LTO_FIXUP_TREE (BINFO_VIRTUALS (t));
|
|
LTO_FIXUP_TREE (BINFO_VPTR_FIELD (t));
|
|
n = vec_safe_length (BINFO_BASE_ACCESSES (t));
|
|
for (i = 0; i < n; i++)
|
|
{
|
|
saved_base = base = BINFO_BASE_ACCESS (t, i);
|
|
LTO_FIXUP_TREE (base);
|
|
if (base != saved_base)
|
|
(*BINFO_BASE_ACCESSES (t))[i] = base;
|
|
}
|
|
LTO_FIXUP_TREE (BINFO_INHERITANCE_CHAIN (t));
|
|
LTO_FIXUP_TREE (BINFO_SUBVTT_INDEX (t));
|
|
LTO_FIXUP_TREE (BINFO_VPTR_INDEX (t));
|
|
n = BINFO_N_BASE_BINFOS (t);
|
|
for (i = 0; i < n; i++)
|
|
{
|
|
saved_base = base = BINFO_BASE_BINFO (t, i);
|
|
LTO_FIXUP_TREE (base);
|
|
if (base != saved_base)
|
|
(*BINFO_BASE_BINFOS (t))[i] = base;
|
|
}
|
|
}
|
|
|
|
/* Fix up fields of a CONSTRUCTOR T. */
|
|
|
|
static void
|
|
lto_ft_constructor (tree t)
|
|
{
|
|
unsigned HOST_WIDE_INT idx;
|
|
constructor_elt *ce;
|
|
|
|
lto_ft_typed (t);
|
|
|
|
for (idx = 0; vec_safe_iterate (CONSTRUCTOR_ELTS (t), idx, &ce); idx++)
|
|
{
|
|
LTO_FIXUP_TREE (ce->index);
|
|
LTO_FIXUP_TREE (ce->value);
|
|
}
|
|
}
|
|
|
|
/* Fix up fields of an expression tree T. */
|
|
|
|
static void
|
|
lto_ft_expr (tree t)
|
|
{
|
|
int i;
|
|
lto_ft_typed (t);
|
|
for (i = TREE_OPERAND_LENGTH (t) - 1; i >= 0; --i)
|
|
LTO_FIXUP_TREE (TREE_OPERAND (t, i));
|
|
}
|
|
|
|
/* Given a tree T fixup fields of T by replacing types with their merged
|
|
variant and other entities by an equal entity from an earlier compilation
|
|
unit, or an entity being canonical in a different way. This includes
|
|
for instance integer or string constants. */
|
|
|
|
static void
|
|
lto_fixup_types (tree t)
|
|
{
|
|
switch (TREE_CODE (t))
|
|
{
|
|
case IDENTIFIER_NODE:
|
|
break;
|
|
|
|
case TREE_LIST:
|
|
LTO_FIXUP_TREE (TREE_VALUE (t));
|
|
LTO_FIXUP_TREE (TREE_PURPOSE (t));
|
|
LTO_FIXUP_TREE (TREE_CHAIN (t));
|
|
break;
|
|
|
|
case FIELD_DECL:
|
|
lto_ft_field_decl (t);
|
|
break;
|
|
|
|
case LABEL_DECL:
|
|
case CONST_DECL:
|
|
case PARM_DECL:
|
|
case RESULT_DECL:
|
|
case IMPORTED_DECL:
|
|
lto_ft_decl_common (t);
|
|
break;
|
|
|
|
case VAR_DECL:
|
|
lto_ft_decl_with_vis (t);
|
|
break;
|
|
|
|
case TYPE_DECL:
|
|
lto_ft_decl_non_common (t);
|
|
break;
|
|
|
|
case FUNCTION_DECL:
|
|
lto_ft_function (t);
|
|
break;
|
|
|
|
case TREE_BINFO:
|
|
lto_ft_binfo (t);
|
|
break;
|
|
|
|
case PLACEHOLDER_EXPR:
|
|
lto_ft_common (t);
|
|
break;
|
|
|
|
case BLOCK:
|
|
case TRANSLATION_UNIT_DECL:
|
|
case OPTIMIZATION_NODE:
|
|
case TARGET_OPTION_NODE:
|
|
break;
|
|
|
|
default:
|
|
if (TYPE_P (t))
|
|
lto_ft_type (t);
|
|
else if (TREE_CODE (t) == CONSTRUCTOR)
|
|
lto_ft_constructor (t);
|
|
else if (CONSTANT_CLASS_P (t))
|
|
LTO_FIXUP_TREE (TREE_TYPE (t));
|
|
else if (EXPR_P (t))
|
|
{
|
|
lto_ft_expr (t);
|
|
}
|
|
else
|
|
{
|
|
remember_with_vars (t);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
/* Return the resolution for the decl with index INDEX from DATA_IN. */
|
|
|
|
static enum ld_plugin_symbol_resolution
|
|
get_resolution (struct data_in *data_in, unsigned index)
|
|
{
|
|
if (data_in->globals_resolution.exists ())
|
|
{
|
|
ld_plugin_symbol_resolution_t ret;
|
|
/* We can have references to not emitted functions in
|
|
DECL_FUNCTION_PERSONALITY at least. So we can and have
|
|
to indeed return LDPR_UNKNOWN in some cases. */
|
|
if (data_in->globals_resolution.length () <= index)
|
|
return LDPR_UNKNOWN;
|
|
ret = data_in->globals_resolution[index];
|
|
return ret;
|
|
}
|
|
else
|
|
/* Delay resolution finding until decl merging. */
|
|
return LDPR_UNKNOWN;
|
|
}
|
|
|
|
/* Map assigning declarations their resolutions. */
|
|
static pointer_map_t *resolution_map;
|
|
|
|
/* We need to record resolutions until symbol table is read. */
|
|
static void
|
|
register_resolution (tree decl, enum ld_plugin_symbol_resolution resolution)
|
|
{
|
|
if (resolution == LDPR_UNKNOWN)
|
|
return;
|
|
if (!resolution_map)
|
|
resolution_map = pointer_map_create ();
|
|
*pointer_map_insert (resolution_map, decl) = (void *)(size_t)resolution;
|
|
}
|
|
|
|
/* Register DECL with the global symbol table and change its
|
|
name if necessary to avoid name clashes for static globals across
|
|
different files. */
|
|
|
|
static void
|
|
lto_register_var_decl_in_symtab (struct data_in *data_in, tree decl)
|
|
{
|
|
tree context;
|
|
|
|
/* Variable has file scope, not local. Need to ensure static variables
|
|
between different files don't clash unexpectedly. */
|
|
if (!TREE_PUBLIC (decl)
|
|
&& !((context = decl_function_context (decl))
|
|
&& auto_var_in_fn_p (decl, context)))
|
|
{
|
|
/* ??? We normally pre-mangle names before we serialize them
|
|
out. Here, in lto1, we do not know the language, and
|
|
thus cannot do the mangling again. Instead, we just
|
|
append a suffix to the mangled name. The resulting name,
|
|
however, is not a properly-formed mangled name, and will
|
|
confuse any attempt to unmangle it. */
|
|
const char *name = IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (decl));
|
|
char *label;
|
|
|
|
ASM_FORMAT_PRIVATE_NAME (label, name, DECL_UID (decl));
|
|
SET_DECL_ASSEMBLER_NAME (decl, get_identifier (label));
|
|
rest_of_decl_compilation (decl, 1, 0);
|
|
}
|
|
|
|
/* If this variable has already been declared, queue the
|
|
declaration for merging. */
|
|
if (TREE_PUBLIC (decl))
|
|
{
|
|
unsigned ix;
|
|
if (!streamer_tree_cache_lookup (data_in->reader_cache, decl, &ix))
|
|
gcc_unreachable ();
|
|
register_resolution (decl, get_resolution (data_in, ix));
|
|
}
|
|
}
|
|
|
|
|
|
/* Register DECL with the global symbol table and change its
|
|
name if necessary to avoid name clashes for static globals across
|
|
different files. DATA_IN contains descriptors and tables for the
|
|
file being read. */
|
|
|
|
static void
|
|
lto_register_function_decl_in_symtab (struct data_in *data_in, tree decl)
|
|
{
|
|
/* Need to ensure static entities between different files
|
|
don't clash unexpectedly. */
|
|
if (!TREE_PUBLIC (decl))
|
|
{
|
|
/* We must not use the DECL_ASSEMBLER_NAME macro here, as it
|
|
may set the assembler name where it was previously empty. */
|
|
tree old_assembler_name = decl->decl_with_vis.assembler_name;
|
|
|
|
/* FIXME lto: We normally pre-mangle names before we serialize
|
|
them out. Here, in lto1, we do not know the language, and
|
|
thus cannot do the mangling again. Instead, we just append a
|
|
suffix to the mangled name. The resulting name, however, is
|
|
not a properly-formed mangled name, and will confuse any
|
|
attempt to unmangle it. */
|
|
const char *name = IDENTIFIER_POINTER (DECL_ASSEMBLER_NAME (decl));
|
|
char *label;
|
|
|
|
ASM_FORMAT_PRIVATE_NAME (label, name, DECL_UID (decl));
|
|
SET_DECL_ASSEMBLER_NAME (decl, get_identifier (label));
|
|
|
|
/* We may arrive here with the old assembler name not set
|
|
if the function body is not needed, e.g., it has been
|
|
inlined away and does not appear in the cgraph. */
|
|
if (old_assembler_name)
|
|
{
|
|
tree new_assembler_name = DECL_ASSEMBLER_NAME (decl);
|
|
|
|
/* Make the original assembler name available for later use.
|
|
We may have used it to indicate the section within its
|
|
object file where the function body may be found.
|
|
FIXME lto: Find a better way to maintain the function decl
|
|
to body section mapping so we don't need this hack. */
|
|
lto_record_renamed_decl (data_in->file_data,
|
|
IDENTIFIER_POINTER (old_assembler_name),
|
|
IDENTIFIER_POINTER (new_assembler_name));
|
|
}
|
|
}
|
|
|
|
/* If this variable has already been declared, queue the
|
|
declaration for merging. */
|
|
if (TREE_PUBLIC (decl) && !DECL_ABSTRACT (decl))
|
|
{
|
|
unsigned ix;
|
|
if (!streamer_tree_cache_lookup (data_in->reader_cache, decl, &ix))
|
|
gcc_unreachable ();
|
|
register_resolution (decl, get_resolution (data_in, ix));
|
|
}
|
|
}
|
|
|
|
|
|
/* Given a streamer cache structure DATA_IN (holding a sequence of trees
|
|
for one compilation unit) go over all trees starting at index FROM until the
|
|
end of the sequence and replace fields of those trees, and the trees
|
|
themself with their canonical variants as per gimple_register_type. */
|
|
|
|
static void
|
|
uniquify_nodes (struct data_in *data_in, unsigned from)
|
|
{
|
|
struct streamer_tree_cache_d *cache = data_in->reader_cache;
|
|
unsigned len = cache->nodes.length ();
|
|
unsigned i;
|
|
|
|
/* Go backwards because children streamed for the first time come
|
|
as part of their parents, and hence are created after them. */
|
|
|
|
/* First register all the types in the cache. This makes sure to
|
|
have the original structure in the type cycles when registering
|
|
them and computing hashes. */
|
|
for (i = len; i-- > from;)
|
|
{
|
|
tree t = cache->nodes[i];
|
|
if (t && TYPE_P (t))
|
|
{
|
|
tree newt = gimple_register_type (t);
|
|
/* Mark non-prevailing types so we fix them up. No need
|
|
to reset that flag afterwards - nothing that refers
|
|
to those types is left and they are collected. */
|
|
if (newt != t)
|
|
TREE_VISITED (t) = 1;
|
|
}
|
|
}
|
|
|
|
/* Second fixup all trees in the new cache entries. */
|
|
for (i = len; i-- > from;)
|
|
{
|
|
tree t = cache->nodes[i];
|
|
tree oldt = t;
|
|
if (!t)
|
|
continue;
|
|
|
|
/* First fixup the fields of T. */
|
|
lto_fixup_types (t);
|
|
|
|
if (!TYPE_P (t))
|
|
continue;
|
|
|
|
/* Now try to find a canonical variant of T itself. */
|
|
t = GIMPLE_REGISTER_TYPE (t);
|
|
|
|
if (t == oldt)
|
|
{
|
|
/* The following re-creates proper variant lists while fixing up
|
|
the variant leaders. We do not stream TYPE_NEXT_VARIANT so the
|
|
variant list state before fixup is broken. */
|
|
tree tem, mv;
|
|
|
|
#ifdef ENABLE_CHECKING
|
|
/* Remove us from our main variant list if we are not the
|
|
variant leader. */
|
|
if (TYPE_MAIN_VARIANT (t) != t)
|
|
{
|
|
tem = TYPE_MAIN_VARIANT (t);
|
|
while (tem && TYPE_NEXT_VARIANT (tem) != t)
|
|
tem = TYPE_NEXT_VARIANT (tem);
|
|
gcc_assert (!tem && !TYPE_NEXT_VARIANT (t));
|
|
}
|
|
#endif
|
|
|
|
/* Query our new main variant. */
|
|
mv = GIMPLE_REGISTER_TYPE (TYPE_MAIN_VARIANT (t));
|
|
|
|
/* If we were the variant leader and we get replaced ourselves drop
|
|
all variants from our list. */
|
|
if (TYPE_MAIN_VARIANT (t) == t
|
|
&& mv != t)
|
|
{
|
|
tem = t;
|
|
while (tem)
|
|
{
|
|
tree tem2 = TYPE_NEXT_VARIANT (tem);
|
|
TYPE_NEXT_VARIANT (tem) = NULL_TREE;
|
|
tem = tem2;
|
|
}
|
|
}
|
|
|
|
/* If we are not our own variant leader link us into our new leaders
|
|
variant list. */
|
|
if (mv != t)
|
|
{
|
|
TYPE_NEXT_VARIANT (t) = TYPE_NEXT_VARIANT (mv);
|
|
TYPE_NEXT_VARIANT (mv) = t;
|
|
if (RECORD_OR_UNION_TYPE_P (t))
|
|
TYPE_BINFO (t) = TYPE_BINFO (mv);
|
|
/* Preserve the invariant that type variants share their
|
|
TYPE_FIELDS. */
|
|
if (RECORD_OR_UNION_TYPE_P (t)
|
|
&& TYPE_FIELDS (mv) != TYPE_FIELDS (t))
|
|
{
|
|
tree f1, f2;
|
|
for (f1 = TYPE_FIELDS (mv), f2 = TYPE_FIELDS (t);
|
|
f1 && f2; f1 = TREE_CHAIN (f1), f2 = TREE_CHAIN (f2))
|
|
{
|
|
unsigned ix;
|
|
gcc_assert (f1 != f2
|
|
&& DECL_NAME (f1) == DECL_NAME (f2));
|
|
if (!streamer_tree_cache_lookup (cache, f2, &ix))
|
|
gcc_unreachable ();
|
|
/* If we're going to replace an element which we'd
|
|
still visit in the next iterations, we wouldn't
|
|
handle it, so do it here. We do have to handle it
|
|
even though the field_decl itself will be removed,
|
|
as it could refer to e.g. integer_cst which we
|
|
wouldn't reach via any other way, hence they
|
|
(and their type) would stay uncollected. */
|
|
/* ??? We should rather make sure to replace all
|
|
references to f2 with f1. That means handling
|
|
COMPONENT_REFs and CONSTRUCTOR elements in
|
|
lto_fixup_types and special-case the field-decl
|
|
operand handling. */
|
|
/* ??? Not sure the above is all relevant in this
|
|
path canonicalizing TYPE_FIELDS to that of the
|
|
main variant. */
|
|
if (ix < i)
|
|
lto_fixup_types (f2);
|
|
streamer_tree_cache_insert_at (cache, f1, ix);
|
|
}
|
|
TYPE_FIELDS (t) = TYPE_FIELDS (mv);
|
|
}
|
|
}
|
|
|
|
/* Finally adjust our main variant and fix it up. */
|
|
TYPE_MAIN_VARIANT (t) = mv;
|
|
|
|
/* The following reconstructs the pointer chains
|
|
of the new pointed-to type if we are a main variant. We do
|
|
not stream those so they are broken before fixup. */
|
|
if (TREE_CODE (t) == POINTER_TYPE
|
|
&& TYPE_MAIN_VARIANT (t) == t)
|
|
{
|
|
TYPE_NEXT_PTR_TO (t) = TYPE_POINTER_TO (TREE_TYPE (t));
|
|
TYPE_POINTER_TO (TREE_TYPE (t)) = t;
|
|
}
|
|
else if (TREE_CODE (t) == REFERENCE_TYPE
|
|
&& TYPE_MAIN_VARIANT (t) == t)
|
|
{
|
|
TYPE_NEXT_REF_TO (t) = TYPE_REFERENCE_TO (TREE_TYPE (t));
|
|
TYPE_REFERENCE_TO (TREE_TYPE (t)) = t;
|
|
}
|
|
}
|
|
|
|
else
|
|
{
|
|
if (RECORD_OR_UNION_TYPE_P (t))
|
|
{
|
|
tree f1, f2;
|
|
if (TYPE_FIELDS (t) != TYPE_FIELDS (oldt))
|
|
for (f1 = TYPE_FIELDS (t), f2 = TYPE_FIELDS (oldt);
|
|
f1 && f2; f1 = TREE_CHAIN (f1), f2 = TREE_CHAIN (f2))
|
|
{
|
|
unsigned ix;
|
|
gcc_assert (f1 != f2 && DECL_NAME (f1) == DECL_NAME (f2));
|
|
if (!streamer_tree_cache_lookup (cache, f2, &ix))
|
|
gcc_unreachable ();
|
|
/* If we're going to replace an element which we'd
|
|
still visit in the next iterations, we wouldn't
|
|
handle it, so do it here. We do have to handle it
|
|
even though the field_decl itself will be removed,
|
|
as it could refer to e.g. integer_cst which we
|
|
wouldn't reach via any other way, hence they
|
|
(and their type) would stay uncollected. */
|
|
/* ??? We should rather make sure to replace all
|
|
references to f2 with f1. That means handling
|
|
COMPONENT_REFs and CONSTRUCTOR elements in
|
|
lto_fixup_types and special-case the field-decl
|
|
operand handling. */
|
|
if (ix < i)
|
|
lto_fixup_types (f2);
|
|
streamer_tree_cache_insert_at (cache, f1, ix);
|
|
}
|
|
}
|
|
|
|
/* If we found a tree that is equal to oldt replace it in the
|
|
cache, so that further users (in the various LTO sections)
|
|
make use of it. */
|
|
streamer_tree_cache_insert_at (cache, t, i);
|
|
}
|
|
}
|
|
|
|
/* Finally compute the canonical type of all TREE_TYPEs and register
|
|
VAR_DECL and FUNCTION_DECL nodes in the symbol table.
|
|
From this point there are no longer any types with
|
|
TYPE_STRUCTURAL_EQUALITY_P and its type-based alias problems.
|
|
This step requires the TYPE_POINTER_TO lists being present, so
|
|
make sure it is done last. */
|
|
for (i = len; i-- > from;)
|
|
{
|
|
tree t = cache->nodes[i];
|
|
if (t == NULL_TREE)
|
|
continue;
|
|
|
|
if (TREE_CODE (t) == VAR_DECL)
|
|
lto_register_var_decl_in_symtab (data_in, t);
|
|
else if (TREE_CODE (t) == FUNCTION_DECL && !DECL_BUILT_IN (t))
|
|
lto_register_function_decl_in_symtab (data_in, t);
|
|
else if (!flag_wpa
|
|
&& TREE_CODE (t) == TYPE_DECL)
|
|
debug_hooks->type_decl (t, !DECL_FILE_SCOPE_P (t));
|
|
else if (TYPE_P (t) && !TYPE_CANONICAL (t))
|
|
TYPE_CANONICAL (t) = gimple_register_canonical_type (t);
|
|
}
|
|
}
|
|
|
|
|
|
/* Read all the symbols from buffer DATA, using descriptors in DECL_DATA.
|
|
RESOLUTIONS is the set of symbols picked by the linker (read from the
|
|
resolution file when the linker plugin is being used). */
|
|
|
|
static void
|
|
lto_read_decls (struct lto_file_decl_data *decl_data, const void *data,
|
|
vec<ld_plugin_symbol_resolution_t> resolutions)
|
|
{
|
|
const struct lto_decl_header *header = (const struct lto_decl_header *) data;
|
|
const int decl_offset = sizeof (struct lto_decl_header);
|
|
const int main_offset = decl_offset + header->decl_state_size;
|
|
const int string_offset = main_offset + header->main_size;
|
|
struct lto_input_block ib_main;
|
|
struct data_in *data_in;
|
|
unsigned int i;
|
|
const uint32_t *data_ptr, *data_end;
|
|
uint32_t num_decl_states;
|
|
|
|
LTO_INIT_INPUT_BLOCK (ib_main, (const char *) data + main_offset, 0,
|
|
header->main_size);
|
|
|
|
data_in = lto_data_in_create (decl_data, (const char *) data + string_offset,
|
|
header->string_size, resolutions);
|
|
|
|
/* We do not uniquify the pre-loaded cache entries, those are middle-end
|
|
internal types that should not be merged. */
|
|
|
|
/* Read the global declarations and types. */
|
|
while (ib_main.p < ib_main.len)
|
|
{
|
|
tree t;
|
|
unsigned from = data_in->reader_cache->nodes.length ();
|
|
t = stream_read_tree (&ib_main, data_in);
|
|
gcc_assert (t && ib_main.p <= ib_main.len);
|
|
uniquify_nodes (data_in, from);
|
|
}
|
|
|
|
/* Read in lto_in_decl_state objects. */
|
|
data_ptr = (const uint32_t *) ((const char*) data + decl_offset);
|
|
data_end =
|
|
(const uint32_t *) ((const char*) data_ptr + header->decl_state_size);
|
|
num_decl_states = *data_ptr++;
|
|
|
|
gcc_assert (num_decl_states > 0);
|
|
decl_data->global_decl_state = lto_new_in_decl_state ();
|
|
data_ptr = lto_read_in_decl_state (data_in, data_ptr,
|
|
decl_data->global_decl_state);
|
|
|
|
/* Read in per-function decl states and enter them in hash table. */
|
|
decl_data->function_decl_states =
|
|
htab_create_ggc (37, lto_hash_in_decl_state, lto_eq_in_decl_state, NULL);
|
|
|
|
for (i = 1; i < num_decl_states; i++)
|
|
{
|
|
struct lto_in_decl_state *state = lto_new_in_decl_state ();
|
|
void **slot;
|
|
|
|
data_ptr = lto_read_in_decl_state (data_in, data_ptr, state);
|
|
slot = htab_find_slot (decl_data->function_decl_states, state, INSERT);
|
|
gcc_assert (*slot == NULL);
|
|
*slot = state;
|
|
}
|
|
|
|
if (data_ptr != data_end)
|
|
internal_error ("bytecode stream: garbage at the end of symbols section");
|
|
|
|
/* Set the current decl state to be the global state. */
|
|
decl_data->current_decl_state = decl_data->global_decl_state;
|
|
|
|
lto_data_in_delete (data_in);
|
|
}
|
|
|
|
/* Custom version of strtoll, which is not portable. */
|
|
|
|
static HOST_WIDEST_INT
|
|
lto_parse_hex (const char *p)
|
|
{
|
|
HOST_WIDEST_INT ret = 0;
|
|
|
|
for (; *p != '\0'; ++p)
|
|
{
|
|
char c = *p;
|
|
unsigned char part;
|
|
ret <<= 4;
|
|
if (c >= '0' && c <= '9')
|
|
part = c - '0';
|
|
else if (c >= 'a' && c <= 'f')
|
|
part = c - 'a' + 10;
|
|
else if (c >= 'A' && c <= 'F')
|
|
part = c - 'A' + 10;
|
|
else
|
|
internal_error ("could not parse hex number");
|
|
ret |= part;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
/* Read resolution for file named FILE_NAME. The resolution is read from
|
|
RESOLUTION. */
|
|
|
|
static void
|
|
lto_resolution_read (splay_tree file_ids, FILE *resolution, lto_file *file)
|
|
{
|
|
/* We require that objects in the resolution file are in the same
|
|
order as the lto1 command line. */
|
|
unsigned int name_len;
|
|
char *obj_name;
|
|
unsigned int num_symbols;
|
|
unsigned int i;
|
|
struct lto_file_decl_data *file_data;
|
|
splay_tree_node nd = NULL;
|
|
|
|
if (!resolution)
|
|
return;
|
|
|
|
name_len = strlen (file->filename);
|
|
obj_name = XNEWVEC (char, name_len + 1);
|
|
fscanf (resolution, " "); /* Read white space. */
|
|
|
|
fread (obj_name, sizeof (char), name_len, resolution);
|
|
obj_name[name_len] = '\0';
|
|
if (filename_cmp (obj_name, file->filename) != 0)
|
|
internal_error ("unexpected file name %s in linker resolution file. "
|
|
"Expected %s", obj_name, file->filename);
|
|
if (file->offset != 0)
|
|
{
|
|
int t;
|
|
char offset_p[17];
|
|
HOST_WIDEST_INT offset;
|
|
t = fscanf (resolution, "@0x%16s", offset_p);
|
|
if (t != 1)
|
|
internal_error ("could not parse file offset");
|
|
offset = lto_parse_hex (offset_p);
|
|
if (offset != file->offset)
|
|
internal_error ("unexpected offset");
|
|
}
|
|
|
|
free (obj_name);
|
|
|
|
fscanf (resolution, "%u", &num_symbols);
|
|
|
|
for (i = 0; i < num_symbols; i++)
|
|
{
|
|
int t;
|
|
unsigned index;
|
|
unsigned HOST_WIDE_INT id;
|
|
char r_str[27];
|
|
enum ld_plugin_symbol_resolution r = (enum ld_plugin_symbol_resolution) 0;
|
|
unsigned int j;
|
|
unsigned int lto_resolution_str_len =
|
|
sizeof (lto_resolution_str) / sizeof (char *);
|
|
res_pair rp;
|
|
|
|
t = fscanf (resolution, "%u " HOST_WIDE_INT_PRINT_HEX_PURE " %26s %*[^\n]\n",
|
|
&index, &id, r_str);
|
|
if (t != 3)
|
|
internal_error ("invalid line in the resolution file");
|
|
|
|
for (j = 0; j < lto_resolution_str_len; j++)
|
|
{
|
|
if (strcmp (lto_resolution_str[j], r_str) == 0)
|
|
{
|
|
r = (enum ld_plugin_symbol_resolution) j;
|
|
break;
|
|
}
|
|
}
|
|
if (j == lto_resolution_str_len)
|
|
internal_error ("invalid resolution in the resolution file");
|
|
|
|
if (!(nd && lto_splay_tree_id_equal_p (nd->key, id)))
|
|
{
|
|
nd = lto_splay_tree_lookup (file_ids, id);
|
|
if (nd == NULL)
|
|
internal_error ("resolution sub id %wx not in object file", id);
|
|
}
|
|
|
|
file_data = (struct lto_file_decl_data *)nd->value;
|
|
/* The indexes are very sparse. To save memory save them in a compact
|
|
format that is only unpacked later when the subfile is processed. */
|
|
rp.res = r;
|
|
rp.index = index;
|
|
file_data->respairs.safe_push (rp);
|
|
if (file_data->max_index < index)
|
|
file_data->max_index = index;
|
|
}
|
|
}
|
|
|
|
/* List of file_decl_datas */
|
|
struct file_data_list
|
|
{
|
|
struct lto_file_decl_data *first, *last;
|
|
};
|
|
|
|
/* Is the name for a id'ed LTO section? */
|
|
|
|
static int
|
|
lto_section_with_id (const char *name, unsigned HOST_WIDE_INT *id)
|
|
{
|
|
const char *s;
|
|
|
|
if (strncmp (name, LTO_SECTION_NAME_PREFIX, strlen (LTO_SECTION_NAME_PREFIX)))
|
|
return 0;
|
|
s = strrchr (name, '.');
|
|
return s && sscanf (s, "." HOST_WIDE_INT_PRINT_HEX_PURE, id) == 1;
|
|
}
|
|
|
|
/* Create file_data of each sub file id */
|
|
|
|
static int
|
|
create_subid_section_table (struct lto_section_slot *ls, splay_tree file_ids,
|
|
struct file_data_list *list)
|
|
{
|
|
struct lto_section_slot s_slot, *new_slot;
|
|
unsigned HOST_WIDE_INT id;
|
|
splay_tree_node nd;
|
|
void **hash_slot;
|
|
char *new_name;
|
|
struct lto_file_decl_data *file_data;
|
|
|
|
if (!lto_section_with_id (ls->name, &id))
|
|
return 1;
|
|
|
|
/* Find hash table of sub module id */
|
|
nd = lto_splay_tree_lookup (file_ids, id);
|
|
if (nd != NULL)
|
|
{
|
|
file_data = (struct lto_file_decl_data *)nd->value;
|
|
}
|
|
else
|
|
{
|
|
file_data = ggc_alloc_lto_file_decl_data ();
|
|
memset(file_data, 0, sizeof (struct lto_file_decl_data));
|
|
file_data->id = id;
|
|
file_data->section_hash_table = lto_obj_create_section_hash_table ();;
|
|
lto_splay_tree_insert (file_ids, id, file_data);
|
|
|
|
/* Maintain list in linker order */
|
|
if (!list->first)
|
|
list->first = file_data;
|
|
if (list->last)
|
|
list->last->next = file_data;
|
|
list->last = file_data;
|
|
}
|
|
|
|
/* Copy section into sub module hash table */
|
|
new_name = XDUPVEC (char, ls->name, strlen (ls->name) + 1);
|
|
s_slot.name = new_name;
|
|
hash_slot = htab_find_slot (file_data->section_hash_table, &s_slot, INSERT);
|
|
gcc_assert (*hash_slot == NULL);
|
|
|
|
new_slot = XDUP (struct lto_section_slot, ls);
|
|
new_slot->name = new_name;
|
|
*hash_slot = new_slot;
|
|
return 1;
|
|
}
|
|
|
|
/* Read declarations and other initializations for a FILE_DATA. */
|
|
|
|
static void
|
|
lto_file_finalize (struct lto_file_decl_data *file_data, lto_file *file)
|
|
{
|
|
const char *data;
|
|
size_t len;
|
|
vec<ld_plugin_symbol_resolution_t>
|
|
resolutions = vNULL;
|
|
int i;
|
|
res_pair *rp;
|
|
|
|
/* Create vector for fast access of resolution. We do this lazily
|
|
to save memory. */
|
|
resolutions.safe_grow_cleared (file_data->max_index + 1);
|
|
for (i = 0; file_data->respairs.iterate (i, &rp); i++)
|
|
resolutions[rp->index] = rp->res;
|
|
file_data->respairs.release ();
|
|
|
|
file_data->renaming_hash_table = lto_create_renaming_table ();
|
|
file_data->file_name = file->filename;
|
|
data = lto_get_section_data (file_data, LTO_section_decls, NULL, &len);
|
|
if (data == NULL)
|
|
{
|
|
internal_error ("cannot read LTO decls from %s", file_data->file_name);
|
|
return;
|
|
}
|
|
/* Frees resolutions */
|
|
lto_read_decls (file_data, data, resolutions);
|
|
lto_free_section_data (file_data, LTO_section_decls, NULL, data, len);
|
|
}
|
|
|
|
/* Finalize FILE_DATA in FILE and increase COUNT. */
|
|
|
|
static int
|
|
lto_create_files_from_ids (lto_file *file, struct lto_file_decl_data *file_data,
|
|
int *count)
|
|
{
|
|
lto_file_finalize (file_data, file);
|
|
if (cgraph_dump_file)
|
|
fprintf (cgraph_dump_file, "Creating file %s with sub id " HOST_WIDE_INT_PRINT_HEX "\n",
|
|
file_data->file_name, file_data->id);
|
|
(*count)++;
|
|
return 0;
|
|
}
|
|
|
|
/* Generate a TREE representation for all types and external decls
|
|
entities in FILE.
|
|
|
|
Read all of the globals out of the file. Then read the cgraph
|
|
and process the .o index into the cgraph nodes so that it can open
|
|
the .o file to load the functions and ipa information. */
|
|
|
|
static struct lto_file_decl_data *
|
|
lto_file_read (lto_file *file, FILE *resolution_file, int *count)
|
|
{
|
|
struct lto_file_decl_data *file_data = NULL;
|
|
splay_tree file_ids;
|
|
htab_t section_hash_table;
|
|
struct lto_section_slot *section;
|
|
struct file_data_list file_list;
|
|
struct lto_section_list section_list;
|
|
|
|
memset (§ion_list, 0, sizeof (struct lto_section_list));
|
|
section_hash_table = lto_obj_build_section_table (file, §ion_list);
|
|
|
|
/* Find all sub modules in the object and put their sections into new hash
|
|
tables in a splay tree. */
|
|
file_ids = lto_splay_tree_new ();
|
|
memset (&file_list, 0, sizeof (struct file_data_list));
|
|
for (section = section_list.first; section != NULL; section = section->next)
|
|
create_subid_section_table (section, file_ids, &file_list);
|
|
|
|
/* Add resolutions to file ids */
|
|
lto_resolution_read (file_ids, resolution_file, file);
|
|
|
|
/* Finalize each lto file for each submodule in the merged object */
|
|
for (file_data = file_list.first; file_data != NULL; file_data = file_data->next)
|
|
lto_create_files_from_ids (file, file_data, count);
|
|
|
|
splay_tree_delete (file_ids);
|
|
htab_delete (section_hash_table);
|
|
|
|
return file_list.first;
|
|
}
|
|
|
|
#if HAVE_MMAP_FILE && HAVE_SYSCONF && defined _SC_PAGE_SIZE
|
|
#define LTO_MMAP_IO 1
|
|
#endif
|
|
|
|
#if LTO_MMAP_IO
|
|
/* Page size of machine is used for mmap and munmap calls. */
|
|
static size_t page_mask;
|
|
#endif
|
|
|
|
/* Get the section data of length LEN from FILENAME starting at
|
|
OFFSET. The data segment must be freed by the caller when the
|
|
caller is finished. Returns NULL if all was not well. */
|
|
|
|
static char *
|
|
lto_read_section_data (struct lto_file_decl_data *file_data,
|
|
intptr_t offset, size_t len)
|
|
{
|
|
char *result;
|
|
static int fd = -1;
|
|
static char *fd_name;
|
|
#if LTO_MMAP_IO
|
|
intptr_t computed_len;
|
|
intptr_t computed_offset;
|
|
intptr_t diff;
|
|
#endif
|
|
|
|
/* Keep a single-entry file-descriptor cache. The last file we
|
|
touched will get closed at exit.
|
|
??? Eventually we want to add a more sophisticated larger cache
|
|
or rather fix function body streaming to not stream them in
|
|
practically random order. */
|
|
if (fd != -1
|
|
&& filename_cmp (fd_name, file_data->file_name) != 0)
|
|
{
|
|
free (fd_name);
|
|
close (fd);
|
|
fd = -1;
|
|
}
|
|
if (fd == -1)
|
|
{
|
|
fd = open (file_data->file_name, O_RDONLY|O_BINARY);
|
|
if (fd == -1)
|
|
{
|
|
fatal_error ("Cannot open %s", file_data->file_name);
|
|
return NULL;
|
|
}
|
|
fd_name = xstrdup (file_data->file_name);
|
|
}
|
|
|
|
#if LTO_MMAP_IO
|
|
if (!page_mask)
|
|
{
|
|
size_t page_size = sysconf (_SC_PAGE_SIZE);
|
|
page_mask = ~(page_size - 1);
|
|
}
|
|
|
|
computed_offset = offset & page_mask;
|
|
diff = offset - computed_offset;
|
|
computed_len = len + diff;
|
|
|
|
result = (char *) mmap (NULL, computed_len, PROT_READ, MAP_PRIVATE,
|
|
fd, computed_offset);
|
|
if (result == MAP_FAILED)
|
|
{
|
|
fatal_error ("Cannot map %s", file_data->file_name);
|
|
return NULL;
|
|
}
|
|
|
|
return result + diff;
|
|
#else
|
|
result = (char *) xmalloc (len);
|
|
if (lseek (fd, offset, SEEK_SET) != offset
|
|
|| read (fd, result, len) != (ssize_t) len)
|
|
{
|
|
free (result);
|
|
fatal_error ("Cannot read %s", file_data->file_name);
|
|
result = NULL;
|
|
}
|
|
#ifdef __MINGW32__
|
|
/* Native windows doesn't supports delayed unlink on opened file. So
|
|
we close file here again. This produces higher I/O load, but at least
|
|
it prevents to have dangling file handles preventing unlink. */
|
|
free (fd_name);
|
|
fd_name = NULL;
|
|
close (fd);
|
|
fd = -1;
|
|
#endif
|
|
return result;
|
|
#endif
|
|
}
|
|
|
|
|
|
/* Get the section data from FILE_DATA of SECTION_TYPE with NAME.
|
|
NAME will be NULL unless the section type is for a function
|
|
body. */
|
|
|
|
static const char *
|
|
get_section_data (struct lto_file_decl_data *file_data,
|
|
enum lto_section_type section_type,
|
|
const char *name,
|
|
size_t *len)
|
|
{
|
|
htab_t section_hash_table = file_data->section_hash_table;
|
|
struct lto_section_slot *f_slot;
|
|
struct lto_section_slot s_slot;
|
|
const char *section_name = lto_get_section_name (section_type, name, file_data);
|
|
char *data = NULL;
|
|
|
|
*len = 0;
|
|
s_slot.name = section_name;
|
|
f_slot = (struct lto_section_slot *) htab_find (section_hash_table, &s_slot);
|
|
if (f_slot)
|
|
{
|
|
data = lto_read_section_data (file_data, f_slot->start, f_slot->len);
|
|
*len = f_slot->len;
|
|
}
|
|
|
|
free (CONST_CAST (char *, section_name));
|
|
return data;
|
|
}
|
|
|
|
|
|
/* Free the section data from FILE_DATA of SECTION_TYPE with NAME that
|
|
starts at OFFSET and has LEN bytes. */
|
|
|
|
static void
|
|
free_section_data (struct lto_file_decl_data *file_data ATTRIBUTE_UNUSED,
|
|
enum lto_section_type section_type ATTRIBUTE_UNUSED,
|
|
const char *name ATTRIBUTE_UNUSED,
|
|
const char *offset, size_t len ATTRIBUTE_UNUSED)
|
|
{
|
|
#if LTO_MMAP_IO
|
|
intptr_t computed_len;
|
|
intptr_t computed_offset;
|
|
intptr_t diff;
|
|
#endif
|
|
|
|
#if LTO_MMAP_IO
|
|
computed_offset = ((intptr_t) offset) & page_mask;
|
|
diff = (intptr_t) offset - computed_offset;
|
|
computed_len = len + diff;
|
|
|
|
munmap ((caddr_t) computed_offset, computed_len);
|
|
#else
|
|
free (CONST_CAST(char *, offset));
|
|
#endif
|
|
}
|
|
|
|
static lto_file *current_lto_file;
|
|
|
|
/* Helper for qsort; compare partitions and return one with smaller size.
|
|
We sort from greatest to smallest so parallel build doesn't stale on the
|
|
longest compilation being executed too late. */
|
|
|
|
static int
|
|
cmp_partitions_size (const void *a, const void *b)
|
|
{
|
|
const struct ltrans_partition_def *pa
|
|
= *(struct ltrans_partition_def *const *)a;
|
|
const struct ltrans_partition_def *pb
|
|
= *(struct ltrans_partition_def *const *)b;
|
|
return pb->insns - pa->insns;
|
|
}
|
|
|
|
/* Helper for qsort; compare partitions and return one with smaller order. */
|
|
|
|
static int
|
|
cmp_partitions_order (const void *a, const void *b)
|
|
{
|
|
const struct ltrans_partition_def *pa
|
|
= *(struct ltrans_partition_def *const *)a;
|
|
const struct ltrans_partition_def *pb
|
|
= *(struct ltrans_partition_def *const *)b;
|
|
int ordera = -1, orderb = -1;
|
|
|
|
if (lto_symtab_encoder_size (pa->encoder))
|
|
ordera = lto_symtab_encoder_deref (pa->encoder, 0)->symbol.order;
|
|
if (lto_symtab_encoder_size (pb->encoder))
|
|
orderb = lto_symtab_encoder_deref (pb->encoder, 0)->symbol.order;
|
|
return orderb - ordera;
|
|
}
|
|
|
|
/* Write all output files in WPA mode and the file with the list of
|
|
LTRANS units. */
|
|
|
|
static void
|
|
lto_wpa_write_files (void)
|
|
{
|
|
unsigned i, n_sets;
|
|
lto_file *file;
|
|
ltrans_partition part;
|
|
FILE *ltrans_output_list_stream;
|
|
char *temp_filename;
|
|
size_t blen;
|
|
|
|
/* Open the LTRANS output list. */
|
|
if (!ltrans_output_list)
|
|
fatal_error ("no LTRANS output list filename provided");
|
|
ltrans_output_list_stream = fopen (ltrans_output_list, "w");
|
|
if (ltrans_output_list_stream == NULL)
|
|
fatal_error ("opening LTRANS output list %s: %m", ltrans_output_list);
|
|
|
|
timevar_push (TV_WHOPR_WPA);
|
|
|
|
FOR_EACH_VEC_ELT (ltrans_partitions, i, part)
|
|
lto_stats.num_output_symtab_nodes += lto_symtab_encoder_size (part->encoder);
|
|
|
|
/* Find out statics that need to be promoted
|
|
to globals with hidden visibility because they are accessed from multiple
|
|
partitions. */
|
|
lto_promote_cross_file_statics ();
|
|
|
|
timevar_pop (TV_WHOPR_WPA);
|
|
|
|
timevar_push (TV_WHOPR_WPA_IO);
|
|
|
|
/* Generate a prefix for the LTRANS unit files. */
|
|
blen = strlen (ltrans_output_list);
|
|
temp_filename = (char *) xmalloc (blen + sizeof ("2147483648.o"));
|
|
strcpy (temp_filename, ltrans_output_list);
|
|
if (blen > sizeof (".out")
|
|
&& strcmp (temp_filename + blen - sizeof (".out") + 1,
|
|
".out") == 0)
|
|
temp_filename[blen - sizeof (".out") + 1] = '\0';
|
|
blen = strlen (temp_filename);
|
|
|
|
n_sets = ltrans_partitions.length ();
|
|
|
|
/* Sort partitions by size so small ones are compiled last.
|
|
FIXME: Even when not reordering we may want to output one list for parallel make
|
|
and other for final link command. */
|
|
ltrans_partitions.qsort (flag_toplevel_reorder
|
|
? cmp_partitions_size
|
|
: cmp_partitions_order);
|
|
for (i = 0; i < n_sets; i++)
|
|
{
|
|
size_t len;
|
|
ltrans_partition part = ltrans_partitions[i];
|
|
|
|
/* Write all the nodes in SET. */
|
|
sprintf (temp_filename + blen, "%u.o", i);
|
|
file = lto_obj_file_open (temp_filename, true);
|
|
if (!file)
|
|
fatal_error ("lto_obj_file_open() failed");
|
|
|
|
if (!quiet_flag)
|
|
fprintf (stderr, " %s (%s %i insns)", temp_filename, part->name, part->insns);
|
|
if (cgraph_dump_file)
|
|
{
|
|
lto_symtab_encoder_iterator lsei;
|
|
|
|
fprintf (cgraph_dump_file, "Writing partition %s to file %s, %i insns\n",
|
|
part->name, temp_filename, part->insns);
|
|
fprintf (cgraph_dump_file, " Symbols in partition: ");
|
|
for (lsei = lsei_start_in_partition (part->encoder); !lsei_end_p (lsei);
|
|
lsei_next_in_partition (&lsei))
|
|
{
|
|
symtab_node node = lsei_node (lsei);
|
|
fprintf (cgraph_dump_file, "%s ", symtab_node_asm_name (node));
|
|
}
|
|
fprintf (cgraph_dump_file, "\n Symbols in boundary: ");
|
|
for (lsei = lsei_start (part->encoder); !lsei_end_p (lsei);
|
|
lsei_next (&lsei))
|
|
{
|
|
symtab_node node = lsei_node (lsei);
|
|
if (!lto_symtab_encoder_in_partition_p (part->encoder, node))
|
|
{
|
|
fprintf (cgraph_dump_file, "%s ", symtab_node_asm_name (node));
|
|
cgraph_node *cnode = dyn_cast <cgraph_node> (node);
|
|
if (cnode
|
|
&& lto_symtab_encoder_encode_body_p (part->encoder, cnode))
|
|
fprintf (cgraph_dump_file, "(body included)");
|
|
else
|
|
{
|
|
varpool_node *vnode = dyn_cast <varpool_node> (node);
|
|
if (vnode
|
|
&& lto_symtab_encoder_encode_initializer_p (part->encoder, vnode))
|
|
fprintf (cgraph_dump_file, "(initializer included)");
|
|
}
|
|
}
|
|
}
|
|
fprintf (cgraph_dump_file, "\n");
|
|
}
|
|
gcc_checking_assert (lto_symtab_encoder_size (part->encoder) || !i);
|
|
|
|
lto_set_current_out_file (file);
|
|
|
|
ipa_write_optimization_summaries (part->encoder);
|
|
|
|
lto_set_current_out_file (NULL);
|
|
lto_obj_file_close (file);
|
|
free (file);
|
|
part->encoder = NULL;
|
|
|
|
len = strlen (temp_filename);
|
|
if (fwrite (temp_filename, 1, len, ltrans_output_list_stream) < len
|
|
|| fwrite ("\n", 1, 1, ltrans_output_list_stream) < 1)
|
|
fatal_error ("writing to LTRANS output list %s: %m",
|
|
ltrans_output_list);
|
|
}
|
|
|
|
lto_stats.num_output_files += n_sets;
|
|
|
|
/* Close the LTRANS output list. */
|
|
if (fclose (ltrans_output_list_stream))
|
|
fatal_error ("closing LTRANS output list %s: %m", ltrans_output_list);
|
|
|
|
free_ltrans_partitions();
|
|
free (temp_filename);
|
|
|
|
timevar_pop (TV_WHOPR_WPA_IO);
|
|
}
|
|
|
|
|
|
/* If TT is a variable or function decl replace it with its
|
|
prevailing variant. */
|
|
#define LTO_SET_PREVAIL(tt) \
|
|
do {\
|
|
if ((tt) && VAR_OR_FUNCTION_DECL_P (tt)) \
|
|
tt = lto_symtab_prevailing_decl (tt); \
|
|
} while (0)
|
|
|
|
/* Ensure that TT isn't a replacable var of function decl. */
|
|
#define LTO_NO_PREVAIL(tt) \
|
|
gcc_assert (!(tt) || !VAR_OR_FUNCTION_DECL_P (tt))
|
|
|
|
/* Given a tree T replace all fields referring to variables or functions
|
|
with their prevailing variant. */
|
|
static void
|
|
lto_fixup_prevailing_decls (tree t)
|
|
{
|
|
enum tree_code code = TREE_CODE (t);
|
|
LTO_NO_PREVAIL (TREE_TYPE (t));
|
|
if (CODE_CONTAINS_STRUCT (code, TS_COMMON))
|
|
LTO_NO_PREVAIL (TREE_CHAIN (t));
|
|
if (DECL_P (t))
|
|
{
|
|
LTO_NO_PREVAIL (DECL_NAME (t));
|
|
LTO_SET_PREVAIL (DECL_CONTEXT (t));
|
|
if (CODE_CONTAINS_STRUCT (code, TS_DECL_COMMON))
|
|
{
|
|
LTO_SET_PREVAIL (DECL_SIZE (t));
|
|
LTO_SET_PREVAIL (DECL_SIZE_UNIT (t));
|
|
LTO_SET_PREVAIL (DECL_INITIAL (t));
|
|
LTO_NO_PREVAIL (DECL_ATTRIBUTES (t));
|
|
LTO_SET_PREVAIL (DECL_ABSTRACT_ORIGIN (t));
|
|
}
|
|
if (CODE_CONTAINS_STRUCT (code, TS_DECL_WITH_VIS))
|
|
{
|
|
LTO_NO_PREVAIL (t->decl_with_vis.assembler_name);
|
|
LTO_NO_PREVAIL (DECL_SECTION_NAME (t));
|
|
}
|
|
if (CODE_CONTAINS_STRUCT (code, TS_DECL_NON_COMMON))
|
|
{
|
|
LTO_NO_PREVAIL (DECL_ARGUMENT_FLD (t));
|
|
LTO_NO_PREVAIL (DECL_RESULT_FLD (t));
|
|
LTO_NO_PREVAIL (DECL_VINDEX (t));
|
|
}
|
|
if (CODE_CONTAINS_STRUCT (code, TS_FUNCTION_DECL))
|
|
LTO_SET_PREVAIL (DECL_FUNCTION_PERSONALITY (t));
|
|
if (CODE_CONTAINS_STRUCT (code, TS_FIELD_DECL))
|
|
{
|
|
LTO_NO_PREVAIL (DECL_FIELD_OFFSET (t));
|
|
LTO_NO_PREVAIL (DECL_BIT_FIELD_TYPE (t));
|
|
LTO_NO_PREVAIL (DECL_QUALIFIER (t));
|
|
LTO_NO_PREVAIL (DECL_FIELD_BIT_OFFSET (t));
|
|
LTO_NO_PREVAIL (DECL_FCONTEXT (t));
|
|
}
|
|
}
|
|
else if (TYPE_P (t))
|
|
{
|
|
LTO_NO_PREVAIL (TYPE_CACHED_VALUES (t));
|
|
LTO_SET_PREVAIL (TYPE_SIZE (t));
|
|
LTO_SET_PREVAIL (TYPE_SIZE_UNIT (t));
|
|
LTO_NO_PREVAIL (TYPE_ATTRIBUTES (t));
|
|
LTO_NO_PREVAIL (TYPE_NAME (t));
|
|
|
|
LTO_SET_PREVAIL (TYPE_MINVAL (t));
|
|
LTO_SET_PREVAIL (TYPE_MAXVAL (t));
|
|
LTO_SET_PREVAIL (t->type_non_common.binfo);
|
|
|
|
LTO_SET_PREVAIL (TYPE_CONTEXT (t));
|
|
|
|
LTO_NO_PREVAIL (TYPE_CANONICAL (t));
|
|
LTO_NO_PREVAIL (TYPE_MAIN_VARIANT (t));
|
|
LTO_NO_PREVAIL (TYPE_NEXT_VARIANT (t));
|
|
}
|
|
else if (EXPR_P (t))
|
|
{
|
|
int i;
|
|
for (i = TREE_OPERAND_LENGTH (t) - 1; i >= 0; --i)
|
|
LTO_SET_PREVAIL (TREE_OPERAND (t, i));
|
|
}
|
|
else
|
|
{
|
|
switch (code)
|
|
{
|
|
case TREE_LIST:
|
|
LTO_SET_PREVAIL (TREE_VALUE (t));
|
|
LTO_SET_PREVAIL (TREE_PURPOSE (t));
|
|
break;
|
|
default:
|
|
gcc_unreachable ();
|
|
}
|
|
}
|
|
}
|
|
#undef LTO_SET_PREVAIL
|
|
#undef LTO_NO_PREVAIL
|
|
|
|
/* Helper function of lto_fixup_decls. Walks the var and fn streams in STATE,
|
|
replaces var and function decls with the corresponding prevailing def. */
|
|
|
|
static void
|
|
lto_fixup_state (struct lto_in_decl_state *state)
|
|
{
|
|
unsigned i, si;
|
|
struct lto_tree_ref_table *table;
|
|
|
|
/* Although we only want to replace FUNCTION_DECLs and VAR_DECLs,
|
|
we still need to walk from all DECLs to find the reachable
|
|
FUNCTION_DECLs and VAR_DECLs. */
|
|
for (si = 0; si < LTO_N_DECL_STREAMS; si++)
|
|
{
|
|
table = &state->streams[si];
|
|
for (i = 0; i < table->size; i++)
|
|
{
|
|
tree *tp = table->trees + i;
|
|
if (VAR_OR_FUNCTION_DECL_P (*tp))
|
|
*tp = lto_symtab_prevailing_decl (*tp);
|
|
}
|
|
}
|
|
}
|
|
|
|
/* A callback of htab_traverse. Just extracts a state from SLOT
|
|
and calls lto_fixup_state. */
|
|
|
|
static int
|
|
lto_fixup_state_aux (void **slot, void *aux ATTRIBUTE_UNUSED)
|
|
{
|
|
struct lto_in_decl_state *state = (struct lto_in_decl_state *) *slot;
|
|
lto_fixup_state (state);
|
|
return 1;
|
|
}
|
|
|
|
/* Fix the decls from all FILES. Replaces each decl with the corresponding
|
|
prevailing one. */
|
|
|
|
static void
|
|
lto_fixup_decls (struct lto_file_decl_data **files)
|
|
{
|
|
unsigned int i;
|
|
htab_iterator hi;
|
|
tree t;
|
|
|
|
FOR_EACH_HTAB_ELEMENT (tree_with_vars, t, tree, hi)
|
|
lto_fixup_prevailing_decls (t);
|
|
|
|
for (i = 0; files[i]; i++)
|
|
{
|
|
struct lto_file_decl_data *file = files[i];
|
|
struct lto_in_decl_state *state = file->global_decl_state;
|
|
lto_fixup_state (state);
|
|
|
|
htab_traverse (file->function_decl_states, lto_fixup_state_aux, NULL);
|
|
}
|
|
}
|
|
|
|
static GTY((length ("lto_stats.num_input_files + 1"))) struct lto_file_decl_data **all_file_decl_data;
|
|
|
|
/* Turn file datas for sub files into a single array, so that they look
|
|
like separate files for further passes. */
|
|
|
|
static void
|
|
lto_flatten_files (struct lto_file_decl_data **orig, int count, int last_file_ix)
|
|
{
|
|
struct lto_file_decl_data *n, *next;
|
|
int i, k;
|
|
|
|
lto_stats.num_input_files = count;
|
|
all_file_decl_data
|
|
= ggc_alloc_cleared_vec_lto_file_decl_data_ptr (count + 1);
|
|
/* Set the hooks so that all of the ipa passes can read in their data. */
|
|
lto_set_in_hooks (all_file_decl_data, get_section_data, free_section_data);
|
|
for (i = 0, k = 0; i < last_file_ix; i++)
|
|
{
|
|
for (n = orig[i]; n != NULL; n = next)
|
|
{
|
|
all_file_decl_data[k++] = n;
|
|
next = n->next;
|
|
n->next = NULL;
|
|
}
|
|
}
|
|
all_file_decl_data[k] = NULL;
|
|
gcc_assert (k == count);
|
|
}
|
|
|
|
/* Input file data before flattening (i.e. splitting them to subfiles to support
|
|
incremental linking. */
|
|
static int real_file_count;
|
|
static GTY((length ("real_file_count + 1"))) struct lto_file_decl_data **real_file_decl_data;
|
|
|
|
/* Read all the symbols from the input files FNAMES. NFILES is the
|
|
number of files requested in the command line. Instantiate a
|
|
global call graph by aggregating all the sub-graphs found in each
|
|
file. */
|
|
|
|
static void
|
|
read_cgraph_and_symbols (unsigned nfiles, const char **fnames)
|
|
{
|
|
unsigned int i, last_file_ix;
|
|
FILE *resolution;
|
|
struct cgraph_node *node;
|
|
int count = 0;
|
|
struct lto_file_decl_data **decl_data;
|
|
|
|
init_cgraph ();
|
|
|
|
timevar_push (TV_IPA_LTO_DECL_IN);
|
|
|
|
real_file_decl_data
|
|
= decl_data = ggc_alloc_cleared_vec_lto_file_decl_data_ptr (nfiles + 1);
|
|
real_file_count = nfiles;
|
|
|
|
/* Read the resolution file. */
|
|
resolution = NULL;
|
|
if (resolution_file_name)
|
|
{
|
|
int t;
|
|
unsigned num_objects;
|
|
|
|
resolution = fopen (resolution_file_name, "r");
|
|
if (resolution == NULL)
|
|
fatal_error ("could not open symbol resolution file: %m");
|
|
|
|
t = fscanf (resolution, "%u", &num_objects);
|
|
gcc_assert (t == 1);
|
|
|
|
/* True, since the plugin splits the archives. */
|
|
gcc_assert (num_objects == nfiles);
|
|
}
|
|
|
|
tree_with_vars = htab_create_ggc (101, htab_hash_pointer, htab_eq_pointer,
|
|
NULL);
|
|
type_hash_cache = htab_create_ggc (512, tree_int_map_hash,
|
|
tree_int_map_eq, NULL);
|
|
type_pair_cache = XCNEWVEC (struct type_pair_d, GIMPLE_TYPE_PAIR_SIZE);
|
|
gimple_type_leader = ggc_alloc_cleared_vec_gimple_type_leader_entry_s
|
|
(GIMPLE_TYPE_LEADER_SIZE);
|
|
gimple_types = htab_create_ggc (16381, gimple_type_hash, gimple_type_eq, 0);
|
|
|
|
if (!quiet_flag)
|
|
fprintf (stderr, "Reading object files:");
|
|
|
|
/* Read all of the object files specified on the command line. */
|
|
for (i = 0, last_file_ix = 0; i < nfiles; ++i)
|
|
{
|
|
struct lto_file_decl_data *file_data = NULL;
|
|
if (!quiet_flag)
|
|
{
|
|
fprintf (stderr, " %s", fnames[i]);
|
|
fflush (stderr);
|
|
}
|
|
|
|
current_lto_file = lto_obj_file_open (fnames[i], false);
|
|
if (!current_lto_file)
|
|
break;
|
|
|
|
file_data = lto_file_read (current_lto_file, resolution, &count);
|
|
if (!file_data)
|
|
{
|
|
lto_obj_file_close (current_lto_file);
|
|
free (current_lto_file);
|
|
current_lto_file = NULL;
|
|
break;
|
|
}
|
|
|
|
decl_data[last_file_ix++] = file_data;
|
|
|
|
lto_obj_file_close (current_lto_file);
|
|
free (current_lto_file);
|
|
current_lto_file = NULL;
|
|
ggc_collect ();
|
|
}
|
|
|
|
lto_flatten_files (decl_data, count, last_file_ix);
|
|
lto_stats.num_input_files = count;
|
|
ggc_free(decl_data);
|
|
real_file_decl_data = NULL;
|
|
|
|
if (resolution_file_name)
|
|
fclose (resolution);
|
|
|
|
/* Free gimple type merging datastructures. */
|
|
htab_delete (gimple_types);
|
|
gimple_types = NULL;
|
|
htab_delete (type_hash_cache);
|
|
type_hash_cache = NULL;
|
|
free (type_pair_cache);
|
|
type_pair_cache = NULL;
|
|
gimple_type_leader = NULL;
|
|
free_gimple_type_tables ();
|
|
ggc_collect ();
|
|
|
|
/* Set the hooks so that all of the ipa passes can read in their data. */
|
|
lto_set_in_hooks (all_file_decl_data, get_section_data, free_section_data);
|
|
|
|
timevar_pop (TV_IPA_LTO_DECL_IN);
|
|
|
|
if (!quiet_flag)
|
|
fprintf (stderr, "\nReading the callgraph\n");
|
|
|
|
timevar_push (TV_IPA_LTO_CGRAPH_IO);
|
|
/* Read the symtab. */
|
|
input_symtab ();
|
|
|
|
/* Store resolutions into the symbol table. */
|
|
if (resolution_map)
|
|
{
|
|
void **res;
|
|
symtab_node snode;
|
|
|
|
FOR_EACH_SYMBOL (snode)
|
|
if (symtab_real_symbol_p (snode)
|
|
&& (res = pointer_map_contains (resolution_map,
|
|
snode->symbol.decl)))
|
|
snode->symbol.resolution
|
|
= (enum ld_plugin_symbol_resolution)(size_t)*res;
|
|
|
|
pointer_map_destroy (resolution_map);
|
|
resolution_map = NULL;
|
|
}
|
|
|
|
timevar_pop (TV_IPA_LTO_CGRAPH_IO);
|
|
|
|
if (!quiet_flag)
|
|
fprintf (stderr, "Merging declarations\n");
|
|
|
|
timevar_push (TV_IPA_LTO_DECL_MERGE);
|
|
/* Merge global decls. In ltrans mode we read merged cgraph, we do not
|
|
need to care about resolving symbols again, we only need to replace
|
|
duplicated declarations read from the callgraph and from function
|
|
sections. */
|
|
if (!flag_ltrans)
|
|
{
|
|
lto_symtab_merge_decls ();
|
|
|
|
/* If there were errors during symbol merging bail out, we have no
|
|
good way to recover here. */
|
|
if (seen_error ())
|
|
fatal_error ("errors during merging of translation units");
|
|
|
|
/* Fixup all decls. */
|
|
lto_fixup_decls (all_file_decl_data);
|
|
}
|
|
htab_delete (tree_with_vars);
|
|
tree_with_vars = NULL;
|
|
ggc_collect ();
|
|
|
|
timevar_pop (TV_IPA_LTO_DECL_MERGE);
|
|
/* Each pass will set the appropriate timer. */
|
|
|
|
if (!quiet_flag)
|
|
fprintf (stderr, "Reading summaries\n");
|
|
|
|
/* Read the IPA summary data. */
|
|
if (flag_ltrans)
|
|
ipa_read_optimization_summaries ();
|
|
else
|
|
ipa_read_summaries ();
|
|
|
|
for (i = 0; all_file_decl_data[i]; i++)
|
|
{
|
|
gcc_assert (all_file_decl_data[i]->symtab_node_encoder);
|
|
lto_symtab_encoder_delete (all_file_decl_data[i]->symtab_node_encoder);
|
|
all_file_decl_data[i]->symtab_node_encoder = NULL;
|
|
}
|
|
|
|
/* Finally merge the cgraph according to the decl merging decisions. */
|
|
timevar_push (TV_IPA_LTO_CGRAPH_MERGE);
|
|
if (cgraph_dump_file)
|
|
{
|
|
fprintf (cgraph_dump_file, "Before merging:\n");
|
|
dump_cgraph (cgraph_dump_file);
|
|
dump_varpool (cgraph_dump_file);
|
|
}
|
|
lto_symtab_merge_cgraph_nodes ();
|
|
ggc_collect ();
|
|
|
|
/* FIXME: ipa_transforms_to_apply holds list of passes that have optimization
|
|
summaries computed and needs to apply changes. At the moment WHOPR only
|
|
supports inlining, so we can push it here by hand. In future we need to stream
|
|
this field into ltrans compilation. */
|
|
if (flag_ltrans)
|
|
FOR_EACH_DEFINED_FUNCTION (node)
|
|
node->ipa_transforms_to_apply.safe_push ((ipa_opt_pass)&pass_ipa_inline);
|
|
|
|
timevar_pop (TV_IPA_LTO_CGRAPH_MERGE);
|
|
|
|
timevar_push (TV_IPA_LTO_DECL_INIT_IO);
|
|
|
|
/* Indicate that the cgraph is built and ready. */
|
|
cgraph_function_flags_ready = true;
|
|
|
|
timevar_pop (TV_IPA_LTO_DECL_INIT_IO);
|
|
ggc_free (all_file_decl_data);
|
|
all_file_decl_data = NULL;
|
|
}
|
|
|
|
|
|
/* Materialize all the bodies for all the nodes in the callgraph. */
|
|
|
|
static void
|
|
materialize_cgraph (void)
|
|
{
|
|
tree decl;
|
|
struct cgraph_node *node;
|
|
unsigned i;
|
|
timevar_id_t lto_timer;
|
|
|
|
if (!quiet_flag)
|
|
fprintf (stderr,
|
|
flag_wpa ? "Materializing decls:" : "Reading function bodies:");
|
|
|
|
/* Now that we have input the cgraph, we need to clear all of the aux
|
|
nodes and read the functions if we are not running in WPA mode. */
|
|
timevar_push (TV_IPA_LTO_GIMPLE_IN);
|
|
|
|
FOR_EACH_FUNCTION (node)
|
|
{
|
|
if (node->symbol.lto_file_data)
|
|
{
|
|
lto_materialize_function (node);
|
|
lto_stats.num_input_cgraph_nodes++;
|
|
}
|
|
}
|
|
|
|
timevar_pop (TV_IPA_LTO_GIMPLE_IN);
|
|
|
|
/* Start the appropriate timer depending on the mode that we are
|
|
operating in. */
|
|
lto_timer = (flag_wpa) ? TV_WHOPR_WPA
|
|
: (flag_ltrans) ? TV_WHOPR_LTRANS
|
|
: TV_LTO;
|
|
timevar_push (lto_timer);
|
|
|
|
current_function_decl = NULL;
|
|
set_cfun (NULL);
|
|
|
|
/* Inform the middle end about the global variables we have seen. */
|
|
FOR_EACH_VEC_ELT (*lto_global_var_decls, i, decl)
|
|
rest_of_decl_compilation (decl, 1, 0);
|
|
|
|
if (!quiet_flag)
|
|
fprintf (stderr, "\n");
|
|
|
|
timevar_pop (lto_timer);
|
|
}
|
|
|
|
|
|
/* Show various memory usage statistics related to LTO. */
|
|
static void
|
|
print_lto_report_1 (void)
|
|
{
|
|
const char *pfx = (flag_lto) ? "LTO" : (flag_wpa) ? "WPA" : "LTRANS";
|
|
fprintf (stderr, "%s statistics\n", pfx);
|
|
|
|
if (gimple_types)
|
|
fprintf (stderr, "[%s] GIMPLE type table: size %ld, %ld elements, "
|
|
"%ld searches, %ld collisions (ratio: %f)\n", pfx,
|
|
(long) htab_size (gimple_types),
|
|
(long) htab_elements (gimple_types),
|
|
(long) gimple_types->searches,
|
|
(long) gimple_types->collisions,
|
|
htab_collisions (gimple_types));
|
|
else
|
|
fprintf (stderr, "[%s] GIMPLE type table is empty\n", pfx);
|
|
if (type_hash_cache)
|
|
fprintf (stderr, "[%s] GIMPLE type hash table: size %ld, %ld elements, "
|
|
"%ld searches, %ld collisions (ratio: %f)\n", pfx,
|
|
(long) htab_size (type_hash_cache),
|
|
(long) htab_elements (type_hash_cache),
|
|
(long) type_hash_cache->searches,
|
|
(long) type_hash_cache->collisions,
|
|
htab_collisions (type_hash_cache));
|
|
else
|
|
fprintf (stderr, "[%s] GIMPLE type hash table is empty\n", pfx);
|
|
|
|
print_gimple_types_stats (pfx);
|
|
print_lto_report (pfx);
|
|
}
|
|
|
|
/* Perform whole program analysis (WPA) on the callgraph and write out the
|
|
optimization plan. */
|
|
|
|
static void
|
|
do_whole_program_analysis (void)
|
|
{
|
|
symtab_node node;
|
|
|
|
timevar_start (TV_PHASE_OPT_GEN);
|
|
|
|
/* Note that since we are in WPA mode, materialize_cgraph will not
|
|
actually read in all the function bodies. It only materializes
|
|
the decls and cgraph nodes so that analysis can be performed. */
|
|
materialize_cgraph ();
|
|
|
|
/* Reading in the cgraph uses different timers, start timing WPA now. */
|
|
timevar_push (TV_WHOPR_WPA);
|
|
|
|
if (pre_ipa_mem_report)
|
|
{
|
|
fprintf (stderr, "Memory consumption before IPA\n");
|
|
dump_memory_report (false);
|
|
}
|
|
|
|
cgraph_function_flags_ready = true;
|
|
|
|
if (cgraph_dump_file)
|
|
{
|
|
dump_cgraph (cgraph_dump_file);
|
|
dump_varpool (cgraph_dump_file);
|
|
}
|
|
bitmap_obstack_initialize (NULL);
|
|
cgraph_state = CGRAPH_STATE_IPA_SSA;
|
|
|
|
execute_ipa_pass_list (all_regular_ipa_passes);
|
|
symtab_remove_unreachable_nodes (false, dump_file);
|
|
|
|
if (cgraph_dump_file)
|
|
{
|
|
fprintf (cgraph_dump_file, "Optimized ");
|
|
dump_cgraph (cgraph_dump_file);
|
|
dump_varpool (cgraph_dump_file);
|
|
}
|
|
#ifdef ENABLE_CHECKING
|
|
verify_cgraph ();
|
|
#endif
|
|
bitmap_obstack_release (NULL);
|
|
|
|
/* We are about to launch the final LTRANS phase, stop the WPA timer. */
|
|
timevar_pop (TV_WHOPR_WPA);
|
|
|
|
timevar_push (TV_WHOPR_PARTITIONING);
|
|
if (flag_lto_partition_1to1)
|
|
lto_1_to_1_map ();
|
|
else if (flag_lto_partition_max)
|
|
lto_max_map ();
|
|
else
|
|
lto_balanced_map ();
|
|
|
|
/* AUX pointers are used by partitioning code to bookkeep number of
|
|
partitions symbol is in. This is no longer needed. */
|
|
FOR_EACH_SYMBOL (node)
|
|
node->symbol.aux = NULL;
|
|
|
|
lto_stats.num_cgraph_partitions += ltrans_partitions.length ();
|
|
timevar_pop (TV_WHOPR_PARTITIONING);
|
|
|
|
timevar_stop (TV_PHASE_OPT_GEN);
|
|
timevar_start (TV_PHASE_STREAM_OUT);
|
|
|
|
if (!quiet_flag)
|
|
{
|
|
fprintf (stderr, "\nStreaming out");
|
|
fflush (stderr);
|
|
}
|
|
lto_wpa_write_files ();
|
|
if (!quiet_flag)
|
|
fprintf (stderr, "\n");
|
|
|
|
timevar_stop (TV_PHASE_STREAM_OUT);
|
|
|
|
ggc_collect ();
|
|
if (post_ipa_mem_report)
|
|
{
|
|
fprintf (stderr, "Memory consumption after IPA\n");
|
|
dump_memory_report (false);
|
|
}
|
|
|
|
/* Show the LTO report before launching LTRANS. */
|
|
if (flag_lto_report)
|
|
print_lto_report_1 ();
|
|
if (mem_report_wpa)
|
|
dump_memory_report (true);
|
|
}
|
|
|
|
|
|
static GTY(()) tree lto_eh_personality_decl;
|
|
|
|
/* Return the LTO personality function decl. */
|
|
|
|
tree
|
|
lto_eh_personality (void)
|
|
{
|
|
if (!lto_eh_personality_decl)
|
|
{
|
|
/* Use the first personality DECL for our personality if we don't
|
|
support multiple ones. This ensures that we don't artificially
|
|
create the need for them in a single-language program. */
|
|
if (first_personality_decl && !dwarf2out_do_cfi_asm ())
|
|
lto_eh_personality_decl = first_personality_decl;
|
|
else
|
|
lto_eh_personality_decl = lhd_gcc_personality ();
|
|
}
|
|
|
|
return lto_eh_personality_decl;
|
|
}
|
|
|
|
/* Set the process name based on the LTO mode. */
|
|
|
|
static void
|
|
lto_process_name (void)
|
|
{
|
|
if (flag_lto)
|
|
setproctitle ("lto1-lto");
|
|
if (flag_wpa)
|
|
setproctitle ("lto1-wpa");
|
|
if (flag_ltrans)
|
|
setproctitle ("lto1-ltrans");
|
|
}
|
|
|
|
|
|
/* Initialize the LTO front end. */
|
|
|
|
static void
|
|
lto_init (void)
|
|
{
|
|
lto_process_name ();
|
|
lto_streamer_hooks_init ();
|
|
lto_reader_init ();
|
|
lto_set_in_hooks (NULL, get_section_data, free_section_data);
|
|
memset (<o_stats, 0, sizeof (lto_stats));
|
|
bitmap_obstack_initialize (NULL);
|
|
gimple_register_cfg_hooks ();
|
|
}
|
|
|
|
|
|
/* Main entry point for the GIMPLE front end. This front end has
|
|
three main personalities:
|
|
|
|
- LTO (-flto). All the object files on the command line are
|
|
loaded in memory and processed as a single translation unit.
|
|
This is the traditional link-time optimization behavior.
|
|
|
|
- WPA (-fwpa). Only the callgraph and summary information for
|
|
files in the command file are loaded. A single callgraph
|
|
(without function bodies) is instantiated for the whole set of
|
|
files. IPA passes are only allowed to analyze the call graph
|
|
and make transformation decisions. The callgraph is
|
|
partitioned, each partition is written to a new object file
|
|
together with the transformation decisions.
|
|
|
|
- LTRANS (-fltrans). Similar to -flto but it prevents the IPA
|
|
summary files from running again. Since WPA computed summary
|
|
information and decided what transformations to apply, LTRANS
|
|
simply applies them. */
|
|
|
|
void
|
|
lto_main (void)
|
|
{
|
|
/* LTO is called as a front end, even though it is not a front end.
|
|
Because it is called as a front end, TV_PHASE_PARSING and
|
|
TV_PARSE_GLOBAL are active, and we need to turn them off while
|
|
doing LTO. Later we turn them back on so they are active up in
|
|
toplev.c. */
|
|
timevar_pop (TV_PARSE_GLOBAL);
|
|
timevar_stop (TV_PHASE_PARSING);
|
|
|
|
timevar_start (TV_PHASE_SETUP);
|
|
|
|
/* Initialize the LTO front end. */
|
|
lto_init ();
|
|
|
|
timevar_stop (TV_PHASE_SETUP);
|
|
timevar_start (TV_PHASE_STREAM_IN);
|
|
|
|
/* Read all the symbols and call graph from all the files in the
|
|
command line. */
|
|
read_cgraph_and_symbols (num_in_fnames, in_fnames);
|
|
|
|
timevar_stop (TV_PHASE_STREAM_IN);
|
|
|
|
if (!seen_error ())
|
|
{
|
|
/* If WPA is enabled analyze the whole call graph and create an
|
|
optimization plan. Otherwise, read in all the function
|
|
bodies and continue with optimization. */
|
|
if (flag_wpa)
|
|
do_whole_program_analysis ();
|
|
else
|
|
{
|
|
struct varpool_node *vnode;
|
|
|
|
timevar_start (TV_PHASE_OPT_GEN);
|
|
|
|
materialize_cgraph ();
|
|
|
|
/* Let the middle end know that we have read and merged all of
|
|
the input files. */
|
|
compile ();
|
|
|
|
timevar_stop (TV_PHASE_OPT_GEN);
|
|
|
|
/* FIXME lto, if the processes spawned by WPA fail, we miss
|
|
the chance to print WPA's report, so WPA will call
|
|
print_lto_report before launching LTRANS. If LTRANS was
|
|
launched directly by the driver we would not need to do
|
|
this. */
|
|
if (flag_lto_report)
|
|
print_lto_report_1 ();
|
|
|
|
/* Record the global variables. */
|
|
FOR_EACH_DEFINED_VARIABLE (vnode)
|
|
vec_safe_push (lto_global_var_decls, vnode->symbol.decl);
|
|
}
|
|
}
|
|
|
|
/* Here we make LTO pretend to be a parser. */
|
|
timevar_start (TV_PHASE_PARSING);
|
|
timevar_push (TV_PARSE_GLOBAL);
|
|
}
|
|
|
|
#include "gt-lto-lto.h"
|