4624 lines
156 KiB
C
4624 lines
156 KiB
C
|
/* Register Transfer Language (RTL) definitions for GCC
|
|||
|
Copyright (C) 1987-2023 Free Software Foundation, Inc.
|
|||
|
|
|||
|
This file is part of GCC.
|
|||
|
|
|||
|
GCC is free software; you can redistribute it and/or modify it under
|
|||
|
the terms of the GNU General Public License as published by the Free
|
|||
|
Software Foundation; either version 3, or (at your option) any later
|
|||
|
version.
|
|||
|
|
|||
|
GCC is distributed in the hope that it will be useful, but WITHOUT ANY
|
|||
|
WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|||
|
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|||
|
for more details.
|
|||
|
|
|||
|
You should have received a copy of the GNU General Public License
|
|||
|
along with GCC; see the file COPYING3. If not see
|
|||
|
<http://www.gnu.org/licenses/>. */
|
|||
|
|
|||
|
#ifndef GCC_RTL_H
|
|||
|
#define GCC_RTL_H
|
|||
|
|
|||
|
/* This file is occasionally included by generator files which expect
|
|||
|
machmode.h and other files to exist and would not normally have been
|
|||
|
included by coretypes.h. */
|
|||
|
#ifdef GENERATOR_FILE
|
|||
|
#include "real.h"
|
|||
|
#include "fixed-value.h"
|
|||
|
#include "statistics.h"
|
|||
|
#include "vec.h"
|
|||
|
#include "hash-table.h"
|
|||
|
#include "hash-set.h"
|
|||
|
#include "input.h"
|
|||
|
#include "is-a.h"
|
|||
|
#endif /* GENERATOR_FILE */
|
|||
|
|
|||
|
#include "hard-reg-set.h"
|
|||
|
|
|||
|
class predefined_function_abi;
|
|||
|
|
|||
|
/* Value used by some passes to "recognize" noop moves as valid
|
|||
|
instructions. */
|
|||
|
#define NOOP_MOVE_INSN_CODE INT_MAX
|
|||
|
|
|||
|
/* Register Transfer Language EXPRESSIONS CODES */
|
|||
|
|
|||
|
#define RTX_CODE enum rtx_code
|
|||
|
enum rtx_code {
|
|||
|
|
|||
|
#define DEF_RTL_EXPR(ENUM, NAME, FORMAT, CLASS) ENUM ,
|
|||
|
#include "rtl.def" /* rtl expressions are documented here */
|
|||
|
#undef DEF_RTL_EXPR
|
|||
|
|
|||
|
LAST_AND_UNUSED_RTX_CODE}; /* A convenient way to get a value for
|
|||
|
NUM_RTX_CODE.
|
|||
|
Assumes default enum value assignment. */
|
|||
|
|
|||
|
/* The cast here, saves many elsewhere. */
|
|||
|
#define NUM_RTX_CODE ((int) LAST_AND_UNUSED_RTX_CODE)
|
|||
|
|
|||
|
/* Similar, but since generator files get more entries... */
|
|||
|
#ifdef GENERATOR_FILE
|
|||
|
# define NON_GENERATOR_NUM_RTX_CODE ((int) MATCH_OPERAND)
|
|||
|
#endif
|
|||
|
|
|||
|
/* Register Transfer Language EXPRESSIONS CODE CLASSES */
|
|||
|
|
|||
|
enum rtx_class {
|
|||
|
/* We check bit 0-1 of some rtx class codes in the predicates below. */
|
|||
|
|
|||
|
/* Bit 0 = comparison if 0, arithmetic is 1
|
|||
|
Bit 1 = 1 if commutative. */
|
|||
|
RTX_COMPARE, /* 0 */
|
|||
|
RTX_COMM_COMPARE,
|
|||
|
RTX_BIN_ARITH,
|
|||
|
RTX_COMM_ARITH,
|
|||
|
|
|||
|
/* Must follow the four preceding values. */
|
|||
|
RTX_UNARY, /* 4 */
|
|||
|
|
|||
|
RTX_EXTRA,
|
|||
|
RTX_MATCH,
|
|||
|
RTX_INSN,
|
|||
|
|
|||
|
/* Bit 0 = 1 if constant. */
|
|||
|
RTX_OBJ, /* 8 */
|
|||
|
RTX_CONST_OBJ,
|
|||
|
|
|||
|
RTX_TERNARY,
|
|||
|
RTX_BITFIELD_OPS,
|
|||
|
RTX_AUTOINC
|
|||
|
};
|
|||
|
|
|||
|
#define RTX_OBJ_MASK (~1)
|
|||
|
#define RTX_OBJ_RESULT (RTX_OBJ & RTX_OBJ_MASK)
|
|||
|
#define RTX_COMPARE_MASK (~1)
|
|||
|
#define RTX_COMPARE_RESULT (RTX_COMPARE & RTX_COMPARE_MASK)
|
|||
|
#define RTX_ARITHMETIC_MASK (~1)
|
|||
|
#define RTX_ARITHMETIC_RESULT (RTX_COMM_ARITH & RTX_ARITHMETIC_MASK)
|
|||
|
#define RTX_BINARY_MASK (~3)
|
|||
|
#define RTX_BINARY_RESULT (RTX_COMPARE & RTX_BINARY_MASK)
|
|||
|
#define RTX_COMMUTATIVE_MASK (~2)
|
|||
|
#define RTX_COMMUTATIVE_RESULT (RTX_COMM_COMPARE & RTX_COMMUTATIVE_MASK)
|
|||
|
#define RTX_NON_COMMUTATIVE_RESULT (RTX_COMPARE & RTX_COMMUTATIVE_MASK)
|
|||
|
|
|||
|
extern const unsigned char rtx_length[NUM_RTX_CODE];
|
|||
|
#define GET_RTX_LENGTH(CODE) (rtx_length[(int) (CODE)])
|
|||
|
|
|||
|
extern const char * const rtx_name[NUM_RTX_CODE];
|
|||
|
#define GET_RTX_NAME(CODE) (rtx_name[(int) (CODE)])
|
|||
|
|
|||
|
extern const char * const rtx_format[NUM_RTX_CODE];
|
|||
|
#define GET_RTX_FORMAT(CODE) (rtx_format[(int) (CODE)])
|
|||
|
|
|||
|
extern const enum rtx_class rtx_class[NUM_RTX_CODE];
|
|||
|
#define GET_RTX_CLASS(CODE) (rtx_class[(int) (CODE)])
|
|||
|
|
|||
|
/* True if CODE is part of the insn chain (i.e. has INSN_UID, PREV_INSN
|
|||
|
and NEXT_INSN fields). */
|
|||
|
#define INSN_CHAIN_CODE_P(CODE) IN_RANGE (CODE, DEBUG_INSN, NOTE)
|
|||
|
|
|||
|
extern const unsigned char rtx_code_size[NUM_RTX_CODE];
|
|||
|
extern const unsigned char rtx_next[NUM_RTX_CODE];
|
|||
|
|
|||
|
/* The flags and bitfields of an ADDR_DIFF_VEC. BASE is the base label
|
|||
|
relative to which the offsets are calculated, as explained in rtl.def. */
|
|||
|
struct addr_diff_vec_flags
|
|||
|
{
|
|||
|
/* Set at the start of shorten_branches - ONLY WHEN OPTIMIZING - : */
|
|||
|
unsigned min_align: 8;
|
|||
|
/* Flags: */
|
|||
|
unsigned base_after_vec: 1; /* BASE is after the ADDR_DIFF_VEC. */
|
|||
|
unsigned min_after_vec: 1; /* minimum address target label is
|
|||
|
after the ADDR_DIFF_VEC. */
|
|||
|
unsigned max_after_vec: 1; /* maximum address target label is
|
|||
|
after the ADDR_DIFF_VEC. */
|
|||
|
unsigned min_after_base: 1; /* minimum address target label is
|
|||
|
after BASE. */
|
|||
|
unsigned max_after_base: 1; /* maximum address target label is
|
|||
|
after BASE. */
|
|||
|
/* Set by the actual branch shortening process - ONLY WHEN OPTIMIZING - : */
|
|||
|
unsigned offset_unsigned: 1; /* offsets have to be treated as unsigned. */
|
|||
|
unsigned : 2;
|
|||
|
unsigned scale : 8;
|
|||
|
};
|
|||
|
|
|||
|
/* Structure used to describe the attributes of a MEM. These are hashed
|
|||
|
so MEMs that the same attributes share a data structure. This means
|
|||
|
they cannot be modified in place. */
|
|||
|
class GTY(()) mem_attrs
|
|||
|
{
|
|||
|
public:
|
|||
|
mem_attrs ();
|
|||
|
|
|||
|
/* The expression that the MEM accesses, or null if not known.
|
|||
|
This expression might be larger than the memory reference itself.
|
|||
|
(In other words, the MEM might access only part of the object.) */
|
|||
|
tree expr;
|
|||
|
|
|||
|
/* The offset of the memory reference from the start of EXPR.
|
|||
|
Only valid if OFFSET_KNOWN_P. */
|
|||
|
poly_int64 offset;
|
|||
|
|
|||
|
/* The size of the memory reference in bytes. Only valid if
|
|||
|
SIZE_KNOWN_P. */
|
|||
|
poly_int64 size;
|
|||
|
|
|||
|
/* The alias set of the memory reference. */
|
|||
|
alias_set_type alias;
|
|||
|
|
|||
|
/* The alignment of the reference in bits. Always a multiple of
|
|||
|
BITS_PER_UNIT. Note that EXPR may have a stricter alignment
|
|||
|
than the memory reference itself. */
|
|||
|
unsigned int align;
|
|||
|
|
|||
|
/* The address space that the memory reference uses. */
|
|||
|
unsigned char addrspace;
|
|||
|
|
|||
|
/* True if OFFSET is known. */
|
|||
|
bool offset_known_p;
|
|||
|
|
|||
|
/* True if SIZE is known. */
|
|||
|
bool size_known_p;
|
|||
|
};
|
|||
|
|
|||
|
/* Structure used to describe the attributes of a REG in similar way as
|
|||
|
mem_attrs does for MEM above. Note that the OFFSET field is calculated
|
|||
|
in the same way as for mem_attrs, rather than in the same way as a
|
|||
|
SUBREG_BYTE. For example, if a big-endian target stores a byte
|
|||
|
object in the low part of a 4-byte register, the OFFSET field
|
|||
|
will be -3 rather than 0. */
|
|||
|
|
|||
|
class GTY((for_user)) reg_attrs {
|
|||
|
public:
|
|||
|
tree decl; /* decl corresponding to REG. */
|
|||
|
poly_int64 offset; /* Offset from start of DECL. */
|
|||
|
};
|
|||
|
|
|||
|
/* Common union for an element of an rtx. */
|
|||
|
|
|||
|
union rtunion
|
|||
|
{
|
|||
|
int rt_int;
|
|||
|
unsigned int rt_uint;
|
|||
|
poly_uint16_pod rt_subreg;
|
|||
|
const char *rt_str;
|
|||
|
rtx rt_rtx;
|
|||
|
rtvec rt_rtvec;
|
|||
|
machine_mode rt_type;
|
|||
|
addr_diff_vec_flags rt_addr_diff_vec_flags;
|
|||
|
struct cselib_val *rt_cselib;
|
|||
|
tree rt_tree;
|
|||
|
basic_block rt_bb;
|
|||
|
mem_attrs *rt_mem;
|
|||
|
class constant_descriptor_rtx *rt_constant;
|
|||
|
struct dw_cfi_node *rt_cfi;
|
|||
|
};
|
|||
|
|
|||
|
/* Describes the properties of a REG. */
|
|||
|
struct GTY(()) reg_info {
|
|||
|
/* The value of REGNO. */
|
|||
|
unsigned int regno;
|
|||
|
|
|||
|
/* The value of REG_NREGS. */
|
|||
|
unsigned int nregs : 8;
|
|||
|
unsigned int unused : 24;
|
|||
|
|
|||
|
/* The value of REG_ATTRS. */
|
|||
|
reg_attrs *attrs;
|
|||
|
};
|
|||
|
|
|||
|
/* This structure remembers the position of a SYMBOL_REF within an
|
|||
|
object_block structure. A SYMBOL_REF only provides this information
|
|||
|
if SYMBOL_REF_HAS_BLOCK_INFO_P is true. */
|
|||
|
struct GTY(()) block_symbol {
|
|||
|
/* The usual SYMBOL_REF fields. */
|
|||
|
rtunion GTY ((skip)) fld[2];
|
|||
|
|
|||
|
/* The block that contains this object. */
|
|||
|
struct object_block *block;
|
|||
|
|
|||
|
/* The offset of this object from the start of its block. It is negative
|
|||
|
if the symbol has not yet been assigned an offset. */
|
|||
|
HOST_WIDE_INT offset;
|
|||
|
};
|
|||
|
|
|||
|
/* Describes a group of objects that are to be placed together in such
|
|||
|
a way that their relative positions are known. */
|
|||
|
struct GTY((for_user)) object_block {
|
|||
|
/* The section in which these objects should be placed. */
|
|||
|
section *sect;
|
|||
|
|
|||
|
/* The alignment of the first object, measured in bits. */
|
|||
|
unsigned int alignment;
|
|||
|
|
|||
|
/* The total size of the objects, measured in bytes. */
|
|||
|
HOST_WIDE_INT size;
|
|||
|
|
|||
|
/* The SYMBOL_REFs for each object. The vector is sorted in
|
|||
|
order of increasing offset and the following conditions will
|
|||
|
hold for each element X:
|
|||
|
|
|||
|
SYMBOL_REF_HAS_BLOCK_INFO_P (X)
|
|||
|
!SYMBOL_REF_ANCHOR_P (X)
|
|||
|
SYMBOL_REF_BLOCK (X) == [address of this structure]
|
|||
|
SYMBOL_REF_BLOCK_OFFSET (X) >= 0. */
|
|||
|
vec<rtx, va_gc> *objects;
|
|||
|
|
|||
|
/* All the anchor SYMBOL_REFs used to address these objects, sorted
|
|||
|
in order of increasing offset, and then increasing TLS model.
|
|||
|
The following conditions will hold for each element X in this vector:
|
|||
|
|
|||
|
SYMBOL_REF_HAS_BLOCK_INFO_P (X)
|
|||
|
SYMBOL_REF_ANCHOR_P (X)
|
|||
|
SYMBOL_REF_BLOCK (X) == [address of this structure]
|
|||
|
SYMBOL_REF_BLOCK_OFFSET (X) >= 0. */
|
|||
|
vec<rtx, va_gc> *anchors;
|
|||
|
};
|
|||
|
|
|||
|
struct GTY((variable_size)) hwivec_def {
|
|||
|
HOST_WIDE_INT elem[1];
|
|||
|
};
|
|||
|
|
|||
|
/* Number of elements of the HWIVEC if RTX is a CONST_WIDE_INT. */
|
|||
|
#define CWI_GET_NUM_ELEM(RTX) \
|
|||
|
((int)RTL_FLAG_CHECK1("CWI_GET_NUM_ELEM", (RTX), CONST_WIDE_INT)->u2.num_elem)
|
|||
|
#define CWI_PUT_NUM_ELEM(RTX, NUM) \
|
|||
|
(RTL_FLAG_CHECK1("CWI_PUT_NUM_ELEM", (RTX), CONST_WIDE_INT)->u2.num_elem = (NUM))
|
|||
|
|
|||
|
struct GTY((variable_size)) const_poly_int_def {
|
|||
|
trailing_wide_ints<NUM_POLY_INT_COEFFS> coeffs;
|
|||
|
};
|
|||
|
|
|||
|
/* RTL expression ("rtx"). */
|
|||
|
|
|||
|
/* The GTY "desc" and "tag" options below are a kludge: we need a desc
|
|||
|
field for gengtype to recognize that inheritance is occurring,
|
|||
|
so that all subclasses are redirected to the traversal hook for the
|
|||
|
base class.
|
|||
|
However, all of the fields are in the base class, and special-casing
|
|||
|
is at work. Hence we use desc and tag of 0, generating a switch
|
|||
|
statement of the form:
|
|||
|
switch (0)
|
|||
|
{
|
|||
|
case 0: // all the work happens here
|
|||
|
}
|
|||
|
in order to work with the existing special-casing in gengtype. */
|
|||
|
|
|||
|
struct GTY((desc("0"), tag("0"),
|
|||
|
chain_next ("RTX_NEXT (&%h)"),
|
|||
|
chain_prev ("RTX_PREV (&%h)"))) rtx_def {
|
|||
|
/* The kind of expression this is. */
|
|||
|
ENUM_BITFIELD(rtx_code) code: 16;
|
|||
|
|
|||
|
/* The kind of value the expression has. */
|
|||
|
ENUM_BITFIELD(machine_mode) mode : 8;
|
|||
|
|
|||
|
/* 1 in a MEM if we should keep the alias set for this mem unchanged
|
|||
|
when we access a component.
|
|||
|
1 in a JUMP_INSN if it is a crossing jump.
|
|||
|
1 in a CALL_INSN if it is a sibling call.
|
|||
|
1 in a SET that is for a return.
|
|||
|
In a CODE_LABEL, part of the two-bit alternate entry field.
|
|||
|
1 in a CONCAT is VAL_EXPR_IS_COPIED in var-tracking.cc.
|
|||
|
1 in a VALUE is SP_BASED_VALUE_P in cselib.cc.
|
|||
|
1 in a SUBREG generated by LRA for reload insns.
|
|||
|
1 in a REG if this is a static chain register.
|
|||
|
Dumped as "/j" in RTL dumps. */
|
|||
|
unsigned int jump : 1;
|
|||
|
/* In a CODE_LABEL, part of the two-bit alternate entry field.
|
|||
|
1 in a MEM if it cannot trap.
|
|||
|
1 in a CALL_INSN logically equivalent to
|
|||
|
ECF_LOOPING_CONST_OR_PURE and DECL_LOOPING_CONST_OR_PURE_P.
|
|||
|
1 in a VALUE is SP_DERIVED_VALUE_P in cselib.cc.
|
|||
|
Dumped as "/c" in RTL dumps. */
|
|||
|
unsigned int call : 1;
|
|||
|
/* 1 in a REG, MEM, or CONCAT if the value is set at most once, anywhere.
|
|||
|
1 in a SUBREG used for SUBREG_PROMOTED_UNSIGNED_P.
|
|||
|
1 in a SYMBOL_REF if it addresses something in the per-function
|
|||
|
constants pool.
|
|||
|
1 in a CALL_INSN logically equivalent to ECF_CONST and TREE_READONLY.
|
|||
|
1 in a NOTE, or EXPR_LIST for a const call.
|
|||
|
1 in a JUMP_INSN of an annulling branch.
|
|||
|
1 in a CONCAT is VAL_EXPR_IS_CLOBBERED in var-tracking.cc.
|
|||
|
1 in a preserved VALUE is PRESERVED_VALUE_P in cselib.cc.
|
|||
|
1 in a clobber temporarily created for LRA.
|
|||
|
Dumped as "/u" in RTL dumps. */
|
|||
|
unsigned int unchanging : 1;
|
|||
|
/* 1 in a MEM or ASM_OPERANDS expression if the memory reference is volatile.
|
|||
|
1 in an INSN, CALL_INSN, JUMP_INSN, CODE_LABEL, BARRIER, or NOTE
|
|||
|
if it has been deleted.
|
|||
|
1 in a REG expression if corresponds to a variable declared by the user,
|
|||
|
0 for an internally generated temporary.
|
|||
|
1 in a SUBREG used for SUBREG_PROMOTED_UNSIGNED_P.
|
|||
|
1 in a LABEL_REF, REG_LABEL_TARGET or REG_LABEL_OPERAND note for a
|
|||
|
non-local label.
|
|||
|
In a SYMBOL_REF, this flag is used for machine-specific purposes.
|
|||
|
In a PREFETCH, this flag indicates that it should be considered a
|
|||
|
scheduling barrier.
|
|||
|
1 in a CONCAT is VAL_NEEDS_RESOLUTION in var-tracking.cc.
|
|||
|
Dumped as "/v" in RTL dumps. */
|
|||
|
unsigned int volatil : 1;
|
|||
|
/* 1 in a REG if the register is used only in exit code a loop.
|
|||
|
1 in a SUBREG expression if was generated from a variable with a
|
|||
|
promoted mode.
|
|||
|
1 in a CODE_LABEL if the label is used for nonlocal gotos
|
|||
|
and must not be deleted even if its count is zero.
|
|||
|
1 in an INSN, JUMP_INSN or CALL_INSN if this insn must be scheduled
|
|||
|
together with the preceding insn. Valid only within sched.
|
|||
|
1 in an INSN, JUMP_INSN, or CALL_INSN if insn is in a delay slot and
|
|||
|
from the target of a branch. Valid from reorg until end of compilation;
|
|||
|
cleared before used.
|
|||
|
|
|||
|
The name of the field is historical. It used to be used in MEMs
|
|||
|
to record whether the MEM accessed part of a structure.
|
|||
|
Dumped as "/s" in RTL dumps. */
|
|||
|
unsigned int in_struct : 1;
|
|||
|
/* At the end of RTL generation, 1 if this rtx is used. This is used for
|
|||
|
copying shared structure. See `unshare_all_rtl'.
|
|||
|
In a REG, this is not needed for that purpose, and used instead
|
|||
|
in `leaf_renumber_regs_insn'.
|
|||
|
1 in a SYMBOL_REF, means that emit_library_call
|
|||
|
has used it as the function.
|
|||
|
1 in a CONCAT is VAL_HOLDS_TRACK_EXPR in var-tracking.cc.
|
|||
|
1 in a VALUE or DEBUG_EXPR is VALUE_RECURSED_INTO in var-tracking.cc. */
|
|||
|
unsigned int used : 1;
|
|||
|
/* 1 in an INSN or a SET if this rtx is related to the call frame,
|
|||
|
either changing how we compute the frame address or saving and
|
|||
|
restoring registers in the prologue and epilogue.
|
|||
|
1 in a REG or MEM if it is a pointer.
|
|||
|
1 in a SYMBOL_REF if it addresses something in the per-function
|
|||
|
constant string pool.
|
|||
|
1 in a VALUE is VALUE_CHANGED in var-tracking.cc.
|
|||
|
Dumped as "/f" in RTL dumps. */
|
|||
|
unsigned frame_related : 1;
|
|||
|
/* 1 in a REG or PARALLEL that is the current function's return value.
|
|||
|
1 in a SYMBOL_REF for a weak symbol.
|
|||
|
1 in a CALL_INSN logically equivalent to ECF_PURE and DECL_PURE_P.
|
|||
|
1 in a CONCAT is VAL_EXPR_HAS_REVERSE in var-tracking.cc.
|
|||
|
1 in a VALUE or DEBUG_EXPR is NO_LOC_P in var-tracking.cc.
|
|||
|
Dumped as "/i" in RTL dumps. */
|
|||
|
unsigned return_val : 1;
|
|||
|
|
|||
|
union {
|
|||
|
/* The final union field is aligned to 64 bits on LP64 hosts,
|
|||
|
giving a 32-bit gap after the fields above. We optimize the
|
|||
|
layout for that case and use the gap for extra code-specific
|
|||
|
information. */
|
|||
|
|
|||
|
/* The ORIGINAL_REGNO of a REG. */
|
|||
|
unsigned int original_regno;
|
|||
|
|
|||
|
/* The INSN_UID of an RTX_INSN-class code. */
|
|||
|
int insn_uid;
|
|||
|
|
|||
|
/* The SYMBOL_REF_FLAGS of a SYMBOL_REF. */
|
|||
|
unsigned int symbol_ref_flags;
|
|||
|
|
|||
|
/* The PAT_VAR_LOCATION_STATUS of a VAR_LOCATION. */
|
|||
|
enum var_init_status var_location_status;
|
|||
|
|
|||
|
/* In a CONST_WIDE_INT (aka hwivec_def), this is the number of
|
|||
|
HOST_WIDE_INTs in the hwivec_def. */
|
|||
|
unsigned int num_elem;
|
|||
|
|
|||
|
/* Information about a CONST_VECTOR. */
|
|||
|
struct
|
|||
|
{
|
|||
|
/* The value of CONST_VECTOR_NPATTERNS. */
|
|||
|
unsigned int npatterns : 16;
|
|||
|
|
|||
|
/* The value of CONST_VECTOR_NELTS_PER_PATTERN. */
|
|||
|
unsigned int nelts_per_pattern : 8;
|
|||
|
|
|||
|
/* For future expansion. */
|
|||
|
unsigned int unused : 8;
|
|||
|
} const_vector;
|
|||
|
} GTY ((skip)) u2;
|
|||
|
|
|||
|
/* The first element of the operands of this rtx.
|
|||
|
The number of operands and their types are controlled
|
|||
|
by the `code' field, according to rtl.def. */
|
|||
|
union u {
|
|||
|
rtunion fld[1];
|
|||
|
HOST_WIDE_INT hwint[1];
|
|||
|
struct reg_info reg;
|
|||
|
struct block_symbol block_sym;
|
|||
|
struct real_value rv;
|
|||
|
struct fixed_value fv;
|
|||
|
struct hwivec_def hwiv;
|
|||
|
struct const_poly_int_def cpi;
|
|||
|
} GTY ((special ("rtx_def"), desc ("GET_CODE (&%0)"))) u;
|
|||
|
};
|
|||
|
|
|||
|
/* A node for constructing singly-linked lists of rtx. */
|
|||
|
|
|||
|
struct GTY(()) rtx_expr_list : public rtx_def
|
|||
|
{
|
|||
|
private:
|
|||
|
/* No extra fields, but adds invariant: (GET_CODE (X) == EXPR_LIST). */
|
|||
|
|
|||
|
public:
|
|||
|
/* Get next in list. */
|
|||
|
rtx_expr_list *next () const;
|
|||
|
|
|||
|
/* Get at the underlying rtx. */
|
|||
|
rtx element () const;
|
|||
|
};
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <rtx_expr_list *>::test (rtx rt)
|
|||
|
{
|
|||
|
return rt->code == EXPR_LIST;
|
|||
|
}
|
|||
|
|
|||
|
struct GTY(()) rtx_insn_list : public rtx_def
|
|||
|
{
|
|||
|
private:
|
|||
|
/* No extra fields, but adds invariant: (GET_CODE (X) == INSN_LIST).
|
|||
|
|
|||
|
This is an instance of:
|
|||
|
|
|||
|
DEF_RTL_EXPR(INSN_LIST, "insn_list", "ue", RTX_EXTRA)
|
|||
|
|
|||
|
i.e. a node for constructing singly-linked lists of rtx_insn *, where
|
|||
|
the list is "external" to the insn (as opposed to the doubly-linked
|
|||
|
list embedded within rtx_insn itself). */
|
|||
|
|
|||
|
public:
|
|||
|
/* Get next in list. */
|
|||
|
rtx_insn_list *next () const;
|
|||
|
|
|||
|
/* Get at the underlying instruction. */
|
|||
|
rtx_insn *insn () const;
|
|||
|
|
|||
|
};
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <rtx_insn_list *>::test (rtx rt)
|
|||
|
{
|
|||
|
return rt->code == INSN_LIST;
|
|||
|
}
|
|||
|
|
|||
|
/* A node with invariant GET_CODE (X) == SEQUENCE i.e. a vector of rtx,
|
|||
|
typically (but not always) of rtx_insn *, used in the late passes. */
|
|||
|
|
|||
|
struct GTY(()) rtx_sequence : public rtx_def
|
|||
|
{
|
|||
|
private:
|
|||
|
/* No extra fields, but adds invariant: (GET_CODE (X) == SEQUENCE). */
|
|||
|
|
|||
|
public:
|
|||
|
/* Get number of elements in sequence. */
|
|||
|
int len () const;
|
|||
|
|
|||
|
/* Get i-th element of the sequence. */
|
|||
|
rtx element (int index) const;
|
|||
|
|
|||
|
/* Get i-th element of the sequence, with a checked cast to
|
|||
|
rtx_insn *. */
|
|||
|
rtx_insn *insn (int index) const;
|
|||
|
};
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <rtx_sequence *>::test (rtx rt)
|
|||
|
{
|
|||
|
return rt->code == SEQUENCE;
|
|||
|
}
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <const rtx_sequence *>::test (const_rtx rt)
|
|||
|
{
|
|||
|
return rt->code == SEQUENCE;
|
|||
|
}
|
|||
|
|
|||
|
struct GTY(()) rtx_insn : public rtx_def
|
|||
|
{
|
|||
|
public:
|
|||
|
/* No extra fields, but adds the invariant:
|
|||
|
|
|||
|
(INSN_P (X)
|
|||
|
|| NOTE_P (X)
|
|||
|
|| JUMP_TABLE_DATA_P (X)
|
|||
|
|| BARRIER_P (X)
|
|||
|
|| LABEL_P (X))
|
|||
|
|
|||
|
i.e. that we must be able to use the following:
|
|||
|
INSN_UID ()
|
|||
|
NEXT_INSN ()
|
|||
|
PREV_INSN ()
|
|||
|
i.e. we have an rtx that has an INSN_UID field and can be part of
|
|||
|
a linked list of insns.
|
|||
|
*/
|
|||
|
|
|||
|
/* Returns true if this insn has been deleted. */
|
|||
|
|
|||
|
bool deleted () const { return volatil; }
|
|||
|
|
|||
|
/* Mark this insn as deleted. */
|
|||
|
|
|||
|
void set_deleted () { volatil = true; }
|
|||
|
|
|||
|
/* Mark this insn as not deleted. */
|
|||
|
|
|||
|
void set_undeleted () { volatil = false; }
|
|||
|
};
|
|||
|
|
|||
|
/* Subclasses of rtx_insn. */
|
|||
|
|
|||
|
struct GTY(()) rtx_debug_insn : public rtx_insn
|
|||
|
{
|
|||
|
/* No extra fields, but adds the invariant:
|
|||
|
DEBUG_INSN_P (X) aka (GET_CODE (X) == DEBUG_INSN)
|
|||
|
i.e. an annotation for tracking variable assignments.
|
|||
|
|
|||
|
This is an instance of:
|
|||
|
DEF_RTL_EXPR(DEBUG_INSN, "debug_insn", "uuBeiie", RTX_INSN)
|
|||
|
from rtl.def. */
|
|||
|
};
|
|||
|
|
|||
|
struct GTY(()) rtx_nonjump_insn : public rtx_insn
|
|||
|
{
|
|||
|
/* No extra fields, but adds the invariant:
|
|||
|
NONJUMP_INSN_P (X) aka (GET_CODE (X) == INSN)
|
|||
|
i.e an instruction that cannot jump.
|
|||
|
|
|||
|
This is an instance of:
|
|||
|
DEF_RTL_EXPR(INSN, "insn", "uuBeiie", RTX_INSN)
|
|||
|
from rtl.def. */
|
|||
|
};
|
|||
|
|
|||
|
struct GTY(()) rtx_jump_insn : public rtx_insn
|
|||
|
{
|
|||
|
public:
|
|||
|
/* No extra fields, but adds the invariant:
|
|||
|
JUMP_P (X) aka (GET_CODE (X) == JUMP_INSN)
|
|||
|
i.e. an instruction that can possibly jump.
|
|||
|
|
|||
|
This is an instance of:
|
|||
|
DEF_RTL_EXPR(JUMP_INSN, "jump_insn", "uuBeiie0", RTX_INSN)
|
|||
|
from rtl.def. */
|
|||
|
|
|||
|
/* Returns jump target of this instruction. The returned value is not
|
|||
|
necessarily a code label: it may also be a RETURN or SIMPLE_RETURN
|
|||
|
expression. Also, when the code label is marked "deleted", it is
|
|||
|
replaced by a NOTE. In some cases the value is NULL_RTX. */
|
|||
|
|
|||
|
inline rtx jump_label () const;
|
|||
|
|
|||
|
/* Returns jump target cast to rtx_code_label *. */
|
|||
|
|
|||
|
inline rtx_code_label *jump_target () const;
|
|||
|
|
|||
|
/* Set jump target. */
|
|||
|
|
|||
|
inline void set_jump_target (rtx_code_label *);
|
|||
|
};
|
|||
|
|
|||
|
struct GTY(()) rtx_call_insn : public rtx_insn
|
|||
|
{
|
|||
|
/* No extra fields, but adds the invariant:
|
|||
|
CALL_P (X) aka (GET_CODE (X) == CALL_INSN)
|
|||
|
i.e. an instruction that can possibly call a subroutine
|
|||
|
but which will not change which instruction comes next
|
|||
|
in the current function.
|
|||
|
|
|||
|
This is an instance of:
|
|||
|
DEF_RTL_EXPR(CALL_INSN, "call_insn", "uuBeiiee", RTX_INSN)
|
|||
|
from rtl.def. */
|
|||
|
};
|
|||
|
|
|||
|
struct GTY(()) rtx_jump_table_data : public rtx_insn
|
|||
|
{
|
|||
|
/* No extra fields, but adds the invariant:
|
|||
|
JUMP_TABLE_DATA_P (X) aka (GET_CODE (INSN) == JUMP_TABLE_DATA)
|
|||
|
i.e. a data for a jump table, considered an instruction for
|
|||
|
historical reasons.
|
|||
|
|
|||
|
This is an instance of:
|
|||
|
DEF_RTL_EXPR(JUMP_TABLE_DATA, "jump_table_data", "uuBe0000", RTX_INSN)
|
|||
|
from rtl.def. */
|
|||
|
|
|||
|
/* This can be either:
|
|||
|
|
|||
|
(a) a table of absolute jumps, in which case PATTERN (this) is an
|
|||
|
ADDR_VEC with arg 0 a vector of labels, or
|
|||
|
|
|||
|
(b) a table of relative jumps (e.g. for -fPIC), in which case
|
|||
|
PATTERN (this) is an ADDR_DIFF_VEC, with arg 0 a LABEL_REF and
|
|||
|
arg 1 the vector of labels.
|
|||
|
|
|||
|
This method gets the underlying vec. */
|
|||
|
|
|||
|
inline rtvec get_labels () const;
|
|||
|
inline scalar_int_mode get_data_mode () const;
|
|||
|
};
|
|||
|
|
|||
|
struct GTY(()) rtx_barrier : public rtx_insn
|
|||
|
{
|
|||
|
/* No extra fields, but adds the invariant:
|
|||
|
BARRIER_P (X) aka (GET_CODE (X) == BARRIER)
|
|||
|
i.e. a marker that indicates that control will not flow through.
|
|||
|
|
|||
|
This is an instance of:
|
|||
|
DEF_RTL_EXPR(BARRIER, "barrier", "uu00000", RTX_EXTRA)
|
|||
|
from rtl.def. */
|
|||
|
};
|
|||
|
|
|||
|
struct GTY(()) rtx_code_label : public rtx_insn
|
|||
|
{
|
|||
|
/* No extra fields, but adds the invariant:
|
|||
|
LABEL_P (X) aka (GET_CODE (X) == CODE_LABEL)
|
|||
|
i.e. a label in the assembler.
|
|||
|
|
|||
|
This is an instance of:
|
|||
|
DEF_RTL_EXPR(CODE_LABEL, "code_label", "uuB00is", RTX_EXTRA)
|
|||
|
from rtl.def. */
|
|||
|
};
|
|||
|
|
|||
|
struct GTY(()) rtx_note : public rtx_insn
|
|||
|
{
|
|||
|
/* No extra fields, but adds the invariant:
|
|||
|
NOTE_P(X) aka (GET_CODE (X) == NOTE)
|
|||
|
i.e. a note about the corresponding source code.
|
|||
|
|
|||
|
This is an instance of:
|
|||
|
DEF_RTL_EXPR(NOTE, "note", "uuB0ni", RTX_EXTRA)
|
|||
|
from rtl.def. */
|
|||
|
};
|
|||
|
|
|||
|
/* The size in bytes of an rtx header (code, mode and flags). */
|
|||
|
#define RTX_HDR_SIZE offsetof (struct rtx_def, u)
|
|||
|
|
|||
|
/* The size in bytes of an rtx with code CODE. */
|
|||
|
#define RTX_CODE_SIZE(CODE) rtx_code_size[CODE]
|
|||
|
|
|||
|
#define NULL_RTX (rtx) 0
|
|||
|
|
|||
|
/* The "next" and "previous" RTX, relative to this one. */
|
|||
|
|
|||
|
#define RTX_NEXT(X) (rtx_next[GET_CODE (X)] == 0 ? NULL \
|
|||
|
: *(rtx *)(((char *)X) + rtx_next[GET_CODE (X)]))
|
|||
|
|
|||
|
/* FIXME: the "NEXT_INSN (PREV_INSN (X)) == X" condition shouldn't be needed.
|
|||
|
*/
|
|||
|
#define RTX_PREV(X) ((INSN_P (X) \
|
|||
|
|| NOTE_P (X) \
|
|||
|
|| JUMP_TABLE_DATA_P (X) \
|
|||
|
|| BARRIER_P (X) \
|
|||
|
|| LABEL_P (X)) \
|
|||
|
&& PREV_INSN (as_a <rtx_insn *> (X)) != NULL \
|
|||
|
&& NEXT_INSN (PREV_INSN (as_a <rtx_insn *> (X))) == X \
|
|||
|
? PREV_INSN (as_a <rtx_insn *> (X)) : NULL)
|
|||
|
|
|||
|
/* Define macros to access the `code' field of the rtx. */
|
|||
|
|
|||
|
#define GET_CODE(RTX) ((enum rtx_code) (RTX)->code)
|
|||
|
#define PUT_CODE(RTX, CODE) ((RTX)->code = (CODE))
|
|||
|
|
|||
|
#define GET_MODE(RTX) ((machine_mode) (RTX)->mode)
|
|||
|
#define PUT_MODE_RAW(RTX, MODE) ((RTX)->mode = (MODE))
|
|||
|
|
|||
|
/* RTL vector. These appear inside RTX's when there is a need
|
|||
|
for a variable number of things. The principle use is inside
|
|||
|
PARALLEL expressions. */
|
|||
|
|
|||
|
struct GTY(()) rtvec_def {
|
|||
|
int num_elem; /* number of elements */
|
|||
|
rtx GTY ((length ("%h.num_elem"))) elem[1];
|
|||
|
};
|
|||
|
|
|||
|
#define NULL_RTVEC (rtvec) 0
|
|||
|
|
|||
|
#define GET_NUM_ELEM(RTVEC) ((RTVEC)->num_elem)
|
|||
|
#define PUT_NUM_ELEM(RTVEC, NUM) ((RTVEC)->num_elem = (NUM))
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff X is an rtx for a register. */
|
|||
|
#define REG_P(X) (GET_CODE (X) == REG)
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff X is an rtx for a memory location. */
|
|||
|
#define MEM_P(X) (GET_CODE (X) == MEM)
|
|||
|
|
|||
|
#if TARGET_SUPPORTS_WIDE_INT
|
|||
|
|
|||
|
/* Match CONST_*s that can represent compile-time constant integers. */
|
|||
|
#define CASE_CONST_SCALAR_INT \
|
|||
|
case CONST_INT: \
|
|||
|
case CONST_WIDE_INT
|
|||
|
|
|||
|
/* Match CONST_*s for which pointer equality corresponds to value
|
|||
|
equality. */
|
|||
|
#define CASE_CONST_UNIQUE \
|
|||
|
case CONST_INT: \
|
|||
|
case CONST_WIDE_INT: \
|
|||
|
case CONST_POLY_INT: \
|
|||
|
case CONST_DOUBLE: \
|
|||
|
case CONST_FIXED
|
|||
|
|
|||
|
/* Match all CONST_* rtxes. */
|
|||
|
#define CASE_CONST_ANY \
|
|||
|
case CONST_INT: \
|
|||
|
case CONST_WIDE_INT: \
|
|||
|
case CONST_POLY_INT: \
|
|||
|
case CONST_DOUBLE: \
|
|||
|
case CONST_FIXED: \
|
|||
|
case CONST_VECTOR
|
|||
|
|
|||
|
#else
|
|||
|
|
|||
|
/* Match CONST_*s that can represent compile-time constant integers. */
|
|||
|
#define CASE_CONST_SCALAR_INT \
|
|||
|
case CONST_INT: \
|
|||
|
case CONST_DOUBLE
|
|||
|
|
|||
|
/* Match CONST_*s for which pointer equality corresponds to value
|
|||
|
equality. */
|
|||
|
#define CASE_CONST_UNIQUE \
|
|||
|
case CONST_INT: \
|
|||
|
case CONST_DOUBLE: \
|
|||
|
case CONST_FIXED
|
|||
|
|
|||
|
/* Match all CONST_* rtxes. */
|
|||
|
#define CASE_CONST_ANY \
|
|||
|
case CONST_INT: \
|
|||
|
case CONST_DOUBLE: \
|
|||
|
case CONST_FIXED: \
|
|||
|
case CONST_VECTOR
|
|||
|
#endif
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff X is an rtx for a constant integer. */
|
|||
|
#define CONST_INT_P(X) (GET_CODE (X) == CONST_INT)
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff X is an rtx for a constant integer. */
|
|||
|
#define CONST_WIDE_INT_P(X) (GET_CODE (X) == CONST_WIDE_INT)
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff X is an rtx for a polynomial constant
|
|||
|
integer. */
|
|||
|
#define CONST_POLY_INT_P(X) \
|
|||
|
(NUM_POLY_INT_COEFFS > 1 && GET_CODE (X) == CONST_POLY_INT)
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff X is an rtx for a constant fixed-point. */
|
|||
|
#define CONST_FIXED_P(X) (GET_CODE (X) == CONST_FIXED)
|
|||
|
|
|||
|
/* Predicate yielding true iff X is an rtx for a double-int
|
|||
|
or floating point constant. */
|
|||
|
#define CONST_DOUBLE_P(X) (GET_CODE (X) == CONST_DOUBLE)
|
|||
|
|
|||
|
/* Predicate yielding true iff X is an rtx for a double-int. */
|
|||
|
#define CONST_DOUBLE_AS_INT_P(X) \
|
|||
|
(GET_CODE (X) == CONST_DOUBLE && GET_MODE (X) == VOIDmode)
|
|||
|
|
|||
|
/* Predicate yielding true iff X is an rtx for a integer const. */
|
|||
|
#if TARGET_SUPPORTS_WIDE_INT
|
|||
|
#define CONST_SCALAR_INT_P(X) \
|
|||
|
(CONST_INT_P (X) || CONST_WIDE_INT_P (X))
|
|||
|
#else
|
|||
|
#define CONST_SCALAR_INT_P(X) \
|
|||
|
(CONST_INT_P (X) || CONST_DOUBLE_AS_INT_P (X))
|
|||
|
#endif
|
|||
|
|
|||
|
/* Predicate yielding true iff X is an rtx for a double-int. */
|
|||
|
#define CONST_DOUBLE_AS_FLOAT_P(X) \
|
|||
|
(GET_CODE (X) == CONST_DOUBLE && GET_MODE (X) != VOIDmode)
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff X is an rtx for a constant vector. */
|
|||
|
#define CONST_VECTOR_P(X) (GET_CODE (X) == CONST_VECTOR)
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff X is a label insn. */
|
|||
|
#define LABEL_P(X) (GET_CODE (X) == CODE_LABEL)
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff X is a jump insn. */
|
|||
|
#define JUMP_P(X) (GET_CODE (X) == JUMP_INSN)
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff X is a call insn. */
|
|||
|
#define CALL_P(X) (GET_CODE (X) == CALL_INSN)
|
|||
|
|
|||
|
/* 1 if RTX is a call_insn for a fake call.
|
|||
|
CALL_INSN use "used" flag to indicate it's a fake call. */
|
|||
|
#define FAKE_CALL_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("FAKE_CALL_P", (RTX), CALL_INSN)->used)
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff X is an insn that cannot jump. */
|
|||
|
#define NONJUMP_INSN_P(X) (GET_CODE (X) == INSN)
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff X is a debug note/insn. */
|
|||
|
#define DEBUG_INSN_P(X) (GET_CODE (X) == DEBUG_INSN)
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff X is an insn that is not a debug insn. */
|
|||
|
#define NONDEBUG_INSN_P(X) (NONJUMP_INSN_P (X) || JUMP_P (X) || CALL_P (X))
|
|||
|
|
|||
|
/* Nonzero if DEBUG_MARKER_INSN_P may possibly hold. */
|
|||
|
#define MAY_HAVE_DEBUG_MARKER_INSNS debug_nonbind_markers_p
|
|||
|
/* Nonzero if DEBUG_BIND_INSN_P may possibly hold. */
|
|||
|
#define MAY_HAVE_DEBUG_BIND_INSNS flag_var_tracking_assignments
|
|||
|
/* Nonzero if DEBUG_INSN_P may possibly hold. */
|
|||
|
#define MAY_HAVE_DEBUG_INSNS \
|
|||
|
(MAY_HAVE_DEBUG_MARKER_INSNS || MAY_HAVE_DEBUG_BIND_INSNS)
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff X is a real insn. */
|
|||
|
#define INSN_P(X) (NONDEBUG_INSN_P (X) || DEBUG_INSN_P (X))
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff X is a note insn. */
|
|||
|
#define NOTE_P(X) (GET_CODE (X) == NOTE)
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff X is a barrier insn. */
|
|||
|
#define BARRIER_P(X) (GET_CODE (X) == BARRIER)
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff X is a data for a jump table. */
|
|||
|
#define JUMP_TABLE_DATA_P(INSN) (GET_CODE (INSN) == JUMP_TABLE_DATA)
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff RTX is a subreg. */
|
|||
|
#define SUBREG_P(RTX) (GET_CODE (RTX) == SUBREG)
|
|||
|
|
|||
|
/* Predicate yielding true iff RTX is a symbol ref. */
|
|||
|
#define SYMBOL_REF_P(RTX) (GET_CODE (RTX) == SYMBOL_REF)
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <rtx_insn *>::test (rtx rt)
|
|||
|
{
|
|||
|
return (INSN_P (rt)
|
|||
|
|| NOTE_P (rt)
|
|||
|
|| JUMP_TABLE_DATA_P (rt)
|
|||
|
|| BARRIER_P (rt)
|
|||
|
|| LABEL_P (rt));
|
|||
|
}
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <const rtx_insn *>::test (const_rtx rt)
|
|||
|
{
|
|||
|
return (INSN_P (rt)
|
|||
|
|| NOTE_P (rt)
|
|||
|
|| JUMP_TABLE_DATA_P (rt)
|
|||
|
|| BARRIER_P (rt)
|
|||
|
|| LABEL_P (rt));
|
|||
|
}
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <rtx_debug_insn *>::test (rtx rt)
|
|||
|
{
|
|||
|
return DEBUG_INSN_P (rt);
|
|||
|
}
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <rtx_nonjump_insn *>::test (rtx rt)
|
|||
|
{
|
|||
|
return NONJUMP_INSN_P (rt);
|
|||
|
}
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <rtx_jump_insn *>::test (rtx rt)
|
|||
|
{
|
|||
|
return JUMP_P (rt);
|
|||
|
}
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <rtx_jump_insn *>::test (rtx_insn *insn)
|
|||
|
{
|
|||
|
return JUMP_P (insn);
|
|||
|
}
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <rtx_call_insn *>::test (rtx rt)
|
|||
|
{
|
|||
|
return CALL_P (rt);
|
|||
|
}
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <rtx_call_insn *>::test (rtx_insn *insn)
|
|||
|
{
|
|||
|
return CALL_P (insn);
|
|||
|
}
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <rtx_jump_table_data *>::test (rtx rt)
|
|||
|
{
|
|||
|
return JUMP_TABLE_DATA_P (rt);
|
|||
|
}
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <rtx_jump_table_data *>::test (rtx_insn *insn)
|
|||
|
{
|
|||
|
return JUMP_TABLE_DATA_P (insn);
|
|||
|
}
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <rtx_barrier *>::test (rtx rt)
|
|||
|
{
|
|||
|
return BARRIER_P (rt);
|
|||
|
}
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <rtx_code_label *>::test (rtx rt)
|
|||
|
{
|
|||
|
return LABEL_P (rt);
|
|||
|
}
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <rtx_code_label *>::test (rtx_insn *insn)
|
|||
|
{
|
|||
|
return LABEL_P (insn);
|
|||
|
}
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <rtx_note *>::test (rtx rt)
|
|||
|
{
|
|||
|
return NOTE_P (rt);
|
|||
|
}
|
|||
|
|
|||
|
template <>
|
|||
|
template <>
|
|||
|
inline bool
|
|||
|
is_a_helper <rtx_note *>::test (rtx_insn *insn)
|
|||
|
{
|
|||
|
return NOTE_P (insn);
|
|||
|
}
|
|||
|
|
|||
|
/* Predicate yielding nonzero iff X is a return or simple_return. */
|
|||
|
#define ANY_RETURN_P(X) \
|
|||
|
(GET_CODE (X) == RETURN || GET_CODE (X) == SIMPLE_RETURN)
|
|||
|
|
|||
|
/* 1 if X is a unary operator. */
|
|||
|
|
|||
|
#define UNARY_P(X) \
|
|||
|
(GET_RTX_CLASS (GET_CODE (X)) == RTX_UNARY)
|
|||
|
|
|||
|
/* 1 if X is a binary operator. */
|
|||
|
|
|||
|
#define BINARY_P(X) \
|
|||
|
((GET_RTX_CLASS (GET_CODE (X)) & RTX_BINARY_MASK) == RTX_BINARY_RESULT)
|
|||
|
|
|||
|
/* 1 if X is an arithmetic operator. */
|
|||
|
|
|||
|
#define ARITHMETIC_P(X) \
|
|||
|
((GET_RTX_CLASS (GET_CODE (X)) & RTX_ARITHMETIC_MASK) \
|
|||
|
== RTX_ARITHMETIC_RESULT)
|
|||
|
|
|||
|
/* 1 if X is an arithmetic operator. */
|
|||
|
|
|||
|
#define COMMUTATIVE_ARITH_P(X) \
|
|||
|
(GET_RTX_CLASS (GET_CODE (X)) == RTX_COMM_ARITH)
|
|||
|
|
|||
|
/* 1 if X is a commutative arithmetic operator or a comparison operator.
|
|||
|
These two are sometimes selected together because it is possible to
|
|||
|
swap the two operands. */
|
|||
|
|
|||
|
#define SWAPPABLE_OPERANDS_P(X) \
|
|||
|
((1 << GET_RTX_CLASS (GET_CODE (X))) \
|
|||
|
& ((1 << RTX_COMM_ARITH) | (1 << RTX_COMM_COMPARE) \
|
|||
|
| (1 << RTX_COMPARE)))
|
|||
|
|
|||
|
/* 1 if X is a non-commutative operator. */
|
|||
|
|
|||
|
#define NON_COMMUTATIVE_P(X) \
|
|||
|
((GET_RTX_CLASS (GET_CODE (X)) & RTX_COMMUTATIVE_MASK) \
|
|||
|
== RTX_NON_COMMUTATIVE_RESULT)
|
|||
|
|
|||
|
/* 1 if X is a commutative operator on integers. */
|
|||
|
|
|||
|
#define COMMUTATIVE_P(X) \
|
|||
|
((GET_RTX_CLASS (GET_CODE (X)) & RTX_COMMUTATIVE_MASK) \
|
|||
|
== RTX_COMMUTATIVE_RESULT)
|
|||
|
|
|||
|
/* 1 if X is a relational operator. */
|
|||
|
|
|||
|
#define COMPARISON_P(X) \
|
|||
|
((GET_RTX_CLASS (GET_CODE (X)) & RTX_COMPARE_MASK) == RTX_COMPARE_RESULT)
|
|||
|
|
|||
|
/* 1 if X is a constant value that is an integer. */
|
|||
|
|
|||
|
#define CONSTANT_P(X) \
|
|||
|
(GET_RTX_CLASS (GET_CODE (X)) == RTX_CONST_OBJ)
|
|||
|
|
|||
|
/* 1 if X is a LABEL_REF. */
|
|||
|
#define LABEL_REF_P(X) \
|
|||
|
(GET_CODE (X) == LABEL_REF)
|
|||
|
|
|||
|
/* 1 if X can be used to represent an object. */
|
|||
|
#define OBJECT_P(X) \
|
|||
|
((GET_RTX_CLASS (GET_CODE (X)) & RTX_OBJ_MASK) == RTX_OBJ_RESULT)
|
|||
|
|
|||
|
/* General accessor macros for accessing the fields of an rtx. */
|
|||
|
|
|||
|
#if defined ENABLE_RTL_CHECKING && (GCC_VERSION >= 2007)
|
|||
|
/* The bit with a star outside the statement expr and an & inside is
|
|||
|
so that N can be evaluated only once. */
|
|||
|
#define RTL_CHECK1(RTX, N, C1) __extension__ \
|
|||
|
(*({ __typeof (RTX) const _rtx = (RTX); const int _n = (N); \
|
|||
|
const enum rtx_code _code = GET_CODE (_rtx); \
|
|||
|
if (_n < 0 || _n >= GET_RTX_LENGTH (_code)) \
|
|||
|
rtl_check_failed_bounds (_rtx, _n, __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
if (GET_RTX_FORMAT (_code)[_n] != C1) \
|
|||
|
rtl_check_failed_type1 (_rtx, _n, C1, __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
&_rtx->u.fld[_n]; }))
|
|||
|
|
|||
|
#define RTL_CHECK2(RTX, N, C1, C2) __extension__ \
|
|||
|
(*({ __typeof (RTX) const _rtx = (RTX); const int _n = (N); \
|
|||
|
const enum rtx_code _code = GET_CODE (_rtx); \
|
|||
|
if (_n < 0 || _n >= GET_RTX_LENGTH (_code)) \
|
|||
|
rtl_check_failed_bounds (_rtx, _n, __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
if (GET_RTX_FORMAT (_code)[_n] != C1 \
|
|||
|
&& GET_RTX_FORMAT (_code)[_n] != C2) \
|
|||
|
rtl_check_failed_type2 (_rtx, _n, C1, C2, __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
&_rtx->u.fld[_n]; }))
|
|||
|
|
|||
|
#define RTL_CHECKC1(RTX, N, C) __extension__ \
|
|||
|
(*({ __typeof (RTX) const _rtx = (RTX); const int _n = (N); \
|
|||
|
if (GET_CODE (_rtx) != (C)) \
|
|||
|
rtl_check_failed_code1 (_rtx, (C), __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
&_rtx->u.fld[_n]; }))
|
|||
|
|
|||
|
#define RTL_CHECKC2(RTX, N, C1, C2) __extension__ \
|
|||
|
(*({ __typeof (RTX) const _rtx = (RTX); const int _n = (N); \
|
|||
|
const enum rtx_code _code = GET_CODE (_rtx); \
|
|||
|
if (_code != (C1) && _code != (C2)) \
|
|||
|
rtl_check_failed_code2 (_rtx, (C1), (C2), __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
&_rtx->u.fld[_n]; }))
|
|||
|
|
|||
|
#define RTL_CHECKC3(RTX, N, C1, C2, C3) __extension__ \
|
|||
|
(*({ __typeof (RTX) const _rtx = (RTX); const int _n = (N); \
|
|||
|
const enum rtx_code _code = GET_CODE (_rtx); \
|
|||
|
if (_code != (C1) && _code != (C2) && _code != (C3)) \
|
|||
|
rtl_check_failed_code3 (_rtx, (C1), (C2), (C3), __FILE__, \
|
|||
|
__LINE__, __FUNCTION__); \
|
|||
|
&_rtx->u.fld[_n]; }))
|
|||
|
|
|||
|
#define RTVEC_ELT(RTVEC, I) __extension__ \
|
|||
|
(*({ __typeof (RTVEC) const _rtvec = (RTVEC); const int _i = (I); \
|
|||
|
if (_i < 0 || _i >= GET_NUM_ELEM (_rtvec)) \
|
|||
|
rtvec_check_failed_bounds (_rtvec, _i, __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
&_rtvec->elem[_i]; }))
|
|||
|
|
|||
|
#define XWINT(RTX, N) __extension__ \
|
|||
|
(*({ __typeof (RTX) const _rtx = (RTX); const int _n = (N); \
|
|||
|
const enum rtx_code _code = GET_CODE (_rtx); \
|
|||
|
if (_n < 0 || _n >= GET_RTX_LENGTH (_code)) \
|
|||
|
rtl_check_failed_bounds (_rtx, _n, __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
if (GET_RTX_FORMAT (_code)[_n] != 'w') \
|
|||
|
rtl_check_failed_type1 (_rtx, _n, 'w', __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
&_rtx->u.hwint[_n]; }))
|
|||
|
|
|||
|
#define CWI_ELT(RTX, I) __extension__ \
|
|||
|
(*({ __typeof (RTX) const _cwi = (RTX); \
|
|||
|
int _max = CWI_GET_NUM_ELEM (_cwi); \
|
|||
|
const int _i = (I); \
|
|||
|
if (_i < 0 || _i >= _max) \
|
|||
|
cwi_check_failed_bounds (_cwi, _i, __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
&_cwi->u.hwiv.elem[_i]; }))
|
|||
|
|
|||
|
#define XCWINT(RTX, N, C) __extension__ \
|
|||
|
(*({ __typeof (RTX) const _rtx = (RTX); \
|
|||
|
if (GET_CODE (_rtx) != (C)) \
|
|||
|
rtl_check_failed_code1 (_rtx, (C), __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
&_rtx->u.hwint[N]; }))
|
|||
|
|
|||
|
#define XCMWINT(RTX, N, C, M) __extension__ \
|
|||
|
(*({ __typeof (RTX) const _rtx = (RTX); \
|
|||
|
if (GET_CODE (_rtx) != (C) || GET_MODE (_rtx) != (M)) \
|
|||
|
rtl_check_failed_code_mode (_rtx, (C), (M), false, __FILE__, \
|
|||
|
__LINE__, __FUNCTION__); \
|
|||
|
&_rtx->u.hwint[N]; }))
|
|||
|
|
|||
|
#define XCNMPRV(RTX, C, M) __extension__ \
|
|||
|
({ __typeof (RTX) const _rtx = (RTX); \
|
|||
|
if (GET_CODE (_rtx) != (C) || GET_MODE (_rtx) == (M)) \
|
|||
|
rtl_check_failed_code_mode (_rtx, (C), (M), true, __FILE__, \
|
|||
|
__LINE__, __FUNCTION__); \
|
|||
|
&_rtx->u.rv; })
|
|||
|
|
|||
|
#define XCNMPFV(RTX, C, M) __extension__ \
|
|||
|
({ __typeof (RTX) const _rtx = (RTX); \
|
|||
|
if (GET_CODE (_rtx) != (C) || GET_MODE (_rtx) == (M)) \
|
|||
|
rtl_check_failed_code_mode (_rtx, (C), (M), true, __FILE__, \
|
|||
|
__LINE__, __FUNCTION__); \
|
|||
|
&_rtx->u.fv; })
|
|||
|
|
|||
|
#define REG_CHECK(RTX) __extension__ \
|
|||
|
({ __typeof (RTX) const _rtx = (RTX); \
|
|||
|
if (GET_CODE (_rtx) != REG) \
|
|||
|
rtl_check_failed_code1 (_rtx, REG, __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
&_rtx->u.reg; })
|
|||
|
|
|||
|
#define BLOCK_SYMBOL_CHECK(RTX) __extension__ \
|
|||
|
({ __typeof (RTX) const _symbol = (RTX); \
|
|||
|
const unsigned int flags = SYMBOL_REF_FLAGS (_symbol); \
|
|||
|
if ((flags & SYMBOL_FLAG_HAS_BLOCK_INFO) == 0) \
|
|||
|
rtl_check_failed_block_symbol (__FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
&_symbol->u.block_sym; })
|
|||
|
|
|||
|
#define HWIVEC_CHECK(RTX,C) __extension__ \
|
|||
|
({ __typeof (RTX) const _symbol = (RTX); \
|
|||
|
RTL_CHECKC1 (_symbol, 0, C); \
|
|||
|
&_symbol->u.hwiv; })
|
|||
|
|
|||
|
extern void rtl_check_failed_bounds (const_rtx, int, const char *, int,
|
|||
|
const char *)
|
|||
|
ATTRIBUTE_NORETURN ATTRIBUTE_COLD;
|
|||
|
extern void rtl_check_failed_type1 (const_rtx, int, int, const char *, int,
|
|||
|
const char *)
|
|||
|
ATTRIBUTE_NORETURN ATTRIBUTE_COLD;
|
|||
|
extern void rtl_check_failed_type2 (const_rtx, int, int, int, const char *,
|
|||
|
int, const char *)
|
|||
|
ATTRIBUTE_NORETURN ATTRIBUTE_COLD;
|
|||
|
extern void rtl_check_failed_code1 (const_rtx, enum rtx_code, const char *,
|
|||
|
int, const char *)
|
|||
|
ATTRIBUTE_NORETURN ATTRIBUTE_COLD;
|
|||
|
extern void rtl_check_failed_code2 (const_rtx, enum rtx_code, enum rtx_code,
|
|||
|
const char *, int, const char *)
|
|||
|
ATTRIBUTE_NORETURN ATTRIBUTE_COLD;
|
|||
|
extern void rtl_check_failed_code3 (const_rtx, enum rtx_code, enum rtx_code,
|
|||
|
enum rtx_code, const char *, int,
|
|||
|
const char *)
|
|||
|
ATTRIBUTE_NORETURN ATTRIBUTE_COLD;
|
|||
|
extern void rtl_check_failed_code_mode (const_rtx, enum rtx_code, machine_mode,
|
|||
|
bool, const char *, int, const char *)
|
|||
|
ATTRIBUTE_NORETURN ATTRIBUTE_COLD;
|
|||
|
extern void rtl_check_failed_block_symbol (const char *, int, const char *)
|
|||
|
ATTRIBUTE_NORETURN ATTRIBUTE_COLD;
|
|||
|
extern void cwi_check_failed_bounds (const_rtx, int, const char *, int,
|
|||
|
const char *)
|
|||
|
ATTRIBUTE_NORETURN ATTRIBUTE_COLD;
|
|||
|
extern void rtvec_check_failed_bounds (const_rtvec, int, const char *, int,
|
|||
|
const char *)
|
|||
|
ATTRIBUTE_NORETURN ATTRIBUTE_COLD;
|
|||
|
|
|||
|
#else /* not ENABLE_RTL_CHECKING */
|
|||
|
|
|||
|
#define RTL_CHECK1(RTX, N, C1) ((RTX)->u.fld[N])
|
|||
|
#define RTL_CHECK2(RTX, N, C1, C2) ((RTX)->u.fld[N])
|
|||
|
#define RTL_CHECKC1(RTX, N, C) ((RTX)->u.fld[N])
|
|||
|
#define RTL_CHECKC2(RTX, N, C1, C2) ((RTX)->u.fld[N])
|
|||
|
#define RTL_CHECKC3(RTX, N, C1, C2, C3) ((RTX)->u.fld[N])
|
|||
|
#define RTVEC_ELT(RTVEC, I) ((RTVEC)->elem[I])
|
|||
|
#define XWINT(RTX, N) ((RTX)->u.hwint[N])
|
|||
|
#define CWI_ELT(RTX, I) ((RTX)->u.hwiv.elem[I])
|
|||
|
#define XCWINT(RTX, N, C) ((RTX)->u.hwint[N])
|
|||
|
#define XCMWINT(RTX, N, C, M) ((RTX)->u.hwint[N])
|
|||
|
#define XCNMWINT(RTX, N, C, M) ((RTX)->u.hwint[N])
|
|||
|
#define XCNMPRV(RTX, C, M) (&(RTX)->u.rv)
|
|||
|
#define XCNMPFV(RTX, C, M) (&(RTX)->u.fv)
|
|||
|
#define REG_CHECK(RTX) (&(RTX)->u.reg)
|
|||
|
#define BLOCK_SYMBOL_CHECK(RTX) (&(RTX)->u.block_sym)
|
|||
|
#define HWIVEC_CHECK(RTX,C) (&(RTX)->u.hwiv)
|
|||
|
|
|||
|
#endif
|
|||
|
|
|||
|
/* General accessor macros for accessing the flags of an rtx. */
|
|||
|
|
|||
|
/* Access an individual rtx flag, with no checking of any kind. */
|
|||
|
#define RTX_FLAG(RTX, FLAG) ((RTX)->FLAG)
|
|||
|
|
|||
|
#if defined ENABLE_RTL_FLAG_CHECKING && (GCC_VERSION >= 2007)
|
|||
|
#define RTL_FLAG_CHECK1(NAME, RTX, C1) __extension__ \
|
|||
|
({ __typeof (RTX) const _rtx = (RTX); \
|
|||
|
if (GET_CODE (_rtx) != C1) \
|
|||
|
rtl_check_failed_flag (NAME, _rtx, __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
_rtx; })
|
|||
|
|
|||
|
#define RTL_FLAG_CHECK2(NAME, RTX, C1, C2) __extension__ \
|
|||
|
({ __typeof (RTX) const _rtx = (RTX); \
|
|||
|
if (GET_CODE (_rtx) != C1 && GET_CODE(_rtx) != C2) \
|
|||
|
rtl_check_failed_flag (NAME,_rtx, __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
_rtx; })
|
|||
|
|
|||
|
#define RTL_FLAG_CHECK3(NAME, RTX, C1, C2, C3) __extension__ \
|
|||
|
({ __typeof (RTX) const _rtx = (RTX); \
|
|||
|
if (GET_CODE (_rtx) != C1 && GET_CODE(_rtx) != C2 \
|
|||
|
&& GET_CODE (_rtx) != C3) \
|
|||
|
rtl_check_failed_flag (NAME, _rtx, __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
_rtx; })
|
|||
|
|
|||
|
#define RTL_FLAG_CHECK4(NAME, RTX, C1, C2, C3, C4) __extension__ \
|
|||
|
({ __typeof (RTX) const _rtx = (RTX); \
|
|||
|
if (GET_CODE (_rtx) != C1 && GET_CODE(_rtx) != C2 \
|
|||
|
&& GET_CODE (_rtx) != C3 && GET_CODE(_rtx) != C4) \
|
|||
|
rtl_check_failed_flag (NAME, _rtx, __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
_rtx; })
|
|||
|
|
|||
|
#define RTL_FLAG_CHECK5(NAME, RTX, C1, C2, C3, C4, C5) __extension__ \
|
|||
|
({ __typeof (RTX) const _rtx = (RTX); \
|
|||
|
if (GET_CODE (_rtx) != C1 && GET_CODE (_rtx) != C2 \
|
|||
|
&& GET_CODE (_rtx) != C3 && GET_CODE (_rtx) != C4 \
|
|||
|
&& GET_CODE (_rtx) != C5) \
|
|||
|
rtl_check_failed_flag (NAME, _rtx, __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
_rtx; })
|
|||
|
|
|||
|
#define RTL_FLAG_CHECK6(NAME, RTX, C1, C2, C3, C4, C5, C6) \
|
|||
|
__extension__ \
|
|||
|
({ __typeof (RTX) const _rtx = (RTX); \
|
|||
|
if (GET_CODE (_rtx) != C1 && GET_CODE (_rtx) != C2 \
|
|||
|
&& GET_CODE (_rtx) != C3 && GET_CODE (_rtx) != C4 \
|
|||
|
&& GET_CODE (_rtx) != C5 && GET_CODE (_rtx) != C6) \
|
|||
|
rtl_check_failed_flag (NAME,_rtx, __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
_rtx; })
|
|||
|
|
|||
|
#define RTL_FLAG_CHECK7(NAME, RTX, C1, C2, C3, C4, C5, C6, C7) \
|
|||
|
__extension__ \
|
|||
|
({ __typeof (RTX) const _rtx = (RTX); \
|
|||
|
if (GET_CODE (_rtx) != C1 && GET_CODE (_rtx) != C2 \
|
|||
|
&& GET_CODE (_rtx) != C3 && GET_CODE (_rtx) != C4 \
|
|||
|
&& GET_CODE (_rtx) != C5 && GET_CODE (_rtx) != C6 \
|
|||
|
&& GET_CODE (_rtx) != C7) \
|
|||
|
rtl_check_failed_flag (NAME, _rtx, __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
_rtx; })
|
|||
|
|
|||
|
#define RTL_INSN_CHAIN_FLAG_CHECK(NAME, RTX) \
|
|||
|
__extension__ \
|
|||
|
({ __typeof (RTX) const _rtx = (RTX); \
|
|||
|
if (!INSN_CHAIN_CODE_P (GET_CODE (_rtx))) \
|
|||
|
rtl_check_failed_flag (NAME, _rtx, __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
_rtx; })
|
|||
|
|
|||
|
extern void rtl_check_failed_flag (const char *, const_rtx, const char *,
|
|||
|
int, const char *)
|
|||
|
ATTRIBUTE_NORETURN ATTRIBUTE_COLD
|
|||
|
;
|
|||
|
|
|||
|
#else /* not ENABLE_RTL_FLAG_CHECKING */
|
|||
|
|
|||
|
#define RTL_FLAG_CHECK1(NAME, RTX, C1) (RTX)
|
|||
|
#define RTL_FLAG_CHECK2(NAME, RTX, C1, C2) (RTX)
|
|||
|
#define RTL_FLAG_CHECK3(NAME, RTX, C1, C2, C3) (RTX)
|
|||
|
#define RTL_FLAG_CHECK4(NAME, RTX, C1, C2, C3, C4) (RTX)
|
|||
|
#define RTL_FLAG_CHECK5(NAME, RTX, C1, C2, C3, C4, C5) (RTX)
|
|||
|
#define RTL_FLAG_CHECK6(NAME, RTX, C1, C2, C3, C4, C5, C6) (RTX)
|
|||
|
#define RTL_FLAG_CHECK7(NAME, RTX, C1, C2, C3, C4, C5, C6, C7) (RTX)
|
|||
|
#define RTL_INSN_CHAIN_FLAG_CHECK(NAME, RTX) (RTX)
|
|||
|
#endif
|
|||
|
|
|||
|
#define XINT(RTX, N) (RTL_CHECK2 (RTX, N, 'i', 'n').rt_int)
|
|||
|
#define XUINT(RTX, N) (RTL_CHECK2 (RTX, N, 'i', 'n').rt_uint)
|
|||
|
#define XSTR(RTX, N) (RTL_CHECK2 (RTX, N, 's', 'S').rt_str)
|
|||
|
#define XEXP(RTX, N) (RTL_CHECK2 (RTX, N, 'e', 'u').rt_rtx)
|
|||
|
#define XVEC(RTX, N) (RTL_CHECK2 (RTX, N, 'E', 'V').rt_rtvec)
|
|||
|
#define XMODE(RTX, N) (RTL_CHECK1 (RTX, N, 'M').rt_type)
|
|||
|
#define XTREE(RTX, N) (RTL_CHECK1 (RTX, N, 't').rt_tree)
|
|||
|
#define XBBDEF(RTX, N) (RTL_CHECK1 (RTX, N, 'B').rt_bb)
|
|||
|
#define XTMPL(RTX, N) (RTL_CHECK1 (RTX, N, 'T').rt_str)
|
|||
|
#define XCFI(RTX, N) (RTL_CHECK1 (RTX, N, 'C').rt_cfi)
|
|||
|
|
|||
|
#define XVECEXP(RTX, N, M) RTVEC_ELT (XVEC (RTX, N), M)
|
|||
|
#define XVECLEN(RTX, N) GET_NUM_ELEM (XVEC (RTX, N))
|
|||
|
|
|||
|
/* These are like XINT, etc. except that they expect a '0' field instead
|
|||
|
of the normal type code. */
|
|||
|
|
|||
|
#define X0INT(RTX, N) (RTL_CHECK1 (RTX, N, '0').rt_int)
|
|||
|
#define X0UINT(RTX, N) (RTL_CHECK1 (RTX, N, '0').rt_uint)
|
|||
|
#define X0STR(RTX, N) (RTL_CHECK1 (RTX, N, '0').rt_str)
|
|||
|
#define X0EXP(RTX, N) (RTL_CHECK1 (RTX, N, '0').rt_rtx)
|
|||
|
#define X0VEC(RTX, N) (RTL_CHECK1 (RTX, N, '0').rt_rtvec)
|
|||
|
#define X0MODE(RTX, N) (RTL_CHECK1 (RTX, N, '0').rt_type)
|
|||
|
#define X0TREE(RTX, N) (RTL_CHECK1 (RTX, N, '0').rt_tree)
|
|||
|
#define X0BBDEF(RTX, N) (RTL_CHECK1 (RTX, N, '0').rt_bb)
|
|||
|
#define X0ADVFLAGS(RTX, N) (RTL_CHECK1 (RTX, N, '0').rt_addr_diff_vec_flags)
|
|||
|
#define X0CSELIB(RTX, N) (RTL_CHECK1 (RTX, N, '0').rt_cselib)
|
|||
|
#define X0MEMATTR(RTX, N) (RTL_CHECKC1 (RTX, N, MEM).rt_mem)
|
|||
|
#define X0CONSTANT(RTX, N) (RTL_CHECK1 (RTX, N, '0').rt_constant)
|
|||
|
|
|||
|
/* Access a '0' field with any type. */
|
|||
|
#define X0ANY(RTX, N) RTL_CHECK1 (RTX, N, '0')
|
|||
|
|
|||
|
#define XCINT(RTX, N, C) (RTL_CHECKC1 (RTX, N, C).rt_int)
|
|||
|
#define XCUINT(RTX, N, C) (RTL_CHECKC1 (RTX, N, C).rt_uint)
|
|||
|
#define XCSUBREG(RTX, N, C) (RTL_CHECKC1 (RTX, N, C).rt_subreg)
|
|||
|
#define XCSTR(RTX, N, C) (RTL_CHECKC1 (RTX, N, C).rt_str)
|
|||
|
#define XCEXP(RTX, N, C) (RTL_CHECKC1 (RTX, N, C).rt_rtx)
|
|||
|
#define XCVEC(RTX, N, C) (RTL_CHECKC1 (RTX, N, C).rt_rtvec)
|
|||
|
#define XCMODE(RTX, N, C) (RTL_CHECKC1 (RTX, N, C).rt_type)
|
|||
|
#define XCTREE(RTX, N, C) (RTL_CHECKC1 (RTX, N, C).rt_tree)
|
|||
|
#define XCBBDEF(RTX, N, C) (RTL_CHECKC1 (RTX, N, C).rt_bb)
|
|||
|
#define XCCFI(RTX, N, C) (RTL_CHECKC1 (RTX, N, C).rt_cfi)
|
|||
|
#define XCCSELIB(RTX, N, C) (RTL_CHECKC1 (RTX, N, C).rt_cselib)
|
|||
|
|
|||
|
#define XCVECEXP(RTX, N, M, C) RTVEC_ELT (XCVEC (RTX, N, C), M)
|
|||
|
#define XCVECLEN(RTX, N, C) GET_NUM_ELEM (XCVEC (RTX, N, C))
|
|||
|
|
|||
|
#define XC2EXP(RTX, N, C1, C2) (RTL_CHECKC2 (RTX, N, C1, C2).rt_rtx)
|
|||
|
#define XC3EXP(RTX, N, C1, C2, C3) (RTL_CHECKC3 (RTX, N, C1, C2, C3).rt_rtx)
|
|||
|
|
|||
|
|
|||
|
/* Methods of rtx_expr_list. */
|
|||
|
|
|||
|
inline rtx_expr_list *rtx_expr_list::next () const
|
|||
|
{
|
|||
|
rtx tmp = XEXP (this, 1);
|
|||
|
return safe_as_a <rtx_expr_list *> (tmp);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx rtx_expr_list::element () const
|
|||
|
{
|
|||
|
return XEXP (this, 0);
|
|||
|
}
|
|||
|
|
|||
|
/* Methods of rtx_insn_list. */
|
|||
|
|
|||
|
inline rtx_insn_list *rtx_insn_list::next () const
|
|||
|
{
|
|||
|
rtx tmp = XEXP (this, 1);
|
|||
|
return safe_as_a <rtx_insn_list *> (tmp);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx_insn *rtx_insn_list::insn () const
|
|||
|
{
|
|||
|
rtx tmp = XEXP (this, 0);
|
|||
|
return safe_as_a <rtx_insn *> (tmp);
|
|||
|
}
|
|||
|
|
|||
|
/* Methods of rtx_sequence. */
|
|||
|
|
|||
|
inline int rtx_sequence::len () const
|
|||
|
{
|
|||
|
return XVECLEN (this, 0);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx rtx_sequence::element (int index) const
|
|||
|
{
|
|||
|
return XVECEXP (this, 0, index);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx_insn *rtx_sequence::insn (int index) const
|
|||
|
{
|
|||
|
return as_a <rtx_insn *> (XVECEXP (this, 0, index));
|
|||
|
}
|
|||
|
|
|||
|
/* ACCESS MACROS for particular fields of insns. */
|
|||
|
|
|||
|
/* Holds a unique number for each insn.
|
|||
|
These are not necessarily sequentially increasing. */
|
|||
|
inline int INSN_UID (const_rtx insn)
|
|||
|
{
|
|||
|
return RTL_INSN_CHAIN_FLAG_CHECK ("INSN_UID",
|
|||
|
(insn))->u2.insn_uid;
|
|||
|
}
|
|||
|
inline int& INSN_UID (rtx insn)
|
|||
|
{
|
|||
|
return RTL_INSN_CHAIN_FLAG_CHECK ("INSN_UID",
|
|||
|
(insn))->u2.insn_uid;
|
|||
|
}
|
|||
|
|
|||
|
/* Chain insns together in sequence. */
|
|||
|
|
|||
|
/* For now these are split in two: an rvalue form:
|
|||
|
PREV_INSN/NEXT_INSN
|
|||
|
and an lvalue form:
|
|||
|
SET_NEXT_INSN/SET_PREV_INSN. */
|
|||
|
|
|||
|
inline rtx_insn *PREV_INSN (const rtx_insn *insn)
|
|||
|
{
|
|||
|
rtx prev = XEXP (insn, 0);
|
|||
|
return safe_as_a <rtx_insn *> (prev);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx& SET_PREV_INSN (rtx_insn *insn)
|
|||
|
{
|
|||
|
return XEXP (insn, 0);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx_insn *NEXT_INSN (const rtx_insn *insn)
|
|||
|
{
|
|||
|
rtx next = XEXP (insn, 1);
|
|||
|
return safe_as_a <rtx_insn *> (next);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx& SET_NEXT_INSN (rtx_insn *insn)
|
|||
|
{
|
|||
|
return XEXP (insn, 1);
|
|||
|
}
|
|||
|
|
|||
|
inline basic_block BLOCK_FOR_INSN (const_rtx insn)
|
|||
|
{
|
|||
|
return XBBDEF (insn, 2);
|
|||
|
}
|
|||
|
|
|||
|
inline basic_block& BLOCK_FOR_INSN (rtx insn)
|
|||
|
{
|
|||
|
return XBBDEF (insn, 2);
|
|||
|
}
|
|||
|
|
|||
|
inline void set_block_for_insn (rtx_insn *insn, basic_block bb)
|
|||
|
{
|
|||
|
BLOCK_FOR_INSN (insn) = bb;
|
|||
|
}
|
|||
|
|
|||
|
/* The body of an insn. */
|
|||
|
inline rtx PATTERN (const_rtx insn)
|
|||
|
{
|
|||
|
return XEXP (insn, 3);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx& PATTERN (rtx insn)
|
|||
|
{
|
|||
|
return XEXP (insn, 3);
|
|||
|
}
|
|||
|
|
|||
|
inline unsigned int INSN_LOCATION (const rtx_insn *insn)
|
|||
|
{
|
|||
|
return XUINT (insn, 4);
|
|||
|
}
|
|||
|
|
|||
|
inline unsigned int& INSN_LOCATION (rtx_insn *insn)
|
|||
|
{
|
|||
|
return XUINT (insn, 4);
|
|||
|
}
|
|||
|
|
|||
|
inline bool INSN_HAS_LOCATION (const rtx_insn *insn)
|
|||
|
{
|
|||
|
return LOCATION_LOCUS (INSN_LOCATION (insn)) != UNKNOWN_LOCATION;
|
|||
|
}
|
|||
|
|
|||
|
/* LOCATION of an RTX if relevant. */
|
|||
|
#define RTL_LOCATION(X) (INSN_P (X) ? \
|
|||
|
INSN_LOCATION (as_a <rtx_insn *> (X)) \
|
|||
|
: UNKNOWN_LOCATION)
|
|||
|
|
|||
|
/* Code number of instruction, from when it was recognized.
|
|||
|
-1 means this instruction has not been recognized yet. */
|
|||
|
#define INSN_CODE(INSN) XINT (INSN, 5)
|
|||
|
|
|||
|
inline rtvec rtx_jump_table_data::get_labels () const
|
|||
|
{
|
|||
|
rtx pat = PATTERN (this);
|
|||
|
if (GET_CODE (pat) == ADDR_VEC)
|
|||
|
return XVEC (pat, 0);
|
|||
|
else
|
|||
|
return XVEC (pat, 1); /* presumably an ADDR_DIFF_VEC */
|
|||
|
}
|
|||
|
|
|||
|
/* Return the mode of the data in the table, which is always a scalar
|
|||
|
integer. */
|
|||
|
|
|||
|
inline scalar_int_mode
|
|||
|
rtx_jump_table_data::get_data_mode () const
|
|||
|
{
|
|||
|
return as_a <scalar_int_mode> (GET_MODE (PATTERN (this)));
|
|||
|
}
|
|||
|
|
|||
|
/* If LABEL is followed by a jump table, return the table, otherwise
|
|||
|
return null. */
|
|||
|
|
|||
|
inline rtx_jump_table_data *
|
|||
|
jump_table_for_label (const rtx_code_label *label)
|
|||
|
{
|
|||
|
return safe_dyn_cast <rtx_jump_table_data *> (NEXT_INSN (label));
|
|||
|
}
|
|||
|
|
|||
|
#define RTX_FRAME_RELATED_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK6 ("RTX_FRAME_RELATED_P", (RTX), DEBUG_INSN, INSN, \
|
|||
|
CALL_INSN, JUMP_INSN, BARRIER, SET)->frame_related)
|
|||
|
|
|||
|
/* 1 if JUMP RTX is a crossing jump. */
|
|||
|
#define CROSSING_JUMP_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("CROSSING_JUMP_P", (RTX), JUMP_INSN)->jump)
|
|||
|
|
|||
|
/* 1 if RTX is a call to a const function. Built from ECF_CONST and
|
|||
|
TREE_READONLY. */
|
|||
|
#define RTL_CONST_CALL_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("RTL_CONST_CALL_P", (RTX), CALL_INSN)->unchanging)
|
|||
|
|
|||
|
/* 1 if RTX is a call to a pure function. Built from ECF_PURE and
|
|||
|
DECL_PURE_P. */
|
|||
|
#define RTL_PURE_CALL_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("RTL_PURE_CALL_P", (RTX), CALL_INSN)->return_val)
|
|||
|
|
|||
|
/* 1 if RTX is a call to a const or pure function. */
|
|||
|
#define RTL_CONST_OR_PURE_CALL_P(RTX) \
|
|||
|
(RTL_CONST_CALL_P (RTX) || RTL_PURE_CALL_P (RTX))
|
|||
|
|
|||
|
/* 1 if RTX is a call to a looping const or pure function. Built from
|
|||
|
ECF_LOOPING_CONST_OR_PURE and DECL_LOOPING_CONST_OR_PURE_P. */
|
|||
|
#define RTL_LOOPING_CONST_OR_PURE_CALL_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("CONST_OR_PURE_CALL_P", (RTX), CALL_INSN)->call)
|
|||
|
|
|||
|
/* 1 if RTX is a call_insn for a sibling call. */
|
|||
|
#define SIBLING_CALL_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("SIBLING_CALL_P", (RTX), CALL_INSN)->jump)
|
|||
|
|
|||
|
/* 1 if RTX is a jump_insn, call_insn, or insn that is an annulling branch. */
|
|||
|
#define INSN_ANNULLED_BRANCH_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("INSN_ANNULLED_BRANCH_P", (RTX), JUMP_INSN)->unchanging)
|
|||
|
|
|||
|
/* 1 if RTX is an insn in a delay slot and is from the target of the branch.
|
|||
|
If the branch insn has INSN_ANNULLED_BRANCH_P set, this insn should only be
|
|||
|
executed if the branch is taken. For annulled branches with this bit
|
|||
|
clear, the insn should be executed only if the branch is not taken. */
|
|||
|
#define INSN_FROM_TARGET_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK3 ("INSN_FROM_TARGET_P", (RTX), INSN, JUMP_INSN, \
|
|||
|
CALL_INSN)->in_struct)
|
|||
|
|
|||
|
/* In an ADDR_DIFF_VEC, the flags for RTX for use by branch shortening.
|
|||
|
See the comments for ADDR_DIFF_VEC in rtl.def. */
|
|||
|
#define ADDR_DIFF_VEC_FLAGS(RTX) X0ADVFLAGS (RTX, 4)
|
|||
|
|
|||
|
/* In a VALUE, the value cselib has assigned to RTX.
|
|||
|
This is a "struct cselib_val", see cselib.h. */
|
|||
|
#define CSELIB_VAL_PTR(RTX) X0CSELIB (RTX, 0)
|
|||
|
|
|||
|
/* Holds a list of notes on what this insn does to various REGs.
|
|||
|
It is a chain of EXPR_LIST rtx's, where the second operand is the
|
|||
|
chain pointer and the first operand is the REG being described.
|
|||
|
The mode field of the EXPR_LIST contains not a real machine mode
|
|||
|
but a value from enum reg_note. */
|
|||
|
#define REG_NOTES(INSN) XEXP(INSN, 6)
|
|||
|
|
|||
|
/* In an ENTRY_VALUE this is the DECL_INCOMING_RTL of the argument in
|
|||
|
question. */
|
|||
|
#define ENTRY_VALUE_EXP(RTX) (RTL_CHECKC1 (RTX, 0, ENTRY_VALUE).rt_rtx)
|
|||
|
|
|||
|
enum reg_note
|
|||
|
{
|
|||
|
#define DEF_REG_NOTE(NAME) NAME,
|
|||
|
#include "reg-notes.def"
|
|||
|
#undef DEF_REG_NOTE
|
|||
|
REG_NOTE_MAX
|
|||
|
};
|
|||
|
|
|||
|
/* Define macros to extract and insert the reg-note kind in an EXPR_LIST. */
|
|||
|
#define REG_NOTE_KIND(LINK) ((enum reg_note) GET_MODE (LINK))
|
|||
|
#define PUT_REG_NOTE_KIND(LINK, KIND) \
|
|||
|
PUT_MODE_RAW (LINK, (machine_mode) (KIND))
|
|||
|
|
|||
|
/* Names for REG_NOTE's in EXPR_LIST insn's. */
|
|||
|
|
|||
|
extern const char * const reg_note_name[];
|
|||
|
#define GET_REG_NOTE_NAME(MODE) (reg_note_name[(int) (MODE)])
|
|||
|
|
|||
|
/* This field is only present on CALL_INSNs. It holds a chain of EXPR_LIST of
|
|||
|
USE, CLOBBER and SET expressions.
|
|||
|
USE expressions list the registers filled with arguments that
|
|||
|
are passed to the function.
|
|||
|
CLOBBER expressions document the registers explicitly clobbered
|
|||
|
by this CALL_INSN.
|
|||
|
SET expressions say that the return value of the call (the SET_DEST)
|
|||
|
is equivalent to a value available before the call (the SET_SRC).
|
|||
|
This kind of SET is used when the return value is predictable in
|
|||
|
advance. It is purely an optimisation hint; unlike USEs and CLOBBERs,
|
|||
|
it does not affect register liveness.
|
|||
|
|
|||
|
Pseudo registers cannot be mentioned in this list. */
|
|||
|
#define CALL_INSN_FUNCTION_USAGE(INSN) XEXP(INSN, 7)
|
|||
|
|
|||
|
/* The label-number of a code-label. The assembler label
|
|||
|
is made from `L' and the label-number printed in decimal.
|
|||
|
Label numbers are unique in a compilation. */
|
|||
|
#define CODE_LABEL_NUMBER(INSN) XINT (INSN, 5)
|
|||
|
|
|||
|
/* In a NOTE that is a line number, this is a string for the file name that the
|
|||
|
line is in. We use the same field to record block numbers temporarily in
|
|||
|
NOTE_INSN_BLOCK_BEG and NOTE_INSN_BLOCK_END notes. (We avoid lots of casts
|
|||
|
between ints and pointers if we use a different macro for the block number.)
|
|||
|
*/
|
|||
|
|
|||
|
/* Opaque data. */
|
|||
|
#define NOTE_DATA(INSN) RTL_CHECKC1 (INSN, 3, NOTE)
|
|||
|
#define NOTE_DELETED_LABEL_NAME(INSN) XCSTR (INSN, 3, NOTE)
|
|||
|
#define SET_INSN_DELETED(INSN) set_insn_deleted (INSN);
|
|||
|
#define NOTE_BLOCK(INSN) XCTREE (INSN, 3, NOTE)
|
|||
|
#define NOTE_EH_HANDLER(INSN) XCINT (INSN, 3, NOTE)
|
|||
|
#define NOTE_BASIC_BLOCK(INSN) XCBBDEF (INSN, 3, NOTE)
|
|||
|
#define NOTE_VAR_LOCATION(INSN) XCEXP (INSN, 3, NOTE)
|
|||
|
#define NOTE_MARKER_LOCATION(INSN) XCUINT (INSN, 3, NOTE)
|
|||
|
#define NOTE_CFI(INSN) XCCFI (INSN, 3, NOTE)
|
|||
|
#define NOTE_LABEL_NUMBER(INSN) XCINT (INSN, 3, NOTE)
|
|||
|
|
|||
|
/* In a NOTE that is a line number, this is the line number.
|
|||
|
Other kinds of NOTEs are identified by negative numbers here. */
|
|||
|
#define NOTE_KIND(INSN) XCINT (INSN, 4, NOTE)
|
|||
|
|
|||
|
/* Nonzero if INSN is a note marking the beginning of a basic block. */
|
|||
|
#define NOTE_INSN_BASIC_BLOCK_P(INSN) \
|
|||
|
(NOTE_P (INSN) && NOTE_KIND (INSN) == NOTE_INSN_BASIC_BLOCK)
|
|||
|
|
|||
|
/* Nonzero if INSN is a debug nonbind marker note,
|
|||
|
for which NOTE_MARKER_LOCATION can be used. */
|
|||
|
#define NOTE_MARKER_P(INSN) \
|
|||
|
(NOTE_P (INSN) && \
|
|||
|
(NOTE_KIND (INSN) == NOTE_INSN_BEGIN_STMT \
|
|||
|
|| NOTE_KIND (INSN) == NOTE_INSN_INLINE_ENTRY))
|
|||
|
|
|||
|
/* Variable declaration and the location of a variable. */
|
|||
|
#define PAT_VAR_LOCATION_DECL(PAT) (XCTREE ((PAT), 0, VAR_LOCATION))
|
|||
|
#define PAT_VAR_LOCATION_LOC(PAT) (XCEXP ((PAT), 1, VAR_LOCATION))
|
|||
|
|
|||
|
/* Initialization status of the variable in the location. Status
|
|||
|
can be unknown, uninitialized or initialized. See enumeration
|
|||
|
type below. */
|
|||
|
#define PAT_VAR_LOCATION_STATUS(PAT) \
|
|||
|
(RTL_FLAG_CHECK1 ("PAT_VAR_LOCATION_STATUS", PAT, VAR_LOCATION) \
|
|||
|
->u2.var_location_status)
|
|||
|
|
|||
|
/* Accessors for a NOTE_INSN_VAR_LOCATION. */
|
|||
|
#define NOTE_VAR_LOCATION_DECL(NOTE) \
|
|||
|
PAT_VAR_LOCATION_DECL (NOTE_VAR_LOCATION (NOTE))
|
|||
|
#define NOTE_VAR_LOCATION_LOC(NOTE) \
|
|||
|
PAT_VAR_LOCATION_LOC (NOTE_VAR_LOCATION (NOTE))
|
|||
|
#define NOTE_VAR_LOCATION_STATUS(NOTE) \
|
|||
|
PAT_VAR_LOCATION_STATUS (NOTE_VAR_LOCATION (NOTE))
|
|||
|
|
|||
|
/* Evaluate to TRUE if INSN is a debug insn that denotes a variable
|
|||
|
location/value tracking annotation. */
|
|||
|
#define DEBUG_BIND_INSN_P(INSN) \
|
|||
|
(DEBUG_INSN_P (INSN) \
|
|||
|
&& (GET_CODE (PATTERN (INSN)) \
|
|||
|
== VAR_LOCATION))
|
|||
|
/* Evaluate to TRUE if INSN is a debug insn that denotes a program
|
|||
|
source location marker. */
|
|||
|
#define DEBUG_MARKER_INSN_P(INSN) \
|
|||
|
(DEBUG_INSN_P (INSN) \
|
|||
|
&& (GET_CODE (PATTERN (INSN)) \
|
|||
|
!= VAR_LOCATION))
|
|||
|
/* Evaluate to the marker kind. */
|
|||
|
#define INSN_DEBUG_MARKER_KIND(INSN) \
|
|||
|
(GET_CODE (PATTERN (INSN)) == DEBUG_MARKER \
|
|||
|
? (GET_MODE (PATTERN (INSN)) == VOIDmode \
|
|||
|
? NOTE_INSN_BEGIN_STMT \
|
|||
|
: GET_MODE (PATTERN (INSN)) == BLKmode \
|
|||
|
? NOTE_INSN_INLINE_ENTRY \
|
|||
|
: (enum insn_note)-1) \
|
|||
|
: (enum insn_note)-1)
|
|||
|
/* Create patterns for debug markers. These and the above abstract
|
|||
|
the representation, so that it's easier to get rid of the abuse of
|
|||
|
the mode to hold the marker kind. Other marker types are
|
|||
|
envisioned, so a single bit flag won't do; maybe separate RTL codes
|
|||
|
wouldn't be a problem. */
|
|||
|
#define GEN_RTX_DEBUG_MARKER_BEGIN_STMT_PAT() \
|
|||
|
gen_rtx_DEBUG_MARKER (VOIDmode)
|
|||
|
#define GEN_RTX_DEBUG_MARKER_INLINE_ENTRY_PAT() \
|
|||
|
gen_rtx_DEBUG_MARKER (BLKmode)
|
|||
|
|
|||
|
/* The VAR_LOCATION rtx in a DEBUG_INSN. */
|
|||
|
#define INSN_VAR_LOCATION(INSN) \
|
|||
|
(RTL_FLAG_CHECK1 ("INSN_VAR_LOCATION", PATTERN (INSN), VAR_LOCATION))
|
|||
|
/* A pointer to the VAR_LOCATION rtx in a DEBUG_INSN. */
|
|||
|
#define INSN_VAR_LOCATION_PTR(INSN) \
|
|||
|
(&PATTERN (INSN))
|
|||
|
|
|||
|
/* Accessors for a tree-expanded var location debug insn. */
|
|||
|
#define INSN_VAR_LOCATION_DECL(INSN) \
|
|||
|
PAT_VAR_LOCATION_DECL (INSN_VAR_LOCATION (INSN))
|
|||
|
#define INSN_VAR_LOCATION_LOC(INSN) \
|
|||
|
PAT_VAR_LOCATION_LOC (INSN_VAR_LOCATION (INSN))
|
|||
|
#define INSN_VAR_LOCATION_STATUS(INSN) \
|
|||
|
PAT_VAR_LOCATION_STATUS (INSN_VAR_LOCATION (INSN))
|
|||
|
|
|||
|
/* Expand to the RTL that denotes an unknown variable location in a
|
|||
|
DEBUG_INSN. */
|
|||
|
#define gen_rtx_UNKNOWN_VAR_LOC() (gen_rtx_CLOBBER (VOIDmode, const0_rtx))
|
|||
|
|
|||
|
/* Determine whether X is such an unknown location. */
|
|||
|
#define VAR_LOC_UNKNOWN_P(X) \
|
|||
|
(GET_CODE (X) == CLOBBER && XEXP ((X), 0) == const0_rtx)
|
|||
|
|
|||
|
/* 1 if RTX is emitted after a call, but it should take effect before
|
|||
|
the call returns. */
|
|||
|
#define NOTE_DURING_CALL_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("NOTE_VAR_LOCATION_DURING_CALL_P", (RTX), NOTE)->call)
|
|||
|
|
|||
|
/* DEBUG_EXPR_DECL corresponding to a DEBUG_EXPR RTX. */
|
|||
|
#define DEBUG_EXPR_TREE_DECL(RTX) XCTREE (RTX, 0, DEBUG_EXPR)
|
|||
|
|
|||
|
/* VAR_DECL/PARM_DECL DEBUG_IMPLICIT_PTR takes address of. */
|
|||
|
#define DEBUG_IMPLICIT_PTR_DECL(RTX) XCTREE (RTX, 0, DEBUG_IMPLICIT_PTR)
|
|||
|
|
|||
|
/* PARM_DECL DEBUG_PARAMETER_REF references. */
|
|||
|
#define DEBUG_PARAMETER_REF_DECL(RTX) XCTREE (RTX, 0, DEBUG_PARAMETER_REF)
|
|||
|
|
|||
|
/* Codes that appear in the NOTE_KIND field for kinds of notes
|
|||
|
that are not line numbers. These codes are all negative.
|
|||
|
|
|||
|
Notice that we do not try to use zero here for any of
|
|||
|
the special note codes because sometimes the source line
|
|||
|
actually can be zero! This happens (for example) when we
|
|||
|
are generating code for the per-translation-unit constructor
|
|||
|
and destructor routines for some C++ translation unit. */
|
|||
|
|
|||
|
enum insn_note
|
|||
|
{
|
|||
|
#define DEF_INSN_NOTE(NAME) NAME,
|
|||
|
#include "insn-notes.def"
|
|||
|
#undef DEF_INSN_NOTE
|
|||
|
|
|||
|
NOTE_INSN_MAX
|
|||
|
};
|
|||
|
|
|||
|
/* Names for NOTE insn's other than line numbers. */
|
|||
|
|
|||
|
extern const char * const note_insn_name[NOTE_INSN_MAX];
|
|||
|
#define GET_NOTE_INSN_NAME(NOTE_CODE) \
|
|||
|
(note_insn_name[(NOTE_CODE)])
|
|||
|
|
|||
|
/* The name of a label, in case it corresponds to an explicit label
|
|||
|
in the input source code. */
|
|||
|
#define LABEL_NAME(RTX) XCSTR (RTX, 6, CODE_LABEL)
|
|||
|
|
|||
|
/* In jump.cc, each label contains a count of the number
|
|||
|
of LABEL_REFs that point at it, so unused labels can be deleted. */
|
|||
|
#define LABEL_NUSES(RTX) XCINT (RTX, 4, CODE_LABEL)
|
|||
|
|
|||
|
/* Labels carry a two-bit field composed of the ->jump and ->call
|
|||
|
bits. This field indicates whether the label is an alternate
|
|||
|
entry point, and if so, what kind. */
|
|||
|
enum label_kind
|
|||
|
{
|
|||
|
LABEL_NORMAL = 0, /* ordinary label */
|
|||
|
LABEL_STATIC_ENTRY, /* alternate entry point, not exported */
|
|||
|
LABEL_GLOBAL_ENTRY, /* alternate entry point, exported */
|
|||
|
LABEL_WEAK_ENTRY /* alternate entry point, exported as weak symbol */
|
|||
|
};
|
|||
|
|
|||
|
#if defined ENABLE_RTL_FLAG_CHECKING && (GCC_VERSION > 2007)
|
|||
|
|
|||
|
/* Retrieve the kind of LABEL. */
|
|||
|
#define LABEL_KIND(LABEL) __extension__ \
|
|||
|
({ __typeof (LABEL) const _label = (LABEL); \
|
|||
|
if (! LABEL_P (_label)) \
|
|||
|
rtl_check_failed_flag ("LABEL_KIND", _label, __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
(enum label_kind) ((_label->jump << 1) | _label->call); })
|
|||
|
|
|||
|
/* Set the kind of LABEL. */
|
|||
|
#define SET_LABEL_KIND(LABEL, KIND) do { \
|
|||
|
__typeof (LABEL) const _label = (LABEL); \
|
|||
|
const unsigned int _kind = (KIND); \
|
|||
|
if (! LABEL_P (_label)) \
|
|||
|
rtl_check_failed_flag ("SET_LABEL_KIND", _label, __FILE__, __LINE__, \
|
|||
|
__FUNCTION__); \
|
|||
|
_label->jump = ((_kind >> 1) & 1); \
|
|||
|
_label->call = (_kind & 1); \
|
|||
|
} while (0)
|
|||
|
|
|||
|
#else
|
|||
|
|
|||
|
/* Retrieve the kind of LABEL. */
|
|||
|
#define LABEL_KIND(LABEL) \
|
|||
|
((enum label_kind) (((LABEL)->jump << 1) | (LABEL)->call))
|
|||
|
|
|||
|
/* Set the kind of LABEL. */
|
|||
|
#define SET_LABEL_KIND(LABEL, KIND) do { \
|
|||
|
rtx const _label = (LABEL); \
|
|||
|
const unsigned int _kind = (KIND); \
|
|||
|
_label->jump = ((_kind >> 1) & 1); \
|
|||
|
_label->call = (_kind & 1); \
|
|||
|
} while (0)
|
|||
|
|
|||
|
#endif /* rtl flag checking */
|
|||
|
|
|||
|
#define LABEL_ALT_ENTRY_P(LABEL) (LABEL_KIND (LABEL) != LABEL_NORMAL)
|
|||
|
|
|||
|
/* In jump.cc, each JUMP_INSN can point to a label that it can jump to,
|
|||
|
so that if the JUMP_INSN is deleted, the label's LABEL_NUSES can
|
|||
|
be decremented and possibly the label can be deleted. */
|
|||
|
#define JUMP_LABEL(INSN) XCEXP (INSN, 7, JUMP_INSN)
|
|||
|
|
|||
|
inline rtx_insn *JUMP_LABEL_AS_INSN (const rtx_insn *insn)
|
|||
|
{
|
|||
|
return safe_as_a <rtx_insn *> (JUMP_LABEL (insn));
|
|||
|
}
|
|||
|
|
|||
|
/* Methods of rtx_jump_insn. */
|
|||
|
|
|||
|
inline rtx rtx_jump_insn::jump_label () const
|
|||
|
{
|
|||
|
return JUMP_LABEL (this);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx_code_label *rtx_jump_insn::jump_target () const
|
|||
|
{
|
|||
|
return safe_as_a <rtx_code_label *> (JUMP_LABEL (this));
|
|||
|
}
|
|||
|
|
|||
|
inline void rtx_jump_insn::set_jump_target (rtx_code_label *target)
|
|||
|
{
|
|||
|
JUMP_LABEL (this) = target;
|
|||
|
}
|
|||
|
|
|||
|
/* Once basic blocks are found, each CODE_LABEL starts a chain that
|
|||
|
goes through all the LABEL_REFs that jump to that label. The chain
|
|||
|
eventually winds up at the CODE_LABEL: it is circular. */
|
|||
|
#define LABEL_REFS(LABEL) XCEXP (LABEL, 3, CODE_LABEL)
|
|||
|
|
|||
|
/* Get the label that a LABEL_REF references. */
|
|||
|
inline rtx_insn *
|
|||
|
label_ref_label (const_rtx ref)
|
|||
|
{
|
|||
|
return as_a<rtx_insn *> (XCEXP (ref, 0, LABEL_REF));
|
|||
|
}
|
|||
|
|
|||
|
/* Set the label that LABEL_REF ref refers to. */
|
|||
|
|
|||
|
inline void
|
|||
|
set_label_ref_label (rtx ref, rtx_insn *label)
|
|||
|
{
|
|||
|
XCEXP (ref, 0, LABEL_REF) = label;
|
|||
|
}
|
|||
|
|
|||
|
/* For a REG rtx, REGNO extracts the register number. REGNO can only
|
|||
|
be used on RHS. Use SET_REGNO to change the value. */
|
|||
|
#define REGNO(RTX) (rhs_regno(RTX))
|
|||
|
#define SET_REGNO(RTX, N) (df_ref_change_reg_with_loc (RTX, N))
|
|||
|
|
|||
|
/* Return the number of consecutive registers in a REG. This is always
|
|||
|
1 for pseudo registers and is determined by TARGET_HARD_REGNO_NREGS for
|
|||
|
hard registers. */
|
|||
|
#define REG_NREGS(RTX) (REG_CHECK (RTX)->nregs)
|
|||
|
|
|||
|
/* ORIGINAL_REGNO holds the number the register originally had; for a
|
|||
|
pseudo register turned into a hard reg this will hold the old pseudo
|
|||
|
register number. */
|
|||
|
#define ORIGINAL_REGNO(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("ORIGINAL_REGNO", (RTX), REG)->u2.original_regno)
|
|||
|
|
|||
|
/* Force the REGNO macro to only be used on the lhs. */
|
|||
|
inline unsigned int
|
|||
|
rhs_regno (const_rtx x)
|
|||
|
{
|
|||
|
return REG_CHECK (x)->regno;
|
|||
|
}
|
|||
|
|
|||
|
/* Return the final register in REG X plus one. */
|
|||
|
inline unsigned int
|
|||
|
END_REGNO (const_rtx x)
|
|||
|
{
|
|||
|
return REGNO (x) + REG_NREGS (x);
|
|||
|
}
|
|||
|
|
|||
|
/* Change the REGNO and REG_NREGS of REG X to the specified values,
|
|||
|
bypassing the df machinery. */
|
|||
|
inline void
|
|||
|
set_regno_raw (rtx x, unsigned int regno, unsigned int nregs)
|
|||
|
{
|
|||
|
reg_info *reg = REG_CHECK (x);
|
|||
|
reg->regno = regno;
|
|||
|
reg->nregs = nregs;
|
|||
|
}
|
|||
|
|
|||
|
/* 1 if RTX is a reg or parallel that is the current function's return
|
|||
|
value. */
|
|||
|
#define REG_FUNCTION_VALUE_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK2 ("REG_FUNCTION_VALUE_P", (RTX), REG, PARALLEL)->return_val)
|
|||
|
|
|||
|
/* 1 if RTX is a reg that corresponds to a variable declared by the user. */
|
|||
|
#define REG_USERVAR_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("REG_USERVAR_P", (RTX), REG)->volatil)
|
|||
|
|
|||
|
/* 1 if RTX is a reg that holds a pointer value. */
|
|||
|
#define REG_POINTER(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("REG_POINTER", (RTX), REG)->frame_related)
|
|||
|
|
|||
|
/* 1 if RTX is a mem that holds a pointer value. */
|
|||
|
#define MEM_POINTER(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("MEM_POINTER", (RTX), MEM)->frame_related)
|
|||
|
|
|||
|
/* 1 if the given register REG corresponds to a hard register. */
|
|||
|
#define HARD_REGISTER_P(REG) (HARD_REGISTER_NUM_P (REGNO (REG)))
|
|||
|
|
|||
|
/* 1 if the given register number REG_NO corresponds to a hard register. */
|
|||
|
#define HARD_REGISTER_NUM_P(REG_NO) ((REG_NO) < FIRST_PSEUDO_REGISTER)
|
|||
|
|
|||
|
/* For a CONST_INT rtx, INTVAL extracts the integer. */
|
|||
|
#define INTVAL(RTX) XCWINT (RTX, 0, CONST_INT)
|
|||
|
#define UINTVAL(RTX) ((unsigned HOST_WIDE_INT) INTVAL (RTX))
|
|||
|
|
|||
|
/* For a CONST_WIDE_INT, CONST_WIDE_INT_NUNITS is the number of
|
|||
|
elements actually needed to represent the constant.
|
|||
|
CONST_WIDE_INT_ELT gets one of the elements. 0 is the least
|
|||
|
significant HOST_WIDE_INT. */
|
|||
|
#define CONST_WIDE_INT_VEC(RTX) HWIVEC_CHECK (RTX, CONST_WIDE_INT)
|
|||
|
#define CONST_WIDE_INT_NUNITS(RTX) CWI_GET_NUM_ELEM (RTX)
|
|||
|
#define CONST_WIDE_INT_ELT(RTX, N) CWI_ELT (RTX, N)
|
|||
|
|
|||
|
/* For a CONST_POLY_INT, CONST_POLY_INT_COEFFS gives access to the
|
|||
|
individual coefficients, in the form of a trailing_wide_ints structure. */
|
|||
|
#define CONST_POLY_INT_COEFFS(RTX) \
|
|||
|
(RTL_FLAG_CHECK1("CONST_POLY_INT_COEFFS", (RTX), \
|
|||
|
CONST_POLY_INT)->u.cpi.coeffs)
|
|||
|
|
|||
|
/* For a CONST_DOUBLE:
|
|||
|
#if TARGET_SUPPORTS_WIDE_INT == 0
|
|||
|
For a VOIDmode, there are two integers CONST_DOUBLE_LOW is the
|
|||
|
low-order word and ..._HIGH the high-order.
|
|||
|
#endif
|
|||
|
For a float, there is a REAL_VALUE_TYPE structure, and
|
|||
|
CONST_DOUBLE_REAL_VALUE(r) is a pointer to it. */
|
|||
|
#define CONST_DOUBLE_LOW(r) XCMWINT (r, 0, CONST_DOUBLE, VOIDmode)
|
|||
|
#define CONST_DOUBLE_HIGH(r) XCMWINT (r, 1, CONST_DOUBLE, VOIDmode)
|
|||
|
#define CONST_DOUBLE_REAL_VALUE(r) \
|
|||
|
((const struct real_value *) XCNMPRV (r, CONST_DOUBLE, VOIDmode))
|
|||
|
|
|||
|
#define CONST_FIXED_VALUE(r) \
|
|||
|
((const struct fixed_value *) XCNMPFV (r, CONST_FIXED, VOIDmode))
|
|||
|
#define CONST_FIXED_VALUE_HIGH(r) \
|
|||
|
((HOST_WIDE_INT) (CONST_FIXED_VALUE (r)->data.high))
|
|||
|
#define CONST_FIXED_VALUE_LOW(r) \
|
|||
|
((HOST_WIDE_INT) (CONST_FIXED_VALUE (r)->data.low))
|
|||
|
|
|||
|
/* For a CONST_VECTOR, return element #n. */
|
|||
|
#define CONST_VECTOR_ELT(RTX, N) const_vector_elt (RTX, N)
|
|||
|
|
|||
|
/* See rtl.texi for a description of these macros. */
|
|||
|
#define CONST_VECTOR_NPATTERNS(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("CONST_VECTOR_NPATTERNS", (RTX), CONST_VECTOR) \
|
|||
|
->u2.const_vector.npatterns)
|
|||
|
|
|||
|
#define CONST_VECTOR_NELTS_PER_PATTERN(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("CONST_VECTOR_NELTS_PER_PATTERN", (RTX), CONST_VECTOR) \
|
|||
|
->u2.const_vector.nelts_per_pattern)
|
|||
|
|
|||
|
#define CONST_VECTOR_DUPLICATE_P(RTX) \
|
|||
|
(CONST_VECTOR_NELTS_PER_PATTERN (RTX) == 1)
|
|||
|
|
|||
|
#define CONST_VECTOR_STEPPED_P(RTX) \
|
|||
|
(CONST_VECTOR_NELTS_PER_PATTERN (RTX) == 3)
|
|||
|
|
|||
|
#define CONST_VECTOR_ENCODED_ELT(RTX, N) XCVECEXP (RTX, 0, N, CONST_VECTOR)
|
|||
|
|
|||
|
/* Return the number of elements encoded directly in a CONST_VECTOR. */
|
|||
|
|
|||
|
inline unsigned int
|
|||
|
const_vector_encoded_nelts (const_rtx x)
|
|||
|
{
|
|||
|
return CONST_VECTOR_NPATTERNS (x) * CONST_VECTOR_NELTS_PER_PATTERN (x);
|
|||
|
}
|
|||
|
|
|||
|
/* For a CONST_VECTOR, return the number of elements in a vector. */
|
|||
|
#define CONST_VECTOR_NUNITS(RTX) GET_MODE_NUNITS (GET_MODE (RTX))
|
|||
|
|
|||
|
/* For a SUBREG rtx, SUBREG_REG extracts the value we want a subreg of.
|
|||
|
SUBREG_BYTE extracts the byte-number. */
|
|||
|
|
|||
|
#define SUBREG_REG(RTX) XCEXP (RTX, 0, SUBREG)
|
|||
|
#define SUBREG_BYTE(RTX) XCSUBREG (RTX, 1, SUBREG)
|
|||
|
|
|||
|
/* in rtlanal.cc */
|
|||
|
/* Return the right cost to give to an operation
|
|||
|
to make the cost of the corresponding register-to-register instruction
|
|||
|
N times that of a fast register-to-register instruction. */
|
|||
|
#define COSTS_N_INSNS(N) ((N) * 4)
|
|||
|
|
|||
|
/* Maximum cost of an rtl expression. This value has the special meaning
|
|||
|
not to use an rtx with this cost under any circumstances. */
|
|||
|
#define MAX_COST INT_MAX
|
|||
|
|
|||
|
/* Return true if CODE always has VOIDmode. */
|
|||
|
|
|||
|
inline bool
|
|||
|
always_void_p (enum rtx_code code)
|
|||
|
{
|
|||
|
return code == SET;
|
|||
|
}
|
|||
|
|
|||
|
/* A structure to hold all available cost information about an rtl
|
|||
|
expression. */
|
|||
|
struct full_rtx_costs
|
|||
|
{
|
|||
|
int speed;
|
|||
|
int size;
|
|||
|
};
|
|||
|
|
|||
|
/* Initialize a full_rtx_costs structure C to the maximum cost. */
|
|||
|
inline void
|
|||
|
init_costs_to_max (struct full_rtx_costs *c)
|
|||
|
{
|
|||
|
c->speed = MAX_COST;
|
|||
|
c->size = MAX_COST;
|
|||
|
}
|
|||
|
|
|||
|
/* Initialize a full_rtx_costs structure C to zero cost. */
|
|||
|
inline void
|
|||
|
init_costs_to_zero (struct full_rtx_costs *c)
|
|||
|
{
|
|||
|
c->speed = 0;
|
|||
|
c->size = 0;
|
|||
|
}
|
|||
|
|
|||
|
/* Compare two full_rtx_costs structures A and B, returning true
|
|||
|
if A < B when optimizing for speed. */
|
|||
|
inline bool
|
|||
|
costs_lt_p (struct full_rtx_costs *a, struct full_rtx_costs *b,
|
|||
|
bool speed)
|
|||
|
{
|
|||
|
if (speed)
|
|||
|
return (a->speed < b->speed
|
|||
|
|| (a->speed == b->speed && a->size < b->size));
|
|||
|
else
|
|||
|
return (a->size < b->size
|
|||
|
|| (a->size == b->size && a->speed < b->speed));
|
|||
|
}
|
|||
|
|
|||
|
/* Increase both members of the full_rtx_costs structure C by the
|
|||
|
cost of N insns. */
|
|||
|
inline void
|
|||
|
costs_add_n_insns (struct full_rtx_costs *c, int n)
|
|||
|
{
|
|||
|
c->speed += COSTS_N_INSNS (n);
|
|||
|
c->size += COSTS_N_INSNS (n);
|
|||
|
}
|
|||
|
|
|||
|
/* Describes the shape of a subreg:
|
|||
|
|
|||
|
inner_mode == the mode of the SUBREG_REG
|
|||
|
offset == the SUBREG_BYTE
|
|||
|
outer_mode == the mode of the SUBREG itself. */
|
|||
|
class subreg_shape {
|
|||
|
public:
|
|||
|
subreg_shape (machine_mode, poly_uint16, machine_mode);
|
|||
|
bool operator == (const subreg_shape &) const;
|
|||
|
bool operator != (const subreg_shape &) const;
|
|||
|
unsigned HOST_WIDE_INT unique_id () const;
|
|||
|
|
|||
|
machine_mode inner_mode;
|
|||
|
poly_uint16 offset;
|
|||
|
machine_mode outer_mode;
|
|||
|
};
|
|||
|
|
|||
|
inline
|
|||
|
subreg_shape::subreg_shape (machine_mode inner_mode_in,
|
|||
|
poly_uint16 offset_in,
|
|||
|
machine_mode outer_mode_in)
|
|||
|
: inner_mode (inner_mode_in), offset (offset_in), outer_mode (outer_mode_in)
|
|||
|
{}
|
|||
|
|
|||
|
inline bool
|
|||
|
subreg_shape::operator == (const subreg_shape &other) const
|
|||
|
{
|
|||
|
return (inner_mode == other.inner_mode
|
|||
|
&& known_eq (offset, other.offset)
|
|||
|
&& outer_mode == other.outer_mode);
|
|||
|
}
|
|||
|
|
|||
|
inline bool
|
|||
|
subreg_shape::operator != (const subreg_shape &other) const
|
|||
|
{
|
|||
|
return !operator == (other);
|
|||
|
}
|
|||
|
|
|||
|
/* Return an integer that uniquely identifies this shape. Structures
|
|||
|
like rtx_def assume that a mode can fit in an 8-bit bitfield and no
|
|||
|
current mode is anywhere near being 65536 bytes in size, so the
|
|||
|
id comfortably fits in an int. */
|
|||
|
|
|||
|
inline unsigned HOST_WIDE_INT
|
|||
|
subreg_shape::unique_id () const
|
|||
|
{
|
|||
|
{ STATIC_ASSERT (MAX_MACHINE_MODE <= 256); }
|
|||
|
{ STATIC_ASSERT (NUM_POLY_INT_COEFFS <= 3); }
|
|||
|
{ STATIC_ASSERT (sizeof (offset.coeffs[0]) <= 2); }
|
|||
|
int res = (int) inner_mode + ((int) outer_mode << 8);
|
|||
|
for (int i = 0; i < NUM_POLY_INT_COEFFS; ++i)
|
|||
|
res += (HOST_WIDE_INT) offset.coeffs[i] << ((1 + i) * 16);
|
|||
|
return res;
|
|||
|
}
|
|||
|
|
|||
|
/* Return the shape of a SUBREG rtx. */
|
|||
|
|
|||
|
inline subreg_shape
|
|||
|
shape_of_subreg (const_rtx x)
|
|||
|
{
|
|||
|
return subreg_shape (GET_MODE (SUBREG_REG (x)),
|
|||
|
SUBREG_BYTE (x), GET_MODE (x));
|
|||
|
}
|
|||
|
|
|||
|
/* Information about an address. This structure is supposed to be able
|
|||
|
to represent all supported target addresses. Please extend it if it
|
|||
|
is not yet general enough. */
|
|||
|
struct address_info {
|
|||
|
/* The mode of the value being addressed, or VOIDmode if this is
|
|||
|
a load-address operation with no known address mode. */
|
|||
|
machine_mode mode;
|
|||
|
|
|||
|
/* The address space. */
|
|||
|
addr_space_t as;
|
|||
|
|
|||
|
/* True if this is an RTX_AUTOINC address. */
|
|||
|
bool autoinc_p;
|
|||
|
|
|||
|
/* A pointer to the top-level address. */
|
|||
|
rtx *outer;
|
|||
|
|
|||
|
/* A pointer to the inner address, after all address mutations
|
|||
|
have been stripped from the top-level address. It can be one
|
|||
|
of the following:
|
|||
|
|
|||
|
- A {PRE,POST}_{INC,DEC} of *BASE. SEGMENT, INDEX and DISP are null.
|
|||
|
|
|||
|
- A {PRE,POST}_MODIFY of *BASE. In this case either INDEX or DISP
|
|||
|
points to the step value, depending on whether the step is variable
|
|||
|
or constant respectively. SEGMENT is null.
|
|||
|
|
|||
|
- A plain sum of the form SEGMENT + BASE + INDEX + DISP,
|
|||
|
with null fields evaluating to 0. */
|
|||
|
rtx *inner;
|
|||
|
|
|||
|
/* Components that make up *INNER. Each one may be null or nonnull.
|
|||
|
When nonnull, their meanings are as follows:
|
|||
|
|
|||
|
- *SEGMENT is the "segment" of memory to which the address refers.
|
|||
|
This value is entirely target-specific and is only called a "segment"
|
|||
|
because that's its most typical use. It contains exactly one UNSPEC,
|
|||
|
pointed to by SEGMENT_TERM. The contents of *SEGMENT do not need
|
|||
|
reloading.
|
|||
|
|
|||
|
- *BASE is a variable expression representing a base address.
|
|||
|
It contains exactly one REG, SUBREG or MEM, pointed to by BASE_TERM.
|
|||
|
|
|||
|
- *INDEX is a variable expression representing an index value.
|
|||
|
It may be a scaled expression, such as a MULT. It has exactly
|
|||
|
one REG, SUBREG or MEM, pointed to by INDEX_TERM.
|
|||
|
|
|||
|
- *DISP is a constant, possibly mutated. DISP_TERM points to the
|
|||
|
unmutated RTX_CONST_OBJ. */
|
|||
|
rtx *segment;
|
|||
|
rtx *base;
|
|||
|
rtx *index;
|
|||
|
rtx *disp;
|
|||
|
|
|||
|
rtx *segment_term;
|
|||
|
rtx *base_term;
|
|||
|
rtx *index_term;
|
|||
|
rtx *disp_term;
|
|||
|
|
|||
|
/* In a {PRE,POST}_MODIFY address, this points to a second copy
|
|||
|
of BASE_TERM, otherwise it is null. */
|
|||
|
rtx *base_term2;
|
|||
|
|
|||
|
/* ADDRESS if this structure describes an address operand, MEM if
|
|||
|
it describes a MEM address. */
|
|||
|
enum rtx_code addr_outer_code;
|
|||
|
|
|||
|
/* If BASE is nonnull, this is the code of the rtx that contains it. */
|
|||
|
enum rtx_code base_outer_code;
|
|||
|
};
|
|||
|
|
|||
|
/* This is used to bundle an rtx and a mode together so that the pair
|
|||
|
can be used with the wi:: routines. If we ever put modes into rtx
|
|||
|
integer constants, this should go away and then just pass an rtx in. */
|
|||
|
typedef std::pair <rtx, machine_mode> rtx_mode_t;
|
|||
|
|
|||
|
namespace wi
|
|||
|
{
|
|||
|
template <>
|
|||
|
struct int_traits <rtx_mode_t>
|
|||
|
{
|
|||
|
static const enum precision_type precision_type = VAR_PRECISION;
|
|||
|
static const bool host_dependent_precision = false;
|
|||
|
/* This ought to be true, except for the special case that BImode
|
|||
|
is canonicalized to STORE_FLAG_VALUE, which might be 1. */
|
|||
|
static const bool is_sign_extended = false;
|
|||
|
static unsigned int get_precision (const rtx_mode_t &);
|
|||
|
static wi::storage_ref decompose (HOST_WIDE_INT *, unsigned int,
|
|||
|
const rtx_mode_t &);
|
|||
|
};
|
|||
|
}
|
|||
|
|
|||
|
inline unsigned int
|
|||
|
wi::int_traits <rtx_mode_t>::get_precision (const rtx_mode_t &x)
|
|||
|
{
|
|||
|
return GET_MODE_PRECISION (as_a <scalar_mode> (x.second));
|
|||
|
}
|
|||
|
|
|||
|
inline wi::storage_ref
|
|||
|
wi::int_traits <rtx_mode_t>::decompose (HOST_WIDE_INT *,
|
|||
|
unsigned int precision,
|
|||
|
const rtx_mode_t &x)
|
|||
|
{
|
|||
|
gcc_checking_assert (precision == get_precision (x));
|
|||
|
switch (GET_CODE (x.first))
|
|||
|
{
|
|||
|
case CONST_INT:
|
|||
|
if (precision < HOST_BITS_PER_WIDE_INT)
|
|||
|
/* Nonzero BImodes are stored as STORE_FLAG_VALUE, which on many
|
|||
|
targets is 1 rather than -1. */
|
|||
|
gcc_checking_assert (INTVAL (x.first)
|
|||
|
== sext_hwi (INTVAL (x.first), precision)
|
|||
|
|| (x.second == BImode && INTVAL (x.first) == 1));
|
|||
|
|
|||
|
return wi::storage_ref (&INTVAL (x.first), 1, precision);
|
|||
|
|
|||
|
case CONST_WIDE_INT:
|
|||
|
return wi::storage_ref (&CONST_WIDE_INT_ELT (x.first, 0),
|
|||
|
CONST_WIDE_INT_NUNITS (x.first), precision);
|
|||
|
|
|||
|
#if TARGET_SUPPORTS_WIDE_INT == 0
|
|||
|
case CONST_DOUBLE:
|
|||
|
return wi::storage_ref (&CONST_DOUBLE_LOW (x.first), 2, precision);
|
|||
|
#endif
|
|||
|
|
|||
|
default:
|
|||
|
gcc_unreachable ();
|
|||
|
}
|
|||
|
}
|
|||
|
|
|||
|
namespace wi
|
|||
|
{
|
|||
|
hwi_with_prec shwi (HOST_WIDE_INT, machine_mode mode);
|
|||
|
wide_int min_value (machine_mode, signop);
|
|||
|
wide_int max_value (machine_mode, signop);
|
|||
|
}
|
|||
|
|
|||
|
inline wi::hwi_with_prec
|
|||
|
wi::shwi (HOST_WIDE_INT val, machine_mode mode)
|
|||
|
{
|
|||
|
return shwi (val, GET_MODE_PRECISION (as_a <scalar_mode> (mode)));
|
|||
|
}
|
|||
|
|
|||
|
/* Produce the smallest number that is represented in MODE. The precision
|
|||
|
is taken from MODE and the sign from SGN. */
|
|||
|
inline wide_int
|
|||
|
wi::min_value (machine_mode mode, signop sgn)
|
|||
|
{
|
|||
|
return min_value (GET_MODE_PRECISION (as_a <scalar_mode> (mode)), sgn);
|
|||
|
}
|
|||
|
|
|||
|
/* Produce the largest number that is represented in MODE. The precision
|
|||
|
is taken from MODE and the sign from SGN. */
|
|||
|
inline wide_int
|
|||
|
wi::max_value (machine_mode mode, signop sgn)
|
|||
|
{
|
|||
|
return max_value (GET_MODE_PRECISION (as_a <scalar_mode> (mode)), sgn);
|
|||
|
}
|
|||
|
|
|||
|
namespace wi
|
|||
|
{
|
|||
|
typedef poly_int<NUM_POLY_INT_COEFFS,
|
|||
|
generic_wide_int <wide_int_ref_storage <false, false> > >
|
|||
|
rtx_to_poly_wide_ref;
|
|||
|
rtx_to_poly_wide_ref to_poly_wide (const_rtx, machine_mode);
|
|||
|
}
|
|||
|
|
|||
|
/* Return the value of a CONST_POLY_INT in its native precision. */
|
|||
|
|
|||
|
inline wi::rtx_to_poly_wide_ref
|
|||
|
const_poly_int_value (const_rtx x)
|
|||
|
{
|
|||
|
poly_int<NUM_POLY_INT_COEFFS, WIDE_INT_REF_FOR (wide_int)> res;
|
|||
|
for (unsigned int i = 0; i < NUM_POLY_INT_COEFFS; ++i)
|
|||
|
res.coeffs[i] = CONST_POLY_INT_COEFFS (x)[i];
|
|||
|
return res;
|
|||
|
}
|
|||
|
|
|||
|
/* Return true if X is a scalar integer or a CONST_POLY_INT. The value
|
|||
|
can then be extracted using wi::to_poly_wide. */
|
|||
|
|
|||
|
inline bool
|
|||
|
poly_int_rtx_p (const_rtx x)
|
|||
|
{
|
|||
|
return CONST_SCALAR_INT_P (x) || CONST_POLY_INT_P (x);
|
|||
|
}
|
|||
|
|
|||
|
/* Access X (which satisfies poly_int_rtx_p) as a poly_wide_int.
|
|||
|
MODE is the mode of X. */
|
|||
|
|
|||
|
inline wi::rtx_to_poly_wide_ref
|
|||
|
wi::to_poly_wide (const_rtx x, machine_mode mode)
|
|||
|
{
|
|||
|
if (CONST_POLY_INT_P (x))
|
|||
|
return const_poly_int_value (x);
|
|||
|
return rtx_mode_t (const_cast<rtx> (x), mode);
|
|||
|
}
|
|||
|
|
|||
|
/* Return the value of X as a poly_int64. */
|
|||
|
|
|||
|
inline poly_int64
|
|||
|
rtx_to_poly_int64 (const_rtx x)
|
|||
|
{
|
|||
|
if (CONST_POLY_INT_P (x))
|
|||
|
{
|
|||
|
poly_int64 res;
|
|||
|
for (unsigned int i = 0; i < NUM_POLY_INT_COEFFS; ++i)
|
|||
|
res.coeffs[i] = CONST_POLY_INT_COEFFS (x)[i].to_shwi ();
|
|||
|
return res;
|
|||
|
}
|
|||
|
return INTVAL (x);
|
|||
|
}
|
|||
|
|
|||
|
/* Return true if arbitrary value X is an integer constant that can
|
|||
|
be represented as a poly_int64. Store the value in *RES if so,
|
|||
|
otherwise leave it unmodified. */
|
|||
|
|
|||
|
inline bool
|
|||
|
poly_int_rtx_p (const_rtx x, poly_int64_pod *res)
|
|||
|
{
|
|||
|
if (CONST_INT_P (x))
|
|||
|
{
|
|||
|
*res = INTVAL (x);
|
|||
|
return true;
|
|||
|
}
|
|||
|
if (CONST_POLY_INT_P (x))
|
|||
|
{
|
|||
|
for (unsigned int i = 0; i < NUM_POLY_INT_COEFFS; ++i)
|
|||
|
if (!wi::fits_shwi_p (CONST_POLY_INT_COEFFS (x)[i]))
|
|||
|
return false;
|
|||
|
for (unsigned int i = 0; i < NUM_POLY_INT_COEFFS; ++i)
|
|||
|
res->coeffs[i] = CONST_POLY_INT_COEFFS (x)[i].to_shwi ();
|
|||
|
return true;
|
|||
|
}
|
|||
|
return false;
|
|||
|
}
|
|||
|
|
|||
|
extern void init_rtlanal (void);
|
|||
|
extern int rtx_cost (rtx, machine_mode, enum rtx_code, int, bool);
|
|||
|
extern int address_cost (rtx, machine_mode, addr_space_t, bool);
|
|||
|
extern void get_full_rtx_cost (rtx, machine_mode, enum rtx_code, int,
|
|||
|
struct full_rtx_costs *);
|
|||
|
extern bool native_encode_rtx (machine_mode, rtx, vec<target_unit> &,
|
|||
|
unsigned int, unsigned int);
|
|||
|
extern rtx native_decode_rtx (machine_mode, const vec<target_unit> &,
|
|||
|
unsigned int);
|
|||
|
extern rtx native_decode_vector_rtx (machine_mode, const vec<target_unit> &,
|
|||
|
unsigned int, unsigned int, unsigned int);
|
|||
|
extern poly_uint64 subreg_lsb (const_rtx);
|
|||
|
extern poly_uint64 subreg_size_lsb (poly_uint64, poly_uint64, poly_uint64);
|
|||
|
extern poly_uint64 subreg_size_offset_from_lsb (poly_uint64, poly_uint64,
|
|||
|
poly_uint64);
|
|||
|
extern bool read_modify_subreg_p (const_rtx);
|
|||
|
|
|||
|
/* Given a subreg's OUTER_MODE, INNER_MODE, and SUBREG_BYTE, return the
|
|||
|
bit offset at which the subreg begins (counting from the least significant
|
|||
|
bit of the operand). */
|
|||
|
|
|||
|
inline poly_uint64
|
|||
|
subreg_lsb_1 (machine_mode outer_mode, machine_mode inner_mode,
|
|||
|
poly_uint64 subreg_byte)
|
|||
|
{
|
|||
|
return subreg_size_lsb (GET_MODE_SIZE (outer_mode),
|
|||
|
GET_MODE_SIZE (inner_mode), subreg_byte);
|
|||
|
}
|
|||
|
|
|||
|
/* Return the subreg byte offset for a subreg whose outer mode is
|
|||
|
OUTER_MODE, whose inner mode is INNER_MODE, and where there are
|
|||
|
LSB_SHIFT *bits* between the lsb of the outer value and the lsb of
|
|||
|
the inner value. This is the inverse of subreg_lsb_1 (which converts
|
|||
|
byte offsets to bit shifts). */
|
|||
|
|
|||
|
inline poly_uint64
|
|||
|
subreg_offset_from_lsb (machine_mode outer_mode,
|
|||
|
machine_mode inner_mode,
|
|||
|
poly_uint64 lsb_shift)
|
|||
|
{
|
|||
|
return subreg_size_offset_from_lsb (GET_MODE_SIZE (outer_mode),
|
|||
|
GET_MODE_SIZE (inner_mode), lsb_shift);
|
|||
|
}
|
|||
|
|
|||
|
extern unsigned int subreg_regno_offset (unsigned int, machine_mode,
|
|||
|
poly_uint64, machine_mode);
|
|||
|
extern bool subreg_offset_representable_p (unsigned int, machine_mode,
|
|||
|
poly_uint64, machine_mode);
|
|||
|
extern unsigned int subreg_regno (const_rtx);
|
|||
|
extern int simplify_subreg_regno (unsigned int, machine_mode,
|
|||
|
poly_uint64, machine_mode);
|
|||
|
extern int lowpart_subreg_regno (unsigned int, machine_mode,
|
|||
|
machine_mode);
|
|||
|
extern unsigned int subreg_nregs (const_rtx);
|
|||
|
extern unsigned int subreg_nregs_with_regno (unsigned int, const_rtx);
|
|||
|
extern unsigned HOST_WIDE_INT nonzero_bits (const_rtx, machine_mode);
|
|||
|
extern unsigned int num_sign_bit_copies (const_rtx, machine_mode);
|
|||
|
extern bool constant_pool_constant_p (rtx);
|
|||
|
extern bool truncated_to_mode (machine_mode, const_rtx);
|
|||
|
extern int low_bitmask_len (machine_mode, unsigned HOST_WIDE_INT);
|
|||
|
extern void split_double (rtx, rtx *, rtx *);
|
|||
|
extern rtx *strip_address_mutations (rtx *, enum rtx_code * = 0);
|
|||
|
extern void decompose_address (struct address_info *, rtx *,
|
|||
|
machine_mode, addr_space_t, enum rtx_code);
|
|||
|
extern void decompose_lea_address (struct address_info *, rtx *);
|
|||
|
extern void decompose_mem_address (struct address_info *, rtx);
|
|||
|
extern void update_address (struct address_info *);
|
|||
|
extern HOST_WIDE_INT get_index_scale (const struct address_info *);
|
|||
|
extern enum rtx_code get_index_code (const struct address_info *);
|
|||
|
|
|||
|
/* 1 if RTX is a subreg containing a reg that is already known to be
|
|||
|
sign- or zero-extended from the mode of the subreg to the mode of
|
|||
|
the reg. SUBREG_PROMOTED_UNSIGNED_P gives the signedness of the
|
|||
|
extension.
|
|||
|
|
|||
|
When used as a LHS, is means that this extension must be done
|
|||
|
when assigning to SUBREG_REG. */
|
|||
|
|
|||
|
#define SUBREG_PROMOTED_VAR_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("SUBREG_PROMOTED", (RTX), SUBREG)->in_struct)
|
|||
|
|
|||
|
/* Valid for subregs which are SUBREG_PROMOTED_VAR_P(). In that case
|
|||
|
this gives the necessary extensions:
|
|||
|
0 - signed (SPR_SIGNED)
|
|||
|
1 - normal unsigned (SPR_UNSIGNED)
|
|||
|
2 - value is both sign and unsign extended for mode
|
|||
|
(SPR_SIGNED_AND_UNSIGNED).
|
|||
|
-1 - pointer unsigned, which most often can be handled like unsigned
|
|||
|
extension, except for generating instructions where we need to
|
|||
|
emit special code (ptr_extend insns) on some architectures
|
|||
|
(SPR_POINTER). */
|
|||
|
|
|||
|
const int SRP_POINTER = -1;
|
|||
|
const int SRP_SIGNED = 0;
|
|||
|
const int SRP_UNSIGNED = 1;
|
|||
|
const int SRP_SIGNED_AND_UNSIGNED = 2;
|
|||
|
|
|||
|
/* Sets promoted mode for SUBREG_PROMOTED_VAR_P(). */
|
|||
|
#define SUBREG_PROMOTED_SET(RTX, VAL) \
|
|||
|
do { \
|
|||
|
rtx const _rtx = RTL_FLAG_CHECK1 ("SUBREG_PROMOTED_SET", \
|
|||
|
(RTX), SUBREG); \
|
|||
|
switch (VAL) \
|
|||
|
{ \
|
|||
|
case SRP_POINTER: \
|
|||
|
_rtx->volatil = 0; \
|
|||
|
_rtx->unchanging = 0; \
|
|||
|
break; \
|
|||
|
case SRP_SIGNED: \
|
|||
|
_rtx->volatil = 0; \
|
|||
|
_rtx->unchanging = 1; \
|
|||
|
break; \
|
|||
|
case SRP_UNSIGNED: \
|
|||
|
_rtx->volatil = 1; \
|
|||
|
_rtx->unchanging = 0; \
|
|||
|
break; \
|
|||
|
case SRP_SIGNED_AND_UNSIGNED: \
|
|||
|
_rtx->volatil = 1; \
|
|||
|
_rtx->unchanging = 1; \
|
|||
|
break; \
|
|||
|
} \
|
|||
|
} while (0)
|
|||
|
|
|||
|
/* Gets the value stored in promoted mode for SUBREG_PROMOTED_VAR_P(),
|
|||
|
including SRP_SIGNED_AND_UNSIGNED if promoted for
|
|||
|
both signed and unsigned. */
|
|||
|
#define SUBREG_PROMOTED_GET(RTX) \
|
|||
|
(2 * (RTL_FLAG_CHECK1 ("SUBREG_PROMOTED_GET", (RTX), SUBREG)->volatil)\
|
|||
|
+ (RTX)->unchanging - 1)
|
|||
|
|
|||
|
/* Returns sign of promoted mode for SUBREG_PROMOTED_VAR_P(). */
|
|||
|
#define SUBREG_PROMOTED_SIGN(RTX) \
|
|||
|
((RTL_FLAG_CHECK1 ("SUBREG_PROMOTED_SIGN", (RTX), SUBREG)->volatil) ? 1\
|
|||
|
: (RTX)->unchanging - 1)
|
|||
|
|
|||
|
/* Predicate to check if RTX of SUBREG_PROMOTED_VAR_P() is promoted
|
|||
|
for SIGNED type. */
|
|||
|
#define SUBREG_PROMOTED_SIGNED_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("SUBREG_PROMOTED_SIGNED_P", (RTX), SUBREG)->unchanging)
|
|||
|
|
|||
|
/* Predicate to check if RTX of SUBREG_PROMOTED_VAR_P() is promoted
|
|||
|
for UNSIGNED type. */
|
|||
|
#define SUBREG_PROMOTED_UNSIGNED_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("SUBREG_PROMOTED_UNSIGNED_P", (RTX), SUBREG)->volatil)
|
|||
|
|
|||
|
/* Checks if RTX of SUBREG_PROMOTED_VAR_P() is promoted for given SIGN. */
|
|||
|
#define SUBREG_CHECK_PROMOTED_SIGN(RTX, SIGN) \
|
|||
|
((SIGN) == SRP_POINTER ? SUBREG_PROMOTED_GET (RTX) == SRP_POINTER \
|
|||
|
: (SIGN) == SRP_SIGNED ? SUBREG_PROMOTED_SIGNED_P (RTX) \
|
|||
|
: SUBREG_PROMOTED_UNSIGNED_P (RTX))
|
|||
|
|
|||
|
/* True if the REG is the static chain register for some CALL_INSN. */
|
|||
|
#define STATIC_CHAIN_REG_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("STATIC_CHAIN_REG_P", (RTX), REG)->jump)
|
|||
|
|
|||
|
/* True if the subreg was generated by LRA for reload insns. Such
|
|||
|
subregs are valid only during LRA. */
|
|||
|
#define LRA_SUBREG_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("LRA_SUBREG_P", (RTX), SUBREG)->jump)
|
|||
|
|
|||
|
/* Access various components of an ASM_OPERANDS rtx. */
|
|||
|
|
|||
|
#define ASM_OPERANDS_TEMPLATE(RTX) XCSTR (RTX, 0, ASM_OPERANDS)
|
|||
|
#define ASM_OPERANDS_OUTPUT_CONSTRAINT(RTX) XCSTR (RTX, 1, ASM_OPERANDS)
|
|||
|
#define ASM_OPERANDS_OUTPUT_IDX(RTX) XCINT (RTX, 2, ASM_OPERANDS)
|
|||
|
#define ASM_OPERANDS_INPUT_VEC(RTX) XCVEC (RTX, 3, ASM_OPERANDS)
|
|||
|
#define ASM_OPERANDS_INPUT_CONSTRAINT_VEC(RTX) XCVEC (RTX, 4, ASM_OPERANDS)
|
|||
|
#define ASM_OPERANDS_INPUT(RTX, N) XCVECEXP (RTX, 3, N, ASM_OPERANDS)
|
|||
|
#define ASM_OPERANDS_INPUT_LENGTH(RTX) XCVECLEN (RTX, 3, ASM_OPERANDS)
|
|||
|
#define ASM_OPERANDS_INPUT_CONSTRAINT_EXP(RTX, N) \
|
|||
|
XCVECEXP (RTX, 4, N, ASM_OPERANDS)
|
|||
|
#define ASM_OPERANDS_INPUT_CONSTRAINT(RTX, N) \
|
|||
|
XSTR (XCVECEXP (RTX, 4, N, ASM_OPERANDS), 0)
|
|||
|
#define ASM_OPERANDS_INPUT_MODE(RTX, N) \
|
|||
|
GET_MODE (XCVECEXP (RTX, 4, N, ASM_OPERANDS))
|
|||
|
#define ASM_OPERANDS_LABEL_VEC(RTX) XCVEC (RTX, 5, ASM_OPERANDS)
|
|||
|
#define ASM_OPERANDS_LABEL_LENGTH(RTX) XCVECLEN (RTX, 5, ASM_OPERANDS)
|
|||
|
#define ASM_OPERANDS_LABEL(RTX, N) XCVECEXP (RTX, 5, N, ASM_OPERANDS)
|
|||
|
#define ASM_OPERANDS_SOURCE_LOCATION(RTX) XCUINT (RTX, 6, ASM_OPERANDS)
|
|||
|
#define ASM_INPUT_SOURCE_LOCATION(RTX) XCUINT (RTX, 1, ASM_INPUT)
|
|||
|
|
|||
|
/* 1 if RTX is a mem that is statically allocated in read-only memory. */
|
|||
|
#define MEM_READONLY_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("MEM_READONLY_P", (RTX), MEM)->unchanging)
|
|||
|
|
|||
|
/* 1 if RTX is a mem and we should keep the alias set for this mem
|
|||
|
unchanged when we access a component. Set to 1, or example, when we
|
|||
|
are already in a non-addressable component of an aggregate. */
|
|||
|
#define MEM_KEEP_ALIAS_SET_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("MEM_KEEP_ALIAS_SET_P", (RTX), MEM)->jump)
|
|||
|
|
|||
|
/* 1 if RTX is a mem or asm_operand for a volatile reference. */
|
|||
|
#define MEM_VOLATILE_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK3 ("MEM_VOLATILE_P", (RTX), MEM, ASM_OPERANDS, \
|
|||
|
ASM_INPUT)->volatil)
|
|||
|
|
|||
|
/* 1 if RTX is a mem that cannot trap. */
|
|||
|
#define MEM_NOTRAP_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("MEM_NOTRAP_P", (RTX), MEM)->call)
|
|||
|
|
|||
|
/* The memory attribute block. We provide access macros for each value
|
|||
|
in the block and provide defaults if none specified. */
|
|||
|
#define MEM_ATTRS(RTX) X0MEMATTR (RTX, 1)
|
|||
|
|
|||
|
/* The register attribute block. We provide access macros for each value
|
|||
|
in the block and provide defaults if none specified. */
|
|||
|
#define REG_ATTRS(RTX) (REG_CHECK (RTX)->attrs)
|
|||
|
|
|||
|
#ifndef GENERATOR_FILE
|
|||
|
/* For a MEM rtx, the alias set. If 0, this MEM is not in any alias
|
|||
|
set, and may alias anything. Otherwise, the MEM can only alias
|
|||
|
MEMs in a conflicting alias set. This value is set in a
|
|||
|
language-dependent manner in the front-end, and should not be
|
|||
|
altered in the back-end. These set numbers are tested with
|
|||
|
alias_sets_conflict_p. */
|
|||
|
#define MEM_ALIAS_SET(RTX) (get_mem_attrs (RTX)->alias)
|
|||
|
|
|||
|
/* For a MEM rtx, the decl it is known to refer to, if it is known to
|
|||
|
refer to part of a DECL. It may also be a COMPONENT_REF. */
|
|||
|
#define MEM_EXPR(RTX) (get_mem_attrs (RTX)->expr)
|
|||
|
|
|||
|
/* For a MEM rtx, true if its MEM_OFFSET is known. */
|
|||
|
#define MEM_OFFSET_KNOWN_P(RTX) (get_mem_attrs (RTX)->offset_known_p)
|
|||
|
|
|||
|
/* For a MEM rtx, the offset from the start of MEM_EXPR. */
|
|||
|
#define MEM_OFFSET(RTX) (get_mem_attrs (RTX)->offset)
|
|||
|
|
|||
|
/* For a MEM rtx, the address space. */
|
|||
|
#define MEM_ADDR_SPACE(RTX) (get_mem_attrs (RTX)->addrspace)
|
|||
|
|
|||
|
/* For a MEM rtx, true if its MEM_SIZE is known. */
|
|||
|
#define MEM_SIZE_KNOWN_P(RTX) (get_mem_attrs (RTX)->size_known_p)
|
|||
|
|
|||
|
/* For a MEM rtx, the size in bytes of the MEM. */
|
|||
|
#define MEM_SIZE(RTX) (get_mem_attrs (RTX)->size)
|
|||
|
|
|||
|
/* For a MEM rtx, the alignment in bits. We can use the alignment of the
|
|||
|
mode as a default when STRICT_ALIGNMENT, but not if not. */
|
|||
|
#define MEM_ALIGN(RTX) (get_mem_attrs (RTX)->align)
|
|||
|
#else
|
|||
|
#define MEM_ADDR_SPACE(RTX) ADDR_SPACE_GENERIC
|
|||
|
#endif
|
|||
|
|
|||
|
/* For a REG rtx, the decl it is known to refer to, if it is known to
|
|||
|
refer to part of a DECL. */
|
|||
|
#define REG_EXPR(RTX) (REG_ATTRS (RTX) == 0 ? 0 : REG_ATTRS (RTX)->decl)
|
|||
|
|
|||
|
/* For a REG rtx, the offset from the start of REG_EXPR, if known, as an
|
|||
|
HOST_WIDE_INT. */
|
|||
|
#define REG_OFFSET(RTX) (REG_ATTRS (RTX) == 0 ? 0 : REG_ATTRS (RTX)->offset)
|
|||
|
|
|||
|
/* Copy the attributes that apply to memory locations from RHS to LHS. */
|
|||
|
#define MEM_COPY_ATTRIBUTES(LHS, RHS) \
|
|||
|
(MEM_VOLATILE_P (LHS) = MEM_VOLATILE_P (RHS), \
|
|||
|
MEM_NOTRAP_P (LHS) = MEM_NOTRAP_P (RHS), \
|
|||
|
MEM_READONLY_P (LHS) = MEM_READONLY_P (RHS), \
|
|||
|
MEM_KEEP_ALIAS_SET_P (LHS) = MEM_KEEP_ALIAS_SET_P (RHS), \
|
|||
|
MEM_POINTER (LHS) = MEM_POINTER (RHS), \
|
|||
|
MEM_ATTRS (LHS) = MEM_ATTRS (RHS))
|
|||
|
|
|||
|
/* 1 if RTX is a label_ref for a nonlocal label. */
|
|||
|
/* Likewise in an expr_list for a REG_LABEL_OPERAND or
|
|||
|
REG_LABEL_TARGET note. */
|
|||
|
#define LABEL_REF_NONLOCAL_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("LABEL_REF_NONLOCAL_P", (RTX), LABEL_REF)->volatil)
|
|||
|
|
|||
|
/* 1 if RTX is a code_label that should always be considered to be needed. */
|
|||
|
#define LABEL_PRESERVE_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK2 ("LABEL_PRESERVE_P", (RTX), CODE_LABEL, NOTE)->in_struct)
|
|||
|
|
|||
|
/* During sched, 1 if RTX is an insn that must be scheduled together
|
|||
|
with the preceding insn. */
|
|||
|
#define SCHED_GROUP_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK4 ("SCHED_GROUP_P", (RTX), DEBUG_INSN, INSN, \
|
|||
|
JUMP_INSN, CALL_INSN)->in_struct)
|
|||
|
|
|||
|
/* For a SET rtx, SET_DEST is the place that is set
|
|||
|
and SET_SRC is the value it is set to. */
|
|||
|
#define SET_DEST(RTX) XC2EXP (RTX, 0, SET, CLOBBER)
|
|||
|
#define SET_SRC(RTX) XCEXP (RTX, 1, SET)
|
|||
|
#define SET_IS_RETURN_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("SET_IS_RETURN_P", (RTX), SET)->jump)
|
|||
|
|
|||
|
/* For a TRAP_IF rtx, TRAP_CONDITION is an expression. */
|
|||
|
#define TRAP_CONDITION(RTX) XCEXP (RTX, 0, TRAP_IF)
|
|||
|
#define TRAP_CODE(RTX) XCEXP (RTX, 1, TRAP_IF)
|
|||
|
|
|||
|
/* For a COND_EXEC rtx, COND_EXEC_TEST is the condition to base
|
|||
|
conditionally executing the code on, COND_EXEC_CODE is the code
|
|||
|
to execute if the condition is true. */
|
|||
|
#define COND_EXEC_TEST(RTX) XCEXP (RTX, 0, COND_EXEC)
|
|||
|
#define COND_EXEC_CODE(RTX) XCEXP (RTX, 1, COND_EXEC)
|
|||
|
|
|||
|
/* 1 if RTX is a symbol_ref that addresses this function's rtl
|
|||
|
constants pool. */
|
|||
|
#define CONSTANT_POOL_ADDRESS_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("CONSTANT_POOL_ADDRESS_P", (RTX), SYMBOL_REF)->unchanging)
|
|||
|
|
|||
|
/* 1 if RTX is a symbol_ref that addresses a value in the file's
|
|||
|
tree constant pool. This information is private to varasm.cc. */
|
|||
|
#define TREE_CONSTANT_POOL_ADDRESS_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("TREE_CONSTANT_POOL_ADDRESS_P", \
|
|||
|
(RTX), SYMBOL_REF)->frame_related)
|
|||
|
|
|||
|
/* Used if RTX is a symbol_ref, for machine-specific purposes. */
|
|||
|
#define SYMBOL_REF_FLAG(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("SYMBOL_REF_FLAG", (RTX), SYMBOL_REF)->volatil)
|
|||
|
|
|||
|
/* 1 if RTX is a symbol_ref that has been the library function in
|
|||
|
emit_library_call. */
|
|||
|
#define SYMBOL_REF_USED(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("SYMBOL_REF_USED", (RTX), SYMBOL_REF)->used)
|
|||
|
|
|||
|
/* 1 if RTX is a symbol_ref for a weak symbol. */
|
|||
|
#define SYMBOL_REF_WEAK(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("SYMBOL_REF_WEAK", (RTX), SYMBOL_REF)->return_val)
|
|||
|
|
|||
|
/* A pointer attached to the SYMBOL_REF; either SYMBOL_REF_DECL or
|
|||
|
SYMBOL_REF_CONSTANT. */
|
|||
|
#define SYMBOL_REF_DATA(RTX) X0ANY ((RTX), 1)
|
|||
|
|
|||
|
/* Set RTX's SYMBOL_REF_DECL to DECL. RTX must not be a constant
|
|||
|
pool symbol. */
|
|||
|
#define SET_SYMBOL_REF_DECL(RTX, DECL) \
|
|||
|
(gcc_assert (!CONSTANT_POOL_ADDRESS_P (RTX)), X0TREE ((RTX), 1) = (DECL))
|
|||
|
|
|||
|
/* The tree (decl or constant) associated with the symbol, or null. */
|
|||
|
#define SYMBOL_REF_DECL(RTX) \
|
|||
|
(CONSTANT_POOL_ADDRESS_P (RTX) ? NULL : X0TREE ((RTX), 1))
|
|||
|
|
|||
|
/* Set RTX's SYMBOL_REF_CONSTANT to C. RTX must be a constant pool symbol. */
|
|||
|
#define SET_SYMBOL_REF_CONSTANT(RTX, C) \
|
|||
|
(gcc_assert (CONSTANT_POOL_ADDRESS_P (RTX)), X0CONSTANT ((RTX), 1) = (C))
|
|||
|
|
|||
|
/* The rtx constant pool entry for a symbol, or null. */
|
|||
|
#define SYMBOL_REF_CONSTANT(RTX) \
|
|||
|
(CONSTANT_POOL_ADDRESS_P (RTX) ? X0CONSTANT ((RTX), 1) : NULL)
|
|||
|
|
|||
|
/* A set of flags on a symbol_ref that are, in some respects, redundant with
|
|||
|
information derivable from the tree decl associated with this symbol.
|
|||
|
Except that we build a *lot* of SYMBOL_REFs that aren't associated with a
|
|||
|
decl. In some cases this is a bug. But beyond that, it's nice to cache
|
|||
|
this information to avoid recomputing it. Finally, this allows space for
|
|||
|
the target to store more than one bit of information, as with
|
|||
|
SYMBOL_REF_FLAG. */
|
|||
|
#define SYMBOL_REF_FLAGS(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("SYMBOL_REF_FLAGS", (RTX), SYMBOL_REF) \
|
|||
|
->u2.symbol_ref_flags)
|
|||
|
|
|||
|
/* These flags are common enough to be defined for all targets. They
|
|||
|
are computed by the default version of targetm.encode_section_info. */
|
|||
|
|
|||
|
/* Set if this symbol is a function. */
|
|||
|
#define SYMBOL_FLAG_FUNCTION (1 << 0)
|
|||
|
#define SYMBOL_REF_FUNCTION_P(RTX) \
|
|||
|
((SYMBOL_REF_FLAGS (RTX) & SYMBOL_FLAG_FUNCTION) != 0)
|
|||
|
/* Set if targetm.binds_local_p is true. */
|
|||
|
#define SYMBOL_FLAG_LOCAL (1 << 1)
|
|||
|
#define SYMBOL_REF_LOCAL_P(RTX) \
|
|||
|
((SYMBOL_REF_FLAGS (RTX) & SYMBOL_FLAG_LOCAL) != 0)
|
|||
|
/* Set if targetm.in_small_data_p is true. */
|
|||
|
#define SYMBOL_FLAG_SMALL (1 << 2)
|
|||
|
#define SYMBOL_REF_SMALL_P(RTX) \
|
|||
|
((SYMBOL_REF_FLAGS (RTX) & SYMBOL_FLAG_SMALL) != 0)
|
|||
|
/* The three-bit field at [5:3] is true for TLS variables; use
|
|||
|
SYMBOL_REF_TLS_MODEL to extract the field as an enum tls_model. */
|
|||
|
#define SYMBOL_FLAG_TLS_SHIFT 3
|
|||
|
#define SYMBOL_REF_TLS_MODEL(RTX) \
|
|||
|
((enum tls_model) ((SYMBOL_REF_FLAGS (RTX) >> SYMBOL_FLAG_TLS_SHIFT) & 7))
|
|||
|
/* Set if this symbol is not defined in this translation unit. */
|
|||
|
#define SYMBOL_FLAG_EXTERNAL (1 << 6)
|
|||
|
#define SYMBOL_REF_EXTERNAL_P(RTX) \
|
|||
|
((SYMBOL_REF_FLAGS (RTX) & SYMBOL_FLAG_EXTERNAL) != 0)
|
|||
|
/* Set if this symbol has a block_symbol structure associated with it. */
|
|||
|
#define SYMBOL_FLAG_HAS_BLOCK_INFO (1 << 7)
|
|||
|
#define SYMBOL_REF_HAS_BLOCK_INFO_P(RTX) \
|
|||
|
((SYMBOL_REF_FLAGS (RTX) & SYMBOL_FLAG_HAS_BLOCK_INFO) != 0)
|
|||
|
/* Set if this symbol is a section anchor. SYMBOL_REF_ANCHOR_P implies
|
|||
|
SYMBOL_REF_HAS_BLOCK_INFO_P. */
|
|||
|
#define SYMBOL_FLAG_ANCHOR (1 << 8)
|
|||
|
#define SYMBOL_REF_ANCHOR_P(RTX) \
|
|||
|
((SYMBOL_REF_FLAGS (RTX) & SYMBOL_FLAG_ANCHOR) != 0)
|
|||
|
|
|||
|
/* Subsequent bits are available for the target to use. */
|
|||
|
#define SYMBOL_FLAG_MACH_DEP_SHIFT 9
|
|||
|
#define SYMBOL_FLAG_MACH_DEP (1 << SYMBOL_FLAG_MACH_DEP_SHIFT)
|
|||
|
|
|||
|
/* If SYMBOL_REF_HAS_BLOCK_INFO_P (RTX), this is the object_block
|
|||
|
structure to which the symbol belongs, or NULL if it has not been
|
|||
|
assigned a block. */
|
|||
|
#define SYMBOL_REF_BLOCK(RTX) (BLOCK_SYMBOL_CHECK (RTX)->block)
|
|||
|
|
|||
|
/* If SYMBOL_REF_HAS_BLOCK_INFO_P (RTX), this is the offset of RTX from
|
|||
|
the first object in SYMBOL_REF_BLOCK (RTX). The value is negative if
|
|||
|
RTX has not yet been assigned to a block, or it has not been given an
|
|||
|
offset within that block. */
|
|||
|
#define SYMBOL_REF_BLOCK_OFFSET(RTX) (BLOCK_SYMBOL_CHECK (RTX)->offset)
|
|||
|
|
|||
|
/* True if RTX is flagged to be a scheduling barrier. */
|
|||
|
#define PREFETCH_SCHEDULE_BARRIER_P(RTX) \
|
|||
|
(RTL_FLAG_CHECK1 ("PREFETCH_SCHEDULE_BARRIER_P", (RTX), PREFETCH)->volatil)
|
|||
|
|
|||
|
/* Indicate whether the machine has any sort of auto increment addressing.
|
|||
|
If not, we can avoid checking for REG_INC notes. */
|
|||
|
|
|||
|
#if (defined (HAVE_PRE_INCREMENT) || defined (HAVE_PRE_DECREMENT) \
|
|||
|
|| defined (HAVE_POST_INCREMENT) || defined (HAVE_POST_DECREMENT) \
|
|||
|
|| defined (HAVE_PRE_MODIFY_DISP) || defined (HAVE_POST_MODIFY_DISP) \
|
|||
|
|| defined (HAVE_PRE_MODIFY_REG) || defined (HAVE_POST_MODIFY_REG))
|
|||
|
#define AUTO_INC_DEC 1
|
|||
|
#else
|
|||
|
#define AUTO_INC_DEC 0
|
|||
|
#endif
|
|||
|
|
|||
|
/* Define a macro to look for REG_INC notes,
|
|||
|
but save time on machines where they never exist. */
|
|||
|
|
|||
|
#if AUTO_INC_DEC
|
|||
|
#define FIND_REG_INC_NOTE(INSN, REG) \
|
|||
|
((REG) != NULL_RTX && REG_P ((REG)) \
|
|||
|
? find_regno_note ((INSN), REG_INC, REGNO (REG)) \
|
|||
|
: find_reg_note ((INSN), REG_INC, (REG)))
|
|||
|
#else
|
|||
|
#define FIND_REG_INC_NOTE(INSN, REG) 0
|
|||
|
#endif
|
|||
|
|
|||
|
#ifndef HAVE_PRE_INCREMENT
|
|||
|
#define HAVE_PRE_INCREMENT 0
|
|||
|
#endif
|
|||
|
|
|||
|
#ifndef HAVE_PRE_DECREMENT
|
|||
|
#define HAVE_PRE_DECREMENT 0
|
|||
|
#endif
|
|||
|
|
|||
|
#ifndef HAVE_POST_INCREMENT
|
|||
|
#define HAVE_POST_INCREMENT 0
|
|||
|
#endif
|
|||
|
|
|||
|
#ifndef HAVE_POST_DECREMENT
|
|||
|
#define HAVE_POST_DECREMENT 0
|
|||
|
#endif
|
|||
|
|
|||
|
#ifndef HAVE_POST_MODIFY_DISP
|
|||
|
#define HAVE_POST_MODIFY_DISP 0
|
|||
|
#endif
|
|||
|
|
|||
|
#ifndef HAVE_POST_MODIFY_REG
|
|||
|
#define HAVE_POST_MODIFY_REG 0
|
|||
|
#endif
|
|||
|
|
|||
|
#ifndef HAVE_PRE_MODIFY_DISP
|
|||
|
#define HAVE_PRE_MODIFY_DISP 0
|
|||
|
#endif
|
|||
|
|
|||
|
#ifndef HAVE_PRE_MODIFY_REG
|
|||
|
#define HAVE_PRE_MODIFY_REG 0
|
|||
|
#endif
|
|||
|
|
|||
|
|
|||
|
/* Some architectures do not have complete pre/post increment/decrement
|
|||
|
instruction sets, or only move some modes efficiently. These macros
|
|||
|
allow us to tune autoincrement generation. */
|
|||
|
|
|||
|
#ifndef USE_LOAD_POST_INCREMENT
|
|||
|
#define USE_LOAD_POST_INCREMENT(MODE) HAVE_POST_INCREMENT
|
|||
|
#endif
|
|||
|
|
|||
|
#ifndef USE_LOAD_POST_DECREMENT
|
|||
|
#define USE_LOAD_POST_DECREMENT(MODE) HAVE_POST_DECREMENT
|
|||
|
#endif
|
|||
|
|
|||
|
#ifndef USE_LOAD_PRE_INCREMENT
|
|||
|
#define USE_LOAD_PRE_INCREMENT(MODE) HAVE_PRE_INCREMENT
|
|||
|
#endif
|
|||
|
|
|||
|
#ifndef USE_LOAD_PRE_DECREMENT
|
|||
|
#define USE_LOAD_PRE_DECREMENT(MODE) HAVE_PRE_DECREMENT
|
|||
|
#endif
|
|||
|
|
|||
|
#ifndef USE_STORE_POST_INCREMENT
|
|||
|
#define USE_STORE_POST_INCREMENT(MODE) HAVE_POST_INCREMENT
|
|||
|
#endif
|
|||
|
|
|||
|
#ifndef USE_STORE_POST_DECREMENT
|
|||
|
#define USE_STORE_POST_DECREMENT(MODE) HAVE_POST_DECREMENT
|
|||
|
#endif
|
|||
|
|
|||
|
#ifndef USE_STORE_PRE_INCREMENT
|
|||
|
#define USE_STORE_PRE_INCREMENT(MODE) HAVE_PRE_INCREMENT
|
|||
|
#endif
|
|||
|
|
|||
|
#ifndef USE_STORE_PRE_DECREMENT
|
|||
|
#define USE_STORE_PRE_DECREMENT(MODE) HAVE_PRE_DECREMENT
|
|||
|
#endif
|
|||
|
|
|||
|
/* Nonzero when we are generating CONCATs. */
|
|||
|
extern int generating_concat_p;
|
|||
|
|
|||
|
/* Nonzero when we are expanding trees to RTL. */
|
|||
|
extern int currently_expanding_to_rtl;
|
|||
|
|
|||
|
/* Generally useful functions. */
|
|||
|
|
|||
|
#ifndef GENERATOR_FILE
|
|||
|
/* Return the cost of SET X. SPEED_P is true if optimizing for speed
|
|||
|
rather than size. */
|
|||
|
|
|||
|
inline int
|
|||
|
set_rtx_cost (rtx x, bool speed_p)
|
|||
|
{
|
|||
|
return rtx_cost (x, VOIDmode, INSN, 4, speed_p);
|
|||
|
}
|
|||
|
|
|||
|
/* Like set_rtx_cost, but return both the speed and size costs in C. */
|
|||
|
|
|||
|
inline void
|
|||
|
get_full_set_rtx_cost (rtx x, struct full_rtx_costs *c)
|
|||
|
{
|
|||
|
get_full_rtx_cost (x, VOIDmode, INSN, 4, c);
|
|||
|
}
|
|||
|
|
|||
|
/* Return the cost of moving X into a register, relative to the cost
|
|||
|
of a register move. SPEED_P is true if optimizing for speed rather
|
|||
|
than size. */
|
|||
|
|
|||
|
inline int
|
|||
|
set_src_cost (rtx x, machine_mode mode, bool speed_p)
|
|||
|
{
|
|||
|
return rtx_cost (x, mode, SET, 1, speed_p);
|
|||
|
}
|
|||
|
|
|||
|
/* Like set_src_cost, but return both the speed and size costs in C. */
|
|||
|
|
|||
|
inline void
|
|||
|
get_full_set_src_cost (rtx x, machine_mode mode, struct full_rtx_costs *c)
|
|||
|
{
|
|||
|
get_full_rtx_cost (x, mode, SET, 1, c);
|
|||
|
}
|
|||
|
#endif
|
|||
|
|
|||
|
/* A convenience macro to validate the arguments of a zero_extract
|
|||
|
expression. It determines whether SIZE lies inclusively within
|
|||
|
[1, RANGE], POS lies inclusively within between [0, RANGE - 1]
|
|||
|
and the sum lies inclusively within [1, RANGE]. RANGE must be
|
|||
|
>= 1, but SIZE and POS may be negative. */
|
|||
|
#define EXTRACT_ARGS_IN_RANGE(SIZE, POS, RANGE) \
|
|||
|
(IN_RANGE ((POS), 0, (unsigned HOST_WIDE_INT) (RANGE) - 1) \
|
|||
|
&& IN_RANGE ((SIZE), 1, (unsigned HOST_WIDE_INT) (RANGE) \
|
|||
|
- (unsigned HOST_WIDE_INT)(POS)))
|
|||
|
|
|||
|
/* In explow.cc */
|
|||
|
extern HOST_WIDE_INT trunc_int_for_mode (HOST_WIDE_INT, machine_mode);
|
|||
|
extern poly_int64 trunc_int_for_mode (poly_int64, machine_mode);
|
|||
|
extern rtx plus_constant (machine_mode, rtx, poly_int64, bool = false);
|
|||
|
extern HOST_WIDE_INT get_stack_check_protect (void);
|
|||
|
|
|||
|
/* In rtl.cc */
|
|||
|
extern rtx rtx_alloc (RTX_CODE CXX_MEM_STAT_INFO);
|
|||
|
inline rtx
|
|||
|
rtx_init (rtx rt, RTX_CODE code)
|
|||
|
{
|
|||
|
memset (rt, 0, RTX_HDR_SIZE);
|
|||
|
PUT_CODE (rt, code);
|
|||
|
return rt;
|
|||
|
}
|
|||
|
#define rtx_alloca(code) \
|
|||
|
rtx_init ((rtx) alloca (RTX_CODE_SIZE ((code))), (code))
|
|||
|
extern rtx rtx_alloc_stat_v (RTX_CODE MEM_STAT_DECL, int);
|
|||
|
#define rtx_alloc_v(c, SZ) rtx_alloc_stat_v (c MEM_STAT_INFO, SZ)
|
|||
|
#define const_wide_int_alloc(NWORDS) \
|
|||
|
rtx_alloc_v (CONST_WIDE_INT, \
|
|||
|
(sizeof (struct hwivec_def) \
|
|||
|
+ ((NWORDS)-1) * sizeof (HOST_WIDE_INT))) \
|
|||
|
|
|||
|
extern rtvec rtvec_alloc (size_t);
|
|||
|
extern rtvec shallow_copy_rtvec (rtvec);
|
|||
|
extern bool shared_const_p (const_rtx);
|
|||
|
extern rtx copy_rtx (rtx);
|
|||
|
extern enum rtx_code classify_insn (rtx);
|
|||
|
extern void dump_rtx_statistics (void);
|
|||
|
|
|||
|
/* In emit-rtl.cc */
|
|||
|
extern rtx copy_rtx_if_shared (rtx);
|
|||
|
|
|||
|
/* In rtl.cc */
|
|||
|
extern unsigned int rtx_size (const_rtx);
|
|||
|
extern rtx shallow_copy_rtx (const_rtx CXX_MEM_STAT_INFO);
|
|||
|
extern int rtx_equal_p (const_rtx, const_rtx);
|
|||
|
extern bool rtvec_all_equal_p (const_rtvec);
|
|||
|
extern bool rtvec_series_p (rtvec, int);
|
|||
|
|
|||
|
/* Return true if X is a vector constant with a duplicated element value. */
|
|||
|
|
|||
|
inline bool
|
|||
|
const_vec_duplicate_p (const_rtx x)
|
|||
|
{
|
|||
|
return (GET_CODE (x) == CONST_VECTOR
|
|||
|
&& CONST_VECTOR_NPATTERNS (x) == 1
|
|||
|
&& CONST_VECTOR_DUPLICATE_P (x));
|
|||
|
}
|
|||
|
|
|||
|
/* Return true if X is a vector constant with a duplicated element value.
|
|||
|
Store the duplicated element in *ELT if so. */
|
|||
|
|
|||
|
template <typename T>
|
|||
|
inline bool
|
|||
|
const_vec_duplicate_p (T x, T *elt)
|
|||
|
{
|
|||
|
if (const_vec_duplicate_p (x))
|
|||
|
{
|
|||
|
*elt = CONST_VECTOR_ENCODED_ELT (x, 0);
|
|||
|
return true;
|
|||
|
}
|
|||
|
return false;
|
|||
|
}
|
|||
|
|
|||
|
/* Return true if X is a vector with a duplicated element value, either
|
|||
|
constant or nonconstant. Store the duplicated element in *ELT if so. */
|
|||
|
|
|||
|
template <typename T>
|
|||
|
inline bool
|
|||
|
vec_duplicate_p (T x, T *elt)
|
|||
|
{
|
|||
|
if (GET_CODE (x) == VEC_DUPLICATE
|
|||
|
&& !VECTOR_MODE_P (GET_MODE (XEXP (x, 0))))
|
|||
|
{
|
|||
|
*elt = XEXP (x, 0);
|
|||
|
return true;
|
|||
|
}
|
|||
|
return const_vec_duplicate_p (x, elt);
|
|||
|
}
|
|||
|
|
|||
|
/* If X is a vector constant with a duplicated element value, return that
|
|||
|
element value, otherwise return X. */
|
|||
|
|
|||
|
template <typename T>
|
|||
|
inline T
|
|||
|
unwrap_const_vec_duplicate (T x)
|
|||
|
{
|
|||
|
if (const_vec_duplicate_p (x))
|
|||
|
x = CONST_VECTOR_ELT (x, 0);
|
|||
|
return x;
|
|||
|
}
|
|||
|
|
|||
|
/* In emit-rtl.cc. */
|
|||
|
extern wide_int const_vector_int_elt (const_rtx, unsigned int);
|
|||
|
extern rtx const_vector_elt (const_rtx, unsigned int);
|
|||
|
extern bool const_vec_series_p_1 (const_rtx, rtx *, rtx *);
|
|||
|
|
|||
|
/* Return true if X is an integer constant vector that contains a linear
|
|||
|
series of the form:
|
|||
|
|
|||
|
{ B, B + S, B + 2 * S, B + 3 * S, ... }
|
|||
|
|
|||
|
for a nonzero S. Store B and S in *BASE_OUT and *STEP_OUT on sucess. */
|
|||
|
|
|||
|
inline bool
|
|||
|
const_vec_series_p (const_rtx x, rtx *base_out, rtx *step_out)
|
|||
|
{
|
|||
|
if (GET_CODE (x) == CONST_VECTOR
|
|||
|
&& CONST_VECTOR_NPATTERNS (x) == 1
|
|||
|
&& !CONST_VECTOR_DUPLICATE_P (x))
|
|||
|
return const_vec_series_p_1 (x, base_out, step_out);
|
|||
|
return false;
|
|||
|
}
|
|||
|
|
|||
|
/* Return true if X is a vector that contains a linear series of the
|
|||
|
form:
|
|||
|
|
|||
|
{ B, B + S, B + 2 * S, B + 3 * S, ... }
|
|||
|
|
|||
|
where B and S are constant or nonconstant. Store B and S in
|
|||
|
*BASE_OUT and *STEP_OUT on sucess. */
|
|||
|
|
|||
|
inline bool
|
|||
|
vec_series_p (const_rtx x, rtx *base_out, rtx *step_out)
|
|||
|
{
|
|||
|
if (GET_CODE (x) == VEC_SERIES)
|
|||
|
{
|
|||
|
*base_out = XEXP (x, 0);
|
|||
|
*step_out = XEXP (x, 1);
|
|||
|
return true;
|
|||
|
}
|
|||
|
return const_vec_series_p (x, base_out, step_out);
|
|||
|
}
|
|||
|
|
|||
|
/* Return true if CONST_VECTORs X and Y, which are known to have the same mode,
|
|||
|
also have the same encoding. This means that they are equal whenever their
|
|||
|
operands are equal. */
|
|||
|
|
|||
|
inline bool
|
|||
|
same_vector_encodings_p (const_rtx x, const_rtx y)
|
|||
|
{
|
|||
|
/* Don't be fussy about the encoding of constant-length vectors,
|
|||
|
since XVECEXP (X, 0) and XVECEXP (Y, 0) list all the elements anyway. */
|
|||
|
if (poly_uint64 (CONST_VECTOR_NUNITS (x)).is_constant ())
|
|||
|
return true;
|
|||
|
|
|||
|
return (CONST_VECTOR_NPATTERNS (x) == CONST_VECTOR_NPATTERNS (y)
|
|||
|
&& (CONST_VECTOR_NELTS_PER_PATTERN (x)
|
|||
|
== CONST_VECTOR_NELTS_PER_PATTERN (y)));
|
|||
|
}
|
|||
|
|
|||
|
/* Return the unpromoted (outer) mode of SUBREG_PROMOTED_VAR_P subreg X. */
|
|||
|
|
|||
|
inline scalar_int_mode
|
|||
|
subreg_unpromoted_mode (rtx x)
|
|||
|
{
|
|||
|
gcc_checking_assert (SUBREG_PROMOTED_VAR_P (x));
|
|||
|
return as_a <scalar_int_mode> (GET_MODE (x));
|
|||
|
}
|
|||
|
|
|||
|
/* Return the promoted (inner) mode of SUBREG_PROMOTED_VAR_P subreg X. */
|
|||
|
|
|||
|
inline scalar_int_mode
|
|||
|
subreg_promoted_mode (rtx x)
|
|||
|
{
|
|||
|
gcc_checking_assert (SUBREG_PROMOTED_VAR_P (x));
|
|||
|
return as_a <scalar_int_mode> (GET_MODE (SUBREG_REG (x)));
|
|||
|
}
|
|||
|
|
|||
|
/* In emit-rtl.cc */
|
|||
|
extern rtvec gen_rtvec_v (int, rtx *);
|
|||
|
extern rtvec gen_rtvec_v (int, rtx_insn **);
|
|||
|
extern rtx gen_reg_rtx (machine_mode);
|
|||
|
extern rtx gen_rtx_REG_offset (rtx, machine_mode, unsigned int, poly_int64);
|
|||
|
extern rtx gen_reg_rtx_offset (rtx, machine_mode, int);
|
|||
|
extern rtx gen_reg_rtx_and_attrs (rtx);
|
|||
|
extern rtx_code_label *gen_label_rtx (void);
|
|||
|
extern rtx gen_lowpart_common (machine_mode, rtx);
|
|||
|
|
|||
|
/* In cse.cc */
|
|||
|
extern rtx gen_lowpart_if_possible (machine_mode, rtx);
|
|||
|
|
|||
|
/* In emit-rtl.cc */
|
|||
|
extern rtx gen_highpart (machine_mode, rtx);
|
|||
|
extern rtx gen_highpart_mode (machine_mode, machine_mode, rtx);
|
|||
|
extern rtx operand_subword (rtx, poly_uint64, int, machine_mode);
|
|||
|
|
|||
|
/* In emit-rtl.cc */
|
|||
|
extern rtx operand_subword_force (rtx, poly_uint64, machine_mode);
|
|||
|
extern int subreg_lowpart_p (const_rtx);
|
|||
|
extern poly_uint64 subreg_size_lowpart_offset (poly_uint64, poly_uint64);
|
|||
|
|
|||
|
/* Return true if a subreg of mode OUTERMODE would only access part of
|
|||
|
an inner register with mode INNERMODE. The other bits of the inner
|
|||
|
register would then be "don't care" on read. The behavior for writes
|
|||
|
depends on REGMODE_NATURAL_SIZE; bits in the same REGMODE_NATURAL_SIZE-d
|
|||
|
chunk would be clobbered but other bits would be preserved. */
|
|||
|
|
|||
|
inline bool
|
|||
|
partial_subreg_p (machine_mode outermode, machine_mode innermode)
|
|||
|
{
|
|||
|
/* Modes involved in a subreg must be ordered. In particular, we must
|
|||
|
always know at compile time whether the subreg is paradoxical. */
|
|||
|
poly_int64 outer_prec = GET_MODE_PRECISION (outermode);
|
|||
|
poly_int64 inner_prec = GET_MODE_PRECISION (innermode);
|
|||
|
gcc_checking_assert (ordered_p (outer_prec, inner_prec));
|
|||
|
return maybe_lt (outer_prec, inner_prec);
|
|||
|
}
|
|||
|
|
|||
|
/* Likewise return true if X is a subreg that is smaller than the inner
|
|||
|
register. Use read_modify_subreg_p to test whether writing to such
|
|||
|
a subreg preserves any part of the inner register. */
|
|||
|
|
|||
|
inline bool
|
|||
|
partial_subreg_p (const_rtx x)
|
|||
|
{
|
|||
|
if (GET_CODE (x) != SUBREG)
|
|||
|
return false;
|
|||
|
return partial_subreg_p (GET_MODE (x), GET_MODE (SUBREG_REG (x)));
|
|||
|
}
|
|||
|
|
|||
|
/* Return true if a subreg with the given outer and inner modes is
|
|||
|
paradoxical. */
|
|||
|
|
|||
|
inline bool
|
|||
|
paradoxical_subreg_p (machine_mode outermode, machine_mode innermode)
|
|||
|
{
|
|||
|
/* Modes involved in a subreg must be ordered. In particular, we must
|
|||
|
always know at compile time whether the subreg is paradoxical. */
|
|||
|
poly_int64 outer_prec = GET_MODE_PRECISION (outermode);
|
|||
|
poly_int64 inner_prec = GET_MODE_PRECISION (innermode);
|
|||
|
gcc_checking_assert (ordered_p (outer_prec, inner_prec));
|
|||
|
return maybe_gt (outer_prec, inner_prec);
|
|||
|
}
|
|||
|
|
|||
|
/* Return true if X is a paradoxical subreg, false otherwise. */
|
|||
|
|
|||
|
inline bool
|
|||
|
paradoxical_subreg_p (const_rtx x)
|
|||
|
{
|
|||
|
if (GET_CODE (x) != SUBREG)
|
|||
|
return false;
|
|||
|
return paradoxical_subreg_p (GET_MODE (x), GET_MODE (SUBREG_REG (x)));
|
|||
|
}
|
|||
|
|
|||
|
/* Return the SUBREG_BYTE for an OUTERMODE lowpart of an INNERMODE value. */
|
|||
|
|
|||
|
inline poly_uint64
|
|||
|
subreg_lowpart_offset (machine_mode outermode, machine_mode innermode)
|
|||
|
{
|
|||
|
return subreg_size_lowpart_offset (GET_MODE_SIZE (outermode),
|
|||
|
GET_MODE_SIZE (innermode));
|
|||
|
}
|
|||
|
|
|||
|
/* Given that a subreg has outer mode OUTERMODE and inner mode INNERMODE,
|
|||
|
return the smaller of the two modes if they are different sizes,
|
|||
|
otherwise return the outer mode. */
|
|||
|
|
|||
|
inline machine_mode
|
|||
|
narrower_subreg_mode (machine_mode outermode, machine_mode innermode)
|
|||
|
{
|
|||
|
return paradoxical_subreg_p (outermode, innermode) ? innermode : outermode;
|
|||
|
}
|
|||
|
|
|||
|
/* Given that a subreg has outer mode OUTERMODE and inner mode INNERMODE,
|
|||
|
return the mode that is big enough to hold both the outer and inner
|
|||
|
values. Prefer the outer mode in the event of a tie. */
|
|||
|
|
|||
|
inline machine_mode
|
|||
|
wider_subreg_mode (machine_mode outermode, machine_mode innermode)
|
|||
|
{
|
|||
|
return partial_subreg_p (outermode, innermode) ? innermode : outermode;
|
|||
|
}
|
|||
|
|
|||
|
/* Likewise for subreg X. */
|
|||
|
|
|||
|
inline machine_mode
|
|||
|
wider_subreg_mode (const_rtx x)
|
|||
|
{
|
|||
|
return wider_subreg_mode (GET_MODE (x), GET_MODE (SUBREG_REG (x)));
|
|||
|
}
|
|||
|
|
|||
|
extern poly_uint64 subreg_size_highpart_offset (poly_uint64, poly_uint64);
|
|||
|
|
|||
|
/* Return the SUBREG_BYTE for an OUTERMODE highpart of an INNERMODE value. */
|
|||
|
|
|||
|
inline poly_uint64
|
|||
|
subreg_highpart_offset (machine_mode outermode, machine_mode innermode)
|
|||
|
{
|
|||
|
return subreg_size_highpart_offset (GET_MODE_SIZE (outermode),
|
|||
|
GET_MODE_SIZE (innermode));
|
|||
|
}
|
|||
|
|
|||
|
extern poly_int64 byte_lowpart_offset (machine_mode, machine_mode);
|
|||
|
extern poly_int64 subreg_memory_offset (machine_mode, machine_mode,
|
|||
|
poly_uint64);
|
|||
|
extern poly_int64 subreg_memory_offset (const_rtx);
|
|||
|
extern rtx make_safe_from (rtx, rtx);
|
|||
|
extern rtx convert_memory_address_addr_space_1 (scalar_int_mode, rtx,
|
|||
|
addr_space_t, bool, bool);
|
|||
|
extern rtx convert_memory_address_addr_space (scalar_int_mode, rtx,
|
|||
|
addr_space_t);
|
|||
|
#define convert_memory_address(to_mode,x) \
|
|||
|
convert_memory_address_addr_space ((to_mode), (x), ADDR_SPACE_GENERIC)
|
|||
|
extern const char *get_insn_name (int);
|
|||
|
extern rtx_insn *get_last_insn_anywhere (void);
|
|||
|
extern rtx_insn *get_first_nonnote_insn (void);
|
|||
|
extern rtx_insn *get_last_nonnote_insn (void);
|
|||
|
extern void start_sequence (void);
|
|||
|
extern void push_to_sequence (rtx_insn *);
|
|||
|
extern void push_to_sequence2 (rtx_insn *, rtx_insn *);
|
|||
|
extern void end_sequence (void);
|
|||
|
#if TARGET_SUPPORTS_WIDE_INT == 0
|
|||
|
extern double_int rtx_to_double_int (const_rtx);
|
|||
|
#endif
|
|||
|
extern void cwi_output_hex (FILE *, const_rtx);
|
|||
|
#if TARGET_SUPPORTS_WIDE_INT == 0
|
|||
|
extern rtx immed_double_const (HOST_WIDE_INT, HOST_WIDE_INT,
|
|||
|
machine_mode);
|
|||
|
#endif
|
|||
|
extern rtx immed_wide_int_const (const poly_wide_int_ref &, machine_mode);
|
|||
|
|
|||
|
/* In varasm.cc */
|
|||
|
extern rtx force_const_mem (machine_mode, rtx);
|
|||
|
|
|||
|
/* In varasm.cc */
|
|||
|
|
|||
|
struct function;
|
|||
|
extern rtx get_pool_constant (const_rtx);
|
|||
|
extern rtx get_pool_constant_mark (rtx, bool *);
|
|||
|
extern fixed_size_mode get_pool_mode (const_rtx);
|
|||
|
extern rtx simplify_subtraction (rtx);
|
|||
|
extern void decide_function_section (tree);
|
|||
|
|
|||
|
/* In emit-rtl.cc */
|
|||
|
extern rtx_insn *emit_insn_before (rtx, rtx_insn *);
|
|||
|
extern rtx_insn *emit_insn_before_noloc (rtx, rtx_insn *, basic_block);
|
|||
|
extern rtx_insn *emit_insn_before_setloc (rtx, rtx_insn *, location_t);
|
|||
|
extern rtx_jump_insn *emit_jump_insn_before (rtx, rtx_insn *);
|
|||
|
extern rtx_jump_insn *emit_jump_insn_before_noloc (rtx, rtx_insn *);
|
|||
|
extern rtx_jump_insn *emit_jump_insn_before_setloc (rtx, rtx_insn *,
|
|||
|
location_t);
|
|||
|
extern rtx_insn *emit_call_insn_before (rtx, rtx_insn *);
|
|||
|
extern rtx_insn *emit_call_insn_before_noloc (rtx, rtx_insn *);
|
|||
|
extern rtx_insn *emit_call_insn_before_setloc (rtx, rtx_insn *, location_t);
|
|||
|
extern rtx_insn *emit_debug_insn_before (rtx, rtx_insn *);
|
|||
|
extern rtx_insn *emit_debug_insn_before_noloc (rtx, rtx_insn *);
|
|||
|
extern rtx_insn *emit_debug_insn_before_setloc (rtx, rtx_insn *, location_t);
|
|||
|
extern rtx_barrier *emit_barrier_before (rtx_insn *);
|
|||
|
extern rtx_code_label *emit_label_before (rtx_code_label *, rtx_insn *);
|
|||
|
extern rtx_note *emit_note_before (enum insn_note, rtx_insn *);
|
|||
|
extern rtx_insn *emit_insn_after (rtx, rtx_insn *);
|
|||
|
extern rtx_insn *emit_insn_after_noloc (rtx, rtx_insn *, basic_block);
|
|||
|
extern rtx_insn *emit_insn_after_setloc (rtx, rtx_insn *, location_t);
|
|||
|
extern rtx_jump_insn *emit_jump_insn_after (rtx, rtx_insn *);
|
|||
|
extern rtx_jump_insn *emit_jump_insn_after_noloc (rtx, rtx_insn *);
|
|||
|
extern rtx_jump_insn *emit_jump_insn_after_setloc (rtx, rtx_insn *, location_t);
|
|||
|
extern rtx_insn *emit_call_insn_after (rtx, rtx_insn *);
|
|||
|
extern rtx_insn *emit_call_insn_after_noloc (rtx, rtx_insn *);
|
|||
|
extern rtx_insn *emit_call_insn_after_setloc (rtx, rtx_insn *, location_t);
|
|||
|
extern rtx_insn *emit_debug_insn_after (rtx, rtx_insn *);
|
|||
|
extern rtx_insn *emit_debug_insn_after_noloc (rtx, rtx_insn *);
|
|||
|
extern rtx_insn *emit_debug_insn_after_setloc (rtx, rtx_insn *, location_t);
|
|||
|
extern rtx_barrier *emit_barrier_after (rtx_insn *);
|
|||
|
extern rtx_insn *emit_label_after (rtx_insn *, rtx_insn *);
|
|||
|
extern rtx_note *emit_note_after (enum insn_note, rtx_insn *);
|
|||
|
extern rtx_insn *emit_insn (rtx);
|
|||
|
extern rtx_insn *emit_debug_insn (rtx);
|
|||
|
extern rtx_insn *emit_jump_insn (rtx);
|
|||
|
extern rtx_insn *emit_call_insn (rtx);
|
|||
|
extern rtx_code_label *emit_label (rtx);
|
|||
|
extern rtx_jump_table_data *emit_jump_table_data (rtx);
|
|||
|
extern rtx_barrier *emit_barrier (void);
|
|||
|
extern rtx_note *emit_note (enum insn_note);
|
|||
|
extern rtx_note *emit_note_copy (rtx_note *);
|
|||
|
extern rtx_insn *gen_clobber (rtx);
|
|||
|
extern rtx_insn *emit_clobber (rtx);
|
|||
|
extern rtx_insn *gen_use (rtx);
|
|||
|
extern rtx_insn *emit_use (rtx);
|
|||
|
extern rtx_insn *make_insn_raw (rtx);
|
|||
|
extern void add_function_usage_to (rtx, rtx);
|
|||
|
extern rtx_call_insn *last_call_insn (void);
|
|||
|
extern rtx_insn *previous_insn (rtx_insn *);
|
|||
|
extern rtx_insn *next_insn (rtx_insn *);
|
|||
|
extern rtx_insn *prev_nonnote_insn (rtx_insn *);
|
|||
|
extern rtx_insn *next_nonnote_insn (rtx_insn *);
|
|||
|
extern rtx_insn *prev_nondebug_insn (rtx_insn *);
|
|||
|
extern rtx_insn *next_nondebug_insn (rtx_insn *);
|
|||
|
extern rtx_insn *prev_nonnote_nondebug_insn (rtx_insn *);
|
|||
|
extern rtx_insn *prev_nonnote_nondebug_insn_bb (rtx_insn *);
|
|||
|
extern rtx_insn *next_nonnote_nondebug_insn (rtx_insn *);
|
|||
|
extern rtx_insn *next_nonnote_nondebug_insn_bb (rtx_insn *);
|
|||
|
extern rtx_insn *prev_real_insn (rtx_insn *);
|
|||
|
extern rtx_insn *next_real_insn (rtx_insn *);
|
|||
|
extern rtx_insn *prev_real_nondebug_insn (rtx_insn *);
|
|||
|
extern rtx_insn *next_real_nondebug_insn (rtx);
|
|||
|
extern rtx_insn *prev_active_insn (rtx_insn *);
|
|||
|
extern rtx_insn *next_active_insn (rtx_insn *);
|
|||
|
extern int active_insn_p (const rtx_insn *);
|
|||
|
|
|||
|
/* In emit-rtl.cc */
|
|||
|
extern int insn_line (const rtx_insn *);
|
|||
|
extern const char * insn_file (const rtx_insn *);
|
|||
|
extern tree insn_scope (const rtx_insn *);
|
|||
|
extern expanded_location insn_location (const rtx_insn *);
|
|||
|
extern int insn_discriminator (const rtx_insn *);
|
|||
|
extern location_t prologue_location, epilogue_location;
|
|||
|
|
|||
|
/* In jump.cc */
|
|||
|
extern enum rtx_code reverse_condition (enum rtx_code);
|
|||
|
extern enum rtx_code reverse_condition_maybe_unordered (enum rtx_code);
|
|||
|
extern enum rtx_code swap_condition (enum rtx_code);
|
|||
|
extern enum rtx_code unsigned_condition (enum rtx_code);
|
|||
|
extern enum rtx_code signed_condition (enum rtx_code);
|
|||
|
extern void mark_jump_label (rtx, rtx_insn *, int);
|
|||
|
|
|||
|
/* Return true if integer comparison operator CODE interprets its operands
|
|||
|
as unsigned. */
|
|||
|
|
|||
|
inline bool
|
|||
|
unsigned_condition_p (enum rtx_code code)
|
|||
|
{
|
|||
|
return unsigned_condition (code) == code;
|
|||
|
}
|
|||
|
|
|||
|
/* In jump.cc */
|
|||
|
extern rtx_insn *delete_related_insns (rtx);
|
|||
|
|
|||
|
/* In recog.cc */
|
|||
|
extern rtx *find_constant_term_loc (rtx *);
|
|||
|
|
|||
|
/* In emit-rtl.cc */
|
|||
|
extern rtx_insn *try_split (rtx, rtx_insn *, int);
|
|||
|
|
|||
|
/* In insn-recog.cc (generated by genrecog). */
|
|||
|
extern rtx_insn *split_insns (rtx, rtx_insn *);
|
|||
|
|
|||
|
/* In simplify-rtx.cc */
|
|||
|
|
|||
|
/* A class that records the context in which a simplification
|
|||
|
is being mode. */
|
|||
|
class simplify_context
|
|||
|
{
|
|||
|
public:
|
|||
|
rtx simplify_unary_operation (rtx_code, machine_mode, rtx, machine_mode);
|
|||
|
rtx simplify_binary_operation (rtx_code, machine_mode, rtx, rtx);
|
|||
|
rtx simplify_ternary_operation (rtx_code, machine_mode, machine_mode,
|
|||
|
rtx, rtx, rtx);
|
|||
|
rtx simplify_relational_operation (rtx_code, machine_mode, machine_mode,
|
|||
|
rtx, rtx);
|
|||
|
rtx simplify_subreg (machine_mode, rtx, machine_mode, poly_uint64);
|
|||
|
|
|||
|
rtx lowpart_subreg (machine_mode, rtx, machine_mode);
|
|||
|
|
|||
|
rtx simplify_merge_mask (rtx, rtx, int);
|
|||
|
|
|||
|
rtx simplify_gen_unary (rtx_code, machine_mode, rtx, machine_mode);
|
|||
|
rtx simplify_gen_binary (rtx_code, machine_mode, rtx, rtx);
|
|||
|
rtx simplify_gen_ternary (rtx_code, machine_mode, machine_mode,
|
|||
|
rtx, rtx, rtx);
|
|||
|
rtx simplify_gen_relational (rtx_code, machine_mode, machine_mode, rtx, rtx);
|
|||
|
rtx simplify_gen_subreg (machine_mode, rtx, machine_mode, poly_uint64);
|
|||
|
rtx simplify_gen_vec_select (rtx, unsigned int);
|
|||
|
|
|||
|
/* Tracks the level of MEM nesting for the value being simplified:
|
|||
|
0 means the value is not in a MEM, >0 means it is. This is needed
|
|||
|
because the canonical representation of multiplication is different
|
|||
|
inside a MEM than outside. */
|
|||
|
unsigned int mem_depth = 0;
|
|||
|
|
|||
|
/* Tracks number of simplify_associative_operation calls performed during
|
|||
|
outermost simplify* call. */
|
|||
|
unsigned int assoc_count = 0;
|
|||
|
|
|||
|
/* Limit for the above number, return NULL from
|
|||
|
simplify_associative_operation after we reach that assoc_count. */
|
|||
|
static const unsigned int max_assoc_count = 64;
|
|||
|
|
|||
|
private:
|
|||
|
rtx simplify_truncation (machine_mode, rtx, machine_mode);
|
|||
|
rtx simplify_byte_swapping_operation (rtx_code, machine_mode, rtx, rtx);
|
|||
|
rtx simplify_associative_operation (rtx_code, machine_mode, rtx, rtx);
|
|||
|
rtx simplify_distributive_operation (rtx_code, machine_mode, rtx, rtx);
|
|||
|
rtx simplify_logical_relational_operation (rtx_code, machine_mode, rtx, rtx);
|
|||
|
rtx simplify_binary_operation_series (rtx_code, machine_mode, rtx, rtx);
|
|||
|
rtx simplify_distribute_over_subregs (rtx_code, machine_mode, rtx, rtx);
|
|||
|
rtx simplify_shift_const_int (rtx_code, machine_mode, rtx, unsigned int);
|
|||
|
rtx simplify_plus_minus (rtx_code, machine_mode, rtx, rtx);
|
|||
|
rtx simplify_cond_clz_ctz (rtx, rtx_code, rtx, rtx);
|
|||
|
|
|||
|
rtx simplify_unary_operation_1 (rtx_code, machine_mode, rtx);
|
|||
|
rtx simplify_binary_operation_1 (rtx_code, machine_mode, rtx, rtx, rtx, rtx);
|
|||
|
rtx simplify_ternary_operation_1 (rtx_code, machine_mode, machine_mode,
|
|||
|
rtx, rtx, rtx);
|
|||
|
rtx simplify_relational_operation_1 (rtx_code, machine_mode, machine_mode,
|
|||
|
rtx, rtx);
|
|||
|
};
|
|||
|
|
|||
|
inline rtx
|
|||
|
simplify_unary_operation (rtx_code code, machine_mode mode, rtx op,
|
|||
|
machine_mode op_mode)
|
|||
|
{
|
|||
|
return simplify_context ().simplify_unary_operation (code, mode, op,
|
|||
|
op_mode);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx
|
|||
|
simplify_binary_operation (rtx_code code, machine_mode mode, rtx op0, rtx op1)
|
|||
|
{
|
|||
|
return simplify_context ().simplify_binary_operation (code, mode, op0, op1);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx
|
|||
|
simplify_ternary_operation (rtx_code code, machine_mode mode,
|
|||
|
machine_mode op0_mode, rtx op0, rtx op1, rtx op2)
|
|||
|
{
|
|||
|
return simplify_context ().simplify_ternary_operation (code, mode, op0_mode,
|
|||
|
op0, op1, op2);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx
|
|||
|
simplify_relational_operation (rtx_code code, machine_mode mode,
|
|||
|
machine_mode op_mode, rtx op0, rtx op1)
|
|||
|
{
|
|||
|
return simplify_context ().simplify_relational_operation (code, mode,
|
|||
|
op_mode, op0, op1);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx
|
|||
|
simplify_subreg (machine_mode outermode, rtx op, machine_mode innermode,
|
|||
|
poly_uint64 byte)
|
|||
|
{
|
|||
|
return simplify_context ().simplify_subreg (outermode, op, innermode, byte);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx
|
|||
|
simplify_gen_unary (rtx_code code, machine_mode mode, rtx op,
|
|||
|
machine_mode op_mode)
|
|||
|
{
|
|||
|
return simplify_context ().simplify_gen_unary (code, mode, op, op_mode);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx
|
|||
|
simplify_gen_binary (rtx_code code, machine_mode mode, rtx op0, rtx op1)
|
|||
|
{
|
|||
|
return simplify_context ().simplify_gen_binary (code, mode, op0, op1);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx
|
|||
|
simplify_gen_ternary (rtx_code code, machine_mode mode, machine_mode op0_mode,
|
|||
|
rtx op0, rtx op1, rtx op2)
|
|||
|
{
|
|||
|
return simplify_context ().simplify_gen_ternary (code, mode, op0_mode,
|
|||
|
op0, op1, op2);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx
|
|||
|
simplify_gen_relational (rtx_code code, machine_mode mode,
|
|||
|
machine_mode op_mode, rtx op0, rtx op1)
|
|||
|
{
|
|||
|
return simplify_context ().simplify_gen_relational (code, mode, op_mode,
|
|||
|
op0, op1);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx
|
|||
|
simplify_gen_subreg (machine_mode outermode, rtx op, machine_mode innermode,
|
|||
|
poly_uint64 byte)
|
|||
|
{
|
|||
|
return simplify_context ().simplify_gen_subreg (outermode, op,
|
|||
|
innermode, byte);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx
|
|||
|
simplify_gen_vec_select (rtx op, unsigned int index)
|
|||
|
{
|
|||
|
return simplify_context ().simplify_gen_vec_select (op, index);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx
|
|||
|
lowpart_subreg (machine_mode outermode, rtx op, machine_mode innermode)
|
|||
|
{
|
|||
|
return simplify_context ().lowpart_subreg (outermode, op, innermode);
|
|||
|
}
|
|||
|
|
|||
|
extern rtx simplify_const_unary_operation (enum rtx_code, machine_mode,
|
|||
|
rtx, machine_mode);
|
|||
|
extern rtx simplify_const_binary_operation (enum rtx_code, machine_mode,
|
|||
|
rtx, rtx);
|
|||
|
extern rtx simplify_const_relational_operation (enum rtx_code,
|
|||
|
machine_mode, rtx, rtx);
|
|||
|
extern rtx simplify_replace_fn_rtx (rtx, const_rtx,
|
|||
|
rtx (*fn) (rtx, const_rtx, void *), void *);
|
|||
|
extern rtx simplify_replace_rtx (rtx, const_rtx, rtx);
|
|||
|
extern rtx simplify_rtx (const_rtx);
|
|||
|
extern rtx avoid_constant_pool_reference (rtx);
|
|||
|
extern rtx delegitimize_mem_from_attrs (rtx);
|
|||
|
extern bool mode_signbit_p (machine_mode, const_rtx);
|
|||
|
extern bool val_signbit_p (machine_mode, unsigned HOST_WIDE_INT);
|
|||
|
extern bool val_signbit_known_set_p (machine_mode,
|
|||
|
unsigned HOST_WIDE_INT);
|
|||
|
extern bool val_signbit_known_clear_p (machine_mode,
|
|||
|
unsigned HOST_WIDE_INT);
|
|||
|
|
|||
|
/* In reginfo.cc */
|
|||
|
extern machine_mode choose_hard_reg_mode (unsigned int, unsigned int,
|
|||
|
const predefined_function_abi *);
|
|||
|
extern const HARD_REG_SET &simplifiable_subregs (const subreg_shape &);
|
|||
|
|
|||
|
/* In emit-rtl.cc */
|
|||
|
extern rtx set_for_reg_notes (rtx);
|
|||
|
extern rtx set_unique_reg_note (rtx, enum reg_note, rtx);
|
|||
|
extern rtx set_dst_reg_note (rtx, enum reg_note, rtx, rtx);
|
|||
|
extern void set_insn_deleted (rtx_insn *);
|
|||
|
|
|||
|
/* Functions in rtlanal.cc */
|
|||
|
|
|||
|
extern rtx single_set_2 (const rtx_insn *, const_rtx);
|
|||
|
extern rtx simple_regno_set (rtx, unsigned int);
|
|||
|
extern bool contains_symbol_ref_p (const_rtx);
|
|||
|
extern bool contains_symbolic_reference_p (const_rtx);
|
|||
|
extern bool contains_constant_pool_address_p (const_rtx);
|
|||
|
extern void add_auto_inc_notes (rtx_insn *, rtx);
|
|||
|
|
|||
|
/* Handle the cheap and common cases inline for performance. */
|
|||
|
|
|||
|
inline rtx single_set (const rtx_insn *insn)
|
|||
|
{
|
|||
|
if (!INSN_P (insn))
|
|||
|
return NULL_RTX;
|
|||
|
|
|||
|
if (GET_CODE (PATTERN (insn)) == SET)
|
|||
|
return PATTERN (insn);
|
|||
|
|
|||
|
/* Defer to the more expensive case. */
|
|||
|
return single_set_2 (insn, PATTERN (insn));
|
|||
|
}
|
|||
|
|
|||
|
extern scalar_int_mode get_address_mode (rtx mem);
|
|||
|
extern int rtx_addr_can_trap_p (const_rtx);
|
|||
|
extern bool nonzero_address_p (const_rtx);
|
|||
|
extern int rtx_unstable_p (const_rtx);
|
|||
|
extern bool rtx_varies_p (const_rtx, bool);
|
|||
|
extern bool rtx_addr_varies_p (const_rtx, bool);
|
|||
|
extern rtx get_call_rtx_from (const rtx_insn *);
|
|||
|
extern tree get_call_fndecl (const rtx_insn *);
|
|||
|
extern HOST_WIDE_INT get_integer_term (const_rtx);
|
|||
|
extern rtx get_related_value (const_rtx);
|
|||
|
extern bool offset_within_block_p (const_rtx, HOST_WIDE_INT);
|
|||
|
extern void split_const (rtx, rtx *, rtx *);
|
|||
|
extern rtx strip_offset (rtx, poly_int64_pod *);
|
|||
|
extern poly_int64 get_args_size (const_rtx);
|
|||
|
extern bool unsigned_reg_p (rtx);
|
|||
|
extern int reg_mentioned_p (const_rtx, const_rtx);
|
|||
|
extern int count_occurrences (const_rtx, const_rtx, int);
|
|||
|
extern int reg_referenced_p (const_rtx, const_rtx);
|
|||
|
extern int reg_used_between_p (const_rtx, const rtx_insn *, const rtx_insn *);
|
|||
|
extern int reg_set_between_p (const_rtx, const rtx_insn *, const rtx_insn *);
|
|||
|
extern int commutative_operand_precedence (rtx);
|
|||
|
extern bool swap_commutative_operands_p (rtx, rtx);
|
|||
|
extern int modified_between_p (const_rtx, const rtx_insn *, const rtx_insn *);
|
|||
|
extern int no_labels_between_p (const rtx_insn *, const rtx_insn *);
|
|||
|
extern int modified_in_p (const_rtx, const_rtx);
|
|||
|
extern int reg_set_p (const_rtx, const_rtx);
|
|||
|
extern int multiple_sets (const_rtx);
|
|||
|
extern int set_noop_p (const_rtx);
|
|||
|
extern int noop_move_p (const rtx_insn *);
|
|||
|
extern bool refers_to_regno_p (unsigned int, unsigned int, const_rtx, rtx *);
|
|||
|
extern int reg_overlap_mentioned_p (const_rtx, const_rtx);
|
|||
|
extern const_rtx set_of (const_rtx, const_rtx);
|
|||
|
extern void record_hard_reg_sets (rtx, const_rtx, void *);
|
|||
|
extern void record_hard_reg_uses (rtx *, void *);
|
|||
|
extern void find_all_hard_regs (const_rtx, HARD_REG_SET *);
|
|||
|
extern void find_all_hard_reg_sets (const rtx_insn *, HARD_REG_SET *, bool);
|
|||
|
extern void note_pattern_stores (const_rtx,
|
|||
|
void (*) (rtx, const_rtx, void *), void *);
|
|||
|
extern void note_stores (const rtx_insn *,
|
|||
|
void (*) (rtx, const_rtx, void *), void *);
|
|||
|
extern void note_uses (rtx *, void (*) (rtx *, void *), void *);
|
|||
|
extern int dead_or_set_p (const rtx_insn *, const_rtx);
|
|||
|
extern int dead_or_set_regno_p (const rtx_insn *, unsigned int);
|
|||
|
extern rtx find_reg_note (const_rtx, enum reg_note, const_rtx);
|
|||
|
extern rtx find_regno_note (const_rtx, enum reg_note, unsigned int);
|
|||
|
extern rtx find_reg_equal_equiv_note (const_rtx);
|
|||
|
extern rtx find_constant_src (const rtx_insn *);
|
|||
|
extern int find_reg_fusage (const_rtx, enum rtx_code, const_rtx);
|
|||
|
extern int find_regno_fusage (const_rtx, enum rtx_code, unsigned int);
|
|||
|
extern rtx alloc_reg_note (enum reg_note, rtx, rtx);
|
|||
|
extern void add_reg_note (rtx, enum reg_note, rtx);
|
|||
|
extern void add_int_reg_note (rtx_insn *, enum reg_note, int);
|
|||
|
extern void add_args_size_note (rtx_insn *, poly_int64);
|
|||
|
extern void add_shallow_copy_of_reg_note (rtx_insn *, rtx);
|
|||
|
extern rtx duplicate_reg_note (rtx);
|
|||
|
extern void remove_note (rtx_insn *, const_rtx);
|
|||
|
extern bool remove_reg_equal_equiv_notes (rtx_insn *, bool = false);
|
|||
|
extern void remove_reg_equal_equiv_notes_for_regno (unsigned int);
|
|||
|
extern int side_effects_p (const_rtx);
|
|||
|
extern int volatile_refs_p (const_rtx);
|
|||
|
extern int volatile_insn_p (const_rtx);
|
|||
|
extern int may_trap_p_1 (const_rtx, unsigned);
|
|||
|
extern int may_trap_p (const_rtx);
|
|||
|
extern int may_trap_or_fault_p (const_rtx);
|
|||
|
extern bool can_throw_internal (const_rtx);
|
|||
|
extern bool can_throw_external (const_rtx);
|
|||
|
extern bool insn_could_throw_p (const_rtx);
|
|||
|
extern bool insn_nothrow_p (const_rtx);
|
|||
|
extern bool can_nonlocal_goto (const rtx_insn *);
|
|||
|
extern void copy_reg_eh_region_note_forward (rtx, rtx_insn *, rtx);
|
|||
|
extern void copy_reg_eh_region_note_backward (rtx, rtx_insn *, rtx);
|
|||
|
extern rtx replace_rtx (rtx, rtx, rtx, bool = false);
|
|||
|
extern void replace_label (rtx *, rtx, rtx, bool);
|
|||
|
extern void replace_label_in_insn (rtx_insn *, rtx_insn *, rtx_insn *, bool);
|
|||
|
extern bool rtx_referenced_p (const_rtx, const_rtx);
|
|||
|
extern bool tablejump_p (const rtx_insn *, rtx_insn **, rtx_jump_table_data **);
|
|||
|
extern rtx tablejump_casesi_pattern (const rtx_insn *insn);
|
|||
|
extern int computed_jump_p (const rtx_insn *);
|
|||
|
extern bool tls_referenced_p (const_rtx);
|
|||
|
extern bool contains_mem_rtx_p (rtx x);
|
|||
|
extern bool register_asm_p (const_rtx);
|
|||
|
|
|||
|
/* Overload for refers_to_regno_p for checking a single register. */
|
|||
|
inline bool
|
|||
|
refers_to_regno_p (unsigned int regnum, const_rtx x, rtx* loc = NULL)
|
|||
|
{
|
|||
|
return refers_to_regno_p (regnum, regnum + 1, x, loc);
|
|||
|
}
|
|||
|
|
|||
|
/* Callback for for_each_inc_dec, to process the autoinc operation OP
|
|||
|
within MEM that sets DEST to SRC + SRCOFF, or SRC if SRCOFF is
|
|||
|
NULL. The callback is passed the same opaque ARG passed to
|
|||
|
for_each_inc_dec. Return zero to continue looking for other
|
|||
|
autoinc operations or any other value to interrupt the traversal and
|
|||
|
return that value to the caller of for_each_inc_dec. */
|
|||
|
typedef int (*for_each_inc_dec_fn) (rtx mem, rtx op, rtx dest, rtx src,
|
|||
|
rtx srcoff, void *arg);
|
|||
|
extern int for_each_inc_dec (rtx, for_each_inc_dec_fn, void *arg);
|
|||
|
|
|||
|
typedef int (*rtx_equal_p_callback_function) (const_rtx *, const_rtx *,
|
|||
|
rtx *, rtx *);
|
|||
|
extern int rtx_equal_p_cb (const_rtx, const_rtx,
|
|||
|
rtx_equal_p_callback_function);
|
|||
|
|
|||
|
typedef int (*hash_rtx_callback_function) (const_rtx, machine_mode, rtx *,
|
|||
|
machine_mode *);
|
|||
|
extern unsigned hash_rtx_cb (const_rtx, machine_mode, int *, int *,
|
|||
|
bool, hash_rtx_callback_function);
|
|||
|
|
|||
|
extern rtx regno_use_in (unsigned int, rtx);
|
|||
|
extern int auto_inc_p (const_rtx);
|
|||
|
extern bool in_insn_list_p (const rtx_insn_list *, const rtx_insn *);
|
|||
|
extern void remove_node_from_insn_list (const rtx_insn *, rtx_insn_list **);
|
|||
|
extern int loc_mentioned_in_p (rtx *, const_rtx);
|
|||
|
extern rtx_insn *find_first_parameter_load (rtx_insn *, rtx_insn *);
|
|||
|
extern bool keep_with_call_p (const rtx_insn *);
|
|||
|
extern bool label_is_jump_target_p (const_rtx, const rtx_insn *);
|
|||
|
extern int pattern_cost (rtx, bool);
|
|||
|
extern int insn_cost (rtx_insn *, bool);
|
|||
|
extern unsigned seq_cost (const rtx_insn *, bool);
|
|||
|
|
|||
|
/* Given an insn and condition, return a canonical description of
|
|||
|
the test being made. */
|
|||
|
extern rtx canonicalize_condition (rtx_insn *, rtx, int, rtx_insn **, rtx,
|
|||
|
int, int);
|
|||
|
|
|||
|
/* Given a JUMP_INSN, return a canonical description of the test
|
|||
|
being made. */
|
|||
|
extern rtx get_condition (rtx_insn *, rtx_insn **, int, int);
|
|||
|
|
|||
|
/* Information about a subreg of a hard register. */
|
|||
|
struct subreg_info
|
|||
|
{
|
|||
|
/* Offset of first hard register involved in the subreg. */
|
|||
|
int offset;
|
|||
|
/* Number of hard registers involved in the subreg. In the case of
|
|||
|
a paradoxical subreg, this is the number of registers that would
|
|||
|
be modified by writing to the subreg; some of them may be don't-care
|
|||
|
when reading from the subreg. */
|
|||
|
int nregs;
|
|||
|
/* Whether this subreg can be represented as a hard reg with the new
|
|||
|
mode (by adding OFFSET to the original hard register). */
|
|||
|
bool representable_p;
|
|||
|
};
|
|||
|
|
|||
|
extern void subreg_get_info (unsigned int, machine_mode,
|
|||
|
poly_uint64, machine_mode,
|
|||
|
struct subreg_info *);
|
|||
|
|
|||
|
/* lists.cc */
|
|||
|
|
|||
|
extern void free_EXPR_LIST_list (rtx_expr_list **);
|
|||
|
extern void free_INSN_LIST_list (rtx_insn_list **);
|
|||
|
extern void free_EXPR_LIST_node (rtx);
|
|||
|
extern void free_INSN_LIST_node (rtx);
|
|||
|
extern rtx_insn_list *alloc_INSN_LIST (rtx, rtx);
|
|||
|
extern rtx_insn_list *copy_INSN_LIST (rtx_insn_list *);
|
|||
|
extern rtx_insn_list *concat_INSN_LIST (rtx_insn_list *, rtx_insn_list *);
|
|||
|
extern rtx_expr_list *alloc_EXPR_LIST (int, rtx, rtx);
|
|||
|
extern void remove_free_INSN_LIST_elem (rtx_insn *, rtx_insn_list **);
|
|||
|
extern rtx remove_list_elem (rtx, rtx *);
|
|||
|
extern rtx_insn *remove_free_INSN_LIST_node (rtx_insn_list **);
|
|||
|
extern rtx remove_free_EXPR_LIST_node (rtx_expr_list **);
|
|||
|
|
|||
|
|
|||
|
/* reginfo.cc */
|
|||
|
|
|||
|
/* Resize reg info. */
|
|||
|
extern bool resize_reg_info (void);
|
|||
|
/* Free up register info memory. */
|
|||
|
extern void free_reg_info (void);
|
|||
|
extern void init_subregs_of_mode (void);
|
|||
|
extern void finish_subregs_of_mode (void);
|
|||
|
extern void reginfo_cc_finalize (void);
|
|||
|
|
|||
|
/* recog.cc */
|
|||
|
extern rtx extract_asm_operands (rtx);
|
|||
|
extern int asm_noperands (const_rtx);
|
|||
|
extern const char *decode_asm_operands (rtx, rtx *, rtx **, const char **,
|
|||
|
machine_mode *, location_t *);
|
|||
|
extern void get_referenced_operands (const char *, bool *, unsigned int);
|
|||
|
|
|||
|
extern enum reg_class reg_preferred_class (int);
|
|||
|
extern enum reg_class reg_alternate_class (int);
|
|||
|
extern enum reg_class reg_allocno_class (int);
|
|||
|
extern void setup_reg_classes (int, enum reg_class, enum reg_class,
|
|||
|
enum reg_class);
|
|||
|
|
|||
|
extern void split_all_insns (void);
|
|||
|
extern unsigned int split_all_insns_noflow (void);
|
|||
|
|
|||
|
#define MAX_SAVED_CONST_INT 64
|
|||
|
extern GTY(()) rtx const_int_rtx[MAX_SAVED_CONST_INT * 2 + 1];
|
|||
|
|
|||
|
#define const0_rtx (const_int_rtx[MAX_SAVED_CONST_INT])
|
|||
|
#define const1_rtx (const_int_rtx[MAX_SAVED_CONST_INT+1])
|
|||
|
#define const2_rtx (const_int_rtx[MAX_SAVED_CONST_INT+2])
|
|||
|
#define constm1_rtx (const_int_rtx[MAX_SAVED_CONST_INT-1])
|
|||
|
extern GTY(()) rtx const_true_rtx;
|
|||
|
|
|||
|
extern GTY(()) rtx const_tiny_rtx[4][(int) MAX_MACHINE_MODE];
|
|||
|
|
|||
|
/* Returns a constant 0 rtx in mode MODE. Integer modes are treated the
|
|||
|
same as VOIDmode. */
|
|||
|
|
|||
|
#define CONST0_RTX(MODE) (const_tiny_rtx[0][(int) (MODE)])
|
|||
|
|
|||
|
/* Likewise, for the constants 1 and 2 and -1. */
|
|||
|
|
|||
|
#define CONST1_RTX(MODE) (const_tiny_rtx[1][(int) (MODE)])
|
|||
|
#define CONST2_RTX(MODE) (const_tiny_rtx[2][(int) (MODE)])
|
|||
|
#define CONSTM1_RTX(MODE) (const_tiny_rtx[3][(int) (MODE)])
|
|||
|
|
|||
|
extern GTY(()) rtx pc_rtx;
|
|||
|
extern GTY(()) rtx ret_rtx;
|
|||
|
extern GTY(()) rtx simple_return_rtx;
|
|||
|
extern GTY(()) rtx_insn *invalid_insn_rtx;
|
|||
|
|
|||
|
/* If HARD_FRAME_POINTER_REGNUM is defined, then a special dummy reg
|
|||
|
is used to represent the frame pointer. This is because the
|
|||
|
hard frame pointer and the automatic variables are separated by an amount
|
|||
|
that cannot be determined until after register allocation. We can assume
|
|||
|
that in this case ELIMINABLE_REGS will be defined, one action of which
|
|||
|
will be to eliminate FRAME_POINTER_REGNUM into HARD_FRAME_POINTER_REGNUM. */
|
|||
|
#ifndef HARD_FRAME_POINTER_REGNUM
|
|||
|
#define HARD_FRAME_POINTER_REGNUM FRAME_POINTER_REGNUM
|
|||
|
#endif
|
|||
|
|
|||
|
#ifndef HARD_FRAME_POINTER_IS_FRAME_POINTER
|
|||
|
#define HARD_FRAME_POINTER_IS_FRAME_POINTER \
|
|||
|
(HARD_FRAME_POINTER_REGNUM == FRAME_POINTER_REGNUM)
|
|||
|
#endif
|
|||
|
|
|||
|
#ifndef HARD_FRAME_POINTER_IS_ARG_POINTER
|
|||
|
#define HARD_FRAME_POINTER_IS_ARG_POINTER \
|
|||
|
(HARD_FRAME_POINTER_REGNUM == ARG_POINTER_REGNUM)
|
|||
|
#endif
|
|||
|
|
|||
|
/* Index labels for global_rtl. */
|
|||
|
enum global_rtl_index
|
|||
|
{
|
|||
|
GR_STACK_POINTER,
|
|||
|
GR_FRAME_POINTER,
|
|||
|
/* For register elimination to work properly these hard_frame_pointer_rtx,
|
|||
|
frame_pointer_rtx, and arg_pointer_rtx must be the same if they refer to
|
|||
|
the same register. */
|
|||
|
#if FRAME_POINTER_REGNUM == ARG_POINTER_REGNUM
|
|||
|
GR_ARG_POINTER = GR_FRAME_POINTER,
|
|||
|
#endif
|
|||
|
#if HARD_FRAME_POINTER_IS_FRAME_POINTER
|
|||
|
GR_HARD_FRAME_POINTER = GR_FRAME_POINTER,
|
|||
|
#else
|
|||
|
GR_HARD_FRAME_POINTER,
|
|||
|
#endif
|
|||
|
#if FRAME_POINTER_REGNUM != ARG_POINTER_REGNUM
|
|||
|
#if HARD_FRAME_POINTER_IS_ARG_POINTER
|
|||
|
GR_ARG_POINTER = GR_HARD_FRAME_POINTER,
|
|||
|
#else
|
|||
|
GR_ARG_POINTER,
|
|||
|
#endif
|
|||
|
#endif
|
|||
|
GR_VIRTUAL_INCOMING_ARGS,
|
|||
|
GR_VIRTUAL_STACK_ARGS,
|
|||
|
GR_VIRTUAL_STACK_DYNAMIC,
|
|||
|
GR_VIRTUAL_OUTGOING_ARGS,
|
|||
|
GR_VIRTUAL_CFA,
|
|||
|
GR_VIRTUAL_PREFERRED_STACK_BOUNDARY,
|
|||
|
|
|||
|
GR_MAX
|
|||
|
};
|
|||
|
|
|||
|
/* Target-dependent globals. */
|
|||
|
struct GTY(()) target_rtl {
|
|||
|
/* All references to the hard registers in global_rtl_index go through
|
|||
|
these unique rtl objects. On machines where the frame-pointer and
|
|||
|
arg-pointer are the same register, they use the same unique object.
|
|||
|
|
|||
|
After register allocation, other rtl objects which used to be pseudo-regs
|
|||
|
may be clobbered to refer to the frame-pointer register.
|
|||
|
But references that were originally to the frame-pointer can be
|
|||
|
distinguished from the others because they contain frame_pointer_rtx.
|
|||
|
|
|||
|
When to use frame_pointer_rtx and hard_frame_pointer_rtx is a little
|
|||
|
tricky: until register elimination has taken place hard_frame_pointer_rtx
|
|||
|
should be used if it is being set, and frame_pointer_rtx otherwise. After
|
|||
|
register elimination hard_frame_pointer_rtx should always be used.
|
|||
|
On machines where the two registers are same (most) then these are the
|
|||
|
same. */
|
|||
|
rtx x_global_rtl[GR_MAX];
|
|||
|
|
|||
|
/* A unique representation of (REG:Pmode PIC_OFFSET_TABLE_REGNUM). */
|
|||
|
rtx x_pic_offset_table_rtx;
|
|||
|
|
|||
|
/* A unique representation of (REG:Pmode RETURN_ADDRESS_POINTER_REGNUM).
|
|||
|
This is used to implement __builtin_return_address for some machines;
|
|||
|
see for instance the MIPS port. */
|
|||
|
rtx x_return_address_pointer_rtx;
|
|||
|
|
|||
|
/* Commonly used RTL for hard registers. These objects are not
|
|||
|
necessarily unique, so we allocate them separately from global_rtl.
|
|||
|
They are initialized once per compilation unit, then copied into
|
|||
|
regno_reg_rtx at the beginning of each function. */
|
|||
|
rtx x_initial_regno_reg_rtx[FIRST_PSEUDO_REGISTER];
|
|||
|
|
|||
|
/* A sample (mem:M stack_pointer_rtx) rtx for each mode M. */
|
|||
|
rtx x_top_of_stack[MAX_MACHINE_MODE];
|
|||
|
|
|||
|
/* Static hunks of RTL used by the aliasing code; these are treated
|
|||
|
as persistent to avoid unnecessary RTL allocations. */
|
|||
|
rtx x_static_reg_base_value[FIRST_PSEUDO_REGISTER];
|
|||
|
|
|||
|
/* The default memory attributes for each mode. */
|
|||
|
class mem_attrs *x_mode_mem_attrs[(int) MAX_MACHINE_MODE];
|
|||
|
|
|||
|
/* Track if RTL has been initialized. */
|
|||
|
bool target_specific_initialized;
|
|||
|
};
|
|||
|
|
|||
|
extern GTY(()) struct target_rtl default_target_rtl;
|
|||
|
#if SWITCHABLE_TARGET
|
|||
|
extern struct target_rtl *this_target_rtl;
|
|||
|
#else
|
|||
|
#define this_target_rtl (&default_target_rtl)
|
|||
|
#endif
|
|||
|
|
|||
|
#define global_rtl \
|
|||
|
(this_target_rtl->x_global_rtl)
|
|||
|
#define pic_offset_table_rtx \
|
|||
|
(this_target_rtl->x_pic_offset_table_rtx)
|
|||
|
#define return_address_pointer_rtx \
|
|||
|
(this_target_rtl->x_return_address_pointer_rtx)
|
|||
|
#define top_of_stack \
|
|||
|
(this_target_rtl->x_top_of_stack)
|
|||
|
#define mode_mem_attrs \
|
|||
|
(this_target_rtl->x_mode_mem_attrs)
|
|||
|
|
|||
|
/* All references to certain hard regs, except those created
|
|||
|
by allocating pseudo regs into them (when that's possible),
|
|||
|
go through these unique rtx objects. */
|
|||
|
#define stack_pointer_rtx (global_rtl[GR_STACK_POINTER])
|
|||
|
#define frame_pointer_rtx (global_rtl[GR_FRAME_POINTER])
|
|||
|
#define hard_frame_pointer_rtx (global_rtl[GR_HARD_FRAME_POINTER])
|
|||
|
#define arg_pointer_rtx (global_rtl[GR_ARG_POINTER])
|
|||
|
|
|||
|
#ifndef GENERATOR_FILE
|
|||
|
/* Return the attributes of a MEM rtx. */
|
|||
|
inline const class mem_attrs *
|
|||
|
get_mem_attrs (const_rtx x)
|
|||
|
{
|
|||
|
class mem_attrs *attrs;
|
|||
|
|
|||
|
attrs = MEM_ATTRS (x);
|
|||
|
if (!attrs)
|
|||
|
attrs = mode_mem_attrs[(int) GET_MODE (x)];
|
|||
|
return attrs;
|
|||
|
}
|
|||
|
#endif
|
|||
|
|
|||
|
/* Include the RTL generation functions. */
|
|||
|
|
|||
|
#ifndef GENERATOR_FILE
|
|||
|
#include "genrtl.h"
|
|||
|
#undef gen_rtx_ASM_INPUT
|
|||
|
#define gen_rtx_ASM_INPUT(MODE, ARG0) \
|
|||
|
gen_rtx_fmt_si (ASM_INPUT, (MODE), (ARG0), 0)
|
|||
|
#define gen_rtx_ASM_INPUT_loc(MODE, ARG0, LOC) \
|
|||
|
gen_rtx_fmt_si (ASM_INPUT, (MODE), (ARG0), (LOC))
|
|||
|
#endif
|
|||
|
|
|||
|
/* There are some RTL codes that require special attention; the
|
|||
|
generation functions included above do the raw handling. If you
|
|||
|
add to this list, modify special_rtx in gengenrtl.cc as well. */
|
|||
|
|
|||
|
extern rtx_expr_list *gen_rtx_EXPR_LIST (machine_mode, rtx, rtx);
|
|||
|
extern rtx_insn_list *gen_rtx_INSN_LIST (machine_mode, rtx, rtx);
|
|||
|
extern rtx_insn *
|
|||
|
gen_rtx_INSN (machine_mode mode, rtx_insn *prev_insn, rtx_insn *next_insn,
|
|||
|
basic_block bb, rtx pattern, int location, int code,
|
|||
|
rtx reg_notes);
|
|||
|
extern rtx gen_rtx_CONST_INT (machine_mode, HOST_WIDE_INT);
|
|||
|
extern rtx gen_rtx_CONST_VECTOR (machine_mode, rtvec);
|
|||
|
extern void set_mode_and_regno (rtx, machine_mode, unsigned int);
|
|||
|
extern rtx init_raw_REG (rtx, machine_mode, unsigned int);
|
|||
|
extern rtx gen_raw_REG (machine_mode, unsigned int);
|
|||
|
#define alloca_raw_REG(mode, regno) \
|
|||
|
init_raw_REG (rtx_alloca (REG), (mode), (regno))
|
|||
|
extern rtx gen_rtx_REG (machine_mode, unsigned int);
|
|||
|
extern rtx gen_rtx_SUBREG (machine_mode, rtx, poly_uint64);
|
|||
|
extern rtx gen_rtx_MEM (machine_mode, rtx);
|
|||
|
extern rtx gen_rtx_VAR_LOCATION (machine_mode, tree, rtx,
|
|||
|
enum var_init_status);
|
|||
|
|
|||
|
#ifdef GENERATOR_FILE
|
|||
|
#define PUT_MODE(RTX, MODE) PUT_MODE_RAW (RTX, MODE)
|
|||
|
#else
|
|||
|
inline void
|
|||
|
PUT_MODE (rtx x, machine_mode mode)
|
|||
|
{
|
|||
|
if (REG_P (x))
|
|||
|
set_mode_and_regno (x, mode, REGNO (x));
|
|||
|
else
|
|||
|
PUT_MODE_RAW (x, mode);
|
|||
|
}
|
|||
|
#endif
|
|||
|
|
|||
|
#define GEN_INT(N) gen_rtx_CONST_INT (VOIDmode, (N))
|
|||
|
|
|||
|
/* Virtual registers are used during RTL generation to refer to locations into
|
|||
|
the stack frame when the actual location isn't known until RTL generation
|
|||
|
is complete. The routine instantiate_virtual_regs replaces these with
|
|||
|
the proper value, which is normally {frame,arg,stack}_pointer_rtx plus
|
|||
|
a constant. */
|
|||
|
|
|||
|
#define FIRST_VIRTUAL_REGISTER (FIRST_PSEUDO_REGISTER)
|
|||
|
|
|||
|
/* This points to the first word of the incoming arguments passed on the stack,
|
|||
|
either by the caller or by the callee when pretending it was passed by the
|
|||
|
caller. */
|
|||
|
|
|||
|
#define virtual_incoming_args_rtx (global_rtl[GR_VIRTUAL_INCOMING_ARGS])
|
|||
|
|
|||
|
#define VIRTUAL_INCOMING_ARGS_REGNUM (FIRST_VIRTUAL_REGISTER)
|
|||
|
|
|||
|
/* If FRAME_GROWS_DOWNWARD, this points to immediately above the first
|
|||
|
variable on the stack. Otherwise, it points to the first variable on
|
|||
|
the stack. */
|
|||
|
|
|||
|
#define virtual_stack_vars_rtx (global_rtl[GR_VIRTUAL_STACK_ARGS])
|
|||
|
|
|||
|
#define VIRTUAL_STACK_VARS_REGNUM ((FIRST_VIRTUAL_REGISTER) + 1)
|
|||
|
|
|||
|
/* This points to the location of dynamically-allocated memory on the stack
|
|||
|
immediately after the stack pointer has been adjusted by the amount
|
|||
|
desired. */
|
|||
|
|
|||
|
#define virtual_stack_dynamic_rtx (global_rtl[GR_VIRTUAL_STACK_DYNAMIC])
|
|||
|
|
|||
|
#define VIRTUAL_STACK_DYNAMIC_REGNUM ((FIRST_VIRTUAL_REGISTER) + 2)
|
|||
|
|
|||
|
/* This points to the location in the stack at which outgoing arguments should
|
|||
|
be written when the stack is pre-pushed (arguments pushed using push
|
|||
|
insns always use sp). */
|
|||
|
|
|||
|
#define virtual_outgoing_args_rtx (global_rtl[GR_VIRTUAL_OUTGOING_ARGS])
|
|||
|
|
|||
|
#define VIRTUAL_OUTGOING_ARGS_REGNUM ((FIRST_VIRTUAL_REGISTER) + 3)
|
|||
|
|
|||
|
/* This points to the Canonical Frame Address of the function. This
|
|||
|
should correspond to the CFA produced by INCOMING_FRAME_SP_OFFSET,
|
|||
|
but is calculated relative to the arg pointer for simplicity; the
|
|||
|
frame pointer nor stack pointer are necessarily fixed relative to
|
|||
|
the CFA until after reload. */
|
|||
|
|
|||
|
#define virtual_cfa_rtx (global_rtl[GR_VIRTUAL_CFA])
|
|||
|
|
|||
|
#define VIRTUAL_CFA_REGNUM ((FIRST_VIRTUAL_REGISTER) + 4)
|
|||
|
|
|||
|
#define LAST_VIRTUAL_POINTER_REGISTER ((FIRST_VIRTUAL_REGISTER) + 4)
|
|||
|
|
|||
|
/* This is replaced by crtl->preferred_stack_boundary / BITS_PER_UNIT
|
|||
|
when finalized. */
|
|||
|
|
|||
|
#define virtual_preferred_stack_boundary_rtx \
|
|||
|
(global_rtl[GR_VIRTUAL_PREFERRED_STACK_BOUNDARY])
|
|||
|
|
|||
|
#define VIRTUAL_PREFERRED_STACK_BOUNDARY_REGNUM \
|
|||
|
((FIRST_VIRTUAL_REGISTER) + 5)
|
|||
|
|
|||
|
#define LAST_VIRTUAL_REGISTER ((FIRST_VIRTUAL_REGISTER) + 5)
|
|||
|
|
|||
|
/* Nonzero if REGNUM is a pointer into the stack frame. */
|
|||
|
#define REGNO_PTR_FRAME_P(REGNUM) \
|
|||
|
((REGNUM) == STACK_POINTER_REGNUM \
|
|||
|
|| (REGNUM) == FRAME_POINTER_REGNUM \
|
|||
|
|| (REGNUM) == HARD_FRAME_POINTER_REGNUM \
|
|||
|
|| (REGNUM) == ARG_POINTER_REGNUM \
|
|||
|
|| ((REGNUM) >= FIRST_VIRTUAL_REGISTER \
|
|||
|
&& (REGNUM) <= LAST_VIRTUAL_POINTER_REGISTER))
|
|||
|
|
|||
|
/* REGNUM never really appearing in the INSN stream. */
|
|||
|
#define INVALID_REGNUM (~(unsigned int) 0)
|
|||
|
|
|||
|
/* REGNUM for which no debug information can be generated. */
|
|||
|
#define IGNORED_DWARF_REGNUM (INVALID_REGNUM - 1)
|
|||
|
|
|||
|
extern rtx output_constant_def (tree, int);
|
|||
|
extern rtx lookup_constant_def (tree);
|
|||
|
|
|||
|
/* Nonzero after end of reload pass.
|
|||
|
Set to 1 or 0 by reload1.cc. */
|
|||
|
|
|||
|
extern int reload_completed;
|
|||
|
|
|||
|
/* Nonzero after thread_prologue_and_epilogue_insns has run. */
|
|||
|
extern int epilogue_completed;
|
|||
|
|
|||
|
/* Set to 1 while reload_as_needed is operating.
|
|||
|
Required by some machines to handle any generated moves differently. */
|
|||
|
|
|||
|
extern int reload_in_progress;
|
|||
|
|
|||
|
/* Set to 1 while in lra. */
|
|||
|
extern int lra_in_progress;
|
|||
|
|
|||
|
/* This macro indicates whether you may create a new
|
|||
|
pseudo-register. */
|
|||
|
|
|||
|
#define can_create_pseudo_p() (!reload_in_progress && !reload_completed)
|
|||
|
|
|||
|
#ifdef STACK_REGS
|
|||
|
/* Nonzero after end of regstack pass.
|
|||
|
Set to 1 or 0 by reg-stack.cc. */
|
|||
|
extern int regstack_completed;
|
|||
|
#endif
|
|||
|
|
|||
|
/* If this is nonzero, we do not bother generating VOLATILE
|
|||
|
around volatile memory references, and we are willing to
|
|||
|
output indirect addresses. If cse is to follow, we reject
|
|||
|
indirect addresses so a useful potential cse is generated;
|
|||
|
if it is used only once, instruction combination will produce
|
|||
|
the same indirect address eventually. */
|
|||
|
extern int cse_not_expected;
|
|||
|
|
|||
|
/* Translates rtx code to tree code, for those codes needed by
|
|||
|
real_arithmetic. The function returns an int because the caller may not
|
|||
|
know what `enum tree_code' means. */
|
|||
|
|
|||
|
extern int rtx_to_tree_code (enum rtx_code);
|
|||
|
|
|||
|
/* In cse.cc */
|
|||
|
extern int delete_trivially_dead_insns (rtx_insn *, int);
|
|||
|
extern int exp_equiv_p (const_rtx, const_rtx, int, bool);
|
|||
|
extern unsigned hash_rtx (const_rtx x, machine_mode, int *, int *, bool);
|
|||
|
|
|||
|
/* In dse.cc */
|
|||
|
extern bool check_for_inc_dec (rtx_insn *insn);
|
|||
|
|
|||
|
/* In jump.cc */
|
|||
|
extern int comparison_dominates_p (enum rtx_code, enum rtx_code);
|
|||
|
extern bool jump_to_label_p (const rtx_insn *);
|
|||
|
extern int condjump_p (const rtx_insn *);
|
|||
|
extern int any_condjump_p (const rtx_insn *);
|
|||
|
extern int any_uncondjump_p (const rtx_insn *);
|
|||
|
extern rtx pc_set (const rtx_insn *);
|
|||
|
extern rtx condjump_label (const rtx_insn *);
|
|||
|
extern int simplejump_p (const rtx_insn *);
|
|||
|
extern int returnjump_p (const rtx_insn *);
|
|||
|
extern int eh_returnjump_p (rtx_insn *);
|
|||
|
extern int onlyjump_p (const rtx_insn *);
|
|||
|
extern int invert_jump_1 (rtx_jump_insn *, rtx);
|
|||
|
extern int invert_jump (rtx_jump_insn *, rtx, int);
|
|||
|
extern int rtx_renumbered_equal_p (const_rtx, const_rtx);
|
|||
|
extern int true_regnum (const_rtx);
|
|||
|
extern unsigned int reg_or_subregno (const_rtx);
|
|||
|
extern int redirect_jump_1 (rtx_insn *, rtx);
|
|||
|
extern void redirect_jump_2 (rtx_jump_insn *, rtx, rtx, int, int);
|
|||
|
extern int redirect_jump (rtx_jump_insn *, rtx, int);
|
|||
|
extern void rebuild_jump_labels (rtx_insn *);
|
|||
|
extern void rebuild_jump_labels_chain (rtx_insn *);
|
|||
|
extern rtx reversed_comparison (const_rtx, machine_mode);
|
|||
|
extern enum rtx_code reversed_comparison_code (const_rtx, const rtx_insn *);
|
|||
|
extern enum rtx_code reversed_comparison_code_parts (enum rtx_code, const_rtx,
|
|||
|
const_rtx, const rtx_insn *);
|
|||
|
extern void delete_for_peephole (rtx_insn *, rtx_insn *);
|
|||
|
extern int condjump_in_parallel_p (const rtx_insn *);
|
|||
|
|
|||
|
/* In emit-rtl.cc. */
|
|||
|
extern int max_reg_num (void);
|
|||
|
extern int max_label_num (void);
|
|||
|
extern int get_first_label_num (void);
|
|||
|
extern void maybe_set_first_label_num (rtx_code_label *);
|
|||
|
extern void delete_insns_since (rtx_insn *);
|
|||
|
extern void mark_reg_pointer (rtx, int);
|
|||
|
extern void mark_user_reg (rtx);
|
|||
|
extern void reset_used_flags (rtx);
|
|||
|
extern void set_used_flags (rtx);
|
|||
|
extern void reorder_insns (rtx_insn *, rtx_insn *, rtx_insn *);
|
|||
|
extern void reorder_insns_nobb (rtx_insn *, rtx_insn *, rtx_insn *);
|
|||
|
extern int get_max_insn_count (void);
|
|||
|
extern int in_sequence_p (void);
|
|||
|
extern void init_emit (void);
|
|||
|
extern void init_emit_regs (void);
|
|||
|
extern void init_derived_machine_modes (void);
|
|||
|
extern void init_emit_once (void);
|
|||
|
extern void push_topmost_sequence (void);
|
|||
|
extern void pop_topmost_sequence (void);
|
|||
|
extern void set_new_first_and_last_insn (rtx_insn *, rtx_insn *);
|
|||
|
extern unsigned int unshare_all_rtl (void);
|
|||
|
extern void unshare_all_rtl_again (rtx_insn *);
|
|||
|
extern void unshare_all_rtl_in_chain (rtx_insn *);
|
|||
|
extern void verify_rtl_sharing (void);
|
|||
|
extern void add_insn (rtx_insn *);
|
|||
|
extern void add_insn_before (rtx_insn *, rtx_insn *, basic_block);
|
|||
|
extern void add_insn_after (rtx_insn *, rtx_insn *, basic_block);
|
|||
|
extern void remove_insn (rtx_insn *);
|
|||
|
extern rtx_insn *emit (rtx, bool = true);
|
|||
|
extern void emit_insn_at_entry (rtx);
|
|||
|
extern rtx gen_lowpart_SUBREG (machine_mode, rtx);
|
|||
|
extern rtx gen_const_mem (machine_mode, rtx);
|
|||
|
extern rtx gen_frame_mem (machine_mode, rtx);
|
|||
|
extern rtx gen_tmp_stack_mem (machine_mode, rtx);
|
|||
|
extern bool validate_subreg (machine_mode, machine_mode,
|
|||
|
const_rtx, poly_uint64);
|
|||
|
|
|||
|
/* In combine.cc */
|
|||
|
extern unsigned int extended_count (const_rtx, machine_mode, int);
|
|||
|
extern rtx remove_death (unsigned int, rtx_insn *);
|
|||
|
extern void dump_combine_stats (FILE *);
|
|||
|
extern void dump_combine_total_stats (FILE *);
|
|||
|
extern rtx make_compound_operation (rtx, enum rtx_code);
|
|||
|
|
|||
|
/* In sched-rgn.cc. */
|
|||
|
extern void schedule_insns (void);
|
|||
|
|
|||
|
/* In sched-ebb.cc. */
|
|||
|
extern void schedule_ebbs (void);
|
|||
|
|
|||
|
/* In sel-sched-dump.cc. */
|
|||
|
extern void sel_sched_fix_param (const char *param, const char *val);
|
|||
|
|
|||
|
/* In print-rtl.cc */
|
|||
|
extern const char *print_rtx_head;
|
|||
|
extern void debug (const rtx_def &ref);
|
|||
|
extern void debug (const rtx_def *ptr);
|
|||
|
extern void debug_rtx (const_rtx);
|
|||
|
extern void debug_rtx_list (const rtx_insn *, int);
|
|||
|
extern void debug_rtx_range (const rtx_insn *, const rtx_insn *);
|
|||
|
extern const rtx_insn *debug_rtx_find (const rtx_insn *, int);
|
|||
|
extern void print_mem_expr (FILE *, const_tree);
|
|||
|
extern void print_rtl (FILE *, const_rtx);
|
|||
|
extern void print_simple_rtl (FILE *, const_rtx);
|
|||
|
extern int print_rtl_single (FILE *, const_rtx);
|
|||
|
extern int print_rtl_single_with_indent (FILE *, const_rtx, int);
|
|||
|
extern void print_inline_rtx (FILE *, const_rtx, int);
|
|||
|
|
|||
|
/* In stmt.cc */
|
|||
|
extern void expand_null_return (void);
|
|||
|
extern void expand_naked_return (void);
|
|||
|
extern void emit_jump (rtx);
|
|||
|
|
|||
|
/* Memory operation built-ins differ by return value. Mapping
|
|||
|
of the enum values is following:
|
|||
|
- RETURN_BEGIN - return destination, e.g. memcpy
|
|||
|
- RETURN_END - return destination + n, e.g. mempcpy
|
|||
|
- RETURN_END_MINUS_ONE - return a pointer to the terminating
|
|||
|
null byte of the string, e.g. strcpy
|
|||
|
*/
|
|||
|
|
|||
|
enum memop_ret
|
|||
|
{
|
|||
|
RETURN_BEGIN,
|
|||
|
RETURN_END,
|
|||
|
RETURN_END_MINUS_ONE
|
|||
|
};
|
|||
|
|
|||
|
/* In expr.cc */
|
|||
|
extern rtx move_by_pieces (rtx, rtx, unsigned HOST_WIDE_INT,
|
|||
|
unsigned int, memop_ret);
|
|||
|
extern poly_int64 find_args_size_adjust (rtx_insn *);
|
|||
|
extern poly_int64 fixup_args_size_notes (rtx_insn *, rtx_insn *, poly_int64);
|
|||
|
|
|||
|
/* In expmed.cc */
|
|||
|
extern void init_expmed (void);
|
|||
|
extern void expand_inc (rtx, rtx);
|
|||
|
extern void expand_dec (rtx, rtx);
|
|||
|
|
|||
|
/* In lower-subreg.cc */
|
|||
|
extern void init_lower_subreg (void);
|
|||
|
|
|||
|
/* In gcse.cc */
|
|||
|
extern bool can_copy_p (machine_mode);
|
|||
|
extern bool can_assign_to_reg_without_clobbers_p (rtx, machine_mode);
|
|||
|
extern rtx_insn *prepare_copy_insn (rtx, rtx);
|
|||
|
|
|||
|
/* In cprop.cc */
|
|||
|
extern rtx fis_get_condition (rtx_insn *);
|
|||
|
|
|||
|
/* In ira.cc */
|
|||
|
extern HARD_REG_SET eliminable_regset;
|
|||
|
extern void mark_elimination (int, int);
|
|||
|
|
|||
|
/* In reginfo.cc */
|
|||
|
extern int reg_classes_intersect_p (reg_class_t, reg_class_t);
|
|||
|
extern int reg_class_subset_p (reg_class_t, reg_class_t);
|
|||
|
extern void globalize_reg (tree, int);
|
|||
|
extern void init_reg_modes_target (void);
|
|||
|
extern void init_regs (void);
|
|||
|
extern void reinit_regs (void);
|
|||
|
extern void init_fake_stack_mems (void);
|
|||
|
extern void save_register_info (void);
|
|||
|
extern void init_reg_sets (void);
|
|||
|
extern void regclass (rtx, int);
|
|||
|
extern void reg_scan (rtx_insn *, unsigned int);
|
|||
|
extern void fix_register (const char *, int, int);
|
|||
|
extern const HARD_REG_SET *valid_mode_changes_for_regno (unsigned int);
|
|||
|
|
|||
|
/* In reload1.cc */
|
|||
|
extern int function_invariant_p (const_rtx);
|
|||
|
|
|||
|
/* In calls.cc */
|
|||
|
enum libcall_type
|
|||
|
{
|
|||
|
LCT_NORMAL = 0,
|
|||
|
LCT_CONST = 1,
|
|||
|
LCT_PURE = 2,
|
|||
|
LCT_NORETURN = 3,
|
|||
|
LCT_THROW = 4,
|
|||
|
LCT_RETURNS_TWICE = 5
|
|||
|
};
|
|||
|
|
|||
|
extern rtx emit_library_call_value_1 (int, rtx, rtx, enum libcall_type,
|
|||
|
machine_mode, int, rtx_mode_t *);
|
|||
|
|
|||
|
/* Output a library call and discard the returned value. FUN is the
|
|||
|
address of the function, as a SYMBOL_REF rtx, and OUTMODE is the mode
|
|||
|
of the (discarded) return value. FN_TYPE is LCT_NORMAL for `normal'
|
|||
|
calls, LCT_CONST for `const' calls, LCT_PURE for `pure' calls, or
|
|||
|
another LCT_ value for other types of library calls.
|
|||
|
|
|||
|
There are different overloads of this function for different numbers
|
|||
|
of arguments. In each case the argument value is followed by its mode. */
|
|||
|
|
|||
|
inline void
|
|||
|
emit_library_call (rtx fun, libcall_type fn_type, machine_mode outmode)
|
|||
|
{
|
|||
|
emit_library_call_value_1 (0, fun, NULL_RTX, fn_type, outmode, 0, NULL);
|
|||
|
}
|
|||
|
|
|||
|
inline void
|
|||
|
emit_library_call (rtx fun, libcall_type fn_type, machine_mode outmode,
|
|||
|
rtx arg1, machine_mode arg1_mode)
|
|||
|
{
|
|||
|
rtx_mode_t args[] = { rtx_mode_t (arg1, arg1_mode) };
|
|||
|
emit_library_call_value_1 (0, fun, NULL_RTX, fn_type, outmode, 1, args);
|
|||
|
}
|
|||
|
|
|||
|
inline void
|
|||
|
emit_library_call (rtx fun, libcall_type fn_type, machine_mode outmode,
|
|||
|
rtx arg1, machine_mode arg1_mode,
|
|||
|
rtx arg2, machine_mode arg2_mode)
|
|||
|
{
|
|||
|
rtx_mode_t args[] = {
|
|||
|
rtx_mode_t (arg1, arg1_mode),
|
|||
|
rtx_mode_t (arg2, arg2_mode)
|
|||
|
};
|
|||
|
emit_library_call_value_1 (0, fun, NULL_RTX, fn_type, outmode, 2, args);
|
|||
|
}
|
|||
|
|
|||
|
inline void
|
|||
|
emit_library_call (rtx fun, libcall_type fn_type, machine_mode outmode,
|
|||
|
rtx arg1, machine_mode arg1_mode,
|
|||
|
rtx arg2, machine_mode arg2_mode,
|
|||
|
rtx arg3, machine_mode arg3_mode)
|
|||
|
{
|
|||
|
rtx_mode_t args[] = {
|
|||
|
rtx_mode_t (arg1, arg1_mode),
|
|||
|
rtx_mode_t (arg2, arg2_mode),
|
|||
|
rtx_mode_t (arg3, arg3_mode)
|
|||
|
};
|
|||
|
emit_library_call_value_1 (0, fun, NULL_RTX, fn_type, outmode, 3, args);
|
|||
|
}
|
|||
|
|
|||
|
inline void
|
|||
|
emit_library_call (rtx fun, libcall_type fn_type, machine_mode outmode,
|
|||
|
rtx arg1, machine_mode arg1_mode,
|
|||
|
rtx arg2, machine_mode arg2_mode,
|
|||
|
rtx arg3, machine_mode arg3_mode,
|
|||
|
rtx arg4, machine_mode arg4_mode)
|
|||
|
{
|
|||
|
rtx_mode_t args[] = {
|
|||
|
rtx_mode_t (arg1, arg1_mode),
|
|||
|
rtx_mode_t (arg2, arg2_mode),
|
|||
|
rtx_mode_t (arg3, arg3_mode),
|
|||
|
rtx_mode_t (arg4, arg4_mode)
|
|||
|
};
|
|||
|
emit_library_call_value_1 (0, fun, NULL_RTX, fn_type, outmode, 4, args);
|
|||
|
}
|
|||
|
|
|||
|
/* Like emit_library_call, but return the value produced by the call.
|
|||
|
Use VALUE to store the result if it is nonnull, otherwise pick a
|
|||
|
convenient location. */
|
|||
|
|
|||
|
inline rtx
|
|||
|
emit_library_call_value (rtx fun, rtx value, libcall_type fn_type,
|
|||
|
machine_mode outmode)
|
|||
|
{
|
|||
|
return emit_library_call_value_1 (1, fun, value, fn_type, outmode, 0, NULL);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx
|
|||
|
emit_library_call_value (rtx fun, rtx value, libcall_type fn_type,
|
|||
|
machine_mode outmode,
|
|||
|
rtx arg1, machine_mode arg1_mode)
|
|||
|
{
|
|||
|
rtx_mode_t args[] = { rtx_mode_t (arg1, arg1_mode) };
|
|||
|
return emit_library_call_value_1 (1, fun, value, fn_type, outmode, 1, args);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx
|
|||
|
emit_library_call_value (rtx fun, rtx value, libcall_type fn_type,
|
|||
|
machine_mode outmode,
|
|||
|
rtx arg1, machine_mode arg1_mode,
|
|||
|
rtx arg2, machine_mode arg2_mode)
|
|||
|
{
|
|||
|
rtx_mode_t args[] = {
|
|||
|
rtx_mode_t (arg1, arg1_mode),
|
|||
|
rtx_mode_t (arg2, arg2_mode)
|
|||
|
};
|
|||
|
return emit_library_call_value_1 (1, fun, value, fn_type, outmode, 2, args);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx
|
|||
|
emit_library_call_value (rtx fun, rtx value, libcall_type fn_type,
|
|||
|
machine_mode outmode,
|
|||
|
rtx arg1, machine_mode arg1_mode,
|
|||
|
rtx arg2, machine_mode arg2_mode,
|
|||
|
rtx arg3, machine_mode arg3_mode)
|
|||
|
{
|
|||
|
rtx_mode_t args[] = {
|
|||
|
rtx_mode_t (arg1, arg1_mode),
|
|||
|
rtx_mode_t (arg2, arg2_mode),
|
|||
|
rtx_mode_t (arg3, arg3_mode)
|
|||
|
};
|
|||
|
return emit_library_call_value_1 (1, fun, value, fn_type, outmode, 3, args);
|
|||
|
}
|
|||
|
|
|||
|
inline rtx
|
|||
|
emit_library_call_value (rtx fun, rtx value, libcall_type fn_type,
|
|||
|
machine_mode outmode,
|
|||
|
rtx arg1, machine_mode arg1_mode,
|
|||
|
rtx arg2, machine_mode arg2_mode,
|
|||
|
rtx arg3, machine_mode arg3_mode,
|
|||
|
rtx arg4, machine_mode arg4_mode)
|
|||
|
{
|
|||
|
rtx_mode_t args[] = {
|
|||
|
rtx_mode_t (arg1, arg1_mode),
|
|||
|
rtx_mode_t (arg2, arg2_mode),
|
|||
|
rtx_mode_t (arg3, arg3_mode),
|
|||
|
rtx_mode_t (arg4, arg4_mode)
|
|||
|
};
|
|||
|
return emit_library_call_value_1 (1, fun, value, fn_type, outmode, 4, args);
|
|||
|
}
|
|||
|
|
|||
|
/* In varasm.cc */
|
|||
|
extern void init_varasm_once (void);
|
|||
|
|
|||
|
extern rtx make_debug_expr_from_rtl (const_rtx);
|
|||
|
|
|||
|
/* In read-rtl.cc */
|
|||
|
#ifdef GENERATOR_FILE
|
|||
|
extern bool read_rtx (const char *, vec<rtx> *);
|
|||
|
#endif
|
|||
|
|
|||
|
/* In alias.cc */
|
|||
|
extern rtx canon_rtx (rtx);
|
|||
|
extern int true_dependence (const_rtx, machine_mode, const_rtx);
|
|||
|
extern rtx get_addr (rtx);
|
|||
|
extern int canon_true_dependence (const_rtx, machine_mode, rtx,
|
|||
|
const_rtx, rtx);
|
|||
|
extern int read_dependence (const_rtx, const_rtx);
|
|||
|
extern int anti_dependence (const_rtx, const_rtx);
|
|||
|
extern int canon_anti_dependence (const_rtx, bool,
|
|||
|
const_rtx, machine_mode, rtx);
|
|||
|
extern int output_dependence (const_rtx, const_rtx);
|
|||
|
extern int canon_output_dependence (const_rtx, bool,
|
|||
|
const_rtx, machine_mode, rtx);
|
|||
|
extern int may_alias_p (const_rtx, const_rtx);
|
|||
|
extern void init_alias_target (void);
|
|||
|
extern void init_alias_analysis (void);
|
|||
|
extern void end_alias_analysis (void);
|
|||
|
extern void vt_equate_reg_base_value (const_rtx, const_rtx);
|
|||
|
extern bool memory_modified_in_insn_p (const_rtx, const_rtx);
|
|||
|
extern bool may_be_sp_based_p (rtx);
|
|||
|
extern rtx gen_hard_reg_clobber (machine_mode, unsigned int);
|
|||
|
extern rtx get_reg_known_value (unsigned int);
|
|||
|
extern bool get_reg_known_equiv_p (unsigned int);
|
|||
|
extern rtx get_reg_base_value (unsigned int);
|
|||
|
extern rtx extract_mem_from_operand (rtx);
|
|||
|
|
|||
|
#ifdef STACK_REGS
|
|||
|
extern int stack_regs_mentioned (const_rtx insn);
|
|||
|
#endif
|
|||
|
|
|||
|
/* In toplev.cc */
|
|||
|
extern GTY(()) rtx stack_limit_rtx;
|
|||
|
|
|||
|
/* In var-tracking.cc */
|
|||
|
extern unsigned int variable_tracking_main (void);
|
|||
|
extern void delete_vta_debug_insns (bool);
|
|||
|
|
|||
|
/* In stor-layout.cc. */
|
|||
|
extern void get_mode_bounds (scalar_int_mode, int,
|
|||
|
scalar_int_mode, rtx *, rtx *);
|
|||
|
|
|||
|
/* In loop-iv.cc */
|
|||
|
extern rtx canon_condition (rtx);
|
|||
|
extern void simplify_using_condition (rtx, rtx *, bitmap);
|
|||
|
|
|||
|
/* In final.cc */
|
|||
|
extern unsigned int compute_alignments (void);
|
|||
|
extern void update_alignments (vec<rtx> &);
|
|||
|
extern int asm_str_count (const char *templ);
|
|||
|
|
|||
|
struct rtl_hooks
|
|||
|
{
|
|||
|
rtx (*gen_lowpart) (machine_mode, rtx);
|
|||
|
rtx (*gen_lowpart_no_emit) (machine_mode, rtx);
|
|||
|
rtx (*reg_nonzero_bits) (const_rtx, scalar_int_mode, scalar_int_mode,
|
|||
|
unsigned HOST_WIDE_INT *);
|
|||
|
rtx (*reg_num_sign_bit_copies) (const_rtx, scalar_int_mode, scalar_int_mode,
|
|||
|
unsigned int *);
|
|||
|
bool (*reg_truncated_to_mode) (machine_mode, const_rtx);
|
|||
|
|
|||
|
/* Whenever you add entries here, make sure you adjust rtlhooks-def.h. */
|
|||
|
};
|
|||
|
|
|||
|
/* Each pass can provide its own. */
|
|||
|
extern struct rtl_hooks rtl_hooks;
|
|||
|
|
|||
|
/* ... but then it has to restore these. */
|
|||
|
extern const struct rtl_hooks general_rtl_hooks;
|
|||
|
|
|||
|
/* Keep this for the nonce. */
|
|||
|
#define gen_lowpart rtl_hooks.gen_lowpart
|
|||
|
|
|||
|
extern void insn_locations_init (void);
|
|||
|
extern void insn_locations_finalize (void);
|
|||
|
extern void set_curr_insn_location (location_t);
|
|||
|
extern location_t curr_insn_location (void);
|
|||
|
extern void set_insn_locations (rtx_insn *, location_t);
|
|||
|
|
|||
|
/* rtl-error.cc */
|
|||
|
extern void _fatal_insn_not_found (const_rtx, const char *, int, const char *)
|
|||
|
ATTRIBUTE_NORETURN ATTRIBUTE_COLD;
|
|||
|
extern void _fatal_insn (const char *, const_rtx, const char *, int, const char *)
|
|||
|
ATTRIBUTE_NORETURN ATTRIBUTE_COLD;
|
|||
|
|
|||
|
#define fatal_insn(msgid, insn) \
|
|||
|
_fatal_insn (msgid, insn, __FILE__, __LINE__, __FUNCTION__)
|
|||
|
#define fatal_insn_not_found(insn) \
|
|||
|
_fatal_insn_not_found (insn, __FILE__, __LINE__, __FUNCTION__)
|
|||
|
|
|||
|
/* reginfo.cc */
|
|||
|
extern tree GTY(()) global_regs_decl[FIRST_PSEUDO_REGISTER];
|
|||
|
|
|||
|
/* Information about the function that is propagated by the RTL backend.
|
|||
|
Available only for functions that has been already assembled. */
|
|||
|
|
|||
|
struct GTY(()) cgraph_rtl_info {
|
|||
|
unsigned int preferred_incoming_stack_boundary;
|
|||
|
|
|||
|
/* Which registers the function clobbers, either directly or by
|
|||
|
calling another function. */
|
|||
|
HARD_REG_SET function_used_regs;
|
|||
|
};
|
|||
|
|
|||
|
/* If loads from memories of mode MODE always sign or zero extend,
|
|||
|
return SIGN_EXTEND or ZERO_EXTEND as appropriate. Return UNKNOWN
|
|||
|
otherwise. */
|
|||
|
|
|||
|
inline rtx_code
|
|||
|
load_extend_op (machine_mode mode)
|
|||
|
{
|
|||
|
scalar_int_mode int_mode;
|
|||
|
if (is_a <scalar_int_mode> (mode, &int_mode)
|
|||
|
&& GET_MODE_PRECISION (int_mode) < BITS_PER_WORD)
|
|||
|
return LOAD_EXTEND_OP (int_mode);
|
|||
|
return UNKNOWN;
|
|||
|
}
|
|||
|
|
|||
|
/* If X is a PLUS of a base and a constant offset, add the constant to *OFFSET
|
|||
|
and return the base. Return X otherwise. */
|
|||
|
|
|||
|
inline rtx
|
|||
|
strip_offset_and_add (rtx x, poly_int64_pod *offset)
|
|||
|
{
|
|||
|
if (GET_CODE (x) == PLUS)
|
|||
|
{
|
|||
|
poly_int64 suboffset;
|
|||
|
x = strip_offset (x, &suboffset);
|
|||
|
*offset = poly_uint64 (*offset) + suboffset;
|
|||
|
}
|
|||
|
return x;
|
|||
|
}
|
|||
|
|
|||
|
/* Return true if X is an operation that always operates on the full
|
|||
|
registers for WORD_REGISTER_OPERATIONS architectures. */
|
|||
|
|
|||
|
inline bool
|
|||
|
word_register_operation_p (const_rtx x)
|
|||
|
{
|
|||
|
switch (GET_CODE (x))
|
|||
|
{
|
|||
|
case CONST_INT:
|
|||
|
case ROTATE:
|
|||
|
case ROTATERT:
|
|||
|
case SIGN_EXTRACT:
|
|||
|
case ZERO_EXTRACT:
|
|||
|
return false;
|
|||
|
|
|||
|
default:
|
|||
|
return true;
|
|||
|
}
|
|||
|
}
|
|||
|
|
|||
|
/* Holds an rtx comparison to simplify passing many parameters pertaining to a
|
|||
|
single comparison. */
|
|||
|
|
|||
|
struct rtx_comparison {
|
|||
|
rtx_code code;
|
|||
|
rtx op0, op1;
|
|||
|
machine_mode mode;
|
|||
|
};
|
|||
|
|
|||
|
/* gtype-desc.cc. */
|
|||
|
extern void gt_ggc_mx (rtx &);
|
|||
|
extern void gt_pch_nx (rtx &);
|
|||
|
extern void gt_pch_nx (rtx &, gt_pointer_operator, void *);
|
|||
|
|
|||
|
#endif /* ! GCC_RTL_H */
|